A Trainable Multiplication Layer for Auto-correlation and Co-occurrence Extraction

  title={A Trainable Multiplication Layer for Auto-correlation and Co-occurrence Extraction},
  author={Hideaki Hayashi and Seiichi Uchida},
In this paper, we propose a trainable multiplication layer (TML) for a neural network that can be used to calculate the multiplication between the input features. Taking an image as an input, the TML raises each pixel value to the power of a weight and then multiplies them, thereby extracting the higher-order local auto-correlation from the input image. The TML can also be used to extract co-occurrence from the feature map of a convolutional network. The training of the TML is formulated based… Expand
Page Segmentation using a Convolutional Neural Network with Trainable Co-Occurrence Features
A method for page segmentation using a CNN with trainable multiplication layers (TMLs) specialized for extracting co-occurrences from feature maps, thereby supporting the detection of objects with similar textures and periodicities is proposed. Expand


Trainable Co-Occurrence Activation Unit for Improving Convnet
  • Takumi Kobayashi
  • Computer Science
  • 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
  • 2018
A co-occurrence activation unit to work across feature channels by extending the element-wise activation function and is trainable by a gradient-based optimization through back-propagation learning and exploits the co- Occurrence relationships among the feature channels. Expand
Deep Co-occurrence Feature Learning for Visual Object Recognition
A new network layer is introduced that can extend a convolutional layer to encode the co-occurrence between the visual parts detected by the numerous neurons, instead of a few pre-specified parts, and is end-to-end trainable. Expand
Very Deep Convolutional Networks for Large-Scale Image Recognition
This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers. Expand
Network In Network
With enhanced local modeling via the micro network, the proposed deep network structure NIN is able to utilize global average pooling over feature maps in the classification layer, which is easier to interpret and less prone to overfitting than traditional fully connected layers. Expand
Learning to Answer Questions from Image Using Convolutional Neural Network
The proposed CNN provides an end-to-end framework with convolutional architectures for learning not only the image and question representations, but also their inter-modal interactions to produce the answer. Expand
Gradient-based learning applied to document recognition
This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task, and Convolutional neural networks are shown to outperform all other techniques. Expand
G2DeNet: Global Gaussian Distribution Embedding Network and Its Application to Visual Recognition
  • Qilong Wang, P. Li, Lei Zhang
  • Mathematics, Computer Science
  • 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
  • 2017
Experimental results on large scale region classification and fine-grained recognition tasks show that G2DeNet is superior to its counterparts, capable of achieving state-of-the-art performance. Expand
Learning Multiple Layers of Features from Tiny Images
It is shown how to train a multi-layer generative model that learns to extract meaningful features which resemble those found in the human visual cortex, using a novel parallelization algorithm to distribute the work among multiple machines connected on a network. Expand
Deep Learning-Based Document Modeling for Personality Detection from Text
This article presents a deep learning based method for determining the author's personality type from text: given a text, the presence or absence of the Big Five traits is detected in theAuthor's psychological profile, and the implementation is freely available for research purposes. Expand
A recurrent log-linearized Gaussian mixture network
A novel neural network, which is named "a recurrent log-linearized Gaussian mixture network (R-LLGMN)," is proposed in this paper for classification of time series, based on a hidden Markov model. Expand