Complaint Classification using Word2Vec Model

@article{Rathore2018ComplaintCU,
  title={Complaint Classification using Word2Vec Model},
  author={M. Rathore and D. Gupta and Dinabandhu Bhandari},
  journal={International journal of engineering and technology},
  year={2018},
  volume={7},
  pages={402}
}
Attempt has been made to develop a versatile, universal complaint grievance segregator by classifying orally acknowledged grievances into one of the predefined categories. The oral complaints are first converted to text and then each word is represented by a vector using word2vec. Each grievance is represented by a single vector using Gated Recurrent Unit (GRU) that implements the hidden state of Recurrent Neural Network (RNN) model. The popular Multi-Layer Perceptron (MLP) has been used as the… Expand

Figures from this paper

Critical Dimension of Word2Vec
TLDR
This work presents an empirical approach to decide the dimension of the word embeddings for a specific set of documents (corpora) to a critical value such that the representation of the words still preserve their original semantic and syntactic meanings. Expand
Powered Hawkes-Dirichlet Process: Challenging Textual Clustering using a Flexible Temporal Prior
  • Gael Poux-M'edard, Julien Velcin, Sabine Loudcher
  • Computer Science
  • 2021
The textual content of a document and its publication date are intertwined. For example, the publication of a news article on a topic is influenced by previous publications on similar issues,Expand

References

SHOWING 1-10 OF 16 REFERENCES
Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling
TLDR
These advanced recurrent units that implement a gating mechanism, such as a long short-term memory (LSTM) unit and a recently proposed gated recurrent unit (GRU), are found to be comparable to LSTM. Expand
Fast Semantic Extraction Using a Novel Neural Network Architecture
TLDR
A novel neural network architecture for the problem of semantic role labeling that learns a direct mapping from source sentence to semantic tags for a given predicate without the aid of a parser or a chunker. Expand
Parametric Representation of Paragraphs and Their Classification
TLDR
A methodology has been proposed based on multi layer perceptron in designing an automatic paragraph classifier and the proposed framework has been tested on large industrial data and found improved performance compare to conventional rule based approach. Expand
Machine learning in automated text categorization
TLDR
This survey discusses the main approaches to text categorization that fall within the machine learning paradigm and discusses in detail issues pertaining to three different problems, namely, document representation, classifier construction, and classifier evaluation. Expand
Efficient Estimation of Word Representations in Vector Space
TLDR
Two novel model architectures for computing continuous vector representations of words from very large data sets are proposed and it is shown that these vectors provide state-of-the-art performance on the authors' test set for measuring syntactic and semantic word similarities. Expand
Bidirectional recurrent neural networks
TLDR
It is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. Expand
Linguistic Regularities in Continuous Space Word Representations
TLDR
The vector-space word representations that are implicitly learned by the input-layer weights are found to be surprisingly good at capturing syntactic and semantic regularities in language, and that each relationship is characterized by a relation-specific vector offset. Expand
Distributed Representations of Words and Phrases and their Compositionality
TLDR
This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling. Expand
Neural Machine Translation by Jointly Learning to Align and Translate
TLDR
It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly. Expand
Adam: A Method for Stochastic Optimization
TLDR
This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework. Expand
...
1
2
...