A multi-model-based deep learning framework for short text multiclass classification with the imbalanced and extremely small data set

  title={A multi-model-based deep learning framework for short text multiclass classification with the imbalanced and extremely small data set},
  author={Jiajun Tong and Zhixi Wang and Xiaobin Rui},
: Text classification plays an important role in many practical applications. In the real world, there are extremely small datasets. Most existing methods adopt pre-trained neural network models to handle this kind of dataset. However, these methods are either difficult to deploy on mobile devices because of their large output size or cannot fully extract the deep semantic information between phrases and clauses. This paper proposes a multimodel-based deep learning framework for short-text… 

Figures and Tables from this paper



A Fine-Tuned BERT-Based Transfer Learning Approach for Text Classification

The proposed work applied transfer learning classification models on both fake news and extremist-non-extremist datasets to check the performance of transfer learning models.

Prompt-Learning for Short Text Classification

This paper proposes a simple short text classification approach that makes use of prompt-learning based on knowledgeable expansion that outperforms the state-of-the-art by up to 6 Accuracy points on three well-known datasets.

Utilizing BERT Intermediate Layers for Aspect Based Sentiment Analysis and Natural Language Inference

The potential of utilizing BERT intermediate layers to enhance the performance of fine-tuning of BERT and applying this approach to a natural language inference task demonstrates the effectiveness and generality of the proposed approach.

DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter

This work proposes a method to pre-train a smaller general-purpose language representation model, called DistilBERT, which can be fine-tuned with good performances on a wide range of tasks like its larger counterparts, and introduces a triple loss combining language modeling, distillation and cosine-distance losses.

Text Classification Algorithms: A Survey

An overview of text classification algorithms is discussed, which covers different text feature extractions, dimensionality reduction methods, existing algorithms and techniques, and evaluations methods.

Efficient, Uncertainty-based Moderation of Neural Networks Text Classifiers

A semi-automated approach that uses prediction uncertainties to pass unconfident, probably incorrect classifications to human moderators to minimize the workload, and can improve the classification F1-scores of modern Neural Networks classifiers.

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

A new language representation model, BERT, designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.

Chinese Language and Literature Online Resource Classification Algorithm Based on Improved SVM

Improved SVM proposed in this paper greatly improves the efficiency of classification and facilitates the rapid search of information with high matching degree among various online resources of Chinese language and literature.

An ensemble based approach using a combination of clustering and classification algorithms to enhance customer churn prediction in telecom industry

The proposed churn prediction model is a hybrid model that is based on a combination of clustering and classification algorithms using an ensemble that performed significantly better than state of the art churn prediction models.

Ensemble learning: A survey

The concept of ensemble learning is introduced, traditional, novel and state‐of‐the‐art ensemble methods are reviewed and current challenges and trends in the field are discussed.