February 25, 2020

1011 words 5 mins read

ShawnyXiao/TextClassification-Keras

ShawnyXiao/TextClassification-Keras

Text classification models implemented in Keras, including: FastText, TextCNN, TextRNN, TextBiRNN, TextAttBiRNN, HAN, RCNN, RCNNVariant, etc.

repo name ShawnyXiao/TextClassification-Keras
repo link https://github.com/ShawnyXiao/TextClassification-Keras
homepage
language Python
size (curr.) 1970 kB
stars (curr.) 428
created 2019-01-03
license MIT License

TextClassification-Keras

This code repository implements a variety of deep learning models for text classification using the Keras framework, which includes: FastText, TextCNN, TextRNN, TextBiRNN, TextAttBiRNN, HAN, RCNN, RCNNVariant, etc. In addition to the model implementation, a simplified application is included.

Guidance

  1. Environment
  2. Usage
  3. Model
    1. FastText
    2. TextCNN
    3. TextRNN
    4. TextBiRNN
    5. TextAttBiRNN
    6. HAN
    7. RCNN
    8. RCNNVariant
    9. To Be Continued…
  4. Reference

Environment

  • Python 3.7
  • NumPy 1.17.2
  • Tensorflow 2.0.1

Usage

All codes are located in the directory /model, and each kind of model has a corresponding directory in which the model and application are placed.

For example, the model and application of FastText are located under /model/FastText, the model part is fast_text.py, and the application part is main.py.

Model

1 FastText

FastText was proposed in the paper Bag of Tricks for Efficient Text Classification.

1.1 Description in Paper

  1. Using a look-up table, bags of ngram covert to word representations.
  2. Word representations are averaged into a text representation, which is a hidden variable.
  3. Text representation is in turn fed to a linear classifier.
  4. Use the softmax function to compute the probability distribution over the predefined classes.

1.2 Implementation Here

Network structure of FastText:

2 TextCNN

TextCNN was proposed in the paper Convolutional Neural Networks for Sentence Classification.

2.1 Description in Paper

  1. Represent sentence with static and non-static channels.
  2. Convolve with multiple filter widths and feature maps.
  3. Use max-over-time pooling.
  4. Use fully connected layer with dropout and softmax ouput.

2.2 Implementation Here

Network structure of TextCNN:

3 TextRNN

TextRNN has been mentioned in the paper Recurrent Neural Network for Text Classification with Multi-Task Learning.

3.1 Description in Paper

3.2 Implementation Here

Network structure of TextRNN:

4 TextBiRNN

TextBiRNN is an improved model based on TextRNN. It improves the RNN layer in the network structure into a bidirectional RNN layer. It is hoped that not only the forward encoding information but also the reverse encoding information can be considered. No related papers have been found yet.

Network structure of TextBiRNN:

5 TextAttBiRNN

TextAttBiRNN is an improved model which introduces attention mechanism based on TextBiRNN. For the representation vectors obtained by bidirectional RNN encoder, the model can focus on the information most relevant to decision making through the attention mechanism. The attention mechanism was first proposed in the paper Neural Machine Translation by Jointly Learning to Align and Translate, and the implementation of the attention mechanism here is referred to this paper Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems.

5.1 Description in Paper

In the paper Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems, the feed forward attention is simplified as follows,

Function a, a learnable function, is recognized as a feed forward network. In this formulation, attention can be seen as producing a fixed-length embedding c of the input sequence by computing an adaptive weighted average of the state sequence h.

5.2 Implementation Here

The implementation of attention is not described here, please refer to the source code directly.

Network structure of TextAttBiRNN:

6 HAN

HAN was proposed in the paper Hierarchical Attention Networks for Document Classification.

6.1 Description in Paper

  1. Word Encoder. Encoding by bidirectional GRU, an annotation for a given word is obtained by concatenating the forward hidden state and backward hidden state, which summarizes the information of the whole sentence centered around word in current time step.
  2. Word Attention. By a one-layer MLP and softmax function, it is enable to calculate normalized importance weights over the previous word annotations. Then, compute the sentence vector as a weighted sum of the word annotations based on the weights.
  3. Sentence Encoder. In a similar way with word encoder, use a bidirectional GRU to encode the sentences to get an annotation for a sentence.
  4. Sentence Attention. Similar with word attention, use a one-layer MLP and softmax function to get the weights over sentence annotations. Then, calculate a weighted sum of the sentence annotations based on the weights to get the document vector.
  5. Document Classification. Use the softmax function to calculate the probability of all classes.

6.2 Implementation Here

The implementation of attention here is based on FeedForwardAttention, which is the same as the attention in TextAttBiRNN.

Network structure of HAN:

The TimeDistributed wrapper is used here, since the parameters of the Embedding, Bidirectional RNN, and Attention layers are expected to be shared on the time step dimension.

7 RCNN

RCNN was proposed in the paper Recurrent Convolutional Neural Networks for Text Classification.

7.1 Description in Paper

  1. Word Representation Learning. RCNN uses a recurrent structure, which is a bi-directional recurrent neural network, to capture the contexts. Then, combine the word and its context to present the word. And apply a linear transformation together with the tanh activation fucntion to the representation.
  2. Text Representation Learning. When all of the representations of words are calculated, it applys a element-wise max-pooling layer in order to capture the most important information throughout the entire text. Finally, do the linear transformation and apply the softmax function.

7.2 Implementation Here

Network structure of RCNN:

8 RCNNVariant

RCNNVariant is an improved model based on RCNN with the following improvements. No related papers have been found yet.

  1. The three inputs are changed to single input. The input of the left and right contexts is removed.
  2. Use bidirectional LSTM/GRU instead of traditional RNN for encoding context.
  3. Use multi-channel CNN to represent the semantic vectors.
  4. Replace the Tanh activation layer with the ReLU activation layer.
  5. Use both AveragePooling and MaxPooling.

Network structure of RCNNVariant:

To Be Continued…

Reference

  1. Bag of Tricks for Efficient Text Classification
  2. Keras Example IMDB FastText
  3. Convolutional Neural Networks for Sentence Classification
  4. Keras Example IMDB CNN
  5. Recurrent Neural Network for Text Classification with Multi-Task Learning
  6. Neural Machine Translation by Jointly Learning to Align and Translate
  7. Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems
  8. cbaziotis’s Attention
  9. Hierarchical Attention Networks for Document Classification
  10. Richard’s HAN
  11. Recurrent Convolutional Neural Networks for Text Classification
  12. airalcorn2’s RCNN
comments powered by Disqus