***
Abstract:
Unstructured text data produced on the internet grows rapidly, and sentiment analysis for short texts becomes a challenge because of the limit of the contextual information they usually contain.
Learning good vector representations for sentences is a challenging task and an ongoing research area.
Moreover, learning long-term dependencies with gradient descent is difficult in neural network language model because of the vanishing gradients problem.
Natural Language Processing (NLP) systems traditionally treat words as discrete atomic symbols; the model can leverage small amounts of information regarding the relationship between the individual symbols.
In this paper, we propose ConvLstm, neural network architecture that employs Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) on top of pre-trained word vectors.
In our experiments, ConvLstm exploit LSTM as a substitute of pooling layer in CNN to reduce the loss of detailed local information and capture long term dependencies in sequence of sentences.
We validate the proposed model on two sentiment datasets IMDB, and Stanford Sentiment Treebank (SSTb).
Empirical results show that ConvLstm achieved comparable performances with less parameters on sentiment analysis tasks.
***
https://ieeexplore.ieee.org/document/7942788
***
No comments:
Post a Comment