SlideShare a Scribd company logo
International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021
DOI: 10.5121/ijnlc.2021.10402 15
SENTIMENT ANALYSIS IN MYANMAR LANGUAGE
USING CONVOLUTIONAL LSTM
NEURAL NETWORK
Nwet Yin Tun Thein and Khin Mar Soe
Department of Computer Science, University of Computer Studies, Yangon, Myanmar
ABSTRACT
In recent years, there has been an increasing use of social media among people in Myanmar and writing
review on social media pages about the product, movie, and trip are also popular among people. Moreover,
most of the people are going to find the review pages about the product they want to buy before deciding
whether they should buy it or not. Extracting and receiving useful reviews over interesting products is very
important and time consuming for people. Sentiment analysis is one of the important processes for extracting
useful reviews of the products. In this paper, the Convolutional LSTM neural network architecture is
proposed to analyse the sentiment classification of cosmetic reviews written in Myanmar Language. The
paper also intends to build the cosmetic reviews dataset for deep learning and sentiment lexicon in Myanmar
Language.
KEYWORDS
Social Media, Sentiment Analysis, Convolutional LSTM
1. INTRODUCTION
Analysing social media contents form Facebook, Twitter, and YouTube are important research area
in Natural Language Processing. There has been increasing use of Sentiment analysis to analyse
the social media contents. Spotting the sentiment of mentions (especially negative) is extremely
helpful in marketing areas as reputation management and customer feedback. From the customer
viewpoint, customers’ reviews and feedbacks about the products are very useful for their
determination whether to buy their interesting products or not. Moreover, deep learning is very
popular in sentiment analysis and text classification. In this paper, convolutional LSTM deep
learning approach is proposed for sentiment analysis on product reviews written in Myanmar
language.
2. RELATED WORK
In the recent years, Deep Learning approach was very popular and achieved the significant results
in computer vision and speech recognition [1]. They are very common to use in NLP applications
from learning word vector representations through building the neural language models. Moreover,
they also perform the composition over the learned word vectors for text classification and solve
the data sparsity problem [2].
Word embedding is the neural representation of a word and is a real vector. Word embedding
allows us to measure similarity between words by simply using the distance between two embedded
International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021
16
vectors [2][3]. Recently, researchers observed that is not necessary for deep neural network to
perform at word level. As long as the document represented as one-hot vector, the
model could work without any change, regardless of if each one-hot vector corresponds to a word.
Character sequence proposed as an alternative to the one-hot vector. Similar ideas also applied to
dependency parsing [4].
Deep CNN for NLP [5] is composed numerous of layers of convolutional and max pooling, it is
identical to the convolutional architecture in the computer vision [6]. CNN was initially designed
for computer vision and exploits layers with convolving filters that are applied to local features.
CNN reached an outstanding result in computer vision where handcrafted features were used, e.g.,
scale invariant features transform (SIFT) followed by a classifier. The main idea is to consider
features extractors and classifier as one jointly trained task [7]. CNN models for NLP achieved
excellent results in semantic parsing [7], sentence modelling [8], search query retrieval [9], and
other NLP tasks.
In this work Convolutional LSTM neural network architecture will be implemented for sentiment
classification of Myanmar Language Cosmetic Review.
3. NEURAL NETWORK ARCHITECTURE
The proposed neural network architecture will be implemented by the combination of
convolutional neural network with LSTM model.
3.1. Convolutional Neural Network (CNN)
A convolutional neural network (CNN) is a neural network that applies convolutional layers to
local features. When the input sentence is received, the CNN can be used to classify the sentiment
of sentence is positive, negative, or neural? The CNN architecture is shown in Figure 1.
Figure 1. CNN Model Architecture
As shown in the figure, the CNN model is created with four layers.
(i) Word Embedding layers
(ii) Convolution layers
(iii)Max-over time pooling layers and
(iv) Fully connected layer.
International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021
17
3.2. Recurrent Neural Network (RNN)
The objective of the RNN is to make use of sequential information, based on the previous
computation results. RNN has a memory that capture information in arbitrary long sequences. In
this model, word by word analysis is taken over the input and then the semantic of all the previous
text are preserved in a fixed-sized hidden layer. However, recurrent networks are biased model
because recent words are more significant than earlier words. Therefore, the efficiency is degraded
when the semantic of whole document is captured. To overcome this problem, Long short-term
memory (LSTM) is applied.
3.3. Long Short-Term Memory (LSTM)
LSTM is more complicated function that learns to control the flow of information, to prevent the
vanishing gradient and to allow the recurrent layer to capture long-term dependencies more easily.
The memory cell consists of four main components: input, output, forget gates and candidate
memory cell. Firstly, the value of input gate ( and candidate are computed according to the
following equations.
(1)
(2)
Then the value of forget gate ( ) is computed using the Eq. (3).
(3)
After getting the input gate activation ( ), forget gate activation ( and the candidate state value
( ), the candidate for new memory cell is computed using the Eq. (4).
(4)
With the new state of the memory cells, the value of output gate is computed using the Eq.(5) and
(6).
(5)
(6)
Where is the input to the memory cell layer at time t. , , , , , , and are weighted
matrices and bi, bf, bc, bo are bias vectors.
4. PROPOSED CONVOLUTIONAL LSTM MODEL FOR
SENTIMENT ANALYSIS
In the proposed model, a recurrent layer is added as the substitution of pooling layer to reduce the
number of convolution layers and capture the long-term dependencies. Therefore, convolutional,
and recurrent layers are considered as the single model in the proposed model. Moreover, rectified
linear (ReLus) was used for nonlinearity, padding was set to zero. All elements that would fall
outside the matrix are taken to be zero. To reduce overfitting, we applied dropout 0.5 only before
the recurrent layer. The Proposed model architecture is shown in Figure. 2.
International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021
18
Figure 2. Proposed Model Architecture
5. EXPERIMENTAL SETUP AND ANALYSIS
The performance of the proposed model will be evaluated over the cosmetic review dataset with
70,000 binary labelled reviews. The sample reviews are shown in Table 1. The reviews are divided
into 50:50 training and testing sets. The average length of each document is 320 tokens, with
standard deviation of 205.8 tokens; the maximum length of a document is 2,840 words. In
compared with other classification models, the proposed model outperforms the sentiment
classification results as shown in the Table 2.
Table 1. Sample Review Text with Result Class
Table 2. Accuracy of Sentiment Classification Models.
Models Accuracy
Naïve Bayes 80.3%
SVM 75%
RNN 85%
CNN 87%
Convolutional LSTM
(Proposed Model)
93.4%
6. CONCLUSION
In this paper, the combination of convolutional and recurrent layer into single neural network model
was proposed. The proposed model was validated on the cosmetic review text collected from social
media cosmetic pages written in Myanmar language. It achieved comparable results with less
number of convolutional layers compared to the convolutional only architecture. The proposed
model will also be applied other NLP such as semantic web search and spam filtering applications.
REFERENCES
[1] Graves, A, A.-r. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural networks. in
2013 IEEE international conference on acoustics, speech and signal processing. 2013. IEEE.
[2] Coliobert, R., et aI., Natural language processing (almost) from scratch. Journal of Machine Learning
Research, 20II. 12(Aug): p.2493-2537 acoustics, speech and signal processing. 2013. IEEE.
International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021
19
[3] Mikolov, T., et al. Distributed representations of words and phrases and their compositionality. in
Advances in neural information processing systems. 2013.
[4] Ling, W., et aI., Finding function in form: Compositional character models for open vocabulary word
representation. arXiv preprint arXiv:1508.02096, 2015.
[5] Conneau, A, et aI., Very Deep Convolutional Networks for Natural Language Processing. arXiv
preprint arXiv:1606.01781, 2016.
[6] Yih, W.-t., X. He, and C. Meek. Semantic Parsing for Single-Relation Question Answering. in ACL(2).
2014. Citeseer.
[7] Kalchbrenner, N., E. Grefenstette, and P. Blunsom, A convolutional neural network for modelling
sentences. arXiv preprint arXiv:1404.2188, 2014.
[8] Sundermeyer, M., H. Ney, and R. SchlUter, From feedfonvard to recurrent LSTM neural networks for
language modeling. IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP),
2015.23(3): p. 517-529.
AUTHORS
Ms. Nwet Yin Tun Thein is a Ph.D candidate of University of Computer Studies,
Yangon. Her research interest is Natural Language Processing, especially in Sentiment
analysis. Currently, she is doing research on sentiment analysis on Myanmar Language
Dr. Khin Mar Soe is a professor of University of Computer Studies, Yangon. Her
research interest are Natural Language Processing, Machine Learning, and Deep
Learning. She has been supervising Master thesis and Ph.D thesis on Natural language
processing such as Information Retrieval, Morphological Analysis, Summarization,
Parsing, Machine Translation and sentiment analysis.

More Related Content

What's hot (17)

PDF
Nlp research presentation
Surya Sg
 
PDF
semantic text doc clustering
Souvik Roy
 
PDF
IRJET- Short-Text Semantic Similarity using Glove Word Embedding
IRJET Journal
 
PPTX
[Paper Reading] Supervised Learning of Universal Sentence Representations fro...
Hiroki Shimanaka
 
PDF
Cleveree: an artificially intelligent web service for Jacob voice chatbot
TELKOMNIKA JOURNAL
 
PDF
Meaning Extraction - IJCTE 2(1)
IT Industry
 
PDF
IRJET- Survey on Generating Suggestions for Erroneous Part in a Sentence
IRJET Journal
 
PDF
An Improved Approach for Word Ambiguity Removal
Waqas Tariq
 
PDF
A Dialogue System for Telugu, a Resource-Poor Language
Sravanthi Mullapudi
 
PDF
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
Journal For Research
 
PDF
IRJET- Survey on Deep Learning Approaches for Phrase Structure Identification...
IRJET Journal
 
PDF
arttt.pdf
ferejadawud
 
PDF
GENERATING SUMMARIES USING SENTENCE COMPRESSION AND STATISTICAL MEASURES
ijnlc
 
PDF
SEMI-AUTOMATIC SIMULTANEOUS INTERPRETING QUALITY EVALUATION
ijnlc
 
PDF
Sentiment analysis by deep learning approaches
TELKOMNIKA JOURNAL
 
PDF
A fuzzy logic based on sentiment
IJDKP
 
PDF
SYLLABLE-BASED NEURAL NAMED ENTITY RECOGNITION FOR MYANMAR LANGUAGE
ijnlc
 
Nlp research presentation
Surya Sg
 
semantic text doc clustering
Souvik Roy
 
IRJET- Short-Text Semantic Similarity using Glove Word Embedding
IRJET Journal
 
[Paper Reading] Supervised Learning of Universal Sentence Representations fro...
Hiroki Shimanaka
 
Cleveree: an artificially intelligent web service for Jacob voice chatbot
TELKOMNIKA JOURNAL
 
Meaning Extraction - IJCTE 2(1)
IT Industry
 
IRJET- Survey on Generating Suggestions for Erroneous Part in a Sentence
IRJET Journal
 
An Improved Approach for Word Ambiguity Removal
Waqas Tariq
 
A Dialogue System for Telugu, a Resource-Poor Language
Sravanthi Mullapudi
 
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
Journal For Research
 
IRJET- Survey on Deep Learning Approaches for Phrase Structure Identification...
IRJET Journal
 
arttt.pdf
ferejadawud
 
GENERATING SUMMARIES USING SENTENCE COMPRESSION AND STATISTICAL MEASURES
ijnlc
 
SEMI-AUTOMATIC SIMULTANEOUS INTERPRETING QUALITY EVALUATION
ijnlc
 
Sentiment analysis by deep learning approaches
TELKOMNIKA JOURNAL
 
A fuzzy logic based on sentiment
IJDKP
 
SYLLABLE-BASED NEURAL NAMED ENTITY RECOGNITION FOR MYANMAR LANGUAGE
ijnlc
 

Similar to Sentiment Analysis In Myanmar Language Using Convolutional Lstm Neural Network (20)

PDF
LSTM Based Sentiment Analysis
ijtsrd
 
PDF
SENTIMENT ANALYSIS FOR MOVIES REVIEWS DATASET USING DEEP LEARNING MODELS
IJDKP
 
PDF
A-STUDY-ON-SENTIMENT-POLARITY.pdf
SUDESHNASANI1
 
PDF
DEEP LEARNING SENTIMENT ANALYSIS OF AMAZON.COM REVIEWS AND RATINGS
ijscai
 
PDF
Survey on Text Prediction Techniques
vivatechijri
 
PDF
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)
Márton Miháltz
 
PDF
Cyber bullying detection and analysis.ppt.pdf
Hunais Abdul Nafi
 
PPT
Ml ppt
Alpna Patel
 
PDF
IRJET - Sentiment Analysis and Rumour Detection in Online Product Reviews
IRJET Journal
 
PPTX
Sentiment analysis
Aditya Kamble
 
PPTX
sentiment analysis using machine learning
RameshPrasadBhatta2
 
PDF
Eat it, Review it: A New Approach for Review Prediction
vivatechijri
 
PPTX
BE-IT01 (1).pptx
Shivam327815
 
PDF
Optimizer algorithms and convolutional neural networks for text classification
IAESIJAI
 
PDF
hardback
Conor Brady
 
PDF
Deep Learning-Based Opinion Mining for Bitcoin Price Prediction with Joyesh ...
Databricks
 
PPTX
Talk from NVidia Developer Connect
Anuj Gupta
 
PDF
Movie Sentiment Analysis using Deep Learning RNN
ijtsrd
 
PPTX
Sentiment Analysis of Bengali text using Gated Recurrent Neural Network
A. Hasib Uddin
 
PDF
Sentimental Analysis For Electronic Product Review
IRJET Journal
 
LSTM Based Sentiment Analysis
ijtsrd
 
SENTIMENT ANALYSIS FOR MOVIES REVIEWS DATASET USING DEEP LEARNING MODELS
IJDKP
 
A-STUDY-ON-SENTIMENT-POLARITY.pdf
SUDESHNASANI1
 
DEEP LEARNING SENTIMENT ANALYSIS OF AMAZON.COM REVIEWS AND RATINGS
ijscai
 
Survey on Text Prediction Techniques
vivatechijri
 
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)
Márton Miháltz
 
Cyber bullying detection and analysis.ppt.pdf
Hunais Abdul Nafi
 
Ml ppt
Alpna Patel
 
IRJET - Sentiment Analysis and Rumour Detection in Online Product Reviews
IRJET Journal
 
Sentiment analysis
Aditya Kamble
 
sentiment analysis using machine learning
RameshPrasadBhatta2
 
Eat it, Review it: A New Approach for Review Prediction
vivatechijri
 
BE-IT01 (1).pptx
Shivam327815
 
Optimizer algorithms and convolutional neural networks for text classification
IAESIJAI
 
hardback
Conor Brady
 
Deep Learning-Based Opinion Mining for Bitcoin Price Prediction with Joyesh ...
Databricks
 
Talk from NVidia Developer Connect
Anuj Gupta
 
Movie Sentiment Analysis using Deep Learning RNN
ijtsrd
 
Sentiment Analysis of Bengali text using Gated Recurrent Neural Network
A. Hasib Uddin
 
Sentimental Analysis For Electronic Product Review
IRJET Journal
 
Ad

More from kevig (20)

PDF
INTERLINGUAL SYNTACTIC PARSING: AN OPTIMIZED HEAD-DRIVEN PARSING FOR ENGLISH ...
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
UNIQUE APPROACH TO CONTROL SPEECH, SENSORY AND MOTOR NEURONAL DISORDER THROUG...
kevig
 
PDF
Call For Papers - 6th International Conference on Natural Language Processing...
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
Natural language processing through the subtractive mountain clustering algor...
kevig
 
PDF
Call For Papers - 4th International Conference on Machine Learning, NLP and D...
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
STREAMING PUNCTUATION: A NOVEL PUNCTUATION TECHNIQUE LEVERAGING BIDIRECTIONAL...
kevig
 
PDF
Call For Papers - 17th International Conference on Networks & Communications ...
kevig
 
PDF
Call For Papers - 6th International Conference on NLP & Big Data (NLPD 2025)
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
LOCATION-BASED SENTIMENT ANALYSIS OF 2019 NIGERIA PRESIDENTIAL ELECTION USING...
kevig
 
PDF
Call For Papers - 6th International Conference on NLP & Big Data (NLPD 2025)
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
HUMAN INTENTION SPACE - NATURAL LANGUAGE PHRASE DRIVEN APPROACH TO PLACE SOCI...
kevig
 
PDF
Call For Papers - 5th International Conference on NLP & Data Mining (NLDM 2025)
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
PDF
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
kevig
 
PDF
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
INTERLINGUAL SYNTACTIC PARSING: AN OPTIMIZED HEAD-DRIVEN PARSING FOR ENGLISH ...
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
UNIQUE APPROACH TO CONTROL SPEECH, SENSORY AND MOTOR NEURONAL DISORDER THROUG...
kevig
 
Call For Papers - 6th International Conference on Natural Language Processing...
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
Natural language processing through the subtractive mountain clustering algor...
kevig
 
Call For Papers - 4th International Conference on Machine Learning, NLP and D...
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
STREAMING PUNCTUATION: A NOVEL PUNCTUATION TECHNIQUE LEVERAGING BIDIRECTIONAL...
kevig
 
Call For Papers - 17th International Conference on Networks & Communications ...
kevig
 
Call For Papers - 6th International Conference on NLP & Big Data (NLPD 2025)
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
LOCATION-BASED SENTIMENT ANALYSIS OF 2019 NIGERIA PRESIDENTIAL ELECTION USING...
kevig
 
Call For Papers - 6th International Conference on NLP & Big Data (NLPD 2025)
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
HUMAN INTENTION SPACE - NATURAL LANGUAGE PHRASE DRIVEN APPROACH TO PLACE SOCI...
kevig
 
Call For Papers - 5th International Conference on NLP & Data Mining (NLDM 2025)
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
kevig
 
Call For Papers - International Journal on Natural Language Computing (IJNLC)
kevig
 
Ad

Recently uploaded (20)

PDF
Electrical Machines and Their Protection.pdf
Nabajyoti Banik
 
PDF
20ES1152 Programming for Problem Solving Lab Manual VRSEC.pdf
Ashutosh Satapathy
 
PDF
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
PPTX
2025 CGI Congres - Surviving agile v05.pptx
Derk-Jan de Grood
 
PPT
New_school_Engineering_presentation_011707.ppt
VinayKumar304579
 
PDF
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
PPTX
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
PDF
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
PPTX
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
PDF
REINFORCEMENT LEARNING IN DECISION MAKING SEMINAR REPORT
anushaashraf20
 
PDF
SERVERLESS PERSONAL TO-DO LIST APPLICATION
anushaashraf20
 
PPT
Testing and final inspection of a solar PV system
MuhammadSanni2
 
PPTX
Worm gear strength and wear calculation as per standard VB Bhandari Databook.
shahveer210504
 
PPTX
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
PPTX
澳洲电子毕业证澳大利亚圣母大学水印成绩单UNDA学生证网上可查学历
Taqyea
 
PDF
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
PPTX
Lecture 1 Shell and Tube Heat exchanger-1.pptx
mailforillegalwork
 
PDF
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
PDF
Pressure Measurement training for engineers and Technicians
AIESOLUTIONS
 
PPTX
Distribution reservoir and service storage pptx
dhanashree78
 
Electrical Machines and Their Protection.pdf
Nabajyoti Banik
 
20ES1152 Programming for Problem Solving Lab Manual VRSEC.pdf
Ashutosh Satapathy
 
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
2025 CGI Congres - Surviving agile v05.pptx
Derk-Jan de Grood
 
New_school_Engineering_presentation_011707.ppt
VinayKumar304579
 
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
REINFORCEMENT LEARNING IN DECISION MAKING SEMINAR REPORT
anushaashraf20
 
SERVERLESS PERSONAL TO-DO LIST APPLICATION
anushaashraf20
 
Testing and final inspection of a solar PV system
MuhammadSanni2
 
Worm gear strength and wear calculation as per standard VB Bhandari Databook.
shahveer210504
 
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
澳洲电子毕业证澳大利亚圣母大学水印成绩单UNDA学生证网上可查学历
Taqyea
 
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
Lecture 1 Shell and Tube Heat exchanger-1.pptx
mailforillegalwork
 
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
Pressure Measurement training for engineers and Technicians
AIESOLUTIONS
 
Distribution reservoir and service storage pptx
dhanashree78
 

Sentiment Analysis In Myanmar Language Using Convolutional Lstm Neural Network

  • 1. International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021 DOI: 10.5121/ijnlc.2021.10402 15 SENTIMENT ANALYSIS IN MYANMAR LANGUAGE USING CONVOLUTIONAL LSTM NEURAL NETWORK Nwet Yin Tun Thein and Khin Mar Soe Department of Computer Science, University of Computer Studies, Yangon, Myanmar ABSTRACT In recent years, there has been an increasing use of social media among people in Myanmar and writing review on social media pages about the product, movie, and trip are also popular among people. Moreover, most of the people are going to find the review pages about the product they want to buy before deciding whether they should buy it or not. Extracting and receiving useful reviews over interesting products is very important and time consuming for people. Sentiment analysis is one of the important processes for extracting useful reviews of the products. In this paper, the Convolutional LSTM neural network architecture is proposed to analyse the sentiment classification of cosmetic reviews written in Myanmar Language. The paper also intends to build the cosmetic reviews dataset for deep learning and sentiment lexicon in Myanmar Language. KEYWORDS Social Media, Sentiment Analysis, Convolutional LSTM 1. INTRODUCTION Analysing social media contents form Facebook, Twitter, and YouTube are important research area in Natural Language Processing. There has been increasing use of Sentiment analysis to analyse the social media contents. Spotting the sentiment of mentions (especially negative) is extremely helpful in marketing areas as reputation management and customer feedback. From the customer viewpoint, customers’ reviews and feedbacks about the products are very useful for their determination whether to buy their interesting products or not. Moreover, deep learning is very popular in sentiment analysis and text classification. In this paper, convolutional LSTM deep learning approach is proposed for sentiment analysis on product reviews written in Myanmar language. 2. RELATED WORK In the recent years, Deep Learning approach was very popular and achieved the significant results in computer vision and speech recognition [1]. They are very common to use in NLP applications from learning word vector representations through building the neural language models. Moreover, they also perform the composition over the learned word vectors for text classification and solve the data sparsity problem [2]. Word embedding is the neural representation of a word and is a real vector. Word embedding allows us to measure similarity between words by simply using the distance between two embedded
  • 2. International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021 16 vectors [2][3]. Recently, researchers observed that is not necessary for deep neural network to perform at word level. As long as the document represented as one-hot vector, the model could work without any change, regardless of if each one-hot vector corresponds to a word. Character sequence proposed as an alternative to the one-hot vector. Similar ideas also applied to dependency parsing [4]. Deep CNN for NLP [5] is composed numerous of layers of convolutional and max pooling, it is identical to the convolutional architecture in the computer vision [6]. CNN was initially designed for computer vision and exploits layers with convolving filters that are applied to local features. CNN reached an outstanding result in computer vision where handcrafted features were used, e.g., scale invariant features transform (SIFT) followed by a classifier. The main idea is to consider features extractors and classifier as one jointly trained task [7]. CNN models for NLP achieved excellent results in semantic parsing [7], sentence modelling [8], search query retrieval [9], and other NLP tasks. In this work Convolutional LSTM neural network architecture will be implemented for sentiment classification of Myanmar Language Cosmetic Review. 3. NEURAL NETWORK ARCHITECTURE The proposed neural network architecture will be implemented by the combination of convolutional neural network with LSTM model. 3.1. Convolutional Neural Network (CNN) A convolutional neural network (CNN) is a neural network that applies convolutional layers to local features. When the input sentence is received, the CNN can be used to classify the sentiment of sentence is positive, negative, or neural? The CNN architecture is shown in Figure 1. Figure 1. CNN Model Architecture As shown in the figure, the CNN model is created with four layers. (i) Word Embedding layers (ii) Convolution layers (iii)Max-over time pooling layers and (iv) Fully connected layer.
  • 3. International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021 17 3.2. Recurrent Neural Network (RNN) The objective of the RNN is to make use of sequential information, based on the previous computation results. RNN has a memory that capture information in arbitrary long sequences. In this model, word by word analysis is taken over the input and then the semantic of all the previous text are preserved in a fixed-sized hidden layer. However, recurrent networks are biased model because recent words are more significant than earlier words. Therefore, the efficiency is degraded when the semantic of whole document is captured. To overcome this problem, Long short-term memory (LSTM) is applied. 3.3. Long Short-Term Memory (LSTM) LSTM is more complicated function that learns to control the flow of information, to prevent the vanishing gradient and to allow the recurrent layer to capture long-term dependencies more easily. The memory cell consists of four main components: input, output, forget gates and candidate memory cell. Firstly, the value of input gate ( and candidate are computed according to the following equations. (1) (2) Then the value of forget gate ( ) is computed using the Eq. (3). (3) After getting the input gate activation ( ), forget gate activation ( and the candidate state value ( ), the candidate for new memory cell is computed using the Eq. (4). (4) With the new state of the memory cells, the value of output gate is computed using the Eq.(5) and (6). (5) (6) Where is the input to the memory cell layer at time t. , , , , , , and are weighted matrices and bi, bf, bc, bo are bias vectors. 4. PROPOSED CONVOLUTIONAL LSTM MODEL FOR SENTIMENT ANALYSIS In the proposed model, a recurrent layer is added as the substitution of pooling layer to reduce the number of convolution layers and capture the long-term dependencies. Therefore, convolutional, and recurrent layers are considered as the single model in the proposed model. Moreover, rectified linear (ReLus) was used for nonlinearity, padding was set to zero. All elements that would fall outside the matrix are taken to be zero. To reduce overfitting, we applied dropout 0.5 only before the recurrent layer. The Proposed model architecture is shown in Figure. 2.
  • 4. International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021 18 Figure 2. Proposed Model Architecture 5. EXPERIMENTAL SETUP AND ANALYSIS The performance of the proposed model will be evaluated over the cosmetic review dataset with 70,000 binary labelled reviews. The sample reviews are shown in Table 1. The reviews are divided into 50:50 training and testing sets. The average length of each document is 320 tokens, with standard deviation of 205.8 tokens; the maximum length of a document is 2,840 words. In compared with other classification models, the proposed model outperforms the sentiment classification results as shown in the Table 2. Table 1. Sample Review Text with Result Class Table 2. Accuracy of Sentiment Classification Models. Models Accuracy Naïve Bayes 80.3% SVM 75% RNN 85% CNN 87% Convolutional LSTM (Proposed Model) 93.4% 6. CONCLUSION In this paper, the combination of convolutional and recurrent layer into single neural network model was proposed. The proposed model was validated on the cosmetic review text collected from social media cosmetic pages written in Myanmar language. It achieved comparable results with less number of convolutional layers compared to the convolutional only architecture. The proposed model will also be applied other NLP such as semantic web search and spam filtering applications. REFERENCES [1] Graves, A, A.-r. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural networks. in 2013 IEEE international conference on acoustics, speech and signal processing. 2013. IEEE. [2] Coliobert, R., et aI., Natural language processing (almost) from scratch. Journal of Machine Learning Research, 20II. 12(Aug): p.2493-2537 acoustics, speech and signal processing. 2013. IEEE.
  • 5. International Journal on Natural Language Computing (IJNLC) Vol.10, No.4, August 2021 19 [3] Mikolov, T., et al. Distributed representations of words and phrases and their compositionality. in Advances in neural information processing systems. 2013. [4] Ling, W., et aI., Finding function in form: Compositional character models for open vocabulary word representation. arXiv preprint arXiv:1508.02096, 2015. [5] Conneau, A, et aI., Very Deep Convolutional Networks for Natural Language Processing. arXiv preprint arXiv:1606.01781, 2016. [6] Yih, W.-t., X. He, and C. Meek. Semantic Parsing for Single-Relation Question Answering. in ACL(2). 2014. Citeseer. [7] Kalchbrenner, N., E. Grefenstette, and P. Blunsom, A convolutional neural network for modelling sentences. arXiv preprint arXiv:1404.2188, 2014. [8] Sundermeyer, M., H. Ney, and R. SchlUter, From feedfonvard to recurrent LSTM neural networks for language modeling. IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP), 2015.23(3): p. 517-529. AUTHORS Ms. Nwet Yin Tun Thein is a Ph.D candidate of University of Computer Studies, Yangon. Her research interest is Natural Language Processing, especially in Sentiment analysis. Currently, she is doing research on sentiment analysis on Myanmar Language Dr. Khin Mar Soe is a professor of University of Computer Studies, Yangon. Her research interest are Natural Language Processing, Machine Learning, and Deep Learning. She has been supervising Master thesis and Ph.D thesis on Natural language processing such as Information Retrieval, Morphological Analysis, Summarization, Parsing, Machine Translation and sentiment analysis.