DOI QR코드

DOI QR Code

Sentiment Analysis of Korean Reviews Using CNN: Focusing on Morpheme Embedding

CNN을 적용한 한국어 상품평 감성분석: 형태소 임베딩을 중심으로

  • 박현정 (이화여자대학교 경영연구소) ;
  • 송민채 (이화여자대학교 빅데이터분석학) ;
  • 신경식 (이화여자대학교 경영대학)
  • Received : 2018.03.28
  • Accepted : 2018.05.30
  • Published : 2018.06.30

Abstract

With the increasing importance of sentiment analysis to grasp the needs of customers and the public, various types of deep learning models have been actively applied to English texts. In the sentiment analysis of English texts by deep learning, natural language sentences included in training and test datasets are usually converted into sequences of word vectors before being entered into the deep learning models. In this case, word vectors generally refer to vector representations of words obtained through splitting a sentence by space characters. There are several ways to derive word vectors, one of which is Word2Vec used for producing the 300 dimensional Google word vectors from about 100 billion words of Google News data. They have been widely used in the studies of sentiment analysis of reviews from various fields such as restaurants, movies, laptops, cameras, etc. Unlike English, morpheme plays an essential role in sentiment analysis and sentence structure analysis in Korean, which is a typical agglutinative language with developed postpositions and endings. A morpheme can be defined as the smallest meaningful unit of a language, and a word consists of one or more morphemes. For example, for a word '예쁘고', the morphemes are '예쁘(= adjective)' and '고(=connective ending)'. Reflecting the significance of Korean morphemes, it seems reasonable to adopt the morphemes as a basic unit in Korean sentiment analysis. Therefore, in this study, we use 'morpheme vector' as an input to a deep learning model rather than 'word vector' which is mainly used in English text. The morpheme vector refers to a vector representation for the morpheme and can be derived by applying an existent word vector derivation mechanism to the sentences divided into constituent morphemes. By the way, here come some questions as follows. What is the desirable range of POS(Part-Of-Speech) tags when deriving morpheme vectors for improving the classification accuracy of a deep learning model? Is it proper to apply a typical word vector model which primarily relies on the form of words to Korean with a high homonym ratio? Will the text preprocessing such as correcting spelling or spacing errors affect the classification accuracy, especially when drawing morpheme vectors from Korean product reviews with a lot of grammatical mistakes and variations? We seek to find empirical answers to these fundamental issues, which may be encountered first when applying various deep learning models to Korean texts. As a starting point, we summarized these issues as three central research questions as follows. First, which is better effective, to use morpheme vectors from grammatically correct texts of other domain than the analysis target, or to use morpheme vectors from considerably ungrammatical texts of the same domain, as the initial input of a deep learning model? Second, what is an appropriate morpheme vector derivation method for Korean regarding the range of POS tags, homonym, text preprocessing, minimum frequency? Third, can we get a satisfactory level of classification accuracy when applying deep learning to Korean sentiment analysis? As an approach to these research questions, we generate various types of morpheme vectors reflecting the research questions and then compare the classification accuracy through a non-static CNN(Convolutional Neural Network) model taking in the morpheme vectors. As for training and test datasets, Naver Shopping's 17,260 cosmetics product reviews are used. To derive morpheme vectors, we use data from the same domain as the target one and data from other domain; Naver shopping's about 2 million cosmetics product reviews and 520,000 Naver News data arguably corresponding to Google's News data. The six primary sets of morpheme vectors constructed in this study differ in terms of the following three criteria. First, they come from two types of data source; Naver news of high grammatical correctness and Naver shopping's cosmetics product reviews of low grammatical correctness. Second, they are distinguished in the degree of data preprocessing, namely, only splitting sentences or up to additional spelling and spacing corrections after sentence separation. Third, they vary concerning the form of input fed into a word vector model; whether the morphemes themselves are entered into a word vector model or with their POS tags attached. The morpheme vectors further vary depending on the consideration range of POS tags, the minimum frequency of morphemes included, and the random initialization range. All morpheme vectors are derived through CBOW(Continuous Bag-Of-Words) model with the context window 5 and the vector dimension 300. It seems that utilizing the same domain text even with a lower degree of grammatical correctness, performing spelling and spacing corrections as well as sentence splitting, and incorporating morphemes of any POS tags including incomprehensible category lead to the better classification accuracy. The POS tag attachment, which is devised for the high proportion of homonyms in Korean, and the minimum frequency standard for the morpheme to be included seem not to have any definite influence on the classification accuracy.

고객과 대중의 니즈를 파악하기 위한 감성분석의 중요성이 커지면서 최근 영어 텍스트를 대상으로 다양한 딥러닝 모델들이 소개되고 있다. 본 연구는 영어와 한국어의 언어적인 차이에 주목하여 딥러닝 모델을 한국어 상품평 텍스트의 감성분석에 적용할 때 부딪히게 되는 기본적인 이슈들에 대하여 실증적으로 살펴본다. 즉, 딥러닝 모델의 입력으로 사용되는 단어 벡터(word vector)를 형태소 수준에서 도출하고, 여러 형태소 벡터(morpheme vector) 도출 대안에 따라 감성분석의 정확도가 어떻게 달라지는지를 비정태적(non-static) CNN(Convolutional Neural Network) 모델을 사용하여 검증한다. 형태소 벡터 도출 대안은 CBOW(Continuous Bag-Of-Words)를 기본적으로 적용하고, 입력 데이터의 종류, 문장 분리와 맞춤법 및 띄어쓰기 교정, 품사 선택, 품사 태그 부착, 고려 형태소의 최소 빈도수 등과 같은 기준에 따라 달라진다. 형태소 벡터 도출 시, 문법 준수도가 낮더라도 감성분석 대상과 같은 도메인의 텍스트를 사용하고, 문장 분리 외에 맞춤법 및 띄어쓰기 전처리를 하며, 분석불능 범주를 포함한 모든 품사를 고려할 때 감성분석의 분류 정확도가 향상되는 결과를 얻었다. 동음이의어 비율이 높은 한국어 특성 때문에 고려한 품사 태그 부착 방안과 포함할 형태소에 대한 최소 빈도수 기준은 뚜렷한 영향이 없는 것으로 나타났다.

Keywords

References

  1. An, J.-y., J.-w. Bae, N.-g. Han, and M. Song, "A Study of 'Emotion Trigger' by Text Mining Techniques," Journal of Intelligence and Information Systems, Vol.21, No.2(2015), 69-92. https://doi.org/10.13088/jiis.2015.21.2.69
  2. Chen, P., Z. Sun, L. Bing, and W. Yang, "Recurrent Attention Network on Memory for Aspect Sentiment Analysis," Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, (2017).
  3. Cho, K., van M. Bart, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, "Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation," arXiv preprint arXiv:1406.1078, (2014).
  4. Collobert, R., J. Weston, L. Bottou, M. Karlen, K. Kavukcuglu, and P. Kuksa, "Natural Language Processing (Almost) from Scratch," Journal of Machine Learning Research, Vol.12, (2011), 2493-2537.
  5. Cui, M.-n., Y.-s. Jin, and O.-b. Kwon, "A Method of Analyzing Sentiment Polarity of Multilingual Social Media: A Case of Korean-Chinese Languages," Journal of Intelligence and Information Systems, Vol.22, No.3(2016), 91-111. https://doi.org/10.13088/jiis.2016.22.3.091
  6. Dhanush, D. and A. K. Thakur, "Aspect-based Sentiment Summarization with Deep Neural Networks," International Journal of Engineering Research and Technology, Vol.5, No.5(2016), 371-375. https://doi.org/10.15623/ijret.2016.0508062
  7. Guan, Z., L. Chen, W. Zhao, Y. Zheng, S. Tan, and D. Cai, "Weakly-supervised Deep Learning for Customer Review Sentiment Classification," Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI'16), (2016), 3719-3725.
  8. Hochreiter, S. and J. Schmidhuber, "Long Short-Term Memory," Neural Computation, Vol.9, No.8(1997), 1735-1780. https://doi.org/10.1162/neco.1997.9.8.1735
  9. Hong, T.-h., T.-w. Lee, and J.-g. Li, "Development of Sentiment Analysis Model for the Hot Topic Detection of Online Stock Forums," Journal of Intelligence and Information Systems, Vol.22, No.1(2016), 187-204. https://doi.org/10.13088/jiis.2016.22.1.187
  10. Jebbara, S. and P. Cimiano, "Aspect-Based Relational Sentiment Analysis Using a Stacked Neural Network Architecture," arXiv preprint arXiv:1709.06309, (2017).
  11. Jebbara, S. and P. Cimiano, "Aspect-Based Sentiment Analysis Using a Two-Step Neural Network Architecture," arXiv preprint arXiv:1709.06311, (2017).
  12. Jeong, J. S., D. S. Kim, and J. W. Kim, "Influence Analysis of Internet Buzz to Corporate Performance: Individual Stock Price Prediction Using Sentiment Analysis of Online News," Journal of Intelligence and Information Systems, Vol.21, No.4(2015), 37-51. https://doi.org/10.13088/JIIS.2015.21.4.037
  13. Joulin, A., E. Grave, P. Bojanowski, and T. Mikolov, "Bag of Tricks for Efficient Text Classification," arXiv preprint arXiv, (2016).
  14. Kang, B.-m., "Aspects of the Use of Homonyms," Language Research, Vol.41, No.1(2005), 1-29.
  15. Kim, B. G., "A Study on the Homonym in Korean," Our Language Study, Vol., No.45 (2013), 181-200.
  16. Kim, J. H., Understanding Linguistics, Yeog-Lag, 2004.
  17. Kim, S. I., D. S. Kim, and J. W. Kim, "Public Sentiment Analysis of Korean Top-10 Companies: Big Data Approach Using Multi-Categorical Sentiment Lexicon," Journal of Intelligence and Information Systems, Vol.22, No.3(2016), 45-69. https://doi.org/10.13088/JIIS.2016.22.3.045
  18. Kim, Y., "Convolutional Neural Networks for Sentence Classification," Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing(EMNLP), (2014), 1746-1751.
  19. Liu, B., Sentiment Analysis: Mining Opinions, Sentiments, and Emotions, The Cambridge University Press, 2015.
  20. Ma, D., S. Li, X. Zhang, and H. Wang, "Interactive Attention Networks for Aspect-Level Sentiment Classification," Proceedings of the Internal Joint Conference on Artificial Intelligence(IJCAI 2017), (2017).
  21. Mikolov, T., I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, "Distributed Representations of Words and Phrases and their Compositionality," Advances in Neural Information Processing Systems, (2013a), 3111-3119.
  22. Mikolov, T., K. Chen, G. Corrado, and J. Dean, "Efficient Estimation of Word Representations in Vector Space," arXiv preprint arXiv:1301.3781, (2013b).
  23. Nam, G.-l. and E.-g. Jo., Korean Text Sentiment Analysis, Communication-Books, 2017.
  24. Ouyang, X., P. Zhou, C. H. Li, and L. Liu, "Sentiment Analysis Using Convolutional Neural Network," 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, (2015), 2359-2364.
  25. Pang, B. and L. Lee, "Opinion Mining and Sentiment Analysis," Foundations and Trends in Information Retrieval, Vol.2, No.1-2(2008), 11-35.
  26. Pennington, J., R. Socher, and C. D. Manning, "GloVe: Global Vectors for Word Representation," Proceedings of the Conference on Empirical Methods on Natural Language Processing(EMNLP 2014), (2014).
  27. Ruder, S., P. Ghaffari, and J. G. Breslin, "A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis," Proceedings of the Conference on Empirical Methods on Natural Language Processing(EMNLP 2016), (2016a).
  28. Ruder, S., P. Ghaffari, and J. G. Breslin, "INSIGHT-1 at SemEval-2016 Task 5: Deep Learning for Multilingual Aspect-based Sentiment Analysis," Proceedings of SemEval, (2016b), 330-336.
  29. Schuster, M. and K. K. Paliwal, "Bidirectional Recurrent Neural Networks," IEEE Transactions on Signal Processing, Vol.45, No.11(1997), 2673-2681. https://doi.org/10.1109/78.650093
  30. Shin, B., T. Lee, and J. D. Choi, "Lexicon Integrated CNN Models with Attention for Sentiment Analysis," Proceedings of the EMNLP Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis of WASSA'17, (2017).
  31. Shirani-Mehr, H., "Applications of Deep Learning to Sentiment Analysis of Movie Reviews," Technical Report, Stanford University, (2015).
  32. Socher, R., C. Lin, A. Y. Ng, and C. D. Manning, "Parsing Natural Scenes and Natural Language with Recursive Neural Networks," Proceedings of the 28th International Conference on Machine Learning(ICML-11), (2011).
  33. Stevenson, M., Word Sense Disambiguation: The Case for Combinations of Knowldedge Sources, Stanford: CSLI Publications, 2003.
  34. Tang, D., B. Qin, X. Feng, and T. Liu, "Effective LSTMs for target-dependent sentiment classification," Proceedings of the International Conference on Computational Linguistics (COLING 2016), (2016).
  35. Tay, Y., L. A. Tuan, and S. C. Hui, "Dyadic Memory Networks for Aspect-based Sentiment Analysis," Proceedings of the International Conference on Information and Knowledge Management(CIKM 2017), (2017).
  36. Wang, B. and M. Liu, "Deep Learning for Aspect-Based Sentiment Analysis," Stanford University Report, 2015. Available at https://cs224d.stanford.edu/reports/WangBo (Downloaded 19 May, 2018).
  37. Wang, Y., M. Huang, X. Zhu, and L. Zhao, "Attention-based LSTM for Aspect-level Sentiment Classification," Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, (2016), 606-615.
  38. Wiebe, J., T. Wilson, R. Bruce, M. Bell, and M. Martin, "Learning Subjective Language," Computational Linguistics, Vol.30, No.3 (2004), 277-308. https://doi.org/10.1162/0891201041850885
  39. Wikibaeggwa, Available at https://ko.wikipedia.org/wiki/%EB%82%B1%EB%A7%90/ (Downloa ded 20 March, 2018).
  40. Yang, Z., D. Yang, C. Dyer, X. He, A. J. Smola, and E. H. Hovy, "Hierarchical Attention Networks for Document Classification," HLT-NAACL(North American Chapter of the Association for Computational Linguistics: Human Language Technologies), (2016).
  41. Zhang, L., S. Wang, and B. Liu, "Deep Learning for Sentiment Analysis: A Survey," arXiv preprint arXiv:1801.07883, (2018).

Cited by

  1. BERT를 활용한 속성기반 감성분석: 속성카테고리 감성분류 모델 개발 vol.26, pp.4, 2020, https://doi.org/10.13088/jiis.2020.26.4.001