Abstract
Acquiring a single sentiment score dependent on all the reviews will benefit the buyers and sellers in making the decision more accurately. The raw format of user-generated content lacks a legitimate language structure. It, therefore, acts as an obstacle for applying the Sentiment analysis task, which aims to predict the true emotion of a sentence by providing a score and its nature. This paper concentrates on obtaining a single sentiment score using a hybrid Long Short-Term Memory encoder–decoder model. This research uses the text normalization process to transform the sentences consisting of noise, appearing as incorrect grammar, abbreviations, freestyle, and typographical errors, into their canonical structure. The experimental outcomes confirm that the proposed hybrid model performs well in standardizing the raw E-commerce website review, enriched with hidden information and provided a single sentiment score influenced by all the review scores for the product.
Similar content being viewed by others
References
Singh, J. P., Irani, S., Rana, N. P., Dwivedi, Y. K., Saumya, S., & Roy, P. K. (2017). Predicting the ôhelpfulnessö of online consumer reviews. Journal of Business Research, 70, 346–355.
Saumya, S., Singh, J. P., & Dwivedi, Y. K. (2020). Predicting the helpfulness score of online reviews using convolutional neural network. Soft Computing, 24(15), 10–11.
Saumya, S., Singh, J. P., Baabdullah, A. M., Rana, N. P., & Dwivedi, Y. K. (2018). Ranking online consumer reviews. Electronic Commerce Research and Applications, 29, 78–89.
Wang, Y., Wang, J., & Yao, T. (2019). What makes a helpful online review? A meta-analysis of review characteristics. Electronic Commerce Research, 19(2), 257–284.
Syamala, M., & Nalini, N. J. (2020). A filter based improved decision tree sentiment classification model for real-time amazon product review data. International Journal of Intelligent Engineering and Systems, 13(1), 191–202.
Kim, R. Y. (2020). When does online review matter to consumers? The effect of product quality information cues. Electronic Commerce Research, 2020, 1–20.
Sudheer, K., & Valarmathi, B. Real time sentiment analysis of e-commerce websites using machine learning algorithms.
Vinodhini, G., & Chandrasekaran, R. (2016). A comparative performance evaluation of neural network based approach for sentiment classification of online reviews. Journal of King Saud University-Computer and Information Sciences, 28(1), 2–12.
Bhatt, A., Patel, A., Chheda, H., & Gawande, K. (2015). Amazon review classification and sentiment analysis. International Journal of Computer Science and Information Technologies, 6(6), 5107–5110.
Zhang, L., Guo, D., Wen, X., & Li, Y. (2020). Effect of other visible reviews’ votes and personality on review helpfulness evaluation: an event-related potentials study. Electronic Commerce Research, 2020, 1–25.
Hu, M., & Liu, B. (2004). Mining opinion features in customer reviews. AAAI, 4(4), 755–760.
Satapathy, R., Li, Y., Cavallari, S., & Cambria, E. (2019). Seq2seq deep learning models for microtext normalization. In 2019 International joint conference on neural networks (IJCNN). IEEE (pp. 1–8).
Matos Veliz, C., De Clercq, O., & Hoste, V. (2019). Benefits of data augmentation for nmt-based text normalization of user-generated content. In 2019 conference on empirical methods in natural language processing and 9th international joint conference on natural language processing. Association for Computational Linguistics (ACL) (pp. 275–285).
Saito, I., Suzuki, J., Nishida, K., Sadamitsu, K., Kobashikawa, S., Masumura, R., Matsumoto, Y., & Tomita, J. (2017). Improving neural text normalization with data augmentation at character-and morphological levels. Proceedings of the Eighth International Joint Conference on Natural Language Processing (Short Papers), 2, 257–262.
Bornás, A. J., & Mateos, G. G. (2019). A character-level approach to the text normalization problem based on a new causal encoder. ar**v preprint ar**v:1903.02642.
Javaloy, A., & García-Mateos, G. (2020). Text normalization using encoder-decoder networks based on the causal feature extractor. Applied Sciences, 10(13), 4551.
Bollmann, M., Bingel, J., & Søgaard, A. (2017). Learning attention for historical text normalization by learning to pronounce. In Proceedings of the 55th annual meeting of the association for computational linguistics (Vol. 1: Long Papers, pp. 332–344).
Lusetti, M., Ruzsics, T., Göhring, A., Samardžić, T., & Stark, E. (2018). Encoder–decoder methods for text normalization. Association for Computational Linguistics, 2018, 18–28.
Li Z., Specia, L. (2019). Improving neural machine translation robustness via data augmentation: Beyond back translation. ar**v preprint ar**v:1910.03009.
Mansfield, C., Sun, M., Liu, Y., Gandhe, A., & Hoffmeister, B. (2019). Neural text normalization with subword units. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies (Vol. 2 (Industry Papers), pp. 190–196).
Zhang, J., Pan, J., Yin, X., Li, C., Liu, S., Zhang, Y., Wang, Y., & Ma Z. (2020). A hybrid text normalization system using multi-head self-attention for mandarin. In ICASSP 2020-2020 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE (pp. 6694–6698).
Min, W., & Mott, B. (2015). Ncsu_sas_wookhee: A deep contextual long-short term memory model for text normalization. In Proceedings of the workshop on noisy user-generated text (pp. 111–119).
Arora, M., & Kansal, V. (2019). Character level embedding with deep convolutional neural network for text normalization of unstructured data for twitter sentiment analysis. Social Network Analysis and Mining, 9(1), 1–14.
Pennell, D., & Liu, Y. (2011). A character-level machine translation approach for normalization of sms abbreviations. In Proceedings of 5th international joint conference on natural language processing (pp. 974–982).
Mager, M., Rosales, M. J., Çetinoğlu, Ö., & Meza, I. (2019). Low-resource neural character-based noisy text normalization. Journal of Intelligent & Fuzzy Systems, 36(5), 4921–4929.
Elli, M. S., & Wang, Y.-F. (2016). Amazon reviews, business analytics with sentiment analysis. Elwalda, Abdulaziz, et al. Perceived Derived Attributes of Online Customer Reviews.
Baldwin, T., de Marneffe, M.-C., Han, B., Kim, Y.-B., Ritter, A., & Xu, W. (2015). Shared tasks of the 2015 workshop on noisy user-generated text: Twitter lexical normalization and named entity recognition. In Proceedings of the workshop on noisy user-generated text (pp. 126–135).
Li, C., & Liu, Y. (2012). Improving text normalization using character-blocks based models and system combination. In Proceedings of COLING, 2012 (pp. 1587–1602).
Yolchuyeva, S., Németh, G., & Gyires-Tóth, B. (2018). Text normalization with convolutional neural networks. International Journal of Speech Technology, 21(3), 589–600.
Sproat, R., Black, A. W., Chen, S., Kumar, S., Ostendorf, M., & Richards, C. (2001). Normalization of non-standard words. Computer speech & language, 15(3), 287–333.
Sonmez, C., & Özgür, A. (2014). A graph-based approach for contextual text normalization. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 313–324).
Sproat, R., & Jaitly, N. (2017). An rnn model of text normalization. In INTERSPEECH. Stockholm (pp. 754–758).
Ruzsics, T., & Samardžić, T. (2019). Multilevel text normalization with sequence-to-sequence networks and multisource learning. ar**v preprint ar**v:1903.11340.
Ruzsics, T., Lusetti, M., Göhring, A., Samardzic, T., & Stark, E. (2019). Neural text normalization with adapted decoding and pos features. Natural Language Engineering, 25(5), 585–605.
Nguyen, H., & Cavallari, S. (2020). Neural multi-task text normalization and sanitization with pointer-generator. In Proceedings of the first workshop on natural language interfaces (pp. 37–47).
Tiwari, A. S., & Naskar, S. K. (2017). Normalization of social media text using deep neural networks. In Proceedings of the 14th international conference on natural language processing (ICON-2017) (pp. 312–321).
Watson, D., Zalmout, N., & Habash, N. (2018). Utilizing character and word embeddings for text normalization with sequence-to-sequence models. ar**v preprint ar**v:1809.01534.
Lourentzou, I., Manghnani, K., & Zhai, C. (2019) Adapting sequence to sequence models for text normalization in social media. In Proceedings of the international AAAI conference on web and social media (Vol. 13, pp. 335–345).
Hochreiter, S., & Schmidhuber, J. (1997). Lstm can solve hard long time lag problems. In Advances in neural information processing systems (pp. 473–479).
Bahdanau, D., Cho, K., & Bengio, Y. (2014). Neural machine translation by jointly learning to align and translate. ar**v preprint ar**v:1409.0473.
Luong, M.-T., Pham, H., & Manning, C. D. (2015). Effective approaches to attention-based neural machine translation. ar**v preprint ar**v:1508.04025.
Huang, Z., Xu, W., & Yu, K. (2015). Bidirectional lstm-crf models for sequence tagging. ar**v preprint ar**v:1508.01991.
Yu, L., Zhang, W., Wang, J., & Yu, Y. (2017). Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI conference on artificial intelligence (vol. 31, no. 1).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Jain, S., Roy, P.K. E-commerce review sentiment score prediction considering misspelled words: a deep learning approach. Electron Commer Res (2022). https://doi.org/10.1007/s10660-022-09582-4
Published:
DOI: https://doi.org/10.1007/s10660-022-09582-4