Log in

E-commerce review sentiment score prediction considering misspelled words: a deep learning approach

  • Published:
Electronic Commerce Research Aims and scope Submit manuscript

Abstract

Acquiring a single sentiment score dependent on all the reviews will benefit the buyers and sellers in making the decision more accurately. The raw format of user-generated content lacks a legitimate language structure. It, therefore, acts as an obstacle for applying the Sentiment analysis task, which aims to predict the true emotion of a sentence by providing a score and its nature. This paper concentrates on obtaining a single sentiment score using a hybrid Long Short-Term Memory encoder–decoder model. This research uses the text normalization process to transform the sentences consisting of noise, appearing as incorrect grammar, abbreviations, freestyle, and typographical errors, into their canonical structure. The experimental outcomes confirm that the proposed hybrid model performs well in standardizing the raw E-commerce website review, enriched with hidden information and provided a single sentiment score influenced by all the review scores for the product.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Germany)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Notes

  1. https://sleeknote.com/blog/e-commerce-statistics#1.

  2. https://noisy-text.github.io/2015/.

  3. http://snap.stanford.edu/.

  4. https://pypi.org/project/vaderSentiment/.

  5. https://textblob.readthedocs.io/en/dev/.

  6. https://colab.research.google.com.

  7. https://keras.io.

References

  1. Singh, J. P., Irani, S., Rana, N. P., Dwivedi, Y. K., Saumya, S., & Roy, P. K. (2017). Predicting the ôhelpfulnessö of online consumer reviews. Journal of Business Research, 70, 346–355.

    Article  Google Scholar 

  2. Saumya, S., Singh, J. P., & Dwivedi, Y. K. (2020). Predicting the helpfulness score of online reviews using convolutional neural network. Soft Computing, 24(15), 10–11.

    Article  Google Scholar 

  3. Saumya, S., Singh, J. P., Baabdullah, A. M., Rana, N. P., & Dwivedi, Y. K. (2018). Ranking online consumer reviews. Electronic Commerce Research and Applications, 29, 78–89.

    Article  Google Scholar 

  4. Wang, Y., Wang, J., & Yao, T. (2019). What makes a helpful online review? A meta-analysis of review characteristics. Electronic Commerce Research, 19(2), 257–284.

    Article  Google Scholar 

  5. Syamala, M., & Nalini, N. J. (2020). A filter based improved decision tree sentiment classification model for real-time amazon product review data. International Journal of Intelligent Engineering and Systems, 13(1), 191–202.

    Article  Google Scholar 

  6. Kim, R. Y. (2020). When does online review matter to consumers? The effect of product quality information cues. Electronic Commerce Research, 2020, 1–20.

    Google Scholar 

  7. Sudheer, K., & Valarmathi, B. Real time sentiment analysis of e-commerce websites using machine learning algorithms.

  8. Vinodhini, G., & Chandrasekaran, R. (2016). A comparative performance evaluation of neural network based approach for sentiment classification of online reviews. Journal of King Saud University-Computer and Information Sciences, 28(1), 2–12.

    Article  Google Scholar 

  9. Bhatt, A., Patel, A., Chheda, H., & Gawande, K. (2015). Amazon review classification and sentiment analysis. International Journal of Computer Science and Information Technologies, 6(6), 5107–5110.

    Google Scholar 

  10. Zhang, L., Guo, D., Wen, X., & Li, Y. (2020). Effect of other visible reviews’ votes and personality on review helpfulness evaluation: an event-related potentials study. Electronic Commerce Research, 2020, 1–25.

    Google Scholar 

  11. Hu, M., & Liu, B. (2004). Mining opinion features in customer reviews. AAAI, 4(4), 755–760.

    Google Scholar 

  12. Satapathy, R., Li, Y., Cavallari, S., & Cambria, E. (2019). Seq2seq deep learning models for microtext normalization. In 2019 International joint conference on neural networks (IJCNN). IEEE (pp. 1–8).

  13. Matos Veliz, C., De Clercq, O., & Hoste, V. (2019). Benefits of data augmentation for nmt-based text normalization of user-generated content. In 2019 conference on empirical methods in natural language processing and 9th international joint conference on natural language processing. Association for Computational Linguistics (ACL) (pp. 275–285).

  14. Saito, I., Suzuki, J., Nishida, K., Sadamitsu, K., Kobashikawa, S., Masumura, R., Matsumoto, Y., & Tomita, J. (2017). Improving neural text normalization with data augmentation at character-and morphological levels. Proceedings of the Eighth International Joint Conference on Natural Language Processing (Short Papers), 2, 257–262.

    Google Scholar 

  15. Bornás, A. J., & Mateos, G. G. (2019). A character-level approach to the text normalization problem based on a new causal encoder. ar**v preprint ar**v:1903.02642.

  16. Javaloy, A., & García-Mateos, G. (2020). Text normalization using encoder-decoder networks based on the causal feature extractor. Applied Sciences, 10(13), 4551.

    Article  Google Scholar 

  17. Bollmann, M., Bingel, J., & Søgaard, A. (2017). Learning attention for historical text normalization by learning to pronounce. In Proceedings of the 55th annual meeting of the association for computational linguistics (Vol. 1: Long Papers, pp. 332–344).

  18. Lusetti, M., Ruzsics, T., Göhring, A., Samardžić, T., & Stark, E. (2018). Encoder–decoder methods for text normalization. Association for Computational Linguistics, 2018, 18–28.

    Google Scholar 

  19. Li Z., Specia, L. (2019). Improving neural machine translation robustness via data augmentation: Beyond back translation. ar**v preprint ar**v:1910.03009.

  20. Mansfield, C., Sun, M., Liu, Y., Gandhe, A., & Hoffmeister, B. (2019). Neural text normalization with subword units. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies (Vol. 2 (Industry Papers), pp. 190–196).

  21. Zhang, J., Pan, J., Yin, X., Li, C., Liu, S., Zhang, Y., Wang, Y., & Ma Z. (2020). A hybrid text normalization system using multi-head self-attention for mandarin. In ICASSP 2020-2020 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE (pp. 6694–6698).

  22. Min, W., & Mott, B. (2015). Ncsu_sas_wookhee: A deep contextual long-short term memory model for text normalization. In Proceedings of the workshop on noisy user-generated text (pp. 111–119).

  23. Arora, M., & Kansal, V. (2019). Character level embedding with deep convolutional neural network for text normalization of unstructured data for twitter sentiment analysis. Social Network Analysis and Mining, 9(1), 1–14.

    Article  Google Scholar 

  24. Pennell, D., & Liu, Y. (2011). A character-level machine translation approach for normalization of sms abbreviations. In Proceedings of 5th international joint conference on natural language processing (pp. 974–982).

  25. Mager, M., Rosales, M. J., Çetinoğlu, Ö., & Meza, I. (2019). Low-resource neural character-based noisy text normalization. Journal of Intelligent & Fuzzy Systems, 36(5), 4921–4929.

    Article  Google Scholar 

  26. Elli, M. S., & Wang, Y.-F. (2016). Amazon reviews, business analytics with sentiment analysis. Elwalda, Abdulaziz, et al. Perceived Derived Attributes of Online Customer Reviews.

  27. Baldwin, T., de Marneffe, M.-C., Han, B., Kim, Y.-B., Ritter, A., & Xu, W. (2015). Shared tasks of the 2015 workshop on noisy user-generated text: Twitter lexical normalization and named entity recognition. In Proceedings of the workshop on noisy user-generated text (pp. 126–135).

  28. Li, C., & Liu, Y. (2012). Improving text normalization using character-blocks based models and system combination. In Proceedings of COLING, 2012 (pp. 1587–1602).

  29. Yolchuyeva, S., Németh, G., & Gyires-Tóth, B. (2018). Text normalization with convolutional neural networks. International Journal of Speech Technology, 21(3), 589–600.

    Article  Google Scholar 

  30. Sproat, R., Black, A. W., Chen, S., Kumar, S., Ostendorf, M., & Richards, C. (2001). Normalization of non-standard words. Computer speech & language, 15(3), 287–333.

    Article  Google Scholar 

  31. Sonmez, C., & Özgür, A. (2014). A graph-based approach for contextual text normalization. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 313–324).

  32. Sproat, R., & Jaitly, N. (2017). An rnn model of text normalization. In INTERSPEECH. Stockholm (pp. 754–758).

  33. Ruzsics, T., & Samardžić, T. (2019). Multilevel text normalization with sequence-to-sequence networks and multisource learning. ar**v preprint ar**v:1903.11340.

  34. Ruzsics, T., Lusetti, M., Göhring, A., Samardzic, T., & Stark, E. (2019). Neural text normalization with adapted decoding and pos features. Natural Language Engineering, 25(5), 585–605.

    Article  Google Scholar 

  35. Nguyen, H., & Cavallari, S. (2020). Neural multi-task text normalization and sanitization with pointer-generator. In Proceedings of the first workshop on natural language interfaces (pp. 37–47).

  36. Tiwari, A. S., & Naskar, S. K. (2017). Normalization of social media text using deep neural networks. In Proceedings of the 14th international conference on natural language processing (ICON-2017) (pp. 312–321).

  37. Watson, D., Zalmout, N., & Habash, N. (2018). Utilizing character and word embeddings for text normalization with sequence-to-sequence models. ar**v preprint ar**v:1809.01534.

  38. Lourentzou, I., Manghnani, K., & Zhai, C. (2019) Adapting sequence to sequence models for text normalization in social media. In Proceedings of the international AAAI conference on web and social media (Vol. 13, pp. 335–345).

  39. Hochreiter, S., & Schmidhuber, J. (1997). Lstm can solve hard long time lag problems. In Advances in neural information processing systems (pp. 473–479).

  40. Bahdanau, D., Cho, K., & Bengio, Y. (2014). Neural machine translation by jointly learning to align and translate. ar**v preprint ar**v:1409.0473.

  41. Luong, M.-T., Pham, H., & Manning, C. D. (2015). Effective approaches to attention-based neural machine translation. ar**v preprint ar**v:1508.04025.

  42. Huang, Z., Xu, W., & Yu, K. (2015). Bidirectional lstm-crf models for sequence tagging. ar**v preprint ar**v:1508.01991.

  43. Yu, L., Zhang, W., Wang, J., & Yu, Y. (2017). Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI conference on artificial intelligence (vol. 31, no. 1).

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pradeep Kumar Roy.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jain, S., Roy, P.K. E-commerce review sentiment score prediction considering misspelled words: a deep learning approach. Electron Commer Res (2022). https://doi.org/10.1007/s10660-022-09582-4

Download citation

  • Published:

  • DOI: https://doi.org/10.1007/s10660-022-09582-4

Keywords

Navigation