Skip to main content

Towards Automated Customer Support

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 11089))

Abstract

Recent years have seen growing interest in conversational agents, such as chatbots, which are a very good fit for automated customer support because the domain in which they need to operate is narrow. This interest was in part inspired by recent advances in neural machine translation, esp. the rise of sequence-to-sequence (seq2seq) and attention-based models such as the Transformer, which have been applied to various other tasks and have opened new research directions in question answering, chatbots, and conversational systems. Still, in many cases, it might be feasible and even preferable to use simple information retrieval techniques. Thus, here we compare three different models: (i) a retrieval model, (ii) a sequence-to-sequence model with attention, and (iii) Transformer. Our experiments with the Twitter Customer Support Dataset, which contains over two million posts from customer support services of twenty major brands, show that the seq2seq model outperforms the other two in terms of semantics and word overlap.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    https://www.kaggle.com/thoughtvector/customer-support-on-twitter.

  2. 2.

    By design, tweets have been strictly limited to 140 characters; this constrain has been relaxed to 280 characters in 2017.

  3. 3.

    In future work, we plan to try byte-pair encoding instead [21].

  4. 4.

    https://www.elastic.co/products/elasticsearch.

  5. 5.

    https://nlp.stanford.edu/projects/glove/.

  6. 6.

    Note that we do not use measures trained on the same data as advised by [10].

References

  1. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)

  2. Banerjee, S., Lavie, A.: METEOR: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, Ann Arbor, Michigan, pp. 65–72 (2005)

    Google Scholar 

  3. Boyanov, M., Nakov, P., Moschitti, A., Da San Martino, G., Koychev, I.: Building chatbots from forum data: model selection using question answering metrics. In: Proceedings of the International Conference Recent Advances in Natural Language Processing, RANLP 2017, Varna, Bulgaria, pp. 121–129 (2017)

    Google Scholar 

  4. Cui, L., Huang, S., Wei, F., Tan, C., Duan, C., Zhou, M.: SuperAgent: a customer service chatbot for e-commerce websites. In: Proceedings of the Association for Computational Linguistics 2017, System Demonstrations, ACL 2017, Vancouver, Canada, pp. 97–102 (2017)

    Google Scholar 

  5. Forgues, G., Pineau, J., Larchevêque, J.M., Tremblay, R.: Bootstrapping dialog systems with word embeddings. In: Proceedings of the NIPS Workshop on Modern Machine Learning and Natural Language Processing, Montreal, Canada (2014)

    Google Scholar 

  6. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  7. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the 2015 International Conference on Learning Representations, ICLR 2015, San Diego, California (2015)

    Google Scholar 

  8. Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Proceedings of the ACL Workshop on Text Summarization Branches Out, Barcelona, Spain, pp. 74–81 (2004)

    Google Scholar 

  9. Lin, C.Y., Och, F.J.: Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics. In: Proceedings of the 42nd Annual Conference of the Association for Computational Linguistics, ACL 2004, Barcelona, Spain, pp. 605–612 (2004)

    Google Scholar 

  10. Liu, C.W., Lowe, R., Serban, I., Noseworthy, M., Charlin, L., Pineau, J.: How NOT to evaluate your dialogue system: an empirical study of unsupervised evaluation metrics for dialogue response generation. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, Texas, pp. 2122–2132 (2016)

    Google Scholar 

  11. Lowe, R., Noseworthy, M., Serban, I.V., Angelard-Gontier, N., Bengio, Y., Pineau, J.: Towards an automatic Turing test: learning to evaluate dialogue responses. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, pp. 1116–1126 (2017)

    Google Scholar 

  12. Lowe, R., Pow, N., Serban, I., Pineau, J.: The Ubuntu dialogue corpus: a large dataset for research in unstructured multi-turn dialogue systems. In: Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL 2015, Prague, Czech Republic, pp. 285–294 (2015)

    Google Scholar 

  13. Lowe, R.T., Pow, N., Serban, I.V., Charlin, L., Liu, C.W., Pineau, J.: Training end-to-end dialogue systems with the Ubuntu dialogue corpus. Dialogue Discourse 8(1), 31–65 (2017)

    Google Scholar 

  14. Manning, C., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S., McClosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, ACL 2014, Baltimore, Maryland, pp. 55–60 (2014)

    Google Scholar 

  15. Nakov, P., et al.: SemEval-2016 task 3: community question answering. In: Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval 2016, San Diego, California, pp. 525–545 (2016)

    Google Scholar 

  16. Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, ACL 2002, Philadelphia, Pennsylvania, pp. 311–318 (2002)

    Google Scholar 

  17. Pennington, J., Socher, R., Manning, C.: GloVe: global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, Doha, Qatar, pp. 1532–1543 (2014)

    Google Scholar 

  18. Qiu, M., et al.: AliMe Chat: a sequence to sequence and rerank based chatbot engine. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, pp. 498–503 (2017)

    Google Scholar 

  19. Robertson, S., Zaragoza, H.: The probabilistic relevance framework: BM25 and beyond. Found. Trends Inf. Retr. 3(4), 333–389 (2009)

    Article  Google Scholar 

  20. Rus, V., Lintean, M.: A comparison of greedy and optimal assessment of natural language student input using word-to-word similarity metrics. In: Proceedings of the Seventh Workshop on Building Educational Applications Using NLP, Montreal, Canada, pp. 157–162 (2012)

    Google Scholar 

  21. Sennrich, R., Haddow, B., Birch, A.: Neural machine translation of rare words with subword units. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016, pp. 1715–1725 (2016)

    Google Scholar 

  22. Serban, I.V., Lowe, R., Henderson, P., Charlin, L., Pineau, J.: A survey of available corpora for building data-driven dialogue systems: the journal version. Dialogue Discourse 9(1), 1–49 (2018)

    Google Scholar 

  23. Serban, I.V., Sordoni, A., Bengio, Y., Courville, A.C., Pineau, J.: Hierarchical neural network generative models for movie dialogues. CoRR, abs/1507.04808 (2015)

    Google Scholar 

  24. Shang, L., Lu, Z., Li, H.: Neural responding machine for short-text conversation. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, ACL-IJCNLP 2015, Beijing, China, pp. 1577–1586 (2015)

    Google Scholar 

  25. Sordoni, A., et al.: A neural network approach to context-sensitive generation of conversational responses. In: Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2015, Denver, Colorado, pp. 196–205 (2015)

    Google Scholar 

  26. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)

    MathSciNet  MATH  Google Scholar 

  27. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Proceedings of the 27th Annual Conference on Neural Information Processing Systems, NIPS 2014, Montreal, Canada, pp. 3104–3112 (2014)

    Google Scholar 

  28. Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 30th Annual Conference on Neural Information Processing Systems, NIPS 2017, Long Beach, California, pp. 5998–6008 (2017)

    Google Scholar 

  29. Vinyals, O., Le, Q.V.: A neural conversational model. CoRR abs/1506.05869 (2015)

    Google Scholar 

Download references

Acknowledgments

This work was supported by the EC under grant no. 763566 and by the Bulgarian National Scientific Fund as project no. DN 12/9,

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Momchil Hardalov .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Hardalov, M., Koychev, I., Nakov, P. (2018). Towards Automated Customer Support. In: Agre, G., van Genabith, J., Declerck, T. (eds) Artificial Intelligence: Methodology, Systems, and Applications. AIMSA 2018. Lecture Notes in Computer Science(), vol 11089. Springer, Cham. https://doi.org/10.1007/978-3-319-99344-7_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-99344-7_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-99343-0

  • Online ISBN: 978-3-319-99344-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics