Designing and Implementing Conversational Intelligent Chat-bot Using Natural Language Processing

Authors

  • Asoke Nath  Department of Computer Science, St. Xavier's College (Autonomous), Kolkata, India
  • Rupamita Sarkar  Department of Computer Science, St. Xavier's College (Autonomous), Kolkata, India
  • Swastik Mitra  Department of Computer Science, St. Xavier's College (Autonomous), Kolkata, India
  • Rohitaswa Pradhan   Department of Computer Science, St. Xavier's College (Autonomous), Kolkata, India

DOI:

https://doi.org//10.32628/CSEIT217351

Keywords:

Natural Language Processing, Natural Language Understanding, Natural Language Generation, Deep Neural Networks, Artificial Intelligence, Transformer Model, Intelligent Agent, Chatbot.

Abstract

In the early days of Artificial Intelligence, it was observed that tasks which humans consider ‘natural’ and ‘commonplace’, such as Natural Language Understanding, Natural Language Generation and Vision were the most difficult task to carry over to computers. Nevertheless, attempts to crack the proverbial NLP nut were made, initially with methods that fall under ‘Symbolic NLP’. One of the products of this era was ELIZA. At present the most promising forays into the world of NLP are provided by ‘Neural NLP’, which uses Representation Learning and Deep Neural networks to model, understand and generate natural language. In the present paper the authors tried to develop a Conversational Intelligent Chatbot, a program that can chat with a user about any conceivable topic, without having domain-specific knowledge programmed into it. This is a challenging task, as it involves both ‘Natural Language Understanding’ (the task of converting natural language user input into representations that a machine can understand) and subsequently ‘Natural Language Generation’ (the task of generating an appropriate response to the user input in natural language). Several approaches exist for building conversational chatbots. In the present paper, two models have been used and their performance has been compared and contrasted. The first model is purely generative and uses a Transformer-based architecture. The second model is retrieval-based, and uses Deep Neural Networks.

References

  1. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, llia Polosukhin, “Attention is All You Need”, Page-2-6, 2017, Conference on Neural Information Processing Systems, Long Beach, CA, USA.
  2. Stephen Roller, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, Myle Ott, Kurt Shuster, Eric M. Smith, Y-Lan Boureau, Jason Weston, “Recipes for building an open-domain chatbot”, Facebook AI Research, 30th April, 2020.
  3. Humeau et al, Facebook AI research, 25th March, 2020.
  4. Miller et al, Facebook AI research, 2017.
  5. Denny Britz, “Recurrent Neural Network Tutorial, Part 4 – Implementing a GRU/LSTM RNN with Python and Theano.,” WILDML, 27-Oct-2015.
  6. Dzmitry Bahdanau, KyungHyun Cho, Yoshua Bengio, “Neural Machine Translation By Jointly Learning to Align and Translation”, Page-3-4, 2015, Conference paper at ICLR.
  7. “Neural Machine Translation with Attention.”, tensorflow.org.
  8. Amir Ali, Muhammad Zain Amin, “Conversational AI Chatbot Based on Encoder-Decoder Architectures with Attention Mechanism”, Page-8-10, 2019, Artificial Intelligence Festival 2.0, NED University of Engineering and Technology, Karachi, Pakistan.
  9. Abonia Sojasingarayar, “Seq2Seq AI Chatbot with Attention Mechanism”, Page-3-10, 2020, IA School/University, Boulogne-Billancourt, France.
  10. J.Prassanna, Khadar Nawas K, Christy Jackson J, Prabakaran R, Sakkaravarthi Ramanath, “Towards Building a Neural Conversation Chatbot through Seq2Seq Model”, International Journal of Scientific & Technology Research, ISSN 2277-8616, Vol-9, Issue-3, Page-2, 2020.
  11. A. Gillioz, J. Casas, E. Mugellini and O. A. Khaled, "Overview of the Transformer-based Models for NLP Tasks," 15th Conference on Computer Science and Information Systems (FedCSIS), Sofia, Bulgaria, 2020, pp. 179-183, doi: 10.15439/2020F20.
  12. Julien Chaumond, Patrick von Platen, Orestis Floros, Anthony MOI, Aditya Malte, “How to train a new language model from scratch using Transformers and Tokenizers”, blog/01_how_to_train.ipynb at master • huggingface/blog (github.com), 11th March, 2021.
  13. Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, Yejin Choi, “The Curious Case Of Neural Text DeGeneration”, Conference Paper at ICLR, 14th February 2020.
  14. Saizheng Zhang, Emily Dinan, Jack Urbanek, Arthur Szlam, Douwe Kiela, Jason Weston, “Personalizing Dialogue Agents: I have a dog, do you have pets too?”, Montreal Institute for Learning Algorithms, MILA and Facebook AI Research, 25th September.

Downloads

Published

2021-06-30

Issue

Section

Research Articles

How to Cite

[1]
Asoke Nath, Rupamita Sarkar, Swastik Mitra, Rohitaswa Pradhan , " Designing and Implementing Conversational Intelligent Chat-bot Using Natural Language Processing , IInternational Journal of Scientific Research in Computer Science, Engineering and Information Technology(IJSRCSEIT), ISSN : 2456-3307, Volume 7, Issue 3, pp.262-266, May-June-2021. Available at doi : https://doi.org/10.32628/CSEIT217351