The Story Of Computational Narratology

2018 ◽  
Author(s):  
Sharath Srivatsa ◽  
Shyam Kumar V N ◽  
Srinath Srinivasa

In recent times, computational modeling of narratives has gained enormous interest in fields like Natural Language Understanding (NLU), Natural Language Generation (NLG), and Artificial General Intelligence (AGI). There is a growing body of literature addressing understanding of narrative structure and generation of narratives. Narrative generation is known to be a far more complex problem than narrative understanding [20].

2020 ◽  
Vol 8 (6) ◽  
pp. 3281-3287

Text is an extremely rich resources of information. Each and every second, minutes, peoples are sending or receiving hundreds of millions of data. There are various tasks involved in NLP are machine learning, information extraction, information retrieval, automatic text summarization, question-answered system, parsing, sentiment analysis, natural language understanding and natural language generation. The information extraction is an important task which is used to find the structured information from unstructured or semi-structured text. The paper presents a methodology for extracting the relations of biomedical entities using spacy. The framework consists of following phases such as data creation, load and converting the data into spacy object, preprocessing, define the pattern and extract the relations. The dataset is downloaded from NCBI database which contains only the sentences. The created model evaluated with performance measures like precision, recall and f-measure. The model achieved 87% of accuracy in retrieving of entities relation.


Author(s):  
Andrew M. Olney ◽  
Natalie K. Person ◽  
Arthur C. Graesser

The authors discuss Guru, a conversational expert ITS. Guru is designed to mimic expert human tutors using advanced applied natural language processing techniques including natural language understanding, knowledge representation, and natural language generation.


Author(s):  
Zhuoxuan Jiang ◽  
Jie Ma ◽  
Jingyi Lu ◽  
Guangyuan Yu ◽  
Yipeng Yu ◽  
...  

We propose a general framework for goal-driven conversation assistant based on Planning methods. It aims to rapidly build a dialogue agent with less handcrafting and make the more interpretable and efficient dialogue management in various scenarios. By employing the Planning method, dialogue actions can be efficiently defined and reusable, and the transition of the dialogue are managed by a Planner. The proposed framework consists of a pipeline of Natural Language Understanding (intent labeler), Planning of Actions (with a World Model), and Natural Language Generation (learned by an attention-based neural network). We demonstrate our approach by creating conversational agents for several independent domains.


Author(s):  
Hima Yeldo

Abstract: Natural Language Processing is the study that focuses the interplay between computer and the human languages NLP has spread its applications in various fields such as an email Spam detection, machine translation, summation, information extraction, and question answering etc. Natural Language Processing classifies two parts i.e. Natural Language Generation and Natural Language understanding which evolves the task to generate and understand the text.


Author(s):  
Shubham Parmar ◽  
Megha Meshram ◽  
Parth Parmar ◽  
Meet Patel ◽  
Payal Desai

Intelligent Chatbot, Natural Language Understanding, Natural Language Generation, NLP, WIT, API, LUIS


2021 ◽  
Author(s):  
Philippe Blache ◽  
Matthis Houlès

This paper presents a dialogue system for training doctors to break bad news. The originality of this work lies in its knowledge representation. All information known before the dialogue (the universe of discourse, the context, the scenario of the dialogue) as well as the knowledge transferred from the doctor to the patient during the conversation is represented in a shared knowledge structure called common ground, that constitute the core of the system. The Natural Language Understanding and the Natural Language Generation modules of the system take advantage on this structure and we present in this paper different original techniques making it possible to implement them efficiently.


Author(s):  
Asoke Nath ◽  
Rupamita Sarkar ◽  
Swastik Mitra ◽  
Rohitaswa Pradhan

In the early days of Artificial Intelligence, it was observed that tasks which humans consider ‘natural’ and ‘commonplace’, such as Natural Language Understanding, Natural Language Generation and Vision were the most difficult task to carry over to computers. Nevertheless, attempts to crack the proverbial NLP nut were made, initially with methods that fall under ‘Symbolic NLP’. One of the products of this era was ELIZA. At present the most promising forays into the world of NLP are provided by ‘Neural NLP’, which uses Representation Learning and Deep Neural networks to model, understand and generate natural language. In the present paper the authors tried to develop a Conversational Intelligent Chatbot, a program that can chat with a user about any conceivable topic, without having domain-specific knowledge programmed into it. This is a challenging task, as it involves both ‘Natural Language Understanding’ (the task of converting natural language user input into representations that a machine can understand) and subsequently ‘Natural Language Generation’ (the task of generating an appropriate response to the user input in natural language). Several approaches exist for building conversational chatbots. In the present paper, two models have been used and their performance has been compared and contrasted. The first model is purely generative and uses a Transformer-based architecture. The second model is retrieval-based, and uses Deep Neural Networks.


2020 ◽  
Vol 34 (05) ◽  
pp. 8327-8335
Author(s):  
Weixin Liang ◽  
Youzhi Tian ◽  
Chengcai Chen ◽  
Zhou Yu

A major bottleneck in training end-to-end task-oriented dialog system is the lack of data. To utilize limited training data more efficiently, we propose Modular Supervision Network (MOSS), an encoder-decoder training framework that could incorporate supervision from various intermediate dialog system modules including natural language understanding, dialog state tracking, dialog policy learning and natural language generation. With only 60% of the training data, MOSS-all (i.e., MOSS with supervision from all four dialog modules) outperforms state-of-the-art models on CamRest676. Moreover, introducing modular supervision has even bigger benefits when the dialog task has a more complex dialog state and action space. With only 40% of the training data, MOSS-all outperforms the state-of-the-art model on a complex laptop network trouble shooting dataset, LaptopNetwork, that we introduced. LaptopNetwork consists of conversations between real customers and customer service agents in Chinese. Moreover, MOSS framework can accommodate dialogs that have supervision from different dialog modules at both framework level and model level. Therefore, MOSS is extremely flexible to update in real-world deployment.


2020 ◽  
Vol 170 ◽  
pp. 02006
Author(s):  
Ravindra Parshuram Bachate ◽  
Ashok Sharma

Natural Language Processing (NLP) deals with the spoken languages by using computer and Artificial Intelligence. As people from different regional areas using different digital platforms and expressing their views in their spoken language, it is now must to focus on working spoken languages in India to make our society smart and digital. NLP research grown tremendously in last decade which results in Siri, Google Assistant, Alexa, Cortona and many more automatic speech recognitions and understanding systems (ASR). Natural Language Processing can be understood by classifying it into Natural Language Generation and Natural Language Understanding. NLP is widely used in various domain such as Health Care, Chatbot, ASR building, HR, Sentiment analysis etc.


Sign in / Sign up

Export Citation Format

Share Document