My_Eliza; a multimodal communication system

title: My_Eliza; a multimodal communication system
author: Siska Fitrianie
published in: August 2002
appeared as: Master of Science thesis
Delft University of Technology
pages 187
PDF (3.596 KB)

Abstract

The field Natural Language Processing (NLP) is an active area of research. One of the possible applications of NLP systems is a question answering system - QA system. QA system is a program that attempts to simulate typed conversation in human natural language. In contrast with it, there are only a few researches involving research in Nonverbal Communication. Even though, it is proved that beyond the speech bodily activities give real substance to face-to-face interaction in real life conversation.

Motivated by this issue, in this report we investigate a design and implementation of a multimodal communication system, called my_Eliza. The design of this system is based on Weizenbaum's Eliza. A user or client can communicate with the system using typed natural language. The system will reply by text-prompts and appropriate facial-expressions. In order to communicate using a nonverbal facial display, the developed system should be able to process natural language and emotional reasoning.

A first prototype as a proof of concept has been developed that consists of a dialog box, an emotional recognizer based on stimulus response, and a facial display generator. To implement the dialog box, the work of Richard Wallace, A.L.I.C.E system, has been used as a starting point. My_Eliza system has a rule engine that determines current system's affective state as reaction to the user's string input and conversation content. This report presents the result of a conversation between my_Eliza and a human user.

 
blue line
University logo