Dissertation/Thesis Abstract

Towards Modeling Collaborative Task Oriented Multimodal Human-human Dialogues
by Chen, Lin, Ph.D., University of Illinois at Chicago, 2014, 134; 3668932
Abstract (Summary)

This research took place in the larger context of building effective multimodal interfaces to help elderly people live independently. The final goal was to build a dialogue manager which could be deployed on a robot. The robot would help elderly people perform Activities of Daily Living (ADLs), such as cooking dinner, and setting a table. In particular, I focused on building dialogue processing modules to understand such multimodal dialogues. Specifically, I investigated the functions of gestures (e.g. Pointing Gestures, and Haptic-Ostensive actions which involve force exchange) in dialogues concerning collaborative tasks in ADLs.

This research employed an empirical approach. The machine learning based modules were built using collected human experiment data. The ELDERLY-AT-HOME corpus was built based on a data collection of human-human collaborative interactions in the elderly care domain. Multiple categories of annotations were further conducted to build the Find corpus, which only contained the experiment episodes where two subjects were collaboratively searching for objects (e.g. a pot, a spoon, etc.), which are essential tasks to perform ADLs.

This research developed three main modules: coreference resolution, Dialogue Act classification, and task state inference. The coreference resolution experiments showed that modalities other than language play an important role in bringing antecedents into the dialogue context. The Dialogue Act classification experiments showed that multimodal features including gestures, Haptic-Ostensive actions, and subject location significantly improve accuracy. They also showed that dialogue games help improve performance, even if the dialogue games were inferred dynamically. A heuristic rule-based task state inference system using the results of Dialogue Act classification and coreference resolution was designed and evaluated; the experiments showed reasonably good results.

Compared to previous work, the contributions of this research are as follows: 1) Built a multimodal corpus focusing on human-human collaborative task-oriented dialogues. 2) Investigated coreference resolution from language to objects in the real world. 3) Experimented with Dialogue Act classification using utterances, gestures and Haptic-Ostensive actions. 4) Implemented and evaluated a task state inference system.

Indexing (document details)
Advisor: Di Eugenio, Barbara
Commitee: Chai, Joyce Y., Di Eugenio, Barbara, Gmytrasiewicz, Piotr, Leigh, Jason, Zefran, Milos
School: University of Illinois at Chicago
Department: Computer Science
School Location: United States -- Illinois
Source: DAI-B 76/05(E), Dissertation Abstracts International
Source Type: DISSERTATION
Subjects: Artificial intelligence, Computer science
Keywords: Activities of Daily Living, Coreference resolution, Dialogue Act classification, Elderly care domain, Haptic-Ostensive actions, Multimodal dialogue system
Publication Number: 3668932
ISBN: 9781321443240
Copyright © 2019 ProQuest LLC. All rights reserved. Terms and Conditions Privacy Policy Cookie Policy
ProQuest