• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to project page

1997 Fiscal Year Final Research Report Summary

Studies on Multimodal Communication by Integrating Speech and Diagram

Research Project

Project/Area Number 08458078
Research Category

Grant-in-Aid for Scientific Research (B)

Allocation TypeSingle-year Grants
Section一般
Research Field Intelligent informatics
Research InstitutionKYOTO UNIVERSITY

Principal Investigator

DOSHITA Shuji  Kyoto University, Graduate School of Engineering, Professor, 工学研究科, 教授 (00025925)

Co-Investigator(Kenkyū-buntansha) ARAKI Masahiro  Kyoto University.Center for for Information and Multimedia Studies.Lecturer, 総合情報メディアセンター, 講師 (50252490)
KAWAHARA Tatsuya  Kyoto University.Graduate School of Engineering.Assistant Professor, 工学研究科, 助教授 (00234104)
Project Period (FY) 1996 – 1997
KeywordsMultimodal Communication / Dialogue / Speech Recognition / Gesture / Drawing System / Information Integration
Research Abstract

The aim of this research is to realize multimodal communication between human and machines through implementing multimodal drawing system with speech input. The results of this research are below :
1. Integration of speech, gesture and diagram
We developed multi-mode integration method without assuming the synchronization of each input elements. The validity of this integration method was examined by evaluating the improvement of accuracy of information input in multimodal dictation system.
2. Implementation of multimodal interface using speech and pointing gesture
We developed the cooperation mechanism which uses the result of pointing gesture recognition, which is reliable channel, for the recognition of speech, which is relatively unreliable channel. By using the result of pointing gesture recognition, the accuracy of speech recognition raised about 6%. Using this integration method, we implemented a multimodal drawing system with speech input.
3. Interpretation of utterance using diagrammatic and contextual information in multimodal drawing system We realized utterance understanding mechanism which uses context management mechanism by stack in order to deal with spontaneous speech (e. g. ellipsis, reference expression, etc.). We integrated this utterance understanding mechanism to multimodal drawing system.

  • Research Products

    (16 results)

All Other

All Publications (16 results)

  • [Publications] Araki, M.: "Evaluating Dialogue Strategies under Various Communication Errors" Proc.of IJCAI Workshop on Collaboration,Cooperation and Conflict in Dialogue Systems. 13-18 (1997)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] 荒木 雅弘: "マルチモーダル作図システムにおける文脈知識を利用した会話理解" 人工知能学会研究会資料SIG-SLUD-9703-1. (1998)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] 河原 達也: "音声言語を用いた仮想空間との対話による試着システム" 情報処理学会論文誌. 採録決定 (1998)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] Kawahara, T.: "Speaking-style dependent lexicalized filler model for key-phrase detection and verification" 電子情報通信学会技術研究報告. SP97-78. (1997)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] Kawahara, T.: "Combining key-phrase detection and subword-based verification for flexible speech understanding" Proc.of ICASSP97. 1159-1162 (1997)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] Kawahara, T.: "Phrase language models for detection and verification-based speech understanding" Proc.IEEE Workshop on Automatic Speech Recognition and Understanding. 49-56 (1997)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] Araki, M.: ""Automatic Evaluation Environment for Spoken Dialogue Systems"in Mayer,E.et al.ed."Dialogue Processing in Spoken Language Systems"" Springer-Verlag, 12 (1997)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] 堂下 修司: "「音声による人間と機械の対話」" オーム社, 383 (1998)

    • Description
      「研究成果報告書概要(和文)」より
  • [Publications] Araki, M.: "Evaluating Dialogue Strategies under Various Communication Errors" Proc.of IJCAI Workshop on Collaboration, Cooperation and Conflict in Dialogue Systems. 13-18 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Araki, M.: "Interpretation of Utterance on Multi-modal Drawing System based on Contextual Knowledge" Tech.report of IJSAI. SIG-SLUD-9703-1. (1998)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Kawahara, T.: "Virtual Fitting Room with Spoken Dialogue Interface" J,of IPSJ. (to appear.).

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Kawahara, T.: "Speaking-style dependent lexicalized filler model for key-phrase detection and verification" Tech.report of IPSJ. SP97-78. (1997)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Kawahara, T.: "Combining key-phrase detection and subword-based verification for flexible speech understanding" Proc.of ICASSP97. 1159-1162 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Kawahara, T.: "Phrase language models for detection and verification-based speech understanding" Proc.IEEE Workshop on Automatic Speech Recognition and Understanding. 49-56 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Araki, M.: "Automatic Evaluation Environment for Spoken Dialogue Systems" in Mayr, E.et al.ed. "Dialogue Processing in Spoken Language Systems". (1997)

    • Description
      「研究成果報告書概要(欧文)」より
  • [Publications] Doshita, S.: "Human-machine communication by speech" Ohmsha inc.(1998)

    • Description
      「研究成果報告書概要(欧文)」より

URL: 

Published: 1999-03-16  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi