• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to previous page

Studies on Multimodal Communication by Integrating Speech and Diagram

Research Project

Project/Area Number 08458078
Research Category

Grant-in-Aid for Scientific Research (B)

Allocation TypeSingle-year Grants
Section一般
Research Field Intelligent informatics
Research InstitutionKYOTO UNIVERSITY

Principal Investigator

DOSHITA Shuji  Kyoto University, Graduate School of Engineering, Professor, 工学研究科, 教授 (00025925)

Co-Investigator(Kenkyū-buntansha) ARAKI Masahiro  Kyoto University.Center for for Information and Multimedia Studies.Lecturer, 総合情報メディアセンター, 講師 (50252490)
KAWAHARA Tatsuya  Kyoto University.Graduate School of Engineering.Assistant Professor, 工学研究科, 助教授 (00234104)
Project Period (FY) 1996 – 1997
Project Status Completed (Fiscal Year 1997)
Budget Amount *help
¥7,300,000 (Direct Cost: ¥7,300,000)
Fiscal Year 1997: ¥1,800,000 (Direct Cost: ¥1,800,000)
Fiscal Year 1996: ¥5,500,000 (Direct Cost: ¥5,500,000)
KeywordsMultimodal Communication / Dialogue / Speech Recognition / Gesture / Drawing System / Information Integration
Research Abstract

The aim of this research is to realize multimodal communication between human and machines through implementing multimodal drawing system with speech input. The results of this research are below :
1. Integration of speech, gesture and diagram
We developed multi-mode integration method without assuming the synchronization of each input elements. The validity of this integration method was examined by evaluating the improvement of accuracy of information input in multimodal dictation system.
2. Implementation of multimodal interface using speech and pointing gesture
We developed the cooperation mechanism which uses the result of pointing gesture recognition, which is reliable channel, for the recognition of speech, which is relatively unreliable channel. By using the result of pointing gesture recognition, the accuracy of speech recognition raised about 6%. Using this integration method, we implemented a multimodal drawing system with speech input.
3. Interpretation of utterance using diagrammatic and contextual information in multimodal drawing system We realized utterance understanding mechanism which uses context management mechanism by stack in order to deal with spontaneous speech (e. g. ellipsis, reference expression, etc.). We integrated this utterance understanding mechanism to multimodal drawing system.

Report

(3 results)
  • 1997 Annual Research Report   Final Research Report Summary
  • 1996 Annual Research Report
  • Research Products

    (28 results)

All Other

All Publications (28 results)

  • [Publications] Araki, M.: "Evaluating Dialogue Strategies under Various Communication Errors" Proc.of IJCAI Workshop on Collaboration,Cooperation and Conflict in Dialogue Systems. 13-18 (1997)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] 荒木 雅弘: "マルチモーダル作図システムにおける文脈知識を利用した会話理解" 人工知能学会研究会資料SIG-SLUD-9703-1. (1998)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] 河原 達也: "音声言語を用いた仮想空間との対話による試着システム" 情報処理学会論文誌. 採録決定 (1998)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Speaking-style dependent lexicalized filler model for key-phrase detection and verification" 電子情報通信学会技術研究報告. SP97-78. (1997)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Combining key-phrase detection and subword-based verification for flexible speech understanding" Proc.of ICASSP97. 1159-1162 (1997)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Phrase language models for detection and verification-based speech understanding" Proc.IEEE Workshop on Automatic Speech Recognition and Understanding. 49-56 (1997)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Araki, M.: ""Automatic Evaluation Environment for Spoken Dialogue Systems"in Mayer,E.et al.ed."Dialogue Processing in Spoken Language Systems"" Springer-Verlag, 12 (1997)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] 堂下 修司: "「音声による人間と機械の対話」" オーム社, 383 (1998)

    • Description
      「研究成果報告書概要(和文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Araki, M.: "Evaluating Dialogue Strategies under Various Communication Errors" Proc.of IJCAI Workshop on Collaboration, Cooperation and Conflict in Dialogue Systems. 13-18 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Araki, M.: "Interpretation of Utterance on Multi-modal Drawing System based on Contextual Knowledge" Tech.report of IJSAI. SIG-SLUD-9703-1. (1998)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Virtual Fitting Room with Spoken Dialogue Interface" J,of IPSJ. (to appear.).

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Speaking-style dependent lexicalized filler model for key-phrase detection and verification" Tech.report of IPSJ. SP97-78. (1997)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Combining key-phrase detection and subword-based verification for flexible speech understanding" Proc.of ICASSP97. 1159-1162 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Kawahara, T.: "Phrase language models for detection and verification-based speech understanding" Proc.IEEE Workshop on Automatic Speech Recognition and Understanding. 49-56 (1997)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Araki, M.: "Automatic Evaluation Environment for Spoken Dialogue Systems" in Mayr, E.et al.ed. "Dialogue Processing in Spoken Language Systems". (1997)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Doshita, S.: "Human-machine communication by speech" Ohmsha inc.(1998)

    • Description
      「研究成果報告書概要(欧文)」より
    • Related Report
      1997 Final Research Report Summary
  • [Publications] Araki,M.: "Evaluating Dialogue Strategies under Various Communication Errors" Proc.of IJCAI Workshop on Collaboration,Cooperation and Conflict in Dialogue Systems. 13-18 (1997)

    • Related Report
      1997 Annual Research Report
  • [Publications] 荒木 雅弘: "マルチモーダル作図システムにおける文脈知識を利用した発話理解" 人工知能学会研究会資料SIG-SLUD-9703-1. (1998)

    • Related Report
      1997 Annual Research Report
  • [Publications] 河原達也: "音声言語を用いた仮想空間との対話による試着システム" 情報処理学会論文誌. (採録決定). (1998)

    • Related Report
      1997 Annual Research Report
  • [Publications] Kawahara,T.: "Speaking-style dependent lexicalized filler model for key-phrase detection and verification" 電子情報通信学会技術研究報告. SP97-78. (1997)

    • Related Report
      1997 Annual Research Report
  • [Publications] Kawahara,T.: "Combining key-phrase detection and subword-based verification for flexible speech understanding" Proc.of ICASSP97. 1159-1162 (1997)

    • Related Report
      1997 Annual Research Report
  • [Publications] Kawahara,T.: "Phrase language models for detection and verification-based speech understanding" Proc.IEEE Workshop on Automatic Speech Recognition and Understanding. 49-56 (1997)

    • Related Report
      1997 Annual Research Report
  • [Publications] Araki,M.: ""Automatic Evaluation Environment for Spoken Dialogue Systems" in Mayer,E.et al.ed. "Dialogue Processing in Spoken Language Systems"" Springer-Verlag, 12 (1997)

    • Related Report
      1997 Annual Research Report
  • [Publications] 堂下 修司: "音声による人間と機械の対話" オーム社, 383 (1998)

    • Related Report
      1997 Annual Research Report
  • [Publications] Doshita,S.: "UNDERSTANDING AND GENERATING DIALOGUE BY INTEGRATING PRO-CESSING OF SPEECH,LANGUAE AND CONCEPT" Proc.of International Symposium on Spoken Dialogue '96. 1-8 (1996)

    • Related Report
      1996 Annual Research Report
  • [Publications] 堂下修司: "音声・言語・概念の統合的処理による対話の理解と生成" 人工知能学会誌. Vol.12,No.1. 3-12 (1997)

    • Related Report
      1996 Annual Research Report
  • [Publications] 荒木雅弘: "音声・ジェスチャ・図像を統合したマルチモーダル情報の理解" 人工知能学会研究会資料. SLUD-9603. 13-18 (1997)

    • Related Report
      1996 Annual Research Report
  • [Publications] 荒木雅弘: "図像情報を利用した講演調音声のディクテーション" 第53回情報処理学会全国大会講演論文集. 357-358 (1996)

    • Related Report
      1996 Annual Research Report

URL: 

Published: 1996-04-01   Modified: 2016-04-21  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi