Research on robust spoken language interfaces for diverse voice variability and expressivity
Project/Area Number |
21300063
|
Research Category |
Grant-in-Aid for Scientific Research (B)
|
Allocation Type | Single-year Grants |
Section | 一般 |
Research Field |
Perception information processing/Intelligent robotics
|
Research Institution | Tokyo Institute of Technology |
Principal Investigator |
KOBAYASHI Takao 東京工業大学, 大学院・総合理工学研究科, 教授 (70153616)
|
Co-Investigator(Renkei-kenkyūsha) |
NAGAHASHI Hiroshi 東京工業大学, 像情報工学研究所, 教授 (20143084)
|
Research Collaborator |
NOSE Takashi 東京工業大学, 大学院・総合理工学研究科, 助教 (90550591)
|
Project Period (FY) |
2009 – 2011
|
Project Status |
Completed (Fiscal Year 2011)
|
Budget Amount *help |
¥9,750,000 (Direct Cost: ¥7,500,000、Indirect Cost: ¥2,250,000)
Fiscal Year 2011: ¥2,470,000 (Direct Cost: ¥1,900,000、Indirect Cost: ¥570,000)
Fiscal Year 2010: ¥3,510,000 (Direct Cost: ¥2,700,000、Indirect Cost: ¥810,000)
Fiscal Year 2009: ¥3,770,000 (Direct Cost: ¥2,900,000、Indirect Cost: ¥870,000)
|
Keywords | 音声情報処理 / HMM音声合成 / テキスト音声合成 / 基本周波数量子化コンテキスト / 自然発話音声 / 重回帰HMM / 音声スタイル制御 / 音声スタイル推定 / モデル適応 / 対話音声合成 / スタイル推定 / 声質変換 / FO量子化 / 発話様式 |
Research Abstract |
The purpose of the research is to develop techniques that make the human-computer interaction using speech input/output more robust for variations of users' emotional states, speaking styles, preferences, and expressivity. We have proposed techniques using a quantized fundamental frequency prosodic context for robust speech synthesis and an extended context set for spontaneous conversational speech synthesis. We have also proposed techniques for robust speech recognition including extraction of paralinguistic information and rapid model adaptation.
|
Report
(4 results)
Research Products
(109 results)