Project/Area Number |
19K20335
|
Research Category |
Grant-in-Aid for Early-Career Scientists
|
Allocation Type | Multi-year Fund |
Review Section |
Basic Section 61030:Intelligent informatics-related
|
Research Institution | National Institute of Advanced Industrial Science and Technology (2021-2022) University of Tsukuba (2019-2020) |
Principal Investigator |
Gatto Bernardo 国立研究開発法人産業技術総合研究所, 情報・人間工学領域, 産総研特別研究員 (10826267)
|
Project Period (FY) |
2019-04-01 – 2023-03-31
|
Project Status |
Discontinued (Fiscal Year 2022)
|
Budget Amount *help |
¥4,160,000 (Direct Cost: ¥3,200,000、Indirect Cost: ¥960,000)
Fiscal Year 2022: ¥1,300,000 (Direct Cost: ¥1,000,000、Indirect Cost: ¥300,000)
Fiscal Year 2021: ¥1,170,000 (Direct Cost: ¥900,000、Indirect Cost: ¥270,000)
Fiscal Year 2020: ¥650,000 (Direct Cost: ¥500,000、Indirect Cost: ¥150,000)
Fiscal Year 2019: ¥1,040,000 (Direct Cost: ¥800,000、Indirect Cost: ¥240,000)
|
Keywords | elderly surveillance / subspace representation / image recognition / deep learning |
Outline of Research at the Start |
We will investigate the fusion of visual and acoustic data to support the safe life of the elderly living alone. We will employ information from visual and acoustic sensors (e.g., cameras and microphones) and recognize events, such as domestic activities or abnormal events. In this research, we develop the following technologies: (1) fast neural networks for extraction of visual characteristics from videos, (2) neural networks for acoustic data analysis and, (3) data fusion for event recognition.
|
Outline of Annual Research Achievements |
Motivated by applications of subspace analysis, two new groups of methods were presented in this project: (1) Shallow networks for image classification; and (2) Subspaces for tensor representation and classification. New representations are proposed to preserve the spatial structure and maintain a fast processing time. A new method to keep the temporal structure was also given.
These solutions were evaluated over problems involving person detection, action, and gesture representation. We focused on the fusion of visual and acoustic data to support the safe life of the elderly living alone.
|