2022 Fiscal Year Research-status Report
Development of learning subspace-based methods for pattern recognition
Project/Area Number |
22K17960
|
Research Institution | National Institute of Advanced Industrial Science and Technology |
Principal Investigator |
SALESDESOUZA LINCON 国立研究開発法人産業技術総合研究所, 情報・人間工学領域, 研究員 (40912481)
|
Project Period (FY) |
2022-04-01 – 2026-03-31
|
Keywords | Subspace learning / Deep neural networks / Manifold optimization |
Outline of Annual Research Achievements |
In fiscal year 2022, we worked to address the problem that traditional deep neural network frameworks process image sets independently, without considering the underlying feature distribution and the variance of the images in the set. To overcome this limitation, we devised a new subspace learning method called Grassmannian learning mutual subspace method (G-LMSM), which is an NN layer that can be integrated into deep neural networks. G-LMSM maps the image set into a low-dimensional input subspace representation, which is then matched with dictionary subspaces using a similarity metric of their canonical angles, an interpretable and computationally efficient metric. The key idea of G-LMSM is to learn dictionary subspaces as points on the Grassmann manifold, which is a smooth, non-linear manifold that captures the geometric structure of subspaces. This learning is optimized with Riemannian stochastic gradient descent, which is stable, efficient, and theoretically well-grounded. The proposed method was evaluated on three different tasks: hand shape recognition, face identification, and facial emotion recognition. Our experimental results showed that G-LMSM outperformed state-of-the-art methods on all three tasks, demonstrating its potential to improve the performance of deep frameworks for object recognition from image sets.
|
Current Status of Research Progress |
Current Status of Research Progress
2: Research has progressed on the whole more than it was originally planned.
Reason
Reason: We were able to combine subspace learning and deep neural networks to improve the performance in tasks of image set recognition.
|
Strategy for Future Research Activity |
We will work on new ways to combine subspace learning and deep neural network that can address their problems and improve performance.
|
Causes of Carryover |
Reason: To cover cloud computing costs to use in the next fiscal year and to attend conferences. Plan: We plan to purchase cloud computing that can process large-scale data at high speed and attend conferences to gather necessary information on the latest technologies and/or present our research.
|