• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to project page

2019 Fiscal Year Final Research Report

High-Order Deep Learning Models: Theoretical Study and Applications

Research Project

  • PDF
Project/Area Number 17K00326
Research Category

Grant-in-Aid for Scientific Research (C)

Allocation TypeMulti-year Fund
Section一般
Research Field Intelligent informatics
Research InstitutionInstitute of Physical and Chemical Research

Principal Investigator

Zhao Qibin  国立研究開発法人理化学研究所, 革新知能統合研究センター, ユニットリーダー (30599618)

Co-Investigator(Kenkyū-buntansha) 曹 建庭  埼玉工業大学, 工学部, 教授 (20306989)
Project Period (FY) 2017-04-01 – 2020-03-31
KeywordsTensor decomposition / Tensor networks / Multi-modal learning / Tensor completion
Outline of Final Research Achievements

Tensor decomposition and tensor networks (TNs) have recently gained increasing attentions in machine learning, data mining and computer vision fields due to its effectiveness in efficient computation and model compression in deep learning. However, there are many open problems that are still unexplored, which limits its impact in machine learning. In this project, we studied the fundamental model and theory of tensor networks and applied it for data representation and model representation. We have introduced a novel tensor decomposition model together with fast and scalable algorithms which can be applied to large-scale data completion and image denoising. In addition, we also developed deep multi-task, multi-model learning and multiple GANs methods based on tensor network representations, which shows powerful expressive ability and economic model complexity.

Free Research Field

Information Science

Academic Significance and Societal Importance of the Research Achievements

Tensor representation and tensor networks have shown to be useful in deep learning models. This project has further promoted to solve the challenging problems in deep learning methods by using TN technology.

URL: 

Published: 2021-02-19  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi