• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to previous page

A new data-driven approach to bring humanity into virtual worlds with computer vision

Research Project

Project/Area Number 23K28129
Project/Area Number (Other) 23H03439 (2023)
Research Category

Grant-in-Aid for Scientific Research (B)

Allocation TypeMulti-year Fund (2024)
Single-year Grants (2023)
Section一般
Review Section Basic Section 61020:Human interface and interaction-related
Basic Section 62040:Entertainment and game informatics-related
Sections That Are Subject to Joint Review: Basic Section61020:Human interface and interaction-related , Basic Section62040:Entertainment and game informatics-related
Research InstitutionKyushu University

Principal Investigator

THOMAS DIEGO  九州大学, システム情報科学研究院, 准教授 (10804651)

Co-Investigator(Kenkyū-buntansha) 鍛冶 静雄  九州大学, マス・フォア・インダストリ研究所, 教授 (00509656)
古賀 靖子  九州大学, 人間環境学研究院, 准教授 (60225399)
川崎 洋  九州大学, システム情報科学研究院, 教授 (80361393)
落合 啓之  九州大学, マス・フォア・インダストリ研究所, 教授 (90214163)
Project Period (FY) 2023-04-01 – 2026-03-31
Project Status Granted (Fiscal Year 2024)
Budget Amount *help
¥18,590,000 (Direct Cost: ¥14,300,000、Indirect Cost: ¥4,290,000)
Fiscal Year 2025: ¥5,200,000 (Direct Cost: ¥4,000,000、Indirect Cost: ¥1,200,000)
Fiscal Year 2024: ¥7,800,000 (Direct Cost: ¥6,000,000、Indirect Cost: ¥1,800,000)
Fiscal Year 2023: ¥5,590,000 (Direct Cost: ¥4,300,000、Indirect Cost: ¥1,290,000)
Keywordsデジタルヒューマン / 4D capture / Immersive VR / Differentiable rendering / AI-based 3D modeling / Animatable Avatar / 3D reconstruction / Deep learning / Weak supervision / Digital humans
Outline of Research at the Start

The core idea is to leverage RGB-D data to achieve realism of digital humans. We propose:
(1) Weakly supervised method to learn 3D shape from in-the-wild RGB-D data.
(2) Real-time animation model to learn motion from RGB-D videos.
(3) Data-driven method to learn semantically appropriate body movements.

Outline of Annual Research Achievements

We proposed a new method for human body animation that generates pose-dependent detailed deformations in real-time on standard animation pipeline. Our proposed method can animate an avatar up to 30 times faster than baselines with better level of details. The results of this research was published in the proceedings of the international conference Computer Graphics International (CGI) 2023.
We proposed a novel AI-based approach to the 3D reconstruction of clothed humans using weak supervision via 2D normal maps. Our results reinforce the notion that less training data is required to train networks that infer normal maps than to train networks that infer 3D geometry. The results were published as on arXiv and submitted the European Conference on Computer Vision (ECCV) 2024.

Current Status of Research Progress
Current Status of Research Progress

2: Research has progressed on the whole more than it was originally planned.

Reason

In FY 2023 we had three main objectives:1) Design an efficient differentiable 3D render from implicit 3D surface representations for 3D reconstruction of clothed humans; 2) Propose a new method to create real-time animatable avatars from RGB-D data; 3) Capture multi-view RGB-D human data at the university.

We achieved our objectives as planned: 1) In order to address the objective of learning detailed 3D clothed human shapes from 2.5D
datasets, we proposed a novel AI-based approach using weak supervision via 2D normal maps; 2) We proposed a new method for animatable avatars that allows to control deformation of body and clothes of the avatar such as wrinkles in real-time; 3) We prepared a 3D capture system in the lab with calibrated RGB-D cameras. We captured some real data using our system.

Strategy for Future Research Activity

Our goal is to propose new methods for creating digital human twins supported by generative AI.

Our future research plan is:
1. Weakly supervised 3D reconstruction.[FY 2024] Employ adversarial learning to learn from both RGB-D and large-scale RGB datasets. [FY 2025] Propose adaptive tessellation of the 3D space to reduce computational cost while maintaining level of details.
2. Real-time photorealistic animatable avatars. [FY 2024] Add detailed animation of hands and face to the animatable avatar.[FY 2025] Capture texture and material properties of skin and clothes .
3. Semantic dynamic bodies.[FY 2024] Design action dependent animated 3D human-scene. [FY 2025] Populate 3D scenes with animated 3D human bodies that interact with the scene in a semantically correct manner.

Report

(1 results)
  • 2023 Annual Research Report
  • Research Products

    (13 results)

All 2024 2023 Other

All Int'l Joint Research (3 results) Journal Article (5 results) (of which Int'l Joint Research: 5 results,  Peer Reviewed: 4 results) Presentation (3 results) (of which Int'l Joint Research: 3 results) Remarks (2 results)

  • [Int'l Joint Research] Universidade Federal Rural de Pernambuco(ブラジル)

    • Related Report
      2023 Annual Research Report
  • [Int'l Joint Research] INRIA/ESIEE, University Gustave Eiffel/University Savoie Mont Blanc(フランス)

    • Related Report
      2023 Annual Research Report
  • [Int'l Joint Research] Stanford University/University of California, Berkeley(米国)

    • Related Report
      2023 Annual Research Report
  • [Journal Article] ActiveNeuS: Neural Signed Distance Fields for Active Stereo2024

    • Author(s)
      Kazuto Ichimaru Takaki Ikeda Diego Thomas Takafumi Iwaguchi Hiroshi Kawasaki
    • Journal Title

      International Conference on 3D Vision

      Volume: 1 Pages: 1-9

    • Related Report
      2023 Annual Research Report
    • Peer Reviewed / Int'l Joint Research
  • [Journal Article] ACT2G: Attention-based Contrastive Learning for Text-to-Gesture Generation2023

    • Author(s)
      Teshima Hitoshi、Wake Naoki、Thomas Diego、Nakashima Yuta、Kawasaki Hiroshi、Ikeuchi Katsushi
    • Journal Title

      Proceedings of the ACM on Computer Graphics and Interactive Techniques

      Volume: 6 Issue: 3 Pages: 1-17

    • DOI

      10.1145/3606940

    • Related Report
      2023 Annual Research Report
    • Peer Reviewed / Int'l Joint Research
  • [Journal Article] A Two-Step Approach for Interactive Animatable Avatars2023

    • Author(s)
      Kitamura Takumi、Iwamoto Naoya、Kawasaki Hiroshi、Thomas Diego
    • Journal Title

      Computer Graphics International Conference

      Volume: 1 Pages: 491-509

    • DOI

      10.1007/978-3-031-50072-5_39

    • ISBN
      9783031500718, 9783031500725
    • Related Report
      2023 Annual Research Report
    • Peer Reviewed / Int'l Joint Research
  • [Journal Article] Toward Unlabeled Multi-View 3D Pedestrian Detection by Generalizable AI: Techniques and Performance Analysis2023

    • Author(s)
      Lima Joao Paulo、Thomas Diego、Uchiyama Hideaki、Teichrieb Veronica
    • Journal Title

      2023 36th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)

      Volume: 1 Pages: 1-6

    • DOI

      10.1109/sibgrapi59091.2023.10347151

    • Related Report
      2023 Annual Research Report
    • Peer Reviewed / Int'l Joint Research
  • [Journal Article] Weakly-Supervised 3D Reconstruction of Clothed Humans via Normal Maps2023

    • Author(s)
      Jane Wu, Diego Thomas, Ronald Fedkiw
    • Journal Title

      arXiv preprint arXiv:2311.16042

      Volume: 1 Pages: 1-15

    • Related Report
      2023 Annual Research Report
    • Int'l Joint Research
  • [Presentation] ActiveNeuS: Neural Signed Distance Fields for Active Stereo2024

    • Author(s)
      Kazuto Ichimaru Takaki Ikeda Diego Thomas Takafumi Iwaguchi Hiroshi Kawasaki
    • Organizer
      International Conference on 3D Vision
    • Related Report
      2023 Annual Research Report
    • Int'l Joint Research
  • [Presentation] A Two-Step Approach for Interactive Animatable Avatars2023

    • Author(s)
      Kitamura Takumi、Iwamoto Naoya、Kawasaki Hiroshi、Thomas Diego
    • Organizer
      Computer Graphics International Conference
    • Related Report
      2023 Annual Research Report
    • Int'l Joint Research
  • [Presentation] ACT2G: Attention-based Contrastive Learning for Text-to-Gesture Generation2023

    • Author(s)
      Teshima Hitoshi、Wake Naoki、Thomas Diego、Nakashima Yuta、Kawasaki Hiroshi、Ikeuchi Katsushi
    • Organizer
      Proceedings of the ACM on Computer Graphics and Interactive Techniques
    • Related Report
      2023 Annual Research Report
    • Int'l Joint Research
  • [Remarks] Interactive Animatable Avatar

    • URL

      https://github.com/diegothomas/Interactive-Animatable-Avatar

    • Related Report
      2023 Annual Research Report
  • [Remarks] Digital Humans Lab

    • URL

      https://diegothomas.github.io/DigitalHumans-lab/

    • Related Report
      2023 Annual Research Report

URL: 

Published: 2023-04-18   Modified: 2024-12-25  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi