• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to previous page

Removing the Burden of Data Labeling: Automatic Surgical Video Understanding with Unsupervised Learning

Research Project

Project/Area Number 20K23343
Research Category

Grant-in-Aid for Research Activity Start-up

Allocation TypeMulti-year Fund
Review Section 1002:Human informatics, applied informatics and related fields
Research InstitutionOsaka University

Principal Investigator

LI LIANGZHI  大阪大学, データビリティフロンティア機構, 特任助教(常勤) (10875545)

Project Period (FY) 2020-09-11 – 2022-03-31
Project Status Completed (Fiscal Year 2021)
Budget Amount *help
¥1,430,000 (Direct Cost: ¥1,100,000、Indirect Cost: ¥330,000)
Fiscal Year 2021: ¥1,170,000 (Direct Cost: ¥900,000、Indirect Cost: ¥270,000)
Fiscal Year 2020: ¥260,000 (Direct Cost: ¥200,000、Indirect Cost: ¥60,000)
KeywordsFew-shot Learning / Semantic Segmentation / Video Understanding / Medical Images / Computer Vision / Surgical Analysis / Deep Learning / Surgical Videos / Unsupervised Learning
Outline of Research at the Start

This project aims to design an unsupervised learning method to perform spatial and temporal segmentations without humans’ input. We will use the hidden information in surgical videos to better differentiate similar objects and surgical phases, both of which are very difficult for existing solutions.

Outline of Final Research Achievements

As most of the existing automatic surgical video analysis models require a large number of manually labeled data for training, this project aims to design a learning method to perform spatial and temporal segmentations with smaller requirements of humans’ input. During this project, I mainly studied the following sub-topics towards the goal.
1.Surgical images/frames semantic segmentation in a weakly-supervised way. I developed a new training strategy for video semantic segmentation models to utilized unlabeled data to improve their segmentation performance.
2.Surgical videos temporal analysis using no labels. I developed a retrieval-based method to automatically predict surgical duration.

Academic Significance and Societal Importance of the Research Achievements

The establishment of techniques capable of understanding what, where, and when is happening with no requirements on human efforts will ease the task of surgery indexing, clinical training, etc.

Report

(3 results)
  • 2021 Annual Research Report   Final Research Report ( PDF )
  • 2020 Research-status Report
  • Research Products

    (5 results)

All 2021 Other

All Journal Article (1 results) (of which Int'l Joint Research: 1 results,  Peer Reviewed: 1 results,  Open Access: 1 results) Presentation (2 results) (of which Int'l Joint Research: 2 results) Remarks (2 results)

  • [Journal Article] Noisy-LSTM: Improving Temporal Awareness for Video Semantic Segmentation2021

    • Author(s)
      Wang Bowen、Li Liangzhi、Nakashima Yuta、Kawasaki Ryo、Nagahara Hajime、Yagi Yasushi
    • Journal Title

      IEEE Access

      Volume: 9 Pages: 46810-46820

    • DOI

      10.1109/access.2021.3067928

    • Related Report
      2020 Research-status Report
    • Peer Reviewed / Open Access / Int'l Joint Research
  • [Presentation] SCOUTER: Slot Attention-based Classifier for Explainable Image Recognition2021

    • Author(s)
      Liangzhi Li, Bowen Wang, Manisha Verma, Yuta Nakashima, Ryo Kawasaki, Hajime Nagahara
    • Organizer
      IEEE/CVF International Conference on Computer Vision (ICCV)
    • Related Report
      2021 Annual Research Report
    • Int'l Joint Research
  • [Presentation] MTUNet: Few-shot Image Classification with Visual Explanations2021

    • Author(s)
      Wang Bowen、Li Liangzhi、Manisha Verma、Nakashima Yuta、Kawasaki Ryo、Nagahara Hajime
    • Organizer
      IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Responsible Computer Vision Workshop
    • Related Report
      2020 Research-status Report
    • Int'l Joint Research
  • [Remarks] Github Code for the Semantic Segmentation

    • URL

      https://github.com/wbw520/NoisyLSTM

    • Related Report
      2020 Research-status Report
  • [Remarks] Github Code for the Few-shot Learning

    • URL

      https://github.com/wbw520/MTUNet

    • Related Report
      2020 Research-status Report

URL: 

Published: 2020-09-29   Modified: 2023-01-30  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi