Visual Concept Modeling of Verbs Based on a Large Scale Set of Tagged Videos Provided by Folksonomy
Project/Area Number |
26730090
|
Research Category |
Grant-in-Aid for Young Scientists (B)
|
Allocation Type | Multi-year Fund |
Research Field |
Perceptual information processing
|
Research Institution | Osaka University |
Principal Investigator |
|
Project Period (FY) |
2014-04-01 – 2017-03-31
|
Project Status |
Completed (Fiscal Year 2016)
|
Budget Amount *help |
¥3,770,000 (Direct Cost: ¥2,900,000、Indirect Cost: ¥870,000)
Fiscal Year 2016: ¥910,000 (Direct Cost: ¥700,000、Indirect Cost: ¥210,000)
Fiscal Year 2015: ¥1,820,000 (Direct Cost: ¥1,400,000、Indirect Cost: ¥420,000)
Fiscal Year 2014: ¥1,040,000 (Direct Cost: ¥800,000、Indirect Cost: ¥240,000)
|
Keywords | 視覚メディア処理 / 視覚概念学習 / 動作認識 / 映像処理 / 画像処理 |
Outline of Final Research Achievements |
This research project investigates a technology for constructing visual models of concepts represented by verbs using tagged videos which are stored in web-based video sharing services such as YouTube. In general, a video consists of several segments, and each of them has a different semantic content. Therefore, we cannot obtain the complete correspondence between tags and segments. To cope with this problem, this project proposes a method to extract a set of segments whose semantic content commonly appears in videos that have the same tag (called “common segments” in this report), and construct a visual model of the tag using the extracted segments. In the process of common segment extraction, it is quite important to measure the similarity between two segments. This project also proposes a method for calculating the similarity based on a large scale set of tagged images.
|
Report
(4 results)
Research Products
(10 results)