研究実績の概要 |
We examined the performance of four recent saliency models EML-NET, SalGAN, DeepGaze IIE, and DeepGaze on images of hand tools. These objects have distinct segments with various roles, and studies suggest that tool segments inherently attract human attention. We tested the models on a dataset containing both tool and non-tool images, then compared their predictions with human gaze data using six criteria. The results show that the models often struggle to predict saliency accurately for tool images compared to non-tool images. This suggests a need to address this limitation in saliency modeling for tool-specific contexts.
|
今後の研究の推進方策 |
We focus on the state-of-the-art visual saliency prediction model of DeepGaze IIE and make an effort to refine it to account for this bias. Since the integration of transfer learning into saliency prediction over the last decade has notably enhanced prediction performance, we will initially curate a custom image data set featuring tools, non-tools and ambiguous images and record empirical gaze data from human participants to be used in fine-tuning. In this way, we will improve the model’s performance for this specific stimulus category and evaluate it by IG and NSS metrics.
|