TY - GEN
T1 - Detecting anomalous regions from an image based on deep captioning
AU - Hatae, Yusuke
AU - Yang, Qingpu
AU - Fadjrimiratno, Muhammad Fikko
AU - Li, Yuanyuan
AU - Matsukawa, Tetsu
AU - Suzuki, Einoshin
N1 - Funding Information:
A part of this work was supported by JSPS KAK-ENHI Grant Number JP18H03290.
Publisher Copyright:
Copyright © 2020 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved
PY - 2020
Y1 - 2020
N2 - In this paper we propose a one-class anomalous region detection method from an image based on deep captioning. Such a method can be installed on an autonomous mobile robot, which reports anomalies from observation without any human supervision and would interest a wide range of researchers, practitioners, and users. In addition to image features, which were used by conventional methods, our method exploits recent advances in deep captioning, which is based on deep neural networks trained on a large-scale data on image - caption pairs, enabling anomaly detection in the semantic level. Incremental clustering is adopted so that the robot is able to model its observation into a set of clusters and report substantially new observations as anomalies. Extensive experiments using two real-world data demonstrate the superiority of our method in terms of recall, precision, F measure, and AUC over the traditional approach. The experiments also show that our method exhibits excellent learning curve and low threshold dependency.
AB - In this paper we propose a one-class anomalous region detection method from an image based on deep captioning. Such a method can be installed on an autonomous mobile robot, which reports anomalies from observation without any human supervision and would interest a wide range of researchers, practitioners, and users. In addition to image features, which were used by conventional methods, our method exploits recent advances in deep captioning, which is based on deep neural networks trained on a large-scale data on image - caption pairs, enabling anomaly detection in the semantic level. Incremental clustering is adopted so that the robot is able to model its observation into a set of clusters and report substantially new observations as anomalies. Extensive experiments using two real-world data demonstrate the superiority of our method in terms of recall, precision, F measure, and AUC over the traditional approach. The experiments also show that our method exhibits excellent learning curve and low threshold dependency.
UR - http://www.scopus.com/inward/record.url?scp=85083497425&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85083497425&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85083497425
T3 - VISIGRAPP 2020 - Proceedings of the 15th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications
SP - 326
EP - 335
BT - VISAPP
A2 - Farinella, Giovanni Maria
A2 - Radeva, Petia
A2 - Braz, Jose
PB - SciTePress
T2 - 15th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, VISIGRAPP 2020
Y2 - 27 February 2020 through 29 February 2020
ER -