TY - GEN
T1 - First-person activity recognition with C3D features from optical flow images
AU - Takamine, Asamichi
AU - Iwashita, Yumi
AU - Kurazume, Ryo
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2016/2/10
Y1 - 2016/2/10
N2 - This paper proposes new features extracted from images derived from optical flow, for first-person activity recognition. Features from convolutional neural network (CNN), which is designed for 2D images, attract attention from computer vision researchers due to its powerful discrimination capability, and recently a convolutional neural network for videos, called C3D (Convolutional 3D), was proposed. Generally CNN / C3D features are extracted directly from original images / videos with pre-trained convolutional neural network, since the network was trained with images / videos. In this paper, on the other hand, we propose the use of images derived from optical flow (we call this image as "optical flow image") as input images into the pre-trained neural network, based on the following reasons; (i) optical flow images give dynamic information which is useful for activity recognition, compared with original images, which give only static information, and (ii) the pre-trained network has chance to extract features with reasonable discrimination capability, since the network was trained with huge amount of images from big categories. We carry out experiments with a dataset named "DogCentric Activity Dataset", and we show the effectiveness of the extracted features.
AB - This paper proposes new features extracted from images derived from optical flow, for first-person activity recognition. Features from convolutional neural network (CNN), which is designed for 2D images, attract attention from computer vision researchers due to its powerful discrimination capability, and recently a convolutional neural network for videos, called C3D (Convolutional 3D), was proposed. Generally CNN / C3D features are extracted directly from original images / videos with pre-trained convolutional neural network, since the network was trained with images / videos. In this paper, on the other hand, we propose the use of images derived from optical flow (we call this image as "optical flow image") as input images into the pre-trained neural network, based on the following reasons; (i) optical flow images give dynamic information which is useful for activity recognition, compared with original images, which give only static information, and (ii) the pre-trained network has chance to extract features with reasonable discrimination capability, since the network was trained with huge amount of images from big categories. We carry out experiments with a dataset named "DogCentric Activity Dataset", and we show the effectiveness of the extracted features.
UR - http://www.scopus.com/inward/record.url?scp=84963736141&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84963736141&partnerID=8YFLogxK
U2 - 10.1109/SII.2015.7405050
DO - 10.1109/SII.2015.7405050
M3 - Conference contribution
AN - SCOPUS:84963736141
T3 - 2015 IEEE/SICE International Symposium on System Integration, SII 2015
SP - 619
EP - 622
BT - 2015 IEEE/SICE International Symposium on System Integration, SII 2015
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 8th Annual IEEE/SICE International Symposium on System Integration, SII 2015
Y2 - 11 December 2015 through 13 December 2015
ER -