TY - GEN
T1 - Video-Based Person Re-identification by 3D Convolutional Neural Networks and Improved Parameter Learning
AU - Kato, Naoki
AU - Hakozaki, Kohei
AU - Tanabiki, Masamoto
AU - Furuyama, Junko
AU - Sato, Yuji
AU - Aoki, Yoshimitsu
N1 - Publisher Copyright:
© 2018, Springer International Publishing AG, part of Springer Nature.
PY - 2018
Y1 - 2018
N2 - In this paper we propose a novel approach for video-based person re-identification that exploits convolutional neural networks to learn the similarity of persons observed from video camera. We take 3-dimensional convolutional neural networks (3D CNN) to extract fine-grained spatiotemporal features from the video sequence of a person. Unlike recurrent neural networks, 3D CNN preserves the spatial patterns of the input, which works well on re-identification problem. The network maps each video sequence of a person to a Euclidean space where distances between feature embeddings directly correspond to measures of person similarity. By our improved parameter learning method called entire triplet loss, all possible triplets in the mini-batch are taken into account to update network parameters. This parameter updating method significantly improves training, enabling the embeddings to be more discriminative. Experimental results show that our model achieves new state of the art identification rate on iLIDS-VID dataset and PRID-2011 dataset with 82.0%, 83.3% at rank 1, respectively.
AB - In this paper we propose a novel approach for video-based person re-identification that exploits convolutional neural networks to learn the similarity of persons observed from video camera. We take 3-dimensional convolutional neural networks (3D CNN) to extract fine-grained spatiotemporal features from the video sequence of a person. Unlike recurrent neural networks, 3D CNN preserves the spatial patterns of the input, which works well on re-identification problem. The network maps each video sequence of a person to a Euclidean space where distances between feature embeddings directly correspond to measures of person similarity. By our improved parameter learning method called entire triplet loss, all possible triplets in the mini-batch are taken into account to update network parameters. This parameter updating method significantly improves training, enabling the embeddings to be more discriminative. Experimental results show that our model achieves new state of the art identification rate on iLIDS-VID dataset and PRID-2011 dataset with 82.0%, 83.3% at rank 1, respectively.
UR - http://www.scopus.com/inward/record.url?scp=85049462398&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85049462398&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-93000-8_18
DO - 10.1007/978-3-319-93000-8_18
M3 - Conference contribution
AN - SCOPUS:85049462398
SN - 9783319929996
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 153
EP - 163
BT - Image Analysis and Recognition - 15th International Conference, ICIAR 2018, Proceedings
A2 - ter Haar Romeny, Bart
A2 - Karray, Fakhri
A2 - Campilho, Aurelio
PB - Springer Verlag
T2 - 15th International Conference on Image Analysis and Recognition, ICIAR 2018
Y2 - 27 June 2018 through 29 June 2018
ER -