Deep metric learning for video-based person re-identification

Naoki Kato, Kohei Hakozaki, Masamoto Tanabiki, Junko Furuyama, Yuji Sato, Yoshimitsu Aoki

研究成果: Article査読

抄録

This paper proposes a novel approach for video-based person re-identification that exploits deep convolutional neural networks to learn the similarity of persons observed from video camera. By Convolutional Neural Networks (CNN), each video sequence of a person is mapped to a Euclidean space where distances between feature embeddings directly correspond to measures of person similarity. By improved parameter learning method called Entire Triplet Loss, all possible triplets in the mini-batch are taken into account to update network parameters at once. This simple change of parameter updating method significantly improves network training, enabling the embeddings to be further discriminative. Experimental results show that proposed model achieves new state of the art identification rate on iLEDS-VID dataset and PRID-2011 dataset with 78.3%, 83.9% at rank 1, respectively.

本文言語English
ページ(範囲)1117-1124
ページ数8
ジャーナルSeimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering
83
12
DOI
出版ステータスPublished - 2017

ASJC Scopus subject areas

  • 機械工学

フィンガープリント

「Deep metric learning for video-based person re-identification」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル