TY - GEN
T1 - A Cost-effective Training Scheme for Person Re-Identification in Soccer Matches
AU - Kao, Yu Yung
AU - Chou, Po Yung
AU - Lin, Cheng Hung
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - In soccer matches, Person Re-identification (ReID) technology plays a crucial role in analyzing the position information and statistics of each player. Besides, multiple cameras are used to capture players during soccer matches. However, the appearance features of the same player in different cameras may exhibit significant differences, which makes person identification difficult. Therefore, a robust ReID model is required to obtain distinctive appearance features and identify each player. Although existing ReID technologies have achieved excellent performance in person identification, they typically require expensive computational resources to train a high-accuracy model for feature extraction. Compared to conventional ReID approaches, this research is built upon the Vision Transformer (ViT), focusing on batch-wise sample imbalance issues to enhance ReID performance within limited batch memory and less training data. Our model training process is performed on a single 24GB Nvidia RTX-3090 graphic card with 23% SoccerNet train data. The experimental results show our approach achieves a ReID performance of 82% in mean average precision (mAP) and 76% in rank1.
AB - In soccer matches, Person Re-identification (ReID) technology plays a crucial role in analyzing the position information and statistics of each player. Besides, multiple cameras are used to capture players during soccer matches. However, the appearance features of the same player in different cameras may exhibit significant differences, which makes person identification difficult. Therefore, a robust ReID model is required to obtain distinctive appearance features and identify each player. Although existing ReID technologies have achieved excellent performance in person identification, they typically require expensive computational resources to train a high-accuracy model for feature extraction. Compared to conventional ReID approaches, this research is built upon the Vision Transformer (ViT), focusing on batch-wise sample imbalance issues to enhance ReID performance within limited batch memory and less training data. Our model training process is performed on a single 24GB Nvidia RTX-3090 graphic card with 23% SoccerNet train data. The experimental results show our approach achieves a ReID performance of 82% in mean average precision (mAP) and 76% in rank1.
KW - cameras
KW - feature extraction
KW - person re-identification
KW - sample balancing
KW - vision transformer
UR - https://www.scopus.com/pages/publications/85205768086
UR - https://www.scopus.com/pages/publications/85205768086#tab=citedBy
U2 - 10.1109/ICCE-Taiwan62264.2024.10674193
DO - 10.1109/ICCE-Taiwan62264.2024.10674193
M3 - Conference contribution
AN - SCOPUS:85205768086
T3 - 11th IEEE International Conference on Consumer Electronics - Taiwan, ICCE-Taiwan 2024
SP - 455
EP - 456
BT - 11th IEEE International Conference on Consumer Electronics - Taiwan, ICCE-Taiwan 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 11th IEEE International Conference on Consumer Electronics - Taiwan, ICCE-Taiwan 2024
Y2 - 9 July 2024 through 11 July 2024
ER -