Cross-modality video person Re-ID with modality-aware cosine-triplet loss
Issued Date
2026-04-01
Resource Type
ISSN
13807501
eISSN
15737721
Scopus ID
2-s2.0-105033434386
Journal Title
Multimedia Tools and Applications
Volume
85
Issue
4
Rights Holder(s)
SCOPUS
Bibliographic Citation
Multimedia Tools and Applications Vol.85 No.4 (2026)
Suggested Citation
Kasantikul R., Kusakunniran W. Cross-modality video person Re-ID with modality-aware cosine-triplet loss. Multimedia Tools and Applications Vol.85 No.4 (2026). doi:10.1007/s11042-026-21477-2 Retrieved from: https://repository.li.mahidol.ac.th/handle/123456789/115886
Title
Cross-modality video person Re-ID with modality-aware cosine-triplet loss
Author(s)
Author's Affiliation
Corresponding Author(s)
Other Contributor(s)
Abstract
Person Re-Identification (Re-ID) is one of the important applications for surveillance. The scenarios where we need to identify the subjects captured by night-vision (infrared) cameras are a significant challenge to the existing Re-ID techniques, where only color footage is available for comparison. This is due to large differences in the composition between color and infrared images, which results in appearance-based information becoming less reliable for Re-ID. For this reason, we hypothesized that motion information from sequences of inputs is vital for cross-modality (visible-to-infrared) Re-ID. From our initial findings, motion information from the sequence of frames significantly improved the cross-modality Re-ID performance. In addition, choices of distance metrics (Euclidean vs. cosine) have a significant effect on the overall performance. As a result, the experimental performance on SYSU-MM01 reached 72.70% in mAP and 73.27% in rank-1 accuracy and yielded significant performance gains of 28.32% in mAP and 29.14% in rank-1 accuracy over our baseline. The performance competes with the existing state-of-the-art techniques tested on the same dataset.
