Abstract
Vehicle re-identification (reID) aims to identify vehicles across different cameras that have non-overlapping views. Most existing vehicle reID approaches train the reID model with well-labeled datasets via a supervised manner, which inevitably causes a severe drop in performance when tested in an unknown domain. Moreover, these supervised approaches require full annotations, which is limiting owing to the amount of unlabeled data. Therefore, with the aim of addressing the aforementioned problems, unsupervised vehicle reID models have attracted considerable attention. It always adopts domain adaptation to transfer discriminative information from supervised domains to unsupervised ones. In this paper, a novel progressive learning method with a multi-scale fusion network is proposed, named PLM, for vehicle reID in the unknown domain, which directly exploits inference from the available abundant data without any annotations. For PLM, a domain adaptation module is employed to smooth the domain bias, which generates images with similar data distribution to unlabeled target domain as “pseudo target samples”. Furthermore, to better exploit the distinct features of vehicle images in the unknown domain, a multi-scale attention network is proposed to train the reID model with the “pseudo target samples” and unlabeled samples; this network embeds low-layer texture features with high-level semantic features to train the reID model. Moreover, a weighted label smoothing (WLS) loss is proposed, which considers the distance between samples and different clusters to balance the confidence of pseudo labels in the feature learning module. Extensive experiments are carried out to verify that our proposed PLM achieves excellent performance on several benchmark datasets.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Wang Y. Survey on deep multi-modal data analytics: collaboration, rivalry, and fusion. ACM Trans Multimedia Comput Commun Appl, 2021, 17: 1–25
Lou Y, Bai Y, Liu J, et al. Embedding adversarial learning for vehicle re-identification. IEEE Trans Image Process, 2019, 28: 3794–3807
Wu L, Wang Y, Gao J, et al. Deep coattention-based comparator for relative representation learning in person re-identification. IEEE Trans Neural Netw Learn Syst, 2021, 32: 722–735
Wang H, Peng J, Chen D, et al. Attribute-guided feature learning network for vehicle reidentification. IEEE Multimedia, 2020, 27: 112–121
Wu Y, Lin Y, Dong X, et al. Exploit the unknown gradually: one-shot video-based person re-identification by stepwise learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018. 5177–5186
Wu Y, Lin Y, Dong X, et al. Progressive learning for person re-identification with one example. IEEE Trans Image Process, 2019, 28: 2872–2881
Deng W, Zheng L, Ye Q, et al. Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018. 994–1003
Wang J, Zhu X, Gong S, et al. Transferable joint attribute-identity deep learning for unsupervised person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018. 2275–2284
Wu L, Wang Y, Yin H, et al. Few-shot deep adversarial learning for video-based person re-identification. IEEE Trans Image Process, 2020, 29: 1233–1245
Song L, Wang C, Zhang L, et al. Unsupervised domain adaptive re-identification: theory and practice. Pattern Recogn, 2020, 102: 107173
Fan H, Zheng L, Yan C, et al. Unsupervised person re-identification. ACM Trans Multimedia Comput Commun Appl, 2018, 14: 1–18
Bashir R M S, Shahzad M, Fraz M M. VR-PROUD: vehicle re-identification using progressive unsupervised deep architecture. Pattern Recogn, 2019, 90: 52–65
Peng J, Wang Y, Wang H, et al. Unsupervised vehicle re-identification with progressive adaptation. In: Proceedings of International Joint Conference on Artificial Intelligence (IJCAI), 2020
Zhou Y, Shao L. Aware attentive multi-view inference for vehicle re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018. 6489–6498
Zhou Y, Liu L, Shao L. Vehicle re-identification by deep hidden multi-view inference. IEEE Trans Image Process, 2018, 27: 3275–3287
Guo H, Zhu K, Tang M, et al. Two-level attention network with multi-grain ranking loss for vehicle re-identification. IEEE Trans Image Process, 2019, 28: 4328–4338
Zhang X, Zhang R, Cao J, et al. Part-guided attention learning for vehicle re-identification. 2019. ArXiv:1909.06023
Liu H, Tian Y, Yang Y, et al. Deep relative distance learning: tell the difference between similar vehicles. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016. 2167–2175
Bai Y, Lou Y, Gao F, et al. Group-sensitive triplet embedding for vehicle reidentification. IEEE Trans Multimedia, 2018, 20: 2385–2399
Liu X, Liu W, Mei T, et al. PROVID: progressive and multimodal vehicle reidentification for large-scale urban surveillance. IEEE Trans Multimedia, 2018, 20: 645–658
Wang Z, Tang L, Liu X, et al. Orientation invariant feature embedding and spatial temporal regularization for vehicle re-identification. In: Proceedings of the IEEE International Conference on Computer Vision, 2017. 379–387
Shen Y, Xiao T, Li H, et al. Learning deep neural networks for vehicle Re-ID with visual-spatio-temporal path proposals. In: Proceedings of the IEEE International Conference on Computer Vision, 2017. 1900–1909
Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets. In: Proceedings of Advances in Neural Information Processing Systems, 2014. 2672–2680
Wu L, Hong R, Wang Y, et al. Cross-entropy adversarial view adaptation for person re-identification. IEEE Trans Circ Syst Video Technol, 2020, 30: 2081–2092
Zhou Y, Shao L. Cross-view GAN based vehicle generation for re-identification. In: Proceedings of the British Machine Vision Conference (BMVC), 2017. 1–12
Zhu J Y, Park T, Isola P, et al. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, 2017. 2223–2232
Wu L, Wang Y, Shao L. Cycle-consistent deep generative hashing for cross-modal retrieval. IEEE Trans Image Process, 2019, 28: 1602–1612
Almahairi A, Rajeswar S, Sordoni A, et al. Augmented CycleGAN: learning many-to-many mappings from unpaired data. 2018. ArXiv:1802.10151
Taigman Y, Polyak A, Wolf L. Unsupervised cross-domain image generation. 2016. ArXiv:1611.02200
He K, Zhang X, Ren S, et al. Deep residual learning for image recognition. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. 770–778
Huang Y, Xu J, Wu Q, et al. Multi-pseudo regularized label for generated data in person re-identification. IEEE Trans Image Process, 2019, 28: 1391–1403
Abadi M, Barham P, Chen J, et al. TensorFlow: a system for large-scale machine learning. In: Proceedings of 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), 2016. 265–283
Vedaldi A, Lenc K. MatConvNet: convolutional neural networks for MATLAB. In: Proceedings of the 23rd ACM International Conference on Multimedia, 2015. 689–692
Liu X, Liu W, Mei T, et al. A deep learning-based approach to progressive vehicle re-identification for urban surveillance. In: Proceedings of European Conference on Computer Vision. Berlin: Springer, 2016. 869–884
Zhong Z, Zheng L, Luo Z, et al. Invariance matters: exemplar memory for domain adaptive person re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019. 598–607
Zheng Z, Zheng L, Yang Y. A discriminatively learned CNN embedding for person reidentification. ACM Trans Multimedia Comput Commun Appl, 2018, 14: 1–20
van Der Maaten L. Accelerating t-SNE using tree-based algorithms. J Machine Learn Res, 2014, 15: 3221–3245
Acknowledgements
This work was partially supported by National Natural Science Foundation of China (Grant Nos. 62172136, U21A20470, U1936217, 61725203, 62002041), Key Research and Technology Development Projects of Anhui Province (Grant No. 202004a5020043), Liaoning Doctoral Research Start-up Fund Project (Grant No. 2021-BS-075), and Dalian Science and Technology Innovation Fund (Grant No. 2021JJ12GX028).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Wang, Y., Peng, J., Wang, H. et al. Progressive learning with multi-scale attention network for cross-domain vehicle re-identification. Sci. China Inf. Sci. 65, 160103 (2022). https://doi.org/10.1007/s11432-021-3383-y
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11432-021-3383-y