Skip to content

Code of 'Interaction-and-aggregation network for person re-identification' (CVPR2019), and 'IAUnet: Global Context-Aware Feature Learning for Person Re-Identification' (TNNLS 2021).

blue-blue272/ImgReID-IAnet

Repository files navigation

IAUnet

This repository contains the code for the paper:
IAUnet: Global Context-Aware Feature Learning for Person Re-Identification
Ruibing Hou, Bingpeng Ma, Hong Chang, Xinqian Gu, Shiguang Shan, Xilin Chen
TNNLS 2020

Interaction-and-aggregation network for person re-identification
Ruibing Hou, Bingpeng Ma, Hong Chang, Xinqian Gu, Shiguang Shan, Xilin Chen
CVPR 2019

Abstract

Person re-identification (reID) by CNNs based networks has achieved favorable performance in recent years. However, most of existing CNNs based methods do not take full advantage of spatial-temporal context modeling. In fact, the global spatial-temporal context can greatly clarify local distractions to enhance the target feature representation. To comprehensively leverage the spatial-temporal context information, in this work, we present a novel block, Interaction-AggregationUpdate (IAU), for high-performance person reID. Firstly, SpatialTemporal IAU (STIAU) module is introduced. STIAU jointly incorporates two types of contextual interactions into a CNN framework for target feature learning. Here the spatial interactions learn to compute the contextual dependencies between different body parts of a single frame. While the temporal interactions are used to capture the contextual dependencies between the same body parts across all frames. Furthermore, a Channel IAU (CIAU) module is designed to model the semantic contextual interactions between channel features to enhance the feature representation, especially for small-scale visual cues and body parts. Therefore, the IAU block enables the feature to incorporate the globally spatial, temporal, and channel context. It is lightweight, end-to-end trainable, and can be easily plugged into existing CNNs to form IAUnet. The experiments show that IAUnet performs favorably against state-of-the-art on both image and video reID tasks and achieves compelling results on a general object categorization task.

Training and test

# For Market
1. we first generate the part masks with the code https://github.com/Engineering-Course/LIP_JPPNet/.
2. python train.py
3. python train.py --resume "path to model.pth" --evaluate

Citation

If you use this code for your research, please cite our paper:

@article{IAUnet,
  title={IAUnet: Global Context-Aware Feature Learning for Person Re-Identification},
  author={Ruibing Hou and Bingpeng Ma and Hong Chang and Xinqian Gu and Shiguang Shan and Xilin Chen},
  journal={IEEE Transactions on Neural Networks and Learning Systems},
  year={2020},
  publisher={IEEE}
}
@inproceedings{IANet,
  title={Interaction-and-Aggregation Network for Person Re-identification},
  author={Ruibing Hou and Bingpeng Ma and Hong Chang and Xinqian Gu and Shiguang Shan and Xilin Chen},
  booktitle={CVPR},
  year={2019}
}

Platform

This code was developed and tested with pytorch version 1.0.1.

Acknowledgments

This code is based on the implementations of Deep person reID.

About

Code of 'Interaction-and-aggregation network for person re-identification' (CVPR2019), and 'IAUnet: Global Context-Aware Feature Learning for Person Re-Identification' (TNNLS 2021).

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published