English | 简体中文
Multi-target multi-camera tracking, or MTMCT, matches the identity of a person in different cameras based on the single-camera tracking. MTMCT is usually applied to the security system and the smart retailing. The MTMCT module of PP-Human aims to provide a multi-target multi-camera pipleline which is simple, and efficient.
-
Download REID model and unzip it to
./output_inference
. For the MOT model, please refer to mot description. -
In the MTMCT mode, input videos are required to be put in the same directory. set the REID "enable: True" in the infer_cfg_pphuman.yml. The command line is:
python3 deploy/pipeline/pipeline.py --config deploy/pipeline/config/infer_cfg_pphuman.yml --video_dir=[your_video_file_directory] --device=gpu
- Configuration can be modified in
./deploy/pipeline/config/infer_cfg_pphuman.yml
.
python3 deploy/pipeline/pipeline.py
--config deploy/pipeline/config/infer_cfg_pphuman.yml -o REID.model_dir=reid_best/
--video_dir=[your_video_file_directory]
--device=gpu
MTMCT module consists of the multi-target multi-camera tracking pipeline and the REID model.
- Multi-Target Multi-Camera Tracking Pipeline
single-camera tracking[id+bbox]
│
capture the target in the original image according to bbox——│
│ │
REID model quality assessment (covered or not, complete or not, brightness, etc.)
│ │
[feature] [quality]
│ │
datacollector—————│
│
sort out and filter features
│
calculate the similarity of IDs in the videos
│
make the IDs cluster together and rearrange them
- The model solution is reid-strong-baseline, with ResNet50 as the backbone.
Under the above circumstances, the REID model used in MTMCT integrates open-source datasets and compresses model features to 128-dimensional features to optimize the generalization. In this way, the actual generalization result becomes much better.
- The provided REID model is obtained from open-source dataset training. It is recommended to add your own data to get a more powerful REID model, notably improving the MTMCT effect.
- The quality assessment is based on simple logic +OpenCV, whose effect is limited. If possible, it is advisable to conduct specific training on the quality assessment model.
- camera 1:
- camera 2:
@InProceedings{Luo_2019_CVPR_Workshops,
author = {Luo, Hao and Gu, Youzhi and Liao, Xingyu and Lai, Shenqi and Jiang, Wei},
title = {Bag of Tricks and a Strong Baseline for Deep Person Re-Identification},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2019}
}
@ARTICLE{Luo_2019_Strong_TMM,
author={H. {Luo} and W. {Jiang} and Y. {Gu} and F. {Liu} and X. {Liao} and S. {Lai} and J. {Gu}},
journal={IEEE Transactions on Multimedia},
title={A Strong Baseline and Batch Normalization Neck for Deep Person Re-identification},
year={2019},
pages={1-1},
doi={10.1109/TMM.2019.2958756},
ISSN={1941-0077},
}