Нет описания

Yijun Fu 2e6f0ccec1 docs(README): 更新项目依赖并调整训练参数 1 месяц назад
data cba558448d Add files via upload 1 год назад
datasets 3ceab2e99d fix num_classes referenced before assignment 1 год назад
images a1d0622c6c add paperswithcode badge 1 год назад
model ce5e5c31d6 fix small bugs 1 год назад
processor 1d10157952 Remove repeated loss calculations. 1 год назад
solver cba558448d Add files via upload 1 год назад
utils 9fff385561 Update metrics.py 1 год назад
.gitignore ecde64d79e chore: 添加 .gitignore 并配置运行脚本 1 месяц назад
LICENSE 1a0d7972e5 Initial commit 1 год назад
README.md 2e6f0ccec1 docs(README): 更新项目依赖并调整训练参数 1 месяц назад
run_irra.sh 2e6f0ccec1 docs(README): 更新项目依赖并调整训练参数 1 месяц назад
test.py 6caff747d6 fix metrics.py ambiguity 1 год назад
train.py cba558448d Add files via upload 1 год назад
visualize.py c698f851e5 Add visualization code 9 месяцев назад

README.md

Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person Retrieval

GitHub PWC

Official PyTorch implementation of the paper Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person Retrieval. (CVPR 2023) arXiv

Updates

  • (3/23/2023) Add arXiv link for our paper.
  • (3/18/2023) Add download links of trained models and logs.
  • (3/17/2023) Ensure the reproducibility of our code.
  • (3/13/2023) Code released!

Highlights

The goal of this work is to enhance global text-to-image person retrieval performance, without requiring any additional supervision and inference cost. To achieve this, we utilize the full CLIP model as our feature extraction backbone. Additionally, we propose a novel cross-modal matching loss (SDM) and an Implicit Relation Reasoning module to mine fine-grained image-text relationships, enabling IRRA to learn more discriminative global image-text representations.

Usage

Requirements

we use single RTX3090 24G GPU for training and evaluation.

pytorch==1.9.0
torchvision==0.10.0
prettytable
easydict
yaml
ftfy
regex
tensorboard
numpy==1.26.4
tqdm
setuptools==59.5.0

Prepare Datasets

Download the CUHK-PEDES dataset from here, ICFG-PEDES dataset from here and RSTPReid dataset form here

Organize them in your dataset root dir folder as follows:

|-- your dataset root dir/
|   |-- <CUHK-PEDES>/
|       |-- imgs
|            |-- cam_a
|            |-- cam_b
|            |-- ...
|       |-- reid_raw.json
|
|   |-- <ICFG-PEDES>/
|       |-- imgs
|            |-- test
|            |-- train 
|       |-- ICFG_PEDES.json
|
|   |-- <RSTPReid>/
|       |-- imgs
|       |-- data_captions.json

Training

python train.py \
--name iira \
--img_aug \
--batch_size 64 \
--MLM \
--loss_names 'sdm+mlm+id' \
--dataset_name 'CUHK-PEDES' \
--root_dir 'your dataset root dir' \
--num_epoch 60

Testing

python test.py --config_file 'path/to/model_dir/configs.yaml'

IRRA on Text-to-Image Person Retrieval Results

CUHK-PEDES dataset

Method Backbone Rank-1 Rank-5 Rank-10 mAP mINP
CMPM/C RN50/LSTM 49.37 - 79.27 - -
DSSL RN50/BERT 59.98 80.41 87.56 - -
SSAN RN50/LSTM 61.37 80.15 86.73 - -
Han et al. RN101/Xformer 64.08 81.73 88.19 60.08 -
LGUR DeiT-Small/BERT 65.25 83.12 89.00 - -
IVT ViT-B-16/BERT 65.59 83.11 89.21 - -
CFine ViT-B-16/BERT 69.57 85.93 91.15 - -
CLIP ViT-B-16/Xformer 68.19 86.47 91.47 61.12 44.86
IRRA (ours) ViT-B-16/Xformer 73.38 89.93 93.71 66.13 50.24

Model & log for CUHK-PEDES

ICFG-PEDES dataset

Method Rank-1 Rank-5 Rank-10 mAP mINP
CMPM/C 43.51 65.44 74.26 - -
SSAN 54.23 72.63 79.53 - -
IVT 56.04 73.60 80.22 - -
CFine 60.83 76.55 82.42 - -
CLIP 56.74 75.72 82.26 31.84 5.03
IRRA (ours) 63.46 80.24 85.82 38.05 7.92

Model & log for ICFG-PEDES

RSTPReid dataset

Method Rank-1 Rank-5 Rank-10 mAP mINP
DSSL 39.05 62.60 73.95 - -
SSAN 43.50 67.80 77.15 - -
IVT 46.70 70.00 78.80 - -
CFine 50.55 72.50 81.60 - -
CLIP 54.05 80.70 88.00 43.41 22.31
IRRA (ours) 60.20 81.30 88.20 47.17 25.28

Model & log for RSTPReid

Acknowledgments

Some components of this code implementation are adopted from CLIP, TextReID and TransReID. We sincerely appreciate for their contributions.

Citation

If you find this code useful for your research, please cite our paper.

@inproceedings{cvpr23crossmodal,
  title={Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person Retrieval},
  author={Jiang, Ding and Ye, Mang},
  booktitle={IEEE International Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2023},
}

Contact

If you have any question, please feel free to contact us. E-mail: jiangding@whu.edu.cn, yemang@whu.edu.cn.