[ICCV 2021] Counterfactual Attention Learning for Fine-Grained Visual Categorization and Re-identification

Related tags

Text Data & NLP CAL
Overview

Counterfactual Attention Learning

Created by Yongming Rao*, Guangyi Chen*, Jiwen Lu, Jie Zhou

This repository contains PyTorch implementation for ICCV 2021 paper Counterfactual Attention Learning for Fine-Grained Visual Categorization and Re-identification [arXiv]

We propose to learn the attention with counterfactual causality, which provides a tool to measure the attention quality and a powerful supervisory signal to guide the learning process.

intro

CAL for Fine-Grained Visual Categorization

See CAL-FGVC.

CAL for Person Re-Identification

See CAL-ReID.

License

MIT License

Citation

If you find our work useful in your research, please consider citing:

@inproceedings{rao2021counterfactual,
  title={Counterfactual Attention Learning for Fine-Grained Visual Categorization and Re-identification},
  author={Rao, Yongming and Chen, Guangyi and Lu, Jiwen and Zhou, Jie},
  booktitle={ICCV},
  year={2021}
}
Comments
  • Pre-trained weights

    Pre-trained weights

    Hello everyone :) thanks for you nice work! Do you provide pre-trained weights for the person re-identification datasets somewhere? Thanks in advance! :)

    opened by JennySeidenschwarz 5
  • cannot reproduce the accuracy on CUB

    cannot reproduce the accuracy on CUB

    I am trying to reproduce the result of the CUB dataset, which is 90.6 acc ( table 1 in the paper). However, I use the same config and startup script as the code repo, but only get 90.03 acc for the last epoch. I notice that the total training epoch for fgvc task is not reported in the paper. So what is the proper epoch to get the 90.6 acc? Are there any other reasons that could affect reproducing the acc?

    Please see attachment for my training log. Thanks train.log !

    opened by yifanpu001 4
  • about feature matrix which is the input of the last fc layer

    about feature matrix which is the input of the last fc layer

    Sorry to bother you, I have a question about the fgvc part. Why the normalized feature_matrix and feature_matrix_hat need to multiply 100 before the fc layer?

    opened by lynlindasy 4
  • Output Feature During Inference Stage

    Output Feature During Inference Stage

    I quickly checked the model script baseline.py and found that you used the cls_score as output when doing inference. I am wondering if your published results were generated by this instead of features before classifier (which is regularly applied in popular reid framework).

    opened by morgenzhang 4
  • Reproduce Result for MSMT dataset

    Reproduce Result for MSMT dataset

    I am trying to reproduce the result as shown in the paper for the MSMT which is mAP@64% and [email protected]%; however, I could not do it. May I ask about the backbone you are using to get these results? Is it the same with the code in your repository or do you use different approach? And do you take the best case during training or the result after training for the whole 160 epoch? I am sorry if these questions bother you. Thank in advance!

    opened by petertran1811 2
Owner
Yongming Rao
Yongming Rao
A framework for evaluating Knowledge Graph Embedding Models in a fine-grained manner.

A framework for evaluating Knowledge Graph Embedding Models in a fine-grained manner.

NEC Laboratories Europe 13 Sep 8, 2022
Line as a Visual Sentence: Context-aware Line Descriptor for Visual Localization

Line as a Visual Sentence with LineTR This repository contains the inference code, pretrained model, and demo scripts of the following paper. It suppo

SungHo Yoon 158 Dec 27, 2022
HAIS_2GNN: 3D Visual Grounding with Graph and Attention

HAIS_2GNN: 3D Visual Grounding with Graph and Attention This repository is for the HAIS_2GNN research project. Tao Gu, Yue Chen Introduction The motiv

Yue Chen 1 Nov 26, 2022
Code for our ACL 2021 (Findings) Paper - Fingerprinting Fine-tuned Language Models in the wild .

?? Fingerprinting Fine-tuned Language Models in the wild This is the code and dataset for our ACL 2021 (Findings) Paper - Fingerprinting Fine-tuned La

LCS2-IIITDelhi 5 Sep 13, 2022
Stand-alone language identification system

langid.py readme Introduction langid.py is a standalone Language Identification (LangID) tool. The design principles are as follows: Fast Pre-trained

null 2k Jan 4, 2023
Stand-alone language identification system

langid.py readme Introduction langid.py is a standalone Language Identification (LangID) tool. The design principles are as follows: Fast Pre-trained

null 1.7k Feb 7, 2021
Stand-alone language identification system

langid.py readme Introduction langid.py is a standalone Language Identification (LangID) tool. The design principles are as follows: Fast Pre-trained

null 1.7k Feb 17, 2021
Framework for fine-tuning pretrained transformers for Named-Entity Recognition (NER) tasks

NERDA Not only is NERDA a mesmerizing muppet-like character. NERDA is also a python package, that offers a slick easy-to-use interface for fine-tuning

Ekstra Bladet 141 Dec 30, 2022
Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP

Pretrain and Fine-tune a T5 model with Flax on GCP This tutorial details how pretrain and fine-tune a FlaxT5 model from HuggingFace using a TPU VM ava

Gabriele Sarti 41 Nov 18, 2022
NLPIR tutorial: pretrain for IR. pre-train on raw textual corpus, fine-tune on MS MARCO Document Ranking

pretrain4ir_tutorial NLPIR tutorial: pretrain for IR. pre-train on raw textual corpus, fine-tune on MS MARCO Document Ranking 用作NLPIR实验室, Pre-training

ZYMa 12 Apr 7, 2022
Fine-tune GPT-3 with a Google Chat conversation history

Google Chat GPT-3 This repo will help you fine-tune GPT-3 with a Google Chat conversation history. The trained model will be able to converse as one o

Nate Baer 7 Dec 10, 2022
spaCy-wrap: For Wrapping fine-tuned transformers in spaCy pipelines

spaCy-wrap: For Wrapping fine-tuned transformers in spaCy pipelines spaCy-wrap is minimal library intended for wrapping fine-tuned transformers from t

Kenneth Enevoldsen 32 Dec 29, 2022
Multilingual Emotion classification using BERT (fine-tuning). Published at the WASSA workshop (ACL2022).

XLM-EMO: Multilingual Emotion Prediction in Social Media Text Abstract Detecting emotion in text allows social and computational scientists to study h

MilaNLP 35 Sep 17, 2022
[ICCV 2021] Instance-level Image Retrieval using Reranking Transformers

Instance-level Image Retrieval using Reranking Transformers Fuwen Tan, Jiangbo Yuan, Vicente Ordonez, ICCV 2021. Abstract Instance-level image retriev

UVA Computer Vision 86 Dec 28, 2022
Learning Spatio-Temporal Transformer for Visual Tracking

STARK The official implementation of the paper Learning Spatio-Temporal Transformer for Visual Tracking Highlights The strongest performances Tracker

Multimedia Research 485 Jan 4, 2023
Visual Automata is a Python 3 library built as a wrapper for Caleb Evans' Automata library to add more visualization features.

Visual Automata Copyright 2021 Lewi Lie Uberg Released under the MIT license Visual Automata is a Python 3 library built as a wrapper for Caleb Evans'

Lewi Uberg 55 Nov 17, 2022
A simple visual front end to the Maya UE4 RBF plugin delivered with MetaHumans

poseWrangler Overview PoseWrangler is a simple UI to create and edit pose-driven relationships in Maya using the MayaUE4RBF plugin. This plugin is dis

Christopher Evans 105 Dec 18, 2022
TalkNet: Audio-visual active speaker detection Model

Is someone talking? TalkNet: Audio-visual active speaker detection Model This repository contains the code for our ACM MM 2021 paper, TalkNet, an acti

null 142 Dec 14, 2022