PyTorch implementation of our ICCV paper DeFRCN: Decoupled Faster R-CNN for Few-Shot Object Detection.

Related tags

Deep Learning DeFRCN
Overview

Introduction

This repo contains the official PyTorch implementation of our ICCV paper DeFRCN: Decoupled Faster R-CNN for Few-Shot Object Detection.

Updates!!

  • 【2021/10/10】 We release the official PyTorch implementation of DeFRCN.
  • 【2021/08/20】 We have uploaded our paper (long version with supplementary material) on arxiv, review it for more details.

Quick Start

1. Check Requirements

  • Linux with Python >= 3.6
  • PyTorch >= 1.6 & torchvision that matches the PyTorch version.
  • CUDA 10.1, 10.2
  • GCC >= 4.9

2. Build DeFRCN

  • Clone Code
    git clone https://github.com/er-muyue/DeFRCN.git
    cd DeFRCN
    
  • Create a virtual environment (optional)
    virtualenv defrcn
    cd /path/to/venv/defrcn
    source ./bin/activate
    
  • Install PyTorch 1.6.0 with CUDA 10.1
    pip3 install torch==1.6.0+cu101 torchvision==0.7.0+cu101 -f https://download.pytorch.org/whl/torch_stable.html
  • Install Detectron2
    python3 -m pip install detectron2==0.3 -f https://dl.fbaipublicfiles.com/detectron2/wheels/cu101/torch1.6/index.html
    
    • If you use other version of PyTorch/CUDA, check the latest version of Detectron2 in this page: Detectron2.
    • Sorry for that I don’t have enough time to test on more versions, if you run into problems with other versions, please let me know.
  • Install other requirements.
    python3 -m pip install -r requirements.txt
    

3. Prepare Data and Weights

  • Data Preparation
    • We evaluate our models on two datasets for both FSOD and G-FSOD settings:

      Dataset Size GoogleDrive BaiduYun Note
      VOC2007 0.8G download download -
      VOC2012 3.5G download download -
      vocsplit <1M download download refer from TFA
      COCO ~19G - - download from offical
      cocosplit 174M download download refer from TFA
    • Unzip the downloaded data-source to datasets and put it into your project directory:

        ...
        datasets
          | -- coco (trainval2014/*.jpg, val2014/*.jpg, annotations/*.json)
          | -- cocosplit
          | -- VOC2007
          | -- VOC2012
          | -- vocsplit
        defrcn
        tools
        ...
      
  • Weights Preparation
    • We use the imagenet pretrain weights to initialize our model. Download the same models from here: GoogleDrive BaiduYun
    • The extract code for all BaiduYun link is 0000

4. Training and Evaluation

For ease of training and evaluation over multiple runs, we integrate the whole pipeline of few-shot object detection into one script run_*.sh, including base pre-training and novel-finetuning (both FSOD and G-FSOD).

  • To reproduce the results on VOC, EXP_NAME can be any string (e.g defrcn, or something) and SPLIT_ID must be 1 or 2 or 3 (we consider 3 random splits like other papers).
    bash run_voc.sh EXP_NAME SPLIT_ID (1, 2 or 3)
    
  • To reproduce the results on COCO, EXP_NAME can be any string (e.g defrcn, or something)
    bash run_coco.sh EXP_NAME
    
  • Please read the details of few-shot object detection pipeline in run_*.sh, you need change IMAGENET_PRETRAIN* to your path.

Results on COCO Benchmark

  • Few-shot Object Detection

    Method mAPnovel
    Shot 1 2 3 5 10 30
    FRCN-ft 1.0* 1.8* 2.8* 4.0* 6.5 11.1
    FSRW - - - - 5.6 9.1
    MetaDet - - - - 7.1 11.3
    MetaR-CNN - - - - 8.7 12.4
    TFA 4.4* 5.4* 6.0* 7.7* 10.0 13.7
    MPSR 5.1* 6.7* 7.4* 8.7* 9.8 14.1
    FSDetView 4.5 6.6 7.2 10.7 12.5 14.7
    DeFRCN (Our Paper) 9.3 12.9 14.8 16.1 18.5 22.6
    DeFRCN (This Repo) 9.7 13.1 14.5 15.6 18.4 22.6
  • Generalized Few-shot Object Detection

    Method mAPnovel
    Shot 1 2 3 5 10 30
    FRCN-ft 1.7 3.1 3.7 4.6 5.5 7.4
    TFA 1.9 3.9 5.1 7 9.1 12.1
    FSDetView 3.2 4.9 6.7 8.1 10.7 15.9
    DeFRCN (Our Paper) 4.8 8.5 10.7 13.6 16.8 21.2
    DeFRCN (This Repo) 4.8 8.5 10.7 13.5 16.7 21.0
  • * indicates that the results are reproduced by us with their source code.
  • It's normal to observe -0.3~+0.3AP noise between your results and this repo.
  • The results of mAPbase and mAPall for G-FSOD are list here GoogleDrive, BaiduYun.
  • If you have any problem of above results in this repo, you can download configs and train logs from GoogleDrive, BaiduYun.

Results on VOC Benchmark

  • Few-shot Object Detection

    Method Split-1 Split-2 Split-3
    Shot 1 2 3 5 10 1 2 3 5 10 1 2 3 5 10
    YOLO-ft 6.6 10.7 12.5 24.8 38.6 12.5 4.2 11.6 16.1 33.9 13.0 15.9 15.0 32.2 38.4
    FRCN-ft 13.8 19.6 32.8 41.5 45.6 7.9 15.3 26.2 31.6 39.1 9.8 11.3 19.1 35.0 45.1
    FSRW 14.8 15.5 26.7 33.9 47.2 15.7 15.2 22.7 30.1 40.5 21.3 25.6 28.4 42.8 45.9
    MetaDet 18.9 20.6 30.2 36.8 49.6 21.8 23.1 27.8 31.7 43.0 20.6 23.9 29.4 43.9 44.1
    MetaR-CNN 19.9 25.5 35.0 45.7 51.5 10.4 19.4 29.6 34.8 45.4 14.3 18.2 27.5 41.2 48.1
    TFA 39.8 36.1 44.7 55.7 56.0 23.5 26.9 34.1 35.1 39.1 30.8 34.8 42.8 49.5 49.8
    MPSR 41.7 - 51.4 55.2 61.8 24.4 - 39.2 39.9 47.8 35.6 - 42.3 48.0 49.7
    DeFRCN (Our Paper) 53.6 57.5 61.5 64.1 60.8 30.1 38.1 47.0 53.3 47.9 48.4 50.9 52.3 54.9 57.4
    DeFRCN (This Repo) 55.1 57.4 61.1 64.6 61.5 32.1 40.5 47.9 52.9 47.5 48.9 51.9 52.3 55.7 59.0
  • Generalized Few-shot Object Detection

    Method Split-1 Split-2 Split-3
    Shot 1 2 3 5 10 1 2 3 5 10 1 2 3 5 10
    FRCN-ft 9.9 15.6 21.6 28.0 52.0 9.4 13.8 17.4 21.9 39.7 8.1 13.9 19 23.9 44.6
    FSRW 14.2 23.6 29.8 36.5 35.6 12.3 19.6 25.1 31.4 29.8 12.5 21.3 26.8 33.8 31.0
    TFA 25.3 36.4 42.1 47.9 52.8 18.3 27.5 30.9 34.1 39.5 17.9 27.2 34.3 40.8 45.6
    FSDetView 24.2 35.3 42.2 49.1 57.4 21.6 24.6 31.9 37.0 45.7 21.2 30.0 37.2 43.8 49.6
    DeFRCN (Our Paper) 40.2 53.6 58.2 63.6 66.5 29.5 39.7 43.4 48.1 52.8 35.0 38.3 52.9 57.7 60.8
    DeFRCN (This Repo) 43.8 57.5 61.4 65.3 67.0 31.5 40.9 45.6 50.1 52.9 38.2 50.9 54.1 59.2 61.9
  • Note that we change the λGDL-RCNN for VOC to 0.001 (0.01 in paper) and get better performance, check the configs for more details.

  • The results of mAPbase and mAPall for G-FSOD are list here GoogleDrive, BaiduYun.

  • If you have any problem of above results in this repo, you can download configs and logs from GoogleDrive, BaiduYun.

Acknowledgement

This repo is developed based on TFA and Detectron2. Please check them for more details and features.

Citing

If you use this work in your research or wish to refer to the baseline results published here, please use the following BibTeX entries:

@inproceedings{qiao2021defrcn,
  title={DeFRCN: Decoupled Faster R-CNN for Few-Shot Object Detection},
  author={Qiao, Limeng and Zhao, Yuxuan and Li, Zhiyuan and Qiu, Xi and Wu, Jianan and Zhang, Chi},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  pages={8681--8690},
  year={2021}
}
Comments
  • About Res5ROIHeads

    About Res5ROIHeads

    Hi! Thanks for your great work! I have noticed that in the config you used Res5ROIHeads(and C4 features) instead of StandardROIHeads(and FPN features) used by TFA. Is there some special consideration behind this design decision? Or have you done some experiments on whether the model works (or even works better) with FPN?

    opened by RyanXLi 10
  • Unable to reproduce the results on Pascal VOC datasets

    Unable to reproduce the results on Pascal VOC datasets

    Hi, I tried to reproduce the G-FSOD results on Pascal VOC using the default config, but I am not able to reproduce the 2/3/5/10 shot results. Only the 1-shot result can be reproduced. The only change I made is the GPU number, which is changed from 8 to 2. Batch size and other config are kept unchanged.

    G-FSOD, split 1, average results of seed 0-9 | shot | reported AP (github repo) | reproduced AP | | ---- | ---- | ---- | | 1-shot | 43.8 | 41.95 | | 2-shot | 57.5 | 44.24 | | 3-shot | 61.4 | 45.09 | | 5-shot | 65.3 | 46.70 | | 10-shot | 67.0 | 47.51 |

    training_script.txt results.txt

    Is there anything wrong with my experiment? Also, could you please provide a 10-shot trained model? Thank you very much.

    opened by Hugotong6425 5
  • where is the `PCB` module?

    where is the `PCB` module?

    Hi, I am sorry that I can not find the code where the PCB module for re-weighting the box-predictor' cls scores refers to. Hope someone who knows could help me :), thx.

    opened by Wei-i 5
  • AssertionError: Checkpoint checkpoints/voc/defrcn/defrcn_det_r101_base3/model_reset_surgery.pth not found

    AssertionError: Checkpoint checkpoints/voc/defrcn/defrcn_det_r101_base3/model_reset_surgery.pth not found

    i used bash run_voc.sh

    File "/.../conda_envs/tpami/lib/python3.7/site-packages/fvcore/common/checkpoint.py", line 153, in load assert os.path.isfile(path), "Checkpoint {} not found!".format(path) AssertionError: Checkpoint checkpoints/voc/defrcn/defrcn_det_r101_base3/model_reset_surgery.pth not found!

    how to solve it? Thx.

    Traceback (most recent call last): File "/.../conda_envs/tpami/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 59, in _wrap fn(i, *args) File "/.../conda_envs/tpami/lib/python3.7/site-packages/detectron2/engine/launch.py", line 126, in _distributed_worker main_func(*args) File "/.../fsod/DeFRCN-main/main.py", line 61, in main trainer.resume_or_load(resume=args.resume) File "/.../fsod/DeFRCN-main/defrcn/engine/defaults.py", line 290, in resume_or_load self.cfg.MODEL.WEIGHTS, resume=resume File "/.../conda_envs/tpami/lib/python3.7/site-packages/fvcore/common/checkpoint.py", line 227, in resume_or_load return self.load(path, checkpointables=[]) File "/.../conda_envs/tpami/lib/python3.7/site-packages/fvcore/common/checkpoint.py", line 153, in load assert os.path.isfile(path), "Checkpoint {} not found!".format(path) AssertionError: Checkpoint checkpoints/voc/defrcn/defrcn_det_r101_base3/model_reset_remove.pth not found!

    opened by zhengfang1997 4
  • ValueError: could not convert string to float

    ValueError: could not convert string to float

    i tried to bash run_voc.sh:

    Traceback (most recent call last):
      File "tools/extract_results.py", line 59, in <module>
        main()
      File "tools/extract_results.py", line 34, in main
        results.append([fid] + [float(x) for x in res_info.split(':')[-1].split(',')])
      File "tools/extract_results.py", line 34, in <listcomp>
        results.append([fid] + [float(x) for x in res_info.split(':')[-1].split(',')])
    ValueError: could not convert string to float: ' [Checkpointer] Loading from checkpoints/voc/defrcn/defrcn_det_r101_base1/model_reset_remove.pth ...'
    
    

    how to solve the problem, thx

    opened by zhengfang1997 4
  • Predicted boxes or scores contain Inf/NaN. Training has diverged

    Predicted boxes or scores contain Inf/NaN. Training has diverged

    When trying to run the training code, I encountered the following error: image In the previous process, my steps were:

    • Refer to the readme, install torch, detectron and other environment dependencies based on conda

    • Data preparation: vocdata in readme is used image

    • Modify operating errors: 1、in the run_*.sh ,change into image 2、as the picture shows(Refer to:https://giters.com/fanq15/FewX/issues/40?amp=1) image

    • Intercept part of the training code and try to start training image

    then i meet the error:Predicted boxes or scores contain Inf/NaN. Training has diverged。I got the same error when I tried on two different servers (cuda10.2, cuda11.1 both torch1.8)

    I tried to modify the learning rate-related settings(to different value , vary from 0.00005 to 0.95) in the configuration file, but no effect image

    opened by henanwg 4
  • problem in Running novels

    problem in Running novels

    Thank you for sharing your code

    I used of DeFRCN code (python3.7 and detectron 0.3, cuda 10.2) but I had AP base 12 on my own data, I downloaded pretrain model and replace path into run_coco.sh but I had AP base 12 on my own data and in novels result.txt is empty and i have this Error: can you help me?

    Traceback (most recent call last): File "tools/extract_results.py", line 59, in main() File "tools/extract_results.py", line 34, in main results.append([fid] + [float(x) for x in res_info.split(':')[-1].split(',')]) File "tools/extract_results.py", line 34, in results.append([fid] + [float(x) for x in res_info.split(':')[-1].split(',')]) ValueError: could not convert string to float: " ResNet.make_stage(first_stride=) is deprecated! Use 'stride_per_block' or 'stride' instead."
    opened by salehnia 3
  • How long will the code be released?

    How long will the code be released?

    Thanks for your great work! The 10-shot COCO nAP can achieve 18 :), by the way, I think DeFRCN is the combination of Transfer-Learning based FSOD and Meta-Learning based FSOD.

    opened by Wei-i 3
  • RuntimeError: The server socket has failed to listen on any local network address. The server socket has failed to bind to [::]:50153 (errno: 98 - Address already in use). The server socket has failed to bind to ?UNKNOWN? (errno: 98 - Address already in use).

    RuntimeError: The server socket has failed to listen on any local network address. The server socket has failed to bind to [::]:50153 (errno: 98 - Address already in use). The server socket has failed to bind to ?UNKNOWN? (errno: 98 - Address already in use).

    I am getting this error while running this command python3 main.py --num-gpus 8 --config-file configs/voc/defrcn_det_r101_base${SPLIT_ID}.yaml
    --opts MODEL.WEIGHTS ${IMAGENET_PRETRAIN}
    OUTPUT_DIR ${SAVE_DIR}/defrcn_det_r101_base${SPLIT_ID}

    opened by MALLI7622 2
  • Whether the PCB is used ?

    Whether the PCB is used ?

    Hello, author. I found "_CC.TEST.PCB_ENABLE = False" in "defrcn/config/defaults.py", does it mean that PCB is not used in the inference process ? When it is False, "outputs = pcb.execute_calibration(inputs, outputs)" is not executed in "evaluator.py". Hope to get a reply soon. thanks.

    opened by Liangbing-Sa 2
  • What if 'RCNN Head' is trainable as well?

    What if 'RCNN Head' is trainable as well?

    Nice paper, thanks for releasing your source code. I got impressed that the cnn backbone is trainable in your theme and make effect. But what if 'RCNN Head' is trainable as well? Have you ever done about it? or is there any reason that you freeze this module?

    opened by mandal4 2
  • Why  the different number of images on inference?

    Why the different number of images on inference?

    My inference log: defrcn.evaluation.evaluator INFO: Start inference on 4952 images Your inference log: defrcn.evaluation.evaluator INFO: Start inference on 619 images

    My config yaml file: [11/20 03:07:39] detectron2 INFO: Command line arguments: Namespace(config_file='configs/voc/defrcn_fsod_r101_novel1_1shot_seed1.yaml', dist_url='tcp://127.0.0.1:50152', end_iter=-1, eval_all=False, eval_during_train=False, eval_iter=-1, eval_only=False, machine_rank=0, num_gpus=0, num_machines=1, opts=['MODEL.WEIGHTS', 'checkpoints/voc/CAA/defrcn_det_r101_base1/model_reset_remove.pth', 'OUTPUT_DIR', 'checkpoints/voc/CAA/defrcn_fsod_r101_novel1/fsrw-like/1shot_seed1', 'TEST.PCB_MODELPATH', '/home/ubuntu/FSOD_DeFRCN/ImageNetPretrained/torchvision/resnet101-5d3b4d8f.pth', 'SOLVER.IMS_PER_BATCH', '6'], resume=False, start_iter=-1) [11/20 03:07:39] detectron2 INFO: Contents of args.config_file=configs/voc/defrcn_fsod_r101_novel1_1shot_seed1.yaml: BASE: "../Base-RCNN.yaml" MODEL: WEIGHTS: "/Path/to/Base/Pretrain/Weight" MASK_ON: False BACKBONE: FREEZE: False RESNETS: DEPTH: 101 RPN: ENABLE_DECOUPLE: True BACKWARD_SCALE: 0.0 FREEZE: False ROI_HEADS: ENABLE_DECOUPLE: True BACKWARD_SCALE: 0.001 NUM_CLASSES: 5 FREEZE_FEAT: True # FREEZE_FEAT: False CLS_DROPOUT: True INPUT:

    MIN_SIZE_TRAIN: (480, 512, 544, 576, 608, 640, 672, 704, 736, 768, 800)

    MIN_SIZE_TRAIN: (608, 640, 672, 704, 736, 768, 800) MIN_SIZE_TEST: 800 DATASETS: TRAIN: ("voc_2007_trainval_novel1_1shot_seed1", ) TEST: ("voc_2007_test_novel1",) SOLVER: IMS_PER_BATCH: 16 BASE_LR: 0.02 STEPS: (600, ) MAX_ITER: 800 CHECKPOINT_PERIOD: 100000 WARMUP_ITERS: 0 TEST: PCB_ENABLE: True PCB_MODELPATH: "/Path/to/ImageNet/Pre-Train/Weight" OUTPUT_DIR: "/Path/to/Output/Dir"

    opened by ZhiXianZ 0
  • AssertionError: Checkpoint checkpoints/voc/mydefrcn/defrcn_det_r101_base1/model_reset_remove.pth not found!

    AssertionError: Checkpoint checkpoints/voc/mydefrcn/defrcn_det_r101_base1/model_reset_remove.pth not found!

    When I run bash run_voc.sh mydefrcn 1, I get the following error: AssertionError: Checkpoint checkpoints/voc/mydefrcn/defrcn_det_r101_base1/model_reset_remove.pth not found!. Where can I get model_reset_remove.pth?

    opened by yewanzhuo 1
  • argument 'alpha' must be Number, not NoneType

    argument 'alpha' must be Number, not NoneType

    Hello, thanks for your work. I met a bug:

    d_p = d_p.add(param, alpha=weight_decay) TypeError: add(): argument 'alpha' must be Number, not NoneType

    The detectron2 version is 0.6.Please tell me how could I solve this problem.

    opened by Yxt1212 2
  • Question about ground truth boxes which are used in training and PrototypicalCalibrationBlock

    Question about ground truth boxes which are used in training and PrototypicalCalibrationBlock

    I found some errors in your code. In meta_voc.py, when we got more objects of a class than a shot(K), the K samples of the class are randomly selected among objects. However, the seed for sampling is not fixed. As a result, the sampled objects for training and the sampled objects for PCB may not be the same.

    When I set a seed to extract the same samples, the performances were dramatically degraded in 2-shot and 3-shot. Avg performance: Performances of VOC(split-1) with original code. | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | --> 1-shot | 30.14 | 51.97 | 31.07 | 30.14 | 51.97 | 31.07 | --> 2-shot | 31.80 | 53.00 | 34.33 | 31.80 | 53.00 | 34.33 | --> 3-shot | 36.79 | 59.16 | 39.64 | 36.79 | 59.16 | 39.64 | --> 5-shot | 36.85 | 61.53 | 40.12 | 36.85 | 61.53 | 40.12 | --> 10-shot | 37.77 | 59.47 | 41.40 | 37.77 | 59.47 | 41.40 |

    Performances of VOC(split-1) with fixing seed for sampling. | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | --> 1-shot | 30.13 | 53.46 | 29.86 | 30.13 | 53.46 | 29.86 | --> 2-shot | 29.09 | 49.41 | 31.12 | 29.09 | 49.41 | 31.12 | --> 3-shot | 30.76 | 50.77 | 33.00 | 30.76 | 50.77 | 33.00 | --> 5-shot | 35.87 | 60.16 | 38.98 | 35.87 | 60.16 | 38.98 | --> 10-shot | 38.04 | 60.19 | 41.74 | 38.04 | 60.19 | 41.74 |

    All performances: Performances of VOC(split-1) with original code. --> 1-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 30.21 | 53.89 | 30.09 | 30.21 | 53.89 | 30.09 | \n| 1 | 29.97 | 50.93 | 31.75 | 29.97 | 50.93 | 31.75 | \n| 2 | 29.98 | 52.97 | 30.34 | 29.98 | 52.97 | 30.34 | \n| 3 | 29.69 | 50.40 | 30.14 | 29.69 | 50.40 | 30.14 | \n| 4 | 30.84 | 53.84 | 31.65 | 30.84 | 53.84 | 31.65 | \n| 5 | 30.22 | 51.04 | 32.58 | 30.22 | 51.04 | 32.58 | \n| 6 | 28.80 | 49.20 | 29.87 | 28.80 | 49.20 | 29.87 | \n| 7 | 30.62 | 52.85 | 31.61 | 30.62 | 52.85 | 31.61 | \n| 8 | 31.46 | 54.46 | 31.72 | 31.46 | 54.46 | 31.72 | \n| 9 | 29.58 | 50.07 | 30.96 | 29.58 | 50.07 | 30.96 | \n| \xce\xbc | 30.14 | 51.97 | 31.07 | 30.14 | 51.97 | 31.07 | \n| c | 0.43 | 1.09 | 0.54 | 0.43 | 1.09 | 0.54 |'

    --> 2-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 31.00 | 50.91 | 34.27 | 31.00 | 50.91 | 34.27 | \n| 1 | 33.64 | 54.77 | 36.80 | 33.64 | 54.77 | 36.80 | \n| 2 | 29.67 | 51.01 | 30.59 | 29.67 | 51.01 | 30.59 | \n| 3 | 33.10 | 55.59 | 34.85 | 33.10 | 55.59 | 34.85 | \n| 4 | 29.30 | 50.05 | 31.24 | 29.30 | 50.05 | 31.24 | \n| 5 | 33.53 | 54.35 | 37.70 | 33.53 | 54.35 | 37.70 | \n| 6 | 33.85 | 56.59 | 36.54 | 33.85 | 56.59 | 36.54 | \n| 7 | 31.73 | 52.02 | 34.55 | 31.73 | 52.02 | 34.55 | \n| 8 | 30.56 | 51.27 | 32.35 | 30.56 | 51.27 | 32.35 | \n| 9 | 31.61 | 53.47 | 34.46 | 31.61 | 53.47 | 34.46 | \n| \xce\xbc | 31.80 | 53.00 | 34.33 | 31.80 | 53.00 | 34.33 | \n| c | 0.99 | 1.33 | 1.39 | 0.99 | 1.33 | 1.39 |'

    --> 3-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 36.01 | 57.80 | 39.41 | 36.01 | 57.80 | 39.41 | \n| 1 | 36.86 | 59.94 | 39.32 | 36.86 | 59.94 | 39.32 | \n| 2 | 37.61 | 60.41 | 40.12 | 37.61 | 60.41 | 40.12 | \n| 3 | 36.32 | 58.89 | 38.80 | 36.32 | 58.89 | 38.80 | \n| 4 | 35.89 | 57.38 | 39.11 | 35.89 | 57.38 | 39.11 | \n| 5 | 37.37 | 59.26 | 40.41 | 37.37 | 59.26 | 40.41 | \n| 6 | 36.53 | 58.73 | 38.78 | 36.53 | 58.73 | 38.78 | \n| 7 | 35.94 | 58.29 | 39.30 | 35.94 | 58.29 | 39.30 | \n| 8 | 37.95 | 60.48 | 40.80 | 37.95 | 60.48 | 40.80 | \n| 9 | 37.39 | 60.38 | 40.37 | 37.39 | 60.38 | 40.37 | \n| \xce\xbc | 36.79 | 59.16 | 39.64 | 36.79 | 59.16 | 39.64 | \n| c | 0.44 | 0.66 | 0.43 | 0.44 | 0.66 | 0.43 |'

    --> 5-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 36.25 | 60.70 | 39.63 | 36.25 | 60.70 | 39.63 | \n| 1 | 37.34 | 61.82 | 40.60 | 37.34 | 61.82 | 40.60 | \n| 2 | 37.53 | 62.76 | 40.59 | 37.53 | 62.76 | 40.59 | \n| 3 | 36.82 | 61.46 | 39.94 | 36.82 | 61.46 | 39.94 | \n| 4 | 36.15 | 60.92 | 39.71 | 36.15 | 60.92 | 39.71 | \n| 5 | 35.80 | 59.40 | 39.25 | 35.80 | 59.40 | 39.25 | \n| 6 | 36.34 | 60.42 | 39.54 | 36.34 | 60.42 | 39.54 | \n| 7 | 37.33 | 62.64 | 40.73 | 37.33 | 62.64 | 40.73 | \n| 8 | 37.63 | 62.97 | 41.25 | 37.63 | 62.97 | 41.25 | \n| 9 | 37.33 | 62.18 | 39.96 | 37.33 | 62.18 | 39.96 | \n| \xce\xbc | 36.85 | 61.53 | 40.12 | 36.85 | 61.53 | 40.12 | \n| c | 0.39 | 0.68 | 0.37 | 0.39 | 0.68 | 0.37 |'

    --> 10-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 38.03 | 59.83 | 42.59 | 38.03 | 59.83 | 42.59 | \n| 1 | 35.99 | 55.93 | 40.42 | 35.99 | 55.93 | 40.42 | \n| 2 | 37.99 | 59.70 | 41.05 | 37.99 | 59.70 | 41.05 | \n| 3 | 37.93 | 59.53 | 41.94 | 37.93 | 59.53 | 41.94 | \n| 4 | 37.91 | 60.19 | 41.63 | 37.91 | 60.19 | 41.63 | \n| 5 | 37.68 | 58.66 | 41.22 | 37.68 | 58.66 | 41.22 | \n| 6 | 37.28 | 58.61 | 40.34 | 37.28 | 58.61 | 40.34 | \n| 7 | 38.28 | 60.35 | 40.89 | 38.28 | 60.35 | 40.89 | \n| 8 | 38.25 | 61.55 | 41.90 | 38.25 | 61.55 | 41.90 | \n| 9 | 38.34 | 60.39 | 41.99 | 38.34 | 60.39 | 41.99 | \n| \xce\xbc | 37.77 | 59.47 | 41.40 | 37.77 | 59.47 | 41.40 | \n| c | 0.41 | 0.89 | 0.43 | 0.41 | 0.89 | 0.43 |'

    Performances of VOC(split-1) with fixing seed for sampling. --> 1-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 30.15 | 53.19 | 30.08 | 30.15 | 53.19 | 30.08 | \n| 1 | 29.85 | 52.90 | 29.37 | 29.85 | 52.90 | 29.37 | \n| 2 | 30.52 | 54.41 | 30.51 | 30.52 | 54.41 | 30.51 | \n| 3 | 30.09 | 52.72 | 29.73 | 30.09 | 52.72 | 29.73 | \n| 4 | 30.55 | 53.91 | 30.96 | 30.55 | 53.91 | 30.96 | \n| 5 | 30.16 | 53.38 | 29.83 | 30.16 | 53.38 | 29.83 | \n| 6 | 30.21 | 53.44 | 29.77 | 30.21 | 53.44 | 29.77 | \n| 7 | 29.89 | 53.39 | 29.30 | 29.89 | 53.39 | 29.30 | \n| 8 | 29.80 | 52.90 | 29.53 | 29.80 | 52.90 | 29.53 | \n| 9 | 30.11 | 54.39 | 29.55 | 30.11 | 54.39 | 29.55 | \n| \xce\xbc | 30.13 | 53.46 | 29.86 | 30.13 | 53.46 | 29.86 | \n| c | 0.15 | 0.35 | 0.31 | 0.15 | 0.35 | 0.31 |'

    --> 2-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 28.62 | 49.23 | 31.43 | 28.62 | 49.23 | 31.43 | \n| 1 | 28.67 | 48.65 | 30.34 | 28.67 | 48.65 | 30.34 | \n| 2 | 29.21 | 49.89 | 30.79 | 29.21 | 49.89 | 30.79 | \n| 3 | 28.80 | 48.55 | 31.55 | 28.80 | 48.55 | 31.55 | \n| 4 | 29.02 | 49.07 | 31.11 | 29.02 | 49.07 | 31.11 | \n| 5 | 29.01 | 50.02 | 30.51 | 29.01 | 50.02 | 30.51 | \n| 6 | 29.05 | 48.80 | 30.78 | 29.05 | 48.80 | 30.78 | \n| 7 | 29.50 | 50.18 | 31.46 | 29.50 | 50.18 | 31.46 | \n| 8 | 29.23 | 49.88 | 31.28 | 29.23 | 49.88 | 31.28 | \n| 9 | 29.79 | 49.81 | 31.95 | 29.79 | 49.81 | 31.95 | \n| \xce\xbc | 29.09 | 49.41 | 31.12 | 29.09 | 49.41 | 31.12 | \n| c | 0.21 | 0.36 | 0.30 | 0.21 | 0.36 | 0.30 |'

    --> 3-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 30.79 | 50.88 | 33.44 | 30.79 | 50.88 | 33.44 | \n| 1 | 31.04 | 50.74 | 33.55 | 31.04 | 50.74 | 33.55 | \n| 2 | 31.12 | 51.27 | 33.69 | 31.12 | 51.27 | 33.69 | \n| 3 | 30.59 | 51.02 | 33.14 | 30.59 | 51.02 | 33.14 | \n| 4 | 30.30 | 50.43 | 32.69 | 30.30 | 50.43 | 32.69 | \n| 5 | 30.71 | 50.90 | 32.63 | 30.71 | 50.90 | 32.63 | \n| 6 | 31.22 | 50.67 | 33.35 | 31.22 | 50.67 | 33.35 | \n| 7 | 30.88 | 50.76 | 32.11 | 30.88 | 50.76 | 32.11 | \n| 8 | 30.80 | 50.80 | 32.76 | 30.80 | 50.80 | 32.76 | \n| 9 | 30.16 | 50.28 | 32.61 | 30.16 | 50.28 | 32.61 | \n| \xce\xbc | 30.76 | 50.77 | 33.00 | 30.76 | 50.77 | 33.00 | \n| c | 0.20 | 0.16 | 0.30 | 0.20 | 0.16 | 0.30 |'

    --> 5-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 35.84 | 60.28 | 39.79 | 35.84 | 60.28 | 39.79 | \n| 1 | 35.89 | 59.92 | 38.88 | 35.89 | 59.92 | 38.88 | \n| 2 | 35.78 | 60.51 | 39.01 | 35.78 | 60.51 | 39.01 | \n| 3 | 35.90 | 60.07 | 39.02 | 35.90 | 60.07 | 39.02 | \n| 4 | 35.88 | 59.74 | 37.86 | 35.88 | 59.74 | 37.86 | \n| 5 | 35.92 | 60.21 | 39.00 | 35.92 | 60.21 | 39.00 | \n| 6 | 35.86 | 60.27 | 39.38 | 35.86 | 60.27 | 39.38 | \n| 7 | 35.89 | 60.11 | 38.39 | 35.89 | 60.11 | 38.39 | \n| 8 | 35.97 | 60.27 | 39.47 | 35.97 | 60.27 | 39.47 | \n| 9 | 35.75 | 60.21 | 39.02 | 35.75 | 60.21 | 39.02 | \n| \xce\xbc | 35.87 | 60.16 | 38.98 | 35.87 | 60.16 | 38.98 | \n| c | 0.04 | 0.13 | 0.32 | 0.04 | 0.13 | 0.32 |'

    --> 10-shot b'| | AP | AP50 | AP75 | nAP | nAP50 | nAP75 | \n|:---|:------|:-------|:-------|:------|:--------|:--------| \n| 0 | 38.03 | 60.52 | 41.19 | 38.03 | 60.52 | 41.19 | \n| 1 | 38.24 | 59.83 | 42.42 | 38.24 | 59.83 | 42.42 | \n| 2 | 38.10 | 59.79 | 41.78 | 38.10 | 59.79 | 41.78 | \n| 3 | 38.29 | 60.53 | 42.64 | 38.29 | 60.53 | 42.64 | \n| 4 | 39.06 | 60.68 | 42.44 | 39.06 | 60.68 | 42.44 | \n| 5 | 37.65 | 60.41 | 40.93 | 37.65 | 60.41 | 40.93 | \n| 6 | 37.30 | 59.76 | 40.94 | 37.30 | 59.76 | 40.94 | \n| 7 | 37.56 | 59.60 | 41.43 | 37.56 | 59.60 | 41.43 | \n| 8 | 37.50 | 60.42 | 41.32 | 37.50 | 60.42 | 41.32 | \n| 9 | 38.70 | 60.39 | 42.31 | 38.70 | 60.39 | 42.31 | \n| \xce\xbc | 38.04 | 60.19 | 41.74 | 38.04 | 60.19 | 41.74 | \n| c | 0.33 | 0.23 | 0.39 | 0.33 | 0.23 | 0.39 |'

    opened by leesb7426 0
Owner
null
A Pytorch Implementation of Source Data-free Domain Adaptation for a Faster R-CNN

A Pytorch Implementation of Source Data-free Domain Adaptation for a Faster R-CNN Please follow Faster R-CNN and DAF to complete the environment confi

null 2 Jan 12, 2022
The official implementation of the CVPR2021 paper: Decoupled Dynamic Filter Networks

Decoupled Dynamic Filter Networks This repo is the official implementation of CVPR2021 paper: "Decoupled Dynamic Filter Networks". Introduction DDF is

F.S.Fire 180 Dec 30, 2022
An implementation for the loss function proposed in Decoupled Contrastive Loss paper.

Decoupled-Contrastive-Learning This repository is an implementation for the loss function proposed in Decoupled Contrastive Loss paper. Requirements P

Ramin Nakhli 71 Dec 4, 2022
Code for our method RePRI for Few-Shot Segmentation. Paper at http://arxiv.org/abs/2012.06166

Region Proportion Regularized Inference (RePRI) for Few-Shot Segmentation In this repo, we provide the code for our paper : "Few-Shot Segmentation Wit

Malik Boudiaf 138 Dec 12, 2022
git《FSCE: Few-Shot Object Detection via Contrastive Proposal Encoding》(CVPR 2021) GitHub: [fig8]

FSCE: Few-Shot Object Detection via Contrastive Proposal Encoding (CVPR 2021) This repo contains the implementation of our state-of-the-art fewshot ob

null 233 Dec 29, 2022
Novel Instances Mining with Pseudo-Margin Evaluation for Few-Shot Object Detection

Novel Instances Mining with Pseudo-Margin Evaluation for Few-Shot Object Detection (NimPme) The official implementation of Novel Instances Mining with

null 12 Sep 8, 2022
Few-NERD: Not Only a Few-shot NER Dataset

Few-NERD: Not Only a Few-shot NER Dataset This is the source code of the ACL-IJCNLP 2021 paper: Few-NERD: A Few-shot Named Entity Recognition Dataset.

THUNLP 319 Dec 30, 2022
Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"

T-Few This repository contains the official code for the paper: "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learni

null 220 Dec 31, 2022
Code for paper Decoupled Dynamic Spatial-Temporal Graph Neural Network for Traffic Forecasting

Decoupled Spatial-Temporal Graph Neural Networks Code for our paper: Decoupled Dynamic Spatial-Temporal Graph Neural Network for Traffic Forecasting.

S22 43 Jan 4, 2023
Official code release for "Learned Spatial Representations for Few-shot Talking-Head Synthesis" ICCV 2021

Official code release for "Learned Spatial Representations for Few-shot Talking-Head Synthesis" ICCV 2021

Moustafa Meshry 16 Oct 5, 2022
Implement Decoupled Neural Interfaces using Synthetic Gradients in Pytorch

disclaimer: this code is modified from pytorch-tutorial Image classification with synthetic gradient in Pytorch I implement the Decoupled Neural Inter

Andrew 114 Dec 22, 2022
Pytorch Implementation for CVPR2018 Paper: Learning to Compare: Relation Network for Few-Shot Learning

LearningToCompare Pytorch Implementation for Paper: Learning to Compare: Relation Network for Few-Shot Learning Howto download mini-imagenet and make

Jackie Loong 246 Dec 19, 2022
Pytorch implementation of the paper "Optimization as a Model for Few-Shot Learning"

Optimization as a Model for Few-Shot Learning This repo provides a Pytorch implementation for the Optimization as a Model for Few-Shot Learning paper.

Albert Berenguel Centeno 238 Jan 4, 2023
Official implementation of our CVPR2021 paper "OTA: Optimal Transport Assignment for Object Detection" in Pytorch.

OTA: Optimal Transport Assignment for Object Detection This project provides an implementation for our CVPR2021 paper "OTA: Optimal Transport Assignme

null 217 Jan 3, 2023
NFT-Price-Prediction-CNN - Using visual feature extraction, prices of NFTs are predicted via CNN (Alexnet and Resnet) architectures.

NFT-Price-Prediction-CNN - Using visual feature extraction, prices of NFTs are predicted via CNN (Alexnet and Resnet) architectures.

null 5 Nov 3, 2022
The implementation of PEMP in paper "Prior-Enhanced Few-Shot Segmentation with Meta-Prototypes"

Prior-Enhanced network with Meta-Prototypes (PEMP) This is the PyTorch implementation of PEMP. Overview of PEMP Meta-Prototypes & Adaptive Prototypes

Jianwei ZHANG 8 Oct 14, 2021
Implementation of the paper "Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning"

Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning This is the implementation of the paper "Self-Promoted Prototype Refinement

Kai Zhu 78 Dec 2, 2022
The official implementation of the CVPR 2021 paper FAPIS: a Few-shot Anchor-free Part-based Instance Segmenter

FAPIS The official implementation of the CVPR 2021 paper FAPIS: a Few-shot Anchor-free Part-based Instance Segmenter Introduction This repo is primari

Khoi Nguyen 8 Dec 11, 2022
Official PyTorch implementation of MX-Font (Multiple Heads are Better than One: Few-shot Font Generation with Multiple Localized Experts)

Introduction Pytorch implementation of Multiple Heads are Better than One: Few-shot Font Generation with Multiple Localized Expert. | paper Song Park1

Clova AI Research 97 Dec 23, 2022