Official code for "Decoupling Zero-Shot Semantic Segmentation"

Overview

Decoupling Zero-Shot Semantic Segmentation

This is the official code for the arxiv.

ZegFormer is the first framework that decouple the zero-shot semantic segmentation into: 1) class-agnostic segmentation and 2) segment-level zero-shot classification

Visualization of semantic segmentation with open vocabularies

ZegFormer is able to segment stuff and things with open vocabularies. The unannotated vocabularies in COCO-Stuff can also be segmented by ZegFormer. visualization

Implementation

Coming soon!

Comments
  • Why novel queries not overfit to the

    Why novel queries not overfit to the "no-object" class?

    Hello authors, thank you for this great work!

    In my understanding, during the training period, instances with novel classes are also exposed to the network, so the queries of these novel instances will match the "no-object" textual embeddings by the cross entropy loss. In this case, the model may tend to ignore the novel instances during inference and cause severe performance drop. But it seems not a problem in the experiements, so I'm wondering if my understanding has something wrong.

    opened by yifliu3 5
  • Request for release of Pascal VOC configs and preprocessing

    Request for release of Pascal VOC configs and preprocessing

    Hello authors, thank you for this great work!

    I was trying to run Zegformer on Pascal VOC, but was facing some problems. I was wondering if you can release the configs and data preprocessing required to reproduce the results on this dataset reported in the paper. Would help a lot!

    Thank you.

    opened by mustafa1728 3
  • Running inference without downloading the datasets

    Running inference without downloading the datasets

    Hi, Thank you for the amazing work, Is there a way to get predictions on a new test image without downloading the whole datsets ? e.g when I run python3 demo/demo.py --config-file configs/coco-stuff/zegformer_R101_bs32_60k_vit16_coco-stuff_gzss_eval.yaml --input figures/dumy.png --output figures/dumdum.png --opts MODEL.WEIGHTS checkpoints/zegformer_R101_bs32_60k_vit16_coco-stuff.pt I get the error message FileNotFoundError: [Errno 2] No such file or directory: 'datasets/coco/coco_stuff/split/seen_classnames.json'

    Can I have these seen_classnames.json files, like the .npy files you have uploaded? Or Can you explain the format of expected .json files so I can generate myself?

    opened by rsadiq 2
  • Is the zegformer trained in an inductive zero shot setting?

    Is the zegformer trained in an inductive zero shot setting?

    Thanks for the good research.

    I have a question about your paper. As far as I know, there are inductive settings and transductive settings for the zero-shot task. In the zero-shot classification task, the inductive setting can be clearly implemented, but in the ZS3 (zero-shot semantic segmentation) task, in the training stage, I know that the part corresponding to the unseen class in the input image is used without masking.

    So I'm confused whether the ZegFormer is inductive or transductive setting. In the case of zegformer, does it belong to inductive setting in ZS3 task?

    opened by Genie-Kim 2
  • Does the direct use of the CLIP model violate the principle of zero-shot learning?

    Does the direct use of the CLIP model violate the principle of zero-shot learning?

    Hello, author, you directly use CLIP model to classify the class-agnostic binary mask during the testing phase. This seems to violate the principle of zero-shot learning, because CLIP already has the information of unseen classes.

    opened by Mamduh-k 2
  • how to split val2017_seen and val2017_unseen?

    how to split val2017_seen and val2017_unseen?

    when i run the command of "python datasets/coco-stuff/prepare_coco_stuff_sem_seg_seen.py", i met the error of "No such file or directory: 'datasets/coco/coco_stuff/annotations/val2017_seen'".

    opened by wly-ai-bj 1
  • Questions about the configuration of build_pixel_decoder?

    Questions about the configuration of build_pixel_decoder?

    Hello author, I didn't find the value of "cfg.MODEL.MASK_FORMER.PIXEL_DECODER_NAME" when checking the configuration file, because it's the first time I read the code related to detectron2, so I don't know it very well.Can you give me some help?

    opened by zkjkak 0
  • Direct evaluate coco-stuff model on ADE-20K

    Direct evaluate coco-stuff model on ADE-20K

    @dingjiansw101 Hi Jian, thanks for your great work! I am wondering did you happen to test your trained coco-stuff model directly on the ADE-20K dataset? Because in the concurrent works, like [1][2], they all report this transfer number. It is very interesting to compare your work with counterparts. Thanks!

    [1] Xu, Mengde, et al. "A simple baseline for zero-shot semantic segmentation with pre-trained vision-language model." arXiv preprint arXiv:2112.14757 (2021). [2] Ghiasi, Golnaz, et al. "Open-vocabulary image segmentation." arXiv preprint arXiv:2112.12143 (2021).

    opened by Jeff-LiangF 3
Owner
Jian Ding
Jian Ding
Official code for Score-Based Generative Modeling through Stochastic Differential Equations

Score-Based Generative Modeling through Stochastic Differential Equations This repo contains the official implementation for the paper Score-Based Gen

Yang Song 818 Jan 6, 2023
Official code for paper "Optimization for Oriented Object Detection via Representation Invariance Loss".

Optimization for Oriented Object Detection via Representation Invariance Loss By Qi Ming, Zhiqiang Zhou, Lingjuan Miao, Xue Yang, and Yunpeng Dong. Th

ming71 56 Nov 28, 2022
This repo provides the official code for TransBTS: Multimodal Brain Tumor Segmentation Using Transformer (https://arxiv.org/pdf/2103.04430.pdf).

TransBTS: Multimodal Brain Tumor Segmentation Using Transformer This repo is the official implementation for TransBTS: Multimodal Brain Tumor Segmenta

Raymond 247 Dec 28, 2022
Official code of the paper "ReDet: A Rotation-equivariant Detector for Aerial Object Detection" (CVPR 2021)

ReDet: A Rotation-equivariant Detector for Aerial Object Detection ReDet: A Rotation-equivariant Detector for Aerial Object Detection (CVPR2021), Jiam

csuhan 334 Dec 23, 2022
Official code implementation for "Personalized Federated Learning using Hypernetworks"

Personalized Federated Learning using Hypernetworks This is an official implementation of Personalized Federated Learning using Hypernetworks paper. [

Aviv Shamsian 121 Dec 25, 2022
Official code for the paper: Deep Graph Matching under Quadratic Constraint (CVPR 2021)

QC-DGM This is the official PyTorch implementation and models for our CVPR 2021 paper: Deep Graph Matching under Quadratic Constraint. It also contain

Quankai Gao 55 Nov 14, 2022
Official code for the ICLR 2021 paper Neural ODE Processes

Neural ODE Processes Official code for the paper Neural ODE Processes (ICLR 2021). Abstract Neural Ordinary Differential Equations (NODEs) use a neura

Cristian Bodnar 50 Oct 28, 2022
Official PyTorch Code of GrooMeD-NMS: Grouped Mathematically Differentiable NMS for Monocular 3D Object Detection (CVPR 2021)

GrooMeD-NMS: Grouped Mathematically Differentiable NMS for Monocular 3D Object Detection GrooMeD-NMS: Grouped Mathematically Differentiable NMS for Mo

Abhinav Kumar 76 Jan 2, 2023
Official code for the CVPR 2021 paper "How Well Do Self-Supervised Models Transfer?"

How Well Do Self-Supervised Models Transfer? This repository hosts the code for the experiments in the CVPR 2021 paper How Well Do Self-Supervised Mod

Linus Ericsson 157 Dec 16, 2022
Official PyTorch code of Holistic 3D Scene Understanding from a Single Image with Implicit Representation (CVPR 2021)

Implicit3DUnderstanding (Im3D) [Project Page] Holistic 3D Scene Understanding from a Single Image with Implicit Representation Cheng Zhang, Zhaopeng C

Cheng Zhang 149 Jan 8, 2023
This is the official code release for the paper Shape and Material Capture at Home

This is the official code release for the paper Shape and Material Capture at Home. The code enables you to reconstruct a 3D mesh and Cook-Torrance BRDF from one or more images captured with a flashlight or camera with flash.

null 89 Dec 10, 2022
Official code of CVPR 2021's PLOP: Learning without Forgetting for Continual Semantic Segmentation

PLOP: Learning without Forgetting for Continual Semantic Segmentation This repository contains all of our code. It is a modified version of Cermelli e

Arthur Douillard 116 Dec 14, 2022
Official code of our work, Unified Pre-training for Program Understanding and Generation [NAACL 2021].

PLBART Code pre-release of our work, Unified Pre-training for Program Understanding and Generation accepted at NAACL 2021. Note. A detailed documentat

Wasi Ahmad 138 Dec 30, 2022
official code for dynamic convolution decomposition

Revisiting Dynamic Convolution via Matrix Decomposition (ICLR 2021) A pytorch implementation of DCD. If you use this code in your research please cons

Yunsheng Li 110 Nov 23, 2022
This repo contains the official code of our work SAM-SLR which won the CVPR 2021 Challenge on Large Scale Signer Independent Isolated Sign Language Recognition.

Skeleton Aware Multi-modal Sign Language Recognition By Songyao Jiang, Bin Sun, Lichen Wang, Yue Bai, Kunpeng Li and Yun Fu. Smile Lab @ Northeastern

Isen (Songyao Jiang) 128 Dec 8, 2022
Official code for "End-to-End Optimization of Scene Layout" -- including VAE, Diff Render, SPADE for colorization (CVPR 2020 Oral)

End-to-End Optimization of Scene Layout Code release for: End-to-End Optimization of Scene Layout CVPR 2020 (Oral) Project site, Bibtex For help conta

Andrew Luo 41 Dec 9, 2022
Official source code to CVPR'20 paper, "When2com: Multi-Agent Perception via Communication Graph Grouping"

When2com: Multi-Agent Perception via Communication Graph Grouping This is the PyTorch implementation of our paper: When2com: Multi-Agent Perception vi

null 34 Nov 9, 2022
Official code repository of the paper Learning Associative Inference Using Fast Weight Memory by Schlag et al.

Learning Associative Inference Using Fast Weight Memory This repository contains the offical code for the paper Learning Associative Inference Using F

Imanol Schlag 18 Oct 12, 2022
Official code for "Mean Shift for Self-Supervised Learning"

MSF Official code for "Mean Shift for Self-Supervised Learning" Requirements Python >= 3.7.6 PyTorch >= 1.4 torchvision >= 0.5.0 faiss-gpu >= 1.6.1 In

UMBC Vision 44 Nov 21, 2022