Dynamic Head: Unifying Object Detection Heads with Attentions

Overview

Dynamic Head: Unifying Object Detection Heads with Attentions

PWC PWC

dyhead_video.mp4

This is the official implementation of CVPR 2021 paper "Dynamic Head: Unifying Object Detection Heads with Attentions".

"In this paper, we present a novel dynamic head framework to unify object detection heads with attentions. By coherently combining multiple self-attention mechanisms between feature levels for scale-awareness, among spatial locations for spatial-awareness, and within output channels for task-awareness, the proposed approach significantly improves the representation ability of object detection heads without any computational overhead."

Dynamic Head: Unifying Object Detection Heads With Attentions

Xiyang Dai, Yinpeng Chen, Bin Xiao, Dongdong Chen, Mengchen Liu, Lu Yuan, Lei Zhang

Model Zoo

Code and Model are under internal review and will release soon. Stay tuned!

In order to open-source, we have ported the implementation from our internal framework to Detectron2 and re-train the models.

We notice better performances on some models compared to original paper.

Config Model Backbone Scheduler COCO mAP Weight
cfg FasterRCNN + DyHead R50 1x 40.3 weight
cfg RetinaNet + DyHead R50 1x 39.9 weight
cfg ATSS + DyHead R50 1x 42.4 weight
cfg ATSS + DyHead Swin-Tiny 2x + ms 49.8 weight

Usage

Dependencies:

Detectron2, timm

Installation:

python -m pip install -e DynamicHead

Train:

To train a config on a single node with 8 gpus, simply use:

DETECTRON2_DATASETS=$DATASET python train_net.py --config configs/dyhead_r50_retina_fpn_1x.yaml --num-gpus 8

Test:

To test a config with a weight on a single node with 8 gpus, simply use:

DETECTRON2_DATASETS=$DATASET python train_net.py --config configs/dyhead_r50_retina_fpn_1x.yaml --num-gpus 8 --eval-only MODEL.WEIGHTS $WEIGHT

Citation

@InProceedings{Dai_2021_CVPR,
    author    = {Dai, Xiyang and Chen, Yinpeng and Xiao, Bin and Chen, Dongdong and Liu, Mengchen and Yuan, Lu and Zhang, Lei},
    title     = {Dynamic Head: Unifying Object Detection Heads With Attentions},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2021},
    pages     = {7373-7382}
}

Contributing

This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.

When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.

This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact [email protected] with any additional questions or comments.

Comments
  • Model train execution fails with a PyTorch error message.

    Model train execution fails with a PyTorch error message.

    Hello, I would like to ask if anyone is facing the below issue: TypeError: add(): argument 'alpha' must be Number, not NoneType

    The steps I followed are:

    1. python -m pip install -e DynamicHead.
    2. Added custom dataset using register_coco_instance
    3. Update config in def setup(args) with custom dataset name and the number of classes: cfg.DATASETS.TRAIN = ('coco_docbank_train',) cfg.MODEL.ROI_HEADS.NUM_CLASSES = 13
    4. Run the model: python train_net.py --config configs/dyhead_r50_rcnn_fpn_1x.yaml --num-gpus 1.

    The detailed error message goes like this:

    File "train_net.py", line 204, in <module> 
        launch(
      File "/opt/conda/lib/python3.8/site-packages/detectron2/engine/launch.py", line 82, in launch
        main_func(*args)
      File "train_net.py", line 198, in main
        return trainer.train()
      File "/opt/conda/lib/python3.8/site-packages/detectron2/engine/defaults.py", line 484, in train
        super().train(self.start_iter, self.max_iter)
      File "/opt/conda/lib/python3.8/site-packages/detectron2/engine/train_loop.py", line 149, in train
        self.run_step()
      File "/opt/conda/lib/python3.8/site-packages/detectron2/engine/defaults.py", line 494, in run_step
        self._trainer.run_step()
      File "/opt/conda/lib/python3.8/site-packages/detectron2/engine/train_loop.py", line 294, in run_step
        self.optimizer.step()
      File "/opt/conda/lib/python3.8/site-packages/torch/optim/lr_scheduler.py", line 65, in wrapper
        return wrapped(*args, **kwargs)
      File "/opt/conda/lib/python3.8/site-packages/torch/optim/optimizer.py", line 88, in wrapper
        return func(*args, **kwargs)
      File "/opt/conda/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 28, in decorate_context
        return func(*args, **kwargs)
      File "/opt/conda/lib/python3.8/site-packages/torch/optim/sgd.py", line 136, in step
        F.sgd(params_with_grad,
      File "/opt/conda/lib/python3.8/site-packages/torch/optim/_functional.py", line 164, in sgd
        d_p = d_p.add(param, alpha=weight_decay)
    TypeError: add(): argument 'alpha' must be Number, not NoneType
    

    Environment Details: sys.platform = linux Python = 3.8.12 numpy = 1.21.2 detectron2 = 0.6 CUDA = 11.4 PyTorch = 1.10.0 torchvision = 0.11.0a0 fvcore = 0.1.5.post20211023 iopath = 0.1.9 cv2 = 4.5.4

    Kindly request for assistance if you are aware of the solution.

    opened by shivamsnaik 10
  • Failed install with `python -m pip install -e DynamicHead`

    Failed install with `python -m pip install -e DynamicHead`

    Installing collected packages: dyhead
      Running setup.py develop for dyhead
        ERROR: Command errored out with exit status 1:
         command: /public/home/hpc2/anaconda3/envs/mask/bin/python -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/public/home/hpc2/rpc/DynamicHead/setup.py'"'"'; __file__='"'"'/public/home/hpc2/rpc/DynamicHead/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' develop --no-deps
             cwd: /public/home/hpc2/rpc/DynamicHead/
        Complete output (71 lines):
        which: no nvcc in (/public/home/hpc2/anaconda3/envs/mask/bin:/public/home/hpc2/anaconda3/condabin:/usr/lib64/qt-3.3/bin:/public/home/hpc2/perl5/bin:/public/software/Modules/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/ibutils/bin:/public/home/hpc2/.local/bin:/public/home/hpc204712181/bin)
        running develop
        running egg_info
        writing dyhead.egg-info/PKG-INFO
        writing dependency_links to dyhead.egg-info/dependency_links.txt
        writing top-level names to dyhead.egg-info/top_level.txt
        reading manifest file 'dyhead.egg-info/SOURCES.txt'
        writing manifest file 'dyhead.egg-info/SOURCES.txt'
        running build_ext
        building 'dyhead._C' extension
        gcc -pthread -B /public/home/hpc2/anaconda3/envs/mask/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -I/public/home/hpc2/rpc/DynamicHead/dyhead/csrc -I/public/home/hpc2/anaconda3/envs/mask/lib/python3.7/site-packages/torch/include -I/public/home/hpc2/anaconda3/envs/mask/lib/python3.7/site-packages/torch/include/torch/csrc/api/include -I/public/home/hpc2/anaconda3/envs/mask/lib/python3.7/site-packages/torch/include/TH -I/public/home/hpc2/anaconda3/envs/mask/lib/python3.7/site-packages/torch/include/THC -I/public/home/hpc2/anaconda3/envs/mask/include/python3.7m -c /public/home/hpc2/rpc/DynamicHead/dyhead/csrc/vision.cpp -o build/temp.linux-x86_64-3.7/public/home/hpc204712181/rpc/DynamicHead/dyhead/csrc/vision.o -DTORCH_API_INCLUDE_EXTENSION_H -DTORCH_EXTENSION_NAME=_C -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++11
        cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
        In file included from /public/home/hpc204712181/rpc/DynamicHead/dyhead/csrc/vision.cpp:2:0:
        /public/home/hpc2/rpc/DynamicHead/dyhead/csrc/deform_conv.h:10:5: error: ‘at’ has not been declared
             at::Tensor input,
             ^
        /public/home/hpc2/rpc/DynamicHead/dyhead/csrc/deform_conv.h:11:5: error: ‘at’ has not been declared
             at::Tensor weight,
             ^
    

    error: command 'gcc' failed with exit status 1 Error seems that at::Tensor can't be recognized, how to solve this problem? My gcc version is 5.4.0

    opened by XiaoyuWant 3
  • cuda 11.1 fail?

    cuda 11.1 fail?

    1、nvcc -V nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2020 NVIDIA Corporation Built on Tue_Sep_15_19:10:02_PDT_2020 Cuda compilation tools, release 11.1, V11.1.74 Build cuda_11.1.TC455_06.29069683_0

    2、pip list Package Version Location


    detectron2 0.6+cu111 torch 1.10.1+cu111 torchaudio 0.10.1+rocm4.1 torchvision 0.11.2+cu111 tqdm 4.63.1

    3、log: sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/structures/boxes.py:158: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at ../torch/csrc/utils/tensor_new.cpp:201.) tensor = torch.as_tensor(tensor, dtype=torch.float32, device=device) error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device /sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/functional.py:3847: UserWarning: nn.functional.upsample_bilinear is deprecated. Use nn.functional.interpolate instead. warnings.warn("nn.functional.upsample_bilinear is deprecated. Use nn.functional.interpolate instead.") error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device

    error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device error in modulated_deformable_im2col_cuda: no kernel image is available for execution on the device THCudaCheck FAIL file=/sda/okay/ai/project/ocrquesseg-svr/dyhead/csrc/cuda/SigmoidFocalLoss_cuda.cu line=139 error=209 : no kernel image is available for execution on the device ERROR [04/01 18:08:11 d2.engine.train_loop]: Exception during training: Traceback (most recent call last): File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/train_loop.py", line 149, in train self.run_step() File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/defaults.py", line 499, in run_step self._trainer.run_step() File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/train_loop.py", line 273, in run_step loss_dict = self.model(data) File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 204, in forward losses = self.losses(anchors, pred_logits, gt_labels, pred_anchor_deltas, gt_boxes, pred_centers) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 236, in losses cls_loss = self.classification_loss_func(box_cls_flatten, labels_flatten.int()) / num_pos_avg_per_gpu File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 48, in forward loss = sigmoid_focal_loss_cuda(logits, targets, self.gamma, self.alpha) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 20, in forward logits, targets, num_classes, gamma, alpha RuntimeError: cuda runtime error (209) : no kernel image is available for execution on the device at /sda/okay/ai/project/ocrquesseg-svr/dyhead/csrc/cuda/SigmoidFocalLoss_cuda.cu:139 [04/01 18:08:11 d2.engine.hooks]: Total training time: 0:00:00 (0:00:00 on hooks) [04/01 18:08:11 d2.utils.events]: iter: 0 lr: N/A max_mem: 6574M Traceback (most recent call last): File "train_net.py", line 224, in args=(args,), File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/launch.py", line 82, in launch main_func(*args) File "train_net.py", line 210, in main return trainer.train() File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/defaults.py", line 489, in train super().train(self.start_iter, self.max_iter) File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/train_loop.py", line 149, in train self.run_step() File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/defaults.py", line 499, in run_step self._trainer.run_step() File "/sda/okay/ai/project/ocrquesseg-svr/detectron2/detectron2/engine/train_loop.py", line 273, in run_step loss_dict = self.model(data) File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 204, in forward losses = self.losses(anchors, pred_logits, gt_labels, pred_anchor_deltas, gt_boxes, pred_centers) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 236, in losses cls_loss = self.classification_loss_func(box_cls_flatten, labels_flatten.int()) / num_pos_avg_per_gpu File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 48, in forward loss = sigmoid_focal_loss_cuda(logits, targets, self.gamma, self.alpha) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 20, in forward logits, targets, num_classes, gamma, alpha RuntimeError: cuda runtime error (209) : no kernel image is available for execution on the device at /sda/okay/ai/project/ocrquesseg-svr/dyhead/csrc/cuda/SigmoidFocalLoss_cuda.cu:139

    opened by vison20080808 2
  • Defining Spatial-aware Attention Layer error

    Defining Spatial-aware Attention Layer error

    when running this line of code spatial_output = spatial_layer(scale_output)

    I got this error ~\AppData\Local\Continuum\anaconda3\lib\site-packages\torch\nn\modules\module.py in _call_impl(self, *input, **kwargs) 725 result = self._slow_forward(*input, **kwargs) 726 else: --> 727 result = self.forward(*input, **kwargs) 728 for hook in itertools.chain( 729 _global_forward_hooks.values(),

    TypeError: forward() takes 3 positional arguments but 4 were given

    opened by Aliweka2020 2
  • This repo is missing important files

    This repo is missing important files

    There are important files that Microsoft projects should all have that are not present in this repository. A pull request has been opened to add the missing file(s). When the pr is merged this issue will be closed automatically.

    Microsoft teams can learn more about this effort and share feedback within the open source guidance available internally.

    Merge this pull request

    opened by microsoft-github-policy-service[bot] 1
  • TypeError: add(): argument 'alpha' must be Number, not NoneType

    TypeError: add(): argument 'alpha' must be Number, not NoneType

    I was trying to run the repo DynamicHead and trying to run the model with a custom dataset and have the following error. In the repo the launch function is called and I get these errors.

    The command I run was

    DETECTRON2_DATASETS=$DATASET python train_net.py --config configs/dyhead_r50_retina_fpn_1x.yaml --num-gpus 8
    

    Logs

    Traceback (most recent call last):
      File "train_net_custom.py", line 222, in <module>
        launch(
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/detectron2/engine/launch.py", line 82, in launch
        main_func(*args)
      File "train_net_custom.py", line 216, in main
        return trainer.train()
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/detectron2/engine/defaults.py", line 484, in train
        super().train(self.start_iter, self.max_iter)
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/detectron2/engine/train_loop.py", line 149, in train
        self.run_step()
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/detectron2/engine/defaults.py", line 494, in run_step
        self._trainer.run_step()
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/detectron2/engine/train_loop.py", line 294, in run_step
        self.optimizer.step()
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/torch/optim/lr_scheduler.py", line 65, in wrapper
        return wrapped(*args, **kwargs)
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/torch/optim/optimizer.py", line 89, in wrapper
        return func(*args, **kwargs)
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context
        return func(*args, **kwargs)
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/torch/optim/sgd.py", line 110, in step
        F.sgd(params_with_grad,
      File "/home/xxx/anaconda3/envs/detectron/lib/python3.8/site-packages/torch/optim/_functional.py", line 160, in sgd
        d_p = d_p.add(param, alpha=weight_decay)
    TypeError: add(): argument 'alpha' must be Number, not NoneType
    

    Detectron2 version: 0.6

    Expected behaviour is the training loop runs without error.

    opened by reubenwenisch 1
  • Adding Microsoft SECURITY.MD

    Adding Microsoft SECURITY.MD

    Please accept this contribution adding the standard Microsoft SECURITY.MD :lock: file to help the community understand the security policy and how to safely report security issues. GitHub uses the presence of this file to light-up security reminders and a link to the file. This pull request commits the latest official SECURITY.MD file from https://github.com/microsoft/repo-templates/blob/main/shared/SECURITY.md.

    Microsoft teams can learn more about this effort and share feedback within the open source guidance available internally.

    opened by microsoft-github-policy-service[bot] 0
  • cuda11.1 train fail?

    cuda11.1 train fail?

    Traceback (most recent call last): File "train_net.py", line 224, in args=(args,), File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/detectron2/engine/launch.py", line 82, in launch main_func(*args) File "train_net.py", line 210, in main return trainer.train() File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/detectron2/engine/defaults.py", line 484, in train super().train(self.start_iter, self.max_iter) File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/detectron2/engine/train_loop.py", line 149, in train self.run_step() File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/detectron2/engine/defaults.py", line 494, in run_step self._trainer.run_step() File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/detectron2/engine/train_loop.py", line 273, in run_step loss_dict = self.model(data) File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 204, in forward losses = self.losses(anchors, pred_logits, gt_labels, pred_anchor_deltas, gt_boxes, pred_centers) File "/sda/okay/ai/project/ocrquesseg-svr/extra/atss.py", line 236, in losses cls_loss = self.classification_loss_func(box_cls_flatten, labels_flatten.int()) / num_pos_avg_per_gpu File "/sda/program/anaconda3/envs/detectron2/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl return forward_call(*input, **kwargs) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 48, in forward loss = sigmoid_focal_loss_cuda(logits, targets, self.gamma, self.alpha) File "/sda/okay/ai/project/ocrquesseg-svr/extra/sigmoid_focal_loss.py", line 20, in forward logits, targets, num_classes, gamma, alpha RuntimeError: cuda runtime error (209) : no kernel image is available for execution on the device at /sda/okay/ai/project/ocrquesseg-svr/dyhead/csrc/cuda/SigmoidFocalLoss_cuda.cu:139

    opened by vison20080808 0
  • ModuleNotFoundError: No module named 'dyhead._C'

    ModuleNotFoundError: No module named 'dyhead._C'

    Thanks for sharing your great work!But I encountered this error for a long time and can't solve it. Traceback (most recent call last): File "train_net.py", line 33, in <module> from dyhead import add_dyhead_config File "/user-data/DynamicHead-master/dyhead/__init__.py", line 2, in <module> from .dyhead import DyHead File "/user-data/DynamicHead-master/dyhead/dyhead.py", line 7, in <module> from .deform import ModulatedDeformConv File "/user-data/DynamicHead-master/dyhead/deform.py", line 8, in <module> import dyhead._C as _C ModuleNotFoundError: No module named 'dyhead._C'

    opened by yaofanji 2
  • DynamicHead/dyhead/_C.cpython-38-x86_64-linux-gnu.so: undefined symbol: _ZN6caffe28TypeMeta21_typeMetaDataInstanceIdEEPKNS_6detail12TypeMetaDataEv

    DynamicHead/dyhead/_C.cpython-38-x86_64-linux-gnu.so: undefined symbol: _ZN6caffe28TypeMeta21_typeMetaDataInstanceIdEEPKNS_6detail12TypeMetaDataEv

    Hello, when I run the train_net.py, I encountered this error: ImportError: XX/Code/DynamicHead/dyhead/_C.cpython-38-x86_64-linux-gnu.so: undefined symbol: _ZN6caffe28TypeMeta21_typeMetaDataInstanceIdEEPKNS_6detail12TypeMetaDataEv,

    It caused by "import dyhead._C as _C", I tried change the torch and torchvision version, but it not works. May you find the solution?

    opened by ChenShawn315 0
  • python -m pip install -e DynamicHead

    python -m pip install -e DynamicHead

    Hi, when I execute this sentence, an error occurs with DynamicHead.
    ERROR: DynamicHead is not a valid editable requirement. It should either be a path to a local p roject or a VCS URL (beginning with bzr+http, bzr+https, bzr+ssh, bzr+sftp, bzr+ftp, bzr+lp, bz r+file, git+http, git+https, git+ssh, git+git, git+file, hg+file, hg+http, hg+https, hg+ssh, hg+static-http, svn+ssh, svn+http, svn+https, svn+svn, svn+file).

    opened by sevenandseven 6
Owner
Microsoft
Open source projects and samples from Microsoft
Microsoft
Official Implementation of "LUNAR: Unifying Local Outlier Detection Methods via Graph Neural Networks"

LUNAR Official Implementation of "LUNAR: Unifying Local Outlier Detection Methods via Graph Neural Networks" Adam Goodge, Bryan Hooi, Ng See Kiong and

Adam Goodge 25 Dec 28, 2022
TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-Captured Scenarios

TPH-YOLOv5 This repo is the implementation of "TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-Captured

cv516Buaa 439 Dec 22, 2022
Code for NAACL 2021 full paper "Efficient Attentions for Long Document Summarization"

LongDocSum Code for NAACL 2021 paper "Efficient Attentions for Long Document Summarization" This repository contains data and models needed to reprodu

null 56 Jan 2, 2023
Job-Recommend-Competition - Vectorwise Interpretable Attentions for Multimodal Tabular Data

SiD - Simple Deep Model Vectorwise Interpretable Attentions for Multimodal Tabul

Jungwoo Park 40 Dec 22, 2022
[CVPRW 2022] Attentions Help CNNs See Better: Attention-based Hybrid Image Quality Assessment Network

Attention Helps CNN See Better: Hybrid Image Quality Assessment Network [CVPRW 2022] Code for Hybrid Image Quality Assessment Network [paper] [code] T

IIGROUP 49 Dec 11, 2022
Dynamic vae - Dynamic VAE algorithm is used for anomaly detection of battery data

Dynamic VAE frame Automatic feature extraction can be achieved by probability di

null 10 Oct 7, 2022
Official PyTorch implementation of MX-Font (Multiple Heads are Better than One: Few-shot Font Generation with Multiple Localized Experts)

Introduction Pytorch implementation of Multiple Heads are Better than One: Few-shot Font Generation with Multiple Localized Expert. | paper Song Park1

Clova AI Research 97 Dec 23, 2022
A coin flip game in which you can put the amount of money below or equal to 1000 and then choose heads or tail

COIN_FLIPPY ##This is a simple example package. You can use Github-flavored Markdown to write your content. Coinflippy A coin flip game in which you c

null 2 Dec 26, 2021
Official implementation of the paper ``Unifying Nonlocal Blocks for Neural Networks'' (ICCV'21)

Spectral Nonlocal Block Overview Official implementation of the paper: Unifying Nonlocal Blocks for Neural Networks (ICCV'21) Spectral View of Nonloca

null 91 Dec 14, 2022
[ICCV'21] UNISURF: Unifying Neural Implicit Surfaces and Radiance Fields for Multi-View Reconstruction

UNISURF: Unifying Neural Implicit Surfaces and Radiance Fields for Multi-View Reconstruction Project Page | Paper | Supplementary | Video This reposit

null 331 Dec 28, 2022
Official repository of OFA. Paper: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework

Paper | Blog OFA is a unified multimodal pretrained model that unifies modalities (i.e., cross-modality, vision, language) and tasks (e.g., image gene

OFA Sys 1.4k Jan 8, 2023
TSDF++: A Multi-Object Formulation for Dynamic Object Tracking and Reconstruction

TSDF++: A Multi-Object Formulation for Dynamic Object Tracking and Reconstruction TSDF++ is a novel multi-object TSDF formulation that can encode mult

ETHZ ASL 130 Dec 29, 2022
Web service for facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation based on OpenFace 2.0

OpenGaze: Web Service for OpenFace Facial Behaviour Analysis Toolkit Overview OpenFace is a fantastic tool intended for computer vision and machine le

Sayom Shakib 4 Nov 3, 2022
OpenFace – a state-of-the art tool intended for facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation.

OpenFace 2.2.0: a facial behavior analysis toolkit Over the past few years, there has been an increased interest in automatic facial behavior analysis

Tadas Baltrusaitis 5.8k Dec 31, 2022
Code for BMVC2021 "MOS: A Low Latency and Lightweight Framework for Face Detection, Landmark Localization, and Head Pose Estimation"

MOS-Multi-Task-Face-Detect Introduction This repo is the official implementation of "MOS: A Low Latency and Lightweight Framework for Face Detection,

null 104 Dec 8, 2022
A light and fast one class detection framework for edge devices. We provide face detector, head detector, pedestrian detector, vehicle detector......

A Light and Fast Face Detector for Edge Devices Big News: LFD, which is a big update of LFFD, now is released (2021.03.09). It is strongly recommended

YonghaoHe 1.3k Dec 25, 2022
Hybrid CenterNet - Hybrid-supervised object detection / Weakly semi-supervised object detection

Hybrid-Supervised Object Detection System Object detection system trained by hybrid-supervision/weakly semi-supervision (HSOD/WSSOD): This project is

null 5 Dec 10, 2022
Yolo object detection - Yolo object detection with python

How to run download required files make build_image make download Docker versio

null 3 Jan 26, 2022
Official implementation of "Dynamic Anchor Learning for Arbitrary-Oriented Object Detection" (AAAI2021).

DAL This project hosts the official implementation for our AAAI 2021 paper: Dynamic Anchor Learning for Arbitrary-Oriented Object Detection [arxiv] [c

ming71 215 Nov 28, 2022