MoCoPnet - Deformable 3D Convolution for Video Super-Resolution

Overview

Deformable 3D Convolution for Video Super-Resolution

Pytorch implementation of local motion and contrast prior driven deep network (MoCoPnet). [PDF]

Overview


Requirements

  • Python 3
  • pytorch >= 1.6
  • numpy, PIL

Datasets

Training & test datasets

Download SAITD dataset.

SAITD dataset is a large-scale high-quality semi-synthetic dataset of infrared small target. We employ the 1st-50th sequences with target annotations as the test datasets and the remaining 300 sequences as the training datasets.

Download Hui and Anti-UAV.

Hui and Anti-UAV datasets are used as the test datasets to test the robustness of our MoCoPnet to real scenes. In Anti-UAV dataset, only the sequences with infrared small target (i.e., The target size is less than 0.12% of the image size) are selected as the test set (21 sequences in total). Note that, we only use the first 100 images of each sequence for test to balance computational/time cost and generalization performance.

For simplicity, you can also Download the test datasets in https://pan.baidu.com/s/1oobhklwIChvNJIBpTcdQRQ?pwd=1113 and put the folder in code/data.

Data format:

  1. The training dataset is in code/data/train/SAITD.
train
  └── SAITD
       └── 1
              ├── 0.png
              ├── 1.png
              ├── ...
       └── 2
              ├── 00001
              ├── 00002
              ├── ...		
       ...
  1. The test datasets are in code/data/test as below:
 test
  └── dataset_1
         └── scene_1
              ├── 0.png  
              ├── 1.png  
              ├── ...
              └── 100.png    
               
         ├── ...		  
         └── scene_M
  ├── ...    
  └── dataset_N      

Results

Quantitative Results of SR performance

Table 1. PSNR/SSIM achieved by different methods.

Table 2. SNR and CR results of different methods achieved on super-resolved LR images and super-resolved HR images.

Qualitative Results of SR performance

Figure 1. Visual results of different SR methods on LR images for 4x SR.

Figure 2. Visual results of different SR methods on LR images for 4x SR.

Quantitative Results of detection

Table 3. Quantitative results of Tophat, ILCM, IPI achieved on super-resolved LR images.

Table 4. Quantitative results of Tophat, ILCM, IPI achieved on super-resolved HR images.

Figure 3. ROC results of Tophat, ILCM and IPI achieved on super-resolved LR images.

Figure 4. ROC results of Tophat, ILCM and IPI achieved on super-resolved HR images.

Qualitative Results of detection

Figure 5. Qualitative results of super-resolved LR image and detection results.

Figure 6. Qualitative results of super-resolved HR image and detection results.

Citiation

@article{MoCoPnet,
  author = {Ying, Xinyi and Wang, Yingqian and Wang, Longguang and Sheng, Weidong and Liu, Li and Lin, Zaipin and Zhou, Shilin},
  title = {MoCoPnet: Exploring Local Motion and Contrast Priors for Infrared Small Target Super-Resolution},
  journal={arXiv preprint arXiv:2201.01014},
  year = {2020},
}

Contact

Please contact us at [email protected] for any question.

You might also like...
BasicVSR++: Improving Video Super-Resolution with Enhanced Propagation and Alignment

BasicVSR++: Improving Video Super-Resolution with Enhanced Propagation and Alignment

VSR-Transformer - This paper proposes a new Transformer for video super-resolution (called VSR-Transformer).
VSR-Transformer - This paper proposes a new Transformer for video super-resolution (called VSR-Transformer).

VSR-Transformer By Jiezhang Cao, Yawei Li, Kai Zhang, Luc Van Gool This paper proposes a new Transformer for video super-resolution (called VSR-Transf

EFENet: Reference-based Video Super-Resolution with Enhanced Flow Estimation

EFENet EFENet: Reference-based Video Super-Resolution with Enhanced Flow Estimation Code is a bit messy now. I woud clean up soon. For training the EF

Fast and Context-Aware Framework for Space-Time Video Super-Resolution (VCIP 2021)

Fast and Context-Aware Framework for Space-Time Video Super-Resolution Preparation Dependencies PyTorch 1.2.0 CUDA 10.0 DCNv2 cd model/DCNv2 bash make

Official repository of "BasicVSR++: Improving Video Super-Resolution with Enhanced Propagation and Alignment"

BasicVSR_PlusPlus (CVPR 2022) [Paper] [Project Page] [Code] This is the official repository for BasicVSR++. Please feel free to raise issue related to

Pytorch implementation of our method for high-resolution (e.g. 2048x1024) photorealistic video-to-video translation.
Pytorch implementation of our method for high-resolution (e.g. 2048x1024) photorealistic video-to-video translation.

vid2vid Project | YouTube(short) | YouTube(full) | arXiv | Paper(full) Pytorch implementation for high-resolution (e.g., 2048x1024) photorealistic vid

An implementation of Video Frame Interpolation via Adaptive Separable Convolution using PyTorch
An implementation of Video Frame Interpolation via Adaptive Separable Convolution using PyTorch

This work has now been superseded by: https://github.com/sniklaus/revisiting-sepconv sepconv-slomo This is a reference implementation of Video Frame I

The VarCNN is an Convolution Neural Network based approach to automate Video Assistant Referee in football.
Super-Fast-Adversarial-Training - A PyTorch Implementation code for developing super fast adversarial training

Super-Fast-Adversarial-Training This is a PyTorch Implementation code for develo

Comments
  • Could you release full checkpoints?

    Could you release full checkpoints?

    Hi, thanks again for your excellent works. Is it possible to release the full checkpoints of the model, so we can make a comparison or finetune it? Thanks.

    opened by Monalissaa 3
  • Testing of other evaluation metrics and target detectionm 其他评估指标的测试及目标检测

    Testing of other evaluation metrics and target detectionm 其他评估指标的测试及目标检测

    感谢您的工作!想问一下,后续是否会开源包含SNRG、BSF、ROC等评估指标的测试代码以及目标检测代码,想作为参考。 再次感谢您的工作,给我带来了很多启发。

    Thanks for your work! I would like to ask whether the testing code containing evaluation metrics such as SNRG, BSF, and ROC and target detection code will be open sourced in the future. I would like to use it as a reference. Thanks again for your work which has inspired me a lot.

    opened by szw811 2
  • 训练到大概 9000 次迭代的时候 loss 变得巨大

    训练到大概 9000 次迭代的时候 loss 变得巨大

    再次感谢您的工作! 我按照您的默认设置(SAIDT数据集是按照论文里的训练测试方式进行划分的)进行了训练,如下所示

    parser.add_argument("--save", default='./log', type=str, help="Save path")
    parser.add_argument("--resume", default="", type=str, help="Resume path (default: none)")
    parser.add_argument("--scale_factor", type=int, default=4, help="scale")
    parser.add_argument("--input_num", type=int, default=7, help="input frame number")
    parser.add_argument("--train_dataset_dir", default='./data/train/SAITD', type=str, help="train_dataset")
    parser.add_argument("--val_dataset_dir", default='./data/test/SAITD', type=str, help="train_dataset")
    parser.add_argument("--batch_size", type=int, default=2, help="Training batch size")
    parser.add_argument('--patch_size', type=int, default=64)
    parser.add_argument('--n_iters', type=int, default=100000, help='number of iterations to train')
    parser.add_argument("--device", default=0, type=int, help="GPU id (default: 0)")
    parser.add_argument("--lr", type=float, default=1e-3, help="Learning Rate. Default=4e-4")
    parser.add_argument('--gamma', type=float, default=0.5, help='gamma')
    parser.add_argument("--milestones", type=int, default=[10000,20000,60000], help="Sets the learning rate to the initial LR decayed by momentum every n epochs, Default: n=6")
    parser.add_argument("--threads", type=int, default=4, help="Number of threads for data loader to use, Default: 1")
    

    可是,训练到大概 9000 次迭代的时候 loss 变得巨大 8999it [32:18, 8.27it/s]Mar 29 23:46:06 iter---9000, loss_epoch---17572159662772990246912.000000, PSNR---5.812408

    请问,您知道这是怎么回事嘛

    opened by Monalissaa 2
Owner
Xinyi Ying
Her current research interests focus on image & video super-resolution and small target detection.
Xinyi Ying
PyTorch implementation of Deformable Convolution

Deformable Convolutional Networks in PyTorch This repo is an implementation of Deformable Convolution. Ported from author's MXNet implementation. Buil

null 411 Dec 16, 2022
PyTorch implementation of Deformable Convolution

PyTorch implementation of Deformable Convolution !!!Warning: There is some issues in this implementation and this repo is not maintained any more, ple

Wei Ouyang 893 Dec 18, 2022
Edge-oriented Convolution Block for Real-time Super Resolution on Mobile Devices, ACM Multimedia 2021

Codes for ECBSR Edge-oriented Convolution Block for Real-time Super Resolution on Mobile Devices Xindong Zhang, Hui Zeng, Lei Zhang ACM Multimedia 202

xindong zhang 236 Dec 26, 2022
[CVPR 2022] Official PyTorch Implementation for "Reference-based Video Super-Resolution Using Multi-Camera Video Triplets"

Reference-based Video Super-Resolution (RefVSR) Official PyTorch Implementation of the CVPR 2022 Paper Project | arXiv | RealMCVSR Dataset This repo c

Junyong Lee 151 Dec 30, 2022
3D2Unet: 3D Deformable Unet for Low-Light Video Enhancement (PRCV2021)

3DDUNET This is the code for 3D2Unet: 3D Deformable Unet for Low-Light Video Enhancement (PRCV2021) Conference Paper Link Dataset We use SMOID dataset

null 1 Jan 7, 2022
The official pytorch implemention of the CVPR paper "Temporal Modulation Network for Controllable Space-Time Video Super-Resolution".

This is the official PyTorch implementation of TMNet in the CVPR 2021 paper "Temporal Modulation Network for Controllable Space-Time VideoSuper-Resolu

Gang Xu 95 Oct 24, 2022
A PyTorch Reimplementation of TecoGAN: Temporally Coherent GAN for Video Super-Resolution

TecoGAN-PyTorch Introduction This is a PyTorch reimplementation of TecoGAN: Temporally Coherent GAN for Video Super-Resolution (VSR). Please refer to

null 165 Dec 17, 2022
PyTorch implementation of EGVSR: Efficcient & Generic Video Super-Resolution (VSR)

This is a PyTorch implementation of EGVSR: Efficcient & Generic Video Super-Resolution (VSR), using subpixel convolution to optimize the inference speed of TecoGAN VSR model. Please refer to the official implementation ESPCN and TecoGAN for more information.

null 789 Jan 4, 2023
Exploit Camera Raw Data for Video Super-Resolution via Hidden Markov Model Inference

RawVSR This repo contains the official codes for our paper: Exploit Camera Raw Data for Video Super-Resolution via Hidden Markov Model Inference Xiaoh

Xiaohong Liu 23 Oct 8, 2022
BasicVSR: The Search for Essential Components in Video Super-Resolution and Beyond

BasicVSR BasicVSR: The Search for Essential Components in Video Super-Resolution and Beyond Ported from https://github.com/xinntao/BasicSR Dependencie

Holy Wu 8 Jun 7, 2022