Official implementation of the ICCV 2021 paper: "The Power of Points for Modeling Humans in Clothing".

Related tags

Deep Learning POP
Overview

The Power of Points for Modeling Humans in Clothing (ICCV 2021)

This repository contains the official PyTorch implementation of the ICCV 2021 paper:

The Power of Points for Modeling Humans in Clothing
Qianli Ma, Jinlong Yang, Siyu Tang, and Michael J. Black
Paper | Supp | Video | Project website

Code

Coming soon...

Citations

@inproceedings{Ma:CVPR:2021,
  title = {{SCALE}: Modeling Clothed Humans with a Surface Codec of Articulated Local Elements},
  author = {Ma, Qianli and Saito, Shunsuke and Yang, Jinlong and Tang, Siyu and Black, Michael J.},
  booktitle = {Proceedings IEEE/CVF Conf.~on Computer Vision and Pattern Recognition (CVPR)},
  month = jun,
  year = {2021},
  month_numeric = {6}
}
Comments
  • UV Coordinate

    UV Coordinate

    https://github.com/qianlim/POP/blob/c7e045e6a701de9581ffd7cfab67475f4eb979f0/lib/utils_io.py#L35 Hi, I found that the above code seems to generate transposed UV coordinates, i.e. the UV coordinates of the first row of the image ([height, width, channel] format) grid are [0, 0], [0, 1], [0, 2], ..., [ 0, H-1].

    opened by caiyongqi 4
  • Data preprocess

    Data preprocess

    Hi,

    Thanks for sharing your great work! I notice that the scale and translation of "scan_pc" and "body_verts" in the provided packed data are different from meshes in the original CAPE dataset. What is the scale and translation parameters you used to normalize the meshes?

    Thanks!

    opened by zhaofang0627 4
  • python main.py --config configs/config_demo.yaml --mode test

    python main.py --config configs/config_demo.yaml --mode test

    (POP) siduo@siduo-NK50S5-SZ:~/下载/POP-main$ python main.py --config configs/config_demo.yaml --mode test

    ------------------------Loading checkpoint POP_pretrained_ReSynthdata_12outfits_epoch00400_model.pt

    ------------------------Test model with checkpoint at epoch 400

    ------------------------Eval on test data, seen outfits, unseen poses... Loading test data... Loading rp_alexandra_posed_006, 0 examples.. 0it [00:00, ?it/s] Data loaded, in total 0 test examples.

    Evaluating... 0it [00:00, ?it/s] Traceback (most recent call last): File "main.py", line 354, in main() File "main.py", line 256, in main test_stats = test_seen_clo( File "/home/siduo/下载/POP-main/lib/infer.py", line 316, in test_seen_clo test_m2s /= n_test_samples ZeroDivisionError: division by zero

    opened by QUMINGZHEMENAN 2
  • The

    The "geom_featmap" seems to be reset during training

    For evaluation during training, the value of the "geom_featmap" seems to be reset to the most recently saved value. https://github.com/qianlim/POP/blob/a05404dd44173809abfee4e2ebef57bbc524cfde/main.py#L193-L195 The "geom_featmap" is saved every 50 epochs, while the default value of the "args.val_every" is set to 20. Therefore, the value of the "geom_featmap" will be reset several times. For example, at the 20th or 40th epoch, the "geom_featmap" will be set to the initial value of the 0th epoch. Is this a bug?

    opened by wenbin-lin 2
  • chamfer distance

    chamfer distance

    Hi. You don't seem to be using the chamfer distance directly as the loss, would the performance be better? https://github.com/qianlim/POP/blob/a05404dd44173809abfee4e2ebef57bbc524cfde/lib/train.py#L87 https://github.com/qianlim/POP/blob/a05404dd44173809abfee4e2ebef57bbc524cfde/lib/train.py#L98

    opened by caiyongqi 2
  • Point rendering with surfels

    Point rendering with surfels

    Hi,

    In Fig 4. of the paper, you mentioned that the point clouds are rendered with surfels. Could you kindly share the code for the point rendering with surfels?

    Thanks!

    opened by zhaofang0627 2
  • Mesh for  ReSynth dataset

    Mesh for ReSynth dataset

    Hi, thanks for sharing this great work, I'm wondering how can i get the original meshes from which the scan point cloud is generated? Are they included somewhere in the dataset we downloaded or they are not provided? Thanks!!

    opened by ray8828 2
  • per-point transformation matrix Ti

    per-point transformation matrix Ti

    Hi. I am confused about the transformation matrix T. The ri in Eq(2) is the displacement of the point on posed human body without outfits, why not add pi directly? That is: xi=ri+pi. Could you explain the T and how to calculate it?

    opened by caiyongqi 1
  • About the ReSynth Dataset

    About the ReSynth Dataset

    Hello, it seems in the ReSynth Dataset download page, the below texts are without hyperlinks, so I cannot download demo data.

    Download links:

    Data_for_POP_demo: packed npz files | high-res point clouds

    opened by fwbx529 1
  • Can I open source my own code that uses part of POP's code?

    Can I open source my own code that uses part of POP's code?

    Hi, thank you for sharing the code. I noticed a "no distribution" term in your license. I wonder whether I can open source my own code which is partly based on POP (for noncommercial research purpose only, of course).

    opened by trisct 4
  • About Resynth Dataset

    About Resynth Dataset

    Hi, great work for human modeling, and I wonder to know what's mean of vtransf, I just found that: "a numpy array of shape (N_vert, 3, 3): the local-to-global transformation matrix of each vertex on the posed SMPL-X body." what's local-to-global mean and how to get it?

    opened by llpspark 0
  • Dimension mismatch of trained weights (ReSynthdata_12outfit)

    Dimension mismatch of trained weights (ReSynthdata_12outfit)

    Thanks for sharing the code of the wonderful work!

    I had tried to test the code with pre-trained weight for ReSynthedata provided in the link in Readme. With the default configuration setting, there is a dimension mismatch between the pre-trained weight and the defined network. Specifically, for unet_posefeat defined by this module, UnetNoCond7DS the layer upconvC5 should have input channel size as 256. However, in the pre-trained model, its dimension is 384 which yields dimension mismatch error during loading that model like below.

    image

    Is there any way to fix this?

    opened by SuwoongHeo 3
  • 运行报错

    运行报错

    python main.py --config configs/config_demo.yaml --mode test ,出现

    RPly: Unable to open file Read PLY failed: unable to open file: /mnt/data2/cv610/anaconda3/envs/SCALEcode/POP/lib/../data/resynth/scans/rp_beatrice_posed_005/96_longshort_flying_eagle/00020_

    ------Step 1: Optimizing w.r.t. UNSEEN scan with full_gt

    Traceback (most recent call last): File "main.py", line 345, in main() File "main.py", line 332, in main **model_config File "/mnt/data2/cv610/anaconda3/envs/SCALEcode/POP/lib/infer.py", line 414, in test_unseen_clo num_unseen_frames=num_unseen_frames File "/mnt/data2/cv610/anaconda3/envs/SCALEcode/POP/lib/infer.py", line 187, in reconstruct random_subsample_scan=random_subsample_scan, num_unseen_frames=num_unseen_frames) File "/mnt/data2/cv610/anaconda3/envs/SCALEcode/POP/lib/infer.py", line 103, in model_forward_and_loss lnormal, closest_target_normals = normal_loss(pred_normals, target_pc_n, idx_closest_gt) File "/mnt/data2/cv610/anaconda3/envs/SCALEcode/POP/lib/losses.py", line 21, in normal_loss target_normals_chosen = torch.gather(target_normals, dim=1, index=nearest_idx) RuntimeError: setStorage: sizes [1, 47911, 3], strides [0, 0, 1], storage offset 0, and itemsize 4 requiring a storage size of 12 are out of bounds for storage of s

    opened by HDYYZDN 0
  • question: how to run on custom data?

    question: how to run on custom data?

    Hi, how to inference on custom data? like a single static scan. Need it be fitted to a spml-x model? If I want to animate this scan, how to get a pose sequence like the "*.npz" file in you demo data? Can you please give the instruction like https://github.com/qianlim/SCALE/issues/6#issuecomment-888125275 ? Again, thanks for your excellent work!

    opened by geniuspatrick 9
Official implementation of the ICCV 2021 paper "Conditional DETR for Fast Training Convergence".

The DETR approach applies the transformer encoder and decoder architecture to object detection and achieves promising performance. In this paper, we handle the critical issue, slow training convergence, and present a conditional cross-attention mechanism for fast DETR training. Our approach is motivated by that the cross-attention in DETR relies highly on the content embeddings and that the spatial embeddings make minor contributions, increasing the need for high-quality content embeddings and thus increasing the training difficulty.

null 281 Dec 30, 2022
The Official Implementation of the ICCV-2021 Paper: Semantically Coherent Out-of-Distribution Detection.

SCOOD-UDG (ICCV 2021) This repository is the official implementation of the paper: Semantically Coherent Out-of-Distribution Detection Jingkang Yang,

Jake YANG 62 Nov 21, 2022
official Pytorch implementation of ICCV 2021 paper FuseFormer: Fusing Fine-Grained Information in Transformers for Video Inpainting.

FuseFormer: Fusing Fine-Grained Information in Transformers for Video Inpainting By Rui Liu, Hanming Deng, Yangyi Huang, Xiaoyu Shi, Lewei Lu, Wenxiu

null 77 Dec 27, 2022
Official implementation of the ICCV 2021 paper "Joint Inductive and Transductive Learning for Video Object Segmentation"

JOINT This is the official implementation of Joint Inductive and Transductive learning for Video Object Segmentation, to appear in ICCV 2021. @inproce

Yunyao 35 Oct 16, 2022
Official PyTorch Implementation of paper "Deep 3D Mask Volume for View Synthesis of Dynamic Scenes", ICCV 2021.

Deep 3D Mask Volume for View Synthesis of Dynamic Scenes Official PyTorch Implementation of paper "Deep 3D Mask Volume for View Synthesis of Dynamic S

Ken Lin 17 Oct 12, 2022
This is the official pytorch implementation for our ICCV 2021 paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering" on VQA Task

?? ERASOR (RA-L'21 with ICRA Option) Official page of "ERASOR: Egocentric Ratio of Pseudo Occupancy-based Dynamic Object Removal for Static 3D Point C

Hyungtae Lim 225 Dec 29, 2022
Official Pytorch implementation of the paper "Action-Conditioned 3D Human Motion Synthesis with Transformer VAE", ICCV 2021

ACTOR Official Pytorch implementation of the paper "Action-Conditioned 3D Human Motion Synthesis with Transformer VAE", ICCV 2021. Please visit our we

Mathis Petrovich 248 Dec 23, 2022
Official Repository for the ICCV 2021 paper "PixelSynth: Generating a 3D-Consistent Experience from a Single Image"

PixelSynth: Generating a 3D-Consistent Experience from a Single Image (ICCV 2021) Chris Rockwell, David F. Fouhey, and Justin Johnson [Project Website

Chris Rockwell 95 Nov 22, 2022
Official code release for ICCV 2021 paper SNARF: Differentiable Forward Skinning for Animating Non-rigid Neural Implicit Shapes.

Official code release for ICCV 2021 paper SNARF: Differentiable Forward Skinning for Animating Non-rigid Neural Implicit Shapes.

null 235 Dec 26, 2022
Official implementation of NPMs: Neural Parametric Models for 3D Deformable Shapes - ICCV 2021

NPMs: Neural Parametric Models Project Page | Paper | ArXiv | Video NPMs: Neural Parametric Models for 3D Deformable Shapes Pablo Palafox, Aljaz Bozic

PabloPalafox 109 Nov 22, 2022
An official implementation of "Exploiting a Joint Embedding Space for Generalized Zero-Shot Semantic Segmentation" (ICCV 2021) in PyTorch.

Exploiting a Joint Embedding Space for Generalized Zero-Shot Semantic Segmentation This is an official implementation of the paper "Exploiting a Joint

CV Lab @ Yonsei University 35 Oct 26, 2022
[ICCV 2021] Official Pytorch implementation for Discriminative Region-based Multi-Label Zero-Shot Learning SOTA results on NUS-WIDE and OpenImages

Discriminative Region-based Multi-Label Zero-Shot Learning (ICCV 2021) [arXiv][Project page >> coming soon] Sanath Narayan*, Akshita Gupta*, Salman Kh

Akshita Gupta 54 Nov 21, 2022
[ICCV 2021] Official Pytorch implementation for Discriminative Region-based Multi-Label Zero-Shot Learning SOTA results on NUS-WIDE and OpenImages

Discriminative Region-based Multi-Label Zero-Shot Learning (ICCV 2021) [arXiv][Project page >> coming soon] Sanath Narayan*, Akshita Gupta*, Salman Kh

Akshita Gupta 54 Nov 21, 2022
[ICCV 2021] Official PyTorch implementation for Deep Relational Metric Learning.

Deep Relational Metric Learning This repository is the official PyTorch implementation of Deep Relational Metric Learning. Framework Datasets CUB-200-

Borui Zhang 39 Dec 10, 2022
[ICCV 2021] Official Tensorflow Implementation for "Single Image Defocus Deblurring Using Kernel-Sharing Parallel Atrous Convolutions"

KPAC: Kernel-Sharing Parallel Atrous Convolutional block This repository contains the official Tensorflow implementation of the following paper: Singl

Hyeongseok Son 50 Dec 29, 2022
Official pytorch implementation of "Scaling-up Disentanglement for Image Translation", ICCV 2021.

Official pytorch implementation of "Scaling-up Disentanglement for Image Translation", ICCV 2021.

Aviv Gabbay 41 Nov 29, 2022
Official PyTorch implementation of N-ImageNet: Towards Robust, Fine-Grained Object Recognition with Event Cameras (ICCV 2021)

N-ImageNet: Towards Robust, Fine-Grained Object Recognition with Event Cameras Official PyTorch implementation of N-ImageNet: Towards Robust, Fine-Gra

null 32 Dec 26, 2022
Official implementation of Protected Attribute Suppression System, ICCV 2021

Official implementation of Protected Attribute Suppression System, ICCV 2021

Prithviraj Dhar 6 Jan 1, 2023
This Repo is the official CUDA implementation of ICCV 2019 Oral paper for CARAFE: Content-Aware ReAssembly of FEatures

Introduction This Repo is the official CUDA implementation of ICCV 2019 Oral paper for CARAFE: Content-Aware ReAssembly of FEatures. @inproceedings{Wa

Jiaqi Wang 42 Jan 7, 2023