title | emoji | colorFrom | colorTo | sdk | app_file | pinned |
---|---|---|---|---|---|---|
AnimeGAN-v2 For Videos
|
|
blue
|
red
|
gradio
|
app.py
|
false
|
AnimeGAN-v2 For Videos
Apply AnimeGAN-v2 across frames of a video
Autogenerated using this template
title | emoji | colorFrom | colorTo | sdk | app_file | pinned |
---|---|---|---|---|---|---|
AnimeGAN-v2 For Videos
|
|
blue
|
red
|
gradio
|
app.py
|
false
|
Apply AnimeGAN-v2 across frames of a video
Autogenerated using this template
Graphlevel-SSL Overview Apply Graph Self-Supervised Learning methods to graph-level task(TUDataset, MolculeNet Dataset). It is unified framework to co
MergeNet - Boost Your Own Depth Boost custom or edited monocular depth maps using MergeNet Input Original result After manual editing of base You can
Raster Perspective Apply a perspective transformation to bitmap image using the selected path as envelope, without the need to use an external softwar
music2video Overview A repo for making a music video with Wav2CLIP and VQGAN-CLIP. The base code was derived from VQGAN-CLIP The CLIP embedding for au
SERank An efficient and effective learning to rank algorithm by mining information across ranking candidates. This repository contains the tensorflow
COTR: Correspondence Transformer for Matching Across Images This repository contains the inference code for COTR. We plan to release the training code
Habitat: A Runtime-Based Computational Performance Predictor for Deep Neural Network Training Habitat is a tool that predicts a deep neural network's
Self-training for Few-shot Transfer Across Extreme Task Differences (STARTUP) Introduction This repo contains the official implementation of the follo
Daft-Exprt - PyTorch Implementation PyTorch Implementation of Daft-Exprt: Robust Prosody Transfer Across Speakers for Expressive Speech Synthesis The
Hey,
first of all it's cool to see someone make a video version of this, already fell in love with the original one :)
It already works with a webcam recording but i can't find a way to import a video file. Does that still need to be added or am i doing something wrong? Tried via Spaces with Chrome.
cheers and cool stuff!
edit: sorry i should probably add that i tried it via HuggingFace, not on my local machine
[Open Source]. The improved version of AnimeGAN. Landscape photos/videos to anime
Panoptic Segmentation Forecasting Colin Graber, Grace Tsai, Michael Firman, Gabriel Brostow, Alexander Schwing - CVPR 2021 [Link to paper] We propose
FuseDream This repo contains code for our paper (paper link): FuseDream: Training-Free Text-to-Image Generation with Improved CLIP+GAN Space Optimizat
CLIP-GEN [简体中文][English] 本项目在萤火二号集群上用 PyTorch 实现了论文 《CLIP-GEN: Language-Free Training of a Text-to-Image Generator with CLIP》。 CLIP-GEN 是一个 Language-F
rigidmask Code for "Learning to Segment Rigid Motions from Two Frames". ** This is a partial release with inference and evaluation code.
Combining Events and Frames using Recurrent Asynchronous Multimodal Networks for Monocular Depth Prediction This is the code for the paper Combining E
Overview | Tutorials | Examples | Installation | FAQ | How to Cite Welcome to ktrain News and Announcements 2020-11-08: ktrain v0.25.x is released and
YOLOV4:You Only Look Once目标检测模型在pytorch当中的实现 2021年2月7日更新: 加入letterbox_image的选项,关闭letterbox_image后网络的map得到大幅度提升。 目录 性能情况 Performance 实现的内容 Achievement
This repository is a toolkit to do machine learning for programming languages. It implements tokenization, dataset preprocessing, model training and m
Applicator Kit for Modo Applicator Kit for Modo allow you to apply Apple ARKit Face Tracking data from your iPhone or iPad with a TrueDepth camera to