AI-generated-characters for Learning and Wellbeing

Overview

AI-generated-characters for Learning and Wellbeing

Click here for the full project page.

This repository contains the source code for the paper AI-generated characters for supporting personalized learning and well-being by Pat Pataranutaporn, Valdemar Danry, Joanne Leong, Parinya Punpongsanon, Dan Novy, Pattie Maes & Misha Sra. This repository is a combination previous work on AI generated characters that include Siarohin et al., Prajwal et al., and Corentin.

Colab Demo

The code is available on google-colab. See: AI_Generated_Characters.ipynb. To run press Open In Colab button. Open In Colab

Examples of Outputs

Screenshot

With the pipeline, one can easily create a video of AI-generated characters from Video, Audio, and Text input (text is coming soon).

You might also like...
Lucid Sonic Dreams syncs GAN-generated visuals to music.

Lucid Sonic Dreams Lucid Sonic Dreams syncs GAN-generated visuals to music. By default, it uses NVLabs StyleGAN2, with pre-trained models lifted from

Utilities to bridge Canvas-generated course rosters with GitLab's API.

gitlab-canvas-utils A collection of scripts originally written for CSE 13S. Oversees everything from GitLab course group creation, student repository

PyKale is a PyTorch library for multimodal learning and transfer learning as well as deep learning and dimensionality reduction on graphs, images, texts, and videos
PyKale is a PyTorch library for multimodal learning and transfer learning as well as deep learning and dimensionality reduction on graphs, images, texts, and videos

PyKale is a PyTorch library for multimodal learning and transfer learning as well as deep learning and dimensionality reduction on graphs, images, texts, and videos. By adopting a unified pipeline-based API design, PyKale enforces standardization and minimalism, via reusing existing resources, reducing repetitions and redundancy, and recycling learning models across areas.

A resource for learning about deep learning techniques from regression to LSTM and Reinforcement Learning using financial data and the fitness functions of algorithmic trading

A tour through tensorflow with financial data I present several models ranging in complexity from simple regression to LSTM and policy networks. The s

Machine Learning From Scratch. Bare bones NumPy implementations of machine learning models and algorithms with a focus on accessibility. Aims to cover everything from linear regression to deep learning.
Machine Learning From Scratch. Bare bones NumPy implementations of machine learning models and algorithms with a focus on accessibility. Aims to cover everything from linear regression to deep learning.

Machine Learning From Scratch About Python implementations of some of the fundamental Machine Learning models and algorithms from scratch. The purpose

Vowpal Wabbit is a machine learning system which pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive learning.
Vowpal Wabbit is a machine learning system which pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive learning.

This is the Vowpal Wabbit fast online learning code. Why Vowpal Wabbit? Vowpal Wabbit is a machine learning system which pushes the frontier of machin

An image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testingAn image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testing
An image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testingAn image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testing

SVM Données Une base d’images contient 490 images pour l’apprentissage (400 voitures et 90 bateaux), et encore 21 images pour fait des tests. Prétrait

🔥 Cogitare - A Modern, Fast, and Modular Deep Learning and Machine Learning framework for Python
🔥 Cogitare - A Modern, Fast, and Modular Deep Learning and Machine Learning framework for Python

Cogitare is a Modern, Fast, and Modular Deep Learning and Machine Learning framework for Python. A friendly interface for beginners and a powerful too

Comments
  • OSError: MoviePy error: the file /content/Wav2Lip/results/result_voice.mp4 could not be found! Please check that you entered the correct path.

    OSError: MoviePy error: the file /content/Wav2Lip/results/result_voice.mp4 could not be found! Please check that you entered the correct path.

    Animating Character with Driving Video: This might take a few minutes.. /content/first-order-model /content Generating speech from text Lipsyncing Character with Audio /content/Wav2Lip /content


    OSError Traceback (most recent call last)

    in () 30 31 # Traceability ---> 32 tracability(final_video_driver) 33 final_video_driver = "marked.mp4" 34 get_ipython().system('ffmpeg -i $final_video_driver -i $audio_driver final_generated.mp4 -y &> /dev/null')

    3 frames

    /usr/local/lib/python3.7/dist-packages/moviepy/video/io/ffmpeg_reader.py in ffmpeg_parse_infos(filename, print_infos, check_duration, fps_source) 270 raise IOError(("MoviePy error: the file %s could not be found!\n" 271 "Please check that you entered the correct " --> 272 "path.")%filename) 273 274 result = dict()

    OSError: MoviePy error: the file /content/Wav2Lip/results/result_voice.mp4 could not be found! Please check that you entered the correct path.

    Could u please tell me why this happened when I use colab to run it?

    opened by kenchikuliu 2
  • Installation of colab utilities  (Documentation)

    Installation of colab utilities (Documentation)

    First of all awesome work! Just discovered this excellent resource and trying to recreate it. I faced the following issue while installing the colab_utilities. And had to utilize an alternative method. Documenting it over here.

    This particular installation didn't work for me

    !pip install https://github.com/tugstugi/dl-colab-notebooks/archive/colab_utils.zip -y &> /dev/null

    I had to use the following alternative -

    !wget https://github.com/tugstugi/dl-colab-notebooks/archive/colab_utils.zip
    !unzip colab_utils.zip
    
    !mv dl-colab-notebooks-colab_utils/dl_colab_notebooks/ .
    !mv dl-colab-notebooks-colab_utils/setup.py . 
    
    !pip install pydub
    

    I am using Colab (free version) Let me know if I am missing anything

    opened by skyprince999 1
  • A bug in the notebook

    A bug in the notebook

    My workflow included:

    • Ran the installation
    • Uploaded an image
    • Left the text box empty
    • Unchecked audio
    • Checked video
    • Uploaded driving video

    When I ran the generate block, the below error occured:

    OSError                                   Traceback (most recent call last)
    [<ipython-input-4-631bb51a4437>](https://localhost:8080/#) in <module>()
         33 
         34 # Traceability
    ---> 35 tracability(final_video_driver)
         36 final_video_driver = "marked.mp4"
         37 get_ipython().system('ffmpeg -i $final_video_driver -i $audio_driver final_generated.mp4 -y &> /dev/null')
    
    3 frames
    [/usr/local/lib/python3.7/dist-packages/moviepy/video/io/ffmpeg_reader.py](https://localhost:8080/#) in ffmpeg_parse_infos(filename, print_infos, check_duration, fps_source)
        270         raise IOError(("MoviePy error: the file %s could not be found!\n"
        271                       "Please check that you entered the correct "
    --> 272                       "path.")%filename)
        273 
        274     result = dict()
    
    OSError: MoviePy error: the file merged.mp4 could not be found!
    Please check that you entered the correct path.
    

    I observed that the If block starting from line 28 uses an audio driver. Which is not present.

    Is this expected, or am I missing something here? @valleballe

    opened by Sreerag-ibtl 2
Owner
MIT Media Lab
MIT Media Lab
A Nim frontend for pytorch, aiming to be mostly auto-generated and internally using ATen.

Master Release Pytorch - Py + Nim A Nim frontend for pytorch, aiming to be mostly auto-generated and internally using ATen. Because Nim compiles to C+

Giovanni Petrantoni 425 Dec 22, 2022
Fast and robust clustering of point clouds generated with a Velodyne sensor.

Depth Clustering This is a fast and robust algorithm to segment point clouds taken with Velodyne sensor into objects. It works with all available Velo

Photogrammetry & Robotics Bonn 957 Dec 21, 2022
[ICLR 2021] Rank the Episodes: A Simple Approach for Exploration in Procedurally-Generated Environments.

[ICLR 2021] RAPID: A Simple Approach for Exploration in Reinforcement Learning This is the Tensorflow implementation of ICLR 2021 paper Rank the Episo

Daochen Zha 48 Nov 21, 2022
Official Pytorch implementation of paper "Reverse Engineering of Generative Models: Inferring Model Hyperparameters from Generated Images"

Reverse_Engineering_GMs Official Pytorch implementation of paper "Reverse Engineering of Generative Models: Inferring Model Hyperparameters from Gener

null 100 Dec 18, 2022
BARTScore: Evaluating Generated Text as Text Generation

This is the Repo for the paper: BARTScore: Evaluating Generated Text as Text Generation Updates 2021.06.28 Release online evaluation Demo 2021.06.25 R

NeuLab 196 Dec 17, 2022
Tracing Versus Freehand for Evaluating Computer-Generated Drawings (SIGGRAPH 2021)

Tracing Versus Freehand for Evaluating Computer-Generated Drawings (SIGGRAPH 2021) Zeyu Wang, Sherry Qiu, Nicole Feng, Holly Rushmeier, Leonard McMill

Zach Zeyu Wang 23 Dec 9, 2022
GAN-generated image detection based on CNNs

GAN-image-detection This repository contains a GAN-generated image detector developed to distinguish real images from synthetic ones. The detector is

Image and Sound Processing Lab 17 Dec 15, 2022
Rename Images with Auto Generated Neural Image Captions

Recaption Images with Generated Neural Image Caption Example Usage: Commandline: Recaption all images from folder /home/feng/Downloads/images to folde

feng wang 3 May 1, 2022
Convolutional neural network that analyzes self-generated images in a variety of languages to find etymological similarities

This project is a convolutional neural network (CNN) that analyzes self-generated images in a variety of languages to find etymological similarities. Specifically, the goal is to prove that computer vision can be used to identify cognates known to exist, and perhaps lead linguists to evidence of unknown cognates.

null 1 Feb 3, 2022