Lyrics generation with GPT2-based Transformer

Overview

HuggingArtists - Train a model to generate lyrics

Create AI-Artist in just 5 minutes!

🚀 Run the demo notebook to train

🚀 Run the GUI demo to test

Disclaimer: this project is not to be used to publish any false generated information or unpleasant words but to perform research on Natural Language Generation.

How to use

Just to generate without training, use Spaces with Streamlit. Open in Streamlit

To train with the demo, click on below link and share your results!

Open In Colab

You can also use it locally by installing the dependencies with pipenv or pip and use huggingartists-demo.ipynb

Results

There are some interesting lyrics. Contact me to add your results:

Eminem:

I am a bit of a nutter
Like a fucking crazy nut
But I just wanted to break and get a little more serious, so be patient
I dont think I need no medical attention
But I do know that these niggas gonna be in my prayers
And they say Im a nutter
Like a fucking crazy nut
But I love you
I am a nutter
Like a fuckin nutter
But I feel like the greatest that I can be
With your body and your spirits
But I dont know how I will act
When I finally get out
And I get pissed off
Ill be right back at ya with the first one on the floor
But I aint ever gonna get the respect youve been getting
And Im a nutter...
I am bipolar while I’m rhyming, bipolar while I am hate
When I’m rhyming, I may seem a little bit crazy
But the thing’s with these rhymes that I pick
I stick to the syllable ç unless it’s on a syllable
Bread is full of charmin I am thinkin all of the garbage
Hear the bass line and ya hear the bass line
But hip hop is what makes’em songs
I try to describe how I’m gettin’ em all, but I’m lazy
When I’m rapin on a song, I’m just making it
If I don’t got the right, In my mind...

Drake:

I am never livin the life that I never lived
I walk around with my Glock 9
And lay down with my feet up
This is my house, bitch
Welcome to Detroit, this is my city
Ive made it all my people proud
I even sold my house down
To get a possible placement as a burnt pit
My little girl, she should ntroller
Wanna flip sometimes on the upside
But its hard to hold a grudge, I dont bite
I say my wholehearted, prayer is my heart
I pray for you when Im gone, but I gotta leave you in peace
That we did not plan to be seeing each other, well
We did plan to be seeing each other, well
And if you ever need anything, I...

Check existing:

Features:

  • Generate lyrics by a specific artist
  • Automatically upload models and datasets to HuggingFace Profile after each run
  • Parse dataset once and use it next time without spending time
  • Each next run by any person will resume training from last checkpoint
  • Training and results are automatically logged into W&B through the HuggingFace integration

About

Built by Aleksey Korshuk

Follow

Follow

Follow

🚀 If you want to contribute to this project OR create something cool together — contact me: link

Star this repository:

GitHub stars

Resources

Got questions about W&B?

If you have any questions about using W&B to track your model performance and predictions, please reach out to the slack community.

You might also like...
Trankit is a Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing
Trankit is a Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing

Trankit: A Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing Trankit is a light-weight Transformer-based Pyth

Transformer-based Text Auto-encoder (T-TA) using TensorFlow 2.

T-TA (Transformer-based Text Auto-encoder) This repository contains codes for Transformer-based Text Auto-encoder (T-TA, paper: Fast and Accurate Deep

A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.
A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.

A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.

Transformer Based Korean Sentence Spacing Corrector
Transformer Based Korean Sentence Spacing Corrector

TKOrrector Transformer Based Korean Sentence Spacing Corrector License Summary This solution is made available under Apache 2 license. See the LICENSE

A Word Level Transformer layer based on PyTorch and 🤗 Transformers.

Transformer Embedder A Word Level Transformer layer based on PyTorch and 🤗 Transformers. How to use Install the library from PyPI: pip install transf

RoNER is a Named Entity Recognition model based on a pre-trained BERT transformer model trained on RONECv2

RoNER RoNER is a Named Entity Recognition model based on a pre-trained BERT transformer model trained on RONECv2. It is meant to be an easy to use, hi

A simple command line tool for text to image generation, using OpenAI's CLIP and a BigGAN
A simple command line tool for text to image generation, using OpenAI's CLIP and a BigGAN

artificial intelligence cosmic love and attention fire in the sky a pyramid made of ice a lonely house in the woods marriage in the mountains lantern

Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow.  This is part of the CASL project: http://casl-project.ai/
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project: http://casl-project.ai/

Texar is a toolkit aiming to support a broad set of machine learning, especially natural language processing and text generation tasks. Texar provides

Comments
  • Dataset does not exist?

    Dataset does not exist?

    Hi I'm trying to find a tutorial that explains this but I'm getting an error of Couldn't find a dataset script. I'm getting stuck at the collect data section under preprocess data. I'm in colab on a free session.

    opened by jscott5811 2
  • [Errno 403] 403 Client Error: Forbidden for url

    [Errno 403] 403 Client Error: Forbidden for url

    There may be an issue relating to the use of LyricsGenius under the Attach Needed Code step:

    [Errno 403] 403 Client Error: Forbidden for url: https://genius.com/api/search/multi?q=Eminem

    opened by Ghirahim101 0
Kinky furry assitant based on GPT2

KinkyFurs-V0 Kinky furry assistant based on GPT2 How to run python3 V0.py then, open web browser and go to localhost:8080 Requirements: Flask trans

Sparki 1 Jun 11, 2022
**NSFW** A chatbot based on GPT2-chitchat

DangBot -- 好怪哦,再来一句 卡群怪话bot,powered by GPT2 for Chinese chitchat Training Example: python train.py --lr 5e-2 --epochs 30 --max_len 300 --batch_size 8

Tommy Yang 11 Jul 21, 2022
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.

Kashgari Overview | Performance | Installation | Documentation | Contributing ?? ?? ?? We released the 2.0.0 version with TF2 Support. ?? ?? ?? If you

Eliyar Eziz 2.3k Dec 29, 2022
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.

Kashgari Overview | Performance | Installation | Documentation | Contributing ?? ?? ?? We released the 2.0.0 version with TF2 Support. ?? ?? ?? If you

Eliyar Eziz 2k Feb 9, 2021
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and GPT-NEO (2.7 B) on a single 16 GB VRAM V100 Google Cloud instance with Huggingface Transformers using DeepSpeed

Guide: Finetune GPT2-XL (1.5 Billion Parameters) and GPT-NEO (2.7 Billion Parameters) on a single 16 GB VRAM V100 Google Cloud instance with Huggingfa

null 289 Jan 6, 2023
Unsupervised text tokenizer for Neural Network-based text generation.

SentencePiece SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabu

Google 6.4k Jan 1, 2023
Unsupervised text tokenizer for Neural Network-based text generation.

SentencePiece SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabu

Google 4.8k Feb 18, 2021
To be a next-generation DL-based phenotype prediction from genome mutations.

Sequence -----------+--> 3D_structure --> 3D_module --+ +--> ? | |

Eric Alcaide 18 Jan 11, 2022
Code for lyric-section-to-comment generation based on huggingface transformers.

CommentGeneration Code for lyric-section-to-comment generation based on huggingface transformers. Migrate Guyu model and code (both 12-layers and 24-l

Yawei Sun 8 Sep 4, 2021