FMA: A Dataset For Music Analysis

Overview

FMA: A Dataset For Music Analysis

Michaël Defferrard, Kirell Benzi, Pierre Vandergheynst, Xavier Bresson.
International Society for Music Information Retrieval Conference (ISMIR), 2017.

We introduce the Free Music Archive (FMA), an open and easily accessible dataset suitable for evaluating several tasks in MIR, a field concerned with browsing, searching, and organizing large music collections. The community's growing interest in feature and end-to-end learning is however restrained by the limited availability of large audio datasets. The FMA aims to overcome this hurdle by providing 917 GiB and 343 days of Creative Commons-licensed audio from 106,574 tracks from 16,341 artists and 14,854 albums, arranged in a hierarchical taxonomy of 161 genres. It provides full-length and high-quality audio, pre-computed features, together with track- and user-level metadata, tags, and free-form text such as biographies. We here describe the dataset and how it was created, propose a train/validation/test split and three subsets, discuss some suitable MIR tasks, and evaluate some baselines for genre recognition. Code, data, and usage examples are available at https://github.com/mdeff/fma.

Data

All metadata and features for all tracks are distributed in fma_metadata.zip (342 MiB). The below tables can be used with pandas or any other data analysis tool. See the paper or the usage.ipynb notebook for a description.

  • tracks.csv: per track metadata such as ID, title, artist, genres, tags and play counts, for all 106,574 tracks.
  • genres.csv: all 163 genres with name and parent (used to infer the genre hierarchy and top-level genres).
  • features.csv: common features extracted with librosa.
  • echonest.csv: audio features provided by Echonest (now Spotify) for a subset of 13,129 tracks.

Then, you got various sizes of MP3-encoded audio data:

  1. fma_small.zip: 8,000 tracks of 30s, 8 balanced genres (GTZAN-like) (7.2 GiB)
  2. fma_medium.zip: 25,000 tracks of 30s, 16 unbalanced genres (22 GiB)
  3. fma_large.zip: 106,574 tracks of 30s, 161 unbalanced genres (93 GiB)
  4. fma_full.zip: 106,574 untrimmed tracks, 161 unbalanced genres (879 GiB)

See the wiki (or #41) for known issues (errata).

Code

The following notebooks, scripts, and modules have been developed for the dataset.

  1. usage.ipynb: shows how to load the datasets and develop, train, and test your own models with it.
  2. analysis.ipynb: exploration of the metadata, data, and features. Creates the figures used in the paper.
  3. baselines.ipynb: baseline models for genre recognition, both from audio and features.
  4. features.py: features extraction from the audio (used to create features.csv).
  5. webapi.ipynb: query the web API of the FMA. Can be used to update the dataset.
  6. creation.ipynb: creation of the dataset (used to create tracks.csv and genres.csv).
  7. creation.py: creation of the dataset (long-running data collection and processing).
  8. utils.py: helper functions and classes.

Usage

Binder   Click the binder badge to play with the code and data from your browser without installing anything.

  1. Clone the repository.

    git clone https://github.com/mdeff/fma.git
    cd fma
  2. Create a Python 3.6 environment.
    # with https://conda.io
    conda create -n fma python=3.6
    conda activate fma
    
    # with https://github.com/pyenv/pyenv
    pyenv install 3.6.0
    pyenv virtualenv 3.6.0 fma
    pyenv activate fma
    
    # with https://pipenv.pypa.io
    pipenv --python 3.6
    pipenv shell
    
    # with https://docs.python.org/3/tutorial/venv.html
    python3.6 -m venv ./env
    source ./env/bin/activate
  3. Install dependencies.

    pip install --upgrade pip setuptools wheel
    pip install numpy==1.12.1  # workaround resampy's bogus setup.py
    pip install -r requirements.txt

    Note: you may need to install ffmpeg or graphviz depending on your usage.
    Note: install CUDA to train neural networks on GPUs (see Tensorflow's instructions).

  4. Download some data, verify its integrity, and uncompress the archives.

    cd data
    
    curl -O https://os.unil.cloud.switch.ch/fma/fma_metadata.zip
    curl -O https://os.unil.cloud.switch.ch/fma/fma_small.zip
    curl -O https://os.unil.cloud.switch.ch/fma/fma_medium.zip
    curl -O https://os.unil.cloud.switch.ch/fma/fma_large.zip
    curl -O https://os.unil.cloud.switch.ch/fma/fma_full.zip
    
    echo "f0df49ffe5f2a6008d7dc83c6915b31835dfe733  fma_metadata.zip" | sha1sum -c -
    echo "ade154f733639d52e35e32f5593efe5be76c6d70  fma_small.zip"    | sha1sum -c -
    echo "c67b69ea232021025fca9231fc1c7c1a063ab50b  fma_medium.zip"   | sha1sum -c -
    echo "497109f4dd721066b5ce5e5f250ec604dc78939e  fma_large.zip"    | sha1sum -c -
    echo "0f0ace23fbe9ba30ecb7e95f763e435ea802b8ab  fma_full.zip"     | sha1sum -c -
    
    unzip fma_metadata.zip
    unzip fma_small.zip
    unzip fma_medium.zip
    unzip fma_large.zip
    unzip fma_full.zip
    
    cd ..

    Note: try 7zip if decompression errors. It might be an unsupported compression issue.

  5. Fill a .env configuration file (at repository's root) with the following content.

    AUDIO_DIR=./data/fma_small/  # the path to a decompressed fma_*.zip
    FMA_KEY=MYKEY  # only if you want to query the freemusicarchive.org API
    
  6. Open Jupyter or run a notebook.

    jupyter notebook
    make usage.ipynb

Impact, coverage, and resources

100+ research papers

Full list on Google Scholar. Some picks below.

2 derived works
~10 posts
5 events
~10 dataset lists

Contributing

Contribute by opening an issue or a pull request. Let this repository be a hub around the dataset!

History

2017-05-09 pre-publication release

  • paper: arXiv:1612.01840v2
  • code: git tag rc1
  • fma_metadata.zip sha1: f0df49ffe5f2a6008d7dc83c6915b31835dfe733
  • fma_small.zip sha1: ade154f733639d52e35e32f5593efe5be76c6d70
  • fma_medium.zip sha1: c67b69ea232021025fca9231fc1c7c1a063ab50b
  • fma_large.zip sha1: 497109f4dd721066b5ce5e5f250ec604dc78939e
  • fma_full.zip sha1: 0f0ace23fbe9ba30ecb7e95f763e435ea802b8ab
  • known issues: see #41

2016-12-06 beta release

  • paper: arXiv:1612.01840v1
  • code: git tag beta
  • fma_small.zip sha1: e731a5d56a5625f7b7f770923ee32922374e2cbf
  • fma_medium.zip sha1: fe23d6f2a400821ed1271ded6bcd530b7a8ea551

Acknowledgments and Licenses

We are grateful to the Swiss Data Science Center (EPFL and ETHZ) for hosting the dataset.

Please cite our work if you use our code or data.

@inproceedings{fma_dataset,
  title = {{FMA}: A Dataset for Music Analysis},
  author = {Defferrard, Micha\"el and Benzi, Kirell and Vandergheynst, Pierre and Bresson, Xavier},
  booktitle = {18th International Society for Music Information Retrieval Conference (ISMIR)},
  year = {2017},
  archiveprefix = {arXiv},
  eprint = {1612.01840},
  url = {https://arxiv.org/abs/1612.01840},
}
@inproceedings{fma_challenge,
  title = {Learning to Recognize Musical Genre from Audio},
  subtitle = {Challenge Overview},
  author = {Defferrard, Micha\"el and Mohanty, Sharada P. and Carroll, Sean F. and Salath\'e, Marcel},
  booktitle = {The 2018 Web Conference Companion},
  year = {2018},
  publisher = {ACM Press},
  isbn = {9781450356404},
  doi = {10.1145/3184558.3192310},
  archiveprefix = {arXiv},
  eprint = {1803.05337},
  url = {https://arxiv.org/abs/1803.05337},
}
Comments
  • Files (much) shorter than 30s in fma-small

    Files (much) shorter than 30s in fma-small

    Hi, there are 6 files that are much shorter than 30s:

    fma_small/098/098565.mp3 --> 1.6s
    fma_small/098/098567.mp3 --> 0.5s
    fma_small/098/098569.mp3 --> 1.5s
    fma_small/099/099134.mp3 --> 0s
    fma_small/108/108925.mp3 --> 0s
    fma_small/133/133297.mp3 --> 0s
    

    , in case it's not a known issue.

    opened by keunwoochoi 18
  • zip: compress with deflate instead of bzip2

    zip: compress with deflate instead of bzip2

    I tried downloading the main metadata file to look at the underlying CSVs: https://os.unil.cloud.switch.ch/fma/fma_metadata.zip

    I'm getting a rejection on both mac and windows when I try to unzip this. Am I missing something?

    opened by mhamberg1 17
  • ValueError: Got an unexpected argument: categories

    ValueError: Got an unexpected argument: categories

    Hi, I've referred to the Usage section in the README as well as #9 and #10. I've checked out rc1 because it's appropriate for the version of fma_metadata.zip and fma_small.zip that I checked out, and also I've set my environment variables.

    Nevertheless, running the line

        tracks = utils.load('tracks.csv')
    

    in either the usage.ipynb file or my own very simple Python script will produce a ValueError about categories:

    Traceback (most recent call last):
      File "proc_fma.py", line 3, in <module>
        tracks = utils.load('fma_metadata/tracks.csv')
      File "/media/datadrive/datasets/fma/utils.py", line 213, in load
        'category', categories=SUBSETS, ordered=True)
      File "/home/shawley/anaconda3/envs/panotti/lib/python3.6/site-packages/pandas/core/generic.py", line 5883, in astype
        dtype=dtype, copy=copy, errors=errors, **kwargs
      File "/home/shawley/anaconda3/envs/panotti/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 581, in astype
        return self.apply("astype", dtype=dtype, **kwargs)
      File "/home/shawley/anaconda3/envs/panotti/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 438, in apply
        applied = getattr(b, f)(**kwargs)
      File "/home/shawley/anaconda3/envs/panotti/lib/python3.6/site-packages/pandas/core/internals/blocks.py", line 557, in astype
        return self._astype(dtype, copy=copy, errors=errors, values=values, **kwargs)
      File "/home/shawley/anaconda3/envs/panotti/lib/python3.6/site-packages/pandas/core/internals/blocks.py", line 598, in _astype
        "Got an unexpected argument: {}".format(deprecated_arg)
    ValueError: Got an unexpected argument: categories
    
    

    I haven't seen this error reported in any of the other issues. Can anyone help, e.g. @mdeff ? Thanks!

    opened by drscotthawley 13
  • Fix baselines.py

    Fix baselines.py

    This pull requests address issue #15 (which was related to corrupt files / time data mismatch) by throwing out corrupt files when being loaded.

    Any help in testing master and code revisions/suggestions are appreciated.

    opened by JustinKavalan 12
  • Corrupted files in FMA Large

    Corrupted files in FMA Large

    I didn't double check, but I couldn't open files with these indices on linux/ffmpeg/librosa. Just wanted to share so that others would get some hints.

    2624,
    3284,
    8669,
    10116,
    11583,
    12838,
    13529,
    14116,
    14180,
    20814,
    22554,
    23429,
    23430,
    23431,
    25173,
    25174,
    25175,
    25176,
    25180,
    29345,
    29346,
    29352,
    29356,
    33411,
    33413,
    33414,
    33417,
    33418,
    33419,
    33425,
    35725,
    39363,
    41745,
    42986,
    43753,
    50594,
    50782,
    53668,
    54569,
    54582,
    61480,
    61822,
    63422,
    63997,
    72656,
    72980,
    73510,
    80553,
    82699,
    84503,
    84504,
    84522,
    84524,
    86656,
    86659,
    86661,
    86664,
    87057,
    90244,
    90245,
    90247,
    90248,
    90250,
    90252,
    90253,
    90442,
    90445,
    91206,
    92479,
    94052,
    94234,
    95253,
    96203,
    96207,
    96210,
    98105,
    98562,
    101265,
    101272,
    101275,
    102241,
    102243,
    102247,
    102249,
    102289,
    106409,
    106412,
    106415,
    106628,
    108920,
    109266,
    110236,
    115610,
    117441,
    127928,
    129207,
    129800,
    130328,
    130748,
    130751,
    131545,
    133641,
    133647,
    134887,
    140449,
    140450,
    140451,
    140452,
    140453,
    140454,
    140455,
    140456,
    140457,
    140458,
    140459,
    140460,
    140461,
    140462,
    140463,
    140464,
    140465,
    140466,
    140467,
    140468,
    140469,
    140470,
    140471,
    140472,
    142614,
    144518,
    144619,
    145056,
    146056,
    147419,
    147424,
    148786,
    148787,
    148788,
    148789,
    148790,
    148791,
    148792,
    148793,
    148794,
    148795,
    151920,
    155051,
    
    opened by keunwoochoi 9
  • can not run deep learning baseline.

    can not run deep learning baseline.

    When I run the baseline, I bump into this problems. Can anyone help me with this?

    Dimensionality: (59953,)
    Epoch 1/2
     1664/19922 [=>............................] - ETA: 2559s - loss: 15.5950 - acc: 0.0325
    Process Process-7:
    Traceback (most recent call last):
      File "/anaconda3/envs/deeplearning3.5/lib/python3.5/multiprocessing/process.py", line 252, in _bootstrap
        self.run()
      File "/anaconda3/envs/deeplearning3.5/lib/python3.5/multiprocessing/process.py", line 93, in run
        self._target(*self._args, **self._kwargs)
      File "/anaconda3/envs/deeplearning3.5/lib/python3.5/site-packages/keras/engine/training.py", line 429, in data_generator_task
        generator_output = next(self._generator)
      File "/Users/dc/Downloads/fma-rc1/utils.py", line 327, in __next__
        self.X[i] = self.loader.load(get_audio_path(audio_dir, tid))
    ValueError: could not broadcast input array from shape (59943) into shape (59953)
     1696/19922 [=>............................] - ETA: 2552s - loss: 15.5954 - acc: 0.0324
    ---------------------------------------------------------------------------
    ValueError                                Traceback (most recent call last)
    <ipython-input-25-373babcd7ac0> in <module>()
         16 model.compile(optimizer, loss='categorical_crossentropy', metrics=['accuracy'])
         17 
    ---> 18 model.fit_generator(SampleLoader(train, batch_size=32), train.size, nb_epoch=2, **params)
         19 loss = model.evaluate_generator(SampleLoader(val, batch_size=32), val.size, **params)
         20 loss = model.evaluate_generator(SampleLoader(test, batch_size=32), test.size, **params)
    
    /anaconda3/envs/deeplearning3.5/lib/python3.5/site-packages/keras/models.py in fit_generator(self, generator, samples_per_epoch, nb_epoch, verbose, callbacks, validation_data, nb_val_samples, class_weight, max_q_size, nb_worker, pickle_safe, initial_epoch, **kwargs)
        933                                         nb_worker=nb_worker,
        934                                         pickle_safe=pickle_safe,
    --> 935                                         initial_epoch=initial_epoch)
        936 
        937     def evaluate_generator(self, generator, val_samples,
    
    /anaconda3/envs/deeplearning3.5/lib/python3.5/site-packages/keras/engine/training.py in fit_generator(self, generator, samples_per_epoch, nb_epoch, verbose, callbacks, validation_data, nb_val_samples, class_weight, max_q_size, nb_worker, pickle_safe, initial_epoch)
       1530                                          '(x, y, sample_weight) '
       1531                                          'or (x, y). Found: ' +
    -> 1532                                          str(generator_output))
       1533                     if len(generator_output) == 2:
       1534                         x, y = generator_output
    
    ValueError: output of generator should be a tuple (x, y, sample_weight) or (x, y). Found: None
    
    
    opened by YaoStriveCode 8
  • Errors in the FMA_large.zip and FMA_full.zip

    Errors in the FMA_large.zip and FMA_full.zip

    There are some errors in the FMA_large.zip and FMA_full.zip. I used multiple download approaches (wget and curl) from multiple links (zstd files are unavailable, the one on github repo and the kaggle version mentioned on other issues), and used multiple approaches for decompression (unzip, 7zip, tar, bzip etc.) on pmultiple linux machines. And there are many files are distorted in all the cases, such as:

    /fma_large/000/000148.mp3 /fma_large/000/000149.mp3 /fma_large/000/000150.mp3 /fma_large/000/000151.mp3 /fma_large/000/000152.mp3 /fma_large/001/001000.mp3 /fma_large/001/001001.mp3 /fma_large/002/002076.mp3 /fma_large/002/002077.mp3 /fma_large/002/002078.mp3 /fma_large/002/002079.mp3 /fma_large/002/002080.mp3 /fma_large/002/002081.mp3 /fma_large/002/002082.mp3

    I believed the kaggle version uploaded 9 months ago from github is a good demo of such noise. Could you please zip the fma_large and fma_full again and release them?

    opened by nicolaus625 6
  • Image data of album covers

    Image data of album covers

    Hi, the FMA dataset looks amazing, thank you so much for sharing this!

    I'm planning a research project that will involve multimodal models trained on album covers as well as the audio signal of tracks from the respective albums. Does FMA include cover art?

    If I understand correctly from the FMA paper, cover art is not yet included. Quoting from the discussion section:

    Cover images for tracks, albums, and artists are another public asset which may be of interest.

    From what I can tell on the freemusicarchive.org website, albums seem to usually (always?) come with a cover image. I think I might be able to automatically download these based on the album names in the FMA dataset. However, I'm wondering if there is a better way. I would appreciate any recommendation.

    opened by leod 6
  • Corrupted Files?

    Corrupted Files?

    Hello, I was trying to convert the small dataset to .wav using pydub and some files gave me errors trying to import. I tried them with librosa and they also failed. The files are as listed:

    fma_small/099/099134.mp3 fma_small/108/108925.mp3 fma_small/133/133297.mp3

    Please let me know if I did something wrong or if you are also getting the error. Thanks.

    opened by albert239825 5
  • How can I split the fma-small dataset with genres?

    How can I split the fma-small dataset with genres?

    I can not find a way to split the source audio zip to 8 genres with 1000 tracks.I can't find a file to help me do this.Would you mind helping me? Please...Thanks

    opened by DennisShaw 5
  • Error while loading tracks.csv

    Error while loading tracks.csv


    KeyError Traceback (most recent call last) in () 3 4 # Load metadata and features. ----> 5 tracks = utils.load('tracks.csv') 6 genres = utils.load('genres.csv') 7 features = utils.load('features.csv')

    ~\Desktop\ML\DeepAudioClassification-master - Copy\utils.py in load(filepath) 201 ('track', 'genres_top')] 202 for column in COLUMNS: --> 203 tracks[column] = tracks[column].map(ast.literal_eval) 204 205 COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'),

    ~\Anaconda3\lib\site-packages\pandas\core\frame.py in getitem(self, key) 2055 if isinstance(i, slice): 2056 return self[i] -> 2057 else: 2058 label = self.index[i] 2059 if isinstance(label, Index):

    ~\Anaconda3\lib\site-packages\pandas\core\frame.py in _getitem_multilevel(self, key) 2099 # a 0-len ndarray. This is effectively catching 2100 # a numpy error (as numpy should really raise) -> 2101 values = self._data.iget(i) 2102 2103 if index_len and not len(values):

    ~\Anaconda3\lib\site-packages\pandas\indexes\multi.py in get_loc(self, key, method)

    pandas\index.pyx in pandas.index.IndexEngine.get_loc (pandas\index.c:4433)()

    pandas\index.pyx in pandas.index.IndexEngine.get_loc (pandas\index.c:4279)()

    pandas\src\hashtable_class_helper.pxi in pandas.hashtable.PyObjectHashTable.get_item (pandas\hashtable.c:13742)()

    pandas\src\hashtable_class_helper.pxi in pandas.hashtable.PyObjectHashTable.get_item (pandas\hashtable.c:13696)()

    KeyError: ('track', 'genres_top')

    Error while loading csv files Does anyone know fix?

    opened by izidov 5
  • Made the logarithmic image of `cross_correlation` more readable

    Made the logarithmic image of `cross_correlation` more readable

    Before modification, the logarithmic image of cross_correlation is as follows:

    cross_correlation_log_old

    This drawing method does not conform to our habit of reading an (X, Y) image. To make it more readable, I committed this pull request.

    After modification, the logarithmic image of cross_correlation is as follows:

    cross_correlation_log_new

    opened by KofClubs 0
  • Deleted redundant `genres_names` definition

    Deleted redundant `genres_names` definition

    In section 4.2 of analysis.ipynb, genres_names is repeatedly defined:

    genres_names = enc.classes_
    genres_names = genres.loc[enc.classes_,'title'].values
    

    I deleted the upper line.

    opened by KofClubs 0
  • Added a prompt not to add comments in `.env`

    Added a prompt not to add comments in `.env`

    I added the following to the .env configuration file:

    AUDIO_DIR=./data/fma_small/  # the path to a decompressed fma_*.zip
    FMA_KEY=MYKEY  # only if you want to query the freemusicarchive.org API
    

    When I run the 12th unit of usage.ipynb, I get an error.

    Input:

    filename = utils.get_audio_path(AUDIO_DIR, 2)
    print('File: {}'.format(filename))
    
    x, sr = librosa.load(filename, sr=None, mono=True)
    print('Duration: {:.2f}s, {} samples'.format(x.shape[-1] / sr, x.size))
    
    start, end = 7, 17
    ipd.Audio(data=x[start*sr:end*sr], rate=sr)
    

    Output:

    File: ./data/fma_small/  # the path to a decompressed fma_*.zip/000/000002.mp3
    ---------------------------------------------------------------------------
    FileNotFoundError                         Traceback (most recent call last)
    <ipython-input-12-858041869145> in <module>()
          2 print('File: {}'.format(filename))
          3 
    ----> 4 x, sr = librosa.load(filename, sr=None, mono=True)
          5 print('Duration: {:.2f}s, {} samples'.format(x.shape[-1] / sr, x.size))
          6 
    
    /opt/homebrew/anaconda3/envs/fma/lib/python3.6/site-packages/librosa/core/audio.py in load(path, sr, mono, offset, duration, dtype, res_type)
        105 
        106     y = []
    --> 107     with audioread.audio_open(os.path.realpath(path)) as input_file:
        108         sr_native = input_file.samplerate
        109         n_channels = input_file.channels
    
    /opt/homebrew/anaconda3/envs/fma/lib/python3.6/site-packages/audioread/__init__.py in audio_open(path)
         76     from . import rawread
         77     try:
    ---> 78         return rawread.RawAudioFile(path)
         79     except DecodeError:
         80         pass
    
    /opt/homebrew/anaconda3/envs/fma/lib/python3.6/site-packages/audioread/rawread.py in __init__(self, filename)
         47     """
         48     def __init__(self, filename):
    ---> 49         self._fh = open(filename, 'rb')
         50 
         51         try:
    
    FileNotFoundError: [Errno 2] No such file or directory: '/Users/zhang/Documents/Repositories.nosync/fma/data/fma_small/  # the path to a decompressed fma_*.zip/000/000002.mp3'
    

    This error is because when the environment variable AUDIO_DIR is read, its comment (# the path to a decompressed fma_*.zip) is also read as part of AUDIO_DIR.

    I added a prompt at the corresponding position of README.md.

    opened by KofClubs 0
  • Known issues (and next release)

    Known issues (and next release)

    Below are issues affecting the rc1 data release that cannot be fixed without a data update. As updating is disruptive (it'll break code and make results non-comparable), it should be done sparingly, e.g., to fix a fatal flaw or many small ones discovered over time.

    • zip decompression fails because of unsupported bzip2 compression (#5)
      • [x] workaround (master): note in README to try with 7zip (5700859)
      • [ ] fix (next): zip with deflate (instead of bzip2) (#5) or zstd (#32)
    • excerpts shorter than 30s and erroneous audio length metadata (#4, #8, #36, #44)
    • erroneous ID3 tags (#27)
      • [x] workaround (master): list (#27)
      • [ ] fix (next): dump ID3 tags with technical metadata and remove from mp3
    • exact duplicate tracks (#23)
      • [ ] workaround (master): list the 937 duplicates
      • [ ] fix (next): remove them (try other methods and detect near duplicates)

    Workarounds are explained in more details in the wiki.

    opened by mdeff 2
Owner
Michaël Defferrard
Research on machine learning and graphs. Open science, source, data.
Michaël Defferrard
Official Implementation and Dataset of "PPR10K: A Large-Scale Portrait Photo Retouching Dataset with Human-Region Mask and Group-Level Consistency", CVPR 2021

Portrait Photo Retouching with PPR10K Paper | Supplementary Material PPR10K: A Large-Scale Portrait Photo Retouching Dataset with Human-Region Mask an

null 184 Dec 11, 2022
This is the dataset and code release of the OpenRooms Dataset.

This is the dataset and code release of the OpenRooms Dataset.

Visual Intelligence Lab of UCSD 95 Jan 8, 2023
A large dataset of 100k Google Satellite and matching Map images, resembling pix2pix's Google Maps dataset.

Larger Google Sat2Map dataset This dataset extends the aerial ⟷ Maps dataset used in pix2pix (Isola et al., CVPR17). The provide script download_sat2m

null 34 Dec 28, 2022
Dataset used in "PlantDoc: A Dataset for Visual Plant Disease Detection" accepted in CODS-COMAD 2020

PlantDoc: A Dataset for Visual Plant Disease Detection This repository contains the Cropped-PlantDoc dataset used for benchmarking classification mode

Pratik Kayal 109 Dec 29, 2022
EMNLP 2021: Single-dataset Experts for Multi-dataset Question-Answering

MADE (Multi-Adapter Dataset Experts) This repository contains the implementation of MADE (Multi-adapter dataset experts), which is described in the pa

Princeton Natural Language Processing 68 Jul 18, 2022
EMNLP 2021: Single-dataset Experts for Multi-dataset Question-Answering

MADE (Multi-Adapter Dataset Experts) This repository contains the implementation of MADE (Multi-adapter dataset experts), which is described in the pa

Princeton Natural Language Processing 39 Oct 5, 2021
LoveDA: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation (NeurIPS2021 Benchmark and Dataset Track)

LoveDA: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation by Junjue Wang, Zhuo Zheng, Ailong Ma, Xiaoyan Lu, and Yanfei Zh

Kingdrone 174 Dec 22, 2022
This is the official source code for SLATE. We provide the code for the model, the training code, and a dataset loader for the 3D Shapes dataset. This code is implemented in Pytorch.

SLATE This is the official source code for SLATE. We provide the code for the model, the training code and a dataset loader for the 3D Shapes dataset.

Gautam Singh 66 Dec 26, 2022
The Habitat-Matterport 3D Research Dataset - the largest-ever dataset of 3D indoor spaces.

Habitat-Matterport 3D Dataset (HM3D) The Habitat-Matterport 3D Research Dataset is the largest-ever dataset of 3D indoor spaces. It consists of 1,000

Meta Research 62 Dec 27, 2022
Delta Conformity Sociopatterns Analysis - Delta Conformity Sociopatterns Analysis

Delta_Conformity_Sociopatterns_Analysis ∆-Conformity is a local homophily measur

null 2 Jan 9, 2022
Streamlit App For Product Analysis - Streamlit App For Product Analysis

Streamlit_App_For_Product_Analysis Здравствуйте! Перед вами дашборд, позволяющий

Grigory Sirotkin 1 Jan 10, 2022
Woosung Choi 63 Nov 14, 2022
Self-Supervised Contrastive Learning of Music Spectrograms

Self-Supervised Music Analysis Self-Supervised Contrastive Learning of Music Spectrograms Dataset Songs on the Billboard Year End Hot 100 were collect

null 27 Dec 10, 2022
Starter kit for getting started in the Music Demixing Challenge.

Music Demixing Challenge - Starter Kit ?? Challenge page This repository is the Music Demixing Challenge Submission template and Starter kit! Clone th

AIcrowd 106 Dec 20, 2022
Source code for "MusCaps: Generating Captions for Music Audio" (IJCNN 2021)

MusCaps: Generating Captions for Music Audio Ilaria Manco1 2, Emmanouil Benetos1, Elio Quinton2, Gyorgy Fazekas1 1 Queen Mary University of London, 2

Ilaria Manco 57 Dec 7, 2022
Source code and data from the RecSys 2020 article "Carousel Personalization in Music Streaming Apps with Contextual Bandits" by W. Bendada, G. Salha and T. Bontempelli

Carousel Personalization in Music Streaming Apps with Contextual Bandits - RecSys 2020 This repository provides Python code and data to reproduce expe

Deezer 48 Jan 2, 2023
Official repository for the paper, MidiBERT-Piano: Large-scale Pre-training for Symbolic Music Understanding.

MidiBERT-Piano Authors: Yi-Hui (Sophia) Chou, I-Chun (Bronwin) Chen Introduction This is the official repository for the paper, MidiBERT-Piano: Large-

null 137 Dec 15, 2022
PyTorch implementation of MuseMorphose, a Transformer-based model for music style transfer.

MuseMorphose This repository contains the official implementation of the following paper: Shih-Lun Wu, Yi-Hsuan Yang MuseMorphose: Full-Song and Fine-

Yating Music, Taiwan AI Labs 142 Jan 8, 2023