Giter Club home page Giter Club logo

atepp's Introduction

ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performances

ATEPP is a dataset of expressive piano performances by virtuoso pianists. The dataset contains 11742 11674 performances (~1000 hours) by 49 pianists and covers 1580 1595 movements by 25 composers. All of the MIDI files in the dataset come from the piano transcription of existing audio recordings of piano performances. Scores in MusicXML format are also available for around half of the tracks. The dataset is organized and aligned by compositions and movements for comparative studies. More details are presented in the paper.

Downloade the ATEPP dataset

Please follow disclaimer.md to agree a disclaimer and download a latest version of ATEPP (~212MB).

Inference

You can inference your own track with the modified code and new checkpoint in piano_transcription-master. The env and setup are the same as https://github.com/bytedance/piano_transcription

python3 pytorch/inference.py --model_type=Regress_onset_offset_frame_velocity_CRNN --checkpoint_path=300000_iterations.pth --audio_path="resources/schumann_romanzen.mp3" --cuda

Released Versions

Version-1.2 (Latest Release!)

There are several issues found with the ATEPP Version-1.1:

  1. Corrupted Transcriptions (CTs) were found because of errors made by the transcription model when dealing with low-quality recordings (usually the live performances or old recordings). To detect the CTs, we compared the total note number and the duration of a performance with the medians for the same composition. We filtered out those which have similar duration to the midian but have much more/less notes with thresholds. We also mannually filtered out low-quality (annotated as low quality) audios by listening to them.

    • In total, 1264 audios were categorised as low-quality, having high possibility to lead to CTs. The corrupted refers to those confirmed corrupted with midi file.
    • In addition to low-quality audios, we annotated 1436 audios with background noise. These audios differ from low-quality audios in the way that they were transcribed with higher accuracy compared to those corresponding to low-quality audios. We suggest to filter out the corresponding midi files (tagged with low quality and background noise) when using the dataset.
    • As for live performances, we annotated applause if the recording contains that. We used high quality to refer to live recordings with good quality (clear, no applause, almost imperceptible background noise).
  2. Combined Movements (CMs) (one midi/audio consists of two or more movements) were found through a similar detection process of comparing the total note number and the duration with the medians. 7 were found, cut and relabeled.

  3. Error Labels (ELs) of composition were found when we manually verify the suspected pieces. 220 performances were found wrongly labelled and corrected.

The quality feature has been added to the metadata for clarifying the audio-related information. The repetition feature (not thoroughly analysed) refers to whether the performer plays the repetition sections or not.

Changed Statistics:

  • 11674 performances
  • 1595 movements

Version-1.1

!Updates: 65 Pieces Removed

When creating ATEPP version-1.0, we only applied movement-wise matching to remove erroneously downloaded audio. Now, we finished detecting repeated audios by audio-wise fingerprint matching. Only 65 audios were detected repeated, and the corresponding transcribed midi files were removed. The repeats.csv lists the repeated transcribed files that have been removed.

Changed Statistics:

  • 11677 performances
  • 1002 hours

Version-1.0

  • 11742 performances (in midi format)
  • 1007 hours
  • 1580 movements
  • 25 composers
  • 49 performers
  • 43% with scores

Related Works

We've released a Python package developed for linking classical music recording & track to the corresponding composition / movement, useful in cleaning up metadata in classical music datasets.

Package on PyPI: https://pypi.org/project/composition-entity-linker/

Contact

Citation

@inproceedings{zhang2022atepp,
  title={ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performance},
  author={Zhang, Huan and Tang, Jingjing and Rafee, Syed Rifat Mahmud and Fazekas, Simon Dixon Gy{\"o}rgy},
  booktitle={ISMIR 2022 Hybrid Conference},
  year={2022}
}

License

CC BY 4.0

atepp's People

Contributors

tangjjbetsy avatar

Stargazers

Hao Hao Tan avatar Yi-Hsuan Yang avatar  avatar  avatar Keshav Bhandari avatar binyued avatar  avatar  avatar Hatori233 avatar  avatar Nero Blackstone​ avatar Lele Liu avatar Emmanouil Karystinaios avatar audiocode avatar  avatar Federico Simonetta avatar Carlos Eduardo Cancino-Chacón avatar uco_physics avatar bgArray avatar  avatar Gerald Golka avatar Ilya Borovik avatar Kenzi NOIKE avatar aaronchen avatar Alex avatar Julian Lenz avatar Neil Lee avatar Alexis Spiliotopoulos avatar Nathan Fradet avatar Yuan-Man avatar Chin-Yun Yu avatar Carlos Hernández Oliván avatar Huan Zhang avatar

Watchers

Gerald Golka avatar  avatar

Forkers

tbergman freddd13

atepp's Issues

Inquiry

Hello, (I'm only a musician, not an expert in programming) but I got interested in the fact that note durations are more accurate in the midis you provide rather than in Bytedance's original system. I was wondering if you altered their method to achieve those results, if you want to comment on that a little! And I also can't wait to try your system, may I ask, will you release it soon? Thanks for the last response.

Copyright concern of this dataset

Hi, can I get some detail of the copyright in this dataset? According to the papar, it said "we download each track from a corresponding open source audio at YouTube Music".

But I check the csv file that most Youtube music downloaded is provided by the Universal Music Group, and the performed pianists are generally not "died over 70 year" so I assuming their arrangements are under the copyright.

Would you detail the "open source audio" about how this open souce? And same concern for the sheet music. For the pieces in MuseScore, does there scores are non-provided licenses (so have copyright issue), or did provide the opensource licenses? MuseScore is open to public and non-careful management the copyright, so made things become sensitive.

This is a very promising dataset, but unfortunately I need to address the Univerisity's concerns regarding copyright before the further research.

Best regards <3

Google Colab

Hello, is there by chance a colab to infer with your model and checkpoint? Thank you so much

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.