Final Research Report on Auto-Tagging of Music

dc.contributor.authorSchwarz, Diemo
dc.contributor.authorPeeters, Geoffroy
dc.contributor.authorCohen-Hadria, Alice
dc.contributor.authorFourer, Dominique
dc.contributor.authorMarchand, Ugo
dc.contributor.authorMignot, Rémi
dc.contributor.authorCornu, Frédéric
dc.contributor.authorLaffitte, Pierre
dc.contributor.authorSchindler, Daniel
dc.contributor.authorHofmann, Robin
dc.contributor.authorSpadaveccia, Rino
dc.date.accessioned2019-02-18T13:42:47Z
dc.date.available2019-02-18T13:42:47Z
dc.date.issued2018-12-12
dc.description.abstractThe deliverable D4.7 concerns the work achieved by IRCAM until M36 for the “auto-tagging of music”. The deliverable is a research report. The software libraries resulting from the research have been integrated into Fincons/HearDis! Music Library Manager or are used by TU Berlin. The final software libraries are described in D4.5. The research work on auto-tagging has concentrated on four aspects: 1) Further improving IRCAM’s machine-learning system ircamclass. This has been done by developing the new MASSS audio features, including audio augmentation and audio segmentation into ircamclass. The system has then been applied to train HearDis! “soft” features (Vocals-1, Vocals-2, Pop-Appeal, Intensity, Instrumentation, Timbre, Genre, Style). This is described in Part 3. 2) Developing two sets of “hard” features (i.e. related to musical or musicological concepts) as specified by HearDis! (for integration into Fincons/HearDis! Music Library Manager) and TU Berlin (as input for the prediction model of the GMBI attributes). Such features are either derived from previously estimated higher-level concepts (such as structure, key or succession of chords) or by developing new signal processing algorithm (such as HPSS) or main melody estimation. This is described in Part 4. 3) Developing audio features to characterize the audio quality of a music track. The goal is to describe the quality of the audio independently of its apparent encoding. This is then used to estimate audio degradation or music decade. This is to be used to ensure that playlists contain tracks with similar audio quality. This is described in Part 5. 4) Developing innovative algorithms to extract specific audio features to improve music mixes. So far, innovative techniques (based on various Blind Audio Source Separation algorithms and Convolutional Neural Network) have been developed for singing voice separation, singing voice segmentation, music structure boundaries estimation, and DJ cue-region estimation. This is described in Part 6.en
dc.description.sponsorshipEC/H2020/688122/EU/Artist-to-Business-to-Business-to-Consumer Audio Branding System/ABC DJen
dc.identifier.urihttps://depositonce.tu-berlin.de/handle/11303/9067
dc.identifier.urihttp://dx.doi.org/10.14279/depositonce-8168
dc.language.isoenen
dc.rights.urihttps://creativecommons.org/licenses/by-nc/4.0/en
dc.subject.ddc780 Musikde
dc.subject.otheraudio brandingen
dc.subject.othermusic brandingen
dc.subject.othermachine learningen
dc.subject.othertaggingen
dc.subject.otherAudiode
dc.subject.otherMusikde
dc.subject.otherMarkenbildungde
dc.subject.othermaschinelles Lernende
dc.titleFinal Research Report on Auto-Tagging of Musicen
dc.typeReporten
dc.type.versionpublishedVersionen
tub.accessrights.dnbfreeen
tub.affiliationFak. 1 Geistes- und Bildungswissenschaften::Inst. Sprache und Kommunikation::FG Audiokommunikationde
tub.affiliation.facultyFak. 1 Geistes- und Bildungswissenschaftende
tub.affiliation.groupFG Audiokommunikationde
tub.affiliation.instituteInst. Sprache und Kommunikationde
tub.publisher.universityorinstitutionTechnische Universität Berlinen

Files

Original bundle
Now showing 1 - 1 of 1
Loading…
Thumbnail Image
Name:
D4.7_ABC_DJ_Final_Research_Report_on_Auto-Tagging_of_Music.pdf
Size:
10.77 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
4.9 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections