Love Me, Love Me, Say (And Write!) That You Love Me: Enriching The
Total Page:16
File Type:pdf, Size:1020Kb
Love Me, Love Me, Say (and Write!) that You Love Me: Enriching the WASABI Song Corpus with Lyrics Annotations Michael Fell, Elena Cabrio, Elmahdi Korfed, Michel Buffa, Fabien Gandon To cite this version: Michael Fell, Elena Cabrio, Elmahdi Korfed, Michel Buffa, Fabien Gandon. Love Me, Love Me, Say (and Write!) that You Love Me: Enriching the WASABI Song Corpus with Lyrics Annotations. LREC 2020 - 12th edition of the Language Resources and Evaluation Conference, May 2020, Marseille, France. hal-03133188 HAL Id: hal-03133188 https://hal.archives-ouvertes.fr/hal-03133188 Submitted on 5 Feb 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020), pages 2138–2147 Marseille, 11–16 May 2020 c European Language Resources Association (ELRA), licensed under CC-BY-NC Love Me, Love Me, Say (and Write!) that You Love Me: Enriching the WASABI Song Corpus with Lyrics Annotations Michael Fell, Elena Cabrio, Elmahdi Korfed, Michel Buffa and Fabien Gandon Universite´ Coteˆ d’Azur, CNRS, Inria, I3S, France fmichael.fell, elena.cabrio, elmahdi.korfed, [email protected], [email protected] Abstract We present the WASABI Song Corpus, a large corpus of songs enriched with metadata extracted from music databases on the Web, and resulting from the processing of song lyrics and from audio analysis. More specifically, given that lyrics encode an important part of the semantics of a song, we focus here on the description of the methods we proposed to extract relevant information from the lyrics, such as their structure segmentation, their topics, the explicitness of the lyrics content, the salient passages of a song and the emotions conveyed. The creation of the resource is still ongoing: so far, the corpus contains 1.73M songs with lyrics (1.41M unique lyrics) annotated at different levels with the output of the above mentioned methods. Such corpus labels and the provided methods can be exploited by music search engines and music professionals (e.g. journalists, radio presenters) to better handle large collections of lyrics, allowing an intelligent browsing, categorization and recommendation of songs. We provide the files of the current version of the WASABI Song Corpus, the models we have built on it as well as updates here: https://github.com/micbuffa/WasabiDataset. Keywords: Corpus (Creation, Annotation, etc.), Information Extraction, Information Retrieval, Music and Song Lyrics 1. Introduction An analysis of the correlations among the above mentioned annotation layers reveals interesting insights about the song Let’s imagine the following scenario: following David corpus. For instance, we demonstrate the change in corpus Bowie’s death, a journalist plans to prepare a radio show annotations diachronically: we show that certain topics be- about the artist’s musical career to acknowledge his quali- come more important over time and others are diminished. ties. To discuss the topic from different angles, she needs We also analyze such changes in explicit lyrics content and to have at her disposal the artist biographical information expressed emotion. to know the history of his career, the song lyrics to know The paper is organized as follows. Section 2. introduces the what he was singing about, his musical style, the emotions WASABI Song Corpus and the metadata initially extracted his songs were conveying, live recordings and interviews. from music databases on the Web. Section 3. describes Similarly, streaming professionals such as Deezer, Spotify, the segmentation method we applied to decompose lyrics Pandora or Apple Music aim at enriching music listening in their building blocks in the corpus. Section 4. explains with artists’ information, to offer suggestions for listening the method used to summarize song lyrics, leveraging their to other songs/albums from the same or similar artists, or structural properties. Section 5. reports on the annotations automatically determining the emotion felt when listening resulting from the explicit content classifier, while Section to a track to propose coherent playlists to the user. To sup- 6. describes how information on the emotions are extracted port such scenarios, the need for rich and accurate musi- from the lyrics. Section 7. describes the topic modeling cal knowledge bases and tools to explore and exploit this algorithm to label each lyrics with the top 10 words, while knowledge becomes evident. Section 8. examines the changes in the annotations over In this paper, we present the WASABI Song Corpus, a large the course of time. Section 9. reports on similar existing corpus of songs (2.10M songs, 1.73M with lyrics) enriched resources, while Section 10. concludes the paper. with metadata extracted from music databases on the Web, and resulting from the processing of song lyrics and from 2. The WASABI Song Corpus audio analysis. The corpus contains songs in 36 different 1 languages, even if the vast majority are in English. As for In the context of the WASABI research project that started the songs genres, the most common ones are Rock, Pop, in 2017, a two million song database has been built, with Country and Hip Hop. metadata on 77k artists, 208k albums, and 2.10M songs More specifically, while an overview of the goals of the (Meseguer-Brocal et al., 2017). The metadata has been i) WASABI project supporting the dataset creation and the aggregated, merged and curated from different data sources description of a preliminary version of the dataset can be on the Web, and ii) enriched by pre-computed or on- found in (Meseguer-Brocal et al., 2017), this paper focuses demand analyses of the lyrics and audio data. on the description of the methods we proposed to annotate We have performed various levels of analysis, and inter- relevant information in the song lyrics. Given that lyrics active Web Audio applications have been built on top of encode an important part of the semantics of a song, we the output. For example, the TimeSide analysis and an- propose to label the WASABI dataset lyrics with their struc- notation framework have been linked (Fillon et al., 2014) ture segmentation, the explicitness of the lyrics content, the to make on-demand audio analysis possible. In connection salient passages of a song, the addressed topics and the emotions conveyed. 1http://wasabihome.i3s.unice.fr/ 2138 Figure 2: The datasources connected to the WASABI Song Corpus. tify, Amazon, AllMusic, GoHear, YouTube. Figure 2 il- lustrates7 all the data sources we have used to create the WASABI Song Corpus. We have also aligned the WASABI Song Corpus with the publicly available LastFM dataset8, Figure 1: The WASABI Interactive Navigator. resulting in 327k tracks in our corpus having a LastFM id. As of today, the corpus contains 1.73M songs with lyrics with the FAST project2, an offline chord analysis of 442k (1.41M unique lyrics). 73k songs have at least an ab- songs has been performed, and both an online enhanced stract on DBpedia, and 11k have been identified as “clas- audio player (Pauwels and Sandler, 2019) and chord search sic songs” (they have been number one, or got a Grammy engine (Pauwels et al., 2018) have been built around it. A award, or have lots of cover versions). About 2k songs have rich set of Web Audio applications and plugins has been a multi-track audio version, and on-demand source sepa- proposed (Buffa and Lebrun, 2017a; Buffa and Lebrun, ration using open-unmix (Stoter¨ et al., 2019) or Spleeter 2017b; Buffa et al., 2018), that allow, for example, songs to (Hennequin et al., 2019) is provided as a TimeSide plugin. be played along with sounds similar to those used by artists. Several Natural Language Processing methods have been All these metadata, computational analyses and Web Au- applied to the lyrics of the songs included in the WASABI dio applications have now been gathered in one easy-to-use Song Corpus, as well as various analyses of the extracted web interface, the WASABI Interactive Navigator3, illus- information have been carried out. After providing some trated4 in Figure 1. statistics on the WASABI corpus, the rest of the article de- We have started building the WASABI Song Corpus by scribes the different annotations we added to the lyrics of collecting for each artist the complete discography, band the songs in the dataset. Based on the research we have con- members with their instruments, time line, equipment they ducted, the following lyrics annotations are added: lyrical use, and so on. For each song we collected its lyrics from structure (Section 3.), summarization (Section 4.), explicit LyricWiki5, the synchronized lyrics when available6, the lyrics (Section 5.), emotion in lyrics (Section 6.) and topics DBpedia abstracts and the categories the song belongs to, in lyrics (Section 7.). e.g. genre, label, writer, release date, awards, produc- 2.1. Statistics on the WASABI Song Corpus ers, artist and band members, the stereo audio track from Deezer, the unmixed audio tracks of the song, its ISRC, This section summarizes key statistics on the corpus, such bpm and duration. as the language and genre distributions, the songs coverage We matched the song ids from the WASABI Song Cor- in terms of publication years, and then gives the technical pus with the ids from MusicBrainz, iTunes, Discogs, Spo- details on its accessibility.