On the Automatic Identification of Music for Common Activities

Research output: Chapter in Book/Conference proceedings/Edited volumeConference contributionScientificpeer-review

7 Citations (Scopus)

Abstract

In this paper, we address the challenge of identifying music suitable to accompany typical daily activities. We first derive a list of common activities by analyzing social media data. Then, an automatic approach is proposed to find music for these activities. Our approach is inspired by our experimentally acquired findings (a) that genre and instrument information, i.e., as appearing in the textual metadata, are not sufficient to distinguish music appropriate for different types of activities, and (b) that existing content-based approaches in the music information retrieval community do not overcome this insufficiency. The main contributions of our work are (a) our analysis of the properties of activity-related music that inspire our use of novel high-level features, e.g., drop-like events, and (b) our approach's novel method of extracting and combining low-level features, and, in particular, the joint optimization of the time window for feature aggregation and the number of features to be used. The effectiveness of the approach method is demonstrated in a comprehensive experimental study including failure analysis.

Original languageEnglish
Title of host publicationProceedings of the 2017 ACM International Conference on Multimedia Retrieval
Place of PublicationNew York
PublisherAssociation for Computing Machinery (ACM)
Pages192-200
Number of pages9
ISBN (Electronic)978-1-4503-4701-3
DOIs
Publication statusPublished - 2017
EventICMR 2017: ACM International Conference on Multimedia Retrieval - Bucharest, Romania
Duration: 6 Jun 20179 Jun 2017
http://www.icmr2017.ro/

Conference

ConferenceICMR 2017
CountryRomania
CityBucharest
Period6/06/179/06/17
Internet address

Keywords

  • Activity
  • Music recommendation
  • Relax music
  • Study music
  • Workout music

Fingerprint Dive into the research topics of 'On the Automatic Identification of Music for Common Activities'. Together they form a unique fingerprint.

Cite this