Βι¶ΉΤΌΕΔ

Making Musical Mood Metadata

Automated metadata generation using music analysis

Published: 1 January 2012

Improving music library navigation and discovery using signal processing and machine learning techniques.

Project from 2012 - 2013

What we are doing

This project has ended. For an overview of all of our audio research partnerships, including those currently active, please see this page.

Making Musical Mood Metadata (M4) was a TSB funded collaborative project with Βι¶ΉΤΌΕΔ R&D, QMUL and I Like Music. Over eighteen months, the team developed new and innovative methods of extracting high-level metadata from music contact, including information about the mood and emotional content of tracks. Having access to this information makes it easier for content producers to find the music they are looking for.

Why it matters

The digital music revolution has seen an explosion in the size of music libraries. TV and radio producers now have a wider choice than ever before over which track to use in their programme, and finding the ideal track can often be a lengthy process.

Using the latest techniques in digital music analysis and machine learning, we can make it easier for people to find the track that is right for their situation. In broadcast, music tracks are often chosen for the emotion and mood that they convey. For this reason, the project focussed on allowing people to search for music by its mood.

Outcomes

By performing an analysis of I Like Music's large music database, we developed a model for representing the mood content of music as a series of numbers. This allows us to create systems which can interpret and compare the mood and emotion of music tracks.

We used the to perform an in-depth musical analysis of 128,000 tracks. This data was combined with the mood model to train and test machine learning systems. Through this work, we discovered are most critical in determining the mood of music.

We have already used this technology to enhance the Βι¶ΉΤΌΕΔ's online music library service with a recommendation function which gives producers a wider variety of music to use in their programmes. We are currently looking for ways to bring the full benefit of this technology to the wider public.

How it works

Machine learning is a method of training a computer to make connections between two pieces of information - in this case, music and mood. This is done by providing the computer with thousands of music tracks of various moods so that it can learn to distinguish between them.

By partnering with music suppliers I Like Music, the project gained access to over 100,000 music tracks that were each hand-labelled with detailed information about their genre, instrumentation and mood. We processed the audio and metadata using sophisticated algorithms and statistical techniques to find underlying structure and to help classify the audio content.

Project Team

  • Chris Baume (MEng CEng PhD)

    Chris Baume (MEng CEng PhD)

    Lead Research Engineer
  • David Marston (BEng)

    David Marston (BEng)

    Senior R&D Engineer
  • Becky Gregory-Clarke

    Becky Gregory-Clarke

    Research Technologist
  • Immersive and Interactive Content section

    IIC section is a group of around 25 researchers, investigating ways of capturing and creating new kinds of audio-visual content, with a particular focus on immersion and interactivity.

Rebuild Page

The page will automatically reload. You may need to reload again if the build takes longer than expected.

Useful links

Theme toggler

Select a theme and theme mode and click "Load theme" to load in your theme combination.

Theme:
Theme Mode: