Abstract
Using solely the information retrieved by audio fingerprinting techniques, we propose methods to treat a possibly large dataset of user-generated audio content, that (1) enable the grouping of several audio files that contain a common audio excerpt (i.e. are relative to the same event), and (2) give information about how those files are correlated in terms of time and quality inside each event. Furthermore, we use supervised learning to detect incorrect matches that may arise from the audio fingerprinting algorithm itself, whilst ensuring our model learns with previous predictions. All the presented methods were further validated by user-generated recordings of several different concerts manually crawled from YouTube.
Original language | English |
---|---|
Title of host publication | 2017 IEEE 19th International Workshop on Multimedia Signal Processing, MMSP 2017 |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Pages | 1-6 |
Number of pages | 6 |
Volume | 2017-January |
ISBN (Electronic) | 9781509036493 |
DOIs | |
Publication status | Published - 27 Nov 2017 |
Event | 19th IEEE International Workshop on Multimedia Signal Processing, MMSP 2017 - Luton, United Kingdom Duration: 16 Oct 2017 → 18 Oct 2017 |
Conference
Conference | 19th IEEE International Workshop on Multimedia Signal Processing, MMSP 2017 |
---|---|
Country/Territory | United Kingdom |
City | Luton |
Period | 16/10/17 → 18/10/17 |
Keywords
- Audio fingerprinting
- Audio synchronisation
- Supervised learning
- User-generated content