Integrating sound and context recognition for acoustic scene analysis http://soundscape.eecs.qmul.ac.uk Wed, 25 Sep 2019 14:40:47 +0000 en-US hourly 1 https://wordpress.org/?v=5.1.6 Attending Interspeech 2019 http://soundscape.eecs.qmul.ac.uk/2019/09/23/attending-interspeech-2019/ Mon, 23 Sep 2019 17:44:31 +0000 http://soundscape.eecs.qmul.ac.uk/?p=175 We had a great week at Interspeech 2019,  here Yogi presented our work on Joint Acoustic Scene Classification & Sound Event Detection entitled “Towards joint sound scene and polyphonic sound event recognition“. yogi_interspeech2019

]]>
DCASE 2019 Acceptance http://soundscape.eecs.qmul.ac.uk/2019/08/23/dcase-2019-acceptance/ Fri, 23 Aug 2019 18:07:32 +0000 http://soundscape.eecs.qmul.ac.uk/?p=183 screenshot-2019-09-23-at-19-09-40Work from Arjun for sound recognition has been accepted into the 4th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE 2019), entitled “Onsets, activity, and events: a multi-task approach for polyphonic sound event modelling“. We’ll see you in NYC soon! 

]]>
WASPAA 2019 Acceptance http://soundscape.eecs.qmul.ac.uk/2019/07/14/waspaa-2019-acceptance/ Sun, 14 Jul 2019 17:29:36 +0000 http://soundscape.eecs.qmul.ac.uk/?p=166 We are delighted to learn that we will be presenting our works on sound scene city classification and multi-task sound event and sound activity detection at the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA) in New York later this year.

For the code of the city classification work please see the github repo.

See you in New Paltz!

screenshot-2019-09-23-at-18-31-51

]]>
Interspeech 2019 acceptance http://soundscape.eecs.qmul.ac.uk/2019/06/17/interspeech-2019-acceptance/ Mon, 17 Jun 2019 17:25:15 +0000 http://soundscape.eecs.qmul.ac.uk/?p=164 screenshot-2019-09-23-at-18-31-39We are delighted to learn that we will be presenting our work on  joint audio scene classification and sound event detection at Interspeech in Graz later this year. For the data check out zenodo, for the code see github, and the camera ready paper will be online soon. 

See you Austria! 

]]>
Attending ICASSP http://soundscape.eecs.qmul.ac.uk/2019/05/12/attending-icassp/ Sun, 12 May 2019 17:34:42 +0000 http://soundscape.eecs.qmul.ac.uk/?p=171 Ines Nolasco and Emmanouil Benetos were both in action this week at ICASSP presenting their works in beehive states and subspectralnets.  ines_icassp_2019 emmanouil_icassp2019

]]>
Pre-print of new work released; sound scene city classification | audio geotagging http://soundscape.eecs.qmul.ac.uk/2019/05/02/pre-print-of-new-work-released-sound-scene-city-classification-audio-geotagging/ Thu, 02 May 2019 17:17:11 +0000 http://soundscape.eecs.qmul.ac.uk/?p=159 We are happy to release a pre-print of our latest work, conducted in partnership with our friends at Technical University of Tampere, Finland online today. screenshot-2019-09-23-at-18-19-30

Abstract: The majority of sound scene analysis work focuses on one of two clearly defined tasks: acoustic scene classification or sound event detection. Whilst this separation of tasks is useful for problem definition, they inherently ignore some subtleties of the real-world, in particular how humans vary in how they describe a scene. Some will describe the weather and features within it, others will use a holistic descriptor like `park’, and others still will use unique identifiers such as cities or names.
In this paper, we undertake the task of automatic city classification to ask whether we can recognize a city from a set of sound scenes? In this problem each city has recordings from multiple scenes.
We test a series of methods for this novel task and show that a simple convolutional neural network (CNN) can achieve accuracy of 50\%. This is less than the acoustic scene classification task baseline in the DCASE 2018 ASC challenge on the same data. A simple adaptation to the class labels of pairing city labels with grouped scenes, accuracy increases to 52\%, closer to the simpler scene classification task. Finally we also formulate the problem in a multi-task learning framework and achieve an accuracy of 56\%, outperforming the aforementioned approaches.

]]>
Pre-print of new work released; joint ASC and SED http://soundscape.eecs.qmul.ac.uk/2019/04/23/pre-print-of-new-work-released-joint-asc-and-sed/ Tue, 23 Apr 2019 17:12:49 +0000 http://soundscape.eecs.qmul.ac.uk/?p=156 A pre-print of our latest screenshot-2019-09-23-at-18-15-47work has been released online

Abstract: Acoustic Scene Classification (ASC) and Sound Event Detection (SED) are two separate tasks in the field of computational sound scene analysis. In this work, we present a new dataset with both sound scene and sound event labels and use this to demonstrate a novel method for jointly classifying sound scenes and recognizing sound events.
We show that by taking a joint approach, learning is more efficient and whilst improvements are still needed for sound event detection, SED results are robust in a dataset where the sample distribution is skewed towards sound scenes.

]]>
Our extensible taxonomy is Live! http://soundscape.eecs.qmul.ac.uk/2019/03/23/our-extensible-taxonomy-is-live/ Sat, 23 Mar 2019 17:08:46 +0000 http://soundscape.eecs.qmul.ac.uk/?p=152 screenshot-2019-09-23-at-18-20-11The great work by Ines Nolasco in automating our extensible taxonomy is now live! Check it out online

 

]]>
TUT Secondment for Post-Doc Yogi http://soundscape.eecs.qmul.ac.uk/2019/03/01/tut-secondment-for-post-doc-yogi/ Fri, 01 Mar 2019 17:00:52 +0000 http://soundscape.eecs.qmul.ac.uk/?p=149 Generously supported by ERC Grant Agreement 637422 EVERYSOUND, our Post-Doc Yogi is spending March 2019 at the Technical University of Tampere, Finland working with Annamaria Mesaros, Toni Heittola, and Tuomas Virtanen to enhance our partnership and investigate audio geotagging. tut-window-view

]]>
New dataset available! http://soundscape.eecs.qmul.ac.uk/2019/02/14/new-dataset-available/ Thu, 14 Feb 2019 16:53:36 +0000 http://soundscape.eecs.qmul.ac.uk/?p=146 We have produced a new dataset, with strong labels for both scene classification and sound event detection. It is available for download from here.  All source recordings are real-world recordings, the background (scene) sources come from prior DCASE challenges, and the foreground (event) sources are from both prior DCASE challenges and freesound.org. The mixtures are synthesised with SCAPER for on/offset labels also for the event classes. We hope you find it useful! 

]]>