Bag-of-Features Models Based on C-DNN Network for Acoustic Scene Classification

Lam Pham, School Computing, The University of Kent, School Computing, The University of Kent, School Computing, The University of Kent, Lang Yue

Publikation: Beitrag in Buch oder TagungsbandVortrag mit Beitrag in TagungsbandBegutachtung

Abstract

This work proposes bag-of-features deep learning models for acoustic scene classi?cation (ASC) – identifying recording locations by analyzing background sound. We explore the effect on classi?cation accuracy of various front-end feature extraction techniques, ensembles of audio channels, and patch sizes from three kinds of spectrogram. The back-end process presents a two-stage learning model with a pre-trained CNN (preCNN) and a post-trained DNN (postDNN). Additionally, data augmentation using the mixup technique is investigated for both the pre-trained and post-trained processes, to improve classi?cation accuracy through increasing class boundary training conditions. Our experiments on the 2018 Challenge on Detection and Classi?cation of Acoustic Scenes and Events - Acoustic Scene Classi?cation (DCASE2018-ASC) subtask 1A and 1B signi?cantly outperform the DCASE2018 reference implementation and approach state-of-the-art performance for each task. Results reveal that the ensemble of multi-spectrogram features and data augmentation is bene?cial to performance.
OriginalspracheEnglisch
TitelAES International Conference on Audio Forensics, 2019
Seitenumfang12
PublikationsstatusVeröffentlicht - Juli 2019
Veranstaltung2019 AES International Conference on Audio Forensics - Porto, Portugal
Dauer: 18 Juni 201920 Juni 2019

Konferenz

Konferenz2019 AES International Conference on Audio Forensics
Land/GebietPortugal
StadtPorto
Zeitraum18/06/1920/06/19

Research Field

  • Ehemaliges Research Field - Data Science

Fingerprint

Untersuchen Sie die Forschungsthemen von „Bag-of-Features Models Based on C-DNN Network for Acoustic Scene Classification“. Zusammen bilden sie einen einzigartigen Fingerprint.

Diese Publikation zitieren