Abstract
This work proposes bag-of-features deep learning models for acoustic scene classi?cation (ASC) – identifying recording locations by analyzing background sound. We explore the effect on classi?cation accuracy of various front-end feature extraction techniques, ensembles of audio channels, and patch sizes from three kinds of spectrogram. The back-end process presents a two-stage learning model with a pre-trained CNN (preCNN) and a post-trained DNN (postDNN). Additionally, data augmentation using the mixup technique is investigated for both the pre-trained and post-trained processes, to improve classi?cation accuracy through increasing class boundary training conditions. Our experiments on the 2018 Challenge on Detection and Classi?cation of Acoustic Scenes and Events - Acoustic Scene Classi?cation (DCASE2018-ASC) subtask 1A and 1B signi?cantly outperform the DCASE2018 reference implementation and approach state-of-the-art performance for each task. Results reveal that the ensemble of multi-spectrogram features and data augmentation is bene?cial to performance.
| Originalsprache | Englisch |
|---|---|
| Titel | AES International Conference on Audio Forensics, 2019 |
| Seitenumfang | 12 |
| Publikationsstatus | Veröffentlicht - Juli 2019 |
| Veranstaltung | 2019 AES International Conference on Audio Forensics - Porto, Portugal Dauer: 18 Juni 2019 → 20 Juni 2019 |
Konferenz
| Konferenz | 2019 AES International Conference on Audio Forensics |
|---|---|
| Land/Gebiet | Portugal |
| Stadt | Porto |
| Zeitraum | 18/06/19 → 20/06/19 |
Research Field
- Ehemaliges Research Field - Data Science
Fingerprint
Untersuchen Sie die Forschungsthemen von „Bag-of-Features Models Based on C-DNN Network for Acoustic Scene Classification“. Zusammen bilden sie einen einzigartigen Fingerprint.Diese Publikation zitieren
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver