Training neural audio classifiers with few data
October 24, 2018 Β· Entered Twilight Β· π IEEE International Conference on Acoustics, Speech, and Signal Processing
"Last commit was 7.0 years ago (β₯5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: README.md, aux, data, requirements.txt, src
Authors
Jordi Pons, Joan SerrΓ , Xavier Serra
arXiv ID
1810.10274
Category
cs.SD: Sound
Cross-listed
cs.AI,
cs.LG,
eess.AS
Citations
66
Venue
IEEE International Conference on Acoustics, Speech, and Signal Processing
Repository
https://github.com/jordipons/neural-classifiers-with-few-audio/
β 60
Last Checked
1 month ago
Abstract
We investigate supervised learning strategies that improve the training of neural network audio classifiers on small annotated collections. In particular, we study whether (i) a naive regularization of the solution space, (ii) prototypical networks, (iii) transfer learning, or (iv) their combination, can foster deep learning models to better leverage a small amount of training examples. To this end, we evaluate (i-iv) for the tasks of acoustic event recognition and acoustic scene classification, considering from 1 to 100 labeled examples per class. Results indicate that transfer learning is a powerful strategy in such scenarios, but prototypical networks show promising results when one does not count with external or validation data.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Sound
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
CNN Architectures for Large-Scale Audio Classification
R.I.P.
π»
Ghosted
Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation
R.I.P.
π»
Ghosted
Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification
R.I.P.
π»
Ghosted
WaveGlow: A Flow-based Generative Network for Speech Synthesis
R.I.P.
π»
Ghosted