End-to-end music source separation: is it possible in the waveform domain?
October 29, 2018 ยท Entered Twilight ยท ๐ Interspeech
"Last commit was 6.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: LICENSE, README.md, audio, config.md, config_multi_instrument.json, config_singing_voice.json, datasets.py, environment.yml, img, layers.py, main.py, models.py, separate.py, sessions, util.py
Authors
Francesc Lluรญs, Jordi Pons, Xavier Serra
arXiv ID
1810.12187
Category
cs.SD: Sound
Cross-listed
cs.LG,
eess.AS
Citations
75
Venue
Interspeech
Repository
https://github.com/francesclluis/source-separation-wavenet
โญ 229
Last Checked
1 month ago
Abstract
Most of the currently successful source separation techniques use the magnitude spectrogram as input, and are therefore by default omitting part of the signal: the phase. To avoid omitting potentially useful information, we study the viability of using end-to-end models for music source separation --- which take into account all the information available in the raw audio signal, including the phase. Although during the last decades end-to-end music source separation has been considered almost unattainable, our results confirm that waveform-based models can perform similarly (if not better) than a spectrogram-based deep learning model. Namely: a Wavenet-based model we propose and Wave-U-Net can outperform DeepConvSep, a recent spectrogram-based deep learning model.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Sound
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
CNN Architectures for Large-Scale Audio Classification
R.I.P.
๐ป
Ghosted
Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation
R.I.P.
๐ป
Ghosted
Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification
R.I.P.
๐ป
Ghosted
WaveGlow: A Flow-based Generative Network for Speech Synthesis
R.I.P.
๐ป
Ghosted