wav2vec - Unsupervised Pre-Training for Speech Recognition

INTERSPEECH, pp. 3465-3469, 2019.

Cited by: 152|Views73
EI

Abstract:

We explore unsupervised pre-training for speech recognition by learning representations of raw audio. wav2vec is trained on large amounts of unlabeled audio data and the resulting representations are then used to improve acoustic model training. We pre-train a simple multi-layer convolutional neural network optimized via a noise contras...More

Code:

Data:

Full Text
Bibtex
Your rating :
0

 

Tags
Comments