Audio-Linguistic Embeddings for Spoken Sentences
ICASSP, 2019.
EI
Abstract:
We propose spoken sentence embeddings which capture both acoustic and linguistic content. While existing works operate at the character, phoneme, or word level, our method learns long-term dependencies by modeling speech at the sentence level. Formulated as an audio-linguistic multitask learning problem, our encoder-decoder model simultan...More
Code:
Data:
Full Text
Tags
Comments