-
Interspeech 20192019This paper proposes a Sub-band Convolutional Neural Network for spoken term classification. Convolutional neural networks(CNNs) have proven to be very effective in acoustic applications such as spoken term classification, keyword spotting, speaker identification, acoustic event detection, etc. Unlike applications in computer vision, the spatial invariance property of 2D convolutional kernels does not fit
-
AAAI 20192019Deep learning models have become state of the art for natural language processing (NLP) tasks, however deploying these models in production system poses significant memory constraints. Existing compression methods are either lossy or introduce significant latency. We propose a compression method that leverages low rank matrix factorization during training,to compress the word embedding layer which represents
-
AAAI 20192019User interaction with voice-powered agents generates large amounts of unlabeled utterances. In this paper, we explore techniques to efficiently transfer the knowledge from these unlabeled utterances to improve model performance on Spoken Language Understanding (SLU) tasks. We use Embeddings from Language Model (ELMo) to take advantage of unlabeled data by learning contextualized word representations. Additionally
-
Interspeech 20192019Any given classification problem can be modeled using multi-class or One-vs-All (OVA) architecture. An OVA system consists of as many OVA models as the number of classes, providing the advantage of asynchrony, where each OVA model can be re-trained independent of other models. This is particularly advantageous in settings where scalable model training is a consideration (for instance in an industrial environment
-
Interspeech 20192019This work aims at bootstrapping the acoustic model training with small amount of the human annotated speech data and large amount of the unlabeled speech data for automatic speech recognition.The technologies of the semi-supervised learning were investigated to select the automatically transcribed training samples.Two semi-supervised learning methods were pro-posed: one is the local-global uncertainty based
Related content
-
January 23, 2020New "Mad Libs" technique for replacing words in individual sentences is grounded in metric differential privacy.
-
January 21, 2020Self-learning system uses customers’ rephrased requests as implicit error signals.
-
January 16, 2020According to listener tests, whispers produced by a new machine learning model sound as natural as vocoded human whispers.
-
December 11, 2019Related data selection techniques yield benefits for both speech recognition and natural-language understanding.
-
November 06, 2019Today is the fifth anniversary of the launch of the Amazon Echo, so in a talk I gave yesterday at the Web Summit in Lisbon, I looked at how far Alexa has come and where we’re heading next.
-
October 28, 2019In a paper we’re presenting at this year’s Conference on Empirical Methods in Natural Language Processing, we describe experiments with a new data selection technique.