Training Keyword Spotters with Limited and Synthesized Speech Data

31 Jan 2020  ·  James Lin, Kevin Kilgour, Dominik Roblek, Matthew Sharifi ·

With the rise of low power speech-enabled devices, there is a growing demand to quickly produce models for recognizing arbitrary sets of keywords. As with many machine learning tasks, one of the most challenging parts in the model creation process is obtaining a sufficient amount of training data. In this paper, we explore the effectiveness of synthesized speech data in training small, spoken term detection models of around 400k parameters. Instead of training such models directly on the audio or low level features such as MFCCs, we use a pre-trained speech embedding model trained to extract useful features for keyword spotting models. Using this speech embedding, we show that a model which detects 10 keywords when trained on only synthetic speech is equivalent to a model trained on over 500 real examples. We also show that a model without our speech embeddings would need to be trained on over 4000 real examples to reach the same accuracy.

PDF Abstract
No code implementations yet. Submit your code now

Datasets


Results from the Paper


Ranked #10 on Keyword Spotting on Google Speech Commands (Google Speech Commands V2 12 metric)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Keyword Spotting Google Speech Commands Embedding + Head Google Speech Commands V2 12 97.7 # 10
Keyword Spotting Google Speech Commands Head without Embedding Google Speech Commands V2 12 97.4 # 12

Methods


No methods listed for this paper. Add relevant methods here