Feature-rich continuous language models for speech recognition

Piotr Mirowski, Sumit Chopra, Suhrid Balakrishnan, Srinivas Bangalore

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

State-of-the-art probabilistic models of text such as n-grams require an exponential number of examples as the size of the context grows, a problem that is due to the discrete word representation. We propose to solve this problem by learning a continuous-valued and lowdimensional mapping of words, and base our predictions for the probabilities of the target word on non-linear dynamics of the latent space representation of the words in context window. We build on neural networks-based language models; by expressing them as energy-based models, we can further enrich the models with additional inputs such as part-of-speech tags, topic information and graphs of word similarity. We demonstrate a significantly lower perplexity on different text corpora, as well as improved word accuracy rate on speech recognition tasks, as compared to Kneser-Ney backoff n-gram-based language models.

Original languageEnglish (US)
Title of host publication2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Proceedings
Pages241-246
Number of pages6
DOIs
StatePublished - 2010
Event2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Berkeley, CA, United States
Duration: Dec 12 2010Dec 15 2010

Publication series

Name2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Proceedings

Conference

Conference2010 IEEE Workshop on Spoken Language Technology, SLT 2010
Country/TerritoryUnited States
CityBerkeley, CA
Period12/12/1012/15/10

Keywords

  • Natural language
  • Neural networks
  • Probability
  • Speech recognition

ASJC Scopus subject areas

  • Language and Linguistics

Fingerprint

Dive into the research topics of 'Feature-rich continuous language models for speech recognition'. Together they form a unique fingerprint.

Cite this