Optimization of Large Scale Neural Networks for Speech Recognition
-
- Michiel Bacchiani
- Google Inc.
この論文をさがす
抄録
Recent years have shown a large scale adoption of speech recognition by the public, in particular around mobile devices. Simultaneously, the state of the art of speech recognition has improved significantly with the adoption of deep neural networks as the key technology for acoustic modeling. This approach is particularly effective when applied at scale. This talk will first provide an overview of how the Google speech group has grown over the years. Starting from a research effort, it has developed into an integral input modality of the Android operating system and is launched in more than 50 languages worldwide. The presentation will provide some historical perspective of that growth, discussing earlier products and outline the current mobile products and future directions of those. The success and adoption of the Google speech technology is evident by the decade of speech our systems receives each day. The talk will then focus on recent advances in training algorithms and infrastructure to train large neural networks on large data sets. Although not exclusively, the talk will provide a Google perspective to approaching this problem. It will describe our asynchronous parallel training infrastructure and how key algorithmic improvements fit within that framework. Particular attention will be paid to optimization using a sequence objective and recurrent network architectures (using Long Short Term Memory models).
収録刊行物
-
- 情報処理学会研究報告. SLP, 音声言語情報処理
-
情報処理学会研究報告. SLP, 音声言語情報処理 2014 (7), 1-, 2014-12-08
一般社団法人情報処理学会
- Tweet
キーワード
詳細情報 詳細情報について
-
- CRID
- 1571417127883926656
-
- NII論文ID
- 110009850956
-
- NII書誌ID
- AN10442647
-
- 本文言語コード
- en
-
- データソース種別
-
- CiNii Articles