Frame-level AnyBoost for LVCSR with the MMI criterion
Ryuki Tachibana, Takashi Fukuda, et al.
ASRU 2011
In this paper we explore discriminative language modeling (DLM) on highly optimized state-of-the-art large vocabulary Arabic broadcast speech recognition systems used for the Phase 5 DARPA GALE Evaluation. In particular, we study in detail a minimum Bayes risk (MBR) criterion for DLM. MBR training outperforms perceptron training. Interestingly, we found that our DLMs generalized to mismatched conditions, such as using a different acoustic model during testing. We also examine the interesting problem of unsupervised DLM training using a Bayes risk metric as a surrogate for word error rate (WER). In some experiments, we were able to obtain about half of the gain of the supervised DLM. © 2011 IEEE.
Ryuki Tachibana, Takashi Fukuda, et al.
ASRU 2011
Tara N. Sainath, David Nahamoo, et al.
ASRU 2011
Dzung Phan, Vinicius Lima
INFORMS 2023
Jehanzeb Mirza, Leonid Karlinsky, et al.
NeurIPS 2023