use of external LM to improve transcriptions #940
-
Could you explain how to used an external LM, N-gram or RNN to improve the trnascriptions ? |
Beta Was this translation helpful? Give feedback.
Replies: 6 comments
-
We provide several decoding methods with external language models. Please have a look at https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/RESULTS.md#pruned_transducer_stateless7-zipformer. |
Beta Was this translation helpful? Give feedback.
-
OK, but why do you use only a BPE-based N-gram LM instead of a word-level N-gram with a lexicon.txt of the type word -> BPE tokens ? |
Beta Was this translation helpful? Give feedback.
-
We support word level N-gram as well, please have a look at here |
Beta Was this translation helpful? Give feedback.
-
Well I tried that a word-level ARPA LM with a lexicon.txt of the type word -> BPE tokens using the BPE models used for training. |
Beta Was this translation helpful? Give feedback.
-
Could you please share the decoding command you are using? BTW, using neural network LMs usually gives a larger performance improvement. If you want to have optimal performance, we recommend you use |
Beta Was this translation helpful? Give feedback.
-
I think the problem was that the script generating LG was using a default path for the ARPA LM, whereas I thought that it would looked for it in the lang_bpe_XXX directory... |
Beta Was this translation helpful? Give feedback.
We provide several decoding methods with external language models. Please have a look at https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/RESULTS.md#pruned_transducer_stateless7-zipformer.