Skip to content

Conversation

@DongjiGao
Copy link
Contributor

@DongjiGao DongjiGao commented Jan 14, 2020

This is for the ASR part of the speech translation project. It is similar to the current GALE Arabic recipe but with more data for acoustic and language model training.

@danpovey
Copy link
Contributor

Is any of the translation stuff somehow baked into this script, i.e. are there aspects that are specific translation? If not I might want you to call it s5c or whatever the next number is, but you should clarify how it differs. E.g. did we do something with subwords in s5b? And how do the WERs differ?

@DongjiGao
Copy link
Contributor Author

DongjiGao commented Jan 14, 2020

Is any of the translation stuff somehow baked into this script, i.e. are there aspects that are specific translation? If not I might want you to call it s5c or whatever the next number is, but you should clarify how it differs. E.g. did we do something with subwords in s5b? And how do the WERs differ?

There's no specific translation stuff. So I think I should call it s5d (s5b use graphemic lexicon and s5c is the subword version of s5b).
s5c (subword) is slightly worse than s5b.
s5d use more training data from GALE than s5b and the WER decrease from 16.47 to 14.89.
I also run a subword version of s5d, it is slightly worse than the word system after RNNLM-rescoring. So I did not check that in.

@danpovey
Copy link
Contributor

Looks OK to me. Do you think it needs testing before merge?

@DongjiGao
Copy link
Contributor Author

Looks OK to me. Do you think it needs testing before merge?

I have already done some tests. I think it's good to be merged.

@danpovey
Copy link
Contributor

Thanks a lot!! Merging.

@danpovey danpovey merged commit f196922 into kaldi-asr:master Jan 14, 2020
wonkyuml pushed a commit to wonkyuml/kaldi that referenced this pull request Jan 16, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

2 participants