-
Notifications
You must be signed in to change notification settings - Fork 37
Conversation
|
This was trained on 960 hours, I assume? |
Yes, with full libri.
A known huge difference is espnet use warm_up scheduler, while I use Noam optimizer in snowfall.
Yes, I have reproduce espnet result and got detail training log, which will be used to diagnose my training process. |
|
You might want to check what data augmentation techniques and settings they are using and compare them with our setup. If we’re missing some techniques in Lhotse we can add them. |
|
So I guess this is ready to merge? |
|
Maybe @csukuangfj is going to review this afternoon. |
snowfall/models/transformer.py
Outdated
| ys_out_pad = pad_list(ys_in, -1) | ||
|
|
||
| else: | ||
| raise VAlueError("Invalid input for decoder self attetion") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
VAlueError -> ValueError
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
fixed.
|
+2 |
|
Thanks! Merging |
This pr release a snowfall trained model together with related decode code.
Wer on test-clean is lower than previously trained model with snowfall, detailed comparison as following:
A thing worth to mention is that: current no-rescore result(3.97 on test-clean) is got WITHOUT a 3-gram.
Maybe the result will get lower with composing currnet ctc_topo with a 3-gram fst (I am working on this).
Another baseline of this model is an espnet released model, detailed comparison as following:
num_paths = 100 when doing n-best rescoring of row 2;
result of row 2 is got using similar techiniques used in #201, by loading espnet released model with snowfall code.
Conclusions:
(related training code will be submited soon in this week; make a promise here to force me to do this quickly).