ASR training from scratch #2967
Unanswered
nata-kostina
asked this question in
Q&A
Replies: 1 comment 2 replies
-
|
Hi, could you open an issue instead? This recipe is very heavy on the augmentation side, mostly used to showcase it rather than train a really good model. That's also maybe why it's using a CRDNN which is an old model. If you open an issue, we'll take a look. Many thanks! |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am trying to retrain an ASR model on LibriSpeech from scratch.
We are training on an A100 GPU, and even with a low batch size of 8, the model is consuming a very large amount of memory. Could you advise why this might be happening?
We are observing that the training seems to not actually progress — the train loss doesn’t change, and CER and WER remain very high. Any insights on what could be causing this?
Thank you.
Below I share ASR configuration:
Beta Was this translation helpful? Give feedback.
All reactions