-
Notifications
You must be signed in to change notification settings - Fork 3.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error when files with same size are in first batch #3327
Comments
Have you tried the documented workaround from #3088 ? |
I checked the thread. From what I could understand, the workaround is experimenting with different versions of |
No, the workaround is an env variable until the proper fix is released by TensorFlow: #3088 (comment) |
I tried this and it worked. Thanks a lot 👍 |
Thanks. TensorFlow team merged the r1.15 PR a few hours ago, so hopefully they are making 1.15.4 and we will be able to bump our dependency |
$ pip install tensorflow-gpu==1.15.2
)python DeepSpeech.py --train_files data/train.csv --train_batch_size 2 --train_cudnn
While training the model using the command shown above, I am getting a strange error that comes only in some particular cases.
The error message
To reproduce the issue, I disabled the sorting of input samples so that I can pinpoint the exact batch which is causing the issue. Here are my findings:
Case 1
Contents of train.csv
The first batch has files that are all of the same sizes. Since the batch size is 2, the model will receive 1.wav and 2.wav in the first step. This configuration of train.csv throws the error above.
Case 2
Now let's rearrange the contents on train.csv to
The first batch will have samples of different sizes (3.wav and 1.wav). Surprisingly, this case does not throw an error. The model trains without any issues.
Can anyone help me in understanding what might be causing this issue? Or is this a bug?
The text was updated successfully, but these errors were encountered: