-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
revert back to using PyAV instead of torch audio #961
Conversation
Will removing torch remove the supposed FFT speedup? |
Great to hear! I work with segments about 10 seconds long, so no benefit from batching. However, I am curious and possibly interested into bumping up to the latest commit because of this FFT speedup and especially interested in GPU acceleration |
these are the latest performance figures for the new FE @joiemoie
|
Do you take into account the overhead of moving long audio tensors to GPU when doing the above measurements on GPU? Update: I tried it and its around 1ms at most. My numpy variant is around ~30ms per 30sec as well. |
Another option could actually be to move the feature extraction layer to CTranslate2 as well, whisper.cpp has some implementations |
|
@MahmoudAshraf97 good job sir. However I have a problem - it does not respect the --hotwords option at all in standard interference and batching mode. I tested on different materials. In the "old" version before the batching mode was introduced, --hotwords worked very well. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good job
This PR reverts the torchaudio code that was added in #856 and removes
torchaudio
dependency but still keepstorch
the reason why
torch
wasn't removed in this PR is that feature extraction still depends on it and I didn't want to include the numpy feature extraction in this PR to keep it simple and to reduce the number of conflicts to be resolved with #936this should be merged before #936, after both are merged, a new PR will be created to completely remove the
torch
dependency