-
Notifications
You must be signed in to change notification settings - Fork 365
Add support for shallow biasing of Whisper #1889
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
Do you have the python wheels for this branch? I would like to use it |
Yes, you can find the links to the wheels via the checks below. |
@minhthuc2502 would you mind taking a look at this PR? Would love to get this released! |
There might be a bug in this PR. When I use sequence bias with a model compiled in int8_float16 I get this error ValueError: expected storage to be of type float16, but is of type float32 which I dont get when I use sequence_bias as None. The exact same input |
Sorry I resolved. Does this increase latency? |
It does, unfortunately. Especially if the number of sequences to bias is large. |
Where does this latency increase come from? |
Attempting to fix #1789 by @zwycl