Replies: 1 comment 1 reply
-
@nikhilro why dont you just use a prebuilt whisper.cpp instead of building it everytime you recieve an audio? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey @ggerganov,
Thanks for Whisper.cpp. It's indispensable for us and I'm excited by everything in your roadmap. Particularly, diarization.
We are making an AI Meeting Notes product and using Whisper.cpp to transcribe meeting audio. We first tried to run locally on user's machines. It worked but was unreliable—most people don't have Apple ARM chips. So, while the toolkit to transcribe locally gets built out, we're thinking of deploying Whisper.cpp in our servers to transcribe the audio.
This has been working well but is expensive. I feel there are ways to implement our setup a lot more efficiently. Hoping you can help.
Here is our current setup: audio comes in -> it's put on the queue -> some worker picks it up / create new worker to accommodate load -> clone Whisper.cpp repo, make, invoke main with audio, return output. We use Node so everything is done using
exec
statements. Usually, audio coming is 1 minute chunks but we have flexibility on this from FE.Two improvements we're thinking:
So, for you,
2
if possible and recommended?Beta Was this translation helpful? Give feedback.
All reactions