Ideas & improvements #4
Replies: 12 comments 19 replies
-
|
So, it is my opinion that the Wyoming protocol is best suited for communication between satellites and so forth. OpenAI is more "globally accepted" for AI backends. That's why I took the approach in this direction - in my lab, I stopped using wyoming_piper and replaced it with FastAPI-Kokoro (but there are Piper options as well, such as the speaches container) With this setup, I am able to share piper or kokoro with both Open WebUI and Home Assistant |
Beta Was this translation helpful? Give feedback.
-
|
Highly recommend replacing wyoming_piper with "Speaches". |
Beta Was this translation helpful? Give feedback.
-
|
I think your suggestion would be a good improvement. It's just not supported yet which is why I suggested the other method |
Beta Was this translation helpful? Give feedback.
-
|
Thanks @roryeckel I'll investigate the options, I think there's no way to integrate STT into HA without using Wyoming. I don't see such integration option. Same applies to TTS, so I think a solution that provides multiple APIs to access the service is a great idea. At the end of the day, all I'm trying to achieve is a single point of failure... I mean single instance of a service to save on VRAM usage as the GPU is also loaded up with Ollama and Frigate TensorRT. 😁 |
Beta Was this translation helpful? Give feedback.
-
This project is your "missing piece" to get a Wyoming server out of an OpenAI compliant endpoint. Speaches and FastAPI-Kokoro can both serve this type of OpenAI endpoint. Then, you would put the IP of this wyoming_openai container into Home Assistant to access it via the proxy I think I need to improve the documentation on how to use it with Home Assistant. I am open to contributions as well. Thanks |
Beta Was this translation helpful? Give feedback.
-
|
Trying it now, first little docker compose fix already in PR ;) |
Beta Was this translation helpful? Give feedback.
-
|
All is working initially |
Beta Was this translation helpful? Give feedback.
-
|
Mind moving this into a discussion? |
Beta Was this translation helpful? Give feedback.
-
|
I set my TTS_VOICES to "af af_bella af_sarah am_adam am_michael bf_emma bf_isabella bm_george bm_lewis af_nicole af_sky" which corresponds to the options I get in the speaches web-ui. Are you seeing the same? |
Beta Was this translation helpful? Give feedback.
-
|
Looks like I copied the voice list from Kokoro 1.0, but Speaches is on 0.19 at the moment. I'll downgrade / make something to fix that. |
Beta Was this translation helpful? Give feedback.
-
|
I am hesitant to add the unofficial /v1/audio/speech/voices endpoint as it may throw off the official OpenAI support. I hope OpenAI improves this voice listing capability in the future. Just added a comment with where to find the voices for now. |
Beta Was this translation helpful? Give feedback.
-
|
One extra thing I think would be beneficial is a pre-built container image as there's a Dockerfile already. Do you mind adding an action for building a container image and rolling out v0.1 (v.0.0.1) at some point? No rush with it, just wondering. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I came across this project when looking for a proxy that can expose Wyoming piper/whisper to an application that expects an OpenAI endpoint.
In other words, I'd like to connect open-webui to my whisper instance, which is hardware accelerated, rather than using OpenAI.
It seems this project does the opposite direction proxying.
Do you plan on adding functionality to provide the proxy in the opposite direction?
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions