Getting: OSError: [Errno -9996] Invalid input device (no default output device) #10
Replies: 7 comments
-
Ya the docker setup is bulky and running the app nativly is best otherwise you might need to do some troubleshooting on some systems. So based on the logs it doesn't seem to be able to pick up your microphone, in windows check mircophone privacy settings to make sure it is toggled on, in privacy and security / microphone. Also make sure apps can access mic, and the app is either going to be python or wsl so make sure those are toggled on. Also make sure your browser microphone settings are good. |
Beta Was this translation helpful? Give feedback.
-
Thanks for the quick reply! I'll try to check the privacy settings and maybe give it a go to run it natively but like I said I'm a complete newb with no knowledge regarding these things. I appreciate you taking the time to respond! |
Beta Was this translation helpful? Give feedback.
-
I managed to manually to install it and I got it to work (with a lot of trail and error). One question, I'm running the Ollama model locally and I noticed the response time takes quite a while (i'd say Also the chatgpt response goes over the 250character limit. Is there a way to give some parameters so it will keep it answers to the max limit. |
Beta Was this translation helpful? Give feedback.
-
yes ollama on first response has to load and once model in memory should run faster and then if you are running it on cpu only takes longer, if you are running on fast gpu runs faster, I have rtx 4090 and using llama3 takes around 4 secs once it is loaded. Openai 4o-mini and good internet connection seems to work quick. If you mean the chatgpt response as in ollama response and not openai, different models follow the prompt better like I noticed llama3.2 is more chatty and doesn't follow a limit as llama3.1 does. So you have downloaded the ollama llama3.1 7b and try that and see if it works better than the 3.2. I have noticed mistal is really bad and doesnt follow as it should. You can also add a token limit to the ollama payload request in app.py ( starts at line 416) just change num_predict from -2 to 250, adjust as needed.
The hard coded prompt gives the the model a limit to not go over, sometimes it doesn't listen to it, the audio also has a limit of characters, if I remember right xtts cuts off after a certain limit as it doesn't do good with longer responses, if elevenlabs or openai it is a bit longer. It all depends on the setup ( providers and models being used) and what type of hardware you are running on. You should be able to make some speed changes like in app.py changing this ( I haven't tried this on this project but tiny.en works on other projects I have used)
to this
|
Beta Was this translation helpful? Give feedback.
-
I actually just tried modifying the num_predict and also setting a max_token perimeter and it doesn't work, I think the only way to limit the model in ollama is by adjusting the models configuration file on the host, The default context size for Ollama is 2048 tokens so would need to adjust for each model as needed. You can also inform each character in the characters/(character)/(character).txt and change from
to a lower number and see if the model follows it's prompt instructions better. So instead of 500 change to 250, ect. |
Beta Was this translation helpful? Give feedback.
-
@gilamonyet just did a big update, no longer need to manually download xtts files now, added docker support for cpu. Easier setup with docker compose. If volume mapping is correct in docker on start should use pyaudio without issues now. |
Beta Was this translation helpful? Give feedback.
-
Thanks for the update! I might give it a go in the next couple of days. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
First of all, I'll be straight and honest - I'm a complete newb/beginner regarding Docker + I'm not a developer by any means. I stumbled across this Github repository because I was looking for ways to have a voice chat locally with an AI to mimic NPC conversations while playing video games.
After some trials and errors, I managed to build the image via Docker.Desktop, and I'm able to see the UI via localhost. However, when I try to talk to it, the logs show me the following error, and I have no clue how to fix it.
I tried asking ChatGPT (I know, don't blame me please) and it mentioned something about Docker not being able to find (or access) my local output device. I've checked my settings in Windows and my headphones are being marked as the output device and my microphone is being marked as the default input device.
I hope someone can help!
See the full log;
Beta Was this translation helpful? Give feedback.
All reactions