Using Langchain with speech input #15930
jaymon0703
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi
Given Multimodal models are becoming more and more prevalent, is there any work on the Langchain roadmap for "reading" speech input natively, or should we use speech-to-text libraries for the foreseeable future?
Example is, i want to verbally ask a question of my app which utilizes a QA Chain, or API Chain or SQL Agent, etc. Being able to speak instead of type will add to productivity, and doing so natively is possible with direct access to the Gemini API for example.
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions