Skip to content

videosdk-community/videosdk-deepgram-voice-agent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AI Copilot with VideoSDK & Deepgram

This project is an AI Voice Agent that uses VideoSDK for video conferencing, Deepgram for speech-to-text (STT), and OpenAI for language model (LLM) capabilities. The AI Copilot can join meetings, transcribe speech, and respond intelligently.

videoSDK + Deepgram Integration

Start with the project

git clone https://github.com/videosdk-community/videosdk-deepgram-voice-agent
cd videosdk-deepgram-voice-agent

Client Setup

  1. Navigate to client dir:
    cd client
  2. Make a copy of the environment configuration file:
    cp .env.example .env
  3. Set the VITE_APP_AUTH_TOKEN in the .env file with your VideoSDK auth token from app.videosdk.live.

Python Setup

  1. Configure the following environment variables in the .env file:

    ROOM_ID=...
    AUTH_TOKEN=...  # (app.videosdk.live)
    LANGUAGE=...
    
    DEEPGRAM_API_KEY=...  # (console.deepgram.com)
    LLM_API_KEY=...  # (platform.openai.com/api-keys)
  2. Create a virtual environment:

    python -m venv venv
  3. Activate the virtual environment:

    • On Unix or MacOS:
      source venv/bin/activate
    • On Windows:
      .\venv\Scripts\activate

Generate a Room ID

Generate a room ID on the client side and add it to the Python configuration. This can be done by running the client application and using the generated room ID in the .env file for the Python setup.

For more information, check out docs.videosdk.live.

About

AI-powered voice agent that joins meetings, transcribes speech in real-time using Deepgram STT, and responds intelligently.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published