Skip to content

πŸš€ Phantom-AI-Interview – A fun experiment exploring how AI can silently assist in virtual interviews, coding challenges, and exams. Uses OCR, LLMs, and automation to extract text, generate responses, and operate without a visible UI. πŸ”₯

Notifications You must be signed in to change notification settings

Abhi5h3k/Phantom-AI-Interview

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸš€ Phantom-AI-Interview: Exploring AI Assistance in Technical Interviews

Abhishek LinkedIn Abhishek StackOverflow

With modern LLMs and AI-powered tools, getting AI assistance in coding interviews is easier than ever. This project explores how AI could seamlessly assist in real-time technical interviewsβ€”and what it means for hiring in the future.

πŸ“Œ Why This Project?


With the rise of large language models (LLMs) and AI-powered tools, how could AI assist in real-time coding interviews?

Phantom-AI-Interview is a proof-of-concept that explores AI's ability to listen, read, and respond to technical questions in a virtual setting. This project aims to understand the boundaries of AI assistance and its implications in hiring and technical assessments.



πŸ‘€ What’s in this Project?

βœ… On-Screen Text Extraction – Using EasyOCR to capture and process text from your screen.
βœ… AI-Assisted Response Automation – Simulating realistic typing for coding and text-based questions.
βœ… Phantom Mode – Operates entirely with hotkey shortcuts, eliminating the need for a visible GUI.
βœ… Active Listening Mode – AI stays alert for real-time interactions and responses.
βœ… Clipboard Jacking – Silently captures copied text, sends it to an LLM, and retrieves a response without detection.

πŸ’‘ Note: This project uses an open-source LLM on CPU. Performance and accuracy can be significantly improved with GPU acceleration, larger models, or by switching to GPT API with optimized prompts.


🌍 AI’s Impact on Tech Hiring

AI-powered tools like LLMs, automated assessments, and interview assistants are changing the way hiring works. Some key trends:

  • Faster Screening: AI can analyze candidates and recommend top matches.
  • Bias Concerns: AI-driven hiring tools risk reinforcing biases in decision-making.
  • Real-Time Assistance: Tools like Phantom-AI-Interview show how AI could provide live support in technical interviews.

πŸ”— Read more: How AI is disrupting tech hiring

πŸ“’ Ethical Considerations

This project is not intended to be used for unethical purposes, including interview fraud. It is a conceptual study on how AI could integrate into virtual hiring processes and technical assessments.


πŸ“Ί Demo (Click for YouTube Video)

πŸ€– All responses are played as audio output, which you can listen to through your connected headset 🎧. Notifications are enabled only for demonstration purposes; by default, they are turned off πŸ’‘.

🎧 Stereo Mode – AI Listens to Interviewer Directly

This mode allows PhantomAI to listen to the interviewer's voice directly from the system audio instead of the microphone. This ensures that the AI captures exactly what the interviewer is saying and generates responses accordingly.

Stereo Mode, Listen to Interviewer / Output Voice


πŸ“ On-Screen Text Extraction (OCR) – Extract Interview Questions

OCR mode is helpful when you need to extract a question from the interviewer's chat, a coding screen, or an online test. PhantomAI reads the text displayed on the screen and processes it for AI-assisted responses.

On-Screen Text Extraction


πŸ“‹ Clipboard Jacking – Quick & Silent AI Assistance

Clipboard Jacking is the fastest way to get help from the AI without any visible UI activity. If an interviewer shares a coding question in chat or an online IDE, and you're stuck on syntax or logic, simply copy the part you need help with and press a hotkey. PhantomAI will silently send it to the LLM and provide a response.

Clipboard Jacking


πŸ‘¨β€πŸ’» AI-Assisted Response

Active Listening Mode

Active Listening Mode


πŸ‘¨β€πŸ’» Technical Details

  • Live Audio Processing: Uses Vosk for real-time speech recognition.
  • File-Based Transcription: Uses Whisper for processing recorded audio files.
  • Customizable Hotkeys: Modify key bindings via config.ini to trigger AI commands silently.
  • Screen Capture: Uses EasyOCR to read input from screen.

πŸ’‘ System Requirements

  • Tested on: Windows 11
  • Processor: Intel(R) Core(TM) i7-10510U CPU @ 1.80GHz 2.30GHz
  • RAM: 32.0 GB (31.8 GB usable)

πŸš€Project Setup

Folder Structure

Your project directory should be structured as follows:

Phantom-AI-Interview/
│── whisper-bin-x64/          # Download [whisper-bin-x64](https://github.com/ggerganov/whisper.cpp/actions/runs/13716448084/) (Place at root)
    │── Some whisper files
    │── models
        │── ggml-base.en.bin  # Download ggml-base.en.bin [Whisper Models on Hugging Face](https://huggingface.co/ggerganov/whisper.cpp/tree/main)
│── vosk-model-en-in-0.5/     # Downloaded Vosk model (Place at root)
│── src/                      # Source code
│── config                    # Config folder containing Configuration file for hotkeys
│── requirements.txt          # Dependencies
│── run.py                    # Main script

Setup Instructions

Since some required files are large and cannot be pushed to GitHub, follow these steps to set up your environment:

image

  1. Download Whisper whisper-bin-x64

    Extract it to the root directory of the project

  2. Download ggml-base.en.bin Whisper Models on Hugging Face

  3. Download the Vosk Model Get the Vosk model (vosk-model-en-in-0.5)

πŸ“‚ Installation & Usage :

1. Install Python 3.11.0

Ensure you have Python 3.11.0 installed. You can download it from python.org.

2. Create a Virtual Environment

py -m venv phantomAI_venv

3. Activate the Virtual Environment

  • Windows:
    phantomAI_venv\Scripts\activate

4. Install Dependencies

pip install -r requirements.txt

5. Configure Hotkeys

Check config.ini for hotkey shortcuts. You can update them as per your preference.

6. Run PhantomAI

py run.py

πŸ“ Technologies Used

1. Vosk for Real-Time Speech Recognition

  • Uses vosk-model-en-in-0.5 for live transcription.
  • Find better models here: Vosk Models

2. Whisper for File-Based Transcription

3. Ollama for LLM Inference

  • Docker setup:
    docker-compose up -d
  • Pull the model:
    docker exec -it ollama ollama pull qwq
  • Find more models: Ollama Search

πŸš€ Future Improvements & Suggestions

If you want to enhance PhantomAI for better performance and a more seamless experience, consider these improvements:

  • Use GPU & Larger Models – For faster and more accurate responses, switch to a GPU-powered setup with a bigger local LLM.
  • Switch to GPT API – If local inference is slow, using the GPT API provides better performance and higher accuracy.
  • Better Prompt Engineering – Optimize response quality by designing improved prompt templates or dynamically selecting templates based on the question type.
  • Move to Containers – This project was a quick Windows-based concept. For better portability and stability, consider switching to Docker/Linux-based deployment.
  • Bundle into a Single Executable – Improve usability by packaging everything into a standalone .exe file for easy distribution.

🌍 Disclaimer

This project is purely for educational purposes and ethical AI research. It highlights the potential risks associated with silent AI in virtual environments but is not intended to be used for unethical activities.


πŸ“š License

Feel free to use and modify as needed.


πŸ” Disclaimer: This project is an experimental exploration of AI's capabilities in real-time assistance. It is meant to study AI-human interaction in technical settings and not to promote unfair practices in interviews or assessments.

About

πŸš€ Phantom-AI-Interview – A fun experiment exploring how AI can silently assist in virtual interviews, coding challenges, and exams. Uses OCR, LLMs, and automation to extract text, generate responses, and operate without a visible UI. πŸ”₯

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages