This project implements a Retrieval-Augmented Generation (RAG) system built completely from scratch.
RAG-LLM combines the power of:
- Custom-built retrieval systems to fetch relevant context from a knowledge base
- Large Language Models for context-aware text generation
- Document Processing: Pipeline for ingesting, cleaning, and chunking documents
- Vector Embedding: Custom embedding system for semantic understanding of content
- Efficient Retrieval: Advanced retrieval mechanisms to find the most relevant information
- Context Integration: Seamless merging of retrieved context with user queries
- Response Generation: High-quality responses based on retrieved information
RAG-LLM/
βββ data/ # Document storage and vector databases
βββ models/ # Model implementations and configs
βββ src/
β βββ embeddings/ # Vector embedding components
β βββ retrieval/ # Retrieval system implementation
β βββ generation/ # Text generation components
β βββ stable_diff/ # Stable Diffusion API integration
β βββ utils/ # Helper functions and utilities
βββ api/ # API endpoints for using the system