Skip to content

neallawson/llm-inference-hub

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

4 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

AI WebHub: Self-Hosted LLM Management Platform πŸš€

GitHub stars GitHub forks GitHub watchers GitHub repo size GitHub language count GitHub top language GitHub last commit Discord

AI WebHub is a high-performance, fully self-hosted AI platform for managing, running, and interacting with LLMs (Large Language Models). Designed to work entirely offline, it supports Ollama, OpenAI-compatible APIs, and provides a built-in inference engine for RAG (Retrieval Augmented Generation), enabling enterprise-grade AI deployment.


Key Highlights 🌟

  • Seamless Deployment: Docker, Docker Compose, Kubernetes (Helm/Kustomize), and native Python installation for flexible deployment.
  • Multi-LLM Support: Integrate Ollama, OpenAI APIs, LMStudio, GroqCloud, Mistral, OpenRouter, and more.
  • Offline & Secure: Fully functional offline mode; supports granular RBAC and SCIM 2.0 provisioning.
  • Modern UI/UX: Responsive, PWA-ready Web UI with Markdown, LaTeX, and live multimedia support.
  • Python Function Integration: Bring Your Own Function (BYOF) for custom Python tools inside the LLM workspace.
  • RAG & Web Integration: Local document RAG, web searches (SearXNG, Google, DuckDuckGo, Bing), and live web content injection.
  • Model Builder & Management: Create, import, and manage Ollama models via a clean Web UI.
  • Image Generation: Integrates AUTOMATIC1111, ComfyUI (local), or DALL-E for rich visual AI content.
  • Pipelines & Plugins: Extend AI WebHub with Python plugins and custom pipelines.
  • Multilingual Support: Full i18n support for global accessibility.

Installation Guide πŸš€

Python Installation

Requires Python 3.11+:

pip install ai-webhub

Start the server:

ai-webhub serve

Access Web UI: http://localhost:8080


Docker Quick Start 🐳

Default Setup with Local Ollama

docker run -d -p 3000:8080 \
  --add-host=host.docker.internal:host-gateway \
  -v ai-webhub:/app/backend/data \
  --name ai-webhub \
  --restart always \
  ghcr.io/karianne50m/ai-webhub:main

GPU-Enabled Ollama Setup

docker run -d -p 3000:8080 --gpus all \
  -v ollama:/root/.ollama \
  -v ai-webhub:/app/backend/data \
  --name ai-webhub \
  --restart always \
  ghcr.io/karianne50m/ai-webhub:ollama

OpenAI-Only API Usage

docker run -d -p 3000:8080 \
  -e OPENAI_API_KEY=your_secret_key \
  -v ai-webhub:/app/backend/data \
  --name ai-webhub \
  --restart always \
  ghcr.io/karianne50m/ai-webhub:main

Offline Mode

export HF_HUB_OFFLINE=1

Troubleshooting & Support

  • Server Connection Errors: Use --network=host if container cannot reach Ollama on localhost.
  • Docker Updates: Use Watchtower for automatic container updates.
  • Join our Discord community for real-time support.

Advanced Features

  • RAG Document Integration: Load documents and query via #doc_name.
  • Web & Image Content Injection: Integrate live web content and images dynamically.
  • Multi-Model Conversations: Simultaneously interact with multiple LLMs.
  • Role-Based Access Control (RBAC): Restrict model creation and access to specific users.
  • Pipelines & Plugins: Automate workflows, rate limiting, monitoring, and real-time translation.

Roadmap & Contribution

Check the roadmap to explore upcoming features. Contributions are welcome β€” report issues or submit PRs on GitHub.


License πŸ“œ

This repository uses a BSD-3-Clause style license with an additional clause preserving the AI WebHub branding. Full license details in LICENSE.


Analytics & Popularity

Star History Chart

Join the Community πŸ’¬

Connect, collaborate, and contribute via Discord.

About

No description, website, or topics provided.

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published