Skip to content

neuralnetworth/aria

 
 

Repository files navigation

Aria

Meet Aria. A local and uncensored AI entity.

Aria

aria_demo.mp4
aria_demo_1.mp4

Installation

Prerequisites

  • Python 3.10 or higher
  • NVIDIA GPU with CUDA support (required for flash-attn)
  • System dependencies:
    • Ubuntu/Debian: sudo apt install python3.12-dev portaudio19-dev libopus-dev
    • Arch Linux: sudo pacman -S python portaudio opus
    • macOS: brew install portaudio opus

Method 1 - Using uv (Recommended)

uv is a fast Python package manager that provides better dependency resolution and faster installations.

# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh

# Clone the repo
git clone https://github.com/neuralnetwork/aria.git
cd aria

# Install dependencies
uv sync

Method 2 - Using pip (Traditional)

# Install from pyproject.toml
pip install -e .
# Note: flash-attn requires CUDA Toolkit and may need manual installation

(Tested on Arch Linux + NVIDIA GPUs with Python 3.12)

Syncing Across Multiple Computers

When using uv, you can easily sync your environment across multiple computers:

# On your first computer (after installation)
git add uv.lock
git commit -m "Add uv.lock for dependency sync"
git push

# On your other computers
git pull
uv sync

The uv.lock file ensures all your computers use exactly the same package versions.

Usage

First run will take a while to download all the required models.
You may edit the default config for your device or use case (change model, specify devices, etc...)
If you have the resources, strongly recommended to use bigger model and/or bigger quant method.

Default Model

Aria now uses Qwen2.5-14B-Instruct-1M-abliterated (Q6_K quantization) by default:

  • Model size: ~12.2 GB
  • Recommended VRAM: 16GB+ (works well on RTX 4080 Super)
  • Downloads automatically on first run to HuggingFace cache
  • Abliterated version (uncensored) for unrestricted responses

Windows Users with WSL

If you're developing with WSL but running Aria on Windows:

  1. Edit code in WSL using your preferred tools
  2. Run Aria directly on Windows for best GPU performance
  3. Use Windows terminal: uv run python main.py
  4. Flash-attn builds properly on native Windows with CUDA

Non server/client mode:

python main.py

Server and Client Mode

server machine:

python server.py 

client machine (edit client target ip in the config):

python client.py

Upcoming Features

  • Android client
  • Raspberry Pi client
  • Ollama support (currently uses GGUF format directly)

Documentation

Work in progress...

Contributions

🌟 We'd love your contribution! Please submit your changes via pull request to join in the fun! 🚀

Disclaimer

Aria is a powerful AI entity designed for local use. Users are advised to exercise caution and responsibility when interacting with Aria, as its capabilities may have unintended consequences if used improperly or without careful consideration.

By engaging with Aria, you understand and agree that the suggestions and responses provided are for informational purposes only, and should be used with caution and discretion.

We cannot be held responsible for any actions, decisions, or outcomes resulting from the use of Aria. We explicitly disclaim liability for any direct, indirect, incidental, consequential, or punitive damages arising from reliance on Aria's responses.

We encourage users to exercise discernment, judgment, and thorough consideration when utilizing information from Aria. Your use of this service constitutes acceptance of these disclaimers and limitations.

Should you have any doubts regarding the accuracy or suitability of Aria's responses, we advise consulting with qualified professionals or experts in the relevant field.

Acknowledgments

License Information

❗ Important Note:

While this project is licensed under GNU AGPLv3, the usage of some of the components it depends on might not and they will be listed below:

TTS MODEL

opuslib

Llama

About

A local and uncensored AI entity.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%