Skip to content

xsharov/enchanted-ollama-openrouter-proxy

 
 

Repository files navigation

Enchanted Proxy for OpenRouter

This repository is specifically made for use with the Enchanted project. The original author of this proxy is marknefedov.

Description

This repository provides a proxy server that emulates Ollama's REST API but forwards requests to OpenRouter. It uses the sashabaranov/go-openai library under the hood, with minimal code changes to keep the Ollama API calls the same. This allows you to use Ollama-compatible tooling and clients, but run your requests on OpenRouter-managed models. Currently, it is enough for usage with Jetbrains AI assistant.

Features

  • Model Filtering: You can provide a models-filter file in the same directory as the proxy. Each line in this file should contain a single model name. The proxy will only show models that match these entries. If the file doesn’t exist or is empty, no filtering is applied.

    Note: OpenRouter model names may sometimes include a vendor prefix, for example deepseek/deepseek-chat-v3-0324:free. To make sure filtering works correctly, remove the vendor part when adding the name to your models-filter file, e.g. deepseek-chat-v3-0324:free.

  • Ollama-like API: The server listens on 11434 and exposes endpoints similar to Ollama (e.g., /api/chat, /api/tags).

  • Model Listing: Fetch a list of available models from OpenRouter.

  • Model Details: Retrieve metadata about a specific model.

  • Streaming Chat: Forward streaming responses from OpenRouter in a chunked JSON format that is compatible with Ollama’s expectations.

Usage

You can provide your OpenRouter (OpenAI-compatible) API key through an environment variable or a command-line argument:

1. Environment Variable

export OPENAI_API_KEY="your-openrouter-api-key"
./ollama-proxy

2. Command Line Argument

./ollama-proxy "your-openrouter-api-key"

Once running, the proxy listens on port 11434. You can make requests to http://localhost:11434 with your Ollama-compatible tooling.

Installation

  1. Clone the Repository:

    git clone https://github.com/your-username/ollama-openrouter-proxy.git
    cd ollama-openrouter-proxy
    
  2. Install Dependencies:

    go mod tidy
    
  3. Build:

    go build -o ollama-proxy
    

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Go 97.3%
  • Dockerfile 2.7%