-
Notifications
You must be signed in to change notification settings - Fork 6
Added the ability to call openai compatible api #14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
go test -bench=. -timeout=0
go test -bench=. -timeout=0
time=2025-08-29T10:22:58.423-04:00 level=INFO msg="Inserting file" rag=light path=docs/christmascarol.txt
time=2025-08-29T10:22:58.524-04:00 level=INFO msg="Upserting sources" rag=light package=golightrag function=Insert count=8
time=2025-08-29T10:22:58.533-04:00 level=INFO msg="Extracting entities" rag=light package=golightrag function=Insert count=8
time=2025-08-29T10:23:25.934-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=1 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:23:52.162-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=1 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:23:58.534-04:00 level=WARN msg="Retry extract" rag=light package=golightrag function=Insert retry=1 error="failed to call LLM: error sending request: Post \"http://localhost:1234/v1/chat/completions\": context deadline exceeded"
time=2025-08-29T10:23:58.534-04:00 level=WARN msg="Retry extract" rag=light package=golightrag function=Insert retry=1 error="failed to call LLM: error sending request: Post \"http://localhost:1234/v1/chat/completions\": context deadline exceeded"
time=2025-08-29T10:23:58.534-04:00 level=WARN msg="Retry extract" rag=light package=golightrag function=Insert retry=1 error="failed to call LLM: error sending request: Post \"http://localhost:1234/v1/chat/completions\": context deadline exceeded"
time=2025-08-29T10:24:21.802-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=2 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:24:34.413-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=2 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:24:43.889-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=2 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:24:55.717-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=2 error="failed to parse llm result: invalid character '<' looking for beginning of value"
time=2025-08-29T10:25:01.537-04:00 level=WARN msg="Retry extract" rag=light package=golightrag function=Insert retry=2 error="failed to call LLM: error sending request: Post \"http://localhost:1234/v1/chat/completions\": context deadline exceeded"
time=2025-08-29T10:25:14.906-04:00 level=WARN msg="Retry parse result" rag=light package=golightrag function=Insert retry=3 error="failed to parse llm result: invalid character '<' looking for beginning of value"
Tested with this config on lm-studio
docker run -p 7687:7687 -p 7474:7474 -e NEO4J_AUTH=neo4j/password neo4j:latest
cat config.yaml
neo4j_uri: "bolt://localhost:7687"
neo4j_user: "neo4j"
neo4j_password: "password"
rag_llm:
type: "openai-compat" # Options: openai, openai-compat, anthropic, ollama, openrouter
api_key: "your-openai-api-key-here"
host: "http://localhost:1234/v1/"
model: "qwen3-0.6b-mlx"
parameters:
temperature: 0.7
eval_llm:
type: "openai-compat" # Options: openai, openai-compat, anthropic, ollama, openrouter
api_key: "your-openai-api-key-here"
host: "http://localhost:1234/v1/"
model: "qwen3-0.6b-mlx"
parameters:
temperature: 0.7
embedding_api_key: "your-openai-api-key-here"
log_level: "info" # Options: debug, info, warn, error
|
I attempted to run embedding models on rag_llm I am not sure how the unittests works. |
|
My machine isn't fast enough to run the unit tests. |
|
lm-studio logs |
|
The program is thinking I cannot turn off think |
|
curl http://localhost:1234/v1/chat/completions |
|
|
Turning off think requires a library upgrade |
go test -bench=. -timeout=0
Tested with this config on lm-studio
docker run -p 7687:7687 -p 7474:7474 -e NEO4J_AUTH=neo4j/password neo4j:latest