Replies: 4 comments
-
@crashr Thank you for the kind words and for suggesting the feature to be able to name the agents. :) I have already added it to the backlog. That should be a quick thing to add for me: #15 Also, you have a really nice project. I really appreciate that you tried it with Paddler. If you add an example to your repo on integrating it with Paddler, I will be happy to feature gppm in Paddler's readme. If you also want to do it, just issue a Pull Request to change the README and add the "Featured Projects" section before the "Changelog" section and list gppm there. |
Beta Was this translation helpful? Give feedback.
-
@crashr I just pushed that feature as a pre-release. You can add a https://github.com/distantmagic/paddler/releases/tag/v0.6.0-rc1 |
Beta Was this translation helpful? Give feedback.
-
@mcharytoniuk Thank you for the quick implementation. This is exactly what I was asking for. As soon as I have my documentation ready in the next few days, there will definitely be an example of how to use it with paddler. It is already a central component of my personal setup. |
Beta Was this translation helpful? Give feedback.
-
This is a real world example that I use: - name: "Codestral-22B-v0.1-Q8_0 (paddler balancer)"
enabled: True
command:
"/usr/local/bin/paddler balancer \
--management-host 0.0.0.0 \
--management-port 8085 \
--management-dashboard-enable=true \
--reverseproxy-host 192.168.178.56 \
--reverseproxy-port 8081"
- name: "Codestral-22B-v0.1-Q8_0 (llama.cpp 01)"
enabled: True
env:
CUDA_VISIBLE_DEVICES: "0,1,2"
command:
"/usr/local/bin/llama-server \
--host 0.0.0.0 \
-ngl 100 \
-m /models/Codestral-22B-v0.1-Q8_0.gguf \
--port 8082 \
-fa \
-sm row \
-mg 0 \
--no-mmap \
--log-format json"
post_launch_hooks:
- name: Codestral-22B-v0.1-Q8_0_(paddler_01)
enabled: True
command:
"/usr/local/bin/paddler agent \
--name 'Codestral-22B-v0.1-Q8_0 (llama.cpp 01)' \
--external-llamacpp-host 192.168.178.56 \
--external-llamacpp-port 8082 \
--local-llamacpp-host 192.168.178.56 \
--local-llamacpp-port 8082 \
--management-host 192.168.178.56 \
--management-port 8085"
- name: "Codestral-22B-v0.1-Q8_0_(llama.cpp_02)"
enabled: True
env:
CUDA_VISIBLE_DEVICES: "0,1,2"
command:
"/usr/local/bin/llama-server \
--host 0.0.0.0 \
-ngl 100 \
-m /models/Codestral-22B-v0.1-Q8_0.gguf \
--port 8083 \
-fa \
-sm row \
-mg 1 \
--no-mmap \
--log-format json"
post_launch_hooks:
- name: "Codestral-22B-v0.1-Q8_0_Paddler_02"
enabled: True
command:
"/usr/local/bin/paddler agent \
--name 'Codestral-22B-v0.1-Q8_0 (llama.cpp 02)' \
--external-llamacpp-host 192.168.178.56 \
--external-llamacpp-port 8083 \
--local-llamacpp-host 192.168.178.56 \
--local-llamacpp-port 8083 \
--management-host 192.168.178.56 \
--management-port 8085" Launching the balancer like this is actually a bit hacky since gppm wasn't intended to do this but it works and I like to have everything bundled that belongs together. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi.
I just tried paddler and it works great. I implemted a nice way in gppm (dev branch) to launch paddler agents via a post ready hook when the llama.cpp instance is up and running. I simply configure it like this:
and a view seconds later an up and running HA setup pops out.
I love it!
But as the title says it would be nice to be able to set a name or identifier via agent flag that is visible in the dashboard.
Greets
Beta Was this translation helpful? Give feedback.
All reactions