New Feature: Multiple Llamafile agent support #1956
fastzombies
started this conversation in
Feature Requests
Replies: 1 comment
-
A Request to include connectivity for LLAMAFILE llamafile in Cline. Pls add the same, to run LLMs locally. Thanks in advance. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Problem
This is an important feature request. There are times where one must work at a client/customer site in an environment where no internet access or sudo are possible but VSCode is available. In this case the best, and perhaps only option is to bring in some vsix and llamafiles.
Proposed
Due to the limited resource nature of this scenario that forces use of of locally served models, it may be helpful to support multiple llamafiles for different agents:
This way context windows can be preserved as much as possible, particularly of the planning agent as it may be asking a lot for even a 16K window in these circumstances.
Notes
Attempted
I set it up with OpenAI-Compatible and
http://127.0.0.1:8080/v1
,NO_KEY
for API key and I cannot change the model. It wants to work, but it gets passed the wrong type.I can use this and get a response:
It seems like it would not take much to make this work.
Duplicates
This is a duplicate of #1041 but I wanted to create and emphasize my own requirements and justification.
Beta Was this translation helpful? Give feedback.
All reactions