Build and deploy llama-server on fly.io #8026
hazelnutcloud
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all, I’ve written a Dockerfile that’ll build and deploy llama-server on fly.io along with the fly.toml configuration file. Here’s the Github repo.
It uses the most minimal dependencies possible to create a small image and downloads model files on initial boot and caches them in a volume for fast subsequent cold starts.
Hope this helps!
Beta Was this translation helpful? Give feedback.
All reactions