Llava models support #1842
Unanswered
iKachaykin
asked this question in
Q&A
Replies: 2 comments
-
Trying to work through this myself now, haven't touched the framework before so I'm probably as lost as you are. Only thing I've found I can do with a GGUF so far thanks to reading gguf_file.rs is reading it like this let model_path = PathBuf::from("models//mistral-7b-instruct-v0.2.Q5_K_M.gguf");
let model_file = std::fs::File::open(model_path)?;
let mmap = unsafe { memmap2::Mmap::map(&model_file)? };
let cursor = std::io::Cursor::new(mmap);
let mut reader = std::io::BufReader::new(cursor);
let gguf = Content::read(&mut reader)?; |
Beta Was this translation helpful? Give feedback.
0 replies
-
It looks like the Llava model is based on a Mistral architecture, so it should be able to run with any Mistral architecture? Perhaps the quantized example or mistral.rs could be used to run the weights. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hey! Thanks a lot for your amazing package! I would like to run Llava models with candle, but I am a bit confused right now how may I do that. Please correct me if I am wrong: to run any custom model not listed in candle-examples, I need to reimplement it from scratch using candle, correct? Or maybe there is any other better way to run a .gguf file with candle?
Beta Was this translation helpful? Give feedback.
All reactions