If fine tuning seemed to work without error and the loss went down does that mean I fine tuned? #5155
Unanswered
MotorCityCobra
asked this question in
Q&A
Replies: 1 comment
-
use the export-lora binary to merge these:
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
The finetune section of the examples dir says the fine tuning only works on llama models. I ran finetune.exe for a few hours not on a llama model, but a quantized Mistral 7B model and the loss went down. No errors. By the end of it I had two guff files and a bin file. I've tried every combination of opening these files in koboldcpp and it usually will not load a model and crashes.
If you could throw me a bone and tell me what the gguf files are for and what the one bin file is for when opening the model I'd appreciate it. Something is the lora adapters? The gguf files are just 125MB but the quantized model is about 6GB. So, does koboldcpp not allow me to run them because finetune.cpp really does not fine tune Mistral models and I haven't fine tuned anything, or do I just not know how to load the files?
If finetune.cpp cannot fine tune a Mistral 7B model where would I begin to modify the code so that it can?
Somewhere here on line 100?
struct my_llama_lora_layer {
// normalization
struct ggml_tensor * attention_norm_a;
struct ggml_tensor * attention_norm_b;
};
struct my_llama_lora {
struct ggml_context * ctx = NULL;
std::vector<uint8_t> data;
};
Beta Was this translation helpful? Give feedback.
All reactions