Does finetune support phi-2 #4995
Replies: 3 comments
-
[1705465454] Log start |
Beta Was this translation helpful? Give feedback.
-
I have same error |
Beta Was this translation helpful? Give feedback.
-
From the Readme in ggerganov/examples/finetune/ I think this time "llama" is literal and applies only to the llama models from Facebook. Today I looked into modifying the code and holy lord, I am out of my depth. Have you looked at the ggml.h and finetune.cpp files? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I recently ran a finetune on a mistral model and all seems great.
However, when I run the same text on the phi-2, I obtain the following log when running a test prompt
<main.log added as comment>
mainly these two lines
[1705465495] llama_apply_lora_from_file_internal: incompatible tensor dimensions (2560 and 4096); are you sure that this adapter is for this model?
[1705465496] main: error: unable to load model
Is there something I have to set to make the tensor dimensions the same. When I run the same test on the mistral LORA
llama_apply_lora_from_file_internal: warning: using a lora adapter with a quantized model may result in poor quality, use a f16 or f32 base model with --lora-base
Also note when I try to merge them I get a general assert
llama.cpp\export-lora.exe --model-base models\phi-2.Q5_K_M.gguf --model-out models\phi-2-ecsql.Q5_K_M.gguf --lora-scaled lora\phi2\ggml-lora-150-f32.gguf 1.0
Device 0: NVIDIA GeForce RTX 2070, compute capability 7.5, VMM: yes
GGML_ASSERT: F:\development\GPT\llama.cpp\ggml.c:3190: ggml_can_repeat(b, a)
phi2.main.log
Beta Was this translation helpful? Give feedback.
All reactions