Releases: furyhawk/llama.cpp
Releases · furyhawk/llama.cpp
b5787
b5599
gguf-py : add add_classifier_output_labels method to writer (#14031) * add add_classifier_output_labels * use add_classifier_output_labels
b5572
gguf: fix failure on version == 0 (#13956)
b5568
sync : ggml ggml-ci
b2998
train : change default FA argument (#7528)
b2961
llama : add phi3 128K model support (#7225) * add phi3 128k support in convert-hf-to-gguf * add phi3 128k support in cuda * address build warnings on llama.cpp * adjust index value in cuda long rope freq factors * add long rope support in ggml cpu backend * make freq factors only depend on ctx size * remove unused rope scaling type 'su' frin gguf converter * fix flint warnings on convert-hf-to-gguf.py * set to the short freq factor when context size is small than trained context size * add one line of comments * metal : support rope freq_factors * ggml : update ggml_rope_ext API to support freq. factors * backends : add dev messages to support rope freq. factors * minor : style * tests : update to use new rope API * backends : fix pragma semicolons * minor : cleanup * llama : move rope factors from KV header to tensors * llama : remove tmp assert * cuda : fix compile warning * convert : read/write n_head_kv * llama : fix uninitialized tensors --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
b2918
ggml : fix quants nans when all the group weights are very close to z…
b2824
cmake : fix typo (#7151)