Skip to content

Commit 03693fa

Browse files
committed
parmas-related fix
1 parent 81f4797 commit 03693fa

File tree

2 files changed

+1
-2
lines changed

2 files changed

+1
-2
lines changed

convert_hf_to_gguf.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3946,7 +3946,6 @@ def set_vocab(self):
39463946

39473947
def set_gguf_parameters(self):
39483948
super().set_gguf_parameters()
3949-
self._try_set_pooling_type()
39503949
self.gguf_writer.add_sliding_window(self.hparams["local_attention"])
39513950
self.gguf_writer.add_rope_freq_base(self.hparams["global_rope_theta"])
39523951
self.gguf_writer.add_rope_freq_base_swa(self.hparams["local_rope_theta"])

src/llama-model.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -724,7 +724,6 @@ void llama_model::load_hparams(llama_model_loader & ml) {
724724
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_EPS, hparams.f_norm_eps);
725725
ml.get_key(LLM_KV_ATTENTION_CAUSAL, hparams.causal_attn);
726726
ml.get_key(LLM_KV_POOLING_TYPE, hparams.pooling_type, false);
727-
ml.get_arr_n(LLM_KV_CLASSIFIER_OUTPUT_LABELS, hparams.n_cls_out, false);
728727

729728
switch (hparams.n_layer) {
730729
case 22:
@@ -2237,6 +2236,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
22372236
cls_norm = create_tensor(tn(LLM_TENSOR_CLS_NORM, "weight"), {n_embd}, TENSOR_NOT_REQUIRED);
22382237

22392238
cls_out = create_tensor(tn(LLM_TENSOR_CLS_OUT, "weight"), {n_embd, hparams.n_cls_out}, TENSOR_NOT_REQUIRED);
2239+
cls_out_b = create_tensor(tn(LLM_TENSOR_CLS_OUT, "bias"), {hparams.n_cls_out}, TENSOR_NOT_REQUIRED);
22402240
} break;
22412241
case LLM_ARCH_JINA_BERT_V2:
22422242
{

0 commit comments

Comments
 (0)