@@ -17357,7 +17357,7 @@ struct llama_data_context {
17357
17357
}
17358
17358
17359
17359
void write_logits(const struct llama_context * ctx) {
17360
- const uint64_t logits_size = std::min(ctx->logits_size, (uint64_t) ctx->n_outputs * ctx->model.hparams.n_vocab);
17360
+ const uint64_t logits_size = std::min((uint64_t) ctx->logits_size, (uint64_t) ctx->n_outputs * ctx->model.hparams.n_vocab);
17361
17361
17362
17362
write(&logits_size, sizeof(logits_size));
17363
17363
@@ -17367,7 +17367,7 @@ struct llama_data_context {
17367
17367
}
17368
17368
17369
17369
void write_embeddings(const struct llama_context * ctx) {
17370
- const uint64_t embeddings_size = std::min(ctx->embd_size, (uint64_t) ctx->n_outputs * ctx->model.hparams.n_embd);
17370
+ const uint64_t embeddings_size = std::min((uint64_t) ctx->embd_size, (uint64_t) ctx->n_outputs * ctx->model.hparams.n_embd);
17371
17371
17372
17372
write(&embeddings_size, sizeof(embeddings_size));
17373
17373
@@ -17461,7 +17461,7 @@ struct llama_data_context {
17461
17461
write(&v_type_i, sizeof(v_type_i));
17462
17462
17463
17463
// Write element size
17464
- const uint64_t v_size_el = ggml_type_size(kv_self.v_l[il]->type);
17464
+ const uint32_t v_size_el = ggml_type_size(kv_self.v_l[il]->type);
17465
17465
write(&v_size_el, sizeof(v_size_el));
17466
17466
17467
17467
// Write GQA embedding size
@@ -17710,7 +17710,7 @@ struct llama_data_read_context {
17710
17710
const size_t k_size_row = ggml_row_size(kv_self.k_l[il]->type, n_embd_k_gqa);
17711
17711
if (k_size_row != k_size_row_ref) {
17712
17712
// llama_kv_cache_seq_rm(kv_self, dest_seq_id, -1, -1);
17713
- LLAMA_LOG_ERROR("%s: mismatched key row size (%zu != %zu, layer %d)\n", __func__, k_size_row, k_size_row_ref, il);
17713
+ LLAMA_LOG_ERROR("%s: mismatched key row size (%zu != %zu, layer %d)\n", __func__, k_size_row, (size_t) k_size_row_ref, il);
17714
17714
return false;
17715
17715
}
17716
17716
@@ -17739,7 +17739,7 @@ struct llama_data_read_context {
17739
17739
read_to(&v_size_row_ref, sizeof(v_size_row_ref));
17740
17740
const size_t v_size_row = ggml_row_size(kv_self.v_l[il]->type, n_embd_v_gqa);
17741
17741
if (v_size_row != v_size_row_ref) {
17742
- LLAMA_LOG_ERROR("%s: mismatched value row size (%zu != %zu, layer %d)\n", __func__, v_size_row, v_size_row_ref, il);
17742
+ LLAMA_LOG_ERROR("%s: mismatched value row size (%zu != %zu, layer %d)\n", __func__, v_size_row, (size_t) v_size_row_ref, il);
17743
17743
return false;
17744
17744
}
17745
17745
@@ -17763,11 +17763,11 @@ struct llama_data_read_context {
17763
17763
}
17764
17764
17765
17765
// Read element size of value
17766
- uint64_t v_size_el_ref;
17766
+ uint32_t v_size_el_ref;
17767
17767
read_to(&v_size_el_ref, sizeof(v_size_el_ref));
17768
17768
const size_t v_size_el = ggml_type_size(kv_self.v_l[il]->type);
17769
17769
if (v_size_el != v_size_el_ref) {
17770
- LLAMA_LOG_ERROR("%s: mismatched value element size (%zu != %llu , layer %d)\n", __func__, v_size_el, v_size_el_ref, il);
17770
+ LLAMA_LOG_ERROR("%s: mismatched value element size (%zu != %zu , layer %d)\n", __func__, v_size_el, (size_t) v_size_el_ref, il);
17771
17771
return false;
17772
17772
}
17773
17773
0 commit comments