Skip to content

Commit ab3de32

Browse files
committed
mamba : remove some useless comments
No code change.
1 parent 2e1bcff commit ab3de32

File tree

1 file changed

+2
-6
lines changed

1 file changed

+2
-6
lines changed

llama.cpp

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4195,9 +4195,6 @@ static bool llm_load_tensors(
41954195

41964196
auto & layer = model.layers[i];
41974197

4198-
// TODO: what's the difference between ctx_layer and ctx_split?
4199-
// A: It seems that ctx_split is for matrices (2d???) while ctx_layer is for other things (like 1D bias and norms, probably.)
4200-
42014198
// norm
42024199
layer.attn_norm = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd});
42034200

@@ -6912,7 +6909,6 @@ struct llm_build_context {
69126909

69136910
const int32_t n_tok = batch.n_tokens;
69146911

6915-
// hopefully the compiler does constant folding
69166912
const int64_t d_model = n_embd;
69176913
const int64_t d_inner = n_head;
69186914
GGML_ASSERT(2 * d_model == d_inner);
@@ -6959,8 +6955,8 @@ struct llm_build_context {
69596955

69606956
// The following tensor is too big in order to avoid an assertion error when making an overlapping view.
69616957
// TODO: in ggml_new_tensor_impl, handle overlapping data range in data size calculation
6962-
// This could then be a tensor with ne[] = {(d_conv-1)+n_tok, d_inner}
6963-
// which is around (d_conv-1) times as small as its current size.
6958+
// This could then be a tensor with ne[] = {(d_conv-1)+n_tok, d_inner},
6959+
// but the size difference is not that big (d_conv is usually 4).
69646960
struct ggml_tensor * conv_x = ggml_new_tensor_1d(ctx0, conv_state->type, d_conv*d_inner*n_tok);
69656961
const size_t conv_x_nb1 = (d_conv - 1 + n_tok) * ggml_element_size(conv_x);
69666962

0 commit comments

Comments
 (0)