@@ -14031,11 +14031,7 @@ struct llm_build_granite_hybrid : public llm_graph_context_mamba, public llm_gra
14031
14031
14032
14032
inpL = build_inp_embd(model.tok_embd);
14033
14033
14034
- const auto * mctx_hyb = static_cast<const llama_memory_hybrid_context *>(mctx);
14035
-
14036
- auto * inp_rs = build_rs_inp(mctx_hyb->get_recr());
14037
-
14038
- auto * inp_attn = build_attn_inp_kv_unified(mctx_hyb->get_attn());
14034
+ auto * inp = build_inp_mem_hybrid();
14039
14035
14040
14036
ggml_tensor * inp_out_ids = build_inp_out_ids();
14041
14037
@@ -14056,11 +14052,11 @@ struct llm_build_granite_hybrid : public llm_graph_context_mamba, public llm_gra
14056
14052
14057
14053
if (hparams.is_recurrent(il)) {
14058
14054
// ssm layer //
14059
- cur = build_mamba2_layer(inp_rs , gf, cur, model, ubatch, il);
14055
+ cur = build_mamba2_layer(inp->get_recr() , gf, cur, model, ubatch, il);
14060
14056
} else {
14061
14057
// attention layer //
14062
14058
cur = build_attention_layer(
14063
- gf, cur, inp_pos, inp_attn , model,
14059
+ gf, cur, inp_pos, inp->get_attn() , model,
14064
14060
n_embd_head, use_rope, il);
14065
14061
}
14066
14062
0 commit comments