File tree Expand file tree Collapse file tree 3 files changed +4
-4
lines changed Expand file tree Collapse file tree 3 files changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -86,7 +86,7 @@ uint32_t llama_hparams::n_embd_s() const {
86
86
return ssm_d_state * ssm_d_inner;
87
87
}
88
88
89
- bool llama_hparams::recurrent_layer (uint32_t il) const {
89
+ bool llama_hparams::is_recurrent (uint32_t il) const {
90
90
return recurrent_layer_arr[il];
91
91
}
92
92
Original file line number Diff line number Diff line change @@ -190,7 +190,7 @@ struct llama_hparams {
190
190
uint32_t n_embd_s () const ;
191
191
192
192
// whether or not the given layer is recurrent (for hybrid models)
193
- bool recurrent_layer (uint32_t il) const ;
193
+ bool is_recurrent (uint32_t il) const ;
194
194
195
195
bool is_swa (uint32_t il) const ;
196
196
};
Original file line number Diff line number Diff line change @@ -32,7 +32,7 @@ llama_memory_hybrid::llama_memory_hybrid(
32
32
mem_attn(new llama_kv_cache_unified(
33
33
model,
34
34
filter_attn == nullptr ?
35
- [&](int32_t il) { return !model.hparams .recurrent_layer (il); }
35
+ [&](int32_t il) { return !model.hparams .is_recurrent (il); }
36
36
: filter_attn,
37
37
type_k,
38
38
type_v,
@@ -47,7 +47,7 @@ llama_memory_hybrid::llama_memory_hybrid(
47
47
mem_recr (new llama_memory_recurrent(
48
48
model,
49
49
filter_recr == nullptr ?
50
- [&](int32_t il) { return model.hparams .recurrent_layer (il); }
50
+ [&](int32_t il) { return model.hparams .is_recurrent (il); }
51
51
: filter_recr,
52
52
type_r,
53
53
type_s,
You can’t perform that action at this time.
0 commit comments