File tree Expand file tree Collapse file tree 2 files changed +3
-8
lines changed Expand file tree Collapse file tree 2 files changed +3
-8
lines changed Original file line number Diff line number Diff line change @@ -268,8 +268,7 @@ def load_llama_from_model_settings(settings: ModelSettings) -> llama_cpp.Llama:
268
268
# Sampling Params
269
269
last_n_tokens_size = settings .last_n_tokens_size ,
270
270
# LoRA Params
271
- lora_base = settings .lora_base ,
272
- lora_path = settings .lora_path ,
271
+ lora_adapters = settings .lora_adapters ,
273
272
# Backend Params
274
273
numa = settings .numa ,
275
274
# Chat Format Params
Original file line number Diff line number Diff line change @@ -113,13 +113,9 @@ class ModelSettings(BaseSettings):
113
113
description = "Last n tokens to keep for repeat penalty calculation." ,
114
114
)
115
115
# LoRA Params
116
- lora_base : Optional [str ] = Field (
116
+ lora_adapters : Optional [Dict [ str , float ]] = Field (
117
117
default = None ,
118
- description = "Optional path to base model, useful if using a quantized base model and you want to apply LoRA to an f16 model." ,
119
- )
120
- lora_path : Optional [str ] = Field (
121
- default = None ,
122
- description = "Path to a LoRA file to apply to the model." ,
118
+ description = "Paths to LoRA adapter files and the scale to apply to them at (scale of 0.0 will not be used during inference)." ,
123
119
)
124
120
# Backend Params
125
121
numa : Union [bool , int ] = Field (
You can’t perform that action at this time.
0 commit comments