@@ -294,6 +294,8 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
294
294
# LLAMA_VOCAB_PRE_TYPE_MPT = 5,
295
295
# LLAMA_VOCAB_PRE_TYPE_STARCODER = 6,
296
296
# LLAMA_VOCAB_PRE_TYPE_GPT2 = 7,
297
+ # LLAMA_VOCAB_PRE_TYPE_REFACT = 8,
298
+ # LLAMA_VOCAB_PRE_TYPE_COMMAND_R = 9,
297
299
# };
298
300
LLAMA_VOCAB_PRE_TYPE_DEFAULT = 0
299
301
LLAMA_VOCAB_PRE_TYPE_LLAMA3 = 1
@@ -303,6 +305,8 @@ def byref(obj: CtypesCData, offset: Optional[int] = None) -> CtypesRef[CtypesCDa
303
305
LLAMA_VOCAB_PRE_TYPE_MPT = 5
304
306
LLAMA_VOCAB_PRE_TYPE_STARCODER = 6
305
307
LLAMA_VOCAB_PRE_TYPE_GPT2 = 7
308
+ LLAMA_VOCAB_PRE_TYPE_REFACT = 8
309
+ LLAMA_VOCAB_PRE_TYPE_COMMAND_R = 9
306
310
307
311
308
312
# // note: these values should be synchronized with ggml_rope
@@ -494,7 +498,7 @@ class llama_token_data_array(ctypes.Structure):
494
498
495
499
llama_token_data_array_p = ctypes .POINTER (llama_token_data_array )
496
500
497
- # typedef bool (*llama_progress_callback)(float progress, void *ctx );
501
+ # typedef bool (*llama_progress_callback)(float progress, void * user_data );
498
502
llama_progress_callback = ctypes .CFUNCTYPE (
499
503
ctypes .c_bool , ctypes .c_float , ctypes .c_void_p
500
504
)
0 commit comments