@@ -1018,6 +1018,7 @@ async def _complete_stream(
1018
1018
* ,
1019
1019
response_format : Literal [None ] = ...,
1020
1020
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1021
+ preset : str | None = ...,
1021
1022
on_message : PredictionMessageCallback | None = ...,
1022
1023
on_first_token : PredictionFirstTokenCallback | None = ...,
1023
1024
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1031,6 +1032,7 @@ async def _complete_stream(
1031
1032
* ,
1032
1033
response_format : Type [ModelSchema ] | DictSchema = ...,
1033
1034
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1035
+ preset : str | None = ...,
1034
1036
on_message : PredictionMessageCallback | None = ...,
1035
1037
on_first_token : PredictionFirstTokenCallback | None = ...,
1036
1038
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1043,6 +1045,7 @@ async def _complete_stream(
1043
1045
* ,
1044
1046
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1045
1047
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1048
+ preset : str | None = None ,
1046
1049
on_message : PredictionMessageCallback | None = None ,
1047
1050
on_first_token : PredictionFirstTokenCallback | None = None ,
1048
1051
on_prediction_fragment : PredictionFragmentCallback | None = None ,
@@ -1057,6 +1060,7 @@ async def _complete_stream(
1057
1060
prompt ,
1058
1061
response_format ,
1059
1062
config ,
1063
+ preset ,
1060
1064
on_message ,
1061
1065
on_first_token ,
1062
1066
on_prediction_fragment ,
@@ -1074,6 +1078,7 @@ async def _respond_stream(
1074
1078
* ,
1075
1079
response_format : Literal [None ] = ...,
1076
1080
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1081
+ preset : str | None = ...,
1077
1082
on_message : PredictionMessageCallback | None = ...,
1078
1083
on_first_token : PredictionFirstTokenCallback | None = ...,
1079
1084
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1087,6 +1092,7 @@ async def _respond_stream(
1087
1092
* ,
1088
1093
response_format : Type [ModelSchema ] | DictSchema = ...,
1089
1094
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1095
+ preset : str | None = ...,
1090
1096
on_message : PredictionMessageCallback | None = ...,
1091
1097
on_first_token : PredictionFirstTokenCallback | None = ...,
1092
1098
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1100,6 +1106,7 @@ async def _respond_stream(
1100
1106
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1101
1107
on_message : PredictionMessageCallback | None = None ,
1102
1108
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1109
+ preset : str | None = None ,
1103
1110
on_first_token : PredictionFirstTokenCallback | None = None ,
1104
1111
on_prediction_fragment : PredictionFragmentCallback | None = None ,
1105
1112
on_prompt_processing_progress : PromptProcessingCallback | None = None ,
@@ -1115,6 +1122,7 @@ async def _respond_stream(
1115
1122
history ,
1116
1123
response_format ,
1117
1124
config ,
1125
+ preset ,
1118
1126
on_message ,
1119
1127
on_first_token ,
1120
1128
on_prediction_fragment ,
@@ -1248,6 +1256,7 @@ async def complete_stream(
1248
1256
* ,
1249
1257
response_format : Literal [None ] = ...,
1250
1258
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1259
+ preset : str | None = ...,
1251
1260
on_message : PredictionMessageCallback | None = ...,
1252
1261
on_first_token : PredictionFirstTokenCallback | None = ...,
1253
1262
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1260,6 +1269,7 @@ async def complete_stream(
1260
1269
* ,
1261
1270
response_format : Type [ModelSchema ] | DictSchema = ...,
1262
1271
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1272
+ preset : str | None = ...,
1263
1273
on_message : PredictionMessageCallback | None = ...,
1264
1274
on_first_token : PredictionFirstTokenCallback | None = ...,
1265
1275
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1272,6 +1282,7 @@ async def complete_stream(
1272
1282
* ,
1273
1283
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1274
1284
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1285
+ preset : str | None = None ,
1275
1286
on_message : PredictionMessageCallback | None = None ,
1276
1287
on_first_token : PredictionFirstTokenCallback | None = None ,
1277
1288
on_prediction_fragment : PredictionFragmentCallback | None = None ,
@@ -1286,6 +1297,7 @@ async def complete_stream(
1286
1297
prompt ,
1287
1298
response_format = response_format ,
1288
1299
config = config ,
1300
+ preset = preset ,
1289
1301
on_message = on_message ,
1290
1302
on_first_token = on_first_token ,
1291
1303
on_prediction_fragment = on_prediction_fragment ,
@@ -1299,6 +1311,7 @@ async def complete(
1299
1311
* ,
1300
1312
response_format : Literal [None ] = ...,
1301
1313
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1314
+ preset : str | None = ...,
1302
1315
on_message : PredictionMessageCallback | None = ...,
1303
1316
on_first_token : PredictionFirstTokenCallback | None = ...,
1304
1317
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1311,6 +1324,7 @@ async def complete(
1311
1324
* ,
1312
1325
response_format : Type [ModelSchema ] | DictSchema = ...,
1313
1326
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1327
+ preset : str | None = ...,
1314
1328
on_message : PredictionMessageCallback | None = ...,
1315
1329
on_first_token : PredictionFirstTokenCallback | None = ...,
1316
1330
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1323,6 +1337,7 @@ async def complete(
1323
1337
* ,
1324
1338
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1325
1339
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1340
+ preset : str | None = None ,
1326
1341
on_message : PredictionMessageCallback | None = None ,
1327
1342
on_first_token : PredictionFirstTokenCallback | None = None ,
1328
1343
on_prediction_fragment : PredictionFragmentCallback | None = None ,
@@ -1337,6 +1352,7 @@ async def complete(
1337
1352
prompt ,
1338
1353
response_format = response_format ,
1339
1354
config = config ,
1355
+ preset = preset ,
1340
1356
on_message = on_message ,
1341
1357
on_first_token = on_first_token ,
1342
1358
on_prediction_fragment = on_prediction_fragment ,
@@ -1355,6 +1371,7 @@ async def respond_stream(
1355
1371
* ,
1356
1372
response_format : Literal [None ] = ...,
1357
1373
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1374
+ preset : str | None = ...,
1358
1375
on_message : PredictionMessageCallback | None = ...,
1359
1376
on_first_token : PredictionFirstTokenCallback | None = ...,
1360
1377
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1367,6 +1384,7 @@ async def respond_stream(
1367
1384
* ,
1368
1385
response_format : Type [ModelSchema ] | DictSchema = ...,
1369
1386
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1387
+ preset : str | None = ...,
1370
1388
on_message : PredictionMessageCallback | None = ...,
1371
1389
on_first_token : PredictionFirstTokenCallback | None = ...,
1372
1390
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1379,6 +1397,7 @@ async def respond_stream(
1379
1397
* ,
1380
1398
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1381
1399
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1400
+ preset : str | None = None ,
1382
1401
on_message : PredictionMessageCallback | None = None ,
1383
1402
on_first_token : PredictionFirstTokenCallback | None = None ,
1384
1403
on_prediction_fragment : PredictionFragmentCallback | None = None ,
@@ -1393,6 +1412,7 @@ async def respond_stream(
1393
1412
history ,
1394
1413
response_format = response_format ,
1395
1414
config = config ,
1415
+ preset = preset ,
1396
1416
on_message = on_message ,
1397
1417
on_first_token = on_first_token ,
1398
1418
on_prediction_fragment = on_prediction_fragment ,
@@ -1406,6 +1426,7 @@ async def respond(
1406
1426
* ,
1407
1427
response_format : Literal [None ] = ...,
1408
1428
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1429
+ preset : str | None = ...,
1409
1430
on_message : PredictionMessageCallback | None = ...,
1410
1431
on_first_token : PredictionFirstTokenCallback | None = ...,
1411
1432
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1418,6 +1439,7 @@ async def respond(
1418
1439
* ,
1419
1440
response_format : Type [ModelSchema ] | DictSchema = ...,
1420
1441
config : LlmPredictionConfig | LlmPredictionConfigDict | None = ...,
1442
+ preset : str | None = ...,
1421
1443
on_message : PredictionMessageCallback | None = ...,
1422
1444
on_first_token : PredictionFirstTokenCallback | None = ...,
1423
1445
on_prediction_fragment : PredictionFragmentCallback | None = ...,
@@ -1430,6 +1452,7 @@ async def respond(
1430
1452
* ,
1431
1453
response_format : Type [ModelSchema ] | DictSchema | None = None ,
1432
1454
config : LlmPredictionConfig | LlmPredictionConfigDict | None = None ,
1455
+ preset : str | None = None ,
1433
1456
on_message : PredictionMessageCallback | None = None ,
1434
1457
on_first_token : PredictionFirstTokenCallback | None = None ,
1435
1458
on_prediction_fragment : PredictionFragmentCallback | None = None ,
@@ -1444,6 +1467,7 @@ async def respond(
1444
1467
history ,
1445
1468
response_format = response_format ,
1446
1469
config = config ,
1470
+ preset = preset ,
1447
1471
on_message = on_message ,
1448
1472
on_first_token = on_first_token ,
1449
1473
on_prediction_fragment = on_prediction_fragment ,
0 commit comments