Skip to content

Commit ece8859

Browse files
authored
fix spelling error (#2827)
1 parent bad53c6 commit ece8859

File tree

2 files changed

+8
-8
lines changed

2 files changed

+8
-8
lines changed

fastdeploy/entrypoints/openai/api_server.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -348,7 +348,7 @@ def reset_scheduler():
348348

349349
if llm_engine is None:
350350
return Response("Engine not loaded", status_code=500)
351-
llm_engine.scheduler.reset_scheduler()
351+
llm_engine.scheduler.reset()
352352
return Response("Scheduler Reset Successfully", status_code=200)
353353

354354

@@ -366,7 +366,7 @@ def control_scheduler(request: ControlSchedulerRequest):
366366
return JSONResponse(content=content.model_dump(), status_code=500)
367367

368368
if request.reset:
369-
llm_engine.scheduler.reset_scheduler()
369+
llm_engine.scheduler.reset()
370370

371371
if request.load_shards_num or request.reallocate_shard:
372372
if hasattr(llm_engine.scheduler, "update_config") and callable(llm_engine.scheduler.update_config):

fastdeploy/scheduler/config.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -110,7 +110,7 @@ def __init__(self,
110110
ttl: int = 900,
111111
min_load_score: float = 3,
112112
max_model_len: int = 8192,
113-
load_shrads_num: int = 1,
113+
load_shards_num: int = 1,
114114
enable_chunked_prefill: bool = False,
115115
max_num_partial_prefills: int = 1,
116116
max_long_partial_prefills: int = 1,
@@ -129,7 +129,7 @@ def __init__(self,
129129
ttl: Time-to-live in seconds for Redis keys (default 900s)
130130
min_load_score: Minimum load score for task assignment (default 3)
131131
max_model_len: Maximum model context length in tokens
132-
load_shrads_num: Number of load balancing shards
132+
load_shards_num: Number of load balancing shards
133133
enable_chunked_prefill: Whether to enable chunked prefill processing
134134
max_num_partial_prefills: Max partial prefill operations allowed
135135
max_long_partial_prefills: Max long-running partial prefill ops
@@ -147,7 +147,7 @@ def __init__(self,
147147
self.topic = topic
148148
self.ttl = ttl
149149
self.min_load_score = min_load_score
150-
self.load_shrads_num = load_shrads_num
150+
self.load_shards_num = load_shards_num
151151

152152
self.max_model_len = max_model_len
153153
self.enable_chunked_prefill = enable_chunked_prefill
@@ -169,8 +169,8 @@ def check(self):
169169
raise ValueError("ttl should be greater than 60")
170170
if self.min_load_score < 1:
171171
raise ValueError("min_load_score should be greater than 0")
172-
if self.load_shrads_num < 1:
173-
raise ValueError("load_shrads_num should be greater than 0")
172+
if self.load_shards_num < 1:
173+
raise ValueError("load_shards_num should be greater than 0")
174174

175175
r = redis.Redis(self.host, self.port, self.db, self.password)
176176
try:
@@ -262,7 +262,7 @@ def scheduler(self):
262262
topic=self.config.topic,
263263
ttl=self.config.ttl,
264264
min_load_score=self.config.min_load_score,
265-
load_shrads_num=self.config.load_shrads_num,
265+
load_shards_num=self.config.load_shards_num,
266266
enable_chunked_prefill=self.config.enable_chunked_prefill,
267267
max_num_partial_prefills=self.config.max_num_partial_prefills,
268268
max_long_partial_prefills=self.config.max_long_partial_prefills,

0 commit comments

Comments
 (0)