Skip to content

Commit 54affdc

Browse files
authored
[Doc] modify offline_inference docs (#2787)
* modify reasoning_output docs * modify offline inference docs * modify offline inference docs * modify offline_inference docs * modify offline_inference docs
1 parent a4fdb39 commit 54affdc

File tree

2 files changed

+8
-8
lines changed

2 files changed

+8
-8
lines changed

docs/offline_inference.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ outputs = llm.chat(
5151
for output in outputs:
5252
prompt = output.prompt
5353
generated_text = output.outputs.text
54-
reasoning_text = output.outputs.resoning_content
54+
reasoning_text = output.outputs.reasoning_content
5555
```
5656

5757
### Text Completion Interface (LLM.generate)
@@ -89,7 +89,7 @@ from PIL import Image
8989

9090
from fastdeploy.entrypoints.llm import LLM
9191
from fastdeploy.engine.sampling_params import SamplingParams
92-
from fastdeploy.input.ernie_tokenizer_v2 import ErnieBotTokenizer
92+
from fastdeploy.input.ernie_tokenizer import ErnieBotTokenizer
9393

9494
PATH = "baidu/ERNIE-4.5-VL-28B-A3B-Paddle"
9595
tokenizer = ErnieBotTokenizer.from_pretrained(os.path.dirname(PATH))
@@ -125,7 +125,7 @@ for message in messages:
125125
})
126126

127127
sampling_params = SamplingParams(temperature=0.1, max_tokens=6400)
128-
llm = LLM(model=PATH, tensor_parallel_size=8, max_model_len=32768, enable_mm=True, limit_mm_per_prompt={"image": 100}, reasoning_parser="ernie-45-vl")
128+
llm = LLM(model=PATH, tensor_parallel_size=1, max_model_len=32768, enable_mm=True, limit_mm_per_prompt={"image": 100}, reasoning_parser="ernie-45-vl")
129129
outputs = llm.generate(prompts={
130130
"prompt": prompt,
131131
"multimodal_data": {
@@ -138,7 +138,7 @@ outputs = llm.generate(prompts={
138138
for output in outputs:
139139
prompt = output.prompt
140140
generated_text = output.outputs.text
141-
reasoning_text = output.outputs.resoning_content
141+
reasoning_text = output.outputs.reasoning_content
142142

143143
```
144144
>Note: The `generate interface` does not currently support passing parameters to control the thinking function (on/off). It always uses the model's default parameters.

docs/zh/offline_inference.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ outputs = llm.chat(
5252
for output in outputs:
5353
prompt = output.prompt
5454
generated_text = output.outputs.text
55-
reasoning_text = output.outputs.resoning_content
55+
reasoning_text = output.outputs.reasoning_content
5656
```
5757

5858
### 续写接口(LLM.generate)
@@ -90,7 +90,7 @@ from PIL import Image
9090

9191
from fastdeploy.entrypoints.llm import LLM
9292
from fastdeploy.engine.sampling_params import SamplingParams
93-
from fastdeploy.input.ernie_tokenizer_v2 import ErnieBotTokenizer
93+
from fastdeploy.input.ernie_tokenizer import ErnieBotTokenizer
9494

9595
PATH = "baidu/ERNIE-4.5-VL-28B-A3B-Paddle"
9696
tokenizer = ErnieBotTokenizer.from_pretrained(os.path.dirname(PATH))
@@ -126,7 +126,7 @@ for message in messages:
126126
})
127127

128128
sampling_params = SamplingParams(temperature=0.1, max_tokens=6400)
129-
llm = LLM(model=PATH, tensor_parallel_size=8, max_model_len=32768, enable_mm=True, limit_mm_per_prompt={"image": 100}, reasoning_parser="ernie-45-vl")
129+
llm = LLM(model=PATH, tensor_parallel_size=1, max_model_len=32768, enable_mm=True, limit_mm_per_prompt={"image": 100}, reasoning_parser="ernie-45-vl")
130130
outputs = llm.generate(prompts={
131131
"prompt": prompt,
132132
"multimodal_data": {
@@ -139,7 +139,7 @@ outputs = llm.generate(prompts={
139139
for output in outputs:
140140
prompt = output.prompt
141141
generated_text = output.outputs.text
142-
reasoning_text = output.outputs.resoning_content
142+
reasoning_text = output.outputs.reasoning_content
143143

144144
```
145145
> 注: `generate` 接口, 暂时不支持思考开关参数控制, 均使用模型默认思考能力。

0 commit comments

Comments
 (0)