We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 684aaad commit 4e16bb9Copy full SHA for 4e16bb9
examples/offline_inference/qwen_1m.py
@@ -56,7 +56,7 @@ def initialize_engine(max_model_len=1048576) -> LLM:
56
enforce_eager=True,
57
enable_chunked_prefill=True,
58
max_num_batched_tokens=131072,
59
- hf_overrides={"max_position_embeddings": max_model_len}
+ hf_overrides={"max_position_embeddings": max_model_len},
60
)
61
return llm
62
0 commit comments