Skip to content

Commit 56fe4be

Browse files
authored
[Deprecation] Remove TokenizerPoolConfig (#20968)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
1 parent d912781 commit 56fe4be

File tree

4 files changed

+4
-62
lines changed

4 files changed

+4
-62
lines changed

docs/api/README.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@ API documentation for vLLM's configuration classes.
88

99
- [vllm.config.ModelConfig][]
1010
- [vllm.config.CacheConfig][]
11-
- [vllm.config.TokenizerPoolConfig][]
1211
- [vllm.config.LoadConfig][]
1312
- [vllm.config.ParallelConfig][]
1413
- [vllm.config.SchedulerConfig][]

tests/async_engine/test_api_server.py

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ def _query_server_long(prompt: str) -> dict:
2929

3030

3131
@pytest.fixture
32-
def api_server(tokenizer_pool_size: int, distributed_executor_backend: str):
32+
def api_server(distributed_executor_backend: str):
3333
script_path = Path(__file__).parent.joinpath(
3434
"api_server_async_engine.py").absolute()
3535
commands = [
@@ -40,8 +40,6 @@ def api_server(tokenizer_pool_size: int, distributed_executor_backend: str):
4040
"facebook/opt-125m",
4141
"--host",
4242
"127.0.0.1",
43-
"--tokenizer-pool-size",
44-
str(tokenizer_pool_size),
4543
"--distributed-executor-backend",
4644
distributed_executor_backend,
4745
]
@@ -54,10 +52,8 @@ def api_server(tokenizer_pool_size: int, distributed_executor_backend: str):
5452
uvicorn_process.terminate()
5553

5654

57-
@pytest.mark.parametrize("tokenizer_pool_size", [0, 2])
5855
@pytest.mark.parametrize("distributed_executor_backend", ["mp", "ray"])
59-
def test_api_server(api_server, tokenizer_pool_size: int,
60-
distributed_executor_backend: str):
56+
def test_api_server(api_server, distributed_executor_backend: str):
6157
"""
6258
Run the API server and test it.
6359

vllm/config.py

Lines changed: 0 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -1730,35 +1730,6 @@ def verify_with_parallel_config(
17301730
logger.warning("Possibly too large swap space. %s", msg)
17311731

17321732

1733-
@config
1734-
@dataclass
1735-
class TokenizerPoolConfig:
1736-
"""This config is deprecated and will be removed in a future release.
1737-
1738-
Passing these parameters will have no effect. Please remove them from your
1739-
configurations.
1740-
"""
1741-
1742-
pool_size: int = 0
1743-
"""This parameter is deprecated and will be removed in a future release.
1744-
Passing this parameter will have no effect. Please remove it from your
1745-
configurations."""
1746-
pool_type: str = "ray"
1747-
"""This parameter is deprecated and will be removed in a future release.
1748-
Passing this parameter will have no effect. Please remove it from your
1749-
configurations."""
1750-
extra_config: dict = field(default_factory=dict)
1751-
"""This parameter is deprecated and will be removed in a future release.
1752-
Passing this parameter will have no effect. Please remove it from your
1753-
configurations."""
1754-
1755-
def __post_init__(self) -> None:
1756-
logger.warning_once(
1757-
"TokenizerPoolConfig is deprecated and will be removed in a "
1758-
"future release. Passing this parameter will have no effect. "
1759-
"Please remove it from your configurations.")
1760-
1761-
17621733
class LoadFormat(str, enum.Enum):
17631734
AUTO = "auto"
17641735
PT = "pt"
@@ -1922,10 +1893,6 @@ class ParallelConfig:
19221893
disable_custom_all_reduce: bool = False
19231894
"""Disable the custom all-reduce kernel and fall back to NCCL."""
19241895

1925-
tokenizer_pool_config: Optional[TokenizerPoolConfig] = None
1926-
"""This parameter is deprecated and will be removed in a future release.
1927-
Please remove it from your configs"""
1928-
19291896
ray_workers_use_nsight: bool = False
19301897
"""Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler."""
19311898

vllm/engine/arg_utils.py

Lines changed: 2 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -32,8 +32,8 @@
3232
ObservabilityConfig, ParallelConfig, PoolerConfig,
3333
PrefixCachingHashAlgo, PromptAdapterConfig,
3434
SchedulerConfig, SchedulerPolicy, SpeculativeConfig,
35-
TaskOption, TokenizerMode, TokenizerPoolConfig,
36-
VllmConfig, get_attr_docs, get_field)
35+
TaskOption, TokenizerMode, VllmConfig, get_attr_docs,
36+
get_field)
3737
from vllm.logger import init_logger
3838
from vllm.platforms import CpuArchEnum, current_platform
3939
from vllm.plugins import load_general_plugins
@@ -373,13 +373,6 @@ class EngineArgs:
373373
enforce_eager: bool = ModelConfig.enforce_eager
374374
max_seq_len_to_capture: int = ModelConfig.max_seq_len_to_capture
375375
disable_custom_all_reduce: bool = ParallelConfig.disable_custom_all_reduce
376-
# The following three fields are deprecated and will be removed in a future
377-
# release. Setting them will have no effect. Please remove them from your
378-
# configurations.
379-
tokenizer_pool_size: int = TokenizerPoolConfig.pool_size
380-
tokenizer_pool_type: str = TokenizerPoolConfig.pool_type
381-
tokenizer_pool_extra_config: dict = \
382-
get_field(TokenizerPoolConfig, "extra_config")
383376
limit_mm_per_prompt: dict[str, int] = \
384377
get_field(MultiModalConfig, "limit_per_prompt")
385378
interleave_mm_strings: bool = MultiModalConfig.interleave_mm_strings
@@ -751,19 +744,6 @@ def add_cli_args(parser: FlexibleArgumentParser) -> FlexibleArgumentParser:
751744
cache_group.add_argument("--calculate-kv-scales",
752745
**cache_kwargs["calculate_kv_scales"])
753746

754-
# Tokenizer arguments
755-
tokenizer_kwargs = get_kwargs(TokenizerPoolConfig)
756-
tokenizer_group = parser.add_argument_group(
757-
title="TokenizerPoolConfig",
758-
description=TokenizerPoolConfig.__doc__,
759-
)
760-
tokenizer_group.add_argument("--tokenizer-pool-size",
761-
**tokenizer_kwargs["pool_size"])
762-
tokenizer_group.add_argument("--tokenizer-pool-type",
763-
**tokenizer_kwargs["pool_type"])
764-
tokenizer_group.add_argument("--tokenizer-pool-extra-config",
765-
**tokenizer_kwargs["extra_config"])
766-
767747
# Multimodal related configs
768748
multimodal_kwargs = get_kwargs(MultiModalConfig)
769749
multimodal_group = parser.add_argument_group(

0 commit comments

Comments
 (0)