From 9c590193a5403661888facb76903135166c02dc7 Mon Sep 17 00:00:00 2001 From: Avasam Date: Wed, 28 May 2025 12:39:34 -0400 Subject: [PATCH] Remove upstream partial transformers stubs --- pyproject.toml | 2 - .../models/auto/auto_factory.pyi | 45 ------------------ .../models/auto/configuration_auto.pyi | 47 ------------------- .../models/auto/tokenization_auto.pyi | 32 ------------- stubs/transformers-stubs/py.typed | 1 - 5 files changed, 127 deletions(-) delete mode 100644 stubs/transformers-stubs/models/auto/auto_factory.pyi delete mode 100644 stubs/transformers-stubs/models/auto/configuration_auto.pyi delete mode 100644 stubs/transformers-stubs/models/auto/tokenization_auto.pyi delete mode 100644 stubs/transformers-stubs/py.typed diff --git a/pyproject.toml b/pyproject.toml index e8321cd9..d2c5b326 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -28,7 +28,6 @@ tests = [ "scikit-learn", "sympy", "traitlets", - "transformers", ] dev = [{ include-group = "hygiene" }, { include-group = "tests" }] @@ -45,7 +44,6 @@ dev = [{ include-group = "hygiene" }, { include-group = "tests" }] "skimage-stubs" = "stubs/skimage" "sklearn-stubs" = "stubs/sklearn" "sympy-stubs" = "stubs/sympy-stubs" -"transformers-stubs" = "stubs/transformers-stubs" "vispy-stubs" = "stubs/vispy" [tool.ruff] diff --git a/stubs/transformers-stubs/models/auto/auto_factory.pyi b/stubs/transformers-stubs/models/auto/auto_factory.pyi deleted file mode 100644 index 3a928f00..00000000 --- a/stubs/transformers-stubs/models/auto/auto_factory.pyi +++ /dev/null @@ -1,45 +0,0 @@ -import os -from collections import OrderedDict -from collections.abc import Iterator -from typing import Any, TypeVar -from typing_extensions import TypeAlias - -from transformers.configuration_utils import PretrainedConfig -from transformers.tokenization_utils_fast import PreTrainedTokenizerFast - -_T = TypeVar("_T") -# Tokenizers will depend on packages installed, too much variance and there are no common base or Protocol -_LazyAutoMappingValue: TypeAlias = tuple[type[Any] | None, type[Any] | None] - -CLASS_DOCSTRING: str -FROM_CONFIG_DOCSTRING: str -FROM_PRETRAINED_TORCH_DOCSTRING: str -FROM_PRETRAINED_TF_DOCSTRING: str -FROM_PRETRAINED_FLAX_DOCSTRING: str - -class _BaseAutoModelClass: - def __init__(self, *args, **kwargs) -> None: ... - @classmethod - def from_config(cls, config, **kwargs): ... - @classmethod - def from_pretrained(cls, pretrained_model_name_or_path: str | os.PathLike[str], *model_args, **kwargs): ... - @classmethod - def register(cls, config_class, model_class, exist_ok=False) -> None: ... - -def insert_head_doc(docstring, head_doc: str = ""): ... -def auto_class_update(cls, checkpoint_for_example: str = "bert-base-cased", head_doc: str = ""): ... -def get_values(model_mapping): ... -def getattribute_from_module(module, attr): ... - -class _LazyAutoMapping(OrderedDict[type[PretrainedConfig], _LazyAutoMappingValue]): - def __init__(self, config_mapping, model_mapping) -> None: ... - def __len__(self) -> int: ... - def __getitem__(self, key: type[PretrainedConfig]) -> _LazyAutoMappingValue: ... - def keys(self) -> list[type[PretrainedConfig]]: ... - def get(self, key: type[PretrainedConfig], default: _T) -> _LazyAutoMappingValue | _T: ... - def __bool__(self) -> bool: ... - def values(self) -> list[_LazyAutoMappingValue]: ... - def items(self) -> list[tuple[type[PretrainedConfig], _LazyAutoMappingValue]]: ... - def __iter__(self) -> Iterator[type[PretrainedConfig]]: ... - def __contains__(self, item: type) -> bool: ... - def register(self, key: type[PretrainedConfig], value: _LazyAutoMappingValue, exist_ok=False) -> None: ... diff --git a/stubs/transformers-stubs/models/auto/configuration_auto.pyi b/stubs/transformers-stubs/models/auto/configuration_auto.pyi deleted file mode 100644 index 5b92d5d3..00000000 --- a/stubs/transformers-stubs/models/auto/configuration_auto.pyi +++ /dev/null @@ -1,47 +0,0 @@ -import os -from collections import OrderedDict -from collections.abc import Callable, Iterator, KeysView, ValuesView -from typing import Any, NoReturn, TypeVar - -from transformers.configuration_utils import PretrainedConfig - -_F = TypeVar("_F", bound=Callable[..., Any]) - -CONFIG_MAPPING_NAMES: OrderedDict[str, str] -MODEL_NAMES_MAPPING: OrderedDict[str, str] -SPECIAL_MODEL_TYPE_TO_MODULE_NAME: OrderedDict[str, str] - -def model_type_to_module_name(key) -> str: ... -def config_class_to_model_type(config) -> str | None: ... - -class _LazyConfigMapping(OrderedDict[str, type[PretrainedConfig]]): - def __init__(self, mapping) -> None: ... - def __getitem__(self, key: str) -> type[PretrainedConfig]: ... - def keys(self) -> list[str]: ... - def values(self) -> list[type[PretrainedConfig]]: ... - def items(self) -> list[tuple[str, type[PretrainedConfig]]]: ... - def __iter__(self) -> Iterator[str]: ... - def __contains__(self, item: object) -> bool: ... - def register(self, key: str, value: type[PretrainedConfig], exist_ok=False) -> None: ... - -CONFIG_MAPPING: _LazyConfigMapping - -class _LazyLoadAllMappings(OrderedDict[str, str]): - def __init__(self, mapping) -> None: ... - def __getitem__(self, key: str) -> str: ... - def keys(self) -> KeysView[str]: ... - def values(self) -> ValuesView[str]: ... - def items(self) -> KeysView[str]: ... - def __iter__(self) -> Iterator[str]: ... - def __contains__(self, item: object) -> bool: ... - -def replace_list_option_in_docstrings(config_to_class=None, use_model_types: bool = True) -> Callable[[_F], _F]: ... - -class AutoConfig: - def __init__(self) -> None: ... - @classmethod - def for_model(cls, model_type: str, *args, **kwargs) -> PretrainedConfig: ... - @classmethod - def from_pretrained(cls, pretrained_model_name_or_path: str | os.PathLike[str], **kwargs): ... - @staticmethod - def register(model_type, config, exist_ok=False) -> None: ... diff --git a/stubs/transformers-stubs/models/auto/tokenization_auto.pyi b/stubs/transformers-stubs/models/auto/tokenization_auto.pyi deleted file mode 100644 index ca62abdf..00000000 --- a/stubs/transformers-stubs/models/auto/tokenization_auto.pyi +++ /dev/null @@ -1,32 +0,0 @@ -import os -from collections import OrderedDict -from typing import Any -from typing_extensions import TypeAlias - -from transformers.models.auto.auto_factory import _LazyAutoMapping -from transformers.tokenization_utils import PreTrainedTokenizer -from transformers.tokenization_utils_fast import PreTrainedTokenizerFast - -TOKENIZER_MAPPING_NAMES: OrderedDict[str, tuple[str | None, str | None]] -TOKENIZER_MAPPING: _LazyAutoMapping -CONFIG_TO_TYPE: dict[str, str] - -def tokenizer_class_from_name(class_name: str) -> type[Any] | None: ... -def get_tokenizer_config( - pretrained_model_name_or_path: str | os.PathLike[str], - cache_dir: str | os.PathLike[str] | None = None, - force_download: bool = False, - resume_download: bool | None = None, - proxies: dict[str, str] | None = None, - token: bool | str | None = None, - revision: str | None = None, - local_files_only: bool = False, - subfolder: str = "", - **kwargs, -) -> dict[str, Any]: ... - -class AutoTokenizer: - def __init__(self) -> None: ... - @classmethod - def from_pretrained(cls, pretrained_model_name_or_path: str | os.PathLike[str], *inputs, **kwargs): ... - def register(config_class, slow_tokenizer_class=None, fast_tokenizer_class=None, exist_ok=False) -> None: ... diff --git a/stubs/transformers-stubs/py.typed b/stubs/transformers-stubs/py.typed deleted file mode 100644 index b648ac92..00000000 --- a/stubs/transformers-stubs/py.typed +++ /dev/null @@ -1 +0,0 @@ -partial