File tree Expand file tree Collapse file tree 9 files changed +6
-27
lines changed Expand file tree Collapse file tree 9 files changed +6
-27
lines changed Original file line number Diff line number Diff line change 29
29
from vllm .distributed .parallel_state import (destroy_distributed_environment ,
30
30
destroy_model_parallel )
31
31
from vllm .inputs import ExplicitEncoderDecoderPrompt , TextPrompt , TokensPrompt
32
- from vllm .logger import init_logger
33
32
from vllm .outputs import RequestOutput
34
33
from vllm .sampling_params import BeamSearchParams
35
34
from vllm .utils import is_list_of
36
35
37
36
from tests .model_utils import (TokensTextLogprobs ,
38
37
TokensTextLogprobsPromptLogprobs )
39
38
40
- logger = init_logger (__name__ )
41
-
42
39
_M = TypeVar ("_M" )
43
40
44
41
_PromptMultiModalInput = Union [List [_M ], List [List [_M ]]]
Original file line number Diff line number Diff line change 22
22
import torch
23
23
import torch_npu # noqa: F401
24
24
import vllm .envs as envs
25
- from vllm .logger import init_logger
25
+ from vllm .logger import logger
26
26
27
27
try :
28
28
# register custom ops into torch_library here
48
48
49
49
os .environ ["RAY_EXPERIMENTAL_NOSET_ASCEND_RT_VISIBLE_DEVICES" ] = "1"
50
50
51
- logger = init_logger (__name__ )
52
-
53
51
54
52
class NPUPlatform (Platform ):
55
53
Original file line number Diff line number Diff line change 21
21
import torch
22
22
import torch_npu # noqa: F401
23
23
from vllm .distributed import get_tensor_model_parallel_rank
24
- from vllm .logger import init_logger
25
24
from vllm .model_executor .layers .fused_moe import (FusedMoE , FusedMoEMethodBase ,
26
25
FusedMoeWeightScaleSupported )
27
26
from vllm .model_executor .layers .fused_moe .layer import \
41
40
42
41
from .quantizer import AscendQuantizer
43
42
44
- logger = init_logger (__name__ )
45
-
46
43
47
44
@register_quantization_config ("ascend" )
48
45
class AscendQuantConfig (QuantizationConfig ):
Original file line number Diff line number Diff line change 17
17
# limitations under the License.
18
18
#
19
19
import torch
20
- from vllm .logger import init_logger
21
-
22
- logger = init_logger (__name__ )
20
+ from vllm .logger import logger
23
21
24
22
25
23
def try_register_lib (lib_name : str , lib_info : str = "" ):
Original file line number Diff line number Diff line change 35
35
from vllm .distributed import get_kv_transfer_group , get_pp_group
36
36
from vllm .forward_context import set_forward_context
37
37
from vllm .inputs import INPUT_REGISTRY , InputRegistry
38
- from vllm .logger import init_logger
38
+ from vllm .logger import logger
39
39
from vllm .lora .layers import LoRAMapping
40
40
from vllm .lora .request import LoRARequest
41
41
from vllm .model_executor import SamplingMetadata , SamplingMetadataCache
63
63
if TYPE_CHECKING :
64
64
from vllm .attention .backends .abstract import AttentionBackend
65
65
66
- logger = init_logger (__name__ )
67
-
68
66
TModelInputForNPU = TypeVar ('TModelInputForNPU' , bound = "ModelInputForNPU" )
69
67
70
68
Original file line number Diff line number Diff line change 32
32
from vllm .distributed .parallel_state import get_pp_group
33
33
from vllm .forward_context import set_forward_context
34
34
from vllm .inputs import INPUT_REGISTRY
35
- from vllm .logger import init_logger
35
+ from vllm .logger import logger
36
36
from vllm .model_executor .layers .fused_moe import FusedMoE
37
37
from vllm .model_executor .model_loader import get_model
38
38
from vllm .multimodal import MULTIMODAL_REGISTRY , MultiModalKwargs
56
56
57
57
NPU_PAGED_ATTENTION_MASK_VALUE = - 10000
58
58
59
- logger = init_logger (__name__ )
60
-
61
59
62
60
class NPUModelRunner :
63
61
Original file line number Diff line number Diff line change 7
7
import torch
8
8
from torch import nn
9
9
from vllm .distributed import get_pp_group
10
- from vllm .logger import init_logger
11
10
from vllm .model_executor .layers .sampler import (PromptLogprobs , SampleLogprobs ,
12
11
SamplerOutput ,
13
12
SamplingMetadata , get_logprobs ,
30
29
if TYPE_CHECKING :
31
30
from vllm .attention .backends .abstract import AttentionBackend
32
31
33
- logger = init_logger (__name__ )
34
-
35
32
36
33
@dataclass (frozen = False )
37
34
class StatefulModelInputForNPU (StatefulModelInput ):
Original file line number Diff line number Diff line change 28
28
from vllm .distributed import (ensure_model_parallel_initialized ,
29
29
init_distributed_environment ,
30
30
set_custom_all_reduce )
31
- from vllm .logger import init_logger
31
+ from vllm .logger import logger
32
32
from vllm .lora .request import LoRARequest
33
33
from vllm .model_executor import set_random_seed
34
34
from vllm .model_executor .layers .sampler import SamplerOutput
48
48
from vllm_ascend .worker .model_runner import NPUModelRunner
49
49
from vllm_ascend .worker .pooling_model_runner import NPUPoolingModelRunner
50
50
51
- logger = init_logger (__name__ )
52
-
53
51
54
52
class NPUWorker (LocalOrDistributedWorkerBase ):
55
53
"""A worker class that executes (a partition of) the model on a NPU.
Original file line number Diff line number Diff line change 29
29
from vllm .distributed import (ensure_model_parallel_initialized ,
30
30
init_distributed_environment ,
31
31
set_custom_all_reduce )
32
- from vllm .logger import init_logger
32
+ from vllm .logger import logger
33
33
from vllm .model_executor import set_random_seed
34
34
from vllm .platforms import current_platform
35
35
from vllm .utils import STR_DTYPE_TO_TORCH_DTYPE
42
42
43
43
from vllm_ascend .worker .model_runner_v1 import NPUModelRunner
44
44
45
- logger = init_logger (__name__ )
46
-
47
45
48
46
class NPUWorker (WorkerBase ):
49
47
You can’t perform that action at this time.
0 commit comments