File tree Expand file tree Collapse file tree 5 files changed +5
-9
lines changed
model_executor/layers/fused_moe Expand file tree Collapse file tree 5 files changed +5
-9
lines changed Original file line number Diff line number Diff line change 6
6
7
7
import pytest
8
8
import torch
9
- import triton .language as tl
10
9
11
10
from tests .kernels .moe .utils import (batched_moe ,
12
11
make_quantized_test_activations ,
18
17
invoke_moe_batched_triton_kernel )
19
18
from vllm .model_executor .layers .fused_moe .fused_moe import fused_topk
20
19
from vllm .platforms import current_platform
20
+ from vllm .triton_utils import tl
21
21
22
22
MNK_FACTORS = [
23
23
(1 , 128 , 128 ),
Original file line number Diff line number Diff line change 8
8
# - Thomas Parnell <tpa@zurich.ibm.com>
9
9
10
10
import torch
11
- import triton
12
- import triton .language as tl
13
11
14
12
from vllm .logger import init_logger
13
+ from vllm .triton_utils import tl , triton
15
14
16
15
logger = init_logger (__name__ )
17
16
Original file line number Diff line number Diff line change 8
8
"""
9
9
10
10
import torch
11
- import triton
12
- import triton .language as tl
13
11
14
12
from vllm .lora .ops .triton_ops .kernel_utils import do_expand_kernel
15
13
from vllm .lora .ops .triton_ops .utils import _get_lora_b_ptr
16
14
from vllm .platforms import current_platform
15
+ from vllm .triton_utils import tl , triton
17
16
from vllm .utils import direct_register_custom_op
18
17
19
18
Original file line number Diff line number Diff line change 8
8
"""
9
9
10
10
import torch
11
- import triton
12
- import triton .language as tl
13
11
14
12
from vllm .lora .ops .triton_ops .kernel_utils import do_shrink_kernel
15
13
from vllm .lora .ops .triton_ops .utils import _get_lora_a_ptr
16
14
from vllm .platforms import current_platform
15
+ from vllm .triton_utils import tl , triton
17
16
from vllm .utils import direct_register_custom_op
18
17
19
18
Original file line number Diff line number Diff line change 4
4
from typing import Optional
5
5
6
6
import torch
7
- import triton
8
- import triton .language as tl
9
7
10
8
import vllm .model_executor .layers .fused_moe .modular_kernel as mk
11
9
from vllm .model_executor .layers .fused_moe .config import FusedMoEQuantConfig
18
16
normalize_scales_shape )
19
17
from vllm .model_executor .layers .quantization .utils .quant_utils import (
20
18
group_broadcast )
19
+ from vllm .triton_utils import tl , triton
21
20
22
21
23
22
@triton .jit
You can’t perform that action at this time.
0 commit comments