File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
vllm/model_executor/layers/mamba Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change 33
33
LoaderFunction , composed_weight_loader , sharded_weight_loader )
34
34
from vllm .model_executor .models .mamba_cache import MambaCacheParams
35
35
from vllm .model_executor .utils import set_weight_attrs
36
- from vllm .platform import current_platform
36
+ from vllm .platforms import current_platform
37
37
from vllm .utils import direct_register_custom_op
38
38
from vllm .v1 .attention .backends .mamba_attn import Mamba2AttentionMetadata
39
39
@@ -746,7 +746,7 @@ def mamba_mixer2(
746
746
output : torch .Tensor ,
747
747
layer_name : str ,
748
748
mup_vector : Optional [torch .Tensor ] = None ,
749
- ):
749
+ ) -> None :
750
750
forward_context : ForwardContext = get_forward_context ()
751
751
self = forward_context .no_compile_layers [layer_name ]
752
752
self .forward_cuda (hidden_states = hidden_states ,
@@ -761,7 +761,7 @@ def mamba_mixer2_fake(
761
761
output : torch .Tensor ,
762
762
layer_name : str ,
763
763
mup_vector : Optional [torch .Tensor ] = None ,
764
- ):
764
+ ) -> None :
765
765
return
766
766
767
767
You can’t perform that action at this time.
0 commit comments