Skip to content

Commit 6943cc9

Browse files
update
Signed-off-by: Congcong Chen <congcongchen@microsoft.com>
1 parent bfa3a19 commit 6943cc9

File tree

1 file changed

+0
-10
lines changed

1 file changed

+0
-10
lines changed

vllm/attention/backends/differential_flash_attn.py

Lines changed: 0 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -678,7 +678,6 @@ def __init__(
678678
self.attn_type = attn_type
679679

680680
self.lambda_full = None
681-
# self.subln = nn.RMSNorm(2 * self.head_size, eps=1e-5, elementwise_affine=True)
682681
self.subln = self.differential_flash_attention_config["subln"]
683682

684683
def split_heads(self, x):
@@ -705,9 +704,6 @@ def populate_kv_cache(self,
705704
if (kv_cache.numel() > 0):
706705
if (key is not None) and (value is not None):
707706
updated_slot_mapping = attn_metadata.slot_mapping
708-
# previous_key_cache_sum = key_cache.sum()
709-
# previous_value_cache_sum = value_cache.sum()
710-
711707
torch.ops._C_cache_ops.reshape_and_cache_flash(
712708
key,
713709
value,
@@ -718,12 +714,6 @@ def populate_kv_cache(self,
718714
layer._k_scale,
719715
layer._v_scale,
720716
)
721-
# assert key_cache.sum() - previous_key_cache_sum == key.sum(), "key_cache sum mismatch"
722-
# assert value_cache.sum() - previous_value_cache_sum == value.sum(), "value_cache sum mismatch"
723-
# if key_cache.sum() - previous_key_cache_sum != key.sum():
724-
# print("key_cache sum mismatch")
725-
# if value_cache.sum() - previous_value_cache_sum != value.sum():
726-
# print("value_cache sum mismatch")
727717

728718
def forward_generate_kv_cache(
729719
self,

0 commit comments

Comments
 (0)