Skip to content

Commit 8561cee

Browse files
committed
first commit
1 parent 5bddd3e commit 8561cee

File tree

3 files changed

+0
-3
lines changed

3 files changed

+0
-3
lines changed

tests/layers/test_append_attention.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -628,7 +628,6 @@ def cmp_append_attention(self, naive_cache_k=None, naive_cache_v=None, attn_mask
628628
12,
629629
(self.q_num_head + 2 * self.kv_num_head) // self.kv_num_head,
630630
self.blocksize,
631-
speculate_max_draft_token_num + 1,
632631
)
633632
if self.use_dynamic_quant:
634633
cache_quant_type = "block_wise_fp8"

tests/layers/test_append_attention_with_output.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -479,7 +479,6 @@ def cmp_append_attention(self, naive_cache_k=None, naive_cache_v=None, attn_mask
479479
12,
480480
(self.q_num_head + 2 * self.kv_num_head) // self.kv_num_head,
481481
self.blocksize,
482-
speculate_max_draft_token_num + 1,
483482
)
484483

485484
# Warm up

tests/operators/test_tree_mask.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -254,7 +254,6 @@ def run_append_c16_attention(
254254
decoder_block_shape_q,
255255
self.num_q_head // self.num_kv_head,
256256
self.block_size,
257-
decoder_step_token_num,
258257
)
259258
s_time = 0
260259
for i in range(self.run_time + self.warm_up):

0 commit comments

Comments
 (0)