Skip to content

Commit af49b41

Browse files
committed
Fix test
Signed-off-by: Chengji Yao <[email protected]>
1 parent e22c472 commit af49b41

File tree

1 file changed

+1
-3
lines changed

1 file changed

+1
-3
lines changed

tests/layers/vllm/test_attention.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -30,9 +30,7 @@
3030
# Number of attention heads (Key/Value) - for Grouped-Query Attention
3131
NUM_KV_HEADS = 4
3232
# Dimension of each attention head
33-
HEAD_DIM = 64
34-
# Padded head dimension
35-
PADDED_HEAD_DIM = 64
33+
HEAD_DIM = 128
3634
# Total number of blocks in the KV cache
3735
NUM_BLOCKS = 32
3836
# Number of tokens per block

0 commit comments

Comments
 (0)