We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent e4c3aca commit 220ceb2Copy full SHA for 220ceb2
lmdeploy/pytorch/engine/devices/ascend.py
@@ -17,7 +17,7 @@ def update_step_context(cls, step_context):
17
single_attention_mask = torch.logical_not(
18
torch.tril(
19
torch.ones(step_context.q_seq_length[i],
20
- (step_context.kv_seq_length[i] + 31) & (~31),
+ step_context.block_offsets.shape[1] * block_size,
21
dtype=torch.bool).cuda(),
22
diagonal=step_context.kv_seq_length[i] -
23
step_context.q_seq_length[i],
0 commit comments