We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 2d654fd commit 052069bCopy full SHA for 052069b
lmdeploy/pytorch/engine/devices/ascend.py
@@ -17,7 +17,7 @@ def update_step_context(cls, step_context):
17
single_attention_mask = torch.logical_not(
18
torch.tril(
19
torch.ones(step_context.q_seq_length[i],
20
- step_context.kv_seq_length[i],
+ (step_context.kv_seq_length[i] + 31) & (~31),
21
dtype=torch.bool).cuda(),
22
diagonal=step_context.kv_seq_length[i] -
23
step_context.q_seq_length[i],
0 commit comments