CUDA: fix race conditions FlashAttention kernels (#13438)
This commit is contained in:
parent
d2a4ef05c6
commit
0208355f42
2 changed files with 3 additions and 0 deletions
|
|
@ -874,6 +874,8 @@ static __device__ __forceinline__ void flash_attn_ext_f16_process_tile(
|
|||
}
|
||||
}
|
||||
|
||||
__syncthreads();
|
||||
|
||||
// Write back combined meta data:
|
||||
#pragma unroll
|
||||
for (int imeta = 0; imeta < nmeta; ++imeta) {
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue