Skip to content

Commit 065eef7

Browse files
authored
[paddle inference ,mmha]head_dim case 10/26 were deleted, because they would trigger misalignedAddr cudaError (#64081)
1 parent 97c4036 commit 065eef7

File tree

1 file changed

+0
-9
lines changed

1 file changed

+0
-9
lines changed

paddle/phi/kernels/fusion/gpu/masked_multihead_attention_kernel.cu

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -709,18 +709,10 @@ void fmha_impl(const phi::GPUContext &dev_ctx,
709709
LoadFunc load_func,
710710
StoreFunc store_func) {
711711
switch (dim_head) {
712-
case 10:
713-
fmha_launch_kernel<T, 10, 32>(
714-
params, dev_ctx.stream(), load_func, store_func);
715-
break;
716712
case 16:
717713
fmha_launch_kernel<T, 16, 32>(
718714
params, dev_ctx.stream(), load_func, store_func);
719715
break;
720-
case 26:
721-
fmha_launch_kernel<T, 26, 32>(
722-
params, dev_ctx.stream(), load_func, store_func);
723-
break;
724716
case 32:
725717
fmha_launch_kernel<T, 32, 32>(
726718
params, dev_ctx.stream(), load_func, store_func);
@@ -729,7 +721,6 @@ void fmha_impl(const phi::GPUContext &dev_ctx,
729721
fmha_launch_kernel<T, 64, 64>(
730722
params, dev_ctx.stream(), load_func, store_func);
731723
break;
732-
// for opt model
733724
case 80:
734725
fmha_launch_kernel<T, 80, 128>(
735726
params, dev_ctx.stream(), load_func, store_func);

0 commit comments

Comments
 (0)