We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 97c4036 commit 065eef7Copy full SHA for 065eef7
paddle/phi/kernels/fusion/gpu/masked_multihead_attention_kernel.cu
@@ -709,18 +709,10 @@ void fmha_impl(const phi::GPUContext &dev_ctx,
709
LoadFunc load_func,
710
StoreFunc store_func) {
711
switch (dim_head) {
712
- case 10:
713
- fmha_launch_kernel<T, 10, 32>(
714
- params, dev_ctx.stream(), load_func, store_func);
715
- break;
716
case 16:
717
fmha_launch_kernel<T, 16, 32>(
718
params, dev_ctx.stream(), load_func, store_func);
719
break;
720
- case 26:
721
- fmha_launch_kernel<T, 26, 32>(
722
723
724
case 32:
725
fmha_launch_kernel<T, 32, 32>(
726
@@ -729,7 +721,6 @@ void fmha_impl(const phi::GPUContext &dev_ctx,
729
fmha_launch_kernel<T, 64, 64>(
730
731
732
- // for opt model
733
case 80:
734
fmha_launch_kernel<T, 80, 128>(
735
0 commit comments