Skip to content

Commit

Permalink
[FFPA] fix some macro typos (#21)
Browse files Browse the repository at this point in the history
* Update faster_prefill_attn_F16F16F16F16_L1.cu

* Update faster_prefill_attn_F32F16F16F32_L1.cu
  • Loading branch information
DefTruth authored Jan 9, 2025
1 parent 7469308 commit e1b3bbc
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 4 deletions.
4 changes: 2 additions & 2 deletions csrc/cuffpa/faster_prefill_attn_F16F16F16F16_L1.cu
Original file line number Diff line number Diff line change
Expand Up @@ -450,7 +450,7 @@ void ffpa_mma_acc_f16_L1(torch::Tensor Q,
}

#else
#ifdef ENBALE_FFPA_ALL_HEADDIM
#ifdef ENABLE_FFPA_ALL_HEADDIM
// multiple of 32
#define DISPATCH_KERNEL_F16_L1_HEADDIM(S) \
{ \
Expand Down Expand Up @@ -523,7 +523,7 @@ void ffpa_mma_acc_f16_L1(torch::Tensor Q,

#endif

#ifdef ENBALE_FFPA_ALL_STAGES
#ifdef ENABLE_FFPA_ALL_STAGES
// dispatch stages
if (stages == 2) {
DISPATCH_KERNEL_F16_L1_HEADDIM(2);
Expand Down
4 changes: 2 additions & 2 deletions csrc/cuffpa/faster_prefill_attn_F32F16F16F32_L1.cu
Original file line number Diff line number Diff line change
Expand Up @@ -452,7 +452,7 @@ void ffpa_mma_acc_f32_L1(torch::Tensor Q,
}

#else
#ifdef ENBALE_FFPA_ALL_HEADDIM
#ifdef ENABLE_FFPA_ALL_HEADDIM
// multiple of 32
#define DISPATCH_KERNEL_F32_L1_HEADDIM(S) \
{ \
Expand Down Expand Up @@ -525,7 +525,7 @@ void ffpa_mma_acc_f32_L1(torch::Tensor Q,

#endif

#ifdef ENBALE_FFPA_ALL_STAGES
#ifdef ENABLE_FFPA_ALL_STAGES
// dispatch stages
if (stages == 2) {
DISPATCH_KERNEL_F32_L1_HEADDIM(2);
Expand Down

0 comments on commit e1b3bbc

Please sign in to comment.