Skip to content

Commit ac29bc4

Browse files
committed
FLAGS_CUTLASS_FP8_GEMM=True
1 parent 1bde9b8 commit ac29bc4

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

paddlenlp/experimental/transformers/fused_transformer_layers.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@
4343
from paddlenlp_ops import rebuild_padding_v2
4444

4545
if core.is_compiled_with_cuda():
46-
if os.getenv("FLAGS_CUTLASS_FP8_GEMM", "False") == "True":
46+
if os.getenv("FLAGS_CUTLASS_FP8_GEMM", "True") == "True":
4747
logger.info("cutlass fp8 gemm is used. you can turn it off by setting FLAGS_CUTLASS_FP8_GEMM to False.")
4848
from paddlenlp_ops import (
4949
cutlass_fp8_fp8_fp8_dual_gemm_fused as fp8_dual_gemm_fused,
@@ -76,7 +76,7 @@
7676

7777

7878
def use_cutlass_fp8_gemm():
79-
return os.getenv("FLAGS_CUTLASS_FP8_GEMM", "False") == "True"
79+
return os.getenv("FLAGS_CUTLASS_FP8_GEMM", "True") == "True"
8080

8181

8282
# for distributed tensor model parallel

0 commit comments

Comments
 (0)