File tree Expand file tree Collapse file tree 2 files changed +6
-1
lines changed Expand file tree Collapse file tree 2 files changed +6
-1
lines changed Original file line number Diff line number Diff line change @@ -890,6 +890,7 @@ function llama_dy2st_auto_bs4_bf16_DP1-MP1-PP4-SD2-VPP3_split_bw() {
890
890
891
891
export CUDA_DEVICE_MAX_CONNECTIONS=1
892
892
export PARALLEL_CROSS_ENTROPY=true
893
+ export FLAGS_enable_pir_api=False # 暂时disable pir,后期修复后打开 @卢畅
893
894
894
895
task_name=" llama_dy2st_auto_bs4_bf16_DP1-MP1-PP4-SD2-VPP3_split_bw"
895
896
case_out_dir=" output/$task_name "
@@ -2256,6 +2257,7 @@ function before_hook_for_gpt() {
2256
2257
env | grep FLAGS
2257
2258
export http_proxy=${proxy}
2258
2259
export https_proxy=${proxy}
2260
+ export no_proxy=bcebos.com
2259
2261
if [[ $FLAGS_install_deps == 0 ]]; then
2260
2262
echo -e " \033[31m ---- Install requirements for GPT auto cases \033[0m"
2261
2263
cp requirements.txt requirements_nlp.txt
@@ -2297,6 +2299,7 @@ function before_hook_for_llama() {
2297
2299
env | grep FLAGS
2298
2300
export http_proxy=${proxy}
2299
2301
export https_proxy=${proxy}
2302
+ export no_proxy=bcebos.com
2300
2303
python -m pip install -r $root_path /requirements.txt
2301
2304
python -m pip install -r $root_path /requirements-dev.txt
2302
2305
if [[ ! $FLAGS_download_data =~ " llama" ]]; then
Original file line number Diff line number Diff line change @@ -42,7 +42,7 @@ function gpt_case_list_dygraph(){
42
42
gpt_generation_345M_single
43
43
gpt_generation_345M_hybrid
44
44
gpt_345M_mp8_qat
45
- gpt_export_345M_mp1
45
+ # gpt_export_345M_mp1
46
46
# gpt_export_345M_mp2
47
47
# gpt_export_qat_345M
48
48
# gpt_inference_345M_single
@@ -512,6 +512,7 @@ function before_hook_for_gpt() {
512
512
env | grep FLAGS
513
513
export http_proxy=${proxy}
514
514
export https_proxy=${proxy}
515
+ export no_proxy=bcebos.com
515
516
if [[ $FLAGS_install_deps == 0 ]]; then
516
517
echo -e " \033[31m ---- Install requirements for GPT dygraph cases \033[0m"
517
518
cp requirements.txt requirements_nlp.txt
@@ -619,6 +620,7 @@ function before_hook_for_llm_gpt() {
619
620
env | grep FLAGS
620
621
export http_proxy=${proxy}
621
622
export https_proxy=${proxy}
623
+ export no_proxy=bcebos.com
622
624
python -m pip install -r $root_path /requirements.txt
623
625
python -m pip install -r $root_path /requirements-dev.txt
624
626
if [[ ! $FLAGS_download_data =~ " llm_gpt" ]]; then
You can’t perform that action at this time.
0 commit comments