From 4f1b128abf591469fa809ed1ba2295bbcc9ff631 Mon Sep 17 00:00:00 2001 From: zjjlivein <1834792141@qq.com> Date: Tue, 14 Apr 2026 14:52:21 +0800 Subject: [PATCH] disable qwen2 --- scripts/regression/config.yaml | 51 ------------------------------- scripts/regression/test_models.py | 2 +- 2 files changed, 1 insertion(+), 52 deletions(-) diff --git a/scripts/regression/config.yaml b/scripts/regression/config.yaml index 0749d8abbaf..e4202995cef 100644 --- a/scripts/regression/config.yaml +++ b/scripts/regression/config.yaml @@ -193,57 +193,6 @@ glm_moe: - [10564, 10564, 102954, 47231, 47231, 47231, 47231, 47231, 47231, 47231] dpo_full_function_call_excepted_result: - [10564, 10564, 102954, 47231, 47231, 47231, 47231, 47231, 47231, 47231] -qwen2: - repo_id: PaddleFormers/tiny-random-qwen2v2 - model_type: text - cli_args: - save_checkpoint_format: flex_checkpoint - load_checkpoint_format: flex_checkpoint - template: qwen - base_loss: - sft_full_loss: 12.751743795 - sft_full_resume_loss: 12.75366116 - dpo_full_loss: 0.69314718 - dpo_full_resume_loss: 0.69314718 - pt_full_loss: 12.9430995 - pt_full_resume_loss: 12.94307518 - sft_lora_loss: 12.751743795 - sft_lora_resume_loss: 12.75366116 - dpo_lora_loss: 0.69314718 - dpo_lora_resume_loss: 0.69314718 - pt_lora_loss: 12.9430995 - pt_lora_resume_loss: 12.94307518 - sft_full_tp_pp_loss: 12.835041045 - sft_full_tp_pp_resume_loss: 12.83503151 - dpo_full_tp_pp_loss: 0.69314718 - dpo_full_tp_pp_resume_loss: 0.69314718 - pt_full_tp_pp_loss: 12.943032265 - pt_full_tp_pp_resume_loss: 12.94280052 - sft_lora_tp_pp_loss: 12.83521986 - sft_lora_tp_pp_resume_loss: 12.83525467 - dpo_lora_tp_pp_loss: 0.69314718 - dpo_lora_tp_pp_resume_loss: 0.69314718 - pt_lora_tp_pp_loss: 12.943032265 - pt_lora_tp_pp_resume_loss: 12.94280052 - sft_full_function_call_loss: 12.800441265 - sft_full_function_call_resume_loss: 12.80294609 - dpo_full_function_call_loss: 0.69314718 - dpo_full_function_call_resume_loss: 0.69314718 - base_result: - pt_full_excepted_result: [] - sft_full_excepted_result: [] - dpo_full_excepted_result: [] - pt_lora_excepted_result: [] - sft_lora_excepted_result: [] - dpo_lora_excepted_result: [] - pt_full_tp_pp_excepted_result: [] - sft_full_tp_pp_excepted_result: [] - dpo_full_tp_pp_excepted_result: [] - pt_lora_tp_pp_excepted_result: [] - sft_lora_tp_pp_excepted_result: [] - dpo_lora_tp_pp_excepted_result: [] - sft_full_function_call_excepted_result: [] - dpo_full_function_call_excepted_result: [] llama: repo_id: PaddleFormers/tiny-random-llama3 model_type: text diff --git a/scripts/regression/test_models.py b/scripts/regression/test_models.py index b15274c3f15..8f70d9f1f4f 100644 --- a/scripts/regression/test_models.py +++ b/scripts/regression/test_models.py @@ -600,7 +600,7 @@ def execute_training_workflow( # Test model generation generate_log_file = os.path.join(LOG_PATH, f"{model_key}_{train_type}_{test_type}_generate.log") - skip_generation = model_key in ["qwen2_moe", "deepseek_v3"] + skip_generation = model_key in ["qwen2", "qwen2_moe", "deepseek_v3"] if skip_generation: result = None else: