From 8ddc56a0bb5b87868056979deee8a9125aab28c0 Mon Sep 17 00:00:00 2001 From: "hongliang.yuan" Date: Mon, 10 Nov 2025 17:06:35 +0800 Subject: [PATCH 1/2] fix vllm train ci --- tests/run_test.py | 1 + 1 file changed, 1 insertion(+) diff --git a/tests/run_test.py b/tests/run_test.py index 3a7a9e05e..c79a2f0ea 100644 --- a/tests/run_test.py +++ b/tests/run_test.py @@ -345,6 +345,7 @@ def run_llm_testcase(model): sout = r.stdout prec = "fp16" metrics = {} + epoch_matches = {} if is_firefly: for match in pattern.finditer(sout): key = match.group(1).strip() -- Gitee From af62f2b957d351d980bac7c4be90c723c0c95ed2 Mon Sep 17 00:00:00 2001 From: "hongliang.yuan" Date: Mon, 10 Nov 2025 17:23:01 +0800 Subject: [PATCH 2/2] fix ci --- models/nlp/llm/chatglm3-6b/pytorch/requirements.txt | 5 ++--- tests/run_test.py | 4 ++-- 2 files changed, 4 insertions(+), 5 deletions(-) diff --git a/models/nlp/llm/chatglm3-6b/pytorch/requirements.txt b/models/nlp/llm/chatglm3-6b/pytorch/requirements.txt index 918c116c9..2d4e1e2a3 100644 --- a/models/nlp/llm/chatglm3-6b/pytorch/requirements.txt +++ b/models/nlp/llm/chatglm3-6b/pytorch/requirements.txt @@ -3,11 +3,10 @@ ruamel_yaml>=0.18.6 rouge_chinese>=1.0.3 jupyter>=1.0.0 datasets>=2.18.0 -peft>=0.10.0 +peft==0.11.1 mpi4py>=3.1.5 transformers==4.40.0 accelerate>=0.33.0 typer>=0.9.0 nltk -sentencepiece - +sentencepiece \ No newline at end of file diff --git a/tests/run_test.py b/tests/run_test.py index c79a2f0ea..64b88d348 100644 --- a/tests/run_test.py +++ b/tests/run_test.py @@ -317,14 +317,14 @@ def run_llm_testcase(model): # {'train_runtime': 84.0969, 'train_samples_per_second': 2.378, 'train_steps_per_second': 1.189, 'train_loss': 0.24943359375, 'epoch': 0.0} pattern = r"({.*?})" prepare_script = f""" - cd ../../{model_path} + cd ../{model_path} pip3 install -r requirements.txt mkdir -p data ln -s /mnt/deepspark/data/datasets/AdvertiseGen data/ python3 process_data.py mkdir -p checkpoint ln -s /mnt/deepspark/data/checkpoints/chatglm3-6b checkpoint/ - timeout 1800 bash run.sh configs/lora.yaml 1 + bash run.sh configs/lora.yaml 4 """ elif is_megatron_deepspeed: # 选择使用llama2-7b作为个例 -- Gitee