Skip to content

Commit 2d46035

Browse files
[ci] testcase bugfix and add more models into testcase (#3463)
* update * update * update * update * update * update * update * updaste * update * update * update * update * update * update * update * update * update * update * update
1 parent 84507b7 commit 2d46035

15 files changed

+146
-55
lines changed

.github/scripts/eval_chat_config.py

Lines changed: 18 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,8 @@
4747
models as lmdeploy_internlm2_chat_7b # noqa: F401, E501
4848
from opencompass.configs.models.hf_internlm.lmdeploy_internlm2_chat_20b import \
4949
models as lmdeploy_internlm2_chat_20b # noqa: F401, E501
50+
from opencompass.configs.models.hf_internlm.lmdeploy_internlm3_8b_instruct import \
51+
models as lmdeploy_internlm3_8b_instruct # noqa: F401, E501
5052
from opencompass.configs.models.hf_internlm.lmdeploy_internlm_chat_7b import \
5153
models as lmdeploy_internlm_chat_7b # noqa: F401, E501
5254
from opencompass.configs.models.hf_llama.hf_llama2_7b_chat import models as hf_llama2_chat_7b # noqa: F401, E501
@@ -66,6 +68,8 @@
6668
models as hf_mixtral_chat_8x7b # noqa: F401, E501
6769
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_7b_instruct import \
6870
models as lmdeploy_qwen2_5_7b_instruct # noqa: F401, E501
71+
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_32b_instruct import \
72+
models as lmdeploy_qwen2_5_32b_instruct # noqa: F401, E501
6973
from opencompass.configs.models.qwen.hf_qwen1_5_7b_chat import models as hf_qwen1_5_chat_7b # noqa: F401, E501
7074
from opencompass.configs.models.qwen.hf_qwen1_5_moe_a2_7b_chat import \
7175
models as hf_qwen1_5_moe_a2_7b_chat # noqa: F401, E501
@@ -116,15 +120,12 @@
116120
turbomind_internlm2_5_7b_chat_batch1 = deepcopy(*lmdeploy_internlm2_5_7b_chat)
117121
turbomind_internlm2_5_7b_chat_batch1_4bits = deepcopy(*lmdeploy_internlm2_5_7b_chat)
118122

119-
turbomind_internlm3_8b_instruct = deepcopy(*lmdeploy_internlm2_5_7b_chat)
120-
turbomind_internlm3_8b_instruct_4bits = deepcopy(*lmdeploy_internlm2_5_7b_chat)
121-
turbomind_internlm3_8b_instruct_kvint4 = deepcopy(*lmdeploy_internlm2_5_7b_chat)
122-
turbomind_internlm3_8b_instruct_kvint8 = deepcopy(*lmdeploy_internlm2_5_7b_chat)
123-
pytorch_internlm3_8b_instruct = deepcopy(*lmdeploy_internlm2_5_7b_chat)
124-
pytorch_internlm3_8b_instruct_w8a8 = deepcopy(*lmdeploy_internlm2_5_7b_chat)
125-
for model in [v for k, v in locals().items() if 'internlm3_8b_instruct' in k]:
126-
model['abbr'] = 'turbomind-internlm3-8b-instruct'
127-
model['path'] = 'internlm/internlm3-8b-instruct'
123+
turbomind_internlm3_8b_instruct = deepcopy(*lmdeploy_internlm3_8b_instruct)
124+
turbomind_internlm3_8b_instruct_4bits = deepcopy(*lmdeploy_internlm3_8b_instruct)
125+
turbomind_internlm3_8b_instruct_kvint4 = deepcopy(*lmdeploy_internlm3_8b_instruct)
126+
turbomind_internlm3_8b_instruct_kvint8 = deepcopy(*lmdeploy_internlm3_8b_instruct)
127+
pytorch_internlm3_8b_instruct = deepcopy(*lmdeploy_internlm3_8b_instruct)
128+
pytorch_internlm3_8b_instruct_w8a8 = deepcopy(*lmdeploy_internlm3_8b_instruct)
128129

129130
# ===== Configs for internlm/internlm2_5_20b_chat =====
130131
turbomind_internlm2_5_20b_chat = deepcopy(*lmdeploy_internlm2_5_20b_chat)
@@ -186,6 +187,14 @@
186187
pytorch_qwen2_5_7b_instruct = deepcopy(*lmdeploy_qwen2_5_7b_instruct)
187188
pytorch_qwen2_5_7b_instruct_w8a8 = deepcopy(*lmdeploy_qwen2_5_7b_instruct)
188189

190+
# ===== Configs for Qwen/Qwen25-32B-Instruct =====
191+
turbomind_qwen2_5_32b_instruct = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
192+
turbomind_qwen2_5_32b_instruct_4bits = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
193+
turbomind_qwen2_5_32b_instruct_kvint4 = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
194+
turbomind_qwen2_5_32b_instruct_kvint8 = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
195+
pytorch_qwen2_5_32b_instruct = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
196+
pytorch_qwen2_5_32b_instruct_w8a8 = deepcopy(*lmdeploy_qwen2_5_32b_instruct)
197+
189198
# ===== Configs for meta-llama/Llama-2-7b-chat-hf =====
190199
turbomind_llama2_7b_chat = deepcopy(*lmdeploy_llama2_7b_chat)
191200
turbomind_llama2_7b_chat_4bits = deepcopy(*lmdeploy_llama2_7b_chat)

.github/scripts/eval_regression_chat_models.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -58,6 +58,8 @@
5858
models as lmdeploy_qwen2_5_3b_instruct_model # noqa: F401, E501
5959
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_14b_instruct import \
6060
models as lmdeploy_qwen2_5_14b_instruct_model # noqa: F401, E501
61+
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_32b_instruct import \
62+
models as lmdeploy_qwen2_5_32b_instruct_model # noqa: F401, E501
6163
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_72b_instruct import \
6264
models as lmdeploy_qwen2_5_72b_instruct_model # noqa: F401, E501
6365
from opencompass.configs.models.qwen.lmdeploy_qwen2_1_5b_instruct import \
@@ -86,6 +88,7 @@
8688
pytorch_mistral_nemo_instruct_2407_model = deepcopy(lmdeploy_mistral_nemo_instruct_2407_model)
8789
pytorch_mistral_small_instruct_2409_model = deepcopy(lmdeploy_mistral_small_instruct_2409_model)
8890
pytorch_qwen2_5_72b_instruct_model = deepcopy(lmdeploy_qwen2_5_72b_instruct_model)
91+
pytorch_qwen2_5_32b_instruct_model = deepcopy(lmdeploy_qwen2_5_32b_instruct_model)
8992
pytorch_qwen2_7b_instruct_model = deepcopy(lmdeploy_qwen2_7b_instruct_model)
9093
pytorch_yi_1_5_34b_chat_model = deepcopy(lmdeploy_yi_1_5_34b_chat_model)
9194

@@ -106,6 +109,7 @@
106109
lmdeploy_nemotron_70b_instruct_hf_model_native = deepcopy(lmdeploy_nemotron_70b_instruct_hf_model)
107110
lmdeploy_qwen2_5_0_5b_instruct_model_native = deepcopy(lmdeploy_qwen2_5_0_5b_instruct_model)
108111
lmdeploy_qwen2_5_14b_instruct_model_native = deepcopy(lmdeploy_qwen2_5_14b_instruct_model)
112+
lmdeploy_qwen2_5_32b_instruct_model_native = deepcopy(lmdeploy_qwen2_5_32b_instruct_model)
109113
lmdeploy_qwen2_5_72b_instruct_model_native = deepcopy(lmdeploy_qwen2_5_72b_instruct_model)
110114
lmdeploy_qwen2_7b_instruct_model_native = deepcopy(lmdeploy_qwen2_7b_instruct_model)
111115
lmdeploy_yi_1_5_6b_chat_model_native = deepcopy(lmdeploy_yi_1_5_6b_chat_model)

.github/workflows/daily_ete_test.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -196,7 +196,7 @@ jobs:
196196
if: ${{!cancelled() && (github.event_name == 'schedule' || contains(fromJSON(github.event.inputs.regression_func), 'tools'))}}
197197
runs-on: [self-hosted, linux-a100]
198198
needs: test_quantization
199-
timeout-minutes: 240
199+
timeout-minutes: 300
200200
strategy:
201201
fail-fast: false
202202
matrix:

.github/workflows/evaluate.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ on:
1717
required: true
1818
description: 'Tested TurboMind models list. eg. [internlm_chat_7b,internlm_chat_7b_w8a16]'
1919
type: string
20-
default: '[turbomind_internlm2_chat_7b, pytorch_internlm2_chat_7b, turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_internlm2_chat_20b, pytorch_internlm2_chat_20b, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_7b_instruct, pytorch_qwen2_7b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it, turbomind_internlm2_chat_7b_4bits, turbomind_internlm2_chat_7b_kvint4, turbomind_internlm2_chat_7b_kvint8, turbomind_internlm2_5_7b_chat_4bits, turbomind_internlm2_5_7b_chat_kvint4, turbomind_internlm2_5_7b_chat_kvint8, pytorch_internlm2_5_7b_chat_w8a8, turbomind_internlm3_8b_instruct_4bits, turbomind_internlm3_8b_instruct_kvint4, turbomind_internlm3_8b_instruct_kvint8, pytorch_internlm3_8b_instruct_w8a8, turbomind_internlm2_5_20b_chat_4bits, turbomind_internlm2_5_20b_chat_kvint4, turbomind_internlm2_5_20b_chat_kvint8, turbomind_llama3_8b_instruct_4bits, turbomind_llama3_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_4bits, turbomind_llama3_1_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_kvint8,turbomind_llama3_8b_instruct_kvint8, pytorch_llama3_1_8b_instruct_w8a8, turbomind_qwen2_7b_instruct_4bits, turbomind_qwen2_7b_instruct_kvint8, pytorch_qwen2_7b_instruct_w8a8, turbomind_qwen2_5_7b_instruct_4bits, turbomind_qwen2_5_7b_instruct_kvint8, pytorch_qwen2_5_7b_instruct_w8a8, turbomind_llama2_7b_chat_4bits, turbomind_llama2_7b_chat_kvint4, turbomind_llama2_7b_chat_kvint8]'
20+
default: '[turbomind_internlm2_chat_7b, pytorch_internlm2_chat_7b, turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_internlm2_chat_20b, pytorch_internlm2_chat_20b, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_7b_instruct, pytorch_qwen2_7b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_qwen2_5_32b_instruct, pytorch_qwen2_5_32b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it, turbomind_internlm2_chat_7b_4bits, turbomind_internlm2_chat_7b_kvint4, turbomind_internlm2_chat_7b_kvint8, turbomind_internlm2_5_7b_chat_4bits, turbomind_internlm2_5_7b_chat_kvint4, turbomind_internlm2_5_7b_chat_kvint8, pytorch_internlm2_5_7b_chat_w8a8, turbomind_internlm3_8b_instruct_4bits, turbomind_internlm3_8b_instruct_kvint4, turbomind_internlm3_8b_instruct_kvint8, pytorch_internlm3_8b_instruct_w8a8, turbomind_internlm2_5_20b_chat_4bits, turbomind_internlm2_5_20b_chat_kvint4, turbomind_internlm2_5_20b_chat_kvint8, turbomind_llama3_8b_instruct_4bits, turbomind_llama3_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_4bits, turbomind_llama3_1_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_kvint8,turbomind_llama3_8b_instruct_kvint8, pytorch_llama3_1_8b_instruct_w8a8, turbomind_qwen2_7b_instruct_4bits, turbomind_qwen2_7b_instruct_kvint8, pytorch_qwen2_7b_instruct_w8a8, turbomind_qwen2_5_7b_instruct_4bits, turbomind_qwen2_5_7b_instruct_kvint8, pytorch_qwen2_5_7b_instruct_w8a8, turbomind_qwen2_5_32b_instruct_4bits, turbomind_qwen2_5_32b_instruct_kvint8, pytorch_qwen2_5_32b_instruct_w8a8,turbomind_llama2_7b_chat_4bits, turbomind_llama2_7b_chat_kvint4, turbomind_llama2_7b_chat_kvint8]'
2121
chat_datasets:
2222
required: true
2323
description: 'Tested datasets list. eg. [*bbh_datasets,*ceval_datasets,*cmmlu_datasets,*GaokaoBench_datasets,*gpqa_datasets,*gsm8k_datasets,*hellaswag_datasets,*humaneval_datasets,*ifeval_datasets,*math_datasets,*sanitized_mbpp_datasets,*mmlu_datasets,*nq_datasets,*race_datasets,*TheoremQA_datasets,*triviaqa_datasets,*winogrande_datasets,*crowspairs_datasets]'

.github/workflows/evaluate_remote.yml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,6 @@ jobs:
8484
matrix:
8585
pyver: [py310]
8686
runs-on: ubuntu-latest
87-
environment: 'prod'
8887
env:
8988
PYTHON_VERSION: ${{ matrix.pyver }}
9089
PLAT_NAME: manylinux2014_x86_64
@@ -118,7 +117,6 @@ jobs:
118117
if: ${{!cancelled()}}
119118
needs: ['build-pypi', 'build-pypi-lmdeploy']
120119
runs-on: lmdeploy-volc
121-
environment: 'prod'
122120
timeout-minutes: 120 #2hours
123121
steps:
124122
- name: Clone repository
@@ -148,6 +146,7 @@ jobs:
148146
pip install -r /fs-computility/llm/qa-llm-cicd/config/requirements.txt --cache-dir ${{env.PIP_CACHE_PATH}}
149147
pip install opencompass*.whl --cache-dir ${{env.PIP_CACHE_PATH}}
150148
pip install opencompass[lmdeploy] --cache-dir ${{env.PIP_CACHE_PATH}}
149+
pip install opencompass[full] --cache-dir ${{env.PIP_CACHE_PATH}}
151150
pip install torch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 --cache-dir ${{env.PIP_CACHE_PATH}}
152151
FLASH_ATTENTION_FORCE_BUILD=TRUE pip install /fs-computility/llm/qa-llm-cicd/packages/flash_attn-2.7.0.post2+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
153152
cp -r /root/nltk_data ${{env.CONDA_PATH}}/envs/${{env.CONDA_ENV}}/nltk_data
@@ -179,7 +178,6 @@ jobs:
179178
regression_func: ${{fromJSON(github.event.inputs.regression_func_volc || '["chat_models","base_models"]')}}
180179
backend: ${{ fromJSON(github.event.inputs.backend || '["turbomind", "pytorch"]')}}
181180
runs-on: lmdeploy-volc
182-
environment: 'prod'
183181
timeout-minutes: 480 #6hours
184182
steps:
185183
- name: Checkout repository

autotest/config.yaml

Lines changed: 14 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,8 @@ tp_config:
2727
DeepSeek-V2-Lite-Chat: 2
2828
DeepSeek-R1-Distill-Qwen-32B: 2
2929
deepseek-vl2: 2
30+
InternVL3-38B: 2
31+
Qwen2.5-32B-Instruct: 4
3032
Llama-3.2-11B-Vision-Instruct: 2 # set tp == 2 for test
3133

3234
turbomind_chat_model:
@@ -44,6 +46,7 @@ turbomind_chat_model:
4446
- internlm/internlm2-chat-20b-4bits
4547
- internlm/internlm-xcomposer2-4khd-7b
4648
- internlm/internlm-xcomposer2d5-7b
49+
- OpenGVLab/InternVL3-38B
4750
- OpenGVLab/InternVL2_5-26B-MPO
4851
- OpenGVLab/InternVL2_5-1B
4952
- OpenGVLab/InternVL2_5-8B
@@ -66,6 +69,7 @@ turbomind_chat_model:
6669
- Qwen/Qwen-VL-Chat
6770
- Qwen/Qwen2.5-0.5B-Instruct
6871
- Qwen/Qwen2.5-7B-Instruct
72+
- Qwen/Qwen2.5-32B-Instruct
6973
- Qwen/Qwen2.5-72B-Instruct
7074
- Qwen/Qwen2-7B-Instruct-GPTQ-Int4
7175
- Qwen/Qwen2-57B-A14B-Instruct-GPTQ-Int4
@@ -85,6 +89,7 @@ turbomind_chat_model:
8589
- deepseek-ai/DeepSeek-V2-Lite-Chat
8690
- codellama/CodeLlama-7b-Instruct-hf
8791
- THUDM/glm-4-9b-chat
92+
- THUDM/codegeex4-all-9b
8893
- openbmb/MiniCPM-Llama3-V-2_5
8994
- openbmb/MiniCPM-V-2_6
9095
- allenai/Molmo-7B-D-0924
@@ -101,6 +106,7 @@ pytorch_chat_model:
101106
- internlm/internlm2_5-7b-chat
102107
- internlm/internlm2_5-20b-chat
103108
- internlm/internlm2-chat-20b
109+
- OpenGVLab/InternVL3-38B
104110
- OpenGVLab/InternVL2_5-26B-MPO
105111
- OpenGVLab/InternVL2_5-1B
106112
- OpenGVLab/InternVL2_5-8B
@@ -121,6 +127,7 @@ pytorch_chat_model:
121127
- Qwen/Qwen2-1.5B-Instruct
122128
- Qwen/Qwen2.5-0.5B-Instruct
123129
- Qwen/Qwen2.5-7B-Instruct
130+
- Qwen/Qwen2.5-32B-Instruct
124131
- Qwen/Qwen2.5-72B-Instruct
125132
- Qwen/Qwen2.5-VL-7B-Instruct
126133
- Qwen/Qwen1.5-7B-Chat
@@ -137,12 +144,11 @@ pytorch_chat_model:
137144
- deepseek-ai/deepseek-moe-16b-chat
138145
- deepseek-ai/deepseek-coder-1.3b-instruct
139146
- deepseek-ai/DeepSeek-V2-Lite-Chat
140-
- deepseek-ai/deepseek-vl2-tiny
141-
- deepseek-ai/deepseek-vl2
142147
- THUDM/chatglm2-6b
143148
- THUDM/cogvlm2-llama3-chinese-chat-19B
144149
- THUDM/glm-4v-9b
145150
- THUDM/glm-4-9b-chat
151+
- THUDM/codegeex4-all-9b
146152
- openbmb/MiniCPM-V-2_6
147153
- microsoft/Phi-3-mini-4k-instruct
148154
- microsoft/Phi-3-vision-128k-instruct
@@ -153,6 +159,7 @@ turbomind_vl_model:
153159
- liuhaotian/llava-v1.6-vicuna-7b
154160
- 01-ai/Yi-VL-6B
155161
- deepseek-ai/deepseek-vl-1.3b-chat
162+
- OpenGVLab/InternVL3-38B
156163
- OpenGVLab/InternVL2_5-26B-MPO
157164
- OpenGVLab/InternVL-Chat-V1-5
158165
- OpenGVLab/Mini-InternVL-Chat-2B-V1-5
@@ -176,6 +183,7 @@ pytorch_vl_model:
176183
- OpenGVLab/InternVL2_5-26B-MPO
177184
- OpenGVLab/InternVL-Chat-V1-5
178185
- OpenGVLab/Mini-InternVL-Chat-2B-V1-5
186+
- OpenGVLab/InternVL3-38B
179187
- OpenGVLab/InternVL2_5-1B
180188
- OpenGVLab/InternVL2_5-8B
181189
- OpenGVLab/InternVL2_5-26B
@@ -196,8 +204,6 @@ pytorch_vl_model:
196204
- openbmb/MiniCPM-V-2_6
197205
- microsoft/Phi-3-vision-128k-instruct
198206
- microsoft/Phi-3.5-vision-instruct
199-
- deepseek-ai/deepseek-vl2-tiny
200-
- deepseek-ai/deepseek-vl2
201207

202208
turbomind_base_model:
203209
- internlm/internlm2_5-7b
@@ -214,7 +220,6 @@ pytorch_base_model:
214220
turbomind_quatization:
215221
no_awq:
216222
- meta-llama/Meta-Llama-3-1-70B-Instruct
217-
- Qwen/Qwen2.5-72B-Instruct
218223
- Qwen/Qwen1.5-MoE-A2.7B-Chat
219224
- Qwen/Qwen2-VL-2B-Instruct
220225
- Qwen/Qwen2-VL-7B-Instruct
@@ -224,6 +229,7 @@ turbomind_quatization:
224229
- deepseek-ai/DeepSeek-V2-Lite-Chat
225230
- codellama/CodeLlama-7b-Instruct-hf
226231
- allenai/Molmo-7B-D-0924
232+
- THUDM/codegeex4-all-9b
227233
gptq:
228234
- internlm/internlm2_5-7b-chat
229235
no_kvint4:
@@ -240,6 +246,7 @@ turbomind_quatization:
240246
- Qwen/Qwen-VL-Chat
241247
- Qwen/Qwen2.5-0.5B-Instruct
242248
- Qwen/Qwen2.5-7B-Instruct
249+
- Qwen/Qwen2.5-32B-Instruct
243250
- Qwen/Qwen2.5-72B-Instruct
244251
- Qwen/Qwen2-VL-2B-Instruct
245252
- Qwen/Qwen2-VL-7B-Instruct
@@ -300,6 +307,7 @@ pytorch_quatization:
300307
- Qwen/Qwen-VL-Chat
301308
- Qwen/Qwen2.5-0.5B-Instruct
302309
- Qwen/Qwen2.5-7B-Instruct
310+
- Qwen/Qwen2.5-32B-Instruct
303311
- Qwen/Qwen2.5-72B-Instruct
304312
- Qwen/Qwen2-VL-2B-Instruct
305313
- Qwen/Qwen2-VL-7B-Instruct
@@ -330,6 +338,7 @@ benchmark_model:
330338
- internlm/internlm2_5-20b-chat
331339
- THUDM/glm-4-9b-chat
332340
- Qwen/Qwen2.5-7B-Instruct
341+
- Qwen/Qwen2.5-32B-Instruct
333342
- Qwen/Qwen2.5-72B-Instruct
334343
- mistralai/Mistral-7B-Instruct-v0.3
335344
- mistralai/Mixtral-8x7B-Instruct-v0.1

autotest/interface/pipeline/test_pipeline_func.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -424,7 +424,7 @@ def run_pipeline_testcase_bad_words(config, model, backend, file_name):
424424
# test bad_words
425425
gen_config = GenerationConfig(bad_words=[' and', '浦', ' to'])
426426
response = pipe(['Hi, pls intro yourself', 'Shanghai is'], gen_config=gen_config)
427-
result = '蒲' in response[0].text or 'sensetime' in response[0].text
427+
result = '蒲' in response[0].text or 'SenseTime' in response[0].text
428428
for i in range(2):
429429
result &= '浦' not in response[i].text and ' and' not in response[i].text and ' to ' not in response[i].text
430430
save_pipeline_common_log(config, file_name, result, response)
@@ -896,7 +896,7 @@ def test_backend_config_validate_pytorch(config, model, backend, worker_id):
896896
del os.environ['CUDA_VISIBLE_DEVICES']
897897

898898

899-
@pytest.mark.parametrize('model', ['internlm/internlm2_5-20b-chat', 'OpenGVLab/InternVL2_5-26B'])
899+
@pytest.mark.parametrize('model', ['internlm/internlm2_5-20b-chat'])
900900
@pytest.mark.parametrize('backend', [TurbomindEngineConfig])
901901
def test_backend_config_tp(config, model, backend, worker_id):
902902
with pytest.raises(AssertionError):

0 commit comments

Comments
 (0)