@@ -27,6 +27,8 @@ tp_config:
2727 DeepSeek-V2-Lite-Chat : 2
2828 DeepSeek-R1-Distill-Qwen-32B : 2
2929 deepseek-vl2 : 2
30+ InternVL3-38B : 2
31+ Qwen2.5-32B-Instruct : 4
3032 Llama-3.2-11B-Vision-Instruct : 2 # set tp == 2 for test
3133
3234turbomind_chat_model :
@@ -44,6 +46,7 @@ turbomind_chat_model:
4446 - internlm/internlm2-chat-20b-4bits
4547 - internlm/internlm-xcomposer2-4khd-7b
4648 - internlm/internlm-xcomposer2d5-7b
49+ - OpenGVLab/InternVL3-38B
4750 - OpenGVLab/InternVL2_5-26B-MPO
4851 - OpenGVLab/InternVL2_5-1B
4952 - OpenGVLab/InternVL2_5-8B
@@ -66,6 +69,7 @@ turbomind_chat_model:
6669 - Qwen/Qwen-VL-Chat
6770 - Qwen/Qwen2.5-0.5B-Instruct
6871 - Qwen/Qwen2.5-7B-Instruct
72+ - Qwen/Qwen2.5-32B-Instruct
6973 - Qwen/Qwen2.5-72B-Instruct
7074 - Qwen/Qwen2-7B-Instruct-GPTQ-Int4
7175 - Qwen/Qwen2-57B-A14B-Instruct-GPTQ-Int4
@@ -85,6 +89,7 @@ turbomind_chat_model:
8589 - deepseek-ai/DeepSeek-V2-Lite-Chat
8690 - codellama/CodeLlama-7b-Instruct-hf
8791 - THUDM/glm-4-9b-chat
92+ - THUDM/codegeex4-all-9b
8893 - openbmb/MiniCPM-Llama3-V-2_5
8994 - openbmb/MiniCPM-V-2_6
9095 - allenai/Molmo-7B-D-0924
@@ -101,6 +106,7 @@ pytorch_chat_model:
101106 - internlm/internlm2_5-7b-chat
102107 - internlm/internlm2_5-20b-chat
103108 - internlm/internlm2-chat-20b
109+ - OpenGVLab/InternVL3-38B
104110 - OpenGVLab/InternVL2_5-26B-MPO
105111 - OpenGVLab/InternVL2_5-1B
106112 - OpenGVLab/InternVL2_5-8B
@@ -121,6 +127,7 @@ pytorch_chat_model:
121127 - Qwen/Qwen2-1.5B-Instruct
122128 - Qwen/Qwen2.5-0.5B-Instruct
123129 - Qwen/Qwen2.5-7B-Instruct
130+ - Qwen/Qwen2.5-32B-Instruct
124131 - Qwen/Qwen2.5-72B-Instruct
125132 - Qwen/Qwen2.5-VL-7B-Instruct
126133 - Qwen/Qwen1.5-7B-Chat
@@ -137,12 +144,11 @@ pytorch_chat_model:
137144 - deepseek-ai/deepseek-moe-16b-chat
138145 - deepseek-ai/deepseek-coder-1.3b-instruct
139146 - deepseek-ai/DeepSeek-V2-Lite-Chat
140- - deepseek-ai/deepseek-vl2-tiny
141- - deepseek-ai/deepseek-vl2
142147 - THUDM/chatglm2-6b
143148 - THUDM/cogvlm2-llama3-chinese-chat-19B
144149 - THUDM/glm-4v-9b
145150 - THUDM/glm-4-9b-chat
151+ - THUDM/codegeex4-all-9b
146152 - openbmb/MiniCPM-V-2_6
147153 - microsoft/Phi-3-mini-4k-instruct
148154 - microsoft/Phi-3-vision-128k-instruct
@@ -153,6 +159,7 @@ turbomind_vl_model:
153159 - liuhaotian/llava-v1.6-vicuna-7b
154160 - 01-ai/Yi-VL-6B
155161 - deepseek-ai/deepseek-vl-1.3b-chat
162+ - OpenGVLab/InternVL3-38B
156163 - OpenGVLab/InternVL2_5-26B-MPO
157164 - OpenGVLab/InternVL-Chat-V1-5
158165 - OpenGVLab/Mini-InternVL-Chat-2B-V1-5
@@ -176,6 +183,7 @@ pytorch_vl_model:
176183 - OpenGVLab/InternVL2_5-26B-MPO
177184 - OpenGVLab/InternVL-Chat-V1-5
178185 - OpenGVLab/Mini-InternVL-Chat-2B-V1-5
186+ - OpenGVLab/InternVL3-38B
179187 - OpenGVLab/InternVL2_5-1B
180188 - OpenGVLab/InternVL2_5-8B
181189 - OpenGVLab/InternVL2_5-26B
@@ -196,8 +204,6 @@ pytorch_vl_model:
196204 - openbmb/MiniCPM-V-2_6
197205 - microsoft/Phi-3-vision-128k-instruct
198206 - microsoft/Phi-3.5-vision-instruct
199- - deepseek-ai/deepseek-vl2-tiny
200- - deepseek-ai/deepseek-vl2
201207
202208turbomind_base_model :
203209 - internlm/internlm2_5-7b
@@ -214,7 +220,6 @@ pytorch_base_model:
214220turbomind_quatization :
215221 no_awq :
216222 - meta-llama/Meta-Llama-3-1-70B-Instruct
217- - Qwen/Qwen2.5-72B-Instruct
218223 - Qwen/Qwen1.5-MoE-A2.7B-Chat
219224 - Qwen/Qwen2-VL-2B-Instruct
220225 - Qwen/Qwen2-VL-7B-Instruct
@@ -224,6 +229,7 @@ turbomind_quatization:
224229 - deepseek-ai/DeepSeek-V2-Lite-Chat
225230 - codellama/CodeLlama-7b-Instruct-hf
226231 - allenai/Molmo-7B-D-0924
232+ - THUDM/codegeex4-all-9b
227233 gptq :
228234 - internlm/internlm2_5-7b-chat
229235 no_kvint4 :
@@ -240,6 +246,7 @@ turbomind_quatization:
240246 - Qwen/Qwen-VL-Chat
241247 - Qwen/Qwen2.5-0.5B-Instruct
242248 - Qwen/Qwen2.5-7B-Instruct
249+ - Qwen/Qwen2.5-32B-Instruct
243250 - Qwen/Qwen2.5-72B-Instruct
244251 - Qwen/Qwen2-VL-2B-Instruct
245252 - Qwen/Qwen2-VL-7B-Instruct
@@ -300,6 +307,7 @@ pytorch_quatization:
300307 - Qwen/Qwen-VL-Chat
301308 - Qwen/Qwen2.5-0.5B-Instruct
302309 - Qwen/Qwen2.5-7B-Instruct
310+ - Qwen/Qwen2.5-32B-Instruct
303311 - Qwen/Qwen2.5-72B-Instruct
304312 - Qwen/Qwen2-VL-2B-Instruct
305313 - Qwen/Qwen2-VL-7B-Instruct
@@ -330,6 +338,7 @@ benchmark_model:
330338 - internlm/internlm2_5-20b-chat
331339 - THUDM/glm-4-9b-chat
332340 - Qwen/Qwen2.5-7B-Instruct
341+ - Qwen/Qwen2.5-32B-Instruct
333342 - Qwen/Qwen2.5-72B-Instruct
334343 - mistralai/Mistral-7B-Instruct-v0.3
335344 - mistralai/Mixtral-8x7B-Instruct-v0.1
0 commit comments