|
1 | 1 | {
|
| 2 | + "yi-1.5": { |
| 3 | + "brief": "Yi 1.5 is a high-performing, bilingual language model.", |
| 4 | + "default": "6b", |
| 5 | + "license": "Apache License Version 2.0", |
| 6 | + "variants": { |
| 7 | + "6b": { |
| 8 | + "default": "q8", |
| 9 | + "quantized": { |
| 10 | + "q8": { |
| 11 | + "size": 6441564880, |
| 12 | + "url": "chatllm_quantized_yi1.5/yi1.5-6b.bin" |
| 13 | + } |
| 14 | + } |
| 15 | + }, |
| 16 | + "9b": { |
| 17 | + "default": "q8", |
| 18 | + "quantized": { |
| 19 | + "q8": { |
| 20 | + "size": 9383354064, |
| 21 | + "url": "chatllm_quantized_yi1.5/yi1.5-9b.bin" |
| 22 | + } |
| 23 | + } |
| 24 | + }, |
| 25 | + "9b-16k": { |
| 26 | + "default": "q4_1", |
| 27 | + "quantized": { |
| 28 | + "q4_1": { |
| 29 | + "size": 5520662224, |
| 30 | + "url": "chatllm_quantized_yi1.5/yi1.5-9b-16k_q4_1.bin" |
| 31 | + } |
| 32 | + } |
| 33 | + } |
| 34 | + } |
| 35 | + }, |
| 36 | + "deepseek-v2": { |
| 37 | + "brief": "A strong, economical, and efficient Mixture-of-Experts language model.", |
| 38 | + "default": "light", |
| 39 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 40 | + "variants": { |
| 41 | + "light": { |
| 42 | + "default": "q8", |
| 43 | + "quantized": { |
| 44 | + "q8": { |
| 45 | + "size": 16691737872, |
| 46 | + "url": "chatllm_quantized_deepseek/deepseekv2-lite.bin" |
| 47 | + } |
| 48 | + } |
| 49 | + } |
| 50 | + } |
| 51 | + }, |
| 52 | + "deepseek-coder-v2": { |
| 53 | + "brief": "An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks.", |
| 54 | + "default": "light", |
| 55 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 56 | + "variants": { |
| 57 | + "light": { |
| 58 | + "default": "q4_1", |
| 59 | + "quantized": { |
| 60 | + "q8": { |
| 61 | + "size": 16691738304, |
| 62 | + "url": "chatllm_quantized_deepseek/deepseek-coder-v2-lite-instruct.bin" |
| 63 | + }, |
| 64 | + "q4_1": { |
| 65 | + "size": 9820206784, |
| 66 | + "url": "chatllm_quantized_deepseek/deepseek-coder-v2-lite-instruct_q4_1.bin" |
| 67 | + } |
| 68 | + } |
| 69 | + } |
| 70 | + } |
| 71 | + }, |
| 72 | + "deepseek-coder-v2-base": { |
| 73 | + "brief": "An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks.", |
| 74 | + "default": "light", |
| 75 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 76 | + "variants": { |
| 77 | + "light": { |
| 78 | + "default": "q4_1", |
| 79 | + "quantized": { |
| 80 | + "q4_1": { |
| 81 | + "size": 9820206784, |
| 82 | + "url": "chatllm_quantized_deepseek/deepseek-coder-v2-lite-base_q4_1.bin" |
| 83 | + } |
| 84 | + } |
| 85 | + } |
| 86 | + } |
| 87 | + }, |
| 88 | + "deepseek-llm": { |
| 89 | + "brief": "An advanced language model crafted with 2 trillion bilingual tokens.", |
| 90 | + "default": "7b", |
| 91 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 92 | + "variants": { |
| 93 | + "7b": { |
| 94 | + "default": "q8", |
| 95 | + "quantized": { |
| 96 | + "q8": { |
| 97 | + "size": 7345830176, |
| 98 | + "url": "chatllm_quantized_deepseek/deepseek-7b.bin" |
| 99 | + } |
| 100 | + } |
| 101 | + } |
| 102 | + } |
| 103 | + }, |
| 104 | + "deepseek-coder": { |
| 105 | + "brief": "DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.", |
| 106 | + "default": "1.3b", |
| 107 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 108 | + "variants": { |
| 109 | + "1.3b": { |
| 110 | + "default": "q8", |
| 111 | + "quantized": { |
| 112 | + "q8": { |
| 113 | + "size": 1431733904, |
| 114 | + "url": "chatllm_quantized_deepseek/deepseekcoder-1.3b.bin" |
| 115 | + } |
| 116 | + } |
| 117 | + }, |
| 118 | + "7b": { |
| 119 | + "default": "q8", |
| 120 | + "quantized": { |
| 121 | + "q8": { |
| 122 | + "size": 7163394192, |
| 123 | + "url": "chatllm_quantized_deepseek/deepseekcoder-7b.bin" |
| 124 | + } |
| 125 | + } |
| 126 | + } |
| 127 | + } |
| 128 | + }, |
| 129 | + "deepseek-coder-base": { |
| 130 | + "brief": "DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.", |
| 131 | + "default": "1.3b", |
| 132 | + "license": "DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023", |
| 133 | + "variants": { |
| 134 | + "1.3b": { |
| 135 | + "default": "q8", |
| 136 | + "quantized": { |
| 137 | + "q8": { |
| 138 | + "size": 1431733904, |
| 139 | + "url": "chatllm_quantized_deepseek/deepseekcoder-1.3b-base.bin" |
| 140 | + } |
| 141 | + } |
| 142 | + }, |
| 143 | + "6.7b": { |
| 144 | + "default": "q8", |
| 145 | + "quantized": { |
| 146 | + "q8": { |
| 147 | + "size": 7163394192, |
| 148 | + "url": "chatllm_quantized_deepseek/deepseekcoder-6.7b-base.bin" |
| 149 | + } |
| 150 | + } |
| 151 | + } |
| 152 | + } |
| 153 | + }, |
2 | 154 | "gemma": {
|
3 | 155 | "brief": "Gemma is a family of lightweight, state-of-the-art open models built by Google DeepMind. Updated to version 1.1.",
|
4 | 156 | "default": "2b",
|
|
25 | 177 | "quantized": {
|
26 | 178 | "q4_1": {
|
27 | 179 | "size": 5025629392,
|
28 |
| - "url": "chatllm_quantized_models/llama3-8b-q4_1.bin" |
| 180 | + "url": "chatllm_quantized_llama3/llama3-8b-q4_1.bin" |
| 181 | + }, |
| 182 | + "q8": { |
| 183 | + "size": 8538752208, |
| 184 | + "url": "chatllm_quantized_llama3/llama3-8b.bin" |
| 185 | + } |
| 186 | + } |
| 187 | + } |
| 188 | + } |
| 189 | + }, |
| 190 | + "llama3-chinese-lora": { |
| 191 | + "brief": "Llama-3-Chinese-8B-Instruct-LoRA, which is further tuned with 5M instruction data on Llama-3-Chinese-8B.", |
| 192 | + "default": "8b", |
| 193 | + "license": "", |
| 194 | + "variants": { |
| 195 | + "8b": { |
| 196 | + "default": "q4_0", |
| 197 | + "quantized": { |
| 198 | + "q4_0": { |
| 199 | + "size": 4523754704, |
| 200 | + "url": "chatllm_quantized_llama3/llama3-8b-lora-q4_0.bin" |
29 | 201 | }
|
30 | 202 | }
|
31 | 203 | }
|
|
0 commit comments