Commit 560bb9c
authored
Adding new MoE e2e tests (#1960)
SUMMARY:
adding e2e tests for MoE
also small fix so scheme: None doesn't error and by default ignore
expert gate layers (model dependent on whether its supported)
TEST PLAN:
in progress:
https://github.com/neuralmagic/llm-compressor-testing/actions/runs/19368818055
local (after disabling the [cadence
skip](https://github.com/vllm-project/llm-compressor/blob/main/tests/e2e/vLLM/test_vllm.py))
export
TEST_DATA_FILE="${REPOS}/llm-compressor/tests/e2e/vLLM/configs/qwen3_fp8_dynamic_per_tensor.yaml"
pytest tests/e2e/vLLM/test_vllm.py -vs 2>&1 | tee log-fp8.log
export
TEST_DATA_FILE="${REPOS}/llm-compressor/tests/e2e/vLLM/configs/qwen3_fp4_nvfp4.yaml"
pytest tests/e2e/vLLM/test_vllm.py -vs 2>&1 | tee log-fp4.log
---------
Signed-off-by: HDCharles <[email protected]>1 parent 6fea888 commit 560bb9c
File tree
4 files changed
+16
-3
lines changed- tests/e2e
- vLLM
- configs
4 files changed
+16
-3
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
84 | 84 | | |
85 | 85 | | |
86 | 86 | | |
87 | | - | |
| 87 | + | |
88 | 88 | | |
89 | 89 | | |
90 | 90 | | |
91 | | - | |
| 91 | + | |
| 92 | + | |
| 93 | + | |
92 | 94 | | |
93 | 95 | | |
94 | 96 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| 5 | + | |
| 6 | + | |
| 7 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
18 | 18 | | |
19 | 19 | | |
20 | 20 | | |
21 | | - | |
| 21 | + | |
22 | 22 | | |
23 | 23 | | |
24 | 24 | | |
| |||
0 commit comments