Skip to content

Add webgpu recipes for qwen3.5 models#307

Open
apsonawane wants to merge 5 commits intomainfrom
asonawane/webgpu
Open

Add webgpu recipes for qwen3.5 models#307
apsonawane wants to merge 5 commits intomainfrom
asonawane/webgpu

Conversation

@apsonawane
Copy link
Copy Markdown
Contributor

Add webgpu recipes for Qwen3.5 models. Export the model using this command:
python optimize.py --config-dir webgpu --device webgpu

Copilot AI review requested due to automatic review settings April 2, 2026 22:48
Copy link
Copy Markdown
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

This PR adds WebGPU export/optimization “recipes” for Qwen3.5 multimodal models (0.8B/2B/4B/9B) so they can be exported via the existing builtin/optimize.py flow using Olive and run with the WebGpuExecutionProvider.

Changes:

  • Add builtin/webgpu/{text,vision,embedding}.json Olive pipelines for Qwen3.5-0.8B/2B/4B/9B targeting WebGpuExecutionProvider.
  • Extend each model’s builtin/optimize.py to accept --device webgpu and emit WebGPU provider options into generated runtime config.

Reviewed changes

Copilot reviewed 16 out of 16 changed files in this pull request and generated 4 comments.

Show a summary per file
File Description
Qwen-Qwen3.5-9B/builtin/webgpu/vision.json New WebGPU vision encoder Olive pipeline targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-9B/builtin/webgpu/text.json New WebGPU text decoder ModelBuilder pipeline (INT4) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-9B/builtin/webgpu/embedding.json New WebGPU embedding/export pipeline (FP16) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-9B/builtin/optimize.py Add webgpu device option and provider-options patching path.
Qwen-Qwen3.5-4B/builtin/webgpu/vision.json New WebGPU vision encoder Olive pipeline targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-4B/builtin/webgpu/text.json New WebGPU text decoder ModelBuilder pipeline (INT4) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-4B/builtin/webgpu/embedding.json New WebGPU embedding/export pipeline (FP16) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-4B/builtin/optimize.py Add webgpu device option and provider-options patching path.
Qwen-Qwen3.5-2B/builtin/webgpu/vision.json New WebGPU vision encoder Olive pipeline targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-2B/builtin/webgpu/text.json New WebGPU text decoder ModelBuilder pipeline (INT4) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-2B/builtin/webgpu/embedding.json New WebGPU embedding/export pipeline (FP16) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-2B/builtin/optimize.py Add webgpu device option and provider-options patching path.
Qwen-Qwen3.5-0.8B/builtin/webgpu/vision.json New WebGPU vision encoder Olive pipeline targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-0.8B/builtin/webgpu/text.json New WebGPU text decoder ModelBuilder pipeline (INT4) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-0.8B/builtin/webgpu/embedding.json New WebGPU embedding/export pipeline (FP16) targeting WebGpuExecutionProvider.
Qwen-Qwen3.5-0.8B/builtin/optimize.py Add webgpu device option and provider-options patching path.

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.

Comment thread Qwen-Qwen3.5-0.8B/builtin/optimize.py
Comment thread Qwen-Qwen3.5-2B/builtin/optimize.py
Comment thread Qwen-Qwen3.5-4B/builtin/optimize.py
Comment thread Qwen-Qwen3.5-9B/builtin/optimize.py
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants