You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Cherry pick fixes to release branch rel-1.3.0 (#3936)
* Fix DirectML nuget creation in Nuget pipeline (#3929)
* Added onnxruntime aarch64 wheel to pypi publishing pipeline (#3903)
* Added onnxruntime aarch64 wheel to pypi publishing pipeline
* Support nightly build flag
* Add support for nightly build
* Fix error handling in LearningModelSession.cpp (#3920)
* Update DML Nuget version and DML EP Doc (#3945)
Update DML Nuget version and DML EP Doc
* Fix ordering of APIs. (#3951)
Co-authored-by: Ryan Lai <[email protected]>
Co-authored-by: Prabhat <[email protected]>
Co-authored-by: Jeff Bloomfield <[email protected]>
Co-authored-by: Pranav Sharma <[email protected]>
Copy file name to clipboardExpand all lines: docs/execution_providers/DirectML-ExecutionProvider.md
+8-7Lines changed: 8 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,16 +1,16 @@
1
-
# DirectML Execution Provider (Preview)
1
+
# DirectML Execution Provider
2
2
3
3
DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning on Windows. DirectML provides GPU acceleration for common machine learning tasks across a broad range of supported hardware and drivers.
4
4
5
5
When used standalone, the DirectML API is a low-level DirectX 12 library and is suitable for high-performance, low-latency applications such as frameworks, games, and other real-time applications. The seamless interoperability of DirectML with Direct3D 12 as well as its low overhead and conformance across hardware makes DirectML ideal for accelerating machine learning when both high performance is desired, and the reliability and predictabiltiy of results across hardware is critical.
6
6
7
7
The *DirectML Execution Provider* is an optional component of ONNX Runtime that uses DirectML to accelerate inference of ONNX models. The DirectML execution provider is capable of greatly improving evaluation time of models using commodity GPU hardware, without sacrificing broad hardware support or requiring vendor-specific extensions to be installed.
8
8
9
-
The DirectML Execution Provider is currently in preview.
9
+
The DirectML Execution Provider currently uses DirectML version 2.1.0.
@@ -48,7 +48,7 @@ To build onnxruntime with the DML EP included, supply the `--use_dml` parameter
48
48
49
49
The DirectML execution provider supports building for both x64 (default) and x86 architectures.
50
50
51
-
Note that building onnxruntime with the DirectML execution provider enabled causes the the DirectML redistributable package to be automatically downloaded as part of the build. This package contains a pre-release version of DirectML, and its use is governed by a license whose text may be found as part of the NuGet package.
51
+
Note that building onnxruntime with the DirectML execution provider enabled causes the the DirectML redistributable package to be automatically downloaded as part of the build. Its use is governed by a license whose text may be found as part of the NuGet package.
52
52
53
53
54
54
@@ -83,7 +83,7 @@ Creates a DirectML Execution Provider using the given DirectML device, and which
83
83
84
84
### ONNX opset support
85
85
86
-
The DirectML execution provider currently supports ONNX opset 9 ([ONNX v1.4](https://github.com/onnx/onnx/releases/tag/v1.4.0)). Evaluating models which require a higher opset version is not supported, and may produce unexpected results.
86
+
The DirectML execution provider currently supports ONNX opset 11 ([ONNX v1.6](https://github.com/onnx/onnx/releases/tag/v1.6.0)). Evaluating models which require a higher opset version is not supported, and may produce unexpected results.
87
87
88
88
### Multi-threading and supported session options
89
89
@@ -114,8 +114,9 @@ The DirectML execution provider works most efficiently when tensor shapes are kn
114
114
115
115
Normally when the shapes of model inputs are known during session creation, the shapes for the rest of the model are inferred by OnnxRuntime when a session is created. However if a model input contains a free dimension (such as for batch size), steps must be taken to retain the above performance benefits.
116
116
117
-
In this case, there are two options:
118
-
- Edit the model to replace an input's free dimension (specified through ONNX using "dim_param") with a fixed size.
117
+
In this case, there are three options:
118
+
- Edit the model to replace an input's free dimension (specified through ONNX using "dim_param") with a fixed size (specified through ONNX using "dim_value").
119
+
- Specify values of named dimensions within model inputs when creating the session using the OnnxRuntime *AddFreeDimensionOverrideByName* ABI.
119
120
- Edit the model to ensure that an input's free dimension has a [denotation](https://github.com/onnx/onnx/blob/master/docs/DimensionDenotation.md) (such as "DATA_BATCH," or a custom denotation). Then when creating the session, specify the dimension size for each denotation. This can be done using the OnnxRuntime *AddFreeDimensionOverride* ABI.
0 commit comments