diff --git a/winarm.html b/winarm.html
index 9b6264717b8da..3eec46ab9885c 100644
--- a/winarm.html
+++ b/winarm.html
@@ -101,19 +101,10 @@
Get started on your Windows Dev Kit 2023 today
Follow these steps to setup your device to use ONNX Runtime (ORT) with the built
in NPU:
- - Request access to the Neural Processing SDK for
- Windows on Snapdragon. Qualcomm may reach out to you via email with further
- registration instructions for approval.
- - Once approved, you will receive an email with links to download SNPE.
-
- - Select the SNPE link which takes you to a Qualcomm login and
- download page.
- - Select the Snapdragon_NPE_SDK.WIN.1.0 Installer link,
- download and install.
-
- - Download and install the ONNX Runtime with SNPE
+
- Download the Qualcomm AI Engine Direct SDK (QNN SDK)
+ - Download and install the ONNX Runtime with QNN
package
- Start using the ONNX Runtime API in your application.
@@ -122,29 +113,14 @@ Optimizing models for the NPU
ONNX is a standard format for
representing ML models authored in frameworks like PyTorch, TensorFlow, and
others. ONNX Runtime can run any ONNX model, however to make use of the NPU, you
- currently need to use the following steps:
-
- - Run the tools provided in the SNPE SDK on your model to generate a binary
- file.
- - Include the contents of the binary file as a node in the ONNX graph.
-
+ currently need to quantize the ONNX model to QDQ model.
See our C# tutorial for an example of how this is done.
Many models can be optimized for the NPU using this process. Even if a model
- cannot be optimized for NPU by the SNPE SDK, it can still be run by ONNX Runtime
+ cannot be optimized for NPU, it can still be run by ONNX Runtime
on the CPU.
- Tutorials
-
-
Getting help
For help with ONNX Runtime, you can Getting help