diff --git a/winarm.html b/winarm.html index 9b6264717b8da..3eec46ab9885c 100644 --- a/winarm.html +++ b/winarm.html @@ -101,19 +101,10 @@

Get started on your Windows Dev Kit 2023 today

Follow these steps to setup your device to use ONNX Runtime (ORT) with the built in NPU:

    -
  1. Request access to the Neural Processing SDK for - Windows on Snapdragon. Qualcomm may reach out to you via email with further - registration instructions for approval.
  2. -
  3. Once approved, you will receive an email with links to download SNPE. -
      -
    1. Select the SNPE link which takes you to a Qualcomm login and - download page.
    2. -
    3. Select the Snapdragon_NPE_SDK.WIN.1.0 Installer link, - download and install.
    4. -
    -
  4. Download and install the ONNX Runtime with SNPE +
  5. Download the Qualcomm AI Engine Direct SDK (QNN SDK)
  6. +
  7. Download and install the ONNX Runtime with QNN package
  8. Start using the ONNX Runtime API in your application.
@@ -122,29 +113,14 @@

Optimizing models for the NPU

ONNX is a standard format for representing ML models authored in frameworks like PyTorch, TensorFlow, and others. ONNX Runtime can run any ONNX model, however to make use of the NPU, you - currently need to use the following steps: -

+ currently need to quantize the ONNX model to QDQ model.

See our C# tutorial for an example of how this is done.

Many models can be optimized for the NPU using this process. Even if a model - cannot be optimized for NPU by the SNPE SDK, it can still be run by ONNX Runtime + cannot be optimized for NPU, it can still be run by ONNX Runtime on the CPU.


-

Tutorials

-
-

Getting help

For help with ONNX Runtime, you can Getting help -f \ No newline at end of file +f