Skip to content

Commit

Permalink
Update winarm.svelte
Browse files Browse the repository at this point in the history
update winarm instruuctions to be current
  • Loading branch information
jeffmend authored Dec 12, 2023
1 parent afa038a commit c08de75
Showing 1 changed file with 2 additions and 6 deletions.
8 changes: 2 additions & 6 deletions src/routes/components/winarm.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -28,16 +28,12 @@
<p class="text-xl text-blue-500">Optimizing models for the NPU</p>
<a class="text-blue-500" href="https://onnx.ai/">ONNX</a> is a standard format for representing ML models authored in frameworks like PyTorch,
TensorFlow, and others. ONNX Runtime can run any ONNX model, however to make use of the NPU,
you currently need to use the following steps:
<ol class="list-disc ml-10">
<li>Run the tools provided in the SNPE SDK on your model to generate a binary file.</li>
<li>Include the contents of the binary file as a node in the ONNX graph.</li>
you currently need to quantize the ONNX model to QDQ model.
<br>
See our <a class="text-blue-500" href="https://github.com/microsoft/onnxruntime-inference-examples/tree/main/c_cxx/QNN_EP/mobilenetv2_classification">C# tutorial</a> for an example of how this is done.
</ol>
<br />
Many models can be optimized for the NPU using this process. Even if a model cannot be optimized
for NPU by the SNPE SDK, it can still be run by ONNX Runtime on the CPU.
for the NPU, it can still be run by ONNX Runtime on the CPU.
<br><br>
<p class="text-xl text-blue-500">Getting Help</p>
For help with ONNX Runtime, you can <a class="text-blue-500" href="https://github.com/microsoft/onnxruntime/discussions">start a discussion</a> on GitHub or <a class="text-blue-500" href="https://github.com/microsoft/onnxruntime/issues">file an issue</a>.
Expand Down

0 comments on commit c08de75

Please sign in to comment.