Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update jetson instruction #18204

Merged
merged 16 commits into from
Nov 2, 2023
71 changes: 38 additions & 33 deletions docs/build/eps.md
Original file line number Diff line number Diff line change
Expand Up @@ -142,58 +142,63 @@ Dockerfile instructions are available [here](https://github.com/microsoft/onnxru
### Build Instructions
{: .no_toc }

These instructions are for JetPack SDK 4.6.1.
These instructions are for the latest [JetPack SDK 5.1.2](https://developer.nvidia.com/embedded/jetpack-sdk-512).

1. Clone the ONNX Runtime repo on the Jetson host

```bash
git clone --recursive https://github.com/microsoft/onnxruntime
```
```bash
git clone --recursive https://github.com/microsoft/onnxruntime
```

2. Specify the CUDA compiler, or add its location to the PATH.

Cmake can't automatically find the correct nvcc if it's not in the PATH.
1. Starting with **CUDA 11.8**, Jetson users on **JetPack 5.0+** can upgrade to the latest CUDA release without updating the JetPack version or Jetson Linux BSP (Board Support Package). CUDA version 11.8 with JetPack 5.1.2 has been tested on Jetson when building ONNX Runtime 1.16.

```bash
export CUDACXX="/usr/local/cuda/bin/nvcc"
1. Check [this official blog](https://developer.nvidia.com/blog/simplifying-cuda-upgrades-for-nvidia-jetson-users/) for CUDA 11.8 upgrade instruction.

```
2. CUDA 12.x is only available to Jetson Orin and newer series (CUDA compute capability >= 8.7). Check [here](https://developer.nvidia.com/cuda-gpus#collapse5) for compute capability datasheet.

or:
2. CMake can't automatically find the correct `nvcc` if it's not in the `PATH`. `nvcc` can be added to `PATH` via:

```bash
export PATH="/usr/local/cuda/bin:${PATH}"
```
```bash
export PATH="/usr/local/cuda/bin:${PATH}"
```

3. Install the ONNX Runtime build dependencies on the Jetpack 4.6.1 host:
or:

```bash
sudo apt install -y --no-install-recommends \
build-essential software-properties-common libopenblas-dev \
libpython3.6-dev python3-pip python3-dev python3-setuptools python3-wheel
```
```bash
export CUDACXX="/usr/local/cuda/bin/nvcc"
```

4. Cmake is needed to build ONNX Runtime. Because the minimum required version is 3.18,
it is necessary to build CMake from source. Download Unix/Linux sources from https://cmake.org/download/
and follow https://cmake.org/install/ to build from source. Version 3.23.0 has been tested on Jetson.
3. Install the ONNX Runtime build dependencies on the Jetpack 5.1.2 host:

5. Build the ONNX Runtime Python wheel:
```bash
sudo apt install -y --no-install-recommends \
build-essential software-properties-common libopenblas-dev \
libpython3.8-dev python3-pip python3-dev python3-setuptools python3-wheel
```

```bash
./build.sh --config Release --update --build --parallel --build_wheel \
--use_cuda --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu
```
4. Cmake is needed to build ONNX Runtime. For ONNX Runtime 1.16, the minimum required CMake version is 3.26 (version 3.27.4 has been tested). This can be either installed by:

Note: You may optionally build with TensorRT support.
1. (Unix/Linux) Build from source. Download sources from [https://cmake.org/download/](https://cmake.org/download/)
and follow [https://cmake.org/install/](https://cmake.org/install/) to build from source.
2. (Ubuntu) Install deb package via apt repository: e.g [https://apt.kitware.com/](https://apt.kitware.com/)

```bash
./build.sh --config Release --update --build --parallel --build_wheel \
--use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu \
--tensorrt_home /usr/lib/aarch64-linux-gnu
```
5. Build the ONNX Runtime Python wheel (update path to CUDA/CUDNN/TensorRT libraries if necessary):

---
1. Build `onnxruntime-gpu` wheel with CUDA and TensorRT support:

```bash
./build.sh --config Release --update --build --parallel --build_wheel \
--use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu \
--tensorrt_home /usr/lib/aarch64-linux-gnu
```

​ Notes:

* By default, `onnxruntime-gpu` wheel file will be captured under `path_to/onnxruntime/build/Linux/Release/dist/` (build path can be customized by adding `--build_dir` followed by a customized path to the build command above).

* For a portion of Jetson devices like the Xavier series, higher power mode involves more cores (up to 6) to compute but it consumes more resource when building ONNX Runtime. Set `--parallel 2` or smaller in the build command if system is hanging and OOM happens.
## oneDNN

See more information on oneDNN (formerly DNNL) [here](../execution-providers/oneDNN-ExecutionProvider.md).
Expand Down
Loading