Skip to content

Commit 20a46e0

Browse files
Update ModelOpt-with-Olive documentation to mention CUDA EP commands (#1099)
### What does this PR do? Type of change: Minor documentation update - Update documentation (olive installation instructions) to mention install commands for CUDA EP packages for ORT / ORT-genai. ### Testing - Locally checked the readme, doc. ### Before your PR is "*Ready for review*" Make sure you read and follow [Contributor guidelines](https://github.com/NVIDIA/Model-Optimizer/blob/main/CONTRIBUTING.md) and your commits are signed (`git commit -s -S`). Make sure you read and follow the [Security Best Practices](https://github.com/NVIDIA/Model-Optimizer/blob/main/SECURITY.md#security-coding-practices-for-contributors) (e.g. avoiding hardcoded `trust_remote_code=True`, `torch.load(..., weights_only=False)`, `pickle`, etc.). - Is this change backward compatible?: ✅ / ❌ / N/A <!--- If ❌, explain why. --> - If you copied code from any other sources or added a new PIP dependency, did you follow guidance in `CONTRIBUTING.md`: ✅ / ❌ / N/A <!--- Mandatory --> - Did you write any new necessary tests?: ✅ / ❌ / N/A <!--- Mandatory for new features or examples. --> - Did you update [Changelog](https://github.com/NVIDIA/Model-Optimizer/blob/main/CHANGELOG.rst)?: ✅ / ❌ / N/A <!--- Only for new features, API changes, critical bug fixes or backward incompatible changes. --> ### Additional Information <!-- E.g. related issue. --> <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit * **Documentation** * Updated Windows Olive installation guidance to recommend CUDA-based ONNX Runtime packages instead of DirectML and added a link to ONNX Runtime’s Execution-Provider docs for alternative EPs and requirements. * Simplified Windows examples by removing explicit package install commands and pointing users to the consolidated Olive setup instructions. <!-- end of auto-generated comment: release notes by coderabbit.ai --> --------- Signed-off-by: vipandya <vipandya@nvidia.com>
1 parent 029593e commit 20a46e0

2 files changed

Lines changed: 9 additions & 8 deletions

File tree

docs/source/getting_started/windows/_installation_with_olive.rst

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -17,14 +17,16 @@ Setup Steps for Olive with ModelOpt-Windows
1717
1818
pip install olive-ai[nvmo]
1919
20-
- **Install Prerequisites:** Ensure all required dependencies are installed. For example, to use DirectML Execution-Provider (EP) based onnxruntime and onnxruntime-genai packages, run the following commands:
20+
- **Install Prerequisites:** Ensure all required dependencies are installed. For example, to install CUDA Execution-Provider (EP) based onnxruntime and onnxruntime-genai packages, run the following commands:
2121

2222
.. code-block:: shell
2323
24-
$ pip install onnxruntime-genai-directml>=0.4.0
25-
$ pip install onnxruntime-directml==1.20.0
24+
$ pip install onnxruntime-genai-cuda
25+
$ pip install onnxruntime-gpu
2626
27-
- Above onnxruntime and onnxruntime-genai packages enable Olive workflow with DirectML Execution-Provider (EP). To use other EPs, install corresponding packages.
27+
- Above onnxruntime and onnxruntime-genai packages enable Olive workflow with CUDA Execution-Provider (EP). To use other EPs, install corresponding packages.
28+
29+
- Refer to the ONNX Runtime's `EP documentation <https://onnxruntime.ai/docs/execution-providers/>`_ for details about different EPs, their requirements, and installation instructions.
2830

2931
- Additionally, ensure that dependencies for Model Optimizer - Windows are met as mentioned in the :ref:`Install-Page-Standalone-Windows`.
3032

examples/windows/README.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -51,15 +51,14 @@ pip install nvidia-modelopt[onnx]
5151

5252
### Installation with Olive
5353

54-
To install ModelOpt-Windows through Microsoft's Olive, use the following commands:
54+
The ModelOpt-Windows is integrated into Microsoft's [Olive](https://microsoft.github.io/Olive/) framework. Run the following commands to install ModelOpt through Olive.
5555

5656
```bash
5757
pip install olive-ai[nvmo]
58-
pip install onnxruntime-genai-directml>=0.4.0
59-
pip install onnxruntime-directml==1.20.0
58+
pip install onnxruntime-genai-cuda
6059
```
6160

62-
For more details, please refer to the [detailed installation instructions](https://nvidia.github.io/Model-Optimizer/getting_started/windows/_installation_for_Windows.html).
61+
For more details, or to use different [ONNX Runtime Execution Providers](https://onnxruntime.ai/docs/execution-providers/), refer to the [detailed installation instructions](https://nvidia.github.io/Model-Optimizer/getting_started/windows/_installation_for_Windows.html).
6362

6463
## Techniques
6564

0 commit comments

Comments
 (0)