![]() |
ONNX-LRE
C++ API documentation
|
Configuration parameters for the inference engine. More...
#include <onnx_lre.hpp>
Public Attributes | |
| Cryption | cryption |
| Credentials for encrypted models (optional) More... | |
| ExecutionProvider | executionProvider = ExecutionProvider::CUDA |
| Hardware acceleration backend. More... | |
| Precision | precision = Precision::Float32 |
| Numerical precision for calculations. More... | |
| std::string | tensorRTTimingCachePath = "" |
| Path to TensorRT timing cache file (persists optimization data) More... | |
| std::string | tensorRTEngineCachePath = "" |
| Path to TensorRT engine cache directory (stores optimized subgraphs) More... | |
| int | deviceID = 0 |
| Device index for multi-GPU systems. More... | |
Configuration parameters for the inference engine.
Comprehensive set of options controlling model loading behavior, hardware acceleration settings, precision, and caching. These options significantly impact inference performance and memory usage.
Different execution providers have different performance characteristics:
Precision settings affect both memory usage and computation speed:
Example usage:
| Cryption OnnxLre::Options::cryption |
Credentials for encrypted models (optional)
| ExecutionProvider OnnxLre::Options::executionProvider = ExecutionProvider::CUDA |
Hardware acceleration backend.
| Precision OnnxLre::Options::precision = Precision::Float32 |
Numerical precision for calculations.
| std::string OnnxLre::Options::tensorRTTimingCachePath = "" |
Path to TensorRT timing cache file (persists optimization data)
| std::string OnnxLre::Options::tensorRTEngineCachePath = "" |
Path to TensorRT engine cache directory (stores optimized subgraphs)
| int OnnxLre::Options::deviceID = 0 |
Device index for multi-GPU systems.