Releases: microsoft/onnxruntime
ONNX Runtime v1.11.1
This is a patch release on 1.11.0 with the following fixes:
- Symbolic shape infer error (#10674)
- Quantization tool bug (#10940)
- Adds missing numpy type when looking for the ort correspondance (#10943)
- Profiling tool JSON format bug (#11046)
- Function bug fix (#11148)
- Add mobile helpers to Python build (#11196)
- Scoped GIL release in run_with_iobinding (#11248)
- Fix output type mapping for JS (#11049)
All official packages are attached, and Python packages are additionally published to PyPi.
ONNX Runtime v1.11.0
Key Updates
General
- Support for ONNX 1.11 with opset 16
- Updated protobuf version to 3.18.x
- Enable usage of Mimalloc (details)
- Transformer model helper scripts
- On Windows, error strings in OrtStatus are now encoded in UTF-8. When you need to print it out to screen, first convert it to a wide char string by using the MultiByteToWideChar Windows API.
Performance
- Memory utilization related performance improvements (e.g. elimination of vectors for small dims)
- Performance variance stability improvement through dynamic cost model session option (details)
- New quantization data format support: S8S8 in QDQ format
- Added s8s8 kernels for ARM64
- Support to convert s8s8 to u8s8 automatically for x64
- Improved performance on ARM64 for quantized CNN model through:
- New kernels for quantized depthwise Conv
- Improved symmetrically quantized Conv by leveraging indirect buffer
- New Gemm kernels for symmetric quantized Conv and MatMul
- General quantization improvements, including new quantized operators (Resize, ArgMax) and quantization tool updates
API
- Java: Only a single OrtEnv can be created in any given execution of the JVM. Previously, the environment could be closed completely and a fresh one could be created with different parameters (e.g. global thread pool, or logging level) (details)
Packages
- Nuget packages
- C# packages now tested with .NET 5. .NET Core 2.1 support is deprecated as it has reached end of life support on August 21, 2021. We will closely follow .NET's support policy
- Removed PDB files. These are attached as release artifacts below.
- Pypi packages
- Python 3.6 is deprecated as it has reached EOL December 2021. Supported Python versions: 3.7-3.9
- Note: Mac M1 builds are not yet available in pypi but can be built from source
- OnnxRuntime with OpenVINO support available at https://pypi.org/project/onnxruntime-openvino/1.11.0/
Execution Providers
- CUDA
- Enable CUDA provider option configuration for C# to support workspace size configuration from and fix binary compatibility of CUDAProviderOptions C API
- Preview support for CUDA Graphs (details)
- TensorRT
- TRT 8.2.3 support
- Memory footprint optimizations
- Support protobuf >= 3.11
- Updated flatbuffers version to 2.0
- Misc Bug Fixes
- DirectML
- Updated more operators to opset 13 (QuantizeLinear, DequantizeLinear, ReduceSum, Split, Squeeze, Unsqueeze, ReduceSum).
- OpenVINO
- OpenVINO™ version upgraded to 2022.1.0 - biggest OpenVINO™ upgrade in 3.5 years. This provides functional bug fixes, API Change 2.0 and capability changes from the previous 2021.4.2 LTS release.
- Performance Optimizations of existing supported models.
- Pre-Built OnnxRuntime Binaries with OpenVINO enabled can be downloaded from https://github.com/intel/onnxruntime/releases/tag/v4.0
https://pypi.org/project/onnxruntime-openvino/1.11.0/
- OpenCL (in preview)
- Introduced the EP for OpenCL to use with Mobile GPUs
- Available in
experimental/opencl
branch for users to try. Provide feedback through Issues and Discussions in the repo. - README is available here.
Mobile
- Added general support for converting a model to NHWC layout at runtime
- Execution provider sets preferred layout and shared infrastructure in ORT will ensure the nodes the execution provider is assigned will be in that layout
- Added support for runtime optimization with minimal binary size impact
- Relevant optimizations are saved in the ORT format model for replay at runtime if applicable
- Added support for QDQ format models to the NNAPI EP
- Will fall back to CPU EP’s QDQ handling if NNAPI is not available using runtime optimizations
- Includes updates to the ORT QDQ optimizers so they work better with mobile scenarios
- Added helpers to:
- Analyze if a model can be used with the pre-built ORT Mobile package
- Update ONNX opset so model can be used with the pre-built package
- Convert dynamic inputs into fixed size inputs so that the model can be used with NNAPI/CoreML
- Optimize a QDQ format model for use with ORT
- Added Android and iOS packages with full ORT builds
- These packages have additional support for the full set of opsets and ops for ONNX models at the cost of a larger binary size.
Web
- Build option to create ONNX Runtime WebAssembly static library
- Support for concurrent creation of multiple inference sessions
- Upgraded emsdk version to 3.1.3 for more stable multi-threads and enables LTO with multi-threads build on WebAssembly.
Known issues
- When using tensor sequences/sparse tensors, the generated profile is not valid JSON. (Fixed in #10974)
- There is a bug in the quantization tool for calibration when choosing percentile algorithm (Fixed in #10940). To fix this, please apply the typo fix in the python file.
- Mac M
Contributions
Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:
snnn, edgchen1, skottmckay, yufenglee, wangyems, yuslepukhin, gwang-msft, iK1D, chilo-ms, fdwr, ytaous, RandySheriffH, hanbitmyths, chenfucn, yihonglyu, ajindal1, fs-eire, souptc, tianleiwu, YUNQIUGUO, hariharans29, oliviajain, xadupre, ashari4, RyanUnderhill, jywu-msft, weixingzhang, baijumeswani, georgen117, natke, Craigacp, jeffdaily, JingqiaoFu, zhanghuanrong, satyajandhyala, smk2007, ryanlai2, askhade, thiagocrepaldi, jingyanwangms, pengwa, scxiao, ashbhandare, BowenBao, SherlockNoMad, sumitsays, sfatimar, mosdav, harshithapv, liqunfu, tiagoshibata, gineshidalgo99, pranavsharma, jcwchen, nkreeger, xkszltl, faxu, suffiank, stevenlix, jeffbloo, feihugis
ONNX Runtime v1.10.0
Announcements
- As noted in the deprecation notice in ORT 1.9, InferenceSession now requires the providers parameters to be set when enabling Execution Providers other than default CPUExecutionProvider.
e.g. InferenceSession('model.onnx', providers=['CUDAExecutionProvider']) - Python 3.6 support removed for Mac builds. Since 3.6 is end-of-life in December 2021, it will no longer be supported from next release (ORT 1.11) onwards
- Removed dependency on optional-lite
- Removed experimental Featurizers code
General
- Support for plug-in custom thread creation and join functions to enable usage of external threads
- Optional type support from op set 15
Performance
- Introduced indirect Convolution method for QLinearConv which has symmetrically quantized filter, i.e., filter type is int8 and zero point of filter is 0. The method leverages in-direct buffer instead of memcpy'ing the original data and doesn’t need to compute the sum of each pixel of output image for quantized Conv.
- X64: new kernels - including avx2, avxvnni, avx512 and avx 512 vnni, for general and depthwise quantized Conv.
- ARM64: new kernels for depthwise quantized Conv.
- Tensor shape optimization to avoid allocating heap memory in most cases - #9542
- Added transpose optimizer to push and cancel transpose ops, significantly improving perf for models requiring layout transformation
API
- Python
- Following through on the deprecation notice in ORT 1.9, InferenceSession now requires the providers parameters to be set when enabling Execution Providers other than default CPUExecutionProvider.
e.g. InferenceSession('model.onnx', providers=['CUDAExecutionProvider'])
- Following through on the deprecation notice in ORT 1.9, InferenceSession now requires the providers parameters to be set when enabling Execution Providers other than default CPUExecutionProvider.
- C/C++
- New API to query CUDA stream to launch a custom kernel for scenarios where custom ops compiled into shared libraries need implicit synchronization with ORT CUDA kernels - #9141
- Updated Invalid -> OrtInvalidAllocator
- Updated every item in OrtCudnnConvAlgoSearch to a safer global name
- WinML
- New APIs to create OrtValues from Windows platform specific ID3D12Resources by exposing DirectML Execution Provider specific APIs. These APIs allow DML to extend the C-API and provide EP specific extensions.
- OrtSessionOptionsAppendExecutionProviderEx_DML
- DmlCreateGPUAllocationFromD3DResource
- DmlFreeGPUAllocation
- DmlGetD3D12ResourceFromAllocation
- Bug fix: LearningModel::LoadFromFilePath in UWP apps
- New APIs to create OrtValues from Windows platform specific ID3D12Resources by exposing DirectML Execution Provider specific APIs. These APIs allow DML to extend the C-API and provide EP specific extensions.
Packages
- Added Mac M1 Universal2 build support for a single binary that runs natively on both Apple silicon and Intel-based Macs. These are included in the official Nuget packages. (build instructions)
- Windows C API Symbols are now uploaded to Microsoft symbol server
- Nuget package now supports ARM64 Linux C#
- Python GPU package now includes both TensorRT and CUDA EPs. Note: EPs need to be explicitly registered to ensure the correct provider is used. e.g. InferenceSession('model.onnx', providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider']). Please also ensure you have appropriate TensorRT dependencies and CUDA dependencies installed.
Execution Providers
- TensorRT EP
- Python GPU release packages now include support for TensorRT 8.0. Enable TensorrtExecutionProvider by explicitly setting providers parameter when creating an InferenceSession. e.g. InferenceSession('model.onnx', providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider'])
- Published quantized BERT model example
- OpenVINO EP
- Add support for OpenVINO 2021.4.x
- Auto Plugin support
- IO Buffer/Copy Avoidance Optimizations for GPU plugin
- Misc fixes
- DNNL EP
- Add Softmaxgrad op
- Add Transpose, Reshape, Pow and LeakyRelu ops
- Add DynamicQuantizeLinear op
- Add squeeze/unsqueeze ops
- DirectML EP
Mobile
- Added Xamarin support to the ORT C# Nuget packages
- Updated target frameworks in native package
- iOS and Android binaries now included in native package
- ORT format models now have backwards compatibility guarantee
Web
- Support WebAssembly SIMD for qgemm kernel to accelerate the performance of quantized models
- Upgraded existing WebGL kernels to the latest opset
- Optimized bundle size to support various production scenarios, such as WebAssembly only or WebGL only
Contributions
Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:
snnn, gineshidalgo99, fs-eire, gwang-msft, edgchen1, hariharans29, skottmckay, jeffdaily, baijumeswani, fdwr, smk2007, suffiank, souptc, RyanUnderhill, iK1D, yuslepukhin, chilo-ms, satyajandhyala, hanbitmyths, thiagocrepaldi, wschin, tianleiwu, pengwa, xadupre, zhanghuanrong, SherlockNoMad, wangyems, RandySheriffH, ashbhandare, tiagoshibata, yufenglee, mindest, sumitsays, MaajidKhan, gramalingam, tracysh, georgen117, jywu-msft, sfatimar, martinb35, nkreeger, ytaous, ashari4, stevenlix, chandru-r, jingyanwangms, mosdav, raviskolli, faxu, liqunfu, kit1980, weixingzhang, pranavsharma, jcwchen, chenfucn, BowenBao, jeffbloo
ONNX Runtime v1.9.1
ONNX Runtime v1.9.0
Announcements
- GCC version < 7 is no longer supported
- CMAKE_SYSTEM_PROCESSOR needs be set when cross-compiling on Linux because pytorch cpuinfo was introduced as a dependency for ARM big.LITTLE support. Set it to the value of
uname -m
output of your target device.
General
- ONNX 1.10 support
- opset 15
- ONNX IR 8 (SparseTensor type, model local functionprotos, Optional type not yet fully supported this release)
- Improved documentation of C/C++ APIs
- IBM Power support
- WinML - DLL dependency fix supports learning models on Windows 8.1
- Support for sub-building onnxruntime-extensions and statically linking into onnxruntime binary for custom builds
- Add
--_use_extensions
option to run models with custom operators implemented in onnxruntime-extensions
- Add
APIs
- Registration of a custom allocator for sharing between multiple sessions. (See RegisterAllocator and UnregisterAllocator APIs in onnxruntime_c_api.h)
- SessionOptionsAppendExecutionProvider_TensorRT API is deprecated; use SessionOptionsAppendExecutionProvider_TensorRT_V2
- New APIs: SessionOptionsAppendExecutionProvider_TensorRT_V2, CreateTensorRTProviderOptions, UpdateTensorRTProviderOptions, GetTensorRTProviderOptionsAsString, ReleaseTensorRTProviderOptions, EnableOrtCustomOps, RegisterAllocator, UnregisterAllocator, IsSparseTensor, CreateSparseTensorAsOrtValue, FillSparseTensorCoo, FillSparseTensorCsr, FillSparseTensorBlockSparse, CreateSparseTensorWithValuesAsOrtValue, UseCooIndices, UseCsrIndices, UseBlockSparseIndices, GetSparseTensorFormat, GetSparseTensorValuesTypeAndShape, GetSparseTensorValues, GetSparseTensorIndicesTypeShape, GetSparseTensorIndices,
Performance and quantization
- Performance improvement on ARM
- Added S8S8 (signed int8, signed int8) matmul kernel. This avoids extending uin8 to int16 for better performance on ARM64 without dot-product instruction
- Expanded GEMM udot kernel to 8x8 accumulator
- Added sgemm and qgemm optimized kernels for ARM64EC
- Operator improvements
- Improved performance for quantized operators: DynamicQuantizeLSTM, QLinearAvgPool
- Added new quantized operator QGemm for quantizing Gemm directly
- Fused HardSigmoid and Conv
- Quantization tool - subgraph support
- Transformers tool improvements
- Fused Attention for BART encoder and Megatron GPT-2
- Integrated mixed precision ONNX conversion and parity test for GPT-2
- Updated graph fusion for embed layer normalization for BERT
- Improved symbolic shape inference for operators: Attention, EmbedLayerNormalization, Einsum and Reciprocal
Packages
- Official ORT GPU packages (except Python) now include both CUDA and TensorRT Execution Providers.
- Python packages will be updated next release. Please note that EPs should be explicitly registered to ensure the correct provider is used.
- GPU packages are built with CUDA 11.4 and should be compatible with 11.x on systems with the minimum required driver version. See: CUDA minor version compatibility
- Pypi
- ORT + DirectML Python packages now available: onnxruntime-directml
- GPU package can be used on both CPU-only and GPU machines
- Nuget
- C#: Added support for using netstandard2.0 as a target framework
- Windows symbol (PDB) files are no longer included in the Nuget package, reducing size of the binary Nuget package by 85%. To download, please see the artifacts below in Github.
Execution Providers
-
CUDA EP
-
TensorRT EP
- Added support for TensorRT 8.0 (x64 Windows/Linux, ARM Jetson), which includes new TensorRT explicit-quantization features (ONNX Q/DQ support)
- General fixes and quality improvements
-
OpenVINO EP
- Added support for OpenVINO 2021.4
-
DirectML EP
- Bug fix for Identity with non-float inputs affecting DynamicQuantizeLinear ONNX backend test
ORT Web
- WebAssembly
- SIMD (Single Instruction, Multiple Data) support
- Option to load WebAssembly from worker thread to avoid blocking main UI thread
- wasm file path override
- WebGL
- Simpler workflow for WebGL kernel implementation
- Improved performance with Conv kernel enhancement
ORT Mobile
- Added more example mobile apps
- CoreML and NNAPI EP enhancements
- Reduced peak memory usage when initializing session with ORT format model as bytes
- Enhanced partitioning to improve performance when using NNAPI and CoreML
- Reduce number of NNAPI/CoreML partitions required
- Add ability to force usage of CPU for post-processing in SSD models
- Improves performance by avoiding expensive device copy to/from NPU for cheap post-processing section of the model
- Changed to using xcframework in the iOS package
- Supports usage of arm64 iPhone simulator on Mac with Apple silicon
ORT Training
- Expanding input formats supported to include dictionaries and lists.
- Enable user defined autograd functions
- Support for fallback to PyTorch for execution
- Added support for deterministic compute to enable reproducibility with ORTModule
- Add DebugOptions and LogLevels to ORTModule API* to improve debuggability
- Improvements additions to kernels/gradients: Concat, Split, MatMul, ReluGrad, PadOp, Tile, BatchNormInternal
- Support for ROCm 4.3.1 on AMD GPU
Contributions
Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:
edgchen1, gwang-msft, tianleiwu, fs-eire, hariharans29, skottmckay, baijumeswani, RyanUnderhill, iK1D, souptc, nkreeger, liqunfu, pengwa, SherlockNoMad, wangyems, chilo-ms, thiagocrepaldi, KeDengMS, suffiank, oliviajain, chenfucn, satyajandhyala, yuslepukhin, pranavsharma, tracysh, yufenglee, hanbitmyths, ytaous, YUNQIUGUO, zhanghuanrong, stevenlix, jywu-msft, chandru-r, duli2012, smk2007, wschin, MaajidKhan, tiagoshibata, xadupre, RandySheriffH, ashbhandare, georgen117, Tixxx, harshithapv, Craigacp, BowenBao, askhade, zhangxiang1993, gramalingam, weixingzhang, natke, tlh20, codemzs, ryanlai2, raviskolli, pranav-prakash, faxu, adtsai, fdwr, wenbingl, jcwchen, neginraoof, cschreib-ibex
ONNX Runtime v1.8.2
This is a minor patch release on 1.8.1 with the following changes:
Inference
- Fix a crash issue when optimizing
Conv->Add->Relu
for CUDA EP - ORT Mobile updates
- Change Pre-built iOS package to static framework to fix App Store submission issue
- Support for metadata in ORT format models
- Additional operators
- Bug fixes
Known issues
- cudnn 8.0.5 causes memory leaks on T4 GPU as indicated by the issue, an upgrade to later version solves the problem.
ONNX Runtime v1.8.1
This release contains fixes and key updates for 1.8.0.
For all package installation details, please refer to https://www.onnxruntime.ai.
Inference
- Fixes for GPU package loading issues
- Fix for memory issue for models with convolution nodes while using the EXHAUSTIVE algo search mode
- ORT Mobile updates
- CoreML EP enabled in iOS mobile package
- Additional operators
- Bug fixes
- React Native package now available
Training
Performance updates for ONNX Runtime for PyTorch (training acceleration for PyTorch models)
- Accelerates most popular Hugging Face models as well as GPT-Neo and Microsoft TNLG and TNLU models
- Support for PyTorch 1.8.1 and 1.9
- Support for CUDA 10.2 and 11.1
- Preview packages for ROCm 4.2
ONNX Runtime v1.8.0
Announcements
- This release
- Building onnxruntime from source now requires a C++ compiler with full C++14 support.
- Builds with OpenMP are no longer published. They can still be built from source if needed. The default threadpool option should provide optimal performance for the majority of models.
- New dependency for Python package: flatbuffers
- Next release (v1.9)
- Builds will require C++ 17 compiler
- GPU build will be updated to CUDA 11.1
General
- ONNX opset 14 support - new and updated operators from the ONNX 1.9 release
- Dynamically loadable CUDA execution provider
- Allows a single build to work for both CPU and GPU (excludes Python packages)
- Profiler tool now includes information on threadpool usage
- multi-threading preparation time
- multi-threading run time
- multi-threading wait time
- [Experimental] onnxruntime-extensions package
- Crowd-sourced library of common/shareable custom operator implementations that can be loaded and run with ONNX Runtime; community contributions are welcome! - microsoft/onnxruntime-extensions
- Currently includes mostly ops and tokenizers for string operations (full list here)
- Tutorials to export and load custom ops from onnxruntime-extensions: TensorFlow, PyTorch
Training
- torch-ort package released as the ONNX Runtime backend in PyTorch
- onnxruntime-training-gpu and onnxruntime-training-rocm packages now available for distributed training on NVIDIA and AMD GPUs
Mobile
- Official package now available
- Pre-built Android and iOS packages with support for selected operators and data types
- Objective-C API for iOS in preview
- Expanded operators supported by NNAPI (Android) and CoreML (iOS) execution providers
- All operators in the ai.onnx domain now support type reduction
- Create ORT format model with
--enable_type_reduction
flag, and perform minimal build--enable_reduced_operator_type_support
flag
- Create ORT format model with
ORT Web
- New ONNX Runtime Javascript API
- ONNX Runtime Web package
- Support WebAssembly and WebGL for CPU and GPU
- Support Web Worker based multi-threaded WebAssembly backend
- Supports ORT model format
- Improved WebGL performance
Performance
-
Memory footprint reduction through shared pre-packed weights for shared initializers
- Pre-packing refers to weights that are pre-processed at model load time
- Allows pre-packed weights of shared initializers to also be shared between sessions, preserving memory savings from using shared initializers
-
Memory footprint reduction through arena shrinkage
- By default, the memory arena doesn't shrink and it holds onto any allocated memory forever. This feature exposes a RunOption that scans the arena and potentially returns unused memory back to the system after the end of a Run. This feature is particularly useful while running a dynamic shape model that may occasionally process an outlier inference request that requires a large amount of memory. If the shrinkage option if invoked as part of these Runs, the memory that was required for that Run is not held on forever by the memory arena.
-
Quantization
- Native support of Quantize-Dequantize (QDQ) format for CPU
- Support for Concat, Transpose, GlobalAveragePool, AveragePool, Resize, Squeeze
- Improved performance on high-end ARM devices by leveraging dot-product instructions
- Improved performance for batched quant GEMM with optimized multi-threading logic
- Per-column quantization for MatMul
-
Transformers
- GPT-2 and beam search integration (example)
APIs
- WinML
- New native WinML API SetIntraOpThreadSpinning for toggling Intra Op thread spin behavior. When enabled, and when there is no current workload, IntraOp threads will continue to spin for some additional time as it waits for any additional work. This can result in better performance for the current workload but may impact performance of other unrelated workloads. This toggle is enabled by default.
- ORT Inferencing
- The following APIs have been added to this release. Please check the API documentation for information.
- KernelInfoGetAttributeArray_float
- KernelInfoGetAttributeArray_int64
- CreateArenaCfgV2
- AddRunConfigEntry
- CreatePrepackedWeightsContainer
- PrepackedWeightsContainer
- CreateSessionWithPrepackedWeightsContainer
- CreateSessionFromArrayWithPrepackedWeightsContainer
- The following APIs have been added to this release. Please check the API documentation for information.
Execution Providers
- TensorRT
- Added support for TensorRT EP configuration using session options instead of environment variables.
- Added support for DLA on Jetson Xavier (AGX, NX)
- General bug fixes and quality improvements.
- OpenVINO
- Added support for OpenVINO 2021.3
- Removed support for OpenVINO 2020.4
- Added support for Loading/Saving of Blobs on MyriadX devices to avoid expensive model blob compilation at runtime.
- DirectML
• Supports ARM/ARM64 architectures now in WinML and ONNX RunTime NuGet packages.
• Support for 8-dimensional tensors to: BatchNormalization, Cast, Join, LpNormalization, MeanVarianceNormalization, Padding, Tile, TopK.
• Substantial performance improvements for several operators.
• Resize nearest_mode “floor” and “round_prefer_ceil”.
• Fusion activations for: Conv, ConvTranspose, BatchNormalization, MeanVarianceNormalization, Gemm, MatMul.
• Decomposes unsupported QLinearSigmoid operation.
• Removes strided 64-bit emulation in Cast.
• Allows empty shapes on constant CPU inputs.
Known issues
- This release has an issue that may result in segmentation faults when deployed on Intel 12th Gen processors with hybrid architecture capabilities with Performance and Efficient-cores (P-core and E-core). This has been fixed in ORT 1.9.
- The CUDA build of this release has a regression in that the memory utilization increases significantly compared to the previous releases. A fix for this will be released shortly as part of 1.8.1 patch. Here is an incomplete list of issues where this was reported - 8287, 8171, 8147.
- GPU part of source code is not compatible with
- Visual Studio 2019 16.10.0 ( which was just released on May 25, 2021). 16.9.x is fine.
- clang 12
- CPU part of source code is not compatible with
- C# OpenVino EP is broken. #7951
- Python and Windows only: if your CUDNN DLLs are not in CUDA's installation dir, then you need to set manually "CUDNN_HOME" variable. Just putting them in %PATH% is not enough. #7965
- onnxruntime-win-gpu-x64-1.8.0.zip on this page misses important DLLs, please don't use it.
Contributions
Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:
snnn, gwang-msft, baijumeswani, fs-eire, edgchen1, zhanghuanrong, yufenglee, thiagocrepaldi, hariharans29, skottmckay, weixingzhang, tianleiwu, SherlockNoMad, ashbhandare, tracysh, satyajandhyala, liqunfu, iK1D, RandySheriffH, suffiank, hanbitmyths, wangyems, askhade, stevenlix, chilo-ms, smk2007, kit1980, codemzs, raviskolli, pranav-prakash, chenfucn, xadupre, gramalingam, harshithapv, oliviajain, xzhu1900, ytaous, MaajidKhan, RyanUnderhill, mrry, orilevari, jingyanwangms, sfatimar, KeDengMS, [jywu-msft](h...
ONNX Runtime v1.7.2
This is a minor patch release on 1.7.1 with the following changes:
- Fix Microsoft.AI.MachineLearning NuGet package to correctly install on C# UWP projects in Visual Studio.
ONNX Runtime v1.7.1
The Microsoft.ML.OnnxRuntime.Gpu and Microsoft.ML.OnnxRuntime.Managed packages are uploaded to Nuget.org. Please note the version numbers for the Microsoft.ML.OnnxRuntime.Managed package.