ONNX Runtime v1.17.1
This patch release includes the following updates:
General
Build System and Packages
- Fix bug that was breaking arm64 build by disabling __cpuid check on arm64 builds since intrinsic is not available (#19574) - @smk2007
Core
- Add capturestate / rundown ETW support logging for session and provider options (#19397) - @ivberg
- Restrict L2 cache core check on Intel devices (#19483) - @smk2007
Performance
- Optimize KahnsTopologicalSort and PriorityNodeCompare to fix performance degradation in session creation time that was affecting many models (#19475) - @smk2007
EPs
QNN
- Fix split index bugs uncovered by QNN SDK 2.19 release (#19381) - @adrianlizarraga
- Add job that builds x64 Python wheels for QNN EP so cached QNN models can be created on Windows x64 (#19499) - @adrianlizarraga
OpenVINO
- Fix bugs for API backwards compatibility (#19482) - @preetha-intel
DirectML
- Fix bug in external data packing that was causing crash (#19415) - @PatriceVignola
- Fix bug in allocation planner by disabling streams for DML EP (#19481) - @PatriceVignola
Web
Training
- Reduce onnxruntime-training package size so it can be published on PyPI (#19486) - @baijumeswani
- Update default std flag used during torch extensions compilation (#19516) - @baijumeswani
- Add ATen fallback support for bicubic interpolation algorithm (#19380) - @prathikr
Quantization
- Update Q/DQ quantization to ensure Microsoft opset (#19335) - @adrianlizarraga
- Add contrib Q/DQ ops to symbolic shape inference tool (#19340) - @adrianlizarraga
- Fix subgraph quantization regression (#19421) - @fxmarty
- Add DefaultTensorType option to specify the default tensor type to quantize (#19455) - @yufenglee
- Fix bug with command line argparse to process --symmetric [True|False] correctly (#19577) - @satyajandhyala
Whisper Model
- Fix bug in BeamSearch implementation of Whisper model that was causing a crash in some scenarios (#19345) - @petermcaughan
- Fix bug in Whisper model timestamps and temperature (#19509) - @kunal-vaishnavi