docs.unity3d.com
Search Results for

    Show / Hide Table of Contents

    Changelog

    All notable changes to this package will be documented in this file.

    The format is based on Keep a Changelog and this project adheres to Semantic Versioning.

    [2.6.1] - 2026-04-02

    Fixed

    • Documentation fix

    [2.6.0] - 2026-03-20

    Added

    • Officially support ONNX opset up to version 25
    • Functional methods for Swish and RMSNorm
    • Support for the alpha argument for the Swish operator
    • Improved analytics and error reporting when importing models with unsupported operators
    • Now supporting Fast Enter Play Mode for CoreCLR compatibility
    • Added support for Buffer in PyTorch import
    • Tokenizer: Generic truncation with support for longestfirst, onlyfirst and onlysecond

    Changed

    • Now using Unity's Mathematics.Random instead of System.Random
    • Improved documentation for the Tensor and Functional APIs
    • Maintenance of documentation links
    • Updated documentation for Cubic interpolation mode support limitations

    Fixed

    • Updated behavior of ReduceL1, ReduceL2, ReduceSumSquare and ReduceLogSum operators when argument noop_with_empty_axes is true and axes are empty
    • Interpolate issue with argument scaleFactor
    • Fix when GPU allocations (ComputeTensorData) are used for tensors without a corresponding backend.
    • Prevent crashes when closing the editor while in Play mode
    • Fixed memory leak in PyTorch Import
    • Fix GPU crash for convolution with padding on GPU compute
    • Fixed issue with Split operator when importing .sentis file

    [2.5.0] - 2026-01-23

    Added

    • PyTorch model import
    • LRN (Local Response Normalization) operator implemented on all backends
    • 3D MaxPool and AveragePool operators implemented on all backends
    • Sentis Importer: Allow users to specify dynamic dimensions as static on Sentis model import, same as we do for ONNX
    • Tokenizer Additions
      • Hugging Face parser
        • Sequence decoder
        • Regex replace decoder
        • String split pre-tokenizer
        • Unigram Mapper
        • Byte-based substring feature to SubString
        • Padding: support "pad multiple of" option
        • Split pre-tokenizers: support "invert"
        • StripAccents normalizer
        • Rune split pre-tokenizer
        • Strip normalizer
        • WordLevel model
        • WhitespaceSplit pre-tokenizer
        • Metaspace pre-tokenizer and decoder
        • Whitespace pre-tokenizer
        • NMT normalizer
        • Punctuation pre-tokenizer
        • Digits pre-tokenizer
        • CharDelimiterSplit pre-tokenizer
        • BPE decoder

    Changed

    • Model Visualizer: Async loading of model
    • Model Visualizer: updating com.unity.dt.app-ui to 1.3.3
    • Resize operator on CPU no longer uses main (mono) thread path
    • All model converters use switch-case instead of if-else cascade
    • Migrate Mono APIs to CoreCLR-compatible APIs

    Fixed

    • Editor crash when quitting in Play Mode
    • Memory Leak in FuseConstantPass
    • Clip operator improvement: no longer need CPU fallback for min/max parameters
    • Mod operator fix: on some platform with float operands, could return incorrect value when one of them was 0
    • ModelLoader: Exception handling now properly bubbles failures for importer integration and analytics reporting instead of silently catching exceptions
    • ModelLoader: Added robust stream reading with full buffer validation to prevent data corruption from partial reads and properly handle EOF scenarios
    • Faulty optimization pass
    • Fix in existing burst code for 2D pooling vectorization calculations
    • TopK issue on GPUCompute when dimension is specified
    • Fix source generator empty array
    • Tokenizer Fixes
      • Special added token decoding condition
      • Fix added token whole word handling
      • Gpt2Splitter subtring length computation
      • Added vocabulary pre-tokenization.
      • ByteLevelDecoder empty-byte guard in string generation
      • DefaultDecoder: joining tokens with whitespace
      • BPE: fix merging, applying on each word instead of the whole string
      • DefaultPostProcessor: apply the proper type id
      • RobertaPostProcessor: fix attention and type id assignment
      • TemplatePostProcessor: fix type id assignment
      • Assign default type id to sequences
      • Better surrogate characters support
      • Fix ByteFallback: inserting the right amount of \ufffd char
      • Fix BertPreTokenizer
      • Default model determination based of chain of responsibility

    [2.4.1] - 2025-10-31

    Fixed

    • Small error in documentation preventing user manual publication

    [2.4.0] - 2025-10-22

    Added

    • LiteRT model import
    • Tokenization API
    • STFT and DFT ONNX operators
    • BlackmanWindow, HammingWindow, HannWindow and MelWeightMatrix ONNX operators
    • BitwiseAnd, BitwiseOr, BitwiseXor, BitwiseNot ONNX operators and functional methods
    • AsStrided, Atan2, Expm1, Log10, Log1p, Log2, Rsqrt, Trunc, ReduceVariance, Diagonal layers, functional methods and optimizer passes
    • NotEqual, FloorDiv, TrueDiv layers and LiteRT operators

    Changed

    • Renamed Inference Engine to Sentis in package name and documentation
    • Improved model import time for ONNX models
    • ONNX model import operator order now consistent with the original model
    • Improved optimization passes to reduce operator count in imported models
    • Improved visualizer loading times and consistency in displaying attributes
    • ScatterND operator can now run on much larger tensors, enabling new models
    • ScatterND operator now allows negative indices
    • ONNX model outputs that are not connected to any inputs are no longer incorrectly pruned
    • Improve model import warning and error display in the inspector

    Fixed

    • Small errors in documentation
    • Faulty optimization passes that could lead to inference issues
    • Memory leaks on model constants
    • Non-matching ProfilerMarker calls
    • Issues in CPU callback which could lead to incorrect inference on some models
    • Enable missing modes for GridSample and Upsample operators

    [2.3.0] - 2025-07-15

    Added

    • Model Visualizer for inspecting models as node-based graphs inside the Unity Editor
    • Support for Tensor<int> input for GatherND operator on GPUPixel backend
    • Support for Tensor<int> input for the base of the Pow operator on all backends
    • Support for the group and dilations arguments for the ConvTranspose operator on all backends
    • Support for value_float, value_floats, value_int and value_ints values in ONNX Constant operators

    Changed

    • Optimized single-argument operators on CPU backend
    • Optimized deserialization of models to avoid reflection at runtime

    Fixed

    • Einsum operator now works correctly on fallback path

    [2.2.1] - 2025-05-28

    Fixed

    • Issue with incorrect TensorShape in Conv layer when dilations are greater than 1 and auto-padding is used
    • Incorrect Third Party Notices

    [2.2.0] - 2025-05-15

    Added

    • First version of Inference Engine
    In This Article
    Back to top
    Copyright © 2026 Unity Technologies — Trademarks and terms of use
    • Legal
    • Privacy Policy
    • Cookie Policy
    • Do Not Sell or Share My Personal Information
    • Your Privacy Choices (Cookie Settings)