com.unity.ai.inference 2.5.0
Source docsNotes: processedReleased January 29, 2026
Browse versions
PreviousNext
Unity Compatibility
Minimum Unity: 6000.0
Package Dependencies
- com.unity.burst (1.8.17)
- com.unity.collections (2.4.3)
- com.unity.dt.app-ui (1.3.3)
- com.unity.modules.imageconversion (1.0.0)
- com.unity.nuget.newtonsoft-json (3.2.1)
✨ Features
- `PyTorch` model import
- `LRN (Local Response Normalization)` operator implemented on all backends
- `3D MaxPool` and `AveragePool` operators implemented on all backends
- Sentis Importer: Allow users to specify dynamic dimensions as static on Sentis model import, same as we do for ONNX
- Tokenizer Additions
- - `Hugging Face` parser
- - Sequence decoder
- - Regex replace decoder
- - String split pre-tokenizer
- - Unigram Mapper
- - Byte-based substring feature to SubString
- - Padding: support "pad multiple of" option
- - Split pre-tokenizers: support "invert"
- - StripAccents normalizer
- - Rune split pre-tokenizer
- - Strip normalizer
- - WordLevel model
- - WhitespaceSplit pre-tokenizer
- - Metaspace pre-tokenizer and decoder
- - Whitespace pre-tokenizer
- - NMT normalizer
- - Punctuation pre-tokenizer
- - Digits pre-tokenizer
- - CharDelimiterSplit pre-tokenizer
- - BPE decoder
📈 Improvements
- Model Visualizer: Async loading of model
- Model Visualizer: updating com.unity.dt.app-ui to 1.3.3
- Resize operator on CPU no longer uses main (mono) thread path
- All model converters use switch-case instead of if-else cascade
- Migrate Mono APIs to CoreCLR-compatible APIs
🔧 Bug Fixes
- Editor crash when quitting in play mode
- Memory Leak in FuseConstantPass
- `Clip` operator improvement: no longer need CPU fallback for min/max parameters
- `Mod` operator fix: on some platform with float operands, could return incorrect value when one of them was 0
- Faulty optimization pass
- Fix in existing burst code for 2D pooling vectorization calculations
- `TopK` issue on `GPUCompute` when dimension is specified
- Fix source generator empty array
- Tokenizer Fixes
- - Special added token decoding condition
- - Fix added token whole word handling
- - Gpt2Splitter subtring length computation
- - Added vocabulary pre-tokenization.
- - ByteLevelDecoder empty-byte guard in string generation
- - DefaultDecoder: joining tokens with whitespace
- - BPE: fix merging, applying on each word instead of the whole string
- - DefaultPostProcessor: apply the proper type id
- - RobertaPostProcessor: fix attention and type id assignment
- - TemplatePostProcessor: fix type id assignment
- - Assign default type id to sequences
- - Better surrogate characters support
- - Fix ByteFallback: inserting the right amount of \ufffd char
- - Fix BertPreTokenizer
- - Default model determination based of chain of responsibility
