SyNAP 版本说明
Version 3.0.0
Release date: 2024.03.01
Runtime
New features
Support for models in .synap format (C++ API fully backward compatible with SyNAP 2.x)
Support for heterogeneous model execution (NPU, CPU, GPU)
Full support and compatibility with legacy SyNAP 2.x models (model.nb and model.json)
Integrate Onnx runtime 1.16.3
Integrate TFlite runtime 2.15
TimVx delegate for TFlite for improved online inference on NPU
Optimized model benchmark binary
benchmark_model
integrating TimVx delegate
Toolkit
New features
Support for heterogeneous model conversion (NPU, CPU, GPU). The desired delegate(s) can be selected at model compilation time.
Generates .synap format by default. The .synap format is a bundle that contains both the model subgraph(s) and the companion meta-information. This replaces the model.nb and model.json files in SyNAP 2.x. It’s still possible to generate the legacy model.nb and model.json files compatible with SyNAP 2.x runtime by specifying the
--out-format nb
option when converting the model.New preprocessing option to accept model input in 32-bits floating point
Improvements and updates
Inference time for some models with mixed quantization
Fixes
Preprocessing support for non-quantized models
Mixed quantization for tflite models (“No layers to requantize” error)
Accuracy issues with some models when using mixed quantization
Version 2.8.1
Toolkit
Fixes
Import of Tensorflow .nb models
Import of ONNX models containing MaxPool layers
Import of ONNX models containing Slice layers
Version 2.8.0
Runtime
New features
Detector now supports ‘yolov8’ format
Tensor now supports assignment from float and int16 data
Improvements and updates
Image Preprocessor now adds horizontal or vertical gray bars when importing an image to preserve proportions. The image is always kept at the center.
Verisilicon software stack upgraded to Ovxlib 1.1.84
Improvements and clarifications in SyNAP.pdf user manual
Fixes
Layer-by-layer profiling now provides more accurate timings
Toolkit
Improvements and updates
Verisilicon Acuity 6.21.2
Conversion docker updated to Tensorflow 2.13.0 and ONNX 1.14.0
Fixes
Issues with mixed quantization with some models
Version 2.7.0
Runtime
New features
Face recognition support
Optional OpenCV support
Model import tutorial: SyNAP_ModelImport.pdf
Improvements and updates
Load network directly from a user buffer (avoid data copy)
Verisilicon software stack upgraded to Unify driver 6.4.13 and ovxlib 1.1.69
Improvements and clarifications in SyNAP.pdf user manual
Fixes
Bounding box scaling in postprocessing for ‘yolov5’ format
Toolkit
New features
Model preprocessing now supports nv12 format
Improvements and updates
Verisilicon Acuity 6.15.0
Conversion docker to ubuntu 22.04 and tensorflow 2.10.0
Fixes
Import of .pb models when post-processing enabled (skip reordering)
Support relative model pathnames in model_convert.py
Version 2.6.0
Runtime
New features
Tensor assign() supports data normalization
Preprocessor supports 16-bits models
Preprocessor supports models with preprocessing and cropping
Preprocessor rescale now preserves the input aspect-ratio by default (a gray band is added on the bottom of the image if needed)
Support for scalar tensors
Detector supports yolov5 output format
Buffer sharing (allows to share the tensor memory between different networks avoiding data copy)
Improvements and updates
Support 64bits compilation
Fixes
Fix Tensor::set_buffer in case the same Buffer is assigned/deassigned multiple times
Fix model json parsing for 16-bits models
Toolkit
New features
Support compilation of models with embedded preprocessing including: format conversion (eg. YUV to RGB), layout conversion (eg. NCHW to NHWC), normalization and cropping
Support “full” model quantization mode
Mixed quantization: the user can mix 8-bits and 16-bits quantization in the same model by specifying the quantization type for each layer
Improvements and updates
Quantization images now rescaled preserving the aspect-ratio of the content
Version 2.5.0
Runtime
New features
Support for NNAPI compilation cache
Move support for Network objects in C++ API
Unified libovxlib.so supporting both VS640 and VS680
Improvements and updates
Faster init time for NNAPI online inference (release mode)
Error checking on out-of-sequence API calls
Accuracy of layer-by-layer metrics
Unify all logcat messages with “SyNAP” tag
Memory optimization: on-demand loading of compressed OpenVX kernels (saves more than 80MB of RAM)
Verisilicon software stack upgraded to Unify driver 6.4.11 and ovxlib 1.1.50
Overall improvements now achieve a score of 33.8 with AIBenchmark 4.0.4
Fixes
Layer-by-layer metrics was not working on some models (inference fail)
Toolkit
New features
Support compilation of Caffe models
Improvements and updates
Verisilicon Acuity 6.9
Error reporting for quantization issues
Version 2.4.0
Runtime
New features
New internal SyNAP model compilation cache This dramatically improves model initialization time during the first inference. Typical speedup of the first inference is by a factor of 3, can be a factor of 20 or more on some models.
Improvements and updates
Further runtime optimizations allowing VS680 to achieve a score of 31.5 in ai-benchmark 4.0.4
SyNAP default log level is now WARNING (instead of no logs)
Operator support table updated in User Manual
Fixes
Correctly support multiple online models at the same time Compiling multiple online models in parallel could in some cases give issues (SyNAP HAL crash) in previous releases.
Toolkit
New features
New internal SyNAP model compilation cache This dramatically improves model compilation time. Typical speedup is by a factor of 3, can be a factor of 20 or more on some models.
Fixes
Conversion of ONNX models when output layer name(s) specified explicitly in metafile
Version 2.3.0
Runtime
New features
By-layer profiling support Low-level driver and runtime binaries and libraries now support layer by layer profiling of any network.
Allocator API in synap device and associated SE-Linux rules This is the default allocator in libsynapnb and the NNAPI is already making use of it. This also enable any user application (native or not) to execute models without root priviledge, including the synap_cli family.
Sample Java support
Improvements and updates
Reorganize libraries. We now have the following libraries:
libsynapnb.so: core EBG execution library
libsynap_preprocess.a: pre-processing
libsynap_postprocess.a: post-processing (classification, detection, etc)
libsynap_img.a: image processing utilities
libsynap_utils.a: common utilities
libsynap_app.a: application support utilities
Repeat mode to synap_cli
EBG for profiling generation to synap_cli_nb
Fixes
Memory leak when running models
Toolkit
New features
By-layer profiling
Secure Model Generation for VS640 (VS680 was already supported) Note: This feature requires special agreement with Synaptics in order to be enabled.
Version 2.2.1
Runtime
New features
New NNHAL architecture (NNAPI) NNRT is now using libsynapnb directly to execute an EBG model; this saves memory and simplify dependencies. VIPBroker dependency was removed from OVXLIB which is now only used as a graph compiler.
Fixes
Memory leak when dellocating Buffers
Version 2.2.0
Runtime
Component |
Type |
Description |
---|---|---|
all |
Add |
Linux Baseline VSSDK support |
lib |
Add |
|
lib |
Add |
|
lib |
Add |
Detector` postprocessors with full support for
|
lib |
Add |
|
lib |
Add |
|
lib |
Fix |
NPU lock functionality |
lib |
Remove |
|
bin |
Add |
|
driver |
Optimize |
Much reduced usage of contiguous memory |
NNAPI |
Update |
VSI OVXLIB to 1.1.37 |
NNAPI |
Update |
VSI NNRT/NHAL to 1.3.1 |
NNAPI |
Add |
More operators supported |
NNAPI |
Optimize |
Much higher score for some AI-benchmark models (ex: PyNET and U-Net) |
NNAPI |
Add |
Android CTS/VTS pass for both VS680 and VS640 |
Toolkit
Type |
Description |
---|---|
Fix |
Crash when importing one TFLite object-detection models |
Add |
Full support for TFLite_Detection_PostProcess layerb |
Add |
Support for ${ANCHOR} and ${FILE:name} variables in tensor format string |
Add |
Support for ${ENV:name} variables substitution in model yaml metafile |
Add |
Support for security.yaml file |
Update |
VSI acuity toolkit to 6.3.1 |
Update |
Improved error checking |
Update |
Layer name and shape are now optional when doing quantization |
Add |
Support for single mean value in metafile |
Remove |
synap_profile tool |
Fix |
Handling of relative paths |
Version 2.1.1
Runtime
Type |
Description |
---|---|
Fix |
Timeout expiration in online model execution (ai-benchmark 4.0.4 now runs correctly) |
Fix |
Issues in |
Change |
On android |
Toolkit
Type |
Description |
---|---|
Fix |
|
Update |
Inference timings section in User Manual now includes y-uv models |
Version 2.1.0
Runtime
Type |
Description |
---|---|
Add |
Full support for SyKURETM: Synaptics secure inference technology |
Improve |
Tensor Buffers for NNAPI and synapnb now allocated in non-contiguous memory by default |
Add |
Buildable source code for |
Change |
Per-target organization of libraries and binaries in the install tree |
Toolkit
Type |
Description |
---|---|
Add |
Support for NHWC tensors in rescale layer |
Fix |
Tensor format in json file for converted models |
Improve |
Reorganize sections in User Manual |
Version 2.0.1
Runtime
Type |
Description |
---|---|
Improve |
Online inference performance |
Add |
Option to show SyNAP version in synap_cli application |
Add |
Buildable source code for all SyNAP sample applications and libraries |
Toolkit
Type |
Description |
---|---|
Update |
Model coversion tool (fixes offline performance drop in some cases) |
Version 2.0.0
Runtime
Type |
Description |
---|---|
Improve |
Inference engine now supports the new EBG (Executable Binary Graph) model format. Compared to previous NBG format, EBG brings several impovements:
NBG models are not supported anymore. |
Toolkit
Type |
Description |
---|---|
Update |
Model coversion tools now support EBG generation |
Version 1.5.0
Runtime
Type |
Description |
---|---|
Add |
Synap device information and statistics in sysfs |
Toolkit
Type |
Description |
---|---|
Update |
Conversion toolkit to v. 5.24.5 |
Improve |
Model quantization algorithm |
Add |
Generate network information file when model is converted |
Add |
Host tools binaries and libraries in toolkit/bin and toolkit/lib |
Version 1.4.0
Runtime
Type |
Description |
---|---|
Fix |
CTS/VTS now run successfully with NNAPI |
Toolkit
Type |
Description |
---|---|
Update |
Conversion toolkit to v. 5.24 |
Add |
Model benchmark binary: /vendor/bin/android_arm_benchmark_model |
Add |
Model test script and specs |
Version 1.3.0
Runtime
Type |
Description |
---|---|
Change |
Update and cleanup object Detector API |
Change |
synap_cli_od allows to specify model |
Add |
synap_cli_od source code |
Add |
Cmake standalone build for synap_cli_ic and synap_cli_od |
Toolkit
Type |
Description |
---|---|
Add |
Import and conversion of ONNX models |
Version 1.2.0
Runtime
Type |
Description |
---|---|
Change |
Remove private implementation details from Buffer.hpp |
Change |
Switch memory allocation to dmabuf |
Fix |
Model pathnames and documentation for object detection |
Add |
Synap device |
Add |
OpenVX headers and librairies |
Toolkit
Type |
Description |
---|---|
New |
Model quantization support |
Version 1.1.0
Runtime
Type |
Description |
---|---|
New |
NNAPI lock support: |
Toolkit
Type |
Description |
---|---|
New |
Model profiling tool: |
New |
NNAPI benchmarking script: |
Version 1.0.0
Initial Version.