Performance Optimizations
- Intel Architecture processors
- Improved performance of int8 compute functionality for future Intel Xeon Scalable processor (code name Sapphire Rapids). The functionality is disabled by default and should be enabled via CPU dispatcher control.
- Improved performance of compute functionality for future Intel Core processor with Intel AVX2 and Intel DL Boost instructions support (code name Alder Lake).
- Improved fp32 inner product forward propagation performance for processors with Intel AVX-512 support.
- Improved
dnnl_gemm
performance for cases withn=1
on all supported processors.
- Intel Graphics products
- Introduced NHWC format support for activations for int8 primitives.
- AArch64-based processors
- Improved performance of fp32 and int8 convolution, and softmax primitives for processors with SVE 512 support.
- Improved performance of fp32 convolution via Arm Compute Library (ACL).
- Improved performance of convolution with a combination of
sum
andrelu
post-ops via ACL.
Functionality
- Extended eltwise primitive with support for
mish
andhardswish
algorithms. - Extended binary primitive with support for comparison operators.
- Introduced support for post-ops in GPU resampling implementation.
- Introduced asymmetric quantization support for int8 deconvolution.
- Introduced binary post-ops support for matmul primitive.
Usability
- Improved presentation of oneDNN primitives in VTune Amplifier.
- Introduced Linux perf support for AArch64.
- Introduced support for Fujitsu C++ compiler.
- Introduced a build time check for minimal supported ACL version. Currently oneDNN requires ACL 21.02 or later.
- Added support for cuDNN 8.x
Thanks to the contributors
This release contains contributions from the project core team as well as Aleksandr Nikolaev @alenik01, araki.kenichi @qnet-araki, Arthur Mitrano @aaraujom, Dr-Noob @Dr-Noob, Gmc2 @GHGmc2, higuchi.motoko @higuchi-motoko, Joe Ramsay @joeramsay, Kentaro Kawakami @kawakami-k, Louie Tsai @louie-tsai, masafumi yamazaki @m-ymzk, Nathan John Sircombe @nSircombe, Takumi-H @Takumi-Honda. We would also like to thank everyone who asked questions and reported issues.