oneAPI Deep Neural Network Library (oneDNN)
-
Updated
Mar 4, 2026 - C++
oneAPI Deep Neural Network Library (oneDNN)
Up to 200x Faster Dot Products & Similarity Metrics — for Python, Rust, C, JS, and Swift, supporting f64, f32, f16 real & complex, i8, and bit vectors using SIMD for both AVX2, AVX-512, NEON, SVE, & SVE2 📐
Half-precision floating point types f16 and bf16 for Rust.
Round matrix elements to lower precision in MATLAB
C++ template library for floating point operations
Floating-Point Arithmetic Library for Z80
CUDA/HIP header-only library for low-precision (16 bit, 8 bit) and vectorized GPU kernel development
IEEE 754-style floating-point converter
A LLaMA2-7b chatbot with memory running on CPU, and optimized using smooth quantization, 4-bit quantization or Intel® Extension For PyTorch with bfloat16.
A JAX implementation of stochastic addition.
A Pytorch implementation of stochastic addition.
Customizable floating point types, with all standard floating point operations implemented from scratch.
Basic linear algebra routines implemented using the chop rounding function
IHP 130nm ASIC tapeout of a 2x2 bfloat16 matrix matrix multiplication with DFT infrastructure. Iteration on the previous accelerator taped out on GF180.
Comparison of vector element sum using various data types.
Comparison of PageRank algorithm using various datatypes.
Add a description, image, and links to the bfloat16 topic page so that developers can more easily learn about it.
To associate your repository with the bfloat16 topic, visit your repo's landing page and select "manage topics."