Lists (1)
Sort Name ascending (A-Z)
Stars
Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.
Generative AI extensions for onnxruntime
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime
Examples for using ONNX Runtime for machine learning inferencing.
Reference implementations of MLPerf® inference benchmarks
Vitis AI is Xilinx’s development stack for AI inference on Xilinx hardware platforms, including both edge devices and Alveo cards.
Board files to build Ultra 96 PYNQ image
HLS based Deep Neural Network Accelerator Library for Xilinx Ultrascale+ MPSoCs