Tensors and Dynamic neural networks in Python with strong GPU acceleration
-
Updated
Nov 6, 2025 - Python
Tensors and Dynamic neural networks in Python with strong GPU acceleration
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals
Open deep learning compiler stack for cpu, gpu and specialized accelerators
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 17+ clouds, or on-prem).
A python library built to empower developers to build applications and systems with self-contained Computer Vision capabilities
Real-Time and Accurate Full-Body Multi-Person Pose Estimation&Tracking System
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, vLLM, DeepSpeed, Axolotl, etc.
An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.
A flexible framework of neural networks for deep learning
A Python framework for accelerated simulation, data generation and spatial computing.
High-performance TensorFlow library for quantitative finance.
Time series forecasting with PyTorch
📊 A simple command-line utility for querying and monitoring GPU status
On-device AI across mobile, embedded and edge for PyTorch
Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators.
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory utilization in both training and inference.
Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.
To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."