Stars
An Open Source Machine Learning Framework for Everyone
GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.
A library for efficient similarity search and clustering of dense vectors.
DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.
Solidity, the Smart Contract Programming Language
Ceph is a distributed object, block, and file storage platform
A toolkit for making real world machine learning and data analysis applications in C++
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. Tensor…
Facebook AI Research's Automatic Speech Recognition Toolkit
Transformer related optimization, including BERT, GPT
🐸STT - The deep learning toolkit for Speech-to-Text. Training and deploying STT models has never been so easy.
A fast communication-overlapping library for tensor/expert parallelism on GPUs.
💻 Solutions to coding problems from popular platforms in multiple programming languages.
janhq / cortex.tensorrt-llm
Forked from NVIDIA/TensorRT-LLMCortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU accelerated inference on NVIDIA's GPUs.