-
Amazon AGI
- New York City
- https://www.linkedin.com/in/shengzha/
- @szha_
Stars
Machine Learning Engineering Open Book
程序员在家做饭方法指南。Programmer's guide about how to cook at home (Simplified Chinese only).
FlashMLA: Efficient Multi-head Latent Attention Kernels
A simple toolchain for moving Remarkable highlights to Readwise
A curated list of projects related to the reMarkable tablet
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
A prize for finding tasks that cause large language models to show inverse scaling
🦜🔗 The platform for reliable agents.
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
Open Academic Research on Improving LLaMA to SOTA LLM
A playbook for systematically maximizing the performance of deep learning models.
Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.
Fast and memory-efficient exact attention
A list of ICs and IPs for AI, Machine Learning and Deep Learning.
Qiskit is an open-source SDK for working with quantum computers at the level of extended quantum circuits, operators, and primitives.
Making large AI models cheaper, faster and more accessible
Full description can be found here: https://discuss.huggingface.co/t/pretrain-gpt-neo-for-open-source-github-copilot-model/7678?u=ncoop57
Training neural networks in TensorFlow 2.0 with 5x less memory
This repository contains implementations and illustrative code to accompany DeepMind publications
Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727
Probabilistic programming with NumPy powered by JAX for autograd and JIT compilation to GPU/TPU/CPU.
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities