A curated list of awesome AI tools, libraries, papers, datasets, and frameworks that accelerate scientific discovery across all disciplines.
Deutsch | English | Español | français | 日本語 | 한국어 | Português | Русский | 中文
AI is revolutionizing scientific research - from drug discovery and materials design to climate modeling and astrophysics. This repository collects the best resources to help researchers leverage AI in their work.
- 🧪 AI Tools for Research
- 📄 Paper→Poster / Slides / Graphical Abstract
- 📊 Chart Understanding & Generation
- 🔄 Paper-to-Code & Reproducibility
- 📋 Scientific Documentation & Parsing
- 🧰 Research Workbench & Plugins
- 🕸 Knowledge Extraction & Scholarly KGs
- 🤖 Research Agents & Autonomous Workflows
- 🏷 Data Labeling & Curation
- ⚗ Scientific Machine Learning
- 📖 Papers & Reviews
- 🔬 Domain-Specific Applications
- 🤖 Foundation Models for Science
- 📈 Datasets & Benchmarks
- 💻 Computing Frameworks
- 🎓 Educational Resources
- 🏛 Research Communities
- 📚 Related Awesome Lists
- Semantic Scholar - AI-powered academic search (Allen AI)
- arXiv - Open-access repository of electronic preprints and postprints
- OpenAlex - Open catalog of scholarly papers and authors
- CORE - Aggregator of open access research papers
- PandasAI - Conversational data analysis using natural language
- DeepAnalyze - First agentic LLM for autonomous data science with end-to-end pipeline from data to analyst-grade reports
- AutoViz - Automated data visualization with minimal code
- Chat2Plot - Secure text-to-visualization through standardized chart specifications
- Label Studio - Multi-type data labeling and annotation tool
- Snorkel - Programmatic data labeling and weak supervision
- Claude Scientific Skills - Comprehensive collection of 125+ ready-to-use scientific skill modules for Claude AI across bioinformatics, cheminformatics, clinical research, ML, and materials science
- Paper2Poster - Multi-agent system with Parser-Planner-Painter architecture converting
paper.pdfto editableposter.pptx, outperforms GPT-4o with 87% fewer tokens - mPLUG-PaperOwl - Multimodal LLM for scientific charts and diagrams understanding/generation
- Auto-Slides - Multi-agent academic paper to high-quality presentation slides with interactive refinement
- PPTAgent - Beyond text-to-slides generation with PPTEval multi-dimensional evaluation (EMNLP 2025)
- paper2slides - Transform arXiv papers into Beamer slides using LLMs
- PaperToSlides - AI-powered tool that automatically converts academic papers (PDF) into presentation slides
- pdf2slides - Convert PDF files into editable slides with three lines of code
- SlideDeck AI - Co-create PowerPoint presentations with Generative AI from documents or topics
- AI Multi-Agent Presentation Builder - Azure Semantic Kernel multi-agent PPT generation reference
- Paper2Video - First benchmark for automatic video generation from scientific papers (NeurIPS 2025)
- paper2video - Transform arXiv research papers into engaging presentations and YouTube-ready videos
- Paper2All - AI-powered pipeline converting papers into interactive websites, posters, and multimedia presentations with "Let's Make Your Paper Alive!" philosophy
- PaperBanana - Automated academic illustration generation for AI scientists, converting research papers into publication-ready figures using VLMs and diffusion models with iterative refinement (PKU & Google Research, 6.2K+ stars, 2026)
Note: For comprehensive chart understanding and code generation tools, see 📊 Chart Understanding & Generation section
- ChartCoder (ACL 2025) - Multimodal LLM for chart-to-code generation, 7B model outperforms larger open-source MLLMs
- ChartAssistant / ChartAst (ACL 2024) - Universal chart comprehension and reasoning model
- Chart-to-Text Datasets - Large-scale chart summarization datasets for training chart description capabilities
- Chat2Plot - Secure text-to-visualization through standardized chart specifications
- AutoViz - Automated data visualization with minimal code
- PlotlyAI - AI-powered data visualization and dashboard creation
- Paper2Code - Automated code generation from machine learning research papers into runnable implementations (4.5K+ stars, 2025)
- AutoP2C - LLM agent framework generating runnable repositories from academic papers
- ResearchCodeAgent - Multi-agent system for automated codification of research methodologies
- ToolMaker - Convert papers with code into callable agent tools
- BioProBench - Comprehensive benchmark for automatic evaluation of LLMs on biological protocols and procedural understanding
- Alhazen - Extract experimental metadata and protocol information from scientific documents
- MinerU (2024/2025) - SOTA multimodal document parsing with 1.2B parameters outperforming GPT-4o, converts PDFs to LLM-ready Markdown/JSON
- PDF-Extract-Kit (2024) - Comprehensive toolkit for high-quality PDF content extraction with layout detection, formula recognition, and OCR
- Docling (IBM, AAAI 2025) - Multi-format (PDF/DOCX/PPTX/HTML/Images) → structured data (Markdown/JSON) with layout reconstruction, table/formula recovery
- Nougat (Meta AI) - Neural optical understanding for academic documents, transforms scientific PDFs to Markdown with mathematical formula support
- PaddleOCR 3.0 (2024/2025) - Advanced OCR with PP-StructureV3 document parsing, 13% accuracy improvement, supports 80+ languages
- Unstructured - Production-grade ETL for transforming complex documents into structured formats, with open-source API
- Marker - High-accuracy PDF→Markdown/JSON/HTML conversion, specialized for tables/formulas/code blocks with benchmark scripts
- S2ORC doc2json (AllenAI) - Large-scale PDF/LaTeX/JATS parsing to standardized JSON for millions of papers
- GROBID - Machine learning software for extracting structured metadata from scholarly documents
- Science-Parse / SPv2 (AllenAI) - Parse scientific papers to structured fields (title/author/sections/references)
- IBM Data Prep Kit: PDF→Parquet - Large-scale scientific document ingestion pipeline with optimization configurations
- Mozilla document-to-markdown - Docling-powered parsing with UI/CLI demonstration for rapid prototyping
- PDFFigures2 - Extract figures, tables, captions, and section titles from scholarly PDFs
- TableBank - Large-scale table detection and recognition dataset with pre-trained models
- PaperQA2 - High-accuracy RAG for scientific PDFs with citation support, agentic RAG, and contradiction detection
- OpenScholar - Retrieval-augmented LM synthesizing scientific literature from 45M papers with human-expert-level citation accuracy, outperforming GPT-4o by 5% on ScholarQABench (Nature 2026, UW & Ai2)
- Valsci - Self-hostable scientific claim-verification and literature-review tool combining Semantic Scholar retrieval, bibliometric scoring, and LLM-based evidence synthesis for large-batch validation workflows
- paper-reviewer - Generate comprehensive reviews from arXiv papers and convert to blog posts
- Jupyter AI (JupyterLab Extension) - Official Jupyter extension with
%%aimagic commands and sidebar chat assistant, connecting multiple model providers and local inference - Notebook Intelligence (NBI) - AI coding assistant for JupyterLab with agent mode, supporting arbitrary LLM providers (2025+)
- Google Colab AI Features - Integrated AI assistance for data science and research notebooks
- PapersGPT for Zotero - Multi-PDF conversation, retrieval, and citation in Zotero with commercial/local models (Ollama), MCP support
- Zotero-GPT (MuiseDestiny) - Classic open-source plugin for document Q&A and summarization within Zotero
- Better BibTeX for Zotero - Enhanced citation key management and LaTeX integration
- Notion AI - AI-powered research note-taking and knowledge management
- Obsidian Smart Connections - AI-powered note linking and research graph navigation
- Research Rabbit - AI-powered literature discovery and research network mapping
- SciWrite - Agent skill for AI-assisted scientific manuscript writing review distilled from Stanford's Writing in the Sciences course, performing five sequential editorial audit passes on clarity, voice, structure, consistency, and integrity (2026)
- iText2KG - Incremental knowledge graph construction using LLMs with entity extraction and Neo4j visualization
- GraphGen - Knowledge graph-guided synthetic data generation for LLM fine-tuning, achieving strong performance on scientific QA (GPQA-Diamond) and math reasoning (AIME)
- KoPA - Structure-aware prefix adaptation for integrating LLMs with knowledge graphs (ACM MM 2024)
- Scholarly KGQA - LLM-powered question answering over scholarly knowledge graphs (ArXiv paper)
- Awesome-LLM-KG - Comprehensive collection of papers on unifying LLMs and knowledge graphs
- FunSearch (DeepMind, Nature 2023) - First system to make novel, verifiable scientific discoveries by pairing LLMs with evolutionary search, solving open problems in combinatorics (cap set problem) and discovering faster matrix multiplication algorithms
- OpenEvolve - Open-source implementation of AlphaEvolve's evolutionary coding agent paradigm, enabling LLMs to autonomously discover and optimize algorithms through iterative evolution, matching the approach behind DeepMind's breakthrough matrix multiplication discovery (6.2K+ stars, 2025)
- The AI Scientist v1 (2024) - First fully autonomous research system: hypothesis→experiment→writing→review simulation
- The AI Scientist v2 (2025) - Enhanced with Agentic Tree Search, reduced template dependency, first workshop-level accepted paper
- DeepScientist - First system progressively surpassing human SOTA on frontier AI tasks (183.7%, 1.9%, 7.9% improvements), month-long autonomous discovery with 20,000+ GPU hours
- Kosmos - Extended autonomy AI scientist with 200 parallel agent rollouts, 42K lines of code execution, 1.5K papers analyzed per run, achieving 79.4% accuracy and 7 scientific discoveries (Edison Scientific)
- AlphaResearch - Autonomous algorithm discovery combining evolutionary search with peer-review reward models, achieving best-known performance on circle packing problems
- AutoResearchClaw - Fully autonomous research from idea to paper with multi-agent debate, citation verification, and OpenClaw integration (11K+ stars, 2026)
- AI-Researcher - Autonomous pipeline from literature review→hypothesis→algorithm implementation→publication-level writing with Scientist-Bench evaluation
- Agent Laboratory - Multi-agent workflows for complete research cycles with AgentRxiv for cumulative discovery
- InternAgent - Closed-loop multi-agent system from hypothesis to verification across 12 scientific tasks, #1 on MLE-Bench (36.44%)
- freephdlabor - First fully customizable open-source multiagent framework automating complete research lifecycle from idea conception to LaTeX papers with dynamic workflows
- ToolUniverse - Democratizing AI scientists by transforming any LLM into research systems with 600+ scientific tools (Harvard MIMS)
- LabClaw - Skill operating layer for biomedical AI agents with 211 production-ready SKILL.md files across 7 domains (biology, pharmacology, medicine, data science, literature search), enabling modular dry-lab reasoning and protocol composition for Stanford LabOS-compatible agents
- Robin - FutureHouse's end-to-end scientific discovery multi-agent system orchestrating literature search (Crow/Falcon) and data analysis (Finch) agents, first AI-generated drug discovery identifying ripasudil as novel dry AMD therapeutic (2025)
- Aviary - Language agent gymnasium for challenging scientific tasks including DNA manipulation, literature search, and protein engineering
- Curie - Automated and rigorous experiments using AI agents for scientific discovery
- POPPER - Automated hypothesis testing with agentic sequential falsifications
- autoresearch - Andrej Karpathy's autonomous LLM research framework: AI agent runs overnight experiments on a real training setup, auto-editing code→5min training→evaluation in a loop, ~100 experiments per night on a single GPU
- UniScientist - Universal scientific research intelligence covering 50+ disciplines, repositioning LLMs as cross-disciplinary generators with human experts as verifiers; 30B model outperforms Claude Opus and GPT on 5 research benchmarks
- EvoScientist - Self-evolving AI scientist with 6 specialized sub-agents (plan/research/code/debug/analyze/write) and persistent memory, #1 on DeepResearch Bench II and AstaBench, supporting multi-provider LLMs and multi-channel deployment (Apache 2.0, 2026)
- ScienceAgentBench (ICLR 2025) - 102 executable tasks from 44 peer-reviewed papers across 4 disciplines with containerized evaluation
- PaperBench (OpenAI, 2025) - Benchmark evaluating AI agents' ability to replicate 20 ICML 2024 Spotlight/Oral papers from scratch, with 8,316 gradable tasks and author-co-developed rubrics
- MLE-Bench (OpenAI, 2024) - Benchmark evaluating AI agents on 75 curated Kaggle-style ML engineering competitions with reproducible Docker-based grading harness, human baselines, and end-to-end task lifecycle, used as a primary benchmark for autonomous ML research agents (e.g., InternAgent #1 at 36.44%)
- ScienceBoard (ICLR 2026) - Evaluating multimodal autonomous agents in realistic scientific workflows across real scientific software environments (KAlgebra, Celestia, Grass GIS, Lean 4, etc.) with VM-based evaluation infrastructure and agent trajectories
- BuildArena - First physics-aligned interactive benchmark for LLM agents in engineering construction, designing rockets/cars/bridges in physics simulator with 3D spatial geometry library
- SciTrust (2024) - Trustworthiness evaluation framework for scientific LLMs (truthfulness, hallucination, sycophancy)
- SciCode - Research coding benchmark curated by scientists with 338 subproblems across 16 subdomains (physics, math, materials, biology, chemistry), evaluating LLMs on realistic scientific programming tasks with gold-standard solutions (NeurIPS 2024)
- SciBench - College-level scientific problem-solving evaluation across multiple domains
- NewtonBench (ICLR 2026) - First benchmark evaluating LLMs' ability to rediscover scientific laws through interactive experimentation across 324 tasks in 12 physics domains, featuring memorization-resistant metaphysical shifts of canonical laws (HKUST)
- AgentReview - LLM agents simulating academic peer review ecosystems
- LLM-Peer-Review - Web application for LLM-assisted manuscript review and annotation
- Aletheia - Google DeepMind's autonomous mathematics research agent powered by Gemini Deep Think, autonomously solving 4 open problems from 700 Erdős conjectures and generating complete research papers without human intervention (February 2026)
- AlphaGeometry - DeepMind's Olympiad-level geometry theorem prover combining neural language model with symbolic deduction engine, AlphaGeometry2 solves 84% of IMO geometry problems (42/50) at gold-medalist level (Nature 2024)
- Goedel-Prover-V2 - Strongest open-source automated theorem prover in Lean 4, 8B model matches DeepSeek-Prover-V2-671B at 84.6% MiniF2F, 32B model achieves 90.4% with self-correction, using scaffolded data synthesis and verifier-guided proof refinement (Princeton, 2025)
- DeepSeek-Prover-V2 - DeepSeek's open-source large language model for formal theorem proving in Lean 4, integrating informal and formal mathematical reasoning through recursive subgoal decomposition and reinforcement learning powered by DeepSeek-V3, with open weights and ProverBench evaluation (2025)
- LeanDojo - Open-source toolkit and benchmark for learning-based theorem proving in Lean, providing programmatic Lean interaction, a 98K+ theorem dataset extracted from 217 Lean projects, and ReProver—the first retrieval-augmented LLM-based theorem prover for Lean—with reproducible training pipelines underpinning much subsequent Lean prover research (Caltech & NVIDIA, NeurIPS 2023 Outstanding Paper, Datasets & Benchmarks)
- Lean Copilot - LLMs as copilots for theorem proving in Lean 4, exposing native tactics (
suggest_tactics,search_proof,select_premises) that embed language model inference and premise retrieval directly inside the Lean proof environment, supporting local CTranslate2/CUDA inference as well as remote model APIs for interactive and automated proof search (Caltech & NVIDIA, NeurIPS 2024, 1.2K+ stars) - BioDiscoveryAgent - AI agent for biological discovery and research automation
- Biomni - General-purpose biomedical AI agent integrating LLM reasoning with retrieval-augmented planning and code-based execution to autonomously execute diverse biomedical research tasks and generate testable hypotheses (Stanford SNAP, bioRxiv 2025)
- STAgent - Multimodal LLM-based AI agent enabling deep research in spatial transcriptomics, automating analysis and interpretation of spatial gene expression data (Harvard LiuLab, bioRxiv 2025)
- Camyla - Fully autonomous medical image segmentation research system that generates complete manuscripts end-to-end from datasets with zero human intervention, beating strongest baselines on 24 of 31 datasets and achieving T1-T2 tier manuscript quality in double-blind evaluations (USTC & Shanghai AI Lab, 2026)
- MOOSE - Large Language Models for automated open-domain scientific hypotheses discovery (ACL 2024, ICML Best Poster)
- ChemCrow - LLM agents for chemistry research with tool integration
- Coscientist - Autonomous chemical experiment planning and execution
- TxAgent - AI agent for therapeutic reasoning across a universe of tools, achieving 92.1% accuracy in drug reasoning and outperforming GPT-4o by 25.8% (Harvard MIMS, 2025)
- Snorkel - Programmatic data labeling and weak supervision for scientific datasets
- PandasAI - Conversational data analysis and visualization using natural language
- torchdiffeq - PyTorch implementation of neural ODEs
- torchdyn - Neural differential equations in PyTorch
- diffrax - Numerical differential equation solving in JAX
- DifferentialEquations.jl - Julia differential equations suite
- DiffEqFlux.jl - Neural differential equations in Julia
- DeepXDE - Deep learning library for solving PDEs
- Lang-PINN - LLM-driven multi-agent system that builds trainable PINNs from natural language task descriptions, achieving 3-5 orders of magnitude MSE reduction and 50%+ execution success improvement (ICLR 2026)
- PINNs - Physics-informed neural networks
- NVIDIA PhysicsNeMo - Open-source framework for building physics-ML models at scale (renamed from Modulus, 2025)
- PINA - Physics-Informed Neural networks for Advanced modeling in PyTorch
- SciANN - Keras-based scientific neural networks
- NeuralPDE.jl - Physics-informed neural networks in Julia
- DeepONet - Learning nonlinear operators
- PySINDy - Sparse identification of nonlinear dynamics
- PySR - High-performance symbolic regression for discovering interpretable scientific equations from data, multi-population evolutionary search with Python/Julia backend, widely used in physics and astronomy (Cambridge, NeurIPS 2023)
- LLM-SR - Scientific equation discovery and symbolic regression using LLMs, combining code generation with evolutionary search (ICLR 2025 Oral)
- PSRN - Parallel symbolic regression network evaluating millions of expressions on GPU with automated subtree reuse, Nature Computational Science cover article (MIT, 2026)
- Fourier Neural Operator - Learning operators in Fourier space
- Poseidon - Efficient foundation models for PDEs with pretrained transformer-based neural operators and downstream task fine-tuning pipelines, HuggingFace integration for models and datasets (ETH Zurich CAMLab, arXiv 2024)
- PhiFlow - Differentiable PDE solving framework for machine learning with built-in fluid simulation, supporting PyTorch/JAX/TensorFlow backends and enabling neural network training within physical simulations (TUM, MIT License)
- Machine Learning for Scientometric Analysis (2021.09) - Comprehensive review
- AI for Science: Progress and Challenges (2023.03) - State of the field
- Foundation Models for Science (2022.05) - Large models in research
- Neural Ordinary Differential Equations (2018.06) - Breakthrough in neural ODEs
- Physics-Informed Neural Networks (2017.11) - Physics-constrained deep learning
- Scientific Discovery in the Age of Artificial Intelligence - Nature review on AI's role in science
- A Survey on AI-assisted Scientific Discovery (2025.02) - Comprehensive overview of LLMs in scientific research lifecycle from literature search to peer review
- AI4Research: A Survey of Artificial Intelligence for Scientific Research (2025.07) - Systematic taxonomy of AI in research
- Artificial Intelligence for Science in Quantum, Atomistic, and Continuum Systems (2023.07) - Unified technical survey across scientific scales with 63 contributors
- From Automation to Autonomy: A Survey on Large Language Models in Scientific Discovery (2025.05) - Three-level taxonomy (Tool, Analyst, Scientist)
- From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery (2025.08) - Comprehensive survey on agentic science across life sciences, chemistry, materials, and physics
- Agentic AI for Scientific Discovery: A Survey of Progress, Challenges, and Future Directions (2025.03) - Comprehensive review of AI agents in science
- Towards Scientific Intelligence: A Survey of LLM-based Scientific Agents (2025.03) - Scientific AI agent systems
- A Comprehensive Survey of Scientific Large Language Models and Their Applications (2024.06) - 260+ scientific LLMs across domains
- A Survey of Scientific Large Language Models: From Data Foundations to Agent Frontiers (2025.08) - Data-centric view of scientific LLMs
- Scientific Large Language Models: A Survey on Biological & Chemical Domains (2024.01) - Domain-specific scientific LLMs
- Scientific Machine Learning through Physics-Informed Neural Networks: Where we are and What's next (2022.01) - Comprehensive PINN review
- Physics-Informed Neural Networks and Extensions (2024.08) - Recent PINN advances and variants
- The frontier of simulation-based inference (PNAS 2020) - Foundational review on SBI for scientific computing by Cranmer et al.
- From Theory to Application: A Practical Introduction to Neural Operators in Scientific Computing (2025.03) - Implementation-focused guide to DeepONet, FNO, and PCANet
- Architectures, variants, and performance of neural operators: A comparative review (2025) - Systematic analysis of DeepONets, integral kernel operators, and transformer-based neural operators
- Foundation Models for Environmental Science: A Survey (2025.04) - Environmental applications
- Foundation Models in Bioinformatics - Biological foundation models
- Foundation Models for Materials Discovery (2025) - Perspective on materials AI
- Uncertainty quantification in scientific machine learning: Methods, metrics, and comparisons (J. Comput. Phys. 2023) - Comprehensive framework for UQ in PINNs and neural operators by Psaros et al.
- A Survey on Uncertainty Quantification Methods for Deep Learning (2023) - Systematic taxonomy of UQ methods from uncertainty source perspective
- Self-Driving Laboratories for Chemistry and Materials Science (Chem. Rev. 2024) - Comprehensive 100-page review on SDL technology, applications, and infrastructure
- Autonomous 'self-driving' laboratories: a review of technology and policy implications (Royal Soc. Open Sci. 2025) - Technology review with policy and safety considerations
- Artificial Intelligence for Science (CSIRO 2022) - Landmark report analyzing AI adoption across 98% of scientific fields over 60 years
- AI for Science 2025 (Fudan University & Nature 2025) - Comprehensive report on AI's transformative impact across 7 scientific fields, 28 research directions, and 90+ challenges
- AI in science evidence review (European Scientific Advice 2024) - Policy-focused evidence review on AI's impact in research
- The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery (2024.08) - First fully autonomous research system
- The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search (2025.04) - Enhanced autonomous research with agentic tree search
- AI-Researcher: Autonomous Scientific Innovation (2025.05) - Autonomous research pipeline from literature to publication with Scientist-Bench evaluation framework
- InternAgent: When Agent Becomes the Scientist -- Building Closed-Loop System from Hypothesis to Verification (2025.05) - Multi-agent system achieving #1 on MLE-Bench with closed-loop research automation
- Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems (2025.07) - Self-evolving multi-agent research systems
- ChemCrow: Augmenting large-language models with chemistry tools (2023.04) - LLM agents for chemistry research
- Autonomous chemical research with large language models - Automated chemical experimentation
- Coscientist: Autonomously planning and executing scientific experiments - Robotic lab automation
- AlphaFold: Protein Structure Prediction
- AI for Materials Discovery
- Large Language Models in Chemistry (2024.02)
- Cell2Sentence: Teaching Large Language Models the Language of Biology (ICML 2024) - LLMs for single-cell transcriptomics
- Scaling Large Language Models for Next-Generation Single-Cell Analysis (2025.04) - 27B parameter biological language models
- Boltz-1: Democratizing Biomolecular Interaction Modeling (bioRxiv 2024) - First fully open-source model achieving AlphaFold3-level accuracy
- MOOSE: Large Language Models for Automated Open-domain Scientific Hypotheses Discovery (ACL 2024) - First work showing LLMs can generate novel and valid scientific hypotheses, ICML Best Poster Award
- Earth-Agent: Unlocking the Full Landscape of Earth Observation with Agents (2025.09) - LLM agent framework for Earth Observation with 104 specialized tools and multi-modal analysis
- MedAgents: Large Language Models as Collaborators for Zero-shot Medical Reasoning (ACL 2024) - Multi-disciplinary collaboration framework for medical reasoning using role-playing LLM agents
- MedAgentGym: A Scalable Agentic Training Environment for Code-Centric Reasoning in Biomedical Data Science (2025.06) - Specialized training environment for biomedical AI agents with code-centric reasoning
- Paper2Web: Let's Make Your Paper Alive! (2025.10) - AI-powered transformation of academic papers into interactive websites with comprehensive evaluation framework
- DeepAnalyze: Agentic Large Language Models for Autonomous Data Science (2025.10) - First agentic LLM for autonomous data science with curriculum-based training
- Democratizing AI scientists using ToolUniverse (2025.09) - Universal ecosystem for building AI scientists from any LLM with 600+ scientific tools
- TxAgent: An AI Agent for Therapeutic Reasoning Across a Universe of Tools (2025.03) - AI agent achieving 92.1% accuracy in drug reasoning, outperforming GPT-4o by 25.8%
- Aviary: Training Language Agents on Challenging Scientific Tasks (2024.12) - Language agent training framework for scientific discovery
- Galactica: A Large Language Model for Science (2022.11)
- ScienceAgentBench (ICLR 2025) - 102 executable tasks from 44 peer-reviewed papers across 4 disciplines with containerized evaluation
- Scientist-Bench - Comprehensive benchmark for comparing LLM Agent-generated research outcomes with high-quality scientific work
- SciTrust: Evaluating the Trustworthiness of Large Language Models for Science (2024) - Scientific LLM trustworthiness evaluation framework
- SciBench: Evaluating College-Level Scientific Problem-Solving Abilities (2023) - Scientific reasoning benchmarks
- ChartCoder Evaluation - Chart-to-code generation benchmarks
- CryoDRGN - Neural network-based cryo-EM heterogeneous reconstruction, modeling continuous 3D structure distributions from single-particle images, with CryoDRGN-ET extending to in-cell cryo-electron tomography (MIT CSAIL, Nature Methods 2021/2024)
- AlphaFold - Protein structure prediction
- AlphaFold3 - AlphaFold 3 inference pipeline for unified biomolecular structure prediction of proteins, nucleic acids, small molecules, ions, and post-translational modifications (Google DeepMind, Nature 2024)
- ColabFold (2025 Updates) - AlphaFold/ESMFold accessible implementation with AF3 JSON export, database updates
- OpenFold3 - Fully open-source (Apache 2.0) biomolecular structure prediction reproducing AlphaFold3, free for academic and commercial use (Columbia AlQuraishi Lab & OpenFold Consortium, 2025)
- Protenix - Trainable PyTorch reproduction of AlphaFold 3
- HelixFold3 - Baidu's open-source reproduction of AlphaFold3 in PaddlePaddle, providing pretrained weights and inference pipelines for unified biomolecular structure prediction across proteins, nucleic acids, ligands, ions, and post-translational modifications within the PaddleHelix biocomputing platform (Baidu, bioRxiv 2024)
- RoseTTAFold-All-Atom - All-atom biomolecular structure prediction for protein-nucleic acid-small molecule-metal ion complexes, enabling accurate modeling of covalent modifications and assemblies beyond proteins (Baker Lab, Science 2024)
- Chai-1 - Multi-modal foundation model for biomolecular structure prediction (proteins, small molecules, DNA, RNA, glycans) achieving SOTA across benchmarks, with optional MSA/template support (Chai Discovery, 2024)
- NeuralPLexer - State-specific protein-ligand complex structure prediction with a multi-scale deep generative model, enabling conformational state-aware modeling of molecular interactions (329+ stars, 2024)
- Boltz - First fully open-source model achieving AlphaFold3-level accuracy with 1000x faster binding affinity prediction (MIT)
- BoltzGen - De novo protein binder design via generative model, achieving nanomolar binding for 66% of novel targets tested (MIT, 2025)
- Proteina-Complexa - Flow-based generative model for atomistic protein binder design with test-time optimization, SOTA on binder benchmarks (ICLR 2026 Oral, NVIDIA)
- xfold - Democratizing AlphaFold3: PyTorch reimplementation to accelerate protein structure prediction research
- MegaFold - Cross-platform system optimizations for accelerating AlphaFold3 training with 1.73x speedup and 1.23x memory reduction
- Graphormer - General-purpose deep learning backbone for molecular modeling
- DiffDock - Diffusion-based molecular docking achieving SOTA blind docking performance, treating ligand pose prediction as generative diffusion over SE(3), with DiffDock-L update for improved generalization (MIT CSAIL, ICLR 2023)
- PLACER - Graph neural network operating entirely at the atomic level for protein-ligand conformational ensemble prediction and docking, generating diverse solutions through rapid stochastic denoising to model conformational heterogeneity (Baker Lab, bioRxiv 2025)
- targetdiff - 3D Equivariant Diffusion for Target-Aware Molecule Generation (ICLR2023)
- ReQFlow - Rectified Quaternion Flow for efficient protein backbone generation, 37× faster than RFDiffusion with 0.972 designability (ICML 2025)
- BioEmu - Microsoft's generative model for sampling protein equilibrium conformations 100,000× faster than MD simulations, predicting domain motions, local unfolding and cryptic binding pockets on a single GPU (Science 2025)
- ProteinMPNN - Deep learning-based protein sequence design (inverse folding) from backbone structures, achieving 52.4% sequence recovery vs 32.9% for Rosetta, core tool in modern protein design pipelines (Baker Lab, Science 2022)
- LigandMPNN - Extension of ProteinMPNN for protein sequence design in the context of small-molecule ligands, metal ions, and nucleic acids, enabling binding site engineering and co-factor redesign (Baker Lab)
- ColabDesign - Accessible protein design platform via Google Colab integrating AlphaFold2, RoseTTAFold, and ProteinMPNN for de novo hallucination, fixed backbone design, and binder design (Sergey Ovchinnikov, 2022+)
- BindCraft - Simple and accurate de novo protein binder design pipeline using AlphaFold2 backpropagation, MPNN, and PyRosetta for automated binder discovery (bioRxiv 2024)
- Genie 2 - Diffusion model for scalable protein structure design with multi-motif scaffolding capabilities, achieving state-of-the-art designability, diversity, and novelty through SE(3)-equivariant attention and massive data augmentation (AlQuraishi Lab, 2024)
- Chroma - Generative model for programmable protein design using diffusion modeling, equivariant graph neural networks, and conditional random fields to efficiently sample diverse all-atom structures; supports conditional generation via composable conditioners for substructure, symmetry, shape, and neural-network predictions; validated crystallographically (Generate Biomedicines, Nature 2023)
- RFdiffusion3 - Latest RFdiffusion for protein structure design with 10× speedup and atom-level precision (December 2025)
- RFantibody - Structure-based de novo antibody design pipeline built on RFdiffusion for computational generation of target-specific antibodies (RosettaCommons, 2025)
- IgGM - Generative foundation model for functional antibody and nanobody design, supporting de novo generation, affinity maturation, inverse design, structure prediction, and humanization (Tencent AI4S, ICLR 2025)
- DrugAssist - LLM-based molecular optimization tool
- GenMol - ICML 2025 drug discovery generalist using masked discrete diffusion and fragment-based generation with molecular context guidance (NVIDIA)
- REINVENT - Industrial-grade reinforcement-learning-based generative platform for de novo molecular design with transformer architectures, supporting multi-objective optimization, scaffold decoration, and curriculum learning (AstraZeneca MolecularAI, REINVENT 4, 2024)
- mint - Learning the language of protein-protein interactions
- Mol-Instructions - Large-scale biomolecular instruction dataset for chemistry/biology LLMs (ICLR2024)
- Uni-Mol - Universal 3D molecular pretraining framework with 209M conformations, scaling to 1.1B parameters (Uni-Mol2) on 800M conformations for molecular property prediction, docking, and quantum chemistry (ICLR 2023, NeurIPS 2024)
- ChemBERTa - Chemical language model
- DeepChem - Machine learning for chemistry
- DeepMol - Unified ML/DL framework for drug discovery workflows, integrating RDKit, DeepChem, and scikit-learn with SHAP explainability
- Chemprop - Message passing neural networks for molecule property prediction, ADMET modeling, and reaction prediction, achieving SOTA on MoleculeNet and widely used in pharmaceutical drug discovery (MIT, 2.3K+ stars)
- RDKit - Cheminformatics toolkit
- ESM3 - 98B-parameter frontier generative model jointly reasoning over protein sequence, structure, and function, trained on 2.78 billion proteins; generated a novel fluorescent protein (esmGFP) with only 58% sequence identity to known GFPs (EvolutionaryScale, 2024)
- ESMFold - Protein structure prediction from ESM models
- SaProt - Structure-aware protein language model using 3D structural vocabulary (Foldseek) for joint sequence-structure pretraining, achieving SOTA on protein engineering and fitness prediction benchmarks (ICML 2024, Westlake University & Repl)
- Foldseek - Fast and accurate protein structure search using a learned 3Di structural alphabet (VQ-VAE) that discretizes tertiary interactions into structural tokens, enabling protein-universe-scale structural alignment at sequence-search speeds (4-5 orders of magnitude faster than DALI/TM-align) and underpinning many AI4S tools such as SaProt, ESMAtlas search, and AFDB clustering pipelines (Steinegger Lab, Nature Biotechnology 2023)
- RhoFold+ - End-to-end RNA 3D structure prediction using RNA language model pretrained on 23.7M sequences, outperforming existing methods and human expert groups on RNA-Puzzles and CASP15 (Nature Methods 2024)
- AIDO.ModelGenerator - GenBio AI's software stack for the AI-Driven Digital Organism, supporting adaptation and finetuning of multiscale biological foundation models across DNA, RNA, protein, structure, and single-cell tasks with reproducible CLIs and pretrained model zoo (2025)
- Evo 2 - Arc Institute's 40B-parameter genome foundation model trained on 9 trillion nucleotides from all domains of life, supporting 1M base pair context for generalist DNA/RNA/protein prediction and design (Nature 2026)
- Nucleotide Transformer - Foundation models for genomics and transcriptomics pretrained on 3,000+ human genomes and 850+ diverse species, enabling chromatin accessibility prediction, splice site detection, and promoter classification across multiple model scales (InstaDeep, NVIDIA & TUM, Nature Methods 2023)
- LucaOne - Generalized biological foundation model with unified nucleic acid and protein language, integrating DNA/RNA/protein sequences (Nature Machine Intelligence 2025)
- Geneformer - Single-cell transformer foundation model pretrained on 104M human transcriptomes via masked gene prediction, enabling transfer learning for cell type classification, gene network analysis, and in silico perturbation with limited labeled data (Nature 2023, V2 2024)
- scFoundation - 100M-parameter foundation model pretrained on 50M+ human single-cell transcriptomes covering ~20,000 genes, achieving SOTA on gene expression enhancement, drug response and perturbation prediction (Nature Methods 2024)
- Tahoe-x1 - Apache 2.0 single-cell foundation model family scaling to 3B parameters, pretrained on 266M cell profiles including perturbation data and released with training, embedding, and downstream benchmarking workflows for disease-relevant single-cell tasks (2025)
- Stack - Arc Institute's single-cell foundation model enabling in-context learning at inference time via a novel tabular attention architecture, trained on 150M uniformly-preprocessed cells for generalizing biological effects and generating unseen cell profiles in novel contexts (2025)
- scvi-tools - Deep probabilistic framework for single-cell and spatial omics analysis, integrating scVI, scANVI, totalVI and other VAE-based models for batch correction, cell annotation, multi-omics integration, and RNA velocity (scverse/NumFOCUS, Nature Methods 2018/2024)
- GEARS - Geometric deep learning model predicting transcriptional outcomes of novel single- and multi-gene perturbations using gene–gene knowledge graphs, 40% higher precision than prior methods on combinatorial perturbation prediction (Stanford, Nature Biotechnology 2024)
- scGPT - Single-cell analysis with transformers
- CellTypist - Automated cell type annotation tool for single-cell transcriptomics using gradient boosting and logistic regression with reference atlases, enabling standardized classification across datasets (Wellcome Sanger Institute, Nature Biotechnology 2022)
- Cell2Sentence - Teaching Large Language Models the Language of Biology through single-cell transcriptomics (ICML 2024)
- OmicVerse - Unified Python framework for bulk, single-cell, and spatial RNA-seq multi-omics analysis with deep learning deconvolution (VAE) and graph neural networks, bridging Bindea, Bindea, scanpy and squidpy ecosystems (Nature Communications 2024)
- ChatSpatial - MCP server enabling spatial transcriptomics analysis via natural language, integrating 60+ methods including SpaGCN, Cell2location, LIANA+, CellRank for Visium, Xenium, MERFISH platforms
- Enformer - Gene expression prediction
- DNABERT - DNA sequence analysis
- scBERT - Single-cell BERT for gene expression
- GenePT - Generative pre-training for genomics
- DNA Claude Analysis - Interactive personal genome analysis toolkit using Claude Code and Python. Parses raw genotyping data from consumer DNA services and analyzes SNPs across 17 categories including health risks, pharmacogenomics, ancestry, and nutrition, with a terminal-style HTML dashboard.
- OpenCRISPR - First open-source AI-generated gene editing systems developed with protein language models, enabling programmable CRISPR-Cas nucleases for synthetic biology and therapeutic genome editing (Profluent, 2024)
- AlphaMissense - Google DeepMind's AlphaFold-derived classifier for proteome-wide missense variant effect prediction, providing pathogenicity scores for all ~71M possible human missense variants and classifying 89% with 90% precision; pre-computed predictions are integrated into Ensembl VEP and UCSC Genome Browser to support clinical variant interpretation (Science 2023)
- AlphaGenome - Google DeepMind's unified DNA sequence foundation model predicting molecular consequences of genetic variants from single-base resolution up to 1 megabase context, jointly outputting thousands of regulatory tracks (RNA expression, splicing, chromatin accessibility, TF binding, contact maps) for human and mouse genomes via a Python client and non-commercial API (2025)
- DeepLabCut - Markerless pose estimation of user-defined features with deep learning for all animals including humans, enabling quantitative behavioral analysis in neuroscience and ethology (Nature Neuroscience 2018, 5.6K+ stars)
- SLEAP - Deep learning-based multi-animal pose tracking and behavior classification, enabling automated quantification of social interactions and collective behavior across species (Nature Methods 2022, 2.2K+ stars)
- CEBRA (Nature 2023) - Learnable latent embeddings for joint behavioral and neural analysis, enabling consistent and interpretable mapping of neural activity to behavior across modalities, species, and experiments (EPFL & Harvard, 1K+ stars)
- TRIBE v2 - Meta FAIR's foundation model of vision, audition, and language for in-silico neuroscience, predicting fMRI brain responses to naturalistic multimodal stimuli (video, audio, text) through unified Transformer architecture mapped to the cortical surface (2026)
- UNI (Nature Medicine 2024) - General-purpose pathology foundation model pretrained on 100K+ diagnostic whole-slide images across 20 major tissue types, achieving state-of-the-art transfer learning across 30+ clinical tasks and serving as a universal feature extractor for digital pathology (Mahmood Lab, 722+ stars)
- Prov-GigaPath (Nature 2024) - Whole-slide pathology foundation model trained on 1.3 billion image tiles from 171K slides using a LongNet-based architecture to encode gigapixel-scale WSIs for cancer subtyping and biomarker prediction (Microsoft Research & Providence, 601+ stars)
- CONCH (Nature Medicine 2024) - Vision-language pathology foundation model using contrastive learning on histopathology image-text pairs, enabling zero-shot classification, slide-level retrieval, and multimodal reasoning across diverse cancer types (Mahmood Lab, 494+ stars)
- PLIP (Nature Medicine 2023) - First vision-and-language foundation model for pathology AI, fine-tuned from CLIP on 249K image-caption pairs, enabling open-ended visual-semantic search and zero-shot diagnosis across histopathology (Pathology Foundation, 376+ stars)
- TITAN (Nature Medicine 2024) - Multimodal whole-slide pathology foundation model jointly pretrained on H&E histology and diagnostic text reports, enabling zero-shot cancer subtyping, biomarker prediction, and multimodal reasoning across diverse cancer types (Mahmood Lab, 341+ stars)
- Virchow (Nature Medicine 2024) - Self-supervised pathology foundation model (ViT-Huge, 632M parameters) pretrained via DINOv2 on 1.5M whole-slide images from Memorial Sloan Kettering across 17 cancer types, with Virchow2 follow-up scaling to 3.1M slides and mixed magnifications, achieving SOTA on biomarker prediction, mutation classification, and rare cancer detection (Paige AI & MSK)
- Cellpose - Generalist deep learning algorithm for cell and nucleus segmentation across diverse image types, with human-in-the-loop training (2.0) and one-click image restoration (3.0), 70K+ training objects (Nature Methods 2021/2022/2025)
- micro-sam - Segment Anything Model for microscopy: interactive and automatic segmentation of light, electron, and fluorescence microscopy images in 2D and 3D, with domain-specific fine-tuning workflows for scientific imaging (1.5K+ stars)
- MedSAM - Universal medical image segmentation foundation model trained on 1.57M image-mask pairs across 10 imaging modalities and 30+ cancer types (Nature Communications 2024)
- MedSAM2 - Segment Anything in 3D medical images and videos, extending SAM2 to volumetric and temporal medical imaging with state-of-the-art zero-shot segmentation performance across CT, MRI, and surgical video (arXiv 2025)
- MedSegX - Generalist foundation model and database for open-world medical image segmentation, enabling universal segmentation of diverse anatomical structures and pathologies with zero-shot generalization to unseen tasks and modalities (Nature Biomedical Engineering 2025)
- BiomedParse - Foundation model for joint segmentation, detection, and recognition of biomedical objects across nine imaging modalities, with v2 introducing BoltzFormer architecture for end-to-end 3D inference (Microsoft, Nature Methods 2025)
- MedAgents - Multi-disciplinary collaboration framework for zero-shot medical reasoning using role-playing LLM agents (ACL 2024)
- MedAgentGym - Scalable agentic training environment for code-centric reasoning in biomedical data science
- MedRAG - Systematic medical RAG toolkit for question answering over PubMed, StatPearls, textbooks, and Wikipedia, supporting multiple retrievers, domain LLMs, and follow-up-query workflows for benchmarked clinical/biomedical QA (ACL Findings 2024)
- NVIDIA Biomedical AI-Q Research Agent - Deployable biomedical deep-research agent blueprint combining on-prem multimodal RAG, report generation, human-in-the-loop editing, and virtual screening with MolMIM and DiffDock for drug discovery workflows (2025)
- nnU-Net - Self-configuring deep learning framework for semantic segmentation of biomedical images requiring no manual hyperparameter tuning; automatically adapts preprocessing, network topology, and training parameters to achieve state-of-the-art results across 120+ international competitions and benchmarks out-of-the-box (DKFZ, Nature Methods 2021, 8.3k+ stars)
- TotalSegmentator - Robust deep learning-based segmentation of >100 anatomical structures in CT and MR images, built on nnU-Net and widely adopted in clinical radiology and surgical planning workflows (2.6K+ stars)
- MONAI - NVIDIA and King's College London's open-source AI toolkit for healthcare imaging, providing foundational frameworks for medical image annotation (MONAI Label), training (MONAI Core), and deployment (MONAI Deploy) across radiology, pathology, and endoscopy (8K+ stars, Apache 2.0)
- LLM4Chemistry - Curated paper list about LLMs for chemistry covering fine-tuning, reasoning, multi-modal models, agents, and benchmarks (COLING 2025)
- ChemMCP - Extensible chemistry toolkit for MCP-enabled AI assistants, exposing molecule analysis, property prediction, and reaction synthesis tools through unified Python/MCP interfaces for chemistry agents and research workflows (Apache 2.0, 2025)
- GNoME - DeepMind's graph neural network for materials exploration, discovering 2.2M new crystal structures (380K most stable) equivalent to 800 years of traditional research, with 520K+ materials dataset open-sourced (Nature 2023)
- FAIRChem (OMat24) - Meta's comprehensive ML ecosystem for materials/chemistry with 118M+ DFT calculations, EquiformerV2 models achieving top Matbench Discovery performance
- NequIP - E(3)-equivariant neural network interatomic potentials achieving DFT accuracy with up to 1000× less training data than invariant models, foundational architecture behind MACE and Allegro (Harvard, MIT, Nature Communications 2022)
- SchNetPack - PyTorch toolkit for deep neural networks in atomistic simulations, implementing SchNet, DimeNet++, PaiNN, and GemNet for molecular dynamics and quantum chemistry (900+ stars)
- pymatgen - Python Materials Genomics: robust materials analysis library defining classes for structures and molecules with support for many electronic structure codes; foundational toolkit powering the Materials Project (Berkeley Lab, 1.8K+ stars)
- MACE - Machine learning interatomic potentials
- CHGNet - Universal pretrained neural network potential with charge and magnetic moment awareness, trained on 1.5M+ Materials Project inorganic structures for charge-informed molecular dynamics and phase diagram prediction (Berkeley, Nature Machine Intelligence 2023 Cover)
- MatterGen - Diffusion-based generative model for inorganic materials design, steering generation by chemistry, symmetry, bulk modulus, band gap, or magnetic properties, 2× more likely to produce stable novel structures than prior methods, experimentally validated with synthesized TaCr₂O₆ (Microsoft, Nature 2025)
- MatterSim - Deep learning atomistic model across elements, temperatures, and pressures
- ORB - Universal machine learning interatomic potential for atomistic simulation of materials, molecules, and biomolecules across the periodic table, with open-source pretrained models and inference tools (Orbital Materials, 2024-2025)
- Crystal Graph CNNs - Crystal property prediction
- MatBench - Materials informatics benchmark
- Best of Atomistic Machine Learning - Curated list of atomistic ML projects for materials science
- AiZynthFinder - AstraZeneca's industrial-grade retrosynthetic planning tool using MCTS to recursively decompose molecules into purchasable precursors, with multi-step route scoring and support for custom one-step models (v4.0, 2024)
- Molecular Transformers - AI for chemical reaction prediction and synthesis planning
- FermiNet - DeepMind's neural network for ab-initio quantum chemistry, directly solving the many-electron Schrödinger equation via variational Monte Carlo with antisymmetric wavefunctions, extended to excited states (Phys. Rev. Research 2020, Science 2024)
- JAX-MD - Molecular dynamics in JAX
- Neural ODEs - Differential equations with neural networks
- Physics-Informed Neural Networks - Physics-constrained ML
- EquiformerV2 - Improved equivariant Transformer for 3D atomic graphs (ICLR2024)
- Equiformer - Equivariant graph attention Transformer (ICLR2023)
- TORAX - Differentiable tokamak core transport simulator for fusion energy research, coupling PDE solvers with JAX auto-differentiation and neural-network surrogates for fast forward modelling, pulse-design, and trajectory optimization (Google DeepMind, Apache 2.0)
- AstroCLIP - Cross-modal self-supervised foundation model for galaxies by Polymathic AI, jointly embedding multi-band galaxy imaging and optical spectra into a shared latent space to enable zero/few-shot redshift estimation, galaxy property prediction, morphology classification, and cross-modal similarity search (MNRAS Letters 2024)
- AION (arXiv 2025) - Polymathic AI's large omnimodal foundation model for astronomical surveys, seamlessly integrating 39 distinct data modalities including imaging, spectra, photometry, and catalog entries for similarity search, property prediction, and generative modeling across legacy surveys (MIT)
- AstroPy - Python astronomy tools
- Gaia Archive - Stellar data for ML
- DeepSphere - Spherical CNNs for astronomy
- GenCast - Google DeepMind's diffusion-based ensemble weather forecasting model at 0.25° resolution, outperforming ECMWF ENS on 97.2% of targets up to 15 days ahead, with open-source code and weights (Nature 2024)
- Aurora - Microsoft's foundation model for the Earth system supporting weather, air pollution, and ocean wave forecasting at multiple resolutions, trained on 1M+ hours of diverse atmospheric data (Nature 2025)
- Earth-Copilot - Microsoft's AI-powered geospatial Earth science application for natural-language exploration, visualization, and analysis of 130+ satellite collections, with STAC integration, multi-agent backend, MCP server, and deployable React/FastAPI stack (MIT, 2025)
- ClimaX - First foundation model for weather and climate by Microsoft, Vision Transformer-based architecture trained on heterogeneous datasets (ICML 2023)
- NeuralGCM - Google Research's hybrid ML/physics atmospheric model combining learned dynamics with physical constraints, outperforming traditional models on 2-15 day forecasts and 40-year climate simulation, developed with ECMWF (Nature 2024)
- NVIDIA Earth-2 - World's first fully open, accelerated weather AI software stack with Medium Range forecasting and Nowcasting models using generative AI (January 2026)
- Pangu-Weather - Huawei's 3D high-resolution global weather forecast model at 0.25° resolution, first AI method to comprehensively outperform traditional NWP across all variables and lead times, integrated into ECMWF operational forecasts (Nature 2023)
- Prithvi WxC - IBM-NASA open-source 2.3B parameter weather and climate foundation model trained on 160 MERRA-2 variables, runs on desktop with fine-tuned variants for climate downscaling and gravity wave parameterization
- ClimateBench - Climate data benchmark for ML models
- WeatherBench - Weather prediction benchmark
- WeatherBench2 - Next-generation benchmark for data-driven global weather models with standardized evaluation framework and curated datasets for ML forecasting (Google Research, 2024)
- WeatherGFT - Physics-AI hybrid modeling for fine-grained weather forecasting (NeurIPS'24)
- GeoAI - High-level open-source geospatial AI package for satellite/aerial imagery analysis, model training, inference, interactive visualization, and QGIS integration, bridging PyTorch/Transformers with remote sensing workflows (MIT, 2026)
- segment-geospatial - Python package for segmenting geospatial data with the Segment Anything Model (SAM), enabling zero-shot object segmentation in satellite and aerial imagery for remote sensing and Earth observation (MIT, 4k+ stars)
- Awesome Large Weather Models - Curated list of large weather models for AI Earth science
- TerraTorch - Python toolkit for fine-tuning geospatial foundation models
- Earth-Agent - LLM agent framework for Earth Observation with 104 specialized tools across 5 functional kits
- AI for Earth - Microsoft's environmental AI
- TorchGeo - PyTorch domain library for geospatial deep learning providing standardized datasets, samplers, transforms, and pre-trained models for remote sensing, land cover mapping, and environmental monitoring (Microsoft, 4K+ stars)
- Clay Foundation Model - Open-source self-supervised vision foundation model for Earth observation by Clay Foundation (non-profit), a Masked Autoencoder ViT pretrained on multimodal satellite imagery (Sentinel-1/2, Landsat 8-9, NAIP, MODIS, LINZ DEM) with location/time embeddings, supporting classification, segmentation, change detection, similarity search, and few-shot downstream geospatial tasks (Apache 2.0, v1.5 2024-2025)
- Satlas - Allen Institute for AI's global geospatial foundation model for satellite imagery analysis, enabling large-scale mapping of buildings, wind turbines, trees, and land cover from Sentinel-2 data with open-source weights and inference tools (2024)
- SkySensePlusPlus - Semantic-enhanced multi-modal remote sensing foundation model for Earth observation (Nature Machine Intelligence 2025), enabling universal interpretation across diverse satellite imagery modalities with open-source weights and benchmarks
- TESSERA (CVPR 2026) - University of Cambridge's foundation model for time-series satellite imagery, enabling efficient extraction of temporal patterns from Earth observation for land classification, canopy height prediction, and other remote sensing tasks
- PlantNet - Plant identification using AI and citizen science
- AgML - Agricultural machine learning platform
- FarmVibes.AI - Multi-modal geospatial ML platform for agriculture and sustainability, fusing satellite imagery (RGB, SAR, multispectral), drone imagery, weather data, and sensor data for crop identification, carbon footprint estimation, and microclimate prediction (Microsoft Research, MIT License)
- BioSimulators - Biological simulation tools
- EcoNet - Ecological modeling and conservation AI
- Galactica - Large language model for science
- Intern-S1 - Open-source scientific multimodal foundation model built on a 235B MoE LLM and 6B vision encoder, continually pretrained on 5T tokens including 2.5T scientific-domain tokens, with strong results across chemistry, materials, life science, and earth science benchmarks (2025)
- Llemma - Open language model for mathematics (7B/34B) trained on Proof-Pile-2, outperforming Minerva at equal scale on MATH benchmark, with tool use and formal theorem proving in Lean without finetuning (EleutherAI, ICLR 2024)
- MinervaAI - Mathematical reasoning
- PaLM-2 - Scientific reasoning capabilities
- ESM - Protein language models
- BioNeMo Framework - NVIDIA's open-source platform for building and adapting biological AI models at scale, bundling ESM-2, Geneformer, MolMIM and DNA embedding models with recipes for single-GPU to multi-node training (2025)
- IBM FM4M - IBM's open foundation model family for materials and chemistry, covering SMILES, SELFIES, molecular graphs, 3D atom positions, and electron density grids, with a unified toolkit for representation learning and downstream prediction/generation (Apache 2.0, 2024-2025)
- ChemGPT - Chemistry-focused language model
- BioGPT - Biomedical text generation
- Hugging Face Datasets - Comprehensive ML research datasets and scientific data collections
- Google Dataset Search - Find scientific datasets
- TDC - Therapeutics Data Commons: 66 AI-ready datasets across 22 drug discovery tasks with 29 leaderboards, covering target identification, molecular generation, ADMET prediction, and clinical trial outcomes (Harvard MIMS, NeurIPS 2021/2024)
- Protein Data Bank - Protein structures
- ChEMBL - Chemical bioactivity data
- Human Protein Atlas - Protein expression data
- Chinese Medical Dataset - Comprehensive collection of Chinese medical datasets for AI research
- Materials Project - Computational materials database
- QM9 - Small molecule properties
- Open Catalyst Project - Catalyst discovery
- The Well - 15TB collection of 16 large-scale numerical simulation datasets spanning fluid dynamics, MHD, astrophysics, biological systems, and acoustic scattering, with unified PyTorch dataloaders and benchmarks for training foundation models on physical sciences (Polymathic AI, NeurIPS 2024)
- LIGO Open Science Center - Gravitational wave data
- Particle Data Group - Particle physics data
- OpenQuantumMaterials - Quantum materials data
- PyTorch - Deep learning framework
- JAX - High-performance ML research
- TensorFlow - End-to-end ML platform
- NumPy - Numerical computing
- SciPy - Scientific computing
- Scikit-learn - Machine learning library
- SciML - Scientific machine learning ecosystem
- DifferentialEquations.jl - Multi-language suite for high-performance differential equation solving and scientific machine learning (3.0k+ stars)
- ModelingToolkit.jl - Acausal modeling framework for automatically parallelized scientific machine learning (1.5k+ stars)
- SciMLBenchmarks.jl - Scientific machine learning benchmarks & differential equation solvers
- NeuralPDE.jl - Physics-informed neural networks (PINNs) for solving partial differential equations (1.1k+ stars)
- DiffEqFlux.jl - Neural ordinary differential equations with O(1) backprop and GPU support (900+ stars)
- Optimization.jl - Unified interface for local, global, gradient-based and derivative-free optimization (800+ stars)
- PaddleScience - SDK & library for AI-driven scientific computing applications
- Flux.jl - Machine learning in Julia
- MDAnalysis - Molecular dynamics analysis
- e3nn - Euclidean neural networks for arbitrary point transformations enabling E(3)-equivariant deep learning, foundational library for building geometry-aware neural networks in molecular dynamics, materials science, and physics
- MDtrajNet - Neural network foundation model that directly generates MD trajectories bypassing force calculations, accelerating simulations by up to 100× with equivariant Transformer architecture (2025)
- ASE - Atomic Simulation Environment for materials modeling
- PyMC - Probabilistic programming
- AI2BMD - Microsoft's AI-powered ab initio biomolecular dynamics simulation achieving quantum-mechanical accuracy for proteins with 10,000+ atoms, orders of magnitude faster than DFT using protein fragmentation and ML force fields (Nature 2024)
- OpenMM - High-performance molecular simulation toolkit
- DeePMD-kit - Deep learning package for many-body potential energy representation and molecular dynamics, achieving quantum-mechanical accuracy with classical MD efficiency (DeepModeling, Gordon Bell Prize 2020, 1.9k+ stars)
- Newton - GPU-accelerated differentiable physics simulation engine built on NVIDIA Warp, supporting rigid/soft body, cloth, and gradient-based optimization for scientific ML, initiated by Disney Research, DeepMind, and NVIDIA (Linux Foundation, Apache 2.0, 2025)
- PennyLane - Cross-platform library for differentiable programming of quantum computers with automatic differentiation, enabling hybrid quantum-classical machine learning for quantum chemistry, quantum physics, and NISQ algorithm research (Xanadu, 3k+ stars)
- AI for Everyone (Coursera) - Basic AI concepts
- CS229 Machine Learning - Stanford ML course
- MIT 6.034 Artificial Intelligence - AI fundamentals
- SciML Book - Parallel Computing and Scientific Machine Learning: MIT 18.337J/6.338J course materials (1.9k+ stars)
- Dive into Deep Learning - Interactive deep learning book with code implementations
- The Elements of Statistical Learning - Classic ML textbook freely available
- Neural Networks and Deep Learning - Free online book by Michael Nielsen
- Awesome Scientific Language Models - Curated scientific LLM papers (260+ models)
- Awesome LLM Scientific Discovery - LLM papers for scientific discovery
- AI4Research Papers - LLM for scientific research papers
- Physics-Informed Neural Networks Papers - PINN research collection
- Scientific Computing with ML Papers - Scientific ML paper repository
- Simulation-Based Inference Papers & Tools - Community-maintained SBI research portal with papers and software
- Awesome AI Scientist Papers - Autonomous AI scientist research
- Awesome Agents for Science - LLM agents across scientific domains
- Two Minute Papers - AI research summaries
- 3Blue1Brown - Mathematical concepts
- AI Coffee Break - AI paper reviews
- Steve Brunton - Data-driven methods
- Nathan Kutz - Applied mathematics
- Physics Informed Machine Learning - SciML tutorials
- NeurIPS - Machine learning conference
- ICML - International Conference on Machine Learning
- AI for Science Workshop - Specialized workshops
- Partnership on AI - AI research collaboration
- Allen Institute for AI - AI research institute
- OpenAI - AI research and deployment
- r/MachineLearning - ML discussions
- AI Alignment Forum - AI safety research
- Distill - Visual explanations of ML
This project builds upon and complements several excellent resources:
- awesome-ai4s - 200+ AI for Science papers with Chinese interpretations
- Awesome AI Scientist Papers - Autonomous AI scientist research
- Awesome Scientific Machine Learning - Physics-informed ML and SciML
- Awesome Agents for Science - LLM agents across scientific domains
- Awesome LLM Agents Scientific Discovery - Biomedical AI agents
- Awesome Foundation Models for Weather and Climate - Comprehensive survey of foundation models for weather and climate data understanding
- Scientific LLM Papers - 260+ scientific language models
- LLM4SR Repository - LLM for scientific research survey materials
- PINNs Paper Collection - Physics-informed neural networks research
- SciML Papers - Scientific computing and machine learning papers
- Current Focus: Shift from tool-level assistance to autonomous scientific agents
- Emerging Trends: Multi-modal scientific models, self-improving research systems
- Research Gaps: Evaluation frameworks, ethical governance, human-AI collaboration
- Future Directions: Fully autonomous discovery cycles, robotic lab integration
We welcome contributions! Please see our Contributing Guidelines for details.
- Fork this repository
- Add your resource in the appropriate section
- Ensure the format matches existing entries
- Submit a pull request with a clear description
- Ensure the resource is actively maintained
- Include a brief, clear description
- Check for duplicates before adding
- Use proper markdown formatting
This project is licensed under the MIT License - see the LICENSE file for details.
Special thanks to all researchers and developers pushing the boundaries of AI for Science. This list is inspired by the awesome community and the transformative potential of AI in scientific discovery.
Star ⭐ this repository if you find it helpful!
Last updated: May 2026 - Enhanced with 2025-2026 breakthroughs in autonomous research, equation discovery, scientific foundation models, spatial transcriptomics agents, biomedical image analysis foundation models, and deep potential molecular dynamics; added FunSearch to Research Agents & Autonomous Workflows; added pymatgen to Chemistry & Materials; added TESSERA to Earth & Climate Science; added nnU-Net to Biology & Medicine; added OpenCRISPR to Biology & Medicine; added Camyla to Research Agents; added TRIBE v2 to Biology & Medicine; added SciWrite to Research Workbench & Plugins; added Chroma to Biology & Medicine; added AlphaMissense to Genomics & Bioinformatics; added Clay Foundation Model to Remote Sensing & Geospatial AI; added MONAI to Medical AI & Clinical Applications; added TotalSegmentator to Medical AI & Clinical Applications; added LeanDojo to Domain-Specific Research Agents; added Biomni to Domain-Specific Research Agents; added TxAgent to Domain-Specific Research Agents; added ScienceBoard to Research Agents Evaluation & Benchmarking; added RFantibody to Biology & Medicine; added NewtonBench to Research Agents Evaluation & Benchmarking; added HelixFold3 to Protein & Drug Discovery; added Genie 2 to Protein & Drug Discovery; added MLE-Bench to Research Agents Evaluation & Benchmarking; added Stack to Genomics & Bioinformatics; added UNI, Prov-GigaPath, CONCH, and PLIP to Computational Pathology & Digital Pathology; added Virchow to Computational Pathology & Digital Pathology; added MedSAM2 to Medical AI & Clinical Applications; added PLACER to Protein & Drug Discovery; added CEBRA to Neuroscience & Behavioral Analysis; added TorchGeo to Remote Sensing & Geospatial AI; added AION to Astronomy & Astrophysics; added SchNetPack to Chemistry & Materials; added micro-sam to Medical AI & Clinical Applications; added LigandMPNN to Protein & Drug Discovery; added AlphaGenome to Genomics & Bioinformatics; added REINVENT to Protein & Drug Discovery; added Poseidon to Scientific Machine Learning; added TITAN to Computational Pathology & Digital Pathology; added PaperBanana to Paper→Poster / Slides / Graphical Abstract; added NeuralPLexer to Protein & Drug Discovery; added Lean Copilot to Domain-Specific Research Agents; added Foldseek to Protein & Drug Discovery; added OpenEvolve to Research Agents & Autonomous Workflows