-
CVS Health
- United States
- in/dylan-bouchard-phd-52594664
Stars
Virtual whiteboard for sketching hand-drawn like diagrams
A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation"
Codebase for reproducing the experiments of the semantic uncertainty paper (short-phrase and sentence-length experiments).
🤗 The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation tools
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
LettuceDetect is a hallucination detection framework for RAG applications.
Supercharge Your LLM Application Evaluations 🚀
Deepchecks: Tests for Continuous Validation of ML Models & Data. Deepchecks is a holistic open-source solution for all of your AI & ML validation needs, enabling to thoroughly test your data and mo…
Python packaging and dependency management made easy
Complete AI governance platform with support for ISO 42001, ISO 27001, NIST AI RMF and EU AI Act. Join our Discord channel: https://discord.com/invite/d3k3E4uEpR
Adversarial Natural Language Inference Benchmark
RAG evaluation without the need for "golden answers"
LLM-powered Conversational AI experience using Vectara
scikit-learn: machine learning in Python
Uncertainty Quantification 360 (UQ360) is an extensible open-source toolkit that can help you estimate, communicate and use uncertainty in machine learning model predictions.
Interpretability and explainability of data and machine learning models
The Granite Guardian models are designed to detect risks in prompts and responses.
This repository contains a collection of surveys, datasets, papers, and codes, for predictive uncertainty estimation in deep learning models.
Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models
nannyml: post-deployment data science in python
A curated list of 120+ LLM libraries category wise.
UQLM: Uncertainty Quantification for Language Models, is a Python package for UQ-based LLM hallucination detection
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights i…