Skip to content

NilayYadav/evals

 
 

Repository files navigation

🤖📊 TrainLoop Evals

Ask DeepWiki

TrainLoop Evals is a comprehensive LLM evaluation framework designed for developers who need simple, vendor-independent evaluation tools.

Core Principles

  • Simplicity First – One environment variable, one function call, one folder of JSON files
  • Vendor Independence – Everything stored as newline-delimited JSON; no databases required
  • Type-Safe & Extensible – In-code tests with full TypeScript support and composable system
  • Meet Developers Where They Are – Works with existing workflows and bespoke loops

Evals Meme

Quick Start

# Install the CLI
pipx install trainloop-cli

# Create a workspace
trainloop init

# Set your data path
export TRAINLOOP_DATA_FOLDER=/path/to/data

# Run evaluations
trainloop eval

# View results
trainloop studio

📚 Documentation

For comprehensive documentation, installation guides, tutorials, and API reference:

👉 evals.docs.trainloop.ai

👉 DeepWiki - lets you chat directly with this codebase rather than wading through documentation. It's the purest form of talking to your docs.

Quick Links

Demo

Support

License

MIT


Need help? Check out our comprehensive documentation or open an issue.

About

Open Source Data Collection and Evaluation Framework

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 53.2%
  • Python 41.8%
  • Go 3.7%
  • JavaScript 0.5%
  • CSS 0.4%
  • Shell 0.3%
  • Dockerfile 0.1%