Skip to content
This repository has been archived by the owner on Oct 10, 2023. It is now read-only.

zeno-ml/zeno-evals-hub

Repository files navigation

Zeno + OpenAI Evals

Github Actions CI tests MIT license Discord

OpenAI's Evals library is a great resource providing evaluation sets for LLMS.

This repo provides a hub for exploring these results using the Zeno evaluation tool.

Add New Evals

To add new evals, add a new entry to evals/evals.yaml with the following fields:

  • results-file: The first .jsonl result from oaievals
  • link: A link to the evals commit for this evaluation
  • description: A succint description of what the evaluation is testing
  • second-results-file: An optional second .jsonl result from oaievals. Must be the same dataset as the first one.
  • functions-file: An optional Python file with Zeno functions for the evaluations.

Make sure you test your evals locally before submitting a PR!

Running

poetry install

python -m zeno-evals-hub evals/evals.yaml