Skip to content

πŸ€— Evaluate: A library for easily evaluating machine learning models and datasets.

License

Notifications You must be signed in to change notification settings

jsondai/evaluate

This branch is 12 commits behind huggingface/evaluate:main.

Folders and files

NameName
Last commit message
Last commit date
Apr 18, 2024
Sep 22, 2022
Mar 14, 2023
Feb 3, 2023
Apr 18, 2024
Apr 18, 2024
Mar 14, 2023
Apr 18, 2024
Apr 1, 2022
Sep 10, 2020
Apr 12, 2021
Dec 27, 2023
Apr 14, 2020
Jul 26, 2022
Dec 12, 2022
May 30, 2022
Nov 12, 2021
Apr 18, 2024

Repository files navigation



Build GitHub Documentation GitHub release Contributor Covenant

πŸ€— Evaluate is a library that makes evaluating and comparing models and reporting their performance easier and more standardized.

It currently contains:

  • implementations of dozens of popular metrics: the existing metrics cover a variety of tasks spanning from NLP to Computer Vision, and include dataset-specific metrics for datasets. With a simple command like accuracy = load("accuracy"), get any of these metrics ready to use for evaluating a ML model in any framework (Numpy/Pandas/PyTorch/TensorFlow/JAX).
  • comparisons and measurements: comparisons are used to measure the difference between models and measurements are tools to evaluate datasets.
  • an easy way of adding new evaluation modules to the πŸ€— Hub: you can create new evaluation modules and push them to a dedicated Space in the πŸ€— Hub with evaluate-cli create [metric name], which allows you to see easily compare different metrics and their outputs for the same sets of references and predictions.

πŸŽ“ Documentation

πŸ”Ž Find a metric, comparison, measurement on the Hub

🌟 Add a new evaluation module

πŸ€— Evaluate also has lots of useful features like:

  • Type checking: the input types are checked to make sure that you are using the right input formats for each metric
  • Metric cards: each metrics comes with a card that describes the values, limitations and their ranges, as well as providing examples of their usage and usefulness.
  • Community metrics: Metrics live on the Hugging Face Hub and you can easily add your own metrics for your project or to collaborate with others.

Installation

With pip

πŸ€— Evaluate can be installed from PyPi and has to be installed in a virtual environment (venv or conda for instance)

pip install evaluate

Usage

πŸ€— Evaluate's main methods are:

  • evaluate.list_evaluation_modules() to list the available metrics, comparisons and measurements
  • evaluate.load(module_name, **kwargs) to instantiate an evaluation module
  • results = module.compute(*kwargs) to compute the result of an evaluation module

Adding a new evaluation module

First install the necessary dependencies to create a new metric with the following command:

pip install evaluate[template]

Then you can get started with the following command which will create a new folder for your metric and display the necessary steps:

evaluate-cli create "Awesome Metric"

See this step-by-step guide in the documentation for detailed instructions.

Credits

Thanks to @marella for letting us use the evaluate namespace on PyPi previously used by his library.

About

πŸ€— Evaluate: A library for easily evaluating machine learning models and datasets.

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 99.9%
  • Makefile 0.1%