File size: 9,349 Bytes
1340f33 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 |
Metadata-Version: 2.1
Name: evaluate
Version: 0.4.2
Summary: HuggingFace community-driven open-source library of evaluation
Home-page: https://github.com/huggingface/evaluate
Author: HuggingFace Inc.
Author-email: [email protected]
License: Apache 2.0
Download-URL: https://github.com/huggingface/evaluate/tags
Keywords: metrics machine learning evaluate evaluation
Platform: UNKNOWN
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Education
Classifier: Intended Audience :: Science/Research
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Operating System :: OS Independent
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.8
Classifier: Programming Language :: Python :: 3.9
Classifier: Programming Language :: Python :: 3.10
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Requires-Python: >=3.8.0
Description-Content-Type: text/markdown
License-File: LICENSE
Requires-Dist: datasets >=2.0.0
Requires-Dist: numpy >=1.17
Requires-Dist: dill
Requires-Dist: pandas
Requires-Dist: requests >=2.19.0
Requires-Dist: tqdm >=4.62.1
Requires-Dist: xxhash
Requires-Dist: multiprocess
Requires-Dist: fsspec[http] >=2021.05.0
Requires-Dist: huggingface-hub >=0.7.0
Requires-Dist: packaging
Requires-Dist: importlib-metadata ; python_version < "3.8"
Provides-Extra: dev
Requires-Dist: absl-py ; extra == 'dev'
Requires-Dist: charcut >=1.1.1 ; extra == 'dev'
Requires-Dist: cer >=1.2.0 ; extra == 'dev'
Requires-Dist: nltk ; extra == 'dev'
Requires-Dist: pytest ; extra == 'dev'
Requires-Dist: pytest-datadir ; extra == 'dev'
Requires-Dist: pytest-xdist ; extra == 'dev'
Requires-Dist: tensorflow !=2.6.0,!=2.6.1,<=2.10,>=2.3 ; extra == 'dev'
Requires-Dist: torch ; extra == 'dev'
Requires-Dist: accelerate ; extra == 'dev'
Requires-Dist: bert-score >=0.3.6 ; extra == 'dev'
Requires-Dist: rouge-score >=0.1.2 ; extra == 'dev'
Requires-Dist: sacrebleu ; extra == 'dev'
Requires-Dist: sacremoses ; extra == 'dev'
Requires-Dist: scipy >=1.10.0 ; extra == 'dev'
Requires-Dist: seqeval ; extra == 'dev'
Requires-Dist: scikit-learn ; extra == 'dev'
Requires-Dist: jiwer ; extra == 'dev'
Requires-Dist: sentencepiece ; extra == 'dev'
Requires-Dist: transformers ; extra == 'dev'
Requires-Dist: mauve-text ; extra == 'dev'
Requires-Dist: trectools ; extra == 'dev'
Requires-Dist: toml >=0.10.1 ; extra == 'dev'
Requires-Dist: requests-file >=1.5.1 ; extra == 'dev'
Requires-Dist: tldextract >=3.1.0 ; extra == 'dev'
Requires-Dist: texttable >=1.6.3 ; extra == 'dev'
Requires-Dist: unidecode >=1.3.4 ; extra == 'dev'
Requires-Dist: Werkzeug >=1.0.1 ; extra == 'dev'
Requires-Dist: six ~=1.15.0 ; extra == 'dev'
Requires-Dist: black ~=22.0 ; extra == 'dev'
Requires-Dist: flake8 >=3.8.3 ; extra == 'dev'
Requires-Dist: isort >=5.0.0 ; extra == 'dev'
Requires-Dist: pyyaml >=5.3.1 ; extra == 'dev'
Provides-Extra: docs
Requires-Dist: s3fs ; extra == 'docs'
Provides-Extra: evaluator
Requires-Dist: transformers ; extra == 'evaluator'
Requires-Dist: scipy >=1.7.1 ; extra == 'evaluator'
Provides-Extra: quality
Requires-Dist: black ~=22.0 ; extra == 'quality'
Requires-Dist: flake8 >=3.8.3 ; extra == 'quality'
Requires-Dist: isort >=5.0.0 ; extra == 'quality'
Requires-Dist: pyyaml >=5.3.1 ; extra == 'quality'
Provides-Extra: template
Requires-Dist: cookiecutter ; extra == 'template'
Requires-Dist: gradio >=3.0.0 ; extra == 'template'
Provides-Extra: tensorflow
Requires-Dist: tensorflow !=2.6.0,!=2.6.1,>=2.2.0 ; extra == 'tensorflow'
Provides-Extra: tensorflow_gpu
Requires-Dist: tensorflow-gpu !=2.6.0,!=2.6.1,>=2.2.0 ; extra == 'tensorflow_gpu'
Provides-Extra: tests
Requires-Dist: absl-py ; extra == 'tests'
Requires-Dist: charcut >=1.1.1 ; extra == 'tests'
Requires-Dist: cer >=1.2.0 ; extra == 'tests'
Requires-Dist: nltk ; extra == 'tests'
Requires-Dist: pytest ; extra == 'tests'
Requires-Dist: pytest-datadir ; extra == 'tests'
Requires-Dist: pytest-xdist ; extra == 'tests'
Requires-Dist: tensorflow !=2.6.0,!=2.6.1,<=2.10,>=2.3 ; extra == 'tests'
Requires-Dist: torch ; extra == 'tests'
Requires-Dist: accelerate ; extra == 'tests'
Requires-Dist: bert-score >=0.3.6 ; extra == 'tests'
Requires-Dist: rouge-score >=0.1.2 ; extra == 'tests'
Requires-Dist: sacrebleu ; extra == 'tests'
Requires-Dist: sacremoses ; extra == 'tests'
Requires-Dist: scipy >=1.10.0 ; extra == 'tests'
Requires-Dist: seqeval ; extra == 'tests'
Requires-Dist: scikit-learn ; extra == 'tests'
Requires-Dist: jiwer ; extra == 'tests'
Requires-Dist: sentencepiece ; extra == 'tests'
Requires-Dist: transformers ; extra == 'tests'
Requires-Dist: mauve-text ; extra == 'tests'
Requires-Dist: trectools ; extra == 'tests'
Requires-Dist: toml >=0.10.1 ; extra == 'tests'
Requires-Dist: requests-file >=1.5.1 ; extra == 'tests'
Requires-Dist: tldextract >=3.1.0 ; extra == 'tests'
Requires-Dist: texttable >=1.6.3 ; extra == 'tests'
Requires-Dist: unidecode >=1.3.4 ; extra == 'tests'
Requires-Dist: Werkzeug >=1.0.1 ; extra == 'tests'
Requires-Dist: six ~=1.15.0 ; extra == 'tests'
Provides-Extra: torch
Requires-Dist: torch ; extra == 'torch'
<p align="center">
<br>
<img src="https://huggingface.co/datasets/evaluate/media/resolve/main/evaluate-banner.png" width="400"/>
<br>
</p>
<p align="center">
<a href="https://github.com/huggingface/evaluate/actions/workflows/ci.yml?query=branch%3Amain">
<img alt="Build" src="https://github.com/huggingface/evaluate/actions/workflows/ci.yml/badge.svg?branch=main">
</a>
<a href="https://github.com/huggingface/evaluate/blob/master/LICENSE">
<img alt="GitHub" src="https://img.shields.io/github/license/huggingface/evaluate.svg?color=blue">
</a>
<a href="https://huggingface.co/docs/evaluate/index">
<img alt="Documentation" src="https://img.shields.io/website/http/huggingface.co/docs/evaluate/index.svg?down_color=red&down_message=offline&up_message=online">
</a>
<a href="https://github.com/huggingface/evaluate/releases">
<img alt="GitHub release" src="https://img.shields.io/github/release/huggingface/evaluate.svg">
</a>
<a href="CODE_OF_CONDUCT.md">
<img alt="Contributor Covenant" src="https://img.shields.io/badge/Contributor%20Covenant-2.0-4baaaa.svg">
</a>
</p>
π€ Evaluate is a library that makes evaluating and comparing models and reporting their performance easier and more standardized.
It currently contains:
- **implementations of dozens of popular metrics**: the existing metrics cover a variety of tasks spanning from NLP to Computer Vision, and include dataset-specific metrics for datasets. With a simple command like `accuracy = load("accuracy")`, get any of these metrics ready to use for evaluating a ML model in any framework (Numpy/Pandas/PyTorch/TensorFlow/JAX).
- **comparisons and measurements**: comparisons are used to measure the difference between models and measurements are tools to evaluate datasets.
- **an easy way of adding new evaluation modules to the π€ Hub**: you can create new evaluation modules and push them to a dedicated Space in the π€ Hub with `evaluate-cli create [metric name]`, which allows you to see easily compare different metrics and their outputs for the same sets of references and predictions.
[π **Documentation**](https://huggingface.co/docs/evaluate/)
π **Find a [metric](https://huggingface.co/evaluate-metric), [comparison](https://huggingface.co/evaluate-comparison), [measurement](https://huggingface.co/evaluate-measurement) on the Hub**
[π **Add a new evaluation module**](https://huggingface.co/docs/evaluate/)
π€ Evaluate also has lots of useful features like:
- **Type checking**: the input types are checked to make sure that you are using the right input formats for each metric
- **Metric cards**: each metrics comes with a card that describes the values, limitations and their ranges, as well as providing examples of their usage and usefulness.
- **Community metrics:** Metrics live on the Hugging Face Hub and you can easily add your own metrics for your project or to collaborate with others.
# Installation
## With pip
π€ Evaluate can be installed from PyPi and has to be installed in a virtual environment (venv or conda for instance)
```bash
pip install evaluate
```
# Usage
π€ Evaluate's main methods are:
- `evaluate.list_evaluation_modules()` to list the available metrics, comparisons and measurements
- `evaluate.load(module_name, **kwargs)` to instantiate an evaluation module
- `results = module.compute(*kwargs)` to compute the result of an evaluation module
# Adding a new evaluation module
First install the necessary dependencies to create a new metric with the following command:
```bash
pip install evaluate[template]
```
Then you can get started with the following command which will create a new folder for your metric and display the necessary steps:
```bash
evaluate-cli create "Awesome Metric"
```
See this [step-by-step guide](https://huggingface.co/docs/evaluate/creating_and_sharing) in the documentation for detailed instructions.
## Credits
Thanks to [@marella](https://github.com/marella) for letting us use the `evaluate` namespace on PyPi previously used by his [library](https://github.com/marella/evaluate).
|