Spaces:
Running
Running
File size: 1,891 Bytes
360f81c cf29557 360f81c 968b189 360f81c 7bdf0cd e3571c1 10ee5bf 195dbfa b10121d 19b22c9 7109f43 19b22c9 d0311ac b10121d a679cf2 b10121d abd945c 7109f43 36fdd36 b10121d 7109f43 b10121d 4e9ddf9 b10121d 4e9ddf9 b10121d 1f4ddf2 87d2c55 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 |
---
title: "ML.ENERGY Leaderboard"
emoji: "⚡"
python_version: "3.9"
app_file: "app.py"
sdk: "gradio"
sdk_version: "3.39.0"
pinned: true
tags: ["energy", "leaderboard"]
---
# ML.ENERGY Leaderboard
[](https://ml.energy/leaderboard)
[](https://github.com/ml-energy/leaderboard/actions/workflows/push_spaces.yaml)
[](/LICENSE)
How much energy do GenAI models like LLMs and Diffusion models consume?
This README focuses on explaining how to run the benchmark yourself.
The actual leaderboard is here: https://ml.energy/leaderboard.
Read our paper [here](https://arxiv.org/abs/2505.06371)!
## Repository Organization
```
leaderboard/
├── benchmark/ # Benchmark scripts & instructions
├── data/ # Benchmark results
├── deployment/ # Colosseum deployment files
├── spitfight/ # Python package for the Colosseum
├── app.py # Leaderboard Gradio app definition
└── index.html # Embeds the leaderboard HuggingFace Space
```
## Colosseum
We instrumented [Hugging Face TGI](https://github.com/huggingface/text-generation-inference) so that it measures and returns GPU energy consumption.
Then, our [controller](/spitfight/colosseum/controller) server receives user prompts from the [Gradio app](/app.py), selects two models randomly, and streams model responses back with energy consumption.
## Running the Benchmark
We open-sourced the entire benchmark with instructions here: [`./benchmark`](./benchmark)
## Citation
Please refer to our BibTeX file: [`citation.bib`](/docs/citation.bib).
|