|
--- |
|
title: Accuracy |
|
emoji: π |
|
colorFrom: blue |
|
colorTo: red |
|
sdk: gradio |
|
sdk_version: 4.44.0 |
|
app_file: app.py |
|
pinned: false |
|
license: apache-2.0 |
|
tags: |
|
- evaluate |
|
- metric |
|
short_description: Accuracy (ACC) |
|
description: >- |
|
The Accuracy (ACC) metric is used to measure the proportion of correctly |
|
predicted sequences compared to the total number of sequences. This metric can |
|
handle both integer and string inputs by converting them to strings for |
|
comparison. The ACC ranges from 0 to 1, where 1 indicates perfect accuracy |
|
(all predictions are correct) and 0 indicates complete failure (no predictions |
|
are correct). It is particularly useful in tasks such as OCR, digit |
|
recognition, sequence prediction, and any task where exact matches are |
|
required. The accuracy can be calculated using the formula: ACC = (Number of |
|
Correct Predictions) / (Total Number of Predictions) Where a prediction is |
|
considered correct if it exactly matches the ground truth sequence after |
|
converting both to strings. |
|
--- |
|
# Metric Card for ACC |
|
|
|
## Metric Description |
|
|
|
The Accuracy (ACC) metric is used to measure the proportion of correctly predicted sequences compared to the total number of sequences. This metric can handle both integer and string inputs by converting them to strings for comparison. |
|
The ACC ranges from 0 to 1, where 1 indicates perfect accuracy (all predictions are correct) and 0 indicates complete failure (no predictions are correct). |
|
It is particularly useful in tasks such as OCR, digit recognition, sequence prediction, and any task where exact matches are required. The accuracy can be calculated using the formula: |
|
|
|
ACC = (Number of Correct Predictions) / (Total Number of Predictions) |
|
|
|
Where: |
|
|
|
A prediction is considered correct if it exactly matches the ground truth sequence after converting both to strings. |