Spaces:
Running
Running
from dataclasses import dataclass | |
from enum import Enum | |
# Your leaderboard name | |
TITLE = """<h1 align="center" id="space-title">ποΈ ACL-25 SpeechIQ Leaderboard</h1>""" | |
# What does your leaderboard evaluate? | |
INTRODUCTION_TEXT = """ | |
## π― Welcome to the SpeechIQ Leaderboard! | |
This leaderboard presents evaluation results for voice understanding large language models (LLM<sub>Voice</sub>) using our novel SpeechIQ evaluation framework. | |
The **Speech IQ Score** provides a unified metric for comparing both cascaded methods (ASR+LLM) and end-to-end models. | |
""" | |
# Which evaluations are you running? how can people reproduce what you have? | |
LLM_BENCHMARKS_TEXT = """ | |
## π About SpeechIQ Evaluation | |
**Speech Intelligence Quotient (SpeechIQ)** represents a first-of-its-kind intelligence examination that bridges cognitive principles with voice-oriented benchmarks. Our framework moves beyond traditional metrics like Word Error Rate (WER) to provide comprehensive evaluation of voice understanding capabilities. | |
### π― Evaluation Framework | |
SpeechIQ evaluates models across three cognitive dimensions inspired by Bloom's Taxonomy: | |
1. **Remember** (Verbatim Accuracy): Tests the model's ability to accurately capture spoken content | |
2. **Understand** (Interpretation Similarity): Evaluates how well the model comprehends the meaning of speech | |
3. **Apply** (Downstream Performance): Measures the model's ability to use speech understanding for practical tasks | |
### π Model Categories | |
- **Agentic (ASR + LLM)**: Cascaded approaches using separate ASR and LLM components | |
- **End2End**: Direct speech-to-text models that process audio end-to-end | |
### π¬ Key Benefits | |
- **Unified Comparisons**: Compare cascaded and end-to-end approaches on equal footing | |
- **Error Detection**: Identify annotation errors in existing benchmarks | |
- **Hallucination Detection**: Detect and quantify hallucinations in voice LLMs | |
- **Cognitive Assessment**: Map model capabilities to human cognitive principles | |
### π Speech IQ Score | |
The final Speech IQ Score combines performance across all three dimensions to provide a comprehensive measure of voice understanding intelligence. | |
## π Reproducibility | |
For detailed methodology and reproduction instructions, please refer to our paper and codebase. | |
""" | |
EVALUATION_QUEUE_TEXT = """ | |
## π Submit Your Model for SpeechIQ Evaluation | |
To submit your voice understanding model for SpeechIQ evaluation: | |
### 1) Ensure Model Compatibility | |
Make sure your model can process audio inputs and generate text outputs in one of these formats: | |
- **ASR + LLM**: Separate ASR and LLM components | |
- **End-to-End**: Direct audio-to-text processing | |
### 2) Model Requirements | |
- Model must be publicly accessible | |
- Provide clear documentation of audio input format and expected outputs | |
- Include information about audio encoder specifications | |
### 3) Evaluation Domains | |
Your model will be evaluated across: | |
- **Remember**: Transcription accuracy | |
- **Understand**: Semantic understanding | |
- **Apply**: Task-specific performance | |
### 4) Documentation | |
Please provide: | |
- Model architecture details | |
- Training data information | |
- Audio preprocessing requirements | |
- Expected input/output formats | |
## π§ Contact | |
For questions about SpeechIQ evaluation or to submit your model, please contact the research team. | |
""" | |
CITATION_BUTTON_LABEL = "Refer to the following ACL 2025 main conference paper." | |
CITATION_BUTTON_TEXT = r"""@article{speechiq2025, | |
title={SpeechIQ: Speech Intelligence Quotient Across Cognitive Levels in Voice Understanding Large Language Models}, | |
author={Zhen Wan, Chao-Han Huck Yang, Yahan Yu, Jinchuan Tian, Sheng Li, Ke Hu, Zhehuai Chen, Shinji Watanabe, Fei Cheng, Chenhui Chu, Sadao Kurohashi}, | |
journal={ACL 2025 main conference}, | |
year={2025} | |
}""" | |