MaxLSB commited on
Commit
9a3c6c1
·
verified ·
1 Parent(s): c815970

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -3
README.md CHANGED
@@ -1,3 +1,31 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - openslr/librispeech_asr
5
+ language:
6
+ - en
7
+ pipeline_tag: automatic-speech-recognition
8
+ ---
9
+ <hr>
10
+ <div align="center" style="line-height: 1;">
11
+ <a href="https://github.com/augustgw/early-exit-transformer" target="_blank" style="margin: 2px;">
12
+ <img alt="GitHub" src="https://img.shields.io/badge/GitHub-Splitformer-181717?logo=github&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
13
+ </a>
14
+ <a href="LICENSE" style="margin: 2px;">
15
+ <img alt="License" src="https://img.shields.io/badge/License-MIT-f5de53?&color=f5de53" style="display: inline-block; vertical-align: middle;"/>
16
+ </a>
17
+ <a href="https://arxiv.org/abs/2501.12948" target="_blank" style="margin: 2px;">
18
+ <img alt="arXiv" src="https://img.shields.io/badge/arXiv-2501.12948-b31b1b?logo=arxiv&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
19
+ </a>
20
+ </div>
21
+
22
+ ---
23
+
24
+ # SplitFormer
25
+
26
+ ## Overview
27
+ Splitformer is a Conformer-based ASR model (36.7M parameters) trained from scratch on 1000 hours of the LibriSpeech dataset with an early‐exit objective.
28
+
29
+ This architecture introduces parallel downsampling layers before the first and last exits to improve performance with minimal extra overhead, while retaining inference speed.
30
+
31
+ Our code is available on our [GitHub](https://github.com/augustgw/early-exit-transformer) repository.