| | --- |
| | license: mit |
| | datasets: |
| | - openslr/librispeech_asr |
| | language: |
| | - en |
| | pipeline_tag: automatic-speech-recognition |
| | --- |
| | |
| | # Splitformer |
| |
|
| | <div align="center" style="line-height: 1;"> |
| | <a href="https://github.com/augustgw/early-exit-transformer" target="_blank" style="margin: 2px;"> |
| | <img alt="GitHub" src="https://img.shields.io/badge/GitHub-Splitformer-181717?logo=github&logoColor=white" style="display: inline-block; vertical-align: middle;"/> |
| | </a> |
| | <a href="https://www.arxiv.org/abs/2506.18035" target="_blank" style="margin: 2px;"> |
| | <img alt="arXiv" src="https://img.shields.io/badge/arXiv-2506.18035-B31B1B?logo=arxiv&logoColor=white" style="display: inline-block; vertical-align: middle;"/> |
| | </a> |
| | </div> |
| | |
| |
|
| | ## 1. Overview |
| |
|
| | **Splitformer** is a 36.7M parameters Conformer-based ASR model trained from scratch on 1000 hours of the **LibriSpeech dataset** with an **early‐exit objective**. |
| |
|
| | This architecture introduces **parallel downsampling layers** before the first and last exits to improve performance with minimal extra overhead, while retaining inference speed. |
| |
|
| | Our code for training and inference is available on our [GitHub](https://github.com/augustgw/early-exit-transformer) repository. |
| |
|
| | ### 2. Results on LibriSpeech |
| |
|
| | <table> |
| | <thead> |
| | <tr> |
| | <th rowspan="2">Layer</th> |
| | <th colspan="2">EE-baseline (31.5M)</th> |
| | <th colspan="2">Splitformer (36.7M)</th> |
| | <th colspan="2">Wav2Vec2 (94.0M)</th> |
| | <th colspan="2">WavLM (94.7M)</th> |
| | </tr> |
| | <tr> |
| | <th>test-clean</th> |
| | <th>test-other</th> |
| | <th>test-clean</th> |
| | <th>test-other</th> |
| | <th>test-clean</th> |
| | <th>test-other</th> |
| | <th>test-clean</th> |
| | <th>test-other</th> |
| | </tr> |
| | </thead> |
| | <tbody> |
| | <tr> |
| | <td>2</td> |
| | <td>31.0</td> |
| | <td>51.0</td> |
| | <td>28.1</td> |
| | <td>48.3</td> |
| | <td>33.7</td> |
| | <td>56.0</td> |
| | <td>28.0</td> |
| | <td>48.5</td> |
| | </tr> |
| | <tr> |
| | <td>4</td> |
| | <td>11.7</td> |
| | <td>27.8</td> |
| | <td>10.8</td> |
| | <td>26.4</td> |
| | <td>17.4</td> |
| | <td>36.7</td> |
| | <td>13.9</td> |
| | <td>27.3</td> |
| | </tr> |
| | <tr> |
| | <td>6</td> |
| | <td>7.1</td> |
| | <td>19.8</td> |
| | <td>6.7</td> |
| | <td>19.2</td> |
| | <td>9.6</td> |
| | <td>23.7</td> |
| | <td>8.7</td> |
| | <td>18.4</td> |
| | </tr> |
| | <tr> |
| | <td>8</td> |
| | <td>5.8</td> |
| | <td>16.6</td> |
| | <td>5.5</td> |
| | <td>16.3</td> |
| | <td>5.8</td> |
| | <td>15.9</td> |
| | <td>4.8</td> |
| | <td>12.4</td> |
| | </tr> |
| | <tr> |
| | <td>10</td> |
| | <td>5.3</td> |
| | <td>15.3</td> |
| | <td>5.1</td> |
| | <td>15.1</td> |
| | <td>4.5</td> |
| | <td>12.6</td> |
| | <td>4.0</td> |
| | <td>9.5</td> |
| | </tr> |
| | <tr> |
| | <td>12</td> |
| | <td>5.1</td> |
| | <td>14.8</td> |
| | <td>4.8</td> |
| | <td>14.7</td> |
| | <td>4.3</td> |
| | <td>12.2</td> |
| | <td>3.6</td> |
| | <td>8.8</td> |
| | </tr> |
| | </tbody> |
| | </table> |
| | |
| | ## 3. Citation |
| |
|
| | ```bibtex |
| | @misc{lasbordes2025splitformer, |
| | title={Splitformer: An improved early-exit architecture for automatic speech recognition on edge devices}, |
| | author={Maxence Lasbordes, Daniele Falavigna and Alessio Brutti}, |
| | year={2025}, |
| | note={Proc. of EUSIPCO 2025}, |
| | } |
| | |