Update README.md
Browse files
README.md
CHANGED
@@ -1,29 +1,68 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
-
|
7 |
|
8 |
-
|
|
|
|
|
9 |
|
10 |
|
11 |
|
12 |
-
|
13 |
|
14 |
-
|
15 |
|
16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
-
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
19 |
|
20 |
-
- **Developed by:** [More Information Needed]
|
21 |
-
- **Funded by [optional]:** [More Information Needed]
|
22 |
-
- **Shared by [optional]:** [More Information Needed]
|
23 |
-
- **Model type:** [More Information Needed]
|
24 |
-
- **Language(s) (NLP):** [More Information Needed]
|
25 |
-
- **License:** [More Information Needed]
|
26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
27 |
|
28 |
### Model Sources [optional]
|
29 |
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
license: other
|
4 |
+
license_name: meralion-public-license
|
5 |
+
license_link: >-
|
6 |
+
https://huggingface.co/MERaLiON/MERaLiON-AudioLLM-Whisper-SEA-LION/blob/main/MERaLiON-Public-Licence-v1.pdf
|
7 |
+
tags:
|
8 |
+
- speech
|
9 |
+
- best-rq
|
10 |
+
- meralion
|
11 |
+
- meralion-2
|
12 |
+
language:
|
13 |
+
- en
|
14 |
+
- zh
|
15 |
+
- ms
|
16 |
+
- ta
|
17 |
+
- id
|
18 |
+
- th
|
19 |
+
- vi
|
20 |
---
|
21 |
|
22 |
+
<h1 align="center">🎧 MERaLiON-SpeechEncoder-2 🎧</h1>
|
23 |
|
24 |
+
<p align="center">
|
25 |
+
<a href="https://meralion.org/demo/">💻 ASR Web Demo (Coming Soon!)</a> |
|
26 |
+
</p>
|
27 |
|
28 |
|
29 |
|
30 |
+
We introduce **MERaLiON-SpeechEncoder-2**, an update of [MERaLiON-SpeechEncoder-v1](https://huggingface.co/MERaLiON/MERaLiON-SpeechEncoder-v1) that greatly expands our pre-training data to **1.4 million hours** of unlabeled audio, with a **strong focus on Southeast Asian (SEA) languages and accents**. As a speech foundation model, it encodes speech into a general-purpose, multilingual acoustic representation that can serve as a high-performance backbone for a wide range of downstream tasks -- including automatic speech recognition (ASR), speech translation, speaker and language identification, and emotion recognition.
|
31 |
|
32 |
+
Unlike many existing models optimized for high-resource, Western languages, MERaLiON-SpeechEncoder-2 is designed from the ground up to reflect the linguistic diversity and complexity of Southeast Asia. The model can be finetuned on custom datasets, allowing developers to build speech systems tailored to their specific needs.
|
33 |
|
34 |
+
## Model Highlights
|
35 |
+
|
36 |
+
### Small model size
|
37 |
+
With only 630M parameters (≈2.5 GB in memory), the model is easily deployable on most commercial GPUs, eliminating the need for distributed or large-scale compute setups.
|
38 |
+
|
39 |
+
### Natively multilingual
|
40 |
+
Building on our [v1 release](https://huggingface.co/MERaLiON/MERaLiON-SpeechEncoder-v1) (which focused on English and Singlish), this version expands to include English, Chinese, Malay, Tamil, Thai, Indonesian, and Vietnamese, along with codeswitching support across these languages. Given the wide coverage of languages in the training corpus, it may also be applicable beyond the officially supported languages.
|
41 |
+
|
42 |
+
### Competitive performance on downstream speech tasks
|
43 |
+
The model retains near state-of-the-art results on the SUPERB benchmark for English, and showcases strong multilingual capabilities deomnstrated through its integration into a high-performance ASR system shown below.
|
44 |
+
|
45 |
+
### Innovative pre-training techniques
|
46 |
+
MERaLiON-SpeechEncoder-2 was trained from scratch with an novel extension of the BEST-RQ self-supervised objective, by using more informative latent targets. We also adopted the Muon optimizer, which has previously only been shown to outperform the popular AdamW for LLM training. We find its advantages also carry over to speech-based models.
|
47 |
+
|
48 |
+
## Model Summary
|
49 |
+
|
50 |
+
- **Developed by:** I<sup>2</sup>R, A\*STAR
|
51 |
+
- **Model type:** Speech Encoder
|
52 |
+
- **Language(s):** Primarily English (Global & Singapore), Chinese, Malay, Tamil, Thai, Indonesian, and Vietnamese. See [pre-training data](#Language coverage of pre-training data) for full breakdown of language coverage.
|
53 |
+
- **License:** [MERaLiON Public License](https://huggingface.co/MERaLiON/MERaLiON-AudioLLM-Whisper-SEA-LION/blob/main/MERaLiON-Public-Licence-v1.pdf)
|
54 |
+
|
55 |
+
The following Hugging Face-compatible models are implemented:
|
56 |
+
|
57 |
+
- **`MeralionBestRqModel`**: The base BEST-RQ Conformer encoder. It outputs the final hidden states and is suitable for feature extraction or as a base for other heads.
|
58 |
+
- **`MeralionBestRqModelForCTC`**: The Conformer model with a linear CTC head for ASR.
|
59 |
+
- **`MeralionBestRqModelForLSTMCTC`**: The Conformer model with a more powerful CTC head that includes two LSTM layers before the final projection layer. This version can also be configured to use a weighted sum of all encoder hidden states.
|
60 |
+
|
61 |
+
For details on background, pre-training, tuning experiments and evaluation, please refer to our [technical report](https://arxiv.org/abs/2412.11538).
|
62 |
+
|
63 |
+
## Language coverage of pre-training data
|
64 |
|
|
|
65 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
|
67 |
### Model Sources [optional]
|
68 |
|