Update README.md
Browse files
README.md
CHANGED
|
@@ -1,29 +1,36 @@
|
|
| 1 |
---
|
|
|
|
| 2 |
language:
|
| 3 |
- ca
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
|
| 7 |
-
# PL-BERT-
|
|
|
|
| 8 |
|
| 9 |
## Overview
|
| 10 |
|
| 11 |
<details>
|
| 12 |
<summary>Click to expand</summary>
|
| 13 |
|
| 14 |
-
-
|
| 15 |
-
-
|
| 16 |
-
-
|
| 17 |
-
-
|
| 18 |
-
-
|
|
|
|
| 19 |
|
| 20 |
</details>
|
| 21 |
|
| 22 |
---
|
| 23 |
|
| 24 |
-
## Model
|
| 25 |
|
| 26 |
-
**PL-BERT-
|
| 27 |
|
| 28 |
This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
|
| 29 |
|
|
@@ -35,7 +42,7 @@ Features of our PL-BERT:
|
|
| 35 |
|
| 36 |
---
|
| 37 |
|
| 38 |
-
## Intended
|
| 39 |
|
| 40 |
### Intended uses
|
| 41 |
|
|
@@ -51,7 +58,7 @@ Features of our PL-BERT:
|
|
| 51 |
|
| 52 |
---
|
| 53 |
|
| 54 |
-
## How to
|
| 55 |
|
| 56 |
Here is an example of how to use this model within the StyleTTS2 framework:
|
| 57 |
|
|
@@ -77,7 +84,7 @@ Note: Although this example uses StyleTTS2, the model is compatible with other T
|
|
| 77 |
|
| 78 |
---
|
| 79 |
|
| 80 |
-
## Training
|
| 81 |
|
| 82 |
### Training data
|
| 83 |
|
|
@@ -114,7 +121,7 @@ Other parameters:
|
|
| 114 |
---
|
| 115 |
|
| 116 |
|
| 117 |
-
|
| 118 |
|
| 119 |
The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
|
| 120 |
|
|
@@ -125,37 +132,35 @@ The model has not been benchmarked via perplexity or extrinsic evaluation, but h
|
|
| 125 |
If this code contributes to your research, please cite the work:
|
| 126 |
|
| 127 |
```
|
| 128 |
-
@misc{
|
| 129 |
-
title={PL-BERT-
|
| 130 |
author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
|
| 131 |
organization={Barcelona Supercomputing Center},
|
| 132 |
-
url={https://huggingface.co/langtech-veu/PL-BERT-
|
| 133 |
year={2025}
|
| 134 |
}
|
| 135 |
```
|
| 136 |
|
| 137 |
-
## Additional information
|
| 138 |
|
| 139 |
-
|
| 140 |
|
| 141 |
-
For questions or feedback, please contact:
|
| 142 | |
| 143 |
|
| 144 |
-
###
|
| 145 |
|
| 146 |
-
|
| 147 |
|
| 148 |
-
###
|
| 149 |
-
|
| 150 |
|
|
|
|
|
|
|
| 151 |
|
|
|
|
| 152 |
|
| 153 |
-
|
| 154 |
|
| 155 |
-
<details>
|
| 156 |
-
<summary>Click to expand</summary>
|
| 157 |
|
| 158 |
-
|
|
|
|
| 159 |
|
| 160 |
-
</details>
|
| 161 |
|
|
|
|
| 1 |
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
language:
|
| 4 |
- ca
|
| 5 |
+
tags:
|
| 6 |
+
- TTS
|
| 7 |
+
- PL-BERT
|
| 8 |
+
- barcelona-supercomputing-center
|
| 9 |
---
|
| 10 |
|
| 11 |
|
| 12 |
+
# PL-BERT-ca
|
| 13 |
+
|
| 14 |
|
| 15 |
## Overview
|
| 16 |
|
| 17 |
<details>
|
| 18 |
<summary>Click to expand</summary>
|
| 19 |
|
| 20 |
+
- [Model Description](#model-description)
|
| 21 |
+
- [Intended Uses and Limitations](#intended-uses-and-limitations)
|
| 22 |
+
- [How to Get Started with the Model](#how-to-get-started-with-the-model)
|
| 23 |
+
- [Training Details](#training-details)
|
| 24 |
+
- [Citation](#citation)
|
| 25 |
+
- [Additional information](#additional-information)
|
| 26 |
|
| 27 |
</details>
|
| 28 |
|
| 29 |
---
|
| 30 |
|
| 31 |
+
## Model Description
|
| 32 |
|
| 33 |
+
**PL-BERT-ca** is a phoneme-level masked language model trained on Catalan text with diverse regional accents. It is based on the [PL-BERT architecture](https://github.com/yl4579/PL-BERT), which learns phoneme representations via a BERT-style masked language modeling objective.
|
| 34 |
|
| 35 |
This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
|
| 36 |
|
|
|
|
| 42 |
|
| 43 |
---
|
| 44 |
|
| 45 |
+
## Intended Uses and Limitations
|
| 46 |
|
| 47 |
### Intended uses
|
| 48 |
|
|
|
|
| 58 |
|
| 59 |
---
|
| 60 |
|
| 61 |
+
## How to Get Started with the Model
|
| 62 |
|
| 63 |
Here is an example of how to use this model within the StyleTTS2 framework:
|
| 64 |
|
|
|
|
| 84 |
|
| 85 |
---
|
| 86 |
|
| 87 |
+
## Training Details
|
| 88 |
|
| 89 |
### Training data
|
| 90 |
|
|
|
|
| 121 |
---
|
| 122 |
|
| 123 |
|
| 124 |
+
### Evaluation
|
| 125 |
|
| 126 |
The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
|
| 127 |
|
|
|
|
| 132 |
If this code contributes to your research, please cite the work:
|
| 133 |
|
| 134 |
```
|
| 135 |
+
@misc{zevallos2025plbertca,
|
| 136 |
+
title={PL-BERT-ca},
|
| 137 |
author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
|
| 138 |
organization={Barcelona Supercomputing Center},
|
| 139 |
+
url={https://huggingface.co/langtech-veu/PL-BERT-ca},
|
| 140 |
year={2025}
|
| 141 |
}
|
| 142 |
```
|
| 143 |
|
|
|
|
| 144 |
|
| 145 |
+
## Additional Information
|
| 146 |
|
|
|
|
|
|
|
| 147 |
|
| 148 |
+
### Author
|
| 149 |
|
| 150 |
+
The [Language Technologies Laboratory](https://huggingface.co/BSC-LT) of the [Barcelona Supercomputing Center](https://www.bsc.es/) by [Rodolfo Zevallos](https://huggingface.co/rjzevallos).
|
| 151 |
|
| 152 |
+
### Contact
|
| 153 |
+
For further information, please send an email to <langtech@bsc.es>.
|
| 154 |
|
| 155 |
+
### Copyright
|
| 156 |
+
Copyright(c) 2025 by Language Technologies Laboratory, Barcelona Supercomputing Center.
|
| 157 |
|
| 158 |
+
### License
|
| 159 |
|
| 160 |
+
[Apache-2.0](https://www.apache.org/licenses/LICENSE-2.0)
|
| 161 |
|
|
|
|
|
|
|
| 162 |
|
| 163 |
+
### Funding
|
| 164 |
+
This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA.
|
| 165 |
|
|
|
|
| 166 |
|