rjzevallos commited on
Commit
c8c13ee
·
verified ·
1 Parent(s): 32967ff

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -28
README.md CHANGED
@@ -1,29 +1,36 @@
1
  ---
 
2
  language:
3
  - ca
 
 
 
 
4
  ---
5
 
6
 
7
- # PL-BERT-cat-multiaccent
 
8
 
9
  ## Overview
10
 
11
  <details>
12
  <summary>Click to expand</summary>
13
 
14
- - **Model type:** Phoneme-level Language Model (PL-BERT)
15
- - **Architecture:** AlBERT-base (12 layers, 768 hidden units, 12 attention heads)
16
- - **Language:** Catalan (multiple accents)
17
- - **License:** Apache 2.0
18
- - **Data:** Crowdsourced phonemized Catalan speech text
 
19
 
20
  </details>
21
 
22
  ---
23
 
24
- ## Model description
25
 
26
- **PL-BERT-cat-multiaccent** is a phoneme-level masked language model trained on Catalan text with diverse regional accents. It is based on the [PL-BERT architecture](https://github.com/yl4579/PL-BERT), which learns phoneme representations via a BERT-style masked language modeling objective.
27
 
28
  This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
29
 
@@ -35,7 +42,7 @@ Features of our PL-BERT:
35
 
36
  ---
37
 
38
- ## Intended uses and limitations
39
 
40
  ### Intended uses
41
 
@@ -51,7 +58,7 @@ Features of our PL-BERT:
51
 
52
  ---
53
 
54
- ## How to use (with StyleTTS2)
55
 
56
  Here is an example of how to use this model within the StyleTTS2 framework:
57
 
@@ -77,7 +84,7 @@ Note: Although this example uses StyleTTS2, the model is compatible with other T
77
 
78
  ---
79
 
80
- ## Training
81
 
82
  ### Training data
83
 
@@ -114,7 +121,7 @@ Other parameters:
114
  ---
115
 
116
 
117
- ## Evaluation
118
 
119
  The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
120
 
@@ -125,37 +132,35 @@ The model has not been benchmarked via perplexity or extrinsic evaluation, but h
125
  If this code contributes to your research, please cite the work:
126
 
127
  ```
128
- @misc{LangtechVeu2025plbertcatmultiaccent,
129
- title={PL-BERT-cat-multiaccent},
130
  author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
131
  organization={Barcelona Supercomputing Center},
132
- url={https://huggingface.co/langtech-veu/PL-BERT-cat-multiaccent},
133
  year={2025}
134
  }
135
  ```
136
 
137
- ## Additional information
138
 
139
- ### Contact
140
 
141
- For questions or feedback, please contact:
142
143
 
144
- ### License
145
 
146
- Distributed under the Apache License, Version 2.0: https://www.apache.org/licenses/LICENSE-2.0
147
 
148
- ### Funding
149
- This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA.
150
 
 
 
151
 
 
152
 
153
- ### Disclaimer
154
 
155
- <details>
156
- <summary>Click to expand</summary>
157
 
158
- This model is released for research and educational use. It may exhibit biases or limitations based on training data characteristics. Users are responsible for ensuring appropriate use in deployed systems and for complying with all applicable regulations.
 
159
 
160
- </details>
161
 
 
1
  ---
2
+ license: apache-2.0
3
  language:
4
  - ca
5
+ tags:
6
+ - TTS
7
+ - PL-BERT
8
+ - barcelona-supercomputing-center
9
  ---
10
 
11
 
12
+ # PL-BERT-ca
13
+
14
 
15
  ## Overview
16
 
17
  <details>
18
  <summary>Click to expand</summary>
19
 
20
+ - [Model Description](#model-description)
21
+ - [Intended Uses and Limitations](#intended-uses-and-limitations)
22
+ - [How to Get Started with the Model](#how-to-get-started-with-the-model)
23
+ - [Training Details](#training-details)
24
+ - [Citation](#citation)
25
+ - [Additional information](#additional-information)
26
 
27
  </details>
28
 
29
  ---
30
 
31
+ ## Model Description
32
 
33
+ **PL-BERT-ca** is a phoneme-level masked language model trained on Catalan text with diverse regional accents. It is based on the [PL-BERT architecture](https://github.com/yl4579/PL-BERT), which learns phoneme representations via a BERT-style masked language modeling objective.
34
 
35
  This model is designed to support **phoneme-based text-to-speech (TTS) systems**, including but not limited to [StyleTTS2](https://github.com/yl4579/StyleTTS2). Thanks to its Catalan-specific phoneme vocabulary and contextual embedding capabilities, it can serve as a phoneme encoder for any TTS architecture requiring phoneme-level features.
36
 
 
42
 
43
  ---
44
 
45
+ ## Intended Uses and Limitations
46
 
47
  ### Intended uses
48
 
 
58
 
59
  ---
60
 
61
+ ## How to Get Started with the Model
62
 
63
  Here is an example of how to use this model within the StyleTTS2 framework:
64
 
 
84
 
85
  ---
86
 
87
+ ## Training Details
88
 
89
  ### Training data
90
 
 
121
  ---
122
 
123
 
124
+ ### Evaluation
125
 
126
  The model has not been benchmarked via perplexity or extrinsic evaluation, but has been successfully integrated into TTS pipelines such as StyleTTS2, where it enables the synthesis of Catalan with regional accent variation.
127
 
 
132
  If this code contributes to your research, please cite the work:
133
 
134
  ```
135
+ @misc{zevallos2025plbertca,
136
+ title={PL-BERT-ca},
137
  author={Rodolfo Zevallos, Jose Giraldo and Carme Armentano-Oller},
138
  organization={Barcelona Supercomputing Center},
139
+ url={https://huggingface.co/langtech-veu/PL-BERT-ca},
140
  year={2025}
141
  }
142
  ```
143
 
 
144
 
145
+ ## Additional Information
146
 
 
 
147
 
148
+ ### Author
149
 
150
+ The [Language Technologies Laboratory](https://huggingface.co/BSC-LT) of the [Barcelona Supercomputing Center](https://www.bsc.es/) by [Rodolfo Zevallos](https://huggingface.co/rjzevallos).
151
 
152
+ ### Contact
153
+ For further information, please send an email to <langtech@bsc.es>.
154
 
155
+ ### Copyright
156
+ Copyright(c) 2025 by Language Technologies Laboratory, Barcelona Supercomputing Center.
157
 
158
+ ### License
159
 
160
+ [Apache-2.0](https://www.apache.org/licenses/LICENSE-2.0)
161
 
 
 
162
 
163
+ ### Funding
164
+ This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA.
165
 
 
166