Update README.md
Browse files
README.md
CHANGED
|
@@ -6,13 +6,13 @@ tags:
|
|
| 6 |
- code-switching
|
| 7 |
- ASR
|
| 8 |
- multilingual
|
| 9 |
-
license:
|
| 10 |
datasets:
|
| 11 |
- common_voice
|
| 12 |
metrics:
|
| 13 |
- wer
|
| 14 |
- cer
|
| 15 |
-
base_model:
|
| 16 |
library_name: transformers
|
| 17 |
model-index:
|
| 18 |
- name: wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experiment
|
|
@@ -31,8 +31,9 @@ model-index:
|
|
| 31 |
value: 0.2400
|
| 32 |
source:
|
| 33 |
name: Internal Evaluation
|
|
|
|
| 34 |
|
| 35 |
-
|
| 36 |
# Enhanced Multilingual Code-Switched Speech Recognition for Low-Resource Languages Using Transformer-Based Models and Dynamic Switching Algorithms
|
| 37 |
|
| 38 |
## Model description
|
|
@@ -56,7 +57,6 @@ The model was fine-tuned using the following parameters:
|
|
| 56 |
- Learning Rate: 3e-4
|
| 57 |
- Mask Time Probability: 0.05
|
| 58 |
|
| 59 |
-
For detailed training logs and experimental tracking, please refer to the [experiment tracking platform](link_to_experiment_tracking_platform).
|
| 60 |
|
| 61 |
## Training dataset
|
| 62 |
The model was trained on the Common Voice dataset, which includes diverse speech samples in both Hindi and Marathi. The dataset was augmented with synthetically generated code-switched speech to improve the model's robustness in handling code-switching scenarios.
|
|
@@ -66,11 +66,3 @@ The model achieved the following performance metrics on the test set:
|
|
| 66 |
- Word Error Rate (WER): 0.2800
|
| 67 |
- Character Error Rate (CER): 0.2400
|
| 68 |
|
| 69 |
-
## Ethical considerations and biases
|
| 70 |
-
The model has potential biases due to the limited representation of accents, dialects, and socio-linguistic variations in the training data. Users should be cautious about deploying this model in critical applications without further evaluation and customization.
|
| 71 |
-
|
| 72 |
-
## CO2 Emissions
|
| 73 |
-
For information about the CO2 impact of training this model, please refer to our [guide on tracking and reporting CO2 emissions](link_to_CO2_impact_guide).
|
| 74 |
-
|
| 75 |
-
## Paper
|
| 76 |
-
If applicable, you can include a link to a paper describing this model. For instance, if your model is described in an arXiv paper, you can add the arXiv ID here:
|
|
|
|
| 6 |
- code-switching
|
| 7 |
- ASR
|
| 8 |
- multilingual
|
| 9 |
+
license: mit
|
| 10 |
datasets:
|
| 11 |
- common_voice
|
| 12 |
metrics:
|
| 13 |
- wer
|
| 14 |
- cer
|
| 15 |
+
base_model: facebook/wav2vec2-large-xls-r-300m
|
| 16 |
library_name: transformers
|
| 17 |
model-index:
|
| 18 |
- name: wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experiment
|
|
|
|
| 31 |
value: 0.2400
|
| 32 |
source:
|
| 33 |
name: Internal Evaluation
|
| 34 |
+
url: "https://huggingface.co/Hemantrao/wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experimentx1/"
|
| 35 |
|
| 36 |
+
---
|
| 37 |
# Enhanced Multilingual Code-Switched Speech Recognition for Low-Resource Languages Using Transformer-Based Models and Dynamic Switching Algorithms
|
| 38 |
|
| 39 |
## Model description
|
|
|
|
| 57 |
- Learning Rate: 3e-4
|
| 58 |
- Mask Time Probability: 0.05
|
| 59 |
|
|
|
|
| 60 |
|
| 61 |
## Training dataset
|
| 62 |
The model was trained on the Common Voice dataset, which includes diverse speech samples in both Hindi and Marathi. The dataset was augmented with synthetically generated code-switched speech to improve the model's robustness in handling code-switching scenarios.
|
|
|
|
| 66 |
- Word Error Rate (WER): 0.2800
|
| 67 |
- Character Error Rate (CER): 0.2400
|
| 68 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|