Hemantrao commited on
Commit
1f37104
·
verified ·
1 Parent(s): baf3ac8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -12
README.md CHANGED
@@ -6,13 +6,13 @@ tags:
6
  - code-switching
7
  - ASR
8
  - multilingual
9
- license: "mit"
10
  datasets:
11
  - common_voice
12
  metrics:
13
  - wer
14
  - cer
15
- base_model: "facebook/wav2vec2-large-xls-r-300m"
16
  library_name: transformers
17
  model-index:
18
  - name: wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experiment
@@ -31,8 +31,9 @@ model-index:
31
  value: 0.2400
32
  source:
33
  name: Internal Evaluation
 
34
 
35
-
36
  # Enhanced Multilingual Code-Switched Speech Recognition for Low-Resource Languages Using Transformer-Based Models and Dynamic Switching Algorithms
37
 
38
  ## Model description
@@ -56,7 +57,6 @@ The model was fine-tuned using the following parameters:
56
  - Learning Rate: 3e-4
57
  - Mask Time Probability: 0.05
58
 
59
- For detailed training logs and experimental tracking, please refer to the [experiment tracking platform](link_to_experiment_tracking_platform).
60
 
61
  ## Training dataset
62
  The model was trained on the Common Voice dataset, which includes diverse speech samples in both Hindi and Marathi. The dataset was augmented with synthetically generated code-switched speech to improve the model's robustness in handling code-switching scenarios.
@@ -66,11 +66,3 @@ The model achieved the following performance metrics on the test set:
66
  - Word Error Rate (WER): 0.2800
67
  - Character Error Rate (CER): 0.2400
68
 
69
- ## Ethical considerations and biases
70
- The model has potential biases due to the limited representation of accents, dialects, and socio-linguistic variations in the training data. Users should be cautious about deploying this model in critical applications without further evaluation and customization.
71
-
72
- ## CO2 Emissions
73
- For information about the CO2 impact of training this model, please refer to our [guide on tracking and reporting CO2 emissions](link_to_CO2_impact_guide).
74
-
75
- ## Paper
76
- If applicable, you can include a link to a paper describing this model. For instance, if your model is described in an arXiv paper, you can add the arXiv ID here:
 
6
  - code-switching
7
  - ASR
8
  - multilingual
9
+ license: mit
10
  datasets:
11
  - common_voice
12
  metrics:
13
  - wer
14
  - cer
15
+ base_model: facebook/wav2vec2-large-xls-r-300m
16
  library_name: transformers
17
  model-index:
18
  - name: wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experiment
 
31
  value: 0.2400
32
  source:
33
  name: Internal Evaluation
34
+ url: "https://huggingface.co/Hemantrao/wav2vec2-large-xls-r-300m-hindi_marathi-code-switching-experimentx1/"
35
 
36
+ ---
37
  # Enhanced Multilingual Code-Switched Speech Recognition for Low-Resource Languages Using Transformer-Based Models and Dynamic Switching Algorithms
38
 
39
  ## Model description
 
57
  - Learning Rate: 3e-4
58
  - Mask Time Probability: 0.05
59
 
 
60
 
61
  ## Training dataset
62
  The model was trained on the Common Voice dataset, which includes diverse speech samples in both Hindi and Marathi. The dataset was augmented with synthetically generated code-switched speech to improve the model's robustness in handling code-switching scenarios.
 
66
  - Word Error Rate (WER): 0.2800
67
  - Character Error Rate (CER): 0.2400
68