Commit
·
17646a7
1
Parent(s):
e655e86
add model
Browse files- config.json +3 -1
- tf_model.h5 +3 -0
config.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
|
|
|
| 2 |
"activation_dropout": 0.1,
|
| 3 |
"apply_spec_augment": true,
|
| 4 |
"architectures": [
|
|
@@ -42,6 +43,7 @@
|
|
| 42 |
"feat_extract_activation": "gelu",
|
| 43 |
"feat_extract_dropout": 0.0,
|
| 44 |
"feat_extract_norm": "layer",
|
|
|
|
| 45 |
"final_dropout": 0.1,
|
| 46 |
"gradient_checkpointing": false,
|
| 47 |
"hidden_act": "gelu",
|
|
@@ -63,6 +65,6 @@
|
|
| 63 |
"num_feat_extract_layers": 7,
|
| 64 |
"num_hidden_layers": 24,
|
| 65 |
"pad_token_id": 0,
|
| 66 |
-
"transformers_version": "4.
|
| 67 |
"vocab_size": 32
|
| 68 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "facebook/hubert-large-ls960-ft",
|
| 3 |
"activation_dropout": 0.1,
|
| 4 |
"apply_spec_augment": true,
|
| 5 |
"architectures": [
|
|
|
|
| 43 |
"feat_extract_activation": "gelu",
|
| 44 |
"feat_extract_dropout": 0.0,
|
| 45 |
"feat_extract_norm": "layer",
|
| 46 |
+
"feat_proj_dropout": 0.1,
|
| 47 |
"final_dropout": 0.1,
|
| 48 |
"gradient_checkpointing": false,
|
| 49 |
"hidden_act": "gelu",
|
|
|
|
| 65 |
"num_feat_extract_layers": 7,
|
| 66 |
"num_hidden_layers": 24,
|
| 67 |
"pad_token_id": 0,
|
| 68 |
+
"transformers_version": "4.10.0.dev0",
|
| 69 |
"vocab_size": 32
|
| 70 |
}
|
tf_model.h5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1fc55856d3faa99c91b1cb91c6b9717d0d967ea09a676b2b20cff7039169a750
|
| 3 |
+
size 1262396792
|