patrickvonplaten
commited on
Commit
•
17646a7
1
Parent(s):
e655e86
add model
Browse files- config.json +3 -1
- tf_model.h5 +3 -0
config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"activation_dropout": 0.1,
|
3 |
"apply_spec_augment": true,
|
4 |
"architectures": [
|
@@ -42,6 +43,7 @@
|
|
42 |
"feat_extract_activation": "gelu",
|
43 |
"feat_extract_dropout": 0.0,
|
44 |
"feat_extract_norm": "layer",
|
|
|
45 |
"final_dropout": 0.1,
|
46 |
"gradient_checkpointing": false,
|
47 |
"hidden_act": "gelu",
|
@@ -63,6 +65,6 @@
|
|
63 |
"num_feat_extract_layers": 7,
|
64 |
"num_hidden_layers": 24,
|
65 |
"pad_token_id": 0,
|
66 |
-
"transformers_version": "4.
|
67 |
"vocab_size": 32
|
68 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "facebook/hubert-large-ls960-ft",
|
3 |
"activation_dropout": 0.1,
|
4 |
"apply_spec_augment": true,
|
5 |
"architectures": [
|
|
|
43 |
"feat_extract_activation": "gelu",
|
44 |
"feat_extract_dropout": 0.0,
|
45 |
"feat_extract_norm": "layer",
|
46 |
+
"feat_proj_dropout": 0.1,
|
47 |
"final_dropout": 0.1,
|
48 |
"gradient_checkpointing": false,
|
49 |
"hidden_act": "gelu",
|
|
|
65 |
"num_feat_extract_layers": 7,
|
66 |
"num_hidden_layers": 24,
|
67 |
"pad_token_id": 0,
|
68 |
+
"transformers_version": "4.10.0.dev0",
|
69 |
"vocab_size": 32
|
70 |
}
|
tf_model.h5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1fc55856d3faa99c91b1cb91c6b9717d0d967ea09a676b2b20cff7039169a750
|
3 |
+
size 1262396792
|