m3hrdadfi commited on
Commit
8656b88
1 Parent(s): 5696c96
config.json CHANGED
@@ -47,7 +47,7 @@
47
  "feat_quantizer_dropout": 0.0,
48
  "final_dropout": 0.0,
49
  "freeze_feat_extract_train": true,
50
- "gradient_checkpointing": true,
51
  "hidden_act": "gelu",
52
  "hidden_dropout": 0.1,
53
  "hidden_size": 768,
 
47
  "feat_quantizer_dropout": 0.0,
48
  "final_dropout": 0.0,
49
  "freeze_feat_extract_train": true,
50
+ "gradient_checkpointing": false,
51
  "hidden_act": "gelu",
52
  "hidden_dropout": 0.1,
53
  "hidden_size": 768,
events.out.tfevents.1626217020.t1v-n-278acf21-w-0.60949.3.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9be44b126823d678b51dec4a8628768f97c45890c1cb22d6a0697054e87c0c5c
3
+ size 662195
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a9dceff9ed264966ac910f740b22fdce31b824bd84cdd17310c5565ea88c6d
3
+ size 190539834