daisyyedda
commited on
Commit
•
f84eff1
1
Parent(s):
2267b4e
End of training
Browse files- README.md +3 -3
- config.json +1 -1
- generation_config.json +1 -1
- model-00001-of-00002.safetensors +1 -1
- model-00002-of-00002.safetensors +1 -1
- runs/Jul12_03-05-31_watvis-whale/events.out.tfevents.1720753555.watvis-whale.343639.0 +3 -0
- runs/Jul12_03-09-27_watvis-whale/events.out.tfevents.1720753789.watvis-whale.344117.0 +3 -0
- training_args.bin +1 -1
README.md
CHANGED
@@ -33,14 +33,14 @@ More information needed
|
|
33 |
|
34 |
The following hyperparameters were used during training:
|
35 |
- learning_rate: 1e-05
|
36 |
-
- train_batch_size:
|
37 |
- eval_batch_size: 8
|
38 |
- seed: 42
|
39 |
- distributed_type: multi-GPU
|
40 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
41 |
- lr_scheduler_type: linear
|
42 |
- lr_scheduler_warmup_steps: 250
|
43 |
-
- num_epochs:
|
44 |
|
45 |
### Training results
|
46 |
|
@@ -48,7 +48,7 @@ The following hyperparameters were used during training:
|
|
48 |
|
49 |
### Framework versions
|
50 |
|
51 |
-
- Transformers 4.42.
|
52 |
- Pytorch 2.3.1+cu118
|
53 |
- Datasets 2.20.0
|
54 |
- Tokenizers 0.19.1
|
|
|
33 |
|
34 |
The following hyperparameters were used during training:
|
35 |
- learning_rate: 1e-05
|
36 |
+
- train_batch_size: 32
|
37 |
- eval_batch_size: 8
|
38 |
- seed: 42
|
39 |
- distributed_type: multi-GPU
|
40 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
41 |
- lr_scheduler_type: linear
|
42 |
- lr_scheduler_warmup_steps: 250
|
43 |
+
- num_epochs: 50
|
44 |
|
45 |
### Training results
|
46 |
|
|
|
48 |
|
49 |
### Framework versions
|
50 |
|
51 |
+
- Transformers 4.42.4
|
52 |
- Pytorch 2.3.1+cu118
|
53 |
- Datasets 2.20.0
|
54 |
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -45,7 +45,7 @@
|
|
45 |
"scale_embedding": false,
|
46 |
"suppress_tokens": [],
|
47 |
"torch_dtype": "float32",
|
48 |
-
"transformers_version": "4.42.
|
49 |
"use_cache": true,
|
50 |
"use_weighted_layer_sum": false,
|
51 |
"vocab_size": 51865
|
|
|
45 |
"scale_embedding": false,
|
46 |
"suppress_tokens": [],
|
47 |
"torch_dtype": "float32",
|
48 |
+
"transformers_version": "4.42.4",
|
49 |
"use_cache": true,
|
50 |
"use_weighted_layer_sum": false,
|
51 |
"vocab_size": 51865
|
generation_config.json
CHANGED
@@ -312,5 +312,5 @@
|
|
312 |
"transcribe": 50359,
|
313 |
"translate": 50358
|
314 |
},
|
315 |
-
"transformers_version": "4.42.
|
316 |
}
|
|
|
312 |
"transcribe": 50359,
|
313 |
"translate": 50358
|
314 |
},
|
315 |
+
"transformers_version": "4.42.4"
|
316 |
}
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4992706480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1552d8d5cec372fa0327455714e3bc9dcaecf220979606781a6a41fe252b2d79
|
3 |
size 4992706480
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1446212088
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:57a8ffefb0eb3708b3c05d9899c20957efc40d67a73c75c2a40e1850821f3fc0
|
3 |
size 1446212088
|
runs/Jul12_03-05-31_watvis-whale/events.out.tfevents.1720753555.watvis-whale.343639.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7071a77ebcafe2a250cdf85f9db4bf0ffe7f28658af831a01b5bf510859bb40c
|
3 |
+
size 22684
|
runs/Jul12_03-09-27_watvis-whale/events.out.tfevents.1720753789.watvis-whale.344117.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3138cb81177e5da16fb975198271078fba2937979148d852009d769d92c0eea5
|
3 |
+
size 7337
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6968
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d7eb94d19dc202a75896815a71938091847daa17e2ec24bfebdd1100d55cc8c
|
3 |
size 6968
|