pszemraj commited on
Commit
59876ca
1 Parent(s): 0b34b03

End of training

Browse files
Files changed (5) hide show
  1. README.md +5 -3
  2. all_results.json +15 -0
  3. eval_results.json +9 -0
  4. train_results.json +10 -0
  5. trainer_state.json +0 -0
README.md CHANGED
@@ -1,5 +1,7 @@
1
  ---
2
  library_name: transformers
 
 
3
  license: apache-2.0
4
  base_model: pszemraj/tFINE-850m-24x24-v0.5-instruct-L1
5
  tags:
@@ -14,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # tFINE-850m-24x24-v0.5-instruct-L1-infinity-instruct-7m-T2T_en-1024-v2
16
 
17
- This model is a fine-tuned version of [pszemraj/tFINE-850m-24x24-v0.5-instruct-L1](https://huggingface.co/pszemraj/tFINE-850m-24x24-v0.5-instruct-L1) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 1.2595
20
- - Num Input Tokens Seen: 568955808
21
 
22
  ## Model description
23
 
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - en
5
  license: apache-2.0
6
  base_model: pszemraj/tFINE-850m-24x24-v0.5-instruct-L1
7
  tags:
 
16
 
17
  # tFINE-850m-24x24-v0.5-instruct-L1-infinity-instruct-7m-T2T_en-1024-v2
18
 
19
+ This model is a fine-tuned version of [pszemraj/tFINE-850m-24x24-v0.5-instruct-L1](https://huggingface.co/pszemraj/tFINE-850m-24x24-v0.5-instruct-L1) on the pszemraj/infinity-instruct-7m-T2T_en dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.2542
22
+ - Num Input Tokens Seen: 750938410
23
 
24
  ## Model description
25
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 1.254249095916748,
4
+ "eval_runtime": 14.3091,
5
+ "eval_samples": 150,
6
+ "eval_samples_per_second": 10.483,
7
+ "eval_steps_per_second": 0.699,
8
+ "num_input_tokens_seen": 750938410,
9
+ "total_flos": 3.6248418467253043e+18,
10
+ "train_loss": 1.2696220230851407,
11
+ "train_runtime": 79988.0702,
12
+ "train_samples": 1013227,
13
+ "train_samples_per_second": 12.667,
14
+ "train_steps_per_second": 0.099
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 1.254249095916748,
4
+ "eval_runtime": 14.3091,
5
+ "eval_samples": 150,
6
+ "eval_samples_per_second": 10.483,
7
+ "eval_steps_per_second": 0.699,
8
+ "num_input_tokens_seen": 750938410
9
+ }
train_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "num_input_tokens_seen": 750938410,
4
+ "total_flos": 3.6248418467253043e+18,
5
+ "train_loss": 1.2696220230851407,
6
+ "train_runtime": 79988.0702,
7
+ "train_samples": 1013227,
8
+ "train_samples_per_second": 12.667,
9
+ "train_steps_per_second": 0.099
10
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff