pszemraj commited on
Commit
1cc4203
1 Parent(s): 59876ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -34
README.md CHANGED
@@ -5,33 +5,21 @@ language:
5
  license: apache-2.0
6
  base_model: pszemraj/tFINE-850m-24x24-v0.5-instruct-L1
7
  tags:
8
- - generated_from_trainer
9
- model-index:
10
- - name: tFINE-850m-24x24-v0.5-instruct-L1-infinity-instruct-7m-T2T_en-1024-v2
11
- results: []
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
 
16
 
17
- # tFINE-850m-24x24-v0.5-instruct-L1-infinity-instruct-7m-T2T_en-1024-v2
18
 
19
- This model is a fine-tuned version of [pszemraj/tFINE-850m-24x24-v0.5-instruct-L1](https://huggingface.co/pszemraj/tFINE-850m-24x24-v0.5-instruct-L1) on the pszemraj/infinity-instruct-7m-T2T_en dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 1.2542
22
  - Num Input Tokens Seen: 750938410
23
 
24
- ## Model description
25
-
26
- More information needed
27
-
28
- ## Intended uses & limitations
29
-
30
- More information needed
31
-
32
- ## Training and evaluation data
33
-
34
- More information needed
35
 
36
  ## Training procedure
37
 
@@ -50,18 +38,3 @@ No additional optimizer arguments
50
  - lr_scheduler_warmup_ratio: 0.03
51
  - num_epochs: 1.0
52
 
53
- ### Training results
54
-
55
- | Training Loss | Epoch | Step | Validation Loss | Input Tokens Seen |
56
- |:-------------:|:------:|:----:|:---------------:|:-----------------:|
57
- | 1.32 | 0.2527 | 2000 | 1.3214 | 189801824 |
58
- | 1.2614 | 0.5053 | 4000 | 1.2815 | 379241088 |
59
- | 1.2367 | 0.7580 | 6000 | 1.2595 | 568955808 |
60
-
61
-
62
- ### Framework versions
63
-
64
- - Transformers 4.46.0.dev0
65
- - Pytorch 2.5.1+cu124
66
- - Datasets 3.1.0
67
- - Tokenizers 0.20.1
 
5
  license: apache-2.0
6
  base_model: pszemraj/tFINE-850m-24x24-v0.5-instruct-L1
7
  tags:
8
+ - instruct
9
+ datasets:
10
+ - pszemraj/infinity-instruct-7m-T2T_en
11
+ pipeline_tag: text2text-generation
12
  ---
13
 
14
+ # tFINE-850m-24x24-instruct-L2
15
+
16
+ This model is a fine-tuned version of [pszemraj/tFINE-850m-24x24-v0.5-instruct-L1](https://huggingface.co/pszemraj/tFINE-850m-24x24-v0.5-instruct-L1) on the pszemraj/infinity-instruct-7m-T2T_en dataset (config `deduped-L2`).
17
 
 
18
 
 
19
  It achieves the following results on the evaluation set:
20
  - Loss: 1.2542
21
  - Num Input Tokens Seen: 750938410
22
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ## Training procedure
25
 
 
38
  - lr_scheduler_warmup_ratio: 0.03
39
  - num_epochs: 1.0
40