k-r-l commited on
Commit
cf50b1f
1 Parent(s): 6985bbc

Model save

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -1,22 +1,22 @@
1
  ---
2
- base_model: bigscience/bloomz-560m
3
  library_name: peft
4
- license: bigscience-bloom-rail-1.0
5
  tags:
6
  - trl
7
  - orpo
8
  - generated_from_trainer
9
  model-index:
10
- - name: bloomz-560m-mergedORPO
11
  results: []
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- # bloomz-560m-mergedORPO
18
 
19
- This model is a fine-tuned version of [bigscience/bloomz-560m](https://huggingface.co/bigscience/bloomz-560m) on an unknown dataset.
20
 
21
  ## Model description
22
 
@@ -36,10 +36,10 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0001
39
- - train_batch_size: 8
40
  - eval_batch_size: 1
41
  - seed: 42
42
- - gradient_accumulation_steps: 4
43
  - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
 
1
  ---
2
+ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
3
  library_name: peft
4
+ license: llama3.1
5
  tags:
6
  - trl
7
  - orpo
8
  - generated_from_trainer
9
  model-index:
10
+ - name: Meta-Llama-3.1-8B-Instruct-mergedORPO
11
  results: []
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ # Meta-Llama-3.1-8B-Instruct-mergedORPO
18
 
19
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0001
39
+ - train_batch_size: 4
40
  - eval_batch_size: 1
41
  - seed: 42
42
+ - gradient_accumulation_steps: 8
43
  - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear