ankitkumar-patel-179 commited on
Commit
6ac2223
1 Parent(s): e78999c

Model save

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: other
3
- base_model: huggyllama/llama-7b
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # llama_instr_tune_fact_checking
15
 
16
- This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on the None dataset.
17
 
18
  ## Model description
19
 
@@ -33,11 +33,11 @@ More information needed
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0002
36
- - train_batch_size: 16
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - gradient_accumulation_steps: 4
40
- - total_train_batch_size: 64
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - lr_scheduler_warmup_ratio: 0.03
@@ -51,5 +51,5 @@ The following hyperparameters were used during training:
51
 
52
  - Transformers 4.34.1
53
  - Pytorch 2.1.0+cu118
54
- - Datasets 2.14.5
55
  - Tokenizers 0.14.1
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: kaist-ai/CoT-T5-11B
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # llama_instr_tune_fact_checking
15
 
16
+ This model is a fine-tuned version of [kaist-ai/CoT-T5-11B](https://huggingface.co/kaist-ai/CoT-T5-11B) on the None dataset.
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0002
36
+ - train_batch_size: 32
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - gradient_accumulation_steps: 4
40
+ - total_train_batch_size: 128
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - lr_scheduler_warmup_ratio: 0.03
 
51
 
52
  - Transformers 4.34.1
53
  - Pytorch 2.1.0+cu118
54
+ - Datasets 2.14.6
55
  - Tokenizers 0.14.1