Weyaxi commited on
Commit
6f70e1f
1 Parent(s): 5cccfb0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -48
README.md CHANGED
@@ -14,7 +14,6 @@ datasets:
14
  - openbookqa
15
  - mandyyyyii/scibench
16
  - derek-thomas/ScienceQA
17
- - wenhu/TheoremQA
18
  - TIGER-Lab/ScienceEval
19
  - jondurbin/airoboros-3.2
20
  - LDJnr/Capybara
@@ -155,50 +154,3 @@ resume_from_checkpoint: Einstein-v4-model/checkpoint-521
155
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
156
  It achieves the following results on the evaluation set:
157
  - Loss: 0.4902
158
-
159
- ## Model description
160
-
161
- More information needed
162
-
163
- ## Intended uses & limitations
164
-
165
- More information needed
166
-
167
- ## Training and evaluation data
168
-
169
- More information needed
170
-
171
- ## Training procedure
172
-
173
- ### Training hyperparameters
174
-
175
- The following hyperparameters were used during training:
176
- - learning_rate: 5e-06
177
- - train_batch_size: 1
178
- - eval_batch_size: 1
179
- - seed: 42
180
- - distributed_type: multi-GPU
181
- - num_devices: 8
182
- - gradient_accumulation_steps: 4
183
- - total_train_batch_size: 32
184
- - total_eval_batch_size: 8
185
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
186
- - lr_scheduler_type: cosine
187
- - lr_scheduler_warmup_steps: 10
188
- - num_epochs: 1.5
189
-
190
- ### Training results
191
-
192
- | Training Loss | Epoch | Step | Validation Loss |
193
- |:-------------:|:-----:|:----:|:---------------:|
194
- | 1.1358 | 0.0 | 1 | 1.2075 |
195
- | 0.5083 | 0.5 | 521 | 0.5161 |
196
- | 0.4759 | 1.0 | 1042 | 0.4902 |
197
-
198
-
199
- ### Framework versions
200
-
201
- - Transformers 4.38.0.dev0
202
- - Pytorch 2.1.2+cu118
203
- - Datasets 2.16.1
204
- - Tokenizers 0.15.0
 
14
  - openbookqa
15
  - mandyyyyii/scibench
16
  - derek-thomas/ScienceQA
 
17
  - TIGER-Lab/ScienceEval
18
  - jondurbin/airoboros-3.2
19
  - LDJnr/Capybara
 
154
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
155
  It achieves the following results on the evaluation set:
156
  - Loss: 0.4902