Text Generation
Transformers
Safetensors
mistral
conversational
text-generation-inference
Inference Endpoints
siddartha-abacus commited on
Commit
4648012
1 Parent(s): 1dbbf47

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -13,4 +13,12 @@ Instruction tuned with the following parameters:
13
  - LORA, Rank 8, Alpha 16, Dropout 0.05, all modules (QKV and MLP)
14
  - 3 epochs
15
  - Micro Batch Size 32 over 4xH100, gradient accumulation steps = 1
16
- - AdamW with learning rate 5e-5
 
 
 
 
 
 
 
 
 
13
  - LORA, Rank 8, Alpha 16, Dropout 0.05, all modules (QKV and MLP)
14
  - 3 epochs
15
  - Micro Batch Size 32 over 4xH100, gradient accumulation steps = 1
16
+ - AdamW with learning rate 5e-5
17
+
18
+ ### Evaluation Results
19
+
20
+ | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
21
+ | --- | --- | --- | --- | --- | --- | --- |
22
+ | 67.33 | 59.64 | 81.82 | 61.69 | 53.23 | 78.45 | 69.14 |
23
+
24
+ For comparison the GSM8K score for the original `metamath/MetaMath-Mistral-7B` was 68.84 and average score was 65.78.