UNA-TheBeagle-7b-v1 / README.md
fblgit's picture
Update README.md
6d0011d
|
raw
history blame
2.15 kB
---
license: cc-by-nc-nd-4.0
tags:
- generated_from_trainer
model-index:
- name: UNA-TheBeagle-7b-v1
results: []
datasets:
- jondurbin/bagel-v0.3
library_name: transformers
---
# UNA-TheBeagle-7b-v1
TheBeagle, a model of 7B parameters trained on The Bagel dataset. DPO & UNA applied over a set of curated DPO Pairs.
The dataset was generated using the original bagel code, including the decontamination step.
As base model, we used the latest Intel's neural-chat model.
It performs very good in many tasks, but its always better that you play with it by yourself.
## Evaluations
Ran with VLLM so expect them to dont be exactly as the one's shown in the board, but not too far :)
```
vllm (pretrained=fblgit/UNA-TheBeagle-7b-v1,dtype=auto,tensor_parallel_size=1,gpu_memory_utilization=0.8,data_parallel_size=8,trust_remote_code=True), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 32
| Tasks |Version| Filter |n-shot| Metric |Value | |Stderr|
|--------------|-------|----------|-----:|-----------|-----:|---|-----:|
|arc_challenge |Yaml |none | 25|acc |0.7090|± |0.0133|
| | |none | 25|acc_norm |0.7329|± |0.0129|
|gsm8k |Yaml |get-answer| 5|exact_match|0.7210|± |0.0124|
|hellaswag |Yaml |none | 10|acc |0.7202|± |0.0045|
| | |none | 10|acc_norm |0.8792|± |0.0033|
|truthfulqa_mc2|Yaml |none | 0|acc |0.7062|± |0.0151|
|winogrande |Yaml |none | 5|acc |0.8366|± |0.0104|
```
## UNA Details
For this release, we only applied UNA thru the perceptrons. It was done at a 3.5e-7 speed, and the training loop code is also the original one of the bagel and transformers-4.35.2-UNA
## Prompt
Im not entirely sure of it, as we used the vanilla version of the bagel training code. But a good model should be able to generalize with different prompt formats, so feel free to give it a shot.
## Citations
Remember if you use UNA's models, cite it in your model card.
## Limitations
Not for commercial use, and only for academic & research purposes.