|
--- |
|
license: llama2 |
|
datasets: |
|
- totally-not-an-llm/EverythingLM-data-V3 |
|
--- |
|
|
|
# EverythingLM-13b-V3-16k |
|
|
|
Introducing EverythingLM, a llama-2 based, general-purpose 13b model with 16k context thanks to LlongMa. The model is trained on the EverythingLM-V3 dataset, more info can be found on the dataset page. |
|
|
|
The model is completely uncensored. |
|
|
|
Despite being "uncensored", the base model might be resistant; you might have to prompt-engineer certain prompts. |
|
|
|
### Quants (Thanks TheBloke!): |
|
https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-GGUF |
|
|
|
https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-GPTQ |
|
|
|
https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-AWQ |
|
|
|
### Notable features: |
|
- Automatically triggered CoT reasoning. |
|
- Verbose and detailed replies. |
|
- Creative stories. |
|
- Good prompt understanding. |
|
|
|
### Differences from V2: |
|
- Much more uncensored. |
|
- Actual roleplaying ability now! |
|
- General all around improvements thanks to the new dataset. Check out the dataset for more info. |
|
|
|
### Prompt format (Alpaca-chat): |
|
|
|
``` |
|
USER: <prompt> |
|
ASSISTANT: |
|
``` |
|
|
|
### Future plans: |
|
- Highest priority right now is V3.1 with more optimized training and iterative dataset improvements based on testing. |
|
|
|
### Note: |
|
Through testing V2, I realized some alignment data had leaked in, causing the model to be less cooperative then intended. This model should do much better due to stricter filetering. |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_totally-not-an-llm__EverythingLM-13b-V3-16k) |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | 44.82 | |
|
| ARC (25-shot) | 58.19 | |
|
| HellaSwag (10-shot) | 80.12 | |
|
| MMLU (5-shot) | 50.48 | |
|
| TruthfulQA (0-shot) | 45.18 | |
|
| Winogrande (5-shot) | 70.72 | |
|
| GSM8K (5-shot) | 1.97 | |
|
| DROP (3-shot) | 7.06 | |
|
|