Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,101 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
---
|
4 |
+
|
5 |
+
# 😼 CatPPT
|
6 |
+
|
7 |
+
Introducing "CatPPT" - the purrfect alternative to that other big cat in town, known for keeping all the secrets to itself! Our feline friend here is created through merging openchat and neuralchat models using Gradient SLERP method (resulting in [rishiraj/CatPPT-base](https://huggingface.co/rishiraj/CatPPT-base)) and then finetuned on no_robots dataset for chat.
|
8 |
+
|
9 |
+
This is the top-performing 7B model on the leaderboard, that's free from any whiff of evaluation data contamination.
|
10 |
+
|
11 |
+
![](https://raw.githubusercontent.com/rishiraj/rishiraj.github.io/main/assets/spider%402x.png)
|
12 |
+
|
13 |
+
## Model date
|
14 |
+
|
15 |
+
rishiraj/CatPPT was trained between 15th and 17th December, 2023.
|
16 |
+
|
17 |
+
## Evaluation
|
18 |
+
|
19 |
+
It achieves the following results on the [Open_LLM_Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). At the time of release, CatPPT is the highest ranked 7B chat model on the leaderboard, that's **free from evaluation data contamination**.
|
20 |
+
|
21 |
+
|Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
|
22 |
+
|------------------------------------|-------|-----|---------|-----|----------|----------|-----|
|
23 |
+
|**rishiraj/CatPPT** |**72.32** |**68.09**|**86.69** |**65.16**|**61.55** |**81.61** |**70.81**|
|
24 |
+
|Intel/neural-chat-7b-v3-3 |69.83 |66.89|85.26 |63.07|63.01 |79.64 |61.11|
|
25 |
+
|openchat/openchat-3.5-1210 |68.89 |64.93|84.92 |64.62|52.15 |80.74 |65.96|
|
26 |
+
|meta-math/MetaMath-Mistral-7B |65.78 |60.67|82.58 |61.95|44.89 |75.77 |68.84|
|
27 |
+
|Deci/DeciLM-7B-instruct |63.19 |61.01|82.37 |60.24|49.75 |79.72 |46.02|
|
28 |
+
|mistralai/Mistral-7B-Instruct-v0.2 |65.71 |63.14|84.88 |60.78|68.26 |77.19 |40.03|
|
29 |
+
|mistralai/Mixtral-8x7B-Instruct-v0.1|72.62 |70.22|87.63 |71.16|64.58 |81.37 |60.73|
|
30 |
+
|meta-llama/Llama-2-70b-hf |67.87 |67.32|87.33 |69.83|44.92 |83.74 |54.06|
|
31 |
+
|tiiuae/falcon-180B |67.85 |69.45|88.86 |70.5 |45.47 |86.9 |45.94|
|
32 |
+
|
33 |
+
## Inference procedure
|
34 |
+
|
35 |
+
Here's how you can run the model using the pipeline() function from 🤗 Transformers:
|
36 |
+
|
37 |
+
```
|
38 |
+
import torch
|
39 |
+
from transformers import pipeline
|
40 |
+
|
41 |
+
pipe = pipeline("text-generation", model="rishiraj/CatPPT", torch_dtype=torch.bfloat16, device_map="auto")
|
42 |
+
|
43 |
+
# We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
|
44 |
+
messages = [
|
45 |
+
{
|
46 |
+
"role": "system",
|
47 |
+
"content": "You are a friendly chatbot who always responds in the style of a pirate"
|
48 |
+
},
|
49 |
+
{
|
50 |
+
"role": "user",
|
51 |
+
"content": "How many helicopters can a human eat in one sitting?"
|
52 |
+
}
|
53 |
+
]
|
54 |
+
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
55 |
+
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
56 |
+
print(outputs[0]["generated_text"])
|
57 |
+
```
|
58 |
+
|
59 |
+
## Training procedure
|
60 |
+
|
61 |
+
### Training hyperparameters
|
62 |
+
|
63 |
+
The following hyperparameters were used during training:
|
64 |
+
- learning_rate: 2e-05
|
65 |
+
- train_batch_size: 4
|
66 |
+
- eval_batch_size: 8
|
67 |
+
- seed: 42
|
68 |
+
- distributed_type: multi-GPU
|
69 |
+
- gradient_accumulation_steps: 128
|
70 |
+
- total_train_batch_size: 512
|
71 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
72 |
+
- lr_scheduler_type: cosine
|
73 |
+
- num_epochs: 1
|
74 |
+
|
75 |
+
### Training results
|
76 |
+
|
77 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
78 |
+
|:-------------:|:-----:|:----:|:---------------:|
|
79 |
+
| 1.9947 | 0.16 | 3 | 2.0093 |
|
80 |
+
|
81 |
+
|
82 |
+
### Framework versions
|
83 |
+
|
84 |
+
- Transformers 4.36.1
|
85 |
+
- Pytorch 2.1.2+cu121
|
86 |
+
- Datasets 2.14.6
|
87 |
+
- Tokenizers 0.15.0
|
88 |
+
- PEFT 0.6.1
|
89 |
+
|
90 |
+
## Citation Information
|
91 |
+
|
92 |
+
```
|
93 |
+
@misc{rishiraj2023catppt,
|
94 |
+
author = {Rishiraj Acharya},
|
95 |
+
title = {CatPPT},
|
96 |
+
year = {2023},
|
97 |
+
publisher = {Hugging Face},
|
98 |
+
journal = {Hugging Face repository},
|
99 |
+
howpublished = {\url{https://huggingface.co/rishiraj/CatPPT}}
|
100 |
+
}
|
101 |
+
```
|