rishiraj commited on
Commit
7b04169
1 Parent(s): 62a9e30

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -0
README.md CHANGED
@@ -1,3 +1,101 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ # 😼 CatPPT
6
+
7
+ Introducing "CatPPT" - the purrfect alternative to that other big cat in town, known for keeping all the secrets to itself! Our feline friend here is created through merging openchat and neuralchat models using Gradient SLERP method (resulting in [rishiraj/CatPPT-base](https://huggingface.co/rishiraj/CatPPT-base)) and then finetuned on no_robots dataset for chat.
8
+
9
+ This is the top-performing 7B model on the leaderboard, that's free from any whiff of evaluation data contamination.
10
+
11
+ ![](https://raw.githubusercontent.com/rishiraj/rishiraj.github.io/main/assets/spider%402x.png)
12
+
13
+ ## Model date
14
+
15
+ rishiraj/CatPPT was trained between 15th and 17th December, 2023.
16
+
17
+ ## Evaluation
18
+
19
+ It achieves the following results on the [Open_LLM_Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). At the time of release, CatPPT is the highest ranked 7B chat model on the leaderboard, that's **free from evaluation data contamination**.
20
+
21
+ |Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
22
+ |------------------------------------|-------|-----|---------|-----|----------|----------|-----|
23
+ |**rishiraj/CatPPT** |**72.32** |**68.09**|**86.69** |**65.16**|**61.55** |**81.61** |**70.81**|
24
+ |Intel/neural-chat-7b-v3-3 |69.83 |66.89|85.26 |63.07|63.01 |79.64 |61.11|
25
+ |openchat/openchat-3.5-1210 |68.89 |64.93|84.92 |64.62|52.15 |80.74 |65.96|
26
+ |meta-math/MetaMath-Mistral-7B |65.78 |60.67|82.58 |61.95|44.89 |75.77 |68.84|
27
+ |Deci/DeciLM-7B-instruct |63.19 |61.01|82.37 |60.24|49.75 |79.72 |46.02|
28
+ |mistralai/Mistral-7B-Instruct-v0.2 |65.71 |63.14|84.88 |60.78|68.26 |77.19 |40.03|
29
+ |mistralai/Mixtral-8x7B-Instruct-v0.1|72.62 |70.22|87.63 |71.16|64.58 |81.37 |60.73|
30
+ |meta-llama/Llama-2-70b-hf |67.87 |67.32|87.33 |69.83|44.92 |83.74 |54.06|
31
+ |tiiuae/falcon-180B |67.85 |69.45|88.86 |70.5 |45.47 |86.9 |45.94|
32
+
33
+ ## Inference procedure
34
+
35
+ Here's how you can run the model using the pipeline() function from 🤗 Transformers:
36
+
37
+ ```
38
+ import torch
39
+ from transformers import pipeline
40
+
41
+ pipe = pipeline("text-generation", model="rishiraj/CatPPT", torch_dtype=torch.bfloat16, device_map="auto")
42
+
43
+ # We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
44
+ messages = [
45
+ {
46
+ "role": "system",
47
+ "content": "You are a friendly chatbot who always responds in the style of a pirate"
48
+ },
49
+ {
50
+ "role": "user",
51
+ "content": "How many helicopters can a human eat in one sitting?"
52
+ }
53
+ ]
54
+ prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
55
+ outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
56
+ print(outputs[0]["generated_text"])
57
+ ```
58
+
59
+ ## Training procedure
60
+
61
+ ### Training hyperparameters
62
+
63
+ The following hyperparameters were used during training:
64
+ - learning_rate: 2e-05
65
+ - train_batch_size: 4
66
+ - eval_batch_size: 8
67
+ - seed: 42
68
+ - distributed_type: multi-GPU
69
+ - gradient_accumulation_steps: 128
70
+ - total_train_batch_size: 512
71
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
72
+ - lr_scheduler_type: cosine
73
+ - num_epochs: 1
74
+
75
+ ### Training results
76
+
77
+ | Training Loss | Epoch | Step | Validation Loss |
78
+ |:-------------:|:-----:|:----:|:---------------:|
79
+ | 1.9947 | 0.16 | 3 | 2.0093 |
80
+
81
+
82
+ ### Framework versions
83
+
84
+ - Transformers 4.36.1
85
+ - Pytorch 2.1.2+cu121
86
+ - Datasets 2.14.6
87
+ - Tokenizers 0.15.0
88
+ - PEFT 0.6.1
89
+
90
+ ## Citation Information
91
+
92
+ ```
93
+ @misc{rishiraj2023catppt,
94
+ author = {Rishiraj Acharya},
95
+ title = {CatPPT},
96
+ year = {2023},
97
+ publisher = {Hugging Face},
98
+ journal = {Hugging Face repository},
99
+ howpublished = {\url{https://huggingface.co/rishiraj/CatPPT}}
100
+ }
101
+ ```