aashish1904 commited on
Commit
b19bd75
1 Parent(s): 1cafb76

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +161 -0
README.md ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: other
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ base_model:
9
+ - Qwen/Qwen2.5-3B
10
+ - Qwen/Qwen2.5-3B-Instruct
11
+ - arcee-ai/raspberry-3B
12
+ license_name: qwen-research
13
+ license_link: https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE
14
+ model-index:
15
+ - name: Rombos-LLM-V2.5.1-Qwen-3b
16
+ results:
17
+ - task:
18
+ type: text-generation
19
+ name: Text Generation
20
+ dataset:
21
+ name: IFEval (0-Shot)
22
+ type: HuggingFaceH4/ifeval
23
+ args:
24
+ num_few_shot: 0
25
+ metrics:
26
+ - type: inst_level_strict_acc and prompt_level_strict_acc
27
+ value: 25.95
28
+ name: strict accuracy
29
+ source:
30
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
31
+ name: Open LLM Leaderboard
32
+ - task:
33
+ type: text-generation
34
+ name: Text Generation
35
+ dataset:
36
+ name: BBH (3-Shot)
37
+ type: BBH
38
+ args:
39
+ num_few_shot: 3
40
+ metrics:
41
+ - type: acc_norm
42
+ value: 14.88
43
+ name: normalized accuracy
44
+ source:
45
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
46
+ name: Open LLM Leaderboard
47
+ - task:
48
+ type: text-generation
49
+ name: Text Generation
50
+ dataset:
51
+ name: MATH Lvl 5 (4-Shot)
52
+ type: hendrycks/competition_math
53
+ args:
54
+ num_few_shot: 4
55
+ metrics:
56
+ - type: exact_match
57
+ value: 8.31
58
+ name: exact match
59
+ source:
60
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
61
+ name: Open LLM Leaderboard
62
+ - task:
63
+ type: text-generation
64
+ name: Text Generation
65
+ dataset:
66
+ name: GPQA (0-shot)
67
+ type: Idavidrein/gpqa
68
+ args:
69
+ num_few_shot: 0
70
+ metrics:
71
+ - type: acc_norm
72
+ value: 3.24
73
+ name: acc_norm
74
+ source:
75
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
76
+ name: Open LLM Leaderboard
77
+ - task:
78
+ type: text-generation
79
+ name: Text Generation
80
+ dataset:
81
+ name: MuSR (0-shot)
82
+ type: TAUR-Lab/MuSR
83
+ args:
84
+ num_few_shot: 0
85
+ metrics:
86
+ - type: acc_norm
87
+ value: 7.82
88
+ name: acc_norm
89
+ source:
90
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: MMLU-PRO (5-shot)
97
+ type: TIGER-Lab/MMLU-Pro
98
+ config: main
99
+ split: test
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 19.1
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
108
+ name: Open LLM Leaderboard
109
+
110
+ ---
111
+
112
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
113
+
114
+
115
+ # QuantFactory/Rombos-LLM-V2.5.1-Qwen-3b-GGUF
116
+ This is quantized version of [rombodawg/Rombos-LLM-V2.5.1-Qwen-3b](https://huggingface.co/rombodawg/Rombos-LLM-V2.5.1-Qwen-3b) created using llama.cpp
117
+
118
+ # Original Model Card
119
+
120
+ # Rombos-LLM-V2.5.1-Qwen-3b
121
+
122
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/642cc1c253e76b4c2286c58e/pNDtgE5FDkxxvbG4qiZ1A.jpeg)
123
+
124
+ A little experiment I threw together to take a really high quality LLM I found (arcee-ai/raspberry-3B) and merge it using the last step of my Continuous Finetuning method outlines in the paper linked bellow.
125
+
126
+ https://docs.google.com/document/d/1OjbjU5AOz4Ftn9xHQrX3oFQGhQ6RDUuXQipnQ9gn6tU/edit?usp=sharing
127
+
128
+ Mergekit.yaml file is as follows:
129
+ ```yaml
130
+ models:
131
+ - model: Qwen2.5-3B-Instruct
132
+ parameters:
133
+ weight: 1
134
+ density: 1
135
+ - model: raspberry-3B
136
+ parameters:
137
+ weight: 1
138
+ density: 1
139
+ merge_method: ties
140
+ base_model: Qwen2.5-3B
141
+ parameters:
142
+ weight: 1
143
+ density: 1
144
+ normalize: true
145
+ int8_mask: true
146
+ dtype: bfloat16
147
+ ```
148
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
149
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_rombodawg__Rombos-LLM-V2.5.1-Qwen-3b)
150
+
151
+ | Metric |Value|
152
+ |-------------------|----:|
153
+ |Avg. |13.22|
154
+ |IFEval (0-Shot) |25.95|
155
+ |BBH (3-Shot) |14.88|
156
+ |MATH Lvl 5 (4-Shot)| 8.31|
157
+ |GPQA (0-shot) | 3.24|
158
+ |MuSR (0-shot) | 7.82|
159
+ |MMLU-PRO (5-shot) |19.10|
160
+
161
+