Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ inference:
|
|
8 |
max_new_tokens: 64
|
9 |
do_sample: true
|
10 |
temperature: 0.1
|
11 |
-
repetition_penalty: 10
|
12 |
no_repeat_ngram_size: 4
|
13 |
eta_cutoff: 0.0006
|
14 |
renormalize_logits: true
|
@@ -17,29 +17,35 @@ widget:
|
|
17 |
example_title: El Microondas
|
18 |
- text: Kennesaw State University is a public
|
19 |
example_title: Kennesaw State University
|
20 |
-
- text:
|
21 |
-
|
22 |
-
|
|
|
23 |
example_title: Bungie
|
24 |
- text: The Mona Lisa is a world-renowned painting created by
|
25 |
example_title: Mona Lisa
|
26 |
-
- text:
|
|
|
|
|
27 |
example_title: Harry Potter Series
|
28 |
-
- text:
|
|
|
29 |
have water, but no fish. What am I?
|
30 |
|
31 |
-
Answer:
|
32 |
example_title: Riddle
|
33 |
- text: The process of photosynthesis involves the conversion of
|
34 |
example_title: Photosynthesis
|
35 |
-
- text:
|
|
|
36 |
and a loaf of bread. When she got home, she realized she forgot
|
37 |
example_title: Story Continuation
|
38 |
-
- text:
|
39 |
-
|
|
|
40 |
they meet if the distance between the stations is 300 miles?
|
41 |
|
42 |
-
To determine
|
43 |
example_title: Math Problem
|
44 |
- text: In the context of computer programming, an algorithm is
|
45 |
example_title: Algorithm Definition
|
@@ -62,7 +68,8 @@ model-index:
|
|
62 |
value: 21.93
|
63 |
name: normalized accuracy
|
64 |
source:
|
65 |
-
url:
|
|
|
66 |
name: Open LLM Leaderboard
|
67 |
- task:
|
68 |
type: text-generation
|
@@ -78,7 +85,8 @@ model-index:
|
|
78 |
value: 27.86
|
79 |
name: normalized accuracy
|
80 |
source:
|
81 |
-
url:
|
|
|
82 |
name: Open LLM Leaderboard
|
83 |
- task:
|
84 |
type: text-generation
|
@@ -95,7 +103,8 @@ model-index:
|
|
95 |
value: 25.34
|
96 |
name: accuracy
|
97 |
source:
|
98 |
-
url:
|
|
|
99 |
name: Open LLM Leaderboard
|
100 |
- task:
|
101 |
type: text-generation
|
@@ -109,9 +118,10 @@ model-index:
|
|
109 |
num_few_shot: 0
|
110 |
metrics:
|
111 |
- type: mc2
|
112 |
-
value: 46
|
113 |
source:
|
114 |
-
url:
|
|
|
115 |
name: Open LLM Leaderboard
|
116 |
- task:
|
117 |
type: text-generation
|
@@ -128,7 +138,8 @@ model-index:
|
|
128 |
value: 50.83
|
129 |
name: accuracy
|
130 |
source:
|
131 |
-
url:
|
|
|
132 |
name: Open LLM Leaderboard
|
133 |
- task:
|
134 |
type: text-generation
|
@@ -142,11 +153,19 @@ model-index:
|
|
142 |
num_few_shot: 5
|
143 |
metrics:
|
144 |
- type: acc
|
145 |
-
value: 0
|
146 |
name: accuracy
|
147 |
source:
|
148 |
-
url:
|
|
|
149 |
name: Open LLM Leaderboard
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
150 |
---
|
151 |
|
152 |
|
@@ -155,7 +174,7 @@ model-index:
|
|
155 |
Inspired by [Phi2](https://huggingface.co/microsoft/phi-2), and open source small language model attempts like [smol_llama-101M-GQA](https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA).
|
156 |
Pre-trained with training 7B token **from scratch**, with application of quality filter to datasets resulting in 0.26B token.
|
157 |
The control is [kenhktsui/nano-phi-115M-control-v0.1](https://huggingface.co/kenhktsui/nano-phi-115M-control-v0.1), where full dataset (0.6B) is used.
|
158 |
-
Not much degradation in performance despite only using **42%** of the data due to the effective quality filter.
|
159 |
In fact, upon inspection, the 6000 steps chkpt achieves similar performance as this model, signaling underlying **effective training due to high quality data**.
|
160 |
It just took 1d to train in Colab with a A100 40GB (**<USD$ 50**).
|
161 |
It achieves quite competitive results in evaluation given its training token, and training data size.
|
@@ -569,5 +588,4 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
569 |
|MMLU (5-Shot) |25.34|
|
570 |
|TruthfulQA (0-shot) |46.00|
|
571 |
|Winogrande (5-shot) |50.83|
|
572 |
-
|GSM8k (5-shot) | 0.00|
|
573 |
-
|
|
|
8 |
max_new_tokens: 64
|
9 |
do_sample: true
|
10 |
temperature: 0.1
|
11 |
+
repetition_penalty: 10
|
12 |
no_repeat_ngram_size: 4
|
13 |
eta_cutoff: 0.0006
|
14 |
renormalize_logits: true
|
|
|
17 |
example_title: El Microondas
|
18 |
- text: Kennesaw State University is a public
|
19 |
example_title: Kennesaw State University
|
20 |
+
- text: >-
|
21 |
+
Bungie Studios is an American video game developer. They are most famous for
|
22 |
+
developing the award winning Halo series of video games. They also made
|
23 |
+
Destiny. The studio was founded
|
24 |
example_title: Bungie
|
25 |
- text: The Mona Lisa is a world-renowned painting created by
|
26 |
example_title: Mona Lisa
|
27 |
+
- text: >-
|
28 |
+
The Harry Potter series, written by J.K. Rowling, begins with the book
|
29 |
+
titled
|
30 |
example_title: Harry Potter Series
|
31 |
+
- text: >-
|
32 |
+
Question: I have cities, but no houses. I have mountains, but no trees. I
|
33 |
have water, but no fish. What am I?
|
34 |
|
35 |
+
Answer:
|
36 |
example_title: Riddle
|
37 |
- text: The process of photosynthesis involves the conversion of
|
38 |
example_title: Photosynthesis
|
39 |
+
- text: >-
|
40 |
+
Jane went to the store to buy some groceries. She picked up apples, oranges,
|
41 |
and a loaf of bread. When she got home, she realized she forgot
|
42 |
example_title: Story Continuation
|
43 |
+
- text: >-
|
44 |
+
Problem 2: If a train leaves Station A at 9:00 AM and travels at 60 mph, and
|
45 |
+
another train leaves Station B at 10:00 AM and travels at 80 mph, when will
|
46 |
they meet if the distance between the stations is 300 miles?
|
47 |
|
48 |
+
To determine
|
49 |
example_title: Math Problem
|
50 |
- text: In the context of computer programming, an algorithm is
|
51 |
example_title: Algorithm Definition
|
|
|
68 |
value: 21.93
|
69 |
name: normalized accuracy
|
70 |
source:
|
71 |
+
url: >-
|
72 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
73 |
name: Open LLM Leaderboard
|
74 |
- task:
|
75 |
type: text-generation
|
|
|
85 |
value: 27.86
|
86 |
name: normalized accuracy
|
87 |
source:
|
88 |
+
url: >-
|
89 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
90 |
name: Open LLM Leaderboard
|
91 |
- task:
|
92 |
type: text-generation
|
|
|
103 |
value: 25.34
|
104 |
name: accuracy
|
105 |
source:
|
106 |
+
url: >-
|
107 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
108 |
name: Open LLM Leaderboard
|
109 |
- task:
|
110 |
type: text-generation
|
|
|
118 |
num_few_shot: 0
|
119 |
metrics:
|
120 |
- type: mc2
|
121 |
+
value: 46
|
122 |
source:
|
123 |
+
url: >-
|
124 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
125 |
name: Open LLM Leaderboard
|
126 |
- task:
|
127 |
type: text-generation
|
|
|
138 |
value: 50.83
|
139 |
name: accuracy
|
140 |
source:
|
141 |
+
url: >-
|
142 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
143 |
name: Open LLM Leaderboard
|
144 |
- task:
|
145 |
type: text-generation
|
|
|
153 |
num_few_shot: 5
|
154 |
metrics:
|
155 |
- type: acc
|
156 |
+
value: 0
|
157 |
name: accuracy
|
158 |
source:
|
159 |
+
url: >-
|
160 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kenhktsui/nano-phi-115M-v0.1
|
161 |
name: Open LLM Leaderboard
|
162 |
+
datasets:
|
163 |
+
- kenhktsui/minipile_quality_score_v1
|
164 |
+
- kenhktsui/simple_wikipedia_LM_quality_score_v1
|
165 |
+
- kenhktsui/refinedweb-3m_quality_score_v1
|
166 |
+
- kenhktsui/TM-DATA_quality_score_v1
|
167 |
+
- kenhktsui/openwebtext_quality_score_v1
|
168 |
+
|
169 |
---
|
170 |
|
171 |
|
|
|
174 |
Inspired by [Phi2](https://huggingface.co/microsoft/phi-2), and open source small language model attempts like [smol_llama-101M-GQA](https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA).
|
175 |
Pre-trained with training 7B token **from scratch**, with application of quality filter to datasets resulting in 0.26B token.
|
176 |
The control is [kenhktsui/nano-phi-115M-control-v0.1](https://huggingface.co/kenhktsui/nano-phi-115M-control-v0.1), where full dataset (0.6B) is used.
|
177 |
+
Not much degradation in performance despite only using **42%** of the data due to the effective quality filter ("quality_score_v1" > 0.5).
|
178 |
In fact, upon inspection, the 6000 steps chkpt achieves similar performance as this model, signaling underlying **effective training due to high quality data**.
|
179 |
It just took 1d to train in Colab with a A100 40GB (**<USD$ 50**).
|
180 |
It achieves quite competitive results in evaluation given its training token, and training data size.
|
|
|
588 |
|MMLU (5-Shot) |25.34|
|
589 |
|TruthfulQA (0-shot) |46.00|
|
590 |
|Winogrande (5-shot) |50.83|
|
591 |
+
|GSM8k (5-shot) | 0.00|
|
|