QwenMath
A generation LLM which can solve math problems.
Training Statistics
training-method: lora
training-time: "5:42"
data-size: 500
epoch: 3
total_flos: "1372250GF"
train_loss: 0.6441
train_samples_per_second: 4.385
train_steps_per_second: 0.544
Validation Set Performance
Dataset used: test split of fdyrd/MATH. Metric: accuracy
Level | Algebra | Intermediate Algebra | Prealgebra | Precalculus | Number Theory | Geometry | Counting & Probability | Average |
---|---|---|---|---|---|---|---|---|
Level 1 | 0.541 : 135 | 0.192 : 52 | 0.477 : 86 | 0.228 : 57 | 0.467 : 30 | 0.263 : 38 | 0.359 : 39 | 0.361 |
Level 2 | 0.323 : 201 | 0.109 : 128 | 0.367 : 177 | 0.044 : 113 | 0.38 : 92 | 0.134 : 82 | 0.248 : 101 | 0.229 |
Level 3 | 0.291 : 261 | 0.046 : 195 | 0.308 : 224 | 0.0 : 127 | 0.262 : 122 | 0.088 : 102 | 0.16 : 100 | 0.165 |
Level 4 | 0.18 : 283 | 0.024 : 248 | 0.22 : 191 | 0.009 : 114 | 0.169 : 142 | 0.064 : 125 | 0.09 : 111 | 0.108 |
Level 5 | 0.088 : 307 | 0.004 : 280 | 0.104 : 193 | 0.0 : 135 | 0.136 : 154 | 0.023 : 132 | 0.065 : 123 | 0.06 |
Average | 0.285 | 0.075 | 0.295 | 0.056 | 0.283 | 0.114 | 0.184 | 0.166 |
Test Set Performance
[
{
"dataset": "MATH500",
"url": "https://huggingface.co/datasets/qq8933/MATH500",
"accuracy": 0.286
},
{
"dataset": "GSM8K",
"url": "https://huggingface.co/datasets/openai/gsm8k",
"accuracy": 0.382
}
]
- Downloads last month
- 10
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for fdyrd/QwenMath-0.5B
Base model
Qwen/Qwen2.5-0.5B