benchmark_results / 2024-06-10 /summaries.json
hf-transformers-bot's picture
Upload folder using huggingface_hub
754823f verified
raw
history blame
1.73 kB
[
{
"model": "google/gemma-2b",
"commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
"config": "backend.cache_implementation=null,backend.torch_compile=False",
"metrics": {
"prefill.latency.mean": 0.020382543563842774,
"prefill.throughput.value": 343.43113155011315,
"decode.latency.mean": 2.425145751953125,
"decode.throughput.value": 52.3679865004892,
"per_token.latency.mean": 0.019167450524130356,
"per_token.throughput.value": 52.17177937885252
}
},
{
"model": "google/gemma-2b",
"commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
"config": "backend.cache_implementation=static,backend.torch_compile=False",
"metrics": {
"prefill.latency.mean": 0.02269580841064453,
"prefill.throughput.value": 308.4269955643853,
"decode.latency.mean": 2.5586441650390626,
"decode.throughput.value": 49.63566319041519,
"per_token.latency.mean": 0.020235656602580556,
"per_token.throughput.value": 49.41771940686495
}
},
{
"model": "google/gemma-2b",
"commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5",
"config": "backend.cache_implementation=static,backend.torch_compile=True",
"metrics": {
"prefill.latency.mean": 0.014055791854858397,
"prefill.throughput.value": 498.01534287664083,
"decode.latency.mean": 1.5724798583984374,
"decode.throughput.value": 80.7641505369416,
"per_token.latency.mean": 0.012435026971719011,
"per_token.throughput.value": 80.41800007947715
}
}
]