Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ This model is the ONNX version of [https://huggingface.co/SamLowe/roberta-base-g
|
|
21 |
|
22 |
`onnx/model.onnx` is the full precision ONNX version
|
23 |
|
24 |
-
- that has identical
|
25 |
- and has the same model size (499MB)
|
26 |
- is faster than inference than normal Transformers, particularly for smaller batch sizes
|
27 |
- in my tests about 2x to 3x as fast for a batch size of 1 on a 8 core 11th gen i7 CPU using ONNXRuntime
|
@@ -32,7 +32,7 @@ This model is the ONNX version of [https://huggingface.co/SamLowe/roberta-base-g
|
|
32 |
|
33 |
- that is one quarter the size (125MB) of the full precision model (above)
|
34 |
- but delivers almost all of the accuracy
|
35 |
-
- is faster
|
36 |
- about 2x as fast for a batch size of 1 on an 8 core 11th gen i7 CPU using ONNXRuntime vs the full precision model above
|
37 |
- which makes it circa 5x as fast as the full precision normal Transformers model (on the above mentioned CPU, for a batch of 1)
|
38 |
|
|
|
21 |
|
22 |
`onnx/model.onnx` is the full precision ONNX version
|
23 |
|
24 |
+
- that has identical accuracy/metrics to the original transformers model
|
25 |
- and has the same model size (499MB)
|
26 |
- is faster than inference than normal Transformers, particularly for smaller batch sizes
|
27 |
- in my tests about 2x to 3x as fast for a batch size of 1 on a 8 core 11th gen i7 CPU using ONNXRuntime
|
|
|
32 |
|
33 |
- that is one quarter the size (125MB) of the full precision model (above)
|
34 |
- but delivers almost all of the accuracy
|
35 |
+
- is faster for inference
|
36 |
- about 2x as fast for a batch size of 1 on an 8 core 11th gen i7 CPU using ONNXRuntime vs the full precision model above
|
37 |
- which makes it circa 5x as fast as the full precision normal Transformers model (on the above mentioned CPU, for a batch of 1)
|
38 |
|