SamLowe commited on
Commit
7e81e16
1 Parent(s): 49ebfbf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -21,7 +21,7 @@ This model is the ONNX version of [https://huggingface.co/SamLowe/roberta-base-g
21
 
22
  `onnx/model.onnx` is the full precision ONNX version
23
 
24
- - that has identical performance to the original transformers model
25
  - and has the same model size (499MB)
26
  - is faster than inference than normal Transformers, particularly for smaller batch sizes
27
  - in my tests about 2x to 3x as fast for a batch size of 1 on a 8 core 11th gen i7 CPU using ONNXRuntime
@@ -32,7 +32,7 @@ This model is the ONNX version of [https://huggingface.co/SamLowe/roberta-base-g
32
 
33
  - that is one quarter the size (125MB) of the full precision model (above)
34
  - but delivers almost all of the accuracy
35
- - is faster than inference
36
  - about 2x as fast for a batch size of 1 on an 8 core 11th gen i7 CPU using ONNXRuntime vs the full precision model above
37
  - which makes it circa 5x as fast as the full precision normal Transformers model (on the above mentioned CPU, for a batch of 1)
38
 
 
21
 
22
  `onnx/model.onnx` is the full precision ONNX version
23
 
24
+ - that has identical accuracy/metrics to the original transformers model
25
  - and has the same model size (499MB)
26
  - is faster than inference than normal Transformers, particularly for smaller batch sizes
27
  - in my tests about 2x to 3x as fast for a batch size of 1 on a 8 core 11th gen i7 CPU using ONNXRuntime
 
32
 
33
  - that is one quarter the size (125MB) of the full precision model (above)
34
  - but delivers almost all of the accuracy
35
+ - is faster for inference
36
  - about 2x as fast for a batch size of 1 on an 8 core 11th gen i7 CPU using ONNXRuntime vs the full precision model above
37
  - which makes it circa 5x as fast as the full precision normal Transformers model (on the above mentioned CPU, for a batch of 1)
38