pronics2004
commited on
Commit
•
4021a5a
1
Parent(s):
be55ba1
Update README.md
Browse files
README.md
CHANGED
@@ -14,14 +14,15 @@ This model is IBM's lightweight, 4-layer toxicity binary classifier for English.
|
|
14 |
- **Release Date**: September 6th, 2024
|
15 |
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|
16 |
|
17 |
-
##
|
|
|
|
|
18 |
This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture. The latency on CPU vs accuracy numbers for this model in comparision to others is shown in the chart below.
|
19 |
|
20 |
![Description of Image](38m_latency.png)
|
21 |
|
22 |
|
23 |
-
|
24 |
-
## Usage
|
25 |
```python
|
26 |
# Example of how to use the model
|
27 |
import torch
|
|
|
14 |
- **Release Date**: September 6th, 2024
|
15 |
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|
16 |
|
17 |
+
## Usage
|
18 |
+
|
19 |
+
### Intended Use
|
20 |
This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture. The latency on CPU vs accuracy numbers for this model in comparision to others is shown in the chart below.
|
21 |
|
22 |
![Description of Image](38m_latency.png)
|
23 |
|
24 |
|
25 |
+
## Prediction
|
|
|
26 |
```python
|
27 |
# Example of how to use the model
|
28 |
import torch
|