pronics2004 commited on
Commit
4021a5a
1 Parent(s): be55ba1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -14,14 +14,15 @@ This model is IBM's lightweight, 4-layer toxicity binary classifier for English.
14
  - **Release Date**: September 6th, 2024
15
  - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
16
 
17
- ## Feature
 
 
18
  This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture. The latency on CPU vs accuracy numbers for this model in comparision to others is shown in the chart below.
19
 
20
  ![Description of Image](38m_latency.png)
21
 
22
 
23
-
24
- ## Usage
25
  ```python
26
  # Example of how to use the model
27
  import torch
 
14
  - **Release Date**: September 6th, 2024
15
  - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
16
 
17
+ ## Usage
18
+
19
+ ### Intended Use
20
  This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture. The latency on CPU vs accuracy numbers for this model in comparision to others is shown in the chart below.
21
 
22
  ![Description of Image](38m_latency.png)
23
 
24
 
25
+ ## Prediction
 
26
  ```python
27
  # Example of how to use the model
28
  import torch