Update README.md
Browse files
README.md
CHANGED
@@ -39,12 +39,8 @@ The embedding module **roberta.embeddings.token_type_embeddings** falls back to
|
|
39 |
|
40 |
### Test result
|
41 |
|
42 |
-
- Batch size = 8
|
43 |
-
- [Amazon Web Services](https://aws.amazon.com/) c6i.xlarge (Intel ICE Lake: 4 vCPUs, 8g Memory) instance.
|
44 |
-
|
45 |
| |INT8|FP32|
|
46 |
|---|:---:|:---:|
|
47 |
-
| **Throughput (samples/sec)** |25.737|13.171|
|
48 |
| **Accuracy (eval-f1)** |0.9247|0.9138|
|
49 |
| **Model size (MB)** |121|476|
|
50 |
|
@@ -56,7 +52,3 @@ int8_model = OptimizedModel.from_pretrained(
|
|
56 |
'Intel/roberta-base-mrpc-int8-static',
|
57 |
)
|
58 |
```
|
59 |
-
|
60 |
-
Notes:
|
61 |
-
- The INT8 model has better performance than the FP32 model when the CPU is fully occupied. Otherwise, there will be the illusion that INT8 is inferior to FP32.
|
62 |
-
|
|
|
39 |
|
40 |
### Test result
|
41 |
|
|
|
|
|
|
|
42 |
| |INT8|FP32|
|
43 |
|---|:---:|:---:|
|
|
|
44 |
| **Accuracy (eval-f1)** |0.9247|0.9138|
|
45 |
| **Model size (MB)** |121|476|
|
46 |
|
|
|
52 |
'Intel/roberta-base-mrpc-int8-static',
|
53 |
)
|
54 |
```
|
|
|
|
|
|
|
|