violetch24 commited on
Commit
e064e32
1 Parent(s): cc68488

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -12,7 +12,7 @@ datasets:
12
  metrics:
13
  - f1
14
  model-index:
15
- - name: bart-large-mrpc-int8-static
16
  results:
17
  - task:
18
  name: Text Classification
@@ -30,7 +30,7 @@ model-index:
30
 
31
  ### Post-training dynamic quantization
32
 
33
- This is an INT8 PyTorch model quantized with [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
34
 
35
  The original fp32 model comes from the fine-tuned model [bart-large-mrpc](https://huggingface.co/Intel/bart-large-mrpc).
36
 
@@ -41,11 +41,11 @@ The original fp32 model comes from the fine-tuned model [bart-large-mrpc](https:
41
  | **Accuracy (eval-f1)** |0.9051|0.9120|
42
  | **Model size (MB)** |547|1556.48|
43
 
44
- ### Load with Intel® Neural Compressor:
45
 
46
  ```python
47
- from neural_compressor.utils.load_huggingface import OptimizedModel
48
- int8_model = OptimizedModel.from_pretrained(
49
  'Intel/bart-large-mrpc-int8-dynamic',
50
  )
51
  ```
 
12
  metrics:
13
  - f1
14
  model-index:
15
+ - name: bart-large-mrpc-int8-dynamic
16
  results:
17
  - task:
18
  name: Text Classification
 
30
 
31
  ### Post-training dynamic quantization
32
 
33
+ This is an INT8 PyTorch model quantized with [huggingface/optimum-intel](https://github.com/huggingface/optimum-intel) through the usage of [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
34
 
35
  The original fp32 model comes from the fine-tuned model [bart-large-mrpc](https://huggingface.co/Intel/bart-large-mrpc).
36
 
 
41
  | **Accuracy (eval-f1)** |0.9051|0.9120|
42
  | **Model size (MB)** |547|1556.48|
43
 
44
+ ### Load with optimum:
45
 
46
  ```python
47
+ from optimum.intel.neural_compressor.quantization import IncQuantizedModelForSequenceClassification
48
+ int8_model = IncQuantizedModelForSequenceClassification.from_pretrained(
49
  'Intel/bart-large-mrpc-int8-dynamic',
50
  )
51
  ```