ivkalgin commited on
Commit
ac04cb3
1 Parent(s): ed99478

fixed typos, updated model size values

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -18,19 +18,19 @@ tags:
18
 
19
  GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
20
 
21
- This repository contains tensorrt engines with mixed precission int8 + fp32. You can find prebuilded engines for next GPUs:
22
  * RTX 4090
23
  * RTX 3080 Ti
24
  * RTX 2080 Ti
25
 
26
- Source ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and will be published soon.
27
 
28
  ## Test result
29
 
30
  | |INT8|FP32|
31
  |---|:---:|:---:|
32
  | **Lambada Acc** |78.50%|79.54%|
33
- | **Model size (GB)** |8.1|23|
34
 
35
 
36
  ## How to use
 
18
 
19
  GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
20
 
21
+ This repository contains TensorRT engines with mixed precission int8 + fp32. You can find prebuilt engines for next GPUs:
22
  * RTX 4090
23
  * RTX 3080 Ti
24
  * RTX 2080 Ti
25
 
26
+ ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and will be published soon.
27
 
28
  ## Test result
29
 
30
  | |INT8|FP32|
31
  |---|:---:|:---:|
32
  | **Lambada Acc** |78.50%|79.54%|
33
+ | **Model size (GB)** |8.5|24.2|
34
 
35
 
36
  ## How to use