JosephusCheung commited on
Commit
56c2d12
1 Parent(s): baeb9de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -16,6 +16,12 @@ tags:
16
 
17
  ![](https://huggingface.co/JosephusCheung/Guanaco/resolve/main/StupidBanner.png)
18
 
 
 
 
 
 
 
19
  **It is highly recommended to use fp16 inference for this model, as 8-bit precision may significantly affect performance. If you require a more Consumer Hardware friendly version, please use the specialized quantized [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).**
20
 
21
  **You are encouraged to use the latest version of transformers from GitHub.**
 
16
 
17
  ![](https://huggingface.co/JosephusCheung/Guanaco/resolve/main/StupidBanner.png)
18
 
19
+ **You need a Colab pro for full version (8bit not well performed, should use fp16)**
20
+
21
+ [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1ocSmoy3ba1EkYu7JWT1oCw9vz8qC2cMk#scrollTo=zLORi5OcPcIJ)
22
+
23
+ Free T4 Colab demo, please check 4bit version [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).**
24
+
25
  **It is highly recommended to use fp16 inference for this model, as 8-bit precision may significantly affect performance. If you require a more Consumer Hardware friendly version, please use the specialized quantized [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).**
26
 
27
  **You are encouraged to use the latest version of transformers from GitHub.**