weiweiz1 commited on
Commit
f4a8e35
·
2 Parent(s): bc917a8 7270565

Merge branch 'main' of https://huggingface.co/OPEA/llama-joycaption-alpha-two-hf-llava-int4-sym-inc into main

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -5,7 +5,7 @@ datasets:
5
 
6
  ## Model Details
7
 
8
- This model is an int4 model with group_size 128 and symmetric quantization of [fancyfeast/llama-joycaption-alpha-two-hf-llava](https://huggingface.co/fancyfeast/llama-joycaption-alpha-two-hf-llava) generated by [intel/auto-round](https://github.com/intel/auto-round). Load the model with revision="" to use AutoGPTQ format.
9
 
10
  ## How To Use
11
 
@@ -25,7 +25,11 @@ quantized_model_path="OPEA/llama-joycaption-alpha-two-hf-llava-int4-sym-inc"
25
 
26
  # Load JoyCaption INT4 Model
27
  processor = AutoProcessor.from_pretrained(quantized_model_path)
28
- model = LlavaForConditionalGeneration.from_pretrained(quantized_model_path, device_map=0)
 
 
 
 
29
  model.eval()
30
 
31
  image_url = "http://images.cocodataset.org/train2017/000000116003.jpg"
 
5
 
6
  ## Model Details
7
 
8
+ This model is an int4 model with group_size 128 and symmetric quantization of [fancyfeast/llama-joycaption-alpha-two-hf-llava](https://huggingface.co/fancyfeast/llama-joycaption-alpha-two-hf-llava) generated by [intel/auto-round](https://github.com/intel/auto-round). Load the model with revision="bc917a8" to use AutoGPTQ format.
9
 
10
  ## How To Use
11
 
 
25
 
26
  # Load JoyCaption INT4 Model
27
  processor = AutoProcessor.from_pretrained(quantized_model_path)
28
+ model = LlavaForConditionalGeneration.from_pretrained(
29
+ quantized_model_path,
30
+ device_map="auto",
31
+ revision="bc917a8" ## ##AutoGPTQ format
32
+ )
33
  model.eval()
34
 
35
  image_url = "http://images.cocodataset.org/train2017/000000116003.jpg"