Merge branch 'main' of https://huggingface.co/OPEA/llama-joycaption-alpha-two-hf-llava-int4-sym-inc into main
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ datasets:
|
|
5 |
|
6 |
## Model Details
|
7 |
|
8 |
-
This model is an int4 model with group_size 128 and symmetric quantization of [fancyfeast/llama-joycaption-alpha-two-hf-llava](https://huggingface.co/fancyfeast/llama-joycaption-alpha-two-hf-llava) generated by [intel/auto-round](https://github.com/intel/auto-round). Load the model with revision="" to use AutoGPTQ format.
|
9 |
|
10 |
## How To Use
|
11 |
|
@@ -25,7 +25,11 @@ quantized_model_path="OPEA/llama-joycaption-alpha-two-hf-llava-int4-sym-inc"
|
|
25 |
|
26 |
# Load JoyCaption INT4 Model
|
27 |
processor = AutoProcessor.from_pretrained(quantized_model_path)
|
28 |
-
model = LlavaForConditionalGeneration.from_pretrained(
|
|
|
|
|
|
|
|
|
29 |
model.eval()
|
30 |
|
31 |
image_url = "http://images.cocodataset.org/train2017/000000116003.jpg"
|
|
|
5 |
|
6 |
## Model Details
|
7 |
|
8 |
+
This model is an int4 model with group_size 128 and symmetric quantization of [fancyfeast/llama-joycaption-alpha-two-hf-llava](https://huggingface.co/fancyfeast/llama-joycaption-alpha-two-hf-llava) generated by [intel/auto-round](https://github.com/intel/auto-round). Load the model with revision="bc917a8" to use AutoGPTQ format.
|
9 |
|
10 |
## How To Use
|
11 |
|
|
|
25 |
|
26 |
# Load JoyCaption INT4 Model
|
27 |
processor = AutoProcessor.from_pretrained(quantized_model_path)
|
28 |
+
model = LlavaForConditionalGeneration.from_pretrained(
|
29 |
+
quantized_model_path,
|
30 |
+
device_map="auto",
|
31 |
+
revision="bc917a8" ## ##AutoGPTQ format
|
32 |
+
)
|
33 |
model.eval()
|
34 |
|
35 |
image_url = "http://images.cocodataset.org/train2017/000000116003.jpg"
|