Image-Text-to-Text
Transformers
Safetensors
English
idefics2
pretraining
multimodal
vision
quantized
4-bit precision
AWQ
Inference Endpoints
awq
VictorSanh's picture
update readme
d32f888
---
license: apache-2.0
datasets:
- HuggingFaceM4/OBELICS
- laion/laion-coco
- wikipedia
- facebook/pmd
- pixparse/idl-wds
- pixparse/pdfa-eng-wds
- wendlerc/RenderedText
- HuggingFaceM4/the_cauldron
- teknium/OpenHermes-2.5
- GAIR/lima
- databricks/databricks-dolly-15k
- meta-math/MetaMathQA
- TIGER-Lab/MathInstruct
- microsoft/orca-math-word-problems-200k
- camel-ai/math
- AtlasUnified/atlas-math-sets
- tiedong/goat
- Lin-Chen/ShareGPT4V
- jxu124/llava_conversation_58k
language:
- en
tags:
- multimodal
- vision
- image-text-to-text
- quantized
- 4-bit
- AWQ
---
4-bit AWQ-quantized version of [HuggingFaceM4/idefics2-8b-base](https://huggingface.co/HuggingFaceM4/idefics2-8b-base). Refer to the original model's card for more information (including inference snippet).