Molmo-7B-D BnB 4bit quant 30GB -> 7GB
approx. 12GB VRAM required
base model for more information:
https://huggingface.co/allenai/Molmo-7B-D-0924
example code:
https://github.com/cyan2k/molmo-7b-bnb-4bit
performance metrics & benchmarks to compare with base will follow over the next week
- Downloads last month
- 118
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The HF Inference API does not support model that require custom code execution.