Same as h2oai/h2ogpt-16k-codellama-34b-instruct but with config.json modified to be 32k for embeddings, which still functions fine as 16k model and allows stretching into 32k in vLLM that otherwise cannot modify maximum sequence length.

Downloads last month
180
Safetensors
Model size
33.7B params
Tensor type
FP16
ยท
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for h2oai/h2ogpt-32k-codellama-34b-instruct

Quantizations
1 model

Space using h2oai/h2ogpt-32k-codellama-34b-instruct 1