Edit model card

mlx-community/Maxine-34B-stock

This model was converted to MLX format from louisbrulenaudet/Maxine-34B-stock using mlx-vlm version 0.15.2. Refer to the original model card for more details on the model.

Use with mlx

pip install -U mlx-vlm
python -m mlx_vlm.generate --model  mlx-community/Maxine-34B-stock --max-tokens 100 --temp 0.0
from mlx_lm import load, generate

model, tokenizer = load("mlx-community/Maxine-34B-stock")
response = generate(model, tokenizer, prompt="hello", verbose=True)

Citing & Authors

If you use this code in your research, please use the following BibTeX entry.

@misc{louisbrulenaudet2024,
  author =       {Louis Brulé Naudet},
  title =        {Maxine-34B-stock, an xtraordinary 34B model},
  year =         {2024}
  howpublished = {\url{https://huggingface.co/mlx-community/Maxine-34B-stock}},
}

Feedback

If you have any feedback, please reach out at louisbrulenaudet@icloud.com.

Downloads last month
25
Safetensors
Model size
5.76B params
Tensor type
FP16
·
U32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mlx-community/Maxine-34B-stock

Finetuned
(1)
this model

Datasets used to train mlx-community/Maxine-34B-stock