phi-2-super-4bit / README.md
muhtasham's picture
Upload folder using huggingface_hub
24ed1f2 verified
metadata
language:
  - en
license: mit
tags:
  - convAI
  - conversational
  - mlx
license_link: https://huggingface.co/microsoft/phi-2/resolve/main/LICENSE
widget:
  - text: Hello who are you?
    example_title: Identity
  - text: What can you do?
    example_title: Capabilities
  - text: Create a fastapi endpoint to retrieve the weather given a zip code.
    example_title: Coding
pipeline_tag: text-generation
model-index:
  - name: phi-2-super
    results:
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: Instruction Following Eval
          type: wis-k/instruction-following-eval
        metrics:
          - type: acc
            value: 0.2717
            name: prompt_level_loose_acc
        source:
          url: https://github.com/huggingface/lighteval
          name: LightEval

mlx-community/phi-2-super-4bit

This model was converted to MLX format from abacaj/phi-2-super. Refer to the original model card for more details on the model.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("mlx-community/phi-2-super-4bit")
response = generate(model, tokenizer, prompt="hello", verbose=True)