Edit model card

CataLlama-v0.2-Instruct-SFT

CataLlama-v0.2-Instruct-SFT is an instruct fine-tune of catallama/CataLlama-v0.2-Base on the catallama/Catalan-Instruct-V2 dataset.

CataLlama-v0.2 was trained on roughly 620 million new tokens which is almost 40% more than CataLlama-v0.1.

This new (V2) SFT Dataset was built mostly from scratch and it only retained parts of the V1.

On top of the existing instructions in Catalan, 250k additional instructions were translated for this model.

All the English instructions existing in the V1 of the dataset were discarded and replaced with high quality instructions scored with RLHFlow/ArmoRM-Llama3-8B-v0.1 reward model.

The model shows improved proficiency with the Catalan language while performing significantly better than CataLlama-v0.1 on all tasks.

This is an instruction fine-tuned model proficient on the following tasks in Catalan

  • Information extraction (suitable for RAG)
  • Named Entity Recognition (NER)
  • Translation from English to Catalan and Catalan to English
  • Summarization - both short form and long form
  • Sentiment analysis
  • Chat

Model developers Laurentiu Petrea based on Llama-3 from Meta.

Model Architecture CataLlama is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and direct preference optimisation (DPO) to align with human preferences for helpfulness and safety.

License The model uses the llama-3 license available at: https://llama.meta.com/llama3/license

Benchmarks

Model CataLlama-v0.1-Instruct-SFT CataLlama-v0.2-Instruct-SFT
MMLU 5 shot 55.28 59.35
GSM8K cot 8 shot 51.63 76.04

Use with transformers

See the snippet below for usage with Transformers:

The model follows the same prompt template as Llama-3 Instruct

import transformers
import torch

model_id = "catallama/CataLlama-v0.2-Instruct-SFT"

pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)

messages = [
    {"role": "user", "content": "Ei com estΓ s avui?"},
]

prompt = pipeline.tokenizer.apply_chat_template(
    messages, 
    tokenize=False, 
    add_generation_prompt=True
)

outputs = pipeline(
    prompt,
    max_new_tokens=1024,
    do_sample=True,
    temperature=0.6,
    top_p=0.9,
)

print(outputs[0]["generated_text"][len(prompt):])

Training procedure

The model was trained with the same prompt template of Llama-3 Instruct.

The model was trained for two epochs on 8x A100 80GB GPUs using DeepSpeed ZeRO State-3 without CPU offloading.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • distributed_type: multi-GPU
  • num_devices: 8
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 100
  • num_epochs: 2

Intended Use

Note: This model is not intended to beat benchmarks, but to demonstrate techniques for augmenting LLMs on new languages and preserve rare languages as part of our world heritage.

Intended Use Cases Llama 3 is intended for commercial and research use in English. Instruction tuned models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks.

Out-of-scope Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in any other way that is prohibited by the Acceptable Use Policy and Llama 3 Community License. Use in languages other than English**.

**Note: Developers may fine-tune Llama 3 models for languages beyond English provided they comply with the Llama 3 Community License and the Acceptable Use Policy.

Downloads last month
8
Safetensors
Model size
8.03B params
Tensor type
BF16
Β·
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for catallama/CataLlama-v0.2-Instruct-SFT

Finetuned
this model
Finetunes
1 model
Quantizations
2 models

Dataset used to train catallama/CataLlama-v0.2-Instruct-SFT

Spaces using catallama/CataLlama-v0.2-Instruct-SFT 4

Collection including catallama/CataLlama-v0.2-Instruct-SFT