Safetensors
llama

license: cc-by-4.0 datasets: - santoshtyss/uk_legislation language: - en base_model: - unsloth/Llama-3.2-3B tags: - legal

Llama 3.2 UK Legislation 3B

Llama 3.2 UK Legislation 3B

This model is a fine-tuned version of the Llama 3.2 UK Legislation 3B base. It was instruction-tuned for Q and A on UK legislation. It was trained as part of a blog series, see the article here

Model Details

Model Description

Model Sources

Uses

Intended Use

This model is designed to serve as Q and A for UK legislation and for further development for tasks such as:

  • Domain-specific applications in law or other fields
  • Research and experimentation in natural language processing
  • General-purpose natural language understanding and generation

Out-of-Scope Use

This model is not suitable for:

  • Providing domain-specific expertise
  • Applications requiring high accuracy or nuanced understanding of UK legislation
  • Tasks involving sensitive or critical real-world applications without rigorous evaluation

Bias, Risks, and Limitations

  • Bias: The model may reflect biases inherent in the pretraining data. Outputs should be critically evaluated for accuracy and fairness.
  • Risks: As a base model, it may generate responses that are overly general or contextually inappropriate for specific tasks.
  • Limitations: The model is not fine-tuned for specific domains, including legal or legislative text, and does not include the most recent developments in any field.

How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained("EryriLabs/llama-3.2-uk-legislation-instruct-3b", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("EryriLabs/llama-3.2-uk-legislation-instruct-3b")

# Sample question
input_text = "What are the main principles of UK legislation?"

# Tokenize and generate response
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(inputs["input_ids"], max_length=50)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)

Technical Specifications

  • Model Architecture: Llama 3.2 3B, a transformer-based model designed for natural language processing tasks.
  • Training Data: Pretrained on a diverse dataset of general text.
  • Compute Infrastructure: Training conducted on high-performance GPUs (e.g., NVIDIA A100).

Citation

If you use this model, please cite:

@misc{llama3.2-uk-legislation-instruct-3b,
  author = {GPT-LABS.AI},
  title = {Llama 3.2 UK Legislation Instruct 3B},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/EryriLabs/llama-3.2-uk-legislation-instruct-3b}
}

Model Card Authors

  • GPT-LABS.AI

Contact

For questions or feedback, please visit gpt-labs.ai

Downloads last month
9
Safetensors
Model size
3.61B params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for EryriLabs/llama-3.2-uk-legislation-instruct-3b

Finetuned
(1)
this model
Quantizations
1 model

Dataset used to train EryriLabs/llama-3.2-uk-legislation-instruct-3b