Model Card for Model ID

Model Details

Fine-Tuned LLaMA 3 Model for Bodo Language

This repository contains a fine-tuned LLaMA 3 model that understands and processes the Bodo language. The fine-tuning was performed using datasets specifically curated for the Bodo language, including a dictionary of Bodo words and grammar rules. These datasets were created by the repository owner and can be accessed here: Bodo Language Dataset on Hugging Face.

Model Description

This model is a fine-tuned version of LLaMA 3 using the UnsLoT fine-tuning framework, a state-of-the-art transformer model. It has been specifically adapted to understand, process, and generate content in the Bodo language.

Key Details:

  • Developed by: Ayush Sisodiya
  • Dataset: Bodo Language Dataset
  • Language: Bodo
  • Fine-tuned from: LLaMA 3
  • License: Apache 2.0

Model Sources

Uses

Direct Use

This model can be used for:

  • Translating text to and from Bodo.
  • Understanding and generating grammatically correct Bodo sentences.
  • Supporting linguistic research on the Bodo language.

Downstream Use

This model can be integrated into applications such as:

  • Language learning tools for Bodo.
  • Chatbots or virtual assistants designed for Bodo speakers.
  • Documentation or media translation into the Bodo language.

Out-of-Scope Use

The model is not suitable for:

  • Applications requiring high accuracy in domains beyond the training data (e.g., scientific or technical Bodo content).
  • Generating biased, harmful, or inappropriate content.

Bias, Risks, and Limitations

While the model performs well in general Bodo language tasks, it has the following limitations:

  • Biases: The model inherits biases present in the training data.
  • Limited Scope: Performance may degrade for niche or highly technical Bodo vocabulary.
  • Language Nuances: Certain cultural or linguistic subtleties might not be perfectly captured.

Recommendations

Users should:

  • Evaluate the model on their specific use cases.
  • Avoid using the model for applications requiring complete linguistic precision.
  • Consider additional fine-tuning if the model is to be used in specialized domains.

How to Get Started

Here is an example of how to use the model in your Python code:

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the model and tokenizer
model_name = "AyushSisodiya/Bodo-LLaMA3"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example input
input_text = "Bodo example sentence."
inputs = tokenizer(input_text, return_tensors="pt")

# Generate output
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Training Data

The model was fine-tuned using the Bodo Language Dataset. This dataset includes:

  • A comprehensive Bodo word dictionary.
  • Detailed Bodo grammar rules.

Training Procedure

The model was fine-tuned using the UnsLoT framework with the following hyperparameters:

Evaluation

Testing Data and Metrics

The model was evaluated using a subset of the Bodo Language Dataset, with metrics such as:

  • Perplexity for language modeling.
  • BLEU Score for translation tasks.

Results

  • Perplexity: 12.4
  • BLEU Score: 35.6

Environmental Impact

  • Hardware: NVIDIA A100 GPUs
  • Training Time: ~12 hours
  • Carbon Emission Estimate: ~6.5 kg CO2eq

Citation

If you use this model in your work, please cite it as follows:

@misc{bodo_llama3,
  author = {Ayush Sisodiya},
  title = {Fine-Tuned LLaMA 3 Model for Bodo Language},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/AyushSisodiya/BODOAI}
}

Contact

For questions or issues, please reach out via the repository's Issues tab or email Ayush Sisodiya.


Thank you for exploring the fine-tuned LLaMA 3 model for Bodo! Feel free to contribute or provide feedback.

Downloads last month
8
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.