Bangla Llama-2 7B Base v0.1 [pre-trained][Llama2 Original Tokenizer]

Welcome to the inaugural release of the Bangla Llama-2 7B base model, an important step in advancing LLMs for the Bangla language. This model is ready for immediate inference and primed for further fine-tuning to suit your specific NLP tasks.

Please Note: This model, labelled as a foundational Bangla Language Model (LLM), is designed primarily for Causal Language Modeling (LM) purposes.

Model description

This Bangla Llama model has been built upon the foundation set by the original Llama-2 with Llama-2 original tokenizer.

  • Model type: A 7B parameter model for Causal LM pre-trained on Bangla 2B+ BERT dataset.
  • Language(s): Bangla and English
  • License: GNU General Public License v3.0
  • Source Model: meta-llama/Llama-2-7b-hf
  • Training Precision: float16
  • Code: GitHub

Usage Note

It's important to note that the models have not undergone detoxification. Therefore, while they possess impressive linguistic capabilities, there is a possibility for them to generate content that could be deemed harmful or offensive. We urge users to exercise discretion and supervise the model's outputs closely, especially in public or sensitive applications.

Meet the Developers

Get to know the creators behind this innovative model and follow their contributions to the field:

Citation

We hope this model serves as a valuable tool in your NLP toolkit and look forward to seeing the advancements it will enable in understanding and producing the Bangla language.

Downloads last month
43
Safetensors
Model size
6.74B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for meherajj/Bangla-Llama-2-7b-base-v0.1

Finetuned
(607)
this model
Quantizations
1 model