Edit model card

Base model: CorticalStack/gemma-7b-ultrachat-sft

This is finetuned from above base model and to be used for multi-turn chat based use-cases. Unlike our AryaBhatta-GemmaOrca model which is skilled in science, literature and finetuned on Orca datasets, this model is fine-tuned on Ultra-Chat datasets. And show improved performance over AryaBhatta-GemmaOrca on Hellaswag datasets and in multi-turn conversations. It is finetuned on 9 Indian languages (Hindi, Tamil, Punjabi, Bengali, Gujarati, Oriya, Telugu, Kannada, Malayalam) plus English.

Benchmarked on Indic LLM leaderboard: https://huggingface.co/spaces/Cognitive-Lab/indic_llm_leaderboard

Release post: https://www.linkedin.com/feed/update/urn:li:activity:7184856055565180928

Downloads last month
4
Safetensors
Model size
8.54B params
Tensor type
FP16
·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.