anudeepadi commited on
Commit
c01a7ad
1 Parent(s): dcf66ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -22
README.md CHANGED
@@ -1,22 +1,41 @@
1
- ---
2
- language:
3
- - en
4
- license: apache-2.0
5
- tags:
6
- - text-generation-inference
7
- - transformers
8
- - unsloth
9
- - gemma
10
- - trl
11
- base_model: unsloth/gemma-7b-bnb-4bit
12
- ---
13
-
14
- # Uploaded model
15
-
16
- - **Developed by:** anudeepadi
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** unsloth/gemma-7b-bnb-4bit
19
-
20
- This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
-
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Telugu Gemma 7B Model
2
+ This is a Telugu language model based on the Gemma 7B model, finetuned on a dataset of Telugu instructions and outputs using the Alpaca format. The model was trained 2x faster using Unsloth and Huggingface's TRL library.
3
+
4
+ Model Details
5
+ Developed by: Anudeep Adi
6
+ Model architecture: Gemma 7B
7
+ Language: Telugu
8
+ License: Apache 2.0
9
+ Base model: unsloth/gemma-7b-bnb-4bit
10
+ Dataset: telugu_teknium_GPTeacher_general_instruct_filtered_romanized
11
+ Finetuning steps: 60
12
+ Finetuning dataset size: 43,614 examples
13
+ Tags: text-generation-inference, transformers, unsloth, gemma, trl
14
+ Usage
15
+ The model can be used for Telugu language text generation tasks following an instruction and optional input prompt in the Alpaca format:
16
+
17
+
18
+ Copy code
19
+ Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
20
+
21
+ ### Instruction:
22
+ {instruction}
23
+
24
+ ### Input:
25
+ {input}
26
+
27
+ ### Response:
28
+ {output}
29
+ Replace {instruction}, {input} and {output} with the relevant text. The model will generate a continuation for the {output} field.
30
+
31
+ Training Procedure
32
+ The model was finetuned on the telugu_teknium_GPTeacher_general_instruct_filtered_romanized dataset which contains 43,614 examples of Telugu instructions and outputs in the Alpaca format.
33
+
34
+ Finetuning was performed for 60 steps using Unsloth and the TRL library from Huggingface to enable 2x faster training. LoRA adapters were added to enable efficient finetuning of the 7B parameter model. Mixed precision training with bfloat16 was used.
35
+
36
+ Unsloth also provides optimized inference code to run the model 2x faster.
37
+
38
+ <img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>
39
+
40
+ Limitations
41
+ As the model was trained on a relatively small Telugu dataset, its knowledge and conversational abilities in Telugu are limited compared to larger language models. It may make factual errors or inconsistent statements. The model should be used as an experimental prototype only.