|
--- |
|
language: |
|
- uz |
|
license: apache-2.0 |
|
tags: |
|
- text-generation-inference |
|
- transformers |
|
- unsloth |
|
- llama |
|
- trl |
|
base_model: llama-3-8b-bnb-4bit |
|
--- |
|
|
|
# Uploaded model |
|
|
|
# Usage model. |
|
|
|
``` |
|
import gradio as gr |
|
from unsloth import FastLanguageModel |
|
|
|
# Load your pre-trained model |
|
max_seq_length = 2048 |
|
dtype = None |
|
load_in_4bit = True |
|
|
|
model, tokenizer = FastLanguageModel.from_pretrained( |
|
model_name="blackhole33/llama-3-8b-bnb-4bit", |
|
max_seq_length=max_seq_length, |
|
dtype=dtype, |
|
load_in_4bit=load_in_4bit, |
|
) |
|
|
|
FastLanguageModel.for_inference(model) # Enable native 2x faster inference |
|
|
|
# Alpaca prompt template |
|
alpaca_prompt = """Quyida vazifani tavsiflovchi ko'rsatma mavjud bo'lib, u qo'shimcha kontekstni ta'minlaydigan kiritish bilan bog'langan. So'rovni to'g'ri to'ldiradigan javob yozing. |
|
|
|
### Instruction: |
|
{} |
|
|
|
### Response: |
|
{}""" |
|
|
|
# Function to generate response |
|
def generate_response(instruction): |
|
inputs = tokenizer( |
|
[ |
|
alpaca_prompt.format( |
|
instruction, # instruction |
|
"" # output - leave this blank for generation! |
|
) |
|
], |
|
return_tensors="pt", |
|
).to("cuda") |
|
|
|
outputs = model.generate(**inputs, max_new_tokens=250, use_cache=True) |
|
res = tokenizer.batch_decode(outputs, skip_special_tokens=True) |
|
return res[0] |
|
|
|
# Gradio interface |
|
interface = gr.Interface( |
|
fn=generate_response, |
|
inputs=[ |
|
gr.Textbox(lines=2, placeholder="Question"), |
|
], |
|
outputs="text", |
|
title="Uzbek Language Model Interface", |
|
description="Enter an instruction and context to get a response from the model.", |
|
) |
|
|
|
# Launch the interface |
|
interface.launch(share=True) |
|
|
|
|
|
``` |
|
|
|
- **Developed by:** blackhole33 |
|
- **License:** apache-2.0 |
|
- **Finetuned from model :** llama-3-8b-bnb-4bit |
|
|