Spaces:
Sleeping
Sleeping
File size: 2,174 Bytes
e2fac8d 6293678 29e0785 6293678 e2fac8d ca0aa0f e2fac8d 42e6f8d 52b129d 42e6f8d e2fac8d 83fe2ae e2fac8d ca0aa0f e2fac8d 83fe2ae e2fac8d ca0aa0f e2fac8d 29e0785 e2fac8d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 |
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
import gradio as gr
import spaces
model_id = "meta-llama/Llama-Guard-3-8B-INT8"
device = "cuda" if torch.cuda.is_available() else "cpu"
dtype = torch.bfloat16
quantization_config = BitsAndBytesConfig(load_in_8bit=True)
def load_model():
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=dtype,
device_map="auto",
quantization_config=quantization_config,
low_cpu_mem_usage=True
)
return tokenizer, model
tokenizer, model = load_model()
@spaces.GPU
def moderate(user_input, assistant_response):
chat = [
{"role": "user", "content": user_input},
{"role": "assistant", "content": assistant_response},
]
input_ids = tokenizer.apply_chat_template(chat, return_tensors="pt").to(device)
with torch.no_grad():
output = model.generate(
input_ids=input_ids,
max_new_tokens=200,
pad_token_id=tokenizer.eos_token_id,
do_sample=False
)
result = tokenizer.decode(output[0], skip_special_tokens=True)
result = result.split(assistant_response)[-1].strip()
is_safe = "safe" in result.lower()
categories = []
if not is_safe and "categories:" in result:
categories = [cat.strip() for cat in result.split("categories:")[1].split(",") if cat.strip()]
return {
"is_safe": "Safe" if is_safe else "Unsafe",
"categories": ", ".join(categories) if categories else "None",
"raw_output": result
}
iface = gr.Interface(
fn=moderate,
inputs=[
gr.Textbox(lines=3, label="User Input"),
gr.Textbox(lines=3, label="Assistant Response")
],
outputs=[
gr.Textbox(label="Safety Status"),
gr.Textbox(label="Violated Categories"),
gr.Textbox(label="Raw Output")
],
title="Llama Guard Moderation",
description="Enter a user input and an assistant response to check for content moderation."
)
if __name__ == "__main__":
iface.launch() |