Model Card for Model ID

TAROT-DPO

Task-Oriented Authorship Obfuscation Using Policy Optimization Methods

Fine-tuned text rewriting model with direct preference optimization for authorship obfuscation.

ArXiv paper: https://arxiv.org/abs/2407.21630v1

Model description

Example use

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("gabrielloiseau/TAROT-DPO")
model = AutoModelForCausalLM.from_pretrained("gabrielloiseau/TAROT-DPO")

paragraph = """I had dinner at Bella's Bistro last night, and it was a delightful experience. 
As soon as I walked in, I was greeted warmly by the hostess, and the cozy, rustic decor made me feel right at home. 
I started with the bruschetta, which was so fresh and flavorful—I could have eaten a whole meal of just that!"""

inputs = tokenizer([paragraph + "<|endoftext|>"], return_tensors="pt", padding=True)
outputs = model.generate(**inputs, do_sample=True, max_new_tokens=128)

outputs = outputs[:, inputs["input_ids"].shape[1]:]
tokenizer.batch_decode(outputs,skip_special_tokens=True)
Downloads last month
24
Safetensors
Model size
355M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for gabrielloiseau/TAROT-DPO

Finetuned
(2)
this model

Dataset used to train gabrielloiseau/TAROT-DPO

Collection including gabrielloiseau/TAROT-DPO