Antoine Chaffin commited on
Commit
995dea4
1 Parent(s): 19b6ec4

Adding padding token

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -32,6 +32,7 @@ If a question does not make any sense, or is not factually coherent, explain why
32
  model = AutoModelForCausalLM.from_pretrained(args.model, use_auth_token=hf_token, torch_dtype=torch.float16,
33
  device_map='auto').to(device)
34
  tokenizer = AutoTokenizer.from_pretrained(args.model, use_auth_token=hf_token)
 
35
 
36
  def embed(user, max_length, window_size, method, prompt):
37
  uid = USERS.index(user)
@@ -40,7 +41,8 @@ def embed(user, max_length, window_size, method, prompt):
40
  watermarked_texts = watermarker.embed(key=args.key, messages=[ uid ],
41
  max_length=max_length, method=method, prompt=prompt)
42
  print("watermarked_texts: ", watermarked_texts)
43
-
 
44
  return watermarked_texts[0].replace(prompt, "")
45
 
46
  def detect(attacked_text, window_size, method, prompt):
 
32
  model = AutoModelForCausalLM.from_pretrained(args.model, use_auth_token=hf_token, torch_dtype=torch.float16,
33
  device_map='auto').to(device)
34
  tokenizer = AutoTokenizer.from_pretrained(args.model, use_auth_token=hf_token)
35
+ tokenizer.pad_token = tokenizer.eos_token
36
 
37
  def embed(user, max_length, window_size, method, prompt):
38
  uid = USERS.index(user)
 
41
  watermarked_texts = watermarker.embed(key=args.key, messages=[ uid ],
42
  max_length=max_length, method=method, prompt=prompt)
43
  print("watermarked_texts: ", watermarked_texts)
44
+ print(watermarked_texts[0].replace(prompt, ""))
45
+ watermarked_texts[0].split("[/INST]")[-1]
46
  return watermarked_texts[0].replace(prompt, "")
47
 
48
  def detect(attacked_text, window_size, method, prompt):