Running forever

#103
by guptaru1 - opened
tokenizer = AutoTokenizer.from_pretrained(
            "google/{}".format(model),
            use_fast=False,
        )
generator = AutoModelForCausalLM.from_pretrained(
            "google/{}".format(model),
            torch_dtype=torch.float16,
            device_map="auto",
        )

user_prompt = "can you provide future suggestions for my a/b test and provide a summary for this a/b test?"
prompt = f"<bos><start_of_turn>user\nPlease respond to user questions based on the analysis of this a/b test experiment.\n\n{system_prompt}\n\n{user_prompt}<end_of_turn>\n<start_of_turn>model\n"
    
token_ids = tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt")
print("starting this")
with torch.no_grad():
        output_ids = generator.generate(
            token_ids.to(generator.device),
            max_new_tokens=50,
            do_sample=True,
            temperature=0.7,
            top_p=1.0,
            pad_token_id=tokenizer.pad_token_id,
            bos_token_id=tokenizer.bos_token_id,
            eos_token_id=tokenizer.eos_token_id,
        )
print("fnished this")
response = tokenizer.decode(output_ids.tolist()[0][token_ids.size(1):])
print("My response from gemma")
print(response)

This was running fine on my macbook but as I am trying to run this on dell, I am having some issues. Any suggestions?

Sign up or log in to comment