Use this inside Hugging face API
#4
by
joncc
- opened
Newbie question - how do I load the tokenizer and ".pt" model into a Huggingface pipeline using the HF API?
I've been trying stuff like "
model = LlamaForCausalLM.from_pretrained("./alpaca7b-4bit.pt")
and
tokenizer = LlamaTokenizer.from_pretrained("./tokenizer.model")
But no success
This is a very old quant at this point, I don't recommend using it. Take a look at TheBloke's work isntead.
ozcur
changed discussion status to
closed
Thanks for the reply