Error: Can't load tokenizer using from_pretrained

#359
by lomi280 - opened

Hello!

First of all, thank you so much for this amazing work! I am new to learning how models works and how to start using them so I tried using the Inference API and the space from "jeffyang123" (https://huggingface.co/spaces/jeffyang123/ctheodoris-Geneformer) to have an idea of how to run it. I downloaded the code, and files to run it locally too but unfortunately , I keep having the following error of not being able to find the tokenizer forBertTokenizerFast .

**
Something went wrong
"Could not complete request to HuggingFace API, Status Code: 400, Error: Can't load tokenizer using from_pretrained, please update its configuration: Can't load tokenizer for 'ctheodoris/Geneformer'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'ctheodoris/Geneformer' is the correct path to a directory containing all relevant files for a BertTokenizerFast tokenizer."
**

I appreciate any guidance for this, thank you so much in advance.

Thank you for your question! We provide examples in the examples folder in this repository that you are welcome to test out. If you would like to use the space you mentioned, I would ask the user who created that space. This space is not related to our lab and we don't have any information about it. Thank you!

ctheodoris changed discussion status to closed

Hi!

Thank you for your prompt reply! This message also occurs when I use it on the Inference API

image.png

Any clues about how to solve this?
Thank you!

That inference API is automatically placed there by Hugging Face and not something we placed there, nor something we can remove. The example in the inference API is natural language text so it is my assumption that this API is built for natural language models.

I see, thank you so much :)

Sign up or log in to comment