It uses the super performant CPU-only models to calculate semantic similarity fully client-side based on Excel or CSV tables.
- App: https://do-me.github.io/semantic-similarity-table/
- Code: https://github.com/do-me/semantic-similarity-table
SentenceTransformer("all-MiniLM-L6-v2", backend="onnx")
. Does your model not have an ONNX or OpenVINO file yet? No worries - it'll be autoexported for you. Thank me later πfrom_model2vec
or with from_distillation
where you do the distillation yourself. It'll only take 5 seconds on GPU & 2 minutes on CPU, no dataset needed.This is absolutely amazing, the speedup is so insane and makes on-device AI much more accessible. Thank you so so much for this!
It would be great to have some kind of "auto" mode for the device param so that devices supporting webGPU use it right away. Anyway, happily waiting for the docs/blog :)
Thanks a lot for your answer, this is confusing. Apparently the other=feature-extraction
covers all pipeline_tag=feature-extraction
as well. There are many popular models tagged in the same way, like https://huggingface.co/Snowflake/snowflake-arctic-embed-xs, which might remain in the dark if you're looking for them this way.
It's 137 vs. 159 models which makes a big difference! It seems indeed that this is the model's authors choice where to tag but it rather seems a mistake here. Maybe HF might want to improve this UI-wise?
fyi @thenlper