English pipeline optimized for CPU. Components: tok2vec, tagger, parser, senter, ner, attribute_ruler, lemmatizer.
Feature | Description |
---|---|
Name | en_food_entity_extractor_v2 |
Version | 3.4.1 |
spaCy | >=3.4.0,<3.5.0 |
Default Pipeline | tok2vec , tagger , parser , attribute_ruler , lemmatizer , ner |
Components | tok2vec , tagger , parser , senter , attribute_ruler , lemmatizer , ner |
Vectors | 514157 keys, 514157 unique vectors (300 dimensions) |
Sources | OntoNotes 5 (Ralph Weischedel, Martha Palmer, Mitchell Marcus, Eduard Hovy, Sameer Pradhan, Lance Ramshaw, Nianwen Xue, Ann Taylor, Jeff Kaufman, Michelle Franchini, Mohammed El-Bachouti, Robert Belvin, Ann Houston) ClearNLP Constituent-to-Dependency Conversion (Emory University) WordNet 3.0 (Princeton University) Explosion Vectors (OSCAR 2109 + Wikipedia + OpenSubtitles + WMT News Crawl) (Explosion) |
License | MIT |
Author | Explosion |
Label Scheme
View label scheme (114 labels for 3 components)
Component | Labels |
---|---|
tagger |
$ , '' , , , -LRB- , -RRB- , . , : , ADD , AFX , CC , CD , DT , EX , FW , HYPH , IN , JJ , JJR , JJS , LS , MD , NFP , NN , NNP , NNPS , NNS , PDT , POS , PRP , PRP$ , RB , RBR , RBS , RP , SYM , TO , UH , VB , VBD , VBG , VBN , VBP , VBZ , WDT , WP , WP$ , WRB , XX , _SP , ```` |
parser |
ROOT , acl , acomp , advcl , advmod , agent , amod , appos , attr , aux , auxpass , case , cc , ccomp , compound , conj , csubj , csubjpass , dative , dep , det , dobj , expl , intj , mark , meta , neg , nmod , npadvmod , nsubj , nsubjpass , nummod , oprd , parataxis , pcomp , pobj , poss , preconj , predet , prep , prt , punct , quantmod , relcl , xcomp |
ner |
CARDINAL , DATE , EVENT , FAC , FOOD , GPE , LANGUAGE , LAW , LOC , MONEY , NORP , ORDINAL , ORG , PERCENT , PERSON , PRODUCT , QUANTITY , TIME , WORK_OF_ART |
Accuracy
Type | Score |
---|---|
TOKEN_ACC |
99.93 |
TOKEN_P |
99.57 |
TOKEN_R |
99.58 |
TOKEN_F |
99.57 |
TAG_ACC |
97.34 |
SENTS_P |
91.79 |
SENTS_R |
89.14 |
SENTS_F |
90.44 |
DEP_UAS |
92.04 |
DEP_LAS |
90.23 |
ENTS_P |
85.35 |
ENTS_R |
85.93 |
ENTS_F |
85.64 |
- Downloads last month
- 3
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Evaluation results
- NER Precisionself-reported0.854
- NER Recallself-reported0.859
- NER F Scoreself-reported0.856
- TAG (XPOS) Accuracyself-reported0.973
- Unlabeled Attachment Score (UAS)self-reported0.920
- Labeled Attachment Score (LAS)self-reported0.902
- Sentences F-Scoreself-reported0.904