metadata
			tags:
  - spacy
  - token-classification
language:
  - en
license: mit
model-index:
  - name: en_core_web_lg
    results:
      - task:
          name: NER
          type: token-classification
        metrics:
          - name: NER Precision
            type: precision
            value: 0.8516398746
          - name: NER Recall
            type: recall
            value: 0.8569711538
          - name: NER F Score
            type: f_score
            value: 0.8542971968
      - task:
          name: TAG
          type: token-classification
        metrics:
          - name: TAG (XPOS) Accuracy
            type: accuracy
            value: 0.9734810915
      - task:
          name: UNLABELED_DEPENDENCIES
          type: token-classification
        metrics:
          - name: Unlabeled Attachment Score (UAS)
            type: f_score
            value: 0.9208198801
      - task:
          name: LABELED_DEPENDENCIES
          type: token-classification
        metrics:
          - name: Labeled Attachment Score (LAS)
            type: f_score
            value: 0.9027174273
      - task:
          name: SENTS
          type: token-classification
        metrics:
          - name: Sentences F-Score
            type: f_score
            value: 0.907098331
Details: https://spacy.io/models/en#en_core_web_lg
English pipeline optimized for CPU. Components: tok2vec, tagger, parser, senter, ner, attribute_ruler, lemmatizer.
| Feature | Description | 
|---|---|
| Name | en_core_web_lg | 
| Version | 3.7.1 | 
| spaCy | >=3.7.2,<3.8.0 | 
| Default Pipeline | tok2vec, tagger, parser, attribute_ruler, lemmatizer, ner | 
| Components | tok2vec, tagger, parser, senter, attribute_ruler, lemmatizer, ner | 
| Vectors | 514157 keys, 514157 unique vectors (300 dimensions) | 
| Sources | OntoNotes 5 (Ralph Weischedel, Martha Palmer, Mitchell Marcus, Eduard Hovy, Sameer Pradhan, Lance Ramshaw, Nianwen Xue, Ann Taylor, Jeff Kaufman, Michelle Franchini, Mohammed El-Bachouti, Robert Belvin, Ann Houston) ClearNLP Constituent-to-Dependency Conversion (Emory University) WordNet 3.0 (Princeton University) Explosion Vectors (OSCAR 2109 + Wikipedia + OpenSubtitles + WMT News Crawl) (Explosion)  | 
| License | MIT | 
| Author | Explosion | 
Label Scheme
View label scheme (113 labels for 3 components)
| Component | Labels | 
|---|---|
tagger | 
$, '', ,, -LRB-, -RRB-, ., :, ADD, AFX, CC, CD, DT, EX, FW, HYPH, IN, JJ, JJR, JJS, LS, MD, NFP, NN, NNP, NNPS, NNS, PDT, POS, PRP, PRP$, RB, RBR, RBS, RP, SYM, TO, UH, VB, VBD, VBG, VBN, VBP, VBZ, WDT, WP, WP$, WRB, XX, _SP, ```` | 
parser | 
ROOT, acl, acomp, advcl, advmod, agent, amod, appos, attr, aux, auxpass, case, cc, ccomp, compound, conj, csubj, csubjpass, dative, dep, det, dobj, expl, intj, mark, meta, neg, nmod, npadvmod, nsubj, nsubjpass, nummod, oprd, parataxis, pcomp, pobj, poss, preconj, predet, prep, prt, punct, quantmod, relcl, xcomp | 
ner | 
CARDINAL, DATE, EVENT, FAC, GPE, LANGUAGE, LAW, LOC, MONEY, NORP, ORDINAL, ORG, PERCENT, PERSON, PRODUCT, QUANTITY, TIME, WORK_OF_ART | 
Accuracy
| Type | Score | 
|---|---|
TOKEN_ACC | 
99.86 | 
TOKEN_P | 
99.57 | 
TOKEN_R | 
99.58 | 
TOKEN_F | 
99.57 | 
TAG_ACC | 
97.35 | 
SENTS_P | 
92.19 | 
SENTS_R | 
89.27 | 
SENTS_F | 
90.71 | 
DEP_UAS | 
92.08 | 
DEP_LAS | 
90.27 | 
ENTS_P | 
85.16 | 
ENTS_R | 
85.70 | 
ENTS_F | 
85.43 |