Named entity recognition (NER) is the task of tagging entities in text with their corresponding type. Approaches typically use BIO notation, which differentiates the beginning (B) and the inside (I) of entities. O is used for non-entity tokens.
Example:
Mark | Watney | visited | Mars |
---|---|---|---|
B-PER | I-PER | O | B-LOC |
(NER definition taken from english/named_entity_recognition.md)
The CANTEMIST-NER 2020 task consists of Spanish oncology clinical reports corpus tagged with one entity type (MORFOLOGIA_NEOPLASIA). Models are evaluated based on span-based F1 on the test set: see evaluation scripts.
The CANTEMIST shared task contains as well an entity linking subtrack (CANTEMIST-NORM) and a document indexing subtrack (CANTEMIST-CODING).
Data link: Zenodo
Model | F1 | Paper / Source | Code |
---|---|---|---|
MRC mBERT-MLP (Xiong et al., 2020) | 87.0 | A Joint Model for Medical Named Entity Recognition and Normalization | Official |
BETO-SciBERT (Garcia-Pablos et al., 2020) | 86.9 | Vicomtech at CANTEMIST 2020 | |
BiLSTM-CRF+GloVe+SME+CWE (López-Úbeda et al., 2020) | 85.5 | Extracting Neoplasms Morphology Mentions in Spanish Clinical Cases through Word Embeddings | |
Biaffine Classifier (Lange et al., 2020) | 85.3 | NLNDE at CANTEMIST: Neural Sequence Labeling and Parsing Approaches for Clinical Concept Extraction | |
BETO (Han et al., 2020) | 85.0 | Pre-trained Language Model for CANTEMIST Named Entity Recognition | |
BiLSTM-CRF+FasText+Char (Carreto Fidalgo et al., 2020) | 84.5 | Recognai’s Working Notes for CANTEMIST-NER Track | Official |
BiLSTM-BiLSTM-CRF+FasText+PoS+Char (Santamaria Carrasco et al., 2020) | 83.4 | Using Embeddings and Bi-LSTM+CRF Model to Detect Tumor Morphology Entities in Spanish Clinical Cases | Official |
The ProfNER-NER 2021 task consists of Spanish COVID-19 related Twitter corpus tagged with four entity types (PROFESION,SITUACION_LABORAL,ACTIVIDAD,FIGURATIVA). Models are evaluated based on span AND label-based F1 on the test set: see Task 7 of Codalab SMM4H competition.
The ProfNER shared task contains as well a tweet classification subtrack (ProfNER-Track A).
Data link: Zenodo
Model | F1 | Paper / Source | Code |
---|---|---|---|
BETO-Linear-CRF (David Carreto Fidalgo et al., 2021) | 83.9 | Recognai | Official |
3xBiLSTM-CRF+BPE+FastText+BETOemb (Usama Yaseen et al., 2021) | 82.4 | MIC-NLP | |
BiLSTM-LSTM-CRF+Char+STE+SME+BETO+Syllabes+POS (Sergio Santamaría Carrasco et al., 2021) | 82.3 | Troy | Official |
BiGRU-BiLSTM-TokenClassification-CRF+STE+Char (David Carreto Fidalgo et al., 2021) | 76.4 | Recognai | Official |
BiLSTM-CRF+Char+STE+SME+WikiFastText (Vasile Pais, et al., 2021) | 75.7 | RACAI | |
30xBETO-BiLSTM (Tong Zhou et al., 2021) | 73.3 | CASIA_Unisound | Official |
Dictionaries-CRF (Alberto Mesa Murgado et al., 2021) | 72.8 | SINAI | Official |
BiLSTM-CRF+FLAIR+FastText (Pedro Ruas et al., 2021) | 72.7 | Lasige-BioTM | Official |