In the following some pre-trained models are provided for different common sequence tagging tasks. These models can be used by executing:
python RunModel.py modelname.h5 input.txt
For the English models, we used the word embeddings by Komninos et al.. For the German files, we used the word embeddings by Reimers et al..
We trained POS-tagger on the Universal Dependencies v1.3 dataset: Trained on universal dependencies v1.3 Englisch:
Language | Development (Accuracy) | Test (Accuracy) |
---|---|---|
English (UD) | 95.58% | 95.58% |
German (UD) | 94.50% | 93.88% |
Further, we trained models on the Wall Street Journal:
Language | Development (Accuracy) | Test (Accuracy) |
---|---|---|
English (WSJ) | 97.33% | 97.39% |
The depicted performance is accuracy.
Trained on CoNLL 2000 Chunking dataset. Performance is F1-score.
Language | Development (F1) | Test(F1) |
---|---|---|
English (CoNLL 2003) | 95.30% | 94.71% |
Trained on CoNLL 2003 and GermEval 2014
Language | Development (F1) | Test (F1) |
---|---|---|
English (CoNLL 2003) | 93.87% | 90.22% |
German (CoNLL 2003) | 81.15% | 77.70% |
German (GermEval 2014) | 80.93% | 78.94% |
Trained on ACE 2005 (https://catalog.ldc.upenn.edu/LDC2006T06)
Language | Development (F1) | Test (F1) |
---|---|---|
English | 83.93% | 85.68% |
Trained on TempEval3 (https://www.cs.york.ac.uk/semeval-2013/task1/)
Language | Development (F1) | Test (F1) |
---|---|---|
English | - | 83.45% |