wiki:NerDataset

Version 7 (modified by xnovot32@fi.muni.cz, 20 měsíci ago) (diff)

--

A Human-Annotated Dataset for Language Modeling and Named Entity Recognition in Medieval Documents

This is an open dataset of sentences from 19th and 20th century letterpress reprints of documents from the Hussite era. The dataset contains human annotations for named entity recognition (NER).

You can download the dataset in the LINDAT/CLARIAH-CZ repository.

Contents

The dataset is stored in archive ner-dataset.zip (1.7 GB) with following structure:

  • 8 files named dataset_mlm_*.txt that contain sentences for unsupervised training of language models.
    We used the following three variables to produce the different files:
    1. The sentences are extracted from book OCR texts and may therefore span several pages.
      However, page boundaries contain pollutants such as running heads, footnotes, and page numbers.
      We either allow the sentences to cross page boundaries (all) or not (non-crossing).
    2. The sentences come from all book pages (all) or just those considered relevant by expert annotators (only-relevant).
    3. We split the sentences roughly into 90% for training (training) and 10% for validation (validation).
  • 16 tuples of files named dataset_ner_*.sentences.txt, .ner_tags.txt, and in two cases also .docx.
    These contain sentences and NER tags for supervised training, validation, and testing of language models.
    The .docx files are authored by human annotators and may contain extra details missing from files .sentences.txt and .ner_tags.txt.
    Here are the five variables that we used to produce the different files:
    1. The sentences may originate from book OCR texts using information retrieval techniques (fuzzy-regex or manatee).
      The sentences may also originate from regests (regests) or both books and regests (fuzzy-regex+regests and fuzzy-regex+manatee).
    2. When sentences originate from book OCR texts, they may span several pages of a book.
      However, page boundaries contain pollutants such as running heads, footnotes, and page numbers.
      We either allow the sentences to cross page boundaries (all) or not (non-crossing).
    3. When sentences originate from book OCR texts, they may come from book pages of different relevance.
      We either use sentences from all book pages (all) or just those considered relevant by expert annotators (only-relevant).
    4. When sentences and NER tags originate from book OCR texts using information retrieval techniques, many entities in the sentences may lack tags.
      Therefore, we also provide NER tags completed by language models (automatically_tagged) and human annotators (tagged).
    5. We split the sentences roughly into 80% for training (training), 10% for validation (validation), and 10% for testing (testing).
      For repeated testing, we subdivide the testing split (testing_001-400 and testing_401-500).

Citing

If you use our dataset in your work, please cite the following article:

TODO

If you use LaTeX, you can use the following BibTeX entry:

TODO

Acknowledgements

This work was funded by TAČR Éta, project number TL03000365.