--- license: mit language: - nl --- # GysBERT v1 This model is a Historical Language Model for Dutch coming from the [MacBERTh project](https://macberth.netlify.app/). The architecture is based on BERT base uncased from the original BERT pre-training codebase. The training material comes mostly from the DBNL and the Delpher newspaper dump. The details can be found in the accompanying publication: [Non-Parametric Word Sense Disambiguation for Historical Languages](https://aclanthology.org/2022.nlp4dh-1.16.pdf) The model has been successfully tested on Word Sense Disambiguation tasks as discussed in the referenced paper above. An updated version with an enlarged pre-training dataset is due soon.