Year: 2,019
Languages: Dutch
Programming languages: Python
Input data:


The transformer-based pre-trained language model BERT has helped to improve state-of-the-art performance on many natural language processing (NLP) tasks. Using the same architecture and parameters, we developed and evaluated a monolingual Dutch BERT model called BERTje. Compared to the multilingual BERT model, which includes Dutch but is only based on Wikipedia text, BERTje is based on a large and diverse dataset of 2.4 billion tokens.

Sign In


Reset Password

Please enter your username or email address, you will receive a link to create a new password via email.