BART

Year: 2,019
Journal: Association for Computational Linguistics
Languages: English
Programming languages: Python
Input data:

Plain text

We present BART, a denoising autoencoder for pretraining sequence-to-sequence models. BART is trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text. It uses a standard Tranformer-based neural machine translation architecture which, despite its simplicity, can be seen as generalizing BERT (due to the bidirectional encoder), GPT (with the left-to-right decoder), and many other more recent pretraining schemes.

Sign In

Register

Reset Password

Please enter your username or email address, you will receive a link to create a new password via email.