Megatron-LM
Year: 2,020
Languages: All Languages
Programming languages: Python
Input data:
sentences
Project website: https://github.com/NVIDIA/Megatron-LM
In this work, we present our techniques for training very large transformer models and implement a simple, efficient intra-layer model parallel approach that enables training transformer models with billions of parameters.