Fairseq(-py) is a sequence modeling toolkit that allows researchers and
developers to train custom models for translation, summarization, language
modeling and other text generation tasks.
We provide reference implementations of various sequence modeling papers:
For large datasets install PyArrow: pip install pyarrow
If you use Docker make sure to increase the shared memory size either with --ipc=host or --shm-size
as command line options to nvidia-docker run .
Getting Started
The full documentation contains instructions
for getting started, training new models and extending fairseq with new model
types and tasks.
Pre-trained models and examples
We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below,
as well as example training and evaluation commands.
Translation: convolutional and transformer models are available
Language Modeling: convolutional and transformer models are available
We also have more detailed READMEs to reproduce results from specific papers:
fairseq(-py) is MIT-licensed.
The license applies to the pre-trained models as well.
Citation
Please cite as:
@inproceedings{ott2019fairseq,
title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
year = {2019},
}
Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks.
We provide reference implementations of various sequence modeling papers:
List of implemented papers
What’s New:
masterbranch renamed tomain.Previous updates
Features:
We also provide pre-trained models for translation and language modeling with a convenient
torch.hubinterface:See the PyTorch Hub tutorials for translation and RoBERTa for more examples.
Requirements and Installation
pip install pyarrow--ipc=hostor--shm-sizeas command line options tonvidia-docker run.Getting Started
The full documentation contains instructions for getting started, training new models and extending fairseq with new model types and tasks.
Pre-trained models and examples
We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below, as well as example training and evaluation commands.
We also have more detailed READMEs to reproduce results from specific papers:
Join the fairseq community
License
fairseq(-py) is MIT-licensed. The license applies to the pre-trained models as well.
Citation
Please cite as: