fairseq

2025-12-11 0 127


Fairseq(-py) is a sequence modeling toolkit that allows researchers and
developers to train custom models for translation, summarization, language
modeling and other text generation tasks.

We provide reference implementations of various sequence modeling papers:

List of implemented papers

  • Convolutional Neural Networks (CNN)

    • Language Modeling with Gated Convolutional Networks (Dauphin et al., 2017)
    • Convolutional Sequence to Sequence Learning (Gehring et al., 2017)
    • Classical Structured Prediction Losses for Sequence to Sequence Learning (Edunov et al., 2018)
    • Hierarchical Neural Story Generation (Fan et al., 2018)
    • wav2vec: Unsupervised Pre-training for Speech Recognition (Schneider et al., 2019)
  • LightConv and DynamicConv models

    • Pay Less Attention with Lightweight and Dynamic Convolutions (Wu et al., 2019)
  • Long Short-Term Memory (LSTM) networks

    • Effective Approaches to Attention-based Neural Machine Translation (Luong et al., 2015)
  • Transformer (self-attention) networks

    • Attention Is All You Need (Vaswani et al., 2017)
    • Scaling Neural Machine Translation (Ott et al., 2018)
    • Understanding Back-Translation at Scale (Edunov et al., 2018)
    • Adaptive Input Representations for Neural Language Modeling (Baevski and Auli, 2018)
    • Lexically constrained decoding with dynamic beam allocation (Post & Vilar, 2018)
    • Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context (Dai et al., 2019)
    • Adaptive Attention Span in Transformers (Sukhbaatar et al., 2019)
    • Mixture Models for Diverse Machine Translation: Tricks of the Trade (Shen et al., 2019)
    • RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)
    • Facebook FAIR\’s WMT19 News Translation Task Submission (Ng et al., 2019)
    • Jointly Learning to Align and Translate with Transformer Models (Garg et al., 2019)
    • Multilingual Denoising Pre-training for Neural Machine Translation (Liu et at., 2020)
    • Neural Machine Translation with Byte-Level Subwords (Wang et al., 2020)
    • Unsupervised Quality Estimation for Neural Machine Translation (Fomicheva et al., 2020)
    • wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations (Baevski et al., 2020)
    • Generating Medical Reports from Patient-Doctor Conversations Using Sequence-to-Sequence Models (Enarvi et al., 2020)
    • Linformer: Self-Attention with Linear Complexity (Wang et al., 2020)
    • Cross-lingual Retrieval for Iterative Self-Supervised Training (Tran et al., 2020)
    • Deep Transformers with Latent Depth (Li et al., 2020)
    • Unsupervised Cross-lingual Representation Learning for Speech Recognition (Conneau et al., 2020)
    • Self-training and Pre-training are Complementary for Speech Recognition (Xu et al., 2020)
    • Robust wav2vec 2.0: Analyzing Domain Shift in Self-Supervised Pre-Training (Hsu, et al., 2021)
    • Unsupervised Speech Recognition (Baevski, et al., 2021)
    • Simple and Effective Zero-shot Cross-lingual Phoneme Recognition (Xu et al., 2021)
    • VideoCLIP: Contrastive Pre-training for Zero-shot Video-Text Understanding (Xu et. al., 2021)
    • VLM: Task-agnostic Video-Language Model Pre-training for Video Understanding (Xu et. al., 2021)
    • NormFormer: Improved Transformer Pretraining with Extra Normalization (Shleifer et. al, 2021)
  • Non-autoregressive Transformers

    • Non-Autoregressive Neural Machine Translation (Gu et al., 2017)
    • Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement (Lee et al. 2018)
    • Insertion Transformer: Flexible Sequence Generation via Insertion Operations (Stern et al. 2019)
    • Mask-Predict: Parallel Decoding of Conditional Masked Language Models (Ghazvininejad et al., 2019)
    • Levenshtein Transformer (Gu et al., 2019)
  • Finetuning

    • Better Fine-Tuning by Reducing Representational Collapse (Aghajanyan et al. 2020)

What\’s New:

  • May 2023 Released models for Scaling Speech Technology to 1,000+ Languages (Pratap, et al., 2023)
  • June 2022 Released code for wav2vec-U 2.0 from Towards End-to-end Unsupervised Speech Recognition (Liu, et al., 2022)
  • May 2022 Integration with xFormers
  • December 2021 Released Direct speech-to-speech translation code
  • October 2021 Released VideoCLIP and VLM models
  • October 2021 Released multilingual finetuned XLSR-53 model
  • September 2021 master branch renamed to main.
  • July 2021 Released DrNMT code
  • July 2021 Released Robust wav2vec 2.0 model
  • June 2021 Released XLMR-XL and XLMR-XXL models
  • May 2021 Released Unsupervised Speech Recognition code
  • March 2021 Added full parameter and optimizer state sharding + CPU offloading
  • February 2021 Added LASER training code
  • December 2020: Added Adaptive Attention Span code
  • December 2020: GottBERT model and code released
  • November 2020: Adopted the Hydra configuration framework
    • see documentation explaining how to use it for new and existing projects
  • November 2020: fairseq 0.10.0 released
  • October 2020: Added R3F/R4F (Better Fine-Tuning) code
  • October 2020: Deep Transformer with Latent Depth code released
  • October 2020: Added CRISS models and code
Previous updates

  • September 2020: Added Linformer code
  • September 2020: Added pointer-generator networks
  • August 2020: Added lexically constrained decoding
  • August 2020: wav2vec2 models and code released
  • July 2020: Unsupervised Quality Estimation code released
  • May 2020: Follow fairseq on Twitter
  • April 2020: Monotonic Multihead Attention code released
  • April 2020: Quant-Noise code released
  • April 2020: Initial model parallel support and 11B parameters unidirectional LM released
  • March 2020: Byte-level BPE code released
  • February 2020: mBART model and code released
  • February 2020: Added tutorial for back-translation
  • December 2019: fairseq 0.9.0 released
  • November 2019: VizSeq released (a visual analysis toolkit for evaluating fairseq models)
  • November 2019: CamemBERT model and code released
  • November 2019: BART model and code released
  • November 2019: XLM-R models and code released
  • September 2019: Nonautoregressive translation code released
  • August 2019: WMT\’19 models released
  • July 2019: fairseq relicensed under MIT license
  • July 2019: RoBERTa models and code released
  • June 2019: wav2vec models and code released

Features:

  • multi-GPU training on one machine or across multiple machines (data and model parallel)
  • fast generation on both CPU and GPU with multiple search algorithms implemented:
    • beam search
    • Diverse Beam Search (Vijayakumar et al., 2016)
    • sampling (unconstrained, top-k and top-p/nucleus)
    • lexically constrained decoding (Post & Vilar, 2018)
  • gradient accumulation enables training with large mini-batches even on a single GPU
  • mixed precision training (trains faster with less GPU memory on NVIDIA tensor cores)
  • extensible: easily register new models, criterions, tasks, optimizers and learning rate schedulers
  • flexible configuration based on Hydra allowing a combination of code, command-line and file based configuration
  • full parameter and optimizer state sharding
  • offloading parameters to CPU

We also provide pre-trained models for translation and language modeling
with a convenient torch.hub interface:

en2de = torch.hub.load(\'pytorch/fairseq\', \'transformer.wmt19.en-de.single_model\')
en2de.translate(\'Hello world\', beam=5)
# \'Hallo Welt\'

See the PyTorch Hub tutorials for translation
and RoBERTa for more examples.

Requirements and Installation

  • PyTorch version >= 1.10.0
  • Python version >= 3.8
  • For training new models, you\’ll also need an NVIDIA GPU and NCCL
  • To install fairseq and develop locally:
git clone https://gith*ub*.c*om/pytorch/fairseq
cd fairseq
pip install --editable ./

# on MacOS:
# CFLAGS=\"-stdlib=libc++\" pip install --editable ./

# to install the latest stable release (0.10.x)
# pip install fairseq
  • For faster training install NVIDIA\’s apex library:
git clone https://g*i*thub*.com/NVIDIA/apex
cd apex
pip install -v --no-cache-dir --global-option=\"--cpp_ext\" --global-option=\"--cuda_ext\" \\
  --global-option=\"--deprecated_fused_adam\" --global-option=\"--xentropy\" \\
  --global-option=\"--fast_multihead_attn\" ./
  • For large datasets install PyArrow: pip install pyarrow
  • If you use Docker make sure to increase the shared memory size either with --ipc=host or --shm-size
    as command line options to nvidia-docker run .

Getting Started

The full documentation contains instructions
for getting started, training new models and extending fairseq with new model
types and tasks.

Pre-trained models and examples

We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below,
as well as example training and evaluation commands.

  • Translation: convolutional and transformer models are available
  • Language Modeling: convolutional and transformer models are available

We also have more detailed READMEs to reproduce results from specific papers:

  • XLS-R: Self-supervised Cross-lingual Speech Representation Learning at Scale (Babu et al., 2021)
  • Cross-lingual Retrieval for Iterative Self-Supervised Training (Tran et al., 2020)
  • wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations (Baevski et al., 2020)
  • Unsupervised Quality Estimation for Neural Machine Translation (Fomicheva et al., 2020)
  • Training with Quantization Noise for Extreme Model Compression ({Fan*, Stock*} et al., 2020)
  • Neural Machine Translation with Byte-Level Subwords (Wang et al., 2020)
  • Multilingual Denoising Pre-training for Neural Machine Translation (Liu et at., 2020)
  • Reducing Transformer Depth on Demand with Structured Dropout (Fan et al., 2019)
  • Jointly Learning to Align and Translate with Transformer Models (Garg et al., 2019)
  • Levenshtein Transformer (Gu et al., 2019)
  • Facebook FAIR\’s WMT19 News Translation Task Submission (Ng et al., 2019)
  • RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)
  • wav2vec: Unsupervised Pre-training for Speech Recognition (Schneider et al., 2019)
  • Mixture Models for Diverse Machine Translation: Tricks of the Trade (Shen et al., 2019)
  • Pay Less Attention with Lightweight and Dynamic Convolutions (Wu et al., 2019)
  • Understanding Back-Translation at Scale (Edunov et al., 2018)
  • Classical Structured Prediction Losses for Sequence to Sequence Learning (Edunov et al., 2018)
  • Hierarchical Neural Story Generation (Fan et al., 2018)
  • Scaling Neural Machine Translation (Ott et al., 2018)
  • Convolutional Sequence to Sequence Learning (Gehring et al., 2017)
  • Language Modeling with Gated Convolutional Networks (Dauphin et al., 2017)

Join the fairseq community

  • Twitter: https://twit*te**r.com/fairseq
  • Facebook page: https://www.fa*c**ebook.com/groups/fairseq.users
  • Google group: https://groups.g*o*og*le.com/forum/#!forum/fairseq-users

License

fairseq(-py) is MIT-licensed.
The license applies to the pre-trained models as well.

Citation

Please cite as:

@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}

下载源码

通过命令行克隆项目:

git clone https://github.com/facebookresearch/fairseq.git

收藏 (0) 打赏

感谢您的支持,我会继续努力的!

打开微信/支付宝扫一扫,即可进行扫码打赏哦,分享从这里开始,精彩与您同在
点赞 (0)

申明:本文由第三方发布,内容仅代表作者观点,与本网站无关。对本文以及其中全部或者部分内容的真实性、完整性、及时性本站不作任何保证或承诺,请读者仅作参考,并请自行核实相关内容。本网发布或转载文章出于传递更多信息之目的,并不意味着赞同其观点或证实其描述,也不代表本网对其真实性负责。

左子网 编程相关 fairseq https://www.zuozi.net/34510.html

Real ESRGAN
上一篇: Real ESRGAN
netron
下一篇: netron
常见问题
  • 1、自动:拍下后,点击(下载)链接即可下载;2、手动:拍下后,联系卖家发放即可或者联系官方找开发者发货。
查看详情
  • 1、源码默认交易周期:手动发货商品为1-3天,并且用户付款金额将会进入平台担保直到交易完成或者3-7天即可发放,如遇纠纷无限期延长收款金额直至纠纷解决或者退款!;
查看详情
  • 1、描述:源码描述(含标题)与实际源码不一致的(例:货不对板); 2、演示:有演示站时,与实际源码小于95%一致的(但描述中有”不保证完全一样、有变化的可能性”类似显著声明的除外); 3、发货:不发货可无理由退款; 4、安装:免费提供安装服务的源码但卖家不履行的; 5、收费:价格虚标,额外收取其他费用的(但描述中有显著声明或双方交易前有商定的除外); 6、其他:如质量方面的硬性常规问题BUG等。 注:经核实符合上述任一,均支持退款,但卖家予以积极解决问题则除外。
查看详情
  • 1、左子会对双方交易的过程及交易商品的快照进行永久存档,以确保交易的真实、有效、安全! 2、左子无法对如“永久包更新”、“永久技术支持”等类似交易之后的商家承诺做担保,请买家自行鉴别; 3、在源码同时有网站演示与图片演示,且站演与图演不一致时,默认按图演作为纠纷评判依据(特别声明或有商定除外); 4、在没有”无任何正当退款依据”的前提下,商品写有”一旦售出,概不支持退款”等类似的声明,视为无效声明; 5、在未拍下前,双方在QQ上所商定的交易内容,亦可成为纠纷评判依据(商定与描述冲突时,商定为准); 6、因聊天记录可作为纠纷评判依据,故双方联系时,只与对方在左子上所留的QQ、手机号沟通,以防对方不承认自我承诺。 7、虽然交易产生纠纷的几率很小,但一定要保留如聊天记录、手机短信等这样的重要信息,以防产生纠纷时便于左子介入快速处理。
查看详情

相关文章

猜你喜欢
发表评论
暂无评论
官方客服团队

为您解决烦忧 - 24小时在线 专业服务