From df2f84ce619edddb80e720a56abc74d5490fed99 Mon Sep 17 00:00:00 2001 From: Myle Ott Date: Tue, 3 Dec 2019 15:18:07 -0800 Subject: [PATCH] v0.8.0 -> v0.9.0 (#1452) Summary: Possibly breaking changes: - Set global numpy seed (4a7cd58) - Split `in_proj_weight` into separate k, v, q projections in MultiheadAttention (fdf4c3e) - TransformerEncoder returns namedtuples instead of dict (27568a7) New features: - Add `--fast-stat-sync` option (e1ba32a) - Add `--empty-cache-freq` option (315c463) - Support criterions with parameters (ba5f829) New papers: - Simple and Effective Noisy Channel Modeling for Neural Machine Translation (49177c9) - Levenshtein Transformer (86857a5, ...) - Cross+Self-Attention for Transformer Models (4ac2c5f) - Jointly Learning to Align and Translate with Transformer Models (1c66792) - Reducing Transformer Depth on Demand with Structured Dropout (dabbef4) - Unsupervised Cross-lingual Representation Learning at Scale (XLM-RoBERTa) (e23e5ea) - BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (a92bcda) - CamemBERT: a French BERT (b31849a) Speed improvements: - Add CUDA kernels for LightConv and DynamicConv (f840564) - Cythonization of various dataloading components (4fc3953, ...) - Don't project mask tokens for MLM training (718677e) Pull Request resolved: https://github.com/pytorch/fairseq/pull/1452 Differential Revision: D18798409 Pulled By: myleott fbshipit-source-id: 860a0d5aaf7377c8c9bd63cdb3b33d464f0e1727 --- docs/conf.py | 4 ++-- examples/__init__.py | 2 +- fairseq/__init__.py | 2 +- setup.py | 2 +- 4 files changed, 5 insertions(+), 5 deletions(-) diff --git a/docs/conf.py b/docs/conf.py index 11358ca2e..d6d150c1f 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -60,9 +60,9 @@ github_doc_root = 'https://github.com/pytorch/fairseq/tree/master/docs/' # built documents. # # The short X.Y version. -version = '0.8.0' +version = '0.9.0' # The full version, including alpha/beta/rc tags. -release = '0.8.0' +release = '0.9.0' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. diff --git a/examples/__init__.py b/examples/__init__.py index 35b0568cb..9369be1b7 100644 --- a/examples/__init__.py +++ b/examples/__init__.py @@ -3,6 +3,6 @@ # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. -__version__ = '0.8.0' +__version__ = '0.9.0' import examples.noisychannel # noqa diff --git a/fairseq/__init__.py b/fairseq/__init__.py index 08e29ab85..cce33469e 100644 --- a/fairseq/__init__.py +++ b/fairseq/__init__.py @@ -4,7 +4,7 @@ # LICENSE file in the root directory of this source tree. __all__ = ['pdb'] -__version__ = '0.8.0' +__version__ = '0.9.0' import fairseq.criterions # noqa import fairseq.models # noqa diff --git a/setup.py b/setup.py index 3dd400599..98a7bdf00 100644 --- a/setup.py +++ b/setup.py @@ -116,7 +116,7 @@ if 'test' in sys.argv[1:]: setup( name='fairseq', - version='0.8.0', + version='0.9.0', description='Facebook AI Research Sequence-to-Sequence Toolkit', url='https://github.com/pytorch/fairseq', classifiers=[