diff --git a/docs/data.rst b/docs/data.rst index a2a464ec..6a390cb3 100644 --- a/docs/data.rst +++ b/docs/data.rst @@ -30,6 +30,8 @@ provide additional functionality: :members: .. autoclass:: fairseq.data.ConcatDataset :members: +.. autoclass:: fairseq.data.ResamplingDataset + :members: .. autoclass:: fairseq.data.RoundRobinZipDatasets :members: .. autoclass:: fairseq.data.TransformEosDataset diff --git a/fairseq/optim/adam.py b/fairseq/optim/adam.py index 80de7f00..e60a7db5 100644 --- a/fairseq/optim/adam.py +++ b/fairseq/optim/adam.py @@ -15,6 +15,12 @@ from . import FairseqOptimizer, register_optimizer @register_optimizer('adam') class FairseqAdam(FairseqOptimizer): + """Adam optimizer for fairseq. + + Important note: this optimizer corresponds to the "AdamW" variant of + Adam in its weight decay behavior. As such, it is most closely + analogous to torch.optim.AdamW from PyTorch. + """ def __init__(self, args, params): super().__init__(args)