Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove fairseq2 AdamW #1084

Merged
merged 1 commit into from
Mar 17, 2025
Merged

Remove fairseq2 AdamW #1084

merged 1 commit into from
Mar 17, 2025

Conversation

cbalioglu
Copy link
Contributor

This PR removes the AdamW optimizer from fairseq2. The main reason of having our own AdamW optimizer in fairseq2 (which was essentially using the same adamw functional API as the PyTorch version) was to parity check fairseq's memory efficient fp16 optimizer. Since nowadays AMP and FSDP mixed precision training have become ubiquitous with almost always better model accuracy and minimal impact on memory, there is no need to maintain our own version. (Beyond fp16 becoming more and more obsolete compared to bf16 and lower-precision data types)

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Mar 17, 2025
@cbalioglu cbalioglu merged commit 579e389 into main Mar 17, 2025
15 checks passed
@cbalioglu cbalioglu deleted the adam branch March 17, 2025 17:34
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants