Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Training] Add an extensive set of optimizer tests #993

Open
vladimirjovanovicTT opened this issue Dec 31, 2024 · 0 comments
Open

[Training] Add an extensive set of optimizer tests #993

vladimirjovanovicTT opened this issue Dec 31, 2024 · 0 comments
Assignees

Comments

@vladimirjovanovicTT
Copy link
Contributor

vladimirjovanovicTT commented Dec 31, 2024

We want to make sure optimizer functionality works for the following: optimizer in vacuum, using a toy model+optimizer pass, and using sth resembling a real-world training scenario.

Proposed testing setup is the following:

  1. Test internal optimizer functionality:
  • Using hardcoded inputs and states, comparing new optimizer outputs/states to golden.
  1. Test optimizer functionality in a training scenario:
  • Hardcode weights and grads; do optimizer step, run a fwd pass with updated weights and compare model outputs to golden.
  1. test e2e
  • Initialize weights in a standard way (xavier or similar), run fwd/bwd passes through model, do optimizer step and check results for the following fwd pass.
  1. We might want to make sure that 3 is stable through a sequence of minibatches (seq length is a hyperparam).

We can create sub-issues for each of these, if necessary.

@vladimirjovanovicTT vladimirjovanovicTT self-assigned this Dec 31, 2024
@vladimirjovanovicTT vladimirjovanovicTT changed the title [Training] Add a set of extensive optimizer tests [Training] Add an extensive set of optimizer tests Dec 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant