You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We want to make sure optimizer functionality works for the following: optimizer in vacuum, using a toy model+optimizer pass, and using sth resembling a real-world training scenario.
Proposed testing setup is the following:
Test internal optimizer functionality:
Using hardcoded inputs and states, comparing new optimizer outputs/states to golden.
Test optimizer functionality in a training scenario:
Hardcode weights and grads; do optimizer step, run a fwd pass with updated weights and compare model outputs to golden.
test e2e
Initialize weights in a standard way (xavier or similar), run fwd/bwd passes through model, do optimizer step and check results for the following fwd pass.
We might want to make sure that 3 is stable through a sequence of minibatches (seq length is a hyperparam).
We can create sub-issues for each of these, if necessary.
The text was updated successfully, but these errors were encountered:
vladimirjovanovicTT
changed the title
[Training] Add a set of extensive optimizer tests
[Training] Add an extensive set of optimizer tests
Dec 31, 2024
We want to make sure optimizer functionality works for the following: optimizer in vacuum, using a toy model+optimizer pass, and using sth resembling a real-world training scenario.
Proposed testing setup is the following:
We can create sub-issues for each of these, if necessary.
The text was updated successfully, but these errors were encountered: