You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This recent commit 51bd072 and pull request #230 which changes how LR schedulers are saved forgot to add the is_zero arg and set it to config.optimizer.zero_stage in src/nanotron/serialize/main.py:
This recent commit 51bd072 and pull request #230 which changes how LR schedulers are saved forgot to add the
is_zero
arg and set it toconfig.optimizer.zero_stage
insrc/nanotron/serialize/main.py
:nanotron/src/nanotron/serialize/main.py
Lines 106 to 110 in cfcdeae
This arg is expected to be passed:
nanotron/src/nanotron/serialize/optimizer.py
Lines 109 to 114 in cfcdeae
This causes training to crash when LR scheduler is saved.
@TJ-Solergibert @NouamaneTazi
The text was updated successfully, but these errors were encountered: