You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have been developing a high-level library for PyTorch and Accelerate for quite some time.
The reason why I'm doing it is because I'm tooooo lazy to write all the logging and IO stuff, but I still want full control with the training loop (nobody tells me how to train my model.
Accelerate has been the core component of my framework now, some of it were extracted out and merge into accelerate, e.g., #488.
There are still sometimes I find it's hard to go, for example, load_accelerator_state and save_accelerator_state. It would be better if there is an additional get_accelerator_state, set_accelerator_state so that I can wrap it inside my checkpoint.
There were a few more examples in my mind... but unfortunately I forget them now.... I'll come back later.
The text was updated successfully, but these errors were encountered:
Hi,
I have been developing a high-level library for PyTorch and Accelerate for quite some time.
The reason why I'm doing it is because I'm tooooo lazy to write all the logging and IO stuff, but I still want full control with the training loop (nobody tells me how to train my model.
Accelerate has been the core component of my framework now, some of it were extracted out and merge into accelerate, e.g., #488.
There are still sometimes I find it's hard to go, for example,
load_accelerator_state
andsave_accelerator_state
. It would be better if there is an additionalget_accelerator_state
,set_accelerator_state
so that I can wrap it inside my checkpoint.There were a few more examples in my mind... but unfortunately I forget them now.... I'll come back later.
The text was updated successfully, but these errors were encountered: