You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For supervised training loops, we provide a convenience output transform which ensures only the model state is returned from the training loop. This means you always lose the entire training state, which might be of interest later on.
I propose instead that we return a tuple:
{loop_state, transformed_state} which always returns the whole state, as well as a transformed version. That way you never accidentally lose the entire state.
The text was updated successfully, but these errors were encountered:
My suggestion is to get rid of output_transform altogether. After all, anyone can transform the output by piping an operation after it.
👍 for the trainer returning {model_state, loop_state} though. The user can even access other metadata inside state.step_state. If you want, you can even add more structure by defining a TrainerStep struct which you then place it as the step_state.
Perhaps it is best to do these changes sooner than later, since they are breaking?
For supervised training loops, we provide a convenience output transform which ensures only the model state is returned from the training loop. This means you always lose the entire training state, which might be of interest later on.
I propose instead that we return a tuple:
{loop_state, transformed_state}
which always returns the whole state, as well as a transformed version. That way you never accidentally lose the entire state.The text was updated successfully, but these errors were encountered: