Open
Description
For supervised training loops, we provide a convenience output transform which ensures only the model state is returned from the training loop. This means you always lose the entire training state, which might be of interest later on.
I propose instead that we return a tuple:
{loop_state, transformed_state}
which always returns the whole state, as well as a transformed version. That way you never accidentally lose the entire state.