You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I clearly wasn't thinking about e-prop when I added support for training with validation in #57 as currently gradients are being accumulated e.g. at https://github.com/genn-team/ml_genn/blob/master/ml_genn/ml_genn/compilers/eprop_compiler.py#L209 during validation and therefore will get applied at the end of the first training batch. In e-prop, the state needs to be fully reset (especially gradients) at the start of each training epoch. I think the cleanest solution is to fully reset state at the beginning of each training epoch. That way eligibility trace/adaptation variable/gradients accumulated during validation won't effect training but model will be fully adapted after a full epoch of training so validation is a bit more realistic
The text was updated successfully, but these errors were encountered:
I still don't understand how to handle adaptation variables though. Resetting makes sense in some ways but it's sub-optimal if our intuition about a "working regime" away from reset values is right?
Yes, it could lead to overestimating validation performance. It would not mean actual overstating of accuracy as long as the test set stays separate. However, how to handle testing on the test set?
I clearly wasn't thinking about e-prop when I added support for training with validation in #57 as currently gradients are being accumulated e.g. at https://github.com/genn-team/ml_genn/blob/master/ml_genn/ml_genn/compilers/eprop_compiler.py#L209 during validation and therefore will get applied at the end of the first training batch. In e-prop, the state needs to be fully reset (especially gradients) at the start of each training epoch. I think the cleanest solution is to fully reset state at the beginning of each training epoch. That way eligibility trace/adaptation variable/gradients accumulated during validation won't effect training but model will be fully adapted after a full epoch of training so validation is a bit more realistic
The text was updated successfully, but these errors were encountered: