Using a pre-trained model from another framework as base #850
-
I'm interested in creating a structure where I can add online learning capabilities to any existing model (EM for short). The EM could have been trained in any framework (eg. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
Hey @ganesh-krishnan, sorry for the slow reply. It's actually pretty use-case dependent, I don't see a super generic way here when implementations don't follow common standards. When transferring a model state from a framework to another, implementation details from both frameworks matter a lot. There is some leg work to do to first to check if implementations are compatible and how to make the conversion. Transformers of your training pipeline matter too. If you have a standard scaler for instance, you probably want to transfer its state too. If you're willing to describe more the context and participate in the solution implementation we could definitely give a hand. If this ends up in a clean solution for your use-case, I think it would make sense to be added in river-extra as a first step. |
Beta Was this translation helpful? Give feedback.
-
@gbolmier Thank you for the response. We are talking about two different things I think. I'm not talking about transferring state which like you pointed out will be very specific to individual frameworks. I'm talking about something like this:
Let me know if that makes sense. This is of course, one way of adding online learning capabilities to any black box offline model. Wondering if there are other ways. |
Beta Was this translation helpful? Give feedback.
Hey @ganesh-krishnan, sorry for the slow reply. It's actually pretty use-case dependent, I don't see a super generic way here when implementations don't follow common standards.
When transferring a model state from a framework to another, implementation details from both frameworks matter a lot. There is some leg work to do to first to check if implementations are compatible and how to make the conversion. Transformers of your training pipeline matter too. If you have a standard scaler for instance, you probably want to transfer its state too.
If you're willing to describe more the context and participate in the solution implementation we could definitely give a hand. If this ends up in a…