Skip to content

Question regarding overfitting a single batch and the weight-tying scheme #61

Answered by qaflan
pavanpreet-gandhi asked this question in Q&A
Discussion options

You must be logged in to vote

I believe weight sharing is a form of regularization and for that reason it reduces overfitting. Weight sharing reduces the number of effective learnable parameters (which reduces overfitting) and also imposes constraints on the network weights which is what regularization methods do to prevent overfitting.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@pavanpreet-gandhi
Comment options

Answer selected by pavanpreet-gandhi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants