NetVLAD first proposed a VLAD layer trained with triplet
loss, and then SARE introduced two softmax-based losses (sare_ind
and sare_joint
) to boost the training. Our SFRS is trained in generations with self-enhanced soft-label losses to achieve state-of-the-art performance.