Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

questions about LLM2CLIP finetune #26

Open
liumingzhu6060 opened this issue Dec 25, 2024 · 1 comment
Open

questions about LLM2CLIP finetune #26

liumingzhu6060 opened this issue Dec 25, 2024 · 1 comment

Comments

@liumingzhu6060
Copy link

when LLM2CLIP finetue, do you try unfreeze the LLM gradients and ViT model simultaneously,do you think this method may generate better results?

@Yif-Yang
Copy link
Collaborator

We have tried using LoRA to fine-tune the gradients of the LLM, and it does lead to a slight performance improvement, but not a substantial one. Perhaps more data is needed to fully unlock its potential. It’s also possible that the distribution of image captions might disrupt the original capabilities of the LLM. Therefore, I believe this requires some new design efforts. We will disclose more related experiments in our next version of the work. If you have tried similar approaches, we would be happy to exchange ideas and share experiences.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants