Skip to content
Discussion options

You must be logged in to vote

@ebsmothers should work for you in this case. You can make this new checkpoint using the code below and then finetune it with any llama3_2_vision config you want.

from torchtune import models, training

with training.set_default_dtype("bf16"), "cpu":
    model = models.llama3_2_vision.llama3_2_vision_11b(...) # same args as in the config

# Load the 3.2 vision checkpoint (use HF or Meta checkpointer depending on your checkpoint)
model_checkpointer = training.FullModelMetaCheckpointer(...) # same args as in the config
state_dict = model_checkpointer.load_checkpoint()
model.load_state_dict(state_dict)

# Load the 3.1 text checkpoint (use HF or Meta checkpointer depending on your checkpoint)
t…

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@Hyperakan
Comment options

@pbontrager
Comment options

Answer selected by Hyperakan
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants