Skip to content

[FEATURE REQUEST] Auto-detect llama2/llama3 from tokenizer.model in runner-aoti/runner-et #706

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
metascroy opened this issue May 6, 2024 · 7 comments

Comments

@metascroy
Copy link
Contributor

metascroy commented May 6, 2024

Remove -l 2 and -l 3 flags and auto-detect model architecture from tokenizer class.

Issue warning if user supplied -v does not match the tokenizer.model inferred vocab size.

@mikekgfb mikekgfb changed the title Auto-detect llama2/llama3 from tokenizer.model in runner-aoti/runner-et [FEATURE REQUEST] Auto-detect llama2/llama3 from tokenizer.model in runner-aoti/runner-et May 12, 2024
@Jack-Khuu
Copy link
Contributor

AOTI support added, ET support flagged in #1484

@larryliu0820
Copy link
Contributor

AOTI support added, ET support flagged in #1484

There's some remaining integration work needs to be done so maybe we should keep this open

@larryliu0820
Copy link
Contributor

Basically right now if we do torchchat export with llama3.2 vision model tag and aoti, there's some breakage. Was hoping someone can enable that path

@larryliu0820
Copy link
Contributor

Oh uh, this issue is not talking about aoti I guess?

@Jack-Khuu
Copy link
Contributor

The original task was pretty old, so context has slightly changed as well

I took it to mean removing the need for manually specifying the tokenizer type during aoti_runs for the Text-Only models (Like angela's AOTI PR and the followup issue #1484)

ModelType model_type = get_model_type(std::stoi(aoti_metadata["tokenizer_type"]));

@Jack-Khuu
Copy link
Contributor

Basically right now if we do torchchat export with llama3.2 vision model tag and aoti, there's some breakage. Was hoping someone can enable that path

Got it, mind spinning up an issue for the 3.2 aoti integration?

@larryliu0820
Copy link
Contributor

Basically right now if we do torchchat export with llama3.2 vision model tag and aoti, there's some breakage. Was hoping someone can enable that path

Got it, mind spinning up an issue for the 3.2 aoti integration?

Here you go #1497

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants