-
Notifications
You must be signed in to change notification settings - Fork 249
[FEATURE REQUEST] Auto-detect llama2/llama3 from tokenizer.model in runner-aoti/runner-et #706
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
AOTI support added, ET support flagged in #1484 |
There's some remaining integration work needs to be done so maybe we should keep this open |
Basically right now if we do torchchat export with llama3.2 vision model tag and aoti, there's some breakage. Was hoping someone can enable that path |
Oh uh, this issue is not talking about aoti I guess? |
The original task was pretty old, so context has slightly changed as well I took it to mean removing the need for manually specifying the tokenizer type during aoti_runs for the Text-Only models (Like angela's AOTI PR and the followup issue #1484) Line 856 in 062dd87
|
Got it, mind spinning up an issue for the 3.2 aoti integration? |
Here you go #1497 |
Remove
-l 2
and-l 3
flags and auto-detect model architecture from tokenizer class.Issue warning if user supplied -v does not match the tokenizer.model inferred vocab size.
The text was updated successfully, but these errors were encountered: