Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

continue training and longer length #23

Open
ypwang61 opened this issue Feb 13, 2025 · 2 comments
Open

continue training and longer length #23

ypwang61 opened this issue Feb 13, 2025 · 2 comments

Comments

@ypwang61
Copy link

Hi, thanks for your really great work! Have you tried on continue training (like on 24k context length) or using longer context length (like 36k or higher)? I just wonder when will this crazy thing saturates, or actually it's still quite powerful if giving more computation budget

@ypwang61 ypwang61 changed the title continue training continue training and longer length Feb 13, 2025
@michaelzhiluo
Copy link
Contributor

We're unsure about the limit of 1.5B, watch out for our future work! There are some new challenges we've encountered ;)

@ypwang61
Copy link
Author

very cool! looking forward to your future works!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants