-
Notifications
You must be signed in to change notification settings - Fork 721
Issues: simplescaling/s1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Issue Reproducing s1.1-32B Training Loss (Observed vs. WandB)
#108
opened Mar 31, 2025 by
dzh19990407
Can I specify the thinking context of the query when s1 inferencing with vllm?
#107
opened Mar 24, 2025 by
Siki-cloud
qwen2.5-32B can not fine tuning on 80GB A100 in max_seq_length=32768
#106
opened Mar 21, 2025 by
Gresham429
The minimum GPU resources needed to fine-tune the 32B model?
#103
opened Mar 19, 2025 by
JaydencoolCC
Exploring LoRA Adapters & Low Precision Training in S1 for Enhanced Test-Time Scaling
#101
opened Mar 19, 2025 by
goravaa
s1-32b keeps generating same trajectories and final answer regardless of changing of seeds
#97
opened Mar 12, 2025 by
nichenshun
Gemini thinking flash API no longer returns thoughts and response separately
#85
opened Mar 3, 2025 by
SusMaria
Script to generate traces via DeepSeek R1 seems to be missing
#73
opened Feb 24, 2025 by
nileshtrivedi
Inference Token Inclusion and SFT Question Loss Calculation Queries
#72
opened Feb 23, 2025 by
ruio248
Previous Next
ProTip!
Updated in the last three days: updated:>2025-04-09.