Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is there any plan for updating vllm to 0.7.x? #43

Open
RyanLiu112 opened this issue Feb 24, 2025 · 1 comment
Open

Is there any plan for updating vllm to 0.7.x? #43

RyanLiu112 opened this issue Feb 24, 2025 · 1 comment

Comments

@RyanLiu112
Copy link

Hi,

Thanks for the great work! I just wanted to share that verl has added support for vLLM 0.7.x (link), and the performance boost is pretty impressive:

For a typical job like examples/ppo_trainer/run_qwen2-7b_seq_balance.sh, the rollout generation time is 115 seconds with vLLM0.6.3, while it is 85 seconds with vLLM0.7.0. By enabling the cudagraph, the generation duration is further reduced to 62 seconds.

Are you planning to add support for this in deepscaler? It would be awesome to see these speed improvements in deepscaler too.

Thanks for considering!

Best regards,
Runze

@lambda7xx
Copy link

we are doing that. In the future release, the vllm will be 0.7.x

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants