Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue: Update VLLM to Version .5.0++, and a few suggestions #83

Open
nerdylive123 opened this issue Jul 3, 2024 · 13 comments
Open

Issue: Update VLLM to Version .5.0++, and a few suggestions #83

nerdylive123 opened this issue Jul 3, 2024 · 13 comments

Comments

@nerdylive123
Copy link

Description

  1. 🌟 Upgrade VLLM: We need to rocket VLLM to version 0.5.0++ or beyond! πŸš€
  2. πŸ€– Tensorize Awesomeness: The tensorize feature is like giving VLLM a turbo boost. 🏎️ Check out the Tensorize VLLM example for a sneak peek.
    • πŸš€ It lets us load the model during download (but remember, the model needs a little conversion magic).
  3. πŸ“¦ Pip It Up: Why build VLLM from scratch when we can summon it with a pip package? Efficiency, my friend! πŸ§™β€β™‚οΈ

Kudos to the stellar maintainer! πŸŒŸπŸ™Œ

@nerdylive123 nerdylive123 changed the title Issue: Update VLLM to Version .5.0 or Later, n a few suggestions Issue: Update VLLM to Version .5.0++, and a few suggestions Jul 3, 2024
@FrederikHandberg
Copy link

FrederikHandberg commented Jul 4, 2024

+1! I really would like to run Phi3VForCausalLM


@Sapessii
Copy link

Sapessii commented Jul 5, 2024

+1!

@shivanker
Copy link

+1, Gemma 2 support has been recently rolled out in vLLM!

@avacaondata
Copy link

+1, it would make much more sense to pip install vllm so that when a new model is released and implemented in vLLM it is automatically integrated in this worker @alpayariyak

@d4rk6un
Copy link

d4rk6un commented Jul 22, 2024

Are there any plans to upgrade the VLLM version and if so, can you provide a date?

@PhoenixSmaug
Copy link

+1, then we could finally run DeepSeek-Coder v2

@harshal-pr
Copy link

+1

Llama 3.1 needs 0.5.3 https://github.com/vllm-project/vllm/releases/tag/v0.5.3

Can we upgrade this worker to support this out of box in runpod serverless vllm ?

@Lhemamou
Copy link

waiting also for the update :) let me know if i can help !

@alpayariyak
Copy link
Contributor

alpayariyak commented Jul 26, 2024

Hi all, thank you so much for the suggestions! I've joined a different company, so @pandyamarut will be taking over. It's been a great pleasure serving you all!

@Lhemamou
Copy link

I wish you an amazing next work experience ;) welcome aboard @pandyamarut !

@pandyamarut
Copy link
Collaborator

pandyamarut commented Jul 26, 2024

Working on it ,Sorry for the delay. Thanks for maintaining the repo @alpayariyak

@TheAlexPG
Copy link

Guys, do we know anything about the approximate time frame for the update? So that you can somehow plan the update of the models in the roadmap. Thanks

@nerdylive123
Copy link
Author

image
Pls support new quantization fp8, refer to this docs:
vllm docs

I've got a whole new menu with a bunch of new options i guess its all of the arguments thats very great thank you for the update staffs and maintainers! just the options value needs to be updated :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests