Skip to content

Supports the vLLM inference framework #18740

@TracyBin

Description

@TracyBin

Does Alluxio support caching and accelerating model weight files based on NFS to achieve faster vLLM model inference startup?

Metadata

Metadata

Assignees

No one assigned

    Labels

    type-featureThis issue is a feature request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions