Skip to content

Popular repositories Loading

  1. petals petals Public

    Forked from panf2333/petals

    🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

    Python

  2. vllm vllm Public

    Forked from vllm-project/vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

    Python

  3. nunchaku nunchaku Public

    Forked from mit-han-lab/nunchaku

    [ICLR2025] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models

    Cuda

Repositories

Showing 3 of 3 repositories
  • vllm Public Forked from vllm-project/vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

    yottalabsai/vllm’s past year of commit activity
    Python 0 Apache-2.0 5,902 0 1 Updated Feb 18, 2025
  • nunchaku Public Forked from mit-han-lab/nunchaku

    [ICLR2025] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models

    yottalabsai/nunchaku’s past year of commit activity
    Cuda 0 Apache-2.0 44 0 0 Updated Feb 12, 2025
  • petals Public Forked from panf2333/petals

    🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

    yottalabsai/petals’s past year of commit activity
    Python 0 MIT 544 0 0 Updated Oct 15, 2024

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…