Skip to content

Conversation

@zkl-ai
Copy link
Contributor

@zkl-ai zkl-ai commented Dec 14, 2025

Summary

  • Switch the quantize_and_pack_int4.ipynb example to the latest compressed-tensors workflow:
    • Use ModelCompressor.compress_model(model) for in-memory compression
    • Remove the outdated compress_quantized_weights usage

Related Issue

…te save flow and docs; Fixes #2105

Signed-off-by: zkl-ai <[email protected]>
@zkl-ai zkl-ai force-pushed the feat/update-int4-notebook-compress-model-2105 branch from 03ec34b to 18697a2 Compare December 15, 2025 10:58
Copy link
Collaborator

@kylesayrs kylesayrs left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Really nice, thank you!

@fynnsu fynnsu merged commit 797d301 into vllm-project:main Dec 15, 2025
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants