From 18697a2eb938719b8433af1e6a9e7da0a7e9cda4 Mon Sep 17 00:00:00 2001 From: zkl-ai <907668776@qq.com> Date: Sun, 14 Dec 2025 20:28:50 +0800 Subject: [PATCH] examples: use compress_model, remove compress_quantized_weights; update save flow and docs; Fixes #2105 Signed-off-by: zkl-ai <907668776@qq.com> --- examples/quantize_and_pack_int4.ipynb | 144 +++++++++----------------- 1 file changed, 51 insertions(+), 93 deletions(-) diff --git a/examples/quantize_and_pack_int4.ipynb b/examples/quantize_and_pack_int4.ipynb index e4d654685..e3889979c 100644 --- a/examples/quantize_and_pack_int4.ipynb +++ b/examples/quantize_and_pack_int4.ipynb @@ -15,7 +15,7 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": 12, "metadata": {}, "outputs": [], "source": [ @@ -25,8 +25,7 @@ "from compressed_tensors.quantization import (\n", " QuantizationConfig,\n", " QuantizationStatus,\n", - " apply_quantization_config,\n", - " compress_quantized_weights\n", + " apply_quantization_config\n", ")\n", "from compressed_tensors.compressors import ModelCompressor\n", "from transformers import AutoModelForCausalLM, AutoTokenizer, DefaultDataCollator\n", @@ -37,51 +36,9 @@ }, { "cell_type": "code", - "execution_count": 3, + "execution_count": 13, "metadata": {}, "outputs": [ - { - "data": { - "application/vnd.jupyter.widget-view+json": { - "model_id": "c883cdc8ecd04866bd01d61796b81c26", - "version_major": 2, - "version_minor": 0 - }, - "text/plain": [ - "config.json: 0%| | 0.00/560 [00:00