-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
is MixDQ a PTQ or QAT method? #11
Comments
Thank you for your interest in our work. MixDQ is a PTQ method that does not require tuning, the code in the |
thanks a lot !
At 2024-07-20 13:00:47, "tianchen" ***@***.***> wrote:
Thank you for your interest in our work. MixDQ is a PTQ method that does not require tuning, the code in the base_quantizer.py is simply for compatibility.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
path: "/share/public/diffusion_quant/calib_dataset/bs32_t30_sdxl.pt" HI, where can I download this file?i need all the file download |
You could generate this file, following the instruction of README.md step 1.1 "Generate Calibration Data"
|
thanks a lot. another question, when I reference, is it much slower since I need if else to see which precision to dequantize? |
I'm not quite sure I fully understand your question. But Yes, the code within this repository is the "algorithm-level" quantization simulation code, and runs slower than FP16. For actual speedup, customized CUDA kernel that utilizes the INT computation is needed (our huggingface demo code, https://huggingface.co/nics-efc/MixDQ ). |
in the base_quantizier.py, there are these words:PyTorch Function that can be used for asymmetric quantization (also called uniform affine
quantization). Quantizes its argument in the forward pass, passes the gradient 'straight
through' on the backward pass, ignoring the quantization that occurred.
Based on https://arxiv.org/abs/1806.08342.,
So is MixDQ a PTQ or QAT method?need backward pass when quantizationing?
The text was updated successfully, but these errors were encountered: