-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
got bad results: quant with 1024 image using sdxl-tubo, using 1.3.1 Normal infer, follow your steps #9
Comments
i am really curious about that whether your result can be wholly reproced. using your quanted config can get good result. but when i quant sdxl-turbo from "scratch", i cannot get expected result. only vague results |
Apologies for the delayed response. There may be an underlying issue in your quantization process. Could you please provide more detailed information about the experimental settings (how does "from scratch" means), so that we can help you with the problem? |
here is my process: after above process, i got result like this: |
then, i tried The Mixed Precision Search Process. |
This process conducts uniform bit-width W8A8 quantization without mixed precision, which would produce unsatisfying result. you could try adding the |
i have tried --act_protect here |
This phenonmenon is probabily due to the sub-optimal mixed precision configuration. does the mixed precision searching process raise any errors? |
I see. The all W8A8 quantization will generate images with visual degradation. Therefore, we identify the top 1% sensitive layers, and preserve them as FP for performance preservation. Therefore, to generate images with good quality, "act_protect" should be specified. We design the mixed precision W8A8, which uses all 8 as W bit-width (weight_8.00.yaml), and average of 7.77 A bit, to account for the FP16 layers to make the average Activation bit-width as 8 bit. (The command we provide as example)
If you want a full W8A8 model, simply changing all the bit-width in act_7.77.yaml as 8 should work. |
i know these commands. using them can produce good results. but my question is how to obtain these configs?
i tried your steps listed in https://github.com/A-suozhang/MixDQ/blob/master/mixed_precision_scripts/mixed_precision_search.md. |
If you simply want to conduct W8A8 quantization, you could set all bit-width in the WEIGHT_MP_CFG & ACT_MP_CFG as 8-bit with act_protect layers. If you want to search your own mixed precision configuration. After acquiring the layer sensitivity, you may need to run multiple times for integer programming with different seed / target bit-width to generate a few candidate mixed precision configurations, and select the optimal one based on visual quality of the actual generated image. |
For more details of the search process, you may refer to the Appendix of our paper. Sorry for the unclear description in |
Also, hope you can disclose more about the process of acquiring the act protect config
in |
The |
No description provided.
The text was updated successfully, but these errors were encountered: