diff --git a/.gitignore b/.gitignore new file mode 100644 index 0000000..938fec3 --- /dev/null +++ b/.gitignore @@ -0,0 +1,2 @@ +.DS_Store +*.pdf diff --git a/index.html b/index.html index 385fd1a..7e55686 100644 --- a/index.html +++ b/index.html @@ -52,28 +52,44 @@
- Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin ullamcorper tellus sed ante aliquam tempus. Etiam porttitor urna feugiat nibh elementum, et tempor dolor mattis. Donec accumsan enim augue, a vulputate nisi sodales sit amet. Proin bibendum ex eget mauris cursus euismod nec et nibh. Maecenas ac gravida ante, nec cursus dui. Vivamus purus nibh, placerat ac purus eget, sagittis vestibulum metus. Sed vestibulum bibendum lectus gravida commodo. Pellentesque auctor leo vitae sagittis suscipit. -
+ The Diffusion model, a prevalent framework for image + generation, encounters significant challenges in terms of + broad applicability due to its extended inference times and + substantial memory requirements. Efficient Post-training + Quantization (PTQ) is pivotal for addressing these issues in + traditional models. Different from traditional models, diffusion models heavily depend on the time-step t to achieve + satisfactory multi-round denoising. Usually, t from the finite set {1, . . . , T} is encoded to a temporal feature by a + few modules totally irrespective of the sampling data. However, existing PTQ methods do not optimize these modules + separately. They adopt inappropriate reconstruction targets + and complex calibration methods, resulting in a severe disturbance of the temporal feature and denoising trajectory, + as well as a low compression efficiency. To solve these, + we propose a Temporal Feature Maintenance Quantization + (TFMQ) framework building upon a Temporal Information + Block which is just related to the time-step t and unrelated + to the sampling data. Powered by the pioneering block design, we devise temporal information aware reconstruction + (TIAR) and finite set calibration (FSC) to align the fullprecision temporal features in a limited time. Equipped with + the framework, we can maintain the most temporal information and ensure the end-to-end generation quality. Extensive experiments on various datasets and diffusion models + prove our state-of-the-art results. Remarkably, our quantization approach, for the first time, achieves model performance nearly on par with the full-precision model under + 4-bit weight quantization. Additionally, our method incurs + almost no extra computational cost and accelerates quantization time by 2.0Ă— on LSUN-Bedrooms 256 Ă— 256 compared to previous works. +
+
+
+ + Random samples from w4a8 quantized and full-precision LDM-4 on CelebA-HQ 256 Ă— 256. The resolution of each sample is + 256 Ă— 256. ++ |
+
+
+
+ + Random samples from w4a8 quantized and full-precision LDM-4 on FFHQ 256 Ă— 256. The resolution of each sample is + 256 Ă— 256. ++ |
+
+
+ + Random samples from w4a8 quantized and full-precision LDM-8 on LSUN-Churches 256Ă—256. The resolution of each sample + is 256 Ă— 256. ++ |
+
+
+ + Random samples from w4a8 quantized and full-precision LDM-4 on LSUN-Bedrooms 256Ă—256. The resolution of each sample + is 256 Ă— 256. ++ |
+
+
+ + Random samples from w4a8 quantized and full-precision DDIM on CIFAR-10 32 Ă— 32. The resolution of each sample is + 32 Ă— 32. ++ |
+
+
+
+ + Random samples from w4a8 quantized and full-precision DDIM on CIFAR-10 32 Ă— 32. The resolution of each sample is + 32 Ă— 32. ++
+
+
+ + Random samples from w4a8 quantized and full-precision Stable Diffusion. (Left) prompt: "A digital illustration of the Babel + tower, detailed, trending in artstation, fantasy vivid colors." (Right) prompt: "A beautiful castle beside a waterfall in the woods." The + resolution of each sample is 512 Ă— 512. ++ |
+
BibTex Code Here
+ @inproceedings{huang2024tfmqdm,
+ title={TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion Models},
+ author={Yushi Huang and Ruihao Gong and Jing Liu and Tianlong Chen and Xianglong Liu},
+ booktitle={The IEEE/CVF Conference on Computer Vision and Pattern Recognition},
+ year={2024}
+}
+