diff --git "a/2025/Text Embedding Knows How to Quantize Text-Guided Diffusion Models/layout.json" "b/2025/Text Embedding Knows How to Quantize Text-Guided Diffusion Models/layout.json" new file mode 100644--- /dev/null +++ "b/2025/Text Embedding Knows How to Quantize Text-Guided Diffusion Models/layout.json" @@ -0,0 +1,9678 @@ +{ + "pdf_info": [ + { + "para_blocks": [ + { + "bbox": [ + 83, + 103, + 529, + 121 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 83, + 103, + 529, + 121 + ], + "spans": [ + { + "bbox": [ + 83, + 103, + 529, + 121 + ], + "type": "text", + "content": "Text Embedding Knows How to Quantize Text-Guided Diffusion Models" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 130, + 143, + 483, + 172 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 130, + 143, + 483, + 172 + ], + "spans": [ + { + "bbox": [ + 130, + 143, + 483, + 172 + ], + "type": "text", + "content": "Hongjae Lee Myungjun Son Dongjea Kang Seung-Won Jung\\* Korea University" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 167, + 173, + 443, + 186 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 167, + 173, + 443, + 186 + ], + "spans": [ + { + "bbox": [ + 167, + 173, + 443, + 186 + ], + "type": "text", + "content": "{jimmy9704,sonbill,kedjk11,swjung83}@korea.ac.kr" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 151, + 213, + 201, + 225 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 151, + 213, + 201, + 225 + ], + "spans": [ + { + "bbox": [ + 151, + 213, + 201, + 225 + ], + "type": "text", + "content": "Abstract" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 54, + 239, + 297, + 479 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 239, + 297, + 479 + ], + "spans": [ + { + "bbox": [ + 54, + 239, + 297, + 479 + ], + "type": "text", + "content": "Despite the success of diffusion models in image generation tasks such as text-to-image, the enormous computational complexity of diffusion models limits their use in resource-constrained environments. To address this, network quantization has emerged as a promising solution for designing efficient diffusion models. However, existing diffusion model quantization methods do not consider input conditions, such as text prompts, as an essential source of information for quantization. In this paper, we propose a novel quantization method dubbed Quantization of Language-to-Image diffusion models using text Prompts (QLIP). QLIP leverages text prompts to guide the selection of bit precision for every layer at each time step. In addition, QLIP can be seamlessly integrated into existing quantization methods to enhance quantization efficiency. Our extensive experiments demonstrate the effectiveness of QLIP in reducing computational complexity and improving the quality of the generated images across various datasets. The source code is available at our project page https://github.com/jimmy9704/QLIP." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 56, + 504, + 137, + 517 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 504, + 137, + 517 + ], + "spans": [ + { + "bbox": [ + 56, + 504, + 137, + 517 + ], + "type": "text", + "content": "1. Introduction" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 55, + 525, + 296, + 693 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 525, + 296, + 693 + ], + "spans": [ + { + "bbox": [ + 55, + 525, + 296, + 693 + ], + "type": "text", + "content": "Diffusion models have achieved remarkable success in various generative tasks, particularly in text-to-image generation [39, 40, 57]. However, the huge computational complexity of diffusion models restricts their practical applicability, especially in resource-constrained environments. Specifically, diffusion models typically require hundreds of denoising steps to produce high-quality samples, making them considerably slower than generative adversarial networks (GANs) [10]. To this end, many studies introduced advanced training-free samplers to reduce the number of denoising iterations [2, 29, 31, 48]. However, the decrease in iterations alone is insufficient since the denoising models used in each iteration possess billions of parameters [4, 37, 42], leading to increased computational com" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 313, + 214, + 474, + 226 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 214, + 474, + 226 + ], + "spans": [ + { + "bbox": [ + 313, + 214, + 474, + 226 + ], + "type": "text", + "content": "plexity and high memory consumption." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 313, + 228, + 555, + 503 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 228, + 555, + 503 + ], + "spans": [ + { + "bbox": [ + 313, + 228, + 555, + 503 + ], + "type": "text", + "content": "Besides other approaches toward light-weight network designs, such as network architecture search [8, 28], network pruning [12, 50], and knowledge distillation [23, 56], network quantization has received particular attention due to its simplicity, the increasing demand for low-precision operations, and related hardware support [9, 55, 58]. Although various network quantization methods have been proposed for convolutional neural networks (CNNs), the direct use of such methods for diffusion models has shown significant performance degradation [45]. As tailored solutions for diffusion models, several methods have been proposed considering the time step of diffusion models for quantization. For example, PTQ4DM [45] performed 8-bit quantization by constructing a calibration dataset based on the time step. Q-Diffusion [24] introduced an advanced method based on BRECQ [25] and performed the performance analysis of quantization methods for text-guided diffusion models. In addition, TDQ [47] performed quantization with varying activation scaling based on the diffusion time step. However, these methods do not use the text prompt provided to the diffusion model as valuable information for quantization, leaving room for further improvement." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 506, + 556, + 651 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 506, + 556, + 651 + ], + "spans": [ + { + "bbox": [ + 313, + 506, + 556, + 651 + ], + "type": "text", + "content": "Meanwhile, recent studies highlight the advantages of dynamic bit precision quantization considering input conditions [30, 52]. For example, in the case of image superresolution, CADyQ [18] and AdaBM [17] performed dynamic quantization according to the complexity of the images and the importance of layers. RefQSR [22] clustered the image patches and assigned high-bit and low-bit precisions to the reference and query patches, respectively. These dynamic quantization methods allow for the cost-effective assignment of bit precision, enhancing quantization efficiency. However, such input-adaptive quantization methods have not yet been explored in diffusion models." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 654, + 556, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 654, + 556, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 654, + 556, + 715 + ], + "type": "text", + "content": "In this paper, we propose a novel quantization method, dubbed Quantization of Language-to-Image diffusion models using text Prompts (QLIP), which determines the bit precision of diffusion models according to the text prompt. Figure 1 shows our observation: The quality of the gener" + } + ] + } + ], + "index": 12 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "type": "header", + "angle": 0, + "lines": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "spans": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "type": "text", + "content": "CVF" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "type": "header", + "angle": 0, + "lines": [ + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "spans": [ + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "type": "text", + "content": "This ICCV paper is the Open Access version, provided by the Computer Vision Foundation. Except for this watermark, it is identical to the accepted version; the final published version of the proceedings is available on IEEE Xplore." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 66, + 703, + 142, + 713 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 66, + 703, + 142, + 713 + ], + "spans": [ + { + "bbox": [ + 66, + 703, + 142, + 713 + ], + "type": "text", + "content": "* corresponding author" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "type": "text", + "content": "15426" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 0 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 73, + 71, + 144, + 142 + ], + "blocks": [ + { + "bbox": [ + 73, + 71, + 144, + 142 + ], + "lines": [ + { + "bbox": [ + 73, + 71, + 144, + 142 + ], + "spans": [ + { + "bbox": [ + 73, + 71, + 144, + 142 + ], + "type": "image", + "image_path": "4eb91b1bf14e292483921a31f8f9a8f863e441e39dd39c885c7b363aaf81194f.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 66, + 147, + 228, + 191 + ], + "lines": [ + { + "bbox": [ + 66, + 147, + 228, + 191 + ], + "spans": [ + { + "bbox": [ + 66, + 147, + 228, + 191 + ], + "type": "text", + "content": "Pikachu, the symbol and eternal mascot of Pokémon, is the most famous of over 1,000 Pokémon and is a Pokémon with an electric attribute. It appeared in various video games, TV shows, movies, and merchandise since its introduction in the late 1990s." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_footnote" + }, + { + "bbox": [ + 54, + 200, + 555, + 243 + ], + "lines": [ + { + "bbox": [ + 54, + 200, + 555, + 243 + ], + "spans": [ + { + "bbox": [ + 54, + 200, + 555, + 243 + ], + "type": "text", + "content": "Figure 1. Examples of the images generated by Stable Diffusion v1.4 using the text prompts given below images. When the text prompt details the shape, texture, and properties for the target object, the generated images have rich details and textures, as shown on the right; whereas the generated images are not very different from the images generated using only the keyword when the text prompt is less specific to the target object, as shown on the left." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 149, + 71, + 220, + 142 + ], + "blocks": [ + { + "bbox": [ + 149, + 71, + 220, + 142 + ], + "lines": [ + { + "bbox": [ + 149, + 71, + 220, + 142 + ], + "spans": [ + { + "bbox": [ + 149, + 71, + 220, + 142 + ], + "type": "image", + "image_path": "ec0073a3ac985fc020281a1356d7493a41b8f94db770caf202fbbe6f261ae37a.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + } + ], + "index": 2 + }, + { + "type": "image", + "bbox": [ + 232, + 71, + 304, + 142 + ], + "blocks": [ + { + "bbox": [ + 232, + 71, + 304, + 142 + ], + "lines": [ + { + "bbox": [ + 232, + 71, + 304, + 142 + ], + "spans": [ + { + "bbox": [ + 232, + 71, + 304, + 142 + ], + "type": "image", + "image_path": "bb842b45fe9d55c3ed0f3fbb581de141ca28649f5f0ab8b643855dd452716366.jpg" + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 294, + 164, + 318, + 173 + ], + "lines": [ + { + "bbox": [ + 294, + 164, + 318, + 173 + ], + "spans": [ + { + "bbox": [ + 294, + 164, + 318, + 173 + ], + "type": "text", + "content": "Pikachu" + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_caption" + } + ], + "index": 3 + }, + { + "type": "image", + "bbox": [ + 309, + 72, + 380, + 141 + ], + "blocks": [ + { + "bbox": [ + 309, + 72, + 380, + 141 + ], + "lines": [ + { + "bbox": [ + 309, + 72, + 380, + 141 + ], + "spans": [ + { + "bbox": [ + 309, + 72, + 380, + 141 + ], + "type": "image", + "image_path": "13febf055745aa57c7ff2933467dfa31a53dd9f95741acbbdc98aed283bc802d.jpg" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 384, + 143, + 545, + 186 + ], + "lines": [ + { + "bbox": [ + 384, + 143, + 545, + 186 + ], + "spans": [ + { + "bbox": [ + 384, + 143, + 545, + 186 + ], + "type": "text", + "content": "Pikachu is a small, yellow, mouse-like creature with large, black-tipped ears that resemble lightning bolts. It has rosy cheeks, a small mouth, and large, expressive eyes. Pikachu's body is covered in soft, yellow fur, and it has a long, lightning bolt-shaped tail." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_footnote" + } + ], + "index": 5 + }, + { + "type": "image", + "bbox": [ + 392, + 72, + 462, + 141 + ], + "blocks": [ + { + "bbox": [ + 392, + 72, + 462, + 141 + ], + "lines": [ + { + "bbox": [ + 392, + 72, + 462, + 141 + ], + "spans": [ + { + "bbox": [ + 392, + 72, + 462, + 141 + ], + "type": "image", + "image_path": "84c03b2ff81d7d3d6496d316cfdd9546f634ee4c1abc10364779983f9d7bca2f.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "image_body" + } + ], + "index": 6 + }, + { + "type": "image", + "bbox": [ + 468, + 72, + 539, + 141 + ], + "blocks": [ + { + "bbox": [ + 468, + 72, + 539, + 141 + ], + "lines": [ + { + "bbox": [ + 468, + 72, + 539, + 141 + ], + "spans": [ + { + "bbox": [ + 468, + 72, + 539, + 141 + ], + "type": "image", + "image_path": "8e769f7a176ca997ed6c3bcd792144e82df40fa92ec0790e889b314586e7ef37.jpg" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "image_body" + } + ], + "index": 7 + }, + { + "bbox": [ + 54, + 265, + 294, + 348 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 265, + 294, + 348 + ], + "spans": [ + { + "bbox": [ + 54, + 265, + 294, + 348 + ], + "type": "text", + "content": "ated images increases as the text prompt has more specific information about the image to be generated. This suggests that high-bit precision operations are necessary to generate images with vivid details and textures corresponding to detailed text prompts; whereas, low-bit precision operations can be sufficient to generate images corresponding to text prompts that lack specific guidance on details and textures." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 54, + 350, + 295, + 529 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 350, + 295, + 529 + ], + "spans": [ + { + "bbox": [ + 54, + 350, + 295, + 529 + ], + "type": "text", + "content": "Motivated by this observation, we quantized the diffusion model [39] using Q-diffusion [24] with high-bit and low-bit settings and compared the quality of the generated images with the full-precision model. As shown in Figure 2, the performance drop of the low-bit model is noticeable when the quality of the generated images is high for the full-precision model. In contrast, when the quality is low for the full-precision model, the high-bit and low-bit models perform similarly to the full-precision model. These results suggest that diffusion models can be efficiently quantized without sacrificing performance if the quality of the generated images can be first predicted and then used to determine bit precision. Accordingly, our proposed QLIP predicts the image quality from the text prompt and determines the quantization bit precision for diffusion models." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 55, + 530, + 295, + 601 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 530, + 295, + 601 + ], + "spans": [ + { + "bbox": [ + 55, + 530, + 295, + 601 + ], + "type": "text", + "content": "Furthermore, previous studies [13, 24, 45, 47] revealed that the activation distributions of diffusion models vary across layers over time steps. Inspired by these studies, QLIP uses learnable parameters to determine the bit precision for each layer and for each timestep. Our contributions can be summarized as follows:" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 55, + 603, + 295, + 711 + ], + "type": "list", + "angle": 0, + "index": 16, + "blocks": [ + { + "bbox": [ + 55, + 603, + 295, + 638 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 603, + 295, + 638 + ], + "spans": [ + { + "bbox": [ + 55, + 603, + 295, + 638 + ], + "type": "text", + "content": "- To the best of our knowledge, QLIP is the first approach that explicitly utilizes text prompts for dynamic quantization in diffusion models." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 55, + 639, + 294, + 674 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 639, + 294, + 674 + ], + "spans": [ + { + "bbox": [ + 55, + 639, + 294, + 674 + ], + "type": "text", + "content": "- We propose leveraging text prompts to predict the image quality, which is used to guide the selection of bit precision for every layer in different denoising time steps." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 55, + 675, + 294, + 711 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 675, + 294, + 711 + ], + "spans": [ + { + "bbox": [ + 55, + 675, + 294, + 711 + ], + "type": "text", + "content": "- Our proposed QLIP can be integrated into existing diffusion quantization methods to enhance quantization efficiency." + } + ] + } + ], + "index": 15 + } + ], + "sub_type": "text" + }, + { + "bbox": [ + 314, + 263, + 398, + 276 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 263, + 398, + 276 + ], + "spans": [ + { + "bbox": [ + 314, + 263, + 398, + 276 + ], + "type": "text", + "content": "2. Related work" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 313, + 284, + 536, + 297 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 284, + 536, + 297 + ], + "spans": [ + { + "bbox": [ + 313, + 284, + 536, + 297 + ], + "type": "text", + "content": "2.1. Computationally efficient diffusion models" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 312, + 302, + 553, + 422 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 312, + 302, + 553, + 422 + ], + "spans": [ + { + "bbox": [ + 312, + 302, + 553, + 422 + ], + "type": "text", + "content": "Diffusion models generate high-quality images by iteratively removing noise from random noise. Despite their capability to generate high-quality images, diffusion models encounter challenges in real-world applications due to the time-consuming iterative denoising process. Thus, many efforts have been made to make the diffusion process computationally efficient, including reducing the number of iterative diffusion processes via a non-Markovian process [48], adjusting the variance scheduling [33], and employing knowledge distillation [32, 43]." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 313, + 422, + 553, + 518 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 422, + 553, + 518 + ], + "spans": [ + { + "bbox": [ + 313, + 422, + 553, + 518 + ], + "type": "text", + "content": "Meanwhile, since the computational complexity of diffusion model architectures also limits their practical applicability, especially in resource-constrained environments, research efforts have been made to lighten denoising models. For example, SnapFusion [26] introduced an efficient U-Net architecture based on Stable Diffusion [39], while BKSDM [19] further lightened the architecture through knowledge distillation." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 313, + 528, + 491, + 540 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 528, + 491, + 540 + ], + "spans": [ + { + "bbox": [ + 313, + 528, + 491, + 540 + ], + "type": "text", + "content": "2.2. Quantization of diffusion models" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 313, + 545, + 555, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 545, + 555, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 545, + 555, + 713 + ], + "type": "text", + "content": "Quantization involves mapping 32-bit floating-point values of weights and activations to lower-bit values. Numerous studies have explored quantization to compress and accelerate neural networks [5, 18], which can be classified into two approaches: quantization-aware training (QAT) and post-training quantization (PTQ). Recent studies [7, 35, 41, 51] on quantizing diffusion models have focused mainly on PTQ due to the inability to use training data for privacy and commercial reasons, as well as the high computational cost of QAT. Specifically, PTQ4DM [45] and Q-diffusion [24] performed quantization of diffusion models by generating a time step-aware calibration set that consists of data sampled across various time steps using the full-precision denoising model. PTQD [13] separated the" + } + ] + } + ], + "index": 22 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15427" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 1 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 65, + 71, + 330, + 220 + ], + "blocks": [ + { + "bbox": [ + 65, + 71, + 330, + 220 + ], + "lines": [ + { + "bbox": [ + 65, + 71, + 330, + 220 + ], + "spans": [ + { + "bbox": [ + 65, + 71, + 330, + 220 + ], + "type": "image", + "image_path": "6ff6701630ba33acf936ede1f816cfe9163784d4bf9677452a8cbfa01de6b369.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 54, + 228, + 555, + 296 + ], + "lines": [ + { + "bbox": [ + 54, + 228, + 555, + 296 + ], + "spans": [ + { + "bbox": [ + 54, + 228, + 555, + 296 + ], + "type": "text", + "content": "Figure 2. Analysis of the effect of quantization and the quality of generated images, where the image quality is measured using the GIQA score [11]. On the left side, the x-axis represents the quality of images generated by the full-precision model, and the y-axis represents the quality difference between the generated images using the full-precision and quantized models. The images are generated by Stable Diffusion v1.4 [39] using 5000 captions from the COCO 2017 dataset [27] with a size of " + }, + { + "bbox": [ + 54, + 228, + 555, + 296 + ], + "type": "inline_equation", + "content": "512 \\times 512" + }, + { + "bbox": [ + 54, + 228, + 555, + 296 + ], + "type": "text", + "content": ". The high-bit (W4A10) and low-bit (W4A6) quantized models are obtained by applying Q-diffusion [24]. On the right side, examples of images generated using models that are unquantized or quantized w/wo QLIP are shown for two different image quality levels." + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 342, + 72, + 419, + 135 + ], + "blocks": [ + { + "bbox": [ + 342, + 72, + 419, + 135 + ], + "lines": [ + { + "bbox": [ + 342, + 72, + 419, + 135 + ], + "spans": [ + { + "bbox": [ + 342, + 72, + 419, + 135 + ], + "type": "image", + "image_path": "80c76b4ebb90825bf8842e67d5e322bd47ed58ffab25c4c264224097e2bc7cd7.jpg" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_body" + } + ], + "index": 1 + }, + { + "type": "image", + "bbox": [ + 419, + 72, + 481, + 134 + ], + "blocks": [ + { + "bbox": [ + 419, + 72, + 481, + 134 + ], + "lines": [ + { + "bbox": [ + 419, + 72, + 481, + 134 + ], + "spans": [ + { + "bbox": [ + 419, + 72, + 481, + 134 + ], + "type": "image", + "image_path": "2f3acb12f551e68b939b3cb4065830534c18ed65279ab62d59565e8ad6c8b99b.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + } + ], + "index": 2 + }, + { + "type": "image", + "bbox": [ + 482, + 72, + 544, + 134 + ], + "blocks": [ + { + "bbox": [ + 482, + 72, + 544, + 134 + ], + "lines": [ + { + "bbox": [ + 482, + 72, + 544, + 134 + ], + "spans": [ + { + "bbox": [ + 482, + 72, + 544, + 134 + ], + "type": "image", + "image_path": "9fe47efa25c24fb108e881a9e4640205b270f3ed8d3eaad9928b690b7a828026.jpg" + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_body" + } + ], + "index": 3 + }, + { + "type": "image", + "bbox": [ + 345, + 137, + 419, + 212 + ], + "blocks": [ + { + "bbox": [ + 345, + 137, + 419, + 212 + ], + "lines": [ + { + "bbox": [ + 345, + 137, + 419, + 212 + ], + "spans": [ + { + "bbox": [ + 345, + 137, + 419, + 212 + ], + "type": "image", + "image_path": "c1d25611b96184d123cbae5f1e8557c49dd2aa2096161f03caad283c5414bc72.jpg" + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_body" + } + ], + "index": 4 + }, + { + "type": "image", + "bbox": [ + 419, + 137, + 481, + 211 + ], + "blocks": [ + { + "bbox": [ + 419, + 137, + 481, + 211 + ], + "lines": [ + { + "bbox": [ + 419, + 137, + 481, + 211 + ], + "spans": [ + { + "bbox": [ + 419, + 137, + 481, + 211 + ], + "type": "image", + "image_path": "a36df503c5555d2cebe04756dd2aec3253cc66cec91e1bb83b8c9178f52becc7.jpg" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "image_body" + } + ], + "index": 5 + }, + { + "type": "image", + "bbox": [ + 482, + 137, + 544, + 211 + ], + "blocks": [ + { + "bbox": [ + 482, + 137, + 544, + 211 + ], + "lines": [ + { + "bbox": [ + 482, + 137, + 544, + 211 + ], + "spans": [ + { + "bbox": [ + 482, + 137, + 544, + 211 + ], + "type": "image", + "image_path": "4cb782038d7fdd814494d66a7f43d49c5a862caf53c38483fc0545de08a3bc84.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "image_body" + } + ], + "index": 6 + }, + { + "bbox": [ + 54, + 316, + 296, + 448 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 316, + 296, + 448 + ], + "spans": [ + { + "bbox": [ + 54, + 316, + 296, + 448 + ], + "type": "text", + "content": "quantization noise from the diffusion noise to reduce quantization errors. Additionally, PTQD introduced a time step-based mixed-precision method but did not employ varying bit precision for each layer. Furthermore, TDQ [47] performed quantization by adaptively scaling the activations for different diffusion time steps and layers, but assigned the same bit precision for all time steps. Although these diffusion model quantization methods have shown progress in reducing the complexity of noise estimation models, they do not adjust quantization parameters based on input conditions such as text prompts." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 55, + 456, + 253, + 468 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 456, + 253, + 468 + ], + "spans": [ + { + "bbox": [ + 55, + 456, + 253, + 468 + ], + "type": "text", + "content": "2.3. Input-adaptive dynamic quantization" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 55, + 473, + 296, + 641 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 473, + 296, + 641 + ], + "spans": [ + { + "bbox": [ + 55, + 473, + 296, + 641 + ], + "type": "text", + "content": "Recent advances in hardware have enabled mixed-precision operations, allowing various layers to use different bit precisions to improve performance and computational efficiency [9, 55]. Many methods have been proposed to determine the bit precision of each layer of CNNs [9, 30, 55]. For example, several input-adaptive dynamic quantization methods [17, 18, 22, 52] have been introduced for super-resolution, which adjust bit precisions based on the complexity of image patches and the sensitivity to quantization. However, diffusion models have not yet explored such input-adaptive quantization methods. Therefore, this paper proposes a novel quantization method for text-guided diffusion models that applies adaptive quantization using input conditions." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 55, + 651, + 111, + 664 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 651, + 111, + 664 + ], + "spans": [ + { + "bbox": [ + 55, + 651, + 111, + 664 + ], + "type": "text", + "content": "3. Method" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 55, + 671, + 141, + 683 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 671, + 141, + 683 + ], + "spans": [ + { + "bbox": [ + 55, + 671, + 141, + 683 + ], + "type": "text", + "content": "3.1. Preliminaries" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 55, + 689, + 296, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 689, + 296, + 713 + ], + "spans": [ + { + "bbox": [ + 55, + 689, + 296, + 713 + ], + "type": "text", + "content": "Diffusion model. The reverse process of a diffusion model gradually removes noise from random data " + }, + { + "bbox": [ + 55, + 689, + 296, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{x}_t" + }, + { + "bbox": [ + 55, + 689, + 296, + 713 + ], + "type": "text", + "content": " to" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "spans": [ + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "content": "produce a high-quality image " + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "inline_equation", + "content": "\\mathbf{x}_0" + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "content": ", which can be expressed as " + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "inline_equation", + "content": "p_{\\theta}(\\mathbf{x}_{t - 1}|\\mathbf{x}_t) = \\mathcal{N}(\\mathbf{x}_{t - 1};\\mu_\\theta (\\mathbf{x}_t,t),\\sigma_t^2\\mathbf{I})" + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "content": ". The denoising process at time step " + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "content": " using the noise estimation model " + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "inline_equation", + "content": "\\epsilon_{\\theta}" + }, + { + "bbox": [ + 313, + 316, + 555, + 365 + ], + "type": "text", + "content": " can then be represented as follows:" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 340, + 369, + 555, + 396 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 340, + 369, + 555, + 396 + ], + "spans": [ + { + "bbox": [ + 340, + 369, + 555, + 396 + ], + "type": "interline_equation", + "content": "\\mu_ {\\theta} (\\mathbf {x} _ {t}, t) = \\frac {1}{\\sqrt {\\alpha_ {t}}} \\left(\\mathbf {x} _ {t} - \\frac {\\beta_ {t}}{\\sqrt {1 - \\bar {\\alpha} _ {t}}} \\epsilon_ {\\theta} (\\mathbf {x} _ {t}, t)\\right), \\quad (1)", + "image_path": "3c90fb4c893235de63b22214178057f721b1e703357a2173303bc4567501f8e0.jpg" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "spans": [ + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "inline_equation", + "content": "\\alpha_{t},\\bar{\\alpha}_{t}" + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "inline_equation", + "content": "\\beta_{t}" + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "text", + "content": " are the parameters that control the strength of the Gaussian noise in each step, and " + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "inline_equation", + "content": "\\sigma_{t}" + }, + { + "bbox": [ + 313, + 399, + 554, + 435 + ], + "type": "text", + "content": " is a variance schedule [16]." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "spans": [ + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "text", + "content": "Diffusion model quantization. Diffusion model quantization reduces the complexity of a noise estimation model that requires hundreds of iterative sampling processes. The denoising process at time step " + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "text", + "content": " using the quantized noise estimation model " + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "inline_equation", + "content": "\\hat{\\epsilon}_{\\theta}" + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "text", + "content": " and the quantized data " + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "inline_equation", + "content": "\\dot{\\mathbf{x}}_t" + }, + { + "bbox": [ + 313, + 435, + 554, + 506 + ], + "type": "text", + "content": " can be expressed as follows:" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 340, + 521, + 553, + 563 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 340, + 521, + 553, + 563 + ], + "spans": [ + { + "bbox": [ + 340, + 521, + 553, + 563 + ], + "type": "interline_equation", + "content": "\\begin{array}{l} \\hat {\\mu} _ {\\theta} \\left(\\hat {\\mathbf {x}} _ {t}, t\\right) = \\frac {1}{\\sqrt {\\alpha_ {t}}} \\left(\\hat {\\mathbf {x}} _ {t} - \\frac {\\beta_ {t}}{\\sqrt {1 - \\bar {\\alpha} _ {t}}} \\hat {\\epsilon} _ {\\theta} \\left(\\hat {\\mathbf {x}} _ {t}, t\\right)\\right), \\tag {2} \\\\ w h e r e \\hat {\\epsilon} _ {\\theta} = Q _ {\\mathbf {b} ^ {t}} (\\epsilon_ {\\theta}). \\\\ \\end{array}", + "image_path": "26c6ae294f78db0dae3599374651d385b878e942ef107e8898b85cd71395e0ea.jpg" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": "Here, " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^t\\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " represents a set of quantization bit precisions for " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "K" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " layers at time step " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "Q_{\\mathbf{b}^t}" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " is a function that performs layer-wise and time step-wise quantization using " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^t" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": ". To our knowledge, all previous diffusion quantization methods set the same bit precision to all layers, i.e., " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^t (1) = \\mathbf{b}^t (2) = \\dots = \\mathbf{b}^t (K)" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": ", and only a few methods [13, 51] performed time step-wise dynamic quantization, i.e., " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^{t_i}\\neq \\mathbf{b}^{t_j}" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " for some time steps " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "t_i" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "t_j" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "t_i\\neq t_j" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": ". However, such methods do not adapt the bit precision for each layer. Our proposed QLIP determines " + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^t" + }, + { + "bbox": [ + 313, + 569, + 555, + 713 + ], + "type": "text", + "content": " by training so that the quantization error can be effectively reduced without sacrificing the performance of diffusion models." + } + ] + } + ], + "index": 19 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15428" + } + ] + } + ], + "index": 20 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 2 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 75, + 71, + 536, + 189 + ], + "blocks": [ + { + "bbox": [ + 75, + 71, + 536, + 189 + ], + "lines": [ + { + "bbox": [ + 75, + 71, + 536, + 189 + ], + "spans": [ + { + "bbox": [ + 75, + 71, + 536, + 189 + ], + "type": "image", + "image_path": "b4b5aca18b19d9b948d6eda78d52618574b5054bc506997ebc73f8da3e0518e8.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 54, + 197, + 555, + 232 + ], + "lines": [ + { + "bbox": [ + 54, + 197, + 555, + 232 + ], + "spans": [ + { + "bbox": [ + 54, + 197, + 555, + 232 + ], + "type": "text", + "content": "Figure 3. Overview of QLIP that consists of two key modules: T2Q module that predicts image quality from the input text prompt and Q2B module that determines bit precisions for all denoising layers and time steps. Only the Q2B module is trained during QLIP training. See Sections 3.3 and 3.4 for the details about the pre-training of the T2Q module and diffusion models." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 251, + 178, + 262 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 251, + 178, + 262 + ], + "spans": [ + { + "bbox": [ + 55, + 251, + 178, + 262 + ], + "type": "text", + "content": "3.2. Framework overview" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "spans": [ + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": "Our observations shown in Figure 2 indicate that bit precisions for quantization can be determined cost-effectively given the quality of the generated image. However, this image quality remains unknown during the inference stage, motivating us to predict it using an available input, i.e., text prompts. Specifically, if text prompts have rich and detailed information, the expected quality of the generated image is high, and vice versa. To this end, we introduce a text-to-quality (T2Q) module " + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "inline_equation", + "content": "\\phi" + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": " that predicts image quality " + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "inline_equation", + "content": "q" + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": " from the embedding vector " + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "inline_equation", + "content": "\\mathbf{z}" + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": ", i.e., " + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "inline_equation", + "content": "q = \\phi(\\mathbf{z})" + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": ", where " + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "inline_equation", + "content": "\\mathbf{z}" + }, + { + "bbox": [ + 55, + 268, + 296, + 400 + ], + "type": "text", + "content": " is obtained by CLIP-Text [38] using the input text prompt." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "spans": [ + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "text", + "content": "Next, as we discussed, high-bit quantization is required to generate images with vivid details and textures corresponding to detailed text prompts; while low-bit quantization is sufficient to generate images using less specific text prompts. Therefore, we design a quality-to-bit (Q2B) module " + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "inline_equation", + "content": "\\psi" + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "text", + "content": " that maps image quality " + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "inline_equation", + "content": "q" + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "text", + "content": " to appropriate bit precision, i.e., " + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "inline_equation", + "content": "\\mathbf{b}^t = \\psi(q)" + }, + { + "bbox": [ + 55, + 400, + 296, + 519 + ], + "type": "text", + "content": ". Note that the Q2B module determines bit precisions for each time step for different layers since the activation distributions of diffusion models vary across layers over time steps." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "spans": [ + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "text", + "content": "In summary, the proposed QLIP obtains a quantized diffusion model " + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "inline_equation", + "content": "\\hat{\\epsilon}_{\\theta} = Q_{\\psi (\\phi (\\mathbf{z}))}(\\epsilon_{\\theta})" + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "text", + "content": " by learning the parameters of " + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "inline_equation", + "content": "\\psi" + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "inline_equation", + "content": "\\phi" + }, + { + "bbox": [ + 55, + 520, + 296, + 591 + ], + "type": "text", + "content": ". The overview of the proposed QLIP is shown in Figure 3. The following subsections detail the T2Q and Q2B modules and their training procedure and loss functions." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 599, + 203, + 612 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 599, + 203, + 612 + ], + "spans": [ + { + "bbox": [ + 55, + 599, + 203, + 612 + ], + "type": "text", + "content": "3.3. Text-to-Quality estimation" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "spans": [ + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": "The T2Q module, denoted as " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\phi" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": ", aims to predict the image quality from the input text prompt. " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\phi" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " consists of three linear layers and outputs an image quality score " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "q" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " as " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "q = \\phi(\\mathbf{z})" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": ", where " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathbf{z} \\in \\mathbb{R}^{C_{clip}}" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " represents the CLIP text embedding, and " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "C_{clip}" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " denotes its channel dimension. For the training of the T2Q module, we collect a dataset " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathcal{T}_{T2Q} = \\{\\mathbf{z}^i, \\mathbf{x}_0^i\\}_{i=1}^{N_{T2Q}}" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": ", where " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathbf{z}^i" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathbf{x}_0^i" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": " are the " + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "inline_equation", + "content": "i" + }, + { + "bbox": [ + 55, + 616, + 296, + 715 + ], + "type": "text", + "content": "-th sample of the CLIP text embedding and its corresponding generated" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "spans": [ + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": "image obtained using the full-precision diffusion model, respectively, and " + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "inline_equation", + "content": "N_{T2Q}" + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": " is the number of images in " + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "inline_equation", + "content": "\\mathcal{T}_{T2Q}" + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": ". Let " + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "inline_equation", + "content": "\\bar{q}^i" + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": " represent the (pseudo) ground-truth image quality of " + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "inline_equation", + "content": "\\mathbf{x}_0^i" + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": ", where we used GIQA [11] as the default quality measure, with an additional normalization to make the range of quality values [0, 1]. See Section 4.4 for the performance analysis with different quality measures. The T2Q module is trained using the loss " + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "inline_equation", + "content": "L_{t2q}" + }, + { + "bbox": [ + 313, + 251, + 555, + 348 + ], + "type": "text", + "content": ":" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 359, + 358, + 555, + 393 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 359, + 358, + 555, + 393 + ], + "spans": [ + { + "bbox": [ + 359, + 358, + 555, + 393 + ], + "type": "interline_equation", + "content": "L _ {t 2 q} = \\frac {1}{N _ {T 2 Q}} \\sum_ {i = 1} ^ {N _ {T 2 Q}} \\left(\\bar {q} ^ {i} - \\phi \\left(\\mathbf {z} ^ {i}\\right)\\right) ^ {2}. \\tag {3}", + "image_path": "f95abbaaee16455618e3d6ef1c54d7bba8eaba15dc2dfcd2c0cf57d6360d5d1e.jpg" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 313, + 401, + 554, + 437 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 401, + 554, + 437 + ], + "spans": [ + { + "bbox": [ + 313, + 401, + 554, + 437 + ], + "type": "text", + "content": "Once the T2Q module is trained, it is frozen during the training of the Q2B module and served as a quality predictor to help determine quantization bit precisions." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 446, + 454, + 459 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 446, + 454, + 459 + ], + "spans": [ + { + "bbox": [ + 313, + 446, + 454, + 459 + ], + "type": "text", + "content": "3.4. Quality-to-Bit estimation" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "spans": [ + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "text", + "content": "In text-guided diffusion models, the richer and more detailed the information in the text prompt is, the higher the image quality tends to be, requiring the use of higher bits. Therefore, the image quality " + }, + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "inline_equation", + "content": "q" + }, + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "text", + "content": " predicted by the T2Q module can serve as a direct hint for bit selection. To this end, the proposed Q2B module obtains a probability vector " + }, + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 464, + 554, + 549 + ], + "type": "text", + "content": " for bit selection as follows:" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 381, + 559, + 555, + 574 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 381, + 559, + 555, + 574 + ], + "spans": [ + { + "bbox": [ + 381, + 559, + 555, + 574 + ], + "type": "interline_equation", + "content": "\\mathbf {p} _ {q} = \\sigma \\left((q - 0. 5) \\mathbf {s} + \\mathbf {o}\\right), \\tag {4}", + "image_path": "6b6a22a95b088d6bd4c711aa6ffaedfb56bf9a2c803f660178a511ed7eedcf1a.jpg" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "spans": [ + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{s} \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{o} \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " represent sets of learnable parameters, and " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\sigma" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " denotes the sigmoid function. Furthermore, to adjust the bit precision for each time step, additional learnable parameters " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_m^t \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_h^t \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " are introduced to obtain probability vectors " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_m^t \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_h^t \\in \\mathbb{R}^K" + }, + { + "bbox": [ + 313, + 583, + 554, + 654 + ], + "type": "text", + "content": " as" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 346, + 656, + 553, + 671 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 346, + 656, + 553, + 671 + ], + "spans": [ + { + "bbox": [ + 346, + 656, + 553, + 671 + ], + "type": "interline_equation", + "content": "\\mathbf {p} _ {j} ^ {t} = \\sigma \\left(\\mathbf {u} _ {j} ^ {t}\\right), \\quad j \\in \\{m, h \\}, t = 1, 2, \\dots , T, \\tag {5}", + "image_path": "a031b6e3b186d1fa75059ace2fa608dff8f82f4117f6465a200e772f9f50f060.jpg" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 677, + 525, + 689 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 677, + 525, + 689 + ], + "spans": [ + { + "bbox": [ + 313, + 677, + 525, + 689 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 313, + 677, + 525, + 689 + ], + "type": "inline_equation", + "content": "T" + }, + { + "bbox": [ + 313, + 677, + 525, + 689 + ], + "type": "text", + "content": " is the total number of denoising time steps." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "content": "These probability vectors, " + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q" + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_m^t" + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_h^t" + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "content": ", are used to determine the bit precision for each layer and each time" + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15429" + } + ] + } + ], + "index": 18 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 3 + }, + { + "para_blocks": [ + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "spans": [ + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "text", + "content": "step. Specifically, we define a set of supported bit precis \nsions as " + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "inline_equation", + "content": "\\mathcal{B} = \\{b_{low},b_{med},b_{high}\\}" + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "text", + "content": " . Then, the probabilities of \nselecting " + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "inline_equation", + "content": "b_{low},b_{med}" + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "text", + "content": " , and " + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "inline_equation", + "content": "b_{high}" + }, + { + "bbox": [ + 55, + 72, + 297, + 109 + ], + "type": "text", + "content": " are obtained as" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 91, + 112, + 218, + 127 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 91, + 112, + 218, + 127 + ], + "spans": [ + { + "bbox": [ + 91, + 112, + 218, + 127 + ], + "type": "interline_equation", + "content": "\\mathbf {p} _ {b _ {l o w}} ^ {t} = \\left(1 - \\mathbf {p} _ {q}\\right) \\odot \\left(1 - \\mathbf {p} _ {m} ^ {t}\\right),", + "image_path": "0a8d72bad006adee4658a427c9fda916502c6dffc1623c86a7cbb6f753b7847a.jpg" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 90, + 129, + 295, + 144 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 90, + 129, + 295, + 144 + ], + "spans": [ + { + "bbox": [ + 90, + 129, + 295, + 144 + ], + "type": "interline_equation", + "content": "\\mathbf {p} _ {b _ {m e d}} ^ {t} = \\left(1 - \\mathbf {p} _ {q}\\right) \\odot \\mathbf {p} _ {m} ^ {t} + \\mathbf {p} _ {q} \\odot \\left(1 - \\mathbf {p} _ {h} ^ {t}\\right), \\tag {6}", + "image_path": "09588836cd754ec0b30c374faefda548356df29f1680f8be7f76fdeca3c6b9aa.jpg" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 89, + 145, + 165, + 161 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 89, + 145, + 165, + 161 + ], + "spans": [ + { + "bbox": [ + 89, + 145, + 165, + 161 + ], + "type": "interline_equation", + "content": "\\mathbf {p} _ {b _ {h i g h}} ^ {t} = \\mathbf {p} _ {q} \\odot \\mathbf {p} _ {h} ^ {t},", + "image_path": "3e2587f67380f0b9ddaa51756ed871af3fd61ccf256e6f8eef2d528186ef8eb8.jpg" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "spans": [ + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "inline_equation", + "content": "\\odot" + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": " represents element-wise multiplication, " + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_{b_{low}}^{t} \\in \\mathbb{R}^{K}" + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_{b_{med}}^{t} \\in \\mathbb{R}^{K}" + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_{b_{high}}^{t} \\in \\mathbb{R}^{K}" + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": " are the probability vectors for bit selection. Finally, the bit precision for each layer at time step " + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 55, + 164, + 296, + 214 + ], + "type": "text", + "content": " is determined as follows:" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 61, + 217, + 296, + 240 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 217, + 296, + 240 + ], + "spans": [ + { + "bbox": [ + 61, + 217, + 296, + 240 + ], + "type": "interline_equation", + "content": "\\mathbf {b} ^ {t} (k) = \\operatorname * {a r g m a x} _ {i \\in \\{b _ {\\text {l o w}}, b _ {\\text {m e d}}, b _ {\\text {h i g h}} \\}} \\mathbf {p} _ {i} ^ {t} (k), k = 1, 2, \\dots , K. \\tag {7}", + "image_path": "f1389e42cc1926cdc714b9edcd3cc236c10fafac47b88119faf4ea7e2825260c.jpg" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "spans": [ + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": "However, the training of " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_m^t" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_h^t" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " for all " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " poses a challenge as " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "T" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " increases. In addition, the activation distributions of the diffusion models are known to be similar to each other for adjacent time steps [13, 47]. Consequently, we learn the parameters for every " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "M" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " time step, i.e., " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_j^{Mt}" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " and use the same parameters for the intermediate time steps, i.e., " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_j^{M(t + 1) - 1} = \\mathbf{u}_j^{M(t + 1) - 2} = \\dots = \\mathbf{u}_j^{Mt}" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "j\\in \\{m,h\\}" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": ". Next, previous studies [1, 6] revealed that the initial stage of the reverse process involves generating the rough context of an image corresponding to the input text prompt; thus, the use of low bits in the initial time steps could decrease the correlation between the text prompt and the generated image. Therefore, we set all values in " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " to 1 for the first " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "m" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " time steps in Equation (6) to enforce the selection of high bits in the initial stage of the reverse process. In summary, the Q2B module has " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "2K" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " (" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{s}" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{o}" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": ") and " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "2KT / M" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " (" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_m^t" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_h^t" + }, + { + "bbox": [ + 55, + 244, + 296, + 449 + ], + "type": "text", + "content": ") trainable parameters." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 449, + 296, + 496 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 449, + 296, + 496 + ], + "spans": [ + { + "bbox": [ + 55, + 449, + 296, + 496 + ], + "type": "text", + "content": "The bit selection in Equation (7) is non-differentiable, making it unable to backpropagate. Thus, we adopt a straight-through estimator [3] to render this process differentiable:" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 65, + 499, + 100, + 513 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 65, + 499, + 100, + 513 + ], + "spans": [ + { + "bbox": [ + 65, + 499, + 100, + 513 + ], + "type": "interline_equation", + "content": "\\hat {\\mathbf {a}} ^ {t} (k) =", + "image_path": "a8f603ac29540f4685d473d17440a2425b348bc599425c50fee50dde839d121f.jpg" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 67, + 516, + 296, + 557 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 516, + 296, + 557 + ], + "spans": [ + { + "bbox": [ + 67, + 516, + 296, + 557 + ], + "type": "interline_equation", + "content": "\\left\\{ \\begin{array}{l l} Q _ {\\mathbf {b} ^ {t} (k)} \\left(\\mathbf {a} ^ {t} (k)\\right) & \\text {f o r w a r d}, \\\\ \\sum_ {i \\in \\left\\{b _ {l o w}, b _ {m e d}, b _ {h i g h} \\right\\}} \\mathbf {p} _ {i} ^ {t} (k) \\cdot Q _ {i} \\left(\\mathbf {a} ^ {t} (k)\\right) & \\text {b a c k w a r d}. \\end{array} \\right. \\tag {8}", + "image_path": "61c8eb81c0e2392d05292835303456065bda41689f7bdbdd21dd8dbb38c0a254.jpg" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "spans": [ + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": "Here, " + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "inline_equation", + "content": "\\mathbf{a}^t (k)" + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "inline_equation", + "content": "\\hat{\\mathbf{a}}^t (k)" + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": " represent the activations of the full-precision and quantized models at time step " + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": " of the " + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "inline_equation", + "content": "k" + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": "-th layer. In other words, during backpropagation, the activations quantized by all bit precisions in " + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "inline_equation", + "content": "\\mathcal{B}" + }, + { + "bbox": [ + 55, + 558, + 296, + 665 + ], + "type": "text", + "content": " are weighted by their corresponding estimated probabilities. Since diffusion models are less sensitive to the quantization of the weights [13, 24], the weights are quantized using a fixed precision, which is empirically chosen as 4 bits for all layers and time steps." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 55, + 665, + 296, + 714 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 665, + 296, + 714 + ], + "spans": [ + { + "bbox": [ + 55, + 665, + 296, + 714 + ], + "type": "text", + "content": "Given the full precision diffusion model, we obtain its quantized weights by applying the baseline diffusion quantization methods [13, 24]. The quantization parameters, including the maximum and minimum values for clipping, are" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "spans": [ + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "text", + "content": "also predetermined for all bit precisions in " + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "inline_equation", + "content": "\\mathcal{B}" + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "text", + "content": ". Then, while freezing all the other modules in Figure 3, only the Q2B module is trained using a small calibration set. The Q2B module is trained at randomly sampled time steps " + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "text", + "content": " using the loss function " + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "inline_equation", + "content": "L_{QLIP}" + }, + { + "bbox": [ + 313, + 72, + 554, + 133 + ], + "type": "text", + "content": ":" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 320, + 139, + 554, + 199 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 320, + 139, + 554, + 199 + ], + "spans": [ + { + "bbox": [ + 320, + 139, + 554, + 199 + ], + "type": "interline_equation", + "content": "\\begin{array}{l} L _ {Q L I P} = \\left(\\epsilon_ {\\theta} (\\mathbf {x} _ {t}, t) - \\hat {\\epsilon} _ {\\theta} (\\hat {\\mathbf {x}} _ {t}, t)\\right) ^ {2} \\\\ + \\lambda_ {b i t} \\left(b _ {h i g h} \\cdot \\sum_ {k = 1} ^ {K} \\mathbf {p} _ {b _ {h i g h}} ^ {t} (k) + b _ {m e d} \\cdot \\sum_ {k = 1} ^ {K} \\mathbf {p} _ {b _ {m e d}} ^ {t} (k)\\right). \\tag {9} \\\\ \\end{array}", + "image_path": "7699f83e50ee27a74dd77af40162f5d64c73a903ccb3833c595e7b94add61e23.jpg" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 200, + 554, + 259 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 200, + 554, + 259 + ], + "spans": [ + { + "bbox": [ + 313, + 200, + 554, + 259 + ], + "type": "text", + "content": "Here, the first term measures the difference between the results of the full-precision and quantized models, and the second term measures the bit lengths for the high- and medium-precision models. " + }, + { + "bbox": [ + 313, + 200, + 554, + 259 + ], + "type": "inline_equation", + "content": "\\lambda_{bit}" + }, + { + "bbox": [ + 313, + 200, + 554, + 259 + ], + "type": "text", + "content": " is a weighting parameter that balances between these two terms." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 313, + 270, + 395, + 283 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 270, + 395, + 283 + ], + "spans": [ + { + "bbox": [ + 313, + 270, + 395, + 283 + ], + "type": "text", + "content": "4. Experiments" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 289, + 429, + 303 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 289, + 429, + 303 + ], + "spans": [ + { + "bbox": [ + 313, + 289, + 429, + 303 + ], + "type": "text", + "content": "4.1. Experimental setup" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 307, + 555, + 629 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 307, + 555, + 629 + ], + "spans": [ + { + "bbox": [ + 313, + 307, + 555, + 629 + ], + "type": "text", + "content": "Datasets and baseline quantization methods. To evaluate the effectiveness of QLIP across different diffusion models, we conducted experiments using two diffusion networks: BK-SDM-Tiny-2M [19] and Stable Diffusion v1.4 [39]. Our experiments involve two datasets, COCO2017 [27] and Conceptual Captions [46]. To train the T2Q module, 10k images were generated by Stable Diffusion v1.4 [39] using 10k text annotations in the COCO2017 training dataset as input text prompts. We generated another 1024 images using unused text annotations in the COCO2017 training dataset to train the Q2B module. We evaluated the effectiveness of our proposed QLIP by applying it to the baseline diffusion quantization methods, Q-diffusion [24] and PTQD [13]. In the implementation of PTQD, we set the noise correction parameter to " + }, + { + "bbox": [ + 313, + 307, + 555, + 629 + ], + "type": "inline_equation", + "content": "b_{med}" + }, + { + "bbox": [ + 313, + 307, + 555, + 629 + ], + "type": "text", + "content": ", while noise correction was not applied if the estimated image quality by the T2Q module is below 0.3 or above 0.7. We trained the T2Q module for 3 epochs using the Adam optimizer [20] with a learning rate of 0.001 and the Q2B module for 5000 iterations using BRECQ [25], as adopted in Q-diffusion and PTQD, using RTX 3090. For performance evaluation, following the baseline quantization method [24], we generated 10k images using 10k text annotations in the COCO2017 validation dataset. Moreover, to evaluate the generalization ability of the quantized models, we used Conceptual Captions to generate additional 10k images." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 313, + 630, + 554, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 630, + 554, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 630, + 554, + 713 + ], + "type": "text", + "content": "Evaluation metrics. We report widely adopted metrics such as FID [15] and sFID [44] to evaluate image quality. In addition, we use CLIP Score [14] to assess the degree of matching between the generated images and their corresponding text prompts. To quantify computational efficiency, we measure BitOPs [22, 53], which account for the required operations in both the denoising model and addi" + } + ] + } + ], + "index": 18 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "text", + "content": "15430" + } + ] + } + ], + "index": 19 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 4 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 56, + 71, + 555, + 163 + ], + "blocks": [ + { + "bbox": [ + 56, + 71, + 555, + 163 + ], + "lines": [ + { + "bbox": [ + 56, + 71, + 555, + 163 + ], + "spans": [ + { + "bbox": [ + 56, + 71, + 555, + 163 + ], + "type": "table", + "html": "
COCO2017
MethodBitoptionsFAB↓BitOPs (T)↓FID↓sFID↓CLIP Score↑
BK-SDM [19]W32A3232.0010.4623.7966.190.3069
Q-diffusion [24]W4A1616.001.0330.0273.250.3068
+QLIPW4A{8,16,32}12.140.8830.0173.240.3063
PTQD [13]W4A1616.001.0330.2777.180.3069
+QLIPW4A{8,16,32}12.140.8830.0273.260.3063
", + "image_path": "ca2bb2ebb84b59dc44eadd91a3fcdfcfb45924e662480e934f03e41c8e6b09b7.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "type": "table", + "bbox": [ + 56, + 163, + 555, + 252 + ], + "blocks": [ + { + "bbox": [ + 56, + 163, + 555, + 252 + ], + "lines": [ + { + "bbox": [ + 56, + 163, + 555, + 252 + ], + "spans": [ + { + "bbox": [ + 56, + 163, + 555, + 252 + ], + "type": "table", + "html": "
Conceptual Captions
MethodBitoptionsFAB↓BitOPs (T)↓FID↓sFID↓CLIP Score↑
BK-SDM [19]W32A3232.0010.4620.0047.800.2983
Q-diffusion [24]W4A1616.001.0324.6862.700.2965
+QLIPW4A{8,16,32}10.580.8224.7259.540.2964
PTQD [13]W4A1616.001.0324.6662.600.2965
+QLIPW4A{8,16,32}10.580.8224.7159.250.2966
", + "image_path": "004a962e7c72cb89b5b418b838a66c76a8ada03a5f368b33ec81e8cf060cb041.jpg" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_body" + } + ], + "index": 1 + }, + { + "type": "table", + "bbox": [ + 57, + 289, + 294, + 472 + ], + "blocks": [ + { + "bbox": [ + 85, + 258, + 523, + 270 + ], + "lines": [ + { + "bbox": [ + 85, + 258, + 523, + 270 + ], + "spans": [ + { + "bbox": [ + 85, + 258, + 523, + 270 + ], + "type": "text", + "content": "Table 1. Quantitative comparisons of various quantization methods at a resolution of " + }, + { + "bbox": [ + 85, + 258, + 523, + 270 + ], + "type": "inline_equation", + "content": "{512} \\times {512}" + }, + { + "bbox": [ + 85, + 258, + 523, + 270 + ], + "type": "text", + "content": " using BK-SDM-Tiny-2M." + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 57, + 289, + 294, + 472 + ], + "lines": [ + { + "bbox": [ + 57, + 289, + 294, + 472 + ], + "spans": [ + { + "bbox": [ + 57, + 289, + 294, + 472 + ], + "type": "table", + "html": "
COCO2017
MethodFAB↓FID↓sFID↓CLIP Score↑
Stable Diffusion [39]32.0022.2365.110.3174
Q-diffusion [24]8.0023.4066.570.3126
+QLIP7.8621.6164.320.3120
PTQD [13]8.0022.7568.630.3126
+QLIP7.8621.3565.810.3120
Conceptual Captions
MethodFAB↓FID↓sFID↓CLIP Score↑
Stable Diffusion [39]32.0015.6745.610.3078
Q-diffusion [24]8.0020.7448.330.2984
+QLIP7.7719.8947.630.2979
PTQD [13]8.0025.4055.010.2972
+QLIP7.7720.4248.830.2975
", + "image_path": "30cfcdc1a6a3afa28ff14724ca66374030c63e9ecdffaffae304df6cda1fca67.jpg" + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "table_body" + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "spans": [ + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "content": "tional modules for QLIP implementation, and feature average bit-width (FAB) [17, 18] for the denoising model. In particular, BitOPs are measured as " + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "inline_equation", + "content": "MACs \\cdot \\frac{b_w}{32} \\cdot \\frac{b_a}{32}" + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "content": ", where " + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "inline_equation", + "content": "MACs" + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "content": " denotes multiply-accumulate computations, and " + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "inline_equation", + "content": "b_w" + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "inline_equation", + "content": "b_a" + }, + { + "bbox": [ + 55, + 545, + 295, + 617 + ], + "type": "text", + "content": " represent the bitwidths of weights and activations, respectively." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "spans": [ + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "content": "Hyperparameter settings. We conducted experiments on various hyperparameter settings for the COCO2017 validation dataset using Q-diffusion as the baseline in Stable Diffusion v1.4. First, Figure 4(a) presents the results obtained with various values of the weight term " + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "inline_equation", + "content": "\\lambda_{bit}" + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "content": " in Equation (9), where " + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "inline_equation", + "content": "\\lambda_{bit} = 1" + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "content": " showed an effective balance between FID and FAB. Next, Figure 4(b) shows the results for different " + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "inline_equation", + "content": "M" + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "content": " values in " + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "inline_equation", + "content": "\\mathbf{u}_j^{Mt}, j\\in \\{m,h\\}" + }, + { + "bbox": [ + 55, + 618, + 295, + 715 + ], + "type": "text", + "content": " when the maximum timestep" + } + ] + } + ], + "index": 6 + }, + { + "type": "table", + "bbox": [ + 315, + 289, + 556, + 354 + ], + "blocks": [ + { + "bbox": [ + 55, + 479, + 295, + 523 + ], + "lines": [ + { + "bbox": [ + 55, + 479, + 295, + 523 + ], + "spans": [ + { + "bbox": [ + 55, + 479, + 295, + 523 + ], + "type": "text", + "content": "Table 2. Quantitative comparisons of various quantization methods at a resolution of " + }, + { + "bbox": [ + 55, + 479, + 295, + 523 + ], + "type": "inline_equation", + "content": "512\\times 512" + }, + { + "bbox": [ + 55, + 479, + 295, + 523 + ], + "type": "text", + "content": " using Stable Diffusion v1.4. For the bit precision options, W4A8 is used for Q-diffusion and PTQD, and W4A{6,8,10} are used for QLIP." + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 315, + 289, + 556, + 354 + ], + "lines": [ + { + "bbox": [ + 315, + 289, + 556, + 354 + ], + "spans": [ + { + "bbox": [ + 315, + 289, + 556, + 354 + ], + "type": "table", + "html": "
MethodBitoptionsRuntime (s)↓FAB↓FID↓
BK-SDM [19]W32A326.5032.0020.00
Q-diffusion [24]W4A84.538.0028.32
Q-diffusion [24]W4A165.5816.0024.68
+QLIPW4A{8,16,32}4.8510.5824.72
", + "image_path": "7d7ab1170578e7defac034e9ae1cc329b60e5430439d381860a725b32f21f9d9.jpg" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_body" + } + ], + "index": 7 + }, + { + "bbox": [ + 313, + 361, + 555, + 385 + ], + "lines": [ + { + "bbox": [ + 313, + 361, + 555, + 385 + ], + "spans": [ + { + "bbox": [ + 313, + 361, + 555, + 385 + ], + "type": "text", + "content": "Table 3. Runtime comparisons, where W4A{8,16,32} is used for QLIP. Measured on the Conceptual Captions dataset." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "text" + }, + { + "type": "image", + "bbox": [ + 315, + 395, + 432, + 485 + ], + "blocks": [ + { + "bbox": [ + 315, + 395, + 432, + 485 + ], + "lines": [ + { + "bbox": [ + 315, + 395, + 432, + 485 + ], + "spans": [ + { + "bbox": [ + 315, + 395, + 432, + 485 + ], + "type": "image", + "image_path": "66a7115134cd2c554b04b9f63bf7ba14c45fb8d1169aae7c49db9a7befd6be79.jpg" + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 381, + 485, + 389, + 493 + ], + "lines": [ + { + "bbox": [ + 381, + 485, + 389, + 493 + ], + "spans": [ + { + "bbox": [ + 381, + 485, + 389, + 493 + ], + "type": "text", + "content": "(a)" + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_caption" + } + ], + "index": 9 + }, + { + "type": "image", + "bbox": [ + 433, + 396, + 553, + 484 + ], + "blocks": [ + { + "bbox": [ + 433, + 396, + 553, + 484 + ], + "lines": [ + { + "bbox": [ + 433, + 396, + 553, + 484 + ], + "spans": [ + { + "bbox": [ + 433, + 396, + 553, + 484 + ], + "type": "image", + "image_path": "92b289e7a2dc3dab7c5a4ddf27e1f7efff5167b6e9fa91450f34cedcdfc7c06d.jpg" + } + ] + } + ], + "index": 11, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 501, + 485, + 510, + 493 + ], + "lines": [ + { + "bbox": [ + 501, + 485, + 510, + 493 + ], + "spans": [ + { + "bbox": [ + 501, + 485, + 510, + 493 + ], + "type": "text", + "content": "(b)" + } + ] + } + ], + "index": 12, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "lines": [ + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "spans": [ + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "type": "text", + "content": "Figure 4. Results on Q-diffusion using different hyperparameter settings of QLIP (W4A{6,8,10}): (a) " + }, + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "type": "inline_equation", + "content": "\\lambda_{bit}" + }, + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "type": "text", + "content": " and (b) " + }, + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "type": "inline_equation", + "content": "M" + }, + { + "bbox": [ + 313, + 502, + 555, + 525 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 13, + "angle": 0, + "type": "image_caption" + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "spans": [ + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "type": "inline_equation", + "content": "T" + }, + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "type": "text", + "content": " was 1000, where we obtained the best scores with " + }, + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "type": "inline_equation", + "content": "M = 200" + }, + { + "bbox": [ + 313, + 544, + 555, + 568 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 313, + 576, + 430, + 588 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 576, + 430, + 588 + ], + "spans": [ + { + "bbox": [ + 313, + 576, + 430, + 588 + ], + "type": "text", + "content": "4.2. Quantitative results" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 594, + 555, + 689 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 594, + 555, + 689 + ], + "spans": [ + { + "bbox": [ + 313, + 594, + 555, + 689 + ], + "type": "text", + "content": "Comparison with existing quantization methods. To demonstrate the effectiveness and applicability of QLIP, we tested QLIP on two diffusion model quantization methods: Q-diffusion [24] and PTQD [13]. Tables 1 and 2 provide comparison results for the COCO2017 [27] and Conceptual Captions [46] datasets using lightweight diffusion model BK-SDM-Tiny-2M [19] and Stable Diffusion v1.4 [39], respectively." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 689, + 554, + 713 + ], + "type": "text", + "content": "As shown in Table 1, for the COCO2017 dataset, QLIP effectively reduced computational complexity of the base" + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15431" + } + ] + } + ], + "index": 18 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 5 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 85, + 70, + 526, + 228 + ], + "blocks": [ + { + "bbox": [ + 85, + 70, + 526, + 228 + ], + "lines": [ + { + "bbox": [ + 85, + 70, + 526, + 228 + ], + "spans": [ + { + "bbox": [ + 85, + 70, + 526, + 228 + ], + "type": "image", + "image_path": "3ba0280fd6a05472baa38fa2c4ace5ffcc10e002c2a9b7dcf1f64e80c2e8de31.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 55, + 236, + 555, + 248 + ], + "lines": [ + { + "bbox": [ + 55, + 236, + 555, + 248 + ], + "spans": [ + { + "bbox": [ + 55, + 236, + 555, + 248 + ], + "type": "text", + "content": "Figure 5. Examples of variations in FAB by QLIP for the texts with different levels of richness and detail, along with the generated images." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 83, + 258, + 526, + 434 + ], + "blocks": [ + { + "bbox": [ + 83, + 258, + 526, + 434 + ], + "lines": [ + { + "bbox": [ + 83, + 258, + 526, + 434 + ], + "spans": [ + { + "bbox": [ + 83, + 258, + 526, + 434 + ], + "type": "image", + "image_path": "9452066d6a2a8f82a7aed4389a807bdf6f4ff3ffd6c433e5465dc93f57dcbc78.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 54, + 443, + 555, + 499 + ], + "lines": [ + { + "bbox": [ + 54, + 443, + 555, + 499 + ], + "spans": [ + { + "bbox": [ + 54, + 443, + 555, + 499 + ], + "type": "text", + "content": "Figure 6. Examples of the bit selection results and generated images. When the text prompt is specifically describing the image to be generated, the predicted image quality is high, and high and medium bits are assigned for most layers and time steps, as shown on the top left. On the other hand, when the text prompt has less detailed information, the predicted image quality is low, and low bits are assigned for many layers and time steps, as shown on the bottom left. On the right side, the generated images using Q-diffusion with and without applying QLIP are shown, demonstrating the effectiveness of QLIP in saving bits while improving the quality of generated images." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_caption" + } + ], + "index": 2 + }, + { + "bbox": [ + 54, + 519, + 294, + 674 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 519, + 294, + 674 + ], + "spans": [ + { + "bbox": [ + 54, + 519, + 294, + 674 + ], + "type": "text", + "content": "line diffusion quantization methods. For example, Q-diffusion+QLIP (W4A{8,16,32}) achieved a reduction in FAB by 3.86 while maintaining similar FID and sFID, compared to Q-diffusion (W4A16). Furthermore, the use of QLIP consistently led to improvements when applied to PTQD, as demonstrated in Table 1. For example, PTQD+QLIP (W4A{8,16,32}) enhanced FID and sFID by 0.25 and 3.92, respectively, while also reducing FAB and BitOPs. The results on Conceptual Captions followed trends similar to those observed in COCO2017. For example, Q-diffusion+QLIP (W4A{8,16,32}) demonstrated reduced computational complexity compared to Q-diffusion on Conceptual Captions." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 677, + 295, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 677, + 295, + 715 + ], + "spans": [ + { + "bbox": [ + 55, + 677, + 295, + 715 + ], + "type": "text", + "content": "Table 2 presents the results of applying lower-bit precision combinations to Stable Diffusion v1.4. Here, PTQD+QLIP (W4A{6,8,10}) achieved a 4.98 improvement" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 313, + 519, + 555, + 628 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 519, + 555, + 628 + ], + "spans": [ + { + "bbox": [ + 313, + 519, + 555, + 628 + ], + "type": "text", + "content": "in FID compared to PTQD (W4A8), while maintaining similar FAB using bit combinations efficiently. Although our proposed QLIP drives the quantization methods to reduce computational complexity and improve image quality, the CLIP Scores remained similar to those of the baseline quantization methods, indicating that the fitness to the text prompts is not significantly influenced by QLIP. Additional results on larger diffusion models can be found in the supplementary material." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 313, + 630, + 556, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 630, + 556, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 630, + 556, + 715 + ], + "type": "text", + "content": "Runtime measurement. Runtime was measured with Cutlass [34], including the T2Q and Q2B modules, as shown in Table 3. The results show that runtime was reduced to a level similar to W4A8 while maintaining an FID score comparable to W4A16. Measurements were conducted on an RTX 3090. Additional analysis on memory requirements can be found in the supplementary material." + } + ] + } + ], + "index": 7 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 758 + ], + "type": "text", + "content": "15432" + } + ] + } + ], + "index": 8 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 6 + }, + { + "para_blocks": [ + { + "bbox": [ + 55, + 72, + 165, + 84 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 72, + 165, + 84 + ], + "spans": [ + { + "bbox": [ + 55, + 72, + 165, + 84 + ], + "type": "text", + "content": "4.3. Qualitative results" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 89, + 296, + 220 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 89, + 296, + 220 + ], + "spans": [ + { + "bbox": [ + 55, + 89, + 296, + 220 + ], + "type": "text", + "content": "Figure 5 visualizes the changes in FAB by the richness and detail level of the text, along with the generated images. Here, ChatGPT-4o was used to create texts of varying complexity that describe a given keyword. Specifically, the prompt used was \"Create four sentences that describe how 'keyword' looks like, gradually getting richer and longer.\" As can be seen, QLIP assigned higher bits as the text describes the keyword more specifically and in greater detail, demonstrating that it utilizes information from the text to determine bit precisions for effective synthesis of vivid details and textures." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 55, + 221, + 296, + 425 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 221, + 296, + 425 + ], + "spans": [ + { + "bbox": [ + 55, + 221, + 296, + 425 + ], + "type": "text", + "content": "Figure 6 shows several results comparing Q-diffusion with and without QLIP when applied to Stable Diffusion v1.4. As can be seen, our QLIP generates more detailed and natural images compared to PTQD by using various bit combinations. Furthermore, Figure 6 visualizes the bit precision maps determined by QLIP. When the T2Q module predicts high image quality from input text prompts, QLIP allocates more bits to enable rendering of vivid and accurate details from rich text prompts. Otherwise, QLIP employs lower bits for layers that are less sensitive to quantization, based on the text prompt and the time step. For example, when given less specific text prompts, the correlation between the generated image and the text prompt is relatively less critical, allowing for low-bit quantization in the cross-attention blocks. These results highlight the effectiveness and significance of dynamically allocating bit precisions according to text prompts." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 432, + 154, + 444 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 432, + 154, + 444 + ], + "spans": [ + { + "bbox": [ + 55, + 432, + 154, + 444 + ], + "type": "text", + "content": "4.4. Ablation studies" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 450, + 295, + 605 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 450, + 295, + 605 + ], + "spans": [ + { + "bbox": [ + 55, + 450, + 295, + 605 + ], + "type": "text", + "content": "Ablation studies on the T2Q module. We report the results on the COCO2017 validation dataset using Stable Diffusion v1.4 with Q-diffusion as the baseline quantization method. Table 4 presents the results of the T2Q modules trained using various perceptual image quality metrics, including Realism score [21], CLIP-IQA [54], and GIQA [11]. The performance of the T2Q module was evaluated using Spearman's rank-order correlation coefficient (SROCC) [49] and Pearson's linear correlation coefficient (PLCC) [36] between the actual quality measures and the predictions of the T2Q module. QLIP with GIQA showed the most substantial improvements in FAB and FID along with the highest SROCC and PLCC scores." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "spans": [ + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": "Ablation studies on the Q2B module. We conducted experiments on several variants of the Q2B module using Q-diffusion and Stable Diffusion v1.4 on the COCO2017 validation dataset. As shown in Table 5, when using either " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": " alone or in combination with " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_h^t" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": ", effective reduction in FAB was achieved at the expense of increase in FID. When using " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_m^t" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": ", FID decreased while FAB increased. Finally, when " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_q" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_h^t" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "inline_equation", + "content": "\\mathbf{p}_m^t" + }, + { + "bbox": [ + 55, + 605, + 295, + 714 + ], + "type": "text", + "content": " were all involved, balanced results were obtained in two metrics. Implementation details of" + } + ] + } + ], + "index": 5 + }, + { + "type": "table", + "bbox": [ + 317, + 70, + 553, + 148 + ], + "blocks": [ + { + "bbox": [ + 317, + 70, + 553, + 148 + ], + "lines": [ + { + "bbox": [ + 317, + 70, + 553, + 148 + ], + "spans": [ + { + "bbox": [ + 317, + 70, + 553, + 148 + ], + "type": "table", + "html": "
MethodSROCC↑PLCC↑FAB↓FID↓
Q-diffusion [24]--8.0023.40
+QLIP
w/ Realism score0.51320.50228.1022.18
w/ CLIP-IQA0.71270.70828.5421.81
w/ GIQA0.80470.81087.8621.61
", + "image_path": "0cd2d20b96951a579726cc2e44876115471f56ef637a10687a1c82b00a242583.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "table_body" + } + ], + "index": 6 + }, + { + "type": "table", + "bbox": [ + 317, + 202, + 552, + 281 + ], + "blocks": [ + { + "bbox": [ + 313, + 155, + 553, + 187 + ], + "lines": [ + { + "bbox": [ + 313, + 155, + 553, + 187 + ], + "spans": [ + { + "bbox": [ + 313, + 155, + 553, + 187 + ], + "type": "text", + "content": "Table 4. Ablation studies on the T2Q module: For the bit precision options, W4A8 in Q-diffusion and +QLIP W4A{6,8,10} were used." + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 317, + 202, + 552, + 281 + ], + "lines": [ + { + "bbox": [ + 317, + 202, + 552, + 281 + ], + "spans": [ + { + "bbox": [ + 317, + 202, + 552, + 281 + ], + "type": "table", + "html": "
MethodBitoptionsFAB↓FID↓
Q-diffusion [24]W4A88.0023.40
+ pqW4A{6,10}7.5726.91
+ pq + ptW4A{6,8,10}6.7329.37
+ pq + pmtW4A{6,8,10}8.6021.96
+ pq + pmt + pth (QLIP)W4A{6,8,10}7.8621.61
", + "image_path": "4ead45ab3e361af5c5db5cb5122037a125bcd30b4fcea725c9d0ffb4c57ed0cc.jpg" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "table_body" + } + ], + "index": 8 + }, + { + "bbox": [ + 350, + 289, + 518, + 300 + ], + "lines": [ + { + "bbox": [ + 350, + 289, + 518, + 300 + ], + "spans": [ + { + "bbox": [ + 350, + 289, + 518, + 300 + ], + "type": "text", + "content": "Table 5. Ablation studies on the Q2B module." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 313, + 325, + 553, + 349 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 325, + 553, + 349 + ], + "spans": [ + { + "bbox": [ + 313, + 325, + 553, + 349 + ], + "type": "text", + "content": "the ablation studies on the Q2B module can be found in the supplementary material." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 365, + 388, + 377 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 365, + 388, + 377 + ], + "spans": [ + { + "bbox": [ + 313, + 365, + 388, + 377 + ], + "type": "text", + "content": "5. Conclusion" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 312, + 387, + 555, + 566 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 312, + 387, + 555, + 566 + ], + "spans": [ + { + "bbox": [ + 312, + 387, + 555, + 566 + ], + "type": "text", + "content": "The proposed quantization method for diffusion models, called QLIP, is the first method that uses text prompts for quantization in diffusion models. From our observation that the bit precision for quantization can be effectively determined based on the quality of the generated images, we designed the T2Q module that predicts image quality from input text prompts. We then introduced the Q2B module that allocates appropriate bit precision for each time step for different layers. By integrating the T2Q and Q2B modules into existing diffusion model quantization methods, we demonstrated the effectiveness of the proposed QLIP. QLIP presents a promising solution for quantizing diffusion models due to its ability to improve the quality of generated images using text prompts while reducing computational costs." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 313, + 567, + 554, + 640 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 567, + 554, + 640 + ], + "spans": [ + { + "bbox": [ + 313, + 567, + 554, + 640 + ], + "type": "text", + "content": "Limitations and future directions. Similar to other generative models, QLIP also has the potential to generate counterfeit images for malicious purposes, potentially undermining trust in visual media. In a future study, we will extend QLIP to consider other input conditions, such as images or segmentation maps, in addition to text prompts." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 656, + 411, + 669 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 656, + 411, + 669 + ], + "spans": [ + { + "bbox": [ + 313, + 656, + 411, + 669 + ], + "type": "text", + "content": "Acknowledgement" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 313, + 677, + 553, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 677, + 553, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 677, + 553, + 713 + ], + "type": "text", + "content": "This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS-2022-NR070077)." + } + ] + } + ], + "index": 15 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15433" + } + ] + } + ], + "index": 16 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 7 + }, + { + "para_blocks": [ + { + "bbox": [ + 56, + 71, + 115, + 83 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 71, + 115, + 83 + ], + "spans": [ + { + "bbox": [ + 56, + 71, + 115, + 83 + ], + "type": "text", + "content": "References" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 57, + 91, + 296, + 714 + ], + "type": "list", + "angle": 0, + "index": 14, + "blocks": [ + { + "bbox": [ + 61, + 91, + 296, + 145 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 91, + 296, + 145 + ], + "spans": [ + { + "bbox": [ + 61, + 91, + 296, + 145 + ], + "type": "text", + "content": "[1] Yogesh Balaji, Seungjun Nah, Xun Huang, Arash Vahdat, Jiaming Song, Qinsheng Zhang, Karsten Kreis, Miika Aittala, Timo Aila, Samuli Laine, et al. eDiff-I: Text-to-image diffusion models with ensemble of expert denoisers. arXiv preprint arXiv:2211.01324, 2022. 5" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 61, + 147, + 296, + 200 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 147, + 296, + 200 + ], + "spans": [ + { + "bbox": [ + 61, + 147, + 296, + 200 + ], + "type": "text", + "content": "[2] Fan Bao, Chongxuan Li, Jun Zhu, and Bo Zhang. Analytic-DPM: An analytic estimate of the optimal reverse variance in diffusion probabilistic models. In Proceedings of the International Conference on Learning Representations, 2022. 1" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 62, + 202, + 295, + 245 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 202, + 295, + 245 + ], + "spans": [ + { + "bbox": [ + 62, + 202, + 295, + 245 + ], + "type": "text", + "content": "[3] Yoshua Bengio, Nicholas Léonard, and Aaron Courville. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432, 2013.5" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 62, + 247, + 294, + 301 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 247, + 294, + 301 + ], + "spans": [ + { + "bbox": [ + 62, + 247, + 294, + 301 + ], + "type": "text", + "content": "[4] James Betker, Gabriel Goh, Li Jing, Tim Brooks, Jianfeng Wang, Linjie Li, Long Ouyang, Juntang Zhuang, Joyce Lee, Yufei Guo, et al. Improving image generation with better captions. Computer Science. https://cdn.openai.com/papers/dall-e-3.pdf, 2023.1" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 62, + 302, + 295, + 356 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 302, + 295, + 356 + ], + "spans": [ + { + "bbox": [ + 62, + 302, + 295, + 356 + ], + "type": "text", + "content": "[5] Jungwook Choi, Zhuo Wang, Swagath Venkataramani, Pierce I-Jen Chuang, Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan. PACT: Parameterized clipping activation for quantized neural networks. arXiv preprint arXiv:1805.06085, 2018. 2" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 62, + 357, + 296, + 413 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 357, + 296, + 413 + ], + "spans": [ + { + "bbox": [ + 62, + 357, + 296, + 413 + ], + "type": "text", + "content": "[6] Jooyoung Choi, Jungbeom Lee, Chaehun Shin, Sungwon Kim, Hyunwoo Kim, and Sungroh Yoon. Perception prioritized training of diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11472-11481, 2022. 5" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 62, + 414, + 294, + 445 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 414, + 294, + 445 + ], + "spans": [ + { + "bbox": [ + 62, + 414, + 294, + 445 + ], + "type": "text", + "content": "[7] Huanpeng Chu, Wei Wu, Chengjie Zang, and Kun Yuan. QNCD: Quantization noise correction for diffusion models. arXiv preprint arXiv:2403.19140, 2024. 2" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 62, + 447, + 294, + 491 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 447, + 294, + 491 + ], + "spans": [ + { + "bbox": [ + 62, + 447, + 294, + 491 + ], + "type": "text", + "content": "[8] Xiangxiang Chu, Shun Lu, Xudong Li, and Bo Zhang. Mix-Path: A unified approach for one-shot neural architecture search. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5972–5981, 2023. 1" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 62, + 491, + 294, + 546 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 491, + 294, + 546 + ], + "spans": [ + { + "bbox": [ + 62, + 491, + 294, + 546 + ], + "type": "text", + "content": "[9] Zhen Dong, Zhewei Yao, Amir Gholami, Michael W Mahoney, and Kurt Keutzer. HAWQ: Hessian aware quantization of neural networks with mixed-precision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 293–302, 2019. 1, 3" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 57, + 547, + 294, + 601 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 547, + 294, + 601 + ], + "spans": [ + { + "bbox": [ + 57, + 547, + 294, + 601 + ], + "type": "text", + "content": "[10] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, pages 2672–2680, 2014. 1" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 57, + 602, + 294, + 645 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 602, + 294, + 645 + ], + "spans": [ + { + "bbox": [ + 57, + 602, + 294, + 645 + ], + "type": "text", + "content": "[11] Shuyang Gu, Jianmin Bao, Dong Chen, and Fang Wen. GIQA: Generated image quality assessment. In Proceedings of the European Conference on Computer Vision, pages 369–385, 2020. 3, 4, 8" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 57, + 647, + 294, + 690 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 647, + 294, + 690 + ], + "spans": [ + { + "bbox": [ + 57, + 647, + 294, + 690 + ], + "type": "text", + "content": "[12] Song Han, Jeff Pool, John Tran, and William Dally. Learning both weights and connections for efficient neural network. In Proceedings of the Advances in Neural Information Processing Systems, pages 1135-1143, 2015. 1" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 57, + 691, + 294, + 714 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 691, + 294, + 714 + ], + "spans": [ + { + "bbox": [ + 57, + 691, + 294, + 714 + ], + "type": "text", + "content": "[13] Yefei He, Luping Liu, Jing Liu, Weijia Wu, Hong Zhou, and Bohan Zhuang. PTQD: Accurate post-training quantization" + } + ] + } + ], + "index": 13 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 316, + 73, + 555, + 713 + ], + "type": "list", + "angle": 0, + "index": 28, + "blocks": [ + { + "bbox": [ + 335, + 73, + 553, + 106 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 335, + 73, + 553, + 106 + ], + "spans": [ + { + "bbox": [ + 335, + 73, + 553, + 106 + ], + "type": "text", + "content": "for diffusion models. In Proceedings of the Advances in Neural Information Processing Systems, pages 13237-13249, 2023. 2, 3, 5, 6" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 316, + 108, + 553, + 163 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 108, + 553, + 163 + ], + "spans": [ + { + "bbox": [ + 316, + 108, + 553, + 163 + ], + "type": "text", + "content": "[14] Jack Hessel, Ari Holtzman, Maxwell Forbes, Ronan Le Bras, and Yejin Choi. CLIPScore: A reference-free evaluation metric for image captioning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7514-7528, 2021. 5" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 317, + 165, + 553, + 220 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 165, + 553, + 220 + ], + "spans": [ + { + "bbox": [ + 317, + 165, + 553, + 220 + ], + "type": "text", + "content": "[15] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the Advances in Neural Information Processing Systems, pages 6629-6640, 2017. 5" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 316, + 222, + 553, + 265 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 222, + 553, + 265 + ], + "spans": [ + { + "bbox": [ + 316, + 222, + 553, + 265 + ], + "type": "text", + "content": "[16] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In Proceedings of the Advances in Neural Information Processing Systems, pages 6840-6851, 2020. 3" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 316, + 268, + 553, + 312 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 268, + 553, + 312 + ], + "spans": [ + { + "bbox": [ + 316, + 268, + 553, + 312 + ], + "type": "text", + "content": "[17] Cheeun Hong and Kyoung Mu Lee. AdaBM: On-the-fly adaptive bit mapping for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2641-2650, 2024. 1, 3, 6" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 317, + 315, + 555, + 369 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 315, + 555, + 369 + ], + "spans": [ + { + "bbox": [ + 317, + 315, + 555, + 369 + ], + "type": "text", + "content": "[18] Cheeun Hong, Sungyong Baik, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee. CADyQ: Content-aware dynamic quantization for image super-resolution. In Proceedings of the European Conference on Computer Vision, pages 367-383, 2022. 1, 2, 3, 6" + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 316, + 371, + 553, + 415 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 371, + 553, + 415 + ], + "spans": [ + { + "bbox": [ + 316, + 371, + 553, + 415 + ], + "type": "text", + "content": "[19] Bo-Kyeong Kim, Hyoung-Kyu Song, Thibault Castells, and Shinkook Choi. BK-SDM: A lightweight, fast, and cheap version of stable diffusion. In Proceedings of the European Conference on Computer Vision, 2024. 2, 5, 6" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 316, + 417, + 553, + 450 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 417, + 553, + 450 + ], + "spans": [ + { + "bbox": [ + 316, + 417, + 553, + 450 + ], + "type": "text", + "content": "[20] Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, 2015. 5" + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 317, + 453, + 553, + 507 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 453, + 553, + 507 + ], + "spans": [ + { + "bbox": [ + 317, + 453, + 553, + 507 + ], + "type": "text", + "content": "[21] Tuomas Kynkänniemi, Tero Karras, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Improved precision and recall metric for assessing generative models. In Proceedings of the Advances in Neural Information Processing Systems, pages 3927-3936, 2019. 8" + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 317, + 509, + 553, + 553 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 509, + 553, + 553 + ], + "spans": [ + { + "bbox": [ + 317, + 509, + 553, + 553 + ], + "type": "text", + "content": "[22] Hongjae Lee, Jun-Sang Yoo, and Seung-Won Jung. RefQSR: Reference-based quantization for image super-resolution networks. IEEE Transactions on Image Processing, 33: 2823-2834, 2024. 1, 3, 5" + } + ] + } + ], + "index": 24 + }, + { + "bbox": [ + 317, + 555, + 553, + 599 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 555, + 553, + 599 + ], + "spans": [ + { + "bbox": [ + 317, + 555, + 553, + 599 + ], + "type": "text", + "content": "[23] Lujun Li, Peijie Dong, Zimian Wei, and Ya Yang. Automated knowledge distillation via monte carlo tree search. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 17413-17424, 2023. 1" + } + ] + } + ], + "index": 25 + }, + { + "bbox": [ + 317, + 601, + 553, + 656 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 601, + 553, + 656 + ], + "spans": [ + { + "bbox": [ + 317, + 601, + 553, + 656 + ], + "type": "text", + "content": "[24] Xiuyu Li, Yijiang Liu, Long Lian, Huanrui Yang, Zhen Dong, Daniel Kang, Shanghang Zhang, and Kurt Keutzer. Q-diffusion: Quantizing diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 17535-17545, 2023. 1, 2, 3, 5, 6, 8" + } + ] + } + ], + "index": 26 + }, + { + "bbox": [ + 317, + 658, + 553, + 713 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 658, + 553, + 713 + ], + "spans": [ + { + "bbox": [ + 317, + 658, + 553, + 713 + ], + "type": "text", + "content": "[25] Yuhang Li, Ruihao Gong, Xu Tan, Yang Yang, Peng Hu, Qi Zhang, Fengwei Yu, Wei Wang, and Shi Gu. BRECQ: Pushing the limit of post-training quantization by block reconstruction. In Proceedings of the International Conference on Learning Representations, 2021. 1, 5" + } + ] + } + ], + "index": 27 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "text", + "content": "15434" + } + ] + } + ], + "index": 29 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 8 + }, + { + "para_blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 713 + ], + "type": "list", + "angle": 0, + "index": 13, + "blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 137 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 72, + 294, + 137 + ], + "spans": [ + { + "bbox": [ + 56, + 72, + 294, + 137 + ], + "type": "text", + "content": "[26] Yanyu Li, Huan Wang, Qing Jin, Ju Hu, Pavlo Chemerys, Yun Fu, Yanzhi Wang, Sergey Tulyakov, and Jian Ren. SnapFusion: Text-to-image diffusion model on mobile devices within two seconds. In Proceedings of the Advances in Neural Information Processing Systems, pages 20662-20678, 2024. 2" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 56, + 140, + 294, + 195 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 140, + 294, + 195 + ], + "spans": [ + { + "bbox": [ + 56, + 140, + 294, + 195 + ], + "type": "text", + "content": "[27] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dólar, and C Lawrence Zitnick. Microsoft COCO: Common objects in context. In Proceedings of the European Conference on Computer Vision, pages 740-755, 2014. 3, 5, 6" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 56, + 197, + 294, + 251 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 197, + 294, + 251 + ], + "spans": [ + { + "bbox": [ + 56, + 197, + 294, + 251 + ], + "type": "text", + "content": "[28] Chenxi Liu, Barret Zoph, Maxim Neumann, Jonathon Shlens, Wei Hua, Li-Jia Li, Li Fei-Fei, Alan Yuille, Jonathan Huang, and Kevin Murphy. Progressive neural architecture search. In Proceedings of the European Conference on Computer Vision, pages 19–34, 2018. 1" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 56, + 252, + 294, + 296 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 252, + 294, + 296 + ], + "spans": [ + { + "bbox": [ + 56, + 252, + 294, + 296 + ], + "type": "text", + "content": "[29] Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. Pseudo numerical methods for diffusion models on manifolds. In Proceedings of the International Conference on Learning Representations, 2022. 1" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 56, + 297, + 294, + 351 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 297, + 294, + 351 + ], + "spans": [ + { + "bbox": [ + 56, + 297, + 294, + 351 + ], + "type": "text", + "content": "[30] Zhenhua Liu, Yunhe Wang, Kai Han, Siwei Ma, and Wen Gao. Instance-aware dynamic neural network quantization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12434-12443, 2022. 1, 3" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 56, + 354, + 294, + 408 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 354, + 294, + 408 + ], + "spans": [ + { + "bbox": [ + 56, + 354, + 294, + 408 + ], + "type": "text", + "content": "[31] Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. DPM-Solver: A fast ODE solver for diffusion probabilistic model sampling in around 10 steps. In Proceedings of the Advances in Neural Information Processing Systems, pages 5775-5787, 2022. 1" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 56, + 410, + 294, + 464 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 410, + 294, + 464 + ], + "spans": [ + { + "bbox": [ + 56, + 410, + 294, + 464 + ], + "type": "text", + "content": "[32] Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik Kingma, Stefano Ermon, Jonathan Ho, and Tim Salimans. On distillation of guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14297-14306, 2023. 2" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 56, + 466, + 294, + 509 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 466, + 294, + 509 + ], + "spans": [ + { + "bbox": [ + 56, + 466, + 294, + 509 + ], + "type": "text", + "content": "[33] Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In Proceedings of the International Conference on Machine Learning, pages 8162-8171, 2021. 2" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 56, + 511, + 294, + 532 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 511, + 294, + 532 + ], + "spans": [ + { + "bbox": [ + 56, + 511, + 294, + 532 + ], + "type": "text", + "content": "[34] NVIDIA. CUTLASS, 2025. https://github.com/NVIDIA/cutlass.7" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 56, + 534, + 294, + 588 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 534, + 294, + 588 + ], + "spans": [ + { + "bbox": [ + 56, + 534, + 294, + 588 + ], + "type": "text", + "content": "[35] Nilesh Prasad Pandey, Marios Fournarakis, Chirag Patel, and Markus Nagel. Softmax bias correction for quantized generative models. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, pages 1453-1458, 2023. 2" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 56, + 590, + 294, + 622 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 590, + 294, + 622 + ], + "spans": [ + { + "bbox": [ + 56, + 590, + 294, + 622 + ], + "type": "text", + "content": "[36] Karl Pearson. Notes on regression and inheritance in the case of two parents. Proceedings of the Royal Society of London, 58:240-242, 1895. 8" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 56, + 624, + 294, + 689 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 624, + 294, + 689 + ], + "spans": [ + { + "bbox": [ + 56, + 624, + 294, + 689 + ], + "type": "text", + "content": "[37] Dustin Podell, *Zion English*, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, and Robin Rombach. *SDXL: Improving latent diffusion models for high-resolution image synthesis*. In *Proceedings of the International Conference on Learning Representations*, 2024. 1" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 56, + 691, + 294, + 713 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 691, + 294, + 713 + ], + "spans": [ + { + "bbox": [ + 56, + 691, + 294, + 713 + ], + "type": "text", + "content": "[38] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry," + } + ] + } + ], + "index": 12 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 316, + 73, + 555, + 713 + ], + "type": "list", + "angle": 0, + "index": 26, + "blocks": [ + { + "bbox": [ + 333, + 73, + 553, + 117 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 333, + 73, + 553, + 117 + ], + "spans": [ + { + "bbox": [ + 333, + 73, + 553, + 117 + ], + "type": "text", + "content": "Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, pages 8748-8763, 2021. 4" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 316, + 119, + 555, + 173 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 119, + 555, + 173 + ], + "spans": [ + { + "bbox": [ + 316, + 119, + 555, + 173 + ], + "type": "text", + "content": "[39] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10674-10685, 2022. 1, 2, 3, 5, 6" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 316, + 175, + 553, + 239 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 175, + 553, + 239 + ], + "spans": [ + { + "bbox": [ + 316, + 175, + 553, + 239 + ], + "type": "text", + "content": "[40] Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. DreamBooth: Fine tuning text-to-image diffusion models for subject-driven generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 22500-22510, 2023. 1" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 316, + 242, + 553, + 285 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 242, + 553, + 285 + ], + "spans": [ + { + "bbox": [ + 316, + 242, + 553, + 285 + ], + "type": "text", + "content": "[41] Hyogon Ryu, Seohyun Lim, and Hyunjung Shim. Memory-efficient personalization using quantized diffusion model. In Proceedings of the European Conference on Computer Vision, 2024. 2" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 316, + 287, + 553, + 363 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 287, + 553, + 363 + ], + "spans": [ + { + "bbox": [ + 316, + 287, + 553, + 363 + ], + "type": "text", + "content": "[42] Chitwan Sahara, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. In Proceedings of the Advances in Neural Information Processing Systems, pages 36479–36494, 2022. 1" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 316, + 365, + 553, + 407 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 365, + 553, + 407 + ], + "spans": [ + { + "bbox": [ + 316, + 365, + 553, + 407 + ], + "type": "text", + "content": "[43] Tim Salimans and Jonathan Ho. Progressive distillation for fast sampling of diffusion models. In Proceedings of the International Conference on Learning Representations, 2022. 2" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 316, + 411, + 553, + 464 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 411, + 553, + 464 + ], + "spans": [ + { + "bbox": [ + 316, + 411, + 553, + 464 + ], + "type": "text", + "content": "[44] Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, Xi Chen, and Xi Chen. Improved techniques for training gans. In Proceedings of the Advances in Neural Information Processing Systems, pages 2234-2242, 2016. 5" + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 316, + 466, + 553, + 520 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 466, + 553, + 520 + ], + "spans": [ + { + "bbox": [ + 316, + 466, + 553, + 520 + ], + "type": "text", + "content": "[45] Yuzhang Shang, Zhihang Yuan, Bin Xie, Bingzhe Wu, and Yan Yan. Post-training quantization on diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1972-1981, 2023. 1, 2" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 316, + 522, + 553, + 578 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 522, + 553, + 578 + ], + "spans": [ + { + "bbox": [ + 316, + 522, + 553, + 578 + ], + "type": "text", + "content": "[46] Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, pages 2556-2565, 2018. 5, 6" + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 316, + 579, + 553, + 633 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 579, + 553, + 633 + ], + "spans": [ + { + "bbox": [ + 316, + 579, + 553, + 633 + ], + "type": "text", + "content": "[47] Junhyuk So, Jungwon Lee, Daehyun Ahn, Hyungjun Kim, and Eunhyeok Park. Temporal dynamic quantization for diffusion models. In Proceedings of the Advances in Neural Information Processing Systems, pages 48686-48698, 2023. 1, 2, 3, 5" + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 316, + 635, + 553, + 677 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 635, + 553, + 677 + ], + "spans": [ + { + "bbox": [ + 316, + 635, + 553, + 677 + ], + "type": "text", + "content": "[48] Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In Proceedings of the International Conference on Learning Representations, 2021. 1, 2" + } + ] + } + ], + "index": 24 + }, + { + "bbox": [ + 316, + 680, + 553, + 713 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 680, + 553, + 713 + ], + "spans": [ + { + "bbox": [ + 316, + 680, + 553, + 713 + ], + "type": "text", + "content": "[49] Charles Spearman. The proof and measurement of association between two things. The American Journal of Psychology, 15(1):72--101, 1961. 8" + } + ] + } + ], + "index": 25 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "15435" + } + ] + } + ], + "index": 27 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 9 + }, + { + "para_blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 552 + ], + "type": "list", + "angle": 0, + "index": 9, + "blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 72, + 294, + 127 + ], + "spans": [ + { + "bbox": [ + 56, + 72, + 294, + 127 + ], + "type": "text", + "content": "[50] Yang Sui, Miao Yin, Yi Xie, Huy Phan, Saman Aliari Zonouz, and Bo Yuan. CHIP: Channel independence-based pruning for compact neural networks. In Proceedings of the Advances in Neural Information Processing Systems, pages 24604-24616, 2021. 1" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 56, + 129, + 294, + 184 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 129, + 294, + 184 + ], + "spans": [ + { + "bbox": [ + 56, + 129, + 294, + 184 + ], + "type": "text", + "content": "[51] Siao Tang, Xin Wang, Hong Chen, Chaoyu Guan, Zewen Wu, Yansong Tang, and Wenwu Zhu. Post-training quantization with progressive calibration and activation relaxing for text-to-image diffusion models. In Proceedings of the European Conference on Computer Vision, 2024. 2, 3" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 56, + 185, + 294, + 250 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 185, + 294, + 250 + ], + "spans": [ + { + "bbox": [ + 56, + 185, + 294, + 250 + ], + "type": "text", + "content": "[52] Senmao Tian, Ming Lu, Jiaming Liu, Yandong Guo, Yurong Chen, and Shunli Zhang. CABM: Content-aware bit mapping for single image super-resolution network with large input. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1756-1765, 2023. 1, 3" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 56, + 251, + 294, + 306 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 251, + 294, + 306 + ], + "spans": [ + { + "bbox": [ + 56, + 251, + 294, + 306 + ], + "type": "text", + "content": "[53] Mart Van Baalen, Christos Louizos, Markus Nagel, Rana Ali Amjad, Ying Wang, Tijmen Blankevoort, and Max Welling. Bayesian bits: Unifying quantization and pruning. In Proceedings of the Advances in Neural Information Processing Systems, pages 5741-5752, 2020. 5" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 56, + 308, + 294, + 351 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 308, + 294, + 351 + ], + "spans": [ + { + "bbox": [ + 56, + 308, + 294, + 351 + ], + "type": "text", + "content": "[54] Jianyi Wang, Kelvin CK Chan, and Chen Change Loy. Exploring clip for assessing the look and feel of images. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 2555-2563, 2023. 8" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 56, + 353, + 294, + 407 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 353, + 294, + 407 + ], + "spans": [ + { + "bbox": [ + 56, + 353, + 294, + 407 + ], + "type": "text", + "content": "[55] Kuan Wang, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. HAQ: Hardware-aware automated quantization with mixed precision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8612-8620, 2019. 1, 3" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 56, + 408, + 294, + 462 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 408, + 294, + 462 + ], + "spans": [ + { + "bbox": [ + 56, + 408, + 294, + 462 + ], + "type": "text", + "content": "[56] Chuanguang Yang, Zhulin An, Helong Zhou, Fuzhen Zhuang, Yongjun Xu, and Qian Zhang. Online knowledge distillation via mutual contrastive learning for visual recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(8):10212-10227, 2023. 1" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 56, + 464, + 294, + 508 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 464, + 294, + 508 + ], + "spans": [ + { + "bbox": [ + 56, + 464, + 294, + 508 + ], + "type": "text", + "content": "[57] Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3836-3847, 2023. 1" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 56, + 509, + 294, + 552 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 509, + 294, + 552 + ], + "spans": [ + { + "bbox": [ + 56, + 509, + 294, + 552 + ], + "type": "text", + "content": "[58] Xuechen Zhang, Zheng Li, Samet Oymak, and Jiasi Chen. Text-to-3d generative ai on mobile devices: Measurements and optimizations. In Proceedings of the Workshop on Emerging Multimedia Systems, pages 8–14, 2023. 1" + } + ] + } + ], + "index": 8 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 749, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 749, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 749, + 317, + 757 + ], + "type": "text", + "content": "15436" + } + ] + } + ], + "index": 10 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 10 + } + ], + "_backend": "vlm", + "_version_name": "2.6.4" +} \ No newline at end of file