File size: 5,253 Bytes
9ff9d53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
from typing import Any
from pathlib import Path
import pickle

from tqdm.notebook import tqdm
from transformers import set_seed
import numpy as np
import openvino as ov
from PIL import Image
import torch
import nncf

from ov_catvton_helper import (
    MODEL_DIR,
    VAE_ENCODER_PATH,
    VAE_DECODER_PATH,
    UNET_PATH,
    DENSEPOSE_PROCESSOR_PATH,
    SCHP_PROCESSOR_ATR,
    SCHP_PROCESSOR_LIP,
)

set_seed(42)
NUM_INFERENCE_STEPS = 50
GUIDANCE_SCALE = 2.5
GENERATOR = torch.Generator(device="cpu").manual_seed(42)

VAE_ENCODER_INT4_PATH = MODEL_DIR / "vae_encoder_int4.xml"
VAE_DECODER_INT4_PATH = MODEL_DIR / "vae_decoder_int4.xml"
UNET_INT8_PATH = MODEL_DIR / "unet_int8.xml"
DENSEPOSE_PROCESSOR_INT4_PATH = MODEL_DIR / "densepose_processor_int4.xml"
SCHP_PROCESSOR_ATR_INT4 = MODEL_DIR / "schp_processor_atr_int4.xml"
SCHP_PROCESSOR_LIP_INT4 = MODEL_DIR / "schp_processor_lip_int4.xml"


class CompiledModelDecorator(ov.CompiledModel):
    def __init__(
        self,
        compiled_model: ov.CompiledModel,
        data_cache: list[Any] = None,
        keep_prob: float = 1.0,
    ):
        super().__init__(compiled_model)
        self.data_cache = data_cache if data_cache is not None else []
        self.keep_prob = keep_prob

    def __call__(self, *args, **kwargs):
        if np.random.rand() <= self.keep_prob:
            self.data_cache.append(*args)
        return super().__call__(*args, **kwargs)


def collect_calibration_data(pipeline, automasker, mask_processor, dataset, subset_size):
    calibration_dataset_filepath = Path("calibration_data") / f"{subset_size}.pkl"
    calibration_dataset_filepath.parent.mkdir(exist_ok=True, parents=True)

    if not calibration_dataset_filepath.exists():
        original_unet = pipeline.unet.unet
        pipeline.unet.unet = CompiledModelDecorator(original_unet)

        calibration_dataset = []
        pbar = tqdm(total=subset_size, desc="Collecting calibration dataset")
        for data in dataset:
            person_image_path, cloth_image_path = data
            person_image = Image.open(person_image_path)
            cloth_image = Image.open(cloth_image_path)
            cloth_type = "upper" if "upper" in person_image_path.as_posix() else "overall"
            mask = automasker(person_image, cloth_type)["mask"]
            mask = mask_processor.blur(mask, blur_factor=9)

            pipeline(
                image=person_image,
                condition_image=cloth_image,
                mask=mask,
                num_inference_steps=NUM_INFERENCE_STEPS,
                guidance_scale=GUIDANCE_SCALE,
                generator=GENERATOR,
            )
            collected_subset_size = len(pipeline.unet.unet.data_cache)
            pbar.update(NUM_INFERENCE_STEPS)
            if collected_subset_size >= subset_size:
                break

        calibration_dataset = pipeline.unet.unet.data_cache
        pipeline.unet.unet = original_unet

        with open(calibration_dataset_filepath, "wb") as f:
            pickle.dump(calibration_dataset, f)
    else:
        with open(calibration_dataset_filepath, "rb") as f:
            calibration_dataset = pickle.load(f)

    return calibration_dataset


def compress_model(core, model_path, save_path, group_size=128, ratio=0.8):
    if not save_path.exists():
        print(f"{model_path.stem} compression started")
        print(f"Compression parameters:\n\tmode = {nncf.CompressWeightsMode.INT4_SYM}\n\tratio = {ratio}\n\tgroup_size = {group_size}")
        model = core.read_model(model_path)
        compressed_model = nncf.compress_weights(
            model,
            mode=nncf.CompressWeightsMode.INT4_SYM,
            ratio=ratio,
            group_size=group_size,
        )
        ov.save_model(compressed_model, save_path)
        print(f"{model_path.stem} compression finished")
    print(f"Compressed {model_path.stem} can be found in {save_path}")


def compress_models(core, group_size=128, ratio=0.8):
    compress_model(core, VAE_ENCODER_PATH, VAE_ENCODER_INT4_PATH, group_size, ratio)
    compress_model(core, VAE_DECODER_PATH, VAE_DECODER_INT4_PATH, group_size, ratio)
    compress_model(core, DENSEPOSE_PROCESSOR_PATH, DENSEPOSE_PROCESSOR_INT4_PATH, group_size, ratio)
    compress_model(core, SCHP_PROCESSOR_ATR, SCHP_PROCESSOR_ATR_INT4, group_size, ratio)
    compress_model(core, SCHP_PROCESSOR_LIP, SCHP_PROCESSOR_LIP_INT4, group_size, ratio)


def compare_models_size():
    fp16_model_paths = [
        VAE_ENCODER_PATH,
        VAE_DECODER_PATH,
        UNET_PATH,
        DENSEPOSE_PROCESSOR_PATH,
        SCHP_PROCESSOR_ATR,
        SCHP_PROCESSOR_LIP,
    ]
    optimized_models = [
        VAE_ENCODER_INT4_PATH,
        VAE_DECODER_INT4_PATH,
        UNET_INT8_PATH,
        DENSEPOSE_PROCESSOR_INT4_PATH,
        SCHP_PROCESSOR_ATR_INT4,
        SCHP_PROCESSOR_LIP_INT4,
    ]

    for fp16_path, optimized_path in zip(fp16_model_paths, optimized_models):
        if not fp16_path.exists():
            continue
        fp16_ir_model_size = fp16_path.with_suffix(".bin").stat().st_size
        optimized_model_size = optimized_path.with_suffix(".bin").stat().st_size
        print(f"{fp16_path.stem} compression rate: {fp16_ir_model_size / optimized_model_size:.3f}")