File size: 8,421 Bytes
ce0c4f3
63d14c6
f26ebbe
63d14c6
f26ebbe
63d14c6
 
 
b5887d5
 
 
 
 
70dd883
fe44b10
b5887d5
ce0c4f3
f26ebbe
 
 
 
 
 
f187eb1
fe44b10
a24abf2
fe44b10
 
 
 
b5887d5
f187eb1
b5887d5
fe44b10
 
 
 
 
 
b5887d5
 
f26ebbe
 
 
 
 
 
 
 
ce0c4f3
b5887d5
f26ebbe
63d14c6
 
f26ebbe
63d14c6
b5887d5
63d14c6
b5887d5
f26ebbe
a24abf2
b5887d5
a24abf2
63d14c6
 
 
f26ebbe
63d14c6
 
 
b5887d5
63d14c6
b5887d5
f26ebbe
b5887d5
f26ebbe
b5887d5
 
f26ebbe
 
b5887d5
 
 
63d14c6
f26ebbe
 
 
 
63d14c6
 
 
 
b5887d5
63d14c6
b5887d5
f26ebbe
 
 
63d14c6
 
f26ebbe
63d14c6
70dd883
f26ebbe
 
 
fe44b10
70dd883
f26ebbe
70dd883
f26ebbe
70dd883
 
f26ebbe
 
70dd883
f26ebbe
b5887d5
 
 
 
 
f26ebbe
b5887d5
70dd883
63d14c6
 
b5887d5
f26ebbe
 
70dd883
63d14c6
f26ebbe
63d14c6
f26ebbe
 
 
63d14c6
b5887d5
63d14c6
b5887d5
f26ebbe
a24abf2
b5887d5
a24abf2
f26ebbe
70dd883
b5887d5
63d14c6
 
f26ebbe
 
70dd883
 
f26ebbe
 
b5887d5
 
f26ebbe
 
 
b5887d5
 
f26ebbe
b5887d5
 
f26ebbe
 
 
 
b5887d5
 
f26ebbe
 
 
 
b5887d5
 
 
f26ebbe
b5887d5
f26ebbe
47efb70
 
f26ebbe
 
b5887d5
f26ebbe
63d14c6
 
f26ebbe
 
 
 
47efb70
63d14c6
f26ebbe
 
 
b5887d5
f26ebbe
 
 
 
 
 
63d14c6
 
70dd883
f26ebbe
63d14c6
 
b5887d5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
import gradio as gr
import torch
from transformers import TorchAoConfig, AutoModel, AutoTokenizer
import tempfile
from huggingface_hub import HfApi, snapshot_download, list_models
from gradio_huggingfacehub_search import HuggingfaceHubSearch
from packaging import version
import os
from torchao.quantization import (
    Int4WeightOnlyConfig,
    Int8WeightOnlyConfig,
    Int8DynamicActivationInt8WeightConfig,
    Float8WeightOnlyConfig,
    Float8DynamicActivationFloat8WeightConfig,
    GemliteUIntXWeightOnlyConfig,
)

# === Load Hugging Face token from environment ===
HF_TOKEN = os.getenv("HF_TOKEN")
if not HF_TOKEN:
    raise ValueError("❌ Missing HF_TOKEN environment variable. Please set it before running the app.")

# === Quantization configuration maps ===
MAP_QUANT_TYPE_TO_NAME = {
    "Int4WeightOnly": "int4wo",
    "GemliteUIntXWeightOnly": "intxwo-gemlite",
    "Int8WeightOnly": "int8wo",
    "Int8DynamicActivationInt8Weight": "int8da8w8",
    "Float8WeightOnly": "float8wo",
    "Float8DynamicActivationFloat8Weight": "float8da8w8",
    "autoquant": "autoquant",
}
MAP_QUANT_TYPE_TO_CONFIG = {
    "Int4WeightOnly": Int4WeightOnlyConfig,
    "GemliteUIntXWeightOnly": GemliteUIntXWeightOnlyConfig,
    "Int8WeightOnly": Int8WeightOnlyConfig,
    "Int8DynamicActivationInt8Weight": Int8DynamicActivationInt8WeightConfig,
    "Float8WeightOnly": Float8WeightOnlyConfig,
    "Float8DynamicActivationFloat8Weight": Float8DynamicActivationFloat8WeightConfig,
}

# === Helper functions ===
def get_username():
    try:
        api = HfApi(token=HF_TOKEN)
        info = api.whoami()
        return info["name"]
    except Exception:
        return "anonymous"


def check_model_exists(username, quantization_type, group_size, model_name, quantized_model_name):
    """Check if a model exists in the user's Hugging Face repository."""
    try:
        models = list_models(author=username, token=HF_TOKEN)
        model_names = [model.id for model in models]
        if quantized_model_name:
            repo_name = f"{username}/{quantized_model_name}"
        else:
            if quantization_type in ["Int4WeightOnly", "GemliteUIntXWeightOnly"] and group_size is not None:
                repo_name = f"{username}/{model_name.split('/')[-1]}-ao-{MAP_QUANT_TYPE_TO_NAME[quantization_type]}-gs{group_size}"
            else:
                repo_name = f"{username}/{model_name.split('/')[-1]}-ao-{MAP_QUANT_TYPE_TO_NAME[quantization_type]}"
        if repo_name in model_names:
            return f"Model '{repo_name}' already exists in your repository."
        else:
            return None
    except Exception as e:
        return f"Error checking model existence: {str(e)}"


def create_model_card(model_name, quantization_type, group_size):
    try:
        model_path = snapshot_download(repo_id=model_name, allow_patterns=["README.md"], repo_type="model", token=HF_TOKEN)
        readme_path = os.path.join(model_path, "README.md")
        original_readme = ""
        if os.path.exists(readme_path):
            with open(readme_path, "r", encoding="utf-8") as f:
                original_readme = f.read()
    except Exception:
        original_readme = ""

    yaml_header = f"""---
base_model:
- {model_name}
tags:
- torchao-my-repo
---
# {model_name} (Quantized)

## Quantization Details
- **Quantization Type**: {quantization_type}
- **Group Size**: {group_size}

"""
    if original_readme:
        yaml_header += "\n\n# πŸ“„ Original Model Info\n\n" + original_readme
    return yaml_header


def quantize_model(model_name, quantization_type, group_size=128, progress=gr.Progress()):
    print(f"Quantizing model: {quantization_type}")
    progress(0, desc="Preparing Quantization")

    if quantization_type == "GemliteUIntXWeightOnly":
        quant_config = MAP_QUANT_TYPE_TO_CONFIG[quantization_type](group_size=group_size)
    elif quantization_type == "Int4WeightOnly":
        from torchao.dtypes import Int4CPULayout
        quant_config = MAP_QUANT_TYPE_TO_CONFIG[quantization_type](group_size=group_size, layout=Int4CPULayout())
    elif quantization_type == "autoquant":
        quant_config = "autoquant"
    else:
        quant_config = MAP_QUANT_TYPE_TO_CONFIG[quantization_type]()

    quantization_config = TorchAoConfig(quant_config)
    progress(0.10, desc="Quantizing model")

    model = AutoModel.from_pretrained(
        model_name,
        torch_dtype="auto",
        quantization_config=quantization_config,
        device_map="cpu",
        token=HF_TOKEN,
    )
    progress(0.45, desc="Quantization completed")
    return model


def save_model(model, model_name, quantization_type, group_size=128, quantized_model_name=None, public=True, progress=gr.Progress()):
    username = get_username()
    progress(0.50, desc="Preparing to push")
    print("Saving quantized model")

    with tempfile.TemporaryDirectory() as tmpdirname:
        tokenizer = AutoTokenizer.from_pretrained(model_name, token=HF_TOKEN)
        tokenizer.save_pretrained(tmpdirname)
        model.save_pretrained(tmpdirname, safe_serialization=False)

        if quantized_model_name:
            repo_name = f"{username}/{quantized_model_name}"
        else:
            if quantization_type in ["Int4WeightOnly", "GemliteUIntXWeightOnly"] and (group_size is not None):
                repo_name = f"{username}/{model_name.split('/')[-1]}-ao-{MAP_QUANT_TYPE_TO_NAME[quantization_type]}-gs{group_size}"
            else:
                repo_name = f"{username}/{model_name.split('/')[-1]}-ao-{MAP_QUANT_TYPE_TO_NAME[quantization_type]}"

        progress(0.70, desc="Creating model card")
        model_card = create_model_card(model_name, quantization_type, group_size)
        with open(os.path.join(tmpdirname, "README.md"), "w") as f:
            f.write(model_card)

        api = HfApi(token=HF_TOKEN)
        api.create_repo(repo_name, exist_ok=True, private=not public)
        progress(0.80, desc="Pushing to Hub")
        api.upload_folder(folder_path=tmpdirname, repo_id=repo_name, repo_type="model")
        progress(1.00, desc="Done")

    repo_link = f"""
    <div class="repo-link">
        <h3>πŸ”— Repository Link</h3>
        <p>Find your repo here: <a href="https://huggingface.co/{repo_name}" target="_blank">{repo_name}</a></p>
    </div>
    """
    return f"<h1>πŸŽ‰ Quantization Completed</h1><br/>{repo_link}"


def quantize_and_save(model_name, quantization_type, group_size, quantized_model_name, public):
    username = get_username()
    if not username or username == "anonymous":
        return "<div class='error-box'><h3>❌ Authentication Error</h3><p>Invalid or missing HF_TOKEN.</p></div>"

    if group_size and group_size.strip():
        try:
            group_size = int(group_size)
        except ValueError:
            group_size = None
    else:
        group_size = None

    exists_message = check_model_exists(username, quantization_type, group_size, model_name, quantized_model_name)
    if exists_message:
        return f"<div class='warning-box'><h3>⚠️ Model Already Exists</h3><p>{exists_message}</p></div>"

    try:
        quantized_model = quantize_model(model_name, quantization_type, group_size)
        return save_model(quantized_model, model_name, quantization_type, group_size, quantized_model_name, public)
    except Exception as e:
        return f"<div class='error-box'><h3>❌ Error</h3><p>{str(e)}</p></div>"


# === Gradio UI ===
with gr.Blocks() as demo:
    gr.Markdown("# πŸ€— TorchAO Quantizer (Token Mode) πŸ”₯")
    gr.Markdown("Uses your environment HF_TOKEN β€” no login required.")

    with gr.Row():
        model_name = HuggingfaceHubSearch(label="πŸ” Hub Model ID", placeholder="Search a model", search_type="model")
        quantization_type = gr.Dropdown(
            choices=list(MAP_QUANT_TYPE_TO_NAME.keys()), value="Int8WeightOnly", label="Quantization Type"
        )
        group_size = gr.Textbox(label="Group Size (optional)", value="128")
        quantized_model_name = gr.Textbox(label="Custom Model Name", value="")
        public = gr.Checkbox(label="Make Public", value=True)
        output_link = gr.Markdown()
        quantize_button = gr.Button("πŸš€ Quantize and Push")

    quantize_button.click(
        fn=quantize_and_save,
        inputs=[model_name, quantization_type, group_size, quantized_model_name, public],
        outputs=output_link,
    )

demo.launch(share=True)