|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import json |
|
|
from collections.abc import Generator |
|
|
from typing import TYPE_CHECKING, Union |
|
|
|
|
|
from ...extras.constants import PEFT_METHODS |
|
|
from ...extras.misc import torch_gc |
|
|
from ...extras.packages import is_gradio_available |
|
|
from ...train.tuner import export_model |
|
|
from ..common import get_save_dir, load_config |
|
|
from ..locales import ALERTS |
|
|
|
|
|
|
|
|
if is_gradio_available(): |
|
|
import gradio as gr |
|
|
|
|
|
|
|
|
if TYPE_CHECKING: |
|
|
from gradio.components import Component |
|
|
|
|
|
from ..engine import Engine |
|
|
|
|
|
|
|
|
GPTQ_BITS = ["8", "4", "3", "2"] |
|
|
|
|
|
|
|
|
def can_quantize(checkpoint_path: Union[str, list[str]]) -> "gr.Dropdown": |
|
|
if isinstance(checkpoint_path, list) and len(checkpoint_path) != 0: |
|
|
return gr.Dropdown(value="none", interactive=False) |
|
|
else: |
|
|
return gr.Dropdown(interactive=True) |
|
|
|
|
|
|
|
|
def save_model( |
|
|
lang: str, |
|
|
model_name: str, |
|
|
model_path: str, |
|
|
finetuning_type: str, |
|
|
checkpoint_path: Union[str, list[str]], |
|
|
template: str, |
|
|
export_size: int, |
|
|
export_quantization_bit: str, |
|
|
export_quantization_dataset: str, |
|
|
export_device: str, |
|
|
export_legacy_format: bool, |
|
|
export_dir: str, |
|
|
export_hub_model_id: str, |
|
|
extra_args: str, |
|
|
) -> Generator[str, None, None]: |
|
|
user_config = load_config() |
|
|
error = "" |
|
|
if not model_name: |
|
|
error = ALERTS["err_no_model"][lang] |
|
|
elif not model_path: |
|
|
error = ALERTS["err_no_path"][lang] |
|
|
elif not export_dir: |
|
|
error = ALERTS["err_no_export_dir"][lang] |
|
|
elif export_quantization_bit in GPTQ_BITS and not export_quantization_dataset: |
|
|
error = ALERTS["err_no_dataset"][lang] |
|
|
elif export_quantization_bit not in GPTQ_BITS and not checkpoint_path: |
|
|
error = ALERTS["err_no_adapter"][lang] |
|
|
elif export_quantization_bit in GPTQ_BITS and checkpoint_path and isinstance(checkpoint_path, list): |
|
|
error = ALERTS["err_gptq_lora"][lang] |
|
|
|
|
|
try: |
|
|
json.loads(extra_args) |
|
|
except json.JSONDecodeError: |
|
|
error = ALERTS["err_json_schema"][lang] |
|
|
|
|
|
if error: |
|
|
gr.Warning(error) |
|
|
yield error |
|
|
return |
|
|
|
|
|
args = dict( |
|
|
model_name_or_path=model_path, |
|
|
cache_dir=user_config.get("cache_dir", None), |
|
|
finetuning_type=finetuning_type, |
|
|
template=template, |
|
|
export_dir=export_dir, |
|
|
export_hub_model_id=export_hub_model_id or None, |
|
|
export_size=export_size, |
|
|
export_quantization_bit=int(export_quantization_bit) if export_quantization_bit in GPTQ_BITS else None, |
|
|
export_quantization_dataset=export_quantization_dataset, |
|
|
export_device=export_device, |
|
|
export_legacy_format=export_legacy_format, |
|
|
trust_remote_code=True, |
|
|
) |
|
|
args.update(json.loads(extra_args)) |
|
|
|
|
|
if checkpoint_path: |
|
|
if finetuning_type in PEFT_METHODS: |
|
|
args["adapter_name_or_path"] = ",".join( |
|
|
[get_save_dir(model_name, finetuning_type, adapter) for adapter in checkpoint_path] |
|
|
) |
|
|
else: |
|
|
args["model_name_or_path"] = get_save_dir(model_name, finetuning_type, checkpoint_path) |
|
|
|
|
|
yield ALERTS["info_exporting"][lang] |
|
|
export_model(args) |
|
|
torch_gc() |
|
|
yield ALERTS["info_exported"][lang] |
|
|
|
|
|
|
|
|
def create_export_tab(engine: "Engine") -> dict[str, "Component"]: |
|
|
with gr.Row(): |
|
|
export_size = gr.Slider(minimum=1, maximum=100, value=5, step=1) |
|
|
export_quantization_bit = gr.Dropdown(choices=["none"] + GPTQ_BITS, value="none") |
|
|
export_quantization_dataset = gr.Textbox(value="data/c4_demo.jsonl") |
|
|
export_device = gr.Radio(choices=["cpu", "auto"], value="cpu") |
|
|
export_legacy_format = gr.Checkbox() |
|
|
|
|
|
with gr.Row(): |
|
|
export_dir = gr.Textbox() |
|
|
export_hub_model_id = gr.Textbox() |
|
|
extra_args = gr.Textbox(value="{}") |
|
|
|
|
|
checkpoint_path: gr.Dropdown = engine.manager.get_elem_by_id("top.checkpoint_path") |
|
|
checkpoint_path.change(can_quantize, [checkpoint_path], [export_quantization_bit], queue=False) |
|
|
|
|
|
export_btn = gr.Button() |
|
|
info_box = gr.Textbox(show_label=False, interactive=False) |
|
|
|
|
|
export_btn.click( |
|
|
save_model, |
|
|
[ |
|
|
engine.manager.get_elem_by_id("top.lang"), |
|
|
engine.manager.get_elem_by_id("top.model_name"), |
|
|
engine.manager.get_elem_by_id("top.model_path"), |
|
|
engine.manager.get_elem_by_id("top.finetuning_type"), |
|
|
engine.manager.get_elem_by_id("top.checkpoint_path"), |
|
|
engine.manager.get_elem_by_id("top.template"), |
|
|
export_size, |
|
|
export_quantization_bit, |
|
|
export_quantization_dataset, |
|
|
export_device, |
|
|
export_legacy_format, |
|
|
export_dir, |
|
|
export_hub_model_id, |
|
|
extra_args, |
|
|
], |
|
|
[info_box], |
|
|
) |
|
|
|
|
|
return dict( |
|
|
export_size=export_size, |
|
|
export_quantization_bit=export_quantization_bit, |
|
|
export_quantization_dataset=export_quantization_dataset, |
|
|
export_device=export_device, |
|
|
export_legacy_format=export_legacy_format, |
|
|
export_dir=export_dir, |
|
|
export_hub_model_id=export_hub_model_id, |
|
|
extra_args=extra_args, |
|
|
export_btn=export_btn, |
|
|
info_box=info_box, |
|
|
) |
|
|
|