| | import argparse |
| | import functools |
| | import os |
| |
|
| | from transformers import WhisperForConditionalGeneration, WhisperFeatureExtractor, WhisperTokenizerFast,\ |
| | WhisperProcessor |
| | from peft import PeftModel, PeftConfig |
| | from utils.utils import print_arguments, add_arguments |
| |
|
| | parser = argparse.ArgumentParser(description=__doc__) |
| | add_arg = functools.partial(add_arguments, argparser=parser) |
| | add_arg("lora_model", type=str, default="output/whisper-tiny/checkpoint-best/", help="model directory") |
| | add_arg('output_dir', type=str, default='models/', help="output directory") |
| | add_arg("local_files_only", type=bool, default=False, help="Choose local file , if not , find the model on HF") |
| | args = parser.parse_args() |
| | print_arguments(args) |
| |
|
| | assert os.path.exists(args.lora_model), f"model file{args.lora_model}does not exist" |
| |
|
| | peft_config = PeftConfig.from_pretrained(args.lora_model) |
| | |
| | base_model = WhisperForConditionalGeneration.from_pretrained(peft_config.base_model_name_or_path, device_map={"": "cuda"}, |
| | local_files_only=args.local_files_only) |
| |
|
| | model = PeftModel.from_pretrained(base_model, args.lora_model, local_files_only=args.local_files_only) |
| | feature_extractor = WhisperFeatureExtractor.from_pretrained(peft_config.base_model_name_or_path, |
| | local_files_only=args.local_files_only) |
| | tokenizer = WhisperTokenizerFast.from_pretrained(peft_config.base_model_name_or_path, |
| | local_files_only=args.local_files_only) |
| | processor = WhisperProcessor.from_pretrained(peft_config.base_model_name_or_path, |
| | local_files_only=args.local_files_only) |
| |
|
| |
|
| | model = model.merge_and_unload() |
| | model.train(False) |
| |
|
| | save_directory = os.path.join(args.output_dir, f'{os.path.basename(peft_config.base_model_name_or_path)}-finetune') |
| | os.makedirs(save_directory, exist_ok=True) |
| |
|
| | model.save_pretrained(save_directory) |
| | feature_extractor.save_pretrained(save_directory) |
| | tokenizer.save_pretrained(save_directory) |
| | processor.save_pretrained(save_directory) |
| | print(f'Merged model save at :{save_directory}') |