| import ipdb | |
| from peft.tuners.tuners_utils import BaseTunerLayer | |
| from typing import List, Any, Optional, Type | |
| def module_active_adapters(module): | |
| if hasattr(module, 'active_adapters'): | |
| result = [i for i in module.active_adapters if i in module.scaling.keys()] | |
| else: | |
| result = [] | |
| return result | |
| class enable_lora: | |
| def __init__(self, lora_modules: List[BaseTunerLayer], enable_adapters: List) -> None: | |
| self.lora_modules: List[BaseTunerLayer] = [ | |
| each for each in lora_modules if isinstance(each, BaseTunerLayer) | |
| ] | |
| self.active_adapter_scales = [ | |
| { | |
| active_adapter: lora_module.scaling[active_adapter] | |
| for active_adapter in module_active_adapters(lora_module) | |
| } | |
| for lora_module in self.lora_modules | |
| ] | |
| self.enable_adapters = enable_adapters | |
| def __enter__(self) -> None: | |
| for lora_module in self.lora_modules: | |
| for active_adapter in module_active_adapters(lora_module): | |
| if active_adapter not in self.enable_adapters: | |
| lora_module.set_scale(active_adapter,0) | |
| def __exit__( | |
| self, | |
| exc_type: Optional[Type[BaseException]], | |
| exc_val: Optional[BaseException], | |
| exc_tb: Optional[Any], | |
| ) -> None: | |
| for i, lora_module in enumerate(self.lora_modules): | |
| for active_adapter in module_active_adapters(lora_module): | |
| lora_module.set_scale(active_adapter,self.active_adapter_scales[i][active_adapter]) | |