File size: 1,560 Bytes
1d7ce38 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
import ipdb
from peft.tuners.tuners_utils import BaseTunerLayer
from typing import List, Any, Optional, Type
def module_active_adapters(module):
if hasattr(module, 'active_adapters'):
result = [i for i in module.active_adapters if i in module.scaling.keys()]
else:
result = []
return result
class enable_lora:
def __init__(self, lora_modules: List[BaseTunerLayer], enable_adapters: List) -> None:
self.lora_modules: List[BaseTunerLayer] = [
each for each in lora_modules if isinstance(each, BaseTunerLayer)
]
self.active_adapter_scales = [
{
active_adapter: lora_module.scaling[active_adapter]
for active_adapter in module_active_adapters(lora_module)
}
for lora_module in self.lora_modules
]
self.enable_adapters = enable_adapters
def __enter__(self) -> None:
for lora_module in self.lora_modules:
for active_adapter in module_active_adapters(lora_module):
if active_adapter not in self.enable_adapters:
lora_module.set_scale(active_adapter,0)
def __exit__(
self,
exc_type: Optional[Type[BaseException]],
exc_val: Optional[BaseException],
exc_tb: Optional[Any],
) -> None:
for i, lora_module in enumerate(self.lora_modules):
for active_adapter in module_active_adapters(lora_module):
lora_module.set_scale(active_adapter,self.active_adapter_scales[i][active_adapter])
|