""" The main body of the ASR model, User: Model: """ import torch import torch.nn as nn from transformers import LlamaForCausalLM, LlamaTokenizer from typing import List try: from .speech_encoder import SpeechEncoder except ImportError: from speech_encoder import SpeechEncoder from transformers import AutoModelForCausalLM, AutoTokenizer from .model import RWKV # from .lora import LinearWithLoRA import pytorch_lightning as pl from torch.nn import functional as F from pytorch_lightning.strategies import DeepSpeedStrategy import os, math, gc, importlib if importlib.util.find_spec('deepspeed'): import deepspeed from deepspeed.ops.adam import DeepSpeedCPUAdam, FusedAdam import time class L2Wrap(torch.autograd.Function): @staticmethod def forward(ctx, loss, y): ctx.save_for_backward(y) return loss @staticmethod def backward(ctx, grad_output): y = ctx.saved_tensors[0] # to encourage the logits to be close to 0 factor = 1e-4 / (y.shape[0] * y.shape[1]) maxx, ids = torch.max(y, -1, keepdim=True) gy = torch.zeros_like(y) gy.scatter_(-1, ids, maxx * factor) return (grad_output, gy) class SLAM_ASR(pl.LightningModule): def __init__( self, args, speech_encoder_model_id,#facebook/hubert-base-ls960 language_model, downsample_K=5, hidden_dim=2048, train_mode="adapter", device="cuda", token = "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", ): assert train_mode in ["adapter", "full"] super().__init__() self.args = args self._device = device self.language_tokenizer = AutoTokenizer.from_pretrained("RWKV/rwkv-6-world-1b6",trust_remote_code=True) ########################################换成RWKV-PEFT的模型结构 self.language_model = language_model ######################################### language_project_dim = args.n_embd #3B language_project_dim = 2560 #7B language_project_dim = 4096 self.speech_encoder = SpeechEncoder( speech_encoder_model_id, language_project_dim, downsample_K=downsample_K, hidden_dim=hidden_dim, train_mode=train_mode, device=device, ).to(self._device) self.set_gradient(train_mode,'state') def gradient_checkpointing_enable(self, **kwargs): self.language_model.gradient_checkpointing_enable(**kwargs) def set_gradient(self, train_mode,tuning): assert train_mode in ["adapter", "full"] # call set_gradient for speech encoder self.speech_encoder.set_gradient(train_mode) print("Parameters that require grad:") for name, param in self.named_parameters(): if param.requires_grad: print(f" {name}: {param.shape}") def remove_padding(self, x, mask): #根据mask去除speech_output的padding部分 x_no_padding = [] # 对于每一个样本和对应的掩码 for x_i, mask_i in zip(x, mask): # 使用掩码来选择非填充部分 x_i_no_padding = x_i[mask_i.bool()] # 将结果添加到列表中 x_no_padding.append(x_i_no_padding) return x_no_padding def concatenate_audio_transcription(self, audio, transcription): #将两个二维/三维向量在第二维度拼起来 result = [] for sublist1, sublist2 in zip(audio, transcription): sub_result = torch.cat((sublist1 ,sublist2), dim=0) result.append(sub_result) return result def _prepare_input_embeds( self, audios: List[float], transcriptions: List[str] = None ): """ First, run audios through speech_encoder to get the embeddings and mask """ speech_output, mask = self.speech_encoder(audios) mask = mask.to(self._device) if transcriptions is not None: ###########处理prompt_embed ############################################################################### #去除speech padding audio_no_padding = self.remove_padding(speech_output,mask) #在speech结尾添加end of audio:# end_of_audio = self.language_tokenizer( "#", return_tensors="pt", ).to(self.device) with torch.no_grad(): end_of_audio = self.language_model.embed(end_of_audio.input_ids) audio_no_padding_eoa = [] for t in audio_no_padding: t = torch.cat((t, end_of_audio.squeeze(0))) audio_no_padding_eoa.append(t) #audio mask 左边添加1 ones = torch.ones(mask.size(0), 1).to(self._device) mask =torch.cat((ones, mask), dim=1) #处理transcription,得到embeded label _labels = self.language_tokenizer( transcriptions, return_tensors="pt", padding=True, truncation=True, add_special_tokens=False, ).to(self.device) with torch.no_grad(): # labels_embeds = self.language_model.rwkv.get_input_embeddings()(_labels.input_ids) labels_embeds = self.language_model.embed(_labels.input_ids) att3 = _labels.attention_mask #拼接speech和label audio_label = self.concatenate_audio_transcription(audio_no_padding_eoa , labels_embeds) # print(f"concatenated inputs:\t{len(audio_label)}-{[len(x) for x in audio_label]}") #对拼接后的内容进行padding max_seq = max([len(x) for x in audio_label]) for i, x in enumerate(audio_label): times = max_seq - len(x) for _ in range(times): x = torch.cat((x,x[len(x)-1].unsqueeze(0))) audio_label[i] = x # print(f"padded inputs:\t{len(audio_label)}-{[len(x) for x in audio_label]}") #转换成tensor audio_label = torch.stack(audio_label) # print(f"padded inputs tensor:\t{audio_label.shape}") prompt_embed = audio_label # print() #####处理prompt_mask ################################################## # 剔除audio mask 右边的0 mask_no_zero = [] for mask_i in mask: mask_i_no_zero = mask_i[mask_i != 0] mask_no_zero.append(mask_i_no_zero) # 将audio mask和transcription mask 拼接 mask_concatenate = self.concatenate_audio_transcription(mask_no_zero, att3) #向mask 填充0 max_mask = max([len(x) for x in mask_concatenate]) for i, x in enumerate(mask_concatenate): times = max_mask - len(x) for _ in range(times): x = torch.cat((x,torch.tensor([0]).to(self.device))) mask_concatenate[i] = x #转换成tensor mask_concatenate = torch.stack(mask_concatenate) prompt_mask = mask_concatenate # #########处理loss mask ##################################################### # import torch.nn.functional as F # loss_mask = [] # for t in mask_no_zero: # pad_len = max_mask - len(t) # pad = F.pad(t, (0, pad_len), "constant", 0) # loss_mask.append(pad) # loss_mask = torch.stack(loss_mask) # loss_mask = prompt_mask - loss_mask # print(f"loss mask:\t{loss_mask.shape}") #########处理true_labels ################################################### # print() # 为transcription 结尾添加 end of sentence: transcriptions_eos = [] for starr in transcriptions: starr = starr + "" transcriptions_eos.append(starr) _labels = self.language_tokenizer( transcriptions_eos, return_tensors="pt", padding=True, truncation=True, add_special_tokens=False, ).to(self.device) true_labels = _labels.input_ids #在ture label左侧填充audio 长度的-100, 同时在右侧填充-100使batch对齐 padded_labels = [] for i,t in enumerate(true_labels): back_padding = max_mask - t.shape[0] - audio_no_padding[i].shape[0] t = torch.cat( [ torch.full( (audio_no_padding[i].shape[0], ), -100, dtype=torch.long, device=self.device, ), t, torch.full( (back_padding, ), -100, dtype=torch.long, device=self.device, ), ] ) padded_labels.append(t) padded_labels = torch.stack(padded_labels) true_labels = padded_labels else: end_of_audio = self.language_tokenizer( "#", return_tensors="pt", ).to(self.device) with torch.no_grad(): end_of_audio = self.language_model.embed(end_of_audio.input_ids) # print(f"speech output:{speech_output.shape}") # print(f"end_of_audio:{end_of_audio.shape}") # exit(0) speech_output = torch.cat((speech_output, end_of_audio), dim= 1) prompt_embed = speech_output prompt_mask = mask true_labels = None return prompt_embed, prompt_mask, true_labels def forward(self, audios: List[float], transcriptions: List[str] = None): prompt_embed, prompt_mask, true_labels = self._prepare_input_embeds( audios, transcriptions ) outputs = self.language_model(inputs_embeds=prompt_embed) return outputs, true_labels, prompt_mask def generate(self, audios: List[float], stopping_criteria=None): """ Generate the transcription """ prompt_embed, prompt_mask, _ = self._prepare_input_embeds(audios) # outputs = self.language_model( # inputs_embeds=prompt_embed, # attention_mask=prompt_mask.bool() # ) self.language_model.to(self._device, dtype=torch.bfloat16) outputs = self.language_model.generate(tokenizer= self.language_tokenizer,inputs_embeds=prompt_embed) return outputs def training_step(self, batch, batch_idx): args = self.args if args.loss_mask: idx, targets, mask = batch mask = mask.view(-1) sum_mask = torch.sum(mask).item() logits = self(idx) loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1), reduction='none') loss = torch.sum(loss * mask) / sum_mask # elif args.my_qa_mask != 1: # idx, targets = batch # logits = self(idx) # loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1)) # if '0' in os.environ["RWKV_MY_TESTING"]: # print('logits', logits) # torch.set_printoptions(threshold=10000) # print('idx', idx) # exit(0) else: ##改动 # idx, transcription = batch idx = [item[0] for item in batch] transcription = [item[1] for item in batch] logits, targets, mask = self(idx, transcription) mask = mask.view(-1) sum_mask = torch.sum(mask).item() ###### # idx, targets, mask = batch # mask = mask.view(-1) # sum_mask = torch.sum(mask).item() # # if sum_mask == 0: # # return torch.tensor([0.0], requires_grad=True) # logits = self(idx) if sum_mask == mask.shape[0]: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1)) # print('rank', self.global_rank, 'loss', loss.item()) else: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1), reduction='none') # loss_raw = loss loss = torch.sum(loss * mask) / sum_mask # torch.set_printoptions(threshold=10000) # if True: #self.global_rank == 1: # tmp = '' # sss = 0 # ccc = 0 # for i in range(mask.shape[0]): # if mask[i] > 0: # tmp += str(idx.view(-1)[i].item()) + ',' # sss += loss_raw.view(-1)[i].float().item() # ccc += 1 # print('rank', self.global_rank, 'loss', loss.item(), 'lavg', sss / ccc)#, 'tmp', tmp, 'input', idx) return L2Wrap.apply(loss, logits) def configure_optimizers(self): args = self.args lr_decay = set() lr_1x = set() lr_2x = set() lr_3x = set() for n, p in self.named_parameters(): if not p.requires_grad: continue if (("_w1" in n) or ("_w2" in n)) and (args.layerwise_lr > 0): lr_1x.add(n) elif (("time_mix" in n) or ("time_maa" in n)) and (args.layerwise_lr > 0): if args.my_pile_stage == 2: lr_2x.add(n) else: lr_1x.add(n) elif (("time_decay" in n) or ("time_daaaa" in n)) and (args.layerwise_lr > 0): if args.my_pile_stage == 2: lr_3x.add(n) else: lr_2x.add(n) elif ("time_faaaa" in n) and (args.layerwise_lr > 0): if args.my_pile_stage == 2: lr_2x.add(n) else: lr_1x.add(n) elif ("time_first" in n) and (args.layerwise_lr > 0): lr_3x.add(n) elif (len(p.squeeze().shape) >= 2) and (args.weight_decay > 0): lr_decay.add(n) else: lr_1x.add(n) lr_decay = sorted(list(lr_decay)) lr_1x = sorted(list(lr_1x)) lr_2x = sorted(list(lr_2x)) lr_3x = sorted(list(lr_3x)) # print('decay', lr_decay) # print('1x', lr_1x) # print('2x', lr_2x) # print('3x', lr_3x) param_dict = {n: p for n, p in self.named_parameters()} if args.layerwise_lr > 0: if args.my_pile_stage == 2: optim_groups = [ {"params": [param_dict[n] for n in lr_1x], "weight_decay": 0.0, "my_lr_scale": 1.0}, {"params": [param_dict[n] for n in lr_2x], "weight_decay": 0.0, "my_lr_scale": 5.0},# test: 2e-3 / args.lr_init}, {"params": [param_dict[n] for n in lr_3x], "weight_decay": 0.0, "my_lr_scale": 5.0},# test: 3e-3 / args.lr_init}, ] else: optim_groups = [ {"params": [param_dict[n] for n in lr_1x], "weight_decay": 0.0, "my_lr_scale": 1.0}, {"params": [param_dict[n] for n in lr_2x], "weight_decay": 0.0, "my_lr_scale": 2.0}, {"params": [param_dict[n] for n in lr_3x], "weight_decay": 0.0, "my_lr_scale": 3.0}, ] else: optim_groups = [{"params": [param_dict[n] for n in lr_1x], "weight_decay": 0.0, "my_lr_scale": 1.0}] if args.weight_decay > 0: optim_groups += [{"params": [param_dict[n] for n in lr_decay], "weight_decay": args.weight_decay, "my_lr_scale": 1.0}] if self.deepspeed_offload: return DeepSpeedCPUAdam(optim_groups, lr=self.args.lr_init, betas=self.args.betas, eps=self.args.adam_eps, bias_correction=True, adamw_mode=True, amsgrad=False) return FusedAdam(optim_groups, lr=self.args.lr_init, betas=self.args.betas, eps=self.args.adam_eps, bias_correction=True, adam_w_mode=True, amsgrad=False) else: if self.deepspeed_offload: return DeepSpeedCPUAdam(optim_groups, lr=self.args.lr_init, betas=self.args.betas, eps=self.args.adam_eps, bias_correction=True, adamw_mode=False, weight_decay=0, amsgrad=False) return FusedAdam(optim_groups, lr=self.args.lr_init, betas=self.args.betas, eps=self.args.adam_eps, bias_correction=True, adam_w_mode=False, weight_decay=0, amsgrad=False) # return ZeroOneAdam(optim_groups, lr=self.args.lr_init, betas=self.args.betas, eps=self.args.adam_eps, bias_correction=True, weight_decay=0, amsgrad=False, cuda_aware=False) def return_tokenizer(self): return self.language_tokenizer @property def config(self): return self.language_model.config @property def device(self): return self._device @device.setter def device(self, value): self._device = value @property def deepspeed_offload(self) -> bool: strategy = self.trainer.strategy if isinstance(strategy, DeepSpeedStrategy): cfg = strategy.config["zero_optimization"] return cfg.get("offload_optimizer") or cfg.get("offload_param") return False