|
|
import argparse |
|
|
import os |
|
|
from types import MethodType |
|
|
|
|
|
import torch |
|
|
from vllm import LLM, SamplingParams |
|
|
|
|
|
|
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument("-m", "--model", type=str, default="meta-llama/Llama-2-7b-hf") |
|
|
parser.add_argument("-i", "--id_path", type=str, required=True) |
|
|
parser.add_argument("-t", "--type", type=str, default="llama") |
|
|
parser.add_argument("-s", "--save_folder", type=str, required=True) |
|
|
parser.add_argument("-n", "--name", type=str, required=True) |
|
|
args = parser.parse_args() |
|
|
|
|
|
model = LLM(model=args.model, tensor_parallel_size=torch.cuda.device_count(), enforce_eager=True) |
|
|
|
|
|
max_length = model.llm_engine.model_config.max_model_len |
|
|
num_layers = model.llm_engine.model_config.hf_config.num_hidden_layers |
|
|
intermediate_size = model.llm_engine.model_config.hf_config.intermediate_size |
|
|
|
|
|
over_zero = torch.zeros(num_layers, intermediate_size, dtype=torch.int32).to('cuda') |
|
|
|
|
|
def extract_lang(id_path): |
|
|
parts = id_path.split('/') |
|
|
id = parts[-1] |
|
|
|
|
|
lang = id.split('.')[1] |
|
|
return lang |
|
|
|
|
|
def factory(idx): |
|
|
def llama_forward(self, x): |
|
|
gate_up, _ = self.gate_up_proj(x) |
|
|
i = gate_up.size(-1) |
|
|
gate_up[:, : i // 2] = torch.nn.SiLU()(gate_up[:, : i // 2]) |
|
|
activation = gate_up[:, : i // 2].float() |
|
|
over_zero[idx, :] += (activation > 0).sum(dim=0) |
|
|
x = gate_up[:, : i // 2] * gate_up[:, i // 2 :] |
|
|
x, _ = self.down_proj(x) |
|
|
return x |
|
|
|
|
|
def qwen_forward(self, x): |
|
|
gate_up, _ = self.gate_up_proj(x) |
|
|
intermediate_size = gate_up.size(-1) // 2 |
|
|
gate = gate_up[..., :intermediate_size] |
|
|
up = gate_up[..., intermediate_size:] |
|
|
gate_activation = torch.nn.functional.silu(gate) |
|
|
over_zero[idx, :] += (gate_activation > 0).sum(dim=0) |
|
|
x, _ = self.down_proj(gate_activation * up) |
|
|
return x |
|
|
|
|
|
if args.type == 'llama': |
|
|
return llama_forward |
|
|
else: |
|
|
return qwen_forward |
|
|
|
|
|
for i in range(num_layers): |
|
|
obj = model.llm_engine.driver_worker.model_runner.model.model.layers[i].mlp |
|
|
obj.forward = MethodType(factory(i), obj) |
|
|
|
|
|
lang = args.lang |
|
|
|
|
|
ids = torch.load(args.id_path) |
|
|
|
|
|
l = ids.size(0) |
|
|
l = min(l, 99999744) // max_length * max_length |
|
|
input_ids = ids[:l].reshape(-1, max_length) |
|
|
|
|
|
output = model.generate(prompt_token_ids=input_ids.tolist(), sampling_params=SamplingParams(max_tokens=1)) |
|
|
|
|
|
output = dict(n=l, over_zero=over_zero.to('cpu')) |
|
|
|
|
|
save_path = os.path.join(args.save_folder, f"activation.{lang}.{args.name}.pt") |
|
|
|
|
|
torch.save(output, save_path) |