| import argparse | |
| import torch | |
| from math import log2 | |
| from text_size import calculate_text_size_per_token | |
| from likelihood import calculate_negative_log_likelihood | |
| from flops import gqa_model_theoretical_flops, mla_model_theoretical_flops | |
| def calculate_information_capacity( | |
| model_path: str, | |
| data_path: str, | |
| max_sample_length: int = 1024, | |
| batch_size: int = 1, | |
| numerator_bias: float = None, | |
| attention_mechanism: str = None, | |
| ) -> float: | |
| if attention_mechanism is None: | |
| attention_mechanism = "mla" if "deepseek" in model_path.lower() else "gqa" | |
| else: | |
| attention_mechanism = attention_mechanism.lower() | |
| if attention_mechanism != "gqa" and attention_mechanism != "mla": | |
| raise NotImplementedError("attention_mechanism argument should be either gqa or mla") | |
| if numerator_bias is None: | |
| if "mixed_text.jsonl" in data_path: numerator_bias = -24 | |
| elif "Ch-FineWeb-Edu.jsonl" in data_path: numerator_bias = -18.7 | |
| else: numerator_bias = -27 | |
| print(f"numerator_bias is not designated, default to {numerator_bias} based on the data_path") | |
| ts_results = calculate_text_size_per_token(data_path, model_path, target_token_length=max_sample_length) | |
| avg_ts = ts_results["mean_text_size"] | |
| for k, v in ts_results.items(): print(f"{k}: {v}") | |
| nlls = calculate_negative_log_likelihood(model_path, data_path, max_sample_length, batch_size=batch_size, num_samples=ts_results["total_valid_lines"]) | |
| avg_nll = torch.nanmean(nlls).item() | |
| print(f"Average negative log-likelihood: {avg_nll}") | |
| cfg_path = model_path + "/config.json" | |
| if attention_mechanism == "gqa": | |
| flops_results = gqa_model_theoretical_flops(cfg_path, gen_len=max_sample_length) | |
| elif attention_mechanism == "mla": | |
| flops_results = mla_model_theoretical_flops(cfg_path, gen_len=max_sample_length) | |
| per_token_flops = flops_results["decode_total_TFLOPs"] * 1e12 / max_sample_length | |
| for k, v in flops_results.items(): print(f"{k}: {v}") | |
| ic = (avg_ts - avg_nll + numerator_bias) / log2(per_token_flops) | |
| print(f"\nInformation capacity: {ic}") | |
| return ic | |
| def main(): | |
| parser = argparse.ArgumentParser( | |
| description="Compute the information capacity of a language model." | |
| ) | |
| parser.add_argument("-m", "--model_path", type=str, required=True, help="Path to the model directory.") | |
| parser.add_argument("-d", "--data_path", type=str, required=True, help="Path to the dataset (JSONL format).") | |
| parser.add_argument("-l", "--max_sample_length", type=int, default=1024, help="Maximum token length for each sample.") | |
| parser.add_argument("-b", "--batch_size", type=int, default=1, help="Batch size for evaluation.") | |
| parser.add_argument("-n", "--numerator_bias", type=float, default=None, | |
| help="Optional numerator bias. If not set, inferred automatically.") | |
| parser.add_argument("-a", "--attention_mechanism", type=str, choices=["gqa", "mla"], default=None, | |
| help="Specify attention mechanism ('gqa' or 'mla'). If not set, inferred automatically.") | |
| args = parser.parse_args() | |
| calculate_information_capacity( | |
| model_path=args.model_path, | |
| data_path=args.data_path, | |
| max_sample_length=args.max_sample_length, | |
| batch_size=args.batch_size, | |
| numerator_bias=args.numerator_bias, | |
| attention_mechanism=args.attention_mechanism, | |
| ) | |
| if __name__ == "__main__": | |
| main() |