""" Inference / demo script for Spatial Context Networks (SCN). Designed for use as a HuggingFace Space or standalone demo. Usage: python inference.py python inference.py --checkpoint path/to/model.pt --input_dim 10 """ import argparse import torch import json from model import SpatialContextNetwork PATTERN_LABELS = ["Mathematics", "Language", "Vision", "Reasoning"] def load_model(checkpoint_path: str | None, input_dim: int, n_neurons: int, output_dim: int): model = SpatialContextNetwork( input_dim=input_dim, n_neurons=n_neurons, output_dim=output_dim, ) if checkpoint_path: state = torch.load(checkpoint_path, map_location="cpu") model.load_state_dict(state) print(f"Loaded checkpoint: {checkpoint_path}") else: print("No checkpoint provided — using randomly initialized weights.") model.eval() return model def run_inference(model: SpatialContextNetwork, x: torch.Tensor) -> dict: """ Run a single forward pass and return rich diagnostic output. Returns: dict with output logits, predicted pattern, network efficiency stats. """ with torch.no_grad(): output = model(x) stats = model.get_network_stats(x) probs = torch.softmax(output, dim=-1) predicted_idx = probs.argmax(dim=-1) results = { "output_logits": output.tolist(), "output_probabilities": probs.tolist(), "predicted_pattern": [PATTERN_LABELS[i] for i in predicted_idx.tolist()], "mean_active_neurons": round(stats["mean_active_neurons"], 2), "network_efficiency": round(stats["network_efficiency"], 4), "mean_context_score": round(stats["mean_context_score"], 4), } return results def demo(args): model = load_model(args.checkpoint, args.input_dim, args.n_neurons, args.output_dim) total_params = sum(p.numel() for p in model.parameters()) print(f"\n{'='*60}") print(" Spatial Context Network — Inference Demo") print(f"{'='*60}") print(f" Input dim : {args.input_dim}") print(f" Hidden neurons: {args.n_neurons}") print(f" Output dim : {args.output_dim}") print(f" Parameters : {total_params}") print(f"{'='*60}\n") torch.manual_seed(42) x = torch.randn(args.batch_size, args.input_dim) print(f"Running inference on {args.batch_size} random samples...\n") results = run_inference(model, x) for i in range(args.batch_size): probs = results["output_probabilities"][i] predicted = results["predicted_pattern"][i] prob_str = " | ".join( f"{label}: {p:.3f}" for label, p in zip(PATTERN_LABELS, probs) ) print(f" Sample {i}: [{prob_str}] → Predicted: {predicted}") print(f"\n Network Stats:") print(f" Active neurons : {results['mean_active_neurons']} / {args.n_neurons}") print(f" Efficiency : {results['network_efficiency']:.1%}") print(f" Context score : {results['mean_context_score']:.4f}") if args.output_json: with open(args.output_json, "w") as f: json.dump(results, f, indent=2) print(f"\n Results saved to {args.output_json}") if __name__ == "__main__": parser = argparse.ArgumentParser(description="SCN Inference Demo") parser.add_argument("--checkpoint", type=str, default=None) parser.add_argument("--input_dim", type=int, default=10) parser.add_argument("--n_neurons", type=int, default=32) parser.add_argument("--output_dim", type=int, default=4) parser.add_argument("--batch_size", type=int, default=8) parser.add_argument("--output_json", type=str, default=None) args = parser.parse_args() demo(args)