File size: 3,730 Bytes
5b25cac
 
 
 
 
 
 
 
 
 
 
 
85c6bd9
5b25cac
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
"""
Inference / demo script for Spatial Context Networks (SCN).
Designed for use as a HuggingFace Space or standalone demo.

Usage:
    python inference.py
    python inference.py --checkpoint path/to/model.pt --input_dim 10
"""

import argparse
import torch
import json
from model import SpatialContextNetwork


PATTERN_LABELS = ["Mathematics", "Language", "Vision", "Reasoning"]


def load_model(checkpoint_path: str | None, input_dim: int, n_neurons: int, output_dim: int):
    model = SpatialContextNetwork(
        input_dim=input_dim,
        n_neurons=n_neurons,
        output_dim=output_dim,
    )
    if checkpoint_path:
        state = torch.load(checkpoint_path, map_location="cpu")
        model.load_state_dict(state)
        print(f"Loaded checkpoint: {checkpoint_path}")
    else:
        print("No checkpoint provided — using randomly initialized weights.")
    model.eval()
    return model


def run_inference(model: SpatialContextNetwork, x: torch.Tensor) -> dict:
    """
    Run a single forward pass and return rich diagnostic output.

    Returns:
        dict with output logits, predicted pattern, network efficiency stats.
    """
    with torch.no_grad():
        output = model(x)
        stats = model.get_network_stats(x)

    probs = torch.softmax(output, dim=-1)
    predicted_idx = probs.argmax(dim=-1)

    results = {
        "output_logits": output.tolist(),
        "output_probabilities": probs.tolist(),
        "predicted_pattern": [PATTERN_LABELS[i] for i in predicted_idx.tolist()],
        "mean_active_neurons": round(stats["mean_active_neurons"], 2),
        "network_efficiency": round(stats["network_efficiency"], 4),
        "mean_context_score": round(stats["mean_context_score"], 4),
    }
    return results


def demo(args):
    model = load_model(args.checkpoint, args.input_dim, args.n_neurons, args.output_dim)

    total_params = sum(p.numel() for p in model.parameters())
    print(f"\n{'='*60}")
    print("  Spatial Context Network — Inference Demo")
    print(f"{'='*60}")
    print(f"  Input dim   : {args.input_dim}")
    print(f"  Hidden neurons: {args.n_neurons}")
    print(f"  Output dim  : {args.output_dim}")
    print(f"  Parameters  : {total_params}")
    print(f"{'='*60}\n")

    torch.manual_seed(42)
    x = torch.randn(args.batch_size, args.input_dim)
    print(f"Running inference on {args.batch_size} random samples...\n")

    results = run_inference(model, x)

    for i in range(args.batch_size):
        probs = results["output_probabilities"][i]
        predicted = results["predicted_pattern"][i]
        prob_str = " | ".join(
            f"{label}: {p:.3f}" for label, p in zip(PATTERN_LABELS, probs)
        )
        print(f"  Sample {i}: [{prob_str}]  → Predicted: {predicted}")

    print(f"\n  Network Stats:")
    print(f"    Active neurons : {results['mean_active_neurons']} / {args.n_neurons}")
    print(f"    Efficiency     : {results['network_efficiency']:.1%}")
    print(f"    Context score  : {results['mean_context_score']:.4f}")

    if args.output_json:
        with open(args.output_json, "w") as f:
            json.dump(results, f, indent=2)
        print(f"\n  Results saved to {args.output_json}")


if __name__ == "__main__":
    parser = argparse.ArgumentParser(description="SCN Inference Demo")
    parser.add_argument("--checkpoint", type=str, default=None)
    parser.add_argument("--input_dim", type=int, default=10)
    parser.add_argument("--n_neurons", type=int, default=32)
    parser.add_argument("--output_dim", type=int, default=4)
    parser.add_argument("--batch_size", type=int, default=8)
    parser.add_argument("--output_json", type=str, default=None)
    args = parser.parse_args()
    demo(args)