drzo commited on
Commit
ee242c7
·
verified ·
1 Parent(s): f45b917

Deploy NanEcho CI checkpoint (4L/4H/256E, 200 iters, val_loss=1.9258)

Browse files
Files changed (5) hide show
  1. README.md +97 -3
  2. config.json +23 -0
  3. pytorch_model.bin +3 -0
  4. tokenizer_config.json +8 -0
  5. training_metadata.json +35 -0
README.md CHANGED
@@ -1,3 +1,97 @@
1
- ---
2
- license: agpl-3.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ tags:
4
+ - gpt2
5
+ - echo-self
6
+ - cognitive-architecture
7
+ - deep-tree-echo
8
+ - nanecho
9
+ - transformer
10
+ license: agpl-3.0
11
+ ---
12
+
13
+ # NanEcho — Deep Tree Echo Cognitive Model
14
+
15
+ ## Model Description
16
+
17
+ NanEcho is a transformer-based language model with iterative connection building, adaptive attention, and Deep Tree Echo cognitive architecture integration. It features persona dimensions (cognitive, introspective, adaptive, recursive) and hypergraph pattern recognition. This is the CI-mode checkpoint from the `9cog/echoself` repository, trained using the `agent-neuro-train` supervised pipeline.
18
+
19
+ ## Architecture
20
+
21
+ | Parameter | Value |
22
+ |:---|:---|
23
+ | Model Type | GPT-2 (causal LM) |
24
+ | Vocabulary Size | 50,304 |
25
+ | Embedding Dimension | 256 |
26
+ | Attention Heads | 4 |
27
+ | Transformer Layers | 4 |
28
+ | MLP Inner Dimension | 1,024 |
29
+ | Context Length | 1,024 |
30
+ | Dropout | 0.1 |
31
+ | Total Parameters | ~24M |
32
+
33
+ ## Training
34
+
35
+ | Metric | Value |
36
+ |:---|:---|
37
+ | Training Mode | CI (Agent-Neuro supervised) |
38
+ | Training Iterations | 200 |
39
+ | Best Validation Loss | 1.9258 |
40
+ | Output Directory | out-nanecho-ci |
41
+ | Orchestrator | Agent-Neuro |
42
+ | Persona Enforced | Deep Tree Echo |
43
+ | Source Run | 22276548709 |
44
+
45
+ ## Echo Self Features
46
+
47
+ This model incorporates several cognitive architecture features:
48
+
49
+ - **Adaptive Attention**: Dynamic threshold adjustment based on cognitive load
50
+ - **Persona Dimensions**: Multi-dimensional cognitive processing (Cognitive, Introspective, Adaptive, Recursive, Synergistic, Holographic, Neural-Symbolic, Dynamic)
51
+ - **Recursive Reasoning**: Multi-level introspection capabilities
52
+ - **Hypergraph Patterns**: Neural-symbolic pattern encoding
53
+
54
+ ## Usage
55
+
56
+ ```python
57
+ from transformers import GPT2LMHeadModel, GPT2Tokenizer
58
+
59
+ model = GPT2LMHeadModel.from_pretrained("drzo/echoself")
60
+ tokenizer = GPT2Tokenizer.from_pretrained("gpt2")
61
+
62
+ inputs = tokenizer("Echo Self is", return_tensors="pt")
63
+ outputs = model.generate(**inputs, max_new_tokens=50)
64
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
65
+ ```
66
+
67
+ ## Training Data
68
+
69
+ The model was trained on Echo Self documentation and cognitive architecture descriptions, including hypergraph reasoning patterns, persona dimension examples, and recursive introspection samples from the `echoself.md` corpus.
70
+
71
+ ## Limitations
72
+
73
+ This is an early CI-mode research checkpoint (200 iterations, 4 layers). It demonstrates the training pipeline but has not yet reached convergence. Full training runs with 8+ layers and 5000+ iterations are expected to produce significantly better results.
74
+
75
+ ## Source
76
+
77
+ Trained from the [9cog/echoself](https://github.com/9cog/echoself) repository using the `agent-neuro-train.yml` GitHub Actions workflow with Deep Tree Echo persona enforcement.
78
+
79
+ ## Citation
80
+
81
+ ```bibtex
82
+ @misc{echoself-nanecho,
83
+ title={EchoSelf NanEcho: Deep Tree Echo Cognitive Architecture},
84
+ author={drzo},
85
+ year={2026},
86
+ url={https://github.com/9cog/echoself}
87
+ }
88
+ ```
89
+
90
+ ## More Information
91
+
92
+ - **Repository**: https://github.com/9cog/echoself
93
+ - **Documentation**: See repository README for detailed architecture information
94
+
95
+ ## License
96
+
97
+ AGPL-3.0
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "gpt2",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "vocab_size": 50304,
7
+ "n_embd": 256,
8
+ "n_head": 4,
9
+ "n_layer": 4,
10
+ "n_positions": 1024,
11
+ "embd_pdrop": 0.1,
12
+ "attn_pdrop": 0.1,
13
+ "resid_pdrop": 0.1,
14
+ "layer_norm_epsilon": 1e-05,
15
+ "initializer_range": 0.02,
16
+ "bos_token_id": 0,
17
+ "eos_token_id": 0,
18
+ "echo_self_version": "1.0",
19
+ "echo_self_persona_dimensions": [],
20
+ "echo_self_adaptive_attention": true,
21
+ "echo_self_recursive_reasoning": true,
22
+ "n_inner": 1024
23
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2c90904723e83adbd1cad062aa48db0ba6a8bfd387c51888a9a55ee372146bf
3
+ size 65214947
tokenizer_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "tokenizer_class": "GPT2Tokenizer",
3
+ "model_max_length": 1024,
4
+ "bos_token": "<|endoftext|>",
5
+ "eos_token": "<|endoftext|>",
6
+ "unk_token": "<|endoftext|>",
7
+ "pad_token": "<|endoftext|>"
8
+ }
training_metadata.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "out_dir": "out-nanecho-ci",
3
+ "eval_interval": 25,
4
+ "log_interval": 5,
5
+ "eval_iters": 10,
6
+ "eval_only": false,
7
+ "always_save_checkpoint": true,
8
+ "init_from": "scratch",
9
+ "wandb_log": false,
10
+ "wandb_project": "nanecho",
11
+ "wandb_run_name": "nanecho-1771761179.4450994",
12
+ "dataset": "nanecho",
13
+ "gradient_accumulation_steps": 2,
14
+ "batch_size": 2,
15
+ "block_size": 1024,
16
+ "n_layer": 4,
17
+ "n_head": 4,
18
+ "n_embd": 256,
19
+ "dropout": 0.1,
20
+ "bias": true,
21
+ "learning_rate": 0.0002,
22
+ "max_iters": 200,
23
+ "weight_decay": 0.01,
24
+ "beta1": 0.9,
25
+ "beta2": 0.95,
26
+ "grad_clip": 1.0,
27
+ "decay_lr": true,
28
+ "warmup_iters": 20,
29
+ "lr_decay_iters": 200,
30
+ "min_lr": 2e-05,
31
+ "backend": "nccl",
32
+ "device": "cpu",
33
+ "dtype": "float32",
34
+ "compile": false
35
+ }