antonypamo commited on
Commit
7f2b3d0
·
verified ·
1 Parent(s): 50e8558

Upload 13 files

Browse files
assets/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.56.1"
6
+ }
assets/outputs_rrf_simbio/checkpoint-63/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adaptive_state": {
3
+ "meta": {
4
+ "fusion_mode": "SAVANT\u2013\u03a6\u2084.1\u221e+ symbiotic",
5
+ "fusion_timestamp": "2025-10-09"
6
+ },
7
+ "resonance_vector": [
8
+ 0.006346969060972333
9
+ ]
10
+ },
11
+ "architectures": [
12
+ "BertForSequenceClassification"
13
+ ],
14
+ "attention_probs_dropout_prob": 0.1,
15
+ "classifier_dropout": null,
16
+ "dtype": "float32",
17
+ "geometry": "icosa\u2013dodeca\u2013resonant",
18
+ "gradient_checkpointing": false,
19
+ "hidden_act": "gelu",
20
+ "hidden_dropout_prob": 0.1,
21
+ "hidden_size": 768,
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "layer_norm_eps": 1e-12,
25
+ "max_position_embeddings": 512,
26
+ "model_type": "bert",
27
+ "num_attention_heads": 12,
28
+ "num_hidden_layers": 12,
29
+ "pad_token_id": 0,
30
+ "position_embedding_type": "absolute",
31
+ "problem_type": "single_label_classification",
32
+ "rrf_mode": "\u03a6\u2084.1\u221e+",
33
+ "symbiotic_energy": -0.695067822933197,
34
+ "transformers_version": "4.57.0",
35
+ "type_vocab_size": 2,
36
+ "use_cache": true,
37
+ "vocab_size": 30522
38
+ }
assets/outputs_rrf_simbio/checkpoint-63/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf8ccf5d63e49ea13a92e38ece92a618d45c32d2a24fed9abb4880ffe23ab9b
3
+ size 437958648
assets/outputs_rrf_simbio/checkpoint-63/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f82501ade94aff2225514cd2c189ab6a9c7b7db42ec6852e611f866e7198ad2
3
+ size 876033163
assets/outputs_rrf_simbio/checkpoint-63/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d9997ac7114ba0fce35b01490f28374d48edc62599868a2f281645498cf19d
3
+ size 14455
assets/outputs_rrf_simbio/checkpoint-63/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1b563dc1a68072e34cb2921b1281f6e4681ac6ed493f4d0b61fa0b6eca73deb
3
+ size 1465
assets/outputs_rrf_simbio/checkpoint-63/trainer_state.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 63,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.31746031746031744,
14
+ "grad_norm": 0.43154916167259216,
15
+ "learning_rate": 2.095238095238095e-05,
16
+ "loss": 0.1419,
17
+ "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.6349206349206349,
21
+ "grad_norm": 0.08845952898263931,
22
+ "learning_rate": 1.1428571428571429e-05,
23
+ "loss": 0.0067,
24
+ "step": 40
25
+ },
26
+ {
27
+ "epoch": 0.9523809523809523,
28
+ "grad_norm": 0.061147306114435196,
29
+ "learning_rate": 1.9047619047619047e-06,
30
+ "loss": 0.0028,
31
+ "step": 60
32
+ }
33
+ ],
34
+ "logging_steps": 20,
35
+ "max_steps": 63,
36
+ "num_input_tokens_seen": 0,
37
+ "num_train_epochs": 1,
38
+ "save_steps": 500,
39
+ "stateful_callbacks": {
40
+ "TrainerControl": {
41
+ "args": {
42
+ "should_epoch_stop": false,
43
+ "should_evaluate": false,
44
+ "should_log": false,
45
+ "should_save": true,
46
+ "should_training_stop": true
47
+ },
48
+ "attributes": {}
49
+ }
50
+ },
51
+ "total_flos": 65777763840000.0,
52
+ "train_batch_size": 4,
53
+ "trial_name": null,
54
+ "trial_params": null
55
+ }
assets/outputs_rrf_simbio/checkpoint-63/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ce2259296581ee2cae5d76a38da0408566b6f0dd78b14cee8f74358de9306dc
3
+ size 5841
assets/rrf_savant_cpu_unified_state.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "meta": {
3
+ "version": "\u03a6\u2084.1\u221e+",
4
+ "mode": "Savant Symbiotic CPU",
5
+ "repo": "antonypamo/Savant_RRF_Simbio",
6
+ "trained_on": "quantum_oscillations"
7
+ },
8
+ "adaptive_state": {
9
+ "resonance_vector": [
10
+ 0.006346969060972333
11
+ ],
12
+ "meta": {
13
+ "fusion_timestamp": "2025-10-09",
14
+ "fusion_mode": "SAVANT\u2013\u03a6\u2084.1\u221e+ symbiotic"
15
+ }
16
+ },
17
+ "memory_fused_keys": [
18
+ "memory",
19
+ "amplitudes",
20
+ "entropy",
21
+ "adaptive_state",
22
+ "persistent_memory"
23
+ ]
24
+ }
assets/rrf_savant_state.json ADDED
The diff for this file is too large to render. See raw diff
 
assets/rrf_savant_symbiotic_adaptive.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "fusion_model": {
3
+ "alpha": 0.000800616224296391,
4
+ "beta": 0.48815417289733887,
5
+ "gamma": 0.5110452175140381,
6
+ "training_steps": 200,
7
+ "learning_rate": 0.01,
8
+ "adaptive_rule": "Homeostasis resonante \u03a6\u2084.3"
9
+ },
10
+ "meta": {
11
+ "version": "\u03a6\u2084.3",
12
+ "description": "Memoria simbi\u00f3tica resonante base para RRF-Savant",
13
+ "device": "cpu"
14
+ }
15
+ }
assets/savant_rrf_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9201822cda404262b9c050ff2a638a01cb7f5b65a89bb085764002987620e36e
3
+ size 103585
assets/trainer_state.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [],
12
+ "logging_steps": 100,
13
+ "max_steps": 3,
14
+ "num_input_tokens_seen": 0,
15
+ "num_train_epochs": 3,
16
+ "save_steps": 10000,
17
+ "stateful_callbacks": {
18
+ "TrainerControl": {
19
+ "args": {
20
+ "should_epoch_stop": false,
21
+ "should_evaluate": false,
22
+ "should_log": false,
23
+ "should_save": true,
24
+ "should_training_stop": true
25
+ },
26
+ "attributes": {}
27
+ }
28
+ },
29
+ "total_flos": 391945125888.0,
30
+ "train_batch_size": 4,
31
+ "trial_name": null,
32
+ "trial_params": null
33
+ }