irodkin commited on
Commit
a7fed0d
·
verified ·
1 Parent(s): 6a0292c

Training checkpoint at step 6000

Browse files
Files changed (1) hide show
  1. config.json +35 -0
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "act_format": "linear",
3
+ "act_on": false,
4
+ "act_type": "associative",
5
+ "architectures": [
6
+ "InnerLoopARMTForCausalLM"
7
+ ],
8
+ "attend_to_previous_input": false,
9
+ "base_model_config": null,
10
+ "base_model_name": "meta-llama/Llama-3.2-1B",
11
+ "constant_depth": false,
12
+ "correction": true,
13
+ "d_mem": 64,
14
+ "dtype": "bfloat16",
15
+ "freeze_mem": false,
16
+ "gating": false,
17
+ "layers_attr": "model.layers",
18
+ "max_hop": 4,
19
+ "model_type": "armt",
20
+ "n_heads": 1,
21
+ "noisy_halting": false,
22
+ "num_mem_tokens": 32,
23
+ "segment_alignment": "left",
24
+ "segment_size": 1024,
25
+ "sliding_window": true,
26
+ "time_penalty": 0.0,
27
+ "transformers_version": "4.57.1",
28
+ "use_denom": true,
29
+ "use_sink": true,
30
+ "wrap_pos": false,
31
+ "auto_map": {
32
+ "AutoConfig": "modeling_armt.ARMTConfig",
33
+ "AutoModelForCausalLM": "modeling_armt.InnerLoopARMTForCausalLM"
34
+ }
35
+ }