michaelwaves's picture
Add files using upload-large-folder tool
cebd1c6 verified
raw
history blame contribute delete
776 Bytes
{"block_size": 1024, "vocab_size": 2048, "n_layer": 8, "n_head": 32, "d_head": 16, "d_model": 512, "dropout": 0.0, "bias": true, "ln_bias": true, "rms_norm": true, "residual_activation_type": "identity", "activation_type": "gelu", "afrac": 0.0625, "afrac_loctypes": "attn_in,attn_out,mlp_in,mlp_out,mlp_neuron,attn_v", "afrac_approx": false, "enable_sparse_kernels": false, "ignore_dw_grad": false, "debug_nans": false, "tied_unembed": false, "bigram_table_rank": null, "tokenizer_name": "tinypython_2k", "pfrac": 1, "expansion_factor": 2, "expansion_factor_mlp": 2, "debug_exact_topk": true, "grad_checkpointing": true, "sparse_matmul_impl": "cuda", "d_mlp": 2048, "flash": true, "enable_bigram_table": true, "learnable_bigram_table": true, "d_pos_emb": null, "n_embd": null}