{ "architectures": [ "ASCHForCausalLM" ], "model_type": "asch", "torch_dtype": "float16", "transformers_version": "4.40.0", "vocab_size": 151936, "hidden_size": 2048, "num_hidden_layers": 24, "num_attention_heads": 16, "max_position_embeddings": 8192 }