{ "architectures": [ "Lfm2AudioForConditionalGeneration" ], "codebooks": 8, "tie_audio_embeddings": false, "semantic_codebook_factor": 100, "codebook_weight": "log", "interleaved_n_text": 6, "interleaved_n_audio": 12, "preprocessor": { "sample_rate": 16000, "normalize": "per_feature", "window_size": 0.025, "window_stride": 0.01, "window": "hann", "features": 128, "n_fft": 512, "log": true, "frame_splicing": 1, "dither": 1.0e-05, "pad_to": 0, "pad_value": 0.0 }, "encoder": { "feat_in": 128, "feat_out": -1, "n_layers": 17, "d_model": 512, "subsampling": "dw_striding", "subsampling_factor": 8, "subsampling_conv_channels": 256, "causal_downsampling": false, "reduction": null, "reduction_position": null, "reduction_factor": 1, "ff_expansion_factor": 4, "self_attention_model": "rel_pos", "n_heads": 8, "att_context_size": [ -1, -1 ], "xscaling": false, "untie_biases": true, "pos_emb_max_len": 5000, "conv_kernel_size": 9, "conv_norm_type": "batch_norm", "conv_context_size": null, "dropout": 0.1, "dropout_pre_encoder": 0.1, "dropout_emb": 0, "dropout_att": 0.1 }, "lfm": { "_name_or_path": "LiquidAI/LFM2-1.2B", "architectures": [ "Lfm2ForCausalLM" ], "block_auto_adjust_ff_dim": true, "block_dim": 2048, "block_ff_dim": 12288, "block_ffn_dim_multiplier": 1, "block_mlp_init_scale": 1, "block_multiple_of": 256, "block_norm_eps": 1e-05, "block_out_init_scale": 1, "block_use_swiglu": true, "block_use_xavier_init": true, "conv_L_cache": 3, "conv_bias": false, "conv_dim": 2048, "conv_dim_out": 2048, "conv_use_xavier_init": true, "eos_token_id": 7, "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 12288, "layer_types": [ "conv", "conv", "full_attention", "conv", "conv", "full_attention", "conv", "conv", "full_attention", "conv", "full_attention", "conv", "full_attention", "conv", "full_attention", "conv" ], "max_position_embeddings": 128000, "model_type": "lfm2", "norm_eps": 1e-05, "num_attention_heads": 32, "num_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "rope_theta": 1000000, "torch_dtype": "bfloat16", "use_cache": true, "use_pos_enc": true, "vocab_size": 65536 }, "depthformer": { "layers": 6, "dim": 1024, "tie": true } }