{ "model_type": "symphonym", "architectures": ["UniversalEncoder"], "version": "v7", "vocab_size": 113280, "num_scripts": 25, "num_langs": 1944, "num_length_buckets": 16, "char_embed_dim": 64, "script_embed_dim": 16, "lang_embed_dim": 16, "length_embed_dim": 8, "hidden_dim": 128, "embed_dim": 128, "num_layers": 2, "num_attention_heads": 2, "dropout": 0.2, "lang_dropout": 0.5, "max_toponym_len": 50, "teacher": { "architectures": ["PhoneticEncoder"], "phonetic_feat_dim": 192, "hidden_dim": 128, "embed_dim": 128, "num_layers": 2, "num_attention_heads": 2, "dropout": 0.2, "note": "PanPhon192: 8-bin positional pooling of 24-dim articulatory features" }, "training": { "phase1_epochs": 50, "phase2_epochs": 50, "phase3_epochs": 30, "phase1_val_loss": 0.0056, "phase2_student_teacher_cosine": 0.942, "phase3_val_loss": 0.021217376736320875, "triplet_margin": 0.3, "phase1_lr": 1e-4, "phase2_lr": 5e-4, "phase3_lr": 5e-5, "phase1_batch_size": 2048, "phase2_batch_size": 1024, "phase3_batch_size": 1024, "weight_decay": 1e-5, "warmup_epochs": 2, "noise_prob": 0.3, "mse_weight": 1.0, "cosine_weight": 1.0, "phase2_alpha": 0.5, "phase2_loss": "alpha * MSE(student, teacher) + (1 - alpha) * (1 - cosine(student, teacher))", "phase3_loss": "triplet_margin(anchor, positive, hard_negative, margin=0.3)" }, "g2p": { "backends": ["epitran", "phonikud", "charsiu_g2p"], "epitran_extensions": 102, "scripts_covered": 20, "ipa_coverage": 0.540, "panphon_bins": 8, "panphon_feat_dim": 24, "panphon192_dim": 192 }, "index": { "total_toponyms": 66924548, "embedding_coverage": 1.0, "quantization": "int8", "ann_backend": "elasticsearch_hnsw" } }