| { |
| "architectures": [ |
| "BioGPTForICD10Classification" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "code_to_idx": { |
| "D62": 0, |
| "D649": 1, |
| "D696": 2, |
| "E039": 3, |
| "E1122": 4, |
| "E119": 5, |
| "E669": 6, |
| "E785": 7, |
| "E871": 8, |
| "E872": 9, |
| "F17210": 10, |
| "F329": 11, |
| "F419": 12, |
| "G4700": 13, |
| "G4733": 14, |
| "G8929": 15, |
| "I10": 16, |
| "I110": 17, |
| "I129": 18, |
| "I130": 19, |
| "I2510": 20, |
| "I252": 21, |
| "I480": 22, |
| "I4891": 23, |
| "I5032": 24, |
| "J189": 25, |
| "J449": 26, |
| "J45909": 27, |
| "J9601": 28, |
| "K219": 29, |
| "K5900": 30, |
| "M109": 31, |
| "N179": 32, |
| "N183": 33, |
| "N189": 34, |
| "N390": 35, |
| "N400": 36, |
| "Y92230": 37, |
| "Y92239": 38, |
| "Y929": 39, |
| "Z23": 40, |
| "Z66": 41, |
| "Z7901": 42, |
| "Z7902": 43, |
| "Z794": 44, |
| "Z86718": 45, |
| "Z8673": 46, |
| "Z87891": 47, |
| "Z951": 48, |
| "Z955": 49 |
| }, |
| "dropout_rate": 0.2, |
| "enable_enhanced_attention": true, |
| "enable_enhanced_classifier": true, |
| "hidden_act": "gelu", |
| "hidden_dropout_prob": 0.1, |
| "hidden_size": 1024, |
| "icd_codes": [ |
| "D62", |
| "D649", |
| "D696", |
| "E039", |
| "E1122", |
| "E119", |
| "E669", |
| "E785", |
| "E871", |
| "E872", |
| "F17210", |
| "F329", |
| "F419", |
| "G4700", |
| "G4733", |
| "G8929", |
| "I10", |
| "I110", |
| "I129", |
| "I130", |
| "I2510", |
| "I252", |
| "I480", |
| "I4891", |
| "I5032", |
| "J189", |
| "J449", |
| "J45909", |
| "J9601", |
| "K219", |
| "K5900", |
| "M109", |
| "N179", |
| "N183", |
| "N189", |
| "N390", |
| "N400", |
| "Y92230", |
| "Y92239", |
| "Y929", |
| "Z23", |
| "Z66", |
| "Z7901", |
| "Z7902", |
| "Z794", |
| "Z86718", |
| "Z8673", |
| "Z87891", |
| "Z951", |
| "Z955" |
| ], |
| "initializer_range": 0.02, |
| "intermediate_size": 4096, |
| "layer_norm_eps": 1e-12, |
| "learnable_icd_embeddings": true, |
| "max_position_embeddings": 1024, |
| "model_type": "biogpt-icd10", |
| "num_attention_heads": 16, |
| "num_attention_layers": 16, |
| "num_hidden_layers": 24, |
| "num_icd_codes": 50, |
| "performance_metrics": { |
| "aggregation_method": "mean", |
| "f1_macro": 0.6921683889110797, |
| "f1_micro": 0.7425923267783542, |
| "hamming_loss": 0.06360780339498354, |
| "precision_macro": 0.6175372765898886, |
| "precision_micro": 0.6570744275501688, |
| "recall_macro": 0.7921473231592516, |
| "recall_micro": 0.8537010843941537, |
| "subset_accuracy": 0.09880922219407144 |
| }, |
| "prediction_threshold": 0.25, |
| "scale_embedding": true, |
| "torch_dtype": "float32", |
| "training_details": { |
| "epoch": 31, |
| "timestamp": "2025-06-09T10:51:55.593755" |
| }, |
| "transformers_version": "4.52.4", |
| "use_attention_pooling": true, |
| "use_cache": true, |
| "use_hierarchical_attention": true, |
| "vocab_size": 42384 |
| } |
|
|