LisaV3 / config.json
Qybera's picture
Upload 14 files
1bc5752 verified
{
"model_type": "lisa",
"architecture": "Lisa_multimodal_transformer",
"lisa_metadata": {
"model_name": "LISA (Learning Intelligence with Sensory Awareness)",
"version": "3.5",
"development_location": "Kenya, Africa",
"development_team": "LISA Team",
"development_country": "Kenya",
"development_continent": "Africa",
"created_date": "2025-08-20T03:07:26.809423",
"architecture_type": "Lisa Multimodal Transformer",
"inspiration": "Vision Transformer (ViT-B/16) architecture, built from scratch",
"capabilities": [
"Multimodal processing (vision, audio, text)",
"Real-time perception and interaction",
"Environmental awareness",
"Lisa object detection",
"Speech recognition and synthesis",
"Emotion detection",
"Autonomous learning"
],
"training_philosophy": "Built from scratch without pretrained models for maximum Lisaization",
"team_location": "Kenya, East Africa",
"cultural_context": "Developed in Africa for global impact"
},
"vision_config": {
"architecture": "Lisa_vit",
"patch_size": 16,
"embed_dim": 768,
"num_layers": 12,
"num_heads": 6,
"image_size": 224,
"num_classes": 80
},
"audio_config": {
"architecture": "Lisa_audio_transformer",
"sample_rate": 16000,
"embed_dim": 512,
"num_layers": 6,
"num_heads": 8,
"vocab_size": 32,
"n_mels": 80
},
"multimodal_config": {
"fusion_strategy": "cross_attention",
"max_sequence_length": 512,
"supports_streaming": true,
"real_time_processing": true,
"hidden_dim": 1024,
"fusion_hidden_dim": 1024
},
"training_config": {},
"torch_dtype": "float32",
"transformers_version": "4.36.0",
"lisa_version": "3.5",
"Lisa_implementation": true,
"pretrained_base": null,
"self_awareness": {
"knows_origin": true,
"development_location": "Kenya, Africa",
"development_team": "LISA Team",
"cultural_identity": "African AI development"
}
}