You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Defense System Animation

LEGION AUTONOMOUS DEFENSE SYSTEM

Made by Death Legion Cyber Team LK

License: MIT Library: SafeTensors Task: Object Detection AWS SageMaker Google Cloud

Overview

Legion is an advanced autonomous defense system powered by artificial intelligence. It provides multi-spectral threat detection and automated response capabilities for defense applications.

Vision System

Core Capabilities

Capability Description
Multi-Spectral Vision Threat detection across RGB, thermal, hyperspectral, and SAR radar spectrums
Real-Time Analysis Machine-speed inference for immediate threat assessment
Automated Response Intelligent action protocols with ethical safety guardrails
Persistent Learning Tactical database for continuous threat signature updates

Performance Metrics

Multi-Spectral Performance Comparison

Latency Comparison Chart Accuracy Comparison Chart Spectrum Coverage Radar

Detailed Performance Table

Spectrum Input Use Case Accuracy Latency
RGB Vision High-resolution imagery Daylight military asset identification 97.5% 15ms
Thermal/Infrared Heat signature maps Night operations, missile detection 96.8% 12ms
Hyperspectral Multi-band spectral data Camouflage penetration, material analysis 94.2% 25ms
SAR Radar Synthetic aperture returns All-weather, cloud/smoke penetration 93.5% 30ms

Quick Start

Installation

pip install torch torchvision transformers safetensors huggingface-hub pillow numpy

Hugging Face Inference API

import requests
import base64

# Encode image
with open("threat_image.jpg", "rb") as f:
    image_bytes = f.read()
    image_b64 = base64.b64encode(image_bytes).decode()

# Call HF Inference API
API_URL = "https://api-inference.huggingface.co/models/Pnny13/legion-defense-system"
headers = {"Authorization": f"Bearer {YOUR_HF_TOKEN}"}

payload = {
    "inputs": image_b64,
    "spectrum": "rgb",
    "confidence_threshold": 0.5
}

response = requests.post(API_URL, headers=headers, json=payload)
result = response.json()

print(f"Detected {len(result['detections'])} threats")
for det in result['detections']:
    print(f"  - {det['label']}: {det['score']:.2%}")

AWS SageMaker Deployment

import boto3
import json

# Deploy to SageMaker
sagemaker = boto3.client('sagemaker')

# Create model
response = sagemaker.create_model(
    ModelName='legion-defense-system',
    PrimaryContainer={
        'Image': '763104351884.dkr.ecr.us-east-1.amazonaws.com/pytorch-inference:2.0.0-cpu-py310',
        'ModelDataUrl': 's3://your-bucket/legion-model.tar.gz',
        'Environment': {
            'SAGEMAKER_PROGRAM': 'serve.py',
            'SAGEMAKER_SUBMIT_DIRECTORY': '/opt/ml/model/code'
        }
    },
    ExecutionRoleArn='arn:aws:iam::YOUR_ACCOUNT:role/SageMakerExecutionRole'
)

# Create endpoint configuration
sagemaker.create_endpoint_config(
    EndpointConfigName='legion-defense-config',
    ProductionVariants=[{
        'VariantName': 'AllTraffic',
        'ModelName': 'legion-defense-system',
        'InitialInstanceCount': 1,
        'InstanceType': 'ml.c5.xlarge'
    }]
)

# Create endpoint
sagemaker.create_endpoint(
    EndpointName='legion-defense-endpoint',
    EndpointConfigName='legion-defense-config'
)

print("SageMaker endpoint deployed successfully!")

Google Cloud Vertex AI

from google.cloud import aiplatform
import base64

# Initialize Vertex AI
aiplatform.init(project='your-project', location='us-central1')

# Deploy model
model = aiplatform.Model.upload(
    display_name='legion-defense-system',
    artifact_uri='gs://your-bucket/legion-model/',
    serving_container_image_uri='us-docker.pkg.dev/vertex-ai/prediction/pytorch-cpu.2-0:latest',
    serving_container_predict_route='/predict',
    serving_container_health_route='/health'
)

# Create endpoint
endpoint = model.deploy(
    deployed_model_display_name='legion-defense-deployed',
    machine_type='n1-standard-4',
    min_replica_count=1,
    max_replica_count=3
)

print(f"Vertex AI endpoint: {endpoint.resource_name}")

# Run prediction
with open("threat_image.jpg", "rb") as f:
    image_b64 = base64.b64encode(f.read()).decode()

prediction = endpoint.predict([{
    "image": image_b64,
    "spectrum": "rgb",
    "confidence_threshold": 0.5
}])

print(prediction)

Local Inference

from handler import LegionInferenceHandler
from PIL import Image

# Initialize handler
handler = LegionInferenceHandler(
    repo_id="Pnny13/legion-defense-system"
)

# Load models
handler.load_models()

# Load image
image = Image.open("threat_image.jpg")

# Run detection on different spectrums
for spectrum in ['rgb', 'thermal', 'hyperspectral', 'sar']:
    detections = handler.predict(
        image=image,
        spectrum=spectrum,
        confidence_threshold=0.5
    )
    print(f"{spectrum.upper()}: {len(detections)} threats detected")

Final Guard Protocol

Guard Protocol

The Final Guard is an automated response system for critical threat scenarios:

Protocol Response Time Description
Nuclear Detection 0.01 seconds Immediate identification of nuclear ignition signatures
Seismic Verification 0.05 seconds Cross-reference with seismic sensor data
Interception Launch 0.10 seconds Automated mid-course interception deployment

Ethical Safety Framework

All automated responses include mandatory ethical verification:

  • Human Oversight: Lethal force requires human authorization
  • Collateral Assessment: Civilian presence evaluation before any strike
  • Civilian Override: Automatic abort if civilians detected in blast radius
  • Audit Trail: Complete decision logging for accountability
  • Multi-Party Authorization: Nuclear protocols require 3-person consent

Natural Language Command Interface

Command Interface

The Point and Attack interface enables natural language control:

python main.py --command "Scan sector 7 for hostile aircraft"

Example Commands:

Command Action
"Scan sector Alpha for tanks" Deploy RGB/Thermal scan
"Track heat signatures in zone 4" Activate thermal tracking
"Detect camouflaged units" Enable hyperspectral analysis
"See through smoke at grid B7" Activate SAR radar
"Intercept incoming missile" Launch Final Guard protocol

Repository Structure

legion-defense-system/
β”œβ”€β”€ README.md                    # Model documentation
β”œβ”€β”€ handler.py                   # HF Inference API handler
β”œβ”€β”€ serve.py                     # Cloud deployment entry point
β”œβ”€β”€ Dockerfile                   # Container for cloud deployment
β”œβ”€β”€ requirements.txt             # Python dependencies
β”œβ”€β”€ rgb.safetensors              # RGB vision model weights
β”œβ”€β”€ thermal.safetensors          # Thermal/IR model weights
β”œβ”€β”€ hyperspectral.safetensors    # Hyperspectral model weights
β”œβ”€β”€ sar.safetensors              # SAR radar model weights
β”œβ”€β”€ manifest.json                # Model metadata
└── analysis/
    └── charts/                  # Performance comparison charts
        β”œβ”€β”€ latency_comparison.png
        β”œβ”€β”€ accuracy_comparison.png
        └── spectrum_coverage_radar.png

Model Specifications

Attribute Value
Architecture Multi-spectral fusion with cross-modal attention
Input Formats RGB (640x640), Thermal (512x512), SAR (512x512), Hyperspectral (128 bands)
Output Bounding boxes, class labels, confidence scores
Classes 50+ military and civilian asset types
Export Format SafeTensors
License MIT
Inference Latency < 30ms (machine speed)
Cloud Support Hugging Face, AWS SageMaker, Google Cloud Vertex AI

Cloud Deployment

Hugging Face Inference Endpoints

Deploy instantly on Hugging Face:

from huggingface_hub import create_inference_endpoint

endpoint = create_inference_endpoint(
    name="legion-defense",
    repository="Pnny13/legion-defense-system",
    framework="pytorch",
    task="object-detection",
    accelerator="cpu",
    instance_type="cpu-small"
)

AWS SageMaker

# Build and push container
docker build -t legion-defense:latest .
docker tag legion-defense:latest your-ecr-repo/legion-defense:latest
docker push your-ecr-repo/legion-defense:latest

# Deploy using AWS CLI
aws sagemaker create-model \
    --model-name legion-defense \
    --primary-container Image=your-ecr-repo/legion-defense:latest

Google Cloud Vertex AI

# Upload model to GCS
gsutil cp -r model/artifacts gs://your-bucket/legion-model/

# Deploy to Vertex AI
gcloud ai models upload \
    --region=us-central1 \
    --display-name=legion-defense \
    --artifact-uri=gs://your-bucket/legion-model/

Credits

Made by Death Legion Cyber Team LK

Team Logo

Advanced Defense Systems Research and Development


License

This project is licensed under the MIT License - see the LICENSE file for details.

Built for Defense. Powered by AI. Protected by Ethics.

Last Updated: 2026-03-08

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support