| # π Neuronales Netzwerk - Visueller Γberblick | |
| ## Netzwerk-Architektur Visualisierung | |
| ``` | |
| BEISPIEL 1: Einfaches Netzwerk fΓΌr Regression | |
| βββββββββββββββββββββββββββββββββββββββββββββββ | |
| Input Layer Hidden Layer 1 Hidden Layer 2 Output Layer | |
| (5) (8) (4) (1) | |
| xβ ββ hββ oββ Ε· (Vorhersage) | |
| β β±ββ¬βββββββ hββ ββ¬βββββββ oββ ββ | |
| xβ ββ€ β hββ β oββ β | |
| β β± β hββ β oββ βββ Ο(z) ββ Ε· | |
| xβ ββ€ β β±ββ¬ββββ hββ β | |
| β β β hββ β | |
| xβ ββ€ β β±ββ€ hββ β | |
| β β hββ ββ€ | |
| xβ ββ΄ββββ βββββββββββββββ΄ββ | |
| Wβ Gewichte: 5Γ8 Wβ Gewichte: 8Γ4 Wβ Gewichte: 4Γ1 | |
| bβ Bias: 8 bβ Bias: 4 bβ Bias: 1 | |
| ``` | |
| ### Mathematik dahinter: | |
| ``` | |
| Input: [xβ, xβ, xβ, xβ, xβ ] | |
| Hidden 1: hβ = ReLU(x Β· Wβ + bβ) | |
| Hidden 2: hβ = ReLU(hβ Β· Wβ + bβ) | |
| Output: Ε· = Sigmoid(hβ Β· Wβ + bβ) | |
| Loss: L = (Ε· - y)Β² | |
| ``` | |
| --- | |
| ## BEISPIEL 2: Klassifizierung (3 Klassen) | |
| βββββββββββββββββββββββββββββββββββββββββ | |
| ``` | |
| Input Features (10) β Hidden 1 (16) β Hidden 2 (8) β Output (3) | |
| Input: Klasse 1 (Katze) | |
| [fβ] [oβ] | |
| [fβ] [oβ] β argmax β Vorhersage | |
| ... ββ Wβ,bβ β ReLU β Wβ,bβ β ReLU β [oβ] | |
| [fββ] | |
| (Prob. fΓΌr jede Klasse) | |
| ``` | |
| --- | |
| ## Trainings-Prozess Visualisierung | |
| ``` | |
| EPOCH 1: Loss = 0.85 | |
| βββββββββββββββββββ | |
| β Batch 1 (16 Samples) | |
| β Forward: Input β HiddenLayers β Output | |
| β Loss: MSE(Ε·, y) | |
| β Backward: βL/βW berechnen | |
| β Update: W := W - Ξ± Γ βL/βW | |
| βββββββββββββββββββ | |
| βββββββββββββββββββ | |
| β Batch 2 (16 Samples) | |
| β (wiederhole) | |
| βββββββββββββββββββ | |
| ... | |
| EPOCH 2: Loss = 0.72 | |
| (Gewichte sind besser, Loss sinkt) | |
| EPOCH 3: Loss = 0.65 | |
| (Netzwerk lernt Muster) | |
| EPOCH 20: Loss = 0.12 β Gut trainiert! β | |
| ``` | |
| --- | |
| ## Gewichte & Bias Learning | |
| ``` | |
| Initiale Gewichte (zufΓ€llig): | |
| βββββββββββββββββββββββ | |
| β W = [0.02, -0.15, β | |
| β 0.08, 0.11, β β Random small values | |
| β -0.03, 0.04] β | |
| βββββββββββββββββββββββ | |
| Nach Training (gelernt): | |
| ββββββββββββββββββββββββ | |
| β W = [1.23, -2.15, β | |
| β 0.89, 1.51, β β GroΓartig angepasst! | |
| β -0.73, 0.94] β Diese Gewichte erkennen | |
| ββββββββββββββββββββββββ jetzt Muster! | |
| ``` | |
| --- | |
| ## Forward & Backward Propagation Fluss | |
| ``` | |
| FORWARD PROPAGATION (Vorhersage): | |
| βββββββββββββββββββββββββββββββββ | |
| x (Input) | |
| β | |
| Layer 1: zβ = xΒ·Wβ + bβ | |
| aβ = ReLU(zβ) | |
| β | |
| Layer 2: zβ = aβΒ·Wβ + bβ | |
| aβ = ReLU(zβ) | |
| β | |
| Layer 3: zβ = aβΒ·Wβ + bβ | |
| Ε· = Sigmoid(zβ) | |
| β | |
| Loss: L = (Ε· - y)Β² | |
| BACKWARD PROPAGATION (Lernen): | |
| ββββββββββββββββββββββββββββββββ | |
| βL/βΕ· (Fehler an Output) | |
| β | |
| βL/βzβ (ΓΌber sigmoid) | |
| βL/βWβ (Grad fΓΌr Wβ) | |
| βL/βbβ (Grad fΓΌr bβ) | |
| β | |
| βL/βaβ (Error rΓΌckwΓ€rts) | |
| βL/βzβ (ΓΌber ReLU) | |
| βL/βWβ (Grad fΓΌr Wβ) | |
| βL/βbβ (Grad fΓΌr bβ) | |
| β | |
| βL/βaβ (Error rΓΌckwΓ€rts) | |
| βL/βzβ (ΓΌber ReLU) | |
| βL/βWβ (Grad fΓΌr Wβ) | |
| βL/βbβ (Grad fΓΌr bβ) | |
| WEIGHT UPDATE: | |
| W := W - learning_rate Γ βL/βW | |
| ``` | |
| --- | |
| ## Integration mit EnhancedMLLearner | |
| ``` | |
| ββββββββββββββββββββββββββββββββββββββββββββββββ | |
| β EnhancedMLLearner β | |
| β β | |
| β Integriert 5 Learning-Module: β | |
| β β | |
| β 1. Context Manager βββββββ β | |
| β 2. Python Analyzer βββββββ€ β | |
| β 3. Google Learner βββββββΌβββ Learning β | |
| β 4. Feedback Learner βββββββ€ Insights β | |
| β 5. Neural Networks βββββββ β | |
| β β β | |
| β π Learning Metrics: β | |
| β - context_awareness: 0.75 β | |
| β - python_quality: 0.82 β | |
| β - web_learning: 0.68 β | |
| β - feedback_quality: 0.91 β | |
| β - neural_network_accuracy: 0.87 β NEW! β | |
| β - overall_improvement: 0.81 β | |
| ββββββββββββββββββββββββββββββββββββββββββββββββ | |
| ``` | |
| --- | |
| ## Datenflusss-Diagramm | |
| ``` | |
| Benutzer Daten | |
| β | |
| βββββββββββββββββββββββ | |
| β Daten Vorbereitung β | |
| β (Normalisierung) β | |
| βββββββββββββββββββββββ | |
| β | |
| βββββββββββββββββββββββ ββββββββββββββββ | |
| β Neural Network β β Gewichte W β | |
| β Training ββββββββ Bias b β | |
| β β β β | |
| β Forward Pass β ββββββββββββββββ | |
| β Backward Pass β | |
| β Weight Update β | |
| βββββββββββββββββββββββ | |
| β | |
| βββββββββββββββββββββββ | |
| β Trainiertes Modell β | |
| β (Gewichte gelernt) β | |
| βββββββββββββββββββββββ | |
| β | |
| βββββββββββββββββββββββ | |
| β Vorhersagen β | |
| β auf neuen Daten β | |
| βββββββββββββββββββββββ | |
| β | |
| Ergebnisse | |
| ``` | |
| --- | |
| ## Aktivierungsfunktionen Visualisierung | |
| ``` | |
| ReLU (Rectified Linear Unit): | |
| ββββββββββββββββββββββββββββ | |
| β / | |
| β / | |
| f(x) β / | |
| β / | |
| β_____ (nur β₯ 0) | |
| βββββΌββββββ x | |
| β | |
| SIGMOID: | |
| ββββββββ | |
| β ___ | |
| f(x) β _/ | |
| β / | |
| ____β__ (zwischen 0-1) | |
| β \ | |
| βββββΌββββββ x | |
| β | |
| TANH: | |
| ββββ | |
| β | |
| f(x) β / ___ | |
| β _/ | |
| βββββΌββ (zwischen -1 to 1) | |
| β \_ | |
| βββββΌββββββ x | |
| ``` | |
| --- | |
| ## Loss-Verlauf wΓ€hrend Training | |
| ``` | |
| Loss | |
| β Epoch 1 | |
| β β² (Hoch - Netzwerk weiΓ noch wenig) | |
| β \ | |
| β \ Epoch 5 | |
| β \βΌ (Sinkt - Lernen findet statt) | |
| β \ | |
| β \ | |
| β \ Epoch 15 (Konvergenz) | |
| β βΌ_____ (Flach - gut trainiert!) | |
| β ββββββββ | |
| βββββββββββββββββββββββββββ Epochs | |
| 0 5 10 15 20 | |
| ``` | |
| --- | |
| ## GrΓΆΓe vs KomplexitΓ€t | |
| ``` | |
| Einfaches Problem: Komplexes Problem: | |
| ββββββββββββββββ βββββββββββββββββ | |
| Input β [Neuron] β Output Input β [32] β[16] β Output | |
| β β | |
| Schnell zu trainieren Langsamer, aber | |
| Weniger Parameter bessere Ergebnisse | |
| Risiko: Underfitting Mehr Parameter | |
| Risiko: Overfitting | |
| ``` | |
| --- | |
| ## Batch Processing Visualisierung | |
| ``` | |
| Training Daten: 100 Samples, Batch Size: 32 | |
| EPOCH 1: | |
| Batch 1: Samples 1-32 β Forward/Backward β Update W | |
| Batch 2: Samples 33-64 β Forward/Backward β Update W | |
| Batch 3: Samples 65-96 β Forward/Backward β Update W | |
| Batch 4: Samples 97-100 β Forward/Backward β Update W | |
| EPOCH 2: | |
| (Wiederhole mit Samples in neuer Reihenfolge) | |
| ``` | |
| --- | |
| ## Konvergenzbeobachtung | |
| ``` | |
| GUTES TRAINING: PROBLEMATISCHES TRAINING: | |
| βββββββββββββββββ ββββββββββββββββββββββ | |
| Loss Loss | |
| β β | |
| β \ β /β²/β² β Oszillation | |
| β \___ β/ββββ (LR zu hoch) | |
| β \___ β | |
| β \___ β βββββ β Stagnation | |
| β ββ β (LR zu niedrig/falsch) | |
| ββββββββββββββββ Epochs β | |
| ββββββββββββ Epochs | |
| ``` | |
| --- | |
| ## Parameter Beziehung | |
| ``` | |
| βββββββββββββββββββ | |
| β Netzwerk GrΓΆΓe β | |
| β (# of Neurons) β | |
| ββββββββββ¬βββββββββ | |
| β | |
| GrΓΆΓer = Komplexer | |
| β | |
| ββββββββββββββββββββΌβββββββββββββββββββ | |
| βΌ βΌ βΌ | |
| LΓ€ngeres Overfitting Bessere | |
| Training Risiko Accuracy | |
| HΓΆher | |
| β | |
| βββββββββββ΄βββββββββ | |
| β Learning Rate β | |
| β (Lerngeschw.) β | |
| ββββββββββ¬βββββββββ | |
| β | |
| Higher = Schneller | |
| β | |
| ββββββββββββββββββββΌβββββββββββββββββββ | |
| βΌ βΌ βΌ | |
| Schneller Oszilation Kann divergieren | |
| Training Risiko | |
| HΓΆher | |
| ``` | |
| --- | |
| ## Metriken Dashboard | |
| ``` | |
| ββββββββββββββββββββββββββββββββββββββββββββββββ | |
| β NEURAL NETWORK STATUS β | |
| β βββββββββββββββββββββββββββββββββββββββββββββββ£ | |
| β β | |
| β Network Name: response_quality β | |
| β Architecture: [12, 8, 4, 1] β | |
| β Status: β TRAINED β | |
| β β | |
| β Metrics: β | |
| β ββ Training Loss: 0.0234 β (Gut!) β | |
| β ββ Test Accuracy: 88.5% β (Gut!) β | |
| β ββ Epochs Trained: 25 β | |
| β ββ Learning Rate: 0.05 β | |
| β ββ Model Size: 142 weights β | |
| β β | |
| β Last Updated: 2026-03-07 14:32:15 β | |
| β β | |
| ββββββββββββββββββββββββββββββββββββββββββββββββ | |
| ``` | |
| --- | |
| ## Checkliste fΓΌr Debugging | |
| ``` | |
| β Training konvergiert nicht | |
| β PrΓΌfe Daten-Normalisierung | |
| β Reduziere Learning Rate | |
| β VergrΓΆΓere Netzwerk | |
| β Overfitting (Test << Training) | |
| β Verkleinere Netzwerk | |
| β Stoppe Training frΓΌher | |
| β Mehr Trainingsdaten | |
| β Underfitting (Test β Training, beide schlecht) | |
| β VergrΓΆΓere Netzwerk | |
| β ErhΓΆhe Learning Rate | |
| β Trainiere lΓ€nger | |
| β Langsames Training | |
| β Kleinere Batch Size | |
| β Vereinfachere Netzwerk-Architektur | |
| β Weniger versteckte Schichten | |
| ``` | |
| --- | |
| **Visualisierungen erstellt:** 7. MΓ€rz 2026 | |
| **Neural Network System:** β VollstΓ€ndig implementiert | |