File size: 3,679 Bytes
01942a7 45f43cb 01942a7 45f43cb 01942a7 45f43cb 01942a7 45f43cb 01942a7 45f43cb 01942a7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 | ---
language: en
license: apache-2.0
tags:
- image-classification
- green-ai
- energy-efficiency
- computer-vision
- inceptionv3
- eden-framework
- e2am
- sustainable-ai
datasets:
- imagenet
metrics:
- accuracy
co2_eq_emissions:
emissions: 77.4491
unit: kg
source: Estimated via CodeCarbon (grid factor 0.475 kg CO2e/kWh)
hardware_used: NVIDIA GeForce GTX 1080 Ti
dataset_info:
dataset_size: "~450,000 images – 300 classes (224 px)"
model-index:
- name: EDEN-InceptionV3-Custom-ImageNet300
results:
- task:
type: image-classification
name: Image Classification
dataset:
name: Custom-ImageNet300
type: imagenet
metrics:
- type: accuracy
value: 0.9994
name: Accuracy
- type: f1
value: 0.9994
name: F1 Score
---
# EDEN-InceptionV3-Custom-ImageNet300 — *Baseline*
> **Primary KPI:** EAG (Energy-to-Accuracy Gradient) = `2.0390e-10` ΔAcc/ΔJoules
## Abstract
This model is part of **Project EDEN (Energy-Driven Evolution of Networks)**, implementing the
**E2AM (Energy Efficient Advanced Model)** Framework. The goal is to shift AI benchmarking from
pure accuracy to *Green SOTA* — maximising predictive power per Joule consumed.
**Applied Technique:** Baseline – Standard Full Training (Reference Study)
## Profiling Environment
| Component | Specification |
|---|---|
| **GPU** | NVIDIA GeForce GTX 1080 Ti (11 GB VRAM, 250 W TDP) |
| **CPU** | Intel Xeon W-2125 (4 cores / 8 threads @ 4.00 GHz) |
| **RAM** | 63.66 GB System RAM |
| **OS** | Windows 10 |
| **Dataset** | Custom-ImageNet300 — ~450,000 images – 300 classes (224 px) |
## 🟢 Green Delta Table
*Comparing this model against the reference baseline (ResNet-50 equivalent)*
| Metric | ResNet50 Baseline | **InceptionV3 (EDEN)** | Δ |
|---|---|---|---|
| Accuracy | 0.9573 | **0.9994** | `+4.21%` |
| Total Energy (J) | 380,392,115 | **586,982,974** | `-54.31% saved` |
| CO₂ Emissions (kg) | 50.1906 | **77.4491** | — |
| **EAG Score** | — | **2.0390e-10** | ΔAcc/ΔJoules |
> A **positive EAG** means this model learns more per Joule than the baseline.
> A **negative EAG** indicates a trade-off where higher accuracy required more energy investment.
## E2AM Algorithm — Applied Phases
Standard full fine-tuning used as the **Brute-Force Baseline** for energy comparison. All layers trained from epoch 1 with a fixed learning rate and no gradient accumulation. Included for transparent EAG benchmarking.
## Training Statistics
| Metric | Value |
|---|---|
| Final Accuracy | 0.9994 (99.94%) |
| Total Energy Consumed | 586,982,974 J (163.0508 kWh) |
| Training Time | 32,248 s (8.96 hrs) |
| Estimated CO₂ | 77.4491 kg CO₂e |
| Training Log | `test2\inceptionv3_CustomImageNet300_stats.csv` |
## 📊 Training Visualizations
### Accuracy & Energy over Training
> Green = accuracy (left axis) · Orange dashed = cumulative energy (right axis)

### EAG Metric Trajectory
> EAG = ΔAccuracy / ΔJoules — positive means learning more per Joule than baseline

### Project-Wide Overview
*All EDEN models: energy vs accuracy*

## Cite This Research
```bibtex
@misc{eden2025,
title = {Project EDEN: Energy-Driven Evolution of Networks},
author = {EDEN Research Team},
year = {2025},
note = {Hugging Face: Shanmuk4622},
url = {https://huggingface.co/Shanmuk4622}
}
```
|