Spaces:
Sleeping
Sleeping
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,16 +1,14 @@
|
|
| 1 |
-
|
| 2 |
-
title: DCLR
|
| 3 |
-
|
| 4 |
-
colorFrom:
|
| 5 |
colorTo: indigo
|
| 6 |
-
|
| 7 |
-
|
| 8 |
app_file: app.py
|
| 9 |
-
|
| 10 |
-
short_description: 'DCLR leads by a significant margin against LION-ADAM '
|
| 11 |
---
|
| 12 |
|
| 13 |
-
|
| 14 |
# π DCLR Optimizer for CIFAR-10 Image Classification
|
| 15 |
|
| 16 |
This Hugging Face Space showcases the **DCLR (Dynamic Consciousness-based Learning Rate)** optimizer applied to a SimpleCNN model for image classification on the CIFAR-10 dataset.
|
|
@@ -23,8 +21,9 @@ In our analysis, the best-tuned DCLR configuration demonstrated superior test ac
|
|
| 23 |
|
| 24 |
## π‘ How to use the Gradio Demo
|
| 25 |
|
| 26 |
-
**
|
| 27 |
-
**
|
|
|
|
| 28 |
|
| 29 |
Try uploading images of planes, cars, birds, cats, deer, dogs, frogs, horses, ships, or trucks to see how well the model classifies them!
|
| 30 |
|
|
@@ -51,10 +50,19 @@ Our analysis identified the following optimal hyperparameters for DCLR on this t
|
|
| 51 |
|
| 52 |
This best-tuned DCLR configuration achieved a final test accuracy of **70.70%** over 20 epochs, significantly outperforming the original DCLR configuration and other optimizers like Adam and DCLRConscious, and performing competitively with Lion and DCLRAdam.
|
| 53 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
## π Acknowledgments
|
| 56 |
|
| 57 |
-
- The DCLR optimizer is inspired by research into dynamic learning rate adaptation based on
|
| 58 |
- CIFAR-10 dataset is provided by the Canadian Institute for Advanced Research.
|
| 59 |
-
- Gradio and Hugging Face for providing an excellent platform for sharing ML demos.
|
| 60 |
-
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
|
| 1 |
+
|
| 2 |
+
title: DCLR Optimizer for CIFAR-10 Classification
|
| 3 |
+
emojis: ["π", "π§ ", "π"]
|
| 4 |
+
colorFrom: green
|
| 5 |
colorTo: indigo
|
| 6 |
+
SDK: gradio
|
| 7 |
+
SDK_version: 4.19.1
|
| 8 |
app_file: app.py
|
| 9 |
+
human_use: This Space demonstrates the DCLR optimizer for image classification on CIFAR-10.
|
|
|
|
| 10 |
---
|
| 11 |
|
|
|
|
| 12 |
# π DCLR Optimizer for CIFAR-10 Image Classification
|
| 13 |
|
| 14 |
This Hugging Face Space showcases the **DCLR (Dynamic Consciousness-based Learning Rate)** optimizer applied to a SimpleCNN model for image classification on the CIFAR-10 dataset.
|
|
|
|
| 21 |
|
| 22 |
## π‘ How to use the Gradio Demo
|
| 23 |
|
| 24 |
+
1. **Upload an Image**: Drag and drop an image (or click to upload) from your local machine into the designated area in the Gradio interface.
|
| 25 |
+
2. **Webcam/Sketch (Optional)**: If enabled, you might be able to use your webcam or draw an image directly.
|
| 26 |
+
3. **Get Predictions**: The model will automatically process your image and display the top 3 predicted classes for the CIFAR-10 dataset along with their confidence scores.
|
| 27 |
|
| 28 |
Try uploading images of planes, cars, birds, cats, deer, dogs, frogs, horses, ships, or trucks to see how well the model classifies them!
|
| 29 |
|
|
|
|
| 50 |
|
| 51 |
This best-tuned DCLR configuration achieved a final test accuracy of **70.70%** over 20 epochs, significantly outperforming the original DCLR configuration and other optimizers like Adam and DCLRConscious, and performing competitively with Lion and DCLRAdam.
|
| 52 |
|
| 53 |
+
## π Performance Visualizations
|
| 54 |
+
|
| 55 |
+
Here are the performance plots comparing DCLR (tuned) against other optimizers:
|
| 56 |
+
|
| 57 |
+
### Training Performance (Loss and Accuracy over Epochs)
|
| 58 |
+

|
| 59 |
+
|
| 60 |
+
### Final Test Accuracy Comparison
|
| 61 |
+

|
| 62 |
+
|
| 63 |
|
| 64 |
## π Acknowledgments
|
| 65 |
|
| 66 |
+
- The DCLR optimizer is inspired by research into dynamic learning rate adaptation based on information theory.
|
| 67 |
- CIFAR-10 dataset is provided by the Canadian Institute for Advanced Research.
|
| 68 |
+
- Gradio and Hugging Face for providing an excellent platform for sharing ML demos.
|
|
|