Spaces:
Sleeping
Sleeping
Update introduction.py
Browse files- introduction.py +3 -3
introduction.py
CHANGED
|
@@ -73,7 +73,7 @@ def Show_introduction():
|
|
| 73 |
if model_choice == "Convolutional Neural Network (CNN)":
|
| 74 |
st.subheader("π Convolutional Neural Network (CNN)")
|
| 75 |
cnn_image = Image.open("Convolutional-Neural-Network.jpg") # Replace with your actual image file
|
| 76 |
-
st.image(cnn_image, caption="Typical CNN architecture"
|
| 77 |
st.markdown("""
|
| 78 |
CNNs are specialized deep learning models for image processing.
|
| 79 |
They consist of layers that automatically learn to detect features like edges, textures, and patterns in images.
|
|
@@ -89,7 +89,7 @@ def Show_introduction():
|
|
| 89 |
elif model_choice == "Vision Transformer (ViT)":
|
| 90 |
st.subheader("π§ Vision Transformer (ViT)")
|
| 91 |
vit_image = Image.open("vit.jpg") # Replace with your actual image file
|
| 92 |
-
st.image(vit_image, caption="Vision Transformer concept"
|
| 93 |
st.markdown("""
|
| 94 |
ViTs bring the power of transformer models to the vision domain by splitting images into patches and processing them using self-attention β a technique originally used in NLP.
|
| 95 |
|
|
@@ -104,7 +104,7 @@ def Show_introduction():
|
|
| 104 |
elif model_choice == "VGG":
|
| 105 |
st.subheader("ποΈ VGG Network")
|
| 106 |
vgg_image = Image.open("new41.jpg") # Replace with your actual image file
|
| 107 |
-
st.image(vgg_image, caption="VGG architecture overview"
|
| 108 |
st.markdown("""
|
| 109 |
The VGG model, introduced by the Visual Geometry Group at Oxford, is known for its deep yet simple architecture using small (3x3) convolution filters.
|
| 110 |
|
|
|
|
| 73 |
if model_choice == "Convolutional Neural Network (CNN)":
|
| 74 |
st.subheader("π Convolutional Neural Network (CNN)")
|
| 75 |
cnn_image = Image.open("Convolutional-Neural-Network.jpg") # Replace with your actual image file
|
| 76 |
+
st.image(cnn_image, caption="Typical CNN architecture")
|
| 77 |
st.markdown("""
|
| 78 |
CNNs are specialized deep learning models for image processing.
|
| 79 |
They consist of layers that automatically learn to detect features like edges, textures, and patterns in images.
|
|
|
|
| 89 |
elif model_choice == "Vision Transformer (ViT)":
|
| 90 |
st.subheader("π§ Vision Transformer (ViT)")
|
| 91 |
vit_image = Image.open("vit.jpg") # Replace with your actual image file
|
| 92 |
+
st.image(vit_image, caption="Vision Transformer concept")
|
| 93 |
st.markdown("""
|
| 94 |
ViTs bring the power of transformer models to the vision domain by splitting images into patches and processing them using self-attention β a technique originally used in NLP.
|
| 95 |
|
|
|
|
| 104 |
elif model_choice == "VGG":
|
| 105 |
st.subheader("ποΈ VGG Network")
|
| 106 |
vgg_image = Image.open("new41.jpg") # Replace with your actual image file
|
| 107 |
+
st.image(vgg_image, caption="VGG architecture overview")
|
| 108 |
st.markdown("""
|
| 109 |
The VGG model, introduced by the Visual Geometry Group at Oxford, is known for its deep yet simple architecture using small (3x3) convolution filters.
|
| 110 |
|