Commit ·
7e12ff9
1
Parent(s): 9be9c31
lfs tracking added
Browse files- .gitattributes +38 -0
- .gitignore +10 -0
- assets/style/fabric1.webp +3 -0
- assets/style/fabric_red.jpg +3 -0
- assets/target/cat0.png +3 -0
- assets/target/cat1.png +3 -0
- cnn_style_transfer.py +4 -2
.gitattributes
CHANGED
|
@@ -1,3 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
*.png filter=lfs diff=lfs merge=lfs -text
|
| 2 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 3 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 1 |
+
# Git LFS Tracking Configuration
|
| 2 |
+
|
| 3 |
+
# Blender project files
|
| 4 |
+
*.blend filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
# 3D model formats
|
| 6 |
+
*.fbx filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.obj filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.glb filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.gltf filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.stl filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
# Textures and images
|
| 12 |
+
*.png filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.tga filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.tif filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.tiff filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.exr filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.hdr filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.webp filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
# Video outputs
|
| 22 |
+
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.mov filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.avi filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.mkv filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
# Audio
|
| 27 |
+
*.wav filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.mp3 filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.flac filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
# Simulation / cache data
|
| 31 |
+
*.abc filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.bphys filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.vdb filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
# Archives
|
| 35 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 39 |
*.png filter=lfs diff=lfs merge=lfs -text
|
| 40 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 41 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Python-generated files
|
| 2 |
+
__pycache__/
|
| 3 |
+
*.py[oc]
|
| 4 |
+
build/
|
| 5 |
+
dist/
|
| 6 |
+
wheels/
|
| 7 |
+
*.egg-info
|
| 8 |
+
|
| 9 |
+
# Virtual environments
|
| 10 |
+
.venv
|
assets/style/fabric1.webp
ADDED
|
Git LFS Details
|
assets/style/fabric_red.jpg
ADDED
|
Git LFS Details
|
assets/target/cat0.png
ADDED
|
Git LFS Details
|
assets/target/cat1.png
ADDED
|
Git LFS Details
|
cnn_style_transfer.py
CHANGED
|
@@ -10,7 +10,7 @@ from pathlib import Path
|
|
| 10 |
import gradio as gr
|
| 11 |
from tqdm.auto import tqdm
|
| 12 |
|
| 13 |
-
SIZE = (
|
| 14 |
|
| 15 |
# Device
|
| 16 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
@@ -177,7 +177,7 @@ def tensor_to_image(image_tensor):
|
|
| 177 |
|
| 178 |
|
| 179 |
def style_transfer(style_image, content_image):
|
| 180 |
-
optimized = training_loop(style_image, content_image, num_steps=
|
| 181 |
return tensor_to_image(optimized)
|
| 182 |
|
| 183 |
|
|
@@ -186,6 +186,8 @@ if __name__ == "__main__":
|
|
| 186 |
|
| 187 |
gr.close_all()
|
| 188 |
with gr.Blocks(theme=gr.themes.Glass()) as interface:
|
|
|
|
|
|
|
| 189 |
with gr.Row():
|
| 190 |
with gr.Column():
|
| 191 |
style_image=gr.Image(type="pil", label="Style Image",height=300,width=300)
|
|
|
|
| 10 |
import gradio as gr
|
| 11 |
from tqdm.auto import tqdm
|
| 12 |
|
| 13 |
+
SIZE = (640, 480) #(H, W)
|
| 14 |
|
| 15 |
# Device
|
| 16 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
|
| 177 |
|
| 178 |
|
| 179 |
def style_transfer(style_image, content_image):
|
| 180 |
+
optimized = training_loop(style_image, content_image, num_steps=2)
|
| 181 |
return tensor_to_image(optimized)
|
| 182 |
|
| 183 |
|
|
|
|
| 186 |
|
| 187 |
gr.close_all()
|
| 188 |
with gr.Blocks(theme=gr.themes.Glass()) as interface:
|
| 189 |
+
with gr.Row():
|
| 190 |
+
gr.Markdown("<h2 style='color: blue;'>Vanilla CNN Style Transfer</h2>")
|
| 191 |
with gr.Row():
|
| 192 |
with gr.Column():
|
| 193 |
style_image=gr.Image(type="pil", label="Style Image",height=300,width=300)
|