Ashrafb commited on
Commit
9b936c1
·
1 Parent(s): c9159f6

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +183 -120
app.py CHANGED
@@ -1,154 +1,217 @@
 
 
 
1
  from share_btn import community_icon_html, loading_icon_html, share_js
2
 
3
- import os, subprocess
4
- import torch
5
 
6
- # (Code for setup and downloading cache files)
 
 
 
7
 
8
- import sys
9
- sys.path.append('src/blip')
10
- sys.path.append('clip-interrogator')
 
 
11
 
12
- import gradio as gr
13
- from clip_interrogator import Config, Interrogator
14
- import io
15
- from PIL import Image
16
- config = Config()
17
- config.device = 'cuda' if torch.cuda.is_available() else 'cpu'
18
- config.blip_offload = False if torch.cuda.is_available() else True
19
- config.chunk_size = 2048
20
- config.flavor_intermediate_count = 512
21
- config.blip_num_beams = 64
22
-
23
- ci = Interrogator(config)
24
-
25
- def inference(input_images, mode, best_max_flavors):
26
- # Process each image in the list and generate prompt results
27
- prompt_texts = []
28
- for image_bytes in input_images:
29
- image = Image.open(io.BytesIO(image_bytes)).convert('RGB')
30
- if mode == 'best':
31
- prompt_result = ci.interrogate(image, max_flavors=int(best_max_flavors))
32
- elif mode == 'classic':
33
- prompt_result = ci.interrogate_classic(image)
34
- else:
35
- prompt_result = ci.interrogate_fast(image)
36
- prompt_texts.append(prompt_result['prompt']) # Extract the prompt text
37
- return prompt_texts
38
-
39
- title = """
40
- <div style="text-align: center; max-width: 500px; margin: 0 auto;">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  <div
42
  style="
43
  display: inline-flex;
44
  align-items: center;
45
  gap: 0.8rem;
46
  font-size: 1.75rem;
47
- margin-bottom: 10px;
48
  "
49
  >
50
- <h1 style="font-weight: 600; margin-bottom: 7px;">
51
- CLIP Interrogator 2.1
52
  </h1>
53
  </div>
54
- <p style="margin-bottom: 10px;font-size: 94%;font-weight: 100;line-height: 1.5em;">
55
- Want to figure out what a good prompt might be to create new images like an existing one?
56
- <br />The CLIP Interrogator is here to get you answers!
57
- <br />This version is specialized for producing nice prompts for use with Stable Diffusion 2.0 using the ViT-H-14 OpenCLIP model!
58
  </p>
 
59
  </div>
60
  """
61
 
62
- article = """
63
- <div style="text-align: center; max-width: 500px; margin: 0 auto;font-size: 94%;">
64
-
 
 
 
 
 
65
  <p>
66
- Server busy? You can also run on <a href="https://colab.research.google.com/github/pharmapsychotic/clip-interrogator/blob/open-clip/clip_interrogator.ipynb">Google Colab</a>
67
  </p>
 
68
  <p>
69
- Has this been helpful to you? Follow Pharma on twitter
70
- <a href="https://twitter.com/pharmapsychotic">@pharmapsychotic</a>
71
- and check out more tools at his
72
  <a href="https://pharmapsychotic.com/tools.html">Ai generative art tools list</a>
73
  </p>
74
  </div>
75
  """
76
 
77
- css = '''
78
- #col-container {width: width: 80%;; margin-left: auto; margin-right: auto;}
79
- a {text-decoration-line: underline; font-weight: 600;}
80
- .animate-spin {
81
- animation: spin 1s linear infinite;
82
- }
83
- @keyframes spin {
84
- from {
85
- transform: rotate(0deg);
86
  }
87
- to {
88
- transform: rotate(360deg);
 
89
  }
90
- }
91
- #share-btn-container {
92
- display: flex; padding-left: 0.5rem !important; padding-right: 0.5rem !important; background-color: #000000; justify-content: center; align-items: center; border-radius: 9999px !important; width: 13rem;
93
- }
94
- #share-btn {
95
- all: initial; color: #ffffff;font-weight: 600; cursor:pointer; font-family: 'IBM Plex Sans', sans-serif; margin-left: 0.5rem !important; padding-top: 0.25rem !important; padding-bottom: 0.25rem !important;
96
- }
97
- #share-btn * {
98
- all: unset;
99
- }
100
- #share-btn-container div:nth-child(-n+2){
101
- width: auto !important;
102
- min-height: 0px !important;
103
- }
104
- #share-btn-container .wrap {
105
- display: none !important;
106
- }
107
- #gallery .caption-label {
108
- font-size: 15px !important;
109
- right: 0 !important;
110
- max-width: 100% !important;
111
- text-overflow: clip !important;
112
- white-space: normal !important;
113
- overflow: auto !important;
114
- height: 20% !important;
115
- }
116
-
117
- #gallery .caption {
118
- padding: var(--size-2) var(--size-3) !important;
119
- text-overflow: clip !important;
120
- white-space: normal !important; /* Allows the text to wrap */
121
- color: var(--block-label-text-color) !important;
122
- font-weight: var(--weight-semibold) !important;
123
- text-align: center !important;
124
- height: 100% !important;
125
- font-size: 17px !important;
126
- }
127
- '''
128
-
129
- with gr.Blocks(css=css) as block:
130
- with gr.Column(elem_id="col-container"):
131
- gr.HTML(title)
132
 
133
- input_image = gr.Files(label = "Inputs", file_count="multiple", type='bytes', elem_id='inputs')
 
134
  with gr.Row():
135
- mode_input = gr.Radio(['best', 'classic', 'fast'], label='Select mode', value='best')
136
- flavor_input = gr.Slider(minimum=2, maximum=24, step=2, value=4, label='best mode max flavors')
137
-
138
- submit_btn = gr.Button("Submit")
139
-
140
- # Display the prompt texts
141
- gallery = gr.HTML(
142
- src="{{' '.join(outputs['prompt_texts'])}}",
143
- type="div",
144
- label="Outputs",
145
- elem_id="gallery"
146
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
147
 
148
- with gr.Group(elem_id="share-btn-container"):
149
- loading_icon = gr.HTML(loading_icon_html, visible=False)
 
 
 
 
 
 
 
 
 
 
 
 
 
150
 
151
- gr.HTML(article)
152
- submit_btn.click(fn=inference, inputs=[input_image,mode_input,flavor_input], outputs=[gallery], api_name="clipi2")
 
 
 
 
153
 
154
- block.queue(max_size=32,concurrency_count=10).launch(show_api=False)
 
1
+ #!/usr/bin/env python3
2
+ import gradio as gr
3
+ from clip_interrogator import Config, Interrogator
4
  from share_btn import community_icon_html, loading_icon_html, share_js
5
 
6
+ MODELS = ['ViT-L (best for Stable Diffusion 1.*)']#, 'ViT-H (best for Stable Diffusion 2.*)']
 
7
 
8
+ # load BLIP and ViT-L https://huggingface.co/openai/clip-vit-large-patch14
9
+ config = Config(clip_model_name="ViT-L-14/openai")
10
+ ci_vitl = Interrogator(config)
11
+ # ci_vitl.clip_model = ci_vitl.clip_model.to("cpu")
12
 
13
+ # load ViT-H https://huggingface.co/laion/CLIP-ViT-H-14-laion2B-s32B-b79K
14
+ # config.blip_model = ci_vitl.blip_model
15
+ # config.clip_model_name = "ViT-H-14/laion2b_s32b_b79k"
16
+ # ci_vith = Interrogator(config)
17
+ # ci_vith.clip_model = ci_vith.clip_model.to("cpu")
18
 
19
+
20
+ def image_analysis(image, clip_model_name):
21
+ # move selected model to GPU and other model to CPU
22
+ # if clip_model_name == MODELS[0]:
23
+ # ci_vith.clip_model = ci_vith.clip_model.to("cpu")
24
+ # ci_vitl.clip_model = ci_vitl.clip_model.to(ci_vitl.device)
25
+ # ci = ci_vitl
26
+ # else:
27
+ # ci_vitl.clip_model = ci_vitl.clip_model.to("cpu")
28
+ # ci_vith.clip_model = ci_vith.clip_model.to(ci_vith.device)
29
+ # ci = ci_vith
30
+ ci = ci_vitl
31
+
32
+ image = image.convert('RGB')
33
+ image_features = ci.image_to_features(image)
34
+
35
+ top_mediums = ci.mediums.rank(image_features, 5)
36
+ top_artists = ci.artists.rank(image_features, 5)
37
+ top_movements = ci.movements.rank(image_features, 5)
38
+ top_trendings = ci.trendings.rank(image_features, 5)
39
+ top_flavors = ci.flavors.rank(image_features, 5)
40
+
41
+ medium_ranks = {medium: sim for medium, sim in zip(top_mediums, ci.similarities(image_features, top_mediums))}
42
+ artist_ranks = {artist: sim for artist, sim in zip(top_artists, ci.similarities(image_features, top_artists))}
43
+ movement_ranks = {movement: sim for movement, sim in zip(top_movements, ci.similarities(image_features, top_movements))}
44
+ trending_ranks = {trending: sim for trending, sim in zip(top_trendings, ci.similarities(image_features, top_trendings))}
45
+ flavor_ranks = {flavor: sim for flavor, sim in zip(top_flavors, ci.similarities(image_features, top_flavors))}
46
+
47
+ return medium_ranks, artist_ranks, movement_ranks, trending_ranks, flavor_ranks
48
+
49
+
50
+ def image_to_prompt(image, clip_model_name, mode):
51
+ # move selected model to GPU and other model to CPU
52
+ # if clip_model_name == MODELS[0]:
53
+ # ci_vith.clip_model = ci_vith.clip_model.to("cpu")
54
+ # ci_vitl.clip_model = ci_vitl.clip_model.to(ci_vitl.device)
55
+ # ci = ci_vitl
56
+ # else:
57
+ # ci_vitl.clip_model = ci_vitl.clip_model.to("cpu")
58
+ # ci_vith.clip_model = ci_vith.clip_model.to(ci_vith.device)
59
+ # ci = ci_vith
60
+ ci = ci_vitl
61
+
62
+ ci.config.blip_num_beams = 64
63
+ ci.config.chunk_size = 2048
64
+ ci.config.flavor_intermediate_count = 2048 if clip_model_name == MODELS[0] else 1024
65
+
66
+ image = image.convert('RGB')
67
+ if mode == 'best':
68
+ prompt = ci.interrogate(image)
69
+ elif mode == 'classic':
70
+ prompt = ci.interrogate_classic(image)
71
+ elif mode == 'fast':
72
+ prompt = ci.interrogate_fast(image)
73
+ elif mode == 'negative':
74
+ prompt = ci.interrogate_negative(image)
75
+
76
+ return prompt, gr.update(visible=True), gr.update(visible=True), gr.update(visible=True)
77
+
78
+
79
+ TITLE = """
80
+ <div style="text-align: center; max-width: 650px; margin: 0 auto;">
81
  <div
82
  style="
83
  display: inline-flex;
84
  align-items: center;
85
  gap: 0.8rem;
86
  font-size: 1.75rem;
 
87
  "
88
  >
89
+ <h1 style="font-weight: 900; margin-bottom: 7px;">
90
+ CLIP Interrogator
91
  </h1>
92
  </div>
93
+ <p style="margin-bottom: 10px; font-size: 94%">
94
+ Want to figure out what a good prompt might be to create new images like an existing one?<br>The CLIP Interrogator is here to get you answers!
 
 
95
  </p>
96
+ <p>You can skip the queue by duplicating this space and upgrading to gpu in settings: <a style='display:inline-block' href='https://huggingface.co/spaces/pharmapsychotic/CLIP-Interrogator?duplicate=true'><img src='https://img.shields.io/badge/-Duplicate%20Space-blue?labelColor=white&style=flat&logo=data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAABAAAAAQCAYAAAAf8/9hAAAAAXNSR0IArs4c6QAAAP5JREFUOE+lk7FqAkEURY+ltunEgFXS2sZGIbXfEPdLlnxJyDdYB62sbbUKpLbVNhyYFzbrrA74YJlh9r079973psed0cvUD4A+4HoCjsA85X0Dfn/RBLBgBDxnQPfAEJgBY+A9gALA4tcbamSzS4xq4FOQAJgCDwV2CPKV8tZAJcAjMMkUe1vX+U+SMhfAJEHasQIWmXNN3abzDwHUrgcRGmYcgKe0bxrblHEB4E/pndMazNpSZGcsZdBlYJcEL9Afo75molJyM2FxmPgmgPqlWNLGfwZGG6UiyEvLzHYDmoPkDDiNm9JR9uboiONcBXrpY1qmgs21x1QwyZcpvxt9NS09PlsPAAAAAElFTkSuQmCC&logoWidth=14' alt='Duplicate Space'></a></p>
97
  </div>
98
  """
99
 
100
+ ARTICLE = """
101
+ <div style="text-align: center; max-width: 650px; margin: 0 auto;">
102
+ <p>
103
+ Example art by <a href="https://pixabay.com/illustrations/watercolour-painting-art-effect-4799014/">Layers</a>
104
+ and <a href="https://pixabay.com/illustrations/animal-painting-cat-feline-pet-7154059/">Lin Tong</a>
105
+ from pixabay.com
106
+ </p>
107
+
108
  <p>
109
+ Server busy? You can also run on <a href="https://colab.research.google.com/github/pharmapsychotic/clip-interrogator/blob/main/clip_interrogator.ipynb">Google Colab</a>
110
  </p>
111
+
112
  <p>
113
+ Has this been helpful to you? Follow me on twitter
114
+ <a href="https://twitter.com/pharmapsychotic">@pharmapsychotic</a><br>
115
+ and check out more tools at my
116
  <a href="https://pharmapsychotic.com/tools.html">Ai generative art tools list</a>
117
  </p>
118
  </div>
119
  """
120
 
121
+ CSS = """
122
+ #col-container {margin-left: auto; margin-right: auto;}
123
+ a {text-decoration-line: underline; font-weight: 600;}
124
+ .animate-spin {
125
+ animation: spin 1s linear infinite;
 
 
 
 
126
  }
127
+ @keyframes spin {
128
+ from { transform: rotate(0deg); }
129
+ to { transform: rotate(360deg); }
130
  }
131
+ #share-btn-container {
132
+ display: flex; padding-left: 0.5rem !important; padding-right: 0.5rem !important; background-color: #000000; justify-content: center; align-items: center; border-radius: 9999px !important; width: 13rem;
133
+ }
134
+ #share-btn {
135
+ all: initial; color: #ffffff;font-weight: 600; cursor:pointer; font-family: 'IBM Plex Sans', sans-serif; margin-left: 0.5rem !important; padding-top: 0.25rem !important; padding-bottom: 0.25rem !important;
136
+ }
137
+ #share-btn * {
138
+ all: unset;
139
+ }
140
+ #share-btn-container div:nth-child(-n+2){
141
+ width: auto !important;
142
+ min-height: 0px !important;
143
+ }
144
+ #share-btn-container .wrap {
145
+ display: none !important;
146
+ }
147
+ """
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148
 
149
+ def analyze_tab():
150
+ with gr.Column():
151
  with gr.Row():
152
+ image = gr.Image(type='pil', label="Image")
153
+ model = gr.Dropdown(MODELS, value=MODELS[0], label='CLIP Model')
154
+ with gr.Row():
155
+ medium = gr.Label(label="Medium", num_top_classes=5)
156
+ artist = gr.Label(label="Artist", num_top_classes=5)
157
+ movement = gr.Label(label="Movement", num_top_classes=5)
158
+ trending = gr.Label(label="Trending", num_top_classes=5)
159
+ flavor = gr.Label(label="Flavor", num_top_classes=5)
160
+
161
+ button = gr.Button("Analyze", api_name="image-analysis")
162
+ button.click(image_analysis, inputs=[image, model], outputs=[medium, artist, movement, trending, flavor])
163
+
164
+ examples=[['example01.jpg', MODELS[0]], ['example02.jpg', MODELS[0]]]
165
+ ex = gr.Examples(
166
+ examples=examples,
167
+ fn=image_analysis,
168
+ inputs=[input_image, input_model],
169
+ outputs=[medium, artist, movement, trending, flavor],
170
+ cache_examples=True,
171
+ run_on_click=True
172
+ )
173
+ ex.dataset.headers = [""]
174
+
175
+
176
+ with gr.Blocks(css=CSS) as block:
177
+ with gr.Column(elem_id="col-container"):
178
+ gr.HTML(TITLE)
179
+
180
+ with gr.Tab("Prompt"):
181
+ with gr.Row():
182
+ input_image = gr.Image(type='pil', elem_id="input-img")
183
+ with gr.Column():
184
+ input_model = gr.Dropdown(MODELS, value=MODELS[0], label='CLIP Model')
185
+ input_mode = gr.Radio(['best', 'fast', 'classic', 'negative'], value='best', label='Mode')
186
+ submit_btn = gr.Button("Submit", api_name="image-to-prompt")
187
+ output_text = gr.Textbox(label="Output", elem_id="output-txt")
188
+
189
+ with gr.Group(elem_id="share-btn-container"):
190
+ community_icon = gr.HTML(community_icon_html, visible=False)
191
+ loading_icon = gr.HTML(loading_icon_html, visible=False)
192
+ share_button = gr.Button("Share to community", elem_id="share-btn", visible=False)
193
 
194
+ examples=[['example01.jpg', MODELS[0], 'best'], ['example02.jpg', MODELS[0], 'best']]
195
+ ex = gr.Examples(
196
+ examples=examples,
197
+ fn=image_to_prompt,
198
+ inputs=[input_image, input_model, input_mode],
199
+ outputs=[output_text, share_button, community_icon, loading_icon],
200
+ cache_examples=True,
201
+ run_on_click=True
202
+ )
203
+ ex.dataset.headers = [""]
204
+
205
+ with gr.Tab("Analyze"):
206
+ analyze_tab()
207
+
208
+ gr.HTML(ARTICLE)
209
 
210
+ submit_btn.click(
211
+ fn=image_to_prompt,
212
+ inputs=[input_image, input_model, input_mode],
213
+ outputs=[output_text, share_button, community_icon, loading_icon]
214
+ )
215
+ share_button.click(None, [], [], _js=share_js)
216
 
217
+ block.queue(max_size=32).launch(show_api=False)