Spaces:
Configuration error
Configuration error
Sreerama
commited on
Commit
·
a00d107
1
Parent(s):
dc54f82
remove swap_text function
Browse files
app.py
CHANGED
|
@@ -40,28 +40,6 @@ if(is_gpu_associated):
|
|
| 40 |
with zipfile.ZipFile("mix.zip", 'r') as zip_ref:
|
| 41 |
zip_ref.extractall(".")
|
| 42 |
|
| 43 |
-
def swap_text(option, base):
|
| 44 |
-
resize_width = 768 if base == "v2-768" else 512
|
| 45 |
-
mandatory_liability = "You must have the right to do so and you are liable for the images you use, example:"
|
| 46 |
-
if(option == "object"):
|
| 47 |
-
instance_prompt_example = "cttoy"
|
| 48 |
-
freeze_for = 30
|
| 49 |
-
return [f"You are going to train `object`(s), upload 5-10 images of each object you are planning on training on from different angles/perspectives. You can use services like <a style='text-decoration: underline' target='_blank' href='https://www.birme.net/?target_width={resize_width}&target_height={resize_width}'>birme</a> for smart cropping. {mandatory_liability}:", '''<img src="file/cat-toy.png" />''', f"You should name your concept with a unique made up word that has low chance of the model already knowing it (e.g.: `{instance_prompt_example}` here). Images will be automatically cropped to {resize_width}x{resize_width}.", freeze_for, gr.update(visible=False)]
|
| 50 |
-
elif(option == "person"):
|
| 51 |
-
instance_prompt_example = "julcto"
|
| 52 |
-
freeze_for = 70
|
| 53 |
-
#show_prior_preservation = True if base != "v2-768" else False
|
| 54 |
-
show_prior_preservation=False
|
| 55 |
-
if(show_prior_preservation):
|
| 56 |
-
prior_preservation_box_update = gr.update(visible=show_prior_preservation)
|
| 57 |
-
else:
|
| 58 |
-
prior_preservation_box_update = gr.update(visible=show_prior_preservation, value=False)
|
| 59 |
-
return [f"You are going to train a `person`(s), upload 10-20 images of each person you are planning on training on from different angles/perspectives. You can use services like <a style='text-decoration: underline' target='_blank' href='https://www.birme.net/?target_width={resize_width}&target_height={resize_width}'>birme</a> for smart cropping. {mandatory_liability}:", '''<img src="file/person.png" />''', f"You should name your concept with a unique made up word that has low chance of the model already knowing it (e.g.: `{instance_prompt_example}` here). Images will be automatically cropped to {resize_width}x{resize_width}.", freeze_for, prior_preservation_box_update]
|
| 60 |
-
elif(option == "style"):
|
| 61 |
-
instance_prompt_example = "trsldamrl"
|
| 62 |
-
freeze_for = 10
|
| 63 |
-
return [f"You are going to train a `style`, upload 10-20 images of the style you are planning on training on. You can use services like <a style='text-decoration: underline' target='_blank' href='https://www.birme.net/?target_width={resize_width}&target_height={resize_width}'>birme</a> for smart cropping. Name the files with the words you would like {mandatory_liability}:", '''<img src="file/trsl_style.png" />''', f"You should name your concept with a unique made up word that has low chance of the model already knowing it (e.g.: `{instance_prompt_example}` here). Images will be automatically cropped to {resize_width}x{resize_width}", freeze_for, gr.update(visible=False)]
|
| 64 |
-
|
| 65 |
def swap_base_model(selected_model):
|
| 66 |
if(is_gpu_associated):
|
| 67 |
global model_to_load
|
|
@@ -606,7 +584,6 @@ with gr.Blocks(css=css) as demo:
|
|
| 606 |
#Swap the examples and the % of text encoder trained depending if it is an object, person or style
|
| 607 |
|
| 608 |
#Swap the base model
|
| 609 |
-
base_model_to_use.change(fn=swap_text, inputs=[base_model_to_use], outputs=[thing_description, thing_image_example, things_naming, perc_txt_encoder, thing_experimental], queue=False, show_progress=False)
|
| 610 |
base_model_to_use.change(fn=swap_base_model, inputs=base_model_to_use, outputs=[])
|
| 611 |
|
| 612 |
#Update the summary box below the UI according to how many images are uploaded and whether users are using custom settings or not
|
|
|
|
| 40 |
with zipfile.ZipFile("mix.zip", 'r') as zip_ref:
|
| 41 |
zip_ref.extractall(".")
|
| 42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 43 |
def swap_base_model(selected_model):
|
| 44 |
if(is_gpu_associated):
|
| 45 |
global model_to_load
|
|
|
|
| 584 |
#Swap the examples and the % of text encoder trained depending if it is an object, person or style
|
| 585 |
|
| 586 |
#Swap the base model
|
|
|
|
| 587 |
base_model_to_use.change(fn=swap_base_model, inputs=base_model_to_use, outputs=[])
|
| 588 |
|
| 589 |
#Update the summary box below the UI according to how many images are uploaded and whether users are using custom settings or not
|