Add files using upload-large-folder tool
Browse files- config.yaml +117 -0
- loss_history.json +207 -0
- optimizer.pt +3 -0
- qwen_2509_object_removal_new_000000500.safetensors +3 -0
- qwen_2509_object_removal_new_000000550.safetensors +3 -0
- qwen_2509_object_removal_new_000000600.safetensors +3 -0
- qwen_2509_object_removal_new_000000650.safetensors +3 -0
- samples/1765913168168__000000050_0.jpg +0 -0
- samples/1765913250652__000000050_1.jpg +0 -0
- samples/1765913333097__000000050_2.jpg +0 -0
- samples/1765916820855__000000100_0.jpg +0 -0
- samples/1765916919610__000000100_1.jpg +0 -0
- samples/1765917018655__000000100_2.jpg +0 -0
- samples/1765920526538__000000150_0.jpg +0 -0
- samples/1765920626672__000000150_1.jpg +0 -0
- samples/1765920727033__000000150_2.jpg +0 -0
- samples/1765924219383__000000200_0.jpg +0 -0
- samples/1765924301805__000000200_1.jpg +0 -0
- samples/1765924384230__000000200_2.jpg +0 -0
- samples/1765927857787__000000250_0.jpg +0 -0
- samples/1765927940222__000000250_1.jpg +0 -0
- samples/1765928022642__000000250_2.jpg +0 -0
- samples/1765931497363__000000300_0.jpg +0 -0
- samples/1765931579810__000000300_1.jpg +0 -0
- samples/1765931662381__000000300_2.jpg +0 -0
- samples/1765935140019__000000350_0.jpg +0 -0
- samples/1765935222456__000000350_1.jpg +0 -0
- samples/1765935305166__000000350_2.jpg +0 -0
- samples/1765938782154__000000400_0.jpg +0 -0
- samples/1765938864681__000000400_1.jpg +0 -0
- samples/1765938947255__000000400_2.jpg +0 -0
- samples/1765942436028__000000450_0.jpg +0 -0
- samples/1765942518822__000000450_1.jpg +0 -0
- samples/1765942601365__000000450_2.jpg +0 -0
- samples/1765952484578__000000500_0.jpg +0 -0
- samples/1765952567599__000000500_1.jpg +0 -0
- samples/1765952650419__000000500_2.jpg +0 -0
- samples/1765956118170__000000550_0.jpg +0 -0
- samples/1765956201084__000000550_1.jpg +0 -0
- samples/1765956284101__000000550_2.jpg +0 -0
- samples/1765959752173__000000600_0.jpg +0 -0
- samples/1765959835054__000000600_1.jpg +0 -0
- samples/1765959917790__000000600_2.jpg +0 -0
- samples/1765963406167__000000650_0.jpg +0 -0
- samples/1765963489019__000000650_1.jpg +0 -0
- samples/1765963571855__000000650_2.jpg +0 -0
config.yaml
ADDED
|
@@ -0,0 +1,117 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
job: extension
|
| 2 |
+
config:
|
| 3 |
+
name: qwen_2509_object_removal_new
|
| 4 |
+
process:
|
| 5 |
+
- type: diffusion_trainer
|
| 6 |
+
training_folder: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/output
|
| 7 |
+
sqlite_db_path: ./aitk_db.db
|
| 8 |
+
device: cuda
|
| 9 |
+
trigger_word: null
|
| 10 |
+
performance_log_every: 10
|
| 11 |
+
network:
|
| 12 |
+
type: lora
|
| 13 |
+
linear: 16
|
| 14 |
+
linear_alpha: 16
|
| 15 |
+
conv: 16
|
| 16 |
+
conv_alpha: 16
|
| 17 |
+
lokr_full_rank: true
|
| 18 |
+
lokr_factor: -1
|
| 19 |
+
network_kwargs:
|
| 20 |
+
ignore_if_contains: []
|
| 21 |
+
save:
|
| 22 |
+
dtype: bf16
|
| 23 |
+
save_every: 50
|
| 24 |
+
max_step_saves_to_keep: 4
|
| 25 |
+
save_format: diffusers
|
| 26 |
+
push_to_hub: false
|
| 27 |
+
datasets:
|
| 28 |
+
- folder_path: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_train/target
|
| 29 |
+
mask_path: null
|
| 30 |
+
mask_min_value: 0.1
|
| 31 |
+
default_caption: Remove Object
|
| 32 |
+
caption_ext: txt
|
| 33 |
+
caption_dropout_rate: 0.05
|
| 34 |
+
cache_latents_to_disk: true
|
| 35 |
+
is_reg: false
|
| 36 |
+
network_weight: 1
|
| 37 |
+
resolution:
|
| 38 |
+
- 512
|
| 39 |
+
controls: []
|
| 40 |
+
shrink_video_to_frames: true
|
| 41 |
+
num_frames: 1
|
| 42 |
+
do_i2v: true
|
| 43 |
+
flip_x: false
|
| 44 |
+
flip_y: false
|
| 45 |
+
control_path_1: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_train/mask_on_background
|
| 46 |
+
control_path_2: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_train/mask
|
| 47 |
+
control_path_3: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_train/mae_output
|
| 48 |
+
train:
|
| 49 |
+
batch_size: 6
|
| 50 |
+
bypass_guidance_embedding: false
|
| 51 |
+
steps: 5500
|
| 52 |
+
gradient_accumulation: 1
|
| 53 |
+
train_unet: true
|
| 54 |
+
train_text_encoder: false
|
| 55 |
+
gradient_checkpointing: true
|
| 56 |
+
noise_scheduler: flowmatch
|
| 57 |
+
optimizer: adamw
|
| 58 |
+
timestep_type: weighted
|
| 59 |
+
content_or_style: balanced
|
| 60 |
+
optimizer_params:
|
| 61 |
+
weight_decay: 0.0001
|
| 62 |
+
unload_text_encoder: false
|
| 63 |
+
cache_text_embeddings: true
|
| 64 |
+
lr: 0.0005
|
| 65 |
+
ema_config:
|
| 66 |
+
use_ema: false
|
| 67 |
+
ema_decay: 0.99
|
| 68 |
+
skip_first_sample: true
|
| 69 |
+
force_first_sample: false
|
| 70 |
+
disable_sampling: false
|
| 71 |
+
dtype: bf16
|
| 72 |
+
diff_output_preservation: false
|
| 73 |
+
diff_output_preservation_multiplier: 1
|
| 74 |
+
diff_output_preservation_class: person
|
| 75 |
+
switch_boundary_every: 1
|
| 76 |
+
loss_type: mse
|
| 77 |
+
model:
|
| 78 |
+
name_or_path: Qwen/Qwen-Image-Edit-2509
|
| 79 |
+
quantize: true
|
| 80 |
+
qtype: uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors
|
| 81 |
+
quantize_te: true
|
| 82 |
+
qtype_te: uint4
|
| 83 |
+
arch: qwen_image_edit_plus
|
| 84 |
+
low_vram: true
|
| 85 |
+
model_kwargs:
|
| 86 |
+
match_target_res: false
|
| 87 |
+
layer_offloading: true
|
| 88 |
+
layer_offloading_text_encoder_percent: 0
|
| 89 |
+
layer_offloading_transformer_percent: 1
|
| 90 |
+
sample:
|
| 91 |
+
sampler: flowmatch
|
| 92 |
+
sample_every: 50
|
| 93 |
+
width: 512
|
| 94 |
+
height: 512
|
| 95 |
+
samples:
|
| 96 |
+
- prompt: Remove Object
|
| 97 |
+
ctrl_img_1: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask_on_background/17.jpg
|
| 98 |
+
ctrl_img_2: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask/17.jpg
|
| 99 |
+
ctrl_img_3: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mae_output/17.jpg
|
| 100 |
+
- prompt: Remove Object
|
| 101 |
+
ctrl_img_1: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask_on_background/45.jpg
|
| 102 |
+
ctrl_img_2: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask/45.jpg
|
| 103 |
+
ctrl_img_3: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mae_output/45.jpg
|
| 104 |
+
- prompt: Remove Object
|
| 105 |
+
ctrl_img_1: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask_on_background/49.jpg
|
| 106 |
+
ctrl_img_2: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mask/49.jpg
|
| 107 |
+
ctrl_img_3: /home/nguyenvanthanhdat1810@gmail.com/ai-toolkit/datasets/object_removal_val/mae_output/49.jpg
|
| 108 |
+
neg: ''
|
| 109 |
+
seed: 42
|
| 110 |
+
walk_seed: true
|
| 111 |
+
guidance_scale: 4
|
| 112 |
+
sample_steps: 10
|
| 113 |
+
num_frames: 1
|
| 114 |
+
fps: 1
|
| 115 |
+
meta:
|
| 116 |
+
name: qwen_2509_object_removal_new
|
| 117 |
+
version: '1.0'
|
loss_history.json
ADDED
|
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"step": 650,
|
| 3 |
+
"epoch": 0,
|
| 4 |
+
"loss_history": [
|
| 5 |
+
0.051893532276153564,
|
| 6 |
+
0.04191002994775772,
|
| 7 |
+
0.03563995659351349,
|
| 8 |
+
0.05188479274511337,
|
| 9 |
+
0.07136755436658859,
|
| 10 |
+
0.0493043027818203,
|
| 11 |
+
0.06096412241458893,
|
| 12 |
+
0.05343471094965935,
|
| 13 |
+
0.05279016122221947,
|
| 14 |
+
0.05165542662143707,
|
| 15 |
+
0.05911332368850708,
|
| 16 |
+
0.046261124312877655,
|
| 17 |
+
0.05145987123250961,
|
| 18 |
+
0.07348300516605377,
|
| 19 |
+
0.06282848119735718,
|
| 20 |
+
0.04710949584841728,
|
| 21 |
+
0.07371529191732407,
|
| 22 |
+
0.04425327107310295,
|
| 23 |
+
0.05758678540587425,
|
| 24 |
+
0.03565998002886772,
|
| 25 |
+
0.054090213030576706,
|
| 26 |
+
0.02724049612879753,
|
| 27 |
+
0.05941595137119293,
|
| 28 |
+
0.06052125617861748,
|
| 29 |
+
0.033973388373851776,
|
| 30 |
+
0.02366490662097931,
|
| 31 |
+
0.03215572610497475,
|
| 32 |
+
0.056784581393003464,
|
| 33 |
+
0.054403841495513916,
|
| 34 |
+
0.031740397214889526,
|
| 35 |
+
0.048682890832424164,
|
| 36 |
+
0.0359157919883728,
|
| 37 |
+
0.03429361432790756,
|
| 38 |
+
0.05047951638698578,
|
| 39 |
+
0.04810015484690666,
|
| 40 |
+
0.05415612459182739,
|
| 41 |
+
0.045266665518283844,
|
| 42 |
+
0.07389190793037415,
|
| 43 |
+
0.07689608633518219,
|
| 44 |
+
0.01617683656513691,
|
| 45 |
+
0.027895715087652206,
|
| 46 |
+
0.04273153468966484,
|
| 47 |
+
0.07595288753509521,
|
| 48 |
+
0.04460812360048294,
|
| 49 |
+
0.06804212182760239,
|
| 50 |
+
0.043365802615880966,
|
| 51 |
+
0.06288819760084152,
|
| 52 |
+
0.053755879402160645,
|
| 53 |
+
0.04600618779659271,
|
| 54 |
+
0.04654350131750107,
|
| 55 |
+
0.04472731798887253,
|
| 56 |
+
0.04098866134881973,
|
| 57 |
+
0.05165230855345726,
|
| 58 |
+
0.046722885221242905,
|
| 59 |
+
0.061621394008398056,
|
| 60 |
+
0.06209284067153931,
|
| 61 |
+
0.028675343841314316,
|
| 62 |
+
0.02694562077522278,
|
| 63 |
+
0.04801797494292259,
|
| 64 |
+
0.07362907379865646,
|
| 65 |
+
0.032962650060653687,
|
| 66 |
+
0.06702133268117905,
|
| 67 |
+
0.040867991745471954,
|
| 68 |
+
0.09054657071828842,
|
| 69 |
+
0.07557031512260437,
|
| 70 |
+
0.05606820434331894,
|
| 71 |
+
0.05454063415527344,
|
| 72 |
+
0.05563917011022568,
|
| 73 |
+
0.031462959945201874,
|
| 74 |
+
0.026764918118715286,
|
| 75 |
+
0.07913929224014282,
|
| 76 |
+
0.03800187259912491,
|
| 77 |
+
0.06417255848646164,
|
| 78 |
+
0.03501094505190849,
|
| 79 |
+
0.019893253222107887,
|
| 80 |
+
0.07155022770166397,
|
| 81 |
+
0.07499650865793228,
|
| 82 |
+
0.08027419447898865,
|
| 83 |
+
0.05345085263252258,
|
| 84 |
+
0.021373271942138672,
|
| 85 |
+
0.059831492602825165,
|
| 86 |
+
0.05794444680213928,
|
| 87 |
+
0.028789382427930832,
|
| 88 |
+
0.05070403218269348,
|
| 89 |
+
0.10934846848249435,
|
| 90 |
+
0.059825070202350616,
|
| 91 |
+
0.035672876983881,
|
| 92 |
+
0.03681573644280434,
|
| 93 |
+
0.05354110151529312,
|
| 94 |
+
0.03043556772172451,
|
| 95 |
+
0.05011315643787384,
|
| 96 |
+
0.03349003195762634,
|
| 97 |
+
0.06471982598304749,
|
| 98 |
+
0.03476070240139961,
|
| 99 |
+
0.059753965586423874,
|
| 100 |
+
0.05377107858657837,
|
| 101 |
+
0.047112029045820236,
|
| 102 |
+
0.05747378617525101,
|
| 103 |
+
0.07838261127471924,
|
| 104 |
+
0.04439996927976608,
|
| 105 |
+
0.051719676703214645,
|
| 106 |
+
0.07081422954797745,
|
| 107 |
+
0.04764055833220482,
|
| 108 |
+
0.06472727656364441,
|
| 109 |
+
0.025721382349729538,
|
| 110 |
+
0.04938185214996338,
|
| 111 |
+
0.0208204947412014,
|
| 112 |
+
0.04679484665393829,
|
| 113 |
+
0.053672343492507935,
|
| 114 |
+
0.04590842127799988,
|
| 115 |
+
0.02758799120783806,
|
| 116 |
+
0.05496695637702942,
|
| 117 |
+
0.052534304559230804,
|
| 118 |
+
0.042658112943172455,
|
| 119 |
+
0.04993772506713867,
|
| 120 |
+
0.04448267072439194,
|
| 121 |
+
0.021319519728422165,
|
| 122 |
+
0.029168587177991867,
|
| 123 |
+
0.037295062094926834,
|
| 124 |
+
0.04790613800287247,
|
| 125 |
+
0.058183394372463226,
|
| 126 |
+
0.03067963197827339,
|
| 127 |
+
0.05868379399180412,
|
| 128 |
+
0.050936199724674225,
|
| 129 |
+
0.04696328565478325,
|
| 130 |
+
0.029223185032606125,
|
| 131 |
+
0.053025759756565094,
|
| 132 |
+
0.04058123379945755,
|
| 133 |
+
0.06434519588947296,
|
| 134 |
+
0.042083121836185455,
|
| 135 |
+
0.043149348348379135,
|
| 136 |
+
0.047702908515930176,
|
| 137 |
+
0.044372182339429855,
|
| 138 |
+
0.030883528292179108,
|
| 139 |
+
0.029037751257419586,
|
| 140 |
+
0.015058952383697033,
|
| 141 |
+
0.03257162496447563,
|
| 142 |
+
0.025359218940138817,
|
| 143 |
+
0.036260537803173065,
|
| 144 |
+
0.05121390148997307,
|
| 145 |
+
0.05242888256907463,
|
| 146 |
+
0.02808375284075737,
|
| 147 |
+
0.07048864662647247,
|
| 148 |
+
0.061022497713565826,
|
| 149 |
+
0.03878414258360863,
|
| 150 |
+
0.042936887592077255,
|
| 151 |
+
0.05563678964972496,
|
| 152 |
+
0.02945074997842312,
|
| 153 |
+
0.020809678360819817,
|
| 154 |
+
0.05034352093935013,
|
| 155 |
+
0.06387125700712204,
|
| 156 |
+
0.028472203761339188,
|
| 157 |
+
0.05118788033723831,
|
| 158 |
+
0.04340597987174988,
|
| 159 |
+
0.07120423763990402,
|
| 160 |
+
0.044768076390028,
|
| 161 |
+
0.06425879895687103,
|
| 162 |
+
0.050438642501831055,
|
| 163 |
+
0.047228530049324036,
|
| 164 |
+
0.0628366768360138,
|
| 165 |
+
0.060967423021793365,
|
| 166 |
+
0.06019226461648941,
|
| 167 |
+
0.07305523753166199,
|
| 168 |
+
0.046951111406087875,
|
| 169 |
+
0.049580540508031845,
|
| 170 |
+
0.03337812051177025,
|
| 171 |
+
0.08303362131118774,
|
| 172 |
+
0.06565201282501221,
|
| 173 |
+
0.02152317762374878,
|
| 174 |
+
0.04263696074485779,
|
| 175 |
+
0.033167771995067596,
|
| 176 |
+
0.050332438200712204,
|
| 177 |
+
0.046692442148923874,
|
| 178 |
+
0.05139806866645813,
|
| 179 |
+
0.06747598946094513,
|
| 180 |
+
0.07169119268655777,
|
| 181 |
+
0.05237269029021263,
|
| 182 |
+
0.0478736087679863,
|
| 183 |
+
0.03808905929327011,
|
| 184 |
+
0.050925932824611664,
|
| 185 |
+
0.06277015060186386,
|
| 186 |
+
0.0564577691257,
|
| 187 |
+
0.05538750812411308,
|
| 188 |
+
0.04669160395860672,
|
| 189 |
+
0.04851044714450836,
|
| 190 |
+
0.08011412620544434,
|
| 191 |
+
0.028662553057074547,
|
| 192 |
+
0.06408864259719849,
|
| 193 |
+
0.06013261526823044,
|
| 194 |
+
0.04410611838102341,
|
| 195 |
+
0.017821576446294785,
|
| 196 |
+
0.06472691148519516,
|
| 197 |
+
0.026214828714728355,
|
| 198 |
+
0.054857559502124786,
|
| 199 |
+
0.04198151081800461,
|
| 200 |
+
0.09210877120494843,
|
| 201 |
+
0.0239681676030159,
|
| 202 |
+
0.022113261744379997,
|
| 203 |
+
0.051053278148174286,
|
| 204 |
+
0.051333315670490265,
|
| 205 |
+
0.04088010638952255
|
| 206 |
+
]
|
| 207 |
+
}
|
optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2314e57b3c2a55f75600e10fbe4ece9db00372e887fdba00d10913a7b831baa2
|
| 3 |
+
size 1176366704
|
qwen_2509_object_removal_new_000000500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6585a769ce8e9b04acde8c9d6d333a8aec96826d9a6540cc1edf1122f3512ee4
|
| 3 |
+
size 295147304
|
qwen_2509_object_removal_new_000000550.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d92511fa48f4eda44f7051025c7595cf87ec452ba566c9b910994fa26131a3b0
|
| 3 |
+
size 295148368
|
qwen_2509_object_removal_new_000000600.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f444924ce4e9807b2957919001cd01723afa4198f58ea6ceba8f7e2679ad43c5
|
| 3 |
+
size 295149440
|
qwen_2509_object_removal_new_000000650.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06f0f5634d7de807b88defe8f9c721436999a4bc93ab4bd4ca50999a1f2eb7d8
|
| 3 |
+
size 295150496
|
samples/1765913168168__000000050_0.jpg
ADDED
|
samples/1765913250652__000000050_1.jpg
ADDED
|
samples/1765913333097__000000050_2.jpg
ADDED
|
samples/1765916820855__000000100_0.jpg
ADDED
|
samples/1765916919610__000000100_1.jpg
ADDED
|
samples/1765917018655__000000100_2.jpg
ADDED
|
samples/1765920526538__000000150_0.jpg
ADDED
|
samples/1765920626672__000000150_1.jpg
ADDED
|
samples/1765920727033__000000150_2.jpg
ADDED
|
samples/1765924219383__000000200_0.jpg
ADDED
|
samples/1765924301805__000000200_1.jpg
ADDED
|
samples/1765924384230__000000200_2.jpg
ADDED
|
samples/1765927857787__000000250_0.jpg
ADDED
|
samples/1765927940222__000000250_1.jpg
ADDED
|
samples/1765928022642__000000250_2.jpg
ADDED
|
samples/1765931497363__000000300_0.jpg
ADDED
|
samples/1765931579810__000000300_1.jpg
ADDED
|
samples/1765931662381__000000300_2.jpg
ADDED
|
samples/1765935140019__000000350_0.jpg
ADDED
|
samples/1765935222456__000000350_1.jpg
ADDED
|
samples/1765935305166__000000350_2.jpg
ADDED
|
samples/1765938782154__000000400_0.jpg
ADDED
|
samples/1765938864681__000000400_1.jpg
ADDED
|
samples/1765938947255__000000400_2.jpg
ADDED
|
samples/1765942436028__000000450_0.jpg
ADDED
|
samples/1765942518822__000000450_1.jpg
ADDED
|
samples/1765942601365__000000450_2.jpg
ADDED
|
samples/1765952484578__000000500_0.jpg
ADDED
|
samples/1765952567599__000000500_1.jpg
ADDED
|
samples/1765952650419__000000500_2.jpg
ADDED
|
samples/1765956118170__000000550_0.jpg
ADDED
|
samples/1765956201084__000000550_1.jpg
ADDED
|
samples/1765956284101__000000550_2.jpg
ADDED
|
samples/1765959752173__000000600_0.jpg
ADDED
|
samples/1765959835054__000000600_1.jpg
ADDED
|
samples/1765959917790__000000600_2.jpg
ADDED
|
samples/1765963406167__000000650_0.jpg
ADDED
|
samples/1765963489019__000000650_1.jpg
ADDED
|
samples/1765963571855__000000650_2.jpg
ADDED
|