Dataset Preview
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed because of a cast error
Error code: DatasetGenerationCastError
Exception: DatasetGenerationCastError
Message: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 38 new columns ({'dora_wd', 'unit', 'use_tucker', 'block_alphas', 'train_norm', 'dim_from_weights', 'constrain', 'module_dropout', 'network_alpha', 'training_comment', 'scale_weight_norms', 'stop_text_encoder_training_pct', 'rescaled', 'down_lr_weight', 'train_on_input', 'decompose_both', 'LoRA_type', 'network_dropout', 'LyCORIS_preset', 'conv_dim', 'block_dims', 'network_dim', 'text_encoder_lr', 'use_cp', 'conv_block_alphas', 'block_lr_zero_threshold', 'rank_dropout_scale', 'unet_lr', 'mid_lr_weight', 'conv_block_dims', 'max_grad_norm', 'up_lr_weight', 'use_scalar', 'bypass_mode', 'network_weights', 'factor', 'rank_dropout', 'conv_alpha'}) and 51 missing columns ({'sd3_cache_text_encoder_outputs', 'logit_std', 'save_last_n_epochs', 'sd3_text_encoder_batch_size', 'weighting_scheme', 'learning_rate_te1', 'no_token_padding', 'discrete_flow_shift', 'cpu_offload_checkpointing', 'log_config', 'timestep_sampling', 'flux1_t5xxl', 'skip_cache_check', 'fused_backward_pass', 'fused_optimizer_groups', 'mem_eff_save', 'clip_l', 'lr_scheduler_type', 'save_t5xxl', 'model_prediction_type', 'blocks_to_swap', 'sd3_cache_text_encoder_outputs_to_disk', 'flux1_clip_l', 'flux1_cache_text_encoder_outputs_to_disk', 'flux_fused_backward_pass', 'ae', 'learning_rate_te', 'logit_mean', 'disable_mmap_load_safetensors', 'mode_scale', 'apply_t5_attn_mask', 'flux1_checkbox', 'blockwise_fused_optimizers', 'single_blocks_to_swap', 'split_mode', 'save_clip', 't5xxl_device', 'clip_g', 'save_last_n_epochs_state', 'flux1_cache_text_encoder_outputs', 'save_as_bool', 'stop_text_encoder_training', 'learning_rate_te2', 'double_blocks_to_swap', 't5xxl_dtype', 't5xxl', 't5xxl_max_token_length', 'train_blocks', 'guidance_scale', 'sd3_checkbox', 'lr_warmup_steps'}).
This happened while the json dataset builder was generating data using
hf://datasets/kratosboy507/kratos_configs/senajuo2idol_noobv75_20241209-003658.json (at revision c8e8663bd8a4d74931fd0031d515da7ca06e6f51)
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Traceback: Traceback (most recent call last):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1870, in _prepare_split_single
writer.write_table(table)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 622, in write_table
pa_table = table_cast(pa_table, self._schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2292, in table_cast
return cast_table_to_schema(table, schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2240, in cast_table_to_schema
raise CastError(
datasets.table.CastError: Couldn't cast
LoRA_type: string
LyCORIS_preset: string
adaptive_noise_scale: int64
additional_parameters: string
async_upload: bool
block_alphas: string
block_dims: string
block_lr_zero_threshold: string
bucket_no_upscale: bool
bucket_reso_steps: int64
bypass_mode: bool
cache_latents: bool
cache_latents_to_disk: bool
caption_dropout_every_n_epochs: int64
caption_dropout_rate: int64
caption_extension: string
clip_skip: int64
color_aug: bool
constrain: int64
conv_alpha: int64
conv_block_alphas: string
conv_block_dims: string
conv_dim: int64
dataset_config: string
debiased_estimation_loss: bool
decompose_both: bool
dim_from_weights: bool
dora_wd: bool
down_lr_weight: string
dynamo_backend: string
dynamo_mode: string
dynamo_use_dynamic: bool
dynamo_use_fullgraph: bool
enable_bucket: bool
epoch: int64
extra_accelerate_launch_args: string
factor: int64
flip_aug: bool
fp8_base: bool
full_bf16: bool
full_fp16: bool
gpu_ids: string
gradient_accumulation_steps: int64
gradient_checkpointing: bool
huber_c: int64
huber_schedule: string
huggingface_path_in_repo: string
huggingface_repo_id: string
huggingface_repo_type: string
huggingface_repo_visibility: string
huggingface_token: string
ip_noise_gamma: int64
ip_noise_gamma_random_strength: bool
keep_tokens: int64
learning_rate: int64
log_tracker_config: string
log_tracker_name: string
log_with: string
logging_dir: string
loss_type: string
lr_scheduler: string
lr_scheduler_args: string
lr_scheduler_num_cycles: int64
lr_scheduler_power: int64
lr_warmup: i
...
etwork_dim: int64
network_dropout: int64
network_weights: string
noise_offset: double
noise_offset_random_strength: bool
noise_offset_type: string
num_cpu_threads_per_process: int64
num_machines: int64
num_processes: int64
optimizer: string
optimizer_args: string
output_dir: string
output_name: string
persistent_data_loader_workers: bool
pretrained_model_name_or_path: string
prior_loss_weight: int64
random_crop: bool
rank_dropout: int64
rank_dropout_scale: bool
reg_data_dir: string
rescaled: bool
resume: string
resume_from_huggingface: string
sample_every_n_epochs: int64
sample_every_n_steps: int64
sample_prompts: string
sample_sampler: string
save_every_n_epochs: int64
save_every_n_steps: int64
save_last_n_steps: int64
save_last_n_steps_state: int64
save_model_as: string
save_precision: string
save_state: bool
save_state_on_train_end: bool
save_state_to_huggingface: bool
scale_v_pred_loss_like_noise_pred: bool
scale_weight_norms: int64
sdxl: bool
sdxl_cache_text_encoder_outputs: bool
sdxl_no_half_vae: bool
seed: int64
shuffle_caption: bool
stop_text_encoder_training_pct: int64
text_encoder_lr: int64
train_batch_size: int64
train_data_dir: string
train_norm: bool
train_on_input: bool
training_comment: string
unet_lr: int64
unit: int64
up_lr_weight: string
use_cp: bool
use_scalar: bool
use_tucker: bool
v2: bool
v_parameterization: bool
v_pred_like_loss: int64
vae: string
vae_batch_size: int64
wandb_api_key: string
wandb_run_name: string
weighted_captions: bool
xformers: string
to
{'adaptive_noise_scale': Value(dtype='int64', id=None), 'additional_parameters': Value(dtype='string', id=None), 'ae': Value(dtype='string', id=None), 'apply_t5_attn_mask': Value(dtype='bool', id=None), 'async_upload': Value(dtype='bool', id=None), 'blocks_to_swap': Value(dtype='int64', id=None), 'blockwise_fused_optimizers': Value(dtype='bool', id=None), 'bucket_no_upscale': Value(dtype='bool', id=None), 'bucket_reso_steps': Value(dtype='int64', id=None), 'cache_latents': Value(dtype='bool', id=None), 'cache_latents_to_disk': Value(dtype='bool', id=None), 'caption_dropout_every_n_epochs': Value(dtype='int64', id=None), 'caption_dropout_rate': Value(dtype='int64', id=None), 'caption_extension': Value(dtype='string', id=None), 'clip_g': Value(dtype='string', id=None), 'clip_l': Value(dtype='string', id=None), 'clip_skip': Value(dtype='int64', id=None), 'color_aug': Value(dtype='bool', id=None), 'cpu_offload_checkpointing': Value(dtype='bool', id=None), 'dataset_config': Value(dtype='string', id=None), 'debiased_estimation_loss': Value(dtype='bool', id=None), 'disable_mmap_load_safetensors': Value(dtype='bool', id=None), 'discrete_flow_shift': Value(dtype='float64', id=None), 'double_blocks_to_swap': Value(dtype='int64', id=None), 'dynamo_backend': Value(dtype='string', id=None), 'dynamo_mode': Value(dtype='string', id=None), 'dynamo_use_dynamic': Value(dtype='bool', id=None), 'dynamo_use_fullgraph': Value(dtype='bool', id=None), 'enable_bucket': Value(dtype='bool', id=None), '
...
'sd3_cache_text_encoder_outputs_to_disk': Value(dtype='bool', id=None), 'sd3_checkbox': Value(dtype='bool', id=None), 'sd3_text_encoder_batch_size': Value(dtype='int64', id=None), 'sdxl': Value(dtype='bool', id=None), 'sdxl_cache_text_encoder_outputs': Value(dtype='bool', id=None), 'sdxl_no_half_vae': Value(dtype='bool', id=None), 'seed': Value(dtype='int64', id=None), 'shuffle_caption': Value(dtype='bool', id=None), 'single_blocks_to_swap': Value(dtype='int64', id=None), 'skip_cache_check': Value(dtype='bool', id=None), 'split_mode': Value(dtype='bool', id=None), 'stop_text_encoder_training': Value(dtype='int64', id=None), 't5xxl': Value(dtype='string', id=None), 't5xxl_device': Value(dtype='string', id=None), 't5xxl_dtype': Value(dtype='string', id=None), 't5xxl_max_token_length': Value(dtype='int64', id=None), 'timestep_sampling': Value(dtype='string', id=None), 'train_batch_size': Value(dtype='int64', id=None), 'train_blocks': Value(dtype='string', id=None), 'train_data_dir': Value(dtype='string', id=None), 'v2': Value(dtype='bool', id=None), 'v_parameterization': Value(dtype='bool', id=None), 'v_pred_like_loss': Value(dtype='int64', id=None), 'vae': Value(dtype='string', id=None), 'vae_batch_size': Value(dtype='int64', id=None), 'wandb_api_key': Value(dtype='string', id=None), 'wandb_run_name': Value(dtype='string', id=None), 'weighted_captions': Value(dtype='bool', id=None), 'weighting_scheme': Value(dtype='string', id=None), 'xformers': Value(dtype='string', id=None)}
because column names don't match
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1417, in compute_config_parquet_and_info_response
parquet_operations = convert_to_parquet(builder)
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1049, in convert_to_parquet
builder.download_and_prepare(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare
self._download_and_prepare(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1000, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1741, in _prepare_split
for job_id, done, content in self._prepare_split_single(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1872, in _prepare_split_single
raise DatasetGenerationCastError.from_cast_error(
datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 38 new columns ({'dora_wd', 'unit', 'use_tucker', 'block_alphas', 'train_norm', 'dim_from_weights', 'constrain', 'module_dropout', 'network_alpha', 'training_comment', 'scale_weight_norms', 'stop_text_encoder_training_pct', 'rescaled', 'down_lr_weight', 'train_on_input', 'decompose_both', 'LoRA_type', 'network_dropout', 'LyCORIS_preset', 'conv_dim', 'block_dims', 'network_dim', 'text_encoder_lr', 'use_cp', 'conv_block_alphas', 'block_lr_zero_threshold', 'rank_dropout_scale', 'unet_lr', 'mid_lr_weight', 'conv_block_dims', 'max_grad_norm', 'up_lr_weight', 'use_scalar', 'bypass_mode', 'network_weights', 'factor', 'rank_dropout', 'conv_alpha'}) and 51 missing columns ({'sd3_cache_text_encoder_outputs', 'logit_std', 'save_last_n_epochs', 'sd3_text_encoder_batch_size', 'weighting_scheme', 'learning_rate_te1', 'no_token_padding', 'discrete_flow_shift', 'cpu_offload_checkpointing', 'log_config', 'timestep_sampling', 'flux1_t5xxl', 'skip_cache_check', 'fused_backward_pass', 'fused_optimizer_groups', 'mem_eff_save', 'clip_l', 'lr_scheduler_type', 'save_t5xxl', 'model_prediction_type', 'blocks_to_swap', 'sd3_cache_text_encoder_outputs_to_disk', 'flux1_clip_l', 'flux1_cache_text_encoder_outputs_to_disk', 'flux_fused_backward_pass', 'ae', 'learning_rate_te', 'logit_mean', 'disable_mmap_load_safetensors', 'mode_scale', 'apply_t5_attn_mask', 'flux1_checkbox', 'blockwise_fused_optimizers', 'single_blocks_to_swap', 'split_mode', 'save_clip', 't5xxl_device', 'clip_g', 'save_last_n_epochs_state', 'flux1_cache_text_encoder_outputs', 'save_as_bool', 'stop_text_encoder_training', 'learning_rate_te2', 'double_blocks_to_swap', 't5xxl_dtype', 't5xxl', 't5xxl_max_token_length', 'train_blocks', 'guidance_scale', 'sd3_checkbox', 'lr_warmup_steps'}).
This happened while the json dataset builder was generating data using
hf://datasets/kratosboy507/kratos_configs/senajuo2idol_noobv75_20241209-003658.json (at revision c8e8663bd8a4d74931fd0031d515da7ca06e6f51)
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
adaptive_noise_scale int64 | additional_parameters string | ae string | apply_t5_attn_mask bool | async_upload bool | blocks_to_swap int64 | blockwise_fused_optimizers bool | bucket_no_upscale bool | bucket_reso_steps int64 | cache_latents bool | cache_latents_to_disk bool | caption_dropout_every_n_epochs int64 | caption_dropout_rate int64 | caption_extension string | clip_g string | clip_l string | clip_skip int64 | color_aug bool | cpu_offload_checkpointing bool | dataset_config string | debiased_estimation_loss bool | disable_mmap_load_safetensors bool | discrete_flow_shift float64 | double_blocks_to_swap int64 | dynamo_backend string | dynamo_mode string | dynamo_use_dynamic bool | dynamo_use_fullgraph bool | enable_bucket bool | epoch int64 | extra_accelerate_launch_args string | flip_aug bool | flux1_cache_text_encoder_outputs bool | flux1_cache_text_encoder_outputs_to_disk bool | flux1_checkbox bool | flux1_clip_l string | flux1_t5xxl string | flux_fused_backward_pass bool | fp8_base bool | full_bf16 bool | full_fp16 bool | fused_backward_pass bool | fused_optimizer_groups int64 | gpu_ids string | gradient_accumulation_steps int64 | gradient_checkpointing bool | guidance_scale int64 | huber_c float64 | huber_schedule string | huggingface_path_in_repo string | huggingface_repo_id string | huggingface_repo_type string | huggingface_repo_visibility string | huggingface_token string | ip_noise_gamma int64 | ip_noise_gamma_random_strength bool | keep_tokens int64 | learning_rate float64 | learning_rate_te int64 | learning_rate_te1 float64 | learning_rate_te2 float64 | log_config bool | log_tracker_config string | log_tracker_name string | log_with string | logging_dir string | logit_mean int64 | logit_std int64 | loss_type string | lr_scheduler string | lr_scheduler_args string | lr_scheduler_num_cycles int64 | lr_scheduler_power int64 | lr_scheduler_type string | lr_warmup int64 | lr_warmup_steps int64 | main_process_port int64 | masked_loss bool | max_bucket_reso int64 | max_data_loader_n_workers int64 | max_resolution string | max_timestep int64 | max_token_length int64 | max_train_epochs int64 | max_train_steps int64 | mem_eff_attn bool | mem_eff_save bool | metadata_author string | metadata_description string | metadata_license string | metadata_tags string | metadata_title string | min_bucket_reso int64 | min_snr_gamma int64 | min_timestep int64 | mixed_precision string | mode_scale float64 | model_list string | model_prediction_type string | multi_gpu bool | multires_noise_discount float64 | multires_noise_iterations int64 | no_token_padding bool | noise_offset float64 | noise_offset_random_strength bool | noise_offset_type string | num_cpu_threads_per_process int64 | num_machines int64 | num_processes int64 | optimizer string | optimizer_args string | output_dir string | output_name string | persistent_data_loader_workers bool | pretrained_model_name_or_path string | prior_loss_weight int64 | random_crop bool | reg_data_dir string | resume string | resume_from_huggingface string | sample_every_n_epochs int64 | sample_every_n_steps int64 | sample_prompts string | sample_sampler string | save_as_bool bool | save_clip bool | save_every_n_epochs int64 | save_every_n_steps int64 | save_last_n_epochs int64 | save_last_n_epochs_state int64 | save_last_n_steps int64 | save_last_n_steps_state int64 | save_model_as string | save_precision string | save_state bool | save_state_on_train_end bool | save_state_to_huggingface bool | save_t5xxl bool | scale_v_pred_loss_like_noise_pred bool | sd3_cache_text_encoder_outputs bool | sd3_cache_text_encoder_outputs_to_disk bool | sd3_checkbox bool | sd3_text_encoder_batch_size int64 | sdxl bool | sdxl_cache_text_encoder_outputs bool | sdxl_no_half_vae bool | seed int64 | shuffle_caption bool | single_blocks_to_swap int64 | skip_cache_check bool | split_mode bool | stop_text_encoder_training int64 | t5xxl string | t5xxl_device string | t5xxl_dtype string | t5xxl_max_token_length int64 | timestep_sampling string | train_batch_size int64 | train_blocks string | train_data_dir string | v2 bool | v_parameterization bool | v_pred_like_loss int64 | vae string | vae_batch_size int64 | wandb_api_key string | wandb_run_name string | weighted_captions bool | weighting_scheme string | xformers string | LoRA_type string | LyCORIS_preset string | block_alphas string | block_dims string | block_lr_zero_threshold string | bypass_mode bool | constrain int64 | conv_alpha int64 | conv_block_alphas string | conv_block_dims string | conv_dim int64 | decompose_both bool | dim_from_weights bool | dora_wd bool | down_lr_weight string | factor int64 | max_grad_norm int64 | mid_lr_weight string | module_dropout int64 | network_alpha int64 | network_dim int64 | network_dropout int64 | network_weights string | rank_dropout int64 | rank_dropout_scale bool | rescaled bool | scale_weight_norms int64 | stop_text_encoder_training_pct int64 | text_encoder_lr int64 | train_norm bool | train_on_input bool | training_comment string | unet_lr int64 | unit int64 | up_lr_weight string | use_cp bool | use_scalar bool | use_tucker bool |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | /home/Ubuntu/Downloads/ae.safetensors | false | false | 0 | false | true | 64 | true | true | 0 | 0 | .txt | 0 | false | false | false | false | 3.1582 | 0 | no | default | false | false | false | 200 | false | true | true | true | /home/Ubuntu/Downloads/clip_l.safetensors | /home/Ubuntu/Downloads/t5xxl_fp16.safetensors | true | false | true | false | false | 0 | 0 | 1 | true | 1 | 0.1 | snr | 0 | false | 0 | 0.000004 | 0 | 0.00001 | 0.00001 | false | 0 | 1 | l2 | constant | 1 | 1 | 0 | 0 | 0 | false | 2,048 | 0 | 1024,1024 | 1,000 | 75 | 0 | 0 | false | true | 256 | 0 | 0 | bf16 | 1.29 | custom | raw | false | 0.3 | 0 | false | 0 | false | Original | 2 | 1 | 1 | Adafactor | scale_parameter=False relative_step=False warmup_init=False weight_decay=0.01 | /home/Ubuntu/apps/StableSwarmUI/Models/diffusion_models | Quality_1 | false | /home/Ubuntu/Downloads/flux1-dev.safetensors | 1 | false | 0 | 0 | euler_a | false | false | 25 | 0 | 0 | 0 | 0 | 0 | safetensors | fp16 | false | false | false | false | false | false | false | false | 1 | false | false | false | 1 | false | 0 | false | false | 0 | bf16 | 512 | sigmoid | 1 | all | /home/Ubuntu/Downloads/training_imgs | false | false | 0 | 4 | false | logit_normal | xformers | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | ||||||||||||||||||||||||||||||
0 | /home/Ubuntu/Downloads/ae.safetensors | false | false | 0 | false | true | 64 | true | true | 0 | 0 | .txt | 0 | false | false | false | false | 3.1582 | 0 | no | default | false | false | false | 200 | false | true | true | true | /home/Ubuntu/Downloads/clip_l.safetensors | /home/Ubuntu/Downloads/t5xxl_fp16.safetensors | true | false | true | false | false | 0 | 0 | 1 | true | 1 | 0.1 | snr | 0 | false | 0 | 0.00001 | 0 | 0.00001 | 0.00001 | false | 0 | 1 | l2 | constant | 1 | 1 | 0 | 0 | 0 | false | 2,048 | 0 | 1024,1024 | 1,000 | 75 | 0 | 0 | false | true | 256 | 0 | 0 | bf16 | 1.29 | custom | raw | false | 0.3 | 0 | false | 0 | false | Original | 2 | 1 | 1 | Adafactor | scale_parameter=False relative_step=False warmup_init=False weight_decay=0.01 | /home/Ubuntu/apps/StableSwarmUI/Models/diffusion_models | Quality_1 | false | /home/Ubuntu/Downloads/flux1-dev.safetensors | 1 | false | 0 | 0 | euler_a | false | false | 25 | 0 | 0 | 0 | 0 | 0 | safetensors | fp16 | false | false | false | false | false | false | false | false | 1 | false | false | false | 1 | false | 0 | false | false | 0 | bf16 | 512 | sigmoid | 7 | all | /home/Ubuntu/Downloads/training_imgs | false | false | 0 | 4 | false | logit_normal | xformers | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | ||||||||||||||||||||||||||||||
0 | --zero_terminal_snr | null | null | false | null | null | true | 32 | false | true | 0 | 0 | .txt | null | null | 0 | false | null | false | null | null | null | no | default | false | false | true | 12 | false | null | null | null | null | null | null | false | false | false | null | null | 1 | true | null | 0 | snr | 0 | false | 3 | 1 | null | null | null | null | D:\lora_model\log | null | null | l2 | cosine_with_restarts | 1 | 1 | null | 5 | null | 0 | false | 2,048 | 0 | 1024,1024 | 1,000 | 225 | 0 | 0 | false | null | 256 | 0 | 0 | bf16 | null | custom | null | false | 0.2 | 6 | null | 0.12 | true | Original | 2 | 1 | 2 | Prodigy | decouple=True weight_decay=0.05 d_coef=1.2 betas=0.9,0.99 use_bias_correction=True safeguard_warmup=True | D:/lora_model/model | senajuo2idol_noobv75 | false | E:/stable-diffusion-webui-master/models/Stable-diffusion/noobaiXLNAIXL_vPred075SVersion.safetensors | 1 | false | 0 | 0 | dpm_2_a | null | null | 1 | 0 | null | null | 0 | 0 | safetensors | fp16 | false | false | false | null | true | null | null | null | null | true | false | true | 4,134 | true | null | null | null | null | null | null | null | null | null | 8 | null | E:\角色相片\學偶\十王\done - 複製 | false | true | 0 | 0 | false | null | xformers | Standard | full | false | 0 | 1 | 1 | false | false | false | -1 | 1 | 0 | 8 | 8 | 0 | 0 | false | false | 0 | 0 | 1 | false | true | 1 | 1 | false | false | false |
No dataset card yet
- Downloads last month
- 6