| --- |
| {{ card_data }} |
| --- |
| |
| <!-- This model card has been generated automatically according to the information the training script had access to. You |
| should probably proofread and complete it, then remove this comment. --> |
|
|
| # {{ model_name | default("Diffusion Model") }} |
| |
| ## Model description |
| |
| This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library |
| on the `{{ dataset_name }}` dataset. |
|
|
| ## Intended uses & limitations |
|
|
| #### How to use |
|
|
| ```python |
| # TODO: add an example code snippet for running this diffusion pipeline |
| ``` |
|
|
| #### Limitations and bias |
|
|
| [TODO: provide examples of latent issues and potential remediations] |
|
|
| ## Training data |
|
|
| [TODO: describe the data used to train the model] |
|
|
| ### Training hyperparameters |
|
|
| The following hyperparameters were used during training: |
| - learning_rate: {{ learning_rate }} |
| - train_batch_size: {{ train_batch_size }} |
| - eval_batch_size: {{ eval_batch_size }} |
| - gradient_accumulation_steps: {{ gradient_accumulation_steps }} |
| - optimizer: AdamW with betas=({{ adam_beta1 }}, {{ adam_beta2 }}), weight_decay={{ adam_weight_decay }} and epsilon={{ adam_epsilon }} |
| - lr_scheduler: {{ lr_scheduler }} |
| - lr_warmup_steps: {{ lr_warmup_steps }} |
| - ema_inv_gamma: {{ ema_inv_gamma }} |
| - ema_inv_gamma: {{ ema_power }} |
| - ema_inv_gamma: {{ ema_max_decay }} |
| - mixed_precision: {{ mixed_precision }} |
| |
| ### Training results |
| |
| 📈 [TensorBoard logs](https://huggingface.co/{{ repo_name }}/tensorboard?#scalars) |
| |
| |
| |