| | --- |
| | {{ card_data }} |
| | --- |
| | |
| | <!-- This model card has been generated automatically according to the information the training script had access to. You |
| | should probably proofread and complete it, then remove this comment. --> |
| |
|
| | # {{ model_name | default("Diffusion Model") }} |
| | |
| | ## Model description |
| | |
| | This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library |
| | on the `{{ dataset_name }}` dataset. |
| |
|
| | ## Intended uses & limitations |
| |
|
| | #### How to use |
| |
|
| | ```python |
| | # TODO: add an example code snippet for running this diffusion pipeline |
| | ``` |
| |
|
| | #### Limitations and bias |
| |
|
| | [TODO: provide examples of latent issues and potential remediations] |
| |
|
| | ## Training data |
| |
|
| | [TODO: describe the data used to train the model] |
| |
|
| | ### Training hyperparameters |
| |
|
| | The following hyperparameters were used during training: |
| | - learning_rate: {{ learning_rate }} |
| | - train_batch_size: {{ train_batch_size }} |
| | - eval_batch_size: {{ eval_batch_size }} |
| | - gradient_accumulation_steps: {{ gradient_accumulation_steps }} |
| | - optimizer: AdamW with betas=({{ adam_beta1 }}, {{ adam_beta2 }}), weight_decay={{ adam_weight_decay }} and epsilon={{ adam_epsilon }} |
| | - lr_scheduler: {{ lr_scheduler }} |
| | - lr_warmup_steps: {{ lr_warmup_steps }} |
| | - ema_inv_gamma: {{ ema_inv_gamma }} |
| | - ema_inv_gamma: {{ ema_power }} |
| | - ema_inv_gamma: {{ ema_max_decay }} |
| | - mixed_precision: {{ mixed_precision }} |
| | |
| | ### Training results |
| | |
| | 📈 [TensorBoard logs](https://huggingface.co/{{ repo_name }}/tensorboard?#scalars) |
| | |
| | |
| | |