Update README.md
Browse files
README.md
CHANGED
|
@@ -1,4 +1,8 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
library_name: sana, sana-sprint
|
| 3 |
tags:
|
| 4 |
- text-to-image
|
|
@@ -71,7 +75,6 @@ Source code is available at https://github.com/NVlabs/Sana.
|
|
| 71 |
- **Model size:** 1.6B parameters
|
| 72 |
- **Model precision:** torch.bfloat16 (BF16)
|
| 73 |
- **Model resolution:** This model is developed to generate 1024px based images with multi-scale heigh and width.
|
| 74 |
-
- **License:** [NSCL v2-custom](./LICENSE.txt). Governing Terms: NVIDIA License. Additional Information: [Gemma Terms of Use | Google AI for Developers](https://ai.google.dev/gemma/terms) for Gemma-2-2B-IT, [Gemma Prohibited Use Policy | Google AI for Developers](https://ai.google.dev/gemma/prohibited_use_policy).
|
| 75 |
- **Model Description:** This is a model that can be used to generate and modify images based on text prompts.
|
| 76 |
It is a Linear Diffusion Transformer that uses one fixed, pretrained text encoders ([Gemma2-2B-IT](https://huggingface.co/google/gemma-2-2b-it))
|
| 77 |
and one 32x spatial-compressed latent feature encoder ([DC-AE](https://hanlab.mit.edu/projects/dc-ae)).
|
|
@@ -85,6 +88,9 @@ For research purposes, we recommend our `generative-models` Github repository (h
|
|
| 85 |
- **Demo:** https://nv-sana.mit.edu/sprint
|
| 86 |
- **Guidance:** https://github.com/NVlabs/Sana/asset/docs/sana_sprint.md
|
| 87 |
|
|
|
|
|
|
|
|
|
|
| 88 |
|
| 89 |
## Uses
|
| 90 |
|
|
|
|
| 1 |
---
|
| 2 |
+
license: other
|
| 3 |
+
license_name: nvidia-open-model-license
|
| 4 |
+
license_link: >-
|
| 5 |
+
https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/
|
| 6 |
library_name: sana, sana-sprint
|
| 7 |
tags:
|
| 8 |
- text-to-image
|
|
|
|
| 75 |
- **Model size:** 1.6B parameters
|
| 76 |
- **Model precision:** torch.bfloat16 (BF16)
|
| 77 |
- **Model resolution:** This model is developed to generate 1024px based images with multi-scale heigh and width.
|
|
|
|
| 78 |
- **Model Description:** This is a model that can be used to generate and modify images based on text prompts.
|
| 79 |
It is a Linear Diffusion Transformer that uses one fixed, pretrained text encoders ([Gemma2-2B-IT](https://huggingface.co/google/gemma-2-2b-it))
|
| 80 |
and one 32x spatial-compressed latent feature encoder ([DC-AE](https://hanlab.mit.edu/projects/dc-ae)).
|
|
|
|
| 88 |
- **Demo:** https://nv-sana.mit.edu/sprint
|
| 89 |
- **Guidance:** https://github.com/NVlabs/Sana/asset/docs/sana_sprint.md
|
| 90 |
|
| 91 |
+
## License/Terms of Use
|
| 92 |
+
|
| 93 |
+
GOVERNING TERMS: This trial service is governed by the [NVIDIA API Trial Terms of Service](https://assets.ngc.nvidia.com/products/api-catalog/legal/NVIDIA%20API%20Trial%20Terms%20of%20Service.pdf). Use of this model is governed by the [NVIDIA Open Model License Agreement](https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/).
|
| 94 |
|
| 95 |
## Uses
|
| 96 |
|