| | --- |
| | license: apache-2.0 |
| | --- |
| | |
| | # Image Restoration |
| |
|
| | This marks the 10th month anniversary of the previous image restoration model. |
| |
|
| | The decoder model redraws the masked area based on the booru tags. |
| |
|
| | The input and the training target are based on: |
| | - LAB color space |
| | - FFT |
| |
|
| | As in another [encoder](https://huggingface.co/nightknocker/sdxs-1b-image-to-longclip-encoder) model, the image was split to 4 equal-sized squares, and then one of the squares was masked by less than 30%. |
| |
|
| | The image goes through the PatchEmbed module. The transformer blocks receive the masked, patched image embeddings along with their Fourier transform, and the [Camie tags](https://huggingface.co/Camais03/camie-tagger-v2) as input. |
| |
|
| | Finally, the latent is [upsampled](https://github.com/zelaki/eqvae/blob/18f2016dd7e9a934c2fb175d6ff54f52422d8928/train_eqvae/ldm/modules/diffusionmodules/model.py) to the original image size. |
| |
|
| | ## Datasets |
| |
|
| | - pixiv rank |