Update README.md
Browse files
README.md
CHANGED
|
@@ -1,16 +1,57 @@
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
| 3 |
---
|
|
|
|
| 4 |
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
-
BHI Filtered Dataset consists of:
|
| 14 |
DF2K -> 12'639 Tiles
|
| 15 |
FFHQ -> 35'112 Tiles
|
| 16 |
HQ50K -> 61'647 Tiles
|
|
@@ -18,4 +59,10 @@ ImageNet -> 4'505 Tiles
|
|
| 18 |
LSDIR -> 116'141 Tiles
|
| 19 |
OST -> 1'048 Tiles
|
| 20 |
|
| 21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
| 3 |
---
|
| 4 |
+
# BHI SISR Dataset
|
| 5 |
|
| 6 |
+
The BHI SISR Dataset's purpose is for training single image super-resolution models and is a result of tests on my BHI filtering method, which I made [a huggingface community blogpost about](https://huggingface.co/blog/Phips/bhi-filtering).
|
| 7 |
+
|
| 8 |
+
TODO it consists of X images, which are all 512x512px dimensions and in the png format.
|
| 9 |
+
TODO visual example of the dataset
|
| 10 |
+
|
| 11 |
+
## Used Datasets
|
| 12 |
+
|
| 13 |
+
This BHI SISR Dataset consists of the following datasets:
|
| 14 |
+
|
| 15 |
+
[HQ50K](https://github.com/littleYaang/HQ-50K)
|
| 16 |
+
[ImageNet](https://www.image-net.org/)
|
| 17 |
+
[FFHQ](https://github.com/NVlabs/ffhq-dataset)
|
| 18 |
+
[LSDIR](https://github.com/ofsoundof/LSDIR)
|
| 19 |
+
[DF2K](https://www.kaggle.com/datasets/thaihoa1476050/df2k-ost)
|
| 20 |
+
[OST](https://www.kaggle.com/datasets/thaihoa1476050/df2k-ost)
|
| 21 |
+
[iNaturalist 2019](https://github.com/visipedia/inat_comp/tree/master/2019)
|
| 22 |
+
[COCO 2017 Train](https://cocodataset.org/#download)
|
| 23 |
+
[COCO 2017 Unlabeled](https://cocodataset.org/#download)
|
| 24 |
+
[Nomosv2](https://github.com/neosr-project/neosr?tab=readme-ov-file#-datasets)
|
| 25 |
+
[HFA2K](https://github.com/neosr-project/neosr?tab=readme-ov-file#-datasets)
|
| 26 |
+
[Nomos_Uni](https://github.com/neosr-project/neosr?tab=readme-ov-file#-datasets)
|
| 27 |
+
[ModernAnimation1080_v3](https://huggingface.co/datasets/Zarxrax/ModernAnimation1080_v3)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
## Tiling
|
| 31 |
+
|
| 32 |
+
These datasets have then been tiled to 512x512px for improved I/O training speed, and normalization of image dimensions is also nice, so it will take consistent ressources if processing.
|
| 33 |
+
|
| 34 |
+
In some cases these led to fewer images in the dataset because they contained images with < 512px dimensions which were filtered out, some examples are:
|
| 35 |
+
COCO 2017 unlabeled from 123'403 images -> 8'814 tiles.
|
| 36 |
+
COCO 2017 train from 118'287 images -> 8'442 tiles.
|
| 37 |
+
|
| 38 |
+
And in some cases this led to more images, because the original images were high resolution and therefore gave multiple 512x512 tiles per single image.
|
| 39 |
+
For example HQ50K -> 213'396 tiles.
|
| 40 |
+
|
| 41 |
+
## Conversion
|
| 42 |
+
|
| 43 |
+
If the images in the dataset were in the jpg format, they have been converted to png format using [Mogrify](https://imagemagick.org/script/mogrify.php).
|
| 44 |
+
|
| 45 |
+
## BHI Filtering
|
| 46 |
+
|
| 47 |
+
I then filtered these sets with the BHI filtering method using the following thresholds:
|
| 48 |
+
|
| 49 |
+
Blockiness < 30
|
| 50 |
+
HyperIQA >= 0.2
|
| 51 |
+
IC9600 >= 0.4
|
| 52 |
+
|
| 53 |
+
Which led to following dataset tile quantities that satisfied the filtering process, which made it into the BHI SISR Dataset:
|
| 54 |
|
|
|
|
| 55 |
DF2K -> 12'639 Tiles
|
| 56 |
FFHQ -> 35'112 Tiles
|
| 57 |
HQ50K -> 61'647 Tiles
|
|
|
|
| 59 |
LSDIR -> 116'141 Tiles
|
| 60 |
OST -> 1'048 Tiles
|
| 61 |
|
| 62 |
+
## Filename normalization
|
| 63 |
+
|
| 64 |
+
All these subsets have been filename normalized, meaning 0.png, 1.png, 2.png, and so forth. They then have been merged into the BHI dataset, which folder has been normalized again.
|
| 65 |
+
|
| 66 |
+
## Optimization
|
| 67 |
+
|
| 68 |
+
Then I used [oxipng](https://github.com/shssoichiro/oxipng) ("oxipng --strip safe --alpha *.png") for optimization.
|