Spaces:
Sleeping
Sleeping
| # Dataset Preparation | |
| The data files tree should be look like: | |
| ``` | |
| data/ | |
| eval/ | |
| dir300/ | |
| 1_in.png | |
| 1_gt.png | |
| ... | |
| kligler/ | |
| jung/ | |
| osr/ | |
| realdae/ | |
| docunet_docaligner/ | |
| dibco18/ | |
| train/ | |
| dewarping/ | |
| doc3d/ | |
| deshadowing/ | |
| fsdsrd/ | |
| tdd/ | |
| appearance/ | |
| clean_pdfs/ | |
| realdae/ | |
| deblurring/ | |
| tdd/ | |
| binarization/ | |
| bickly/ | |
| dibco/ | |
| noise_office/ | |
| phibd/ | |
| msi/ | |
| ``` | |
| ## Evaluation Dataset | |
| You can find the links for downloading the dataset we used for evaluation (Tables 1 and 2) in [this](https://github.com/ZZZHANG-jx/Recommendations-Document-Image-Processing/tree/master) repository, including DIR300 (300 samples), Kligler (300 samples), Jung (87 samples), OSR (237 samples), RealDAE (150 samples), DocUNet_DocAligner (150 samples), TDD (16000 samples) and DIBCO18 (10 samples). After downloading, add the suffix of `_in` and `_gt` to the input image and gt image respectively, and place them in the folder of the corresponding dataset | |
| ## Training Dataset | |
| You can find the links for downloading the dataset we used for training in [this](https://github.com/ZZZHANG-jx/Recommendations-Document-Image-Processing/tree/master) repository. | |
| ### Dewarping | |
| - Doc3D | |
| - Mask extraction: you should extract the mask for each image from the uv data in Doc3D | |
| - Background preparation: you can download the background data from [here](https://www.robots.ox.ac.uk/~vgg/data/dtd/) and specify it for self.background_paths in `loaders/docres_loader.py` | |
| - JSON preparation: | |
| ``` | |
| [ | |
| ## you need to specify the paths of 'in_path', 'mask_path and 'gt_path': | |
| { | |
| "in_path": "dewarping/doc3d/img/1/102_1-pp_Page_048-xov0001.png", | |
| "mask_path": "dewarping/doc3d/mask/1/102_1-pp_Page_048-xov0001.png", | |
| "gt_path": "dewarping/doc3d/bm/1/102_1-pp_Page_048-xov0001.npy" | |
| } | |
| ] | |
| ``` | |
| ### Deshadowing | |
| - RDD | |
| - FSDSRD | |
| - JSON preparation | |
| ``` | |
| [ ## you need to specify the paths of 'in_path' and 'gt_path', for example: | |
| { | |
| "in_path": "deshadowing/fsdsrd/im/00004.png", | |
| "gt_path": "deshadowing/fsdsrd/gt/00004.png" | |
| }, | |
| { | |
| "in_path": "deshadowing/rdd/im/00004.png", | |
| "gt_path": "deshadowing/rdd/gt/00004.png" | |
| } | |
| ] | |
| ``` | |
| ### Appearance enhancement | |
| - Doc3DShade | |
| - Clean PDFs collection: You should collection PDFs files from the internet and convert them as images to serve as the source for synthesis. | |
| - Extract shadows from Doc3DShade by using `data/preprocess/shadow_extract.py` and dewarp the obtained shadows by using `data/MBD/infer.py`. Then you should specify self.shadow_paths in `loaders/docres_loader.py` | |
| - RealDAE | |
| - JSON preparation: | |
| ``` | |
| [ | |
| ## for Doc3DShade dataset, you only need to specify the path of image from PDF, for example: | |
| { | |
| 'gt_path':'appearance/clean_pdfs/1.jpg' | |
| }, | |
| ## for RealDAE dataset, you need to specify the paths of both input and gt, for example: | |
| { | |
| 'in_path': 'appearance/realdae/1_in.jpg', | |
| 'gt_path': 'appearance/realdae/1_gt.jpg' | |
| } | |
| ] | |
| ``` | |
| ### Debluring | |
| - TDD | |
| - JSON preparation | |
| ``` | |
| [ ## you need to specify the paths of 'in_path' and 'gt_path', for example: | |
| { | |
| "in_path": "debluring/tdd/im/00004.png", | |
| "gt_path": "debluring/tdd/gt/00004.png" | |
| }, | |
| ] | |
| ``` | |
| ### Binarization | |
| - Bickly | |
| - DTPrompt preparation: Since the DTPrompt for binarization is time-expensive, we obtain it offline before training. Use `data/preprocess/sauvola_binarize.py` | |
| - DIBCO | |
| - DTPrompt preparation: the same as Bickly | |
| - Noise Office | |
| - DTPrompt preparation: the same as Bickly | |
| - PHIDB | |
| - DTPrompt preparation: the same as Bickly | |
| - MSI | |
| - DTPrompt preparation: the same as Bickly | |
| - JSON preparation | |
| ``` | |
| [ | |
| ## you need to specify the paths of 'in_path', 'gt_path', 'bin_path', 'thr_path' and 'gradient_path', for example: | |
| { | |
| "in_path": "binarization/noise_office/imgs/1.png", | |
| "gt_path": "binarization/noise_office/gt_imgs/1.png", | |
| "bin_path": "binarization/noise_office/imgs/1_bin.png", | |
| "thr_path": "binarization/noise_office/imgs/1_thr.png", | |
| "gradient_path": "binarization/noise_office/imgs/1_gradient.png" | |
| }, | |
| ] | |
| ``` | |
| After all the data are prepared, you should specify the dataset_setting in `train.py`. | |