| | --- |
| | license: mit |
| | tags: |
| | - image |
| | size_categories: |
| | - 1K<n<10K -> 7K |
| | task_categories: |
| | - image-style-transfer |
| | --- |
| | # DeepDigitalFilm |
| |
|
| | DigitalFilm: Use a neural network to simulate film style. |
| |
|
| | --- |
| |
|
| | <!-- PROJECT LOGO --> |
| | <br /> |
| |
|
| | <p align="center"> |
| | <a href="./readme.md"> |
| | </a> |
| |
|
| | <h3 align="center">"DigitalFilm" Digital Film</h3> |
| | <p align="center"> |
| | Use a neural network to simulate film style. |
| | <br /> |
| | <a href="https://github.com/shaojintian/Best_README_template"><strong>Explore the documentation of this project »</strong></a> |
| | <br /> |
| | <br /> |
| | <a href="./app/digitalFilm.py">View the demo</a> |
| | · |
| | <a href="https://github.com/SongZihui-sudo/digitalFilm/issues">Report a bug</a> |
| | · |
| | <a href="https://github.com/SongZihui-sudo/digitalFilm/issues">Propose a new feature</a> |
| | </p> |
| |
|
| | </p> |
| |
|
| | This README.md is for developers and users |
| | [简体中文](./chinese.md) |
| |
|
| | [PowerPoint Presentation |
| | ](https://incandescent-salmiakki-063eb6.netlify.app/) |
| |
|
| | [source code](https://github.com/SongZihui-sudo/digitalFilm) |
| |
|
| | ## Table of Contents |
| |
|
| | - [DeepDigitalFilm](#deepdigitalfilm) |
| | - [Table of Contents](#table-of-contents) |
| | - [Sample](#sample) |
| | - [Run Demo](#run-demo) |
| | - [training model](#training-model) |
| | - [**Installation steps**](#installation-steps) |
| | - [Overall architecture](#overall-architecture) |
| | - [Dataset](#dataset) |
| | - [File directory description](#file-directory-description) |
| | - [Version Control](#version-control) |
| | - [Author](#author) |
| | - [Copyright](#copyright) |
| |
|
| | ### Sample |
| |
|
| |  |
| | <center style="font-size:14px;color:#C0C0C0;text-decoration:underline">Figure 1 Sample rollei_infrared_400</center> |
| |
|
| |  |
| | <center style="font-size:14px;color:#C0C0C0;text-decoration:underline">Figure 2 Sample kodak gold 200</center> |
| |
|
| |  |
| | <center style="font-size:14px;color:#C0C0C0;text-decoration:underline">Figure 3 Sample fuji color 200</center> |
| |
|
| | ### Run Demo |
| |
|
| | > The length and width of the input photo need to be divisible by **32**. |
| |
|
| | ```bash |
| | python digitalFilm.py [-v/-h/-g] -i <input> -o <ouput> -m <model> |
| | ``` |
| | - -v print version information |
| | - -h help information |
| | - -g graphical image selection |
| | - -i input image directory |
| | - -o output image directory |
| | - -m model directory |
| |
|
| | ### training model |
| |
|
| | training model directly use cyclegan.ipynb. |
| | But you need to download the pre-trained model of resnet18 in advance. |
| | Prepare digital photos and film photos in two folders. |
| | The model are included in the Release. |
| |
|
| | ###### **Installation steps** |
| |
|
| | ```sh |
| | git clone https://github.com/SongZihui-sudo/digitalFilm.git |
| | ``` |
| |
|
| | It is best to create an environment in conda now and then install various dependencies. |
| |
|
| | ```sh |
| | pip install -r requirement.txt |
| | ``` |
| |
|
| | ### Overall architecture |
| |
|
| | Converting digital photos to film style can be regarded as an image style conversion task. Therefore, the overall architecture adopts the cycleGAN network. |
| | [pytorch-CycleGAN-and-pix2pix](https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix) |
| | In addition, it is difficult to obtain large-scale digital photos and film-style photos, so an unsupervised approach is adopted to use unpaired data for training. |
| |
|
| | ### Dataset |
| |
|
| | The dataset consists of dual-source image data, the main part of which is collected from high-quality digital photos taken by Xiaomi 13 Ultra mobile phone, and the rest is selected from professional HDR image dataset. |
| | Film samples are collected from the Internet. |
| |
|
| | ### File directory description |
| |
|
| | - DigitalFilm.ipynb is used to train the model |
| | - app is a demo |
| | - digitalFilm.py |
| | - mynet.py |
| | - mynet2.py |
| |
|
| | ### Version Control |
| |
|
| | This project uses Git for version management. You can view the currently available version in the repository. |
| |
|
| | ### Author |
| |
|
| | 151122876@qq.com SongZihui-sudo |
| |
|
| | Zhihu:Dr.who   qq:1751122876 |
| |
|
| | *You can also view all the developers involved in the project in the list of contributors. * |
| |
|
| | ### Copyright |
| |
|
| | This project is licensed under GPLv3. For details, please refer to [LICENSE.txt](./LICENSE.txt) |