Datasets:
File size: 8,768 Bytes
0ba52cd 7ffae14 0ba52cd aa2bcc0 0ba52cd 7ffae14 47295d2 7ffae14 0ba52cd aa2bcc0 7ffae14 aa2bcc0 7ffae14 aa2bcc0 7ffae14 f53d9a5 1378110 f53d9a5 1378110 f53d9a5 1378110 f53d9a5 0ba52cd f53d9a5 7ffae14 0ba52cd 47295d2 f53d9a5 0ba52cd aa2bcc0 0ba52cd f53d9a5 0ba52cd aa2bcc0 0ba52cd aa2bcc0 f53d9a5 0ba52cd f53d9a5 0ba52cd f53d9a5 0ba52cd f53d9a5 c6abc5d 0ba52cd b82e466 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 |
# HyperForensics++ Dataset Contribution Guide
This guide outlines the steps to contribute to the HyperForensics++ dataset repository.
---
## 1. Clone the Repository
First, clone the repository from Hugging Face to your local machine:
```bash
# Make sure git-lfs is installed (https://git-lfs.com)
git lfs install
git clone https://huggingface.co/datasets/OtoroLin/HyperForensics-plus-plus
cd ./HyperForensics-plus-plus
```
## 2. Extract the Dataset
Decompress the `.tar.gz` files into a separate local hierarchy. You can utilize the provided `unzipping.sh` script. First, modify the `ROOT_DIR` variable to the `local_data` path
```bash
# In zipping.sh
# Defined the root directory of the whole dataset
ROOT_DIR="/root/to/local_data"
```
Due to the redundancy of our dataset, we use **`pigz`** program instead of standar `gipz` program to compress the files. Make sure you install it first.
```bash
sudo apt install pigz
# If you are using miniconda
conda install pigz
```
### 2.1 Decompressing the specify method and configuration:
```bash
./unzipping.sh --method <method name> --config <config name>
```
Replace `<method name>` with the forgery method and `<config name>` with the configuration index. This will automatically decompressed `.tar.gz` file and put it under hierarchy of the local dataset directory.
For instance, if I want to decompress the images under `ADMM-ADAM` forgery method, `config0` configuration:
```bash
./unzipping.sh --method ADMM-ADAM --config 0
```
This will decompress `HyperForensics-plus-plus/data/ADMM-ADAM/config0/config0.tar.gz` to `/path/to/local/ADMM-ADAM/config0/*`.
### 2.2 Decompress the entire hierarchy
```bash
./unzipping.sh --all
```
This will recursively decompress all `tar.gz` file in the `HyperForensics-plus-plus/data` and automatically put the decompressed files under hierarchy of the local dataset directory.
## 3. [Create a pull request](https://huggingface.co/docs/hub/repositories-pull-requests-discussions#pull-requests-advanced-usage)
The pull requests on Hugging Face do not use forks and branches, but instead custom “branches” called `refs` that are stored directly on the source repo.
The advantage of using custom `refs` (like `refs/pr/42` for instance) instead of branches is that they’re not fetched (by default) by people cloning the repo, but they can still be fetched on demand.
### 3.1 Create a new PR on the hugging face
Go to *community* $\to$ *New pull request* $\to$ *On your machine* $\to$ enter *Pull request title* $\to$ *Creat PR branch*
Then, you shall see a brand new PR page with the title of your PR on the top left. Make note of the index number next to the title `#{index}`.
> In this step, you actually create a new custom branch on **remote** under reference of `ref/pr/#{index}`.
The PR is still in **draft mode**, the maintainer can not merge the PR untill you publish the PR.
### 3.2 Create a new PR branch on local
```bash
git fetch origin refs/pr/<index>
git checkout pr/<index>
```
Fetch the remote PR branch `ref/pr/<index>` to local PR branch. Then checkout to the newly created custom branch under references `pr/`, and just for unambiguity, using the PR index as the name of that reference. You can actually create a new branch locally here as usual.
### 3.3 Push the PR branch
After you finish your modification, push the local PR branch to remote Hugging Face. Check it out [here at 6](#6-create-a-pull-request).
## 4. Modify or Generate Data
Make changes or generate new data in the extracted directory. Ensure the data follows the required structure:
```
local_data/
|-{attacking method}/
| |-config0/
| | |-images.npy
| |-config1/
| | ...
| |-config4/
|-{attacking method}/
| |-config{0-4}/
|-Origin
| |-images.npy
HyperForensics-plus-plus/
|-data/
```
## 5. Zip the Directory
After modifying or generating data, zip the directory into `.tar.gz` files and place it in the repository.
You can utilize the provided `zipping.sh` script. First, modify the `ROOT_DIR` variable to the `local_data` path
```bash
# In zipping.sh
# Defined the root directory of the whole dataset
ROOT_DIR="/root/to/local_data"
```
Due to the redundancy of our dataset, we use **`pigz`** program instead of standar `gipz` program to compress the files. Make sure you install it first.
```bash
sudo apt install pigz
# If you are using miniconda
conda install pigz
```
### 5.1 Compressing the specify method and configuration:
```bash
./zipping.sh --method <method name> --config <config name>
```
Replace <method name> with the forgery method and <config name> with the configuration index. This will automatically put the compressed `.tar.gz` file under hierarchy of the dataset repo.
For instance, if I want to zip the images under `ADMM-ADAM` forgery method, `config0` configuration:
```bash
./zipping.sh --method ADMM-ADAM --config 0
```
This will compress `/path/to/local/ADMM-ADAM/config0/*` to `HyperForensics-plus-plus/data/ADMM-ADAM/config0/config0.tar.gz`.
### 5.2 Compressing the specify direcotry
```bash
./zipping.sh --dir-path <path/to/directory>
```
Replace `<path/to/directory>` with the intended directory path. This will compress the directory and put it in the working directory.
For instance,
```bash
./zipping.sh --dir-path /path/to/local/ADMM-ADAM/config0
```
This will compress `/path/to/local/ADMM-ADAM/config0/*` to `HyperForensics-plus-plus/config0.tar.gz`.
### 5.3 Compress the entire hierarchy
```bash
./zipping.sh --all
```
This will compress the entire local data hierarchy seperately and automatically put the compressed files under hierarchy of the dataset repo.
> **Note** that I did not imiplement the method/config-specified compressing for **Origin** since it seems unlikely people needs to compress/decompress it regularly. You can either do it manually or use the path-specified compressing method and put it in to hierarchy by hand.
## 6. Create a Pull Request
Push your changes to created PR:
```bash
git push origin pr/<index>:refs/pr/<index>
```
Replace `<index>` with the pull request index. And click the ***Publish*** button on the buttom of the page.
> You can aso leave comment for your PR to showcase where you have modified.
## 6. Wait for merging the Pull Request
Once the pull request is published, the repository maintainer will review and merge it.
---
Thank you for contributing to the HyperForensics++ dataset!
## Divide my study into three main scope
1. The data generation
2. The quality matricies to determind the quality of our dataset
3. The forgery detection baseline models
### The data generation
#### Removal Method
This scope has been almost done by min-zuo with two algorithm 1. ADMM-ADAM; 2. FasyHyIn.
> However, since both strategy looks into spectral domain to inpaint the area. I have to look carefully to make sure it actually **remove** the data instead of recovering it.
> How to do that? I need a **classifier or segmentation model** to detect the image making sure it has a wrong output
#### Object replacement
This scope is under min-zuo's hand. He use CRS_diff diffusion model + harmonization to generate the replacemnt example.
:::warning
The images example above water are hard the current bottleneck.
Needs more attention
:::
#### Adv attack
This is implemented by 恩召, I did not actually ask much about it. Choose one adv-attack strategy. Focusing on highly [transibale](https://zhuanlan.zhihu.com/p/686734868) attack
### Qaulity matrics
#### Spectral Consistency Metrics
These metrics are used in HSI-tampering, which are used to ensure that the forged region's spectrum matches the surrounding context
1. SAM
2. Spectral RMSE
3. Pearson correlation (need more info)
4. EGRAS (need more info)
#### Forgery Detectability Metrics
I think this is the most important one since we are making forgery dataset. We want to show:
- rate of forgery images being mis-classified is high
1. FNR
2. Attack Success Rate
3. Detection Score Drop (detector confidence before/after forgery)
We can use either a classification model or segmentation model
- rate of forgeries not detected by a detection model is low
#### Visual Inspection
1. User studies - using false color?
2. Visual result figures (origin vs. forged vs. mask)
#### Visual Quality Matrics
This is relatively not important since this is a forgery dataset, how similar the forgery image is to the ground truth can not evaluate the goodness of the forgery
1. PSNR
2. SSIM
3. LPIPS(Learned Perceptual Image Patch Similarity)
### Forgery Detection Baseline Model
From previous paragraph [Forgery Detectability Metrics](#forgery-detectability-metrics). We need:
1. Pixel wise existing forgery detection model
2. Our own modern SOTA detection model for each attacking method |