File size: 8,768 Bytes
0ba52cd
 
 
 
 
 
 
 
 
 
 
 
7ffae14
0ba52cd
 
aa2bcc0
0ba52cd
7ffae14
 
 
 
47295d2
 
 
 
 
 
7ffae14
 
 
0ba52cd
aa2bcc0
7ffae14
 
 
 
 
aa2bcc0
7ffae14
 
 
aa2bcc0
7ffae14
f53d9a5
1378110
 
 
f53d9a5
 
1378110
f53d9a5
 
1378110
f53d9a5
 
 
 
 
 
 
 
 
 
 
0ba52cd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f53d9a5
7ffae14
0ba52cd
 
 
 
 
 
47295d2
 
 
 
 
 
f53d9a5
0ba52cd
 
 
aa2bcc0
0ba52cd
 
 
 
 
f53d9a5
0ba52cd
 
 
 
 
 
 
 
 
aa2bcc0
0ba52cd
 
 
aa2bcc0
 
f53d9a5
 
 
0ba52cd
f53d9a5
0ba52cd
f53d9a5
 
0ba52cd
f53d9a5
c6abc5d
0ba52cd
b82e466
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
# HyperForensics++ Dataset Contribution Guide

This guide outlines the steps to contribute to the HyperForensics++ dataset repository.

---

## 1. Clone the Repository
First, clone the repository from Hugging Face to your local machine:
```bash
# Make sure git-lfs is installed (https://git-lfs.com)
git lfs install
git clone https://huggingface.co/datasets/OtoroLin/HyperForensics-plus-plus
cd ./HyperForensics-plus-plus
```
## 2. Extract the Dataset
Decompress the `.tar.gz` files into a separate local hierarchy. You can utilize the provided `unzipping.sh` script. First, modify the `ROOT_DIR` variable to the `local_data` path
```bash
# In zipping.sh
# Defined the root directory of the whole dataset
ROOT_DIR="/root/to/local_data"
```
Due to the redundancy of our dataset, we use **`pigz`** program instead of standar `gipz` program to compress the files. Make sure you install it first.
```bash
sudo apt install pigz
# If you are using miniconda
conda install pigz
```
### 2.1 Decompressing the specify method and configuration:
```bash
./unzipping.sh --method <method name> --config <config name>
```
Replace `<method name>` with the forgery method and `<config name>` with the configuration index. This will automatically decompressed `.tar.gz` file and put it under hierarchy of the local dataset directory.
For instance, if I want to decompress the images under `ADMM-ADAM` forgery method, `config0` configuration:
```bash
./unzipping.sh --method ADMM-ADAM --config 0
```
This will decompress `HyperForensics-plus-plus/data/ADMM-ADAM/config0/config0.tar.gz` to `/path/to/local/ADMM-ADAM/config0/*`.
### 2.2 Decompress the entire hierarchy
```bash
./unzipping.sh --all
```
This will recursively decompress all `tar.gz` file in the `HyperForensics-plus-plus/data` and automatically put the decompressed files under hierarchy of the local dataset directory.

## 3. [Create a pull request](https://huggingface.co/docs/hub/repositories-pull-requests-discussions#pull-requests-advanced-usage)
The pull requests on Hugging Face do not use forks and branches, but instead custom “branches” called `refs` that are stored directly on the source repo.

The advantage of using custom `refs` (like `refs/pr/42` for instance) instead of branches is that they’re not fetched (by default) by people cloning the repo, but they can still be fetched on demand.
### 3.1 Create a new PR on the hugging face
Go to *community* $\to$ *New pull request* $\to$ *On your machine* $\to$ enter *Pull request title* $\to$ *Creat PR branch*

Then, you shall see a brand new PR page with the title of your PR on the top left. Make note of the index number next to the title `#{index}`.
> In this step, you actually create a new custom branch on **remote** under reference of `ref/pr/#{index}`.

The PR is still in **draft mode**, the maintainer can not merge the PR untill you publish the PR.
### 3.2 Create a new PR branch on local
```bash
git fetch origin refs/pr/<index>
git checkout pr/<index>
```
Fetch the remote PR branch `ref/pr/<index>` to local PR branch. Then checkout to the newly created custom branch under references `pr/`, and just for unambiguity, using the PR index as the name of that reference. You can actually create a new branch locally here as usual.
### 3.3 Push the PR branch
After you finish your modification, push the local PR branch to remote Hugging Face. Check it out [here at 6](#6-create-a-pull-request).

## 4. Modify or Generate Data
Make changes or generate new data in the extracted directory. Ensure the data follows the required structure:
```
local_data/
    |-{attacking method}/
    |   |-config0/
    |   |   |-images.npy
    |   |-config1/
    |   | ...
    |   |-config4/
    |-{attacking method}/
    |   |-config{0-4}/
    |-Origin
    |   |-images.npy
HyperForensics-plus-plus/
    |-data/
```
## 5. Zip the Directory
After modifying or generating data, zip the directory into `.tar.gz` files and place it in the repository.
You can utilize the provided `zipping.sh` script. First, modify the `ROOT_DIR` variable to the `local_data` path
```bash
# In zipping.sh
# Defined the root directory of the whole dataset
ROOT_DIR="/root/to/local_data"
```
Due to the redundancy of our dataset, we use **`pigz`** program instead of standar `gipz` program to compress the files. Make sure you install it first.
```bash
sudo apt install pigz
# If you are using miniconda
conda install pigz
```
### 5.1 Compressing the specify method and configuration:
```bash
./zipping.sh --method <method name> --config <config name>
```
Replace <method name> with the forgery method and <config name> with the configuration index. This will automatically put the compressed `.tar.gz` file under hierarchy of the dataset repo.
For instance, if I want to zip the images under `ADMM-ADAM` forgery method, `config0` configuration:
```bash
./zipping.sh --method ADMM-ADAM --config 0
```
This will compress `/path/to/local/ADMM-ADAM/config0/*` to `HyperForensics-plus-plus/data/ADMM-ADAM/config0/config0.tar.gz`.
### 5.2 Compressing the specify direcotry
```bash
./zipping.sh --dir-path <path/to/directory>
```
Replace `<path/to/directory>` with the intended directory path. This will compress the directory and put it in the working directory.
For instance,
```bash
./zipping.sh --dir-path /path/to/local/ADMM-ADAM/config0
```
This will compress `/path/to/local/ADMM-ADAM/config0/*` to `HyperForensics-plus-plus/config0.tar.gz`.
### 5.3 Compress the entire hierarchy
```bash
./zipping.sh --all
```
This will compress the entire local data hierarchy seperately and automatically put the compressed files under hierarchy of the dataset repo.
> **Note** that I did not imiplement the method/config-specified compressing for **Origin** since it seems unlikely people needs to compress/decompress it regularly. You can either do it manually or use the path-specified compressing method and put it in to hierarchy by hand.

## 6. Create a Pull Request
Push your changes to created PR:
```bash
git push origin pr/<index>:refs/pr/<index>
```
Replace `<index>` with the pull request index. And click the ***Publish*** button on the buttom of the page.
> You can aso leave comment for your PR to showcase where you have modified.
## 6. Wait for merging the Pull Request
Once the pull request is published, the repository maintainer will review and merge it.

---
Thank you for contributing to the HyperForensics++ dataset!

## Divide my study into three main scope
1. The data generation
2. The quality matricies to determind the quality of our dataset
3. The forgery detection baseline models
### The data generation
#### Removal Method
This scope has been almost done by min-zuo with two algorithm 1. ADMM-ADAM; 2. FasyHyIn.
> However, since both strategy looks into spectral domain to inpaint the area. I have to look carefully to make sure it actually **remove** the data instead of recovering it.
> How to do that? I need a **classifier or segmentation model** to detect the image making sure it has a wrong output
#### Object replacement
This scope is under min-zuo's hand. He use CRS_diff diffusion model + harmonization to generate the replacemnt example.
:::warning
The images example above water are hard the current bottleneck.
Needs more attention
:::
#### Adv attack
This is implemented by 恩召, I did not actually ask much about it. Choose one adv-attack strategy. Focusing on highly [transibale](https://zhuanlan.zhihu.com/p/686734868) attack

### Qaulity matrics
#### Spectral Consistency Metrics
These metrics are used in HSI-tampering, which are used to ensure that the forged region's spectrum matches the surrounding context
1. SAM
2. Spectral RMSE
3. Pearson correlation (need more info)
4. EGRAS (need more info)
#### Forgery Detectability Metrics
I think this is the most important one since we are making forgery dataset. We want to show:
- rate of forgery images being mis-classified is high
    1. FNR
    2. Attack Success Rate
    3. Detection Score Drop (detector confidence before/after forgery)
    We can use either a classification model or segmentation model
- rate of forgeries not detected by a detection model is low
#### Visual Inspection
1. User studies - using false color?
2. Visual result figures (origin vs. forged vs. mask)
#### Visual Quality Matrics
This is relatively not important since this is a forgery dataset, how similar the forgery image is to the ground truth can not evaluate the goodness of the forgery
1. PSNR
2. SSIM
3. LPIPS(Learned Perceptual Image Patch Similarity)
### Forgery Detection Baseline Model
From previous paragraph [Forgery Detectability Metrics](#forgery-detectability-metrics). We need:
1. Pixel wise existing forgery detection model
2. Our own modern SOTA detection model for each attacking method