Datasets:
testing (#1)
Browse files- extend contributor_README (1378110021a7783e2f68b0452a2acf853b942a35)
- init branch commit (4fcb9f32a4d370bfa00d0967b08bf781de8629f6)
- contributor_README.md +8 -1
- unzipping.sh +1 -1
contributor_README.md
CHANGED
|
@@ -41,6 +41,13 @@ This will decompress `HyperForensics-plus-plus/data/ADMM-ADAM/config0/config0.ta
|
|
| 41 |
```
|
| 42 |
This will recursively decompress all `tar.gz` file in the `HyperForensics-plus-plus/data` and automatically put the decompressed files under hirarchy of the local dataset directory.
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
## 3. Modify or Generate Data
|
| 45 |
Make changes or generate new data in the extracted directory. Ensure the data follows the required structure:
|
| 46 |
```
|
|
@@ -97,7 +104,7 @@ This will compress `/path/to/local/ADMM-ADAM/config0/*` to `HyperForensics-plus-
|
|
| 97 |
./zipping.sh --all
|
| 98 |
```
|
| 99 |
This will compress the entire local data hirarchy seperately and automatically put the compressed files under hirarchy of the dataset repo.
|
| 100 |
-
> Note that I
|
| 101 |
## 5. Create a Pull Request
|
| 102 |
Push your changes to create a pull request. Hugging Face uses `git push` to implement pull requests:
|
| 103 |
```bash
|
|
|
|
| 41 |
```
|
| 42 |
This will recursively decompress all `tar.gz` file in the `HyperForensics-plus-plus/data` and automatically put the decompressed files under hirarchy of the local dataset directory.
|
| 43 |
|
| 44 |
+
## 3. Create a pull request
|
| 45 |
+
The pull requests on Hugging Face do not use forks and branches, but instead custom “branches” called `refs` that are stored directly on the source repo.
|
| 46 |
+
|
| 47 |
+
The advantage of using custom `refs` (like `refs/pr/42` for instance) instead of branches is that they’re not fetched (by default) by people cloning the repo, but they can still be fetched on demand.
|
| 48 |
+
### .1 Create a new PR on the hugging face
|
| 49 |
+
|
| 50 |
+
|
| 51 |
## 3. Modify or Generate Data
|
| 52 |
Make changes or generate new data in the extracted directory. Ensure the data follows the required structure:
|
| 53 |
```
|
|
|
|
| 104 |
./zipping.sh --all
|
| 105 |
```
|
| 106 |
This will compress the entire local data hirarchy seperately and automatically put the compressed files under hirarchy of the dataset repo.
|
| 107 |
+
> **Note** that I did not imiplement the method/config-specified compressing for **Origin** since it seems unlikely people needs to compress/decompress it regularly. You can either do it manually or use the path-specified compressing method and put it in to hirarchy by hand.
|
| 108 |
## 5. Create a Pull Request
|
| 109 |
Push your changes to create a pull request. Hugging Face uses `git push` to implement pull requests:
|
| 110 |
```bash
|
unzipping.sh
CHANGED
|
@@ -7,7 +7,7 @@ if ! command -v pigz &> /dev/null; then
|
|
| 7 |
fi
|
| 8 |
|
| 9 |
# Defined the root directory of the whole dataset
|
| 10 |
-
ROOT_DIR="
|
| 11 |
|
| 12 |
# Function to display usage
|
| 13 |
usage() {
|
|
|
|
| 7 |
fi
|
| 8 |
|
| 9 |
# Defined the root directory of the whole dataset
|
| 10 |
+
ROOT_DIR="../data_test"
|
| 11 |
|
| 12 |
# Function to display usage
|
| 13 |
usage() {
|