| # Export Data outside of JZ | |
| ## Upload to the Hub | |
| First go to https://huggingface.co/bigscience/ and via your username (right upper corner) create "new Model" | |
| while choosing the `bigscience` as org. | |
| Say you created https://huggingface.co/bigscience/misc-test-data/ | |
| Now on JZ side | |
| ``` | |
| module load git-lfs | |
| git lfs install | |
| git clone https://huggingface.co/bigscience/misc-test-data/ | |
| cd misc-test-data/ | |
| ``` | |
| Now you can add files which are less than 10M, commit and push. | |
| Make sure that if the file is larger than 10M its extension is tracked by git LFS, e.g. if you're adding `foo.tar.gz` make sure `*gz` is in `.gitattributes` like so: | |
| ``` | |
| *.gz filter=lfs diff=lfs merge=lfs -text | |
| ``` | |
| if it isn't add it: | |
| ``` | |
| git lfs track "*.gz" | |
| git commit -m "compressed files" .gitattributes | |
| git push | |
| ``` | |
| only now add your large file `foo.tar.gz` | |
| ``` | |
| cp /some/place/foo.tar.gz . | |
| git add foo.tar.gz | |
| git commit -m "foo.tar.gz" foo.tar.gz | |
| git push | |
| ``` | |
| Now you can tell the contributor on the other side where they can download the files you have just uploaded by sending them to the corresponding hub repo. | |
| ## Automated upload to the hub | |
| Once a repo has been cloned and is used as a destination for checkpoints and log files, the following process will automatically push any new files into it. | |
| 1. Auth. | |
| Typically you can skip directly to the stage 2 as stage 1 should already work. | |
| We use a shared auth file located at `$six_ALL_CCFRWORK/auth/.hub_info.json` for all processes syncing to the hub. We use a special account of the `bigscience-bot` user so that the process doesn't depend on personal accounts. | |
| If for some reason you need to override this shared file with a different auth data for a specific project, simply run: | |
| ``` | |
| tools/hub-auth.py | |
| ``` | |
| And enter login and password, and email, at prompt. This will create `tools/.hub_info.json` with the username, email and then auth token locally. | |
| 2. Now for each tracking repo, run the script with the desired pattern, e.g.: | |
| ``` | |
| module load git-lfs | |
| DATA_OUTPUT_PATH=$six_ALL_CCFRSCRATCH/checkpoints/tr1-13B | |
| CHECKPOINT_PATH=$DATA_OUTPUT_PATH/checkpoints | |
| TENSORBOARD_PATH=$DATA_OUTPUT_PATH/tensorboard | |
| CODECARBON_PATH=$DATA_OUTPUT_PATH/codecarbon | |
| BIG_SCIENCE_REPO_PATH=$six_ALL_CCFRWORK/code/bigscience | |
| $BIG_SCIENCE_REPO_PATH/tools/hub-sync.py --repo-path $TENSORBOARD_PATH --patterns '*tfevents*' | |
| $BIG_SCIENCE_REPO_PATH/tools/hub-sync.py --repo-path $CODECARBON_PATH --patterns '*csv' | |
| $BIG_SCIENCE_REPO_PATH/tools/hub-sync.py --repo-path $CHECKPOINT_PATH --patterns '*pt' | |
| ``` | |
| Of course this needs to be automated, so we will create slurm jobs to perform all these. These must be run on the `prepost` partition, since it has a limited Internet access. | |
| ``` | |
| $ cat tr1-13B-hub-sync-tensorboard.slurm | |
| #!/bin/bash | |
| #SBATCH --job-name=tr1-13B-hub-sync-tensorboard # job name | |
| #SBATCH --ntasks=1 # number of MP tasks | |
| #SBATCH --nodes=1 # number of nodes | |
| #SBATCH --cpus-per-task=1 # number of cores per task | |
| #SBATCH --hint=nomultithread # we get physical cores not logical | |
| #SBATCH --time=20:00:00 # maximum execution time (HH:MM:SS) | |
| #SBATCH --output=%x-%j.out # output file name | |
| #SBATCH --partition=prepost | |
| echo "START TIME: $(date)" | |
| module load git-lfs | |
| DATA_OUTPUT_PATH=$six_ALL_CCFRSCRATCH/checkpoints/tr1-13B | |
| TENSORBOARD_PATH=$DATA_OUTPUT_PATH/tensorboard | |
| BIG_SCIENCE_REPO_PATH=$six_ALL_CCFRWORK/code/bigscience | |
| $BIG_SCIENCE_REPO_PATH/tools/hub-sync.py --repo-path $TENSORBOARD_PATH --patterns '*tfevents*' -d | |
| echo "END TIME: $(date)" | |
| ``` | |
| XXX: create a slurm script for codecarbon when it starts operating | |
| XXX: create a slurm script for checkpoints once we figure out how to share those | |
| XXX: concern: if this is run from `cron.hourly` what if the first `git push` is still uploading when the next round is pushed? | |
| ## Large Text files | |
| Normally `*txt` files aren't LFS tracked, so if your log file gets synced to he hub an it has grown over 10M you will get the next push fail with: | |
| ``` | |
| * Pushing 1 files | |
| remote: ------------------------------------------------------------------------- | |
| remote: Your push was rejected because it contains files larger than 10M. | |
| remote: Please use https://git-lfs.github.com/ to store larger files. | |
| remote: ------------------------------------------------------------------------- | |
| remote: Offending files: | |
| remote: - logs/main_log.txt (ref: refs/heads/main) | |
| To https://huggingface.co/bigscience/tr3n-1B3-pile-fancy-logs | |
| ! [remote rejected] main -> main (pre-receive hook declined) | |
| error: failed to push some refs to 'https://bigscience-bot:api_gyGezHBUDEGfyBxlAYTHCxQIbkjMUUEpaK@huggingface.co/bigscience/tr3n-1B3-pile-fancy-logs' | |
| ``` | |
| So you need to do the following from the cloned repo dir in question: | |
| 1. Unstage the commits that weren't pushed: | |
| ``` | |
| git reset --soft origin/HEAD | |
| ``` | |
| 2. Add `*txt` to LFS-tracking | |
| ``` | |
| git lfs track "**.txt" | |
| gc -am text .gitattributes | |
| ``` | |
| this will automatically switch to LFS on the next commit | |
| 3. commit/push normally | |
| ``` | |
| git commit -m "update file" logs/main_log.txt | |
| git push | |
| ``` | |
| In order to avoid this issue in the first place, it's best to set it up to: | |
| ``` | |
| git lfs track "**.txt" | |
| gc -am text .gitattributes | |
| ``` | |
| when you first setup the repo clone. | |