text stringlengths 0 131 |
|---|
# data_transfer |
A Python tool to upload files and directories to cloud backends with a focus on speed and simplicity. |
Supported backends: |
- Hugging Face Hub |
- Google Drive |
Features: |
- Simple CLI and Python API |
- Resumable, chunked uploads |
- Parallel transfer where supported |
- Credential storage via file store at `~/.config/data_transfer/credentials.json` (macOS & Linux) |
- Configurable data root (default `./data`) and remote paths/filenames |
- Timing summary with `-v` |
## Quickstart |
1. Install |
```bash |
pip install -e . |
``` |
2. Configure credentials |
```bash |
# For Hugging Face |
data-transfer auth hf |
# For Google Drive (opens browser for OAuth once) |
# Optionally specify path to client_secrets.json |
# export GDRIVE_CLIENT_SECRETS=/path/to/client_secrets.json |
data-transfer auth gdrive |
``` |
3. Upload |
```bash |
# Upload a single file to Hugging Face Hub repo under path `datasets/mydir/file.bin` |
# By default source is resolved under ./data; override with --data-root |
data-transfer upload --backend hf --source test.pt --remote datasets/mydir/file.bin --repo your-username/your-repo --create-repo -v |
# Upload directory recursively to Google Drive under a folder id |
# Remote path supports nested folders; will be created on demand under the given folder id |
# Use --workers to speed up parallel uploads |
data-transfer upload --backend gdrive --source ./myfolder --remote nested1/nested2 --parent-id <FOLDER_ID> --workers 8 -v |
``` |
See `data_transfer/config.example.toml` for optional defaults. |
No dataset card yet
- Downloads last month
- 19