cluster-scripts / setup_cluster.sh
szang18's picture
Upload setup_cluster.sh with huggingface_hub
9b87049 verified
#!/bin/zsh
# Cluster environment setup for GR00T N1.7 training
# Usage: source setup_cluster.sh
# --- Redirect caches to scratch ---
export HOME=/var/tmp/szang
export HF_HOME=/var/tmp/szang
export TORCH_HOME=/var/tmp/szang
export TRANSFORMERS_CACHE=/var/tmp/szang
export UV_CACHE_DIR=/var/tmp/szang/.cache/uv
mkdir -p /var/tmp/szang/.triton/autotune
mkdir -p /var/tmp/szang/.cache/torch/kernels
mkdir -p $UV_CACHE_DIR
# --- Load modules (compatible with zsh) ---
if typeset -f module > /dev/null 2>&1; then
module load gcc/10.5.0 cuda/11.8.0
elif [[ -f /etc/profile.d/modules.sh ]]; then
source /etc/profile.d/modules.sh
module load gcc/10.5.0 cuda/11.8.0
fi
# --- Library paths ---
export LD_LIBRARY_PATH=/global/scratch/users/ghr/Projects/szang/envs/video_act/lib:$LD_LIBRARY_PATH
# --- NCCL settings ---
export NCCL_P2P_DISABLE=1
export NCCL_IB_DISABLE=1
export NCCL_DEBUG=INFO
export NCCL_NVLS_ENABLE=1
export NCCL_P2P_LEVEL=NVL
export NCCL_TIMEOUT=1200
# --- Training settings ---
export OMP_NUM_THREADS=4
export TOKENIZERS_PARALLELISM=false
# --- tmux ---
tmux set mouse on 2>/dev/null || true
# --- Activate venv ---
cd /global/scratch/users/ghr/Projects/szang/GR00TN1.7
source .venv/bin/activate
# --- Verify ---
nvcc --version && echo "CUDA_HOME: $CUDA_HOME"
python -c "
import torch
print('PyTorch version:', torch.__version__)
print('CUDA available:', torch.cuda.is_available())
print('CUDA version:', torch.version.cuda)
print('cuDNN version:', torch.backends.cudnn.version())
print('GPU count:', torch.cuda.device_count())
for i in range(torch.cuda.device_count()):
print(f' GPU {i}: {torch.cuda.get_device_name(i)} ({torch.cuda.get_device_properties(i).total_memory / 1024**3:.1f} GB)')
print('Current device:', torch.cuda.current_device() if torch.cuda.is_available() else 'N/A')
"
python -c "import transformers; print('Transformers version:', transformers.__version__)"
python -c "import flash_attn; print('Flash Attention version:', flash_attn.__version__)"