| # Script to benchmark inference performance, without bases precomputation | |
| # CLI args with defaults | |
| BATCH_SIZE=${1:-240} | |
| AMP=${2:-true} | |
| CUDA_VISIBLE_DEVICES=0 python -m se3_transformer.runtime.inference \ | |
| --amp "$AMP" \ | |
| --batch_size "$BATCH_SIZE" \ | |
| --use_layer_norm \ | |
| --norm \ | |
| --task homo \ | |
| --seed 42 \ | |
| --benchmark | |