| BASE_DIR=../../data/seq_truncated | |
| TASKS=("DST" "T2S" "SF" "RRR" "QCR" "NLI" "MRC" "MCQA" "DS" "DCRG" "ER" "ID" "DT" "CC" "CI" "ABSA") | |
| # TASKS=("DT") | |
| TOKENIZER_PATH="../../ckpts/t5-base" | |
| OUTPUT_DIR=../../data | |
| DATASETS=() | |
| for TASK in ${TASKS[*]}; do | |
| for dataset in `ls ${BASE_DIR} | grep "^${TASK}-"`; do | |
| DATASETS+=(${BASE_DIR}/${dataset}) | |
| done | |
| done | |
| python get_statistics.py \ | |
| --input-dir-list ${DATASETS[@]} \ | |
| --tokenizer-path ${TOKENIZER_PATH} \ | |
| --output-path ${OUTPUT_DIR} \ | |