kai-2054's picture
Initial commit: add code
cb0ad2d
import os
import torch
from libs.utils.vocab import Vocab
device = torch.device('cuda')
train_lrcs_path = [
"/yrfs1/intern/pfhu6/TSR/Dataset/SciTSR/train/table.lrc"
]
train_data_dir = ''
train_max_pixel_nums = 400 * 400 * 5
train_bucket_seps = (50, 50, 50)
train_max_batch_size = 8
train_num_workers = 4
valid_lrc_path = '/yrfs1/intern/pfhu6/TSR/Dataset/SciTSR/test/table.lrc'
valid_data_dir = ''
valid_num_workers = 0
valid_batch_size = 1
vocab = Vocab()
# model params
# backbone
arch = "res34"
pretrained_backbone = True
backbone_out_channels = (64, 128, 256, 512)
# fpn
fpn_out_channels = 256
# pan
pan_num_levels = 4
pan_in_dim = 256
pan_out_dim = 256
# row segment predictor
rs_scale = 1
# col segment predictor
cs_scale = 1
# divide predictor
dp_head_nums = 8
dp_scale = 1
# cells extractor params
ce_scale = 1 / 8
ce_pool_size = (3, 3)
ce_dim = 512
ce_head_nums = 8
ce_heads = 1
# decoder
embed_dim = 512
feat_dim = 512
lm_state_dim = 512
proj_dim = 512
cover_kernel = 7
att_threshold = 0.5
spatial_att_weight_loss_wight = 1.0
# train params
base_lr = 0.0001
min_lr = 1e-6
weight_decay = 0
num_epochs = 20
sync_rate = 20
log_sep = 20
work_dir = './experiments/heads_1'
train_checkpoint = None
eval_checkpoint = os.path.join(work_dir, 'best_f1_model.pth')