code
stringlengths
3
6.57k
model.loss(e_inputs, d_inputs, grounds)
loss.backward()
optimizer.step()
splits_loss.item()
nucs_loss.item()
rels_loss.item()
loss.item()
writer.add_scalar("train/split_loss", log_splits_loss, niter)
writer.add_scalar("train/nuclear_loss", log_nucs_loss, niter)
writer.add_scalar("train/relation_loss", log_rels_loss, niter)
writer.add_scalar("train/loss", log_loss, niter)
parse_and_eval(cdtb.validate, model)
logging.info("validation on %d instances" % num_instances)
logging.info(gen_parse_report(*validate_scores)
writer.add_scalar("validate/span_f1", validate_scores[0][2], niter)
writer.add_scalar("validate/nuclear_f1", validate_scores[1][2], niter)
writer.add_scalar("validate/coarse_relation_f1", validate_scores[2][2], niter)
writer.add_scalar("validate/fine_relation_f1", validate_scores[3][2], niter)
model_score(validate_scores)
copy.deepcopy(model)
logging.info("test on new best model")
parse_and_eval(cdtb.test, best_model)
logging.info("test on %d instances" % num_instances)
logging.info(gen_parse_report(*test_scores)
writer.add_scalar("test/span_f1", test_scores[0][2], niter)
writer.add_scalar("test/nuclear_f1", test_scores[1][2], niter)
writer.add_scalar("test/coarse_relation_f1", test_scores[2][2], niter)
writer.add_scalar("test/fine_relation_f1", test_scores[3][2], niter)
logging.info("final test result")
parse_and_eval(cdtb.test, best_model)
logging.info("test on %d instances" % num_instances)
logging.info(gen_parse_report(*test_scores)
logging.info("save best model to %s" % args.model_save)
open(args.model_save, "wb+")
torch.save(best_model, model_fd)
writer.close()
logging.basicConfig(level=logging.INFO)
argparse.ArgumentParser()
arg_parser.add_argument("--data", default="data/CDTB")
arg_parser.add_argument("--ctb_dir", default="data/CTB")
arg_parser.add_argument("--cache_dir", default="data/cache")
arg_parser.add_argument("-hidden_size", default=512, type=int)
arg_parser.add_argument("-dropout", default=0.33, type=float)
arg_parser.add_mutually_exclusive_group(required=True)
arg_parser.add_argument("-pretrained", default="data/pretrained/sgns.renmin.word")
arg_parser.add_argument("-w2v_size", type=int)
arg_parser.add_argument("-pos_size", default=30, type=int)
arg_parser.add_argument("-split_mlp_size", default=64, type=int)
arg_parser.add_argument("-nuc_mlp_size", default=32, type=int)
arg_parser.add_argument("-rel_mlp_size", default=128, type=int)
arg_parser.add_argument("--w2v_freeze", dest="w2v_freeze", action="store_true")
arg_parser.set_defaults(w2v_freeze=True)
arg_parser.add_argument("-epoch", default=20, type=int)
arg_parser.add_argument("-batch_size", default=64, type=int)
arg_parser.add_argument("-lr", default=0.001, type=float)
arg_parser.add_argument("-l2", default=0.0, type=float)
arg_parser.add_argument("-log_every", default=10, type=int)
arg_parser.add_argument("-validate_every", default=10, type=int)
arg_parser.add_argument("-a_split_loss", default=0.3, type=float)
arg_parser.add_argument("-a_nuclear_loss", default=1.0, type=float)
arg_parser.add_argument("-a_relation_loss", default=1.0, type=float)
arg_parser.add_argument("-log_dir", default="data/log")
arg_parser.add_argument("-model_save", default="data/models/treebuilder.partptr.model")
arg_parser.add_argument("--seed", default=21, type=int)
arg_parser.add_argument("--use_gpu", dest="use_gpu", action="store_true")
arg_parser.set_defaults(use_gpu=True)
main(arg_parser.parse_args()
fileConfig(config.config_file_name)
logging.getLogger('alembic.env')
replace('%', '%%')
config.get_main_option("my_important_option")
run_migrations_offline()
context.execute()
config.get_main_option("sqlalchemy.url")
context.begin_transaction()
context.run_migrations()
run_migrations_online()
process_revision_directives(context, revision, directives)
getattr(config.cmd_opts, 'autogenerate', False)
script.upgrade_ops.is_empty()
logger.info('No changes in schema detected.')
config.get_section(config.config_ini_section)
connectable.connect()
context.begin_transaction()
context.run_migrations()
context.is_offline_mode()
run_migrations_offline()
run_migrations_online()
Copyright (c)
len(sys.argv)
len(sys.argv)
Database("numbers.sqlite")
conn.execute("DELETE FROM %s WHERE number=?" % SECONDARY,(NUMBER,)
conn.execute("DELETE FROM %s WHERE number=?" % PRIMARY, (NUMBER,)
conn.execute("INSERT OR REPLACE INTO %s(number)
VALUES (?)
conn.execute("SELECT DISTINCT p.number,h.name FROM %s AS p LEFT JOIN history AS h ON p.number=h.number" % PRIMARY)
os.path.abspath(os.path.dirname(__file__)
open(os.path.join(here, 'README.rst')
f.read()