code stringlengths 3 6.57k |
|---|
model.loss(e_inputs, d_inputs, grounds) |
loss.backward() |
optimizer.step() |
splits_loss.item() |
nucs_loss.item() |
rels_loss.item() |
loss.item() |
writer.add_scalar("train/split_loss", log_splits_loss, niter) |
writer.add_scalar("train/nuclear_loss", log_nucs_loss, niter) |
writer.add_scalar("train/relation_loss", log_rels_loss, niter) |
writer.add_scalar("train/loss", log_loss, niter) |
parse_and_eval(cdtb.validate, model) |
logging.info("validation on %d instances" % num_instances) |
logging.info(gen_parse_report(*validate_scores) |
writer.add_scalar("validate/span_f1", validate_scores[0][2], niter) |
writer.add_scalar("validate/nuclear_f1", validate_scores[1][2], niter) |
writer.add_scalar("validate/coarse_relation_f1", validate_scores[2][2], niter) |
writer.add_scalar("validate/fine_relation_f1", validate_scores[3][2], niter) |
model_score(validate_scores) |
copy.deepcopy(model) |
logging.info("test on new best model") |
parse_and_eval(cdtb.test, best_model) |
logging.info("test on %d instances" % num_instances) |
logging.info(gen_parse_report(*test_scores) |
writer.add_scalar("test/span_f1", test_scores[0][2], niter) |
writer.add_scalar("test/nuclear_f1", test_scores[1][2], niter) |
writer.add_scalar("test/coarse_relation_f1", test_scores[2][2], niter) |
writer.add_scalar("test/fine_relation_f1", test_scores[3][2], niter) |
logging.info("final test result") |
parse_and_eval(cdtb.test, best_model) |
logging.info("test on %d instances" % num_instances) |
logging.info(gen_parse_report(*test_scores) |
logging.info("save best model to %s" % args.model_save) |
open(args.model_save, "wb+") |
torch.save(best_model, model_fd) |
writer.close() |
logging.basicConfig(level=logging.INFO) |
argparse.ArgumentParser() |
arg_parser.add_argument("--data", default="data/CDTB") |
arg_parser.add_argument("--ctb_dir", default="data/CTB") |
arg_parser.add_argument("--cache_dir", default="data/cache") |
arg_parser.add_argument("-hidden_size", default=512, type=int) |
arg_parser.add_argument("-dropout", default=0.33, type=float) |
arg_parser.add_mutually_exclusive_group(required=True) |
arg_parser.add_argument("-pretrained", default="data/pretrained/sgns.renmin.word") |
arg_parser.add_argument("-w2v_size", type=int) |
arg_parser.add_argument("-pos_size", default=30, type=int) |
arg_parser.add_argument("-split_mlp_size", default=64, type=int) |
arg_parser.add_argument("-nuc_mlp_size", default=32, type=int) |
arg_parser.add_argument("-rel_mlp_size", default=128, type=int) |
arg_parser.add_argument("--w2v_freeze", dest="w2v_freeze", action="store_true") |
arg_parser.set_defaults(w2v_freeze=True) |
arg_parser.add_argument("-epoch", default=20, type=int) |
arg_parser.add_argument("-batch_size", default=64, type=int) |
arg_parser.add_argument("-lr", default=0.001, type=float) |
arg_parser.add_argument("-l2", default=0.0, type=float) |
arg_parser.add_argument("-log_every", default=10, type=int) |
arg_parser.add_argument("-validate_every", default=10, type=int) |
arg_parser.add_argument("-a_split_loss", default=0.3, type=float) |
arg_parser.add_argument("-a_nuclear_loss", default=1.0, type=float) |
arg_parser.add_argument("-a_relation_loss", default=1.0, type=float) |
arg_parser.add_argument("-log_dir", default="data/log") |
arg_parser.add_argument("-model_save", default="data/models/treebuilder.partptr.model") |
arg_parser.add_argument("--seed", default=21, type=int) |
arg_parser.add_argument("--use_gpu", dest="use_gpu", action="store_true") |
arg_parser.set_defaults(use_gpu=True) |
main(arg_parser.parse_args() |
fileConfig(config.config_file_name) |
logging.getLogger('alembic.env') |
replace('%', '%%') |
config.get_main_option("my_important_option") |
run_migrations_offline() |
context.execute() |
config.get_main_option("sqlalchemy.url") |
context.begin_transaction() |
context.run_migrations() |
run_migrations_online() |
process_revision_directives(context, revision, directives) |
getattr(config.cmd_opts, 'autogenerate', False) |
script.upgrade_ops.is_empty() |
logger.info('No changes in schema detected.') |
config.get_section(config.config_ini_section) |
connectable.connect() |
context.begin_transaction() |
context.run_migrations() |
context.is_offline_mode() |
run_migrations_offline() |
run_migrations_online() |
Copyright (c) |
len(sys.argv) |
len(sys.argv) |
Database("numbers.sqlite") |
conn.execute("DELETE FROM %s WHERE number=?" % SECONDARY,(NUMBER,) |
conn.execute("DELETE FROM %s WHERE number=?" % PRIMARY, (NUMBER,) |
conn.execute("INSERT OR REPLACE INTO %s(number) |
VALUES (?) |
conn.execute("SELECT DISTINCT p.number,h.name FROM %s AS p LEFT JOIN history AS h ON p.number=h.number" % PRIMARY) |
os.path.abspath(os.path.dirname(__file__) |
open(os.path.join(here, 'README.rst') |
f.read() |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.