code stringlengths 3 6.57k |
|---|
self._urlopener.open('http://icanhazip.com') |
res.read() |
decode('utf-8') |
self.err(ex) |
_get_local_ip(self) |
socket.socket(socket.AF_INET, socket.SOCK_DGRAM) |
s.connect(("8.8.8.8", 80) |
s.getsockname() |
self.err(ex) |
License (MIT) |
main() |
ArgumentParser(add_help=False) |
parser.add_argument('-c', '--config_path', type=str, default='./src/configs/CIFAR10/ContraGAN.json') |
parser.add_argument('--checkpoint_folder', type=str, default=None) |
parser.add_argument('-current', '--load_current', action='store_true', help='whether you load the current or best checkpoint') |
parser.add_argument('--log_output_path', type=str, default=None) |
parser.add_argument('-DDP', '--distributed_data_parallel', action='store_true') |
parser.add_argument('-n', '--nodes', default=1, type=int, metavar='N') |
parser.add_argument('-nr', '--nr', default=0, type=int, help='ranking within the nodes') |
parser.add_argument('--seed', type=int, default=-1, help='seed for generating random numbers') |
parser.add_argument('--num_workers', type=int, default=8, help='') |
parser.add_argument('-sync_bn', '--synchronized_bn', action='store_true', help='whether turn on synchronized batchnorm') |
parser.add_argument('-mpc', '--mixed_precision', action='store_true', help='whether turn on mixed precision training') |
parser.add_argument('-LARS', '--LARS_optimizer', action='store_true', help='whether turn on LARS optimizer') |
parser.add_argument('-rm_API', '--disable_debugging_API', action='store_true', help='whether disable pytorch autograd debugging mode') |
parser.add_argument('--reduce_train_dataset', type=float, default=1.0, help='control the number of train dataset') |
parser.add_argument('--truncated_factor', type=float, default=-1.0, help='factor for truncation trick') |
parser.add_argument('-stat_otf', '--bn_stat_OnTheFly', action='store_true', help='when evaluating, use the statistics of a batch') |
parser.add_argument('-std_stat', '--standing_statistics', action='store_true') |
parser.add_argument('--standing_step', type=int, default=-1, help='# of steps for accumulation batchnorm') |
parser.add_argument('--freeze_layers', type=int, default=-1, help='# of layers for freezing discriminator') |
parser.add_argument('-l', '--load_all_data_in_memory', action='store_true') |
parser.add_argument('-t', '--train', action='store_true') |
parser.add_argument('-e', '--eval', action='store_true') |
parser.add_argument('-s', '--save_images', action='store_true') |
parser.add_argument('-iv', '--image_visualization', action='store_true', help='select whether conduct image visualization') |
parser.add_argument('-knn', '--k_nearest_neighbor', action='store_true', help='select whether conduct k-nearest neighbor analysis') |
parser.add_argument('-itp', '--interpolation', action='store_true', help='whether conduct interpolation analysis') |
parser.add_argument('-fa', '--frequency_analysis', action='store_true', help='whether conduct frequency analysis') |
parser.add_argument('-tsne', '--tsne_analysis', action='store_true', help='whether conduct tsne analysis') |
parser.add_argument('--nrow', type=int, default=10, help='number of rows to plot image canvas') |
parser.add_argument('--ncol', type=int, default=8, help='number of cols to plot image canvas') |
parser.add_argument('--print_every', type=int, default=100, help='control log interval') |
parser.add_argument('--save_every', type=int, default=2000, help='control evaluation and save interval') |
parser.add_argument('--eval_type', type=str, default='test', help='[train/valid/test]') |
update_parser_defaults_from_yaml(parser=parser) |
parser.parse_args() |
parser.print_help(sys.stderr) |
sys.exit(1) |
open(args.config_path) |
json.load(f) |
vars(args) |
make_hdf5(model_configs['data_processing'], train_configs, mode="train") |
random.randint(1,4096) |
fix_all_seed(train_configs['seed']) |
torch.cuda.device_count() |
torch.cuda.current_device() |
warnings.warn('You have chosen a specific GPU. This will completely disable data parallelism.') |
make_run_name(RUN_NAME_FORMAT, framework=train_configs['config_path'].split('/') |
torch.autograd.set_detect_anomaly(False) |
check_flags(train_configs, model_configs, world_size) |
print("Train the models through DistributedDataParallel (DDP) |
prepare_train_eval(rank, gpus_per_node, world_size, run_name, train_configs, model_configs, hdf5_path_train=hdf5_path_train) |
main() |
scan
(i.e., for which it doesn't have a specific scanner in its dictionary) |
TestSCons.TestSCons() |
open(sys.argv[1], 'w') |
open(infile, 'r') |
process(infp, outfp, include_prefix=include_prefix) |
infp.readlines() |
len(include_prefix) |
len(include_prefix) |
open(file, 'r') |
process(f, outfp) |
outfp.write(line) |
process(ifp, ofp) |
sys.exit(0) |
SConscript('SConscript') |
re.compile(r'^include1\s+(\S+) |
re.compile(r'^include2\s+(\S+) |
re.compile(r'^include3\s+(\S+) |
k1_scan(node, env, scanpaths, arg=None) |
node.get_text_contents() |
include1_re.findall(contents) |
k2_scan(node, env, scanpaths, arg=None) |
node.get_text_contents() |
include2_re.findall(contents) |
k3_scan(node, env, scanpaths, arg=None) |
node.get_text_contents() |
include3_re.findall(contents) |
Scanner({'.k1' : Scanner(k1_scan) |
Scanner(k2_scan) |
Builder(action=r'%(_python_) |
Environment(BUILDERS={'Build':b}) |
kscanner.add_scanner('.k3', Scanner(k3_scan) |
env.Build('aaa', 'aaa.k1') |
env.Build('bbb', 'bbb.k2') |
env.Build('ccc', 'ccc.k3') |
env.Build('ddd', ['ddd.k4', 'aaa.k1', 'bbb.k2', 'ccc.k3']) |
locals() |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.