code
stringlengths
3
6.57k
label.after(3000,clear_label)
window.mainloop()
ctypes.windll.shell32.ShellExecuteW(None, "runas", sys.executable, __file__, None, 1)
is_admin()
open('error.log','a+')
error_file.write(e+'\n')
error_file.close()
open("README.md","r")
fh.read()
folder (MyLib)
logging.basicConfig(level=logging.INFO)
logging.getLogger('text2dict')
safe_pickle(obj, filename)
os.path.isfile(filename)
logger.info("Overwriting %s." % filename)
logger.info("Saving to %s." % filename)
open(filename, 'wb')
cPickle.dump(obj, f, protocol=cPickle.HIGHEST_PROTOCOL)
_itersplit(l, splitters)
current.append(item)
magicsplit(l, *splitters)
_itersplit(l, splitters)
argparse.ArgumentParser()
parser.add_argument("input", type=str, help="Binarized dialogue corpus (pkl file)
parser.add_argument("token_id", type=int, help="Token index to split examples by (e.g. to split by end-of-dialogue set this to 2)
parser.add_argument("consecutive_examples_to_merge", type=int, default='1', help="After splitting these number of examples will be merged.")
parser.add_argument("output", type=str, help="Filename of processed binarized dialogue corpus (pkl file)
parser.parse_args()
os.path.isfile(args.input)
Exception("Input file not found!")
logger.info("Loading dialogue corpus")
cPickle.load(open(args.input, 'r')
len(data)
logger.info('Corpus loaded... Data len is %d' % data_len)
range(data_len)
len(data[i])
logger.info('Tokens count %d' % tokens_count)
logger.info("Splitting corpus examples by token id... ")
range(data_len)
logger.info(' Example %d ' % i)
magicsplit(data[i], int(args.token_id)
len(new_examples)
len(new_examples)
len(new_examples)
len(new_examples)
processed_binarized_corpus.append(new_example + [int(args.token_id)
int(math.floor(len(new_examples)
range(1, s)
reversed(range(args.consecutive_examples_to_merge)
int(args.token_id)
processed_binarized_corpus.append(merged_example)
range((s-1)
len(new_examples)
int(args.token_id)
processed_binarized_corpus.append(merged_example)
range(len(new_examples)
int(args.token_id)
processed_binarized_corpus.append(merged_example)
logger.info('New data len is %d' % len(processed_binarized_corpus)
range(len(processed_binarized_corpus)
len(processed_binarized_corpus[i])
logger.info('New tokens count %d' % processed_tokens_count)
assert(processed_tokens_count == tokens_count)
logger.info("Reshuffling corpus.")
numpy.random.RandomState(13248)
rng.shuffle(processed_binarized_corpus)
logger.info("Saving corpus.")
safe_pickle(processed_binarized_corpus, args.output + ".pkl")
logger.info("Corpus saved. All done!")
SystemsRolesShow(SystemsFormatOne, ServiceIdentifier, Username)
get_parser(self, prog_name)
super(SystemsRolesShow, self)
get_parser(prog_name)
ServiceIdentifier.extend_parser(self, parser)
Username.extend_parser(self, parser)
take_action(self, parsed_args)
self.preprocess_args(parsed_args)
self.requests_client.setup(API_NAME, SERVICE_VERSION)
self.update_payload(parsed_args)
self.render_headers(SystemRole, parsed_args)
self.render_value(rec.get(key, None)
data.append(val)
return (tuple(headers)
tuple(data)
TestExample(TestCase)
test_example(self)
UserFactory()
self.client.get('/')
self.assertEqual(resp.status_code, 200)
TestParent(unittest.TestCase)
setUp(self)
tearDown(self)
testParent(self)
petstore_api.Parent()
unittest.main()
testmod.build_model( ['Quantization'] , ['LinearTrend'] , ['Seasonal_Hour'] , ['SVR'] )
sorted()
sorted()
multiprocessing.Pool()
map()