code stringlengths 3 6.57k |
|---|
label.after(3000,clear_label) |
window.mainloop() |
ctypes.windll.shell32.ShellExecuteW(None, "runas", sys.executable, __file__, None, 1) |
is_admin() |
open('error.log','a+') |
error_file.write(e+'\n') |
error_file.close() |
open("README.md","r") |
fh.read() |
folder (MyLib) |
logging.basicConfig(level=logging.INFO) |
logging.getLogger('text2dict') |
safe_pickle(obj, filename) |
os.path.isfile(filename) |
logger.info("Overwriting %s." % filename) |
logger.info("Saving to %s." % filename) |
open(filename, 'wb') |
cPickle.dump(obj, f, protocol=cPickle.HIGHEST_PROTOCOL) |
_itersplit(l, splitters) |
current.append(item) |
magicsplit(l, *splitters) |
_itersplit(l, splitters) |
argparse.ArgumentParser() |
parser.add_argument("input", type=str, help="Binarized dialogue corpus (pkl file) |
parser.add_argument("token_id", type=int, help="Token index to split examples by (e.g. to split by end-of-dialogue set this to 2) |
parser.add_argument("consecutive_examples_to_merge", type=int, default='1', help="After splitting these number of examples will be merged.") |
parser.add_argument("output", type=str, help="Filename of processed binarized dialogue corpus (pkl file) |
parser.parse_args() |
os.path.isfile(args.input) |
Exception("Input file not found!") |
logger.info("Loading dialogue corpus") |
cPickle.load(open(args.input, 'r') |
len(data) |
logger.info('Corpus loaded... Data len is %d' % data_len) |
range(data_len) |
len(data[i]) |
logger.info('Tokens count %d' % tokens_count) |
logger.info("Splitting corpus examples by token id... ") |
range(data_len) |
logger.info(' Example %d ' % i) |
magicsplit(data[i], int(args.token_id) |
len(new_examples) |
len(new_examples) |
len(new_examples) |
len(new_examples) |
processed_binarized_corpus.append(new_example + [int(args.token_id) |
int(math.floor(len(new_examples) |
range(1, s) |
reversed(range(args.consecutive_examples_to_merge) |
int(args.token_id) |
processed_binarized_corpus.append(merged_example) |
range((s-1) |
len(new_examples) |
int(args.token_id) |
processed_binarized_corpus.append(merged_example) |
range(len(new_examples) |
int(args.token_id) |
processed_binarized_corpus.append(merged_example) |
logger.info('New data len is %d' % len(processed_binarized_corpus) |
range(len(processed_binarized_corpus) |
len(processed_binarized_corpus[i]) |
logger.info('New tokens count %d' % processed_tokens_count) |
assert(processed_tokens_count == tokens_count) |
logger.info("Reshuffling corpus.") |
numpy.random.RandomState(13248) |
rng.shuffle(processed_binarized_corpus) |
logger.info("Saving corpus.") |
safe_pickle(processed_binarized_corpus, args.output + ".pkl") |
logger.info("Corpus saved. All done!") |
SystemsRolesShow(SystemsFormatOne, ServiceIdentifier, Username) |
get_parser(self, prog_name) |
super(SystemsRolesShow, self) |
get_parser(prog_name) |
ServiceIdentifier.extend_parser(self, parser) |
Username.extend_parser(self, parser) |
take_action(self, parsed_args) |
self.preprocess_args(parsed_args) |
self.requests_client.setup(API_NAME, SERVICE_VERSION) |
self.update_payload(parsed_args) |
self.render_headers(SystemRole, parsed_args) |
self.render_value(rec.get(key, None) |
data.append(val) |
return (tuple(headers) |
tuple(data) |
TestExample(TestCase) |
test_example(self) |
UserFactory() |
self.client.get('/') |
self.assertEqual(resp.status_code, 200) |
TestParent(unittest.TestCase) |
setUp(self) |
tearDown(self) |
testParent(self) |
petstore_api.Parent() |
unittest.main() |
testmod.build_model( ['Quantization'] , ['LinearTrend'] , ['Seasonal_Hour'] , ['SVR'] ) |
sorted() |
sorted() |
multiprocessing.Pool() |
map() |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.