source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(zero_ok=False, more_ok=True, **kwargs): return _find_one(find_projects, zero_ok=zero_ok, more_ok=more_ok, **kwargs)
find_one_project
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(zero_ok=False, more_ok=True, **kwargs):\n return _find_one(find_projects, zero_ok=zero_ok, more_ok=more_ok, **kwargs)\n", "role": "user" }, { "c...
def x(zero_ok=False, more_ok=True, **kwargs): return _find_one(find_apps, zero_ok=zero_ok, more_ok=more_ok, **kwargs)
find_one_app
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(zero_ok=False, more_ok=True, **kwargs):\n return _find_one(find_apps, zero_ok=zero_ok, more_ok=more_ok, **kwargs)\n", "role": "user" }, { "conte...
def x(api_method, org_id, query, first_page_size=100): if 'limit' not in query: query['limit'] = min(first_page_size, 1000) while True: resp = api_method(org_id, query) for result in resp['results']: yield result if resp['next'] is not None: query['starting'] = resp['next'] query['limit'] = min(query['limit'] * 2, 1000) else: break
org_find
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(api_method, org_id, query, first_page_size=100):\n if 'limit' not in query:\n query['limit'] = min(first_page_size, 1000)\n while True:\n r...
def x(org_id=None, level=None, describe=False): query = {} if level is not None: query['level'] = level query['describe'] = describe return _org_find(dxpy.api.org_find_members, org_id, query)
org_find_members
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(org_id=None, level=None, describe=False):\n query = {}\n if level is not None:\n query['level'] = level\n query['describe'] = describe\n ret...
def x(org_id=None, name=None, name_mode='exact', ids=None, properties=None, tags=None, describe=False, public=None, created_after=None, created_before=None, region=None, containsPHI=None): query = {} if name is not None: if name_mode == 'exact': query['name'] = name elif name_mode == 'glob': query['name'] = {'glob': name} elif name_mode == 'regexp': query['name'] = {'regexp': name} else: raise DXError( 'org_find_projects: Unexpected value found for argument name_mode' ) if ids is not None: query['id'] = ids if properties is not None: if len(properties.keys()) == 1: query['properties'] = properties else: query['properties'] = {'$and': [{k: v} for k, v in properties. iteritems()]} if tags is not None: if len(tags) == 1: query['tags'] = tags[0] else: query['tags'] = {'$and': tags} query['describe'] = describe if public is not None: query['public'] = public if created_after is not None or created_before is not None: query['created'] = {} if created_after is not None: query['created']['after'] = dxpy.utils.normalize_time_input( created_after) if created_before is not None: query['created']['before'] = dxpy.utils.normalize_time_input( created_before) if region is not None: query['region'] = region if containsPHI is not None: query['containsPHI'] = containsPHI return _org_find(dxpy.api.org_find_projects, org_id, query)
org_find_projects
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(org_id=None, name=None, name_mode='exact', ids=None, properties=None,\n tags=None, describe=False, public=None, created_after=None,\n created_before=None...
def x(org_id, name=None, name_mode='exact', category=None, all_versions= None, published=None, created_by=None, developer=None, authorized_user= None, created_after=None, created_before=None, modified_after=None, modified_before=None, describe=False, limit=None, return_handler=False, first_page_size=100, **kwargs): query = {} if name is not None: if name_mode == 'exact': query['name'] = name elif name_mode == 'glob': query['name'] = {'glob': name} elif name_mode == 'regexp': query['name'] = {'regexp': name} else: raise DXError( 'find_apps: Unexpected value found for argument name_mode') if category is not None: query['category'] = category if all_versions is not None: query['allVersions'] = all_versions if published is not None: query['published'] = published if created_by is not None: query['createdBy'] = created_by if developer is not None: query['developer'] = developer if authorized_user is not None: query['authorizedUser'] = authorized_user if modified_after is not None or modified_before is not None: query['modified'] = {} if modified_after is not None: query['modified']['after'] = dxpy.utils.normalize_time_input( modified_after) if modified_before is not None: query['modified']['before'] = dxpy.utils.normalize_time_input( modified_before) if created_after is not None or created_before is not None: query['created'] = {} if created_after is not None: query['created']['after'] = dxpy.utils.normalize_time_input( created_after) if created_before is not None: query['created']['before'] = dxpy.utils.normalize_time_input( created_before) if describe is not None and describe is not False: query['describe'] = describe if limit is not None: query['limit'] = limit return _org_find(dxpy.api.org_find_apps, org_id, query)
org_find_apps
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(org_id, name=None, name_mode='exact', category=None, all_versions=\n None, published=None, created_by=None, developer=None, authorized_user=\n None, crea...
def x(query, first_page_size=10): return _find(dxpy.api.system_find_orgs, query, limit=None, return_handler=False, first_page_size=first_page_size)
find_orgs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(query, first_page_size=10):\n return _find(dxpy.api.system_find_orgs, query, limit=None,\n return_handler=False, first_page_size=first_page_size)\n",...
def x(idir, dirs): file_load_utils.ensure_dir(idir) for d in dirs: file_load_utils.ensure_dir(os.path.join(idir, d))
create_dirs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(idir, dirs):\n file_load_utils.ensure_dir(idir)\n for d in dirs:\n file_load_utils.ensure_dir(os.path.join(idir, d))\n", "role": "user" }, ...
def x(filtered_inputs): file_key_descs, _ignore = file_load_utils.analyze_bash_vars(file_load_utils .get_input_json_file(), None) flattened_dict = {} def x(key, value, dict_): if key not in dict_: dict_[key] = value for input_ in filtered_inputs: if input_ not in file_key_descs: continue input_var_dict = file_key_descs[input_] add_if_no_collision(input_ + '_path', input_var_dict['path'], flattened_dict) add_if_no_collision(input_ + '_name', input_var_dict['basename'], flattened_dict) add_if_no_collision(input_ + '_prefix', input_var_dict['prefix'], flattened_dict) return flattened_dict
gen_helper_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filtered_inputs):\n file_key_descs, _ignore = file_load_utils.analyze_bash_vars(file_load_utils\n .get_input_json_file(), None)\n flattened_dict =...
def x(max_threads, num_cores, mem_available_mb): return min(max_threads, num_cores, max(int(mem_available_mb / 1200), 1))
get_num_parallel_threads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_threads, num_cores, mem_available_mb):\n return min(max_threads, num_cores, max(int(mem_available_mb / 1200), 1))\n", "role": "user" }, { "c...
def x(exclude=None, parallel=False, max_threads=8): idir = file_load_utils.get_input_dir() try: job_input_file = file_load_utils.get_input_json_file() dirs, inputs, rest = file_load_utils.get_job_input_filenames( job_input_file) except IOError: msg = 'Error: Could not find the input json file: {0}.\n'.format( job_input_file) msg += ( ' This function should only be called from within a running job.' ) print(msg) raise dirs_to_create = [] for d in dirs: keep = True if exclude is not None and d is not None: if d.split('/')[0] in exclude: keep = False if keep: dirs_to_create.append(d) _create_dirs(idir, dirs_to_create) if exclude: inputs = file_load_utils.filter_dict(inputs, exclude) to_download = [] for ival_list in inputs.values(): to_download.extend(ival_list) if parallel: total_mem = psutil.virtual_memory().total >> 20 num_cores = multiprocessing.cpu_count() max_num_parallel_downloads = _get_num_parallel_threads(max_threads, num_cores, total_mem) sys.stderr.write('Downloading files using {} threads'.format( max_num_parallel_downloads)) _parallel_file_download(to_download, idir, max_num_parallel_downloads) else: _sequential_file_download(to_download, idir) helper_vars = _gen_helper_dict(inputs) return helper_vars
download_all_inputs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(exclude=None, parallel=False, max_threads=8):\n idir = file_load_utils.get_input_dir()\n try:\n job_input_file = file_load_utils.get_input_json_fi...
def x(string, ignored_prefix=' '): paragraphs = string.split('\n\n') refilled_paragraphs = [(fill(paragraph) if not paragraph.startswith( ignored_prefix) else paragraph) for paragraph in paragraphs] return '\n\n'.join(refilled_paragraphs).strip('\n')
refill_paragraphs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, ignored_prefix=' '):\n paragraphs = string.split('\\n\\n')\n refilled_paragraphs = [(fill(paragraph) if not paragraph.startswith(\n ign...
def x(args, results): if args.json: print(json.dumps(list(results), indent=4)) elif args.brief: for result in results: print(result['id']) elif args.func.__name__ in ('find_projects', 'org_find_projects'): _format_find_projects_results(results) elif args.func.__name__ in 'org_find_members': _format_find_org_members_results(results) elif args.func.__name__ in 'org_find_apps': _format_find_apps_results(results, verbose=args.verbose)
format_find_results
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, results):\n if args.json:\n print(json.dumps(list(results), indent=4))\n elif args.brief:\n for result in results:\n print...
def x(**kwargs): draw_lines() draw_histogram() draw_bar_chart() destination = '-r /report' if use_html: generate_html() command = 'dx-build-report-html {h} {d}'.format(h=html_filename, d= destination) else: command = 'dx-build-report-html {l} {b} {h} {d}'.format(l= lines_filename, b=bars_filename, h=histogram_filename, d= destination) sub_output = json.loads(subprocess.check_output(command, shell=True)) output = {} output['report'] = dxpy.dxlink(sub_output['recordId']) return output
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n draw_lines()\n draw_histogram()\n draw_bar_chart()\n destination = '-r /report'\n if use_html:\n generate_html()\n co...
def x(): mu, sigma = 100, 15 x = mu + sigma * numpy.random.randn(10000) fig = pyplot.figure() ax = fig.add_subplot(111) n, bins, patches = ax.hist(x, 50, normed=1, facecolor='green', alpha=0.75) ax.set_xlabel('Random number generated') ax.set_ylabel('Probability') ax.set_xlim(40, 160) ax.set_ylim(0, 0.03) ax.grid(True) pyplot.savefig(histogram_filename)
draw_histogram
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n mu, sigma = 100, 15\n x = mu + sigma * numpy.random.randn(10000)\n fig = pyplot.figure()\n ax = fig.add_subplot(111)\n n, bins, patches = a...
def x(): N = 5 menMeans = 20, 35, 30, 35, 27 menStd = 2, 3, 4, 1, 2 ind = numpy.arange(N) width = 0.35 fig = pyplot.figure() ax = fig.add_subplot(111) rects1 = ax.bar(ind, menMeans, width, color='r', yerr=menStd) womenMeans = 25, 32, 34, 20, 25 womenStd = 3, 5, 2, 3, 3 rects2 = ax.bar(ind + width, womenMeans, width, color='y', yerr=womenStd) ax.set_ylabel('Scores') ax.set_title('Scores by group and color') ax.set_xticks(ind + width) ax.set_xticklabels(('G1', 'G2', 'G3', 'G4', 'G5')) ax.legend((rects1[0], rects2[0]), ('Red', 'Yellow')) def x(rects): for rect in rects: height = rect.get_height() ax.text(rect.get_x() + rect.get_width() / 2.0, 1.05 * height, '%d' % int(height), ha='center', va='bottom') autolabel(rects1) autolabel(rects2) pyplot.savefig(bars_filename)
draw_bar_chart
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n N = 5\n menMeans = 20, 35, 30, 35, 27\n menStd = 2, 3, 4, 1, 2\n ind = numpy.arange(N)\n width = 0.35\n fig = pyplot.figure()\n ax = ...
def x(): r = numpy.random.randn(200) fig = pyplot.figure() ax = fig.add_subplot(111) ax.plot(r) ax.grid(True) pyplot.savefig(lines_filename)
draw_lines
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n r = numpy.random.randn(200)\n fig = pyplot.figure()\n ax = fig.add_subplot(111)\n ax.plot(r)\n ax.grid(True)\n pyplot.savefig(lines_file...
def x(): html_file = open(html_filename, 'w') html_file.write('<html><body>') html_file.write('<h1>Here are some graphs for you!</h1>') for image in [lines_filename, bars_filename, histogram_filename]: html_file.write("<div><h2>{0}</h2><img src='{0}' /></div>".format( image)) html_file.write('</body></html>') html_file.close()
generate_html
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n html_file = open(html_filename, 'w')\n html_file.write('<html><body>')\n html_file.write('<h1>Here are some graphs for you!</h1>')\n for image...
def x(o): def x(s): return u''.join([(c if c in ASCII_PRINTABLE else '?') for c in s]) if USING_PYTHON2: try: return unicode(o) except: try: s = str(o) try: return s.decode('utf-8') except: return clean(s[:2048] ) + u' [Raw error message: ' + unicode(s.encode( 'hex'), 'utf-8') + u']' except: return u'(Unable to decode Python exception message)' else: return str(o)
safe_unicode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(o):\n\n def x(s):\n return u''.join([(c if c in ASCII_PRINTABLE else '?') for c in s])\n if USING_PYTHON2:\n try:\n return unico...
def x(e): if isinstance(e, dxpy.AppError): return _safe_unicode(e) if USING_PYTHON2: return unicode(e.__class__.__name__, 'utf-8') + ': ' + _safe_unicode(e) else: return e.__class__.__name__ + ': ' + _safe_unicode(e)
format_exception_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(e):\n if isinstance(e, dxpy.AppError):\n return _safe_unicode(e)\n if USING_PYTHON2:\n return unicode(e.__class__.__name__, 'utf-8') + ': '...
def x(function_name=None, function_input=None): global RUN_COUNT RUN_COUNT += 1 dx_working_dir = os.getcwd() if dxpy.JOB_ID is not None: logging.basicConfig() try: logging.getLogger().addHandler(dxpy.DXLogHandler()) except dxpy.exceptions.DXError: print( 'TODO: FIXME: the EE client should die if logging is not available' ) job = dxpy.describe(dxpy.JOB_ID) else: if function_name is None: function_name = os.environ.get('DX_TEST_FUNCTION', 'main') if function_input is None: with open('job_input.json', 'r') as fh: function_input = json.load(fh) job = {'function': function_name, 'input': function_input} with open('job_error_reserved_space', 'w') as fh: fh.write( """This file contains reserved space for writing job errors in case the filesystem becomes full. """ + ' ' * 1024 * 64) print('Invoking', job.get('function'), 'with', job.get('input')) try: result = ENTRY_POINT_TABLE[job['function']](**job['input']) except dxpy.AppError as e: save_error(e, dx_working_dir, error_type='AppError') raise except Exception as e: save_error(e, dx_working_dir) raise if result is not None: os.chdir(dx_working_dir) if USING_PYTHON2: with open('job_output.json', 'wb') as fh: json.dump(result, fh, indent=2, cls=DXJSONEncoder) fh.write(b'\n') else: with open('job_output.json', 'w') as fh: json.dump(result, fh, indent=2, cls=DXJSONEncoder) fh.write('\n') return result
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(function_name=None, function_input=None):\n global RUN_COUNT\n RUN_COUNT += 1\n dx_working_dir = os.getcwd()\n if dxpy.JOB_ID is not None:\n ...
def x(entry_point_name): def x(f): ENTRY_POINT_TABLE[entry_point_name] = f @wraps(f) def x(*args, **kwargs): return f(*args, **kwargs) return wrapped_f return wrap
entry_point
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(entry_point_name):\n\n def x(f):\n ENTRY_POINT_TABLE[entry_point_name] = f\n\n @wraps(f)\n def x(*args, **kwargs):\n return ...
def x(path): project, _folderpath, entity_result = try_call(resolve_existing_path, path, expected='entity') try: if entity_result is None or not entity_result['id'].startswith( 'workflow-'): raise DXCLIError('Could not resolve "' + path + '" to a workflow object') except: err_exit() return entity_result['id'], project
get_workflow_id_and_project
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path):\n project, _folderpath, entity_result = try_call(resolve_existing_path,\n path, expected='entity')\n try:\n if entity_result is None...
def x(authserver_host=None, authserver_port=None): authserver = get_auth_server_name(authserver_host, authserver_port) return DXHTTPRequest(authserver + '/system/getUserInfo', {}, prepend_srv=False)
user_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(authserver_host=None, authserver_port=None):\n authserver = get_auth_server_name(authserver_host, authserver_port)\n return DXHTTPRequest(authserver + '/...
def x(job_homedir=None): if job_homedir is not None: home_dir = job_homedir else: home_dir = environ.get('HOME') idir = os.path.join(home_dir, 'in') return idir
get_input_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_homedir=None):\n if job_homedir is not None:\n home_dir = job_homedir\n else:\n home_dir = environ.get('HOME')\n idir = os.path.join...
def x(job_homedir=None): if job_homedir is not None: home_dir = job_homedir else: home_dir = environ.get('HOME') odir = os.path.join(home_dir, 'out') return odir
get_output_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_homedir=None):\n if job_homedir is not None:\n home_dir = job_homedir\n else:\n home_dir = environ.get('HOME')\n odir = os.path.join...
def x(): path = get_output_json_file() try: os.remove(path) except OSError as e: if e.errno == errno.ENOENT: pass else: raise
rm_output_json_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n path = get_output_json_file()\n try:\n os.remove(path)\n except OSError as e:\n if e.errno == errno.ENOENT:\n pass\n ...
def x(path): if not os.path.exists(path): os.mkdir(path) elif os.path.isfile(path): raise Exception( 'Path %s already exists, and it is a file, not a directory' % path)
ensure_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path):\n if not os.path.exists(path):\n os.mkdir(path)\n elif os.path.isfile(path):\n raise Exception(\n 'Path %s already exists...
def x(fname): bad_filenames = ['.', '..'] if fname in bad_filenames: raise DXError('Invalid filename {}'.format(fname)) return fname.replace('/', '%2F')
make_unix_filename
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fname):\n bad_filenames = ['.', '..']\n if fname in bad_filenames:\n raise DXError('Invalid filename {}'.format(fname))\n return fname.replace(...
def x(job_input_file): def x(): with open(job_input_file) as fh: job_input = json.load(fh) return job_input job_input = get_input_hash() files = collections.defaultdict(list) dirs = [] def x(iname, subdir, value): if not dxpy.is_dxlink(value): return handler = dxpy.get_handler(value) if not isinstance(handler, dxpy.DXFile): return filename = make_unix_filename(handler.name) trg_dir = iname if subdir is not None: trg_dir = os.path.join(trg_dir, subdir) files[iname].append({'trg_fname': os.path.join(trg_dir, filename), 'handler': handler, 'src_file_id': handler.id}) dirs.append(trg_dir) def x(input_name, links): num_files = len(links) if num_files == 0: return num_digits = len(str(num_files - 1)) dirs.append(input_name) for i, link in enumerate(links): subdir = str(i).zfill(num_digits) add_file(input_name, subdir, link) for input_name, value in list(job_input.items()): if isinstance(value, list): add_file_array(input_name, value) else: add_file(input_name, None, value) rest_hash = {key: val for key, val in list(job_input.items()) if key not in files} return dirs, files, rest_hash
get_job_input_filenames
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_input_file):\n\n def x():\n with open(job_input_file) as fh:\n job_input = json.load(fh)\n return job_input\n job_input ...
def x(): input_spec = None if 'DX_JOB_ID' in environ: job_desc = dxpy.describe(dxpy.JOB_ID) if job_desc['function'] == 'main': desc = dxpy.describe(job_desc.get('app', job_desc.get('applet'))) if 'inputSpec' in desc: input_spec = desc['inputSpec'] elif 'DX_TEST_DXAPP_JSON' in environ: path_to_dxapp_json = environ['DX_TEST_DXAPP_JSON'] with open(path_to_dxapp_json) as fd: dxapp_json = json.load(fd) input_spec = dxapp_json.get('inputSpec') if input_spec is None: return {} def x(pattern): return '*' in pattern patterns_dict = {} for spec in input_spec: name = spec['name'] if 'patterns' in spec: patterns_dict[name] = [] for p in spec['patterns']: if is_legal_pattern(p): patterns_dict[name].append(p) return patterns_dict
get_input_spec_patterns
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n input_spec = None\n if 'DX_JOB_ID' in environ:\n job_desc = dxpy.describe(dxpy.JOB_ID)\n if job_desc['function'] == 'main':\n ...
def x(job_input_file, job_homedir): _, file_entries, rest_hash = get_job_input_filenames(job_input_file) patterns_dict = get_input_spec_patterns() def x(basename, key): best_prefix = None patterns = patterns_dict.get(key) if patterns is not None: for pattern in patterns: if fnmatch.fnmatch(basename, pattern): _, _, right_piece = pattern.rpartition('*') best_prefix = choose_shorter_string(best_prefix, basename[:-len(right_piece)]) if best_prefix is not None: return best_prefix else: parts = os.path.splitext(basename) if parts[1] == '.gz': parts = os.path.splitext(parts[0]) return parts[0] def x(): return {'handler': [], 'basename': [], 'prefix': [], 'path': []} file_key_descs = collections.defaultdict(factory) rel_home_dir = get_input_dir(job_homedir) for key, entries in list(file_entries.items()): for entry in entries: filename = entry['trg_fname'] basename = os.path.basename(filename) prefix = get_prefix(basename, key) k_desc = file_key_descs[key] k_desc['handler'].append(entry['handler']) k_desc['basename'].append(basename) k_desc['prefix'].append(prefix) k_desc['path'].append(os.path.join(rel_home_dir, filename)) return file_key_descs, rest_hash
analyze_bash_vars
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_input_file, job_homedir):\n _, file_entries, rest_hash = get_job_input_filenames(job_input_file)\n patterns_dict = get_input_spec_patterns()\n\n d...
def x(job_input_file, job_homedir=None, check_name_collision=True): file_key_descs, rest_hash = analyze_bash_vars(job_input_file, job_homedir) def x(elem): result = None if isinstance(elem, basestring): result = elem elif isinstance(elem, dxpy.DXFile): result = json.dumps(dxpy.dxlink(elem)) else: result = json.dumps(elem) return pipes.quote(result) def x(val): if isinstance(val, list): string = ' '.join([string_of_elem(vitem) for vitem in val]) return '( {} )'.format(string) else: return string_of_elem(val) var_defs_hash = {} def x(key, val): if check_name_collision: if key not in environ and key not in var_defs_hash: var_defs_hash[key] = val else: sys.stderr.write(dxpy.utils.printing.fill( 'Creating environment variable ({}) would cause a name collision' .format(key)) + '\n') else: var_defs_hash[key] = val for key, desc in list(rest_hash.items()): gen_text_line_and_name_collision(key, string_of_value(desc)) for file_key, desc in list(file_key_descs.items()): gen_text_line_and_name_collision(file_key, string_of_value(desc[ 'handler'])) gen_text_line_and_name_collision(file_key + '_name', string_of_value(desc['basename'])) gen_text_line_and_name_collision(file_key + '_prefix', string_of_value(desc['prefix'])) gen_text_line_and_name_collision(file_key + '_path', string_of_value(desc['path'])) return var_defs_hash
gen_bash_vars
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job_input_file, job_homedir=None, check_name_collision=True):\n file_key_descs, rest_hash = analyze_bash_vars(job_input_file, job_homedir)\n\n def x(elem...
def x(futures, print_traceback=False): while True: try: future = next(concurrent.futures.as_completed(futures, timeout= THREAD_TIMEOUT_MAX)) break except concurrent.futures.TimeoutError: pass except KeyboardInterrupt: if print_traceback: traceback.print_stack() else: print('') os._exit(os.EX_IOERR) return future
wait_for_a_future
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(futures, print_traceback=False):\n while True:\n try:\n future = next(concurrent.futures.as_completed(futures, timeout=\n T...
def x(futures, print_traceback=False): try: while True: waited_futures = concurrent.futures.wait(futures, timeout=60) if len(waited_futures.not_done) == 0: break except KeyboardInterrupt: if print_traceback: traceback.print_stack() else: print('') os._exit(os.EX_IOERR)
wait_for_all_futures
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(futures, print_traceback=False):\n try:\n while True:\n waited_futures = concurrent.futures.wait(futures, timeout=60)\n if len(...
def x(request_iterator, thread_pool, max_active_tasks=None, do_first_task_sequentially=True): tasks_in_progress = collections.deque() if max_active_tasks is None: max_active_tasks = cpu_count() def x(task_iterator, executor, futures_queue): retval = next(task_iterator, None) if retval is None: return False task_callable, task_args, task_kwargs = retval task_future = executor.submit(task_callable, *task_args, **task_kwargs) futures_queue.append(task_future) return True def x(tasks_in_progress): future = tasks_in_progress.popleft() try: result = future.result(timeout=THREAD_TIMEOUT_MAX) except KeyboardInterrupt: print('') os._exit(os.EX_IOERR) return result if do_first_task_sequentially: task_callable, task_args, task_kwargs = next(request_iterator) yield task_callable(*task_args, **task_kwargs) for _i in range(max_active_tasks): retval = submit_task(request_iterator, thread_pool, tasks_in_progress) if not retval: break while len(tasks_in_progress) > 0: result = next_result(tasks_in_progress) submit_task(request_iterator, thread_pool, tasks_in_progress) yield result del result
response_iterator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request_iterator, thread_pool, max_active_tasks=None,\n do_first_task_sequentially=True):\n tasks_in_progress = collections.deque()\n if max_active_ta...
def x(t, future=False, default_unit='ms'): error_msg = ( 'Error: Expected an int timestamp, a date format (e.g. YYYY-MM-DD), or an int with a single-letter suffix (s=seconds, m=minutes, h=hours, d=days, w=weeks, M=months, y=years; e.g. "-10d" indicates 10 days ago); but got {t}' ) if isinstance(t, basestring) and t.isdigit(): t = int(t) if isinstance(t, basestring): try: t = normalize_timedelta(t) except ValueError: try: t = int(time.mktime(dateutil.parser.parse(t).timetuple()) * 1000) assert t > 0 except (ValueError, OverflowError, AssertionError): raise ValueError(error_msg.format(t=t)) elif isinstance(t, numbers.Integral): units_multipliers = {'ms': 1, 's': 1000} if default_unit not in units_multipliers: raise ValueError("Expected default_unit to be one of 's' or 'ms'") t = t * units_multipliers[default_unit] else: raise ValueError(error_msg.format(t=t)) now = int(time.time() * 1000) if t < 0 or future and t < now: t += now return t
normalize_time_input
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(t, future=False, default_unit='ms'):\n error_msg = (\n 'Error: Expected an int timestamp, a date format (e.g. YYYY-MM-DD), or an int with a single-le...
def x(timedelta): try: return int(timedelta) * 1000 except ValueError as e: t, suffix = timedelta[:-1], timedelta[-1:] suffix_multipliers = {'s': 1000, 'm': 1000 * 60, 'h': 1000 * 60 * 60, 'd': 1000 * 60 * 60 * 24, 'w': 1000 * 60 * 60 * 24 * 7, 'M': 1000 * 60 * 60 * 24 * 30, 'y': 1000 * 60 * 60 * 24 * 365} if suffix not in suffix_multipliers: raise ValueError() return int(t) * suffix_multipliers[suffix]
normalize_timedelta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(timedelta):\n try:\n return int(timedelta) * 1000\n except ValueError as e:\n t, suffix = timedelta[:-1], timedelta[-1:]\n suffix_mu...
def x(ordered_pairs): d = {} for k, v in ordered_pairs: if k in d: raise ValueError('duplicate key: %r' % (k,)) else: d[k] = v return d
dict_raise_on_duplicates
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ordered_pairs):\n d = {}\n for k, v in ordered_pairs:\n if k in d:\n raise ValueError('duplicate key: %r' % (k,))\n else:\n ...
def x(input_params): input_cp = input_params.copy() if len(input_cp.get('nonce', '')) == 0: input_cp['nonce'] = str(Nonce()) return input_cp
update_nonce
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_params):\n input_cp = input_params.copy()\n if len(input_cp.get('nonce', '')) == 0:\n input_cp['nonce'] = str(Nonce())\n return input_cp\...
def x(json_spec, src_dir): if 'description' not in json_spec: readme_filename = None for filename in ('README.md', 'Readme.md', 'readme.md'): if os.path.exists(os.path.join(src_dir, filename)): readme_filename = filename break if readme_filename is not None: with open(os.path.join(src_dir, readme_filename)) as fh: json_spec['description'] = fh.read() if 'developerNotes' not in json_spec: for filename in ('README.developer.md', 'Readme.developer.md', 'readme.developer.md'): if os.path.exists(os.path.join(src_dir, filename)): with open(os.path.join(src_dir, filename)) as fh: json_spec['developerNotes'] = fh.read() break
inline_documentation_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, src_dir):\n if 'description' not in json_spec:\n readme_filename = None\n for filename in ('README.md', 'Readme.md', 'readme.md'):\...
def x(prefixed_name): assert prefixed_name.startswith('app-') or prefixed_name.startswith( 'globalworkflow-') if prefixed_name.partition('-')[0] == 'app': exception_type = dxpy.app_builder.AppBuilderException describe_method = dxpy.api.app_describe exception_msg = ( 'An app with the given name already exists and you are not a developer of that app' ) else: exception_type = dxpy.workflow_builder.WorkflowBuilderException describe_method = dxpy.api.global_workflow_describe exception_msg = ( 'A global workflow with the given name already exists and you are not a developer of that workflow' ) name_already_exists = True is_developer = False version = None executable_id = None FoundExecutable = collections.namedtuple('FoundExecutable', ['name', 'version', 'id']) try: describe_output = describe_method(prefixed_name, input_params={ 'fields': {'isDeveloperFor': True, 'version': True, 'id': True}}) is_developer = describe_output['isDeveloperFor'] version = describe_output['version'] executable_id = describe_output['id'] except dxpy.exceptions.DXAPIError as e: if e.name == 'ResourceNotFound': name_already_exists = False elif e.name == 'PermissionDenied': raise exception_type(exception_msg) else: raise e if not name_already_exists: return FoundExecutable(name=None, version=None, id=None) name_without_prefix = prefixed_name.partition('-')[2] if not is_developer: raise exception_type('You are not a developer for {n}'.format(n= name_without_prefix)) return FoundExecutable(name=name_without_prefix, version=version, id= executable_id)
verify_developer_rights
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prefixed_name):\n assert prefixed_name.startswith('app-') or prefixed_name.startswith(\n 'globalworkflow-')\n if prefixed_name.partition('-')[0] =...
def x(from_spec, from_command_line, builder_exception): if from_spec is None or from_command_line is None: return if set(from_spec) != set(from_command_line): raise builder_exception( "--region and the 'regionalOptions' key in the JSON file do not agree" )
assert_consistent_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(from_spec, from_command_line, builder_exception):\n if from_spec is None or from_command_line is None:\n return\n if set(from_spec) != set(from_co...
def x(exec_type, json_spec, executable_builder_exeception): reg_options_spec = json_spec.get('regionalOptions') json_fn = 'dxapp.json' if exec_type == 'app' else 'dxworkflow.json' if not isinstance(reg_options_spec, dict): raise executable_builder_exeception( "The field 'regionalOptions' in must be a mapping") if not reg_options_spec: raise executable_builder_exeception( "The field 'regionalOptions' in " + json_fn + ' must be a non-empty mapping') regional_options_list = list(reg_options_spec.items()) for region, opts_for_region in regional_options_list: if not isinstance(opts_for_region, dict): raise executable_builder_exeception( "The field 'regionalOptions['" + region + "']' in " + json_fn + ' must be a mapping') if set(opts_for_region.keys()) != set(regional_options_list[0][1]. keys()): if set(opts_for_region.keys()) - set(regional_options_list[0][1 ].keys()): with_key, without_key = region, regional_options_list[0][0] key_name = next(iter(set(opts_for_region.keys()) - set( regional_options_list[0][1].keys()))) else: with_key, without_key = regional_options_list[0][0], region key_name = next(iter(set(regional_options_list[0][1].keys() ) - set(opts_for_region.keys()))) raise executable_builder_exeception( 'All regions in regionalOptions must specify the same options; ' + '%s was given for %s but not for %s' % (key_name, with_key, without_key)) if exec_type == 'app': for key in opts_for_region: if key in json_spec.get('runSpec', {}): raise executable_builder_exeception(key + ' cannot be given in both runSpec and in regional options for ' + region)
assert_consistent_reg_options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(exec_type, json_spec, executable_builder_exeception):\n reg_options_spec = json_spec.get('regionalOptions')\n json_fn = 'dxapp.json' if exec_type == 'app...
def x(exec_type, json_spec, from_command_line, executable_builder_exeception): from_spec = json_spec.get('regionalOptions') if from_spec is not None: assert_consistent_reg_options(exec_type, json_spec, executable_builder_exeception) assert_consistent_regions(from_spec, from_command_line, executable_builder_exeception) enabled_regions = None if from_spec is not None: enabled_regions = from_spec.keys() elif from_command_line is not None: enabled_regions = from_command_line return enabled_regions
get_enabled_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(exec_type, json_spec, from_command_line, executable_builder_exeception):\n from_spec = json_spec.get('regionalOptions')\n if from_spec is not None:\n ...
def x(self, fields=None, **kwargs): describe_input = {} if fields is not None: describe_input['fields'] = fields self._desc = dxpy.api.analysis_describe(self._dxid, describe_input, ** kwargs) return self._desc
describe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fields=None, **kwargs):\n describe_input = {}\n if fields is not None:\n describe_input['fields'] = fields\n self._desc = dxpy.api.analys...
def x(self, tags, **kwargs): dxpy.api.analysis_add_tags(self._dxid, {'tags': tags}, **kwargs)
add_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tags, **kwargs):\n dxpy.api.analysis_add_tags(self._dxid, {'tags': tags}, **kwargs)\n", "role": "user" }, { "content": "add_tags", "ro...
def x(self, tags, **kwargs): dxpy.api.analysis_remove_tags(self._dxid, {'tags': tags}, **kwargs)
remove_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tags, **kwargs):\n dxpy.api.analysis_remove_tags(self._dxid, {'tags': tags}, **kwargs)\n", "role": "user" }, { "content": "remove_tags", ...
def x(self, properties, **kwargs): dxpy.api.analysis_set_properties(self._dxid, {'properties': properties}, **kwargs)
set_properties
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, properties, **kwargs):\n dxpy.api.analysis_set_properties(self._dxid, {'properties': properties},\n **kwargs)\n", "role": "user" }, { ...
def x(self, field, index=None, metadata=None): link = {'$dnanexus_link': {'analysis': self._dxid, 'field': field}} if index is not None: link['$dnanexus_link']['index'] = index if metadata is not None: link['$dnanexus_link']['metadata'] = metadata return link
get_output_ref
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, field, index=None, metadata=None):\n link = {'$dnanexus_link': {'analysis': self._dxid, 'field': field}}\n if index is not None:\n link['$dn...
def x(app_json, publish=False): for input_field in app_json.get('inputSpec', []): for suggestion in input_field.get('suggestions', []): if 'project' in suggestion: try: project = dxpy.api.project_describe(suggestion[ 'project'], {'permissions': True}) if 'PUBLIC' not in project['permissions'] and publish: logger.warn('Project {name} NOT PUBLIC!'.format( name=project['name'])) except dxpy.exceptions.DXAPIError as e: if e.code == 404: logger.warn( 'Suggested project {name} does not exist, or not accessible by user' .format(name=suggestion['project'])) if 'path' in suggestion: try: check_folder_exists(suggestion['project'], suggestion['path'], '') except ResolutionError as e: logger.warn( 'Folder {path} could not be found in project {project}' .format(path=suggestion['path'], project= suggestion['project'])) if '$dnanexus_link' in suggestion: if suggestion['$dnanexus_link'].startswith(('file-', 'record-') ): try: dnanexus_link = dxpy.describe(suggestion[ '$dnanexus_link']) except dxpy.exceptions.DXAPIError as e: if e.code == 404: raise dxpy.app_builder.AppBuilderException( 'Suggested object {name} could not be found' .format(name=suggestion['$dnanexus_link'])) except Exception as e: raise dxpy.app_builder.AppBuilderException(str(e)) if 'value' in suggestion and isinstance(suggestion['value'], dict): if '$dnanexus_link' in suggestion['value']: if isinstance(suggestion['value']['$dnanexus_link'], dict): if 'project' in suggestion['value']['$dnanexus_link']: try: dxpy.api.project_describe(suggestion[ 'value']['$dnanexus_link']['project']) except dxpy.exceptions.DXAPIError as e: if e.code == 404: logger.warn( 'Suggested project {name} does not exist, or not accessible by user' .format(name=suggestion['value'][ '$dnanexus_link']['project'])) elif isinstance(suggestion['value']['$dnanexus_link'], basestring): if suggestion['value']['$dnanexus_link'].startswith(( 'file-', 'record-')): try: dnanexus_link = dxpy.describe(suggestion[ 'value']['$dnanexus_link']) except dxpy.exceptions.DXAPIError as e: if e.code == 404: raise dxpy.app_builder.AppBuilderException( 'Suggested object {name} could not be found' .format(name=suggestion['value'][ '$dnanexus_link'])) except Exception as e: raise dxpy.app_builder.AppBuilderException(str (e))
check_suggestions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app_json, publish=False):\n for input_field in app_json.get('inputSpec', []):\n for suggestion in input_field.get('suggestions', []):\n if...
def x(dxapp_json_filename, mode): def x(dirname): for basename in ['README.md', 'Readme.md', 'readme.md']: if os.path.exists(os.path.join(dirname, basename)): return os.path.join(dirname, basename) return None app_spec = json.load(open(dxapp_json_filename)) dirname = os.path.basename(os.path.dirname(os.path.abspath( dxapp_json_filename))) if mode == 'app': if 'title' not in app_spec: logger.warn( 'app is missing a title, please add one in the "title" field of dxapp.json' ) if 'summary' in app_spec: if app_spec['summary'].endswith('.'): logger.warn( 'summary "%s" should be a short phrase not ending in a period' % (app_spec['summary'],)) else: logger.warn( 'app is missing a summary, please add one in the "summary" field of dxapp.json' ) readme_filename = _find_readme(os.path.dirname(dxapp_json_filename)) if 'description' in app_spec: if readme_filename: raise dxpy.app_builder.AppBuilderException( 'Description was provided both in Readme.md and in the "description" field of {file}. Please consolidate content in Readme.md and remove the "description" field.' .format(file=dxapp_json_filename)) if not app_spec['description'].strip().endswith('.'): logger.warn( '"description" field should be written in complete sentences and end with a period' ) elif readme_filename is None: logger.warn( 'app is missing a description, please supply one in README.md') if 'categories' in app_spec: for category in app_spec['categories']: if category not in APP_CATEGORIES: logger.warn('app has unrecognized category "%s"' % ( category,)) if category == 'Import': if 'title' in app_spec and not app_spec['title'].endswith( 'Importer'): logger.warn('title "%s" should end in "Importer"' % (app_spec['title'],)) if category == 'Export': if 'title' in app_spec and not app_spec['title'].endswith( 'Exporter'): logger.warn('title "%s" should end in "Exporter"' % (app_spec['title'],)) if 'name' in app_spec: if app_spec['name'] != app_spec['name'].lower(): logger.warn('name "%s" should be all lowercase' % (app_spec[ 'name'],)) if dirname != app_spec['name']: logger.warn( 'app name "%s" does not match containing directory "%s"' % (app_spec['name'], dirname)) else: logger.warn( 'app is missing a name, please add one in the "name" field of dxapp.json' ) if 'version' in app_spec: if not dxpy.executable_builder.GLOBAL_EXEC_VERSION_RE.match(app_spec ['version']): logger.warn( '"version" %s should be semver compliant (e.g. of the form X.Y.Z)' % (app_spec['version'],)) if 'inputSpec' in app_spec: for i, input_field in enumerate(app_spec['inputSpec']): if not re.match('^[a-zA-Z_][0-9a-zA-Z_]*$', input_field['name']): logger.error( 'input %d has illegal name "%s" (must match ^[a-zA-Z_][0-9a-zA-Z_]*$)' % (i, input_field['name'])) else: logger.warn( 'dxapp.json contains no input specification (inputSpec). Your applet will not be usable as an ' + 'app, runnable from the GUI, or composable using workflows.') if 'outputSpec' in app_spec: for i, output_field in enumerate(app_spec['outputSpec']): if not re.match('^[a-zA-Z_][0-9a-zA-Z_]*$', output_field['name']): logger.error( 'output %d has illegal name "%s" (must match ^[a-zA-Z_][0-9a-zA-Z_]*$)' % (i, output_field['name'])) else: logger.warn( 'dxapp.json contains no output specification (outputSpec). Your applet will not be usable as an ' + 'app, runnable from the GUI, or composable using workflows.')
lint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dxapp_json_filename, mode):\n\n def x(dirname):\n for basename in ['README.md', 'Readme.md', 'readme.md']:\n if os.path.exists(os.path.joi...
def x(code, lang, temp_dir, enforce=True): if lang == 'python2.7': temp_basename = 'inlined_code_from_dxapp_json.py' elif lang == 'bash': temp_basename = 'inlined_code_from_dxapp_json.sh' else: raise ValueError('lang must be one of "python2.7" or "bash"') with open(os.path.join(temp_dir, temp_basename), 'w') as ofile: ofile.write(code) _check_file_syntax(os.path.join(temp_dir, temp_basename), temp_dir, override_lang=lang, enforce=enforce)
check_syntax
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(code, lang, temp_dir, enforce=True):\n if lang == 'python2.7':\n temp_basename = 'inlined_code_from_dxapp_json.py'\n elif lang == 'bash':\n ...
def x(filename, temp_dir, override_lang=None, enforce=True): def x(filename): pyc_path = os.path.join(temp_dir, os.path.basename(filename) + '.pyc') try: if USING_PYTHON2: filename = filename.encode(sys.getfilesystemencoding()) py_compile.compile(filename, cfile=pyc_path, doraise=True) finally: try: os.unlink(pyc_path) except OSError: pass def x(filename): if platform.system() == 'Windows': logging.warn( 'Skipping bash syntax check due to unavailability of bash on Windows.' ) else: subprocess.check_output(['/bin/bash', '-n', filename], stderr= subprocess.STDOUT) if override_lang == 'python2.7': checker_fn = check_python elif override_lang == 'bash': checker_fn = check_bash elif filename.endswith('.py'): checker_fn = check_python elif filename.endswith('.sh'): checker_fn = check_bash else: return open(filename) try: checker_fn(filename) except subprocess.CalledProcessError as e: print(filename + ' has a syntax error! Interpreter output:', file= sys.stderr) for line in e.output.strip('\n').split('\n'): print(' ' + line.rstrip('\n'), file=sys.stderr) if enforce: raise DXSyntaxError(filename + ' has a syntax error') except py_compile.PyCompileError as e: print(filename + ' has a syntax error! Interpreter output:', file= sys.stderr) print(' ' + e.msg.strip(), file=sys.stderr) if enforce: raise DXSyntaxError(e.msg.strip())
check_file_syntax
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, temp_dir, override_lang=None, enforce=True):\n\n def x(filename):\n pyc_path = os.path.join(temp_dir, os.path.basename(filename) + '.pyc')\...
def x(src_dir, temp_dir, mode, enforce=True): _lint(os.path.join(src_dir, 'dxapp.json'), mode) manifest = json.load(open(os.path.join(src_dir, 'dxapp.json'))) if 'runSpec' in manifest: if 'interpreter' not in manifest['runSpec']: raise dxpy.app_builder.AppBuilderException( 'runSpec.interpreter field was not present') if 'release' not in manifest['runSpec' ] or 'distribution' not in manifest['runSpec']: warn_message = ( 'runSpec.distribution or runSpec.release was not present. These fields ' ) warn_message += ( 'will be required in a future version of the API. Recommended value ' ) warn_message += ( 'for distribution is "Ubuntu" and release - "14.04".') logger.warn(warn_message) if manifest['runSpec']['interpreter'] in ['python2.7', 'bash']: if 'file' in manifest['runSpec']: entry_point_file = os.path.abspath(os.path.join(src_dir, manifest['runSpec']['file'])) try: _check_file_syntax(entry_point_file, temp_dir, override_lang=manifest['runSpec']['interpreter'], enforce=enforce) except IOError as e: raise dxpy.app_builder.AppBuilderException( 'Could not open runSpec.file=%r. The problem was: %s' % (entry_point_file, e)) except DXSyntaxError: raise dxpy.app_builder.AppBuilderException( 'Entry point file %s has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.' % (entry_point_file,)) elif 'code' in manifest['runSpec']: try: _check_syntax(manifest['runSpec']['code'], manifest[ 'runSpec']['interpreter'], temp_dir, enforce=enforce) except DXSyntaxError: raise dxpy.app_builder.AppBuilderException( 'Code in runSpec.code has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.' ) cluster_bootstrap_scripts = ( _get_all_cluster_bootstrap_script_names(manifest)) for filename in cluster_bootstrap_scripts: abs_filename = os.path.abspath(os.path.join(src_dir, filename)) try: _check_file_syntax(abs_filename, temp_dir, override_lang=manifest['runSpec']['interpreter'], enforce=enforce) except IOError as e: raise dxpy.app_builder.AppBuilderException( 'Could not open cluster bootstrap script %r. The problem was: %s' % (abs_filename, e)) except DXSyntaxError: raise dxpy.app_builder.AppBuilderException( 'Code in cluster bootstrapScript %r has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.' % filename) if 'execDepends' in manifest['runSpec']: if not isinstance(manifest['runSpec']['execDepends'], list): raise dxpy.app_builder.AppBuilderException( 'Expected runSpec.execDepends to be an array. Rerun with --no-check-syntax to proceed anyway.' ) if not all(isinstance(dep, dict) for dep in manifest['runSpec'] ['execDepends']): raise dxpy.app_builder.AppBuilderException( 'Expected runSpec.execDepends to be an array of hashes. Rerun with --no-check-syntax to proceed anyway.' ) if any(dep.get('package_manager', 'apt') != 'apt' for dep in manifest['runSpec']['execDepends']): if not isinstance(manifest.get('access'), dict ) or 'network' not in manifest['access']: msg = '\n'.join([ 'runSpec.execDepends specifies non-APT dependencies, but no network access spec is given.' , 'Add {"access": {"network": ["*"]}} to allow dependencies to install.' , 'See https://wiki.dnanexus.com/Developer-Tutorials/Request-Additional-App-Resources#Network-Access.' , 'Rerun with --no-check-syntax to proceed anyway.']) raise dxpy.app_builder.AppBuilderException(msg) if 'authorizedUsers' in manifest: if not isinstance(manifest['authorizedUsers'], list) or isinstance( manifest['authorizedUsers'], basestring): raise dxpy.app_builder.AppBuilderException( 'Expected authorizedUsers to be a list of strings') for thing in manifest['authorizedUsers']: if thing != 'PUBLIC' and (not isinstance(thing, basestring) or not re.match('^(org-|user-)', thing)): raise dxpy.app_builder.AppBuilderException( 'authorizedUsers field contains an entry which is not either the string "PUBLIC" or a user or org ID' ) if 'pricingPolicy' in manifest: error_message = ( '"pricingPolicy" at the top level is not accepted. It must be specified ' ) error_message += ( 'under the "regionalOptions" field in all enabled regions of the app' ) raise dxpy.app_builder.AppBuilderException(error_message) files_with_problems = [] for dirpath, dirnames, filenames in os.walk(os.path.abspath(os.path. join(src_dir, 'resources'))): for filename in filenames: if not filename.startswith('._'): try: _check_file_syntax(os.path.join(dirpath, filename), temp_dir, enforce=True) except IOError as e: raise dxpy.app_builder.AppBuilderException( 'Could not open file in resources directory %r. The problem was: %s' % (os.path.join(dirpath, filename), e)) except DXSyntaxError: files_with_problems.append(os.path.join(dirpath, filename)) if files_with_problems: files_str = files_with_problems[0] if len(files_with_problems ) == 1 else files_with_problems[0] + ' and ' + str(len( files_with_problems) - 1) + ' other file' + ('s' if len( files_with_problems) > 2 else '') logging.warn('%s contained syntax errors, see above for details' % (files_str,))
verify_app_source_dir_impl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_dir, temp_dir, mode, enforce=True):\n _lint(os.path.join(src_dir, 'dxapp.json'), mode)\n manifest = json.load(open(os.path.join(src_dir, 'dxapp.json'...
def x(src_dir, mode, enforce=True): temp_dir = tempfile.mkdtemp(prefix='dx-build_tmp') try: _verify_app_source_dir_impl(src_dir, temp_dir, mode, enforce=enforce) finally: shutil.rmtree(temp_dir)
verify_app_source_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_dir, mode, enforce=True):\n temp_dir = tempfile.mkdtemp(prefix='dx-build_tmp')\n try:\n _verify_app_source_dir_impl(src_dir, temp_dir, mode, e...
def x(src_dir): if not os.path.isdir(src_dir): parser.error('%s is not a directory' % src_dir) if not os.path.exists(os.path.join(src_dir, 'dxapp.json')): raise dxpy.app_builder.AppBuilderException( 'Directory %s does not contain dxapp.json: not a valid DNAnexus app source directory' % src_dir) with open(os.path.join(src_dir, 'dxapp.json')) as app_desc: try: return json_load_raise_on_duplicates(app_desc) except Exception as e: raise dxpy.app_builder.AppBuilderException( 'Could not parse dxapp.json file as JSON: ' + str(e.args))
parse_app_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_dir):\n if not os.path.isdir(src_dir):\n parser.error('%s is not a directory' % src_dir)\n if not os.path.exists(os.path.join(src_dir, 'dxapp....
def x(args, extra_args): if not args.remote: try: output = build_and_upload_locally(args.src_dir, args.mode, overwrite=args.overwrite, archive=args.archive, publish= args.publish, destination_override=args.destination, version_override=args.version_override, bill_to_override= args.bill_to, use_temp_build_project=args. use_temp_build_project, do_parallel_build=args. parallel_build, do_version_autonumbering=args. version_autonumbering, do_try_update=args.update, dx_toolkit_autodep=args.dx_toolkit_autodep, do_check_syntax =args.check_syntax, ensure_upload=args.ensure_upload, force_symlinks=args.force_symlinks, dry_run=args.dry_run, confirm=args.confirm, return_object_dump=args.json, region= args.region, **extra_args) if output is not None and args.run is None: print(json.dumps(output)) except dxpy.app_builder.AppBuilderException as e: print('Error: %s' % (e.args,), file=sys.stderr) sys.exit(3) except dxpy.exceptions.DXAPIError as e: print('Error: %s' % (e,), file=sys.stderr) sys.exit(3) if args.dry_run: return None return output['id'] else: try: app_json = _parse_app_spec(args.src_dir) _check_suggestions(app_json, publish=args.publish) _verify_app_source_dir(args.src_dir, args.mode) if args.mode == 'app' and not args.dry_run: dxpy.executable_builder.verify_developer_rights('app-' + app_json['name']) except dxpy.app_builder.AppBuilderException as e: print('Error: %s' % (e.args,), file=sys.stderr) sys.exit(3) if args.dry_run: parser.error('--remote cannot be combined with --dry-run') if args.overwrite: parser.error('--remote cannot be combined with --overwrite/-f') if args.archive: parser.error('--remote cannot be combined with --archive/-a') if args.json: parser.error('--remote cannot be combined with --json') if not args.use_temp_build_project: parser.error( '--remote cannot be combined with --no-temp-build-project') if isinstance(args.region, list) and len(args.region) > 1: parser.error( '--region can only be specified once for remote builds') region = args.region[0] if args.region is not None else None more_kwargs = {} if args.version_override: more_kwargs['version_override'] = args.version_override if args.bill_to: more_kwargs['bill_to_override'] = args.bill_to if not args.version_autonumbering: more_kwargs['do_version_autonumbering'] = False if not args.update: more_kwargs['do_try_update'] = False if not args.parallel_build: more_kwargs['do_parallel_build'] = False if not args.check_syntax: more_kwargs['do_check_syntax'] = False return _build_app_remote(args.mode, args.src_dir, destination_override=args.destination, publish=args.publish, dx_toolkit_autodep=args.dx_toolkit_autodep, region=region, watch=args.watch, **more_kwargs)
build_app
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, extra_args):\n if not args.remote:\n try:\n output = build_and_upload_locally(args.src_dir, args.mode,\n overwrite=ar...
def x(**kwargs): if len(sys.argv) > 0: if sys.argv[0].endswith('dx-build-app'): logging.warn( 'Warning: dx-build-app has been replaced with "dx build --create-app". Please update your scripts.' ) elif sys.argv[0].endswith('dx-build-applet'): logging.warn( 'Warning: dx-build-applet has been replaced with "dx build". Please update your scripts.' ) exit(0)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n if len(sys.argv) > 0:\n if sys.argv[0].endswith('dx-build-app'):\n logging.warn(\n 'Warning: dx-build-app has ...
def x(self, dxid=None, name=None, alias=None): self._dxid = None self._name = None self._alias = None if dxid is not None: if name is not None or alias is not None: raise DXError( 'Did not expect name or alias to be given if dxid is given') verify_string_dxid(dxid, self._class) self._dxid = dxid elif name is not None: self._name = name if not isinstance(name, basestring): raise DXError('App name needs to be a string: %r' % (name,)) if alias is not None: if not isinstance(alias, basestring): raise DXError('App alias needs to be a string: %r' % (alias,)) self._alias = alias else: self._alias = 'default'
set_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dxid=None, name=None, alias=None):\n self._dxid = None\n self._name = None\n self._alias = None\n if dxid is not None:\n if name is no...
def x(self): if self._dxid is not None: return self._dxid else: return 'app-' + self._name + '/' + self._alias
get_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._dxid is not None:\n return self._dxid\n else:\n return 'app-' + self._name + '/' + self._alias\n", "role": "user" }...
def x(self, **kwargs): dx_hash = {} if 'applet' not in kwargs and 'initializeFrom' not in kwargs: raise DXError( '%s: One of the keyword arguments %s and %s is required' % ( self.__class__.__name__, 'applet', 'initializeFrom')) for field in ['version']: if field not in kwargs: raise DXError('%s: Keyword argument %s is required' % (self. __class__.__name__, field)) dx_hash[field] = kwargs[field] del kwargs[field] for field in ('initializeFrom', 'applet', 'name', 'title', 'summary', 'description', 'billing', 'access', 'resources'): if field in kwargs: dx_hash[field] = kwargs[field] del kwargs[field] if 'bill_to' in kwargs: dx_hash['billTo'] = kwargs['bill_to'] del kwargs['bill_to'] resp = dxpy.api.app_new(dx_hash, **kwargs) self.set_id(dxid=resp['id'])
new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n dx_hash = {}\n if 'applet' not in kwargs and 'initializeFrom' not in kwargs:\n raise DXError(\n '%s: One of the keyw...
def x(self, fields=None, **kwargs): describe_input = {} if fields: describe_input['fields'] = fields if self._dxid is not None: self._desc = dxpy.api.app_describe(self._dxid, input_params= describe_input, **kwargs) else: self._desc = dxpy.api.app_describe('app-' + self._name, alias=self. _alias, input_params=describe_input, **kwargs) return self._desc
describe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fields=None, **kwargs):\n describe_input = {}\n if fields:\n describe_input['fields'] = fields\n if self._dxid is not None:\n self...
def x(self, **kwargs): updates = {} for field in ('applet', 'billing', 'access', 'resources', 'details'): if field in kwargs: updates[field] = kwargs[field] del kwargs[field] if self._dxid is not None: resp = dxpy.api.app_update(self._dxid, input_params=updates, **kwargs) else: resp = dxpy.api.app_update('app-' + self._name, alias=self._alias, input_params=updates, **kwargs)
update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n updates = {}\n for field in ('applet', 'billing', 'access', 'resources', 'details'):\n if field in kwargs:\n updates...
def x(self, tags, **kwargs): if self._dxid is not None: return dxpy.api.app_add_tags(self._dxid, input_params={'tags': tags }, **kwargs) else: return dxpy.api.app_add_tags('app-' + self._name, alias=self._alias, input_params={'tags': tags}, **kwargs)
add_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tags, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_add_tags(self._dxid, input_params={'tags': tags\n }, **kwargs)\n...
def x(self, tags, **kwargs): if self._dxid is not None: return dxpy.api.app_remove_tags(self._dxid, input_params={'tags': tags}, **kwargs) else: return dxpy.api.app_remove_tags('app-' + self._name, alias=self. _alias, input_params={'tags': tags}, **kwargs)
remove_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tags, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_remove_tags(self._dxid, input_params={'tags':\n tags}, **kwargs)...
def x(self, **kwargs): if self._dxid is not None: return dxpy.api.app_install(self._dxid, **kwargs) else: return dxpy.api.app_install('app-' + self._name, alias=self._alias, **kwargs)
install
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_install(self._dxid, **kwargs)\n else:\n return dxpy.api.app_install('ap...
def x(self, **kwargs): if self._dxid is not None: return dxpy.api.app_uninstall(self._dxid, **kwargs) else: return dxpy.api.app_uninstall('app-' + self._name, alias=self. _alias, **kwargs)
uninstall
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_uninstall(self._dxid, **kwargs)\n else:\n return dxpy.api.app_uninstall...
def x(self, **kwargs): if self._dxid is not None: return dxpy.api.app_get(self._dxid, **kwargs) else: return dxpy.api.app_get('app-' + self._name, alias=self._alias, ** kwargs)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_get(self._dxid, **kwargs)\n else:\n return dxpy.api.app_get('app-' + se...
def x(self, **kwargs): if self._dxid is not None: return dxpy.api.app_delete(self._dxid, **kwargs) else: return dxpy.api.app_delete('app-' + self._name, alias=self._alias, **kwargs)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_delete(self._dxid, **kwargs)\n else:\n return dxpy.api.app_delete('app-...
def x(self, app_input, *args, **kwargs): return super(DXApp, self).run(app_input, *args, **kwargs)
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, app_input, *args, **kwargs):\n return super(DXApp, self).run(app_input, *args, **kwargs)\n", "role": "user" }, { "content": "run", "ro...
def x(src_dir, json_file_name, parser): if not os.path.isdir(src_dir): parser.error('{} is not a directory'.format(src_dir)) if not os.path.exists(os.path.join(src_dir, json_file_name)): raise WorkflowBuilderException( 'Directory {} does not contain dxworkflow.json: not a valid DNAnexus workflow source directory' .format(src_dir)) with open(os.path.join(src_dir, json_file_name)) as desc: try: return json_load_raise_on_duplicates(desc) except Exception as e: raise WorkflowBuilderException( 'Could not parse {} file as JSON: {}'.format(json_file_name, e.args))
parse_executable_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_dir, json_file_name, parser):\n if not os.path.isdir(src_dir):\n parser.error('{} is not a directory'.format(src_dir))\n if not os.path.exists...
def x(json_spec, args, build_project_id=None): if build_project_id: return build_project_id if 'project' in json_spec: return json_spec['project'] if dxpy.WORKSPACE_ID: return dxpy.WORKSPACE_ID error_msg = ( "Can't create a workflow without specifying a destination project; ") error_msg += ( 'please use the -d/--destination flag to explicitly specify a project') raise WorkflowBuilderException(error_msg)
get_destination_project
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args, build_project_id=None):\n if build_project_id:\n return build_project_id\n if 'project' in json_spec:\n return json_spec['...
def x(json_spec, folder_name=None): dest_folder = folder_name or json_spec.get('folder') or '/' if not dest_folder.endswith('/'): dest_folder = dest_folder + '/' return dest_folder
get_destination_folder
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, folder_name=None):\n dest_folder = folder_name or json_spec.get('folder') or '/'\n if not dest_folder.endswith('/'):\n dest_folder = de...
def x(global_workflow_id, categories_to_set): assert isinstance(categories_to_set, list) existing_categories = dxpy.api.global_workflow_list_categories( global_workflow_id)['categories'] categories_to_add = set(categories_to_set).difference(set( existing_categories)) categories_to_remove = set(existing_categories).difference(set( categories_to_set)) if categories_to_add: dxpy.api.global_workflow_add_categories(global_workflow_id, input_params={'categories': list(categories_to_add)}) if categories_to_remove: dxpy.api.global_workflow_remove_categories(global_workflow_id, input_params={'categories': list(categories_to_remove)})
set_categories_on_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(global_workflow_id, categories_to_set):\n assert isinstance(categories_to_set, list)\n existing_categories = dxpy.api.global_workflow_list_categories(\n ...
def x(json_spec, name=None, version=None): requested_name = json_spec['name'] requested_version = json_spec['version'] if requested_name == name and requested_version == version: return True else: try: desc_output = dxpy.api.global_workflow_describe( 'globalworkflow-' + json_spec['name'], alias=json_spec[ 'version'], input_params={'fields': {'name': True, 'version': True}}) return desc_output['name'] == json_spec['name'] and desc_output[ 'version'] == json_spec['version'] except dxpy.exceptions.DXAPIError: return False except: raise
version_exists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, name=None, version=None):\n requested_name = json_spec['name']\n requested_version = json_spec['version']\n if requested_name == name and r...
def x(stages, ignore_reuse_stages): if not isinstance(ignore_reuse_stages, list): raise WorkflowBuilderException( '"IgnoreReuse must be a list of strings - stage IDs or "*"') ignore_reuse_set = set(ignore_reuse_stages) if '*' in ignore_reuse_set and ignore_reuse_set == 1: return stage_ids = set([stage.get('id') for stage in stages]) for ignored in ignore_reuse_set: if ignored not in stage_ids: raise WorkflowBuilderException( 'Stage with ID {} not found. Add a matching "id" for the stage you wish to set ignoreReuse for' .format(ignored))
validate_ignore_reuse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(stages, ignore_reuse_stages):\n if not isinstance(ignore_reuse_stages, list):\n raise WorkflowBuilderException(\n '\"IgnoreReuse must be a...
def x(stages): if not isinstance(stages, list): raise WorkflowBuilderException( 'Stages must be specified as a list of dictionaries') validated_stages = [] for index, stage in enumerate(stages): validated_stages.append(_get_validated_stage(stage, index)) return validated_stages
get_validated_stages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(stages):\n if not isinstance(stages, list):\n raise WorkflowBuilderException(\n 'Stages must be specified as a list of dictionaries')\n ...
def x(json_spec, args): validated = {} override_project_id, override_folder, override_workflow_name = (dxpy. executable_builder.get_parsed_destination(args.destination)) validated['project'] = _get_destination_project(json_spec, args, override_project_id) validated['folder'] = _get_destination_folder(json_spec, override_folder) workflow_name = _get_workflow_name(json_spec, override_workflow_name) if not workflow_name: print('Warning: workflow name is not specified') else: validated['name'] = workflow_name return validated
validate_json_for_regular_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n validated = {}\n override_project_id, override_folder, override_workflow_name = (dxpy.\n executable_builder.get_parsed_destina...
def x(json_spec, args): if 'name' not in json_spec: raise WorkflowBuilderException( "dxworkflow.json contains no 'name' field, but it is required to build a global workflow" ) if not dxpy.executable_builder.GLOBAL_EXEC_NAME_RE.match(json_spec['name'] ): raise WorkflowBuilderException( 'The name of your workflow must match /^[a-zA-Z0-9._-]+$/') if json_spec['name'] != json_spec['name'].lower(): logger.warn('workflow name "{}" should be all lowercase'.format( json_spec['name'])) if 'version' not in json_spec: raise WorkflowBuilderException( "dxworkflow.json contains no 'version' field, but it is required to build a global workflow" ) if not dxpy.executable_builder.GLOBAL_EXEC_VERSION_RE.match(json_spec[ 'version']): logger.warn( '"version" {} should be semver compliant (e.g. of the form X.Y.Z)' .format(json_spec['version'])) if 'details' in json_spec: if not isinstance(json_spec['details'], dict): raise WorkflowBuilderException( 'The field "details" must be a dictionary') if 'regionalOptions' in json_spec: if not (isinstance(json_spec['regionalOptions'], dict) and json_spec['regionalOptions'] and all([isinstance(i, dict) for i in json_spec['regionalOptions'].values()])): raise WorkflowBuilderException( 'The field "regionalOptions" must be a non-empty dictionary whose values are dictionaries' ) if args.bill_to: json_spec['billTo'] = args.bill_to
validate_json_for_global_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n if 'name' not in json_spec:\n raise WorkflowBuilderException(\n \"dxworkflow.json contains no 'name' field, but it is ...
def x(json_spec, args): if not json_spec: return if not args: return validated_spec = copy.deepcopy(json_spec) unsupported_keys = _get_unsupported_keys(validated_spec.keys(), SUPPORTED_KEYS) if len(unsupported_keys) > 0: logger.warn( 'Warning: the following root level fields are not supported and will be ignored: {}' .format(', '.join(unsupported_keys))) if 'stages' in validated_spec: validated_spec['stages'] = _get_validated_stages(validated_spec[ 'stages']) if 'name' in validated_spec: if args.src_dir != validated_spec['name']: logger.warn( 'workflow name "%s" does not match containing directory "%s"' % (validated_spec['name'], args.src_dir)) if 'ignoreReuse' in validated_spec: validate_ignore_reuse(validated_spec['stages'], validated_spec[ 'ignoreReuse']) validated_documentation_fields = _get_validated_json_for_build_or_update( validated_spec, args) validated_spec.update(validated_documentation_fields) if args.mode == 'workflow': validated = _validate_json_for_regular_workflow(json_spec, args) validated_spec.update(validated) if args.mode == 'globalworkflow': _validate_json_for_global_workflow(validated_spec, args) return validated_spec
get_validated_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n if not json_spec:\n return\n if not args:\n return\n validated_spec = copy.deepcopy(json_spec)\n unsupported_keys...
def x(json_spec, args): validated = copy.deepcopy(json_spec) dxpy.executable_builder.inline_documentation_files(validated, args.src_dir) if 'title' not in json_spec: logger.warn( "dxworkflow.json is missing a title, please add one in the 'title' field" ) if 'summary' not in json_spec: logger.warn( "dxworkflow.json is missing a summary, please add one in the 'summary' field" ) elif json_spec['summary'].endswith('.'): logger.warn( 'summary {} should be a short phrase not ending in a period'. format(json_spec['summary'])) return validated
get_validated_json_for_build_or_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n validated = copy.deepcopy(json_spec)\n dxpy.executable_builder.inline_documentation_files(validated, args.src_dir)\n if 'title' no...
def x(workflow_enabled_regions, workflow_spec): executables = [i.get('executable') for i in workflow_spec.get('stages')] for exect in executables: if exect.startswith('applet-') and len(workflow_enabled_regions) > 1: raise WorkflowBuilderException( 'Building a global workflow with applets in more than one region is not yet supported.' ) elif exect.startswith('app-'): app_regional_options = dxpy.api.app_describe(exect, input_params={'fields': {'regionalOptions': True}}) app_regions = set(app_regional_options['regionalOptions'].keys()) if not workflow_enabled_regions.issubset(app_regions): additional_workflow_regions = (workflow_enabled_regions - app_regions) mesg = ( 'The app {} is enabled in regions {} while the global workflow in {}.' .format(exect, ', '.join(app_regions), ', '.join( workflow_enabled_regions))) mesg += ' The workflow will not be able to run in {}.'.format( ', '.join(additional_workflow_regions)) mesg += ( ' If you are a developer of the app, you can enable the app in {} to run the workflow in that region(s).' .format(', '.join(additional_workflow_regions))) logger.warn(mesg) elif exect.startswith('workflow-'): inner_workflow_spec = dxpy.api.workflow_describe(exect) _assert_executable_regions_match(workflow_enabled_regions, inner_workflow_spec) elif exect.startswith('globalworkflow-'): raise WorkflowBuilderException( 'Building a global workflow with nested global workflows is not yet supported' )
assert_executable_regions_match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(workflow_enabled_regions, workflow_spec):\n executables = [i.get('executable') for i in workflow_spec.get('stages')]\n for exect in executables:\n ...
def x(json_spec): workflow_id = dxpy.api.workflow_new(json_spec)['id'] dxpy.api.workflow_close(workflow_id) return workflow_id
build_regular_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec):\n workflow_id = dxpy.api.workflow_new(json_spec)['id']\n dxpy.api.workflow_close(workflow_id)\n return workflow_id\n", "role": "user" ...
def x(json_spec, from_command_line): enabled_regions = dxpy.executable_builder.get_enabled_regions( 'globalworkflow', json_spec, from_command_line, WorkflowBuilderException) if not enabled_regions: enabled_regions = [] if not dxpy.WORKSPACE_ID: msg = ( "A context project must be selected to enable a workflow in the project's region." ) msg += ( " You can use 'dx select' to select a project. Otherwise you can use --region option" ) msg += ( ' to select a region in which the workflow should be enabled') raise WorkflowBuilderException(msg) region = dxpy.api.project_describe(dxpy.WORKSPACE_ID, input_params= {'fields': {'region': True}})['region'] enabled_regions.append(region) if not enabled_regions: raise AssertionError( 'This workflow should be enabled in at least one region') return set(enabled_regions)
get_validated_enabled_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, from_command_line):\n enabled_regions = dxpy.executable_builder.get_enabled_regions(\n 'globalworkflow', json_spec, from_command_line,\n ...
def x(enabled_regions, args): projects_by_region = {} for region in enabled_regions: try: project_input = {'name': 'Temporary build project for dx build global workflow', 'region': region} if args.bill_to: project_input['billTo'] = args.bill_to temp_project = dxpy.api.project_new(project_input)['id'] projects_by_region[region] = temp_project logger.debug('Created temporary project {} to build in'.format( temp_project)) except: if projects_by_region: dxpy.executable_builder.delete_temporary_projects( projects_by_region.values()) err_exit() return projects_by_region
create_temporary_projects
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(enabled_regions, args):\n projects_by_region = {}\n for region in enabled_regions:\n try:\n project_input = {'name':\n '...
def x(enabled_regions, json_spec, args): projects_by_region = _create_temporary_projects(enabled_regions, args) workflows_by_region = {} try: for region, project in projects_by_region.items(): json_spec['project'] = project workflow_id = _build_regular_workflow(json_spec) logger.debug('Created workflow ' + workflow_id + ' successfully') workflows_by_region[region] = workflow_id except: if projects_by_region: dxpy.executable_builder.delete_temporary_projects( projects_by_region.values()) raise return workflows_by_region, projects_by_region
build_underlying_workflows
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(enabled_regions, json_spec, args):\n projects_by_region = _create_temporary_projects(enabled_regions, args)\n workflows_by_region = {}\n try:\n ...
def x(json_spec, args): enabled_regions = _get_validated_enabled_regions(json_spec, args.region) _assert_executable_regions_match(enabled_regions, json_spec) workflows_by_region, projects_by_region = {}, {} try: workflows_by_region, projects_by_region = _build_underlying_workflows( enabled_regions, json_spec, args) regional_options = {} for region, workflow_id in workflows_by_region.items(): regional_options[region] = {'workflow': workflow_id} json_spec.update({'regionalOptions': regional_options}) gwf_provided_keys = GLOBALWF_SUPPORTED_KEYS.intersection(set( json_spec.keys())) gwf_final_json = dict((k, v) for k, v in json_spec.items() if k in gwf_provided_keys) print_spec = copy.deepcopy(gwf_final_json) if 'description' in gwf_final_json: del print_spec['description'] if 'developerNotes' in gwf_final_json: del print_spec['developerNotes'] logger.info('Will create global workflow with spec: {}'.format(json .dumps(print_spec))) global_workflow_id = dxpy.api.global_workflow_new(gwf_final_json)['id'] logger.info('Uploaded global workflow {n}/{v} successfully'.format( n=gwf_final_json['name'], v=gwf_final_json['version'])) logger.info('You can publish this workflow with:') logger.info(' dx publish {n}/{v}'.format(n=gwf_final_json['name'], v=gwf_final_json['version'])) finally: if projects_by_region: dxpy.executable_builder.delete_temporary_projects( projects_by_region.values()) try: _set_categories_on_workflow(global_workflow_id, gwf_final_json.get( 'categories', [])) except: logger.warn( 'The workflow {n}/{v} was created but setting categories failed' .format(n=gwf_final_json['name'], v=gwf_final_json['version'])) raise return global_workflow_id
build_global_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n enabled_regions = _get_validated_enabled_regions(json_spec, args.region)\n _assert_executable_regions_match(enabled_regions, json_spe...
def x(json_spec, args): try: if args.mode == 'workflow': json_spec = _get_validated_json(json_spec, args) workflow_id = _build_regular_workflow(json_spec) elif args.mode == 'globalworkflow': existing_workflow = (dxpy.executable_builder. verify_developer_rights('globalworkflow-' + json_spec['name'])) if existing_workflow and _version_exists(json_spec, existing_workflow.name, existing_workflow.version): workflow_id = _update_global_workflow(json_spec, args, existing_workflow.id) else: json_spec = _get_validated_json(json_spec, args) workflow_id = _build_global_workflow(json_spec, args) else: raise WorkflowBuilderException('Unrecognized workflow type: {}' .format(args.mode)) except dxpy.exceptions.DXAPIError as e: raise e return workflow_id
build_or_update_workflow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(json_spec, args):\n try:\n if args.mode == 'workflow':\n json_spec = _get_validated_json(json_spec, args)\n workflow_id = _buil...
def x(args, parser): if args is None: raise Exception('arguments not provided') try: json_spec = _parse_executable_spec(args.src_dir, 'dxworkflow.json', parser) workflow_id = _build_or_update_workflow(json_spec, args) _print_output(workflow_id, args) except WorkflowBuilderException as e: print('Error: %s' % (e.args,), file=sys.stderr) sys.exit(3)
build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, parser):\n if args is None:\n raise Exception('arguments not provided')\n try:\n json_spec = _parse_executable_spec(args.src_dir, 'dx...
def x(args, dest_path, dx_dest, dest_proj): if len(args.sources) != 1: raise DXCLIError('The destination folder does not exist') last_slash_pos = get_last_pos_of_char('/', dest_path) if last_slash_pos == 0: dest_folder = '/' else: dest_folder = dest_path[:last_slash_pos] dest_name = dest_path[last_slash_pos + 1:].replace('\\/', '/') try: dx_dest.list_folder(folder=dest_folder, only='folders') except dxpy.DXAPIError as details: if details.code == requests.codes['not_found']: raise DXCLIError('The destination folder does not exist') else: raise except: err_exit() src_proj, src_path, src_results = try_call(resolve_existing_path, args. sources[0], allow_mult=True, all_mult=args.all) if src_proj == dest_proj: if is_hashid(args.sources[0]): raise DXCLIError(fill( 'Error: You must specify a source project for ' + args. sources[0])) else: raise DXCLIError(fill( 'A source path and the destination path resolved to the ' + 'same project or container. Please specify different source ' + 'and destination containers, e.g.') + """ dx cp source-project:source-id-or-path dest-project:dest-path""" ) if src_results is None: try: contents = dxpy.api.project_list_folder(src_proj, {'folder': src_path, 'includeHidden': True}) dxpy.api.project_new_folder(dest_proj, {'folder': dest_path}) exists = dxpy.api.project_clone(src_proj, {'folders': contents[ 'folders'], 'objects': [result['id'] for result in contents ['objects']], 'project': dest_proj, 'destination': dest_path})[ 'exists'] if len(exists) > 0: print(fill( 'The following objects already existed in the destination ' + 'container and were not copied:') + '\n ' + '\n '. join(exists)) return except: err_exit() else: try: exists = dxpy.api.project_clone(src_proj, {'objects': [result[ 'id'] for result in src_results], 'project': dest_proj, 'destination': dest_folder})['exists'] if len(exists) > 0: print(fill( 'The following objects already existed in the destination ' + 'container and were not copied:') + '\n ' + '\n '. join(exists)) for result in src_results: if result['id'] not in exists: dxpy.DXHTTPRequest('/' + result['id'] + '/rename', { 'project': dest_proj, 'name': dest_name}) return except: err_exit()
cp_to_noexistent_destination
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, dest_path, dx_dest, dest_proj):\n if len(args.sources) != 1:\n raise DXCLIError('The destination folder does not exist')\n last_slash_pos = ...
def x(num_bytes): B = num_bytes KB = float(1024) MB = float(KB * 1024) GB = float(MB * 1024) TB = float(GB * 1024) if B < KB: return '{0} {1}'.format(B, 'bytes' if B != 1 else 'byte') elif KB <= B < MB: return '{0:.2f} KiB'.format(B / KB) elif MB <= B < GB: return '{0:.2f} MiB'.format(B / MB) elif GB <= B < TB: return '{0:.2f} GiB'.format(B / GB) elif TB <= B: return '{0:.2f} TiB'.format(B / TB)
readable_part_size
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_bytes):\n B = num_bytes\n KB = float(1024)\n MB = float(KB * 1024)\n GB = float(MB * 1024)\n TB = float(GB * 1024)\n if B < KB:\n ...
def x(self, dx_hash, media_type=None, **kwargs): if media_type is not None: dx_hash['media'] = media_type resp = dxpy.api.file_new(dx_hash, **kwargs) self.set_ids(resp['id'], dx_hash['project'])
new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dx_hash, media_type=None, **kwargs):\n if media_type is not None:\n dx_hash['media'] = media_type\n resp = dxpy.api.file_new(dx_hash, **kwar...
def x(self, dxid, project=None): if self._dxid is not None: self.flush() DXDataObject.set_ids(self, dxid, project) self._pos = 0 self._file_length = None self._cur_part = 1 self._num_uploaded_parts = 0
set_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dxid, project=None):\n if self._dxid is not None:\n self.flush()\n DXDataObject.set_ids(self, dxid, project)\n self._pos = 0\n self._f...
def x(self, offset, from_what=os.SEEK_SET): if from_what == os.SEEK_SET: reference_pos = 0 elif from_what == os.SEEK_CUR: reference_pos = self._pos elif from_what == os.SEEK_END: if self._file_length == None: desc = self.describe() self._file_length = int(desc['size']) reference_pos = self._file_length else: raise DXFileError('Invalid value supplied for from_what') orig_pos = self._pos self._pos = reference_pos + offset in_buf = False orig_buf_pos = self._read_buf.tell() if offset < orig_pos: if orig_buf_pos > orig_pos - offset: in_buf = True else: buf_len = dxpy.utils.string_buffer_length(self._read_buf) if buf_len - orig_buf_pos > offset - orig_pos: in_buf = True if in_buf: self._read_buf.seek(orig_buf_pos - orig_pos + offset) elif offset == orig_pos: pass else: self._read_buf = BytesIO() self._request_iterator, self._response_iterator = None, None
seek
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, offset, from_what=os.SEEK_SET):\n if from_what == os.SEEK_SET:\n reference_pos = 0\n elif from_what == os.SEEK_CUR:\n reference_pos =...
def x(self, multithread=True, **kwargs): if self._write_buf.tell() > 0: data = self._write_buf.getvalue() self._write_buf = BytesIO() if multithread: self._async_upload_part_request(data, index=self._cur_part, ** kwargs) else: self.upload_part(data, self._cur_part, **kwargs) self._cur_part += 1 if len(self._http_threadpool_futures) > 0: dxpy.utils.wait_for_all_futures(self._http_threadpool_futures) try: for future in self._http_threadpool_futures: if future.exception() != None: raise future.exception() finally: self._http_threadpool_futures = set()
flush
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, multithread=True, **kwargs):\n if self._write_buf.tell() > 0:\n data = self._write_buf.getvalue()\n self._write_buf = BytesIO()\n ...
def x(self, data, multithread=True, **kwargs): if not USING_PYTHON2: assert isinstance(data, bytes) self._ensure_write_bufsize(**kwargs) def x(data_for_write_req): if multithread: self._async_upload_part_request(data_for_write_req, index=self. _cur_part, **kwargs) else: self.upload_part(data_for_write_req, self._cur_part, **kwargs) self._cur_part += 1 if self._write_buf.tell() == 0 and self._write_bufsize == len(data): write_request(data) return remaining_space = self._write_bufsize - self._write_buf.tell() if len(data) <= remaining_space: self._write_buf.write(data) else: self._write_buf.write(data[:remaining_space]) temp_data = self._write_buf.getvalue() self._write_buf = BytesIO() write_request(temp_data) self.write(data[remaining_space:], **kwargs)
write2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, multithread=True, **kwargs):\n if not USING_PYTHON2:\n assert isinstance(data, bytes)\n self._ensure_write_bufsize(**kwargs)\n\n de...
def x(self, data, multithread=True, **kwargs): if USING_PYTHON2: self._write2(data, multithread=multithread, **kwargs) else: if isinstance(data, str): bt = data.encode('utf-8') elif isinstance(data, bytearray): bt = bytes(data) elif isinstance(data, bytes): bt = data elif isinstance(data, mmap.mmap): bt = bytes(data) else: raise DXFileError('Invalid type {} for write data argument'. format(type(data))) assert isinstance(bt, bytes) self._write2(bt, multithread=multithread, **kwargs)
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, multithread=True, **kwargs):\n if USING_PYTHON2:\n self._write2(data, multithread=multithread, **kwargs)\n else:\n if isinstanc...
def x(self, block=False, **kwargs): self.flush(**kwargs) self._ensure_write_bufsize(**kwargs) if self._num_uploaded_parts == 0 and self._empty_last_part_allowed: try: if USING_PYTHON2: self.upload_part('', 1, **kwargs) else: self.upload_part(b'', 1, **kwargs) except dxpy.exceptions.InvalidState: pass if 'report_progress_fn' in kwargs: del kwargs['report_progress_fn'] dxpy.api.file_close(self._dxid, **kwargs) if block: self._wait_on_close(**kwargs)
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block=False, **kwargs):\n self.flush(**kwargs)\n self._ensure_write_bufsize(**kwargs)\n if self._num_uploaded_parts == 0 and self._empty_last_pa...
def x(self, data, index=None, display_progress=False, report_progress_fn= None, **kwargs): if not USING_PYTHON2: assert isinstance(data, bytes) req_input = {} if index is not None: req_input['index'] = int(index) md5 = hashlib.md5() if hasattr(data, 'seek') and hasattr(data, 'tell'): rewind_input_buffer_offset = data.tell() while True: bytes_read = data.read(MD5_READ_CHUNK_SIZE) if bytes_read: md5.update(bytes_read) else: break data.seek(rewind_input_buffer_offset) else: md5.update(data) req_input['md5'] = md5.hexdigest() req_input['size'] = len(data) def x(): if 'max_retries' not in kwargs: kwargs['max_retries'] = dxpy.DEFAULT_RETRIES elif kwargs['max_retries'] > 0: kwargs['max_retries'] -= 1 if 'timeout' not in kwargs: kwargs['timeout'] = FILE_REQUEST_TIMEOUT resp = dxpy.api.file_upload(self._dxid, req_input, **kwargs) url = resp['url'] return url, _validate_headers(resp.get('headers', {})) dxpy.DXHTTPRequest(get_upload_url_and_headers, data, jsonify_data=False, prepend_srv=False, always_retry=True, timeout=FILE_REQUEST_TIMEOUT, auth=None, method='PUT') self._num_uploaded_parts += 1 if display_progress: warn('.') if report_progress_fn is not None: report_progress_fn(self, len(data))
upload_part
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, index=None, display_progress=False, report_progress_fn=\n None, **kwargs):\n if not USING_PYTHON2:\n assert isinstance(data, bytes)\n ...