source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(zero_ok=False, more_ok=True, **kwargs):
return _find_one(find_projects, zero_ok=zero_ok, more_ok=more_ok, **kwargs)
| find_one_project | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(zero_ok=False, more_ok=True, **kwargs):\n return _find_one(find_projects, zero_ok=zero_ok, more_ok=more_ok, **kwargs)\n",
"role": "user"
},
{
"c... |
def x(zero_ok=False, more_ok=True, **kwargs):
return _find_one(find_apps, zero_ok=zero_ok, more_ok=more_ok, **kwargs)
| find_one_app | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(zero_ok=False, more_ok=True, **kwargs):\n return _find_one(find_apps, zero_ok=zero_ok, more_ok=more_ok, **kwargs)\n",
"role": "user"
},
{
"conte... |
def x(api_method, org_id, query, first_page_size=100):
if 'limit' not in query:
query['limit'] = min(first_page_size, 1000)
while True:
resp = api_method(org_id, query)
for result in resp['results']:
yield result
if resp['next'] is not None:
query['starting'] = resp['next']
query['limit'] = min(query['limit'] * 2, 1000)
else:
break
| org_find | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_method, org_id, query, first_page_size=100):\n if 'limit' not in query:\n query['limit'] = min(first_page_size, 1000)\n while True:\n r... |
def x(org_id=None, level=None, describe=False):
query = {}
if level is not None:
query['level'] = level
query['describe'] = describe
return _org_find(dxpy.api.org_find_members, org_id, query)
| org_find_members | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(org_id=None, level=None, describe=False):\n query = {}\n if level is not None:\n query['level'] = level\n query['describe'] = describe\n ret... |
def x(org_id=None, name=None, name_mode='exact', ids=None, properties=None,
tags=None, describe=False, public=None, created_after=None,
created_before=None, region=None, containsPHI=None):
query = {}
if name is not None:
if name_mode == 'exact':
query['name'] = name
elif name_mode == 'glob':
query['name'] = {'glob': name}
elif name_mode == 'regexp':
query['name'] = {'regexp': name}
else:
raise DXError(
'org_find_projects: Unexpected value found for argument name_mode'
)
if ids is not None:
query['id'] = ids
if properties is not None:
if len(properties.keys()) == 1:
query['properties'] = properties
else:
query['properties'] = {'$and': [{k: v} for k, v in properties.
iteritems()]}
if tags is not None:
if len(tags) == 1:
query['tags'] = tags[0]
else:
query['tags'] = {'$and': tags}
query['describe'] = describe
if public is not None:
query['public'] = public
if created_after is not None or created_before is not None:
query['created'] = {}
if created_after is not None:
query['created']['after'] = dxpy.utils.normalize_time_input(
created_after)
if created_before is not None:
query['created']['before'] = dxpy.utils.normalize_time_input(
created_before)
if region is not None:
query['region'] = region
if containsPHI is not None:
query['containsPHI'] = containsPHI
return _org_find(dxpy.api.org_find_projects, org_id, query)
| org_find_projects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(org_id=None, name=None, name_mode='exact', ids=None, properties=None,\n tags=None, describe=False, public=None, created_after=None,\n created_before=None... |
def x(org_id, name=None, name_mode='exact', category=None, all_versions=
None, published=None, created_by=None, developer=None, authorized_user=
None, created_after=None, created_before=None, modified_after=None,
modified_before=None, describe=False, limit=None, return_handler=False,
first_page_size=100, **kwargs):
query = {}
if name is not None:
if name_mode == 'exact':
query['name'] = name
elif name_mode == 'glob':
query['name'] = {'glob': name}
elif name_mode == 'regexp':
query['name'] = {'regexp': name}
else:
raise DXError(
'find_apps: Unexpected value found for argument name_mode')
if category is not None:
query['category'] = category
if all_versions is not None:
query['allVersions'] = all_versions
if published is not None:
query['published'] = published
if created_by is not None:
query['createdBy'] = created_by
if developer is not None:
query['developer'] = developer
if authorized_user is not None:
query['authorizedUser'] = authorized_user
if modified_after is not None or modified_before is not None:
query['modified'] = {}
if modified_after is not None:
query['modified']['after'] = dxpy.utils.normalize_time_input(
modified_after)
if modified_before is not None:
query['modified']['before'] = dxpy.utils.normalize_time_input(
modified_before)
if created_after is not None or created_before is not None:
query['created'] = {}
if created_after is not None:
query['created']['after'] = dxpy.utils.normalize_time_input(
created_after)
if created_before is not None:
query['created']['before'] = dxpy.utils.normalize_time_input(
created_before)
if describe is not None and describe is not False:
query['describe'] = describe
if limit is not None:
query['limit'] = limit
return _org_find(dxpy.api.org_find_apps, org_id, query)
| org_find_apps | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(org_id, name=None, name_mode='exact', category=None, all_versions=\n None, published=None, created_by=None, developer=None, authorized_user=\n None, crea... |
def x(query, first_page_size=10):
return _find(dxpy.api.system_find_orgs, query, limit=None,
return_handler=False, first_page_size=first_page_size)
| find_orgs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query, first_page_size=10):\n return _find(dxpy.api.system_find_orgs, query, limit=None,\n return_handler=False, first_page_size=first_page_size)\n",... |
def x(idir, dirs):
file_load_utils.ensure_dir(idir)
for d in dirs:
file_load_utils.ensure_dir(os.path.join(idir, d))
| create_dirs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(idir, dirs):\n file_load_utils.ensure_dir(idir)\n for d in dirs:\n file_load_utils.ensure_dir(os.path.join(idir, d))\n",
"role": "user"
},
... |
def x(filtered_inputs):
file_key_descs, _ignore = file_load_utils.analyze_bash_vars(file_load_utils
.get_input_json_file(), None)
flattened_dict = {}
def x(key, value, dict_):
if key not in dict_:
dict_[key] = value
for input_ in filtered_inputs:
if input_ not in file_key_descs:
continue
input_var_dict = file_key_descs[input_]
add_if_no_collision(input_ + '_path', input_var_dict['path'],
flattened_dict)
add_if_no_collision(input_ + '_name', input_var_dict['basename'],
flattened_dict)
add_if_no_collision(input_ + '_prefix', input_var_dict['prefix'],
flattened_dict)
return flattened_dict
| gen_helper_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filtered_inputs):\n file_key_descs, _ignore = file_load_utils.analyze_bash_vars(file_load_utils\n .get_input_json_file(), None)\n flattened_dict =... |
def x(max_threads, num_cores, mem_available_mb):
return min(max_threads, num_cores, max(int(mem_available_mb / 1200), 1))
| get_num_parallel_threads | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(max_threads, num_cores, mem_available_mb):\n return min(max_threads, num_cores, max(int(mem_available_mb / 1200), 1))\n",
"role": "user"
},
{
"c... |
def x(exclude=None, parallel=False, max_threads=8):
idir = file_load_utils.get_input_dir()
try:
job_input_file = file_load_utils.get_input_json_file()
dirs, inputs, rest = file_load_utils.get_job_input_filenames(
job_input_file)
except IOError:
msg = 'Error: Could not find the input json file: {0}.\n'.format(
job_input_file)
msg += (
' This function should only be called from within a running job.'
)
print(msg)
raise
dirs_to_create = []
for d in dirs:
keep = True
if exclude is not None and d is not None:
if d.split('/')[0] in exclude:
keep = False
if keep:
dirs_to_create.append(d)
_create_dirs(idir, dirs_to_create)
if exclude:
inputs = file_load_utils.filter_dict(inputs, exclude)
to_download = []
for ival_list in inputs.values():
to_download.extend(ival_list)
if parallel:
total_mem = psutil.virtual_memory().total >> 20
num_cores = multiprocessing.cpu_count()
max_num_parallel_downloads = _get_num_parallel_threads(max_threads,
num_cores, total_mem)
sys.stderr.write('Downloading files using {} threads'.format(
max_num_parallel_downloads))
_parallel_file_download(to_download, idir, max_num_parallel_downloads)
else:
_sequential_file_download(to_download, idir)
helper_vars = _gen_helper_dict(inputs)
return helper_vars
| download_all_inputs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(exclude=None, parallel=False, max_threads=8):\n idir = file_load_utils.get_input_dir()\n try:\n job_input_file = file_load_utils.get_input_json_fi... |
def x(string, ignored_prefix=' '):
paragraphs = string.split('\n\n')
refilled_paragraphs = [(fill(paragraph) if not paragraph.startswith(
ignored_prefix) else paragraph) for paragraph in paragraphs]
return '\n\n'.join(refilled_paragraphs).strip('\n')
| refill_paragraphs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, ignored_prefix=' '):\n paragraphs = string.split('\\n\\n')\n refilled_paragraphs = [(fill(paragraph) if not paragraph.startswith(\n ign... |
def x(args, results):
if args.json:
print(json.dumps(list(results), indent=4))
elif args.brief:
for result in results:
print(result['id'])
elif args.func.__name__ in ('find_projects', 'org_find_projects'):
_format_find_projects_results(results)
elif args.func.__name__ in 'org_find_members':
_format_find_org_members_results(results)
elif args.func.__name__ in 'org_find_apps':
_format_find_apps_results(results, verbose=args.verbose)
| format_find_results | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, results):\n if args.json:\n print(json.dumps(list(results), indent=4))\n elif args.brief:\n for result in results:\n print... |
def x(**kwargs):
draw_lines()
draw_histogram()
draw_bar_chart()
destination = '-r /report'
if use_html:
generate_html()
command = 'dx-build-report-html {h} {d}'.format(h=html_filename, d=
destination)
else:
command = 'dx-build-report-html {l} {b} {h} {d}'.format(l=
lines_filename, b=bars_filename, h=histogram_filename, d=
destination)
sub_output = json.loads(subprocess.check_output(command, shell=True))
output = {}
output['report'] = dxpy.dxlink(sub_output['recordId'])
return output
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(**kwargs):\n draw_lines()\n draw_histogram()\n draw_bar_chart()\n destination = '-r /report'\n if use_html:\n generate_html()\n co... |
def x():
mu, sigma = 100, 15
x = mu + sigma * numpy.random.randn(10000)
fig = pyplot.figure()
ax = fig.add_subplot(111)
n, bins, patches = ax.hist(x, 50, normed=1, facecolor='green', alpha=0.75)
ax.set_xlabel('Random number generated')
ax.set_ylabel('Probability')
ax.set_xlim(40, 160)
ax.set_ylim(0, 0.03)
ax.grid(True)
pyplot.savefig(histogram_filename)
| draw_histogram | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n mu, sigma = 100, 15\n x = mu + sigma * numpy.random.randn(10000)\n fig = pyplot.figure()\n ax = fig.add_subplot(111)\n n, bins, patches = a... |
def x():
N = 5
menMeans = 20, 35, 30, 35, 27
menStd = 2, 3, 4, 1, 2
ind = numpy.arange(N)
width = 0.35
fig = pyplot.figure()
ax = fig.add_subplot(111)
rects1 = ax.bar(ind, menMeans, width, color='r', yerr=menStd)
womenMeans = 25, 32, 34, 20, 25
womenStd = 3, 5, 2, 3, 3
rects2 = ax.bar(ind + width, womenMeans, width, color='y', yerr=womenStd)
ax.set_ylabel('Scores')
ax.set_title('Scores by group and color')
ax.set_xticks(ind + width)
ax.set_xticklabels(('G1', 'G2', 'G3', 'G4', 'G5'))
ax.legend((rects1[0], rects2[0]), ('Red', 'Yellow'))
def x(rects):
for rect in rects:
height = rect.get_height()
ax.text(rect.get_x() + rect.get_width() / 2.0, 1.05 * height,
'%d' % int(height), ha='center', va='bottom')
autolabel(rects1)
autolabel(rects2)
pyplot.savefig(bars_filename)
| draw_bar_chart | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n N = 5\n menMeans = 20, 35, 30, 35, 27\n menStd = 2, 3, 4, 1, 2\n ind = numpy.arange(N)\n width = 0.35\n fig = pyplot.figure()\n ax = ... |
def x():
r = numpy.random.randn(200)
fig = pyplot.figure()
ax = fig.add_subplot(111)
ax.plot(r)
ax.grid(True)
pyplot.savefig(lines_filename)
| draw_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n r = numpy.random.randn(200)\n fig = pyplot.figure()\n ax = fig.add_subplot(111)\n ax.plot(r)\n ax.grid(True)\n pyplot.savefig(lines_file... |
def x():
html_file = open(html_filename, 'w')
html_file.write('<html><body>')
html_file.write('<h1>Here are some graphs for you!</h1>')
for image in [lines_filename, bars_filename, histogram_filename]:
html_file.write("<div><h2>{0}</h2><img src='{0}' /></div>".format(
image))
html_file.write('</body></html>')
html_file.close()
| generate_html | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n html_file = open(html_filename, 'w')\n html_file.write('<html><body>')\n html_file.write('<h1>Here are some graphs for you!</h1>')\n for image... |
def x(o):
def x(s):
return u''.join([(c if c in ASCII_PRINTABLE else '?') for c in s])
if USING_PYTHON2:
try:
return unicode(o)
except:
try:
s = str(o)
try:
return s.decode('utf-8')
except:
return clean(s[:2048]
) + u' [Raw error message: ' + unicode(s.encode(
'hex'), 'utf-8') + u']'
except:
return u'(Unable to decode Python exception message)'
else:
return str(o)
| safe_unicode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(o):\n\n def x(s):\n return u''.join([(c if c in ASCII_PRINTABLE else '?') for c in s])\n if USING_PYTHON2:\n try:\n return unico... |
def x(e):
if isinstance(e, dxpy.AppError):
return _safe_unicode(e)
if USING_PYTHON2:
return unicode(e.__class__.__name__, 'utf-8') + ': ' + _safe_unicode(e)
else:
return e.__class__.__name__ + ': ' + _safe_unicode(e)
| format_exception_message | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(e):\n if isinstance(e, dxpy.AppError):\n return _safe_unicode(e)\n if USING_PYTHON2:\n return unicode(e.__class__.__name__, 'utf-8') + ': '... |
def x(function_name=None, function_input=None):
global RUN_COUNT
RUN_COUNT += 1
dx_working_dir = os.getcwd()
if dxpy.JOB_ID is not None:
logging.basicConfig()
try:
logging.getLogger().addHandler(dxpy.DXLogHandler())
except dxpy.exceptions.DXError:
print(
'TODO: FIXME: the EE client should die if logging is not available'
)
job = dxpy.describe(dxpy.JOB_ID)
else:
if function_name is None:
function_name = os.environ.get('DX_TEST_FUNCTION', 'main')
if function_input is None:
with open('job_input.json', 'r') as fh:
function_input = json.load(fh)
job = {'function': function_name, 'input': function_input}
with open('job_error_reserved_space', 'w') as fh:
fh.write(
"""This file contains reserved space for writing job errors in case the filesystem becomes full.
"""
+ ' ' * 1024 * 64)
print('Invoking', job.get('function'), 'with', job.get('input'))
try:
result = ENTRY_POINT_TABLE[job['function']](**job['input'])
except dxpy.AppError as e:
save_error(e, dx_working_dir, error_type='AppError')
raise
except Exception as e:
save_error(e, dx_working_dir)
raise
if result is not None:
os.chdir(dx_working_dir)
if USING_PYTHON2:
with open('job_output.json', 'wb') as fh:
json.dump(result, fh, indent=2, cls=DXJSONEncoder)
fh.write(b'\n')
else:
with open('job_output.json', 'w') as fh:
json.dump(result, fh, indent=2, cls=DXJSONEncoder)
fh.write('\n')
return result
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(function_name=None, function_input=None):\n global RUN_COUNT\n RUN_COUNT += 1\n dx_working_dir = os.getcwd()\n if dxpy.JOB_ID is not None:\n ... |
def x(entry_point_name):
def x(f):
ENTRY_POINT_TABLE[entry_point_name] = f
@wraps(f)
def x(*args, **kwargs):
return f(*args, **kwargs)
return wrapped_f
return wrap
| entry_point | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(entry_point_name):\n\n def x(f):\n ENTRY_POINT_TABLE[entry_point_name] = f\n\n @wraps(f)\n def x(*args, **kwargs):\n return ... |
def x(path):
project, _folderpath, entity_result = try_call(resolve_existing_path,
path, expected='entity')
try:
if entity_result is None or not entity_result['id'].startswith(
'workflow-'):
raise DXCLIError('Could not resolve "' + path +
'" to a workflow object')
except:
err_exit()
return entity_result['id'], project
| get_workflow_id_and_project | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n project, _folderpath, entity_result = try_call(resolve_existing_path,\n path, expected='entity')\n try:\n if entity_result is None... |
def x(authserver_host=None, authserver_port=None):
authserver = get_auth_server_name(authserver_host, authserver_port)
return DXHTTPRequest(authserver + '/system/getUserInfo', {},
prepend_srv=False)
| user_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(authserver_host=None, authserver_port=None):\n authserver = get_auth_server_name(authserver_host, authserver_port)\n return DXHTTPRequest(authserver + '/... |
def x(job_homedir=None):
if job_homedir is not None:
home_dir = job_homedir
else:
home_dir = environ.get('HOME')
idir = os.path.join(home_dir, 'in')
return idir
| get_input_dir | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_homedir=None):\n if job_homedir is not None:\n home_dir = job_homedir\n else:\n home_dir = environ.get('HOME')\n idir = os.path.join... |
def x(job_homedir=None):
if job_homedir is not None:
home_dir = job_homedir
else:
home_dir = environ.get('HOME')
odir = os.path.join(home_dir, 'out')
return odir
| get_output_dir | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_homedir=None):\n if job_homedir is not None:\n home_dir = job_homedir\n else:\n home_dir = environ.get('HOME')\n odir = os.path.join... |
def x():
path = get_output_json_file()
try:
os.remove(path)
except OSError as e:
if e.errno == errno.ENOENT:
pass
else:
raise
| rm_output_json_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n path = get_output_json_file()\n try:\n os.remove(path)\n except OSError as e:\n if e.errno == errno.ENOENT:\n pass\n ... |
def x(path):
if not os.path.exists(path):
os.mkdir(path)
elif os.path.isfile(path):
raise Exception(
'Path %s already exists, and it is a file, not a directory' % path)
| ensure_dir | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n if not os.path.exists(path):\n os.mkdir(path)\n elif os.path.isfile(path):\n raise Exception(\n 'Path %s already exists... |
def x(fname):
bad_filenames = ['.', '..']
if fname in bad_filenames:
raise DXError('Invalid filename {}'.format(fname))
return fname.replace('/', '%2F')
| make_unix_filename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname):\n bad_filenames = ['.', '..']\n if fname in bad_filenames:\n raise DXError('Invalid filename {}'.format(fname))\n return fname.replace(... |
def x(job_input_file):
def x():
with open(job_input_file) as fh:
job_input = json.load(fh)
return job_input
job_input = get_input_hash()
files = collections.defaultdict(list)
dirs = []
def x(iname, subdir, value):
if not dxpy.is_dxlink(value):
return
handler = dxpy.get_handler(value)
if not isinstance(handler, dxpy.DXFile):
return
filename = make_unix_filename(handler.name)
trg_dir = iname
if subdir is not None:
trg_dir = os.path.join(trg_dir, subdir)
files[iname].append({'trg_fname': os.path.join(trg_dir, filename),
'handler': handler, 'src_file_id': handler.id})
dirs.append(trg_dir)
def x(input_name, links):
num_files = len(links)
if num_files == 0:
return
num_digits = len(str(num_files - 1))
dirs.append(input_name)
for i, link in enumerate(links):
subdir = str(i).zfill(num_digits)
add_file(input_name, subdir, link)
for input_name, value in list(job_input.items()):
if isinstance(value, list):
add_file_array(input_name, value)
else:
add_file(input_name, None, value)
rest_hash = {key: val for key, val in list(job_input.items()) if key not in
files}
return dirs, files, rest_hash
| get_job_input_filenames | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_input_file):\n\n def x():\n with open(job_input_file) as fh:\n job_input = json.load(fh)\n return job_input\n job_input ... |
def x():
input_spec = None
if 'DX_JOB_ID' in environ:
job_desc = dxpy.describe(dxpy.JOB_ID)
if job_desc['function'] == 'main':
desc = dxpy.describe(job_desc.get('app', job_desc.get('applet')))
if 'inputSpec' in desc:
input_spec = desc['inputSpec']
elif 'DX_TEST_DXAPP_JSON' in environ:
path_to_dxapp_json = environ['DX_TEST_DXAPP_JSON']
with open(path_to_dxapp_json) as fd:
dxapp_json = json.load(fd)
input_spec = dxapp_json.get('inputSpec')
if input_spec is None:
return {}
def x(pattern):
return '*' in pattern
patterns_dict = {}
for spec in input_spec:
name = spec['name']
if 'patterns' in spec:
patterns_dict[name] = []
for p in spec['patterns']:
if is_legal_pattern(p):
patterns_dict[name].append(p)
return patterns_dict
| get_input_spec_patterns | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n input_spec = None\n if 'DX_JOB_ID' in environ:\n job_desc = dxpy.describe(dxpy.JOB_ID)\n if job_desc['function'] == 'main':\n ... |
def x(job_input_file, job_homedir):
_, file_entries, rest_hash = get_job_input_filenames(job_input_file)
patterns_dict = get_input_spec_patterns()
def x(basename, key):
best_prefix = None
patterns = patterns_dict.get(key)
if patterns is not None:
for pattern in patterns:
if fnmatch.fnmatch(basename, pattern):
_, _, right_piece = pattern.rpartition('*')
best_prefix = choose_shorter_string(best_prefix,
basename[:-len(right_piece)])
if best_prefix is not None:
return best_prefix
else:
parts = os.path.splitext(basename)
if parts[1] == '.gz':
parts = os.path.splitext(parts[0])
return parts[0]
def x():
return {'handler': [], 'basename': [], 'prefix': [], 'path': []}
file_key_descs = collections.defaultdict(factory)
rel_home_dir = get_input_dir(job_homedir)
for key, entries in list(file_entries.items()):
for entry in entries:
filename = entry['trg_fname']
basename = os.path.basename(filename)
prefix = get_prefix(basename, key)
k_desc = file_key_descs[key]
k_desc['handler'].append(entry['handler'])
k_desc['basename'].append(basename)
k_desc['prefix'].append(prefix)
k_desc['path'].append(os.path.join(rel_home_dir, filename))
return file_key_descs, rest_hash
| analyze_bash_vars | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_input_file, job_homedir):\n _, file_entries, rest_hash = get_job_input_filenames(job_input_file)\n patterns_dict = get_input_spec_patterns()\n\n d... |
def x(job_input_file, job_homedir=None, check_name_collision=True):
file_key_descs, rest_hash = analyze_bash_vars(job_input_file, job_homedir)
def x(elem):
result = None
if isinstance(elem, basestring):
result = elem
elif isinstance(elem, dxpy.DXFile):
result = json.dumps(dxpy.dxlink(elem))
else:
result = json.dumps(elem)
return pipes.quote(result)
def x(val):
if isinstance(val, list):
string = ' '.join([string_of_elem(vitem) for vitem in val])
return '( {} )'.format(string)
else:
return string_of_elem(val)
var_defs_hash = {}
def x(key, val):
if check_name_collision:
if key not in environ and key not in var_defs_hash:
var_defs_hash[key] = val
else:
sys.stderr.write(dxpy.utils.printing.fill(
'Creating environment variable ({}) would cause a name collision'
.format(key)) + '\n')
else:
var_defs_hash[key] = val
for key, desc in list(rest_hash.items()):
gen_text_line_and_name_collision(key, string_of_value(desc))
for file_key, desc in list(file_key_descs.items()):
gen_text_line_and_name_collision(file_key, string_of_value(desc[
'handler']))
gen_text_line_and_name_collision(file_key + '_name',
string_of_value(desc['basename']))
gen_text_line_and_name_collision(file_key + '_prefix',
string_of_value(desc['prefix']))
gen_text_line_and_name_collision(file_key + '_path',
string_of_value(desc['path']))
return var_defs_hash
| gen_bash_vars | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job_input_file, job_homedir=None, check_name_collision=True):\n file_key_descs, rest_hash = analyze_bash_vars(job_input_file, job_homedir)\n\n def x(elem... |
def x(futures, print_traceback=False):
while True:
try:
future = next(concurrent.futures.as_completed(futures, timeout=
THREAD_TIMEOUT_MAX))
break
except concurrent.futures.TimeoutError:
pass
except KeyboardInterrupt:
if print_traceback:
traceback.print_stack()
else:
print('')
os._exit(os.EX_IOERR)
return future
| wait_for_a_future | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(futures, print_traceback=False):\n while True:\n try:\n future = next(concurrent.futures.as_completed(futures, timeout=\n T... |
def x(futures, print_traceback=False):
try:
while True:
waited_futures = concurrent.futures.wait(futures, timeout=60)
if len(waited_futures.not_done) == 0:
break
except KeyboardInterrupt:
if print_traceback:
traceback.print_stack()
else:
print('')
os._exit(os.EX_IOERR)
| wait_for_all_futures | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(futures, print_traceback=False):\n try:\n while True:\n waited_futures = concurrent.futures.wait(futures, timeout=60)\n if len(... |
def x(request_iterator, thread_pool, max_active_tasks=None,
do_first_task_sequentially=True):
tasks_in_progress = collections.deque()
if max_active_tasks is None:
max_active_tasks = cpu_count()
def x(task_iterator, executor, futures_queue):
retval = next(task_iterator, None)
if retval is None:
return False
task_callable, task_args, task_kwargs = retval
task_future = executor.submit(task_callable, *task_args, **task_kwargs)
futures_queue.append(task_future)
return True
def x(tasks_in_progress):
future = tasks_in_progress.popleft()
try:
result = future.result(timeout=THREAD_TIMEOUT_MAX)
except KeyboardInterrupt:
print('')
os._exit(os.EX_IOERR)
return result
if do_first_task_sequentially:
task_callable, task_args, task_kwargs = next(request_iterator)
yield task_callable(*task_args, **task_kwargs)
for _i in range(max_active_tasks):
retval = submit_task(request_iterator, thread_pool, tasks_in_progress)
if not retval:
break
while len(tasks_in_progress) > 0:
result = next_result(tasks_in_progress)
submit_task(request_iterator, thread_pool, tasks_in_progress)
yield result
del result
| response_iterator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request_iterator, thread_pool, max_active_tasks=None,\n do_first_task_sequentially=True):\n tasks_in_progress = collections.deque()\n if max_active_ta... |
def x(t, future=False, default_unit='ms'):
error_msg = (
'Error: Expected an int timestamp, a date format (e.g. YYYY-MM-DD), or an int with a single-letter suffix (s=seconds, m=minutes, h=hours, d=days, w=weeks, M=months, y=years; e.g. "-10d" indicates 10 days ago); but got {t}'
)
if isinstance(t, basestring) and t.isdigit():
t = int(t)
if isinstance(t, basestring):
try:
t = normalize_timedelta(t)
except ValueError:
try:
t = int(time.mktime(dateutil.parser.parse(t).timetuple()) *
1000)
assert t > 0
except (ValueError, OverflowError, AssertionError):
raise ValueError(error_msg.format(t=t))
elif isinstance(t, numbers.Integral):
units_multipliers = {'ms': 1, 's': 1000}
if default_unit not in units_multipliers:
raise ValueError("Expected default_unit to be one of 's' or 'ms'")
t = t * units_multipliers[default_unit]
else:
raise ValueError(error_msg.format(t=t))
now = int(time.time() * 1000)
if t < 0 or future and t < now:
t += now
return t
| normalize_time_input | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(t, future=False, default_unit='ms'):\n error_msg = (\n 'Error: Expected an int timestamp, a date format (e.g. YYYY-MM-DD), or an int with a single-le... |
def x(timedelta):
try:
return int(timedelta) * 1000
except ValueError as e:
t, suffix = timedelta[:-1], timedelta[-1:]
suffix_multipliers = {'s': 1000, 'm': 1000 * 60, 'h': 1000 * 60 *
60, 'd': 1000 * 60 * 60 * 24, 'w': 1000 * 60 * 60 * 24 * 7, 'M':
1000 * 60 * 60 * 24 * 30, 'y': 1000 * 60 * 60 * 24 * 365}
if suffix not in suffix_multipliers:
raise ValueError()
return int(t) * suffix_multipliers[suffix]
| normalize_timedelta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(timedelta):\n try:\n return int(timedelta) * 1000\n except ValueError as e:\n t, suffix = timedelta[:-1], timedelta[-1:]\n suffix_mu... |
def x(ordered_pairs):
d = {}
for k, v in ordered_pairs:
if k in d:
raise ValueError('duplicate key: %r' % (k,))
else:
d[k] = v
return d
| dict_raise_on_duplicates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ordered_pairs):\n d = {}\n for k, v in ordered_pairs:\n if k in d:\n raise ValueError('duplicate key: %r' % (k,))\n else:\n ... |
def x(input_params):
input_cp = input_params.copy()
if len(input_cp.get('nonce', '')) == 0:
input_cp['nonce'] = str(Nonce())
return input_cp
| update_nonce | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_params):\n input_cp = input_params.copy()\n if len(input_cp.get('nonce', '')) == 0:\n input_cp['nonce'] = str(Nonce())\n return input_cp\... |
def x(json_spec, src_dir):
if 'description' not in json_spec:
readme_filename = None
for filename in ('README.md', 'Readme.md', 'readme.md'):
if os.path.exists(os.path.join(src_dir, filename)):
readme_filename = filename
break
if readme_filename is not None:
with open(os.path.join(src_dir, readme_filename)) as fh:
json_spec['description'] = fh.read()
if 'developerNotes' not in json_spec:
for filename in ('README.developer.md', 'Readme.developer.md',
'readme.developer.md'):
if os.path.exists(os.path.join(src_dir, filename)):
with open(os.path.join(src_dir, filename)) as fh:
json_spec['developerNotes'] = fh.read()
break
| inline_documentation_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, src_dir):\n if 'description' not in json_spec:\n readme_filename = None\n for filename in ('README.md', 'Readme.md', 'readme.md'):\... |
def x(prefixed_name):
assert prefixed_name.startswith('app-') or prefixed_name.startswith(
'globalworkflow-')
if prefixed_name.partition('-')[0] == 'app':
exception_type = dxpy.app_builder.AppBuilderException
describe_method = dxpy.api.app_describe
exception_msg = (
'An app with the given name already exists and you are not a developer of that app'
)
else:
exception_type = dxpy.workflow_builder.WorkflowBuilderException
describe_method = dxpy.api.global_workflow_describe
exception_msg = (
'A global workflow with the given name already exists and you are not a developer of that workflow'
)
name_already_exists = True
is_developer = False
version = None
executable_id = None
FoundExecutable = collections.namedtuple('FoundExecutable', ['name',
'version', 'id'])
try:
describe_output = describe_method(prefixed_name, input_params={
'fields': {'isDeveloperFor': True, 'version': True, 'id': True}})
is_developer = describe_output['isDeveloperFor']
version = describe_output['version']
executable_id = describe_output['id']
except dxpy.exceptions.DXAPIError as e:
if e.name == 'ResourceNotFound':
name_already_exists = False
elif e.name == 'PermissionDenied':
raise exception_type(exception_msg)
else:
raise e
if not name_already_exists:
return FoundExecutable(name=None, version=None, id=None)
name_without_prefix = prefixed_name.partition('-')[2]
if not is_developer:
raise exception_type('You are not a developer for {n}'.format(n=
name_without_prefix))
return FoundExecutable(name=name_without_prefix, version=version, id=
executable_id)
| verify_developer_rights | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prefixed_name):\n assert prefixed_name.startswith('app-') or prefixed_name.startswith(\n 'globalworkflow-')\n if prefixed_name.partition('-')[0] =... |
def x(from_spec, from_command_line, builder_exception):
if from_spec is None or from_command_line is None:
return
if set(from_spec) != set(from_command_line):
raise builder_exception(
"--region and the 'regionalOptions' key in the JSON file do not agree"
)
| assert_consistent_regions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(from_spec, from_command_line, builder_exception):\n if from_spec is None or from_command_line is None:\n return\n if set(from_spec) != set(from_co... |
def x(exec_type, json_spec, executable_builder_exeception):
reg_options_spec = json_spec.get('regionalOptions')
json_fn = 'dxapp.json' if exec_type == 'app' else 'dxworkflow.json'
if not isinstance(reg_options_spec, dict):
raise executable_builder_exeception(
"The field 'regionalOptions' in must be a mapping")
if not reg_options_spec:
raise executable_builder_exeception(
"The field 'regionalOptions' in " + json_fn +
' must be a non-empty mapping')
regional_options_list = list(reg_options_spec.items())
for region, opts_for_region in regional_options_list:
if not isinstance(opts_for_region, dict):
raise executable_builder_exeception(
"The field 'regionalOptions['" + region + "']' in " +
json_fn + ' must be a mapping')
if set(opts_for_region.keys()) != set(regional_options_list[0][1].
keys()):
if set(opts_for_region.keys()) - set(regional_options_list[0][1
].keys()):
with_key, without_key = region, regional_options_list[0][0]
key_name = next(iter(set(opts_for_region.keys()) - set(
regional_options_list[0][1].keys())))
else:
with_key, without_key = regional_options_list[0][0], region
key_name = next(iter(set(regional_options_list[0][1].keys()
) - set(opts_for_region.keys())))
raise executable_builder_exeception(
'All regions in regionalOptions must specify the same options; '
+ '%s was given for %s but not for %s' % (key_name,
with_key, without_key))
if exec_type == 'app':
for key in opts_for_region:
if key in json_spec.get('runSpec', {}):
raise executable_builder_exeception(key +
' cannot be given in both runSpec and in regional options for '
+ region)
| assert_consistent_reg_options | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(exec_type, json_spec, executable_builder_exeception):\n reg_options_spec = json_spec.get('regionalOptions')\n json_fn = 'dxapp.json' if exec_type == 'app... |
def x(exec_type, json_spec, from_command_line, executable_builder_exeception):
from_spec = json_spec.get('regionalOptions')
if from_spec is not None:
assert_consistent_reg_options(exec_type, json_spec,
executable_builder_exeception)
assert_consistent_regions(from_spec, from_command_line,
executable_builder_exeception)
enabled_regions = None
if from_spec is not None:
enabled_regions = from_spec.keys()
elif from_command_line is not None:
enabled_regions = from_command_line
return enabled_regions
| get_enabled_regions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(exec_type, json_spec, from_command_line, executable_builder_exeception):\n from_spec = json_spec.get('regionalOptions')\n if from_spec is not None:\n ... |
def x(self, fields=None, **kwargs):
describe_input = {}
if fields is not None:
describe_input['fields'] = fields
self._desc = dxpy.api.analysis_describe(self._dxid, describe_input, **
kwargs)
return self._desc
| describe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fields=None, **kwargs):\n describe_input = {}\n if fields is not None:\n describe_input['fields'] = fields\n self._desc = dxpy.api.analys... |
def x(self, tags, **kwargs):
dxpy.api.analysis_add_tags(self._dxid, {'tags': tags}, **kwargs)
| add_tags | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tags, **kwargs):\n dxpy.api.analysis_add_tags(self._dxid, {'tags': tags}, **kwargs)\n",
"role": "user"
},
{
"content": "add_tags",
"ro... |
def x(self, tags, **kwargs):
dxpy.api.analysis_remove_tags(self._dxid, {'tags': tags}, **kwargs)
| remove_tags | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tags, **kwargs):\n dxpy.api.analysis_remove_tags(self._dxid, {'tags': tags}, **kwargs)\n",
"role": "user"
},
{
"content": "remove_tags",
... |
def x(self, properties, **kwargs):
dxpy.api.analysis_set_properties(self._dxid, {'properties': properties},
**kwargs)
| set_properties | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, properties, **kwargs):\n dxpy.api.analysis_set_properties(self._dxid, {'properties': properties},\n **kwargs)\n",
"role": "user"
},
{
... |
def x(self, field, index=None, metadata=None):
link = {'$dnanexus_link': {'analysis': self._dxid, 'field': field}}
if index is not None:
link['$dnanexus_link']['index'] = index
if metadata is not None:
link['$dnanexus_link']['metadata'] = metadata
return link
| get_output_ref | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, field, index=None, metadata=None):\n link = {'$dnanexus_link': {'analysis': self._dxid, 'field': field}}\n if index is not None:\n link['$dn... |
def x(app_json, publish=False):
for input_field in app_json.get('inputSpec', []):
for suggestion in input_field.get('suggestions', []):
if 'project' in suggestion:
try:
project = dxpy.api.project_describe(suggestion[
'project'], {'permissions': True})
if 'PUBLIC' not in project['permissions'] and publish:
logger.warn('Project {name} NOT PUBLIC!'.format(
name=project['name']))
except dxpy.exceptions.DXAPIError as e:
if e.code == 404:
logger.warn(
'Suggested project {name} does not exist, or not accessible by user'
.format(name=suggestion['project']))
if 'path' in suggestion:
try:
check_folder_exists(suggestion['project'],
suggestion['path'], '')
except ResolutionError as e:
logger.warn(
'Folder {path} could not be found in project {project}'
.format(path=suggestion['path'], project=
suggestion['project']))
if '$dnanexus_link' in suggestion:
if suggestion['$dnanexus_link'].startswith(('file-', 'record-')
):
try:
dnanexus_link = dxpy.describe(suggestion[
'$dnanexus_link'])
except dxpy.exceptions.DXAPIError as e:
if e.code == 404:
raise dxpy.app_builder.AppBuilderException(
'Suggested object {name} could not be found'
.format(name=suggestion['$dnanexus_link']))
except Exception as e:
raise dxpy.app_builder.AppBuilderException(str(e))
if 'value' in suggestion and isinstance(suggestion['value'], dict):
if '$dnanexus_link' in suggestion['value']:
if isinstance(suggestion['value']['$dnanexus_link'], dict):
if 'project' in suggestion['value']['$dnanexus_link']:
try:
dxpy.api.project_describe(suggestion[
'value']['$dnanexus_link']['project'])
except dxpy.exceptions.DXAPIError as e:
if e.code == 404:
logger.warn(
'Suggested project {name} does not exist, or not accessible by user'
.format(name=suggestion['value'][
'$dnanexus_link']['project']))
elif isinstance(suggestion['value']['$dnanexus_link'],
basestring):
if suggestion['value']['$dnanexus_link'].startswith((
'file-', 'record-')):
try:
dnanexus_link = dxpy.describe(suggestion[
'value']['$dnanexus_link'])
except dxpy.exceptions.DXAPIError as e:
if e.code == 404:
raise dxpy.app_builder.AppBuilderException(
'Suggested object {name} could not be found'
.format(name=suggestion['value'][
'$dnanexus_link']))
except Exception as e:
raise dxpy.app_builder.AppBuilderException(str
(e))
| check_suggestions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(app_json, publish=False):\n for input_field in app_json.get('inputSpec', []):\n for suggestion in input_field.get('suggestions', []):\n if... |
def x(dxapp_json_filename, mode):
def x(dirname):
for basename in ['README.md', 'Readme.md', 'readme.md']:
if os.path.exists(os.path.join(dirname, basename)):
return os.path.join(dirname, basename)
return None
app_spec = json.load(open(dxapp_json_filename))
dirname = os.path.basename(os.path.dirname(os.path.abspath(
dxapp_json_filename)))
if mode == 'app':
if 'title' not in app_spec:
logger.warn(
'app is missing a title, please add one in the "title" field of dxapp.json'
)
if 'summary' in app_spec:
if app_spec['summary'].endswith('.'):
logger.warn(
'summary "%s" should be a short phrase not ending in a period'
% (app_spec['summary'],))
else:
logger.warn(
'app is missing a summary, please add one in the "summary" field of dxapp.json'
)
readme_filename = _find_readme(os.path.dirname(dxapp_json_filename))
if 'description' in app_spec:
if readme_filename:
raise dxpy.app_builder.AppBuilderException(
'Description was provided both in Readme.md and in the "description" field of {file}. Please consolidate content in Readme.md and remove the "description" field.'
.format(file=dxapp_json_filename))
if not app_spec['description'].strip().endswith('.'):
logger.warn(
'"description" field should be written in complete sentences and end with a period'
)
elif readme_filename is None:
logger.warn(
'app is missing a description, please supply one in README.md')
if 'categories' in app_spec:
for category in app_spec['categories']:
if category not in APP_CATEGORIES:
logger.warn('app has unrecognized category "%s"' % (
category,))
if category == 'Import':
if 'title' in app_spec and not app_spec['title'].endswith(
'Importer'):
logger.warn('title "%s" should end in "Importer"' %
(app_spec['title'],))
if category == 'Export':
if 'title' in app_spec and not app_spec['title'].endswith(
'Exporter'):
logger.warn('title "%s" should end in "Exporter"' %
(app_spec['title'],))
if 'name' in app_spec:
if app_spec['name'] != app_spec['name'].lower():
logger.warn('name "%s" should be all lowercase' % (app_spec[
'name'],))
if dirname != app_spec['name']:
logger.warn(
'app name "%s" does not match containing directory "%s"' %
(app_spec['name'], dirname))
else:
logger.warn(
'app is missing a name, please add one in the "name" field of dxapp.json'
)
if 'version' in app_spec:
if not dxpy.executable_builder.GLOBAL_EXEC_VERSION_RE.match(app_spec
['version']):
logger.warn(
'"version" %s should be semver compliant (e.g. of the form X.Y.Z)'
% (app_spec['version'],))
if 'inputSpec' in app_spec:
for i, input_field in enumerate(app_spec['inputSpec']):
if not re.match('^[a-zA-Z_][0-9a-zA-Z_]*$', input_field['name']):
logger.error(
'input %d has illegal name "%s" (must match ^[a-zA-Z_][0-9a-zA-Z_]*$)'
% (i, input_field['name']))
else:
logger.warn(
'dxapp.json contains no input specification (inputSpec). Your applet will not be usable as an '
+ 'app, runnable from the GUI, or composable using workflows.')
if 'outputSpec' in app_spec:
for i, output_field in enumerate(app_spec['outputSpec']):
if not re.match('^[a-zA-Z_][0-9a-zA-Z_]*$', output_field['name']):
logger.error(
'output %d has illegal name "%s" (must match ^[a-zA-Z_][0-9a-zA-Z_]*$)'
% (i, output_field['name']))
else:
logger.warn(
'dxapp.json contains no output specification (outputSpec). Your applet will not be usable as an '
+ 'app, runnable from the GUI, or composable using workflows.')
| lint | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dxapp_json_filename, mode):\n\n def x(dirname):\n for basename in ['README.md', 'Readme.md', 'readme.md']:\n if os.path.exists(os.path.joi... |
def x(code, lang, temp_dir, enforce=True):
if lang == 'python2.7':
temp_basename = 'inlined_code_from_dxapp_json.py'
elif lang == 'bash':
temp_basename = 'inlined_code_from_dxapp_json.sh'
else:
raise ValueError('lang must be one of "python2.7" or "bash"')
with open(os.path.join(temp_dir, temp_basename), 'w') as ofile:
ofile.write(code)
_check_file_syntax(os.path.join(temp_dir, temp_basename), temp_dir,
override_lang=lang, enforce=enforce)
| check_syntax | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(code, lang, temp_dir, enforce=True):\n if lang == 'python2.7':\n temp_basename = 'inlined_code_from_dxapp_json.py'\n elif lang == 'bash':\n ... |
def x(filename, temp_dir, override_lang=None, enforce=True):
def x(filename):
pyc_path = os.path.join(temp_dir, os.path.basename(filename) + '.pyc')
try:
if USING_PYTHON2:
filename = filename.encode(sys.getfilesystemencoding())
py_compile.compile(filename, cfile=pyc_path, doraise=True)
finally:
try:
os.unlink(pyc_path)
except OSError:
pass
def x(filename):
if platform.system() == 'Windows':
logging.warn(
'Skipping bash syntax check due to unavailability of bash on Windows.'
)
else:
subprocess.check_output(['/bin/bash', '-n', filename], stderr=
subprocess.STDOUT)
if override_lang == 'python2.7':
checker_fn = check_python
elif override_lang == 'bash':
checker_fn = check_bash
elif filename.endswith('.py'):
checker_fn = check_python
elif filename.endswith('.sh'):
checker_fn = check_bash
else:
return
open(filename)
try:
checker_fn(filename)
except subprocess.CalledProcessError as e:
print(filename + ' has a syntax error! Interpreter output:', file=
sys.stderr)
for line in e.output.strip('\n').split('\n'):
print(' ' + line.rstrip('\n'), file=sys.stderr)
if enforce:
raise DXSyntaxError(filename + ' has a syntax error')
except py_compile.PyCompileError as e:
print(filename + ' has a syntax error! Interpreter output:', file=
sys.stderr)
print(' ' + e.msg.strip(), file=sys.stderr)
if enforce:
raise DXSyntaxError(e.msg.strip())
| check_file_syntax | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, temp_dir, override_lang=None, enforce=True):\n\n def x(filename):\n pyc_path = os.path.join(temp_dir, os.path.basename(filename) + '.pyc')\... |
def x(src_dir, temp_dir, mode, enforce=True):
_lint(os.path.join(src_dir, 'dxapp.json'), mode)
manifest = json.load(open(os.path.join(src_dir, 'dxapp.json')))
if 'runSpec' in manifest:
if 'interpreter' not in manifest['runSpec']:
raise dxpy.app_builder.AppBuilderException(
'runSpec.interpreter field was not present')
if 'release' not in manifest['runSpec'
] or 'distribution' not in manifest['runSpec']:
warn_message = (
'runSpec.distribution or runSpec.release was not present. These fields '
)
warn_message += (
'will be required in a future version of the API. Recommended value '
)
warn_message += (
'for distribution is "Ubuntu" and release - "14.04".')
logger.warn(warn_message)
if manifest['runSpec']['interpreter'] in ['python2.7', 'bash']:
if 'file' in manifest['runSpec']:
entry_point_file = os.path.abspath(os.path.join(src_dir,
manifest['runSpec']['file']))
try:
_check_file_syntax(entry_point_file, temp_dir,
override_lang=manifest['runSpec']['interpreter'],
enforce=enforce)
except IOError as e:
raise dxpy.app_builder.AppBuilderException(
'Could not open runSpec.file=%r. The problem was: %s' %
(entry_point_file, e))
except DXSyntaxError:
raise dxpy.app_builder.AppBuilderException(
'Entry point file %s has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.'
% (entry_point_file,))
elif 'code' in manifest['runSpec']:
try:
_check_syntax(manifest['runSpec']['code'], manifest[
'runSpec']['interpreter'], temp_dir, enforce=enforce)
except DXSyntaxError:
raise dxpy.app_builder.AppBuilderException(
'Code in runSpec.code has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.'
)
cluster_bootstrap_scripts = (
_get_all_cluster_bootstrap_script_names(manifest))
for filename in cluster_bootstrap_scripts:
abs_filename = os.path.abspath(os.path.join(src_dir, filename))
try:
_check_file_syntax(abs_filename, temp_dir,
override_lang=manifest['runSpec']['interpreter'],
enforce=enforce)
except IOError as e:
raise dxpy.app_builder.AppBuilderException(
'Could not open cluster bootstrap script %r. The problem was: %s'
% (abs_filename, e))
except DXSyntaxError:
raise dxpy.app_builder.AppBuilderException(
'Code in cluster bootstrapScript %r has syntax errors, see above for details. Rerun with --no-check-syntax to proceed anyway.'
% filename)
if 'execDepends' in manifest['runSpec']:
if not isinstance(manifest['runSpec']['execDepends'], list):
raise dxpy.app_builder.AppBuilderException(
'Expected runSpec.execDepends to be an array. Rerun with --no-check-syntax to proceed anyway.'
)
if not all(isinstance(dep, dict) for dep in manifest['runSpec']
['execDepends']):
raise dxpy.app_builder.AppBuilderException(
'Expected runSpec.execDepends to be an array of hashes. Rerun with --no-check-syntax to proceed anyway.'
)
if any(dep.get('package_manager', 'apt') != 'apt' for dep in
manifest['runSpec']['execDepends']):
if not isinstance(manifest.get('access'), dict
) or 'network' not in manifest['access']:
msg = '\n'.join([
'runSpec.execDepends specifies non-APT dependencies, but no network access spec is given.'
,
'Add {"access": {"network": ["*"]}} to allow dependencies to install.'
,
'See https://wiki.dnanexus.com/Developer-Tutorials/Request-Additional-App-Resources#Network-Access.'
, 'Rerun with --no-check-syntax to proceed anyway.'])
raise dxpy.app_builder.AppBuilderException(msg)
if 'authorizedUsers' in manifest:
if not isinstance(manifest['authorizedUsers'], list) or isinstance(
manifest['authorizedUsers'], basestring):
raise dxpy.app_builder.AppBuilderException(
'Expected authorizedUsers to be a list of strings')
for thing in manifest['authorizedUsers']:
if thing != 'PUBLIC' and (not isinstance(thing, basestring) or
not re.match('^(org-|user-)', thing)):
raise dxpy.app_builder.AppBuilderException(
'authorizedUsers field contains an entry which is not either the string "PUBLIC" or a user or org ID'
)
if 'pricingPolicy' in manifest:
error_message = (
'"pricingPolicy" at the top level is not accepted. It must be specified '
)
error_message += (
'under the "regionalOptions" field in all enabled regions of the app'
)
raise dxpy.app_builder.AppBuilderException(error_message)
files_with_problems = []
for dirpath, dirnames, filenames in os.walk(os.path.abspath(os.path.
join(src_dir, 'resources'))):
for filename in filenames:
if not filename.startswith('._'):
try:
_check_file_syntax(os.path.join(dirpath, filename),
temp_dir, enforce=True)
except IOError as e:
raise dxpy.app_builder.AppBuilderException(
'Could not open file in resources directory %r. The problem was: %s'
% (os.path.join(dirpath, filename), e))
except DXSyntaxError:
files_with_problems.append(os.path.join(dirpath, filename))
if files_with_problems:
files_str = files_with_problems[0] if len(files_with_problems
) == 1 else files_with_problems[0] + ' and ' + str(len(
files_with_problems) - 1) + ' other file' + ('s' if len(
files_with_problems) > 2 else '')
logging.warn('%s contained syntax errors, see above for details' %
(files_str,))
| verify_app_source_dir_impl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_dir, temp_dir, mode, enforce=True):\n _lint(os.path.join(src_dir, 'dxapp.json'), mode)\n manifest = json.load(open(os.path.join(src_dir, 'dxapp.json'... |
def x(src_dir, mode, enforce=True):
temp_dir = tempfile.mkdtemp(prefix='dx-build_tmp')
try:
_verify_app_source_dir_impl(src_dir, temp_dir, mode, enforce=enforce)
finally:
shutil.rmtree(temp_dir)
| verify_app_source_dir | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_dir, mode, enforce=True):\n temp_dir = tempfile.mkdtemp(prefix='dx-build_tmp')\n try:\n _verify_app_source_dir_impl(src_dir, temp_dir, mode, e... |
def x(src_dir):
if not os.path.isdir(src_dir):
parser.error('%s is not a directory' % src_dir)
if not os.path.exists(os.path.join(src_dir, 'dxapp.json')):
raise dxpy.app_builder.AppBuilderException(
'Directory %s does not contain dxapp.json: not a valid DNAnexus app source directory'
% src_dir)
with open(os.path.join(src_dir, 'dxapp.json')) as app_desc:
try:
return json_load_raise_on_duplicates(app_desc)
except Exception as e:
raise dxpy.app_builder.AppBuilderException(
'Could not parse dxapp.json file as JSON: ' + str(e.args))
| parse_app_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_dir):\n if not os.path.isdir(src_dir):\n parser.error('%s is not a directory' % src_dir)\n if not os.path.exists(os.path.join(src_dir, 'dxapp.... |
def x(args, extra_args):
if not args.remote:
try:
output = build_and_upload_locally(args.src_dir, args.mode,
overwrite=args.overwrite, archive=args.archive, publish=
args.publish, destination_override=args.destination,
version_override=args.version_override, bill_to_override=
args.bill_to, use_temp_build_project=args.
use_temp_build_project, do_parallel_build=args.
parallel_build, do_version_autonumbering=args.
version_autonumbering, do_try_update=args.update,
dx_toolkit_autodep=args.dx_toolkit_autodep, do_check_syntax
=args.check_syntax, ensure_upload=args.ensure_upload,
force_symlinks=args.force_symlinks, dry_run=args.dry_run,
confirm=args.confirm, return_object_dump=args.json, region=
args.region, **extra_args)
if output is not None and args.run is None:
print(json.dumps(output))
except dxpy.app_builder.AppBuilderException as e:
print('Error: %s' % (e.args,), file=sys.stderr)
sys.exit(3)
except dxpy.exceptions.DXAPIError as e:
print('Error: %s' % (e,), file=sys.stderr)
sys.exit(3)
if args.dry_run:
return None
return output['id']
else:
try:
app_json = _parse_app_spec(args.src_dir)
_check_suggestions(app_json, publish=args.publish)
_verify_app_source_dir(args.src_dir, args.mode)
if args.mode == 'app' and not args.dry_run:
dxpy.executable_builder.verify_developer_rights('app-' +
app_json['name'])
except dxpy.app_builder.AppBuilderException as e:
print('Error: %s' % (e.args,), file=sys.stderr)
sys.exit(3)
if args.dry_run:
parser.error('--remote cannot be combined with --dry-run')
if args.overwrite:
parser.error('--remote cannot be combined with --overwrite/-f')
if args.archive:
parser.error('--remote cannot be combined with --archive/-a')
if args.json:
parser.error('--remote cannot be combined with --json')
if not args.use_temp_build_project:
parser.error(
'--remote cannot be combined with --no-temp-build-project')
if isinstance(args.region, list) and len(args.region) > 1:
parser.error(
'--region can only be specified once for remote builds')
region = args.region[0] if args.region is not None else None
more_kwargs = {}
if args.version_override:
more_kwargs['version_override'] = args.version_override
if args.bill_to:
more_kwargs['bill_to_override'] = args.bill_to
if not args.version_autonumbering:
more_kwargs['do_version_autonumbering'] = False
if not args.update:
more_kwargs['do_try_update'] = False
if not args.parallel_build:
more_kwargs['do_parallel_build'] = False
if not args.check_syntax:
more_kwargs['do_check_syntax'] = False
return _build_app_remote(args.mode, args.src_dir,
destination_override=args.destination, publish=args.publish,
dx_toolkit_autodep=args.dx_toolkit_autodep, region=region,
watch=args.watch, **more_kwargs)
| build_app | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, extra_args):\n if not args.remote:\n try:\n output = build_and_upload_locally(args.src_dir, args.mode,\n overwrite=ar... |
def x(**kwargs):
if len(sys.argv) > 0:
if sys.argv[0].endswith('dx-build-app'):
logging.warn(
'Warning: dx-build-app has been replaced with "dx build --create-app". Please update your scripts.'
)
elif sys.argv[0].endswith('dx-build-applet'):
logging.warn(
'Warning: dx-build-applet has been replaced with "dx build". Please update your scripts.'
)
exit(0)
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(**kwargs):\n if len(sys.argv) > 0:\n if sys.argv[0].endswith('dx-build-app'):\n logging.warn(\n 'Warning: dx-build-app has ... |
def x(self, dxid=None, name=None, alias=None):
self._dxid = None
self._name = None
self._alias = None
if dxid is not None:
if name is not None or alias is not None:
raise DXError(
'Did not expect name or alias to be given if dxid is given')
verify_string_dxid(dxid, self._class)
self._dxid = dxid
elif name is not None:
self._name = name
if not isinstance(name, basestring):
raise DXError('App name needs to be a string: %r' % (name,))
if alias is not None:
if not isinstance(alias, basestring):
raise DXError('App alias needs to be a string: %r' % (alias,))
self._alias = alias
else:
self._alias = 'default'
| set_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dxid=None, name=None, alias=None):\n self._dxid = None\n self._name = None\n self._alias = None\n if dxid is not None:\n if name is no... |
def x(self):
if self._dxid is not None:
return self._dxid
else:
return 'app-' + self._name + '/' + self._alias
| get_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._dxid is not None:\n return self._dxid\n else:\n return 'app-' + self._name + '/' + self._alias\n",
"role": "user"
}... |
def x(self, **kwargs):
dx_hash = {}
if 'applet' not in kwargs and 'initializeFrom' not in kwargs:
raise DXError(
'%s: One of the keyword arguments %s and %s is required' % (
self.__class__.__name__, 'applet', 'initializeFrom'))
for field in ['version']:
if field not in kwargs:
raise DXError('%s: Keyword argument %s is required' % (self.
__class__.__name__, field))
dx_hash[field] = kwargs[field]
del kwargs[field]
for field in ('initializeFrom', 'applet', 'name', 'title', 'summary',
'description', 'billing', 'access', 'resources'):
if field in kwargs:
dx_hash[field] = kwargs[field]
del kwargs[field]
if 'bill_to' in kwargs:
dx_hash['billTo'] = kwargs['bill_to']
del kwargs['bill_to']
resp = dxpy.api.app_new(dx_hash, **kwargs)
self.set_id(dxid=resp['id'])
| new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n dx_hash = {}\n if 'applet' not in kwargs and 'initializeFrom' not in kwargs:\n raise DXError(\n '%s: One of the keyw... |
def x(self, fields=None, **kwargs):
describe_input = {}
if fields:
describe_input['fields'] = fields
if self._dxid is not None:
self._desc = dxpy.api.app_describe(self._dxid, input_params=
describe_input, **kwargs)
else:
self._desc = dxpy.api.app_describe('app-' + self._name, alias=self.
_alias, input_params=describe_input, **kwargs)
return self._desc
| describe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fields=None, **kwargs):\n describe_input = {}\n if fields:\n describe_input['fields'] = fields\n if self._dxid is not None:\n self... |
def x(self, **kwargs):
updates = {}
for field in ('applet', 'billing', 'access', 'resources', 'details'):
if field in kwargs:
updates[field] = kwargs[field]
del kwargs[field]
if self._dxid is not None:
resp = dxpy.api.app_update(self._dxid, input_params=updates, **kwargs)
else:
resp = dxpy.api.app_update('app-' + self._name, alias=self._alias,
input_params=updates, **kwargs)
| update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n updates = {}\n for field in ('applet', 'billing', 'access', 'resources', 'details'):\n if field in kwargs:\n updates... |
def x(self, tags, **kwargs):
if self._dxid is not None:
return dxpy.api.app_add_tags(self._dxid, input_params={'tags': tags
}, **kwargs)
else:
return dxpy.api.app_add_tags('app-' + self._name, alias=self._alias,
input_params={'tags': tags}, **kwargs)
| add_tags | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tags, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_add_tags(self._dxid, input_params={'tags': tags\n }, **kwargs)\n... |
def x(self, tags, **kwargs):
if self._dxid is not None:
return dxpy.api.app_remove_tags(self._dxid, input_params={'tags':
tags}, **kwargs)
else:
return dxpy.api.app_remove_tags('app-' + self._name, alias=self.
_alias, input_params={'tags': tags}, **kwargs)
| remove_tags | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tags, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_remove_tags(self._dxid, input_params={'tags':\n tags}, **kwargs)... |
def x(self, **kwargs):
if self._dxid is not None:
return dxpy.api.app_install(self._dxid, **kwargs)
else:
return dxpy.api.app_install('app-' + self._name, alias=self._alias,
**kwargs)
| install | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_install(self._dxid, **kwargs)\n else:\n return dxpy.api.app_install('ap... |
def x(self, **kwargs):
if self._dxid is not None:
return dxpy.api.app_uninstall(self._dxid, **kwargs)
else:
return dxpy.api.app_uninstall('app-' + self._name, alias=self.
_alias, **kwargs)
| uninstall | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_uninstall(self._dxid, **kwargs)\n else:\n return dxpy.api.app_uninstall... |
def x(self, **kwargs):
if self._dxid is not None:
return dxpy.api.app_get(self._dxid, **kwargs)
else:
return dxpy.api.app_get('app-' + self._name, alias=self._alias, **
kwargs)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_get(self._dxid, **kwargs)\n else:\n return dxpy.api.app_get('app-' + se... |
def x(self, **kwargs):
if self._dxid is not None:
return dxpy.api.app_delete(self._dxid, **kwargs)
else:
return dxpy.api.app_delete('app-' + self._name, alias=self._alias,
**kwargs)
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._dxid is not None:\n return dxpy.api.app_delete(self._dxid, **kwargs)\n else:\n return dxpy.api.app_delete('app-... |
def x(self, app_input, *args, **kwargs):
return super(DXApp, self).run(app_input, *args, **kwargs)
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, app_input, *args, **kwargs):\n return super(DXApp, self).run(app_input, *args, **kwargs)\n",
"role": "user"
},
{
"content": "run",
"ro... |
def x(src_dir, json_file_name, parser):
if not os.path.isdir(src_dir):
parser.error('{} is not a directory'.format(src_dir))
if not os.path.exists(os.path.join(src_dir, json_file_name)):
raise WorkflowBuilderException(
'Directory {} does not contain dxworkflow.json: not a valid DNAnexus workflow source directory'
.format(src_dir))
with open(os.path.join(src_dir, json_file_name)) as desc:
try:
return json_load_raise_on_duplicates(desc)
except Exception as e:
raise WorkflowBuilderException(
'Could not parse {} file as JSON: {}'.format(json_file_name,
e.args))
| parse_executable_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_dir, json_file_name, parser):\n if not os.path.isdir(src_dir):\n parser.error('{} is not a directory'.format(src_dir))\n if not os.path.exists... |
def x(json_spec, args, build_project_id=None):
if build_project_id:
return build_project_id
if 'project' in json_spec:
return json_spec['project']
if dxpy.WORKSPACE_ID:
return dxpy.WORKSPACE_ID
error_msg = (
"Can't create a workflow without specifying a destination project; ")
error_msg += (
'please use the -d/--destination flag to explicitly specify a project')
raise WorkflowBuilderException(error_msg)
| get_destination_project | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args, build_project_id=None):\n if build_project_id:\n return build_project_id\n if 'project' in json_spec:\n return json_spec['... |
def x(json_spec, folder_name=None):
dest_folder = folder_name or json_spec.get('folder') or '/'
if not dest_folder.endswith('/'):
dest_folder = dest_folder + '/'
return dest_folder
| get_destination_folder | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, folder_name=None):\n dest_folder = folder_name or json_spec.get('folder') or '/'\n if not dest_folder.endswith('/'):\n dest_folder = de... |
def x(global_workflow_id, categories_to_set):
assert isinstance(categories_to_set, list)
existing_categories = dxpy.api.global_workflow_list_categories(
global_workflow_id)['categories']
categories_to_add = set(categories_to_set).difference(set(
existing_categories))
categories_to_remove = set(existing_categories).difference(set(
categories_to_set))
if categories_to_add:
dxpy.api.global_workflow_add_categories(global_workflow_id,
input_params={'categories': list(categories_to_add)})
if categories_to_remove:
dxpy.api.global_workflow_remove_categories(global_workflow_id,
input_params={'categories': list(categories_to_remove)})
| set_categories_on_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(global_workflow_id, categories_to_set):\n assert isinstance(categories_to_set, list)\n existing_categories = dxpy.api.global_workflow_list_categories(\n ... |
def x(json_spec, name=None, version=None):
requested_name = json_spec['name']
requested_version = json_spec['version']
if requested_name == name and requested_version == version:
return True
else:
try:
desc_output = dxpy.api.global_workflow_describe(
'globalworkflow-' + json_spec['name'], alias=json_spec[
'version'], input_params={'fields': {'name': True,
'version': True}})
return desc_output['name'] == json_spec['name'] and desc_output[
'version'] == json_spec['version']
except dxpy.exceptions.DXAPIError:
return False
except:
raise
| version_exists | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, name=None, version=None):\n requested_name = json_spec['name']\n requested_version = json_spec['version']\n if requested_name == name and r... |
def x(stages, ignore_reuse_stages):
if not isinstance(ignore_reuse_stages, list):
raise WorkflowBuilderException(
'"IgnoreReuse must be a list of strings - stage IDs or "*"')
ignore_reuse_set = set(ignore_reuse_stages)
if '*' in ignore_reuse_set and ignore_reuse_set == 1:
return
stage_ids = set([stage.get('id') for stage in stages])
for ignored in ignore_reuse_set:
if ignored not in stage_ids:
raise WorkflowBuilderException(
'Stage with ID {} not found. Add a matching "id" for the stage you wish to set ignoreReuse for'
.format(ignored))
| validate_ignore_reuse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(stages, ignore_reuse_stages):\n if not isinstance(ignore_reuse_stages, list):\n raise WorkflowBuilderException(\n '\"IgnoreReuse must be a... |
def x(stages):
if not isinstance(stages, list):
raise WorkflowBuilderException(
'Stages must be specified as a list of dictionaries')
validated_stages = []
for index, stage in enumerate(stages):
validated_stages.append(_get_validated_stage(stage, index))
return validated_stages
| get_validated_stages | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(stages):\n if not isinstance(stages, list):\n raise WorkflowBuilderException(\n 'Stages must be specified as a list of dictionaries')\n ... |
def x(json_spec, args):
validated = {}
override_project_id, override_folder, override_workflow_name = (dxpy.
executable_builder.get_parsed_destination(args.destination))
validated['project'] = _get_destination_project(json_spec, args,
override_project_id)
validated['folder'] = _get_destination_folder(json_spec, override_folder)
workflow_name = _get_workflow_name(json_spec, override_workflow_name)
if not workflow_name:
print('Warning: workflow name is not specified')
else:
validated['name'] = workflow_name
return validated
| validate_json_for_regular_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n validated = {}\n override_project_id, override_folder, override_workflow_name = (dxpy.\n executable_builder.get_parsed_destina... |
def x(json_spec, args):
if 'name' not in json_spec:
raise WorkflowBuilderException(
"dxworkflow.json contains no 'name' field, but it is required to build a global workflow"
)
if not dxpy.executable_builder.GLOBAL_EXEC_NAME_RE.match(json_spec['name']
):
raise WorkflowBuilderException(
'The name of your workflow must match /^[a-zA-Z0-9._-]+$/')
if json_spec['name'] != json_spec['name'].lower():
logger.warn('workflow name "{}" should be all lowercase'.format(
json_spec['name']))
if 'version' not in json_spec:
raise WorkflowBuilderException(
"dxworkflow.json contains no 'version' field, but it is required to build a global workflow"
)
if not dxpy.executable_builder.GLOBAL_EXEC_VERSION_RE.match(json_spec[
'version']):
logger.warn(
'"version" {} should be semver compliant (e.g. of the form X.Y.Z)'
.format(json_spec['version']))
if 'details' in json_spec:
if not isinstance(json_spec['details'], dict):
raise WorkflowBuilderException(
'The field "details" must be a dictionary')
if 'regionalOptions' in json_spec:
if not (isinstance(json_spec['regionalOptions'], dict) and
json_spec['regionalOptions'] and all([isinstance(i, dict) for i in
json_spec['regionalOptions'].values()])):
raise WorkflowBuilderException(
'The field "regionalOptions" must be a non-empty dictionary whose values are dictionaries'
)
if args.bill_to:
json_spec['billTo'] = args.bill_to
| validate_json_for_global_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n if 'name' not in json_spec:\n raise WorkflowBuilderException(\n \"dxworkflow.json contains no 'name' field, but it is ... |
def x(json_spec, args):
if not json_spec:
return
if not args:
return
validated_spec = copy.deepcopy(json_spec)
unsupported_keys = _get_unsupported_keys(validated_spec.keys(),
SUPPORTED_KEYS)
if len(unsupported_keys) > 0:
logger.warn(
'Warning: the following root level fields are not supported and will be ignored: {}'
.format(', '.join(unsupported_keys)))
if 'stages' in validated_spec:
validated_spec['stages'] = _get_validated_stages(validated_spec[
'stages'])
if 'name' in validated_spec:
if args.src_dir != validated_spec['name']:
logger.warn(
'workflow name "%s" does not match containing directory "%s"' %
(validated_spec['name'], args.src_dir))
if 'ignoreReuse' in validated_spec:
validate_ignore_reuse(validated_spec['stages'], validated_spec[
'ignoreReuse'])
validated_documentation_fields = _get_validated_json_for_build_or_update(
validated_spec, args)
validated_spec.update(validated_documentation_fields)
if args.mode == 'workflow':
validated = _validate_json_for_regular_workflow(json_spec, args)
validated_spec.update(validated)
if args.mode == 'globalworkflow':
_validate_json_for_global_workflow(validated_spec, args)
return validated_spec
| get_validated_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n if not json_spec:\n return\n if not args:\n return\n validated_spec = copy.deepcopy(json_spec)\n unsupported_keys... |
def x(json_spec, args):
validated = copy.deepcopy(json_spec)
dxpy.executable_builder.inline_documentation_files(validated, args.src_dir)
if 'title' not in json_spec:
logger.warn(
"dxworkflow.json is missing a title, please add one in the 'title' field"
)
if 'summary' not in json_spec:
logger.warn(
"dxworkflow.json is missing a summary, please add one in the 'summary' field"
)
elif json_spec['summary'].endswith('.'):
logger.warn(
'summary {} should be a short phrase not ending in a period'.
format(json_spec['summary']))
return validated
| get_validated_json_for_build_or_update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n validated = copy.deepcopy(json_spec)\n dxpy.executable_builder.inline_documentation_files(validated, args.src_dir)\n if 'title' no... |
def x(workflow_enabled_regions, workflow_spec):
executables = [i.get('executable') for i in workflow_spec.get('stages')]
for exect in executables:
if exect.startswith('applet-') and len(workflow_enabled_regions) > 1:
raise WorkflowBuilderException(
'Building a global workflow with applets in more than one region is not yet supported.'
)
elif exect.startswith('app-'):
app_regional_options = dxpy.api.app_describe(exect,
input_params={'fields': {'regionalOptions': True}})
app_regions = set(app_regional_options['regionalOptions'].keys())
if not workflow_enabled_regions.issubset(app_regions):
additional_workflow_regions = (workflow_enabled_regions -
app_regions)
mesg = (
'The app {} is enabled in regions {} while the global workflow in {}.'
.format(exect, ', '.join(app_regions), ', '.join(
workflow_enabled_regions)))
mesg += ' The workflow will not be able to run in {}.'.format(
', '.join(additional_workflow_regions))
mesg += (
' If you are a developer of the app, you can enable the app in {} to run the workflow in that region(s).'
.format(', '.join(additional_workflow_regions)))
logger.warn(mesg)
elif exect.startswith('workflow-'):
inner_workflow_spec = dxpy.api.workflow_describe(exect)
_assert_executable_regions_match(workflow_enabled_regions,
inner_workflow_spec)
elif exect.startswith('globalworkflow-'):
raise WorkflowBuilderException(
'Building a global workflow with nested global workflows is not yet supported'
)
| assert_executable_regions_match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(workflow_enabled_regions, workflow_spec):\n executables = [i.get('executable') for i in workflow_spec.get('stages')]\n for exect in executables:\n ... |
def x(json_spec):
workflow_id = dxpy.api.workflow_new(json_spec)['id']
dxpy.api.workflow_close(workflow_id)
return workflow_id
| build_regular_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec):\n workflow_id = dxpy.api.workflow_new(json_spec)['id']\n dxpy.api.workflow_close(workflow_id)\n return workflow_id\n",
"role": "user"
... |
def x(json_spec, from_command_line):
enabled_regions = dxpy.executable_builder.get_enabled_regions(
'globalworkflow', json_spec, from_command_line,
WorkflowBuilderException)
if not enabled_regions:
enabled_regions = []
if not dxpy.WORKSPACE_ID:
msg = (
"A context project must be selected to enable a workflow in the project's region."
)
msg += (
" You can use 'dx select' to select a project. Otherwise you can use --region option"
)
msg += (
' to select a region in which the workflow should be enabled')
raise WorkflowBuilderException(msg)
region = dxpy.api.project_describe(dxpy.WORKSPACE_ID, input_params=
{'fields': {'region': True}})['region']
enabled_regions.append(region)
if not enabled_regions:
raise AssertionError(
'This workflow should be enabled in at least one region')
return set(enabled_regions)
| get_validated_enabled_regions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, from_command_line):\n enabled_regions = dxpy.executable_builder.get_enabled_regions(\n 'globalworkflow', json_spec, from_command_line,\n ... |
def x(enabled_regions, args):
projects_by_region = {}
for region in enabled_regions:
try:
project_input = {'name':
'Temporary build project for dx build global workflow',
'region': region}
if args.bill_to:
project_input['billTo'] = args.bill_to
temp_project = dxpy.api.project_new(project_input)['id']
projects_by_region[region] = temp_project
logger.debug('Created temporary project {} to build in'.format(
temp_project))
except:
if projects_by_region:
dxpy.executable_builder.delete_temporary_projects(
projects_by_region.values())
err_exit()
return projects_by_region
| create_temporary_projects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(enabled_regions, args):\n projects_by_region = {}\n for region in enabled_regions:\n try:\n project_input = {'name':\n '... |
def x(enabled_regions, json_spec, args):
projects_by_region = _create_temporary_projects(enabled_regions, args)
workflows_by_region = {}
try:
for region, project in projects_by_region.items():
json_spec['project'] = project
workflow_id = _build_regular_workflow(json_spec)
logger.debug('Created workflow ' + workflow_id + ' successfully')
workflows_by_region[region] = workflow_id
except:
if projects_by_region:
dxpy.executable_builder.delete_temporary_projects(
projects_by_region.values())
raise
return workflows_by_region, projects_by_region
| build_underlying_workflows | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(enabled_regions, json_spec, args):\n projects_by_region = _create_temporary_projects(enabled_regions, args)\n workflows_by_region = {}\n try:\n ... |
def x(json_spec, args):
enabled_regions = _get_validated_enabled_regions(json_spec, args.region)
_assert_executable_regions_match(enabled_regions, json_spec)
workflows_by_region, projects_by_region = {}, {}
try:
workflows_by_region, projects_by_region = _build_underlying_workflows(
enabled_regions, json_spec, args)
regional_options = {}
for region, workflow_id in workflows_by_region.items():
regional_options[region] = {'workflow': workflow_id}
json_spec.update({'regionalOptions': regional_options})
gwf_provided_keys = GLOBALWF_SUPPORTED_KEYS.intersection(set(
json_spec.keys()))
gwf_final_json = dict((k, v) for k, v in json_spec.items() if k in
gwf_provided_keys)
print_spec = copy.deepcopy(gwf_final_json)
if 'description' in gwf_final_json:
del print_spec['description']
if 'developerNotes' in gwf_final_json:
del print_spec['developerNotes']
logger.info('Will create global workflow with spec: {}'.format(json
.dumps(print_spec)))
global_workflow_id = dxpy.api.global_workflow_new(gwf_final_json)['id']
logger.info('Uploaded global workflow {n}/{v} successfully'.format(
n=gwf_final_json['name'], v=gwf_final_json['version']))
logger.info('You can publish this workflow with:')
logger.info(' dx publish {n}/{v}'.format(n=gwf_final_json['name'],
v=gwf_final_json['version']))
finally:
if projects_by_region:
dxpy.executable_builder.delete_temporary_projects(
projects_by_region.values())
try:
_set_categories_on_workflow(global_workflow_id, gwf_final_json.get(
'categories', []))
except:
logger.warn(
'The workflow {n}/{v} was created but setting categories failed'
.format(n=gwf_final_json['name'], v=gwf_final_json['version']))
raise
return global_workflow_id
| build_global_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n enabled_regions = _get_validated_enabled_regions(json_spec, args.region)\n _assert_executable_regions_match(enabled_regions, json_spe... |
def x(json_spec, args):
try:
if args.mode == 'workflow':
json_spec = _get_validated_json(json_spec, args)
workflow_id = _build_regular_workflow(json_spec)
elif args.mode == 'globalworkflow':
existing_workflow = (dxpy.executable_builder.
verify_developer_rights('globalworkflow-' + json_spec['name']))
if existing_workflow and _version_exists(json_spec,
existing_workflow.name, existing_workflow.version):
workflow_id = _update_global_workflow(json_spec, args,
existing_workflow.id)
else:
json_spec = _get_validated_json(json_spec, args)
workflow_id = _build_global_workflow(json_spec, args)
else:
raise WorkflowBuilderException('Unrecognized workflow type: {}'
.format(args.mode))
except dxpy.exceptions.DXAPIError as e:
raise e
return workflow_id
| build_or_update_workflow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(json_spec, args):\n try:\n if args.mode == 'workflow':\n json_spec = _get_validated_json(json_spec, args)\n workflow_id = _buil... |
def x(args, parser):
if args is None:
raise Exception('arguments not provided')
try:
json_spec = _parse_executable_spec(args.src_dir, 'dxworkflow.json',
parser)
workflow_id = _build_or_update_workflow(json_spec, args)
_print_output(workflow_id, args)
except WorkflowBuilderException as e:
print('Error: %s' % (e.args,), file=sys.stderr)
sys.exit(3)
| build | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, parser):\n if args is None:\n raise Exception('arguments not provided')\n try:\n json_spec = _parse_executable_spec(args.src_dir, 'dx... |
def x(args, dest_path, dx_dest, dest_proj):
if len(args.sources) != 1:
raise DXCLIError('The destination folder does not exist')
last_slash_pos = get_last_pos_of_char('/', dest_path)
if last_slash_pos == 0:
dest_folder = '/'
else:
dest_folder = dest_path[:last_slash_pos]
dest_name = dest_path[last_slash_pos + 1:].replace('\\/', '/')
try:
dx_dest.list_folder(folder=dest_folder, only='folders')
except dxpy.DXAPIError as details:
if details.code == requests.codes['not_found']:
raise DXCLIError('The destination folder does not exist')
else:
raise
except:
err_exit()
src_proj, src_path, src_results = try_call(resolve_existing_path, args.
sources[0], allow_mult=True, all_mult=args.all)
if src_proj == dest_proj:
if is_hashid(args.sources[0]):
raise DXCLIError(fill(
'Error: You must specify a source project for ' + args.
sources[0]))
else:
raise DXCLIError(fill(
'A source path and the destination path resolved to the ' +
'same project or container. Please specify different source '
+ 'and destination containers, e.g.') +
"""
dx cp source-project:source-id-or-path dest-project:dest-path"""
)
if src_results is None:
try:
contents = dxpy.api.project_list_folder(src_proj, {'folder':
src_path, 'includeHidden': True})
dxpy.api.project_new_folder(dest_proj, {'folder': dest_path})
exists = dxpy.api.project_clone(src_proj, {'folders': contents[
'folders'], 'objects': [result['id'] for result in contents
['objects']], 'project': dest_proj, 'destination': dest_path})[
'exists']
if len(exists) > 0:
print(fill(
'The following objects already existed in the destination '
+ 'container and were not copied:') + '\n ' + '\n '.
join(exists))
return
except:
err_exit()
else:
try:
exists = dxpy.api.project_clone(src_proj, {'objects': [result[
'id'] for result in src_results], 'project': dest_proj,
'destination': dest_folder})['exists']
if len(exists) > 0:
print(fill(
'The following objects already existed in the destination '
+ 'container and were not copied:') + '\n ' + '\n '.
join(exists))
for result in src_results:
if result['id'] not in exists:
dxpy.DXHTTPRequest('/' + result['id'] + '/rename', {
'project': dest_proj, 'name': dest_name})
return
except:
err_exit()
| cp_to_noexistent_destination | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, dest_path, dx_dest, dest_proj):\n if len(args.sources) != 1:\n raise DXCLIError('The destination folder does not exist')\n last_slash_pos = ... |
def x(num_bytes):
B = num_bytes
KB = float(1024)
MB = float(KB * 1024)
GB = float(MB * 1024)
TB = float(GB * 1024)
if B < KB:
return '{0} {1}'.format(B, 'bytes' if B != 1 else 'byte')
elif KB <= B < MB:
return '{0:.2f} KiB'.format(B / KB)
elif MB <= B < GB:
return '{0:.2f} MiB'.format(B / MB)
elif GB <= B < TB:
return '{0:.2f} GiB'.format(B / GB)
elif TB <= B:
return '{0:.2f} TiB'.format(B / TB)
| readable_part_size | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_bytes):\n B = num_bytes\n KB = float(1024)\n MB = float(KB * 1024)\n GB = float(MB * 1024)\n TB = float(GB * 1024)\n if B < KB:\n ... |
def x(self, dx_hash, media_type=None, **kwargs):
if media_type is not None:
dx_hash['media'] = media_type
resp = dxpy.api.file_new(dx_hash, **kwargs)
self.set_ids(resp['id'], dx_hash['project'])
| new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dx_hash, media_type=None, **kwargs):\n if media_type is not None:\n dx_hash['media'] = media_type\n resp = dxpy.api.file_new(dx_hash, **kwar... |
def x(self, dxid, project=None):
if self._dxid is not None:
self.flush()
DXDataObject.set_ids(self, dxid, project)
self._pos = 0
self._file_length = None
self._cur_part = 1
self._num_uploaded_parts = 0
| set_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dxid, project=None):\n if self._dxid is not None:\n self.flush()\n DXDataObject.set_ids(self, dxid, project)\n self._pos = 0\n self._f... |
def x(self, offset, from_what=os.SEEK_SET):
if from_what == os.SEEK_SET:
reference_pos = 0
elif from_what == os.SEEK_CUR:
reference_pos = self._pos
elif from_what == os.SEEK_END:
if self._file_length == None:
desc = self.describe()
self._file_length = int(desc['size'])
reference_pos = self._file_length
else:
raise DXFileError('Invalid value supplied for from_what')
orig_pos = self._pos
self._pos = reference_pos + offset
in_buf = False
orig_buf_pos = self._read_buf.tell()
if offset < orig_pos:
if orig_buf_pos > orig_pos - offset:
in_buf = True
else:
buf_len = dxpy.utils.string_buffer_length(self._read_buf)
if buf_len - orig_buf_pos > offset - orig_pos:
in_buf = True
if in_buf:
self._read_buf.seek(orig_buf_pos - orig_pos + offset)
elif offset == orig_pos:
pass
else:
self._read_buf = BytesIO()
self._request_iterator, self._response_iterator = None, None
| seek | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, offset, from_what=os.SEEK_SET):\n if from_what == os.SEEK_SET:\n reference_pos = 0\n elif from_what == os.SEEK_CUR:\n reference_pos =... |
def x(self, multithread=True, **kwargs):
if self._write_buf.tell() > 0:
data = self._write_buf.getvalue()
self._write_buf = BytesIO()
if multithread:
self._async_upload_part_request(data, index=self._cur_part, **
kwargs)
else:
self.upload_part(data, self._cur_part, **kwargs)
self._cur_part += 1
if len(self._http_threadpool_futures) > 0:
dxpy.utils.wait_for_all_futures(self._http_threadpool_futures)
try:
for future in self._http_threadpool_futures:
if future.exception() != None:
raise future.exception()
finally:
self._http_threadpool_futures = set()
| flush | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, multithread=True, **kwargs):\n if self._write_buf.tell() > 0:\n data = self._write_buf.getvalue()\n self._write_buf = BytesIO()\n ... |
def x(self, data, multithread=True, **kwargs):
if not USING_PYTHON2:
assert isinstance(data, bytes)
self._ensure_write_bufsize(**kwargs)
def x(data_for_write_req):
if multithread:
self._async_upload_part_request(data_for_write_req, index=self.
_cur_part, **kwargs)
else:
self.upload_part(data_for_write_req, self._cur_part, **kwargs)
self._cur_part += 1
if self._write_buf.tell() == 0 and self._write_bufsize == len(data):
write_request(data)
return
remaining_space = self._write_bufsize - self._write_buf.tell()
if len(data) <= remaining_space:
self._write_buf.write(data)
else:
self._write_buf.write(data[:remaining_space])
temp_data = self._write_buf.getvalue()
self._write_buf = BytesIO()
write_request(temp_data)
self.write(data[remaining_space:], **kwargs)
| write2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, multithread=True, **kwargs):\n if not USING_PYTHON2:\n assert isinstance(data, bytes)\n self._ensure_write_bufsize(**kwargs)\n\n de... |
def x(self, data, multithread=True, **kwargs):
if USING_PYTHON2:
self._write2(data, multithread=multithread, **kwargs)
else:
if isinstance(data, str):
bt = data.encode('utf-8')
elif isinstance(data, bytearray):
bt = bytes(data)
elif isinstance(data, bytes):
bt = data
elif isinstance(data, mmap.mmap):
bt = bytes(data)
else:
raise DXFileError('Invalid type {} for write data argument'.
format(type(data)))
assert isinstance(bt, bytes)
self._write2(bt, multithread=multithread, **kwargs)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, multithread=True, **kwargs):\n if USING_PYTHON2:\n self._write2(data, multithread=multithread, **kwargs)\n else:\n if isinstanc... |
def x(self, block=False, **kwargs):
self.flush(**kwargs)
self._ensure_write_bufsize(**kwargs)
if self._num_uploaded_parts == 0 and self._empty_last_part_allowed:
try:
if USING_PYTHON2:
self.upload_part('', 1, **kwargs)
else:
self.upload_part(b'', 1, **kwargs)
except dxpy.exceptions.InvalidState:
pass
if 'report_progress_fn' in kwargs:
del kwargs['report_progress_fn']
dxpy.api.file_close(self._dxid, **kwargs)
if block:
self._wait_on_close(**kwargs)
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block=False, **kwargs):\n self.flush(**kwargs)\n self._ensure_write_bufsize(**kwargs)\n if self._num_uploaded_parts == 0 and self._empty_last_pa... |
def x(self, data, index=None, display_progress=False, report_progress_fn=
None, **kwargs):
if not USING_PYTHON2:
assert isinstance(data, bytes)
req_input = {}
if index is not None:
req_input['index'] = int(index)
md5 = hashlib.md5()
if hasattr(data, 'seek') and hasattr(data, 'tell'):
rewind_input_buffer_offset = data.tell()
while True:
bytes_read = data.read(MD5_READ_CHUNK_SIZE)
if bytes_read:
md5.update(bytes_read)
else:
break
data.seek(rewind_input_buffer_offset)
else:
md5.update(data)
req_input['md5'] = md5.hexdigest()
req_input['size'] = len(data)
def x():
if 'max_retries' not in kwargs:
kwargs['max_retries'] = dxpy.DEFAULT_RETRIES
elif kwargs['max_retries'] > 0:
kwargs['max_retries'] -= 1
if 'timeout' not in kwargs:
kwargs['timeout'] = FILE_REQUEST_TIMEOUT
resp = dxpy.api.file_upload(self._dxid, req_input, **kwargs)
url = resp['url']
return url, _validate_headers(resp.get('headers', {}))
dxpy.DXHTTPRequest(get_upload_url_and_headers, data, jsonify_data=False,
prepend_srv=False, always_retry=True, timeout=FILE_REQUEST_TIMEOUT,
auth=None, method='PUT')
self._num_uploaded_parts += 1
if display_progress:
warn('.')
if report_progress_fn is not None:
report_progress_fn(self, len(data))
| upload_part | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, index=None, display_progress=False, report_progress_fn=\n None, **kwargs):\n if not USING_PYTHON2:\n assert isinstance(data, bytes)\n ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.