source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, message, color=None): output_to = stderr if color == 'red' else stdout secho(self.indent(message), fg=color, file=output_to)
output
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message, color=None):\n output_to = stderr if color == 'red' else stdout\n secho(self.indent(message), fg=color, file=output_to)\n", "role": "u...
def x(self, url, params=None, **kwargs): return requests.get(url, params=params, headers=self.add_headers(**kwargs))
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url, params=None, **kwargs):\n return requests.get(url, params=params, headers=self.add_headers(**kwargs))\n", "role": "user" }, { "conten...
def x(self, url, data=None, **kwargs): return requests.patch(url, data=data, headers=self.add_headers(**kwargs))
patch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url, data=None, **kwargs):\n return requests.patch(url, data=data, headers=self.add_headers(**kwargs))\n", "role": "user" }, { "content": ...
def x(self, url, data=None, **kwargs): return requests.post(url, data=data, headers=self.add_headers(**kwargs))
post
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url, data=None, **kwargs):\n return requests.post(url, data=data, headers=self.add_headers(**kwargs))\n", "role": "user" }, { "content": "...
def x(self, content): if os.path.exists(self.file_path) and not self.assume_yes: message = 'Overwrite existing {}? (y/n) ' if not confirm(message.format(self.filename)): self.backup() self.output('Saving ' + self.filename) with open(self.file_path, 'wb') as handler: if not isinstance(content, bytes): content = bytes(content, 'utf-8') handler.write(content) self.yeah('Done!')
save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, content):\n if os.path.exists(self.file_path) and not self.assume_yes:\n message = 'Overwrite existing {}? (y/n) '\n if not confirm(mess...
def x(self): count = 0 name = '{}.bkp'.format(self.filename) backup = os.path.join(self.cwd, name) while os.path.exists(backup): count += 1 name = '{}.bkp{}'.format(self.filename, count) backup = os.path.join(self.cwd, name) self.hey('Moving existing {} to {}'.format(self.filename, name)) os.rename(os.path.join(self.cwd, self.filename), backup)
backup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n count = 0\n name = '{}.bkp'.format(self.filename)\n backup = os.path.join(self.cwd, name)\n while os.path.exists(backup):\n count +...
def x(self, file_path=None): if not file_path: file_path = self.file_path if not os.path.exists(file_path): self.oops('Sorry, but {} does not exist'.format(file_path)) return False if not os.path.isfile(file_path): self.oops('Sorry, but {} is not a file'.format(file_path)) return False with open(file_path) as handler: return handler.read()
read
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_path=None):\n if not file_path:\n file_path = self.file_path\n if not os.path.exists(file_path):\n self.oops('Sorry, but {} does...
def x(mode): def x(inp_str, pos): return mode if inp_str[pos] is mode else None def x(inp_str, pos): ch = inp_str[pos] for each in mode: if ch is each: return ch return None if isinstance(mode, str): return f_raw if len(mode) is 1: mode = mode[0] return f_raw return f_collection
char_matcher
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mode):\n\n def x(inp_str, pos):\n return mode if inp_str[pos] is mode else None\n\n def x(inp_str, pos):\n ch = inp_str[pos]\n for e...
def x(mode): def x(inp_str, pos): return unique_literal_cache_pool[mode] if inp_str.startswith(mode, pos ) else None def x(inp_str, pos): for each in mode: if inp_str.startswith(each, pos): return unique_literal_cache_pool[each] return None if isinstance(mode, str): return f_raw if len(mode) is 1: mode = mode[0] return f_raw return f_collection
str_matcher
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mode):\n\n def x(inp_str, pos):\n return unique_literal_cache_pool[mode] if inp_str.startswith(mode, pos\n ) else None\n\n def x(inp_st...
def x(regex_pat): if isinstance(regex_pat, str): regex_pat = re.compile(regex_pat) def x(inp_str, pos): m = regex_pat.match(inp_str, pos) return m.group() if m else None return f
regex_matcher
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(regex_pat):\n if isinstance(regex_pat, str):\n regex_pat = re.compile(regex_pat)\n\n def x(inp_str, pos):\n m = regex_pat.match(inp_str, po...
def x(self, stmts: T) ->None: if not stmts: raise ValueError('no ast found!') head, *equals = stmts if head.name is NameEnum.TokenDef: self.ast_for_token_def(head) elif head.name is NameEnum.TokenIgnore: self.ast_for_token_ignore(head) else: self.ast_for_equals(head) for each in equals: self.ast_for_equals(each) if not self.compile_helper.alone and self._current__combined_parser_name: self.compile_helper.alone.add(self._current__combined_parser_name)
ast_for_stmts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, stmts: T) ->None:\n if not stmts:\n raise ValueError('no ast found!')\n head, *equals = stmts\n if head.name is NameEnum.TokenDef:\n ...
def x(self, method, resource_uri, **kwargs): data = kwargs.get('data') response = method(self.API_BASE_URL + resource_uri, json=data, headers= self.headers) response.raise_for_status() return response.json()
request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, method, resource_uri, **kwargs):\n data = kwargs.get('data')\n response = method(self.API_BASE_URL + resource_uri, json=data, headers=\n sel...
def x(self, endpoint, **kwargs): return self._request(requests.get, endpoint, **kwargs)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, endpoint, **kwargs):\n return self._request(requests.get, endpoint, **kwargs)\n", "role": "user" }, { "content": "get", "role": "assis...
def x(self, endpoint, **kwargs): return self._request(requests.post, endpoint, **kwargs)
post
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, endpoint, **kwargs):\n return self._request(requests.post, endpoint, **kwargs)\n", "role": "user" }, { "content": "post", "role": "ass...
def x(self, endpoint, **kwargs): return self._request(requests.put, endpoint, **kwargs)
put
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, endpoint, **kwargs):\n return self._request(requests.put, endpoint, **kwargs)\n", "role": "user" }, { "content": "put", "role": "assis...
def x(cls, customer_id, **kwargs): return cls().requests.put('customer/{customer_id}'.format(**locals()), data=kwargs)
update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, customer_id, **kwargs):\n return cls().requests.put('customer/{customer_id}'.format(**locals()),\n data=kwargs)\n", "role": "user" }, { ...
def x(txt): txt = re.sub('<#[^\\|]*\\|(.*)>', '#\\g<1>', txt) txt = re.sub('<(@.*)>', '\\g<1>', txt) txt = re.sub('<([^\\|]*)\\|([^\\|]*)>', '<a href="\\g<1>" target="blank">\\g<2></a>', txt) txt = re.sub('<([^a|/a].*)>', '<a href="\\g<1>" target="blank">\\g<1></a>', txt) for delimeter in LIST_DELIMITERS: slack_tag = delimeter class_name = LIST_DELIMITERS[delimeter] list_regex = u'(?:^|\n){}\\s?(.*)'.format(slack_tag) list_repl = '<li class="list-item-{}">\\g<1></li>'.format(class_name) txt = re.sub(list_regex, list_repl, txt) txt = re.sub(u'(^|\n)(?:&gt;){3}\\s?(.*)$', '\\g<1><blockquote>\\g<2></blockquote>', txt, flags=re.DOTALL) txt = re.sub(u'(?:^|\n)&gt;\\s?(.*)\n?', '<blockquote>\\g<1></blockquote>', txt) txt = re.sub('```\\n?(.*)```', '<pre>\\g<1></pre>', txt, flags=re.DOTALL) txt = re.sub('\\n(</pre>)', '\\g<1>', txt) for wrapper in FORMATTERS: slack_tag = wrapper html_tag = FORMATTERS[wrapper] regex = '(?<!\\\\)\\{t}([^\\{t}|\\n]*)\\{t}'.format(t=slack_tag) repl = '<{t}>\\g<1></{t}>'.format(t=html_tag) txt = re.sub(regex, repl, txt) txt = txt.replace('\n', '<br />') parser = CustomSlackdownHTMLParser(txt) txt = parser.clean() txt = txt.replace(' ', ' &nbsp') return txt
render
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(txt):\n txt = re.sub('<#[^\\\\|]*\\\\|(.*)>', '#\\\\g<1>', txt)\n txt = re.sub('<(@.*)>', '\\\\g<1>', txt)\n txt = re.sub('<([^\\\\|]*)\\\\|([^\\\\|]*...
def x(self, list_type): if list_type in LIST_TYPES.keys(): tag = LIST_TYPES[list_type] else: raise Exception( 'CustomSlackdownHTMLParser:_open_list: Not a valid list type.') html = '<{t} class="list-container-{c}">'.format(t=tag, c=list_type) self.cleaned_html += html self.current_parent_element['tag'] = LIST_TYPES[list_type] self.current_parent_element['attrs'] = {'class': list_type}
open_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, list_type):\n if list_type in LIST_TYPES.keys():\n tag = LIST_TYPES[list_type]\n else:\n raise Exception(\n 'CustomSlackdo...
def x(self): list_type = self.current_parent_element['attrs']['class'] tag = LIST_TYPES[list_type] html = '</{t}>'.format(t=tag) self.cleaned_html += html self.current_parent_element['tag'] = '' self.current_parent_element['attrs'] = {}
close_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n list_type = self.current_parent_element['attrs']['class']\n tag = LIST_TYPES[list_type]\n html = '</{t}>'.format(t=tag)\n self.cleaned_htm...
def x(self, tag, attrs): attrs_dict = dict(t for t in attrs) if tag in PARENT_ELEMENTS: if self.current_parent_element['tag'] != '': self.cleaned_html += '</{}>'.format(self.current_parent_element ['tag']) self.current_parent_element['tag'] = tag self.current_parent_element['attrs'] = {} self.cleaned_html += '<{}>'.format(tag) elif tag == 'li': self.parsing_li = True class_name = attrs_dict['class'] list_type = class_name[10:] if self.current_parent_element['tag' ] == 'ul' or self.current_parent_element['tag'] == 'ol': cur_list_type = self.current_parent_element['attrs']['class'] if cur_list_type != list_type: self._close_list() self._open_list(list_type) else: if self.current_parent_element['tag'] != '': self.cleaned_html += '</{}>'.format(self. current_parent_element['tag']) self._open_list(list_type) self.cleaned_html += '<{}>'.format(tag) elif tag == 'br': if self.current_parent_element['tag'] == 'p': self.cleaned_html += '</p>' self.current_parent_element['tag'] = '' self.current_parent_element['attrs'] = {} elif self.current_parent_element['tag' ] == 'ul' or self.current_parent_element['tag'] == 'ol': self._close_list() elif self.current_parent_element['tag'] in PARENT_ELEMENTS: self.cleaned_html += '<br />' else: self.cleaned_html += '<p></p>' else: if self.current_parent_element['tag'] == '': self.cleaned_html += '<p>' self.current_parent_element['tag'] = 'p' self.cleaned_html += '<{}'.format(tag) for attr in sorted(attrs_dict.keys()): self.cleaned_html += ' {k}="{v}"'.format(k=attr, v=attrs_dict[attr] ) self.cleaned_html += '>'
handle_starttag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag, attrs):\n attrs_dict = dict(t for t in attrs)\n if tag in PARENT_ELEMENTS:\n if self.current_parent_element['tag'] != '':\n ...
def x(self, tag): if tag in PARENT_ELEMENTS: self.current_parent_element['tag'] = '' self.current_parent_element['attrs'] = '' if tag == 'li': self.parsing_li = True if tag != 'br': self.cleaned_html += '</{}>'.format(tag)
handle_endtag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag):\n if tag in PARENT_ELEMENTS:\n self.current_parent_element['tag'] = ''\n self.current_parent_element['attrs'] = ''\n if tag == ...
def x(self, data): if self.current_parent_element['tag'] == '': self.cleaned_html += '<p>' self.current_parent_element['tag'] = 'p' self.cleaned_html += data
handle_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n if self.current_parent_element['tag'] == '':\n self.cleaned_html += '<p>'\n self.current_parent_element['tag'] = 'p'\n self....
def x(self): preformatted_wrappers = ['pre', 'code'] for wrapper in preformatted_wrappers: for formatter in FORMATTERS: tag = FORMATTERS[formatter] character = formatter regex = '(<{w}>.*)<{t}>(.*)</{t}>(.*</{w}>)'.format(t=tag, w= wrapper) repl = '\\g<1>{c}\\g<2>{c}\\g<3>'.format(c=character) self.cleaned_html = re.sub(regex, repl, self.cleaned_html)
remove_pre_formatting
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n preformatted_wrappers = ['pre', 'code']\n for wrapper in preformatted_wrappers:\n for formatter in FORMATTERS:\n tag = FORMATT...
def x(self): self.feed() if self.current_parent_element['tag'] != '': self.cleaned_html += '</{}>'.format(self.current_parent_element['tag']) self.cleaned_html = re.sub('(</[u|o]l>)<p></p>', '\\g<1>', self. cleaned_html) self._remove_pre_formatting() return self.cleaned_html
clean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.feed()\n if self.current_parent_element['tag'] != '':\n self.cleaned_html += '</{}>'.format(self.current_parent_element['tag'])\n ...
def x(**kwargs): html = '<div class="user">' author_avatar = kwargs.get('author_avatar', None) if author_avatar: html += '<img class="user-avatar" src="{}">'.format(author_avatar) author_name = kwargs.get('author_name', None) if author_name: html += '<p class="user-name">' author_link = kwargs.get('author_link', None) if author_link: html += '<a href="{author_link}">{author_name}</a>'.format( author_link=author_link, author_name=author_name) else: html += author_name html += '</p>' author_handle = kwargs.get('author_handle', None) if author_handle: html += '<p class="user-handle">{}</p>'.format(author_handle) html += '</div>'
render_author
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n html = '<div class=\"user\">'\n author_avatar = kwargs.get('author_avatar', None)\n if author_avatar:\n html += '<img class=\"user...
def x(**kwargs): html = '<div class="metadata">' service_logo = kwargs.get('service_logo', None) if service_logo: html += '<img class="metadata-logo" src="{}">'.format(service_logo) service_name = kwargs.get('service_name', None) if service_name: html += '<p class="metadata-name">{}</p>'.format(service_name) timestamp = kwargs.get('timestamp', None) if timestamp: html += '<p class="user-name">' timestamp_link = kwargs.get('timestamp_link', None) if timestamp_link: html += '<a href="{timestamp_link}">{timestamp}</a>'.format( timestamp_link=timestamp_link, timestamp=timestamp) else: html += timestamp html += '</p>' html += '</div>'
render_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n html = '<div class=\"metadata\">'\n service_logo = kwargs.get('service_logo', None)\n if service_logo:\n html += '<img class=\"met...
def x(**kwargs): html = '' url = kwargs.get('url', None) if url: html = '<img' alt_text = kwargs.get('alt_text', None) if alt_text: html += ' alt="{}"'.format(alt_text) title = kwargs.get('title', None) if title: html += ' title="{}"'.format(title) html += ' src="{}">'.format(url) return html
render_image
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n html = ''\n url = kwargs.get('url', None)\n if url:\n html = '<img'\n alt_text = kwargs.get('alt_text', None)\n if a...
def x(text, **kwargs): author = render_author(**kwargs['author']) metadata = render_metadata(**kwargs['metadata']) image = render_image(**kwargs['image']) html = ( """ <div class="attachment attachment-twitter"> {author} <p class="twitter-content">{text}</p> {metadata} {image} </div> """ .format(author=author, text=text, metadata=metadata, image=image). strip()) return html
render_twitter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, **kwargs):\n author = render_author(**kwargs['author'])\n metadata = render_metadata(**kwargs['metadata'])\n image = render_image(**kwargs['imag...
def x(LAB_DIR): coeffs = np.load('%s/coeffs.npz' % LAB_DIR)['arr_0'] scatters = np.load('%s/scatters.npz' % LAB_DIR)['arr_0'] chisqs = np.load('%s/chisqs.npz' % LAB_DIR)['arr_0'] pivots = np.load('%s/pivots.npz' % LAB_DIR)['arr_0'] return coeffs, scatters, chisqs, pivots
get_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(LAB_DIR):\n coeffs = np.load('%s/coeffs.npz' % LAB_DIR)['arr_0']\n scatters = np.load('%s/scatters.npz' % LAB_DIR)['arr_0']\n chisqs = np.load('%s/chi...
def x(ids_find): a = pyfits.open('%s/lamost_catalog_full.fits' % LAB_DIR) data = a[1].data a.close() id_all = data['lamost_id'] id_all = np.array(id_all) id_all = np.array([val.strip() for val in id_all]) snr_all = data['cannon_snrg'] chisq_all = data['cannon_chisq'] teff = data['cannon_teff'] logg = data['cannon_logg'] feh = data['cannon_m_h'] afe = data['cannon_alpha_m'] ak = data['cannon_a_k'] labels = np.vstack((teff, logg, feh, afe, ak)) choose = np.in1d(id_all, ids_find) id_choose = id_all[choose] label_choose = labels[:, choose] snr_choose = snr_all[choose] chisq_choose = chisq_all[choose] inds = np.array([np.where(id_choose == val)[0][0] for val in ids_find]) print(id_choose[inds][100]) print(ids_find[100]) return label_choose[:, inds], snr_choose[inds], chisq_choose[inds]
get_labels
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ids_find):\n a = pyfits.open('%s/lamost_catalog_full.fits' % LAB_DIR)\n data = a[1].data\n a.close()\n id_all = data['lamost_id']\n id_all = np....
def x(): wl = np.load('%s/wl.npz' % LAB_DIR)['arr_0'] filenames = np.array([(SPEC_DIR + '/Spectra' + '/' + val) for val in lamost_id]) grid, fluxes, ivars, npix, SNRs = lamost.load_spectra(lamost_id, input_grid=wl) ds = dataset.Dataset(wl, lamost_id, fluxes, ivars, [1], lamost_id[0:2], fluxes[0:2], ivars[0:2]) ds.continuum_normalize_gaussian_smoothing(L=50) np.savez(SPEC_DIR + '/' + 'norm_flux.npz', ds.tr_flux) np.savez(SPEC_DIR + '/' + 'norm_ivar.npz', ds.tr_ivar) return ds.tr_flux, ds.tr_ivar
get_normed_spectra
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n wl = np.load('%s/wl.npz' % LAB_DIR)['arr_0']\n filenames = np.array([(SPEC_DIR + '/Spectra' + '/' + val) for val in\n lamost_id])\n grid, ...
def x(): for f in lamost_id: short = f.split('-')[2].split('_')[0] filename = '%s/%s.gz' % (short, f) DIR = '/Users/annaho/Data/Li_Giants/Spectra_APOKASC' searchfor = '%s/%s.gz' % (DIR, f) if glob.glob(searchfor): print('done') else: os.system('wget http://dr2.lamost.org/sas/fits/%s' % filename) new_filename = filename.split('_')[0] + '_' + filename.split('_')[2 ] os.system('wget http://dr2.lamost.org/sas/fits/%s' % new_filename)
wget_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n for f in lamost_id:\n short = f.split('-')[2].split('_')[0]\n filename = '%s/%s.gz' % (short, f)\n DIR = '/Users/annaho/Data/Li_Gi...
def x(): cannon_teff = data['cannon_teff_2'] cannon_logg = data['cannon_logg_2'] cannon_m_h = data['cannon_m_h'] cannon_alpha_m = data['cannon_alpha_m'] cannon_a_k = data['cannon_a_k'] labels = np.vstack((cannon_teff, cannon_logg, cannon_m_h, cannon_alpha_m, cannon_a_k)) cannon_chisq = data['cannon_chisq'] np.savez(DATA_DIR + 'chisq.npz', labels) np.savez(DATA_DIR + 'labels.npz', labels) snrg = data['cannon_snrg'] np.savez('snr.npz', snrg) return labels.T
get_labels
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n cannon_teff = data['cannon_teff_2']\n cannon_logg = data['cannon_logg_2']\n cannon_m_h = data['cannon_m_h']\n cannon_alpha_m = data['cannon_al...
def x(spec_raw): spec = np.array([spec_raw]) wl = np.arange(0, spec.shape[1]) w = continuum_normalization.gaussian_weight_matrix(wl, L=50) ivar = np.ones(spec.shape) * 0.5 cont = continuum_normalization._find_cont_gaussian_smooth(wl, spec, ivar, w ) norm_flux, norm_ivar = continuum_normalization._cont_norm(spec, ivar, cont) return norm_flux[0]
cannon_normalize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(spec_raw):\n spec = np.array([spec_raw])\n wl = np.arange(0, spec.shape[1])\n w = continuum_normalization.gaussian_weight_matrix(wl, L=50)\n ivar =...
def x(grid, wl, flux): flux_rs = interpolate.interp1d(wl, flux)(grid) return flux_rs
resample
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(grid, wl, flux):\n flux_rs = interpolate.interp1d(wl, flux)(grid)\n return flux_rs\n", "role": "user" }, { "content": "resample", "role":...
def x(choose, coeffs, pivots): base_labels = [4800, 2.5, 0.03, 0.1, -0.17, -0.17, 0, -0.16, -0.13, - 0.15, 0.13, 0.08, 0.17, -0.062] label_names = np.array(['TEFF', 'LOGG', 'AK', 'Al', 'Ca', 'C', 'Fe', 'Mg', 'Mn', 'Ni', 'N', 'O', 'Si', 'Ti']) label_atnum = np.array([0, 1, -1, 13, 20, 6, 26, 12, 25, 28, 7, 8, 14, 22]) ind = np.where(label_atnum == choose)[0][0] low_lab = copy.copy(base_labels) high = base_labels[ind] if choose > 0: low = base_labels[ind] - 0.2 else: if choose != 0: print('warning...') low = base_labels[ind] - 200 low_lab[ind] = low lvec = train_model._get_lvec(np.array([low_lab]), pivots)[0] model_low = np.dot(coeffs, lvec) lvec = train_model._get_lvec(np.array([base_labels]), pivots)[0] model_high = np.dot(coeffs, lvec) grad_spec = (model_high - model_low) / (high - low) return grad_spec
gen_cannon_grad_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(choose, coeffs, pivots):\n base_labels = [4800, 2.5, 0.03, 0.1, -0.17, -0.17, 0, -0.16, -0.13, -\n 0.15, 0.13, 0.08, 0.17, -0.062]\n label_names =...
def x(atomic_number): DATA_DIR = '/Users/annaho/Data/LAMOST/Mass_And_Age' temp = np.load('%s/X_u_template_KGh_res=1800.npz' % DATA_DIR) X_u_template = temp['X_u_template'] wl = temp['wavelength'] grad_spec = X_u_template[:, atomic_number] return wl, grad_spec
get_model_spec_ting
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(atomic_number):\n DATA_DIR = '/Users/annaho/Data/LAMOST/Mass_And_Age'\n temp = np.load('%s/X_u_template_KGh_res=1800.npz' % DATA_DIR)\n X_u_template =...
def x(ds, m): model_spectra = get_model_spectra(ds, m) resid = ds.test_flux - model_spectra return resid
get_residuals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ds, m):\n model_spectra = get_model_spectra(ds, m)\n resid = ds.test_flux - model_spectra\n return resid\n", "role": "user" }, { "content"...
def x(): direc = '/home/annaho/TheCannon/code/lamost/mass_age/cn' m = model.CannonModel(2) m.coeffs = np.load(direc + '/coeffs.npz')['arr_0'][0:3626, :] m.scatters = np.load(direc + '/scatters.npz')['arr_0'][0:3626] m.chisqs = np.load(direc + '/chisqs.npz')['arr_0'][0:3626] m.pivots = np.load(direc + '/pivots.npz')['arr_0'] return m
load_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n direc = '/home/annaho/TheCannon/code/lamost/mass_age/cn'\n m = model.CannonModel(2)\n m.coeffs = np.load(direc + '/coeffs.npz')['arr_0'][0:3626, ...
def x(date): LAB_DIR = '/home/annaho/TheCannon/data/lamost' WL_DIR = '/home/annaho/TheCannon/code/lamost/mass_age/cn' SPEC_DIR = ( '/home/annaho/TheCannon/code/apogee_lamost/xcalib_4labels/output') wl = np.load(WL_DIR + '/wl_cols.npz')['arr_0'][0:3626] ds = dataset.Dataset(wl, [], [], [], [], [], [], []) test_label = np.load('%s/%s_all_cannon_labels.npz' % (LAB_DIR, date))[ 'arr_0'] ds.test_label_vals = test_label a = np.load('%s/%s_norm.npz' % (SPEC_DIR, date)) ds.test_flux = a['arr_0'] ds.test_ivar = a['arr_1'] ds.test_ID = np.load('%s/%s_ids.npz' % (SPEC_DIR, date))['arr_0'] return ds
load_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(date):\n LAB_DIR = '/home/annaho/TheCannon/data/lamost'\n WL_DIR = '/home/annaho/TheCannon/code/lamost/mass_age/cn'\n SPEC_DIR = (\n '/home/ann...
def x(x, y, yerr, p0): try: popt, pcov = curve_fit(gaussian, x, y, sigma=yerr, p0=p0, absolute_sigma=True) except RuntimeError: return [0], [0] return popt, pcov
fit_gaussian
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, y, yerr, p0):\n try:\n popt, pcov = curve_fit(gaussian, x, y, sigma=yerr, p0=p0,\n absolute_sigma=True)\n except RuntimeError:\n ...
def x(yerrs, amps, amp_errs, widths): keep_1 = np.logical_and(amps < 0, widths > 1) keep_2 = np.logical_and(np.abs(amps) > 3 * yerrs, amp_errs < 3 * np.abs (amps)) keep = np.logical_and(keep_1, keep_2) return keep
select
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(yerrs, amps, amp_errs, widths):\n keep_1 = np.logical_and(amps < 0, widths > 1)\n keep_2 = np.logical_and(np.abs(amps) > 3 * yerrs, amp_errs < 3 * np.abs...
def x(): DATA_DIR = '/home/annaho/TheCannon/code/apogee_lamost/xcalib_4labels' dates = os.listdir('/home/share/LAMOST/DR2/DR2_release') dates = np.array(dates) dates = np.delete(dates, np.where(dates == '.directory')[0][0]) dates = np.delete(dates, np.where(dates == 'all_folders.list')[0][0]) dates = np.delete(dates, np.where(dates == 'dr2.lis')[0][0]) for date in dates: if glob.glob('*%s*.txt' % date): print('%s done' % date) else: print('running %s' % date) run_one_date(date)
run_all
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n DATA_DIR = '/home/annaho/TheCannon/code/apogee_lamost/xcalib_4labels'\n dates = os.listdir('/home/share/LAMOST/DR2/DR2_release')\n dates = np.arr...
def x(base_labels, choose, low, high, coeffs, pivots): low_lab = copy.copy(base_labels) low_lab[choose] = low lvec = train_model._get_lvec(np.array([low_lab]), pivots)[0] model_low = np.dot(coeffs, lvec) high_lab = copy.copy(base_labels) high_lab[choose] = high lvec = train_model._get_lvec(np.array([high_lab]), pivots)[0] model_high = np.dot(coeffs, lvec) grad_spec = (model_high - model_low) / (high - low) return grad_spec
gen_cannon_grad_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(base_labels, choose, low, high, coeffs, pivots):\n low_lab = copy.copy(base_labels)\n low_lab[choose] = low\n lvec = train_model._get_lvec(np.array([l...
def x(snr): quad_terms = np.array([0.00311, 1.1e-05, 6.95e-06, 5.05e-06, 4.65e-06, 4.1e-06]) lin_terms = np.array([-0.869, -0.00207, -0.0014, -0.00103, -0.00113, - 0.000729]) consts = np.array([104, 0.2, 0.117, 0.114, 0.156, 0.0624]) err = quad_terms[:, None] * snr ** 2 + lin_terms[:, None] * snr + consts[ :, None] min_snr = -lin_terms / (2 * quad_terms) min_err = quad_terms * min_snr ** 2 + lin_terms * min_snr + consts mask = (snr[:, None] > min_snr).T for i in range(0, len(min_err)): err[i][mask[i]] = min_err[i] return err
get_err
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(snr):\n quad_terms = np.array([0.00311, 1.1e-05, 6.95e-06, 5.05e-06, 4.65e-06, \n 4.1e-06])\n lin_terms = np.array([-0.869, -0.00207, -0.0014, -0....
def x(catalog): print('Get Colors') a = pyfits.open(catalog) data = a[1].data a.close() all_ids = data['LAMOST_ID_1'] all_ids = np.array([val.strip() for val in all_ids]) gmag = data['gpmag'] gmag_err = data['e_gpmag'] rmag = data['rpmag'] rmag_err = data['e_rpmag'] imag = data['ipmag'] imag_err = data['e_ipmag'] W1 = data['W1mag'] W1_err = data['e_W1mag'] W2 = data['W2mag'] W2_err = data['e_W2mag'] Jmag = data['Jmag'] Jmag_err = data['e_Jmag'] Hmag = data['Hmag'] Hmag_err = data['e_Hmag'] Kmag = data['Kmag'] Kmag_err = data['e_Kmag'] mag = np.vstack((gmag, rmag, imag, Jmag, Hmag, Kmag, W2, W1)) mag_err = np.vstack((gmag_err, rmag_err, imag_err, Jmag_err, Hmag_err, Kmag_err, W2_err, W1_err)) col = mag[:-1] - mag[1:] col_ivar = 1 / (mag_err[:-1] ** 2 + mag_err[1:] ** 2) return all_ids, col, col_ivar
get_colors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(catalog):\n print('Get Colors')\n a = pyfits.open(catalog)\n data = a[1].data\n a.close()\n all_ids = data['LAMOST_ID_1']\n all_ids = np.arra...
def x(md, ds): coeffs_all, covs, scatters, red_chisqs, pivots, label_vector = model.model nstars = len(dataset.test_SNR) cannon_flux = np.zeros(dataset.test_flux.shape) cannon_ivar = np.zeros(dataset.test_ivar.shape) for i in range(nstars): x = label_vector[:, i, :] spec_fit = np.einsum('ij, ij->i', x, coeffs_all) cannon_flux[i, :] = spec_fit bad = dataset.test_ivar[i, :] == SMALL ** 2 cannon_ivar[i, :][~bad] = 1.0 / scatters[~bad] ** 2 return cannon_flux, cannon_ivar
draw_spectra
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(md, ds):\n coeffs_all, covs, scatters, red_chisqs, pivots, label_vector = model.model\n nstars = len(dataset.test_SNR)\n cannon_flux = np.zeros(datase...
def x(model, dataset): best_flux, best_ivar = draw_spectra(model, dataset) coeffs_all, covs, scatters, all_chisqs, pivots, label_vector = model.model print('Overplotting spectra for ten random stars') res = dataset.test_flux - best_flux lambdas = dataset.wl npix = len(lambdas) nstars = best_flux.shape[0] pickstars = [] for i in range(10): pickstars.append(random.randrange(0, nstars - 1)) for i in pickstars: print('Star %s' % i) ID = dataset.test_ID[i] spec_orig = dataset.test_flux[i, :] bad = dataset.test_flux[i, :] == 0 lambdas = np.ma.array(lambdas, mask=bad, dtype=float) npix = len(lambdas.compressed()) spec_orig = np.ma.array(dataset.test_flux[i, :], mask=bad) spec_fit = np.ma.array(best_flux[i, :], mask=bad) ivars_orig = np.ma.array(dataset.test_ivar[i, :], mask=bad) ivars_fit = np.ma.array(best_ivar[i, :], mask=bad) red_chisq = np.sum(all_chisqs[:, i], axis=0) / (npix - coeffs_all. shape[1]) red_chisq = np.round(red_chisq, 2) fig, axarr = plt.subplots(2) ax1 = axarr[0] im = ax1.scatter(lambdas, spec_orig, label='Orig Spec', c=1 / np. sqrt(ivars_orig), s=10) ax1.scatter(lambdas, spec_fit, label='Cannon Spec', c='r', s=10) ax1.errorbar(lambdas, spec_fit, yerr=1 / np.sqrt(ivars_fit), fmt= 'ro', ms=1, alpha=0.7) ax1.set_xlabel('Wavelength $\\lambda (\\AA)$') ax1.set_ylabel('Normalized flux') ax1.set_title('Spectrum Fit: %s' % ID) ax1.set_title('Spectrum Fit') ax1.set_xlim(min(lambdas.compressed()) - 10, max(lambdas.compressed ()) + 10) ax1.legend(loc='lower center', fancybox=True, shadow=True) ax2 = axarr[1] ax2.scatter(spec_orig, spec_fit, c=1 / np.sqrt(ivars_orig), alpha=0.7) ax2.errorbar(spec_orig, spec_fit, yerr=1 / np.sqrt(ivars_fit), ecolor='k', fmt='none', ms=1, alpha=0.7) fig.colorbar() xlims = ax2.get_xlim() ylims = ax2.get_ylim() lims = [np.min([xlims, ylims]), np.max([xlims, ylims])] ax2.plot(lims, lims, 'k-', alpha=0.75) textstr = 'Red Chi Sq: %s' % red_chisq props = dict(boxstyle='round', facecolor='palevioletred', alpha=0.5) ax2.text(0.05, 0.95, textstr, transform=ax2.transAxes, fontsize=14, verticalalignment='top', bbox=props) ax2.set_xlim(xlims) ax2.set_ylim(ylims) ax2.set_xlabel('Orig Fluxes') ax2.set_ylabel('Fitted Fluxes') plt.tight_layout() filename = 'best_fit_spec_Star%s.png' % i print('Saved as %s' % filename) fig.savefig(filename) plt.close(fig)
overlay_spectra
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, dataset):\n best_flux, best_ivar = draw_spectra(model, dataset)\n coeffs_all, covs, scatters, all_chisqs, pivots, label_vector = model.model\n ...
def x(cannon_set, dataset): print('Stacking spectrum fit residuals') res = dataset.test_fluxes - cannon_set.test_fluxes bad = dataset.test_ivars == SMALL ** 2 err = np.zeros(len(dataset.test_ivars)) err = np.sqrt(1.0 / dataset.test_ivars + 1.0 / cannon_set.test_ivars) res_norm = res / err res_norm = np.ma.array(res_norm, mask=np.ones_like(res_norm) * (np.std( res_norm, axis=0) == 0)) res_norm = np.ma.compress_cols(res_norm) for i in range(len(cannon_set.get_plotting_labels())): label_name = cannon_set.get_plotting_labels()[i] print('Plotting residuals sorted by %s' % label_name) label_vals = cannon_set.tr_label_vals[:, i] sorted_res = res_norm[np.argsort(label_vals)] mu = np.mean(sorted_res.flatten()) sigma = np.std(sorted_res.flatten()) left, width = 0.1, 0.65 bottom, height = 0.1, 0.65 bottom_h = left_h = left + width + 0.1 rect_scatter = [left, bottom, width, height] rect_histx = [left, bottom_h, width, 0.1] rect_histy = [left_h, bottom, 0.1, height] plt.figure() axScatter = plt.axes(rect_scatter) axHistx = plt.axes(rect_histx) axHisty = plt.axes(rect_histy) im = axScatter.imshow(sorted_res, cmap=plt.cm.bwr_r, interpolation= 'nearest', vmin=mu - 3.0 * sigma, vmax=mu + 3.0 * sigma, aspect ='auto', origin='lower', extent=[0, len(dataset.wl), min( label_vals), max(label_vals)]) cax, kw = colorbar.make_axes(axScatter.axes, location='bottom') plt.colorbar(im, cax=cax, orientation='horizontal') axScatter.set_title('Spectral Residuals Sorted by ${0:s}$'.format( label_name)) axScatter.set_xlabel('Pixels') axScatter.set_ylabel('$%s$' % label_name) axHisty.hist(np.std(res_norm, axis=1)[~np.isnan(np.std(res_norm, axis=1))], orientation='horizontal', range=[0, 2]) axHisty.axhline(y=1, c='k', linewidth=3, label='y=1') axHisty.legend(bbox_to_anchor=(0.0, 0.8, 1.0, 0.102), prop={ 'family': 'serif', 'size': 'small'}) axHisty.text(1.0, 0.5, 'Distribution of Stdev of Star Residuals', verticalalignment='center', transform=axHisty.transAxes, rotation=270) axHisty.set_ylabel('Standard Deviation') start, end = axHisty.get_xlim() axHisty.xaxis.set_ticks(np.linspace(start, end, 3)) axHisty.set_xlabel('Number of Stars') axHisty.xaxis.set_label_position('top') axHistx.hist(np.std(res_norm, axis=0)[~np.isnan(np.std(res_norm, axis=0))], range=[0.8, 1.1]) axHistx.axvline(x=1, c='k', linewidth=3, label='x=1') axHistx.set_title('Distribution of Stdev of Pixel Residuals') axHistx.set_xlabel('Standard Deviation') axHistx.set_ylabel('Number of Pixels') start, end = axHistx.get_ylim() axHistx.yaxis.set_ticks(np.linspace(start, end, 3)) axHistx.legend() filename = 'residuals_sorted_by_label_%s.png' % i plt.savefig(filename) print('File saved as %s' % filename) plt.close() print('Plotting Auto-Correlation of Mean Residuals') mean_res = res_norm.mean(axis=0) autocorr = np.correlate(mean_res, mean_res, mode='full') pkwidth = int(len(autocorr) / 2 - np.argmin(autocorr)) xmin = int(len(autocorr) / 2) - pkwidth xmax = int(len(autocorr) / 2) + pkwidth zoom_x = np.linspace(xmin, xmax, len(autocorr[xmin:xmax])) fig, axarr = plt.subplots(2) axarr[0].plot(autocorr) axarr[0].set_title('Autocorrelation of Mean Spectral Residual') axarr[0].set_xlabel('Lag (# Pixels)') axarr[0].set_ylabel('Autocorrelation') axarr[1].plot(zoom_x, autocorr[xmin:xmax]) axarr[1].set_title('Central Peak, Zoomed') axarr[1].set_xlabel('Lag (# Pixels)') axarr[1].set_ylabel('Autocorrelation') filename = 'residuals_autocorr.png' plt.savefig(filename) print('saved %s' % filename) plt.close()
residuals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cannon_set, dataset):\n print('Stacking spectrum fit residuals')\n res = dataset.test_fluxes - cannon_set.test_fluxes\n bad = dataset.test_ivars == SM...
def x(f_cut, sig_cut, wl, fluxes, ivars): f_bar = np.median(fluxes, axis=0) sigma_f = np.var(fluxes, axis=0) bad = np.logical_and(f_bar == 0, sigma_f == 0) cont1 = np.abs(f_bar - 1) <= f_cut cont2 = sigma_f <= sig_cut contmask = np.logical_and(cont1, cont2) contmask[bad] = False return contmask
find_contpix_given_cuts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f_cut, sig_cut, wl, fluxes, ivars):\n f_bar = np.median(fluxes, axis=0)\n sigma_f = np.var(fluxes, axis=0)\n bad = np.logical_and(f_bar == 0, sigma_f ...
def x(wl, fluxes, ivars, target_frac): print('Target frac: %s' % target_frac) bad1 = np.median(ivars, axis=0) == SMALL bad2 = np.var(ivars, axis=0) == 0 bad = np.logical_and(bad1, bad2) npixels = len(wl) - sum(bad) f_cut = 0.0001 stepsize = 0.0001 sig_cut = 0.0001 contmask = _find_contpix_given_cuts(f_cut, sig_cut, wl, fluxes, ivars) if npixels > 0: frac = sum(contmask) / float(npixels) else: frac = 0 while frac < target_frac: f_cut += stepsize sig_cut += stepsize contmask = _find_contpix_given_cuts(f_cut, sig_cut, wl, fluxes, ivars) if npixels > 0: frac = sum(contmask) / float(npixels) else: frac = 0 if frac > 0.1 * npixels: print('Warning: Over 10% of pixels identified as continuum.') print('%s out of %s pixels identified as continuum' % (sum(contmask), npixels)) print('Cuts: f_cut %s, sig_cut %s' % (f_cut, sig_cut)) return contmask
find_contpix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, target_frac):\n print('Target frac: %s' % target_frac)\n bad1 = np.median(ivars, axis=0) == SMALL\n bad2 = np.var(ivars, axis=0) ==...
def x(wl, fluxes, ivars, frac, ranges): contmask = np.zeros(len(wl), dtype=bool) for chunk in ranges: start = chunk[0] stop = chunk[1] contmask[start:stop] = _find_contpix(wl[start:stop], fluxes[:, start:stop], ivars[:, start:stop], frac) return contmask
find_contpix_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, frac, ranges):\n contmask = np.zeros(len(wl), dtype=bool)\n for chunk in ranges:\n start = chunk[0]\n stop = chunk[1]\n ...
def x(): tr_obj = np.load('%s/ref_id.npz' % direc_ref)['arr_0'] groups = np.random.randint(0, 8, size=len(tr_obj)) np.savez('ref_groups.npz', groups)
group_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n tr_obj = np.load('%s/ref_id.npz' % direc_ref)['arr_0']\n groups = np.random.randint(0, 8, size=len(tr_obj))\n np.savez('ref_groups.npz', groups)\...
def x(ds, ii): print('Loading model') m = model.CannonModel(2) print('Training...') m.fit(ds) np.savez('./ex%s_coeffs.npz' % ii, m.coeffs) np.savez('./ex%s_scatters.npz' % ii, m.scatters) np.savez('./ex%s_chisqs.npz' % ii, m.chisqs) np.savez('./ex%s_pivots.npz' % ii, m.pivots) fig = m.diagnostics_leading_coeffs(ds) plt.savefig('ex%s_leading_coeffs.png' % ii) return m
train
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ds, ii):\n print('Loading model')\n m = model.CannonModel(2)\n print('Training...')\n m.fit(ds)\n np.savez('./ex%s_coeffs.npz' % ii, m.coeffs)\n...
def x(): print('Loading data') groups = np.load('ref_groups.npz')['arr_0'] ref_label = np.load('%s/ref_label.npz' % direc_ref)['arr_0'] ref_id = np.load('%s/ref_id.npz' % direc_ref)['arr_0'] ref_flux = np.load('%s/ref_flux.npz' % direc_ref)['arr_0'] ref_ivar = np.load('%s/ref_ivar.npz' % direc_ref)['arr_0'] wl = np.load('%s/wl.npz' % direc_ref)['arr_0'] num_models = 8 for ii in np.arange(num_models): print('Leaving out group %s' % ii) train_on = groups != ii test_on = groups == ii tr_label = ref_label[train_on] tr_id = ref_id[train_on] tr_flux = ref_flux[train_on] tr_ivar = ref_ivar[train_on] print('Training on %s objects' % len(tr_id)) test_label = ref_label[test_on] test_id = ref_id[test_on] test_flux = ref_flux[test_on] test_ivar = ref_ivar[test_on] print('Testing on %s objects' % len(test_id)) print('Loading dataset...') ds = dataset.Dataset(wl, tr_id, tr_flux, tr_ivar, tr_label, test_id, test_flux, test_ivar) ds.set_label_names(['T_{eff}', '\\log g', '[M/H]', '[\\alpha/Fe]', 'AKWISE']) fig = ds.diagnostics_SNR() plt.savefig('ex%s_SNR.png' % ii) fig = ds.diagnostics_ref_labels() plt.savefig('ex%s_ref_label_triangle.png' % ii) np.savez('ex%s_tr_snr.npz' % ii, ds.tr_SNR) m = train(ds, ii) ds.tr_label = test_label test(ds, m, ii)
xvalidate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n print('Loading data')\n groups = np.load('ref_groups.npz')['arr_0']\n ref_label = np.load('%s/ref_label.npz' % direc_ref)['arr_0']\n ref_id = ...
def x(values, weights): average = np.average(values, weights=weights) variance = np.average((values - average) ** 2, weights=weights) return np.sqrt(variance)
weighted_std
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(values, weights):\n average = np.average(values, weights=weights)\n variance = np.average((values - average) ** 2, weights=weights)\n return np.sqrt(v...
def x(fluxes, contmask): nstars = fluxes.shape[0] scatter = np.zeros(nstars) for i, spec in enumerate(fluxes): cont = spec[contmask] scatter[i] = stats.funcs.mad_std(cont) return scatter
estimate_noise
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fluxes, contmask):\n nstars = fluxes.shape[0]\n scatter = np.zeros(nstars)\n for i, spec in enumerate(fluxes):\n cont = spec[contmask]\n ...
def x(): data_dir = '/Users/annaho/Data/AAOmega/ref_spectra' ff = glob.glob('%s/*.txt' % data_dir) nstars = len(ff) print('We have %s training objects' % nstars) f = ff[0] data = Table.read(f, format='ascii.fast_no_header') wl = data['col1'] npix = len(wl) print('We have %s pixels' % npix) tr_flux = np.zeros((nstars, npix)) tr_ivar = np.zeros(tr_flux.shape) for i, f in enumerate(ff): data = Table.read(f, format='ascii.fast_no_header') flux = data['col2'] tr_flux[i, :] = flux sigma = data['col3'] tr_ivar[i, :] = 1.0 / sigma ** 2 return np.array(ff), wl, tr_flux, tr_ivar
load_ref_spectra
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n data_dir = '/Users/annaho/Data/AAOmega/ref_spectra'\n ff = glob.glob('%s/*.txt' % data_dir)\n nstars = len(ff)\n print('We have %s training ob...
def x(): data_dir = '/Users/annaho/Data/AAOmega' out_dir = '%s/%s' % (data_dir, 'Run_13_July') """ Use all the above functions to set data up for The Cannon """ ff, wl, tr_flux, tr_ivar = load_ref_spectra() """ pick one that doesn't have extra dead pixels """ skylines = tr_ivar[4, :] np.savez('%s/skylines.npz' % out_dir, skylines) contmask = np.load('%s/contmask_regions.npz' % data_dir)['arr_0'] scatter = estimate_noise(tr_flux, contmask) ids, labels = load_labels() inds = [] ff_short = [] for fname in ff: val = fname.split('/')[-1] short = val.split('.')[0] + '.' + val.split('.')[1] ff_short.append(short) if short in ids: ind = np.where(ids == short)[0][0] inds.append(ind) tr_id = ids[inds] tr_label = labels[inds] ff_short = np.array(ff_short) inds = np.array([np.where(ff_short == val)[0][0] for val in tr_id]) tr_flux_choose = tr_flux[inds] tr_ivar_choose = tr_ivar[inds] scatter_choose = scatter[inds] np.savez('%s/wl.npz' % out_dir, wl) np.savez('%s/ref_id_all.npz' % out_dir, tr_id) np.savez('%s/ref_flux_all.npz' % out_dir, tr_flux_choose) np.savez('%s/ref_ivar_all.npz' % out_dir, tr_ivar_choose) np.savez('%s/ref_label_all.npz' % out_dir, tr_label) np.savez('%s/ref_spec_scat_all.npz' % out_dir, scatter_choose) test_id, test_flux = load_test_spectra() scatter = estimate_noise(test_flux, contmask) np.savez('%s/test_id.npz' % out_dir, test_id) np.savez('%s/test_flux.npz' % out_dir, test_flux) np.savez('%s/test_spec_scat.npz' % out_dir, scatter)
load_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n data_dir = '/Users/annaho/Data/AAOmega'\n out_dir = '%s/%s' % (data_dir, 'Run_13_July')\n \"\"\" Use all the above functions to set data up for T...
def x(): ref_flux = np.load('%s/ref_flux_all.npz' % DATA_DIR)['arr_0'] ref_scat = np.load('%s/ref_spec_scat_all.npz' % DATA_DIR)['arr_0'] test_flux = np.load('%s/test_flux.npz' % DATA_DIR)['arr_0'] test_scat = np.load('%s/test_spec_scat.npz' % DATA_DIR)['arr_0'] ref_ivar = np.ones(ref_flux.shape) / ref_scat[:, None] ** 2 test_ivar = np.ones(test_flux.shape) / test_scat[:, None] ** 2 ref_bad = np.logical_or(ref_flux <= 0, ref_flux > 1.1) test_bad = np.logical_or(test_flux <= 0, test_flux > 1.1) SMALL = 1.0 / 1000000000.0 ref_ivar[ref_bad] = SMALL test_ivar[test_bad] = SMALL np.savez('%s/ref_ivar_corr.npz' % DATA_DIR, ref_ivar) np.savez('%s/test_ivar_corr.npz' % DATA_DIR, test_ivar)
make_full_ivar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n ref_flux = np.load('%s/ref_flux_all.npz' % DATA_DIR)['arr_0']\n ref_scat = np.load('%s/ref_spec_scat_all.npz' % DATA_DIR)['arr_0']\n test_flux = ...
def x(x, p, L, y): N = int(len(p) / 2) n = np.linspace(0, N, N + 1) k = n * np.pi / L func = 0 for n in range(0, N): func += p[2 * n] * np.sin(k[n] * x) + p[2 * n + 1] * np.cos(k[n] * x) return func
sinusoid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, p, L, y):\n N = int(len(p) / 2)\n n = np.linspace(0, N, N + 1)\n k = n * np.pi / L\n func = 0\n for n in range(0, N):\n func += p[2 * ...
def x(values, weights, quantile): sindx = np.argsort(values) cvalues = 1.0 * np.cumsum(weights[sindx]) if cvalues[-1] == 0: return values[0] cvalues = cvalues / cvalues[-1] foo = sindx[cvalues > quantile] if len(foo) == 0: return values[0] indx = foo[0] return values[indx]
weighted_median
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(values, weights, quantile):\n sindx = np.argsort(values)\n cvalues = 1.0 * np.cumsum(weights[sindx])\n if cvalues[-1] == 0:\n return values[0]\...
def x(wl, fluxes, ivars, w): print('Finding the continuum') bot = np.dot(ivars, w.T) top = np.dot(fluxes * ivars, w.T) bad = bot == 0 cont = np.zeros(top.shape) cont[~bad] = top[~bad] / bot[~bad] return cont
find_cont_gaussian_smooth
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, w):\n print('Finding the continuum')\n bot = np.dot(ivars, w.T)\n top = np.dot(fluxes * ivars, w.T)\n bad = bot == 0\n cont =...
def x(dataset, L): print('Gaussian smoothing the entire dataset...') w = gaussian_weight_matrix(dataset.wl, L) print('Gaussian smoothing the training set') cont = _find_cont_gaussian_smooth(dataset.wl, dataset.tr_flux, dataset. tr_ivar, w) norm_tr_flux, norm_tr_ivar = _cont_norm(dataset.tr_flux, dataset. tr_ivar, cont) print('Gaussian smoothing the test set') cont = _find_cont_gaussian_smooth(dataset.wl, dataset.test_flux, dataset.test_ivar, w) norm_test_flux, norm_test_ivar = _cont_norm(dataset.test_flux, dataset. test_ivar, cont) return norm_tr_flux, norm_tr_ivar, norm_test_flux, norm_test_ivar
cont_norm_gaussian_smooth
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dataset, L):\n print('Gaussian smoothing the entire dataset...')\n w = gaussian_weight_matrix(dataset.wl, L)\n print('Gaussian smoothing the training ...
def x(fluxes, ivars, contmask, deg, ffunc, n_proc=1): nstars = fluxes.shape[0] npixels = fluxes.shape[1] cont = np.zeros(fluxes.shape) if n_proc == 1: for jj in range(nstars): flux = fluxes[jj, :] ivar = ivars[jj, :] pix = np.arange(0, npixels) y = flux[contmask] x = pix[contmask] yivar = ivar[contmask] yivar[yivar == 0] = SMALL ** 2 if ffunc == 'sinusoid': p0 = np.ones(deg * 2) L = max(x) - min(x) pcont_func = _partial_func(_sinusoid, L=L, y=flux) popt, pcov = opt.curve_fit(pcont_func, x, y, p0=p0, sigma= 1.0 / np.sqrt(yivar)) elif ffunc == 'chebyshev': fit = np.polynomial.chebyshev.Chebyshev.fit(x=x, y=y, w= yivar, deg=deg) for element in pix: if ffunc == 'sinusoid': cont[jj, element] = _sinusoid(element, popt, L=L, y=flux) elif ffunc == 'chebyshev': cont[jj, element] = fit(element) else: pool = mp.Pool(processes=n_proc) mp_results = [] for i in xrange(nstars): mp_results.append(pool.apply_async(_find_cont_fitfunc, (fluxes[ i, :].reshape((1, -1)), ivars[i, :].reshape((1, -1)), contmask[:]), {'deg': deg, 'ffunc': ffunc})) pool.close() pool.join() cont = np.array([mp_results[i].get().flatten() for i in xrange(nstars)] ) return cont
find_cont_fitfunc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fluxes, ivars, contmask, deg, ffunc, n_proc=1):\n nstars = fluxes.shape[0]\n npixels = fluxes.shape[1]\n cont = np.zeros(fluxes.shape)\n if n_proc ...
def x(fluxes, ivars, contmask, deg, ranges, ffunc, n_proc=1): nstars = fluxes.shape[0] npixels = fluxes.shape[1] cont = np.zeros(fluxes.shape) for chunk in ranges: start = chunk[0] stop = chunk[1] if ffunc == 'chebyshev': output = _find_cont_fitfunc(fluxes[:, start:stop], ivars[:, start:stop], contmask[start:stop], deg=deg, ffunc= 'chebyshev', n_proc=n_proc) elif ffunc == 'sinusoid': output = _find_cont_fitfunc(fluxes[:, start:stop], ivars[:, start:stop], contmask[start:stop], deg=deg, ffunc= 'sinusoid', n_proc=n_proc) cont[:, start:stop] = output return cont
find_cont_fitfunc_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fluxes, ivars, contmask, deg, ranges, ffunc, n_proc=1):\n nstars = fluxes.shape[0]\n npixels = fluxes.shape[1]\n cont = np.zeros(fluxes.shape)\n fo...
def x(wl, fluxes, ivars, q, delta_lambda, verbose=False): cont = np.zeros(fluxes.shape) nstars = fluxes.shape[0] for jj in range(nstars): if verbose: print('cont_norm_q(): working on star [%s/%s]...' % (jj + 1, nstars)) flux = fluxes[jj, :] ivar = ivars[jj, :] for ll, lam in enumerate(wl): indx = np.where(abs(wl - lam) < delta_lambda)[0] flux_cut = flux[indx] ivar_cut = ivar[indx] cont[jj, ll] = _weighted_median(flux_cut, ivar_cut, q) return cont
find_cont_running_quantile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, q, delta_lambda, verbose=False):\n cont = np.zeros(fluxes.shape)\n nstars = fluxes.shape[0]\n for jj in range(nstars):\n if ...
def x(wl, fluxes, ivars, q, delta_lambda, n_proc=2, verbose=False): nStar = fluxes.shape[0] mp_results = [] pool = mp.Pool(processes=n_proc) for i in xrange(nStar): mp_results.append(pool.apply_async(_find_cont_running_quantile, (wl, fluxes[i, :].reshape((1, -1)), ivars[i, :].reshape((1, -1)), q, delta_lambda), {'verbose': False})) if verbose: print('@Bo Zhang: continuum normalizing star [%d/%d] ...' % (i + 1, nStar)) pool.close() pool.join() cont = np.zeros_like(fluxes) for i in xrange(nStar): cont[i, :] = mp_results[i].get() norm_fluxes = np.ones(fluxes.shape) norm_fluxes[cont != 0] = fluxes[cont != 0] / cont[cont != 0] norm_ivars = cont ** 2 * ivars print('@Bo Zhang: continuum normalization finished!') return norm_fluxes, norm_ivars
cont_norm_running_quantile_mp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, q, delta_lambda, n_proc=2, verbose=False):\n nStar = fluxes.shape[0]\n mp_results = []\n pool = mp.Pool(processes=n_proc)\n for ...
def x(wl, fluxes, ivars, q, delta_lambda, ranges, verbose=True): print('contnorm.py: continuum norm using running quantile') print('Taking spectra in %s chunks' % len(ranges)) nstars = fluxes.shape[0] norm_fluxes = np.zeros(fluxes.shape) norm_ivars = np.zeros(ivars.shape) for chunk in ranges: start = chunk[0] stop = chunk[1] output = _cont_norm_running_quantile(wl[start:stop], fluxes[:, start:stop], ivars[:, start:stop], q, delta_lambda) norm_fluxes[:, start:stop] = output[0] norm_ivars[:, start:stop] = output[1] return norm_fluxes, norm_ivars
cont_norm_running_quantile_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, q, delta_lambda, ranges, verbose=True):\n print('contnorm.py: continuum norm using running quantile')\n print('Taking spectra in %s ch...
def x(wl, fluxes, ivars, q, delta_lambda, ranges, n_proc=2, verbose=False): print('contnorm.py: continuum norm using running quantile') print('Taking spectra in %s chunks' % len(ranges)) nchunks = len(ranges) norm_fluxes = np.zeros(fluxes.shape) norm_ivars = np.zeros(ivars.shape) for i in xrange(nchunks): chunk = ranges[i, :] start = chunk[0] stop = chunk[1] if verbose: print( '@Bo Zhang: Going to normalize Chunk [%d/%d], pixel:[%d, %d] ...' % (i + 1, nchunks, start, stop)) output = _cont_norm_running_quantile_mp(wl[start:stop], fluxes[:, start:stop], ivars[:, start:stop], q, delta_lambda, n_proc= n_proc, verbose=verbose) norm_fluxes[:, start:stop] = output[0] norm_ivars[:, start:stop] = output[1] return norm_fluxes, norm_ivars
cont_norm_running_quantile_regions_mp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(wl, fluxes, ivars, q, delta_lambda, ranges, n_proc=2, verbose=False):\n print('contnorm.py: continuum norm using running quantile')\n print('Taking spect...
def x(fluxes, ivars, cont): nstars = fluxes.shape[0] npixels = fluxes.shape[1] norm_fluxes = np.ones(fluxes.shape) norm_ivars = np.zeros(ivars.shape) bad = cont == 0.0 norm_fluxes = np.ones(fluxes.shape) norm_fluxes[~bad] = fluxes[~bad] / cont[~bad] norm_ivars = cont ** 2 * ivars return norm_fluxes, norm_ivars
cont_norm
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fluxes, ivars, cont):\n nstars = fluxes.shape[0]\n npixels = fluxes.shape[1]\n norm_fluxes = np.ones(fluxes.shape)\n norm_ivars = np.zeros(ivars.sh...
def x(fluxes, ivars, cont, ranges): nstars = fluxes.shape[0] norm_fluxes = np.zeros(fluxes.shape) norm_ivars = np.zeros(ivars.shape) for chunk in ranges: start = chunk[0] stop = chunk[1] output = _cont_norm(fluxes[:, start:stop], ivars[:, start:stop], cont[:, start:stop]) norm_fluxes[:, start:stop] = output[0] norm_ivars[:, start:stop] = output[1] for jj in range(nstars): bad = norm_ivars[jj, :] == 0.0 norm_fluxes[jj, :][bad] = 1.0 return norm_fluxes, norm_ivars
cont_norm_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fluxes, ivars, cont, ranges):\n nstars = fluxes.shape[0]\n norm_fluxes = np.zeros(fluxes.shape)\n norm_ivars = np.zeros(ivars.shape)\n for chunk in...
def x(self, ds): if self.useErrors: (self.coeffs, self.scatters, self.new_tr_labels, self.chisqs, self. pivots, self.scales) = _train_model_new(ds) else: (self.coeffs, self.scatters, self.chisqs, self.pivots, self.scales ) = _train_model(ds)
train
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ds):\n if self.useErrors:\n (self.coeffs, self.scatters, self.new_tr_labels, self.chisqs, self.\n pivots, self.scales) = _train_mode...
def x(self, ds): lvec_all = _get_lvec(ds.test_label_vals, self.pivots, self.scales, derivs=False) self.model_spectra = np.dot(lvec_all, self.coeffs.T)
infer_spectra
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ds):\n lvec_all = _get_lvec(ds.test_label_vals, self.pivots, self.scales,\n derivs=False)\n self.model_spectra = np.dot(lvec_all, self.coeff...
def x(self, x, y, contpix_x, contpix_y, figname): fig, axarr = plt.subplots(2, sharex=True) plt.xlabel('Wavelength $\\lambda (\\AA)$') plt.xlim(min(x), max(x)) ax = axarr[0] ax.step(x, y, where='mid', c='k', linewidth=0.3, label='$\\theta_0$' + '= the leading fit coefficient') ax.scatter(contpix_x, contpix_y, s=1, color='r', label='continuum pixels') ax.legend(loc='lower right', prop={'family': 'serif', 'size': 'small'}) ax.set_title('Baseline Spectrum with Continuum Pixels') ax.set_ylabel('$\\theta_0$') ax = axarr[1] ax.step(x, y, where='mid', c='k', linewidth=0.3, label='$\\theta_0$' + '= the leading fit coefficient') ax.scatter(contpix_x, contpix_y, s=1, color='r', label='continuum pixels') ax.set_title('Baseline Spectrum with Continuum Pixels, Zoomed') ax.legend(loc='upper right', prop={'family': 'serif', 'size': 'small'}) ax.set_ylabel('$\\theta_0$') ax.set_ylim(0.95, 1.05) print('Diagnostic plot: fitted 0th order spec w/ cont pix') print('Saved as %s.png' % figname) plt.savefig(figname) plt.close()
plot_contpix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x, y, contpix_x, contpix_y, figname):\n fig, axarr = plt.subplots(2, sharex=True)\n plt.xlabel('Wavelength $\\\\lambda (\\\\AA)$')\n plt.xlim(mi...
def x(self, data, nchunks=10, fig='baseline_spec_with_cont_pix'): if data.contmask is None: print('No contmask set') else: coeffs_all = self.coeffs wl = data.wl baseline_spec = coeffs_all[:, 0] contmask = data.contmask contpix_x = wl[contmask] contpix_y = baseline_spec[contmask] rem = len(wl) % nchunks wl_split = np.array(np.split(wl[0:len(wl) - rem], nchunks)) baseline_spec_split = np.array(np.split(baseline_spec[0:len(wl) - rem], nchunks)) nchunks = wl_split.shape[0] for i in range(nchunks): fig_chunk = fig + '_%s' % str(i) wl_chunk = wl_split[i, :] baseline_spec_chunk = baseline_spec_split[i, :] take = np.logical_and(contpix_x > wl_chunk[0], contpix_x < wl_chunk[-1]) self.plot_contpix(wl_chunk, baseline_spec_chunk, contpix_x[take ], contpix_y[take], fig_chunk)
diagnostics_contpix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, nchunks=10, fig='baseline_spec_with_cont_pix'):\n if data.contmask is None:\n print('No contmask set')\n else:\n coeffs_all = s...
def x(self, ds, figname='modelfit_chisqs.png'): label_names = ds.get_plotting_labels() lams = ds.wl pivots = self.pivots npixels = len(lams) nlabels = len(pivots) chisqs = self.chisqs coeffs = self.coeffs scatters = self.scatters plt.hist(np.sum(chisqs, axis=0), color='lightblue', alpha=0.7, bins=int (np.sqrt(len(chisqs)))) dof = len(lams) - coeffs.shape[1] plt.axvline(x=dof, c='k', linewidth=2, label='DOF') plt.legend() plt.title('Distribution of ' + '$\\chi^2$' + ' of the Model Fit') plt.ylabel('Count') plt.xlabel('$\\chi^2$' + ' of Individual Star') print('Diagnostic plot: histogram of the red chi squareds of the fit') print('Saved as %s' % figname) plt.savefig(figname) plt.close()
diagnostics_plot_chisq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ds, figname='modelfit_chisqs.png'):\n label_names = ds.get_plotting_labels()\n lams = ds.wl\n pivots = self.pivots\n npixels = len(lams)\n ...
def x(nu_max, delta_nu, teff): NU_MAX = 3140.0 DELTA_NU = 135.03 TEFF = 5777.0 return (nu_max / NU_MAX) ** 3 * (delta_nu / DELTA_NU) ** -4 * (teff / TEFF ) ** 1.5
calc_mass
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(nu_max, delta_nu, teff):\n NU_MAX = 3140.0\n DELTA_NU = 135.03\n TEFF = 5777.0\n return (nu_max / NU_MAX) ** 3 * (delta_nu / DELTA_NU) ** -4 * (tef...
def x(mh, cm, nm, teff, logg): CplusN = calc_sum(mh, cm, nm) t = teff / 4000.0 return (95.8689 - 10.4042 * mh - 0.7266 * mh ** 2 + 41.3642 * cm - 5.3242 * cm * mh - 46.7792 * cm ** 2 + 15.0508 * nm - 0.9342 * nm * mh - 30.5159 * nm * cm - 1.6083 * nm ** 2 - 67.6093 * CplusN + 7.0486 * CplusN * mh + 133.5775 * CplusN * cm + 38.9439 * CplusN * nm - 88.9948 * CplusN ** 2 - 144.1765 * t + 5.118 * t * mh - 73.769 * t * cm - 15.2927 * t * nm + 101.7482 * t * CplusN + 27.769 * t ** 2 - 9.4246 * logg + 1.5159 * logg * mh + 16.0412 * logg * cm + 1.3549 * logg * nm - 18.6527 * logg * CplusN + 28.8015 * logg * t - 4.0982 * logg ** 2)
calc_mass_2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mh, cm, nm, teff, logg):\n CplusN = calc_sum(mh, cm, nm)\n t = teff / 4000.0\n return (95.8689 - 10.4042 * mh - 0.7266 * mh ** 2 + 41.3642 * cm - \n ...
def x(x, q, weights=None): if weights is None: return np.percentile(x, [(100.0 * qi) for qi in q]) else: idx = np.argsort(x) xsorted = x[idx] cdf = np.add.accumulate(weights[idx]) cdf /= cdf[-1] return np.interp(q, cdf, xsorted).tolist()
quantile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, q, weights=None):\n if weights is None:\n return np.percentile(x, [(100.0 * qi) for qi in q])\n else:\n idx = np.argsort(x)\n xso...
def x(x, y, bins=20, range=None, weights=None, levels=None, smooth=None, ax =None, color=None, plot_datapoints=True, plot_density=True, plot_contours=True, no_fill_contours=False, fill_contours=False, contour_kwargs=None, contourf_kwargs=None, data_kwargs=None, **kwargs): if ax is None: ax = pl.gca() if range is None: if 'extent' in kwargs: logging.warn( "Deprecated keyword argument 'extent'. Use 'range' instead.") range = kwargs['extent'] else: range = [[x.min(), x.max()], [y.min(), y.max()]] if color is None: color = 'k' if levels is None: levels = 1.0 - np.exp(-0.5 * np.arange(0.5, 2.1, 0.5) ** 2) density_cmap = LinearSegmentedColormap.from_list('density_cmap', [color, (1, 1, 1, 0)]) white_cmap = LinearSegmentedColormap.from_list('white_cmap', [(1, 1, 1), (1, 1, 1)], N=2) rgba_color = colorConverter.to_rgba(color) contour_cmap = [list(rgba_color) for l in levels] + [rgba_color] for i, l in enumerate(levels): contour_cmap[i][-1] *= float(i) / (len(levels) + 1) try: H, X, Y = np.histogram2d(x.flatten(), y.flatten(), bins=bins, range =range, weights=weights) except ValueError: raise ValueError( "It looks like at least one of your sample columns have no dynamic range. You could try using the 'range' argument." ) if smooth is not None: if gaussian_filter is None: raise ImportError('Please install scipy for smoothing') H = gaussian_filter(H, smooth) Hflat = H.flatten() inds = np.argsort(Hflat)[::-1] Hflat = Hflat[inds] sm = np.cumsum(Hflat) sm /= sm[-1] V = np.empty(len(levels)) for i, v0 in enumerate(levels): try: V[i] = Hflat[sm <= v0][-1] except: V[i] = Hflat[0] V.sort() m = np.diff(V) == 0 if np.any(m): logging.warning('Too few points to create valid contours') while np.any(m): V[np.where(m)[0][0]] *= 1.0 - 0.0001 m = np.diff(V) == 0 V.sort() X1, Y1 = 0.5 * (X[1:] + X[:-1]), 0.5 * (Y[1:] + Y[:-1]) H2 = H.min() + np.zeros((H.shape[0] + 4, H.shape[1] + 4)) H2[2:-2, 2:-2] = H H2[2:-2, 1] = H[:, 0] H2[2:-2, -2] = H[:, -1] H2[1, 2:-2] = H[0] H2[-2, 2:-2] = H[-1] H2[1, 1] = H[0, 0] H2[1, -2] = H[0, -1] H2[-2, 1] = H[-1, 0] H2[-2, -2] = H[-1, -1] X2 = np.concatenate([X1[0] + np.array([-2, -1]) * np.diff(X1[:2]), X1, X1[-1] + np.array([1, 2]) * np.diff(X1[-2:])]) Y2 = np.concatenate([Y1[0] + np.array([-2, -1]) * np.diff(Y1[:2]), Y1, Y1[-1] + np.array([1, 2]) * np.diff(Y1[-2:])]) if plot_datapoints: if data_kwargs is None: data_kwargs = dict() data_kwargs['color'] = data_kwargs.get('color', color) data_kwargs['ms'] = data_kwargs.get('ms', 2.0) data_kwargs['mec'] = data_kwargs.get('mec', 'none') data_kwargs['alpha'] = data_kwargs.get('alpha', 0.1) ax.plot(x, y, 'o', zorder=-1, rasterized=True, **data_kwargs) if (plot_contours or plot_density) and not no_fill_contours: ax.contourf(X2, Y2, H2.T, [V.min(), H.max()], cmap=white_cmap, antialiased=False) if plot_contours and fill_contours: if contourf_kwargs is None: contourf_kwargs = dict() contourf_kwargs['colors'] = contourf_kwargs.get('colors', contour_cmap) contourf_kwargs['antialiased'] = contourf_kwargs.get('antialiased', False) ax.contourf(X2, Y2, H2.T, np.concatenate([[0], V, [H.max() * (1 + 0.0001)]]), **contourf_kwargs) elif plot_density: ax.pcolor(X, Y, H.max() - H.T, cmap=density_cmap) if plot_contours: if contour_kwargs is None: contour_kwargs = dict() contour_kwargs['colors'] = contour_kwargs.get('colors', color) ax.contour(X2, Y2, H2.T, V, **contour_kwargs) ax.set_xlim(range[0]) ax.set_ylim(range[1])
hist2d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, y, bins=20, range=None, weights=None, levels=None, smooth=None, ax\n =None, color=None, plot_datapoints=True, plot_density=True,\n plot_contours=True,...
def x(lamost_point, training_points, coeffs): diff2 = (training_points - lamost_point) ** 2 dist = np.sqrt(np.sum(diff2 * coeffs, axis=1)) return np.mean(dist[dist.argsort()][0:10])
calc_dist
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(lamost_point, training_points, coeffs):\n diff2 = (training_points - lamost_point) ** 2\n dist = np.sqrt(np.sum(diff2 * coeffs, axis=1))\n return np.m...
def x(self, name, ids, labels): if not all(np.in1d(ids, self.ids)): raise ValueError( 'Hay ids de textos que no se encuentran almacenados.' ) setattr(self, name, SGDClassifier()) classifier = getattr(self, name) indices = np.searchsorted(self.ids, ids) classifier.fit(self.tfidf_mat[indices, :], labels)
make_classifier
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, ids, labels):\n if not all(np.in1d(ids, self.ids)):\n raise ValueError(\n 'Hay ids de textos que no se encuentran ...
def x(self, name, ids, labels): if not all(np.in1d(ids, self.ids)): raise ValueError( 'Hay ids de textos que no se encuentran almacenados.' ) try: classifier = getattr(self, name) except AttributeError: raise AttributeError('No hay ningun clasificador con ese nombre.') indices = np.in1d(self.ids, ids) if isinstance(labels, str): labels = [labels] classifier.partial_fit(self.tfidf_mat[indices, :], labels)
retrain
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, ids, labels):\n if not all(np.in1d(ids, self.ids)):\n raise ValueError(\n 'Hay ids de textos que no se encuentran ...
def x(self, classifier_name, examples, max_labels=None, goodness_of_fit=False): classifier = getattr(self, classifier_name) texts_vectors = self._make_text_vectors(examples) return classifier.classes_, classifier.decision_function(texts_vectors)
classify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, classifier_name, examples, max_labels=None, goodness_of_fit=False):\n classifier = getattr(self, classifier_name)\n texts_vectors = self._make_text...
def x(self, examples): if isinstance(examples, str): if examples in self.ids: textvec = self.tfidf_mat[self.ids == examples, :] else: textvec = self.vectorizer.transform([examples]) textvec = self.transformer.transform(textvec) elif type(examples) is list: if all(np.in1d(examples, self.ids)): textvec = self.tfidf_mat[np.in1d(self.ids, examples)] elif not any(np.in1d(examples, self.ids)): textvec = self.vectorizer.transform(examples) textvec = self.transformer.transform(textvec) else: raise ValueError( 'Las listas de ejemplos deben ser todos ids de textos almacenados o todos textos planos' ) else: raise TypeError('Los ejemplos no son del tipo de dato adecuado.') return textvec
make_text_vectors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, examples):\n if isinstance(examples, str):\n if examples in self.ids:\n textvec = self.tfidf_mat[self.ids == examples, :]\n e...
def x(self, example, max_similars=3, similarity_cutoff=None, term_diff_max_rank=10, filter_list=None, term_diff_cutoff=None): if term_diff_cutoff: warnings.warn('Deprecado. Quedo sin uso. Se quitara en el futuro.', DeprecationWarning) if filter_list: if max_similars > len(filter_list): raise ValueError( 'No se pueden pedir mas sugerencias que la cantidad de textos en `filter_list`.' ) else: filt_idx = np.in1d(self.ids, filter_list) elif max_similars > self.term_mat.shape[0]: raise ValueError( 'No se pueden pedir mas sugerencias que la cantidad de textos que hay almacenados.' ) else: filt_idx = np.ones(len(self.ids), dtype=bool) good_ids = np.array(np.sum(self.term_mat, 1) > 0).squeeze() filt_idx = filt_idx & good_ids filt_idx_to_general_idx = np.flatnonzero(filt_idx) if example in self.ids: index = self.ids == example exmpl_vec = self.tfidf_mat[index, :] distances = np.squeeze(pairwise_distances(self.tfidf_mat[filt_idx], exmpl_vec)) if filter_list and example in filter_list: distances[filter_list.index(example)] = np.inf elif not filter_list: idx_example = np.searchsorted(self.ids, example) filt_idx_example = np.searchsorted(np.flatnonzero(filt_idx), idx_example) distances[filt_idx_example] = np.inf else: exmpl_vec = self.vectorizer.transform([example]) exmpl_vec = self.transformer.transform(exmpl_vec) distances = np.squeeze(pairwise_distances(self.tfidf_mat[filt_idx], exmpl_vec)) if np.sum(exmpl_vec) == 0: return [], [], [] sorted_indices = np.argsort(distances) closest_n = sorted_indices[:max_similars] sorted_dist = distances[closest_n] if similarity_cutoff: closest_n = closest_n[sorted_dist < similarity_cutoff] sorted_dist = sorted_dist[sorted_dist < similarity_cutoff] best_words = [] best_example = np.squeeze(exmpl_vec.toarray()) sorted_example_weights = np.flipud(np.argsort(best_example)) truncated_max_rank = min(term_diff_max_rank, np.sum(best_example > 0)) best_example_words = sorted_example_weights[:truncated_max_rank] for suggested in closest_n: suggested_idx = filt_idx_to_general_idx[suggested] test_vec = np.squeeze(self.tfidf_mat[suggested_idx, :].toarray()) sorted_test_weights = np.flipud(np.argsort(test_vec)) truncated_max_rank = min(term_diff_max_rank, np.sum(test_vec > 0)) best_test = sorted_test_weights[:truncated_max_rank] best_words_ids = np.intersect1d(best_example_words, best_test) best_words.append([k for k, v in self.vectorizer.vocabulary_.items( ) if v in best_words_ids]) if filter_list: text_ids = self.ids[filt_idx_to_general_idx[closest_n]] else: text_ids = self.ids[closest_n] return list(text_ids), list(sorted_dist), best_words
get_similar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, example, max_similars=3, similarity_cutoff=None,\n term_diff_max_rank=10, filter_list=None, term_diff_cutoff=None):\n if term_diff_cutoff:\n ...
def x(self, texts, ids, vocabulary=None): self._check_id_length(ids) self.ids = np.array(sorted(ids)) if vocabulary: self.vectorizer.vocabulary = vocabulary sorted_texts = [x for y, x in sorted(zip(ids, texts))] self.term_mat = self.vectorizer.fit_transform(sorted_texts) self._update_tfidf()
reload_texts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, texts, ids, vocabulary=None):\n self._check_id_length(ids)\n self.ids = np.array(sorted(ids))\n if vocabulary:\n self.vectorizer.vocabula...
def x(q=None, datasetKey=None, rank=None, limit=100, offset=None, **kwargs): url = gbif_baseurl + 'species/suggest' args = {'q': q, 'rank': rank, 'offset': offset, 'limit': limit} return gbif_GET(url, args, **kwargs)
name_suggest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(q=None, datasetKey=None, rank=None, limit=100, offset=None, **kwargs):\n url = gbif_baseurl + 'species/suggest'\n args = {'q': q, 'rank': rank, 'offset':...
def x(uuid, **kwargs): def x(x, **kwargs): url = gbif_baseurl + 'dataset/' + x + '/metrics' return gbif_GET(url, {}, **kwargs) if len2(uuid) == 1: return getdata(uuid) else: return [getdata(x) for x in uuid]
dataset_metrics
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(uuid, **kwargs):\n\n def x(x, **kwargs):\n url = gbif_baseurl + 'dataset/' + x + '/metrics'\n return gbif_GET(url, {}, **kwargs)\n if len2(...
def x(data='all', type=None, uuid=None, query=None, id=None, limit=100, offset=None, **kwargs): args = {'q': query, 'type': type, 'limit': limit, 'offset': offset} data_choices = ['all', 'organization', 'contact', 'endpoint', 'identifier', 'tag', 'machinetag', 'comment', 'constituents', 'document', 'metadata', 'deleted', 'duplicate', 'subDataset', 'withNoEndpoint'] check_data(data, data_choices) if len2(data) == 1: return datasets_fetch(data, uuid, args, **kwargs) else: return [datasets_fetch(x, uuid, args, **kwargs) for x in data]
datasets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data='all', type=None, uuid=None, query=None, id=None, limit=100,\n offset=None, **kwargs):\n args = {'q': query, 'type': type, 'limit': limit, 'offset':...
def x(q=None, type=None, keyword=None, owningOrg=None, publishingOrg=None, hostingOrg=None, publishingCountry=None, decade=None, limit=100, offset =None, **kwargs): url = gbif_baseurl + 'dataset/suggest' args = {'q': q, 'type': type, 'keyword': keyword, 'publishingOrg': publishingOrg, 'hostingOrg': hostingOrg, 'owningOrg': owningOrg, 'decade': decade, 'publishingCountry': publishingCountry, 'limit': limit, 'offset': offset} out = gbif_GET(url, args, **kwargs) return out
dataset_suggest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(q=None, type=None, keyword=None, owningOrg=None, publishingOrg=None,\n hostingOrg=None, publishingCountry=None, decade=None, limit=100, offset\n =None, *...
def x(q=None, type=None, keyword=None, owningOrg=None, publishingOrg=None, hostingOrg=None, decade=None, publishingCountry=None, facet=None, facetMincount=None, facetMultiselect=None, hl=False, limit=100, offset= None, **kwargs): url = gbif_baseurl + 'dataset/search' args = {'q': q, 'type': type, 'keyword': keyword, 'owningOrg': owningOrg, 'publishingOrg': publishingOrg, 'hostingOrg': hostingOrg, 'decade': decade, 'publishingCountry': publishingCountry, 'facet': facet, 'facetMincount': facetMincount, 'facetMultiselect': facetMultiselect, 'hl': hl, 'limit': limit, 'offset': offset} gbif_kwargs = {key: kwargs[key] for key in kwargs if key not in requests_argset} if gbif_kwargs is not None: xx = dict(zip([re.sub('_', '.', x) for x in gbif_kwargs.keys()], gbif_kwargs.values())) args.update(xx) kwargs = {key: kwargs[key] for key in kwargs if key in requests_argset} out = gbif_GET(url, args, **kwargs) return out
dataset_search
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(q=None, type=None, keyword=None, owningOrg=None, publishingOrg=None,\n hostingOrg=None, decade=None, publishingCountry=None, facet=None,\n facetMincount=...
def x(x, digits=None): z = wkt.loads(x) if digits is None: coords = z['coordinates'] nums = __flatten(coords) dec_n = [decimal.Decimal(str(w)).as_tuple().exponent for w in nums] digits = abs(statistics.mean(dec_n)) elif not isinstance(digits, int): raise TypeError("'digits' must be an int") wound = rewind(z) back_to_wkt = wkt.dumps(wound, decimals=digits) return back_to_wkt
wkt_rewind
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, digits=None):\n z = wkt.loads(x)\n if digits is None:\n coords = z['coordinates']\n nums = __flatten(coords)\n dec_n = [decimal.D...
def x(issue=None, code=None): if code is None: bb = [trymatch(issue, x) for x in gbifissues['issue']] tmp = filter(None, bb) else: bb = [trymatch(code, x) for x in gbifissues['code']] tmp = filter(None, bb) return tmp
occ_issues_lookup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(issue=None, code=None):\n if code is None:\n bb = [trymatch(issue, x) for x in gbifissues['issue']]\n tmp = filter(None, bb)\n else:\n ...
def x(taxonKey=None, repatriated=None, kingdomKey=None, phylumKey=None, classKey=None, orderKey=None, familyKey=None, genusKey=None, subgenusKey=None, scientificName=None, country=None, publishingCountry= None, hasCoordinate=None, typeStatus=None, recordNumber=None, lastInterpreted=None, continent=None, geometry=None, recordedBy=None, basisOfRecord=None, datasetKey=None, eventDate=None, catalogNumber=None, year=None, month=None, decimalLatitude=None, decimalLongitude=None, elevation=None, depth=None, institutionCode=None, collectionCode=None, hasGeospatialIssue=None, issue=None, q=None, spellCheck=None, mediatype =None, limit=300, offset=0, establishmentMeans=None, facet=None, facetMincount=None, facetMultiselect=None, **kwargs): url = gbif_baseurl + 'occurrence/search' args = {'taxonKey': taxonKey, 'repatriated': repatriated, 'kingdomKey': kingdomKey, 'phylumKey': phylumKey, 'classKey': classKey, 'orderKey': orderKey, 'familyKey': familyKey, 'genusKey': genusKey, 'subgenusKey': subgenusKey, 'scientificName': scientificName, 'country': country, 'publishingCountry': publishingCountry, 'hasCoordinate': hasCoordinate, 'typeStatus': typeStatus, 'recordNumber': recordNumber, 'lastInterpreted': lastInterpreted, 'continent': continent, 'geometry': geometry, 'recordedBy': recordedBy, 'basisOfRecord': basisOfRecord, 'datasetKey': datasetKey, 'eventDate': eventDate, 'catalogNumber': catalogNumber, 'year': year, 'month': month, 'decimalLatitude': decimalLatitude, 'decimalLongitude': decimalLongitude, 'elevation': elevation, 'depth': depth, 'institutionCode': institutionCode, 'collectionCode': collectionCode, 'hasGeospatialIssue': hasGeospatialIssue, 'issue': issue, 'q': q, 'spellCheck': spellCheck, 'mediatype': mediatype, 'limit': limit, 'offset': offset, 'establishmentMeans': establishmentMeans, 'facetMincount': facetMincount, 'facet': facet, 'facetMultiselect': facetMultiselect} gbif_kwargs = {key: kwargs[key] for key in kwargs if key not in requests_argset} if gbif_kwargs is not None: xx = dict(zip([re.sub('_', '.', x) for x in gbif_kwargs.keys()], gbif_kwargs.values())) args.update(xx) kwargs = {key: kwargs[key] for key in kwargs if key in requests_argset} out = gbif_GET(url, args, **kwargs) return out
search
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(taxonKey=None, repatriated=None, kingdomKey=None, phylumKey=None,\n classKey=None, orderKey=None, familyKey=None, genusKey=None,\n subgenusKey=None, scie...
def x(data='all', uuid=None, q=None, identifier=None, identifierType=None, limit=100, offset=None, **kwargs): args = {'q': q, 'limit': limit, 'offset': offset, 'identifier': identifier, 'identifierType': identifierType} data_choices = ['all', 'contact', 'endpoint', 'identifier', 'tag', 'machineTag', 'comment', 'constituents'] check_data(data, data_choices) def x(x, uuid, args, **kwargs): if x is not 'all' and uuid is None: stop('You must specify a uuid if data does not equal "all"') if uuid is None: url = gbif_baseurl + 'network' elif x is 'all': url = gbif_baseurl + 'network/' + uuid else: url = gbif_baseurl + 'network/' + uuid + '/' + x res = gbif_GET(url, args, **kwargs) return {'meta': get_meta(res), 'data': parse_results(res, uuid)} if len2(data) == 1: return getdata(data, uuid, args, **kwargs) else: return [getdata(x, uuid, args, **kwargs) for x in data]
networks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data='all', uuid=None, q=None, identifier=None, identifierType=None,\n limit=100, offset=None, **kwargs):\n args = {'q': q, 'limit': limit, 'offset': off...
def x(source='density', z=0, x=0, y=0, format='@1x.png', srs='EPSG:4326', bin=None, hexPerTile=None, style='classic.point', taxonKey=None, country=None, publishingCountry=None, publisher=None, datasetKey=None, year=None, basisOfRecord=None, **kwargs): if format not in ['.mvt', '@Hx.png', '@1x.png', '@2x.png', '@3x.png', '@4x.png']: raise ValueError("'format' not in allowed set, see docs") if source not in ['density', 'adhoc']: raise ValueError("'source' not in allowed set, see docs") if srs not in ['EPSG:3857', 'EPSG:4326', 'EPSG:3575', 'EPSG:3031']: raise ValueError("'srs' not in allowed set, see docs") if bin is not None: if bin not in ['square', 'hex']: raise ValueError("'bin' not in allowed set, see docs") if style is not None: if style not in map_styles: raise ValueError("'style' not in allowed set, see docs") maps_baseurl = 'https://api.gbif.org' url = maps_baseurl + '/v2/map/occurrence/%s/%s/%s/%s%s' url = url % (source, z, x, y, format) year = __handle_year(year) basisOfRecord = __handle_bor(basisOfRecord) args = {'srs': srs, 'bin': bin, 'hexPerTile': hexPerTile, 'style': style, 'taxonKey': taxonKey, 'country': country, 'publishingCountry': publishingCountry, 'publisher': publisher, 'datasetKey': datasetKey, 'year': year, 'basisOfRecord': basisOfRecord} kw = {key: kwargs[key] for key in kwargs if key not in requests_argset} if kw is not None: xx = dict(zip([re.sub('_', '.', x) for x in kw.keys()], kw.values())) args.update(xx) kwargs = {key: kwargs[key] for key in kwargs if key in requests_argset} ctype = 'image/png' if has(format, 'png') else 'application/x-protobuf' out = gbif_GET_map(url, args, ctype, **kwargs) return GbifMap(out)
map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source='density', z=0, x=0, y=0, format='@1x.png', srs='EPSG:4326',\n bin=None, hexPerTile=None, style='classic.point', taxonKey=None,\n country=None, pu...
def x(key=None, name=None, data='all', language=None, datasetKey=None, uuid =None, sourceId=None, rank=None, shortname=None, limit=100, offset=None, **kwargs): args = {'language': language, 'name': name, 'datasetKey': datasetKey, 'rank': rank, 'sourceId': sourceId, 'limit': limit, 'offset': offset} data_choices = ['all', 'verbatim', 'name', 'parents', 'children', 'related', 'synonyms', 'descriptions', 'distributions', 'media', 'references', 'speciesProfiles', 'vernacularNames', 'typeSpecimens', 'root'] check_data(data, data_choices) if len2(data) == 1: return name_usage_fetch(data, key, shortname, uuid, args, **kwargs) else: return [name_usage_fetch(x, key, shortname, uuid, args, **kwargs) for x in data]
name_usage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key=None, name=None, data='all', language=None, datasetKey=None, uuid\n =None, sourceId=None, rank=None, shortname=None, limit=100, offset=None,\n **kwar...
def x(variable, value): if is_not_none(value): return value else: value = os.environ.get(variable) if is_none(value): stop(''.join([variable, """ not supplied and no entry in environmental variables""" ])) else: return value
check_environ
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(variable, value):\n if is_not_none(value):\n return value\n else:\n value = os.environ.get(variable)\n if is_none(value):\n ...