source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self):
data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup)
for key, val in data.items():
if key == 'location_area_encounters' and self.endpoint == 'pokemon':
params = val.split('/')[-3:]
ep, id_, subr = params
encounters = get_data(ep, int(id_), subr)
data[key] = [_make_obj(enc) for enc in encounters]
continue
if isinstance(val, dict):
data[key] = _make_obj(val)
elif isinstance(val, list):
data[key] = [_make_obj(i) for i in val]
self.__dict__.update(data)
return None
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup)\n for key, val in data.items():\n if key == 'location_area_enc... |
def x(path, mode=511):
try:
os.makedirs(path, mode)
except OSError as error:
if error.errno != 17:
raise
return path
| safe_make_dirs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, mode=511):\n try:\n os.makedirs(path, mode)\n except OSError as error:\n if error.errno != 17:\n raise\n return path\n"... |
def x():
xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os.
path.expanduser('~'), '.cache')
return os.path.join(xdg_cache_home, 'pokebase')
| get_default_cache | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os.\n path.expanduser('~'), '.cache')\n return os.path.join(xdg_cache_home,... |
def x(new_path=None):
global CACHE_DIR, API_CACHE, SPRITE_CACHE
if new_path is None:
new_path = get_default_cache()
CACHE_DIR = safe_make_dirs(os.path.abspath(new_path))
API_CACHE = os.path.join(CACHE_DIR, 'api.cache')
SPRITE_CACHE = safe_make_dirs(os.path.join(CACHE_DIR, 'sprite'))
return CACHE_DIR, API_CACHE, SPRITE_CACHE
| set_cache | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(new_path=None):\n global CACHE_DIR, API_CACHE, SPRITE_CACHE\n if new_path is None:\n new_path = get_default_cache()\n CACHE_DIR = safe_make_dir... |
def x(self, lun_or_snap, skip_hlu_0=False):
try:
return self._attach_with_retry(lun_or_snap, skip_hlu_0)
except ex.SystemAPINotSupported:
raise
except ex.UnityAttachExceedLimitError:
raise
except:
self.detach(lun_or_snap)
raise
| attach | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, skip_hlu_0=False):\n try:\n return self._attach_with_retry(lun_or_snap, skip_hlu_0)\n except ex.SystemAPINotSupported:\n ... |
def x(self, lun_or_snap, cg_member=None):
hlu = self.get_hlu(lun_or_snap, cg_member=cg_member)
return hlu is not None
| has_hlu | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, cg_member=None):\n hlu = self.get_hlu(lun_or_snap, cg_member=cg_member)\n return hlu is not None\n",
"role": "user"
},
{
"... |
def x(self, lun_or_snap, cg_member=None):
import storops.unity.resource.lun as lun_module
import storops.unity.resource.snap as snap_module
which = None
if isinstance(lun_or_snap, lun_module.UnityLun):
which = self._get_host_luns(lun=lun_or_snap)
elif isinstance(lun_or_snap, snap_module.UnitySnap):
if lun_or_snap.is_cg_snap():
if cg_member is None:
log.debug(
'None host lun for CG snap {}. Use its member snap instead or pass in cg_member.'
.format(lun_or_snap.id))
return None
lun_or_snap = lun_or_snap.get_member_snap(cg_member)
which = self._get_host_luns(lun=cg_member, snap=lun_or_snap)
else:
which = self._get_host_luns(snap=lun_or_snap)
if not which:
log.debug('Resource(LUN or Snap) {} is not attached to host {}'.
format(lun_or_snap.name, self.name))
return None
return which[0]
| get_host_lun | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, cg_member=None):\n import storops.unity.resource.lun as lun_module\n import storops.unity.resource.snap as snap_module\n which = No... |
def x(self, resource, cg_member=None):
host_lun = self.get_host_lun(resource, cg_member=cg_member)
return host_lun if host_lun is None else host_lun.hlu
| get_hlu | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resource, cg_member=None):\n host_lun = self.get_host_lun(resource, cg_member=cg_member)\n return host_lun if host_lun is None else host_lun.hlu\n"... |
def x(self, iqns=None, wwns=None):
iqns = set(iqns) if iqns else set()
current_iqns = set()
if self.iscsi_host_initiators:
current_iqns = {initiator.initiator_id for initiator in self.
iscsi_host_initiators}
wwns = set(wwns) if wwns else set()
current_wwns = set()
if self.fc_host_initiators:
current_wwns = {initiator.initiator_id for initiator in self.
fc_host_initiators}
updater = UnityHostInitiatorUpdater(self, current_iqns | current_wwns,
iqns | wwns)
return updater.update()
| update_initiators | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, iqns=None, wwns=None):\n iqns = set(iqns) if iqns else set()\n current_iqns = set()\n if self.iscsi_host_initiators:\n current_iqns = {in... |
def x(tag=None, sat_id=None, data_path=None, format_str=None,
supported_tags=None, fake_daily_files_from_monthly=False,
two_digit_year_break=None):
if data_path is not None:
if format_str is None:
try:
format_str = supported_tags[sat_id][tag]
except KeyError:
raise ValueError('Unknown tag')
out = pysat.Files.from_os(data_path=data_path, format_str=format_str)
if not out.empty and fake_daily_files_from_monthly:
out.ix[out.index[-1] + pds.DateOffset(months=1) - pds.
DateOffset(days=1)] = out.iloc[-1]
out = out.asfreq('D', 'pad')
out = out + '_' + out.index.strftime('%Y-%m-%d')
return out
return out
else:
estr = (
'A directory must be passed to the loading routine for <Instrument Code>'
)
raise ValueError(estr)
| list_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None,\n supported_tags=None, fake_daily_files_from_monthly=False,\n two_digit_year_break=None):\n if... |
def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False,
flatten_twod=True):
import pysatCDF
if len(fnames) <= 0:
return pysat.DataFrame(None), None
elif fake_daily_files_from_monthly:
fname = fnames[0][0:-11]
date = pysat.datetime.strptime(fnames[0][-10:], '%Y-%m-%d')
with pysatCDF.CDF(fname) as cdf:
data, meta = cdf.to_pysat(flatten_twod=flatten_twod)
data = data.ix[date:date + pds.DateOffset(days=1) - pds.
DateOffset(microseconds=1), :]
return data, meta
else:
with pysatCDF.CDF(fnames[0]) as cdf:
return cdf.to_pysat(flatten_twod=flatten_twod)
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False,\n flatten_twod=True):\n import pysatCDF\n if len(fnames) <= 0:\n return pys... |
def x(supported_tags, date_array, tag, sat_id, ftp_site=
'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None,
fake_daily_files_from_monthly=False):
import os
import ftplib
ftp = ftplib.FTP(ftp_site)
ftp.login()
try:
ftp_dict = supported_tags[tag]
except KeyError:
raise ValueError('Tag name unknown.')
ftp.cwd(ftp_dict['dir'])
remote_fname = ftp_dict['remote_fname']
local_fname = ftp_dict['local_fname']
for date in date_array:
formatted_remote_fname = remote_fname.format(year=date.year, month=
date.month, day=date.day)
formatted_local_fname = local_fname.format(year=date.year, month=
date.month, day=date.day)
saved_local_fname = os.path.join(data_path, formatted_local_fname)
try:
print('Attempting to download file for ' + date.strftime('%x'))
sys.stdout.flush()
ftp.retrbinary('RETR ' + formatted_remote_fname, open(
saved_local_fname, 'wb').write)
print('Finished.')
except ftplib.error_perm as exception:
if str(exception.args[0]).split(' ', 1)[0] != '550':
raise
else:
os.remove(saved_local_fname)
print('File not available for ' + date.strftime('%x'))
ftp.close()
| download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(supported_tags, date_array, tag, sat_id, ftp_site=\n 'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None,\n fake_daily_files_from_monthly=Fa... |
def x(response, method, url):
req_id = response.headers.get('x-openstack-request-id')
if not req_id:
req_id = response.headers.get('x-compute-request-id')
kwargs = {'http_status': response.status_code, 'response': response,
'method': method, 'url': url, 'request_id': req_id}
if 'retry-after' in response.headers:
kwargs['retry_after'] = response.headers['retry-after']
content_type = response.headers.get('Content-Type', '')
if content_type.startswith('application/json'):
try:
body = response.json()
except ValueError:
pass
else:
if isinstance(body, dict):
error = body.get(list(body)[0])
if isinstance(error, dict):
kwargs['message'] = error.get('message') or error.get(
'faultstring')
kwargs['details'] = error.get('details') or six.text_type(
body)
elif content_type.startswith('text/'):
kwargs['details'] = response.text
try:
cls = _code_map[response.status_code]
except KeyError:
if 500 <= response.status_code < 600:
cls = HttpServerError
elif 400 <= response.status_code < 500:
cls = HTTPClientError
else:
cls = HttpError
return cls(**kwargs)
| from_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(response, method, url):\n req_id = response.headers.get('x-openstack-request-id')\n if not req_id:\n req_id = response.headers.get('x-compute-requ... |
def x(self, name, raid_groups, description=None, **kwargs):
return UnityPool.create(self._cli, name=name, description=description,
raid_groups=raid_groups, **kwargs)
| create_pool | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, raid_groups, description=None, **kwargs):\n return UnityPool.create(self._cli, name=name, description=description,\n raid_groups=raid_gro... |
def x(self):
eths = self.get_ethernet_port(bond=False)
las = self.get_link_aggregation()
return eths + las
| get_file_port | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n eths = self.get_ethernet_port(bond=False)\n las = self.get_link_aggregation()\n return eths + las\n",
"role": "user"
},
{
"conten... |
def x(self, management_address, local_username=None, local_password=None,
remote_username=None, remote_password=None, connection_type=None):
return UnityRemoteSystem.create(self._cli, management_address,
local_username=local_username, local_password=local_password,
remote_username=remote_username, remote_password=remote_password,
connection_type=connection_type)
| create_remote_system | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, management_address, local_username=None, local_password=None,\n remote_username=None, remote_password=None, connection_type=None):\n return UnityRe... |
def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None,
gateway=None, vlan_id=None):
return UnityReplicationInterface.create(self._cli, sp, ip_port,
ip_address, netmask=netmask, v6_prefix_length=v6_prefix_length,
gateway=gateway, vlan_id=vlan_id)
| create_replication_interface | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None,\n gateway=None, vlan_id=None):\n return UnityReplicationInterface.create(self._cli, s... |
def x(self, res_id, backup_snap=None):
name = self._get_name()
out = self._cli.restore_snap(name, res_id, backup_snap)
ex.raise_if_err(out, 'failed to restore snap {}.'.format(name), default
=ex.VNXSnapError)
| restore | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, res_id, backup_snap=None):\n name = self._get_name()\n out = self._cli.restore_snap(name, res_id, backup_snap)\n ex.raise_if_err(out, 'failed to... |
def x(tag='', sat_id=None, data_path=None, format_str=None):
if format_str is None and data_path is not None:
file_base = 'supermag_magnetometer'
if tag == 'indices' or tag == 'all':
file_base += '_all'
if tag == 'indices':
psplit = path.split(data_path[:-1])
data_path = path.join(psplit[0], 'all', '')
if tag == 'stations':
min_fmt = '_'.join([file_base, '{year:4d}.???'])
doff = pds.DateOffset(years=1)
else:
min_fmt = '_'.join([file_base, '{year:4d}{month:02d}{day:02d}.???']
)
doff = pds.DateOffset(days=1)
files = pysat.Files.from_os(data_path=data_path, format_str=min_fmt)
if not files.empty:
files = files.sort_index()
if tag == 'stations':
orig_files = files.copy()
new_files = []
for orig in orig_files.iteritems():
files.ix[orig[0] + doff - pds.DateOffset(days=1)] = orig[1]
files = files.sort_index()
new_files.append(files.ix[orig[0]:orig[0] + doff - pds.
DateOffset(days=1)].asfreq('D', method='pad'))
files = pds.concat(new_files)
files = files.dropna()
files = files.sort_index()
files = files + '_' + files.index.strftime('%Y-%m-%d')
return files
elif format_str is None:
estr = 'A directory must be passed to the loading routine for SuperMAG'
raise ValueError(estr)
else:
return pysat.Files.from_os(data_path=data_path, format_str=format_str)
| list_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and data_path is not None:\n file_base = 'supermag_magnetometer'\n ... |
def x(fnames, tag='', sat_id=None):
if len(fnames) <= 0:
return pysat.DataFrame(None), pysat.Meta(None)
if isinstance(fnames, str):
fnames = [fnames]
data = pds.DataFrame()
baseline = list()
for fname in fnames:
fname = fname[:-11]
file_type = path.splitext(fname)[1].lower()
if file_type == '.csv':
if tag != 'indices':
temp = load_csv_data(fname, tag)
else:
temp, bline = load_ascii_data(fname, tag)
if bline is not None:
baseline.append(bline)
if len(temp.columns) > 0:
data = pds.concat([data, temp], axis=0)
del temp
if len(data.columns) > 0:
meta = pysat.Meta()
for cc in data.columns:
meta[cc] = update_smag_metadata(cc)
meta.info = {'baseline': format_baseline_list(baseline)}
else:
meta = pysat.Meta(None)
return data, meta
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag='', sat_id=None):\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str):\n fname... |
def x(fname, tag):
import re
if tag == 'stations':
ddict = dict()
dkeys = list()
date_list = list()
with open(fname, 'r') as fopen:
dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0
], '%Y')
for fline in fopen.readlines():
sline = [ll for ll in re.split('[,\\n]+', fline) if len(ll) > 0
]
if len(ddict.items()) == 0:
for kk in sline:
kk = re.sub('-', '_', kk)
ddict[kk] = list()
dkeys.append(kk)
else:
date_list.append(dtime)
for i, ll in enumerate(sline):
if i >= 1 and i <= 4:
ddict[dkeys[i]].append(float(ll))
elif i == 6:
ddict[dkeys[i]].append(int(ll))
elif i < len(dkeys):
ddict[dkeys[i]].append(ll)
else:
ddict[dkeys[-1]][-1] += ' {:s}'.format(ll)
data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())
else:
def x(dd):
return pysat.datetime.strptime(dd, '%Y-%m-%d %H:%M:%S')
data = pds.read_csv(fname, parse_dates={'datetime': [0]},
date_parser=parse_smag_date, index_col='datetime')
return data
| load_csv_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname, tag):\n import re\n if tag == 'stations':\n ddict = dict()\n dkeys = list()\n date_list = list()\n with open(fname, 'r... |
def x(fname, tag):
import re
ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8}
dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']}
data = pds.DataFrame(None)
baseline = None
if not tag in ndata.keys():
return data, baseline
with open(fname, 'r') as fopen:
hflag = True
pflag = False
dflag = False if tag == 'stations' else True
snum = 0
ddict = dict()
date_list = list()
if tag == 'stations':
dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0
], '%Y')
for fline in fopen.readlines():
line_len = len(fline)
if hflag:
if pflag:
pflag = False
if fline.find('-mlt') > 0:
ndata[''] += 2
dkeys[''].extend(['MLT', 'MLAT'])
if fline.find('-sza') > 0:
ndata[''] += 1
dkeys[''].append('SZA')
if fline.find('-decl') > 0:
ndata[''] += 1
dkeys[''].append('IGRF_DECL')
if tag == 'indices' and fline.find('-envelope') < 0:
break
lsplit = fline.split()
idelta = lsplit.index('-delta') + 1
ibase = lsplit.index('-baseline') + 1
isd = lsplit.index('-sd') + 1
ist = lsplit.index('-st') + 1
iex = lsplit.index('-ex') + 1
baseline = ' '.join([lsplit[ibase], lsplit[idelta],
lsplit[isd], lsplit[ist], lsplit[iex]])
if fline.find('Selected parameters:') >= 0:
pflag = True
if fline.count('=') == line_len - 1 and line_len > 2:
hflag = False
else:
lsplit = [ll for ll in re.split('[\\t\\n]+', fline) if len(
ll) > 0]
if dflag:
dflag = False
dstring = ' '.join(lsplit[:6])
dtime = pysat.datetime.strptime(dstring,
'%Y %m %d %H %M %S')
snum = int(lsplit[6])
if tag == 'indices':
date_list.append(dtime)
else:
date_list.extend([dtime for i in range(snum)])
elif len(lsplit) == ndata['indices']:
if tag is not '':
if lsplit[0] not in ddict.keys():
ddict[lsplit[0]] = list()
if tag == 'indices':
ddict[lsplit[0]].append(int(lsplit[1]))
else:
ddict[lsplit[0]].extend([int(lsplit[1]) for i in
range(snum)])
elif tag == 'stations' and len(lsplit) >= ndata[tag]:
if len(dkeys[tag]) == 0:
for ll in lsplit:
ll = re.sub('-', '_', ll)
dkeys[tag].append(ll)
ddict[ll] = list()
else:
date_list.append(dtime)
for i, ll in enumerate(lsplit):
if i >= 1 and i <= 4:
ddict[dkeys[tag][i]].append(float(ll))
elif i == 6:
ddict[dkeys[tag][i]].append(int(ll))
elif i < len(dkeys[tag]):
ddict[dkeys[tag][i]].append(ll)
else:
ddict[dkeys[tag][-1]][-1] += ' {:s}'.format(ll)
elif len(lsplit) == ndata['']:
snum -= 1
if tag != 'indices':
if len(ddict.keys()) < ndata['']:
for kk in dkeys['']:
ddict[kk] = list()
for i, kk in enumerate(dkeys['']):
if i == 0:
ddict[kk].append(lsplit[i])
else:
ddict[kk].append(float(lsplit[i]))
if tag != 'stations' and snum == 0 and len(ddict.items()) >= 2:
dflag = True
data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())
fopen.close()
return data, baseline
| load_ascii_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname, tag):\n import re\n ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8}\n dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']}\n dat... |
def x(col_name):
smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT':
'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL':
'degrees', 'SMU': 'none', 'SML': 'none', 'datetime':
'YYYY-MM-DD HH:MM:SS', 'GEOLON': 'degrees', 'GEOLAT': 'degrees',
'AACGMLON': 'degrees', 'AACGMLAT': 'degrees', 'STATION_NAME':
'none', 'OPERATOR_NUM': 'none', 'OPERATORS': 'none'}
smag_name = {'IAGA': 'Station Code', 'N':
'B along local magnetic North', 'E': 'B along local magnetic East',
'Z': 'B vertically downward', 'MLT': 'Magnetic Local Time', 'MLAT':
'Magnetic Latitude', 'SZA': 'Solar Zenith Angle', 'IGRF_DECL':
'IGRF magnetic declination', 'SMU':
"""Maximum eastward auroral electrojets strength.
Upper envelope of N-component for stations between 40 and 80 degrees magnetic north."""
, 'SML':
"""Maximum westward auroral electrojets strength.
Lower envelope of N-component for stations between 40 and 80 degrees magnetic north."""
, 'datetime': 'UT date and time', 'GEOLON': 'geographic longitude',
'GEOLAT': 'geographic latitude', 'AACGMLON':
'Altitude-Adjusted Corrected Geomagnetic longitude', 'AACGMLAT':
'Altitude-Adjusted Corrected Geomagnetic latitude', 'STATION_NAME':
'Long form station name', 'OPERATOR_NUM':
'Number of station operators', 'OPERATORS': 'Station operator name(s)'}
ackn = 'When using this data please include the following reference:\n'
ackn += 'Gjerloev, J. W., The SuperMAG data processing technique, '
ackn += 'Geophys. Res., 117, A09213, doi:10.1029/2012JA017683, 2012\n\n'
ackn += 'For publications and presentations, please include the following'
ackn += 'acknowledgement:\nFor the ground magnetometer data we gratefully '
ackn += 'acknowledge: Intermagnet; USGS, Jeffrey J. Love; CARISMA, PI Ian '
ackn += 'Mann; CANMOS; The S-RAMP Database, PI K. Yumoto and Dr. K. '
ackn += 'Shiokawa; The SPIDR database; AARI, PI Oleg Troshichev; The '
ackn += 'MACCS program, PI M. Engebretson, Geomagnetism Unit of the '
ackn += 'Geological Survey of Canada; GIMA; MEASURE, UCLA IGPP and Florida'
ackn += ' Institute of Technology; SAMBA, PI Eftyhia Zesta; 210 Chain, PI '
ackn += 'K. Yumoto; SAMNET, PI Farideh Honary; The institutes who maintain'
ackn += (
' the IMAGE magnetometer array, PI Eija Tanskanen; PENGUIN; AUTUMN,')
ackn += (
' PI Martin Connors; DTU Space, PI Dr. Rico Behlke; South Pole and ')
ackn += " McMurdo Magnetometer, PI's Louis J. Lanzarotti and Alan T. "
ackn += (
'Weatherwax; ICESTAR; RAPIDMAG; PENGUIn; British Artarctic Survey; ')
ackn += 'McMac, PI Dr. Peter Chi; BGS, PI Dr. Susan Macmillan; Pushkov '
ackn += 'Institute of Terrestrial Magnetism, Ionosphere and Radio Wave '
ackn += 'Propagation (IZMIRAN); GFZ, PI Dr. Juergen Matzka; MFGI, PI B. '
ackn += 'Heilig; IGFPAS, PI J. Reda; University of L’Aquila, PI M. '
ackn += 'Vellante; BCMT, V. Lesur and A. Chambodut; Data obtained in '
ackn += 'cooperation with Geoscience Australia, PI Marina Costelloe; '
ackn += 'SuperMAG, PI Jesper W. Gjerloev.'
col_dict = {'units': smag_units[col_name], 'long_name': smag_name[
col_name], 'acknowledgements': ackn}
return col_dict
| update_smag_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(col_name):\n smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT':\n 'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL':\n ... |
def x(baseline_list):
uniq_base = dict()
uniq_delta = dict()
for bline in baseline_list:
bsplit = bline.split()
bdate = ' '.join(bsplit[2:])
if bsplit[0] not in uniq_base.keys():
uniq_base[bsplit[0]] = ''
if bsplit[1] not in uniq_delta.keys():
uniq_delta[bsplit[1]] = ''
uniq_base[bsplit[0]] += '{:s}, '.format(bdate)
uniq_delta[bsplit[1]] += '{:s}, '.format(bdate)
if len(uniq_base.items()) == 1:
base_string = 'Baseline {:s}'.format(list(uniq_base.keys())[0])
else:
base_string = 'Baseline '
for i, kk in enumerate(uniq_base.keys()):
if i == 1:
base_string += '{:s}: {:s}'.format(kk, uniq_base[kk][:-2])
else:
base_string += ' {:s}: {:s}'.format(kk, uniq_base[
kk][:-2])
else:
base_string += 'unknown'
if len(uniq_delta.items()) == 1:
base_string += '\nDelta {:s}'.format(list(uniq_delta.keys())[0])
else:
base_string += '\nDelta '
for i, kk in enumerate(uniq_delta.keys()):
if i == 1:
base_string += '{:s}: {:s}'.format(kk, uniq_delta[kk][:-2])
else:
base_string += ' {:s}: {:s}'.format(kk, uniq_delta[
kk][:-2])
else:
base_string += 'unknown'
return base_string
| format_baseline_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(baseline_list):\n uniq_base = dict()\n uniq_delta = dict()\n for bline in baseline_list:\n bsplit = bline.split()\n bdate = ' '.join(bsp... |
def x(file_strings, file_fmt, tag):
if file_fmt.lower() == 'csv':
return append_csv_data(file_strings)
else:
return append_ascii_data(file_strings, tag)
| append_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings, file_fmt, tag):\n if file_fmt.lower() == 'csv':\n return append_csv_data(file_strings)\n else:\n return append_ascii_data(fil... |
def x(file_strings, tag):
import re
out_lines = file_strings[0].split('\n')
iparam = -1
ihead = -1
idates = list()
date_list = list()
num_stations = list()
ind_num = 2 if tag in ['all', 'indices', ''] else 0
for i, line in enumerate(out_lines):
if line == 'Selected parameters:':
iparam = i + 1
elif line.count('=') == len(line) and len(line) > 2:
ihead = i
break
i = ihead + 1
while i < len(out_lines) - 1:
idates.append(i)
lsplit = re.split('\t+', out_lines[i])
dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),
'%Y %m %d %H %M %S')
date_list.append(dtime)
num_stations.append(int(lsplit[-1]))
i += num_stations[-1] + 1 + ind_num
idates = np.array(idates)
station_names = list()
for ff in range(len(file_strings) - 1):
file_lines = file_strings[ff + 1].split('\n')
head = True
snum = 0
for i, line in enumerate(file_lines):
if head:
if line.count('=') == len(line) and len(line) > 2:
head = False
elif len(line) > 0:
lsplit = re.split('\t+', line)
if snum == 0:
dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),
'%Y %m %d %H %M %S')
try:
idate = date_list.index(dtime)
except:
raise ValueError('Unexpected date ', dtime)
snum = int(lsplit[-1])
onum = num_stations[idate]
inum = ind_num
idates[idate + 1:] += snum
num_stations[idate] += snum
oline = '{:s}\t{:d}'.format(dtime.strftime(
'%Y\t%m\t%d\t%H\t%M\t%S'), num_stations[idate])
out_lines[idates[idate]] = oline
elif inum > 0:
inum -= 1
else:
onum += 1
snum -= 1
out_lines.insert(idates[idate] + onum, line)
if not lsplit[0] in station_names:
station_names.append(lsplit[0])
out_lines[iparam] += ',' + ','.join(station_names)
out_string = '\n'.join(out_lines)
return out_string
| append_ascii_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings, tag):\n import re\n out_lines = file_strings[0].split('\\n')\n iparam = -1\n ihead = -1\n idates = list()\n date_list = list()\... |
def x(file_strings):
out_lines = list()
head_line = None
for fstrings in file_strings:
file_lines = fstrings.split('\n')
head_line = file_lines.pop(0)
out_lines.extend(file_lines)
out_lines.sort()
i = 0
while len(out_lines[i]) == 0:
out_lines.pop(i)
out_lines.insert(0, head_line)
out_lines.append('')
out_string = '\n'.join(out_lines)
return out_string
| append_csv_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings):\n out_lines = list()\n head_line = None\n for fstrings in file_strings:\n file_lines = fstrings.split('\\n')\n head_line ... |
def x(self):
self.custom.add(add_quasi_dipole_coordinates, 'modify')
self.custom.add(add_aacgm_coordinates, 'modify')
self.custom.add(calculate_ecef_velocity, 'modify')
self.custom.add(add_sc_attitude_vectors, 'modify')
self.custom.add(add_iri_thermal_plasma, 'modify')
self.custom.add(add_hwm_winds_and_ecef_vectors, 'modify')
self.custom.add(add_igrf, 'modify')
in_meta = {'desc': 'IGRF geomagnetic field expressed in the s/c basis.',
'units': 'nT'}
self.custom.add(project_ecef_vector_onto_sc, 'modify', 'end',
'B_ecef_x', 'B_ecef_y', 'B_ecef_z', 'B_sc_x', 'B_sc_y', 'B_sc_z',
meta=[in_meta.copy(), in_meta.copy(), in_meta.copy()])
self.custom.add(project_hwm_onto_sc, 'modify')
self.custom.add(add_msis, 'modify')
| init | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.custom.add(add_quasi_dipole_coordinates, 'modify')\n self.custom.add(add_aacgm_coordinates, 'modify')\n self.custom.add(calculate_ecef_v... |
def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0,
TLE1=None, TLE2=None):
import sgp4
from sgp4.earth_gravity import wgs72
from sgp4.io import twoline2rv
import ephem
import pysatMagVect
line1 = (
'1 25544U 98067A 18135.61844383 .00002728 00000-0 48567-4 0 9998'
)
line2 = (
'2 25544 51.6402 181.0633 0004018 88.8954 22.2246 15.54059185113452'
)
if TLE1 is not None:
line1 = TLE1
if TLE2 is not None:
line2 = TLE2
satellite = twoline2rv(line1, line2, wgs72)
parts = os.path.split(fnames[0])[-1].split('-')
yr = int(parts[0])
month = int(parts[1])
day = int(parts[2][0:2])
date = pysat.datetime(yr, month, day)
times = pds.date_range(start=date, end=date + pds.DateOffset(seconds=
86399), freq='1S')
on_travis = os.environ.get('ONTRAVIS') == 'True'
if on_travis:
times = times[0:100]
position = []
velocity = []
for time in times:
pos, vel = satellite.propagate(time.year, time.month, time.day,
time.hour, time.minute, time.second)
position.extend(pos)
velocity.extend(vel)
data = pysat.DataFrame({'position_eci_x': position[::3],
'position_eci_y': position[1::3], 'position_eci_z': position[2::3],
'velocity_eci_x': velocity[::3], 'velocity_eci_y': velocity[1::3],
'velocity_eci_z': velocity[2::3]}, index=times)
data.index.name = 'Epoch'
site = ephem.Observer()
site.lon = str(obs_long)
site.lat = str(obs_lat)
site.elevation = obs_alt
sat = ephem.readtle('pysat', line1, line2)
output_params = []
for time in times:
lp = {}
site.date = time
sat.compute(site)
lp['obs_sat_az_angle'] = ephem.degrees(sat.az)
lp['obs_sat_el_angle'] = ephem.degrees(sat.alt)
lp['obs_sat_slant_range'] = sat.range
lp['glat'] = np.degrees(sat.sublat)
lp['glong'] = np.degrees(sat.sublong)
lp['alt'] = sat.elevation / 1000.0
lp['x'], lp['y'], lp['z'] = pysatMagVect.geodetic_to_ecef(lp['glat'
], lp['glong'], lp['alt'])
output_params.append(lp)
output = pds.DataFrame(output_params, index=times)
data[['glong', 'glat', 'alt']] = output[['glong', 'glat', 'alt']]
data[['position_ecef_x', 'position_ecef_y', 'position_ecef_z']] = output[[
'x', 'y', 'z']]
data['obs_sat_az_angle'] = output['obs_sat_az_angle']
data['obs_sat_el_angle'] = output['obs_sat_el_angle']
data['obs_sat_slant_range'] = output['obs_sat_slant_range']
return data, meta.copy()
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0,\n TLE1=None, TLE2=None):\n import sgp4\n from sgp4.earth_gravity import wgs72\n... |
def x(tag=None, sat_id=None, data_path=None, format_str=None):
index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018,
12, 1))
names = [(data_path + date.strftime('%Y-%m-%d') + '.nofile') for date in
index]
return pysat.Series(names, index=index)
| list_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018,\n 12, 1))\n na... |
def x(inst):
import pysatMagVect
inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'
] = pysatMagVect.normalize_vector(inst['velocity_ecef_x'], inst[
'velocity_ecef_y'], inst['velocity_ecef_z'])
inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'
] = pysatMagVect.normalize_vector(-inst['position_ecef_x'], -inst[
'position_ecef_y'], -inst['position_ecef_z'])
inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'
] = pysatMagVect.cross_product(inst['sc_zhat_ecef_x'], inst[
'sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'], inst['sc_xhat_ecef_x'],
inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'])
inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'
] = pysatMagVect.normalize_vector(inst['sc_yhat_ecef_x'], inst[
'sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])
inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'
] = pysatMagVect.cross_product(inst['sc_xhat_ecef_x'], inst[
'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],
inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])
inst.meta['sc_xhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_xhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_xhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, z-component'
}
inst.meta['sc_zhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_zhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_zhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, z-component'
}
inst.meta['sc_yhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_yhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_yhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, z-component'
}
mag = np.sqrt(inst['sc_zhat_ecef_x'] ** 2 + inst['sc_zhat_ecef_y'] ** 2 +
inst['sc_zhat_ecef_z'] ** 2)
idx, = np.where((mag < 0.999999999) | (mag > 1.000000001))
if len(idx) > 0:
print(mag[idx])
raise RuntimeError(
'Unit vector generation failure. Not sufficently orthogonal.')
return
| add_sc_attitude_vectors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst):\n import pysatMagVect\n inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'\n ] = pysatMagVect.normalize_vector(inst['vel... |
def x(inst):
x = inst['position_ecef_x']
vel_x = (x.values[2:] - x.values[0:-2]) / 2.0
y = inst['position_ecef_y']
vel_y = (y.values[2:] - y.values[0:-2]) / 2.0
z = inst['position_ecef_z']
vel_z = (z.values[2:] - z.values[0:-2]) / 2.0
inst[1:-1, 'velocity_ecef_x'] = vel_x
inst[1:-1, 'velocity_ecef_y'] = vel_y
inst[1:-1, 'velocity_ecef_z'] = vel_z
inst.meta['velocity_ecef_x'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
inst.meta['velocity_ecef_y'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
inst.meta['velocity_ecef_z'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
return
| calculate_ecef_velocity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst):\n x = inst['position_ecef_x']\n vel_x = (x.values[2:] - x.values[0:-2]) / 2.0\n y = inst['position_ecef_y']\n vel_y = (y.values[2:] - y.valu... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import apexpy
ap = apexpy.Apex(date=inst.date)
qd_lat = []
qd_lon = []
mlt = []
for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],
inst[alt_label], inst.data.index):
tlat, tlon = ap.geo2qd(lat, lon, alt)
qd_lat.append(tlat)
qd_lon.append(tlon)
mlt.append(ap.mlon2mlt(tlon, time))
inst['qd_lat'] = qd_lat
inst['qd_long'] = qd_lon
inst['mlt'] = mlt
inst.meta['qd_lat'] = {'units': 'degrees', 'long_name':
'Quasi dipole latitude'}
inst.meta['qd_long'] = {'units': 'degrees', 'long_name':
'Quasi dipole longitude'}
inst.meta['qd_mlt'] = {'units': 'hrs', 'long_name': 'Magnetic local time'}
return
| add_quasi_dipole_coordinates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import apexpy\n ap = apexpy.Apex(date=inst.date)\n qd_lat = []\n qd_lon = []\n ... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import aacgmv2
aalat = []
aalon = []
mlt = []
for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],
inst[alt_label], inst.data.index):
tlat, tlon, tmlt = aacgmv2.get_aacgm_coord(lat, lon, alt, time)
aalat.append(tlat)
aalon.append(tlon)
mlt.append(tmlt)
inst['aacgm_lat'] = aalat
inst['aacgm_long'] = aalon
inst['aacgm_mlt'] = mlt
inst.meta['aacgm_lat'] = {'units': 'degrees', 'long_name': 'AACGM latitude'
}
inst.meta['aacgm_long'] = {'units': 'degrees', 'long_name':
'AACGM longitude'}
inst.meta['aacgm_mlt'] = {'units': 'hrs', 'long_name':
'AACGM Magnetic local time'}
return
| add_aacgm_coordinates | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import aacgmv2\n aalat = []\n aalon = []\n mlt = []\n for lat, lon, alt, time ... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
iri_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_iri()
iri = {}
iri['ion_temp'] = pt.Ti
iri['e_temp'] = pt.Te
iri['ion_dens'] = pt.ni['O+'] + pt.ni['H+'] + pt.ni['HE+']
iri['frac_dens_o'] = pt.ni['O+'] / iri['ion_dens']
iri['frac_dens_h'] = pt.ni['H+'] / iri['ion_dens']
iri['frac_dens_he'] = pt.ni['HE+'] / iri['ion_dens']
iri_params.append(iri)
iri = pds.DataFrame(iri_params)
iri.index = inst.data.index
inst[iri.keys()] = iri
inst.meta['ion_temp'] = {'units': 'Kelvin', 'long_name': 'Ion Temperature'}
inst.meta['ion_dens'] = {'units': 'N/cc', 'long_name': 'Ion Density',
'desc': 'Total ion density including O+ and H+ from IRI model run.'}
inst.meta['frac_dens_o'] = {'units': '', 'long_name':
'Fractional O+ Density'}
inst.meta['frac_dens_h'] = {'units': '', 'long_name':
'Fractional H+ Density'}
| add_iri_thermal_plasma | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n iri_params = []\n for time, lat... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
import pysatMagVect
hwm_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = pyglow.Point(time, lat, lon, alt)
pt.run_hwm()
hwm = {}
hwm['zonal_wind'] = pt.u
hwm['meridional_wind'] = pt.v
hwm_params.append(hwm)
hwm = pds.DataFrame(hwm_params)
hwm.index = inst.data.index
inst[['zonal_wind', 'meridional_wind']] = hwm[['zonal_wind',
'meridional_wind']]
mag = np.sqrt(inst['position_ecef_x'] ** 2 + inst['position_ecef_y'] ** 2)
inst['unit_zonal_wind_ecef_x'] = -inst['position_ecef_y'] / mag
inst['unit_zonal_wind_ecef_y'] = inst['position_ecef_x'] / mag
inst['unit_zonal_wind_ecef_z'] = 0 * inst['position_ecef_x']
unit_pos_x, unit_pos_y, unit_pos_z = pysatMagVect.normalize_vector(-
inst['position_ecef_x'], -inst['position_ecef_y'], -inst[
'position_ecef_z'])
inst['unit_mer_wind_ecef_x'], inst['unit_mer_wind_ecef_y'], inst[
'unit_mer_wind_ecef_z'] = pysatMagVect.cross_product(unit_pos_x,
unit_pos_y, unit_pos_z, inst['unit_zonal_wind_ecef_x'], inst[
'unit_zonal_wind_ecef_y'], inst['unit_zonal_wind_ecef_z'])
inst.meta['zonal_wind'] = {'units': 'm/s', 'long_name': 'Zonal Wind',
'desc': 'HWM model zonal wind'}
inst.meta['meridional_wind'] = {'units': 'm/s', 'long_name':
'Meridional Wind', 'desc': 'HWM model meridional wind'}
inst.meta['unit_zonal_wind_ecef_x'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF x-vector', 'desc':
'x-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_zonal_wind_ecef_y'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF y-vector', 'desc':
'y-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_zonal_wind_ecef_z'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF z-vector', 'desc':
'z-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_x'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF x-vector', 'desc':
'x-value of meridional wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_y'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF y-vector', 'desc':
'y-value of meridional wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_z'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF z-vector', 'desc':
'z-value of meridional wind unit vector in ECEF co ordinates'}
return
| add_hwm_winds_and_ecef_vectors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n import pysatMagVect\n hwm_params = []\n for time, lat, lon, alt i... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
import pysatMagVect
igrf_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_igrf()
igrf = {}
igrf['B'] = pt.B
igrf['B_east'] = pt.Bx
igrf['B_north'] = pt.By
igrf['B_up'] = pt.Bz
igrf_params.append(igrf)
igrf = pds.DataFrame(igrf_params)
igrf.index = inst.data.index
inst[igrf.keys()] = igrf
x, y, z = pysatMagVect.enu_to_ecef_vector(inst['B_east'], inst[
'B_north'], inst['B_up'], inst[glat_label], inst[glong_label])
inst['B_ecef_x'] = x
inst['B_ecef_y'] = y
inst['B_ecef_z'] = z
inst.meta['B'] = {'units': 'nT', 'desc':
'Total geomagnetic field from IGRF.'}
inst.meta['B_east'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_north'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_up'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_ecef_x'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
inst.meta['B_ecef_y'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
inst.meta['B_ecef_z'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
return
| add_igrf | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n import pysatMagVect\n igrf_para... |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
msis_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_msis()
msis = {}
total = 0
for key in pt.nn.keys():
total += pt.nn[key]
msis['Nn'] = total
msis['Nn_N'] = pt.nn['N']
msis['Nn_N2'] = pt.nn['N2']
msis['Nn_O'] = pt.nn['O']
msis['Nn_O2'] = pt.nn['O2']
msis['Tn_msis'] = pt.Tn_msis
msis_params.append(msis)
msis = pds.DataFrame(msis_params)
msis.index = inst.data.index
inst[msis.keys()] = msis
inst.meta['Nn'] = {'units': 'cm^-3', 'desc':
'Total neutral number particle density from MSIS.'}
inst.meta['Nn_N'] = {'units': 'cm^-3', 'desc':
'Total nitrogen number particle density from MSIS.'}
inst.meta['Nn_N2'] = {'units': 'cm^-3', 'desc':
'Total N2 number particle density from MSIS.'}
inst.meta['Nn_O'] = {'units': 'cm^-3', 'desc':
'Total oxygen number particle density from MSIS.'}
inst.meta['Nn_O2'] = {'units': 'cm^-3', 'desc':
'Total O2 number particle density from MSIS.'}
inst.meta['Tn_msis'] = {'units': 'K', 'desc':
'Neutral temperature from MSIS.'}
return
| add_msis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n msis_params = []\n for time, la... |
def x(inst, x_label, y_label, z_label, new_x_label, new_y_label,
new_z_label, meta=None):
import pysatMagVect
x, y, z = pysatMagVect.project_ecef_vector_onto_basis(inst[x_label],
inst[y_label], inst[z_label], inst['sc_xhat_ecef_x'], inst[
'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],
inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'], inst[
'sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'])
inst[new_x_label] = x
inst[new_y_label] = y
inst[new_z_label] = z
if meta is not None:
inst.meta[new_x_label] = meta[0]
inst.meta[new_y_label] = meta[1]
inst.meta[new_z_label] = meta[2]
return
| project_ecef_vector_onto_sc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, x_label, y_label, z_label, new_x_label, new_y_label,\n new_z_label, meta=None):\n import pysatMagVect\n x, y, z = pysatMagVect.project_ecef_vect... |
def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None):
plt.ioff()
figs = []
axs = []
if type(data_label) is str:
figs.append(plt.figure())
ax1 = figs[0].add_subplot(211, projection='3d')
ax2 = figs[0].add_subplot(212)
axs.append((ax1, ax2))
plt.suptitle(data_label)
if xlim is not None:
ax1.set_xlim(xlim)
ax2.set_xlim(xlim)
if ylim is not None:
ax1.set_ylim(ylim)
ax2.set_ylim(ylim)
data_label = [data_label]
plt.hold(True)
else:
for i in np.arange(len(data_label)):
figs.append(plt.figure())
ax1 = figs[i].add_subplot(211, projection='3d')
ax2 = figs[i].add_subplot(212)
axs.append((ax1, ax2))
plt.suptitle(data_label[i])
if xlim is not None:
ax1.set_xlim(xlim)
ax2.set_xlim(xlim)
if ylim is not None:
ax1.set_ylim(ylim)
ax2.set_ylim(ylim)
plt.hold(True)
norm = colors.Normalize(vmin=datalim[0], vmax=datalim[1])
p = [i for i in np.arange(len(figs))]
for i, inst in enumerate(inst):
for j, (fig, ax) in enumerate(zip(figs, axs)):
if len(inst.data) > 0:
if (len(inst.data[labelx]) > 0) & (len(inst.data[labely]) > 0
) & (len(inst.data[data_label[j]]) > 0):
p[j] = ax[0].scatter(inst.data[labelx], inst.data[
labely], inst.data[data_label[j]], zdir='z', c=inst
.data[data_label[j]], cmap=cm.jet, norm=norm,
linewidth='0', edgecolors=None)
ax[1].scatter(inst.data[labelx], inst.data[labely], c=
inst.data[data_label[j]], cmap=cm.jet, norm=norm,
linewidth=1e-11, alpha=0.5, edgecolor=None)
for j, (fig, ax) in enumerate(zip(figs, axs)):
try:
cbar = plt.colorbar(p[j], ax=ax[0], label='Amplitude (m/s)')
except:
print('Tried colorbar but failed, thus no colorbar.')
ax[0].elev = 30.0
plt.ion()
return figs
| scatterplot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None):\n plt.ioff()\n figs = []\n axs = []\n if type(data_label) is str:\n figs.a... |
def x(level=logging.DEBUG):
logger = logging.getLogger(__name__)
logger.setLevel(level)
if not logger.handlers:
logger.info('enabling logging to console.')
logger.addHandler(logging.StreamHandler(sys.stdout))
| enable_log | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(level=logging.DEBUG):\n logger = logging.getLogger(__name__)\n logger.setLevel(level)\n if not logger.handlers:\n logger.info('enabling logging... |
def x(tag=None, sat_id=None, data_path=None, format_str=None):
import sys
estr = 'Building a list of COSMIC files, which can possibly take time. '
estr = '{:s}~1s per 100K files'.format(estr)
print(estr)
sys.stdout.flush()
cosmicFiles = glob.glob(os.path.join(data_path, '*/*_nc'))
num = len(cosmicFiles)
if num != 0:
print('Estimated time:', num * 1e-05, 'seconds')
sys.stdout.flush()
year = [None] * num
days = [None] * num
hours = [None] * num
minutes = [None] * num
microseconds = [None] * num
for i, f in enumerate(cosmicFiles):
f2 = f.split('.')
year[i] = f2[-6]
days[i] = f2[-5]
hours[i] = f2[-4]
minutes[i] = f2[-3]
microseconds[i] = i
year = np.array(year).astype(int)
days = np.array(days).astype(int)
uts = np.array(hours).astype(int) * 3600.0 + np.array(minutes).astype(
int) * 60.0
uts += np.mod(np.array(microseconds).astype(int) * 4, 8000) * 1e-05
index = pysat.utils.create_datetime_index(year=year, day=days, uts=uts)
file_list = pysat.Series(cosmicFiles, index=index)
return file_list
else:
print('Found no files, check your path or download them.')
return pysat.Series(None)
| list_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n import sys\n estr = 'Building a list of COSMIC files, which can possibly take time. '\n est... |
def x(files, tag=None, sat_id=None, altitude_bin=None):
output = [None] * len(files)
drop_idx = []
for i, file in enumerate(files):
try:
data = netcdf_file(file, mode='r', mmap=False)
new = {}
ncattrsList = data._attributes.keys()
for d in ncattrsList:
new[d] = data._attributes[d]
loadedVars = {}
keys = data.variables.keys()
for key in keys:
if data.variables[key][:].dtype.byteorder != '=':
loadedVars[key] = data.variables[key][:].byteswap(
).newbyteorder()
else:
loadedVars[key] = data.variables[key][:]
new['profiles'] = pysat.DataFrame(loadedVars)
output[i] = new
data.close()
except RuntimeError:
drop_idx.append(i)
drop_idx.reverse()
for i in drop_idx:
del output[i]
if tag == 'ionprf':
if altitude_bin is not None:
for out in output:
out['profiles'].index = (out['profiles']['MSL_alt'] /
altitude_bin).round().values * altitude_bin
out['profiles'] = out['profiles'].groupby(out['profiles'].
index.values).mean()
else:
for out in output:
out['profiles'].index = out['profiles']['MSL_alt']
return output
| load_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(files, tag=None, sat_id=None, altitude_bin=None):\n output = [None] * len(files)\n drop_idx = []\n for i, file in enumerate(files):\n try:\n ... |
def x(value):
t = 60
if value is not None:
r = value % t
if r > t / 2:
ret = value + (t - r)
else:
ret = value - r
else:
ret = NaN
return ret
| round_60 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n t = 60\n if value is not None:\n r = value % t\n if r > t / 2:\n ret = value + (t - r)\n else:\n ret... |
def x(prev, curr, counters):
busy_prop, idle_prop = counters
pb = getattr(prev, busy_prop)
pi = getattr(prev, idle_prop)
cb = getattr(curr, busy_prop)
ci = getattr(curr, idle_prop)
db = minus(cb, pb)
di = minus(ci, pi)
return mul(div(db, add(db, di)), 100)
| utilization | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n busy_prop, idle_prop = counters\n pb = getattr(prev, busy_prop)\n pi = getattr(prev, idle_prop)\n cb = getattr(curr, busy_... |
def x(prev, curr, counters):
counter = get_counter(counters)
pv = getattr(prev, counter)
cv = getattr(curr, counter)
return minus(cv, pv)
| delta_ps | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n counter = get_counter(counters)\n pv = getattr(prev, counter)\n cv = getattr(curr, counter)\n return minus(cv, pv)\n",
... |
def x(prev, curr, counters):
bw_stats, io_stats = counters
size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats))
return mul(size_mb, 1024)
| io_size_kb | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n bw_stats, io_stats = counters\n size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats))\n return mul(size_mb, 1024)\n... |
def x(tag='', sat_id=None, data_path=None, format_str=None):
if format_str is None and tag is not None:
if tag == '' or tag == 'ascii':
ascii_fmt = 'Density_3deg_{year:02d}_{doy:03d}.ascii'
return pysat.Files.from_os(data_path=data_path, format_str=
ascii_fmt)
else:
raise ValueError('Unrecognized tag name for CHAMP STAR')
elif format_str is None:
estr = 'A tag name must be passed to the loading routine for CHAMP'
raise ValueError(estr)
else:
return pysat.Files.from_os(data_path=data_path, format_str=format_str)
| list_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and tag is not None:\n if tag == '' or tag == 'ascii':\n as... |
def x(fnames, tag=None, sat_id=None):
import re
if len(fnames) <= 0:
return pysat.DataFrame(None), pysat.Meta(None)
if isinstance(fnames, str):
fnames = [fnames]
champ_labels = {'Two-digit Year (years)': 'year',
'Day of the Year (days)': 'doy', 'Second of the Day (GPS time,sec)':
'sod', 'Center Latitude of 3-degree Bin (deg)': 'bin_lat',
'Satellite Geodetic Latitude (deg)': 'sat_glat',
'Satellite Longitude (deg)': 'sat_lon', 'Satellite Height (km)':
'sat_h', 'Satellite Local Time (hours)': 'sat_lt',
'Satellite Quasi-Dipole Latitude (deg)': 'sat_qdlat',
'Satellite Magnetic Longitude (deg)': 'sat_mlon',
'Satellite Magnetic Local Time (hours)': 'sat_mlt',
'Neutral Density (kg/m^3)': 'ndens',
'Neutral Density Normalized to 400km using NRLMSISe00': 'ndens400',
'Neutral Density Normalized to 410km using NRLMSISe00': 'ndens410',
'NRLMSISe00 Neutral Density at Satellite Height': 'nrlmsis_ndens',
'Uncertainty in Neutral Density (kg/m^3)': 'ndens_err',
'Number of Data Points in Current Averaging Bin': 'npnts',
'Number of Points in Current Averaging Bin that Required Interpolation'
: 'npnts_interp',
'Average Coefficient of Drag Used in Current Averaging Bin':
'avg_drag_coeff'}
champ_dtypes = {'year': np.int32, 'doy': np.int32, 'sod': float,
'bin_lat': float, 'sat_glat': float, 'sat_lon': float, 'sat_h':
float, 'sat_lt': float, 'sat_qdlat': float, 'sat_mlon': float,
'sat_mlt': float, 'ndens': float, 'ndens400': float, 'ndens410':
float, 'nrlmsis_ndens': float, 'ndens_err': float, 'npnts': int,
'npnts_interp': float, 'avg_drag_coeff': float}
champ_units = {'year': '2-digit years', 'doy': 'day of year', 'sod':
'seconds of day', 'bin_lat': 'degrees', 'sat_glat': 'degrees',
'sat_lon': 'degrees', 'sat_h': 'km', 'sat_lt': 'hours', 'sat_qdlat':
'degrees', 'sat_mlon': 'degrees', 'sat_mlt': 'hours', 'ndens':
'km m^{-3}', 'ndens400': 'km m^{-3}', 'ndens410': 'km m^{-3}',
'nrlmsis_ndens': 'km m^{-3}', 'ndens_err': 'km m^{-3}', 'npnts':
'number', 'npnts_interp': 'number', 'avg_drag_coeff': 'unitless'}
def x(y, d, s):
import datetime as dt
t = dt.datetime.strptime('{:02d} {:03d}'.format(int(y), int(d)),
'%y %j')
fsec = float(s)
isec = np.floor(fsec)
microsec = (fsec - isec) * 1000000.0
t += dt.timedelta(seconds=isec, microseconds=microsec)
return t
f = open(fnames[0], 'r')
vdata = re.split(';|\n', f.readline())
hdata = re.split(';|\n', f.readline())
f.close()
try:
hdata.pop(hdata.index(''))
except:
pass
data = pds.read_csv(fnames[0], delim_whitespace=True, skiprows=2,
header=None, names=[champ_labels[h] for h in hdata], keep_date_col=
True, index_col='datetime', parse_dates={'datetime': [0, 1, 2]},
date_parser=parse_champdate)
meta = pysat.Meta()
for h in hdata:
col = champ_labels[h]
data[col].astype(champ_dtypes[col])
meta[col] = {'units': champ_units[col], 'long_name': h}
return data, meta
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n import re\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str... |
def x(self, by_name=False, inst_module=None):
import importlib
self._list_rtn = self._pass_func
self._load_rtn = self._pass_func
self._default_rtn = self._pass_func
self._clean_rtn = self._pass_func
self._init_rtn = self._pass_func
self._download_rtn = self._pass_func
self.directory_format = None
self.file_format = None
self.multi_file_day = False
self.orbit_info = None
if by_name:
inst = importlib.import_module(''.join(('.', self.platform, '_',
self.name)), package='pysat.instruments')
elif inst_module is not None:
inst = inst_module
else:
return
try:
self._load_rtn = inst.load
self._list_rtn = inst.list_files
self._download_rtn = inst.download
except AttributeError:
estr = 'A load, file_list, and download routine are required for '
raise AttributeError('{:s}every instrument.'.format(estr))
try:
self._default_rtn = inst.default
except AttributeError:
pass
try:
self._init_rtn = inst.init
except AttributeError:
pass
try:
self._clean_rtn = inst.clean
except AttributeError:
pass
try:
self.directory_format = inst.directory_format
except AttributeError:
pass
try:
self.multi_file_day = inst.multi_file_day
except AttributeError:
pass
try:
self.orbit_info = inst.orbit_info
except AttributeError:
pass
return
| assign_funcs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, by_name=False, inst_module=None):\n import importlib\n self._list_rtn = self._pass_func\n self._load_rtn = self._pass_func\n self._default_rt... |
def x(self, date=None, fid=None):
if fid is not None:
fname = self.files[fid:fid + 1]
elif date is not None:
fname = self.files[date:date + pds.DateOffset(days=1)]
else:
raise ValueError('Must supply either a date or file id number.')
if len(fname) > 0:
load_fname = [os.path.join(self.files.data_path, f) for f in fname]
data, mdata = self._load_rtn(load_fname, tag=self.tag, sat_id=self.
sat_id, **self.kwargs)
mdata.accept_default_labels(self)
else:
data = DataFrame(None)
mdata = _meta.Meta(units_label=self.units_label, name_label=self.
name_label, notes_label=self.notes_label, desc_label=self.
desc_label, plot_label=self.plot_label, axis_label=self.
axis_label, scale_label=self.scale_label, min_label=self.
min_label, max_label=self.max_label, fill_label=self.fill_label)
output_str = '{platform} {name} {tag} {sat_id}'
output_str = output_str.format(platform=self.platform, name=self.name,
tag=self.tag, sat_id=self.sat_id)
if not data.empty:
if not isinstance(data, DataFrame):
raise TypeError(' '.join(('Data returned by instrument load',
'routine must be a pandas.DataFrame')))
if not isinstance(mdata, _meta.Meta):
raise TypeError('Metadata returned must be a pysat.Meta object')
if date is not None:
output_str = ' '.join(('Returning', output_str, 'data for',
date.strftime('%x')))
elif len(fname) == 1:
output_str = ' '.join(('Returning', output_str, 'data from',
fname[0]))
else:
output_str = ' '.join(('Returning', output_str, 'data from',
fname[0], '::', fname[-1]))
else:
output_str = ' '.join(('No', output_str, 'data for', date.strftime(
'%m/%d/%y')))
output_str = ' '.join(output_str.split())
print(output_str)
return data, mdata
| load_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, date=None, fid=None):\n if fid is not None:\n fname = self.files[fid:fid + 1]\n elif date is not None:\n fname = self.files[date:date... |
def x(self):
if self._load_by_date:
next_date = self.date + pds.DateOffset(days=1)
return self._load_data(date=next_date)
else:
return self._load_data(fid=self._fid + 1)
| load_next | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._load_by_date:\n next_date = self.date + pds.DateOffset(days=1)\n return self._load_data(date=next_date)\n else:\n ... |
def x(self):
if self._load_by_date:
prev_date = self.date - pds.DateOffset(days=1)
return self._load_data(date=prev_date)
else:
return self._load_data(fid=self._fid - 1)
| load_prev | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._load_by_date:\n prev_date = self.date - pds.DateOffset(days=1)\n return self._load_data(date=prev_date)\n else:\n ... |
def x(self, start, stop, freq='D', user=None, password=None, **kwargs):
import errno
try:
os.makedirs(self.files.data_path)
except OSError as e:
if e.errno != errno.EEXIST:
raise
print('Downloading data to: ', self.files.data_path)
date_array = utils.season_date_range(start, stop, freq=freq)
if user is None:
self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,
data_path=self.files.data_path, **kwargs)
else:
self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,
data_path=self.files.data_path, user=user, password=password,
**kwargs)
first_date = self.files.start_date
last_date = self.files.stop_date
print('Updating pysat file list')
self.files.refresh()
if len(self.bounds[0]) == 1:
if self.bounds[0][0] == first_date and self.bounds[1][0] == last_date:
print('Updating instrument object bounds.')
self.bounds = None
| download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, start, stop, freq='D', user=None, password=None, **kwargs):\n import errno\n try:\n os.makedirs(self.files.data_path)\n except OSError as... |
def x(self, verifyPad=False):
if self._iter_type == 'date':
if self.date is not None:
idx, = np.where(self._iter_list == self.date)
if len(idx) == 0:
raise StopIteration('File list is empty. Nothing to be done.')
elif idx[-1] + 1 >= len(self._iter_list):
raise StopIteration('Outside the set date boundaries.')
else:
idx += 1
self.load(date=self._iter_list[idx[0]], verifyPad=verifyPad)
else:
self.load(date=self._iter_list[0], verifyPad=verifyPad)
elif self._iter_type == 'file':
if self._fid is not None:
first = self.files.get_index(self._iter_list[0])
last = self.files.get_index(self._iter_list[-1])
if (self._fid < first) | (self._fid + 1 > last):
raise StopIteration('Outside the set file boundaries.')
else:
self.load(fname=self._iter_list[self._fid + 1 - first],
verifyPad=verifyPad)
else:
self.load(fname=self._iter_list[0], verifyPad=verifyPad)
| next | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verifyPad=False):\n if self._iter_type == 'date':\n if self.date is not None:\n idx, = np.where(self._iter_list == self.date)\n ... |
def x(self, coltype):
if type(coltype) is np.dtype:
var_type = coltype.kind + str(coltype.itemsize)
return var_type
elif coltype is np.int64:
return 'i8'
elif coltype is np.int32:
return 'i4'
elif coltype is np.int16:
return 'i2'
elif coltype is np.int8:
return 'i1'
elif coltype is np.uint64:
return 'u8'
elif coltype is np.uint32:
return 'u4'
elif coltype is np.uint16:
return 'u2'
elif coltype is np.uint8:
return 'u1'
elif coltype is np.float64:
return 'f8'
elif coltype is np.float32:
return 'f4'
elif issubclass(coltype, basestring):
return 'S1'
else:
raise TypeError('Unknown Variable Type' + str(coltype))
| get_var_type_code | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, coltype):\n if type(coltype) is np.dtype:\n var_type = coltype.kind + str(coltype.itemsize)\n return var_type\n elif coltype is np.in... |
def x(self, data, file_format):
data_type = data.dtype
if file_format != 'NETCDF4':
old_format = True
else:
old_format = False
if data_type != np.dtype('O'):
if (data_type == np.int64) & old_format:
data = data.astype(np.int32)
data_type = np.int32
if data_type == np.dtype('<M8[ns]'):
if not old_format:
data_type = np.int64
else:
data_type = np.float
datetime_flag = True
else:
datetime_flag = False
else:
data_type = type(data.iloc[0])
for i in np.arange(len(data)):
if len(data.iloc[i]) > 0:
data_type = type(data.iloc[i])
if not isinstance(data_type, np.float):
break
datetime_flag = False
return data, data_type, datetime_flag
| get_data_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, file_format):\n data_type = data.dtype\n if file_format != 'NETCDF4':\n old_format = True\n else:\n old_format = False\n ... |
def x(self, mdata_dict, coltype, remove=False):
for key in mdata_dict:
if type(mdata_dict[key]) == bool:
mdata_dict[key] = int(mdata_dict[key])
if coltype == type(' ') or coltype == type(u' '):
remove = True
if u'_FillValue' in mdata_dict.keys():
if remove:
mdata_dict.pop('_FillValue')
else:
mdata_dict['_FillValue'] = np.array(mdata_dict['_FillValue']
).astype(coltype)
if u'FillVal' in mdata_dict.keys():
if remove:
mdata_dict.pop('FillVal')
else:
mdata_dict['FillVal'] = np.array(mdata_dict['FillVal']).astype(
coltype)
return mdata_dict
| filter_netcdf4_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mdata_dict, coltype, remove=False):\n for key in mdata_dict:\n if type(mdata_dict[key]) == bool:\n mdata_dict[key] = int(mdata_dict[... |
def x(self, meta_to_translate):
export_dict = {}
if self._meta_translation_table is not None:
translation_table = {}
for key in self._meta_translation_table:
translation_table[getattr(self, key)
] = self._meta_translation_table[key]
else:
translation_table = None
for key in meta_to_translate.data.index:
if translation_table is None:
export_dict[key] = meta_to_translate.data.loc[key].to_dict()
else:
export_dict[key] = {}
meta_dict = meta_to_translate.data.loc[key].to_dict()
for original_key in meta_dict:
if original_key in translation_table:
for translated_key in translation_table[original_key]:
export_dict[key][translated_key] = meta_dict[
original_key]
else:
export_dict[key][original_key] = meta_dict[original_key]
for key in meta_to_translate.ho_data:
if key not in export_dict:
export_dict[key] = {}
for ho_key in meta_to_translate.ho_data[key].data.index:
if translation_table is None:
export_dict[key + '_' + ho_key] = meta_to_translate.ho_data[key
].data.loc[ho_key].to_dict()
else:
export_dict[key + '_' + ho_key] = {}
meta_dict = meta_to_translate.ho_data[key].data.loc[ho_key
].to_dict()
for original_key in meta_dict:
if original_key in translation_table:
for translated_key in translation_table[original_key]:
export_dict[key + '_' + ho_key][translated_key
] = meta_dict[original_key]
else:
export_dict[key + '_' + ho_key][original_key
] = meta_dict[original_key]
return export_dict
| generic_meta_translator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, meta_to_translate):\n export_dict = {}\n if self._meta_translation_table is not None:\n translation_table = {}\n for key in self._met... |
def x(input_str):
ret = {}
if input_str is not None:
pattern = re.compile('(\\d+)\\s*(\\d+)')
for line in input_str.split('\n'):
line = line.strip()
if len(line) == 0:
continue
matched = re.search(pattern, line)
if matched is None or len(matched.groups()) < 2:
continue
else:
hlu = matched.group(1)
alu = matched.group(2)
ret[int(alu)] = int(hlu)
return ret
| to_alu_hlu_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_str):\n ret = {}\n if input_str is not None:\n pattern = re.compile('(\\\\d+)\\\\s*(\\\\d+)')\n for line in input_str.split('\\n'):\n... |
def x(value):
ret = []
p = re.compile('Bus\\s+(\\w+)\\s+Enclosure\\s+(\\w+)\\s+Disk\\s+(\\w+)')
if value is not None:
for line in value.split('\n'):
line = line.strip()
if len(line) == 0:
continue
matched = re.search(p, line)
if matched is None or len(matched.groups()) < 3:
continue
else:
ret.append('{}_{}_{}'.format(*matched.groups()))
return ret
| to_disk_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n ret = []\n p = re.compile('Bus\\\\s+(\\\\w+)\\\\s+Enclosure\\\\s+(\\\\w+)\\\\s+Disk\\\\s+(\\\\w+)')\n if value is not None:\n for lin... |
def x(url):
regex_url = ('([a-z][a-z0-9+\\-.]*://)?' +
"([a-z0-9\\-._~%!$&'()*+,;=]+@)?" + '([a-z0-9\\-._~%]+' +
"|\\[[a-z0-9\\-._~%!$&'()*+,;=:]+\\])?" + '(:(?P<port>[0-9]+))?')
m = re.match(regex_url, url, re.IGNORECASE)
if m and m.group(3):
return url[m.start(3):m.end(3)]
else:
raise ValueError('URL without a valid host or ip')
| url_to_host | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url):\n regex_url = ('([a-z][a-z0-9+\\\\-.]*://)?' +\n \"([a-z0-9\\\\-._~%!$&'()*+,;=]+@)?\" + '([a-z0-9\\\\-._~%]+' +\n \"|\\\\[[a-z0-9\\\\-.... |
def x(addr):
if addr.startswith('[') and addr.endswith(']'):
addr = addr[1:-1]
parts = addr.split('/')
if len(parts) == 1:
return parts[0], None, None
if len(parts) > 2:
raise ValueError('Illegal host address')
else:
domain_or_ip, prefix = parts
prefix = int(prefix)
if re.match('^(?:[0-9]{1,3}\\.){3}[0-9]{1,3}$', domain_or_ip):
return domain_or_ip, prefix, ipv4_prefix_to_mask(prefix)
elif ':' in domain_or_ip:
return domain_or_ip, prefix, ipv6_prefix_to_mask(prefix)
else:
return domain_or_ip, None, None
| parse_host_address | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(addr):\n if addr.startswith('[') and addr.endswith(']'):\n addr = addr[1:-1]\n parts = addr.split('/')\n if len(parts) == 1:\n return pa... |
def x(prefix):
if prefix > 32 or prefix < 0:
raise ValueError('invalid cidr prefix for ipv4')
else:
mask = (1 << 32) - 1 ^ (1 << 32 - prefix) - 1
eight_ones = 255
mask_str = ''
for i in range(0, 4):
mask_str = str(mask & eight_ones) + mask_str
mask = mask >> 8
if i != 3:
mask_str = '.' + mask_str
return mask_str
| ipv4_prefix_to_mask | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prefix):\n if prefix > 32 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv4')\n else:\n mask = (1 << 32) - 1 ^ (1 << 32 - pre... |
def x(prefix):
if prefix > 128 or prefix < 0:
raise ValueError('invalid cidr prefix for ipv6')
else:
mask = (1 << 128) - 1 ^ (1 << 128 - prefix) - 1
f = 15
hex_mask_str = ''
for i in range(0, 32):
hex_mask_str = format(mask & f, 'x') + hex_mask_str
mask = mask >> 4
if i != 31 and i & 3 == 3:
hex_mask_str = ':' + hex_mask_str
return hex_mask_str
| ipv6_prefix_to_mask | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prefix):\n if prefix > 128 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv6')\n else:\n mask = (1 << 128) - 1 ^ (1 << 128 - ... |
def x(self, new_size):
ret = self.size_total
resp = self.modify(size=new_size)
resp.raise_if_err()
return ret
| expand | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_size):\n ret = self.size_total\n resp = self.modify(size=new_size)\n resp.raise_if_err()\n return ret\n",
"role": "user"
},
{
... |
def x(self, host_names):
if self.host_access:
curr_hosts = [access.host.name for access in self.host_access]
else:
curr_hosts = []
if set(curr_hosts) == set(host_names):
log.info(
'Hosts for updating is equal to current hosts, skip modification.')
return None
new_hosts = [UnityHostList.get(cli=self._cli, name=host_name)[0] for
host_name in host_names]
new_access = [{'host': item, 'accessMask': HostLUNAccessEnum.PRODUCTION
} for item in new_hosts]
resp = self.modify(host_access=new_access)
resp.raise_if_err()
return resp
| update_hosts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, host_names):\n if self.host_access:\n curr_hosts = [access.host.name for access in self.host_access]\n else:\n curr_hosts = []\n i... |
def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None,
replicate_existing_snaps=None, remote_system=None):
return UnityReplicationSession.create(self._cli, self.get_id(),
dst_lun_id, max_time_out_of_sync, name=replication_name,
replicate_existing_snaps=replicate_existing_snaps, remote_system=
remote_system)
| replicate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None,\n replicate_existing_snaps=None, remote_system=None):\n return UnityReplicationSession.cre... |
def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None,
remote_system=None, replication_name=None, dst_size=None, dst_sp=None,
is_dst_thin=None, dst_tiering_policy=None, is_dst_compression=None):
dst_size = self.size_total if dst_size is None else dst_size
dst_resource = UnityResourceConfig.to_embedded(name=dst_lun_name,
pool_id=dst_pool_id, size=dst_size, default_sp=dst_sp,
tiering_policy=dst_tiering_policy, is_thin_enabled=is_dst_thin,
is_compression_enabled=is_dst_compression)
return UnityReplicationSession.create_with_dst_resource_provisioning(self
._cli, self.get_id(), dst_resource, max_time_out_of_sync,
remote_system=remote_system, name=replication_name)
| replicate_with_dst_resource_provisioning | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None,\n remote_system=None, replication_name=None, dst_size=None, dst_sp=None,\n is_dst_thin=None, ... |
def x(self):
obj = None
if self.is_link_aggregation():
obj = UnityLinkAggregation.get(self._cli, self.get_id())
else:
obj = UnityEthernetPort.get(self._cli, self.get_id())
return obj
| get_physical_port | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n obj = None\n if self.is_link_aggregation():\n obj = UnityLinkAggregation.get(self._cli, self.get_id())\n else:\n obj = UnityEth... |
def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None,
is_compression_enabled=None, is_backup_only=None, size=None,
tiering_policy=None, request_id=None, src_id=None, name=None,
default_sp=None, replication_resource_type=None):
return {'poolId': pool_id, 'isThinEnabled': is_thin_enabled,
'isDeduplicationEnabled': is_deduplication_enabled,
'isCompressionEnabled': is_compression_enabled, 'isBackupOnly':
is_backup_only, 'size': size, 'tieringPolicy': tiering_policy,
'requestId': request_id, 'srcId': src_id, 'name': name, 'defaultSP':
default_sp, 'replicationResourceType': replication_resource_type}
| to_embedded | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None,\n is_compression_enabled=None, is_backup_only=None, size=None,\n tiering_policy=None, ... |
def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync,
name=None, members=None, auto_initiate=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
replicate_existing_snaps=None, remote_system=None, src_spa_interface=
None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface
=None):
req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceId=
dst_resource_id, maxTimeOutOfSync=max_time_out_of_sync, members=
members, autoInitiate=auto_initiate, name=name,
hourlySnapReplicationPolicy=hourly_snap_replication_policy,
dailySnapReplicationPolicy=daily_snap_replication_policy,
replicateExistingSnaps=replicate_existing_snaps, remoteSystem=
remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface)
resp = cli.post(cls().resource_class, **req_body)
resp.raise_if_err()
return cls.get(cli, resp.resource_id)
| create | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync,\n name=None, members=None, auto_initiate=None,\n hourly_snap_replication_policy=None, d... |
def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync,
name=None, remote_system=None, src_spa_interface=None,
src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None,
dst_resource_element_configs=None, auto_initiate=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
replicate_existing_snaps=None):
req_body = cli.make_body(srcResourceId=src_resource_id,
dstResourceConfig=dst_resource_config, maxTimeOutOfSync=
max_time_out_of_sync, name=name, remoteSystem=remote_system,
srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface, dstResourceElementConfigs=
dst_resource_element_configs, autoInitiate=auto_initiate,
hourlySnapReplicationPolicy=hourly_snap_replication_policy,
dailySnapReplicationPolicy=daily_snap_replication_policy,
replicateExistingSnaps=replicate_existing_snaps)
resp = cli.type_action(cls().resource_class,
'createReplicationSessionWDestResProvisioning', **req_body)
resp.raise_if_err()
session_resp = resp.first_content['id']
return cls.get(cli, _id=session_resp['id'])
| create_with_dst_resource_provisioning | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync,\n name=None, remote_system=None, src_spa_interface=None,\n src_spb_interface=None, ... |
def x(self, max_time_out_of_sync=None, name=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None,
dst_spb_interface=None):
req_body = self._cli.make_body(maxTimeOutOfSync=max_time_out_of_sync,
name=name, hourlySnapReplicationPolicy=
hourly_snap_replication_policy, dailySnapReplicationPolicy=
daily_snap_replication_policy, srcSPAInterface=src_spa_interface,
srcSPBInterface=src_spb_interface, dstSPAInterface=
dst_spa_interface, dstSPBInterface=dst_spb_interface)
resp = self.action('modify', **req_body)
resp.raise_if_err()
return resp
| modify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_time_out_of_sync=None, name=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n src_spa_interface=None, src_spb_... |
def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface
=None, dst_spa_interface=None, dst_spb_interface=None):
req_body = self._cli.make_body(forceFullCopy=force_full_copy,
srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface)
resp = self.action('resume', **req_body)
resp.raise_if_err()
return resp
| resume | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface\n =None, dst_spa_interface=None, dst_spb_interface=None):\n req_body = self._cli.ma... |
def x(self, sync=None, force=None):
req_body = self._cli.make_body(sync=sync, force=force)
resp = self.action('failover', **req_body)
resp.raise_if_err()
return resp
| failover | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sync=None, force=None):\n req_body = self._cli.make_body(sync=sync, force=force)\n resp = self.action('failover', **req_body)\n resp.raise_if_er... |
def x(self, force_full_copy=None):
req_body = self._cli.make_body(forceFullCopy=force_full_copy)
resp = self.action('failback', **req_body)
resp.raise_if_err()
return resp
| failback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force_full_copy=None):\n req_body = self._cli.make_body(forceFullCopy=force_full_copy)\n resp = self.action('failback', **req_body)\n resp.raise... |
def x(self):
if len(self._orbit_breaks) == 0:
self._detBreaks()
self._fullDayData = self.sat.data.copy()
self._current = 0
| calcOrbits | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if len(self._orbit_breaks) == 0:\n self._detBreaks()\n self._fullDayData = self.sat.data.copy()\n self._current = 0\n",
"r... |
def x(self, orbit_index_period=24.0):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not exist in ' +
'loaded data')
lt_diff = self.sat[self.orbit_index].diff()
ut_vals = Series(self.sat.data.index)
ut_diff = ut_vals.diff()
ind, = np.where(lt_diff < -0.1)
if len(ind) > 0:
ind = np.hstack((ind, np.array([len(self.sat[self.orbit_index])])))
dist = ind[1:] - ind[0:-1]
if len(ind) > 1:
if min(dist) == 1:
print('There are orbit breaks right next to each other')
ind = ind[:-1][dist > 1]
new_ind = []
for idx in ind:
tidx, = np.where(lt_diff[idx - 5:idx + 6] > 0.1)
if len(tidx) != 0:
for tidx in tidx:
if ut_diff[idx - 5:idx + 6].iloc[tidx] < lt_diff[idx -
5:idx + 6].iloc[tidx
] / orbit_index_period * self.orbit_period:
pass
else:
new_ind.append(idx)
break
else:
new_ind.append(idx)
ind = np.array(new_ind)
ut_change_vs_period = ut_diff > self.orbit_period
norm_ut = ut_diff / self.orbit_period
norm_ut_vs_norm_lt = norm_ut.gt(np.abs(lt_diff.values / orbit_index_period)
)
ut_ind, = np.where(ut_change_vs_period | norm_ut_vs_norm_lt & (norm_ut >
0.95))
if len(ut_ind) > 0:
ind = np.hstack((ind, ut_ind))
ind = np.sort(ind)
ind = np.unique(ind)
print('Time Gap')
orbit_ut_diff = ut_vals[ind].diff()
orbit_lt_diff = self.sat[self.orbit_index][ind].diff()
idx, = np.where(orbit_ut_diff / self.orbit_period - orbit_lt_diff.
values / orbit_index_period > 0.97)
if len(idx) > 0:
if idx[0] != 0:
idx = np.hstack((0, idx))
else:
idx = np.array([0])
if len(ind) > 0:
ind = ind[idx]
if ind[0] != 0:
ind = np.hstack((np.array([0]), ind))
else:
ind = np.array([0])
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
| equaBreaks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit_index_period=24.0):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instru... |
def x(self):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not appear to ' +
'exist in loaded data')
pos = self.sat[self.orbit_index] >= 0
npos = -pos
change = pos.values[:-1] & npos.values[1:] | npos.values[:-1] & pos.values[
1:]
ind, = np.where(change)
ind += 1
ut_diff = Series(self.sat.data.index).diff()
ut_ind, = np.where(ut_diff / self.orbit_period > 0.95)
if len(ut_ind) > 0:
ind = np.hstack((ind, ut_ind))
ind = np.sort(ind)
ind = np.unique(ind)
if ind[0] != 0:
ind = np.hstack((np.array([0]), ind))
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
| polarBreaks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation... |
def x(self):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not appear to ' +
'exist in loaded data')
uniq_vals = self.sat[self.orbit_index].unique()
orbit_index = []
for val in uniq_vals:
idx, = np.where(val == self.sat[self.orbit_index].values)
orbit_index.append(idx[0])
if orbit_index[0] != 0:
ind = np.hstack((np.array([0]), orbit_index))
else:
ind = orbit_index
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
| orbitNumberBreaks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation... |
def x(self, orbit=None):
if not self.sat.empty:
self._calcOrbits()
if orbit is not None:
if orbit == -1:
self.sat.data = self._fullDayData[self._orbit_breaks[self.
num + orbit]:]
self._current = self.num + orbit + 1
elif (orbit < 0) & (orbit >= -self.num):
self.sat.data = self._fullDayData[self._orbit_breaks[self.
num + orbit]:self._orbit_breaks[self.num + orbit + 1]]
self._current = self.num + orbit + 1
elif (orbit < self.num) & (orbit != 0):
self.sat.data = self._fullDayData[self._orbit_breaks[orbit -
1]:self._orbit_breaks[orbit]]
self._current = orbit
elif orbit == self.num:
self.sat.data = self._fullDayData[self._orbit_breaks[orbit -
1]:]
self._current = orbit
elif orbit == 0:
raise ValueError('Orbits internally indexed by 1, 0 not ' +
'allowed')
else:
self.sat.data = []
raise ValueError('Requested an orbit past total orbits ' +
'for day')
else:
raise ValueError('Must set an orbit')
| getBasicOrbit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit == -1:\n self.sa... |
def x(self, orbit=None):
if not self.sat.empty:
self._calcOrbits()
if orbit is not None:
if orbit < 0:
orbit = self.num + 1 + orbit
if orbit == 1:
try:
true_date = self.sat.date
self.sat.prev()
if not self.sat.empty:
self.load(orbit=-1)
else:
self.sat.next()
self._getBasicOrbit(orbit=1)
delta = true_date - self.sat.data.index[0]
if delta >= self.orbit_period:
self.next()
except StopIteration:
self._getBasicOrbit(orbit=1)
print('Loaded Orbit:%i' % (self._current - 1))
elif orbit == self.num:
if self.num != 1:
self._getBasicOrbit(self.num - 1)
self.next()
else:
self._getBasicOrbit(orbit=-1)
elif orbit < self.num:
self._getBasicOrbit(orbit)
print('Loaded Orbit:%i' % (self._current - 1))
else:
self.sat.data = DataFrame()
raise Exception('Requested an orbit past total orbits for day')
else:
raise Exception('Must set an orbit')
else:
print('No data loaded in instrument object to determine orbits.')
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit < 0:\n orbit = s... |
def x(self, *arg, **kwarg):
if not self.sat.empty:
self._calcOrbits()
if self._current == self.num - 1:
self._getBasicOrbit(orbit=-1)
load_next = True
if self.sat._iter_type == 'date':
delta = self.sat.date - self.sat.data.index[-1
] + pds.Timedelta('1 day')
if delta >= self.orbit_period:
load_next = False
if load_next:
temp_orbit_data = self.sat.data.copy()
try:
self.sat.next()
if not self.sat.empty:
self.sat.data = pds.concat([temp_orbit_data[:self.
sat.data.index[0] - pds.DateOffset(microseconds
=1)], self.sat.data])
self._getBasicOrbit(orbit=1)
else:
self.sat.prev()
self._getBasicOrbit(orbit=-1)
except StopIteration:
pass
del temp_orbit_data
print('Loaded Orbit:%i' % (self._current - 1))
elif self._current == self.num:
temp_orbit_data = self.sat.data.copy()
self.sat.next()
if not self.sat.empty:
pad_next = True
if self.sat._iter_type == 'date':
delta = self.sat.date - temp_orbit_data.index[-1]
if delta >= self.orbit_period:
pad_next = False
if pad_next:
self.sat.data = pds.concat([temp_orbit_data[:self.sat.
data.index[0] - pds.DateOffset(microseconds=1)],
self.sat.data])
self._getBasicOrbit(orbit=2)
else:
self._getBasicOrbit(orbit=1)
if self.sat._iter_type == 'date':
delta = self.sat.date + pds.DateOffset(days=1
) - self.sat.data.index[0]
if delta < self.orbit_period:
self._current = self.num - 1
self.next()
else:
while self.sat.empty:
self.sat.next()
self._getBasicOrbit(orbit=1)
del temp_orbit_data
print('Loaded Orbit:%i' % (self._current - 1))
elif self._current == 0:
self.load(orbit=1)
elif self._current < self.num - 1:
self._getBasicOrbit(orbit=self._current + 1)
print('Loaded Orbit:%i' % (self._current - 1))
else:
raise Exception('You ended up where nobody should ever be. ' +
'Talk to someone about this fundamental ' + 'failure.')
else:
while self.sat.empty:
self.sat.next()
self.next()
| next | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *arg, **kwarg):\n if not self.sat.empty:\n self._calcOrbits()\n if self._current == self.num - 1:\n self._getBasicOrbit(orbit... |
def x(fnames, tag=None, sat_id=None):
data = pds.DataFrame()
for filename in fnames:
fname = filename[0:-11]
with open(fname) as f:
lines = f.readlines()
idx = 0
max_lines = 0
for line in lines:
if len(line) > 1:
max_lines += 1
yr = np.zeros(max_lines * 24, dtype=int)
mo = np.zeros(max_lines * 24, dtype=int)
day = np.zeros(max_lines * 24, dtype=int)
ut = np.zeros(max_lines * 24, dtype=int)
dst = np.zeros(max_lines * 24, dtype=int)
for line in lines:
if len(line) > 1:
temp_year = int(line[14:16] + line[3:5])
if temp_year > 57:
temp_year += 1900
else:
temp_year += 2000
yr[idx:idx + 24] = temp_year
mo[idx:idx + 24] = int(line[5:7])
day[idx:idx + 24] = int(line[8:10])
ut[idx:idx + 24] = np.arange(24)
temp = line.strip()[20:-4]
temp2 = [temp[4 * i:4 * (i + 1)] for i in np.arange(24)]
dst[idx:idx + 24] = temp2
idx += 24
start = pds.datetime(yr[0], mo[0], day[0], ut[0])
stop = pds.datetime(yr[-1], mo[-1], day[-1], ut[-1])
dates = pds.date_range(start, stop, freq='H')
new_data = pds.DataFrame(dst, index=dates, columns=['dst'])
new_date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')
idx, = np.where((new_data.index >= new_date) & (new_data.index <
new_date + pds.DateOffset(days=1)))
new_data = new_data.iloc[idx, :]
data = pds.concat([data, new_data], axis=0)
return data, pysat.Meta()
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n for filename in fnames:\n fname = filename[0:-11]\n with open(fname) as f:\n ... |
def x(yr, mo, day):
yr = '20' + yr
yr = int(yr)
mo = int(mo)
day = int(day)
return pds.datetime(yr, mo, day)
| parse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(yr, mo, day):\n yr = '20' + yr\n yr = int(yr)\n mo = int(mo)\n day = int(day)\n return pds.datetime(yr, mo, day)\n",
"role": "user"
},
{... |
def x(fnames, tag=None, sat_id=None):
data = pds.DataFrame()
colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19
), (19, 23), (23, 26), (26, 29), (29, 32), (32, 50)]
for filename in fnames:
fname = filename[0:-11]
date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')
temp = pds.read_fwf(fname, colspecs=colspec, skipfooter=4, header=
None, parse_dates=[[0, 1, 2]], date_parser=_parse, index_col=
'0_1_2')
idx, = np.where((temp.index >= date) & (temp.index < date + pds.
DateOffset(days=1)))
temp = temp.iloc[idx, :]
data = pds.concat([data, temp], axis=0)
data = data.iloc[:, 0:-1]
s = pds.Series()
for i in np.arange(8):
temp = pds.Series(data.iloc[:, i].values, index=data.index + pds.
DateOffset(hours=int(3 * i)))
s = s.append(temp)
s = s.sort_index()
s.index.name = 'time'
first = np.array([float(x[0]) for x in s])
flag = np.array([x[1] for x in s])
ind, = np.where(flag == '+')
first[ind] += 1.0 / 3.0
ind, = np.where(flag == '-')
first[ind] -= 1.0 / 3.0
result = pds.DataFrame(first, columns=['kp'], index=s.index)
return result, pysat.Meta()
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19\n ), (19, 23), ... |
def x(date_array, tag, sat_id, data_path, user=None, password=None):
import ftplib
from ftplib import FTP
import sys
ftp = FTP('ftp.gfz-potsdam.de')
ftp.login()
ftp.cwd('/pub/home/obs/kp-ap/tab')
for date in date_array:
fname = 'kp{year:02d}{month:02d}.tab'
fname = fname.format(year=date.year - date.year // 100 * 100, month
=date.month)
local_fname = fname
saved_fname = os.path.join(data_path, local_fname)
try:
print('Downloading file for ' + date.strftime('%D'))
sys.stdout.flush()
ftp.retrbinary('RETR ' + fname, open(saved_fname, 'wb').write)
except ftplib.error_perm as exception:
if str(exception.args[0]).split(' ', 1)[0] != '550':
raise
else:
os.remove(saved_fname)
print('File not available for ' + date.strftime('%D'))
ftp.close()
return
| download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(date_array, tag, sat_id, data_path, user=None, password=None):\n import ftplib\n from ftplib import FTP\n import sys\n ftp = FTP('ftp.gfz-potsdam.d... |
def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None):
if kp_inst is not None:
kp_inst.load(date=sat.date, verifyPad=True)
kpData = kp_inst
elif kpData is None:
kp = pysat.Instrument('sw', 'kp', pad=pds.DateOffset(days=1))
kp.load(date=sat.date, verifyPad=True)
kpData = kp
if maxKp is None:
maxKp = 3 + 1.0 / 3.0
if filterTime is None:
filterTime = 24
date = sat.date
selData = kpData[date - pds.DateOffset(days=1):date + pds.DateOffset(
days=1)]
ind, = np.where(selData['kp'] >= maxKp)
for lind in ind:
sat.data[selData.index[lind]:selData.index[lind] + pds.DateOffset(
hours=filterTime)] = np.NaN
sat.data = sat.data.dropna(axis=0, how='all')
return
| filter_geoquiet | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None):\n if kp_inst is not None:\n kp_inst.load(date=sat.date, verifyPad=True)\n kpData... |
def x(self, converter_str):
ret = None
if converter_str is not None:
converter_desc_list = converter_str.split('.')
if len(converter_desc_list) == 1:
converter = converter_desc_list[0]
ret = getattr(cvt, converter, None)
if ret is None:
ret = self.get_converter(converter)
if ret is None:
ret = self.get_resource_clz_by_name(converter)
if ret is None:
ret = self.get_enum_by_name(converter)
if ret is None:
ret = self.get(converter)
if ret is None and converter_str is not None:
raise ValueError('Specified converter not supported: {}'.format
(converter_str))
return ret
| get_converter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, converter_str):\n ret = None\n if converter_str is not None:\n converter_desc_list = converter_str.split('.')\n if len(converter_desc... |
def x(self, local_path, remote_path):
sftp_client = self.transport.open_sftp_client()
LOG.debug(
'Copy the local file to remote. Source=%(src)s. Target=%(target)s.' %
{'src': local_path, 'target': remote_path})
try:
sftp_client.put(local_path, remote_path)
except Exception as ex:
LOG.error('Failed to copy the local file to remote. Reason: %s.' %
six.text_type(ex))
raise SFtpExecutionError(err=ex)
| copy_file_to_remote | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, local_path, remote_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug(\n 'Copy the local file to remote. Source=%(src)s. T... |
def x(self, remote_path, local_path):
sftp_client = self.transport.open_sftp_client()
LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' % {
'src': remote_path, 'target': local_path})
try:
sftp_client.get(remote_path, local_path)
except Exception as ex:
LOG.error('Failed to secure copy. Reason: %s.' % six.text_type(ex))
raise SFtpExecutionError(err=ex)
| get_remote_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, remote_path, local_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' %... |
def x(self):
if 'isLive' in self.__dict__ and self.isLive:
self.transport.close()
self.isLive = False
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if 'isLive' in self.__dict__ and self.isLive:\n self.transport.close()\n self.isLive = False\n",
"role": "user"
},
{
"con... |
def x(check_object=False, check_invalid_data_mover=False):
def x(f):
@functools.wraps(f)
def x(self, *argv, **kwargs):
request = f(self, *argv, **kwargs)
return self.request(request, check_object=check_object,
check_invalid_data_mover=check_invalid_data_mover)
return func_wrapper
return decorator
| xml_request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(check_object=False, check_invalid_data_mover=False):\n\n def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n reques... |
def x(f):
@functools.wraps(f)
def x(self, *argv, **kwargs):
commands = f(self, *argv, **kwargs)
return self.ssh_execute(['env', 'NAS_DB=/nas'] + commands)
return func_wrapper
| nas_command | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n commands = f(self, *argv, **kwargs)\n return self.ssh_execute(['env', 'NAS_DB... |
def x(self, backup=None, delete_backup=False):
resp = self._cli.action(self.resource_class, self.get_id(), 'restore',
copyName=backup)
resp.raise_if_err()
backup = resp.first_content['backup']
backup_snap = UnitySnap(_id=backup['id'], cli=self._cli)
if delete_backup:
log.info('Deleting the backup snap {} as the restoration succeeded.'
.format(backup['id']))
backup_snap.delete()
return backup_snap
| restore | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, backup=None, delete_backup=False):\n resp = self._cli.action(self.resource_class, self.get_id(), 'restore',\n copyName=backup)\n resp.raise_... |
def x(self, name, io_limit_policy=None, description=None):
if self.is_member_snap():
raise UnityCGMemberActionNotSupportError()
if self.lun and not self.lun.is_thin_enabled:
raise UnityThinCloneNotAllowedError()
return TCHelper.thin_clone(self._cli, self, name, io_limit_policy,
description)
| thin_clone | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, io_limit_policy=None, description=None):\n if self.is_member_snap():\n raise UnityCGMemberActionNotSupportError()\n if self.lun and no... |
def x(self, async_mode=False, even_attached=False):
try:
return super(UnitySnap, self).delete(async_mode=async_mode)
except UnityDeleteAttachedSnapError:
if even_attached:
log.debug(
'Force delete the snapshot even if it is attached. First detach the snapshot from hosts, then delete again.'
)
self.detach_from(None)
return super(UnitySnap, self).delete(async_mode=async_mode)
else:
raise
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, async_mode=False, even_attached=False):\n try:\n return super(UnitySnap, self).delete(async_mode=async_mode)\n except UnityDeleteAttachedSna... |
def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None):
fields = self.get_fields(type_name, base_fields, nested_fields)
the_filter = self.dict_to_filter_string(the_filter)
url = '/api/types/{}/instances'.format(type_name)
resp = self.rest_get(url, fields=fields, filter=the_filter)
ret = resp
while resp.has_next_page:
resp = self.rest_get(url, fields=fields, filter=the_filter, page=
resp.next_page)
ret.entries.extend(resp.entries)
return ret
| get_all | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None):\n fields = self.get_fields(type_name, base_fields, nested_fields)\n the_filter =... |
def x(self, type_name, obj_id, base_fields=None, nested_fields=None):
base_fields = self.get_fields(type_name, base_fields, nested_fields)
url = '/api/instances/{}/{}'.format(type_name, obj_id)
return self.rest_get(url, fields=base_fields)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_name, obj_id, base_fields=None, nested_fields=None):\n base_fields = self.get_fields(type_name, base_fields, nested_fields)\n url = '/api/inst... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.