text stringlengths 81 112k |
|---|
Used to do a service full check when saving it.
def service_post_save(instance, *args, **kwargs):
"""
Used to do a service full check when saving it.
"""
# check service
if instance.is_monitored and settings.REGISTRY_SKIP_CELERY:
check_service(instance.id)
elif instance.is_monitored:
check_service.delay(instance.id) |
Used to check layer validity.
def layer_pre_save(instance, *args, **kwargs):
"""
Used to check layer validity.
"""
is_valid = True
# we do not need to check validity for WM layers
if not instance.service.type == 'Hypermap:WorldMap':
# 0. a layer is invalid if its service its invalid as well
if not instance.service.is_valid:
is_valid = False
LOGGER.debug('Layer with id %s is marked invalid because its service is invalid' % instance.id)
# 1. a layer is invalid with an extent within (-2, -2, +2, +2)
if instance.bbox_x0 > -2 and instance.bbox_x1 < 2 and instance.bbox_y0 > -2 and instance.bbox_y1 < 2:
is_valid = False
LOGGER.debug(
'Layer with id %s is marked invalid because its extent is within (-2, -2, +2, +2)' % instance.id
)
instance.is_valid = is_valid |
Used to do a layer full check when saving it.
def layer_post_save(instance, *args, **kwargs):
"""
Used to do a layer full check when saving it.
"""
if instance.is_monitored and instance.service.is_monitored: # index and monitor
if not settings.REGISTRY_SKIP_CELERY:
check_layer.delay(instance.id)
else:
check_layer(instance.id)
else: # just index
index_layer(instance.id) |
Used to do reindex layers/services when a issue is removed form them.
def issue_post_delete(instance, *args, **kwargs):
"""
Used to do reindex layers/services when a issue is removed form them.
"""
LOGGER.debug('Re-adding layer/service to search engine index')
if isinstance(instance.content_object, Service):
if not settings.REGISTRY_SKIP_CELERY:
index_service.delay(instance.content_object.id)
else:
index_service(instance.content_object.id)
else:
if not settings.REGISTRY_SKIP_CELERY:
index_layer.delay(instance.content_object.id)
else:
index_layer(instance.content_object.id) |
When service Realiability is going down users should go to the
the check history to find problem causes.
:return: admin url with check list for this instance
def get_checks_admin_reliability_warning_url(self):
"""
When service Realiability is going down users should go to the
the check history to find problem causes.
:return: admin url with check list for this instance
"""
# TODO: cache this.
path = self.get_checks_admin_url()
content_type = ContentType.objects.get_for_model(self)
params = "?content_type__id__exact={0}&q={1}&success__exact=0".format(
content_type.id,
self.id
)
url = path + params
return url |
Update layers for a service.
def update_layers(self):
"""
Update layers for a service.
"""
signals.post_save.disconnect(layer_post_save, sender=Layer)
try:
LOGGER.debug('Updating layers for service id %s' % self.id)
if self.type == 'OGC:WMS':
update_layers_wms(self)
elif self.type == 'OGC:WMTS':
update_layers_wmts(self)
elif self.type == 'ESRI:ArcGIS:MapServer':
update_layers_esri_mapserver(self)
elif self.type == 'ESRI:ArcGIS:ImageServer':
update_layers_esri_imageserver(self)
elif self.type == 'Hypermap:WorldMapLegacy':
update_layers_wm_legacy(self)
elif self.type == 'Hypermap:WorldMap':
update_layers_geonode_wm(self)
elif self.type == 'Hypermap:WARPER':
update_layers_warper(self)
except:
LOGGER.error('Error updating layers for service %s' % self.uuid)
signals.post_save.connect(layer_post_save, sender=Layer) |
Check for availability of a service and provide run metrics.
def check_available(self):
"""
Check for availability of a service and provide run metrics.
"""
success = True
start_time = datetime.datetime.utcnow()
message = ''
LOGGER.debug('Checking service id %s' % self.id)
try:
title = None
abstract = None
keywords = []
wkt_geometry = None
srs = '4326'
if self.type == 'OGC:CSW':
ows = CatalogueServiceWeb(self.url)
title = ows.identification.title
abstract = ows.identification.abstract
keywords = ows.identification.keywords
if self.type == 'OGC:WMS':
ows = get_wms_version_negotiate(self.url)
title = ows.identification.title
abstract = ows.identification.abstract
keywords = ows.identification.keywords
for c in ows.contents:
if ows.contents[c].parent is None:
wkt_geometry = bbox2wktpolygon(ows.contents[c].boundingBoxWGS84)
break
if self.type == 'OGC:WMTS':
ows = WebMapTileService(self.url)
title = ows.identification.title
abstract = ows.identification.abstract
keywords = ows.identification.keywords
if self.type == 'OSGeo:TMS':
ows = TileMapService(self.url)
title = ows.identification.title
abstract = ows.identification.abstract
keywords = ows.identification.keywords
if self.type == 'ESRI:ArcGIS:MapServer':
esri = ArcMapService(self.url)
extent, srs = get_esri_extent(esri)
title = esri.mapName
if len(title) == 0:
title = get_esri_service_name(self.url)
wkt_geometry = bbox2wktpolygon([
extent['xmin'],
extent['ymin'],
extent['xmax'],
extent['ymax']
])
if self.type == 'ESRI:ArcGIS:ImageServer':
esri = ArcImageService(self.url)
extent, srs = get_esri_extent(esri)
title = esri._json_struct['name']
if len(title) == 0:
title = get_esri_service_name(self.url)
wkt_geometry = bbox2wktpolygon([
extent['xmin'],
extent['ymin'],
extent['xmax'],
extent['ymax']
])
if self.type == 'Hypermap:WorldMap':
urllib2.urlopen(self.url)
if self.type == 'Hypermap:WorldMapLegacy':
urllib2.urlopen(self.url)
title = 'Harvard WorldMap Legacy'
if self.type == 'Hypermap:WARPER':
urllib2.urlopen(self.url)
# update title without raising a signal and recursion
if title:
self.title = title
Service.objects.filter(id=self.id).update(title=title)
if abstract:
self.abstract = abstract
Service.objects.filter(id=self.id).update(abstract=abstract)
if keywords:
for kw in keywords:
# FIXME: persist keywords to Django model
self.keywords.add(kw)
if wkt_geometry:
self.wkt_geometry = wkt_geometry
Service.objects.filter(id=self.id).update(wkt_geometry=wkt_geometry)
xml = create_metadata_record(
identifier=self.id_string,
source=self.url,
links=[[self.type, self.url]],
format=self.type,
type='service',
title=title,
abstract=abstract,
keywords=keywords,
wkt_geometry=self.wkt_geometry,
srs=srs
)
anytexts = gen_anytext(title, abstract, keywords)
Service.objects.filter(id=self.id).update(anytext=anytexts, xml=xml, csw_type='service')
except Exception, e:
LOGGER.error(e, exc_info=True)
message = str(e)
success = False
end_time = datetime.datetime.utcnow()
delta = end_time - start_time
response_time = '%s.%s' % (delta.seconds, delta.microseconds)
check = Check(
content_object=self,
success=success,
response_time=response_time,
message=message
)
check.save()
LOGGER.debug('Service checked in %s seconds, status is %s' % (response_time, success)) |
Update validity of a service.
def update_validity(self):
"""
Update validity of a service.
"""
# WM is always valid
if self.type == 'Hypermap:WorldMap':
return
signals.post_save.disconnect(service_post_save, sender=Service)
try:
# some service now must be considered invalid:
# 0. any service not exposed in SUPPORTED_SRS
# 1. any WMTS service
# 2. all of the NOAA layers
is_valid = True
# 0. any service not exposed in SUPPORTED_SRS
if self.srs.filter(code__in=SUPPORTED_SRS).count() == 0:
LOGGER.debug('Service with id %s is marked invalid because in not exposed in SUPPORTED_SRS' % self.id)
is_valid = False
# 1. any WMTS service
if self.type == 'OGC:WMTS':
LOGGER.debug('Service with id %s is marked invalid because it is of type OGC:WMTS' % self.id)
is_valid = False
# 2. all of the NOAA layers
if 'noaa' in self.url.lower():
LOGGER.debug('Service with id %s is marked invalid because it is from NOAA' % self.id)
is_valid = False
# now we save the service
self.is_valid = is_valid
self.save()
except:
LOGGER.error('Error updating validity of the service!')
signals.post_save.connect(service_post_save, sender=Service) |
resolve the search url no matter if local or remote.
:return: url or exception
def get_search_url(self):
"""
resolve the search url no matter if local or remote.
:return: url or exception
"""
if self.is_remote:
return self.url
return reverse('search_api', args=[self.slug]) |
Returns the Hypermap endpoint for a layer.
This endpoint will be the WMTS MapProxy endpoint, only for WM we use the original endpoint.
def get_url_endpoint(self):
"""
Returns the Hypermap endpoint for a layer.
This endpoint will be the WMTS MapProxy endpoint, only for WM we use the original endpoint.
"""
endpoint = self.url
if self.type not in ('Hypermap:WorldMap',):
endpoint = 'registry/%s/layer/%s/map/wmts/1.0.0/WMTSCapabilities.xml' % (
self.catalog.slug,
self.id
)
return endpoint |
Check for availability of a layer and provide run metrics.
def check_available(self):
"""
Check for availability of a layer and provide run metrics.
"""
success = True
start_time = datetime.datetime.utcnow()
message = ''
LOGGER.debug('Checking layer id %s' % self.id)
signals.post_save.disconnect(layer_post_save, sender=Layer)
try:
self.update_thumbnail()
except ValueError, err:
# caused by update_thumbnail()
# self.href is empty in arcserver.ExportMap
if str(err).startswith("unknown url type:"):
LOGGER.debug('Thumbnail can not be updated: %s' % str(err))
except Exception, err:
message = str(err)
success = False
signals.post_save.connect(layer_post_save, sender=Layer)
end_time = datetime.datetime.utcnow()
delta = end_time - start_time
response_time = '%s.%s' % (delta.seconds, delta.microseconds)
check = Check(
content_object=self,
success=success,
response_time=response_time,
message=message
)
check.save()
LOGGER.debug('Layer checked in %s seconds, status is %s' % (response_time, success))
return success, message |
Get extra metadata tagged with a registry keyword.
For example:
<registry:property name="nomination/serviceOwner" value="True"/>
<registry:property name="nominator/name" value="Random Person"/>
<registry:property name="nominator/email" value="contact@example.com"/>
<registry:property name="lastmodifiedby" value="2016-10-23"/>
<registry:property name="updateFreq" value="as needed"/>
<registry:property name="mission" value="RIO"/>
<registry:property name="authentication" value="Open"/>
<registry:property name="crisis" value="False"/>
<registry:property name="intlAgreement/multi" value="none"/>
<registry:property name="intlAgreement/bilateral" value="none"/>
<registry:property name="classificationRecord/classification" value="Unclassified"/>
<registry:property name="classificationData/classification" value="Unclassified"/>
<registry:property name="serviceName/classification/classification" value="Unclassified"/>
<registry:property name="serviceName/classification/classifiedBy" value="TNT"/>
<registry:property name="description/classification/classification" value="Unclassified"/>
<registry:property name="description/classification/classifiedBy" value="TNT"/>
<registry:property name="ContactInformation/Primary/owner" value="Help Desk"/>
<registry:property name="ContactInformation/Primary/organization" value="Three Letter One"/>
<registry:property name="ContactInformation/Email" value="contact@example.com"/>
<registry:property name="ContactInformation/Phone" value="Toll-free: 1 800 555-5555"/>
<registry:property name="license/restrictions" value="none"/>
<registry:property name="license/copyright" value="Private. For testing purposes."/>
def registry_tags(self, query_string='{http://gis.harvard.edu/HHypermap/registry/0.1}property'):
"""
Get extra metadata tagged with a registry keyword.
For example:
<registry:property name="nomination/serviceOwner" value="True"/>
<registry:property name="nominator/name" value="Random Person"/>
<registry:property name="nominator/email" value="contact@example.com"/>
<registry:property name="lastmodifiedby" value="2016-10-23"/>
<registry:property name="updateFreq" value="as needed"/>
<registry:property name="mission" value="RIO"/>
<registry:property name="authentication" value="Open"/>
<registry:property name="crisis" value="False"/>
<registry:property name="intlAgreement/multi" value="none"/>
<registry:property name="intlAgreement/bilateral" value="none"/>
<registry:property name="classificationRecord/classification" value="Unclassified"/>
<registry:property name="classificationData/classification" value="Unclassified"/>
<registry:property name="serviceName/classification/classification" value="Unclassified"/>
<registry:property name="serviceName/classification/classifiedBy" value="TNT"/>
<registry:property name="description/classification/classification" value="Unclassified"/>
<registry:property name="description/classification/classifiedBy" value="TNT"/>
<registry:property name="ContactInformation/Primary/owner" value="Help Desk"/>
<registry:property name="ContactInformation/Primary/organization" value="Three Letter One"/>
<registry:property name="ContactInformation/Email" value="contact@example.com"/>
<registry:property name="ContactInformation/Phone" value="Toll-free: 1 800 555-5555"/>
<registry:property name="license/restrictions" value="none"/>
<registry:property name="license/copyright" value="Private. For testing purposes."/>
"""
from pycsw.core.etree import etree
parsed = etree.fromstring(self.xml, etree.XMLParser(resolve_entities=False))
registry_tags = parsed.findall(query_string)
registry_dict = {}
for tag in registry_tags:
try:
registry_dict[tag.attrib['name']] = tag.attrib['value']
except Exception, e:
LOGGER.error(e, exc_info=True)
return registry_dict |
Grabs input from the user and saves
it as their trytravis target repo
def _input_github_repo(url=None):
""" Grabs input from the user and saves
it as their trytravis target repo """
if url is None:
url = user_input('Input the URL of the GitHub repository '
'to use as a `trytravis` repository: ')
url = url.strip()
http_match = _HTTPS_REGEX.match(url)
ssh_match = _SSH_REGEX.match(url)
if not http_match and not ssh_match:
raise RuntimeError('That URL doesn\'t look like a valid '
'GitHub URL. We expect something '
'of the form: `https://github.com/[USERNAME]/'
'[REPOSITORY]` or `ssh://git@github.com/'
'[USERNAME]/[REPOSITORY]')
# Make sure that the user actually made a new repository on GitHub.
if http_match:
_, name = http_match.groups()
else:
_, name = ssh_match.groups()
if 'trytravis' not in name:
raise RuntimeError('You must have `trytravis` in the name of your '
'repository. This is a security feature to reduce '
'chances of running git push -f on a repository '
'you don\'t mean to.')
# Make sure that the user actually wants to use this repository.
accept = user_input('Remember that `trytravis` will make commits on your '
'behalf to `%s`. Are you sure you wish to use this '
'repository? Type `y` or `yes` to accept: ' % url)
if accept.lower() not in ['y', 'yes']:
raise RuntimeError('Operation aborted by user.')
if not os.path.isdir(config_dir):
os.makedirs(config_dir)
with open(os.path.join(config_dir, 'repo'), 'w+') as f:
f.truncate()
f.write(url)
print('Repository saved successfully.') |
Loads the GitHub repository from the users config.
def _load_github_repo():
""" Loads the GitHub repository from the users config. """
if 'TRAVIS' in os.environ:
raise RuntimeError('Detected that we are running in Travis. '
'Stopping to prevent infinite loops.')
try:
with open(os.path.join(config_dir, 'repo'), 'r') as f:
return f.read()
except (OSError, IOError):
raise RuntimeError('Could not find your repository. '
'Have you ran `trytravis --repo`?') |
Temporarily commits local changes and submits them to
the GitHub repository that the user has specified. Then
reverts the changes to the git repository if a commit was
necessary.
def _submit_changes_to_github_repo(path, url):
""" Temporarily commits local changes and submits them to
the GitHub repository that the user has specified. Then
reverts the changes to the git repository if a commit was
necessary. """
try:
repo = git.Repo(path)
except Exception:
raise RuntimeError('Couldn\'t locate a repository at `%s`.' % path)
commited = False
try:
try:
repo.delete_remote('trytravis')
except Exception:
pass
print('Adding a temporary remote to '
'`%s`...' % url)
remote = repo.create_remote('trytravis', url)
print('Adding all local changes...')
repo.git.add('--all')
try:
print('Committing local changes...')
timestamp = datetime.datetime.now().isoformat()
repo.git.commit(m='trytravis-' + timestamp)
commited = True
except git.exc.GitCommandError as e:
if 'nothing to commit' in str(e):
commited = False
else:
raise
commit = repo.head.commit.hexsha
committed_at = repo.head.commit.committed_datetime
print('Pushing to `trytravis` remote...')
remote.push(force=True)
finally:
if commited:
print('Reverting to old state...')
repo.git.reset('HEAD^')
try:
repo.delete_remote('trytravis')
except Exception:
pass
return commit, committed_at |
Waits for a Travis build to appear with the given commit SHA
def _wait_for_travis_build(url, commit, committed_at):
""" Waits for a Travis build to appear with the given commit SHA """
print('Waiting for a Travis build to appear '
'for `%s` after `%s`...' % (commit, committed_at))
import requests
slug = _slug_from_url(url)
start_time = time.time()
build_id = None
while time.time() - start_time < 60:
with requests.get('https://api.travis-ci.org/repos/%s/builds' % slug,
headers=_travis_headers()) as r:
if not r.ok:
raise RuntimeError('Could not reach the Travis API '
'endpoint. Additional information: '
'%s' % str(r.content))
# Search through all commits and builds to find our build.
commit_to_sha = {}
json = r.json()
for travis_commit in sorted(json['commits'],
key=lambda x: x['committed_at']):
travis_committed_at = datetime.datetime.strptime(
travis_commit['committed_at'], '%Y-%m-%dT%H:%M:%SZ'
).replace(tzinfo=utc)
if travis_committed_at < committed_at:
continue
commit_to_sha[travis_commit['id']] = travis_commit['sha']
for build in json['builds']:
if (build['commit_id'] in commit_to_sha and
commit_to_sha[build['commit_id']] == commit):
build_id = build['id']
print('Travis build id: `%d`' % build_id)
print('Travis build URL: `https://travis-ci.org/'
'%s/builds/%d`' % (slug, build_id))
if build_id is not None:
break
time.sleep(3.0)
else:
raise RuntimeError('Timed out while waiting for a Travis build '
'to start. Is Travis configured for `%s`?' % url)
return build_id |
Watches and progressively outputs information
about a given Travis build
def _watch_travis_build(build_id):
""" Watches and progressively outputs information
about a given Travis build """
import requests
try:
build_size = None # type: int
running = True
while running:
with requests.get('https://api.travis-ci.org/builds/%d' % build_id,
headers=_travis_headers()) as r:
json = r.json()
if build_size is not None:
if build_size > 1:
sys.stdout.write('\r\x1b[%dA' % build_size)
else:
sys.stdout.write('\r')
build_size = len(json['jobs'])
running = False
current_number = 1
for job in json['jobs']: # pragma: no coverage
color, state, is_running = _travis_job_state(job['state'])
if is_running:
running = True
platform = job['config']['os']
if platform == 'osx':
platform = ' osx '
env = job['config'].get('env', '')
sudo = 's' if job['config'].get('sudo', True) else 'c'
lang = job['config'].get('language', 'generic')
padding = ' ' * (len(str(build_size)) -
len(str(current_number)))
number = str(current_number) + padding
current_number += 1
job_display = '#' + ' '.join([number,
state,
platform,
sudo,
lang,
env])
print(color + job_display + colorama.Style.RESET_ALL)
time.sleep(3.0)
except KeyboardInterrupt:
pass |
Converts a Travis state into a state character, color,
and whether it's still running or a stopped state.
def _travis_job_state(state):
""" Converts a Travis state into a state character, color,
and whether it's still running or a stopped state. """
if state in [None, 'queued', 'created', 'received']:
return colorama.Fore.YELLOW, '*', True
elif state in ['started', 'running']:
return colorama.Fore.LIGHTYELLOW_EX, '*', True
elif state == 'passed':
return colorama.Fore.LIGHTGREEN_EX, 'P', False
elif state == 'failed':
return colorama.Fore.LIGHTRED_EX, 'X', False
elif state == 'errored':
return colorama.Fore.LIGHTRED_EX, '!', False
elif state == 'canceled':
return colorama.Fore.LIGHTBLACK_EX, 'X', False
else:
raise RuntimeError('unknown state: %s' % str(state)) |
Parses a project slug out of either an HTTPS or SSH URL.
def _slug_from_url(url):
""" Parses a project slug out of either an HTTPS or SSH URL. """
http_match = _HTTPS_REGEX.match(url)
ssh_match = _SSH_REGEX.match(url)
if not http_match and not ssh_match:
raise RuntimeError('Could not parse the URL (`%s`) '
'for your repository.' % url)
if http_match:
return '/'.join(http_match.groups())
else:
return '/'.join(ssh_match.groups()) |
Gets the output for `trytravis --version`.
def _version_string():
""" Gets the output for `trytravis --version`. """
platform_system = platform.system()
if platform_system == 'Linux':
os_name, os_version, _ = platform.dist()
else:
os_name = platform_system
os_version = platform.version()
python_version = platform.python_version()
return 'trytravis %s (%s %s, python %s)' % (__version__,
os_name.lower(),
os_version,
python_version) |
Function that acts just like main() except
doesn't catch exceptions.
def _main(argv):
""" Function that acts just like main() except
doesn't catch exceptions. """
repo_input_argv = len(argv) == 2 and argv[0] in ['--repo', '-r', '-R']
# We only support a single argv parameter.
if len(argv) > 1 and not repo_input_argv:
_main(['--help'])
# Parse the command and do the right thing.
if len(argv) == 1 or repo_input_argv:
arg = argv[0]
# Help/usage
if arg in ['-h', '--help', '-H']:
print(_USAGE)
# Version
elif arg in ['-v', '--version', '-V']:
print(_version_string())
# Token
elif arg in ['-r', '--repo', '-R']:
if len(argv) == 2:
url = argv[1]
else:
url = None
_input_github_repo(url)
# No wait
elif arg in ['--no-wait', '-nw']:
url = _load_github_repo()
commit, committed = _submit_changes_to_github_repo(os.getcwd(),
url)
build_id = _wait_for_travis_build(url, commit, committed)
# Help string
else:
_main(['--help'])
# No arguments means we're trying to submit to Travis.
elif len(argv) == 0:
url = _load_github_repo()
commit, committed = _submit_changes_to_github_repo(os.getcwd(), url)
build_id = _wait_for_travis_build(url, commit, committed)
_watch_travis_build(build_id) |
Main entry point when the user runs the `trytravis` command.
def main(argv=None): # pragma: no coverage
""" Main entry point when the user runs the `trytravis` command. """
try:
colorama.init()
if argv is None:
argv = sys.argv[1:]
_main(argv)
except RuntimeError as e:
print(colorama.Fore.RED + 'ERROR: ' +
str(e) + colorama.Style.RESET_ALL)
sys.exit(1)
else:
sys.exit(0) |
pycsw wrapper
def csw_global_dispatch(request, url=None, catalog_id=None):
"""pycsw wrapper"""
if request.user.is_authenticated(): # turn on CSW-T
settings.REGISTRY_PYCSW['manager']['transactions'] = 'true'
env = request.META.copy()
# TODO: remove this workaround
# HH should be able to pass env['wsgi.input'] without hanging
# details at https://github.com/cga-harvard/HHypermap/issues/94
if request.method == 'POST':
from StringIO import StringIO
env['wsgi.input'] = StringIO(request.body)
env.update({'local.app_root': os.path.dirname(__file__),
'REQUEST_URI': request.build_absolute_uri()})
# if this is a catalog based CSW, then update settings
if url is not None:
settings.REGISTRY_PYCSW['server']['url'] = url
if catalog_id is not None:
settings.REGISTRY_PYCSW['repository']['filter'] = 'catalog_id = %d' % catalog_id
csw = server.Csw(settings.REGISTRY_PYCSW, env)
content = csw.dispatch_wsgi()
# pycsw 2.0 has an API break:
# pycsw < 2.0: content = xml_response
# pycsw >= 2.0: content = [http_status_code, content]
# deal with the API break
if isinstance(content, list): # pycsw 2.0+
content = content[1]
response = HttpResponse(content, content_type=csw.contenttype)
# TODO: Fix before 1.0 release. CORS should not be enabled blindly like this.
response['Access-Control-Allow-Origin'] = '*'
return response |
pycsw wrapper for catalogs
def csw_global_dispatch_by_catalog(request, catalog_slug):
"""pycsw wrapper for catalogs"""
catalog = get_object_or_404(Catalog, slug=catalog_slug)
if catalog: # define catalog specific settings
url = settings.SITE_URL.rstrip('/') + request.path.rstrip('/')
return csw_global_dispatch(request, url=url, catalog_id=catalog.id) |
OpenSearch wrapper
def opensearch_dispatch(request):
"""OpenSearch wrapper"""
ctx = {
'shortname': settings.REGISTRY_PYCSW['metadata:main']['identification_title'],
'description': settings.REGISTRY_PYCSW['metadata:main']['identification_abstract'],
'developer': settings.REGISTRY_PYCSW['metadata:main']['contact_name'],
'contact': settings.REGISTRY_PYCSW['metadata:main']['contact_email'],
'attribution': settings.REGISTRY_PYCSW['metadata:main']['provider_name'],
'tags': settings.REGISTRY_PYCSW['metadata:main']['identification_keywords'].replace(',', ' '),
'url': settings.SITE_URL.rstrip('/')
}
return render_to_response('search/opensearch_description.xml', ctx,
content_type='application/opensearchdescription+xml') |
passed a string array
def good_coords(coords):
""" passed a string array """
if (len(coords) != 4):
return False
for coord in coords[0:3]:
try:
num = float(coord)
if (math.isnan(num)):
return False
if (math.isinf(num)):
return False
except ValueError:
return False
return True |
Clear all indexes in the es core
def clear_es():
"""Clear all indexes in the es core"""
# TODO: should receive a catalog slug.
ESHypermap.es.indices.delete(ESHypermap.index_name, ignore=[400, 404])
LOGGER.debug('Elasticsearch: Index cleared') |
Create ES core indices
def create_indices(catalog_slug):
"""Create ES core indices """
# TODO: enable auto_create_index in the ES nodes to make this implicit.
# https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-index_.html#index-creation
# http://support.searchly.com/customer/en/portal/questions/
# 16312889-is-automatic-index-creation-disabled-?new=16312889
mapping = {
"mappings": {
"layer": {
"properties": {
"layer_geoshape": {
"type": "geo_shape",
"tree": "quadtree",
"precision": REGISTRY_MAPPING_PRECISION
}
}
}
}
}
ESHypermap.es.indices.create(catalog_slug, ignore=[400, 404], body=mapping) |
kill WSGI processes that may be running in development
def kill_process(procname, scriptname):
"""kill WSGI processes that may be running in development"""
# from http://stackoverflow.com/a/2940878
import signal
import subprocess
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.decode().splitlines():
if procname in line and scriptname in line:
pid = int(line.split()[1])
info('Stopping %s %s %d' % (procname, scriptname, pid))
os.kill(pid, signal.SIGKILL) |
Populate a fresh installed Hypermap instances with basic services.
def populate_initial_services():
"""
Populate a fresh installed Hypermap instances with basic services.
"""
services_list = (
(
'Harvard WorldMap',
'Harvard WorldMap open source web geospatial platform',
'Hypermap:WorldMap',
'http://worldmap.harvard.edu'
),
(
'NYPL MapWarper',
'The New York Public Library (NYPL) MapWarper web site',
'Hypermap:WARPER',
'http://maps.nypl.org/warper/maps'
),
(
'Map Warper',
'The MapWarper web site developed, hosted and maintained by Tim Waters',
'Hypermap:WARPER',
'http://mapwarper.net/maps'
),
(
'WorldMap Warp',
'The MapWarper instance part of the Harvard WorldMap project',
'Hypermap:WARPER',
'http://warp.worldmap.harvard.edu/maps'
),
(
'WFP GeoNode',
'World Food Programme GeoNode',
'OGC:WMS',
'http://geonode.wfp.org/geoserver/ows?'
),
(
'NASA EARTHDATA',
'NASA EARTHDATA, powered by EOSDIS',
'OGC:WMTS',
'http://map1.vis.earthdata.nasa.gov/wmts-geo/1.0.0/WMTSCapabilities.xml'
),
)
esri_endpoint = 'https://gis.ngdc.noaa.gov/arcgis/rest/services'
LOGGER.debug('*** Importing esri endpoint: %s' % esri_endpoint)
create_services_from_endpoint(esri_endpoint)
for service in services_list:
LOGGER.debug('*** Importing %s' % service[0])
service = Service(
title=service[0],
abstract=service[1],
type=service[2],
url=service[3]
)
service.save() |
https://www.elastic.co/guide/en/elasticsearch/reference/current/_the_search_api.html
:param serializer:
:return:
def elasticsearch(serializer, catalog):
"""
https://www.elastic.co/guide/en/elasticsearch/reference/current/_the_search_api.html
:param serializer:
:return:
"""
search_engine_endpoint = "{0}/{1}/_search".format(SEARCH_URL, catalog.slug)
q_text = serializer.validated_data.get("q_text")
q_time = serializer.validated_data.get("q_time")
q_geo = serializer.validated_data.get("q_geo")
q_user = serializer.validated_data.get("q_user")
d_docs_sort = serializer.validated_data.get("d_docs_sort")
d_docs_limit = int(serializer.validated_data.get("d_docs_limit"))
d_docs_page = int(serializer.validated_data.get("d_docs_page"))
a_text_limit = serializer.validated_data.get("a_text_limit")
a_user_limit = serializer.validated_data.get("a_user_limit")
a_time_gap = serializer.validated_data.get("a_time_gap")
a_time_limit = serializer.validated_data.get("a_time_limit")
original_response = serializer.validated_data.get("original_response")
# Dict for search on Elastic engine
must_array = []
filter_dic = {}
aggs_dic = {}
# get ES version to make the query builder to be backward compatible with
# diffs versions.
# TODO: move this to a proper place. maybe ES client?.
# TODO: cache it to avoid overwhelm ES with this call.
# TODO: ask for ES_VERSION when building queries with an elegant way.
ES_VERSION = 2
response = requests.get(SEARCH_URL)
if response.ok:
# looks ugly but will work on normal ES response for "/".
ES_VERSION = int(response.json()["version"]["number"][0])
# String searching
if q_text:
# Wrapping query string into a query filter.
if ES_VERSION >= 2:
query_string = {
"query_string": {
"query": q_text
}
}
else:
query_string = {
"query": {
"query_string": {
"query": q_text
}
}
}
# add string searching
must_array.append(query_string)
if q_time:
# check if q_time exists
q_time = str(q_time) # check string
shortener = q_time[1:-1]
shortener = shortener.split(" TO ")
gte = shortener[0] # greater than
lte = shortener[1] # less than
layer_date = {}
if gte == '*' and lte != '*':
layer_date["lte"] = lte
range_time = {
"layer_date": layer_date
}
range_time = {"range": range_time}
must_array.append(range_time)
if gte != '*' and lte == '*':
layer_date["gte"] = gte
range_time = {
"layer_date": layer_date
}
range_time = {"range": range_time}
must_array.append(range_time)
if gte != '*' and lte != '*':
layer_date["gte"] = gte
layer_date["lte"] = lte
range_time = {
"layer_date": layer_date
}
range_time = {"range": range_time}
must_array.append(range_time)
# geo_shape searching
if q_geo:
q_geo = str(q_geo)
q_geo = q_geo[1:-1]
Ymin, Xmin = q_geo.split(" TO ")[0].split(",")
Ymax, Xmax = q_geo.split(" TO ")[1].split(",")
geoshape_query = {
"layer_geoshape": {
"shape": {
"type": "envelope",
"coordinates": [[Xmin, Ymax], [Xmax, Ymin]]
},
"relation": "intersects"
}
}
filter_dic["geo_shape"] = geoshape_query
if q_user:
# Using q_user
user_searching = {
"match": {
"layer_originator": q_user
}
}
must_array.append(user_searching)
if ES_VERSION >= 2:
dic_query = {
"query": {
"bool": {
"must": must_array,
"filter": filter_dic
}
}
}
else:
dic_query = {
"query": {
"filtered": {
"filter": {
"bool": {
"must": must_array,
"should": filter_dic
}
}
}
}
}
# Page
if d_docs_limit:
dic_query["size"] = d_docs_limit
if d_docs_page:
dic_query["from"] = d_docs_limit * d_docs_page - d_docs_limit
if d_docs_sort == "score":
dic_query["sort"] = {"_score": {"order": "desc"}}
if d_docs_sort == "time":
dic_query["sort"] = {"layer_date": {"order": "desc"}}
if d_docs_sort == "distance":
if q_geo:
# distance_x = float(((float(Xmin) - float(Xmax)) ** 2.0) ** (0.5))
# distance_y = float(((float(Ymin) - float(Ymax)) ** 2.0) ** (0.5))
msg = ("Sorting by distance is different on ElasticSearch than Solr, because this"
"feature on elastic is unavailable to geo_shape type.ElasticSearch docs said:"
"Due to the complex input structure and index representation of shapes,"
"it is not currently possible to sort shapes or retrieve their fields directly."
"The geo_shape value is only retrievable through the _source field."
" Link: https://www.elastic.co/guide/en/elasticsearch/reference/current/geo-shape.html")
return {"error": {"msg": msg}}
else:
msg = "q_qeo MUST BE NO ZERO if you wanna sort by distance"
return {"error": {"msg": msg}}
if a_text_limit:
# getting most frequently occurring users.
text_limit = {
"terms": {
"field": "abstract",
"size": a_text_limit
}
}
aggs_dic['popular_text'] = text_limit
if a_user_limit:
# getting most frequently occurring users.
users_limit = {
"terms": {
"field": "layer_originator",
"size": a_user_limit
}
}
aggs_dic['popular_users'] = users_limit
if a_time_limit:
# TODO: Work in progress, a_time_limit is incomplete.
# TODO: when times are * it does not work. also a a_time_gap is not required.
if q_time:
if not a_time_gap:
# getting time limit histogram.
time_limt = {
"date_range": {
"field": "layer_date",
"format": "yyyy-MM-dd'T'HH:mm:ssZ",
"ranges": [
{"from": gte, "to": lte}
]
}
}
aggs_dic['range'] = time_limt
else:
pass
else:
msg = "If you want to use a_time_limit feature, q_time MUST BE initialized"
return {"error": {"msg": msg}}
if a_time_gap:
interval = gap_to_elastic(a_time_gap)
time_gap = {
"date_histogram": {
"field": "layer_date",
"format": "yyyy-MM-dd'T'HH:mm:ssZ",
"interval": interval
}
}
aggs_dic['articles_over_time'] = time_gap
# adding aggreations on body query
if aggs_dic:
dic_query['aggs'] = aggs_dic
try:
res = requests.post(search_engine_endpoint, data=json.dumps(dic_query))
except Exception as e:
return 500, {"error": {"msg": str(e)}}
es_response = res.json()
if original_response:
return es_response
data = {}
if 'error' in es_response:
data["error"] = es_response["error"]
return 400, data
data["request_url"] = res.url
data["request_body"] = json.dumps(dic_query)
data["a.matchDocs"] = es_response['hits']['total']
docs = []
# aggreations response: facets searching
if 'aggregations' in es_response:
aggs = es_response['aggregations']
# getting the most frequently occurring users.
if 'popular_users' in aggs:
a_users_list_array = []
users_resp = aggs["popular_users"]["buckets"]
for item in users_resp:
temp = {}
temp['count'] = item['doc_count']
temp['value'] = item['key']
a_users_list_array.append(temp)
data["a.user"] = a_users_list_array
# getting most frequently ocurring words
if 'popular_text' in aggs:
a_text_list_array = []
text_resp = es_response["aggregations"]["popular_text"]["buckets"]
for item in text_resp:
temp = {}
temp['count'] = item['doc_count']
temp['value'] = item['key']
a_text_list_array.append(temp)
data["a.text"] = a_text_list_array
if 'articles_over_time' in aggs:
gap_count = []
a_gap = {}
gap_resp = aggs["articles_over_time"]["buckets"]
start = "*"
end = "*"
if len(gap_resp) > 0:
start = gap_resp[0]['key_as_string'].replace('+0000', 'z')
end = gap_resp[-1]['key_as_string'].replace('+0000', 'z')
a_gap['start'] = start
a_gap['end'] = end
a_gap['gap'] = a_time_gap
for item in gap_resp:
temp = {}
if item['doc_count'] != 0:
temp['count'] = item['doc_count']
temp['value'] = item['key_as_string'].replace('+0000', 'z')
gap_count.append(temp)
a_gap['counts'] = gap_count
data['a.time'] = a_gap
if 'range' in aggs:
# Work in progress
# Pay attention in the following code lines: Make it better!!!!
time_count = []
time_resp = aggs["range"]["buckets"]
a_time = {}
a_time['start'] = gte
a_time['end'] = lte
a_time['gap'] = None
for item in time_resp:
temp = {}
if item['doc_count'] != 0:
temp['count'] = item['doc_count']
temp['value'] = item['key'].replace('+0000', 'z')
time_count.append(temp)
a_time['counts'] = time_count
data['a.time'] = a_time
if not int(d_docs_limit) == 0:
for item in es_response['hits']['hits']:
# data
temp = item['_source']['abstract']
temp = temp.replace(u'\u201c', "\"")
temp = temp.replace(u'\u201d', "\"")
temp = temp.replace('"', "\"")
temp = temp.replace("'", "\'")
temp = temp.replace(u'\u2019', "\'")
item['_source']['abstract'] = temp
docs.append(item['_source'])
data["d.docs"] = docs
return data |
Search on solr endpoint
:param serializer:
:return:
def solr(serializer):
"""
Search on solr endpoint
:param serializer:
:return:
"""
search_engine_endpoint = serializer.validated_data.get("search_engine_endpoint")
q_time = serializer.validated_data.get("q_time")
q_geo = serializer.validated_data.get("q_geo")
q_text = serializer.validated_data.get("q_text")
q_user = serializer.validated_data.get("q_user")
d_docs_limit = serializer.validated_data.get("d_docs_limit")
d_docs_page = serializer.validated_data.get("d_docs_page")
d_docs_sort = serializer.validated_data.get("d_docs_sort")
a_time_limit = serializer.validated_data.get("a_time_limit")
a_time_gap = serializer.validated_data.get("a_time_gap")
a_time_filter = serializer.validated_data.get("a_time_filter")
a_hm_limit = serializer.validated_data.get("a_hm_limit")
a_hm_gridlevel = serializer.validated_data.get("a_hm_gridlevel")
a_hm_filter = serializer.validated_data.get("a_hm_filter")
a_text_limit = serializer.validated_data.get("a_text_limit")
a_user_limit = serializer.validated_data.get("a_user_limit")
original_response = serializer.validated_data.get("original_response")
# query params to be sent via restful solr
params = {
"q": "*:*",
"indent": "on",
"wt": "json",
"rows": d_docs_limit,
"facet": "off",
"facet.field": [],
"debug": "timing"
}
if q_text:
params["q"] = q_text
if d_docs_limit >= 0:
d_docs_page -= 1
d_docs_page = d_docs_limit * d_docs_page
params["start"] = d_docs_page
# query params for filters
filters = []
if q_time:
# TODO: when user sends incomplete dates like 2000, its completed: 2000-(TODAY-MONTH)-(TODAY-DAY)T00:00:00Z
# TODO: "Invalid Date in Date Math String:'[* TO 2000-12-05T00:00:00Z]'"
# Kotlin like: "{!field f=layer_date tag=layer_date}[* TO 2000-12-05T00:00:00Z]"
# then do it simple:
filters.append("{0}:{1}".format(TIME_FILTER_FIELD, q_time))
if q_geo:
filters.append("{0}:{1}".format(GEO_FILTER_FIELD, q_geo))
if q_user:
filters.append("{{!field f={0} tag={0}}}{1}".format(USER_FIELD, q_user))
if filters:
params["fq"] = filters
# query params for ordering
if d_docs_sort == 'score' and q_text:
params["sort"] = 'score desc'
elif d_docs_sort == 'time':
params["sort"] = '{} desc'.format(TIME_SORT_FIELD)
elif d_docs_sort == 'distance':
rectangle = parse_geo_box(q_geo)
params["sort"] = 'geodist() asc'
params["sfield"] = GEO_SORT_FIELD
params["pt"] = '{0},{1}'.format(rectangle.centroid.x, rectangle.centroid.y)
# query params for facets
if a_time_limit > 0:
params["facet"] = 'on'
time_filter = a_time_filter or q_time or None
# traduce * to actual min/max dates.
time_filter = asterisk_to_min_max(TIME_FILTER_FIELD, time_filter, search_engine_endpoint)
# create the range faceting params.
facet_parms = request_time_facet(TIME_FILTER_FIELD, time_filter, a_time_gap, a_time_limit)
params.update(facet_parms)
if a_hm_limit > 0:
params["facet"] = 'on'
hm_facet_params = request_heatmap_facet(GEO_HEATMAP_FIELD, a_hm_filter, a_hm_gridlevel, a_hm_limit)
params.update(hm_facet_params)
if a_text_limit > 0:
params["facet"] = 'on'
params["facet.field"].append(TEXT_FIELD)
params["f.{}.facet.limit".format(TEXT_FIELD)] = a_text_limit
if a_user_limit > 0:
params["facet"] = 'on'
params["facet.field"].append("{{! ex={0}}}{0}".format(USER_FIELD))
params["f.{}.facet.limit".format(USER_FIELD)] = a_user_limit
try:
res = requests.get(
search_engine_endpoint, params=params
)
except Exception as e:
return 500, {"error": {"msg": str(e)}}
print '>', res.url
solr_response = res.json()
solr_response["solr_request"] = res.url
if original_response > 0:
return solr_response
# create the response dict following the swagger model:
data = {}
if 'error' in solr_response:
data["error"] = solr_response["error"]
return 400, data
response = solr_response["response"]
data["a.matchDocs"] = response.get("numFound")
if response.get("docs"):
data["d.docs"] = response.get("docs")
if a_time_limit > 0:
date_facet = solr_response["facet_counts"]["facet_ranges"][TIME_FILTER_FIELD]
counts = []
value_count = iter(date_facet.get("counts"))
for value, count in zip(value_count, value_count):
counts.append({
"value": value,
"count": count
})
a_time = {
"start": date_facet.get("start"),
"end": date_facet.get("end"),
"gap": date_facet.get("gap"),
"counts": counts
}
data["a.time"] = a_time
if a_hm_limit > 0:
hm_facet_raw = solr_response["facet_counts"]["facet_heatmaps"][GEO_HEATMAP_FIELD]
hm_facet = {
'gridLevel': hm_facet_raw[1],
'columns': hm_facet_raw[3],
'rows': hm_facet_raw[5],
'minX': hm_facet_raw[7],
'maxX': hm_facet_raw[9],
'minY': hm_facet_raw[11],
'maxY': hm_facet_raw[13],
'counts_ints2D': hm_facet_raw[15],
'projection': 'EPSG:4326'
}
data["a.hm"] = hm_facet
if a_user_limit > 0:
user_facet = solr_response["facet_counts"]["facet_fields"][USER_FIELD]
counts = []
value_count = iter(user_facet)
for value, count in zip(value_count, value_count):
counts.append({
"value": value,
"count": count
})
data["a.user"] = counts
if a_text_limit > 0:
text_facet = solr_response["facet_counts"]["facet_fields"][TEXT_FIELD]
counts = []
value_count = iter(text_facet)
for value, count in zip(value_count, value_count):
counts.append({
"value": value,
"count": count
})
data["a.text"] = counts
subs = []
for label, values in solr_response["debug"]["timing"].iteritems():
if type(values) is not dict:
continue
subs_data = {"label": label, "subs": []}
for label, values in values.iteritems():
if type(values) is not dict:
subs_data["millis"] = values
continue
subs_data["subs"].append({
"label": label,
"millis": values.get("time")
})
subs.append(subs_data)
timing = {
"label": "requests.get.elapsed",
"millis": res.elapsed,
"subs": [{
"label": "QTime",
"millis": solr_response["responseHeader"].get("QTime"),
"subs": subs
}]
}
data["timing"] = timing
data["request_url"] = res.url
return data |
parse all url get params that contains dots in a representation of
serializer field names, for example: d.docs.limit to d_docs_limit.
that makes compatible an actual API client with django-rest-framework
serializers.
:param request:
:return: QueryDict with parsed get params.
def parse_get_params(request):
"""
parse all url get params that contains dots in a representation of
serializer field names, for example: d.docs.limit to d_docs_limit.
that makes compatible an actual API client with django-rest-framework
serializers.
:param request:
:return: QueryDict with parsed get params.
"""
get = request.GET.copy()
new_get = request.GET.copy()
for key in get.iterkeys():
if key.count(".") > 0:
new_key = key.replace(".", "_")
new_get[new_key] = get.get(key)
del new_get[key]
return new_get |
For testing purpose
def main():
"""For testing purpose"""
tcp_adapter = TcpAdapter("192.168.1.3", name="HASS", activate_source=False)
hdmi_network = HDMINetwork(tcp_adapter)
hdmi_network.start()
while True:
for d in hdmi_network.devices:
_LOGGER.info("Device: %s", d)
time.sleep(7) |
Compute difference in bits between digest1 and digest2
returns -127 to 128; 128 is the same, -127 is different
def compare_hexdigests( digest1, digest2 ):
"""Compute difference in bits between digest1 and digest2
returns -127 to 128; 128 is the same, -127 is different"""
# convert to 32-tuple of unsighed two-byte INTs
digest1 = tuple([int(digest1[i:i+2],16) for i in range(0,63,2)])
digest2 = tuple([int(digest2[i:i+2],16) for i in range(0,63,2)])
bits = 0
for i in range(32):
bits += POPC[255 & digest1[i] ^ digest2[i]]
return 128 - bits |
Get accumulator for a transition n between chars a, b, c.
def tran3(self, a, b, c, n):
"""Get accumulator for a transition n between chars a, b, c."""
return (((TRAN[(a+n)&255]^TRAN[b]*(n+n+1))+TRAN[(c)^TRAN[n]])&255) |
Add data to running digest, increasing the accumulators for 0-8
triplets formed by this char and the previous 0-3 chars.
def update(self, data):
"""Add data to running digest, increasing the accumulators for 0-8
triplets formed by this char and the previous 0-3 chars."""
for character in data:
if PY3:
ch = character
else:
ch = ord(character)
self.count += 1
# incr accumulators for triplets
if self.lastch[1] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[1], 0)] +=1
if self.lastch[2] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[2], 1)] +=1
self.acc[self.tran3(ch, self.lastch[1], self.lastch[2], 2)] +=1
if self.lastch[3] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[3], 3)] +=1
self.acc[self.tran3(ch, self.lastch[1], self.lastch[3], 4)] +=1
self.acc[self.tran3(ch, self.lastch[2], self.lastch[3], 5)] +=1
self.acc[self.tran3(self.lastch[3], self.lastch[0], ch, 6)] +=1
self.acc[self.tran3(self.lastch[3], self.lastch[2], ch, 7)] +=1
# adjust last seen chars
self.lastch = [ch] + self.lastch[:3] |
Get digest of data seen thus far as a list of bytes.
def digest(self):
"""Get digest of data seen thus far as a list of bytes."""
total = 0 # number of triplets seen
if self.count == 3: # 3 chars = 1 triplet
total = 1
elif self.count == 4: # 4 chars = 4 triplets
total = 4
elif self.count > 4: # otherwise 8 triplets/char less
total = 8 * self.count - 28 # 28 'missed' during 'ramp-up'
threshold = total / 256 # threshold for accumulators, using the mean
code = [0]*32 # start with all zero bits
for i in range(256): # for all 256 accumulators
if self.acc[i] > threshold: # if it meets the threshold
code[i >> 3] += 1 << (i&7) # set corresponding digest bit, equivalent to i/8, 2 ** (i % 8)
return code[::-1] |
Update running digest with content of named file.
def from_file(self, filename):
"""Update running digest with content of named file."""
f = open(filename, 'rb')
while True:
data = f.read(10480)
if not data:
break
self.update(data)
f.close() |
Compute difference in bits between own digest and another.
returns -127 to 128; 128 is the same, -127 is different
def compare(self, otherdigest, ishex=False):
"""Compute difference in bits between own digest and another.
returns -127 to 128; 128 is the same, -127 is different"""
bits = 0
myd = self.digest()
if ishex:
# convert to 32-tuple of unsighed two-byte INTs
otherdigest = tuple([int(otherdigest[i:i+2],16) for i in range(0,63,2)])
for i in range(32):
bits += POPC[255 & myd[i] ^ otherdigest[i]]
return 128 - bits |
JD Output function. Does quick pretty printing of a CloudGenix Response body. This function returns a string
instead of directly printing content.
**Parameters:**
- **api_response:** A CloudGenix-attribute extended `requests.Response` object
**Returns:** Pretty-formatted text of the Response body
def jdout(api_response):
"""
JD Output function. Does quick pretty printing of a CloudGenix Response body. This function returns a string
instead of directly printing content.
**Parameters:**
- **api_response:** A CloudGenix-attribute extended `requests.Response` object
**Returns:** Pretty-formatted text of the Response body
"""
try:
# attempt to output the cgx_content. should always be a Dict if it exists.
output = json.dumps(api_response.cgx_content, indent=4)
except (TypeError, ValueError, AttributeError):
# cgx_content did not exist, or was not JSON serializable. Try pretty output the base obj.
try:
output = json.dumps(api_response, indent=4)
except (TypeError, ValueError, AttributeError):
# Same issue, just raw output the passed data. Let any exceptions happen here.
output = api_response
return output |
JD Output Detailed function. Meant for quick DETAILED pretty-printing of CloudGenix Request and Response
objects for troubleshooting. This function returns a string instead of directly printing content.
**Parameters:**
- **api_response:** A CloudGenix-attribute extended `requests.Response` object
- **sensitive:** Boolean, if True will print sensitive content (specifically, authentication cookies/headers).
**Returns:** Pretty-formatted text of the Request, Request Headers, Request body, Response, Response Headers,
and Response Body.
def jdout_detailed(api_response, sensitive=False):
"""
JD Output Detailed function. Meant for quick DETAILED pretty-printing of CloudGenix Request and Response
objects for troubleshooting. This function returns a string instead of directly printing content.
**Parameters:**
- **api_response:** A CloudGenix-attribute extended `requests.Response` object
- **sensitive:** Boolean, if True will print sensitive content (specifically, authentication cookies/headers).
**Returns:** Pretty-formatted text of the Request, Request Headers, Request body, Response, Response Headers,
and Response Body.
"""
try:
# try to be super verbose.
output = "REQUEST: {0} {1}\n".format(api_response.request.method, api_response.request.path_url)
output += "REQUEST HEADERS:\n"
for key, value in api_response.request.headers.items():
# look for sensitive values
if key.lower() in ['cookie'] and not sensitive:
# we need to do some work to watch for the AUTH_TOKEN cookie. Split on cookie separator
cookie_list = value.split('; ')
muted_cookie_list = []
for cookie in cookie_list:
# check if cookie starts with a permutation of AUTH_TOKEN/whitespace.
if cookie.lower().strip().startswith('auth_token='):
# first 11 chars of cookie with whitespace removed + mute string.
newcookie = cookie.strip()[:11] + "\"<SENSITIVE - NOT SHOWN BY DEFAULT>\""
muted_cookie_list.append(newcookie)
else:
muted_cookie_list.append(cookie)
# got list of cookies, muted as needed. recombine.
muted_value = "; ".join(muted_cookie_list)
output += "\t{0}: {1}\n".format(key, muted_value)
elif key.lower() in ['x-auth-token'] and not sensitive:
output += "\t{0}: {1}\n".format(key, "<SENSITIVE - NOT SHOWN BY DEFAULT>")
else:
output += "\t{0}: {1}\n".format(key, value)
# if body not present, output blank.
if not api_response.request.body:
output += "REQUEST BODY:\n{0}\n\n".format({})
else:
try:
# Attempt to load JSON from string to make it look beter.
output += "REQUEST BODY:\n{0}\n\n".format(json.dumps(json.loads(api_response.request.body), indent=4))
except (TypeError, ValueError, AttributeError):
# if pretty call above didn't work, just toss it to jdout to best effort it.
output += "REQUEST BODY:\n{0}\n\n".format(jdout(api_response.request.body))
output += "RESPONSE: {0} {1}\n".format(api_response.status_code, api_response.reason)
output += "RESPONSE HEADERS:\n"
for key, value in api_response.headers.items():
output += "\t{0}: {1}\n".format(key, value)
try:
# look for CGX content first.
output += "RESPONSE DATA:\n{0}".format(json.dumps(api_response.cgx_content, indent=4))
except (TypeError, ValueError, AttributeError):
# look for standard response data.
output += "RESPONSE DATA:\n{0}".format(json.dumps(json.loads(api_response.content), indent=4))
except (TypeError, ValueError, AttributeError, UnicodeDecodeError):
# cgx_content did not exist, or was not JSON serializable. Try pretty output the base obj.
try:
output = json.dumps(api_response, indent=4)
except (TypeError, ValueError, AttributeError):
# Same issue, just raw output the passed data. Let any exceptions happen here.
output = api_response
return output |
Check for a new version of the SDK on API constructor instantiation. If new version found, print
Notification to STDERR.
On failure of this check, fail silently.
**Returns:** No item returned, directly prints notification to `sys.stderr`.
def notify_for_new_version(self):
"""
Check for a new version of the SDK on API constructor instantiation. If new version found, print
Notification to STDERR.
On failure of this check, fail silently.
**Returns:** No item returned, directly prints notification to `sys.stderr`.
"""
# broad exception clause, if this fails for any reason just return.
try:
recommend_update = False
update_check_resp = requests.get(self.update_info_url, timeout=3)
web_version = update_check_resp.json()["info"]["version"]
api_logger.debug("RETRIEVED_VERSION: %s", web_version)
available_version = SDK_BUILD_REGEX.search(web_version).groupdict()
current_version = SDK_BUILD_REGEX.search(self.version).groupdict()
available_major = available_version.get('major')
available_minor = available_version.get('minor')
available_patch = available_version.get('patch')
available_build = available_version.get('build')
current_major = current_version.get('major')
current_minor = current_version.get('minor')
current_patch = current_version.get('patch')
current_build = current_version.get('build')
api_logger.debug("AVAILABLE_VERSION: %s", available_version)
api_logger.debug("CURRENT_VERSION: %s", current_version)
# check for major/minor version differences, do not alert for build differences.
if available_major > current_major:
recommend_update = True
elif available_major >= current_major and available_minor > current_minor:
recommend_update = True
elif available_major >= current_major and available_minor >= current_minor and \
available_patch > current_patch:
recommend_update = True
api_logger.debug("NEED_UPDATE: %s", recommend_update)
# notify.
if recommend_update:
sys.stderr.write("WARNING: CloudGenix Python SDK upgrade available. SDKs are typically deprecated 6 "
"months after release of a new version.\n"
"\tLatest Version: {0}\n"
"\tCurrent Version: {1}\n"
"\tFor more info, see 'https://github.com/cloudgenix/sdk-python'. Additionally, this "
"message can be suppressed by instantiating the API with API(update_check=False).\n\n"
"".format(web_version, self.version))
return
except Exception:
# just return and continue.
return |
Modify ssl verification settings
**Parameters:**
- ssl_verify:
- True: Verify using builtin BYTE_CA_BUNDLE.
- False: No SSL Verification.
- Str: Full path to a x509 PEM CA File or bundle.
**Returns:** Mutates API object in place, no return.
def ssl_verify(self, ssl_verify):
"""
Modify ssl verification settings
**Parameters:**
- ssl_verify:
- True: Verify using builtin BYTE_CA_BUNDLE.
- False: No SSL Verification.
- Str: Full path to a x509 PEM CA File or bundle.
**Returns:** Mutates API object in place, no return.
"""
self.verify = ssl_verify
# if verify true/false, set ca_verify_file appropriately
if isinstance(self.verify, bool):
if self.verify: # True
if os.name == 'nt':
# Windows does not allow tmpfile access w/out close. Close file then delete it when done.
self._ca_verify_file_handle = temp_ca_bundle(delete=False)
self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)
self._ca_verify_file_handle.flush()
self.ca_verify_filename = self._ca_verify_file_handle.name
self._ca_verify_file_handle.close()
# Other (POSIX/Unix/Linux/OSX)
else:
self._ca_verify_file_handle = temp_ca_bundle()
self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)
self._ca_verify_file_handle.flush()
self.ca_verify_filename = self._ca_verify_file_handle.name
# register cleanup function for temp file.
atexit.register(self._cleanup_ca_temp_file)
else: # False
# disable warnings for SSL certs.
urllib3.disable_warnings()
self.ca_verify_filename = False
else: # Not True/False, assume path to file/dir for Requests
self.ca_verify_filename = self.verify
return |
Modify retry parameters for the SDK's rest call object.
Parameters are directly from and passed directly to `urllib3.util.retry.Retry`, and get applied directly to
the underlying `requests.Session` object.
Default retry with total=8 and backoff_factor=0.705883:
- Try 1, 0 delay (0 total seconds)
- Try 2, 0 delay (0 total seconds)
- Try 3, 0.705883 delay (0.705883 total seconds)
- Try 4, 1.411766 delay (2.117649 total seconds)
- Try 5, 2.823532 delay (4.941181 total seconds)
- Try 6, 5.647064 delay (10.588245 total seconds)
- Try 7, 11.294128 delay (21.882373 total seconds)
- Try 8, 22.588256 delay (44.470629 total seconds)
- Try 9, 45.176512 delay (89.647141 total seconds)
- Try 10, 90.353024 delay (180.000165 total seconds)
**Parameters:**
- **total:** int, Total number of retries to allow. Takes precedence over other counts.
- **connect:** int, How many connection-related errors to retry on.
- **read:** int, How many times to retry on read errors.
- **redirect:** int, How many redirects to perform. loops.
- **status:** int, How many times to retry on bad status codes.
- **method_whitelist:** iterable, Set of uppercased HTTP method verbs that we should retry on.
- **status_forcelist:** iterable, A set of integer HTTP status codes that we should force a retry on.
- **backoff_factor:** float, A backoff factor to apply between attempts after the second try.
- **raise_on_redirect:** bool, True = raise a MaxRetryError, False = return latest 3xx response.
- **raise_on_status:** bool, Similar logic to ``raise_on_redirect`` but for status responses.
- **respect_retry_after_header:** bool, Whether to respect Retry-After header on status codes.
- **adapter_url:** string, URL match for these retry values (default `https://`)
**Returns:** No return, mutates the session directly
def modify_rest_retry(self, total=8, connect=None, read=None, redirect=None, status=None,
method_whitelist=urllib3.util.retry.Retry.DEFAULT_METHOD_WHITELIST, status_forcelist=None,
backoff_factor=0.705883, raise_on_redirect=True, raise_on_status=True,
respect_retry_after_header=True, adapter_url="https://"):
"""
Modify retry parameters for the SDK's rest call object.
Parameters are directly from and passed directly to `urllib3.util.retry.Retry`, and get applied directly to
the underlying `requests.Session` object.
Default retry with total=8 and backoff_factor=0.705883:
- Try 1, 0 delay (0 total seconds)
- Try 2, 0 delay (0 total seconds)
- Try 3, 0.705883 delay (0.705883 total seconds)
- Try 4, 1.411766 delay (2.117649 total seconds)
- Try 5, 2.823532 delay (4.941181 total seconds)
- Try 6, 5.647064 delay (10.588245 total seconds)
- Try 7, 11.294128 delay (21.882373 total seconds)
- Try 8, 22.588256 delay (44.470629 total seconds)
- Try 9, 45.176512 delay (89.647141 total seconds)
- Try 10, 90.353024 delay (180.000165 total seconds)
**Parameters:**
- **total:** int, Total number of retries to allow. Takes precedence over other counts.
- **connect:** int, How many connection-related errors to retry on.
- **read:** int, How many times to retry on read errors.
- **redirect:** int, How many redirects to perform. loops.
- **status:** int, How many times to retry on bad status codes.
- **method_whitelist:** iterable, Set of uppercased HTTP method verbs that we should retry on.
- **status_forcelist:** iterable, A set of integer HTTP status codes that we should force a retry on.
- **backoff_factor:** float, A backoff factor to apply between attempts after the second try.
- **raise_on_redirect:** bool, True = raise a MaxRetryError, False = return latest 3xx response.
- **raise_on_status:** bool, Similar logic to ``raise_on_redirect`` but for status responses.
- **respect_retry_after_header:** bool, Whether to respect Retry-After header on status codes.
- **adapter_url:** string, URL match for these retry values (default `https://`)
**Returns:** No return, mutates the session directly
"""
# Cloudgenix responses with 502/504 are usually recoverable. Use them if no list specified.
if status_forcelist is None:
status_forcelist = (413, 429, 502, 503, 504)
retry = urllib3.util.retry.Retry(total=total,
connect=connect,
read=read,
redirect=redirect,
status=status,
method_whitelist=method_whitelist,
status_forcelist=status_forcelist,
backoff_factor=backoff_factor,
raise_on_redirect=raise_on_redirect,
raise_on_status=raise_on_status,
respect_retry_after_header=respect_retry_after_header)
adapter = requests.adapters.HTTPAdapter(max_retries=retry)
self._session.mount(adapter_url, adapter)
return |
View current rest retry settings in the `requests.Session()` object
**Parameters:**
- **url:** URL to use to determine retry methods for. Defaults to 'https://'
**Returns:** Dict, Key header, value is header value.
def view_rest_retry(self, url=None):
"""
View current rest retry settings in the `requests.Session()` object
**Parameters:**
- **url:** URL to use to determine retry methods for. Defaults to 'https://'
**Returns:** Dict, Key header, value is header value.
"""
if url is None:
url = "https://"
return vars(self._session.get_adapter(url).max_retries) |
View current cookies in the `requests.Session()` object
**Returns:** List of Dicts, one cookie per Dict.
def view_cookies(self):
"""
View current cookies in the `requests.Session()` object
**Returns:** List of Dicts, one cookie per Dict.
"""
return_list = []
for cookie in self._session.cookies:
return_list.append(vars(cookie))
return return_list |
Change the debug level of the API
**Returns:** No item returned.
def set_debug(self, debuglevel):
"""
Change the debug level of the API
**Returns:** No item returned.
"""
if isinstance(debuglevel, int):
self._debuglevel = debuglevel
if self._debuglevel == 1:
logging.basicConfig(level=logging.INFO,
format="%(levelname)s [%(name)s.%(funcName)s:%(lineno)d] %(message)s")
api_logger.setLevel(logging.INFO)
elif self._debuglevel == 2:
logging.basicConfig(level=logging.DEBUG,
format="%(levelname)s [%(name)s.%(funcName)s:%(lineno)d] %(message)s")
requests.cookies.cookielib.debug = True
api_logger.setLevel(logging.DEBUG)
elif self._debuglevel >= 3:
logging.basicConfig(level=logging.DEBUG,
format="%(levelname)s [%(name)s.%(funcName)s:%(lineno)d] %(message)s")
requests.cookies.cookielib.debug = True
api_logger.setLevel(logging.DEBUG)
urllib3_logger = logging.getLogger("requests.packages.urllib3")
urllib3_logger.setLevel(logging.DEBUG)
urllib3_logger.propagate = True
else:
# Remove all handlers
for handler in logging.root.handlers[:]:
logging.root.removeHandler(handler)
# set logging level to default
requests.cookies.cookielib.debug = False
api_logger.setLevel(logging.WARNING)
return |
Call subclasses via function to allow passing parent namespace to subclasses.
**Returns:** dict with subclass references.
def _subclass_container(self):
"""
Call subclasses via function to allow passing parent namespace to subclasses.
**Returns:** dict with subclass references.
"""
_parent_class = self
class GetWrapper(Get):
def __init__(self):
self._parent_class = _parent_class
class PostWrapper(Post):
def __init__(self):
self._parent_class = _parent_class
class PutWrapper(Put):
def __init__(self):
self._parent_class = _parent_class
class PatchWrapper(Patch):
def __init__(self):
self._parent_class = _parent_class
class DeleteWrapper(Delete):
def __init__(self):
self._parent_class = _parent_class
class InteractiveWrapper(Interactive):
def __init__(self):
self._parent_class = _parent_class
return {"get": GetWrapper,
"post": PostWrapper,
"put": PutWrapper,
"patch": PatchWrapper,
"delete": DeleteWrapper,
"interactive": InteractiveWrapper} |
Generic REST call worker function
**Parameters:**
- **url:** URL for the REST call
- **method:** METHOD for the REST call
- **data:** Optional DATA for the call (for POST/PUT/etc.)
- **sensitive:** Flag if content request/response should be hidden from logging functions
- **timeout:** Requests Timeout
- **content_json:** Bool on whether the Content-Type header should be set to application/json
- **retry:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
- **max_retry:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
- **retry_sleep:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
**Returns:** Requests.Response object, extended with:
- **cgx_status**: Bool, True if a successful CloudGenix response, False if error.
- **cgx_content**: Content of the response, guaranteed to be in Dict format. Empty/invalid responses
will be converted to a Dict response.
def rest_call(self, url, method, data=None, sensitive=False, timeout=None, content_json=True,
retry=None, max_retry=None, retry_sleep=None):
"""
Generic REST call worker function
**Parameters:**
- **url:** URL for the REST call
- **method:** METHOD for the REST call
- **data:** Optional DATA for the call (for POST/PUT/etc.)
- **sensitive:** Flag if content request/response should be hidden from logging functions
- **timeout:** Requests Timeout
- **content_json:** Bool on whether the Content-Type header should be set to application/json
- **retry:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
- **max_retry:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
- **retry_sleep:** DEPRECATED - please use `cloudgenix.API.modify_rest_retry` instead.
**Returns:** Requests.Response object, extended with:
- **cgx_status**: Bool, True if a successful CloudGenix response, False if error.
- **cgx_content**: Content of the response, guaranteed to be in Dict format. Empty/invalid responses
will be converted to a Dict response.
"""
# pull retry related items from Constructor if not specified.
if timeout is None:
timeout = self.rest_call_timeout
if retry is not None:
# Someone using deprecated retry code. Notify.
sys.stderr.write("WARNING: 'retry' option of rest_call() has been deprecated. "
"Please use 'API.modify_rest_retry()' instead.")
if max_retry is not None:
# Someone using deprecated retry code. Notify.
sys.stderr.write("WARNING: 'max_retry' option of rest_call() has been deprecated. "
"Please use 'API.modify_rest_retry()' instead.")
if retry_sleep is not None:
# Someone using deprecated retry code. Notify.
sys.stderr.write("WARNING: 'max_retry' option of rest_call() has been deprecated. "
"Please use 'API.modify_rest_retry()' instead.")
# Get logging level, use this to bypass logging functions with possible large content if not set.
logger_level = api_logger.getEffectiveLevel()
# populate headers and cookies from session.
if content_json and method.lower() not in ['get', 'delete']:
headers = {
'Content-Type': 'application/json'
}
else:
headers = {}
# add session headers
headers.update(self._session.headers)
cookie = self._session.cookies.get_dict()
# make sure data is populated if present.
if isinstance(data, (list, dict)):
data = json.dumps(data)
api_logger.debug('REST_CALL URL = %s', url)
# make request
try:
if not sensitive:
api_logger.debug('\n\tREQUEST: %s %s\n\tHEADERS: %s\n\tCOOKIES: %s\n\tDATA: %s\n',
method.upper(), url, headers, cookie, data)
# Actual request
response = self._session.request(method, url, data=data, verify=self.ca_verify_filename,
stream=True, timeout=timeout, headers=headers, allow_redirects=False)
# Request complete - lets parse.
# if it's a non-CGX-good response, return with cgx_status = False
if response.status_code not in [requests.codes.ok,
requests.codes.no_content,
requests.codes.found,
requests.codes.moved]:
# Simple JSON debug
if not sensitive:
try:
api_logger.debug('RESPONSE HEADERS: %s\n', json.dumps(
json.loads(text_type(response.headers)), indent=4))
except ValueError:
api_logger.debug('RESPONSE HEADERS: %s\n', text_type(response.headers))
try:
api_logger.debug('RESPONSE: %s\n', json.dumps(response.json(), indent=4))
except ValueError:
api_logger.debug('RESPONSE: %s\n', text_type(response.text))
else:
api_logger.debug('RESPONSE NOT LOGGED (sensitive content)')
api_logger.debug("Error, non-200 response received: %s", response.status_code)
# CGX extend requests.Response for return
response.cgx_status = False
response.cgx_content = self._catch_nonjson_streamresponse(response.text)
return response
else:
# Simple JSON debug
if not sensitive and (logger_level <= logging.DEBUG and logger_level != logging.NOTSET):
try:
api_logger.debug('RESPONSE HEADERS: %s\n', json.dumps(
json.loads(text_type(response.headers)), indent=4))
api_logger.debug('RESPONSE: %s\n', json.dumps(response.json(), indent=4))
except ValueError:
api_logger.debug('RESPONSE HEADERS: %s\n', text_type(response.headers))
api_logger.debug('RESPONSE: %s\n', text_type(response.text))
elif sensitive:
api_logger.debug('RESPONSE NOT LOGGED (sensitive content)')
# CGX extend requests.Response for return
response.cgx_status = True
response.cgx_content = self._catch_nonjson_streamresponse(response.text)
return response
except (requests.exceptions.Timeout, requests.exceptions.ConnectionError, urllib3.exceptions.MaxRetryError)\
as e:
api_logger.info("Error, %s.", text_type(e))
# make a requests.Response object for return since we didn't get one.
response = requests.Response
# CGX extend requests.Response for return
response.cgx_status = False
response.cgx_content = {
'_error': [
{
'message': 'REST Request Exception: {}'.format(e),
'data': {},
}
]
}
return response |
Function to clean up ca temp file for requests.
**Returns:** Removes TEMP ca file, no return
def _cleanup_ca_temp_file(self):
"""
Function to clean up ca temp file for requests.
**Returns:** Removes TEMP ca file, no return
"""
if os.name == 'nt':
if isinstance(self.ca_verify_filename, (binary_type, text_type)):
# windows requires file to be closed for access. Have to manually remove
os.unlink(self.ca_verify_filename)
else:
# other OS's allow close and delete of file.
self._ca_verify_file_handle.close() |
Break auth_token up into it's constituent values.
**Parameters:**
- **auth_token:** Auth_token string
**Returns:** dict with Auth Token constituents
def parse_auth_token(self, auth_token):
"""
Break auth_token up into it's constituent values.
**Parameters:**
- **auth_token:** Auth_token string
**Returns:** dict with Auth Token constituents
"""
# remove the random security key value from the front of the auth_token
auth_token_cleaned = auth_token.split('-', 1)[1]
# URL Decode the Auth Token
auth_token_decoded = self.url_decode(auth_token_cleaned)
# Create a new dict to hold the response.
auth_dict = {}
# Parse the token
for key_value in auth_token_decoded.split("&"):
key_value_list = key_value.split("=")
# check for valid token parts
if len(key_value_list) == 2 and type(key_value_list[0]) in [text_type, binary_type]:
auth_dict[key_value_list[0]] = key_value_list[1]
# Return the dict of key/values in the token.
return auth_dict |
Update the controller string with dynamic region info.
Controller string should end up as `<name[-env]>.<region>.cloudgenix.com`
**Parameters:**
- **region:** region string.
**Returns:** No return value, mutates the controller in the class namespace
def update_region_to_controller(self, region):
"""
Update the controller string with dynamic region info.
Controller string should end up as `<name[-env]>.<region>.cloudgenix.com`
**Parameters:**
- **region:** region string.
**Returns:** No return value, mutates the controller in the class namespace
"""
# default region position in a list
region_position = 1
# Check for a global "ignore region" flag
if self.ignore_region:
# bypass
api_logger.debug("IGNORE_REGION set, not updating controller region.")
return
api_logger.debug("Updating Controller Region")
api_logger.debug("CONTROLLER = %s", self.controller)
api_logger.debug("CONTROLLER_ORIG = %s", self.controller_orig)
api_logger.debug("CONTROLLER_REGION = %s", self.controller_region)
# Check if this is an initial region use or an update region use
if self.controller_orig:
controller_base = self.controller_orig
else:
controller_base = self.controller
self.controller_orig = self.controller
# splice controller string
controller_full_part_list = controller_base.split('.')
for idx, part in enumerate(controller_full_part_list):
# is the region already in the controller string?
if region == part:
# yes, controller already has apropriate region
api_logger.debug("REGION %s ALREADY IN CONTROLLER AT INDEX = %s", region, idx)
# update region if it is not already set.
if self.controller_region != region:
self.controller_region = region
api_logger.debug("UPDATED_CONTROLLER_REGION = %s", self.controller_region)
return
controller_part_count = len(controller_full_part_list)
# handle short domain case
if controller_part_count > 1:
# insert region
controller_full_part_list[region_position] = region
self.controller = ".".join(controller_full_part_list)
else:
# short domain, just add region
self.controller = ".".join(controller_full_part_list) + '.' + region
# update SDK vars with region info
self.controller_orig = controller_base
self.controller_region = region
api_logger.debug("UPDATED_CONTROLLER = %s", self.controller)
api_logger.debug("UPDATED_CONTROLLER_ORIG = %s", self.controller_orig)
api_logger.debug("UPDATED_CONTROLLER_REGION = %s", self.controller_region)
return |
Return region from a successful login response.
**Parameters:**
- **login_response:** requests.Response from a successful login.
**Returns:** region name.
def parse_region(self, login_response):
"""
Return region from a successful login response.
**Parameters:**
- **login_response:** requests.Response from a successful login.
**Returns:** region name.
"""
auth_token = login_response.cgx_content['x_auth_token']
auth_token_dict = self.parse_auth_token(auth_token)
auth_region = auth_token_dict.get('region')
return auth_region |
Sometimes, login cookie gets sent with region info instead of api.cloudgenix.com. This function
re-parses the original login request and applies cookies to the session if they now match the new region.
**Parameters:**
- **login_response:** requests.Response from a non-region login.
**Returns:** updates API() object directly, no return.
def reparse_login_cookie_after_region_update(self, login_response):
"""
Sometimes, login cookie gets sent with region info instead of api.cloudgenix.com. This function
re-parses the original login request and applies cookies to the session if they now match the new region.
**Parameters:**
- **login_response:** requests.Response from a non-region login.
**Returns:** updates API() object directly, no return.
"""
login_url = login_response.request.url
api_logger.debug("ORIGINAL REQUEST URL = %s", login_url)
# replace old controller with new controller.
login_url_new = login_url.replace(self.controller_orig, self.controller)
api_logger.debug("UPDATED REQUEST URL = %s", login_url_new)
# reset login url with new region
login_response.request.url = login_url_new
# prep cookie jar parsing
req = requests.cookies.MockRequest(login_response.request)
res = requests.cookies.MockResponse(login_response.raw._original_response.msg)
# extract cookies to session cookie jar.
self._session.cookies.extract_cookies(res, req)
return |
Validate a streamed response is JSON. Return a Python dictionary either way.
**Parameters:**
- **rawresponse:** Streamed Response from Requests.
**Returns:** Dictionary
def _catch_nonjson_streamresponse(rawresponse):
"""
Validate a streamed response is JSON. Return a Python dictionary either way.
**Parameters:**
- **rawresponse:** Streamed Response from Requests.
**Returns:** Dictionary
"""
# attempt to load response for return.
try:
response = json.loads(rawresponse)
except (ValueError, TypeError):
if rawresponse:
response = {
'_error': [
{
'message': 'Response not in JSON format.',
'data': rawresponse,
}
]
}
else:
# in case of null response, return empty dict.
response = {}
return response |
URL Decode function using REGEX
**Parameters:**
- **url:** URLENCODED text string
**Returns:** Non URLENCODED string
def url_decode(url):
"""
URL Decode function using REGEX
**Parameters:**
- **url:** URLENCODED text string
**Returns:** Non URLENCODED string
"""
return re.compile('%([0-9a-fA-F]{2})', re.M).sub(lambda m: chr(int(m.group(1), 16)), url) |
Get optimal file system buffer size (in bytes) for I/O calls.
def blksize(path):
"""
Get optimal file system buffer size (in bytes) for I/O calls.
"""
diskfreespace = win32file.GetDiskFreeSpace
dirname = os.path.dirname(fullpath(path))
try:
cluster_sectors, sector_size = diskfreespace(dirname)[:2]
size = cluster_sectors * sector_size
except win32file.error as e:
if e.winerror != winerror.ERROR_NOT_READY:
raise
sleep(3)
size = blksize(dirname)
return size |
Pass email hash, return Gravatar URL. You can get email hash like this::
import hashlib
avatar_hash = hashlib.md5(email.lower().encode('utf-8')).hexdigest()
Visit https://en.gravatar.com/site/implement/images/ for more information.
:param hash: The email hash used to generate avatar URL.
:param size: The size of the avatar, default to 100 pixel.
:param rating: The rating of the avatar, default to ``g``
:param default: The type of default avatar, default to ``identicon``.
:param include_extension: Append a '.jpg' extension at the end of URL, default to ``False``.
:param force_default: Force to use default avatar, default to ``False``.
def gravatar(hash, size=100, rating='g', default='identicon', include_extension=False, force_default=False):
"""Pass email hash, return Gravatar URL. You can get email hash like this::
import hashlib
avatar_hash = hashlib.md5(email.lower().encode('utf-8')).hexdigest()
Visit https://en.gravatar.com/site/implement/images/ for more information.
:param hash: The email hash used to generate avatar URL.
:param size: The size of the avatar, default to 100 pixel.
:param rating: The rating of the avatar, default to ``g``
:param default: The type of default avatar, default to ``identicon``.
:param include_extension: Append a '.jpg' extension at the end of URL, default to ``False``.
:param force_default: Force to use default avatar, default to ``False``.
"""
if include_extension:
hash += '.jpg'
default = default or current_app.config['AVATARS_GRAVATAR_DEFAULT']
query_string = urlencode({'s': int(size), 'r': rating, 'd': default})
if force_default:
query_string += '&q=y'
return 'https://gravatar.com/avatar/' + hash + '?' + query_string |
Return avatar URL at social media.
Visit https://avatars.io for more information.
:param username: The username of the social media.
:param platform: One of facebook, instagram, twitter, gravatar.
:param size: The size of avatar, one of small, medium and large.
def social_media(username, platform='twitter', size='medium'):
"""Return avatar URL at social media.
Visit https://avatars.io for more information.
:param username: The username of the social media.
:param platform: One of facebook, instagram, twitter, gravatar.
:param size: The size of avatar, one of small, medium and large.
"""
return 'https://avatars.io/{platform}/{username}/{size}'.format(
platform=platform, username=username, size=size) |
Load jcrop css file.
:param css_url: The custom CSS URL.
def jcrop_css(css_url=None):
"""Load jcrop css file.
:param css_url: The custom CSS URL.
"""
if css_url is None:
if current_app.config['AVATARS_SERVE_LOCAL']:
css_url = url_for('avatars.static', filename='jcrop/css/jquery.Jcrop.min.css')
else:
css_url = 'https://cdn.jsdelivr.net/npm/jcrop-0.9.12@0.9.12/css/jquery.Jcrop.min.css'
return Markup('<link rel="stylesheet" href="%s">' % css_url) |
Load jcrop Javascript file.
:param js_url: The custom JavaScript URL.
:param with_jquery: Include jQuery or not, default to ``True``.
def jcrop_js(js_url=None, with_jquery=True):
"""Load jcrop Javascript file.
:param js_url: The custom JavaScript URL.
:param with_jquery: Include jQuery or not, default to ``True``.
"""
serve_local = current_app.config['AVATARS_SERVE_LOCAL']
if js_url is None:
if serve_local:
js_url = url_for('avatars.static', filename='jcrop/js/jquery.Jcrop.min.js')
else:
js_url = 'https://cdn.jsdelivr.net/npm/jcrop-0.9.12@0.9.12/js/jquery.Jcrop.min.js'
if with_jquery:
if serve_local:
jquery = '<script src="%s"></script>' % url_for('avatars.static', filename='jcrop/js/jquery.min.js')
else:
jquery = '<script src="https://cdn.jsdelivr.net/npm/jcrop-0.9.12@0.9.12/js/jquery.min.js"></script>'
else:
jquery = ''
return Markup('''%s\n<script src="%s"></script>
''' % (jquery, js_url)) |
Create a crop box.
:param endpoint: The endpoint of view function that serve avatar image file.
:param filename: The filename of the image that need to be crop.
def crop_box(endpoint=None, filename=None):
"""Create a crop box.
:param endpoint: The endpoint of view function that serve avatar image file.
:param filename: The filename of the image that need to be crop.
"""
crop_size = current_app.config['AVATARS_CROP_BASE_WIDTH']
if endpoint is None or filename is None:
url = url_for('avatars.static', filename='default/default_l.jpg')
else:
url = url_for(endpoint, filename=filename)
return Markup('<img src="%s" id="crop-box" style="max-width: %dpx; display: block;">' % (url, crop_size)) |
Create a preview box.
:param endpoint: The endpoint of view function that serve avatar image file.
:param filename: The filename of the image that need to be crop.
def preview_box(endpoint=None, filename=None):
"""Create a preview box.
:param endpoint: The endpoint of view function that serve avatar image file.
:param filename: The filename of the image that need to be crop.
"""
preview_size = current_app.config['AVATARS_CROP_PREVIEW_SIZE'] or current_app.config['AVATARS_SIZE_TUPLE'][2]
if endpoint is None or filename is None:
url = url_for('avatars.static', filename='default/default_l.jpg')
else:
url = url_for(endpoint, filename=filename)
return Markup('''
<div id="preview-box">
<div class="preview-box" style="width: %dpx; height: %dpx; overflow: hidden;">
<img src="%s" class="jcrop-preview" alt="Preview"/>
</div>
</div>''' % (preview_size, preview_size, url)) |
Initialize jcrop.
:param min_size: The minimal size of crop area.
def init_jcrop(min_size=None):
"""Initialize jcrop.
:param min_size: The minimal size of crop area.
"""
init_x = current_app.config['AVATARS_CROP_INIT_POS'][0]
init_y = current_app.config['AVATARS_CROP_INIT_POS'][1]
init_size = current_app.config['AVATARS_CROP_INIT_SIZE'] or current_app.config['AVATARS_SIZE_TUPLE'][2]
if current_app.config['AVATARS_CROP_MIN_SIZE']:
min_size = min_size or current_app.config['AVATARS_SIZE_TUPLE'][2]
min_size_js = 'jcrop_api.setOptions({minSize: [%d, %d]});' % (min_size, min_size)
else:
min_size_js = ''
return Markup('''
<script type="text/javascript">
jQuery(function ($) {
// Create variables (in this scope) to hold the API and image size
var jcrop_api,
boundx,
boundy,
// Grab some information about the preview pane
$preview = $('#preview-box'),
$pcnt = $('#preview-box .preview-box'),
$pimg = $('#preview-box .preview-box img'),
xsize = $pcnt.width(),
ysize = $pcnt.height();
$('#crop-box').Jcrop({
onChange: updatePreview,
onSelect: updateCoords,
setSelect: [%s, %s, %s, %s],
aspectRatio: 1
}, function () {
// Use the API to get the real image size
var bounds = this.getBounds();
boundx = bounds[0];
boundy = bounds[1];
// Store the API in the jcrop_api variable
jcrop_api = this;
%s
jcrop_api.focus();
// Move the preview into the jcrop container for css positioning
$preview.appendTo(jcrop_api.ui.holder);
});
function updatePreview(c) {
if (parseInt(c.w) > 0) {
var rx = xsize / c.w;
var ry = ysize / c.h;
$pimg.css({
width: Math.round(rx * boundx) + 'px',
height: Math.round(ry * boundy) + 'px',
marginLeft: '-' + Math.round(rx * c.x) + 'px',
marginTop: '-' + Math.round(ry * c.y) + 'px'
});
}
}
});
function updateCoords(c) {
$('#x').val(c.x);
$('#y').val(c.y);
$('#w').val(c.w);
$('#h').val(c.h);
}
</script>
''' % (init_x, init_y, init_size, init_size, min_size_js)) |
Resize an avatar.
:param img: The image that needs to be resize.
:param base_width: The width of output image.
def resize_avatar(self, img, base_width):
"""Resize an avatar.
:param img: The image that needs to be resize.
:param base_width: The width of output image.
"""
w_percent = (base_width / float(img.size[0]))
h_size = int((float(img.size[1]) * float(w_percent)))
img = img.resize((base_width, h_size), PIL.Image.ANTIALIAS)
return img |
Save an avatar as raw image, return new filename.
:param image: The image that needs to be saved.
def save_avatar(self, image):
"""Save an avatar as raw image, return new filename.
:param image: The image that needs to be saved.
"""
path = current_app.config['AVATARS_SAVE_PATH']
filename = uuid4().hex + '_raw.png'
image.save(os.path.join(path, filename))
return filename |
Crop avatar with given size, return a list of file name: [filename_s, filename_m, filename_l].
:param filename: The raw image's filename.
:param x: The x-pos to start crop.
:param y: The y-pos to start crop.
:param w: The crop width.
:param h: The crop height.
def crop_avatar(self, filename, x, y, w, h):
"""Crop avatar with given size, return a list of file name: [filename_s, filename_m, filename_l].
:param filename: The raw image's filename.
:param x: The x-pos to start crop.
:param y: The y-pos to start crop.
:param w: The crop width.
:param h: The crop height.
"""
x = int(x)
y = int(y)
w = int(w)
h = int(h)
sizes = current_app.config['AVATARS_SIZE_TUPLE']
if not filename:
path = os.path.join(self.root_path, 'static/default/default_l.jpg')
else:
path = os.path.join(current_app.config['AVATARS_SAVE_PATH'], filename)
print(path)
raw_img = Image.open(path)
base_width = current_app.config['AVATARS_CROP_BASE_WIDTH']
if raw_img.size[0] >= base_width:
raw_img = self.resize_avatar(raw_img, base_width=base_width)
cropped_img = raw_img.crop((x, y, x + w, y + h))
filename = uuid4().hex
avatar_s = self.resize_avatar(cropped_img, base_width=sizes[0])
avatar_m = self.resize_avatar(cropped_img, base_width=sizes[1])
avatar_l = self.resize_avatar(cropped_img, base_width=sizes[2])
filename_s = filename + '_s.png'
filename_m = filename + '_m.png'
filename_l = filename + '_l.png'
path_s = os.path.join(current_app.config['AVATARS_SAVE_PATH'], filename_s)
path_m = os.path.join(current_app.config['AVATARS_SAVE_PATH'], filename_m)
path_l = os.path.join(current_app.config['AVATARS_SAVE_PATH'], filename_l)
avatar_s.save(path_s, optimize=True, quality=85)
avatar_m.save(path_m, optimize=True, quality=85)
avatar_l.save(path_l, optimize=True, quality=85)
return [filename_s, filename_m, filename_l] |
Byte representation of a PNG image
def get_image(self, string, width, height, pad=0):
"""
Byte representation of a PNG image
"""
hex_digest_byte_list = self._string_to_byte_list(string)
matrix = self._create_matrix(hex_digest_byte_list)
return self._create_image(matrix, width, height, pad) |
Create a pastel colour hex colour string
def _get_pastel_colour(self, lighten=127):
"""
Create a pastel colour hex colour string
"""
def r():
return random.randint(0, 128) + lighten
return r(), r(), r() |
Determine the liminanace of an RGB colour
def _luminance(self, rgb):
"""
Determine the liminanace of an RGB colour
"""
a = []
for v in rgb:
v = v / float(255)
if v < 0.03928:
result = v / 12.92
else:
result = math.pow(((v + 0.055) / 1.055), 2.4)
a.append(result)
return a[0] * 0.2126 + a[1] * 0.7152 + a[2] * 0.0722 |
Creates a hex digest of the input string given to create the image,
if it's not already hexadecimal
Returns:
Length 16 list of rgb value range integers
(each representing a byte of the hex digest)
def _string_to_byte_list(self, data):
"""
Creates a hex digest of the input string given to create the image,
if it's not already hexadecimal
Returns:
Length 16 list of rgb value range integers
(each representing a byte of the hex digest)
"""
bytes_length = 16
m = self.digest()
m.update(str.encode(data))
hex_digest = m.hexdigest()
return list(int(hex_digest[num * 2:num * 2 + 2], bytes_length)
for num in range(bytes_length)) |
Check if the n (index) of hash_bytes is 1 or 0.
def _bit_is_one(self, n, hash_bytes):
"""
Check if the n (index) of hash_bytes is 1 or 0.
"""
scale = 16 # hexadecimal
if not hash_bytes[int(n / (scale / 2))] >> int(
(scale / 2) - ((n % (scale / 2)) + 1)) & 1 == 1:
return False
return True |
Generates a PNG byte list
def _create_image(self, matrix, width, height, pad):
"""
Generates a PNG byte list
"""
image = Image.new("RGB", (width + (pad * 2),
height + (pad * 2)), self.bg_colour)
image_draw = ImageDraw.Draw(image)
# Calculate the block width and height.
block_width = float(width) / self.cols
block_height = float(height) / self.rows
# Loop through blocks in matrix, draw rectangles.
for row, cols in enumerate(matrix):
for col, cell in enumerate(cols):
if cell:
image_draw.rectangle((
pad + col * block_width, # x1
pad + row * block_height, # y1
pad + (col + 1) * block_width - 1, # x2
pad + (row + 1) * block_height - 1 # y2
), fill=self.fg_colour)
stream = BytesIO()
image.save(stream, format="png", optimize=True)
# return the image byte data
return stream.getvalue() |
This matrix decides which blocks should be filled fg/bg colour
True for fg_colour
False for bg_colour
hash_bytes - array of hash bytes values. RGB range values in each slot
Returns:
List representation of the matrix
[[True, True, True, True],
[False, True, True, False],
[True, True, True, True],
[False, False, False, False]]
def _create_matrix(self, byte_list):
"""
This matrix decides which blocks should be filled fg/bg colour
True for fg_colour
False for bg_colour
hash_bytes - array of hash bytes values. RGB range values in each slot
Returns:
List representation of the matrix
[[True, True, True, True],
[False, True, True, False],
[True, True, True, True],
[False, False, False, False]]
"""
# Number of rows * cols halfed and rounded
# in order to fill opposite side
cells = int(self.rows * self.cols / 2 + self.cols % 2)
matrix = [[False] * self.cols for num in range(self.rows)]
for cell_number in range(cells):
# If the bit with index corresponding to this cell is 1
# mark that cell as fg_colour
# Skip byte 1, that's used in determining fg_colour
if self._bit_is_one(cell_number, byte_list[1:]):
# Find cell coordinates in matrix.
x_row = cell_number % self.rows
y_col = int(cell_number / self.cols)
# Set coord True and its opposite side
matrix[x_row][self.cols - y_col - 1] = True
matrix[x_row][y_col] = True
return matrix |
Generate and save avatars, return a list of file name: [filename_s, filename_m, filename_l].
:param text: The text used to generate image.
def generate(self, text):
"""Generate and save avatars, return a list of file name: [filename_s, filename_m, filename_l].
:param text: The text used to generate image.
"""
sizes = current_app.config['AVATARS_SIZE_TUPLE']
path = current_app.config['AVATARS_SAVE_PATH']
suffix = {sizes[0]: 's', sizes[1]: 'm', sizes[2]: 'l'}
for size in sizes:
image_byte_array = self.get_image(
string=str(text),
width=int(size),
height=int(size),
pad=int(size * 0.1))
self.save(image_byte_array, save_location=os.path.join(path, '%s_%s.png' % (text, suffix[size])))
return [text + '_s.png', text + '_m.png', text + '_l.png'] |
Read values.
Args:
vals (list): list of strings representing values
def read(self, vals):
"""Read values.
Args:
vals (list): list of strings representing values
"""
i = 0
if len(vals[i]) == 0:
self.city = None
else:
self.city = vals[i]
i += 1
if len(vals[i]) == 0:
self.state_province_region = None
else:
self.state_province_region = vals[i]
i += 1
if len(vals[i]) == 0:
self.country = None
else:
self.country = vals[i]
i += 1
if len(vals[i]) == 0:
self.source = None
else:
self.source = vals[i]
i += 1
if len(vals[i]) == 0:
self.wmo = None
else:
self.wmo = vals[i]
i += 1
if len(vals[i]) == 0:
self.latitude = None
else:
self.latitude = vals[i]
i += 1
if len(vals[i]) == 0:
self.longitude = None
else:
self.longitude = vals[i]
i += 1
if len(vals[i]) == 0:
self.timezone = None
else:
self.timezone = vals[i]
i += 1
if len(vals[i]) == 0:
self.elevation = None
else:
self.elevation = vals[i]
i += 1 |
Corresponds to IDD Field `city`
Args:
value (str): value for IDD Field `city`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def city(self, value=None):
"""Corresponds to IDD Field `city`
Args:
value (str): value for IDD Field `city`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError('value {} need to be of type str '
'for field `city`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `city`')
self._city = value |
Corresponds to IDD Field `state_province_region`
Args:
value (str): value for IDD Field `state_province_region`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def state_province_region(self, value=None):
"""Corresponds to IDD Field `state_province_region`
Args:
value (str): value for IDD Field `state_province_region`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError(
'value {} need to be of type str '
'for field `state_province_region`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `state_province_region`')
self._state_province_region = value |
Corresponds to IDD Field `country`
Args:
value (str): value for IDD Field `country`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def country(self, value=None):
"""Corresponds to IDD Field `country`
Args:
value (str): value for IDD Field `country`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError('value {} need to be of type str '
'for field `country`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `country`')
self._country = value |
Corresponds to IDD Field `source`
Args:
value (str): value for IDD Field `source`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def source(self, value=None):
"""Corresponds to IDD Field `source`
Args:
value (str): value for IDD Field `source`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError('value {} need to be of type str '
'for field `source`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `source`')
self._source = value |
Corresponds to IDD Field `wmo` usually a 6 digit field. Used as
alpha in EnergyPlus.
Args:
value (str): value for IDD Field `wmo`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def wmo(self, value=None):
"""Corresponds to IDD Field `wmo` usually a 6 digit field. Used as
alpha in EnergyPlus.
Args:
value (str): value for IDD Field `wmo`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError('value {} need to be of type str '
'for field `wmo`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `wmo`')
self._wmo = value |
Corresponds to IDD Field `latitude`
+ is North, - is South, degree minutes represented in decimal (i.e. 30 minutes is .5)
Args:
value (float): value for IDD Field `latitude`
Unit: deg
Default value: 0.0
value >= -90.0
value <= 90.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def latitude(self, value=0.0):
"""Corresponds to IDD Field `latitude`
+ is North, - is South, degree minutes represented in decimal (i.e. 30 minutes is .5)
Args:
value (float): value for IDD Field `latitude`
Unit: deg
Default value: 0.0
value >= -90.0
value <= 90.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `latitude`'.format(value))
if value < -90.0:
raise ValueError('value need to be greater or equal -90.0 '
'for field `latitude`')
if value > 90.0:
raise ValueError('value need to be smaller 90.0 '
'for field `latitude`')
self._latitude = value |
Corresponds to IDD Field `longitude`
- is West, + is East, degree minutes represented in decimal (i.e. 30 minutes is .5)
Args:
value (float): value for IDD Field `longitude`
Unit: deg
Default value: 0.0
value >= -180.0
value <= 180.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def longitude(self, value=0.0):
"""Corresponds to IDD Field `longitude`
- is West, + is East, degree minutes represented in decimal (i.e. 30 minutes is .5)
Args:
value (float): value for IDD Field `longitude`
Unit: deg
Default value: 0.0
value >= -180.0
value <= 180.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `longitude`'.format(value))
if value < -180.0:
raise ValueError('value need to be greater or equal -180.0 '
'for field `longitude`')
if value > 180.0:
raise ValueError('value need to be smaller 180.0 '
'for field `longitude`')
self._longitude = value |
Corresponds to IDD Field `timezone` Time relative to GMT.
Args:
value (float): value for IDD Field `timezone`
Unit: hr - not on standard units list???
Default value: 0.0
value >= -12.0
value <= 12.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def timezone(self, value=0.0):
"""Corresponds to IDD Field `timezone` Time relative to GMT.
Args:
value (float): value for IDD Field `timezone`
Unit: hr - not on standard units list???
Default value: 0.0
value >= -12.0
value <= 12.0
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `timezone`'.format(value))
if value < -12.0:
raise ValueError('value need to be greater or equal -12.0 '
'for field `timezone`')
if value > 12.0:
raise ValueError('value need to be smaller 12.0 '
'for field `timezone`')
self._timezone = value |
Corresponds to IDD Field `elevation`
Args:
value (float): value for IDD Field `elevation`
Unit: m
Default value: 0.0
value >= -1000.0
value < 9999.9
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def elevation(self, value=0.0):
"""Corresponds to IDD Field `elevation`
Args:
value (float): value for IDD Field `elevation`
Unit: m
Default value: 0.0
value >= -1000.0
value < 9999.9
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `elevation`'.format(value))
if value < -1000.0:
raise ValueError('value need to be greater or equal -1000.0 '
'for field `elevation`')
if value >= 9999.9:
raise ValueError('value need to be smaller 9999.9 '
'for field `elevation`')
self._elevation = value |
Exports object to its string representation.
Args:
top (bool): if True appends `internal_name` before values.
All non list objects should be exported with value top=True,
all list objects, that are embedded in as fields inlist objects
should be exported with `top`=False
Returns:
str: The objects string representation
def export(self, top=True):
"""Exports object to its string representation.
Args:
top (bool): if True appends `internal_name` before values.
All non list objects should be exported with value top=True,
all list objects, that are embedded in as fields inlist objects
should be exported with `top`=False
Returns:
str: The objects string representation
"""
out = []
if top:
out.append(self._internal_name)
out.append(self._to_str(self.city))
out.append(self._to_str(self.state_province_region))
out.append(self._to_str(self.country))
out.append(self._to_str(self.source))
out.append(self._to_str(self.wmo))
out.append(self._to_str(self.latitude))
out.append(self._to_str(self.longitude))
out.append(self._to_str(self.timezone))
out.append(self._to_str(self.elevation))
return ",".join(out) |
Read values.
Args:
vals (list): list of strings representing values
def read(self, vals):
"""Read values.
Args:
vals (list): list of strings representing values
"""
i = 0
if len(vals[i]) == 0:
self.title_of_design_condition = None
else:
self.title_of_design_condition = vals[i]
i += 1
if len(vals[i]) == 0:
self.unkown_field = None
else:
self.unkown_field = vals[i]
i += 1
if len(vals[i]) == 0:
self.design_stat_heating = None
else:
self.design_stat_heating = vals[i]
i += 1
if len(vals[i]) == 0:
self.coldestmonth = None
else:
self.coldestmonth = vals[i]
i += 1
if len(vals[i]) == 0:
self.db996 = None
else:
self.db996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db990 = None
else:
self.db990 = vals[i]
i += 1
if len(vals[i]) == 0:
self.dp996 = None
else:
self.dp996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hr_dp996 = None
else:
self.hr_dp996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_dp996 = None
else:
self.db_dp996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.dp990 = None
else:
self.dp990 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hr_dp990 = None
else:
self.hr_dp990 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_dp990 = None
else:
self.db_dp990 = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws004c = None
else:
self.ws004c = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_ws004c = None
else:
self.db_ws004c = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws010c = None
else:
self.ws010c = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_ws010c = None
else:
self.db_ws010c = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws_db996 = None
else:
self.ws_db996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wd_db996 = None
else:
self.wd_db996 = vals[i]
i += 1
if len(vals[i]) == 0:
self.design_stat_cooling = None
else:
self.design_stat_cooling = vals[i]
i += 1
if len(vals[i]) == 0:
self.hottestmonth = None
else:
self.hottestmonth = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbr = None
else:
self.dbr = vals[i]
i += 1
if len(vals[i]) == 0:
self.db004 = None
else:
self.db004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb_db004 = None
else:
self.wb_db004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db010 = None
else:
self.db010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb_db010 = None
else:
self.wb_db010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db020 = None
else:
self.db020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb_db020 = None
else:
self.wb_db020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb004 = None
else:
self.wb004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_wb004 = None
else:
self.db_wb004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb010 = None
else:
self.wb010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_wb010 = None
else:
self.db_wb010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wb020 = None
else:
self.wb020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_wb020 = None
else:
self.db_wb020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws_db004 = None
else:
self.ws_db004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wd_db004 = None
else:
self.wd_db004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.dp004 = None
else:
self.dp004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hr_dp004 = None
else:
self.hr_dp004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_dp004 = None
else:
self.db_dp004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.dp010 = None
else:
self.dp010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hr_dp010 = None
else:
self.hr_dp010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_dp010 = None
else:
self.db_dp010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.dp020 = None
else:
self.dp020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hr_dp020 = None
else:
self.hr_dp020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_dp020 = None
else:
self.db_dp020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.en004 = None
else:
self.en004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_en004 = None
else:
self.db_en004 = vals[i]
i += 1
if len(vals[i]) == 0:
self.en010 = None
else:
self.en010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_en010 = None
else:
self.db_en010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.en020 = None
else:
self.en020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.db_en020 = None
else:
self.db_en020 = vals[i]
i += 1
if len(vals[i]) == 0:
self.hrs_84_and_db12_8_or_20_6 = None
else:
self.hrs_84_and_db12_8_or_20_6 = vals[i]
i += 1
if len(vals[i]) == 0:
self.design_stat_extremes = None
else:
self.design_stat_extremes = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws010 = None
else:
self.ws010 = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws025 = None
else:
self.ws025 = vals[i]
i += 1
if len(vals[i]) == 0:
self.ws050 = None
else:
self.ws050 = vals[i]
i += 1
if len(vals[i]) == 0:
self.wbmax = None
else:
self.wbmax = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin_mean = None
else:
self.dbmin_mean = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax_mean = None
else:
self.dbmax_mean = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin_stddev = None
else:
self.dbmin_stddev = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax_stddev = None
else:
self.dbmax_stddev = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin05years = None
else:
self.dbmin05years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax05years = None
else:
self.dbmax05years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin10years = None
else:
self.dbmin10years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax10years = None
else:
self.dbmax10years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin20years = None
else:
self.dbmin20years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax20years = None
else:
self.dbmax20years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmin50years = None
else:
self.dbmin50years = vals[i]
i += 1
if len(vals[i]) == 0:
self.dbmax50years = None
else:
self.dbmax50years = vals[i]
i += 1 |
Corresponds to IDD Field `title_of_design_condition`
Args:
value (str): value for IDD Field `title_of_design_condition`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def title_of_design_condition(self, value=None):
"""Corresponds to IDD Field `title_of_design_condition`
Args:
value (str): value for IDD Field `title_of_design_condition`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError(
'value {} need to be of type str '
'for field `title_of_design_condition`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `title_of_design_condition`')
self._title_of_design_condition = value |
Corresponds to IDD Field `unkown_field` Empty field in data.
Args:
value (str): value for IDD Field `unkown_field`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def unkown_field(self, value=None):
"""Corresponds to IDD Field `unkown_field` Empty field in data.
Args:
value (str): value for IDD Field `unkown_field`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError('value {} need to be of type str '
'for field `unkown_field`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `unkown_field`')
self._unkown_field = value |
Corresponds to IDD Field `design_stat_heating`
Args:
value (str): value for IDD Field `design_stat_heating`
Accepted values are:
- Heating
Default value: Heating
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def design_stat_heating(self, value="Heating"):
"""Corresponds to IDD Field `design_stat_heating`
Args:
value (str): value for IDD Field `design_stat_heating`
Accepted values are:
- Heating
Default value: Heating
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = str(value)
except ValueError:
raise ValueError(
'value {} need to be of type str '
'for field `design_stat_heating`'.format(value))
if ',' in value:
raise ValueError('value should not contain a comma '
'for field `design_stat_heating`')
vals = set()
vals.add("Heating")
if value not in vals:
raise ValueError('value {} is not an accepted value for '
'field `design_stat_heating`'.format(value))
self._design_stat_heating = value |
Corresponds to IDD Field `coldestmonth`
Args:
value (int): value for IDD Field `coldestmonth`
value >= 1
value <= 12
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def coldestmonth(self, value=None):
"""Corresponds to IDD Field `coldestmonth`
Args:
value (int): value for IDD Field `coldestmonth`
value >= 1
value <= 12
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = int(value)
except ValueError:
raise ValueError('value {} need to be of type int '
'for field `coldestmonth`'.format(value))
if value < 1:
raise ValueError('value need to be greater or equal 1 '
'for field `coldestmonth`')
if value > 12:
raise ValueError('value need to be smaller 12 '
'for field `coldestmonth`')
self._coldestmonth = value |
Corresponds to IDD Field `db996`
Dry-bulb temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def db996(self, value=None):
""" Corresponds to IDD Field `db996`
Dry-bulb temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `db996`'.format(value))
self._db996 = value |
Corresponds to IDD Field `db990`
Dry-bulb temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def db990(self, value=None):
""" Corresponds to IDD Field `db990`
Dry-bulb temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `db990`'.format(value))
self._db990 = value |
Corresponds to IDD Field `dp996`
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `dp996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def dp996(self, value=None):
""" Corresponds to IDD Field `dp996`
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `dp996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `dp996`'.format(value))
self._dp996 = value |
Corresponds to IDD Field `hr_dp996`
humidity ratio, calculated at standard atmospheric pressure
at elevation of station, corresponding to
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `hr_dp996`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def hr_dp996(self, value=None):
""" Corresponds to IDD Field `hr_dp996`
humidity ratio, calculated at standard atmospheric pressure
at elevation of station, corresponding to
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `hr_dp996`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `hr_dp996`'.format(value))
self._hr_dp996 = value |
Corresponds to IDD Field `db_dp996`
mean coincident drybulb temperature corresponding to
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db_dp996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def db_dp996(self, value=None):
""" Corresponds to IDD Field `db_dp996`
mean coincident drybulb temperature corresponding to
Dew-point temperature corresponding to 99.6% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db_dp996`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `db_dp996`'.format(value))
self._db_dp996 = value |
Corresponds to IDD Field `dp990`
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `dp990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def dp990(self, value=None):
""" Corresponds to IDD Field `dp990`
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `dp990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `dp990`'.format(value))
self._dp990 = value |
Corresponds to IDD Field `hr_dp990`
humidity ratio, calculated at standard atmospheric pressure
at elevation of station, corresponding to
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `hr_dp990`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def hr_dp990(self, value=None):
""" Corresponds to IDD Field `hr_dp990`
humidity ratio, calculated at standard atmospheric pressure
at elevation of station, corresponding to
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `hr_dp990`
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `hr_dp990`'.format(value))
self._hr_dp990 = value |
Corresponds to IDD Field `db_dp990`
mean coincident drybulb temperature corresponding to
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db_dp990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
def db_dp990(self, value=None):
""" Corresponds to IDD Field `db_dp990`
mean coincident drybulb temperature corresponding to
Dew-point temperature corresponding to 90.0% annual cumulative
frequency of occurrence (cold conditions)
Args:
value (float): value for IDD Field `db_dp990`
Unit: C
if `value` is None it will not be checked against the
specification and is assumed to be a missing value
Raises:
ValueError: if `value` is not a valid value
"""
if value is not None:
try:
value = float(value)
except ValueError:
raise ValueError('value {} need to be of type float '
'for field `db_dp990`'.format(value))
self._db_dp990 = value |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.