blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 3 281 | content_id stringlengths 40 40 | detected_licenses listlengths 0 57 | license_type stringclasses 2 values | repo_name stringlengths 6 116 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringclasses 313 values | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 18.2k 668M ⌀ | star_events_count int64 0 102k | fork_events_count int64 0 38.2k | gha_license_id stringclasses 17 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 107 values | src_encoding stringclasses 20 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 4 6.02M | extension stringclasses 78 values | content stringlengths 2 6.02M | authors listlengths 1 1 | author stringlengths 0 175 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
8fad0e1ffc2107d073e1e36d04a28101f3819657 | 741663bc3d7dfc49c4b881eafd24e387e925392a | /src/multimodal/models/predict_model.py | feed72242acc25cc48d02f3229d36ee81ff20aca | [
"MIT"
] | permissive | markrofail/multi-modal-deep-learning-for-vehicle-sensor-data-abstraction-and-attack-detection | e957731ca293eb65d1b56dfce6abe607ec20bbf7 | 2f252c072f3091bb27506978dd90311f7f82f386 | refs/heads/master | 2023-07-30T02:22:28.412955 | 2020-06-20T17:12:54 | 2020-06-20T17:12:54 | 273,745,779 | 0 | 0 | MIT | 2021-09-08T02:12:12 | 2020-06-20T16:37:18 | HTML | UTF-8 | Python | false | false | 6,173 | py | import os
import click
import cv2
import matplotlib.pyplot as plt
import numpy as np
from src.helpers import paths
from src.helpers.flags import AttackModes, Verbose
from src.multimodal import multimodal
from src.multimodal.data import make_dataset
###############################################################################
# DATA PARAMETERS
###############################################################################
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'
np.set_printoptions(suppress=True, precision=4, sign=' ')
config = paths.config.read(paths.config.multimodal())
VERBOSE = config['ENVIROMENT_CONFIG']['VERBOSE']
DEFAULT_DRIVE = '2011_09_26'
DEFAULT_NUMBER = 1
DEFAULT_FRAME = 1
DEFAULT_ATTACK = 1
DEFAULT_ATTACK_FLAG = True
def print_results(**kargs):
keys = np.array(list(kargs.keys()))
keys = np.sort(keys)
print('# results:')
for key in keys:
value = kargs[key]
print('## {k} = {v}'.format(k=key, v=value))
print()
def display_results(drive_date, drive_number, drive_frame, attack, result):
if attack:
img_path = paths.attack.interim_frame(drive_date, drive_number, drive_frame)
else:
img_path = paths.rgb.interim_frame(drive_date, drive_number, drive_frame)
img = cv2.imread(str(img_path))
if img is None:
raise Exception("could not load image !")
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
plt.imshow(img, interpolation="bicubic")
result = 'result: {}!'.format('PASS' if result else 'FAIL')
plt.xlabel(result)
plt.show()
def feed_forward(input_image, input_depth, label):
# Create the network
net = multimodal.Multimodal()
# Load pretrained model
model_path = str(paths.checkpoints.multimodal()) # ./checkpoints/multimodal/train
net.model.load_weights(model_path)
# Predict
pred = net.model.predict([[input_image], [input_depth]])
print_args = dict()
print_args['predict_raw'] = pred
notations = ['Normal', 'Attack']
pred = np.argmax(pred)
label = np.argmax(label)
verdict = 'PASS' if label == pred else 'FAIL'
print_args['label'] = str(notations[label])
print_args['predict'] = str(notations[pred])
print_args['verdict'] = verdict
print_results(**print_args)
return label == pred
def predict(drive_date, drive_number, frame, attack, attack_type):
# create the frame data
print('# generating data...')
frame_data = [(drive_date, drive_number, frame)]
make_dataset.make_data(
frame_data, attack_type=attack_type, verbose=Verbose.SILENT, keep=True)
if attack:
rgb_path = paths.attack.processed_tensor(drive_date, drive_number, frame)
label_data = np.array([0, 1])
else:
rgb_path = paths.rgb.processed_tensor(drive_date, drive_number, frame)
label_data = np.array([1, 0])
rgb_data = np.load(rgb_path)
depth_path = paths.depth.processed_tensor(drive_date, drive_number, frame)
depth_data = np.load(depth_path)
print('# feedforward data...')
return feed_forward(input_image=rgb_data, input_depth=depth_data, label=label_data)
def get_arguments_interactively():
args_dict = dict()
print('\nEnter drive details:')
default_drive = DEFAULT_DRIVE
input_date = input(
'# drive date (format: \'yyyy_mm_dd\') [\'{}\']:'.format(
default_drive))
if input_date:
args_dict['drive_date'] = input_date
default_number = DEFAULT_NUMBER
input_number = input('# drive number (int) [{}]:'.format(default_number))
if input_number:
args_dict['drive_number'] = int(input_number)
default_frame = DEFAULT_FRAME
input_frame = input('# drive frame (int) [{}]:'.format(default_frame))
if input_frame:
args_dict['drive_frame'] = int(input_frame)
default_attack_flag = DEFAULT_ATTACK_FLAG
input_attack_flag = input('# normal/attack (0/1) [{}]:'.format(int(default_attack_flag)))
if input_attack_flag:
args_dict['attack'] = bool(int(input_attack_flag))
if input_attack_flag and not args_dict['attack']:
print()
return args_dict
default_attack = DEFAULT_ATTACK
input_attack = input('# inpainting/translation (1/2) [{}]:'.format(int(default_attack)))
if input_attack:
args_dict['attack_type'] = int(input_attack)
print()
return args_dict
@click.option(
'--drive_date', type=str, default=DEFAULT_DRIVE, help='date of the drive')
@click.option(
'--drive_number',
type=int,
default=DEFAULT_NUMBER,
help='date of the drive')
@click.option(
'--drive_frame',
type=int,
default=DEFAULT_FRAME,
help='frame within the drive')
@click.option(
'--attack_type',
type=int,
default=DEFAULT_ATTACK,
help='frame within the drive')
@click.option(
'--attack/--normal',
help='normal or attack',
)
@click.option(
'--interactive',
'-i',
is_flag=True,
help='interactively',
)
@click.option(
'--display',
'-d',
is_flag=True,
help='display image',
)
@click.command()
def main(interactive=False,
display=False,
drive_date=DEFAULT_DRIVE,
drive_number=DEFAULT_NUMBER,
drive_frame=DEFAULT_FRAME,
attack_type=DEFAULT_ATTACK,
attack=True):
if interactive:
args_dict = get_arguments_interactively()
if 'drive_date' in args_dict:
drive_date = args_dict['drive_date']
if 'drive_number' in args_dict:
drive_number = args_dict['drive_number']
if 'drive_frame' in args_dict:
drive_frame = args_dict['drive_frame']
if 'attack_type' in args_dict:
attack_type = args_dict['attack_type']
if 'attack' in args_dict:
attack = args_dict['attack']
path = paths.rgb.external_frame(drive_date, drive_number, drive_frame)
assert path.exists(), 'frame does not exist'
result = predict(drive_date, drive_number, drive_frame, attack, attack_type)
if display:
display_results(drive_date, drive_number, drive_frame, attack, result)
if __name__ == '__main__':
main()
'''
# for attack:
python -m src.multimodal.models.predict_model -d \
--drive_date 2011_09_26 --drive_number 1 --drive_frame 1 --attack --attack_type 2
# for normal:
python -m src.multimodal.models.predict_model -d \
--drive_date 2011_09_26 --drive_number 1 --drive_frame 1 --normal --attack_type 2
'''
| [
"markm.rofail@gmail.com"
] | markm.rofail@gmail.com |
f9a070cd4f54bb5db6172fe1754a0c3ce4c8635e | 32c1f5e296e3fdfd5583b767ba388450c22d4920 | /xclib/tests/prosody_test.py | 0105858d2c138d2284a48ac4bc09dbd37cfca878 | [
"MIT"
] | permissive | wenzhuoz/xmpp-cloud-auth | 848a930677a282e942211a129bff801d85e13ca1 | 8359b80197d2d7743a41cc3c60475d8c8cd2920d | refs/heads/master | 2020-03-16T21:57:13.370802 | 2018-05-11T09:46:13 | 2018-05-11T09:46:13 | 133,021,003 | 0 | 0 | null | 2018-05-11T09:41:24 | 2018-05-11T09:41:24 | null | UTF-8 | Python | false | false | 1,001 | py | import sys
import unittest
from xclib.prosody_io import prosody_io
from xclib.tests.iostub import iostub
class TestProsody(unittest.TestCase, iostub):
def test_input(self):
self.stub_stdin('isuser:login:\n' +
'auth:log:dom:pass\n')
tester = iter(prosody_io.read_request())
output = tester.next()
assert output == ('isuser', 'login', '')
output = tester.next()
assert output == ('auth', 'log', 'dom', 'pass')
try:
output = tester.next()
assert False # Should raise StopIteration
except StopIteration:
pass
def test_output_false(self):
self.stub_stdout()
prosody_io.write_response(False)
self.assertEqual(sys.stdout.getvalue(), '0\n')
# Cannot be merged, as getvalue() returns the aggregate value
def test_output_true(self):
self.stub_stdout()
prosody_io.write_response(True)
self.assertEqual(sys.stdout.getvalue(), '1\n')
| [
"marcel.waldvogel@uni-konstanz.de"
] | marcel.waldvogel@uni-konstanz.de |
273419d8074c1ba7bc5bf2633a855f024edd0d7c | 00e3bc6d538cb0f1cfcb14cb725c11e1be2503b0 | /steps/offer_steps.py | 71e28eadf0a67c74d36002e46472f77f44d463a0 | [] | no_license | minhwang/carousell_appium_automation | 960e9a433efdd61026cadffdf58f403c7a71759c | 554fda4c1fcf7cecefc2e513a0b6ebad042c6203 | refs/heads/master | 2020-12-30T14:19:21.096863 | 2017-08-21T07:00:48 | 2017-08-21T07:00:48 | 91,303,260 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 610 | py | from behave import *
from carousell import App, Platform
use_step_matcher("re")
@when('I submit an offer')
def step_impl(context):
app = App(Platform.ANDROID)
context.chat = app\
.welcome_view\
.create(context.wd)\
.login_with_email()\
.login(context.user_id, context.user_pwd)\
.browse()\
.browse_category('Cars')\
.view_product(0)\
.buy()\
.submit()\
.yes()
@then('The app brings me chat')
def step_impl(context):
assert context.chat | [
"min81.hwang@gmail.com"
] | min81.hwang@gmail.com |
d57b2b6ba4918837e7dbca3a7ab328705da2f0df | 0593fbd857b5286f93c60056982e9b4d2b23eff7 | /automaton_t.py | 8b8fbfe222d8adc889c56645509ec4df7b0ec6bf | [] | no_license | Moysenko/NFA-converter | 5c74145b081eff6efb834a8a7458f4bb0f83419c | 82de98b2d6bf7602387b5e82a0b246ef406334d9 | refs/heads/main | 2023-01-05T16:12:31.640233 | 2020-10-16T21:52:31 | 2020-10-16T21:52:31 | 300,353,590 | 0 | 2 | null | 2020-10-16T21:52:33 | 2020-10-01T16:51:01 | null | UTF-8 | Python | false | false | 10,366 | py | from vertex_t import Vertex
from collections import defaultdict, namedtuple
import sys, os
import json
class Automaton:
def __init__(self, start=0, vertices=None, alphabet=None, other_automaton=None):
if other_automaton is None:
self.start = start
self.vertices = vertices or dict()
self._free_vertex_id = (max(self.vertices) + 1) if len(self.vertices) else 0
self.alphabet = alphabet or set()
else:
self.start = other_automaton.start
self._free_vertex_id = other_automaton._free_vertex_id
self.vertices = dict(other_automaton.vertices)
self.alphabet = other_automaton.alphabet.copy()
def __getitem__(self, vertex_id):
if vertex_id not in self.vertices:
self._free_vertex_id = max(self._free_vertex_id, vertex_id + 1)
self.vertices[vertex_id] = Vertex(vertex_id)
return self.vertices[vertex_id]
def add_edge(self, vertex_from, vertex_to, word):
self[vertex_to] # in order to add vertex_to in self.vertices
self[vertex_from].add_edge(word, vertex_to)
def scan(self):
self.alphabet = set(input('Alphabet: '))
number_of_edges = int(input('Number of edges: '))
print('Edges: (in format "{from} {to} {word}", symbol "-" stands for empty string)')
self.vertices = dict()
self._free_vertex_id = 0
for edge_id in range(number_of_edges):
vertex_from, vertex_to, word = input('Edge #{0}: '.format(edge_id)).split()
if word == '-': # null edge
word = None
self.add_edge(int(vertex_from), int(vertex_to), word)
self.start = int(input('Start state: '))
for terminal_vertex_id in list(map(int, input('Terminal states: ').split())):
self[terminal_vertex_id].is_terminal = True
def read_from_json(self, filename):
with open(filename, 'r') as input_file, open(os.devnull, 'w') as output_file:
sys.stdin = input_file
sys.stdout = output_file
self.scan()
sys.stdin = sys.__stdin__
sys.stdout = sys.__stdout__
def __str__(self):
output = 'Automaton:\n'
prefix = ' ' * 4
output += prefix + 'Edges:\n'
terminal_vertices = []
for vertex in self.vertices:
if self[vertex].is_terminal:
terminal_vertices.append(vertex)
for word, neighbors in self[vertex].edges.items():
for vertex_to in neighbors:
output += prefix * 2 + 'From {0} to {1} by {2}\n'.format(vertex, vertex_to, word or '-')
output += prefix + 'Start state: {0}'.format(self.start)
output += prefix + 'Terminal states: ' + ', '.join(str(v) for v in terminal_vertices) + '\n'
return output
def _split_long_edges(self): # replaces all edges with keys longer than 1 with multiple edges
edges_to_delete = []
for vertex_id in self.vertices:
for word in self[vertex_id].edges:
if word is not None and len(word) > 1:
edges_to_delete.append((vertex_id, word))
for vertex_id, word in edges_to_delete:
for edge_end in self[vertex_id].neighbors_by_word(word):
last_vertex = vertex_id
for i, letter in enumerate(word):
if i + 1 == len(word):
vertex_to = edge_end
else:
vertex_to = self._free_vertex_id
self._free_vertex_id += 1
self.add_edge(last_vertex, vertex_to, letter)
last_vertex = vertex_to
self[vertex_id].remove_edge(word)
def _shorten_path(self, vertex_from, word, visited_vertices): # dfs in wich every step except first is using null edge
if word in vertex_from.edges:
for vertex_to in vertex_from.edges[word]:
if vertex_to not in visited_vertices:
visited_vertices.add(vertex_to)
self._shorten_path(self[vertex_to], None, visited_vertices)
def _get_shortened_null_paths(self, vertex):
new_edges = defaultdict(set)
reached_by_null_edges = set()
self._shorten_path(vertex, None, reached_by_null_edges)
for vertex_to in reached_by_null_edges:
for word in self[vertex_to].edges:
if word is not None:
new_edges[word] |= self[vertex_to].edges[word]
return new_edges
def _remove_null_edges(self):
for vertex in self.vertices.values(): # add new terminal vertices
if not vertex.is_terminal:
reached_by_null_edges = set()
self._shorten_path(vertex, None, reached_by_null_edges)
for terminal_vertex in reached_by_null_edges:
vertex.is_terminal |= self[terminal_vertex].is_terminal
new_edges = dict()
for vertex_id, vertex in self.vertices.items(): # add new adges and delete null edges
new_edges[vertex_id] = self._get_shortened_null_paths(vertex)
for vertex_id, edges in new_edges.items():
vertex = self[vertex_id]
if None in vertex.edges:
del vertex.edges[None]
for word, vertices_to in edges.items():
vertex.edges[word] |= vertices_to
def _reachable_from_vertex(self, current_vertex, visited):
for neighbors in current_vertex.edges.values():
for vertex_to in neighbors:
if vertex_to not in visited:
visited.add(vertex_to)
self._reachable_from_vertex(self[vertex_to], visited)
def _init_from_automaton_subsets(self, other):
for subset in range(2**other._free_vertex_id): # build automaton on subsets
for vertex_id, vertex in other.vertices.items():
if (2**vertex_id) & subset:
if other.start == vertex_id and (2**vertex_id) == subset:
self.start = subset
self[subset].is_terminal |= vertex.is_terminal
for word in vertex.edges:
self[subset].edges[word] |= vertex.edges[word]
def _replace_edges_with_subsets(self):
for vertex in self.vertices:
for word in self[vertex].edges:
subset_to = 0
for vertex_to in self[vertex].edges[word]:
subset_to += 2**vertex_to
self[vertex].edges[word] = {subset_to}
def _init_from_useful_vertices(self, automaton):
useful_vertices = {automaton.start}
automaton._reachable_from_vertex(automaton[automaton.start], useful_vertices)
useful_vertex_id = {old_vertex_id: vertex_id
for vertex_id, old_vertex_id in enumerate(useful_vertices)}
self.vertices = dict()
self._free_vertex_id = len(useful_vertices)
self.start = useful_vertex_id[automaton.start]
for vertex in useful_vertices:
self[useful_vertex_id[vertex]].is_terminal |= automaton[vertex].is_terminal
for word, neighbors in automaton[vertex].edges.items():
for vertex_to in neighbors:
self.add_edge(useful_vertex_id[vertex], useful_vertex_id[vertex_to], word)
def _remove_duplicate_edges(self):
new_automaton = Automaton(start=0, vertices=dict(), alphabet=set())
new_automaton._init_from_automaton_subsets(self)
new_automaton._replace_edges_with_subsets()
self._init_from_useful_vertices(new_automaton)
def to_dfa(self):
self._split_long_edges()
self._remove_null_edges()
self._remove_duplicate_edges()
def accept_string(self, word):
current_state = self.start
for letter in word:
try:
current_state = self[current_state].go(letter)
except KeyError:
return False
return self[current_state].is_terminal
def to_cdfa(self): # it is assumed that automaton is already deterministic
missing_edges = []
Edge = namedtuple('Edge', 'vertex, letter')
for vertex in self.vertices.values():
missing_edges += [Edge(vertex=vertex, letter=letter) for letter in self.alphabet if letter not in vertex.edges]
if missing_edges:
dummy_vertex = self._free_vertex_id
self._free_vertex_id += 1
for edge in missing_edges:
self.add_edge(edge.vertex.id, dummy_vertex, edge.letter)
for letter in self.alphabet:
self.add_edge(dummy_vertex, dummy_vertex, letter)
self._free_vertex_id += 1
def reverse_cdfa(self):
for vertex in self.vertices.values():
vertex.is_terminal ^= 1
def _equivalence_groups(self):
group = dict()
for vertex in self.vertices.values():
group[vertex.id] = int(vertex.is_terminal)
old_number_of_classes = 1
current_number_of_classes = 2
step_id = 0
while old_number_of_classes != current_number_of_classes:
step_id += 1
output_groups = defaultdict(list)
for vertex in self.vertices.values():
key = tuple([group[vertex.id]] + [group[vertex.go(letter)] for letter in self.alphabet])
output_groups[key].append(vertex.id)
old_number_of_classes = current_number_of_classes
current_number_of_classes = len(output_groups)
group = dict()
for group_id, vertices in enumerate(output_groups.values()):
for vertex in vertices:
group[vertex] = group_id
return group
def to_minimal_cdfa(self):
group = self._equivalence_groups()
new_automaton = Automaton(start=group[self.start], alphabet=self.alphabet, vertices={})
for vertex in self.vertices.values():
new_automaton[group[vertex.id]].is_terminal |= vertex.is_terminal
for word in self.alphabet:
new_automaton.add_edge(group[vertex.id], group[vertex.go(word)], word)
self.__init__(other_automaton=new_automaton)
| [
"moysenkom@gmail.com"
] | moysenkom@gmail.com |
f6b1a60691f6f53ffd82e4d7f873cb546ad3dc71 | 25c9290b26cf39888aad8890c7778b4f4f5c5daa | /Triode-Car/microbit(only)/1.1 Control motor with button/main.py | fd61a508c41d0449300ed51c09c856f5e028b8b9 | [] | no_license | vtt-info/Micropython-10 | 0ca3feaf2297ce5b9c24e46ea0e538316a9e5980 | 1dfc63af4ea863c9c61170974bd022e10ee59a1c | refs/heads/main | 2023-08-29T02:45:49.280754 | 2021-11-03T09:36:45 | 2021-11-03T09:36:45 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,193 | py | #用micro:bit的AB按钮控制triodecar电机
'''
一切与硬件交互直接相关的东西都存在于 microbit函数库中,
一般直接从中引用全部功能。
'''
from microbit import *
'''
使用def来创建自定义函数,将可能重复应用的某些功能的代码写入其中以便调用。
micro:bit的pin14引脚控制着triodecar的左电机,pin15引脚控制着右电机,
引脚输出高电平将使电机停转,低电平将使电机运行。
'''
def direction_stop():
pin14.write_digital(1)
pin15.write_digital(1)
def direction_foward():
pin14.write_digital(0)
pin15.write_digital(0)
'''右电机转,左电机停转,triodecar向左行驶。'''
def direction_left():
pin14.write_digital(1)
pin15.write_digital(0)
'''右电机停转,左电机转,triodecar向右行驶。'''
def direction_right():
pin14.write_digital(0)
pin15.write_digital(1)
'''开始主循环,进入循环后如果不执行break语句将在系统关机前永远循环下去'''
while True:
'''if条件判断,若满足条件则执行。
当按钮A和按钮B都被按下时,显示北向箭头,控制triodecar向前行驶。'''
if button_a.is_pressed() and button_b.is_pressed():
display.show(Image.ARROW_N, delay=0, wait=True, loop=False, clear=False)
direction_foward()
'''elif表示“否则如果”,“不满足if的条件但满足这个条件”的意思。
此处即为:若仅按钮A被按下,则显示西向箭头,控制triodecar向右行驶。'''
elif button_a.is_pressed():
display.show(Image.ARROW_W, delay=0, wait=True, loop=False, clear=False)
direction_right()
'''若仅按钮B被按下,则显示东向箭头,控制triodecar向左行驶。'''
elif button_b.is_pressed():
display.show(Image.ARROW_E, delay=0, wait=True, loop=False, clear=False)
direction_left()
'''else为if和elif都判定为否的情况下将执行。
此处为:若按钮A或按钮B都没被按下,则显示困倦图标,控制triodecar停车。'''
else:
display.show(Image.ASLEEP, delay=0, wait=True, loop=False, clear=False)
direction_stop()
| [
"76462385+Wind-stormger@users.noreply.github.com"
] | 76462385+Wind-stormger@users.noreply.github.com |
d2ca8ecfa155dc2a582b16faf400c3ff2d5b08a0 | 3999521ed32fc384ecb0446f2956cd44734bc989 | /apiconfig.py | 6424b51cb212f1139d52645e4fea6f38cd245aa0 | [
"Apache-2.0"
] | permissive | yangyzp/shadowsocksdd | 56789be92be33e45044900caf5b6eb08dc924b21 | 2ecc23db49b28a51e07758b4968d8bcd54ad18d3 | refs/heads/manyuser | 2021-06-19T06:13:08.005075 | 2018-05-07T03:39:52 | 2018-05-07T03:39:52 | 132,402,930 | 2 | 1 | Apache-2.0 | 2021-06-01T22:22:08 | 2018-05-07T03:36:02 | Python | UTF-8 | Python | false | false | 887 | py | # Config
NODE_ID = 1
# hour,set 0 to disable
SPEEDTEST = 0
CLOUDSAFE = 1
ANTISSATTACK = 0
AUTOEXEC = 0
MU_SUFFIX = 'v1m3cc2x'
MU_REGEX = '%5m'
SERVER_PUB_ADDR = '127.0.0.1' # mujson_mgr need this to generate ssr link
API_INTERFACE = 'glzjinmod' # glzjinmod, modwebapi
WEBAPI_URL = 'https://zhaoj.in'
WEBAPI_TOKEN = 'glzjin'
# mudb
MUDB_FILE = 'mudb.json'
# Mysql
MYSQL_HOST = '23.234.197.24'
MYSQL_PORT = 3306
MYSQL_USER = 'sspanel'
MYSQL_PASS = 'qq469566135'
MYSQL_DB = 'sspanel'
MYSQL_SSL_ENABLE = 0
MYSQL_SSL_CA = ''
MYSQL_SSL_CERT = ''
MYSQL_SSL_KEY = ''
# API
API_HOST = '127.0.0.1'
API_PORT = 80
API_PATH = '/mu/v2/'
API_TOKEN = 'abcdef'
API_UPDATE_TIME = 60
# Manager (ignore this)
MANAGE_PASS = 'ss233333333'
# if you want manage in other server you should set this value to global ip
MANAGE_BIND_IP = '127.0.0.1'
# make sure this port is idle
MANAGE_PORT = 23333
| [
"noreply@github.com"
] | noreply@github.com |
49b2a27304baa0fd8ed633b0cf8b738d64a581b3 | 1345dc02bbe664db2a8804a0fc0e8a5a7129147b | /sale.py | 43b89520cc39bbaaae2ae1f6992dadd73933e3d1 | [] | no_license | gavindav/DataScienceRepo | e2f9c68a9352c51f72678c7bc2247fd5b83a2a28 | 6378e854ff2428962f8f2d571eeea92a1b3d3edd | refs/heads/main | 2023-02-17T01:26:04.980578 | 2021-01-11T08:59:55 | 2021-01-11T08:59:55 | 328,593,942 | 0 | 0 | null | 2021-01-11T08:59:56 | 2021-01-11T08:18:28 | Python | UTF-8 | Python | false | false | 3,700 | py | # (C) Copyright IBM Corp. 2020
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
"""SaleOrder."""
from odoo import api, fields, models
from odoo.tools.config import config
import re
class SaleOrder(models.Model):
"""SaleOrder."""
_inherit = 'sale.order'
case_number = fields.Char('case_number', required=False)
def _get_ddb_service_types(self):
return self.env['delivery.carrier']._get_ddb_service_types()
ddb_service_type = fields.Selection(
_get_ddb_service_types, string="DDB Service Type")
@api.onchange('carrier_id')
def _onchange_carrier_id(self):
"""Return the element value."""
orders = self.env['sale.order'].sudo().search(
[('package_delivery_group', '=', self.package_delivery_group)])
delivery_type = self.carrier_id.delivery_type
for order in orders:
order.carrier_id = self.carrier_id
if hasattr(
self.carrier_id, delivery_type + "_default_service_type"):
expr = "order." + delivery_type + \
"_service_type = self.carrier_id." + \
delivery_type + "_default_service_type"
exec(expr)
order.ibmorder_data.deliverymethod = self.carrier_id.name
def generate_next_case_number_in_sequence(self, base_initialisation):
ddbcase_obj = self.env['ibmddb.case_number'].with_user(2).create({'ordno': self.name, })
case_num = ddbcase_obj.id
while case_num < base_initialisation:
ddbcase_obj.unlink()
ddbcase_obj = self.env['ibmddb.case_number'].with_user(2).create({'ordno': self.name, })
case_num = ddbcase_obj.id
return case_num
def int2base36(self, x, alphabet='0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ'):
"""Convert an integer to its string representation to base 36."""
val = ''
while x > 0:
x, idx = divmod(x, 36)
val = alphabet[idx] + val
return "0" * (5 - len(val)) + val
def get_ddb_case_number(self):
"""Return the element value."""
# format is 76ZSWAaaaaa00
# initialise from 00000 so sumber is from 76ZSWA0000000 to
# 76ZSWAZZZZZ00
ddbprefix = config.get("ddbprefix")
ddb_base_initialisation = config.get("ddb_base_initialisation")
if ddb_base_initialisation and re.match("^[0-9]*$", ddb_base_initialisation.strip()):
ddb_base_initialisation = int(ddb_base_initialisation.strip(), 36)
else:
ddb_base_initialisation = 0 # 10 = 'AXXXXX', 11='BXXXXX'
if ddbprefix and re.match("^[0-9]*$", ddbprefix.strip()):
ddbprefix = int(ddbprefix.strip())
else:
ddbprefix = 10 # 10 = 'AXXXXX', 11='BXXXXX'
foundation = 36 * ddbprefix * 36 * 36 * 36 * 36
if self.case_number:
return self.case_number
else:
case_number_prefix = "76ZSW"
case_num = self.generate_next_case_number_in_sequence(ddb_base_initialisation)
case_num = self.int2base36(case_num + foundation)
case_number_suffix = "00"
self.case_number = case_number_prefix + case_num + \
case_number_suffix
pdg_sale_objects = self.env['sale.order'].sudo().search(
[('package_delivery_group', '=', self.package_delivery_group)])
for pdg_so in pdg_sale_objects:
pdg_so.case_number = case_number_prefix + case_num + \
case_number_suffix
return self.case_number
| [
"noreply@github.com"
] | noreply@github.com |
9adcc12b7ba970cf3f19bbad83bbd0ecb835aa85 | f15c8b3a6b093c3b70a900221f485d74a1bc1f95 | /0_joan_stark/golf.py | 2f646ee459477d0f80708844a426c3b1cdd2b1bf | [
"MIT"
] | permissive | wang0618/ascii-art | 2955023e47b988f491b9d46bc8a300ba4a6cdd60 | 7ce6f152541716034bf0a22d341a898b17e2865f | refs/heads/master | 2023-07-17T23:17:31.187906 | 2021-09-04T12:46:31 | 2021-09-04T12:46:31 | 400,987,004 | 4 | 0 | null | null | null | null | UTF-8 | Python | false | false | 15,978 | py | # Hole in One!
# https://web.archive.org/web/20000307135811/http://geocities.com/SoHo/Gallery/6446/golfanim.htm
duration = 200
name = "Golf"
frames = [
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" /|\\o |\n" +
" | | |\n" +
" ,|/| |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^",
" \n" +
" |>18>>\n" +
" |\n" +
" __O |\n" +
" / /\\o |\n" +
" ,/ | |\n" +
" | |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^",
" \n" +
" |>18>>\n" +
" |\n" +
" |\n" +
" __O |\n" +
" \\ \\ |\n" +
" / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" __O |\n" +
" / /\\ |\n" +
" ,/ |\\ |\n" +
" |/ o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" |\\ |\n" +
" /\\| |\n" +
" / ||o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" |\\ |\n" +
" |\\| |\n" +
" / ||o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" /> |\n" +
" //\\ |\n" +
" ,// / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" ,___/| |\n" +
" /\\ |\n" +
" / / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" `\\ |>18>>\n" +
" \\ |\n" +
" <<O |\n" +
" | |\n" +
" |\\ |\n" +
" / | o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" /` |>18>>\n" +
" / |\n" +
" <<O |\n" +
" \\ |\n" +
" /\\ |\n" +
" / / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" /` |>18>>\n" +
" / |\n" +
" <<O |\n" +
" \\ |\n" +
" /\\ |\n" +
" / / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" /` |>18>>\n" +
" / |\n" +
" <<O |\n" +
" \\ |\n" +
" /\\ |\n" +
" / / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" `\\ |>18>>\n" +
" \\ |\n" +
" <<O |\n" +
" | |\n" +
" |\\ |\n" +
" / | o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" ,___/| |\n" +
" /\\ |\n" +
" / / o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" |\\ |\n" +
" |\\| |\n" +
" / ||o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" `/ |\n" +
" O__/ |\n" +
" \\-` o |\n" +
" /\\ . |\n" +
" / / .' |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . o |>18>>\n" +
" \\ . |\n" +
" O>> . |\n" +
" \\ . |\n" +
" /\\ . |\n" +
" / / .' |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . 'o |\n" +
" \\ . |\n" +
" /\\ . |\n" +
" / / .' |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . ' |\n" +
" \\ . ' . |\n" +
" /\\ . . |\n" +
" / / .' o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . ' |\n" +
" \\ . ' . |\n" +
" /\\ . . . o |\n" +
" / / .' . |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . ' |\n" +
" \\ . ' . |\n" +
" /\\ . . . ' . |\n" +
" / / .' . o |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . ' |\n" +
" \\ . ' . |\n" +
" /\\ . . . ' . |\n" +
" / / .' . . o\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" '\\ . . |>18>>\n" +
" \\ . ' . |\n" +
" O>> . ' |\n" +
" \\ . ' . |\n" +
" /\\ . . . ' . |\n" +
" / / .' . . .|\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" `/ . . |>18>>\n" +
" / . ' . |\n" +
" \\O/ . ' |\n" +
" | . ' . |\n" +
" /\\ . . . ' . |\n" +
" / | .' . . .|\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" `/ . . |>18>>\n" +
" / . ' . |\n" +
" __O/ . ' |\n" +
" | . ' . |\n" +
" /\\ . . . ' . |\n" +
" | \\ .' . . .|\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" . . |>18>>\n" +
" `/ . ' . |\n" +
" O__/ . ' |\n" +
" /| . ' . |\n" +
" /\\ . . . ' . |\n" +
" / / .' . . .|\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" \\O |\n" +
" |\\ |\n" +
" /\\\\ |\n" +
" / | \\, |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" /|\\ |\n" +
" |\\\\ |\n" +
" / | \\, |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" yipee! |>18>>\n" +
" |\n" +
" \\O |\n" +
" |\\ |\n" +
" /\\\\ |\n" +
" / | \\, |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" yipee! |>18>>\n" +
" |\n" +
" O |\n" +
" /|\\ |\n" +
" |\\\\ |\n" +
" / | \\, |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" yipee! |>18>>\n" +
" |\n" +
" O |\n" +
" /|\\ |\n" +
" / |\\ |\n" +
" /,/ | |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ",
" \n" +
" |>18>>\n" +
" |\n" +
" O |\n" +
" /|\\o |\n" +
" | | |\n" +
" ,|/| |\n" +
" jgs^^^^^^^`^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ "
]
| [
"wang0.618@qq.com"
] | wang0.618@qq.com |
fb004d77dcce82d96824cd25b93bca814fdb7157 | e4aa60a4f8d5f7d85be02b10f05aa956dc92ca68 | /venv/bin/pip | a288e2a0ae32ac4cbff413bfd22b155817b4fa31 | [] | no_license | SerenaPaley/WordUp | 37c29fb2f673421a26ad89a70ae62e75ae2e79a3 | 86b4bb82382abc241eaf8d8e1847d1796365936c | refs/heads/master | 2023-03-18T08:54:53.396244 | 2021-03-02T18:12:50 | 2021-03-02T18:12:50 | 343,645,743 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 259 | #!/Users/spaley/PycharmProjects/WordGame/venv/bin/python
# -*- coding: utf-8 -*-
import re
import sys
from pip._internal.cli.main import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"serena.paley8@gmail.com"
] | serena.paley8@gmail.com | |
3a409d438db536481f5aef2b5b889704149e23f9 | 0d6ecf1dd8d84b1a8ab8f929c0ea4e08531a63fa | /alexa_auth.py | caea5f70132e7c23c54d6f2da3f5a58195ddd124 | [
"LicenseRef-scancode-unknown-license-reference",
"MIT"
] | permissive | molodoj88/AlexaDevice | 846f1d99fc8e918999c3c47a1996e3eeaaaff13f | 957bdf79970a04f7e85fd0f7835801d0cca08d92 | refs/heads/master | 2021-01-22T21:07:07.390869 | 2017-02-24T13:54:17 | 2017-02-24T13:54:17 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,232 | py | #!/usr/bin/env python3
import alexa_params
import alexa_control
import alexa_http_config
import socket
import threading
from zeroconf import raw_input, ServiceInfo, Zeroconf
from http.server import HTTPServer
localHTTP = None
zeroconf = None
info = None
def get_local_address():
s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
s.connect(("www.amazon.com", 80))
res = s.getsockname()[0]
s.close()
return res
def start():
global localHTTP, zeroconf, info, httpthread
ip = get_local_address()
print("Local IP is " + ip)
desc = {'version': '0.1'}
info = ServiceInfo("_http._tcp.local.",
"Alexa Device._http._tcp.local.",
socket.inet_aton(ip), alexa_params.LOCAL_PORT, 0, 0,
desc, alexa_params.LOCAL_HOST + ".")
zeroconf = Zeroconf()
zeroconf.registerService(info)
print("Local mDNS is started, domain is " + alexa_params.LOCAL_HOST)
localHTTP = HTTPServer(("", alexa_params.LOCAL_PORT), alexa_http_config.AlexaConfig)
httpthread = threading.Thread(target=localHTTP.serve_forever)
httpthread.start()
print("Local HTTP is " + alexa_params.BASE_URL)
alexa_control.start()
def close():
localHTTP.shutdown()
httpthread.join()
zeroconf.unregisterService(info)
zeroconf.close()
alexa_control.close()
| [
"2xl@mail.ru"
] | 2xl@mail.ru |
a02ede264cfadd79a5af66ac1ef5a8d3658d7477 | 37ccb985555670df4a9df2650b41b7cc49c4cecd | /WEEK4/Hamburger.py | b858d3968a4b65a31b4f1171d97e7fd9d66f339a | [] | no_license | PeravitK/PSIT | f525620ad72fdcf55c84b115e8a0484f9f941f60 | f31522199fff312b26d2e3f5d54b0d34a2bec675 | refs/heads/main | 2023-08-18T04:36:09.966788 | 2021-09-19T13:55:37 | 2021-09-19T13:55:37 | 405,643,641 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 171 | py | '''Hambergur'''
def main():
'''docstring'''
num1 = int(input())
num2 = int(input())
print("%s%s%s" %("|"*num1, "*"*(num1+num2)*2, "|"*num2))
main()
| [
"noreply@github.com"
] | noreply@github.com |
348fc47cef3dc9dc96c748af7cf91394fd8222e7 | 2d7c21a793c8080a090ce8c9f05df38f6477c7c7 | /tests/data_templates/test_field_definitions.py | c4f05eb9f57000460d8661f4d47b2a554f7826ea | [
"Apache-2.0"
] | permissive | kids-first/kf-api-study-creator | c40e0a8a514fd52a857e9a588635ef76d16d5bc7 | ba62b369e6464259ea92dbb9ba49876513f37fba | refs/heads/master | 2023-08-17T01:09:38.789364 | 2023-08-15T14:06:29 | 2023-08-15T14:06:29 | 149,347,812 | 3 | 0 | Apache-2.0 | 2023-09-08T15:33:40 | 2018-09-18T20:25:38 | Python | UTF-8 | Python | false | false | 5,204 | py | import os
import json
import pytest
import pandas
from marshmallow import ValidationError
from pprint import pprint
from creator.data_templates.models import TemplateVersion
from creator.data_templates.field_definitions_schema import (
coerce_number,
coerce_bool,
FieldDefinitionSchema,
FieldDefinitionsSchema
)
@pytest.mark.parametrize(
"in_value, expected_out",
[
("0.0", 0.0),
(0.0, 0.0),
("0", 0),
(0, 0),
("10.0", 10),
(10.0, 10),
("200", 200),
(200, 200),
("1.234", 1.234),
(1.234, 1.234),
("foo", "foo"),
(None, None),
]
)
def test_coerce_number(in_value, expected_out):
"""
Test helper function that coerces strings to float/int
"""
assert coerce_number(in_value) == expected_out
@pytest.mark.parametrize(
"in_value, expected_out",
[
(True, True),
(False, False),
("foo", "foo"),
("0.0", False),
("1", True),
("True", True),
("FALSE", False),
("Yes", True),
("no", False),
("Required", True),
("Not Required", False),
(None, False),
]
)
def test_coerce_bool(in_value, expected_out):
"""
Test helper function that coerces strings to booleans
"""
assert coerce_bool(in_value) == expected_out
def test_schema_clean():
"""
Test FieldDefinitionSchema.clean method
"""
schema = FieldDefinitionSchema()
# Test keys are all snake cased
in_data = {
"Label": None,
"Data Type": None,
}
out_data = schema.clean(in_data)
assert {"label", "data_type"} == set(out_data.keys())
# Test data_type default
assert out_data["data_type"] == "string"
# Test data_type casing
in_data["data_type"] = "Number"
out_data = schema.clean(in_data)
assert out_data["data_type"] == "number"
# Test accepted_values
in_data["accepted_values"] = None
out_data = schema.clean(in_data)
assert out_data["accepted_values"] is None
in_data["data_type"] = "foobar"
in_data["accepted_values"] = "1.0, 2.0, 3.0"
out_data = schema.clean(in_data)
assert out_data["accepted_values"] == ["1.0", "2.0", "3.0"]
assert out_data["data_type"] == "enum"
# Test missing values
in_data["missing_values"] = None
out_data = schema.clean(in_data)
assert out_data["missing_values"] is None
in_data["missing_values"] = "None, Unknown"
out_data = schema.clean(in_data)
assert ["None", "Unknown"] == out_data["missing_values"]
# Test empty strings handled properly
in_data["accepted_values"] = " "
in_data["missing_values"] = ""
in_data["required"] = " "
in_data["data_type"] = " "
out_data = schema.clean(in_data)
assert out_data["accepted_values"] is None
assert out_data["missing_values"] is None
assert out_data["required"] == False # noqa
assert out_data["data_type"] == "string"
def test_validation_error():
"""
Test custom handling of validation errors
"""
in_fields = {
"fields": [
{
"Key": "person.id",
"Label": "Person ID",
# Missing description, but has required keys
},
{
"Key": "specimen.id",
"Description": "Identifier for specimen"
# Missing label but has other required keys
}
]
}
schema = FieldDefinitionsSchema()
# Test custom validation message
with pytest.raises(ValidationError) as e:
schema.load(in_fields)
errors = e.value.messages[0]
assert "fields" not in errors
assert "Field Definition [1]" in errors
assert "Field Definition [Person ID]" in errors
# Test normal validation message
with pytest.raises(ValidationError) as e:
schema.load("foo")
assert {'_schema': ['Invalid input type.']} == e.value.messages
def test_schema_load():
"""
End to end test using the field definitions schema to clean and validate
input data
"""
in_fields = {
"fields": [
{
"Key": "person.id",
"Label": "Person ID",
"Description": "Identifier for person"
},
{
"Key": "specimen.id",
"Label": "Specimen ID",
"Description": "Identifier for specimen"
}
]
}
schema = FieldDefinitionsSchema()
data = schema.load(in_fields)
out_fields = data["fields"]
# Check version
assert data["schema_version"]["number"] == schema.SCHEMA_VERSION["number"]
# Check all fields are in output
assert len(out_fields) == len(in_fields["fields"])
# Check that defaults were set right and all components of a field
# definition are present in each field definition instance
for out in out_fields:
assert set(FieldDefinitionsSchema.key_order) == set(out.keys())
assert out["data_type"] == "string"
assert out["required"] == False # noqa
assert out["accepted_values"] is None
assert out["instructions"]is None
| [
"dukedesi22@gmail.com"
] | dukedesi22@gmail.com |
9044ef4372639411577526182e2432646d7b8420 | 1b0934b52c2db1ae96b74b51efc3747989b763e1 | /contacts/serializers.py | 8d5176072ed96dfcbc4b5430efc648b411275355 | [] | no_license | swarajgaidhane15/real_estate_app | 96f7f491ed0994c5f09f3dc654e9266e43a0b0cb | c2473194dcf1d160f54ae4e26446c037290b2775 | refs/heads/master | 2023-04-12T06:30:00.560571 | 2021-05-01T13:24:12 | 2021-05-01T13:24:12 | 359,822,992 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 191 | py | from rest_framework import serializers
from .models import Contact
class ContactSerializer(serializers.ModelSerializer):
class Meta:
model = Contact
fields = "__all__"
| [
"rutvikgaidhane1508@gmail.com"
] | rutvikgaidhane1508@gmail.com |
c091fb2e440475a0d6d7fdded18bfbd31026ea4e | 85398ab1933641284edb49f77ab2c8f21411129e | /gui/visualization.py | d146d659d4919965a7f2f37ae5ea136604a5d04a | [] | no_license | olavvatne/EANN | 4dcd030fab40548175f236c80cbaa8f13b4ba3aa | dd58375256dc83237f1678506ddb4e9eb53651f9 | refs/heads/master | 2021-01-16T22:51:49.080191 | 2015-04-08T18:45:48 | 2015-04-08T18:45:48 | 32,599,494 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 10,101 | py | from tkinter import Toplevel, Button
from simulator.environment import Environment
from tkinter import *
from tkinter import ttk
from enum import Enum
from math import fabs, floor
from gui.elements import LabelledSelect
from collections import deque
#Subclass of the tkinters Canvas object. Contains methods
#for setting a graph model and drawing a graph, and changing
#the vertices' colors.
class PixelDisplay(Canvas):
cWi = 500
cHi = 500
def __init__(self, parent):
self.queue = deque([])
self.model = None
self.width = self.cWi
self.height = self.cHi
self.padding = int(self.width/64)
self.parent = parent
self.offset = 1
self.event_rate = 400
self._callback_id = None
super().__init__(parent, bg='white', width=self.width, height=self.height, highlightthickness=0)
def set_rate(self, n):
self.event_rate = n
def set_model(self, model):
self.model = model
def get_model(self):
return self.model
def draw(self):
'''
Draw will call itself and redraw (colorize nodes) as long
as the display is in running mode or there are timeslices left
in the queue. The queue of timeslices allow the algorithm to run at
full speed while the display is delaying the rendering, so it is easy to
watch it's progress
Draw will pop a timeslice from the draw queue, and
use it's data to draw the partial solution on screen.
Each cell will be assigned a color, and a arrow/point to indicate
direction the cell gives its output.
'''
if len(self.queue)>0:
timeslice = self.queue.popleft()
if timeslice:
self.draw_model(timeslice)
if not self.stopped or len(self.queue) > 0:
self._callback_id =self.after(self.event_rate, self.draw)
def colorize_item(self, item, color):
self.itemconfig(item, fill=color)
def draw_label(self, x_pos, y_pos, w, h, text,t="label", c="black"):
x = self.translate_x(x_pos)
y = self.translate_y(y_pos)
w = self.translate_y(x_pos + w)
h = self.translate_y(y_pos + h)
penalty = len(text)
font_size = 35 -penalty*2
font = ("Helvetica", font_size, "bold")
self.create_text((x+w)/2, (y+h)/2, text=text, tags=t, fill=c, font=font)
#Method for drawing a graph from a ProblemModel.
#Draws the model and add tags so individual nodes can later
#be changed.
def draw_model(self, timeslice):
pass
def start(self):
self.stopped = False
self.draw()
def stop(self):
self.after_cancel(self._callback_id)
self.stopped = True
#The actual x position of the graph element on screen
def translate_x(self, x):
self.padding = 0
x_norm = fabs(self.min_x) + x
available_width = min(self.width, self.height)
x_screen = (self.padding/2) + x_norm*(float((available_width-self.padding)/self.w))
return x_screen
#The actual y position of the graph element on screen
def translate_y(self, y):
self.padding = 0
available_height= min(self.width, self.height)
y_norm = fabs(self.min_y) + y
y_screen = (self.padding/2) + y_norm*(float((available_height-self.padding)/self.h))
return y_screen
def reset(self):
self.delete(ALL)
def set_padding(self, padding):
self.padding = padding
#draws a cell.
def draw_pixel(self, x,y, w, h, c, tag=""):
self.create_rectangle(self.translate_x(x),
self.translate_y(y),
self.translate_x(x+w),
self.translate_y(y+h),
fill=c,
tags=tag)
def draw_rounded(self, x_pos, y_pos, width, height, color, rad=5, tags="", padding=0, line="black"):
x = self.translate_x(x_pos)+padding
y = self.translate_y(y_pos)+padding
w = self.translate_x(x_pos+width)-padding
h = self.translate_y(y_pos+height)-padding
self.create_oval(x, y, x +rad, y + rad, fill=color, tag=tags, width=1, outline=line)
self.create_oval(w -rad, y, w, y + rad, fill=color, tag=tags, width=1, outline=line)
self.create_oval(x, h-rad, x +rad, h, fill=color, tag=tags, width=1, outline=line)
self.create_oval(w-rad, h-rad, w , h, fill=color, tag=tags, width=1, outline=line)
self.create_rectangle(x + (rad/2.0), y, w-(rad/2.0), h, fill=color, tag=tags, width=0)
self.create_rectangle(x , y + (rad/2.0), w, h-(rad/2.0), fill=color, tag=tags, width=0)
def set_dimension(self, max_x, max_y, min_x, min_y):
self.w = fabs(min_x) + max_x
self.h = fabs(min_y) + max_y
self.max_x = max_x
self.max_y = max_y
self.min_y = min_y
self.min_x = min_x
def set_queue(self, data):
self.queue.clear()
self.queue.extend(data)
def event(self, data):
self.queue.append(data)
class FlatlandsDisplay(PixelDisplay):
def __init__(self, parent, dim):
super().__init__(parent)
self.dim = dim
self.bg = "#bbada0"
self.empty_cell = "#ccc0b3"
self.set_dimension(self.dim, self.dim, 0, 0 )
self.draw_board()
def draw_board(self):
self.reset()
self.draw_pixel(0, 0, self.dim, self.dim, self.bg, tag="bg")
for i in range(self.dim):
for j in range(self.dim):
self.draw_rounded(i,j, 1, 1, self.empty_cell, padding=2, line=self.bg, tags="bg")
def draw_model(self, timeslice):
if timeslice:
t, x,y,dir,b = timeslice
self.delete("Piece")
for i in range(self.dim):
for j in range(self.dim):
tile = b[i][j]
if tile > 0:
self.draw_piece("Piece", j, i, tile)
self.draw_piece("Piece", x,y, 3)
self.draw_direction("Piece", x, y, dir)
self.create_text(20, 20, font=("Arial",20), text=str(t+1), fill="white", tags="Piece")
def draw_direction(self,id, tx, ty ,dir):
x = self.translate_x(tx)
y = self.translate_y(ty)
x2 = self.translate_x(tx+1)
y2 = self.translate_y(ty+1)
if dir == Environment.WEST:
self.create_line(x, (y+y2)/2,(x+x2)/2 ,(y+y2)/2, tags=id, fill="#F5C60A", width=3)
elif dir == Environment.NORTH:
self.create_line((x+x2)/2, y,(x+x2)/2 ,(y+y2)/2, tags=id, fill="#F5C60A", width=3)
elif dir == Environment.SOUTH:
self.create_line((x+x2)/2, y2,(x+x2)/2 ,(y+y2)/2, tags=id, fill="#F5C60A", width=3)
else:
self.create_line((x+x2)/2, (y+y2)/2, x2 ,(y+y2)/2, tags=id, fill="#F5C60A", width=3)
def draw_piece(self, piece_id, x, y, piece_type):
self.draw_rounded(x,y, 1, 1, self._get_color(piece_type), padding=8, line=self.bg, tags=piece_id)
#self.draw_label( x,y, 1,1, str(piece_id), t=piece_id)
def _get_color(self, type):
c = {1:"green", 2:"red", 3:"blue"}
return c.get(type)
class ResultDialog(object):
'''
The flatlands agent can be visualized by the resultDialog. The dialog consists of a pixel display, speed adjuster,
restart button, scenario list box and a new scenario button.
'''
def __init__(self, parent, individual, scenarios, config):
self.config = config
self.individual = individual
self.s = scenarios
#TODO: Generate a new scenario. But what to do for static?
self.dim = self.config["fitness"]["flatlands"]["parameters"]["grid_dimension"]
dynamic = self.config["fitness"]["flatlands"]["parameters"]["dynamic"]
if dynamic:
self.scenarios = [Environment(self.dim)]
self.current = self.scenarios[0]
else:
self.scenarios = scenarios
self.current = self.scenarios[0]
top = self.top = Toplevel(parent)
top.title("Flatlands - results")
top.grid()
self.canvas = FlatlandsDisplay(top, self.dim)
self.canvas.set_model(self.current)
self.canvas.grid(row=0, column=0, columnspan=5, sticky=N ,padx=4, pady=4)
self.v = StringVar()
speed_adjuster = Scale(top, from_=200, to=1000, command=self.set_speed,orient=HORIZONTAL, variable=self.v)
speed_adjuster.set(400)
speed_adjuster.grid(row=1, column=0,padx=4, pady=4)
self.scenario_select = LabelledSelect(top, list(range(len(self.scenarios)+1)), "Scenario", command=self.change_scenario)
self.scenario_select.grid(row=1, column=1,padx=4, pady=4)
restart_button = Button(top, text="Restart", command=self.reset)
restart_button.grid(row=1, column=4,padx=4, pady=4)
new_button = Button(top, text="New scenario", command=self.new_scenario)
new_button.grid(row=1, column=2,padx=4, pady=4)
finish_button = Button(top, text="OK", command=self.ok)
finish_button.grid(row=2, column=4,padx=4, pady=10)
self.record_agent()
def reset(self):
self.canvas.stop()
self.canvas.set_queue(self.recording)
self.canvas.start()
def set_speed(self, *args):
self.canvas.set_rate(int(self.v.get()))
def new_scenario(self):
self.scenarios.append(Environment(self.dim))
self.scenario_select.add_option(len(self.scenarios))
self.canvas.stop()
self.change_scenario(len(self.scenarios))
def change_scenario(self, picked):
print(picked)
self.current = self.scenarios[picked-1]
self.canvas.set_model(self.current)
self.canvas.stop()
self.record_agent()
def record_agent(self):
p = self.individual.phenotype_container.get_ANN()
self.current.score_agent(p, 60)
self.recording = self.current.get_recording()
print(self.recording)
self.canvas.set_queue(self.recording)
self.canvas.start()
def ok(self):
self.top.destroy() | [
"olavvatne@gmail.com"
] | olavvatne@gmail.com |
9a94ff41afbb90503d4a11767ceae7cee397f000 | 2a764b51f3c6b8d0032b31228bf3d2f9ae3d4ab7 | /firstPythonProject/globalAndLocalScopes.py | ec80d5bc97d87d29cf1be29d84620cec6fbdbcae | [] | no_license | Okles/python | c67a7007106778a3828077660846696dfa72e43d | 60eba9e5e6013be280f3b1c165580de1a60e6594 | refs/heads/master | 2020-03-16T23:17:50.907071 | 2018-06-29T14:23:45 | 2018-06-29T14:23:45 | 133,072,135 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 212 | py | def spam():
eggs = 99
bacon()
print(eggs)
#global variableName creates a global variable inside a function
def bacon():
ham = 101
eggs = 0
print(eggs)
spam()
eggs = 42
print(eggs) | [
"l.oklesinski@gmail.com"
] | l.oklesinski@gmail.com |
8b84f45fa19f09a482011666f1125ef4cea7f135 | a2cdc4b934d57da60190d250784af59d94ce6767 | /led8x8m/__init__.py | da50e3ff9e4d2345325ae3a97d5679d89cafdd90 | [
"MIT"
] | permissive | soundmaking/led8x8m | 1a9152a02e9d80b1b25200d7bd087d00b88ce5b8 | 383fe39c9e328951a25fd23298a4a4c11e8c964e | refs/heads/main | 2023-01-12T00:37:47.832057 | 2020-11-20T09:15:29 | 2020-11-20T09:15:29 | 310,889,233 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,485 | py | # led8x8m/__init__.py
import RPi.GPIO as IO
from time import sleep
class LedMatrix():
PIN_X = (12, 22, 27, 25, 17, 24, 23, 18)
PIN_Y = (21, 20, 26, 16, 19, 13, 6, 5)
matrix_buffer = [
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1, 1, 1],
]
def __init__(self):
IO.setwarnings(False)
IO.setmode(IO.BCM)
for pin_number in self.PIN_X + self.PIN_Y:
IO.setup(pin_number, IO.OUT)
def xy_on(self, x, y):
"""x and y are int 0-7"""
for index, pin in enumerate(self.PIN_X):
IO.output(pin, 1 if index == x else 0)
for index, pin in enumerate(self.PIN_Y):
IO.output(pin, 0 if index == y else 1)
def buffer_to_pins(self):
for y, row in enumerate(self.matrix_buffer):
for x, val in enumerate(row):
if val:
self.xy_on(x, y)
# sleep(0.00125)
if __name__ == '__main__':
ledmx = LedMatrix()
mode = input("\n/! choose mode, xy or buffer (x or b): ")
while mode == 'x':
print('\n/! set x and y ... ')
x = int(input('x (0-7): '))
y = int(input('y (0-7): '))
ledmx.xy_on(x, y)
while mode == 'b':
ledmx.buffer_to_pins()
print('error')
| [
"soundmaking@merfasmean.com"
] | soundmaking@merfasmean.com |
26b5977a9cdd6177a1d5b17a3b207e49b4f83eb4 | da6e2f791c8b4610775ecc4db273f2a8fc80a015 | /pixivrss.py | 4dba695605fe8a75fa939246eadbcf3bcb114dc7 | [
"CC0-1.0"
] | permissive | tsudoko/pixivrss | 5faf3e035c59bfe23fbed60498f677d75920ecca | b0b62a5f929f7365492e250bc9b1ee8c4afef5a7 | refs/heads/master | 2021-01-18T22:59:13.844762 | 2019-09-06T11:26:38 | 2019-09-06T11:26:38 | 41,524,243 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 4,465 | py | #!/usr/bin/env python3
# This work is subject to the CC0 1.0 Universal (CC0 1.0) Public Domain
# Dedication license. Its contents can be found in the LICENSE file or at:
# http://creativecommons.org/publicdomain/zero/1.0/
from datetime import datetime, timezone
from email.utils import formatdate as rfc822
from xml.sax.saxutils import escape
import argparse
import getpass
import hashlib
import os.path
import platform
import sys
import requests
CLIENT_ID = "bYGKuGVw91e0NMfPGp44euvGt59s"
CLIENT_SECRET = "HP3RmkgAmEGro0gn1x9ioawQE8WMfvLXDz3ZqxpK"
TIME_SECRET = "28c1fdd170a5204386cb1313c7077b34f83e4aaf4aa829ce78c231e05b0bae2c"
API_URL = "https://public-api.secure.pixiv.net/v1"
ILLUST_URL = "https://pixiv.net/i/{illust_id}"
THUMB_URL = "https://embed.pixiv.net/decorate.php?illust_id={}"
def get_access_token(username, password):
#timestamp = datetime.now(timezone.utc).isoformat(timespec='seconds') # python 3.6+
timestamp = datetime.now(timezone.utc).isoformat().rsplit('.')[0] + "+00:00"
timehash = hashlib.md5(timestamp.encode() + TIME_SECRET.encode()).hexdigest()
headers = {
"x-client-time": timestamp,
"x-client-hash": timehash,
}
auth = {
"username": username,
"password": password,
"grant_type": "password",
"client_id": CLIENT_ID,
"client_secret": CLIENT_SECRET,
}
r = requests.post("https://oauth.secure.pixiv.net/auth/token", data=auth, headers=headers)
r = r.json()
if "response" not in r or "access_token" not in r['response']:
raise Exception("unexpected json:\n" + str(r))
return r['response']['access_token']
def get_following(access_token):
headers = {"Authorization": "Bearer " + access_token}
r = requests.get(API_URL + "/me/following/works.json", headers=headers)
r = r.json()
if "response" not in r:
raise Exception("unexpected json:\n" + str(r))
return r['response']
def make_rss(works):
def mkdate(d):
format_string = "%Y-%m-%d %H:%M:%S %z"
r = rfc822(datetime.strptime(d + " +0900", format_string).timestamp())
return r
now = rfc822(datetime.now().timestamp())
ver = platform.python_version()
print('<?xml version="1.0"?>')
print('<rss version="2.0">')
print("<channel>")
print(" <title>[pixiv] フォロー新着作品</title>")
print(" <link>http://www.pixiv.net/bookmark_new_illust.php</link>")
print(" <pubDate>" + now + "</pubDate>")
print(" <description />")
print(" <generator>pixivrss (Python " + ver + ")</generator>")
for i in works:
title = "「%s」/「%s」" % (i['title'], i['user']['name'])
url = ILLUST_URL.format(user_id=str(i['user']['id']), illust_id=str(i['id']))
thumb_available = i['age_limit'] == "all-age"
print("\n <item>")
print(" <title>" + escape(title) + "</title>")
print(" <link>" + escape(url) + "</link>")
print(" <description><![CDATA[")
if i['caption']:
print(" " + i['caption'].replace("\r\n", "<br />"))
if thumb_available:
print(" <br />")
if thumb_available:
print(' <img src="' + escape(THUMB_URL.format(i['id'])) + '" />')
print(" ]]></description>")
print(" <pubDate>" + mkdate(i['created_time']) + "</pubDate>")
print(" <guid>" + escape(url) + "</guid>")
print(" </item>")
print("</channel>")
print("</rss>")
def main():
parser = argparse.ArgumentParser()
parser.add_argument("-u", "--username")
parser.add_argument("-p", "--password")
parser.add_argument("-n", "--unattended", action="store_true", help="don't ask for credentials")
parser.add_argument("-t", "--token", help="use the API token instead of username/password; it's generated after logging in through the API and is usually valid for an hour")
args = parser.parse_args()
if not args.unattended and not args.token:
if not args.username:
args.username = input("Pixiv ID: ")
if not args.password:
args.password = getpass.getpass("Password: ")
if not ((args.username and args.password) or args.token):
raise Exception("not enough credentials")
if not args.token:
args.token = get_access_token(args.username, args.password)
make_rss(get_following(args.token))
if __name__ == "__main__":
main()
| [
"flan@flande.re"
] | flan@flande.re |
fb34fef5a994d27aa76b19358a3979cd8df5aaf8 | 26ce347a68ce7ab060ee6d60459b0fb33a61b155 | /lesson1_4_name_spaces_and_scopes.py | 483e54fc2942f0da50f476a2922395f4ee595eb5 | [] | no_license | ekstash/stepik_python_course | d7864bdeb1bd501cf69f85702f1a0c5f48917a85 | 342b2b6b196af90a891957ac8f8300f2361451f6 | refs/heads/main | 2023-02-05T15:48:14.425728 | 2020-12-28T20:55:26 | 2020-12-28T20:55:26 | 321,792,701 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 707 | py | # большая часть урока понятна и так, но есть интересные моменты, о которых я не задумывалась.
for i in range(5):
x = i * i
print(x) # здесь x создастся как глобальная переменная.
x = 0
def function():
global x # работаем с глобальной переменной x. Именно глобальной, это не работает для функции в функции
x = 1
print(x) # x=1
def f():
x = 0
def g():
nonlocal x # возьмет x из ближайшего пространства имен, содержащего x
x = 1 | [
"ekstash@yandex.ru"
] | ekstash@yandex.ru |
df42974f8feaba118778818317b26cfb50904473 | 94c9bcf833689a7f64099531bfb67ca827eb9bf8 | /bot/modules/rclone.py | d7cf48216f67dccb84926ea1e0cadbedffd9f717 | [] | no_license | boyerharold033/ARPT-Bot | f9685140de2eafb0f5a99c9eabad276c9907621a | 644dc7c9018bb8586fcbbc27bd3dff2cccd20c93 | refs/heads/main | 2023-08-28T09:06:51.171170 | 2021-11-10T13:16:48 | 2021-11-10T13:16:48 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,776 | py | import time
import subprocess
import sys
import re
import json
import os
import threading
import requests
from config import Rclone_share,Aria2_secret
from modules.control import cal_time,only_progessbar
def hum_convert(value):
value=float(value)
units = ["B", "KB", "MB", "GB", "TB", "PB"]
size = 1024.0
for i in range(len(units)):
if (value / size) < 1:
return "%.2f%s" % (value, units[i])
value = value / size
async def start_rclonecopy(client, message):
try:
firstdir = message.text.split()[1]
seconddir= message.text.split()[2]
print(f"rclone {firstdir} {seconddir}")
sys.stdout.flush()
rc_url = f"http://root:{Aria2_secret}@127.0.0.1:5572"
info = await client.send_message(chat_id=message.chat.id, text=f"添加任务:", parse_mode='markdown')
rcd_copyfile_url = f"{rc_url}/sync/copy"
data = {
"srcFs": firstdir,
"dstFs": seconddir,
"createEmptySrcDirs": True,
"_async": True,
}
html = requests.post(url=rcd_copyfile_url, json=data)
result = html.json()
jobid = result["jobid"]
rcd_status_url = f"{rc_url}/job/status"
while requests.post(url=rcd_status_url, json={"jobid": jobid}).json()['finished'] == False:
job_status = requests.post(url=f"{rc_url}/core/stats", json={"group": f"job/{jobid}"}).json()
print(job_status)
if "transferring" in job_status:
if job_status['eta'] == None:
eta = "暂无"
else:
eta = cal_time(job_status['eta'])
print(f"剩余时间:{eta}")
text = f"任务ID:`{jobid}`\n" \
f"源地址:`{firstdir}`\n" \
f"目标地址:`{seconddir}`\n" \
f"传输部分:`{hum_convert(job_status['bytes'])}/{hum_convert(job_status['totalBytes'])}`\n" \
f"传输进度:`{only_progessbar(job_status['bytes'], job_status['totalBytes'])}%`\n" \
f"传输速度:`{hum_convert(job_status['speed'])}/s`\n" \
f"剩余时间:`{eta}`"
try:
await client.edit_message_text(text=text, chat_id=info.chat.id, message_id=info.message_id,
parse_mode='markdown')
except:
continue
else:
print("等待信息")
time.sleep(1)
requests.post(url=f"{rc_url}/core/stats-delete", json={"group": f"job/{jobid}"}).json()
requests.post(url=f"{rc_url}/fscache/clear").json()
except Exception as e:
print(f"rclonecopy :{e}")
sys.stdout.flush()
async def start_rclonecopyurl(client, message):
try:
url = message.text.split()[1]
print(f"rclonecopyurl {url} ")
sys.stdout.flush()
rc_url = f"http://root:{Aria2_secret}@127.0.0.1:5572"
Rclone_remote = os.environ.get('Remote')
Upload = os.environ.get('Upload')
title = os.path.basename(url)
info = await client.send_message(chat_id=message.chat.id, text=f"添加任务:`{title}`", parse_mode='markdown')
rcd_copyfile_url = f"{rc_url}/operations/copyurl"
data = {
"fs": f"{Rclone_remote}:{Upload}",
"remote": "",
"url": url,
"autoFilename": True,
"_async": True,
}
html = requests.post(url=rcd_copyfile_url, json=data)
result = html.json()
jobid = result["jobid"]
rcd_status_url = f"{rc_url}/job/status"
while requests.post(url=rcd_status_url, json={"jobid": jobid}).json()['finished'] == False:
job_status = requests.post(url=f"{rc_url}/core/stats", json={"group": f"job/{jobid}"}).json()
if "transferring" in job_status:
if job_status['transferring'][0]['eta'] == None:
eta = "暂无"
else:
eta = cal_time(job_status['transferring'][0]['eta'])
text = f"任务ID:`{jobid}`\n" \
f"任务名称:`{title}`\n" \
f"传输部分:`{hum_convert(job_status['transferring'][0]['bytes'])}/{hum_convert(job_status['transferring'][0]['size'])}`\n" \
f"传输进度:`{job_status['transferring'][0]['percentage']}%`\n" \
f"传输速度:`{hum_convert(job_status['transferring'][0]['speed'])}/s`\n" \
f"平均速度:`{hum_convert(job_status['transferring'][0]['speedAvg'])}/s`\n" \
f"剩余时间:`{eta}`"
try:
await client.edit_message_text(text=text, chat_id=info.chat.id, message_id=info.message_id,
parse_mode='markdown')
except:
continue
else:
print("等待信息加载")
time.sleep(1)
requests.post(url=f"{rc_url}/core/stats-delete", json={"group": f"job/{jobid}"}).json()
requests.post(url=f"{rc_url}/fscache/clear").json()
print("上传结束")
except Exception as e:
print(f"rclonecopy :{e}")
sys.stdout.flush()
async def start_rclonelsd(client, message):
try:
firstdir = message.text.split()[1]
child1 = subprocess.Popen(f'rclone lsd {firstdir}',shell=True, stdout=subprocess.PIPE)
out = child1.stdout.read()
print(out)
i = str(out,encoding='utf-8').replace(" ","")
print(i)
await client.send_message(chat_id=message.chat.id,text=f"`{str(i)}`",parse_mode='markdown')
except Exception as e:
print(f"rclonelsd :{e}")
sys.stdout.flush()
async def start_rclonels(client, message):
try:
firstdir = message.text.split()[1]
child1 = subprocess.Popen(f'rclone lsjson {firstdir}',shell=True, stdout=subprocess.PIPE)
out = child1.stdout.read()
print(out)
i = str(out,encoding='utf-8').replace("","")
print(i)
info=i.replace("[\n","").replace("\n]","")
print(info)
info_list=info.split(",\n")
print(info_list)
text=""
for a in info_list:
new=json.loads(a)
print(new)
filetime=str(new['ModTime']).replace("T"," ").replace("Z"," ")
text=text+f"{filetime}--{new['Name']}\n"
await client.send_message(chat_id=message.chat.id,text=f"`{text}`",parse_mode='markdown')
except Exception as e:
print(f"rclone :{e}")
sys.stdout.flush()
| [
"769020367@qq.com"
] | 769020367@qq.com |
dba3ee426705b4ed70feb471665fa78dfd7174d7 | 73bebf4fc3cf6b72193292c92ae2228f62facb3f | /utcnormal.py | 3468b870b7447a5982c82bb5dba4fa761bda47d6 | [] | no_license | Atari-Frosch/utcnormalise | 4bfe9720e34c2a67c790994d97b4ccc1b2f54f22 | 2fb5869dace892011a31b310ae6d61ec5b42d303 | refs/heads/master | 2016-09-11T12:43:04.417008 | 2015-04-16T17:57:49 | 2015-04-16T17:57:49 | 34,066,257 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,933 | py | #!/usr/bin/env python
# coding: utf8
import datetime
'''
This works only as long as the difference to UTC is in full hours!
'''
class date_normaliser(object):
def invert_utcdiff(self, utc):
if "+" in utc:
if utc == "+0":
utcdiff = 0
else:
utcdiff = 0 - int(utc[1:])
elif "-" in utc:
utcdiff = int(utc[1:])
else:
utcdiff = "ERROR"
return utcdiff
def check_leapyear(self, myDate):
myYear = int(myDate[0] + myDate[1] + myDate[2] + myDate[3])
leapyear = False
if myYear % 4 == 0:
leapyear = True
if myYear % 100 == 0:
leapyear = False
return leapyear
def utc_normalise(self, myTimestamp):
myTimestamp = myTimestamp.split(" ")
myDate = myTimestamp[0] + " " + myTimestamp[1]
utc = myTimestamp[2]
utcdiff = self.invert_utcdiff(utc)
myYear = myDate[0] + myDate[1] + myDate[2] + myDate[3]
myMonth = int(myDate[5] + myDate[6])
myDay = int(myDate[8] + myDate[9])
myHour = int(myDate[11] + myDate[12])
myMinute = int(myDate[14] + myDate[15])
mySecond = int(myDate[17] + myDate[18])
if utcdiff == 0:
outTimestamp = myDate
elif utcdiff > 0:
if myHour + utcdiff > 23:
myHour = 24 - myHour - utcdiff
myDay +=1
leapyear = self.check_leapyear(myDate)
if leapyear:
if myMonth == 2:
if myDay == 29:
myDay = 1
myMonth = 3
elif myDay == 28:
myDay = 29
else:
if myMonth == 2 and myDay == 28:
myDay = 1
myMonth = 3
if myMonth == 1 or myMonth == 3 or myMonth == 5 or myMonth == 7 or myMonth == 8 or myMonth == 10 or myMonth == 12:
if myDay > 31:
myDay = 1
myMonth += 1
if myMonth > 12:
myMonth = 1
myYear += 1
elif myMonth == 4 or myMonth == 6 or myMonth == 9 or myMonth == 11:
if myDay > 30:
myDay = 1
myMonth += 1
else:
myHour = myHour + utcdiff
elif utcdiff < 0:
if myHour + utcdiff < 0:
myHour = 24 - myHour + utcdiff
myDay -= 1
if myDay == 0:
if myMonth == 1 or myMonth == 2 or myMonth == 4 or myMonth == 6 or myMonth == 8 or myMonth == 9 or myMonth == 11:
myDay = 31
myMonth -= 1
if myMonth == 0:
myMonth = 1
myYear -= 1
elif myMonth == 5 or myMonth == 7 or myMonth == 10 or myMonth == 12:
myDay = 30
myMonth -= 1
elif myMonth == 3:
leapyear = self.check_leapyear(myDate)
myMonth = 2
if leapyear:
myDay = 29
else:
myDay = 28
else:
myHour = myHour + utcdiff
myYear = str(myYear)
myMonth = str(myMonth)
if len(myMonth) == 1:
myMonth = "0" + myMonth
myDay = str(myDay)
if len(myDay) == 1:
myDay = "0" + myDay
myHour = str(myHour)
if len(myHour) == 1:
myHour = "0" + myHour
myMinute = str(myMinute)
if len(myMinute) == 1:
myMinute = "0" + myMinute
mySecond = str(mySecond)
if len(mySecond) == 1:
mySecond = "0" + mySecond
outTimestamp = myYear + "-" + myMonth + "-" + myDay + " " + myHour + ":" + myMinute + ":" + mySecond + " +0"
return outTimestamp
myNormaliser = date_normaliser()
sourcefile = "spammail.lst"
source = open(sourcefile, "r")
fulltext = source.read()
source.close()
myLines = fulltext.split("\n")
targetfile = "spammail-utc.lst"
target = open(targetfile, "w")
for i in range(len(myLines)):
# tmpout = "myLines[" + str(i) + "] = " + myLines[i]
# print tmpout
if myLines[i] != "":
currentLine = myLines[i].split(" ")
myTimestamp = currentLine[0] + " " + currentLine[1] + " " + currentLine[2].replace(",", "")
outTimestamp = myNormaliser.utc_normalise(myTimestamp)
outLine = outTimestamp + ", " + currentLine[3] + " " + currentLine[4] + " " + currentLine[5] + " " + currentLine[6] + "\n"
target.write(outLine)
target.close()
| [
"frosch@atari-frosch.de"
] | frosch@atari-frosch.de |
7e8548d8a6de1e92bfe9ae3a868ce13ea025c259 | cc4f5f736adef13c5d259b4d061681bdef238576 | /great_expectations/expectations/metrics/column_aggregate_metrics/column_min.py | 7b82733a813ca0aafa863bd643df5dce22d26e90 | [
"Apache-2.0"
] | permissive | chsigjan/great_expectations | 3f133271abd24c4ea1d9d6b8cfada83fd58f520a | 56d048f4cca073154ce67ca9696a75917ad8c81c | refs/heads/main | 2023-08-12T08:41:52.709139 | 2021-09-30T15:51:40 | 2021-09-30T15:51:40 | 412,717,483 | 1 | 0 | Apache-2.0 | 2021-10-02T06:59:22 | 2021-10-02T06:59:21 | null | UTF-8 | Python | false | false | 1,054 | py | from great_expectations.execution_engine import (
PandasExecutionEngine,
SparkDFExecutionEngine,
)
from great_expectations.execution_engine.sqlalchemy_execution_engine import (
SqlAlchemyExecutionEngine,
)
from great_expectations.expectations.metrics.column_aggregate_metric_provider import (
ColumnAggregateMetricProvider,
column_aggregate_partial,
column_aggregate_value,
)
from great_expectations.expectations.metrics.column_aggregate_metric_provider import (
sa as sa,
)
from great_expectations.expectations.metrics.import_manager import F
class ColumnMin(ColumnAggregateMetricProvider):
metric_name = "column.min"
@column_aggregate_value(engine=PandasExecutionEngine)
def _pandas(cls, column, **kwargs):
return column.min()
@column_aggregate_partial(engine=SqlAlchemyExecutionEngine)
def _sqlalchemy(cls, column, **kwargs):
return sa.func.min(column)
@column_aggregate_partial(engine=SparkDFExecutionEngine)
def _spark(cls, column, **kwargs):
return F.min(column)
| [
"noreply@github.com"
] | noreply@github.com |
e695d0df27a389029524c11afe920770636319d2 | d6d2b61aeb44b3e9fbed8fd9839f1e8abaaf74a8 | /tests/unit/peapods/runtimes/container/test_container_runtime.py | 42143d96d76a7c4656cd9f57f01a92854cdd3430 | [
"Apache-2.0"
] | permissive | 10zinten/jina | b9776b8e28fc546e7eb205b67857ebbf56bcc51d | f18b04eb82d18a3c554e2892bbae4b95fc0cb13e | refs/heads/master | 2023-08-26T22:01:58.578591 | 2021-10-25T23:06:49 | 2021-10-25T23:06:49 | 400,064,091 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,029 | py | import os
import time
from sys import platform
import multiprocessing
import pytest
from jina.checker import NetworkChecker
from jina.executors import BaseExecutor
from jina.executors.decorators import requests
from jina import Flow, __windows__
from jina.helper import random_name
from jina.parsers import set_pea_parser
from jina.parsers.ping import set_ping_parser
from jina.peapods import Pea
from jina.peapods.runtimes.container import ContainerRuntime
from jina.peapods.runtimes.container.helper import get_gpu_device_requests
from tests import random_docs, validate_callback
if __windows__:
pytest.skip(msg='Windows containers are not supported yet', allow_module_level=True)
cur_dir = os.path.dirname(os.path.abspath(__file__))
img_name = 'jina/mwu-encoder'
defaulthost = '0.0.0.0'
@pytest.fixture
def _logforward():
class _LogForward(BaseExecutor):
@requests
def foo(self, **kwargs):
pass
return _LogForward
@pytest.fixture(scope='module')
def docker_image_built():
import docker
client = docker.from_env()
client.images.build(path=os.path.join(cur_dir, 'mwu-encoder/'), tag=img_name)
client.close()
yield
time.sleep(2)
client = docker.from_env()
client.containers.prune()
def test_simple_container(docker_image_built):
args = set_pea_parser().parse_args(['--uses', f'docker://{img_name}'])
with Pea(args):
pass
time.sleep(2)
Pea(args).start().close()
def test_flow_with_one_container_pod(docker_image_built):
f = Flow().add(name='dummyEncoder1', uses=f'docker://{img_name}')
with f:
f.post(on='/index', inputs=random_docs(10))
def test_flow_with_one_container_pod_shards(docker_image_built):
f = Flow().add(name='dummyEncoder1', shards=2, uses=f'docker://{img_name}')
with f:
pod = f._pod_nodes['dummyEncoder1']
assert pod.args.shards == pod.args.parallel == 2
for idx, shard in enumerate(pod.shards):
assert shard.args.pea_id == shard.args.shard_id == idx
assert shard.args.shards == shard.args.parallel == 2
f.post(on='/index', inputs=random_docs(10))
def test_flow_with_replica_container_ext_yaml(docker_image_built):
f = Flow().add(
name='dummyEncoder3',
uses=f'docker://{img_name}',
shards=3,
entrypoint='jina pea',
)
with f:
f.post(on='/index', inputs=random_docs(10))
f.post(on='/index', inputs=random_docs(10))
f.post(on='/index', inputs=random_docs(10))
def test_flow_topo1(docker_image_built):
f = (
Flow()
.add(
name='d0',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
)
.add(
name='d1',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
)
.add(
name='d2',
uses='docker://jinaai/jina:test-pip',
needs='d0',
entrypoint='jina executor',
)
.join(['d2', 'd1'])
)
with f:
f.post(on='/index', inputs=random_docs(10))
def test_flow_topo_mixed(docker_image_built, _logforward):
f = (
Flow()
.add(
name='d4',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
)
.add(name='d5', uses=_logforward)
.add(
name='d6',
uses='docker://jinaai/jina:test-pip',
needs='d4',
entrypoint='jina executor',
)
.join(['d6', 'd5'])
)
with f:
f.post(on='/index', inputs=random_docs(10))
def test_flow_topo_shards():
f = (
Flow()
.add(
name='d7',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
shards=3,
)
.add(name='d8', shards=3)
.add(
name='d9',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
needs='d7',
)
.join(['d9', 'd8'])
)
with f:
f.post(on='/index', inputs=random_docs(10))
def test_flow_topo_ldl_shards():
f = (
Flow()
.add(name='d10')
.add(
name='d11',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
shards=3,
)
.add(name='d12')
)
with f:
f.post(on='/index', inputs=random_docs(10))
def test_container_ping(docker_image_built):
a4 = set_pea_parser().parse_args(['--uses', f'docker://{img_name}'])
a5 = set_ping_parser().parse_args(
['0.0.0.0', str(a4.port_ctrl), '--print-response']
)
# test with container
with pytest.raises(SystemExit) as cm:
with Pea(a4):
NetworkChecker(a5)
assert cm.value.code == 0
def test_tail_host_docker2local_shards():
f = (
Flow()
.add(
name='d10',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
shards=3,
)
.add(name='d11')
)
with f:
assert getattr(f._pod_nodes['d10'].tail_args, 'host_out') == defaulthost
def test_tail_host_docker2local():
f = (
Flow()
.add(
name='d12',
uses='docker://jinaai/jina:test-pip',
entrypoint='jina executor',
)
.add(name='d13')
)
with f:
assert getattr(f._pod_nodes['d12'].tail_args, 'host_out') == defaulthost
def test_pass_arbitrary_kwargs(monkeypatch, mocker):
import docker
mock = mocker.Mock()
mocker.patch(
'jina.peapods.runtimes.container.ContainerRuntime.is_ready',
return_value=True,
)
class MockContainers:
class MockContainer:
def reload(self):
pass
def logs(self, **kwargs):
return []
def __init__(self):
pass
def get(self, *args):
pass
def run(self, *args, **kwargs):
mock_kwargs = {k: kwargs[k] for k in ['hello', 'environment']}
mock(**mock_kwargs)
assert 'ports' in kwargs
assert 'environment' in kwargs
assert kwargs['environment'] == ['VAR1=BAR', 'VAR2=FOO']
assert 'hello' in kwargs
assert kwargs['hello'] == 0
return MockContainers.MockContainer()
class MockClient:
def __init__(self, *args, **kwargs):
pass
def close(self):
pass
def version(self):
return {'Version': '20.0.1'}
@property
def networks(self):
return {'bridge': None}
@property
def containers(self):
return MockContainers()
@property
def images(self):
return {}
monkeypatch.setattr(docker, 'from_env', MockClient)
args = set_pea_parser().parse_args(
[
'--uses',
'docker://jinahub/pod',
'--docker-kwargs',
'hello: 0',
'environment: ["VAR1=BAR", "VAR2=FOO"]',
]
)
_ = ContainerRuntime(args, ctrl_addr='', ready_event=multiprocessing.Event())
expected_args = {'hello': 0, 'environment': ['VAR1=BAR', 'VAR2=FOO']}
mock.assert_called_with(**expected_args)
def test_pass_arbitrary_kwargs_from_yaml():
f = Flow.load_config(os.path.join(cur_dir, 'flow.yml'))
assert f._pod_nodes['executor1'].args.docker_kwargs == {
'hello': 0,
'environment': ['VAR1=BAR', 'VAR2=FOO'],
}
def test_container_override_params(docker_image_built, tmpdir, mocker):
def validate_response(resp):
assert len(resp.docs) > 0
for doc in resp.docs:
assert doc.tags['greetings'] == 'overriden greetings'
mock = mocker.Mock()
abc_path = os.path.join(tmpdir, 'abc')
f = Flow().add(
name=random_name(),
uses=f'docker://{img_name}',
volumes=abc_path + ':' + '/mapped/here/abc',
uses_with={'greetings': 'overriden greetings'},
uses_metas={
'name': 'ext-mwu-encoder',
'workspace': '/mapped/here/abc',
},
)
with f:
f.index(random_docs(10), on_done=mock)
assert os.path.exists(
os.path.join(abc_path, 'ext-mwu-encoder', '0', 'ext-mwu-encoder.bin')
)
validate_callback(mock, validate_response)
def test_container_volume(docker_image_built, tmpdir):
abc_path = os.path.join(tmpdir, 'abc')
f = Flow().add(
name=random_name(),
uses=f'docker://{img_name}',
volumes=abc_path + ':' + '/mapped/here/abc',
uses_metas={
'name': 'ext-mwu-encoder',
'workspace': '/mapped/here/abc',
},
)
with f:
f.index(random_docs(10))
assert os.path.exists(
os.path.join(abc_path, 'ext-mwu-encoder', '0', 'ext-mwu-encoder.bin')
)
@pytest.mark.parametrize(
(
'gpus_value',
'expected_count',
'expected_device',
'expected_driver',
'expected_capabilities',
),
[
('all', -1, [], '', [['gpu']]), # all gpus
('2', 2, [], '', [['gpu']]), # use two gpus
(
'device=GPU-fake-gpu-id',
0,
['GPU-fake-gpu-id'],
'',
[['gpu']],
), # gpu by one device id
(
'device=GPU-fake-gpu-id1,device=GPU-fake-gpu-id2',
0,
['GPU-fake-gpu-id1', 'GPU-fake-gpu-id2'],
'',
[['gpu']],
), # gpu by 2 device id
(
'device=GPU-fake-gpu-id,driver=nvidia,capabilities=utility,capabilities=display',
0,
['GPU-fake-gpu-id'],
'nvidia',
[['gpu', 'utility', 'display']],
), # gpu with id, driver and capability
(
'device=GPU-fake-gpu-id1,device=GPU-fake-gpu-id2,driver=nvidia,capabilities=utility',
0,
['GPU-fake-gpu-id1', 'GPU-fake-gpu-id2'],
'nvidia',
[['gpu', 'utility']],
), # multiple ids
],
)
def test_gpu_container(
gpus_value, expected_count, expected_device, expected_driver, expected_capabilities
):
args = set_pea_parser().parse_args(
['--uses', f'docker://{img_name}', '--gpus', gpus_value]
)
device_requests = get_gpu_device_requests(args.gpus)
assert device_requests[0]['Count'] == expected_count
assert device_requests[0]['DeviceIDs'] == expected_device
assert device_requests[0]['Driver'] == expected_driver
assert device_requests[0]['Capabilities'] == expected_capabilities
def test_pass_native_arg(monkeypatch, mocker):
import docker
mocker.patch(
'jina.peapods.runtimes.container.ContainerRuntime.is_ready',
return_value=True,
)
class MockContainers:
class MockContainer:
def reload(self):
pass
def logs(self, **kwargs):
return []
def __init__(self):
pass
def get(self, *args):
pass
def run(self, *args, **kwargs):
assert '--native' in args[1]
return MockContainers.MockContainer()
class MockClient:
def __init__(self, *args, **kwargs):
pass
def close(self):
pass
def version(self):
return {'Version': '20.0.1'}
@property
def networks(self):
return {'bridge': None}
@property
def containers(self):
return MockContainers()
@property
def images(self):
return {}
monkeypatch.setattr(docker, 'from_env', MockClient)
args = set_pea_parser().parse_args(
[
'--uses',
'docker://jinahub/pod',
]
)
_ = ContainerRuntime(args, ctrl_addr='', ready_event=multiprocessing.Event())
| [
"noreply@github.com"
] | noreply@github.com |
fb4a0f6047e8b49dd942360dcb017c97f5fe04f5 | a4b2e0612d765429e4dccd037daa3763eb54c12c | /code/model/comment.py | 92ee39692d72148be2a45e0a16ac3b0f2eb6de33 | [] | no_license | evertheylen/SmartHome | b98238ed43d1a39e08ae7e8f7deaea93505384db | 542468e7ba03f81be765a876797cdd0c7094f22b | refs/heads/master | 2020-04-10T19:09:14.047037 | 2016-05-19T11:08:59 | 2016-05-19T11:08:59 | 51,506,014 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 466 | py |
from sparrow import *
from .owentity import *
from .status import Status
from .user import User
class Comment(RTOwEntity):
key = CID = KeyProperty()
status = RTReference(Status)
author = RTReference(User)
date = Property(int)
date_edited = Property(int) # If they are the same, no edits
text = Property(str)
async def can_delete(self, usr_uid, db):
u = await User.find_by_key(self.author, db)
return u.key == usr_uid
| [
"hermans.anthony@hotmail.com"
] | hermans.anthony@hotmail.com |
845e06146026e7a00fd10824220dd35e50e2ccab | 127d8c209b00978f4f660534363e95eca3f514f2 | /backend/home/migrations/0002_load_initial_data.py | 110b21901b630cf3f96ad807523e091bfc8ac157 | [] | no_license | crowdbotics-apps/sitespace-19938 | afd070e64d32ab455f9b2b05e376152e9e28e5ad | 416b5cef0bdb25018ec3b634bf3096e61fe8b662 | refs/heads/master | 2022-12-10T15:52:58.601025 | 2020-09-02T15:20:15 | 2020-09-02T15:20:15 | 292,319,517 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,290 | py | from django.db import migrations
def create_customtext(apps, schema_editor):
CustomText = apps.get_model("home", "CustomText")
customtext_title = "Sitespace"
CustomText.objects.create(title=customtext_title)
def create_homepage(apps, schema_editor):
HomePage = apps.get_model("home", "HomePage")
homepage_body = """
<h1 class="display-4 text-center">Sitespace</h1>
<p class="lead">
This is the sample application created and deployed from the Crowdbotics app.
You can view list of packages selected for this application below.
</p>"""
HomePage.objects.create(body=homepage_body)
def create_site(apps, schema_editor):
Site = apps.get_model("sites", "Site")
custom_domain = "sitespace-19938.botics.co"
site_params = {
"name": "Sitespace",
}
if custom_domain:
site_params["domain"] = custom_domain
Site.objects.update_or_create(defaults=site_params, id=1)
class Migration(migrations.Migration):
dependencies = [
("home", "0001_initial"),
("sites", "0002_alter_domain_unique"),
]
operations = [
migrations.RunPython(create_customtext),
migrations.RunPython(create_homepage),
migrations.RunPython(create_site),
]
| [
"team@crowdbotics.com"
] | team@crowdbotics.com |
82dd296e5b4516cc622e3a206cb39c14e931ec3c | 84123de245cbea1c3d299421fec03401b27c274c | /Lecture 8 programming exercise/lecture8.3.largest.py | 43be0d2b64ebb00699f40518cbcb854c80297170 | [] | no_license | udanzo-p/Lectures38 | ae43801f6770a6f05a4fa7ef4bc57b597181e992 | 65128508931573517f4ef850d699ea95c780258c | refs/heads/master | 2020-12-12T21:31:29.949256 | 2020-02-04T08:54:41 | 2020-02-04T08:54:41 | 234,233,096 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 275 | py | import math
k=int(input())
target=int(input())
print("To find largest digit no. divisible by ",target)
pow=math.pow
a=(pow(10,k)-1)
while(int x=target):
if(a%target=0):
print("The largest no. divisible by is ",target,a)
else:
a=a-1
x=x-1
| [
"noreply@github.com"
] | noreply@github.com |
ba508a2958f5325258855671103405bc641ebe97 | a5e591dc09e11e88af56fb5a881fae064fb9c495 | /recruitment/recruitment/doctype/interview/interview.py | 0449ed7ff48f9261f3c429e7522f6aad25c3b49d | [
"MIT"
] | permissive | barathprathosh/recruitment | 6b61dd1ee9c0b9d7851b0b3e5bab307f7ee2d1b5 | 9660944856e72288e47960e6802ec97a220a656d | refs/heads/master | 2020-04-29T03:03:51.722972 | 2019-03-15T08:58:32 | 2019-03-15T08:58:32 | 175,794,797 | 0 | 0 | NOASSERTION | 2019-03-15T10:00:32 | 2019-03-15T10:00:31 | null | UTF-8 | Python | false | false | 250 | py | # -*- coding: utf-8 -*-
# Copyright (c) 2015, VHRS and contributors
# For license information, please see license.txt
from __future__ import unicode_literals
import frappe
from frappe.model.document import Document
class Interview(Document):
pass
| [
"abdulla.pi@voltechgroup.com"
] | abdulla.pi@voltechgroup.com |
b63af80e07a7db51998cd62a6b59a52a243d8216 | ec0fe7fa7951a42dcf8fb9e3a910cc78ea58ab3d | /aeroctf/solution.py | 56610d813c8ea411df6bbbdb4ac618819f5cd4ca | [
"MIT"
] | permissive | S7uXN37/CTF-Writeups | 6111f70f0d811734cd8b4fcc41c73a0a4d582dc4 | ffbbae7e3c98bf47786d9996e114412325117a3a | refs/heads/master | 2021-02-08T18:59:25.115539 | 2020-04-20T16:04:36 | 2020-04-20T16:04:36 | 244,186,440 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,643 | py | from pwn import *
from binascii import *
from time import sleep
from z3 import *
ciphers = [
"7685737a9f7895737a9f84857b769f7a" + "657b769f78898378",
"717785747885858d6f7e917364686776",
"7393a992708d8fad708d83aa7273707d" + "6f3939856b7d398bb53b8b34b573b6c5618e7135"
]
plains = [
"test_test_test_t" + "est_test",
"qwertyuiopasdfgh",
"skIllaoInasJjklq" + "o19akq9k13k45k69alq1"
]
secret = "8185748f7b3b3a3565454584b8babbb8" + "b441323ebc8b3a86b5899283b9c2c56d64388889b781"
#for i in range(len(ciphers)):
# ciphers[i] = unhexlify(ciphers[i])
keysize = 16
# allowed key chars: 0-9, a-h
solver = Solver()
key = [BitVec("key" + str(i), 8) for i in range(keysize)]
for (p,c) in zip(plains, ciphers[:3]):
for pos in range(len(p)):
solver.add(int(c[2*pos:2*pos+2], 16) == (key[pos % 16] + (key[pos % 16] ^ ord(p[pos]))) & 0xff)
# constrain to aero{...}
start = "Aero{"
for i in range(len(start)):
v = int(secret[2*i:2*i+2], 16)
dec = ((v - key[i]) & 0xff) ^ key[i]
dec = dec & 0xff
solver.add( dec == ord(start[i]))
print("Solving...")
# Solve the equations
while solver.check() == sat:
modl = solver.model()
keyVals = [modl[key[i]].as_long() for i in range(keysize)]
#print("Found key: " + str(keyVals))
flag = [int(secret[2*pos:2*pos+2], 16) for pos in range(len(secret) // 2)]
flag = [((flag[i] - keyVals[i % 16]) & 0xff) ^ keyVals[i % 16] for i in range(len(flag))]
f = ""
for i in range(len(flag)):
f += chr(flag[i])
full_ascii = True
for c in f:
if ord(c) < 0x20 or ord(c) > 0x80:
full_ascii = False
if full_ascii:
print(f)
# Encryption works like:
# output byte = key[mod 16] + key[mod 16] ^ plain byte | [
"S7uXN37@users.noreply.github.com"
] | S7uXN37@users.noreply.github.com |
268ab70d2790ee69fd6584adc601abb2cb0fafbc | 05487c52248d7b185fadd7824d709c23c4877e43 | /modules/GuiWidgetManager.py | 291596dc304137d9d45439efd37e177776cb3ed0 | [
"MIT"
] | permissive | reality3d/molefusion | 6b3751bcb0c4350eccced2a2ffa1f3190cd610b0 | df6e3486a412777117aa63d68d169102065051fd | refs/heads/master | 2016-08-04T04:58:38.925897 | 2013-09-02T19:42:20 | 2013-09-02T19:42:20 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,027 | py | import pygame
from pygame.locals import *
from GW_TextInput import GW_TextInput
from Event import Event
from Constants import Constants
class GuiWidgetManager:
"Widget Manager Container"
def __init__(self , widgetlist):
"Sets up the widget Manager"
self.background=Constants.BACKGROUND
self.screen=Constants.SCREEN
self.widgets = pygame.sprite.RenderUpdates() #RenderUpdates Sprite Group type
self.widgetlist = widgetlist #internal list for indexable access
for widget in widgetlist:
self.widgets.add(widget)
self.currentfocus=-1 #by default no widgets focused
self.draw=True
def set_draw(self,value):
self.draw=value
def get_widgets(self):
return self.widgets
def has_input_focus(self):
return isinstance(self.currentfocus,GW_TextInput)
def run(self):
for event in pygame.event.get(): #Process events
if event.type == MOUSEMOTION:
for widget in self.widgetlist:
if(widget.get_active()): #Is active or disabled?
if(widget.get_rect().collidepoint(pygame.mouse.get_pos()[0],pygame.mouse.get_pos()[1])):
if(not widget.get_mouse_over()): #Was already mouse over it?
widget.notify(Event("onmouseover",event))
#print "onmouseover"
elif widget.get_mouse_over(): #Widget was mouseover and mouse has gone out
widget.notify(Event("onmouseout",event))
#print "onmouseout"
elif event.type == MOUSEBUTTONDOWN and event.button==1: #Left Mouse click
for widget in self.widgetlist:
if(widget.get_active()):
if(widget.get_rect().collidepoint(pygame.mouse.get_pos()[0],pygame.mouse.get_pos()[1])):
widget.notify(Event("onmouseclick",event))
#print "onmouseclick"
if(not widget.get_focus()):
widget.notify(Event("onfocus",event))
self.currentfocus=widget
#print "onfocus"
elif widget.get_focus():
widget.notify(Event("onblur",event))
if(widget==self.currentfocus):#We could have put the focus already on another widget!
self.currentfocus=-1
#print "onblur"
elif event.type == MOUSEBUTTONUP and event.button==1:
for widget in self.widgetlist:
if(widget.get_active()):
if(widget.get_rect().collidepoint(pygame.mouse.get_pos()[0],pygame.mouse.get_pos()[1])):
widget.notify(Event("onmouseclickup",event))
elif event.type == KEYDOWN:
if(self.currentfocus!=-1):
if(self.currentfocus.get_active()):
if(isinstance(self.currentfocus,GW_TextInput)):
if(event.key!=K_ESCAPE and event.key!=K_RETURN):
self.currentfocus.notify(Event("onkeydown",event))
else:
self.currentfocus.notify(Event("onblur",event))
self.currentfocus=-1
else:
pygame.event.post(event) #Reinject the event into the queue for maybe latter process
if self.draw==True:
self.widgets.clear(self.screen,self.background)
self.widgets.update()
self.rectlist = self.widgets.draw(self.screen)
pygame.display.update(self.rectlist)
| [
"reality3d@gmail.com"
] | reality3d@gmail.com |
7a80572cd829ac7a970a7e869f590fbca51c840e | b410a506bd4bdbbc55a770ab76e0507625ebb52e | /app/admin/forms.py | 651179a851d4b8765a8b9efd99cd96820ecce34e | [] | no_license | fadebowaley/MO-App | 6a219a984c89c2dd0601a6f3f3d46810b644dcb3 | b7de965390d69e349533765db0ac190e8a67684b | refs/heads/main | 2023-08-24T06:17:17.510741 | 2021-09-14T12:31:30 | 2021-09-14T12:31:30 | 403,637,924 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 694 | py | from flask_wtf import FlaskForm
from wtforms import Form, FieldList, StringField, StringField, SubmitField, BooleanField, \
PasswordField, TextAreaField, SubmitField, FormField, SelectField,\
FileField, IntegerField, TextField, FieldList
from wtforms.validators import DataRequired, Length,ValidationError, InputRequired
from flask_wtf.file import FileField, FileAllowed
from wtforms.fields.html5 import DateTimeField, DateField
from wtforms.ext.sqlalchemy.fields import QuerySelectField
from wtforms.validators import ValidationError, DataRequired, Length
#from flask_babel import _, lazy_gettext as _l
from flask_login import current_user, login_required
from app.models import User
| [
"fadebowaley@gmail.com"
] | fadebowaley@gmail.com |
1b97ca333f9999257c9f67250e6f7ea46ba7df58 | 605936a6e51061b25dd04b688ddab41f10c24153 | /Easter Competition.py | 9bacdd0f037d3c209eb241e3a89588453b2e855f | [] | no_license | EmanuilManev/Python | d8e16f8c48ef55d9ef6063eb7702fea94b5078f5 | 8f0f454ebba08857c3bfcd76db75e6dd3bb6de84 | refs/heads/master | 2021-04-11T02:36:55.146516 | 2020-03-23T14:08:47 | 2020-03-23T14:08:47 | 248,986,319 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 533 | py | kozunak = int(input())
index = 1
vhod = str("")
score = 0
m_score = 0
top_chef = str("")
for index in range(kozunak):
chef = input()
while vhod != "Stop":
vhod = input()
if vhod != "Stop":
score += int(vhod)
print(str(chef) + " has " + str(score) + " points.")
if score > m_score:
m_score = score
top_chef = chef
print(str(chef) + " is the new number 1!")
vhod = str("")
score = 0
print(str(top_chef) + " won competition with " + str(m_score) + " points!")
| [
"noreply@github.com"
] | noreply@github.com |
3d033f53f447828b966051b34b7913e92aa9b31b | 772a1fad7ff949c41920004579166fcab58af7a5 | /app/main/views.py | 198cfee8dab5b104352c62526b6aa218495243ed | [
"MIT"
] | permissive | saudahabib/news_highlight | c29a461ba185be23fd0a183239af245557f11d6c | 6d59b85b93497e14e988be2bf9e0fb358d871622 | refs/heads/master | 2020-05-09T10:02:56.993083 | 2019-04-17T08:00:00 | 2019-04-17T08:00:00 | 181,026,997 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 924 | py | from flask import render_template
from . import main
from ..request import get_sources, get_source, get_articles
from ..models import Article, Source
#Views
@main.route('/')
def index():
'''
View root page that returns the index page and its data
'''
sources = get_sources()
articles = get_articles('kenya')
title = "All the spice, under one roof"
return render_template('index.html', title = title, sources = sources, articles = articles)
@main.route('/news/<int:news_id>')
def news(news_id):
'''
View news page function that returns page with news
'''
return render_template('news.html', id = news_id)
@main.route('/source/<int:id>')
def source(id):
'''
View source page function that returns the source details page and its data
'''
source = get_source(id)
name = f'{source.name}'
return render_template('news.html',name = name, source = source)
| [
"saudababs00@gmail.com"
] | saudababs00@gmail.com |
19caab41b1e7e5822d71d8e70217b1ac6dda3b67 | 847273de4b1d814fab8b19dc651c651c2d342ede | /.history/Sudoku_II_005_20180620141234.py | 396d0dea7f396d2fdc9165bfceb7cd75b20f3c37 | [] | no_license | Los4U/sudoku_in_python | 0ba55850afcffeac4170321651620f3c89448b45 | 7d470604962a43da3fc3e5edce6f718076197d32 | refs/heads/master | 2020-03-22T08:10:13.939424 | 2018-07-04T17:21:13 | 2018-07-04T17:21:13 | 139,749,483 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 4,622 | py | from random import randint
sudoku1 = [
[5, 9, 8, 6, 1, 2, 3, 4, 7],
[2, 1, 7, 9, 3, 4, 8, 6, 5],
[6, 4, 3, 5, 8, 7, 1, 2, 9],
[1, 6, 5, 4, 9, 8, 2, 7, 3],
[3, 2, 9, 7, 6, 5, 4, 1, 8],
[7, 8, 4, 3, 2, 1, 5, 9, 6],
[8, 3, 1, 2, 7, 6, 9, 5, 4],
[4, 7, 2, 8, 5, 9, 6, 3, 1],
[9, 5, 6, 1, 4, 3, 7, 8, " "]
]
sudoku2 = [
[9, 8, 7, 4, 3, 2, 5, 6, 1],
[2, 4, 3, 5, 1, 6, 8, 7, 9],
[5, 6, 1, 7, 9, 8, 4, 3, 2],
[3, 9, 5, 6, 4, 7, 2, 1, 8],
[8, 2, 4, 3, 5, 1, 6, 9, 7],
[1, 7, 6, 2, 8, 9, 3, 4, 5],
[7, 1, 2, 8, 6, 3, 9, 5, 4],
[4, 3, 8, 9, 7, 5, 1, 2, 6],
[' ', 5, ' ', ' ', 2, ' ', 7, ' ', ' ']
]
sudoku3 = [
[9, 8, 7, 4, 3, 2, 5, 6, 1],
[2, 4, 3, 5, 1, 6, 8, 7, 9],
[5, 6, 1, 7, 9, 8, 4, 3, 2],
[3, 9, 5, 6, 4, 7, 2, 1, 8],
[8, 2, 4, 3, 5, 1, 6, 9, 7],
[1, 7, 6, 2, 8, 9, 3, 4, 5],
[7, 1, 2, 8, 6, 3, 9, 5, 4],
[4, 3, 8, 9, 7, 5, 1, 2, 6],
[' ', 5, ' ', ' ', 2, ' ', 7, ' ', ' ']
]
def printSudoku():
i = 0
while i < 10:
if i == 0:
print(" 1 2 3 4 5 6 7 8 9")
print(" -------------------------")
elif i == 3 or i == 6 or i == 9:
print(" -------------------------")
line = "|"
if i < 9:
print('{2} {1} {0[0]} {0[1]} {0[2]} {1} {0[3]} {0[4]} {0[5]} {1} {0[6]} {0[7]} {0[8]} {1}'.format(sudoku[i], line, i+1))
i = i + 1
print(" ")
print(" %@@@@@@@ @@@ @@@ (@@@@@@@@@ ,@@@@2@@@@@ @@@, /@@@/ @@@, @@@ ")
print(" @@@* @@@ @@@ (@@( /@@@# .@@@% (@@@ @@@, @@@% @@@, @@@. ")
print(" @@@& @@@ @@@ (@@( @@@* @@@% #@@% @@@,.@@@. @@@, @@@. ")
print(" ,@@@@@@* @@@ @@@ (@@( (@@% .@@@* ,@@@ @@@%@@% @@@, @@@. ")
print(" /@@@@@# @@@ @@@ (@@( (@@% .@@@* ,@@@ @@@,@@@( @@@, @@@. ")
print(" *@@@. @@@ .@@& (@@( @@@. @@@% &@@( @@@, &@@@. @@@* .@@@. ")
print(" &, &@@@ #@@@. ,@@@, (@@( ,&@@@* ,@@@& .@@@@ @@@, (@@@/ #@@@* @@@# ")
print(",@@@@@@@@( (@@@@@@@@% (@@@@@@@@@( #@@@@@@@@@, @@@, ,@@@% ,@@@@@@@@@. \n ")
print("To start game input:")
print(" r - to load random puzzle:")
print(" 1 - to load chart nr 1:")
print(" 2 - to load chart nr 2:")
print(" 3 - to load chart nr 3:")
choice = input("Input here: ")
s = 0
if choice == "R" or choice == "r":
listaSudoku = [sudoku1, sudoku2, sudoku3]
sudoku_number = randint(0, 2)
print("dupa", sudoku_number)
sudoku = listaSudoku[sudoku_number]
#print("ktore = ", sudoku)
elif int(choice) == 1:
s = 1
sudoku = sudoku1
elif int(choice) == 2:
s = 2
sudoku = sudoku2
elif int(choice) == 3:
s = 3
sudoku = sudoku3
while True: # prints Sudoku until is solved
print("Your sudoku to solve:")
printSudoku()
print("Input 3 numbers in format a b c, np. 4 5 8")
print(" a - row number")
print(" b - column number ")
print(" c - value")
# vprint(" r - reset chart to start\n ")
x = input("Input a b c: ")
print("")
numbers = " 0123456789" # conditions of entering the numbers !
if (len(x) != 5) or (str(x[0]) not in numbers) or (str(x[2]) not in numbers) or (
str(x[4]) not in numbers) or (str(x[1]) != " ") or (str(x[3]) != " "):
if x == "r": # reset
if s == 1 :
sudoku = sudoku1
elif s == 1 :
sudoku = sudoku1
s == 1 :
sudoku = sudoku1
print(" Function reset() will be ready in Next Week")
else:
print("Error - wrong number format \n ")
continue
sudoku[int(x[0])-1][int(x[2])-1] = int(x[4])
column1 = 0
column2 = 0
try:
i = 0
list = []
while i < 9:
column = 0
for item in sudoku:
column = column + item[i]
list.append(column)
#p rint(list)
# print("Suma columny ", i, " = ", column)
i += 1
is45 = 0
for listElement in list:
if listElement == 45:
is45 = is45 + 1
# print("Ile kolumen OK", is45)
i = 0
for item in sudoku:
if sum(item) == 45 and is45 == 9:
i = i + 1
if i == 9:
printSudoku()
print("@@@@@@@@@@ YOU WIN @@@@@@@@@@")
break
except TypeError:
print()
| [
"inz.kamil.wos@gmail.com"
] | inz.kamil.wos@gmail.com |
cac8cca8bbafc756a771cbbd21f316a640e98cd7 | 6b4a48fb6142789326654c48d32acda3eb5e7b08 | /formationproject/wsgi.py | a9ea3c0a982ffb7af95cba5e2211d90796a89dd1 | [] | no_license | mwesterhof/formationproject | 0d9795c218b5010bfbb716216d3d8f4fa5bd4799 | 1b4a057b996829609e308c78721aca840ec58ee7 | refs/heads/master | 2023-08-19T00:08:58.282341 | 2021-10-08T16:19:18 | 2021-10-08T16:19:18 | 401,425,998 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 413 | py | """
WSGI config for formationproject project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/3.2/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "formationproject.settings.dev")
application = get_wsgi_application()
| [
"m.westerhof@lukkien.com"
] | m.westerhof@lukkien.com |
b236c8c4e623cf11c90bb5c8d15b7e0df7aba00b | 7d66ec851a0bbba1403848e2caf7323d89797936 | /abra/inference/__init__.py | 40fcd85e690de6a5a0222958aa03e7e2ddf5acce | [
"MIT"
] | permissive | seekshreyas/abracadabra | d19aceb9de2dbbcfb7feb3c3e14ccb1c4c164095 | a79f03bba2917ab4cc99d4c3ec0459401842ee9f | refs/heads/master | 2022-10-05T07:41:21.163255 | 2020-06-12T17:43:04 | 2020-06-12T17:43:04 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 899 | py | from abra.inference.inference_base import InferenceProcedure, FrequentistProcedure
def get_inference_procedure(method, **infer_params):
_method = method.lower().replace('-', '').replace('_', '').replace(' ', '')
if _method in ('meansdelta'):
from abra import MeansDelta as IP
elif _method in ('proportionsdelta'):
from abra import ProportionsDelta as IP
elif _method in ('ratesratio'):
from abra import RatesRatio as IP
elif method in (
'gaussian',
'bernoulli',
'binomial',
'beta_binomial',
'gamma_poisson'
):
from abra import BayesianDelta as IP
infer_params.update({"model_name": method})
else:
raise ValueError('Unknown inference method {!r}'.format(method))
return IP(method=method, **infer_params)
__all__ = [
"InferenceProcedure",
"FrequentistProcedure"
]
| [
"dustin@quizlet.com"
] | dustin@quizlet.com |
89b1685f529264b86004c272eb59419b27a1315b | 4a42fefd8945c73402ddf36f8943e011cd9c4151 | /projects/myhellowebapp/hellowebapp/wsgi.py | 2b6fe00b6b8875c39ed849cf147b0eb94f51d25b | [] | no_license | momentum-cohort-2018-10/hello-web-app-SowmyaAji | c2c1374b460232822ff91fc1d034f1d89a400332 | 2cfe7fd6d22db4f9b9ac0d8fdc611787cb1372c5 | refs/heads/master | 2020-04-06T11:53:35.991478 | 2018-11-18T20:48:49 | 2018-11-18T20:48:49 | 157,434,877 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 399 | py | """
WSGI config for hellowebapp project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/2.0/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "hellowebapp.settings")
application = get_wsgi_application()
| [
"sowmya.aji@gmail.com"
] | sowmya.aji@gmail.com |
bc87958ada6e6684a8aae3abd9d9e9d963c1b983 | 92f5057eeab3b5fb6952f478917bd84858907588 | /socialbayes.py | 0d959c7bb67bcedaecf9d8530d00f1e641ff07e9 | [] | no_license | enjoylife/mlsocial | 6e16c8f5fd836e34bbca1192b1b94cf1e2914f8a | 0ac877e4ffa013cd075ee65a09247b8a65cb8185 | refs/heads/master | 2021-01-22T23:33:40.081502 | 2012-02-18T06:16:16 | 2012-02-18T06:16:16 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 11,261 | py | # -*- coding: UTF-8 -*-
import Stemmer
#twitter_search
import requests
from requests import RequestException
from requests import async
from urllib import urlencode
from urllib2 import urlopen
import json
from multiprocessing import Process
from thread import Thread
#redisbayes
import re
import math
import redis
import random
from redis import WatchError
stemmer = Stemmer.Stemmer('english')
english_ignore = set()
with open('stoplist.txt', 'r') as stops:
for word in stops:
english_ignore.add(stemmer.stemWord(word.strip()))
class DataConsumer(Process):
""" Consumer process that will extract data given a Joinable queue """
def __init__(self, q, stoplist, stemmer, data):
Process.__init__(self)
self.input_q = q
self.stemmer = stemmer
self.stoplist = stoplist
self.data = data
def run(self):
while True:
data = self.input_q.get()
# Replace with real extraction work later
print data
self.input_q.task_done()
class RedisExtract(object):
def __init__(self, redis=None, prefix='data:'):
if not redis: import redis; self.redis = redis.Redis()
else: self.redis = redis
self.prefix = prefix
def flush(self):
for cat in self.redis.smembers(self.prefix + 'categories'):
self.rpipe.delete(self.prefix + cat)
self.rpipe.delete(self.prefix + 'categories')
self.rpipe.execute()
def train(self, category, text):
self.redis.sadd(self.prefix + 'categories', category)
## TODO create a super class that has a whole set of text extraction
#tools
def tidy(text):
if not isinstance(text, basestring):
text = str(text)
if not isinstance(text, unicode):
text = text.decode('utf8')
text = text.lower()
return re.sub(r'[\_.,<>:;~+|\[\]?`"!@#$%^&*()\s]', ' ', text, re.UNICODE)
def english_tokenizer(text, stem=True, stops=english_ignore):
if stem:
words = stemmer.stemWords(tidy(text).split())
else:
words = text.split()
return [w for w in words if len(w) > 2 and w not in stops ]
def occurances(words):
counts = {}
for word in words:
if word in counts:
counts[word] += 1
else:
counts[word] = 1
return counts
class RedisBayes(object):
u""" Naïve Bayesian Text Classifier on Redis.
redisbayes
~~~~~~~~~~
I wrote this to filter spammy comments from a high traffic forum website
and it worked pretty well. It can work for you too :)
For example::
>>> rb = RedisBayes(redis.Redis(), prefix='bayes:test:')
>>> rb.flush()
>>> rb.classify('nothing trained yet') is None
True
>>> rb.train('good', 'sunshine drugs love sex lobster sloth')
>>> rb.train('bad', 'fear death horror government zombie god')
>>> rb.classify('sloths are so cute i love them')
'good'
>>> rb.classify('i fear god and love the government')
'bad'
>>> int(rb.score('i fear god and love the government')['bad'])
-9
>>> int(rb.score('i fear god and love the government')['good'])
-14
>>> rb.untrain('good', 'sunshine drugs love sex lobster sloth')
>>> rb.untrain('bad', 'fear death horror government zombie god')
>>> rb.score('lolcat')
{}
Words are lowercased and unicode is supported::
>>> print tidy(english_tokenizer("Æther")[0])
æther
Common english words and 1-2 character words are ignored::
>>> english_tokenizer("greetings mary a b aa bb")
[u'mari']
Some characters are removed::
>>> print english_tokenizer("contraction's")[0]
contract
>>> print english_tokenizer("what|is|goth")[0]
goth
"""
def __init__(self, redis=None, prefix='bayes:', correction=0.1,
tokenizer=None):
self.redis = redis
self.prefix = prefix
self.correction = correction
self.tokenizer = tokenizer or english_tokenizer
if not self.redis:
import redis
redis = redis.Redis()
self.rpipe = redis.pipeline()
def flush(self):
for cat in self.redis.smembers(self.prefix + 'categories'):
self.rpipe.delete(self.prefix + cat)
self.rpipe.delete(self.prefix + 'categories')
self.rpipe.execute()
def train(self, category, text, twitdata=False, multi=False):
self.redis.sadd(self.prefix + 'categories', category)
words = self.tokenizer(text)
for word, count in occurances(words).iteritems():
self.rpipe.hincrby(self.prefix + category, word, count)
self.rpipe.execute()
if twitdata and not multi:
words = set(self.tokenizer(text, stem=False))
sample = random.sample(words, 2)
tweets = twitter_search(sample, twit_pages=5)
data = self.tokenizer(tweets,stops=(words|english_ignore))
for word, count in occurances(data).iteritems():
self.rpipe.hincrby(self.prefix + category, word, count)
self.rpipe.execute()
if multi:
words = set(self.tokenizer(text, stem=False))
sample = random.sample(words, 2)
twitter_search_async(sample, twit_pages=5)
def untrain(self, category, text):
for word, count in occurances(self.tokenizer(text)).iteritems():
cur = self.redis.hget(self.prefix + category, word)
if cur:
new = int(cur) - count
if new > 0:
self.redis.hset(self.prefix + category, word, new)
else:
self.redis.hdel(self.prefix + category, word)
with self.rpipe:
while 1:
try:
self.rpipe.watch(self.prefix + category)
if self.rpipe.hlen(self.prefix + category) == 0:
self.rpipe.multi()
self.rpipe.delete(self.prefix + category)
self.rpipe.srem(self.prefix + 'categories', category)
self.rpipe.execute()
break
except WatchError:
continue
def classify(self, text):
score = self.score(text)
if not score:
return None
return sorted(score.iteritems(), key=lambda v: v[1])[-1][0]#hackish?
def guess(self, text):
score = self.score(text)
if not score:
return None
values = sorted(score.iteritems(), key=lambda v: v[1], reverse=True)#hackish?
for key, value in values:
print key
def score(self, text):
occurs = occurances(self.tokenizer(text))
scores = {}
for category in self.redis.smembers(self.prefix + 'categories'):
tally = self.tally(category)
scores[category] = 0.0
for word, count in occurs.iteritems():
score = self.redis.hget(self.prefix + category, word)
assert not score or score > 0, "corrupt bayesian database"
score = score or self.correction
scores[category] += math.log(float(score) / tally)
return scores
def tally(self, category):
tally = sum(int(x) for x in self.redis.hvals(self.prefix + category))
assert tally >= 0, "corrupt bayesian database" #TODO better error check
return tally
def twitter_search(words, twit_pages):
"""input: list of words, returns a list of results
Filtering out the multiple retweets and named entities
TODO: Error checking and better func params"""
sentences=[]
pagenum = twit_pages
twit_base = "http://search.twitter.com/search.json?"
for page in range(1,pagenum+1):
twit_ops = {'lang':'en', 'result_type':'mixed','include_entities':1,'page':page, 'rpp':100,
'q': ' '.join(words)}
try:
r =requests.get(twit_base+urllib.urlencode(twit_ops), prefetch=True,
timeout=.4)
content = json.loads(r.text)
except RequestException:
continue
print "adding new sentences for page %s" % page
for sents in content['results']:
# Dont want retweets
if not sents['text'].startswith('RT'):
new_sent = sents['text']
# Test if their is urls in string, if so remove them before appending
if sents['entities'].get('urls', False):
for urls in sents['entities'].get('urls'):
i = urls['indices']
# Hackish with indices, but it works :)
new_sent=sents['text'].replace(sents['text'][i[0]:i[1]],'')
sentences.append(new_sent.encode('utf-8'))
return sentences
red = redis.Redis()
rp = red.pipeline()
def twitter_search_async(words, twit_pages):
"""input: list of words, returns a list of results
Filtering out the multiple retweets and named entities
TODO: Error checking and better func params"""
urls=[]
pagenum = twit_pages
twit_base = "http://search.twitter.com/search.json?"
for page in range(1,pagenum+1):
twit_ops = {'lang':'en', 'result_type':'mixed','include_entities':1,'page':page, 'rpp':100,
'q': ' '.join(words)}
urls.append(async.get(twit_base+urllib.urlencode(twit_ops),
timeout=.4, hooks={'response': parse_tweets}))
async.map(urls)
def parse_tweets(data):
sentences=[]
tweets = json.load(data.text)
for sents in tweets['results']:
print decoding
# Dont want retweets
if not sents['text'].startswith('RT'):
new_sent = sents['text']
# Test if their is urls in string, if so remove them before appending
if sents['entities'].get('urls', False):
for urls in sents['entities'].get('urls'):
i = urls['indices']
# Hackish with indices, but it works :)
new_sent=sents['text'].replace(sents['text'][i[0]:i[1]],'')
sentences.append(new_sent.encode('utf-8'))
for word, count in occurances(sentences).iteritems():
rp.hincrby(self.prefix + category, word, count)
rp.execute()
if __name__ =='__main__':
#import doctest
#doctest.testmod()
#rb = RedisBayes(redis.Redis(), prefix="bayes:test:")
#rb.train('food', ' pizza apples yogurt water' , True,)
#rb.train('sports', ' baseball hockey football soccer' , True)
#rb.train('sports', ' bat glove goal field rink hoop helmet ' , True)
#rb.train('school', ' paper pencil laptop hw' , True)
#rb.train('book', 'read table of contents paper cover', True )
#rb.train('car', 'ford nissan gmc wheel headlight ', True)
# rb.train('math', ' subtract times multiple divide log ', True , True)
# print rb.classify('sit down at the table')
# print rb.classify(' i read')
# print rb.classify('puck')
# print rb.classify('honda')
# print rb.classify('division')
| [
"mclemens66+github@gmail.com"
] | mclemens66+github@gmail.com |
63b940320659fb0fc6204694ab667047cfce35e3 | 7f957ab6a6a0d9b6d21c126134aad956eaad8fb0 | /mod/data_encoding/__init__.py | b7d241a4e2f630278f21584a37a01f09889618bf | [
"MIT"
] | permissive | wangludewdrop/algorithm-tools | a3fc96118c38306e017812c2b93fd930985c50f8 | 094fd0c8587efbead704ff1b49e06fb7a00468b7 | refs/heads/main | 2023-04-12T18:39:50.240765 | 2021-05-06T10:32:56 | 2021-05-06T10:32:56 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 168 | py | # -*- coding: utf-8 -*-
'''
Created on 2021/02/27 18:02:32
@File -> __init__.py
@Author: luolei
@Email: dreisteine262@163.com
@Describe: 初始化
'''
__all__ = [] | [
"dreisteine@stu.scu.edu.cn"
] | dreisteine@stu.scu.edu.cn |
b9e4db8d80d9a37b23a0d85cebf5894410c987cd | 3b97c64fcee27cb19faab732ff1b7d71347881c8 | /Consumption_Prediction/Codes/Data_Processing.py | 40e615d8149cb51b73ce5732771b5c9319d8fb1c | [] | no_license | pedrobranco0410/E4C-Forecast | 43a6048ac2520401fab152103b971090f2864bdb | 5807bd8f4053f18298387cf596cc88f518b0e402 | refs/heads/master | 2023-07-02T07:29:40.076176 | 2021-08-06T09:44:46 | 2021-08-06T09:44:46 | 375,277,477 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,500 | py | import numpy as np
import pandas as pd
from sklearn.preprocessing import MinMaxScaler
def read_csv(path, interval):
"""
Reads the csv file in the corresponding path and returns a dataframe with all data rearranged in desired intervals
"""
df = pd.read_csv(path, usecols=[0,5, 6,7,8,21,24], index_col=0, parse_dates=True)
df = df.fillna(method='ffill')
df = df.resample(interval).mean()
df = df.fillna(method='ffill')
return df
def feature_and_targets(data):
"""
It removes from the dataset the features that will be used in the prediction model and the data that must be predicted so that we can validate the model.
"""
data['day of the week'] = data.index.dayofweek
data['day of the year'] = data.index.dayofyear
data['hour of the day'] = data.index.hour
data['minute of the hour'] = data.index.minute
data["Consumption"] = data['T1']+data['T2']+data['T3']+data['T4']
#data["Consumption"] = data["TGBT"]
features = ['day of the week','day of the year','hour of the day','minute of the hour', 'AirTemp','rh']#, 'wd', 'ws','rh', 'rain']
labels = ["Consumption"]
inputs = features + labels
data = data[inputs]
return data, features,labels,len(features), len(labels)
def normalize(data):
"""
Normalizes the dataset individually for each column between -1 and 1
"""
scaler = MinMaxScaler(feature_range=(-1, 1))
data_scaled = pd.DataFrame(scaler.fit_transform(data.values), columns=data.columns, index=data.index)
return data_scaled,scaler
def split_data(data, sequence_length, features, labels, test_size=0.25):
"""
splits data to training and testing parts
"""
#Cut the dataset into 2 parts: the first will be for training and the second for validation
ntest = int(round(len(data) * (1 - test_size)))
df_train, df_test = data.iloc[:ntest], data.iloc[ntest:]
#Separates the data between the features that will be used and the results that should be predicted
x_train = np.asarray(df_train[features].iloc[:-sequence_length])
x_test = np.asarray(df_test[features].iloc[:-sequence_length])
y_test = np.asarray(df_test[labels].iloc[sequence_length:])
y_train = np.asarray(df_train[labels].iloc[sequence_length:])
return x_train, x_test, y_test, y_train,df_train, df_test
def batch_generator(batch_size, sequence_length, num_features, num_labels, x, y):
"""
Generator function for creating random batches of training-data.
"""
while True:
# Allocate a new array for the batch of input-signals.
x_shape = (batch_size, sequence_length, num_features)
x_batch = np.zeros(shape=x_shape, dtype=np.float16)
# Allocate a new array for the batch of output-signals.
y_shape = (batch_size, sequence_length, num_labels)
y_batch = np.zeros(shape=y_shape, dtype=np.float16)
# Fill the batch with random sequences of data.
for i in range(batch_size):
# Get a random start-index.
# This points somewhere into the training-data.
if len(x)<sequence_length:
print("there will be a problem test too short", len(x))
idx = np.random.randint(len(x) - 2*sequence_length)
# Copy the sequences of data starting at this index.
x_batch[i] = x[idx:idx+sequence_length]
y_batch[i] = y[idx:idx+sequence_length]
yield (x_batch, y_batch) | [
"pedrobrancondrade@gmail.com"
] | pedrobrancondrade@gmail.com |
433dc5780c6bf966236e507e8947e87df83870a2 | 43e900f11e2b230cdc0b2e48007d40294fefd87a | /Amazon/VideoOnsite/926.flip-string-to-monotone-increasing.py | d4efde64ddbe2e4540f93d5acfa3516e947730ab | [] | no_license | DarkAlexWang/leetcode | 02f2ed993688c34d3ce8f95d81b3e36a53ca002f | 89142297559af20cf990a8e40975811b4be36955 | refs/heads/master | 2023-01-07T13:01:19.598427 | 2022-12-28T19:00:19 | 2022-12-28T19:00:19 | 232,729,581 | 3 | 1 | null | null | null | null | UTF-8 | Python | false | false | 472 | py | #
# @lc app=leetcode id=926 lang=python3
#
# [926] Flip String to Monotone Increasing
#
# @lc code=start
class Solution:
def minFlipsMonoIncr(self, s: str) -> int:
n = len(s)
cnt0 = s.count('0')
cnt1 = 0
res = n - cnt0
for i in range(n):
if s[i] == '0':
cnt0 -= 1
elif s[i] == '1':
res = min(res, cnt1 + cnt0)
cnt1 += 1
return res
# @lc code=end
| [
"wangzhihuan0815@gmail.com"
] | wangzhihuan0815@gmail.com |
fc3a2dd07ead6d429cbfb7267d00260f4065db16 | e6a8e129c14c641072645f55dad5248d67a6b7ef | /Project/backend/accounts/urls.py | c6ba47a7d390ed7facf66cbfd465f6f50c374589 | [] | no_license | Dongock/aiosk | 6fcb3f5b5d360f9cead3aaf3bac07a615b5b86f2 | ca761ad258938214ceee09c42b6152c9d8a50198 | refs/heads/master | 2023-02-10T00:05:23.172411 | 2020-12-21T03:53:45 | 2020-12-21T03:53:45 | 323,223,272 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 182 | py | from django.urls import path
from . import views
urlpatterns = [
path('coupon/', views.coupon),
path('class5/', views.class5),
# path('coupon/use/', views.use_coupon),
] | [
"okdong23@naver.com"
] | okdong23@naver.com |
533417aa0ac2c08e47b14885a90b9b910347f2d1 | e9fc61eff5ef4f73dd7cb810dff10ea71d402b16 | /find_sum.py | b85dc25e48b59b5d12f74d2ded3c0f7ca6149455 | [] | no_license | syasuhiro2019/08_if_and_for | 7720d7e4d9f06bcba012bb7379ef2eed03bd92c5 | 092b0515f9a0541e86ef712e1e54281b5d585242 | refs/heads/master | 2020-04-27T20:31:15.285801 | 2019-03-09T08:34:36 | 2019-03-09T08:34:36 | 174,661,095 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 305 | py | numbers = [34, 432, 1, 99]
total = 0
total = total + numbers[0]
total = total + numbers[1]
total = total + numbers[2]
total = total + numbers[3]
print(total)
total = 0
for number in numbers:
total = total + number
total = 0
print(total)
for number in numbers:
total += number
print(total)
| [
"s_ychm@yahoo.co.jp"
] | s_ychm@yahoo.co.jp |
2e309a5345ef496934edd7635284f37e09219c00 | 8c67f4d3df0dcb9c94420239f0e5dca0d054371c | /functions/soccer_results.py | e91217277e9cbeeef99ff1c877d380f73614a4d2 | [] | no_license | Sanusi1997/python_power_of_computing_exercises | 49ea74ac83d513f4703a9026176882db49e24810 | cd1626b5ab6d5d5e88a7eecf8c723965a4caf5ef | refs/heads/master | 2021-01-14T23:40:20.245877 | 2020-07-03T20:15:16 | 2020-07-03T20:15:16 | 242,799,419 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 615 | py | def soccer_scores(first_team_scores, second_team_scores ):
""" A function that takes two integer arguments that represents a
soccer result and prints the winner from the score """
if first_team_scores > second_team_scores:
print(f"Final score is {first_team_scores}:{second_team_scores} \nTeam one won the match")
elif first_team_scores == second_team_scores:
print(f"Final score is {first_team_scores}:{second_team_scores} \nMatch ended in a draw")
else:
print(f"Final score is {second_team_scores}:{first_team_scores} \nTeam two won the match")
soccer_scores(3,2) | [
"sanusihameedolayiwola@gmail.com"
] | sanusihameedolayiwola@gmail.com |
bd3316480a59c494040f990aefb5deb06677a1bf | e44301fb87ecce8defe68fdaa5f540753e8dbccf | /server/app/services/base/views/file_system.py | c951f50d81399432afdfbde5d532b90b72bbe4df | [
"Apache-2.0"
] | permissive | dlsurainflow/actorcloud_BE | d2766a6937ca87a3fc77de849d43938abd3276e9 | c7eff034185d6f39c408f2b6c75c138c7febdf47 | refs/heads/master | 2023-06-16T19:53:46.188814 | 2020-11-11T07:00:02 | 2020-11-11T07:00:02 | 297,188,795 | 0 | 0 | Apache-2.0 | 2021-07-19T07:39:45 | 2020-09-21T00:25:36 | Vue | UTF-8 | Python | false | false | 2,327 | py | import hashlib
import time
from flask import jsonify, g, request, send_from_directory, current_app
from flask_uploads import UploadNotAllowed
from actor_libs.decorators import limit_upload_file
from actor_libs.errors import APIException, ParameterInvalid
from app import auth, images, packages
from app.models import UploadInfo
from . import bp
@bp.route('/download')
def download_file():
file_type = request.args.get('fileType', None, type=str)
filename = request.args.get('filename', None, type=str)
download_path = {
'template': 'DOWNLOAD_TEMPLATE_EXCEL_DEST',
'export_excel': 'EXPORT_EXCEL_PATH',
'image': 'UPLOADED_IMAGES_DEST',
'package': 'UPLOADED_PACKAGES_DEST'
}
if not file_type:
raise ParameterInvalid(field='fileType')
if not filename:
raise ParameterInvalid(field='filename')
path = current_app.config.get(download_path.get(file_type))
return send_from_directory(path, filename)
@bp.route('/upload', methods=['POST'])
@auth.login_required(permission_required=False)
@limit_upload_file()
def upload_file():
file_type = request.args.get('fileType', None, type=str)
file_type_dict = {
'package': {
'type': 1,
'upload_set': packages
},
'image': {
'type': 2,
'upload_set': images
}
}
if file_type not in file_type_dict.keys():
raise ParameterInvalid(field='fileType')
try:
unique_name = hashlib.md5((str(g.user_id) + str(time.time())).encode()).hexdigest()
upload_set = file_type_dict.get(file_type).get('upload_set')
request_file = request.files.get('file')
file_name = upload_set.save(request_file, name=unique_name + '.')
file_url = '/api/v1/download?fileType=%s&filename=%s' % (file_type, file_name)
except UploadNotAllowed:
raise APIException()
request_dict = {
'fileName': file_name,
'displayName': request_file.filename,
'userIntID': g.user_id,
'fileType': file_type_dict.get(file_type).get('type')
}
upload_info = UploadInfo()
created_upload = upload_info.create(request_dict)
return jsonify({
'name': created_upload.displayName,
'url': file_url,
'uploadID': created_upload.id
}), 201
| [
"wilyfreddie@github.com"
] | wilyfreddie@github.com |
a9e8d88a96e19be6e971f475c10c84ebd1e981f2 | cb6e5c8a91dce5911afbbbb7a8a4b55bc0c7687e | /scripts/SuperRod/superrod_GUI_pyqtgraph_chemlab.py | a439ead1fc9fcc5dd2d90dc7e86ea2982ea0dcae | [] | no_license | jackey-qiu/DaFy_P23 | 3870d4e436b0e9df7f1dcb747caaf38589274f92 | ad2ca8e16e92935233e84c2d9fe2b59f4f114444 | refs/heads/master | 2022-04-10T18:32:24.392046 | 2020-03-22T18:22:46 | 2020-03-22T18:22:46 | 198,180,139 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 26,159 | py | import sys,os
from PyQt5.QtWidgets import QApplication, QMainWindow, QFileDialog
from PyQt5 import uic
import random
import numpy as np
import pandas as pd
import types
import matplotlib.pyplot as plt
try:
from . import locate_path
except:
import locate_path
script_path = locate_path.module_path_locator()
DaFy_path = os.path.dirname(os.path.dirname(script_path))
sys.path.append(DaFy_path)
sys.path.append(os.path.join(DaFy_path,'dump_files'))
sys.path.append(os.path.join(DaFy_path,'EnginePool'))
sys.path.append(os.path.join(DaFy_path,'FilterPool'))
sys.path.append(os.path.join(DaFy_path,'util'))
from fom_funcs import *
import parameters
import data_superrod as data
import model
import solvergui
import time
import matplotlib
matplotlib.use("Qt5Agg")
import pyqtgraph as pg
import pyqtgraph.exporters
from PyQt5 import QtCore
from PyQt5.QtWidgets import QCheckBox, QRadioButton, QTableWidgetItem, QHeaderView, QAbstractItemView
from PyQt5.QtCore import Qt, QTimer
from PyQt5.QtGui import QTransform, QFont, QBrush, QColor
from pyqtgraph.Qt import QtGui
import syntax_pars
from chemlab.graphics.renderers import AtomRenderer
from chemlab.db import ChemlabDB
from PyQt5.QtOpenGL import *
from superrod_new import *
#from matplotlib.backends.backend_qt5agg import (NavigationToolbar2QT as NavigationToolbar)
class RunFit(QtCore.QObject):
updateplot = QtCore.pyqtSignal(str,object)
def __init__(self,solver):
super(RunFit, self).__init__()
self.solver = solver
self.running = True
def run(self):
if self.running:
self.solver.optimizer.stop = False
self.solver.StartFit(self.updateplot)
def stop(self):
self.solver.optimizer.stop = True
class MyMainWindow(QMainWindow, Ui_MainWindow):
def __init__(self, parent = None):
super(MyMainWindow, self).__init__(parent)
self.setupUi(self)
context = QGLContext(QGLFormat())
self.widget_edp = MSViewer(context,self.tab_4)
context.makeCurrent()
self.widget_edp.setObjectName("widget_edp")
self.horizontalLayout_9.addWidget(self.widget_edp)
pg.setConfigOptions(imageAxisOrder='row-major')
pg.mkQApp()
#uic.loadUi(os.path.join(DaFy_path,'scripts','SuperRod','superrod_new.ui'),self)
self.setWindowTitle('Data analysis factory: CTR data modeling')
self.stop = False
self.show_checkBox_list = []
#set fom_func
#self.fom_func = chi2bars_2
#parameters
#self.parameters = parameters.Parameters()
#scripts
#self.script = ''
#script module
#self.script_module = types.ModuleType('genx_script_module')
self.model = model.Model()
# self.solver = solvergui.SolverController(self)
self.run_fit = RunFit(solvergui.SolverController(self.model))
self.fit_thread = QtCore.QThread()
self.structure_view_thread = QtCore.QThread()
self.widget_edp.moveToThread(self.structure_view_thread)
self.run_fit.moveToThread(self.fit_thread)
self.run_fit.updateplot.connect(self.update_plot_data_view_upon_simulation)
self.run_fit.updateplot.connect(self.update_par_during_fit)
self.run_fit.updateplot.connect(self.update_status)
#self.run_fit.updateplot.connect(self.update_structure_view)
# self.run_fit.updateplot.connect(self.start_timer_structure_view)
self.fit_thread.started.connect(self.run_fit.run)
#tool bar buttons to operate modeling
self.actionNew.triggered.connect(self.init_new_model)
self.actionOpen.triggered.connect(self.open_model)
self.actionSave.triggered.connect(self.save_model)
self.actionSimulate.triggered.connect(self.simulate_model)
self.actionRun.triggered.connect(self.run_model)
self.actionStop.triggered.connect(self.stop_model)
#pushbuttons for data handeling
self.pushButton_load_data.clicked.connect(self.load_data_ctr)
self.pushButton_append_data.clicked.connect(self.append_data)
self.pushButton_delete_data.clicked.connect(self.delete_data)
self.pushButton_save_data.clicked.connect(self.save_data)
self.pushButton_calculate.clicked.connect(self.calculate)
#pushbutton for changing plotting style
self.pushButton_plot_style.clicked.connect(self.change_plot_style)
#pushbutton to load/save script
self.pushButton_load_script.clicked.connect(self.load_script)
self.pushButton_save_script.clicked.connect(self.save_script)
#pushbutton to load/save parameter file
self.pushButton_load_table.clicked.connect(self.load_par)
self.pushButton_save_table.clicked.connect(self.save_par)
#select dataset in the viewer
self.comboBox_dataset.activated.connect(self.update_data_view)
#syntax highlight
self.plainTextEdit_script.setStyleSheet("""QPlainTextEdit{
font-family:'Consolas';
font-size:11pt;
color: #ccc;
background-color: #2b2b2b;}""")
self.plainTextEdit_script.setTabStopWidth(self.plainTextEdit_script.fontMetrics().width(' ')*4)
#self.data = data.DataList()
#table view for parameters set to selecting row basis
#self.tableWidget_pars.itemChanged.connect(self.update_par_upon_change)
self.tableWidget_pars.setSelectionBehavior(QAbstractItemView.SelectRows)
self.timer_save_data = QtCore.QTimer(self)
self.timer_save_data.timeout.connect(self.save_model)
self.timer_update_structure = QtCore.QTimer(self)
self.timer_update_structure.timeout.connect(self.update_structure_view)
self.setup_plot()
print(self.widget_edp.update)
def setup_plot(self):
self.selected_data_profile = self.widget_data.addPlot()
self.fom_evolution_profile = self.widget_fom.addPlot()
self.par_profile = self.widget_pars.addPlot()
self.fom_scan_profile = self.widget_fom_scan.addPlot()
# water = ChemlabDB().get('molecule', 'example.water')
# ar = AtomRenderer(self.widget_edp, water.r_array, water.type_array)
# ar = self.widget_edp.renderers.append(AtomRenderer(self.widget_edp, water.r_array, water.type_array))
#self.widget_edp.setup_view()
def update(self):
super(MyMainWindow, self).update()
self.widget_edp.update()
def update_plot_data_view(self):
plot_data_index = []
for i in range(len(self.model.data)):
if self.tableWidget_data.cellWidget(i,1).isChecked():
# self.selected_data_profile.plot(self.data[i].x, self.data[i].y, clear = True)
self.selected_data_profile.plot(self.model.data[i].x, self.model.data[i].y,pen={'color': 'y', 'width': 1}, symbolBrush=(255,0,0), symbolSize=5,symbolPen='w', clear = (len(plot_data_index) == 0))
plot_data_index.append(i)
self.selected_data_profile.setLogMode(x=False,y=True)
self.selected_data_profile.autoRange()
def update_plot_data_view_upon_simulation(self):
plot_data_index = []
for i in range(len(self.model.data)):
if self.tableWidget_data.cellWidget(i,1).isChecked():
# self.selected_data_profile.plot(self.data[i].x, self.data[i].y, clear = True)
self.selected_data_profile.plot(self.model.data[i].x, self.model.data[i].y,pen={'color': 'y', 'width': 0}, symbolBrush=(255,0,0), symbolSize=5,symbolPen='w', clear = (len(plot_data_index) == 0))
self.selected_data_profile.plot(self.model.data[i].x, self.model.data[i].y_sim,pen={'color': 'r', 'width': 2}, clear = False)
plot_data_index.append(i)
self.selected_data_profile.setLogMode(x=False,y=True)
self.selected_data_profile.autoRange()
fom_log = np.array(self.run_fit.solver.optimizer.fom_log)
#print(fom_log)
self.fom_evolution_profile.plot(fom_log[:,0],fom_log[:,1],pen={'color': 'r', 'width': 2}, clear = True)
self.fom_evolution_profile.autoRange()
def update_plot(self):
pass
def init_new_model(self):
pass
def open_model(self):
options = QFileDialog.Options()
options |= QFileDialog.DontUseNativeDialog
fileName, _ = QFileDialog.getOpenFileName(self,"QFileDialog.getOpenFileName()", "","rod file (*.rod);;zip Files (*.rar)", options=options)
if fileName:
self.model.load(fileName)
self.update_table_widget_data()
self.update_combo_box_dataset()
self.update_plot_data_view()
self.update_par_upon_load()
self.update_script_upon_load()
def save_model(self):
path, _ = QFileDialog.getSaveFileName(self, "Save file", "", "rod file (*.rod);zip files (*.rar)")
if path:
self.model.script = (self.plainTextEdit_script.toPlainText())
self.model.save(path)
def simulate_model(self):
# self.update_par_upon_change()
self.model.script = (self.plainTextEdit_script.toPlainText())
self.model.simulate()
'''
self.compile_script()
# self.update_pars()
(funcs, vals) = self.get_sim_pars()
# Set the parameter values in the model
#[func(val) for func,val in zip(funcs, vals)]
i = 0
for func, val in zip(funcs,vals):
try:
func(val)
except Exception as e:
(sfuncs_tmp, vals_tmp) = self.parameters.get_sim_pars()
raise ParameterError(sfuncs_tmp[i], i, str(e), 1)
i += 1
self.evaluate_sim_func()
'''
self.update_plot_data_view_upon_simulation()
self.init_structure_view()
def run_model(self):
# self.solver.StartFit()
self.start_timer_structure_view()
self.structure_view_thread.start()
self.fit_thread.start()
def stop_model(self):
self.run_fit.stop()
self.fit_thread.terminate()
self.stop_timer_structure_view()
def load_data(self, loader = 'ctr'):
exec('self.load_data_{}()'.format(loader))
def load_data_ctr(self):
#8 columns in total
#X, H, K, Y, I, eI, LB, dL
#for CTR data, X column is L column, Y column all 0
#for RAXR data, X column is energy column, Y column is L column
# self.data = data.DataList()
options = QFileDialog.Options()
options |= QFileDialog.DontUseNativeDialog
fileName, _ = QFileDialog.getOpenFileName(self,"QFileDialog.getOpenFileName()", "","csv Files (*.csv);;data Files (*.dat);txt Files (*.txt)", options=options)
if fileName:
with open(fileName,'r') as f:
data_loaded = np.loadtxt(f,comments = '#',delimiter=None)
data_loaded_pd = pd.DataFrame(data_loaded, columns = ['X','h','k','Y','I','eI','LB','dL'])
data_loaded_pd['h'] = data_loaded_pd['h'].apply(lambda x:int(np.round(x)))
data_loaded_pd['k'] = data_loaded_pd['k'].apply(lambda x:int(np.round(x)))
data_loaded_pd.sort_values(by = ['h','k'], inplace = True)
# print(data_loaded_pd)
hk_unique = list(set(zip(list(data_loaded_pd['h']), list(data_loaded_pd['k']))))
hk_unique.sort()
h_unique = [each[0] for each in hk_unique]
k_unique = [each[1] for each in hk_unique]
for i in range(len(h_unique)):
h_temp, k_temp = h_unique[i], k_unique[i]
name = 'Data-{}{}L'.format(h_temp, k_temp)
self.model.data.add_new(name = name)
self.model.data.items[-1].x = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['X'].to_numpy()
self.model.data.items[-1].y = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['I'].to_numpy()
self.model.data.items[-1].error = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['eI'].to_numpy()
self.model.data.items[-1].x_raw = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['X'].to_numpy()
self.model.data.items[-1].y_raw = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['I'].to_numpy()
self.model.data.items[-1].error_raw = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['eI'].to_numpy()
self.model.data.items[-1].set_extra_data(name = 'h', value = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['h'].to_numpy())
self.model.data.items[-1].set_extra_data(name = 'k', value = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['k'].to_numpy())
self.model.data.items[-1].set_extra_data(name = 'Y', value = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['Y'].to_numpy())
self.model.data.items[-1].set_extra_data(name = 'LB', value = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['LB'].to_numpy())
self.model.data.items[-1].set_extra_data(name = 'dL', value = data_loaded_pd[(data_loaded_pd['h']==h_temp) & (data_loaded_pd['k']==k_temp)]['dL'].to_numpy())
#now remove the empty datasets
empty_data_index = []
i=0
for each in self.model.data.items:
if len(each.x_raw) == 0:
empty_data_index.append(i)
i += 1
for i in range(len(empty_data_index)):
self.model.data.delete_item(empty_data_index[i])
for ii in range(len(empty_data_index)):
if empty_data_index[ii]>empty_data_index[i]:
empty_data_index[ii] = empty_data_index[ii]-1
else:
pass
#update script_module
#self.model.script_module.__dict__['data'] = self.data
#update the view
self.update_table_widget_data()
self.update_combo_box_dataset()
self.update_plot_data_view()
def update_table_widget_data(self):
self.tableWidget_data.clear()
self.tableWidget_data.setRowCount(len(self.model.data))
self.tableWidget_data.setColumnCount(4)
self.tableWidget_data.setHorizontalHeaderLabels(['DataID','Show','Use','Errors'])
# self.tableWidget_pars.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch)
for i in range(len(self.model.data)):
current_data = self.model.data[i]
name = current_data.name
for j in range(4):
if j == 0:
qtablewidget = QTableWidgetItem(name)
self.tableWidget_data.setItem(i,j,qtablewidget)
else:
check_box = QCheckBox()
#self.show_checkBox_list.append(check_box)
check_box.setChecked(True)
check_box.stateChanged.connect(self.update_plot_data_view)
self.tableWidget_data.setCellWidget(i,j,check_box)
def update_combo_box_dataset(self):
new_items = [each.name for each in self.model.data]
self.comboBox_dataset.clear()
self.comboBox_dataset.addItems(new_items)
def update_data_view(self):
dataset_name = self.comboBox_dataset.currentText()
dataset = None
for each in self.model.data:
if each.name == dataset_name:
dataset = each
break
else:
pass
column_labels_main = ['x','y','error']
extra_labels = ['h', 'k', 'dL', 'LB']
all_labels = ['x','y','error','h','k','dL','LB','mask']
self.tableWidget_data_view.setRowCount(len(dataset.x))
self.tableWidget_data_view.setColumnCount(len(all_labels))
self.tableWidget_data_view.setHorizontalHeaderLabels(all_labels)
for i in range(len(dataset.x)):
for j in range(len(all_labels)):
if all_labels[j] in column_labels_main:
# print(getattr(dataset,'x')[i])
qtablewidget = QTableWidgetItem(str(getattr(dataset,all_labels[j])[i]))
elif all_labels[j] in extra_labels:
qtablewidget = QTableWidgetItem(str(dataset.get_extra_data(all_labels[j])[i]))
else:
qtablewidget = QTableWidgetItem('True')
self.tableWidget_data_view.setItem(i,j,qtablewidget)
def init_structure_view(self):
xyz, e = self.model.script_module.sample.extract_exyz(1)
ar = self.widget_edp.renderers.append(AtomRenderer(self.widget_edp, xyz, e))
#self.widget_edp.show_structure(xyz)
def update_structure_view(self):
#xyz = self.model.script_module.sample.extract_xyz(1)
xyz, e = self.model.script_module.sample.extract_exyz(1)
self.widget_edp.renderers[0].update_positions(xyz)
#ar = self.widget_edp.renderers.append(AtomRenderer(self.widget_edp, water.r_array, water.type_array))
def start_timer_structure_view(self):
self.timer_update_structure.start(2000)
def stop_timer_structure_view(self):
self.timer_update_structure.stop()
def append_data(self):
pass
def delete_data(self):
pass
def save_data(self):
pass
def calculate(self):
pass
def change_plot_style(self):
pass
def load_script(self):
options = QFileDialog.Options()
options |= QFileDialog.DontUseNativeDialog
fileName, _ = QFileDialog.getOpenFileName(self,"QFileDialog.getOpenFileName()", "","script Files (*.py);;text Files (*.txt)", options=options)
if fileName:
with open(fileName,'r') as f:
self.plainTextEdit_script.setPlainText(f.read())
self.model.script = (self.plainTextEdit_script.toPlainText())
#self.compile_script()
def update_script_upon_load(self):
self.plainTextEdit_script.setPlainText(self.model.script)
def save_script(self):
pass
def update_par_upon_load(self):
vertical_labels = []
lines = self.model.parameters.data
how_many_pars = len(lines)
self.tableWidget_pars.clear()
self.tableWidget_pars.setRowCount(how_many_pars)
self.tableWidget_pars.setColumnCount(6)
self.tableWidget_pars.setHorizontalHeaderLabels(['Parameter','Value','Fit','Min','Max','Error'])
# self.tableWidget_pars.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch)
for i in range(len(lines)):
items = lines[i]
#items = line.rstrip().rsplit('\t')
j = 0
if items[0] == '':
#self.model.parameters.data.append([items[0],0,False,0, 0,'-'])
vertical_labels.append('')
j += 1
else:
#add items to parameter attr
#self.model.parameters.data.append([items[0],float(items[1]),items[2]=='True',float(items[3]), float(items[4]),items[5]])
#add items to table view
if len(vertical_labels)==0:
vertical_labels.append('1')
else:
if vertical_labels[-1] != '':
vertical_labels.append('{}'.format(int(vertical_labels[-1])+1))
else:
vertical_labels.append('{}'.format(int(vertical_labels[-2])+1))
for item in items:
if j == 2:
check_box = QCheckBox()
check_box.setChecked(item==True)
self.tableWidget_pars.setCellWidget(i,2,check_box)
else:
qtablewidget = QTableWidgetItem(str(item))
# qtablewidget.setTextAlignment(Qt.AlignCenter)
if j == 0:
qtablewidget.setFont(QFont('Times',10,QFont.Bold))
elif j == 1:
qtablewidget.setForeground(QBrush(QColor(255,0,255)))
self.tableWidget_pars.setItem(i,j,qtablewidget)
j += 1
self.tableWidget_pars.resizeColumnsToContents()
self.tableWidget_pars.resizeRowsToContents()
self.tableWidget_pars.setShowGrid(False)
self.tableWidget_pars.setVerticalHeaderLabels(vertical_labels)
def load_par(self):
options = QFileDialog.Options()
options |= QFileDialog.DontUseNativeDialog
fileName, _ = QFileDialog.getOpenFileName(self,"QFileDialog.getOpenFileName()", "","Table Files (*.tab);;text Files (*.txt)", options=options)
vertical_labels = []
if fileName:
with open(fileName,'r') as f:
lines = f.readlines()
# self.parameters.set_ascii_input(f)
lines = [each for each in lines if not each.startswith('#')]
how_many_pars = len(lines)
self.tableWidget_pars.setRowCount(how_many_pars)
self.tableWidget_pars.setColumnCount(6)
self.tableWidget_pars.setHorizontalHeaderLabels(['Parameter','Value','Fit','Min','Max','Error'])
# self.tableWidget_pars.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch)
for i in range(len(lines)):
line = lines[i]
items = line.rstrip().rsplit('\t')
j = 0
if items[0] == '':
self.model.parameters.data.append([items[0],0,False,0, 0,'-'])
vertical_labels.append('')
j += 1
else:
#add items to parameter attr
self.model.parameters.data.append([items[0],float(items[1]),items[2]=='True',float(items[3]), float(items[4]),items[5]])
#add items to table view
if len(vertical_labels)==0:
vertical_labels.append('1')
else:
if vertical_labels[-1] != '':
vertical_labels.append('{}'.format(int(vertical_labels[-1])+1))
else:
vertical_labels.append('{}'.format(int(vertical_labels[-2])+1))
for item in items:
if j == 2:
check_box = QCheckBox()
check_box.setChecked(item=='True')
self.tableWidget_pars.setCellWidget(i,2,check_box)
else:
qtablewidget = QTableWidgetItem(item)
# qtablewidget.setTextAlignment(Qt.AlignCenter)
if j == 0:
qtablewidget.setFont(QFont('Times',10,QFont.Bold))
elif j == 1:
qtablewidget.setForeground(QBrush(QColor(255,0,255)))
self.tableWidget_pars.setItem(i,j,qtablewidget)
j += 1
self.tableWidget_pars.resizeColumnsToContents()
self.tableWidget_pars.resizeRowsToContents()
self.tableWidget_pars.setShowGrid(False)
self.tableWidget_pars.setVerticalHeaderLabels(vertical_labels)
@QtCore.pyqtSlot(str,object)
def update_par_during_fit(self,string,model):
#labels = [data[0] for each in self.model.parameters.data]
for i in range(len(model.parameters.data)):
if model.parameters.data[i][0]!='':
# print(self.model.parameters.data[i][0])
#print(len(self.model.parameters.data))
# print(model.parameters.data[i][0])
item_temp = self.tableWidget_pars.item(i,1)
#print(type(item_temp))
item_temp.setText(str(model.parameters.data[i][1]))
self.tableWidget_pars.resizeColumnsToContents()
self.tableWidget_pars.resizeRowsToContents()
self.tableWidget_pars.setShowGrid(False)
# self.update_structure_view()
def update_par_upon_change(self):
self.model.parameters.data = []
for each_row in range(self.tableWidget_pars.rowCount()):
if self.tableWidget_pars.item(each_row,0)==None:
items = ['',0,False,0,0,'-']
elif self.tableWidget_pars.item(each_row,0).text()=='':
items = ['',0,False,0,0,'-']
else:
# print(each_row,type(self.tableWidget_pars.item(each_row,0)))
items = [self.tableWidget_pars.item(each_row,0).text()] + [float(self.tableWidget_pars.item(each_row,i).text()) for i in [1,3,4]] + [self.tableWidget_pars.item(each_row,5).text()]
items.insert(2, self.tableWidget_pars.cellWidget(each_row,2).isChecked())
self.model.parameters.data.append(items)
@QtCore.pyqtSlot(str,object)
def update_status(self,string,model):
self.statusbar.clearMessage()
self.statusbar.showMessage(string)
self.label_2.setText('FOM {}:{}'.format(self.model.fom_func.__name__,self.run_fit.solver.optimizer.best_fom))
def save_par(self):
pass
if __name__ == "__main__":
QApplication.setStyle("windows")
app = QApplication(sys.argv)
myWin = MyMainWindow()
myWin.setWindowIcon(QtGui.QIcon('dafy.PNG'))
hightlight = syntax_pars.PythonHighlighter(myWin.plainTextEdit_script.document())
myWin.plainTextEdit_script.show()
myWin.plainTextEdit_script.setPlainText(myWin.plainTextEdit_script.toPlainText())
myWin.show()
sys.exit(app.exec_())
| [
"cqiu@alaska.edu"
] | cqiu@alaska.edu |
9dec5036ce036c0f4f8ece900dbe9b0c7f4709df | 0e8be64523c1fca6594118f61e8f9ef6d26ccb00 | /2017/0924_boys_names.py | a51ffd5cca53989959741ae2ed9071c5999921a7 | [
"CC0-1.0"
] | permissive | boisvert42/npr-puzzle-python | 782184540d8e9f62a9f8ab8a39ef51088839c1e8 | 5e3761e7cf3d9f6a05c9f32d6a26444375e73aff | refs/heads/master | 2020-05-22T01:14:51.732606 | 2019-12-29T17:32:40 | 2019-12-29T17:32:40 | 56,475,366 | 1 | 2 | null | null | null | null | UTF-8 | Python | false | false | 721 | py | #!/usr/bin/env python
"""
NPR 2017-09-24
http://www.npr.org/2017/09/24/553147004/sunday-puzzle-what-s-in-a-name
Think of a familiar 6-letter boy's name starting with a vowel.
Change the first letter to a consonant to get another familiar boy's name.
Then change the first letter to another consonant to get another familiar boy's name.
What names are these?
"""
from nltk.corpus import names
from collections import defaultdict
#%%
name_dict = defaultdict(set)
for name in names.words('male.txt'):
name_dict[name[1:]].add(name[0])
#%%
for k,v in name_dict.iteritems():
if len(v) >= 3 and len(k) == 5 and set('AEIOU').intersection(v):
for letter in v:
print letter+k,
print | [
"boisvert42@gmail.com"
] | boisvert42@gmail.com |
35de3051422fc58eeb8bcd5a8d99d3ce2a367973 | 107e7aca94b888b36f67c24a740c4aae7f827341 | /tacker/tests/unit/vm/test_plugin.py | 2ba2c91015a21cd35c15845285bbd463d7abc9a1 | [
"Apache-2.0"
] | permissive | vmehmeri/tacker | b397292511242bfe50540d10c5e116c23f312749 | d05cbaa26361090f36275a942ad87f6ecd3dd480 | refs/heads/master | 2020-05-29T10:23:21.900489 | 2015-10-08T15:46:15 | 2015-10-08T15:46:15 | 57,904,655 | 0 | 0 | null | 2016-05-02T16:41:15 | 2016-05-02T16:41:15 | null | UTF-8 | Python | false | false | 6,776 | py | # Copyright 2015 Brocade Communications System, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
import uuid
from tacker import context
from tacker.db.vm import vm_db
from tacker.tests.unit.db import base as db_base
from tacker.tests.unit.db import utils
from tacker.vm import plugin
class FakeDriverManager(mock.Mock):
def invoke(self, *args, **kwargs):
if 'create' in args:
return str(uuid.uuid4())
class FakeDeviceStatus(mock.Mock):
pass
class FakeGreenPool(mock.Mock):
pass
class TestVNFMPlugin(db_base.SqlTestCase):
def setUp(self):
super(TestVNFMPlugin, self).setUp()
self.addCleanup(mock.patch.stopall)
self.context = context.get_admin_context()
self._mock_device_manager()
self._mock_device_status()
self._mock_green_pool()
self.vnfm_plugin = plugin.VNFMPlugin()
def _mock_device_manager(self):
self._device_manager = mock.Mock(wraps=FakeDriverManager())
self._device_manager.__contains__ = mock.Mock(
return_value=True)
fake_device_manager = mock.Mock()
fake_device_manager.return_value = self._device_manager
self._mock(
'tacker.common.driver_manager.DriverManager', fake_device_manager)
def _mock_device_status(self):
self._device_status = mock.Mock(wraps=FakeDeviceStatus())
fake_device_status = mock.Mock()
fake_device_status.return_value = self._device_status
self._mock(
'tacker.vm.monitor.DeviceStatus', fake_device_status)
def _mock_green_pool(self):
self._pool = mock.Mock(wraps=FakeGreenPool())
fake_green_pool = mock.Mock()
fake_green_pool.return_value = self._pool
self._mock(
'eventlet.GreenPool', fake_green_pool)
def _mock(self, target, new=mock.DEFAULT):
patcher = mock.patch(target, new)
return patcher.start()
def _insert_dummy_device_template(self):
session = self.context.session
device_template = vm_db.DeviceTemplate(
id='eb094833-995e-49f0-a047-dfb56aaf7c4e',
tenant_id='ad7ebc56538745a08ef7c5e97f8bd437',
name='fake_template',
description='fake_template_description',
infra_driver='fake_driver',
mgmt_driver='fake_mgmt_driver')
session.add(device_template)
session.flush()
return device_template
def _insert_dummy_device(self):
session = self.context.session
device_db = vm_db.Device(id='6261579e-d6f3-49ad-8bc3-a9cb974778ff',
tenant_id='ad7ebc56538745a08ef7c5e97f8bd437',
name='fake_device',
description='fake_device_description',
instance_id=
'da85ea1a-4ec4-4201-bbb2-8d9249eca7ec',
template_id=
'eb094833-995e-49f0-a047-dfb56aaf7c4e',
status='ACTIVE')
session.add(device_db)
session.flush()
return device_db
def test_create_vnfd(self):
vnfd_obj = utils.get_dummy_vnfd_obj()
result = self.vnfm_plugin.create_vnfd(self.context, vnfd_obj)
self.assertIsNotNone(result)
self.assertIn('id', result)
self.assertIn('service_types', result)
self.assertIn('attributes', result)
self._device_manager.invoke.assert_called_once_with(mock.ANY,
mock.ANY,
plugin=mock.ANY,
context=mock.ANY,
device_template=
mock.ANY)
def test_create_vnf(self):
device_template_obj = self._insert_dummy_device_template()
vnf_obj = utils.get_dummy_vnf_obj()
vnf_obj['vnf']['vnfd_id'] = device_template_obj['id']
result = self.vnfm_plugin.create_vnf(self.context, vnf_obj)
self.assertIsNotNone(result)
self.assertIn('id', result)
self.assertIn('instance_id', result)
self.assertIn('status', result)
self.assertIn('attributes', result)
self.assertIn('mgmt_url', result)
self._device_manager.invoke.assert_called_with(mock.ANY, mock.ANY,
plugin=mock.ANY,
context=mock.ANY,
device=mock.ANY)
self._pool.spawn_n.assert_called_once_with(mock.ANY)
def test_delete_vnf(self):
self._insert_dummy_device_template()
dummy_device_obj = self._insert_dummy_device()
self.vnfm_plugin.delete_vnf(self.context, dummy_device_obj[
'id'])
self._device_manager.invoke.assert_called_with(mock.ANY, mock.ANY,
plugin=mock.ANY,
context=mock.ANY,
device_id=mock.ANY)
self._device_status.delete_hosting_device.assert_called_with(mock.ANY)
self._pool.spawn_n.assert_called_once_with(mock.ANY, mock.ANY,
mock.ANY)
def test_update_vnf(self):
self._insert_dummy_device_template()
dummy_device_obj = self._insert_dummy_device()
vnf_config_obj = utils.get_dummy_vnf_config_obj()
result = self.vnfm_plugin.update_vnf(self.context, dummy_device_obj[
'id'], vnf_config_obj)
self.assertIsNotNone(result)
self.assertEqual(dummy_device_obj['id'], result['id'])
self.assertIn('instance_id', result)
self.assertIn('status', result)
self.assertIn('attributes', result)
self.assertIn('mgmt_url', result)
self._pool.spawn_n.assert_called_once_with(mock.ANY, mock.ANY,
mock.ANY) | [
"sseetha@brocade.com"
] | sseetha@brocade.com |
76d04754ce5210635d52dd9cebffe88b8f70157b | 14a4864b10c64ed0f2d9662dae0dfe7bfb9b367e | /blog/urls.py | 268192c06ad190b9c2de92b9b58b875a26eef86f | [] | no_license | PungentLemon/My-portfolio | 0e54126992474e74c31d3d02aa8653f45350d99d | f7603cbc68f7a1db9ef57c78e2a7739e5e7637be | refs/heads/master | 2020-05-22T23:19:58.083478 | 2019-05-14T06:09:47 | 2019-05-14T06:09:47 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 176 | py |
from django.urls import path
from . import views
urlpatterns = [
path('',views.allblogs, name='allblogs'),
path('<int:blog_id>/',views.detail,name='detail'),
]
| [
"kukurureloaded@gmail.com"
] | kukurureloaded@gmail.com |
0ec26afbb8f3455b69bdcaa43e81f526289e0d08 | 97c418826161258cd6eb6c636689a6e649104cb4 | /Codigo/Aula-10-Processos/threads_e_sockets/app/aula8_fixed_client.py | fcb9abb07914e59c0f6f0e120d3faec734ef14e6 | [] | no_license | mairags/cet-100 | b53347799df3c5aee4010c53adf9b2eda6492274 | 21567e3f0623c34995c8dc2252c9c77f4053a9b6 | refs/heads/master | 2023-04-17T20:52:50.293209 | 2021-05-04T14:37:31 | 2021-05-04T14:37:31 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 660 | py | from socket import socket
from threading import Thread, Lock
TAM_BUFFER = 1000
NUM_REQ = 10
LOCK = Lock()
def enviar():
sock = socket()
server_info = ('127.0.0.1', 5000)
sock.connect(server_info)
dados_recebidos = sock.recv(TAM_BUFFER)
if LOCK.acquire():
print(dados_recebidos, flush=True)
LOCK.release()
sock.close()
def requisicoes():
try:
count = 0
while count < NUM_REQ:
count += 1
th = Thread(target=enviar)
th.start()
except KeyboardInterrupt:
print("Interrompido!")
def main():
requisicoes()
if __name__ == '__main__':
main()
| [
"mathias.brito@me.com"
] | mathias.brito@me.com |
c718408ccc29e4bca88b5deef7e84bb586acddfc | ea0c0b8d67a42086f840149b3dbe1c0e4f58e56f | /members_area/forms.py | 06d19b868f16f535ae4172f3cc5f191a2c75b8b0 | [
"MIT"
] | permissive | AzeezBello/raodoh | 78b27e0886f8882144a4def160d9c3f53bcc6af9 | 296bd44069bd750557bf49995374601f5052d695 | refs/heads/master | 2022-05-03T05:07:21.632642 | 2020-02-26T10:16:08 | 2020-02-26T10:16:08 | 235,878,080 | 0 | 0 | MIT | 2022-04-22T23:01:27 | 2020-01-23T20:15:39 | JavaScript | UTF-8 | Python | false | false | 194 | py | from django.forms import ModelForm
from .models import Lesson
class LessonForm(ModelForm):
class Meta:
model = Lesson
fields = ('title', 'course', 'body', 'url', 'video')
| [
"azeez@scholarx.co"
] | azeez@scholarx.co |
2a0e0bf2a10cc41a975d515cc6c614e82a56b20b | 2655a633b6c5f89400901179a7c19d76a7edc127 | /smili2/uvdata/uvdata/antable.py | 252d488532abdda27b395382fdea81720c9e9832 | [] | no_license | astrosmili/smili2_dev | c069d943405bd78ac95addc4d36ee0c9eb419cf6 | 06a389ebe054d9b285adf820e570484f56489fdb | refs/heads/master | 2021-09-17T06:13:40.090548 | 2021-03-18T06:18:17 | 2021-03-18T06:18:17 | 238,537,673 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,903 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
'''
'''
__author__ = "Smili Developer Team"
# ------------------------------------------------------------------------------
# Modules
# ------------------------------------------------------------------------------
# internal
from ...util.table import DataTable, DataSeries, TableHeader
# ------------------------------------------------------------------------------
# functions
# ------------------------------------------------------------------------------
class ANTable(DataTable):
header = TableHeader([
dict(name="antname", dtype="U32", unit="", comment="Antenna Name"),
dict(name="antid", dtype="int32", unit="", comment="Antenna ID"),
dict(name="mjd", dtype="float64", unit="day",
comment="Modified Jurian Day"),
dict(name="gst", dtype="float64", unit="hourangle",
comment="Greenwich Sidereal Time"),
dict(name="ra", dtype="float64", unit="deg",
comment="GCRS Right Ascention"),
dict(name="dec", dtype="float64", unit="deg", comment="GCRS Declination"),
dict(name="x", dtype="float64", unit="m",
comment="Geocenric Coordinate x"),
dict(name="y", dtype="float64", unit="m",
comment="Geocenric Coordinate y"),
dict(name="z", dtype="float64", unit="m",
comment="Geocenric Coordinate z"),
dict(name="az", dtype="float64", unit="deg", comment="Azimuthal Angle"),
dict(name="el", dtype="float64", unit="deg", comment="Elevation Angle"),
dict(name="par", dtype="float64", unit="deg",
comment="Parallactic Angle"),
dict(name="fra", dtype="float64", unit="deg",
comment="Field Roation Angle"),
dict(name="sefd1", dtype="float64", unit="Jy", comment="SEFD at Pol 1"),
dict(name="sefd2", dtype="float64", unit="Jy", comment="SEFD at Pol 2"),
dict(name="d1", dtype="float128", unit="", comment="D-term at Pol 1"),
dict(name="d2", dtype="float128", unit="", comment="D-term at Pol 2")
])
@property
def _constructor(self):
return ANTable
@property
def _constructor_sliced(self):
return ANSeries
@classmethod
def make(cls, utc, array, source):
from pandas import concat
from astropy.coordinates import GCRS
# number of time and utc bins
Nant = len(array.table)
# compute apparent source coordinates and GST
skycoord = source.skycoord.transform_to(GCRS(obstime=utc))
gst = utc.sidereal_time(
kind="apparent", longitude="greenwich", model="IAU2006A")
# run loop
def map_func(iant): return _antable_make_iant(
iant=iant, utc=utc, gst=gst, array=array, skycoord=skycoord)
antab = concat([map_func(iant) for iant in range(Nant)])
return antab
class ANSeries(DataSeries):
@property
def _constructor(self):
return ANSeries
@property
def _constructor_expanddim(self):
return ANTable
# define internal function to compute antenna based tables
def _antable_make_iant(iant, utc, gst, array, skycoord):
from numpy import exp, cos, sin, tan, arctan2
from astropy.coordinates import AltAz, EarthLocation
from ...util.units import DEG, RAD, M, DIMLESS
# get values
ant = array.table.loc[iant, :]
location = EarthLocation(x=ant.x, y=ant.y, z=ant.z, unit=M)
lon = location.lon
lat = location.lat
ra = skycoord.ra
dec = skycoord.dec
# compute LST
lst = utc.sidereal_time(kind="apparent", longitude=lon, model="IAU2006A")
# compute AZ / alt
site = AltAz(location=location, obstime=utc)
altaz = skycoord.transform_to(site)
el = altaz.alt
az = altaz.az
secz = altaz.secz
# compute sefd
sefd1 = ant.sefd1 * exp(-ant.tau1*secz)
sefd2 = ant.sefd2 * exp(-ant.tau2*secz)
# compute pallactic angle
H = lst.radian - ra.radian
cosH = cos(H)
sinH = sin(H)
tanlat = tan(lat.radian)
cosdec = cos(dec.radian)
sindec = sin(dec.radian)
par = arctan2(sinH, cosdec*tanlat - sindec*cosH)*RAD
# compute field rotation angle
fra = (ant.fr_pa_coeff * DIMLESS) * par
fra += (ant.fr_el_coeff * DIMLESS) * el
fra += ant.fr_offset * DEG
# antab
antab = ANTable(
dict(
antid=iant,
antname=ant.antname,
mjd=utc.mjd,
gst=gst.hour,
ra=ra.deg,
dec=dec.deg,
x=ant.x,
y=ant.y,
z=ant.z,
az=az.deg,
el=el.deg,
sefd1=sefd1,
sefd2=sefd2,
par=par.to_value(DEG),
fra=fra.to_value(DEG),
d1=ant.d1,
d2=ant.d2
),
columns=ANTable.header.name.to_list()
)
antab.convert_format()
return antab
| [
"kazunori.akiyama.kazu@gmail.com"
] | kazunori.akiyama.kazu@gmail.com |
415003695854767481118785de3b9da037c212d6 | 8db966ce80c6f5ae18243e42f8912bbc9e2bafa3 | /BPNN/BPNN_Regression/polution2/no2.py | 75e8cd74f81027481c28777156ba0022fe5872d3 | [] | no_license | soar200/PythonWorkSpace | 720d2066355089873651e926804de410157b7c93 | 95c90253981a6ea650f6191d1036e11eb2c936a6 | refs/heads/master | 2021-01-30T13:13:24.570947 | 2018-11-26T08:28:04 | 2018-11-26T08:28:04 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 9,043 | py | #-*- coding:utf-8 -*-
# &Author AnFany
# 适用于多维输出
from HB_Data_Reg import model_data as R_data
from HB_Data_Reg import L as LL
from HB_Data_Reg import n as N
import numpy as np
import tensorflow as tf
import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'
import matplotlib.pyplot as plt
'''第一部分:数据准备'''
'''第二部分: 基于TensorFlow构建训练函数'''
# 创建激活函数
def activate(input_layer, weights, biases, actfunc):
layer = tf.add(tf.matmul(input_layer, weights), biases)
if actfunc == 'relu':
return tf.nn.relu(layer)
elif actfunc == 'tanh':
return tf.nn.tanh(layer)
elif actfunc == 'sigmoid':
return tf.nn.sigmoid(layer)
# 权重初始化的方式和利用激活函数的关系很大
# sigmoid: xavir tanh: xavir relu: he
# 构建训练函数
def Ten_train(xdata, ydata,prexdata,key,hiddenlayers=3, hiddennodes=5, \
learn_rate=0.0012, itertimes=10000, batch_size=3, activate_func='sigmoid', break_error=0.005):
# 开始搭建神经网络
Input_Dimen = len(xdata[0])
Unit_Layers = [Input_Dimen] + [hiddennodes] * hiddenlayers + [len(ydata[0])] # 输入的维数,隐层的神经数,输出的维数1
# 创建占位符
x_data = tf.placeholder(shape=[None, Input_Dimen], dtype=tf.float32)
y_target = tf.placeholder(shape=[None, len(ydata[0])], dtype=tf.float32)
# 实现动态命名变量
VAR_NAME = locals()
for jj in range(hiddenlayers + 1):
VAR_NAME['weight%s' % jj] = tf.Variable(np.random.rand(Unit_Layers[jj], Unit_Layers[jj + 1]), dtype=tf.float32,\
name='weight%s' % jj) / np.sqrt(Unit_Layers[jj]) # sigmoid tanh
# VAR_NAME['weight%s'%jj] = tf.Variable(np.random.rand(Unit_Layers[jj], Unit_Layers[jj + 1]), dtype=tf.float32,name='weight%s' % jj) \/ np.sqrt(Unit_Layers[jj] / 2) # relu
VAR_NAME['bias%s' % jj] = tf.Variable(tf.random_normal([Unit_Layers[jj + 1]], stddev=10, name='bias%s' % jj),
dtype=tf.float32)
if jj == 0:
VAR_NAME['ooutda%s' % jj] = activate(x_data, eval('weight%s' % jj), eval('bias%s' % jj), actfunc=activate_func)
else:
VAR_NAME['ooutda%s' % jj] = activate(eval('ooutda%s' % (jj - 1)), eval('weight%s' % jj), \
eval('bias%s' % jj), actfunc=activate_func)
# 均方误差
loss = tf.reduce_mean(tf.reduce_sum(tf.square(y_target - eval('ooutda%s' % (hiddenlayers))), reduction_indices=[1]))
# 优化的方法
my_opt = tf.train.AdamOptimizer(learn_rate)
train_step = my_opt.minimize(loss)
# 初始化
init = tf.global_variables_initializer()
loss_vec = [] # 训练误差
with tf.Session() as sess:
saver = tf.train.Saver()
sess.run(init)
for i in range(itertimes):
rand_index = np.random.choice(len(xdata), size=batch_size, replace=False)
rand_x = xdata[rand_index]
rand_y = ydata[rand_index]
sess.run(train_step, feed_dict={x_data: rand_x, y_target: rand_y})
temp_loss = sess.run(loss, feed_dict={x_data: xdata, y_target: ydata})
loss_vec.append(temp_loss)
# 根据输出的误差,判断训练的情况
if (i + 1) % 50 == 0:
print(key+' '+'Generation: ' + str(i + 1) + '. 训练误差:Loss = ' + str(temp_loss))
# 提前退出的判断
if temp_loss < break_error: # 根据经验获得此数值, 因为采用的是随机下降,因此误差在前期可能出现浮动
break
# 计算预测数据的输出
pre_in_data0 = np.array(prexdata, dtype=np.float32)
for ipre in range(hiddenlayers + 1):
VAR_NAME['pre_in_data%s' % (ipre + 1)] = activate(eval('pre_in_data%s' % ipre), eval('weight%s' % ipre).eval(),\
eval('bias%s' % ipre).eval(), actfunc=activate_func)
# 计算训练数据的输出
train_in_data0 = np.array(xdata, dtype=np.float32)
for ipre in range(hiddenlayers + 1):
VAR_NAME['train_in_data%s' % (ipre + 1)] = activate(eval('train_in_data%s' % ipre), eval('weight%s' % ipre).eval(),\
eval('bias%s' % ipre).eval(), actfunc=activate_func)
# path = 'model2/'+key+'checkpoint/model.ckpt'
# saver.save(sess,path)
return eval('train_in_data%s'%(hiddenlayers+1)).eval(), eval('pre_in_data%s'%(hiddenlayers+1)).eval(), loss_vec
'''第三部分: 结果展示函数'''
#import matplotlib.pyplot as plt
from pylab import mpl # 作图显示中文
mpl.rcParams['font.sans-serif'] = ['FangSong'] # 设置中文字体新宋体
# 绘制图像
def figure(real, net, le='训练', real_line='ko-', net_line='r.-', width=3):
length = len(real[0])
# 绘制每个维度的对比图
for iwe in range(length):
plt.subplot(length, 1, iwe+1)
plt.plot(list(range(len(real.T[iwe]))), real.T[iwe], real_line, linewidth=width)
plt.plot(list(range(len(net.T[iwe]))), net.T[iwe], net_line, linewidth=width - 1)
plt.legend(['%s真实值'%le, '网络输出值'])
if length == 1:
plt.title('%s结果对比'%le)
else:
if iwe == 0:
plt.title('%s结果: %s维度对比'%(le, iwe))
else:
plt.title('%s维度对比'%iwe)
plt.show()
# 绘制成本函数曲线图
def costfig(errlist, le='成本函数曲线图'):
plt.plot(list(range(len(errlist))), errlist, linewidth=3)
plt.title(le)
plt.xlabel('迭代次数')
plt.ylabel('成本函数值')
plt.show()
def batchSizefigure(batch,error,le='batchSize分析图'):
plt.plot(batch,error,linewidth=3)
plt.title(le)
plt.xlabel('batchSize')
plt.ylabel('error')
plt.show()
# 因为训练数据较多,为了不影响展示效果,按序随机选取一定数量的数据,便于展示
def select(datax, datay, count=200):
sign = list(range(len(datax)))
selectr_sign = np.random.choice(sign, count, replace=False)
dx=datax[selectr_sign]
dy=datay[selectr_sign]
return dx,dy
# 将输出的数据转换尺寸,变为原始数据的尺寸
# def trans(ydata, minumber=R_data[4][0], maxumber=R_data[4][1]):
# return ydata * (maxumber - minumber) + minumber
#处理训练和测试的输入数据
def InputHandler(dat,LL):
rows = dat.shape[0]
cols = dat.shape[1]
data = []
for index in range(rows-LL+1):
dd = []
for k in range(index,index+LL):
dd.append(dat[k])
dt = [0]*36
for p in range(len(dd)):
nd = dd[p]
for q in range(cols):
dt[q+p*cols] = nd[q]
data.append(dt)
return data
if __name__ == '__main__':
errordic = {}
data = R_data.get('NO2')
train_x_data = data[0] # 训练输入
train_y_data = data[1] # 训练输出
predict_x_data = data[2] # 测试输入
predict_y_data = data[3] # 测试输出
for nodeNum in range(5,20):
er = []
batchSizelist = []
for batchSize in range(1,100):
# 训练
tfrelu = Ten_train(train_x_data, train_y_data, predict_x_data,'NO2', hiddennodes=nodeNum,batch_size=batchSize)
minumber = data[4][1]
maxumber = data[4][0]
#训练和测试的真实输出
train_y_data_tran = train_y_data*(maxumber - minumber) + minumber
predict_y_data_tran = predict_y_data*(maxumber - minumber) + minumber
#训练和测试的预测输出
train_output = tfrelu[0] * (maxumber - minumber) + minumber
predict_output = tfrelu[1] * (maxumber - minumber) + minumber
error = [0] * len(predict_output)
for index in range(len(predict_output)):
error[index] = abs(predict_y_data_tran[index] - predict_output[index]) / predict_y_data_tran[index]
count = 0
for k in range(len(error)):
if error[k] <= 0.4:
count = count + 1
auc = count / len(predict_output)
key = str(nodeNum)+'and'+str(batchSize)
errordic.setdefault(key, auc)
print(key+","+str(auc))
er.append(auc)
batchSizelist.append(batchSize)
if auc >= 0.75:
break
batchSizefigure(batchSizelist,er)
print(errordic)
# 数据多影响展示,随机挑选100条数据
# random_train_x_data = select(train_output, train_y_data_tran, 200)
# random_predict_x_data = select(predict_output, predict_y_data_tran, 100)
# figure(random_train_x_data[1], random_train_x_data[0], le='训练')
# figure(random_predict_x_data[1], random_predict_x_data[0], le='预测')
# costfig(tfrelu[2])
| [
"595001741@qq.com"
] | 595001741@qq.com |
c04f34aa199ad324a1340a32baf2c128a229bff8 | 1b0b44c8cb454d0241a732766f381ac908a3db96 | /merchant/backend/src/currency/__init__.py | f474fb25e8ed449f118c1ee2a0623071fd559e61 | [] | no_license | zeeis/reference-merchant | 35515ffba1b26fed29789cd26d26eaf265fee30c | f62c4185a0b53cf7512862a50c5a1bb74a86b377 | refs/heads/master | 2023-08-21T22:10:39.781860 | 2021-07-18T12:51:45 | 2021-07-18T12:51:45 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 87 | py | from .amount import Amount
from .currency import FiatCurrency
from .price import Price
| [
"ericnakagawa@gmail.com"
] | ericnakagawa@gmail.com |
f4dc6c951770b51ced2bc9a8a162b854c7b97c23 | 8ddd9445d773e8f10c0726a7503968557413d363 | /veri.py | 0fc85fe190025887014dad1816edcf31bbba9b56 | [] | no_license | Tuuf/telegrambot | 50b928794213b185dc2c51d9b21e0fc5dc801a94 | 241665629bc4ef4f2c2b60ba6e30f2d04373070c | refs/heads/master | 2022-12-16T01:46:04.775399 | 2020-09-02T14:45:18 | 2020-09-02T14:45:18 | 288,405,358 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 859 | py | import httpx
url="https://finans.truncgil.com/today.json"
response=httpx.get(url)
veri=response.json()
x = veri['ABD DOLARI']
amerika = x['Alış']
x=veri['İNGİLİZ STERLİNİ']
sterlin = x['Alış']
x=veri['İSVİÇRE FRANGI']
frank = x['Alış']
#
x=veri['KANADA DOLARI']
kanada = x['Alış']
#
x=veri['KUVEYT DİNARI']
kuveytdinar = x['Alış']
#
x=veri['NORVEÇ KRONU']
norveckron = x['Alış']
x=veri['SUUDİ ARABİSTAN RİYALİ']
sudiriyal = x['Alış']
#
x=veri['JAPON YENİ']
japyen= x['Alış']
#
x=veri['BULGAR LEVASI']
bulgarleva = x['Alış']
#
x=veri['RUMEN LEYİ']
romenleyi= x['Alış']
x=veri['RUS RUBLESİ']
ruble= x['Alış']
x=veri['İRAN RİYALİ']
riyaliran= x['Alış']
x=veri['ÇİN YUANI']
yuan= x['Alış']
x = veri['PAKİSTAN RUPİSİ']
pakistanrubi = x['Alış']
x = veri['KATAR RİYALİ']
katarriyali = x['Alış']
| [
"noreply@github.com"
] | noreply@github.com |
dd06fdb0ad1fa5a35fe8493e6cb4d7907be97e88 | 493915d74f4ad666a3673b8cf55d3521c8e10fad | /GMail_Alerts.py | e3e53abe4500a17fb21380f27002396d514cc7d9 | [] | no_license | tifabi/allTheGiNeed | e106e92cb91d8f5cfca27d4de09dfc759f7cc311 | 8a5c8f92113938c7b84e4a9c3f70b770078293c6 | refs/heads/master | 2020-03-22T06:36:53.772508 | 2018-07-03T23:39:36 | 2018-07-03T23:39:36 | 139,646,250 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 44,149 | py | '''
Tiffany Fabianac Modified code from:
Reading GMAIL using Python
- https://github.com/abhishekchhibber/Gmail-Api-through-Python
- Abhishek Chhibber
'''
'''
This script does the following:
- Go to Gmal inbox
- Find and read all the Google Alert messages
- Extract details (Date, Snippet,URL) and export them to a .csv file / DB
'''
'''
Before running this script, the user should get the authentication by following
the link: https://developers.google.com/gmail/api/quickstart/python
Also, client_secret.json should be saved in the same directory as this file
'''
# Importing required libraries
import base64
import email
from apiclient import discovery
from httplib2 import Http
from oauth2client import file, client, tools
import urllib
import re
import dateutil.parser as parser
import json
# Creating a storage.JSON file with authentication details
SCOPES = 'https://www.googleapis.com/auth/gmail.modify' # we are using modify and not readonly, as we will be marking the messages Read
store = file.Storage('storage.json')
creds = store.get()
if not creds or creds.invalid:
flow = client.flow_from_clientsecrets('client_secret.json', SCOPES)
creds = tools.run_flow(flow, store)
GMAIL = discovery.build('gmail', 'v1', http=creds.authorize(Http()))
user_id = 'me'
label_id_one = 'INBOX'
# Getting Google Alert messages from Inbox
#maxResults=1, q='from:googlealerts-noreply@google.com is:unread'
alert_msgs = GMAIL.users().messages().list(userId='me', labelIds=[label_id_one], maxResults=1,
q='from:googlealerts-noreply@google.com').execute()
##RETURNS: {'messages': [{'id': '1645db225f22f01b', 'threadId': '1644e1a892b315c7'}], 'nextPageToken': '16344340453457345441', 'resultSizeEstimate': 4}
# Read values for the key 'messages'
mssg_list = alert_msgs['messages']
##RETURNS: [{'id': '1645db225f22f01b', 'threadId': '1644e1a892b315c7'}]
final_list = []
for mssg in mssg_list:
temp_dict = {}
# get id of individual message
m_id = mssg['id']
# fetch the message using API
# format='raw'
message = GMAIL.users().messages().get(userId=user_id, id=m_id).execute()
'''
RETURNS: {
'id': '1645db225f22f01b',
'threadId': '1644e1a892b315c7',
'labelIds': ['IMPORTANT', 'CATEGORY_UPDATES', 'INBOX'],
'snippet': 'Google Phase III trial As-it-happens update ⋅ July 3, 2018 NEWS Dr. Yardley on the Role of Biosimilars in Breast Cancer OncLive The phase III trial showed an equivalent pathologic complete response',
'historyId': '3276761',
'internalDate': '1530580312000',
'payload': {
'partId': '',
'mimeType': 'multipart/alternative',
'filename': '',
'headers': [
{'name': 'Delivered-To',
'value': 'tiffanyfabianac@gmail.com'
},
{'name': 'Received', 'value': 'by 2002:a67:7cc7:0:0:0:0:0 with SMTP id x190-v6csp496217vsc; Mon, 2 Jul 2018 18:11:53 -0700 (PDT)'},
{'name': 'X-Received', 'value': 'by 2002:a25:e7c8:: with SMTP id e191-v6mr14135883ybh.358.1530580313504; Mon, 02 Jul 2018 18:11:53 -0700 (PDT)'},
{'name': 'ARC-Seal', 'value': 'i=1; a=rsa-sha256; t=1530580313; cv=none; d=google.com; s=arc-20160816; b=mU8UUtj7ocw8+KCZLIUQeoVIqGlkGEjdt5y/rHB9zIu9WivLp4nhm/EQDwhSEiKWQz fu/ooKubGTQnsa80kfxPkkGQI5n6KCgyiZ3lrCbNO6LLq0vvmp4C/5IR+pMlT8Eim+5t h6gEc8ssn7pPg+r0JrwmD3A5TA8G7XhP9Iz910icHKiPfyg3yhXXYkGXrpCUqyuDXTnD SpXGa5OWwIAF0hh/s2hZPwewvQr+UtjZvjHG8Q5bzCBQsnVDhxDNbvRTK7zMwyG41OM9 l2bV+wJbxmqeVKolU/Qrm5t/B+4pH2M6JF0RfP5AVX6+ZFr/LV4PJCdArmdTJn2ZgtE7 a07A=='},
{'name': 'ARC-Message-Signature', 'value': 'i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=to:from:subject:message-id:list-unsubscribe:list-id:date :mime-version:dkim-signature:arc-authentication-results; bh=9r86Eg1CwIukfK0DSWCms++zhn5NAd7rfZJut81wR5k=; b=Ecn/KekxBObsgIke5n9Wale55ksCncSFWNlv/xpcek2o/kKm+pBSYC2HsXEOR9Uk9n dmaC0c9ueAUYCtL0nTypZ+gMgcka/vsiZeYoo+TS0cd08Z0CW2rRYVRWCPe/FfVUB8HP Oh7zTixDjiCBMvil7ktG/iSjVgxFkE40QyRmCB95yr5BMF/iQJ8iiLoVybqy+KmmRXf2 eC+e8ifmyv+T/dJ3/X+AdSzMnWBReAKJCGqfrLxKDFpQwdnMiPAtSnwYJzB8reEMgCuz hnPJ0QczzYp06/m8u8nPTbUW3OoQ9XxnpNLlvLv6HFuq9ku9X9RNRU8jx8ByeDVyRh1d Z6WQ=='},
{'name': 'ARC-Authentication-Results', 'value': 'i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=BqSripFi; spf=pass (google.com: domain of 3wm06wxqkal0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com designates 209.85.220.69 as permitted sender) smtp.mailfrom=3WM06WxQKAL0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com'},
{'name': 'Return-Path', 'value': '<3WM06WxQKAL0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com>'}, {'name': 'Received', 'value': 'from mail-sor-f69.google.com (mail-sor-f69.google.com. [209.85.220.69]) by mx.google.com with SMTPS id 82-v6sor4071997ybz.185.2018.07.02.18.11.53 for <tiffanyfabianac@gmail.com> (Google Transport Security); Mon, 02 Jul 2018 18:11:53 -0700 (PDT)'},
{'name': 'Received-SPF', 'value': 'pass (google.com: domain of 3wm06wxqkal0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com designates 209.85.220.69 as permitted sender) client-ip=209.85.220.69;'},
{'name': 'Authentication-Results', 'value': 'mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=BqSripFi; spf=pass (google.com: domain of 3wm06wxqkal0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com designates 209.85.220.69 as permitted sender) smtp.mailfrom=3WM06WxQKAL0jrrjohdohuwv-qruhso1jrrjoh.frp@alerts.bounces.google.com; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com'},
{'name': 'DKIM-Signature', 'value': 'v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:date:list-id:list-unsubscribe:message-id:subject:from :to; bh=9r86Eg1CwIukfK0DSWCms++zhn5NAd7rfZJut81wR5k=; b=BqSripFiLq2sqIN/PHts9wLXr9/hlqCJpYmx97azZzfGJrUCcJVHrFj+H+ft7a2F47 Piu+fkmprHFN1NanT0OuYWjTVmTKDvzwTSJ2gSfDR49jpYveXZSrxjoQXBesoXQUxh98 94FvRCzsw0Vem24g0lUnTvjmcATlse1J+OL+BCT/6uv1cpCWkBMWtoCNBp2bb8Todx/C CJNenfuBDjoFIoE9XcpmEuVmaNvk36m3rWdAr3XBIt0rVDMABMbQujWm+QimF/M040gG ao0ehCb+QKgI+So+vUwRK7aioCOGngsxAjCTVnFlGlhoTnAF9HieLcpTt4tIxdvgi6yu fR5Q=='},
{'name': 'X-Google-DKIM-Signature', 'value': 'v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:date:list-id:list-unsubscribe :message-id:subject:from:to; bh=9r86Eg1CwIukfK0DSWCms++zhn5NAd7rfZJut81wR5k=; b=GfY1Y5PqRYnF7NMu15paoNqquSXnDYOM/VsbWBm1x/i5auUs+FhBNOmrLZo/Kn5PaS 4aH+x6tEldP5V0OmS4lLL2NDZwGN5Z57R2e6b5GMfN60nrL2TLZih+D7mMpIcauq5eWb FGZJYivEM4z+COSpCSB/K3o4jCQKRSGs5o070HwhbNIB3eB2XHFfbyC57h+xlnfTssxD 9Si4I0naSPc4aFkD5HzXUZ8c80wdVpjZQSUy25K4W1WZQ5lTp7fLrXm5JX7YlCPV9CcE 4sccIXiUGKql021JhbiorMYf52yP8w2uZm5MeeucHiPeBOdbSrxqu4zRy0XLCkoJT5zT RfMQ=='}, {'name': 'X-Gm-Message-State', 'value': 'APt69E27szFihIx5AGfr09D5H3zL9qmvELdzWXRPVMCX8g4Ylmmrfujp hlPEW7ugivU='}, {'name': 'X-Google-Smtp-Source', 'value': 'ADUXVKJk9ZZV72gBSPr06HYC5DTltujUq4EoDXGBs0mvbz2wiFZMzKQnzlUC3eKAMN0/GfE20c4='},
{'name': 'MIME-Version', 'value': '1.0'},
{'name': 'X-Received', 'value': 'by 2002:a25:8706:: with SMTP id a6-v6mr8168201ybl.49.1530580312922; Mon, 02 Jul 2018 18:11:52 -0700 (PDT)'}, {'name': 'Date', 'value': 'Mon, 02 Jul 2018 18:11:52 -0700'}, {'name': 'List-Id', 'value': '<11057473802345533838.alerts.google.com>'}, {'name': 'List-Unsubscribe', 'value': '<mailto:ur@unsubscribe.alerts.google.com?subject=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA>'}, {'name': 'Message-ID', 'value': '<000000000000da1bf605700dffc3@google.com>'},
{'name': 'Subject', 'value': 'Google Alert - Phase III trial'},
{'name': 'From', 'value': 'Google Alerts <googlealerts-noreply@google.com>'},
{'name': 'To', 'value': 'tiffanyfabianac@gmail.com'},
{'name': 'Content-Type', 'value': 'multipart/alternative; boundary="000000000000da1bcf05700dffc0"'}],
'body': {'size': 0},
'parts': [{'partId': '0', 'mimeType': 'text/plain', 'filename': '',
'headers': [
{'name': 'Content-Type', 'value': 'text/plain; charset="UTF-8"; format=flowed; delsp=yes'},
{'name': 'Content-Transfer-Encoding', 'value': 'base64'}],
'body': {'size': 2131, 'data': 'PT09IE5ld3MgLSAzIG5ldyByZXN1bHRzIGZvciBbUGhhc2UgSUlJIHRyaWFsXSA9PT0NCg0KRHIuIFlhcmRsZXkgb24gdGhlIFJvbGUgb2YgQmlvc2ltaWxhcnMgaW4gQnJlYXN0IENhbmNlcg0KT25jTGl2ZQ0KVGhlIHBoYXNlIElJSSB0cmlhbCBzaG93ZWQgYW4gZXF1aXZhbGVudCBwYXRob2xvZ2ljIGNvbXBsZXRlIHJlc3BvbnNlIHJhdGUNCmJldHdlZW4gdHJhc3R1enVtYWIgKEhlcmNlcHRpbikgYW5kIHRoZSBiaW9zaW1pbGFyIEFCUCA5ODAuIEFsdGhvdWdoIC4uLg0KPGh0dHBzOi8vd3d3Lmdvb2dsZS5jb20vdXJsP3JjdD1qJnNhPXQmdXJsPWh0dHBzOi8vd3d3Lm9uY2xpdmUuY29tL29uY2xpdmUtdHYvZHIteWFyZGxleS1vbi10aGUtcm9sZS1vZi1iaW9zaW1pbGFycy1pbi1icmVhc3QtY2FuY2VyJmN0PWdhJmNkPUNBRVlBQ29UTWpRNE9Ea3pNVGd3TlRZM01UTTJPRFF5TVRJY04yUTRNR0V4T1dObE1UbGhNVGM0TmpwamIyMDZaVzQ2VlZNNlVnJnVzZz1BRlFqQ05Ha3U3R19TN3Z2UzlmMENkRVlBRmVhdmdxT3dBPg0KDQpOZXcgUmFkaW90aGVyYXB5IFByb3N0YXRlIENhbmNlciBUcmVhdG1lbnQsIFNwYWNlT0FSwq4gSHlkcm9nZWwsIE5vdw0KQXZhaWxhYmxlIGluIEphcGFuDQpDaXRpemVudHJpYnVuZQ0KQ29udGludWVkIEJlbmVmaXQgdG8gUmVjdGFsIFNlcGFyYXRpb24gZm9yIFByb3N0YXRlIFJUOiBGaW5hbCBSZXN1bHRzIG9mIGENClBoYXNlIElJSSBUcmlhbC4gSW50IEogUmFkaWF0IE9uY29sIEJpb2wgUGh5czsgMjAxNyBWb2x1bWUgOTcsIElzc3VlIDUsDQpQYWdlcyAuLi4NCjxodHRwczovL3d3dy5nb29nbGUuY29tL3VybD9yY3Q9aiZzYT10JnVybD1odHRwczovL3d3dy5jaXRpemVudHJpYnVuZS5jb20vbmV3cy9idXNpbmVzcy9uZXctcmFkaW90aGVyYXB5LXByb3N0YXRlLWNhbmNlci10cmVhdG1lbnQtc3BhY2VvYXItaHlkcm9nZWwtbm93LWF2YWlsYWJsZS1pbi9hcnRpY2xlXzcwYWEyOGMzLTA3N2QtNWJlMS1iMThmLTRjNDcyZjY3ZjhjNy5odG1sJmN0PWdhJmNkPUNBRVlBU29UTWpRNE9Ea3pNVGd3TlRZM01UTTJPRFF5TVRJY04yUTRNR0V4T1dObE1UbGhNVGM0TmpwamIyMDZaVzQ2VlZNNlVnJnVzZz1BRlFqQ05FcGFJV0dmWmhXMlcycnlhZS1JMHo3WjZSNU1RPg0KDQpEci4gSHVtcGhyZXkgb24gTW9nYW11bGl6dW1hYiBmb3IgQ3V0YW5lb3VzIFQtQ2VsbCBMeW1waG9tYQ0KT25jTGl2ZQ0KLi4uIHdpdGggY3V0YW5lb3VzIFQtY2VsbCBseW1waG9tYSAoQ1RDTCkgdGhhdCB3ZXJlIHJlcG9ydGVkIGluIHRoZSBwaGFzZQ0KSUlJIE1BVk9SSUMgdHJpYWwsIHdoaWNoIHdhcyBwcmVzZW50ZWQgYXQgdGhlIDIwMTggQVNDTyBBbm51YWwgTWVldGluZy4NCjxodHRwczovL3d3dy5nb29nbGUuY29tL3VybD9yY3Q9aiZzYT10JnVybD1odHRwczovL3d3dy5vbmNsaXZlLmNvbS9vbmNsaXZlLXR2L2RyLWh1bXBocmV5LW9uLW1vZ2FtdWxpenVtYWItZm9yLWN1dGFuZW91cy10Y2VsbC1seW1waG9tYSZjdD1nYSZjZD1DQUVZQWlvVE1qUTRPRGt6TVRnd05UWTNNVE0yT0RReU1USWNOMlE0TUdFeE9XTmxNVGxoTVRjNE5qcGpiMjA2Wlc0NlZWTTZVZyZ1c2c9QUZRakNORUlWVHVKZTkxaHhDLVMtX1doQkFMdXUyMHB6UT4NCg0KDQotIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtIC0gLSAtDQpVbnN1YnNjcmliZSBmcm9tIHRoaXMgR29vZ2xlIEFsZXJ0Og0KPGh0dHBzOi8vd3d3Lmdvb2dsZS5jb20vYWxlcnRzL3JlbW92ZT9zb3VyY2U9YWxlcnRzbWFpbCZobD1lbiZnbD1VUyZtc2dpZD1NalE0T0Rrek1UZ3dOVFkzTVRNMk9EUXlNUSZzPUFCMlhxNGh2akk1OHdyOVVLNHRVN3EwcXNQMXhmdGlVNmtPbDNTQT4NCg0KQ3JlYXRlIGFub3RoZXIgR29vZ2xlIEFsZXJ0Og0KPGh0dHBzOi8vd3d3Lmdvb2dsZS5jb20vYWxlcnRzP3NvdXJjZT1hbGVydHNtYWlsJmhsPWVuJmdsPVVTJm1zZ2lkPU1qUTRPRGt6TVRnd05UWTNNVE0yT0RReU1RPg0KDQpTaWduIGluIHRvIG1hbmFnZSB5b3VyIGFsZXJ0czoNCjxodHRwczovL3d3dy5nb29nbGUuY29tL2FsZXJ0cz9zb3VyY2U9YWxlcnRzbWFpbCZobD1lbiZnbD1VUyZtc2dpZD1NalE0T0Rrek1UZ3dOVFkzTVRNMk9EUXlNUT4NCg=='}},
{'partId': '1', 'mimeType': 'text/html', 'filename': '',
'headers': [{'name': 'Content-Type', 'value': 'text/html; charset="UTF-8"'},
{'name': 'Content-Transfer-Encoding', 'value': 'quoted-printable'}],
'body': {'size': 22634, 'data': '<html lang="en-US"> <head>  </head> <body> <div>  <script data-scope="inboxmarkup" type="application/json">{
  "api_version": "1.0",
  "publisher": {
    "api_key": "668269e72cfedea31b22524041ff21d9",
    "name": "Google Alerts"
  },
  "entity": {
    "external_key": "Google Alert - Phase III trial",
    "title": "Google Alert - Phase III trial",
    "subtitle": "Latest: July 3, 2018",
    "avatar_image_url": "https://www.gstatic.com/images/branding/product/1x/gsa_512dp.png",
    "main_image_url": "https://www.gstatic.com/bt/C3341AA7A1A076756462EE2E5CD71C11/smartmail/mobile/il_newspaper_header_r1.png"
  },
  "updates": {
    "snippets": [ {
      "icon": "BOOKMARK",
      "message": "Dr. Yardley on the Role of Biosimilars in Breast Cancer"
    }, {
      "icon": "BOOKMARK",
      "message": "New Radiotherapy Prostate Cancer Treatment, SpaceOAR® Hydrogel, Now Available in Japan"
    }, {
      "icon": "BOOKMARK",
      "message": "Dr. Humphrey on Mogamulizumab for Cutaneous T-Cell Lymphoma"
    } ]
  },
  "cards": [ {
    "title": "Google Alert - Phase III trial",
    "subtitle": "Highlights from the latest email",
    "actions": [ {
      "name": "See more results",
      "url": "https://www.google.com/alerts?s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA\u0026start=1530565046\u0026end=1530580312\u0026source=alertsmail\u0026hl=en\u0026gl=US\u0026msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ#history"
    } ],
    "widgets": [ {
      "type": "LINK",
      "title": "Dr. Yardley on the Role of Biosimilars in Breast Cancer",
      "description": "The phase III trial showed an equivalent pathologic complete response rate between trastuzumab (Herceptin) and the biosimilar ABP 980. Although ...",
      "image_url": "http://img.youtube.com/vi/Nv6OaazbXC8/default.jpg",
      "url": "https://www.google.com/url?rct=j\u0026sa=t\u0026url=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer\u0026ct=ga\u0026cd=CAEYACoTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug\u0026usg=AFQjCNGku7G_S7vvS9f0CdEYAFeavgqOwA"
    }, {
      "type": "LINK",
      "title": "New Radiotherapy Prostate Cancer Treatment, SpaceOAR® Hydrogel, Now Available in Japan",
      "description": "Continued Benefit to Rectal Separation for Prostate RT: Final Results of a Phase III Trial. Int J Radiat Oncol Biol Phys; 2017 Volume 97, Issue 5, Pages ...",
      "url": "https://www.google.com/url?rct=j\u0026sa=t\u0026url=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html\u0026ct=ga\u0026cd=CAEYASoTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug\u0026usg=AFQjCNEpaIWGfZhW2W2ryae-I0z7Z6R5MQ"
    }, {
      "type": "LINK",
      "title": "Dr. Humphrey on Mogamulizumab for Cutaneous T-Cell Lymphoma",
      "description": "... with cutaneous T-cell lymphoma (CTCL) that were reported in the phase III MAVORIC trial, which was presented at the 2018 ASCO Annual Meeting.",
      "image_url": "http://img.youtube.com/vi/B8Pqnwuak0g/default.jpg",
      "url": "https://www.google.com/url?rct=j\u0026sa=t\u0026url=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma\u0026ct=ga\u0026cd=CAEYAioTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug\u0026usg=AFQjCNEIVTuJe91hxC-S-_WhBALuu20pzQ"
    } ]
  } ]
}
</script> <!--[if mso]>
 <table><tr><td width=650>
<![endif]-->
 <div style="width:100%;max-width:650px"> <div style="font-family:Arial"> <table style="border-collapse:collapse;border-left:1px solid #e4e4e4;border-right:1px solid #e4e4e4"> <tr> <td style="background-color:#f8f8f8;padding-left:18px;border-bottom:1px solid #e4e4e4;border-top:1px solid #e4e4e4"></td> <td valign="middle" style="padding:13px 10px 8px 0px;background-color:#f8f8f8;border-top:1px solid #e4e4e4;border-bottom:1px solid #e4e4e4"> <a href="https://www.google.com/alerts?source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ" style="text-decoration:none"> <img src="https://www.google.com/intl/en_us/alerts/logo.png?cd=KhMyNDg4OTMxODA1NjcxMzY4NDIx" alt="Google" border="0" height="25"> </a> </td> <td style="background-color:#f8f8f8;padding-left:18px;border-top:1px solid #e4e4e4;border-bottom:1px solid #e4e4e4"></td> </tr>  <tr>  <td style="padding-left:32px"></td> <td style="padding:18px 0px 0px 0px;vertical-align:middle;line-height:20px;font-family:Arial"> <span style="color:#262626;font-size:22px">Phase III trial</span> <div style="vertical-align:top;padding-top:6px;color:#aaa;font-size:12px;line-height:16px"> <span>As-it-happens update</span> <span style="padding:0px 4px 0px 4px">&sdot;</span> <a style="color:#aaa;text-decoration:none">July 3, 2018</a> </div> </td> <td style="padding-left:32px"></td>   </tr>  <tr> <td style="padding-left:18px"></td> <td style="padding:16px 0px 12px 0px;border-bottom:1px solid #e4e4e4"> <span style="font-size:12px;color:#737373"> NEWS </span> </td> <td style="padding-right:18px"></td> </tr>   <tr itemscope="" itemtype="http://schema.org/Article"> <td style="padding-left:18px"></td> <td style="padding:18px 0px 12px 0px;vertical-align:top;font-family:Arial"> <a href="https://www.google.com/url?rct=j&amp;sa=t&amp;url=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;ct=ga&amp;cd=CAEYACoTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug&amp;usg=AFQjCNGku7G_S7vvS9f0CdEYAFeavgqOwA" style="text-decoration:none"> <table align="right" style="display:inline;border-collapse:collapse"> <tr> <td style="padding-left:18px"></td> <td background="https://img.youtube.com/vi/Nv6OaazbXC8/default.jpg" height="100" width="100" valign="bottom" style="padding:0px 0px 0px 0px;background-repeat:no-repeat;text-align:center;border:1px solid #e4e4e4"> <link href="https://img.youtube.com/vi/Nv6OaazbXC8/default.jpg" itemprop="image"> <!--[if gte mso 9]>
       <v:image xmlns:v="urn:schemas-microsoft-com:vml" id="theImage"
                style='behavior: url(#default#VML);
                       display:inline-block; position:absolute;
                       height: 100px; width: 100px;
                       top:0; left:0; border:0; z-index:1;' src="https://img.youtube.com/vi/Nv6OaazbXC8/default.jpg"/>
       <v:rect xmlns:v="urn:schemas-microsoft-com:vml"
               style='behavior: url(#default#VML);
                      display:inline-block;
                      position: absolute; top: 82px; left: 0px;
                      width:100px; border:0; z-index:2;' strokecolor="none">
       <v:fill opacity="50%" color="#000000"/>
       <v:textbox style="mso-fit-shape-to-text:t;
                         mso-column-margin: 0pt;
                         letter-spacing: 0.8px;"
                         inset="0pt,0pt,0pt,0pt">
         <font size="-1" color="#ffffff"></font>
       </v:textbox>
       </v:rect>
       <div style="display: none">
       <![endif]--> <div style="color:#fff;font-size:9px;letter-spacing:0.8px"> <div style="padding:3px 0px 4px 4px;background:rgb(255,255,255);background-color:rgba(0,0,0,0.5);width:96px"></div> </div> <!--[if gte mso 9]></div><![endif]--> </td> </tr> </table> </a> <div>  <span style="padding:0px 6px 0px 0px"> <a href="https://www.google.com/url?rct=j&amp;sa=t&amp;url=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;ct=ga&amp;cd=CAEYACoTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug&amp;usg=AFQjCNGku7G_S7vvS9f0CdEYAFeavgqOwA" itemprop="url" style="color:#427fed;display:inline;text-decoration:none;font-size:16px;line-height:20px"> <span itemprop="name">Dr. Yardley on the Role of Biosimilars in Breast Cancer</span> </a> </span>  <div> <div style="padding:2px 0px 8px 0px"> <div itemprop="publisher" itemscope="" itemtype="http://schema.org/Organization" style="color:#737373;font-size:12px"> <a style="text-decoration:none;color:#737373"> <span itemprop="name">OncLive</span> </a> </div> <div itemprop="description" style="color:#252525;padding:2px 0px 0px 0px;font-size:12px;line-height:18px">The <b>phase III trial</b> showed an equivalent pathologic complete response rate between trastuzumab (Herceptin) and the biosimilar ABP 980. Although&nbsp;...</div> </div>   <table> <tr> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;ss=gp&amp;rt=Dr.+Yardley+on+the+Role+of+Biosimilars+in+Breast+Cancer&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmUyMRNWvwfisKgikDrtu5dLgTgxrw" style="text-decoration:none"> <img alt="Google Plus" src="https://www.gstatic.com/alerts/images/gp-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;ss=fb&amp;rt=Dr.+Yardley+on+the+Role+of+Biosimilars+in+Breast+Cancer&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmUyMRNWvwfisKgikDrtu5dLgTgxrw" style="text-decoration:none"> <img alt="Facebook" src="https://www.gstatic.com/alerts/images/fb-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;ss=tw&amp;rt=Dr.+Yardley+on+the+Role+of+Biosimilars+in+Breast+Cancer&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmUyMRNWvwfisKgikDrtu5dLgTgxrw" style="text-decoration:none"> <img alt="Twitter" src="https://www.gstatic.com/alerts/images/tw-24.png" border="0" height="16" width="16"></a> </td> <td style="padding:0px 0px 6px 15px;font-family:Arial"> <a href="https://www.google.com/alerts/feedback?ffu=https://www.onclive.com/onclive-tv/dr-yardley-on-the-role-of-biosimilars-in-breast-cancer&amp;source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA" style="text-decoration:none;vertical-align:middle;color:#aaa;font-size:10px"> Flag as irrelevant </a> </td> </tr> </table>  </div> </div> </td> <td style="padding-right:18px"></td> </tr>    <tr itemscope="" itemtype="http://schema.org/Article"> <td style="padding-left:18px"></td> <td style="padding:18px 0px 12px 0px;vertical-align:top;border-top:1px solid #e4e4e4;font-family:Arial"> <a></a> <div>  <span style="padding:0px 6px 0px 0px"> <a href="https://www.google.com/url?rct=j&amp;sa=t&amp;url=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html&amp;ct=ga&amp;cd=CAEYASoTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug&amp;usg=AFQjCNEpaIWGfZhW2W2ryae-I0z7Z6R5MQ" itemprop="url" style="color:#427fed;display:inline;text-decoration:none;font-size:16px;line-height:20px"> <span itemprop="name">New Radiotherapy Prostate Cancer Treatment, SpaceOAR® Hydrogel, Now Available in Japan</span> </a> </span>  <div> <div style="padding:2px 0px 8px 0px"> <div itemprop="publisher" itemscope="" itemtype="http://schema.org/Organization" style="color:#737373;font-size:12px"> <a style="text-decoration:none;color:#737373"> <span itemprop="name">Citizentribune</span> </a> </div> <div itemprop="description" style="color:#252525;padding:2px 0px 0px 0px;font-size:12px;line-height:18px">Continued Benefit to Rectal Separation for Prostate RT: Final Results of a <b>Phase</b> <b>III Trial</b>. Int J Radiat Oncol Biol Phys; 2017 Volume 97, Issue 5, Pages&nbsp;...</div> </div>   <table> <tr> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html&amp;ss=gp&amp;rt=New+Radiotherapy+Prostate+Cancer+Treatment,+SpaceOAR%C2%AE+Hydrogel,+Now+Available+in+Japan&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmVWKJ7n0WfkJ51hiflfA7cFF-jOMA" style="text-decoration:none"> <img alt="Google Plus" src="https://www.gstatic.com/alerts/images/gp-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html&amp;ss=fb&amp;rt=New+Radiotherapy+Prostate+Cancer+Treatment,+SpaceOAR%C2%AE+Hydrogel,+Now+Available+in+Japan&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmVWKJ7n0WfkJ51hiflfA7cFF-jOMA" style="text-decoration:none"> <img alt="Facebook" src="https://www.gstatic.com/alerts/images/fb-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html&amp;ss=tw&amp;rt=New+Radiotherapy+Prostate+Cancer+Treatment,+SpaceOAR%C2%AE+Hydrogel,+Now+Available+in+Japan&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmVWKJ7n0WfkJ51hiflfA7cFF-jOMA" style="text-decoration:none"> <img alt="Twitter" src="https://www.gstatic.com/alerts/images/tw-24.png" border="0" height="16" width="16"></a> </td> <td style="padding:0px 0px 6px 15px;font-family:Arial"> <a href="https://www.google.com/alerts/feedback?ffu=https://www.citizentribune.com/news/business/new-radiotherapy-prostate-cancer-treatment-spaceoar-hydrogel-now-available-in/article_70aa28c3-077d-5be1-b18f-4c472f67f8c7.html&amp;source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA" style="text-decoration:none;vertical-align:middle;color:#aaa;font-size:10px"> Flag as irrelevant </a> </td> </tr> </table>  </div> </div> </td> <td style="padding-right:18px"></td> </tr>    <tr itemscope="" itemtype="http://schema.org/Article"> <td style="padding-left:18px"></td> <td style="padding:18px 0px 12px 0px;vertical-align:top;border-top:1px solid #e4e4e4;font-family:Arial"> <a href="https://www.google.com/url?rct=j&amp;sa=t&amp;url=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;ct=ga&amp;cd=CAEYAioTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug&amp;usg=AFQjCNEIVTuJe91hxC-S-_WhBALuu20pzQ" style="text-decoration:none"> <table align="right" style="display:inline;border-collapse:collapse"> <tr> <td style="padding-left:18px"></td> <td background="https://img.youtube.com/vi/B8Pqnwuak0g/default.jpg" height="100" width="100" valign="bottom" style="padding:0px 0px 0px 0px;background-repeat:no-repeat;text-align:center;border:1px solid #e4e4e4"> <link href="https://img.youtube.com/vi/B8Pqnwuak0g/default.jpg" itemprop="image"> <!--[if gte mso 9]>
       <v:image xmlns:v="urn:schemas-microsoft-com:vml" id="theImage"
                style='behavior: url(#default#VML);
                       display:inline-block; position:absolute;
                       height: 100px; width: 100px;
                       top:0; left:0; border:0; z-index:1;' src="https://img.youtube.com/vi/B8Pqnwuak0g/default.jpg"/>
       <v:rect xmlns:v="urn:schemas-microsoft-com:vml"
               style='behavior: url(#default#VML);
                      display:inline-block;
                      position: absolute; top: 82px; left: 0px;
                      width:100px; border:0; z-index:2;' strokecolor="none">
       <v:fill opacity="50%" color="#000000"/>
       <v:textbox style="mso-fit-shape-to-text:t;
                         mso-column-margin: 0pt;
                         letter-spacing: 0.8px;"
                         inset="0pt,0pt,0pt,0pt">
         <font size="-1" color="#ffffff"></font>
       </v:textbox>
       </v:rect>
       <div style="display: none">
       <![endif]--> <div style="color:#fff;font-size:9px;letter-spacing:0.8px"> <div style="padding:3px 0px 4px 4px;background:rgb(255,255,255);background-color:rgba(0,0,0,0.5);width:96px"></div> </div> <!--[if gte mso 9]></div><![endif]--> </td> </tr> </table> </a> <div>  <span style="padding:0px 6px 0px 0px"> <a href="https://www.google.com/url?rct=j&amp;sa=t&amp;url=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;ct=ga&amp;cd=CAEYAioTMjQ4ODkzMTgwNTY3MTM2ODQyMTIcN2Q4MGExOWNlMTlhMTc4Njpjb206ZW46VVM6Ug&amp;usg=AFQjCNEIVTuJe91hxC-S-_WhBALuu20pzQ" itemprop="url" style="color:#427fed;display:inline;text-decoration:none;font-size:16px;line-height:20px"> <span itemprop="name">Dr. Humphrey on Mogamulizumab for Cutaneous T-Cell Lymphoma</span> </a> </span>  <div> <div style="padding:2px 0px 8px 0px"> <div itemprop="publisher" itemscope="" itemtype="http://schema.org/Organization" style="color:#737373;font-size:12px"> <a style="text-decoration:none;color:#737373"> <span itemprop="name">OncLive</span> </a> </div> <div itemprop="description" style="color:#252525;padding:2px 0px 0px 0px;font-size:12px;line-height:18px">... with cutaneous T-cell lymphoma (CTCL) that were reported in the <b>phase III</b> MAVORIC <b>trial</b>, which was presented at the 2018 ASCO Annual Meeting.</div> </div>   <table> <tr> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;ss=gp&amp;rt=Dr.+Humphrey+on+Mogamulizumab+for+Cutaneous+T-Cell+Lymphoma&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmWM1eUGOQO0fEhX7DKYYsnTD73J1A" style="text-decoration:none"> <img alt="Google Plus" src="https://www.gstatic.com/alerts/images/gp-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;ss=fb&amp;rt=Dr.+Humphrey+on+Mogamulizumab+for+Cutaneous+T-Cell+Lymphoma&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmWM1eUGOQO0fEhX7DKYYsnTD73J1A" style="text-decoration:none"> <img alt="Facebook" src="https://www.gstatic.com/alerts/images/fb-24.png" border="0" height="16" width="16"></a> </td> <td width="16" style="padding-right:6px"> <a href="https://www.google.com/alerts/share?hl=en&amp;gl=US&amp;ru=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;ss=tw&amp;rt=Dr.+Humphrey+on+Mogamulizumab+for+Cutaneous+T-Cell+Lymphoma&amp;cd=KhMyNDg4OTMxODA1NjcxMzY4NDIxMhw3ZDgwYTE5Y2UxOWExNzg2OmNvbTplbjpVUzpS&amp;ssp=AMJHsmWM1eUGOQO0fEhX7DKYYsnTD73J1A" style="text-decoration:none"> <img alt="Twitter" src="https://www.gstatic.com/alerts/images/tw-24.png" border="0" height="16" width="16"></a> </td> <td style="padding:0px 0px 6px 15px;font-family:Arial"> <a href="https://www.google.com/alerts/feedback?ffu=https://www.onclive.com/onclive-tv/dr-humphrey-on-mogamulizumab-for-cutaneous-tcell-lymphoma&amp;source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA" style="text-decoration:none;vertical-align:middle;color:#aaa;font-size:10px"> Flag as irrelevant </a> </td> </tr> </table>  </div> </div> </td> <td style="padding-right:18px"></td> </tr>    <tr> <td colspan="3" valign="middle" style="background-color:#f8f8f8;font-size:14px;vertical-align:middle;text-align:center;padding:10px 10px 10px 10px;line-height:20px;border:1px solid #e4e4e4;font-family:Arial"> <a href="https://www.google.com/alerts?s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA&amp;start=1530565046&amp;end=1530580312&amp;source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ#history" style="text-decoration:none;vertical-align:middle;color:#427fed">  See more results  </a> <span style="font-size:12px;padding-left:15px;padding-right:15px;color:#aaa">|</span> <a href="https://www.google.com/alerts/edit?source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA&amp;email=tiffanyfabianac%40gmail.com" style="text-decoration:none;vertical-align:middle;color:#427fed">Edit this alert</a>  </td> </tr>  </table> <table style="padding-top:6px;font-size:12px;color:#252525;text-align:center;width:100%"> <tr> <td style="font-family:Arial">  You have received this email because you have subscribed to <b>Google Alerts</b>. <div> <a href="https://www.google.com/alerts/remove?source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA" style="text-decoration:none;color:#427fed">Unsubscribe</a> <span style="padding:0px 4px 0px 4px;color:#252525">|</span> <a href="https://www.google.com/alerts?source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ" style="text-decoration:none;color:#427fed">  View all your alerts  </a> </div> </td> </tr> <tr> <td style="padding:6px 10px 0px 0px;font-family:Arial"> <a href="https://www.google.com/alerts/feeds/14520785336048043613/3506681673875074342" style="text-decoration:none;color:#427fed"> <img src="https://www.gstatic.com/alerts/images/rss-16.gif" alt="RSS" border="0" style="padding:0px 8px 0px 0px;vertical-align:middle"> <span style="display:inline;line-height:16px;vertical-align:middle"> Receive this alert as RSS feed </span> </a> </td> </tr> <tr> <td style="padding:6px 0px 0px 0px;font-family:Arial"> <a href="https://www.google.com/alerts?source=alertsmail&amp;hl=en&amp;gl=US&amp;msgid=MjQ4ODkzMTgwNTY3MTM2ODQyMQ&amp;s=AB2Xq4hvjI58wr9UK4tU7q0qsP1xftiU6kOl3SA&amp;ffu=" style="text-decoration:none;color:#427fed"> <div style="display:inline;line-height:16px;vertical-align:middle"> Send Feedback </div> </a> </td> </tr> </table> </div> </div> <!--[if mso]>
</td></tr></table>
<![endif]-->
   </div>  </body> </html>'}}]}, 'sizeEstimate': 32554}
'''
# Get the data from the message
payld = message['payload']
# Get Date
headr = payld['headers'] # get header of the payload
for two in headr:
if two['name'] == 'Date':
msg_date = two['value']
date_parse = (parser.parse(msg_date))
m_date = (date_parse.date())
msg_date = str(m_date)
else:
pass
# Get Message
parts = payld['parts']
for i in parts:
if i['partId'] == '0':
msg_str = base64.urlsafe_b64decode(i['body'].get('data'))
mime_msg = email.message_from_string(bytes.decode(msg_str))
else:
pass
# Remove the Unsubscribe and Google related links from message
msg_string = mime_msg.as_string()
msg_string = msg_string.rsplit('- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -', 1)[0]
msg_string = re.sub(r'===*.*===\n\n', '', msg_string)
# Extract Hyperlinks
myList =[]
# summaries = []
#urls = re.sub(r'>', '',re.search("(?P<url>https?://[^\s]+)", msg_string).group("url"))
urls = re.findall("(=https://www.*.*?>)",msg_string)
for i in range(len(urls)):
sURL = urls[i][1:-1]
summaries = re.sub(r'\n', ' ', msg_string.partition(urls[i])[0]).rsplit('<https://www.google.com', 1)[0]
if i != 0:
temp_link = urls[i-1][1:-1]
summaries = summaries.partition(temp_link)[2]
temp_dict = ({'Date': msg_date, 'URL': sURL, 'Summary': summaries})
myList.append(temp_dict)
print (myList) | [
"noreply@github.com"
] | noreply@github.com |
44f0a16cb6fcf836890a2a7383410fd334e0908d | e45c6f36a065b6a44e873a773428105de4d3758e | /bases/br_me_caged/code/caged_antigo/scpts/caged.py | 5a26195ed1c67ee805eb3133d1c765f73239f8c8 | [
"MIT"
] | permissive | basedosdados/mais | 080cef1de14376699ef65ba71297e40784410f12 | 2836c8cfad11c27191f7a8aca5ca26b94808c1da | refs/heads/master | 2023-09-05T20:55:27.351309 | 2023-09-02T03:21:02 | 2023-09-02T03:21:02 | 294,702,369 | 376 | 98 | MIT | 2023-08-30T21:17:28 | 2020-09-11T13:26:45 | SQL | UTF-8 | Python | false | false | 25,196 | py | """
Extract and clean CAGED data
"""
# pylint: disable=invalid-name,unnecessary-comprehension,import-error
import sys
import time
import os
import shutil
from contextlib import closing
from urllib import request
import pandas as pd
import numpy as np
import manipulation
sys.path.insert(0, "../")
def create_folder_structure():
"""
Create folder structure for CAGED data
"""
### cria pasta data caso n exista
if not os.path.exists("../data"):
os.mkdir("../data")
if not os.path.exists("../data/caged/"):
os.mkdir("../data/caged/")
if not os.path.exists("../data/caged/raw"):
os.mkdir("../data/caged/raw")
if not os.path.exists("../data/caged/raw/caged_antigo"):
os.mkdir("../data/caged/raw/caged_antigo")
if not os.path.exists("../data/caged/raw/caged_novo"):
os.mkdir("../data/caged/raw/caged_novo")
if not os.path.exists("../data/caged/raw/caged_antigo_ajustes"):
os.mkdir("../data/caged/raw/caged_antigo_ajustes")
if not os.path.exists("../data/caged/clean/"):
os.mkdir("../data/caged/clean")
if not os.path.exists("../data/caged/clean/caged_antigo"):
os.mkdir("../data/caged/clean/caged_antigo")
if not os.path.exists("../data/caged/clean/caged_novo"):
os.mkdir("../data/caged/clean/caged_novo")
if not os.path.exists("../data/caged/clean/caged_antigo_ajustes"):
os.mkdir("../data/caged/clean/caged_antigo_ajustes")
def download_data(download_link, download_path, filename):
"""
Download data from link and save it in path
"""
## download do arquivo
with closing(request.urlopen(download_link)) as r:
with open(os.path.join(download_path, filename), "wb") as f:
shutil.copyfileobj(r, f)
#####======================= ANTIGO CAGED DOWNLOAD =======================#####
def download_caged_normal(download_link, download_path_year, ano, mes):
"""
download dos arquivos do caged para os anos 2007 a 2019.
Tambem vale para arquivos de ajustes entre 2010 e 2019
"""
download_path_month = download_path_year + f"/{int(mes)}/"
## cria pastas
if os.path.exists(download_path_year):
if not os.path.exists(download_path_month):
os.mkdir(download_path_month)
else:
os.mkdir(download_path_year)
if not os.path.exists(download_path_month):
os.mkdir(download_path_month)
if "AJUSTE" not in download_link:
filename = f"CAGEDEST_{mes}{ano}.7z"
else:
filename = f"CAGEDEST_AJUSTES_{mes}{ano}.7z"
## verifica se arquivo ja existe
if os.path.exists(os.path.join(download_path_month, filename)):
print(f"{mes}/{ano} | já existe")
else:
try:
ti = time.time()
download_data(download_link, download_path_month, filename)
t = time.strftime("%M:%S", time.gmtime((time.time() - ti)))
print(f"{mes}/{ano} | criado em {t}")
except Exception:
print(f"{mes}/{ano} | não conseguiu baixar")
def download_caged_ajustes_2002a2009(download_link, download_path_year, ano):
"""
download dos arquivos de ajustes do caged para os anos 2002 a 2009
"""
if not os.path.exists(download_path_year):
os.mkdir(download_path_year)
os.mkdir(download_path_year + "/1")
download_path_year = download_path_year + "/1"
## verifica se arquivo ja existe
filename = filename = f"CAGEDEST_AJUSTES_{ano}.7z"
if os.path.exists(os.path.join(download_path_year, filename)):
print(f"{ano} | já existe")
else:
try:
ti = time.time()
download_data(download_link, download_path_year, filename)
t = time.strftime("%M:%S", time.gmtime((time.time() - ti)))
print(f"{ano} | criado em {t}")
except Exception:
print(f"{ano} | não conseguiu baixar")
def download_caged_file(download_link, ano=None, mes=None, raw_path=None):
"""
Download CAGED file
"""
# cria estrutura de pastas
create_folder_structure()
## cria link e path das pastas
download_path_year = raw_path + f"/{ano}"
if isinstance(ano, int):
download_caged_normal(download_link, download_path_year, ano, mes)
else:
download_caged_ajustes_2002a2009(download_link, download_path_year, ano)
def caged_antigo_download():
"""
Download CAGED antigo
"""
## define caminhos
raw_path = "../data/caged/raw/caged_antigo"
# seleciona anos e meses a serem baixados
anos = [i for i in range(2007, 2020)]
meses = ["01", "02", "03", "04", "05", "06", "07", "08", "09", "10", "11", "12"]
print("#===== CAGED ANTIGO =====#\n")
for ano in anos:
for mes in meses:
download_link = f"ftp://ftp.mtps.gov.br/pdet/microdados/CAGED/{ano}/CAGEDEST_{mes}{ano}.7z"
download_caged_file(download_link, ano, mes, raw_path)
print("\n")
def caged_antigo_ajustes_download():
"""
Download CAGED antigo ajustes
"""
## define caminhos
raw_path = "../data/caged/raw/caged_antigo_ajustes"
# seleciona anos e meses a serem baixados
anos = [i for i in range(2010, 2020)]
meses = ["01", "02", "03", "04", "05", "06", "07", "08", "09", "10", "11", "12"]
print("#===== CAGED ANTIGO AJUSTES =====#\n")
for ano in anos:
for mes in meses:
download_link = f"ftp://ftp.mtps.gov.br/pdet/microdados/CAGED_AJUSTES/{ano}/CAGEDEST_AJUSTES_{mes}{ano}.7z"
download_caged_file(download_link, ano, mes, raw_path)
print("\n")
def caged_antigo_ajustes_2002a2009_download():
"""
Ajustes CAGED antigo
"""
raw_path = "../data/caged/raw/caged_antigo_ajustes"
anos = [str(i) for i in range(2007, 2010)]
print("#===== CAGED ANTIGO AJUSTES 2002a2009 =====#\n")
for ano in anos:
download_link = f"ftp://ftp.mtps.gov.br/pdet/microdados/CAGED_AJUSTES/2002a2009/CAGEDEST_AJUSTES_{ano}.7z"
download_caged_file(download_link, ano, mes=None, raw_path=raw_path)
print("\n")
def caged_antigo_ajustes_2002a2009_extract_organize(folders, force_remove_csv=False):
"""
Os anos de 2002 a 2009 nao possuem arquivos para cada mes.
Essa funcão extrai o arquivo .7z e cria o .csv na mesma
estrutura de ano/mes
"""
folders = [
folder
for folder in folders
for ano in ["2007", "2008", "2009"]
if ano in folder
]
for folder in folders:
ano = folder.split("/")[-2]
filename_7z = get_file_names_and_clean_residues(folder, force_remove_csv)
extract_file(folder, filename_7z, save_rows=None)
print(f"{ano} | extraido")
filename = [file for file in os.listdir(folder) if ".csv" in file][0][:-4]
df = pd.read_csv(f"{folder}{filename}.csv", dtype="str")
df["ano"] = df["competencia_declarada"].apply(lambda x: str(x)[:4])
df["mes"] = df["competencia_declarada"].apply(lambda x: str(x)[4:])
for mes in df["mes"].unique():
if not os.path.exists(f"{folder}{str(int(mes))}"):
os.mkdir(f"{folder}{str(int(mes))}")
mask = df["mes"] == mes
dd = df[mask]
dd = dd.drop(["ano", "mes"], 1)
dd.to_csv(
f"{folder}{str(int(mes))}/{filename_7z[:-5]}_{mes}{ano}.csv",
index=False,
encoding="utf-8",
)
#####======================= MANIPULA ARQUIVOS =======================#####
def extract_file(path_month, filename, save_rows=10):
"""
Extrai arquivo .7z
"""
if not os.path.exists(f"{path_month}{filename}.csv"):
filename_txt = [
file for file in os.listdir(path_month) if ".txt" in file.lower()
][0]
try:
df = pd.read_csv(
f"{path_month}{filename_txt}",
sep=";",
encoding="latin-1",
nrows=save_rows,
dtype="str",
)
except Exception:
## caso de erro de bad lines por conter um ; extra no arquivo txt
with open(
f"{path_month}{filename_txt}",
encoding="latin-1",
) as f:
newText = f.read().replace(";99;", ";99")
with open(f"{path_month}{filename_txt}", "w", encoding="latin-1") as f:
f.write(newText)
df = pd.read_csv(
f"{path_month}{filename_txt}",
sep=";",
encoding="latin-1",
nrows=save_rows,
dtype="str",
)
df.columns = manipulation.normalize_cols(df.columns)
df.to_csv(f"{path_month}{filename}.csv", index=False, encoding="utf-8")
os.remove(f"{path_month}{filename_txt}")
def get_file_names_and_clean_residues(path_month, force_remove_csv=True):
"""
Get file names and clean residues
"""
filename_txt = [file for file in os.listdir(path_month) if ".txt" in file.lower()]
filename_csv = [file for file in os.listdir(path_month) if ".csv" in file]
try:
filename_7z = [file for file in os.listdir(path_month) if ".7z" in file][0][:-3]
except Exception:
filename_7z = filename_csv[0][:-4]
if filename_txt != []:
os.remove(f"{path_month}{filename_txt[0]}")
if filename_csv != [] and force_remove_csv is True:
os.remove(f"{path_month}{filename_csv[0][:-4]}.csv")
return filename_7z
def make_dirs(path, folder, var):
"""
Make dirs
"""
if not os.path.exists(f"{path}{var}={folder}/"):
os.mkdir(f"{path}{var}={folder}/")
def make_folder_tree(clean_path, ano, mes, uf="SP"):
"""
Make folder tree
"""
make_dirs(clean_path, ano, var="ano")
path_ano = f"{clean_path}/ano={ano}/"
make_dirs(path_ano, mes, var="mes")
path_mes = f"{clean_path}/ano={ano}/mes={mes}/"
make_dirs(path_mes, uf, var="sigla_uf")
return f"{clean_path}/ano={ano}/mes={mes}/sigla_uf={uf}/"
def save_partitioned_file(df, clean_save_path, ano, mes, file_name):
"""
Save partitioned file
"""
for uf in df.sigla_uf.unique():
## filtra apenas o estado de interesse
df_uf = df[df["sigla_uf"] == uf]
## exclui colunas particionadas
df_uf = df_uf.drop(["sigla_uf"], 1)
## cria estrutura de pastas
path_uf = make_folder_tree(clean_save_path, ano, mes, uf)
df_uf.to_csv(f"{path_uf}{file_name}.csv", index=False)
def caged_antigo_padronize_and_partitioned(
folders, clean_save_path, municipios, force_remove_csv=True
):
"""
Padroniza e particiona arquivos CAGED
"""
# all_cols = pd.DataFrame()
for folder in folders:
ano = folder.split("/")[-3]
mes = folder.split("/")[-2]
if "ajustes" in clean_save_path:
mode = "ajustes"
save_name = "caged_antigo_ajustes"
else:
mode = "padrao"
save_name = "caged_antigo"
if mode == "padrao" or (int(ano)) > 2009:
filename_7z = get_file_names_and_clean_residues(folder, force_remove_csv)
else:
filename_7z = get_file_names_and_clean_residues(folder, False)
## verifica se o arquivo ja foi tratado
if (
os.path.exists(f"{clean_save_path}/ano={ano}/mes={mes}")
and len(os.listdir(f"{clean_save_path}/ano={ano}/mes={mes}")) == 27
):
print(f"{ano}-{mes} | já tratado | arquivo {mode}\n")
else:
## extrai o arquivo zipado, cria um novo arquivo .csv e deleta o arquivo extraido (.txt)
if mode == "padrao" or (int(ano)) > 2009:
extract_file(folder, filename_7z, save_rows=None)
print(f"{ano}-{mes} | extraido | arquivo {mode}")
## le o arquivo
filename = [file for file in os.listdir(folder) if ".csv" in file][0][:-4]
df = pd.read_csv(
f"{folder}{filename}.csv",
dtype="str",
)
## padronizacao dos dados | caso seja o Path de ajustes chama funcao de padronizacao de ajustes
if mode == "ajustes":
df = padroniza_caged_antigo_ajustes(df, municipios)
else:
df = padroniza_caged_antigo(df, municipios)
print(f"{ano}-{mes} | padronizado | arquivo {mode}")
save_partitioned_file(df, clean_save_path, ano, mes, file_name=save_name)
# remove arquivo csv do raw files
if mode == "padrao" or (int(ano)) > 2009:
os.remove(f"{folder}{filename}.csv")
print(f"{ano}-{mes} | finalizado | arquivo {mode}\n")
# # checa nome de todas as colunas
# dd = pd.DataFrame(df.columns.tolist(), columns=["cols"])
# dd = dd.transpose().reset_index(drop=True)
# cols = dd.columns.tolist()
# dd["ano"] = ano
# dd["mes"] = mes
# dd = dd[["ano", "mes"] + cols]
# all_cols = pd.concat([all_cols, dd])
#####======================= PADRONIZA CAGED FUNCOES =======================#####
def clean_caged(df, municipios):
"""
Clean CAGED
"""
## cria coluna ano e mes apartir da competencia declarada
df["competencia_declarada"] = df["competencia_declarada"].apply(
lambda x: str(x)[:4] + "-" + str(x)[4:]
)
if "competencia_movimentacao" in df.columns.tolist():
df["competencia_movimentacao"] = df["competencia_movimentacao"].apply(
lambda x: str(x)[:4] + "-" + str(x)[4:]
)
# renomeia municipio para padrao do diretorio de municipios
rename_cols = {
"municipio": "id_municipio_6",
}
df = df.rename(columns=rename_cols)
# adiciona id_municio do diretorio de municipios
df = df.merge(municipios, on="id_municipio_6", how="left")
# remove strings do tipo {ñ e converte salario e tempo de emprego para float
objct_cols = df.select_dtypes(include=["object"]).columns.tolist()
for col in objct_cols:
if col in ["salario_mensal", "tempo_emprego"]:
df[col] = pd.to_numeric(
df[col].str.replace(",", "."), downcast="float", errors="coerce"
)
else:
df[col] = np.where(df[col].str.contains("{ñ"), np.nan, df[col])
df = df.drop(["mesorregiao", "microrregiao", "uf", "competencia_declarada"], 1)
df = df[df["sigla_uf"].notnull()]
return df
#####======================= PADRONIZA DADOS CAGED AJUSTES =======================#####
def padroniza_caged_antigo_ajustes(df, municipios):
"""
Padroniza dados CAGED ajustes
"""
## cria colunas que nao existem em outros arquivos
check_cols = ["ind_trab_parcial", "ind_trab_intermitente"]
create_cols = [col for col in check_cols if col not in df.columns.tolist()]
for col in create_cols:
df[col] = np.nan
hard_coded_cols = [
"admitidos_desligados",
"competencia_movimentacao",
"municipio",
"ano_movimentacao",
"cbo_94_ocupacao",
"cbo_2002_ocupacao",
"cnae_10_classe",
"faixa_empr_inicio_jan",
"grau_instrucao",
"qtd_hora_contrat",
"ibge_subsetor",
"idade",
"ind_aprendiz",
"salario_mensal",
"saldo_mov",
"sexo",
"tempo_emprego",
"tipo_estab",
"tipo_mov_desagregado",
"uf",
"competencia_declarada",
"bairros_sp",
"bairros_fortaleza",
"bairros_rj",
"distritos_sp",
"regioes_adm_df",
"mesorregiao",
"microrregiao",
"regiao_adm_rj",
"regiao_adm_sp",
"regiao_corede",
"regiao_corede_04",
"regiao_gov_sp",
"regiao_senac_pr",
"regiao_senai_pr",
"regiao_senai_sp",
"subregiao_senai_pr",
"regiao_metro_mte",
"cnae_20_subclas",
"raca_cor",
"ind_portador_defic",
"tipo_defic",
"ind_trab_parcial",
"ind_trab_intermitente",
]
df = df[hard_coded_cols]
# df.columns = hard_coded_cols
#### remove typos e define tipos
df = clean_caged(df, municipios)
df = df.drop(["ano_movimentacao"], 1)
organize_cols = [
"competencia_movimentacao",
"sigla_uf",
"id_municipio",
"id_municipio_6",
"admitidos_desligados",
"tipo_estab",
"tipo_mov_desagregado",
"faixa_empr_inicio_jan",
"tempo_emprego",
"qtd_hora_contrat",
"salario_mensal",
"saldo_mov",
"ind_aprendiz",
"ind_trab_intermitente",
"ind_trab_parcial",
"ind_portador_defic",
"tipo_defic",
"cbo_94_ocupacao",
"cnae_10_classe",
"cbo_2002_ocupacao",
"cnae_20_subclas",
"grau_instrucao",
"idade",
"sexo",
"raca_cor",
"ibge_subsetor",
"bairros_sp",
"bairros_fortaleza",
"bairros_rj",
"distritos_sp",
"regioes_adm_df",
"regiao_adm_rj",
"regiao_adm_sp",
"regiao_corede",
"regiao_corede_04",
"regiao_gov_sp",
"regiao_senac_pr",
"regiao_senai_pr",
"regiao_senai_sp",
"subregiao_senai_pr",
"regiao_metro_mte",
]
df = df[organize_cols]
columns_rename = {
"cbo_2002_ocupacao": "cbo_2002",
"cbo_94_ocupacao": "cbo_1994",
"cnae_10_classe": "cnae_1",
"cnae_20_subclas": "cnae_2_subclasse",
"faixa_empr_inicio_jan": "faixa_emprego_inicio_janeiro",
"qtd_hora_contrat": "quantidade_horas_contratadas",
"ibge_subsetor": "subsetor_ibge",
"ind_aprendiz": "indicador_aprendiz",
"ind_portador_defic": "indicador_portador_deficiencia",
"saldo_mov": "saldo_movimentacao",
"tipo_estab": "tipo_estabelecimento",
"tipo_defic": "tipo_deficiencia",
"tipo_mov_desagregado": "tipo_movimentacao_desagregado",
"regioes_adm_df": "regiao_administrativas_df",
"regiao_adm_rj": "regiao_administrativas_rj",
"regiao_adm_sp": "regiao_administrativas_sp",
"ind_trab_parcial": "indicador_trabalho_parcial",
"ind_trab_intermitente": "indicador_trabalho_intermitente",
"regiao_metro_mte": "regiao_metropolitana_mte",
}
df = df.rename(columns=columns_rename)
return df
#####======================= PADRONIZA DADOS CAGED ANTIGO =======================#####
def padroniza_caged_antigo(df, municipios):
"""
Padroniza dados do CAGED antigo
"""
## cria colunas que nao existem em outros arquivos
check_cols = ["ind_trab_parcial", "ind_trab_intermitente"]
create_cols = [col for col in check_cols if col not in df.columns.tolist()]
for col in create_cols:
df[col] = np.nan
## renomeia colunas para alguns casos typo nos arquivos originais
rename_typo_columns = {
"competaancia_declarada": "competencia_declarada",
"municapio": "municipio",
"cbo_2002_ocupaaao": "cbo_2002_ocupacao",
"faixa_empr_inacio_jan": "faixa_empr_inicio_jan",
"grau_instruaao": "grau_instrucao",
"raaa_cor": "raca_cor",
"regiaes_adm_df": "regioes_adm_df",
}
df = df.rename(columns=rename_typo_columns)
hard_coded_cols = [
"admitidos_desligados",
"competencia_declarada",
"municipio",
"ano_declarado",
"cbo_2002_ocupacao",
"cnae_10_classe",
"cnae_20_classe",
"cnae_20_subclas",
"faixa_empr_inicio_jan",
"grau_instrucao",
"qtd_hora_contrat",
"ibge_subsetor",
"idade",
"ind_aprendiz",
"ind_portador_defic",
"raca_cor",
"salario_mensal",
"saldo_mov",
"sexo",
"tempo_emprego",
"tipo_estab",
"tipo_defic",
"tipo_mov_desagregado",
"uf",
"bairros_sp",
"bairros_fortaleza",
"bairros_rj",
"distritos_sp",
"regioes_adm_df",
"mesorregiao",
"microrregiao",
"regiao_adm_rj",
"regiao_adm_sp",
"regiao_corede",
"regiao_corede_04",
"regiao_gov_sp",
"regiao_senac_pr",
"regiao_senai_pr",
"regiao_senai_sp",
"subregiao_senai_pr",
"ind_trab_parcial",
"ind_trab_intermitente",
]
df = df[hard_coded_cols]
# df.columns = hard_coded_cols
#### remove typos e define tipos
df = clean_caged(df, municipios)
df = df.drop(["ano_declarado"], 1)
organize_cols = [
"sigla_uf",
"id_municipio",
"id_municipio_6",
"admitidos_desligados",
"tipo_estab",
"tipo_mov_desagregado",
"faixa_empr_inicio_jan",
"tempo_emprego",
"qtd_hora_contrat",
"salario_mensal",
"saldo_mov",
"ind_aprendiz",
"ind_trab_intermitente",
"ind_trab_parcial",
"ind_portador_defic",
"tipo_defic",
"cnae_10_classe",
"cbo_2002_ocupacao",
"cnae_20_classe",
"cnae_20_subclas",
"grau_instrucao",
"idade",
"sexo",
"raca_cor",
"ibge_subsetor",
"bairros_sp",
"bairros_fortaleza",
"bairros_rj",
"distritos_sp",
"regioes_adm_df",
"regiao_adm_rj",
"regiao_adm_sp",
"regiao_corede",
"regiao_corede_04",
"regiao_gov_sp",
"regiao_senac_pr",
"regiao_senai_pr",
"regiao_senai_sp",
"subregiao_senai_pr",
]
df = df[organize_cols]
columns_rename = {
"cbo_2002_ocupacao": "cbo_2002",
"cnae_10_classe": "cnae_1",
"cnae_20_classe": "cnae_2",
"cnae_20_subclas": "cnae_2_subclasse",
"faixa_empr_inicio_jan": "faixa_emprego_inicio_janeiro",
"qtd_hora_contrat": "quantidade_horas_contratadas",
"ibge_subsetor": "subsetor_ibge",
"ind_aprendiz": "indicador_aprendiz",
"ind_portador_defic": "indicador_portador_deficiencia",
"saldo_mov": "saldo_movimentacao",
"tipo_estab": "tipo_estabelecimento",
"tipo_defic": "tipo_deficiencia",
"tipo_mov_desagregado": "tipo_movimentacao_desagregado",
"regioes_adm_df": "regiao_administrativas_df",
"regiao_adm_rj": "regiao_administrativas_rj",
"regiao_adm_sp": "regiao_administrativas_sp",
"ind_trab_parcial": "indicador_trabalho_parcial",
"ind_trab_intermitente": "indicador_trabalho_intermitente",
}
df = df.rename(columns=columns_rename)
return df
#####======================= PADRONIZA DADOS NOVO CAGED =======================#####
def padroniza_caged_novo(df, municipios):
"""
Padroniza dados do CAGED novo
"""
## cria coluna ano e mes apartir da competencia declarada
df["ano"] = df["competencia_declarada"].apply(lambda x: int(str(x)[:4]))
df["mes"] = df["competencia_declarada"].apply(lambda x: int(str(x)[4:]))
## cria colunas que nao existem em outros arquivos
df = df.drop(
["competencia_declarada", "uf", "regiao"],
1,
)
df = clean_caged(df, municipios)
return df
def rename_caged_novo(df, option):
"""
Renomeia colunas do CAGED novo
"""
rename_cols_estabelecimentos = {
"competaancia": "competencia_declarada",
"regiao": "regiao",
"uf": "uf",
"municapio": "municipio",
"seaao": "cnae_20_classe",
"subclasse": "cnae_20_subclas",
"admitidos": "admitidos",
"desligados": "desligados",
"fonte_desl": "fonte_desligamento",
"saldomovimentaaao": "saldo_movimentacao",
"tipoempregador": "tipo_empregador",
"tipoestabelecimento": "tipo_estab",
"tamestabjan": "faixa_empr_inicio_jan",
}
rename_cols_movimentacao = {
"competaancia": "competencia_declarada",
"regiao": "regiao",
"uf": "uf",
"municapio": "municipio",
"seaao": "cnae_20_classe",
"subclasse": "cnae_20_subclas",
"saldomovimentaaao": "saldo_mov",
"cbo2002ocupaaao": "cbo_2002_ocupacao",
"categoria": "categoria_trabalhador",
"graudeinstruaao": "grau_instrucao",
"idade": "idade",
"horascontratuais": "qtd_hora_contrat",
"raaacor": "raca_cor",
"sexo": "sexo",
"tipoempregador": "tipo_empregador",
"tipoestabelecimento": "tipo_estab",
"tipomovimentaaao": "tipo_movimentacao",
"tipodedeficiaancia": "tipo_defic",
"indtrabintermitente": "ind_trab_intermitente",
"indtrabparcial": "ind_trab_parcial",
"salario": "salario_mensal",
"tamestabjan": "faixa_empr_inicio_jan",
"indicadoraprendiz": "ind_aprendiz",
"fonte": "fonte_movimentacao",
}
if option == "movimentacao":
df = df.rename(columns=rename_cols_movimentacao)
else:
df = df.rename(columns=rename_cols_estabelecimentos)
return df
| [
"noreply@github.com"
] | noreply@github.com |
cc5e512c40eeb3d7ecfa986f9e8245d9a2c05f43 | 48b2fd9da2b8c67b1d732dae445c21cd6d878504 | /warrior/test_WarriorCore/test_defects_driver.py | 37eaf27fccb5059193d351b85a125d7a5aa51870 | [
"Apache-2.0"
] | permissive | shubhendra-tomar/warriorframework_py3 | 6ca4f59cd44c2420c48c7ccab505ffdcc76820fd | fc268c610c429f5a60e5627c2405aa66036487dd | refs/heads/master | 2023-07-17T09:34:35.025678 | 2021-05-10T06:43:57 | 2021-05-10T06:43:57 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,423 | py | """
Copyright 2017, Fujitsu Network Communications, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import sys
import os
from os.path import abspath, dirname
from unittest import TestCase
try:
import warrior
# except ModuleNotFoundError as error:
except Exception as e:
WARRIORDIR = dirname(dirname(dirname(abspath(__file__))))
print(WARRIORDIR)
sys.path.append(WARRIORDIR)
import warrior
from warrior.WarriorCore.defects_driver import DefectsDriver
temp_cwd = os.path.split(__file__)[0]
path = os.path.join(temp_cwd, 'UT_results')
try:
os.makedirs(path, exist_ok=True)
result_dir = os.path.join(dirname(abspath(__file__)), 'UT_results')
except OSError as error:
pass
class test_DefectsDriver(TestCase):
""" Defects Driver Class """
def test_get_defect_json_list(self):
"""Gets the list of defect json files for the testcase execution """
wt_resultfile = os.path.join(os.path.split(__file__)[0], "defects_driver_results_tc.xml")
wt_defectsdir = result_dir
with open(result_dir+'/'+'myfile.log', 'w'):
pass
wt_logsdir = os.path.join(result_dir, 'myfile.log')
wt_testcase_filepath = os.path.join(os.path.split(__file__)[0], "defects_driver_tc.xml")
data_repository = {'wt_resultfile':wt_resultfile, 'wt_defectsdir':wt_defectsdir, \
'wt_logsdir':wt_logsdir, 'wt_testcase_filepath': wt_testcase_filepath, 'jiraproj':None}
cls_obj = DefectsDriver(data_repository)
result = cls_obj.get_defect_json_list()
assert type(result) == list
def test_create_failing_kw_json_without_keywords(self):
"""Create a json file each failing keyword """
wt_resultfile = os.path.join(os.path.split(__file__)[0], "defects_driver_results_tc1.xml")
wt_defectsdir = result_dir
with open(result_dir+'/'+'myfile.log', 'w'):
pass
wt_logsdir = os.path.join(result_dir, 'myfile.log')
wt_testcase_filepath = os.path.join(os.path.split(__file__)[0], "defects_driver_tc1.xml")
data_repository = {'wt_resultfile':wt_resultfile, 'wt_defectsdir':wt_defectsdir, \
'wt_logsdir':wt_logsdir, 'wt_testcase_filepath': wt_testcase_filepath, 'jiraproj':None}
cls_obj = DefectsDriver(data_repository)
result = cls_obj.create_failing_kw_json()
assert result == None
def test_create_failing_kw_json_no_failed_keywords(self):
"""Create a json file each failing keyword """
wt_resultfile = os.path.join(os.path.split(__file__)[0], "defects_driver_results_tc2.xml")
wt_defectsdir = result_dir
with open(result_dir+'/'+'myfile.log', 'w'):
pass
wt_logsdir = os.path.join(result_dir, 'myfile.log')
wt_testcase_filepath = os.path.join(os.path.split(__file__)[0], "defects_driver_tc2.xml")
data_repository = {'wt_resultfile':wt_resultfile, 'wt_defectsdir':wt_defectsdir, \
'wt_logsdir':wt_logsdir, 'wt_testcase_filepath': wt_testcase_filepath, 'jiraproj':None}
cls_obj = DefectsDriver(data_repository)
result = cls_obj.create_failing_kw_json()
assert result == False
def test_create_failing_kw_json(self):
"""Create a json file each failing keyword """
wt_resultfile = os.path.join(os.path.split(__file__)[0], "defects_driver_results_tc.xml")
wt_defectsdir = result_dir
with open(result_dir+'/'+'myfile.log', 'w'):
pass
wt_logsdir = os.path.join(result_dir, 'myfile.log')
wt_testcase_filepath = os.path.join(os.path.split(__file__)[0], "defects_driver_tc.xml")
data_repository = {'wt_resultfile':wt_resultfile, 'wt_defectsdir':wt_defectsdir, \
'wt_logsdir':wt_logsdir, 'wt_testcase_filepath': wt_testcase_filepath, 'jiraproj':None}
cls_obj = DefectsDriver(data_repository)
result = cls_obj.create_failing_kw_json()
assert result == True | [
"venkatadhri.kotakonda@us.fujitsu.com"
] | venkatadhri.kotakonda@us.fujitsu.com |
878f1d4c667f843e3a31085b1d0d6aba7e6df757 | 3deaa908ee9bd781921f5581182b5537b6a340ea | /P2-plagiarism-detection/source_sklearn/train.py | 24656b134ac5337d07152183af799aa2c4706d5a | [
"MIT"
] | permissive | suryasanchez/machine-learning-engineer-nanodegree | b68372d73b59eb7cb5d0a11e9a55dca0e361bbd1 | 8bb4c7b1258dd1aad95011d9fcb25546ed9d9324 | refs/heads/master | 2020-12-27T20:53:33.373136 | 2020-03-29T03:31:21 | 2020-03-29T03:31:21 | 238,049,325 | 7 | 3 | null | null | null | null | UTF-8 | Python | false | false | 2,130 | py | from __future__ import print_function
import argparse
import os
import pandas as pd
from sklearn.externals import joblib
## TODO: Import any additional libraries you need to define a model
from sklearn import svm
# Provided model load function
def model_fn(model_dir):
"""Load model from the model_dir. This is the same model that is saved
in the main if statement.
"""
print("Loading model.")
# load using joblib
model = joblib.load(os.path.join(model_dir, "model.joblib"))
print("Done loading model.")
return model
## TODO: Complete the main code
if __name__ == '__main__':
# All of the model parameters and training parameters are sent as arguments
# when this script is executed, during a training job
# Here we set up an argument parser to easily access the parameters
parser = argparse.ArgumentParser()
# SageMaker parameters, like the directories for training data and saving models; set automatically
# Do not need to change
parser.add_argument('--output-data-dir', type=str, default=os.environ['SM_OUTPUT_DATA_DIR'])
parser.add_argument('--model-dir', type=str, default=os.environ['SM_MODEL_DIR'])
parser.add_argument('--data-dir', type=str, default=os.environ['SM_CHANNEL_TRAIN'])
## TODO: Add any additional arguments that you will need to pass into your model
parser.add_argument('--epochs', type=int, default=1000)
# args holds all passed-in arguments
args = parser.parse_args()
# Read in csv training file
training_dir = args.data_dir
train_data = pd.read_csv(os.path.join(training_dir, "train.csv"), header=None, names=None)
# Labels are in the first column
train_y = train_data.iloc[:,0]
train_x = train_data.iloc[:,1:]
## --- Your code here --- ##
## TODO: Define a model
model = svm.LinearSVC(max_iter=args.epochs)
## TODO: Train the model
model.fit(train_x, train_y)
## --- End of your code --- ##
# Save the trained model
joblib.dump(model, os.path.join(args.model_dir, "model.joblib")) | [
"suryasanchez@outlook.com"
] | suryasanchez@outlook.com |
43c9fbe98cd91e26a7efe54a86ba1b5103909a08 | 328581c2430f061e379192dacefc0f272f3ef27a | /scifin/exceptions/exceptions.py | 3d5423cf70e34417429b357d0aede8f648625eec | [
"MIT"
] | permissive | sg48/SciFin | e6b70d03d86d4e44caf977b70981678a1c6ed7ce | e4e3d1a32e060c911e43d89df833b3ad078ba6df | refs/heads/master | 2023-03-20T03:49:23.190132 | 2020-10-17T07:29:10 | 2020-10-17T07:29:10 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 421 | py | # Created on 2020/8/14
# This module is for storing built-in classes for exceptions.
class AccessError(Exception):
"""Raised when a file or url cannot be accessed."""
pass
class SamplingError(Exception):
"""Raised when the sampling of a time series is not uniform."""
pass
class ArgumentsError(Exception):
"""Raised when provided arguments of a function are not satisfactory."""
pass
| [
"fabien.nugier@googlemail.com"
] | fabien.nugier@googlemail.com |
7ea0060a4ae2c77bf6b62d6084705a5bd61a261b | ef458faffdf14ee518f479f0db0b19ba18e8596f | /send.py | ea5159063fbae4a59c92126919d8a6025dcc4520 | [] | no_license | shadowkernel/PyMailStress | eafc2fe87db25b60957ea3ab0fe35dd99b747e4c | 1dce4f066ea7f38ddbdc94b574b37a260fd19cc5 | refs/heads/master | 2021-01-10T04:20:00.086020 | 2013-03-05T06:30:49 | 2013-03-05T06:30:49 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,984 | py | #!/usr/bin/python
# -*- coding: utf-8 -*-
# Send mail script
# Author: Xiaoyang Li <hsiaoyonglee@gmail.com>
import smtplib
import time
import threading
import ConfigParser
from email.mime.text import MIMEText
from email.mime.image import MIMEImage
from email.mime.multipart import MIMEMultipart
config = ConfigParser.ConfigParser()
config.read('config')
server_addr = config.get('server', 'smtp')
port = 587
username = config.get('account', 'username')
password = config.get('account', 'password')
To = config.get('account', 'to')
From = username
thread_count = int(config.get('send_model', 'thread_count'))
send_count = int(config.get('send_model', 'send_count'))
login_times = int(config.get('send_model', 'login_times'))
login_interval=int(config.get('send_model', 'login_interval'))
fp = open('att.png', 'rb')
img = MIMEImage(fp.read())
fp.close()
msg = MIMEMultipart()
msg['Subject'] = 'TEST'
msg['From'] = "Joe Jeong"
msg['To'] = "Whatever"
content = "It's a Long Way to the Top if You Wanna Rock'N'Roll!!!!"
msg.attach(MIMEText(content,'plain'));
msg.attach(img)
class SendMail(threading.Thread):
def __init__(self):
threading.Thread.__init__(self)
def send_mail(self, usingTLS=True):
if usingTLS == True :
server = smtplib.SMTP(server_addr, port)
server.ehlo()
server.starttls()
server.ehlo()
server.login(username, password)
for x in xrange(send_count):
try:
server.sendmail(From, To, msg.as_string())
except:
pass
server.close()
def run(self):
for i in range(login_times):
# logging
self.send_mail()
time.sleep(login_interval)
def go():
threads = [SendMail() for i in xrange(thread_count)]
for thread in threads:
thread.start()
for thread in threads:
thread.join()
if __name__ == "__main__":
go()
| [
"Joe.Jeong@icloud.com"
] | Joe.Jeong@icloud.com |
f1143c5015ebcd88c7912f5c54bdc340466fa1fe | 2172a26b843ee1466d6c33e805335dadbd853ed4 | /p097.py | 3b11cf862ac80ce2d27fa3bf79ff3838bfa82801 | [] | no_license | stewSquared/project-euler | aca2b8a610954c7e0eb2cab90d48cbbca839ad4a | 9369b4a6b6d97b2baeb44e81424e4f2e0aa0f870 | refs/heads/master | 2020-12-30T11:03:35.004325 | 2018-01-31T06:36:57 | 2018-01-31T06:36:57 | 21,247,626 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 142 | py | from functools import reduce
from itertools import repeat
ans = reduce(lambda m, n: m*n % 10**10, repeat(2, 7830457), 28433) + 1
print(ans)
| [
"stewinsalot@gmail.com"
] | stewinsalot@gmail.com |
e077975f79cef8246e26e09b0c121db7baa3577b | 5256a2412950c9c9df5ee7e37fdf748f34e0ff1c | /tests/system/springboothystrixbeat.py | 296d59ade0581021a0d74c0a812ced19fdf9fd23 | [
"Apache-2.0"
] | permissive | defus/springboothystrixbeat | 8276abdd687c43aedfb0ebb788000ba28383cfa5 | a85f8c33789669a2773e77126cf139d1ece4561a | refs/heads/master | 2020-08-23T02:13:25.761545 | 2019-10-22T15:50:06 | 2019-10-22T15:50:06 | 216,521,528 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 386 | py | import os
import sys
sys.path.append('../../vendor/github.com/elastic/beats/libbeat/tests/system')
from beat.beat import TestCase
class BaseTest(TestCase):
@classmethod
def setUpClass(self):
self.beat_name = "springboothystrixbeat"
self.beat_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../"))
super(BaseTest, self).setUpClass()
| [
"ldefokuate@octo.com"
] | ldefokuate@octo.com |
cdf6845bb895f5e4820d2059d293dc3b6e70ad83 | 58af4e8d6a41fd10d3e8d9a1e36e589729a50dd2 | /Networker.py | 85651068695f321012e4154402acfda8c31c39ce | [] | no_license | intrabit/blacklight | ff3087e8d58cb9f8857423fd13fabe3e75325bbd | 53397f709332a5fed3cfafd835d1c691d7ca68fa | refs/heads/master | 2020-12-30T09:58:30.844038 | 2019-09-05T19:41:21 | 2019-09-05T19:41:21 | 99,246,659 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,216 | py | # Connects to webservers requesting page data.
# Written by Louis Kennedy
import gzip
import re
import ssl
import socket
import threading
import Settings as settings
def finddata(source):
lastfind = 0
current = 0
while current != -1:
lastfind = current
current = source.find(b"\r\n", lastfind + 1)
else:
return lastfind + 2
def getheader(data, header):
start_index = data.find(header)
end_index = data.find(b"\r\n", start_index)
try:
return (data[start_index + len(header):end_index]).strip()
except Exception as exception:
raise Exception("Header Doesn't Exist")
def decoderesponse(data):
encoding_type = getheader(data, b"Content-Encoding:")
temp_data = data[0:int(len(data) / 10)]
data_end = finddata(temp_data)
data = data[data_end:]
data = data[:len(data) - 7]
if encoding_type == b"gzip":
decoded_response = gzip.decompress(data)
decoded_response = str(decoded_response, settings.DEFAULT_ENCODING)
return decoded_response
else:
data = str(data, settings.DEFAULT_ENCODING)
return data
def requestpage(address, retryCounter = 0):
resource = address[address.find("/"):]
csocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
port = 80
if settings.ENCRYPTED:
csocket = ssl.wrap_socket(csocket)
port = 443
csocket.settimeout(settings.SOCKET_TIMEOUT)
try:
csocket.connect((socket.gethostbyname(settings.BASE_SERVER), port))
except socket.timeout:
if retryCounter < settings.MAX_RETRIES:
print("Network Error: Attempt to connect failed. Retrying...\n")
requestpage(address, retryCounter + 1)
else:
print("Network Error: Max retries reached. Aborting server request.\n")
request = "GET " + resource + " HTTP/1.1\r\n" + "User-Agent: Blacklight/" + \
settings.VERSION + "\r\nHost: " + settings.BASE_SERVER + \
"\r\nAccept-Language: en-us\r\nAccept-Encoding: gzip\r\nAccept-Charset: utf-8\r\n\r\n"
csocket.send(request.encode(settings.DEFAULT_ENCODING))
encoded_response = b''
while True:
try:
chunk = csocket.recv(settings.RESPONSE_BUFFER_SIZE)
if not chunk:
break
encoded_response += chunk
except socket.timeout:
break
csocket.close()
if encoded_response != b'':
code = re.search(b"([0-9]{3} .+?)\r\n", encoded_response)
if code:
code = (code.group()).rstrip(b"\n\r")
if code == b"200 OK":
try:
return decoderesponse(encoded_response)
except:
return None
elif code == b"404 Not Found":
raise Exception("Page Not Found")
elif code == b"301 Moved Permanently" or code == b"302 Moved Temporarily":
try:
location = getheader(encoded_response, b"Location:")
if location.find(settings.MAIN_BRANCH) != -1:
check = (b"https://" if settings.ENCRYPTED else b"http://") + bytes(settings.BASE_SERVER, settings.DEFAULT_ENCODING) + bytes(resource, settings.DEFAULT_ENCODING)
if location != check:
if location.find(b"https") != -1:
settings.ENCRYPTED = True
elif location.find(b"http") != -1:
settings.ENCRYPTED = False
return requestpage(settings.BASE_SERVER + resource)
else:
raise Exception("Cyclical Redirection Detected") # The relocation link is identical to the original link followed.
else:
raise Exception("Invalid Link")
except:
raise Exception("Relocation URI Is Invalid")
else:
raise Exception("Unhandled Code: " + str(code))
else:
raise Exception("Corrupt HTTP Response")
else:
raise Exception("Couldn't Find HTTP Status Code Header")
| [
"noreply@github.com"
] | noreply@github.com |
adaefc6cb4c5988401db3a5e39c183ad7f465be6 | d4d162ee2ff1fd68121ec237bf8d19107abc9ae8 | /Arsen_Osipyan/game_of_life/styles.py | 4216118276e7ebabc28f5a49f5f031c3e941ad36 | [] | no_license | droidroot1995/DAFE_Python_015 | 48e1571cebd27a6bd978e3020b878f699a54c92c | f6c739e3be40e7737a173e1fdf327e10c5ae0cab | refs/heads/main | 2023-03-05T14:47:51.014141 | 2020-12-17T17:54:53 | 2020-12-17T17:54:53 | 308,549,551 | 1 | 21 | null | 2021-02-23T18:29:47 | 2020-10-30T06:57:00 | Python | UTF-8 | Python | false | false | 1,169 | py | # Window styles
WINDOW_CSS = "background-color: #fff;" \
"font-family: Arial;" \
"text-transform: uppercase;" \
"font-weight: 800;"
# Buttons styles
BUTTON_CSS = "border-radius: 0.4em;" \
"font-size: 14px;" \
"color: #fff;" \
"border: 0;"
START_CSS = BUTTON_CSS + \
"background-color: #007bff;"
CLEAR_CSS = BUTTON_CSS + \
"background-color: #007bff;"
CLOSE_CSS = BUTTON_CSS + \
"background-color: #dc3545;"
CONFIRM_CSS = BUTTON_CSS + \
"background-color: #007bff;"
# Status bar styles
STATUS_CSS = "color: #333;" \
"font-size: 12px;" \
"font-weight: 800;"
# Input styles
INPUT_CSS = "background-color: #ddd;" \
"border: 1px solid #eee;" \
"border-radius: 0.4em;" \
"color: #333;" \
"font-size: 14px;"
# Checkbox styles
CHECKBOX_CSS = "color: #333;" \
"align: center;"
# Cells styles
CELL_CSS = "border: 1px solid #eee;"
DEAD_CSS = CELL_CSS + \
"background-color: #ddd;"
ALIVE_CSS = CELL_CSS + \
"background-color: #28A745;"
| [
"arsenosipan02work@gmail.com"
] | arsenosipan02work@gmail.com |
0083e93218f93976537335daf228c1432b37276d | 0f2b51f85685b8f165b05854bf657f918bafb976 | /NER_train_predict/bert.py | 9a396b7af665652b701661bde9f5ce664a9986f1 | [] | no_license | Zenodia/NLP_Bert_Pytorch_NER_task | dfade7086fb2a2ae101db43f13d551e56a5cbe24 | 087efbb86dd0e84203dd07a507551caa23689f3a | refs/heads/master | 2022-12-14T21:13:04.091027 | 2020-09-16T11:46:59 | 2020-09-16T11:46:59 | 296,013,848 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,560 | py | """BERT NER Inference."""
from __future__ import absolute_import, division, print_function
import json
import os
import torch
import torch.nn.functional as F
from nltk import word_tokenize
from pytorch_transformers import (BertConfig, BertForTokenClassification,
BertTokenizer)
class BertNer(BertForTokenClassification):
def forward(self, input_ids, token_type_ids=None, attention_mask=None, valid_ids=None):
sequence_output = self.bert(input_ids, token_type_ids, attention_mask, head_mask=None)[0]
batch_size,max_len,feat_dim = sequence_output.shape
valid_output = torch.zeros(batch_size,max_len,feat_dim,dtype=torch.float32,device='cuda' if torch.cuda.is_available() else 'cpu')
for i in range(batch_size):
jj = -1
for j in range(max_len):
if valid_ids[i][j].item() == 1:
jj += 1
valid_output[i][jj] = sequence_output[i][j]
sequence_output = self.dropout(valid_output)
logits = self.classifier(sequence_output)
return logits
class Ner:
def __init__(self,model_dir: str):
self.model , self.tokenizer, self.model_config = self.load_model(model_dir)
self.label_map = self.model_config["label_map"]
self.max_seq_length = self.model_config["max_seq_length"]
self.label_map = {int(k):v for k,v in self.label_map.items()}
self.device = "cuda" if torch.cuda.is_available() else "cpu"
self.model = self.model.to(self.device)
self.model.eval()
def load_model(self, model_dir: str, model_config: str = "model_config.json"):
model_config = os.path.join(model_dir,model_config)
model_config = json.load(open(model_config))
model = BertNer.from_pretrained(model_dir)
tokenizer = BertTokenizer.from_pretrained(model_dir, do_lower_case=model_config["do_lower"])
return model, tokenizer, model_config
def tokenize(self, text: str):
""" tokenize input"""
words = word_tokenize(text)
tokens = []
valid_positions = []
for i,word in enumerate(words):
token = self.tokenizer.tokenize(word)
tokens.extend(token)
for i in range(len(token)):
if i == 0:
valid_positions.append(1)
else:
valid_positions.append(0)
return tokens, valid_positions
def preprocess(self, text: str):
""" preprocess """
tokens, valid_positions = self.tokenize(text)
## insert "[CLS]"
tokens.insert(0,"[CLS]")
valid_positions.insert(0,1)
## insert "[SEP]"
tokens.append("[SEP]")
valid_positions.append(1)
segment_ids = []
for i in range(len(tokens)):
segment_ids.append(0)
input_ids = self.tokenizer.convert_tokens_to_ids(tokens)
input_mask = [1] * len(input_ids)
while len(input_ids) < self.max_seq_length:
input_ids.append(0)
input_mask.append(0)
segment_ids.append(0)
valid_positions.append(0)
return input_ids,input_mask,segment_ids,valid_positions
def predict(self, text: str):
input_ids,input_mask,segment_ids,valid_ids = self.preprocess(text)
input_ids = torch.tensor([input_ids],dtype=torch.long,device=self.device)
input_mask = torch.tensor([input_mask],dtype=torch.long,device=self.device)
segment_ids = torch.tensor([segment_ids],dtype=torch.long,device=self.device)
valid_ids = torch.tensor([valid_ids],dtype=torch.long,device=self.device)
with torch.no_grad():
logits = self.model(input_ids, segment_ids, input_mask,valid_ids)
logits = F.softmax(logits,dim=2)
logits_label = torch.argmax(logits,dim=2)
logits_label = logits_label.detach().cpu().numpy().tolist()[0]
logits_confidence = [values[label].item() for values,label in zip(logits[0],logits_label)]
logits = []
pos = 0
for index,mask in enumerate(valid_ids[0]):
if index == 0:
continue
if mask == 1:
logits.append((logits_label[index-pos],logits_confidence[index-pos]))
else:
pos += 1
logits.pop()
labels = [(self.label_map[label],confidence) for label,confidence in logits]
words = word_tokenize(text)
assert len(labels) == len(words)
Person = []
Location = []
Organization = []
Miscelleneous = []
for word, (label, confidence) in zip(words, labels):
if label=="PER" :
Person.append(word)
elif label=="LOC" :
Location.append(word)
elif label=="ORG" :
Organization.append(word)
elif label=="MISC" :
Miscelleneous.append(word)
else:
output = None
output = []
for word, (label, confidence) in zip(words, labels):
if label == "PER":
output.append(' '.join(Person) + ": Person")
if label=="LOC":
output.append(' '.join(Location) + ": Location")
if label=="MISC":
output.append(' '.join(Miscelleneous) + ": Miscelleneous Entity")
if label=="ORG":
output.append(' '.join(Organization) + ": Organization")
return output
| [
"noreply@github.com"
] | noreply@github.com |
16d37fe91e6e6174ecc5ebf06d10063687980ee8 | 97e54e4b18c1d696926678f1e320b2fc9cef5436 | /jaraco/text/strip-prefix.py | 761717a9b9e1f837eeacf0e888822f6fad881361 | [
"MIT"
] | permissive | jaraco/jaraco.text | 8ff2d7d49b3af0ca5e98c1cb337562bde9d3ba72 | 460dc329b799b88adb32ea95435d3a9e03cbdc00 | refs/heads/main | 2023-09-04T06:57:23.624303 | 2023-07-30T01:01:42 | 2023-07-30T01:01:42 | 48,551,451 | 15 | 8 | MIT | 2023-07-30T14:52:20 | 2015-12-24T17:20:06 | Python | UTF-8 | Python | false | false | 412 | py | import sys
import autocommand
from jaraco.text import Stripper
def strip_prefix():
r"""
Strip any common prefix from stdin.
>>> import io, pytest
>>> getfixture('monkeypatch').setattr('sys.stdin', io.StringIO('abcdef\nabc123'))
>>> strip_prefix()
def
123
"""
sys.stdout.writelines(Stripper.strip_prefix(sys.stdin).lines)
autocommand.autocommand(__name__)(strip_prefix)
| [
"jaraco@jaraco.com"
] | jaraco@jaraco.com |
dc72573a696b1184ae2cf899bda0ecd956d49f9d | 0931b32140ba932b3ba02f5109a087c6c70a244d | /frappe/desk/desk_page.py | fc7281e06c18d9766c2efcb8f939fa6938c5c494 | [
"MIT"
] | permissive | cstkyrilos/frappe | b60ed4e95ce929c74c2fc46000080d10b343190e | 27d9306bc5924c11c2749503454cc6d11a8cc654 | refs/heads/main | 2023-03-23T10:35:42.732385 | 2021-03-22T21:55:58 | 2021-03-22T21:55:58 | 350,292,784 | 0 | 0 | MIT | 2021-03-22T10:01:08 | 2021-03-22T10:01:07 | null | UTF-8 | Python | false | false | 1,569 | py | # Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import frappe
from frappe.translate import send_translations
@frappe.whitelist()
def get(name):
"""
Return the :term:`doclist` of the `Page` specified by `name`
"""
page = frappe.get_doc('Page', name)
if page.is_permitted():
page.load_assets()
docs = frappe._dict(page.as_dict())
if getattr(page, '_dynamic_page', None):
docs['_dynamic_page'] = 1
return docs
else:
frappe.response['403'] = 1
raise frappe.PermissionError, 'No read permission for Page %s' % \
(page.title or name)
@frappe.whitelist(allow_guest=True)
def getpage():
"""
Load the page from `frappe.form` and send it via `frappe.response`
"""
page = frappe.form_dict.get('name')
doc = get(page)
# load translations
if frappe.lang != "en":
send_translations(frappe.get_lang_dict("page", page))
frappe.response.docs.append(doc)
def has_permission(page):
if frappe.session.user == "Administrator" or "System Manager" in frappe.get_roles():
return True
page_roles = [d.role for d in page.get("roles")]
if page_roles:
if frappe.session.user == "Guest" and "Guest" not in page_roles:
return False
elif not set(page_roles).intersection(set(frappe.get_roles())):
# check if roles match
return False
if not frappe.has_permission("Page", ptype="read", doc=page):
# check if there are any user_permissions
return False
else:
# hack for home pages! if no Has Roles, allow everyone to see!
return True
| [
"cst.kyrilos@gmail.com"
] | cst.kyrilos@gmail.com |
91ea1c1fcfcc6577bf717c2abd059bc968643776 | 5e84763c16bd6e6ef06cf7a129bb4bd29dd61ec5 | /blimgui/dist/pyglet/libs/darwin/cocoapy/runtime.py | b692ce130d04d7d7af0a3e1daa11e437a71c142c | [
"MIT"
] | permissive | juso40/bl2sdk_Mods | 8422a37ca9c2c2bbf231a2399cbcb84379b7e848 | 29f79c41cfb49ea5b1dd1bec559795727e868558 | refs/heads/master | 2023-08-15T02:28:38.142874 | 2023-07-22T21:48:01 | 2023-07-22T21:48:01 | 188,486,371 | 42 | 110 | MIT | 2022-11-20T09:47:56 | 2019-05-24T20:55:10 | Python | UTF-8 | Python | false | false | 51,751 | py | # objective-ctypes
#
# Copyright (c) 2011, Phillip Nguyen
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions
# are met:
#
# Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# Neither the name of objective-ctypes nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
# FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
# COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
# BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
# ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
# POSSIBILITY OF SUCH DAMAGE.
import sys
import platform
import struct
from ctypes import *
from ctypes import util
from .cocoatypes import *
__LP64__ = (8*struct.calcsize("P") == 64)
__i386__ = (platform.machine() == 'i386')
__arm64__ = (platform.machine() == 'arm64')
if sizeof(c_void_p) == 4:
c_ptrdiff_t = c_int32
elif sizeof(c_void_p) == 8:
c_ptrdiff_t = c_int64
######################################################################
lib = util.find_library('objc')
# Hack for compatibility with macOS > 11.0
if lib is None:
lib = '/usr/lib/libobjc.dylib'
objc = cdll.LoadLibrary(lib)
######################################################################
# BOOL class_addIvar(Class cls, const char *name, size_t size, uint8_t alignment, const char *types)
objc.class_addIvar.restype = c_bool
objc.class_addIvar.argtypes = [c_void_p, c_char_p, c_size_t, c_uint8, c_char_p]
# BOOL class_addMethod(Class cls, SEL name, IMP imp, const char *types)
objc.class_addMethod.restype = c_bool
# BOOL class_addProtocol(Class cls, Protocol *protocol)
objc.class_addProtocol.restype = c_bool
objc.class_addProtocol.argtypes = [c_void_p, c_void_p]
# BOOL class_conformsToProtocol(Class cls, Protocol *protocol)
objc.class_conformsToProtocol.restype = c_bool
objc.class_conformsToProtocol.argtypes = [c_void_p, c_void_p]
# Ivar * class_copyIvarList(Class cls, unsigned int *outCount)
# Returns an array of pointers of type Ivar describing instance variables.
# The array has *outCount pointers followed by a NULL terminator.
# You must free() the returned array.
objc.class_copyIvarList.restype = POINTER(c_void_p)
objc.class_copyIvarList.argtypes = [c_void_p, POINTER(c_uint)]
# Method * class_copyMethodList(Class cls, unsigned int *outCount)
# Returns an array of pointers of type Method describing instance methods.
# The array has *outCount pointers followed by a NULL terminator.
# You must free() the returned array.
objc.class_copyMethodList.restype = POINTER(c_void_p)
objc.class_copyMethodList.argtypes = [c_void_p, POINTER(c_uint)]
# objc_property_t * class_copyPropertyList(Class cls, unsigned int *outCount)
# Returns an array of pointers of type objc_property_t describing properties.
# The array has *outCount pointers followed by a NULL terminator.
# You must free() the returned array.
objc.class_copyPropertyList.restype = POINTER(c_void_p)
objc.class_copyPropertyList.argtypes = [c_void_p, POINTER(c_uint)]
# Protocol ** class_copyProtocolList(Class cls, unsigned int *outCount)
# Returns an array of pointers of type Protocol* describing protocols.
# The array has *outCount pointers followed by a NULL terminator.
# You must free() the returned array.
objc.class_copyProtocolList.restype = POINTER(c_void_p)
objc.class_copyProtocolList.argtypes = [c_void_p, POINTER(c_uint)]
# id class_createInstance(Class cls, size_t extraBytes)
objc.class_createInstance.restype = c_void_p
objc.class_createInstance.argtypes = [c_void_p, c_size_t]
# Method class_getClassMethod(Class aClass, SEL aSelector)
# Will also search superclass for implementations.
objc.class_getClassMethod.restype = c_void_p
objc.class_getClassMethod.argtypes = [c_void_p, c_void_p]
# Ivar class_getClassVariable(Class cls, const char* name)
objc.class_getClassVariable.restype = c_void_p
objc.class_getClassVariable.argtypes = [c_void_p, c_char_p]
# Method class_getInstanceMethod(Class aClass, SEL aSelector)
# Will also search superclass for implementations.
objc.class_getInstanceMethod.restype = c_void_p
objc.class_getInstanceMethod.argtypes = [c_void_p, c_void_p]
# size_t class_getInstanceSize(Class cls)
objc.class_getInstanceSize.restype = c_size_t
objc.class_getInstanceSize.argtypes = [c_void_p]
# Ivar class_getInstanceVariable(Class cls, const char* name)
objc.class_getInstanceVariable.restype = c_void_p
objc.class_getInstanceVariable.argtypes = [c_void_p, c_char_p]
# const char *class_getIvarLayout(Class cls)
objc.class_getIvarLayout.restype = c_char_p
objc.class_getIvarLayout.argtypes = [c_void_p]
# IMP class_getMethodImplementation(Class cls, SEL name)
objc.class_getMethodImplementation.restype = c_void_p
objc.class_getMethodImplementation.argtypes = [c_void_p, c_void_p]
# The function is marked as OBJC_ARM64_UNAVAILABLE.
if not __arm64__:
# IMP class_getMethodImplementation_stret(Class cls, SEL name)
objc.class_getMethodImplementation_stret.restype = c_void_p
objc.class_getMethodImplementation_stret.argtypes = [c_void_p, c_void_p]
# const char * class_getName(Class cls)
objc.class_getName.restype = c_char_p
objc.class_getName.argtypes = [c_void_p]
# objc_property_t class_getProperty(Class cls, const char *name)
objc.class_getProperty.restype = c_void_p
objc.class_getProperty.argtypes = [c_void_p, c_char_p]
# Class class_getSuperclass(Class cls)
objc.class_getSuperclass.restype = c_void_p
objc.class_getSuperclass.argtypes = [c_void_p]
# int class_getVersion(Class theClass)
objc.class_getVersion.restype = c_int
objc.class_getVersion.argtypes = [c_void_p]
# const char *class_getWeakIvarLayout(Class cls)
objc.class_getWeakIvarLayout.restype = c_char_p
objc.class_getWeakIvarLayout.argtypes = [c_void_p]
# BOOL class_isMetaClass(Class cls)
objc.class_isMetaClass.restype = c_bool
objc.class_isMetaClass.argtypes = [c_void_p]
# IMP class_replaceMethod(Class cls, SEL name, IMP imp, const char *types)
objc.class_replaceMethod.restype = c_void_p
objc.class_replaceMethod.argtypes = [c_void_p, c_void_p, c_void_p, c_char_p]
# BOOL class_respondsToSelector(Class cls, SEL sel)
objc.class_respondsToSelector.restype = c_bool
objc.class_respondsToSelector.argtypes = [c_void_p, c_void_p]
# void class_setIvarLayout(Class cls, const char *layout)
objc.class_setIvarLayout.restype = None
objc.class_setIvarLayout.argtypes = [c_void_p, c_char_p]
# Class class_setSuperclass(Class cls, Class newSuper)
objc.class_setSuperclass.restype = c_void_p
objc.class_setSuperclass.argtypes = [c_void_p, c_void_p]
# void class_setVersion(Class theClass, int version)
objc.class_setVersion.restype = None
objc.class_setVersion.argtypes = [c_void_p, c_int]
# void class_setWeakIvarLayout(Class cls, const char *layout)
objc.class_setWeakIvarLayout.restype = None
objc.class_setWeakIvarLayout.argtypes = [c_void_p, c_char_p]
######################################################################
# const char * ivar_getName(Ivar ivar)
objc.ivar_getName.restype = c_char_p
objc.ivar_getName.argtypes = [c_void_p]
# ptrdiff_t ivar_getOffset(Ivar ivar)
objc.ivar_getOffset.restype = c_ptrdiff_t
objc.ivar_getOffset.argtypes = [c_void_p]
# const char * ivar_getTypeEncoding(Ivar ivar)
objc.ivar_getTypeEncoding.restype = c_char_p
objc.ivar_getTypeEncoding.argtypes = [c_void_p]
######################################################################
# char * method_copyArgumentType(Method method, unsigned int index)
# You must free() the returned string.
objc.method_copyArgumentType.restype = c_char_p
objc.method_copyArgumentType.argtypes = [c_void_p, c_uint]
# char * method_copyReturnType(Method method)
# You must free() the returned string.
objc.method_copyReturnType.restype = c_char_p
objc.method_copyReturnType.argtypes = [c_void_p]
# void method_exchangeImplementations(Method m1, Method m2)
objc.method_exchangeImplementations.restype = None
objc.method_exchangeImplementations.argtypes = [c_void_p, c_void_p]
# void method_getArgumentType(Method method, unsigned int index, char *dst, size_t dst_len)
# Functionally similar to strncpy(dst, parameter_type, dst_len).
objc.method_getArgumentType.restype = None
objc.method_getArgumentType.argtypes = [c_void_p, c_uint, c_char_p, c_size_t]
# IMP method_getImplementation(Method method)
objc.method_getImplementation.restype = c_void_p
objc.method_getImplementation.argtypes = [c_void_p]
# SEL method_getName(Method method)
objc.method_getName.restype = c_void_p
objc.method_getName.argtypes = [c_void_p]
# unsigned method_getNumberOfArguments(Method method)
objc.method_getNumberOfArguments.restype = c_uint
objc.method_getNumberOfArguments.argtypes = [c_void_p]
# void method_getReturnType(Method method, char *dst, size_t dst_len)
# Functionally similar to strncpy(dst, return_type, dst_len)
objc.method_getReturnType.restype = None
objc.method_getReturnType.argtypes = [c_void_p, c_char_p, c_size_t]
# const char * method_getTypeEncoding(Method method)
objc.method_getTypeEncoding.restype = c_char_p
objc.method_getTypeEncoding.argtypes = [c_void_p]
# IMP method_setImplementation(Method method, IMP imp)
objc.method_setImplementation.restype = c_void_p
objc.method_setImplementation.argtypes = [c_void_p, c_void_p]
######################################################################
# Class objc_allocateClassPair(Class superclass, const char *name, size_t extraBytes)
objc.objc_allocateClassPair.restype = c_void_p
objc.objc_allocateClassPair.argtypes = [c_void_p, c_char_p, c_size_t]
# Protocol **objc_copyProtocolList(unsigned int *outCount)
# Returns an array of *outcount pointers followed by NULL terminator.
# You must free() the array.
objc.objc_copyProtocolList.restype = POINTER(c_void_p)
objc.objc_copyProtocolList.argtypes = [POINTER(c_int)]
# id objc_getAssociatedObject(id object, void *key)
objc.objc_getAssociatedObject.restype = c_void_p
objc.objc_getAssociatedObject.argtypes = [c_void_p, c_void_p]
# id objc_getClass(const char *name)
objc.objc_getClass.restype = c_void_p
objc.objc_getClass.argtypes = [c_char_p]
# int objc_getClassList(Class *buffer, int bufferLen)
# Pass None for buffer to obtain just the total number of classes.
objc.objc_getClassList.restype = c_int
objc.objc_getClassList.argtypes = [c_void_p, c_int]
# id objc_getMetaClass(const char *name)
objc.objc_getMetaClass.restype = c_void_p
objc.objc_getMetaClass.argtypes = [c_char_p]
# Protocol *objc_getProtocol(const char *name)
objc.objc_getProtocol.restype = c_void_p
objc.objc_getProtocol.argtypes = [c_char_p]
# You should set return and argument types depending on context.
# id objc_msgSend(id theReceiver, SEL theSelector, ...)
# id objc_msgSendSuper(struct objc_super *super, SEL op, ...)
# The function is marked as OBJC_ARM64_UNAVAILABLE.
if not __arm64__:
# void objc_msgSendSuper_stret(struct objc_super *super, SEL op, ...)
objc.objc_msgSendSuper_stret.restype = None
# double objc_msgSend_fpret(id self, SEL op, ...)
# objc.objc_msgSend_fpret.restype = c_double
# The function is marked as OBJC_ARM64_UNAVAILABLE.
if not __arm64__:
# void objc_msgSend_stret(void * stretAddr, id theReceiver, SEL theSelector, ...)
objc.objc_msgSend_stret.restype = None
# void objc_registerClassPair(Class cls)
objc.objc_registerClassPair.restype = None
objc.objc_registerClassPair.argtypes = [c_void_p]
# void objc_removeAssociatedObjects(id object)
objc.objc_removeAssociatedObjects.restype = None
objc.objc_removeAssociatedObjects.argtypes = [c_void_p]
# void objc_setAssociatedObject(id object, void *key, id value, objc_AssociationPolicy policy)
objc.objc_setAssociatedObject.restype = None
objc.objc_setAssociatedObject.argtypes = [c_void_p, c_void_p, c_void_p, c_int]
######################################################################
# id object_copy(id obj, size_t size)
objc.object_copy.restype = c_void_p
objc.object_copy.argtypes = [c_void_p, c_size_t]
# id object_dispose(id obj)
objc.object_dispose.restype = c_void_p
objc.object_dispose.argtypes = [c_void_p]
# Class object_getClass(id object)
objc.object_getClass.restype = c_void_p
objc.object_getClass.argtypes = [c_void_p]
# const char *object_getClassName(id obj)
objc.object_getClassName.restype = c_char_p
objc.object_getClassName.argtypes = [c_void_p]
# Ivar object_getInstanceVariable(id obj, const char *name, void **outValue)
objc.object_getInstanceVariable.restype = c_void_p
objc.object_getInstanceVariable.argtypes=[c_void_p, c_char_p, c_void_p]
# id object_getIvar(id object, Ivar ivar)
objc.object_getIvar.restype = c_void_p
objc.object_getIvar.argtypes = [c_void_p, c_void_p]
# Class object_setClass(id object, Class cls)
objc.object_setClass.restype = c_void_p
objc.object_setClass.argtypes = [c_void_p, c_void_p]
# Ivar object_setInstanceVariable(id obj, const char *name, void *value)
# Set argtypes based on the data type of the instance variable.
objc.object_setInstanceVariable.restype = c_void_p
# void object_setIvar(id object, Ivar ivar, id value)
objc.object_setIvar.restype = None
objc.object_setIvar.argtypes = [c_void_p, c_void_p, c_void_p]
######################################################################
# const char *property_getAttributes(objc_property_t property)
objc.property_getAttributes.restype = c_char_p
objc.property_getAttributes.argtypes = [c_void_p]
# const char *property_getName(objc_property_t property)
objc.property_getName.restype = c_char_p
objc.property_getName.argtypes = [c_void_p]
######################################################################
# BOOL protocol_conformsToProtocol(Protocol *proto, Protocol *other)
objc.protocol_conformsToProtocol.restype = c_bool
objc.protocol_conformsToProtocol.argtypes = [c_void_p, c_void_p]
class OBJC_METHOD_DESCRIPTION(Structure):
_fields_ = [ ("name", c_void_p), ("types", c_char_p) ]
# struct objc_method_description *protocol_copyMethodDescriptionList(Protocol *p, BOOL isRequiredMethod, BOOL isInstanceMethod, unsigned int *outCount)
# You must free() the returned array.
objc.protocol_copyMethodDescriptionList.restype = POINTER(OBJC_METHOD_DESCRIPTION)
objc.protocol_copyMethodDescriptionList.argtypes = [c_void_p, c_bool, c_bool, POINTER(c_uint)]
# objc_property_t * protocol_copyPropertyList(Protocol *protocol, unsigned int *outCount)
objc.protocol_copyPropertyList.restype = c_void_p
objc.protocol_copyPropertyList.argtypes = [c_void_p, POINTER(c_uint)]
# Protocol **protocol_copyProtocolList(Protocol *proto, unsigned int *outCount)
objc.protocol_copyProtocolList = POINTER(c_void_p)
objc.protocol_copyProtocolList.argtypes = [c_void_p, POINTER(c_uint)]
# struct objc_method_description protocol_getMethodDescription(Protocol *p, SEL aSel, BOOL isRequiredMethod, BOOL isInstanceMethod)
objc.protocol_getMethodDescription.restype = OBJC_METHOD_DESCRIPTION
objc.protocol_getMethodDescription.argtypes = [c_void_p, c_void_p, c_bool, c_bool]
# const char *protocol_getName(Protocol *p)
objc.protocol_getName.restype = c_char_p
objc.protocol_getName.argtypes = [c_void_p]
######################################################################
# const char* sel_getName(SEL aSelector)
objc.sel_getName.restype = c_char_p
objc.sel_getName.argtypes = [c_void_p]
# SEL sel_getUid(const char *str)
# Use sel_registerName instead.
# BOOL sel_isEqual(SEL lhs, SEL rhs)
objc.sel_isEqual.restype = c_bool
objc.sel_isEqual.argtypes = [c_void_p, c_void_p]
# SEL sel_registerName(const char *str)
objc.sel_registerName.restype = c_void_p
objc.sel_registerName.argtypes = [c_char_p]
######################################################################
def ensure_bytes(x):
if isinstance(x, bytes):
return x
return x.encode('ascii')
######################################################################
def get_selector(name):
return c_void_p(objc.sel_registerName(ensure_bytes(name)))
def get_class(name):
return c_void_p(objc.objc_getClass(ensure_bytes(name)))
def get_object_class(obj):
return c_void_p(objc.object_getClass(obj))
def get_metaclass(name):
return c_void_p(objc.objc_getMetaClass(ensure_bytes(name)))
def get_superclass_of_object(obj):
cls = c_void_p(objc.object_getClass(obj))
return c_void_p(objc.class_getSuperclass(cls))
# http://www.sealiesoftware.com/blog/archive/2008/10/30/objc_explain_objc_msgSend_stret.html
# http://www.x86-64.org/documentation/abi-0.99.pdf (pp.17-23)
# executive summary: on x86-64, who knows?
def x86_should_use_stret(restype):
"""Try to figure out when a return type will be passed on stack."""
if type(restype) != type(Structure):
return False
if not __LP64__ and sizeof(restype) <= 8:
return False
if __LP64__ and sizeof(restype) <= 16: # maybe? I don't know?
return False
return True
# http://www.sealiesoftware.com/blog/archive/2008/11/16/objc_explain_objc_msgSend_fpret.html
def should_use_fpret(restype):
"""Determine if objc_msgSend_fpret is required to return a floating point type."""
if not __i386__:
# Unneeded on non-intel processors
return False
if __LP64__ and restype == c_longdouble:
# Use only for long double on x86_64
return True
if not __LP64__ and restype in (c_float, c_double, c_longdouble):
return True
return False
# By default, assumes that restype is c_void_p
# and that all arguments are wrapped inside c_void_p.
# Use the restype and argtypes keyword arguments to
# change these values. restype should be a ctypes type
# and argtypes should be a list of ctypes types for
# the arguments of the message only.
def send_message(receiver, selName, *args, **kwargs):
if isinstance(receiver, str):
receiver = get_class(receiver)
selector = get_selector(selName)
restype = kwargs.get('restype', c_void_p)
#print 'send_message', receiver, selName, args, kwargs
argtypes = kwargs.get('argtypes', [])
# Choose the correct version of objc_msgSend based on return type.
if should_use_fpret(restype):
objc.objc_msgSend_fpret.restype = restype
objc.objc_msgSend_fpret.argtypes = [c_void_p, c_void_p] + argtypes
result = objc.objc_msgSend_fpret(receiver, selector, *args)
elif x86_should_use_stret(restype):
objc.objc_msgSend_stret.argtypes = [POINTER(restype), c_void_p, c_void_p] + argtypes
result = restype()
objc.objc_msgSend_stret(byref(result), receiver, selector, *args)
else:
objc.objc_msgSend.restype = restype
objc.objc_msgSend.argtypes = [c_void_p, c_void_p] + argtypes
result = objc.objc_msgSend(receiver, selector, *args)
if restype == c_void_p:
result = c_void_p(result)
return result
class OBJC_SUPER(Structure):
_fields_ = [ ('receiver', c_void_p), ('class', c_void_p) ]
OBJC_SUPER_PTR = POINTER(OBJC_SUPER)
# http://stackoverflow.com/questions/3095360/what-exactly-is-super-in-objective-c
#
# `superclass_name` is optional and can be used to force finding the superclass
# by name. It is used to circumvent a bug in which the superclass was resolved
# incorrectly which lead to an infinite recursion:
# https://github.com/pyglet/pyglet/issues/5
def send_super(receiver, selName, *args, superclass_name=None, **kwargs):
if hasattr(receiver, '_as_parameter_'):
receiver = receiver._as_parameter_
if superclass_name is None:
superclass = get_superclass_of_object(receiver)
else:
superclass = get_class(superclass_name)
super_struct = OBJC_SUPER(receiver, superclass)
selector = get_selector(selName)
restype = kwargs.get('restype', c_void_p)
argtypes = kwargs.get('argtypes', None)
objc.objc_msgSendSuper.restype = restype
if argtypes:
objc.objc_msgSendSuper.argtypes = [OBJC_SUPER_PTR, c_void_p] + argtypes
else:
objc.objc_msgSendSuper.argtypes = None
result = objc.objc_msgSendSuper(byref(super_struct), selector, *args)
if restype == c_void_p:
result = c_void_p(result)
return result
######################################################################
cfunctype_table = {}
def parse_type_encoding(encoding):
"""Takes a type encoding string and outputs a list of the separated type codes.
Currently does not handle unions or bitfields and strips out any field width
specifiers or type specifiers from the encoding. For Python 3.2+, encoding is
assumed to be a bytes object and not unicode.
Examples:
parse_type_encoding('^v16@0:8') --> ['^v', '@', ':']
parse_type_encoding('{CGSize=dd}40@0:8{CGSize=dd}16Q32') --> ['{CGSize=dd}', '@', ':', '{CGSize=dd}', 'Q']
"""
type_encodings = []
brace_count = 0 # number of unclosed curly braces
bracket_count = 0 # number of unclosed square brackets
typecode = b''
for c in encoding:
# In Python 3, c comes out as an integer in the range 0-255. In Python 2, c is a single character string.
# To fix the disparity, we convert c to a bytes object if necessary.
if isinstance(c, int):
c = bytes([c])
if c == b'{':
# Check if this marked the end of previous type code.
if typecode and typecode[-1:] != b'^' and brace_count == 0 and bracket_count == 0:
type_encodings.append(typecode)
typecode = b''
typecode += c
brace_count += 1
elif c == b'}':
typecode += c
brace_count -= 1
assert(brace_count >= 0)
elif c == b'[':
# Check if this marked the end of previous type code.
if typecode and typecode[-1:] != b'^' and brace_count == 0 and bracket_count == 0:
type_encodings.append(typecode)
typecode = b''
typecode += c
bracket_count += 1
elif c == b']':
typecode += c
bracket_count -= 1
assert(bracket_count >= 0)
elif brace_count or bracket_count:
# Anything encountered while inside braces or brackets gets stuck on.
typecode += c
elif c in b'0123456789':
# Ignore field width specifiers for now.
pass
elif c in b'rnNoORV':
# Also ignore type specifiers.
pass
elif c in b'^cislqCISLQfdBv*@#:b?':
if typecode and typecode[-1:] == b'^':
# Previous char was pointer specifier, so keep going.
typecode += c
else:
# Add previous type code to the list.
if typecode:
type_encodings.append(typecode)
# Start a new type code.
typecode = c
# Add the last type code to the list
if typecode:
type_encodings.append(typecode)
return type_encodings
# Limited to basic types and pointers to basic types.
# Does not try to handle arrays, arbitrary structs, unions, or bitfields.
# Assume that encoding is a bytes object and not unicode.
def cfunctype_for_encoding(encoding):
# Check if we've already created a CFUNCTYPE for this encoding.
# If so, then return the cached CFUNCTYPE.
if encoding in cfunctype_table:
return cfunctype_table[encoding]
# Otherwise, create a new CFUNCTYPE for the encoding.
typecodes = {b'c':c_char, b'i':c_int, b's':c_short, b'l':c_long, b'q':c_longlong,
b'C':c_ubyte, b'I':c_uint, b'S':c_ushort, b'L':c_ulong, b'Q':c_ulonglong,
b'f':c_float, b'd':c_double, b'B':c_bool, b'v':None, b'*':c_char_p,
b'@':c_void_p, b'#':c_void_p, b':':c_void_p, NSPointEncoding:NSPoint,
NSSizeEncoding:NSSize, NSRectEncoding:NSRect, NSRangeEncoding:NSRange,
PyObjectEncoding:py_object}
argtypes = []
for code in parse_type_encoding(encoding):
if code in typecodes:
argtypes.append(typecodes[code])
elif code[0:1] == b'^' and code[1:] in typecodes:
argtypes.append(POINTER(typecodes[code[1:]]))
else:
raise Exception('unknown type encoding: ' + code)
cfunctype = CFUNCTYPE(*argtypes)
# Cache the new CFUNCTYPE in the cfunctype_table.
# We do this mainly because it prevents the CFUNCTYPE
# from being garbage-collected while we need it.
cfunctype_table[encoding] = cfunctype
return cfunctype
######################################################################
# After calling create_subclass, you must first register
# it with register_subclass before you may use it.
# You can add new methods after the class is registered,
# but you cannot add any new ivars.
def create_subclass(superclass, name):
if isinstance(superclass, str):
superclass = get_class(superclass)
return c_void_p(objc.objc_allocateClassPair(superclass, ensure_bytes(name), 0))
def register_subclass(subclass):
objc.objc_registerClassPair(subclass)
# types is a string encoding the argument types of the method.
# The first type code of types is the return type (e.g. 'v' if void)
# The second type code must be '@' for id self.
# The third type code must be ':' for SEL cmd.
# Additional type codes are for types of other arguments if any.
def add_method(cls, selName, method, types):
type_encodings = parse_type_encoding(types)
assert(type_encodings[1] == b'@') # ensure id self typecode
assert(type_encodings[2] == b':') # ensure SEL cmd typecode
selector = get_selector(selName)
cfunctype = cfunctype_for_encoding(types)
imp = cfunctype(method)
objc.class_addMethod.argtypes = [c_void_p, c_void_p, cfunctype, c_char_p]
objc.class_addMethod(cls, selector, imp, types)
return imp
def add_ivar(cls, name, vartype):
return objc.class_addIvar(cls, ensure_bytes(name), sizeof(vartype), alignment(vartype), encoding_for_ctype(vartype))
def set_instance_variable(obj, varname, value, vartype):
objc.object_setInstanceVariable.argtypes = [c_void_p, c_char_p, vartype]
objc.object_setInstanceVariable(obj, ensure_bytes(varname), value)
def get_instance_variable(obj, varname, vartype):
variable = vartype()
objc.object_getInstanceVariable(obj, ensure_bytes(varname), byref(variable))
return variable.value
######################################################################
class ObjCMethod:
"""This represents an unbound Objective-C method (really an IMP)."""
# Note, need to map 'c' to c_byte rather than c_char, because otherwise
# ctypes converts the value into a one-character string which is generally
# not what we want at all, especially when the 'c' represents a bool var.
typecodes = {b'c':c_byte, b'i':c_int, b's':c_short, b'l':c_long, b'q':c_longlong,
b'C':c_ubyte, b'I':c_uint, b'S':c_ushort, b'L':c_ulong, b'Q':c_ulonglong,
b'f':c_float, b'd':c_double, b'B':c_bool, b'v':None, b'Vv':None, b'*':c_char_p,
b'@':c_void_p, b'#':c_void_p, b':':c_void_p, b'^v':c_void_p, b'?':c_void_p,
NSPointEncoding:NSPoint, NSSizeEncoding:NSSize, NSRectEncoding:NSRect,
NSRangeEncoding:NSRange,
PyObjectEncoding:py_object}
cfunctype_table = {}
def __init__(self, method):
"""Initialize with an Objective-C Method pointer. We then determine
the return type and argument type information of the method."""
self.selector = c_void_p(objc.method_getName(method))
self.name = objc.sel_getName(self.selector)
self.pyname = self.name.replace(b':', b'_')
self.encoding = objc.method_getTypeEncoding(method)
self.return_type = objc.method_copyReturnType(method)
self.nargs = objc.method_getNumberOfArguments(method)
self.imp = c_void_p(objc.method_getImplementation(method))
self.argument_types = []
for i in range(self.nargs):
buffer = c_buffer(512)
objc.method_getArgumentType(method, i, buffer, len(buffer))
self.argument_types.append(buffer.value)
# Get types for all the arguments.
try:
self.argtypes = [self.ctype_for_encoding(t) for t in self.argument_types]
except:
#print 'no argtypes encoding for %s (%s)' % (self.name, self.argument_types)
self.argtypes = None
# Get types for the return type.
try:
if self.return_type == b'@':
self.restype = ObjCInstance
elif self.return_type == b'#':
self.restype = ObjCClass
else:
self.restype = self.ctype_for_encoding(self.return_type)
except:
#print 'no restype encoding for %s (%s)' % (self.name, self.return_type)
self.restype = None
self.func = None
def ctype_for_encoding(self, encoding):
"""Return ctypes type for an encoded Objective-C type."""
if encoding in self.typecodes:
return self.typecodes[encoding]
elif encoding[0:1] == b'^' and encoding[1:] in self.typecodes:
return POINTER(self.typecodes[encoding[1:]])
elif encoding[0:1] == b'^' and encoding[1:] in [CGImageEncoding, NSZoneEncoding]:
# special cases
return c_void_p
elif encoding[0:1] == b'r' and encoding[1:] in self.typecodes:
# const decorator, don't care
return self.typecodes[encoding[1:]]
elif encoding[0:2] == b'r^' and encoding[2:] in self.typecodes:
# const pointer, also don't care
return POINTER(self.typecodes[encoding[2:]])
else:
raise Exception('unknown encoding for %s: %s' % (self.name, encoding))
def get_prototype(self):
"""Returns a ctypes CFUNCTYPE for the method."""
if self.restype == ObjCInstance or self.restype == ObjCClass:
# Some hacky stuff to get around ctypes issues on 64-bit. Can't let
# ctypes convert the return value itself, because it truncates the pointer
# along the way. So instead, we must do set the return type to c_void_p to
# ensure we get 64-bit addresses and then convert the return value manually.
self.prototype = CFUNCTYPE(c_void_p, *self.argtypes)
else:
self.prototype = CFUNCTYPE(self.restype, *self.argtypes)
return self.prototype
def __repr__(self):
return "<ObjCMethod: %s %s>" % (self.name, self.encoding)
def get_callable(self):
"""Returns a python-callable version of the method's IMP."""
if not self.func:
prototype = self.get_prototype()
self.func = cast(self.imp, prototype)
if self.restype == ObjCInstance or self.restype == ObjCClass:
self.func.restype = c_void_p
else:
self.func.restype = self.restype
self.func.argtypes = self.argtypes
return self.func
def __call__(self, objc_id, *args):
"""Call the method with the given id and arguments. You do not need
to pass in the selector as an argument since it will be automatically
provided."""
f = self.get_callable()
try:
result = f(objc_id, self.selector, *args)
# Convert result to python type if it is a instance or class pointer.
if self.restype == ObjCInstance:
result = ObjCInstance(result)
elif self.restype == ObjCClass:
result = ObjCClass(result)
return result
except ArgumentError as error:
# Add more useful info to argument error exceptions, then reraise.
error.args += ('selector = ' + str(self.name),
'argtypes =' + str(self.argtypes),
'encoding = ' + str(self.encoding))
raise
######################################################################
class ObjCBoundMethod:
"""This represents an Objective-C method (an IMP) which has been bound
to some id which will be passed as the first parameter to the method."""
def __init__(self, method, objc_id):
"""Initialize with a method and ObjCInstance or ObjCClass object."""
self.method = method
self.objc_id = objc_id
def __repr__(self):
return '<ObjCBoundMethod %s (%s)>' % (self.method.name, self.objc_id)
def __call__(self, *args):
"""Call the method with the given arguments."""
return self.method(self.objc_id, *args)
######################################################################
class ObjCClass:
"""Python wrapper for an Objective-C class."""
# We only create one Python object for each Objective-C class.
# Any future calls with the same class will return the previously
# created Python object. Note that these aren't weak references.
# After you create an ObjCClass, it will exist until the end of the
# program.
_registered_classes = {}
def __new__(cls, class_name_or_ptr):
"""Create a new ObjCClass instance or return a previously created
instance for the given Objective-C class. The argument may be either
the name of the class to retrieve, or a pointer to the class."""
# Determine name and ptr values from passed in argument.
if isinstance(class_name_or_ptr, str):
name = class_name_or_ptr
ptr = get_class(name)
else:
ptr = class_name_or_ptr
# Make sure that ptr value is wrapped in c_void_p object
# for safety when passing as ctypes argument.
if not isinstance(ptr, c_void_p):
ptr = c_void_p(ptr)
name = objc.class_getName(ptr)
# Check if we've already created a Python object for this class
# and if so, return it rather than making a new one.
if name in cls._registered_classes:
return cls._registered_classes[name]
# Otherwise create a new Python object and then initialize it.
objc_class = super(ObjCClass, cls).__new__(cls)
objc_class.ptr = ptr
objc_class.name = name
objc_class.instance_methods = {} # mapping of name -> instance method
objc_class.class_methods = {} # mapping of name -> class method
objc_class._as_parameter_ = ptr # for ctypes argument passing
# Store the new class in dictionary of registered classes.
cls._registered_classes[name] = objc_class
# Not sure this is necessary...
objc_class.cache_instance_methods()
objc_class.cache_class_methods()
return objc_class
def __repr__(self):
return "<ObjCClass: %s at %s>" % (self.name, str(self.ptr.value))
def cache_instance_methods(self):
"""Create and store python representations of all instance methods
implemented by this class (but does not find methods of superclass)."""
count = c_uint()
method_array = objc.class_copyMethodList(self.ptr, byref(count))
for i in range(count.value):
method = c_void_p(method_array[i])
objc_method = ObjCMethod(method)
self.instance_methods[objc_method.pyname] = objc_method
def cache_class_methods(self):
"""Create and store python representations of all class methods
implemented by this class (but does not find methods of superclass)."""
count = c_uint()
method_array = objc.class_copyMethodList(objc.object_getClass(self.ptr), byref(count))
for i in range(count.value):
method = c_void_p(method_array[i])
objc_method = ObjCMethod(method)
self.class_methods[objc_method.pyname] = objc_method
def get_instance_method(self, name):
"""Returns a python representation of the named instance method,
either by looking it up in the cached list of methods or by searching
for and creating a new method object."""
if name in self.instance_methods:
return self.instance_methods[name]
else:
# If method name isn't in the cached list, it might be a method of
# the superclass, so call class_getInstanceMethod to check.
selector = get_selector(name.replace(b'_', b':'))
method = c_void_p(objc.class_getInstanceMethod(self.ptr, selector))
if method.value:
objc_method = ObjCMethod(method)
self.instance_methods[name] = objc_method
return objc_method
return None
def get_class_method(self, name):
"""Returns a python representation of the named class method,
either by looking it up in the cached list of methods or by searching
for and creating a new method object."""
if name in self.class_methods:
return self.class_methods[name]
else:
# If method name isn't in the cached list, it might be a method of
# the superclass, so call class_getInstanceMethod to check.
selector = get_selector(name.replace(b'_', b':'))
method = c_void_p(objc.class_getClassMethod(self.ptr, selector))
if method.value:
objc_method = ObjCMethod(method)
self.class_methods[name] = objc_method
return objc_method
return None
def __getattr__(self, name):
"""Returns a callable method object with the given name."""
# If name refers to a class method, then return a callable object
# for the class method with self.ptr as hidden first parameter.
name = ensure_bytes(name)
method = self.get_class_method(name)
if method:
return ObjCBoundMethod(method, self.ptr)
# If name refers to an instance method, then simply return the method.
# The caller will need to supply an instance as the first parameter.
method = self.get_instance_method(name)
if method:
return method
# Otherwise, raise an exception.
raise AttributeError('ObjCClass %s has no attribute %s' % (self.name, name))
######################################################################
class ObjCInstance:
"""Python wrapper for an Objective-C instance."""
_cached_objects = {}
def __new__(cls, object_ptr):
"""Create a new ObjCInstance or return a previously created one
for the given object_ptr which should be an Objective-C id."""
# Make sure that object_ptr is wrapped in a c_void_p.
if not isinstance(object_ptr, c_void_p):
object_ptr = c_void_p(object_ptr)
# If given a nil pointer, return None.
if not object_ptr.value:
return None
# Check if we've already created an python ObjCInstance for this
# object_ptr id and if so, then return it. A single ObjCInstance will
# be created for any object pointer when it is first encountered.
# This same ObjCInstance will then persist until the object is
# deallocated.
if object_ptr.value in cls._cached_objects:
return cls._cached_objects[object_ptr.value]
# Otherwise, create a new ObjCInstance.
objc_instance = super(ObjCInstance, cls).__new__(cls)
objc_instance.ptr = object_ptr
objc_instance._as_parameter_ = object_ptr
# Determine class of this object.
class_ptr = c_void_p(objc.object_getClass(object_ptr))
objc_instance.objc_class = ObjCClass(class_ptr)
# Store new object in the dictionary of cached objects, keyed
# by the (integer) memory address pointed to by the object_ptr.
cls._cached_objects[object_ptr.value] = objc_instance
# Create a DeallocationObserver and associate it with this object.
# When the Objective-C object is deallocated, the observer will remove
# the ObjCInstance corresponding to the object from the cached objects
# dictionary, effectively destroying the ObjCInstance.
observer = send_message(send_message('DeallocationObserver', 'alloc'), 'initWithObject:', objc_instance)
objc.objc_setAssociatedObject(objc_instance, observer, observer, 0x301)
# The observer is retained by the object we associate it to. We release
# the observer now so that it will be deallocated when the associated
# object is deallocated.
send_message(observer, 'release')
return objc_instance
def __repr__(self):
if self.objc_class.name == b'NSCFString':
# Display contents of NSString objects
from .cocoalibs import cfstring_to_string
string = cfstring_to_string(self)
return "<ObjCInstance %#x: %s (%s) at %s>" % (id(self), self.objc_class.name, string, str(self.ptr.value))
return "<ObjCInstance %#x: %s at %s>" % (id(self), self.objc_class.name, str(self.ptr.value))
def __getattr__(self, name):
"""Returns a callable method object with the given name."""
# Search for named instance method in the class object and if it
# exists, return callable object with self as hidden argument.
# Note: you should give self and not self.ptr as a parameter to
# ObjCBoundMethod, so that it will be able to keep the ObjCInstance
# alive for chained calls like MyClass.alloc().init() where the
# object created by alloc() is not assigned to a variable.
name = ensure_bytes(name)
method = self.objc_class.get_instance_method(name)
if method:
return ObjCBoundMethod(method, self)
# Else, search for class method with given name in the class object.
# If it exists, return callable object with a pointer to the class
# as a hidden argument.
method = self.objc_class.get_class_method(name)
if method:
return ObjCBoundMethod(method, self.objc_class.ptr)
# Otherwise raise an exception.
raise AttributeError('ObjCInstance %s has no attribute %s' % (self.objc_class.name, name))
######################################################################
def convert_method_arguments(encoding, args):
"""Used by ObjCSubclass to convert Objective-C method arguments to
Python values before passing them on to the Python-defined method."""
new_args = []
arg_encodings = parse_type_encoding(encoding)[3:]
for e, a in zip(arg_encodings, args):
if e == b'@':
new_args.append(ObjCInstance(a))
elif e == b'#':
new_args.append(ObjCClass(a))
else:
new_args.append(a)
return new_args
# ObjCSubclass is used to define an Objective-C subclass of an existing
# class registered with the runtime. When you create an instance of
# ObjCSubclass, it registers the new subclass with the Objective-C
# runtime and creates a set of function decorators that you can use to
# add instance methods or class methods to the subclass.
#
# Typical usage would be to first create and register the subclass:
#
# MySubclass = ObjCSubclass('NSObject', 'MySubclassName')
#
# then add methods with:
#
# @MySubclass.method('v')
# def methodThatReturnsVoid(self):
# pass
#
# @MySubclass.method('Bi')
# def boolReturningMethodWithInt_(self, x):
# return True
#
# @MySubclass.classmethod('@')
# def classMethodThatReturnsId(self):
# return self
#
# It is probably a good idea to organize the code related to a single
# subclass by either putting it in its own module (note that you don't
# actually need to expose any of the method names or the ObjCSubclass)
# or by bundling it all up inside a python class definition, perhaps
# called MySubclassImplementation.
#
# It is also possible to add Objective-C ivars to the subclass, however
# if you do so, you must call the __init__ method with register=False,
# and then call the register method after the ivars have been added.
# But rather than creating the ivars in Objective-C land, it is easier
# to just define python-based instance variables in your subclass's init
# method.
#
# This class is used only to *define* the interface and implementation
# of an Objective-C subclass from python. It should not be used in
# any other way. If you want a python representation of the resulting
# class, create it with ObjCClass.
#
# Instances are created as a pointer to the objc object by using:
#
# myinstance = send_message('MySubclassName', 'alloc')
# myinstance = send_message(myinstance, 'init')
#
# or wrapped inside an ObjCInstance object by using:
#
# myclass = ObjCClass('MySubclassName')
# myinstance = myclass.alloc().init()
#
class ObjCSubclass:
"""Use this to create a subclass of an existing Objective-C class.
It consists primarily of function decorators which you use to add methods
to the subclass."""
def __init__(self, superclass, name, register=True):
self._imp_table = {}
self.name = name
self.objc_cls = create_subclass(superclass, name)
self._as_parameter_ = self.objc_cls
if register:
self.register()
def register(self):
"""Register the new class with the Objective-C runtime."""
objc.objc_registerClassPair(self.objc_cls)
# We can get the metaclass only after the class is registered.
self.objc_metaclass = get_metaclass(self.name)
def add_ivar(self, varname, vartype):
"""Add instance variable named varname to the subclass.
varname should be a string.
vartype is a ctypes type.
The class must be registered AFTER adding instance variables."""
return add_ivar(self.objc_cls, varname, vartype)
def add_method(self, method, name, encoding):
imp = add_method(self.objc_cls, name, method, encoding)
self._imp_table[name] = imp
# http://iphonedevelopment.blogspot.com/2008/08/dynamically-adding-class-objects.html
def add_class_method(self, method, name, encoding):
imp = add_method(self.objc_metaclass, name, method, encoding)
self._imp_table[name] = imp
def rawmethod(self, encoding):
"""Decorator for instance methods without any fancy shenanigans.
The function must have the signature f(self, cmd, *args)
where both self and cmd are just pointers to objc objects."""
# Add encodings for hidden self and cmd arguments.
encoding = ensure_bytes(encoding)
typecodes = parse_type_encoding(encoding)
typecodes.insert(1, b'@:')
encoding = b''.join(typecodes)
def decorator(f):
name = f.__name__.replace('_', ':')
self.add_method(f, name, encoding)
return f
return decorator
def method(self, encoding):
"""Function decorator for instance methods."""
# Add encodings for hidden self and cmd arguments.
encoding = ensure_bytes(encoding)
typecodes = parse_type_encoding(encoding)
typecodes.insert(1, b'@:')
encoding = b''.join(typecodes)
def decorator(f):
def objc_method(objc_self, objc_cmd, *args):
py_self = ObjCInstance(objc_self)
py_self.objc_cmd = objc_cmd
args = convert_method_arguments(encoding, args)
result = f(py_self, *args)
if isinstance(result, ObjCClass):
result = result.ptr.value
elif isinstance(result, ObjCInstance):
result = result.ptr.value
return result
name = f.__name__.replace('_', ':')
self.add_method(objc_method, name, encoding)
return objc_method
return decorator
def classmethod(self, encoding):
"""Function decorator for class methods."""
# Add encodings for hidden self and cmd arguments.
encoding = ensure_bytes(encoding)
typecodes = parse_type_encoding(encoding)
typecodes.insert(1, b'@:')
encoding = b''.join(typecodes)
def decorator(f):
def objc_class_method(objc_cls, objc_cmd, *args):
py_cls = ObjCClass(objc_cls)
py_cls.objc_cmd = objc_cmd
args = convert_method_arguments(encoding, args)
result = f(py_cls, *args)
if isinstance(result, ObjCClass):
result = result.ptr.value
elif isinstance(result, ObjCInstance):
result = result.ptr.value
return result
name = f.__name__.replace('_', ':')
self.add_class_method(objc_class_method, name, encoding)
return objc_class_method
return decorator
######################################################################
# Instances of DeallocationObserver are associated with every
# Objective-C object that gets wrapped inside an ObjCInstance.
# Their sole purpose is to watch for when the Objective-C object
# is deallocated, and then remove the object from the dictionary
# of cached ObjCInstance objects kept by the ObjCInstance class.
#
# The methods of the class defined below are decorated with
# rawmethod() instead of method() because DeallocationObservers
# are created inside of ObjCInstance's __new__ method and we have
# to be careful to not create another ObjCInstance here (which
# happens when the usual method decorator turns the self argument
# into an ObjCInstance), or else get trapped in an infinite recursion.
class DeallocationObserver_Implementation:
DeallocationObserver = ObjCSubclass('NSObject', 'DeallocationObserver', register=False)
DeallocationObserver.add_ivar('observed_object', c_void_p)
DeallocationObserver.register()
@DeallocationObserver.rawmethod('@@')
def initWithObject_(self, cmd, anObject):
self = send_super(self, 'init')
self = self.value
set_instance_variable(self, 'observed_object', anObject, c_void_p)
return self
@DeallocationObserver.rawmethod('v')
def dealloc(self, cmd):
anObject = get_instance_variable(self, 'observed_object', c_void_p)
ObjCInstance._cached_objects.pop(anObject, None)
send_super(self, 'dealloc')
@DeallocationObserver.rawmethod('v')
def finalize(self, cmd):
# Called instead of dealloc if using garbage collection.
# (which would have to be explicitly started with
# objc_startCollectorThread(), so probably not too much reason
# to have this here, but I guess it can't hurt.)
anObject = get_instance_variable(self, 'observed_object', c_void_p)
ObjCInstance._cached_objects.pop(anObject, None)
send_super(self, 'finalize')
| [
"justin.sostmann@googlemail.com"
] | justin.sostmann@googlemail.com |
a84a107c0e275202b4cbb1b877c37f3825b42b1d | 36cf4465f576b2a7a2648f71177333c199f84796 | /src/analysis/document_similarities.py | 6b92aa2c2afefd204b3c88ea62f38dba55efd079 | [] | no_license | stevenrouk/evolution-of-machine-learning | d0cd5f6535db876f97445d04a3ad78db27a82f03 | 3be74301ce425377ec7a2e7aeffb35a1764225ff | refs/heads/master | 2023-02-20T08:37:32.932015 | 2022-08-10T22:14:10 | 2022-08-10T22:14:10 | 215,085,244 | 2 | 1 | null | 2023-02-15T22:58:25 | 2019-10-14T15:54:48 | Jupyter Notebook | UTF-8 | Python | false | false | 353 | py | import numpy as np
from sklearn.metrics.pairwise import cosine_similarity
def get_similar_doc_idxs_to_loadings(loadings, W):
sims = cosine_similarity(loadings, W)[0]
return np.argsort(sims)[::-1]
def get_similar_doc_idxs_to_tfidf(tfidf_vec, all_docs):
sims = cosine_similarity(tfidf_vec, all_docs)[0]
return np.argsort(sims)[::-1]
| [
"stevenrouk@gmail.com"
] | stevenrouk@gmail.com |
5fa9d6c34be022eb700f187f8a95d6366b9ee254 | a5ae57b44064e6bb2fa17b5e7a8e551ec278345f | /src/core/runtime/control_structure/block.py | 00143a8d061be0bc1fd9baa0369f77e4f6ebaa9d | [
"MIT"
] | permissive | thomasmf/nomenine | 5cfc06ecfbacd73938b3ab0c52a2ebd3a0349faf | ead48185b150fdc07a5019499511f696c5326d45 | refs/heads/master | 2021-01-10T02:35:06.465381 | 2015-11-11T10:00:57 | 2015-11-11T10:00:57 | 44,127,966 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,245 | py |
TEST( """ with ( function x1 [ . 40 ] ) [ x1 ] == 40 """ )
TEST( """ let x1 30 [ x1 * 10 ] == 300 """ )
TEST( """ use [ [ definition x 387 ] [ definition y 4123 ] [ function f [ x + 1000 ] ] ] [ f * ( y ) ] == 5718601 """ )
TEST( """ use [ [ definition TestFactory ( factory ( Integer ) [ : that + 10000 ] ) ] [ function f ( TestFactory ) [ : that ] ] ] [ f ( TestFactory 100 ) + 1 ] == 10101 """ )
ROOT_SCOPE_METHOD(
MC( ARG( CW( 'with' ), CG( 'ANY', 'scope' ), CG( 'LIST', 'phrase' ) ), """
$NOM( CONTEXT,
$CA(UNION_new( $LISTNEW(
nom_definition( $CA(WORD_new( "scope" )), PARAM_scope ),
nom_definition( $CA(WORD_new( "phrase" )), PARAM_phrase )
) )),
: that phrase evaluate ( union
( : that scope )
( : this )
)
) ;
""" ),
MC( ARG( CW( 'let' ), CG( 'WORD', 'name' ), CG( 'ANY', 'value' ), CG( 'LIST', 'phrase' ) ), """
$NOM( CONTEXT,
$CA(UNION_new( $LISTNEW(
nom_definition( $CA(WORD_new( "name" )), PARAM_name ),
nom_definition( $CA(WORD_new( "value" )), PARAM_value ),
nom_definition( $CA(WORD_new( "phrase" )), PARAM_phrase )
) )),
: this with ( definition ( : that name ) ( : that value ) ) ( : that phrase )
) ;
""" ),
MC( ARG( CW( 'block' ), CG( 'ANY', 'scope' ), CG( 'LIST', 'components' ) ), """
$NOM( CONTEXT,
$CA(UNION_new( $LISTNEW(
nom_definition( $CA(WORD_new( "scope" )), PARAM_scope ),
nom_definition( $CA(WORD_new( "components" )), PARAM_components )
) )),
if value [ : that components value ]
then [
let scope ( union ( : that scope ) ( value evaluate ( : that scope ) ) ) [
: this block ( scope ) ( : that components next )
]
]
else [
: that scope
]
) ;
""" ),
MC( ARG( CW( 'use' ), CG( 'LIST', 'components' ), CG( 'LIST', 'phrase' ) ), """
$NOM( CONTEXT,
$CA(UNION_new( $LISTNEW(
nom_definition( $CA(WORD_new( "components" )), PARAM_components ),
nom_definition( $CA(WORD_new( "phrase" )), PARAM_phrase )
) )),
: this with ( : this block ( : this ) ( : that components ) ) ( : that phrase )
) ;
""" )
)
| [
"thomas@metatools.org"
] | thomas@metatools.org |
e78a07d5a9ac0d6375bab50be733a669fac273ff | e5b6d2e79d6593587fa8f5854def9ebf4d47a9e1 | /djangocli/wsgi.py | 8e9c0ba06187289fb8d23d2abffc8b6bcf5721d6 | [] | no_license | redeyed-archive/DjangoSiteCheckerExample | 35756664f0b9667e151d4608c6ebd5d279523534 | e53b2fad15d2a768e75bc853c69113c0d54c2ed2 | refs/heads/master | 2023-03-17T06:22:46.129989 | 2019-02-17T05:48:43 | 2019-02-17T05:48:43 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 395 | py | """
WSGI config for djangocli project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/2.1/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'djangocli.settings')
application = get_wsgi_application()
| [
"unconfigured@null.spigotmc.org"
] | unconfigured@null.spigotmc.org |
39b0fd52bc50bd9b15dfb1bcdb7536b39cdacdc5 | 0abb03d472094ea9c2b3c672cba53b4d603b19dd | /ansible/plugins/modules/test1.py | de2bace0bbbea56e32b1bc8d2ecdb1939c8aaca0 | [
"Unlicense"
] | permissive | mamercad/sandbox | 29502faa7f7b5bdbb17bcf87fae7ec0a0cefa568 | 2e69094452a7e2ecbf6500bb658c94fbaba2395b | refs/heads/main | 2023-08-21T21:09:03.867299 | 2023-05-22T18:57:27 | 2023-05-22T18:57:27 | 280,487,343 | 0 | 0 | Unlicense | 2023-05-22T18:52:03 | 2020-07-17T17:35:17 | Python | UTF-8 | Python | false | false | 21 | py | test1
test
test
test
| [
"mamercad@gmail.com"
] | mamercad@gmail.com |
3d26d362ff03dbd5bd9189d208e4bcad515f2f2a | e5d04f7f5e67e186dcb78a256341c3ab3afe796d | /twidder/helper.py | 44cc176fb3c2a497f691b2e0e481b8ebcb5bf55b | [] | no_license | bavaria95/twidder | c7ce456d1dd6b4713553608534cd2d19b7f3b168 | 496ed5c2324ab07db36f4f3cfe22ef4dd471d37c | refs/heads/master | 2016-08-11T21:58:01.284283 | 2016-03-08T00:16:08 | 2016-03-08T00:16:08 | 51,746,949 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,325 | py | import os
import binascii
import json
import database_helper
import random
import hmac
import hashlib
import time
from flask_sockets import Sockets
# implementing Diffie-Hellman key exchange algorithm
# generator
g = 3
# divider
p = 17
def log(msg):
f = open('log.txt', 'a')
f.write(str(msg) + '\n')
f.close()
def compute_public_key():
y = random.randrange(50, 100)
public_key = g**y % p
return {'public_key': public_key, 'secret_variable': y}
def compute_secret_key(client_public, y):
return client_public**y % p
def is_legid(message, exp_hash, timestamp):
'''
checks whether received message is correct(expected and actual hash-sum match)
'''
actual_timestamp = int(time.time())
# if message is older than 10 seconds - drop it
if abs(actual_timestamp - timestamp) > 10:
return False
token = message['token']
secret_key = database_helper.storage.get_user_secret(token)
message_string = ''.join([message[x] for x in sorted(message.keys())])
message_string += str(timestamp)
actual_hash = hmac.new(str(secret_key),
message_string,
hashlib.sha1).hexdigest()
return exp_hash == actual_hash
# to store tokens and corresponded emails to it
class Storage():
def __init__(self):
self.d = {}
def add_user(self, token, email, secret):
if token in self.d:
raise 'Token is used.'
self.d[token] = {'email': email, 'secret': secret}
def remove_user(self, token):
self.d.pop(token, None)
def get_user_email(self, token):
res = self.d.get(token)
if res:
return res['email']
return None
def get_user_secret(self, token):
res = self.d.get(token)
if res:
return res['secret']
return None
def is_token_presented(self, token):
return token in self.d
def get_all_storage(self):
return self.d
def get_token_by_email(self, email):
res = []
for k,v in self.d.iteritems():
if v['email'] == email:
res.append(k)
return res
def remove_token_by_email(self, email):
for k,v in list(self.d.iteritems())[:]:
if v['email'] == email:
self.d.pop(k, None)
class SocketPool():
def __init__(self):
self.d = {}
def add_socket(self, email, sock):
self.d[email] = sock
def get_socket(self, email):
return self.d.get(email, None)
def is_socket_presented(self, email):
return email in self.d
def remove_socket(self, email):
return self.d.pop(email, None)
def get_all_sockets(self):
return self.d
def size(self):
return len(self.d)
def change_socket_by_email(self, email, sock):
self.d[email] = sock
class StatsInfo():
def __init__(self):
self.d = {}
def add_entry(self, token, sock):
self.d[token] = {'socket': sock, 'prev': {'all_users': None,
'online': None,
'posts': None,
'all_posts': None}}
def is_entry_presented(self, token):
return token in self.d
def get_entry(self, token):
return self.d.get(token, None)
def remove_entry(self, token):
return self.d.pop(token)
def get_all_entries(self):
return self.d
def notify_by_token(self, token, data):
'''
Also takes care about necessity of notifying
(sends nothing if nothing changed)
'''
if not token in self.d:
return
if self.d[token]['prev'] != data:
try:
self.d[token]['socket'].send(json.dumps(data))
self.d[token]['prev'] = data
except:
pass
def all_subscribers(self):
return self.d.keys()
def generate_random_token():
token_length = 36
return binascii.hexlify(os.urandom(token_length))
def allowed_file(filename):
ALLOWED_EXTENSIONS = set(['png', 'jpg', 'jpeg', 'gif', 'avi', 'mp4', 'mp3'])
return '.' in filename and \
filename.rsplit('.', 1)[1] in ALLOWED_EXTENSIONS
| [
"bavaria95@gmail.com"
] | bavaria95@gmail.com |
12bde3b49a3da09feeaeba0f56b55e4a8f07c0de | 47d5fe99e3cb28d3e0aed9564b774889d3a3d1cc | /ask/ask/urls.py | 2f115fd2b1d7167cb3452f53e1af2107688aaaff | [] | no_license | EgrethName/Web | 5a19999342dcfba445ba1b4b963c2c01e283b2bd | 928dc006f8155e28cc16ef490d47a671053d7c27 | refs/heads/master | 2023-04-14T16:34:34.856734 | 2021-04-28T12:37:56 | 2021-04-28T12:37:56 | 327,976,788 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,103 | py | """ask URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.1/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path
2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
"""
from django.contrib import admin
from django.urls import include, re_path
urlpatterns = [
re_path('qa/', include('qa.urls')),
re_path('admin/', admin.site.urls),
re_path('', include('qa.urls')),
re_path('login/', include('qa.urls')),
re_path('signup/', include('qa.urls')),
re_path('ask/', include('qa.urls')),
re_path(r'question/\d+/', include('qa.urls')),
re_path('popular/', include('qa.urls')),
re_path('new/', include('qa.urls')),
]
| [
"egreth@mail.ru"
] | egreth@mail.ru |
1f009db8fd1167365c68a8e8ac41c0e88857d046 | c219339c6f1818685cccd71a9ca88ecf0d9bfcf2 | /Module 8/demo22ex3/demo22ex3.py | 01743c4aa432d1b030ee9fc40f872bda5bbc2885 | [] | no_license | UmeshDeshmukh/DSP_Lab_ECE-GY_6183 | fc2b01e2a65ae16c243f7f747072ffaec689e179 | 7ad7988fcf7ac947dcb8a90c148938be008d7021 | refs/heads/main | 2023-05-08T17:05:44.054708 | 2021-06-02T04:34:25 | 2021-06-02T04:34:25 | 334,494,205 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,424 | py | # find_blue_in_image.py
# Detect pixels similar to a prescribed color.
# This can be done usg HSV color space.
import cv2
import numpy as np
img = cv2.imread('input_image.jpg', 1)
# 1 : import image in color
# Convert to different color space
img_hsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)
print(type(img_hsv))
print(img_hsv.shape)
print(img_hsv.dtype)
green = np.uint8([[[0, 255, 0]]]) # 3D array
green_hsv = cv2.cvtColor(green, cv2.COLOR_BGR2HSV)
h = green_hsv[0,0,0]
print('Green in HSV color space:', green_hsv)
print('Hue = ', h) # see that h = 120
# lower = np.array([90, 50, 22])
# upper = np.array([180, 255, 90])
lower = np.array([h-50, 50, 50])
upper = np.array([h+50, 255, 255])
print('lower = ', lower)
print('upper = ', upper)
# quit()
# Determine binary mask
green_mask = cv2.inRange(img_hsv, lower, upper)
# Apply mask to color image
output = cv2.bitwise_and(img, img, mask = green_mask)
# Show images:
cv2.imshow('Original image', img)
cv2.imshow('Mask', green_mask)
cv2.imshow('Segmented image', output)
print('Switch to images. Then press any key to stop')
cv2.waitKey(0)
cv2.destroyAllWindows()
# Write the image to a file
# cv2.imwrite('tiger_mask.jpg', green_mask)
# cv2.imwrite('tiger_green.jpg', output)
cv2.imwrite('mask_image.jpg', green_mask)
cv2.imwrite('detected_pixels.jpg', output)
# Reference
# http://docs.opencv.org/3.2.0/df/d9d/tutorial_py_colorspaces.html
| [
"noreply@github.com"
] | noreply@github.com |
36c646e5fe820af4bb83f2371e13d24460aa2f1f | 29bae218ba2662d8c700b48f7f256298de08d9f2 | /env/bin/f2py | 493f3036cc4786365a324ee6baf2959080e2f80a | [] | no_license | bradyb/deepgraminterview | c4986cb9ff090eb89a119b61faab4845d24ca7c0 | f48a4c5a17ca62195656e5839f4e8b60e33d4cdc | refs/heads/main | 2023-07-05T05:47:37.931140 | 2021-08-11T06:00:55 | 2021-08-11T06:00:55 | 394,879,774 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 252 | #!/home/turist/interviews/deepgram/env/bin/python3
# -*- coding: utf-8 -*-
import re
import sys
from numpy.f2py.f2py2e import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"brady.ben12@gmail.com"
] | brady.ben12@gmail.com | |
2b3a01df19d821d47f7af67c938295d1d18bcc2e | 28922bb1e78165a6d6fd02a45eca4acbdf202cab | /0x0F-python-object_relational_mapping/relationship_state.py | 740f17ebad5cdfa3478b85d68f147f312fe9a214 | [] | no_license | dalejoroc11/holbertonschool-higher_level_programming | 03d261a59971421b42fc0def730cc43ad478d2cb | e3195595f3237df4f0f8f519888927f07e86d959 | refs/heads/master | 2022-12-23T13:32:21.378580 | 2020-09-23T00:30:09 | 2020-09-23T00:30:09 | 259,437,432 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 520 | py | #!/usr/bin/python3
# Class definition of a State and an instance relationship
from sqlalchemy import Column, Integer, String
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.orm import relationship
from relationship_city import Base, City
class State(Base):
"""Represents a state for database"""
__tablename__ = "states"
id = Column(Integer, primary_key=True)
name = Column(String(128), nullable=False)
cities = relationship("City", backref="state", cascade="all, delete")
| [
"diegorojas279@hotmail.com"
] | diegorojas279@hotmail.com |
303f9eed263c05da330921097549a5347a38de5e | 7f9f44689af2d52b4bc0384f934d74ce86fe7deb | /src/magic1.py | 7a18fd8bc1f41e7470ac0d05aaf44be6dfa5e52e | [] | no_license | Nuked88/excbot | af94c8f16e145288e7d5b812d2f15bc7d9c0acce | 9b3b70a345c3dba396d3f92bfabebec2fe70c516 | refs/heads/master | 2021-05-11T00:56:25.097351 | 2018-02-11T15:54:54 | 2018-02-11T15:54:54 | 118,316,939 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,283 | py | from __future__ import division
from itertools import count
import matplotlib.pyplot as plt
from numpy import linspace, loadtxt, ones, convolve
import numpy as np
import pandas as pd
import collections
from random import randint
from matplotlib import style
import pymongo
from pymongo import MongoClient
from pprint import pprint
#cdb = MongoClient('173.249.9.155', 27017)
cdb = MongoClient('localhost', 27017)
db = cdb.excbot
data = db.data2
score = db.score
maxRes=40000
sym = "ETHUSDT"
def way(array):
prevValue = 0
wWay=0
for a in array:
if a > prevValue:
wWay=wWay+1
else:
wWay=wWay-1
prevValue=a
return wWay
def getData():
datacount= data.aggregate([
{ "$match": { "sym": sym} },
{ "$project": {
"year": { "$year": '$date'},
"month": { "$month": '$date'},
"day": { "$dayOfMonth": '$date'},
"hour": { "$hour": '$date'},
"minute": { "$minute": '$date'},
"price": 1,
"date":1
}},
{
"$group": {
"_id": { "date":"$date"},
"price":{
"$avg": "$price"
}}
},
{
"$limit":maxRes
},
{ "$sort" : { '_id.date': 1 } }])
score=0
i=0
armonica=0
list1=[]
b=[]
for a in datacount:
if str(a["price"]) != 'None':
i=i+1
tarr=[]
#pprint(str(a["price"])+"-"+str(a["_id"]))
tarr.append(i)
tarr.append(a["price"])
tarr.append(a["_id"]["date"])
b.append(a["price"])
list1.append(tarr)
#list1['SunSpots'].append(a["price"])
armonica=float(armonica)+(1/float(a["price"]))
#PRINT DATE
pprint("Ultima data:")
pprint(list1)
#mi dice se va su o giù
res=(100*way(b))/maxRes
pprint("Guadagno/Perdita")
pprint(str(res)+"%")
#media armonica prezzi
pprint("Numero risultati:")
pprint(i)
armonica=i/float(armonica)
return [list1,armonica]
# 3. Lets define some use-case specific UDF(User Defined Functions)
def moving_average(data, window_size):
""" Computes moving average using discrete linear convolution of two one dimensional sequences.
Args:
-----
data (pandas.Series): independent variable
window_size (int): rolling window size
Returns:
--------
ndarray of linear convolution
References:
------------
[1] Wikipedia, "Convolution", http://en.wikipedia.org/wiki/Convolution.
[2] API Reference: https://docs.scipy.org/doc/numpy/reference/generated/numpy.convolve.html
"""
window = np.ones(int(window_size))/float(window_size)
return np.convolve(data, window, 'same')
def explain_anomalies(y, window_size, sigma=1.0):
""" Helps in exploring the anamolies using stationary standard deviation
Args:
-----
y (pandas.Series): independent variable
window_size (int): rolling window size
sigma (int): value for standard deviation
Returns:
--------
a dict (dict of 'standard_deviation': int, 'anomalies_dict': (index: value))
containing information about the points indentified as anomalies
"""
avg = moving_average(y, window_size).tolist()
residual = y - avg
# Calculate the variation in the distribution of the residual
std = np.std(residual)
return {'standard_deviation': round(std, 8),
'ad': collections.OrderedDict([(index, y_i) for
index, y_i, avg_i in zip(count(), y, avg)
if (y_i > avg_i + (sigma*std)) | (y_i < avg_i - (sigma*std))])}
def explain_anomalies_rolling_std(y, window_size, sigma=1.0):
""" Helps in exploring the anamolies using rolling standard deviation
Args:
-----
y (pandas.Series): independent variable
window_size (int): rolling window size
sigma (int): value for standard deviation
Returns:
--------
a dict (dict of 'standard_deviation': int, 'anomalies_dict': (index: value))
containing information about the points indentified as anomalies
"""
avg = moving_average(y, window_size)
avg_list = avg.tolist()
residual = y - avg
# Calculate the variation in the distribution of the residual
testing_std = pd.rolling_std(residual, window_size)
testing_std_as_df = pd.DataFrame(testing_std)
rolling_std = testing_std_as_df.replace(np.nan,
testing_std_as_df.ix[window_size - 1]).round(3).iloc[:,0].tolist()
std = np.std(residual)
return {'stationary standard_deviation': round(std, 3),
'anomalies_dict': collections.OrderedDict([(index, y_i)
for index, y_i, avg_i, rs_i in zip(count(),
y, avg_list, rolling_std)
if (y_i > avg_i + (sigma * rs_i)) | (y_i < avg_i - (sigma * rs_i))])}
# This function is repsonsible for displaying how the function performs on the given dataset.
def plot_results(x, y, window_size, sigma_value=1,
text_xlabel="X Axis", text_ylabel="Y Axis", applying_rolling_std=False):
""" Helps in generating the plot and flagging the anamolies.
Supports both moving and stationary standard deviation. Use the 'applying_rolling_std' to switch
between the two.
Args:
-----
x (pandas.Series): dependent variable
y (pandas.Series): independent variable
window_size (int): rolling window size
sigma_value (int): value for standard deviation
text_xlabel (str): label for annotating the X Axis
text_ylabel (str): label for annotatin the Y Axis
applying_rolling_std (boolean): True/False for using rolling vs stationary standard deviation
"""
#plt.figure(figsize=(15, 8))
#plt.plot(x, y, "k.")
#y_av = moving_average(y, window_size)
#plt.plot(x, y_av, color='green')
#plt.xlim(0, 600)
#plt.xlabel(text_xlabel)
#plt.ylabel(text_ylabel)
# Query for the anomalies and plot the same
events = {}
if applying_rolling_std:
events = explain_anomalies_rolling_std(y, window_size=window_size, sigma=sigma_value)
else:
events = explain_anomalies(y, window_size=window_size, sigma=sigma_value)
x_anomaly = np.fromiter(events['ad'].keys(), dtype=int, count=len(events['ad']))
y_anomaly = np.fromiter(events['ad'].values(), dtype=float,
count=len(events['ad']))
return events
def calcPer(old,new):
#http://www.marcolazzari.it/blog/2010/08/24/come-calcolare-le-percentuali-con-excel-o-con-calc/
diffP= (1-float(new)/old)* 100
return diffP
# 4. Lets play with the functions
data = getData()
data_as_frame = pd.DataFrame(data[0], columns=['Months', 'SunSpots','Date'])
data_as_frame.head()
pprint("Media Armonica:")
pprint(data[1])
x = data_as_frame['Months']
Y = data_as_frame['SunSpots']
# plot the results
anomaly=plot_results(x, y=Y, window_size=10, text_xlabel="Minutes", sigma_value=2,text_ylabel="No. of Sun spots")
pprint(anomaly)
final = anomaly['ad']
rev= list(final.items())[-2]
pprint("Ultimo spike:")
pprint(rev)
#getData()
| [
"nuked8@gmail.com"
] | nuked8@gmail.com |
1613cc4678e8069425e88c9c92f2e3f813b95770 | 7e3fbcd8c130c5f98e45f20cfebee1d4d9531503 | /SERE/catalogos/urls.py | 6ff633552962a16b06c656edb6ff70d2deff891e | [] | no_license | hmachuca22/sere | 3653cc17a9fe8814b0733af6af364bb7e43169ae | ae318004729bb76277422d264a4fef6cdb108daa | refs/heads/master | 2020-11-25T16:45:29.576663 | 2019-12-20T00:53:49 | 2019-12-20T00:53:49 | 228,761,298 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,247 | py | from django.urls import path
from catalogos.views import CategoriaView
from catalogos.views import CategoriaNew
from catalogos.views import CategoriaEdit
from catalogos.views import CategoriaDel
from catalogos.views import ProductoViewSINREGISTRO,ProductoNewSINREGISTRO,ProductoViewINTERNOS,ProductoNewINTERNOS,ProductoEditINTERNOS
from catalogos.views import SubCategoriaView,SubCategoriaNew,SubCategoriaEdit,SubCategoriaDel,ProductoView,ProductoNew,ProductoEdit,categoria_print,historial_list
urlpatterns = [
path('categorias', CategoriaView.as_view(), name='categoria_list'),
path('categorias/new', CategoriaNew.as_view(), name='categoria_new'),
path('categoria/edit/<int:pk>', CategoriaEdit.as_view(), name='categoria_edit'),
path('categoria/delete/<int:pk>', CategoriaDel.as_view(), name='categoria_delete'),
path('categoria/dprint', categoria_print, name='categoria_print'),
path('categoria/dprint/<int:pk>', categoria_print, name='categoria_print_one'),
path('subcategorias', SubCategoriaView.as_view(), name='subcategoria_list'),
path('subcategorias/new', SubCategoriaNew.as_view(), name='subcategoria_new'),
path('subcategoria/edit/<int:pk>', SubCategoriaEdit.as_view(), name='subcategoria_edit'),
path('subcategoria/delete/<int:pk>', SubCategoriaDel.as_view(), name='subcategoria_delete'),
path('productos', ProductoView.as_view(), name='producto_list'),
path('producto/new', ProductoNew.as_view(), name='producto_new'),
path('producto/edit/<int:pk>', ProductoEdit.as_view(), name='producto_edit'),
path('productos/internos', ProductoViewINTERNOS.as_view(), name='producto_list_internos'),
path('producto/new/internos', ProductoNewINTERNOS.as_view(), name='producto_new_internos'),
path('producto/edit/internos/<int:pk>', ProductoEditINTERNOS.as_view(), name='producto_edit_internos'),
path('productos/sinregistro', ProductoViewSINREGISTRO.as_view(), name='producto_list_sinregistro'),
path('producto/new/sinregistro', ProductoNewSINREGISTRO.as_view(), name='producto_new_sinregistro'),
path('producto/edit/sinregistro/<int:pk>', ProductoEdit.as_view(), name='producto_edit_sinregistro'),
path('historial', historial_list, name='historial_list'),
] | [
"hmachuca19@gmail.com"
] | hmachuca19@gmail.com |
6a62b0983738172fa91c493cf2ac540bc6800c3f | 5df58c0a5796092d621486bc42b0754daac9d36f | /flashcardproject/flashcardapp/tests.py | 92868e1695605be05d35615c9f21b7479931b497 | [] | no_license | mcnguyenvn/Flashcard_team02 | 4af7fc4de825461daffc501d07fd28e0a4a25263 | fcfd0dc9e4ab6d1da7f876ccc1197dca92254280 | refs/heads/master | 2022-09-18T20:59:12.736488 | 2012-05-22T01:39:32 | 2012-05-22T01:39:32 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 722 | py | from django.test import TestCase
from django.test.client import Client
class UserFunctionTest(TestCase):
def setUp(self):
self.client = Client()
def test_creatingflashcard(self):
response = self.client.login(username='demo', password='123456')
self.assertTrue(response)
data = {
'title' : 'test',
'description':'test creating flashcard',
'grade' : 'first',
'subject' : 'art',
'Prompt 01':'test01',
'Answer 01' : 'answer01',
'Prompt 02':'test',
'Answer 02' : 'answer01'
}
response = self.client.post('/create/', data)
self.assertEqual(response.status_code, 200) | [
"cuongnm92@hotmail.com.vn"
] | cuongnm92@hotmail.com.vn |
be751f1b34a9337ee7ddbdecf8faf42c9f798399 | 09445ee10edc71baf61c0e286e7a79531fb76ba0 | /main.py | a1c4a1a93221990ed28908ebd856b9136e778957 | [] | no_license | yonkshi/master_thesis | e8d40fc0bd41d937946406bd9d4073bdffcaa05b | 0309061e7b3770b669adf15031ad76a672e2c1e6 | refs/heads/master | 2020-04-24T15:33:31.187952 | 2019-03-29T13:24:05 | 2019-03-29T13:24:05 | 172,072,683 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,577 | py | # -*- coding: utf-8 -*-
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
import tensorflow as tf
import random
import os
import datetime
from scipy.misc import imsave
from model import VAE
from data_manager import DataManager
tf.app.flags.DEFINE_integer("epoch_size", 2000, "epoch size")
tf.app.flags.DEFINE_integer("batch_size", 64, "batch size")
tf.app.flags.DEFINE_float("gamma", 1000.0, "gamma param for latent loss")
tf.app.flags.DEFINE_float("capacity_limit", 1000.0,
"encoding capacity limit param for latent loss")
tf.app.flags.DEFINE_integer("capacity_change_duration", 100000,
"encoding capacity change duration")
tf.app.flags.DEFINE_float("learning_rate", 5e-4, "learning rate")
tf.app.flags.DEFINE_string("checkpoint_dir", "checkpoints", "checkpoint directory")
tf.app.flags.DEFINE_string("log_file", "./log", "log file directory")
tf.app.flags.DEFINE_boolean("training", True, "training or not")
tf.app.flags.DEFINE_integer("input_width", 128, "input image pixel width")
tf.app.flags.DEFINE_integer("input_height", 128, "input image pixel height")
tf.app.flags.DEFINE_integer("input_channels", 3, "input image color channels (RGB default)")
tf.app.flags.DEFINE_integer("latent_dim", 256, "Dimension of latent space")
tf.app.flags.DEFINE_string("dataset", "Megaman", "log file directory")
flags = tf.app.flags.FLAGS
run_name = "gamma={0}, capacity_lim={1}, latent_dim={2}, input_dim={3}x{4}x{5}, dataset={6}, " \
"date={7}".format(flags.gamma,
flags.capacity_limit,
flags.latent_dim,
flags.input_width,
flags.input_height,
flags.input_channels,
flags.dataset,
datetime.datetime.now(),
)
run_logpath = os.path.join(flags.log_file, run_name)
run_checkpoint_path = os.path.join(flags.checkpoint_dir, run_name)
if not os.path.exists(run_logpath):
os.mkdir(run_logpath)
def train(sess,
model,
manager,
saver):
summary_writer = tf.summary.FileWriter(run_logpath, sess.graph)
n_samples = manager.sample_size
np.random.seed(1231)
reconstruct_check_images = manager.get_random_images(10)
indices = list(range(n_samples))
step = 0
# Training cycle
for epoch in range(flags.epoch_size):
print('\n===== EPOCH %d =====' % epoch)
# Shuffle image indices
random.shuffle(indices)
avg_cost = 0.0
total_batch = n_samples // flags.batch_size
print('>> Total Batch Size: %d' % total_batch)
# Loop over all batches
print('>> Training ', end='')
for i in range(total_batch):
# Generate image batch
print(".", end='')
batch_indices = indices[flags.batch_size * i: flags.batch_size * (i + 1)]
batch_xs = manager.get_images(batch_indices)
# Fit training using batch data
reconstr_loss, latent_loss, summary_str = model.partial_fit(sess, batch_xs, step)
summary_writer.add_summary(summary_str, step)
step += 1
# Image reconstruction check
print('')
print('>> Reconstruction check ... ', end='')
img_summary = reconstruct_check(sess, model, reconstruct_check_images)
print('Done')
# # Disentangle check
# print('>> Disentanglement check...', end='')
# disentangle_check(sess, model, manager)
# print('Done')
summary_writer.add_summary(img_summary, step)
# Save checkpoint
saver.save(sess, run_checkpoint_path + '/' + 'checkpoint', global_step=step)
def reconstruct_check(sess, model, images):
# Check image reconstruction
x_reconstruct, img_summary = model.reconstruct(sess, images)
if not os.path.exists("reconstr_img"):
os.mkdir("reconstr_img")
for i in range(len(images)):
print('>>>> Reconstructing image %d ' % i)
org_img = images[i].reshape([flags.input_width, flags.input_height, flags.input_channels])
org_img = org_img.astype(np.float32)
reconstr_img = x_reconstruct[i].reshape([flags.input_width, flags.input_height, flags.input_channels])
imsave("reconstr_img/org_{0}.png".format(i), org_img)
imsave("reconstr_img/reconstr_{0}.png".format(i), reconstr_img)
return img_summary
def disentangle_check(sess, model, manager, save_original=False):
'''
This code appears to be running disentanglement check (specified in the paper) with a preselected image.
So in my case, I am running with the preselected image 1337
:param sess:
:param model:
:param manager:
:param save_original:
:return:
'''
img = manager.get_image(1337)
if save_original:
imsave("original.png",
img.reshape([flags.input_width, flags.input_height, flags.input_channels]).astype(np.float32))
batch_xs = [img]
z_mean, z_log_sigma_sq = model.transform(sess, batch_xs)
z_sigma_sq = np.exp(z_log_sigma_sq)[0]
# Print variance
zss_str = ""
for i, zss in enumerate(z_sigma_sq):
str = "z{0}={1:.4f}".format(i, zss)
zss_str += str + ", "
# print(zss_str)
# Save disentangled images
z_m = z_mean[0]
n_z = 256 # Latent space dim
if not os.path.exists("disentangle_img"):
os.mkdir("disentangle_img")
for target_z_index in range(n_z):
for ri in range(n_z):
value = -3.0 + (6.0 / 9.0) * ri
z_mean2 = np.zeros((1, n_z))
for i in range(n_z):
if (i == target_z_index):
z_mean2[0][i] = value
else:
z_mean2[0][i] = z_m[i]
reconstr_img = model.generate(sess, z_mean2)
rimg = reconstr_img[0].reshape([flags.input_width, flags.input_height, flags.input_channels])
imsave("disentangle_img/check_z{0}_{1}.png".format(target_z_index, ri), rimg)
def load_checkpoints(sess):
saver = tf.train.Saver()
checkpoint = tf.train.get_checkpoint_state(run_checkpoint_path)
if checkpoint and checkpoint.model_checkpoint_path:
saver.restore(sess, checkpoint.model_checkpoint_path)
print("loaded checkpoint: {0}".format(checkpoint.model_checkpoint_path))
else:
print("Could not find old checkpoint")
if not os.path.exists(run_checkpoint_path):
os.mkdir(run_checkpoint_path)
return saver
def main(argv):
manager = DataManager()
manager.load()
sess = tf.Session()
model = VAE(
input_width=flags.input_width,
input_height=flags.input_height,
input_channels=flags.input_channels,
gamma=flags.gamma,
capacity_limit=flags.capacity_limit,
capacity_change_duration=flags.capacity_change_duration,
learning_rate=flags.learning_rate,
)
sess.run(tf.global_variables_initializer())
saver = load_checkpoints(sess)
if flags.training:
# Train
train(sess, model, manager, saver)
else:
reconstruct_check_images = manager.get_random_images(10)
# Image reconstruction check
reconstruct_check(sess, model, reconstruct_check_images)
# Disentangle check
disentangle_check(sess, model, manager)
if __name__ == '__main__':
tf.app.run()
| [
"nopctoday@gmail.com"
] | nopctoday@gmail.com |
97b2511118f4f47bd460d91c403d4e82b77fff9f | 316fef235cb8e446ea29f226418f0e9a79f2b2fa | /convex_hull/1405075_PlotScript.py | be4ac1ed76b699319c64322aa0eec8beceacc6f1 | [] | no_license | azmainadel/computational-geometry | a4a88719075bdb7759cf1cd3b3e94055cb316bc8 | ab0ef74d0d9557ca7f9dcfed82047d2ba421e032 | refs/heads/master | 2023-08-28T07:30:42.076334 | 2023-08-09T04:31:06 | 2023-08-09T04:31:06 | 253,682,150 | 3 | 1 | null | 2020-04-08T06:03:30 | 2020-04-07T03:56:39 | C++ | UTF-8 | Python | false | false | 602 | py | import matplotlib.pyplot as plt
p = []
q = []
with open('1405075_input1.txt') as f2:
next(f2)
for line in f2:
xC, yC = line.split()
p.append(xC)
q.append(yC)
fig = plt.figure()
ax = fig.add_subplot(111)
plt.scatter(p, q, s=10, c='r')
for xy in zip(p,q):
ax.annotate('(%s, %s)' % xy, xy=xy, textcoords='data')
plt.grid()
x = []
y = []
with open('1405075_output.txt') as f1:
for line in f1:
xC, yC = line.split()
x.append(xC)
y.append(yC)
plt.plot(x, y, linewidth=4)
plt.title("Convex Hull")
plt.xlabel("X")
plt.ylabel("Y")
plt.show()
| [
"azmainadel47@gmail.com"
] | azmainadel47@gmail.com |
5e32aa296d6c958b4c340ef0e125637e341d4fdf | 25e48619b6157be79a0cb3051f7b59af4e7a48bb | /main_name.py | 8d6bc8217700cfdc49b3b936482e765111907e08 | [] | no_license | Nana-Antwi/UVM-CS-21 | 8fdb2125f01820f063e7a2b3e40c4a0b3bd64c73 | 535b8e7efb61a0e4071766b4986e5d9b97952456 | refs/heads/master | 2020-04-17T09:29:27.027534 | 2019-01-18T19:19:18 | 2019-01-18T19:19:18 | 166,459,805 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 151 | py | #program that display personal information
def main():
print_my_name()
#function
def print_my_name():
print('Nana')
main():
| [
"noreply@github.com"
] | noreply@github.com |
2bc4f1ab2384a7e76f74641976a53715c495cc2a | b0c528e2650dec1ff011215537fc5ea536627966 | /main/urls.py | 58a80f586c83f786718a9f83bb105e9b11210f7e | [] | no_license | trinhgliedt/Python_Great_number_game | 9cb84a1bd95333df15140cc2e1c466d0911b7b19 | 8358c84012981b8dfaafb9017fc9a92450a98e7b | refs/heads/master | 2023-02-08T21:14:23.124896 | 2021-01-01T06:18:02 | 2021-01-01T06:18:02 | 325,926,745 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 135 | py | from django.urls import path
from . import views
urlpatterns = [
path('', views.index),
path('result/', views.process_form),
] | [
"chuot2008@gmail.com"
] | chuot2008@gmail.com |
e8da4fa3f3886a0e95df838344b3c25934dc0c78 | 9d87541ce623eeb1a82987c62d8ddb293b21ecdf | /groups/migrations/0001_initial.py | 764aa2174250e12c8c9337223696ff50aea2e2b0 | [] | no_license | rajasbhadke/code.fun.do | 24e76785a2689d28b84b0b188346a39ba3b12834 | c61b5e5d6c6ed251d597e8df1c3e336c3f55a662 | refs/heads/master | 2021-09-19T05:58:32.543269 | 2018-07-24T06:01:11 | 2018-07-24T06:01:11 | 107,448,756 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,313 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.11.3 on 2017-10-20 13:00
from __future__ import unicode_literals
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Events',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('EventType', models.CharField(max_length=255)),
('description', models.CharField(max_length=255)),
('date', models.DateField()),
('time', models.TimeField()),
],
),
migrations.CreateModel(
name='Group',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=255, unique=True)),
('slug', models.SlugField(allow_unicode=True, unique=True)),
('description', models.TextField(blank=True, default='')),
('description_html', models.TextField(blank=True, default='', editable=False)),
],
options={
'ordering': ['name'],
},
),
migrations.CreateModel(
name='GroupMember',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('group', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='memberships', to='groups.Group')),
('user', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='user_groups', to=settings.AUTH_USER_MODEL)),
],
),
migrations.AddField(
model_name='group',
name='members',
field=models.ManyToManyField(through='groups.GroupMember', to=settings.AUTH_USER_MODEL),
),
migrations.AlterUniqueTogether(
name='groupmember',
unique_together=set([('group', 'user')]),
),
]
| [
"rajasbhadke@gmail.com"
] | rajasbhadke@gmail.com |
96e52d0093401f85706495e10a173ccc42cff38e | 10571744731d20bbb463ae0559e1e5462319c2c3 | /ModelTesting/venv/Scripts/easy_install-3.7-script.py | 6e487f8ffbc499a22648a2dd081ce0e8a45cc5cd | [] | no_license | monsadan/MachineLearning | 6731c7a88e886525e31d47bc4e52e827f87af69a | c5d4eceb099a467cb5ae3bac6d039b9b4ccc2aff | refs/heads/master | 2022-11-05T23:15:08.634848 | 2020-08-26T22:57:01 | 2020-08-26T22:57:01 | 246,444,686 | 0 | 1 | null | 2022-10-06T19:11:06 | 2020-03-11T01:18:46 | null | UTF-8 | Python | false | false | 480 | py | #!C:\Users\dmmpc\Documents\GitHub\MachineLearning\ModelTesting\venv\Scripts\python.exe
# EASY-INSTALL-ENTRY-SCRIPT: 'setuptools==40.8.0','console_scripts','easy_install-3.7'
__requires__ = 'setuptools==40.8.0'
import re
import sys
from pkg_resources import load_entry_point
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(
load_entry_point('setuptools==40.8.0', 'console_scripts', 'easy_install-3.7')()
)
| [
"54322798+monsadan@users.noreply.github.com"
] | 54322798+monsadan@users.noreply.github.com |
e5e0ff738ca6d3556fc24b219daa7846f1e584f6 | e4d532362315fa21e9d403a5e04f27c5046c95cc | /app/core/migrations/0001_initial.py | 5c306375b4d4a859034f1000c1a32bcbe51eee15 | [
"MIT"
] | permissive | yvsreenivas/recipe-app-api | 722055d0abcd07a63f483c38ba6a3a419d55721e | 84c68edcbbc4060255589ea7666df76eed5fad53 | refs/heads/main | 2023-01-23T17:42:24.686340 | 2020-11-27T11:43:50 | 2020-11-27T11:43:50 | 297,965,550 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,709 | py | # Generated by Django 2.1.15 on 2020-11-26 09:24
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
('auth', '0009_alter_user_last_name_max_length'),
]
operations = [
migrations.CreateModel(
name='User',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('password', models.CharField(max_length=128, verbose_name='password')),
('last_login', models.DateTimeField(blank=True, null=True, verbose_name='last login')),
('is_superuser', models.BooleanField(default=False, help_text='Designates that this user has all permissions without explicitly assigning them.', verbose_name='superuser status')),
('email', models.EmailField(max_length=255, unique=True)),
('name', models.CharField(max_length=255)),
('is_active', models.BooleanField(default=True)),
('is_staff', models.BooleanField(default=False)),
('groups', models.ManyToManyField(blank=True, help_text='The groups this user belongs to. A user will get all permissions granted to each of their groups.', related_name='user_set', related_query_name='user', to='auth.Group', verbose_name='groups')),
('user_permissions', models.ManyToManyField(blank=True, help_text='Specific permissions for this user.', related_name='user_set', related_query_name='user', to='auth.Permission', verbose_name='user permissions')),
],
options={
'abstract': False,
},
),
]
| [
"yvsreenivas@yahoo.co.in"
] | yvsreenivas@yahoo.co.in |
f67df4a1e6f69b4ed8c0be1156e7f043268c888d | 5cf49cea129b36e7999bb5eed4a79a539dbc622c | /app/user/serializers.py | cd3b85465de2c1b98571250680297d34ea0c8124 | [
"MIT"
] | permissive | mydjangoprojects/recipe-app-api | 449674317f72b760b33d4e2bfc934b123b0c904b | 95f402e825b3bd22d1f2c63bf3de113e8c9044d7 | refs/heads/master | 2020-04-29T03:22:17.678663 | 2019-05-01T13:09:35 | 2019-05-01T13:09:35 | 175,806,784 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,696 | py | from django.contrib.auth import get_user_model, authenticate
from django.utils.translation import ugettext_lazy as _
from rest_auth.registration.serializers import \
RegisterSerializer as BaseRegisterSerializer
from rest_auth.serializers import LoginSerializer as BaseLoginSerializer
from allauth.account import app_settings as allauth_settings
from allauth.utils import email_address_exists
from allauth.account.adapter import get_adapter
from allauth.account.utils import setup_user_email
from rest_framework import serializers
class UserSerializer(serializers.ModelSerializer):
"""Serializer for the user object"""
class Meta:
model = get_user_model()
fields = ('email', 'password1', 'password2')
extra_kwargs = {'password': {'write_only': True, 'min_length': 5}}
def create(self, validated_data):
"""Create a new user with encrypted password and return it"""
return get_user_model().objects.create_user(**validated_data)
def update(self, instance, validated_data):
"""Update a user, setting the password correctly and return the user"""
password = validated_data.pop('password', None)
user = super().update(instance, validated_data)
if password:
user.set_password(password)
user.save()
return user
class AuthTokenSerializer(serializers.Serializer):
"""Serializer for the user authentication object"""
email = serializers.CharField()
password = serializers.CharField(
style={'input_type': 'password'},
trim_whitespace=False,
)
def validate(self, attrs):
"""Validate and authenticate the user"""
email = attrs.get('email')
password = attrs.get('password')
user = authenticate(
request=self.context.get('request'),
username=email,
password=password,
)
if not user:
msg = _('Unable to authenticate with provided credentials.')
raise serializers.ValidationError(msg, code='authentication')
attrs['user'] = user
return attrs
# Rest-Auth Override
class RegisterSerializer(BaseRegisterSerializer):
"""Serializer for register new User"""
username = None
email = serializers.EmailField(required=allauth_settings.EMAIL_REQUIRED)
password1 = serializers.CharField(required=True, write_only=True)
password2 = serializers.CharField(required=True, write_only=True)
def validate_email(self, email):
email = get_adapter().clean_email(email)
if allauth_settings.UNIQUE_EMAIL:
if email and email_address_exists(email):
raise serializers.ValidationError(
_("A user is already registered with this e-mail address.")
)
return email
def validate_password1(self, password):
return get_adapter().clean_password(password)
def validate(self, data):
if data['password1'] != data['password2']:
raise serializers.ValidationError(
_("The two password fields didn't match."))
return data
def get_cleaned_data(self):
return {
'password1': self.validated_data.get('password1', ''),
'email': self.validated_data.get('email', ''),
}
def save(self, request):
adapter = get_adapter()
user = adapter.new_user(request)
self.cleaned_data = self.get_cleaned_data()
adapter.save_user(request, user, self)
setup_user_email(request, user, [])
return user
class LoginSerializer(BaseLoginSerializer):
# Removing username to use only email
username = None
| [
"psykepro@abv.bg"
] | psykepro@abv.bg |
5b2b9b5810a6117b1af4440eda79ec7c15ae4f0b | c537ab4ad9769454c3bc894ae56e650ce74e96a3 | /lesson_07/code/06.文件的读取.py | 466343210c431990d28681bba33847f92b12a646 | [] | no_license | ZhouYao0627/course_python | f8b940ddf0b71c0d62b8e6f13121b7bd1de8a11b | 8236a3e431d45fd8353c55c181b5d9f84b9d4757 | refs/heads/master | 2023-08-15T06:22:15.325444 | 2021-10-01T08:38:27 | 2021-10-01T08:38:27 | 412,095,996 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 964 | py | file_name = 'demo2 .txt'
try:
with open(file_name) as file_obj:
# 如果直接调用read()他会将文本文件的所有内容全部都读出来
# 对于较大的文件,不要直接调用read()
# help(file_obj.read())
# read()可以接受一个size作为参数,该参数用来指定要读取的字符的数量
# 默认值是-1,它会读取文件中的所有字符
# 可以为size指定一个值,这样read()会读取指定数量的字符
# 每一次的读取都是从上一次读取到的位置开始读取的
# 如果字符的数量小于size,则会读取剩余所有的
# content = file_obj.read(-1)
content = file_obj.read(6)
content = file_obj.read(6)
content = file_obj.read(6)
content = file_obj.read(6)
print(content)
print(len(content))
except FileNotFoundError:
print(f'{file_name} 文件不存在···')
| [
"1607435943@qq.com"
] | 1607435943@qq.com |
f77c214c3496e3927ac4f3539fc029b10bbe7302 | f6c41523b1f03f447a0ace4fd01ddf9d5ba85a24 | /getCountriesbyRegion.py | 1f5d81a6df742fab4ba97a8288ab2ecf7eda9c18 | [] | no_license | AdroitAnandAI/Automated-Reasoning-with-ML-Knowledge-Graph-and-Machine-Comprehension-with-AI | cc39d75c99700ace3fd0f94d97665547b57dc947 | 252448c9885e1d47daa30aec7113fa732326e03a | refs/heads/main | 2023-01-03T04:38:21.228788 | 2020-10-18T07:26:33 | 2020-10-18T07:26:33 | 305,045,678 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,665 | py | from grakn.client import GraknClient
with GraknClient(uri="localhost:48555") as client:
with client.session(keyspace = "globe") as session:
with session.transaction().read() as transaction:
query = [
'match $country isa country, has countryname $cname,' +
' has indepyear $iy, has region "Caribbean"; $iy > 1980; get $cname;'
]
# has indepyear $iy, has population $p, has surfacearea $sc; $sc > 10000; $p < 1000000; get $cname, $p, $sc;
print("\nQuery:\n", "\n".join(query))
query = "".join(query)
iterator = transaction.query(query)
#can also use compute max like this but will get only highest value.
# if you want to find country with maximum lifeexpectancy, then u need
# to execute 2 queries, first to compute max and
# then the above query to "== max"
#compute max of lifeexpectancy, in country;
countries = [[ans.get("cname")] for ans in iterator]
print(countries)
# result = [ answer.id for answer in answers ]
for country in countries:
print(str(country[0].value()))
# print(": " country[1].value())
# print("\nResult:\n", result)
# answers = [ans.get("city") for ans in iterator]
# print(answers)
# result = [ answer.id for answer in answers ]
# for answer in answers:
# print(answer.type().__getattribute__('cityname'))
# print(answer.type().label())
# print("\nResult:\n", result) | [
"noreply@github.com"
] | noreply@github.com |
01b2b94a2a0eedae0205ccea536b5edd468b7316 | 2ab6afdcc194efa65f6122ff4eb9c32843d1611b | /visualize.py | 6b7464a64706b0353a58e9255d1fb255955b8f9d | [] | no_license | liuliuOD/Data-Visualization | 641b98d8bffc98ee6a6ed010309ee9b752081dcf | ed8e20a61f1105ea2513a17bfd592c153daf9e7a | refs/heads/master | 2020-05-17T14:36:42.355886 | 2019-04-28T00:57:14 | 2019-04-28T00:57:14 | 183,768,823 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,009 | py | import matplotlib.pyplot as plt
class visualize () :
# def __init__ (self) :
def scatter (self, x, y, color = 'g', marker = 'o', size = 8) :
plt.scatter(x, y, c = color, marker = marker, s = size)
return self
# def plot (self, x, y, color = 'g') :
# plt.plot(x, y, color = color, marker = marker)
# return self
def show (self) :
plt.show()
return self
def draw (self) :
plt.draw()
return self
def pause (self, pauseZone = 0.1) :
plt.pause(pauseZone)
return self
def close (self, window = "all") :
plt.close(window)
return self
def interactive (self, turn = True) :
if turn :
plt.ion()
else :
plt.ioff()
return self
def clearWindow (self) :
plt.clf()
return self
def save (self, imgName = "test.png") :
plt.savefig(imgName)
return self
| [
"liuliugit@gmail.com"
] | liuliugit@gmail.com |
2e8acef5d561c60f976847083658fb070708a1dd | c18ee10367a6b8bd3efa20d3e5f1345f8a065a73 | /hw2/code/src/regression.py | 4ec54b283222b2d08f8387ee925f81550a819ae6 | [] | no_license | michaelwu756/CSM146 | 9f6bf604a20b20483d268a9ec2dfbb21731d2d4e | 46599acfa792e9dc68f40b956b7583c6f23c623b | refs/heads/master | 2021-05-12T03:33:56.246498 | 2018-03-16T11:40:02 | 2018-03-16T11:40:02 | 117,620,119 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,292 | py | #!/usr/bin/python
# This code was adapted from course material by Jenna Wiens (UMichigan).
# python libraries
import os
import math
# numpy libraries
import numpy as np
# matplotlib libraries
import matplotlib.pyplot as plt
######################################################################
# classes
######################################################################
class Data :
def __init__(self, X=None, y=None) :
"""
Data class.
Attributes
--------------------
X -- numpy array of shape (n,d), features
y -- numpy array of shape (n,), targets
"""
# n = number of examples, d = dimensionality
self.X = X
self.y = y
def load(self, filename) :
"""
Load csv file into X array of features and y array of labels.
Parameters
--------------------
filename -- string, filename
"""
# determine filename
dir = os.path.dirname('__file__')
f = os.path.join(dir, '..', 'data', filename)
# load data
with open(f, 'r') as fid :
data = np.loadtxt(fid, delimiter=",")
# separate features and labels
self.X = data[:,:-1]
self.y = data[:,-1]
def plot(self, name, **kwargs) :
"""Plot data."""
if 'color' not in kwargs :
kwargs['color'] = 'b'
plt.scatter(self.X, self.y, **kwargs)
plt.xlabel('x', fontsize = 16)
plt.ylabel('y', fontsize = 16)
if name==None:
plt.show()
else:
plt.savefig(name)
plt.clf()
# wrapper functions around Data class
def load_data(filename) :
data = Data()
data.load(filename)
return data
def plot_data(X, y, name=None, **kwargs) :
data = Data(X, y)
data.plot(name, **kwargs)
class PolynomialRegression() :
def __init__(self, m=1, reg_param=0) :
"""
Ordinary least squares regression.
Attributes
--------------------
coef_ -- numpy array of shape (d,)
estimated coefficients for the linear regression problem
m_ -- integer
order for polynomial regression
lambda_ -- float
regularization parameter
"""
self.coef_ = None
self.m_ = m
self.lambda_ = reg_param
def generate_polynomial_features(self, X) :
"""
Maps X to an mth degree feature vector e.g. [1, X, X^2, ..., X^m].
Parameters
--------------------
X -- numpy array of shape (n,1), features
Returns
--------------------
Phi -- numpy array of shape (n,(m+1)), mapped features
"""
n,d = X.shape
# part b: modify to create matrix for simple linear model
# part g: modify to create matrix for polynomial model
m = self.m_
Phi = np.zeros((n,m+1))
for i in range(0,n):
val=[1]
index=[(m+1)*i]
for j in range(0,m):
val.append(val[j]*X.flat[i])
index.append(index[j]+1)
np.put(Phi, index, val)
return Phi
def fit_GD(self, X, y, eta=None,
eps=0, tmax=10000, verbose=False) :
"""
Finds the coefficients of a {d-1}^th degree polynomial
that fits the data using least squares batch gradient descent.
Parameters
--------------------
X -- numpy array of shape (n,d), features
y -- numpy array of shape (n,), targets
eta -- float, step size
eps -- float, convergence criterion
tmax -- integer, maximum number of iterations
verbose -- boolean, for debugging purposes
Returns
--------------------
self -- an instance of self
"""
if self.lambda_ != 0 :
raise Exception("GD with regularization not implemented")
if verbose :
plt.subplot(1, 2, 2)
plt.xlabel('iteration')
plt.ylabel(r'$J(\theta)$')
plt.ion()
plt.show()
X = self.generate_polynomial_features(X) # map features
n,d = X.shape
eta_input = eta
self.coef_ = np.zeros(d) # coefficients
err_list = np.zeros((tmax,1)) # errors per iteration
# GD loop
for t in xrange(tmax) :
# part f: update step size
# change the default eta in the function signature to 'eta=None'
# and update the line below to your learning rate function
if eta_input is None :
eta = 1/float(1+t)
else :
eta = eta_input
# part d: update theta (self.coef_) using one step of GD
# hint: you can write simultaneously update all theta using vector math
# track error
# hint: you cannot use self.predict(...) to make the predictions
y_pred = np.zeros(n)
for i in range(0,n):
np.put(y_pred,i,np.dot(X[i],self.coef_))
self.coef_=self.coef_-2*eta*np.dot(y_pred-y,X)
err_list[t] = np.sum(np.power(y - y_pred, 2)) / float(n)
# stop?
if t > 0 and abs(err_list[t] - err_list[t-1]) <= eps :
break
# debugging
if verbose :
x = np.reshape(X[:,1], (n,1))
cost = self.cost(x,y)
plt.subplot(1, 2, 1)
plt.cla()
plot_data(x, y)
self.plot_regression()
plt.subplot(1, 2, 2)
plt.plot([t+1], [cost], 'bo')
plt.suptitle('iteration: %d, cost: %f' % (t+1, cost))
plt.draw()
plt.pause(0.05) # pause for 0.05 sec
print 'number of iterations: %d' % (t+1)
return self
def fit(self, X, y, l2regularize = None ) :
"""
Finds the coefficients of a {d-1}^th degree polynomial
that fits the data using the closed form solution.
Parameters
--------------------
X -- numpy array of shape (n,d), features
y -- numpy array of shape (n,), targets
l2regularize -- set to None for no regularization. set to positive double for L2 regularization
Returns
--------------------
self -- an instance of self
"""
X = self.generate_polynomial_features(X) # map features
# part e: implement closed-form solution
# hint: use np.dot(...) and np.linalg.pinv(...)
# be sure to update self.coef_ with your solution
self.coef_=np.dot(np.linalg.pinv(np.dot(X.T,X)),np.dot(X.T,y))
def predict(self, X) :
"""
Predict output for X.
Parameters
--------------------
X -- numpy array of shape (n,d), features
Returns
--------------------
y -- numpy array of shape (n,), predictions
"""
if self.coef_ is None :
raise Exception("Model not initialized. Perform a fit first.")
X = self.generate_polynomial_features(X) # map features
# part c: predict y
n,d = X.shape
y = np.zeros(n)
for i in range(0,n):
np.put(y,i,np.dot(X[i],self.coef_))
return y
def cost(self, X, y) :
"""
Calculates the objective function.
Parameters
--------------------
X -- numpy array of shape (n,d), features
y -- numpy array of shape (n,), targets
Returns
--------------------
cost -- float, objective J(theta)
"""
# part d: compute J(theta)
n,d = X.shape
y_pred=self.predict(X)
cost = 0
for i in range(0,n):
cost+=(y_pred.flat[i]-y.flat[i])**2
return cost
def rms_error(self, X, y) :
"""
Calculates the root mean square error.
Parameters
--------------------
X -- numpy array of shape (n,d), features
y -- numpy array of shape (n,), targets
Returns
--------------------
error -- float, RMSE
"""
# part h: compute RMSE
n,d = X.shape
error = math.sqrt(self.cost(X,y)/n)
return error
def plot_regression(self, xmin=0, xmax=1, n=50, **kwargs) :
"""Plot regression line."""
if 'color' not in kwargs :
kwargs['color'] = 'r'
if 'linestyle' not in kwargs :
kwargs['linestyle'] = '-'
X = np.reshape(np.linspace(0,1,n), (n,1))
y = self.predict(X)
plot_data(X, y, **kwargs)
plt.show()
######################################################################
# main
######################################################################
def main() :
# load data
train_data = load_data('regression_train.csv')
test_data = load_data('regression_test.csv')
# part a: main code for visualizations
print 'Visualizing data...'
#plot_data(train_data.X, train_data.y, 'trainData.pdf')
#plot_data(test_data.X, test_data.y, 'testData.pdf')
# parts b-f: main code for linear regression
print 'Investigating linear regression...'
import time
reg=PolynomialRegression(1)
eta=0.0001
print("eta="+str(eta))
start=time.time()
reg.fit_GD(train_data.X, train_data.y, eta)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
eta=0.001
print("eta="+str(eta))
start=time.time()
reg.fit_GD(train_data.X, train_data.y, eta)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
eta=0.01
print("eta="+str(eta))
start=time.time()
reg.fit_GD(train_data.X, train_data.y, eta)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
eta=0.0407
print("eta="+str(eta))
start=time.time()
reg.fit_GD(train_data.X, train_data.y, eta)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
print("eta=1/(1+k)")
start=time.time()
reg.fit_GD(train_data.X, train_data.y)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
print("Closed form fit")
start=time.time()
reg.fit(train_data.X, train_data.y)
end=time.time()
print("coefficients="+str(reg.coef_))
print("cost="+str(reg.cost(train_data.X, train_data.y)))
print("time="+str(end-start))
print("")
# parts g-i: main code for polynomial regression
print 'Investigating polynomial regression...'
mPlot=[]
trainRMSE=[]
testRMSE=[]
for m in range(0,11):
reg=PolynomialRegression(m)
reg.fit(train_data.X, train_data.y)
print("m="+str(m))
print("train RMSE="+str(reg.rms_error(train_data.X, train_data.y)))
print("test RMSE="+str(reg.rms_error(test_data.X, test_data.y)))
print("")
mPlot.append(m)
trainRMSE.append(reg.rms_error(train_data.X, train_data.y))
testRMSE.append(reg.rms_error(test_data.X, test_data.y))
#line1, =plt.plot(mPlot, trainRMSE, '-', label='Training RMSE')
#line2, =plt.plot(mPlot, testRMSE, '-', label='Testing RMSE')
#plt.axis('auto')
#plt.xlabel('m')
#plt.ylabel('Root Mean Squared Error')
#plt.legend(loc='best')
#plt.title('Errors for Polynomial Regression of Degree m')
#plt.savefig("polynomialRegression.pdf")
#plt.clf()
print "Done!"
if __name__ == "__main__" :
main()
| [
"cheeserules43@gmail.com"
] | cheeserules43@gmail.com |
270280b5e47a20fdb3e29373151542e0bac4ad1e | a1f63894e73369a4649be4cc792c407b671f9547 | /src/brown_drivers/irobot_create_2_1/src/irobot_create_2_1/srv/_Dock.py | 86383b2435da8ebdf3092f8c38b5a4851722a6f1 | [] | no_license | parksj92/rosie | 5b21329a4d14f2e21cebd71f85bae154d714ff6f | 8f8b70a2cdb351c9a8ae1dbcfd8e9f242c1c7e03 | refs/heads/master | 2020-06-04T12:55:28.007459 | 2015-05-06T19:57:54 | 2015-05-06T19:57:54 | 31,720,276 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,149 | py | """autogenerated by genpy from irobot_create_2_1/DockRequest.msg. Do not edit."""
import sys
python3 = True if sys.hexversion > 0x03000000 else False
import genpy
import struct
class DockRequest(genpy.Message):
_md5sum = "d41d8cd98f00b204e9800998ecf8427e"
_type = "irobot_create_2_1/DockRequest"
_has_header = False #flag to mark the presence of a Header object
_full_text = """
"""
__slots__ = []
_slot_types = []
def __init__(self, *args, **kwds):
"""
Constructor. Any message fields that are implicitly/explicitly
set to None will be assigned a default value. The recommend
use is keyword arguments as this is more robust to future message
changes. You cannot mix in-order arguments and keyword arguments.
The available fields are:
:param args: complete set of field values, in .msg order
:param kwds: use keyword arguments corresponding to message field names
to set specific fields.
"""
if args or kwds:
super(DockRequest, self).__init__(*args, **kwds)
def _get_types(self):
"""
internal API method
"""
return self._slot_types
def serialize(self, buff):
"""
serialize message into buffer
:param buff: buffer, ``StringIO``
"""
try:
pass
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(_x))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(_x))))
def deserialize(self, str):
"""
unpack serialized message in str into this message instance
:param str: byte array of serialized message, ``str``
"""
try:
end = 0
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
def serialize_numpy(self, buff, numpy):
"""
serialize message with numpy array types into buffer
:param buff: buffer, ``StringIO``
:param numpy: numpy python module
"""
try:
pass
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(_x))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(_x))))
def deserialize_numpy(self, str, numpy):
"""
unpack serialized message in str into this message instance using numpy for array types
:param str: byte array of serialized message, ``str``
:param numpy: numpy python module
"""
try:
end = 0
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
_struct_I = genpy.struct_I
"""autogenerated by genpy from irobot_create_2_1/DockResponse.msg. Do not edit."""
import sys
python3 = True if sys.hexversion > 0x03000000 else False
import genpy
import struct
class DockResponse(genpy.Message):
_md5sum = "358e233cde0c8a8bcfea4ce193f8fc15"
_type = "irobot_create_2_1/DockResponse"
_has_header = False #flag to mark the presence of a Header object
_full_text = """bool success
"""
__slots__ = ['success']
_slot_types = ['bool']
def __init__(self, *args, **kwds):
"""
Constructor. Any message fields that are implicitly/explicitly
set to None will be assigned a default value. The recommend
use is keyword arguments as this is more robust to future message
changes. You cannot mix in-order arguments and keyword arguments.
The available fields are:
success
:param args: complete set of field values, in .msg order
:param kwds: use keyword arguments corresponding to message field names
to set specific fields.
"""
if args or kwds:
super(DockResponse, self).__init__(*args, **kwds)
#message fields cannot be None, assign default values for those that are
if self.success is None:
self.success = False
else:
self.success = False
def _get_types(self):
"""
internal API method
"""
return self._slot_types
def serialize(self, buff):
"""
serialize message into buffer
:param buff: buffer, ``StringIO``
"""
try:
buff.write(_struct_B.pack(self.success))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(_x))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(_x))))
def deserialize(self, str):
"""
unpack serialized message in str into this message instance
:param str: byte array of serialized message, ``str``
"""
try:
end = 0
start = end
end += 1
(self.success,) = _struct_B.unpack(str[start:end])
self.success = bool(self.success)
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
def serialize_numpy(self, buff, numpy):
"""
serialize message with numpy array types into buffer
:param buff: buffer, ``StringIO``
:param numpy: numpy python module
"""
try:
buff.write(_struct_B.pack(self.success))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(_x))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(_x))))
def deserialize_numpy(self, str, numpy):
"""
unpack serialized message in str into this message instance using numpy for array types
:param str: byte array of serialized message, ``str``
:param numpy: numpy python module
"""
try:
end = 0
start = end
end += 1
(self.success,) = _struct_B.unpack(str[start:end])
self.success = bool(self.success)
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
_struct_I = genpy.struct_I
_struct_B = struct.Struct("<B")
class Dock(object):
_type = 'irobot_create_2_1/Dock'
_md5sum = '358e233cde0c8a8bcfea4ce193f8fc15'
_request_class = DockRequest
_response_class = DockResponse
| [
"rbtying@aeturnalus.com"
] | rbtying@aeturnalus.com |
4c557ef7d1519e016f479c44b7cdfb3e045d59ab | 3681d5b2786f60dafdcac95fb2421307c0a59975 | /blogapp/blog/admin.py | 517dec69d08d243e68cb9ca116bd59ade671195f | [] | no_license | conectabell/Django_SimpleBlog | 45cd4e43bec87a4c435a7345618ca28a419149db | 498f5b5432ffc15be5adbd7b053ce92023430101 | refs/heads/master | 2021-01-09T05:53:30.002624 | 2017-04-23T23:32:47 | 2017-04-23T23:32:47 | 80,858,258 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 156 | py | from django.contrib import admin
#from django.db import models
from .models import Post
#admin.site.register(Rules, RulesAdmin)
admin.site.register(Post)
| [
"conectabell@gmail.com"
] | conectabell@gmail.com |
c118f02e997f6b98c12d016a62028a6e2795e9da | 30606e113697002f7a2a44a315d44796b428b944 | /src/neural_net_tests.py | 0ff0b44abfd150bcefd8474fe40e2fad66248ab6 | [
"MIT"
] | permissive | EoinM95/FYP | 2d0ac824e921bf45a502bbaba964a29c4be8a714 | 8b3bb73a0776dc65a8b577867d6a1d2f369fc7c5 | refs/heads/master | 2021-01-11T07:58:08.640167 | 2017-03-30T14:59:57 | 2017-03-30T14:59:57 | 72,131,784 | 1 | 0 | null | 2017-02-22T18:39:50 | 2016-10-27T17:27:49 | Python | UTF-8 | Python | false | false | 1,274 | py | """Tests for NeuralNetwork class"""
import unittest
import numpy as np
from neural_net import NeuralNetwork
class NeuralNetworkTests(unittest.TestCase):
"""Tests for NeuralNetwork class"""
def test_xor_gate(self):
"""Simulate XOR gate and ensure working"""
inputs = [[1.0, 1.0],
[1.0, 0.0],
[0.0, 1.0],
[0.0, 0.0]]
output_vector = [[0.0],
[1.0],
[1.0],
[0.0]]
inputs = np.array(inputs, dtype='float32')
output_vector = np.array(output_vector)
net = NeuralNetwork(inputs, output_vector)
net.train()
output = net.feed(np.array([[0, 1]], dtype='float32'))[0][0]
output = round(output, 3)
self.assertAlmostEqual(output, 1)
output = net.feed(np.array([[1, 0]], dtype='float32'))[0][0]
output = round(output, 3)
self.assertAlmostEqual(output, 1)
output = net.feed(np.array([[0, 0]], dtype='float32'))[0][0]
output = round(output, 3)
self.assertAlmostEqual(output, 0)
output = net.feed(np.array([[1, 1]], dtype='float32'))[0][0]
output = round(output, 3)
self.assertAlmostEqual(output, 0)
| [
"murphe43@tcd.ie"
] | murphe43@tcd.ie |
39d52659f57b9560c8b853368de4c568f3f77ca1 | 757d26801d42764be6d5368a5dbbe175f3ceeef4 | /venv/bin/mailmail | dcbd886bccad5d70e80dc0583034fcc2b0744ac4 | [] | no_license | luyuehm/scrapy_v1 | d1265eaad7836348be5911c06556ed23340d6de9 | 69da04ccfea06822f6bcd74f3abc3a44daabd300 | refs/heads/master | 2023-03-15T04:10:34.668307 | 2021-03-07T08:58:53 | 2021-03-07T08:58:53 | 346,115,989 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 267 | #!/Users/macbook/PycharmProjects/scrapy_v1/venv/bin/python3
# -*- coding: utf-8 -*-
import re
import sys
from twisted.mail.scripts.mailmail import run
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
sys.exit(run())
| [
"macbook@macbookdeMacBook-Pro.local"
] | macbook@macbookdeMacBook-Pro.local | |
edf40036c2140cfcdf728abbeb3fcfcbf0fdae99 | caf8d5ee09ef4412cd0c14b6a479714e50c68c1e | /small.py | 3fbd8ffd9d98384e49ed26d60a07560960e4f135 | [
"MIT"
] | permissive | victorShawFan/distill_BERT_into_RNN-CNN | 0dde6e89c8df5d8a9edf74aa528448d6f3214494 | 0ecbb033d229a2cc4f611964fd249a8eafdd710f | refs/heads/main | 2023-05-24T03:32:24.500902 | 2021-06-13T09:30:49 | 2021-06-13T09:30:49 | 377,155,136 | 1 | 0 | MIT | 2021-06-15T12:28:51 | 2021-06-15T12:28:50 | null | UTF-8 | Python | false | false | 4,352 | py | import torch, numpy as np
import torch.nn as nn
import torch.optim as optim
import torch.nn.functional as F
from torch.autograd import Variable
from keras.preprocessing import sequence
from utils import load_data
USE_CUDA = torch.cuda.is_available()
if USE_CUDA: torch.cuda.set_device(0)
LTensor = torch.cuda.LongTensor if USE_CUDA else torch.LongTensor
class RNN(nn.Module):
def __init__(self, x_dim, e_dim, h_dim, o_dim):
super(RNN, self).__init__()
self.h_dim = h_dim
self.dropout = nn.Dropout(0.2)
self.emb = nn.Embedding(x_dim, e_dim, padding_idx=0)
self.lstm = nn.LSTM(e_dim, h_dim, bidirectional=True, batch_first=True)
self.fc = nn.Linear(h_dim * 2, o_dim)
self.softmax = nn.Softmax(dim=1)
self.log_softmax = nn.LogSoftmax(dim=1)
def forward(self, x, lens):
embed = self.dropout(self.emb(x))
out, _ = self.lstm(embed)
hidden = self.fc(out[:, -1, :])
return self.softmax(hidden), self.log_softmax(hidden)
class CNN(nn.Module):
def __init__(self, x_dim, e_dim, h_dim, o_dim):
super(CNN, self).__init__()
self.emb = nn.Embedding(x_dim, e_dim, padding_idx=0)
self.dropout = nn.Dropout(0.2)
self.conv1 = nn.Conv2d(1, h_dim, (3, e_dim))
self.conv2 = nn.Conv2d(1, h_dim, (4, e_dim))
self.conv3 = nn.Conv2d(1, h_dim, (5, e_dim))
self.fc = nn.Linear(h_dim * 3, o_dim)
self.softmax = nn.Softmax(dim=1)
self.log_softmax = nn.LogSoftmax(dim=1)
def forward(self, x, lens):
embed = self.dropout(self.emb(x)).unsqueeze(1)
c1 = torch.relu(self.conv1(embed).squeeze(3))
p1 = torch.max_pool1d(c1, c1.size()[2]).squeeze(2)
c2 = torch.relu(self.conv2(embed).squeeze(3))
p2 = torch.max_pool1d(c2, c2.size()[2]).squeeze(2)
c3 = torch.relu(self.conv3(embed).squeeze(3))
p3 = torch.max_pool1d(c3, c3.size()[2]).squeeze(2)
pool = self.dropout(torch.cat((p1, p2, p3), 1))
hidden = self.fc(pool)
return self.softmax(hidden), self.log_softmax(hidden)
class Model(object):
def __init__(self, v_size):
self.model = None
self.b_size = 64
self.lr = 0.001
self.model = RNN(v_size, 256, 256, 2)
# self.model = CNN(v_size,256,128,2)
def train(self, x_tr, y_tr, l_tr, x_te, y_te, l_te, epochs=15):
assert self.model is not None
if USE_CUDA: self.model = self.model.cuda()
loss_func = nn.NLLLoss()
opt = optim.Adam(self.model.parameters(), lr=self.lr)
for epoch in range(epochs):
losses = []
accu = []
self.model.train()
for i in range(0, len(x_tr), self.b_size):
self.model.zero_grad()
bx = Variable(LTensor(x_tr[i:i + self.b_size]))
by = Variable(LTensor(y_tr[i:i + self.b_size]))
bl = Variable(LTensor(l_tr[i:i + self.b_size]))
_, py = self.model(bx, bl)
loss = loss_func(py, by)
loss.backward()
opt.step()
losses.append(loss.item())
self.model.eval()
with torch.no_grad():
for i in range(0, len(x_te), self.b_size):
bx = Variable(LTensor(x_te[i:i + self.b_size]))
by = Variable(LTensor(y_te[i:i + self.b_size]))
bl = Variable(LTensor(l_te[i:i + self.b_size]))
_, py = torch.max(self.model(Variable(LTensor(bx)), bl)[1], 1)
accu.append((py == by).float().mean().item())
print(np.mean(losses), np.mean(accu))
if __name__ == '__main__':
x_len = 50
# ----- ----- ----- ----- -----
# from keras.datasets import imdb
# v_size = 10000
# (x_tr,y_tr),(x_te,y_te) = imdb.load_data(num_words=v_size)
# ----- ----- ----- ----- -----
name = 'hotel' # clothing, fruit, hotel, pda, shampoo
(x_tr, y_tr, _), _, (x_te, y_te, _), v_size, _ = load_data(name)
l_tr = list(map(lambda x: min(len(x), x_len), x_tr))
l_te = list(map(lambda x: min(len(x), x_len), x_te))
x_tr = sequence.pad_sequences(x_tr, maxlen=x_len)
x_te = sequence.pad_sequences(x_te, maxlen=x_len)
clf = Model(v_size)
clf.train(x_tr, y_tr, l_tr, x_te, y_te, l_te)
| [
"luxuantao@126.com"
] | luxuantao@126.com |
83f87eada15ab8d5a659a10f018b7e6ffd3ba2ef | ca68dec0fa417f3276454cc27ed30066246947da | /stats.py | b250261d121b5c0d88c8c0fb2e7a03a2fa7cad1c | [] | no_license | jpanag1213/Project_stock | 0f431c6c83102957fe2617222c546c73a74435a7 | cbad6729850869d5fa53aab85f9812eebdfc5cad | refs/heads/master | 2020-04-19T08:59:24.160458 | 2019-04-11T08:15:34 | 2019-04-11T08:15:34 | 168,096,311 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 82,142 | py | # -*- coding: utf-8 -*-
"""
Created on 2019-02-23
to stats the order imbalance feature
#导出日线数据
@author: jiaxiong
"""
import numpy as np
import Data
import SignalTester
import pandas as pd
import os
import configparser
import time
from Utils import *
import matplotlib.pyplot as plt
import datetime
from numba import jit
from multiprocessing.dummy import Pool as mdP
from multiprocessing.pool import Pool as mpP
from functools import partial
class Stats(object):
def __init__(self, symbol, tradedate, quoteData,tradeData = None,futureData =None, outputpath = 'E://stats_test/'):
self.symbol = symbol
self.tradeDate = tradedate
if isinstance(quoteData,dict):
self.quoteData = quoteData
self.tradeData = tradeData
else:
self.quoteData = dict()
self.quoteData[symbol] = quoteData
self.tradeData = dict()
self.tradeData[symbol] = tradeData
self.outputpath = outputpath
self.futureData = futureData
if os.path.exists(outputpath) is False:
os.makedirs(outputpath)
if futureData == None:
self.columns =[
'exchangeCode', 'exchangeTime', 'latest', 'pre_close', 'open', 'high', 'low', 'upper_limit',
'lower_limit', 'status', 'tradeNos', 'tradeVolume', 'totalTurnover',
'bidPrice1', 'bidPrice2', 'bidPrice3', 'bidPrice4', 'bidPrice5', 'bidPrice6', 'bidPrice7',
'bidPrice8', 'bidPrice9', 'bidPrice10',
'askPrice1', 'askPrice2', 'askPrice3', 'askPrice4', 'askPrice5', 'askPrice6', 'askPrice7',
'askPrice8', 'askPrice9', 'askPrice10',
'bidVolume1', 'bidVolume2', 'bidVolume3', 'bidVolume4', 'bidVolume5', 'bidVolume6', 'bidVolume7',
'bidVolume8', 'bidVolume9', 'bidVolume10',
'askVolume1', 'askVolume2', 'askVolume3', 'askVolume4', 'askVolume5', 'askVolume6', 'askVolume7',
'askVolume8', 'askVolume9', 'askVolume10',
'total_bid_qty', 'total_ask_qty', 'weighted_avg_bid_price', 'weighted_avg_ask_price',
'iopv', 'yield_to_maturity']
else:
self.columns = [
'id', 'exchangeTime', 'latest', 'pre_close', 'settle','pre_settle','open', 'high', 'low','close', 'upper_limit',
'lower_limit', 'status', 'buy', 'tradeVolume', 'totalTurnover',
'bidPrice1', 'bidPrice2', 'bidPrice3', 'bidPrice4', 'bidPrice5',
'askPrice1', 'askPrice2', 'askPrice3', 'askPrice4', 'askPrice5',
'bidVolume1', 'bidVolume2', 'bidVolume3', 'bidVolume4', 'bidVolume5',
'askVolume1', 'askVolume2', 'askVolume3', 'askVolume4', 'askVolume5',
'open_interest','pre_open_interest', 'open_qty','close_qty']
def check_file(self,file,symbol = " "):
file.to_csv(self.outputpath+str(self.tradeDate)+'_'+symbol + '_checkquote.csv')
return 0
def Evaluation_times(self,tick = 30):
##监测前30分钟的因子值。订单结构等。
#定义:订单结构,因子设计等等。
#input quotedata tradedata
#output return?volatility?
eva_duration = 20 * tick
quoteData = self.quoteData[self.symbol[0]]
quoteData = pd.concat([quoteData.loc[datetime.datetime.strptime(str(self.tradeDate + ' 09:30:00'),'%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate + ' 9:45:00'), '%Y%m%d %H:%M:%S'):],quoteData.loc[datetime.datetime.strptime(str(self.tradeDate + ' 13:00:00'),'%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate + ' 13:15:00'), '%Y%m%d %H:%M:%S'):]])
columns = self.columns
stats.check_file(quoteData)
return 0
def lastNday(self,tradingday,tradingDates):
tradingDayFile ='./ref_data/TradingDay.csv'
tradingDays = pd.read_csv(tradingDayFile)
def opentime(self,symbol,closetime = ' 14:50:00'):
quoteData = self.quoteData[symbol]
quoteData.loc[:,'opentime'] = 0
quoteData.loc[datetime.datetime.strptime(str(self.tradeDate + ' 09:30:00'), '%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate +closetime), '%Y%m%d %H:%M:%S'),'opentime'] = 1
quoteData.loc[
datetime.datetime.strptime(str(self.tradeDate + closetime), '%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate + ' 14:57:00'), '%Y%m%d %H:%M:%S'), 'opentime'] =2
#stats.check_file(quoteData)
return quoteData
def time_cut(self,symbol,closetime = ' 14:50:00'):
quoteData = self.quoteData[symbol]
quoteData.loc[:,'opentime'] = 0
quoteData.loc[datetime.datetime.strptime(str(self.tradeDate + ' 09:30:03'), '%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate +closetime), '%Y%m%d %H:%M:%S'),'opentime'] = 1
quoteData = quoteData.loc[datetime.datetime.strptime(str(self.tradeDate + ' 09:30:03'), '%Y%m%d %H:%M:%S'):datetime.datetime.strptime(str(self.tradeDate +closetime), '%Y%m%d %H:%M:%S'),:]
#stats.check_file(quoteData)
return quoteData
def quote_cut(self,quoteData,num = 10):
price_column_ask= list()
price_column_bid= list()
volunme_column_ask = list()
volunme_column_bid = list()
for Nos in range(1,num+1):
price_column_ask.append('askPrice'+ str(Nos))
volunme_column_ask.append('askVolume' + str(Nos))
for Nos in range(1,num+1):
price_column_bid.append('bidPrice' + str(Nos))
volunme_column_bid.append('bidVolume'+ str(Nos))
return quoteData.loc[:,price_column_ask],quoteData.loc[:,volunme_column_ask],quoteData.loc[:,price_column_bid],quoteData.loc[:,volunme_column_bid]
def responseFun(self,symbol):
quoteData = self.quoteData[symbol]
quoteD = pd.DataFrame()
return 0
def spread_width(self):
quotedata.loc[:,'bid_spread'] = quotedata.loc[:,'bidPrice1'] - quotedata.loc[:,'bidPrice10']
quotedata.loc[:, 'bid_weight'] =( (quotedata.loc[:,'bidPrice1'] - quotedata.loc[:,'weighted_avg_bid_price'])**2*(quotedata.loc[:,'bidVolume1'])
+(quotedata.loc[:, 'bidPrice2'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume2'])
+(quotedata.loc[:, 'bidPrice3'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume3'])
+(quotedata.loc[:, 'bidPrice4'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume4'])
+(quotedata.loc[:, 'bidPrice5'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume5'])
+(quotedata.loc[:, 'bidPrice6'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume6'])
+(quotedata.loc[:, 'bidPrice7'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume7'])
+(quotedata.loc[:, 'bidPrice8'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume8'])
+(quotedata.loc[:, 'bidPrice9'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume9'])
+(quotedata.loc[:, 'bidPrice10'] - quotedata.loc[:, 'weighted_avg_bid_price'])**2 * (quotedata.loc[:, 'bidVolume10']))/ (quotedata.loc[:, 'total_bid_qty'])
quotedata.loc[:, 'ask_weight'] = ((quotedata.loc[:,'askPrice1'] - quotedata.loc[:,'weighted_avg_ask_price'])**2*(quotedata.loc[:,'askVolume1'])
+(quotedata.loc[:, 'askPrice2'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume2'])
+(quotedata.loc[:, 'askPrice3'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume3'])
+(quotedata.loc[:, 'askPrice4'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume4'])
+(quotedata.loc[:, 'askPrice5'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume5'])
+(quotedata.loc[:, 'askPrice6'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume6'])
+(quotedata.loc[:, 'askPrice7'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume7'])
+(quotedata.loc[:, 'askPrice8'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume8'])
+(quotedata.loc[:, 'askPrice9'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume9'])
+(quotedata.loc[:, 'askPrice10'] - quotedata.loc[:, 'weighted_avg_ask_price'])**2 * (quotedata.loc[:, 'askVolume10']))/ (quotedata.loc[:, 'total_ask_qty'])
quotedata.loc[:,'weight_ratio'] = quotedata.loc[:, 'bid_weight']/quotedata.loc[:, 'ask_weight']
'''
quotedata.loc[:,'ratio_mean'] = quotedata.loc[:,'weight_ratio'].ewm(10).mean()
quotedata.loc[:,'ratio_std'] = quotedata.loc[:,'weight_ratio'].ewm(10).std()
positivePos = quotedata.loc[:,'weight_ratio'] > (quotedata.loc[:,'ratio_mean'] + 2*quotedata.loc[:,'ratio_std'])
negativePos = quotedata.loc[:,'weight_ratio'] < (quotedata.loc[:,'ratio_mean'] -2* quotedata.loc[:,'ratio_std'])
quotedata.loc[:,'signal'] = 0
quotedata.loc[positivePos,'signal'] = 1
quotedata.loc[negativePos,'signal'] = -1
'''
quotedata.loc[:,'ask_spread'] = quotedata.loc[:,'askPrice1'] -quotedata.loc[:,'askPrice10']
quotedata.loc[:,'spread'] = quotedata.loc[:,'askPrice1'] -quotedata.loc[:,'bidPrice1']
return 0
def morning_session(self,symbol):
opendata = stats.time_cut(symbol,closetime = ' 09:30:06')
base_price_ask,base_volume_ask = stats.rolling_dealer(opendata,num=10,name='ask')
base_price_bid,base_volume_bid = stats.rolling_dealer(opendata,num=10,name='bid')
quotedata = stats.time_cut(symbol, closetime=' 14:50:00')
quotedata = stats.rolling_dealer(quotedata,num=5,name='ask',base_price = base_price_ask,base_qty= base_volume_ask)
quotedata = stats.rolling_dealer(quotedata,num=5,name='bid',base_price = base_price_bid,base_qty= base_volume_bid)
quotedata.loc[:,'ret_bid'] = np.log(quotedata.loc[:,'bid_leftPrice'] / quotedata.loc[:, 'pre_close'])
quotedata.loc[:,'ret_ask'] = np.log(quotedata.loc[:,'ask_leftPrice'] / quotedata.loc[:, 'pre_close'])
quotedata.loc[:, 'open_ret'] = np.log(quotedata.loc[:, 'open'] / quotedata.loc[:, 'pre_close'])
stats.check_file(quotedata,symbol)
return quotedata
def rolling_dealer(self,quotedata,num = 10,name = 'ask',base_price = 0,base_qty = 0):
#统计开盘时的订单情况。
#计算开盘时刻,盘口之外的挂单情况。
total_qty_name = 'total_'+name+'_qty'
avg_price_name = 'weighted_avg_'+name+'_price'
temp = quotedata.loc[:, avg_price_name]*quotedata.loc[:, total_qty_name] - base_price *1*base_qty/2
temp2 = quotedata.loc[:, total_qty_name] -1* base_qty/2
for number in range(1,num+1):
price_name = name+'Price'+str(number)
Volume_name = name+'Volume'+str(number)
temp = temp -quotedata.loc[:,price_name]*quotedata.loc[:,Volume_name]
temp2 = temp2 -quotedata.loc[:,Volume_name]
quotedata.loc[:,name+'_leftPrice'] = temp/temp2
quotedata.loc[:, name + '_BasePrice'] = base_price
quotedata.loc[:,name+'_leftVolume'] = temp2
quotedata.loc[:,name+'_BaseVolume'] =base_qty*1/2
if base_price ==0:
return (temp/temp2).mean(),temp2.mean()
else:
return quotedata
def cancel_order(self,symbol):
quotedata = self.quoteData[symbol]
tradeData = self.tradeData[symbol]
quote_time = pd.to_datetime(quotedata.exchangeTime.values).values
quotedata.loc[:,'tradeVolume'] =quotedata.loc[:, 'tradeVolume'].diff()
quotedata.loc[:,'Turnover'] =quotedata.loc[:, 'totalTurnover'].diff()
quotedata.index = pd.to_datetime(quotedata.loc[:, 'exchangeTime'].values, format='%Y-%m-%d %H:%M:%S')
temp_1 = pd.to_datetime(tradeData.loc[:,' nTime'], format='%Y-%m-%d %H:%M:%S.%f')
qqq = temp_1[0].microsecond
bid_order = tradeData.loc[:, ' nBSFlag'] == 'B'
ask_order = tradeData.loc[:, ' nBSFlag'] == 'S'
can_order = tradeData.loc[:, ' nBSFlag'] == ' '
tradeData.loc[bid_order, 'numbs_flag'] = 1
tradeData.loc[ask_order, 'numbs_flag'] = -1
tradeData.loc[can_order, 'numbs_flag'] = 0
tradeData.loc[:, 'temp'] = tradeData.loc[:, ' nPrice']
#tradeData.loc[pos, 'temp'] = np.nan
tradeData.temp.fillna(method='ffill', inplace=True)
lastrep = list(tradeData.temp.values[:-1])
lastrep.insert(0, 0)
lastrep = np.asarray(lastrep)
tradeData_quote = pd.merge(quotedata.loc[:, ['bidPrice1', 'askPrice1','tradeVolume','Turnover']],tradeData, left_index=True,right_index=True, how='outer')
tradeData_quote['bidPrice1'].fillna(method='ffill', inplace=True)
tradeData_quote['askPrice1'].fillna(method='ffill', inplace=True)
# tradeData_quote.to_csv(self.dataSavePath + './' + str(self.tradeDate.date()) + signal + ' ' + symbol + '.csv')
ActiveBuy = (tradeData_quote.loc[:, 'numbs_flag'] == 1)
ActiveSell = (tradeData_quote.loc[:, 'numbs_flag'] == -1)
tradeData_quote.loc[ActiveBuy, 'abVolume'] = tradeData_quote.loc[ActiveBuy, ' nVolume']
tradeData_quote.loc[ActiveSell, 'asVolume'] = tradeData_quote.loc[ActiveSell, ' nVolume']
tradeData_quote.loc[ActiveBuy, 'abPrice'] = tradeData_quote.loc[ActiveBuy, ' nTurnover']
tradeData_quote.loc[ActiveSell, 'asPrice'] = tradeData_quote.loc[ActiveSell, ' nTurnover']
#stats.check_file(tradeData_quote)
temp_quote_time = np.asarray(list(quote_time))
Columns_ = ['abVolume', 'asVolume','abPrice', 'asPrice']
resample_tradeData = tradeData_quote.loc[:, Columns_].resample('1S', label='right', closed='right').sum()
resample_tradeData = resample_tradeData.cumsum()
resample_tradeData = resample_tradeData.loc[temp_quote_time, :]
r_tradeData = resample_tradeData.diff()
r_tradeData.loc[:, 'abPrice'] = r_tradeData.loc[:, 'abPrice'] / r_tradeData.loc[:, 'abVolume']
r_tradeData.loc[:, 'asPrice'] = r_tradeData.loc[:, 'asPrice']/ r_tradeData.loc[:, 'asVolume']
r_tradeData.loc[:, 'VWAP'] =(r_tradeData.loc[:, 'asPrice']+ r_tradeData.loc[:, 'asPrice'])/ (r_tradeData.loc[:, 'asVolume']+r_tradeData.loc[:, 'abVolume'])
r_tradeData.loc[:,'timecheck'] =quotedata.loc[:, 'exchangeTime']
#stats.check_file(r_tradeData)
#stats.check_file(r_tradeData)
'''
quote_order = pd.merge(self.quoteData[symbol].loc[:, ['midp', 'midp_10', 'spread']], r_tradeData, left_index=True,
right_index=True, how='left')
# .loc[:,'midp'] =self.quoteData[symbol].loc[:,'midp']
quote_order.to_csv(self.outputpath + './ quote_order.csv')
# self.quoteData[symbol].loc[:, ['midp', 'bidVolume1', 'askVolume1']].to_csv(self.outputpath + './ quote_o.csv')
'''
r_tradeData.loc[:,'diff'] = r_tradeData.loc[:, 'abVolume'] - r_tradeData.loc[:, 'asVolume']
r_tradeData.loc[:,'cum_diff'] = r_tradeData.loc[:,'diff'].cumsum()
return r_tradeData
def plot(self):
return 0
def price_filter(self):
symbol = self.symbol[0]
midp = self.quoteData[symbol].loc[:, 'midp']
quotedata = self.quoteData[symbol]
bid_Volume10 = (quotedata.loc[:, 'bidVolume1'] + quotedata.loc[:, 'bidVolume2'] + quotedata.loc[:,
'bidVolume3']) * 1 / 10
ask_Volume10 = (quotedata.loc[:, 'askVolume1'] + quotedata.loc[:, 'askVolume2'] + quotedata.loc[:,
'askVolume3']) * 1 / 10
bid_Volume10_2 = (quotedata.loc[:, 'bidVolume1'] + quotedata.loc[:, 'bidVolume2'])
ask_Volume10_2 = (quotedata.loc[:, 'askVolume1'] + quotedata.loc[:, 'askVolume2'])
bid_price = (bid_Volume10 < quotedata.loc[:, 'bidVolume1']) + 2 * (
(bid_Volume10 > quotedata.loc[:, 'bidVolume1']) & (bid_Volume10 < bid_Volume10_2))
ask_price = (ask_Volume10 < quotedata.loc[:, 'askVolume1']) + 2 * (
(ask_Volume10 > quotedata.loc[:, 'askVolume1']) & (ask_Volume10 < ask_Volume10_2))
quotedata.loc[:, 'bid_per10'] = quotedata.loc[:, 'bidPrice1']
quotedata.loc[:, 'ask_per10'] = quotedata.loc[:, 'askPrice1']
quotedata.loc[:, 'bid_vol10'] = quotedata.loc[:, 'bidVolume1']
quotedata.loc[:, 'ask_vol10'] = quotedata.loc[:, 'askVolume1']
quotedata.loc[bid_price == 2, 'bid_per10'] = quotedata.loc[bid_price == 2, 'bidPrice2']
quotedata.loc[bid_price == 0, 'bid_per10'] = quotedata.loc[bid_price == 0, 'bidPrice3']
quotedata.loc[ask_price == 2, 'ask_per10'] = quotedata.loc[ask_price == 2, 'askPrice2']
quotedata.loc[ask_price == 0, 'ask_per10'] = quotedata.loc[ask_price == 0, 'askPrice3']
quotedata.loc[quotedata.loc[:, 'ask_per10'] == 0, 'ask_per10'] = np.nan
quotedata.loc[quotedata.loc[:, 'bid_per10'] == 0, 'bid_per10'] = np.nan
quotedata.loc[bid_price == 2, 'bid_vol10'] = quotedata.loc[bid_price == 2, 'bidVolume2']
quotedata.loc[bid_price == 0, 'bid_vol10'] = quotedata.loc[bid_price == 0, 'bidVolume3']
quotedata.loc[ask_price == 2, 'ask_vol10'] = quotedata.loc[ask_price == 2, 'askVolume2']
quotedata.loc[ask_price == 0, 'ask_vol10'] = quotedata.loc[ask_price == 0, 'askVolume3']
midp_2 = (quotedata.loc[:, 'ask_per10'] * quotedata.loc[:, 'bid_vol10'] + quotedata.loc[:,
'bid_per10'] * quotedata.loc[:,
'ask_vol10']) / (
quotedata.loc[:, 'bid_vol10'] + quotedata.loc[:, 'ask_vol10'])
self.quoteData[symbol].loc[:, 'midp_2'] = midp_2
# midp = (quotedata.loc[:, 'askPrice1'] * quotedata.loc[:, 'bidVolume1'] + quotedata.loc[:, 'bidPrice1'] * quotedata.loc[:,'askVolume1']) / (quotedata.loc[:, 'bidVolume1'] + quotedata.loc[:, 'askVolume1'])
mean_midp = midp_2.rolling(20).mean()
Minute = 6
ewm_midp = mean_midp.ewm(6 * 20).mean()
fig, ax = plt.subplots(1, figsize=(20, 12), sharex=True)
mean_midp_ = midp.rolling(20).mean()
ewm_midp_ = mean_midp_.ewm(6 * 20).mean()
not_point = list()
kp1_point = list()
kp2_point = list()
kp_1 = 0
kp_2 = 0
id_1 = 0
id_2 = 0
count = 0
std_ = mean_midp.ewm(6 * 20).std()
# std = mean_midp.ewm(M2* T)
STATE_test = list()
kp_list = list()
for row in zip(ewm_midp, std_):
count = count + 1
i = row[0]
j = row[1]
if i is not np.nan:
if (kp_1 != 0) & (kp_2 != 0) & (kp_1 == kp_1) & (kp_2 == kp_2):
kp_diff = kp_1 - kp_2
if (kp_diff * (i - kp_2) < 0):
if ((abs(i - kp_2)) > 4 * j):
# print(id_2-id_1)
not_point.append(i)
kp_1 = kp_2
kp_2 = i
id_1 = id_2
id_2 = count
STATE_test.append(3)
else:
not_point.append(kp_2)
STATE_test.append(2)
else:
kp_2 = i
id_2 = count
not_point.append(kp_1)
STATE_test.append(1)
else:
not_point.append(np.nan)
kp_1 = i
kp_2 = i
id_1 = count
id_2 = count
STATE_test.append(0)
kp1_point.append(kp_1)
kp2_point.append(kp_2)
else:
not_point.append(np.nan)
kp1_point.append(np.nan)
kp2_point.append(np.nan)
STATE_test.append(np.nan)
self.quoteData[symbol].loc[:, 'ewm'] = ewm_midp_
self.quoteData[symbol].loc[:, 'filter_ewm'] = ewm_midp
self.quoteData[symbol].loc[:, 'not'] = not_point
# self.quoteData[symbol].loc[:,'kp_1'] = kp1_point
# self.quoteData[symbol].loc[:,'kp_2'] = kp2_point
self.quoteData[symbol].loc[:, 'std_'] = std_
# self.quoteData[symbol].loc[:,'std_'] = std_
# self.quoteData[symbol].loc[:,'state'] = STATE_test
self.quoteData[symbol].loc[:, 'upper_bound'] = self.quoteData[symbol].loc[:, 'not'] + 3 * \
self.quoteData[symbol].loc[:, 'std_']
self.quoteData[symbol].loc[:, 'lower_bound'] = self.quoteData[symbol].loc[:, 'not'] - 3 * \
self.quoteData[symbol].loc[:, 'std_']
# negativePos = (self.quoteData[symbol].loc[:,'ewm']> (self.quoteData[symbol].loc[:,'not'] +3*self.quoteData[symbol].loc[:,'std_']))&(self.quoteData[symbol].loc[:,'ewm'].shift(-1) <(self.quoteData[symbol].loc[:,'not'].shift(-1) + 3*self.quoteData[symbol].loc[:,'std_'].shift(-1)))
# negativePos = (self.quoteData[symbol].loc[:,'ewm'].shift(1) > (self.quoteData[symbol].loc[:,'not'].shift(1) +3*self.quoteData[symbol].loc[:,'std_'].shift(1) ))&(self.quoteData[symbol].loc[:,'ewm'] <(self.quoteData[symbol].loc[:,'not'] + 3*self.quoteData[symbol].loc[:,'std_']))
positivePos = (self.quoteData[symbol].loc[:, 'ewm'].shift(1) < (
self.quoteData[symbol].loc[:, 'not'].shift(1) + 3 * self.quoteData[symbol].loc[:,
'std_'].shift(1))) & (
self.quoteData[symbol].loc[:, 'ewm'] > (
self.quoteData[symbol].loc[:, 'not'] + 3 * self.quoteData[symbol].loc[:,
'std_']))
# positivePos = (self.quoteData[symbol].loc[:,'ewm']< (self.quoteData[symbol].loc[:,'not'] -3*self.quoteData[symbol].loc[:,'std_']))&(self.quoteData[symbol].loc[:,'ewm'].shift(-1)>(self.quoteData[symbol].loc[:,'not'].shift(-1) - 3*self.quoteData[symbol].loc[:,'std_'].shift(-1)))
# positivePos = (self.quoteData[symbol].loc[:,'ewm'].shift(1) < (self.quoteData[symbol].loc[:,'not'].shift(1) -3*self.quoteData[symbol].loc[:,'std_'].shift(1) ))&(self.quoteData[symbol].loc[:,'ewm']>(self.quoteData[symbol].loc[:,'not'] - 3*self.quoteData[symbol].loc[:,'std_']))
negativePos = (self.quoteData[symbol].loc[:, 'ewm'].shift(1) > (
self.quoteData[symbol].loc[:, 'not'].shift(1) - 3 * self.quoteData[symbol].loc[:,
'std_'].shift(1))) & (
self.quoteData[symbol].loc[:, 'ewm'] < (
self.quoteData[symbol].loc[:, 'not'] - 3 * self.quoteData[symbol].loc[:,
'std_']))
'''
y_value = list(midp.iloc[:])
yvalue =list(ewm_midp)
yvalue_3 = list(ewm_midp_)
ax.plot(yvalue,label = '1')
ax.plot(y_value,label = '2')
ax.plot(not_point, marker='^', c='red')
#plt.savefig(self.dataSavePath + '/'+ str(self.tradeDate.date()) +symbol +signal+ '.jpg')
'''
quotedata = self.quoteData[symbol]
stats.check_file(quotedata,symbol)
return 0
def obi_fixedprice(self,symbol):
'''
定义一个相对的量:
bv1 = 1
股 if bv1 < 主动卖量
意味着:盘口的量只要按1个tick内就很有可能被打掉了。 ##这个事情怎么衡量呢 或者说打掉的分布。
av1 = 1
股 if av1 < 主动买量
obi = log(bv / av)
去除obi过分大的情况
large_obi = abs(obi) > exp(2)(7.3
倍左右)
obi[large_obi] = 0 ##todo 这里可以修改 。这里我想可能设置的不够好。如果是0的话。也体现不了obi从-2到2变动的情况。
##备选方式1 obi[large_obi] = k * obi[large_obi] ,k<<1
初始化盘口量计数。last_obi
价格不变的情况下,计算obi_change = obi - last_obi ##这里有个计数的方式,表示跟踪N tick的 obi的变化。
midp变化的情况下,重置last_obi。
##这里有个trick,主要的原因在于当obi信号触发的时候,薄弱的盘口不稳定。导致midp的多次变化->last_obi重复出现->last_obi不一定可靠。
##不过有一点不变的是 厚单快被打掉的时候就进去了。
##厚单的情况仍然需要继续统计。
## 突然被打掉的情况是不清楚的。如下方的青岛港
'''
self.quoteData[symbol] = self.high_obi(symbol)
small_obi_bid = self.quoteData[symbol].loc[:, 'bidVolume1']<self.quoteData[symbol].loc[:, 'asVolume']
small_obi_ask = self.quoteData[symbol].loc[:, 'askVolume1']<self.quoteData[symbol].loc[:, 'abVolume']
self.quoteData[symbol].loc[small_obi_bid, 'bidVolume1'] = 1
self.quoteData[symbol].loc[small_obi_ask, 'askVolume1'] = 1
self.quoteData[symbol] .loc[:, 'obi'] = np.log(self.quoteData[symbol] .loc[:, 'bidVolume1']) - np.log(
self.quoteData[symbol] .loc[:, 'askVolume1'])
large_obi = np.abs(self.quoteData[symbol].loc[:,'obi']) > 2
self.quoteData[symbol].loc[large_obi, 'obi'] = 0
self.quoteData[symbol].loc[:, 'obi'] = np.log(self.quoteData[symbol].loc[:, 'bidVolume1']) - np.log(
self.quoteData[symbol].loc[:, 'askVolume1'])
# self.quoteData[symbol].loc[:, 'obi_' + str(window) + '_min'] = self.quoteData[symbol].loc[:,
# 'obi'].rolling(window * 60).mean()
self.quoteData[symbol].loc[:, 'obi_' + '_min'] = self.quoteData[symbol].loc[:, 'obi'].diff()
askPriceDiff = self.quoteData[symbol]['askPrice1'].diff()
bidPriceDiff = self.quoteData[symbol]['bidPrice1'].diff()
midPriceChange = self.quoteData[symbol]['midp'].diff()
self.quoteData[symbol].loc[:, 'priceChange'] = 1
self.quoteData[symbol].loc[midPriceChange == 0, 'priceChange'] = 0
obi_change_list = list()
last_obi = self.quoteData[symbol]['obi'].iloc[0]
tick_count = 0
row_count = 0
for row in zip(self.quoteData[symbol]['priceChange'], self.quoteData[symbol]['obi']):
priceStatus = row[0]
obi = row[1]
if (priceStatus == 1) or np.isnan(priceStatus):
tick_count = 0
last_obi = obi
else:
last_obi = self.quoteData[symbol]['obi'].iloc[row_count - tick_count]
tick_count = tick_count + 1
row_count = row_count + 1
obi_change = obi - last_obi
obi_change_list.append(obi_change)
self.quoteData[symbol].loc[:, 'obi'] = obi_change_list
return self.quoteData[symbol]
def volume_change(self,symbol):
#检测对应价格的变动量以及变动率。记录变动时间等等。这个后面有个更好更快的版本了。可以删去。
quotedata = stats.time_cut(symbol)
askPrice1 = quotedata.loc[:,'askPrice2']
bidPrice1 = quotedata.loc[:,'bidPrice2']
askVolume1 = quotedata.loc[:,'askVolume2']
bidVolume1 = quotedata.loc[:,'bidVolume2']
Time = quotedata.loc[:, 'exchangeTime']
MAX_PRICE_VOLUME_ASK = dict()
MAX_PRICE_VOLUME_BID = dict()
for row in zip(askPrice1,bidPrice1,askVolume1,bidVolume1,Time):
ap1 = row[0]
bp1 = row[1]
av1 = row[2]
bv1 = row[3]
time = row[4]
key_ask = list(MAX_PRICE_VOLUME_ASK.keys())
key_bid = list(MAX_PRICE_VOLUME_BID.keys())
if ap1 in key_ask:
MAX_PRICE_VOLUME_ASK[ap1][1] =(av1 -MAX_PRICE_VOLUME_ASK[ap1][0])
MAX_PRICE_VOLUME_ASK[ap1][0] = av1
MAX_PRICE_VOLUME_ASK[ap1][2] = MAX_PRICE_VOLUME_ASK[ap1][1]/(av1+1)
MAX_PRICE_VOLUME_ASK[ap1][4] = MAX_PRICE_VOLUME_ASK[ap1][4] + 1
if MAX_PRICE_VOLUME_ASK[ap1][2] > 0.5:
MAX_PRICE_VOLUME_ASK[ap1][3] = MAX_PRICE_VOLUME_ASK[ap1][3]+1
MAX_PRICE_VOLUME_ASK[ap1][5] = time
else:
MAX_PRICE_VOLUME_ASK[ap1] = [av1,0,0,0,0,time]
if bp1 in key_bid:
MAX_PRICE_VOLUME_BID[bp1][0] = max(bv1,MAX_PRICE_VOLUME_BID[bp1][0])
if MAX_PRICE_VOLUME_BID[bp1][0] == bv1:
MAX_PRICE_VOLUME_BID[bp1][1] = time
else:
MAX_PRICE_VOLUME_BID[bp1] = [bv1 ,time]
if bp1 in key_bid:
MAX_PRICE_VOLUME_BID[bp1][1] =(bv1 -MAX_PRICE_VOLUME_BID[bp1][0])
MAX_PRICE_VOLUME_BID[bp1][0] = bv1
MAX_PRICE_VOLUME_BID[bp1][2] = MAX_PRICE_VOLUME_BID[bp1][1]/(bv1+1)
MAX_PRICE_VOLUME_BID[bp1][4] = MAX_PRICE_VOLUME_BID[bp1][4] + 1
if MAX_PRICE_VOLUME_BID[bp1][2] > 0.5:
MAX_PRICE_VOLUME_BID[bp1][3] = MAX_PRICE_VOLUME_BID[bp1][3]+1
MAX_PRICE_VOLUME_ASK[bp1][5] = time
else:
MAX_PRICE_VOLUME_BID[bp1] = [av1,0,0,0,0,time]
ask_df = pd.DataFrame.from_dict(MAX_PRICE_VOLUME_ASK,orient='index',columns = ['volume_ask','diff_v_ask','diff_r_ask','diff_t_ask','diff_l_ask','time_ask'])
bid_df = pd.DataFrame.from_dict(MAX_PRICE_VOLUME_BID,orient='index',columns = ['volume_bid','diff_v_bid','diff_r_bid','diff_t_bid','diff_l_bid','time_bid'])
df = pd.merge(ask_df,bid_df,left_index=True,right_index= True,how = 'outer')
return df
def large_order(self,symbol,price = 5.9, closetime=' 14:50:00'):
'''
检测某一个价位当天的情况。
##用于检测有无单一大挂单。可能是虚假单。
包括所在当前tick盘口位置(location) bid是负号,ask是正号,
交易量(nvolume,从tradeorder接入) 对应该price的交易数目
加单量(add_cancel) 正的表示加单 负的表示撤单
加单绝对值(add_cancel_abs)
订单变化量(order diff)
##todo 和order变化不同的是 这里去除了交易量的影响来计算。
###按每个价格计算一次也太慢了
例子:20190409的青岛港 601298.SH price = 10.85
10.85这个价位出现bp10到bp1到转变到ap4,以第117行为例:
买单减少116手,交易是165手,那么新增订单是49手
再以123行为例
此时价格是卖3,卖单增加138手,交易数是108手,那么实际在10.85新增下单数目是246手。
'''
window = 50
quotedata = stats.time_cut(symbol,closetime=closetime )
Price_form = dict()
price_ask,volume_ask,price_bid,volume_bid = stats.quote_cut(quotedata)
quote_index = quotedata.index
price_quote = price_ask.loc[:,:] ==price
ask_cum_quote = 11 - pd.DataFrame(price_quote.cumsum(axis=1).sum(axis = 1),columns = ['location'])
ask_cum_quote.loc[ ask_cum_quote.loc[:,'location'] == 11,'location'] = 0
volume_ask.columns = price_ask.columns
price_quote = pd.DataFrame(price_quote,dtype=int)
vol_ask = price_quote*volume_ask
vol_ask = vol_ask.sum(axis = 1)
vol_ask = pd.DataFrame(vol_ask,columns = ['order'])
price_quote = price_bid.loc[:,:] ==price
volume_bid.columns = price_bid.columns
bid_cum_quote = -(11 - pd.DataFrame(price_quote.cumsum(axis=1).sum(axis=1), columns=['location']))
bid_cum_quote.loc[bid_cum_quote.loc[:, 'location'] == -11, 'location'] = 0
price_quote = pd.DataFrame(price_quote,dtype=int)
vol_bid = price_quote*volume_bid
vol_bid = -1*vol_bid.sum(axis = 1)
vol_bid = pd.DataFrame(vol_bid,columns = ['order'])
vol = vol_ask+vol_bid
inorderbook = vol.loc[:,'order']==0
vol.loc[inorderbook,:] = np.nan
vol.fillna(method='ffill', inplace=True)
vol_diff = vol.loc[:,'order'].diff()
vol.loc[:, 'order_diff'] = vol_diff
vol.loc[:, 'location'] = bid_cum_quote+ ask_cum_quote
tv = quotedata.loc[:,'tradeVolume'].diff()
#vol.loc[:,'tv'] = tv
tradedata_price= stats.tradeorder(symbol,quotedata,price = price)
vol.loc[:,'nVolume'] = tradedata_price.loc[:,' nVolume']
vol.loc[:, 'nVolume'] = vol.loc[:,'nVolume'].fillna(0)
vol.loc[:,'pre_order'] = vol.loc[:,'order']-vol.loc[:, 'order_diff']
tempPositive = vol.loc[:,'pre_order'] >0
tempNegative = vol.loc[:,'pre_order'] <0
vol.loc[:,'temp'] = 0
vol.loc[tempPositive,'temp'] = vol.loc[tempPositive, 'order_diff']
vol.loc[tempNegative,'temp'] = -1*vol.loc[tempNegative, 'order_diff']
vol.loc[:,'add_cancel'] =vol.loc[:,'temp'] +vol.loc[:,'nVolume']
vol.loc[:,'add_cancel_abs'] =vol.loc[:,'add_cancel'].abs()
diff_where = np.sign(vol.loc[:, 'order']).diff() !=0
last_time = (list(vol.loc[diff_where,:].index)[-1])
return vol#max(vol.loc[last_time:,'add_cancel_abs'])
def tradeorder(self,symbol,quotedata,price = 6.25):
'''
tradeorder
给定价格
输出当前以该价格交易的交易量。
### todo 这里没有识别是买单卖单。这个重要吗? (可以仔细思考下)
'''
tradeData = self.tradeData[symbol]
#print(tradeData.columns)
price_loc = tradeData.loc[:,' nPrice'] == price
tradeData_loc = tradeData.loc[price_loc,:]
temp_quote_time = quotedata.index
tradeData_ = tradeData_loc.resample('1S').sum()
tradeData_ = tradeData_.cumsum()
#print(tradeData_)
#stats.check_file(pd.DataFrame(tradeData_))
try:
tradeData_ = tradeData_.loc[temp_quote_time, :]
r_tradeData = tradeData_.diff()
return r_tradeData
except:
print('AError')
return tradeData_
def high_obi(self,symbol):
'''
##接入large_order_count(self,quotedata,large_margin_buy,large_margin_sell,num = 10)
定义量阈值,认为这是大单:
比如
N个tick的主动买量:abVolume: active
Buy
volume
N个tick的主动卖量:asVolume: active
Sell
volume
以askVolume
卖盘为例,如果买的交易量很小,那么一个小的挂单可以“卡住”价格。
如果
存在大于abVolume的量,返回满足条件的最大askPrice,Max_ask_price, 以及对应的askVolume, Max_ask_volume, 若无
则返回askPrice = np.nan
askVolume = 0
对应asVolume
和bid盘
有
Min_bid_volume, Min_bid_price
###
因子:
如果Min_bid_volume == 0 & Max_ask_volume != 0 = > bid侧无“大单”,ask侧有一个或以上的“大单”
## todo 这里涉及到交易量的预测,预期一个大单的出现可能是一个方向?
## todo 上面考虑的因子其实和obi有点违背,obi本身有设计到吃掉大单的行为,上面的因子是一种“远离”大单的行为。
'''
quotedata = self.quoteData[symbol]
tradeData = self.cancel_order(symbol)
tradeData.loc[:, 'cum_buy'] =tradeData.loc[:, 'abVolume'].rolling(10).sum()
tradeData.loc[:, 'cum_sell'] =tradeData.loc[:, 'asVolume'].rolling(10).sum()
quotedata.loc[:,'obi'] = np.log(quotedata.loc[:,'askVolume1'] /quotedata.loc[:,'bidVolume1'])
#quotedata.loc[:, 'tv'] = quotedata.loc[:,'tradeVolume'].diff(10)
volume_bid, volume_ask,bid_loc,ask_loc = self.large_order_count(quotedata,tradeData.loc[:, 'cum_buy'],tradeData.loc[:, 'cum_sell'],num= 10 )
vb = list()
va = list()
pb = list()
pb = list()
quotedata.loc[:,'large_bid'] = volume_bid
quotedata.loc[:,'large_ask'] = volume_ask
#quotedata.loc[large_ask,'large_ask'] = quotedata.loc[large_ask,'askPrice1']
#quotedata.loc[:,'large_bid'].fillna(method='ffill', inplace=True)
#quotedata.loc[:,'large_ask'].fillna(method='ffill', inplace=True)
quotedata.loc[:,'large_width'] = quotedata.loc[:,'large_ask'] - quotedata.loc[:,'large_bid']
'''
for row in zip(quotedata.index ,volume_bid,volume_ask):
times = str(row[0])[10:]
print(times)
bp = row[1]
ap = row[2]
bid_max.append(stats.large_order(symbol,bp,times))
ask_max.append(stats.large_order(symbol,ap,times))
'''
#quotedata.loc[:,'bid_max'] = bid_max
#quotedata.loc[:,'ask_max'] = ask_max
quotedata.loc[:,'bid_loc'] = bid_loc
quotedata.loc[:,'ask_loc'] = ask_loc
quotedata.loc[:,'abVolume'] = tradeData.loc[:, 'abVolume']
quotedata.loc[:,'asVolume'] = tradeData.loc[:, 'asVolume']
return quotedata
def large_order_count(self,quotedata,large_margin_buy,large_margin_sell,num = 10):
###large_order_count
###用于统计大于某个阈值large_margin_buy(ask volume)/large_margin_sell(bid volume)的最小/大价格
###num :默认统计档数
###
###待优化效率
price_ask, volume_ask, price_bid, volume_bid = self.quote_cut(quotedata,num = num)
bid_ = (volume_bid).apply(lambda x : x - np.asarray(large_margin_sell))> 0
bid_ = pd.DataFrame(bid_,dtype= int)
price_bid.columns = bid_.columns
bid_ = price_bid *bid_
volume_zero = bid_ ==0
bid_[volume_zero] = np.nan
bid_ = bid_.max(axis =1)
bid_loc = (price_bid).apply(lambda x : x == np.asarray(bid_))
bid_loc = pd.DataFrame(bid_loc,columns=volume_bid.columns,dtype= int)
bid_loc =( bid_loc * volume_bid ).sum(axis=1)
#zero_bid_loc = bid_loc ==0
#bid_loc[zero_bid_loc] = np.nan
ask_ = (volume_ask).apply(lambda x : x - np.asarray(large_margin_buy))> 0
ask_ = pd.DataFrame(ask_,dtype= int)
#volume_loc = (volume_bid).apply(lambda x: x == np.asarray(large_margin_buy))
price_ask.columns = ask_.columns
ask_ = price_ask *ask_
volume_zero = ask_ ==0
ask_[volume_zero] = np.nan
ask_ = ask_.min(axis =1)
ask_loc = (price_ask).apply(lambda x : x == np.asarray(ask_))
ask_loc = pd.DataFrame(ask_loc,columns=volume_ask.columns,dtype= int)
ask_loc =( ask_loc * volume_ask ).sum(axis=1)
#zero_ask_loc = ask_loc ==0
#ask_loc[~zero_ask_loc] = np.nan
#bid_.fillna(method='ffill', inplace=True)
#ask_.fillna(method='ffill', inplace=True)
#bid_loc.fillna(method='ffill', inplace=True)
#ask_loc.fillna(method='ffill', inplace=True)
return bid_,ask_,bid_loc,ask_loc
def point_monitor(self, symbol, point_list):
###链接过滤后的序列化处理,类似于分钟数据的采样后的策略
### todo 如果有合适的key point 序列,可以考虑他们之间的统计性质 类似分钟数据下的策略构造。
### 建议用作统计两个或多个key point 之间的性质。而不仅仅是一个信号过滤器。
quotedata = self.quoteData[symbol]
tradeData = self.cancel_order(symbol)
quotedata.loc[:, 'kp'] = point_list
positivePos = quotedata.loc[:, 'kp'] == 1
negativePos = quotedata.loc[:, 'kp'] == -1
quotedata.loc[~positivePos & ~negativePos, 'kp'] = np.nan
quotedata.loc[:, 'kp'].fillna(method='ffill', inplace=True)
quotedata.loc[:, 'kp_diff'] = quotedata.loc[:, 'kp'].diff()
quotedata.loc[:, 'asVolume_cum'] = quotedata.loc[:, 'asVolume'].cumsum()
quotedata.loc[:, 'abVolume_cum'] = quotedata.loc[:, 'abVolume'].cumsum()
tick_count = 0
row_count = 0
last_as = quotedata.loc[:, 'asVolume_cum'].iloc[0]
last_ab = quotedata.loc[:, 'abVolume_cum'].iloc[0]
cum_as = list()
cum_ab = list()
midp_change = list()
last_mid = quotedata.loc[:, 'midp'].iloc[0]
for row in zip(quotedata.loc[:, 'asVolume_cum'], quotedata.loc[:, 'abVolume_cum'], quotedata.loc[:, 'kp_diff'],quotedata.loc[:, 'midp']):
ak = row[0]
ab = row[1]
kp = row[2]
midp = row[3]
if (kp != 0):
tick_count = 0
last_as = ak
last_ab = ab
last_mid = midp
else:
last_as = quotedata.loc[:, 'asVolume_cum'].iloc[row_count - tick_count]
last_ab = quotedata.loc[:, 'abVolume_cum'].iloc[row_count - tick_count]
last_mid= quotedata.loc[:, 'midp'].iloc[row_count - tick_count]
tick_count = tick_count+1
#print(last_as)
row_count = row_count + 1
as_change = ak - last_as
ab_change = ab - last_ab
mid_change = midp - last_mid
cum_as.append( ak - last_as)
cum_ab.append( ab - last_ab)
midp_change.append(mid_change)
quotedata.loc[:, 'midp_change'] = midp_change
quotedata.loc[:, 'cum_as'] = cum_as
quotedata.loc[:, 'cum_ab'] = cum_ab
quotedata.loc[:, 'ab_as'] = quotedata.loc[:, 'cum_ab'] -quotedata.loc[:, 'cum_as']
tick_count = 0
row_count = 0
grad = 0
vol_cum = 0
pri_cum = 0
vm = list()
vs= list()
for row in zip(quotedata.loc[:, 'ab_as'],quotedata.loc[:, 'midp_change'], quotedata.loc[:, 'kp'], quotedata.loc[:, 'kp_diff']):
volume_change = row[0]
price_change = row[1]
key_point = row[2]
kp = row[3]
if (kp!= 0):
grad = 0
tick_count = 0
vol_mean = 0
vol_std = 0
else:
vol_mean = quotedata.loc[:, 'midp_change'].iloc[(row_count - tick_count):row_count].mean()
vol_std = quotedata.loc[:, 'midp_change'].iloc[(row_count - tick_count):row_count].std()
tick_count = tick_count + 1
row_count = row_count + 1
vm.append(vol_mean)
vs.append(vol_std)
quotedata.loc[:, 'vm'] = vm
quotedata.loc[:, 'vs'] = vs
return quotedata
def volume_imbalance_bar(self,symbol):
### vol imbalance bar
### 采样的一种方式
#tradeData = self.tradeData
T = 50
a = 1
exp_para = 10
trade_list = self.cancel_order(symbol)
count = 0
pre_bar = 0
pre_bar_list = list()
theta_bar_list = list()
bar_label = list()
temp_list = list()
count_list = list()
std_list = list()
theta_bar = 0
pre_bar_std = 0
#trade_list.loc[:, 'abVolume'] = np.exp(trade_list.loc[:,'abVolume'] /1000 )
#trade_list.loc[:, 'asVolume'] = np.exp(trade_list.loc[:,'asVolume'] /1000)
for row in zip(trade_list.loc[:,'abVolume'],trade_list.loc[:,'asVolume']):
buy_volume = row[0]
sell_volume = row[1]
if np.isnan(buy_volume):
buy_volume = 0
if np.isnan(sell_volume):
sell_volume = 0
if count < T:
pre_bar = pre_bar + buy_volume - sell_volume
bar_label.append(0)
else:
theta_bar = theta_bar +buy_volume - sell_volume
if ((theta_bar)- (pre_bar) )>a * pre_bar_std:
bar_label.append(1)
temp_list.append(theta_bar)
pre_bar_df = (pd.DataFrame(temp_list))
pre_bar_df_ewm = (pre_bar_df.ewm(exp_para).mean())
theta_bar = 0.0
if pre_bar_df_ewm.shape[0]>1:
pre_bar_std = (pre_bar_df.ewm(exp_para).std()).iloc[-1,0]
else:
pre_bar_std = 0
#print(pre_bar_df_ewm.iloc[-1,0])
pre_bar = pre_bar_df_ewm.iloc[-1,0]
##print(theta_bar)
#theta_bar = 0.0
elif ((theta_bar)- (pre_bar) )<-a* pre_bar_std:
bar_label.append(-1)
theta_bar = 0.0
temp_list.append(theta_bar)
pre_bar_df = (pd.DataFrame(temp_list))
pre_bar_df_ewm = (pre_bar_df.ewm(exp_para).mean())
if pre_bar_df_ewm.shape[0]>1:
pre_bar_std = (pre_bar_df.ewm(exp_para).std()).iloc[-1,0]
else:
pre_bar_std = 0
#print(pre_bar_df_ewm.iloc[-1,0])
pre_bar = pre_bar_df_ewm.iloc[-1,0]
else:
bar_label.append(0)
count = count + 1
std_list.append(pre_bar_std)
pre_bar_list.append(pre_bar)
theta_bar_list.append(theta_bar)
count_list.append(count)
trade_list.loc[:, 'pre_bar_list'] = pre_bar_list
trade_list.loc[:, 'theta_bar_list'] = theta_bar_list
trade_list.loc[:, 'bar_label'] = bar_label
trade_list.loc[:, 'count_list'] = count_list
trade_list.loc[:, 'pre_bar_std'] = std_list
return trade_list
def response_fun(self,symbol):
quotedata = self.time_cut(symbol)
quotedata = quotedata[~quotedata.index.duplicated(keep='first')]
tradeData = self.tradeData[symbol]
quote_time = pd.to_datetime(quotedata.exchangeTime.values).values
quotedata.loc[:, 'tradeVolume'] = quotedata.loc[:, 'tradeVolume'].diff()
quotedata.loc[:, 'Turnover'] = quotedata.loc[:, 'totalTurnover'].diff()
quotedata.loc[:,'VWAP'] = quotedata.loc[:,'Turnover'] / quotedata.loc[:, 'tradeVolume']
quotedata.loc[:,'spd'] = quotedata.loc[:, 'askPrice1'] - quotedata.loc[:, 'bidPrice1']
quotedata.loc[quotedata.loc[:,'spd'] == 0 , 'spd'] = 0.01
bid_order = tradeData.loc[:, ' nBSFlag'] == 'B'
ask_order = tradeData.loc[:, ' nBSFlag'] == 'S'
can_order = tradeData.loc[:, ' nBSFlag'] == ' '
tradeData.loc[bid_order, 'numbs_flag'] = 1
tradeData.loc[ask_order, 'numbs_flag'] = -1
tradeData.loc[can_order, 'numbs_flag'] = 0
temp_quote_time = np.asarray(list(quote_time))
tradeData.loc[:, 'temp'] = tradeData.loc[:, ' nPrice']
# tradeData.loc[pos, 'temp'] = np.nan
tradeData.temp.fillna(method='ffill', inplace=True)
lastrep = list(tradeData.temp.values[:-1])
lastrep.insert(0, 0)
lastrep = np.asarray(lastrep)
tradeData_quote = pd.merge(quotedata.loc[:, ['spd', 'tradeVolume', 'Turnover']], tradeData,
left_index=True, right_index=True, how='outer')
tradeData_quote.loc[:, 'spd'].fillna(method='ffill', inplace=True)
#ActiveBuy = (tradeData_quote.loc[:, 'numbs_flag'] == 1)
#ActiveSell = (tradeData_quote.loc[:, 'numbs_flag'] == -1)
# tradeData_quote.loc[ActiveBuy, 'abVolume'] = tradeData_quote.loc[ActiveBuy, ' nVolume']
#tradeData_quote.loc[ActiveSell, 'asVolume'] = tradeData_quote.loc[ActiveSell, ' nVolume']
#tradeData_quote.loc[ActiveBuy, 'abPrice'] = tradeData_quote.loc[ActiveBuy, ' nTurnover']
l_dict = dict()
for l in range(1,301,10):
#quotedata.loc[:, 'VWAP_l'] = quotedata.loc[:, 'VWAP'].shift(-l)
tradeData_quote.loc[:, 'VWAP_l'] = np.nan
tradeData_quote.loc[temp_quote_time,'VWAP_l'] = quotedata.loc[:, 'VWAP'].shift(-l)
tradeData_quote.loc[:,'VWAP_l'].fillna(method='ffill', inplace=True)
tradeData_quote.loc[:, 'rsp_func'] = tradeData_quote.loc[:, ' nVolume']* (- tradeData_quote.loc[:, ' nPrice'] + tradeData_quote.loc[:, 'VWAP_l']) *tradeData_quote.loc[:, 'numbs_flag'] / tradeData_quote.loc[:, 'spd']
#print(np.nanmean(tradeData_quote.loc[:, 'rsp_func'])/np.nanmean(tradeData_quote.loc[:, ' nVolume']))
l_dict[l] = np.nanmean(tradeData_quote.loc[:, 'rsp_func'])/np.nanmean(tradeData_quote.loc[:, ' nVolume'])
'''
temp_quote_time = np.asarray(list(quote_time))
Columns_ = ['Rp_func']
resample_tradeData = tradeData_quote.loc[:, Columns_].resample('1S', label='right', closed='right').mean()
resample_tradeData = resample_tradeData.cumsum()
resample_tradeData = resample_tradeData.loc[temp_quote_time, :]
r_tradeData = resample_tradeData.diff()
'''
'''
r_tradeData.loc[:, 'abPrice'] = r_tradeData.loc[:, 'abPrice'] / r_tradeData.loc[:, 'abVolume']
r_tradeData.loc[:, 'asPrice'] = r_tradeData.loc[:, 'asPrice'] / r_tradeData.loc[:, 'asVolume']
r_tradeData.loc[:, 'timecheck'] = quotedata.loc[:, 'exchangeTime']
'''
# stats.check_file(r_tradeData)
# stats.check_file(r_tradeData)
'''
quote_order = pd.merge(self.quoteData[symbol].loc[:, ['midp', 'midp_10', 'spread']], r_tradeData, left_index=True,
right_index=True, how='left')
# .loc[:,'midp'] =self.quoteData[symbol].loc[:,'midp']
quote_order.to_csv(self.outputpath + './ quote_order.csv')
# self.quoteData[symbol].loc[:, ['midp', 'bidVolume1', 'askVolume1']].to_csv(self.outputpath + './ quote_o.csv')
'''
#r_tradeData.loc[:, 'diff'] = r_tradeData.loc[:, 'abVolume'] - r_tradeData.loc[:, 'asVolume']
#r_tradeData.loc[:, 'cum_diff'] = r_tradeData.loc[:, 'diff'].cumsum()
#print(l_dict)
return l_dict
def price_concat(self,price_ask,price_bid):
## 统计当天出现的所有价格
ask_column = price_ask.columns
bid_column = price_bid.columns
price_list = list()
price_ = list()
for column in zip(ask_column,bid_column):
ask = column[0]
bid = column[1]
price_list.append (pd.Series(price_ask.loc[:,ask]).unique())
price_list.append(pd.Series(price_bid.loc[:,bid]).unique())
price_.append(list(price_ask.loc[:,ask]))
price_.append(list(price_bid.loc[:,bid]))
price_concat = [x for j in price_ for x in j]
price_ = pd.DataFrame(pd.Series(price_concat).unique(),columns=['today_price'])
return price_
def price_volume_fun(self,symbol):
'''
dataframe test:算出每个tick下的对应价格的volume,并放在一个以time 为index,price 为column 的dataframe内
对于该tick前未出现过的价位,其对应的volume 为nan,若出现,则对应的volume 为最新出现的volume,
为了区分bid和ask,其中askVolume 为正 volume,bidVolume 为负volume
在bp10 到ap10之间的价位,若不出现在盘口,则记为0,因为被吃掉了
计算每一个tick之间的订单变化情况:
order_change = diff(test)
计算posChange :正向变化总和 正向变化包括:ask方加单 ask方吃bid单,bid方撤单
posChange = (order_change*((order_change> 0 ).astype(int))).sum(axis = 1)
计算negChange :负向变化总和 负向变化包括:bid方加单 bid方吃ask单,ask方撤单
negChange = (order_change*((order_change<0).astype(int))).sum(axis = 1)
计算净变化totalchange和绝对变化abschange
totalchange = posChange - negchange
abschange = posChange + negchange
观测 20 tick内的一致性:
consistence = quotedata.loc[:, 'TOTALchange'].rolling(20).sum() / quotedata.loc[:, 'abs_change'].rolling(20).sum()
假如consistence 绝对高的情况下,totalchange占abschange的比重很大,那么意味着poschange 或者negchange其中一个相对很小。这里使用滚动均值方差来表示绝对高,参数为2
posMark = consistence > consistence_mean + 2 * consistence_std ##过高,一致poschange,ask方一致增强,表示卖信号
negMark = consistence < consistence_mean - 2 * consistence_std ##过低,一致negchange,bid方一致增强,表示买信号
'''
quotedata =self.quoteData[symbol]
quotedata = quotedata[~quotedata.index.duplicated(keep='first')]
#print(quotedata.index.duplicated(keep='first'))
price_ask, volume_ask, price_bid, volume_bid = self.quote_cut(quotedata,num= 10)
## price_today 当日的所有的价格序列。
len_time = len(quotedata.index)
count = 0
dict_ = [dict() for i in range(len_time) ]
volume_ask.columns = price_ask.columns
volume_bid.columns = price_bid.columns
dict_ask = self.dict_merge(price_ask.to_dict('index'),volume_ask.to_dict('index'))
dict_bid = self.dict_merge(price_bid.to_dict('index'),volume_bid.to_dict('index'))
pool = mdP(4)
zip_list = zip(dict_,dict_ask.items(), dict_bid.items())
test =pool.map(self.volume_loading, zip_list)
pool.close()
pool.join()
test = pd.DataFrame(test,index = price_ask.index)
test.fillna(method = 'ffill',inplace = True)
order_change = test.diff()
quotedata.loc[:,'posChange'] = (order_change*((order_change> 0 ).astype(int))).sum(axis = 1)
quotedata.loc[:,'negChange'] = (order_change*((order_change<0).astype(int))).sum(axis = 1)
quotedata.loc[:,'tradeVol'] = quotedata.loc[:,'tradeVolume'].diff()*1
quotedata.loc[:, 'TOTALchange'] = (quotedata.loc[:,'posChange'] + quotedata.loc[:,'negChange'])
quotedata.loc[:, 'abs_change'] =( abs(quotedata.loc[:,'posChange']) + abs(quotedata.loc[:,'negChange']))
over_vol = quotedata.loc[:, 'tradeVol']>= quotedata.loc[:, 'abs_change'] /2
quotedata.loc[:,'over_vol'] = 0
quotedata.loc[over_vol,'over_vol'] =1
quotedata.loc[:, 'cum_over_vol'] = (quotedata.loc[:,'over_vol']* quotedata.loc[:, 'tradeVol']).cumsum()
quotedata.loc[:, 'cum_over_vol_diff'] = quotedata.loc[:,'tradeVolume'] - quotedata.loc[:, 'cum_over_vol']
quotedata.loc[:, 'consistence'] = quotedata.loc[:, 'TOTALchange'].rolling(20).mean() / quotedata.loc[:, 'abs_change'].rolling(20).mean()
quotedata.loc[:,'consistence_mean'] = quotedata.loc[:, 'consistence'].rolling(20).mean()
quotedata.loc[:,'consistence_std'] = quotedata.loc[:, 'consistence'].rolling(20).std()
posMark =(quotedata.loc[:, 'consistence']> quotedata.loc[:,'consistence_mean']+2*quotedata.loc[:,'consistence_std'])#&((quotedata.loc[:, 'consistence']< quotedata.loc[:,'consistence_mean']+3*quotedata.loc[:,'consistence_std']))
negMark =(quotedata.loc[:, 'consistence']< quotedata.loc[:,'consistence_mean']-2*quotedata.loc[:,'consistence_std'])#&(quotedata.loc[:, 'consistence']> quotedata.loc[:,'consistence_mean']-3*quotedata.loc[:,'consistence_std'])
quotedata.loc[:,'upper_'] = quotedata.loc[:,'consistence_mean']+2*quotedata.loc[:,'consistence_std']
quotedata.loc[:,'lower_'] = quotedata.loc[:,'consistence_mean']-2*quotedata.loc[:,'consistence_std']
#negMark = quotedata.loc[:,'posChange'] > av_sum
#posMark = quotedata.loc[:,'negChange'] <- bv_sum
quotedata.loc[posMark, 'marker'] = 1
quotedata.loc[negMark, 'marker'] = -1
quotedata.loc[(~posMark) & (~negMark), 'marker'] =0
#quotedata.loc[:,'change_cum'] = quotedata.loc[:, 'TOTALchange'] .cumsum()
'''
key_point = quotedata.loc[:, 'marker'] !=0
quotedata_kp = quotedata.loc[key_point, :]
quotedata_kp.loc[:,'tc_change'] = quotedata_kp.loc[:,'change_cum'].diff()
quotedata_kp.loc[:,'tv_change'] = quotedata_kp.loc[:,'tradeVolume'].diff()
#quotedata.loc[:, 'consistence_diff'] = quotedata.loc[:, 'consistence_5'] - quotedata.loc[:, 'consistence_20']
para_matrix = pd.DataFrame()
para_matrix.loc[:,'midp'] =quotedata.loc[:,'midp']
para_matrix.loc[:,'posChange'] =quotedata.loc[:,'posChange']
para_matrix.loc[:,'negChange'] =quotedata.loc[:,'negChange']
para_matrix.loc[:,'price_shift_20'] = para_matrix.loc[:,'midp'].shift(20)
para_matrix.loc[:,'price_shift_50'] = para_matrix.loc[:,'midp'].shift(50)
para_matrix.loc[:,'price_diff_20'] = para_matrix.loc[:,'midp'] - para_matrix.loc[:,'price_shift_20']
para_matrix.loc[:,'price_diff_50'] = para_matrix.loc[:,'midp'] - para_matrix.loc[:,'price_shift_50']
large_change = abs(para_matrix.loc[:,'price_diff_20'])> para_matrix.loc[:,'midp'].iloc[0]*15/10000 + 0.01
para_matrix = para_matrix.loc[large_change,:]
'''
return quotedata
def price_volume_fun3(self,symbol):
'''
dataframe test:算出每个tick下的对应价格的volume,并放在一个以time 为index,price 为column 的dataframe内
对于该tick前未出现过的价位,其对应的volume 为nan,若出现,则对应的volume 为最新出现的volume,
为了区分bid和ask,其中askVolume 为正 volume,bidVolume 为负volume
在bp10 到ap10之间的价位,若不出现在盘口,则记为0,因为被吃掉了
计算每一个tick之间的订单变化情况:
order_change = diff(test)
计算posChange :正向变化总和 正向变化包括:ask方加单 ask方吃bid单,bid方撤单
posChange = (order_change*((order_change> 0 ).astype(int))).sum(axis = 1)
计算negChange :负向变化总和 负向变化包括:bid方加单 bid方吃ask单,ask方撤单
negChange = (order_change*((order_change<0).astype(int))).sum(axis = 1)
计算净变化totalchange和绝对变化abschange
totalchange = posChange - negchange
abschange = posChange + negchange
观测 20 tick内的一致性:
consistence = quotedata.loc[:, 'TOTALchange'].rolling(20).sum() / quotedata.loc[:, 'abs_change'].rolling(20).sum()
假如consistence 绝对高的情况下,totalchange占abschange的比重很大,那么意味着poschange 或者negchange其中一个相对很小。这里使用滚动均值方差来表示绝对高,参数为2
posMark = consistence > consistence_mean + 2 * consistence_std ##过高,一致poschange,ask方一致增强,表示卖信号
negMark = consistence < consistence_mean - 2 * consistence_std ##过低,一致negchange,bid方一致增强,表示买信号
'''
quotedata =self.quoteData[symbol]
quotedata = quotedata[~quotedata.index.duplicated(keep='first')]
#print(quotedata.index.duplicated(keep='first'))
price_ask, volume_ask, price_bid, volume_bid = self.quote_cut(quotedata,num= 10)
## price_today 当日的所有的价格序列。
len_time = len(quotedata.index)
count = 0
dict_ = [dict() for i in range(len_time) ]
volume_ask.columns = price_ask.columns
volume_bid.columns = price_bid.columns
dict_ask = self.dict_merge(price_ask.to_dict('index'),volume_ask.to_dict('index'))
dict_bid = self.dict_merge(price_bid.to_dict('index'),volume_bid.to_dict('index'))
pool = mdP(4)
zip_list = zip(dict_,dict_ask.items(), dict_bid.items())
test =pool.map(self.volume_loading, zip_list)
pool.close()
pool.join()
test = pd.DataFrame(test,index = price_ask.index)
test.fillna(method = 'ffill',inplace = True)
order_change = test.diff()
quotedata.loc[:,'posChange'] = (order_change*((order_change> 0 ).astype(int))).sum(axis = 1)
quotedata.loc[:,'negChange'] = (order_change*((order_change<0).astype(int))).sum(axis = 1)
quotedata.loc[:,'tradeVol'] = quotedata.loc[:,'tradeVolume'].diff()*1
quotedata.loc[:, 'TOTALchange'] = (quotedata.loc[:,'posChange'] + quotedata.loc[:,'negChange'])
quotedata.loc[:, 'abs_change'] =( abs(quotedata.loc[:,'posChange']) + abs(quotedata.loc[:,'negChange']))
over_vol = quotedata.loc[:, 'tradeVol']>= quotedata.loc[:, 'abs_change'] /2
quotedata.loc[:,'over_vol'] = 0
quotedata.loc[over_vol,'over_vol'] =1
quotedata.loc[:, 'cum_over_vol'] = (quotedata.loc[:,'over_vol']* quotedata.loc[:, 'tradeVol']).cumsum()
quotedata.loc[:, 'cum_over_vol_diff'] = quotedata.loc[:,'tradeVolume'] - quotedata.loc[:, 'cum_over_vol']
quotedata.loc[:, 'consistence'] = quotedata.loc[:, 'TOTALchange'].rolling(20).sum() / quotedata.loc[:, 'abs_change'].rolling(20).sum()
quotedata.loc[:, 'consistence'] = quotedata.loc[:, 'consistence'].ewm(20).mean()
quotedata.loc[:,'consistence_mean'] = quotedata.loc[:, 'consistence'].rolling(20).mean()
quotedata.loc[:,'consistence_std'] = quotedata.loc[:, 'consistence'].rolling(20).std()
quotedata.loc[:, 'regular'] = 0
quotedata.loc[quotedata.loc[:,'consistence_mean']>quotedata.loc[:,'consistence_std'], 'regular'] = 0.5
quotedata.loc[quotedata.loc[:,'consistence_mean']<-quotedata.loc[:,'consistence_std'], 'regular'] = -0.5
posMark =(quotedata.loc[:, 'consistence']> quotedata.loc[:,'consistence_mean']+(2-quotedata.loc[:, 'regular'])*quotedata.loc[:,'consistence_std'])#&((quotedata.loc[:, 'consistence']< quotedata.loc[:,'consistence_mean']+3*quotedata.loc[:,'consistence_std']))
negMark =(quotedata.loc[:, 'consistence']< quotedata.loc[:,'consistence_mean']-(2+quotedata.loc[:, 'regular'])*quotedata.loc[:,'consistence_std'])#&(quotedata.loc[:, 'consistence']> quotedata.loc[:,'consistence_mean']-3*quotedata.loc[:,'consistence_std'])
quotedata.loc[:,'upper_'] = quotedata.loc[:,'consistence_mean']+(2+quotedata.loc[:, 'regular'])*quotedata.loc[:,'consistence_std']
quotedata.loc[:,'lower_'] = quotedata.loc[:,'consistence_mean']-(2-quotedata.loc[:, 'regular'])*quotedata.loc[:,'consistence_std']
#negMark = quotedata.loc[:,'posChange'] > av_sum
#posMark = quotedata.loc[:,'negChange'] <- bv_sum
quotedata.loc[posMark, 'marker'] = 1
quotedata.loc[negMark, 'marker'] = -1
quotedata.loc[(~posMark) & (~negMark), 'marker'] =0
quotedata.loc[:,'change_cum'] = quotedata.loc[:, 'TOTALchange'] .cumsum()
'''
key_point = quotedata.loc[:, 'marker'] !=0
quotedata_kp = quotedata.loc[key_point, :]
quotedata_kp.loc[:,'tc_change'] = quotedata_kp.loc[:,'change_cum'].diff()
quotedata_kp.loc[:,'tv_change'] = quotedata_kp.loc[:,'tradeVolume'].diff()
#quotedata.loc[:, 'consistence_diff'] = quotedata.loc[:, 'consistence_5'] - quotedata.loc[:, 'consistence_20']
para_matrix = pd.DataFrame()
para_matrix.loc[:,'midp'] =quotedata.loc[:,'midp']
para_matrix.loc[:,'posChange'] =quotedata.loc[:,'posChange']
para_matrix.loc[:,'negChange'] =quotedata.loc[:,'negChange']
para_matrix.loc[:,'price_shift_20'] = para_matrix.loc[:,'midp'].shift(20)
para_matrix.loc[:,'price_shift_50'] = para_matrix.loc[:,'midp'].shift(50)
para_matrix.loc[:,'price_diff_20'] = para_matrix.loc[:,'midp'] - para_matrix.loc[:,'price_shift_20']
para_matrix.loc[:,'price_diff_50'] = para_matrix.loc[:,'midp'] - para_matrix.loc[:,'price_shift_50']
large_change = abs(para_matrix.loc[:,'price_diff_20'])> para_matrix.loc[:,'midp'].iloc[0]*15/10000 + 0.01
para_matrix = para_matrix.loc[large_change,:]
'''
return quotedata
def dict_merge(self,dict1,dict2):
for k in dict1.keys():
if k in dict2:
temp = dict()
for i in dict1[k].keys():
temp[(dict1[k][i])] = dict2[k][i]
dict1[k] = temp
return dict1
def volume_loading(self,row):
dict_y = row[0]
dict_ask = row[1][1]
dict_bid = row[2][1]
bid_nonzero = [k for k in dict_bid.keys() if k >0]
ask_nonzero = [k for k in dict_ask.keys() if k >0]
ask_max_volume =max(dict_ask.values())
#print(ask_max_volume)
bid_max_volume =max(dict_bid.values())
if len(bid_nonzero)>0:
lower_bound = min(bid_nonzero)
upper_bound = max(ask_nonzero)
price_range = np.linspace(lower_bound,upper_bound,num =round((upper_bound - lower_bound)/0.01) + 1).round(2)
for price in price_range:
if price in dict_ask.keys():
#dict_y[price] = dict_ask[price]/np.log(ask_max_volume+1)
dict_y[price] = dict_ask[price]
elif price in dict_bid.keys():
#dict_y[price] = -dict_bid[price]/ np.log(bid_max_volume+1)
dict_y[price] = - dict_bid[price]
else:
dict_y[price] = 0
else:
dict_y[0] = 0
return dict_y
def flow_detect(self,symbol):
### 检测流动性用 公式 Spread_k = 2 * (Dk-MK)/MK
quotedata = self.quoteData[symbol]
tradeData = self.tradeData[symbol]
quotedata.loc[:,'midp'] =(quotedata.loc[:,'bidPrice1'] * quotedata.loc[:,'askVolume1'] + quotedata.loc[:,'bidVolume1'] * quotedata.loc[:,'askPrice1'] ) / (quotedata.loc[:,'bidVolume1']+ quotedata.loc[:,'askVolume1'] )
quote_time = pd.to_datetime(quotedata.exchangeTime.values).values
quotedata.loc[:, 'tradeVolume'] = quotedata.loc[:, 'tradeVolume'].diff()
quotedata.loc[:, 'Turnover'] = quotedata.loc[:, 'totalTurnover'].diff()
quotedata.index = pd.to_datetime(quotedata.loc[:, 'exchangeTime'].values, format='%Y-%m-%d %H:%M:%S')
temp_1 = pd.to_datetime(tradeData.loc[:, ' nTime'], format='%Y-%m-%d %H:%M:%S.%f')
qqq = temp_1[0].microsecond
bid_order = tradeData.loc[:, ' nBSFlag'] == 'B'
ask_order = tradeData.loc[:, ' nBSFlag'] == 'S'
can_order = tradeData.loc[:, ' nBSFlag'] == ' '
tradeData.loc[bid_order, 'numbs_flag'] = 1
tradeData.loc[ask_order, 'numbs_flag'] = -1
tradeData.loc[can_order, 'numbs_flag'] = 0
cancel_order = tradeData.loc[:, ' nPrice'] == 0
tradeData.loc[:, 'temp'] = tradeData.loc[:, ' nPrice']
# tradeData.loc[pos, 'temp'] = np.nan
tradeData.temp.fillna(method='ffill', inplace=True)
lastrep = list(tradeData.temp.values[:-1])
lastrep.insert(0, 0)
lastrep = np.asarray(lastrep)
tradeData_quote = pd.merge(quotedata.loc[:, [ 'bidPrice1', 'askPrice1','midp', 'tradeVolume', 'Turnover']], tradeData,
left_index=True, right_index=True, how='outer')
tradeData_quote['midp'].fillna(method='ffill', inplace=True)
tradeData_quote['askPrice1'].fillna(method='ffill', inplace=True)
tradeData_quote['bidPrice1'].fillna(method='ffill', inplace=True)
# tradeData_quote.to_csv(self.dataSavePath + './' + str(self.tradeDate.date()) + signal + ' ' + symbol + '.csv')
ActiveBuy = (tradeData_quote.loc[:, 'numbs_flag'] == 1)
ActiveSell = (tradeData_quote.loc[:, 'numbs_flag'] == -1)
tradeData_quote.loc[ActiveBuy, 'abVolume'] = tradeData_quote.loc[ActiveBuy, ' nVolume']
tradeData_quote.loc[ActiveSell, 'asVolume'] = tradeData_quote.loc[ActiveSell, ' nVolume']
tradeData_quote.loc[ActiveBuy, 'abPrice'] = (tradeData_quote.loc[ActiveBuy, ' nTurnover'] - tradeData_quote.loc[ActiveBuy, ' nVolume']* tradeData_quote.loc[ActiveBuy, 'bidPrice1'])
tradeData_quote.loc[ActiveSell, 'asPrice'] = (tradeData_quote.loc[ActiveSell, ' nTurnover'] - tradeData_quote.loc[ActiveSell, ' nVolume']*tradeData_quote.loc[ActiveSell, 'askPrice1'])*-1
# stats.check_file(tradeData_quote)
temp_quote_time = np.asarray(list(quote_time))
Columns_ = ['abVolume', 'asVolume', 'abPrice', 'asPrice']
resample_tradeData = tradeData_quote.loc[:, Columns_].resample('1S', label='right', closed='right').sum()
resample_tradeData = resample_tradeData.cumsum()
resample_tradeData = resample_tradeData.loc[temp_quote_time, :]
r_tradeData = resample_tradeData.diff()
r_tradeData.loc[:, 'abSpread'] = r_tradeData.loc[:, 'abPrice'] / r_tradeData.loc[:, 'abVolume']
r_tradeData.loc[:, 'asSpread'] = r_tradeData.loc[:, 'asPrice'] / r_tradeData.loc[:, 'asVolume']
r_tradeData.loc[:, 'timecheck'] = quotedata.loc[:, 'exchangeTime']
# stats.check_file(r_tradeData)
# stats.check_file(r_tradeData)
'''
quote_order = pd.merge(self.quoteData[symbol].loc[:, ['midp', 'midp_10', 'spread']], r_tradeData, left_index=True,
right_index=True, how='left')
# .loc[:,'midp'] =self.quoteData[symbol].loc[:,'midp']
quote_order.to_csv(self.outputpath + './ quote_order.csv')
# self.quoteData[symbol].loc[:, ['midp', 'bidVolume1', 'askVolume1']].to_csv(self.outputpath + './ quote_o.csv')
'''
r_tradeData.loc[:, 'diff'] = r_tradeData.loc[:, 'abVolume'] - r_tradeData.loc[:, 'asVolume']
r_tradeData.loc[:, 'cum_diff'] = r_tradeData.loc[:, 'diff'].cumsum()
return r_tradeData
def PV_summary(self,symbol):
quotedata = self.price_volume_fun(symbol)
quotedata_2 = self.high_obi(symbol)
#quotedata_3 = self.obi_fixedprice(symbol)
quotedata_2 = quotedata_2[~quotedata_2.index.duplicated(keep='first')]
#quotedata_3 = quotedata_3[~quotedata_3.index.duplicated(keep='first')]
quotedata.loc[:,'large_bid'] = quotedata_2.loc[:,'large_bid']
quotedata.loc[:,'large_ask'] = quotedata_2.loc[:,'large_ask']
quotedata.loc[:,'bid_loc'] = quotedata_2.loc[:,'bid_loc']
quotedata.loc[:,'ask_loc'] = quotedata_2.loc[:,'ask_loc']
quotedata.loc[:,'obi'] = quotedata_2.loc[:,'obi']
quotedata.loc[:,'spread'] =quotedata.loc[:,'askPrice1'] - quotedata.loc[:,'bidPrice1']
negativePos =(quotedata.loc[:, 'marker'] == 1)&(quotedata.loc[:,'bid_loc']==0)& (quotedata.loc[:,'ask_loc']!=0)
positivePos = (quotedata.loc[:, 'marker'] == -1)&(quotedata.loc[:,'ask_loc']==0)& (quotedata.loc[:,'bid_loc']!=0)
quotedata.loc[negativePos,'signal_'] = -1
quotedata.loc[positivePos,'signal_'] = 1
quotedata.loc[(~positivePos) & (~negativePos),'signal_'] = 0
return quotedata
def vol_detect(self, symbol):
###算一算突破成功率?
###
quotedata = self.price_volume_fun(symbol)
tradeData = self.cancel_order(symbol)
tradeData.loc[:,'up_down'] =0
tradeData.loc[tradeData.loc[:,'diff'] <0,'up_down'] =-1
tradeData.loc[tradeData.loc[:,'diff'] >0,'up_down'] =1
tradeData.loc[:,'up_down_mean_long'] = tradeData.loc[:,'up_down'].rolling(30).mean()
tradeData.loc[:,'up_down_std_long'] = tradeData.loc[:,'up_down_mean_long'].rolling(30).std()
tradeData.loc[:,'up_down_mean_short'] = tradeData.loc[:,'up_down'].rolling(10).mean()
tradeData.loc[:,'up_down_std_short'] = tradeData.loc[:,'up_down_mean_short'].rolling(10).std()
return tradeData
def ProcessOrderInfo(self, sampleData, orderType = ' nBidOrder'):
#整合订单信息
##todo 可以考察一个订单被吃掉以后的情况用作验证想法( 可以结合到订单的分析,注意情况)
"""
This function is used to aggregate the orderInfo
:param orderInfo: the groupby object which is group by the bidorder or askorder
:return:整合得到以下信息:主动报单方向,主动报单量,主动成交量,被动成交量,撤单量,主动报单金额,主动成交金额,被动成交金额
撤单金额,主动报单价格
"""
# activeBuy = sampleData.groupby([orderType, ' nBSFlag'])
# activeBuyTime = activeBuy.first().loc[:, [' nTime']]
# activeBuyPrice= activeBuy.last().loc[:, [' nPrice']]
# activeBuy = pd.concat([activeBuy.sum().loc[:, [' nVolume', ' nTurnover']], activeBuyTime, activeBuyPrice], 1) # here, sort by level = 2 due to that level = 2 is the time index level, first two levels is order and bs flag
activeBuy = sampleData.groupby([orderType, ' nBSFlag']).agg({' nVolume': 'sum', ' nTurnover': 'sum', ' nTime': 'first', ' nPrice': 'last'}) # use agg can apply different type of
if orderType == ' nBidOrder':
orderDirection = 'B'
otherSideDirection = 'S'
else:
orderDirection = 'S'
otherSideDirection = 'B'
# start = time.time()
# activeBuy = activeBuy.sort_values(' nTime')
# activeBuy = activeBuy.reset_index()
# activeBuy.index = pd.to_datetime(pd.Series(map(lambda stime: self.tradeDate + str(stime),
# activeBuy.loc[:, ' nTime'])), format='%Y%m%d%H%M%S%f')
# activeBuy.index = list(map(lambda stime: datetime.datetime.strptime(self.tradeDate + str(stime), '%Y%m%d%H%M%S%f'),
# activeBuy.loc[:, ' nTime']))
# activeBuyB = activeBuy.loc[activeBuy.loc[:, ' nBSFlag'] == orderDirection, [orderType, ' nPrice', ' nVolume', ' nTurnover']] # which is the part of active buying
# activeBuyB.columns = ['order', 'auctionPrice', 'activeVolume', 'activeTurnover']
# activeBuyS = activeBuy.loc[activeBuy.loc[:, ' nBSFlag'] == otherSideDirection, [orderType, ' nPrice', ' nVolume', ' nTurnover']] # which is the part of active buying
# activeBuyS.columns = ['order', 'tradePrice', 'passiveVolume', 'passiveTurnover']
# activeBuyC = activeBuy.loc[activeBuy.loc[:, ' nBSFlag'] == ' ', [orderType, ' nVolume']] # which is the part of active buying and cancel
# activeBuyC.columns = ['order', 'cancelVolume']
# activeBuy = pd.merge(activeBuyB, activeBuyS, on='order', sort=False, how='left')
# activeBuy = pd.merge(activeBuy, activeBuyC, on='order', sort=False, how='left')
# activeBuy.index = activeBuyB.index
# end = time.time()
# start = time.time()
activeBuyB = activeBuy.iloc[activeBuy.index.get_level_values(' nBSFlag') == orderDirection]
if activeBuyB.shape[0] == 0:
return None
activeBuyB.columns = ['activeVolume', 'activeTurnover', ' nTime', 'auctionPrice']
activeBuyS = activeBuy.iloc[activeBuy.index.get_level_values(' nBSFlag') == otherSideDirection]
activeBuyS.columns = ['passiveVolume', 'passiveTurnover', ' nTime', 'tradePrice']
activeBuyC = activeBuy.iloc[activeBuy.index.get_level_values(' nBSFlag') == ' '].loc[:, [' nVolume', ' nTime']]
activeBuyC.columns = ['cancelVolume', ' nTime']
activeBuy = pd.merge(activeBuyB.reset_index(), activeBuyS.loc[:,['passiveVolume', 'passiveTurnover', 'tradePrice']].reset_index(), on=orderType, sort=False, how='left')
activeBuy = pd.merge(activeBuy, activeBuyC.loc[:, 'cancelVolume'].reset_index(), on=orderType, sort=False, how='left')
activeBuy.index = pd.to_datetime(activeBuy.loc[:, ' nTime'])
# end = time.time()
# print(end - start,' s')
activeBuy = activeBuy.rename(columns = {orderType:'order'})
activeBuy = activeBuy.loc[:, ['order', 'auctionVolume', 'auctionPrice', 'auctionTurnover',
'activeVolume', 'activeTurnover', 'passiveVolume', 'passiveTurnover', 'cancelVolume']].fillna(0)
activeBuy.loc[:, 'auctionVolume'] = activeBuy.loc[:, 'activeVolume'] + activeBuy.loc[:, 'passiveVolume'] + activeBuy.loc[:, 'cancelVolume']
activeBuy.loc[:, 'auctionTurnover'] = activeBuy.loc[:, 'auctionPrice'] * activeBuy.loc[:, 'auctionVolume'] / 100
activeBuy.loc[:, 'orderDirection'] = orderDirection
return activeBuy.loc[:, ['order', 'orderDirection', 'auctionVolume', 'auctionPrice', 'auctionTurnover',
'activeVolume', 'activeTurnover', 'passiveVolume', 'passiveTurnover', 'cancelVolume']]
def run(self,symbol):
print(symbol)
t1 = time.time()
#quotedata = stats.zaopan_stats(symbol)
#stats.cancel_order(symbol)
#stats.price_filter()
price_situation =pd.DataFrame.from_dict(self.response_fun(symbol),orient='index')
t2 = time.time()
#self.check_file(price_situation)
#price_situation
#price_situation = stats.high_obi(symbol,' 14:55:00')
#self.check_file(price_situation,symbol = symbol)
#t3 = time.time()
#print('cal time:'+str(t2-t1))
#print('writing time:'+str(t3-t2))
return price_situation
if __name__ == '__main__':
"""
test the class
['2019-01-02', '2019-01-03', '2019-01-04', '2019-01-07', '2019-01-08', '2019-01-09', '2019-01-10', '2019-01-11', '2019-01-14',
'2019-01-15', '2019-01-16', '2019-01-17', '2019-01-18', '2019-01-21', '2019-01-22', '2019-01-23', '2019-01-24', '2019-01-25',
'2019-01-28', '2019-01-29', '2019-01-30', '2019-01-31', '2019-02-01', '2019-02-11', '2019-02-12', '2019-02-13', '2019-02-14',
'2019-02-15', '2019-02-18', '2019-02-19', '2019-02-20', '2019-02-21', '2019-02-22', '2019-02-25', '2019-02-26', '2019-02-27',
'2019-02-28', '2019-03-01', '2019-03-04', '2019-03-05', '2019-03-06', '2019-03-07', '2019-03-08', '2019-03-11', '2019-03-12',
'2019-03-13', '2019-03-14', '2019-03-15', '2019-03-18', '2019-03-19', '2019-03-20', '2019-03-21', '2019-03-22', '2019-03-25',
'2019-03-26', '2019-03-27', '2019-03-28', '2019-03-29', '2019-04-01', '2019-04-02', '2019-04-03', '2019-04-04', '2019-04-08',
'2019-04-09', '2019-04-10']
"""
# data = Data('E:/personalfiles/to_zhixiong/to_zhixiong/level2_data_with_factor_added','600030.SH','20170516')
dataPath = '//192.168.0.145/data/stock/wind'
## /sh201707d/sh_20170703
t1 = time.time()
tradeDate = '20190410'
symbols_path = 'D:/SignalTest/SignalTest/ref_data/sh50.csv'
symbol_list = pd.read_csv(symbols_path)
symbols = symbol_list.loc[:,'secucode']
print(symbols)
symbols = ['600366.SH']
tradingDay = ['20190102', '20190103', '20190104', '20190107', '20190108', '20190109', '20190110',
'20190111', '20190114', '20190115', '20190116', '20190117', '20190118', '20190121',
'20190122', '20190123', '20190124', '20190125', '20190128', '20190129', '20190130',
'20190131', '20190201', '20190211', '20190212', '20190213', '20190214', '20190215',
'20190218', '20190219', '20190220', '20190221', '20190222', '20190225', '20190226',
'20190227',
'20190228', '20190301', '20190304', '20190305', '20190306', '20190307', '20190308',
'20190311',
'20190312', '20190314', '20190315', '20190318', '20190319', '20190320', '20190321',
'20190322',
'20190325',]
'''
'20190326', '20190328', '20190329', '20190401', '20190402', '20190403',
'20190404',
'20190408', '20190409']
'''
# price_situation =self.ProcessOrderInfo(data.tradeData[symbol],orderType = ' nAskOrder')
t2 = time.time()
stats_df = pd.DataFrame()
for tradeDate in tradingDay:
print(tradeDate)
data = Data.Data(dataPath,symbols, tradeDate,'' ,dataReadType= 'gzip', RAWDATA = 'True')
stats = Stats(symbols,tradeDate,data.quoteData,data.tradeData)
temp = stats.run(symbols[0])
stats_df.loc[:, tradeDate] = temp[0]
#print(data.tradeData[symbols[0]])
#file_ = stats_df.loc[:,tradeDate] = temp
stats.check_file(stats_df)
'''
q = pd.DataFrame()
multi_pool = mpP(4)
multi_pool.map(stats.run,symbols)
multi_pool.close()
multi_pool.join()
'''
t3 = time.time()
print('total:' + str(t3 - t2))
print('readData_time:' + str(t2 - t1))
#
print('Test end')
| [
"jpanag@163.com"
] | jpanag@163.com |
a554fe02fd9428bee0aa30e20f5fc252b6d25627 | d4f19cbcd4179b84069dcc588edc970fe30909c0 | /supervisors_commands/add_user.py | 707ed49ad48becef19fc21862d2dd4ffd959cd35 | [] | no_license | chypppre/opvs_bot | d77d7aa1d104d51fcc9d53e22207dfbaa1332d32 | e6e1c784d32747ff98d783cd60b6ac6f1155cf44 | refs/heads/master | 2020-09-14T12:30:52.774310 | 2019-12-02T09:40:33 | 2019-12-02T09:40:33 | 222,877,939 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 15,334 | py | import sqlite3
from telebot import types
from misc import DB_DIR, CHATS_ID
def add_to_tempo(uid, last_name, first_name, department):
"""Добавить во временную таблицу в БД"""
db = sqlite3.connect(DB_DIR)
cursor = db.cursor()
check_sql = "SELECT uid FROM tempo WHERE uid={}".format(uid)
cursor.execute(check_sql)
if len(cursor.fetchall()) == 1:
add_sql = "UPDATE tempo SET uid={}, last_name='{}', first_name='{}', department={}".format(
uid, last_name, first_name, department)
else:
add_sql = "INSERT INTO tempo VALUES ({}, '{}', '{}', {})".format(uid, last_name, first_name, department)
cursor.execute(add_sql)
db.commit()
def send_add_request(bot, uid):
"""Отправить запрос с данными в бота"""
kbrd = types.InlineKeyboardMarkup()
uc_btn = types.InlineKeyboardButton(text="КЭ", callback_data="add_uc")
uc_helper_btn = types.InlineKeyboardButton(text="Пом.КЭ", callback_data="add_uc_helper")
uc_super_btn = types.InlineKeyboardButton(text="Супер.КЭ", callback_data="add_uc_super")
#kb_btn = types.InlineKeyboardButton(text="КБухглатерия", callback_data="add_kb")
#kb_helper_btn = types.InlineKeyboardButton(text="Пом.КБухглатерия", callback_data="add_kb_helper")
#kb_super_btn = types.InlineKeyboardButton(text="Супер.КБухглатерия", callback_data="add_kb_super")
#elba_btn = types.InlineKeyboardButton(text="Эльба", callback_data="add_elba")
#elba_helper_btn = types.InlineKeyboardButton(text="Пом.Эльба", callback_data="add_elba_helper")
#elba_super_btn = types.InlineKeyboardButton(text="Супер.Эльба", callback_data="add_elba_super")
# fms_btn = types.InlineKeyboardButton(text="ФМС", callback_data="add_fms")
# fms_helper_btn = types.InlineKeyboardButton(text="Пом.ФМС", callback_data="add_fms_helper")
# fms_super_btn = types.InlineKeyboardButton(text="Супер.ФМС", callback_data="add_fms_super")
intern_btn = types.InlineKeyboardButton(text="Стажер", callback_data="add_intern")
mentor_btn = types.InlineKeyboardButton(text="Наставник", callback_data="add_intern_helper")
oo_btn = types.InlineKeyboardButton(text="ОО", callback_data="add_oo")
pip_btn = types.InlineKeyboardButton(text="ПиП", callback_data="add_pip")
decline_btn = types.InlineKeyboardButton(text="Отклонить", callback_data="decline")
db = sqlite3.connect(DB_DIR)
cursor = db.cursor()
sql = "SELECT uid, last_name, first_name, department FROM tempo WHERE uid={}".format(uid)
cursor.execute(sql)
data = cursor.fetchall()[0]
uid, last_name, first_name, department= data[0], data[1], data[2], data[3]
if department in [1,3]:
chat_id = CHATS_ID[0]
kbrd.row(uc_btn, uc_helper_btn, uc_super_btn)
kbrd.row(intern_btn, mentor_btn)
kbrd.row(oo_btn, pip_btn)
elif department == 5:
chat_id = CHATS_ID[6]
kbrd.row(kb_btn, kb_helper_btn, kb_super_btn)
elif department == 6:
chat_id = CHATS_ID[7]
kbrd.row(elba_btn, elba_helper_btn, elba_super_btn)
kbrd.row(decline_btn)
full_name = "{} {}".format(
last_name,
first_name)
text_for_admins = "{} просит добавить его в бота.\nUID = {}.".format(
full_name,
uid)
bot.send_message(
chat_id=chat_id,
text=text_for_admins,
reply_markup=kbrd)
def add_user(bot, call):
"""Добавить юзера согласно нажатой кнопки"""
text = call.message.text
text = text.split(" ")
cons_last_name, cons_first_name, uid = text[0], text[1], text[-1]
cons_full_name = "{} {}".format(cons_last_name, cons_first_name)
db = sqlite3.connect(DB_DIR)
cursor = db.cursor()
try:
if call.data == "add_uc":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс'.")
elif call.data == "add_kb":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс'.")
elif call.data == "add_elba":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс'.")
# elif call.data == "add_fms":
# sql = """INSERT INTO staff (uid, department, last_name, first_name)
# VALUES ({}, 2, '{}', '{}')""".format(uid, last_name, first_name)
elif call.data == "add_intern":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 3, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Набор дд.мм ххх кабинет'.")
elif call.data == "add_uc_helper":
try:
sql = """INSERT INTO helpers (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
elif call.data == "add_kb_helper":
try:
sql = """INSERT INTO helpers (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
elif call.data == "add_elba_helper":
try:
sql = """INSERT INTO helpers (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
# elif call.data == "add_fms_helper":
# sql = """INSERT INTO helpers (uid, department, last_name, first_name)
# VALUES ({}, 2, '{}', '{}')""".format(uid, last_name, first_name)
# cursor.execute(sql)
# sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
# VALUES ({}, 2, '{}', '{}')""".format(uid, last_name, first_name)
# cursor.execute(sql2)
elif call.data == "add_intern_helper":
try:
sql2 = """INSERT INTO helpers (uid, department, last_name, first_name)
VALUES ({}, 3, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql2)
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
elif call.data == "add_uc_super":
try:
sql = """INSERT INTO supers (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
elif call.data == "add_kb_super":
try:
sql = """INSERT INTO supers (uid, department, last_name, first_name)
VALUES ({}, {}}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
elif call.data == "add_elba_super":
try:
sql = """INSERT INTO supers (uid, department, last_name, first_name)
VALUES ({}, {}, '{}', '{}')""".format(uid, department, cons_last_name, cons_first_name)
cursor.execute(sql)
sql2 = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 1, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql2)
except:
pass
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс' и "\
"Установить помогаторский адрес Врн, ххх опенспейс'.")
# elif call.data == "add_fms_super":
# sql = """INSERT INTO supers (uid, department, last_name, first_name)
# VALUES ({}, 2, '{}', '{}')""".format(uid, last_name, first_name)
elif call.data == "add_oo":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 4, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
bot.send_message(
chat_id=uid,
text="Тебя добавили. Для налача работы напиши боту 'Установить адрес Врн, ххх опенспейс'.")
elif call.data == "add_pip":
sql = """INSERT INTO staff (uid, department, last_name, first_name)
VALUES ({}, 5, '{}', '{}')""".format(uid, cons_last_name, cons_first_name)
cursor.execute(sql)
except Exception as e:
bot.edit_message_text(
chat_id=call.message.chat.id,
message_id=call.message.message_id,
text="Скорее всего пользователь уже добавлен, но вот ошибка\n{}".format(e))
who_sql = "SELECT last_name, first_name FROM supers WHERE uid={}".format(call.from_user.id)
cursor.execute(who_sql)
resp = cursor.fetchall()[0]
responser_full_name = "{} {} (@{})".format(resp[0], resp[1], call.from_user.username)
bot.edit_message_text(
chat_id=call.message.chat.id,
message_id=call.message.message_id,
text="{} успешно добавлен супервизором {}!".format(
cons_full_name,
responser_full_name))
delete_sql = "DELETE FROM tempo WHERE uid={}".format(uid)
cursor.execute(delete_sql)
db.commit()
cursor.close()
db.close()
| [
"noreply@github.com"
] | noreply@github.com |
c9ee8812a25fefe8c101523fb066299ab9ab3dd6 | 92c8737a6c4b967fd43885b97626ef25deb3e983 | /Burgers/burgers.py | f2feb0fa28ccaed55811dd23b18373c044983697 | [] | no_license | JMLipsmeyer/SA-PINNs | 42d53ffc3c3383dd6bbaea113303696e65348ea5 | a180801354e7db02bdda41f9c433008387a1c2f7 | refs/heads/master | 2022-12-27T18:45:35.179715 | 2020-10-12T18:51:08 | 2020-10-12T18:51:08 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,219 | py | import tensorflow as tf
import numpy as np
import matplotlib.pyplot as plt
import time
import scipy.io
import math
import matplotlib.gridspec as gridspec
from plotting import newfig
from mpl_toolkits.axes_grid1 import make_axes_locatable
from tensorflow import keras
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, Input
from tensorflow.keras import layers, activations
from scipy.interpolate import griddata
from eager_lbfgs import lbfgs, Struct
from pyDOE import lhs
layer_sizes = [2, 20, 20, 20, 20, 20, 20, 20, 20, 1]
sizes_w = []
sizes_b = []
for i, width in enumerate(layer_sizes):
if i != 1:
sizes_w.append(int(width * layer_sizes[1]))
sizes_b.append(int(width if i != 0 else layer_sizes[1]))
def set_weights(model, w, sizes_w, sizes_b):
for i, layer in enumerate(model.layers[0:]):
start_weights = sum(sizes_w[:i]) + sum(sizes_b[:i])
end_weights = sum(sizes_w[:i+1]) + sum(sizes_b[:i])
weights = w[start_weights:end_weights]
w_div = int(sizes_w[i] / sizes_b[i])
weights = tf.reshape(weights, [w_div, sizes_b[i]])
biases = w[end_weights:end_weights + sizes_b[i]]
weights_biases = [weights, biases]
layer.set_weights(weights_biases)
def get_weights(model):
w = []
for layer in model.layers[0:]:
weights_biases = layer.get_weights()
weights = weights_biases[0].flatten()
biases = weights_biases[1]
w.extend(weights)
w.extend(biases)
w = tf.convert_to_tensor(w)
return w
def neural_net(layer_sizes):
model = Sequential()
model.add(layers.InputLayer(input_shape=(layer_sizes[0],)))
for width in layer_sizes[1:-1]:
model.add(layers.Dense(
width, activation=tf.nn.tanh,
kernel_initializer="glorot_normal"))
model.add(layers.Dense(
layer_sizes[-1], activation=None,
kernel_initializer="glorot_normal"))
return model
u_model = neural_net(layer_sizes)
u_model.summary()
def loss(x_f_batch, t_f_batch,
x0, t0, u0, x_lb,
t_lb, x_ub, t_ub, col_weights, u_weights):
f_u_pred = f_model(x_f_batch, t_f_batch)
u0_pred = u_model(tf.concat([x0, t0],1))
u_lb_pred, _ = u_x_model(x_lb, t_lb)
u_ub_pred, _ = u_x_model(x_ub, t_ub)
mse_0_u = tf.reduce_mean(tf.square(u_weights*(u0 - u0_pred)))
mse_b_u = tf.reduce_mean(tf.square(u_lb_pred - 0)) + \
tf.reduce_mean(tf.square(u_ub_pred - 0)) #since ub/lb is 0
mse_f_u = tf.reduce_mean(tf.square(col_weights*f_u_pred))
return mse_0_u + mse_b_u + mse_f_u , mse_0_u, mse_f_u
def f_model(x, t):
# keep track of our gradients
with tf.GradientTape(persistent=True) as tape:
tape.watch(x)
tape.watch(t)
u = u_model(tf.concat([x, t],1))
u_x = tape.gradient(u, x)
u_xx = tape.gradient(u_x, x)
u_t = tape.gradient(u, t)
del tape
#hsq = (u**2 + v**2)
f_u = u_t + u*u_x - (0.01/tf.constant(math.pi))*u_xx
#f_v = v_t - .5*u_xx - hsq*u
return f_u
def u_x_model(x, t):
with tf.GradientTape(persistent=True) as tape:
tape.watch(x)
tape.watch(t)
X = tf.concat([x,t],1)
u = u_model(X)
u_x = tape.gradient(u, x)
del tape
return u, u_x
def fit(x_f, t_f, x0, t0, u0, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights, tf_iter, newton_iter):
# Built in support for mini-batch, set to N_f (i.e. full batch) by default
batch_sz = N_f
n_batches = N_f // batch_sz
start_time = time.time()
tf_optimizer = tf.keras.optimizers.Adam(lr = 0.005, beta_1=.90)
tf_optimizer_coll = tf.keras.optimizers.Adam(lr = 0.005, beta_1=.90)
tf_optimizer_u = tf.keras.optimizers.Adam(lr = 0.005, beta_1=.90)
print("starting Adam training")
for epoch in range(tf_iter):
for i in range(n_batches):
x0_batch = x0#[i*batch_sz:(i*batch_sz + batch_sz),]
t0_batch = t0#[i*batch_sz:(i*batch_sz + batch_sz),]
u0_batch = u0#[i*batch_sz:(i*batch_sz + batch_sz),]
x_f_batch = x_f[i*batch_sz:(i*batch_sz + batch_sz),]
t_f_batch = t_f[i*batch_sz:(i*batch_sz + batch_sz),]
with tf.GradientTape(persistent=True) as tape:
loss_value, mse_0, mse_f = loss(x_f_batch, t_f_batch, x0_batch, t0_batch, u0_batch, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights)
grads = tape.gradient(loss_value, u_model.trainable_variables)
grads_col = tape.gradient(loss_value, col_weights)
grads_u = tape.gradient(loss_value, u_weights)
tf_optimizer.apply_gradients(zip(grads, u_model.trainable_variables))
tf_optimizer_coll.apply_gradients(zip([-grads_col], [col_weights]))
tf_optimizer_u.apply_gradients(zip([-grads_u], [u_weights]))
del tape
if epoch % 10 == 0:
elapsed = time.time() - start_time
print('It: %d, Time: %.2f' % (epoch, elapsed))
tf.print(f"mse_0: {mse_0} mse_f: {mse_f} total loss: {loss_value}")
start_time = time.time()
print(col_weights)
#l-bfgs-b optimization
print("Starting L-BFGS training")
loss_and_flat_grad = get_loss_and_flat_grad(x_f_batch, t_f_batch, x0_batch, t0_batch, u0_batch, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights)
lbfgs(loss_and_flat_grad,
get_weights(u_model),
Struct(), maxIter=newton_iter, learningRate=0.8)
# L-BFGS implementation from https://github.com/pierremtb/PINNs-TF2.0
def get_loss_and_flat_grad(x_f_batch, t_f_batch, x0_batch, t0_batch, u0_batch, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights):
def loss_and_flat_grad(w):
with tf.GradientTape() as tape:
set_weights(u_model, w, sizes_w, sizes_b)
loss_value, _, _ = loss(x_f_batch, t_f_batch, x0_batch, t0_batch, u0_batch, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights)
grad = tape.gradient(loss_value, u_model.trainable_variables)
grad_flat = []
for g in grad:
grad_flat.append(tf.reshape(g, [-1]))
grad_flat = tf.concat(grad_flat, 0)
#print(loss_value, grad_flat)
return loss_value, grad_flat
return loss_and_flat_grad
def predict(X_star):
X_star = tf.convert_to_tensor(X_star, dtype=tf.float32)
u_star, _ = u_x_model(X_star[:,0:1],
X_star[:,1:2])
f_u_star = f_model(X_star[:,0:1],
X_star[:,1:2])
return u_star.numpy(), f_u_star.numpy()
lb = np.array([-1.0]) #x upper boundary
ub = np.array([1.0]) #x lower boundary
N0 = 100
N_b = 25 #25 per upper and lower boundary, so 50 total
N_f = 10000
col_weights = tf.Variable(tf.random.uniform([N_f, 1]))
u_weights = tf.Variable(100*tf.random.uniform([N0, 1]))
#load data, from Raissi et. al
data = scipy.io.loadmat('burgers_shock.mat')
t = data['t'].flatten()[:,None]
x = data['x'].flatten()[:,None]
Exact = data['usol']
Exact_u = np.real(Exact)
#grab random points off the initial condition
idx_x = np.random.choice(x.shape[0], N0, replace=False)
x0 = x[idx_x,:]
u0 = Exact_u[idx_x,0:1]
idx_t = np.random.choice(t.shape[0], N_b, replace=False)
tb = t[idx_t,:]
# Sample collocation points via LHS
X_f = lb + (ub-lb)*lhs(2, N_f)
x_f = tf.convert_to_tensor(X_f[:,0:1], dtype=tf.float32)
t_f = tf.convert_to_tensor(np.abs(X_f[:,1:2]), dtype=tf.float32)
#generate point vectors for training
X0 = np.concatenate((x0, 0*x0), 1) # (x0, 0)
X_lb = np.concatenate((0*tb + lb[0], tb), 1) # (lb[0], tb)
X_ub = np.concatenate((0*tb + ub[0], tb), 1) # (ub[0], tb)
#seperate point vectors
x0 = X0[:,0:1]
t0 = X0[:,1:2]
x_lb = tf.convert_to_tensor(X_lb[:,0:1], dtype=tf.float32)
t_lb = tf.convert_to_tensor(X_lb[:,1:2], dtype=tf.float32)
x_ub = tf.convert_to_tensor(X_ub[:,0:1], dtype=tf.float32)
t_ub = tf.convert_to_tensor(X_ub[:,1:2], dtype=tf.float32)
# Begin training, modify 10000/10000 for varying levels of adam/L-BFGS respectively
fit(x_f, t_f, x0, t0, u0, x_lb, t_lb, x_ub, t_ub, col_weights, u_weights, tf_iter = 10000, newton_iter = 10000)
#generate mesh to find U0-pred for the whole domain
X, T = np.meshgrid(x,t)
X_star = np.hstack((X.flatten()[:,None], T.flatten()[:,None]))
u_star = Exact_u.T.flatten()[:,None]
lb = np.array([-1.0, 0.0])
ub = np.array([1.0, 1])
# Get preds
u_pred, f_u_pred = predict(X_star)
#find L2 error
error_u = np.linalg.norm(u_star-u_pred,2)/np.linalg.norm(u_star,2)
print('Error u: %e' % (error_u))
U_pred = griddata(X_star, u_pred.flatten(), (X, T), method='cubic')
FU_pred = griddata(X_star, f_u_pred.flatten(), (X, T), method='cubic')
#plotting script in the style of Raissi et al
######################################################################
############################# Plotting ###############################
######################################################################
X0 = np.concatenate((x0, 0*x0), 1) # (x0, 0)
X_lb = np.concatenate((0*tb + lb[0], tb), 1) # (lb[0], tb)
X_ub = np.concatenate((0*tb + ub[0], tb), 1) # (ub[0], tb)
X_u_train = np.vstack([X0, X_lb, X_ub])
fig, ax = newfig(1.3, 1.0)
ax.axis('off')
####### Row 0: h(t,x) ##################
gs0 = gridspec.GridSpec(1, 2)
gs0.update(top=1-0.06, bottom=1-1/3, left=0.15, right=0.85, wspace=0)
ax = plt.subplot(gs0[:, :])
h = ax.imshow(U_pred.T, interpolation='nearest', cmap='YlGnBu',
extent=[lb[1], ub[1], lb[0], ub[0]],
origin='lower', aspect='auto')
divider = make_axes_locatable(ax)
cax = divider.append_axes("right", size="5%", pad=0.05)
fig.colorbar(h, cax=cax)
line = np.linspace(x.min(), x.max(), 2)[:,None]
ax.plot(t[25]*np.ones((2,1)), line, 'k--', linewidth = 1)
ax.plot(t[50]*np.ones((2,1)), line, 'k--', linewidth = 1)
ax.plot(t[75]*np.ones((2,1)), line, 'k--', linewidth = 1)
ax.set_xlabel('$t$')
ax.set_ylabel('$x$')
leg = ax.legend(frameon=False, loc = 'best')
# plt.setp(leg.get_texts(), color='w')
ax.set_title('$u(t,x)$', fontsize = 10)
####### Row 1: h(t,x) slices ##################
gs1 = gridspec.GridSpec(1, 3)
gs1.update(top=1-1/3, bottom=0, left=0.1, right=0.9, wspace=0.5)
ax = plt.subplot(gs1[0, 0])
ax.plot(x,Exact_u[:,25], 'b-', linewidth = 2, label = 'Exact')
ax.plot(x,U_pred[25,:], 'r--', linewidth = 2, label = 'Prediction')
ax.set_xlabel('$x$')
ax.set_ylabel('$u(t,x)$')
ax.set_title('$t = %.2f$' % (t[25]), fontsize = 10)
ax.axis('square')
ax.set_xlim([-1.1,1.1])
ax.set_ylim([-1.1,1.1])
ax = plt.subplot(gs1[0, 1])
ax.plot(x,Exact_u[:,50], 'b-', linewidth = 2, label = 'Exact')
ax.plot(x,U_pred[50,:], 'r--', linewidth = 2, label = 'Prediction')
ax.set_xlabel('$x$')
ax.set_ylabel('$u(t,x)$')
ax.axis('square')
ax.set_xlim([-1.1,1.1])
ax.set_ylim([-1.1,1.1])
ax.set_title('$t = %.2f$' % (t[50]), fontsize = 10)
ax.legend(loc='upper center', bbox_to_anchor=(0.5, -0.3), ncol=5, frameon=False)
ax = plt.subplot(gs1[0, 2])
ax.plot(x,Exact_u[:,75], 'b-', linewidth = 2, label = 'Exact')
ax.plot(x,U_pred[75,:], 'r--', linewidth = 2, label = 'Prediction')
ax.set_xlabel('$x$')
ax.set_ylabel('$u(t,x)$')
ax.axis('square')
ax.set_xlim([-1.1,1.1])
ax.set_ylim([-1.1,1.1])
ax.set_title('$t = %.2f$' % (t[75]), fontsize = 10)
#show u_pred across domain
fig, ax = plt.subplots()
ec = plt.imshow(U_pred.T, interpolation='nearest', cmap='rainbow',
extent=[0.0, 1.0, -1.0, 1.0],
origin='lower', aspect='auto')
ax.autoscale_view()
ax.set_xlabel('$t$')
ax.set_ylabel('$x$')
cbar = plt.colorbar(ec)
cbar.set_label('$u(x,t)$')
plt.title("Predicted $u(x,t)$",fontdict = {'fontsize': 14})
plt.show()
# Show F_U_pred across domain, should be close to 0
fig, ax = plt.subplots()
ec = plt.imshow(FU_pred.T, interpolation='nearest', cmap='rainbow',
extent=[0.0, math.pi/2, -5.0, 5.0],
origin='lower', aspect='auto')
ax.autoscale_view()
ax.set_xlabel('$x$')
ax.set_ylabel('$t$')
cbar = plt.colorbar(ec)
cbar.set_label('$\overline{f}_u$ prediction')
plt.show()
# collocation point weights
plt.scatter(t_f, x_f, c = col_weights.numpy(), s = col_weights.numpy()/5)
plt.show()
| [
"levimcclenny@tamu.edu"
] | levimcclenny@tamu.edu |
c9a91552c1b8f4b8a2ff609676b81cd11cf08ead | 48df99f4358be7a51becd3d685e1ec825d295ba4 | /dentalstate/models.py | 36c642462ac4cabb367d2fe592fdd0be94d557a6 | [
"Apache-2.0"
] | permissive | kuyesu/tscharts | 21d2aedeea4aad3b126defaa1703f60f44f14de6 | 9ed4e4bb0a6d296e1156afca5b55d0f71dfb894b | refs/heads/master | 2023-06-03T04:50:15.282855 | 2021-06-12T19:50:51 | 2021-06-12T19:50:51 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,258 | py | #(C) Copyright Syd Logan 2020
#(C) Copyright Thousand Smiles Foundation 2020
#
#Licensed under the Apache License, Version 2.0 (the "License");
#you may not use this file except in compliance with the License.
#
#You may obtain a copy of the License at
#http://www.apache.org/licenses/LICENSE-2.0
#
#Unless required by applicable law or agreed to in writing, software
#distributed under the License is distributed on an "AS IS" BASIS,
#WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
#See the License for the specific language governing permissions and
#limitations under the License.
from __future__ import unicode_literals
from django.db import models
from patient.models import Patient
from clinic.models import Clinic
from dentalcdt.models import DentalCDT
class DentalState(models.Model):
clinic = models.ForeignKey(Clinic)
patient = models.ForeignKey(Patient)
username = models.CharField(max_length=64, default = "") # user supplied name
time = models.DateTimeField(auto_now=True)
'''
tooth location is relative to location (top or bottom). Zero
indicates the treatment applies to whole mouth (and location
is ignored
'''
tooth = models.IntegerField(default = 0)
DENTAL_LOCATION_TOP = 't'
DENTAL_LOCATION_BOTTOM = 'b'
DENTAL_LOCATION_CHOICES = ((DENTAL_LOCATION_TOP, "top"), (DENTAL_LOCATION_BOTTOM, "bottom"))
location = models.CharField(max_length = 1, choices = DENTAL_LOCATION_CHOICES, default = DENTAL_LOCATION_TOP)
code = models.ForeignKey(DentalCDT)
DENTAL_STATE_NONE = 'n'
DENTAL_STATE_UNTREATED = 'u'
DENTAL_STATE_TREATED = 't'
DENTAL_STATE_OTHER = 'o'
DENTAL_STATE_MISSING = 'm'
DENTAL_STATE_CHOICES = ((DENTAL_STATE_MISSING, "missing"), (DENTAL_STATE_NONE, "none"), (DENTAL_STATE_UNTREATED, "untreated"), (DENTAL_STATE_TREATED, "treated"), (DENTAL_STATE_OTHER, "other"))
state = models.CharField(max_length = 1, choices = DENTAL_STATE_CHOICES, default = DENTAL_STATE_NONE)
DENTAL_SURFACE_NONE = 'n'
DENTAL_SURFACE_BUCCAL = 'b'
DENTAL_SURFACE_LINGUAL = 'u'
DENTAL_SURFACE_MESIAL = 'm'
DENTAL_SURFACE_OCCLUSAL = 'c'
DENTAL_SURFACE_LABIAL = 'a'
DENTAL_SURFACE_INCISAL = 'i'
DENTAL_SURFACE_WHOLE_MOUTH_OR_VISIT = 'w'
DENTAL_SURFACE_OTHER = 'o'
DENTAL_SURFACE_CHOICES = ((DENTAL_SURFACE_NONE, "none"), (DENTAL_SURFACE_BUCCAL, "buccal"), (DENTAL_SURFACE_LINGUAL, "lingual"), (DENTAL_SURFACE_MESIAL, "mesial"), (DENTAL_SURFACE_OCCLUSAL, 'occlusal'), (DENTAL_SURFACE_LABIAL, 'labial'), (DENTAL_SURFACE_INCISAL, 'incisal'), (DENTAL_SURFACE_WHOLE_MOUTH_OR_VISIT, 'whole_mouth_or_visit'), (DENTAL_SURFACE_OTHER, 'other'))
# here we define a charfield as a string to hold a set of surfaces
# this won't work with forms, but since we are just a REST API, doesn't
# matter much. The DENTAL_STATE_CHOICES tuple will be useful as we
# serialize/unserialize values between the client and the model. We
# could also have done this as an integer bitmask, but a string of chars
# facilitates debugging.
surface = models.CharField(max_length = 10, choices = DENTAL_SURFACE_CHOICES, default = DENTAL_SURFACE_NONE)
comment = models.TextField(default = "")
| [
"slogan621@gmail.com"
] | slogan621@gmail.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.