hexsha
stringlengths
40
40
size
int64
5
2.06M
ext
stringclasses
10 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
248
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
listlengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
248
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
listlengths
1
10
max_issues_count
int64
1
67k
max_issues_repo_issues_event_min_datetime
stringdate
2015-01-01 00:00:47
2022-03-31 23:42:18
max_issues_repo_issues_event_max_datetime
stringdate
2015-01-01 17:43:30
2022-03-31 23:59:58
max_forks_repo_path
stringlengths
3
248
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
listlengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
2.06M
avg_line_length
float64
1
1.02M
max_line_length
int64
3
1.03M
alphanum_fraction
float64
0
1
a01762ca3e759a9a379ad71578ccb40a1edcad3d
738
py
Python
contests_atcoder/abc153/abc153_f.py
takelifetime/competitive-programming
e7cf8ef923ccefad39a1727ca94c610d650fcb76
[ "BSD-2-Clause" ]
null
null
null
contests_atcoder/abc153/abc153_f.py
takelifetime/competitive-programming
e7cf8ef923ccefad39a1727ca94c610d650fcb76
[ "BSD-2-Clause" ]
1
2021-01-02T06:36:51.000Z
2021-01-02T06:36:51.000Z
contests_atcoder/abc153/abc153_f.py
takelifetime/competitive-programming
e7cf8ef923ccefad39a1727ca94c610d650fcb76
[ "BSD-2-Clause" ]
null
null
null
from bisect import bisect_left, bisect_right from collections import deque, Counter from itertools import combinations, permutations from math import gcd, sin, cos, tan, degrees, radians import sys input = lambda: sys.stdin.readline().rstrip() MOD = 10 ** 9 + 7 INF = float("inf") n, d, a = map(int, input().split()) monsters = [tuple(map(int, input().split())) for _ in range(n)] monsters.sort() now = 0 ans = 0 bomb = deque() for m in monsters: x = m[0] attack_count = -(-m[1] // a) while len(bomb) and bomb[0][0] < x: b = bomb.popleft() now -= b[1] if attack_count > now: ans += attack_count - now bomb.append((x + 2 * d, attack_count - now)) now = attack_count print(ans)
23.0625
63
0.624661
a017a1ab05231fbc634e10328c46e53e752448d8
16,532
py
Python
sistema_experto.py
Erubeyy/SistemaExperto-
6194f798fad684eb83635fe85bf3f1a7d70ed2a2
[ "MIT" ]
null
null
null
sistema_experto.py
Erubeyy/SistemaExperto-
6194f798fad684eb83635fe85bf3f1a7d70ed2a2
[ "MIT" ]
null
null
null
sistema_experto.py
Erubeyy/SistemaExperto-
6194f798fad684eb83635fe85bf3f1a7d70ed2a2
[ "MIT" ]
null
null
null
from tkinter import* from tkinter import font from experta import * raiz = Tk() raiz.title("Sistema experto- Tipos de covid") raiz.config(bg="#f4f7fa") #raiz.resizable(0,0) mi0Frame = Frame(raiz)#, width="1200", height="700") mi0Frame.grid(row=1, column=0) mi0Frame.config(bg="#f4f7fa") mi3Frame = Frame(raiz)#, width="1200", height="700") mi3Frame.grid(row=1, column=1) mi3Frame.config(bg="#f4f7fa") miFrame = Frame(raiz)#, width="1200", height="700") miFrame.grid(row=2, column=0) miFrame.config(bg="#f4f7fa") mi2Frame = Frame(raiz, highlightbackground="black", highlightthickness=0.5) mi2Frame.grid(row=2, column=1) mi2Frame.config(bg="#f4f7fa") mi4Frame = Frame(raiz, highlightbackground="black", highlightthickness=0.5) mi4Frame.grid(row=0, column=0) mi4Frame.config(bg="#f4f7fa") reinicio = 0 #-----------------------------------------------INPUTS DE LOS SÍNTOMAS------------------------------------------------------------ sin0 = Label(miFrame, text="Dolor de cabeza:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin0.grid(row=0, column=0,padx=10, pady=10,sticky="e") in_sin0 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin0.grid(row=0, column=1,padx=10, pady=10) sin1 = Label(miFrame, text="Perdida del olfato:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin1.grid(row=1, column=0,padx=10, pady=10,sticky="e") in_sin1 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin1.grid(row=1, column=1,padx=10, pady=10) sin2 = Label(miFrame, text="Dolor muscular:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin2.grid(row=2, column=0,padx=10, pady=10,sticky="e") in_sin2 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin2.grid(row=2, column=1,padx=10, pady=10) sin3 = Label(miFrame, text="Tos:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin3.grid(row=3, column=0,padx=10, pady=10,sticky="e") in_sin3 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin3.grid(row=3, column=1,padx=10, pady=10) sin4 = Label(miFrame, text="Dolor de garganta:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin4.grid(row=4, column=0,padx=10, pady=10,sticky="e") in_sin4 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin4.grid(row=4, column=1,padx=10, pady=10) sin5 = Label(miFrame, text="Dolor en el pecho:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin5.grid(row=5, column=0,padx=10, pady=10,sticky="e") in_sin5 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin5.grid(row=5, column=1,padx=10, pady=10) sin6 = Label(miFrame, text="Fiebre:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin6.grid(row=6, column=0,padx=10, pady=10,sticky="e") in_sin6 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin6.grid(row=6, column=1,padx=10, pady=10) sin7 = Label(miFrame, text="Ronquera:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin7.grid(row=7, column=0,padx=10, pady=10,sticky="e") in_sin7 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin7.grid(row=7, column=1,padx=10, pady=10) sin8 = Label(miFrame, text="Pérdida del apetito:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin8.grid(row=8, column=0,padx=10, pady=10,sticky="e") in_sin8 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin8.grid(row=8, column=1,padx=10, pady=10) sin9 = Label(miFrame, text="Diarrea:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin9.grid(row=9, column=0,padx=10, pady=10,sticky="e") in_sin9 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin9.grid(row=9, column=1,padx=10, pady=10) sin10 = Label(miFrame, text="Fatiga:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin10.grid(row=10, column=0,padx=10, pady=10,sticky="e") in_sin10 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin10.grid(row=10, column=1,padx=10, pady=10) sin11 = Label(miFrame, text="Confusión:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin11.grid(row=11, column=0,padx=10, pady=10,sticky="e") in_sin11 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin11.grid(row=11, column=1,padx=10, pady=10) sin12 = Label(miFrame, text="Dificultad para respirar:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sin12.grid(row=12, column=0,padx=10, pady=10,sticky="e") in_sin12 = Entry(miFrame, width=10, font=('CASTELLAR', 9, font.BOLD), justify='center') in_sin12.grid(row=12, column=1,padx=10, pady=10) #------Cuadros de los resultados-------- tipo_final_lbl = Label(mi2Frame, text="Tipo de covid diagnosticado:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) tipo_final_lbl.grid(row=2, column=0,padx=10, pady=10,sticky="n") tipo_final = Entry(mi2Frame, width=35, justify='center', font=('FELIX TITLING', 10, font.BOLD)) tipo_final.grid(row=3, column=0, padx=1, pady=1) blank = Label(mi2Frame, bg="#F0F8FF") blank.grid(row=4, column=0,padx=10, pady=10,sticky="n") descripcion_tipo_lbl = Label(mi2Frame, text="Descripción del tipo de covid diagnosticado:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) descripcion_tipo_lbl.grid(row=5, column=0,padx=10, pady=10,sticky="n") descripcion_tipo = Text(mi2Frame, width=60, height=10) descripcion_tipo.grid(row=6, column=0, padx=10, pady=10) sugerencias_lbl = Label(mi2Frame, text="Sugerencias para tratar la enfermedad:", bg="#F0F8FF", font=('Century Ghotic', 10, font.BOLD)) sugerencias_lbl.grid(row=7, column=0,padx=10, pady=10,sticky="n") sugerencias = Text(mi2Frame, width=60, height=10) sugerencias.grid(row=8, column=0, padx=10, pady=10) #------HEADER-------- head1 = Label(mi0Frame, text="\nSÍNTOMAS", bg="#F0F8FF", font=('Elephant', 15)) head1.grid(row=0, column=0, sticky="n") head1_0 = Label(mi3Frame, text="DIAGNÓSTICO", bg="#F0F8FF", font=('Elephant', 15)) head1_0.grid(row=0, column=0, sticky="n") head1 = Label(mi0Frame, bg="#F0F8FF") head1.grid(row=1, column=0, sticky="n") head2 = Label(mi0Frame, text=" -Introduce un 'si' o un 'no' dependiendo de los síntomas que presentes", bg="#F0F8FF", font=('Century Ghotic', 11)) head2.grid(row=2, column=0, sticky="n" ) head3 = Label(mi4Frame, text="Sistema experto - Tipos de COVID", bg="#F0F8FF", font=('Elephant', 15)) head3.grid(row=0) #-----------------------------------------^^^^^^INPUTS DE LOS SÍNTOMAS^^^^^^------------------------------------------------------ lista_tipos = [] sintomas_tipo = [] map_sintomas = {} d_desc_map = {} d_tratamiento_map = {} def preprocess(): global lista_tipos,sintomas_tipo,map_sintomas,d_desc_map,d_tratamiento_map tipos = open("tipos.txt") tipos_t = tipos.read() lista_tipos = tipos_t.split("\n") tipos.close() for tipo in lista_tipos: tipo_s_file = open("Sintomas tipo/" + tipo + ".txt") tipo_s_data = tipo_s_file.read() s_list = tipo_s_data.split("\n") sintomas_tipo.append(s_list) map_sintomas[str(s_list)] = tipo tipo_s_file.close() tipo_s_file = open("Descripcion tipo/" + tipo + ".txt") tipo_s_data = tipo_s_file.read() d_desc_map[tipo] = tipo_s_data tipo_s_file.close() tipo_s_file = open("Tratamientos tipo/" + tipo + ".txt") tipo_s_data = tipo_s_file.read() d_tratamiento_map[tipo] = tipo_s_data tipo_s_file.close() def identificar_tipo(*arguments): lista_sintomas = [] for sintoma in arguments: lista_sintomas.append(sintoma) # Handle key error return map_sintomas[str(lista_sintomas)] def get_details(tipo): return d_desc_map[tipo] def get_tratamiento(tipo): return d_tratamiento_map[tipo] def no_coincide(tipo): tipo_final.delete("1.0", END) descripcion_tipo.delete("1.0", END) sugerencias.delete("1.0", END) id_tipo = tipo tipo_details = get_details(id_tipo) tratamientos = get_tratamiento(id_tipo) tipo_final.insert("1.0", id_tipo) descripcion_tipo.insert("1.0", tipo_details) sugerencias.insert("1.0", tratamientos) #def identificar_tipo(dolor_cabeza, perdida_olfato, dolor_muscular, tos, dolor_garganta, dolor_pecho, fiebre, ronquera, perdida_apetito , diarrea, fatiga, confusión, dificultad_respiratoria): class Covid(KnowledgeEngine): @DefFacts() def _initial_action(self): yield Fact(action="encontrar_tipo") @Rule(Fact(action='encontrar_tipo'), NOT(Fact(dolor_cabeza=W())),salience = 1) def sintoma_0(self): self.declare(Fact(dolor_cabeza=in_sin0.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(perdida_olfato=W())),salience = 1) def sintoma_1(self): self.declare(Fact(perdida_olfato=in_sin1.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(dolor_muscular=W())),salience = 1) def sintoma_2(self): self.declare(Fact(dolor_muscular=in_sin2.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(tos=W())),salience = 1) def sintoma_3(self): self.declare(Fact(tos=in_sin3.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(dolor_garganta=W())),salience = 1) def sintoma_4(self): self.declare(Fact(dolor_garganta=in_sin4.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(dolor_pecho=W())),salience = 1) def sintoma_5(self): self.declare(Fact(dolor_pecho=in_sin5.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(fiebre=W())),salience = 1) def sintoma_6(self): self.declare(Fact(fiebre=in_sin6.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(ronquera=W())),salience = 1) def sintoma_7(self): self.declare(Fact(ronquera=in_sin7.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(perdida_apetito=W())),salience = 1) def sintoma_8(self): self.declare(Fact(perdida_apetito=in_sin8.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(diarrea=W())),salience = 1) def sintoma_9(self): self.declare(Fact(diarrea=in_sin9.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(fatiga=W())),salience = 1) def sintoma_10(self): self.declare(Fact(fatiga=in_sin10.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(confusion=W())),salience = 1) def sintoma_11(self): self.declare(Fact(confusion=in_sin11.get())) @Rule(Fact(action='encontrar_tipo'), NOT(Fact(dificultad_respiratoria=W())),salience = 1) def sintoma_12(self): self.declare(Fact(dificultad_respiratoria=in_sin12.get())) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="si"),Fact(tos="si"),Fact(dolor_garganta="si"),Fact(dolor_pecho="si"),Fact(fiebre="no"),Fact(ronquera="no"),Fact(perdida_apetito="no"),Fact(diarrea="no"),Fact(fatiga="no"),Fact(confusion="no"),Fact(dificultad_respiratoria="no")) def tipo_0(self): self.declare(Fact(tipo="Gripal sin fiebre")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="no"),Fact(tos="si"),Fact(dolor_garganta="si"),Fact(dolor_pecho="no"),Fact(fiebre="si"),Fact(ronquera="si"),Fact(perdida_apetito="si"),Fact(diarrea="no"),Fact(fatiga="no"),Fact(confusion="no"),Fact(dificultad_respiratoria="no")) def tipo_1(self): self.declare(Fact(tipo="Gripal con fiebre")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="no"),Fact(tos="no"),Fact(dolor_garganta="si"),Fact(dolor_pecho="si"),Fact(fiebre="no"),Fact(ronquera="no"),Fact(perdida_apetito="no"),Fact(diarrea="si"),Fact(fatiga="no"),Fact(confusion="no"),Fact(dificultad_respiratoria="no")) def tipo_2(self): self.declare(Fact(tipo="Gastro Intestinal")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="no"),Fact(tos="si"),Fact(dolor_garganta="no"),Fact(dolor_pecho="si"),Fact(fiebre="si"),Fact(ronquera="si"),Fact(perdida_apetito="no"),Fact(diarrea="no"),Fact(fatiga="si"),Fact(confusion="no"),Fact(dificultad_respiratoria="no")) def tipo_3(self): self.declare(Fact(tipo="Nivel severo uno")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="si"),Fact(tos="si"),Fact(dolor_garganta="si"),Fact(dolor_pecho="si"),Fact(fiebre="si"),Fact(ronquera="si"),Fact(perdida_apetito="si"),Fact(diarrea="no"),Fact(fatiga="si"),Fact(confusion="si"),Fact(dificultad_respiratoria="no")) def tipo_4(self): self.declare(Fact(tipo="Nivel severo dos")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="si"),Fact(perdida_olfato="si"),Fact(dolor_muscular="si"),Fact(tos="si"),Fact(dolor_garganta="si"),Fact(dolor_pecho="si"),Fact(fiebre="si"),Fact(ronquera="si"),Fact(perdida_apetito="si"),Fact(diarrea="si"),Fact(fatiga="si"),Fact(confusion="si"),Fact(dificultad_respiratoria="si")) def tipo_5(self): self.declare(Fact(tipo="Nivel severo tres")) @Rule(Fact(action='encontrar_tipo'),Fact(dolor_cabeza="no"),Fact(perdida_olfato="no"),Fact(dolor_muscular="no"),Fact(tos="no"),Fact(dolor_garganta="no"),Fact(dolor_pecho="no"),Fact(fiebre="no"),Fact(ronquera="no"),Fact(perdida_apetito="no"),Fact(diarrea="no"),Fact(fatiga="no"),Fact(confusion="no"),Fact(dificultad_respiratoria="no")) def tipo_6(self): self.declare(Fact(tipo="No es covid")) @Rule(Fact(action='encontrar_tipo'),Fact(tipo=MATCH.tipo),salience = -998) def tipo(self, tipo): tipo_final.delete("0", END) descripcion_tipo.delete("1.0", END) sugerencias.delete("1.0", END) id_tipo = tipo tipo_details = get_details(id_tipo) tratamientos = get_tratamiento(id_tipo) tipo_final.insert("0", id_tipo) descripcion_tipo.insert("1.0", tipo_details) sugerencias.insert("1.0",tratamientos) @Rule(Fact(action='encontrar_tipo'), Fact(dolor_cabeza=MATCH.dolor_cabeza), Fact(perdida_olfato=MATCH.perdida_olfato), Fact(dolor_muscular=MATCH.dolor_muscular), Fact(tos=MATCH.tos), Fact(dolor_garganta=MATCH.dolor_garganta), Fact(dolor_pecho=MATCH.dolor_pecho), Fact(fiebre=MATCH.fiebre), Fact(ronquera=MATCH.ronquera), Fact(perdida_apetito=MATCH.perdida_apetito), Fact(diarrea=MATCH.diarrea), Fact(fatiga=MATCH.fatiga), Fact(confusion=MATCH.confusion), Fact(dificultad_respiratoria=MATCH.dificultad_respiratoria),NOT(Fact(tipo=MATCH.tipo)),salience = -999) def not_matched(self,dolor_cabeza, perdida_olfato, dolor_muscular, tos, dolor_garganta, dolor_pecho, fiebre, ronquera,perdida_apetito ,diarrea ,fatiga ,confusion ,dificultad_respiratoria): global reinicio if reinicio == 0: tipo_final.delete("0", END) descripcion_tipo.delete("1.0", END) sugerencias.delete("1.0", END) tipo_final.insert("0", "Sin coincidencia") descripcion_tipo.insert("1.0", "No se encontró un tipo de covid que se relacione con los síntomas presentados") sugerencias.insert("1.0", "Se sugiere consultar a un médico que le ayude a descubrir su tipo de enfermedad") else: reinicio = 0 def iniciar_sistema(): if __name__ == "__main__": preprocess() engine = Covid() engine.reset() engine.run() def reiniciar(): global reinicio reinicio = 1 in_sin0.delete("0", END) in_sin1.delete("0", END) in_sin2.delete("0", END) in_sin3.delete("0", END) in_sin4.delete("0", END) in_sin5.delete("0", END) in_sin6.delete("0", END) in_sin7.delete("0", END) in_sin8.delete("0", END) in_sin9.delete("0", END) in_sin10.delete("0", END) in_sin11.delete("0", END) in_sin12.delete("0", END) tipo_final.delete("0", END) descripcion_tipo.delete('1.0', END) sugerencias.delete('1.0', END) preprocess() engine = Covid() engine.reset() engine.run() def salir(): exit() #------------------BOTONES--------------------------------------- generarTabla = Button( miFrame, text="RESULTADO", command=iniciar_sistema, bg="#7fd1ff", font=("Eurostile", 10, font.BOLD), padx=20, pady=5 ) generarTabla.grid(row=13, column=1, padx=10, pady=15) reiniciar = Button( mi2Frame, text="REINICIAR", command=reiniciar, bg="#7fd1ff", font=("Eurostile", 10, font.BOLD), padx=20, pady=5 ) reiniciar.grid(row=9, column=0, padx=10, pady=15) salir = Button( mi2Frame, text="SALIR", command=salir, bg="#ea9999", font=("Eurostile", 9), border='2p', padx=20, pady=3 ) salir.grid(row=10, column=0, padx=10, pady=15) raiz.mainloop()
43.851459
335
0.687576
a017aa5f81d90682eeec3d31e4bdb2e999666f4b
6,105
py
Python
socket_temperature_connect.py
MeowMeowZi/PPLTestTool
576f28fb20680b1ed33520d92c552ccafc93d716
[ "MIT" ]
null
null
null
socket_temperature_connect.py
MeowMeowZi/PPLTestTool
576f28fb20680b1ed33520d92c552ccafc93d716
[ "MIT" ]
null
null
null
socket_temperature_connect.py
MeowMeowZi/PPLTestTool
576f28fb20680b1ed33520d92c552ccafc93d716
[ "MIT" ]
null
null
null
import socket import time import shelve preset_command = { 1: ['MB0023,1', 'MI0695,'], 2: ['MB0024,1', 'MI0696,'], 3: ['MB0076,1', 'MI0697,'], 4: ['MB0026,1', 'MI0698,'], } force_command = 'MB0336,1' start_command = 'MB0020,0' stop_command = 'MB0020,1' class Temperature: def __init__(self): # 是否打印log信息 self.is_info = False # 打印log信息 self.info = '' # temp测试任务 self.task = [] # 打开配置文件 self.init_temp = shelve.open('init/init_temp') self.ip = self.init_temp['temp_ip'] self.channel1_temp = self.init_temp['temp_channel1_temp'] self.channel2_temp = self.init_temp['temp_channel2_temp'] self.channel3_temp = self.init_temp['temp_channel3_temp'] self.channel4_temp = self.init_temp['temp_channel4_temp'] self.is_channel1_temp = self.init_temp['temp_is_channel1_temp'] self.is_channel2_temp = self.init_temp['temp_is_channel2_temp'] self.is_channel3_temp = self.init_temp['temp_is_channel3_temp'] self.is_channel4_temp = self.init_temp['temp_is_channel4_temp'] # 关闭配置文件 self.init_temp.close() self.channel1 = (self.channel1_temp, 1) self.channel2 = (self.channel2_temp, 2) self.channel3 = (self.channel3_temp, 3) self.channel4 = (self.channel4_temp, 4) # 创造套接字 self.server = socket.socket() # self.ip = '192.168.0.14' self.port = 5000 try: self.server.connect((self.ip, self.port)) # print('[INFO-TEMP]connect successfully') self.send_info('[INFO-TEMP]connect successfully') time.sleep(1) except: # print('[FAIL-TEMP]connect fail') self.send_info('[FAIL-TEMP]connect fail') # 向设备发送数据 def send(self, data): try: self.server.send(bytes(data, encoding='ASCII')) except ConnectionError: # print('[FAIL-TEMP]send data fail') self.send_info('[FAIL-TEMP]send data fail') # 向设备接受数据 def recv(self): try: text = str(self.server.recv(1024), encoding='UTF-8') # print(text) except ConnectionError: # print('[FAIL-TEMP]receive error') self.send_info('[FAIL-TEMP]receive error') text = ',9990' return text # 指令 (发送指令) def command(self, command): self.send('m') time.sleep(1) self.send(command) time.sleep(1) # 写入指令 (无返回值) def write_command(self, command): self.command(command) self.ack() # 询问指令 (有返回值) def query_command(self, command): self.command(command) return self.recv() # 设备应答 def ack(self): while True: if self.recv() == 'OK': break # 温度预设 (四个通道) def preset(self, channel): temp = int(channel[0]) temp_command = '' if temp == 0: temp_command = '0000' elif (temp > 0) and (temp < 10): temp_command = '00' + str(temp) + '0' elif (temp > 9) and (temp < 100): temp_command = '0' + str(temp) + '0' elif temp > 99: temp_command = str(temp) + '0' elif (temp < 0) and (temp > -10): temp_command = '0' + str(temp) + '0' elif temp < -9: temp_command = '' + str(temp) + '0' elif temp >= 175: temp_command = '1750' elif temp <= -75: temp_command = '-750' channel_command = preset_command[channel[1]][1] command = channel_command + temp_command self.write_command(command) # print('[INFO-TEMP]channel%s, %s℃ set successfully!' % (channel[1], channel[0])) self.send_info('[INFO-TEMP]channel' + str(channel[1]) + ', ' + str(channel[0]) +'℃ set successfully!') # 选择温度预设为当前值 def change_channel(self, channel): state_command = preset_command[channel[1]][0] self.write_command(state_command) # print('[INFO-TEMP]change channel:', channel[1]) self.send_info('[INFO-TEMP]change channel ' + str(channel[1]) + " " + str(channel[0]) + '℃') # 将测试项添加到任务列表中 def task_generate(self): if self.is_channel1_temp: self.preset((self.channel1_temp, 1)) self.task.append(self.channel1) if self.is_channel2_temp: self.preset((self.channel2_temp, 2)) self.task.append(self.channel2) if self.is_channel3_temp: self.preset((self.channel3_temp, 3)) self.task.append(self.channel3) if self.is_channel4_temp: self.preset((self.channel4_temp, 4)) self.task.append(self.channel4) self.write_command(force_command) # print('[INFO-TEMP]force on') self.send_info('[INFO-TEMP]force on') # 检查设备温度 (1秒询问一次) def check_temp(self, channel): while True: for i in range(3): text = self.query_command('MI0006?') # 获取格式为 MI6,250 temp1 = int(text.split(',')[1]) # 250 整数位+小数位 # print('[INFO-TEMP]temp: ', temp1 / 10.0, '℃') self.send_info('[INFO-TEMP]temp: ' + str(temp1 / 10.0) + '℃') temp = int(channel[0]) if (temp1 == temp * 10) and (i == 2): return elif temp1 == temp * 10: pass else: break # 启动设备 def start(self): self.write_command(start_command) # print('[INFO-TEMP]running!') self.send_info('[INFO-TEMP]running!') # 关闭设备 def stop(self): self.write_command(stop_command) # print('[INFO-TEMP]close!') self.send_info('[INFO-TEMP]close!') # 用于切换不同task def run(self, task): self.change_channel(task) self.check_temp(task) time.sleep(1) # 向主线程发送数据 def send_info(self, info): self.info = info self.is_info = True if __name__ == '__main__': temperature = Temperature()
30.678392
110
0.552826
a0187e302825ea7cb1c14461fb74435494c1cd4b
12,938
py
Python
wwwdccn/chair_mail/models.py
marvinxu99/dccnsys
8f53728d06b859cace42cc84bc190bc89950d252
[ "MIT" ]
16
2020-03-15T15:33:30.000Z
2021-11-26T21:57:27.000Z
wwwdccn/chair_mail/models.py
marvinxu99/dccnsys
8f53728d06b859cace42cc84bc190bc89950d252
[ "MIT" ]
11
2019-04-27T19:15:43.000Z
2022-03-11T23:43:08.000Z
wwwdccn/chair_mail/models.py
marvinxu99/dccnsys
8f53728d06b859cace42cc84bc190bc89950d252
[ "MIT" ]
10
2020-03-14T09:25:39.000Z
2022-02-21T16:46:33.000Z
from django.conf import settings from django.core.mail import send_mail from django.db import models from django.db.models import ForeignKey, OneToOneField, TextField, CharField, \ SET_NULL, CASCADE, BooleanField, UniqueConstraint from django.db.models.signals import post_save from django.dispatch import receiver from django.template import Template, Context from django.utils import timezone from markdown import markdown from html2text import html2text from chair_mail.context import get_conference_context, get_user_context, \ get_submission_context, get_frame_context from conferences.models import Conference from submissions.models import Submission from users.models import User MSG_TYPE_USER = 'user' MSG_TYPE_SUBMISSION = 'submission' MESSAGE_TYPE_CHOICES = ( (MSG_TYPE_USER, 'Message to users'), (MSG_TYPE_SUBMISSION, 'Message to submissions'), ) class EmailFrame(models.Model): text_html = models.TextField() text_plain = models.TextField() created_at = models.DateTimeField(auto_now_add=True) updated_at = models.DateTimeField(auto_now_add=True) created_by = models.ForeignKey(User, on_delete=models.SET_NULL, null=True) conference = models.ForeignKey(Conference, on_delete=models.CASCADE) @staticmethod def render(frame_template, conference, subject, body): context_data = get_frame_context(conference, subject, body) context = Context(context_data, autoescape=False) return Template(frame_template).render(context) def render_html(self, subject, body): return EmailFrame.render( self.text_html, self.conference, subject, body ) def render_plain(self, subject, body): text_plain = self.text_plain if not text_plain: text_plain = html2text(self.text_html) return EmailFrame.render( text_plain, self.conference, subject, body ) class EmailSettings(models.Model): frame = models.ForeignKey(EmailFrame, on_delete=models.SET_NULL, null=True) conference = models.OneToOneField( Conference, null=True, blank=True, on_delete=models.CASCADE, related_name='email_settings', ) class GroupMessage(models.Model): subject = models.CharField(max_length=1024) body = models.TextField() conference = models.ForeignKey( Conference, on_delete=models.CASCADE, related_name='sent_group_emails', ) sent_at = models.DateTimeField(auto_now_add=True) sent_by = models.ForeignKey( User, on_delete=models.SET_NULL, null=True, related_name='sent_group_emails' ) sent = models.BooleanField(default=False) @property def message_type(self): return '' class UserMessage(GroupMessage): recipients = models.ManyToManyField(User, related_name='group_emails') group_message = models.OneToOneField( GroupMessage, on_delete=models.CASCADE, parent_link=True) @property def message_type(self): return MSG_TYPE_USER @staticmethod def create(subject, body, conference, objects_to): msg = UserMessage.objects.create( subject=subject, body=body, conference=conference) for user in objects_to: msg.recipients.add(user) msg.save() return msg def send(self, sender): # 1) Update status and save sender chair user: self.sent = False self.sent_by = sender self.save() # 2) For each user, we render this template with the given context, # and then build the whole message by inserting this body into # the frame. Plain-text version is also formed from HTML. frame = self.conference.email_settings.frame conference_context = get_conference_context(self.conference) for user in self.recipients.all(): context = Context({ **conference_context, **get_user_context(user, self.conference) }, autoescape=False) email = EmailMessage.create( group_message=self.group_message, user_to=user, context=context, frame=frame ) email.send(sender) # 3) Update self status, write sending timestamp self.sent_at = timezone.now() self.sent = True self.save() return self class SubmissionMessage(GroupMessage): recipients = models.ManyToManyField( Submission, related_name='group_emails') group_message = models.OneToOneField( GroupMessage, on_delete=models.CASCADE, parent_link=True) @property def message_type(self): return MSG_TYPE_SUBMISSION @staticmethod def create(subject, body, conference, objects_to): msg = SubmissionMessage.objects.create( subject=subject, body=body, conference=conference) for submission in objects_to: msg.recipients.add(submission) msg.save() return msg def send(self, sender): # 1) Update status and save sender chair user: self.sent = False self.sent_by = sender self.save() # 2) For each user, we render this template with the given context, # and then build the whole message by inserting this body into # the frame. Plain-text version is also formed from HTML. frame = self.conference.email_settings.frame conference_context = get_conference_context(self.conference) for submission in self.recipients.all(): submission_context = get_submission_context(submission) for author in submission.authors.all(): user = author.user context = Context({ **conference_context, **submission_context, **get_user_context(user, self.conference) }, autoescape=False) email = EmailMessage.create( group_message=self.group_message, user_to=user, context=context, frame=frame ) email.send(sender) # 3) Update self status, write sending timestamp self.sent_at = timezone.now() self.sent = True self.save() return self def get_group_message_model(msg_type): return { MSG_TYPE_USER: UserMessage, MSG_TYPE_SUBMISSION: SubmissionMessage, }[msg_type] def get_message_leaf_model(msg): """If provided a `GroupMessage` instance, check the inheritance, find the most descent child and return it. Now the possible leaf models are `UserMessage` and `SubmissionMessage`.""" if hasattr(msg, 'usermessage'): return msg.usermessage elif hasattr(msg, 'submissionmessage'): return msg.submissionmessage # Also check, maybe a message is already a leaf: if isinstance(msg, UserMessage) or isinstance(msg, SubmissionMessage): return msg # If neither succeeded, raise an error: raise TypeError(f'Not a group message: type(msg)') class EmailMessage(models.Model): subject = models.TextField(max_length=1024) text_plain = models.TextField() text_html = models.TextField() user_to = models.ForeignKey( User, on_delete=models.CASCADE, related_name='emails' ) sent_at = models.DateTimeField(auto_now_add=True) sent = models.BooleanField(default=False) sent_by = models.ForeignKey( User, on_delete=models.SET_NULL, null=True, related_name='sent_emails' ) group_message = models.ForeignKey( GroupMessage, on_delete=models.SET_NULL, null=True, related_name='messages', ) @staticmethod def create(group_message, user_to, context, frame): template_body = Template(group_message.body) template_subject = Template(group_message.subject) body_md = template_body.render(context) body_html = markdown(body_md) subject = template_subject.render(context) return EmailMessage.objects.create( user_to=user_to, group_message=group_message, subject=subject, text_html=frame.render_html(subject, body_html), text_plain=frame.render_plain(subject, body_md), ) def send(self, sender): if not self.sent: from_email = settings.DEFAULT_FROM_EMAIL send_mail(self.subject, self.text_plain, from_email, [self.user_to], html_message=self.text_html) self.sent_at = timezone.now() self.sent_by = sender self.sent = True self.save() return self class SystemNotification(models.Model): """This model represents a system notification fired on a specific event. The model itself doesn't define the circumstances in which the message must be sent, which are subject to views. Notification is defined with a mandatory name, optional description, subject and template. If template is not assigned or subject is not specified, messages won't be sent. Notification can also be turned off with `is_active` flag field. """ ASSIGN_STATUS_SUBMIT = 'assign_status_submit' ASSIGN_STATUS_REVIEW = 'assign_status_review' ASSIGN_STATUS_ACCEPT = 'assign_status_accept' ASSIGN_STATUS_REJECT = 'assign_status_reject' ASSIGN_STATUS_INPRINT = 'assign_status_inprint' ASSIGN_STATUS_PUBLISHED = 'assign_status_publish' NAME_CHOICES = ( (ASSIGN_STATUS_REVIEW, 'Assign status REVIEW to the paper'), (ASSIGN_STATUS_SUBMIT, 'Assign status SUBMIT to the paper'), (ASSIGN_STATUS_ACCEPT, 'Assign status ACCEPT to the paper'), (ASSIGN_STATUS_REJECT, 'Assign status REJECT to the paper'), (ASSIGN_STATUS_INPRINT, 'Assign status IN-PRINT to the paper'), (ASSIGN_STATUS_PUBLISHED, 'Assign status PUBLISHED to the paper'), ) name = CharField(max_length=64, choices=NAME_CHOICES) subject = CharField(max_length=1024, blank=True) is_active = BooleanField(default=False) type = CharField(max_length=64, choices=MESSAGE_TYPE_CHOICES, blank=False) body = TextField(blank=True) conference = ForeignKey(Conference, related_name='notifications', on_delete=CASCADE) class Meta: constraints = [ UniqueConstraint(fields=['conference', 'name'], name='unique_name'), ] def send(self, recipients, sender=None): if self.is_active and self.body and self.subject: message_class = get_group_message_model(self.type) message = message_class.create( self.subject, self.body, self.conference, recipients) message.send(sender) DEFAULT_NOTIFICATIONS_DATA = { SystemNotification.ASSIGN_STATUS_REVIEW: { 'subject': 'Submission #{{ paper_id }} is under review', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, your submission #{{ paper_id }} **"{{ paper_title }}"** is assigned for the review. Reviews are expected to be ready at **{{ rev_end_date|time:"H:i:s" }}**.''' }, SystemNotification.ASSIGN_STATUS_SUBMIT: { 'subject': 'Submission #{{ paper_id }} is in draft editing state', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, your submission #{{ paper_id }} **"{{ paper_title }}"** is in draft editing state. At this point you can modify review manuscript, title and other data if you need.''' }, SystemNotification.ASSIGN_STATUS_ACCEPT: { 'subject': 'Submission #{{ paper_id }} was accepted', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, congratulations, your submission #{{ paper_id }} **"{{ paper_title }}"** was accepted for the conference.''' }, SystemNotification.ASSIGN_STATUS_REJECT: { 'subject': 'Submission #{{ paper_id }} was rejected', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, unfortunately your submission #{{ paper_id }} **"{{ paper_title }}"** was rejected according to the double-blinded review. ''' }, SystemNotification.ASSIGN_STATUS_INPRINT: { 'subject': 'Submission #{{ paper_id }} was rejected', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, your submission #{{ paper_id }} **"{{ paper_title }}"** camera-ready was sent to the publisher. We will let you know when the paper will be published. ''' }, SystemNotification.ASSIGN_STATUS_PUBLISHED: { 'subject': 'Submission #{{ paper_id }} was rejected', 'type': MSG_TYPE_SUBMISSION, 'body': '''Dear {{ username }}, we are glad to inform you that your submission #{{ paper_id }} **"{{ paper_title }}"** was published. ''' }, }
35.157609
83
0.659762
a01a6cd80a71c68a6da168b3758e9d7078688990
100
py
Python
Pruebas.py
MacoChave/Server-Iniciales
035d98793a1c20738b7af885d455fd62197988bd
[ "Apache-2.0" ]
null
null
null
Pruebas.py
MacoChave/Server-Iniciales
035d98793a1c20738b7af885d455fd62197988bd
[ "Apache-2.0" ]
null
null
null
Pruebas.py
MacoChave/Server-Iniciales
035d98793a1c20738b7af885d455fd62197988bd
[ "Apache-2.0" ]
null
null
null
from datetime import date from datetime import datetime dateToday = date.today() print(dateToday)
14.285714
29
0.8
a01dc69fc961ecf3abcdcc4efc76fa8f20eeb48a
1,753
py
Python
translator/model.py
marco-nicola/python-translator
6a559874c9899e52a4cac9c2954dcca6b638f002
[ "Apache-2.0" ]
null
null
null
translator/model.py
marco-nicola/python-translator
6a559874c9899e52a4cac9c2954dcca6b638f002
[ "Apache-2.0" ]
null
null
null
translator/model.py
marco-nicola/python-translator
6a559874c9899e52a4cac9c2954dcca6b638f002
[ "Apache-2.0" ]
null
null
null
# Copyright 2021 Marco Nicola # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging from typing import Optional from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, MarianMTModel, \ MarianTokenizer from .config import ConfigLanguageModel class Model: def __init__(self, conf: ConfigLanguageModel, models_path: str): self._conf: ConfigLanguageModel = conf self._models_path: str = models_path self._tokenizer: Optional[MarianTokenizer] = None self._model: Optional[MarianMTModel] = None def load(self) -> None: logging.info(f'[{self._conf.model}] - Loading tokenizer...') self._tokenizer = AutoTokenizer.from_pretrained( self._conf.model, cache_dir=self._models_path) logging.info(f'[{self._conf.model}] - Loading model...') self._model = AutoModelForSeq2SeqLM.from_pretrained( self._conf.model, cache_dir=self._models_path) logging.info(f'[{self._conf.model}] - Loaded.') def translate(self, text: str) -> str: tokenized = self._tokenizer(text, return_tensors="pt", padding=True) outputs = self._model.generate(**tokenized) return self._tokenizer.decode(outputs[0], skip_special_tokens=True)
38.108696
79
0.718768
a01f36af66f5f536cdcfeccc977f7396f86e0837
15,082
py
Python
home/.local/share/tkthemes/clearlooks/create_imgs.py
ssokolow/profile
09f2a842077909d883a08b546659516deec7d719
[ "MIT" ]
9
2015-04-14T22:27:40.000Z
2022-02-23T05:33:00.000Z
home/.local/share/tkthemes/clearlooks/create_imgs.py
ssokolow/profile
09f2a842077909d883a08b546659516deec7d719
[ "MIT" ]
10
2018-06-18T07:57:56.000Z
2021-10-04T06:47:19.000Z
home/.local/share/tkthemes/clearlooks/create_imgs.py
ssokolow/profile
09f2a842077909d883a08b546659516deec7d719
[ "MIT" ]
9
2015-04-14T22:27:42.000Z
2017-11-21T11:34:23.000Z
#!/usr/bin/env python # -*- mode: python; coding: koi8-r; -*- import os import gtk, gobject imdir = 'images' imtype = 'png' background = '#efebe7' #fill_color = 0xff000000 # red fill_color = int('ff000000', 16) if not os.path.exists(imdir): os.mkdir(imdir) gc = None def draw_rect(): global gc if gc is None: gc = drawing_area.window.new_gc() colormap = gtk.gdk.colormap_get_system() gc.set_colormap(colormap) color = gtk.gdk.color_parse('red') colormap.alloc_color(color) gc.set_rgb_fg_color(color) drawing_area.window.draw_rectangle(gc, True, 0,0, 800,800) def save_image(fn, w, h, x=0, y=0): pixbuf = gtk.gdk.Pixbuf(gtk.gdk.COLORSPACE_RGB, True, 8, w, h) pixbuf.fill(fill_color) pb = pixbuf.get_from_drawable(drawing_area.window, drawing_area.get_colormap(), x,y, 0,0, w,h) pb.save(os.path.join(imdir, fn+"."+imtype), imtype) drawing_area.window.clear() draw_rect() done = False def save_callback(*args): global done if done: return done = True print 'create images' style = drawing_area.get_style() draw_rect() # separator w = 20 style.paint_vline(drawing_area.window, gtk.STATE_NORMAL, None, drawing_area, "frame", 0, w, 0) save_image('sep-v', 2, w) style.paint_hline(drawing_area.window, gtk.STATE_NORMAL, None, drawing_area, "frame", 0, w, 0) save_image('sep-h', w, 2) # tree w, h = 32, 32 w, h = 24, 24 for fn, state, shadow in ( ("tree-n", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("tree-h", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("tree-p", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("tree-d", gtk.STATE_INSENSITIVE, gtk.SHADOW_IN), ): style.paint_box(drawing_area.window, state, shadow, None, drawing_area, "stepper", 0,0, w,h) save_image(fn, w, h) # sizegrip w, h = 16, 16 fn = 'sizegrip' style.paint_resize_grip(drawing_area.window, gtk.STATE_NORMAL, None, drawing_area, "statusbar", gtk.gdk.WINDOW_EDGE_SOUTH_EAST, 0,0, w,h) save_image(fn, w, h) # progress w, h = 37+3, 16+3 progress_style = progress.get_style() fn = 'progress-h' progress_style.paint_box(drawing_area.window, gtk.STATE_PRELIGHT, gtk.SHADOW_NONE, None, progress, "bar", 0,0, w,h) save_image(fn, w, h) # button w, h = 32, 32 w, h = 28, 28 for fn, state, shadow in ( ("button-n", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("button-a", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("button-p", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("button-d", gtk.STATE_INSENSITIVE, gtk.SHADOW_OUT), ): style.paint_box(drawing_area.window, state, shadow, None, drawing_area, "buttondefault", 0,0, w,h) save_image(fn, w, h) style.paint_box(drawing_area.window, gtk.STATE_PRELIGHT, gtk.SHADOW_IN, None, togglebutton, "buttondefault", 0,0, w,h) save_image("button-pa", w, h) # toolbar w, h = 16, 16 w, h = 24, 24 fn = "blank" pixbuf = gtk.gdk.Pixbuf(gtk.gdk.COLORSPACE_RGB, True, 8, w, h) pixbuf.fill(fill_color) pixbuf.save(os.path.join(imdir, fn+"."+imtype), imtype) for fn, state, shadow in ( ("toolbutton-n", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("toolbutton-a", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("toolbutton-p", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("toolbutton-d", gtk.STATE_INSENSITIVE, gtk.SHADOW_IN), ): style.paint_box(drawing_area.window, state, shadow, None, drawing_area, "buttondefault", 0,0, w,h) save_image(fn, w, h) style.paint_box(drawing_area.window, gtk.STATE_PRELIGHT, gtk.SHADOW_IN, None, togglebutton, "buttondefault", 0,0, w,h) save_image("toolbutton-pa", w, h) # slider msl = hscroll.style_get_property("min_slider_length") msl = 20 sw = hscroll.style_get_property("slider_width") print '>>', msl, sw for t, w, h, state, orient in ( ('hn', msl,sw, gtk.STATE_NORMAL, gtk.ORIENTATION_HORIZONTAL), ('ha', msl,sw, gtk.STATE_PRELIGHT, gtk.ORIENTATION_HORIZONTAL), ('hp', msl,sw, gtk.STATE_NORMAL, gtk.ORIENTATION_HORIZONTAL), ('hd', msl,sw, gtk.STATE_INSENSITIVE, gtk.ORIENTATION_HORIZONTAL), ('vn', sw,msl, gtk.STATE_NORMAL, gtk.ORIENTATION_VERTICAL), ('va', sw,msl, gtk.STATE_PRELIGHT, gtk.ORIENTATION_VERTICAL), ('vp', sw,msl, gtk.STATE_NORMAL, gtk.ORIENTATION_VERTICAL), ('vd', sw,msl, gtk.STATE_INSENSITIVE, gtk.ORIENTATION_VERTICAL), ): fn = 'sbthumb-'+t if 0: style.paint_slider(drawing_area.window, state, gtk.SHADOW_OUT, None, drawing_area, "slider", 0,0, w,h, orient) else: if orient == gtk.ORIENTATION_VERTICAL: w, h = h, w style.paint_box(drawing_area.window, state, shadow, None, drawing_area, "stepper", 0,0, w,h) save_image(fn, w, h) msl = hscroll.style_get_property("min_slider_length") sw = hscroll.style_get_property("slider_width") # scale for t, w, h, state, orient in ( ('hn', msl,sw, gtk.STATE_NORMAL, gtk.ORIENTATION_HORIZONTAL), ('ha', msl,sw, gtk.STATE_PRELIGHT, gtk.ORIENTATION_HORIZONTAL), ('hd', msl,sw, gtk.STATE_INSENSITIVE, gtk.ORIENTATION_HORIZONTAL), ('vn', sw,msl, gtk.STATE_NORMAL, gtk.ORIENTATION_VERTICAL), ('va', sw,msl, gtk.STATE_PRELIGHT, gtk.ORIENTATION_VERTICAL), ('vd', sw,msl, gtk.STATE_INSENSITIVE, gtk.ORIENTATION_VERTICAL), ): fn = 'scale-'+t if orient == gtk.ORIENTATION_HORIZONTAL: detail = "hscale" else: detail = "vscale" style.paint_slider(drawing_area.window, state, gtk.SHADOW_OUT, None, drawing_area, detail, 0,0, w+2,h+2, orient) save_image(fn, w, h, 1, 1) w, h = msl, sw fn = 'scaletrough-h' style.paint_box(drawing_area.window, gtk.STATE_ACTIVE, gtk.SHADOW_IN, None, scale, "trough", 0,0, w,h) save_image(fn, w, h) # arrow w = h = hscroll.style_get_property("stepper_size") #w = h = 15 arrow_width = w / 2 arrow_height = h / 2 arrow_x = (w - arrow_width) / 2 arrow_y = (h - arrow_height) / 2 alloc = hscroll.get_allocation() x0 = alloc.x x1 = alloc.x+alloc.width-w alloc = vscroll.get_allocation() y0 = alloc.y y1 = alloc.y+alloc.height-h sn = gtk.STATE_NORMAL sp = gtk.STATE_PRELIGHT sa = gtk.STATE_ACTIVE si = gtk.STATE_INSENSITIVE for fn, x, y, state, shadow, arrow_type, widget in ( ("arrowleft-n", x0, 0, sn, gtk.SHADOW_OUT, gtk.ARROW_LEFT, hscroll), ("arrowleft-a", x0, 0, sp, gtk.SHADOW_OUT, gtk.ARROW_LEFT, hscroll), ("arrowleft-p", x0, 0, sa, gtk.SHADOW_IN, gtk.ARROW_LEFT, hscroll), ("arrowleft-d", x0, 0, si, gtk.SHADOW_OUT, gtk.ARROW_LEFT, hscroll), ("arrowright-n", x1, 0, sn, gtk.SHADOW_OUT, gtk.ARROW_RIGHT, hscroll), ("arrowright-a", x1, 0, sp, gtk.SHADOW_OUT, gtk.ARROW_RIGHT, hscroll), ("arrowright-p", x1, 0, sa, gtk.SHADOW_IN, gtk.ARROW_RIGHT, hscroll), ("arrowright-d", x1, 0, si, gtk.SHADOW_OUT, gtk.ARROW_RIGHT, hscroll), ("arrowup-n", 0, y0, sn, gtk.SHADOW_OUT, gtk.ARROW_UP, vscroll), ("arrowup-a", 0, y0, sp, gtk.SHADOW_OUT, gtk.ARROW_UP, vscroll), ("arrowup-p", 0, y0, sa, gtk.SHADOW_IN, gtk.ARROW_UP, vscroll), ("arrowup-d", 0, y0, si, gtk.SHADOW_OUT, gtk.ARROW_UP, vscroll), ("arrowdown-n", 0, y1, sn, gtk.SHADOW_OUT, gtk.ARROW_DOWN, vscroll), ("arrowdown-a", 0, y1, sp, gtk.SHADOW_OUT, gtk.ARROW_DOWN, vscroll), ("arrowdown-p", 0, y1, sa, gtk.SHADOW_IN, gtk.ARROW_DOWN, vscroll), ("arrowdown-d", 0, y1, si, gtk.SHADOW_OUT, gtk.ARROW_DOWN, vscroll), ): if 0: detail = 'hscrollbar' if widget is vscroll: detail = 'vscrollbar' else: x, y = 0, 0 detail = 'stepper' widget = drawing_area style.paint_box(drawing_area.window, state, shadow, None, widget, detail, x,y, w,h) style.paint_arrow(drawing_area.window, state, shadow, None, widget, detail, arrow_type, True, x+arrow_x, y+arrow_y, arrow_width, arrow_height) save_image(fn, w, h, x, y) # combobox w, h = w, 24 w, h = 16, 24 alloc = hscroll.get_allocation() x1 = alloc.x+alloc.width-w arrow_width = w / 2 arrow_height = h / 2 arrow_x = (w - arrow_width) / 2 arrow_y = (h - arrow_height) / 2 detail = 'hscrollbar' widget = hscroll for fn, state, shadow, arrow_type in ( ("comboarrow-n", gtk.STATE_NORMAL, gtk.SHADOW_OUT, gtk.ARROW_DOWN), ("comboarrow-a", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT, gtk.ARROW_DOWN), ("comboarrow-p", gtk.STATE_ACTIVE, gtk.SHADOW_IN, gtk.ARROW_DOWN), ("comboarrow-d", gtk.STATE_INSENSITIVE, gtk.SHADOW_IN, gtk.ARROW_DOWN), ): style.paint_box(drawing_area.window, state, shadow, None, widget, detail, x1,0, w,h) style.paint_arrow(drawing_area.window, state, shadow, None, drawing_area, "stepper", arrow_type, True, x1+arrow_x, arrow_y, arrow_width, arrow_height) save_image(fn, w, h, x1, 0) w = 24 for fn, state, shadow in ( ("combo-rn", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("combo-ra", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("combo-rp", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("combo-rd", gtk.STATE_INSENSITIVE, gtk.SHADOW_OUT), ): style.paint_box(drawing_area.window, state, shadow, None, drawing_area, "button", 0,0, w+2,h) save_image(fn, w, h) style.paint_box(drawing_area.window, gtk.STATE_NORMAL, gtk.SHADOW_OUT, None, drawing_area, "button", 0,0, w+2,h) d = 3 style.paint_focus(drawing_area.window, gtk.STATE_NORMAL, None, drawing_area, "button", d,d, w-2*d,h-2*d) save_image('combo-rf', w, h) style.paint_shadow(drawing_area.window, gtk.STATE_NORMAL, gtk.SHADOW_IN, None, drawing_area, "entry", 0,0, w+2,h) save_image('combo-n', w, h) # checkbutton #define INDICATOR_SIZE 13 #define INDICATOR_SPACING 2 x, y = 2, 2 w, h = 13, 13 #w = h = checkbutton.style_get_property("indicator_size") for fn, state, shadow in ( ("check-nc", gtk.STATE_NORMAL, gtk.SHADOW_IN), ("check-nu", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("check-ac", gtk.STATE_PRELIGHT, gtk.SHADOW_IN), ("check-au", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("check-pc", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("check-pu", gtk.STATE_ACTIVE, gtk.SHADOW_OUT), ("check-dc", gtk.STATE_INSENSITIVE, gtk.SHADOW_IN), ("check-du", gtk.STATE_INSENSITIVE, gtk.SHADOW_OUT), ): ## style.paint_flat_box(drawing_area.window, ## gtk.STATE_PRELIGHT, ## gtk.SHADOW_ETCHED_OUT, ## gtk.gdk.Rectangle(0,0,w,h), drawing_area, ## "checkbutton", 0,0, w,h) style.paint_check(drawing_area.window, state, shadow, None, drawing_area, "checkbutton", x,y, w,h) save_image(fn, w+2*x, h+2*y) # radiobutton for fn, state, shadow in ( ("radio-nc", gtk.STATE_NORMAL, gtk.SHADOW_IN), ("radio-nu", gtk.STATE_NORMAL, gtk.SHADOW_OUT), ("radio-ac", gtk.STATE_PRELIGHT, gtk.SHADOW_IN), ("radio-au", gtk.STATE_PRELIGHT, gtk.SHADOW_OUT), ("radio-pc", gtk.STATE_ACTIVE, gtk.SHADOW_IN), ("radio-pu", gtk.STATE_ACTIVE, gtk.SHADOW_OUT), ("radio-dc", gtk.STATE_INSENSITIVE, gtk.SHADOW_IN), ("radio-du", gtk.STATE_INSENSITIVE, gtk.SHADOW_OUT), ): ## style.paint_flat_box(drawing_area.window, ## gtk.STATE_PRELIGHT, ## gtk.SHADOW_ETCHED_OUT, ## gtk.gdk.Rectangle(0,0,w,h), drawing_area, ## "checkbutton", 0,0, w,h) style.paint_option(drawing_area.window, state, shadow, None, drawing_area, "radiobutton", x,y, w,h) save_image(fn, w+2*x, h+2*y) # notebook w, h = 28, 22 state = gtk.STATE_NORMAL shadow = gtk.SHADOW_OUT for fn, gap_h, state in ( ("tab-n", 0, gtk.STATE_NORMAL), ("tab-a", 2, gtk.STATE_ACTIVE), ): ## style.paint_box_gap(drawing_area.window, state, shadow, ## gtk.gdk.Rectangle(0,0,w,gap_h), drawing_area, ## "notebook", 0,0, w,gap_h, gtk.POS_TOP, 0, w) y = gap_h hh = h - y style.paint_extension(drawing_area.window, state, gtk.SHADOW_OUT, None, drawing_area, "tab", 0,y, w,hh, gtk.POS_BOTTOM) save_image(fn, w, h+2) print 'done' gtk.main_quit() def pack(w, row, col): table.attach(w, col, col+1, row, row+1, gtk.EXPAND | gtk.FILL, gtk.EXPAND | gtk.FILL, 0, 0) win = gtk.Window() win.connect("destroy", gtk.main_quit) table = gtk.Table() win.add(table) row, col = 0, 0 drawing_area = gtk.DrawingArea() #drawing_area.set_size_request(100, 100) pack(drawing_area, row, col) row += 1 vscroll = gtk.VScrollbar() pack(vscroll, 0, 1) hscroll = gtk.HScrollbar() pack(hscroll, row, col) row += 1 notebook = gtk.Notebook() label = gtk.Label("Label") notebook.append_page(label) label = gtk.Label("Label") notebook.append_page(label) pack(notebook, row, col) row += 1 button = gtk.Button("Button") pack(button, row, col) row += 1 checkbutton = gtk.CheckButton("CheckButton") pack(checkbutton, row, col) row += 1 progress = gtk.ProgressBar() pack(progress, row, col) row += 1 scale = gtk.HScale() pack(scale, row, col) row += 1 entry = gtk.Entry() pack(entry, row, col) row += 1 togglebutton = gtk.ToggleButton() pack(togglebutton, row, col) togglebutton.set_active(True) row += 1 drawing_area.connect("expose-event", save_callback) #gobject.timeout_add(2000, save_callback) win.show_all() #drawing_area.modify_bg(gtk.STATE_NORMAL, gtk.gdk.color_parse('red')) gtk.main()
34.045147
79
0.579631
a01fef4f0de34d121af8b1cbdb955d6acb92f36a
4,687
py
Python
coil_COMdistance.py
Johanu/MDAnalysis_scripts
f44d2b32bb916daa3bd61e9f3ad636db503293bf
[ "MIT" ]
null
null
null
coil_COMdistance.py
Johanu/MDAnalysis_scripts
f44d2b32bb916daa3bd61e9f3ad636db503293bf
[ "MIT" ]
null
null
null
coil_COMdistance.py
Johanu/MDAnalysis_scripts
f44d2b32bb916daa3bd61e9f3ad636db503293bf
[ "MIT" ]
null
null
null
from __future__ import division import matplotlib.pyplot as plt import MDAnalysis as md import numpy as np def calculate_dists(gro_file, xtc_file): u = md.Universe(gro_file, xtc_file) select_group1 = u.selectAtoms("backbone and (resnum 50 or resnum 51)") select_group2 = u.selectAtoms("backbone and (resnum 149 or resnum 150)") select_group3 = u.selectAtoms("backbone and (resnum 50 or resnum 51 or resnum 149 or resnum 150)") select_group4 = u.selectAtoms("backbone and (resnum 25 or resnum 124)") for i in select_group1: print "Loop1 ", i for i in select_group2: print "Loop2 ", i for i in select_group4: print "ASP ", i COM_distance = [] COM_distance_ASP = [] COM_distance_ASP1 = [] COM_distance_ASP2 = [] max_dist = 0 index = 0 min_dist = 100 index_min = 0 max_dist_1 = 0 index_1 = 0 min_dist_1 = 100 index_min_1 = 0 max_dist_2 = 0 index_2 = 0 min_dist_2 = 100 index_min_2 = 0 max_dist_3 = 0 index_3 = 0 min_dist_3 = 100 index_min_3 = 0 #group1_COM = select_group1.centerOfMass() #group2_COM = select_group2.centerOfMass() #print group1_COM #print group2_COM #print np.sqrt(np.dot(group1_COM-group2_COM, group1_COM-group2_COM)) #print np.linalg.norm(group1_COM - group2_COM) for i in u.trajectory: group1_COM = select_group1.centerOfMass() group2_COM = select_group2.centerOfMass() dist = np.linalg.norm(group1_COM - group2_COM) COM_distance.append(dist) if dist > max_dist: max_dist = dist index = i.frame if dist < min_dist: min_dist = dist index_min = i.frame group3_COM = select_group3.centerOfMass() group4_COM = select_group4.centerOfMass() dist1 = np.linalg.norm(group3_COM - group4_COM) COM_distance_ASP.append(dist1) if dist1 > max_dist_1: max_dist_1 = dist1 index_1 = i.frame if dist1 < min_dist_1: min_dist_1 = dist1 index_min_1 = i.frame dist2 = np.linalg.norm(group1_COM - group4_COM) dist3 = np.linalg.norm(group2_COM - group4_COM) COM_distance_ASP1.append(dist2) COM_distance_ASP2.append(dist3) if dist2 > max_dist_2: max_dist_2 = dist2 index_2 = i.frame if dist2 < min_dist_2: min_dist_2 = dist2 index_min_2 = i.frame if dist3 > max_dist_3: max_dist_3 = dist3 index_3 = i.frame if dist3 < min_dist_3: min_dist_3 = dist3 index_min_3 = i.frame print 'Max interloop distance: ', max_dist, index print 'Min interloop distance: ', min_dist, index_min print 'Max loops-ASP distance: ', max_dist_1, index_1 print 'Min loops-ASP distance: ', min_dist_1, index_min_1 print 'Max loop1-ASP distance: ', max_dist_2, index_2 print 'Min loop1-ASP distance: ', min_dist_2, index_min_2 print 'Max loop2-ASP distance: ', max_dist_3, index_3 print 'Min loop2-ASP distance: ', min_dist_3, index_min_3 return COM_distance, COM_distance_ASP, COM_distance_ASP1, COM_distance_ASP2 coil_distance, ASP_distance, ASP_distance1, ASP_distance2 = calculate_dists('structure.pdb', 'equ.dcd') x_vals = [x / 10 for x in range(0, len(coil_distance))] plt.plot(x_vals, coil_distance, linewidth=0.5) #leg = plt.legend(ncol=3, loc=9, fancybox=True) #leg.get_frame().set_alpha(0.5) plt.xlabel('Time / ns') plt.ylabel(ur'Loop COM distance / $\AA$') plt.axhline(y=9.84, linewidth=1, color = 'red') plt.axhline(y=11.11, linewidth=1, color = 'green') plt.savefig('coil_COMdistance.png', dpi=300) plt.close() plt.plot(x_vals, ASP_distance, linewidth=0.5) plt.plot(x_vals, ASP_distance1, linewidth=0.5) plt.plot(x_vals, ASP_distance2, linewidth=0.5) print 'Loop1 average: ', np.average(ASP_distance1[500:]), np.std(ASP_distance1[500:]) print 'Loop2 average: ', np.average(ASP_distance2[500:]), np.std(ASP_distance2[500:]) plt.xlabel('Time / ns') plt.ylabel(ur'Loop COM distance / $\AA$') plt.axhline(y=21.29, linewidth=1, color = '#C45AEC', label='PR20') plt.axhline(y=15.18, linewidth=1, color = '#C45AEC') plt.axhline(y=20.36, linewidth=1, color = '#EAC117', label='PR') plt.axhline(y=15.11, linewidth=1, color = '#EAC117') plt.axhline(y=np.average(ASP_distance1), linewidth=1, color = 'green', label='Loop1 average') plt.axhline(y=np.average(ASP_distance2), linewidth=1, color = 'red', label='Loop2 average') leg = plt.legend(fancybox=True, loc=2, framealpha=0.5) #leg.get_frame().set_alpha(0.5) plt.savefig('ASP_COMdistance.png', dpi=300) plt.close()
39.058333
103
0.667164
a021e7c81cd72a8cb8466d95bea774bd4667239f
1,692
py
Python
src/api/content_flag.py
Viewly/alpha-2
6b6d827197489164d8c4bde4f4d591dcec5a2163
[ "MIT" ]
null
null
null
src/api/content_flag.py
Viewly/alpha-2
6b6d827197489164d8c4bde4f4d591dcec5a2163
[ "MIT" ]
1
2021-05-07T06:26:16.000Z
2021-05-07T06:26:16.000Z
src/api/content_flag.py
Viewly/alpha-2
6b6d827197489164d8c4bde4f4d591dcec5a2163
[ "MIT" ]
null
null
null
import datetime as dt import json from flask_restful import ( Resource, reqparse, ) from flask_security import current_user from marshmallow_sqlalchemy import ModelSchema from .utils import auth_required from .. import db from ..core.utils import log_exception from ..models import ContentFlag class FlagSchema(ModelSchema): class Meta: model = ContentFlag include_fk = True flag_schema = FlagSchema() parser = reqparse.RequestParser() parser.add_argument('video_id', type=str, required=True) parser.add_argument('flag_type', type=str) class FlagApi(Resource): method_decorators = [auth_required] def get(self): args = parser.parse_args() flag = \ (db.session.query(ContentFlag) .filter(ContentFlag.video_id == args['video_id'], ContentFlag.user_id == current_user.id) .first()) return flag_schema.dump(flag).data or ({}, 404) def put(self): args = parser.parse_args() try: assert args['flag_type'] in \ ['xxx', 'hate', 'scam', 'spam', 'plagiarism'], 'Invalid flag' flag = ContentFlag( user_id=current_user.id, video_id=args['video_id'], flag_type=args['flag_type'], created_at=dt.datetime.utcnow(), ) db.session.add(flag) db.session.commit() except AssertionError as e: log_exception() return dict(message=str(e)), 400 except Exception as e: log_exception() return dict(message=str(e)), 500 return flag_schema.dump(flag).data
26.030769
77
0.60461
a0220e4b4dae9e864bc6a43965e05ecf1eb56be9
13,231
py
Python
cgmodsel/utils.py
franknu/cgmodsel
b008ed88e4f10205ee0ff5e9433d5426c1d5ff6a
[ "MIT" ]
1
2020-09-01T08:39:14.000Z
2020-09-01T08:39:14.000Z
cgmodsel/utils.py
franknu/cgmodsel
b008ed88e4f10205ee0ff5e9433d5426c1d5ff6a
[ "MIT" ]
null
null
null
cgmodsel/utils.py
franknu/cgmodsel
b008ed88e4f10205ee0ff5e9433d5426c1d5ff6a
[ "MIT" ]
1
2020-09-04T13:35:41.000Z
2020-09-04T13:35:41.000Z
# -*- coding: utf-8 -*- """ Copyright: Frank Nussbaum (frank.nussbaum@uni-jena.de) This file contains various functions used in the module including - sparse norms and shrinkage operators - a stable logsumexp implementation - array printing-method that allows pasting the output into Python code """ import numpy as np ################################################################################# # norms and shrinkage operators ################################################################################# try: # the following requires setup # import os # os.system('python cyshrink/setup.py build_ext --inplace') # TODO(franknu): configure n_threads/interface from cyshrink.shrink.shrink import grp as grp_soft_shrink from cyshrink.shrink.shrink import grp_weight as grp_soft_shrink_weight print('successfully imported shrink.shrink') except Exception as e: print(e) # from cyshrink.shrink.shrink import grp_weight as grp_soft_shrink_weight2 # naive and slow implementations print(''' Failed to import Cython shrink functions, setup is required... using slower native Python functions instead''') def grp_soft_shrink(mat, tau, glims, off=False): """just a wrapper for grp_soft_shrink_weight with weiths=None""" return grp_soft_shrink_weight(mat, tau, glims, off=False, weights=None) def grp_soft_shrink_weight(mat, tau, glims, off=False, weights=None): """ calculate (group-)soft-shrinkage. Args: mat (np.array): matrix. tau (float): non-negative shrinkage parameter. off (bool): if True, do not shrink diagonal entries. glims: group delimiters (cumulative sizes of groups). weights (optional): weights for weighted l_{1,2} norm/shrinkage. Returns: tuple: shrunken matrix, (group) l_{1,2}-norm of shrunken matrix. Note: this code could be made much faster (by parallizing loops, efficient storage access). """ shrinkednorm = 0 # if glims is None: n_groups = len(glims) - 1 if glims[-1] == n_groups: # each group has size 1 tmp = np.abs(mat) if not weights is None: # weighted l1-norm # tmp = np.multiply(tmp, weights).flatten tmp -= tau * weights else: tmp -= tau tmp[tmp < 1e-25] = 0 shrinked = np.multiply(np.sign(mat), tmp) l1norm = np.sum(np.abs(shrinked.flatten())) if off: l1norm -= np.sum(np.abs(np.diag(shrinked))) shrinked -= np.diag(np.diag(shrinked)) shrinked += np.diag(np.diag(mat)) return shrinked, l1norm # group soft shrink if weights is None: weights = np.ones(mat.shape) # TODO(franknu): improve style tmp = np.empty(mat.shape) for i in range(n_groups): for j in range(n_groups): # TODO(franknu): use symmetry group = mat[glims[i]:glims[i + 1], glims[j]:glims[j + 1]] if (i == j) and off: tmp[glims[i]:glims[i + 1], glims[i]:glims[i + 1]] = group continue gnorm = np.linalg.norm(group, 'fro') w_ij = tau * weights[i,j] if gnorm <= w_ij: tmp[glims[i]:glims[i + 1], glims[j]:glims[j + 1]] = np.zeros(group.shape) else: tmp[glims[i]:glims[i+1], glims[j]:glims[j+1]] = \ group * (1 - w_ij / gnorm) shrinkednorm += weights[i,j] * (1 - w_ij / gnorm) * gnorm return tmp, shrinkednorm def l21norm(mat, glims=None, off=False, weights=None): """ calculate l_{1,2}-norm. Args: mat (np.array): matrix. off (bool): if True, do not shrink diagonal entries. glims: group delimiters (cumulative sizes of groups). n_groups: # groups per row/column (if this is given, perform group soft shrink instead of soft shrink). weights (optional): weights for weighted l_{1,2} norm. Returns: float: (group) l_{1,2}-norm. """ if glims is None: # calculate regular l1-norm tmp = np.abs(mat) # tmp is copy, can do this inplace by specifying out if not weights is None: # weighted l1-norm tmp = np.multiply(tmp, weights).flatten tmp = np.sum(tmp) if off: tmp -= np.sum(np.diag(np.abs(mat))) return tmp n_groups = len(glims) - 1 l21sum = 0 if weights is None: for i in range(n_groups): for j in range(i): group = mat[glims[i]:glims[i + 1], glims[j]:glims[j + 1]] l21sum += np.linalg.norm(group, 'fro') else: for i in range(n_groups): for j in range(i): group = mat[glims[i]:glims[i + 1], glims[j]:glims[j + 1]] l21sum += weights[i,j] * np.linalg.norm(group, 'fro') l21sum *= 2 # use symmetry if not off: for i in range(n_groups): group = mat[glims[i]:glims[i + 1], glims[i]:glims[i + 1]] l21sum += np.linalg.norm(group, 'fro') return l21sum ############################################################################### # stable implementation of logsumexp etc. ############################################################################### #from scipy.special import logsumexp def _exp_shiftedmax(array, axis=None): """calculate exponentials of array shifted by its max, avoiding overflow by subtracting maximum before""" a_max = np.amax(array, axis=axis, keepdims=True) if a_max.ndim > 0: a_max[~np.isfinite(a_max)] = 0 elif not np.isfinite(a_max): a_max = 0 # print((a-a_max).shape) exp_shiftedamax = np.exp(array - a_max) # last line: a_max is repeated columnwise (if axis = 1) return exp_shiftedamax, a_max def logsumexp(array, axis=None, keepdims=True): """Compute the log of the sum of exponentials of input elements. Args: array (np.array): array on which to compute logsumexp. axis (int): axis along which to compute logsupexp. keepdims (bool): passed to np.sum. Returns: np.array: logsumexp Note: This is an adaptation of logsumexp in scipy.special (v1.1.0) """ exp_shifted, a_max = _exp_shiftedmax(array, axis=axis) # suppress warnings about log of zero with np.errstate(divide='ignore'): summed = np.sum(exp_shifted, axis=axis, keepdims=keepdims) out = np.log(summed) if not keepdims: a_max = np.squeeze(a_max, axis=axis) out += a_max return out def _logsumexp_and_conditionalprobs(array): """return logsumexp and conditional probabilities from array a that has the same shape as the discrete data in dummy-representation""" exp_shifted, a_max = _exp_shiftedmax(array, axis=1) summed = np.sum(exp_shifted, axis=1, keepdims=True) # entries always > 1 # suppress warnings about log of zero with np.errstate(divide='ignore'): out_logsumexp = np.log(summed) out_logsumexp += a_max # node conditional probabilities size = array.shape[1] out_conditionalprobs = np.divide(exp_shifted, np.dot(summed, np.ones((1, size)))) # unstable = np.log(np.sum(np.exp(a), axis = 1)).reshape((a.shape[0], 1)) # diff = unstable - out_logsumexp # print (unstable) # for i in range(unstable.shape[0]): # if abs(diff[i, 0]) > 10e-5: # print('a', a[i, :]) # print('unstable', unstable[i, 0]) # print('stable', out_logsumexp[i, 0]) # break # assert np.linalg.norm(unstable - out_logsumexp) < 10E-5 # print(out_logsumexp) # print(out_logsumexp[:1, 0]) # assert 1 == 0 out_logsumexp = np.squeeze(out_logsumexp) return out_logsumexp, out_conditionalprobs def _logsumexp_condprobs_red(array): """normalization and conditional probabilities for reduced levels, a ... two-dimensional array""" a_max = np.amax(array, axis=1, keepdims=True) a_max = np.maximum(a_max, 0) # last line: account for missing column with probs exp(0) for 0th level if a_max.ndim > 0: a_max[~np.isfinite(a_max)] = 0 elif not np.isfinite(a_max): a_max = 0 exp_shifted = np.exp(array - a_max) # a_max is repeated columnwise (axis=1) # calc column vector s of (shifted) normalization sums # note that entries always > 1, since one summand in each col is exp(0) summed = np.sum(exp_shifted, axis=1, keepdims=True) summed += np.exp(-a_max) # add values from missing 0th column # suppress warnings about log of zero with np.errstate(divide='ignore'): out_logsumexp = np.log(summed) out_logsumexp += a_max out_logsumexp = np.squeeze(out_logsumexp) # node conditional probabilities, required for gradient size = array.shape[1] out_conditionalprobs = np.divide(exp_shifted, np.dot(summed, np.ones((1, size)))) # note: log of this is not stable if probabilities close to zero # - use logsumexp instead for calculating plh value return out_logsumexp, out_conditionalprobs ############################################################################### # some conversion functions for representations of discrete data ############################################################################### def dummy_to_index_single(dummy_x, sizes): """convert dummy to index representation""" offset = 0 ind = np.empty(len(sizes), dtype=np.int) for i, size_r in enumerate(sizes): for j in range(size_r): if dummy_x[offset + j] == 1: ind[i] = j break offset += size_r return ind def dummy_to_index(dummy_data, sizes): """convert dummy to index representation""" n_data, ltot = dummy_data.shape assert ltot == sum(sizes) n_cat = len(sizes) index_data = np.empty((n_data, n_cat), dtype=np.int) for k in range(n_data): offset = 0 for i, size_r in enumerate(sizes): for j in range(size_r): if dummy_data[offset + j] == 1: index_data[k, i] = j break offset += size_r return index_data #def dummypadded_to_unpadded(dummy_data, n_cat): # """remove convert dummy to index representation""" # unpadded = np.empty(n_cat) # for i,x in enumerate(dummy_data): # if i % 2 == 1: # unpadded[i // 2] = x # return unpadded def index_to_dummy(idx, glims, ltot): """convert index to dummy representation""" dummy_data = np.zeros(ltot) for i, ind in enumerate(idx): dummy_data[glims[i] + ind] = 1 return dummy_data def dummy2dummyred(dummy_data, glims): """convert dummy to reduced dummy representation""" return np.delete(dummy_data, glims[:-1], 1) ############################################################################### # testing utilities ############################################################################### def strlistfrom(array, rnd=2): """a convenient representation for printing out numpy array s.t. it can be reused as a list""" string = np.array2string(array, precision=rnd, separator=',') string = 'np.array(' + string.translate({ord(c): None for c in '\n '}) + ')' return string def tomatlabmatrix(mat): """print numpy matrix in a way that can be pasted into MATLAB code.""" nrows, ncols = mat.shape string = "[" for i in range(nrows): string += "[" for j in range(ncols): string += str(mat[i, j]) + " " string += "];" string = string[:-1] + "]" print(string) def frange(start, stop, step): """a float range function""" i = start while i < stop: yield i i += step if __name__ == '__main__': SIZES = [2, 2, 2] GLIMS = [0, 2, 4, 6] LTOT = 6 IND = [0, 0, 1] DUMMY = index_to_dummy(IND, GLIMS, LTOT) IND2 = dummy_to_index_single(DUMMY, SIZES) MAT = np.arange(6).reshape((3, 2)) RES = _logsumexp_condprobs_red(MAT) print(RES) # res should be # (array([ 1.55144471, 3.34901222, 5.31817543]), array([[ 0.21194156, 0.57611688], # [ 0.25949646, 0.70538451], # [ 0.26762315, 0.72747516]]))
33.752551
90
0.542438
4e41eecc288939d5378c49ce5811a41875918b72
1,091
py
Python
authorization/migrations/0002_auto_20200207_2011.py
KariSpace/CRM_Sedicomm
cb19e90ca99c7a50a1841afbfb878191f62dec5c
[ "MIT" ]
null
null
null
authorization/migrations/0002_auto_20200207_2011.py
KariSpace/CRM_Sedicomm
cb19e90ca99c7a50a1841afbfb878191f62dec5c
[ "MIT" ]
null
null
null
authorization/migrations/0002_auto_20200207_2011.py
KariSpace/CRM_Sedicomm
cb19e90ca99c7a50a1841afbfb878191f62dec5c
[ "MIT" ]
null
null
null
# Generated by Django 2.2.4 on 2020-02-07 18:11 from django.db import migrations, models import django.db.models.deletion import jsonfield.fields class Migration(migrations.Migration): dependencies = [ ('authorization', '0001_initial'), ] operations = [ migrations.CreateModel( name='Course', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(blank=True, max_length=100)), ('price', jsonfield.fields.JSONField(blank=True, null=True)), ], ), migrations.AlterField( model_name='daily', name='course', field=models.ForeignKey(on_delete=django.db.models.deletion.DO_NOTHING, to='authorization.Course'), ), migrations.AddField( model_name='group', name='course', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.DO_NOTHING, to='authorization.Course'), ), ]
32.088235
122
0.604033
4e428e1c353f9ae16acebfc45bfab7a9a4bd2704
2,081
py
Python
ssd/modeling/head/ssd_head.py
tkhe/ssd-family
a797ec36fda59549aff54419c105813c33d8cdd3
[ "MIT" ]
1
2019-07-12T02:21:24.000Z
2019-07-12T02:21:24.000Z
ssd/modeling/head/ssd_head.py
tkhe/ssd-family
a797ec36fda59549aff54419c105813c33d8cdd3
[ "MIT" ]
null
null
null
ssd/modeling/head/ssd_head.py
tkhe/ssd-family
a797ec36fda59549aff54419c105813c33d8cdd3
[ "MIT" ]
null
null
null
import torch.nn as nn import torch.nn.functional as F from ssd.modeling.anchor import make_anchor_generator from ssd.utils import bbox from .inference import make_post_processor from .loss import make_loss_evaluator from .predictor import make_ssd_predictor class SSDHead(nn.Module): def __init__(self, cfg, in_channels): super(SSDHead, self).__init__() num_classes = cfg.MODEL.NUM_CLASSES anchors_per_location = [ len(aspect_ratio) * 2 + 2 for aspect_ratio in cfg.MODEL.ANCHOR.ASPECT_RATIOS ] self.predictor = make_ssd_predictor(cfg, in_channels, anchors_per_location, num_classes) self.loss_evaluator = make_loss_evaluator(cfg) self.post_processor = make_post_processor(cfg) self.anchor_generator = make_anchor_generator(cfg) self.center_variance = cfg.MODEL.CENTER_VARIANCE self.size_variance = cfg.MODEL.SIZE_VARIANCE self.size = cfg.INPUT.SIZE def forward(self, features, targets=None): cls_logits, bbox_pred = self.predictor(features) if self.training: return self._forward_train(cls_logits, bbox_pred, targets) else: return self._forward_test(cls_logits, bbox_pred) def _forward_train(self, cls_logits, bbox_pred, targets): gt_boxes, gt_labels = targets[0], targets[1] cls_loss, reg_loss = self.loss_evaluator(cls_logits, bbox_pred, gt_labels, gt_boxes) loss_dict = dict( cls_loss=cls_loss, reg_loss=reg_loss, ) return {}, loss_dict def _forward_test(self, cls_logits, bbox_pred): anchors = self.anchor_generator.generate_anchors() anchors = anchors.to(cls_logits.device) scores = F.softmax(cls_logits, dim=2) boxes = bbox.convert_locations_to_boxes( bbox_pred, anchors, self.center_variance, self.size_variance ) boxes = bbox.xywh2xyxy(boxes) detections = self.post_processor(boxes, scores) return detections, {}
35.87931
96
0.675156
4e42bcb647690572f850059e2f35498edac0af13
415
py
Python
find_max_occurence_simple.py
swatmantis/my-pyscripts
e16af5879b101c30e34e82727292849d1d33f440
[ "Apache-2.0" ]
null
null
null
find_max_occurence_simple.py
swatmantis/my-pyscripts
e16af5879b101c30e34e82727292849d1d33f440
[ "Apache-2.0" ]
null
null
null
find_max_occurence_simple.py
swatmantis/my-pyscripts
e16af5879b101c30e34e82727292849d1d33f440
[ "Apache-2.0" ]
null
null
null
"""Find max element""" #!/usr/bin/env python3 """Find max element""" import random from collections import Counter List = [random.randrange(1, 15) for num in range(10)] def most_frequent(List): occurence_count = Counter(List) return occurence_count.most_common() frequent_number, frequency = most_frequent(List)[0] print(f"List {List}: \nMost frequent number {frequent_number} \nFrequency: {frequency}")
27.666667
88
0.742169
4e4b385ebb874ffc51cb3af951c49e948dbf2c97
1,659
py
Python
plugin.video.SportsDevil/lib/dialogs/dialogProgress.py
akuala/REPO.KUALA
ea9a157025530d2ce8fa0d88431c46c5352e89d4
[ "Apache-2.0" ]
105
2015-11-28T00:03:11.000Z
2021-05-05T20:47:42.000Z
plugin.video.SportsDevil/lib/dialogs/dialogProgress.py
akuala/REPO.KUALA
ea9a157025530d2ce8fa0d88431c46c5352e89d4
[ "Apache-2.0" ]
918
2015-11-28T14:12:40.000Z
2022-03-23T20:24:49.000Z
plugin.video.SportsDevil/lib/dialogs/dialogProgress.py
akuala/REPO.KUALA
ea9a157025530d2ce8fa0d88431c46c5352e89d4
[ "Apache-2.0" ]
111
2015-12-01T14:06:10.000Z
2020-08-01T10:44:39.000Z
# -*- coding: utf-8 -*- import xbmcgui class DialogProgress: def __init__(self): self.dlg = xbmcgui.DialogProgress() self.__reset__() def __reset__(self): self.head = '' self.firstline = '' self.secondline = None self.thirdline = None self.percent = 0 def isCanceled(self): return self.dlg.iscanceled() def update(self, percent=None, firstline=None, secondline=None, thirdline=None): if firstline: self.firstline = firstline if secondline: self.secondline = secondline if thirdline: self.thirdline = thirdline if percent: self.percent = percent if self.secondline and self.thirdline: self.dlg.update(self.percent, self.firstline, self.secondline, self.thirdline) elif self.secondline: self.dlg.update(self.percent, self.firstline, self.secondline) else: self.dlg.update(self.percent, self.firstline) def create(self, head, firstline = None, secondline=None, thirdline=None): if firstline: self.firstline = firstline if secondline: self.secondline = secondline if thirdline: self.thirdline = thirdline if self.secondline and self.thirdline: self.dlg.create(head, self.firstline, self.secondline, self.thirdline) elif self.secondline: self.dlg.create(head, self.firstline, self.secondline) else: self.dlg.create(head, self.firstline) def close(self): self.dlg.close() self.__reset__()
28.603448
90
0.603978
4e4b7d98eca7eba2d20b079df0bbd0eb0b4e7a32
3,828
py
Python
bitbake/lib/bb/manifest.py
KDAB/OpenEmbedded-Archos
a525c5629a57ccb8656c22fe5528ce264003f9d8
[ "MIT" ]
3
2015-05-25T10:56:21.000Z
2021-11-27T17:25:26.000Z
bitbake/lib/bb/manifest.py
KDAB/OpenEmbedded-Archos
a525c5629a57ccb8656c22fe5528ce264003f9d8
[ "MIT" ]
1
2021-11-27T17:24:21.000Z
2021-11-27T17:24:21.000Z
bitbake/lib/bb/manifest.py
KDAB/OpenEmbedded-Archos
a525c5629a57ccb8656c22fe5528ce264003f9d8
[ "MIT" ]
2
2016-08-13T08:40:48.000Z
2021-03-26T03:01:03.000Z
# ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (C) 2003, 2004 Chris Larson # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. import os, sys import bb, bb.data def getfields(line): fields = {} fieldmap = ( "pkg", "src", "dest", "type", "mode", "uid", "gid", "major", "minor", "start", "inc", "count" ) for f in xrange(len(fieldmap)): fields[fieldmap[f]] = None if not line: return None splitline = line.split() if not len(splitline): return None try: for f in xrange(len(fieldmap)): if splitline[f] == '-': continue fields[fieldmap[f]] = splitline[f] except IndexError: pass return fields def parse (mfile, d): manifest = [] while 1: line = mfile.readline() if not line: break if line.startswith("#"): continue fields = getfields(line) if not fields: continue manifest.append(fields) return manifest def emit (func, manifest, d): #str = "%s () {\n" % func str = "" for line in manifest: emittedline = emit_line(func, line, d) if not emittedline: continue str += emittedline + "\n" # str += "}\n" return str def mangle (func, line, d): import copy newline = copy.copy(line) src = bb.data.expand(newline["src"], d) if src: if not os.path.isabs(src): src = "${WORKDIR}/" + src dest = newline["dest"] if not dest: return if dest.startswith("/"): dest = dest[1:] if func is "do_install": dest = "${D}/" + dest elif func is "do_populate": dest = "${WORKDIR}/install/" + newline["pkg"] + "/" + dest elif func is "do_stage": varmap = {} varmap["${bindir}"] = "${STAGING_DIR}/${HOST_SYS}/bin" varmap["${libdir}"] = "${STAGING_DIR}/${HOST_SYS}/lib" varmap["${includedir}"] = "${STAGING_DIR}/${HOST_SYS}/include" varmap["${datadir}"] = "${STAGING_DATADIR}" matched = 0 for key in varmap.keys(): if dest.startswith(key): dest = varmap[key] + "/" + dest[len(key):] matched = 1 if not matched: newline = None return else: newline = None return newline["src"] = src newline["dest"] = dest return newline def emit_line (func, line, d): import copy newline = copy.deepcopy(line) newline = mangle(func, newline, d) if not newline: return None str = "" type = newline["type"] mode = newline["mode"] src = newline["src"] dest = newline["dest"] if type is "d": str = "install -d " if mode: str += "-m %s " % mode str += dest elif type is "f": if not src: return None if dest.endswith("/"): str = "install -d " str += dest + "\n" str += "install " else: str = "install -D " if mode: str += "-m %s " % mode str += src + " " + dest del newline return str
26.4
112
0.544148
4e4e48bf1020755d5adf17a1c4aa85cf738609d6
23,209
py
Python
riglib/bmi/robot_arms.py
sgowda/brain-python-interface
708e2a5229d0496a8ce9de32bda66f0925d366d9
[ "Apache-2.0" ]
7
2015-08-25T00:28:49.000Z
2020-04-14T22:58:51.000Z
riglib/bmi/robot_arms.py
sgowda/brain-python-interface
708e2a5229d0496a8ce9de32bda66f0925d366d9
[ "Apache-2.0" ]
89
2020-08-03T16:54:08.000Z
2022-03-09T19:56:19.000Z
riglib/bmi/robot_arms.py
sgowda/brain-python-interface
708e2a5229d0496a8ce9de32bda66f0925d366d9
[ "Apache-2.0" ]
4
2016-10-05T17:54:26.000Z
2020-08-06T15:37:09.000Z
''' Classes implementing various kinematic chains. This module is perhaps mis-located as it does not have a direct BMI role but rather contains code which is useful in supporting BMI control of kinematic chains. This code depends on the 'robot' module (https://github.com/sgowda/robotics_toolbox) ''' import numpy as np try: import robot except ImportError: import warnings warnings.warn("The 'robot' module cannot be found! See https://github.com/sgowda/robotics_toolbox") import matplotlib.pyplot as plt from collections import OrderedDict import time pi = np.pi class KinematicChain(object): ''' Arbitrary kinematic chain (i.e. spherical joint at the beginning of each joint) ''' def __init__(self, link_lengths=[10., 10.], name='', base_loc=np.array([0., 0., 0.]), rotation_convention=1): ''' Docstring Parameters ---------- link_lengths: iterable Lengths of all the distances between joints base_loc: np.array of shape (3,), default=np.array([0, 0, 0]) Location of the base of the kinematic chain in an "absolute" reference frame ''' self.n_links = len(link_lengths) self.link_lengths = link_lengths self.base_loc = base_loc assert rotation_convention in [-1, 1] self.rotation_convention = rotation_convention # Create the robot object. Override for child classes with different types of joints self._init_serial_link() self.robot.name = name def _init_serial_link(self): links = [] for link_length in self.link_lengths: link1 = robot.Link(alpha=-pi/2) link2 = robot.Link(alpha=pi/2) link3 = robot.Link(d=-link_length) links += [link1, link2, link3] # By convention, we start the arm in the XY-plane links[1].offset = -pi/2 self.robot = robot.SerialLink(links) def calc_full_joint_angles(self, joint_angles): ''' Override in child classes to perform static transforms on joint angle inputs. If some joints are always static (e.g., if the chain only operates in a plane) this can avoid unclutter joint angle specifications. ''' return self.rotation_convention * joint_angles def full_angles_to_subset(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' return joint_angles def plot(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' joint_angles = self.calc_full_joint_angles(joint_angles) self.robot.plot(joint_angles) def forward_kinematics(self, joint_angles, **kwargs): ''' Calculate forward kinematics using D-H parameter convention Parameters ---------- Returns ------- ''' joint_angles = self.calc_full_joint_angles(joint_angles) t, allt = self.robot.fkine(joint_angles, **kwargs) self.joint_angles = joint_angles self.t = t self.allt = allt return t, allt def apply_joint_limits(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' return joint_angles def inverse_kinematics(self, target_pos, q_start=None, method='pso', **kwargs): ''' Docstring Parameters ---------- Returns ------- ''' if q_start == None: q_start = self.random_sample() return self.inverse_kinematics_pso(target_pos, q_start, **kwargs) # ik_method = getattr(self, 'inverse_kinematics_%s' % method) # return ik_method(q_start, target_pos) def inverse_kinematics_grad_descent(self, target_pos, starting_config, n_iter=1000, verbose=False, eps=0.01, return_path=False): ''' Default inverse kinematics method is RRT since for redundant kinematic chains, an infinite number of inverse kinematics solutions exist Docstring Parameters ---------- Returns ------- ''' q = starting_config start_time = time.time() endpoint_traj = np.zeros([n_iter, 3]) joint_limited = np.zeros(len(q)) for k in range(n_iter): # print k # calc endpoint position of the manipulator endpoint_traj[k] = self.endpoint_pos(q) current_cost = np.linalg.norm(endpoint_traj[k] - target_pos, 2) if current_cost < eps: print("Terminating early") break # calculate the jacobian J = self.jacobian(q) J_pos = J[0:3,:] # for joints that are at their limit, zero out the jacobian? # J_pos[:, np.nonzero(self.calc_full_joint_angles(joint_limited))] = 0 # take a step from the current position toward the target pos using the inverse Jacobian J_inv = np.linalg.pinv(J_pos) # J_inv = J_pos.T xdot = (target_pos - endpoint_traj[k])#/np.linalg.norm(endpoint_traj[k] - target_pos) # if current_cost < 3 or k > 10: # stepsize = 0.001 # else: # stepsize = 0.01 xdot = (target_pos - endpoint_traj[k])#/np.linalg.norm(endpoint_traj[k] - target_pos) # xdot = (endpoint_traj[k] - target_pos)/np.linalg.norm(endpoint_traj[k] - target_pos) qdot = 0.001*np.dot(J_inv, xdot) qdot = self.full_angles_to_subset(np.array(qdot).ravel()) q += qdot # apply joint limits q, joint_limited = self.apply_joint_limits(q) end_time = time.time() runtime = end_time - start_time if verbose: print("Runtime: %g" % runtime) print("# of iterations: %g" % k) if return_path: return q, endpoint_traj[:k] else: return q def jacobian(self, joint_angles): ''' Return the full jacobian Docstring Parameters ---------- Returns ------- ''' joint_angles = self.calc_full_joint_angles(joint_angles) J = self.robot.jacobn(joint_angles) return J def endpoint_pos(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' t, allt = self.forward_kinematics(joint_angles) pos_rel_to_base = np.array(t[0:3,-1]).ravel() return pos_rel_to_base + self.base_loc def ik_cost(self, q, q_start, target_pos, weight=100): ''' Docstring Parameters ---------- Returns ------- ''' q_diff = q - q_start return np.linalg.norm(q_diff[0:2]) + weight*np.linalg.norm(self.endpoint_pos(q) - target_pos) def inverse_kinematics_pso(self, target_pos, q_start, time_limit=np.inf, verbose=False, eps=0.5, n_particles=10, n_iter=10): ''' Docstring Parameters ---------- Returns ------- ''' # Initialize the particles; n_joints = self.n_joints particles_q = np.tile(q_start, [n_particles, 1]) # if 0: # # initialize the velocities to be biased around the direction the jacobian tells you is correct # current_pos = self.endpoint_pos(q_start) # int_displ = target_pos - current_pos # print int_displ, target_pos # J = self.jacobian(q_start) # endpoint_vel = np.random.randn(n_particles, 3)# + int_displ # particles_v = np.dot(J[0:3,1::3].T, endpoint_vel.T).T # else: # # initialize particle velocities randomly particles_v = np.random.randn(n_particles, n_joints) #/ np.array([1., 1., 1, 1]) #np.array(self.link_lengths) cost_fn = lambda q: self.ik_cost(q, q_start, target_pos) gbest = particles_q.copy() gbestcost = np.array(list(map(cost_fn, gbest))) pbest = gbest[np.argmin(gbestcost)] pbestcost = cost_fn(pbest) min_limits = np.array([x[0] for x in self.joint_limits]) max_limits = np.array([x[1] for x in self.joint_limits]) min_limits = np.tile(min_limits, [n_particles, 1]) max_limits = np.tile(max_limits, [n_particles, 1]) start_time = time.time() for k in range(n_iter): if time.time() - start_time > time_limit: break # update positions of particles particles_q += particles_v # apply joint limits min_viol = particles_q < min_limits max_viol = particles_q > max_limits particles_q[min_viol] = min_limits[min_viol] particles_q[max_viol] = max_limits[max_viol] # update the costs costs = np.array(list(map(cost_fn, particles_q))) # update the 'bests' gbest[gbestcost > costs] = particles_q[gbestcost > costs] gbestcost[gbestcost > costs] = costs[gbestcost > costs] idx = np.argmin(gbestcost) pbest = gbest[idx] pbestcost = gbestcost[idx] # update the velocity phi1 = 1#np.random.rand() phi2 = 1#np.random.rand() w=0.25 c1=0.5 c2=0.25 particles_v = w*particles_v + c1*phi1*(pbest - particles_q) + c2*phi2*(gbest - particles_q) error = np.linalg.norm(self.endpoint_pos(pbest) - target_pos) if error < eps: break end_time = time.time() if verbose: print("Runtime = %g, error = %g, n_iter=%d" % (end_time-start_time, error, k)) return pbest def spatial_positions_of_joints(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' _, allt = self.forward_kinematics(joint_angles, return_allt=True) pos = (allt[0:3, -1,:].T + self.base_loc).T # pos = np.hstack([np.zeros([3,1]), pos]) return pos class PlanarXZKinematicChain(KinematicChain): ''' Kinematic chain restricted to movement in the XZ-plane ''' def _init_serial_link(self): base = robot.Link(alpha=pi/2, d=0, a=0) links = [base] for link_length in self.link_lengths: link1 = robot.Link(alpha=0, d=0, a=link_length) links.append(link1) # link2 = robot.Link(alpha=pi/2) # link3 = robot.Link(d=-link_length) # links += [link1, link2, link3] # By convention, we start the arm in the XY-plane # links[1].offset = -pi/2 self.robot = robot.SerialLink(links) def calc_full_joint_angles(self, joint_angles): ''' only some joints rotate in the planar kinematic chain Parameters ---------- joint_angles : np.ndarray of shape (self.n_links) Joint angles without the angle for the base link, which is fixed at 0 Returns ------- joint_angles_full : np.ndarray of shape (self.n_links+1) Add on the 0 at the proximal end for the base link angle ''' if not len(joint_angles) == self.n_links: raise ValueError("Incorrect number of joint angles specified!") # # There are really 3 angles per joint to allow 3D rotation at each joint # joint_angles_full = np.zeros(self.n_links * 3) # joint_angles_full[1::3] = joint_angles joint_angles_full = np.hstack([0, joint_angles]) return self.rotation_convention * joint_angles_full def random_sample(self): ''' Sample the joint configuration space within the limits of each joint Parameters ---------- None Returns ------- None ''' if hasattr(self, 'joint_limits'): joint_limits = self.joint_limits else: joint_limits = [(-np.pi, np.pi)] * self.n_links q_start = [] for lim_min, lim_max in joint_limits: q_start.append(np.random.uniform(lim_min, lim_max)) return np.array(q_start) def full_angles_to_subset(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' # return joint_angles[1::3] return joint_angles[1:] def apply_joint_limits(self, joint_angles): ''' Docstring Parameters ---------- Returns ------- ''' if not hasattr(self, 'joint_limits'): return joint_angles else: angles = [] limit_hit = [] for angle, (lim_min, lim_max) in zip(joint_angles, self.joint_limits): limit_hit.append(angle < lim_min or angle > lim_max) angle = max(lim_min, angle) angle = min(angle, lim_max) angles.append(angle) return np.array(angles), np.array(limit_hit) @property def n_joints(self): ''' In a planar arm, the number of joints equals the number of links ''' return len(self.link_lengths) def spatial_positions_of_joints(self, *args, **kwargs): ''' Docstring Parameters ---------- Returns ------- ''' pos_all_joints = super(PlanarXZKinematicChain, self).spatial_positions_of_joints(*args, **kwargs) return pos_all_joints #(pos_all_joints[:,::3].T + self.base_loc).T def create_ik_subchains(self): ''' Docstring Parameters ---------- Returns ------- ''' proximal_link_lengths = self.link_lengths[:2] distal_link_lengths = self.link_lengths[2:] self.proximal_chain = PlanarXZKinematicChain2Link(proximal_link_lengths) if len(self.link_lengths) > 2: self.distal_chain = PlanarXZKinematicChain(distal_link_lengths) else: self.distal_chain = None def inverse_kinematics(self, target_pos, **kwargs): ''' Docstring Parameters ---------- Returns ------- ''' target_pos = target_pos.copy() target_pos -= self.base_loc if not hasattr(self, 'proximal_chain') or not hasattr(self, 'distal_chain'): self.create_ik_subchains() if len(self.link_lengths) > 2: distal_angles = kwargs.pop('distal_angles', None) if distal_angles is None: # Sample randomly from the joint limits (-pi, pi) if not specified if not hasattr(self, 'joint_limits') or len(self.joint_limits) < len(self.link_lengths): joint_limits = [(-pi, pi)] * len(self.distal_chain.link_lengths) else: joint_limits = self.joint_limits[2:] distal_angles = np.array([np.random.uniform(*limits) for limits in joint_limits]) distal_displ = self.distal_chain.endpoint_pos(distal_angles) proximal_endpoint_pos = target_pos - distal_displ proximal_angles = self.proximal_chain.inverse_kinematics(proximal_endpoint_pos).ravel() angles = distal_angles.copy() joint_angles = proximal_angles.tolist() angles[0] -= np.sum(proximal_angles) ik_angles = np.hstack([proximal_angles, angles]) ik_angles = np.array([np.arctan2(np.sin(angle), np.cos(angle)) for angle in ik_angles]) return ik_angles else: return self.proximal_chain.inverse_kinematics(target_pos).ravel() def jacobian(self, theta, old=False): ''' Returns the first derivative of the forward kinematics function for x and z endpoint positions: [[dx/dtheta_1, ..., dx/dtheta_N] [dz/dtheta_1, ..., dz/dtheta_N]] Parameters ---------- theta : np.ndarray of shape (N,) Valid configuration for the arm (the jacobian calculations are specific to the configuration of the arm) Returns ------- J : np.ndarray of shape (2, N) Manipulator jacobian in the format above ''' if old: # Calculate jacobian based on hand calculation specific to this type of chain l = self.link_lengths N = len(theta) J = np.zeros([2, len(l)]) for m in range(N): for i in range(m, N): J[0, m] += -l[i]*np.sin(sum(self.rotation_convention*theta[:i+1])) J[1, m] += l[i]*np.cos(sum(self.rotation_convention*theta[:i+1])) return J else: # Use the robotics toolbox and the generic D-H convention jacobian J = self.robot.jacob0(self.calc_full_joint_angles(theta)) return np.array(J[[0,2], 1:]) def endpoint_potent_null_split(self, q, vel, return_J=False): ''' (Approximately) split joint velocities into an endpoint potent component, which moves the endpoint, and an endpoint null component which only causes self-motion ''' J = self.jacobian(q) J_pinv = np.linalg.pinv(J) J_task = np.dot(J_pinv, J) J_null = np.eye(self.n_joints) - J_task vel_task = np.dot(J_task, vel) vel_null = np.dot(J_null, vel) if return_J: return vel_task, vel_null, J, J_pinv else: return vel_task, vel_null def config_change_nullspace_workspace(self, config1, config2): ''' For two configurations, determine how much joint displacement is in the "null" space and how much is in the "task" space Docstring Parameters ---------- Returns ------- ''' config = config1 vel = config2 - config1 endpt1 = self.endpoint_pos(config1) endpt2 = self.endpoint_pos(config2) task_displ = np.linalg.norm(endpt1 - endpt2) # compute total displ of individual joints total_joint_displ = 0 n_joints = len(config1) for k in range(n_joints): jnt_k_vel = np.zeros(n_joints) jnt_k_vel[k] = vel[k] single_joint_displ_pos = self.endpoint_pos(config + jnt_k_vel) total_joint_displ += np.linalg.norm(endpt1 - single_joint_displ_pos) return task_displ, total_joint_displ def detect_collision(self, theta, obstacle_pos): ''' Detect a collision between the chain and a circular object ''' spatial_joint_pos = self.spatial_positions_of_joints(theta).T + self.base_loc plant_segments = [(x, y) for x, y in zip(spatial_joint_pos[:-1], spatial_joint_pos[1:])] dist_to_object = np.zeros(len(plant_segments)) for k, segment in enumerate(plant_segments): dist_to_object[k] = point_to_line_segment_distance(obstacle_pos, segment) return dist_to_object def plot_joint_pos(self, joint_pos, ax=None, flip=False, **kwargs): if ax == None: plt.figure() ax = plt.subplot(111) if isinstance(joint_pos, dict): joint_pos = np.vstack(list(joint_pos.values())) elif isinstance(joint_pos, np.ndarray) and np.ndim(joint_pos) == 1: joint_pos = joint_pos.reshape(1, -1) elif isinstance(joint_pos, tuple): joint_pos = np.array(joint_pos).reshape(1, -1) for pos in joint_pos: spatial_pos = self.spatial_positions_of_joints(pos).T shoulder_anchor = np.array([2., 0., -15.]) spatial_pos = spatial_pos# + shoulder_anchor if flip: ax.plot(-spatial_pos[:,0], spatial_pos[:,2], **kwargs) else: ax.plot(spatial_pos[:,0], spatial_pos[:,2], **kwargs) return ax def point_to_line_segment_distance(point, segment): ''' Determine the distance between a point and a line segment. Used to determine collisions between robot arm links and virtual obstacles. Adapted from http://stackoverflow.com/questions/849211/shortest-distance-between-a-point-and-a-line-segment ''' v, w = segment l2 = np.sum(np.abs(v - w)**2) if l2 == 0: return np.linalg.norm(v - point) t = np.dot(point - v, w - v)/l2 if t < 0: return np.linalg.norm(point - v) elif t > 1: return np.linalg.norm(point - w) else: projection = v + t*(w-v) return np.linalg.norm(projection - point) class PlanarXZKinematicChain2Link(PlanarXZKinematicChain): ''' Docstring ''' def __init__(self, link_lengths, *args, **kwargs): ''' Docstring Parameters ---------- Returns ------- ''' if not len(link_lengths) == 2: raise ValueError("Can't instantiate a 2-link arm with > 2 links!") super(PlanarXZKinematicChain2Link, self).__init__(link_lengths, *args, **kwargs) def inverse_kinematics(self, pos, **kwargs): ''' Inverse kinematics for a two-link kinematic chain. These equations can be solved deterministically. Docstring Parameters ---------- pos : np.ndarray of shape (3,) Desired endpoint position where the coordinate system origin is the base of the arm. y coordinate must be 0 Returns ------- np.ndarray of shape (2,) Joint angles which yield the endpoint position with the forward kinematics of this manipulator ''' pos -= self.base_loc l_upperarm, l_forearm = self.link_lengths if np.ndim(pos) == 1: pos = pos.reshape(1,-1) # require the y-coordinate to be 0, i.e. flat on the screen x, y, z = pos[:,0], pos[:,1], pos[:,2] assert np.all(np.abs(np.array(y)) < 1e-10) L = np.sqrt(x**2 + z**2) cos_el_pflex = (L**2 - l_forearm**2 - l_upperarm**2) / (2*l_forearm*l_upperarm) cos_el_pflex[ (cos_el_pflex > 1) & (cos_el_pflex < 1 + 1e-9)] = 1 el_pflex = np.arccos(cos_el_pflex) sh_pabd = np.arctan2(z, x) - np.arcsin(l_forearm * np.sin(np.pi - el_pflex) / L) return np.array([sh_pabd, el_pflex])
32.190014
138
0.561463
4e4e7f677ab2a0f132b93e4b1dfb1c29e362f6de
3,622
py
Python
src/utils/tilemap.py
Magicalbat/Metroidvania-Month-15
a0a30fb3f531a597ced69bf76568ef26e5e88019
[ "MIT" ]
null
null
null
src/utils/tilemap.py
Magicalbat/Metroidvania-Month-15
a0a30fb3f531a597ced69bf76568ef26e5e88019
[ "MIT" ]
null
null
null
src/utils/tilemap.py
Magicalbat/Metroidvania-Month-15
a0a30fb3f531a597ced69bf76568ef26e5e88019
[ "MIT" ]
null
null
null
import pygame from pygame.math import Vector2 import json, math class Tilemap: def __init__(self, tileSize, imgs, chunkSize=8): self.tileSize = tileSize self.imgs = imgs self.drawTiles = [] self.chunks = {} self.chunkSize = chunkSize def toChunkScale(self, p): return math.floor(p/self.tileSize/self.chunkSize) def toChunkPos(self, p): return (self.toChunkScale(p[0]), self.toChunkScale(p[1])) def collidePoint(self, p:Vector2): cp = self.toChunkPos(p) if cp in self.chunks: for rect in self.chunks[cp]: if rect.collidepoint(p): return True return False def _getColRects(self, testPointsX, testPointsY, colRects): minX = self.toChunkScale(min(testPointsX)) maxX = self.toChunkScale(max(testPointsX)) minY = self.toChunkScale(min(testPointsY)) maxY = self.toChunkScale(max(testPointsY)) testChunkPositions = { (minX, minY), (minX, maxY), (maxX, minY), (maxX, maxY) } if colRects is None: colRects = [] for pos in testChunkPositions: if pos in self.chunks: colRects += self.chunks[pos] return colRects def getEntityColRects(self, pos, width, height, vel, colRects=None): return self._getColRects( \ ( pos.x, pos.x + width, pos.x + vel.x, pos.x + width + vel.x ), ( pos.y, pos.y + height, pos.y + vel.y, pos.y + height + vel.y ), \ colRects) def getRectColRects(self, rect, colRects=None): return self._getColRects((rect.x, rect.right), (rect.y, rect.bottom), colRects) def draw(self, win, scroll=None): if scroll is None: scroll = Vector2(0, 0) winDim = win.get_size() for layer in self.drawTiles: for tile in layer: if tile[0] < winDim[0]+scroll.x and tile[0] > scroll.x-self.tileSize and \ tile[1] < winDim[1]+scroll.y and tile[1] > scroll.y-self.tileSize: win.blit(self.imgs[tile[2]], (tile[0] - scroll.x, tile[1] - scroll.y)) def drawCollision(self, win, scroll=None): if scroll is None: scroll = Vector2(0, 0) cols = ((255,0,0), (0,255,0), (0,0,255)) for pos, rects in self.chunks.items(): pygame.draw.rect(win, (255,255,255), \ (pos[0] * self.tileSize * self.chunkSize - scroll.x,\ pos[1] * self.tileSize * self.chunkSize - scroll.y,\ self.tileSize * self.chunkSize, self.tileSize * self.chunkSize), 1) for i, rect in enumerate(rects): pygame.draw.rect(win, cols[i%len(cols)], \ (rect.x - scroll.x, rect.y - scroll.y, \ rect.w, rect.h), width=1) def loadLevel(self, filepath): with open(filepath, 'r') as f: data = json.loads(f.read()) for layer in data["drawTiles"]: tempLayer = [] for key, item in layer.items(): pStr = key.split(';') x, y = int(pStr[0]), int(pStr[1]) tempLayer.append((x*self.tileSize, y*self.tileSize, item)) self.drawTiles.append(tempLayer) for pos, rects in data["chunks"].items(): tempRects = [] for rect in rects: tempRects.append(pygame.Rect(rect)) pStr = pos.split(';') self.chunks[(int(pStr[0]), int(pStr[1]))] = tempRects if "extraData" in data: return data["extraData"]
38.126316
90
0.549144
4e508c95181eba9329a23ec0f597dadfe33c7e09
7,295
py
Python
src/dsnt/util.py
anibali/dsnt-pose2d
f453331a6b120f02948336555b996ac0d95bf4be
[ "Apache-2.0" ]
12
2018-10-18T06:41:00.000Z
2021-07-31T08:19:41.000Z
src/dsnt/util.py
anibali/dsnt-pose2d
f453331a6b120f02948336555b996ac0d95bf4be
[ "Apache-2.0" ]
2
2019-07-15T13:36:08.000Z
2020-03-09T04:39:08.000Z
src/dsnt/util.py
anibali/dsnt-pose2d
f453331a6b120f02948336555b996ac0d95bf4be
[ "Apache-2.0" ]
5
2019-01-08T01:32:18.000Z
2020-08-04T07:42:12.000Z
""" Miscellaneous utility functions. """ import random import time from contextlib import contextmanager import math import numpy as np import torch from PIL.ImageDraw import Draw # Joints to connect for visualisation, giving the effect of drawing a # basic "skeleton" of the pose. BONES = { 'right_lower_leg': (0, 1), 'right_upper_leg': (1, 2), 'right_pelvis': (2, 6), 'left_lower_leg': (4, 5), 'left_upper_leg': (3, 4), 'left_pelvis': (3, 6), 'center_lower_torso': (6, 7), 'center_upper_torso': (7, 8), 'center_head': (8, 9), 'right_lower_arm': (10, 11), 'right_upper_arm': (11, 12), 'right_shoulder': (12, 8), 'left_lower_arm': (14, 15), 'left_upper_arm': (13, 14), 'left_shoulder': (13, 8), } def draw_skeleton(img, coords, joint_mask=None): '''Draw a pose skeleton connecting joints (for visualisation purposes). Left-hand-side joints are connected with blue lines. Right-hand-size joints are connected with red lines. Center joints are connected with magenta lines. Args: img (PIL.Image.Image): PIL image which the skeleton will be drawn over. coords (Tensor): 16x2 tensor containing 0-based pixel coordinates of joint locations. Joints indices are expected to match http://human-pose.mpi-inf.mpg.de/#download joint_mask (Tensor, optional): Mask of valid joints (invalid joints will be drawn with grey lines). ''' draw = Draw(img) for bone_name, (j1, j2) in BONES.items(): if bone_name.startswith('center_'): colour = (255, 0, 255) # Magenta elif bone_name.startswith('left_'): colour = (0, 0, 255) # Blue elif bone_name.startswith('right_'): colour = (255, 0, 0) # Red else: colour = (255, 255, 255) if joint_mask is not None: # Change colour to grey if either vertex is not masked in if joint_mask[j1] == 0 or joint_mask[j2] == 0: colour = (100, 100, 100) draw.line([coords[j1, 0], coords[j1, 1], coords[j2, 0], coords[j2, 1]], fill=colour) def draw_gaussian(img_tensor, x, y, sigma, normalize=False, clip_size=None): '''Draw a Gaussian in a single-channel 2D image. Args: img_tensor: Image tensor to draw to. x: x-coordinate of Gaussian centre (in pixels). y: y-coordinate of Gaussian centre (in pixels). sigma: Standard deviation of Gaussian (in pixels). normalize: Ensures values sum to 1 when True. clip_size: Restrict the size of the draw region. ''' # To me it makes more sense to round() these, but hey - I'm just following the example # of others. x = int(x) y = int(y) if img_tensor.dim() == 2: height, width = list(img_tensor.size()) elif img_tensor.dim() == 3: n_chans, height, width = list(img_tensor.size()) assert n_chans == 1, 'expected img_tensor to have one channel' img_tensor = img_tensor[0] else: raise Exception('expected img_tensor to have 2 or 3 dimensions') radius = max(width, height) if clip_size is not None: radius = clip_size / 2 if radius < 0.5 or x <= -radius or y <= -radius or \ x >= (width - 1) + radius or y >= (height - 1) + radius: return start_x = max(0, math.ceil(x - radius)) end_x = min(width, int(x + radius + 1)) start_y = max(0, math.ceil(y - radius)) end_y = min(height, int(y + radius + 1)) w = end_x - start_x h = end_y - start_y subimg = img_tensor[start_y:end_y, start_x:end_x] xs = torch.arange(start_x, end_x).type_as(img_tensor).view(1, w).expand_as(subimg) ys = torch.arange(start_y, end_y).type_as(img_tensor).view(h, 1).expand_as(subimg) k = -0.5 * (1 / sigma)**2 subimg.copy_((xs - x)**2) subimg.add_((ys - y)**2) subimg.mul_(k) subimg.exp_() if normalize: val_sum = subimg.sum() if val_sum > 0: subimg.div_(val_sum) def encode_heatmaps(coords, width, height, sigma=1): '''Convert normalised coordinates into heatmaps.''' # Normalised coordinates to pixel coordinates coords.add_(1) coords[:, :, 0].mul_(width / 2) coords[:, :, 1].mul_(height / 2) coords.add_(-0.5) batch_size = coords.size(0) n_chans = coords.size(1) target = torch.FloatTensor(batch_size, n_chans, height, width).zero_() for i in range(batch_size): for j in range(n_chans): x = round(coords[i, j, 0]) y = round(coords[i, j, 1]) draw_gaussian(target[i, j], x, y, sigma, normalize=False, clip_size=7) return target def get_preds(heatmaps): batch_size, n_chans, height, width = list(heatmaps.size()) maxval, idx = torch.max(heatmaps.view(batch_size, n_chans, -1), 2) maxval = maxval.view(batch_size, n_chans, 1) idx = idx.view(batch_size, n_chans, 1) coords = idx.repeat(1, 1, 2) coords[:, :, 0] = coords[:, :, 0] % width coords[:, :, 1] = coords[:, :, 1] / height coords = coords.float() # When maxval is zero, select coords (0, 0) pred_mask = maxval.gt(0).repeat(1, 1, 2).float() torch.mul(coords, pred_mask, out=coords) return coords def decode_heatmaps(heatmaps, use_neighbours=True): '''Convert heatmaps into normalised coordinates.''' coords = get_preds(heatmaps) _, _, height, width = list(heatmaps.size()) if use_neighbours: # "To improve performance at high precision thresholds the prediction # is offset by a quarter of a pixel in the direction of its next highest # neighbor before transforming back to the original coordinate space # of the image" # - Stacked Hourglass Networks for Human Pose Estimation for i, joint_coords in enumerate(coords): for j, (x, y) in enumerate(joint_coords): x = int(x) y = int(y) if x > 0 and x < width - 1 and y > 0 and y < height - 1: hm = heatmaps[i, j] joint_coords[j, 0] += (0.25 * np.sign(hm[y, x + 1] - hm[y, x - 1])) joint_coords[j, 1] += (0.25 * np.sign(hm[y + 1, x] - hm[y - 1, x])) # Pixel coordinates to normalised coordinates coords.add_(0.5) coords[:, :, 0].mul_(2 / width) coords[:, :, 1].mul_(2 / height) coords.add_(-1) return coords def type_as_index(indices, tensor): if tensor.is_cuda: return indices.type(torch.cuda.LongTensor) return indices.type(torch.LongTensor) def reverse_tensor(tensor, dim): indices = torch.arange(tensor.size(dim) - 1, -1, -1) indices = type_as_index(indices, tensor) return tensor.index_select(dim, indices) @contextmanager def timer(meter): start_time = time.perf_counter() yield time_elapsed = time.perf_counter() - start_time meter.add(time_elapsed) def generator_timer(generator, meter): while True: with timer(meter): vals = next(generator) yield vals def seed_random_number_generators(seed): """Seed all random number generators.""" random.seed(seed) np.random.seed(seed) torch.manual_seed(seed) torch.cuda.manual_seed_all(seed)
31.042553
92
0.615216
4e50c9eaddb3cc1ea6331eb13dca8d92f32d04fe
1,321
py
Python
lect01_codes/lect01_eg/multiprocess_eg/mp_main.py
radiumweilei/chinahadoop-python-ai-2
f45271b073f99b9c46de150aa87bcf4adc5feca2
[ "Apache-2.0" ]
null
null
null
lect01_codes/lect01_eg/multiprocess_eg/mp_main.py
radiumweilei/chinahadoop-python-ai-2
f45271b073f99b9c46de150aa87bcf4adc5feca2
[ "Apache-2.0" ]
null
null
null
lect01_codes/lect01_eg/multiprocess_eg/mp_main.py
radiumweilei/chinahadoop-python-ai-2
f45271b073f99b9c46de150aa87bcf4adc5feca2
[ "Apache-2.0" ]
1
2019-11-11T09:42:06.000Z
2019-11-11T09:42:06.000Z
import os import time from datetime import datetime from multiprocessing import Process, Pool def run_proc(n): print('第{}次循环,子进程id:{},父进程id:{}'.format(n, os.getpid(), os.getppid())) time.sleep(1) if __name__ == '__main__': print('父进程id', os.getpid()) # 1. 顺序执行任务 # start = datetime.now() # for i in range(10): # run_proc(i) # print('耗时:', datetime.now() - start) # 2. 多进程并行执行 # 2.1 多进程异步并行执行,进程间没有先后顺序 # start = datetime.now() # for i in range(10): # p = Process(target=run_proc, args=(i,)) # p.start() # print('耗时:', datetime.now() - start) # 2.2 多进程同步并行执行,进程间有先后顺序 # start = datetime.now() # for i in range(10): # p = Process(target=run_proc, args=(i,)) # p.start() # p.join() # print('耗时:', datetime.now() - start) # 3. 进程池管理多进程 # 3.1 使用Pool管理多个进程,同步执行 # pool = Pool() # start = datetime.now() # for i in range(10): # pool.apply(func=run_proc, args=(i,)) # pool.close() # pool.join() # print('耗时:', datetime.now() - start) # 3.2 使用Pool管理多个进程,异步执行 # pool = Pool() # start = datetime.now() # for i in range(10): # pool.apply_async(func=run_proc, args=(i,)) # pool.close() # pool.join() # print('耗时:', datetime.now() - start)
23.175439
74
0.551098
4e51a3f60a853dcb91ad39c536974879ba250f9f
268
py
Python
strings_12/tests/test_change_case.py
njoroge33/py_learn
6ad55f37789045bc5c03f3dd668cf1ea497f4e84
[ "MIT" ]
null
null
null
strings_12/tests/test_change_case.py
njoroge33/py_learn
6ad55f37789045bc5c03f3dd668cf1ea497f4e84
[ "MIT" ]
2
2019-04-15T06:29:55.000Z
2019-04-19T17:34:32.000Z
strings_12/tests/test_change_case.py
njoroge33/py_learn
6ad55f37789045bc5c03f3dd668cf1ea497f4e84
[ "MIT" ]
1
2019-11-19T04:51:18.000Z
2019-11-19T04:51:18.000Z
import pytest from ..change_case import change_case @pytest.mark.parametrize("x_str, expected", [ ("BLACKstar", "blackSTAR"), ("jOhn", "JoHN") ]) def test_change_case(x_str, expected): actual = change_case(x_str) assert actual == expected
22.333333
46
0.664179
4e533f314e5c3e66781f51a4229383e5a116f3ac
803
py
Python
Curso_Python_3_UDEMY/POO/desafio_carro.py
DanilooSilva/Cursos_de_Python
8f167a4c6e16f01601e23b6f107578aa1454472d
[ "MIT" ]
null
null
null
Curso_Python_3_UDEMY/POO/desafio_carro.py
DanilooSilva/Cursos_de_Python
8f167a4c6e16f01601e23b6f107578aa1454472d
[ "MIT" ]
null
null
null
Curso_Python_3_UDEMY/POO/desafio_carro.py
DanilooSilva/Cursos_de_Python
8f167a4c6e16f01601e23b6f107578aa1454472d
[ "MIT" ]
null
null
null
class Carro: def __init__(self, velocidadeMaxima): self.velocidadeMaxima = velocidadeMaxima self.velocidade = 0 def acelerar(self, delta=5): if self.velocidade < self.velocidadeMaxima: self.velocidade += delta else: self.velocidade = 180 return self.velocidade if self.velocidade <= self.velocidadeMaxima else 180 def frear(self, delta=20): if self.velocidade > 0: self.velocidade -= delta else: self.velocidade = 0 return self.velocidade if self.velocidade >= 0 else 0 if __name__ == '__main__': c1 = Carro(180) for _ in range(25): print(f'Acelerando {c1.acelerar(8)}') for _ in range(10): print(f' reduzindo a velocidade {c1.frear(delta=20)}')
28.678571
83
0.608966
4e55091a618968c01ac26471c9cd251dd97a71d7
9,578
py
Python
argocd_client/models/application_application_sync_request.py
thepabloaguilar/argocd-client
a6c4ff268a63ee6715f9f837b9225b798aa6bde2
[ "BSD-3-Clause" ]
1
2021-09-29T11:57:07.000Z
2021-09-29T11:57:07.000Z
argocd_client/models/application_application_sync_request.py
thepabloaguilar/argocd-client
a6c4ff268a63ee6715f9f837b9225b798aa6bde2
[ "BSD-3-Clause" ]
1
2020-09-09T00:28:57.000Z
2020-09-09T00:28:57.000Z
argocd_client/models/application_application_sync_request.py
thepabloaguilar/argocd-client
a6c4ff268a63ee6715f9f837b9225b798aa6bde2
[ "BSD-3-Clause" ]
2
2020-10-13T18:31:59.000Z
2021-02-15T12:52:33.000Z
# coding: utf-8 """ Consolidate Services Description of all APIs # noqa: E501 The version of the OpenAPI document: version not set Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six from argocd_client.configuration import Configuration class ApplicationApplicationSyncRequest(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { 'dry_run': 'bool', 'infos': 'list[V1alpha1Info]', 'manifests': 'list[str]', 'name': 'str', 'prune': 'bool', 'resources': 'list[V1alpha1SyncOperationResource]', 'retry_strategy': 'V1alpha1RetryStrategy', 'revision': 'str', 'strategy': 'V1alpha1SyncStrategy' } attribute_map = { 'dry_run': 'dryRun', 'infos': 'infos', 'manifests': 'manifests', 'name': 'name', 'prune': 'prune', 'resources': 'resources', 'retry_strategy': 'retryStrategy', 'revision': 'revision', 'strategy': 'strategy' } def __init__(self, dry_run=None, infos=None, manifests=None, name=None, prune=None, resources=None, retry_strategy=None, revision=None, strategy=None, local_vars_configuration=None): # noqa: E501 """ApplicationApplicationSyncRequest - a model defined in OpenAPI""" # noqa: E501 if local_vars_configuration is None: local_vars_configuration = Configuration() self.local_vars_configuration = local_vars_configuration self._dry_run = None self._infos = None self._manifests = None self._name = None self._prune = None self._resources = None self._retry_strategy = None self._revision = None self._strategy = None self.discriminator = None if dry_run is not None: self.dry_run = dry_run if infos is not None: self.infos = infos if manifests is not None: self.manifests = manifests if name is not None: self.name = name if prune is not None: self.prune = prune if resources is not None: self.resources = resources if retry_strategy is not None: self.retry_strategy = retry_strategy if revision is not None: self.revision = revision if strategy is not None: self.strategy = strategy @property def dry_run(self): """Gets the dry_run of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The dry_run of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: bool """ return self._dry_run @dry_run.setter def dry_run(self, dry_run): """Sets the dry_run of this ApplicationApplicationSyncRequest. :param dry_run: The dry_run of this ApplicationApplicationSyncRequest. # noqa: E501 :type: bool """ self._dry_run = dry_run @property def infos(self): """Gets the infos of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The infos of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: list[V1alpha1Info] """ return self._infos @infos.setter def infos(self, infos): """Sets the infos of this ApplicationApplicationSyncRequest. :param infos: The infos of this ApplicationApplicationSyncRequest. # noqa: E501 :type: list[V1alpha1Info] """ self._infos = infos @property def manifests(self): """Gets the manifests of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The manifests of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: list[str] """ return self._manifests @manifests.setter def manifests(self, manifests): """Sets the manifests of this ApplicationApplicationSyncRequest. :param manifests: The manifests of this ApplicationApplicationSyncRequest. # noqa: E501 :type: list[str] """ self._manifests = manifests @property def name(self): """Gets the name of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The name of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: str """ return self._name @name.setter def name(self, name): """Sets the name of this ApplicationApplicationSyncRequest. :param name: The name of this ApplicationApplicationSyncRequest. # noqa: E501 :type: str """ self._name = name @property def prune(self): """Gets the prune of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The prune of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: bool """ return self._prune @prune.setter def prune(self, prune): """Sets the prune of this ApplicationApplicationSyncRequest. :param prune: The prune of this ApplicationApplicationSyncRequest. # noqa: E501 :type: bool """ self._prune = prune @property def resources(self): """Gets the resources of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The resources of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: list[V1alpha1SyncOperationResource] """ return self._resources @resources.setter def resources(self, resources): """Sets the resources of this ApplicationApplicationSyncRequest. :param resources: The resources of this ApplicationApplicationSyncRequest. # noqa: E501 :type: list[V1alpha1SyncOperationResource] """ self._resources = resources @property def retry_strategy(self): """Gets the retry_strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The retry_strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: V1alpha1RetryStrategy """ return self._retry_strategy @retry_strategy.setter def retry_strategy(self, retry_strategy): """Sets the retry_strategy of this ApplicationApplicationSyncRequest. :param retry_strategy: The retry_strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :type: V1alpha1RetryStrategy """ self._retry_strategy = retry_strategy @property def revision(self): """Gets the revision of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The revision of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: str """ return self._revision @revision.setter def revision(self, revision): """Sets the revision of this ApplicationApplicationSyncRequest. :param revision: The revision of this ApplicationApplicationSyncRequest. # noqa: E501 :type: str """ self._revision = revision @property def strategy(self): """Gets the strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :return: The strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :rtype: V1alpha1SyncStrategy """ return self._strategy @strategy.setter def strategy(self, strategy): """Sets the strategy of this ApplicationApplicationSyncRequest. :param strategy: The strategy of this ApplicationApplicationSyncRequest. # noqa: E501 :type: V1alpha1SyncStrategy """ self._strategy = strategy def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, ApplicationApplicationSyncRequest): return False return self.to_dict() == other.to_dict() def __ne__(self, other): """Returns true if both objects are not equal""" if not isinstance(other, ApplicationApplicationSyncRequest): return True return self.to_dict() != other.to_dict()
29.112462
200
0.618292
4e5750ff5296717b749da87c576b380ec5a0ca38
1,818
py
Python
kedro/extras/decorators/retry_node.py
hfwittmann/kedro
b0d4fcd8f19b49a7916d78fd09daeb6209a7b6c6
[ "Apache-2.0" ]
1
2021-11-25T12:33:13.000Z
2021-11-25T12:33:13.000Z
kedro/extras/decorators/retry_node.py
MerelTheisenQB/kedro
1eaa2e0fa5d80f96e18ea60b9f3d6e6efc161827
[ "Apache-2.0" ]
null
null
null
kedro/extras/decorators/retry_node.py
MerelTheisenQB/kedro
1eaa2e0fa5d80f96e18ea60b9f3d6e6efc161827
[ "Apache-2.0" ]
null
null
null
""" This module contains the retry decorator, which can be used as ``Node`` decorators to retry nodes. See ``kedro.pipeline.node.decorate`` """ import logging from functools import wraps from time import sleep from typing import Callable, Type def retry( exceptions: Type[Exception] = Exception, n_times: int = 1, delay_sec: float = 0 ) -> Callable: """ Catches exceptions from the wrapped function at most n_times and then bundles and propagates them. **Make sure your function does not mutate the arguments** Args: exceptions: The superclass of exceptions to catch. By default catch all exceptions. n_times: At most let the function fail n_times. The bundle the errors and propagate them. By default retry only once. delay_sec: Delay between failure and next retry in seconds Returns: The original function with retry functionality. """ def _retry(func: Callable): @wraps(func) def _wrapper(*args, **kwargs): counter = n_times errors = [] while counter >= 0: try: return func(*args, **kwargs) # pylint: disable=broad-except except exceptions as exc: errors.append(exc) if counter != 0: sleep(delay_sec) counter -= 1 if errors: log = logging.getLogger(__name__) log.error( "Function `%s` failed %i times. Errors:\n", func.__name__, n_times ) log.error("\n".join(str(err) for err in errors)) log.error("Raising last exception") raise errors[-1] return _wrapper return _retry
30.3
86
0.573707
4e57932c1bf27e86e563c5240b4f42764bb1b0f4
1,470
py
Python
test/lmp/dset/_base/test_download_file.py
ProFatXuanAll/char-RNN
531f101b3d1ba20bafd28ca060aafe6f583d1efb
[ "Beerware" ]
null
null
null
test/lmp/dset/_base/test_download_file.py
ProFatXuanAll/char-RNN
531f101b3d1ba20bafd28ca060aafe6f583d1efb
[ "Beerware" ]
null
null
null
test/lmp/dset/_base/test_download_file.py
ProFatXuanAll/char-RNN
531f101b3d1ba20bafd28ca060aafe6f583d1efb
[ "Beerware" ]
null
null
null
"""Test the ability to download files. Test target: - :py:meth:`lmp.dset._base.BaseDset.download`. """ import os from typing import Callable import pytest import lmp.dset._base import lmp.util.path @pytest.fixture def file_url() -> str: """Download target file URL.""" return 'https://raw.githubusercontent.com/ProFatXuanAll/language-model-playground/main/README.rst' @pytest.fixture def file_path(clean_dir_finalizer_factory: Callable[[str], None], exp_name: str, file_url: str, request) -> str: """Download file path. After testing, clean up files and directories created during test. """ # Create temporary directory. abs_dir_path = os.path.join(lmp.util.path.DATA_PATH, exp_name) if not os.path.exists(abs_dir_path): os.makedirs(abs_dir_path) abs_file_path = os.path.join(abs_dir_path, file_url.split(r'/')[-1]) request.addfinalizer(clean_dir_finalizer_factory(abs_dir_path)) return abs_file_path def test_download_as_text_file(file_path: str, file_url: str) -> None: """Must be able to download file and output as text file.""" lmp.dset._base.BaseDset.download_file(mode='text', download_path=file_path, url=file_url) assert os.path.exists(file_path) def test_download_as_binary_file(file_path: str, file_url: str) -> None: """Must be able to download file and output as binary file.""" lmp.dset._base.BaseDset.download_file(mode='binary', download_path=file_path, url=file_url) assert os.path.exists(file_path)
30
112
0.755782
4e584909a422d8166030333d2adb063c0ced43a9
1,019
py
Python
server/account/models.py
istommao/fakedataset
365ef0c68d1ecac30ab2c9908e6a5efa1da5d81e
[ "MIT" ]
null
null
null
server/account/models.py
istommao/fakedataset
365ef0c68d1ecac30ab2c9908e6a5efa1da5d81e
[ "MIT" ]
null
null
null
server/account/models.py
istommao/fakedataset
365ef0c68d1ecac30ab2c9908e6a5efa1da5d81e
[ "MIT" ]
null
null
null
"""account models.""" from django.contrib.auth.hashers import ( check_password, make_password ) from django.db import models from extension.modelutils import RandomFixedCharField class Account(models.Model): uid = RandomFixedCharField('编号', max_length=32, unique=True) username = models.CharField('用户名', max_length=32, unique=True) password = models.CharField('密码', max_length=80) create_time = models.DateTimeField('创建时间', auto_now_add=True) update_time = models.DateTimeField('修改时间', auto_now=True) def __str__(self): return self.username class Meta: """Meta.""" verbose_name = '用户' verbose_name_plural = '用户' def set_password(self, raw_password): self.password = make_password(raw_password) def check_password(self, raw_password): def setter(raw_password): self.set_password(raw_password) self.save(update_fields=['password']) return check_password(raw_password, self.password, setter)
28.305556
66
0.693817
4e59aa0341bec9390bf565218344a25a6e72bf84
2,420
py
Python
tests/views/test_delete.py
fvalverd/AutoApi
3ceb320fe6a36d24032df121e335a8470fb929af
[ "MIT" ]
6
2015-04-28T13:03:04.000Z
2021-08-24T19:15:53.000Z
tests/views/test_delete.py
fvalverd/AutoApi
3ceb320fe6a36d24032df121e335a8470fb929af
[ "MIT" ]
6
2017-06-19T20:59:10.000Z
2020-05-22T16:22:28.000Z
tests/views/test_delete.py
fvalverd/AutoApi
3ceb320fe6a36d24032df121e335a8470fb929af
[ "MIT" ]
2
2015-11-10T14:38:39.000Z
2017-05-18T05:46:03.000Z
# -*- coding: utf-8 -*- import json import unittest from .. import MoviesTest class TestDeleteResource(MoviesTest): def setUp(self): self.setUpClass() super(TestDeleteResource, self).setUp() def test_delete_invalid_id(self): response = self.app.delete('/%s/movies/a1' % self.api, headers=self.headers) self.assertEqual(response.status_code, 409) self.assertDictContainsSubset({'message': u'Resource "a1" is invalid'}, json.loads(response.data)) def test_delete_not_found_id(self): response = self.app.delete('/%s/actors/%s' % (self.api, self.movies[0]['id']), headers=self.headers) self.assertEqual(response.status_code, 404) response_json = json.loads(response.data) self.assertDictContainsSubset( {'message': u'Resource "%s" not found' % self.movies[0]['id']}, response_json ) def test_delete(self): response = self.app.delete('/%s/movies/%s' % (self.api, self.movies[0]['id']), headers=self.headers) self.assertEqual(response.status_code, 204) response_json = self._count_test('movies', 2) self.assertNotIn(self.movies[0], response_json.get('items')) def test_delete_nested(self): response = self.app.delete( '/%s/actors/%s/movies/%s' % (self.api, self.actors[0]['id'], self.movies[0]['id']), headers=self.headers ) self.assertEqual(response.status_code, 204) response_json = self._count_test('movies', 2) self.assertNotIn(self.movies[0], response_json.get('items')) class TestDeleteCollection(MoviesTest): def setUp(self): self.setUpClass() super(TestDeleteCollection, self).setUp() def test_delete(self): response = self.app.delete('/%s/movies' % self.api, headers=self.headers) self.assertEqual(response.status_code, 204) response_json = self._count_test('movies', 0) self.assertEqual([], response_json.get('items')) def test_delete_nested(self): response = self.app.delete( '/%s/actors/%s/movies' % (self.api, self.actors[0]['id']), headers=self.headers ) self.assertEqual(response.status_code, 204) response_json = self._count_test('movies', 2) self.assertNotIn(self.movies[0], response_json.get('items')) if __name__ == '__main__': unittest.main()
35.588235
108
0.638843
4e5ac8d618f9e77a2b39df9c4c03557f518e532c
1,347
py
Python
src/api/urls.py
Karim-Valeev/django-myfoods
e8750a05461616a2e7740230177a139749daac73
[ "MIT" ]
null
null
null
src/api/urls.py
Karim-Valeev/django-myfoods
e8750a05461616a2e7740230177a139749daac73
[ "MIT" ]
null
null
null
src/api/urls.py
Karim-Valeev/django-myfoods
e8750a05461616a2e7740230177a139749daac73
[ "MIT" ]
null
null
null
from django.urls import path, re_path from drf_yasg import openapi from drf_yasg.views import get_schema_view from rest_framework.routers import SimpleRouter, DefaultRouter from rest_framework_simplejwt import views as jwt_views from api.views import * # роутер нужен, чтобы сгенерить урлы под вью сет и самому их не прописывать соотвественно router = SimpleRouter() router.register("baskets", BasketViewSet, "baskets") schema_view = get_schema_view( openapi.Info( title="Snippets API", default_version="v1", description="Test description", terms_of_service="https://www.google.com/policies/terms/", contact=openapi.Contact(email="contact@snippets.local"), license=openapi.License(name="BSD License"), ), public=True, ) urlpatterns = [ path("check/", check_api_view, name="check-api"), path("token/", jwt_views.TokenObtainPairView.as_view(), name="token-obtain-pair"), path("token/refresh/", jwt_views.TokenRefreshView.as_view(), name="token-refresh"), *router.urls, re_path(r"swagger(?P<format>\.json|\.yaml)$", schema_view.without_ui(cache_timeout=0), name="schema-json"), path("swagger/", schema_view.with_ui("swagger", cache_timeout=0), name="schema-swagger-ui"), path("redoc/", schema_view.with_ui("redoc", cache_timeout=0), name="schema-redoc"), ]
38.485714
111
0.723831
4e5b531d4dae58f3f455001978beda2d6160593c
18,417
py
Python
second/pytorch/inference_ros.py
neolixcn/nutonomy_pointpillars
03f46f6de97c0c97d7bc98d7af3daee215d81a30
[ "MIT" ]
1
2021-06-11T00:54:48.000Z
2021-06-11T00:54:48.000Z
second/pytorch/inference_ros.py
neolixcn/nutonomy_pointpillars
03f46f6de97c0c97d7bc98d7af3daee215d81a30
[ "MIT" ]
null
null
null
second/pytorch/inference_ros.py
neolixcn/nutonomy_pointpillars
03f46f6de97c0c97d7bc98d7af3daee215d81a30
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import os import argparse import pathlib import pickle import shutil import time from functools import partial import sys sys.path.append('../') from pathlib import Path import fire import numpy as np import torch import torch.nn as nn import os print(torch.__version__) print(os.environ['PYTHONPATH']) from google.protobuf import text_format import rospy from sensor_msgs.msg import PointCloud2 import sensor_msgs.point_cloud2 as pc2 from std_msgs.msg import Header from jsk_recognition_msgs.msg import BoundingBox, BoundingBoxArray import torchplus import second.data.kitti_common as kitti from second.builder import target_assigner_builder, voxel_builder from second.data.preprocess import merge_second_batch from second.protos import pipeline_pb2 from second.pytorch.builder import (box_coder_builder, input_reader_builder, lr_scheduler_builder, optimizer_builder, second_builder) from second.utils.eval import get_coco_eval_result, get_official_eval_result from second.utils.progress_bar import ProgressBar def get_paddings_indicator(actual_num, max_num, axis=0): """ Create boolean mask by actually number of a padded tensor. :param actual_num: :param max_num: :param axis: :return: [type]: [description] """ actual_num = torch.unsqueeze(actual_num, axis+1) max_num_shape = [1] * len(actual_num.shape) max_num_shape[axis+1] = -1 max_num = torch.arange(max_num, dtype=torch.int, device=actual_num.device).view(max_num_shape) # tiled_actual_num : [N, M, 1] # tiled_actual_num : [[3,3,3,3,3], [4,4,4,4,4], [2,2,2,2,2]] # title_max_num : [[0,1,2,3,4], [0,1,2,3,4], [0,1,2,3,4]] paddings_indicator = actual_num.int() > max_num # paddings_indicator shape : [batch_size, max_num] return paddings_indicator def _get_pos_neg_loss(cls_loss, labels): # cls_loss: [N, num_anchors, num_class] # labels: [N, num_anchors] batch_size = cls_loss.shape[0] if cls_loss.shape[-1] == 1 or len(cls_loss.shape) == 2: cls_pos_loss = (labels > 0).type_as(cls_loss) * cls_loss.view( batch_size, -1) cls_neg_loss = (labels == 0).type_as(cls_loss) * cls_loss.view( batch_size, -1) cls_pos_loss = cls_pos_loss.sum() / batch_size cls_neg_loss = cls_neg_loss.sum() / batch_size else: cls_pos_loss = cls_loss[..., 1:].sum() / batch_size cls_neg_loss = cls_loss[..., 0].sum() / batch_size return cls_pos_loss, cls_neg_loss def _flat_nested_json_dict(json_dict, flatted, sep=".", start=""): for k, v in json_dict.items(): if isinstance(v, dict): _flat_nested_json_dict(v, flatted, sep, start + sep + k) else: flatted[start + sep + k] = v def flat_nested_json_dict(json_dict, sep=".") -> dict: """flat a nested json-like dict. this function make shadow copy. """ flatted = {} for k, v in json_dict.items(): if isinstance(v, dict): _flat_nested_json_dict(v, flatted, sep, k) else: flatted[k] = v return flatted def example_convert_to_torch(example, dtype=torch.float32, device=None) -> dict: # device = device or torch.device("cuda:0") example_torch = {} # float_names = ["voxels", "anchors", "reg_targets", "reg_weights", "bev_map", "rect", "Trv2c", "P2"] float_names = ["voxels", "anchors", "reg_targets", "reg_weights", "bev_map"] for k, v in example.items(): if k in float_names: example_torch[k] = torch.as_tensor(v, dtype=dtype).cuda() elif k in ["coordinates", "labels", "num_points"]: example_torch[k] = torch.as_tensor(v, dtype=torch.int32).cuda() elif k in ["anchors_mask"]: example_torch[k] = torch.as_tensor(v, dtype=torch.uint8).cuda() # torch.uint8 is now deprecated, please use a dtype torch.bool instead else: example_torch[k] = v return example_torch def _make_point_field(num_field): msg_pf1 = pc2.PointField() msg_pf1.name = np.str('x') msg_pf1.offset = np.uint32(0) msg_pf1.datatype = np.uint8(7) msg_pf1.count = np.uint32(1) msg_pf2 = pc2.PointField() msg_pf2.name = np.str('y') msg_pf2.offset = np.uint32(4) msg_pf2.datatype = np.uint8(7) msg_pf2.count = np.uint32(1) msg_pf3 = pc2.PointField() msg_pf3.name = np.str('z') msg_pf3.offset = np.uint32(8) msg_pf3.datatype = np.uint8(7) msg_pf3.count = np.uint32(1) msg_pf4 = pc2.PointField() msg_pf4.name = np.str('intensity') msg_pf4.offset = np.uint32(16) msg_pf4.datatype = np.uint8(7) msg_pf4.count = np.uint32(1) if num_field == 4: return [msg_pf1, msg_pf2, msg_pf3, msg_pf4] msg_pf5 = pc2.PointField() msg_pf5.name = np.str('label') msg_pf5.offset = np.uint32(20) msg_pf5.datatype = np.uint8(4) msg_pf5.count = np.uint32(1) return [msg_pf1, msg_pf2, msg_pf3, msg_pf4, msg_pf5] def publish_test(np_p_ranged, frame_id): header = Header() header.stamp = rospy.Time() header.frame_id = frame_id x = np_p_ranged[:, 0].reshape(-1) y = np_p_ranged[:, 1].reshape(-1) z = np_p_ranged[:, 2].reshape(-1) if np_p_ranged.shape[1] == 4: i = np_p_ranged[:, 3].reshape(-1) else: i = np.zeros(np_p_ranged.shape[0], 1).reshape(-1) cloud = np.stack((x, y, z, i)) msg_segment = pc2.create_cloud(header=header, fields=_make_point_field(4), points=cloud.T) pub_points.publish(msg_segment) def predict_kitti_to_anno(net, example, class_names, center_limit_range=None, lidar_input=False, global_set=None): # eval example : [0: 'voxels', 1: 'num_points', 2: 'coordinates', 3: 'rect' # 4: 'Trv2c', 5: 'P2', 6: 'anchors', 7: 'anchors_mask' # 8: 'image_idx', 9: 'image_shape'] # eval example [0: 'voxels', 1: 'num_points', 2: 'coordinate', 3: 'anchors', # 4: 'anchor_mask', 5: 'pc_idx'] pillar_x = example[0][:, :, 0].unsqueeze(0).unsqueeze(0) pillar_y = example[0][:, :, 1].unsqueeze(0).unsqueeze(0) pillar_z = example[0][:, :, 2].unsqueeze(0).unsqueeze(0) pillar_i = example[0][:, :, 3].unsqueeze(0).unsqueeze(0) num_points_per_pillar = example[1].float().unsqueeze(0) # Find distance of x, y, and z from pillar center # assuming xyres_16.proto coors_x = example[2][:, 3].float() coors_y = example[2][:, 2].float() x_sub = coors_x.unsqueeze(1) * 0.16 -22.96 #+ 0.08#-22.96#+ 0.08#-22.96#-19.76 y_sub = coors_y.unsqueeze(1) * 0.16 -22.96#- 19.76 #-22.96#-19.76#-22.96#-19.76 ones = torch.ones([1, 100], dtype=torch.float32, device=pillar_x.device) x_sub_shaped = torch.mm(x_sub, ones).unsqueeze(0).unsqueeze(0) y_sub_shaped = torch.mm(y_sub, ones).unsqueeze(0).unsqueeze(0) num_points_for_a_pillar = pillar_x.size()[3] mask = get_paddings_indicator(num_points_per_pillar, num_points_for_a_pillar, axis=0) mask = mask.permute(0, 2, 1) mask = mask.unsqueeze(1) mask = mask.type_as(pillar_x) coors = example[2] anchors = example[3] anchors_mask = example[4] anchors_mask = torch.as_tensor(anchors_mask, dtype=torch.uint8, device=pillar_x.device) anchors_mask = anchors_mask.byte() # rect = example[3] # Trv2c = example[4] # P2 = example[5] pc_idx = example[5] input = [pillar_x, pillar_y, pillar_z, pillar_i, num_points_per_pillar, x_sub_shaped, y_sub_shaped, mask, coors, anchors, anchors_mask, pc_idx] predictions_dicts = net(input) # lidar_box, final_score, label_preds, pc_idx annos = [] for i, preds_dict in enumerate(predictions_dicts): # image_shape = batch_image_shape[i] pc_idx = preds_dict[3] if preds_dict[0] is not None: # bbox list # box_2d_preds = preds_dict[0].detach().cpu().numpy() # bbox # box_preds = preds_dict[1].detach().cpu().numpy() # bbox3d_camera scores = preds_dict[1].detach().cpu().numpy() # scores box_preds_lidar = preds_dict[0].detach().cpu().numpy() # box3d_lidar # write pred to file label_preds = preds_dict[2].detach().cpu().numpy() # label_preds anno = kitti.get_start_result_anno() num_example = 0 content = '' for box_lidar, score, label in zip( box_preds_lidar, scores, label_preds): if center_limit_range is not None: limit_range = np.array(center_limit_range) if (np.any(box_lidar[:3] < limit_range[:3]) or np.any(box_lidar[:3] > limit_range[3:])): continue content += str(label) + " 0.0 0 0.0 0.0 0.0 0.0 0.0 " + str(box_lidar[5]) + " " + str(box_lidar[3]) + " "\ + str(box_lidar[4]) + " " + str(box_lidar[0]) + " " + str(box_lidar[1]) + " " + str(box_lidar[2]) + " " + str(box_lidar[6]) + " " + str(score) + "\n" anno["name"].append(class_names[int(label)]) anno["truncated"].append(0.0) anno["occluded"].append(0) anno["alpha"].append(-np.arctan2(-box_lidar[1], box_lidar[0]) + box_lidar[6]) anno["bbox"].append(np.array([0, 0, 0, 0])) anno["dimensions"].append([box_lidar[4], box_lidar[5], box_lidar[3]]) # annotate by shl # anno["dimensions"].append(box_lidar[3:6]) anno["location"].append(box_lidar[:3]) anno["rotation_y"].append(box_lidar[6]) if global_set is not None: for i in range(100000): if score in global_set: score -= 1 / 100000 else: global_set.add(score) break anno["score"].append(score) num_example += 1 content = content.strip() def delete_nan_points(points): new_poins = np.array([]) print(points) for i in range(points.shape[0]): if (np.isnan(points[i][0])) | (np.isnan(points[i][1])) | (np.isnan(points[i][2])): pass else: np.row_stack((new_poins, points[i])) return new_poins def callback(msg): arr_bbox = BoundingBoxArray() # pcl_msg = pc2.read_points(msg, skip_nans=False, field_names=("x", "y", "z", "intensity", "ring")) pcl_msg = pc2.read_points(msg, skip_nans=False, field_names=("x", "y", "z")) # pcl_msg = pc2.read_points(msg, skip_nans=True) # print(pcl_msg) np_p = np.array(list(pcl_msg), dtype=np.float32) np_p = np.column_stack((np_p, np.zeros((np_p.shape[0], 1)))) print(np_p) # np_p = np.delete(np_p, -1, 1) # delete "ring" field # np_p = delete_nan_points(np_p) eval_dataset = input_reader_builder.build( input_cfg, model_cfg, training=False, voxel_generator=voxel_generator, target_assigner=target_assigner, inference=True, points=np_p) eval_dataloader = torch.utils.data.DataLoader( eval_dataset, batch_size=input_cfg.batch_size, shuffle=False, num_workers=input_cfg.num_workers, pin_memory=False, collate_fn=merge_second_batch) net.eval() global_set = None eval_data = iter(eval_dataloader) example = next(eval_data) example = example_convert_to_torch(example, torch.float32) example_tuple = list(example.values()) example_tuple[5] = torch.from_numpy(example_tuple[5]) result = predict_kitti_to_anno(net, example_tuple, class_names, center_limit_range, model_cfg.lidar_input, global_set) print("result", result) # def evaluate(config_path, # model_dir, # result_path=None, # predict_test=False, # ckpt_path=None, # ref_detfile=None, # pickle_result=True, # read_predict_pkl_path=None): # # model_dir = str(Path(model_dir).resolve()) # if predict_test: # result_name = 'predict_test' # else: # result_name = 'eval_results' # if result_path is None: # model_dir = Path(model_dir) # result_path = model_dir / result_name # else: # result_path = pathlib.Path(result_path) # # if isinstance(config_path, str): # config = pipeline_pb2.TrainEvalPipelineConfig() # with open(config_path, "r") as f: # proto_str = f.read() # text_format.Merge(proto_str, config) # else: # config = config_path # # input_cfg = config.eval_input_reader # model_cfg = config.model.second # train_cfg = config.train_config # class_names = list(input_cfg.class_names) # center_limit_range = model_cfg.post_center_limit_range # ######################### # # Build Voxel Generator # ######################### # voxel_generator = voxel_builder.build(model_cfg.voxel_generator) # bv_range = voxel_generator.point_cloud_range[[0, 1, 3, 4]] # box_coder = box_coder_builder.build(model_cfg.box_coder) # target_assigner_cfg = model_cfg.target_assigner # target_assigner = target_assigner_builder.build(target_assigner_cfg, # bv_range, box_coder) # # net = second_builder.build(model_cfg, voxel_generator, target_assigner, input_cfg.batch_size) # net.cuda() # if train_cfg.enable_mixed_precision: # net.half() # net.metrics_to_float() # net.convert_norm_to_float(net) # # if ckpt_path is None: # torchplus.train.try_restore_latest_checkpoints(model_dir, [net]) # else: # torchplus.train.restore(ckpt_path, net) # # eval_dataset = input_reader_builder.build( # input_cfg, # model_cfg, # training=False, # voxel_generator=voxel_generator, # target_assigner=target_assigner) # # eval_dataloader = torch.utils.data.DataLoader( # eval_dataset, # batch_size=input_cfg.batch_size, # shuffle=False, # num_workers=input_cfg.num_workers, # pin_memory=False, # collate_fn=merge_second_batch) # # if train_cfg.enable_mixed_precision: # float_dtype = torch.float16 # else: # float_dtype = torch.float32 # # net.eval() # result_path_step = result_path / f"step_{net.get_global_step()}" # result_path_step.mkdir(parents=True, exist_ok=True) # t = time.time() # dt_annos = [] # global_set = None # eval_data = iter(eval_dataloader) # example = next(eval_data) # example = example_convert_to_torch(example, float_dtype) # example_tuple = list(example.values()) # example_tuple[5] = torch.from_numpy(example_tuple[5]) # if (example_tuple[3].size()[0] != input_cfg.batch_size): # continue # # dt_annos += predict_kitti_to_anno( # net, example_tuple, class_names, center_limit_range, # model_cfg.lidar_input, global_set) # for example in iter(eval_dataloader): # # eval example [0: 'voxels', 1: 'num_points', 2: 'coordinates', 3: 'rect' # # 4: 'Trv2c', 5: 'P2', 6: 'anchors', 7: 'anchors_mask' # # 8: 'image_idx', 9: 'image_shape'] # # # eval example [0: 'voxels', 1: 'num_points', 2: 'coordinate', 3: 'anchors', # # 4: 'anchor_mask', 5: 'pc_idx'] # example = example_convert_to_torch(example, float_dtype) # # eval example [0: 'voxels', 1: 'num_points', 2: 'coordinate', 3: 'anchors', # # 4: 'anchor_mask', 5: 'pc_idx'] # # example_tuple = list(example.values()) # example_tuple[5] = torch.from_numpy(example_tuple[5]) # # example_tuple[9] = torch.from_numpy(example_tuple[9]) # # if (example_tuple[3].size()[0] != input_cfg.batch_size): # continue # # dt_annos += predict_kitti_to_anno( # net, example_tuple, class_names, center_limit_range, # model_cfg.lidar_input, global_set) if __name__ == '__main__': parser = argparse.ArgumentParser(description='testing') args = parser.parse_args() model_dir = "/nfs/nas/model/songhongli/neolix_shanghai_3828/" config_path = "/home/songhongli/Projects/pointpillars2/second/configs/pointpillars/xyres_16_4cls.proto" if isinstance(config_path, str): config = pipeline_pb2.TrainEvalPipelineConfig() with open(config_path, "r") as f: proto_str = f.read() text_format.Merge(proto_str, config) else: config = config_path input_cfg = config.eval_input_reader model_cfg = config.model.second train_cfg = config.train_config class_names = list(input_cfg.class_names) center_limit_range = model_cfg.post_center_limit_range ######################### # Build Voxel Generator ######################### voxel_generator = voxel_builder.build(model_cfg.voxel_generator) bv_range = voxel_generator.point_cloud_range[[0, 1, 3, 4]] box_coder = box_coder_builder.build(model_cfg.box_coder) target_assigner_cfg = model_cfg.target_assigner target_assigner = target_assigner_builder.build(target_assigner_cfg, bv_range, box_coder) net = second_builder.build(model_cfg, voxel_generator, target_assigner, input_cfg.batch_size) net.cuda() torchplus.train.try_restore_latest_checkpoints(model_dir, [net]) # code added for using ROS rospy.init_node('pointpillars_ros_node') sub_ = rospy.Subscriber("/sensor/velodyne16/all/compensator/PointCloud2", PointCloud2, callback, queue_size=1) pub_points = rospy.Publisher("points_modified", PointCloud2, queue_size=1) pub_arr_bbox = rospy.Publisher("pre_arr_bbox", BoundingBoxArray, queue_size=10) # pub_bbox = rospy.Publisher("voxelnet_bbox", BoundingBox, queue_size=1) print("[+] voxelnet_ros_node has started!") rospy.spin()
37.432927
176
0.617147
4e5c7dba2e2083dcb5bc4c5689df3f572c63510f
3,112
py
Python
agent.py
AdamMiltonBarker/TassAI
61ae4f208f06ea39cc5b58079175f17bf1fca4c4
[ "MIT" ]
1
2021-06-29T09:46:47.000Z
2021-06-29T09:46:47.000Z
agent.py
AdamMiltonBarker/TassAI
61ae4f208f06ea39cc5b58079175f17bf1fca4c4
[ "MIT" ]
4
2021-06-27T16:06:43.000Z
2021-06-27T16:09:53.000Z
agent.py
AdamMiltonBarker/TassAI
61ae4f208f06ea39cc5b58079175f17bf1fca4c4
[ "MIT" ]
2
2020-09-28T02:11:43.000Z
2020-10-13T15:27:41.000Z
#!/usr/bin/env python3 """ HIAS TassAI Facial Recognition Agent. HIAS TassAI Facial Recognition Agent processes streams from local or remote cameras to identify known and unknown humans. MIT License Copyright (c) 2021 Asociación de Investigacion en Inteligencia Artificial Para la Leucemia Peter Moss Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files(the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and / or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. Contributors: - Adam Milton-Barker """ import sys from abc import ABC, abstractmethod from modules.AbstractAgent import AbstractAgent from modules.helpers import helpers from modules.model import model from modules.read import read from modules.stream import stream from modules.sockets import sockets from threading import Thread class agent(AbstractAgent): """ HIAS TassAI Facial Recognition Agent HIAS TassAI Facial Recognition Agent processes streams from local or remote cameras to identify known and unknown humans. """ def set_model(self, mtype): # Inititializes the TassAI model self.model = model(self.helpers) def load_model(self): """ Loads the trained model """ # Prepares the network and data self.model.prepare_network() self.model.prepare_data() def server(self): """ Loads the API server """ # Starts the MQTT connection self.mqtt_start() # Inititializes the socket self.sockets = sockets(self.helpers) # Loads the TassAI model self.load_model() # Camera read and stream threads Thread(target=read.run, args=(self, ), daemon=True).start() Thread(target=stream.run, args=(self, ), daemon=True).start() def signal_handler(self, signal, frame): self.helpers.logger.info("Disconnecting") self.mqtt.disconnect() sys.exit(1) agent = agent() def main(): if len(sys.argv) < 2: agent.helpers.logger.info( "You must provide an argument") exit() elif sys.argv[1] not in agent.helpers.confs["agent"]["params"]: agent.helpers.logger.info( "Mode not supported! server, train or inference") exit() mode = sys.argv[1] if mode == "classify": agent.set_model("") agent.inference() elif mode == "server": agent.set_model("") agent.server() if __name__ == "__main__": main()
26.151261
78
0.754499
4e5f6c409675e74bac8adf5ea0c951c284a25d25
181
py
Python
asyncy/constants/LineConstants.py
rashmi43/platform-engine
dd9a22742bc8dc43a530ea5edef39b3c35db57c1
[ "Apache-2.0" ]
null
null
null
asyncy/constants/LineConstants.py
rashmi43/platform-engine
dd9a22742bc8dc43a530ea5edef39b3c35db57c1
[ "Apache-2.0" ]
null
null
null
asyncy/constants/LineConstants.py
rashmi43/platform-engine
dd9a22742bc8dc43a530ea5edef39b3c35db57c1
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- class LineConstants: method = 'method' output = 'output' enter = 'enter' parent = 'parent' command = 'command' service = 'service'
16.454545
23
0.563536
4e604e0888e3f4c9cd3d2b535fdc4b7f1eabfe77
2,576
py
Python
Payload_Types/apfell/mythic/agent_functions/terminals_send.py
xorrior/Mythic
ea348b66e1d96e88e0e7fbabff182945cbdf12b6
[ "BSD-3-Clause" ]
2
2021-01-28T19:35:46.000Z
2021-04-08T12:01:48.000Z
Payload_Types/apfell/mythic/agent_functions/terminals_send.py
xorrior/Mythic
ea348b66e1d96e88e0e7fbabff182945cbdf12b6
[ "BSD-3-Clause" ]
null
null
null
Payload_Types/apfell/mythic/agent_functions/terminals_send.py
xorrior/Mythic
ea348b66e1d96e88e0e7fbabff182945cbdf12b6
[ "BSD-3-Clause" ]
2
2020-12-29T02:34:13.000Z
2021-06-24T04:07:38.000Z
from CommandBase import * import json from MythicResponseRPC import * class TerminalsSendArguments(TaskArguments): def __init__(self, command_line): super().__init__(command_line) self.args = { "window": CommandParameter( name="window", type=ParameterType.Number, description="window # to send command to", ), "tab": CommandParameter( name="tab", type=ParameterType.Number, description="tab # to send command to", ), "command": CommandParameter( name="command", type=ParameterType.String, description="command to execute", ), } async def parse_arguments(self): if len(self.command_line) > 0: if self.command_line[0] == "{": self.load_args_from_json_string(self.command_line) else: raise ValueError("Missing JSON arguments") else: raise ValueError("Missing arguments") class TerminalsSendCommand(CommandBase): cmd = "terminals_send" needs_admin = False help_cmd = "terminals_send" description = """ This uses AppleEvents to inject the shell command, {command}, into the specified terminal shell as if the user typed it from the keyboard. This is pretty powerful. Consider the instance where the user is SSH-ed into another machine via terminal - with this you can inject commands to run on the remote host. Just remember, the user will be able to see the command, but you can always see what they see as well with the "terminals_read contents" command. """ version = 1 is_exit = False is_file_browse = False is_process_list = False is_download_file = False is_remove_file = False is_upload_file = False author = "@its_a_feature_" attackmapping = ["T1059", "T1184"] argument_class = TerminalsSendArguments async def create_tasking(self, task: MythicTask) -> MythicTask: resp = await MythicResponseRPC(task).register_artifact( artifact_instance="{}".format( task.args.get_arg("command"), ), artifact_type="Process Create", ) resp = await MythicResponseRPC(task).register_artifact( artifact_instance="Target Application of Terminal", artifact_type="AppleEvent Sent", ) return task async def process_response(self, response: AgentResponse): pass
36.8
457
0.622671
4e6138843998a3ede92abaaa70a1ae3fc7c18aae
711
py
Python
losses.py
dhaulagiri0/AniGen
bd845a29e771544ade1f64b94f967d8e178952f8
[ "MIT" ]
null
null
null
losses.py
dhaulagiri0/AniGen
bd845a29e771544ade1f64b94f967d8e178952f8
[ "MIT" ]
null
null
null
losses.py
dhaulagiri0/AniGen
bd845a29e771544ade1f64b94f967d8e178952f8
[ "MIT" ]
null
null
null
import tensorflow as tf from tensorflow.keras import backend #DEPRECATED # An implementation of wasserstein used for a naive implementation of WGAN # calculate wasserstein loss def wasserstein_loss(y_true, y_pred): return backend.mean(y_true * y_pred) # Define the loss functions for the discriminator, # which should be (fake_loss - real_loss). # We will add the gradient penalty later to this loss function. def discriminator_loss(real_img, fake_img): real_loss = tf.reduce_mean(real_img) fake_loss = tf.reduce_mean(fake_img) return real_loss, fake_loss, fake_loss - real_loss # Define the loss functions for the generator. def generator_loss(fake_img): return -tf.reduce_mean(fake_img)
32.318182
74
0.78481
4e6142fd70771c11fbb624c19a0644bc6c708693
623
py
Python
mephisto/plugins/math_expr.py
Kenton1989/mephisto-bot
50a8008c99b984a453713f480fa578bf5a8353c8
[ "MIT" ]
null
null
null
mephisto/plugins/math_expr.py
Kenton1989/mephisto-bot
50a8008c99b984a453713f480fa578bf5a8353c8
[ "MIT" ]
null
null
null
mephisto/plugins/math_expr.py
Kenton1989/mephisto-bot
50a8008c99b984a453713f480fa578bf5a8353c8
[ "MIT" ]
null
null
null
import re import math import numexpr as ne MATH_CONST = { 'pi': math.pi, 'π': math.pi, 'e': math.e, 'inf': math.inf, 'i': 1j, 'j': 1j, } SUB_MAP = { # replace UTF char with ASCII char '(': '(', ')': ')', ',': ',', '-': '-', '÷': '/', '×': '*', '+': '+', # replace common synonym 'ln': 'log', 'lg': 'log10', '∞': 'inf', 'mod': '%', } SUB_RE = re.compile('|'.join(re.escape(s) for s in SUB_MAP.keys())) def evaluate(txt: str): txt = SUB_RE.sub(lambda m: SUB_MAP[m.group(0)], txt) return ne.evaluate(txt, local_dict=MATH_CONST).item()
16.394737
67
0.473515
4e61d7b3b7f328b277a5ef816c4995021aeb1703
1,185
py
Python
testemu/client/testemu_client/network.py
advaoptical/netemu
a418503d3829f206602e9360c05235626fa8bec5
[ "Apache-2.0" ]
null
null
null
testemu/client/testemu_client/network.py
advaoptical/netemu
a418503d3829f206602e9360c05235626fa8bec5
[ "Apache-2.0" ]
null
null
null
testemu/client/testemu_client/network.py
advaoptical/netemu
a418503d3829f206602e9360c05235626fa8bec5
[ "Apache-2.0" ]
null
null
null
from collections import Mapping from . import yang_models class Meta(type): class Interface( yang_models .com_adva_netemu_testemu_client_TestInterface_YangListModel): """ Pythonizer for Java class ``TestInterface``. From Java package ``com.adva.netemu.testemu.client`` """ class TestNetwork( yang_models.com_adva_netemu_testemu_client_TestNetwork_YangModel, metaclass=Meta): """ Pythonizer for Java class ``TestNetwork``. From Java package ``com.adva.netemu.testemu.client`` """ @property def interfaces(self): class List(Mapping): @staticmethod def __len__(): return len(self._java_object.getInterfaces()) @staticmethod def __iter__(): for intf in self._java_object.getInterfaces(): yield type(self).Interface(intf) @staticmethod def __getitem__(name): for intf in List.__iter__(): if intf.name() == name: return intf raise KeyError(name) return List()
23.7
73
0.571308
4e61d7b801c1c3cd496fc2afd8c46c182f86ceda
666
py
Python
asym_rlpo/representations/identity.py
abaisero/asym-porl
8a76d920e51d783bbeeeea3cd2b02efffbb33c72
[ "MIT" ]
2
2021-08-24T22:41:36.000Z
2021-10-31T01:55:37.000Z
asym_rlpo/representations/identity.py
abaisero/asym-porl
8a76d920e51d783bbeeeea3cd2b02efffbb33c72
[ "MIT" ]
null
null
null
asym_rlpo/representations/identity.py
abaisero/asym-porl
8a76d920e51d783bbeeeea3cd2b02efffbb33c72
[ "MIT" ]
1
2021-10-13T12:27:40.000Z
2021-10-13T12:27:40.000Z
import gym import torch from asym_rlpo.utils.debugging import checkraise from .base import Representation class IdentityRepresentation(Representation): def __init__(self, input_space: gym.spaces.Box): super().__init__() checkraise( isinstance(input_space, gym.spaces.Box) and len(input_space.shape) == 1, TypeError, 'input_space must be Box', ) (self.__out_dim,) = input_space.shape @property def dim(self): return self.__out_dim def forward( # pylint: disable=no-self-use self, inputs: torch.Tensor ) -> torch.Tensor: return inputs
22.2
52
0.630631
4e630265553913112faaae0a442558c6d77373c7
8,885
py
Python
src/ggplib/db/lookup.py
richemslie/ggplib
8388678f311db4a9906d8a3aff71d3f0037b623b
[ "MIT" ]
11
2019-03-02T13:49:07.000Z
2021-12-21T17:03:05.000Z
src/ggplib/db/lookup.py
ggplib/ggplib
8388678f311db4a9906d8a3aff71d3f0037b623b
[ "MIT" ]
2
2019-05-15T18:23:50.000Z
2019-05-19T08:13:19.000Z
src/ggplib/db/lookup.py
ggplib/ggplib
8388678f311db4a9906d8a3aff71d3f0037b623b
[ "MIT" ]
1
2020-04-02T17:35:35.000Z
2020-04-02T17:35:35.000Z
import sys import traceback from ggplib.util import log from ggplib.statemachine import builder from ggplib.db import signature class GameInfo(object): def __init__(self, game, gdl_str): self.game = game self.gdl_str = gdl_str # might be None, depends on whether we grab it from sig.json self.idx = None # lazy loads in get_symbol_map() self.sigs = None self.symbol_map = None # lazy loads self.sm = None self.model = None def get_symbol_map(self): if self.sigs is None: idx, self.sigs = signature.get_index(self.gdl_str, verbose=False) if self.idx is not None: assert self.idx == idx else: self.idx = idx self.symbol_map = signature.build_symbol_map(self.sigs, verbose=False) def lazy_load(self, the_game_store): if self.sm is None: # ok here we can cache the game XXX self.model, self.sm = builder.build_sm(self.gdl_str, the_game_store=the_game_store, add_to_game_store=True) log.verbose("Lazy loading done for %s" % self.game) def get_sm(self): return self.sm.dupe() class TempGameInfo(object): def __init__(self, game, gdl_str, sm, model): self.game = game self.gdl_str = gdl_str self.sm = sm self.model = model def get_sm(self): return self.sm.dupe() class GameInfoBypass(GameInfo): ''' bypass everything, special case statemachine that doesn't have any GDL ''' special_game = True def __init__(self, game, sm, model): self.game = game self.sm = sm self.model = model def get_symbol_map(self): pass def lazy_load(self, the_game_store): pass def get_sm(self): return self.sm.dupe() ############################################################################### class LookupFailed(Exception): pass class GameDatabase: def __init__(self, root_store): self.root_store = root_store self.rulesheets_store = root_store.get_directory("rulesheets") self.games_store = root_store.get_directory("games", create=True) self.idx_mapping = {} self.game_mapping = {} @property def all_games(self): return self.game_mapping.keys() def load(self, verbose=True): if verbose: log.info("Building the database") filenames = self.rulesheets_store.listdir("*.kif") for fn in sorted(filenames): # skip tmp files if fn.startswith("tmp"): continue game = fn.replace(".kif", "") # get the gdl gdl_str = self.rulesheets_store.load_contents(fn) info = GameInfo(game, gdl_str) # first does the game directory exist? the_game_store = self.games_store.get_directory(game, create=True) if the_game_store.file_exists("sig.json"): info.idx = the_game_store.load_json("sig.json")['idx'] else: if verbose: log.verbose("Creating signature for %s" % game) info.get_symbol_map() if info.symbol_map is None: log.warning("FAILED to add: %s" % game) raise Exception("FAILED TO add %s" % game) # save as json assert info.idx is not None the_game_store.save_json("sig.json", dict(idx=info.idx)) assert info.idx is not None if info.idx in self.idx_mapping: other_info = self.idx_mapping[info.idx] log.warning("DUPE GAMES: %s %s!=%s" % (info.idx, game, other_info.game)) raise Exception("Dupes not allowed in database") self.idx_mapping[info.idx] = info self.game_mapping[info.game] = info def get_by_name(self, name): if name not in self.game_mapping: raise LookupFailed("Did not find game: %s" % name) info = self.game_mapping[name] if getattr(info, "special_game", False): return info # for side effects info.get_symbol_map() the_game_store = self.games_store.get_directory(name) info.lazy_load(the_game_store) return info def lookup(self, gdl_str): idx, sig = signature.get_index(gdl_str, verbose=False) if idx not in self.idx_mapping: raise LookupFailed("Did not find game : %s" % idx) info = self.idx_mapping[idx] info.get_symbol_map() # create the symbol map for this gdl_str symbol_map = signature.build_symbol_map(sig, verbose=False) new_mapping = {} # remap the roles back roles = info.sigs.roles.items() for ii in range(len(roles)): match = "role%d" % ii for k1, v1 in roles: if v1 == match: for k2, v2 in sig.roles.items(): if v2 == match: new_mapping[k2] = k1 break # remap the other symbols for k1, v1 in info.symbol_map.items(): new_mapping[symbol_map[k1]] = v1 # remove if the keys/values all the same in new_mapping all_same = True for k, v in new_mapping.items(): if k != v: all_same = False break if all_same: new_mapping = None # log.info("Lookup - found game %s in database" % info.game) the_game_store = self.games_store.get_directory(info.game) info.lazy_load(the_game_store) return info, new_mapping ############################################################################### def install_draughts(add_game): ' load custom c++ statemachine for draughts ' from ggplib import interface from ggplib.non_gdl_games.draughts import desc, model desc10 = desc.BoardDesc(10) cpp_statemachines = interface.CppStateMachines() model = model.create_sm_model(desc10) for game_variant in ["draughts_10x10", "draughts_killer_10x10", "draughts_bt_10x10"]: sm_create_meth = getattr(cpp_statemachines, game_variant) add_game(game_variant, sm_create_meth(), model) def install_hex(add_game): ' load custom c++ statemachine for draughts ' from ggplib import interface from ggplib.non_gdl_games.hex.model import create_sm_model cpp_statemachines = interface.CppStateMachines() for sz in [9, 11, 13, 15, 19]: cpp_sm = cpp_statemachines.get_hex(sz) model = create_sm_model(sz) add_game("hex_lg_%s" % sz, cpp_sm, model) ############################################################################### # The API: the_database = None def get_database(verbose=True): global the_database def add_game_to_db(game, sm, model): info = GameInfoBypass(game, sm, model) the_database.game_mapping[game] = info if the_database is None: from ggplib.db.store import get_root the_database = GameDatabase(get_root()) the_database.load(verbose=verbose) try: install_draughts(add_game_to_db) except Exception as err: log.error("Failed to install draughts: %s" % err) try: install_hex(add_game_to_db) except Exception as err: log.error("Failed to install hex: %s" % err) return the_database def get_all_game_names(): return get_database().all_games # XXX build_sm not used. def by_name(name, build_sm=True): try: db = get_database(verbose=False) return db.get_by_name(name) except Exception as exc: # creates temporary files msg = "Lookup of %s failed: %s" % (name, exc) log.error(msg) log.error(traceback.format_exc()) raise LookupFailed(msg) def by_gdl(gdl): try: gdl_str = gdl if not isinstance(gdl, str): lines = [] for s in gdl: lines.append(str(s)) gdl_str = "\n".join(lines) db = get_database() try: info, mapping = db.lookup(gdl_str) except LookupFailed as exc: etype, value, tb = sys.exc_info() traceback.print_exc() raise LookupFailed("Did not find game %s" % exc) return mapping, info except Exception as exc: # creates temporary files log.error("Lookup failed: %s" % exc) model, sm = builder.build_sm(gdl) info = TempGameInfo("unknown", gdl, sm, model) return None, info
28.206349
88
0.566798
4e63749234da693d5c1f2625bba0bf9c3d524e3f
274
py
Python
testproject/fiber_test/tests/test_templatetags/test_fiber_version.py
bsimons/django-fiber
0f4b03217a4aeba6b48908825507fbe8c5732c8d
[ "Apache-2.0" ]
143
2015-01-06T01:15:22.000Z
2017-07-08T04:10:08.000Z
testproject/fiber_test/tests/test_templatetags/test_fiber_version.py
bsimons/django-fiber
0f4b03217a4aeba6b48908825507fbe8c5732c8d
[ "Apache-2.0" ]
44
2015-01-22T14:21:32.000Z
2017-05-31T16:59:23.000Z
testproject/fiber_test/tests/test_templatetags/test_fiber_version.py
bsimons/django-fiber
0f4b03217a4aeba6b48908825507fbe8c5732c8d
[ "Apache-2.0" ]
53
2015-01-21T21:48:49.000Z
2017-06-12T07:33:13.000Z
import fiber from django.test import SimpleTestCase from ...test_util import RenderMixin class TestFiberVersion(RenderMixin, SimpleTestCase): def test_fiber_version(self): self.assertRendered('{% load fiber_tags %}{% fiber_version %}', str(fiber.__version__))
30.444444
95
0.766423
4e658277a9b24094cf1e76fa7c348cccc93b01df
7,352
py
Python
main.py
dogerish/pic2html
cca9d032fb2325cb8c220cd0f5f632235d0f8c94
[ "MIT" ]
null
null
null
main.py
dogerish/pic2html
cca9d032fb2325cb8c220cd0f5f632235d0f8c94
[ "MIT" ]
null
null
null
main.py
dogerish/pic2html
cca9d032fb2325cb8c220cd0f5f632235d0f8c94
[ "MIT" ]
null
null
null
#! /usr/bin/python3 import sys, re from PIL import Image # return the argument if it exists (converted to the same type as the default), otherwise default default = lambda arg, defa: type(defa)(sys.argv[arg]) if len(sys.argv) > arg and sys.argv[arg] else defa # filename of image to evaluate, default is image.jpg IMAGE = default(1, "image.jpg") # filename of output, default just prints it to stdout OUTPUT = default(2, "") # outputs in defined way based on whether or not an output file is given if OUTPUT == "": output = print else: def output(*args, **kwargs): with open(OUTPUT, "w+") as ofile: ofile.write(*args, **kwargs) # output columns (width) COLS = default(3, 200) # color hues (degrees, [0-360)) COLORS = dict() with open('colors.txt') as f: # each line in the file for line in f.readlines(): # means comment if line.startswith('#'): continue # name: hue saturation # split bt name and values line = line.split(':') # split values with whitespace characters line = [line[0], *line[1].strip().split('\t')] # strip blank things from each piece for i, piece in enumerate(line): line[i] = piece.strip() # add key to COLORS name, hue, sat = line COLORS[name] = (None if hue == '*' else int(hue), None if sat == '*' else float(sat)) # characters for lightness values (ascending) CHARS = " -+:!?%#&$@" # color class class Color: def __init__(self, r=0, g=0, b=0, name=None): self.r, self.g, self.b = r, g, b self.vals = ('r', 'g', 'b') self.name = name # reduce the color to accumulator def reduce(self, reducer, accumulator=0): for v in self.vals: accumulator = reducer(accumulator, getattr(self, v)) return accumulator # executes f for each value of this color, returns a list of results def for_each(self, f): return [f(getattr(self, v)) for v in self.vals] # executes f on each color value, returns list of results def on_each(self, other, f): return [f(getattr(self, v), getattr(other, v)) for v in self.vals] # add with another color def __add__(self, color2): if type(color2) == Color: return Color(*self.on_each(color2, lambda a, b: a + b)) else: return Color(*self.for_each(lambda x: x + color2)) # multiply with another color def __mul__(self, color2): if type(color2) == Color: return Color(*self.on_each(color2, lambda a, b: a * b)) else: return Color(*self.for_each(lambda x: x * color2)) # subtract another color def __sub__(self, color2): return self + -1*color2 # divide by another color def __truediv__(self, color2): if type(color2) == Color: return Color(*self.on_each(color2, lambda a, b: a / b)) else: return Color(*self.for_each(lambda x: x / color2)) # get the difference between 2 colors (like subtraction but with no negatives) def diff(self, color2): return Color(*self.on_each(color2, lambda a, b: abs(a - b))) # get the sum of the rgb values def sum(self): return self.reduce(lambda a, b: a + b) # get the lightness of this color as a decimal percent # 1 means brightest, 0 means darkest, 0.5 means middle... def graylightness(self): return self.sum() / 765 # returns the hsl version of this color def hsl(self): ## setup # normalized version of self nself = self / 255 # rgb values vals = nself.for_each(lambda x: x) x, n = max(vals), min(vals) # max value d = x - n # difference bt max and min ## hue hue = 0; if d == 0: pass # max and min same elif x == nself.r: hue = 60*( (nself.g - nself.b) / d % 6 ) # r is max elif x == nself.g: hue = 60*( (nself.b - nself.r) / d + 2 ) # g is max else: hue = 60*( (nself.r - nself.g) / d + 4 ) # b is max lightness = (x + n) / 2 ## lightness saturation = 0 if d == 0 else d / (1 - abs(2*lightness - 1)) ## saturation # add 360 to hue if it's negative return (hue < 0)*360 + hue, saturation, lightness # approximate a given color to be one of the colors listed in COLORS # works by comparing hue values. lowest difference wins def approx(self, hsl=None): if hsl == None: hsl = self.hsl() hue, sat = hsl[:2] # the best one so far: (score, name, diff) best = (None, None, None) for name in COLORS.keys(): chue, csat = COLORS[name] a, am, b, bm = 0, 2, 0, 2 # if hue does matter if chue != None: a, bm = abs(hue - chue)/360, 1 # if saturation does matter if csat != None: b, am = abs(sat - csat), 1 # sum of difference in hue and saturation is the score score = a*am + b*bm # if this is a new best score if best[0] == None or score < best[0]: best = (score, name) # return the name of the best color return best[1] # color the string the color that the name describes def color_str(self, string, colorName): return f'<font color="{colorName}">{string}' # where the output will be accumulated to accumulator = '<body style="background-color: #000"><pre>' # open the image with Image.open(IMAGE) as img: # the step to increment by each time step = img.size[0] / COLS # the vertical step, to account for characters not being squares vstep = step * 15/7.81 # the current color curcolor = None # each row for row in range(int(img.size[1]/vstep)): row *= vstep # add newline character to go to next row if this isn't the first row accumulator += '\n' # each column for col in range(COLS): col *= step # average the colors for this location avgcolor = Color() colorc = 0 # color count # within this tile/area for y in range(int(row), int(row + vstep)): for x in range(int(col), int(col + step)): if x >= img.size[0]: break # break if it's out of range # add this pixel's color to the average avgcolor += Color(*img.getpixel((x, y))) colorc += 1 if y >= img.size[1]: break # break if it's out of range # turn sum into average avgcolor /= colorc # get the hsl version hsl = avgcolor.hsl() # approximate the color apcolor = avgcolor.approx(hsl) # pick the right character based on the lightness char = CHARS[round(hsl[2]*(len(CHARS) - 1))] # if it isn't already in the right color, change it if apcolor != curcolor: # add colored string to accumulator accumulator += "</font>" + avgcolor.color_str(char, apcolor) # new color curcolor = apcolor else: # add character accumulator += char # end the elements accumulator += "</font></pre></body>" # output the result output(accumulator)
37.131313
104
0.569777
4e673a1840258d5931483218139042ef6091e9ee
485
py
Python
coding_challenge/conftest.py
jojacobsen/coding_challenge
94335f00f57a6c4d64cbc2b282a0ca099445e866
[ "MIT" ]
1
2022-03-06T15:40:56.000Z
2022-03-06T15:40:56.000Z
coding_challenge/conftest.py
jojacobsen/coding_challenge
94335f00f57a6c4d64cbc2b282a0ca099445e866
[ "MIT" ]
null
null
null
coding_challenge/conftest.py
jojacobsen/coding_challenge
94335f00f57a6c4d64cbc2b282a0ca099445e866
[ "MIT" ]
null
null
null
import pytest from coding_challenge.users.models import User from coding_challenge.users.tests.factories import UserFactory from coding_challenge.ship_manager.models import Ship from coding_challenge.ship_manager.tests.factories import ShipFactory @pytest.fixture(autouse=True) def media_storage(settings, tmpdir): settings.MEDIA_ROOT = tmpdir.strpath @pytest.fixture def user() -> User: return UserFactory() @pytest.fixture def ship() -> Ship: return ShipFactory()
22.045455
69
0.797938
4e699a095a31850ba8f572b972a6266ffa8b3893
1,909
py
Python
swagger.py
edwardw1987/swagger
69b868523834811537af5265b47eb0bd94c42c2f
[ "MIT" ]
null
null
null
swagger.py
edwardw1987/swagger
69b868523834811537af5265b47eb0bd94c42c2f
[ "MIT" ]
null
null
null
swagger.py
edwardw1987/swagger
69b868523834811537af5265b47eb0bd94c42c2f
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # @Author: edward # @Date: 2016-05-12 14:11:21 # @Last Modified by: edward # @Last Modified time: 2016-05-12 17:29:48 from functools import partial # api = swagger.docs(Api(app), apiVersion='0.1', # basePath='http://localhost:5000', # resourcePath='/', # produces=["application/json", "text/html"], # api_spec_url='/api/spec', # description='A Basic API') class _APIs: def __init__(self): self.apis = [] self.make_docs() def add_api(self, method, **kw): self.apis.append(self.pre_api(method, **kw)) def operation(self, **kw): return partial(self.add_api, **kw) def pre_api(self, fn, **kw): d = dict( method=fn.__name__, ) d.update(kw) return d def get_spec(self): return self.apis def make_docs(self, apiVersion='1.0', swaggerVersion='1.2', basePath='http://localhost:9999', resourcePath='/', produces=["application/json"], api_spec_url='/api/spec', description='Auto generated API docs by swagger'): self.API_VERSION = apiVersion self.SWAGGER_VERSION = swaggerVersion self.BASE_PATH = basePath self.RESOURCE_PATH = resourcePath self.PRODUCES = produces self.API_SPEC_URL = api_spec_url self.DESCRIPTION = description apis = _APIs() operation = apis.operation docs = apis.make_docs get_api_spec = apis.get_spec def main(): class Handler: @operation( nickname="get something from api" ) def get(self): """this is get notes """ print 'get' def post(self): print 'post' for i in apis.apis: print i if __name__ == '__main__': main()
25.118421
64
0.550026
4e69ad09d076ffa2812d0d72b1983c1392ea46e1
29,469
py
Python
wicarproject/carapi/carupload_tests.py
todhm/wicarproject
5a3ea7b70ba6649af75d9e9bb49683eb6f94b570
[ "MIT" ]
1
2018-04-20T04:58:50.000Z
2018-04-20T04:58:50.000Z
wicarproject/carapi/carupload_tests.py
todhm/wicarproject
5a3ea7b70ba6649af75d9e9bb49683eb6f94b570
[ "MIT" ]
7
2021-02-08T20:24:49.000Z
2022-03-11T23:26:33.000Z
wicarproject/carapi/carupload_tests.py
todhm/wicarproject
5a3ea7b70ba6649af75d9e9bb49683eb6f94b570
[ "MIT" ]
null
null
null
import os import unittest import sqlalchemy from flask import Flask,session,url_for,redirect from flask_sqlalchemy import SQLAlchemy from application import create_app ,db import unittest import json from caruser.models import User, UserBank from carupload.models import CarOption,Car,CarImage from flask_testing import TestCase from utilities.dao.userdao import UserDao from utilities.dao.cardao import CarDao from utilities.testutil import TestUtil from freezegun import freeze_time from datetime import datetime as dt from datetime import timedelta from settings import TEST_DB_URI,MONGO_URI import urllib from utilities.flask_tracking.documents import Tracking from mongoengine.queryset.visitor import Q import os TEST_UPLOADED_FOLDER='/static/images/test_images' class CaruploadTest(TestUtil): def validate_caroption(self,data,carOption): for key in data: option_data = getattr(carOption,key) if getattr(carOption,key) is not None else "" self.assertEqual(data[key],option_data) # def test_mongoDB_api(self): # rv = self.client.get( # '/api/getCarBrand', # content_type='application/json', # follow_redirects=True # ) # data = json.loads(rv.get_data().decode('utf-8')) # self.assertTrue(len(data) >50) # self.check_mongoData('brandName','api/getCarClass','기아') # self.check_mongoData('brandName','api/getCarClass','현대') # # self.check_mongoData('className','api/getCarModel','K9') # self.check_mongoData('className','api/getCarModel','쏘울') def test_carWith_coordinate(self): with self.client.session_transaction() as session: rv1= self.register_user() session['email'] = 'todhm@nate.com' session['verification_code'] = '12345' rv2 = self.verify_code(code = '123445') carWithCoord = self.return_car_with_coord() rv = self.client.post( '/api/add_basic_info', data=json.dumps(carWithCoord), content_type='application/json', follow_redirects=True ) data = json.loads(rv.get_data().decode('utf-8')) user_id = self.userdao.get_user_id(session['email']) car = self.cardao.get_car_obj(user_id) self.verify_car_basic(car,carWithCoord) self.assertAlmostEqual(float(car.lng),carWithCoord['address']['pointX']) self.assertAlmostEqual(float(car.lat),carWithCoord['address']['pointY']) def test_update_car(self): with self.client.session_transaction() as session: rv1= self.register_user() session['email'] = 'todhm@nate.com' session['verification_code'] = '12345' rv2 = self.verify_code(code = '123445') car_without_coord = self.return_car_without_coord() rv = self.client.post( '/api/add_basic_info', data=json.dumps(car_without_coord), content_type='application/json', follow_redirects=True ) data = json.loads(rv.get_data().decode('utf-8')) user_id = self.userdao.get_user_id(session['email']) car = self.cardao.get_car_obj(user_id) car_id = car.id car_without_coord['brandName'] = "르노삼성" car_without_coord['model']="sm5" car_without_coord['transmission'] = "manual" rv = self.client.post( '/api/add_basic_info' + '/'+str(car_id), data=json.dumps(car_without_coord), content_type='application/json', follow_redirects= True ) user_id = self.userdao.get_user_id(session['email']) car = self.cardao.get_car_obj(user_id) self.verify_car_basic(car,car_without_coord) def test_car_without_coordinate(self): with self.client.session_transaction() as session: self.register_user_with_phone() car_without_coord = self.return_car_without_coord() session['email'] = 'todhm@nate.com' rv = self.client.post( '/api/add_basic_info', data=json.dumps(car_without_coord), content_type='application/json', follow_redirects=True ) user_id = self.userdao.get_user_id(session['email']) car = self.cardao.get_car_obj(user_id) self.verify_car_basic(car,car_without_coord) self.assertTrue(isinstance(float(car.lng), float)) self.assertTrue(isinstance(float(car.lat),float)) def test_proper_address(self): self.verify_address('평촌대로40번길 100') self.verify_address('신사동 502') self.verify_address('제주시 도령로 129') def test_proper_region(self): self.verify_region('과천도서관') self.verify_region('연세대학교') self.verify_region('고려대학교') self.verify_region('패스트캠퍼스') def test_liscence_addwith_correctemail(self): with self.client.session_transaction() as session: self.register_user_with_phone('todhm@naver.com') session['email'] = 'todhm@naver.com' liscence_data = self.return_full_liscence_info() rv = self.client.post( '/api/add_liscence_info', data=json.dumps(liscence_data), content_type='application/json', follow_redirects=True ) user = self.userdao.get_user_obj(session['email']) self.assertEqual(user.liscence_1 , liscence_data['liscence_1']) def test_liscence_with_uncomplete(self): with self.client.session_transaction() as session: self.register_user_with_phone('todhm@naver.com') session['email'] = 'todhm@naver.com' liscence_data = self.return_full_liscence_info() liscence_data.pop('liscence_1') rv = self.client.post( '/api/add_liscence_info', data=json.dumps(liscence_data), content_type='application/json', follow_redirects=True ) self.assertTrue(json.loads(rv.data.decode())['message'] !="success") def test_liscence_addwith_withoutsession(self): self.register_user_with_phone(email = 'todhm@naver.com') self.client.get('/logout') self.register_user_with_phone(email = 'gmlaud14@nate.com') liscence_data = self.return_full_liscence_info() rv = self.client.post( '/api/add_liscence_info', data=json.dumps(liscence_data), content_type='application/json', follow_redirects=True ) with self.client.session_transaction() as session: user = self.userdao.get_user_obj('todhm@naver.com') self.assertTrue(user.liscence_1 != liscence_data['liscence_1']) def test_get_liscence(self): self.register_user_with_phone(email = 'todhm@naver.com') liscence_data = self.return_full_liscence_info() rv = self.client.post( '/api/add_liscence_info', data=json.dumps(liscence_data), content_type='application/json', follow_redirects=True ) with self.client.session_transaction() as session: rv = self.client.get('/api/get_liscence') liscence_json = json.loads(rv.get_data().decode('utf-8')) self.assertTrue(liscence_json['liscence_1'] == liscence_data['liscence_1']) self.assertTrue(liscence_json['liscence_2'] == liscence_data['liscence_2']) self.assertTrue(liscence_json['liscence_3'] == liscence_data['liscence_3']) self.assertTrue(liscence_json['liscence_4'] == liscence_data['liscence_4']) self.assertTrue(liscence_json['birth'] == liscence_data['birth']) self.assertTrue(liscence_json['serialNumber'] == liscence_data['serialNumber']) def test_empty_liscence(self): self.register_user_with_phone(email = 'todhm@naver.com',name='강희명') with self.client.session_transaction() as session: rv = self.client.get('/api/get_liscence') liscence_json = json.loads(rv.get_data().decode('utf-8')) self.assertTrue(liscence_json.get('liscecne_1')is None) def test_caroption_add(self): car = self.return_car_obj() data = self.get_caroption_data() url = '/api/add_car_option'+ "/" + str(car.id) rv2=self.client.post( url, data=json.dumps(data), content_type='application/json', follow_redirects=True ) carOption = CarOption.query.filter(CarOption.id==car.id).first() self.validate_caroption(data,carOption) def test_add_caroption_without_carData(self): self.register_user_with_phone(email = 'todhm@naver.com',name='강희명') data = self.get_caroption_data() rv2=self.client.post('/api/add_car_option/1', data=json.dumps(data), content_type='application/json', follow_redirects=True) self.assertTrue(rv2.status_code==403) def test_add_caroption_with_unvalid_carData(self): car = self.return_car_obj() data = self.get_caroption_data() data.pop('price') url = '/api/add_car_option'+ "/" + str(car.id) rv2=self.client.post( url, data=json.dumps(data), content_type='application/json', follow_redirects=True ) self.assertEqual(rv2.status_code,403) def test_update_caroption(self): car = self.return_car_obj() data = self.get_caroption_data() url = '/api/add_car_option'+ "/" + str(car.id) rv2=self.client.post( url, data=json.dumps(data), content_type='application/json', follow_redirects=True ) data['price'] = 400000 data['description']='자동차가 업드레이드되서 돈을 조금 더 받아야할것 같습니다. ' rv3 = self.client.post( url, data=json.dumps(data), content_type='application/json', follow_redirects=True ) carOption = CarOption.query.filter(CarOption.id==car.id).first() self.validate_caroption(data,carOption) def test_get_caroption(self): car = self.return_car_obj() data = self.get_caroption_data() url = '/api/add_car_option'+ "/" + str(car.id) rv2=self.client.post( url, data=json.dumps(data), content_type='application/json', follow_redirects=True ) url = '/api/get_car_option'+ "/" + str(car.id) carOption = CarOption.query.filter(CarOption.id==car.id).first() rv3 = self.client.get( url, content_type='application/json', follow_redirects=True ) json_data = json.loads(rv3.get_data().decode('utf-8')) self.validate_caroption(json_data,carOption) def test_empty_caroption(self): car = self.return_car_obj() data = self.get_caroption_data() url = '/api/get_car_option'+ "/" + str(car.id) carOption = CarOption.query.filter(CarOption.id==car.id).first() rv3 = self.client.get( url, content_type='application/json', follow_redirects=True ) json_data = json.loads(rv3.get_data().decode('utf-8')) self.assertTrue(isinstance(json_data,dict)) def test_add_proper_image(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" with open(test_image_url, 'r+b') as f: rv= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) carImage = CarImage.query.filter(CarImage.car_id==car_id).first() imgUrl =json.loads(rv.get_data().decode('utf-8'))['imgList'] self.assertTrue(len(imgUrl)==1) self.assertTrue(carImage.active ==1) self.assertTrue(carImage.image_index == 0) with open("."+carImage.imgsrc,"r+b") as f2: self.assertTrue(f2 is not None ) self.remove_image(carImage.image,car.id) def test_update_image(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) imgUrl =json.loads(rv1.get_data().decode('utf-8'))['imgList'] carImage = CarImage.query.filter(CarImage.car_id==car_id).first() with open("."+carImage.imgsrc,"r+b") as f: self.assertTrue(f is not None ) self.remove_image(carImage.image,car.id) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url, 'r+b') as f2: rv2 = self.client.post('/api/update_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'image_index':0}) imgUrl2 =json.loads(rv2.get_data().decode('utf-8'))['imgList'] carImage = CarImage.query.filter(CarImage.car_id==car_id).first() self.assertTrue(len(imgUrl)==1) self.assertTrue(carImage.active ==1) self.assertTrue(carImage.image_index == 0) print(imgUrl,imgUrl2) self.assertTrue(imgUrl[0]['url'] != imgUrl2[0]['url']) with open("."+carImage.imgsrc,"r+b") as f: self.assertTrue(f is not None ) self.remove_image(carImage.image,car.id) def test_update_image_multiple_times(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) imgUrl =json.loads(rv1.get_data().decode('utf-8'))['imgList'] carImage = CarImage.query.filter(CarImage.car_id==car_id).first() self.remove_image(carImage.image,car.id) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url, 'r+b') as f2: rv2 = self.client.post('/api/update_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'image_index':0}) imgUrl2 =json.loads(rv2.get_data().decode('utf-8'))['imgList'] carImage = CarImage.query.filter(CarImage.car_id==car_id).first() self.remove_image(carImage.image,car.id) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url, 'r+b') as f2: rv3 = self.client.post('/api/update_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'image_index':'0'}) imgUrl3 =json.loads(rv3.get_data().decode('utf-8'))['imgList'] carImage = CarImage.query.filter(CarImage.car_id==car_id).first() self.remove_image(carImage.image,car.id) carImage = CarImage.query.filter(CarImage.car_id==car_id).first() self.assertTrue(len(imgUrl)==1) self.assertTrue(carImage.active ==1) self.assertTrue(carImage.image_index == 0) self.assertTrue(imgUrl[0]['url'] != imgUrl2[0]['url']) self.assertTrue(imgUrl[0]['url'] != imgUrl3[0]['url']) self.assertTrue(imgUrl2[0]['url'] != imgUrl3[0]['url']) def test_add_multiple_valid_image(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" test_image_url2 = "."+TEST_UPLOADED_FOLDER + "/background4.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1 =self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) frozen_datetime.tick(delta=timedelta(seconds=1)) with open(test_image_url2, 'r+b') as f2: rv2 = self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'image_index':1}) imgUrl1 =json.loads(rv1.get_data().decode('utf-8'))['imgList'] imgUrl2 =json.loads(rv2.get_data().decode('utf-8'))['imgList'] self.assertTrue(len(imgUrl1)==1) self.assertTrue(len(imgUrl2)==2) for idx,img in enumerate(imgUrl2): with open( "."+img['url'], 'r+b') as f3: self.assertTrue(f3 is not None) self.assertTrue(img['image_index'] ==idx) os.remove("."+img['url']) @freeze_time("Jan 14th, 2020", tick=True) def test_get_car_img(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) frozen_datetime.tick(delta=timedelta(seconds=1)) with open(test_image_url, 'r+b') as f2: rv2 = self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'car_index':1}) img_response = json.loads(self.client.get('/api/get_images/'+car_id).get_data().decode('utf-8')) self.assertTrue(img_response['message']=="success") self.assertTrue(len(img_response['imgList'])==2) for idx,img in enumerate(img_response['imgList']): with open("."+img['url'],"r+b") as f3: self.assertTrue(img['image_index']==idx) self.assertTrue(f3 is not None ) os.remove("."+img['url']) def test_remove_img(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" with open(test_image_url, 'r+b') as f: rv= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) removeUrl={} removeUrl['image_index'] = 0 rv = self.client.post('/api/remove_image/' + car_id, data=json.dumps(removeUrl), content_type='application/json', follow_redirects= True ) carImage = CarImage.query.filter(CarImage.car_id==car_id).filter(CarImage.active==False).all() self.assertTrue(len(carImage)==1) self.assertTrue(rv.status_code==200) for carimg in carImage: self.remove_image(carimg.image,car_id) @freeze_time("Jan 14th, 2020", tick=True) def test_remove_multiple_img(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" test_image_url2 = "."+TEST_UPLOADED_FOLDER + "/background4.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url2, 'r+b') as f2: rv2 = self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2}) imgUrl =json.loads(rv2.get_data().decode('utf-8'))['imgList'] removeUrl={} removeUrl['image_index'] = 0 for img in imgUrl: os.remove("."+img['url']) with freeze_time(dt.now()) as frozen_datetime: rv = self.client.post('/api/remove_image/' + car_id, data=json.dumps(removeUrl), content_type='application/json', follow_redirects= True ) removeUrl['image_index'] = 1 frozen_datetime.tick(delta=timedelta(hours=1)) rv2 = self.client.post('/api/remove_image/' + car_id, data=json.dumps(removeUrl), content_type='application/json', follow_redirects= True ) carImage = CarImage.query.filter(CarImage.car_id==car_id).filter(CarImage.active==False).all() self.assertTrue(len(carImage)==2) self.assertTrue(rv.status_code==200) self.assertTrue(rv2.status_code==200) @freeze_time("Jan 14th, 2020", tick=True) def test_add_invalid_img(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/hm.jpg" with open(test_image_url, 'r+b') as f: rv1= self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) data = json.loads(rv1.data.decode()) self.assertTrue(data['message']=="fail") carImage = CarImage.query.filter(CarImage.car_id==car_id).all() self.assertTrue(len(carImage)==0) def test_get_inactive_car(self): car = self.return_car_obj() car_id = str(car.id) rv = self.client.get('/api/get_images/'+car_id) data = json.loads(rv.data.decode()) self.assertTrue(data['active']==False) def test_get_active_car(self): car = self.return_car_obj() car_id = str(car.id) self.assertTrue(car.active ==0) rv = self.client.post('/api/activate_car/'+car_id) rv = self.client.get('/api/get_images/'+car_id) data = json.loads(rv.data.decode()) self.assertTrue(data['active']==True) def test_car_activate(self): car = self.return_car_obj() car_id = str(car.id) self.assertTrue(car.active ==0) rv = self.client.post('/api/activate_car/'+car_id) self.assertTrue("success" in rv.get_data().decode('utf-8')) car = Car.query.filter(Car.id==car_id).first() self.assertTrue(rv.status_code ==200) self.assertTrue(car.active ==1) #한사용자가 여러대의 차량을 등록해놓고 사진을 업데이트 시키면 다른자동차의 사진이 없어지는 문제를 확인하기 위한 테스트. def test_update_multiple_images(self): car = self.return_car_obj() car_id = str(car.id) test_image_url = "."+TEST_UPLOADED_FOLDER + "/background.jpg" test_image_url2 = "."+TEST_UPLOADED_FOLDER + "/background4.jpg" with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1 =self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f}) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url2, 'r+b') as f2: rv2 = self.client.post('/api/upload_image/' + car_id, buffered=True, content_type='multipart/form-data', data={'image':f2,'image_index':1}) car2 = self.return_car_obj(remainUser= True) car_2id = str(car2.id) with freeze_time(dt.now()) as frozen_datetime: with open(test_image_url, 'r+b') as f: rv1 =self.client.post('/api/upload_image/' + car_2id, buffered=True, content_type='multipart/form-data', data={'image':f}) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url2, 'r+b') as f2: rv2 = self.client.post('/api/upload_image/' + car_2id, buffered=True, content_type='multipart/form-data', data={'image':f2}) frozen_datetime.tick(delta=timedelta(hours=1)) with open(test_image_url2, 'r+b') as f: rv2 = self.client.post('/api/update_image/' + car_2id, buffered=True, content_type='multipart/form-data', data={'image':f,'image_index':0}) carImage = CarImage.query.filter(CarImage.car_id==car_id).filter(CarImage.active==True).filter(CarImage.image_index==0).first() car2Image = CarImage.query.filter(CarImage.car_id==car_2id).filter(CarImage.active==True).filter(CarImage.image_index==0).first() with open("."+carImage.imgsrc,"r+b") as f: self.assertTrue(f is not None ) with open("."+car2Image.imgsrc,"r+b") as f: self.assertTrue(f is not None ) #차량등록시 사용자들이 알맞은 위치에 존재하고 있는지 확인 def test_get_last_status(self): #차량 처음등록일경우. self.register_user_with_phone(email = 'todhm@naver.com') rv = self.client.get('/api/getLastStatus') data = json.loads(rv.data.decode()) self.assertEqual(data['stage_name'], ["자동차 등록","면허*계좌등록","세부사항조정","사진 등록","최종확인"]) #은행계좌등록. def test_bank_account(self): # 제대로된 계좌가 아닌경우. user = self.register_user_with_phone(email='todhm@naver.com') bank_info = self.get_bank_info() origin_birth = bank_info['account_holder_info'] original_account = bank_info['account_num'] bank_info['account_holder_info'] = origin_birth[1:] rv = self.client.post( '/api/add_bank_account', data=json.dumps(bank_info), content_type='application/json' ) result = json.loads(rv.data.decode()) self.assertTrue(result['message']!="success") self.assertEqual(rv.status_code,400) #은행계좌등록. def test_get_bank_account(self): user = self.register_user_with_phone(email='todhm@naver.com') bank = self.add_bank(user.id) rv = self.client.get('/api/get_bank_account') bank_info = json.loads(rv.data.decode()) bank_info.pop('message') self.verify_data(bank,bank_info) # 제대로된 계좌가 아닌경우. #가격불러오기. def test_get_car_price(self): car = self.activate_car_without_img() rv = self.client.get("/api/get_car_price/"+car.id) response = json.loads(rv.data.decode()) self.assertTrue(car.caroption.price==response['ordinaryPrice']) self.assertTrue(0==response['weeklyDiscount']) self.assertTrue(0==response['monthlyDiscount']) #가격 및 주별 월별 할인율 추가 def test_add_car_ordinary_price(self): car = self.activate_car_without_img() price_info = dict( ordinaryPrice=5000, weeklyDiscount=10, monthlyDiscount=30 ) rv = self.client.post( '/api/add_car_ordinary_price/'+car.id, data=json.dumps(price_info), content_type='application/json' ) response = json.loads(rv.data.decode()) self.assertTrue(response['message']=="success") self.assertTrue(price_info['ordinaryPrice']==car.caroption.price) self.assertTrue(price_info['weeklyDiscount']==car.caroption.weekly_discount) self.assertTrue(price_info['monthlyDiscount']==car.caroption.monthly_discount)
43.083333
137
0.574468
4e69d114d4d24a38c5a1cd7288544f6fdd2af296
1,947
py
Python
test/train_net.py
gregdeon/simple-ann
80f1d239d15b820162d5de93766290bca81f7bd3
[ "MIT" ]
1
2018-08-07T03:27:23.000Z
2018-08-07T03:27:23.000Z
test/train_net.py
gregdeon/simple-ann
80f1d239d15b820162d5de93766290bca81f7bd3
[ "MIT" ]
null
null
null
test/train_net.py
gregdeon/simple-ann
80f1d239d15b820162d5de93766290bca81f7bd3
[ "MIT" ]
null
null
null
# train-net.py # Use the neural network module to detect simple signals import numpy as np import matplotlib.pyplot as plt import random from src.net import Net def main(): """ Step 1: make dataset """ random.seed() # Make 3 inputs - 1 base and 2 added inputs sig_len = 10 y_base = np.array([1, 2, 3, 2, 6, 5, 0, -1, 2, 4]) y_add1 = np.array([0, 0, 1, 0, -2, 0, 0, 1, 1, 0]) y_add2 = np.array([1, 0, 0, 1, 2, -1, 0, 0, 0, 0]) # Set up a bunch of random signals to detect y_num = 100 signal1 = np.array([random.randint(0,1) for i in range(y_num)]) signal2 = np.array([random.randint(0,1) for i in range(y_num)]) signal = np.array([signal1, signal2]) # Add up the inputs accordingly y_list = np.zeros([y_num, len(y_base)]) for i in range(y_num): y_sum = np.array([y_base[j] + signal1[i]*y_add1[j] + signal2[i]*y_add2[j] for j in range(sig_len)]) y_list[i] = y_sum # Add noise noise = np.random.random([y_num, len(y_base)]) / 10 y_list += noise """ Step 2: train neural network """ # Set up input and signals input = np.array(y_list) signal = signal.transpose() # Set up min and max for each input # Can give the network a good idea of input ranges or just a rough range limits = [[y_base[i]-2, y_base[i]+2] for i in range(10)] #limits = [[-20, 20]]*10 # Make network net = Net(limits, 2, 2) errorList = net.train_many(input, signal, 0.1, 100, 0.001, True) print "\n".join(map(str, errorList)) """ Step 3: check results """ # Print results by hand #for i in range(y_num): # print y_list[i] # print signal1[i] # print signal2[i] # print net.sim(y_list[i, :]) # Plot error vs. training epochs plt.semilogy(errorList) plt.grid() plt.xlabel('Epochs') plt.ylabel('SSE') plt.show()
30.421875
82
0.580894
4e6a00926225b1d573a3d0b1379d0479bee2d1ba
2,165
py
Python
neural-network/model/gcn_lib/sparse/data_util.py
BuildJet/hongbomiao.com
33bcbf63c51f9df725811ddbdef49d0db46146ef
[ "MIT" ]
104
2019-08-09T21:27:48.000Z
2022-03-29T11:58:36.000Z
neural-network/model/gcn_lib/sparse/data_util.py
BuildJet/hongbomiao.com
33bcbf63c51f9df725811ddbdef49d0db46146ef
[ "MIT" ]
4,187
2019-08-04T08:19:36.000Z
2022-03-31T22:43:20.000Z
neural-network/model/gcn_lib/sparse/data_util.py
BuildJet/hongbomiao.com
33bcbf63c51f9df725811ddbdef49d0db46146ef
[ "MIT" ]
19
2019-08-06T00:53:05.000Z
2022-01-04T05:55:48.000Z
# allowable multiple choice node and edge features # code from https://github.com/snap-stanford/ogb/blob/master/ogb/utils/features.py allowable_features = { "possible_atomic_num_list": list(range(1, 119)) + ["misc"], # type: ignore "possible_chirality_list": [ "CHI_UNSPECIFIED", "CHI_TETRAHEDRAL_CW", "CHI_TETRAHEDRAL_CCW", "CHI_OTHER", ], "possible_degree_list": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, "misc"], "possible_formal_charge_list": [-5, -4, -3, -2, -1, 0, 1, 2, 3, 4, 5, "misc"], "possible_numH_list": [0, 1, 2, 3, 4, 5, 6, 7, 8, "misc"], "possible_number_radical_e_list": [0, 1, 2, 3, 4, "misc"], "possible_hybridization_list": ["SP", "SP2", "SP3", "SP3D", "SP3D2", "misc"], "possible_is_aromatic_list": [False, True], "possible_is_in_ring_list": [False, True], "possible_bond_type_list": ["SINGLE", "DOUBLE", "TRIPLE", "AROMATIC", "misc"], "possible_bond_stereo_list": [ "STEREONONE", "STEREOZ", "STEREOE", "STEREOCIS", "STEREOTRANS", "STEREOANY", ], "possible_is_conjugated_list": [False, True], } def get_atom_feature_dims(): return list( map( len, [ allowable_features["possible_atomic_num_list"], allowable_features["possible_chirality_list"], allowable_features["possible_degree_list"], allowable_features["possible_formal_charge_list"], allowable_features["possible_numH_list"], allowable_features["possible_number_radical_e_list"], allowable_features["possible_hybridization_list"], allowable_features["possible_is_aromatic_list"], allowable_features["possible_is_in_ring_list"], ], ) ) def get_bond_feature_dims(): return list( map( len, [ allowable_features["possible_bond_type_list"], allowable_features["possible_bond_stereo_list"], allowable_features["possible_is_conjugated_list"], ], ) )
35.491803
82
0.590762
4e6b5fa33d845c1a9c9c556d16d04f10c237dd56
3,401
py
Python
ising_model/hamiltonian.py
FeiQuantumSoftware/ising_model
6d8b177678aa953840fc01616dc7c789d9531b93
[ "BSD-3-Clause" ]
null
null
null
ising_model/hamiltonian.py
FeiQuantumSoftware/ising_model
6d8b177678aa953840fc01616dc7c789d9531b93
[ "BSD-3-Clause" ]
null
null
null
ising_model/hamiltonian.py
FeiQuantumSoftware/ising_model
6d8b177678aa953840fc01616dc7c789d9531b93
[ "BSD-3-Clause" ]
null
null
null
"""coupling Hamiltonian class def""" from math import exp import numpy as np from .spinconfig import SpinConfig class Hamiltonian(): """Create a class of Hamiltonian of 2-d Ising model. Parameters ---------- J: float, optional Coupling parameter, default J=-2 . u: float, optional External field strength, default u=1.1 . Returns ------- Hamiltonian: class A Hamiltonian of a Ising model with J: coupling strength, u: external field factor. Examples -------- >>>ham = Hamiltonian(-2,1.1) >>>ham. J -2 """ def __init__(self, J=-2, u=1.1): self.u = u self.J = J def energy(self, spinlist): """Calculate the energy of a given spinconfiguration. Parameters ---------- spinlist : list Spin configuration represented in '1': spin up, '0': spin down. Returns ------- energy : float Total energy out from both the external filed and coupling from neighbor spins. Examples -------- >>>ham = Hamiltonian(-2,1.1) >>>ham. energy([0,1,0,1,1]) -4.9 """ self.spinlist = spinlist E = 0 # Energy from the external field: # H_external = Sum over i of u * spin[i] for eachspin in self.spinlist: if eachspin == 1: E += self.u * 1 elif eachspin == 0: E += self.u * (-1) else: print("Spin input error") # Energy from coupling the nearest neighbor spin: # H_c = -J/k * spin[i] * spin[i+1] newList = self.spinlist[1:] newList.append(self.spinlist[0]) for spinx, spiny in zip(self.spinlist, newList): if spinx == spiny: E += -self.J * 1 elif spinx != spiny: E += -self.J * (-1) else: print("Type error spininput") return E def average(self, T=1, N=0): """Calculate the oberservables of a given spin list with N sites. Parameters ---------- T : float, optional Temperature of the system. N : interger, optional The site number of a spin list. Returns ------- E, m, C, ms : set Average energy, average magnetism, heat capacibility, magnetic susceptbility. Examples -------- >>>ham = Hamiltonian(-2,1.1) >>>ham. average(10, 4) (-1.894905381126034, -0.29386784002835087, 0.17850826588133842, 0.26682385808137565) """ mySpin = SpinConfig(N) Zsum = 0 E = 0 EE = 0 m = 0 mm = 0 for i in range(mySpin.iMax): myspinlist = mySpin.input_decimal(i) mi = mySpin.magnetization() Ei = self.energy(myspinlist) Zi = exp(-Ei/T) Zsum += Zi E += Zi * Ei EE += Zi * Ei*Ei m += Zi * mi mm += Zi * mi * mi # get average energy E = E/Zsum EE = EE/Zsum # get average magnetism m = m/Zsum mm = mm/Zsum # get capacity C = (EE - E**2)/(T*T) # get magnetic susceptibility ms = (mm - m**2)/(T) return E, m, C, ms
25.007353
91
0.486622
4e6b7f322b294819ab5b4b631d1bcc760d50a431
618
py
Python
examples/testpetsc.py
DDMGNI/viVlasov1D
901dd058711f6943eb6497b941bc115a64e822de
[ "MIT" ]
2
2018-09-13T12:39:07.000Z
2019-04-05T04:55:59.000Z
examples/testpetsc.py
DDMGNI/viVlasov1D
901dd058711f6943eb6497b941bc115a64e822de
[ "MIT" ]
null
null
null
examples/testpetsc.py
DDMGNI/viVlasov1D
901dd058711f6943eb6497b941bc115a64e822de
[ "MIT" ]
null
null
null
import numpy as np def temperature(x): pass def density(x, L): # weak Jeans instability # density = 1. + 0.01 * np.cos(0.8 * (x - 0.5 * L)) # strong Jeans instability # density = 1. + 0.01 * np.cos(0.1 * (x - 0.5 * L)) # linear Landau damping # return 1. + 0.01 * np.cos(0.5 * (x - 0.5 * L)) # nonlinear Landau damping density = 1. + (np.cos(0.5 * x)) return density def distribution(x, y): # if y <= 2. and y >= -2.: # f = 1. # else: # f = 0. f = 1. + 0.5 * (np.cos(x + y*np.pi) - np.cos(x - y*np.pi)) return f
16.263158
62
0.469256
4e6dab17c24195e50268c9d3fbe8629caaa109c4
17,866
py
Python
tcm-py/tcm_fabric.py
Datera/lio-utils
0ac9091c1ff7a52d5435a4f4449e82637142e06e
[ "Apache-2.0" ]
8
2015-04-02T21:44:47.000Z
2021-07-15T08:31:28.000Z
tcm-py/tcm_fabric.py
Datera/lio-utils
0ac9091c1ff7a52d5435a4f4449e82637142e06e
[ "Apache-2.0" ]
null
null
null
tcm-py/tcm_fabric.py
Datera/lio-utils
0ac9091c1ff7a52d5435a4f4449e82637142e06e
[ "Apache-2.0" ]
8
2015-06-18T14:30:21.000Z
2021-03-25T19:51:03.000Z
#!/usr/bin/python import os, sys, shutil import subprocess as sub import string import re import datetime, time import optparse target_root = "/sys/kernel/config/target/" spec_root = "/var/target/fabric/" def fabric_err(msg): print >> sys.stderr, msg sys.exit(1) def fabric_configfs_dump(fabric_name, fabric_root, module_name): if not os.path.isdir(fabric_root): print "Unable to access fabric_root: " + fabric_root sys.exit(1) iqn_root = os.listdir(fabric_root) # This will load up the fabric module print "modprobe " + module_name print "mkdir " + fabric_root # print "#### " + fabric_name + " Discovery authentication information" auth_dir = fabric_root + "/discovery_auth" if os.path.isdir(auth_dir) == True: for auth in os.listdir(auth_dir): if auth == "authenticate_target": continue auth_file = auth_dir + "/" + auth p = os.open(auth_file, 0) value = os.read(p, 256) ret = value.isspace() if ret: os.close(p) continue print "echo -n " + value.rstrip() + " > " + auth_file os.close(p) iqn_root = os.listdir(fabric_root) # Loop through LIO-Target IQN list for iqn in iqn_root: if not os.path.isdir(fabric_root + "/" + iqn): continue if iqn == "lio_version": continue if iqn == "discovery_auth": continue # Loop through LIO-Target IQN+TPGT list tpg_root = os.listdir(fabric_root + "/" + iqn); for tpgt_tmp in tpg_root: if tpgt_tmp == "fabric_statistics": continue tpgt_tmp2 = tpgt_tmp.split('_') tpgt = tpgt_tmp2[1] # print "#### Network portals for iSCSI Target Portal Group" # np_root = os.listdir(fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/np") # for np in np_root: # print "mkdir -p " + fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/np/" + np # Dump Nexus attribute (when available) nexus_file = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/nexus" if os.path.isfile(nexus_file): print "mkdir -p " + fabric_root + "/" + iqn + "/tpgt_" + tpgt p = os.open(nexus_file, 0) value = os.read(p, 256) print "echo " + value.rstrip() + " > " + nexus_file os.close(p) print "#### " + fabric_name + " Target Ports" lun_root = os.listdir(fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/lun") for lun_tmp in lun_root: lun_tmp2 = lun_tmp.split('_') lun = lun_tmp2[1] lun_dir = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/lun/lun_" + lun print "mkdir -p " + lun_dir port_root = os.listdir(lun_dir) for port in port_root: if port == "alua_tg_pt_gp": continue if port == "alua_tg_pt_offline": continue if port == "alua_tg_pt_status": continue if port == "alua_tg_pt_write_md": continue if not os.path.islink(lun_dir + "/" + port): continue port_link = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/lun/lun_" + lun + "/" + port sourcelink = os.readlink(port_link) sourcelink2 = os.path.join(os.path.dirname(port_link), sourcelink) print "ln -s " + sourcelink2 + " " + port_link # Dump ALUA Target Port Group tg_pt_gp_file = lun_dir + "/alua_tg_pt_gp" p = os.open(tg_pt_gp_file, 0) try: value = os.read(p, 512) except: os.close(p) continue os.close(p) if value: tg_pt_gp_tmp = value.split('\n') tg_pt_gp_out = tg_pt_gp_tmp[0] off = tg_pt_gp_out.index('Alias: ') off += 7 # Skip over "Alias: " tg_pt_gp_name = tg_pt_gp_out[off:] # Only need to dump if LIO-Target Port is NOT partof # the 'default_tg_pt_gp' if not re.search(tg_pt_gp_name, 'default_tg_pt_gp'): print "#### ALUA Target Port Group" print "echo " + tg_pt_gp_name + " > " + tg_pt_gp_file #FIXME: --aluasecmd support # print "lio_node --aluasecmd " + iqn + " " + tpgt + " " + lun # Dump values of iscsi/iqn/tpgt/attrib/ print "#### Attributes for " + fabric_name + " Target Portal Group" attrib_dir = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/attrib/" attrib_root = os.listdir(attrib_dir) for attrib in attrib_root: attrib_file = attrib_dir + attrib p = os.open(attrib_file, 0) value = os.read(p, 16) print "echo " + value.rstrip() + " > " + attrib_file os.close(p) # Dump values for iscsi/iqn/tpgt/param print "#### Parameters for " + fabric_name + " Target Portal Group" param_dir = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/param/" param_root = os.listdir(param_dir) for param in param_root: param_file = param_dir + param p = os.open(param_file, 0) value = os.read(p, 256) print "echo \"" + value.rstrip() + "\" > " + param_file os.close(p) if os.path.isfile(nexus_file): continue # Dump fabric Initiator Node ACLs from fabric_root/$WWN/tpgt_$TPGT/acls/ print "#### " + fabric_name + " Initiator ACLs for " + fabric_name + " Target Portal Group" nacl_dir = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/acls/" nacl_root = os.listdir(nacl_dir) for nacl in nacl_root: print "mkdir -p " + nacl_dir + nacl # Dump fabric Initiator ACL authentication info from fabric_root/$WWN/tpgt_$TPGT/acls//$INITIATOR/auth print "#### " + fabric_name + " Initiator ACL authentication information" auth_dir = nacl_dir + nacl + "/auth" for auth in os.listdir(auth_dir): if auth == "authenticate_target": continue auth_file = auth_dir + "/" + auth p = os.open(auth_file, 0) value = os.read(p, 256) ret = value.isspace() if ret: os.close(p) continue print "echo -n " + value.rstrip() + " > " + auth_file os.close(p) # Dump fabric Initiator ACL TPG attributes from fabric_root/$WWN/tpgt_$TPGT/acls/$INITIATOR/attrib print "#### " + fabric_name + " Initiator ACL TPG attributes" nacl_attrib_dir = nacl_dir + nacl + "/attrib" for nacl_attrib in os.listdir(nacl_attrib_dir): nacl_attrib_file = nacl_attrib_dir + "/" + nacl_attrib p = os.open(nacl_attrib_file, 0) value = os.read(p, 8) print "echo " + value.rstrip() + " > " + nacl_attrib_file os.close(p) # Dump fabric Initiator LUN ACLs from fabric_root/$WWN/tpgt_$TPGT//acls/$INITIATOR/lun print "#### " + fabric_name + " Initiator LUN ACLs for iSCSI Target Portal Group" lun_acl_dir = nacl_dir + nacl for lun_acl in os.listdir(lun_acl_dir): ret = re.search('lun_', lun_acl) if not ret: continue lun_link_dir = nacl_dir + nacl + "/" + lun_acl print "mkdir -p " + lun_link_dir for lun_acl_link in os.listdir(lun_link_dir): if lun_acl_link == "write_protect": p = os.open(lun_link_dir + "/write_protect", 0) value = os.read(p, 4) print "echo " + value.rstrip() + " > " + lun_link_dir + "/write_protect" os.close(p) continue if not os.path.islink(lun_link_dir + "/" + lun_acl_link): continue sourcelink = os.readlink(lun_link_dir + "/" + lun_acl_link) sourcelink2 = os.path.join(os.path.dirname(lun_link_dir + "/" + lun_acl_link), sourcelink) print "ln -s " + sourcelink2 + " " + lun_link_dir + "/" + lun_acl_link # Dump value of fabric_root/$WWN/tpgt_$TPGT//enable print "#### Trigger to enable " + fabric_name + " Target Portal Group" enable_file = fabric_root + "/" + iqn + "/tpgt_" + tpgt + "/enable" if os.path.isfile(enable_file): p = os.open(enable_file, 0) value = os.read(p, 1) print "echo " + value.rstrip() + " > " + enable_file os.close(p) return def fabric_configfs_dump_all(): for fabric_name in os.listdir(target_root): if fabric_name == "version": continue if fabric_name == "core": continue # FIXME: currently using lio_dump --stdout if fabric_name == "iscsi": continue fabric_root = target_root + fabric_name # print "Using fabric_configfs_dump_all: " + fabric_name + ", " + fabric_root module_name = fabric_get_module_name(fabric_name) # print "module_name: "+ module_name fabric_configfs_dump(fabric_name, fabric_root, module_name); return def fabric_backup_to_file(date_time, fabric_name, fabric_root, module_name): now = date_time if not os.path.isdir(fabric_root): print "Unable to access fabric_root: " + fabric_root sys.exit(1) current_dir = "/etc/target" backup_dir = "/etc/target/backup" if not os.path.isdir(backup_dir): op = "mkdir " + backup_dir ret = os.system(op) if ret: print "Unable to open backup_dir" sys.exit(1) op = "tcm_fabric --stdout --fabric-name=" + fabric_name + " --fabric-root=" + fabric_root + " --module-name=" + module_name # print "Using op: " + op p = sub.Popen(op, shell=True, stdout=sub.PIPE).stdout if not p: print "Unable to dump " + fabric_name + "/ConfigFS running state" sys.exit(1) orig_file = current_dir + "/" + fabric_name + "_start.sh" print "Making backup of " + fabric_name + "/ConfigFS with timestamp: " + now backup_file = backup_dir + "/" + fabric_name + "_backup-" + now + ".sh" if os.path.isfile(backup_file): print "" + fabric_name + " backup_file: " + backup_file + "already exists, exiting" p.close() sys.exit(1) back = open(backup_file, 'w') line = p.readline() while line: print >>back, line.rstrip() line = p.readline() p.close() back.close() ret = shutil.copyfile(backup_file, orig_file) if ret: print "Unable to copy " + back_file sys.exit(1) print "Successfully updated default config " + orig_file return backup_file def fabric_backup_to_file_all(date_time): if not os.path.isdir(target_root): print "Unable to open target_root: " + target_root sys.exit(1) for fabric_name in os.listdir(target_root): if fabric_name == "version": continue if fabric_name == "core": continue # FIXME: currently using lio_dump if fabric_name == "iscsi": continue fabric_root = target_root + fabric_name # print "Using fabric_backup_to_file: " + date_time + ", " + fabric_name + ", " + fabric_root module_name = fabric_get_module_name(fabric_name) # print "Using module_name: "+ module_name fabric_backup_to_file(date_time, fabric_name, fabric_root, module_name) return def fabric_unload(fabric_name, fabric_root, module_name): if not os.path.isdir(fabric_root): print "Unable to access fabric_root: " + fabric_root sys.exit(1) wwn_root = os.listdir(fabric_root) for wwn in wwn_root: if not os.path.isdir(fabric_root + "/" + wwn): continue if wwn == "discovery_auth": continue tpg_root = fabric_root + "/" + wwn for tpgt_tmp in os.listdir(tpg_root): if tpgt_tmp == "fabric_statistics": continue tpgt_tmp2 = tpgt_tmp.split('_') tpgt = tpgt_tmp2[1] if os.path.isfile(fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/enable"): disable_op = "echo 0 > " + fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/enable" ret = os.system(disable_op) if ret: print "Unable to disable TPG: " + wwn + " TPGT: " + tpgt nacl_root = fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/acls" for nacl in os.listdir(nacl_root): lun_acl_root = nacl_root + "/" + nacl + "/" for lun_acl in os.listdir(lun_acl_root): ret = re.search('lun_', lun_acl) if not ret: continue mapped_lun = lun_acl[4:] lun_link_dir = lun_acl_root + "/" + lun_acl + "/" for lun_acl_link in os.listdir(lun_link_dir): if lun_acl_link == "write_protect": continue if os.path.islink(lun_link_dir + "/" + lun_acl_link): unlink_op = lun_link_dir + "/" + lun_acl_link ret = os.unlink(unlink_op) if ret: print "Unable to unlink MappedLUN: " + lun_link_dir + "/" + lun_acl_link dellunacl_op = "rmdir " + lun_link_dir ret = os.system(dellunacl_op) if ret: print "Unable to rmdir fabric mapped_lun" delnodeacl_op = "rmdir " + nacl_root + "/" + nacl + "/" ret = os.system(delnodeacl_op) if ret: print "Unable to remove NodeACL: " + nacl_root + "/" + nacl + "/" lun_root = fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/lun" for lun_tmp in os.listdir(lun_root): lun_tmp2 = lun_tmp.split('_') lun = lun_tmp2[1] lun_dir = lun_root + "/lun_" + lun for port in os.listdir(lun_dir): if not os.path.islink(lun_dir + "/" + port): continue unlink_op = lun_dir + "/" + port ret = os.unlink(unlink_op) if ret: print "Unable to unlink fabric port/lun" rmdir_op= "rmdir " + lun_dir ret = os.system(rmdir_op); if ret: print "Unable to rmdir fabric port/lun: " + lun_dir rmdir_op = "rmdir " + fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/" ret = os.system(rmdir_op) if ret: print "Unable to rmdir fabric tpg: " + fabric_root + "/" + wwn + "/tpgt_" + tpgt + "/" rmdir_op = "rmdir " + fabric_root + "/" + wwn + "/" ret = os.system(rmdir_op) if ret: print "Unable to rmdir fabric wwn: " + fabric_root + "/" + wwn + "/" rmdir_op = "rmdir " + fabric_root ret = os.system(rmdir_op) if ret: print "Unable to release fabric_root: " + fabric_root rmmod_op = "rmmod " + module_name ret = os.system(rmmod_op) if ret: print "Unable to unload " + module_name print "Successfully released fabric: " + fabric_root return def fabric_get_module_name(fabric_name): kernel_module = "" for specs in os.listdir(spec_root): if specs == "README": continue # print "specs: " + specs + ", fabric_name: " + fabric_name if not re.search(fabric_name + ".spec", specs) and not re.search("tcm_" + fabric_name + ".spec", specs) and not re.search(fabric_name, specs): continue op = "cat " + spec_root + specs p = sub.Popen(op, shell=True, stdout=sub.PIPE).stdout if not p: print "Unable to dump " + fabric_name + "/ConfigFS running state" sys.exit(1) line = p.readline() while line: tmp = line.rstrip() # Check for 'kernel_module' line in $FABRIC.spec if re.search('kernel_module', tmp): tmp_list = tmp.split('= ') p.close() return tmp_list[1] line = p.readline() p.close() return kernel_module def fabric_unloadall(): module_name = "" try: for fabric_name in os.listdir(target_root): if fabric_name == "version": continue if fabric_name == "core": continue # FIXME: currently using lio_node --unload if fabric_name == "iscsi": continue fabric_root = target_root + fabric_name module_name = fabric_get_module_name(fabric_name) #print "fabric_get_module_name() using: " + module_name if module_name == "": continue fabric_unload(fabric_name, fabric_root, module_name) except OSError, (errno, strerror): if errno == 2: fabric_err("%s %s\n%s" % (target_root, strerror, "Is kernel module loaded?") ) def do_work(stdout_enable, stdout_enable_all, date_time, unload, unloadall, fabric_name, fabric_root, module_name): if not stdout_enable == "None": fabric_configfs_dump(fabric_name, fabric_root, module_name) elif not stdout_enable_all == "None": fabric_configfs_dump_all() elif not date_time == "None": fabric_backup_to_file(date_time, fabric_name, fabric_root, module_name) elif not unload == "None": fabric_unload(fabric_name, fabric_root, module_name) elif not unloadall == "None": fabric_unloadall() return 0 def main(): parser_fabric = optparse.OptionParser() parser_fabric.add_option("--s","--stdout", dest='stdout_enable', action='store', nargs=0, help="Dump running Fabric/ConfigFS syntax to STDOUT", type='string') parser_fabric.add_option("--z","--stdoutall", dest='stdout_enable_all', action='store', nargs=0, help="Dump all running Fabric/ConfigFS syntax to STDOUT", type='string') parser_fabric.add_option("--t", "--tofile", dest="date_time", action='store', nargs=1, help="Backup running Fabric/ConfigFS syntax to /etc/target/backup/fabricname_backup-<DATE_TIME>.sh", type='string') parser_fabric.add_option("--u", "--unload", dest="unload", action='store', nargs=0, help="Unload running Fabric/ConfigFS", type='string') parser_fabric.add_option("--a", "--unloadall", dest="unloadall", action='store', nargs=0, help="Unload all running Fabric/ConfigFS", type='string') parser_fabric.add_option("--f", "--fabric-name", dest='fabric_name', action='store', nargs=1, help="Target fabric name", type='string') parser_fabric.add_option("--r", "--fabric-root", dest='fabric_root', action='store', nargs=1, help="Target fabric configfs root", type='string') parser_fabric.add_option("--m", "--module-name", dest='module_name', action='store', nargs=1, help="Target fabric module name ", type='string') (opts_fabric, args_fabric) = parser_fabric.parse_args() mandatories = ['fabric_name', 'fabric_root', 'module_name'] for m in mandatories: if not opts_fabric.__dict__[m]: unloadall = str(opts_fabric.__dict__['unloadall']) stdout_enable = str(opts_fabric.__dict__['stdout_enable']) stdout_enable_all = str(opts_fabric.__dict__['stdout_enable_all']) date_time = str(opts_fabric.__dict__['date_time']) if unloadall == "None" and stdout_enable == "None" and stdout_enable_all == "None" and date_time == "None": print "mandatory option is missing\n" parser_fabric.print_help() exit(-1) do_work(str(opts_fabric.stdout_enable), str(opts_fabric.stdout_enable_all), str(opts_fabric.date_time), str(opts_fabric.unload), str(opts_fabric.unloadall), str(opts_fabric.fabric_name), str(opts_fabric.fabric_root), str(opts_fabric.module_name)) if __name__ == "__main__": main()
32.9631
144
0.640378
4e6eb09628c21128f6237f44dd57b2bfd0a093f8
12,321
py
Python
vmflib2/games/hl2mp.py
Trainzack/vmflib2
9bc9803b3c8c644346b5a5eb864c0deaf544d8a6
[ "BSD-2-Clause" ]
1
2021-02-11T17:52:48.000Z
2021-02-11T17:52:48.000Z
vmflib2/games/hl2mp.py
Trainzack/vmflib2
9bc9803b3c8c644346b5a5eb864c0deaf544d8a6
[ "BSD-2-Clause" ]
null
null
null
vmflib2/games/hl2mp.py
Trainzack/vmflib2
9bc9803b3c8c644346b5a5eb864c0deaf544d8a6
[ "BSD-2-Clause" ]
1
2021-02-12T18:56:51.000Z
2021-02-12T18:56:51.000Z
""" Helper classes for creating maps in any Source Engine game that uses hl2mp.fgd. This file was auto-generated by import_fgd.py on 2020-01-19 09:11:14.977620. """ from vmflib2.vmf import * class FilterActivatorTeam(Entity): """ Auto-generated from hl2mp.fgd, line 30. A filter that filters by the team of the activator. """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", targetname: str="", Negated="Allow entities that match criteria", filterteam=2): Entity.__init__(self, "filter_activator_team", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Name : The name that other entities refer to this entity by. self.targetname: str = targetname # Filter mode : If set to Allow, only entities who match the criteria will pass the filter. self.Negated = Negated # Filter Team Number : self.filterteam = filterteam self.auto_properties.extend(["origin", "targetname", "Negated", "filterteam"]) class InfoPlayerCombine(Entity): """ Auto-generated from hl2mp.fgd, line 17. This entity indicates the position and facing direction at which the player will spawn during a deathmatch map. Any number of info_player_deathmatch entities may be placed in a map. """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", angles: "Origin"="0 0 0"): Entity.__init__(self, "info_player_combine", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Pitch Yaw Roll (Y Z X) : This entity's orientation in the world. Pitch is rotation around the Y axis, self.angles: "Origin" = angles self.auto_properties.extend(["origin", "angles"]) class InfoPlayerDeathmatch(Entity): """ Auto-generated from hl2mp.fgd, line 10. This entity indicates the position and facing direction at which the player will spawn during a deathmatch map. Any number of info_player_deathmatch entities may be placed in a map. """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", angles: "Origin"="0 0 0"): Entity.__init__(self, "info_player_deathmatch", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Pitch Yaw Roll (Y Z X) : This entity's orientation in the world. Pitch is rotation around the Y axis, self.angles: "Origin" = angles self.auto_properties.extend(["origin", "angles"]) class InfoPlayerRebel(Entity): """ Auto-generated from hl2mp.fgd, line 24. This entity indicates the position and facing direction at which the player will spawn during a deathmatch map. Any number of info_player_deathmatch entities may be placed in a map. """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", angles: "Origin"="0 0 0"): Entity.__init__(self, "info_player_rebel", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Pitch Yaw Roll (Y Z X) : This entity's orientation in the world. Pitch is rotation around the Y axis, self.angles: "Origin" = angles self.auto_properties.extend(["origin", "angles"]) class PropPhysicsRespawnable(Entity): """ Auto-generated from hl2mp.fgd, line 43. This class is the same as prop_physics, except it respawns after it breaks """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", globalname: str="", angles: "Origin"="0 0 0", model: str="", skin: int=0, modelscale: float="1.0", targetname: str="", damagefilter: str="", disableshadows=0, ExplodeDamage: float=0, ExplodeRadius: float=0, PerformanceMode=0, BreakModelMessage: str="", pressuredelay: float=0, mindxlevel=0, maxdxlevel=0, fademindist: float=-1, fademaxdist: float=0, fadescale: float=1, spawnflags="", minhealthdmg: int=0, shadowcastdist: int=0, physdamagescale: float="0.1", Damagetype=0, nodamageforces=0, inertiaScale: float="1.0", massScale: float="0", overridescript: str="", damagetoenablemotion: int=0, forcetoenablemotion: float=0, puntsound: str="", renderfx=0, rendermode=0, renderamt: int=255, rendercolor: "RGB"="255 255 255", disablereceiveshadows=0, RespawnTime: float=60): Entity.__init__(self, "prop_physics_respawnable", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Global Entity Name : Name by which this entity is linked to another entity in a different map. When the player transitions to a new map, entities in the new map with globalnames matching entities in the previous map will have the previous map's state copied over their state. self.globalname: str = globalname # Pitch Yaw Roll (Y Z X) : This entity's orientation in the world. Pitch is rotation around the Y axis, self.angles: "Origin" = angles # World Model : self.model: str = model # Skin : Some models have multiple versions of their textures, called skins. Set this to a number other than 0 to use that skin instead of the default. self.skin: int = skin # Model Scale : A multiplier for the size of the model. self.modelscale: float = modelscale # Name : The name that other entities refer to this entity by. self.targetname: str = targetname # Damage Filter : Name of the filter entity that controls which entities can damage us. self.damagefilter: str = damagefilter # Disable shadows : self.disableshadows = disableshadows # Explosion Damage : If non-zero, when this entity breaks it will create an explosion that causes the specified amount of damage. See also 'Explosion Radius'. self.ExplodeDamage: float = ExplodeDamage # Explosion Radius : If non-zero, when this entity breaks it will create an explosion with a radius of the specified amount. See also 'Explosion Damage'. self.ExplodeRadius: float = ExplodeRadius # Performance Mode : Used to limit the amount of gibs produced when this entity breaks, for performance reasons. self.PerformanceMode = PerformanceMode # Break Model Message : If set, will use this break model message instead of the normal break behavior. self.BreakModelMessage: str = BreakModelMessage # Pressure Delay : Delay, in seconds, after 'broken' by pressure before breaking apart (allows for sound to play before breaking apart). self.pressuredelay: float = pressuredelay # Minimum DX Level : self.mindxlevel = mindxlevel # Maximum DX Level : self.maxdxlevel = maxdxlevel # Start Fade Dist : Distance at which the prop starts to fade (<0 = use fademaxdist). self.fademindist: float = fademindist # End Fade Dist : Max fade distance at which the prop is visible (0 = don't fade out) self.fademaxdist: float = fademaxdist # Fade Scale : If you specify a fade in the worldspawn, or if the engine is running under dx7, then the engine will forcibly fade out props even if fademindist/fademaxdist isn't specified. self.fadescale: float = fadescale # TODO: Replace this filler. : self.spawnflags = spawnflags # Min Damage to Hurt : The prop will ignore any damage events if the damage is less than this amount. self.minhealthdmg: int = minhealthdmg # Shadow Cast Distance : Use this to override how far this object casts shadows. 0 = default distance. self.shadowcastdist: int = shadowcastdist # Physics Impact Damage Scale : Scales damage energy when this object is hit by a physics object. NOTE: 0 means this feature is disabled for backwards compatibility.\nSet to 1.0 for materials as strong as flesh, smaller numbers indicate stronger materials. self.physdamagescale: float = physdamagescale # Impact damage type : self.Damagetype = Damagetype # Damaging it Doesn't Push It : Used to determine whether or not damage should cause the brush to move. self.nodamageforces = nodamageforces # Scale Factor For Inertia : Scales the angular mass of an object. Used to hack angular damage and collision response. self.inertiaScale: float = inertiaScale # Mass Scale : A scale multiplier for the object's mass. self.massScale: float = massScale # Override Parameters : A list of physics key/value pairs that are usually in a physics prop .qc file. Format is 'key,value,key,value,etc'. self.overridescript: str = overridescript # Health Level to Override Motion : If specified, this object will start motion disabled. Once its health has dropped below this specified amount, it will enable motion. self.damagetoenablemotion: int = damagetoenablemotion # Physics Impact Force to Override Motion : If specified, this object will start motion disabled. Any impact that imparts a force greater than this value on the physbox will enable motion. self.forcetoenablemotion: float = forcetoenablemotion # Sound to make when punted : self.puntsound: str = puntsound # Render FX : self.renderfx = renderfx # Render Mode : Used to set a non-standard rendering mode on this entity. See also 'FX Amount' and 'FX Color'. self.rendermode = rendermode # FX Amount (0 - 255) : The FX amount is used by the selected Render Mode. self.renderamt: int = renderamt # FX Color (R G B) : The FX color is used by the selected Render Mode. self.rendercolor: "RGB" = rendercolor # Disable Receiving Shadows : self.disablereceiveshadows = disablereceiveshadows # Respawn Time : Ammount in seconds this prop will respawn after it breaks. self.RespawnTime: float = RespawnTime self.auto_properties.extend(["origin", "globalname", "angles", "model", "skin", "modelscale", "targetname", "damagefilter", "disableshadows", "ExplodeDamage", "ExplodeRadius", "PerformanceMode", "BreakModelMessage", "pressuredelay", "mindxlevel", "maxdxlevel", "fademindist", "fademaxdist", "fadescale", "spawnflags", "minhealthdmg", "shadowcastdist", "physdamagescale", "Damagetype", "nodamageforces", "inertiaScale", "massScale", "overridescript", "damagetoenablemotion", "forcetoenablemotion", "puntsound", "renderfx", "rendermode", "renderamt", "rendercolor", "disablereceiveshadows", "RespawnTime"]) class WeaponSlam(Entity): """ Auto-generated from hl2mp.fgd, line 50. S.L.A.M. - Selectable Lightweight Attack Munition """ def __init__(self, vmf_map: "ValveMap", origin: "Origin"="0 0 0", targetname: str="", angles: "Origin"="0 0 0", spawnflags="", fademindist: float=-1, fademaxdist: float=0, fadescale: float=1): Entity.__init__(self, "weapon_slam", vmf_map) # Origin : This entity's location in 3D space. self.origin: "Origin" = origin # Name : The name that other entities refer to this entity by. self.targetname: str = targetname # Pitch Yaw Roll (Y Z X) : This entity's orientation in the world. Pitch is rotation around the Y axis, self.angles: "Origin" = angles # TODO: Replace this filler. : self.spawnflags = spawnflags # Start Fade Dist/Pixels : Distance at which the prop starts to fade (<0 = use fademaxdist). If 'Screen Space Fade' is selected, this represents the number of pixels wide covered by the prop when it starts to fade. self.fademindist: float = fademindist # End Fade Dist/Pixels : Maximum distance at which the prop is visible (0 = don't fade out). If 'Screen Space Fade' is selected, this represents the *minimum* number of pixels wide covered by the prop when it fades. self.fademaxdist: float = fademaxdist # Fade Scale : If you specify a fade in the worldspawn, or if the engine is running under dx7, then the engine will forcibly fade out props even if fademindist/fademaxdist isn't specified. self.fadescale: float = fadescale self.auto_properties.extend(["origin", "targetname", "angles", "spawnflags", "fademindist", "fademaxdist", "fadescale"])
57.306977
840
0.692152
4e6f333b107f38c8aaa09dc165be5dc797f0e6b5
317
py
Python
src/constants.py
heyhpython/desktop
e75ffddf9526e8fd1adaca69c315005e202bf84b
[ "MIT" ]
null
null
null
src/constants.py
heyhpython/desktop
e75ffddf9526e8fd1adaca69c315005e202bf84b
[ "MIT" ]
null
null
null
src/constants.py
heyhpython/desktop
e75ffddf9526e8fd1adaca69c315005e202bf84b
[ "MIT" ]
null
null
null
import os BASE_DIR = os.path.dirname(__file__) __config__ = os.path.abspath(os.path.join(BASE_DIR, "../config.cfg")) __template__ = os.path.abspath(os.path.join(BASE_DIR, "templates")) __static__ = os.path.abspath(os.path.join(BASE_DIR, "static")) __upload__ = os.path.abspath(os.path.join(__static__, "uploads"))
31.7
69
0.744479
4e7096429698d4dbcba3e4c9717842932c8154f8
1,363
py
Python
app.py
IamSilentBot/Guardzilla
8ca9dcda2d99cba1628b708a770a34dd726acd9e
[ "MIT" ]
1
2022-02-05T22:55:50.000Z
2022-02-05T22:55:50.000Z
app.py
IamSilentBot/Guardzilla
8ca9dcda2d99cba1628b708a770a34dd726acd9e
[ "MIT" ]
null
null
null
app.py
IamSilentBot/Guardzilla
8ca9dcda2d99cba1628b708a770a34dd726acd9e
[ "MIT" ]
1
2022-02-21T17:47:39.000Z
2022-02-21T17:47:39.000Z
import nextcord from nextcord.ext import commands import json import os import pymongo import os from keep_alive import keep_alive # Set environment variables # os.environ['info'] = "test:pass123" # os.environ['TOKEN'] = "MY-AWSOME-TOKEN" intents = nextcord.Intents.all() TOKEN = os.environ['TOKEN'] async def prefix_d(_, message): f = pymongo.MongoClient( f"mongodb+srv://{os.environ['info']}@cluster0.o0xc5.mongodb.net/myFirstDatabase?retryWrites=true&w=majority") cluster = f["Guardzilla"] prefix = cluster["prefix"] prefix_x = prefix.find_one({"_id": 0}) if not prefix_x or str(message.guild.id) not in prefix_x: prefix.delete_one({"_id": 0}) prefix.insert_one({"_id": 0, str(message.guild.id): "."}) prefix_x = prefix.find_one({"_id": 0}) if str(message.content).startswith(prefix_x[str(message.guild.id)]): return prefix_x[str(message.guild.id)] else: return str(client.user.id) client = nextcord.ext.commands.Bot( command_prefix=prefix_d, intents=intents, help_command=None) @client.event async def on_ready(): print(f'{client.user} has connected to Discord!') for pyFile in os.listdir("./commands"): if pyFile.endswith(".py"): client.load_extension(f"commands.{pyFile[:-3]}") print(f"{pyFile[:-3]} | Loaded") keep_alive() client.run(TOKEN)
27.816327
117
0.681585
4e71fbacfde29db685d3ecdd9b0c31b58a97a5ef
1,841
py
Python
generator/contact.py
bloodes/adressbook
52582bc8c4825987db668ab084dff32202f1e2e5
[ "Apache-2.0" ]
null
null
null
generator/contact.py
bloodes/adressbook
52582bc8c4825987db668ab084dff32202f1e2e5
[ "Apache-2.0" ]
null
null
null
generator/contact.py
bloodes/adressbook
52582bc8c4825987db668ab084dff32202f1e2e5
[ "Apache-2.0" ]
null
null
null
from models.model_contact import Contact import random import string import os.path import jsonpickle import getopt import sys def random_string(prefix, maxlen): symbols = string.ascii_letters + string.digits return prefix + ''.join([random.choice(symbols) for i in range(random.randrange(maxlen))]) try: opts, args = getopt.getopt(sys.argv[1:], "n:f:", ["number_of_groups", "file"]) except getopt.GetoptError as err: getopt.usage() sys.exit(2) n = 5 f = "data/contacts.json" for o, a in opts: if o == '-n': n = int(a) elif o == '-f': f = a testdata = [Contact(firstname='Stepan', middlename='Barantsev', lastname='Lol', nickname='Bloodes', email1='stepan.barantsev@gmail.com')] +\ [Contact(firstname=random_string('', 10), middlename=random_string('', 20), lastname=random_string('', 20), nickname=random_string('', 20), homephone=random_string('', 20), mobilephone=random_string('', 20), workphone=random_string('', 20), secondaryphone=random_string('', 20), email1=random_string('', 20), email2=random_string('', 20), email3=random_string('', 20), title=random_string('', 20), notes=random_string('', 20), company=random_string('', 20), homepage=random_string('', 20), fax=random_string('', 20)) for i in range(5) ] file = os.path.join(os.path.dirname(os.path.abspath(__file__)), "..", f) with open(file, 'w') as out: jsonpickle.set_encoder_options("json", indent=2) out.write(jsonpickle.encode(testdata))
33.472727
96
0.551331
4e72b83d81906c88261b3bf53646c5b537fd803e
13,598
py
Python
tick/array_test/tests/array_memory_test.py
sumau/tick
1b56924a35463e12f7775bc0aec182364f26f2c6
[ "BSD-3-Clause" ]
411
2017-03-30T15:22:05.000Z
2022-03-27T01:58:34.000Z
tick/array_test/tests/array_memory_test.py
saurabhdash/tick
bbc561804eb1fdcb4c71b9e3e2d83a66e7b13a48
[ "BSD-3-Clause" ]
345
2017-04-13T14:53:20.000Z
2022-03-26T00:46:22.000Z
tick/array_test/tests/array_memory_test.py
saurabhdash/tick
bbc561804eb1fdcb4c71b9e3e2d83a66e7b13a48
[ "BSD-3-Clause" ]
102
2017-04-25T11:47:53.000Z
2022-02-15T11:45:49.000Z
# License: BSD 3 clause import gc import unittest import weakref import numpy as np import scipy from scipy.sparse import csr_matrix from tick.array.build.array import tick_double_sparse2d_from_file from tick.array.build.array import tick_double_sparse2d_to_file from tick.array_test.build import array_test as test class Test(unittest.TestCase): def test_varray_smart_pointer_in_cpp(self): """...Test C++ reference counter """ vcc = test.VarrayContainer() self.assertEqual(vcc.nRef(), 0) vcc.initVarray() self.assertEqual(vcc.nRef(), 1) cu1 = test.VarrayUser() cu1.setArray(vcc) self.assertEqual(vcc.nRef(), 2) cu1.setArray(vcc) self.assertEqual(vcc.nRef(), 2) cu2 = test.VarrayUser() cu2.setArray(vcc) self.assertEqual(vcc.nRef(), 3) del cu1 self.assertEqual(vcc.nRef(), 2) cu3 = test.VarrayUser() cu3.setArray(vcc) self.assertEqual(vcc.nRef(), 3) del cu3, cu2 self.assertEqual(vcc.nRef(), 1) # we cannot check it will go to 0 after vcc deletion in Python cu4 = test.VarrayUser() cu4.setArray(vcc) self.assertEqual(vcc.nRef(), 2) del vcc self.assertEqual(cu4.nRef(), 1) # we cannot check it will go to 0 after cu4 deletion in Python del cu4 def test_varray_smart_pointer_deletion1(self): """...Test that varray is still alive after deletion in Python """ vcc = test.VarrayContainer() vcc.initVarray() # Now mix with some Python a = vcc.varrayPtr # This does not increment C++ reference counter self.assertEqual(vcc.nRef(), 1) # Get a weak ref of the array r = weakref.ref(a) del a np.testing.assert_array_almost_equal(r(), vcc.varrayPtr) del vcc self.assertIsNone(r()) def test_varray_smart_pointer_deletion2(self): """...Test that base is deleted after a double assignment in Python """ vcc = test.VarrayContainer() vcc.initVarray() a = vcc.varrayPtr b = vcc.varrayPtr r = weakref.ref(b) del a, vcc, b self.assertIsNone(r()) def test_varray_smart_pointer_deletion3(self): """...Test that base is deleted after a double assignment in Python """ vcc = test.VarrayContainer() vcc.initVarray() # Now mix with some Python a = vcc.varrayPtr a_sum = np.sum(a) # This does not increment C++ reference counter self.assertEqual(vcc.nRef(), 1) # Get a weak ref of the array r = weakref.ref(vcc.varrayPtr) del vcc np.testing.assert_array_almost_equal(a_sum, np.sum(a)) self.assertIsNone(r()) del a def test_sarray_memory_leaks(self): """...Test brute force method in order to see if we have a memory leak during typemap out """ import os try: import psutil except ImportError: print('Without psutils we cannot ensure we have no memory leaks') return def get_memory_used(): """Returns memory used by current process """ process = psutil.Process(os.getpid()) return process.memory_info()[0] initial_memory = get_memory_used() size = int(1e6) # The size in memory of an array of ``size`` doubles bytes_size = size * 8 a = test.test_typemap_out_SArrayDoublePtr(size) first_filled_memory = get_memory_used() # Check that new memory is of the correct order (10%) self.assertAlmostEqual(first_filled_memory - initial_memory, bytes_size, delta=1.1 * bytes_size) for _ in range(10): del a a = test.test_typemap_out_SArrayDoublePtr(size) filled_memory = get_memory_used() # Check memory is not increasing self.assertAlmostEqual(first_filled_memory - initial_memory, filled_memory - initial_memory, delta=1.1 * bytes_size) #print("\nfirst_filled_memory %.2g, filled_memory %.2g, initial_memory %.2g, array_bytes_size %.2g" % (first_filled_memory, filled_memory, initial_memory, bytes_size)) def test_sarray_memory_leaks2(self): """...Test brute force method in order to see if we have a memory leak during typemap in or out """ import os try: import psutil except ImportError: print('Without psutils we cannot ensure we have no memory leaks') return def get_memory_used(): """Returns memory used by current process """ process = psutil.Process(os.getpid()) return process.memory_info()[0] size = int(1e6) a, b = np.ones(size), np.arange(size, dtype=float) initial_memory = get_memory_used() # The size in memory of an array of ``size`` doubles bytes_size = 2 * size * 8 c = test.test_VArrayDouble_append(a, b) first_filled_memory = get_memory_used() # Check that new memory is of the correct order (10%) self.assertAlmostEqual(first_filled_memory, initial_memory + bytes_size, delta=1.1 * bytes_size) for _ in range(10): del c c = test.test_VArrayDouble_append(a, b) filled_memory = get_memory_used() # Check memory is not increasing self.assertAlmostEqual(first_filled_memory - initial_memory, filled_memory - initial_memory, delta=1.1 * bytes_size) def test_sarray2d_memory_leaks(self): """...Test brute force method in order to see if we have a memory leak during typemap out """ import os try: import psutil except ImportError: print('Without psutils we cannot ensure we have no memory leaks') return def get_memory_used(): """Returns memory used by current process """ process = psutil.Process(os.getpid()) return process.memory_info()[0] initial_memory = get_memory_used() n_rows = int(1e2) n_cols = int(1e3) # The size in memory of an array of ``size`` doubles bytes_size = n_rows * n_cols * 8 a = test.test_typemap_out_SArrayDouble2dPtr(n_rows, n_cols) first_filled_memory = get_memory_used() # Check that new memory is of the correct order (10%) self.assertAlmostEqual(first_filled_memory - initial_memory, bytes_size, delta=1.1 * bytes_size) for _ in range(10): del a a = test.test_typemap_out_SArrayDouble2dPtr(n_rows, n_cols) filled_memory = get_memory_used() # Check memory is not increasing self.assertAlmostEqual(first_filled_memory - initial_memory, filled_memory - initial_memory, delta=1.1 * bytes_size) def test_s_sparse_array2d_memory_leaks(self): """...Test brute force method in order to see if we have a memory leak during typemap out """ import os try: import psutil except ImportError: print('Without psutils we cannot ensure we have no memory leaks') return def get_memory_used(): """Returns memory used by current process """ process = psutil.Process(os.getpid()) return process.memory_info()[0] cereal_file = "sparse.gen.cereal" try: n_rows = int(1e3) n_cols = int(1e2) s_spar = int((n_rows * n_cols) * .3) data_size = (s_spar * 8) # The size in memory of an array of ``size`` doubles bytes_size = (data_size * 2) + ((n_rows + 1) * 8) sparsearray_double = scipy.sparse.rand( n_rows, n_cols, 0.3, format="csr", dtype=np.float64) tick_double_sparse2d_to_file(cereal_file, sparsearray_double) initial_memory = get_memory_used() a = tick_double_sparse2d_from_file(cereal_file) first_filled_memory = get_memory_used() # Check that new memory is of the correct order (10%) self.assertAlmostEqual(first_filled_memory - initial_memory, bytes_size, delta=1.1 * bytes_size) del a for i in range(10): # Check memory is not increasing gc.collect() filled_memory = get_memory_used() self.assertAlmostEqual(filled_memory, initial_memory, delta=1.1 * bytes_size) X = tick_double_sparse2d_from_file(cereal_file) del X gc.collect() end = get_memory_used() self.assertAlmostEqual(end, initial_memory, delta=1.1 * bytes_size) finally: if os.path.exists(cereal_file): os.remove(cereal_file) def test_varray_share_same_support(self): """...Test that modifications on Varray of in Python affect the same support """ vcc = test.VarrayContainer() vcc.initVarray() # Now mix with some Python a = vcc.varrayPtr a[0] = 99.0 self.assertEqual(vcc.varrayPtr[0], 99.0) vcc.varrayPtr[1] = 999.0 self.assertEqual(a[1], 999.0) def test_sbasearrayptr(self): sparsearray_double = csr_matrix( (np.array([1., 2, 3, 4, 5]), np.array([2, 4, 6, 8, 10]), np.array([0, 5])), shape=(1, 12)) test.test_sbasearray_container_new(sparsearray_double) self.assertEqual(test.test_sbasearray_container_compute(), 45) test.test_sbasearray_container_clear() self.assertEqual(test.test_sbasearray_container_compute(), -1) array_double = np.arange(2, 14, dtype=float) test.test_sbasearray_container_new(array_double) self.assertEqual(test.test_sbasearray_container_compute(), array_double.sum()) test.test_sbasearray_container_clear() self.assertEqual(test.test_sbasearray_container_compute(), -1) def test_ref_sbasearrayptr(self): sparsearray_double = csr_matrix( (np.array([1., 2, 3, 4, 5]), np.array([2, 4, 6, 8, 10]), np.array([0, 5])), shape=(1, 12)) refdata = weakref.ref(sparsearray_double.data) refindices = weakref.ref(sparsearray_double.indices) refindptr = weakref.ref(sparsearray_double.indptr) test.test_sbasearray_container_new(sparsearray_double) del sparsearray_double self.assertIsNone(refindptr()) self.assertIsNotNone(refdata()) self.assertIsNotNone(refindices()) test.test_sbasearray_container_clear() self.assertIsNone(refdata()) self.assertIsNone(refindices()) array_double = np.arange(2, 14, dtype=float) ref = weakref.ref(array_double) test.test_sbasearray_container_new(array_double) del array_double self.assertIsNotNone(ref()) test.test_sbasearray_container_clear() self.assertIsNone(ref()) def test_sbasearray2dptr(self): sparsearray2d_double = csr_matrix( (np.array([1., 2, 3, 4, 5]), np.array([2, 4, 6, 1, 3]), np.array([0, 3, 5])), shape=(2, 4)) test.test_sbasearray2d_container_new(sparsearray2d_double) self.assertEqual(test.test_sbasearray2d_container_compute(), 39) test.test_sbasearray2d_container_clear() self.assertEqual(test.test_sbasearray2d_container_compute(), -1) array2d_double = np.array([[1.2, 3], [4, 5]]) test.test_sbasearray2d_container_new(array2d_double) self.assertEqual(test.test_sbasearray2d_container_compute(), array2d_double.sum()) test.test_sbasearray2d_container_clear() self.assertEqual(test.test_sbasearray2d_container_compute(), -1) def test_ref_sbasearray2dptr(self): sparsearray2d_double = csr_matrix( (np.array([1., 2, 3, 4, 5]), np.array([2, 4, 6, 1, 3]), np.array([0, 3, 5])), shape=(2, 4)) refdata = weakref.ref(sparsearray2d_double.data) refindices = weakref.ref(sparsearray2d_double.indices) refindptr = weakref.ref(sparsearray2d_double.indptr) test.test_sbasearray2d_container_new(sparsearray2d_double) del sparsearray2d_double self.assertIsNotNone(refindptr()) self.assertIsNotNone(refdata()) self.assertIsNotNone(refindices()) test.test_sbasearray2d_container_clear() self.assertIsNone(refindptr()) self.assertIsNone(refdata()) self.assertIsNone(refindices()) array2d_double = np.array([[1.2, 3], [4, 5]]) ref = weakref.ref(array2d_double) test.test_sbasearray2d_container_new(array2d_double) del array2d_double self.assertIsNotNone(ref()) test.test_sbasearray2d_container_clear() self.assertIsNone(ref()) if __name__ == "__main__": unittest.main()
36.652291
175
0.599941
4e73eaef843757f7ea7a8bbd35f9c54ff770774c
6,878
py
Python
chatbot/interact.py
VictorDebray/RoadBuddy
9c62e2acd2d540caa0ebefc50af5446c0d4f864f
[ "MIT" ]
null
null
null
chatbot/interact.py
VictorDebray/RoadBuddy
9c62e2acd2d540caa0ebefc50af5446c0d4f864f
[ "MIT" ]
null
null
null
chatbot/interact.py
VictorDebray/RoadBuddy
9c62e2acd2d540caa0ebefc50af5446c0d4f864f
[ "MIT" ]
null
null
null
# Author: DINDIN Meryll # Date: 15 September 2019 # Project: RoadBuddy try: from chatbot.imports import * except: from imports import * class Contextualizer: def __init__(self): try: self._load_models() except: drc = ['models', 'datasets'] for d in drc: if not os.path.exists(d): os.mkdir(d) self._download_models() self._load_models() def _download_models(self): s3 = boto3.client('s3') # Download dataset fle = ('datasets.huggingface.co', 'personachat/personachat_self_original.json') s3.download_file(*fle, 'datasets/persona-chat.json') # Download model fle = ('models.huggingface.co', 'transfer-learning-chatbot/finetuned_chatbot_gpt.tar.gz') s3.download_file(fle, 'models/gpt.tar.gpz') with tarfile.open('models/gpt.tar.gpz', 'r:gz') as archive: archive.extractall('models') # Remove tar file os.remove('models/gpt.tar.gpz') def _load_models(self): self.token = OpenAIGPTTokenizer.from_pretrained('models') self.model = OpenAIGPTLMHeadModel.from_pretrained('models') def tokenize_personnalities(self): with open('datasets/persona-chat.json', encoding='utf-8') as f: dtb = json.loads(f.read()) def tokenize(obj): if isinstance(obj, str): return self.token.convert_tokens_to_ids(self.token.tokenize(obj)) if isinstance(obj, dict): return dict((n, tokenize(o)) for n, o in obj.items()) return list(tokenize(o) for o in obj) dtb = tokenize(dtb) torch.save(dtb, 'datasets/persona-cached') class Trigger: def __init__(self): self.url_jokes = 'https://icanhazdadjoke.com' self.url_facts = 'https://some-random-api.ml/facts' def get(self, message): if 'joke' in message: return requests.get(self.url_jokes, headers={"Accept":"application/json"}).json()['joke'] elif ('fun' in message) and ('fact' in message): animal = np.random.choice(['panda', 'cat', 'dog', 'fox', 'bird', 'koala']) return json.loads(requests.get('/'.join([self.url_facts, animal])).content)['fact'] else: return '' class Runner: SPECIAL_TOKENS = ["<bos>", "<eos>", "<speaker1>", "<speaker2>", "<pad>"] def __init__(self, directory='models'): self.hists = [] self.trigs = Trigger() self.token = OpenAIGPTTokenizer.from_pretrained(directory) self.model = OpenAIGPTLMHeadModel.from_pretrained(directory) def set_background(self, characteristics): self.perso = [self.token.convert_tokens_to_ids(self.token.tokenize(e)) for e in characteristics] def read_background(self): for e in self.token.decode(chain(*self.perso)): print('-', e) def input_from_segments(self, history, reply): bos, eos, speaker1, speaker2 = self.token.convert_tokens_to_ids(self.SPECIAL_TOKENS[:-1]) instance = {} sequence = [[bos] + list(chain(*self.perso))] + history + [reply] sequence = [sequence[0]] + [[speaker2 if (len(sequence)-i) % 2 else speaker1] + s for i, s in enumerate(sequence[1:])] instance["input_ids"] = list(chain(*sequence)) instance["token_type_ids"] = [speaker2 if i % 2 else speaker1 for i, s in enumerate(sequence) for _ in s] instance["mc_token_ids"] = len(instance["input_ids"]) - 1 instance["lm_labels"] = [-1] * len(instance["input_ids"]) return instance, sequence @staticmethod def top_filtering(logits, top_k=0, top_p=0.9, threshold=-float('Inf'), filter_value=-float('Inf')): assert logits.dim() == 1 top_k = min(top_k, logits.size(-1)) if top_k > 0: # Remove all tokens with a probability less than the last token in the top-k tokens indices_to_remove = logits < torch.topk(logits, top_k)[0][..., -1, None] logits[indices_to_remove] = filter_value if top_p > 0.0: # Compute cumulative probabilities of sorted tokens sorted_logits, sorted_indices = torch.sort(logits, descending=True) cumulative_probabilities = torch.cumsum(F.softmax(sorted_logits, dim=-1), dim=-1) # Remove tokens with cumulative probability above the threshold sorted_indices_to_remove = cumulative_probabilities > top_p # Shift the indices to the right to keep also the first token above the threshold sorted_indices_to_remove[..., 1:] = sorted_indices_to_remove[..., :-1].clone() sorted_indices_to_remove[..., 0] = 0 # Back to unsorted indices and set them to -infinity indices_to_remove = sorted_indices[sorted_indices_to_remove] logits[indices_to_remove] = filter_value indices_to_remove = logits < threshold logits[indices_to_remove] = filter_value return logits def sample_sequence(self, history, min_length=1, max_length=30, temperature=0.7, current_output=None): special_tokens_ids = self.token.convert_tokens_to_ids(self.SPECIAL_TOKENS) if current_output is None: current_output = [] for i in range(max_length): instance, sequence = self.input_from_segments(history, current_output) input_ids = torch.tensor(instance["input_ids"], device='cpu').unsqueeze(0) token_type_ids = torch.tensor(instance["token_type_ids"], device='cpu').unsqueeze(0) logits = self.model(input_ids, token_type_ids=token_type_ids) logits = logits[0, -1, :] / temperature logits = self.top_filtering(logits) probs = F.softmax(logits, dim=-1) prev = torch.multinomial(probs, 1) if i < 1 and prev.item() in special_tokens_ids: while prev.item() in special_tokens_ids: prev = torch.multinomial(probs, num_samples=1) if prev.item() in special_tokens_ids: break current_output.append(prev.item()) return current_output def answer(self, message, time=4): self.hists.append(self.token.encode(message)) with torch.no_grad(): out_ids = self.sample_sequence(self.hists) response = self.token.decode(out_ids, skip_special_tokens=True) response = ' '.join([response, self.trigs.get(message)]) self.hists.append(self.token.encode(response)) self.hists = self.hists[-time:] return response
38
126
0.604536
4e744f117ca425c6d830404575c231f03329052e
20,731
py
Python
workspace/plug/maya/lynxinode/scripts/python/lxCommand/template/nodeTemplate.py
no7hings/Lynxi
43c745198a714c2e5aca86c6d7a014adeeb9abf7
[ "MIT" ]
2
2018-03-06T03:33:55.000Z
2019-03-26T03:25:11.000Z
workspace/plug/maya/lynxinode/scripts/python/lxCommand/template/nodeTemplate.py
no7hings/lynxi
43c745198a714c2e5aca86c6d7a014adeeb9abf7
[ "MIT" ]
null
null
null
workspace/plug/maya/lynxinode/scripts/python/lxCommand/template/nodeTemplate.py
no7hings/lynxi
43c745198a714c2e5aca86c6d7a014adeeb9abf7
[ "MIT" ]
null
null
null
# encoding=utf-8 import re # import types # noinspection PyUnresolvedReferences import maya.mel as mel # noinspection PyUnresolvedReferences import maya.cmds as cmds # _objectStore = {} # def pyToMelProc(pyObj, args=(), returnType=None, procName=None, useName=False, procPrefix='pyToMel_'): melParams = [] pyParams = [] melReturn = returnType if returnType else '' # for t, n in args: melParams.append('%s $%s' % (t, n)) # if t == 'string': pyParams.append(r"""'"+$%s+"'""" % n) else: pyParams.append(r'"+$%s+"' % n) # objId = id(pyObj) # d = {} # if procName: d['procName'] = procName elif useName: d['procName'] = pyObj.__name__ else: if isinstance(pyObj, types.LambdaType): procPrefix += '_lambda' elif isinstance(pyObj, (types.FunctionType, types.BuiltinFunctionType)): try: procPrefix += '_' + pyObj.__name__ except (AttributeError, TypeError): pass elif isinstance(pyObj, types.MethodType): try: procPrefix += '_' + pyObj.im_class.__name__ + '_' + pyObj.__name__ except (AttributeError, TypeError): pass d['procName'] = '%s%s' % (procPrefix, objId) # d['procName'] = d['procName'].replace('<', '_').replace('>', '_').replace('-', '_') d['melParams'] = ', '.join(melParams) d['pyParams'] = ', '.join(pyParams) d['melReturn'] = melReturn d['thisModule'] = __name__ d['id'] = objId # contents = '''global proc %(melReturn)s %(procName)s(%(melParams)s){''' if melReturn: contents += 'return ' contents += '''python("import %(thisModule)s;%(thisModule)s._objectStore[%(id)s](%(pyParams)s)");}''' mel.eval(contents % d) _objectStore[objId] = pyObj return d['procName'] # def capitalize(s): return s[0].upper() + s[1:] if s else s # def prettify(s): return ' '.join([capitalize(x) for x in re.findall('[a-zA-Z][a-z]*[0-9]*', s)]) # def toCamelCase(s): parts = s.split('_') return ''.join([parts[0]] + [capitalize(x) for x in parts[1:]]) # def aeCallback(func): return pyToMelProc(func, [('string', 'nodeName')], procPrefix='AECallback') # def attrTextFieldGrp(*args, **kwargs): attribute = kwargs.pop('attribute', kwargs.pop('a', None)) assert attribute is not None, "You Must Passed an Attribute" # changeCommand = kwargs.pop('changeCommand', kwargs.pop('cc', None)) if changeCommand: # noinspection PyCallingNonCallable def cc(newVal): cmds.setAttr(attribute, newVal, type="string") changeCommand(newVal) else: def cc(newVal): cmds.setAttr(attribute, newVal, type="string") # if kwargs.pop('edit', kwargs.pop('e', False)): ctrl = args[0] cmds.textFieldGrp( ctrl, edit=True, text=cmds.getAttr(attribute), changeCommand=cc ) cmds.scriptJob( parent=ctrl, replacePrevious=True, attributeChange=[attribute, lambda: cmds.textFieldGrp(ctrl, edit=True, text=cmds.getAttr(attribute))] ) elif kwargs.pop('query', kwargs.pop('q', False)): pass else: labelText = kwargs.pop('label', None) if not labelText: labelText = mel.eval('interToUI(\"{}\")'.format(attribute.split('.')[-1])) # ctrl = None if len(args) > 0: ctrl = args[0] cmds.textFieldGrp( ctrl, label=labelText, text=cmds.getAttr(attribute), changeCommand=cc ) else: ctrl = cmds.textFieldGrp( label=labelText, text=cmds.getAttr(attribute), changeCommand=cc ) # cmds.scriptJob( parent=ctrl, attributeChange=[attribute, lambda: cmds.textFieldGrp(ctrl, edit=True, text=cmds.getAttr(attribute))] ) return ctrl # def attrType(attr): t = cmds.getAttr(attr, type=True) if t == 'float3': node, at = attr.split('.', 1) if cmds.attributeQuery(at, node=node, usedAsColor=1): t = 'color' return t # def modeMethod(func): def wrapped(self, *args, **kwargs): modeFunc = getattr(self._mode, func.__name__) if self._record: self._actions.append((modeFunc, args, kwargs)) else: modeFunc(*args, **kwargs) # wrapped.__doc__ = func.__doc__ wrapped.__name__ = func.__name__ wrapped._orig = func return wrapped # def modeAttrMethod(func): def wrapped(self, attr, *args, **kwargs): assert isinstance(attr, basestring), "%r.%s: attr argument must be a string, got %s" % (self, func.__name__, type(attr).__name__) modeFunc = getattr(self._mode, func.__name__) if self.convertToMayaStyle: attr = toCamelCase(attr) if self._record: self._actions.append((modeFunc, (attr,) + args, kwargs)) else: modeFunc(attr, *args, **kwargs) self._attributes.append(attr) # wrapped.__doc__ = func.__doc__ wrapped.__name__ = func.__name__ wrapped._orig = func return wrapped # def swatchLabel(nodeName): nodeType = cmds.nodeType(nodeName) classificationsList = cmds.getClassification(nodeType) for classification in classificationsList: allClassList = classification.split(':') for allClass in allClassList: classList = allClass.split('/') if 'swatch' == classList[0]: continue else: if classList: if 'shader' != classList[-1]: classList = filter(lambda x: x != 'shader', classList) return "\n".join(map(lambda x: x.capitalize(), classList)) else: return "Sample" # def swatchDisplayNew(plugName): nodeAndAttrs = plugName.split(".") node = nodeAndAttrs[0] cmds.formLayout('swatchDisplayForm') cmds.text('swatchLabel', label=swatchLabel(node)) cmds.swatchDisplayPort('swatchDisplay', wh=(64, 64), rs=64) # cmds.popupMenu('swatchPopup', button=3) cmds.menuItem('swatchSmall', label='Small') cmds.menuItem('swatchMedium', label='Medium') cmds.menuItem('swatchLarge', label='Large') # cmds.setParent(upLevel=True) gTextColumnWidthIndex = mel.eval("$tempVar=$gTextColumnWidthIndex;") cmds.formLayout( 'swatchDisplayForm', edit=True, af=[ ('swatchLabel', "top", 0), ('swatchLabel', "bottom", 0), ('swatchDisplay', "top", 0), ('swatchDisplay', "bottom", 0) ], aof=[ ('swatchLabel', "right", -gTextColumnWidthIndex) ], an=[ ('swatchLabel', "left"), ('swatchDisplay', "right") ], ac=[ ('swatchDisplay', "left", 5, 'swatchLabel') ] ) swatchDisplayReplace(plugName) # def swatchDisplayReplace(plugName): nodeAndAttrs = plugName.split(".") node = nodeAndAttrs[0] # cmds.swatchDisplayPort( 'swatchDisplay', edit=True, shadingNode=node, annotation='Refresh Swatch', pressCommand=lambda *args: mel.eval("updateFileNodeSwatch " + node) ) cmds.popupMenu('swatchPopup', edit=True, button=3) cmds.menuItem( 'swatchSmall', edit=True, command=lambda *args: cmds.swatchDisplayPort('swatchDisplay', edit=True, wh=(64, 64), rs=64) ) cmds.menuItem( 'swatchMedium', edit=True, command=lambda *args: cmds.swatchDisplayPort('swatchDisplay', edit=True, wh=(96, 96), rs=96) ) cmds.menuItem( 'swatchLarge', edit=True, command=lambda *args: cmds.swatchDisplayPort('swatchDisplay', edit=True, wh=(128, 128), rs=128) ) cmds.text('swatchLabel', edit=True, label=swatchLabel(node)) # class baseMode(object): def __init__(self, template): self.template = template @property def nodeName(self): return self.template.nodeName @property def attr(self): return self.template.attr # def nodeType(self): self.template.nodeType() # def nodeAttr(self, attr): return self.template.nodeAttr(attr) # def nodeAttrExists(self, attr): return self.template.nodeAttrExists(attr) # class rootMode(baseMode): def __init__(self, template): super(rootMode, self).__init__(template) # self._attr = None # self._nodeName = None self._type = self.template.nodeType() # def _updateCallback(self, nodeAttr): self.template._doUpdate(nodeAttr.split('.')[0]) # def preSetup(self): self.addCustom('message', self._updateCallback, self._updateCallback) # def postSetup(self): pass # def update(self): pass # def addTemplate(self, attr, template): if template._isRootMode(): template._doSetup(self.nodeAttr(attr)) else: self.addChildTemplate(attr, template) @staticmethod def addChildTemplate(attr, template): template._setToChildMode() template._record = True template.setup() for attr in template._attributes: try: cmds.editorTemplate(suppress=attr) except RuntimeError: pass cmds.editorTemplate( aeCallback(template._doSetup), aeCallback(template._doUpdate), attr, callCustom=True ) @staticmethod def addControl(attr, label=None, changeCommand=None, annotation=None, preventOverride=False, dynamic=False, enumeratedItem=None): if not label: label = prettify(attr) # args = [attr] kwargs = {} # if dynamic: kwargs['addDynamicControl'] = True else: kwargs['addControl'] = True if changeCommand: if hasattr(changeCommand, '__call__'): changeCommand = aeCallback(changeCommand) args.append(changeCommand) if label: kwargs['label'] = label if annotation: kwargs['annotation'] = annotation cmds.editorTemplate(*args, **kwargs) @staticmethod def suppress(attr): cmds.editorTemplate(suppress=attr) @staticmethod def addCustom(attr, newFunc, replaceFunc): if hasattr(newFunc, '__call__'): newFunc = aeCallback(newFunc) if hasattr(replaceFunc, '__call__'): replaceFunc = aeCallback(replaceFunc) args = (newFunc, replaceFunc, attr) cmds.editorTemplate(callCustom=1, *args) @staticmethod def addSeparator(): cmds.editorTemplate(addSeparator=True) @staticmethod def dimControl(nodeName, control, state): cmds.editorTemplate(dimControl=(nodeName, control, state)) @staticmethod def beginLayout(name, collapse=True): cmds.editorTemplate(beginLayout=name, collapse=collapse) @staticmethod def endLayout(): cmds.editorTemplate(endLayout=True) @staticmethod def beginScrollLayout(): cmds.editorTemplate(beginScrollLayout=True) @staticmethod def endScrollLayout(): cmds.editorTemplate(endScrollLayout=True) @staticmethod def beginNoOptimize(): cmds.editorTemplate(beginNoOptimize=True) @staticmethod def endNoOptimize(): cmds.editorTemplate(endNoOptimize=True) @staticmethod def interruptOptimize(): cmds.editorTemplate(interruptOptimize=True) @staticmethod def addComponents(): cmds.editorTemplate(addComponents=True) @staticmethod def addExtraControls(label=None): kwargs = {} if label: kwargs['extraControlsLabel'] = label cmds.editorTemplate(addExtraControls=True, **kwargs) # class AttrControlGrp(object): uiTypeDic = { 'float': cmds.attrFieldSliderGrp, 'float2': cmds.attrFieldGrp, 'float3': cmds.attrFieldGrp, 'color': cmds.attrColorSliderGrp, 'bool': cmds.attrControlGrp, 'long': cmds.attrFieldSliderGrp, 'byte': cmds.attrFieldSliderGrp, 'long2': cmds.attrFieldGrp, 'long3': cmds.attrFieldGrp, 'short': cmds.attrFieldSliderGrp, 'short2': cmds.attrFieldGrp, 'short3': cmds.attrFieldGrp, 'enum': cmds.attrEnumOptionMenuGrp, 'double': cmds.attrFieldSliderGrp, 'double2': cmds.attrFieldGrp, 'double3': cmds.attrFieldGrp, 'string': attrTextFieldGrp, 'message': cmds.attrNavigationControlGrp } def __init__(self, attribute, *args, **kwargs): self.attribute = attribute self.type = kwargs.pop('type', kwargs.pop('typ', None)) if not self.type: self.type = attrType(self.attribute) if self.type in ['color', 'enum', 'message']: self.callback = kwargs.pop('changeCommand', None) else: self.callback = None kwargs['attribute'] = self.attribute if self.type not in self.uiTypeDic: return cmd = self.uiTypeDic[self.type] try: self.control = cmd(*args, **kwargs) except RuntimeError: print "Error creating %s:" % cmd.__name__ raise if self.callback: cmds.scriptJob( attributeChange=[self.attribute, self.callback], replacePrevious=True, parent=self.control ) # def edit(self, **kwargs): kwargs['edit'] = True if self.type not in self.uiTypeDic: return self.uiTypeDic[self.type](self.control, **kwargs) # def setAttribute(self, attribute): self.attribute = attribute if self.type not in self.uiTypeDic: return self.uiTypeDic[self.type](self.control, edit=True, attribute=self.attribute) if self.callback: cmds.scriptJob( attributeChange=[self.attribute, self.callback], replacePrevious=True, parent=self.control ) # class childMode(baseMode): def __init__(self, template): super(childMode, self).__init__(template) self._controls = [] self._layoutStack = [] # def preSetup(self): cmds.setUITemplate('attributeEditorTemplate', pushTemplate=True) self._layoutStack = [cmds.setParent(query=True)] @staticmethod def postSetup(): cmds.setUITemplate(popTemplate=True) # def update(self): cmds.setUITemplate('attributeEditorTemplate', pushTemplate=True) try: for attr, updateFunc, parent in self._controls: cmds.setParent(parent) updateFunc(self.nodeAttr(attr)) except: print("Template %r Failed to Update Attribute '%s'" % (self.template, self.attr)) raise finally: cmds.setUITemplate(popTemplate=True) # def addTemplate(self, attr, template): self.addChildTemplate(attr, template) # def addChildTemplate(self, attr, template): template._setToChildMode() template._record = True template.setup() for attr in template._attributes: try: cmds.editorTemplate(suppress=attr) except RuntimeError: pass self.addCustom(attr, template._doSetup, template._doUpdate) # def addControl(self, attr, label=None, changeCommand=None, annotation=None, preventOverride=False, dynamic=False, enumeratedItem=None): if not label: label = prettify(attr) # kwargs = {'label': label, 'attribute': self.nodeAttr(attr)} if annotation: kwargs['annotation'] = annotation if changeCommand: kwargs['changeCommand'] = changeCommand if enumeratedItem: kwargs['enumeratedItem'] = enumeratedItem parent = self._layoutStack[-1] cmds.setParent(parent) control = AttrControlGrp(**kwargs) self._controls.append((attr, control.setAttribute, parent)) # def addCustom(self, attr, createFunc, updateFunc): parent = self._layoutStack[-1] cmds.setParent(parent) col = cmds.columnLayout(adj=True) # createFunc(self.nodeAttr(attr)) cmds.setParent(parent) self._controls.append((attr, updateFunc, col)) @staticmethod def addSeparator(): cmds.separator() # def beginLayout(self, label, **kwargs): kwargs['label'] = label cmds.setParent(self._layoutStack[-1]) cmds.frameLayout(**kwargs) self._layoutStack.append(cmds.columnLayout(adjustableColumn=True)) # def endLayout(self): self._layoutStack.pop() cmds.setParent(self._layoutStack[-1]) # def beginNoOptimize(self): pass # def endNoOptimize(self): pass # def beginScrollLayout(self): pass # def endScrollLayout(self): pass # def addExtraControls(self): pass # class baseTemplate(object): def __init__(self, nodeType): self._type = nodeType self._nodeName = None self._attr = None # def __repr__(self): return '%s(%r)' % (self.__class__.__name__, self._type) @property def nodeName(self): return self._nodeName @property def attr(self): return self._attr # def nodeType(self): if self._type is None: self._type = cmds.objectType(self.nodeName) return self._type # def nodeAttr(self, attr=None): if attr is None: attr = self.attr return self.nodeName + '.' + attr # def nodeAttrExists(self, attr): return cmds.addAttr(self.nodeAttr(attr), q=1, ex=1) # class attributeTemplate(baseTemplate): convertToMayaStyle = False def __init__(self, nodeType): super(attributeTemplate, self).__init__(nodeType) self._rootMode = rootMode(self) self._childMode = childMode(self) self._mode = self._rootMode self._actions = [] self._attributes = [] self._record = False # def _setToRootMode(self): self._mode = self._rootMode # def _isRootMode(self): return self._mode == self._rootMode # def _setToChildMode(self): self._mode = self._childMode # def _isChildMode(self): return self._mode == self._childMode # def _setActiveNodeAttr(self, nodeName): parts = nodeName.split('.', 1) self._nodeName = parts[0] if len(parts) > 1: self._attr = parts[1] # def _doSetup(self, nodeAttr): self._setActiveNodeAttr(nodeAttr) self._mode.preSetup() if self._record: for func, args, kwargs in self._actions: func(*args, **kwargs) else: self.setup() self._mode.postSetup() # def _doUpdate(self, nodeAttr): self._setActiveNodeAttr(nodeAttr) self._mode.update() @modeMethod def update(self): pass @modeAttrMethod def addTemplate(self, attr, template): pass @modeAttrMethod def addChildTemplate(self, attr, template): pass @modeAttrMethod def addControl(self, attr, label=None, changeCommand=None, annotation=None, preventOverride=False, dynamic=False, enumeratedItem=None): pass @modeMethod def suppress(self, attr): pass @modeMethod def addSeparator(self): pass @modeAttrMethod def addCustom(self, attr, createFunc, updateFunc): pass @modeMethod def beginLayout(self, label, **kwargs): pass @modeMethod def endLayout(self): pass @modeMethod def beginNoOptimize(self): pass @modeMethod def endNoOptimize(self): pass @modeMethod def beginScrollLayout(self): pass @modeMethod def endScrollLayout(self): pass @modeMethod def addExtraControls(self): pass # def addSwatch(self): self.addCustom("message", swatchDisplayNew, swatchDisplayReplace) # For Override def setup(self): pass
29.871758
139
0.589262
4e751966b10b05f698edd3d37469d6c2ff784045
31
py
Python
bubble_io/__init__.py
jasontyping/bubble-io-python
487dd253e85814a012df4a5a5a6a08f023517641
[ "MIT" ]
null
null
null
bubble_io/__init__.py
jasontyping/bubble-io-python
487dd253e85814a012df4a5a5a6a08f023517641
[ "MIT" ]
null
null
null
bubble_io/__init__.py
jasontyping/bubble-io-python
487dd253e85814a012df4a5a5a6a08f023517641
[ "MIT" ]
1
2020-10-25T08:31:59.000Z
2020-10-25T08:31:59.000Z
from .bubbleio import BubbleIo
15.5
30
0.83871
4e769aee426de55532dd683d9dd832dcae724616
68
py
Python
python/pandas_pbf/core.py
ccharlesgb/pandas-pbf
8c5b1af2c291cfd485b1296a1a5ba34ddc93d995
[ "MIT" ]
null
null
null
python/pandas_pbf/core.py
ccharlesgb/pandas-pbf
8c5b1af2c291cfd485b1296a1a5ba34ddc93d995
[ "MIT" ]
null
null
null
python/pandas_pbf/core.py
ccharlesgb/pandas-pbf
8c5b1af2c291cfd485b1296a1a5ba34ddc93d995
[ "MIT" ]
null
null
null
import pandas as pd def dump(df: pd.DataFrame) -> bytes: pass
11.333333
36
0.661765
4e76f4bcaf6c2b3ef6bdb2c9d12ef79f80ffb1ec
13,152
py
Python
iceprod/server/rest/datasets.py
WIPACrepo/iceprod
83615da9b0e764bc2498ac588cc2e2b3f5277235
[ "MIT" ]
2
2017-01-23T17:12:41.000Z
2019-01-14T13:38:17.000Z
iceprod/server/rest/datasets.py
WIPACrepo/iceprod
83615da9b0e764bc2498ac588cc2e2b3f5277235
[ "MIT" ]
242
2016-05-09T18:46:51.000Z
2022-03-31T22:02:29.000Z
iceprod/server/rest/datasets.py
WIPACrepo/iceprod
83615da9b0e764bc2498ac588cc2e2b3f5277235
[ "MIT" ]
2
2017-03-27T09:13:40.000Z
2019-01-27T10:55:30.000Z
import logging import json import uuid from collections import defaultdict import tornado.web import tornado.httpclient from tornado.platform.asyncio import to_asyncio_future import pymongo import motor from rest_tools.client import RestClient from iceprod.server.rest import RESTHandler, RESTHandlerSetup, authorization from iceprod.server.util import nowstr, dataset_statuses, dataset_status_sort logger = logging.getLogger('rest.datasets') def setup(config, *args, **kwargs): """ Setup method for Dataset REST API. Sets up any database connections or other prerequisites. Args: config (dict): an instance of :py:class:`iceprod.server.config`. Returns: list: Routes for dataset, which can be passed to :py:class:`tornado.web.Application`. """ cfg_rest = config.get('rest',{}).get('datasets',{}) db_cfg = cfg_rest.get('database',{}) # add indexes db = pymongo.MongoClient(**db_cfg).datasets if 'dataset_id_index' not in db.datasets.index_information(): db.datasets.create_index('dataset_id', name='dataset_id_index', unique=True) handler_cfg = RESTHandlerSetup(config, *args, **kwargs) handler_cfg.update({ 'database': motor.motor_tornado.MotorClient(**db_cfg).datasets, }) return [ (r'/datasets', MultiDatasetHandler, handler_cfg), (r'/datasets/(?P<dataset_id>\w+)', DatasetHandler, handler_cfg), (r'/datasets/(?P<dataset_id>\w+)/description', DatasetDescriptionHandler, handler_cfg), (r'/datasets/(?P<dataset_id>\w+)/status', DatasetStatusHandler, handler_cfg), (r'/datasets/(?P<dataset_id>\w+)/priority', DatasetPriorityHandler, handler_cfg), (r'/datasets/(?P<dataset_id>\w+)/jobs_submitted', DatasetJobsSubmittedHandler, handler_cfg), (r'/dataset_summaries/status', DatasetSummariesStatusHandler, handler_cfg), ] class BaseHandler(RESTHandler): """ Base handler for Dataset REST API. """ def initialize(self, database=None, **kwargs): super(BaseHandler, self).initialize(**kwargs) self.db = database class MultiDatasetHandler(BaseHandler): """ Handle multi-group requests. """ @authorization(roles=['admin','client','system','user']) #TODO: figure out how to do auth for each dataset in the list async def get(self): """ Get a dict of datasets. Params (optional): status: | separated list of status filters groups: | separated list of groups to filter on users: | separated list of users to filter on keys: | separated list of keys to return for each dataset Returns: dict: {<dataset_id>: metadata} """ query = {} status = self.get_argument('status', None) if status: query['status'] = {'$in': status.split('|')} groups = self.get_argument('groups', None) if groups: query['group'] = {'$in': groups.split('|')} users = self.get_argument('users', None) if users: query['username'] = {'$in': users.split('|')} projection = {'_id': False} keys = self.get_argument('keys', None) if keys: projection.update({x:True for x in keys.split('|') if x}) ret = {} async for row in self.db.datasets.find(query, projection=projection): k = row['dataset_id'] ret[k] = row self.write(ret) self.finish() @authorization(roles=['admin','user']) # anyone should be able to create a dataset async def post(self): """ Add a dataset. Body should contain all necessary fields for a dataset. """ data = json.loads(self.request.body) # validate first req_fields = { 'description': str, 'jobs_submitted': int, 'tasks_submitted': int, 'tasks_per_job': int, 'group': str, } for k in req_fields: if k not in data: raise tornado.web.HTTPError(400, reason='missing key: '+k) if not isinstance(data[k], req_fields[k]): r = 'key "{}" should be of type {}'.format(k, req_fields[k].__name__) raise tornado.web.HTTPError(400, reason=r) opt_fields = { 'priority': int, 'debug': bool, 'jobs_immutable': bool, 'status': str, } for k in opt_fields: if k in data and not isinstance(data[k], opt_fields[k]): r = 'key "{}" should be of type {}'.format(k, opt_fields[k].__name__) raise tornado.web.HTTPError(400, reason=r) bad_fields = set(data).difference(set(opt_fields).union(req_fields)) if bad_fields: r = 'invalid keys found' raise tornado.web.HTTPError(400, reason=r) if data['jobs_submitted'] == 0 and data['tasks_per_job'] <= 0: r = '"tasks_per_job" must be > 0' raise tornado.web.HTTPError(400, reason=r) elif data['tasks_submitted'] != 0 and data['tasks_submitted'] / data['jobs_submitted'] != data['tasks_per_job']: r = '"tasks_per_job" does not match "tasks_submitted"/"jobs_submitted"' raise tornado.web.HTTPError(400, reason=r) # generate dataset number ret = await self.db.settings.find_one_and_update( {'name': 'dataset_num'}, {'$inc': {'num': 1}}, projection={'num': True, '_id': False}, upsert=True, return_document=pymongo.ReturnDocument.AFTER) dataset_num = ret['num'] # set some fields data['dataset_id'] = uuid.uuid1().hex data['dataset'] = dataset_num if 'status' not in data: data['status'] = 'processing' data['start_date'] = nowstr() data['username'] = self.auth_data['username'] if 'priority' not in data: data['priority'] = 0.5 if 'debug' not in data: data['debug'] = False if 'jobs_immutable' not in data: data['jobs_immutable'] = False # insert ret = await self.db.datasets.insert_one(data) # set auth rules url = '/auths/'+data['dataset_id'] http_client = RestClient(self.auth_url, token=self.module_auth_key) auth_data = { 'read_groups':['admin',data['group'],'users'], 'write_groups':['admin',data['group']], } logger.info('Authorization header: %s', 'bearer '+self.module_auth_key) await http_client.request('PUT', url, auth_data) # return success self.set_status(201) self.set_header('Location', '/datasets/'+data['dataset_id']) self.write({'result': '/datasets/'+data['dataset_id']}) self.finish() class DatasetHandler(BaseHandler): """ Handle dataset requests. """ @authorization(roles=['admin','client','system','pilot'], attrs=['dataset_id:read']) async def get(self, dataset_id): """ Get a dataset. Args: dataset_id (str): the dataset Returns: dict: dataset metadata """ ret = await self.db.datasets.find_one({'dataset_id':dataset_id}, projection={'_id':False}) if not ret: self.send_error(404, reason="Dataset not found") else: self.write(ret) self.finish() class DatasetDescriptionHandler(BaseHandler): """ Handle dataset description updates. """ @authorization(roles=['admin'], attrs=['dataset_id:write']) async def put(self, dataset_id): """ Set a dataset description. Args: dataset_id (str): the dataset Returns: dict: empty dict """ data = json.loads(self.request.body) if 'description' not in data: raise tornado.web.HTTPError(400, reason='missing description') elif not isinstance(data['description'],str): raise tornado.web.HTTPError(400, reason='bad description') ret = await self.db.datasets.find_one_and_update({'dataset_id':dataset_id}, {'$set':{'description': data['description']}}, projection=['_id']) if not ret: self.send_error(404, reason="Dataset not found") else: self.write({}) self.finish() class DatasetStatusHandler(BaseHandler): """ Handle dataset status updates. """ @authorization(roles=['admin','system','client'], attrs=['dataset_id:write']) async def put(self, dataset_id): """ Set a dataset status. Args: dataset_id (str): the dataset Returns: dict: empty dict """ data = json.loads(self.request.body) if 'status' not in data: raise tornado.web.HTTPError(400, reason='missing status') elif data['status'] not in dataset_statuses: raise tornado.web.HTTPError(400, reason='bad status') ret = await self.db.datasets.find_one_and_update({'dataset_id':dataset_id}, {'$set':{'status': data['status']}}, projection=['_id']) if not ret: self.send_error(404, reason="Dataset not found") else: self.write({}) self.finish() class DatasetPriorityHandler(BaseHandler): """ Handle dataset priority updates. """ @authorization(roles=['admin','system','client'], attrs=['dataset_id:write']) async def put(self, dataset_id): """ Set a dataset priority. Args: dataset_id (str): the dataset Returns: dict: empty dict """ data = json.loads(self.request.body) if 'priority' not in data: raise tornado.web.HTTPError(400, reason='missing priority') elif not isinstance(data['priority'], (int, float)): raise tornado.web.HTTPError(400, reason='priority is not a number') ret = await self.db.datasets.find_one_and_update({'dataset_id':dataset_id}, {'$set':{'priority': data['priority']}}, projection=['_id']) if not ret: self.send_error(404, reason="Dataset not found") else: self.write({}) self.finish() class DatasetJobsSubmittedHandler(BaseHandler): """ Handle dataset jobs_submitted updates. """ @authorization(roles=['admin'], attrs=['dataset_id:write']) async def put(self, dataset_id): """ Set a dataset's jobs_submitted. Only allows increases, if the jobs_immutable flag is not set. Args: dataset_id (str): the dataset Json body: jobs_submitted (int): the number of jobs submitted Returns: dict: empty dict """ data = json.loads(self.request.body) if 'jobs_submitted' not in data: raise tornado.web.HTTPError(400, reason='missing jobs_submitted') try: jobs_submitted = int(data['jobs_submitted']) except Exception: raise tornado.web.HTTPError(400, reason='jobs_submitted is not an int') ret = await self.db.datasets.find_one({'dataset_id':dataset_id}) if not ret: raise tornado.web.HTTPError(404, reason='Dataset not found') if ret['jobs_immutable']: raise tornado.web.HTTPError(400, reason='jobs_submitted is immutable') if ret['jobs_submitted'] > jobs_submitted: raise tornado.web.HTTPError(400, reason='jobs_submitted must be larger than before') if 'tasks_per_job' not in ret or ret['tasks_per_job'] <= 0: raise tornado.web.HTTPError(400, reason='tasks_per_job not valid') ret = await self.db.datasets.find_one_and_update({'dataset_id':dataset_id}, {'$set':{ 'jobs_submitted': jobs_submitted, 'tasks_submitted': int(jobs_submitted*ret['tasks_per_job']), }}, projection=['_id']) if not ret: self.send_error(404, reason="Dataset not found") else: self.write({}) self.finish() class DatasetSummariesStatusHandler(BaseHandler): """ Handle dataset summary grouping by status. """ @authorization(roles=['admin','system','client','user']) #TODO: figure out how to do auth for each dataset in the list async def get(self): """ Get the dataset summary for all datasets, group by status. Returns: dict: {<status>: [<dataset_id>,]} """ cursor = self.db.datasets.find( projection={'_id':False,'status':True,'dataset_id':True}) ret = defaultdict(list) async for row in cursor: ret[row['status']].append(row['dataset_id']) ret2 = {} for k in sorted(ret, key=dataset_status_sort): ret2[k] = ret[k] self.write(ret2) self.finish()
34.978723
122
0.586603
4e79f990859c3061f129402b3a92ec843ee5ea60
2,938
py
Python
backend/utils/n9e_api.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
null
null
null
backend/utils/n9e_api.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
6
2021-03-19T10:20:05.000Z
2021-09-22T19:30:21.000Z
backend/utils/n9e_api.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # author: itimor import requests import json try: from urllib.parse import urlencode except ImportError: from urllib import urlencode class FalconClient(object): def __init__(self, endpoint=None, user=None, token=None, keys=[], session=None, ssl_verify=True): self._endpoint = endpoint self._job_prex = 'job/' self._url_suffix = 'api/json' self._keys = keys self._session = session self.ssl_verify = ssl_verify if not session: params = { "name": user, "password": token } self._session = requests.Session() ret = self.do_request('get', '/', params=params) print(ret) api_token = { "name": user, "sig": ret.get("sig") } self._session.auth = (user, token) self._session.headers.update({ 'Content-Type': 'application/json; charset=utf-8', 'Accept': 'application/json', 'Apitoken': json.dumps(api_token) }) def __getattr__(self, key): if key in self.__dict__: return self.__dict__[key] return self.__class__( endpoint=self._endpoint, keys=self._keys + [key], session=self._session, ssl_verify=self.ssl_verify) def __getitem__(self, key): """Look up an option value and perform string substitution.""" return self.__getattr__(key) def __call__(self, **kwargs): method = self._keys[-1] url = "/".join(self._keys[0:-1]) url = url.replace("_", "-") return self.do_request(method, url, **kwargs) def do_request(self, method, url, params=None, data=None): url = self._endpoint + url + self._url_suffix if data: print(data) if params is None: params = {} if method == 'get' or method == 'list': response = self._session.get(url, params=params, verify=self.ssl_verify) if method == 'post' or method == 'create': response = self._session.post(url, params=params, json=data, verify=self.ssl_verify) if method == 'put' or method == 'update': response = self._session.put(url, json=data, verify=self.ssl_verify) if method == 'delete': response = self._session.delete(url, params=params, json=data, verify=self.ssl_verify) try: body = json.loads(response.text) except ValueError: body = "Get unknow error is [%s]" % response.reason return body if __name__ == '__main__': cli = FalconClient(endpoint="http://n9e.xxoo.com", user='admin', token='11871bd159bd19da9ab624d161c569e3c8') params = {"idents": ["192.168.0.112"]} r = cli.node['2'].endpoint_unbind.post(data=params) print(r)
31.255319
112
0.566372
4e7a0f3a33cc53fa5588171ed4ecc80f2e9996b8
411
py
Python
python/RASCUNHOS/criargrafico.py
raquelmachado4993/omundodanarrativagit
eb8cebcc74514ba8449fab5f9dc5e9a93a826850
[ "MIT" ]
null
null
null
python/RASCUNHOS/criargrafico.py
raquelmachado4993/omundodanarrativagit
eb8cebcc74514ba8449fab5f9dc5e9a93a826850
[ "MIT" ]
null
null
null
python/RASCUNHOS/criargrafico.py
raquelmachado4993/omundodanarrativagit
eb8cebcc74514ba8449fab5f9dc5e9a93a826850
[ "MIT" ]
null
null
null
import matplotlib.pyplot meses = ['Janeiro','Fevereiro','Marco','Abril','Maio','Junho'] valores = [105235, 107697, 110256, 109236, 108859, 109986] matplotlib.pyplot.plot(meses, valores) matplotlib.pyplot.title('Faturamento no primeiro semestre de 2017') matplotlib.pyplot.xlabel('Meses') matplotlib.pyplot.ylabel('Faturamento em R$') matplotlib.pyplot.savefig('grafico.png', dpi=100) matplotlib.pyplot.show()
34.25
67
0.766423
4e7aae2cfbb07486e48d60f16bce3de6949f6366
3,213
py
Python
getHelmstedtGNDs.py
hbeyer/pcp-vd
e8b4903b4188fea5295e7709e25216f10954f23f
[ "MIT" ]
null
null
null
getHelmstedtGNDs.py
hbeyer/pcp-vd
e8b4903b4188fea5295e7709e25216f10954f23f
[ "MIT" ]
null
null
null
getHelmstedtGNDs.py
hbeyer/pcp-vd
e8b4903b4188fea5295e7709e25216f10954f23f
[ "MIT" ]
null
null
null
#!/usr/bin/python3 # -*- coding: utf-8 -*- import mysql.connector mydb = mysql.connector.connect( host = "localhost", user = "root", passwd = "schleichkatze", database = "helmstedt" ) mycursor = mydb.cursor() mycursor.execute("SELECT id, gnd FROM helmstedt.temp_prof_kat") myresult = mycursor.fetchall() gnds = [x[1] for x in myresult if x[1] != None] print('|'.join(gnds)) """ # Eine Liste (geordnet, indexiert und veränderlich) mylist = ['Lerche', 'Schneider', 'Zimmermann', 'Kästner', 'Raabe', 'Schmidt-Glintzer', 'bURSCHEL'] mylist[len(mylist) - 1] = mylist[len(mylist) - 1].swapcase() mylist.append('Ritter Rost') mylist.insert(0, 'Zimmermann') print(mylist) """ """ # Ein Tupel (ist unveränderlich) mytuple = ('Montag', 'Dienstag', 'Mittwoch', 'Donnerstag', 'Freitag', 'Samstag', 'Sonntag') #print(mytuple[3:6]) """ """ # Ein Set (unindexiert und ungeordnet, Elemente sind unveränderlich, können aber vermehrt oder reduziert werden) myset = {'Adenauer', 'Erhard', 'Kiesinger', 'Brandt', 'Schmidt', 'Kohl', 'Schröder', 'Merkel', 'Schulz'} myset.remove('Schulz') myset.add('Kramp-Karrenbauer') for i in myset: print(i) """ """ # Ein Dictionary mydict = {'Mann':'vyras', 'Frau':'moteris','Fisch':'žuvis', 'Biber':'bebras', 'Stadt':'miestas', 'König':'karalius'} for x, y in mydict.items(): print(x + ' heißt auf Litauisch ' + y) """ """ # Eine Datumsoperation import time import datetime time = time.localtime(time.time()) print(time) """ """ # Eine Funktion def makeName(forename, surname, title=""): result = forename + " " + surname if title: result = title + " " + result return result print(makeName("Hartmut", "Beyer", "Magister artium")) """ """ # Eine Klasse class Person: def __init__(self, forename, surname): self.forename = forename self.surename = surname person = Person('Ben', 'Gurion') print(person.forename) """ """ # Eine Klasse class Language: def __init__(self, code): self.codes = { "eng":"Englisch", "ger":"Deutsch", "fre":"Französisch", "rus":"Russisch" } if code not in self.codes: self.name = code return self.name = self.codes[code] lang = Language("rus") print(lang.name) """ """ # Eine Datei aus dem Netz auslesen import urllib.request as ur url = "http://diglib.hab.de/edoc/ed000228/1623_06.xml" fileobject = ur.urlopen(url) string = fileobject.read() print(string) """ """ # Eine XML-Datei parsen import xml.etree.ElementTree as et tree = et.parse('test.xml') root = tree.getroot() nbs = root.findall('.//{http://www.tei-c.org/ns/1.0}rs') name = "" for ent in nbs: if ent.get('type') == 'person': name = str(ent.text).strip() ref = str(ent.get('ref')).strip() print(name + ' - ' + ref) """ """ # Laden und Auslesen einer XML-Datei im Netz import urllib.request as ur import xml.etree.ElementTree as et url = "http://diglib.hab.de/edoc/ed000228/1623_08.xml" fileobject = ur.urlopen(url) tree = et.parse(fileobject) root = tree.getroot() nbs = root.findall('.//{http://www.tei-c.org/ns/1.0}rs') name = "" for ent in nbs: if ent.get('type') == 'person': name = str(ent.text).strip() ref = str(ent.get('ref')).strip() print(name + ' - ' + ref) """
22.787234
116
0.647059
4e7ab87b888bbbe2383cfa6903a948e5d52465e7
9,035
py
Python
tests/test_mapexplorer.py
OCHA-DAP/hdx-scraper-mapexplorer
3fef67376815611657657c6d53ce904b8f9e4550
[ "MIT" ]
null
null
null
tests/test_mapexplorer.py
OCHA-DAP/hdx-scraper-mapexplorer
3fef67376815611657657c6d53ce904b8f9e4550
[ "MIT" ]
null
null
null
tests/test_mapexplorer.py
OCHA-DAP/hdx-scraper-mapexplorer
3fef67376815611657657c6d53ce904b8f9e4550
[ "MIT" ]
null
null
null
#!/usr/bin/python # -*- coding: utf-8 -*- ''' Unit tests for scrapername. ''' import difflib import filecmp from datetime import datetime from os.path import join from tempfile import gettempdir import pytest from hdx.hdx_configuration import Configuration import hdx.utilities.downloader from hdx.utilities.compare import assert_files_same from hdx.utilities.loader import load_json from src.acled import update_lc_acled, update_ssd_acled from mapexplorer import get_valid_names from src.cbpf import update_cbpf from src.fts import update_fts #from src.rowca import update_rowca class TestScraperName: @pytest.fixture(scope='class') def configuration(self): Configuration._create(user_agent='test', hdx_read_only=True, project_config_yaml=join('tests', 'config', 'project_configuration.yml')) @pytest.fixture(scope='class') def folder(self, configuration): return gettempdir() @pytest.fixture(scope='class') def downloader(self): return hdx.utilities.downloader.Download() @pytest.fixture(scope='class') def today(self): return datetime.strptime('2018-01-16', '%Y-%m-%d') @pytest.fixture(scope='class') def lc_country_list(self, configuration): return ['Nigeria'] @pytest.fixture(scope='class') def ssd_country_list(self, configuration): return ['South Sudan'] @pytest.fixture(scope='class') def valid_lc_names(self, downloader): lc_names_url = Configuration.read()['lc_names_url'] return get_valid_names(downloader, lc_names_url, headers=['ISO', 'Name']) @pytest.fixture(scope='class') def replace_lc_values(self, downloader): lc_mappings_url = Configuration.read()['lc_mappings_url'] return downloader.download_tabular_key_value(lc_mappings_url) @pytest.fixture(scope='class') def valid_ssd_adm1_names(self, downloader): ssd_adm1_names_url = Configuration.read()['ssd_adm1_names_url'] return get_valid_names(downloader, ssd_adm1_names_url, headers=['Name']) @pytest.fixture(scope='class') def valid_ssd_adm2_names(self, downloader): ssd_adm2_names_url = Configuration.read()['ssd_adm2_names_url'] return get_valid_names(downloader, ssd_adm2_names_url, headers=['Name']) @pytest.fixture(scope='class') def replace_ssd_values(self, downloader): ssd_mappings_url = Configuration.read()['ssd_mappings_url'] return downloader.download_tabular_key_value(ssd_mappings_url) @pytest.fixture(scope='function') def downloaderfts(self): class Response: @staticmethod def json(): pass class Download: @staticmethod def download(url): response = Response() if url == 'http://lala/plan/country/NGA': def fn(): return load_json(join('tests', 'fixtures', 'FTS_plan_NGA.json')) response.json = fn elif url == 'http://lala/fts/flow?groupby=plan&countryISO3=NGA': def fn(): return load_json(join('tests', 'fixtures', 'FTS_flow_NGA.json')) response.json = fn return response return Download() @pytest.fixture(scope='function') def downloaderrowca(self): class Response: @staticmethod def json(): pass class Download: @staticmethod def download(url): response = Response() if url == 'http://haha/country=3,4,8,9&subcat=4&inclids=yes&final=1&format=json&lng=en': def fn(): return load_json(join('tests', 'fixtures', 'ROWCA_population.json')) response.json = fn elif url == 'http://haha/country=3,4,8,9&subcat=9,10&inclids=yes&final=1&format=json&lng=en': def fn(): return load_json(join('tests', 'fixtures', 'ROWCA_movement.json')) response.json = fn return response return Download() @pytest.fixture(scope='function') def downloadercbpf(self): class Response: @staticmethod def json(): pass class Download: @staticmethod def download(url): response = Response() if url == 'http://mama/ProjectSummary?poolfundAbbrv=SSD19': def fn(): return load_json(join('tests', 'fixtures', 'CBPF_ProjectSummary_SSD.json')) response.json = fn elif url == 'http://mama/Location?poolfundAbbrv=SSD19': def fn(): return load_json(join('tests', 'fixtures', 'CBPF_Location_SSD.json')) response.json = fn return response return Download() def test_lc_acled(self, folder, today, lc_country_list, valid_lc_names, replace_lc_values): resource_updates = dict() filename = 'Lake_Chad_Basin_Recent_Conflict_Events.csv' expected_events = join('tests', 'fixtures', filename) actual_events = join(folder, filename) resource_updates['acled_events'] = {'path': actual_events} filename = 'Lake_Chad_Basin_Recent_Conflict_Event_Total_Fatalities.csv' expected_fatalities = join('tests', 'fixtures', filename) actual_fatalities = join(folder, filename) resource_updates['acled_fatalities'] = {'path': actual_fatalities} update_lc_acled(today, 'https://raw.githubusercontent.com/mcarans/hdxscraper-mapexplorer/master/tests/fixtures/ACLEDNigeria.csv?', lc_country_list, valid_lc_names, replace_lc_values, resource_updates) assert_files_same(expected_events, actual_events) assert_files_same(expected_fatalities, actual_fatalities) def test_ssd_acled(self, folder, today, ssd_country_list, valid_ssd_adm2_names, replace_ssd_values): resource_updates = dict() filename = 'South_Sudan_Recent_Conflict_Events.csv' expected_events = join('tests', 'fixtures', filename) actual_events = join(folder, filename) resource_updates['acled_events'] = {'path': actual_events} filename = 'South_Sudan_Recent_Conflict_Event_Total_Fatalities.csv' expected_fatalities = join('tests', 'fixtures', filename) actual_fatalities = join(folder, filename) resource_updates['acled_fatalities'] = {'path': actual_fatalities} update_ssd_acled(today, 'https://raw.githubusercontent.com/mcarans/hdxscraper-mapexplorer/master/tests/fixtures/ACLEDSouthSudan.csv?', ssd_country_list, valid_ssd_adm2_names, replace_ssd_values, resource_updates) assert_files_same(expected_events, actual_events) assert_files_same(expected_fatalities, actual_fatalities) def test_fts(self, folder, downloaderfts, lc_country_list): resource_updates = dict() filename = 'Lake_Chad_Basin_Appeal_Status.csv' expected = join('tests', 'fixtures', filename) actual = join(folder, filename) resource_updates['fts'] = {'path': actual} update_fts('http://lala/', downloaderfts, lc_country_list, resource_updates) assert_files_same(expected, actual) def test_cbpf(self, folder, today, downloadercbpf, valid_ssd_adm1_names, replace_ssd_values): resource_updates = dict() filename = 'South_Sudan_Country_Based_Pool_Funds.csv' expected = join('tests', 'fixtures', filename) actual = join(folder, filename) resource_updates['cbpf'] = {'path': actual} update_cbpf('http://mama/', downloadercbpf, 'SSD19', today, valid_ssd_adm1_names, replace_ssd_values, resource_updates) assert_files_same(expected, actual) # def test_rowca(self, folder, downloaderrowca, valid_lc_names, replace_lc_values): # resource_updates = dict() # filename = 'Lake_Chad_Basin_Estimated_Population.csv' # expected_population = join('tests', 'fixtures', filename) # actual_population = join(folder, filename) # resource_updates['rowca_population'] = {'path': actual_population} # filename = 'Lake_Chad_Basin_Displaced.csv' # expected_displaced = join('tests', 'fixtures', filename) # actual_displaced = join(folder, filename) # resource_updates['rowca_displaced'] = {'path': actual_displaced} # update_rowca('http://haha/', downloaderrowca, valid_lc_names, replace_lc_values, resource_updates) # assert filecmp.cmp(expected_population, actual_population, shallow=False) is True, 'Expected: %s and Actual: %s do not match!' % (expected_population, actual_population) # assert filecmp.cmp(expected_displaced, actual_displaced, shallow=False) is True, 'Expected: %s and Actual: %s do not match!' % (expected_displaced, actual_displaced)
44.727723
220
0.657554
4e7bf6a5635ca159b39d2d3d8c59aa1c3e27375b
255
py
Python
configs/_base_/schedules/imagenet_bs256_140e.py
sty16/cell_transformer
fc3d8dd8363664381617c76fb016f14c704749d8
[ "Apache-2.0" ]
1
2022-03-15T07:36:04.000Z
2022-03-15T07:36:04.000Z
configs/_base_/schedules/imagenet_bs256_140e.py
sty16/cell_transformer
fc3d8dd8363664381617c76fb016f14c704749d8
[ "Apache-2.0" ]
null
null
null
configs/_base_/schedules/imagenet_bs256_140e.py
sty16/cell_transformer
fc3d8dd8363664381617c76fb016f14c704749d8
[ "Apache-2.0" ]
null
null
null
# optimizer optimizer = dict(type='SGD', lr=0.001, momentum=0.9, weight_decay=0.001) optimizer_config = dict(grad_clip=None) # learning policy lr_config = dict(policy='step', step=[40, 70, 90]) runner = dict(type='EpochBasedRunner', max_epochs=100)
36.428571
73
0.721569
4e81b7168137c8ead27ea61e73b96364b565fc1e
708
py
Python
2016/day_06.py
nabiirah/advent-of-code
9c7e7cae437c024aa05d9cb7f9211fd47f5226a2
[ "MIT" ]
24
2020-12-08T20:07:52.000Z
2022-01-18T20:08:06.000Z
2016/day_06.py
nestorhf/advent-of-code
1bb827e9ea85e03e0720e339d10b3ed8c44d8f27
[ "MIT" ]
null
null
null
2016/day_06.py
nestorhf/advent-of-code
1bb827e9ea85e03e0720e339d10b3ed8c44d8f27
[ "MIT" ]
10
2020-12-04T10:04:15.000Z
2022-02-21T22:22:26.000Z
""" Advent of Code Day 6 - Signals and Noise""" with open('inputs/day_06.txt', 'r') as f: rows = [row.strip() for row in f.readlines()] flipped = zip(*rows) message = '' mod_message = '' for chars in flipped: most_freq = '' least_freq = '' highest = 0 lowest = 100 for char in chars: if chars.count(char) > highest: highest = chars.count(char) most_freq = char if chars.count(char) < lowest: # Part Two lowest = chars.count(char) least_freq = char message += most_freq mod_message += least_freq # Answer One print("Error Corrected Message:", message) # Answer Two print("Modified Message:", mod_message)
22.83871
51
0.601695
4e83689fcdf6c1f0b2f2c351aa1c6fe2dad28771
1,846
py
Python
tasks.py
chmp/misc-exp
2edc2ed598eb59f4ccb426e7a5c1a23343a6974b
[ "MIT" ]
6
2017-10-31T20:54:37.000Z
2020-10-23T19:03:00.000Z
tasks.py
chmp/misc-exp
2edc2ed598eb59f4ccb426e7a5c1a23343a6974b
[ "MIT" ]
7
2020-03-24T16:14:34.000Z
2021-03-18T20:51:37.000Z
tasks.py
chmp/misc-exp
2edc2ed598eb59f4ccb426e7a5c1a23343a6974b
[ "MIT" ]
1
2019-07-29T07:55:49.000Z
2019-07-29T07:55:49.000Z
import hashlib import json import os import pathlib import shlex import nbformat from invoke import task files_to_format = ["chmp/src", "tasks.py", "chmp/setup.py"] inventories = [ "http://daft-pgm.org", "https://matplotlib.org", "http://www.numpy.org", "https://pandas.pydata.org", "https://docs.python.org/3", "https://pytorch.org/docs/stable", ] directories_to_test = ["chmp", "20170813-KeywordDetection/chmp-app-kwdetect"] @task def precommit(c): format(c) docs(c) test(c) @task def test(c): run(c, "pytest", *directories_to_test) @task def docs(c): run( c, *["python", "-m", "chmp.tools", "mddocs"], *(part for inventory in inventories for part in ["--inventory", inventory]), *["chmp/docs/src", "chmp/docs"], ) self_path = pathlib.Path(__file__).parent.resolve() for p in self_path.glob("*/Post.ipynb"): run( c, *["python", "-m", "chmp.tools", "blog"], *[str(p), str(p.with_suffix(".md"))], ) @task def format(c): run(c, "black", *files_to_format) @task def release(c, yes=False): import packaging.version with c.cd("chmp"): run(c, "python", "setup.py", "bdist_wheel") latest_package = max( ( package for package in os.listdir("chmp/dist") if not package.startswith(".") and package.endswith(".whl") ), key=packaging.version.parse, ) if not yes: answer = input(f"upload {latest_package} [yN] ") if answer != "y": print("stop") return with c.cd("chmp/dist"): run(c, "twine", "upload", latest_package) def run(c, *args, **kwargs): args = [shlex.quote(arg) for arg in args] args = " ".join(args) return c.run(args, **kwargs)
20.511111
84
0.566089
4e83cc92299ecc7a687b5a70cfeda857351d4ef2
1,016
py
Python
WeChat/translation.py
satoukiCk/SummerRobot
a22b17fb1927dcc1aa7316e2b892f7daee484583
[ "MIT" ]
null
null
null
WeChat/translation.py
satoukiCk/SummerRobot
a22b17fb1927dcc1aa7316e2b892f7daee484583
[ "MIT" ]
null
null
null
WeChat/translation.py
satoukiCk/SummerRobot
a22b17fb1927dcc1aa7316e2b892f7daee484583
[ "MIT" ]
null
null
null
# -*- coding:utf-8 -*- import requests import json import random import hashlib KEY = '' APPID = '' API = 'http://api.fanyi.baidu.com/api/trans/vip/translate' class translation(): def __init__(self,src, fromlang, tolang): self.src = src self.fromlang = fromlang self.tolang = tolang def trans(self): salt = random.randint(32768,65535) sign = APPID+self.src+str(salt)+KEY m1 = hashlib.md5() m1.update(sign) sign = m1.hexdigest() paras = { 'q':self.src, 'from':self.fromlang, 'to':self.tolang, 'appid':APPID, 'salt':salt, 'sign':sign } result = requests.get(API,params=paras,timeout=50) tdata = json.loads(result.text) res_msg = '' src = tdata['trans_result'][0]['src'] dst = tdata['trans_result'][0]['dst'] res_msg += '源语言: %s\n翻译结果: %s' % (src.encode('utf8'), dst.encode('utf8')) return res_msg
25.4
81
0.541339
4e85a767dc1a77745686895d8e0fa92531e23594
210
py
Python
bolinette/defaults/__init__.py
bolinette/bolinette
b35a7d828c7d9617da6a8d7ac066e3b675a65252
[ "MIT" ]
4
2020-11-02T15:16:32.000Z
2022-01-11T11:19:24.000Z
bolinette/defaults/__init__.py
bolinette/bolinette
b35a7d828c7d9617da6a8d7ac066e3b675a65252
[ "MIT" ]
14
2021-01-04T11:06:59.000Z
2022-03-23T17:01:49.000Z
bolinette/defaults/__init__.py
bolinette/bolinette
b35a7d828c7d9617da6a8d7ac066e3b675a65252
[ "MIT" ]
null
null
null
import bolinette.defaults.models import bolinette.defaults.mixins import bolinette.defaults.services import bolinette.defaults.middlewares import bolinette.defaults.controllers import bolinette.defaults.topics
30
37
0.885714
4e864e061007124f810efb595fdd8cc9331ec714
2,040
py
Python
kicost/currency_converter/currency_converter.py
mdeweerd/KiCost
2f67dad0f8d3335590835a6790181fc6428086d5
[ "MIT" ]
null
null
null
kicost/currency_converter/currency_converter.py
mdeweerd/KiCost
2f67dad0f8d3335590835a6790181fc6428086d5
[ "MIT" ]
null
null
null
kicost/currency_converter/currency_converter.py
mdeweerd/KiCost
2f67dad0f8d3335590835a6790181fc6428086d5
[ "MIT" ]
null
null
null
#!/usr/bin/python3 # -*- coding: utf-8 -*- # Copyright (c) 2021 Salvador E. Tropea # Copyright (c) 2021 Instituto Nacional de Tecnología Industrial # License: Apache 2.0 # Project: KiCost # Adapted from: https://github.com/alexprengere/currencyconverter """ CurrencyConverter: This is reduced version of the 'Currency Converter' by Alex Prengère. Original project: https://github.com/alexprengere/currencyconverter This version only supports conversions for the last exchange rates, not historic ones. On the other hand this version always tries to get the last rates. """ try: from .default_rates import default_rates, default_date except ImportError: # Only useful to boostrap default_rates = {} default_date = '' from .download_rates import download_rates # Author information. __author__ = 'Salvador Eduardo Tropea' __webpage__ = 'https://github.com/set-soft/' __company__ = 'INTI-CMNB - Argentina' class CurrencyConverter(object): def __init__(self): self.initialized = False def _do_init(self): if self.initialized: return self.date, self.rates = download_rates() if not self.date: self.date = default_date self.rates = default_rates self.initialized = True def convert(self, amount, currency, new_currency='EUR'): """Convert amount from a currency to another one. :param float amount: The amount of `currency` to convert. :param str currency: The currency to convert from. :param str new_currency: The currency to convert to. :return: The value of `amount` in `new_currency`. :rtype: float >>> c = CurrencyConverter() >>> c.convert(100, 'EUR', 'USD') """ self._do_init() for c in currency, new_currency: if c not in self.rates: raise ValueError('{0} is not a supported currency'.format(c)) r0 = self.rates[currency] r1 = self.rates[new_currency] return float(amount) / r0 * r1
30.447761
77
0.666667
4e893aecc42c83f372f87792977e579561f4f1e5
385
py
Python
apps/vector.py
HayesAJ83/LeafMapAppTest
5da65d5c1958f47934453124a72ec800c0ce6a93
[ "MIT" ]
22
2021-08-10T05:11:47.000Z
2022-02-27T14:35:30.000Z
apps/vector.py
HayesAJ83/LeafMapAppTest
5da65d5c1958f47934453124a72ec800c0ce6a93
[ "MIT" ]
null
null
null
apps/vector.py
HayesAJ83/LeafMapAppTest
5da65d5c1958f47934453124a72ec800c0ce6a93
[ "MIT" ]
8
2021-10-04T13:10:32.000Z
2021-11-17T12:32:57.000Z
import streamlit as st import leafmap def app(): st.title("Add vector datasets") url = "https://raw.githubusercontent.com/giswqs/data/main/world/world_cities.csv" in_csv = st.text_input("Enter a URL to a vector file", url) m = leafmap.Map() if in_csv: m.add_xy_data(in_csv, x="longitude", y="latitude", layer_name="World Cities") m.to_streamlit()
22.647059
85
0.675325
4e89a4bc599fa5786586aa1889dcbfc722b2e1e8
1,225
py
Python
Book_Ladder/web/test.py
Rdjroot/BookLadder
d4e1f90572f2dda2e7c25890b99c965ded0f02c8
[ "MIT" ]
null
null
null
Book_Ladder/web/test.py
Rdjroot/BookLadder
d4e1f90572f2dda2e7c25890b99c965ded0f02c8
[ "MIT" ]
null
null
null
Book_Ladder/web/test.py
Rdjroot/BookLadder
d4e1f90572f2dda2e7c25890b99c965ded0f02c8
[ "MIT" ]
null
null
null
# -*- coding = utf-8 -*- # @Time:2021/3/1417:56 # @Author:Linyu # @Software:PyCharm from web.pageutils import BooksScore from web.pageutils import BooksCount from web.pageutils import pointsDraw from web.pageutils import scoreRelise from web.pageutils import messBarInfo from web.pageutils import tagRader from web.models import tagThree from web.wdCloud import infoCloud from web.priceSpider import spider from web.models import Dict from web.models import Modle from web.priceSpider import spiderDD #用围城做测试 isbn = "'9787020090006'" dd = "http://search.dangdang.com/?key=%s&act=input&sort_type=sort_xlowprice_asc#J_tab"%(isbn) ddPrice = spiderDD(dd) print(ddPrice) # sql = 'select title from allbook where isbn = %s'%(isbn) # print(sql) # testData = Modle().query(sql) # print(testData[0][0]) # title = "'活着'" # sqlNum = 'select id_num from corebook where title = %s'%(title) # id_num = Modle().query(sqlNum) # print(id_num[0][0]) # print(scoreRelise()) # print(BooksScore()) # print(BooksCount()) # print(pointsDraw()) # messBar() # print(messBar()) # tagRader() # tagThree("小说") # infoCloud('代码大全(第2版)') # print(spider('9787108009821')) # dic = Dict() # for key in dic.keys(): # print(key) # print(dic[key])
24.5
93
0.719184
4e8e84d17eb6d1c3fde6f5fcc279f8f8f53c1518
2,252
py
Python
news/spiders/bbc.py
azzuwan/ScraperExample
6de382df0b20414f2a55b70837b5fd41d76e8712
[ "MIT" ]
null
null
null
news/spiders/bbc.py
azzuwan/ScraperExample
6de382df0b20414f2a55b70837b5fd41d76e8712
[ "MIT" ]
null
null
null
news/spiders/bbc.py
azzuwan/ScraperExample
6de382df0b20414f2a55b70837b5fd41d76e8712
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import scrapy from scrapy.linkextractors import LinkExtractor from scrapy.spiders import CrawlSpider, Rule from readability import Document import datetime from pprint import pprint class Articles(scrapy.Item): url = scrapy.Field() title = scrapy.Field() author = scrapy.Field() published = scrapy.Field() body = scrapy.Field() agency = scrapy.Field() class BbcSpider(CrawlSpider): name = 'bbc' allowed_domains = ['bbc.com'] start_urls = ['http://bbc.com/news'] rules = ( Rule(LinkExtractor(), callback='parse_item', follow=True), ) def parse_item(self, res): title = self.get_title(res) article = Article() # Only do further processing if there is a title element in the page if title != None: article = Articles() article['url'] = res.url article['title']=title article['body']= self.get_body(res) article['published']= self.get_published(res) article['author']= self.get_author(res) article['agency']= self.get_agency(res) # self.log(article) return article else: return None def get_title(self, res): """ Get the title of the article """ title = res.css('h1.story-body__h1 ::text').extract_first() return title def get_body(self, res): """ Get the actual text of the article """ raw = res.css('div.story-body__inner p ::text') body = ''.join(raw.extract()) return body def get_published(self, res): """ Get the article timestamp """ timestamp = res.css('div.story-body div.date ::attr(data-seconds)').extract_first() published = datetime.datetime.fromtimestamp(int(timestamp)) return published def get_author(self, res): """ Get the author of the article. BBC is somewhat shy about putting a name on articles So we just return the string "bbc" """ return 'bbc' def get_agency(self, res): """ Get the agency name """ return 'bbc'
29.246753
96
0.572824
4e8fedac1be1849239d5c90fd7ed2234086dcfe6
698
py
Python
profiles/models.py
ev-agelos/acr-server
dba11b001ae4aae6dcbb761a5c0222c6fb3b939d
[ "MIT" ]
1
2021-03-11T04:25:07.000Z
2021-03-11T04:25:07.000Z
profiles/models.py
ev-agelos/acr-server
dba11b001ae4aae6dcbb761a5c0222c6fb3b939d
[ "MIT" ]
7
2020-03-06T17:37:01.000Z
2021-09-22T17:40:10.000Z
profiles/models.py
ev-agelos/ac-rank
dba11b001ae4aae6dcbb761a5c0222c6fb3b939d
[ "MIT" ]
null
null
null
from django.db import models from django.contrib.auth.models import User from django.db.models.signals import post_save from django.dispatch import receiver from django_countries.fields import CountryField class Profile(models.Model): """Extend user model with a country field.""" user = models.OneToOneField(User, on_delete=models.CASCADE) country = CountryField(blank_label='(select country)', blank=True) @receiver(post_save, sender=User) def create_user_profile(sender, instance, created, **kwargs): if created: Profile.objects.create(user=instance) @receiver(post_save, sender=User) def save_user_profile(sender, instance, **kwargs): instance.profile.save()
29.083333
70
0.767908
4e90e63aeba9851ed0445a458eb6eb560cabb51f
5,684
py
Python
tests/unit/test_command.py
shintaii/flower
fdeb135ddb3718404c0f1e9cca73fc45181f611a
[ "BSD-3-Clause" ]
4,474
2015-01-01T18:34:36.000Z
2022-03-29T06:02:38.000Z
tests/unit/test_command.py
shintaii/flower
fdeb135ddb3718404c0f1e9cca73fc45181f611a
[ "BSD-3-Clause" ]
835
2015-01-06T21:29:48.000Z
2022-03-31T04:35:10.000Z
tests/unit/test_command.py
shintaii/flower
fdeb135ddb3718404c0f1e9cca73fc45181f611a
[ "BSD-3-Clause" ]
980
2015-01-02T21:41:28.000Z
2022-03-31T08:30:52.000Z
import os import sys import tempfile import unittest import subprocess from unittest.mock import Mock, patch import mock from prometheus_client import Histogram from flower.command import apply_options, warn_about_celery_args_used_in_flower_command, apply_env_options from tornado.options import options from tests.unit import AsyncHTTPTestCase class TestFlowerCommand(AsyncHTTPTestCase): def test_task_runtime_metric_buckets_read_from_cmd_line(self): apply_options('flower', argv=['--task_runtime_metric_buckets=1,10,inf']) self.assertEqual([1.0, 10.0, float('inf')], options.task_runtime_metric_buckets) def test_task_runtime_metric_buckets_no_cmd_line_arg(self): apply_options('flower', argv=[]) self.assertEqual(Histogram.DEFAULT_BUCKETS, options.task_runtime_metric_buckets) def test_task_runtime_metric_buckets_read_from_env(self): os.environ["FLOWER_TASK_RUNTIME_METRIC_BUCKETS"] = "2,5,inf" apply_env_options() self.assertEqual([2.0, 5.0, float('inf')], options.task_runtime_metric_buckets) def test_task_runtime_metric_buckets_no_env_value_provided(self): apply_env_options() self.assertEqual(Histogram.DEFAULT_BUCKETS, options.task_runtime_metric_buckets) def test_port(self): with self.mock_option('port', 5555): apply_options('flower', argv=['--port=123']) self.assertEqual(123, options.port) def test_address(self): with self.mock_option('address', '127.0.0.1'): apply_options('flower', argv=['--address=foo']) self.assertEqual('foo', options.address) def test_autodiscovery(self): """ Simulate basic Django setup: - creating celery app - run app.autodiscover_tasks() - create flower command """ celery_app = self._get_celery_app() with mock.patch.object(celery_app, '_autodiscover_tasks') as autodiscover: celery_app.autodiscover_tasks() self.get_app(capp=celery_app) self.assertTrue(autodiscover.called) class TestWarnAboutCeleryArgsUsedInFlowerCommand(AsyncHTTPTestCase): @patch('flower.command.logger.warning') def test_does_not_log_warning(self, mock_warning): mock_app_param = Mock(name='app_param', opts=('-A', '--app')) mock_broker_param = Mock(name='broker_param', opts=('-b', '--broker')) class FakeContext: parent = Mock(command=Mock(params=[mock_app_param, mock_broker_param])) ctx = FakeContext() warn_about_celery_args_used_in_flower_command( ctx=ctx, flower_args=('--port=5678', '--address=0.0.0.0') ) mock_warning.assert_not_called() @patch('flower.command.logger.warning') def test_logs_warning(self, mock_warning): mock_app_param = Mock(name='app_param', opts=('-A', '--app')) mock_broker_param = Mock(name='broker_param', opts=('-b', '--broker')) class FakeContext: parent = Mock(command=Mock(params=[mock_app_param, mock_broker_param])) ctx = FakeContext() warn_about_celery_args_used_in_flower_command( ctx=ctx, flower_args=('--app=proj', '-b', 'redis://localhost:6379/0') ) mock_warning.assert_called_once_with( "You have incorrectly specified the following celery arguments after flower command: " "[\'--app\', \'-b\']. Please specify them after celery command instead following" " this template: celery [celery args] flower [flower args]." ) class TestConfOption(AsyncHTTPTestCase): def test_error_conf(self): with self.mock_option('conf', None): self.assertRaises(IOError, apply_options, 'flower', argv=['--conf=foo']) self.assertRaises(IOError, apply_options, 'flower', argv=['--conf=/tmp/flower/foo']) def test_default_option(self): apply_options('flower', argv=[]) self.assertEqual('flowerconfig.py', options.conf) def test_empty_conf(self): with self.mock_option('conf', None): apply_options('flower', argv=['--conf=/dev/null']) self.assertEqual('/dev/null', options.conf) def test_conf_abs(self): with tempfile.NamedTemporaryFile() as cf: with self.mock_option('conf', cf.name), self.mock_option('debug', False): cf.write('debug=True\n'.encode('utf-8')) cf.flush() apply_options('flower', argv=['--conf=%s' % cf.name]) self.assertEqual(cf.name, options.conf) self.assertTrue(options.debug) def test_conf_relative(self): with tempfile.NamedTemporaryFile(dir='.') as cf: with self.mock_option('conf', cf.name), self.mock_option('debug', False): cf.write('debug=True\n'.encode('utf-8')) cf.flush() apply_options('flower', argv=['--conf=%s' % os.path.basename(cf.name)]) self.assertTrue(options.debug) @unittest.skipUnless(not sys.platform.startswith("win"), 'skip windows') def test_all_options_documented(self): def grep(patter, filename): return int(subprocess.check_output( 'grep "%s" %s|wc -l' % (patter, filename), shell=True)) defined = grep('^define(', 'flower/options.py') - 4 documented = grep('^~~', 'docs/config.rst') self.assertEqual(defined, documented, msg='Missing option documentation. Make sure all options ' 'are documented in docs/config.rst')
39.748252
106
0.645144
4e91557146c36922257c5f4c9ff456b0ce8b407c
534
py
Python
Trojan.py
alrocks29/alpha-backdoor
16a2d0ffdb183005f687bdf19b25cc918a1f12a0
[ "MIT" ]
null
null
null
Trojan.py
alrocks29/alpha-backdoor
16a2d0ffdb183005f687bdf19b25cc918a1f12a0
[ "MIT" ]
null
null
null
Trojan.py
alrocks29/alpha-backdoor
16a2d0ffdb183005f687bdf19b25cc918a1f12a0
[ "MIT" ]
null
null
null
#!/usr/bin/env python import requests import subprocess import os import tempfile def download(url): get_response = requests.get(url) file_name = url.split("/")[-1] with open(file_name, "wb") as out_file: out_file.write(get_response.content) temp_directory = tempfile.gettempdir() os.chdir(temp_directory) download("http://ip/image.jpg") subprocess.Popen("image.jpg", shell=True) download("http://ip/backdoor.exe") subprocess.call("backdoor.exe", shell=True) os.remove("image.jpg") os.remove("backdoor.exe")
21.36
44
0.724719
4e91fe4c0f8f01c97da338f53c30caedc69665c2
3,524
py
Python
Assignments/Assignment_1/Q1/task1.py
Kaustubh1Verma/CS671_Deep-Learning_2019
062002a1369dc962feb52d3c9561a3f1153e0f84
[ "MIT" ]
null
null
null
Assignments/Assignment_1/Q1/task1.py
Kaustubh1Verma/CS671_Deep-Learning_2019
062002a1369dc962feb52d3c9561a3f1153e0f84
[ "MIT" ]
null
null
null
Assignments/Assignment_1/Q1/task1.py
Kaustubh1Verma/CS671_Deep-Learning_2019
062002a1369dc962feb52d3c9561a3f1153e0f84
[ "MIT" ]
1
2019-06-12T14:02:33.000Z
2019-06-12T14:02:33.000Z
import numpy as np import cv2 import math import random import os from tempfile import TemporaryFile from sklearn.model_selection import train_test_split # Creating classes. length=[7,15] width=[1,3] col=[] col.append([0,0,255]) #Blue col.append([255,0,0]) #Red interval=15 angles=[] x=0 while x<180: angles.append(x) x+=interval dirn=1 a1=0 os.mkdir("/home/aj/Desktop/DL2") for l in length: a2=0 #a1 0->7,1->15 for w in width: a3=0 #a2 0->1,1->3 for co in col: a4=0 #a3 0->red,1->blue for ang in angles: flag=0 m=0 os.mkdir("/home/aj/Desktop/DL2/"+str(dirn)) while flag<1000: img=np.zeros((28,28,3),np.uint8) x=random.randrange((28-math.ceil(l*math.sin(math.radians(180-ang))))) y=random.randrange((28-math.ceil(l*math.sin(math.radians(180-ang))))) endy = y+l*math.sin(math.radians(180-ang)) endy=math.floor(endy) endx = x+l*math.cos(math.radians(180-ang)) endx=math.floor(endx) if(0<=endx<=28 and 0<=endy<=28): cv2.line(img,(x,y),(endx,endy),co,w) flag=flag+1 cv2.imwrite("/home/aj/Desktop/DL2/"+str(dirn)+"/"+str(a1)+"_"+str(a2)+"_"+str(a4)+"_"+str(a3)+"_"+str(flag)+".png",img) dirn+=1 a4+=1 a3=a3+1 a2=a2+1 a1=a1+1 outfile = TemporaryFile() # Creating Frames train=[] train_class=[] test_class=[] allimg=[] label=[] flag=0 # os.mkdir("/home/aj/Desktop/DL2/frames") for count in range (1,97): f=[] # os.mkdir("/home/aj/Desktop/DL2/frames/frame_"+str(count)) f=os.listdir("/home/aj/Desktop/DL2/"+str(count)) for fi in f: # print(fi) n=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+fi) n = n.reshape(2352) allimg.append(n) label.append(flag) flag+=1 for i in range (0,10): img1=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i],1) img2=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+1],1) img3=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+2],1) img1f=np.concatenate((img1,img2,img3),axis=1) img4=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+3],1) img5=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+4],1) img6=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+5],1) img2f=np.concatenate((img4,img5,img6),axis=1) img7=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+6],1) img8=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+7],1) img9=cv2.imread("/home/aj/Desktop/DL2/"+str(count)+"/"+f[i+8],1) img3f=np.concatenate((img7,img8,img9),axis=1) imgf=np.concatenate((img1f,img2f,img3f),axis=0) cv2.imwrite("/home/aj/Desktop/DL2/frames/frame_"+str(count)+"/"+"f"+str(i+1)+".png",imgf) # print(allimg[0]) # print(label[0:97]) X_train, X_test, y_oldtrain, y_oldtest = train_test_split(allimg, label, test_size=0.40, random_state=42) # print(y_oldtrain[0:10]) y_oldtrain = np.array(y_oldtrain).reshape(-1) y_train=np.eye(96)[y_oldtrain] y_oldtest = np.array(y_oldtest).reshape(-1) y_test=np.eye(96)[y_oldtest] np.savez_compressed("/home/aj/Desktop/DL2/outfile",X_train=X_train,X_test=X_test,y_train=y_train,y_test=y_test) # Creating Video # img_frame=[] # for i in range (1,97): # f=[] # f=os.listdir("/home/aj/Desktop/DL2/frames/frame_"+str(i)) # path="/home/aj/Desktop/DL2/frames/frame_"+str(i)+"/" # for file in f: # img = cv2.imread(path+file) # height,width,layers = img.shape # size = (width,height) # img_frame.append(img) # out = cv2.VideoWriter("/home/aj/Desktop/DL2/assign1.mp4",0x7634706d,5, size) # for i in range(len(img_frame)): # out.write(img_frame[i]) # out.release()
32.036364
126
0.653235
4e92e1a84cbde5bdb7f5c55409d43bffcb5d668d
17,703
py
Python
octopus_deploy_swagger_client/models/user_role_resource.py
cvent/octopus-deploy-api-client
0e03e842e1beb29b132776aee077df570b88366a
[ "Apache-2.0" ]
null
null
null
octopus_deploy_swagger_client/models/user_role_resource.py
cvent/octopus-deploy-api-client
0e03e842e1beb29b132776aee077df570b88366a
[ "Apache-2.0" ]
null
null
null
octopus_deploy_swagger_client/models/user_role_resource.py
cvent/octopus-deploy-api-client
0e03e842e1beb29b132776aee077df570b88366a
[ "Apache-2.0" ]
null
null
null
# coding: utf-8 """ Octopus Server API No description provided (generated by Swagger Codegen https://github.com/swagger-api/swagger-codegen) # noqa: E501 OpenAPI spec version: 2019.6.7+Branch.tags-2019.6.7.Sha.aa18dc6809953218c66f57eff7d26481d9b23d6a Generated by: https://github.com/swagger-api/swagger-codegen.git """ import pprint import re # noqa: F401 import six class UserRoleResource(object): """NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ """ Attributes: swagger_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ swagger_types = { 'id': 'str', 'name': 'str', 'description': 'str', 'supported_restrictions': 'list[str]', 'space_permission_descriptions': 'list[str]', 'system_permission_descriptions': 'list[str]', 'granted_space_permissions': 'list[str]', 'granted_system_permissions': 'list[str]', 'can_be_deleted': 'bool', 'last_modified_on': 'datetime', 'last_modified_by': 'str', 'links': 'dict(str, str)' } attribute_map = { 'id': 'Id', 'name': 'Name', 'description': 'Description', 'supported_restrictions': 'SupportedRestrictions', 'space_permission_descriptions': 'SpacePermissionDescriptions', 'system_permission_descriptions': 'SystemPermissionDescriptions', 'granted_space_permissions': 'GrantedSpacePermissions', 'granted_system_permissions': 'GrantedSystemPermissions', 'can_be_deleted': 'CanBeDeleted', 'last_modified_on': 'LastModifiedOn', 'last_modified_by': 'LastModifiedBy', 'links': 'Links' } def __init__(self, id=None, name=None, description=None, supported_restrictions=None, space_permission_descriptions=None, system_permission_descriptions=None, granted_space_permissions=None, granted_system_permissions=None, can_be_deleted=None, last_modified_on=None, last_modified_by=None, links=None): # noqa: E501 """UserRoleResource - a model defined in Swagger""" # noqa: E501 self._id = None self._name = None self._description = None self._supported_restrictions = None self._space_permission_descriptions = None self._system_permission_descriptions = None self._granted_space_permissions = None self._granted_system_permissions = None self._can_be_deleted = None self._last_modified_on = None self._last_modified_by = None self._links = None self.discriminator = None if id is not None: self.id = id if name is not None: self.name = name if description is not None: self.description = description if supported_restrictions is not None: self.supported_restrictions = supported_restrictions if space_permission_descriptions is not None: self.space_permission_descriptions = space_permission_descriptions if system_permission_descriptions is not None: self.system_permission_descriptions = system_permission_descriptions if granted_space_permissions is not None: self.granted_space_permissions = granted_space_permissions if granted_system_permissions is not None: self.granted_system_permissions = granted_system_permissions if can_be_deleted is not None: self.can_be_deleted = can_be_deleted if last_modified_on is not None: self.last_modified_on = last_modified_on if last_modified_by is not None: self.last_modified_by = last_modified_by if links is not None: self.links = links @property def id(self): """Gets the id of this UserRoleResource. # noqa: E501 :return: The id of this UserRoleResource. # noqa: E501 :rtype: str """ return self._id @id.setter def id(self, id): """Sets the id of this UserRoleResource. :param id: The id of this UserRoleResource. # noqa: E501 :type: str """ self._id = id @property def name(self): """Gets the name of this UserRoleResource. # noqa: E501 :return: The name of this UserRoleResource. # noqa: E501 :rtype: str """ return self._name @name.setter def name(self, name): """Sets the name of this UserRoleResource. :param name: The name of this UserRoleResource. # noqa: E501 :type: str """ self._name = name @property def description(self): """Gets the description of this UserRoleResource. # noqa: E501 :return: The description of this UserRoleResource. # noqa: E501 :rtype: str """ return self._description @description.setter def description(self, description): """Sets the description of this UserRoleResource. :param description: The description of this UserRoleResource. # noqa: E501 :type: str """ self._description = description @property def supported_restrictions(self): """Gets the supported_restrictions of this UserRoleResource. # noqa: E501 :return: The supported_restrictions of this UserRoleResource. # noqa: E501 :rtype: list[str] """ return self._supported_restrictions @supported_restrictions.setter def supported_restrictions(self, supported_restrictions): """Sets the supported_restrictions of this UserRoleResource. :param supported_restrictions: The supported_restrictions of this UserRoleResource. # noqa: E501 :type: list[str] """ self._supported_restrictions = supported_restrictions @property def space_permission_descriptions(self): """Gets the space_permission_descriptions of this UserRoleResource. # noqa: E501 :return: The space_permission_descriptions of this UserRoleResource. # noqa: E501 :rtype: list[str] """ return self._space_permission_descriptions @space_permission_descriptions.setter def space_permission_descriptions(self, space_permission_descriptions): """Sets the space_permission_descriptions of this UserRoleResource. :param space_permission_descriptions: The space_permission_descriptions of this UserRoleResource. # noqa: E501 :type: list[str] """ self._space_permission_descriptions = space_permission_descriptions @property def system_permission_descriptions(self): """Gets the system_permission_descriptions of this UserRoleResource. # noqa: E501 :return: The system_permission_descriptions of this UserRoleResource. # noqa: E501 :rtype: list[str] """ return self._system_permission_descriptions @system_permission_descriptions.setter def system_permission_descriptions(self, system_permission_descriptions): """Sets the system_permission_descriptions of this UserRoleResource. :param system_permission_descriptions: The system_permission_descriptions of this UserRoleResource. # noqa: E501 :type: list[str] """ self._system_permission_descriptions = system_permission_descriptions @property def granted_space_permissions(self): """Gets the granted_space_permissions of this UserRoleResource. # noqa: E501 :return: The granted_space_permissions of this UserRoleResource. # noqa: E501 :rtype: list[str] """ return self._granted_space_permissions @granted_space_permissions.setter def granted_space_permissions(self, granted_space_permissions): """Sets the granted_space_permissions of this UserRoleResource. :param granted_space_permissions: The granted_space_permissions of this UserRoleResource. # noqa: E501 :type: list[str] """ allowed_values = ["None", "AdministerSystem", "ProjectEdit", "ProjectView", "ProjectCreate", "ProjectDelete", "ProcessView", "ProcessEdit", "VariableEdit", "VariableEditUnscoped", "VariableView", "VariableViewUnscoped", "ReleaseCreate", "ReleaseView", "ReleaseEdit", "ReleaseDelete", "DefectReport", "DefectResolve", "DeploymentCreate", "DeploymentDelete", "DeploymentView", "EnvironmentView", "EnvironmentCreate", "EnvironmentEdit", "EnvironmentDelete", "MachineCreate", "MachineEdit", "MachineView", "MachineDelete", "ArtifactView", "ArtifactCreate", "ArtifactEdit", "ArtifactDelete", "FeedView", "EventView", "LibraryVariableSetView", "LibraryVariableSetCreate", "LibraryVariableSetEdit", "LibraryVariableSetDelete", "ProjectGroupView", "ProjectGroupCreate", "ProjectGroupEdit", "ProjectGroupDelete", "TeamCreate", "TeamView", "TeamEdit", "TeamDelete", "UserView", "UserInvite", "UserRoleView", "UserRoleEdit", "TaskView", "TaskCreate", "TaskCancel", "TaskEdit", "InterruptionView", "InterruptionSubmit", "InterruptionViewSubmitResponsible", "BuiltInFeedPush", "BuiltInFeedAdminister", "BuiltInFeedDownload", "ActionTemplateView", "ActionTemplateCreate", "ActionTemplateEdit", "ActionTemplateDelete", "LifecycleCreate", "LifecycleView", "LifecycleEdit", "LifecycleDelete", "AccountView", "AccountEdit", "AccountCreate", "AccountDelete", "TenantCreate", "TenantEdit", "TenantView", "TenantDelete", "TagSetCreate", "TagSetEdit", "TagSetDelete", "MachinePolicyCreate", "MachinePolicyView", "MachinePolicyEdit", "MachinePolicyDelete", "ProxyCreate", "ProxyView", "ProxyEdit", "ProxyDelete", "SubscriptionCreate", "SubscriptionView", "SubscriptionEdit", "SubscriptionDelete", "TriggerCreate", "TriggerView", "TriggerEdit", "TriggerDelete", "CertificateView", "CertificateCreate", "CertificateEdit", "CertificateDelete", "CertificateExportPrivateKey", "UserEdit", "ConfigureServer", "FeedEdit", "WorkerView", "WorkerEdit", "RunSystem", "SpaceEdit", "SpaceView", "SpaceDelete", "SpaceCreate", "PackageMetadataPush"] # noqa: E501 if not set(granted_space_permissions).issubset(set(allowed_values)): raise ValueError( "Invalid values for `granted_space_permissions` [{0}], must be a subset of [{1}]" # noqa: E501 .format(", ".join(map(str, set(granted_space_permissions) - set(allowed_values))), # noqa: E501 ", ".join(map(str, allowed_values))) ) self._granted_space_permissions = granted_space_permissions @property def granted_system_permissions(self): """Gets the granted_system_permissions of this UserRoleResource. # noqa: E501 :return: The granted_system_permissions of this UserRoleResource. # noqa: E501 :rtype: list[str] """ return self._granted_system_permissions @granted_system_permissions.setter def granted_system_permissions(self, granted_system_permissions): """Sets the granted_system_permissions of this UserRoleResource. :param granted_system_permissions: The granted_system_permissions of this UserRoleResource. # noqa: E501 :type: list[str] """ allowed_values = ["None", "AdministerSystem", "ProjectEdit", "ProjectView", "ProjectCreate", "ProjectDelete", "ProcessView", "ProcessEdit", "VariableEdit", "VariableEditUnscoped", "VariableView", "VariableViewUnscoped", "ReleaseCreate", "ReleaseView", "ReleaseEdit", "ReleaseDelete", "DefectReport", "DefectResolve", "DeploymentCreate", "DeploymentDelete", "DeploymentView", "EnvironmentView", "EnvironmentCreate", "EnvironmentEdit", "EnvironmentDelete", "MachineCreate", "MachineEdit", "MachineView", "MachineDelete", "ArtifactView", "ArtifactCreate", "ArtifactEdit", "ArtifactDelete", "FeedView", "EventView", "LibraryVariableSetView", "LibraryVariableSetCreate", "LibraryVariableSetEdit", "LibraryVariableSetDelete", "ProjectGroupView", "ProjectGroupCreate", "ProjectGroupEdit", "ProjectGroupDelete", "TeamCreate", "TeamView", "TeamEdit", "TeamDelete", "UserView", "UserInvite", "UserRoleView", "UserRoleEdit", "TaskView", "TaskCreate", "TaskCancel", "TaskEdit", "InterruptionView", "InterruptionSubmit", "InterruptionViewSubmitResponsible", "BuiltInFeedPush", "BuiltInFeedAdminister", "BuiltInFeedDownload", "ActionTemplateView", "ActionTemplateCreate", "ActionTemplateEdit", "ActionTemplateDelete", "LifecycleCreate", "LifecycleView", "LifecycleEdit", "LifecycleDelete", "AccountView", "AccountEdit", "AccountCreate", "AccountDelete", "TenantCreate", "TenantEdit", "TenantView", "TenantDelete", "TagSetCreate", "TagSetEdit", "TagSetDelete", "MachinePolicyCreate", "MachinePolicyView", "MachinePolicyEdit", "MachinePolicyDelete", "ProxyCreate", "ProxyView", "ProxyEdit", "ProxyDelete", "SubscriptionCreate", "SubscriptionView", "SubscriptionEdit", "SubscriptionDelete", "TriggerCreate", "TriggerView", "TriggerEdit", "TriggerDelete", "CertificateView", "CertificateCreate", "CertificateEdit", "CertificateDelete", "CertificateExportPrivateKey", "UserEdit", "ConfigureServer", "FeedEdit", "WorkerView", "WorkerEdit", "RunSystem", "SpaceEdit", "SpaceView", "SpaceDelete", "SpaceCreate", "PackageMetadataPush"] # noqa: E501 if not set(granted_system_permissions).issubset(set(allowed_values)): raise ValueError( "Invalid values for `granted_system_permissions` [{0}], must be a subset of [{1}]" # noqa: E501 .format(", ".join(map(str, set(granted_system_permissions) - set(allowed_values))), # noqa: E501 ", ".join(map(str, allowed_values))) ) self._granted_system_permissions = granted_system_permissions @property def can_be_deleted(self): """Gets the can_be_deleted of this UserRoleResource. # noqa: E501 :return: The can_be_deleted of this UserRoleResource. # noqa: E501 :rtype: bool """ return self._can_be_deleted @can_be_deleted.setter def can_be_deleted(self, can_be_deleted): """Sets the can_be_deleted of this UserRoleResource. :param can_be_deleted: The can_be_deleted of this UserRoleResource. # noqa: E501 :type: bool """ self._can_be_deleted = can_be_deleted @property def last_modified_on(self): """Gets the last_modified_on of this UserRoleResource. # noqa: E501 :return: The last_modified_on of this UserRoleResource. # noqa: E501 :rtype: datetime """ return self._last_modified_on @last_modified_on.setter def last_modified_on(self, last_modified_on): """Sets the last_modified_on of this UserRoleResource. :param last_modified_on: The last_modified_on of this UserRoleResource. # noqa: E501 :type: datetime """ self._last_modified_on = last_modified_on @property def last_modified_by(self): """Gets the last_modified_by of this UserRoleResource. # noqa: E501 :return: The last_modified_by of this UserRoleResource. # noqa: E501 :rtype: str """ return self._last_modified_by @last_modified_by.setter def last_modified_by(self, last_modified_by): """Sets the last_modified_by of this UserRoleResource. :param last_modified_by: The last_modified_by of this UserRoleResource. # noqa: E501 :type: str """ self._last_modified_by = last_modified_by @property def links(self): """Gets the links of this UserRoleResource. # noqa: E501 :return: The links of this UserRoleResource. # noqa: E501 :rtype: dict(str, str) """ return self._links @links.setter def links(self, links): """Sets the links of this UserRoleResource. :param links: The links of this UserRoleResource. # noqa: E501 :type: dict(str, str) """ self._links = links def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value if issubclass(UserRoleResource, dict): for key, value in self.items(): result[key] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, UserRoleResource): return False return self.__dict__ == other.__dict__ def __ne__(self, other): """Returns true if both objects are not equal""" return not self == other
42.555288
2,033
0.673106
4e949a6f4b8f9d86c879098cae8dde8d91b75f85
10,163
py
Python
helpers.py
jchanke/mixtape50
68d03034b503fd0374b9fcba1c1d5207ed7f0170
[ "MIT" ]
1
2022-03-15T11:49:54.000Z
2022-03-15T11:49:54.000Z
helpers.py
jchanke/mixtape50
68d03034b503fd0374b9fcba1c1d5207ed7f0170
[ "MIT" ]
null
null
null
helpers.py
jchanke/mixtape50
68d03034b503fd0374b9fcba1c1d5207ed7f0170
[ "MIT" ]
null
null
null
""" Does the legwork of searching for matching tracks. Contains: (1) Search functions: - search_message - search_spotipy - search_db - search_lookup (2) String parsers (to clean title name): - clean_title - remove_punctuation (3) Creates new Spotify playlist. - create_playlist """ from typing import Any, List, Dict, Union import os import re import sqlite3 import time import spotipy from spotipy.oauth2 import SpotifyOAuth from announcer import MessageAnnouncer, format_sse # Localhost URL to access the application; Flask runs on port 5000 by default # Adapated from https://github.com/Deffro/statify/blob/dd15a6e70428bd36ecddb5d4a8ac3d82b85c9339/code/server.py#L553 CLIENT_SIDE_URL = "http://127.0.0.1" PORT = 5000 # Get environment variables SPOTIPY_CLIENT_ID = os.getenv("SPOTIPY_CLIENT_ID") SPOTIPY_CLIENT_SECRET = os.getenv("SPOTIFY_CLIENT_SECRET") SPOTIPY_REDIRECT_URI = f"{CLIENT_SIDE_URL}:{PORT}/callback" SCOPE = "playlist-modify-public playlist-modify-private playlist-read-private" # Set up Spotipy sp = spotipy.Spotify(auth_manager = SpotifyOAuth(client_id = SPOTIPY_CLIENT_ID, client_secret = SPOTIPY_CLIENT_SECRET, redirect_uri = SPOTIPY_REDIRECT_URI, scope = SCOPE, )) # Create ('instantiate') a MessageAnnouncer object announcer = MessageAnnouncer() """ (1) Search functions: - search_message - search_spotipy - search_db - search_lookup """ def search_message(message: str, max_search_length: int = 10, query_lookup: Dict[str, list] = dict(), failed_queries: set = set()) -> List[Union[list, Any]]: """ search_message(message, max_search_length = 10) Returns a list of song names (change to ids) matching the message. Uses regex-style greedy search. Song names will be limited to [max_search_length] words (default is 10, can be adjusted.) Returns songs from Spotify API via spotipy library; if not, checks Spotify 1.2M songs dataset via an sqlite3 query. Memoizes successful queries (to query_lookup) and failured queries (to failed_queries). https://www.kaggle.com/rodolfofigueroa/spotify-12m-songs """ # Split message into list of lower-case words message = remove_punctuation(message.casefold()).split() # Gets up to max_search_length words of message query_length = min(max_search_length, len(message)) # List containing search functions to iterate over search_functions = [ search_lookup, search_spotipy, search_db, ] # Wait 0.2 seconds to ensure /creating has loaded time.sleep(0.2) # Splits query into prefix and suffix, decrementing prefix, until # - prefix exactly matches a song # - suffix can be expressed as a list of songs for i in range(query_length): prefix, suffix = message[:query_length - i], message[query_length - i:] prefix, suffix = " ".join(prefix), " ".join(suffix) announcer.announce(format_sse(event = "add", data = prefix)) # Only search if suffix is not known to fail if suffix in failed_queries: time.sleep(0.1) announcer.announce(format_sse(event = "drop", data = prefix)) continue # back to the start of the 'for' loop # Looping through search functions, for search_function in search_functions: # Search for tracks matching prefix prefix_results = search_function(prefix, query_lookup = query_lookup) if prefix_results: query_lookup[prefix] = prefix_results print(f"Try: {prefix} in {search_function.__name__.replace('search_', '')}") # In announcer: replace prefix, add each track in prefix_results announcer.announce(format_sse(event = "drop", data = prefix)) for track in map(lambda tracks: tracks[0]["name"], prefix_results): announcer.announce(format_sse(event = "add", data = remove_punctuation(clean_title(track.casefold())))) time.sleep(0.1) # Base case: if prefix is whole message, suffix == "", so we should just return prefix if suffix == "": print(f"All done!") announcer.announce(format_sse(event = "lock in")) return prefix_results # Recursive case: make sure suffix it can be split into songs as well suffix_results = search_message(suffix, max_search_length = max_search_length, query_lookup = query_lookup, failed_queries = failed_queries) # If both are valid, return joined list if suffix_results: results = prefix_results + suffix_results query_lookup[" ".join([prefix, suffix])] = results return results # Suffix cannot be split into songs, drop prefix for track in map(lambda tracks: tracks[0]["name"], prefix_results): announcer.announce(format_sse(event = "drop", data = remove_punctuation(clean_title(track.casefold())))) time.sleep(0.1) print(f"\"{suffix}\" suffix can't be split.") break # suffix doesn't work, try next prefix-suffix pair # Prefix not found in all search functions, drop it else: print(f"\"{prefix}\" doesn't work, moving on.") announcer.announce(format_sse(data = "prefix doesn't work, dropping it")) announcer.announce(format_sse(event = "drop", data = prefix)) # Recursive case: failure failed_queries.add(" ".join(message)) return [] def search_lookup(query: str, query_lookup: Dict[str, list]) -> list: """ Checks query_lookup (a dictionary created at the initial function call of search_message) and returns the results of the query if it has already been found. """ # Checks query_lookup dict if query in query_lookup: return query_lookup[query] else: return [] def search_spotipy(query: str, query_lookup: Dict[str, list]) -> list: """ Uses Spotify API via spotipy library to return a list of songs (name & id) which match the query. Note: the query_lookup parameter is not used. It is only included in the definition because query_lookup is passed to search_functions. """ # Attributes to return attributes = ["name", "id"] # Search for tracks where the name matches query results = sp.search(q=f"track:\"{query}\"", type="track", limit=50) results = results["tracks"]["items"] results = [{ attr: item[attr] for attr in attributes } for item in results if remove_punctuation(clean_title(item["name"].casefold())) == remove_punctuation(query)] # If no results, return empty list: if results == []: return [] else: return [results] def search_db(query: str, query_lookup: Dict[str, list]) -> list: """ Searches tracks.db (1.2 million songs from Spotify from the Kaggle database) to return a list of songs (name & id) which match the query. https://www.kaggle.com/rodolfofigueroa/spotify-12m-songs """ # Import sqlite database tracks = sqlite3.connect("tracks.db") db = sqlite3.Cursor(tracks) # SQLite3 query results = db.execute("SELECT name, id FROM tracks WHERE name_cleaned = ?", [remove_punctuation(query)]).fetchall() results = list(map(lambda item: { "name": item[0], "id": item[1], }, results)) # If no results, return empty list if results == []: return [] else: return [results] """ (2) String parsers (to clean title name): - clean_title - remove_punctuation """ def clean_title(title): """ Cleans title by performing the following transformations in order: - Remove substrings enclosed in (...) or [...] and preceding whitespace (using regex greedy matching) - Remove " - " and substring after - Remove " feat.", " ft(.)", or " featuring" and substring after https://stackoverflow.com/questions/14596884/remove-text-between-and """ # (Greedy) replace substrings between (...) and [] title = re.sub(r"\s+\(.+\)", "", title) title = re.sub(r"\s+\[.+\]", "", title) # Remove " - " and subsequent substring title = re.sub(r" - .*", "", title) # Remove " feat(.) ", " ft(.) ", or " featuring " (but not "feature") and substring after title = re.sub(r"\W+(ft[:.]?|feat[:.]|featuring)\s.*", "", title) return title def remove_punctuation(title): """ Removes punctuation by performing the following transformations: - Delete XML escape sequences: &amp; &quot; &lt; &gt; &apos; - Replace "/", "//", etc. and surrounding whitespace with " " (in medley titles) - Replace "&" and surrounding whitespace with " and " - Remove the following characters from the string: !"#$%'‘’“”()*+,-.:;<=>?@[\]^_—`{|}~ - Strips surrounding whitespace """ title = re.sub(r"&[amp|quot|lt|gt|apos];", "", title) title = re.sub(r"\s*\/+\s*", " ", title) title = re.sub(r"\s*&\s*", " and ", title) title = re.sub(r"[!\"#$%'‘’“”()*+,-.:;<=>?@[\\\]^_—`{|}~]", "", title) title = re.sub(r"\s{2,}", " ", title) return title.strip() """ (3) Creates new Spotify playlist. """ def create_playlist(results): """ Takes the result of search_message as input. Constructs a playlist (via the spotipy library). Returns the Spotify id of the playlist. """ # Process items items = list(map(lambda songs: songs[0]["id"], results)) # Create playlist playlist = sp.user_playlist_create( user=sp.me()["id"], name="mixtape50", public=False, collaborative=False, description="Created with Mixtape50: https://github.com/jchanke/mixtape50." ) sp.playlist_add_items(playlist_id=playlist["id"], items=items) return playlist["id"]
34.686007
168
0.630326
4e96757c37df00a4561207275579e02e7d774aeb
3,836
py
Python
molly/routing/providers/cyclestreets.py
mollyproject/mollyproject
3247c6bac3f39ce8d275d19aa410b30c6284b8a7
[ "Apache-2.0" ]
7
2015-05-16T13:27:21.000Z
2019-08-06T11:09:24.000Z
molly/routing/providers/cyclestreets.py
mollyproject/mollyproject
3247c6bac3f39ce8d275d19aa410b30c6284b8a7
[ "Apache-2.0" ]
null
null
null
molly/routing/providers/cyclestreets.py
mollyproject/mollyproject
3247c6bac3f39ce8d275d19aa410b30c6284b8a7
[ "Apache-2.0" ]
4
2015-11-27T13:36:36.000Z
2021-03-09T17:55:53.000Z
from urllib import urlencode from urllib2 import urlopen import simplejson from django.conf import settings from django.contrib.gis.geos import Point, LineString from django.utils.text import capfirst from django.utils.translation import ugettext as _ from molly.apps.places.models import bearing_to_compass from molly.utils.templatetags.molly_utils import humanise_distance, humanise_seconds CYCLESTREETS_URL = 'http://www.cyclestreets.net/api/journey.json?%s' if 'cyclestreets' not in settings.API_KEYS: # Cyclestreets not configured raise ImportError() def generate_route(points, type): """ Given 2 Points, this will return a route between them. The route consists of a dictionary with the following keys: * error (optional, and if set means that the object contains no route), which is a string describing any errors that occurred in plotting the route * total_time: An int of the number of seconds this route is estimated to take * total_distance: An int of the number of metres this route is expected to take * waypoints: A list of dictionaries, where each dictionary has 2 keys: 'instruction', which is a human-readable description of the steps to be taken here, and 'location', which is a Point describing the route to be taken @param points: An ordered list of points to be included in this route @type points: [Point] @param type: The type of route to generate (foot, car or bike) @type type: str @return: A dictionary containing the route and metadata associated with it @rtype: dict """ # Build Cyclestreets request: url = CYCLESTREETS_URL % urlencode({ 'key': settings.API_KEYS['cyclestreets'], 'plan': 'balanced', 'itinerarypoints': '|'.join('%f,%f' % (p[0], p[1]) for p in points) }) json = simplejson.load(urlopen(url)) if not json: return { 'error': _('Unable to plot route') } else: summary = json['marker'][0]['@attributes'] waypoints = [] for i, waypoint in enumerate(json['marker'][1:]): segment = waypoint['@attributes'] waypoints.append({ 'instruction': _('%(instruction)s at %(name)s') % { 'instruction': capfirst(segment['turn']), 'name': segment['name'] }, 'additional': _('%(direction)s for %(distance)s (taking approximately %(time)s)') % { 'direction': bearing_to_compass(int(segment['startBearing'])), 'distance': humanise_distance(segment['distance'], False), 'time': humanise_seconds(segment['time']) }, 'waypoint_type': { 'straight on': 'straight', 'turn left': 'left', 'bear left': 'slight-left', 'sharp left': 'sharp-left', 'turn right': 'right', 'bear right': 'slight-right', 'sharp right': 'sharp-right', 'double-back': 'turn-around', }.get(segment['turn']), 'location': Point(*map(float, segment['points'].split(' ')[0].split(','))), 'path': LineString(map(lambda ps: Point(*map(float, ps.split(','))), segment['points'].split(' '))) }) return { 'total_time': summary['time'], 'total_distance': summary['length'], 'waypoints': waypoints, 'path': LineString(map(lambda ps: Point(*map(float, ps.split(','))), summary['coordinates'].split(' '))) }
40.808511
116
0.569343
4e99de297b6c41fb361b034e5f59be29d6569791
316
py
Python
exercises/exc_G1.py
dataXcode/IPP
c9b94ad2d7dc14b01e6657a4fa555507bbc7e93b
[ "MIT" ]
null
null
null
exercises/exc_G1.py
dataXcode/IPP
c9b94ad2d7dc14b01e6657a4fa555507bbc7e93b
[ "MIT" ]
null
null
null
exercises/exc_G1.py
dataXcode/IPP
c9b94ad2d7dc14b01e6657a4fa555507bbc7e93b
[ "MIT" ]
null
null
null
num_list = [10,50,30,12,6,8,100] def max_min_first_last(nlist): ____________________ # Maximum number ____________________ # Minimum number ____________________ # First number ____________________ # Last number return _____________________________________ print( max_min_first_last(num_list) )
28.727273
48
0.78481
4e9c35d7a10e21f257f971c50e260fb397455462
5,829
py
Python
web/impact/impact/tests/api_test_case.py
masschallenge/impact-api
81075ced8fcc95de9390dd83c15e523e67fc48c0
[ "MIT" ]
5
2017-10-19T15:11:52.000Z
2020-03-08T07:16:21.000Z
web/impact/impact/tests/api_test_case.py
masschallenge/impact-api
81075ced8fcc95de9390dd83c15e523e67fc48c0
[ "MIT" ]
182
2017-06-21T19:32:13.000Z
2021-03-22T13:38:16.000Z
web/impact/impact/tests/api_test_case.py
masschallenge/impact-api
81075ced8fcc95de9390dd83c15e523e67fc48c0
[ "MIT" ]
1
2018-06-23T11:53:18.000Z
2018-06-23T11:53:18.000Z
# MIT License # Copyright (c) 2017 MassChallenge, Inc. import json from oauth2_provider.models import get_application_model from rest_framework.test import APIClient from test_plus.test import TestCase from django.core import mail from django.conf import settings from django.contrib.auth.models import Group from django.urls import reverse from accelerator_abstract.models.base_clearance import ( CLEARANCE_LEVEL_GLOBAL_MANAGER, CLEARANCE_LEVEL_STAFF ) from impact.tests.factories import ( ClearanceFactory, UserFactory, ) OAuth_App = get_application_model() API_GROUPS = [settings.V0_API_GROUP, settings.V1_API_GROUP] DESCRIPTION_CONTENT = 'DESCRIPTION:Topics: {topics}' LOCATION_CONTENT = 'LOCATION:{location}\\;' LOCATION_INFO = 'LOCATION:{location}\\;{meeting_info}' class APITestCase(TestCase): SOME_SITE_NAME = "somesite.com" _user_count = 0 client_class = APIClient user_factory = UserFactory @classmethod def setUpClass(cls): [Group.objects.get_or_create(name=name) for name in API_GROUPS] @classmethod def tearDownClass(cls): [Group.objects.get(name=name).delete() for name in API_GROUPS] def basic_user(self): user = self.make_user('basic_user{}@test.com'.format(self._user_count), perms=["mc.view_startup"]) self._user_count += 1 for group in Group.objects.filter(name__in=API_GROUPS): user.groups.add(group) user.set_password('password') user.save() return user def staff_user(self, program_family=None, level=CLEARANCE_LEVEL_STAFF): user = self.make_user('basic_user{}@test.com'.format(self._user_count)) self._user_count += 1 kwargs = {"level": level, "user": user} if program_family: kwargs['program_family'] = program_family clearance = ClearanceFactory(**kwargs) return clearance.user def global_operations_manager(self, program_family): user = self.staff_user() ClearanceFactory(user=user, level=CLEARANCE_LEVEL_GLOBAL_MANAGER, program_family=program_family) return user def get_access_token(self, user): app = OAuth_App.objects.create( user=user, name="Test666", client_type=OAuth_App.CLIENT_PUBLIC, authorization_grant_type=OAuth_App.GRANT_PASSWORD, redirect_uris="http://thirdparty.com/exchange/", ) response = self.client.post( self.reverse("oauth2_provider:token"), data={ "password": 'password', "client_id": app.client_id, "username": user.username, "grant_type": "password", }, headers={'Content-Type': 'application/x-www-form-urlencoded'} ) response_json = json.loads(response.content) return response_json['access_token'] def assert_options_include(self, method, expected_options, object_id=None): if object_id: args = [object_id] else: args = [] url = reverse(self.view.view_name, args=args) with self.login(email=self.basic_user().email): response = self.client.options(url) result = json.loads(response.content) assert method in result['actions'] options = result['actions'][method]['properties'] for key, params in expected_options.items(): self.assertIn(key, options) self.assertEqual(options[key], params) def assert_ui_notification(self, response, success, notification): data = response.data detail = notification if notification else "" header = self.success_header if success else self.fail_header self.assertTrue(all([ data['success'] == success, data['header'] == header, data['detail'] == detail ]), msg='Notification data was not as expected') def assert_notified(self, user, message="", subject="", check_alternative=False): '''Assert that the user received a notification. If `message` is specified, assert that the message appears in one of the outgoing emails to this user ''' emails = [email for email in mail.outbox if user.email in email.to] self.assertGreater(len(emails), 0) if message: if check_alternative: self.assertTrue(any([_message_included_in_email_alternative( email, message) for email in emails])) else: self.assertTrue(any([ message in email.body for email in emails])) if subject: self.assertIn(subject, [email.subject for email in emails]) def assert_ics_email_attachments(self, user): '''assert that the ics email attachment exists ''' emails = [email for email in mail.outbox if user.email in email.to] for email in emails: attachments = email.attachments self.assertGreater(len(email.attachments), 0) self.assertIn("reminder.ics", [attachment[0] for attachment in attachments]) def assert_not_notified(self, user): '''Assert that the specified user did not receive a notification. ''' if mail.outbox: self.assertNotIn(user.email, [email.to for email in mail.outbox], msg="Found an email sent to user") def _message_included_in_email_alternative(email, message): return any([message in alt[0] for alt in email.alternatives])
36.892405
79
0.621891
4e9d488202a407ec4de3fade6bfb2e435ba6bb6b
607
py
Python
pydis_site/apps/api/models/bot/aoc_link.py
Robin5605/site
81aa42aa748cb228d7a09e6cf6b211484b654496
[ "MIT" ]
13
2018-02-03T22:57:41.000Z
2018-05-17T07:38:36.000Z
pydis_site/apps/api/models/bot/aoc_link.py
Robin5605/site
81aa42aa748cb228d7a09e6cf6b211484b654496
[ "MIT" ]
61
2018-02-07T21:34:39.000Z
2018-06-05T16:15:28.000Z
pydis_site/apps/api/models/bot/aoc_link.py
Robin5605/site
81aa42aa748cb228d7a09e6cf6b211484b654496
[ "MIT" ]
16
2018-02-03T12:37:48.000Z
2018-06-02T17:14:55.000Z
from django.db import models from pydis_site.apps.api.models.bot.user import User from pydis_site.apps.api.models.mixins import ModelReprMixin class AocAccountLink(ModelReprMixin, models.Model): """An AoC account link for a Discord User.""" user = models.OneToOneField( User, on_delete=models.CASCADE, help_text="The user that is blocked from getting the AoC Completionist Role", primary_key=True ) aoc_username = models.CharField( max_length=120, help_text="The AoC username associated with the Discord User.", blank=False )
27.590909
85
0.698517
4e9e6122ed3109b35f3efe158b363d95df381cc6
10,892
py
Python
server/tree_pickler.py
michaelpeterswa/CPSC322Project-WildfireAnalysis
872727e8c59619fcfc11aaa70367762271207dbd
[ "MIT" ]
null
null
null
server/tree_pickler.py
michaelpeterswa/CPSC322Project-WildfireAnalysis
872727e8c59619fcfc11aaa70367762271207dbd
[ "MIT" ]
null
null
null
server/tree_pickler.py
michaelpeterswa/CPSC322Project-WildfireAnalysis
872727e8c59619fcfc11aaa70367762271207dbd
[ "MIT" ]
1
2021-04-16T21:21:25.000Z
2021-04-16T21:21:25.000Z
import pickle best_trees = [ {'accuracy': 0.36416184971098264, 'tree': ['Attribute', 'att1', ['Value', 'Pend Oreille', ['Leaf', 2.0, 0, 69] ], ['Value', 'Okanogan', ['Leaf', 3.0, 0, 314] ], ['Value', 'Lincoln', ['Leaf', 5.0, 0, 55] ], ['Value', 'Grant', ['Leaf', 5.0, 0, 4] ], ['Value', 'Chelan', ['Leaf', 3.0, 0, 136]], ['Value', 'Stevens', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 18]], ['Value', 'Miscellaneou', ['Leaf', 2.0, 0, 83]], ['Value', 'Lightning', ['Leaf', 2.0, 0, 43]], ['Value', 'Under Invest', ['Leaf', 5.0, 0, 6]], ['Value', 'Debris Burn', ['Leaf', 3.0, 0, 120]], ['Value', 'Children', ['Leaf', 3.0, 0, 8]], ['Value', 'None', ['Leaf', 5.0, 1, 308]], ['Value', 'Smoker', ['Leaf', 2.0, 0, 7]], ['Value', 'Logging', ['Leaf', 3.0, 0, 8]], ['Value', 'Arson', ['Leaf', 2.0, 0, 5]], ['Value', 'Undetermined', ['Leaf', 9.0, 2, 308]], ['Value', 'Railroad', ['Leaf', 4.0, 0, 7]]]], ['Value', 'Clark', ['Leaf', 3.0, 0, 20]], ['Value', 'Yakima', ['Leaf', 3.0, 0, 97]], ['Value', 'Spokane', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 23]], ['Value', 'Miscellaneou', ['Leaf', 2.0, 0, 142]], ['Value', 'Lightning', ['Leaf', 3.0, 0, 24]], ['Value', 'Under Invest', ['Leaf', 3.0, 0, 4]], ['Value', 'Debris Burn', ['Leaf', 2.0, 0, 54]], ['Value', 'Children', ['Leaf', 3.0, 0, 20]], ['Value', 'None', ['Leaf', 3.0, 3, 326]], ['Value', 'Smoker', ['Leaf', 2.0, 0, 2]], ['Value', 'Logging', ['Leaf', 2.0, 0, 3]], ['Value', 'Arson', ['Leaf', 2.0, 0, 29]], ['Value', 'Undetermined', ['Leaf', 2.0, 0, 7]], ['Value', 'Railroad', ['Leaf', 2.0, 0, 15]]]], ['Value', 'Pierce', ['Leaf', 3.0, 0, 55]], ['Value', 'Skagit', ['Leaf', 3.0, 0, 34]], ['Value', 'Grays Harbor', ['Leaf', 3.0, 0, 52]], ['Value', 'Skamania', ['Leaf', 3.0, 0, 28]], ['Value', 'King', ['Leaf', 3.0, 0, 41]], ['Value', 'Island', ['Leaf', 3.0, 0, 7]], ['Value', 'Klickitat', ['Leaf', 3.0, 0, 180]], ['Value', 'Whitman', ['Leaf', 7.0, 0, 5]], ['Value', 'Cowlitz', ['Leaf', 3.0, 0, 68]], ['Value', 'Douglas', ['Leaf', 5.0, 0, 27]], ['Value', 'Ferry', ['Leaf', 3.0, 0, 72]], ['Value', 'Mason', ['Leaf', 3.0, 0, 66]], ['Value', 'Kittitas', ['Leaf', 3.0, 0, 99]], ['Value', 'Jefferson', ['Leaf', 3.0, 0, 30]], ['Value', 'Franklin', ['Leaf', 5.0, 3, 2503]], ['Value', 'Clallam', ['Leaf', 3.0, 0, 44]], ['Value', 'Pacific', ['Leaf', 3.0, 0, 51]], ['Value', 'Lewis', ['Leaf', 3.0, 0, 93]], ['Value', 'Thurston', ['Leaf', 2.0, 0, 59]], ['Value', 'Walla Walla', ['Leaf', 3.0, 0, 18]], ['Value', 'Snohomish', ['Leaf', 3.0, 0, 38]], ['Value', 'Asotin', ['Leaf', 4.0, 0, 23]], ['Value', 'Adams', ['Leaf', 5.0, 1, 2503]], ['Value', 'Whatcom', ['Leaf', 2.0, 0, 40]], ['Value', 'San Juan', ['Leaf', 3.0, 0, 7]], ['Value', 'Garfield', ['Leaf', 3.0, 0, 10]], ['Value', 'Columbia', ['Leaf', 3.0, 0, 14]], ['Value', 'Benton', ['Leaf', 7.0, 1, 2503]], ['Value', 'Wahkiakum', ['Leaf', 3.0, 5, 2503]], ['Value', 'No Data', ['Leaf', 4.0, 1, 2503]], ['Value', 'Kitsap', ['Leaf', 3.0, 0, 2]]]}, {'accuracy': 0.34375, 'tree': ['Attribute', 'att1', ['Value', 'Klickitat', ['Leaf', 2.0, 0, 150]], ['Value', 'Ferry', ['Leaf', 3.0, 0, 66]], ['Value', 'Okanogan', ['Leaf', 3.0, 0, 341]], ['Value', 'Clallam', ['Leaf', 3.0, 0, 53]], ['Value', 'Lewis', ['Leaf', 3.0, 0, 105]], ['Value', 'Kittitas', ['Leaf', 3.0, 0, 115]], ['Value', 'Spokane', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 31]], ['Value', 'Arson', ['Leaf', 2.0, 0, 37]], ['Value', 'Lightning', ['Leaf', 3.0, 0, 25]], ['Value', 'Miscellaneou', ['Leaf', 3.0, 0, 122]], ['Value', 'Logging', ['Leaf', 3.0, 1, 318]], ['Value', 'Under Invest', ['Leaf', 5.0, 4, 318]], ['Value', 'Debris Burn', ['Leaf', 3.0, 0, 51]], ['Value', 'Railroad', ['Leaf', 2.0, 0, 25]], ['Value', 'Children', ['Leaf', 4.0, 0, 12]], ['Value', 'Undetermined', ['Leaf', 5.0, 0, 5]], ['Value', 'Smoker', ['Leaf', 6.0, 0, 4]], ['Value', 'None', ['Leaf', 3.0, 1, 318]]]], ['Value', 'Chelan', ['Leaf', 3.0, 0, 142]], ['Value', 'Mason', ['Leaf', 3.0, 0, 69]], ['Value', 'Lincoln', ['Leaf', 3.0, 0, 79]], ['Value', 'Yakima', ['Leaf', 3.0, 0, 82]], ['Value', 'Jefferson', ['Leaf', 3.0, 0, 32]], ['Value', 'Pend Oreille', ['Leaf', 2.0, 0, 61]], ['Value', 'Stevens', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 15]], ['Value', 'Arson', ['Leaf', 2.0, 0, 11]], ['Value', 'Lightning', ['Leaf', 3.0, 0, 33]], ['Value', 'Miscellaneou', ['Leaf', 3.0, 0, 84]], ['Value', 'Logging', ['Leaf', 3.0, 4, 290]], ['Value', 'Under Invest', ['Leaf', 5.0, 0, 4]], ['Value', 'Debris Burn', ['Leaf', 2.0, 0, 117]], ['Value', 'Railroad', ['Leaf', 2.0, 0, 6]], ['Value', 'Children', ['Leaf', 2.0, 0, 4]], ['Value', 'Undetermined', ['Leaf', 9.0, 1, 290]], ['Value', 'Smoker', ['Leaf', 2.0, 0, 10]], ['Value', 'None', ['Leaf', 5.0, 1, 290]]]], ['Value', 'Cowlitz', ['Leaf', 3.0, 0, 77]], ['Value', 'Pierce', ['Leaf', 3.0, 0, 58]], ['Value', 'King', ['Leaf', 2.0, 0, 23]], ['Value', 'Walla Walla', ['Leaf', 3.0, 0, 24]], ['Value', 'Douglas', ['Leaf', 6.0, 0, 17]], ['Value', 'Island', ['Leaf', 3.0, 0, 9]], ['Value', 'Skamania', ['Leaf', 3.0, 0, 27]], ['Value', 'Thurston', ['Leaf', 2.0, 0, 52]], ['Value', 'Columbia', ['Leaf', 3.0, 0, 15]], ['Value', 'Snohomish', ['Leaf', 3.0, 0, 36]], ['Value', 'Skagit', ['Leaf', 3.0, 0, 47]], ['Value', 'Pacific', ['Leaf', 3.0, 0, 36]], ['Value', 'Grays Harbor', ['Leaf', 2.0, 0, 56]], ['Value', 'Whatcom', ['Leaf', 3.0, 0, 37]], ['Value', 'Clark', ['Leaf', 3.0, 0, 30]], ['Value', 'Kitsap', ['Leaf', 3.0, 2, 2503]], ['Value', 'San Juan', ['Leaf', 3.0, 0, 9]], ['Value', 'Asotin', ['Leaf', 4.0, 0, 20]], ['Value', 'Garfield', ['Leaf', 3.0, 0, 7]], ['Value', 'Adams', ['Leaf', 5.0, 2, 2503]], ['Value', 'Wahkiakum', ['Leaf', 2.0, 0, 7]], ['Value', 'Whitman', ['Leaf', 5.0, 0, 5]], ['Value', 'Grant', ['Leaf', 5.0, 1, 2503]], ['Value', 'No Data', ['Leaf', 4.0, 0, 2]], ['Value', 'Benton', ['Leaf', 7.0, 1, 2503]]]}, {'accuracy': 0.33568904593639576, 'tree': ['Attribute', 'att1', ['Value', 'Stevens', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 24]], ['Value', 'Debris Burn', ['Leaf', 2.0, 0, 105]], ['Value', 'Children', ['Leaf', 3.0, 0, 4]], ['Value', 'Miscellaneou', ['Leaf', 3.0, 0, 80]], ['Value', 'Railroad', ['Leaf', 2.0, 0, 6]], ['Value', 'Undetermined', ['Leaf', 9.0, 3, 300]], ['Value', 'Logging', ['Leaf', 3.0, 0, 9]], ['Value', 'Lightning', ['Leaf', 2.0, 0, 39]], ['Value', 'Smoker', ['Leaf', 2.0, 0, 8]], ['Value', 'None', ['Leaf', 5.0, 2, 300]], ['Value', 'Arson', ['Leaf', 3.0, 0, 15]], ['Value', 'Under Invest', ['Leaf', 3.0, 0, 5]]]], ['Value', 'Grays Harbor', ['Leaf', 2.0, 0, 49]], ['Value', 'Chelan', ['Leaf', 3.0, 0, 143]], ['Value', 'Okanogan', ['Leaf', 3.0, 0, 306]], ['Value', 'Spokane', ['Attribute', 'att2', ['Value', 'Recreation', ['Leaf', 2.0, 0, 27]], ['Value', 'Debris Burn', ['Leaf', 3.0, 0, 66]], ['Value', 'Children', ['Leaf', 2.0, 0, 10]], ['Value', 'Miscellaneou', ['Leaf', 3.0, 0, 152]], ['Value', 'Railroad', ['Leaf', 2.0, 0, 21]], ['Value', 'Undetermined', ['Leaf', 5.0, 0, 8]], ['Value', 'Logging', ['Leaf', 2.0, 0, 2]], ['Value', 'Lightning', ['Leaf', 3.0, 0, 25]], ['Value', 'Smoker', ['Leaf', 3.0, 0, 3]], ['Value', 'None', ['Leaf', 2.0, 0, 5]], ['Value', 'Arson', ['Leaf', 2.0, 0, 24]], ['Value', 'Under Invest', ['Leaf', 5.0, 2, 345]]]], ['Value', 'Cowlitz', ['Leaf', 3.0, 0, 74]], ['Value', 'Lincoln', ['Leaf', 3.0, 0, 66]], ['Value', 'Kittitas', ['Leaf', 3.0, 0, 122]], ['Value', 'Pacific', ['Leaf', 3.0, 0, 61]], ['Value', 'Skagit', ['Leaf', 3.0, 0, 57]], ['Value', 'Lewis', ['Leaf', 3.0, 0, 111]], ['Value', 'Island', ['Leaf', 3.0, 0, 8]], ['Value', 'Klickitat', ['Leaf', 2.0, 0, 193]], ['Value', 'Walla Walla', ['Leaf', 4.0, 0, 19]], ['Value', 'Jefferson', ['Leaf', 3.0, 0, 23]], ['Value', 'Garfield', ['Leaf', 7.0, 0, 6]], ['Value', 'Thurston', ['Leaf', 2.0, 0, 50]], ['Value', 'King', ['Leaf', 3.0, 0, 33]], ['Value', 'Douglas', ['Leaf', 6.0, 0, 28]], ['Value', 'Yakima', ['Leaf', 3.0, 0, 90]], ['Value', 'Mason', ['Leaf', 3.0, 0, 55]], ['Value', 'Snohomish', ['Leaf', 3.0, 0, 27]], ['Value', 'Pierce', ['Leaf', 3.0, 0, 44]], ['Value', 'Kitsap', ['Leaf', 3.0, 0, 6]], ['Value', 'Clark', ['Leaf', 3.0, 0, 18]], ['Value', 'Columbia', ['Leaf', 3.0, 0, 17]], ['Value', 'Pend Oreille', ['Leaf', 3.0, 0, 45]], ['Value', 'Skamania', ['Leaf', 3.0, 0, 27]], ['Value', 'Asotin', ['Leaf', 7.0, 0, 17]], ['Value', 'Whatcom', ['Leaf', 3.0, 0, 39]], ['Value', 'Ferry', ['Leaf', 3.0, 0, 72]], ['Value', 'Wahkiakum', ['Leaf', 3.0, 1, 2503]], ['Value', 'Clallam', ['Leaf', 3.0, 0, 38]], ['Value', 'Adams', ['Leaf', 5.0, 3, 2503]], ['Value', 'San Juan', ['Leaf', 2.0, 0, 3]], ['Value', 'Grant', ['Leaf', 6.0, 1, 2503]], ['Value', 'No Data', ['Leaf', 4.0, 0, 2]], ['Value', 'Whitman', ['Leaf', 5.0, 0, 4]]]}, {'accuracy': 0.33390705679862304, 'tree': ['Attribute', 'att1', ['Value', 'Spokane', ['Leaf', 3.0, 0, 364]], ['Value', 'Stevens', ['Leaf', 2.0, 0, 298]], ['Value', 'Klickitat', ['Leaf', 3.0, 0, 165]], ['Value', 'Okanogan', ['Leaf', 3.0, 0, 340]], ['Value', 'Yakima', ['Leaf', 5.0, 0, 88]], ['Value', 'Chelan', ['Leaf', 3.0, 0, 110]], ['Value', 'Cowlitz', ['Leaf', 3.0, 0, 84]], ['Value', 'Thurston', ['Leaf', 2.0, 0, 78]], ['Value', 'Pend Oreille', ['Leaf', 2.0, 0, 46]], ['Value', 'Pierce', ['Leaf', 3.0, 0, 45]], ['Value', 'Mason', ['Leaf', 3.0, 0, 69]], ['Value', 'Grays Harbor', ['Leaf', 2.0, 0, 58]], ['Value', 'Douglas', ['Leaf', 6.0, 0, 33]], ['Value', 'Ferry', ['Leaf', 3.0, 0, 77]], ['Value', 'Skagit', ['Leaf', 3.0, 0, 39]], ['Value', 'Clark', ['Leaf', 2.0, 0, 28]], ['Value', 'Kittitas', ['Leaf', 3.0, 0, 108]], ['Value', 'Lewis', ['Leaf', 3.0, 0, 106]], ['Value', 'Skamania', ['Leaf', 3.0, 0, 25]], ['Value', 'King', ['Leaf', 3.0, 0, 23]], ['Value', 'Asotin', ['Leaf', 3.0, 0, 24]], ['Value', 'Snohomish', ['Leaf', 3.0, 0, 26]], ['Value', 'Pacific', ['Leaf', 2.0, 0, 36]], ['Value', 'Jefferson', ['Leaf', 3.0, 0, 29]], ['Value', 'Clallam', ['Leaf', 3.0, 0, 44]], ['Value', 'Lincoln', ['Leaf', 3.0, 0, 56]], ['Value', 'Walla Walla', ['Leaf', 3.0, 0, 18]], ['Value', 'Island', ['Leaf', 3.0, 6, 2503]], ['Value', 'Whatcom', ['Leaf', 3.0, 0, 26]], ['Value', 'Benton', ['Leaf', 7.0, 1, 2503]], ['Value', 'Kitsap', ['Leaf', 3.0, 0, 8]], ['Value', 'San Juan', ['Leaf', 2.0, 0, 14]], ['Value', 'Columbia', ['Leaf', 3.0, 0, 16]], ['Value', 'Franklin', ['Leaf', 5.0, 1, 2503]], ['Value', 'Grant', ['Leaf', 5.0, 4, 2503]], ['Value', 'Garfield', ['Leaf', 3.0, 0, 5]], ['Value', 'Whitman', ['Leaf', 7.0, 2, 2503]], ['Value', 'Wahkiakum', ['Leaf', 2.0, 1, 2503]], ['Value', 'No Data', ['Leaf', 3.0, 1, 2503]], ['Value', 'Adams', ['Leaf', 5.0, 1, 2503]]]}] packaged_object = best_trees # pickle packaged_object outfile = open("trees.p", "wb") pickle.dump(packaged_object, outfile) outfile.close()
495.090909
10,287
0.484209
4e9e7a69ae46de63cdefe46d785a1f6e94dac1e1
624
py
Python
parse.py
Mimori256/kdb-parse
45f7aca85fea9a7db612da86e9c31daaec52a580
[ "MIT" ]
3
2021-06-20T04:35:05.000Z
2021-10-05T06:30:09.000Z
parse.py
Mimori256/kdb-parse
45f7aca85fea9a7db612da86e9c31daaec52a580
[ "MIT" ]
2
2021-06-13T01:19:12.000Z
2022-03-23T04:27:05.000Z
parse.py
Mimori256/kdb-parse
45f7aca85fea9a7db612da86e9c31daaec52a580
[ "MIT" ]
null
null
null
import json #Parse csv to kdb.json with open("kdb.csv", "r", encoding="utf_8") as f: l=[] lines = f.readlines() # remove the header lines.pop(0) for line in lines: tmp1 = line.split('"') if tmp1[15] == "": tmp1[15] = " " if not "" in set([tmp1[1], tmp1[3], tmp1[11], tmp1[13], tmp1[15], tmp1[21]]): l.append([tmp1[1], tmp1[3], tmp1[11], tmp1[13], tmp1[15], tmp1[21]]) json_data = {} l.pop(0) for i in l: json_data[i[0]] = i[1:] enc = json.dumps(json_data,ensure_ascii=False) with open("kdb.json", "w") as f: f.write(enc) print("complete")
20.8
85
0.540064
4e9e92e9363d4d32c2609f2f36539abe9b27e294
2,600
py
Python
DLFrameWork/dataset/CIFAR_10.py
Mostafa-ashraf19/TourchPIP
a5090a0ec9cc81a91fe1fd6af41d77841361cec1
[ "MIT" ]
null
null
null
DLFrameWork/dataset/CIFAR_10.py
Mostafa-ashraf19/TourchPIP
a5090a0ec9cc81a91fe1fd6af41d77841361cec1
[ "MIT" ]
null
null
null
DLFrameWork/dataset/CIFAR_10.py
Mostafa-ashraf19/TourchPIP
a5090a0ec9cc81a91fe1fd6af41d77841361cec1
[ "MIT" ]
null
null
null
import os import shutil import tarfile import urllib.request import pandas as pd CIFAR10_URL = 'https://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz' class CIFAR_10: def __init__(self, path, download=True, train=True): self.path = path self.download = download self.train = train self.csv_list = [] if self.download: self._Download() self.path = os.getcwd() + '/' + self.path self.toCSV() self.TrainFile = self.path + '/' + 'cifar-10-batches-py/train_cifar.csv' self.TestFile = self.path + '/' + 'cifar-10-batches-py/test_batch.csv' def _Download(self): if not os.path.exists(os.getcwd() + '/' + self.path): os.mkdir(self.path) file_name = 'CIFAR-10.tar.gz' with urllib.request.urlopen(CIFAR10_URL) as response, open(os.getcwd() + '/' + self.path + '/' + file_name, 'wb') as out_file: shutil.copyfileobj(response, out_file) tar = tarfile.open(os.getcwd() + '/' + self.path + '/' + file_name, "r:gz") tar.extractall(os.getcwd() + '/' + self.path + '/') tar.close() def unpickle(self, file): import pickle with open(file, 'rb') as fo: dict = pickle.load(fo, encoding='bytes') return dict def toCSV(self): file_names = ['data_batch_1', 'data_batch_2', 'data_batch_3', 'data_batch_4', 'data_batch_5', 'test_batch'] for name in file_names: df_labels = pd.DataFrame(self.unpickle(self.path + '/' + 'cifar-10-batches-py/' + name)[b'labels']) df_data = pd.DataFrame(self.unpickle(self.path + '/' + 'cifar-10-batches-py/' + name)[b'data']) new = pd.concat([df_labels, df_data], axis=1) if not os.path.exists(self.path + '/' + 'cifar-10-batches-py/' + name + '.csv'): new.to_csv(self.path + '/' + 'cifar-10-batches-py/' + name + '.csv', index=False) for name in file_names[0:5]: self.csv_list.append(self.path + '/' + 'cifar-10-batches-py/' + name + '.csv') if not os.path.exists(self.path + '/' + 'cifar-10-batches-py/train_cifar.csv'): df_from_each_file = (pd.read_csv(f, sep=',', header=None) for f in self.csv_list) df_merged = pd.concat(df_from_each_file, ignore_index=True) df_merged.to_csv(self.path + '/' + 'cifar-10-batches-py/train_cifar.csv', index=False) def __repr__(self): return self.TrainFile if self.train == True else self.TestFile
40.625
115
0.575385
4ea10581f2a6479a2145424512ce3b01dbcd78d5
367
py
Python
Python_Codes_for_BJ/stage12 큐 사용하기/프린터 큐.py
ch96an/BaekJoonSolution
25594fda5ba1c0c4d26ff0828ec8dcf2f6572d33
[ "MIT" ]
null
null
null
Python_Codes_for_BJ/stage12 큐 사용하기/프린터 큐.py
ch96an/BaekJoonSolution
25594fda5ba1c0c4d26ff0828ec8dcf2f6572d33
[ "MIT" ]
null
null
null
Python_Codes_for_BJ/stage12 큐 사용하기/프린터 큐.py
ch96an/BaekJoonSolution
25594fda5ba1c0c4d26ff0828ec8dcf2f6572d33
[ "MIT" ]
null
null
null
def printer(n,k,order): lst = [(order[x],False if x==k else True) for x in range(len(order))] flag, i = True, 0 while flag: if lst[0][0] == max(lst,key=lambda x:x[0])[0]: flag = lst.pop(0)[1] i +=1 else: lst.append(lst.pop(0)) print(i) for _ in range(int(input())): n,k=map(int,input().split()) lst=list(map(int,input().split())) printer(n,k,lst)
24.466667
70
0.59673
4ea295d60036c358d4e3b22a59c7d5d5aba282d3
4,119
py
Python
grpc/clients/python/vegaapiclient/generated/commands/v1/oracles_pb2.py
ConnorChristie/api
7e585d47bad1a5ef95ca932045b0ce70962b029a
[ "MIT" ]
6
2021-05-20T15:30:46.000Z
2022-02-22T12:06:39.000Z
grpc/clients/python/vegaapiclient/generated/commands/v1/oracles_pb2.py
ConnorChristie/api
7e585d47bad1a5ef95ca932045b0ce70962b029a
[ "MIT" ]
29
2021-03-16T11:58:05.000Z
2021-10-05T14:04:45.000Z
grpc/clients/python/vegaapiclient/generated/commands/v1/oracles_pb2.py
ConnorChristie/api
7e585d47bad1a5ef95ca932045b0ce70962b029a
[ "MIT" ]
6
2021-05-07T06:43:02.000Z
2022-03-29T07:18:01.000Z
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: commands/v1/oracles.proto """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() DESCRIPTOR = _descriptor.FileDescriptor( name='commands/v1/oracles.proto', package='vega.commands.v1', syntax='proto3', serialized_options=b'\n io.vegaprotocol.vega.commands.v1Z+code.vegaprotocol.io/vega/proto/commands/v1', create_key=_descriptor._internal_create_key, serialized_pb=b'\n\x19\x63ommands/v1/oracles.proto\x12\x10vega.commands.v1\"\xcb\x01\n\x14OracleDataSubmission\x12K\n\x06source\x18\x01 \x01(\x0e\x32\x33.vega.commands.v1.OracleDataSubmission.OracleSourceR\x06source\x12\x18\n\x07payload\x18\x02 \x01(\x0cR\x07payload\"L\n\x0cOracleSource\x12\x1d\n\x19ORACLE_SOURCE_UNSPECIFIED\x10\x00\x12\x1d\n\x19ORACLE_SOURCE_OPEN_ORACLE\x10\x01\x42O\n io.vegaprotocol.vega.commands.v1Z+code.vegaprotocol.io/vega/proto/commands/v1b\x06proto3' ) _ORACLEDATASUBMISSION_ORACLESOURCE = _descriptor.EnumDescriptor( name='OracleSource', full_name='vega.commands.v1.OracleDataSubmission.OracleSource', filename=None, file=DESCRIPTOR, create_key=_descriptor._internal_create_key, values=[ _descriptor.EnumValueDescriptor( name='ORACLE_SOURCE_UNSPECIFIED', index=0, number=0, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), _descriptor.EnumValueDescriptor( name='ORACLE_SOURCE_OPEN_ORACLE', index=1, number=1, serialized_options=None, type=None, create_key=_descriptor._internal_create_key), ], containing_type=None, serialized_options=None, serialized_start=175, serialized_end=251, ) _sym_db.RegisterEnumDescriptor(_ORACLEDATASUBMISSION_ORACLESOURCE) _ORACLEDATASUBMISSION = _descriptor.Descriptor( name='OracleDataSubmission', full_name='vega.commands.v1.OracleDataSubmission', filename=None, file=DESCRIPTOR, containing_type=None, create_key=_descriptor._internal_create_key, fields=[ _descriptor.FieldDescriptor( name='source', full_name='vega.commands.v1.OracleDataSubmission.source', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, json_name='source', file=DESCRIPTOR, create_key=_descriptor._internal_create_key), _descriptor.FieldDescriptor( name='payload', full_name='vega.commands.v1.OracleDataSubmission.payload', index=1, number=2, type=12, cpp_type=9, label=1, has_default_value=False, default_value=b"", message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, json_name='payload', file=DESCRIPTOR, create_key=_descriptor._internal_create_key), ], extensions=[ ], nested_types=[], enum_types=[ _ORACLEDATASUBMISSION_ORACLESOURCE, ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=48, serialized_end=251, ) _ORACLEDATASUBMISSION.fields_by_name['source'].enum_type = _ORACLEDATASUBMISSION_ORACLESOURCE _ORACLEDATASUBMISSION_ORACLESOURCE.containing_type = _ORACLEDATASUBMISSION DESCRIPTOR.message_types_by_name['OracleDataSubmission'] = _ORACLEDATASUBMISSION _sym_db.RegisterFileDescriptor(DESCRIPTOR) OracleDataSubmission = _reflection.GeneratedProtocolMessageType('OracleDataSubmission', (_message.Message,), { 'DESCRIPTOR' : _ORACLEDATASUBMISSION, '__module__' : 'commands.v1.oracles_pb2' # @@protoc_insertion_point(class_scope:vega.commands.v1.OracleDataSubmission) }) _sym_db.RegisterMessage(OracleDataSubmission) DESCRIPTOR._options = None # @@protoc_insertion_point(module_scope)
38.495327
480
0.786113
4ea392c525c167ab6e4487bc47072399df2ebdf7
2,788
py
Python
non-regression-tests/config.py
etalab-ia/piaf-ml
cd006b905d4c3e6a358326a42b84179724b00e5f
[ "MIT" ]
5
2021-06-22T08:51:53.000Z
2021-12-14T17:26:32.000Z
non-regression-tests/config.py
etalab-ia/piaf-ml
cd006b905d4c3e6a358326a42b84179724b00e5f
[ "MIT" ]
55
2021-06-16T07:58:16.000Z
2021-08-30T10:30:26.000Z
non-regression-tests/config.py
etalab-ia/piaf-ml
cd006b905d4c3e6a358326a42b84179724b00e5f
[ "MIT" ]
null
null
null
import os parameter_tuning_options = { "experiment_name": "non-regression-tests", # Tuning method alternatives: # - "optimization": use bayesian optimisation # - "grid_search" "tuning_method": "grid_search", # Additionnal options for the grid search method "use_cache": False, # Additionnal options for the optimization method "optimization_ncalls": 10, } parameters_fquad = { "k_retriever": [5], "k_title_retriever" : [1], # must be present, but only used when retriever_type == title_bm25 "k_reader_per_candidate": [20], "k_reader_total": [10], "reader_model_version": ["053b085d851196110d7a83d8e0f077d0a18470be"], "retriever_model_version": ["1a01b38498875d45f69b2a6721bf6fe87425da39"], "dpr_model_version": ["v1.0"], "retriever_type": ["bm25"], # Can be bm25, sbert, dpr, title or title_bm25 "squad_dataset": [ os.getenv("DATA_DIR") + "/non-regression-tests/fquad_dataset.json" ], "filter_level": [None], "preprocessing": [False], "boosting" : [1], #default to 1 "split_by": ["word"], # Can be "word", "sentence", or "passage" "split_length": [1000], } # A dictionnary specifying the criteria a test result must pass. Keys are # metrics names and keys are predicates on the corresponding metric which must # return true if the value is satisfying. pass_criteria_fquad = { "reader_topk_accuracy_has_answer": # metric ~= 0.747 +/- 1% lambda metric: abs(metric / 0.747 - 1) < 0.01 } parameters_dila = { "k_retriever": [1], "k_title_retriever" : [1], # must be present, but only used when retriever_type == title_bm25 "k_reader_per_candidate": [20], "k_reader_total": [10], "reader_model_version": ["053b085d851196110d7a83d8e0f077d0a18470be"], "retriever_model_version": ["1a01b38498875d45f69b2a6721bf6fe87425da39"], "dpr_model_version": ["v1.0"], "retriever_type": ["bm25"], # Can be bm25, sbert, dpr, title or title_bm25 "squad_dataset": [ os.getenv("SRC_DIR") + "/piaf-ml/clients/dila/knowledge_base/squad.json"], "filter_level": [None], "preprocessing": [False], "boosting" : [1], #default to 1 "split_by": ["word"], # Can be "word", "sentence", or "passage" "split_length": [1000], } # A dictionnary specifying the criteria a test result must pass. Keys are # metrics names and keys are predicates on the corresponding metric which must # return true if the value is satisfying. pass_criteria_dila = { "reader_topk_accuracy_has_answer": # metric ~= 0.427 +/- 1% lambda metric: abs(metric / 0.427 - 1) < 0.01 } tests = [ (parameters_fquad, parameter_tuning_options, pass_criteria_fquad), (parameters_dila, parameter_tuning_options, pass_criteria_dila), ]
35.74359
97
0.681133
4ea47fc79c5dcbec42ef206e57f938c9dff9b024
2,101
py
Python
zhang.py
AndrewQuijano/Treespace_REU_2017
e1aff2224ad5152d82f529675444146a70623bca
[ "MIT" ]
2
2021-06-07T12:22:46.000Z
2021-09-14T00:19:03.000Z
zhang.py
AndrewQuijano/Treespace_REU_2017
e1aff2224ad5152d82f529675444146a70623bca
[ "MIT" ]
null
null
null
zhang.py
AndrewQuijano/Treespace_REU_2017
e1aff2224ad5152d82f529675444146a70623bca
[ "MIT" ]
null
null
null
import networkx as nx from misc import maximum_matching_all from networkx import get_node_attributes def is_tree_based(graph): if is_binary(graph): # print("Graph is not binary! Zhang's won't work!") return None unmatched_reticulation = zhang_graph(graph) if len(unmatched_reticulation) == 0: return True else: return False def is_binary(graph): for node in graph.nodes(): if graph.out_degree(node) > 2 or graph.in_degree(node) > 2: return False return True # Use this for non-binary graph def zhang_graph(graph): try: zhang = zhang_bipartite(graph) max_match = maximum_matching_all(zhang) reticulations = [n for n, d in zhang.nodes(data=True) if d['biparite'] == 0] data = get_node_attributes(zhang, 'biparite') matched_reticulations = set() for s, t in max_match.items(): try: if data[s] == 1: matched_reticulations.add(s) if data[t] == 1: matched_reticulations.add(t) except KeyError: continue except nx.exception.NetworkXPointlessConcept: return list() set_minus = set(reticulations) - matched_reticulations return list(set_minus) def zhang_bipartite(graph): zhang = nx.Graph() for node in graph.nodes(): # This is a reticulation vertex if graph.in_degree(node) == 2 and graph.out_degree(node) == 1: zhang.add_node(node, bipartite=0) # BE CAREFUL NOT TO ADD RETICULATIONS AGAIN ON OTHER SIDE! for parent in graph.predecessors(node): if graph.in_degree(parent) == 1 and graph.out_degree(parent) == 2: zhang.add_node(parent, bipartite=1) for e in graph.edges(parent): # Add the edge only if we know the child is a reticulation if graph.in_degree(e[1]) == 2 and graph.out_degree(e[1]) == 1: zhang.add_edge(node, parent) return zhang
32.828125
86
0.595907
4ea5498deec294ffeeebf2d2ad50bbf782de71a8
141
py
Python
esteid/idcard/__init__.py
thorgate/django-esteid
4a4227b20dca7db5441a3514f724f1404575562c
[ "BSD-3-Clause" ]
17
2016-03-30T09:20:19.000Z
2022-01-17T12:04:03.000Z
esteid/idcard/__init__.py
thorgate/django-esteid
4a4227b20dca7db5441a3514f724f1404575562c
[ "BSD-3-Clause" ]
15
2016-02-22T22:49:07.000Z
2021-11-09T12:29:35.000Z
esteid/idcard/__init__.py
thorgate/django-esteid
4a4227b20dca7db5441a3514f724f1404575562c
[ "BSD-3-Clause" ]
2
2016-07-27T10:57:52.000Z
2017-10-05T13:15:59.000Z
__all__ = ["BaseIdCardAuthenticationView", "IdCardSigner"] from .signer import IdCardSigner from .views import BaseIdCardAuthenticationView
28.2
58
0.836879
4eaa032d7c85557301f7b3de83a688e4d6c318a3
101
py
Python
placeable_interface.py
Alex92rus/funkyBlue
747fdbfc72edd85556465204f0f654a5cac32c2a
[ "MIT" ]
2
2020-03-07T19:46:52.000Z
2020-03-08T09:11:02.000Z
placeable_interface.py
Alex92rus/funkyBlue
747fdbfc72edd85556465204f0f654a5cac32c2a
[ "MIT" ]
3
2020-03-07T10:09:31.000Z
2021-01-14T08:40:27.000Z
placeable_interface.py
Alex92rus/funkyBlue
747fdbfc72edd85556465204f0f654a5cac32c2a
[ "MIT" ]
null
null
null
from arcade import Sprite class PlaceableInterface: def place(self, *args): pass
7.769231
27
0.643564
4eaa0630211dc9678f367337b57ebf1618235962
3,765
py
Python
sme_financing/main/apis/document_api.py
BuildForSDG/team-214-backend
f1aff9c27d7b7588b4bbb2bc68956b35051d4506
[ "MIT" ]
1
2020-05-20T16:32:33.000Z
2020-05-20T16:32:33.000Z
sme_financing/main/apis/document_api.py
BuildForSDG/team-214-backend
f1aff9c27d7b7588b4bbb2bc68956b35051d4506
[ "MIT" ]
23
2020-05-19T07:12:53.000Z
2020-06-21T03:57:54.000Z
sme_financing/main/apis/document_api.py
BuildForSDG/team-214-backend
f1aff9c27d7b7588b4bbb2bc68956b35051d4506
[ "MIT" ]
1
2020-05-18T14:18:12.000Z
2020-05-18T14:18:12.000Z
"""RESTful API Document resource.""" from flask_restx import Resource, reqparse from flask_restx._http import HTTPStatus from werkzeug.datastructures import FileStorage from ..service.document_service import ( delete_document, edit_document, get_all_documents, get_document, save_document, ) from .dto import DocumentDTO api = DocumentDTO.document_api _document = DocumentDTO.document parser = reqparse.RequestParser() parser.add_argument("document_name", type=str, help="Document name", location="form") parser.add_argument("file", type=FileStorage, location="files") @api.route("/") class DocumentList(Resource): @api.doc("list of documents") @api.marshal_list_with(_document, envelope="data") def get(self): """List all documents.""" return get_all_documents() @api.doc("Create a new Document") @api.expect(parser, validate=True) @api.response(HTTPStatus.CREATED, "Document successfully saved") @api.response(HTTPStatus.NOT_FOUND, "File not found") @api.response(HTTPStatus.BAD_REQUEST, "File empty") @api.response(HTTPStatus.NOT_ACCEPTABLE, "File extension not allowed") @api.response(HTTPStatus.REQUEST_ENTITY_TOO_LARGE, "File exceeds max upload size") def post(self): """Create a new Document.""" parse_data = parser.parse_args() document_name = parse_data["document_name"] file = parse_data["file"] if not file or not document_name: self.api.abort( code=HTTPStatus.NOT_FOUND, message="File not found or document name empty", ) else: return save_document(document_name, file) @api.route("/<int:doc_id>") @api.param("doc_id", "The ID of the docuemnt to process") @api.response(HTTPStatus.NOT_FOUND, "Document not found") @api.response(HTTPStatus.NOT_ACCEPTABLE, "File and document name empty") class DocumentByID(Resource): @api.doc("Get a single document") @api.marshal_with(_document) def get(self, doc_id): """Retrieve a document.""" document = get_document(doc_id) if not document: self.api.abort(code=HTTPStatus.NOT_FOUND, message="Document not found") else: return document @api.doc("Patch a document") @api.expect(parser) def patch(self, doc_id): """Patch a document.""" document = get_document(doc_id) if not document: self.api.abort(code=HTTPStatus.NOT_FOUND, message="Document not found") else: parse_data = parser.parse_args() document_name = parse_data["document_name"] file = parse_data["file"] if not file and not document_name: self.api.abort(HTTPStatus.NOT_ACCEPTABLE, message="Both inputs empty") else: return edit_document(document, document_name, file) # return self.get(doc_id) @api.doc("Delete a document") @api.response(HTTPStatus.BAD_REQUEST, "Can't delete document") def delete(self, doc_id): """Delete a document.""" document = get_document(doc_id) if not document: self.api.abort(code=HTTPStatus.NOT_FOUND, message="Document not found") else: return delete_document(document) # @api.route("/smes/<sme_id>") # @api.param("sme_id", "The SME id") # @api.response(HTTPStatus.NOT_FOUND, "SME not found") # class DocumentSME(Resource): # @api.doc("List all documents of an SME") # @api.marshal_list_with(_document, envelope="data") # def get(self, sme_id): # """List all documents of an SME.""" # if not get_sme_by_id(sme_id): # api.abort(404) # return get_all_sme_documents(sme_id)
35.186916
86
0.657371
4eaadef4bc857f47d228828cdfd23ca47dfe5099
1,405
py
Python
column_name_renaming.py
strathclyde-rse/strathclyde-software-survey
1dd3805a416f1da6cbfa27958ae96a5ad685fe19
[ "CC-BY-4.0", "BSD-3-Clause" ]
null
null
null
column_name_renaming.py
strathclyde-rse/strathclyde-software-survey
1dd3805a416f1da6cbfa27958ae96a5ad685fe19
[ "CC-BY-4.0", "BSD-3-Clause" ]
null
null
null
column_name_renaming.py
strathclyde-rse/strathclyde-software-survey
1dd3805a416f1da6cbfa27958ae96a5ad685fe19
[ "CC-BY-4.0", "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python # encoding: utf-8 col_shortener = { 'Q1':'confirm', 'Q2':'faculty', 'Q3':'department', 'Q4':'funders', 'Q5':'position', 'Q6':'use_software', 'Q7':'importance_software', 'Q8':'develop_own_code', 'Q9':'development_expertise', 'Q10':'sufficient_training', 'Q11':'want_to_commercialise', 'Q12':'ready_to_release', 'Q13':'hpc_use', 'Q14_1':'version_control', 'Q14_2':'unit_regression_testing', 'Q14_3':'continuous_integration', 'Q14_4':'compilation', 'Q14_5':'documentation', 'Q15':'uni_support', 'Q16':'hired_developer', 'Q17':'costed_developer', 'Q18_1':'hire_full_time_developer', 'Q18_2':'hire_pool_developer', 'Q19':'voucher', 'Q20':'consulting', 'Q21':'mailing' } add_an_other_category = [ 'funders', 'position', 'hpc_use' ] sort_no_further_analysis = [ 'faculty', 'funders', 'position', 'hpc_use' ] yes_no_analysis = [ 'use_software', 'develop_own_code', 'sufficient_training', 'want_to_commercialise', 'ready_to_release', 'hired_developer' ] scale_analysis = [ 'importance_software', 'development_expertise', 'sufficient_training' ] worded_scale_analysis = [ 'version_control', 'continuous_integration', 'unit_regression_testing', 'hire_full_time_developer', 'hire_pool_developer' ]
19.788732
39
0.635587
4eae8bae94764d3c5a64b90797dd929834fa6067
1,974
py
Python
scanpy/tests/test_scaling.py
alexcwsmith/scanpy
b69015e9e7007193c9ac461d5c6fbf845b3d6962
[ "BSD-3-Clause" ]
1,171
2017-01-17T14:01:02.000Z
2022-03-31T23:02:57.000Z
scanpy/tests/test_scaling.py
alexcwsmith/scanpy
b69015e9e7007193c9ac461d5c6fbf845b3d6962
[ "BSD-3-Clause" ]
1,946
2017-01-22T10:19:04.000Z
2022-03-31T17:13:03.000Z
scanpy/tests/test_scaling.py
alexcwsmith/scanpy
b69015e9e7007193c9ac461d5c6fbf845b3d6962
[ "BSD-3-Clause" ]
499
2017-01-21T11:39:29.000Z
2022-03-23T13:57:35.000Z
import pytest import numpy as np from anndata import AnnData from scipy.sparse import csr_matrix import scanpy as sc # test "data" for 3 cells * 4 genes X = [ [-1, 2, 0, 0], [1, 2, 4, 0], [0, 2, 2, 0], ] # with gene std 1,0,2,0 and center 0,2,2,0 X_scaled = [ [-1, 2, 0, 0], [1, 2, 2, 0], [0, 2, 1, 0], ] # with gene std 1,0,1,0 and center 0,2,1,0 X_centered = [ [-1, 0, -1, 0], [1, 0, 1, 0], [0, 0, 0, 0], ] # with gene std 1,0,1,0 and center 0,0,0,0 @pytest.mark.parametrize('typ', [np.array, csr_matrix], ids=lambda x: x.__name__) @pytest.mark.parametrize('dtype', ['float32', 'int64']) def test_scale(typ, dtype): # test AnnData arguments # test scaling with default zero_center == True adata0 = AnnData(typ(X), dtype=dtype) sc.pp.scale(adata0) assert np.allclose(csr_matrix(adata0.X).toarray(), X_centered) # test scaling with explicit zero_center == True adata1 = AnnData(typ(X), dtype=dtype) sc.pp.scale(adata1, zero_center=True) assert np.allclose(csr_matrix(adata1.X).toarray(), X_centered) # test scaling with explicit zero_center == False adata2 = AnnData(typ(X), dtype=dtype) sc.pp.scale(adata2, zero_center=False) assert np.allclose(csr_matrix(adata2.X).toarray(), X_scaled) # test bare count arguments, for simplicity only with explicit copy=True # test scaling with default zero_center == True data0 = typ(X, dtype=dtype) cdata0 = sc.pp.scale(data0, copy=True) assert np.allclose(csr_matrix(cdata0).toarray(), X_centered) # test scaling with explicit zero_center == True data1 = typ(X, dtype=dtype) cdata1 = sc.pp.scale(data1, zero_center=True, copy=True) assert np.allclose(csr_matrix(cdata1).toarray(), X_centered) # test scaling with explicit zero_center == False data2 = typ(X, dtype=dtype) cdata2 = sc.pp.scale(data2, zero_center=False, copy=True) assert np.allclose(csr_matrix(cdata2).toarray(), X_scaled)
35.890909
81
0.662614
4eae8be82d67b6164b7865425e58eaf76d1e1eba
7,810
py
Python
wsireg/tmpSaves/demo_self6_complete unit.py
luweishuang/wsireg
344af8585932e3e0f5df3ce40a7dc75846a0214b
[ "MIT" ]
null
null
null
wsireg/tmpSaves/demo_self6_complete unit.py
luweishuang/wsireg
344af8585932e3e0f5df3ce40a7dc75846a0214b
[ "MIT" ]
null
null
null
wsireg/tmpSaves/demo_self6_complete unit.py
luweishuang/wsireg
344af8585932e3e0f5df3ce40a7dc75846a0214b
[ "MIT" ]
null
null
null
import cv2 import numpy as np import bilinear import patchreg from skimage.util import view_as_windows def bilinear_interpolation_of_patch_registration(master_srcdata, target_srcdata): print("Beginning bilinear_interpolation_of_patch_registration...") w_shape = (1000, 1000, 4) # window_size w_step = (500, 500, 4) # 步长 padding = w_step[0] # must do step padding master_data = cv2.copyMakeBorder(master_srcdata, padding, padding, padding, padding, cv2.BORDER_REFLECT) target_data = cv2.copyMakeBorder(target_srcdata, padding, padding, padding, padding, cv2.BORDER_REFLECT) master_img = cv2.cvtColor(master_data, code=cv2.COLOR_BGRA2RGBA) target_img = cv2.cvtColor(target_data, code=cv2.COLOR_BGRA2RGBA) # Stage One: Low-precision feature alignment h, _ = patchreg.alignFeatures(target_img, master_img) height, width = target_img.shape[:2] master_aligned = cv2.warpPerspective(master_img, h, (width, height)) # Stage Two: Calculate patch-level registrations stack1 = np.concatenate((target_img, master_aligned), axis=-1) # (2000, 40000, 8) patches = view_as_windows(stack1, window_shape=w_shape, step=w_step) morphs = patchreg.calcPlateMorphs(patches) # (3,7,2,3,3) # Stage Three: Compute patch-level DVFs=dense displacement vector field id_patches = patchreg.calc_id_patches(img_shape=master_aligned.shape, patch_size=1000) # (3,7,3,2000,2000,1) map_morphs = np.append(morphs, morphs[:, :, 1, None], axis=2) # (3,7,3,3,3) reg_patches_src = patchreg.applyMorphs(id_patches, map_morphs) # (3,7,3,2000,2000,1) map_patches = reg_patches_src[:, :, 1:, 500:1500, 500:1500, :] # Stage Four: Merge patch-level DVFs into a single global transform. quilts = bilinear.quilter(map_patches) wquilts = bilinear.bilinear_wquilts(map_patches) qmaps = [q * w for q, w in zip(quilts, wquilts)] # 对应位置的元素相乘 qmaps_sum = qmaps[0] + qmaps[1] + qmaps[2] + qmaps[3] summed = (qmaps_sum).reshape(qmaps_sum.shape[:-1]).astype(np.float32) master_remap = cv2.remap(master_img, summed[0], summed[1], interpolation=cv2.INTER_LINEAR) # summed 是坐标映射关系 master_reg = master_remap[padding:height-padding, padding:width-padding, :] return master_reg def draw_img(): master_srcdata = cv2.imread("../data/OK1_1_32.jpg") padding = 500 master_data = cv2.copyMakeBorder(master_srcdata, padding, padding, padding, padding, cv2.BORDER_CONSTANT,value=(255, 255, 255)) cv2.line(master_data, (0, 1000), (5000, 1000), (0, 255, 0), 2) cv2.line(master_data, (0, 2000), (5000, 2000), (0, 255, 0), 2) cv2.line(master_data, (1000, 0), (1000, 3000), (0, 255, 0), 2) cv2.line(master_data, (2000, 0), (2000, 3000), (0, 255, 0), 2) cv2.line(master_data, (3000, 0), (3000, 3000), (0, 255, 0), 2) cv2.line(master_data, (4000, 0), (4000, 3000), (0, 255, 0), 2) cv2.imwrite("master_data.jpg", master_data) def pad_imgs(master3, target3): master_h, master_w, _ = master3.shape target_h, target_w, _ = target3.shape assert master_h == target_h and master_w == target_w src_w = master_w src_h = master_h mid_h = int(max(2000, np.ceil(src_h/1000)*1000)) mid_w = int(max(2000, np.ceil(src_w/1000)*1000)) assert mid_w >= src_w and mid_h >= src_h left_pad = int((mid_w-src_w)/2) right_pad = int(mid_w - src_w - left_pad) top_pad = int((mid_h - src_h) / 2) down_pad = int(mid_h - src_h - top_pad) master3_pad = cv2.copyMakeBorder(master3, top_pad, down_pad, left_pad, right_pad, cv2.BORDER_REFLECT) target3_pad = cv2.copyMakeBorder(target3, top_pad, down_pad, left_pad, right_pad, cv2.BORDER_REFLECT) return master3_pad, target3_pad, top_pad, down_pad, left_pad, right_pad MAX_FEATURES = 5000 GOOD_MATCH_PERCENT = 0.45 def alignImages_Perspective(img1, img2): im1Gray = cv2.cvtColor(img1, cv2.COLOR_BGR2GRAY) im2Gray = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY) # Detect ORB features and compute descriptors. orb = cv2.ORB_create(MAX_FEATURES) keypoints1, descriptors1 = orb.detectAndCompute(im1Gray, None) keypoints2, descriptors2 = orb.detectAndCompute(im2Gray, None) # Match features. matcher = cv2.DescriptorMatcher_create(cv2.DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMING) matches = list(matcher.match(descriptors1, descriptors2, None)) # Sort matches by score matches.sort(key=lambda x: x.distance, reverse=False) # Remove not so good matches numGoodMatches = int(len(matches) * GOOD_MATCH_PERCENT) matches = matches[:numGoodMatches] # Extract location of good matches points1 = np.zeros((len(matches), 2), dtype=np.float32) points2 = np.zeros((len(matches), 2), dtype=np.float32) for i, match in enumerate(matches): points1[i, :] = keypoints1[match.queryIdx].pt points2[i, :] = keypoints2[match.trainIdx].pt height, width, channels = img2.shape # Perspective h, mask = cv2.findHomography(points1, points2, cv2.RANSAC) im1Reg_Perspective = cv2.warpPerspective(img1, h, (width, height)) # 透视变换 return im1Reg_Perspective def process_single_imgpart(img_master, target_img): master_height, master_width, _ = img_master.shape cur_height, cur_width, _ = target_img.shape assert cur_width == master_width top_pad, down_pad = 0, 0 target_imgpad = target_img.copy() if master_height > cur_height: top_pad = int((master_height - cur_height)/2) down_pad = master_height - cur_height - top_pad target_imgpad = cv2.copyMakeBorder(target_img, top_pad, down_pad, 0, 0, cv2.BORDER_CONSTANT, value=(255, 255, 255)) elif master_height < cur_height: print("cur_height > master_height", cur_height, master_height) img_show = target_imgpad.copy() im2Gray = cv2.cvtColor(target_imgpad, cv2.COLOR_BGR2GRAY) im1Reg_Perspective = alignImages_Perspective(img_master, target_imgpad) imRegGray = cv2.cvtColor(im1Reg_Perspective, cv2.COLOR_BGR2GRAY) diff = cv2.absdiff(imRegGray, im2Gray) # cv2.imwrite("diff.jpg", diff) ret, thresh = cv2.threshold(diff, 120, 255, cv2.THRESH_BINARY) # 120 # cv2.imwrite("thresh.jpg", thresh) contours, hierarchy = cv2.findContours(thresh, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_NONE) for cnt in contours: area = cv2.contourArea(cnt) if area > 1 and area < max(cur_height, cur_width): cv2.drawContours(img_show, cnt, -1, (0, 0, 255), 2) img_out = img_show[top_pad: master_height - down_pad, :, : ] return img_out if __name__ == "__main__": # draw_img() # exit() root = "../data/" master_srcdata = cv2.imread(root + "OK1_1.jpg") target_srcdata = cv2.imread(root + "NG1_1.jpg") master3 = master_srcdata[300:4850,:,:] # cv2.imwrite("master3.jpg", master3) target3 = target_srcdata[720:5270,:,:] # cv2.imwrite("target3.jpg", target3) # padding to 1000s, at least 2000 master3_pad, target3_pad, top_pad, down_pad, left_pad, right_pad = pad_imgs(master3, target3) # cv2.imwrite("master3_pad.jpg", master3_pad) # cv2.imwrite("target3_pad.jpg", target3_pad) masterpad_h, masterpad_w, _ = master3_pad.shape master_reg_pad = bilinear_interpolation_of_patch_registration(master3_pad, target3_pad) master3_reg = master_reg_pad[top_pad: masterpad_h-down_pad, left_pad:masterpad_w-right_pad, : ] cv2.imwrite("master3_reg.jpg", master3_reg) cv2.imwrite("master3.jpg", master3) cv2.imwrite("target3.jpg", target3) # Stage Five: high-precision feature alignment master_reg_out = process_single_imgpart(master3_reg, target3) cv2.imwrite("master_reg_out.jpg", master_reg_out) master_out = process_single_imgpart(master3, target3) cv2.imwrite("master_out.jpg", master_out)
43.631285
131
0.705634
4eaf9ec2243bbc0b3558c08de925bc43b8365f96
1,253
py
Python
src/sprites/weapon_vfx.py
mgear2/undervoid
6c91a5786d29d766223831190952fd90ddc6a1e8
[ "MIT" ]
1
2020-08-29T06:41:03.000Z
2020-08-29T06:41:03.000Z
src/sprites/weapon_vfx.py
mgear2/undervoid
6c91a5786d29d766223831190952fd90ddc6a1e8
[ "MIT" ]
10
2019-07-15T05:15:38.000Z
2020-11-25T03:14:03.000Z
src/sprites/weapon_vfx.py
mgear2/undervoid
6c91a5786d29d766223831190952fd90ddc6a1e8
[ "MIT" ]
1
2020-11-22T08:25:26.000Z
2020-11-22T08:25:26.000Z
# Copyright (c) 2020 # [This program is licensed under the "MIT License"] # Please see the file LICENSE in the source # distribution of this software for license terms. import pygame as pg import ruamel.yaml from random import choice vec = pg.math.Vector2 class Weapon_VFX(pg.sprite.Sprite): """ Weapon_VFX appear when the player is shooting. Cycling between available img options provides animation effect. """ def __init__( self, settings: ruamel.yaml.comments.CommentedMap, game_client_data_weaponvfx: list, pos: vec, ): self.settings = settings self._layer = self.settings["layer"]["vfx"] pg.sprite.Sprite.__init__(self) self.image = pg.transform.scale( choice(game_client_data_weaponvfx), ( self.settings["gen"]["tilesize"], self.settings["gen"]["tilesize"], ), ) self.rect = self.image.get_rect() self.pos = self.rect.center = pos self.spawn_time = pg.time.get_ticks() def update(self): if ( pg.time.get_ticks() - self.spawn_time > self.settings["weapon"]["vbullet"]["fx_life"] ): self.kill()
27.23913
59
0.60016
4eafa55a23b75bd6783941216b9d9087a84c8b15
9,860
py
Python
game/blenderpanda/pman.py
Moguri/prototype-nitrogen
607f78219fcfbd55dfcd1611684107a2922f635d
[ "Apache-2.0" ]
1
2017-05-29T23:03:13.000Z
2017-05-29T23:03:13.000Z
game/blenderpanda/pman.py
Moguri/prototype-nitrogen
607f78219fcfbd55dfcd1611684107a2922f635d
[ "Apache-2.0" ]
null
null
null
game/blenderpanda/pman.py
Moguri/prototype-nitrogen
607f78219fcfbd55dfcd1611684107a2922f635d
[ "Apache-2.0" ]
null
null
null
import fnmatch import os import shutil import subprocess import sys import time from collections import OrderedDict try: import configparser except ImportError: import ConfigParser as configparser class PManException(Exception): pass class NoConfigError(PManException): pass class CouldNotFindPythonError(PManException): pass class BuildError(PManException): pass class FrozenEnvironmentError(PManException): def __init__(self): PManException.__init__(self, "Operation not supported in frozen applications") if '__file__' not in globals(): __is_frozen = True __file__ = '' else: __is_frozen = False _config_defaults = OrderedDict([ ('general', OrderedDict([ ('name', 'Game'), ('render_plugin', ''), ])), ('build', OrderedDict([ ('asset_dir', 'assets/'), ('export_dir', 'game/assets/'), ('ignore_patterns', '*.blend1, *.blend2'), ])), ('run', OrderedDict([ ('main_file', 'game/main.py'), ('auto_build', True), ('auto_save', True), ])), ]) _user_config_defaults = OrderedDict([ ('blender', OrderedDict([ ('last_path', 'blender'), ('use_last_path', True), ])), ]) def __py2_read_dict(config, d): for section, options in d.items(): config.add_section(section) for option, value in options.items(): config.set(section, option, value) def _get_config(startdir, conf_name, defaults): try: if startdir is None: startdir = os.getcwd() except FileNotFoundError: # The project folder was deleted on us raise NoConfigError("Could not find config file") dirs = os.path.abspath(startdir).split(os.sep) while dirs: cdir = os.sep.join(dirs) if cdir.strip() and conf_name in os.listdir(cdir): configpath = os.path.join(cdir, conf_name) config = configparser.ConfigParser() if hasattr(config, 'read_dict'): config.read_dict(defaults) else: __py2_read_dict(config, defaults) config.read(configpath) config.add_section('internal') config.set('internal', 'projectdir', os.path.dirname(configpath)) return config dirs.pop() # No config found raise NoConfigError("Could not find config file") def get_config(startdir=None): return _get_config(startdir, '.pman', _config_defaults) def get_user_config(startdir=None): try: return _get_config(startdir, '.pman.user', _user_config_defaults) except NoConfigError: # No user config, just create one config = get_config(startdir) fp = os.path.join(config.get('internal', 'projectdir'), '.pman.user') print("Creating user config at {}".format(fp)) with open(fp, 'w') as f: pass return _get_config(startdir, '.pman.user', _user_config_defaults) def _write_config(config, conf_name): writecfg = configparser.ConfigParser() writecfg.read_dict(config) writecfg.remove_section('internal') with open(os.path.join(config.get('internal', 'projectdir'), conf_name), 'w') as f: writecfg.write(f) def write_config(config): _write_config(config, '.pman') def write_user_config(user_config): _write_config(user_config, '.pman.user') def is_frozen(): return __is_frozen def get_python_program(config): python_programs = [ 'ppython', 'python3', 'python', 'python2', ] # Check to see if there is a version of Python that can import panda3d for pyprog in python_programs: args = [ pyprog, '-c', 'import panda3d.core; import direct', ] with open(os.devnull, 'w') as fp: try: retcode = subprocess.call(args, stderr=fp) except FileNotFoundError: retcode = 1 if retcode == 0: return pyprog # We couldn't find a python program to run raise CouldNotFindPythonError('Could not find a usable Python install') def create_project(projectdir): if is_frozen(): raise FrozenEnvironmentError() confpath = os.path.join(projectdir, '.pman') if os.path.exists(confpath): print("Updating project in {}".format(projectdir)) else: print("Creating new project in {}".format(projectdir)) # Touch config file to make sure it is present with open(confpath, 'a') as f: pass config = get_config(projectdir) write_config(config) templatedir = os.path.join(os.path.dirname(__file__), 'templates') print("Creating directories...") dirs = [ 'assets', 'game', ] bpanda_mod_files = [ os.path.join(templatedir, '__init__.py'), os.path.join(templatedir, 'bpbase.py'), 'rendermanager.py', 'pman.py', 'pman_build.py', ] dirs = [os.path.join(projectdir, i) for i in dirs] for d in dirs: if os.path.exists(d): print("\tSkipping existing directory: {}".format(d)) else: print("\tCreating directory: {}".format(d)) os.mkdir(d) print("Creating main.py") with open(os.path.join(templatedir, 'main.py')) as f: main_data = f.read() mainpath = os.path.join(projectdir, 'game', 'main.py') if os.path.exists(mainpath): print("\tmain.py already exists at {}".format(mainpath)) else: with open(mainpath, 'w') as f: f.write(main_data) print("\tmain.py created at {}".format(mainpath)) bpmodpath = os.path.join(projectdir, 'game/blenderpanda') if os.path.exists(bpmodpath): print("Updating blenderpanda module") shutil.rmtree(bpmodpath) else: print("Creating blenderpanda module") os.mkdir(bpmodpath) for cf in bpanda_mod_files: bname = os.path.basename(cf) print("\tCopying over {}".format(bname)) cfsrc = os.path.join(os.path.dirname(__file__), cf) cfdst = os.path.join(projectdir, 'game', 'blenderpanda', bname) shutil.copy(cfsrc, cfdst) print("\t\t{} created at {}".format(bname, cfdst)) def get_abs_path(config, path): return os.path.join( config.get('internal', 'projectdir'), path ) def get_rel_path(config, path): return os.path.relpath(path, config.get('internal', 'projectdir')) def build(config=None): if is_frozen(): raise FrozenEnvironmentError() if config is None: config = get_config() user_config = get_user_config(config.get('internal', 'projectdir')) if hasattr(time, 'perf_counter'): stime = time.perf_counter() else: stime = time.time() print("Starting build") srcdir = get_abs_path(config, config.get('build', 'asset_dir')) dstdir = get_abs_path(config, config.get('build', 'export_dir')) if not os.path.exists(srcdir): raise BuildError("Could not find asset directory: {}".format(srcdir)) if not os.path.exists(dstdir): print("Creating asset export directory at {}".format(dstdir)) os.makedirs(dstdir) print("Read assets from: {}".format(srcdir)) print("Export them to: {}".format(dstdir)) ignore_patterns = [i.strip() for i in config.get('build', 'ignore_patterns').split(',')] print("Ignoring file patterns: {}".format(ignore_patterns)) num_blends = 0 for root, dirs, files in os.walk(srcdir): for asset in files: src = os.path.join(root, asset) dst = src.replace(srcdir, dstdir) ignore_pattern = None for pattern in ignore_patterns: if fnmatch.fnmatch(asset, pattern): ignore_pattern = pattern break if ignore_pattern is not None: print('Skip building file {} that matched ignore pattern {}'.format(asset, ignore_pattern)) continue if asset.endswith('.blend'): dst = dst.replace('.blend', '.bam') if os.path.exists(dst) and os.stat(src).st_mtime <= os.stat(dst).st_mtime: print('Skip building up-to-date file: {}'.format(dst)) continue if asset.endswith('.blend'): # Handle with Blender num_blends += 1 else: print('Copying non-blend file from "{}" to "{}"'.format(src, dst)) if not os.path.exists(os.path.dirname(dst)): os.makedirs(os.path.dirname(dst)) shutil.copyfile(src, dst) if num_blends > 0: blender_path = user_config.get('blender', 'last_path') if user_config.getboolean('blender', 'use_last_path') else 'blender' args = [ blender_path, '-b', '-P', os.path.join(os.path.dirname(__file__), 'pman_build.py'), '--', srcdir, dstdir, ] #print("Calling blender: {}".format(' '.join(args))) subprocess.call(args, env=os.environ.copy()) if hasattr(time, 'perf_counter'): etime = time.perf_counter() else: etime = time.time() print("Build took {:.4f}s".format(etime - stime)) def run(config=None): if is_frozen(): raise FrozenEnvironmentError() if config is None: config = get_config() if config.getboolean('run', 'auto_build'): build(config) mainfile = get_abs_path(config, config.get('run', 'main_file')) print("Running main file: {}".format(mainfile)) args = [get_python_program(config), mainfile] #print("Args: {}".format(args)) subprocess.Popen(args, cwd=config.get('internal', 'projectdir'))
27.853107
131
0.600406
4eb07bf2ab74b26ed4d8db65e2b44e12fd9bf220
1,326
py
Python
src/createGraph.py
AJMFactsheets/NetworkSpeedGrapher
86e755e8831ab22394719520713d4949ed3d018e
[ "Apache-2.0" ]
null
null
null
src/createGraph.py
AJMFactsheets/NetworkSpeedGrapher
86e755e8831ab22394719520713d4949ed3d018e
[ "Apache-2.0" ]
null
null
null
src/createGraph.py
AJMFactsheets/NetworkSpeedGrapher
86e755e8831ab22394719520713d4949ed3d018e
[ "Apache-2.0" ]
null
null
null
import sys import plotly import plotly.plotly as py import plotly.graph_objs as go #Argument 1 must be your plotly username, argument 2 is your api key. Get those by registering for a plotly account. #Argument 3 is the name of the input file to input data from. Must be in the form: Date \n Download \n Upload \n plotly.tools.set_credentials_file(username=sys.argv[1], api_key=sys.argv[2]) time = [] download = [] upload = [] lnum = 1 x = 1 file = open(sys.argv[3], 'r') for line in file: if lnum == 1: #time.append(line[11:13]) time.append(x) x += 1 lnum = 2 elif lnum == 2: download.append(line[10:15]) lnum = 3 elif lnum == 3: upload.append(line[8:12]) lnum = 1 else: raise SystemError('lnum internal error', lnum) #trace1 = go.Histogram( # x=time, # y=download, # opacity=0.75 #) #trace2 = go.Histogram( # x=time, # y=upload, # opacity=0.75 #) #data = [trace1, trace2] #layout = go.Layout(barmode='overlay') #fig = go.Figure(data=data, layout=layout) #py.iplot(fig, filename='Network Speed Graph') trace1 = go.Bar( x=time, y=download, name='Download Speed' ) trace2 = go.Bar( x=time, y=upload, name='Upload Speed' ) data = [trace1, trace2] layout = go.Layout( barmode='group' ) fig = go.Figure(data=data, layout=layout) py.iplot(fig, filename='Network Speed Graph')
17
116
0.6727
4eb27769bbc6f1af6058f15f8a964479f5a48ebc
484
py
Python
crosshair/libimpl/__init__.py
mristin/CrossHair
66a44a0d10021e0b1e2d847a677274e62ddd1e9d
[ "MIT" ]
null
null
null
crosshair/libimpl/__init__.py
mristin/CrossHair
66a44a0d10021e0b1e2d847a677274e62ddd1e9d
[ "MIT" ]
null
null
null
crosshair/libimpl/__init__.py
mristin/CrossHair
66a44a0d10021e0b1e2d847a677274e62ddd1e9d
[ "MIT" ]
null
null
null
from crosshair.libimpl import builtinslib from crosshair.libimpl import collectionslib from crosshair.libimpl import datetimelib from crosshair.libimpl import mathlib from crosshair.libimpl import randomlib from crosshair.libimpl import relib def make_registrations(): builtinslib.make_registrations() collectionslib.make_registrations() datetimelib.make_registrations() mathlib.make_registrations() randomlib.make_registrations() relib.make_registrations()
30.25
44
0.82438
4eb3cdb71cd64b8402ec42bddffc7a5a65442095
690
py
Python
schedulesy/apps/ade_legacy/serializers.py
unistra/schedulesy
bcd8c42281013f02ecd5c89fba9b622f20e47761
[ "Apache-2.0" ]
1
2020-07-24T19:17:56.000Z
2020-07-24T19:17:56.000Z
schedulesy/apps/ade_legacy/serializers.py
unistra/schedulesy
bcd8c42281013f02ecd5c89fba9b622f20e47761
[ "Apache-2.0" ]
1
2020-07-09T10:23:28.000Z
2020-07-09T10:23:28.000Z
schedulesy/apps/ade_legacy/serializers.py
unistra/schedulesy
bcd8c42281013f02ecd5c89fba9b622f20e47761
[ "Apache-2.0" ]
null
null
null
import re from django.urls import reverse from rest_framework import serializers from schedulesy.apps.ade_legacy.models import Customization class CustomizationSerializer(serializers.ModelSerializer): configuration = serializers.SerializerMethodField() class Meta: model = Customization fields = '__all__' def to_internal_value(self, data): d = super().to_internal_value(data) if 'configuration' in data and type(data['configuration'] == dict): d['configuration'] = data['configuration'] return d def get_configuration(self, obj): lc = obj.local_customization return lc.configuration if lc else {}
26.538462
75
0.701449
4eb6277eff4239146619cfdeeb4696e25cfb8808
927
py
Python
tests/test_operator_filter.py
gva-jjoyce/gva_data
cda990d0abb4b175025aaf16e75192bd9cc213af
[ "Apache-2.0" ]
null
null
null
tests/test_operator_filter.py
gva-jjoyce/gva_data
cda990d0abb4b175025aaf16e75192bd9cc213af
[ "Apache-2.0" ]
24
2020-12-24T12:21:42.000Z
2021-01-28T14:22:38.000Z
tests/test_operator_filter.py
gva-jjoyce/gva_data
cda990d0abb4b175025aaf16e75192bd9cc213af
[ "Apache-2.0" ]
null
null
null
""" Test Filter Operator """ import os import sys sys.path.insert(1, os.path.join(sys.path[0], '..')) from gva.flows.operators import FilterOperator try: from rich import traceback traceback.install() except ImportError: pass def test_filter_operator_default(): in_d = {'a':1} n = FilterOperator() d, c = n.execute(in_d) assert d == in_d def test_filter_operator(): ds = [ {"value":1}, {"value":2}, {"value":3}, {"value":4} ] def is_even(val): return val.get('value') % 2 == 0 op = FilterOperator(condition=is_even) res = [op(row) for row in ds] assert res[0] is None assert res[1] == ({'value': 2}, {}) assert res[2] is None assert res[3] == ({'value': 4}, {}) if __name__ == "__main__": test_filter_operator_default() test_filter_operator() print('okay')
20.152174
52
0.558792
4eb65c2f13bb97d8948357e8ad1093ac25bd46cd
5,620
py
Python
python/pynamics/quaternion.py
zmpatel19/Foldable-Robotics
97590ec7d173cc1936cc8ff0379b16ad63bcda23
[ "MIT" ]
2
2018-08-20T22:01:18.000Z
2021-04-19T00:50:56.000Z
python/pynamics/quaternion.py
zmpatel19/Foldable-Robotics
97590ec7d173cc1936cc8ff0379b16ad63bcda23
[ "MIT" ]
3
2017-10-24T03:10:17.000Z
2017-10-24T03:15:27.000Z
python/pynamics/quaternion.py
zmpatel19/Foldable-Robotics
97590ec7d173cc1936cc8ff0379b16ad63bcda23
[ "MIT" ]
2
2017-03-03T23:04:17.000Z
2021-03-20T20:33:53.000Z
# -*- coding: utf-8 -*- """ Created on Tue May 25 10:24:05 2021 @author: danaukes https://en.wikipedia.org/wiki/Rotation_formalisms_in_three_dimensions https://en.wikipedia.org/wiki/Quaternions_and_spatial_rotation https://en.wikipedia.org/wiki/Conversion_between_quaternions_and_Euler_angles """ import sympy sympy.init_printing(pretty_print=False) from sympy import sin,cos,tan,pi,acos import numpy def array(input1): # return numpy.array(input1) return sympy.Matrix(input1) def cross(a,b): # return numpy.cross(a,b) return a.cross(b) def dot(a,b): return a.dot(b) class Quaternion(object): def __init__(self,e0,e1,e2,e3): self.e = [e0,e1,e2,e3] @classmethod def build_from_axis_angle(cls,theta,x,y,z): e0 = cos(theta/2) s = sin(theta/2) e1 = s*x e2 = s*y e3 = s*z return UnitQuaternion(e0,e1,e2,e3) def norm(self): return self.norm_squared()**.5 def norm_squared(self): e = sympy.Matrix(self.e) return sum([item**2 for item in e]) def __mul__(self,other): if type(other) in [int,float]: other = Quaternion(other,0,0,0) return self.hamilton_product(other) def __truediv__(self,other): if type(other) in [int,float]: other = Quaternion(1/other,0,0,0) return self.hamilton_product(other) else: other = Quaternion(1/other,0,0,0) return self.hamilton_product(other) # raise TypeError def __str__(self): e = [] for item in self.e: if type(item) in [int,float]: a='{0:.3f}'.format(item) else: a=str(item) e.append(a) s = 'Q({0},{1},{2},{3})'.format(*e) return s def __repr__(self): return str(self) def hamilton_product(self,other): e01 = self.e[0] e02 = other.e[0] e11 = self.e[1] e12 = other.e[1] e21 = self.e[2] e22 = other.e[2] e31 = self.e[3] e32 = other.e[3] e0 = e01*e02-e11*e12-e21*e22-e31*e32 e1 = e01*e12+e11*e02+e21*e32-e31*e22 e2 = e01*e22-e11*e32+e21*e02+e31*e12 e3 = e01*e32+e11*e22-e21*e12+e31*e02 return Quaternion(e0,e1,e2,e3) def scalar(self): return self.e0 def vector(self): vector = self.e[1:] return array(vector) def conjugate(self): new = type(self)(self.e[0],-self.e[1],-self.e[2],-self.e[3]) return new def inv(self): new = self.conjugate()/self.norm_squared() return new def unit(self): result = self/self.norm() return UnitQuaternion(*result) def rotate_by(self,q): new = q.rotate(self) return new def sum(self,other): new = Quaternion(self.e[0]+other.e[0],self.e[1]+other.e[1],self.e[2]+other.e[2],self.e[3]+other.e[3]) return new # def multiply(self,other): # e0 = self.e0*other.e0-dot(self.vector(),other.vector()) # v = self.e0*other.vector()+other.e0*self.vector()+cross(self.vector(),other.vector()) # new = Quaternion(e0,*v) # return new def expand(self): e = sympy.Matrix(self.e) new = Quaternion(*(e.expand())) return new def conjugation(self,other): result = other*self*other.inv() return result def __getitem__(self, index): if isinstance(index, int): return self.e[index] elif isinstance(index, slice): return self.e[index] def __setitem__(self, index, v): if isinstance(index, int): self.e[index] = v elif isinstance(index, slice): if isinstance(v,Quaternion): self.list[index] = v.e elif isinstance(v,list): self.list[index] = v else: raise(Exception()) def __iter__(self): for item in self.e: yield item def __len__(self): return len(self.e) class UnitQuaternion(Quaternion): @classmethod def build_from_axis_angle(cls,theta,x,y,z): e0 = cos(theta/2) s = sin(theta/2) e1 = s*x e2 = s*y e3 = s*z return cls(e0,e1,e2,e3) def hamilton_product(self, other): result = super(UnitQuaternion,self).hamilton_product(other) if isinstance(other,UnitQuaternion): result = UnitQuaternion(*result) return result def rotate(self,other): l=len(other) if l==3: other = Quaternion(0,*other) result = other.conjugation(self) if l==3: result=result.vector() if isinstance(other,UnitQuaternion): return UnitQuaternion(*result) else: return result # q = self # t = 2*cross(q.vector(),v.vector()) # new = v.vector()+q.e[0]*t+cross(q.vector(),t) # new = Quaternion(sympy.Number(0),*new) # return new def inv(self): return self.conjugate() def unit(self): return self class VectorQuaternion(Quaternion): def __init__(self,e1,e2,e3): self.e=[0,e1,e2,e3] import sympy a,b,c,d = sympy.symbols('a,b,c,d') e,f,g,h = sympy.symbols('e,f,g,h') q = sympy.Symbol('q') v1 = Quaternion(a,b,c,d) v12 = [b,c,d] q = UnitQuaternion(e,f,g,h) # q = Quaternion.build_from_axis_angle(q, 0,0,1) # v1 = Quaternion(0,2,3,4) v2 = v1.rotate_by(q) v22 = q*v1*q.inv() v3 = q.rotate(v12)
27.54902
109
0.55694
4eb77f4c11a2d3ec08d7055fbeacf7a5223e4aad
630
py
Python
src/spellbot/migrations/versions/6e982c9318a6_adds_voice_category_per_channel.py
lexicalunit/spellbot
17a4999d5e1def06246727ac5481230aa4a4557d
[ "MIT" ]
13
2020-07-03T01:20:54.000Z
2021-11-22T06:06:21.000Z
src/spellbot/migrations/versions/6e982c9318a6_adds_voice_category_per_channel.py
lexicalunit/spellbot
17a4999d5e1def06246727ac5481230aa4a4557d
[ "MIT" ]
660
2020-06-26T02:52:18.000Z
2022-03-31T14:14:02.000Z
src/spellbot/migrations/versions/6e982c9318a6_adds_voice_category_per_channel.py
lexicalunit/spellbot
17a4999d5e1def06246727ac5481230aa4a4557d
[ "MIT" ]
3
2020-07-12T06:18:39.000Z
2021-06-22T06:54:47.000Z
"""Adds voice category per channel Revision ID: 6e982c9318a6 Revises: ef54f035a75c Create Date: 2021-12-03 13:18:57.468342 """ import sqlalchemy as sa from alembic import op # revision identifiers, used by Alembic. revision = "6e982c9318a6" down_revision = "ef54f035a75c" branch_labels = None depends_on = None def upgrade(): op.add_column( "channels", sa.Column( "voice_category", sa.String(length=50), nullable=True, server_default=sa.text("'SpellBot Voice Channels'"), ), ) def downgrade(): op.drop_column("channels", "voice_category")
19.6875
64
0.655556