diff --git "a/4171.jsonl" "b/4171.jsonl" new file mode 100644--- /dev/null +++ "b/4171.jsonl" @@ -0,0 +1,669 @@ +{"seq_id":"72039623616","text":"import requests\nimport json, os, time\n\nEASY = \"https://opentdb.com/api.php?amount=10&category=31&difficulty=easy&type=multiple\"\nMEDIUM = \"https://opentdb.com/api.php?amount=10&category=31&difficulty=medium&type=multiple\"\nHARD = \"https://opentdb.com/api.php?amount=10&category=31&difficulty=hard&type=multiple\"\n\ndef set_questions():\n difficulty = get_difficulty()\n if difficulty == \"1\" or difficulty.lower() == \"easy\": response = requests.get(EASY)\n elif difficulty == \"2\" or difficulty.lower() == \"medium\": response = requests.get(MEDIUM)\n elif difficulty == \"3\" or difficulty.lower() == \"hard\": response = requests.get(HARD)\n else: print(\"Invalid Difficulty\"); exit(0)\n Questions_Dict = json.loads(response.text)\n\n questions = []\n correct_answer = []\n incorrect_answers = []\n for data in Questions_Dict[\"results\"]:\n questions.append(data[\"question\"])\n correct_answer.append(data[\"correct_answer\"])\n incorrect_answers.append(data[\"incorrect_answers\"])\n\n return list(zip(questions, correct_answer, incorrect_answers))\n\n\ndef get_difficulty():\n print(\"1 - Easy\")\n print(\"2 - Medium\")\n print(\"3 - Hard\")\n difficulty = input(\"Choose Difficulty: \")\n return difficulty\n\ndef game():\n QUESTIONS = set_questions()\n SLEEP = 1.5\n points = 0\n\n question_count = 0\n while len(QUESTIONS) > question_count:\n os.system(\"cls\")\n correct_answer = []\n correct_answer.append(QUESTIONS[question_count][1])\n\n print(QUESTIONS[question_count][0] + \"\\n\")\n print(\" \".join(n + \" - \" for n in correct_answer + QUESTIONS[question_count][2])[0:-2])\n your_answer = input(\"Enter Your Answer: \")\n\n if your_answer.title() == QUESTIONS[question_count][1]:\n print(\"Good Job! You Earn 10 Points\")\n points += 10\n print(\"Total Points: {}\".format(points))\n question_count += 1\n time.sleep(SLEEP)\n elif your_answer.title() in QUESTIONS[question_count][2]:\n print(\"Sorry, Wrong Answer\")\n question_count += 1\n time.sleep(SLEEP)\n else:\n print(\"There is No Such Choice\")\n time.sleep(SLEEP)\n\ndef main():\n game()\n \n \n\nmain()\n","repo_name":"cetinboran/QuestionGame","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2241,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"35773214640","text":"import pytest\n\n\nclass EmptyListError(Exception):\n pass\n\n\ndef roman_to_int(s: str) -> int:\n \"\"\"\n :param s: A roman form number\n :return: The corresponding integer\n \"\"\"\n if len(s) < 1:\n raise EmptyListError('The roman number should have at least 1 letter')\n return -1\n d = {'I': 1, 'V': 5, 'X': 10, 'L': 50, 'C': 100, 'D': 500, 'M': 1000}\n i, res = 0, 0\n while i < len(s):\n if i < len(s) - 1 and d[s[i]] < d[s[i + 1]]:\n res += d[s[i + 1]] - d[s[i]]\n i += 2\n else:\n res += d[s[i]]\n i += 1\n return res\n\n\n@pytest.mark.parametrize('roman, res', [\n ('III', 3),\n ('LVIII', 58),\n ('MCMXCIV', 1994),\n])\ndef test_roman_to_int(roman, res):\n assert roman_to_int(roman) == res\n\n\n@pytest.mark.parametrize('roman, res', [\n ('', -1),\n])\ndef test_two_sum_not_enough_items(roman, res):\n with pytest.raises(EmptyListError):\n assert roman_to_int(roman) == res\n","repo_name":"dmb225/algorithms","sub_path":"source/strings/roman_to_int.py","file_name":"roman_to_int.py","file_ext":"py","file_size_in_byte":966,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38625349012","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Apr 12 12:18:57 2020\n\n@author: ekshe\n\"\"\"\nimport requests\nimport pandas as pd\nfrom selenium import webdriver\n\nfrom bs4 import BeautifulSoup as bs\nfrom bs4 import Comment\n\n#res = requests.get('https://www.pro-football-reference.com/boxscores/201909050chi.htm')\n\nteams = ['nwe', 'buf', 'nyj', 'mia', 'rav', 'pit', 'cle', 'cin', 'phi', 'dal', 'nyg', 'was',\\\n 'gnb', 'min', 'chi', 'det', 'htx', 'jax', 'clt', 'oti', 'nor', 'atl', 'car', 'tam',\\\n 'kan', 'sdg', 'den', 'rai', 'crd', 'sfo', 'ram', 'sea']\n\ndata = pd.DataFrame()\nfor team in teams: \n driver = webdriver.Chrome()\n driver.get('https://www.pro-football-reference.com/teams/{}/2019_advanced.htm'.format(team))\n rest = driver.page_source\n driver.quit()\n \n \n p = pd.DataFrame(pd.read_html(rest)[-1])\n p.columns = p.columns.get_level_values(1)\n p['team'] = team\n data = data.append(p)\n \n\ndata.to_csv('data/defensive_stats.csv')\n\n\n\n\n","repo_name":"ekethio/NFL-Analysis","sub_path":"defensive_stats_scraper.py","file_name":"defensive_stats_scraper.py","file_ext":"py","file_size_in_byte":974,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"14217168636","text":"from netsquid.protocols import NodeProtocol\nfrom netsquid.components import QSource,Clock,QuantumProgram\nfrom netsquid.components.qsource import SourceStatus\nfrom netsquid.components.instructions import INSTR_H,INSTR_CNOT,INSTR_MEASURE,INSTR_MEASURE_X\n\nimport sys\nscriptpath = \"../../lib/\"\nsys.path.append(scriptpath)\nfrom functions import Compare_basis,Random_basis_gen\n\n\n\n\nclass QG_A_qPrepare(QuantumProgram):\n \n def __init__(self,num_bits=1):\n self.num_bits=num_bits\n super().__init__()\n \n def program(self):\n qList_idx=self.get_qubit_indices(2*self.num_bits)\n # create multiEPR\n for i in range(2*self.num_bits):\n if i%2==0: # List A case\n self.apply(INSTR_H, qList_idx[i])\n else: # List B case\n self.apply(INSTR_CNOT, [qList_idx[i-1], qList_idx[i]])\n yield self.run(parallel=False)\n\n\nclass QG_A_measure(QuantumProgram):\n def __init__(self,basisList,num_bits):\n self.basisList=basisList\n self.num_bits=num_bits\n super().__init__()\n\n\n def program(self): \n for i in range(0,len(self.basisList*2),2):\n if self.basisList[int(i/2)] == 0: # only even slot \n self.apply(INSTR_MEASURE, \n qubit_indices=i, output_key=str(i),physical=True) # standard basis\n else: # 1 case \n self.apply(INSTR_MEASURE_X, \n qubit_indices=i, output_key=str(i),physical=True) # Hadamard basis\n yield self.run(parallel=False)\n \n\n\n\nclass AliceProtocol(NodeProtocol):\n \n def __init__(self,node,processor,num_bits,sourceFreq,\n port_names=[\"portQA_1\",\"portCA_1\",\"portCA_2\"]):\n super().__init__()\n self.num_bits=num_bits\n self.node=node\n self.processor=processor\n self.portNameQ1=port_names[0]\n self.portNameC1=port_names[1]\n self.portNameC2=port_names[2]\n self.EPRList=None\n self.basisList=Random_basis_gen(self.num_bits)\n self.loc_measRes=[]\n self.key=None\n self.sourceQList=[]\n self.sourceFreq=sourceFreq\n \n #generat qubits from source\n self.A_Source = QSource(\"Alice_source\"\n ,status=SourceStatus.EXTERNAL) # enable frequency\n self.A_Source.ports[\"qout0\"].bind_output_handler(self.storeSourceOutput)\n \n # =======================================A run ============================\n def run(self):\n\n \n # A generat qubits\n self.A_genQubits(self.num_bits,self.sourceFreq)\n \n # wait\n yield self.await_program(processor=self.processor)\n \n \n #yield self.await_program(processor=self.processor)\n # send qubits\n self.A_sendEPR()\n \n\n #self.A_measure()\n self.myQG_A_measure=QG_A_measure(\n basisList=self.basisList,num_bits=self.num_bits)\n self.processor.execute_program(\n self.myQG_A_measure,qubit_mapping=[i for i in range(0, 2*self.num_bits)])\n \n yield self.await_program(processor=self.processor)\n\n\n # receive B basis\n port=self.node.ports[self.portNameC1]\n yield self.await_port_input(port)\n basis_B = port.rx_input().items\n \n \n \n\n\n # get A meas\n #self.processor.set_program_done_callback(self.A_getPGoutput,once=True)\n for i in range(2*self.num_bits):\n if i%2 == 0: # only even slot\n tmp=self.myQG_A_measure.output[str(i)][0]\n self.loc_measRes.append(tmp)\n \n # send A basis to B\n self.node.ports[self.portNameC2].tx_output(self.basisList)\n \n \n # compare basis\n self.loc_measRes=Compare_basis(self.basisList,basis_B,self.loc_measRes)\n \n self.key=''.join(map(str, self.loc_measRes))\n #print(\"A key:\",self.key)\n\n def storeSourceOutput(self,qubit):\n self.sourceQList.append(qubit.items[0])\n if len(self.sourceQList)==2*self.num_bits:\n self.processor.put(qubits=self.sourceQList)\n \n # apply H detector\n PG_qPrepare=QG_A_qPrepare(num_bits=self.num_bits)\n self.processor.execute_program(\n PG_qPrepare,qubit_mapping=[i for i in range(0, 2*self.num_bits)])\n\n\n def A_genQubits(self,num,freq=8e7):\n \n \n #set clock\n clock = Clock(\"clock\", frequency=freq, max_ticks=2*num)\n try:\n clock.ports[\"cout\"].connect(self.A_Source.ports[\"trigger\"])\n except:\n pass\n #print(\"alread connected\") \n \n clock.start()\n \n \n def A_sendEPR(self):\n #print(\"A_sendEPR\")\n inx=list(range(1,2*self.num_bits+1,2))\n payload=self.processor.pop(inx)\n self.node.ports[self.portNameQ1].tx_output(payload)\n \n","repo_name":"LiaoChinTe/netsquid-simulation","sub_path":"QKD/E91/E91_Alice.py","file_name":"E91_Alice.py","file_ext":"py","file_size_in_byte":5000,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"79"} +{"seq_id":"71413356734","text":"import os\nimport time\n\ndef clears():\n\tif os.name == \"posix\":\n\t\tos.system(\"clear\")\n\telse:\n\t\tos.system(\"cls\")\n\ntry:\n\tfrom colorama import init, Fore, Back, Style\nexcept ModuleNotFoundError:\n\tos.system(\"pip install colorama --progress-bar off\")\n\tfrom colorama import init, Fore, Back, Style\ntry:\n\timport speedtest\nexcept ModuleNotFoundError:\n\tos.system(\"pip install speedtest-cli --progress-bar off\")\n\timport speedtest\nclears()\ninit(autoreset=True)\n\ndef listsum(numList):\n if len(numList) == 1:\n return numList[0]\n else:\n return numList[0] + listsum(numList[1:])\n\ndef starttest(n):\n\tsd, su, sp = [], [], []\n\tfor x in range(n):\n\t\tx = x+1\n\t\tprint(Fore.YELLOW + f\"Testing {x}...\")\n\t\ttry:\n\t\t\tst = speedtest.Speedtest()\n\t\t\tdownload = st.download()\n\t\t\tdown = round(download/(1024*1024), 3)\n\t\t\tdow = download/(1024*1024)\n\t\t\tsd.append(down)\n\t\texcept speedtest.ConfigRetrievalError:\n\t\t\tprint(Fore.RED + \"Error: No Internet connection!\")\n\t\t\treturn\n\t\tclears()\n\t\tif down < 10:\n\t\t\tprint(Fore.RED + f\"\tDownload: {down} mb/s\")\n\t\telif down < 25:\n\t\t\tprint(Fore.YELLOW + f\"\tDownload: {down} mb/s\")\n\t\telif down < 45:\n\t\t\tprint(Fore.GREEN + f\"\tDownload: {down} mb/s\")\n\t\telse:\n\t\t\tprint(Fore.GREEN + Style. BRIGHT + f\"\tDownload: {down} mb/s\")\n\t\ttry:\n\t\t\tupload = st.upload()\n\t\t\tupl = round(upload/(1024*1024), 3)\n\t\t\tup = upload/(1024*1024)\n\t\t\tsu.append(up)\n\t\texcept speedtest.ConfigRetrievalError:\n\t\t\tprint(Fore.RED + \"Error: No Internet connection!\")\n\t\t\treturn\n\t\tif upl < 8:\n\t\t\tprint(Fore.RED + f\"\tUpload: {upl} mb/s\")\n\t\telif upl < 20:\n\t\t\tprint(Fore.YELLOW + f\"\tUpload: {upl} mb/s\")\n\t\telif upl < 40:\n\t\t\tprint(Fore.GREEN + f\"\tUpload: {upl} mb/s\")\n\t\telse:\n\t\t\tprint(Fore.GREEN + Style. BRIGHT + f\"\tUpload: {upl} mb/s\")\n\t\ttry:\n\t\t\tservernames = \"\"\n\t\t\tst.get_servers(servernames)\n\t\t\tping = st.results.ping\n\t\t\tsp.append(ping)\n\t\texcept speedtest.ConfigRetrievalError:\n\t\t\tprint(Fore.RED + \"Error: No Internet connection!\")\n\t\t\treturn\n\t\tprint(Fore.GREEN + f\"\tPing: {ping} ms\")\n\t\tif ping > 0:\n\t\t\tprint(Fore.GREEN + \"\tCommentar: Good ping!\")\n\t\telif ping > 110:\n\t\t\tprint(Fore.YELLOW + \"\tComentar: min. ping!\")\n\t\telse:\n\t\t\tprint(Fore.RED + Back.GREEN + \"\tCommentar: ping > 300 ms!\")\n\t\tif x == n:\n\t\t\ttime.sleep(5)\n\t\t\tclears()\n\t\t\tnsd = round((listsum(sd))/n, 3)\n\t\t\tif nsd < 10:\n\t\t\t\tprint(Fore.RED + f\"\tAverage download speed: {nsd} mb/s\")\n\t\t\telif nsd < 25:\n\t\t\t\tprint(Fore.YELLOW + f\"\tAverage download speed: {nsd} mb/s\")\n\t\t\telif nsd < 45:\n\t\t\t\tprint(Fore.GREEN + f\"\tAverage download speed: {nsd} mb/s\")\n\t\t\telse:\n\t\t\t\tprint(Fore.GREEN + Style. BRIGHT + f\"\tAverage download speed: {nsd} mb/s\")\n\t\t\tnsu = round((listsum(su))/n, 3)\n\t\t\tif nsu < 8:\n\t\t\t\tprint(Fore.RED + f\"\tAverage upload speed: {nsu} mb/s\")\n\t\t\telif nsu < 20:\n\t\t\t\tprint(Fore.YELLOW + f\"\tAverage upload speed: {nsu} mb/s\")\n\t\t\telif nsu < 40:\n\t\t\t\tprint(Fore.GREEN + f\"\tAverage upload speed: {nsu} mb/s\")\n\t\t\telse:\n\t\t\t\tprint(Fore.GREEN + Style. BRIGHT + f\"\tAverage upload speed: {nsu} mb/s\")\n\t\t\tnsp = round((listsum(sp))/n, 3)\n\t\t\tif nsp < 50:\n\t\t\t\tprint(Fore.GREEN + Style. BRIGHT + f\"\tAverage ping: {nsp} ms\")\n\t\t\telif nsp < 100:\n\t\t\t\tprint(Fore.GREEN + f\"\tAverage ping: {nsp} ms\")\n\t\t\telif nsp < 130:\n\t\t\t\tprint(Fore.YELLOW + f\"\tAverage ping: {nsp} ms\")\n\t\t\telse:\n\t\t\t\tprint(Fore.RED + f\"\tAverage ping: {nsp} ms\")\n\ndef main():\n\twhile True:\n\t\tUS = input(Fore.YELLOW + \"Start/Restart? (Y/N/About): \")\n\t\ttry:\n\t\t\tif US.lower() == \"y\":\n\t\t\t\tne = input(Fore.BLUE + \"Enter the number of tests: \")\n\t\t\t\tn = int(ne)\n\t\t\t\tif n < 1:\n\t\t\t\t\tprint(Fore.RED + \"You cannot test less than once!\")\n\t\t\t\tif n > 10:\n\t\t\t\t\tprint(Fore.RED + \"You cannot test more than ten times!\")\n\t\t\t\telse:\n\t\t\t\t\tstarttest(n)\n\t\t\telif US.lower() == \"n\":\n\t\t\t\tbreak\n\t\t\t\tprint(Fore.GREEN + \"Completing the test... (7s)\")\n\t\t\t\ttime.sleep(7)\n\t\t\t\tos.system(\"exit\")\n\t\t\telif US.lower() == \"about\":\n\t\t\t\tprint(Fore.RED + Back.GREEN + \"Telegram code creator: @DepyXa.\\nDonatello website: http://donatello.to/DepyXa\")\n\t\texcept ValueError:\n\t\t\tprint(Fore.GREEN + \"A letter is entered in place of numbers, enter the number...\")\n\t\t\tpass\n\nmain()\n","repo_name":"DepyXa/SpeedTestS","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4003,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"74612078335","text":"import random\n\n\ndef jogar():\n print(\"**********************************\")\n print(\"Bem-vindo no jogo de Adivinhação!\")\n print(\"**********************************\")\n\n numero_secreto = random.randrange(1, 101)\n total_tentativas = 0\n pontos = 1000\n # print(numero_secreto)\n\n print(\"Qual nivel de dificuldade você quer?\")\n print(\"(1) Facil, (2) Médio, (3) Difícil\")\n nivel = int(input(\"Defina o nível: \"))\n\n if nivel == 1:\n total_tentativas = 20\n elif nivel == 2:\n total_tentativas = 10\n else:\n total_tentativas = 5\n\n for rodada in range(1, total_tentativas + 1):\n print(\"Tentativa: {} de {}\".format(rodada, total_tentativas))\n chute = int(input(\"digite o numero: \"))\n print(\"Você digitou\", chute)\n\n acertou = numero_secreto == chute\n maior = chute > numero_secreto\n menor = chute < numero_secreto\n if chute < 1 or chute > 100:\n print(\"Você deve digitar um valor entre 1 e 100\")\n continue\n\n if acertou:\n print(\"Você acertou e fez {} pontos\".format(pontos))\n break\n else:\n if maior:\n print(\"Você digitou um numero maior\")\n elif menor:\n print(\"Você digitou um numero menor\")\n pontos = abs(pontos - chute)\n\n rodada = rodada + 1\n print(\"Fim de jogo\")\n\n\nif __name__ == \"__main__\":\n jogar()\n","repo_name":"anisiabarros/jogoForca","sub_path":"adivinhacao.py","file_name":"adivinhacao.py","file_ext":"py","file_size_in_byte":1438,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"33744793313","text":"from mxnet import gluon\nfrom mxnet import ndarray as nd\nfrom mxnet import autograd\nimport matplotlib.pyplot as plt\nimport sys\nimport random\n\ndef transform(data, label):\n return data.astype('float32')/255, label.astype('float32')\n\nminist_train = gluon.data.vision.FashionMNIST(train=True, transform=transform)\nminist_test = gluon.data.vision.FashionMNIST(train=False, transform=transform)\n\ndef show_image(images):\n n = images.shape[0] # nums of images\n _, figs = plt.subplots(1, n, figsize=(15, 15))\n for i in range(n):\n figs[i].imshow(images[i].reshape((28, 28)).asnumpy())\n figs[i].axes.get_xaxis().set_visible(False)\n figs[i].axes.get_yaxis().set_visible(False)\n plt.show()\n\ndef get_text_labels(label):\n text_labels = [\n 't-shirt', 'trouser', 'pullover', 'dress,', 'coat',\n 'sandal', 'shirt', 'sneaker', 'bag', 'ankle boot'\n ]\n return [text_labels[int(i)] for i in label]\n\n# data, label = minist_train[0:9]\n# show_image(data)\n# print(get_text_labels(label))\n\n# load data\nbatch_size = 256\ntrain_data = gluon.data.DataLoader(minist_train, batch_size, shuffle = True)\ntest_data = gluon.data.DataLoader(minist_test, batch_size, shuffle = False)\n\n# initialize parameters\nnum_inputs = 784 # 28*28 num of features\nnum_outputs = 10 # num of class\n\nW = nd.random_normal(shape=(num_inputs, num_outputs))\nb = nd.random_normal(shape=num_outputs) # row vec\n\nparams = [W, b]\n\nfor param in params:\n param.attach_grad()\n\ndef softmax(X):\n exp = nd.exp(X)\n partition = exp.sum(axis = 1, keepdims=True) # return (nrows, 1) matrix\n return exp / partition\n\ndef net(X):\n return softmax(nd.dot(X.reshape((-1, num_inputs)), W) + b)\n # the second dim is num_inputs, the first dim will be automatically calculated\n\ndef cross_entropy(yhat, y):\n return - nd.pick(nd.log(yhat), y)\n\ndef accurancy(output, label):\n return nd.mean(output.argmax(axis=1) == label).asscalar() # output: batch_size * nums_classes\n# output.argmax(axis=1) 返回一个 1*10 矩阵,每个元素代表output每行最大值的列号\ndef evaluate_accurancy(data_iterator, net):\n acc = 0\n for data, label in data_iterator:\n output = net(data)\n acc += accurancy(output, label)\n #print(\"acc: %f\" % acc)\n return acc / len(data_iterator)\n\ndef SGD(params, lr):\n for param in params:\n param[:] = param - lr * param.grad;\n\n\nlearning_rate = 0.05\nfor epoch in range(10):\n train_loss = 0\n train_acc = 0\n for data, label in train_data:\n with autograd.record():\n output = net(data)\n loss = cross_entropy(output, label)\n loss.backward()\n SGD(params, learning_rate/batch_size) #将梯度做平均,这样学习率对batch_size不会那么敏感\n\n train_loss += nd.mean(loss).asscalar()\n train_acc += accurancy(output, label)\n\n test_acc = evaluate_accurancy(test_data, net)\n print(\"Epoch %d. Loss: %f, Train acc %f, Test acc %f\" % (epoch, train_loss/len(train_data), train_acc/len(train_data),\n test_acc)\n )\n\ndata, label = minist_test[0:9]\nshow_image(data)\nprint('true labels')\nprint(get_text_labels(label))\n\npredicted_labels = net(data).argmax(axis=1)\nprint('predicted labels')\nprint(get_text_labels(predicted_labels.asnumpy()))","repo_name":"gonglixue/PRML_Python","sub_path":"MXNET/SoftmaxRegression.py","file_name":"SoftmaxRegression.py","file_ext":"py","file_size_in_byte":3324,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"43183976414","text":"import pygame as pg\nimport Code.pyshooterClient as pyshooterClient\nimport Libs.textbox.TextBox as TextBox\nimport Code.constants as constants\n\nKEY_REPEAT_SETTING = (200, 70)\n\n\nclass Menu():\n def __init__(self):\n #load menu images\n self.MENU_IMAGE = pg.image.load(\"Assets/Images/menu/background.png\")\n self.MENU_IMAGE = pg.transform.scale(self.MENU_IMAGE, (800, 600))\n self.SINGLE_OFF = pg.image.load(\"Assets/Images/menu/button_singleplayer_off.png\")\n self.SINGLE_OFF = pg.transform.scale(self.SINGLE_OFF, (200, 50))\n self.SINGLE_ON = pg.image.load(\"Assets/Images/menu/button_singleplayer_on.png\")\n self.SINGLE_ON = pg.transform.scale(self.SINGLE_ON, (200, 50))\n self.MULTI_OFF = pg.image.load(\"Assets/Images/menu/button_multiplayer_off.png\")\n self.MULTI_OFF = pg.transform.scale(self.MULTI_OFF, (200, 50))\n self.MULTI_ON = pg.image.load(\"Assets/Images/menu/button_multiplayer_on.png\")\n self.MULTI_ON = pg.transform.scale(self.MULTI_ON, (200, 50))\n self.SERVER_OFF = pg.image.load(\"Assets/Images/menu/button_create_server_off.png\")\n self.SERVER_OFF = pg.transform.scale(self.SERVER_OFF, (200, 50))\n self.SERVER_ON = pg.image.load(\"Assets/Images/menu/button_create_server_on.png\")\n self.SERVER_ON = pg.transform.scale(self.SERVER_ON, (200, 50))\n self.CONNECT_OFF = pg.image.load(\"Assets/Images/menu/button_connect_off.png\")\n self.CONNECT_OFF = pg.transform.scale(self.CONNECT_OFF, (200, 50))\n self.CONNECT_ON = pg.image.load(\"Assets/Images/menu/button_connect_on.png\")\n self.CONNECT_ON = pg.transform.scale(self.CONNECT_ON, (200, 50))\n\n #### new menu\n self.is_selected = False\n self.switch_sound = pg.mixer.Sound('Assets/Sounds/switch_sound.wav')\n self.switch_sound.set_volume(1)\n self.options_width = 300 #background options width\n self.spacing = 50 #space between the center of two options\n self.first_option_position = constants.SCREEN_SIZE[0]/2, constants.SCREEN_SIZE[1]/2 #center position of first menu option\n self.font_text = pg.font.Font(\"Assets/Fonts/BebasNeue-Regular.otf\", 35)\n \n ### single player\n self.single_off = self.font_text.render(\"SINGLEPLAYER\", 1, constants.WHITE)\n self.rect_single = self.single_off.get_rect()\n \n self.surface_single_off = pg.Surface((self.options_width, self.rect_single.height))\n self.surface_single_rect = self.surface_single_off.get_rect(center=self.first_option_position)\n crop_area = self.surface_single_rect.topleft[0], self.surface_single_rect[1], self.surface_single_rect.width, self.surface_single_rect.height\n self.surface_single_rect.topleft = (0,0)\n center_of_surface = self.surface_single_rect.center \n self.surface_single_off.blit(self.MENU_IMAGE, (0,0), crop_area)\n self.rect_single.center = center_of_surface\n self.surface_single_off.blit(self.single_off, self.rect_single)\n \n self.surface_single_on = pg.Surface((self.options_width, self.rect_single.height))\n self.surface_single_rect.topleft = (0,0)\n center_of_surface = self.surface_single_rect.center \n self.rect_single.center = center_of_surface\n self.surface_single_on.fill(constants.BLACK)\n self.surface_single_on.blit(self.single_off, self.rect_single)\n \n self.surface_single_rect.center = self.first_option_position\n\n ###################################################\n ## multiplayer\n self.multi_off = self.font_text.render(\"MULTIPLAYER\", 1, constants.WHITE)\n self.rect_multi = self.multi_off.get_rect()\n \n self.surface_multi_off = pg.Surface((self.options_width, self.rect_multi.height))\n self.surface_multi_rect = self.surface_multi_off.get_rect(center=(self.first_option_position[0], self.first_option_position[1]+self.spacing))\n crop_area = self.surface_multi_rect.topleft[0], self.surface_multi_rect[1], self.surface_multi_rect.width, self.surface_multi_rect.height\n self.surface_multi_rect.topleft = (0,0)\n center_of_surface = self.surface_multi_rect.center \n self.surface_multi_off.blit(self.MENU_IMAGE, (0,0), crop_area)\n self.rect_multi.center = center_of_surface\n self.surface_multi_off.blit(self.multi_off, self.rect_multi)\n\n self.surface_multi_on = pg.Surface((self.options_width, self.rect_multi.height))\n self.surface_multi_rect.topleft = (0,0)\n center_of_surface = self.surface_multi_rect.center \n self.rect_multi.center = center_of_surface\n self.surface_multi_on.fill(constants.BLACK)\n self.surface_multi_on.blit(self.multi_off, self.rect_multi)\n \n self.surface_multi_rect.center = (self.first_option_position[0], self.first_option_position[1]+self.spacing)\n\n self.mouse_rect = pg.Rect(0,0,3,3)\n ####\n\n def on_init(self):\n #FlowCrontoller inital value\n self.menu_state= \"singleplayer/multiplayer\"\n self._in_menu = True\n\n #Multiplayer Variables\n self.name = None\n self.server_ip = None\n self.have_client = False\n self.is_host = False\n\n #surface\n self.surface = pg.display.get_surface()\n self.color = (100, 100, 100)\n self.surface.blit(self.MENU_IMAGE, (0, 0))\n\n #input text configuration\n pg.key.set_repeat(*KEY_REPEAT_SETTING)\n\n #music configuration\n self.music = pg.mixer.Sound('Assets/Sounds/menu_sound_loop.wav')\n self.music.set_volume(0.5)\n self.music.play(-1)\n\n def singleplayer_multiplayer_display(self, event):\n self.singleplayer_multiplayer_interactive(event)\n\n def multiplayer_get_name(self, event):\n self.input.update()\n self.input.draw(self.surface)\n prompt = self.make_prompt('Type your name :')\n self.surface.blit(*prompt)\n self.input.get_event(event)\n\n def multiplayer_create_connect(self, event):\n self.multiplayer_create_connect_interactive(event)\n\n def multiplayer_connect_get_ip(self, event):\n self.input.update()\n self.input.draw(self.surface)\n prompt = self.make_prompt('Type server ip :')\n self.surface.blit(*prompt)\n self.input.get_event(event)\n\n def on_render(self):\n pg.display.update()\n\n def on_event(self, event):\n if event.type == pg.QUIT or (event.type == pg.KEYDOWN and event.key == pg.K_ESCAPE):\n self._in_menu = False\n self._in_menu_multiplayer = False\n\n def singleplayer_multiplayer_interactive(self, event):\n mouse = pg.mouse.get_pos()\n self.mouse_rect.center = mouse \n \n single_is_selected = self.mouse_rect.colliderect(self.surface_single_rect)\n multi_is_selected = self.mouse_rect.colliderect(self.surface_multi_rect)\n if single_is_selected or multi_is_selected:\n if not self.is_selected:\n self.is_selected = True\n self.switch_sound.play()\n else:\n self.is_selected = False\n\n if single_is_selected:\n self.surface.blit(self.surface_single_on, self.surface_single_rect)\n if event.type == pg.MOUSEBUTTONDOWN:\n self._in_menu = False\n else:\n self.surface.blit(self.surface_single_off, self.surface_single_rect)\n\n if multi_is_selected:\n self.surface.blit(self.surface_multi_on, self.surface_multi_rect)\n if event.type == pg.MOUSEBUTTONDOWN:\n self.input = TextBox.TextBox((300, 300, 200, 30), \n command=self.change_name,\n clear_on_enter=True,\n inactive_on_enter=False)\n self.surface.blit(self.MENU_IMAGE, (0, 0))\n self.menu_state = \"multiplayer_get_name\"\n else:\n self.surface.blit(self.surface_multi_off, self.surface_multi_rect)\n\n def multiplayer_create_connect_interactive(self, event):\n mouse = pg.mouse.get_pos()\n\n if 150 + 200 > mouse[0] > 150 and 450 + 50 > mouse[1] > 450:\n self.surface.blit(self.SERVER_ON, (150, 450))\n if event.type == pg.MOUSEBUTTONDOWN:\n self.server_client = pyshooterClient.pyshooterClient(self.name)\n self.server_client.start()\n self.have_client = True\n self.is_host = True\n self._in_menu = False\n else:\n self.surface.blit(self.SERVER_OFF, (150, 450))\n\n if 450 + 200 > mouse[0] > 450 and 450 + 50 > mouse[1] > 450:\n self.surface.blit(self.CONNECT_ON, (450, 450))\n if event.type == pg.MOUSEBUTTONDOWN:\n self.input = TextBox.TextBox((300, 300, 200, 30), command=self.change_ip,\n clear_on_enter=True, inactive_on_enter=False)\n self.menu_state = \"multiplayer_connect_get_ip\"\n self.surface.blit(self.MENU_IMAGE, (0, 0))\n else:\n self.surface.blit(self.CONNECT_OFF, (450, 450))\n\n def change_name(self, id, name):\n self.name = str(name)\n self.menu_state = \"multiplayer_create/connect\"\n self.surface.blit(self.MENU_IMAGE, (0, 0))\n\n def change_ip(self,id,ip):\n self.server_ip = str(ip)\n self.menu_state = \"connect\"\n self.surface.blit(self.MENU_IMAGE, (0, 0))\n\n def connect(self):\n self.server_client = pyshooterClient.pyshooterClient(self.name)\n self.have_client = self.server_client.start_connect(self.server_ip)\n self._in_menu = False\n\n\n def make_prompt(self, message):\n # font = pg.font.SysFont(\"arial\", 20)\n font = pg.font.Font(\"Assets/Fonts/BebasNeue-Regular.otf\", 20)\n rend = font.render(message, True, pg.Color(\"black\"))\n return (rend, rend.get_rect(topleft=(270, 270)))\n\n\n def intro(self):\n #music and other variables\n self.on_init()\n\n #while in menu select correct display case\n while(self._in_menu):\n for event in pg.event.get():\n self.on_event(event)\n if self.menu_state == \"singleplayer/multiplayer\":\n self.singleplayer_multiplayer_display(event)\n elif self.menu_state == \"multiplayer_get_name\":\n self.multiplayer_get_name(event)\n elif self.menu_state == \"multiplayer_create/connect\":\n self.multiplayer_create_connect(event)\n elif self.menu_state == \"multiplayer_connect_get_ip\":\n self.multiplayer_connect_get_ip(event)\n elif self.menu_state == \"connect\":\n self.connect()\n self.on_render()\n\n self.music.stop()\n","repo_name":"Duhamb/pyshooter","sub_path":"Code/Menu.py","file_name":"Menu.py","file_ext":"py","file_size_in_byte":10851,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"74569907774","text":"from abc import ABC, abstractmethod\n\nimport numpy as np\n\nimport rospy\nfrom sensor_msgs.msg import NavSatFix\nfrom geometry_msgs.msg import Pose as ROSPose\n\nfrom pose import Pose\nfrom trajectory import Trajectory\nfrom controller import Controller\nfrom world import World\n\n\nclass StanleyController(Controller):\n \"\"\"\n Stanley Controller\n Calculations based off FRONT axle of Buggy\n \"\"\"\n\n LOOK_AHEAD_DIST_CONST = 0.1\n MIN_LOOK_AHEAD_DIST = 0.1\n MAX_LOOK_AHEAD_DIST = 2\n\n CROSS_TRACK_GAIN = 1\n HEADING_GAIN = 0.75\n\n def __init__(self, buggy_name, start_index=0) -> None:\n super(StanleyController, self).__init__(start_index, buggy_name)\n self.debug_reference_pos_publisher = rospy.Publisher(\n buggy_name + \"/auton/debug/reference_navsat\", NavSatFix, queue_size=1\n )\n self.debug_error_publisher = rospy.Publisher(\n buggy_name + \"/auton/debug/error\", ROSPose, queue_size=1\n )\n\n def compute_control(\n self, current_pose: Pose, trajectory: Trajectory, current_speed: float\n ):\n \"\"\"Computes the steering angle necessary for stanley controller.\n Does this by looking at the crosstrack error + heading error\n\n Args:\n current_pose (Pose): current pose (x, y, theta) (UTM coordinates)\n trajectory (Trajectory): reference trajectory\n current_speed (float): current speed of the buggy\n\n Returns:\n float (desired steering angle)\n \"\"\"\n if self.current_traj_index >= trajectory.get_num_points() - 1:\n return 0\n\n heading = current_pose.theta # in radians\n x = current_pose.x\n y = current_pose.y\n\n # Assume current pose is rear of buggy, project it to center of front axle\n # 10 is a good number to search forward along the index\n front_x = x + StanleyController.WHEELBASE * np.cos(heading)\n front_y = y + StanleyController.WHEELBASE * np.sin(heading)\n\n traj_index = trajectory.get_closest_index_on_path(\n front_x,\n front_y,\n start_index=self.current_traj_index,\n end_index=self.current_traj_index + 10,\n )\n self.current_traj_index = max(traj_index, self.current_traj_index)\n\n # Calculate heading error\n\n lookahead_dist = np.clip(\n self.LOOK_AHEAD_DIST_CONST * current_speed,\n self.MIN_LOOK_AHEAD_DIST,\n self.MAX_LOOK_AHEAD_DIST,\n )\n traj_dist = (\n trajectory.get_distance_from_index(self.current_traj_index) + lookahead_dist\n )\n ref_heading = trajectory.get_heading_by_index(\n trajectory.get_index_from_distance(traj_dist)\n )\n error_heading = ref_heading - current_pose.theta\n error_heading = (\n np.arctan2(np.sin(error_heading), np.cos(error_heading))\n * StanleyController.HEADING_GAIN\n )\n\n # Calculate cross track error by finding the distance from the buggy to the tangent line of\n # the reference trajectory\n closest_position = trajectory.get_position_by_index(self.current_traj_index)\n next_position = trajectory.get_position_by_index(\n self.current_traj_index + 0.0001\n )\n x1 = closest_position[0]\n y1 = closest_position[1]\n x2 = next_position[0]\n y2 = next_position[1]\n error_dist = -((x - x1) * (y2 - y1) - (y - y1) * (x2 - x1)) / np.sqrt(\n (y2 - y1) ** 2 + (x2 - x1) ** 2\n )\n\n speed = current_speed\n if current_speed < 1:\n speed = 1\n\n cross_track_error = -np.arctan2(\n StanleyController.CROSS_TRACK_GAIN * error_dist, speed\n )\n\n steering_cmd = error_heading + cross_track_error\n steering_cmd = np.clip(steering_cmd, -np.pi / 9, np.pi / 9)\n\n reference_position = trajectory.get_position_by_index(self.current_traj_index)\n reference_error = current_pose.convert_point_from_global_to_local_frame(\n reference_position\n )\n reference_error -= np.array([StanleyController.WHEELBASE, 0])\n\n # Publish error for debugging\n error_pose = ROSPose()\n error_pose.position.x = reference_error[0]\n error_pose.position.y = reference_error[1]\n self.debug_error_publisher.publish(error_pose)\n\n # Publish reference position for debugging\n reference_navsat = NavSatFix()\n ref_gps = World.world_to_gps(*closest_position)\n reference_navsat.latitude = ref_gps[0]\n reference_navsat.longitude = ref_gps[1]\n self.debug_reference_pos_publisher.publish(reference_navsat)\n\n return steering_cmd\n","repo_name":"CMU-Robotics-Club/RoboBuggy2","sub_path":"rb_ws/src/buggy/scripts/auton/stanley_controller.py","file_name":"stanley_controller.py","file_ext":"py","file_size_in_byte":4696,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"36303262999","text":"#!/usr/bin/env python\nimport rospy\nfrom sensor_msgs.msg import Image\nimport cv2\nfrom cv_bridge import CvBridge, CvBridgeError\n\nbridge = CvBridge()\npub = rospy.Publisher('/depth', Image, queue_size=10)\n\n\ndef callback1(msg):\n global flag1\n global imgL\n imgL = bridge.imgmsg_to_cv2(msg, desired_encoding='bgr8')\n imgL = cv2.cvtColor(imgL, cv2.COLOR_BGR2GRAY)\n flag1 = True\n\n\ndef callback2(msg):\n global flag2\n global imgR\n imgR = bridge.imgmsg_to_cv2(msg, desired_encoding='bgr8')\n imgR = cv2.cvtColor(imgR, cv2.COLOR_BGR2GRAY)\n flag2 = True\n\n\ndef talker():\n global imgL\n global imgR\n global flag1\n global flag2\n flag1 = False\n flag2 = False\n rospy.init_node('depthmap', anonymous=True)\n rospy.Subscriber('contour_right', Image, callback1)\n rospy.Subscriber('contour_left', Image, callback2)\n rate = rospy.Rate(10)\n while not rospy.is_shutdown():\n if flag1 and flag2:\n stereo = cv2.StereoBM_create(numDisparities=16,\n blockSize=15)\n disparity = stereo.compute(imgL, imgR)\n depth_img = bridge.cv2_to_imgmsg(disparity, encoding=\"16SC1\")\n pub.publish(depth_img)\n rate.sleep()\n\n\nif __name__ == '__main__':\n talker()\n","repo_name":"KeerthivasanIITMadras/viraj_simualtion_task","sub_path":"src/my_simulations/scripts/depthmap.py","file_name":"depthmap.py","file_ext":"py","file_size_in_byte":1276,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"24810714130","text":"import django_filters\nfrom rest_framework import viewsets, permissions, status, filters\nfrom django.db import transaction as trn\nfrom rest_framework.response import Response\n\nfrom transactions.models import Transaction\nfrom transactions.serializers import TransactionSerializer\nfrom user.models import User\n\n\nclass IsAuthorOrIsAuthenticated(permissions.BasePermission):\n\n def has_object_permission(self, request, view, obj):\n return obj.user == request.user\n\n\nclass TransactionsViewSet(viewsets.ModelViewSet):\n permission_classes = (IsAuthorOrIsAuthenticated, )\n serializer_class = TransactionSerializer\n filter_backends = [django_filters.rest_framework.DjangoFilterBackend, filters.OrderingFilter]\n filterset_fields = ['date_done', 'sum']\n ordering_fields = ['date_done', 'sum']\n\n def get_queryset(self):\n return Transaction.objects.filter(user__email=self.request.user.email)\n\n def create(self, request, *args, **kwargs):\n serializer = TransactionSerializer(data=request.data)\n if serializer.is_valid():\n with trn.atomic():\n transaction = serializer.save(user=request.user)\n user = User.objects.get(email=request.user.email)\n user.balance += transaction.sum\n user.save()\n return Response(data=serializer.data)\n else:\n return Response(status=status.HTTP_400_BAD_REQUEST)\n","repo_name":"Ultralightmemes/Expense-Manager","sub_path":"transactions/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1424,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"35551709280","text":"# *****************************************************************************\n# *****************************************************************************\n#\n#\t\tName:\t\tmakepgm.py\n#\t\tAuthor:\t\tPaul Robson (paul@robsons.org.uk)\n#\t\tDate:\t\t26th February 2021\n#\t\tPurpose:\tTokenises complete program.\n#\n# *****************************************************************************\n# *****************************************************************************\n\nimport os,re,sys\nfrom tokeniser import *\nfrom tokens import *\n\n# *****************************************************************************\n#\n#\t\t\t\t\t\t\tBuild a complete program\n#\n# *****************************************************************************\n\nclass ProgramBuilder(object):\n\tdef __init__(self):\n\t\tself.tokens = Tokens()\t\t\t\t\t\t\t\t\t\n\t\tself.tokeniser = Tokeniser()\n\t\tself.code = [] \t\t\t\t\t\t\t\t\t\t\t\n\t\tself.nextLineNumber = 1000\n\t#\n\t#\t\tConvert line number + line text to byte sequence.\n\t#\n\tdef makeLine(self,lineNumber,lineText):\n\t\tcode = self.tokeniser.tokenise(lineText)\n\t\tcode = [ lineNumber & 0xFF,lineNumber >> 8] + code + [ self.tokens.getEOLToken() ]\n\t\tcode.insert(0,len(code)+1)\n\t\treturn code\n\t#\n\t#\t\tAppend line number/program code. Line number optional\n\t#\n\tdef add(self,lineNumber,lineText = None):\n\t\tif lineText is None:\n\t\t\tlineText = lineNumber \t\t\t\t\t\t\t\t\t\t\t# take advantage of typing :)\n\t\t\tlineNumber = self.nextLineNumber\n\t\telse:\n\t\t\tassert lineNumber >= self.nextLineNumber,\"Line number sequence\"\n\t\tself.nextLineNumber = lineNumber+10\t\n\t\t#print(lineNumber,lineText)\n\t\tself.code += self.makeLine(lineNumber,lineText)\n\t#\n\t#\t\tLoad source file.\n\t#\n\tdef load(self,srcFile):\n\t\tfor l in open(srcFile).readlines():\n\t\t\tif not l.strip().startswith(\";\"):\n\t\t\t\tm = re.match(\"^(\\\\d+)(.*)$\",l)\n\t\t\t\tif m is not None:\n\t\t\t\t\tself.add(int(m.group(1)),m.group(2).strip())\n\t\t\t\telse:\n\t\t\t\t\tif l.strip() != \"\":\n\t\t\t\t\t\tself.add(l.strip())\n\t#\n\t#\t\tExport as assembly include file\n\t#\n\tdef exportAsm(self,tgtFile):\n\t\th = open(tgtFile,\"w\")\n\t\theader= \";\\n;\\tAutomatically generated\\n;\\n\"\t\n\t\th.write(header)\n\t\th.write(\"\\t.byte {0}\\n\\n\".format(\",\".join([str(x) for x in self.code+[0]])))\n\t\th.close()\n\t#\n\t#\t\tExport binary dump\n\t#\n\tdef exportBin(self,tgtFile):\n\t\th = open(tgtFile,\"wb\")\n\t\th.write(bytes(self.code+[0]))\n\t\th.close()\n\nif __name__ == \"__main__\":\n\tpb = ProgramBuilder()\n\tfor f in sys.argv[1:]:\n\t\tpb.load(f)\n\t\tpb.exportAsm(\"../source/generated/testcode.inc\".replace(\"/\",os.sep))\n\t\tpb.exportBin(\"../source/generated/testcode.bin\".replace(\"/\",os.sep))\n","repo_name":"paulscottrobson/6502-basic","sub_path":"scripts/build/makepgm.py","file_name":"makepgm.py","file_ext":"py","file_size_in_byte":2501,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"22138855398","text":"from datetime import datetime, timedelta\nfrom itertools import groupby\nimport json\n\nfrom odoo import api, fields, models, SUPERUSER_ID, _\nfrom odoo.exceptions import AccessError, UserError, ValidationError\nfrom odoo.osv import expression\nfrom odoo.tools import float_is_zero, html_keep_url, is_html_empty\nfrom odoo.addons.payment import utils as payment_utils\n\n\nclass realestateorder(models.Model):\n _name = \"real_estate.order\"\n _inherit = \"mail.thread\", \"mail.activity.mixin\", \"mail.composer.mixin\"\n _description = \"Real Estate Order\"\n _order = \"id desc\"\n\n name = fields.Char(string='Title', required=True)\n description = fields.Text(string='Description', required=False)\n postcode = fields.Char(string='Postcode', required=False)\n date_availability = fields.Date(string='Available Date', optional='hide')\n expected_price = fields.Float(string='Expected Price', store=True, required=False)\n selling_price = fields.Float(string='Selling Price', store=True)\n bedrooms = fields.Integer(string='Bedrooms')\n living_area = fields.Integer(string='Living Area')\n facades = fields.Integer(string='Facades')\n garage = fields.Boolean(string='Garage')\n garden = fields.Boolean(string='Garden')\n garden_area = fields.Integer(string='Garden Area')\n garden_orientation = fields.Selection([\n ('north', 'North'),\n ('south', 'South'),\n ('east', 'East'),\n ('west', 'West'),\n ], copy=False, index=True, tracking=3, default='draft')\n property_type_id = fields.Many2one('property.type', string='Property Type')\n other_info = fields.Text(string='Other Info', required=False)\n salesman = fields.Many2one('res.users', string='Salesman', default=lambda self: self.env.user)\n buyer = fields.Many2one('res.partner', string='Buyer')\n # salesman_id = fields.Many2one('res.users', string='Salesman', default='Mitchell Admin')\n # buyer_id = fields.Many2one('res.partner', string='Buyer')\n tag_id = fields.Many2many('property.tag', string='Property Tag')\n offer_ids = fields.One2many('property.offer', 'property_id', string='Offers')\n total = fields.Float(compute='_compute_total', string='Total Area')\n best_offer = fields.Float(compute='_compute_best_offer_price', string='Best Offer', optional='hide')\n state = fields.Selection([\n ('new', 'New'),\n ('offer', 'Offer'),\n ('received', 'Received'),\n ('offer_accepted', 'Offer Accepted'),\n ('sold', 'Sold'),\n ('canceled', 'Canceled')\n ], copy=False, string='Status', default='new')\n # active = fields.Boolean(string=\"Active\", default=False)\n company_id = fields.Many2one(\n 'res.company', string='Company',\n default=lambda self: self.env.user.company_id,\n required=True)\n email_id = fields.Char(string=\"Email\")\n\n\n @api.onchange(\"garden\")\n def _onchange_garden(self):\n if self.garden == True:\n self.garden_area = 10\n self.garden_orientation = \"north\"\n else:\n self.garden_area = 0\n self.garden_orientation = None\n\n @api.depends(\"garden_area\", \"living_area\")\n def _compute_total(self):\n for rec in self:\n rec.total = 0\n rec.total = rec.garden_area + rec.living_area\n\n @api.depends(\"offer_ids.price\")\n def _compute_best_offer_price(self):\n for rec in self:\n rec.best_offer = 0\n rec.best_offer = max(rec.offer_ids.mapped(\"price\")) if rec.offer_ids else 0.0\n\n def action_cancel(self):\n if self.state != \"sold\":\n self.state = \"canceled\"\n else:\n raise UserError(\"A Sold Properties cannot be Canceled!!!\")\n\n\n def action_sold(self):\n if self.state != \"canceled\":\n self.state = \"sold\"\n else:\n raise UserError(\"A canceled property cannot be sold!!!\")\n\n\n @api.constrains('selling_price')\n def _check_selling_price(self):\n for rec in self:\n if rec.selling_price < rec.expected_price * 0.9:\n raise ValidationError(_(\"The selling price cannot be lower than 90% of the expected price.\"))\n\n\n @api.ondelete(at_uninstall=False)\n def _unlink_open(self):\n for rec in self:\n if rec.state not in ['new', 'canceled']:\n raise UserError(_(\"Only new and canceled properties can be deleted.\"))\n\n\n @api.model\n def _property_cron_job(self):\n pass\n\n def action_mail_send(self):\n # ''' Opens a wizard to compose an email, with relevant mail template loaded by default '''\n # self.ensure_one()\n # template_id = self._find_mail_template()\n # lang = self.env.context.get('lang')\n # template = self.env['mail.template'].browse(template_id)\n # if template.lang:\n # lang = template._render_lang(self.ids)[self.id]\n # ctx = {\n # 'default_model': 'real_estate.order',\n # 'default_res_id': self.ids[0],\n # 'default_use_template': bool(template_id),\n # 'default_template_id': template_id,\n # 'default_composition_mode': 'comment',\n # 'mark_so_as_sent': True,\n # 'custom_layout': \"mail.mail_notification_paynow\",\n # 'proforma': self.env.context.get('proforma', False),\n # 'force_email': True,\n # 'model_description': self.with_context(lang=lang).type_name,\n # }\n return {\n 'type': 'ir.actions.act_window',\n 'view_mode': 'form',\n 'res_model': 'mail.compose.message',\n 'views': [(False, 'form')],\n 'view_id': False,\n 'target': 'new',\n # 'context': ctx,\n }\n\n\n\n\n\n # name_search method\n\n\n # @api.model\n # def name_search(self, name='', args=None, operator='ilike', limit=100):\n # if args in self:\n # args = []\n # domain = args + ['|', ('property_type_id', operator, name), ('tag_id', operator, name)]\n # return super(realestateorder, self).search(domain, limit=limit).name_get()\n\n def action_send_by_mail(self):\n template = self.env.ref('real_estate.email_template_real_estate')\n for rec in self:\n template.send_mail(rec.id)\n\n # name_get_method\n\n\n # @api.multi\n # def name_get(self):\n # return [(rec.id, \"%s:%s\" % (rec.name, rec.postcode)) for rec in self]\n\n\n\n\n\n","repo_name":"divyangi7/custom_15","sub_path":"custom_addons/Real_Estate/models/real_estate_order.py","file_name":"real_estate_order.py","file_ext":"py","file_size_in_byte":6390,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31540195739","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n'''\n @File : Hcf_lcd.py\n @Time : 2019/04/25 15:21:52\n @Author : R.\n @Version : 1.0\n @Contact : 827710637@qq.com\n @Desc : None\n'''\n\n'''\n 分别编写求两个整数的最大公约数的函数hcf和求最小公倍数的函数lcd。主函数已给出,\n 其从键盘接收两个整数,调用这两个函数后输出结果。\n (提示:求最大公约数可用辗转相除法。即将大数作为被除数,小数作为除数,若二者\n 余数不为0,则将小数作为被除数,余数作为除数,…直到余数为0。求最小公倍数则用\n 两数的积除以最大公约数即可。)\n'''\n\n\n# here put the import lib\n\ndef hcf(u,v):\n if v>u:\n u,v=v,u\n r=u%v\n while r!=0:\n u=v\n v=r\n r=u%v\n return v\n \ndef lcd(u,v,h):\n return u*v/h\n \nu=int(input(\"请输入第一个整数:\"))\nv=int(input(\"请输入第二个整数:\"))\nh=hcf(u,v)\nprint(\"%d和%d的最大公约数为%d:\"%(u,v,h))\nl=lcd(u,v,h)\nprint(\"%d和%d的最小公倍数为%d:\"%(u,v,l))\n","repo_name":"PiKaChu-R/code-learn","sub_path":"program/niuke-python/Hcf_lcd.py","file_name":"Hcf_lcd.py","file_ext":"py","file_size_in_byte":1110,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"73237599295","text":"import math\n\nnumberHash = dict()\nduplicated = dict()\nN = 0\nanswer = 0\n\n\ndef determine(n):\n if n == 1 or n == 0:\n return False\n\n for i in range(2, int(math.sqrt(n)) + 1):\n if n % i == 0:\n return False\n return True\n\n\ndef DFS(index, combination):\n global answer\n global N\n\n if index == N:\n if combination == \"\":\n return\n\n combination = int(combination)\n if determine(combination):\n if combination in duplicated:\n return\n\n duplicated[combination] = 1\n answer += 1\n return\n\n for i in numberHash:\n if numberHash[i] != 0:\n numberHash[i] -= 1\n DFS(index + 1, combination + i)\n numberHash[i] += 1\n DFS(index + 1, combination)\n\n return 0\n\n\ndef solution(numbers):\n global numberHash\n global N\n\n N = len(numbers)\n for i in numbers:\n if i not in numberHash:\n numberHash[i] = 0\n numberHash[i] += 1\n\n DFS(0, \"\")\n\n return answer\n","repo_name":"Interview-study/algosalza","sub_path":"프로그래머스/완전탐색/소수_찾기/subway.py","file_name":"subway.py","file_ext":"py","file_size_in_byte":1043,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"30850864415","text":"class Solution(object):\n def numSubmatrixSumTarget(self, matrix, target):\n \"\"\"\n :type matrix: List[List[int]]\n :type target: int\n :rtype: int\n \"\"\"\n n, m = len(matrix), len(matrix[0]) # n columns rows\n run = [[matrix[i][j] for j in range(m)] for i in range(n)]\n for i in range(1, n):\n for j in range(m):\n run[i][j] += run[i - 1][j]\n run = [[0] * m] + run\n\n # print(run)\n\n def helper(a):\n d = {0: 1}\n s = res = 0\n for i in range(len(a)):\n s += a[i]\n res += d.get(s - target, 0)\n d[s] = d.get(s, 0) + 1\n return res\n\n res = 0\n for i in range(n + 1):\n for j in range(i + 1, n + 1):\n a = [k2 - k1 for k1, k2 in zip(run[i], run[j])]\n # print(a)\n res += helper(a)\n return res\n\n\ns = Solution()\nprint(s.numSubmatrixSumTarget(matrix=[[0, 1, 0], [1, 1, 1], [0, 1, 0]], target=0))\nprint(s.numSubmatrixSumTarget(matrix=[[1, -1], [-1, 1]], target=0))\nprint(s.numSubmatrixSumTarget(matrix=[[0, 0, 0, 1, 1], [1, 1, 1, 0, 1], [1, 1, 1, 1, 0], [0, 0, 0, 1, 0], [0, 0, 0, 1, 1]], target=0))\n\n","repo_name":"zhangzjn/DTVNet","sub_path":"utils/csv_cal.py","file_name":"csv_cal.py","file_ext":"py","file_size_in_byte":1266,"program_lang":"python","lang":"en","doc_type":"code","stars":59,"dataset":"github-code","pt":"79"} +{"seq_id":"11753967831","text":"import time\nimport random\n\nfrom django.forms import widgets\nfrom django.db import models as m\n\nfrom .strutil import to36\n\n\ntry:\n from south.modelsinspector import add_introspection_rules\nexcept ImportError:\n pass\nelse:\n add_introspection_rules([], [\"^fi\\.dj\\.fields\\.IDField\"])\n\n\nclass IDField(m.Field):\n \"\"\" An ID field which is a 12 byte string, which is a base36 encoded 62\n bit integer, where the leading bits are time, and the trailing bits\n are random. \"\"\"\n description = \"A unique ID 12 byte string, suitable for use as a unique ID\"\n max_length = 12\n\n # Note: because of http://code.djangoproject.com/ticket/13696, IDField will\n # do Bad Things when it's used with an inline admin.\n\n @classmethod\n def new(cls):\n \"\"\" Returns a unique 62 bit ID which is based on a random number and\n the current time. \"\"\"\n # Truncate the current unix time to 30 bits...\n curtime = int(time.time()) & ((1<<30)-1)\n # ... then slap some random bits on the end.\n # Do this to help the database maintain temporal locality.\n # (it's possible that these should be swapped - with the\n # random bits coming first and the time bits coming second)\n return to36((curtime << 32) | random.getrandbits(32))\n\n def __init__(self, primary_key=True, auto=True, **kwargs):\n self.auto = auto\n super(IDField, self).__init__(primary_key=primary_key,\n default=None,\n max_length=self.max_length,\n blank=True)\n\n def db_type(self, connection):\n return \"CHAR(%s)\" %(self.max_length, )\n\n def to_python(self, value):\n if isinstance(value, basestring):\n return value\n elif isinstance(value, (int, long)):\n return to36(value)\n elif value is None:\n # When an object is being deleted, its primary key is null'd\n return value\n raise TypeError(\"Bad ID Field value: %r (bad type)\" %(value, ))\n\n def pre_save(self, model_instance, add):\n # Note: only set on save because Django's admin (or possibly just the\n # 'TabularInline' admin) gets upset when the primary key changes (as it\n # would if a 'default' value was set (eg, 'default=self.new').\n cur_val = super(IDField, self).pre_save(model_instance, add)\n if not cur_val and self.auto:\n cur_val = self.new()\n setattr(model_instance, self.attname, cur_val)\n if cur_val and len(cur_val) > self.max_length:\n raise ValueError(\"invalid %r on %r: %r is longer than %r\"\n %(self.attname, model_instance, cur_val,\n self.max_length))\n return cur_val\n\n def get_prep_value(self, value):\n if isinstance(value, (int, long)):\n # This can happen when using 'dumpdata', for some reason\n return to36(value)\n return value\n\n def formfield(self, **kwargs):\n defaults = {'widget': ReadOnlyWidget}\n defaults.update(kwargs)\n return super(IDField, self).formfield(**defaults)\n\n\nclass ReadOnlyWidget(widgets.Widget):\n def render(self, name, value, attrs=None):\n if value is None:\n value = ''\n final_attrs = self.build_attrs(attrs, type='hidden', name=name)\n if value != '':\n final_attrs['value'] = widgets.force_text(value)\n return widgets.format_html('{1}',\n widgets.flatatt(final_attrs),\n value)\n","repo_name":"wolever/dwdj","sub_path":"dwdj/fields.py","file_name":"fields.py","file_ext":"py","file_size_in_byte":3639,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"79"} +{"seq_id":"39551085613","text":"import random\n\nwant_more = 'хочу'\n\nwhile want_more == 'хочу':\n answer = random.randint(1, 10)\n guess = -1\n while answer != guess:\n guess = input('Введите число: ')\n try:\n guess = int(guess)\n except ValueError:\n print('Где у вас башка? Я вам сказала введите ЧИСЛО! Вы слышите, ЧИСЛО!!!!!!!')\n continue\n\n if answer < guess:\n print('Загаданое число меньше')\n elif answer > guess:\n print('Загаданое число больше')\n else:\n print('Бинго!!!!')\n want_more = input('Хотите сыграть ещё? \"хочу\" или \"не хочу\": ')\n\nprint()\nprint('До свиданья(не в том смысле!), приходите ещё!')\nprint()","repo_name":"stacy2010/guess_number","sub_path":"guess_number.py","file_name":"guess_number.py","file_ext":"py","file_size_in_byte":878,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"16312180899","text":"import addons.LevelSystem.handlers.handlerUser as handlerUser\n\nfrom settings.settingColors import *\nimport settings.settingThumbnail as settingThumbnail\n\nimport services.serviceBot as serviceBot\ndiscord = serviceBot.classBot.getDiscord()\n\n#Get the level and xp of a user\nasync def getUserLevel(ctx, user):\n\n # PERMISSIONS CHECK\n import addons.LevelSystem.functions.services.servicePermission as servicePermission\n if await servicePermission.permissionCheck(ctx, \"cmdLevel\") == False:\n return\n\n # If no user is specified, the command author is used\n if user == None:\n user = ctx.author\n\n # Get the user data\n userData = handlerUser.getUser(ctx.guild.id, user.id)[0]\n\n # Extract the data\n userLevel = userData[3]\n userXP = userData[4]\n\n # Calculate the next level\n nextLevel = userLevel * 100\n\n # Create the embed\n embed = discord.Embed(title=\"Level \"+ user.name, color=green)\n embed.add_field(name=\"Level\", value=f\"```{userLevel}```\", inline=True)\n embed.add_field(name=\"XP\", value=f\"```{userXP}/{nextLevel}```\", inline=True)\n embed.set_thumbnail(url=user.display_avatar)\n\n # Create the progress bar\n progress = int(userXP / nextLevel * 10)\n progressBar = \"🟩\" * progress + \"⬜\" * (10 - progress)\n embed.add_field(name=\"Progression\", value=f\"{progressBar}\", inline=False)\n\n # Send the embed\n await ctx.respond(embed=embed)","repo_name":"Ted-18/Bot.Assistant-LevelSystem","sub_path":"LevelSystem/functions/commands/commandLevel.py","file_name":"commandLevel.py","file_ext":"py","file_size_in_byte":1407,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"14466878650","text":"VARIABLE_TYPES = [\n ' 1:\n # if we aren't at the last key then go a level deeper\n return get_from_key_list(data[keys[0]], keys[1:])\n else:\n # return the value we want\n return data[keys[0]]\n\ndef set_from_key_list(data, keys, value):\n # if the key doesn't exist then return None\n if not keys[0] in data.keys():\n if len(keys) == 1:\n data[keys[0]] = value\n return data\n else:\n return None\n return None\n if len(keys) > 1:\n # if we aren't at the last key then go a level deeper\n ret = set_from_key_list(data[keys[0]], keys[1:], value)\n if ret == None:\n return None\n else:\n data[keys[0]] = ret\n else:\n # return the value we want\n data[keys[0]] = value\n return data\n\ndef add_in_values(data, values):\n for k in values:\n if values[k] != None:\n keys = k.split('.')\n data = set_from_key_list(data, keys, values[k])\n return data\n\ndef stringify(obj):\n name = type(obj).__name__\n variables = vars(obj)\n str_rep = ''\n for v in variables:\n str_rep += '{}={}, '.format(v, getattr(obj, v))\n return name + '(' + str_rep[:-2] + ')'\n\ndef clean_null(d):\n clean = {}\n if type(d) == dict:\n for k, v in d.items():\n if type(v) == dict:\n nested = clean_null(v)\n if len(nested.keys()) > 0:\n clean[k] = nested\n elif type(v) == list:\n for i in range(0, len(v)):\n v[i] = clean_null(v[i])\n v = [i for i in v if i]\n if len(v) > 0:\n clean[k] = v\n elif v or v == 0 or v == \"\":\n clean[k] = v\n for k in clean:\n if clean[k] == {} or clean[k] == []:\n del clean[k]\n else:\n clean = d\n return clean\n\ndef clean_unset(data):\n if type(data) == dict:\n for k in data:\n if type(data[k]) == dict:\n data[k] = clean_unset(data[k])\n elif type(data[k]) == list:\n data[k] = clean_unset(data[k])\n elif type(data[k]) == str:\n for vt in VARIABLE_TYPES:\n if data[k].startswith(vt):\n data[k] = None\n break\n # if data[k].startswith('<') and data[k].endswith('>'):\n # data[k] = None\n else:\n for k in range(0, len(data)):\n if type(data[k]) == dict:\n data[k] = clean_unset(data[k])\n elif type(data[k]) == list:\n data[k] = clean_unset(data[k])\n elif type(data[k]) == str:\n for vt in VARIABLE_TYPES:\n if data[k].startswith(vt):\n data[k] = None\n break\n # if data[k].startswith('<') and data[k].endswith('>'):\n # data[k] = None\n return data\n\ndef recurse_expand(data, components_list, indent=0):\n # print(' ' * indent + str(data))\n if type(data) == dict:\n for k in data:\n if type(data[k]).__name__ in components_list:\n data[k] = data[k].to_json()\n else:\n if type(data[k]) == dict:\n data[k] = recurse_expand(data[k], components_list, indent = indent+2)\n elif type(data[k]) == list:\n data[k] = recurse_expand(data[k], components_list, indent = indent+2)\n elif type(data[k]) == str:\n for vt in VARIABLE_TYPES:\n if data[k].startswith(vt):\n data[k] = None\n break\n # if data[k].startswith('<') and data[k].endswith('>'):\n # data[k] = None\n else:\n for k in range(0, len(data)):\n if type(data[k]).__name__ in components_list:\n data[k] = data[k].to_json()\n else:\n if type(data[k]) == dict:\n data[k] = recurse_expand(data[k], components_list, indent = indent+2)\n elif type(data[k]) == list:\n data[k] = recurse_expand(data[k], components_list, indent = indent+2)\n elif type(data[k]) == str:\n for vt in VARIABLE_TYPES:\n if data[k].startswith(vt):\n data[k] = None\n break\n # if data[k].startswith('<') and data[k].endswith('>'):\n # data[k] = None\n return data\n\ndef recurse_build(data, key_list, elements, indent=0):\n # print(' ' * indent + str(data))\n if type(data) == dict:\n for k in data:\n key = '.'.join(key_list + [k])\n if key in elements.keys():\n data[k] = elements[key]\n else:\n if type(data[k]) == dict:\n data[k] = recurse_build(data[k], key_list + [k], elements, indent = indent+2)\n elif type(data[k]) == list:\n data[k] = recurse_build(data[k], key_list + [k], elements, indent = indent+2)\n else:\n for k in range(0, len(data)):\n key = '.'.join(key_list)\n if key in elements.keys():\n data[k] = elements[key]\n else:\n if type(data[k]) == dict:\n data[k] = recurse_build(data[k], key_list, elements, indent = indent+2)\n elif type(data[k]) == list:\n data[k] = recurse_build(data[k], key_list, elements, indent = indent+2)\n return data\n\ndef get_key_string(data):\n temp = list(get_paths(data))\n ret = ['.'.join(a) for i, a in enumerate(temp) if a not in temp[:i]]\n return ret\n\ndef get_paths(d, current = []):\n for a, b in d.items():\n yield current+[a]\n if isinstance(b, dict):\n yield from get_paths(b, current+[a])\n elif isinstance(b, list):\n for i in b:\n yield from get_paths(i, current+[a])\n\ndef fix_brace_strings(text):\n text = text.replace('\\'{}\\'', '{}')\n text = text.replace('\"{}\"', '{}')\n return text","repo_name":"jfcarter2358/k8sGen","sub_path":"k8sgen/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":6737,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"973155302","text":"import bisect\n\n\ndef kthSmallest(matrix, k):\n \"\"\"\n :type matrix: List[List[int]]\n :type k: int\n :rtype: int\n \"\"\"\n # oneD = []\n # for arr in matrix:\n # oneD += arr\n\n # oneD = sorted(oneD)\n\n # return oneD[k-1]\n n = len(matrix)\n high, low = matrix[n-1][n-1], matrix[0][0]\n while low < high:\n cnt = 0\n mid = low + (high-low)//2\n for i in range(0, n):\n # cnt += bisect_right(matrix[i], mid)\n if cnt < k:\n low = mid+1\n else:\n high = mid\n return low\n\n\nprint(kthSmallest([[1, 5, 9], [10, 11, 13], [12, 13, 15]], 8))\n","repo_name":"meqmeq/LeetCodeProblems","sub_path":"kthSmallest.py","file_name":"kthSmallest.py","file_ext":"py","file_size_in_byte":653,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"75097429694","text":"from django.shortcuts import render, redirect\n\n# Create your views here.\nfrom django.shortcuts import render\nfrom .forms import UserRegForm, UserLoginForm\nfrom django.contrib import messages\nfrom django.contrib.auth import login, logout\n\n\n# Create your views here.\n\ndef sign_in(request):\n if request.method == 'POST':\n form = UserLoginForm(data=request.POST)\n if form.is_valid():\n user = form.get_user()\n login(request, user)\n return redirect('/')\n else:\n messages.error(request, 'Ошибка входа')\n else:\n form = UserLoginForm()\n return render(request, 'auth/auth.html', {\n 'title': 'Вход',\n 'form': form,\n })\n\n\ndef sign_up(request):\n if request.method == 'POST':\n form = UserRegForm(request.POST)\n if form.is_valid():\n form.save()\n messages.success(request, 'Вы успешно зарегистрированы')\n return redirect('sign-in')\n else:\n messages.error(request, 'Ошибка регистрации')\n else:\n form = UserRegForm()\n return render(request, 'auth/reg.html', {\n \"form\": form\n })\n\n\ndef user_logout(request):\n logout(request)\n return redirect('/')\n","repo_name":"FunckingCat/global-cource-work","sub_path":"app/regauth/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1279,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"15177840268","text":"from django.shortcuts import render, redirect\nfrom .models import Product, OrderItem, Order, Pembayaran\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import get_object_or_404\nfrom django.contrib import messages\nfrom django.urls import reverse, reverse_lazy\nfrom datetime import datetime\nfrom users.models import Profile\nfrom django.conf import settings\nfrom django.db.models import Avg, Sum, Count\nfrom .forms import PembayaranForm, OrderUpdateForm,OrderBayarForm, OrderItemForm, OrderItemForm2, MemberForm\nfrom rajaongkir import RajaOngkirApi\nfrom django.db.models import Q\nfrom django.views.generic.edit import CreateView, UpdateView, DeleteView\nfrom django.http import HttpResponse\n\nimport http.client\nimport requests\nimport datetime\nimport json\n\nfrom django.core.paginator import Paginator\n\nfrom django.template.loader import render_to_string\nfrom django.core.mail import EmailMessage\nfrom django.core.mail import send_mail\n\n\n# Create your views here.\n\n\ndef cek_ongkir(request,kota_id, kecamatan_tujuan_id, berat, jasa_ongkir):\n\tkota_id = kota_id\n\tkecamatan_tujuan_id = kecamatan_tujuan_id\n\tberat = berat\n\tjasa_ongkir = jasa_ongkir\n\turl = \"https://pro.rajaongkir.com/api/cost\" #origin=501&originType=city&destination=574&destinationType=subdistrict&weight=1700&courier=jne\n\tpayload = \"origin=\" + kota_id + \"&originType=city&destination=\" + kecamatan_tujuan_id + \"&destinationType=subdistrict&weight=\" + str(berat) + \"&courier=\" + jasa_ongkir\n\tr = requests.post(url,payload, headers = { 'key': settings.API_KEY_SECRET, 'content-type': \"application/x-www-form-urlencoded\" })\n\tongkir = r.json()\n\tcontext = {\n\t\t'ongkir':ongkir,\n\t\t'kota_asal':kota_id,\n\t\t'kecamatan_tujuan':kecamatan_tujuan_id,\n\n\t}\n\n\treturn render(request, 'shop/ongkir.html', context)\n\n\n\n\ndef cek_kecamatan_tujuan(request, kota_tujuan_id):\n\tkota_tujuan_id = kota_tujuan_id\n\turl = \"https://pro.rajaongkir.com/api/subdistrict?city=\" + str(kota_tujuan_id)\n\tr = requests.get(url, headers = { 'key': settings.API_KEY_SECRET })\n\tkecamatan = r.json()\n\tcontext = {\n\t\t'kecamatan':kecamatan,\n\t}\n\treturn render(request, 'shop/kecamatan.html', context)\n\n\n\n\n\ndef cek_kota_tujuan(request, prov_tujuan_id):\n\tprov_tujuan_id = prov_tujuan_id\n\turl = \"https://pro.rajaongkir.com/api/city?province=\" + prov_tujuan_id\n\t# print(url)\n\theaders = { 'key': settings.API_KEY_SECRET }\n\tr = requests.get(url, headers)\n\tkota = r.json()\n\n\tprint(json.dumps(kota, indent=2))\n\n\tcontext = {\n\t\t'city':kota,\n\n\t}\n\treturn render(request, 'shop/kota_tujuan.html', context)\n\ndef cek_kota(request, prov_id):\n\tprov_id = prov_id\n\turl = \"https://pro.rajaongkir.com/api/city?province=\" + prov_id\n\tprint(url)\n\theaders = { 'key': settings.API_KEY_SECRET }\n\tr = requests.get(url, headers)\n\tkota = r.json()\n\t# print(json.dumps(kota, indent=3))\n\tcontext = {\n\t\t'city':kota,\n\t\t'prov_id': prov_id\n\t}\n\treturn render(request, 'shop/kab.html', context)\n\n\ndef cek_provinsi(request):\n\turl = \"https://pro.rajaongkir.com/api/province\"\n\theaders = { 'key': settings.API_KEY_SECRET }\n\tr = requests.get(url, headers)\n\t\n\tprovinces = r.json()\n\t# print(json.dumps(provinces, indent=3))\n\n\tcontext = {\n\t\t'provinces':provinces,\n\t\t# 'city':list_of_city,\n\t}\n\ttemplate = 'shop/checkout_page.html'\n\treturn render(request, template, context)\n\ndef dashboard(request):\n\tpl = Product.objects.all()\n\n\tcontext = {\n\t\t'products' : pl,\n\t}\n\treturn render(request, 'shop/dashboard.html', context)\n\n@login_required\ndef product_list(request):\n\tpl = Product.objects.all()\n\titems = OrderItem.objects.filter(order__user = request.user or None, order__status_order = False)\n\ttotal = OrderItem.objects.filter(order__user = request.user or None, order__status_order = False).aggregate(Sum('price'))['price__sum'] or 0.00\n\tcount_items = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tform = OrderItemForm(request.POST or None)\n\n\tif form.is_valid():\n\t\tform.save()\n\t\treturn redirect(reverse('shop-checkout', kwargs = {'id':self.items.order.id}))\n\n\tcontext = {\n\t\t'products' : pl,\n\t\t'items':items,\n\t\t'subtotal':total,\n\t\t'count_items':count_items,\n\t\t'form':form,\n\t}\n\ttemplate = 'shop/home.html'\n\treturn render(request, template, context)\n\ndef product_detail(request, slug):\n\tobj = get_object_or_404(Product, slug=slug)\n\tcount_items = OrderItem.objects.filter(order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tcontext = {\n\t\t'obj':obj,\n\t\t'count_items':count_items,\n\t}\n\ttemplate = 'shop/product-details.html'\n\treturn render(request, template, context)\n\n\n@login_required\ndef add_to_cart(request,pk):\n\tpk = pk\n\tnow = datetime.datetime.now()\n\tproduct = get_object_or_404(Product, pk=pk)\n\tcart,created = Order.objects.get_or_create(\n user=request.user,\n status_order = False,\n )\n\torderitem,created = OrderItem.objects.get_or_create(product=product,order=cart, price=product.harga)\n\ttotal = OrderItem.objects.filter(order__status_order = False).aggregate(Sum('price'))['price__sum'] or 0.00\n\tif created:\n\t\tkode = 'INV'+now.strftime('%Y')+now.strftime('%m')+now.strftime('%d')+str(orderitem.order.pk)\n\t\torderitem.order.kode_nota = kode\n\t\torderitem.order.harga = total\n\t\torderitem.save()\n\t\tcart.save()\n\tmessages.success(request, \"Cart updated!\")\n\treturn redirect('shop-showcart')\n\n\n# 2020-06-06 05:24:29.900785\n\n\ndef delete_from_cart(request, **kwargs):\n item_to_delete = OrderItem.objects.filter(product__slug=kwargs['slug'])\n if item_to_delete.exists():\n item_to_delete[0].delete()\n messages.info(request, \"Item has been deleted\")\n return redirect('shop-showcart')\n\ndef show_cart(request):\n\titems = OrderItem.objects.filter(order__user = request.user, order__status_order = False)\n\torder = Order.objects.filter(user = request.user, status_order=False)\n\tall_items = OrderItem.objects.filter(order__user = request.user)\n\ttotal = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Sum('price'))['price__sum'] or 0.00\n\tcount_items = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tcontext = {\n\t\t'orders':order,\n\t\t'items':items,\n\t\t'item_true':all_items,\n\t\t'total':total,\n\t\t'count_items':count_items,\n\t}\n\treturn render(request, 'shop/cart.html', context)\n\nclass OrderItemDeleteView(DeleteView):\n model = OrderItem\n # template_name = \"shop/delete.html\"\n\n\n def get(self, request, *args, **kwargs):\n return self.post(request, *args, **kwargs)\n\n def get_success_url(self):\n return reverse('shop-showcart')\n\n# def pembayarancreate(request, kode_nota):\n# \tkode_nota = get_object_or_404(Order, kode_nota=kode_nota)\n# \tkode_nota = kode_nota\n# \t# pembayaran, created = Pembayaran.objects.get_or_create(kode_nota = kode_nota)\n# \t# print(pembayaran.kode_nota)\n\n# \t# pembayaran.kode_nota = kode_nota\n# \t# pembayaran.save()\n# \tforms = PembayaranForm(request.POST or None)\n\n# \tif forms.is_valid():\n# \t\tforms.kode_nota =kode_nota\n# \t\tprint(forms.kode_nota)\n# \t\tform.save()\n# \t\treturn redirect('shop-list')\n# \tcontext = {\n# \t\t'form':forms,\n# \t\t'kode_nota':kode_nota,\n# \t}\n# \ttemplate = 'shop/pembayarancreate.html'\n# \treturn render(request, template, context)\n\nclass OrderUpdateView(UpdateView):\n\tmodel = Order\n\ttemplate_name = 'shop/checkout_page.html'\n\t# template_name = 'shop/form.html'\n\tform_class = OrderUpdateForm\n\n\tdef get_context_data(self, **kwargs):\n\t\tcontext = super().get_context_data(**kwargs)\n\t\turl = \"https://pro.rajaongkir.com/api/province\"\n\t\theaders = { 'key': settings.API_KEY_SECRET }\n\t\tr = requests.get(url, headers)\n\t\tprovinces = r.json()\n\t\tcontext['provinces'] = provinces\n\t\tobj = super(OrderUpdateView, self).get_object(queryset = OrderItem.objects.filter(order__status_order = False))\n\t\tif obj.bukti_pembayaran:\n\t\t\tobj.status_order = True\n\t\t\tobj.save()\n\t\treturn context\n\n\tdef get_success_url(self, **kwargs):\n\t\treturn reverse('shop-bayar', kwargs={'pk':self.object.id})\n\n\ndef orderupdate(request, pk):\n\torder = get_object_or_404(Order, pk=pk)\n\tform = OrderUpdateForm(request.POST or None, instance = order)\n\turl = \"https://pro.rajaongkir.com/api/province\"\n\theaders = { 'key': settings.API_KEY_SECRET }\n\titems = OrderItem.objects.filter(order__user = request.user or None, order__status_order = False)\n\tsum_price_item = OrderItem.objects.filter(order__user = request.user or None, order__status_order = False).aggregate(Sum('total_harga'))['total_harga__sum'] or 0.00\n\t# print(\"order :\", order.total_harga)\n\torder.harga = sum_price_item\n\torder.save()\n\n\tr = requests.get(url, headers)\n\tprovinces = r.json()\n\tif form.is_valid():\n\t\torderitem = OrderItem.objects.filter(order__kode_nota = order.kode_nota)\n\t\torder.total_harga = order.harga + order.harga_ongkir\n\t\t# order.status_order = True\n\t\torder.save()\n\t\tform.save()\n\t\tuser = request.user\n\t\tto_email = user.email\n\t\tmessage = render_to_string('shop/order_email.html', {\n 'user': user,\n 'order':order,\n 'orderitem':orderitem,\n # 'domain': current_site.domain,\n # 'uid':urlsafe_base64_encode(force_bytes(user.pk)),\n # 'token':account_activation_token.make_token(user),\n })\n\t\tmail_subject = 'Mennunggu Pembayaran Anda'\n\t\temail = EmailMessage(mail_subject, message, to=[to_email])\n\t\temail.send()\n\t\tsend_mail(\n\t\t mail_subject,\n\t\t message,\n\t\t 'alifvianm@gmail.com',\n\t\t [to_email],\n\t\t fail_silently=False,\n\t\t)\n\t\t\t# return redirect(reverse('shop-list'))\n\t\treturn redirect(reverse('shop-orderdetail', kwargs={'pk':pk}))\n\n\tcontext = {\n\t\t'total':sum_price_item,\n\t\t'form':form,\n\t\t'provinces':provinces,\n\t\t'items':items,\n\t}\n\treturn render(request, 'shop/checkout_page.html', context)\n\n\n\nclass OrderBayarUpdateView(UpdateView):\n\tmodel = Order\n\ttemplate_name = 'shop/bayar.html'\n\tform_class = OrderBayarForm\n\n\tdef get_context_data(self, **kwargs):\n\t\tcontext = super().get_context_data(**kwargs)\n\t\tcontext['users'] = self.request.user\n\t\tcontext['bayar'] = Order.objects.filter(user = self.request.user, status_order = False)\n\t\tobj = super(OrderBayarUpdateView, self).get_object(queryset = context['bayar'])\n\t\tprint(\"obj = \", obj)\n\t\tprint(obj.bukti_pembayaran)\n\t\tif obj.bukti_pembayaran:\n\t\t\tobj.status_order = True\n\t\t\tprint(obj.status_order)\n\t\t\tobj.save()\n\t\t\tprint(obj.status_order)\n\n\t\t\n\t\treturn context\n\n\t# def form_valid(self, form):\n\t# \tOrder.objects.filter(pk = self.id).update(status_order = True)\n\t# \treturn super().form_valid(form)\n\n\tdef get_success_url(self, **kwargs):\n\t\treturn reverse('shop-list')\n\ndef order_update(request, pk):\n\turl = \"https://pro.rajaongkir.com/api/province\"\n\theaders = { 'key': settings.API_KEY_SECRET }\n\tr = requests.get(url, headers)\n\tprovinces = r.json()\n\ttemplate = 'shop/form.html'\n\torder = get_object_or_404(Order, pk=pk)\n\tform = OrderUpdateForm(request.POST or None, instance=order)\n\torderitem = OrderItem.objects.filter(order__pk = pk, status_order = False, order__user = request.user)\n\n\n\tif form.is_valid():\n\t\t\n\t\tform.save()\n\t\treturn redirect('shop-list')\n\n\tcontext = {\n \t'form':form,\n\t\t'provinces':provinces,\n\t}\n\treturn render(request, template, context)\n\n# def orderitem_update(request, pk):\n# \tord_item = get_object_or_404(OrderItem, pk=pk)\n# \tform = OrderItemForm(request.POST or None)\n\t\n# \tif form.is_valid():\n# \t\tform.save()\n\n\t\n# \tcontext = {\n# \t\t\"form\": form\n# \t\t}\n# \treturn redirect(reverse('shop-checkout', kwargs = {'pk':pk}))\n\t# return redirect('/')\n\n\ndef orderitem_update(request, order_id):\n\t\t\t\n\t\n\tcontext = {\n\t\t\"form\": form\n\t\t}\n\treturn redirect(reverse('shop-checkout', kwargs = {'pk':order_id}))\n\n\ndef updateorderitem(request, value, pk_orderitem):\n\torderitem = get_object_or_404(OrderItem, pk = pk_orderitem)\n\tOrderItem.objects.filter(order__user = request.user, order__status_order = False, pk = pk_orderitem).update(quantity=value)\n\toi = OrderItem.objects.get(pk = pk_orderitem, order__status_order = False, order__user = request.user)\n\tprint(\"quantity : \", oi.quantity)\n\toi.total_harga = oi.quantity * oi.price\n\tprint('harga :', oi.total_harga)\n\tOrderItem.objects.filter(order__user = request.user, order__status_order = False, pk = pk_orderitem).update(total_harga=oi.total_harga)\n\n\treturn HttpResponse(value)\n\nclass OrderItemUpdateView(UpdateView):\n\tmodel = OrderItem\n\tform_class = OrderItemForm\n\t# queryset = OrderItem.objects.filter(order__status_order = False)\n\t# success_url = reverse_lazy('shop-list')\n\n\t# def post(self, request, *args, **kwargs):\n\t# \tself.object = self.get_object()\n\t# \treturn super().post(request, *args, **kwargs)\n\tdef form_valid(self, form):\n\t\tqty = request.POST.getlist('quantity')\n\t\tself.object.order_set.update(quantity =qty )\n\n\tdef get_success_url(self):\n\t\treturn reverse('shop-checkout', kwargs={'pk':object.order.id})\n\ndef transaction(request):\n\torder = Order.objects.filter( user = request.user).order_by('-created_at')\n\tcount_items = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tcontext = {\n\t\t'orders':order,\n\t\t'count_items':count_items\n\t}\n\treturn render(request, 'shop/transaction.html', context)\n\n\n# def transaction(request):\n# \torder = Order.objects.filter(user = request.user, status_order = ).order_by('-created_at')\n\n\n\n# def OrderBayarUpdate(request, pk):\n# \torder = get_object_or_404(Order, pk=pk)\n# \tform = OrderBayarForm(request.POST or None, instance = order)\n# \titems = OrderItem.objects.filter(order__pk = pk, order__user = request.user)\n# \torders = Order.objects.filter(pk = pk, user = request.user)\n# \tif form.is_valid():\n# \t\tOrder.objects.filter(pk = pk).update(status_order = True)\n# \t\treturn redirect(reverse('shop-list'))\n# \tcontext = {\n# \t\t'orders':orders,\n# \t\t'items':items,\n# \t\t'form':form,\n# \t}\t\t\n# \treturn render(request, 'shop/bayar.html', context)\n\ndef OrderBayarUpdate(request, pk):\n\torder = get_object_or_404(Order, pk=pk)\n\tform = OrderBayarForm(request.POST, request.FILES, instance = order)\n\tform.kode_nota = order.kode_nota\n\tprint(form.kode_nota)\n\titems = OrderItem.objects.filter(order__pk = pk, order__user = request.user)\n\torders = Order.objects.filter(pk = pk, user = request.user)\n# \tparam = {\n# \"transaction_details\": {\n# \"order_id\": order.kode_nota ,\n# \"gross_amount\": 10000\n# },\n# \"item_details\": [{\n# \"id\": \"ITEM1\",\n# \"price\": 10000,\n# \"quantity\": 1,\n# \"name\": \"Midtrans Bear\",\n# \"brand\": \"Midtrans\",\n# \"category\": \"Toys\",\n# \"merchant_name\": \"Midtrans\"\n# }],\n# \"customer_details\": {\n# \"first_name\": \"John\",\n# \"last_name\": \"Watson\",\n# \"email\": \"test@example.com\",\n# \"phone\": \"+628123456\",\n# \"billing_address\": {\n# \"first_name\": \"John\",\n# \"last_name\": \"Watson\",\n# \"email\": \"test@example.com\",\n# \"phone\": \"081 2233 44-55\",\n# \"address\": \"Sudirman\",\n# \"city\": \"Jakarta\",\n# \"postal_code\": \"12190\",\n# \"country_code\": \"IDN\"\n# },\n# \"shipping_address\": {\n# \"first_name\": \"John\",\n# \"last_name\": \"Watson\",\n# \"email\": \"test@example.com\",\n# \"phone\": \"0 8128-75 7-9338\",\n# \"address\": \"Sudirman\",\n# \"city\": \"Jakarta\",\n# \"postal_code\": \"12190\",\n# \"country_code\": \"IDN\"\n# }\n# },\n \n# \"credit_card\": {\n# \"secure\": True,\n# \"bank\": \"bca\",\n# \"installment\": {\n# \"required\": False,\n# \"terms\": {\n# \"bni\": [3, 6, 12],\n# \"mandiri\": [3, 6, 12],\n# \"cimb\": [3],\n# \"bca\": [3, 6, 12],\n# \"offline\": [6, 12]\n# }\n# },\n# \"whitelist_bins\": [\n# \"48111111\",\n# \"41111111\"\n# ]\n# },\n# \"bca_va\": {\n# \"va_number\": \"12345678911\",\n# \"free_text\": {\n# \"inquiry\": [\n# {\n# \"en\": \"text in English\",\n# \"id\": \"text in Bahasa Indonesia\"\n# }\n# ],\n# \"payment\": [\n# {\n# \"en\": \"text in English\",\n# \"id\": \"text in Bahasa Indonesia\"\n# }\n# ]\n# }\n# },\n# \"bni_va\": {\n# \"va_number\": \"12345678\"\n# },\n# \"permata_va\": {\n# \"va_number\": \"1234567890\",\n# \"recipient_name\": \"SUDARSONO\"\n# },\n# \"callbacks\": {\n# \"finish\": \"https://demo.midtrans.com\"\n# },\n# \"expiry\": {\n# \"start_time\": \"2020-12-20 18:11:08 +0700\",\n# \"unit\": \"minute\",\n# \"duration\": 9000\n# },\n# \"custom_field1\": \"custom field 1 content\",\n# \"custom_field2\": \"custom field 2 content\",\n# \"custom_field3\": \"custom field 3 content\"\n# }\n# \tclient = 'SB-Mid-client-HcNQMZtgAkCva4_F'\n\n# \ttransaction = settings.SNAP.create_transaction(param)\n# \ttransaction_token = transaction['token']\n# \tprint('transaction_token:')\n# \tprint(transaction_token)\n# \t# transaction_redirect_url = transaction['redirect_url']\n# \ttransaction_redirect_url = reverse('shop-orderdetail', kwargs={'pk':pk})\n# \tprint('transaction_redirect_url:')\n# \tprint(transaction_redirect_url)\n\n\tif form.is_valid():\n\t\tform.save()\n\t\tOrder.objects.filter(pk = pk).update(status_bayar = 'SUDAH', status_order = True)\n\t\torderitem = OrderItem.objects.filter(order__kode_nota = order.kode_nota)\n\t\tuser = request.user\n\t\tto_email = user.email\n\t\tmessage = render_to_string('shop/bayar_email.html', {\n 'user': user,\n 'order':order,\n 'orderitem':orderitem,\n })\n\t\tmail_subject = 'Pembayaran Berhasil'\n\t\temail = EmailMessage(mail_subject, message, to=[to_email])\n\t\temail.send()\n\t\t\t# return redirect(reverse('shop-list'))\n\t\treturn redirect(reverse('shop-orderdetail', kwargs={'pk':pk}))\n\tcontext = {\n\t\t'orders':orders,\n\t\t'items':items,\n\t\t'form':form,\n\t\t# 'token':transaction_token,\n\t\t# 'transaction_redirect_url':transaction_redirect_url,\n\t\t# 'clk':client,\n\t}\t\t\n\treturn render(request, 'shop/bayar.html', context)\n\ndef orderdetail(request, pk):\n\torder = get_object_or_404(Order, pk=pk)\n\t\n\tkota_id = order.kota_asal\n\tkecamatan_tujuan_id = order.kecamatan_tujuan\n\tberat = 1000\n\tjasa_ongkir = order.jasa_ongkir\n\turl = \"https://pro.rajaongkir.com/api/cost\" #origin=501&originType=city&destination=574&destinationType=subdistrict&weight=1700&courier=jne\n\tpayload = \"origin=\" + kota_id + \"&originType=city&destination=\" + kecamatan_tujuan_id + \"&destinationType=subdistrict&weight=\" + str(berat) + \"&courier=\" + jasa_ongkir\n\tr = requests.post(url,payload, headers = { 'key': settings.API_KEY_SECRET, 'content-type': \"application/x-www-form-urlencoded\" })\n\tongkir = r.json()\n\n\torderitem = OrderItem.objects.filter(order__kode_nota = order.kode_nota)\n\tcount_items = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tcontext = {\n\t\t'orders':order,\n\t\t'items':orderitem,\n\t\t'count_items':count_items,\n\t\t'ongkir':ongkir\n\t}\n\treturn render(request, 'shop/order_detail.html', context)\n\n@login_required\ndef show_product(request):\n\tproduct = Product.objects.all().order_by('-created_at')\n\tcount_items = OrderItem.objects.filter(order__user = request.user, order__status_order = False).aggregate(Count('id'))['id__count'] or 0\n\tpaginator = Paginator(product, 1)\n\tpage_number = request.GET.get('page')\n\tpage_obj = paginator.get_page(page_number)\n\tcontext = {\n\t\t'products':product,\n\t\t'count_items':count_items,\n\t\t'page_obj':page_obj\n\t}\n\treturn render(request, 'shop/products.html', context)\n\n\ndef show_product_home(request):\n\tproduct = Product.objects.all().order_by('-created_at')\n\tpaginator = Paginator(product, 20)\n\tpage_number = request.GET.get('page')\n\tpage_obj = paginator.get_page(page_number)\n\tcontext = {\n\t\t'products':product,\n\t\t'page_obj':page_obj\n\t}\n\treturn render(request, 'shop/products.html', context)\n\n\ndef member(request):\n\tif request.POST:\n\t\tform = MemberForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tuser = form.cleaned_data.get('nama')\n\t\t\tto_email = form.cleaned_data.get('email')\n\t\t\tmessage = render_to_string('shop/member_email.html', {\n 'user': user,\n })\n\t\t\tmail_subject = 'Thanks For Being Member Of DRWSkincare Banyuwangi.'\n\t\t\temail = EmailMessage(mail_subject, message, to=[to_email])\n\t\t\temail.send()\n\t\t\treturn redirect(reverse('shop-list'))\n\telse:\n\t\tform = MemberForm()\n\treturn render(request, 'shop/member.html', {'form':form})\n\ndef search(request):\n\ttemplate = 'shop/products.html'\n\tquery = request.GET.get('q', None)\n\tif query:\n\t\tproduct = Product.objects.filter(Q(nama__icontains = query)).order_by('-created_at')\n\telse:\n\t\tproduct = Product.objects.all()\n\n\tpaginator = Paginator(product, 20)\n\tpage_number = request.GET.get('page')\n\tpage_obj = paginator.get_page(page_number)\n\tcontext = {\n\t\t'products':product,\n\t\t'page_obj':page_obj,\n\t}\n\treturn render(request, template, context)\n\n\n","repo_name":"AlifvianM/meybeauty","sub_path":"shop/views copy.py","file_name":"views copy.py","file_ext":"py","file_size_in_byte":21002,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28463553804","text":"# Esempio slides raddoppio\ndef raddoppio():\n while True:\n x = yield\n x = yield x * 2\n\n\ng = raddoppio()\nr = next(g)\nif r is None:\n print(\"niente\")\nr = g.send(5)\nprint(\"r = {}\". format(r))\nr = next(g)\nif r is None:\n print(\"niente\")\nr = g.send(8)\nprint(\"r = {}\". format(r))\n","repo_name":"SerenaFraD/UNI-PA","sub_path":"Esempi-altro/raddoppioGeneratore.py","file_name":"raddoppioGeneratore.py","file_ext":"py","file_size_in_byte":294,"program_lang":"python","lang":"it","doc_type":"code","stars":4,"dataset":"github-code","pt":"79"} +{"seq_id":"71486216255","text":"'''\nCreated on Aug 24, 2012\n\n@author: Anthony Honstain\n\nThis is collection of higher level scripts for administrator or key section of code\nto process or cleanup data in the database.\n\n'''\nfrom django.db import connection\n\nfrom rcstats.rcdata.models import LapTimes, SingleRaceDetails, SingleRaceResults, RacerId, TrackName\nfrom rcstats.rcdata.models import SupportedTrackName, OfficialClassNames, AliasClassNames\n\nimport re\n\n\nclass _ProcessRacerId():\n '''\n _ProcessRacerId handles the work of identifying alias names and collapsing them\n to the alias with the most results. \n \n I have a number of different racer names, and many of them represent\n the same person but have slightly different spelling, structure, letter case.\n \n For the time being I need a quick and easy way to push these to a single name, so\n that people can get their data in a single place. \n '''\n def __init__(self, racerid_data):\n '''\n The constructor does the bulk of the work for this object.\n \n racerid_data: has a very specific format. Each element should have the form:\n [, , ]\n \n Example: [ [1, 'hovarter,kevin', 5], [2, 'gripentrog, kurt', 5], [3, 'Collins, Brandon', 2]\n '''\n \n # This will contain the primary names that all others will be aliases of.\n # This should be the name which is most frequently raced under.\n self._primaryName_dict = {} \n \n # I need to start identifying duplicates, these are several of the\n # most obvious reasons:\n # Character Case - upper/lower\n # FName - LName order\n #\n # I think I can save myself some grief if I assume names with more races should be the \n # original, so that when I run this in the future, I am less likely to have updates.\n #for result in results:\n # print result\n\n for result in racerid_data: \n # We need to calculate the formated name (no case, reverse ordering of fname/lname).\n tempname = result[1].lower()\n \n comma_index = tempname.find(',')\n if (comma_index > -1):\n # This probably contains a name of the format \", \"\n fname = tempname[comma_index + 1:].strip() \n lname = tempname[0 : comma_index].strip()\n tempname = fname + \" \" + lname\n\n #print result\n #print '\\t' + tempname\n \n # Now with the name formated, we can see it belongs in the dictionary\n if (not self._primaryName_dict.has_key(tempname)):\n \n # Before we add this as a new key, we are going to check if there\n # are any similarly spelled names.\n actual_spelling = self._check_edit_distance(tempname)\n \n # if _check_edit_distance found a key, that is the key will will\n # put this name under.\n if (actual_spelling != \"\"):\n self._primaryName_dict[actual_spelling].append(result)\n else:\n self._primaryName_dict[tempname] = [result]\n \n else:\n # This means it was an unkown name with no aliases found.\n self._primaryName_dict[tempname].append(result)\n \n # Filter the results so that we have just possible aliases \n self.likely_aliases = filter(lambda x: len(x[1]) > 1, self._primaryName_dict.items())\n \n #print \"Prining all the likely_aliases\"\n #for key in likely_aliases:\n # print '\\t', key\n \n def _check_edit_distance(self, tempname): \n for key in self._primaryName_dict.keys():\n if (len(key) == len(tempname) and +\\\n self._misspell_count(key, tempname) == len(key) - 1):\n return key\n return \"\"\n \n def _misspell_count(self, origional_str, test_str):\n same_count = 0\n for i in range(min(len(origional_str), len(test_str))):\n if (origional_str[i] == test_str[i]):\n same_count += 1\n return same_count\n\n \ndef collapse_racer_names():\n '''\n collapsenames uses the _ProcessRacerId object to identify aliases in the RacerId's \n and collapse them to a single name.\n \n This modifies and cleans up several of the rcdata tables.\n \n WARNING - this does not touch anything in ranking (if rankings are computed and the\n names are later aliased and collapsed, there is nothing that can be done but \n RECOMPUTE the ranking). \n '''\n \n # First step is to get all the racer names and their ids. \n # Info - We use the singleraceresults table to give us an\n # idea of how many racers their are for each name, more popular\n # names will be the root (that others are collapsed to).\n get_racers_cmd = '''\n SELECT racerid.id, racerid.racerpreferredname, COUNT(rresult.id) \n FROM rcdata_racerid as racerid , \n rcdata_singleraceresults as rresult \n WHERE racerid.id = rresult.racerid_id \n GROUP BY racerid.id \n ORDER BY COUNT(rresult.id) desc;\n ''' \n cursor = connection.cursor() \n cursor.execute(get_racers_cmd)\n results = cursor.fetchall() \n # Example results\n # [ [1, 'hovarter,kevin', 5], [2, 'gripentrog, kurt', 5], [3, 'Collins, Brandon', 2]\n \n processRacerObj = _ProcessRacerId(results)\n\n for canidate in processRacerObj.likely_aliases: \n \n primary_name = canidate[1][0]\n alias_list = canidate[1][1:]\n \n # Helpful information to look at when investigating problems \n #print\n #print \"Primary:\", primary_name\n #print \"Lkely Alias:\", alias_list \n #print 'COLLAPSING IN SQL'\n '''\n Example output:\n Primary: (3, u'Charlie, Jon', 2L)\n Lkely Alias: [(7, u'Charlee, Jon', 1L)]\n COLLAPSING IN SQL\n \n Primary: (1, u'Anthony Honstain', 2L)\n Lkely Alias: [(10, u'Honstain, Anthony', 1L)]\n COLLAPSING IN SQL\n \n Primary: (2, u'lowercase jim', 2L)\n Lkely Alias: [(6, u'LOWERCASE JIM', 1L)]\n COLLAPSING IN SQL\n Anthony Honstain\n lowercase jim\n Charlie, Jon\n Delta, Jon\n Echo, Jon\n Hotel, Jon\n Golf, Jon\n '''\n \n for alias in alias_list:\n \n alias_id = alias[0]\n new_racerid_obj = RacerId.objects.get(pk=primary_name[0])\n # ===================================\n # Update the race results\n # ===================================\n raceresult_set = SingleRaceResults.objects.filter(racerid__exact=alias_id).update(racerid = new_racerid_obj)\n \n # ===================================\n # Update the lap times\n # ===================================\n laptimes_set = LapTimes.objects.filter(racerid__exact=alias_id).update(racerid = new_racerid_obj)\n \n # ===================================\n # Remove the alias racerid\n # ===================================\n racerid_obj = RacerId.objects.get(pk=alias_id)\n racerid_obj.delete()\n\n\ndef collapse_alias_classnames(queryset):\n '''\n collapse_alias_classnames takes a queryset of SingleRaceDetails and will use\n the mapping between OfficialClassNames and AliasClassNames to clean up\n the classnames.\n \n This means setting the correct case on the class name if it is different from the official.\n This means changing an alias to the offical class name if they differ.\n \n History:\n Oct 25, 2012 - Anthony: I move the information about the main events\n to a separate column. No reason to have dead code in here.\n '''\n \n # First we need to construct an efficient structure for checking\n \n # Note - all the of the keys in lookup and offical_classnames are LOWERCASE\n lookup = {}\n official_classnames = {} # This is where the official names (unmodified case) are stored.\n \n officalclass_queryset = OfficialClassNames.objects.all()\n for officialclass in officalclass_queryset:\n lookup[officialclass.raceclass.lower()] = None\n official_classnames[officialclass.raceclass.lower()] = officialclass.raceclass\n \n alliasclass_queryset = AliasClassNames.objects.all()\n for alliasclass in alliasclass_queryset:\n lookup[alliasclass.raceclass.lower()] = alliasclass.officialclass.raceclass\n \n #\n # Look at each of the race details from the query set, and update the\n # class name if needed.\n #\n for racedetail in queryset:\n raceclass = racedetail.racedata\n \n if raceclass.lower() in lookup:\n #print \"HIT:\", raceclass.lower(), racedetail.racedata\n #\n # We found a hit, this is either an official class, or we will change it.\n #\n if (lookup[raceclass.lower()] == None):\n # This is an official class, we just need to check if case is good.\n if (official_classnames[raceclass.lower()] != raceclass):\n # The class is named correctly, but we are going to fix the case.\n racedetail.racedata = official_classnames[raceclass.lower()] \n #print \"CASE FIXED:\", racedetail.racedata\n racedetail.save() \n else:\n # We need to UPDATE the race to use the official class name.\n racedetail.racedata = lookup[raceclass.lower()]\n #print \"ALIAS FIXED:\", racedetail.racedata\n racedetail.save()","repo_name":"AnthonyHonstain/rcstats","sub_path":"rcdata/database_cleanup.py","file_name":"database_cleanup.py","file_ext":"py","file_size_in_byte":10088,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"29893891369","text":"import scrapy\nfrom scrapy_selenium import SeleniumRequest\nimport time\n\nclass DealsSpider(scrapy.Spider):\n name = 'deals'\n\n def start_requests(self):\n yield SeleniumRequest(\n url = 'https://slickdeals.net/computer-deals/',\n wait_time = 3,\n screenshot = True,\n callback = self.parse\n\n )\n\n def parse(self, response):\n driver = response.meta['driver']\n driver.set_window_size(1920,1080)\n time.sleep(60)\n \n items = response.xpath(\"//li[@class='fpGridBox grid altDeal hasPrice']\")\n for item in items:\n yield{\n \"item store\" : response.xpath(\".//span[@class='blueprint']/button/text()\").get() ,\n \"description\" : response.xpath(\".//a[contains(@class, 'itemTitle bp-p-dealLink bp-c-link')]/text()\").get() ,\n \"price\" : response.xpath(\"normalize-space(.//div[@class='itemPrice wide ']/text())\").get()\n }\n\n # driver.save_screenshot(\"test.png\")\n # driver.close()\n next_page = response.xpath(\"//*[@id='fpMainContent']/div[6]/a[position() = last()]/@href\").get()\n\n if next_page:\n next_page_url = response.urljoin(next_page)\n yield SeleniumRequest(\n url = next_page_url,\n wait_time = 3,\n callback = self.parse\n )","repo_name":"DevJuma10/Slickdeals-","sub_path":"slikdeals/spiders/deals.py","file_name":"deals.py","file_ext":"py","file_size_in_byte":1385,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"74550784896","text":"import tensorflow.compat.v1 as tf\nimport numpy as np\nimport collections\nimport random\nimport math\nimport datetime\nimport time\n\nword_map = {}\ndata = []\nside_info = np.loadtxt('./data/side_info_feature', dtype=int)\nitem_size, feature_size = side_info.shape\nembedding_size = 128\nn_sampled = 50\nnum_gpus = 2\nbatch_size = 256\nnum_steps = 200001 # data_size / batch_size * n_epoch\nevery_k_step = 5000\nnum_skips = 4 # batch_size % num_skips == 0\nwindow_size = 4\ntf.disable_eager_execution()\n\nitem_set = set()\ndef read_data(filename):\n global item_set\n with open(filename) as f:\n for line in f.readlines():\n line = line.strip().split(' ')\n data.extend(line)\n item_set = set(data)\n return data\n\n\ndata_index = 0\ndef generate_batch(batch_size):\n global data_index\n batch = np.ndarray(shape=(batch_size), dtype=np.int32)\n label = np.ndarray(shape=(batch_size, 1), dtype=np.int32)\n span = 2 * window_size + 1\n buffer = collections.deque(maxlen=span)\n if data_index + span > len(data):\n data_index = 0\n\n buffer.extend(data[data_index: data_index + span])\n data_index += span\n for i in range(batch_size // num_skips):\n tgt = window_size\n visited_tgt = [tgt]\n for j in range(num_skips):\n while tgt in visited_tgt:\n tgt = random.randint(0, span - 1)\n visited_tgt.append(tgt)\n batch[i * num_skips + j] = buffer[window_size]\n label[i * num_skips + j, 0] = buffer[tgt]\n if data_index == len(data):\n for k in range(span):\n buffer.append(k)\n data_index = span\n else:\n buffer.append(data[data_index])\n data_index += 1\n data_index = (data_index + len(data) - span) % len(data)\n return batch, label\n\n\ndef _variable_on_cpu(name, shape, initializer, dtype=np.float32):\n with tf.device('/cpu:0'):\n var = tf.get_variable(name, shape, initializer=initializer, dtype=dtype)\n return var\n\n\ndef tower_loss(scope, inputs, labels):\n embedding_list = []\n\n for i in range(feature_size):\n embedding = _variable_on_cpu('side_info_{0}_embeddings'.format(i), [max(side_info[:, i]) + 1, embedding_size],\n tf.random_uniform_initializer(-1.0, 1.0))\n side_info_index = tf.nn.embedding_lookup(side_info[:, i], inputs)\n side_info_embed = tf.nn.embedding_lookup(embedding, tf.cast(side_info_index[:], dtype=tf.int32))\n embedding_list.append(side_info_embed)\n \n alpha_embedding = _variable_on_cpu('alpha_embeddings', [item_size, feature_size],\n tf.random_uniform_initializer(-1.0, 1.0))\n stacked_embed = tf.stack(embedding_list, axis=-1)\n alpha_index = tf.nn.embedding_lookup(side_info[:, 0], inputs)\n alpha_embed = tf.nn.embedding_lookup(alpha_embedding, alpha_index)\n alpha_embed_expand = tf.expand_dims(alpha_embed, 1)\n alpha_i_sum = tf.reduce_sum(tf.exp(alpha_embed_expand), axis=-1)\n merge_embedding = tf.reduce_sum(stacked_embed * tf.exp(alpha_embed_expand), axis=-1) / alpha_i_sum\n\n ''' cold start item\n stacked_embed = tf.stack(embedding_list[1:], axis=-1)\n alpha_index = tf.nn.embedding_lookup(side_info[:, 1], inputs)\n alpha_embed = tf.nn.embedding_lookup(alpha_embedding, alpha_index[:])\n alpha_embed_expand = tf.expand_dims(alpha_embed, 1)\n alpha_i_sum = tf.reduce_sum(tf.exp(alpha_embed_expand), axis=-1)\n merge_embedding = tf.reduce_sum(stacked_embed * tf.exp(alpha_embed_expand), axis=-1) / alpha_i_sum\n cold_start_embedding = tf.reduce_sum(stacked_embed * tf.exp(alpha_embed_expand), axis=-1) / alpha_i_sum\n '''\n weights = _variable_on_cpu('w', [item_size, embedding_size], tf.truncated_normal_initializer(stddev=1.0/math.sqrt(embedding_size)))\n biases = _variable_on_cpu('b', [item_size], tf.zeros_initializer())\n loss = tf.reduce_mean(tf.nn.nce_loss(\n weights=weights,\n biases=biases,\n labels=labels,\n inputs=merge_embedding,\n num_sampled=n_sampled,\n num_classes=item_size\n ))\n return loss, merge_embedding\n\n\ndef average_gradient(tower_grads):\n avg_grads = []\n for grads_vars in zip(*tower_grads):\n values = tf.concat([g.values / num_gpus for g, _ in grads_vars], 0)\n indices = tf.concat([g.indices for g, _ in grads_vars], 0)\n grad = tf.IndexedSlices(values, indices)\n\n var = grads_vars[0][1]\n cur_grad_and_var = (grad, var)\n avg_grads.append(cur_grad_and_var)\n return avg_grads\n\n\ndef get_final_embedding():\n cnt = item_size // batch_size\n remain = item_size % batch_size\n final_embedding = {}\n all_item = side_info[:, 0]\n all_item = np.concatenate([all_item, [0] * remain], axis=0)\n\n for i in range(cnt):\n eval_input = all_item[i * batch_size: (i + 1) * batch_size]\n eval_label = np.zeros((batch_size, 1))\n eval_embedding = sess.run(merged_embedding, feed_dict={train_input: eval_input, train_label: eval_label})\n # for cold start item\n # cold_start_embedding = sess.run(cold_start_embedding, feed_dict={train_input: eval_input, train_label: eval_label})\n eval_embedding = eval_embedding.tolist()\n if i == cnt - 1:\n eval_embedding = eval_embedding[:-remain]\n final_embedding.update({all_item[i*batch_size+k]: eval_embedding[k] for k in range(len(eval_embedding))})\n dump_embedding(final_embedding, 'data/item_embeddings')\n\n\ndef dump_embedding(embedding_result, output_file):\n with open(output_file, 'w') as f:\n for k, v in embedding_result.items():\n f.write(\"{0} {1}\\n\".format(k, \" \".join(list(map(lambda x: str(x), v)))))\n\n\nif __name__ == '__main__':\n d = read_data('data/walk_seq')\n\n graph = tf.Graph()\n with graph.as_default(), tf.device('/cpu:0'):\n train_input = tf.placeholder(tf.int32, shape=[batch_size])\n train_label = tf.placeholder(tf.int32, shape=[batch_size, 1])\n\n train_opt = tf.train.GradientDescentOptimizer(1.0)\n #train_opt = tf.train.AdamOptimizer(1.0)\n\n tower_grads = []\n batch_size_gpu = batch_size // num_gpus\n with tf.variable_scope(tf.get_variable_scope()):\n for i in range(num_gpus):\n with tf.device('/gpu:{0}'.format(i)):\n with tf.name_scope('tower_{0}'.format(i)) as scope:\n train_input_gpu = tf.slice(train_input, [i * batch_size_gpu], [batch_size_gpu])\n train_label_gpu = tf.slice(train_label, [i * batch_size_gpu, 0], [batch_size_gpu, 1])\n\n loss, merged_embedding = tower_loss(scope, train_input_gpu, train_label_gpu)\n tf.get_variable_scope().reuse_variables()\n\n grads = train_opt.compute_gradients(loss)\n tower_grads.append(grads)\n\n grads = average_gradient(tower_grads)\n apply_gradient_op = train_opt.apply_gradients(grads)\n\n init = tf.global_variables_initializer()\n\n config = tf.ConfigProto(allow_soft_placement=True)\n config.gpu_options.allow_growth = True\n\n with tf.Session(graph=graph, config=config) as sess:\n start_time = datetime.datetime.now()\n init.run()\n print('Init finished')\n saver = tf.train.Saver(max_to_keep=4)\n\n avg_loss = 0\n final_loss = 0\n for step in range(1, num_steps):\n batch_input, batch_label = generate_batch(batch_size)\n feed_dict = {train_input: batch_input, train_label: batch_label}\n _, loss_val, batch_res = sess.run([apply_gradient_op, loss, merged_embedding], feed_dict=feed_dict)\n\n avg_loss += loss_val\n final_loss += loss_val\n\n if step % every_k_step == 0:\n end_time = datetime.datetime.now()\n avg_loss /= every_k_step\n print(\"step: {0}, loss: {1}, time: {2}s\".format(step, avg_loss, (end_time-start_time).seconds))\n avg_loss = 0\n start_time = datetime.datetime.now()\n\n get_final_embedding()\n","repo_name":"bound2020/EGES_multiGPU","sub_path":"eges_multigpu.py","file_name":"eges_multigpu.py","file_ext":"py","file_size_in_byte":8127,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"79"} +{"seq_id":"19269577656","text":"import scrapy\nimport string\n\nclass ItemNews(scrapy.Item):\n date = scrapy.Field()\n title = scrapy.Field()\n content = scrapy.Field()\n# url = scrapy.Field()\n\nclass PostKotaSpider(scrapy.Spider):\n max_page = 7605\n cur_page = 1\n page_url = 'https://poskota.id/index-berita/page/'\n\n name = \"jakarta_poskota\"\n allowed_domains = [ 'poskota.id' ]\n start_urls = [\n 'https://poskota.id/index-berita/page/1',\n #'https://poskota.id/2020/3/10/bertemu-komisi-viii-dpr-dubes-arab-saudi-kami-batasi-umrah-warga-lokal',\n ]\n \n def parse(self, response):\n \"\"\" Retrieve article list \"\"\"\n self.logger.info('>> PROCESSING in parse %s\\n', response.url)\n for link in response.xpath('//div[@class=\"content-artikel-list-box\"]//a/@href').getall():\n self.logger.info('>> link in indekx page %s\\n', link)\n yield scrapy.Request(link, callback=self.parse_article_page)\n\n self.cur_page = self.cur_page + 1\n if self.cur_page <= self.max_page :\n next_url = self.page_url + str(self.cur_page)\n self.logger.info(\">>> NEXT URL: %s \", next_url)\n yield scrapy.Request(next_url, callback=self.parse)\n\n def parse_article_page(self, response):\n self.logger.info('>> PROCESSING in parse_article_page %s\\n', response.url)\n item = ItemNews()\n item['title'] = response.xpath('//h1[@class=\"title\"]/text()').get()\n item['date'] = response.xpath('//div[@class=\"date\"]/text()').get()\n content = response.xpath('//div[@class=\"text-box\"]//text()').getall()\n clean_content = [ s.rstrip() for s in content ]\n item['content'] = \"\".join(clean_content)\n #item['url'] = response.url\n yield item\n \n","repo_name":"gusman/web-crawler","sub_path":"text_collector/spiders/jakarta_poskota.py","file_name":"jakarta_poskota.py","file_ext":"py","file_size_in_byte":1749,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"7935155224","text":"\n_dict = {\n 0: 'job initialized',\n 1: 'uploading the file',\n 2: 'start to recoginse speech',\n 3: 'ASR completed',\n 4: 'ASR failed',\n 5: 'not job found'\n}\n\ndef get_job_status_message(code):\n return _dict.get(code)\n\n","repo_name":"wangjinyu124419/long-audio-asr","sub_path":"yitu/yitu_speech/common/exception/job_status_message.py","file_name":"job_status_message.py","file_ext":"py","file_size_in_byte":235,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"42810693107","text":"from libs.s3_client import s3_client\nimport os, csv, json\n\n\ndef handler(event, context):\n try:\n file_content = json.loads(event['body'])\n \n with open('/tmp/data.json', 'w') as f:\n json.dump(file_content, f)\n \n with open('/tmp/data.json', 'r') as f:\n csv_file = csv.writer(f)\n \n \n header = file_content[0].keys()\n csv_data = [header] + [[row[key] for key in header] for row in file_content]\n\n with open('/tmp/dump.csv', 'w') as f:\n _csv = csv.writer(f)\n _csv.writerows(csv_data)\n \n # s3_client().put_object(\n # Bucket=os.environ.get('BUCKET_NAME'),\n # Key=os.environ.get('CSV_FILE_NAME'),\n # Body=json_data\n # )\n\n return {\n \"statusCode\": 200,\n \"headers\": {\n \"Content-Type\": \"application/json\"\n },\n \"body\": \"Job completed\"\n }\n except Exception as e:\n print(e)\n return {\n \"statusCode\": 500,\n \"headers\": {\n \"Content-Type\": \"text/plain\"\n },\n \"body\": \"Something went wrong\"\n }\n","repo_name":"sebastiantuyu/python-monorepo","sub_path":"apps/lambda-json-to-csv/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1028,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"2435291444","text":"#from mpl_toolkits.axes_grid1.inset_locator import inset_axes\nimport matplotlib.cm as cm\nimport bzmagPy as bzmag\n\n#-----------------------------------------------------------------------------\nclass FluxlineArtist():\n # ------------------------------------------------------------------------\n # based on Data coordinates\n def __init__(self, parent):\n self.parent_ = parent\n self.fluxline_ = None\n self.contour_ = None\n self.nLevels_ = 20\n self.pot_min_ = 0\n self.pot_max_ = 0\n \n # ------------------------------------------------------------------------\n def setPotentials(self, mesh, pot):\n # 기존 데이터 삭제\n if self.fluxline_ != None:\n for flux in self.fluxline_.collections:\n flux.remove()\n if self.contour_ != None:\n for flux in self.contour_.collections:\n flux.remove()\n \n axes = self.parent_.axes_\n \n vets = mesh['vertices']\n triangles = mesh['triangles']\n xs = vets[:, 0]\n ys = vets[:, 1]\n \n self.pot_min_ = min(pot)\n self.pot_max_ = max(pot)\n #pot_del = (pot_max - pot_min) / 20\n #levels = np.arange(pot_min, pot_max, pot_del)\n cmap = cm.get_cmap(name='Blues', lut=None)\n self.fluxline_ = axes.tricontour(xs, ys, triangles, pot, levels=20, linewidths=0.2, colors='k')\n self.contour_ = axes.tricontourf(xs, ys, triangles, pot, levels=20, cmap=\"RdBu_r\")\n \n # Legend\n #cbaxes = inset_axes(self.axes_, width=\"3%\", height=\"40%\", loc='upper left') \n #self.fig_.colorbar(self.contour_, cax=cbaxes, shrink=0.5)\n \n def setLevels(self, levels):\n self.nLevels_ = levels\n self.fluxline_.set_levels(levels)\n self.contour_.set_levels(levels)\n \n # ------------------------------------------------------------------------\n def hide(self):\n if self.fluxline_ != None: \n for flux in self.fluxline_.collections:\n flux.set_visible(False)\n if self.contour_ != None: \n for flux in self.contour_.collections:\n flux.set_visible(False)\n \n # ------------------------------------------------------------------------\n def show(self):\n if self.fluxline_ != None: \n for flux in self.fluxline_.collections:\n flux.set_visible(True)\n if self.contour_ != None: \n for flux in self.contour_.collections:\n flux.set_visible(True)\n \n","repo_name":"JaenamBae/bzmagEditor","sub_path":"lib/FluxlineArtist.py","file_name":"FluxlineArtist.py","file_ext":"py","file_size_in_byte":2578,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"9787099176","text":"import datetime\nimport uuid\nfrom src.common.database import Database\n\n\nclass Match(object):\n\n def __init__(self, p1, p2, p1score, p2score, room_id, date=None, _id=None):\n self._id = self._id = uuid.uuid4().hex if _id is None else _id\n self.p1 = p1\n self.date = datetime.datetime.utcnow()\n self.p1score = p1score\n self.p2 = p2\n self.p2score = p2score\n self.room_id = room_id\n\n def save_to_mongo(self):\n Database.insert(collection='matches', data=self.json())\n\n def json(self):\n return {\n \"p1\":self.p1,\n \"p2\":self.p2,\n \"p1score\":self.p1score,\n \"p2score\":self.p2score,\n \"date\": self.date,\n \"room_id\":self.room_id,\n \"_id\":self._id\n }\n\n @classmethod\n def from_mongo(cls, _id):\n match_data = Database.find_one(collection=\"matches\", query={'_id': _id})\n return cls(**match_data)\n\n\n @classmethod\n def find_by_room_id(cls, _id):\n match_data = Database.sort(\"matches\", {\"room_id\": _id}, 'date')\n if match_data is None:\n return None\n else:\n return [cls(**data) for data in match_data]\n\n","repo_name":"sid-hant/tables","sub_path":"src/models/match.py","file_name":"match.py","file_ext":"py","file_size_in_byte":1200,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8841053069","text":"'''\nMops up the remaining connections that were not calculated by traverse\nDraws lines between existing points\nDoes not calculate Easements - these are filtered\nChecks if both ends of the Observation to be drawn are calc'd otherwise observation is not drawn\n'''\n\nimport genericFunctions as funcs\nfrom LandXML.RefMarks import RefMarkQueries\nfrom LandXML.Easements import RemoveEasements\nfrom LandXML import PointClass, SharedOperations, TraverseSideCalcs\nfrom DrawingObjects import LinesPoints, Arcs\n\nimport CadastreClasses as DataObjects\n\ndef main(gui, LandXML_Obj):\n '''\n Coordinates calculation of remaining connections\n Should be all single connections\n Either draws the connection (if connects to already calculated points)\n - trigger error when too different to calc'd point\n :param CadastralPlan: contains the calculated Cadastral data\n :param LandXML_Obj: contains observations to mop up\n :return Cadastral Plan\n '''\n\n #Create Connection Mopper Object\n ConnectionMop = ObservationMop(gui, LandXML_Obj)\n\n for obs in LandXML_Obj.ReducedObs:\n #ignore Easements\n if CheckIfEasement(LandXML_Obj, obs) or CheckObsAllowed(obs, LandXML_Obj.TraverseProps):\n continue\n\n ConnectionMop.ObservationCalculator(obs)\n\n print(\"Observations left after connection mopper: \" + str(len(LandXML_Obj.ReducedObs)))\n\ndef CheckObsAllowed(obs, TraverseProps):\n '''\n\n :param obs:\n :param TraverseProps:\n :return:\n '''\n for NotAllowedType in TraverseProps.NotAllowedObs:\n if NotAllowedType == obs.get(\"desc\"):\n return True\n return False\n\ndef CheckIfEasement(LandXML_Obj, Observation):\n '''\n Checks if observation is an easment\n :param LandXML_Obj:\n :param Observation:\n :return:\n '''\n try:\n PntRefNum = Observation.get(\"setupID\").replace(LandXML_Obj.TraverseProps.tag, \"\")\n TargetID = Observation.get(\"targetSetupID\").replace(LandXML_Obj.TraverseProps.tag, \"\")\n RemoveEasObj = RemoveEasements.RemoveEasementObservations(Observation,\n PntRefNum,\n LandXML_Obj)\n if RemoveEasObj.CheckLotParcel(TargetID):\n return False\n elif RemoveEasObj.SearchEasementParcels(TargetID):\n return True\n else:\n return False\n except AttributeError:\n return False\n\n\n\nclass ObservationMop:\n\n def __init__(self, gui, LandXML_Obj):\n self.gui = gui\n self.CadastralPlan = gui.CadastralPlan\n self.LandXML_Obj = LandXML_Obj\n\n def ObservationCalculator(self, Observation):\n '''\n Coordinates the seach for observation not drawn on the UI\n or added to the Cadastral Plan\n These will be the only connections remaining in LandXML_Obj.ReducedObs\n :return:\n '''\n #get observation name\n self.ObsName = Observation.get(\"name\")\n #Set whether observation is arc or line\n if Observation.tag.replace(self.LandXML_Obj.TraverseProps.Namespace,\"\") == \\\n 'ReducedObservation':\n self.Arc = False\n else:\n self.Arc = True\n #Get start and end point of Observation\n self.StartRef, self.EndRef, Flip, self.connection = self.FindEndPoint(Observation)\n if self.StartRef is not None:\n self.SrcPoint = self.CadastralPlan.Points.__getattribute__(self.StartRef)\n #create a traverse object to store connection\n TravStart = SharedOperations.TraverseStartPoint(self.StartRef, self.CadastralPlan)\n self.traverse = SharedOperations.initialiseTraverse(TravStart, TravStart.Layer, False)\n self.distance = self.GetDistance(Observation)\n self.bearing = self.GetBearing(Observation, Flip)\n if self.distance is None or self.bearing is None:\n Observation.getparent().remove(Observation)\n else:\n #calculate point\n self.PointCalculation(Observation)\n self.LineObj(Flip)\n self.DrawObservation()\n Observation.getparent().remove(Observation)\n \n\n\n #print(\"Done\")\n\n def FindEndPoint(self, Observation):\n '''\n Determines what the start end point of the observation is\n If both end and start of Observation are already calculated - just takes setupID\n If RM with no end not calc'd - sets already calc'd point as startpoint\n :return:\n '''\n try:\n SetupID = Observation.get(\"setupID\").replace(self.LandXML_Obj.TraverseProps.tag, \"\")\n TargetID = Observation.get(\"targetSetupID\").replace(self.LandXML_Obj.TraverseProps.tag, \"\")\n except AttributeError:\n Observation.getparent().remove(Observation)\n return None, None, None, None\n\n if hasattr(self.CadastralPlan.Points, SetupID) and \\\n hasattr(self.CadastralPlan.Points, TargetID):\n #Observation between 2 already calc'd points\n return SetupID, TargetID, False, True\n elif hasattr(self.CadastralPlan.Points, SetupID):\n #CadastralPlan only contains the setupid point\n return SetupID, TargetID, False, False\n elif hasattr(self.CadastralPlan.Points, TargetID):\n #Contains\n return TargetID, SetupID, True, False\n else:\n return None, None, None, None\n\n\n\n def GetDistance(self, Observation):\n '''\n Gets distance for obs\n If arc true, calculates chord distance. LandXML gives the arc\n :return:\n '''\n if not self.Arc:\n return Observation.get(\"horizDistance\")\n else:\n ArcLength = Observation.get(\"length\")\n self.radius = Observation.get(\"radius\")\n self.rotation = Observation.get(\"rot\")\n return funcs.CalcChordLength(self.radius, ArcLength)\n\n def GetBearing(self, Observation, Flip):\n '''\n Gets bearing of obs, rotates if targetID is start of connection\n :param Observation:\n :return:\n '''\n #retrieve bearing of observation\n if self.Arc:\n bearing = Observation.get(\"chordAzimuth\")\n else:\n bearing = Observation.get(\"azimuth\")\n\n #Flip bearing if required\n if Flip:\n bearing = str(funcs.FlipBearing(float(bearing)))\n if len(bearing.split(\".\")[1]) == 1 or \\\n len(bearing.split(\".\")[1]) == 3:\n bearing += \"0\"\n\n return bearing\n\n def PointCalculation(self, Observation):\n '''\n Calculates the end point of the connection\n :return: \n '''\n\n self.PointObj = PointClass.Points(self.LandXML_Obj, self.CadastralPlan.Points)\n self.point = self.PointObj.CalcPoints(self.bearing, self.distance,\n self.StartRef, self.EndRef,\n Observation)\n #create line object\n \n def LineObj(self, Flip):\n '''\n Creates a Line object between the start and end point\n :return: \n '''\n \n # get line properties\n self.SetLinePointGuiProps()\n \n if not self.Arc:\n self.line = DataObjects.Line(self.StartRef, self.EndRef,\n self.LineLayer, self.distance,\n self.PointObj.deltaE, self.PointObj.deltaN,\n self.bearing, self.LineColour)\n else:\n self.CreateArcLine(Flip)\n\n self.CreateTraverse()\n\n def CreateArcLine(self, Flip):\n\n # set rotation\n ArcObjCls = Arcs.ArcClass(self.rotation)\n if Flip:\n self.rotation = ArcObjCls.SetArcRotation(True)\n else:\n self.rotation = ArcObjCls.SetArcRotation(False)\n # get coordinates for centre of the arc\n CentreArcCoords = funcs.ArcCentreCoords(self.SrcPoint, self.point,\n self.radius, self.rotation)\n ArcAngles = funcs.ArcAngles(self.SrcPoint, self.point,\n CentreArcCoords, self.rotation)\n\n\n #create arc object\n self.line = DataObjects.Arc(self.StartRef, self.EndRef,\n self.LineLayer, self.radius,\n CentreArcCoords, self.rotation, self.distance,\n self.bearing, self.PointObj.deltaE,\n self.PointObj.deltaN, self.LineColour, ArcAngles)\n \n def CheckClose(self):\n '''\n For when calculation is between 2 points already calc'd\n Throw message when >5mm. Colour and show\n :return: \n '''\n def CreateTraverse(self):\n '''\n Crates a traverse instance for ref marks - required for Survey Automation output\n :return:\n '''\n\n point = DataObjects.Point(self.point.PntNum, self.point.E, self.point.N,\n self.point.NorthingScreen, self.point.Elev,\n self.point.Code, self.point.Layer)\n traverse = SharedOperations.initialiseTraverse(point, \"REFERENCE MARKS\", False)\n setattr(traverse.PointsRaw, point.PntNum, point)\n setattr(self.traverse.Lines, self.ObsName, self.line)\n setattr(self.traverse.LinesRaw, self.ObsName, self.line)\n self.traverse.Lines.LineNum += 1\n\n\n def DrawObservation(self):\n '''\n Draws the Observation in gui and adds it to the Cadastral Plan\n :return:\n '''\n \n if self.connection:\n self.DrawLineMeth()\n else:\n self.DrawPoint()\n setattr(self.CadastralPlan.Points, self.EndRef, self.point)\n setattr(self.CadastralPlan.PointsRaw, self.EndRef, self.point)\n self.DrawLineMeth()\n \n setattr(self.CadastralPlan.Lines, self.ObsName, self.line)\n setattr(self.CadastralPlan.LinesRaw, self.ObsName, self.line)\n\n #Add traverse\n TraverseNum = self.CadastralPlan.Traverses.TraverseCounter\n try:\n TravName = \"Traverse\" + \"_\" + self.traverse.type + \"_\" + str(TraverseNum)\n except TypeError:\n TravName = \"Traverse\" + \"_BOUNDARY_\" + str(TraverseNum)\n setattr(self.CadastralPlan.Traverses, TravName, self.traverse)\n setattr(self.CadastralPlan.Traverses, \"TraverseCounter\", (TraverseNum + 1))\n\n \n def DrawPoint(self): \n '''\n Draws point on UI\n '''\n pointObj = LinesPoints.AddPointToScene(self.gui.view, self.point,\n self.point.Layer)\n \n def DrawLineMeth(self):\n '''\n Draw Line coordination method\n :return: \n '''\n \n \n if not self.Arc:\n self.DrawLine()\n else:\n self.DrawArc()\n \n def DrawLine(self):\n '''\n Draws line in UI\n '''\n self.GraphLine = self.gui.view.Line(self.SrcPoint.E*1000, self.SrcPoint.NorthingScreen*1000,\n self.point.E*1000, self.point.NorthingScreen*1000,\n self.LinePen)\n\n # add QGraphicsLine to line.GraphicsItems\n setattr(self.line.GraphicsItems, \"Line\", self.GraphLine)\n self.line.BoundingRect = self.GraphLine.boundingRect()\n \n def DrawArc(self):\n '''\n Draws arc defined by SrcPoint and EndPoint coords on drawing canvas\n '''\n #Get Arc drawing params\n ArcParamsObj = ArcParams(self.line)\n #create path object to draw arc\n # get arc path object\n ArcPath = Arcs.DrawArc(self.SrcPoint, self.point, self.line.CentreCoords, ArcParamsObj)\n #draw arc\n ArcLine = self.gui.view.scene.addPath(ArcPath.arcPath, self.LinePen)\n self.line.BoundingRect = ArcLine.boundingRect()\n \n def SetLinePointGuiProps(self):\n '''\n Sets properties of line and points to be drawn on the drawing canvas\n '''\n\n #get line props\n LineProps = LinesPoints.LinePointProperties()\n\n self.LinePen, self.LineLayer, \\\n self.LineColour = LineProps.SetLineProperties(self.SrcPoint.Layer,\n self.point.Layer)\n \n \nclass ArcParams:\n def __init__(self, ArcObject):\n self.Radius = ArcObject.Radius\n self.ArcRotation = ArcObject.Rotation","repo_name":"sparkes-intrax/CadastreCalcsUI","sub_path":"src/LandXML/ConnectionMopper.py","file_name":"ConnectionMopper.py","file_ext":"py","file_size_in_byte":12647,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"23905763157","text":"'''\nСтворіть масив А [1..7] за допомогою генератора випадкових чисел і\nвиведіть його на екран. Збільште всі його елементи в 2 рази.\nВиконав : Канюка Р. 122В\n'''\nimport random\nimport numpy as np\nwhile True:\n # Рандомна генерація чисел\n while True:\n try:\n A = [random.randint(0,10) for i in range(7)]\n print(A)\n break\n except ValueError :\n print('Введіть число!')\n # Вивід масиву помноженого на 2\n for i in range(len(A)):\n A[i] *= 2\n print(A)\n quest = input('Завершити програму? Y/N : ')\n if(quest == 'Y' or quest == 'y'):\n break\n","repo_name":"OctavianusAvg/Donny","sub_path":"5.py","file_name":"5.py","file_ext":"py","file_size_in_byte":809,"program_lang":"python","lang":"uk","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31246716580","text":"#Dustin Smith\r\n#Project 4a\r\n#Takes customers names and addresses in loops\r\n\r\n#Establish library of names and addresses\r\nlistOfNames = []\r\nlistOfAddresses = []\r\n\r\n#input loop\r\nrepeat = 'yes'\r\nwhile repeat == 'yes':\r\n listOfNames.append(input(\"Input customers name: \"))\r\n listOfAddresses.append(input(\"Input customers address: \"))\r\n repeat = input(\"Would you like to repeat? Type 'yes' to continue: \")\r\n\r\n#output loop for each customer set\r\nfor names in listOfNames:\r\n customer = listOfNames.index(names)\r\n print(\"Customer name: \", listOfNames[customer], \"\\\r\n Customer address: \", listOfAddresses[customer], \"\\n\")","repo_name":"DustinSmith37/Python-Stuffs","sub_path":"Smith_D_4a.py","file_name":"Smith_D_4a.py","file_ext":"py","file_size_in_byte":629,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22892082460","text":"import json\nimport pandas as pd\nimport numpy as np \nimport random as rd \nimport os\n\n\nclass Tablero:\n \"\"\"\n La clase tablero genera un tablero con un numero de filas y \n columnas pasados como argumentos, así como el nombre del \n jugador asociado. Posee los métodos para comprobar su estado, \n comprobar una jugada introducida, actualizar un barco o una\n jugada, comprobar la evolución de los aciertos o del juego y\n abrir y guardar su estado en un archivo json.\n \"\"\"\n def __init__(self, columnas, filas, nombre):\n self.columnas = columnas\n self.filas = filas\n self.nombre = nombre\n\n fila = ['~'] * self.filas\n tablero = {i: fila for i in range(1, self.columnas + 1)}\n index = list(range(1,11))\n tablero = pd.DataFrame(tablero, index=index)\n\n self.tablero = tablero\n\n\n def comprobar_tablero(self):\n \"\"\"\n Muestra por pantalla el estado del tablero.\n \"\"\"\n print('\\n')\n print(f'El estado actual de {self.nombre} es: \\n')\n print(self.tablero, '\\n')\n\n\n def comprobar_jugada(self, casilla):\n \"\"\"\n Comprueba si una jugada es acertada o no y devuelve el reslutado.\n \"\"\"\n if self.tablero.loc[casilla[0]][casilla[1]] == '#':\n acierto = 'barco'\n print('\\n')\n print('¡Acierto! El disparo ha dado en un barco')\n elif self.tablero.loc[casilla[0]][casilla[1]] == '~':\n acierto = 'agua'\n print('\\n')\n print('El disparo ha dado en agua')\n else:\n acierto = 'ya atacado'\n print('\\n')\n print('La casilla elegida ya se había bombardeado')\n\n return acierto\n\n\n def actualizar_jugada(self, acierto, casilla):\n \"\"\"\n El tablero actualiza el tablero con el resultado de la \n jugada, ambos como argumentos.\n \"\"\"\n if acierto == 'barco':\n self.tablero.loc[casilla[0]][casilla[1]] = 'x'\n elif acierto == 'agua':\n self.tablero.loc[casilla[0]][casilla[1]] = 'o'\n\n return self.tablero\n\n\n def actualizar_barco(self, casillas):\n \"\"\"\n La función actualiza el tablero con la posición del barco si \n no había uno previamente, y devulve un estado de error por \n si ya había un barco en alguna de las casillas pasadas como \n argumento.\n \"\"\"\n errores = False\n for casilla in casillas:\n if self.tablero.loc[casilla[0]][casilla[1]] == '#':\n errores = True\n break\n if not errores:\n for casilla in casillas:\n self.tablero.loc[casilla[0]][casilla[1]] = '#' \n \n return self.tablero, errores\n\n \n def guarda_json(self, nombre_archivo):\n \"\"\"\n La función guarda en un archivo json su estado.\n \"\"\"\n ruta = os.path.dirname(__file__) + os.sep + 'Partidas_Batalla_naval' + os.sep + nombre_archivo\n\n with open(ruta, 'w+') as outfile:\n tablero_json= self.tablero.to_dict(orient='list')\n json.dump(tablero_json, outfile, indent=4)\n\n\n def abrir_json(self, archivo, jugador):\n \"\"\"\n La función carga como estado una judagada previa\n desde un archivo json.\n \"\"\"\n ruta = os.path.dirname(__file__) + os.sep + 'Partidas_Batalla_naval' + os.sep + archivo\n with open(ruta, 'r+') as outfile:\n partida = json.load(outfile)\n tablero_guardado = pd.DataFrame.from_dict(partida)\n tablero_guardado.rename(index=dict(zip(range(0,10), range(1,11))), inplace=True)\n self.tablero = tablero_guardado\n\n return self.tablero\n\n\n def comprobar_estado_aciertos(self):\n \"\"\"\n Cuenta el número de aciertos del tablero\n y los devuelve.\n \"\"\"\n aciertos = self.tablero[self.tablero == 'x'].count().sum()\n return aciertos\n\n\n def comprobar_evolucion_juego(self, tablero_contrincante):\n \"\"\"\n Dado el tablero del cotrincante como argumento muestra\n por pantalla la evolución de tu tablero y de tus aciertos.\n \"\"\"\n print(f\"\"\"\n###################################################\n# #\n# AQUI TIENES EL RESUMEN DE TU JUGADA #\n# #\"\"\")\n aciertos = self.comprobar_estado_aciertos()\n self.comprobar_tablero()\n print(f\"\"\"\nLlevas {aciertos} aciertos\"\"\")\n aciertos = tablero_contrincante.comprobar_estado_aciertos()\n tablero_contrincante.comprobar_tablero()\n print(f\"\"\"\n{self.nombre} lleva {aciertos} aciertos\"\"\")\n\n print(\"\"\"\n###################################################\"\"\")","repo_name":"GinaGarridoCasas/Gina","sub_path":"week3_course_python_III/day5_probability_statistics/exercise/juego_barcos/exercise-valido/tablero.py","file_name":"tablero.py","file_ext":"py","file_size_in_byte":4771,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27119304482","text":"import re\nimport logging\nimport os\nimport math\nimport configparser\nimport subprocess\n\nfrom ..exceptions import SetConfigError\nfrom ..common import Configurator\n\nLOGGER = logging.getLogger(__name__)\n\n\nclass Network(Configurator):\n _module = \"NETWORK\"\n _submod = \"NETWORK\"\n\n\n def __init__(self, user=None):\n Configurator.__init__(self, user)\n self._cmd_map = {'xps': ['all', 'off', 'half', 'separate'],\n 'rps': ['all', 'off', 'half', 'separate']}\n self._nic = self._get_nic()\n self._queue_dir = '/sys/class/net/{}/queues'.format(self._nic)\n\n def _get_nic(self):\n if not os.path.isfile('/etc/atuned/atuned.cnf'):\n raise SetConfigError('Cannot get network name')\n config = configparser.ConfigParser()\n config.read('/etc/atuned/atuned.cnf')\n return config.get('system', 'network')\n\n def _set_cpus(self, key, value, pattern):\n dir_strs = shell_cmd(['ls', self._queue_dir], \n 'Failed to get dir under {}'.format(self._queue_dir))\n dir_list = re.findall(pattern, dir_strs)\n core_num = os.cpu_count()\n for index, dir_name in enumerate(dir_list):\n file_path = '{}/{}/{}_cpus'.format(self._queue_dir, dir_name, key)\n shell_cmd(['cat', file_path],\n 'Failed to set {}={}: does not support for {}'.format(key, value, key))\n\n if value == 'off':\n set_value = '0'\n elif value == 'all':\n set_value = 'f' * (core_num // 4)\n if core_num % 4 != 0:\n bin_num = int('0b' + '1' * (core_num % 4), 2)\n set_value = f\"{bin_num:x}\" + set_value\n elif value == 'half':\n half_num = core_num // 2\n if core_num == 1:\n val_format = 1\n elif key == 'rps':\n offset = index % (half_num + core_num % 2)\n val_format = 1 << offset\n else:\n offset = index % half_num\n val_format = 1 << (offset + half_num + core_num % 2)\n set_value = f\"{val_format:x}\"\n else: # value == 'separate'\n num = 1 << (index % core_num)\n set_value = f\"{num:x}\"\n\n shell_cmd(['sh', '-c', 'echo {} > {}'.format(set_value, file_path)],\n 'Failed to set {} to {}'.format(key, file_path))\n return 0\n\n def _set(self, key, value):\n if not key.lower() in self._cmd_map or not value.lower() in self._cmd_map[key.lower()]:\n raise SetConfigError(\"Invalid value {}={}\".format(key, value))\n\n if key == 'xps':\n self._set_cpus(key.lower(), value.lower(), r'tx-\\d+')\n elif key == 'rps':\n self._set_cpus(key.lower(), value.lower(), r'rx-\\d+')\n\n return 0\n \n def _get(self, key, _):\n pass\n \n def _backup(self, config, _):\n return str(config)\n\n @staticmethod\n def check(config1, config2):\n return True\n\n\ndef shell_cmd(cmd, error_message):\n output = subprocess.run(cmd, shell=False, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n if output.returncode != 0:\n raise SetConfigError(error_message)\n return output.stdout.decode()\n\n","repo_name":"openeuler-mirror/A-Tune-Collector","sub_path":"atune_collector/plugin/configurator/network/network.py","file_name":"network.py","file_ext":"py","file_size_in_byte":3299,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"8841235989","text":"'''\nSets up, adds and removes data from traverse objects\nAlso commits a traverse to the lxmlObj\n'''\n\nfrom PyQt5.QtGui import QPen, QBrush\n\nclass TraverseObject:\n def __init__(self, gui, lineWidth, Colour):\n '''\n Colours graphitems in traverse object\n :param gui: contains traverse object\n :param lineWidth:\n :param Colour:\n '''\n\n #lines\n if len(gui.travese.Lines.__dict__.keys()) > 0:\n self.ColourLines(gui, lineWidth, Colour)\n\n # Points\n if len(gui.travese.Points.__dict__.keys()) > 0:\n self.ColourLines(gui, lineWidth, Colour)\n\n # Arcs\n if len(gui.travese.Arcs.__dict__.keys()) > 0:\n self.ColourLines(gui, lineWidth, Colour)\n\n def ColourLines(self, gui, lineWidth, Colour):\n '''\n Colours lines in traverse object\n :param gui:\n :param lineWidth:\n :param Colour:\n :return:\n '''\n\n for key in gui.traverse.Lines.__dict__.keys():\n if key != \"LineNum\":\n line = gui.traverse.Lines.__getattribute__(key)\n lineItem = line.GraphicsItems.Line\n self.SetObjectPen(lineItem, Colour, lineWidth)\n gui.scene.update()\n\n def ColourPoints(self, gui, lineWidth, Colour):\n '''\n Colours points in traverse object\n :param gui:\n :param lineWidth:\n :param Colour:\n :return:\n '''\n\n for key in gui.traverse.Points.__dict__.keys():\n if key != \"LineNum\":\n point = gui.traverse.Lines.__getattribute__(key)\n PointItem = point.GraphicsItems.Point\n self.SetObjectPen(PointItem, Colour, lineWidth)\n self.SetObjectBrush(PointItem, Colour)\n gui.scene.update()\n\n\n\n def ColourArcs(self, gui, lineWidth, Colour):\n '''\n Colourts arcs in traverse object\n :param gui:\n :param lineWidth:\n :param Colour:\n :return:\n '''\n for key in self.traverse.Arcs.__dict__.keys():\n if key != \"ArcNum\":\n arc = self.traverse.Arcs.__getattribute__(key)\n ArcItem = arc.GraphicsItems.Arc\n self.SetLinePen(ArcItem, Colour, lineWidth)\n\n\n\n def SetObjectPen(self, item, colour, linewidth):\n '''\n Sets colour of item pen\n :param line:\n :return:\n '''\n # QtGui.QColor.\n Pen = QPen(colour)\n Pen.setWidth(linewidth)\n item.setPen(Pen)\n\n def SetObjectBrush(self, item, colour):\n '''\n Sets colour ofitem brush\n :param line:\n :return:\n '''\n # QtGui.QColor.\n Brush = QBrush(colour)\n item.setBrush(Brush)\n\n\n","repo_name":"sparkes-intrax/CadastreCalcsUI","sub_path":"src/TraverseOperations/TraverseObject.py","file_name":"TraverseObject.py","file_ext":"py","file_size_in_byte":2766,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38586085844","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom sklearn import tree\nfrom sklearn.datasets import load_iris\n\n\ndef load_dataset():\n \"\"\"\n Loads iris dataset.\n :return: iris dataset\n \"\"\"\n dataset = load_iris()\n\n return dataset\n\n\ndef describe_dataset(dataset):\n \"\"\"\n Prints dataset description.\n :param dataset: dataset to describe.\n :return: None\n \"\"\"\n print(f\"Dataset feature names: {dataset.feature_names}\")\n print(f\"Dataset target names: {dataset.target_names}\")\n\n\ndef prepare_data_and_target(dataset, to_remove):\n \"\"\"\n Prepares data and target from the dataset by removing part of it.\n :param dataset: dataset to use.\n :param to_remove: list of indexes to remove from the dataset.\n :return: data and target\n \"\"\"\n data = np.delete(dataset.data, to_remove, axis=0)\n target = np.delete(dataset.target, to_remove)\n return data, target\n\n\ndef train(data, target):\n \"\"\"\n Creates and trains the decision tree classifier on the provided data and target.\n :param data: data to use for training.\n :param target: target to use for training.\n :return: trained decision tree classifier.\n \"\"\"\n classifier = tree.DecisionTreeClassifier()\n classifier = classifier.fit(data, target)\n return classifier\n\n\ndef predict(classifier, data, target):\n \"\"\"\n Predict using the provided classifier and data.\n :param classifier: trained decision tree classifier.\n :param data: data to use for prediction.\n :param target: expected result.\n :return: prediction.\n \"\"\"\n prediction = classifier.predict(data)\n print(\"Original Data:\\n\", data)\n print(\"Original Labels:\", target)\n print(\"Labels Predicted:\", prediction)\n return prediction\n\n\ndef draw(classifier, dataset):\n \"\"\"\n Draws decision tree from the trained classifier.\n :param classifier: trained decision tree classifier.\n :param dataset: dataset used for training.\n :return: None\n \"\"\"\n\n plt.figure(figsize=(50, 50))\n tree.plot_tree(classifier,\n feature_names=dataset.feature_names,\n class_names=dataset.target_names,\n filled=True)\n plt.show()\n\n\nif __name__ == \"__main__\":\n dataset = load_dataset()\n describe_dataset(dataset)\n data, target = prepare_data_and_target(dataset, [0, 50, 100])\n classifier = train(data, target)\n prediction = predict(classifier, dataset.data[[0, 50, 100]], dataset.target[[0, 50, 100]])\n draw(classifier, dataset)\n","repo_name":"blackbat13/MachineLearningPython","sub_path":"classifiers/decision_tree.py","file_name":"decision_tree.py","file_ext":"py","file_size_in_byte":2499,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39693825150","text":"from typing import List\nfrom fastapi import FastAPI, Depends, HTTPException, status, APIRouter\nfrom sqlalchemy.orm import Session\nfrom starlette.responses import JSONResponse\nfrom user import queries, schemas\nfrom database import SessionLocal, engine, get_db\n\n\nuser_router = APIRouter()\n\n\n@user_router.get('/user', response_model=List[schemas.User])\ndef get_users(db: Session = Depends(get_db)):\n users = queries.get_users(db)\n return users\n\n\n@user_router.post('/user', response_model=schemas.User)\ndef create_user(user: schemas.UserCreate, db: Session = Depends(get_db)):\n db_user = queries.get_user_by_email(db, user_email=user.email)\n if db_user:\n raise HTTPException(\n status_code=status.HTTP_400_BAD_REQUEST,\n detail=\"Email already registered\"\n )\n return queries.create_user(db=db, user=user)","repo_name":"jonnaschristian/Lead-Mentoring","sub_path":"12_Exercício/app/user/endpoints.py","file_name":"endpoints.py","file_ext":"py","file_size_in_byte":848,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"21460997081","text":"class Solution:\n def myPow(self, x: float, n: int) -> float:\n\n def helper(x, n):\n\n if x == 0:\n return 0\n if n == 0:\n return 1\n\n answer = helper(x, n//2)\n answer = answer * answer\n return answer * x if n % 2 else answer\n\n result = helper(x, abs(n))\n return result if n > 0 else 1/result","repo_name":"rajat698/Leetcode-Practice","sub_path":"Solutions/LC50.py","file_name":"LC50.py","file_ext":"py","file_size_in_byte":394,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"14644343328","text":"from __future__ import absolute_import, division, print_function\n\nimport numpy as np\n\nimport torch\nfrom scipy.spatial.transform import Rotation\n\n# ----------- Modification ----------\nclass Embedder(object):\n # Positional encoding (section 5.1)\n def __init__(self, **kwargs):\n self.kwargs = kwargs\n self.create_embedding_fn()\n def create_embedding_fn(self):\n embed_fns = []\n d = self.kwargs[\"input_dims\"]\n out_dim = 0\n if self.kwargs[\"include_input\"]:\n embed_fns.append(lambda x: x)\n out_dim += d\n max_freq = self.kwargs[\"max_freq_log2\"]\n N_freqs = self.kwargs[\"num_freqs\"]\n if self.kwargs[\"log_sampling\"]:\n freq_bands = 2.**torch.linspace(0., max_freq, steps=N_freqs)\n else:\n freq_bands = torch.linspace(2.**0., 2.**max_freq, steps=N_freqs)\n for freq in freq_bands:\n for p_fn in self.kwargs[\"periodic_fns\"]:\n embed_fns.append(lambda x, p_fn=p_fn, freq=freq: p_fn(x * freq))\n out_dim += d\n self.embed_fns = embed_fns\n self.out_dim = out_dim\n def embed(self, inputs):\n return torch.cat([fn(inputs) for fn in self.embed_fns], -1)\n\ndef get_embedder(multires):\n embed_kwargs = {\"include_input\": True,\n \"input_dims\": 1,\n \"max_freq_log2\": multires - 1,\n \"num_freqs\": multires,\n \"log_sampling\": True,\n \"periodic_fns\": [torch.sin, torch.cos],}\n embedder_obj = Embedder(**embed_kwargs)\n def embed(x, eo=embedder_obj):\n return eo.embed(x)\n return embed, embedder_obj.out_dim\n\n# ---------- Modification ----------\nclass HomographySample:\n def __init__(self, H, W, device=None):\n if device is None:\n self.device = torch.device(\"cpu\")\n else:\n self.device = device\n self.Height_tgt = H\n self.Width_tgt = W\n self.meshgrid = self.grid_generation(self.Height_tgt, self.Width_tgt, self.device)\n self.meshgrid = self.meshgrid.permute(2, 0, 1).contiguous() # 3xHxW\n self.n = self.plane_normal_generation(self.device)\n @staticmethod\n def grid_generation(H, W, device):\n x = np.linspace(0, W-1, W)\n y = np.linspace(0, H-1, H)\n # Inversing the order due to the usage of numpy\n xv, yv = np.meshgrid(x, y) # HxW\n xv = torch.from_numpy(xv.astype(np.float32)).to(dtype=torch.float32, device=device)\n yv = torch.from_numpy(yv.astype(np.float32)).to(dtype=torch.float32, device=device)\n ones = torch.ones_like(xv)\n meshgrid = torch.stack((xv, yv, ones), dim=2) # HxWx3\n return meshgrid #(H,W,3); x - [0, W]; y - [0, H]\n @staticmethod\n def plane_normal_generation(device):\n n = torch.tensor([0, 0, 1], dtype=torch.float32, device=device)\n return n\n def sample(self, src_BCHW, d_src_B, G_tgt_src,\n K_src_inv, K_tgt):\n \"\"\"\n Coordinate system: x, y are the image directions, z is pointing to depth direction\n :param src_BCHW: torch tensor float, 0-1, rgb/rgba. BxCxHxW\n Assume to be at position P=[I|0]\n :param d_src_B: distance of image plane to src camera origin\n :param G_tgt_src: Bx4x4\n :param K_src_inv: Bx3x3\n :param K_tgt: Bx3x3\n :return: tgt_BCHW\n \"\"\"\n # parameter processing ------ begin ------\n B, channels, Height_src, Width_src = src_BCHW.size(0), src_BCHW.size(1), src_BCHW.size(2), src_BCHW.size(3)\n R_tgt_src = G_tgt_src[:, 0:3, 0:3]\n t_tgt_src = G_tgt_src[:, 0:3, 3]\n Height_tgt = self.Height_tgt\n Width_tgt = self.Width_tgt\n R_tgt_src = R_tgt_src.to(device=src_BCHW.device)\n t_tgt_src = t_tgt_src.to(device=src_BCHW.device)\n K_src_inv = K_src_inv.to(device=src_BCHW.device)\n K_tgt = K_tgt.to(device=src_BCHW.device)\n # the goal is compute H_src_tgt, that maps a tgt pixel to src pixel\n # so we compute H_tgt_src first, and then inverse\n n = self.n.to(device=src_BCHW.device)\n n = n.unsqueeze(0).repeat(B, 1) # Bx3\n # Bx3x3 - (Bx3x1 * Bx1x3)\n d_src_B33 = d_src_B.reshape(B, 1, 1).repeat(1, 3, 3) # B -> Bx3x3\n R_tnd = R_tgt_src - torch.matmul(t_tgt_src.unsqueeze(2), n.unsqueeze(1)) / -d_src_B33\n H_tgt_src = torch.matmul(K_tgt,\n torch.matmul(R_tnd, K_src_inv))\n # From source to Target\n with torch.no_grad():\n H_src_tgt = inverse(H_tgt_src)\n # create tgt image grid, and map to src\n meshgrid_tgt_homo = self.meshgrid.to(src_BCHW.device)\n # 3xHxW -> Bx3xHxW\n meshgrid_tgt_homo = meshgrid_tgt_homo.unsqueeze(0).expand(B, 3, Height_tgt, Width_tgt)\n # wrap meshgrid_tgt_homo to meshgrid_src\n meshgrid_tgt_homo_B3N = meshgrid_tgt_homo.view(B, 3, -1) # Bx3xHW\n meshgrid_src_homo_B3N = torch.matmul(H_src_tgt, meshgrid_tgt_homo_B3N) # Bx3x3 * Bx3xHW -> Bx3xHW\n # Bx3xHW -> Bx3xHxW -> BxHxWx3\n meshgrid_src_homo = meshgrid_src_homo_B3N.view(B, 3, Height_tgt, Width_tgt).permute(0, 2, 3, 1)\n meshgrid_src = meshgrid_src_homo[:, :, :, 0:2] / meshgrid_src_homo[:, :, :, 2:] # BxHxWx2\n np_meshgrid_src = meshgrid_src.cpu().detach().numpy()\n valid_mask_x = np.logical_and(np_meshgrid_src[:, :, :, 0] < Width_src, np_meshgrid_src[:, :, :, 0] > -1)\n valid_mask_y = np.logical_and(np_meshgrid_src[:, :, :, 1] < Height_src, np_meshgrid_src[:, :, :, 1] > -1)\n valid_mask = np.logical_and(valid_mask_x, valid_mask_y) # BxHxW\n valid_mask = torch.tensor(valid_mask).to(src_BCHW.device)\n # sample from src_BCHW\n # normalize meshgrid_src to [-1,1]\n meshgrid_src[:, :, :, 0] = (meshgrid_src[:, :, :, 0]+0.5) / (Width_src * 0.5) - 1\n meshgrid_src[:, :, :, 1] = (meshgrid_src[:, :, :, 1]+0.5) / (Height_src * 0.5) - 1\n tgt_BCHW = torch.nn.functional.grid_sample(src_BCHW, grid=meshgrid_src, padding_mode='border',\n align_corners=False)\n # BxCxHxW, BxHxW\n return tgt_BCHW, valid_mask\n\ndef inverse(matrices):\n inverse = None\n max_tries = 5\n while (inverse is None) or (torch.isnan(inverse)).any():\n torch.cuda.synchronize()\n inverse = torch.inverse(matrices)\n # Break out of the loop when the inverse is successful or there\"re no more tries\n max_tries -= 1\n if max_tries == 0:\n break\n # Raise an Exception if the inverse contains nan\n if (torch.isnan(inverse)).any():\n raise Exception(\"Matrix inverse contains nan!\")\n return inverse\n\ndef get_disparity_list(opt, B, device):\n S_coarse = opt.num_bins\n disparity_start, disparity_end = opt.disparity_start, opt.disparity_end\n if not opt.uniform_disparity:\n disparity_coarse = torch.linspace(\n disparity_start, disparity_end, S_coarse, dtype=torch.float32,\n device=device).unsqueeze(0).repeat(B, 1)\n return disparity_coarse # B, S\n elif opt.uniform_disparity:\n disparity_coarse = uniformly_sample_disparity_from_linspace_bins(\n batch_size=B, num_bins=S_coarse,\n start=disparity_start,\n end=disparity_end, device=device)\n return disparity_coarse # B, S\n\ndef uniformly_sample_disparity_from_linspace_bins(batch_size, num_bins, start, end, device):\n assert start > end\n B, S = batch_size, num_bins\n bin_edges = torch.linspace(start, end, num_bins+1, dtype=torch.float32, device=device) # S+1\n interval = bin_edges[1] - bin_edges[0] # scalar\n bin_edges_start = bin_edges[0:-1].unsqueeze(0).repeat(B, 1) # S -> BxS\n random_float = torch.rand((B, S), dtype=torch.float32, device=device) # BxS\n disparity_array = bin_edges_start + interval * random_float\n return disparity_array # BxS\n\n'''----------------Transformation for pose matrix'''\ndef transformation_from_parameters(axisangle, translation, invert=False):\n R = rot_from_axisangle(axisangle)\n t = translation.clone()\n if invert:\n R = R.transpose(1, 2)\n t *= -1\n T = get_translation_matrix(t)\n if invert:\n M = torch.matmul(R, T)\n else:\n M = torch.matmul(T, R)\n return M\n\ndef get_translation_matrix(translation_vector):\n T = torch.zeros(translation_vector.shape[0], 4, 4).cuda()\n t = translation_vector.contiguous().view(-1, 3, 1)\n T[:, 0, 0] = 1\n T[:, 1, 1] = 1\n T[:, 2, 2] = 1\n T[:, 3, 3] = 1\n T[:, :3, 3, None] = t\n return T\n\ndef rot_from_axisangle(vec):\n angle = torch.norm(vec, 2, 2, True)\n axis = vec / (angle + 1e-7)\n ca = torch.cos(angle)\n sa = torch.sin(angle)\n C = 1 - ca\n x = axis[..., 0].unsqueeze(1)\n y = axis[..., 1].unsqueeze(1)\n z = axis[..., 2].unsqueeze(1)\n xs = x * sa\n ys = y * sa\n zs = z * sa\n xC = x * C\n yC = y * C\n zC = z * C\n xyC = x * yC\n yzC = y * zC\n zxC = z * xC\n rot = torch.zeros((vec.shape[0], 4, 4)).cuda()\n rot[:, 0, 0] = torch.squeeze(x * xC + ca)\n rot[:, 0, 1] = torch.squeeze(xyC - zs)\n rot[:, 0, 2] = torch.squeeze(zxC + ys)\n rot[:, 1, 0] = torch.squeeze(xyC + zs)\n rot[:, 1, 1] = torch.squeeze(y * yC + ca)\n rot[:, 1, 2] = torch.squeeze(yzC - xs)\n rot[:, 2, 0] = torch.squeeze(zxC - ys)\n rot[:, 2, 1] = torch.squeeze(yzC + xs)\n rot[:, 2, 2] = torch.squeeze(z * zC + ca)\n rot[:, 3, 3] = 1\n return rot","repo_name":"gitkaichenzhou/DevNet","sub_path":"networks_dev/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":9708,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"79"} +{"seq_id":"37703946070","text":"import argparse\nfrom argparse import RawTextHelpFormatter\nimport os.path\nimport sys\nimport shutil\nimport subprocess\nimport json\nimport itertools\nimport time\nimport http.server\nimport socketserver\nimport re\n# local imports\nfrom fbksd_cli.common import *\n\n\n#=============================================\n# Common paths #\n#=============================================\nscenes_dir = os.path.join(os.getcwd(), 'scenes')\nrenderers_dir = os.path.join(os.getcwd(), 'renderers')\ndenoisers_dir = os.path.join(os.getcwd(), 'denoisers')\nsamplers_dir = os.path.join(os.getcwd(), 'samplers')\niqa_metrics_dir = os.path.join(os.getcwd(), 'iqa')\nconfigs_dir = os.path.join(os.getcwd(), 'configs')\ncurrent_config = os.path.join(configs_dir, '.current.json')\npage_dir = os.path.join(os.getcwd(), '.page')\nresults_dir = os.path.join(os.getcwd(), 'results')\ncurrent_slot_dir = os.path.join(results_dir, '.current')\nscenes_file = os.path.join(scenes_dir, '.fbksd-scenes-cache.json')\ninstall_prefix_dir = os.getenv('FBKSD_INSTALL_DIR',\n os.path.abspath(os.path.join(os.path.dirname(os.path.abspath(__file__)), os.pardir)))\n\n\n#=============================================\n# Global variables #\n#=============================================\ng_filters = {}\ng_filters_names = {}\ng_filters_versions = {}\ng_samplers = {}\ng_samplers_names = {}\ng_samplers_versions = {}\ng_renderers = {}\ng_scenes = {}\ng_scenes_names = {}\ng_iqa_metrics = {}\ng_persistent_state = None\ng_scenes_loaded = False\ng_filters_loaded = False\ng_samplers_loaded = False\ng_results_loaded = False\ng_samplers_results_loaded = False\ng_iqa_metrics_loaded = False\n\n\ndef load_scenes_g():\n global g_scenes, g_scenes_names, g_renderers, g_scenes_loaded\n if not g_scenes_loaded:\n g_scenes, g_scenes_names, g_renderers = load_scenes(scenes_file, renderers_dir)\n g_scenes_loaded = True\n\n\n\ndef load_filters_g():\n global g_filters, g_filters_names, g_filters_versions, g_filters_loaded\n if not g_filters_loaded:\n g_filters, g_filters_names, g_filters_versions = load_filters(denoisers_dir)\n g_filters_loaded = True\n\n\ndef load_samplers_g():\n global g_samplers, g_samplers_names, g_samplers_versions, g_samplers_loaded\n if not g_samplers_loaded:\n g_samplers, g_samplers_names, g_samplers_versions = load_samplers(samplers_dir)\n g_samplers_loaded = True\n\n\ndef load_results_g():\n global g_filters, g_scenes, g_results_loaded\n if not g_results_loaded:\n load_filters_results(g_filters, g_scenes, g_iqa_metrics, current_slot_dir)\n g_results_loaded = True\n\n\ndef load_samplers_results_g():\n global g_samplers, g_scenes, g_samplers_results_loaded\n if not g_samplers_results_loaded:\n load_samplers_results(g_samplers, g_scenes, g_iqa_metrics, current_slot_dir)\n g_samplers_results_loaded = True\n\n\ndef load_metrics_g():\n global g_iqa_metrics, g_iqa_metrics_loaded\n if not g_iqa_metrics_loaded:\n g_iqa_metrics = load_iqa_metrics(iqa_metrics_dir)\n g_iqa_metrics_loaded = True\n\n\ndef save_scenes_file_g(scenes):\n save_scenes_file(scenes, current_slot_dir)\n\n\ndef save_filters_file_g(scenes, filters_ids):\n save_techniques_file(scenes, filters_ids, g_filters, os.path.join(current_slot_dir, 'filters.json'))\n\n\ndef save_samplers_file_g(scenes, samplers_ids):\n save_techniques_file(scenes, samplers_ids, g_samplers, os.path.join(current_slot_dir, 'samplers.json'))\n\n\ndef save_results_file_g(scenes, filters_ids):\n save_filters_result_file(current_slot_dir, scenes, g_filters, filters_ids)\n\n\ndef save_samplers_results_file_g(scenes, samplers_ids):\n save_samplers_result_file(current_slot_dir, scenes, g_samplers, samplers_ids)\n\n\n#=============================================\n# Procedures #\n#=============================================\n\n\ndef cmd_init(args):\n if not os.path.isdir(renderers_dir):\n os.mkdir(renderers_dir)\n\n if not os.path.isdir(denoisers_dir):\n os.mkdir(denoisers_dir)\n\n if not os.path.isdir(samplers_dir):\n os.mkdir(samplers_dir)\n\n if not os.path.isdir(iqa_metrics_dir):\n os.mkdir(iqa_metrics_dir)\n\n if not os.path.isdir(configs_dir):\n os.mkdir(configs_dir)\n\n if not os.path.isdir(results_dir):\n new_slot_name = 'Results 1'\n if args.slot_name:\n new_slot_name = args.slot_name\n new_slot_path = os.path.join(results_dir, new_slot_name)\n os.makedirs(new_slot_path)\n os.symlink(new_slot_name, os.path.join(results_dir, '.current'))\n\n if not os.path.isdir(scenes_dir):\n if args.scenes_dir:\n if os.path.isdir(args.scenes_dir):\n os.symlink(os.path.abspath(args.scenes_dir), 'scenes')\n else:\n print('ERROR: \\\"{}\\\" is not an existing directory.\\n'.format(args.scenes_dir))\n else:\n os.mkdir(scenes_dir)\n\n # install build-in IQA metrics\n save_native_iqa(os.path.join(iqa_metrics_dir, 'MSE'), 'fbksd-mse')\n save_native_iqa(os.path.join(iqa_metrics_dir, 'RMSE'), 'fbksd-rmse')\n save_native_iqa(os.path.join(iqa_metrics_dir, 'PSNR'), 'fbksd-psnr')\n save_native_iqa(os.path.join(iqa_metrics_dir, 'SSIM'), 'fbksd-ssim')\n\n\ndef cmd_config(args):\n check = check_configs(configs_dir)\n if not check:\n return\n\n configs, current = get_configs(configs_dir, current_config)\n if not configs:\n print('No configuration found.')\n return\n\n print('{0}{1:<4s}{2:<16}{3:<20s}'.format(' ', 'Id', 'Creation Date', 'Name'))\n print('{0:75s}'.format('-'*75))\n for i, config in enumerate(configs):\n pretty_ctime = time.strftime('%d/%m/%Y', time.gmtime(config['ctime']))\n name = config['name']\n if current == config['name']:\n print('* {0:<3s}{1:<16}{2:<20s}'.format(str(i+1), pretty_ctime, name))\n else:\n print(' {0:<3s}{1:<16}{2:<20s}'.format(str(i+1), pretty_ctime, name))\n print('{0:75s}'.format('-'*75))\n\n\ndef cmd_config_new(args):\n check = check_scenes(scenes_dir) and check_filters(denoisers_dir) and check_samplers(samplers_dir)\n if not check:\n return\n\n new_config_file = os.path.join(configs_dir, args.name + '.json')\n if os.path.exists(new_config_file):\n print('ERROR: Config {} already exists.'.format(args.name))\n return\n if args.name[0] == '.':\n print('ERROR: Config name can not start with dot.')\n return\n\n load_scenes_g()\n scenes = []\n if args.scenes_all:\n scenes = g_scenes.values()\n else:\n scenes = scenesFromIds(args.scenes, g_scenes)\n\n load_filters_g()\n filters = []\n if args.filters_all:\n filters = g_filters_versions.values()\n else:\n filters = techniqueVersionsFromIds(args.filters, g_filters_versions)\n\n load_samplers_g()\n samplers = []\n if args.samplers_all:\n samplers = g_samplers_versions.values()\n else:\n samplers = techniqueVersionsFromIds(args.samplers, g_samplers_versions)\n\n spps = []\n if args.spps:\n spps = list(set(args.spps))\n spps.sort()\n else:\n spps = [2, 4, 8, 16]\n\n root_node = new_config(scenes, filters, samplers, spps)\n with open(new_config_file, 'w') as outfile:\n json.dump(root_node, outfile, indent=4)\n\n if args.select:\n if os.path.islink(current_config):\n os.unlink(current_config)\n os.symlink(args.name + '.json', current_config)\n\n\ndef cmd_config_select(args):\n check = check_configs(configs_dir)\n if not check:\n return\n\n configs, current = get_configs(configs_dir, current_config)\n if args.id < 1 or args.id > len(configs):\n print('ERROR: Invalid config ID.')\n return\n\n if os.path.islink(current_config):\n os.unlink(current_config)\n name = configs[args.id - 1]['name']\n os.symlink(name + '.json', current_config)\n print('Selected configuration: ' + name)\n\n\ndef cmd_config_show(args):\n check = check_configs(configs_dir)\n if not check:\n return\n\n if args.id is not None:\n configs, current = get_configs(configs_dir, current_config)\n if args.id < 1 or args.id > len(configs):\n print('ERROR: Invalid config ID.')\n return\n config_file = os.path.join(configs_dir, configs[args.id - 1]['name'] + '.json')\n else:\n config_file = current_config\n\n if not os.path.exists(config_file):\n print('Configuration not found.')\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n config.print()\n\n\ndef cmd_config_add_scenes(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n scenes = []\n if args.all:\n scenes = g_scenes.values()\n else:\n scenes = scenesFromIds(args.scenes, g_scenes)\n\n spps = list(set(args.spps))\n spps.sort()\n\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for scene in scenes:\n config.add_scene(scene, spps)\n config.save()\n\n\ndef cmd_config_remove_scenes(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n scenes = scenesFromIds(args.scenes, g_scenes)\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for scene in scenes:\n config.remove_scene(scene)\n config.save()\n\n\ndef cmd_config_add_filters(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n filters = []\n if args.all:\n filters = g_filters_versions.values()\n else:\n filters = techniqueVersionsFromIds(args.filters, g_filters_versions)\n\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for f in filters:\n config.add_filter(f)\n config.save()\n\n\ndef cmd_config_remove_filters(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n filters = techniqueVersionsFromIds(args.filters, g_filters_versions)\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for f in filters:\n config.remove_filter(f)\n config.save()\n\n\ndef cmd_config_add_samplers(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n samplers = []\n if args.all:\n samplers = g_samplers_versions.values()\n else:\n samplers = techniqueVersionsFromIds(args.samplers, g_samplers_versions)\n\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for s in samplers:\n config.add_sampler(s)\n config.save()\n\n\ndef cmd_config_remove_samplers(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n samplers = techniqueVersionsFromIds(args.samplers, g_samplers_versions)\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n for s in samplers:\n config.remove_sampler(s)\n config.save()\n\n\ndef cmd_config_add_spps(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n spps = set(args.spps)\n\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n config_scenes = []\n if args.scenes:\n scenes = scenesFromIds(args.scenes, g_scenes)\n for s in config.scenes:\n if s.scene in scenes:\n config_scenes.append(s)\n else:\n config_scenes = config.scenes\n\n for s in config_scenes:\n s.spps = list(set(s.spps) | spps)\n s.spps.sort()\n config.save()\n\n\ndef cmd_config_remove_spps(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n\n spps = set(args.spps)\n\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n config_scenes = []\n if args.scenes:\n scenes = scenesFromIds(args.scenes, g_scenes)\n for s in config.scenes:\n if s.scene in scenes:\n config_scenes.append(s)\n else:\n config_scenes = config.scenes\n\n for s in config_scenes:\n s.spps = list(set(s.spps) - spps)\n s.spps.sort()\n config.save()\n\n\ndef cmd_page_update(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n load_metrics_g()\n load_results_g()\n load_samplers_results_g()\n\n if args.all:\n scenes = g_scenes\n filters_ids = g_filters_versions.keys()\n samplers_ids = g_samplers_versions.keys()\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = {s.scene.id : s for s in config.scenes}\n filters_ids = [v.id for v in config.filter_versions]\n samplers_ids = [v.id for v in config.sampler_versions]\n\n save_iqa_metrics(g_iqa_metrics, current_slot_dir)\n save_scenes_file_g(scenes)\n save_filters_file_g(scenes, filters_ids)\n save_samplers_file_g(scenes, samplers_ids)\n save_results_file_g(scenes, filters_ids)\n save_samplers_results_file_g(scenes, samplers_ids)\n\n\ndef cmd_filters(args):\n if not check_filters(denoisers_dir):\n return\n\n load_filters_g()\n if not g_filters_versions:\n print('No filters found.')\n return\n\n print('{0:<5s}{1:<20s}{2:<10s}'.format('Id', 'Name', 'Status'))\n print('{0:75s}'.format('-'*75))\n for id, v in g_filters_versions.items():\n if v.status != 'ready' and args.ready:\n continue\n if v.tag == 'default':\n print('{0:<5s}{1:<20s}{2:<10s}'.format(str(id), v.technique.name, v.status))\n else:\n print('{0:<5s}{1:<20s}{2:<10s}'.format(str(id), v.technique.name + '-{}'.format(v.tag), v.status))\n print('{0:75s}'.format('-'*75))\n\n\ndef cmd_samplers(args):\n if not check_samplers(samplers_dir):\n return\n\n load_samplers_g()\n if not g_samplers_versions:\n print('No samplers found.')\n return\n\n print('{0:<5s}{1:<20s}{2:<10s}'.format('Id', 'Name', 'Status'))\n print('{0:75s}'.format('-'*75))\n for id, v in g_samplers_versions.items():\n if v.tag == 'default':\n print('{0:<5s}{1:<20s}{2:<10s}'.format(str(id), v.technique.name, v.status))\n else:\n print('{0:<5s}{1:<20s}{2:<10s}'.format(str(id), v.technique.name + '-{}'.format(v.tag), v.status))\n print('{0:75s}'.format('-'*75))\n\n\ndef cmd_filter_info(args):\n if not check_filters(denoisers_dir):\n return\n\n load_filters_g()\n filters = techniqueVersionsFromIds(args.filters, g_filters_versions)\n\n for f in filters:\n print('Name: {}'.format(f.technique.name))\n print('Full name: {}'.format(f.technique.full_name))\n print('Comment: {}'.format(f.technique.comment))\n print('Citation: {}'.format(f.technique.citation))\n print('Versions:')\n for version in f.technique.versions:\n print('{}'.format(version.id))\n print(' Name: {}'.format(version.tag))\n print(' Message: {}'.format(version.message))\n\n\ndef cmd_sampler_info(args):\n if not check_samplers(samplers_dir):\n return\n\n load_samplers_g()\n samplers = techniqueVersionsFromIds(args.samplers, g_samplers_versions)\n\n for f in samplers:\n print('Name: {}'.format(f.technique.name))\n print('Full name: {}'.format(f.technique.full_name))\n print('Comment: {}'.format(f.technique.comment))\n print('Citation: {}'.format(f.technique.citation))\n print('Versions:')\n for version in f.technique.versions:\n print('{}'.format(version.id))\n print(' Name: {}'.format(version.tag))\n print(' Message: {}'.format(version.message))\n\n\ndef cmd_scenes(args):\n if args.set:\n if os.path.isdir(args.set):\n if os.path.islink(scenes_dir):\n os.unlink(scenes_dir)\n elif os.path.exists(scenes_dir):\n print('ERROR: scenes folder already exists.')\n return\n os.symlink(os.path.abspath(args.set), 'scenes')\n print('Scenes path set.\\n')\n return\n else:\n print('ERROR: \\\"{}\\\" is not an existing directory.\\n'.format(args.set))\n return\n\n needs_update = False\n if not check_scenes(scenes_dir):\n return\n elif not os.path.exists(scenes_file):\n needs_update = True\n\n if args.update or needs_update:\n print('Updating scenes cache file...')\n scan_scenes(scenes_dir)\n print('Done.\\n')\n\n ready_only = args.ready\n\n load_scenes_g()\n scenes = g_scenes\n if ready_only:\n scenes = {sid: g_scenes[sid] for sid in g_scenes.keys() if g_scenes[sid].renderer.is_ready}\n\n if not scenes:\n print('No scenes.')\n return\n\n print('{0:<5s}{1:<50s}{2:<20s}'.format('Id', 'Name', 'Renderer'))\n print('{0:75s}'.format('-'*75))\n for sid, scene in scenes.items():\n print('{0:<5s}{1:<50s}{2:<20s}'.format(str(sid), scene.name, scene.renderer.name))\n print('{0:75s}'.format('-'*75))\n\n\ndef cmd_metrics(args):\n metrics = load_iqa_metrics(iqa_metrics_dir)\n if not metrics:\n print('No IQA metrics.')\n return\n print('-'*60)\n print('{0:<10}{1:<50s}'.format('Acronym', 'Name'))\n print('-'*60)\n for key, metric in metrics.items():\n print('{0:<10s}{1:<50s}'.format(key, metric.name))\n print('-'*60)\n\n\ndef cmd_run(args):\n check = check_configs(configs_dir) and \\\n check_current_config(current_config) and \\\n check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n config_filename = '/tmp/benchmark_config.json'\n config = write_tmp_config(config_filename, current_config, renderers_dir, g_renderers, scenes_dir, g_scenes_names)\n if not config:\n print('Nothing to run.')\n print('Running configuration \\'' + current_config_name(current_config) + '\\'\\n')\n\n benchmark_exec = os.path.join(install_prefix_dir, 'bin/fbksd-benchmark')\n run_techniques(\n benchmark_exec,\n denoisers_dir,\n config['filters'],\n g_filters_names,\n os.path.join(current_slot_dir, 'denoisers'),\n config_filename,\n args.overwrite\n )\n\n run_techniques(\n benchmark_exec,\n samplers_dir,\n config['samplers'],\n g_samplers_names,\n os.path.join(current_slot_dir, 'samplers'),\n config_filename,\n args.overwrite\n )\n\n print('Benchmark finished. Run \\'fbksd results compute\\' to compute result errors.')\n\n\n# computes errors for each result image and saves a corresponding log file for each one.\n# overwrite logs that are older then the corresponding result images\ndef cmd_results_compute(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n load_metrics_g()\n\n exr2png_exec = os.path.join(install_prefix_dir, 'bin/fbksd-exr2png')\n\n if args.all:\n scenes = g_scenes_names\n # NOTE: The compare_techniques_results function computes all results in the slot\n # for the given scenes when the versions arg is empty.\n filters = None\n samplers = None\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = {s.get_name():s for s in config.scenes}\n filters = {v.get_name():v for v in config.filter_versions}\n samplers = {v.get_name():v for v in config.sampler_versions}\n\n print('Comparing results...')\n compare_techniques_results(\n os.path.join(current_slot_dir, 'denoisers'),\n scenes,\n filters,\n scenes_dir,\n exr2png_exec,\n g_iqa_metrics,\n args.overwrite\n )\n\n compare_techniques_results(\n os.path.join(current_slot_dir, 'samplers'),\n scenes,\n samplers,\n scenes_dir,\n exr2png_exec,\n g_iqa_metrics,\n args.overwrite\n )\n\n print('Updating visualization page cache...')\n cmd_page_update(args)\n\n\ndef cmd_results_show(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_metrics_g()\n load_results_g()\n load_samplers_g()\n load_samplers_results_g()\n\n if args.metrics is None:\n metrics = load_iqa_metrics(iqa_metrics_dir).keys()\n else:\n metrics = args.metrics\n\n if args.all:\n scenes = g_scenes.values()\n filters = g_filters_versions.values()\n samplers = g_samplers_versions.values()\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = config.scenes\n filters = config.filter_versions\n samplers = config.sampler_versions\n\n if filters:\n print('DENOISERS')\n print_results(filters, scenes, metrics)\n if samplers:\n print('SAMPLERS')\n print_results(samplers, scenes, metrics)\n\n\ndef cmd_results_rank(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_metrics_g()\n load_results_g()\n load_samplers_g()\n load_samplers_results_g()\n\n metrics = load_iqa_metrics(iqa_metrics_dir).values()\n if args.metrics is not None:\n metrics = [m for m in metrics if m.acronym in args.metrics]\n\n if args.all:\n scenes = g_scenes.values()\n filters = g_filters_versions.values()\n samplers = g_samplers_versions.values()\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = [s.scene for s in config.scenes]\n filters = config.filter_versions\n samplers = config.sampler_versions\n\n rank_techniques('Denoisers Ranking', scenes, filters, metrics)\n rank_techniques('Samplers Ranking', scenes, samplers, metrics)\n\n\ndef cmd_results_export_csv(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n load_metrics_g()\n load_results_g()\n load_samplers_results_g()\n\n if args.all:\n scenes = g_scenes.values()\n filters = g_filters_versions.values()\n samplers = g_samplers_versions.values()\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = [s.scene for s in config.scenes]\n filters = config.filter_versions\n samplers = config.sampler_versions\n\n if args.metrics is None:\n metrics = {k:1.0 for k in load_iqa_metrics(iqa_metrics_dir).keys()}\n else:\n metrics = {}\n regex = r'^([a-zA-Z0-9_-]+)({(\\d*\\.?\\d*)})?$'\n for m in args.metrics:\n match = re.match(regex, m)\n metric = match.group(1)\n scale = match.group(3)\n if scale:\n metrics[metric] = float(scale)\n else:\n metrics[metric] = 1.0\n\n if filters:\n print('Filters')\n print_csv(scenes, filters, metrics)\n if samplers:\n print('Samplers')\n print_csv(scenes, samplers, metrics)\n\n\ndef cmd_page_export(args):\n check = check_scenes(scenes_dir) and \\\n check_filters(denoisers_dir) and \\\n check_samplers(samplers_dir) and \\\n check_results(results_dir) and \\\n check_current_slot(current_slot_dir)\n if not check:\n return\n\n if os.path.isdir(args.dest):\n if args.overwrite:\n print('Overwriting existing files.')\n else:\n print('Destination already exists (see option \\'--overwrite\\').')\n return\n\n load_scenes_g()\n load_filters_g()\n load_samplers_g()\n load_metrics_g()\n load_results_g()\n load_samplers_results_g()\n\n if args.all:\n scenes = g_scenes.values()\n filters = g_filters_versions.values()\n samplers = g_samplers_versions.values()\n scene_config_spps = None\n else:\n check = check_configs(configs_dir) and \\\n check_current_config(current_config)\n if not check:\n return\n config = Config(current_config, g_scenes_names, g_filters_names, g_samplers_names)\n scenes = [s.scene for s in config.scenes]\n filters = config.filter_versions\n samplers = config.sampler_versions\n scene_config_spps = {s.scene.id : s.spps for s in config.scenes}\n\n def copy_results(versions, scene_config_spps, tech_path):\n for version in versions:\n for result in version.results:\n if scene_config_spps:\n if result.scene.id not in scene_config_spps:\n continue\n spps = scene_config_spps[result.scene.id]\n if result.spp not in spps:\n continue\n path = os.path.join(tech_path, version.technique.name, version.tag, result.scene.name)\n dest_path = os.path.join(args.dest, 'data', path)\n if not os.path.exists(dest_path):\n os.makedirs(dest_path)\n\n suffixes = [ os.path.join(path, '{}_0.png'.format(result.spp)) ]\n for metric in result.metrics:\n suffixes.append(os.path.join(path, '{}_0_{}_map.png'.format(result.spp, metric)))\n\n for suffix in suffixes:\n img = os.path.join(current_slot_dir, suffix)\n if not os.path.isfile(img):\n continue\n print('{} -> {}'.format(img, os.path.join(args.dest, 'data', suffix)))\n shutil.copyfile(img, os.path.join(args.dest, 'data', suffix))\n\n # copy filters and samplers result images\n copy_results(filters, scene_config_spps, 'denoisers')\n copy_results(samplers, scene_config_spps, 'samplers')\n # copy data files\n data_dest = os.path.join(args.dest, 'data')\n shutil.copyfile(os.path.join(current_slot_dir, 'iqa_metrics.json'), os.path.join(data_dest, 'iqa_metrics.json'))\n shutil.copyfile(os.path.join(current_slot_dir, 'scenes.json'), os.path.join(data_dest, 'scenes.json'))\n shutil.copyfile(os.path.join(current_slot_dir, 'filters.json'), os.path.join(data_dest, 'filters.json'))\n shutil.copyfile(os.path.join(current_slot_dir, 'results.json'), os.path.join(data_dest, 'results.json'))\n shutil.copyfile(os.path.join(current_slot_dir, 'samplers.json'), os.path.join(data_dest, 'samplers.json'))\n shutil.copyfile(os.path.join(current_slot_dir, 'samplers_results.json'), os.path.join(data_dest, 'samplers_results.json'))\n # copy scenes images\n for scene in scenes:\n orig_gt = os.path.join(scenes_dir, scene.get_reference())\n orig_png = os.path.splitext(orig_gt)[0] + '.png'\n orig_thumb = os.path.splitext(orig_gt)[0] + '_thumb256.jpg'\n dest_gt = os.path.join(args.dest, 'scenes', scene.get_reference())\n dest_path = os.path.dirname(dest_gt)\n dest_png = os.path.splitext(dest_gt)[0] + '.png'\n dest_thumb = os.path.splitext(dest_gt)[0] + '_thumb256.jpg'\n if not os.path.exists(dest_path):\n os.makedirs(dest_path)\n print('{} -> {}'.format(orig_png, dest_png))\n shutil.copyfile(orig_png, dest_png)\n print('{} -> {}'.format(orig_thumb, dest_thumb))\n shutil.copyfile(orig_thumb, dest_thumb)\n # copy page files\n if not args.results_only:\n src = os.path.join(install_prefix_dir, 'share/fbksd/page/')\n dest = args.dest if args.dest[-1] == '/' else args.dest + '/'\n subprocess.run(['rsync', '-a', '--exclude', 'data', '--exclude', 'scenes', src, dest])\n\n\ndef cmd_slots(args):\n if not check_results(results_dir):\n return\n\n slots, current_slot = get_slots(results_dir)\n print('{0}{1:<4s}{2:<16}{3:<20s}'.format(' ', 'Id', 'Creation Date', 'Name'))\n print('{0:75s}'.format('-'*75))\n for i, slot in enumerate(slots):\n pretty_ctime = time.strftime('%d/%m/%Y', time.gmtime(slot['ctime']))\n name = os.path.basename(slot['name'])\n if current_slot == slot['name']:\n print('* {0:<4s}{1:<16}{2:<20s}'.format(str(i+1), pretty_ctime, name))\n else:\n print(' {0:<4s}{1:<16}{2:<20s}'.format(str(i+1), pretty_ctime, name))\n print('{0:75s}'.format('-'*75))\n\n\ndef cmd_slots_new(args):\n if not check_results(results_dir):\n return\n\n invalid_chars = ['.', '/', '\\\\']\n for char in invalid_chars:\n if char in args.name:\n print('ERROR: Slot name can not contain \\'{}\\'.'.format(char))\n return\n\n new_path = os.path.join(results_dir, args.name)\n if os.path.exists(new_path):\n print('ERROR: Slot {} already exists.'.format(args.name))\n return\n\n if not os.path.islink(current_slot_dir):\n print('ERROR: {} is not a symlink.'.format(current_slot_dir))\n return\n\n os.mkdir(new_path)\n if args.select:\n if os.path.islink(current_slot_dir):\n os.unlink(current_slot_dir)\n os.symlink(args.name, current_slot_dir)\n\n\ndef cmd_slots_select(args):\n if not check_results(results_dir):\n return\n\n slots, current_slot = get_slots(results_dir)\n if args.id < 1 or args.id > len(slots):\n print('ERROR: Invalid ID.')\n return\n\n if os.path.isdir(current_slot_dir):\n if not os.path.islink(current_slot_dir):\n print('ERROR: {} is not a symlink.'.format(current_slot_dir))\n return\n os.unlink(current_slot_dir)\n os.symlink(slots[args.id - 1]['name'], current_slot_dir)\n\n\ndef cmd_serve(args):\n check = check_results(results_dir) and \\\n check_current_slot(current_slot_dir) and \\\n check_scenes(scenes_dir)\n if not check:\n return\n\n stdout = sys.stderr\n if not args.no_update:\n print('Updating visualization page cache...')\n cmd_page_update(args)\n orig_page = os.path.join(install_prefix_dir, 'share/fbksd/page/')\n subprocess.run(['rsync', '-a', '--delete', orig_page, page_dir])\n try:\n os.chdir(page_dir)\n Handler = http.server.SimpleHTTPRequestHandler\n httpd = socketserver.TCPServer(('', args.port), Handler)\n print('serving at port {}. Press Ctrl-C to finish.'.format(args.port))\n f = open(os.devnull, 'w')\n sys.stderr = f\n httpd.serve_forever()\n except KeyboardInterrupt:\n sys.stderr = stdout\n print('\\b\\bfinished.')\n\n\n\n#=============================================\n# Main #\n#=============================================\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(prog='fbksd', description='fbksd system cli interface.')\n parser.add_argument('--version', action='version', version='%(prog)s version 2.3.0')\n subparsers = parser.add_subparsers(title='subcommands')\n\n # init\n parserInit = subparsers.add_parser('init', formatter_class=RawTextHelpFormatter,\n help='Initialize a workspace in the current directory.',\n description=\n 'Initialize a workspace in the current directory.\\n\\n'\n 'A workspace is the folder where all the renderers, techniques, and benchmarking data is kept.\\n'\n 'You should always call \\'fbksd\\' having a workspace as working directory.')\n parserInit.add_argument('--slot-name', help='Name of the result slot (default: \\\"Results 1\\\")')\n parserInit.add_argument('--scenes-dir', help='Path to a scenes folder. A link to that folder is created in the current directory.')\n parserInit.set_defaults(func=cmd_init)\n\n # config\n parserConfig = subparsers.add_parser('config', formatter_class=RawTextHelpFormatter,\n help=\"Manage configurations.\",\n description=\n 'A configuration is a set of techniques and scenes to be executed. Each technique will be executed\\n'\n 'for each scene with a set of samples-per-pixel (spp), which can also be configured.\\n'\n 'Subsequent commands like \\'run\\', \\'results\\', etc. will act on the current configuration.\\n\\n'\n 'Each configuration has a corresponding file in the \\'configs\\' folder. You can edit them directly.')\n parserConfig.set_defaults(func=cmd_config)\n configSubparsers = parserConfig.add_subparsers(title='subcommands')\n ## config new\n parserConfigNew = configSubparsers.add_parser('new', formatter_class=RawTextHelpFormatter,\n help='Creates a new configuration.',\n description=\n 'Creates a new configuration with the given name, containing the scenes, techniques,\\n'\n 'and spps provided using the corresponding options.\\n\\n'\n 'The created config is not selected by default. You can select it using the \\'--select\\' option.')\n parserConfigNew.add_argument('name', metavar='NAME', help='Configuration name.')\n parserConfigNew.add_argument('--scenes', nargs='+', type=int, metavar='SCENE_ID', help='Scenes to be included.')\n parserConfigNew.add_argument('--scenes-all', action='store_true', help='Include all scenes.')\n parserConfigNew.add_argument('--filters', nargs='+', type=int, metavar='FILTER_ID', help='Filters to be included.')\n parserConfigNew.add_argument('--filters-all', action='store_true', help='Include all filters.')\n parserConfigNew.add_argument('--samplers', nargs='+', type=int, metavar='SAMPLER_ID', help='Samplers to be included.')\n parserConfigNew.add_argument('--samplers-all', action='store_true', help='Include all samplers.')\n parserConfigNew.add_argument('--spps', metavar='SPP', type=int, nargs='+', help='List of spps to use.')\n parserConfigNew.add_argument('--select', action='store_true', help='Select the created config.')\n parserConfigNew.set_defaults(func=cmd_config_new)\n ## config select\n parserConfigSelect = configSubparsers.add_parser('select', formatter_class=RawTextHelpFormatter,\n help='Select a configuration.',\n description=\n 'Select a configuration given its ID.\\n\\n'\n 'The selected configuration is used by subsequent commands like \\'run\\' and \\'results\\'.')\n parserConfigSelect.add_argument('id', type=int, metavar='CONFIG_ID', help='Id of the desired configuration.')\n parserConfigSelect.set_defaults(func=cmd_config_select)\n ## config show\n parserConfigShow = configSubparsers.add_parser('show', formatter_class=RawTextHelpFormatter,\n help='Show configuration details.',\n description=\n 'Show the contents of a configuration\\'s config file. If no ID is given, the current config is shown.')\n parserConfigShow.add_argument('--id', type=int, metavar='CONFIG_ID', help='Id of the desired configuration.')\n parserConfigShow.set_defaults(func=cmd_config_show)\n ## config add-scenes\n parserConfigAddScenes = configSubparsers.add_parser('add-scenes', formatter_class=RawTextHelpFormatter,\n help='Add scenes to the current config.',\n description='Add scenes to the current config.')\n parserConfigAddScenes.add_argument('scenes', nargs='*', type=int, metavar='SCENE_ID', help='Scene ID.')\n parserConfigAddScenes.add_argument('--spps', metavar='SPP', type=int, nargs='+', required=True, help='spps used for the added scenes.')\n parserConfigAddScenes.add_argument('--all', action='store_true', help='Add all available scenes.')\n parserConfigAddScenes.set_defaults(func=cmd_config_add_scenes)\n ## config rm-scenes\n parserConfigRemoveScenes = configSubparsers.add_parser('rm-scenes', formatter_class=RawTextHelpFormatter,\n help='Remove scenes from the current config.',\n description='Remove scenes from the current config.')\n parserConfigRemoveScenes.add_argument('scenes', nargs='+', type=int, metavar='SCENE_ID', help='Scene ID.')\n parserConfigRemoveScenes.set_defaults(func=cmd_config_remove_scenes)\n ## config add-filters\n parserConfigAddFilters = configSubparsers.add_parser('add-filters', formatter_class=RawTextHelpFormatter,\n help='Add filters to the current config.',\n description='Add filters to the current config.')\n parserConfigAddFilters.add_argument('filters', nargs='*', type=int, metavar='FILTER_ID', help='Filter ID.')\n parserConfigAddFilters.add_argument('--all', action='store_true', help='Add all available filters.')\n parserConfigAddFilters.set_defaults(func=cmd_config_add_filters)\n ## config rm-filters\n parserConfigRemoveFilters = configSubparsers.add_parser('rm-filters', formatter_class=RawTextHelpFormatter,\n help='Remove filters from the current config.',\n description='Remove filters from the current config.')\n parserConfigRemoveFilters.add_argument('filters', nargs='+', type=int, metavar='FILTER_ID', help='Filter ID.')\n parserConfigRemoveFilters.set_defaults(func=cmd_config_remove_filters)\n ## config add-samplers\n parserConfigAddSamplers = configSubparsers.add_parser('add-samplers', formatter_class=RawTextHelpFormatter,\n help='Add samplers to the current config.',\n description='Add samplers to the current config.')\n parserConfigAddSamplers.add_argument('samplers', nargs='*', type=int, metavar='FILTER_ID', help='Sampler ID.')\n parserConfigAddSamplers.add_argument('--all', action='store_true', help='Add all available samplers.')\n parserConfigAddSamplers.set_defaults(func=cmd_config_add_samplers)\n ## config rm-samplers\n parserConfigRemoveSamplers = configSubparsers.add_parser('rm-samplers', formatter_class=RawTextHelpFormatter,\n help='Remove samplers from the current config.',\n description='Remove samplers from the current config.')\n parserConfigRemoveSamplers.add_argument('samplers', nargs='+', type=int, metavar='FILTER_ID', help='Sampler ID.')\n parserConfigRemoveSamplers.set_defaults(func=cmd_config_remove_samplers)\n ## config add-spps\n parserConfigAddSpps = configSubparsers.add_parser('add-spps', formatter_class=RawTextHelpFormatter,\n help='Add spps to the current config.',\n description=\n 'Add spps to the current config.\\n\\n'\n 'By default, the spps are added to all scenes in the config. You can add to specific scenes using\\n'\n 'the \\'--scenes\\' option.')\n parserConfigAddSpps.add_argument('spps', nargs='+', type=int, metavar='SPP', help='spp value.')\n parserConfigAddSpps.add_argument('--scenes', metavar='SCENE_ID', type=int, nargs='+',\n help='Add the spps to specific scenes.')\n parserConfigAddSpps.set_defaults(func=cmd_config_add_spps)\n ## config rm-spps\n parserConfigRemoveSpps = configSubparsers.add_parser('rm-spps', formatter_class=RawTextHelpFormatter,\n help='Remove spps from the current config.',\n description=\n 'Remove spps from the current config.\\n\\n'\n 'By default, the spps are removed from all scenes in the config. You can remove from specific scenes using\\n'\n 'the \\'--scenes\\' option.')\n parserConfigRemoveSpps.add_argument('spps', nargs='+', type=int, metavar='SCENE_ID', help='spps.')\n parserConfigRemoveSpps.add_argument('--scenes', metavar='SCENE_ID', type=int, nargs='+',\n help='Remove the spps from specific scenes.')\n parserConfigRemoveSpps.set_defaults(func=cmd_config_remove_spps)\n\n # filters\n parserFilters = subparsers.add_parser('filters', formatter_class=RawTextHelpFormatter,\n help='List all filters.', description='List all filters.')\n parserFilters.add_argument('--ready', action='store_true', help='List only ready (compiled) denoisers.')\n parserFilters.set_defaults(func=cmd_filters)\n\n # samplers\n parserSamplers = subparsers.add_parser('samplers', formatter_class=RawTextHelpFormatter,\n help='List all samplers.', description=\"List all samplers.\")\n parserSamplers.add_argument('--ready', action='store_true', help='List only ready (compiled) samplers.')\n parserSamplers.set_defaults(func=cmd_samplers)\n\n # scenes\n parserScenes = subparsers.add_parser('scenes', formatter_class=RawTextHelpFormatter,\n help='Manage the scenes cache.',\n description=\n 'Without options, this command list all scenes.\\n\\n'\n 'The scenes are read from the a cache file located in the \\'scenes\\' folder. This file is generated automatically,\\n'\n 'by scanning the \\'scenes\\' folder. You can update the cache using the \\'--update\\' option.\\n\\n'\n 'If you keep you scenes folder somewhere else, you can create a link for it in the workspace using the\\n'\n '\\'--set\\' option.')\n parserScenes.add_argument('--set', metavar='SCENES_DIR', help='Create/change the \\'scenes\\' link to the given scenes folder.')\n parserScenes.add_argument('--update', action='store_true', help='Re-scan the scenes directory and re-generate the scenes cache file.')\n parserScenes.add_argument('--ready', action='store_true', help='List only scenes from ready renderers.')\n parserScenes.set_defaults(func=cmd_scenes)\n\n # metrics\n parserMetrics = subparsers.add_parser('metrics', formatter_class=RawTextHelpFormatter,\n help='List available IQA (Image Quality Assessment) metrics.',\n description='List available IQA (Image Quality Assessment) metrics.')\n parserMetrics.set_defaults(func=cmd_metrics)\n\n # filter-info\n parserFilterInfo = subparsers.add_parser('filter-info', formatter_class=RawTextHelpFormatter,\n help='Show details about filters.', description='Show details about filters.')\n parserFilterInfo.add_argument('filters', metavar='FILTER_ID', type=int, nargs='+', help='Filter ID.')\n parserFilterInfo.set_defaults(func=cmd_filter_info)\n # sampler-info\n parserSamplerInfo = subparsers.add_parser('sampler-info', formatter_class=RawTextHelpFormatter,\n help='Shows details about samplers.', description='Shows details about samplers.')\n parserSamplerInfo.add_argument('samplers', metavar='SAMPLER_ID', type=int, nargs='+', help='Sampler ID.')\n parserSamplerInfo.set_defaults(func=cmd_sampler_info)\n\n # run\n parserRun = subparsers.add_parser('run', formatter_class=RawTextHelpFormatter,\n help='Run the benchmark with the current configuration.',\n description='Run the benchmark with the current configuration.')\n parserRun.set_defaults(func=cmd_run)\n parserRun.add_argument('--overwrite', action='store_true', help='Overwrites previous results.')\n\n # results\n parserResults = subparsers.add_parser('results', help='Manipulate results.')\n resultsSubparsers = parserResults.add_subparsers(title='results subcommands')\n ## results compute\n parserResultsCompute = resultsSubparsers.add_parser('compute',formatter_class=RawTextHelpFormatter,\n help='Compute errors for the saved result images.',\n description=\n 'Compute errors for the saved result images.\\n\\n'\n 'This action should be performed after running a benchmark.\\n'\n 'By default, only results for the current config are computed. You can compute all the results\\n'\n 'in the slot using the \\'--all\\' option.')\n parserResultsCompute.add_argument('--all', action='store_true', help='Compute all results available in the slot.')\n parserResultsCompute.add_argument('--overwrite', action='store_true', dest='overwrite', help='Overwrite previous results.')\n parserResultsCompute.set_defaults(func=cmd_results_compute)\n ## results show\n parserResultsShow = resultsSubparsers.add_parser('show', formatter_class=RawTextHelpFormatter,\n help='Show results.',\n description=\n 'Show results.\\n\\n'\n 'By default, only errors metrics for the current config are shown. You can show all the results\\n'\n 'in the slot using the \\'--all\\' options, and also specify individual error metrics.')\n parserResultsShow.add_argument('--all', action='store_true', help='Show all results available in the slot.')\n parserResultsShow.add_argument('--metrics', metavar='IQA_METRIC', type=str, nargs='+', help='IQA metrics values to show.')\n parserResultsShow.set_defaults(func=cmd_results_show)\n ## results rank\n parserResultsRank = resultsSubparsers.add_parser('rank', formatter_class=RawTextHelpFormatter,\n help='Show the techniques ranking.',\n description=\n 'The ranking is computed by ordering the techniques for each scene/spp/metric and averaging\\n'\n 'all ranks for each technique.\\n\\n'\n 'Techniques with larger rank value are worst, while techniques with rank closer to 1.0 are better.')\n parserResultsRank.add_argument('--all', action='store_true', help='Rank all results available in the workspace.')\n parserResultsRank.add_argument('--metrics', metavar='IQA_METRIC', type=str, nargs='+', help='IQA metrics.')\n parserResultsRank.set_defaults(func=cmd_results_rank)\n ## result print-csv\n parserResultsExportCSV = resultsSubparsers.add_parser('print-csv', formatter_class=RawTextHelpFormatter,\n help='Print CSV tables with the results.',\n description=\n 'Print CSV tables with the results.\\n\\n'\n 'The results are printed in the folowing layout (using \\',\\' to separete cells):\\n\\n'\n '│ │ │ technique 1 │ │ technique 2 │ │\\n'\n '│ │ │ metric 1 │ metric 2 │ metric 1 │ metric 2 │\\n'\n '│ scene 1 │ spp 1 │ e1 │ e2 │ e3 │ e4 │\\n'\n '│ │ spp 2 │ e5 │ e6 │ e7 │ e8 │\\n'\n '│ scene 2 │ spp 1 │ e9 │ e10 │ e11 │ e12 │\\n'\n '│ │ spp 2 │ e13 │ e14 │ e15 │ e16 │\\n\\n'\n 'By default, only results for the current config are included.\\n'\n 'You can use the option \\'--all\\' to include all results available in the slot.\\n\\n'\n 'You cal also choose which IQA metrics are used (all by default) using the \\'--metrics\\' option,\\n'\n 'passing the metrics acronyms. When passing the metrics, you can use the syntax \\'{}\\'\\n'\n 'to specify a scale to be applied to all values of a specific metric: Ex: \\'--metrics RMSE{10}\\' will\\n'\n 'apply a scale of 10 to all RMSE values.')\n parserResultsExportCSV.add_argument('--all', action='store_true', help='Include all available results.')\n parserResultsExportCSV.add_argument('--metrics', metavar='IQA_METRIC', type=str, nargs='+', help='IQA metrics.')\n parserResultsExportCSV.set_defaults(func=cmd_results_export_csv)\n\n # slots\n parserSlots = subparsers.add_parser('slots', help='Manage results slots.')\n parserSlots.set_defaults(func=cmd_slots)\n slotsSubparsers = parserSlots.add_subparsers(title='slots subcommands')\n ## slots new\n parserSlotsNew = slotsSubparsers.add_parser('new', formatter_class=RawTextHelpFormatter,\n help='Create a new results slot.',\n description=\n 'Create a new results slot.\\n\\n'\n 'A slot is basically a folder in the \\'/results\\' directory where all results data\\n'\n '(images, error logs, etc) will be saved.'\n 'The created slot is not selected by default. You can select it using the \\'--select\\' option.')\n parserSlotsNew.add_argument('name', metavar='NAME', help='New slot name.')\n parserSlotsNew.add_argument('--select', action='store_true', help='Select the created slot.')\n parserSlotsNew.set_defaults(func=cmd_slots_new)\n ## slots select\n parserSlotsSelect = slotsSubparsers.add_parser('select', help='Select a slot.')\n parserSlotsSelect.add_argument('id', type=int, metavar='SLOT_ID', help='Id of the desired slot.')\n parserSlotsSelect.set_defaults(func=cmd_slots_select)\n\n # page\n parserPage = subparsers.add_parser('page', formatter_class=RawTextHelpFormatter,\n help=\"Manage the visualization page.\",\n description=\n 'Once you have run a benchmark (\\'fbksd run\\') and computed the result errors (\\'fbksd results compute\\'),\\n'\n 'you can visualize the results using the included visualization page.')\n pageSubparsers = parserPage.add_subparsers(title='subcommands')\n pageSubparsers.required = True\n pageSubparsers.dest = 'subcommand'\n ## page serve\n parserServe = pageSubparsers.add_parser('serve', formatter_class=RawTextHelpFormatter,\n help='Serve the visualization page on a given port (use Ctrl-C to exit).',\n description=\n 'Serve the visualization page on a given port (use Ctrl-C to exit).\\n\\n'\n 'By default, a \\'page update\\' command is performed before serving the page.\\n'\n 'You can disable the update by using the \\'--no-update\\' option.')\n parserServe.add_argument('--port', help='Port number (default is 8000).', type=int, default=8000)\n parserServe.add_argument('--no-update', action='store_true', help='Don\\'t call the \\'page update\\' command.')\n parserServe.add_argument('--all', action='store_true', help='Passes the \\'--all\\' option to the \\'page update\\' command.')\n parserServe.set_defaults(func=cmd_serve)\n ## page export\n parserPageExport = pageSubparsers.add_parser('export', formatter_class=RawTextHelpFormatter,\n help='Export the results and the visualization page.',\n description=\n 'Export the results and the visualization page.\\n\\n'\n 'The exported page can be statically served by any webserver.\\n'\n 'By default, only results for the current config are included.\\n'\n 'You can use the option \\'--all\\' to include all results available in the slot.')\n parserPageExport.add_argument('dest', metavar='DEST', help='Destination folder.')\n parserPageExport.add_argument('--all', action='store_true', help='Export all available results.')\n parserPageExport.add_argument('--overwrite', action='store_true', help='Overwrites DEST if it already exists.')\n parserPageExport.add_argument('--results-only', action='store_true', help='Exports only the results (not the html/javascrip/etc. page files.)')\n parserPageExport.set_defaults(func=cmd_page_export)\n ## page update\n parserPageUpdate = pageSubparsers.add_parser('update', formatter_class=RawTextHelpFormatter,\n help='Update the cache used by the visualization page.',\n description=\n 'Update the cache used by the visualization page.\\n\\n'\n 'The update consists in scanning all computed results and generating the set of files used by\\n'\n 'the visualization page.\\n'\n 'By default, only results for the current config are included.\\n'\n 'You can use the option \\'--all\\' to include all results available in the slot.\\n\\n'\n 'This action is performed automatically when you run a \\'serve\\' or \\'results compute\\'')\n parserPageUpdate.add_argument('--all', action='store_true', help='Include all available results.')\n parserPageUpdate.set_defaults(func=cmd_page_update)\n\n args = parser.parse_args()\n if hasattr(args, 'func'):\n args.func(args)\n else:\n parser.print_help()\n","repo_name":"fbksd/fbksd","sub_path":"share/fbksd/python/fbksd_cli.py","file_name":"fbksd_cli.py","file_ext":"py","file_size_in_byte":55354,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"79"} +{"seq_id":"8593196399","text":"def city_country(city_f, country_f):\n formatted = f\"{city_f}, {country_f}\"\n return formatted.title()\n\n\nwhile True:\n print(\"Enter 'q' at any time for exit\")\n city = input(\"Enter city: \")\n if city == 'q':\n break\n country = input(\"Enter country for city: \")\n if country == 'q':\n break\n print(f\"\\nFormatted string : {city_country(city, country)}\\n\")\n","repo_name":"Lykalon/Python-learning","sub_path":"8_th_chapter/8.6_cities_2.py","file_name":"8.6_cities_2.py","file_ext":"py","file_size_in_byte":384,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28480024842","text":"import os\nimport time\nimport argparse\nimport pandas as pd\nfrom pickle import load, dump\nfrom nltk import corpus\nfrom bm25 import BM25\nfrom tabulate import tabulate\n\npd.options.plotting.backend = \"plotly\"\n\n\ndef get_args():\n parser = argparse.ArgumentParser(description='Information Retrieval system based on BM25.\\n'\n 'This program is a demo of the BM25 algorithm.')\n parser.add_argument('--ndocs', type=int, default=2500,\n help='Maximum number of documents to load from the corpus')\n parser.add_argument('--k1', type=float, default=1.2,\n help='Coefficient in the BM25 formula')\n parser.add_argument('--b', type=float, default=0.75,\n help='Coefficient in the BM25 formula')\n parser.add_argument('--delta', type=float, default=0.0,\n help='Coefficient in the BM25+ formula')\n parser.add_argument('--rf_docs', type=int, default=15,\n help='Number of documents to use for pseudo-relevance feedback')\n parser.add_argument('--rf_terms', type=int, default=10,\n help='Number of terms to use for pseudo-relevance feedback')\n parser.add_argument('--nresults', type=int, default=30,\n help='Number of documents to return')\n parser.add_argument('--expand', type=bool, default=False,\n help='Expand the query with pseudo-relevance feedback')\n parser.add_argument('--load', type=str, default=None,\n help='Path to a pickle file containing a previously-saved BM25 object')\n parser.add_argument('--dump', type=str, default=None,\n help='Path to a pickle file to save the BM25 object')\n parser.add_argument('--plot', type=bool, default=False,\n help='Plot the score distribution')\n args = parser.parse_args()\n return args\n\n\ndef extract_title(text: str) -> str:\n i = 0\n while i < len(text) and text[i].upper() == text[i]:\n i += 1\n i = min(i, 40)\n return text[:i]\n\n\nif __name__ == \"__main__\":\n\n args = get_args()\n\n # Load the corpus\n if args.load is None:\n file_ids = corpus.reuters.fileids()\n data = []\n for i in range(min(args.ndocs, len(file_ids))):\n words = corpus.reuters.words(file_ids[i])\n text = \" \".join(words)\n title = extract_title(text)\n data.append([file_ids[i], title, text])\n df = pd.DataFrame(data, columns=['id', 'title', 'text'])\n bm25 = BM25(df, args.k1, args.b, args.delta,\n args.rf_docs, args.rf_terms)\n if args.dump is not None:\n with open(args.dump, 'wb') as f:\n dump(bm25, f)\n else:\n with open(args.load, 'rb') as f:\n bm25 = load(f)\n\n # Perform queries\n while True:\n os.system('clear')\n query = input(\"Enter a plain-text query: \")\n start = time.time()\n results = bm25.query(query, args.nresults, expand=args.expand)\n end = time.time()\n print(\"\\nQuery time: %.2f seconds\\n\" % (end - start))\n print(tabulate(results, headers=[\n \"id\", \"title\", \"score\"], showindex=False))\n if args.plot:\n results.drop(['id', 'title'], axis=1, inplace=True)\n results.index = list(range(len(results)))\n fig = results.plot.bar()\n fig.update_layout(legend_title_text=\"Legend\")\n fig.update_xaxes(title_text=\"Result Index\")\n fig.update_yaxes(title_text=\"Score\")\n fig.update_layout(title_text=f\"Query Results for '{query}'\")\n fig.show()\n if input(\"\\nPress enter to continue or q to quit: \") == \"q\":\n break\n","repo_name":"AlessandroPierro/BM25","sub_path":"demo.py","file_name":"demo.py","file_ext":"py","file_size_in_byte":3769,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38128390408","text":"from llama_index import download_loader, GPTVectorStoreIndex, ServiceContext, StorageContext, load_index_from_storage\nfrom pathlib import Path\n\nyears = [2022, 2021, 2020, 2019]\nUnstructuredReader = download_loader(\"UnstructuredReader\", refresh_cache=True)\n\nloader = UnstructuredReader()\ndoc_set = {}\nall_docs = []\nfor year in years:\n print(\"loading year: \", year)\n year_docs = loader.load_data(file=Path(f'./data/UBER/UBER_{year}.html'), split_documents=False)\n # insert year metadata into each year\n for d in year_docs:\n d.extra_info = {\"year\": year}\n doc_set[year] = year_docs\n all_docs.extend(year_docs)\n\n# initialize simple vector indices + global vector index\nservice_context = ServiceContext.from_defaults(chunk_size=512)\nindex_set = {}\nfor year in years:\n print(\"building index for year: \", year)\n storage_context = StorageContext.from_defaults()\n cur_index = GPTVectorStoreIndex.from_documents(\n doc_set[year],\n service_context=service_context,\n storage_context=storage_context,\n )\n index_set[year] = cur_index\n print(\"persisting index for year: \", year)\n storage_context.persist(persist_dir=f'./storage/{year}')\n\n\n# Load indices from disk\nindex_set = {}\nfor year in years:\n storage_context = StorageContext.from_defaults(persist_dir=f'./storage/{year}')\n cur_index = load_index_from_storage(storage_context=storage_context)\n index_set[year] = cur_index\n\n","repo_name":"wkmcyz/openai_demo","sub_path":"llama_index_demo/chat_bot.py","file_name":"chat_bot.py","file_ext":"py","file_size_in_byte":1439,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"26275037061","text":"import requests\n\nBASE_URL = \"http://127.0.0.1:8000/\"\nENDPOINT = \"api/updates\"\n\n\ndef get_list():\n r = requests.get(BASE_URL + ENDPOINT)\n data = r.json()\n for obj in data:\n print(obj['id'])\n return data\n\n\nprint(get_list())\n","repo_name":"fabienallade/workSimply","sub_path":"scripts/work_pure_api.py","file_name":"work_pure_api.py","file_ext":"py","file_size_in_byte":240,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18745581892","text":"\"\"\"add table case and tag\n\nRevision ID: 19951205f73c\nRevises: 119bb1687ffa\nCreate Date: 2015-09-23 17:26:22.896415\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '19951205f73c'\ndown_revision = '119bb1687ffa'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.create_table('tag',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('name', sa.String(length=100), nullable=True),\n sa.PrimaryKeyConstraint('id')\n )\n op.create_table('bra_case',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('type', sa.Integer(), nullable=True),\n sa.Column('name', sa.String(length=100), nullable=True),\n sa.Column('url', sa.String(length=300), nullable=True),\n sa.Column('medium_id', sa.Integer(), nullable=True),\n sa.Column('brand', sa.String(length=100), nullable=True),\n sa.Column('industry', sa.String(length=100), nullable=True),\n sa.Column('create_time', sa.DateTime(), nullable=True),\n sa.Column('creator_id', sa.Integer(), nullable=True),\n sa.ForeignKeyConstraint(['creator_id'], ['user.id'], ),\n sa.ForeignKeyConstraint(['medium_id'], ['medium.id'], ),\n sa.PrimaryKeyConstraint('id')\n )\n op.create_table('bra_tag_case',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('tag_id', sa.Integer(), nullable=True),\n sa.Column('case_id', sa.Integer(), nullable=True),\n sa.ForeignKeyConstraint(['case_id'], ['bra_case.id'], ),\n sa.ForeignKeyConstraint(['tag_id'], ['tag.id'], ),\n sa.PrimaryKeyConstraint('id')\n )\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.drop_table('bra_tag_case')\n op.drop_table('bra_case')\n op.drop_table('tag')\n ### end Alembic commands ###\n","repo_name":"giveme168/braavos","sub_path":"migrations/versions/19951205f73c_add_table_case_and_tag.py","file_name":"19951205f73c_add_table_case_and_tag.py","file_ext":"py","file_size_in_byte":1836,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"18081132963","text":"import pandas as pd\n\n# Load the first CSV file\ndf1 = pd.read_csv('Amir/Data/DraftedRB_14-23.csv')\n\n# Load the second CSV file\ndf2 = pd.read_csv('Amir/Data/merged_RB_data.csv')\n\n# Save the modified DataFrame to a new CSV file\ndf1.to_csv('Amir/Data/DraftedRB_14-23.csv', index=False)\n\n# Extract the names to compare from both DataFrames\nnames_to_compare = df1['Names'].tolist()\n\n# Add a new column 'Match' to the second DataFrame and initialize it with 0\ndf2['Match'] = 0\n\n# Iterate through the names in the second DataFrame and update 'Match' to 1 if it matches any name from the first DataFrame\nfor index, row in df2.iterrows():\n if row['player'] in names_to_compare:\n df2.at[index, 'Match'] = 1\n\n# Save the updated second DataFrame to a new CSV file\ndf2.to_csv('Amir/data/merged_RB_data.csv', index=False)\n\nprint(\"Comparison and update complete. The updated data is saved.\")\n","repo_name":"Excaleobur/Capstone","sub_path":"Amir/IsDraftedCheck.py","file_name":"IsDraftedCheck.py","file_ext":"py","file_size_in_byte":886,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"10534889506","text":"__author__ = \"CuiYu\"\n\nimport numpy as np\nimport urllib3\nimport gzip\nimport subprocess\nimport pickle\nimport pandas as pd\nimport datetime\nimport sklearn\nimport matplotlib.pyplot as plt\nimport matplotlib\nimport xgboost as xgb\nimport optuna\nfrom sklearn.metrics import mean_squared_error\nimport torch\nfrom sklearn import preprocessing\n#read times series data\ndef percentile(data,n):\n data2=np.array(data).copy()\n P=np.arange(n)\n category =dict()\n percentage=[a*int(100/(n-1)) for a in np.arange(n-1)][1:]\n for index,p in enumerate(percentage):\n if p==percentage[0]:\n Index=list(set(np.where(data>=np.percentile(np.array(data),0))[0]).intersection( set(np.where(data<=np.percentile(np.array(data),percentage[index+1]))[0])))\n data2[Index]=index\n elif p==percentage[-1]:\n Index=list(set(np.where(data>np.percentile(np.array(data),percentage[index]))[0]).intersection( set(np.where(data<=np.percentile(np.array(data),100))[0])))\n data2[Index]=index\n else:\n Index=list(set(np.where(data>np.percentile(np.array(data),percentage[index]))[0]).intersection( set(np.where(data<=np.percentile(np.array(data),percentage[index+1]))[0])))\n data2[Index]=index\n return data2.tolist()\ndef compute_transition_matrix(data, n, step = 1):\n P = np.zeros((n, n))\n m = len(data)\n for i in range(m):\n initial, final = i, i + step\n if final < m:\n P[int(data[initial])][int(data[final])] += 1\n sums = np.sum(P, axis = 1)\n for i in range(n):\n if sums[i] != 0: # Added this check\n for j in range(n):\n P[i][j] = P[i][j] / sums[i]\n return P\n# 定义GetLoader类,继承Dataset方法,并重写__getitem__()和__len__()方法\nclass GetLoader(torch.utils.data.Dataset):\n # 初始化函数,得到数据\n def __init__(self, data_root, data_label, data_label2):\n self.data = data_root\n self.label = data_label\n self.label2 = data_label2\n # index是根据batchsize划分数据后得到的索引,最后将data和对应的labels进行一起返回\n def __getitem__(self, index):\n data = self.data[index]\n labels = self.label[index]\n labels2 = self.label2[index]\n return data, labels, labels2\n # 该函数返回数据大小长度,目的是DataLoader方便划分,如果不知道大小,DataLoader会一脸懵逼\n def __len__(self):\n return len(self.data)\n\ndef data_to_label(datay1):\n dataylabel=[]\n for x in datay1:\n if np.mean(datay1)+0*np.std(datay1)= 0:\n adjacency_list[vertex.pos].append((vertex.pos[0], vertex.pos[1] - 1))\n # Down\n if vertex.pos[1] + 1 <= self.pathfinder.res[1] // self.pathfinder.vertex_dimension - 1:\n adjacency_list[vertex.pos].append((vertex.pos[0], vertex.pos[1] + 1))\n # Left\n if vertex.pos[0] - 1 >= 0:\n adjacency_list[vertex.pos].append((vertex.pos[0] - 1, vertex.pos[1]))\n # Right\n if vertex.pos[0] + 1 <= self.pathfinder.res[0] // self.pathfinder.vertex_dimension - 1:\n adjacency_list[vertex.pos].append((vertex.pos[0] + 1, vertex.pos[1]))\n\n while algo_state:\n # Quit\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n\n if len(Q) != 0:\n current_vertex_pos = Q.pop(0)\n current_vertex = self.pathfinder.vertex_list[current_vertex_pos[1]][current_vertex_pos[0]]\n # If end is reached\n if current_vertex.colour == END:\n algo_state = False\n self.light_up_path(current_vertex.pos)\n else:\n for adjacent_vertex_pos in adjacency_list[current_vertex_pos]:\n adjacent_vertex = self.pathfinder.vertex_list[adjacent_vertex_pos[1]][adjacent_vertex_pos[0]]\n # If end is reached\n if adjacent_vertex.colour == END:\n adjacent_vertex.pred = current_vertex.pos\n algo_state = False\n self.light_up_path(adjacent_vertex.pos)\n elif adjacent_vertex.colour == UNEXPLORED:\n adjacent_vertex.pred = current_vertex.pos\n adjacent_vertex.colour = EXPLORED\n Q.append(adjacent_vertex_pos)\n else: \n algo_state = False\n\n # Draw all vertices\n self.pathfinder.draw_vertices()\n pygame.display.update()\n time.sleep(0.025)","repo_name":"ChunKaiC/Pathfinder","sub_path":"BFSCommand.py","file_name":"BFSCommand.py","file_ext":"py","file_size_in_byte":3522,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"19724938331","text":"def parse_molecule (formula):\n \n starts=('[','{','(')\n ends=(']','}',')')\n elements=[]\n index=0\n result={}\n \n for item in formula:\n if item in starts:\n if len(elements):\n if elements[-1][0].isalpha():\n elements.append(['1',index,1])\n index+=1\n elements.append(['(',index,0])\n elif item in ends:\n if elements[-1][0].isalpha():\n elements.append(['1',index,1])\n index-=1\n elements.append([')',index,0])\n elif item.isupper():\n if len(elements):\n if elements[-1][0].isalpha():\n elements.append(['1',index,1])\n elements.append([item,index,0])\n elif item.islower():\n elements[-1][0]=elements[-1][0]+item\n elif item.isdigit():\n if elements[-1][0].isdigit():\n elements[-1][0]=elements[-1][0]+item\n elements[-1][2]=int(elements[-1][0])\n else:\n elements.append([item,index,int(item)])\n\n if elements[-1][0].isalpha():\n elements.append(['1',0,1])\n \n for i in range(len(elements)-1,1,-1):\n if elements[i][0].isdigit() and elements[i-1][0]==')':\n i_index=elements[i][1]\n multiplier=elements[i][2]\n isPeak=False\n for j in range(i-1,-1,-1):\n if isPeak and elements[j][1]==i_index:\n break\n if elements[j][1]==i_index+1:\n isPeak=True\n elements[j][2]*=multiplier\n \n \n for i in range(len(elements)):\n if elements[i][0].isalpha():\n if elements[i][0] in result.keys():\n result[elements[i][0]]+=elements[i+1][2]\n else:\n result[elements[i][0]]=elements[i+1][2]\n return dict(sorted(result.items(), key=lambda item: item[0]))","repo_name":"Awerroes569/codewars","sub_path":"Python/molecule_to_atoms.py","file_name":"molecule_to_atoms.py","file_ext":"py","file_size_in_byte":1947,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18454615254","text":"\"\"\"\n This is a script for Internet Of Things Worksheet 3 Task 1\n\"\"\"\n\"\"\"\n import python modules\n\"\"\"\nimport asyncio\nimport websockets\nimport json\nimport time\nimport base64\n\nfrom struct import *\nfrom aioconsole import ainput, aprint\n\nasync def recv_packet(websocket):\n \"\"\"recv_packet(websocket) -> b64decoded_packet\"\"\"\n\n # wait for packet to be received\n packet = await websocket.recv()\n\n # print the base64 packet to meet task 1 requirements\n print(f\"Base64: {packet}\")\n\n # return decoded message\n return base64.b64decode(packet)\n\nasync def recv_and_decode_packet(websocket):\n \"\"\" recv_and_decode_packet(websocket) -> None, awaits for receive of packet from websocket and outputs its contents\"\"\"\n\n # await for packet receive and decode\n packet = await recv_packet(websocket)\n\n # get the length of the packet\n length = int.from_bytes(packet[4:6],'little')\n\n # get the source port\n source_port = int.from_bytes(packet[0:2],'little')\n\n # get the dest port\n dest_port = int.from_bytes(packet[2:4],'little')\n\n # the packet checksum\n checksum = int.from_bytes(packet[6:8],'little')\n\n # content of the packet\n payload = packet[8:(length+8)].decode(\"utf-8\")\n\n print(f\"UDP: {packet}\\nsource-port: {source_port}\\ndest-port: {dest_port}\\ndata-len: {length}\\nchecksum: {checksum}\\npayload: '{payload}'\")\n\n\nasync def main():\n \"\"\"main() -> None, establishes connection with uri using a web socket and awaits for recv_and_decode_packet(websocket)\"\"\"\n\n # uri to connect to the server\n uri = \"ws://localhost:5612\"\n\n # an asynchronous web socket\n async with websockets.connect(uri) as websocket:\n # await for recv and decode function to execute\n await recv_and_decode_packet(websocket)\n\n# begin running async code\nasyncio.get_event_loop().run_until_complete(main())","repo_name":"Auzlex/uwe-iot-worksheet3","sub_path":"Task 1/udp.py","file_name":"udp.py","file_ext":"py","file_size_in_byte":1843,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22648164023","text":"#!/usr/bin/env python2.7\nimport sys\nimport re\nimport subprocess\nimport shlex\nimport snapconf\nimport snapconfshared \nimport snaputil\n\n#for speedy multi-sample ID searching per jx\nimport ahocorasick\n\nRegionArgs = snapconfshared.RegionArgs\ndefault_region_args = snapconfshared.default_region_args\n\nEITHER_START=snapconfshared.EITHER_START\nEITHER_END=snapconfshared.EITHER_END\nreturn_formats = snaputil.return_formats\n\n#return formats:\nTSV=snapconfshared.TSV\nUCSC_BED=snapconfshared.UCSC_BED\nUCSC_URL=snapconfshared.UCSC_URL\n\ndef build_sid_ahoc_queries(sample_ids):\n acs = ahocorasick.Automaton()\n #need to add \",\" and \":\" to make sure we dont match coverage counts or within other IDs\n [acs.add_word(\",\"+sid+\":\", sid) for sid in sample_ids]\n acs.make_automaton()\n return acs\n\nclass RunExternalQueryEngine:\n\n def __init__(self,cmd,qargs,rangeq,snaptron_ids,region_args=default_region_args,additional_cmd=\"\"):\n self.cmd = cmd\n self.qargs = qargs\n self.ra = region_args\n self.snaptron_ids = snaptron_ids\n #this trumps whatever stream_back instructions we were given\n if self.ra.result_count:\n self.ra = ra._replace(stream_back=False)\n self.ra = ra._replace(save_introns=True)\n\n self.filter_by_introns = (self.ra.intron_filter != None and len(self.ra.intron_filter) > 0)\n self.filter_by_samples = (self.ra.sample_filter != None and len(self.ra.sample_filter) > 0)\n\n (header_method,streamer_method) = return_formats[self.ra.return_format]\n self.header = header_method(sys.stdout,region_args=self.ra,interval=self.qargs)\n self.streamer_method = streamer_method\n #exit early as we only want the ucsc_url\n if self.ra.return_format == UCSC_URL:\n return\n\n if cmd == snapconf.TABIX:\n self.delim = '\\t'\n m = snapconf.TABIX_PATTERN.search(self.qargs)\n self.start = int(m.group(2))\n self.end = int(m.group(3))\n self.ra = self.ra._replace(tabix_db_file = \"%s/%s\" % (snapconf.TABIX_DB_PATH,self.ra.tabix_db_file))\n if self.ra.debug:\n sys.stderr.write(\"running %s %s %s\\n\" % (cmd,self.ra.tabix_db_file,self.qargs))\n additional_cmd = ''\n if len(self.ra.additional_cmd) > 0:\n additional_cmd = \" | %s\" % (self.ra.additional_cmd)\n #self.full_cmd = \"%s %s %s | cut -f %d- %s\" % (cmd,self.ra.tabix_db_file,self.qargs,self.ra.cut_start_col,additional_cmd)\n self.full_cmd = \"%s %s %s %s\" % (cmd,self.ra.tabix_db_file,self.qargs,additional_cmd)\n self.range_filters = self.ra.range_filters if self.ra.range_filters is not None and len(self.ra.range_filters) > 0 else None\n self.extern_proc = subprocess.Popen(self.full_cmd, stdout=subprocess.PIPE, shell=True, bufsize=-1)\n\n elif cmd == snapconf.SQLITE:\n self.delim = '\\t'\n arguments = []\n where = []\n (chrom,start,end) = snaputil.sqlite3_interval_query_parse(self.qargs,where,arguments,self.ra)\n self.chrom = chrom\n self.start = start\n self.end = end\n snaputil.sqlite3_range_query_parse(rangeq,where,arguments)\n #force sqlite3 to 3 decimal places\n select_fields = snapconf.INTRON_HEADER_FIELDS\n select_fields[snapconf.CHROM_COL]='chrom'\n select_fields[snapconf.DONOR_COL]='donor'\n select_fields[snapconf.ACCEPTOR_COL]='acceptor'\n select_fields[snapconf.COV_AVG_COL]=\"printf('%.3f',coverage_avg)\"\n select_fields[snapconf.COV_MED_COL]=\"printf('%.3f',coverage_median)\"\n select = \"SELECT %s from intron WHERE %s\" % (\",\".join(select_fields), ' AND '.join(where))\n if self.ra.debug:\n sys.stderr.write(\"%s\\t%s\\n\" % (select,arguments))\n query_ = select\n chr_patt = re.compile('(chr)|[+-]')\n for (i,arg_) in enumerate(arguments):\n arg_ = str(arg_)\n if chr_patt.search(arg_):\n query_ = re.sub('\\?',\"\\'%s\\'\" % arg_,query_,count=1)\n else:\n query_ = re.sub('\\?',arg_,query_,count=1)\n full_cmd_args = [self.cmd, '-separator \\'\t\\'', \"%s.noindices\" % snapconf.SNAPTRON_SQLITE_DB, '\"%s\"' % query_]\n self.full_cmd = \" \".join(full_cmd_args)\n full_cmd_args = shlex.split(self.full_cmd)\n if self.ra.debug:\n sys.stderr.write(\"%s\\n\" % (self.full_cmd))\n self.range_filters = None\n self.extern_proc = subprocess.Popen(full_cmd_args, stdout=subprocess.PIPE, shell=False, bufsize=-1)\n #self.extern_proc = subprocess.Popen(\" \".join(full_cmd_args), stdout=subprocess.PIPE, shell=True, bufsize=-1)\n\n\n def run_query(self):\n ids_found = set()\n sample_set = set()\n #exit early as we only want the ucsc_url\n if self.ra.return_format == UCSC_URL:\n return (ids_found,sample_set)\n for line in self.extern_proc.stdout:\n fields = line.rstrip().split(self.delim)\n snaptron_id = str(fields[snapconf.INTRON_ID_COL])\n lstart = int(fields[self.ra.region_start_col])\n lend = int(fields[self.ra.region_end_col])\n #first attempt to filter by violation of containment (if in effect)\n if self.ra.exact and (lstart != self.start or lend != self.end):\n continue\n #2nd attempt to filter by violation of containment (if in effect)\n if self.ra.contains and (lstart < self.start or lend > self.end):\n continue\n #third attempt to filter by violation of within one end or the other (if in effect)\n if (self.ra.either == EITHER_START and lstart < self.start) or (self.ra.either == EITHER_END and lend > self.end):\n continue\n #now filter, this order is important (filter first, than save ids/print)\n if self.filter_by_introns and snaptron_id not in self.ra.intron_filter and snaptron_id not in self.snaptron_ids:\n continue\n #filter by M (sample IDs), recalculate summaries from subset of samples, and update fields\n samples_found_iter = None\n if self.ra.sid_search_object is not None:\n samples_found_iter = self.ra.sid_search_object.iter(fields[snapconf.SAMPLE_IDS_COL])\n #check to see if this jx has one or more of the sample IDs\n (found_np, fields) = snaputil.extract_sids_and_covs_from_search_iter(samples_found_iter, fields)\n if fields is None:\n continue\n #print fields\n #not used unless testing Tabix or doing a F + M query\n if (self.cmd == snapconf.TABIX or samples_found_iter is not None) and self.range_filters and snaputil.filter_by_ranges(fields,self.range_filters):\n continue\n #combine these two so we only have to split sample <= 1 times\n if self.ra.save_samples:\n samples = set(fields[snapconf.SAMPLE_IDS_COL].split(\",\"))\n #due to start prefixed \",\" delete empty string in set\n if '' in samples:\n samples.remove('')\n sample_set.update(samples)\n #TODO: use samples_found_iter to get a projection of just the samples\n #searched for and then recalculate summary stats\n #filter return stream based on range queries (if any)\n if self.ra.stream_back:\n if samples_found_iter is not None:\n self.streamer_method(sys.stdout,None,fields,region_args=self.ra)\n else:\n self.streamer_method(sys.stdout,line,fields,region_args=self.ra)\n if self.ra.save_introns:\n ids_found.add(snaptron_id)\n exitc=self.extern_proc.wait() \n if exitc != 0:\n raise RuntimeError(\"%s returned non-0 exit code\\n\" % (self.full_cmd))\n return (ids_found, sample_set)\n","repo_name":"ChristopherWilks/snaptron","sub_path":"performance_benchmark/snquery_noindex.py","file_name":"snquery_noindex.py","file_ext":"py","file_size_in_byte":8101,"program_lang":"python","lang":"en","doc_type":"code","stars":24,"dataset":"github-code","pt":"79"} +{"seq_id":"150512947","text":"import json\nimport os\nimport glob\n\n\ndef mergeJson():\n BASE_FILEPATH = \"./api/*\"\n db = dict()\n\n filepaths = glob.glob(BASE_FILEPATH)\n for path in filepaths:\n file = open(path, 'r')\n data = json.load(file)\n filename = os.path.splitext(os.path.basename(path))[0]\n db[filename] = data\n file.close()\n\n MERGED_FILEPATH = \"./db.json\"\n file = open(MERGED_FILEPATH, 'w')\n file.write(json.dumps(db, ensure_ascii=False))\n file.close()\n\n\nmergeJson()","repo_name":"tsushiy/codernote-api-mock","sub_path":"merge_json.py","file_name":"merge_json.py","file_ext":"py","file_size_in_byte":460,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"7252618832","text":"\n\"\"\"\n\nApril 25th. This note added to document change to Test2 Branch\n\nApril 15, 2020\n\nMove Test Model and Axes Example to GraphsTab.py\n\nRename \"Test Area\" to 2L-PR\n\n\"\"\"\n\nfrom tkinter import *\nfrom tkinter.ttk import Notebook\nfrom tkinter import filedialog\nfrom datetime import datetime\nimport DataModel as dm\nimport GraphsTab as gt\nimport TestArea as ta\nimport TextTab as tt\nimport ExcelStuff\nimport stream01\nimport math\nimport os\nimport GraphLib\nimport model\nimport Examples\nimport numpy as np\nimport ListLib\n\nfrom scipy.optimize import curve_fit\nfrom scipy.stats.stats import pearsonr\n\nimport matplotlib\nmatplotlib.use('TkAgg')\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg, NavigationToolbar2Tk\nfrom matplotlib.lines import Line2D\nfrom matplotlib.figure import Figure\nimport matplotlib.patches as patches\nimport matplotlib.lines as lines\nfrom matplotlib import gridspec\nimport matplotlib.pyplot as plt\nfrom matplotlib.ticker import (MultipleLocator, MaxNLocator, FormatStrFormatter, AutoMinorLocator)\nimport matplotlib.ticker as ticker\n\n\"\"\"\nModels, Views and Controllers (MCV) design: keep the representation of the data separate\nfrom the parts of the program that the user interacts with.\n\nView: displays information to the user (Graphical User Interface (tkinter and graphs)\n\nModels: store and retrieve data from databases and files.\nControllers: convert user input into calls on functions that manipulate data\n\"\"\"\n\ndef main(argv=None):\n if argv is None:\n argv = sys.argv\n gui = myGUI()\n gui.go()\n return 0\n\n\nclass myGUI(object):\n def __init__(self):\n \"\"\"\n This object controls all aspects of the Graphical User Interface:\n It uses the Tk tookkit imported from tkinter.\n \"root\" is the base level; all other frames and widgets are in relation to \"root\".\n\n Note that widgets can be called from Tk or ttk. Here the default is to Tk widgets. \n\n \"\"\"\n self.version = \"Analysis\"\n self.root = Tk()\n self.root.title(self.version)\n canvas_width = 800\n canvas_height = 600\n self.initialDir = \"\"\n if (os.name == \"posix\"):\n self.initialDir = \"/Users/daveroberts/Documents\"\n else:\n self.initialDir = \"C:/\"\n print(\"Initial Directory:\", self.initialDir)\n\n # **********************************************************************\n # ********* Variables and Lists associated with the View *********\n # **********************************************************************\n \n #Construct ten empty dataRecords\n self.record0 = dm.DataRecord([],\"empty\") #Data Record defined in DataModel.py \n self.record1 = dm.DataRecord([],\"empty\")\n self.record2 = dm.DataRecord([],\"empty\")\n self.record3 = dm.DataRecord([],\"empty\")\n self.record4 = dm.DataRecord([],\"empty\")\n self.record5 = dm.DataRecord([],\"empty\")\n self.record6 = dm.DataRecord([],\"empty\")\n self.record7 = dm.DataRecord([],\"empty\")\n self.record8 = dm.DataRecord([],\"empty\")\n self.record9 = dm.DataRecord([],\"empty\")\n \n # Create a list of these dataRecords so that one can be \"selected\" with self.fileChoice.get()\n self.recordList = [self.record0,self.record1,self.record2,self.record3,self.record4, \\\n self.record5,self.record6,self.record7,self.record8,self.record9]\n\n # Header row\n self.showOn_tkCanvas = BooleanVar(value = True) # Either tk Canvas or pyplot\n self.clockTimeStringVar = StringVar(value=\"0:00:00\")\n \n self.fileChoice = IntVar(value=0)\n self.fileName0 = StringVar(value = self.recordList[0].fileName)\n self.fileName1 = StringVar(value = self.recordList[1].fileName)\n self.fileName2 = StringVar(value = self.recordList[2].fileName)\n self.fileName3 = StringVar(value = self.recordList[3].fileName)\n self.fileName4 = StringVar(value = self.recordList[4].fileName)\n self.fileName5 = StringVar(value = self.recordList[5].fileName)\n self.fileName6 = StringVar(value = self.recordList[6].fileName)\n self.fileName7 = StringVar(value = self.recordList[7].fileName)\n self.fileName8 = StringVar(value = self.recordList[8].fileName)\n self.fileName9 = StringVar(value = self.recordList[9].fileName)\n \n self.fileNameList = [self.fileName0,self.fileName1,self.fileName2,self.fileName3,self.fileName4,\\\n self.fileName5,self.fileName6,self.fileName7,self.fileName8,self.fileName9]\n\n # Graphs Tab\n self.showBPVar = BooleanVar(value = True)\n self.max_x_scale = IntVar(value=360)\n self.max_y_scale = IntVar(value=500)\n\n # Threshold stuff\n self.printReportVar = BooleanVar(value = True)\n self.pumpTimes = IntVar() # Use OMNI or M0 pumptimes\n self.pumpTimes.set(0) # Default to OMNI pumptimes\n self.logXVar = BooleanVar(value = True) \n self.logYVar = BooleanVar(value = True)\n self.showPmaxLine = BooleanVar(value = True)\n self.showOmaxLine = BooleanVar(value = True)\n self.manualCurveFitVar = BooleanVar(value = False)\n self.QzeroVar = DoubleVar() # Qzero\n self.alphaVar = DoubleVar() # alpha\n self.k_Var = DoubleVar(value=3.0) # k \n self.rangeBegin = IntVar() # First Point\n self.rangeBegin.set(1)\n self.rangeEnd = IntVar() # Last Point\n self.rangeEnd.set(11)\n self.responseCurveVar = BooleanVar(value = True) # Show Response Curve\n self.respMax = IntVar()\n self.respMax.set(200)\n self.average_TH_FilesVar = BooleanVar(value=False) # Not associated with widget\n\n # Text Tab\n self.startTimeVar = IntVar() # Associated with startTimeScale, initialized to zero \n self.endTimeVar = IntVar() # Associated with endTimeScale, initialized to 360\n self.drugConcStr = StringVar(value=\"5.0\")\n self.weightStr = StringVar(value=\"350\")\n\n # Test Area Tab\n self.leverCount = IntVar() # Associated with L1Button & L2Button\n self.leverCount.set(2) \n\n # ******************************************************************************\n # ************************** Root Frame *********************\n # ******************************************************************************\n \n self.rootFrame = Frame(self.root, borderwidth=2, relief=\"sunken\")\n self.rootFrame.grid(column = 0, row = 0)\n headerFrame= Frame(self.root,borderwidth=2, relief=\"sunken\")\n headerFrame.grid(row=0,column=0,sticky=EW)\n fileSelectorFrame = Frame(self.root, borderwidth=2, relief=\"sunken\")\n fileSelectorFrame.grid(row=1,column=0,sticky=NSEW) \n noteBookFrame = Frame(self.root, borderwidth=2, relief=\"sunken\")\n noteBookFrame.grid(row=2,column=0)\n myNotebook = Notebook(noteBookFrame)\n self.graphTab = Frame(myNotebook)\n self.thresholdTab = Frame(myNotebook)\n self.textTab = Frame(myNotebook)\n self.testAreaTab = Frame(myNotebook)\n myNotebook.add(self.textTab,text = \"Text\")\n myNotebook.add(self.graphTab,text = \"Graphs\")\n myNotebook.add(self.thresholdTab,text = \"Threshold\")\n myNotebook.add(self.testAreaTab,text = \"Test Area\")\n myNotebook.grid(row=0,column=0)\n\n # ************** Header Row ****************** \n openFilesButton = Button(headerFrame, text=\"Open Files\", command= lambda: self.openWakeFiles(\"\")).grid(row=0,column=0, sticky=W) \n spacer1Label = Label(headerFrame, text=\" \").grid(row=0,column=1)\n clockTimeLabel = Label(headerFrame, textvariable = self.clockTimeStringVar).grid(row = 0, column=2)\n spacer2Label = Label(headerFrame, text=\" \").grid(row=0,column=3)\n loadTestButton1 = Button(headerFrame, text=\"1L_PR.str\", command= lambda: \\\n self.openWakeFiles(\"1L_PR.str\")).grid(row=0,column=4,sticky=N, padx = 20)\n loadTestButton2 = Button(headerFrame, text=\"TH_FEATHER.dat\", command= lambda: \\\n self.openWakeFiles(\"TH_FEATHER.dat\")).grid(row=0,column=5,sticky=N, padx = 20)\n loadTestButton3 = Button(headerFrame, text=\"TH_OMNI1.str\", command= lambda: \\\n self.openWakeFiles(\"TH_OMNI1.str\")).grid(row=0,column=6,sticky=N, padx = 20)\n loadTestButton4 = Button(headerFrame, text=\"2L-PR-HD1.str\", command= lambda: \\\n self.openWakeFiles(\"2L-PR-HD1.str\")).grid(row=0,column=7,sticky=N, padx = 20)\n \n spacer2Label = Label(headerFrame, text=\" \").grid(row = 0,column = 8)\n canvasButton = Radiobutton(headerFrame, text = \"tk Canvas\", variable = self.showOn_tkCanvas, value = 1).grid(row = 0, column = 9, sticky = E)\n pyplotButton = Radiobutton(headerFrame, text = \"pyplot \", variable = self.showOn_tkCanvas, value = 0).grid(row = 0, column = 10, sticky = E)\n\n #*************** FileSelectorFrame stuff ****************\n padding = 20\n radiobutton1 = Radiobutton(fileSelectorFrame, textvariable = self.fileName0, variable = self.fileChoice, \\\n value = 0, command =lambda: self.selectList()).grid(column=0, row=2, padx=padding)\n radiobutton2 = Radiobutton(fileSelectorFrame, textvariable = self.fileName1, variable = self.fileChoice, \\\n value = 1, command =lambda: self.selectList()).grid(column=1, row=2,padx=padding)\n radiobutton3 = Radiobutton(fileSelectorFrame, textvariable = self.fileName2, variable = self.fileChoice, \\\n value = 2, command =lambda: self.selectList()).grid(column=2, row=2,padx=padding)\n radiobutton4 = Radiobutton(fileSelectorFrame, textvariable = self.fileName3, variable = self.fileChoice, \\\n value = 3, command =lambda: self.selectList()).grid(column=3, row=2,padx=padding)\n radiobutton5 = Radiobutton(fileSelectorFrame, textvariable = self.fileName4, variable = self.fileChoice, \\\n value = 4, command =lambda: self.selectList()).grid(column=4, row=2,padx=padding)\n radiobutton6 = Radiobutton(fileSelectorFrame, textvariable = self.fileName5, variable = self.fileChoice, \\\n value = 5, command =lambda: self.selectList()).grid(column=0, row=3,padx=padding)\n radiobutton7 = Radiobutton(fileSelectorFrame, textvariable = self.fileName6, variable = self.fileChoice, \\\n value = 6, command =lambda: self.selectList()).grid(column=1, row=3,padx=padding)\n radiobutton8 = Radiobutton(fileSelectorFrame, textvariable = self.fileName7, variable = self.fileChoice, \\\n value = 7, command =lambda: self.selectList()).grid(column=2, row=3,padx=padding)\n radiobutton9 = Radiobutton(fileSelectorFrame, textvariable = self.fileName8, variable = self.fileChoice, \\\n value = 8, command =lambda: self.selectList()).grid(column=3, row=3,padx=padding)\n radiobutton10 = Radiobutton(fileSelectorFrame, textvariable = self.fileName9, variable = self.fileChoice, \\\n value = 9, command =lambda: self.selectList()).grid(column=4, row=3,padx=padding)\n\n\n # *************************************************************\n # ************** Graph Tab *************************\n # *************************************************************\n \n self.columnFrame = Frame(self.graphTab, borderwidth=2, relief=\"sunken\")\n self.columnFrame.grid(column = 0, row = 0, columnspan= 1, sticky=NS)\n \n self.graphButtonFrame = Frame(self.columnFrame, borderwidth=2, relief=\"sunken\")\n self.graphButtonFrame.grid(column = 0, row = 0, sticky=N)\n clearCanvasButton = Button(self.graphButtonFrame, text=\"Clear\", command= lambda: \\\n self.clearGraphTabCanvas()).grid(row=0,column=0,sticky=N)\n cumRecButton = Button(self.graphButtonFrame, text=\"cumulativeRecord()\", command= lambda: \\\n self.cumulativeRecord()).grid(row=2,column=0,sticky=N)\n showBPButton = Checkbutton(self.graphButtonFrame, text = \"show BP\", variable = self.showBPVar, onvalue = True, offvalue = False, \\\n command= lambda:self.cumulativeRecord()).grid(row = 3,column=0)\n eventRecButton = Button(self.graphButtonFrame, text=\"eventRecords()\", command= lambda: \\\n self.eventRecords()).grid(row=4,column=0,sticky=N)\n timeStampButton = Button(self.graphButtonFrame, text=\"timeStamps()\", command= lambda: \\\n self.timeStamps()).grid(row=5,column=0,sticky=N) \n modelButton = Button(self.graphButtonFrame, text=\"cocaineModel()\", command= lambda: \\\n self.cocaineModel()).grid(row=6,column=0,sticky=N)\n histogramButton = Button(self.graphButtonFrame, text=\"histogram()\", command= lambda: \\\n self.histogram()).grid(row=7,column=0,sticky=N)\n\n # ****** IntA Frame ************\n self.graph_IntA_frame = Frame(self.columnFrame, borderwidth=2, relief=\"sunken\")\n self.graph_IntA_frame.grid(column = 0, row = 1)\n IntA_frame_lable = Label(self.graph_IntA_frame, text = \"IntA\").grid(row = 0, column=0)\n IntA_event_button = Button(self.graph_IntA_frame, text=\"eventRecordsIntA()\", command= lambda: \\\n self.eventRecordsIntA()).grid(row=1,column=0,sticky=N)\n IntA_durations_button = Button(self.graph_IntA_frame, text=\"pumpDurationsIntA()\", command= lambda: \\\n self.pumpDurationsIntA()).grid(row=2,column=0,sticky=N) \n IntA_histogram_block_Button = Button(self.graph_IntA_frame, text=\"Histogram (blocks)\", command= lambda: \\\n self.IntAHistogram_blocks()).grid(row=3,column=0,sticky=N)\n IntA_histogram_all_Button = Button(self.graph_IntA_frame, text=\"Histogram (All)\", command= lambda: \\\n self.IntAHistogram_all()).grid(row=4,column=0,sticky=N)\n\n # ****** 2L - PR Frame *********\n self.graph_2LPR_frame = Frame(self.columnFrame, borderwidth=2, relief=\"sunken\")\n self.graph_2LPR_frame.grid(column = 0, row = 2)\n TwoLever_frame_lable = Label(self.graph_2LPR_frame, text = \"2L-PR\").grid(row = 0, column=0)\n TwoLever_CR_button = Button(self.graph_2LPR_frame, text=\"Cum Rec\", command= lambda: \\\n self.TwoLeverCR()).grid(row=1,column=0,sticky=N)\n TwoLever_Test2_button = Button(self.graph_2LPR_frame, text=\"Test 1\", command= lambda: \\\n self.TwoLeverGraphTest1()).grid(row=2,column=0,sticky=N)\n\n # ****** Example Frame *********\n self.graph_example_frame = Frame(self.columnFrame, borderwidth=2, relief=\"sunken\")\n self.graph_example_frame.grid(column = 0, row = 3)\n example_frame_lable = Label(self.graph_example_frame, text = \"Examples\").grid(row = 0, column=0)\n model_example_button = Button(self.graph_example_frame, text=\"Test Model\", command= lambda: \\\n self.testModel()).grid(row=1,column=0,sticky=N)\n axes_example_button = Button(self.graph_example_frame, text=\"Axes\", command= lambda: \\\n self.test()).grid(row=2,column=0,sticky=N)\n\n # ******* Y axis frame *********\n self.graph_YaxisRadioButtonFrame = Frame(self.columnFrame, borderwidth=2, relief=\"sunken\")\n self.graph_YaxisRadioButtonFrame.grid(column = 0, row = 4)\n y_axisButtonLabel = Label(self.graph_YaxisRadioButtonFrame, text = \"Y axis\").grid(row = 0, column=0)\n y_scaleRadiobutton250 = Radiobutton(self.graph_YaxisRadioButtonFrame, text=\"250\", variable=self.max_y_scale, value=250)\n y_scaleRadiobutton250.grid(column = 0, row = 1)\n y_scaleRadiobutton500 = Radiobutton(self.graph_YaxisRadioButtonFrame, text=\"500\", variable=self.max_y_scale, value=500)\n y_scaleRadiobutton500.grid(column = 0, row = 2)\n y_scaleRadiobutton1000 = Radiobutton(self.graph_YaxisRadioButtonFrame, text=\"1000\", variable=self.max_y_scale, value=1000)\n y_scaleRadiobutton1000.grid(column = 0, row = 3)\n y_scaleRadiobutton1500 = Radiobutton(self.graph_YaxisRadioButtonFrame, text=\"1500\", variable=self.max_y_scale, value=1500)\n y_scaleRadiobutton1500.grid(column = 0, row = 4)\n \n # *************************************\n \n self.graph_XaxisRadioButtonFrame = Frame(self.graphTab, borderwidth=2, relief=\"sunken\")\n self.graph_XaxisRadioButtonFrame.grid(column = 1, row = 1, sticky=EW)\n x_axisButtonLabel = Label(self.graph_XaxisRadioButtonFrame, text = \"X axis\").grid(row = 0, column=0)\n\n x_scaleRadiobutton10 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"10\", variable=self.max_x_scale, value=10)\n x_scaleRadiobutton10.grid(column = 1, row = 0)\n x_scaleRadiobutton30 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"30\", variable=self.max_x_scale, value=30)\n x_scaleRadiobutton30.grid(column = 2, row = 0)\n x_scaleRadiobutton60 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"60\", variable=self.max_x_scale, value=60)\n x_scaleRadiobutton60.grid(column = 3, row = 0)\n x_scaleRadiobutton120 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"120\", variable=self.max_x_scale, value=120)\n x_scaleRadiobutton120.grid(column = 4, row = 0)\n x_scaleRadiobutton180 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"180\", variable=self.max_x_scale, value=180)\n x_scaleRadiobutton180.grid(column = 5, row = 0)\n x_scaleRadiobutton360 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"360\", variable=self.max_x_scale, value=360)\n x_scaleRadiobutton360.grid(column = 6, row = 0)\n x_scaleRadiobutton720 = Radiobutton(self.graph_XaxisRadioButtonFrame, text=\"720\", variable=self.max_x_scale, value=720)\n x_scaleRadiobutton720.grid(column = 7, row = 0)\n \n self.graphCanvasFrame = Frame(self.graphTab, borderwidth=2, relief=\"sunken\")\n self.graphCanvasFrame.grid(column = 1, row = 0)\n self.graphCanvas = Canvas(self.graphCanvasFrame, width = canvas_width, height = canvas_height)\n self.graphCanvas.grid(row=0,column=0)\n self.graphCanvas.create_text(100,10,text = \"Graph Canvas\")\n\n\n # *****************************************************************\n # ************** Threshold Tab *************************\n # *****************************************************************\n\n # Two subframes: \n #\n # Column 0\n # thresholdButtonFrame\n # lots of widgets\n # Draw Demand Curve/ Clear Canvas/ OMNI vs M0 etc \n # drawThresholdFrame\n # more widgets\n # Qzero / alpha / k\n # firstPointFrame\n # 1 2 3 4\n # 9 10 11 12\n # Save Figure\n # testStuff2() testStuff3()\n # responseButtonFrame\n # Show Response Curve\n # 25 50 100 200\n #\n # Column 1\n # TH_FigureFrame - tk container\n # self.tkCanvas - drawing space for things loke event records\n # self.matPlotCanvas - container for the MatPlotLib Figure\n # self.figure - the thing that axes and lines are drawn on\n\n self.thresholdButtonFrame = Frame(self.thresholdTab, borderwidth=2, relief=\"sunken\")\n self.thresholdButtonFrame.grid(column = 0, row = 0, sticky=N)\n\n self.thresholdFigureFrame = Frame(self.thresholdTab, borderwidth=2, relief=\"sunken\")\n self.thresholdFigureFrame.grid(column = 1, row = 0, sticky=N)\n\n #************** Threshold Button Frame *****\n \n clearTHCanvasButton = Button(self.thresholdButtonFrame, text=\"Clear Canvas\", \\\n command = lambda: self.clearFigure()).grid(row=0,column=0, columnspan = 2, sticky = EW)\n thresholdButton = Button(self.thresholdButtonFrame, text=\"Draw Demand Curve\", command= lambda: \\\n self.drawThreshold()).grid(row=1,column=0,columnspan = 2)\n CreateReportButton = Checkbutton(self.thresholdButtonFrame, text = \"Print Report\", variable = self.printReportVar, \\\n onvalue = True, offvalue = False).grid(row = 2, column = 0, sticky = W)\n pumpTimeLabel = Label(self.thresholdButtonFrame, text = \"Pump Times\").grid(row=3,column=0,sticky=W)\n pumpTimesOMNI = Radiobutton(self.thresholdButtonFrame, text = \"OMNI\", variable = self.pumpTimes, value = 0).grid(row =4,column = 0, sticky = W)\n pumpTimesM0 = Radiobutton(self.thresholdButtonFrame, text = \"M0 \", variable = self.pumpTimes, value = 1).grid(row = 5,column = 0, sticky = W)\n\n # ********** k slider *********************\n self.kFrame = Frame(self.thresholdButtonFrame, borderwidth=2, relief=\"sunken\")\n self.kFrame.grid(row = 6, column = 0, columnspan=2, sticky=EW)\n k_Label = Label(self.kFrame, text = \"k = \").grid(row=0,column=0,sticky=W)\n self.scale_k = Scale(self.kFrame, orient=HORIZONTAL, length=150, resolution = 0.1, \\\n from_= 0.0, to = 9.9, variable = self.k_Var)\n self.scale_k.grid(row=0,column=1, columnspan = 1,stick = W)\n\n # ********* Log Buttons ******************\n logLogXCheckButton = Checkbutton(self.thresholdButtonFrame, text = \"Log X\", variable = self.logXVar, \\\n onvalue = True, offvalue = False).grid(row = 7, column = 0, sticky = W)\n logLogYCheckButton = Checkbutton(self.thresholdButtonFrame, text = \"Log Y\", variable = self.logYVar, \\\n onvalue = True, offvalue = False).grid(row = 8, column = 0, sticky = W) \n showPmaxCheckButton = Checkbutton(self.thresholdButtonFrame, text = \"Show Pmax line\", variable = self.showPmaxLine, \\\n onvalue = True, offvalue = False).grid(row = 9, column = 0, stick = W)\n showOmaxCheckButton = Checkbutton(self.thresholdButtonFrame, text = \"Show Omax line\", variable = self.showOmaxLine, \\\n onvalue = True, offvalue = False).grid(row = 10, column = 0, sticky = W)\n\n # ********* Start / Stop Buttons ********\n self.startStopFrame = Frame(self.thresholdButtonFrame, borderwidth=2, relief=\"sunken\")\n self.startStopFrame.grid(row = 11, column = 0, columnspan=2, sticky=EW)\n \n self.firstPointFrame = Frame(self.startStopFrame, borderwidth=2, relief=\"sunken\")\n self.firstPointFrame.grid(row = 0, column = 0, sticky=W)\n \n rangeBeginLable = Label(self.firstPointFrame, text=\"First Point\").grid(row=0,column=0, sticky=N)\n cumRecButton1 = Radiobutton(self.firstPointFrame, text = \"1 \", variable= self.rangeBegin, value = 0).grid(row=1,column=0)\n cumRecButton2 = Radiobutton(self.firstPointFrame, text = \"2 \", variable= self.rangeBegin, value = 1).grid(row=2,column=0)\n cumRecButton3 = Radiobutton(self.firstPointFrame, text = \"3 \", variable= self.rangeBegin, value = 2).grid(row=3,column=0)\n cumRecButton4 = Radiobutton(self.firstPointFrame, text = \"4 \", variable= self.rangeBegin, value = 3).grid(row=4,column=0)\n\n self.lastPointFrame = Frame(self.startStopFrame, borderwidth=2, relief=\"sunken\")\n self.lastPointFrame.grid(row = 0, column = 1, sticky=E)\n \n rangeEndLable = Label(self.lastPointFrame, text=\"Last Point\").grid(row=0,column=0, sticky=N)\n cumRecButton1 = Radiobutton(self.lastPointFrame, text = \"9 \", variable= self.rangeEnd, value = 8).grid(row=1,column=0)\n cumRecButton2 = Radiobutton(self.lastPointFrame, text = \"10 \", variable= self.rangeEnd, value = 9).grid(row=2,column=0)\n cumRecButton3 = Radiobutton(self.lastPointFrame, text = \"11 \", variable= self.rangeEnd, value = 10).grid(row=3,column=0)\n cumRecButton4 = Radiobutton(self.lastPointFrame, text = \"12 \", variable= self.rangeEnd, value = 11).grid(row=4,column=0)\n\n # Responses Curve ***********************\n responseButtonFrame = Frame(self.thresholdButtonFrame, borderwidth=2, relief=\"sunken\")\n responseButtonFrame.grid(row = 12, column = 0, sticky = EW)\n responseCurveCheckButton = Checkbutton(responseButtonFrame, text = \"Show Response Curve\", variable = self.responseCurveVar, \\\n onvalue = True, offvalue = False).grid(row = 0, column = 0, columnspan = 2)\n respMaxLable = Label(responseButtonFrame, text=\"Responses (Y Scale)\").grid(row=1,column = 0, columnspan = 2, sticky=(N))\n respMaxButton1 = Radiobutton(responseButtonFrame, text = \"25 \", variable=self.respMax, value = 25).grid(row=2,column=0)\n respMaxButton2 = Radiobutton(responseButtonFrame, text = \"50 \", variable=self.respMax, value = 50).grid(row=3,column=0)\n respMaxButton3 = Radiobutton(responseButtonFrame, text = \"100 \", variable=self.respMax, value = 100).grid(row=2,column=1)\n respMaxButton4 = Radiobutton(responseButtonFrame, text = \"200 \", variable=self.respMax, value = 200).grid(row=3,column=1)\n\n \n test2Button = Button(self.thresholdButtonFrame, text=\"Save Figure.png\", command= lambda: \\\n self.save_TH_Figure()).grid(row=13,column=0,sticky=S)\n test3Button = Button(self.thresholdButtonFrame, text=\"testStuff2()\", command= lambda: \\\n self.testStuff2()).grid(row=14,column=0,sticky=S)\n test4Button = Button(self.thresholdButtonFrame, text=\"testStuff3()\", command = lambda: \\\n self.testStuff3()).grid(row=15,column=0,sticky=N)\n\n #************* drawThresholdFrame within thresholdButtonFrame ******** \n self.manualFrame = Frame(self.thresholdButtonFrame, borderwidth=2, relief=\"sunken\")\n self.manualFrame.grid(row = 16, column = 0, columnspan = 1, sticky = EW)\n \n curveFitCheckButton = Checkbutton(self.manualFrame, text = \"Manual Curve Fit\", \\\n variable = self.manualCurveFitVar, onvalue = True, \\\n offvalue = False).grid(row = 0, column = 0, columnspan = 2, stick=W)\n\n self.QzeroLabel = Label(self.manualFrame, text = \"Qzero\").grid(row=1,column=0,sticky=EW)\n self.alphaLabel = Label(self.manualFrame, text = \"alpha\").grid(row=1,column=1,sticky=EW) \n self.scale_Q_zero = Scale(self.manualFrame, orient=HORIZONTAL, length=150, resolution = 0.05, \\\n from_=0.25, to=5.0, variable = self.QzeroVar)\n self.scale_Q_zero.grid(row=2,column=0, columnspan = 1)\n self.scale_Q_zero.set(1.0)\n self.scale_alpha = Scale(self.manualFrame, orient=HORIZONTAL, length=150, resolution = 0.00025, \\\n from_= 0.0005, to = 0.02, variable = self.alphaVar)\n self.scale_alpha.grid(row=2,column=1, columnspan = 1)\n self.scale_alpha.set(0.005)\n \n\n #****************\n\n # TH_FigureFrame - tk container (a Frame)\n # self.matPlotFigure - the thing that axes and lines are drawn on \n # self.threshold_matPlot_Canvas - container for the MatPlotLib Figure\n # - This is the thing that gets redrawn after things are changed.\n\n self.matPlotFigure = Figure(figsize=(7,8), dpi=80)\n \n self.threshold_matPlot_Canvas = FigureCanvasTkAgg(self.matPlotFigure, master=self.thresholdFigureFrame)\n self.threshold_matPlot_Canvas.get_tk_widget().grid(row=0,column=0)\n\n\n # *****************************************************************\n # ******************* Text Tab ******************************\n # *****************************************************************\n\n \n \n #dose report frame\n self.dosereportFrame = Frame(self.textTab, borderwidth=1, relief=\"solid\") # add , relief=\"sunken\" if you want sunken border\n self.dosereportFrame.grid(column = 0, row = 0, sticky=S)\n \n \n self.textButtonFrame = Frame(self.textTab, borderwidth=1)\n self.textButtonFrame.grid(column = 0, row = 0, sticky=N)\n\n self.textBox = Text(self.textTab, width=100, height=47)\n self.textBox.grid(column = 1, row = 0, rowspan = 2)\n \n cleartextButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Clear\", command= lambda: \\\n self.clearText()).grid(row=0,column=0,columnspan = 2, sticky=EW)\n summarytextButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Summary\", command= lambda: \\\n self.summaryText()).grid(row=1,column=0,columnspan = 2,sticky=EW)\n TH_text_button = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"IntA/TH Text\", command= lambda: \\\n self.IntA_TH_text()).grid(row = 3,column = 0, columnspan = 2,sticky=EW)\n## injectionTimesButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Injection Times\", command= lambda: \\\n## self.injectionTimesText()).grid(row=2,column=0,columnspan = 2,sticky=EW)\n## #SeaChange 03012020\n## injectionTimesMinButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Injection Times (min only)\", command= lambda: \\\n## self.injectionTimesTextMin()).grid(row=13,column=0,columnspan = 2,sticky=EW)\n## #SeaChange 03012020\n## intervalButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Interval Times (min only)\", command= lambda: \\\n## self.intervalText()).grid(row=14,column=0,columnspan = 2,sticky=EW)\n## #SeaChange 03162020\n## bintimebutton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Bin Times (min only)\", command= lambda: \\\n## self.bintimeText()).grid(row=15,column=0,columnspan = 2,sticky=EW)\n\n pyPlotEventButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"PyPlot Event Record\", command= lambda: \\\n self.pyPlotEventRecord()).grid(row=16,column=0,columnspan=2,sticky=EW)\n\n pushInjTimesToExcelButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Push Inj Times To Excel\", command= lambda: \\\n self.pushInjTimesToExcel()).grid(row=17,column=0,columnspan=2,sticky=EW)\n\n pushTHToExcelButton = Button(self.textButtonFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Push TH To Excel\", command= lambda: \\\n self.pushTHToExcel()).grid(row=18,column=0,columnspan=2,sticky=EW)\n\n #*************\n\n doseReportButton = Button(self.dosereportFrame, bg=\"white\", font=('Helvetica', 12, 'bold'), text=\"Dose Report\", command= lambda: \\\n self.doseReport()).grid(row=20,column=0,columnspan = 2,sticky=EW)\n\n self.startTimeLabel = Label(self.dosereportFrame, font=('Helvetica', 10, 'bold'), text = \"T1\").grid(row=5,column=0,sticky=E) \n\n self.startTimeScale = Scale(self.dosereportFrame, orient=HORIZONTAL, length=100, resolution = 1, \\\n from_=0, to=360, variable = self.startTimeVar)\n self.startTimeScale.grid(row=5,column=1)\n self.startTimeScale.set(0)\n\n #entry for T1 here\n self.starttimeEntry = Entry(self.dosereportFrame, width=6,textvariable = self.startTimeVar)\n self.starttimeEntry.grid(row = 6, column = 1)\n self.starttimeEntryLabel = Label(self.dosereportFrame, font=('Helvetica', 12, 'bold'), text = \"T1 Entry\").grid(row=6,column=0,sticky=E)\n\n self.endTimeLabel = Label(self.dosereportFrame, font=('Helvetica', 10, 'bold'), text = \"T2\").grid(row=8,column=0,sticky=E) \n\n self.endTimeScale = Scale(self.dosereportFrame, orient=HORIZONTAL, length=100, resolution = 1, \\\n from_=0, to=360, variable = self.endTimeVar)\n self.endTimeScale.grid(row=8,column=1)\n self.endTimeScale.set(360)\n\n #entry for T2 here\n self.endtimeEntry = Entry(self.dosereportFrame, width=6,textvariable = self.endTimeVar)\n self.endtimeEntry.grid(row = 7, column = 1)\n self.starttimeEntryLabel = Label(self.dosereportFrame, font=('Helvetica', 12, 'bold'), text = \"T2 Entry\").grid(row=7,column=0,sticky=E)\n \n concentrationLabel = Label(self.dosereportFrame, font=('Helvetica', 10, 'bold'), text=\"Conc (mg/ml)\")\n concentrationLabel.grid(row = 9, column = 0, sticky=E)\n \n self.concentrationEntry = Entry(self.dosereportFrame, width=6,textvariable = self.drugConcStr)\n self.concentrationEntry.grid(row = 9, column = 1)\n\n weightLabel = Label(self.dosereportFrame, font=('Helvetica', 10, 'bold'), text=\"Body Wt (gms)\")\n weightLabel.grid(row = 10, column = 0)\n\n self.weightEntry = Entry(self.dosereportFrame, width=6,textvariable = self.weightStr)\n self.weightEntry.grid(row = 10, column = 1)\n\n #***************** 2L-PR stuff **************\n self.text_2LPR_Frame = Frame(self.textTab, borderwidth=5, relief=\"sunken\")\n self.text_2LPR_Frame.grid(row = 8, column = 0, sticky=(N))\n \n TwoLeverTextButton = Button(self.text_2LPR_Frame, text=\"2L-PR Summary\", command= lambda: \\\n self.TwoLeverTextReport()).grid(row=0,column=0,sticky=W)\n TestButton = Button(self.text_2LPR_Frame, text=\"TH Test\", command= lambda: \\\n self.THTest()).grid(row=1,column=0,sticky=W)\n TwoLeverTest2Button = Button(self.text_2LPR_Frame, text=\"2L-PR Test2\", command= lambda: \\\n self.TwoLeverTest2()).grid(row=3,column=0,sticky=\"W\")\n testText1Button = Button(self.text_2LPR_Frame, text=\"Text Formatting Examples\", command= lambda: \\\n self.testText1()).grid(row=4,column=0,columnspan = 2,sticky=N)\n\n # *****************************************************************\n # ******************* Test Area Tab ******************************\n # *****************************************************************\n\n \n # Contains testAreaButtonFrame and testAreaFigureFrame\n #\n # testAreaFigureFrame - tk container (a Frame)\n # self.matPlotTestFigure - the thing that axes and lines are drawn on \n # self.threshold_tk_Canvas - drawing space for things like event records\n # self.testArea_matPlot_Canvas - container for the MatPlotLib Figure\n # - This is the thing that gets redrawn after things are changed.\n\n self.testAreaButtonFrame = Frame(self.testAreaTab, borderwidth=5, relief=\"sunken\")\n self.testAreaButtonFrame.grid(column = 0, row = 0, sticky=N)\n\n self.testAreaFigureFrame = Frame(self.testAreaTab, borderwidth=5, relief=\"sunken\")\n self.testAreaFigureFrame.grid(column = 1, row = 0, sticky=N)\n \n self.matPlotTestFigure = Figure(figsize=(9,8), dpi=80)\n self.matPlotTestFigure.set_edgecolor(\"white\") #see help(colors)\n self.matPlotTestFigure.set_facecolor(\"white\") #Set whether the figure frame (background) is displayed or invisible\n self.matPlotTestFigure.set_frameon(True)\n \n self.testArea_MatPlot_Canvas = FigureCanvasTkAgg(self.matPlotTestFigure, master=self.testAreaFigureFrame)\n self.testArea_MatPlot_Canvas.get_tk_widget().grid(row=0,column=0)\n \n\n Button1 = Button(self.testAreaButtonFrame, text=\"fig1_2L_PR()\", command= lambda: \\\n self.fig1_2L_PR()).grid(row=0,column=0,columnspan=2,sticky=N)\n Button2 = Button(self.testAreaButtonFrame, text=\"TwoLeverFig()\", command= lambda: \\\n self.TwoLeverFig()).grid(row=1,column=0,columnspan=2,sticky=N)\n Button3 = Button(self.testAreaButtonFrame, text=\"MatPlot Event Record\", command= lambda: \\\n self.Record()).grid(row=3,column=0,columnspan=2, sticky=N)\n Button4 = Button(self.testAreaButtonFrame, text=\"bin_HD_Records()\", command= lambda: \\\n self.bin_HD_Records()).grid(row=4,column=0,columnspan=2, sticky=N)\n Button5 = Button(self.testAreaButtonFrame, text=\"bin_HD_10SecCount()\", command= lambda: \\\n self.bin_HD_10SecCount()).grid(row=5,column=0,columnspan=2, sticky=N)\n Button6 = Button(self.testAreaButtonFrame, text=\"Load 2L_PR Files\", command= lambda: \\\n self.load_2L_PR_Files()).grid(row=6,column=0,columnspan=2,sticky=N)\n\n L1Button = Radiobutton(self.testAreaButtonFrame, text = \"1L\", variable = self.leverCount, value = 1).grid(row = 7, column = 0, sticky = E)\n L2Button = Radiobutton(self.testAreaButtonFrame, text = \"2L\", variable = self.leverCount, value = 2).grid(row = 8, column = 0, sticky = E)\n \n # Button5 = Button(self.testAreaButtonFrame, text=\"unused\", command= lambda: \\\n # self.someCommand()).grid(row=5,column=0,columnspan=2,sticky=N)\n\n # *************** End of __init__(self) *************************\n \n\n # ************** Procedures called from Graphs Tab *************\n def clearGraphTabCanvas(self):\n self.graphCanvas.delete('all')\n \n def cumulativeRecord(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n showBP = self.showBPVar.get()\n max_x_scale = self.max_x_scale.get()\n max_y_scale = self.max_y_scale.get() \n gt.cumulativeRecord(aCanvas,aRecord,showBP,max_x_scale,max_y_scale)\n\n def eventRecords(self):\n aCanvas = self.graphCanvas\n aRecordList = self.recordList\n max_x_scale = self.max_x_scale.get()\n gt.eventRecords(aCanvas,aRecordList,max_x_scale)\n\n def timeStamps(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n max_x_scale = self.max_x_scale.get()\n gt.timeStamps(aCanvas,aRecord,max_x_scale)\n\n def cocaineModel(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n max_x_scale = self.max_x_scale.get()\n gt.cocaineModel(aCanvas,aRecord,max_x_scale)\n\n def histogram(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n max_x_scale = self.max_x_scale.get()\n gt.histogram(aCanvas,aRecord,max_x_scale)\n\n def eventRecordsIntA(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n gt.eventRecordsIntA(aCanvas,aRecord)\n\n def pumpDurationsIntA(self):\n aCanvas = self.graphCanvas\n aRecord = self.recordList[self.fileChoice.get()]\n gt.pumpDurationsIntA(aCanvas,aRecord)\n\n # ************** Procedures called from Threshold Tab *************\n\n\n # ************** Procedures called from Text Tab *************\n\n def summaryText(self):\n aTextBox = self.textBox\n aRecord = self.recordList[self.fileChoice.get()]\n tt.summaryText(aTextBox,aRecord)\n\n def injectionTimesText(self):\n aTextBox = self.textBox\n aRecord = self.recordList[self.fileChoice.get()]\n tt.injectionTimesText(aTextBox,aRecord)\n \n def doseReport(self):\n aTextBox = self.textBox\n aRecord = self.recordList[self.fileChoice.get()]\n aStartTimeScale = self.startTimeScale.get()\n aEndTimeScale = self.endTimeScale.get()\n aConcentrationEntry = self.concentrationEntry.get()\n aWeightEntry = self.weightEntry.get()\n tt.doseReport(aTextBox,aRecord,aStartTimeScale,aEndTimeScale,aConcentrationEntry,aWeightEntry)\n\n def IntA_TH_text(self):\n aTextBox = self.textBox\n aRecord = self.recordList[self.fileChoice.get()]\n tt.threshold_text(aTextBox,aRecord)\n\n## def injectionTimesTextMin(self):\n## aTextBox = self.textBox\n## aRecord = self.recordList[self.fileChoice.get()]\n## tt.injectionTimesTextMin(aTextBox,aRecord)\n##\n## def intervalText(self):\n## aTextBox = self.textBox\n## aRecord = self.recordList[self.fileChoice.get()]\n## tt.intervalText(aTextBox,aRecord)\n##\n## def bintimeText(self):\n## aTextBox = self.textBox\n## aRecord = self.recordList[self.fileChoice.get()]\n## tt.bintimeText(aTextBox,aRecord)\n##\n## def injectionTimesText(self):\n## aTextBox = self.textBox\n## aRecord = self.recordList[self.fileChoice.get()]\n## tt.injectionTimesText(aTextBox,aRecord)\n\n def pushInjTimesToExcel(self):\n aRecordList = self.recordList\n ExcelStuff.pushInjTimesToExcel(aRecordList)\n\n def pushTHToExcel(self):\n aRecordList = self.recordList\n ExcelStuff.pushTHToExcel(aRecordList)\n\n\n \n # Steven: Note that the above could be written as a single line:\n #\n # tt.summaryText(self.textBox,self.recordList[self.fileChoice.get())\n #\n # but I find the expanded version (with obvious definitions) a bit easier to read\n\n \n \n # ************** Procedures called from Test Area Tab *************\n\n def fig1_2L_PR(self):\n \"\"\"\n This draws Figure 1 for the 2L-PR-HD paper.\n \"\"\"\n if (self.showOn_tkCanvas.get()):\n aFigure = self.matPlotTestFigure # Previously defined Figure containing matPlotCanvas\n aFigure.clf()\n else:\n aFigure = plt.figure(figsize=(6,6), dpi=150, constrained_layout = False) # Newly instantaited pyplot figure\n\n aRecordList = self.recordList \n ta.fig1_2L_PR(aFigure,aRecordList)\n\n if (self.showOn_tkCanvas.get()):\n self.testArea_MatPlot_Canvas.draw()\n else:\n plt.show()\n\n def TwoLeverFig(self):\n \"\"\"\n This draws a figure for 2L data\n \"\"\"\n if (self.showOn_tkCanvas.get()):\n fig = self.matPlotTestFigure # Previously defined Figure containing matPlotCanvas\n fig.clf()\n else:\n fig = plt.figure(figsize=(6,6), dpi=150, constrained_layout = False) # Newly instantaited pyplot figure\n levers = self.leverCount.get()\n print(\"Number of Levers =\",levers) # Defaults to Two Lever\n aRecord = self.recordList[self.fileChoice.get()] \n max_x_scale = self.max_x_scale.get()\n max_y_scale = self.max_y_scale.get()\n \n ta.TwoLeverFig(fig, aRecord, levers,max_x_scale, max_y_scale)\n\n print(\"A directory: \",self.initialDir)\n\n if (self.showOn_tkCanvas.get()):\n self.testArea_MatPlot_Canvas.draw()\n else:\n plt.show()\n # plt.savefig('/Users/daveroberts/Documents/Two Lever PR/Figures/Figure5/Figure5.eps', format='eps', dpi=1000)\n\n def matPlotEventRecord(self):\n aCanvas = self.testArea_MatPlot_Canvas\n aFigure = self.matPlotTestFigure\n aRecord = self.recordList[self.fileChoice.get()]\n startTime = self.startTimeScale.get()\n endTime = self.endTimeScale.get()\n ta.matPlotEventRecord(aCanvas,aFigure,aRecord,startTime,endTime)\n\n def bin_HD_Records(self):\n if (self.showOn_tkCanvas.get()):\n aFigure = self.matPlotTestFigure # Previously defined Figure containing matPlotCanvas\n aFigure.clf()\n else:\n aFigure = plt.figure(figsize=(9,8), dpi=80, constrained_layout = True) # Newly instantaited pyplot figure\n aRecord = self.recordList[self.fileChoice.get()]\n ta.bin_HD_Records(aFigure,aRecord)\n if (self.showOn_tkCanvas.get()):\n self.testArea_MatPlot_Canvas.draw()\n else:\n plt.show()\n\n def bin_HD_10SecCount(self):\n if (self.showOn_tkCanvas.get()):\n aFigure = self.matPlotTestFigure # Previously defined Figure containing matPlotCanvas\n aFigure.clf()\n else:\n aFigure = plt.figure(figsize=(9,8), dpi=80, constrained_layout = True) # Newly instantaited pyplot figure\n aRecord = self.recordList[self.fileChoice.get()]\n ta.bin_HD_10SecCount(aFigure,aRecord) \n if (self.showOn_tkCanvas.get()):\n self.testArea_MatPlot_Canvas.draw()\n else:\n plt.show()\n\n def load_2L_PR_Files(self):\n \"\"\"\n Load 4 files of subject H383\n 8_H383_Mar_10.str - 1L-FR\n 8_H383_Mar_10.str - 1L-HD\n 8_H383_Mar_10.str - 2L-FR-HD \n 8_H383_Mar_10.str - 2L-PR-HD\n \"\"\"\n self.fileChoice.set(0) \n self.openWakeFiles(\"/Users/daveroberts/Documents/Two Lever PR/Raw Data/Final/H383/8_H383_Mar_10.str\")\n self.fileChoice.set(1) \n self.openWakeFiles(\"/Users/daveroberts/Documents/Two Lever PR/Raw Data/Final/H383/8_H383_Mar_11.str\")\n self.fileChoice.set(2) \n self.openWakeFiles(\"/Users/daveroberts/Documents/Two Lever PR/Raw Data/Final/H383/8_H383_Mar_21.str\")\n self.fileChoice.set(3) \n self.openWakeFiles(\"/Users/daveroberts/Documents/Two Lever PR/Raw Data/Final/H383/8_H383_Mar_27.str\")\n\n # ********************************************************************* \n \n def save_TH_Figure(self):\n \"\"\"\n None > save a \"Figure.png\" in current directory.\n This will overwrite current file. Rename if you want to keep it. \n\n self.fig is defined in myGUI and used as a container in all pyplot plots.\n This procedure saves the current self.fig to Figure.png\n Changing the extension will change the format: eg. \".pdf\" \n\n \"\"\"\n print(\"Saving Figure.png\")\n self.matPlotFigure.savefig('Figure.png')\n\n\n def openWakeFiles(self,filename):\n \"\"\"\n The procedure will read Wake datafiles that originate either from OMNI (.str) or from the\n Feather system (.dat). If a filename is passed to this procedure then it will be opened. This\n is how filename Speed Buttons are handled.\n If no filename (\"\") is passed, then a File Open Dialog is spawned. One or several files can\n be selected and loaded.\n \"\"\" \n fileList = []\n fPath = \"\"\n if filename == \"\":\n fileList = filedialog.askopenfilenames(initialdir=self.initialDir)\n else:\n fileList.append(filename)\n \"\"\"\n filenum = 0\n for file in fileList:\n filenum = filenum + 1\n fName = file[file.rfind('/')+1:]\n fPath = file[0:file.rfind('/')+1]\n print('File ',str(filenum), file)\n self.initialDir = fPath\n print(\"Path =\", self.initialDir)\n \"\"\"\n selected = self.fileChoice.get()\n for fName in fileList:\n if (selected < 10):\n print(\"Selection number:\",selected) \n self.recordList[selected].datalist = []\n name = fName[fName.rfind('/')+1:]\n path = fName[0:fName.rfind('/')+1]\n self.initialDir = path\n # print('path =',path)\n self.recordList[selected].fileName = name\n self.fileNameList[selected].set(name)\n # OMNI pump times\n # self.recordList[selected].TH_PumpTimes = [3.162,1.780,1.000,0.562,0.316,0.188, \\\n # 0.100,0.056,0.031,0.018,0.010,0.0056]\n self.recordList[selected].TH_PumpTimes = [3.160,2.000,1.260,0.790,0.500,0.320, \\\n 0.200,0.130,0.080,0.050,0.030,0.020]\n self.recordList[selected].cocConc = 5.0\n self.recordList[selected].pumpSpeed = 0.025 # Wake default 0.1 mls/4 sec = 0.025 / sec \n # textBox.insert('1.0', name+\" opened \\n\\n\")\n if fName.find(\".str\") > 0:\n self.recordList[selected].datalist = stream01.read_str_file(fName) \n elif fName.find(\".dat\") > 0:\n aFile = open(fName,'r')\n for line in aFile:\n if line[0] == \"#\":\n self.recordList[selected].iniLine = line\n params = line.split()\n print(\"IDstr =\", params[1])\n print(\"Weight =\", params[2])\n print(\"Protocol =\", params[3])\n print(\"FR value =\", params[4])\n print(\"SessionLength =\", params[5])\n print(\"PumpTime =\", params[6])\n print(\"Calculate Pump Time =\", params[7]) \n else:\n pair = line.split()\n pair[0] = int(pair[0])\n self.recordList[selected].datalist.append(pair)\n aFile.close()\n self.recordList[selected].extractStatsFromList()\n\n # ------------ fillLists ---------\n verbose = True\n pumpStarttime = 0\n blockNum = -1 \n pumpOn = False\n leverTotal = 0 \n pumpTimeList = [0,0,0,0,0,0,0,0,0,0,0,0] #Temp list of 12 pairs: price and total pump time\n responseList = [0,0,0,0,0,0,0,0,0,0,0,0]\n \"\"\"\n This procedure assumes the datafile if a Threshold file and fills the\n response and consumption lists accordingly - i.e. 12 bins.\n But a PR daatfile could have many more bins which could throw an error.\n So for now, if the bin number does not count higher than 11.\n\n Eventually, \n\n \"\"\"\n for pairs in self.recordList[selected].datalist:\n if pairs[1] == 'B':\n if blockNum < 11:\n blockNum= blockNum + 1\n elif pairs[1] == 'P':\n pumpStartTime = pairs[0]\n pumpOn = True\n responseList[blockNum] = responseList[blockNum] + 1 # inc Bin_responses\n leverTotal = leverTotal + 1 # using pump for responses\n elif pairs[1] == 'p':\n if pumpOn:\n duration = pairs[0]-pumpStartTime\n if blockNum <= 12:\n pumpTimeList[blockNum] = pumpTimeList[blockNum] + duration\n pumpOn = False\n # else no nothing\n # print(\"responseList = \", responseList)\n consumptionList = [0,0,0,0,0,0,0,0,0,0,0,0]\n mgPerSec = self.recordList[selected].cocConc * (self.recordList[selected].pumpSpeed * 0.001)\n if verbose:\n print(\"Cocaine Conc (mg/ml):\", self.recordList[selected].cocConc)\n print(\"Pump Speed ( mls/msec):\", self.recordList[selected].pumpSpeed)\n print(\"cocaine mg/sec:\", mgPerSec)\n for i in range(12):\n consumptionList[i] = pumpTimeList[i] * mgPerSec\n if consumptionList[i] == 0:\n consumptionList[i] = 0.01 #so as not to have a zero value that would crash in a log function\n totalResp = 0\n totalIntake = 0\n for i in range(12):\n totalResp = totalResp + responseList[i]\n totalIntake = totalIntake + consumptionList[i]\n print('Total Intake = ',totalIntake,'; Total Responses = ',totalResp)\n priceList = [] \n for i in range(12):\n # dosePerResponse = pumptime(mSec) * mg/ml * ml/sec)\n dosePerResponse = self.recordList[selected].TH_PumpTimes[i] * \\\n self.recordList[selected].cocConc * \\\n (self.recordList[selected].pumpSpeed)\n price = round(1/dosePerResponse,2)\n priceList.append(price)\n self.recordList[selected].priceList = priceList\n self.recordList[selected].consumptionList = consumptionList\n self.recordList[selected].responseList = responseList\n\n # ------------- end fillLists -----------------\n print(self.recordList[selected])\n selected = selected + 1\n else:\n print(\"More files selected than spots available\")\n print(\"Path =\", self.initialDir)\n\n\n # ********************** The Controllers ***********************************\n # Controllers converts user input into calls on functions that manipulate data\n # ****************************************************************************\n\n def pyPlotEventRecord(self):\n injNum = 0\n injTimeList = []\n \n aRecord = self.recordList[self.fileChoice.get()]\n for pairs in aRecord.datalist:\n if pairs[1] == 'P': \n injNum = injNum + 1\n injTimeList.append(pairs[0]/60000) # Min\n\n plt.figure(figsize=(9,3))\n plt.subplot(111)\n plt.axis([-0.1,185,0.0,1.0])\n plt.eventplot(injTimeList,lineoffsets = 0, linelengths=1.5)\n plt.show() \n\n def clearFigure(self):\n self.matPlotFigure.clf()\n self.threshold_matPlot_Canvas.draw()\n\n def testStuff2(self):\n print(\"testStuff2\")\n aRecord = self.recordList[self.fileChoice.get()]\n print(aRecord.datalist[0])\n\n def testStuff3(self):\n print(\"testStuff3\")\n\n def drawThreshold(self):\n \"\"\"\n TH_FigureFrame - tk Frame that will contain the Figure\n self.threshold_matPlot_Canvas - container for the MatPlotLib Figure\n - This must be redrawn after things are changed\n self.matPlotFigure - This is the Figure that axes and lines are drawn on\n \"\"\"\n verbose = True\n\n def demandFunction(x,alpha):\n \"\"\"\n Demand function described by Hursh\n \n\n y = np.e**(np.log10(Qzero)+k*(np.exp(-alpha*Qzero*x)-1))\n\n # Some day experiment with a different equation like:\n # y = Qzero * np.e**(-x * alpha)\n \n \"\"\"\n \n Qzero = self.Qzero\n #k = self.k_Var.get()\n\n k = 4.19\n y = 10**(np.log10(Qzero)+k*(np.exp(-alpha*Qzero*x)-1))\n \n return y \n \n if (self.pumpTimes.get() == 0):\n pumpTimesString = \"Using OMNI pumpTimes\"\n TH_PumpTimes = [3.162,1.780,1.000,0.562,0.316,0.188, 0.100,0.056,0.031,0.017,0.010,0.0056]\n else:\n pumpTimesString = \"Using Feather M0 pumpTimes\"\n TH_PumpTimes = [3.160,2.000,1.260,0.790,0.500,0.320, 0.200,0.130,0.080,0.050,0.030,0.020]\n\n\n # Generate a price list based on which pump times were selected\n # This assumes a standard cocaine concentration of 5 mg/ml and a Razel pump with 5 RPM motor\n # The pump speed (0.025 ml/sec) was determined as an average across several pumps being\n # switched intermittently on a PR. Might be worth checking against the total fluid delivered\n # during a TH session.\n # Pump Speed was empirically determined. Razel tables show 0.0275 ml/sec\n priceList = []\n for i in range(12):\n dosePerResponse = TH_PumpTimes[i] * 5.0 * 0.025 # pumptime(mSec) * mg/ml * ml/sec)\n print(dosePerResponse) \n price = round(1/dosePerResponse,2)\n priceList.append(price)\n\n print(priceList)\n\n # Retrieve the consumption and response lists from the selected dataRecord.\n # These lists are extracted from the datafile when initially opened.\n # Not that 0.01 is substituted for zero to avoid errors in log functions.\n\n aDataRecord = self.recordList[self.fileChoice.get()]\n datalist = aDataRecord.datalist # Event record needs this.\n consumptionList = aDataRecord.consumptionList\n responseList = aDataRecord.responseList\n\n # ****************************************************************\n #\n # SUBSTITUTIONS\n #\n # ****************************************************************\n # There is a very small discrepancy in how the mg/resp is calculated in Hursh's spreadsheet\n # Apparently Cody came up with the following mg/resp:\n mgPerRespList = [0.421,0.237,0.133,0.075, 0.041,0.024,0.013,0.0075,0.0041,0.0024,0.0013,0.000715]\n # I've estimated a 12th entry\n # This corresponds to the following pricelist from Hursh's spreadsheet. \n priceList = [2.37,4.21,7.5,13.3,24.2,42.1,75,133.9,241.9,416.7,750,1398]\n\n # substitute the consumptionList that would be used by Steven\n for i in range(12):\n consumptionList[i] = responseList[i] * mgPerRespList[i]\n\n # *****************************************************************\n\n # Truncate the range of priceList and consumptionList according the radio button settings\n startRange = self.rangeBegin.get()\n endRange = self.rangeEnd.get()\n truncPriceList = []\n truncConsumptionList = []\n for t in range(startRange,endRange+1): \n truncPriceList.append(priceList[t])\n truncConsumptionList.append(consumptionList[t])\n\n # Calculate Pmax as price with the highest response rate\n maxResp = max(responseList)\n binNum = 0\n for i in range(len(responseList)):\n if (responseList[i] == maxResp):\n binNum = i\n graphicalPmaxString = \"Pmax (Graphically determine) = \"+str(priceList[binNum])\n \n # Create list of injection times for event record\n injNum = 0\n injTimeList = []\n aRecord = self.recordList[self.fileChoice.get()]\n for pairs in aRecord.datalist:\n if pairs[1] == 'P': \n injNum = injNum + 1\n injTimeList.append(pairs[0]/60000) # Min\n\n firstInjTime = injTimeList[0] \n firstInjString = \"First injection at {0:5.1f} min\".format(firstInjTime)\n\n # Create a list to show the beginnings of each block\n blockNum = 0\n finishTime = 0\n blockTimeList = []\n for pairs in aRecord.datalist:\n if pairs[1] == 'B': \n blockNum = blockNum + 1\n blockTimeList.append(pairs[0]/60000) # Min\n if pairs[1] == 'b':\n finishTime = pairs[0]/60000 + 10\n # The end of the (presumably) eleventh block plus ten minutes\n # It appears that the last block does not have a corresponding \"b\" \n finishTimeString = \"Finish Time = {0:5.1f} (min)\".format(finishTime)\n \n # Sometimes and animal won't start for an hour. One could erase this interval\n # by subtracting the time of the first injection from all times\n \"\"\"\n adjustedTimeList = []\n for t in injTimeList:\n adjustedTimeList.append(t - firstInjTime)\n \"\"\"\n\n # Create the beginnings of the graph\n # The user can control whether the graph is embedded into the program ...\n # Or pushed to a separate window with its own widgets.\n # The graph in the separate window can be saved as a separate publication quality .prn file\n if (self.showOn_tkCanvas.get()):\n fig = self.matPlotFigure # Previously defined Figure containing matPlotCanvas\n fig.clf()\n else:\n fig = plt.figure(figsize=(6,6), dpi=80, constrained_layout = True) # Newly instantaited pyplot figure\n\n # Patch is used to configure colors, lines etc. \n fig.patch.set_facecolor(\"azure\")\n fig.patch.set_edgecolor(\"blue\") \n fig.patch.set_linewidth(5.0) # 0.5 would be very thin\n fig.set_frameon(True) # Set whether the figure background is displayed or not\n # fig.suptitle(\"Threshold Title\", fontsize = 16, x = 0.2, y = 0.94)\n # gridspec allows greater control over the positioning of plots\n from matplotlib import gridspec \n gs = gridspec.GridSpec(nrows = 20, ncols= 1)\n\n # ************** EVENT RECORD **********************\n # Create a subplot for event record\n eventRecord = fig.add_subplot(gs[0,0],label=\"1\") # row [0] and col [0]]\n eventRecord.axes.get_yaxis().set_visible(False)\n \n eventRecord.text(0.5, 3, aDataRecord.fileName, ha = 'center', transform=eventRecord.transAxes, \\\n fontsize=12, color='black')\n \n eventRecord.set_ylabel('')\n eventRecord.set_yticklabels(\"\") # Suppress tick labels\n eventRecord.set_xlabel('Time (minutes)')\n eventRecord.spines[\"top\"].set_color('none')\n eventRecord.spines[\"left\"].set_color('none')\n eventRecord.spines[\"right\"].set_color('none')\n startTime = 1 # could start from time of second bin\n eventRecord.set_xlim(startTime, finishTime) \n eventRecord.set_ylim(0, 3)\n eventRecord.eventplot(injTimeList,lineoffsets = 0, linelengths=3)\n eventRecord.eventplot(blockTimeList,lineoffsets = 0, linelengths=5, color = 'red')\n\n # ************** DEMAND CURVE *************************\n\n # The k value is derived from the slider. Defaults to 3 (see definition on line 232)\n # Roberts thinks there is no justification for the function to have this parameter.\n\n k = self.scale_k.get() \n kString = \"Using k from slider = {0:4.1f}\".format(k)\n\n # The manual curve fit checkbox allows users to play with k and alpha\n # If unchecked, it uses curvit to calculate alpha\n if (self.manualCurveFitVar.get() == True):\n\n # Override for now and use Steven's values:\n self.Qzero = 1.0\n QzeroString = \"Qzero (mean of bins 1..3) = {0:6.3f}\".format(self.Qzero)\n self.alpha = 0.00060899\n alphaString = \"alpha (curve fit) = {0:7.5f}\".format(self.alpha)\n k = 4.190\n kString = \"Using k from Hursh = {0:4.3f}\".format(k)\n\n \"\"\"\n self.alpha = self.alphaVar.get()\n alphaString = \"alpha (from slider) = {0:7.5f}\".format(self.alpha)\n self.Qzero = self.QzeroVar.get()\n QzeroString = \"Qzero (from slider) = {0:6.3f}\".format(self.Qzero)\n \"\"\"\n else:\n # Calculate Qzero as the average of the first three bins\n self.Qzero = (consumptionList[1]+consumptionList[2]+consumptionList[3])/3\n QzeroString = \"Qzero (mean of bins 1..3) = {0:6.3f}\".format(self.Qzero)\n # Fit the curve - find alpha\n param_bounds=([0.001],[0.02])\n fitParams, fitCovariances = curve_fit(demandFunction, truncPriceList, truncConsumptionList, bounds=param_bounds)\n self.alpha = fitParams[0]\n alphaString = \"alpha (curve fit) = {0:7.5f}\".format(self.alpha)\n #print (fitCovariances)\n\n # Create y values for best fit line\n fitLine = []\n Qzero = self.Qzero\n alpha = self.alpha \n for x in truncPriceList:\n y = demandFunction(x,alpha)\n fitLine.append(y)\n\n # **** Create the demand Curve Plot and set configurations **** \n\n demandCurve = fig.add_subplot(gs[1:9,0],label=\"2\")\n position = [0.15, 0.1, 0.70, 0.6] # X1,Y1,X2,Y2 - lower left corner, top right corner\n demandCurve.set_position(position) \n if (self.logXVar.get() == True):\n demandCurve.set_xscale(\"log\")\n else:\n demandCurve.set_xscale(\"linear\")\n if (self.logYVar.get() == True):\n demandCurve.set_yscale(\"log\")\n yMax = 5\n else:\n demandCurve.set_yscale(\"linear\")\n yMax = 2\n xMin = 1\n xMax = 1500\n yMin = 0.01\n yMax = 10.0\n demandCurve.set_xlim(xMin,xMax)\n demandCurve.set_ylim(yMin,yMax) \n demandCurve.set_ylabel('Consumption', fontsize = 14)\n demandCurve.yaxis.labelpad = 15 # Move label left or right\n demandCurve.set_xlabel('Price', fontsize = 14)\n demandCurve.plot(truncPriceList, fitLine, color ='red') # Draw a loglog line \n demandCurve.scatter(truncPriceList, truncConsumptionList) # and a scatter plot\n\n # Calculate r as an indicator of goodness of fit\n r = pearsonr(truncPriceList,fitLine)\n rString = \"r = {:.3f}, N = {}\".format(r[0],len(truncPriceList))\n\n\n \"\"\"\n Bentxley et al. (2013) offers the following formula for 1st derivative so presumably:\n slope = -alpha*Qzero*x*k*np.exp(-alpha*Qzero*x)\n equivalent to:\n slope = -alpha*Qzero*x*k*np.e**(-alpha*Qzero*x)\n \"\"\"\n\n # If possible, calculate Pmax and Omax\n PmaxFound = False\n OmaxFound = False\n PmaxString = \"Pmax not found\"\n curveFitPmaxString = \"Pmax not found\"\n OmaxString = \"Omax not found\"\n for p in range(10,1500):\n if (PmaxFound != True):\n slope = -np.log(10**k) * Qzero * p * alpha * np.exp(-alpha * Qzero * p)\n \"\"\"\n # Uncomment this section if you want to see it work\n if (slope < -0.98) and (slope > -1.02):\n print(x, slope)\n \"\"\"\n if slope < -1.0:\n Pmax = p \n PmaxFound = True\n curveFitPmaxString = \"Pmax (curve fit) = {0:6.0f}\".format(Pmax)\n if PmaxFound:\n OmaxFound = True\n Omax = demandFunction(Pmax,alpha)\n OmaxString = \"Omax = {0:6.3f}\".format(Omax)\n if self.showPmaxLine.get():\n x = [Pmax,Pmax]\n y = [yMin,Omax]\n pmaxLine = Line2D(x,y, color = 'green')\n PmaxString = \"Pmax = {0:3.0f} \".format(Pmax)\n demandCurve.add_line(pmaxLine)\n demandCurve.text(Pmax, Omax, PmaxString, ha = 'left', color = 'green',transform=demandCurve.transData)\n\n if self.showOmaxLine.get():\n if PmaxFound:\n x = [xMin,Pmax]\n y = [Omax,Omax]\n OmaxLine = Line2D(x,y, color = 'blue')\n demandCurve.add_line(OmaxLine)\n OmaxString = \"Omax = {0:6.4f}\".format(Omax)\n demandCurve.text(3, Omax - 0.05, OmaxString, ha = 'left', color = 'blue',transform=demandCurve.transData)\n \n \"\"\"\n x = [1e0, 1e4]\n y = [Qzero,Qzero]\n QzeroLine = Line2D(x,y, color = 'blue')\n demandCurve.add_line(QzeroLine)\n \"\"\" \n \n # Show responses on second axis\n\n if (self.responseCurveVar.get() == True):\n respCurve = demandCurve.twinx() # create a 2nd axes that shares the same x-axis\n respCurve.set_position(position)\n respCurve.set_ylabel('Responses', fontsize = 14)\n respCurve.yaxis.labelpad = 20 # Move label left or right\n respCurve.set_ylim(0,self.respMax.get()) # Y axis from 0 to 250\n respCurve.plot(priceList,responseList, color = 'black')\n\n # Display stuff on screen\n demandCurve.text(0.05, 0.95, QzeroString, ha = 'left', transform=demandCurve.transAxes)\n demandCurve.text(0.05, 0.91, alphaString, ha = 'left', transform=demandCurve.transAxes)\n demandCurve.text(0.05, 0.87, kString, ha = 'left', transform=demandCurve.transAxes)\n demandCurve.text(0.65, 0.95, rString, ha = 'left', transform=demandCurve.transAxes)\n \n if (self.showOn_tkCanvas.get()):\n self.threshold_matPlot_Canvas.draw()\n else:\n plt.show()\n\n if (self.printReportVar.get()):\n print(\"********************************\")\n print(pumpTimesString)\n print(\"Number of data points plotted = \",len(truncPriceList))\n print(firstInjString)\n print(finishTimeString)\n print(kString)\n print(QzeroString) \n print(alphaString)\n print(graphicalPmaxString)\n print(curveFitPmaxString)\n print(\"priceList\", priceList)\n tempString = \"Consumption: \"\n for i in range(len(consumptionList)):\n tempString = tempString+\"{0:5.3f}, \".format(consumptionList[i])\n print(tempString)\n print(\"responses\", responseList)\n tempString = \"FitLine: \"\n for i in range(len(fitLine)):\n tempString = tempString+\"{0:5.3f}, \".format(fitLine[i])\n print(tempString)\n print(\"********************************\")\n\n # *************** Two Lever ********************\n\n def TwoLeverCR(self):\n\n def draw_bar(x,y, pixel_height, width, color = \"black\"):\n self.graphCanvas.create_line(x, y, x, y-pixel_height, fill=color)\n self.graphCanvas.create_line(x, y-pixel_height, x+width, y - pixel_height, fill=color)\n self.graphCanvas.create_line(x+width, y-pixel_height, x+width, y, fill=color)\n \n self.graphCanvas.delete('all')\n # label = \"TwoLever Cum Rec\"\n # self.graphCanvas.create_text(300,200, text=label)\n aRecord = self.recordList[self.fileChoice.get()]\n # print(aRecord)\n # canvas is 800 x 600\n x_zero = 50\n y_zero = 400\n x_pixel_width = 700 \n y_pixel_height = 350\n x_divisions = 12\n max_x_scale = self.max_x_scale.get()\n if (max_x_scale == 10) or (max_x_scale == 30): x_divisions = 10\n max_y_scale = self.max_y_scale.get()\n y_divisions = 10\n aTitle = aRecord.fileName\n # def cumRecord(aCanvas, x_zero, y_zero, x_pixel_width, y_pixel_height, max_x_scale, max_y_scale, datalist, aTitle, leverChar = 'L')\n self.graphCanvas.create_line(x_zero, y_zero, x_pixel_width, y_zero)\n GraphLib.drawYaxis(self.graphCanvas, x_zero, y_zero, y_pixel_height, max_y_scale, y_divisions, True)\n GraphLib.cumRecord(self.graphCanvas, x_zero, y_zero, x_pixel_width, y_pixel_height, max_x_scale, max_y_scale, \\\n aRecord.datalist, self.showBPVar.get(), aTitle, leverChar = 'J')\n # Get pump times\n binPumpTime = 0\n pumpStarttime = 0\n pumpOn = False \n pumpTimeList = []\n for pairs in aRecord.datalist:\n if pairs[1] == 'B': # Start of Drug Access\n binStartTime = pairs[0]\n elif pairs[1] == 'P':\n pumpStartTime = pairs[0]\n pumpOn = True\n elif pairs[1] == 'p':\n if pumpOn:\n pumpDuration = pairs[0]-pumpStartTime\n binPumpTime = binPumpTime + pumpDuration\n pumpOn = False\n elif pairs[1] == 'b': # End of Drug Access Period\n dataPoint = [binStartTime,binPumpTime]\n pumpTimeList.append(dataPoint)\n binPumpTime = 0\n # print(pumpTimeList)\n # adapted from GraphLib.eventRecord()\n y_zero = 525\n x = x_zero\n y = y_zero\n scale_height = 90\n scale_max = 10000\n x_scaler = x_pixel_width / (max_x_scale*60*1000)\n for pairs in pumpTimeList:\n x = (x_zero + pairs[0] * x_scaler // 1)\n y = (pairs[1]/scale_max * scale_height) // 1\n draw_bar(x,y_zero,y,5)\n if pairs[1] == 0:\n self.graphCanvas.create_text(x, y_zero-100, fill=\"blue\", text = '*') \n # self.graphCanvas.create_line(x, y, newX, y)\n # self.graphCanvas.create_line(newX, y, newX, y-10) \n # x = newX\n GraphLib.drawXaxis(self.graphCanvas, x_zero, y_zero, x_pixel_width, max_x_scale, x_divisions)\n self.graphCanvas.create_text(400, y_zero+50, fill=\"blue\", text = 'Session Time (min)')\n self.graphCanvas.create_text(200, y_zero+65, fill=\"blue\", text = 'asterisk (*) indicates zero pump time during access period ')\n \n def TwoLeverGraphTest1(self):\n label = \"TwoLeverGraphTest1\"\n self.graphCanvas.delete('all')\n self.graphCanvas.create_text(300,200, text=label)\n\n\n def TwoLeverTextReport(self):\n \"\"\"\n As of July 2018, all two lever PR files were collected through OMNI. Therefore\n the data analyzed here would have been read from an \".str\" file through streamIO.read_str _file()\n Codes are as follows: \n A, a = Start and stop of a PR lever block\n B, b = Start and stop of a drug lever block\n L, l = First lever down and up\n J, j = Second lever down and up\n =, . = First lever extend (=) and retract (.)\n -, , = Second lever extend (-) and retract (,)\n T, t = Start and stop of a trial\n S, s = Stimulus light on and off\n P, p = Pump on and off\n E = End of session\n R = Restart session\n\n It appears that the \"j\" (lever \n \n Access to drug is defined by 'B' and 'b'\n Don't know what 'T' and 't' represent\n \n\n \"\"\"\n self.textBox.insert(END,'\\n****************************\\n')\n self.textBox.insert(END,\"Two Lever PR Summary for \"+ self.recordList[self.fileChoice.get()].fileName+ '\\n')\n\n printPumpTimes = True\n # PR Lever\n PR_accessStarts = 0\n PR_accessStops = 0\n PR_BinResponses = 0 # Responses in an access period \n PR_LeverDownResponses = 0 \n PR_LeverUpResponses = 0\n PR_LeverDownTime = 0 # Time of each press\n PR_LeverDuration = 0 # Duration of each press\n PR_LeverTotalTime = 0 # Total of all presses\n PR_LeverDown = False \n \n drugAccessStarts = 0 # Number of Drug access periods started \n drugAccessStops = 0\n drugAccessStartTime = 0\n totalDrugAccessTime = 0\n drugLeverDownResponses = 0\n drugLeverUpResponses = 0\n drugLeverDownTime = 0\n drugLeverDown = False\n totalDrugLeverTime = 0\n minAccessTime = 10000000\n maxAccessTime = 0\n\n # Pump\n pumpStarts = 0\n pumpStops = 0\n pumpStartTime = 0\n pumpOn = False\n binPumpTime = 0\n totalPumpTime = 0\n\n finalRatio = 0\n \n # Total = 0 \n pumpTimeList = []\n responseList = []\n aRecord = self.recordList[self.fileChoice.get()]\n for pairs in aRecord.datalist:\n\n # PR lever block\n if pairs[1] == 'A': PR_accessStarts= PR_accessStarts + 1\n elif pairs[1] == 'a': PR_accessStops= PR_accessStops + 1\n elif pairs[1] == 'J':\n PR_BinResponses = PR_BinResponses + 1\n PR_LeverDownResponses = PR_LeverDownResponses + 1\n PR_LeverDownTime = pairs[0]\n PR_LeverDown = True\n elif pairs[1] == 'j':\n PR_LeverUpResponses = PR_LeverUpResponses + 1\n if PR_LeverDown:\n PR_LeverDuration = pairs[0]-PR_LeverDownTime\n PR_LeverTotalTime = PR_LeverTotalTime + PR_LeverDuration\n PR_LeverDown = False\n \n # Drug access Block\n elif pairs[1] == 'B': # Start of Drug Access - (also end of PR lever access)\n drugAccessStarts = drugAccessStarts + 1\n responseList.append(PR_BinResponses) \n PR_BinResponses = 0 # Clear PR_BinResponses\n drugAccessStartTime = pairs[0]\n elif pairs[1] == 'b': # End of Drug Access Period\n drugAccessStops = drugAccessStops + 1\n pumpTimeList.append(binPumpTime)\n binPumpTime = 0 # Clear binPumpTime\n binAccessTime = pairs[0]-drugAccessStartTime\n #print(binAccessTime)\n if binAccessTime < minAccessTime: minAccessTime = binAccessTime\n if binAccessTime > maxAccessTime: maxAccessTime = binAccessTime\n \n totalDrugAccessTime = totalDrugAccessTime + binAccessTime\n \n # Drug Lever Up/Down\n elif pairs[1] == 'L':\n drugLeverDownResponses = drugLeverDownResponses + 1\n drugLeverDownTime = pairs[0]\n leverDown = True\n elif pairs[1] == 'l':\n drugLeverUpResponses = drugLeverUpResponses + 1\n if leverDown:\n drugLeverDuration = pairs[0]-drugLeverDownTime\n totalDrugLeverTime = totalDrugLeverTime + drugLeverDuration\n leverDown = False\n\n # Pump On/Off \n elif pairs[1] == 'P':\n pumpStarts = pumpStarts + 1\n pumpStartTime = pairs[0]\n pumpOn = True\n elif pairs[1] == 'p':\n # The program will try to stop the pump at the end of the session. This will ignore\n # that command if pump is already off.\n if pumpOn:\n pumpStops = pumpStops +1 \n pumpDuration = pairs[0]-pumpStartTime\n binPumpTime = binPumpTime + pumpDuration\n totalPumpTime = totalPumpTime + pumpDuration\n pumpOn = False\n \n responseList.append(PR_BinResponses) # Non reinforced responses at the end of the session\n self.textBox.insert(END,'PR Access Intervals (Starts, Stops): '+str(PR_accessStarts)+', '+str(PR_accessStops)+'\\n')\n self.textBox.insert(END,\"Total PR Lever Down Responses = \"+str(PR_LeverDownResponses)+'\\n')\n self.textBox.insert(END,\"Total PR Lever Up Responses = \"+str(PR_LeverUpResponses)+'\\n')\n self.textBox.insert(END,\"Total PR Lever Down Time = \"+str(PR_LeverTotalTime)+'\\n')\n if PR_LeverDownResponses > 0:\n PR_LeverAverage = int(PR_LeverTotalTime/PR_LeverDownResponses)\n self.textBox.insert(END,\"Average PR Lever Down Time = \"+str(PR_LeverAverage)+'\\n\\n')\n\n self.textBox.insert(END,\"Drug Access Intervals (Starts, Stops): \"+str(drugAccessStarts)+', '+str(drugAccessStops)+'\\n')\n accessIntervalLength = int((totalDrugAccessTime/drugAccessStarts)/1000)\n self.textBox.insert(END,\"Average Drug Access Interval = \"+str(accessIntervalLength)+' seconds \\n')\n self.textBox.insert(END,\"Minimum Access Duration = \"+str(minAccessTime)+'\\n')\n self.textBox.insert(END,\"Maximum Access Duration = \"+str(maxAccessTime)+'\\n')\n self.textBox.insert(END,\"Total Drug Lever Down Responses = \"+str(drugLeverDownResponses)+'\\n')\n self.textBox.insert(END,\"Total Drug Lever Up Responses = \"+str(drugLeverUpResponses)+'\\n')\n self.textBox.insert(END,\"Total Drug Lever Down Time = \"+str(totalDrugLeverTime)+'\\n\\n')\n\n\n self.textBox.insert(END,\"Total Pump Starts = \"+str(pumpStarts)+'\\n')\n self.textBox.insert(END,\"Total Pump Stops = \"+str(pumpStops)+'\\n')\n self.textBox.insert(END,\"Total Pump Time = \"+str(totalPumpTime)+'\\n')\n if pumpStarts > 0:\n averagePumpDuration = int(totalPumpTime/pumpStarts)\n self.textBox.insert(END,\"Average Pump Duration = \"+str(averagePumpDuration)+'\\n\\n')\n\n\n if drugAccessStarts > 0:\n self.textBox.insert(END,\"Final Ratio = \"+str(responseList[drugAccessStarts-1])+'\\n')\n\n if len(pumpTimeList) > 0:\n self.textBox.insert(END,\"Pumptimes in each bin: \\n\")\n if printPumpTimes:\n for i in range (len(pumpTimeList)):\n self.textBox.insert(END,str(pumpTimeList[i])+'\\n')\n print(pumpTimeList) \n \n def THTest(self):\n self.textBox.insert(\"1.0\",\"THTest\\n\")\n\n def TwoLeverTest2(self):\n self.textBox.insert(\"1.0\",\"TwoLeverTest2\\n\")\n\n # ************ End Two Lever *******************\n\n def testText1(self):\n Examples.showTextFormatExamples(self.textBox)\n\n def demandFunction(self, x, alpha, Qzero):\n k = 2\n y = np.e**(np.log(Qzero)+k*(np.exp(-alpha*Qzero*x)-1))\n return y\n\n def testFunction(self, x, k_value, alpha, Qzero):\n k = k_value\n y = np.exp(np.log(Qzero)+k*(np.exp(-alpha*Qzero*x)-1))\n return y \n \n \n \n def testAxisExamples(self):\n \"\"\"\n Test the drawLogYAxis() function with various parameters\n \"\"\" \n aCanvas = self.thresholdCanvas\n x_zero = 100\n y_zero = 550\n x_pixel_width = 600\n y_pixel_height = 500\n x_startValue = 0.001\n y_startValue = 0.001\n x_logRange = 5\n y_logRange = 5\n x_caption = \"Price (responses/mg cocaine)\"\n y_caption = \"Y\"\n #leftLabel = True\n GraphLib.drawLog_X_Axis(aCanvas,x_zero,y_zero,x_pixel_width,x_startValue,x_logRange,x_caption, test = True)\n GraphLib.drawLog_Y_Axis(aCanvas,x_zero,y_zero,y_pixel_height,y_startValue,y_logRange,y_caption, test = True) \n x_zero = 200\n y_zero = 450\n x_pixel_width = 500\n y_pixel_height = 400\n x_startValue = 0.01\n y_startValue = 0.01\n x_logRange = 5\n logRange = 5\n GraphLib.drawLog_X_Axis(aCanvas,x_zero,y_zero,x_pixel_width,x_startValue,x_logRange,x_caption, test = True)\n GraphLib.drawLog_Y_Axis(aCanvas, x_zero,y_zero,y_pixel_height,y_startValue,y_logRange, y_caption, test = True)\n x_zero = 300\n y_zero = 350\n x_pixel_width = 400\n y_pixel_height = 300\n x_startValue = 0.1\n y_startValue = 0.1\n x_logRange = 2\n y_logRange = 2\n GraphLib.drawLog_X_Axis(aCanvas,x_zero,y_zero,x_pixel_width,x_startValue,x_logRange,x_caption)\n GraphLib.drawLog_Y_Axis(aCanvas,x_zero+x_pixel_width,y_zero,y_pixel_height,y_startValue,y_logRange,y_caption, test = True, leftLabel=False)\n \n## def injectionTimesText(self):\n## aRecord = self.recordList[self.fileChoice.get()]\n## injection = 0\n## previousInjTime = 0\n## self.textBox.insert(END,\"Inj Duration Time (sec) Time (min) Interval (sec)\\n\")\n## pumpOn = False\n## for pairs in aRecord.datalist:\n## if pairs[1] == 'P':\n## pumpStartTime = pairs[0]\n## injection = injection + 1\n## secTime = pairs[0]/1000\n## minTime = secTime/60\n## interval = secTime - previousInjTime\n## previousInjTime = secTime\n## pumpOn = True\n## if pairs[1] == 'p':\n## if pumpOn:\n## pumpOn = False\n## duration = pairs[0]-pumpStartTime\n## if injection == 1:\n## tempString = \"{0} {1:10.2f} {2:10.2f} {3:10.2f}\".format(injection,duration,secTime,minTime,interval)\n## else:\n## tempString = \"{0} {1:10.2f} {2:10.2f} {3:10.2f} {4:10.2f}\".format(injection,duration,secTime,minTime,interval)\n## self.textBox.insert(END,tempString+\"\\n\")\n## \n## self.textBox.insert(END,\"Number of injections: \"+str(injection)+\"\\n\")\n\n## def doseReport(self):\n## aRecord = self.recordList[self.fileChoice.get()]\n## pumpOn = False\n## injections = 0\n## totalPumpDuration = 0\n## lastTime = 0\n## time1 = self.startTimeScale.get()\n## time2 = self.endTimeScale.get()\n## for pairs in aRecord.datalist:\n## if pairs[1] == 'P':\n## minTime = pairs[0]/60000\n## if (minTime >= time1) and (minTime < time2):\n## injections = injections + 1\n## pumpStartTime = pairs[0]\n## lastTime = pumpStartTime\n## pumpOn = True\n## if pairs[1] == 'p':\n## if pumpOn:\n## duration = pairs[0]-pumpStartTime\n## pumpOn = False\n## totalPumpDuration = totalPumpDuration + duration\n## \n## \n## aString = \"Injections between \"+str(time1)+\" and \"+str(time2)+\" minutes = \"+str(injections)+\"\\n\"\n## self.textBox.insert(END,aString)\n##\n## try:\n## conc = float(self.concentrationEntry.get())\n## weight = int(self.weightEntry.get()) # in grams\n## aString = \"Drug Concentration = {0:5.3f} mg/ml\\nWeight = {1:3d} grams \\n\".format(conc,weight) \n## except ValueError:\n## aString = \"Error getting Conc and/or Body weight \\n\"\n## self.textBox.insert(END,aString)\n##\n## if injections > 0:\n## aString = \"Total Pump Duration = {0:6d} mSec \\n\".format(totalPumpDuration)\n## self.textBox.insert(END,aString)\n## averagePumpTime = round(totalPumpDuration / injections,2)\n## aString = \"Average Pump Time = {0:5.3f} mSec \\n\".format(averagePumpTime)\n## self.textBox.insert(END,aString)\n## totalDose = (totalPumpDuration/1000) * conc * 0.025 # pumptime(mSec) * mg/ml * ml/sec)\n## totalDosePerKg = totalDose/(weight/1000)\n## aString = \"Total Dose = {0:5.3f} mg; {1:5.3f} mg/kg \\n\".format(totalDose, totalDosePerKg)\n## self.textBox.insert(END,aString)\n## averageDose = (totalDose / injections)\n## averageDosePerKg = averageDose / (weight/1000)\n## aString = \"Average Dose = {0:5.3f} mg; {1:5.3f} mg/kg \\n\".format(averageDose, averageDosePerKg)\n## self.textBox.insert(END,aString)\n##\n## \n## self.textBox.insert(END,\"********************************\\n\")\n\n def selectList(self):\n \"\"\"\n Dummy function associated with radiobuttons that selects the filename textvariable.\n \"\"\"\n # print(\"fileChoice: \", self.fileChoice.get())\n pass\n\n\n def clearTHCanvas(self):\n self.graphCanvas.delete('all')\n self.fig = plt.figure(clear=True) # clear contents \n \n \n def IntAHistogram_blocks(self):\n '''\n\n '''\n self.clearGraphTabCanvas()\n aRecord = self.recordList[self.fileChoice.get()]\n pump_total = 0\n for b in range (12):\n total_pump_time = 0\n pump_timelist = ListLib.get_pump_duration_list(aRecord.datalist, block = b)\n pumptimes_per_bin = ListLib.get_pumptimes_per_bin(pump_timelist, bin_size = 5000)\n for t in range(len(pumptimes_per_bin)):\n total_pump_time = total_pump_time + pumptimes_per_bin[t]\n pump_total = pump_total + total_pump_time\n y = (b*45)+50\n GraphLib.histogram(self.graphCanvas,pumptimes_per_bin, y_zero = y, y_pixel_height = 35, clear = False)\n self.graphCanvas.create_text(750, y, fill=\"blue\", text = \"Sum = \"+str(total_pump_time))\n dose = (total_pump_time * 5 * 0.000025) / 0.33\n aString = \"{0:6.2f} mg/kg\".format(dose) #eg. 4000 mSec * 5 mg/ml *0.000025 mls/mSec / 0.330 kg = 1.5 mg/kg\n self.graphCanvas.create_text(750, y + 12, fill=\"blue\", text = aString)\n self.graphCanvas.create_text(750, y + 45, fill=\"blue\", text = \"Total \"+str(pump_total))\n\n def IntAHistogram_all(self):\n self.clearGraphTabCanvas()\n aRecord = self.recordList[self.fileChoice.get()]\n pump_total = 0\n x_zero = 75\n y_zero = 550\n x_pixel_width = 600\n y_pixel_height = 400\n max_x_scale = 300\n max_y_scale = 20000\n x_divisions = 5\n y_divisions = 10\n labelLeft = True\n GraphLib.drawYaxis(self.graphCanvas, x_zero, y_zero, y_pixel_height, max_y_scale, y_divisions, labelLeft, \\\n format_int = True, color = \"black\")\n pump_duration_list = ListLib.get_pump_duration_list(aRecord.datalist, block = -1)\n pumptimes_per_bin = ListLib.get_pumptimes_per_bin(pump_duration_list, bin_size = 5000)\n for t in range(len(pumptimes_per_bin)):\n pump_total = pump_total + pumptimes_per_bin[t]\n GraphLib.histogram(self.graphCanvas,pumptimes_per_bin, y_zero = 550, y_pixel_height = 400, clear = False, \\\n max_y_scale = 20000, y_divisions = 4)\n\n self.graphCanvas.create_text(300, 100, fill=\"blue\", text = \"Total Pump Time: \"+str(pump_total))\n dose = (pump_total * 5 * 0.000025) / 0.33\n aString = \"Total: {0:6.2f} mg/kg\".format(dose) # Format float to 2 decimal points in 6 character field\n self.graphCanvas.create_text(300, 130, fill=\"blue\", text = aString) \n\n\n def testModel(self):\n WakePumpTimes = [3.162,1.780,1.000,0.562,0.316,0.188,0.100,0.056,0.031,0.018,0.010,0.0056]\n \"\"\"\n This compares the same dose over 3 different time periods 5,25 and 50 sec\n It does this by changing the concentration, but perhpas it would be\n better to change the pump speed.\n\n eg. 5000 mSec * 4 mg/ml *0.000025 mls/mSec / 0.330 kg = 1.5 mg/kg\n # model.calculateCocConc defaults to bodyweight 0.330 \n\n \"\"\" \n # testRecord1 5 sec infusion\n testRecord1 = dm.DataRecord([],\"5 sec\") \n testRecord1.datalist = [[10000, 'P'],[15000, 'p']]\n testRecord1.pumpSpeed = 0.025 # Wake default 0.1 mls/4 sec = 0.025 / sec\n testRecord1.cocConc = 4.0\n testRecord1.TH_PumpTimes = WakePumpTimes\n testRecord1.extractStatsFromList()\n duration = testRecord1.totalPumpDuration\n dose = (testRecord1.totalPumpDuration * testRecord1.cocConc * (testRecord1.pumpSpeed/1000)/0.330)\n print(\"testRecord1 Duration = {0}; Total Dose = {1:2.1f}\".format(duration,dose))\n # testRecord2 50 sec infusion\n duration = 50\n testRecord2 = dm.DataRecord([],\"50 sec\") \n testRecord2.datalist = [[10000, 'P'],[15000, 'p'], [15000, 'P'],[20000, 'p'], \\\n [20000, 'P'],[25000, 'p'], [25000, 'P'],[30000, 'p'], \\\n [30000, 'P'],[35000, 'p'], [35000, 'P'],[40000, 'p'], \\\n [40000, 'P'],[45000, 'p'], [45000, 'P'],[50000, 'p'], \\\n [50000, 'P'],[55000, 'p'], [55000, 'P'],[60000, 'p']] \n testRecord2.pumpSpeed = 0.025 # Wake default 0.1 mls/4 sec = 0.025 / sec\n testRecord2.cocConc = 4.0 / (duration / 5.0)\n testRecord2.TH_PumpTimes = WakePumpTimes\n testRecord2.extractStatsFromList()\n duration = testRecord2.totalPumpDuration\n dose = (testRecord2.totalPumpDuration * testRecord2.cocConc * (testRecord2.pumpSpeed/1000)/0.330)\n print(\"testRecord2 Duration = {0}; Total Dose = {1:2.1f}\".format(duration,dose))\n # testRecord3 90 sec infusion\n duration = 90\n testRecord3 = dm.DataRecord([],\"90 sec\") \n testRecord3.datalist = [[10000, 'P'],[15000, 'p'], [15000, 'P'],[20000, 'p'], \\\n [20000, 'P'],[25000, 'p'], [25000, 'P'],[30000, 'p'], \\\n [30000, 'P'],[35000, 'p'], [35000, 'P'],[40000, 'p'], \\\n [40000, 'P'],[45000, 'p'], [45000, 'P'],[50000, 'p'], \\\n [50000, 'P'],[55000, 'p'], [55000, 'P'],[60000, 'p'], \\\n [60000, 'P'],[65000, 'p'], [65000, 'P'],[70000, 'p'], \\\n [70000, 'P'],[75000, 'p'], [75000, 'P'],[80000, 'p'], \\\n [80000, 'P'],[85000, 'p'], [85000, 'P'],[90000, 'p'], \\\n [90000, 'P'],[95000, 'p'], [95000, 'P'],[100000, 'p']]\n testRecord3.pumpSpeed = 0.025 # Wake default 0.1 mls/4 sec = 0.025 / sec\n testRecord3.cocConc = 4.0 / (duration / 5.0)\n testRecord3.TH_PumpTimes = WakePumpTimes\n testRecord3.extractStatsFromList()\n duration = testRecord3.totalPumpDuration\n dose = (testRecord3.totalPumpDuration * testRecord3.cocConc * (testRecord3.pumpSpeed/1000)/0.330)\n print(\"testRecord3 Duration = {0}; Total Dose = {1:2.1f}\".format(duration,dose))\n\n self.showModel(testRecord1, resolution = 5, aColor = \"black\", max_y_scale = 10)\n self.showModel(testRecord2, resolution = 5, aColor = \"red\", clear = False, max_y_scale = 10)\n self.showModel(testRecord3, resolution = 5, aColor = \"blue\", clear = False, max_y_scale = 10)\n\n def test(self):\n self.clearGraphTabCanvas()\n x_zero = 50\n y_zero = 550\n x_pixel_width = 700 \n y_pixel_height = 500\n max_x_scale = self.max_x_scale.get()\n x_divisions = 12\n if (max_x_scale == 10) or (max_x_scale == 30): x_divisions = 10\n max_y_scale = self.max_y_scale.get() \n y_divisions = 10\n GraphLib.drawXaxis(self.graphCanvas, x_zero, y_zero, x_pixel_width, max_x_scale, x_divisions, color = \"red\")\n offset = 0 \n GraphLib.drawYaxis(self.graphCanvas, x_zero, y_zero, y_pixel_height, max_y_scale, y_divisions, True, color = \"blue\")\n GraphLib.drawYaxis(self.graphCanvas, x_zero+x_pixel_width +10, y_zero, y_pixel_height, max_y_scale, y_divisions, False)\n\n # ******************** End Draw Threshold ***********\n\n def clearText(self):\n self.textBox.delete(\"1.0\",END)\n\n\n def periodic_check(self):\n thisTime = datetime.now()\n self.clockTimeStringVar.set(thisTime.strftime(\"%H:%M:%S\")) \n self.root.after(100, self.periodic_check) \n\n def go(self):\n self.root.after(100, self.periodic_check)\n self.root.mainloop()\n \n\nif __name__ == \"__main__\":\n sys.exit(main())\n\n","repo_name":"DCSR/Analysis","sub_path":"Analysis.py","file_name":"Analysis.py","file_ext":"py","file_size_in_byte":96877,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"1414116401","text":"class Node:\n\n def __init__(temp, data):\n temp.left = None\n temp.right = None\n temp.data = data\n\n\nclass Tree:\n\n def levelordertraverse(self,root):\n queue=[]\n queue.append(root)\n\n while(len(queue)>0):\n front=queue[0].data\n print(front)\n current=queue.pop(0)\n if(current.left!=None):\n queue.append(current.left)\n if(current.right!=None):\n queue.append(current.right)\n\n\n\nroot = Node(1)\nroot.left = Node(2)\nroot.right = Node(3)\nroot.left.left = Node(4)\nroot.left.right = Node(5)\nroot.right.left = Node(6)\nroot.right.right = Node(7)\nroot.right.left.right = Node(8)\nroot.right.right.right = Node(9)\n\nif __name__ == \"__main__\":\n fp = Tree()\n\n # for i in a:\n # data = fp.insert(i)\n # print(data)\n\n data=fp.levelordertraverse(root)\n print(data)\n","repo_name":"tanu312000/pyChapter","sub_path":"org/netsetos/python/Tree/BinaryTree/levelordertraversal.py","file_name":"levelordertraversal.py","file_ext":"py","file_size_in_byte":886,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"4166602247","text":"#!/usr/bin/python3\n\n#\n# dividy.py\n#\n# Description: given two numbers (dividend and a divisor) divide the number without the use of any of the following operators: \"+ / * %\"\n#\n# Created: 6/03/2020\n#\n# Developer: codex040\n#\n# Version: 1.0\n#\n\n######################### Function definition ######################### \ndef divide(dividend, divisor):\n\n\t#local variables\n\tneg = False if ((dividend < 0) and (divisor < 0) or (dividend > 0 and divisor > 0)) else True\n\tquotient = 0\n\tTOP_MIN = -2147483648\n\tTOP_MAX = 2147483647\n\n\t#normalize the dividend and the divisor\n\tdividend, divisor = abs(dividend), abs(divisor)\n\n\t#determine the quotient if the divisor isn't \"0\"\n\tif divisor != 1:\n\t\twhile dividend >= divisor: \n\t\t\tquotient += 1\n\t\t\tdividend -= divisor\n\telse:\n\t\tquotient = dividend\n\n\t#determine what sign needs to be assigned as well if over/underflow occured\n\tif neg: \n\t\treturn max(-quotient, TOP_MIN)\n\telse: \n\t\treturn min(quotient, TOP_MAX)\n\n\n######################### Driver definition ######################### \n\n#local variables \ndividend = 313\ndivisor = 3\nresult = divide(dividend, divisor)\n\n#display info to the user \nprint(\"Dividend is : \" + str(dividend))\nprint(\"Divisor is : \" + str(divisor))\nprint(\"Result is : \" + str(result))\n","repo_name":"maihan040/Python_Random_Scripts","sub_path":"divide.py","file_name":"divide.py","file_ext":"py","file_size_in_byte":1233,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22467694614","text":"class Solution:\n def areAlmostEqual(self, s1: str, s2: str) -> bool:\n num_diff = 0\n diff_idx1 = -1\n diff_idx2 = -1\n n = len(s1)\n for i in range(n):\n if s1[i] != s2[i]:\n num_diff += 1\n if diff_idx1 < 0:\n diff_idx1 = i\n else:\n diff_idx2 = i\n if num_diff > 2:\n return False\n if (s1[diff_idx1], s1[diff_idx2]) == (s2[diff_idx2], s2[diff_idx1]):\n return True\n return False","repo_name":"Aden-Q/LeetCode","sub_path":"code/1790.Check-if-One-String-Swap-Can-Make-Strings-Equal.py","file_name":"1790.Check-if-One-String-Swap-Can-Make-Strings-Equal.py","file_ext":"py","file_size_in_byte":542,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22499958400","text":"# 广度优先搜索\r\n\r\nfrom collections import deque\r\n\r\ngraph = {}\r\ngraph['you'] = ['alice','bob','claire']\r\ngraph['bob'] = ['anuj','peggy']\r\ngraph['alice'] = ['peggy']\r\ngraph['claire'] = ['thom','jonny']\r\ngraph['anuj'] = []\r\ngraph['peggy'] = []\r\ngraph['thom'] = []\r\ngraph['jonny'] = []\r\n\r\ndef person_is_seller(name):\r\n return name[-1] == 'm'\r\n\r\ndef search(name):\r\n search_queue = deque() #创建一个队列\r\n search_queue += graph[name] #将节点都加入到搜索队列中\r\n searched = [] #用于记录检查过的人\r\n while search_queue: #只要队列不为空\r\n person = search_queue.popleft() #就取出第一个节点\r\n if person not in searched: #仅当这个节点没检查过时才检查\r\n if person_is_seller(person):\r\n print (person + ' is a mango seller!')\r\n return True\r\n else:\r\n search_queue +=graph[person]\r\n searched.append(person) #将这个节点标记为检查过\r\n return False\r\n\r\nsearch('you')\r\n\r\n\r\n","repo_name":"Queenzhang/notes","sub_path":"arithmetic/BFS.py","file_name":"BFS.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31757316404","text":"from elliot.run import run_experiment\nimport argparse\n\nparser = argparse.ArgumentParser(description=\"Run sample main.\")\nparser.add_argument('--type_experiment', type=str, default='baselines')\nparser.add_argument('--dataset', type=str, default='amazon_baby_alexnet')\nargs = parser.parse_args()\n\nrun_experiment(\"config_files/{0}_{1}.yml\".format(args.type_experiment, args.dataset))\n\n","repo_name":"peternara/Elliot-Reproducible-VRSs-Recommender","sub_path":"sample_main.py","file_name":"sample_main.py","file_ext":"py","file_size_in_byte":381,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"74798139776","text":"#!/usr/bin/env python\n\n\"\"\" A ham radio keyer for raspberry pi\"\"\"\n\nimport RPi.GPIO as GPIO\nimport time\nfrom cmd import Cmd\n\n# Set defaults here\nCALL = \"w1aw\"\nRST = \"r 5nn tu\"\ndefault_WPM = 25\nBCM_pin = 24\n\n\ndef dit():\n tx(DitLength)\n time.sleep(DitLength)\n\n\ndef dah():\n tx(DahLength)\n time.sleep(DitLength)\n\n\ndef space():\n time.sleep(CharSpaceExtra)\n\n\ndef word():\n time.sleep(WordSpaceExtra)\n\n\ndef tx(keylength):\n \"\"\" Keys the TX \"\"\"\n # set the output to Key Down...\n GPIO.output(BCM_pin, True)\n time.sleep(keylength)\n # clear the output ...\n GPIO.output(BCM_pin, False)\n return\n\n\ndef send(code):\n for element in morse[code]:\n if element == \".\":\n dit()\n elif element == \"-\":\n dah()\n\n\ndef lookup(message):\n sendspace = True\n for char in message:\n if char == \"<\":\n sendspace = False\n continue\n if char == \">\":\n sendspace = True\n continue\n if char == \" \":\n sendspace = True\n if char == \" \":\n word()\n elif char not in morse.keys():\n print(\"\")\n print(\"unknown char: %s\" % char)\n pass\n else:\n send(char)\n if sendspace:\n space()\n print(\"sent message: '%s'\" % message)\n\n\nclass MyPrompt(Cmd):\n def emptyline(self):\n pass\n\n def do_PROSIGNS(self, args):\n \"\"\"Anything enclosed in angle brackets will have no separating space: \"\"\"\n pass\n\n def do_CALL(self, args):\n \"\"\"Passing arguments sets the CALL message.\nWith no arguments, the CALL message is sent.\n'CALL ?' displays the current CALL message.\"\"\"\n global CALL\n if not args:\n lookup(CALL.lower())\n elif args == \"?\":\n print(\"current CALL message: '%s'\" % CALL)\n else:\n print(\"Setting CALL to '%s'\" % args.lower())\n CALL = args.lower()\n\n def do_RST(self, args):\n \"\"\"Passing arguments sets the RST message.\nWith no arguments, the RST message is sent.\n'RST ?' displays the current RST message.\"\"\"\n global RST\n if not args:\n lookup(RST.lower())\n elif args == \"?\":\n print(\"current RST message: '%s'\" % RST)\n else:\n print(\"Setting RST to '%s'\" % args.lower())\n RST = args.lower()\n\n def do_EOF(self, args):\n \"\"\"Quits the program. Type 'EOF' or press 'Cntrl-D'\"\"\"\n print(\"Quitting.\")\n raise SystemExit\n\n def default(self, line):\n lookup(line.lower())\n\n def do_WPM(self, args):\n \"\"\"Sets WPM to argument passed. No argument will set it to the default.\"\"\"\n global WPM\n global DitLength\n global DahLength\n if not args:\n WPM = default_WPM\n else:\n if args.isdigit():\n WPM = int(args)\n else:\n print(\"Invalid value for WPM: %s. WPM must be an integer.\")\n prompt.prompt = str(WPM) + ' WPM> '\n DitLength = float(60 / float(WPM * 50))\n DahLength = 3 * DitLength\n\n# Initialize GPIO\nGPIO.setmode(GPIO.BCM)\nGPIO.setwarnings(False)\nGPIO.setup(BCM_pin, GPIO.OUT)\nGPIO.output(BCM_pin, GPIO.LOW)\n\nWPM = default_WPM\nDitLength = float(60 / float(WPM * 50))\nDahLength = 3 * DitLength\n# Dits and Dahs will have one ditlength space after each already\n# CharSpace and WordSpace should be 3 and 7, so here is the extra\nCharSpaceExtra = 2 * DitLength\nWordSpaceExtra = 6 * DitLength\n\nmorse = {\n \"a\": \".-\",\n \"b\": \"-...\",\n \"c\": \"-.-.\",\n \"d\": \"-..\",\n \"e\": \".\",\n \"f\": \"..-.\",\n \"g\": \"--.\",\n \"h\": \"....\",\n \"i\": \"..\",\n \"j\": \".---\",\n \"k\": \"-.-\",\n \"l\": \".-..\",\n \"m\": \"--\",\n \"n\": \"-.\",\n \"o\": \"---\",\n \"p\": \".--.\",\n \"q\": \"--.-\",\n \"r\": \".-.\",\n \"s\": \"...\",\n \"t\": \"-\",\n \"u\": \"..-\",\n \"v\": \"...-\",\n \"w\": \".--\",\n \"x\": \"-..-\",\n \"y\": \"-.--\",\n \"z\": \"--..\",\n \"0\": \"-----\",\n \"1\": \".----\",\n \"2\": \"..---\",\n \"3\": \"...--\",\n \"4\": \"....-\",\n \"5\": \".....\",\n \"6\": \"-....\",\n \"7\": \"--...\",\n \"8\": \"---..\",\n \"9\": \"----.\",\n \"?\": \"..--..\",\n \".\": \".-.-.-\",\n \",\": \"--..--\",\n \"/\": \"-..-.\"\n}\n\nif __name__ == '__main__':\n prompt = MyPrompt()\n prompt.prompt = str(WPM) + ' WPM> '\n prompt.cmdloop('Starting keyer. Type \"help\" for help...')\n","repo_name":"adodson/pi_keyer","sub_path":"pi_keyer.py","file_name":"pi_keyer.py","file_ext":"py","file_size_in_byte":4350,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"19793834613","text":"# -*- coding: utf-8 -*-\nimport scrapy\nfrom scrapy.http import HtmlResponse\nfrom jobparser.items import JobparserItem\n\n\nclass HhruSpider(scrapy.Spider):\n name = 'hhru'\n allowed_domains = ['hh.ru']\n start_urls = ['https://saratov.hh.ru/search/vacancy?L_is_autosearch=false&clusters=/'\n 'true&enable_snippets=true&items_on_page=20&text=python']\n\n def parse(self, response: HtmlResponse):\n next_page = response.css('a.HH-Pager-Controls-Next::attr(href)').extract_first()\n if next_page:\n yield response.follow(next_page, callback=self.parse)\n\n vacansy = response.xpath(\n \"//div[@class='vacancy-serp-item__info']//a[@data-qa='vacancy-serp__vacancy-title']/@href\").extract()\n\n if vacansy:\n for link in vacansy:\n yield response.follow(link, callback=self.vacansy_parse)\n\n def vacansy_parse(self, response: HtmlResponse):\n url = response.url\n name = response.xpath(\n \"//div[@class=\\'bloko-columns-row\\']//h1[@data-qa=\\'vacancy-title\\']//text()\").extract_first()\n salary = response.xpath(\"//div[@class=\\'bloko-columns-row\\']//p[@class=\\'vacancy-salary\\']//text()\").extract()\n yield JobparserItem(link=url, name=name, salary=salary)\n","repo_name":"Brutalfacepalm/data_mining","sub_path":"lesson_5/jobparser/spiders/hhru.py","file_name":"hhru.py","file_ext":"py","file_size_in_byte":1269,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"16496346461","text":"from tkinter import *\r\nfrom tkinter import ttk\r\nfrom Adafruit_IO import Client,Data\r\nfrom PIL import Image,ImageTk\r\naio = Client('ak47av', 'bda36d622b874920902764079deb0631')\r\n\r\nclass product:\r\n\r\n def __init__(self, name, price, quantity):\r\n self.price = price\r\n self.name = name\r\n self.quantity = quantity\r\n\r\n def adagetdata(self):\r\n Quantity = aio.receive(self.name)\r\n Price = aio.receive((self.name+'price'))\r\n self.quantity = IntVar()\r\n self.quantity.set(Quantity.value)\r\n self.price = IntVar()\r\n self.price.set(Price.value)\r\n\r\nproduct1 = product('lays',10,10)\r\nproduct2 = product('coke',25,25)\r\nproduct3 = product('candy',1,50)\r\n\r\nwin = Tk()\r\nwin.geometry('1386x250')\r\nwin.title('Update Vending Details')\r\n\r\ndef main():\r\n\r\n product1.adagetdata()\r\n product2.adagetdata()\r\n product3.adagetdata()\r\n\r\n fr1 = Frame(win).grid(row=0,column=0)\r\n #Display existing attributes\r\n ttk.Separator(fr1, orient=VERTICAL).grid(row=0,column=3,rowspan=5,sticky='ns')\r\n ttk.Separator(fr1, orient=VERTICAL).grid(row=0,column=8,rowspan=5,sticky='ns')\r\n load = Image.open('/home/akav/Downloads/Spark-Logo_Navin.png')\r\n load = load.resize((150, 150), Image.ANTIALIAS)\r\n sparkimg = ImageTk.PhotoImage(load)\r\n spark = Label(fr1,image = sparkimg)\r\n spark.image = sparkimg\r\n spark.grid(row=0,rowspan=5,column=9,padx=10)\r\n\r\n #display update options\r\n\r\n prodlabel = Label(fr1,text=' Products ').grid(row=0,column=0,padx=10,pady=10)\r\n pricelabel = Label(fr1,text=' Prices ').grid(row=0,column=1)\r\n quantitylabel = Label(fr1,text=' Quantity Available ').grid(row=0,column=2)\r\n\r\n p1 = Label(fr1,text=(product1.name)).grid(row=1,column=0)\r\n price1 = Label(fr1,textvariable = str(product1.price)).grid(row=1,column=1)\r\n q1 = Label(fr1,textvariable=str(product1.quantity)).grid(row=1,column=2)\r\n\r\n p2 = Label(fr1, text=(product2.name)).grid(row=2, column=0)\r\n price2 = Label(fr1, textvariable=str(product2.price)).grid(row=2, column=1)\r\n q2 = Label(fr1, textvariable=str(product2.quantity)).grid(row=2, column=2)\r\n\r\n p3 = Label(fr1, text=(product3.name)).grid(row=3, column=0)\r\n price3 = Label(fr1, textvariable=str(product3.price)).grid(row=3, column=1)\r\n q3 = Label(fr1, textvariable=str(product3.quantity)).grid(row=3, column=2)\r\n\r\n\r\n def updateprice1():\r\n product1.price.set(pr1.get())\r\n data = Data(value=product1.price.get())\r\n aio.create_data((product1.name+'price'),data)\r\n print('Updated !')\r\n pr1.set(0)\r\n\r\n def updatequan1():\r\n product1.quantity.set(product1.quantity.get()+quantity1.get())\r\n data1 = Data(value=product1.quantity.get())\r\n aio.create_data(product1.name,data1)\r\n print('Updated !')\r\n quantity1.set(0)\r\n\r\n def updateprice2():\r\n product2.price.set(pr2.get())\r\n data = Data(value=product2.price.get())\r\n aio.create_data((product2.name+'price'), data)\r\n print('Updated !')\r\n pr2.set(0)\r\n\r\n def updatequan2():\r\n product2.quantity.set(product2.quantity.get()+quantity2.get())\r\n data1 = Data(value=product2.quantity.get())\r\n aio.create_data(product2.name, data1)\r\n print('Updated !')\r\n quantity2.set(0)\r\n\r\n def updateprice3():\r\n product3.price.set(pr3.get())\r\n data = Data(value=product3.price.get())\r\n aio.create_data((product3.name+'price'), data)\r\n print('Updated !')\r\n pr3.set(0)\r\n\r\n def updatequan3():\r\n product3.quantity.set(product3.quantity.get()+quantity3.get())\r\n data1 = Data(value=product3.quantity.get())\r\n aio.create_data(product3.name, data1)\r\n print('Updated !')\r\n quantity3.set(0)\r\n\r\n editPrice = Label(fr1,text=' Edit Prices ').grid(row=0,column=4,padx=10)\r\n editQuantity = Label(fr1,text=' Add Quantity ').grid(row=0,column=5)\r\n\r\n pr1 = IntVar()\r\n pre1 = Entry(fr1,textvariable=str(pr1)).grid(row=1,column=4,padx=20)\r\n quantity1 = IntVar()\r\n quan1 = Entry(fr1,textvariable=str(quantity1)).grid(row=1,column=5,padx=10)\r\n okprice1 = Button(fr1,text='Update Price',command=updateprice1).grid(row=1,column=6,padx=15)\r\n okquan1 = Button(fr1,text='Update Quantity',command=updatequan1).grid(row=1,column=7,padx=15)\r\n\r\n pr2 = IntVar()\r\n pre2 = Entry(fr1,textvariable=str(pr2)).grid(row=2, column=4)\r\n quantity2 = IntVar()\r\n quan2 = Entry(fr1,textvariable=str(quantity2)).grid(row=2, column=5)\r\n okprice2 = Button(fr1, text='Update Price',command=updateprice2).grid(row=2, column=6, padx=15)\r\n okquan2 = Button(fr1, text='Update Quantity',command=updatequan2).grid(row=2, column=7, padx=15)\r\n\r\n pr3 = IntVar()\r\n pre3 = Entry(fr1,textvariable=str(pr3)).grid(row=3, column=4)\r\n quantity3 = IntVar()\r\n quan3 = Entry(fr1,textvariable=str(quantity3)).grid(row=3, column=5)\r\n okprice3 = Button(fr1, text='Update Price',command=updateprice3).grid(row=3, column=6, padx=15)\r\n okquan3 = Button(fr1, text='Update Quantity',command=updatequan3).grid(row=3, column=7, padx=15)\r\n\r\n refresh = Button(fr1,text='Refresh Contents',command=main).grid(row=4,column=3)\r\n\r\nmain()\r\n\r\nwin.mainloop()","repo_name":"ak47av/Spark_internship","sub_path":"editadafruit.py","file_name":"editadafruit.py","file_ext":"py","file_size_in_byte":5225,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"26118847752","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep 28 16:55:35 2017\n@author: Daryl Larsen\n\"\"\"\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndef poly1(Xin, XYparams):\n '''\n Includes polynomial terms up to order 'pord' for each element and quadratic \n cross terms One observation (row) at a time\n '''\n (pord, nx, ny, nz) = XYparams\n nX = nx + nz\n Xbasis = np.ones((1, 1))\n # generate polynomial terms for each element\n for i in range(1, pord+1):\n Xbasis = np.append(Xbasis, Xin**i)\n # generate cross terms\n for i in range (0, nX):\n for j in range(i+1, nX):\n temp = Xin[i]*Xin[j]\n Xbasis = np.append(Xbasis, temp)\n return Xbasis\n\ndef XYfunc(Xm, Zn, XYparams, coeffs):\n (pord, nx, ny, nz) = XYparams\n An = np.exp(Zn)\n XZin = np.append(Xm, An)\n XYbasis = np.append(1., XZin)\n for i in range(1, pord+1):\n XYbasis = poly1(XZin, XYparams)\n Xn = np.dot(XYbasis, coeffs)\n return Xn\n \ndef MVOLS(Y, X):\n '''\n OLS regression with observations in rows\n '''\n XX = np.dot(np.transpose(X), X)\n XY = np.dot(np.transpose(X), Y)\n coeffs = np.dot(np.linalg.inv(XX), XY)\n return coeffs\n\ndef GSSA(params, kbar, GSSAparams, old_coeffs): \n regtype = 'poly1' # functional form for X & Y functions \n fittype = 'MVOLS' # regression fitting method\n ccrit = 1.0E-8 # convergence criteria for XY change\n damp = 0.01 # damping paramqter for fixed point algorithm\n \n [alpha, beta, tau, rho, sigma] = params\n (T, nx, ny, nz, pord, old) = GSSAparams\n cnumb = int((pord+1)*(nx+nz) + .5*(nx+nz-1)*(nx+nz-2))\n cnumb2 = int(3*(nx+nz) + .5*(nx+nz-1)*(nx+nz-2))\n Xstart = kbar\n \n #create history of Z's\n Z = np.zeros([T,nz])\n for t in range(1,T):\n Z[t,:] = rho*Z[t-1] + np.random.randn(1)*sigma\n \n if regtype == 'poly1' and old == False:\n coeffs = np.array([[ -2.04961035e-02], \\\n [ 2.26920891e-01], \\\n [ 1.17409797e-01], \\\n [ -6.27573544e-01], \\\n [ -4.88424960e-05], \\\n [ 3.49581228e-01]])\n elif old == True:\n coeffs = old_coeffs\n \n if old == False and pord > 2:\n A = np.zeros((cnumb - cnumb2, nx+ny))\n coeffs = np.insert(coeffs, cnumb2 - 1, A)\n \n dist = 1.\n distold = 2.\n count = 0\n Xold = np.ones((T-1, nx+ny))\n XYparams = (pord, nx, ny, nz)\n\n \n while dist > ccrit:\n count = count + 1\n X = np.zeros((T+1, nx))\n Xin = np.zeros((T, nx+nz))\n A = np.exp(Z)\n x = np.zeros((T,(pord*2+2)))\n X[0] = XYfunc(Xstart, Z[0], XYparams, coeffs)\n for t in range(1,T+1):\n X[t] = XYfunc(X[t-1], Z[t-1], XYparams, coeffs)\n Xin[t-1,:] = np.concatenate((X[t-1], A[t-1]))\n x[t-1,:] = poly1(Xin[t-1,:], XYparams)\n # plot time series\n if count % 100 == 0:\n X1 = X[0:T]\n timeperiods = np.asarray(range(0,T))\n plt.plot(timeperiods, X1, label='X')\n plt.axhline(y=kbar, color='r')\n plt.title('time series')\n plt.xlabel('time')\n plt.legend(loc=9, ncol=(nx+ny))\n plt.show() \n \n # Generate consumption and gamma series \n c = (1-tau)*X[0:T]**alpha*A[0:T] - X[1:T+1]\n Gam = (beta*c[1:T]**(-1)*(alpha*X[1:T]**(alpha-1)*A[1:T]*(1-tau))) / (c[0:T-1]**(-1))\n \n # update values for X and Y\n Xnew = Gam*X[1:T]\n x = x[0:T-1,:]\n \n if fittype == 'MVOLS':\n coeffsnew = MVOLS(Xnew, x)\n \n if dist < distold:\n damp = damp*1.05\n if damp > 1.:\n damp = 1.\n else:\n damp = damp*.8\n if damp < 0.001:\n damp = 0.001\n \n distold = 1.*dist\n\n # calculate distance between X and Xold\n dist = np.mean(np.abs(1-(Xnew/Xold)))\n print('count ', count, 'distance', dist, 'damp', damp)\n \n # update coeffs\n Xold = 1*Xnew\n coeffs = (1-damp)*coeffs + damp*coeffsnew\n if count % 100 == 0:\n print('coeffs', coeffs)\n coeffs = (1-damp)*coeffs + damp*coeffsnew\n \n return coeffs","repo_name":"econdaryl/PolicyUncertainty","sub_path":"BrockMirman/gssa.py","file_name":"gssa.py","file_ext":"py","file_size_in_byte":4301,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"14451123392","text":"import os \nimport json\ndir_path = os.path.dirname(os.path.realpath(__file__))\nprint(dir_path)\n# call listdir() method\n# path is a directory of which you want to list\ndirectories = os.listdir( dir_path )\n \ndef main():\n # This would print all the files and directories\n for file in directories:\n if file == \"eg.tf.json\":\n print(file)\n file_path = dir_path+'/eg.tf.json'\n new_file_path = dir_path+'/update.json'\n print(file_path)\n dict_object = python_json_file_to_dict(file_path)\n python_dict_to_json_file(new_file_path, dict_object)\n\ndef python_json_file_to_dict(file_path):\n try:\n # Get a file object with write permission.\n file_object = open(file_path, 'r')\n # Load JSON file data to a python dict object.\n dict_object = json.load(file_object)\n update_engress(dict_object)\n print(dict_object)\n # all_values = list(nested_dict_values_iterator(dict_object))\n # print(all_values)\n # print(dict_object)\n return dict_object\n except FileNotFoundError:\n print(file_path + \" not found. \")\n\ndef update_engress(dict_object):\n engressValues = dict_object['resource']['aws_security_group_rule'].values()\n count = 0\n for v in engressValues:\n count += 1\n print(count)\n if count>0:\n new_engress_str = f'engress{count+1}'\n old_engress_str = f'engress{count}'\n dict_object['resource']['aws_security_group_rule'][new_engress_str]=dict_object['resource']['aws_security_group_rule'][old_engress_str]\n dict_object['resource']['aws_security_group_rule'][new_engress_str]['vpc-id']='xyz' \n\ndef nested_dict_values_iterator(dict_obj):\n ''' This function accepts a nested dictionary as argument\n and iterate over all values of nested dictionaries\n '''\n # Iterate over all values of given dictionary\n for value in dict_obj.values():\n # Check if value is of dict type\n if isinstance(value, dict):\n # If value is dict then iterate over all its values\n for v in nested_dict_values_iterator(value):\n yield v\n else:\n # If value is not dict type then yield the value\n yield value\n\n\ndef python_dict_to_json_file(file_path, dict_object):\n try:\n # Get a file object with write permission.\n file_object = open(file_path, 'w')\n # Save dict data into the JSON file.\n json.dump(dict_object, file_object)\n print(file_path + \" created. \") \n except FileNotFoundError:\n print(file_path + \" not found. \")\n\nif __name__ == '__main__':\n main()","repo_name":"prem-v/poc-tf","sub_path":"loadFile.py","file_name":"loadFile.py","file_ext":"py","file_size_in_byte":2602,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"885800333","text":"import maya.cmds as cmds\nimport maya.OpenMaya as OpenMaya\nimport sg.convert\nimport math\n\n\ndef makeSymmetry( side ):\n \n def setRotationMirror( stdOther, std ):\n trMtx = OpenMaya.MTransformationMatrix()\n trMtx.rotateTo( OpenMaya.MEulerRotation( math.radians(180), 0, 0 ) )\n mirrorRotMtx = trMtx.asMatrix()\n \n mtx = sg.convert.listToMatrix( cmds.getAttr( otherStd + '.wm' ) )\n mtx *= mirrorRotMtx\n mtxList = sg.convert.matrixToList( mtx )\n pos = cmds.xform( otherStd, q=1, t=1, ws=1 )\n mtxList[12] = -pos[0]\n mtxList[13] = pos[1]\n mtxList[14] = pos[2]\n cmds.xform( std, ws=1, matrix=mtxList )\n \n def setPositionMirror( stdOther, std ):\n pos = cmds.xform( stdOther, q=1, t=1, ws=1 )\n cmds.xform( std, ws=1, t=[-pos[0], pos[1], pos[2]])\n \n \n\n otherSide = side.replace( '_R_', '_L_' )\n if otherSide == side:\n otherSide = side.replace( '_L_', '_R_' )\n \n sels = cmds.ls( sl=1 )\n target = sels[-1] \n ns = target.split( 'Std_' )[0]\n stdList = cmds.ls( ns + 'Std_*', type='transform' )\n \n for std in stdList:\n if std.find( side ) != -1: continue\n if std.find( otherSide ) != -1: continue\n cmds.setAttr( std + '.tx', 0 )\n \n \n for std in stdList:\n if std.find( side ) == -1: continue\n otherStd = std.replace( side, otherSide )\n \n if not cmds.objExists( std ) or not cmds.objExists( otherStd ): continue\n \n if std.find( '_Wrist_' ) != -1:\n setRotationMirror( otherStd, std )\n\n\n for std in stdList:\n if std.find( side ) == -1: continue\n otherStd = std.replace( side, otherSide )\n \n if not cmds.objExists( std ) or not cmds.objExists( otherStd ): continue\n \n if std.find( '_Thumb_' ) != -1 or std.find( '_Index_' ) != -1 or std.find( '_Middle_' ) != -1 or std.find( '_Ring_' ) != -1 or std.find( '_Pinky_' ) != -1:\n setRotationMirror( otherStd, std )\n \n for std in stdList:\n if std.find( side ) == -1: continue\n otherStd = std.replace( side, otherSide )\n \n if not cmds.objExists( std ) or not cmds.objExists( otherStd ): continue\n \n setPositionMirror( otherStd, std )\n \n \n\ndef makeSymmetryToR( evt=0 ):\n makeSymmetry( '_R_' )\n\n\ndef makeSymmetryToL( evt=0 ):\n makeSymmetry( '_L_' )","repo_name":"kimsung9k/public","sub_path":"maya_tools/sg/python/sg/Function_autoRig/popupCommand.py","file_name":"popupCommand.py","file_ext":"py","file_size_in_byte":2416,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"34947575633","text":"n=int(input(\"enter n value:\"))\r\nf1=0\r\nf2=1\r\nprint(f1,end=\" \")\r\nprint(f2,end=\" \")\r\ni=1\r\nwhile(i<=n): #while((f1+f2)<=n):\r\n f3=f2+f1\r\n print(f3,end=\" \")\r\n f1=f2\r\n f2=f3\r\n i=i+1","repo_name":"Nenichan/python-programs","sub_path":"fibonacciseries.py","file_name":"fibonacciseries.py","file_ext":"py","file_size_in_byte":179,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"5830235602","text":"import requests\nfrom bs4 import BeautifulSoup\nfrom jsonpath import jsonpath\nfrom time import sleep\nfrom prettytable import PrettyTable\nimport argparse\nimport urllib3\n\nurllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)\n\n\ndef ip_shudi(ip):\n url = \"https://www.ipshudi.com/\" + ip + \".htm\"\n web = requests.get(url)\n bs = BeautifulSoup(web.content, 'lxml')\n item = bs.find('tbody').find_all('span')\n info = []\n for i in item:\n info.append(i.text)\n return info\n\n\ndef ip_chinaz(ip):\n url = \"https://ip.tool.chinaz.com/\" + ip\n web = requests.get(url)\n bs = BeautifulSoup(web.content, 'lxml')\n item = bs.find('span', class_='Whwtdhalf w45-0 lh45').find('em').text\n item = item.split()\n info = []\n for i in item:\n info.append(i)\n return info\n\n\ndef ip_threatbook(ip):\n url = \"https://api.threatbook.cn/v3/scene/ip_reputation\"\n query = {\n \"apikey\": \"\",\n \"resource\": ip,\n \"lang\": \"zh\"\n }\n item = query.items()\n key, value = list(item)[0]\n if key == \"apikey\" and value != \"\":\n response = requests.request(\"GET\", url, params=query)\n result = response.json()\n a2 = jsonpath(result, \"$..severity\") # 严重级别\n a3 = jsonpath(result, \"$..is_malicious\") # 是否恶意IP\n a4 = jsonpath(result, \"$..confidence_level\") # 可信度\n a5 = \",\".join(jsonpath(result, \"$..judgments\")[0]) # 威胁类型\n a6 = \"-\".join(jsonpath(result, \"$..location.*\")[0:3]) # IP归属地\n a7 = jsonpath(result, \"$..update_time\") # 最近更新时间\n a8 = jsonpath(result, \"$..scene\") # 应用场景\n info = [ip, *a2, *a3, *a4, a5, a6, *a7, *a8]\n return info\n else:\n print(\"未设置微步API,前往第38行设置API\")\n\n\ndef domain_138(ip):\n try:\n headers = {\n 'User-Agent': \"Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1\"\n }\n url = \"https://ipchaxun.com/\" + ip + \"/\"\n web = requests.post(url, headers=headers)\n bs = BeautifulSoup(web.content, 'lxml')\n try:\n item = bs.find('span', class_='date').find_next_sibling().text\n except:\n item = \"暂无结果\"\n return item\n except:\n item = 0\n return item\n\n\ndef ip_whois(domain):\n url = \"https://api.devopsclub.cn/api/whoisquery\"\n query = {\n \"domain\": domain, # 你要查询域名\n \"type\": \"json\", # 数据类型\n \"standard\": \"true\"\n }\n response = requests.request(\"POST\", url, params=query, verify=False)\n result = response.json()\n a = jsonpath(result, \"$..status\") # 返回状态\n a1 = jsonpath(result, \"$...contactEmail\") # 联系邮箱\n a2 = jsonpath(result, \"$...contactPhone\") # 联系电话\n a3 = jsonpath(result, \"$...domainName\") # 域名\n a5 = jsonpath(result, \"$...expirationTime\") # 过期时间\n a6 = jsonpath(result, \"$...registrant\") # 联系人\n a7 = jsonpath(result, \"$...registrar\") # 注册商\n a10 = jsonpath(result, \"$...updatedDate\") # 更新时间\n info = [a3[0], a1[0], a2[0], a5[0], a6[0], a7[0], a10[0], a[0]]\n return info\n\n\ndef ip1_out_put(ip):\n x = PrettyTable()\n info = ip_shudi(ip)\n if len(info) == 1:\n x.title = 'IP属地 ipshudi.com'\n x.field_names = [\"IP\", \"归属地\"]\n x.add_row([ip, info[0]])\n print(x)\n elif len(info) == 2:\n x.title = 'IP属地 ipshudi.com'\n x.field_names = [\"IP\", \"归属地\", \"运营商\"]\n x.add_row([ip, info[0], info[1]])\n print(x)\n elif len(info) == 3:\n x.title = 'IP属地 ipshudi.com'\n x.field_names = [\"IP\", \"归属地\", \"运营商\", \"网络类型\"]\n x.add_row([ip, info[0], info[1], info[2]])\n print(x)\n sleep(1)\n\n\ndef ip2_out_put(ip):\n x = PrettyTable()\n info = ip_chinaz(ip)\n if len(info) == 1:\n x.title = '站长工具 chinaz.com'\n x.field_names = [\"IP\", \"归属地\"]\n x.add_row([ip, info[0]])\n print(x)\n elif len(info) == 2:\n x.title = '站长工具 chinaz.com'\n x.field_names = [\"IP\", \"归属地\", \"运营商\"]\n x.add_row([ip, info[0], info[1]])\n print(x)\n elif len(info) == 3:\n x.title = '站长工具 chinaz.com'\n x.field_names = [\"IP\", \"归属地\", \"运营商\", \"网络类型\"]\n x.add_row([ip, info[0], info[1], info[2]])\n print(x)\n sleep(1)\n\n\ndef domain_out_put(ip):\n if domain_138(ip) != 0:\n x = PrettyTable()\n info = domain_138(ip)\n x.title = '域名信息查询'\n x.field_names = [\"IP\", \"历史绑定域名\"]\n x.add_row([ip, info])\n print(x)\n sleep(1)\n else:\n print(\"域名查询失败\")\n\n\ndef whois_out_put(ip):\n if domain_138(ip) != \"暂无结果\":\n x = PrettyTable()\n info = ip_whois(domain_138(ip))\n x.title = '域名WHOIS信息'\n x.field_names = [\"域名\", \"联系邮箱\", \"联系电话\", \"过期时间\", \"联系人\", \"注册商\", \"更新时间\"]\n x.add_row([info[0], info[1], info[2], info[3], info[4], info[5], info[6]])\n print(x)\n\n\ndef threatbook(ip):\n x = PrettyTable()\n info = ip_threatbook(ip)\n x.title = '微步情报'\n x.field_names = [\"IP\", \"严重级别\", \"是否恶意IP\", \"可信度\", \"威胁类型\", \"ip归属地\", \"最近更新时间\", \"应用场景\"]\n x.add_row([info[0], info[1], info[2], info[3], info[4], info[5], info[6], info[7]])\n print(x)\n sleep(1)\n\n\ndef out_put_csv():\n a = 1\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('-t', metavar='target', help='查询单个IP', default='')\n parser.add_argument('-f', metavar='file', help='按文件查询多个IP', default='')\n parser.add_argument('-wb', help='使用微步查询,t使用 f不使用,默认f不使用', choices=['t', 'f'], default='f')\n parser.add_argument('-o', metavar='output', help='输出xls文件', default='')\n args = parser.parse_args()\n target = args.t\n file = args.f\n weibu = args.wb\n out_put = args.o\n if target != \"\" and file == \"\":\n if weibu == \"t\":\n ip1_out_put(target)\n ip2_out_put(target)\n threatbook(target)\n domain_out_put(target)\n whois_out_put(target)\n else:\n ip1_out_put(target)\n ip2_out_put(target)\n domain_out_put(target)\n whois_out_put(target)\n if file != \"\" and target == \"\":\n filename = file\n with open(filename) as f:\n for line in f:\n line = line.replace('\\n', '')\n if weibu == \"t\":\n ip1_out_put(line)\n ip2_out_put(line)\n threatbook(line)\n domain_out_put(line)\n whois_out_put(line)\n else:\n ip1_out_put(line)\n ip2_out_put(line)\n domain_out_put(line)\n whois_out_put(line)\n print('\\n')\n","repo_name":"IamShensy/ip_chck","sub_path":"IP_Check.py","file_name":"IP_Check.py","file_ext":"py","file_size_in_byte":7073,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"73157862655","text":"'''\n Need to deduce whether a ransom can be created from whole words available in a magazine\n'''\ndef main(magazine, ransom):\n '''\n @param magazine - an array of Strings (words available)\n @param ransom - an array of Strings (message wish to form)\n\n @output print \"Yes\" or \"No\" (depending on whether the ransom could be formed from the words in the magazine)\n '''\n # sorting optimises the searching function\n magazine.sort()\n ransom.sort()\n\n temp = []\n for word in ransom:\n if word in magazine:\n magazine.remove(word)\n temp.append(word)\n else:\n print(\"No\")\n break\n\n if len(ransom) == len(temp):\n print(\"Yes\")\n\n # NOTE: issubset does not cater for duplicates (e.g. ransom = [\"a\", \"a\", \"b\"] and magazine = [\"a\", \"b\"] will return True\n # e.g: if set(ransom).issubset(magazine):\n # print(\"Yes\")\n\n\nif __name__ == '__main__':\n magazine = [\"word2\", \"word1\"]\n ransom = [\"word1\", \"word2\", \"word1\"]\n main(magazine, ransom)\n","repo_name":"ginalamp/Coding_Challenges","sub_path":"hacker_rank/hash_dicts/ransom_note.py","file_name":"ransom_note.py","file_ext":"py","file_size_in_byte":1045,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39121976811","text":"'''\n输入一个链表,按链表从尾到头的顺序返回一个ArrayList。\n'''\n# class ListNode:\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution:\n # 返回从尾部到头部的列表值序列,例如[1,2,3]\n def printListFromTailToHead(self, head):\n res = []\n while head:\n res.append(head.val)\n head = head.next\n return res[::-1]\n","repo_name":"cuiboautotest/learnpython3","sub_path":"算法练习/剑指offer/6_从尾到头打印链表.py","file_name":"6_从尾到头打印链表.py","file_ext":"py","file_size_in_byte":432,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"71423509055","text":"import requests\nimport argparse\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-u\")\nparser.add_argument(\"-f\", type=argparse.FileType('r'))\nargs = parser.parse_args()\ntarget = args.u\nwordlist = args.f.readlines()\n\ndef loop():\n for word in wordlist:\n word = word.strip() \n print(word)\n res = requests.get(url=f\"{target}/{word}\")\n print(res)\n if res.status_code == 404:\n continue\n else:\n data = res.json()\n print(data)\n print(res.status_code)\n print(word)\n\nloop()","repo_name":"Berkan-dev/API-fuzzer","sub_path":"fuzzer.py","file_name":"fuzzer.py","file_ext":"py","file_size_in_byte":570,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"51473794973","text":"def Jump_Searching( list_array , x , n ):\r\n langkah = n**(1/2)\r\n prev = 0\r\n for y in range(len(list_array)):\r\n if type(list_array[y]) == list: \r\n output1 = Jump_Searching(list_array[int(y)],x,len(list_array[int(y)]))\r\n if output1 == -1:\r\n list_array[int(y)] = 'z'\r\n else:\r\n print(x,\"ada di indeks\",int(y),\"kolom \",output1)\r\n exit()\r\n while list_array[int(min(langkah, n)-1)] < x:\r\n prev = langkah\r\n langkah += n**(1/2)\r\n if prev >= n:\r\n return -1\r\n while list_array[int(prev)] < x: \r\n prev += 1\r\n if prev == min(langkah, n):\r\n return -1\r\n if list_array[int(prev)] == x:\r\n return int(prev)\r\n return -1\r\ndef Linear_Search(arr,x):\r\n for j in range(len(arr)):\r\n if type(arr[j]) == list:\r\n hasil_x = Linear_Search(arr[j],x)\r\n if hasil_x == -1:\r\n return -1\r\n else:\r\n print(f'{x} ditemukan pada indeks {j} kolom {hasil_x}')\r\n exit()\r\n elif arr[j] == x:\r\n return j\r\n return -1\r\n\r\nlist_NAMA = ['Arsel','Avivah','Daiva',['Wahyu','Wibi']]\r\npanjanglist = len(list_NAMA)\r\nwhile True:\r\n print(f'''\r\nDaftar ASLAB\r\n-------------\r\n|No | Nama |\r\n-------------\r\n| 1 Arsel |\r\n| 2 Avivah |\r\n| 3 Daiva |\r\n| 4 Wahyu |\r\n| 5 Wibi |\r\n-------------\r\n''')\r\n print('''\r\n ----------------------\r\n | 1. Linear Search |\r\n | 2. Jump Search |\r\n ----------------------\r\n ''')\r\n p1 = int(input(\"Masukan pilihan search yang ingin anda gunakan : \"))\r\n input_nama = input(\"Masukan nama ASLAB yang ingin dicari indeksnya : \")\r\n if p1 == 1:\r\n search_linear = Linear_Search(list_NAMA,input_nama)\r\n if search_linear == -1:\r\n print(input_nama,\" tidak ditemukan\")\r\n else:\r\n print(input_nama,\" ditemukan di indeks \",search_linear)\r\n exit()\r\n elif p1 == 2:\r\n search_fib = Jump_Searching(list_NAMA,input_nama,panjanglist)\r\n if search_fib == -1:\r\n print(input_nama,\" tidak ditemukan\")\r\n else:\r\n print(input_nama,\" ditemukan di indeks \",search_fib)\r\n exit()\r\n\r\n else:\r\n print(\"Input Yang Anda Masukan Salah Atau Tidak Ada\")","repo_name":"PutriNadilla/Posttest-2-ASD-","sub_path":"POSTTEST2_PUTRI.py","file_name":"POSTTEST2_PUTRI.py","file_ext":"py","file_size_in_byte":2380,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"35814486923","text":"import cv2 as cv\r\nimport numpy as np\r\nimport random as rnd\r\nimport sys\r\n\r\nfrom matplotlib import pyplot as plt\r\nfrom matplotlib.widgets import Button as btn\r\nfrom scipy.spatial.distance import cosine as csn\r\nfrom scipy.ndimage import gaussian_gradient_magnitude as ggm\r\nfrom skimage.metrics import structural_similarity as ssim\r\nfrom PyQt5 import QtCore, QtGui, QtWidgets\r\n\r\n# ----------------------------------------------------------------------------\r\n\r\nclass Ui_MainWindow(object):\r\n def setupUi(self, MainWindow):\r\n MainWindow.setObjectName(\"MainWindow\")\r\n MainWindow.resize(227, 78)\r\n MainWindow.setMinimumSize(QtCore.QSize(227, 78))\r\n MainWindow.setMaximumSize(QtCore.QSize(227, 78))\r\n MainWindow.setTabShape(QtWidgets.QTabWidget.Rounded)\r\n self.centralwidget = QtWidgets.QWidget(MainWindow)\r\n self.centralwidget.setObjectName(\"centralwidget\")\r\n self.label = QtWidgets.QLabel(self.centralwidget)\r\n self.label.setGeometry(QtCore.QRect(9, 9, 148, 16))\r\n font = QtGui.QFont()\r\n font.setPointSize(10)\r\n self.label.setFont(font)\r\n self.label.setObjectName(\"label\")\r\n self.lineEdit = QtWidgets.QLineEdit(self.centralwidget)\r\n self.lineEdit.setGeometry(QtCore.QRect(163, 9, 51, 20))\r\n font = QtGui.QFont()\r\n font.setPointSize(10)\r\n self.lineEdit.setFont(font)\r\n self.lineEdit.setObjectName(\"lineEdit\")\r\n self.pushButton = QtWidgets.QPushButton(self.centralwidget)\r\n self.pushButton.setEnabled(True)\r\n self.pushButton.setGeometry(QtCore.QRect(9, 35, 201, 31))\r\n font = QtGui.QFont()\r\n font.setPointSize(10)\r\n self.pushButton.setFont(font)\r\n self.pushButton.setObjectName(\"pushButton\")\r\n self.pushButton.clicked.connect(self.button_click)\r\n MainWindow.setCentralWidget(self.centralwidget)\r\n\r\n self.retranslateUi(MainWindow)\r\n QtCore.QMetaObject.connectSlotsByName(MainWindow)\r\n\r\n def retranslateUi(self, MainWindow):\r\n _translate = QtCore.QCoreApplication.translate\r\n MainWindow.setWindowTitle(_translate(\"MainWindow\", \"Программа\"))\r\n self.label.setText(_translate(\"MainWindow\", \"Введите кол-во человек:\"))\r\n self.pushButton.setText(_translate(\"MainWindow\", \"Принять\"))\r\n \r\n def button_click(self):\r\n Program(int(self.lineEdit.text())).show_result()\r\n \r\nclass Sketch():\r\n def __init__(self, i, tp):\r\n self.tp = tp\r\n self.sketch = cv.imread(f\"cuhk\\sketches\\{i}.jpg\")\r\n self.sketch_gray = cv.imread(f\"cuhk\\sketches\\{i}.jpg\", 0)\r\n \r\n def get_sketch(self):\r\n if self.tp == 'normal':\r\n return self.sketch\r\n \r\n elif self.tp == 'gray':\r\n return self.sketch_gray\r\n\r\nclass Photo():\r\n def __init__(self, i, tp):\r\n self.tp = tp\r\n self.photo = cv.imread(f\"cuhk\\photos\\{i}.jpg\")\r\n self.photo_gray = cv.imread(f\"cuhk\\photos\\{i}.jpg\", 0)\r\n \r\n def get_photo(self):\r\n if self.tp == 'normal':\r\n return self.photo\r\n \r\n elif self.tp == 'gray':\r\n return self.photo_gray\r\n \r\n# ----------------------------------------------------------------------------\r\n\r\nclass Program():\r\n def __init__(self, count):\r\n self.photo_dataset = []\r\n self.sketch_dataset = []\r\n self.count = count\r\n self.get_dataset()\r\n \r\n def get_dataset(self):\r\n for i in range(188):\r\n sketch = Sketch(i + 1, 'gray').get_sketch()\r\n self.sketch_dataset.append(sketch)\r\n \r\n for i in range(self.count):\r\n photo = Photo(i + 1, 'gray').get_photo()\r\n self.photo_dataset.append(photo)\r\n \r\n def create_populations(self, count):\r\n self.sketch_populations = [[] for i in range(len(self.sketch_dataset))]\r\n \r\n for i in range(len(self.sketch_dataset)):\r\n p1 = self.sketch_dataset[i].shape[0]\r\n p2 = self.sketch_dataset[i].shape[1]\r\n \r\n rnd1 = rnd.randint(5, 25)\r\n rnd2 = rnd.randint(5, 25)\r\n rnd3 = rnd.randint(-15, 15)\r\n \r\n if rnd3 < 0:\r\n x = int(rnd1 / 2)\r\n y = x\r\n z = int(rnd2 / 2 + rnd3)\r\n \r\n elif rnd3 > 0:\r\n x = int(rnd1 / 2)\r\n y = int(rnd2 / 2 + rnd3)\r\n z = x\r\n\r\n else:\r\n x = int(rnd1 / 2)\r\n y = int(rnd2 / 2)\r\n z = y\r\n \r\n sketch = self.sketch_dataset[i][x:p1 - x, y:p2 - z]\r\n \r\n new_sketch1 = cv.resize(sketch, (p2, p1))\r\n new_sketch2 = np.mean([self.sketch_dataset[i], new_sketch1], \r\n axis=0)\r\n \r\n self.sketch_populations[i].append(new_sketch1)\r\n self.sketch_populations[i].append(new_sketch2)\r\n\r\n def dct(self, photo, sketch):\r\n dct1 = cv.dct(np.float32(photo))\r\n dct2 = cv.dct(np.float32(sketch))\r\n\r\n sim = 1 - csn(dct1.flatten(), dct2.flatten())\r\n\r\n return sim * 100\r\n \r\n def dft(self, photo, sketch):\r\n dft1 = cv.dft(np.float32(photo), flags=cv.DFT_COMPLEX_OUTPUT)\r\n dft2 = cv.dft(np.float32(sketch), flags=cv.DFT_COMPLEX_OUTPUT)\r\n \r\n dist = np.linalg.norm(dft1 - dft2)\r\n max_dist = np.sqrt(photo.shape[0] * photo.shape[1] * 2) * 255\r\n\r\n return (max_dist / dist) * 1000\r\n \r\n def compare_images(self, count):\r\n self.result = [[] for i in range(count)]\r\n \r\n for i in range(count):\r\n compared = []\r\n \r\n for j in range(len(self.sketch_dataset)):\r\n compared.append(self.dct(self.photo_dataset[i],\r\n self.sketch_dataset[j]))\r\n \r\n sim1 = ssim(self.photo_dataset[i], self.sketch_populations[i][0], data_range=1.0)\r\n sim2 = ssim(self.photo_dataset[i], self.sketch_populations[i][1], data_range=1.0)\r\n \r\n self.result[i] = [np.argmax(compared), np.max(compared),\r\n sim1, sim2]\r\n \r\n def get_accuracy(self, count):\r\n self.accuracy = [[] for i in range(count)]\r\n \r\n trues = 0\r\n alls = 0\r\n\r\n for i in range(count):\r\n alls += 1\r\n\r\n if self.result[i][0] == i:\r\n trues += 1\r\n\r\n self.accuracy[i].append((trues / alls) * 100)\r\n \r\n def show_result(self):\r\n self.create_populations(self.count)\r\n self.compare_images(self.count)\r\n self.get_accuracy(self.count)\r\n \r\n self.flag = True\r\n \r\n def stop(event):\r\n self.flag = not self.flag\r\n \r\n fig = plt.figure('Результат', figsize=(16, 8))\r\n \r\n ax1 = fig.add_subplot(2, 4, 1)\r\n ax2 = fig.add_subplot(2, 4, 2)\r\n ax3 = fig.add_subplot(2, 4, 5)\r\n ax4 = fig.add_subplot(2, 4, 6)\r\n ax5 = fig.add_subplot(1, 4, 3)\r\n ax6 = fig.add_subplot(1, 4, 4)\r\n \r\n ax7 = plt.axes([0.85, 0.05, 0.1, 0.05])\r\n button = btn(ax7, 'Остановить', color='gray')\r\n button.on_clicked(stop)\r\n \r\n y = [[], []]\r\n x = []\r\n z = []\r\n \r\n for i in range(self.count):\r\n y[0].append(self.result[i][2])\r\n y[1].append(self.result[i][3])\r\n x.append(i + 1)\r\n z.append(self.accuracy[i])\r\n \r\n ax1.cla()\r\n ax1.imshow(Photo(i + 1, 'gray').get_photo(), cmap='gray')\r\n ax1.set_xticks([])\r\n ax1.set_yticks([])\r\n ax1.set_title('Фото')\r\n ax1.set_xlabel('Класс = ' + str(i + 1))\r\n \r\n ax2.cla()\r\n ax2.imshow(Sketch(self.result[i][0] + 1, 'gray').get_sketch(), \r\n cmap='gray')\r\n ax2.set_xticks([])\r\n ax2.set_yticks([])\r\n ax2.set_title('Скетч')\r\n ax2.set_xlabel('Класс = ' + str(self.result[i][0] + 1) + '\\nСходство с фото = ' +\r\n str(round(self.result[i][1], 1)) + '%')\r\n \r\n ax3.cla()\r\n ax3.imshow(self.sketch_populations[self.result[i][0]][0], \r\n cmap='gray')\r\n ax3.set_xticks([])\r\n ax3.set_yticks([])\r\n ax3.set_title('Популяция 1')\r\n \r\n ax4.cla()\r\n ax4.imshow(self.sketch_populations[self.result[i][0]][1], \r\n cmap='gray')\r\n ax4.set_xticks([])\r\n ax4.set_yticks([])\r\n ax4.set_title('Популяция 2')\r\n \r\n ax5.cla()\r\n ax5.plot(x, y[0], label='П1')\r\n ax5.plot(x, y[1], label='П2')\r\n ax5.set_xlabel('Кол-во человек')\r\n ax5.set_ylabel('SSIM')\r\n ax5.set_title('Индекс структурного сходства')\r\n ax5.legend()\r\n \r\n ax6.cla()\r\n ax6.plot(x, z)\r\n ax6.set_xlabel('Кол-во человек')\r\n ax6.set_ylabel('Точность (%)')\r\n ax6.set_title('Точность распознавания')\r\n ax6.set_yticks(np.arange(0, 110, 10))\r\n \r\n plt.subplots_adjust(wspace=0.3, hspace=0.5, top=0.95,\r\n bottom=0.2, left=0.01, right=0.96)\r\n plt.show()\r\n plt.pause(0.01)\r\n \r\n if not self.flag: \r\n break\r\n\r\n# ----------------------------------------------------------------------------\r\n\r\napp = QtWidgets.QApplication(sys.argv)\r\nMainWindow = QtWidgets.QMainWindow()\r\nui = Ui_MainWindow()\r\nui.setupUi(MainWindow)\r\nMainWindow.show()\r\nsys.exit(app.exec_())","repo_name":"donikorch/photo_sketch_detection","sub_path":"photo_sketch.py","file_name":"photo_sketch.py","file_ext":"py","file_size_in_byte":10019,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"41459096835","text":"# Finds the largest prime factor of a given number.\r\nimport math\r\ndef largestPrimeFactor(num):\r\n largestPrime = -1\r\n # First we will remove all even factors by repetetively dividing by 2. \r\n while num % 2 == 0:\r\n largestPrime = 2\r\n num /= 2\r\n # We then divide by 3 repetetively\r\n while num % 3 == 0:\r\n largestPrime = 3\r\n num /= 3\r\n # We can then divide by factrs between 5 and the square root of the original number, since the largest prime factor cannot be greater than the square root of the number itself.\r\n # We can also increment our loop by 6 each time because we have already removed all factors of 2 and 3 and therefore 6.\r\n # The code will record the largest prime during this process.\r\n for factor in range(5, int(math.sqrt(num)) + 1, 6):\r\n while num % factor == 0:\r\n largestPrime = factor\r\n num /= factor\r\n while num % (factor + 2) == 0:\r\n largestPrime = factor + 2\r\n num /= factor + 2\r\n # Either the remaining factor will be the largest prime or the largest prime will already have been recorded.\r\n if num > 4:\r\n largestPrime = num\r\n return largestPrime\r\n\r\nprint(largestPrimeFactor(600851475143))\r\n","repo_name":"sutherland11/projectEuler","sub_path":"p3.py","file_name":"p3.py","file_ext":"py","file_size_in_byte":1239,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"15230760399","text":"import streamlit as st\nfrom navigation import nav_page\nfrom edit_prepare import status\nimport os\nimport sys\nfrom streamlit_extras.switch_page_button import switch_page\n\nif \"a\" not in st.session_state:\n st.session_state[\"a\"] = 0\n\nst.set_page_config(page_title=\"Notification System\",page_icon=\"🧊\",initial_sidebar_state=\"collapsed\")\nst.session_state \n\ndef get_platform(): #These three lines are reponsible to display appropate links\n platforms = {\n 'linux' : 'https://askubuntu.com/questions/225666/copy-file-and-folder-path-from-nautilus',\n 'linux1' : 'https://askubuntu.com/questions/225666/copy-file-and-folder-path-from-nautilus',\n 'linux2' : 'https://askubuntu.com/questions/225666/copy-file-and-folder-path-from-nautilus',\n 'darwin' : 'https://support.apple.com/en-in/guide/mac-help/mchlp1774/mac',\n 'win32' : 'https://techcommunity.microsoft.com/t5/windows-11/how-do-i-copy-file-paths-on-windows-11/m-p/3061750'\n }\n return platforms[sys.platform]\ndef check_path(up_file):\n if(os.path.exists(up_file)==False): \n st.text(\"Input the Correct and Full Path\")\n st.stop()\n if(up_file[-5:]!=\".xlsx\"): \n st.text(\"Input the correct file with .xlsx extension\")\n st.stop()\n\nf_name = \"file_path.txt\"\nif os.path.exists(f_name):\n file = open(f_name,\"r+\")\nelse:\n file = open(f_name,\"a+\")\n\n\nif(\"open_page\" not in st.session_state): #It is a flag that represents the flag that tells the next page to enable its function\n st.session_state[\"open_page\"]=False\n\n\nst.title(\"Parent Notification System\")\n\nst.write(\"#\") #This hash is for the newline in the frontend\nvalue = st.slider(\"Input the value for filteration of students : \",0,100,75,key=\"value_for_filteration\")\nst.caption(\"Students below \"+str(value)+\" percent will be having their parents sent message to!\")\n \n\nst.write(\"#\")\nst.subheader(\"Upload Students Attendance File\")\nuploaded_file = st.text_input(\"Choose a Student Data File\",key=\"upld_file\",placeholder=\"Insert Path or Click The Question Button\"\n ,help=\"If you are not able to, then click the following link \"+get_platform())\nif uploaded_file is not None: #These if else statement check if the file is of xlsx extension\n check_path(uploaded_file)\n\n\nst.write(\"#\")\nst.subheader(\"Select Parent Data File\")\n\ndata = [\"Select... \"]\ndata_path = []\na=0\nfor i in file.readlines():\n # if(os.path.exists(i)==True): #Even if the path exists this function returns false \n data_path.append(i)\n # print(a,\"path :\",i,\"flag : \",os.path.exists(i),\"array : \",data_path) This line is for the debugging\n # a=a+1\ndata.extend(data_path)\n\n\nopt = st.selectbox(\"Select one of the following files : \",options=data,index=0)\n\nst.subheader(\"Type Path to the Parent Data File ! \")\ntext_input = st.text_input(\"Input Parent Data File : \",placeholder=\"Insert Path or Click The Question Button\"\n ,help=\"If you are not able to, then click the following link \"+get_platform())\nst.info('Make Sure that the File is Closed. ', icon=\"ℹ️\")\n\nbutton = st.button(\"Insert\")\ntext_flag = False\nif button:\n if text_input: #Assuming the file is correct\n check_path(text_input)\n text_flag = True\n # if text_input not in file.readlines()[:][:-3]: #This is not working\n file.write(text_input+'\\n') \n data.append(text_input)\n else:\n st.warning(\"There is nothing in the Input Box\",icon=\"🚨\")\n\nstatus_flag = False\nif(opt is not None and opt!=data[0] and not status_flag and st.session_state[\"a\"]==0):\n st.session_state[\"open_page\"]=status(uploaded_file,opt[:-1],value)\n st.session_state['a'] = st.session_state[\"a\"]+1\n\nif st.session_state[\"open_page\"]:\n with open(\"var.txt\",\"w\") as file:\n file.write(uploaded_file+\"\\n\")\n file.write(str(value)+\"\\n\")\n\n\nif(st.session_state[\"open_page\"]):\n st.subheader(\"Go to the next Page : \")\n if st.button(\"Next >\",key=\"idk\"):\n status_flag = True\n # st.session_state[\"a\"] = 0\n switch_page(\"upload student file\")\n \n\n # nav_page(\"Upload_Student_File\")\n\nif(text_flag and not st.session_state[\"open_page\"]):\n st.error(\"The Editing Was NOT Done Properly. Check if the File is close otherwise editing will not be possible\")\n\nfile.close()","repo_name":"shubh-zs/Streamlit_Project","sub_path":"Parent_Notification/Parent_Notification_v1.1/Parent_Notification_System.py","file_name":"Parent_Notification_System.py","file_ext":"py","file_size_in_byte":4454,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"40584897952","text":"from __future__ import print_function\n\nimport sys\n\n\ntry:\n import urlparse as urllib\nexcept ImportError:\n import urllib.parse as urllib\n\ntry:\n import StringIO\n BytesIO = StringIO.StringIO\nexcept ImportError:\n import io\n BytesIO = io.BytesIO\n\n\nurlparse = urllib.urlparse\nurljoin = urllib.urljoin\n\nif sys.version_info < (3,):\n text = unicode\nelse:\n text = str\n\nif sys.version_info < (3,):\n # exceptions in python2, cannot contain unicode chars\n # or they will error when being printed\n # replace all characters with ascii-safe equivalents\n def clean_exc_message(msg):\n return msg.encode('ascii', 'replace').decode('ascii')\nelse:\n def clean_exc_message(msg):\n return msg\n\nif sys.version_info < (3,):\n import codecs\n\n def replace_with_question_mark(exc):\n return (u'?', exc.end)\n\n codecs.register_error('replace_with_question_mark',\n replace_with_question_mark)\n\n def make_text(val):\n if isinstance(val, str):\n u_val = unicode(val, errors='replace_with_question_mark')\n return u\"b'\" + u_val + u\"'\"\n else:\n return unicode(val)\nelse:\n make_text = str\n\n\nprintf = print\n","repo_name":"Paczesiowa/screenshots","sub_path":"dom2img/_compat.py","file_name":"_compat.py","file_ext":"py","file_size_in_byte":1208,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"3871280357","text":"import pyautogui\nimport time\nfrom circles_calibration import get_circles\nfrom filter_function import create_range\nfrom get_resolution import get_resolution\nfrom settings_folder import settings_folder\n\n\ndef calibrate_colors(image_name):\n color_names = ['iridescent.txt', 'purple.txt', 'green.txt', 'yellow.txt', 'brown.txt']\n for name in color_names:\n name = f'{settings_folder}{name}'\n create_range(image_name, name)\n\n\ndef calibrate_circles(image_name):\n color_name = f'{settings_folder}/neighbours.txt'\n output = create_range(image_name, color_name)\n get_circles(output)\n\n\nif __name__ == '__main__':\n time.sleep(5)\n image = pyautogui.screenshot()\n crop_img = get_resolution(image)\n calibrate_circles(crop_img)\n calibrate_colors(crop_img)\n","repo_name":"Dikzamen/dbd_bloodweb_python","sub_path":"calibration.py","file_name":"calibration.py","file_ext":"py","file_size_in_byte":783,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"34060789837","text":"import io\nimport os\nimport yaml\nimport pymysql\nfrom logger import setup_log\n\nlogger = setup_log(__name__)\nDEFAULT_CONFIG_LOCATION = 'config.yml'\n\n\nclass mysql:\n\n def __init__(self, config):\n \"\"\"\n 实例化链接对象\n :param config: 数据配置参数\n \"\"\"\n self.connection = pymysql.connect(user=config['user'],\n password=config['password'],\n port=config['port'],\n host=config['host'],\n db=config['db'],\n charset=config['charset'])\n\n def fetchone_db(self, sql):\n \"\"\"\n 数据查询\n :param sql: sql语句\n :return: sql结果\n \"\"\"\n c = self.connection.cursor(cursor=pymysql.cursors.DictCursor)\n self.connection.ping(reconnect=True)\n c.execute(sql)\n return c.fetchone()\n\n def fetchall_db(self, sql):\n \"\"\"\n 数据查询\n :param sql: sql语句\n :return: sql结果\n \"\"\"\n c = self.connection.cursor(cursor=pymysql.cursors.DictCursor)\n self.connection.ping(reconnect=True)\n c.execute(sql)\n return c.fetchall()\n\n def exe(self, sql):\n \"\"\"\n 数据添加\n :param sql: sql语句\n \"\"\"\n c = self.connection.cursor(cursor=pymysql.cursors.DictCursor)\n self.connection.ping(reconnect=True)\n c.execute(sql)\n self.connection.commit()\n\n def rollback(self):\n \"\"\"\n 回滚\n \"\"\"\n self.connection.ping(reconnect=True)\n self.connection.rollback()\n\n\nclass InvalidConfigError(ValueError):\n \"\"\"如果遇到无效的配置.就会引发此异常\n \"\"\"\n\n def __init__(self, message):\n super(InvalidConfigError, self).__init__(message)\n\n\ndef read_file(filename, encoding=\"utf-8-sig\"):\n \"\"\"\n 从本地读入一个文件\n \"\"\"\n with io.open(filename, encoding=encoding) as f:\n return f.read()\n\n\ndef fix_yaml_loader():\n \"\"\"确保读出的yaml文件内容\n 可以被unicode编码\n \"\"\"\n from yaml import Loader, SafeLoader\n\n def construct_yaml_str(self, node):\n # Override the default string handling function\n # to always return unicode objects\n return self.construct_scalar(node)\n\n Loader.add_constructor(u'tag:yaml.org,2002:str', construct_yaml_str)\n SafeLoader.add_constructor(u'tag:yaml.org,2002:str', construct_yaml_str)\n\n\ndef read_yaml(content):\n \"\"\"读入yaml文件\n \"\"\"\n fix_yaml_loader()\n return yaml.safe_load(content)\n\n\ndef read_yaml_file(filename):\n \"\"\"\n 从本地读入yaml文件\n \"\"\"\n fix_yaml_loader()\n return yaml.safe_load(read_file(filename, \"utf-8\"))\n\n\ndef load_config(filename=DEFAULT_CONFIG_LOCATION, **kwargs):\n \"\"\"\n input:\n filename:\n \"/Config/config.yml\"\n output:\n items:\n {'language': 'zh',\n 'neo4j':\n {'host': '127.0.0.1',\n 'http_port': 7474,\n 'user': 'neo4j',\n 'password': '123456'\n }\n }\n \"\"\"\n if filename is None and os.path.isfile(DEFAULT_CONFIG_LOCATION):\n filename = DEFAULT_CONFIG_LOCATION\n if filename is not None:\n try:\n file_config = read_yaml_file(filename)\n except Exception as e:\n error = \"Failed to read configuration file '{}'. Error: {}\".format(filename, e)\n logger.error(error)\n raise InvalidConfigError(error)\n if kwargs:\n file_config.update(kwargs)\n return file_config\n else:\n return kwargs\n","repo_name":"lsq960124/Flask-BookRecommend-Mysql","sub_path":"web/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":3831,"program_lang":"python","lang":"en","doc_type":"code","stars":219,"dataset":"github-code","pt":"7"} +{"seq_id":"28522305857","text":"import numpy as np\nfrom boo.utils import (\n _as_pairs,\n _chop_off_ends,\n _gather_list_of_von_neumann_neighbors,\n _gather_list_of_moore_neighbors,\n _gather_list_of_1D_neighbors,\n _multiconvolve,\n)\nfrom typing import Type\n\n\nclass GhostArray:\n def __init__(\n self,\n interior: np.ndarray,\n pad_width: list,\n mode: str = \"dirichlet\",\n constant_values: list = 0,\n ghost_array: np.ndarray = None,\n ):\n \"\"\"\n args:\n interior numpy or cupy array, arbitrary size, no ghost cells\n ghost_array numpy or cupy array, arbitrary size, includes ghost cells\n overwritten if interior is defined\n pad_width int a\n applies a ghost width of a around the entire array\n tuple (l,r)\n applies a ghost width of l on the left of each dimension,\n r on the right of each dimension\n iterable of tuples ((l0, r0), ...)\n applies a ghost width of l1 on the left of axis 0, ...,\n r1 on the right of axis 0, ...\n iterable of tuples ((s0,), ...)\n applies a ghost width of s1 on the left and right\n of axis 0, ...\n mode 'dirichlet', 'periodic'\n constant_values same format as pad_width, for 'dirichlet' mode\n \"\"\"\n self.interior = interior\n self.ndim = interior.ndim if self.interior is not None else ghost_array.ndim\n\n # configure arguments for np.pad()\n self.mode = mode\n pad_width_np = np.asarray(pad_width)\n if np.any(pad_width_np < 0):\n raise BaseException(\"Negative pad width.\")\n self.pad_width = _as_pairs(pad_width_np, self.ndim, as_index=True)\n constant_values_np = np.asarray(constant_values)\n self.constant_values = _as_pairs(constant_values_np, self.ndim, as_index=False)\n\n # define ghost array\n if self.interior is not None:\n self.shape = self.interior.shape\n self._compute_ghost_zone()\n else:\n self.ghost_array = ghost_array\n self.dtype = ghost_array.dtype\n self.slices = None # compute slices only as needed\n\n def _config_np_pad(self) -> None:\n \"\"\"\n helper method for _compute_ghost_zone\n args:\n self.mode\n self.pad_width\n self.constant_values\n returns:\n self._np_pad_config argument dictionary for np.pad()\n \"\"\"\n np_pad_configs = {\n \"periodic\": {\"pad_width\": self.pad_width, \"mode\": \"wrap\"},\n \"dirichlet\": {\n \"pad_width\": self.pad_width,\n \"mode\": \"constant\",\n \"constant_values\": self.constant_values,\n },\n }\n self._np_pad_config = np_pad_configs[self.mode]\n\n def _compute_ghost_zone(self) -> None:\n \"\"\"\n args:\n self.interior\n self._np_pad_config\n returns:\n self.ghost_array same shape as self.interior, includes ghost zones\n self.dtype\n \"\"\"\n self._config_np_pad()\n self.ghost_array = np.pad(array=self.interior, **self._np_pad_config)\n self.dtype = self.ghost_array.dtype\n\n def _compute_slices(self) -> None:\n \"\"\"\n helper method for _compute_interior\n args:\n self.pad_width\n returns:\n self.slices list of slices that index ghost_array into interior\n \"\"\"\n if self.slices is None:\n self.slices = [\n slice(left, -right or None) for left, right, in self.pad_width\n ]\n\n def _compute_interior(self) -> None:\n \"\"\"\n args:\n self.ghost_array\n self.slices\n returns:\n self.interior defined region of ghost_array with no padding\n \"\"\"\n if self.interior is None:\n self._compute_slices()\n self.interior = self.ghost_array[tuple(self.slices)]\n self.shape = self.interior.shape\n\n def to_numpy(self) -> np.ndarray:\n \"\"\"\n returns\n self.interior\n \"\"\"\n self._compute_interior()\n return self.interior\n\n def add_along_axis(\n self, axis: list, pad_width: list, constant_values: list = 0\n ) -> Type[\"GhostArray\"]:\n \"\"\"\n modify the padding along specified axes\n args:\n axis\n pad_width\n constant_values\n returns:\n self.ghost_array\n self.pad_width\n self.constant_values\n self.slices\n \"\"\"\n # configure pad_width and constant_values\n axes = [axis] if isinstance(axis, int) else axis\n nadd = len(axes)\n pad_width_np = np.asarray(pad_width)\n delta_pad_width = _as_pairs(pad_width_np, nadd, as_index=True)\n constant_values_np = np.asarray(constant_values)\n delta_constant_values = _as_pairs(constant_values_np, nadd, as_index=False)\n # new attributes and arguments for np.pad()\n new_pad_width = self.pad_width.copy()\n new_constant_values = self.constant_values.copy()\n np_pad_argument = [(0, 0)] * self.ndim\n constant_values_argument = [(0, 0)] * self.ndim\n for i, ax in enumerate(axes):\n l_old, r_old = self.pad_width[ax]\n if l_old != 0 or r_old != 0:\n self.remove_along_axis(axis=ax)\n new_pad_width[ax] = [delta_pad_width[i][0], delta_pad_width[i][1]]\n np_pad_argument[ax] = [delta_pad_width[i][0], delta_pad_width[i][1]]\n if self.mode == \"dirichlet\":\n new_constant_values[ax] = [\n delta_constant_values[i][0],\n delta_constant_values[i][1],\n ]\n constant_values_argument[ax] = [\n delta_constant_values[i][0],\n delta_constant_values[i][1],\n ]\n # temporarily set pad_width and constant_values to deltas to trick pad_config\n self.pad_width = np_pad_argument\n if self.mode == \"dirichlet\":\n self.constant_values = constant_values_argument\n self._config_np_pad()\n self.ghost_array = np.pad(array=self.ghost_array, **self._np_pad_config)\n # reset pad_width and constant_values to true values\n self.pad_width = new_pad_width\n if self.mode == \"dirichlet\":\n self.constant_values = new_constant_values\n self.slices = None # slices are no longer accurate\n return self\n\n def remove_along_axis(self, axis: list = None) -> Type[\"GhostArray\"]:\n \"\"\"\n remove the padding along specified axes\n args:\n axis removes all padding if 'None'\n returns:\n self.ghost_array\n self.pad_width\n self.constant_values\n self.slices\n \"\"\"\n if axis is None:\n self._compute_slices()\n self.ghost_array = self.ghost_array[tuple(self.slices)]\n self.pad_width = [(0, 0)] * self.ndim\n self.constant_values = [(0, 0)] * self.ndim\n self.slices = None\n return self\n if isinstance(axis, int):\n self.ghost_array = _chop_off_ends(\n array=self.ghost_array, chop_size=self.pad_width[axis], axis=axis\n )\n self.pad_width[axis] = [0, 0]\n self.constant_values[axis] = [0, 0]\n self.slices = None\n return self\n rm_slice = [slice(None)] * self.ndim\n new_pad_width = self.pad_width.copy()\n new_constant_values = self.constant_values.copy()\n for i in axis:\n rm_slice[i] = slice(self.pad_width[i][0], -self.pad_width[i][1] or None)\n new_pad_width[i] = [0, 0]\n new_constant_values[i] = [0, 0]\n self.ghost_array = self.ghost_array[tuple(rm_slice)]\n self.pad_width = new_pad_width\n self.constant_values = new_constant_values\n self.slices = None\n return self\n\n def convolve(\n self, kernel: np.ndarray, axis: int = 0, bias_shift: int = 0\n ) -> Type[\"GhostArray\"]:\n \"\"\"\n 1d convolution of a kernel on an array\n args:\n kernel 1d array of weights\n axis int\n bias_shift int, defaults to 0\n [X, _, _, _, X] for even kernels\n [X, _, _, _, X + 1] for odd kernels\n returns:\n new GhostArray instance with less padding in the axis dimension\n \"\"\"\n # find kernel dimensions\n kernel_np = np.asarray(kernel)\n n_kernel = kernel_np.size\n # perform convolutions\n convolved_array = _multiconvolve(\n array=self.ghost_array, kernels=kernel_np[np.newaxis], axis=axis\n )[0]\n # find amount to chop off either end of array\n # if n_kernel is even and bias_shift = 0, right_rm > left_rm by 1\n left_rm = n_kernel // 2 + n_kernel % 2 - 1 - bias_shift\n right_rm = n_kernel // 2 + bias_shift\n # remove spent padding\n new_pad_width = np.asarray(self.pad_width)\n new_pad_width[axis] -= np.array([left_rm, right_rm])\n # remove constant if padding is depleted\n new_constant_values = np.asarray(self.constant_values)\n new_constant_values[axis] = np.where(\n new_pad_width[axis] == 0, 0, new_constant_values[axis]\n )\n out = self.__class__(\n ghost_array=convolved_array,\n interior=None,\n pad_width=new_pad_width,\n mode=self.mode,\n constant_values=new_constant_values,\n )\n return out\n\n def multiconvolve(\n self, kernels: np.ndarray, axis: int = 0, bias_shift: int = 0\n ) -> Type[\"GhostArray\"]:\n \"\"\"\n 1d convolution of a multiple kernels on an array\n args:\n kernels 2d array of weights, shape (N_kernels, n_kernel)\n axis int\n bias_shift int, defaults to 0\n [X, _, _, _, X] for even kernels\n [X, _, _, _, X + 1] for odd kernels\n returns:\n new GhostArray instance with a new first dimension of length N_kernels and\n an axis + 1 dimension with reduced length\n \"\"\"\n # find kernel dimensions\n kernels_np = np.asarray(kernels)\n N_kernels, n_kernel = kernels_np.shape\n # perform convolutions\n convolved_array = _multiconvolve(\n array=self.ghost_array, kernels=kernels_np, axis=axis\n )\n # find amount to chop off either end of array\n # if n_kernel is even and bias_shift = 0, right_rm > left_rm by 1\n left_rm = n_kernel // 2 + n_kernel % 2 - 1 - bias_shift\n right_rm = n_kernel // 2 + bias_shift\n # remove spent padding\n new_pad_width = np.concatenate((np.array([[0, 0]]), np.asarray(self.pad_width)))\n new_pad_width[axis + 1] -= np.array([left_rm, right_rm])\n # remove constant if padding is depleted\n new_constant_values = np.concatenate(\n (np.array([[0, 0]]), np.asarray(self.constant_values))\n )\n new_constant_values[axis + 1] = np.where(\n new_pad_width[axis + 1] == 0, 0, new_constant_values[axis + 1]\n )\n out = self.__class__(\n ghost_array=convolved_array,\n interior=None,\n pad_width=new_pad_width,\n mode=self.mode,\n constant_values=new_constant_values,\n )\n return out\n\n def apply_f_to_neighbors(\n self, f: callable, axis: list = None, mode: str = \"neumann\", **kwargs\n ) -> Type[\"GhostArray\"]:\n \"\"\"\n ars:\n axis spatial dimensions\n mode 'neumann' (no corners), 'moore' (corners)\n returns:\n new GhostArray instance with neighbors arranged along a new first dimension\n \"\"\"\n if self.ndim == 1:\n axis = 0 if axis is None else axis\n neighbors = _gather_list_of_1D_neighbors(array=self.ghost_array, axis=axis)\n elif mode == \"neumann\":\n axis = list(range(self.ndim)) if axis is None else axis\n neighbors = _gather_list_of_von_neumann_neighbors(\n array=self.ghost_array, axes=axis\n )\n elif mode == \"moore\":\n axis = list(range(self.ndim)) if axis is None else axis\n neighbors = _gather_list_of_moore_neighbors(\n array=self.ghost_array, axes=axis\n )\n # apply f\n new_array = f(neighbors, axis=0, **kwargs)\n # remove spent padding\n new_pad_width = np.asarray(self.pad_width)\n new_pad_width[axis] -= np.array([1, 1])\n # remove constant if padding is depleted\n new_constant_values = np.asarray(self.constant_values)\n new_constant_values[axis] = np.where(\n new_pad_width[axis] == 0, 0, new_constant_values[axis]\n )\n out = self.__class__(\n ghost_array=new_array,\n interior=None,\n pad_width=new_pad_width,\n mode=self.mode,\n constant_values=new_constant_values,\n )\n return out\n\n def __repr__(self) -> str:\n return (\n f\"GhostArray(interior = {self.interior}\"\n + f\", pad_width = {self.pad_width}, mode = {self.mode})\"\n )\n\n def __eq__(self, other) -> bool:\n interior_is_eq = np.all(self.interior == other.interior)\n ghost_array_is_eq = np.all(self.ghost_array == other.ghost_array)\n return interior_is_eq and ghost_array_is_eq\n\n def __neg__(self) -> Type[\"GhostArray\"]:\n self._compute_interior()\n if self.mode == \"dirichlet\":\n return self.__class__(\n interior=-self.interior,\n pad_width=self.pad_width,\n mode=self.mode,\n constant_values=-np.asarray(self.constant_values),\n )\n return self.__class__(\n interior=-self.interior,\n pad_width=self.pad_width,\n mode=self.mode,\n constant_values=self.constant_values,\n )\n\n def __add__(self, other):\n raise NotImplementedError(f\"{self.__class__} + {type(other)}\")\n\n def __radd__(self, other):\n raise NotImplementedError(f\"{type(other)} + {self.__class__}\")\n\n def __sub__(self, other):\n raise NotImplementedError(f\"{self.__class__} - {type(other)}\")\n\n def __rsub__(self, other):\n raise NotImplementedError(f\"{type(other)} - {self.__class__}\")\n\n def __mul__(self, other) -> Type[\"GhostArray\"]:\n self._compute_interior()\n if isinstance(other, int) or isinstance(other, float):\n if self.mode == \"dirichlet\":\n return self.__class__(\n interior=self.interior * other,\n pad_width=self.pad_width,\n mode=self.mode,\n constant_values=np.asarray(self.constant_values) * other,\n )\n return self.__class__(\n interior=self.interior * other,\n pad_width=self.pad_width,\n mode=self.mode,\n constant_values=self.constant_values,\n )\n raise NotImplementedError(f\"{self.__class__} * {type(other)}\")\n\n def __rmul__(self, other):\n return self.__mul__(other)\n\n def __floordiv__(self, other):\n raise NotImplementedError(f\"{self.__class__} // {type(other)}\")\n\n def __rfloordiv__(other, self):\n raise NotImplementedError(f\"{type(other)} // {self.__class__}\")\n\n def __truediv__(self, other):\n if isinstance(other, float) or isinstance(other, int):\n return self.__mul__(1 / other)\n raise NotImplementedError(f\"{type(other)} / {self.__class__}\")\n\n def __rtruediv__(self, other):\n raise NotImplementedError(f\"{self.__class__} / {type(other)}\")\n","repo_name":"jpalafou/boopy","sub_path":"boo/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":16155,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"17089184559","text":"import os\nfrom typing import Any\nfrom conans import ConanFile, CMake\n\n\nVALID_MAYA_CONFIGS: dict[tuple[str, str], set[str]] = {\n ('Visual Studio', '16'): { '2022', '2023' },\n ('gcc', '7'): { '2022', '2023' },\n ('gcc', '9'): { '2022', '2023' },\n ('apple-clang', '10.0'): { '2022', '2023' }\n}\n\nSETTINGS: dict[str, Any] = {\n 'os': ['Windows', 'Linux', 'Macos'],\n 'compiler': {\n 'Visual Studio': {'version': ['16']},\n 'gcc': {'version': ['7', '9']},\n 'apple-clang': {'version': ['10.0']}\n },\n 'build_type': None,\n 'arch': 'x86_64'\n}\n\nTOOL_REQUIRES: list[str] = [\n 'cmake/3.24.1',\n 'thinkboxcmlibrary/1.0.0'\n]\n\nREQUIRES: list[str] = [\n 'thinkboxlibrary/1.0.0',\n 'mayasdk/1.0.0',\n 'tinyxml2/9.0.0'\n]\n\n\nclass ThinkboxMYLibraryConan(ConanFile):\n name: str = 'thinkboxmylibrary'\n version: str = '1.0.0'\n license: str = 'Apache-2.0'\n description: str = 'Shared code for Thinkbox\\'s Maya plugins'\n settings: dict[str, Any] = SETTINGS\n requires: list[str] = REQUIRES\n tool_requires: list[str] = TOOL_REQUIRES\n generators: str | list[str] = 'cmake_find_package'\n options: dict[str, Any] = {\n 'maya_version': ['2022', '2023']\n }\n \n def configure(self) -> None:\n if self.options.maya_version == None:\n self.options.maya_version = '2022'\n self.options['mayasdk'].maya_version = self.options.maya_version\n\n def validate(self) -> None:\n if self.options.maya_version != self.options['mayasdk'].maya_version:\n raise Exception('Option \\'maya_version\\' must be the same as mayasdk')\n compiler = str(self.settings.compiler)\n compiler_version = str(self.settings.compiler.version)\n compiler_tuple = (compiler, compiler_version)\n maya_version = str(self.options.maya_version)\n if maya_version not in VALID_MAYA_CONFIGS[compiler_tuple]:\n raise Exception(f'{str(compiler_tuple)} is not a valid configuration for Maya {maya_version}')\n\n def build(self) -> None:\n cmake = CMake(self)\n cmake.configure(defs={\n 'MAYA_VERSION': self.options.maya_version\n })\n cmake.build()\n\n def export_sources(self) -> None:\n self.copy('**.h', src='', dst='')\n self.copy('**.hpp', src='', dst='')\n self.copy('**.cpp', src='', dst='')\n self.copy('**.cmake', src='', dst='')\n self.copy('CMakeLists.txt', src='', dst='')\n self.copy('NOTICE.txt', src='', dst='')\n self.copy('LICENSE.txt', src='', dst='')\n\n def package(self) -> None:\n cmake = CMake(self)\n cmake.install()\n\n with open(os.path.join(self.source_folder, 'NOTICE.txt'), 'r', encoding='utf8') as notice_file:\n notice_contents = notice_file.readlines()\n with open(os.path.join(self.source_folder, 'LICENSE.txt'), 'r', encoding='utf8') as license_file:\n license_contents = license_file.readlines()\n os.makedirs(os.path.join(self.package_folder, 'licenses'), exist_ok=True)\n with open(os.path.join(self.package_folder, 'licenses', 'LICENSE'), 'w', encoding='utf8') as cat_license_file:\n cat_license_file.writelines(notice_contents)\n cat_license_file.writelines(license_contents)\n\n def deploy(self) -> None:\n self.copy('*', dst='bin', src='bin')\n self.copy('*', dst='lib', src='lib')\n self.copy('*', dst='include', src='include')\n\n def package_info(self) -> None:\n self.cpp_info.libs = [\"thinkboxmylibrary\"]\n","repo_name":"aws/thinkbox-my-library","sub_path":"conanfile.py","file_name":"conanfile.py","file_ext":"py","file_size_in_byte":3527,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"7"} +{"seq_id":"70414983582","text":"import numpy as np\nfrom numpy.random import choice\nfrom scipy.stats import multivariate_normal, norm\n\nfrom kerehmm.util import random_simplex\n\n\nclass Distribution(object):\n def get_probability(self, observation, *args, **kwargs):\n raise NotImplementedError()\n\n def __getitem__(self, item, *args, **kwargs):\n return self.get_probability(item, *args, **kwargs)\n\n def emit(self):\n \"\"\"\n Emits a randomly drawn output.\n :return:\n \"\"\"\n raise NotImplementedError()\n\n\nclass DiscreteDistribution(Distribution):\n \"\"\"\n I am a discrete distribution.\n \"\"\"\n\n def __init__(self, n, randomize=False):\n self.n = n\n # initialize to 1/n\n self.probabilities = np.array([1. / n] * n)\n if randomize:\n self.probabilities = random_simplex(n)\n\n def get_probability(self, observation, *args, **kwargs):\n return self.probabilities[observation]\n\n def emit(self):\n return choice(range(self.n), p=self.probabilities)\n\n def b_coefficient(self, other_dist):\n \"\"\"\n Bhattacharyya coefficient\n\n :return:\n\n >>> d1 = DiscreteDistribution(3)\n >>> d1.probabilities\n array([ 0.33333333, 0.33333333, 0.33333333])\n >>> d2 = DiscreteDistribution(3)\n >>> d2.probabilities\n array([ 0.33333333, 0.33333333, 0.33333333])\n >>> d1.b_coefficient(d2) == d2.b_coefficient(d1)\n True\n >>> expected = np.sqrt(np.sum([1./3 ** 2] * 3))\n >>> expected\n 0.57735026918962573\n >>> d1.b_coefficient(d2) == expected\n True\n \"\"\"\n assert isinstance(other_dist, DiscreteDistribution)\n assert other_dist.n == self.n\n running_sum = np.sum(self.probabilities * other_dist.probabilities)\n return np.sqrt(running_sum)\n\n def b_distance(self, other_dist):\n coef = self.b_coefficient(other_dist)\n return -np.log(coef)\n\n\n\n\nclass GaussianDistribution(Distribution):\n \"\"\"\n I am a gaussian distribution.\n\n It is initialized to mean 0 across dimensions, and\n the covariance is set to a scalar matrix with 1\n across the diagonal.\n >>> m = GaussianDistribution(2)\n >>> m.mean\n array([ 0., 0.])\n >>> m.variance\n array([[ 1., 0.],\n [ 0., 1.]])\n\n You can simply use bracket indexing to get the probability of\n an observation. Note that all probabilities are in log scale.\n >>> from scipy import stats\n >>> m[(0, 0)] == stats.multivariate_normal(mean=m.mean, cov=m.variance).pdf((0,0))\n True\n\n It also supports univariate gaussians.\n >>> m = GaussianDistribution(1)\n >>> m.mean\n 0\n >>> m.variance\n 1\n >>> m[0] == stats.norm(loc=0, scale=1).pdf(0)\n True\n\n You can also randomize the means upon initialization by passing random=True.\n Optionally, you can also pass lower_bounds=[..] and upper_bounds[..] to specify\n boundaries of each dimension when initializing the distribution.\n >>> m = GaussianDistribution(2, random=True, lower_bounds=[1,1], upper_bounds=[2,2])\n >>> all([1 <= d <= 2 for d in m.mean])\n True\n \"\"\"\n\n def __init__(self, dimensions, random=False, lower_bounds=0, upper_bounds=100,\n mean=None, variance=None):\n self.dimensions = dimensions\n\n def mean_():\n if random:\n if dimensions == 1:\n return np.random.uniform(lower_bounds, upper_bounds, size=self.dimensions)[0]\n else:\n return np.random.uniform(lower_bounds, upper_bounds, size=self.dimensions)\n else:\n if mean is not None:\n try:\n if dimensions != 1:\n assert mean.shape == (self.dimensions,)\n return mean\n except AssertionError:\n raise ValueError(\"Mean has invalid shape: \\nmean\\t=\\t{}\".format(mean))\n else:\n if self.dimensions == 1:\n return 0.\n else:\n return np.zeros(shape=(self.dimensions,))\n\n def variance_():\n if variance is not None:\n return variance\n if self.dimensions == 1:\n if random:\n return 1.\n # return np.random.random()\n else:\n return 1.\n else:\n variances = np.zeros((self.dimensions, self.dimensions))\n np.fill_diagonal(variances, 1)\n return variances\n\n self.mean = mean_()\n self.variance = variance_()\n\n def get_probability(self, observation, *args, **kwargs):\n if self.dimensions != 1:\n return multivariate_normal(mean=self.mean, cov=self.variance).pdf(observation)\n else:\n return norm(loc=self.mean, scale=self.variance).pdf(observation)\n\n def emit(self):\n if self.dimensions != 1:\n return multivariate_normal(mean=self.mean, cov=self.variance).rvs()\n else:\n return norm(loc=self.mean, scale=self.variance).rvs()\n\n def __str__(self):\n string = \\\n \"\"\"\n Gaussian(ndim={}, mean={}, covar={})\n \"\"\".format(self.dimensions, self.mean, self.variance)\n return string\n\n def __repr__(self):\n return self.__str__()\n\nclass GaussianMixture(Distribution):\n \"\"\"\n I am a mixture of continuous gaussians.\n\n A single component mixture is essentially a Gaussian.\n It is initialized to mean 0 across dimensions, and\n the covariance is set to a scalar matrix with 1\n across the diagonal.\n >>> m = GaussianMixture(1, 2)\n >>> m.means\n array([[ 0., 0.]])\n >>> m.variances\n array([[[ 1., 0.],\n [ 0., 1.]]])\n\n You can simply use bracket indexing to get the probability of\n an observation. Note that all probabilities are in log scale.\n >>> from numpy import exp\n >>> exp(m[(0, 0)])\n 0.15915494309189535\n >>> exp(m[(1, 2)])\n 0.013064233284684921\n \"\"\"\n\n def __init__(self, nmixtures, dimensions):\n self.nmixtures = nmixtures\n self.dimensions = dimensions\n self.means = np.zeros(shape=(nmixtures, dimensions))\n self.variances = np.zeros(shape=(nmixtures, dimensions, dimensions))\n map(lambda x: np.fill_diagonal(x, 1), self.variances)\n self.weights = np.zeros(shape=(nmixtures,))\n self.weights[:] = np.log(1. / nmixtures)\n\n def get_probability(self, observation, *args, **kwargs):\n running_sum = -np.inf\n for weight, mean, variance in zip(self.weights, self.means, self.variances):\n dist = multivariate_normal(mean=mean,\n cov=variance)\n prob = np.log(np.sum(dist.pdf(observation)))\n running_sum = np.logaddexp(weight + prob,\n running_sum)\n return running_sum\n\n def emit(self):\n mixture = choice(range(self.nmixtures), p=np.exp(self.weights))\n dist = multivariate_normal()\n return dist.rvs(mean=self.means[mixture], cov=self.variances[mixture])\n","repo_name":"keryil/kereHMM","sub_path":"kerehmm/distribution.py","file_name":"distribution.py","file_ext":"py","file_size_in_byte":7210,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"31588696035","text":"#!/usr/bin/env python3\n\"\"\"\nA sample pybowler pipeline that demonstrates replacement of\nvarious types of string-interpolation with f-strings.\n\nImplemented:\n * 'stringliteral %s' % name\n --> f'stringliteral {name}'\n\n * 'stringliteral %s %d' % (foo, bar)\n --> f'stringliteral {foo} {bar}'\n\n * 'stringliteral {} {bar}'.format(foo, bar=bar)\n --> f'stringliteral {foo} {bar}'\n\nTODO: plenty:\n * Handle other printf-style things other than %s/%d/%f, e.g. '%20d' etc.\n * Same for .format() - handle '!20d' etc\n * Handle old-style dict interpolation, e.g. '%(a)s' % {'a': 'a'}\n\"\"\"\n\nimport argparse\nimport re\nimport sys\n\nfrom bowler import Query, TOKEN, SYMBOL\nfrom bowler.types import Leaf, Node, STARS\n\nflags = {}\n\n\nRE_OLD_INTERPOLATION_BASIC = re.compile(r'(? f'{xyz}'\n \"\"\"\n formatstring = capture['formatstring']\n interpolation_args = capture['interpolation_args']\n if isinstance(interpolation_args, Leaf):\n # string interpolation (old style), where the thing on the right is a name.\n # e.g. `'foo %s' % bar\n interpolation_args = [interpolation_args.value]\n elif isinstance(interpolation_args, list):\n # string interpolation (old style), where the thing on the right is a tuple.\n # e.g. `'foo %s %s' % (bar, baz)\n # first, find the 'bar' and 'baz' bits:\n interpolation_args = [\n o.value\n for o in interpolation_args\n if isinstance(o, Leaf) and o.type == TOKEN.NAME\n ]\n\n if len(RE_OLD_INTERPOLATION_BASIC.findall(formatstring.value)) != len(interpolation_args):\n # TODO: The arguments don't line up 1:1 with the number of '%s' bits.\n # This could be a bug in the program.\n # More likely, it's because our regex isn't that inclusive.\n # e.g. if one of them is '%.20f' we'll miss that one and skip the whole expression.\n # We could implement this, by using f'{foo!.20f}' for that case.\n return node\n\n # Replace all occurrences of '%s' in the formatstring with the matching '{argumentname}'\n replacement_value = RE_OLD_INTERPOLATION_BASIC.sub(\n lambda matchobj: ('{%s}' % interpolation_args.pop(0)),\n formatstring.value,\n )\n\n # Make sure we consumed all the arguments, otherwise something went wrong.\n assert not interpolation_args\n\n # Convert to an f-string.\n try:\n replacement_value = add_f_prefix(replacement_value)\n except SkipString:\n return node\n\n if flags['debug']:\n print(f\"Interpolating (old-style) format-string:\\n\\t{formatstring}\")\n print(f\"With arguments:\\n\\t{interpolation_args}\")\n print(f\"Replacement formatstring: {replacement_value}\")\n print()\n\n # Finally, replace the formatstring node in the CST, and remove the operator & operand.\n formatstring.value = replacement_value\n node.children[1:] = []\n\n return node\n\n\ndef _interpret_format_arguments(arg):\n \"\"\"\n Recursive generator.\n\n Given a single name, an argument, an arglist or a vararg expression,\n returns actual argument values, in order.\n\n Yields one of:\n positional args: yields the arg value\n keyword args: yields a dict {k: v}\n other (complex nested structures or varargs): yields None.\n )\n \"\"\"\n if isinstance(arg, Leaf) and arg.type == TOKEN.COMMA:\n # Skip comma tokens between actual args\n return\n\n if isinstance(arg, list):\n # Handle top-level list of args. Also handles there being no args at all: .format()\n for sub_arg in arg:\n yield from _interpret_format_arguments(sub_arg)\n return\n\n if isinstance(arg, Node) and arg.type == SYMBOL.arglist:\n # Multiple arguments, may be either keyword or positional\n for child in arg.children:\n yield from _interpret_format_arguments(child)\n return\n\n if isinstance(arg, Node) and arg.type == SYMBOL.argument:\n if arg.children[0].type in STARS:\n # *args, or **kwargs.\n # Not useful for f-stringing. Give up.\n raise SkipString\n\n # Single keyword argument: .format(keyword=value)\n # The three child nodes here are (keyword, '=', value).\n value = arg.children[2]\n if not isinstance(value, Leaf):\n # Might be complex? Give up. This stops parsing of the entire expression,\n # beacuse having an f-string *and* a .format() is pretty nasty.\n raise SkipString\n else:\n yield {\n arg.children[0].value: value.value\n }\n elif isinstance(arg, Leaf) and arg.type == TOKEN.NAME:\n # Single positional argument, which is just a name.\n yield arg.value\n else:\n # Something else.\n # Might be a complex expression? Give up. This stops parsing of the entire expression,\n # because having an f-string *and* a .format() is pretty nasty.\n raise SkipString\n\n\ndef format_method_to_fstrings(node, capture, filename):\n \"\"\"\n '{}'.format(xyz)\n --> f'{xyz}'\n \"\"\"\n\n if flags['debug']:\n print(\"Selected expression: \", list(node.children))\n\n formatstring = capture['formatstring']\n interpolation_args = capture['interpolation_args']\n\n # We only convert .format() stuff to an f-string if the arguments are all simple-ish.\n # Currently that means we ignore dotted names, dict['things'] and literals.\n positional_args = []\n keyword_args = {\n # Maps kwarg names (strings) to the kwarg *value*\n # e.g. for .format(a=b) this would be {'a': 'b'}\n }\n try:\n for parsed_arg in _interpret_format_arguments(interpolation_args):\n if parsed_arg is None:\n # This arg was deemed too complex to bother pushing into an f-string.\n # Give up.\n return node\n elif isinstance(parsed_arg, dict):\n keyword_args.update(parsed_arg)\n else:\n positional_args.append(parsed_arg)\n\n # Actually push the new names into a new formatstring. Wrap each value with curly braces.\n replacement_value = formatstring.value.format(\n *['{%s}' % a for a in positional_args],\n **{k: '{%s}' % v for (k, v) in keyword_args.items()}\n )\n\n # Convert to an f-string\n replacement_value = add_f_prefix(replacement_value)\n\n if flags['debug']:\n print(f\"Interpolating (new-style) format-string:\\n\\t{formatstring}\")\n print(f\"With arguments:\\n\\t{positional_args}, {keyword_args}\")\n print(f\"Replacement formatstring: {replacement_value}\")\n print()\n\n except SkipString:\n return node\n\n # Finally, apply the whole thing\n formatstring.value = replacement_value\n capture['trailer1'].remove()\n capture['trailer2'].remove()\n return node\n\n\ndef main():\n parser = argparse.ArgumentParser(\n description=\"Converts string interpolation expressions to use f-strings where possible.\"\n )\n parser.add_argument(\n '--no-input',\n dest='interactive',\n default=True,\n action='store_false',\n help=\"Non-interactive mode\"\n )\n parser.add_argument(\n '--no-write',\n dest='write',\n default=True,\n action='store_false',\n help=\"Don't write the changes to the source file, just output a diff to stdout\"\n )\n parser.add_argument(\n '--debug',\n dest='debug',\n default=False,\n action='store_true',\n help=\"Spit out debugging information\"\n )\n parser.add_argument(\n 'files',\n nargs='+',\n help=\"The python source file(s) to operate on.\"\n )\n args = parser.parse_args()\n\n # No way to pass this to .modify() callables, so we just set it at module level\n flags['debug'] = args.debug\n\n query = (\n # Look for files in the current working directory\n Query(*args.files)\n\n # NOTE: You can append as many .select().modify() bits as you want to one query.\n # Each .modify() acts only on the .select[_*]() immediately prior.\n\n # 1. String interpolation (old style):\n # ... where the thing on the right is a variable name\n # ... where the thing on the right is a tuple of variable names.\n .select('''\n (\n term<\n formatstring=STRING '%' interpolation_args=NAME >\n |\n term< formatstring=STRING '%' atom< '('\n (testlist_gexp< interpolation_args=((NAME ',')* NAME [',']) >)\n ')' > >\n )\n ''')\n .modify(callback=old_interpolation_to_fstrings)\n\n # 2. New-style interpolation (.format(...))\n # The 'power<>' thing is confusing to me. What's 'power' mean in this context?\n # NOTE: this selector is quite loose; it accepts 'any*' in the arguments to format().\n # i.e. this happily accepts: ''.format(a, 2, b=3, c=d[e], *x, **y)\n # We'll need to be careful handling each of these in the modify callback,\n # since not all of those args make much sense shoved into an fstring.\n .select('''\n function_call=power<\n formatstring=STRING\n trailer1=trailer<\n '.' 'format'\n >\n trailer2=trailer< '(' interpolation_args=any* ')' >\n any*\n >\n ''')\n .modify(callback=format_method_to_fstrings)\n\n # Actually run both of the above.\n .execute(\n # interactive diff implies write (for the bits the user says 'y' to)\n interactive=(args.interactive and args.write),\n write=args.write,\n )\n )\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"craigds/decrapify","sub_path":"fstrings.py","file_name":"fstrings.py","file_ext":"py","file_size_in_byte":11005,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"7"} +{"seq_id":"26946242927","text":"\"\"\"\nMain module where all the neural network magic happens\n\"\"\"\n\nfrom __future__ import print_function\n#from __future__ import absolute_import\nimport os\nimport sys\n#sys.path.insert(0,os.path.join(os.path.dirname(os.path.realpath(__file__)), \"..\"))\nimport numpy as np\nimport tensorflow as tf\n#from sklearn.base import BaseEstimator, RegressorMixin\nfrom sklearn.utils.validation import check_X_y, check_array\nfrom sklearn.metrics import r2_score, mean_squared_error, mean_absolute_error\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\n#import inverse_dist as inv\n#from tensorflow.python.framework import ops\n#from tensorflow.python.training import saver as saver_lib\n#from tensorflow.python.framework import graph_io\n#from tensorflow.python.tools import freeze_graph\n\n#from .utils import is_positive, is_positive_integer, is_positive_integer_or_zero, \\\n# is_bool, is_string, is_positive_or_zero, InputError, ceil\nfrom .utils import InputError, ceil, is_positive_or_zero, is_positive_integer, is_positive, \\\n is_bool, is_positive_integer_or_zero, is_string, is_positive_integer_array\nfrom .tf_utils import TensorBoardLogger\n\nclass _NN(object):\n\n \"\"\"\n Parent class for training multi-layered neural networks on molecular or atomic properties via Tensorflow\n \"\"\"\n\n def __init__(self, hidden_layer_sizes = [5], l1_reg = 0.0, l2_reg = 0.0001, batch_size = 'auto', learning_rate = 0.001,\n iterations = 500, tensorboard = False, store_frequency = 200, tf_dtype = tf.float32, scoring_function = 'mae',\n activation_function = tf.sigmoid, tensorboard_subdir = os.getcwd() + '/tensorboard', **kwargs):\n \"\"\"\n :param hidden_layer_sizes: Number of hidden layers. The n'th element represents the number of neurons in the n'th\n hidden layer.\n :type hidden_layer_size: Tuple of integers\n :param l1_reg: L1-regularisation parameter for the neural network weights\n :type l1_reg: float\n :param l2_reg: L2-regularisation parameter for the neural network weights\n :type l2_reg: float\n :param batch_size: Size of minibatches for the ADAM optimizer. If set to 'auto' ``batch_size = min(200,n_samples)``\n :type batch_size: integer\n :param learning_rate: The learning rate in the numerical minimisation.\n :type learning_rate: float\n :param iterations: Total number of iterations that will be carried out during the training process.\n :type iterations: integer\n :param tf_dtype: Accuracy to use for floating point operations in tensorflow. 64 and 'float64' is recognised as tf.float64\n and similar for tf.float32 and tf.float16.\n :type tf_dtype: Tensorflow datatype\n :param scoring_function: Scoring function to use. Available choices are 'mae', 'rmse', 'r2'.\n :type scoring_function: string\n :param activation_function: Activation function to use in the neural network. Currently 'sigmoid', 'tanh', 'elu', 'softplus',\n 'softsign', 'relu', 'relu6', 'crelu' and 'relu_x' is supported.\n :type activation_function: Tensorflow datatype\n :param tensorboard: Store summaries to tensorboard or not\n :type tensorboard: boolean\n :param store_frequency: How often to store summaries to tensorboard.\n :type store_frequency: integer\n :param tensorboard_subdir: Directory to store tensorboard data\n :type tensorboard_subdir: string\n \"\"\"\n\n super(_NN,self).__init__()\n\n # Catch unrecognised passed variables\n if len(kwargs) > 0:\n msg = \"Warning: unrecognised input variable(s): \"\n msg += \", \".join([str(x for x in kwargs.keys())])\n print(msg)\n\n\n # Initialising the parameters\n self._set_hidden_layers_sizes(hidden_layer_sizes)\n self._set_l1_reg(l1_reg)\n self._set_l2_reg(l2_reg)\n self._set_batch_size(batch_size)\n self._set_learning_rate(learning_rate)\n self._set_iterations(iterations)\n self._set_tf_dtype(tf_dtype)\n self._set_scoring_function(scoring_function)\n self._set_tensorboard(tensorboard, store_frequency, tensorboard_subdir)\n self._set_activation_function(activation_function)\n\n\n # Placeholder variables\n self.n_features = None\n self.n_samples = None\n self.training_cost = []\n self.session = None\n #self.test_cost = []\n #self.loaded_model = False\n #self.is_vis_ready = False\n\n def _set_activation_function(self, activation_function):\n if activation_function in ['sigmoid', tf.nn.sigmoid]:\n self.activation_function = tf.nn.sigmoid\n elif activation_function in ['tanh', tf.nn.tanh]:\n self.activation_function = tf.nn.tanh\n elif activation_function in ['elu', tf.nn.elu]:\n self.activation_function = tf.nn.elu\n elif activation_function in ['softplus', tf.nn.softplus]:\n self.activation_function = tf.nn.softplus\n elif activation_function in ['softsign', tf.nn.softsign]:\n self.activation_function = tf.nn.softsign\n elif activation_function in ['relu', tf.nn.relu]:\n self.activation_function = tf.nn.relu\n elif activation_function in ['relu6', tf.nn.relu6]:\n self.activation_function = tf.nn.relu6\n elif activation_function in ['crelu', tf.nn.crelu]:\n self.activation_function = tf.nn.crelu\n elif activation_function in ['relu_x', tf.nn.relu_x]:\n self.activation_function = tf.nn.relu_x\n else:\n raise InputError(\"Unknown activation function. Got %s\" % str(activation_function))\n\n def _set_l1_reg(self, l1_reg):\n if not is_positive_or_zero(l1_reg):\n raise InputError(\"Expected positive float value for variable 'l1_reg'. Got %s\" % str(l1_reg))\n self.l1_reg = l1_reg\n\n def _set_l2_reg(self, l2_reg):\n if not is_positive_or_zero(l2_reg):\n raise InputError(\"Expected positive float value for variable 'l2_reg'. Got %s\" % str(l2_reg))\n self.l2_reg = l2_reg\n\n def _set_batch_size(self, batch_size):\n if batch_size != \"auto\":\n if not is_positive_integer(batch_size):\n raise InputError(\"Expected 'batch_size' to be a positive integer. Got %s\" % str(batch_size))\n elif batch_size == 1:\n raise InputError(\"batch_size must be larger than 1. Got %s\" % str(batch_size))\n self.batch_size = int(batch_size)\n else:\n self.batch_size = batch_size\n\n def _set_learning_rate(self, learning_rate):\n if not is_positive(learning_rate):\n raise InputError(\"Expected positive float value for variable learning_rate. Got %s\" % str(learning_rate))\n self.learning_rate = float(learning_rate)\n\n def _set_iterations(self, iterations):\n if not is_positive_integer(iterations):\n raise InputError(\"Expected positive integer value for variable iterations. Got %s\" % str(iterations))\n self.iterations = int(iterations)\n\n def _set_tf_dtype(self, tf_dtype):\n # 2 == tf.float64 and 1 == tf.float32 for some reason\n # np.float64 recognised as tf.float64 as well\n if tf_dtype in ['64', 64, 'float64', tf.float64]:\n self.tf_dtype = tf.float64\n elif tf_dtype in ['32', 32, 'float32', tf.float32]:\n self.tf_dtype = tf.float32\n elif tf_dtype in ['16', 16, 'float16', tf.float16]:\n self.tf_dtype = tf.float16\n else:\n raise InputError(\"Unknown tensorflow data type. Got %s\" % str(tf_dtype))\n\n def _set_optimiser_param(self, beta1, beta2, epsilon, rho, initial_accumulator_value, initial_gradient_squared_accumulator_value,\n l1_regularization_strength, l2_regularization_strength):\n \"\"\"\n This function sets all the parameters that are required by all the optimiser functions. In the end, only the parameters\n for the optimiser chosen will be used.\n\n :param beta1:\n :param beta2:\n :param epsilon:\n :param rho:\n :param initial_accumulator_value:\n :param initial_gradient_squared_accumulator_value:\n :param l1_regularization_strength:\n :param l2_regularization_strength:\n :return: None\n \"\"\"\n if not is_positive(beta1) and not is_positive(beta2):\n raise InputError(\"Expected positive float values for variable beta1 and beta2. Got %s and %s.\" % (str(beta1),str(beta2)))\n self.beta1 = float(beta1)\n self.beta2 = float(beta2)\n\n if not is_positive(epsilon):\n raise InputError(\"Expected positive float value for variable epsilon. Got %s\" % str(epsilon))\n self.epsilon = float(epsilon)\n\n if not is_positive(rho):\n raise InputError(\"Expected positive float value for variable rho. Got %s\" % str(rho))\n self.epsilon = float(rho)\n\n if not is_positive(initial_accumulator_value) and not is_positive(initial_gradient_squared_accumulator_value):\n raise InputError(\"Expected positive float value for accumulator values. Got %s and %s\" %\n (str(initial_accumulator_value), str(initial_gradient_squared_accumulator_value)))\n self.initial_accumulator_value = float(initial_accumulator_value)\n self.initial_gradient_squared_accumulator_value = float(initial_gradient_squared_accumulator_value)\n\n if not is_positive_or_zero(l1_regularization_strength) and not is_positive_or_zero(l2_regularization_strength):\n raise InputError(\"Expected positive or zero float value for regularisation variables. Got %s and %s\" %\n (str(l1_regularization_strength), str(l2_regularization_strength)))\n self.l1_regularization_strength = float(l1_regularization_strength)\n self.l2_regularization_strength = float(l2_regularization_strength)\n\n def _set_scoring_function(self, scoring_function):\n if not is_string(scoring_function):\n raise InputError(\"Expected a string for variable 'scoring_function'. Got %s\" % str(scoring_function))\n if scoring_function.lower() not in ['mae', 'rmse', 'r2']:\n raise InputError(\"Available scoring functions are 'mae', 'rmse', 'r2'. Got %s\" % str(scoring_function))\n\n self.scoring_function = scoring_function\n\n def _set_hidden_layers_sizes(self, hidden_layer_sizes):\n if not is_positive_integer_array(hidden_layer_sizes):\n raise InputError(\"'hidden_layer_sizes' must be an array of positive integers\")\n\n self.hidden_layer_sizes = np.asarray(hidden_layer_sizes, dtype = int)\n\n def _set_tensorboard(self, tensorboard, store_frequency, tensorboard_subdir):\n\n if not is_bool(tensorboard):\n raise InputError(\"Expected boolean value for variable tensorboard. Got %s\" % str(tensorboard))\n self.tensorboard = bool(tensorboard)\n\n if not self.tensorboard:\n return\n\n if not is_string(tensorboard_subdir):\n raise InputError('Expected string value for variable tensorboard_subdir. Got %s' % str(self.tensorboard_subdir))\n\n # TensorBoardLogger will handle all tensorboard related things\n self.tensorboard_logger = TensorBoardLogger(tensorboard_subdir)\n\n if not is_positive_integer(store_frequency):\n raise InputError(\"Expected positive integer value for variable store_frequency. Got %s\" % str(store_frequency))\n\n if store_frequency > self.iterations:\n print(\"Only storing final iteration for tensorboard\")\n self.tensorboard_logger.set_store_frequency(self.iterations)\n else:\n self.tensorboard_logger.set_store_frequency(store_frequency)\n\n\n def _init_weight(self, n1, n2, name):\n \"\"\"\n Generate a tensor of weights of size (n1, n2)\n\n \"\"\"\n\n w = tf.Variable(tf.truncated_normal([n1,n2], stddev = 1.0 / np.sqrt(n2), dtype = self.tf_dtype),\n dtype = self.tf_dtype, name = name)\n\n return w\n\n def _init_bias(self, n, name):\n \"\"\"\n Generate a tensor of biases of size n.\n\n \"\"\"\n\n b = tf.Variable(tf.zeros([n], dtype = self.tf_dtype), name=name, dtype = self.tf_dtype)\n\n return b\n\n def _generate_weights(self, n_out):\n \"\"\"\n Generates the weights and the biases, by looking at the size of the hidden layers,\n the number of features in the descriptor and the number of outputs. The weights are initialised from\n a zero centered normal distribution with precision :math:`\\\\tau = a_{m}`, where :math:`a_{m}` is the number\n of incoming connections to a neuron. Weights larger than two standard deviations from the mean is\n redrawn.\n\n :param n_out: Number of outputs\n :type n_out: integer\n :return: tuple of weights and biases, each being of length (n_hidden_layers + 1)\n :rtype: tuple\n \"\"\"\n\n weights = []\n biases = []\n\n # Weights from input layer to first hidden layer\n weights.append(self._init_weight(self.hidden_layer_sizes[0], self.n_features, 'weight_in'))\n biases.append(self._init_bias(self.hidden_layer_sizes[0], 'bias_in'))\n\n # Weights from one hidden layer to the next\n for i in range(1, self.hidden_layer_sizes.size):\n weights.append(self._init_weight(self.hidden_layer_sizes[i], self.hidden_layer_sizes[i-1], 'weight_hidden_%d' %i))\n biases.append(self._init_bias(self.hidden_layer_sizes[i], 'bias_hidden_%d' % i))\n\n # Weights from last hidden layer to output layer\n weights.append(self._init_weight(n_out, self.hidden_layer_sizes[-1], 'weight_out'))\n biases.append(self._init_bias(n_out, 'bias_out'))\n\n return weights, biases\n\n def _l2_loss(self, weights):\n \"\"\"\n Creates the expression for L2-regularisation on the weights\n\n :param weights: tensorflow tensors representing the weights\n :type weights: list\n :return: tensorflow scalar representing the regularisation contribution to the cost function\n :rtype: tf.float32\n \"\"\"\n\n reg_term = tf.zeros([], name=\"l2_loss\")\n\n for i in range(self.hidden_layer_sizes.size):\n reg_term += tf.reduce_sum(tf.square(weights[i]))\n\n return self.l2_reg * reg_term\n\n def _l1_loss(self, weights):\n \"\"\"\n Creates the expression for L1-regularisation on the weights\n\n :param weights: tensorflow tensors representing the weights\n :type weights: list\n :return: tensorflow scalar representing the regularisation contribution to the cost function\n :rtype: tf.float32\n \"\"\"\n\n reg_term = tf.zeros([], name=\"l1_loss\")\n\n for i in range(self.hidden_layer_sizes.size):\n reg_term += tf.reduce_sum(tf.abs(weights[i]))\n\n return self.l1_reg * reg_term\n\n def model(self, x, weights, biases):\n \"\"\"\n Constructs the actual network.\n\n :param x: Input\n :type x: tf.placeholder of shape (None, n_features)\n :param weights: Weights used in the network.\n :type weights: list of tf.Variables of length hidden_layer_sizes.size + 1\n :param biases: Biases used in the network.\n :type biases: list of tf.Variables of length hidden_layer_sizes.size + 1\n :return: Output\n :rtype: tf.Variable of size (None, n_targets)\n \"\"\"\n\n # Calculate the activation of the first hidden layer\n z = tf.add(tf.matmul(x, tf.transpose(weights[0])), biases[0])\n h = self.activation_function(z)\n\n # Calculate the activation of the remaining hidden layers\n for i in range(self.hidden_layer_sizes.size-1):\n z = tf.add(tf.matmul(h, tf.transpose(weights[i+1])), biases[i+1])\n h = self.activation_function(z)\n\n # Calculating the output of the last layer\n z = tf.add(tf.matmul(h, tf.transpose(weights[-1])), biases[-1], name = \"output\")\n\n return z\n\n def _get_batch_size(self):\n \"\"\"\n Determines the actual batch size. If set to auto, the batch size will be set to 100.\n If the batch size is larger than the number of samples, it is truncated and a warning\n is printed.\n\n Furthermore the returned batch size will be slightly modified from the user input if\n the last batch would be tiny compared to the rest.\n\n :return: Batch size\n :rtype: integer\n \"\"\"\n\n if self.batch_size == 'auto':\n batch_size = min(100, self.n_samples)\n else:\n if self.batch_size > self.n_samples:\n print(\"Warning: batch_size larger than sample size. It is going to be clipped\")\n return min(self.n_samples, self.batch_size)\n else:\n batch_size = self.batch_size\n\n # see if the batch size can be modified slightly to make sure the last batch is similar in size\n # to the rest of the batches\n # This is always less that the requested batch size, so no memory issues should arise\n better_batch_size = ceil(self.n_samples, ceil(self.n_samples, batch_size))\n\n return better_batch_size\n\n def plot_cost(self, filename = None):\n \"\"\"\n Plots the value of the cost function as a function of the iterations.\n\n :param filename: File to save the plot to. If '' the plot is shown instead of saved.\n :type filename: string\n \"\"\"\n\n try:\n import pandas as pd\n import seaborn as sns\n except ModuleNotFoundError:\n raise ModuleNotFoundError(\"Plotting functions require the modules 'seaborn' and 'pandas'\")\n\n sns.set()\n df = pd.DataFrame()\n df[\"Iterations\"] = range(len(self.training_cost))\n df[\"Training cost\"] = self.training_cost\n f = sns.lmplot('Iterations', 'Training cost', data=df, scatter_kws={\"s\": 20, \"alpha\": 0.6}, line_kws={\"alpha\": 0.5}, fit_reg=False)\n f.set(yscale = \"log\")\n\n if filename == None:\n plt.show()\n elif is_string(filename):\n plt.savefig(filename)\n else:\n raise InputError(\"Wrong data type of variable 'filename'. Expected string\")\n\n def correlation_plot(self, y_nn, y_true, filename = ''):\n \"\"\"\n Creates a correlation plot between predictions and true values.\n\n :param y_predicted: Values predicted by the neural net\n :type y_predicted: list\n :param y_true: True values\n :type y_true: list\n :param filename: File to save the plot to. If '' the plot is shown instead of saved.\n If the dimensionality of y is higher than 1, the filename will be prefixed\n by the dimension.\n :type filename: string\n \"\"\"\n\n try:\n import pandas as pd\n import seaborn as sns\n except ModuleNotFoundError:\n raise ModuleNotFoundError(\"Plotting functions require the modules 'seaborn' and 'pandas'\")\n\n if y_nn.shape != y_true.shape:\n raise InputError(\"Shape mismatch between predicted and true values. %s and %s\" % (str(y_nn.shape), str(y_true.shape)))\n\n if y_nn.ndim == 1 or y_nn.shape[1] == 1:\n df = pd.DataFrame()\n df[\"Predictions\"] = y_nn.ravel()\n df[\"True\"] = y_true.ravel()\n sns.set()\n lm = sns.lmplot('True', 'Predictions', data=df, scatter_kws={\"s\": 20, \"alpha\": 0.6}, line_kws={\"alpha\": 0.5})\n if filename == '':\n plt.show()\n elif is_string(filename):\n plt.savefig(filename)\n else:\n raise InputError(\"Wrong data type of variable 'filename'. Expected string\")\n else:\n for i in range(y_nn.shape[0]):\n df = pd.DataFrame()\n df[\"Predictions\"] = y_nn[:,i]\n df[\"True\"] = y_true[:,i]\n sns.set()\n lm = sns.lmplot('True', 'Predictions', data=df, scatter_kws={\"s\": 20, \"alpha\": 0.6}, line_kws={\"alpha\": 0.5})\n if filename == '':\n plt.show()\n elif is_string(filename):\n tokens = filename.split(\"/\")\n file_ = str(i) + \"_\" + tokens[-1]\n if len(tokens) > 1:\n file_ = \"/\".join(tokens[:-1]) + \"/\" + file_\n plt.savefig(file_)\n else:\n raise InputError(\"Wrong data type of variable 'filename'. Expected string\")\n\n def score(self, *args):\n return self._score(*args)\n\n # TODO test\n def _score(self, *args):\n if self.scoring_function == 'mae':\n return self._score_mae(*args)\n if self.scoring_function == 'rmse':\n return self._score_rmse(*args)\n if self.scoring_function == 'r2':\n return self._score_r2(*args)\n\n def predict(self, x):\n predictions = self._predict(x)\n\n if predictions.ndim > 1 and predictions.shape[1] == 1:\n return predictions.ravel()\n else:\n return predictions\n\n # TODO test\n def _predict(self, x):\n \"\"\"\n Use the trained network to make predictions on the data x.\n\n :param x: The input data of shape (n_samples, n_features)\n :type x: array\n\n :return: Predictions for the target values corresponding to the samples contained in x.\n :rtype: array\n\n \"\"\"\n\n if self.session == None:\n raise InputError(\"Model needs to be fit before predictions can be made.\")\n\n check_array(x, warn_on_dtype = True)\n\n graph = tf.get_default_graph()\n\n with graph.as_default():\n tf_x = graph.get_tensor_by_name(\"Data/Descriptors:0\")\n model = graph.get_tensor_by_name(\"Model/output:0\")\n y_pred = self.session.run(model, feed_dict = {tf_x : x})\n return y_pred\n\n# TODO: Rename to something more sensible\nclass NN(_NN):\n \"\"\"\n Neural network for either\n 1) predicting global properties, such as energies, using molecular representations, or\n 2) predicting local properties, such as chemical shieldings, using atomic representations.\n \"\"\"\n\n def __init__(self, **kwargs):\n \"\"\"\n Descriptors is used as input to a single or multi layered feed-forward neural network with a single output.\n This class inherits from the _NN class and all inputs not unique to the NN class is passed to the _NN\n parent.\n\n \"\"\"\n\n super(NN,self).__init__(**kwargs)\n\n #TODO test\n def fit(self, x, y):\n \"\"\"\n Fit the neural network to molecular descriptors x and target y.\n\n :param x: Input data with samples in the rows and features in the columns.\n :type x: array\n :param y: Target values for each sample.\n :type y: array\n\n \"\"\"\n\n return self._fit(x, y)\n\n def _fit(self, x, y):\n\n # Check that X and y have correct shape\n x, y = check_X_y(x, y, multi_output = False, y_numeric = True, warn_on_dtype = True)\n\n # reshape to tensorflow friendly shape\n y = np.atleast_2d(y).T\n\n # Useful quantities\n self.n_features = x.shape[1]\n self.n_samples = x.shape[0]\n #self.n_atoms = int(X.shape[1]/3)\n\n # Set the batch size\n batch_size = self._get_batch_size()\n\n # Initial set up of the NN\n with tf.name_scope(\"Data\"):\n tf_x = tf.placeholder(self.tf_dtype, [None, self.n_features], name=\"Descriptors\")\n tf_y = tf.placeholder(self.tf_dtype, [None, 1], name=\"Properties\")\n\n # Either initialise the weights and biases or restart training from wherever it was stopped\n with tf.name_scope(\"Weights\"):\n weights, biases = self._generate_weights(n_out = 1)\n\n # Log weights for tensorboard\n if self.tensorboard:\n self.tensorboard_logger.write_weight_histogram(weights)\n\n with tf.name_scope(\"Model\"):\n y_pred = self.model(tf_x, weights, biases)\n\n with tf.name_scope(\"Cost_func\"):\n cost = self.cost(y_pred, tf_y, weights)\n\n if self.tensorboard:\n cost_summary = tf.summary.scalar('cost', cost)\n\n optimizer = tf.train.AdamOptimizer(learning_rate=self.learning_rate).minimize(cost)\n\n # Initialisation of the variables\n init = tf.global_variables_initializer()\n #self.initialised = True\n\n if self.tensorboard:\n self.tensorboard_logger.initialise()\n\n # This is the total number of batches in which the training set is divided\n n_batches = ceil(self.n_samples, batch_size)\n\n self.session = tf.Session()\n\n # Running the graph\n if self.tensorboard:\n self.tensorboard_logger.set_summary_writer(self.session)\n\n self.session.run(init)\n\n indices = np.arange(0,self.n_samples, 1)\n\n for i in range(self.iterations):\n # This will be used to calculate the average cost per iteration\n avg_cost = 0\n # Learning over the batches of data\n for j in range(n_batches):\n batch_x = x[indices][j * batch_size:(j+1) * batch_size]\n batch_y = y[indices][j * batch_size:(j+1) * batch_size]\n feed_dict = {tf_x: batch_x, tf_y: batch_y}\n opt, c = self.session.run([optimizer, cost], feed_dict=feed_dict)\n avg_cost += c * batch_x.shape[0] / x.shape[0]\n\n if self.tensorboard:\n if i % self.tensorboard_logger.store_frequency == 0:\n self.tensorboard_logger.write_summary(self.session, feed_dict, i, j)\n\n self.training_cost.append(avg_cost)\n\n # Shuffle the dataset at each iteration\n np.random.shuffle(indices)\n\n # TODO test\n def _score_r2(self, x, y, sample_weight=None):\n \"\"\"\n Calculate the coefficient of determination (R^2).\n Larger values corresponds to a better prediction.\n\n :param x: The input data.\n :type x: array of shape (n_samples, n_features)\n :param y: The target values for each sample in x.\n :type y: array of shape (n_samples,)\n\n :param sample_weight: Weights of the samples. None indicates that that each sample has the same weight.\n :type sample_weight: array of shape (n_samples,)\n\n :return: R^2\n :rtype: float\n\n \"\"\"\n\n y_pred = self.predict(x)\n r2 = r2_score(y, y_pred, sample_weight = sample_weight)\n return r2\n\n # TODO test\n def _score_mae(self, x, y, sample_weight=None):\n \"\"\"\n Calculate the mean absolute error.\n Smaller values corresponds to a better prediction.\n\n :param x: The input data.\n :type x: array of shape (n_samples, n_features)\n :param y: The target values for each sample in x.\n :type y: array of shape (n_samples,)\n\n :param sample_weight: Weights of the samples. None indicates that that each sample has the same weight.\n :type sample_weight: array of shape (n_samples,)\n\n :return: Mean absolute error\n :rtype: float\n\n \"\"\"\n\n y_pred = self.predict(x)\n mae = mean_absolute_error(y, y_pred, sample_weight = sample_weight)\n return mae\n\n # TODO test\n def _score_rmse(self, x, y, sample_weight=None):\n \"\"\"\n Calculate the root mean squared error.\n Smaller values corresponds to a better prediction.\n\n :param x: The input data.\n :type x: array of shape (n_samples, n_features)\n :param y: The target values for each sample in x.\n :type y: array of shape (n_samples,)\n\n :param sample_weight: Weights of the samples. None indicates that that each sample has the same weight.\n :type sample_weight: array of shape (n_samples,)\n\n :return: Mean absolute error\n :rtype: float\n\n \"\"\"\n\n y_pred = self.predict(x)\n rmse = np.sqrt(mean_squared_error(y, y_pred, sample_weight = sample_weight))\n return rmse\n\n def cost(self, y_pred, y, weights):\n \"\"\"\n Constructs the cost function\n\n :param y_pred: Predicted output\n :type y_pred: tf.Variable of size (None, 1)\n :param y: True output\n :type y: tf.placeholder of shape (None, 1)\n :param weights: Weights used in the network.\n :type weights: list of tf.Variables of length hidden_layer_sizes.size + 1\n :return: Cost\n :rtype: tf.Variable of size (1,)\n \"\"\"\n\n err = tf.square(tf.subtract(y,y_pred))\n loss = tf.reduce_mean(err, name=\"loss\")\n cost = loss\n if self.l2_reg > 0:\n l2_loss = self._l2_loss(weights)\n cost = cost + l2_loss\n if self.l1_reg > 0:\n l1_loss = self._l1_loss(weights)\n cost = cost + l1_loss\n\n return cost\n\n","repo_name":"SilviaAmAm/Aglaia","sub_path":"aglaia/aglaia.py","file_name":"aglaia.py","file_ext":"py","file_size_in_byte":29049,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"38098826904","text":"import os\nimport re\nimport time\n\nimport numpy as np\n\nfrom features import AutoEncoderFE, ColorLayoutFE, FeatureExtractor\nfrom keyframes import KeyframeSelector, MaxHistDiffKS, FPSReductionKS\nfrom utils.files import get_features_dir, get_videos_dir, group_features, log_persistent\n\n\ndef extract_features_videos(\n selector: KeyframeSelector,\n extractor: FeatureExtractor,\n force=False\n):\n \"\"\"\n Extracts features for the all the videos in the directory and saves them in a new directory obtained using\n get_features_path.\n\n :param selector: .\n :param extractor: .\n :param force: when True, calculates features even if it was done previously.\n \"\"\"\n\n # create directory when necessary\n feats_path = get_features_dir(selector=selector, extractor=extractor)\n if not os.path.isdir(feats_path):\n os.makedirs(feats_path)\n\n # create or empty log files\n if not os.path.isfile(f'{feats_path}/extraction_log.txt') or force:\n open(f'{feats_path}/extraction_log.txt', 'w').close()\n open(f'{feats_path}/selection_log.txt', 'w').close()\n\n # obtain all files in the directory\n videos_path = get_videos_dir()\n videos = os.listdir(videos_path)\n\n # extract features from each video\n for video in videos:\n if video.endswith('.mp4'):\n extract_features(\n file_path=f'{videos_path}/{video}', save_dir=feats_path,\n selector=selector, extractor=extractor, force=force)\n\n # group all features into 2 files (features and tags)\n group_features(\n selector=selector,\n extractor=extractor,\n force=force\n )\n return\n\n\ndef extract_features(\n file_path: str,\n save_dir: str,\n selector: KeyframeSelector,\n extractor: FeatureExtractor,\n force=False\n):\n \"\"\"\n Extracts features for the video and saves them in the given dir.\n\n :param file_path: video path.\n :param save_dir: directory to save the features.\n :param selector: .\n :param extractor: .\n :param force: when True, calculates features even if it was done previously.\n \"\"\"\n\n video_name = re.split('[/.]', file_path)[-2]\n save_path_feats = f'{save_dir}/{video_name}-feats.npy'\n save_path_tags = f'{save_dir}/{video_name}-tags.npy'\n\n # skip already processed videos\n if not force and os.path.isfile(save_path_feats) and os.path.isfile(save_path_tags):\n print(f'Skipping video {video_name}')\n return\n\n print(f'Extracting features from video {video_name}')\n\n # obtain keyframes\n t0 = time.time()\n keyframes, timestamps, total_frames = selector.select_keyframes(file_path)\n\n selection = time.time() - t0\n print(f'selected {len(keyframes)} of {total_frames} frames in {selection:.1f} secs')\n\n # log selection time\n log_persistent(f'{len(timestamps)}\\t{selection:.2f}\\n', f'{save_dir}/selection_log.txt')\n\n # measure time\n t0 = time.time()\n\n # extract features and save\n features = extractor.extract_features(keyframes)\n np.save(save_path_feats, features)\n\n # generate tags and save\n tags = np.empty(timestamps.shape[0], dtype=' ^img_\\d+.jpg$\n # Flow patteren example: {}_{:05d}.jpg -> ^x_\\d+.jpg$\n # 獲取檔案名稱的模板\n pattern = f'^{filename_tmpl}$'\n if modality == 'Flow':\n # 如果是光流資料就會到這裡\n pattern = pattern.replace('{}', 'x')\n pattern = pattern.replace(\n pattern[pattern.find('{'):pattern.find('}') + 1], '\\\\d+')\n total_frames = len(\n list(\n filter(lambda x: re.match(pattern, x) is not None,\n os.listdir(video))))\n data = dict(\n frame_dir=video,\n total_frames=total_frames,\n label=-1,\n start_index=start_index,\n filename_tmpl=filename_tmpl,\n modality=modality)\n if 'Init' in test_pipeline[0]['type']:\n test_pipeline = test_pipeline[1:]\n for i in range(len(test_pipeline)):\n if 'Decode' in test_pipeline[i]['type']:\n test_pipeline[i] = dict(type='RawFrameDecode')\n if input_flag == 'audio':\n # 如果input_flag是audio就會到這裡\n data = dict(\n audio_path=video,\n total_frames=len(np.load(video)),\n start_index=cfg.data.test.get('start_index', 1),\n label=-1)\n\n # 構建影片處理流水線\n test_pipeline = Compose(test_pipeline)\n # 將data放入到流水線當中處理\n data = test_pipeline(data)\n # 透過collate將資料整理成一個batch\n data = collate([data], samples_per_gpu=1)\n\n if next(model.parameters()).is_cuda:\n # 如果有使用gpu就會到這裡\n # scatter to specified GPU\n data = scatter(data, [device])[0]\n\n # forward the model\n with OutputHook(model, outputs=outputs, as_tensor=as_tensor) as h:\n # 在OutputHook下\n with torch.no_grad():\n # 將模型的反向傳遞關閉,進行正向傳遞\n scores = model(return_loss=False, **data)[0]\n # 如果有需要獲取層結構輸出就會到這裡獲取\n returned_features = h.layer_outputs if outputs else None\n\n # 獲取總共是多少分類\n num_classes = scores.shape[-1]\n # 將一個類別對應上一個置信度分數,score_tuples = tuple(tuple(對應類別的index, 置信度分數)),第一個tuple長度就會是num_classes\n score_tuples = tuple(zip(range(num_classes), scores))\n # 將score_tuples進行排序,這裡會從置信度大排到小\n score_sorted = sorted(score_tuples, key=itemgetter(1), reverse=True)\n\n # 獲取前5大概率值的置信度以及分類類別\n top5_label = score_sorted[:5]\n if outputs:\n # 如果有需要中途的輸出就會到這裡\n return top5_label, returned_features\n # 其他就會輸出前5大預測結果\n return top5_label\n","repo_name":"chris901003/DeepLearning","sub_path":"mmaction2/mmaction/apis/inference.py","file_name":"inference.py","file_ext":"py","file_size_in_byte":12408,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"7"} +{"seq_id":"13538416749","text":"import vk\nimport config\nimport del_name\n\nsession = vk.Session(config.tokenVK)\napi = vk.API(session, v='5.85')\nid_name = \"0\"\n\ndef mama(s):\n result = s.replace(\"маме \", \"\")\n return result\n\ndef papa(s):\n result = s.replace(\"папе \", \"\")\n return result\n\ndef inna(s):\n result = s.replace(\"инне \", \"\")\n return result\n\ndef roma(s):\n result = s.replace(\"роме \", \"\")\n return result\n\ndef seva(s):\n result = s.replace(\"севе \", \"\")\n return result\n\ndef mne(s):\n result = s.replace(\"мне \", \"\")\n return result\n\ndef mama2(s):\n result = s.replace(\"Маме \", \"\")\n return result\n\ndef papa2(s):\n result = s.replace(\"Папе \", \"\")\n return result\n\ndef inna2(s):\n result = s.replace(\"Инне \", \"\")\n return result\n\ndef roma2(s):\n result = s.replace(\"Роме \", \"\")\n return result\n\ndef seva2(s):\n result = s.replace(\"Севе \", \"\")\n return result\n\ndef mne2(s):\n result = s.replace(\"Мне \", \"\")\n return result\n\ndef remove_char(s):\n result = s.replace(\"Новый пост \", \"\")\n return result\n\ndef remove_char1(s):\n result = s.replace(\"новый пост \", \"\")\n return result\n\ndef del_name(s):\n a = mama(s)\n b = papa(a)\n c = inna(b)\n d = roma(c)\n e = seva(d)\n f = mne(e)\n g = mama2(f)\n h = papa2(g)\n i = inna2(h)\n j = roma2(i)\n k = seva2(j)\n result = mne2(k)\n return result\n\ndef name_to_id(s):\n global id_name\n if s == \"Маме\" or s == \"маме\":\n id_name = \"516717987\"\n elif s == \"Папе\" or s == \"папе\":\n id_name = \"3883837\"\n elif s == \"Инне\" or s == \"инне\":\n id_name = \"25070022\"\n elif s == \"Роме\" or s == \"роме\":\n id_name = \"397205762\"\n elif s == \"Севе\" or s == \"севе\":\n id_name = \"507564168\"\n elif s == \"Мне\" or s == \"мне\":\n id_name = \"267303926\"\n return id_name\n\ndef remove_charmess(s):\n result = s.replace(\"Отправь сообщение \", \"\")\n return result\n\ndef remove_charmess1(s):\n result = s.replace(\"отправь сообщение \", \"\")\n return result\n\ndef message_send(event, context):\n intents = event.get('request', {}).get('nlu', {}).get('intents', {})\n command = event.get('request', {}).get('command')\n chto = event.get('request', {}).get('nlu', {}).get('intents', {}).get('vkmessage', {}).get('slots', {}).get('Chto', {}).get('value')\n chtostatus = event.get('request', {}).get('nlu', {}).get('intents', {}).get('statusset', {}).get('slots', {}).get('Chto', {}).get('value')\n komu = event.get('request', {}).get('nlu', {}).get('intents', {}).get('vkmessage', {}).get('slots', {}).get('Komu', {}).get('value')\n \n if intents.get('yes'):\n text = str(config.MESS_TEXT1) + str(chto) + str(config.MESS_TEXT2) + str(komu)\n komu_id = name_to_id(komu)\n api.messages.send(user_id = komu_id, message = chto)\n end_session = 'true'\n else:\n text = \"Хорошо, отменил отправку.\"\n end_session = 'true'\n\ndef handler(event, context):\n intents = event.get('request', {}).get('nlu', {}).get('intents', {})\n command = event.get('request', {}).get('command')\n chto = event.get('request', {}).get('nlu', {}).get('intents', {}).get('vkmessage', {}).get('slots', {}).get('Chto', {}).get('value')\n chtostatus = event.get('request', {}).get('nlu', {}).get('intents', {}).get('statusset', {}).get('slots', {}).get('Chto', {}).get('value')\n komu = event.get('request', {}).get('nlu', {}).get('intents', {}).get('vkmessage', {}).get('slots', {}).get('Komu', {}).get('value')\n ev = event\n cont = context\n\n\n text = config.INTRO_TEXT\n end_session = 'false'\n\n if intents.get('exit'):\n text = 'Ну пока) Жду снова!'\n end_session = 'true'\n elif intents.get('help'):\n text = config.INTRO_TEXT\n end_session = 'true'\n elif intents.get('vkpost'):\n messagepost1 = remove_char(command)\n messagepost = remove_char1(messagepost1)\n textold = config.POST_TEXT1 + messagepost + config.POST_TEXT2\n text = textold\n api.wall.post(message = messagepost)\n end_session = 'true'\n elif intents.get('vkmessage'):\n text_mess1 = del_name(command)\n text_mess2 = remove_charmess(text_mess1)\n text_mess = remove_charmess1(text_mess2)\n if text_mess == None:\n text = \"Вы не сказали что отправить...\"\n end_session = 'true'\n else:\n text = \"Вы подтверждаете отправку? Скажите 'Да' или 'Нет'.\"\n end_session = 'false'\n message_send(ev, cont)\n\n elif intents.get('statusset'):\n api.status.set(text = chtostatus)\n text = str(config.STAT_TEXT1) + str(chtostatus) + str(config.STAT_TEXT2)\n end_session = 'true'\n elif command:\n text = 'Не поняла тебя. Для выхода, скажи \"Хватит\".'\n end_session = 'true'\n\n return {\n 'version': event['version'],\n 'session': event['session'],\n 'response': {\n 'text': text,\n 'end_session': end_session\n },\n }\n","repo_name":"maxivimax/alice-vk","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5229,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"71084825504","text":"from django.shortcuts import get_object_or_404, render, redirect\nfrom django.http import HttpResponse, Http404\nfrom .models import Message\n\n\n# Create your views here.\ndef index(request):\n if request.method == 'POST':\n new_message = Message.objects.create(\n send_text=request.POST['contents']\n )\n return redirect('/chat/')\n\n latest_message_list = Message.objects.order_by('pub_date')[:20]\n context = {\n 'latest_message_list':latest_message_list\n }\n\n return render(request, 'chat/index.html', context)\n\ndef detail(request, message_id):\n message = get_object_or_404(Message, pk=message_id)\n return render(request, 'chat/detail.html', {'message':message})\n","repo_name":"daheejo/Sinabro","sub_path":"woori-origin/chat/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":712,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"32447371027","text":"# get list of Ballers\nimport requests\nfrom bs4 import BeautifulSoup\n# scrape top 50 players from here\nurl = \"https://www.ranker.com/crowdranked-list/top-10-current-nba-players?var=6&utm_expid=16418821-388.pwrOe5-lSJetLqzNW0S00A.1&utm_referrer=https%3A%2F%2Fwww.google.com%2F\"\n\nresp = requests.get(url)\nprint(resp.status_code)\nsoup = BeautifulSoup(resp.text, 'lxml')\n\n# this gets first 50 out of 100 ballers\n# 51-100 only load when scroll to bottom of page\n# These could be gathered using selenium but for now 50 will do.\nnames = soup.findAll('a', class_='listItem__title listItem__title--link black $tkl')\nnames = [n.text for n in names]\nlen(names)\n\n\nwith open('ballers.txt', 'w') as f:\n f.write('\\n'.join(names))\n\n\n# This is a list of all players\n# will use for named entity recognition for FB\n# NM these HAVE to be gathered with selenium\nimport os\nfrom selenium import webdriver\nfrom selenium.webdriver.support.ui import WebDriverWait\nchromedriver = '/Users/rick.shapiro/Documents/Python_Programs/Selenium/Drivers/chromedriver'\nos.environ['webdriver.chrome.driver'] = chromedriver\n\nbrowser = webdriver.Chrome(chromedriver)\n\nurl = \"https://stats.nba.com/players/list/\"\nbrowser.get(url) # spins for a while...\nlis = browser.find_elements_by_css_selector('li.players-list__name a')\nnames = [l.text for l in lis]\nlen(names)\nbrowser.close()\n\nwith open('all_ballers.txt', 'w') as f:\n f.write('\\n'.join(names))\n","repo_name":"rshap91/nba_social","sub_path":"getBallerNames.py","file_name":"getBallerNames.py","file_ext":"py","file_size_in_byte":1412,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7603507501","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n@File : main.py\n@Author: XuYaoJian\n@Date : 2021/11/15 22:42\n@Desc : \n\"\"\"\n# Definition for a binary tree node.\nfrom typing import List\n\n\nclass TreeNode:\n def __init__(self, val=0, left=None, right=None):\n self.val = val\n self.left = left\n self.right = right\n\n\nclass Solution:\n def buildTree(self, preorder: List[int], inorder: List[int]) -> TreeNode:\n inMap = {value: idx for idx, value in enumerate(inorder)}\n pre_end = len(preorder) - 1\n in_end = len(inorder) - 1\n def buildTree(pre_start, pre_end, in_start, in_end):\n \"\"\"\n :param pre_start:前序遍历中的 树的左边界下标\n :param pre_end: 前序遍历中的 树的右边界下标\n :param in_start: 中序遍历中的 树的左边界下标\n :param in_end: 中序遍历中的 树的右边界下标\n :return:\n \"\"\"\n if pre_start > pre_end:\n return None\n root = TreeNode(preorder[pre_start])\n index = inMap[root.val]\n nums = index - in_start\n # 左子树\n root.left = buildTree(pre_start+1 , nums + pre_start, in_start, index-1)\n # 右子树\n root.right = buildTree(nums+pre_start+1, pre_end, index+1, in_end)\n return root\n root = buildTree(0, pre_end, 0, in_end)\n return root\n\nif __name__ == \"__main__\":\n s = Solution()\n preorder = [3, 9, 20, 15, 7]\n inorder = [9, 3, 15, 20, 7]\n root = s.buildTree(preorder, inorder)\n print(root.val)\n\n","repo_name":"xuyaojian123/shuaTi","sub_path":"105从前序与中序遍历序列构造二叉树/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1621,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"73947898462","text":"from subprocess import check_output, CalledProcessError, PIPE, STDOUT\nimport os\nimport shutil\nimport logging\nfrom .utils import execute\n\n\nclass SourceManager:\n def __init__(self, repo_url):\n self.repo_url = repo_url\n self.src_dir = os.path.abspath(\".src\")\n self.logger = logging.getLogger(\"core.SourceManager\")\n\n def check(self, repo_url, repo_dir):\n if not os.path.exists(repo_dir) or not os.path.isdir(repo_dir):\n return False\n wd = os.getcwd()\n try:\n os.chdir(repo_dir)\n try:\n self._execute(f\"git status\")\n except CalledProcessError:\n return False\n\n return self._get_origin_url() == repo_url\n finally:\n os.chdir(wd)\n\n def _get_origin_url(self):\n cmd = f\"git remote get-url origin\"\n output = execute(cmd)\n self.logger.debug(\"$ %s\\n%s\", cmd, output)\n return output.strip()\n\n def _clone_repo(self, repo_url, repo_dir):\n self._execute(f\"git clone {repo_url} {repo_dir}\")\n\n def ensure_repo(self, repo_url, repo_dir):\n if not self.check(repo_url, repo_dir):\n if os.path.exists(repo_dir):\n shutil.rmtree(repo_dir)\n\n if not os.path.exists(repo_dir):\n self._clone_repo(repo_url, repo_dir)\n\n def ensure(self, version):\n repo_dir = self.src_dir\n self.ensure_repo(self.repo_url, repo_dir)\n self.checkout(repo_dir, version)\n\n def get_ref(self, version):\n if version == \"latest\":\n return \"master\"\n else:\n return \"v\" + version\n\n def get_dockerfile(self, version):\n return \"Dockerfile\"\n\n def get_build_args(self, version):\n return {}\n\n def _execute(self, cmd):\n output = execute(cmd)\n self.logger.debug(\"$ %s\\n%s\", cmd, output)\n\n def checkout_repo(self, repo_dir, ref):\n wd = os.getcwd()\n try:\n os.chdir(repo_dir)\n self._execute(f\"git fetch\")\n self._execute(f\"git checkout {ref}\")\n self._execute(f\"git pull origin {ref}\")\n self._execute(f\"git clean -xfd\")\n finally:\n os.chdir(wd)\n\n def checkout(self, repo_dir, version):\n ref = self.get_ref(version)\n self.checkout_repo(repo_dir, ref)\n\n def get_revision(self, repo_dir):\n wd = os.getcwd()\n try:\n os.chdir(repo_dir)\n output = execute(f\"git rev-parse HEAD\")\n return output.strip()\n finally:\n os.chdir(wd)\n\n def get_application_revision(self, version):\n return self.get_revision(self.src_dir)\n","repo_name":"ExchangeUnion/xud-docker","sub_path":"tools/core/src.py","file_name":"src.py","file_ext":"py","file_size_in_byte":2658,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"7"} +{"seq_id":"13736287762","text":"import stadv\r\nimport tensorflow as tf\r\nimport numpy as np\r\nimport pickle\r\nimport datetime\r\n\r\ndef save_obj(obj, name ):\r\n with open(name, 'wb') as f:\r\n pickle.dump(obj, f, pickle.HIGHEST_PROTOCOL)\r\n \r\n \r\n \r\ndef load_obj( name):\r\n with open(name, 'rb') as f:\r\n return pickle.load(f)\r\n\r\n\r\ndef targetedAdvRate(target, perb_preds, orig_preds, groundTruth):\r\n \r\n succCount = 0\r\n potentialCount =0\r\n for i in range(orig_preds.shape[0]):\r\n if orig_preds[i]!=target and orig_preds[i] == groundTruth[i]:\r\n potentialCount = potentialCount+1\r\n if orig_preds[i]!=perb_preds[i] and perb_preds[i]== target:\r\n succCount=succCount+1\r\n \r\n return round(succCount/potentialCount,2)\r\n\r\n\r\ndef unTargetedAdvRate(perb_preds, orig_preds,groundTruth):\r\n \r\n succCount = 0\r\n potentialCount =0\r\n for i in range(orig_preds.shape[0]):\r\n if orig_preds[i] == groundTruth[i]:\r\n potentialCount = potentialCount+1\r\n if orig_preds[i]!=perb_preds[i]:\r\n succCount=succCount+1\r\n \r\n return round(succCount/potentialCount,3)\r\n\r\ndef advRate(targeted, target, perb_preds, orig_preds, groundTruth):\r\n if targeted: \r\n return targetedAdvRate(target, perb_preds, orig_preds, groundTruth)\r\n else:\r\n return unTargetedAdvRate(perb_preds, orig_preds, groundTruth)\r\n\r\ndef calcPerbPreds(X_testData, perbFlow, model, modelFile):\r\n \r\n saver = tf.train.Saver()\r\n images = tf.constant(X_testData, dtype = tf.float32)\r\n perbFlow_tf = tf.constant(perbFlow)\r\n perb_images = stadv.layers.flow_st(images, perbFlow_tf)\r\n perb_preds = model.make_pred(perb_images)\r\n \r\n with tf.Session() as sess:\r\n sess.run(tf.global_variables_initializer())\r\n saver.restore(sess,modelFile)\r\n perb_preds_np = sess.run(perb_preds, feed_dict = {images:X_testData})\r\n \r\n return perb_preds_np\r\n\r\n\r\n\r\ndef calcPreds(X_testData, model, modelFile):\r\n \r\n saver = tf.train.Saver()\r\n images = tf.constant(X_testData, dtype = tf.float32)\r\n \r\n preds = model.make_pred(images)\r\n \r\n with tf.Session() as sess:\r\n sess.run(tf.global_variables_initializer())\r\n saver.restore(sess,modelFile)\r\n preds_np = preds.eval()\r\n #preds_np = sess.run(preds, feed_dict = {images:X_testData})\r\n \r\n return preds_np\r\n \r\ndef convertOneHot2Labels(oneHotLabels):\r\n return [np.where(r==1)[0][0] for r in oneHotLabels]\r\n\r\ndef getNowString(date = True):\r\n \r\n if date:\r\n out = str(datetime.datetime.now()).replace(\":\",\"-\").replace(\" \",\"_\")[0:19]\r\n else:\r\n out = str(datetime.datetime.now().time()).replace(\":\",\"-\")[0:8]\r\n \r\n return out\r\n\r\ndef convert2OneHot(indices, numClasses):\r\n \r\n return np.eye(numClasses)[indices]","repo_name":"tomstelk/Universal-Adversarial-Flow-Network","sub_path":"uan_utils.py","file_name":"uan_utils.py","file_ext":"py","file_size_in_byte":2856,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70707443743","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('kitbuilder', '0003_auto_20141222_1050'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='customkit',\n name='user',\n field=models.ForeignKey(related_name='custom_kits', to='userprofile.UserProfile'),\n preserve_default=True,\n ),\n ]\n","repo_name":"bantonelli/CustomDrumSamples","sub_path":"kitbuilder/migrations/0004_auto_20141223_0621.py","file_name":"0004_auto_20141223_0621.py","file_ext":"py","file_size_in_byte":485,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70331452063","text":"#Kadane’s Algorithm:\n\ndef maxSubArraySum(a, size):\n max = a[0]\n sum = 0\n\n for i in range(0, size):\n sum += a[i]\n if sum < 0:\n max = sum\n sum = 0\n\n\n # Do not compare for all elements. Compare only\n # when max_ending_here > 0\n elif (max < sum):\n max = sum\n\n return max\n\n\n# Driver function to check the above function\na = [-2,1,-3,4,-1,2,1,-5,4]\nans = maxSubArraySum(a, len(a))\nprint(ans)\n","repo_name":"amit-223/PythonCodes","sub_path":"pythonProject/LeetCode/Arrays/Maximum Subarray.py","file_name":"Maximum Subarray.py","file_ext":"py","file_size_in_byte":468,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"23114098459","text":"import pandas as pd\nimport urllib.request\nfrom bs4 import BeautifulSoup as bs\nimport matplotlib.pyplot as plt\nfrom lifelines import KaplanMeierFitter\n\nplt.style.use('ggplot')\n###looping to get data from 12 different pages\nlist=[]\nfor i in range(1,12):\n ### MAKING SOUP ###\n url = \"http://stats.espncricinfo.com/ci/engine/stats/index.html?class=2;filter=advanced;orderby=runs;page=\" + str(\n i) + \";size=200;template=results;type=batting\"\n url = urllib.request.urlopen(url).read()\n soup = bs(url, 'html.parser')\n table = soup.find_all('table')\n table = table[2]\n\n ### COLUMN_NAMES ###\n thead = table.findAll('thead')\n thead = thead[0].text\n thead = thead.splitlines()\n thead = thead[2:15]\n\n ### TABLE FORMING ###\n tbody = table.findAll('tbody')\n row = tbody[0].findAll('tr')\n for i in range(len(row)):\n single_row = row[i].find_all('td')\n single_row = [single_row[int(j)].text for j in range(len(single_row))][:13]\n list.append(single_row)\n\ndata = pd.DataFrame(list, columns=thead)\n\n### ELIMINATING CURRENT PLAYER STATS ###\nspan_data = data.ix[:, 1]\nindex = []\nfor i in range(len(span_data)):\n if (span_data[i][5:9] == '2016') or (span_data[i][5:9] == '2017' ):\n index.append(i)\n\ndata = data.drop(index)\ndata = data.reset_index(drop=True)\n\n###CREATING CAREER SPAN YEAR DATA (start,end,span)\nspan_data1 = data.ix[:, 1]\nfor i in range(len(span_data1)):\n start = span_data1[i].split('-')[0]\n end = span_data1[i].split('-')[1]\n data.ix[i, 'start'] = start\n data.ix[i, 'end'] = end\n\nfor i in range(len(data)):\n data.ix[i, 'span'] = int(data.ix[i, 'end']) - int(data.ix[i, 'start'])\n\n\n###seprating player-name and team_name(ICC/ASIA/INDIA)\nname = data.ix[:, 0]\ncountry_list = []\nfor i in range(len(name)):\n namelist = name[i].split('(')[0]\n data.ix[i, 'player'] = namelist\n countrylist = name[i].split('(')[1]\n countrylist = countrylist.split(')')[0]\n countrylist = countrylist.split('/')\n country_list.append(countrylist)\n\ncountries = ['INDIA', 'SL', 'AUS', 'ENG', 'USA', 'NZ', 'PAK', 'KENYA', 'BMUDA', 'SCOT', 'WI', 'BDESH', 'AFG', 'IRE',\n 'NAM', 'ZIM', 'HK', 'Neth', 'UAE', 'Can', 'EAf', 'SA']\n\n###more than 1 team is present in teams_name(ICC/ASIA/INDIA),so seprating and adding\n### -- country name to the team\nfor i in range(len(country_list)):\n for j in range(len(country_list[i])):\n for country in countries:\n if (country_list[i][j] == country):\n data.ix[i, 'country'] = country_list[i][j]\n\n###droping original form of span and name\ndata = data.drop('Player', axis=1)\ndata = data.drop('Span', axis=1)\n\n###renaming the column_headings\ndata.columns = ['mat', 'inns', 'no', 'runs', 'hs', 'avg', 'bf', 'sr', '100', '50', '0', 'start', 'end', 'span','player',\n 'country']\n\n###rearranging the columns\ndata = data[['player', 'country', 'mat', 'inns', 'no', 'runs', 'hs', 'avg', 'bf', 'sr', '100', '50', '0', 'start', 'end',\n 'span']]\n\n#batsmen_data = data\n\n#data.to_csv('data.csv', sep=',')\n#print(batsmen_data)\n\n\n#----------------------------------(i) Player's Country vs Career Length -------------------------------------------------------\n\ndata = pd.read_csv(\"data.csv\")\ndata.ix[:,'censor']=1\n\ndata = pd.DataFrame(data)\nduration = data['span']\nobserved = data.ix[:,'censor']\n\nkmf = KaplanMeierFitter()\nkmf.fit(duration,observed,label='kmf_mean')\n#kmf.plot()\n#plt.show()\n\n###INDIA kmf\n\nindia_data = data.ix[data['country']=='INDIA']\nindia_duration = india_data['span']\nindia_observed = india_data['censor']\n\nkmfind = KaplanMeierFitter()\nkmfind.fit(india_duration,india_observed,label=\"india\")\n\n###simillarly for other countries\nkmfpak =KaplanMeierFitter()\nkmfpak.fit((data.ix[data['country']=='PAK'])['span'],(data.ix[data['country']=='PAK'])['censor'],label='pakistan')\n\nkmfaus =KaplanMeierFitter()\nkmfaus.fit((data.ix[data['country']=='AUS'])['span'],(data.ix[data['country']=='AUS'])['censor'],label='australia')\n\nkmfsa =KaplanMeierFitter()\nkmfsa.fit((data.ix[data['country']=='SA'])['span'],(data.ix[data['country']=='SA'])['censor'],label='south_africa')\n\nkmfbdesh =KaplanMeierFitter()\nkmfbdesh.fit((data.ix[data['country']=='BDESH'])['span'],(data.ix[data['country']=='BDESH'])['censor'],label='bangladesh')\n\nkmfnz =KaplanMeierFitter()\nkmfnz.fit((data.ix[data['country']=='NZ'])['span'],(data.ix[data['country']=='NZ'])['censor'],label='newzealand')\n\nkmfwi =KaplanMeierFitter()\nkmfwi.fit((data.ix[data['country']=='WI'])['span'],(data.ix[data['country']=='WI'])['censor'],label='westindies')\n\nkmfeng =KaplanMeierFitter()\nkmfeng.fit((data.ix[data['country']=='ENG'])['span'],(data.ix[data['country']=='ENG'])['censor'],label='england')\n\nkmfsl =KaplanMeierFitter()\nkmfsl.fit((data.ix[data['country']=='SL'])['span'],(data.ix[data['country']=='SL'])['censor'],label='srilanka')\n\nkmfcan =KaplanMeierFitter()\nkmfcan.fit((data.ix[data['country']=='Can'])['span'],(data.ix[data['country']=='Can'])['censor'],label='canada')\n\nkmfhk =KaplanMeierFitter()\nkmfhk.fit((data.ix[data['country']=='HK'])['span'],(data.ix[data['country']=='HK'])['censor'],label='hongkong')\n\nkmfscot =KaplanMeierFitter()\nkmfscot.fit((data.ix[data['country']=='SCOT'])['span'],(data.ix[data['country']=='SCOT'])['censor'],label='scotland')\n\nkmfuae =KaplanMeierFitter()\nkmfuae.fit((data.ix[data['country']=='UAE'])['span'],(data.ix[data['country']=='UAE'])['censor'],label='uae')\n\nkmfeaf =KaplanMeierFitter()\nkmfeaf.fit((data.ix[data['country']=='EAf'])['span'],(data.ix[data['country']=='EAf'])['censor'],label='EAf')\n\nkmfzim =KaplanMeierFitter()\nkmfzim.fit((data.ix[data['country']=='ZIM'])['span'],(data.ix[data['country']=='ZIM'])['censor'],label='zimbabwe')\n\nkmfneth =KaplanMeierFitter()\nkmfneth.fit((data.ix[data['country']=='Neth'])['span'],(data.ix[data['country']=='Neth'])['censor'],label='netherland')\n\nkmfkenya =KaplanMeierFitter()\nkmfkenya.fit((data.ix[data['country']=='KENYA'])['span'],(data.ix[data['country']=='KENYA'])['censor'],label='kenya')\n\nkmfbmuda =KaplanMeierFitter()\nkmfbmuda.fit((data.ix[data['country']=='BMUDA'])['span'],(data.ix[data['country']=='BMUDA'])['censor'],label='burmuda')\n\nkmfire =KaplanMeierFitter()\nkmfire.fit((data.ix[data['country']=='IRE'])['span'],(data.ix[data['country']=='IRE'])['censor'],label='ireland')\n\nkmfnam =KaplanMeierFitter()\nkmfnam.fit((data.ix[data['country']=='NAM'])['span'],(data.ix[data['country']=='NAM'])['censor'],label='namibia')\n\nkmfusa =KaplanMeierFitter()\nkmfusa.fit((data.ix[data['country']=='USA'])['span'],(data.ix[data['country']=='USA'])['censor'],label='usa')\n\n\nkmfafg =KaplanMeierFitter()\nkmfafg.fit((data.ix[data['country']=='AFG'])['span'],(data.ix[data['country']=='AFG'])['censor'],label='afghanistan')\n\n\n\n####PLOTING kmf in PLOT\n\nax=plt.subplot(111)\n#kmf.survival_function_.plot(ax=ax)\n#kmfaus.survival_function_.plot(ax=ax)\n#kmfpak.survival_function_.plot(ax=ax)\nkmfind.survival_function_.plot(ax=ax)\nkmfeng.survival_function_.plot(ax=ax)\n#kmfsa.survival_function_.plot(ax=ax)\nkmfwi.survival_function_.plot(ax=ax)\n#kmfbdesh.survival_function_.plot(ax=ax)\n#kmfsl.survival_function_.plot(ax=ax)\n#kmfnz.survival_function_.plot(ax=ax)\n#kmfusa.survival_function_.plot(ax=ax)\n#kmfzim.survival_function_.plot(ax=ax)\n#kmfhk.survival_function_.plot(ax=ax)\n#kmfbmuda.survival_function_.plot(ax=ax)\n#kmfeaf.survival_function_.plot(ax=ax)\n#kmfneth.survival_function_.plot(ax=ax)\n#kmfscot.survival_function_.plot(ax=ax)\n#kmfafg.survival_function_.plot(ax=ax)\n#kmfnam.survival_function_.plot(ax=ax)\n#kmfkenya.survival_function_.plot(ax=ax)\n#kmfuae.survival_function_.plot(ax=ax)\n#kmfcan.survival_function_.plot(ax=ax)\n#kmfire.survival_function_.plot(ax=ax)\n\nplt.show()\n\n\n#-------------------------------------(ii) Strike Rate vs Career length------------------------------------------------------------\n## creating the probability of career length of player vs their strike rate\n##atleast 50 innings player\ndata['inns'] = pd.to_numeric(data['inns'])\ninnings = data[:][data['inns']>50]\n\n\ninnings['sr'] = pd.to_numeric(innings['sr'])\n\nvhighsr = innings.ix[innings['sr']>=100]\nhighsr = innings.ix[innings['sr']>=90]\nhighsr = highsr.ix[highsr['sr']<100]\ngoodsr = innings.ix[innings['sr']>=80]\ngoodsr = goodsr.ix[goodsr['sr']<90]\navgsr = innings.ix[innings['sr']>=70]\navgsr = avgsr.ix[avgsr['sr']<80]\npoorsr = innings.ix[innings['sr']>=55]\npoorsr = poorsr.ix[innings['sr']<70]\nvpoorsr = innings.ix[innings['sr']<55]\n\n\nkmfvhighsr = KaplanMeierFitter()\nkmfvhighsr.fit(vhighsr['span'],vhighsr['censor'],label = ' sr > 100' )\n\nkmfhighsr = KaplanMeierFitter()\nkmfhighsr.fit(highsr['span'],highsr['censor'],label = '100 > sr > 90' )\n\nkmfgoodsr = KaplanMeierFitter()\nkmfgoodsr.fit(goodsr['span'],goodsr['censor'],label='90 > sr > 80')\n\nkmfavgsr = KaplanMeierFitter()\nkmfavgsr.fit(avgsr['span'],avgsr['censor'] , label = '80 > sr > 70')\n\nkmfpoorsr = KaplanMeierFitter()\nkmfpoorsr.fit(poorsr['span'],poorsr['censor'],label='70 > sr > 55')\n\nkmfvpoorsr = KaplanMeierFitter()\nkmfvpoorsr.fit(vpoorsr['span'],vpoorsr['censor'],label='sr < 55')\n\n\n\nkmfmat = KaplanMeierFitter()\nkmfmat.fit(innings['span'],innings['censor'],label = 'mean')\n\nbx = plt.subplot(111)\n\nkmfmat.survival_function_.plot(ax=bx)\nkmfvhighsr.survival_function_.plot(ax=bx)\nkmfhighsr.survival_function_.plot(ax=bx)\nkmfgoodsr.survival_function_.plot(ax=bx)\nkmfavgsr.survival_function_.plot(ax=bx)\nkmfpoorsr.survival_function_.plot(ax=bx)\nkmfvpoorsr.survival_function_.plot(ax=bx)\n\n\nplt.show()\n\n\n#-------------------------------- (iii) Runs vs Career_Length ------------------------------------------------------------------------------------------------\n\n\ndata['runs'] = pd.to_numeric(data['runs'])\n\nruns8000 = data.ix[data['runs']>=8000]\nruns3000 = data.ix[data['runs']<=3000]\n#runs3000 = runs3000.ix[runs3000['runs']< 3000]\n\nkmfruns8000 = KaplanMeierFitter()\nkmfruns8000.fit(runs8000['span'],runs8000['censor'],label = ' runs > 8000' )\n\nkmfruns3000 = KaplanMeierFitter()\nkmfruns3000.fit(runs3000['span'],runs3000['censor'],label = ' runs < 3000')\n\nbx = plt.subplot(111)\nkmfruns8000.survival_function_.plot(ax=bx)\nkmfruns3000.survival_function_.plot(ax=bx)\n\n\nplt.xlabel(\" career length ( in years )\")\nplt.ylabel(\" probability of players \")\nplt.title(\"probability of players with specific runs vs their career length\")\nplt.show()","repo_name":"rushil83/Cricket_Survival_Analysis","sub_path":"Cricket_plus_Batsman.py","file_name":"Cricket_plus_Batsman.py","file_ext":"py","file_size_in_byte":10296,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"44367140229","text":"import re\nimport shutil\nimport os.path\nfrom xml.etree.ElementTree import ElementTree\nfrom concurrent_tree_crawler.abstract_tree_navigator import NavigationException\nfrom concurrent_tree_crawler.common.file_helper import lenient_makedir\nfrom concurrent_tree_crawler.html_multipage_navigator.abstract_page_analyzer \\\n\timport PageLinks, AbstractPageAnalyzer, Level, AbstractLevelsCreator\n\nclass PageAnalyzerException(NavigationException):\n\tpass\n\nclass MagazinePageAnalyzer(AbstractPageAnalyzer):\n\t\"\"\"A class that parses magazine-level pages\"\"\"\n\t\n\tdef get_links(self, page_file, child_links_retrieved_so_far):\n\t\tchildren = []\n\t\tdoc = ElementTree(file=page_file)\n\t\tlinks = doc.findall(\"body/div/div[@id='elements']/table/tr/td/a\")\n\t\tfor link in links:\n\t\t\tlink_text = self.__convert_date(link.text)\n\t\t\tchildren.append((link_text, link.attrib[\"href\"]))\n\t\tnext_page_elem = \\\n\t\t\tdoc.find(\"body/div/div[@id='navigation']/table/tr/td[3]/a\")\n\t\tnext_page_link = None\n\t\tif next_page_elem is not None:\n\t\t\tnext_page_link = next_page_elem.attrib[\"href\"]\n\t\treturn PageLinks(children, next_page_link)\n\n\t@staticmethod\n\tdef __convert_date(text):\n\t\tmonths = {\"January\": 1, \"February\": 2, \"March\": 3, \"April\": 4,\n\t\t\t\"May\": 5, \"June\": 6, \"July\": 7, \"August\": 8, \"September\": 9,\n\t\t\t\"October\": 10, \"November\": 11, \"December\": 12}\n\t\t(month_str, day_str, year_str) = \\\n\t\t\tre.match(\"^(\\w+) (\\d+), (\\d+)$\", text).group(1, 2, 3)\n\t\tmonth = _convert_to_2_digit_number(\n\t\t\tint(months[month_str]))\n\t\tday = _convert_to_2_digit_number(int(day_str))\n\t\treturn \"{}-{}-{}\".format(year_str, month, day)\n\nclass IssuePageAnalyzer(AbstractPageAnalyzer):\n\t\"\"\"A class that parses issues-level pages\"\"\"\n\n\tdef __init__(self, dst_dir_path):\n\t\tself.__dst_dir_path = dst_dir_path\t\n\n\tdef process(self, tree_path, page_file):\n\t\tassert len(tree_path) > 0\n\t\tdir_path = os.path.join(self.__dst_dir_path,\n\t\t\t_convert_tree_path_to_dir_path(tree_path))\n\t\tlenient_makedir(dir_path)\n\t\terror_page_path = os.path.join(dir_path, \"error.txt\")\n\t\t_handle_error_page(page_file, error_page_path)\n\n\tdef get_links(self, page_file, child_links_retrieved_so_far):\n\t\tchildren = []\n\t\tdoc = ElementTree(file=page_file)\n\t\tlinks = doc.findall(\"body/div/div[@id='elements']/table/tr/td/a\")\n\t\tfor i, link in enumerate(links):\n\t\t\t#link_text = self.__convert_title(link.text)\n\t\t\tlink_text = _convert_to_2_digit_number(\n\t\t\t\tchild_links_retrieved_so_far + i+1)\n\t\t\tchildren.append((link_text, link.attrib[\"href\"]))\n\t\tnext_page_elem = \\\n\t\t\tdoc.find(\"body/div/div[@id='navigation']/table/tr/td[3]/a\")\n\t\tnext_page_link = None\n\t\tif next_page_elem is not None:\n\t\t\tnext_page_link = next_page_elem.attrib[\"href\"]\n\t\treturn PageLinks(children, next_page_link)\n\nclass ArticlePageAnalyzer(AbstractPageAnalyzer):\n\t\"\"\"A class that downloads article pages\"\"\"\n\t\n\tdef __init__(self, dst_dir_path):\n\t\tself.__dst_dir_path = dst_dir_path\n\n\tdef process(self, tree_path, page_file):\n\t\tassert len(tree_path) > 0\n\t\tdir_path = os.path.join(self.__dst_dir_path,\n\t\t\t_convert_tree_path_to_dir_path(tree_path[:-1]))\n\t\tlenient_makedir(dir_path)\n\t\terror_page_path = os.path.join(dir_path, tree_path[-1]+\"-error.txt\")\n\t\t_handle_error_page(page_file, error_page_path)\n\t\tfile_path = os.path.join(dir_path, tree_path[-1]+\".html\")\n\t\tself.__download_page(page_file, file_path)\n\n\tdef __download_page(self, page_file, dst_file):\n\t\tf = open(dst_file, 'wb')\n\t\tshutil.copyfileobj(page_file, f)\n\t\tf.close()\n\nclass LevelsCreator(AbstractLevelsCreator):\n\tdef __init__(self, download_dir_path):\n\t\tself.__download_dir_path = download_dir_path\n\n\tdef create(self):\n\t\treturn [Level(\"magazine\", MagazinePageAnalyzer()),\n\t\t\t\tLevel(\"issue\", IssuePageAnalyzer(self.__download_dir_path)),\n\t\t\t\tLevel(\"article\", ArticlePageAnalyzer(self.__download_dir_path))]\n\ndef _handle_error_page(page_file, file_path):\n\t\"\"\"@return: C{True} iff the given page is an error page\"\"\"\n\tif not _is_error_page(page_file):\n\t\treturn\n\twith open(file_path, \"w\") as f:\n\t\tprint >>f, \"An error message shown after requesting the page\"\n\traise PageAnalyzerException(\"Error page encountered\")\n\ndef _is_error_page(page_file):\n\tdoc = ElementTree(file=page_file)\n\ttitle = doc.find(\"head/title\")\n\tpage_file.seek(0) ## reset file's current position\n\tif title.text.find(\"not available\") != -1:\n\t\treturn True\n\treturn False\n\ndef _convert_tree_path_to_dir_path(tree_path):\n\t## Skip root element\n\treturn '/'.join(tree_path[1:])\n\ndef _convert_to_2_digit_number(number):\n\tassert number > 0 and number < 99\n\tstr_ = str(number)\n\tif number < 10:\n\t\treturn \"0\"+str_\n\treturn str_\n","repo_name":"mkobos/tree_crawler","sub_path":"concurrent_tree_crawler/html_multipage_navigator/sample_page_analyzer.py","file_name":"sample_page_analyzer.py","file_ext":"py","file_size_in_byte":4490,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"7"} +{"seq_id":"21991588417","text":"import subprocess\nimport numpy as np\nfrom argparse import ArgumentParser as arp\nimport os\nimport sys\n\ndef run(cmd):\n p = subprocess.run(cmd,\n capture_output=True,\n text=True,\n shell=True)\n if p.stdout or p.stderr:\n print(p.stdout, p.stderr)\n\ndef run2(cmd):\n p = subprocess.Popen(cmd,shell=True)\n p.communicate()\n p.wait()\n\ndef get_args():\n p = arp(description=\"read videos given a master list\")\n p.add_argument(\"inp\",\n help=\"input master file\")\n p.add_argument(\"--height\", default=1080,\n help=\"desired height\")\n args = p.parse_args()\n return args\n\ndef main():\n args = get_args()\n afile = args.inp\n height = args.height\n # print(afile, height)\n # sys.exit(1)\n with open(afile,\"r\") as f:\n count = 1\n for line in f.readlines():\n url = line.strip()\n outdir = f\"{count:02}\"\n os.makedirs(outdir, exist_ok=True)\n os.chdir(outdir)\n cmd = [ \n 'node',\n '../vimeo-downloader.js',\n url,\n height]\n # print(\" \".join(map(str, cmd)))\n # run2(cmd)\n cmd = [\n \"mkvmerge\",\n \"-o\",\n f\"{count:02}.mkv\",\n \"*.m4v\",\n \"*.m4a\"\n ]\n # here we must join so that globbing works correctly\n cmd = \" \".join(cmd)\n print(cmd)\n run2(cmd)\n os.chdir(\"..\")\n count += 1\n\nmain()\n","repo_name":"frankieliu/vimeo","sub_path":"college-solution/merge-files.py","file_name":"merge-files.py","file_ext":"py","file_size_in_byte":1604,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"28837431377","text":"from openupgradelib import openupgrade\n\n\n@openupgrade.migrate()\ndef migrate(env, version):\n if env.ref(\n \"l10n_es_facturae_face.facturae_face_exchange_type_rule\",\n raise_if_not_found=False,\n ):\n # In this case, the upgrade passed on a previous version\n return\n for exchange_type in env.ref(\n \"l10n_es_facturae_face.facturae_exchange_type\"\n ) | env.ref(\"l10n_es_facturae_face.facturae_face_update_exchange_type\"):\n exchange_type.write(\n {\n \"model_ids\": [(5, 0, 0)],\n \"enable_snippet\": False,\n \"enable_domain\": False,\n }\n )\n exchange_type.rule_ids.unlink()\n","repo_name":"curonny/l10n-spain","sub_path":"l10n_es_facturae_face/migrations/16.0.1.0.0/pre-migration.py","file_name":"pre-migration.py","file_ext":"py","file_size_in_byte":690,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"7"} +{"seq_id":"26673425763","text":"from django.shortcuts import render, redirect\nfrom .models import League, Team, Player\nfrom django.db.models import Count\n\nfrom . import team_maker\n\ndef index(request):\n\tcontext = {\n\t\t\"all\": League.objects.all(),\n\t\t\"baseballLeagues\": League.objects.filter(sport=\"Baseball\"),\n\t\t\"womensLeagues\": League.objects.filter(name__contains=\"Women\"),\n\t\t\"hockeyLeagues\": League.objects.filter(sport__contains=\"Hockey\"),\n\t\t\"nonFootballLeagues\": League.objects.exclude(sport=\"Football\"),\n\t\t\"conferences\": League.objects.filter(name__contains=\"Conference\"),\n\t\t\"atlantic\": League.objects.filter(name__contains=\"Atlantic\"),\n\t\t\"teams\": Team.objects.all().order_by(\"location\", \"team_name\"),\n\t\t\"dallasTeams\": Team.objects.filter(location=\"Dallas\"),\n\t\t\"raptorTeams\": Team.objects.filter(team_name=\"Raptors\"),\n\t\t\"cityTeams\": Team.objects.filter(location__contains=\"City\"),\n\t\t\"tTeams\": Team.objects.filter(team_name__startswith=\"T\"),\n\t\t\"orderedByLocation\": Team.objects.all().order_by(\"location\"),\n\t\t\"reverseOrderTeam\": Team.objects.all().order_by(\"team_name\").reverse(),\n\t\t\"players\": Player.objects.all().order_by(\"last_name\"),\n\t\t\"cooperPlayers\": Player.objects.filter(last_name=\"Cooper\"),\n\t\t\"joshuaPlayers\": Player.objects.filter(first_name=\"Joshua\"),\n\t\t\"cooperNotJoshua\": Player.objects.filter(last_name=\"Cooper\").exclude(first_name=\"Joshua\"),\n\t\t\"alexanderOrWyatt\": Player.objects.filter(first_name=\"Alexander\") | Player.objects.filter(first_name=\"Wyatt\"),\n\t\t\"atlanticHockey\": League.objects.filter(name__contains=\"Atlantic\", sport__contains=\"Hockey\"),\n\t\t\"teamBosSta\": Team.objects.filter(location=\"Boston\", team_name=\"Stallions\"),\n\t\t\"intBaseConf\": Team.objects.filter(league=League.objects.get(name=\"International Baseball Conference\")),\n\t\t\"amFedSoc\": Team.objects.filter(league=League.objects.get(name=\"Atlantic League of Ice Hockey\")),\n\t\t\"footballLeagues\": League.objects.filter(sport=\"Football\"),\n\t\t\"teamsSophia\": Player.objects.filter(first_name=\"Sophia\"),\n\t\t\"teamsGonzales\": Player.objects.filter(last_name=\"Gonzales\"),\n\t\t\"johnHarris\": Player.objects.get(first_name=\"John\", last_name=\"Harris\"),\n\t\t\"ontarioColts\": Team.objects.get(location=\"Ontario\", team_name=\"Colts\"),\n\t\t\"nevadaBJ\": Team.objects.get(location=\"Nevada\", team_name=\"Blue Jays\"),\n\t\t\"jacksonWhite\": Player.objects.get(first_name=\"Jackson\", last_name=\"White\"),\n\t\t\"teamPlayerCount\": Team.objects.annotate(player_count=Count('all_players')).order_by('-player_count'),\n\t\t\"allPlayers\": Player.objects.annotate(numTeams=Count('all_teams')).order_by('-numTeams')\n\n\t}\n\treturn render(request, \"leagues/index.html\", context)\n\ndef make_data(request):\n\tteam_maker.gen_leagues(10)\n\tteam_maker.gen_teams(50)\n\tteam_maker.gen_players(200)\n\n\treturn redirect(\"index\")","repo_name":"Alastair-W/sports_orm","sub_path":"leagues/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2699,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"7"} +{"seq_id":"21055172609","text":"import json\nimport requests\nimport pickle\n\nclass Questions:\n\n def __init__(self,category,level,question,correct_answer,incorrect_answers):\n self.category = category\n self.level = level\n self.question = question\n self.correct_answer = correct_answer\n self.incorrect_answer = incorrect_answers\n self.isAsk = False\n\n\ndef selectQuestion(list,level):\n for question in list:\n if question.level == level and not question.isAsk:\n return question\n\ndef QuestionList(category):\n question_list = []\n\n science_URL = \"https://opentdb.com/api.php?amount=50&type=multiple\"\n sports_URL = \"https://opentdb.com/api.php?amount=50&category=21&type=multiple\"\n history_URL = \"https://opentdb.com/api.php?amount=50&category=23&type=multiple\"\n\n API_URL = \"https://opentdb.com/api.php?amount=50&type=multiple\" #default\n\n if category == \"science\":\n API_URL = science_URL\n elif category == \"sports\":\n API_URL = sports_URL\n elif category == \"history\":\n API_URL = history_URL\n\n response = requests.get(API_URL)\n questionsResponse = response.json()\n\n for i in questionsResponse['results']:\n question = Questions(i['category'],i['difficulty'],i['question'],i['correct_answer'],i['incorrect_answers'])\n question_list.append(question)\n\n return question_list\n\n\n\nlist = QuestionList(\"sports\")\n\nwith open('fileSports.pickle', 'wb') as questionList:\n pickle.dump(list, questionList)\n\nwith open('fileSports.pickle', 'rb') as questionList:\n result = pickle.load(questionList)\n","repo_name":"ahmetzumber/kivy-quiz-application-","sub_path":"datas.py","file_name":"datas.py","file_ext":"py","file_size_in_byte":1579,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"28988574405","text":"import itertools\nimport logging\nimport random\nimport sys\nfrom typing import List\nfrom functools import cached_property\nimport pickle\nimport os\nimport subprocess\n\nimport numpy as np\nimport shortuuid\nfrom itertools import islice\nimport re\n\nfrom config import save_path, strat_path, VERIFYTA\n\nfrom ETCetera.Scheduling.NTGA import network, controlloop\nfrom ETCetera import pyuppaal\nimport ETCetera.Abstractions as abstr\n\n\nclass nta:\n\n def __init__(self, net: network, control_loops: List[controlloop], synchronization=None):\n\n # TODO: Check compatibility w.r.t. synchronization signals\n self.net = net\n self.control_loops = control_loops\n self.ns = len(control_loops)\n self.synchronization = synchronization or [\"up\", \"down\", \"timeout\", \"ack\", \"nack\"]\n # self.max_early_triggers = max_early_triggers\n # Calculate common scaling factor to make UPPAAL happy\n import math\n scale = 1\n l = [a.scale for a in self.control_loops]\n for i in l:\n scale = scale * i // math.gcd(scale, i)\n self.common_scale = scale\n for s in self.control_loops:\n s.scale = scale\n logging.info(f'Set common scale to: {self.common_scale}')\n\n def _generate_declarations(self):\n if len(self.synchronization) == 0:\n return \"\"\n else:\n # return f'broadcast chan {\", \".join(self.actions)};\\n'\n decl = f'chan {\", \".join(self.synchronization)};\\n'\n decl += f'int EarNum;\\n'#const int EarMax = {self.max_early_triggers};\\n'\n return decl\n\n def _generate_system(self):\n system_ass = ''\n systems = []\n for cl in self.control_loops:\n system_ass += f'{cl.name}{cl.index} = {cl.name}();\\n'\n systems.append(f'{cl.name}{cl.index}')\n\n system_ass += f'{self.net.name}{self.net.index} = {self.net.name}();\\n'\n systems += [f'{self.net.name}{self.net.index}']\n return system_ass + f\"system {', '.join(systems)};\"\n\n def _generate_template(self):\n templates = [cl.template for cl in self.control_loops]\n templates += [self.net.template]\n return pyuppaal.NTA(templates=templates,\n declaration=self._generate_declarations(),\n system=self._generate_system())\n\n @cached_property\n def template(self):\n return self._generate_template()\n\n def to_xml(self):\n return self.template.to_xml()\n\n def export(self, file_name: str = None, export_type: str = 'txt'):\n \"\"\"\n Exports the timed automaton to a specified file_type. Default is plain text\n \"\"\"\n export_type = export_type.lower()\n if file_name is None:\n file_name = self.__class__.__name__\n elif export_type in ['uppaal', 'xml']:\n res = self.to_xml()\n if not file_name.endswith('.xml'):\n file_name += '.xml'\n with open(os.path.join(save_path, file_name), 'w') as f:\n f.write(res)\n elif export_type in ['pickle', 'bytes', 'byte_stream']:\n self._export_pickle(file_name)\n\n def _export_pickle(self, file_name: str):\n if not file_name.endswith('.pickle'):\n file_name += '.pickle'\n\n with open(os.path.join(save_path, file_name), 'wb') as f:\n pickle.dump(self, f, pickle.HIGHEST_PROTOCOL)\n\n def generate_strategy(self, parse_strategy=True, delete_files=True):\n\n # Throw warning if more than 2 control loops\n if len(self.control_loops) > 2:\n import warnings\n logging.warning('Using more than two control loops might take a lot of time and/or memory(!). ')\n warnings.warn('Using more than two control loops might take a lot of time and/or memory(!). ')\n\n idx = shortuuid.uuid()[:4]\n print(\"Exporting NTA to UPPAAL...\")\n # First export NTA to xml format\n self.export(f'NTA_{idx}', 'xml')\n\n strat_name = f'Strategy_{idx}'\n spath = os.path.join(strat_path, strat_name) + '.q'\n with open(spath, 'w') as f:\n f.write(f'strategy {strat_name} = control: A[] not ({self.net.name}{self.net.index}.Bad)')\n\n temp = os.path.join(save_path, f'NTA_{idx}.xml')\n arg_list = [VERIFYTA, '-u', '-s', '--generate-strategy', '2', '--print-strategies',\n strat_path, temp, f'{spath}']\n\n # arg_list = [VERIFYTA, '-h']\n\n # import resource\n # def limit_mem():\n # resource.setrlimit(resource.RLIMIT_AS, (4*1024**3, resource.RLIM_INFINITY))\n\n print('Generating Strategy using UPPAAL...')\n verify_ta = subprocess.Popen(arg_list, stdout=subprocess.PIPE)#, preexec_fn=limit_mem)\n try:\n # Run the query in UPPAAL to generate the strategy file\n verify_ta.wait()\n # result = verify_ta.stdout.decode('utf-8')\n except KeyboardInterrupt:\n verify_ta.terminate()\n sys.exit()\n except:\n print(verify_ta.communicate())\n sys.exit()\n\n logging.info(f'Strategy {idx} generated. Saved in {strat_path}')\n\n if delete_files:\n os.remove(os.path.join(save_path, f'NTA_{idx}.xml'))\n os.remove(spath)\n\n if parse_strategy:\n\n print(\"Parsing Strategy...\")\n # This regex string represents all states from which\n # an ack! signal is sent, i.e. when triggering must occur\n # Groups: (Region, CL1), (Region, CL2), ... , (clock conditions), (CL to trigger)\n REGEX_STRING = r'\\nState: \\( .*\\).*'\n for cl in self.control_loops:\n REGEX_STRING += rf'{cl.name}{cl.index}+\\.from_region=([0-9]+).*'\n\n REGEX_STRING += r'\\n.*[\\n]*When you are in \\((.*)\\).*[\\n]+([A-z0-9]+)\\.Trans.*, from_region := to_region.*}\\n'\n\n with open(os.path.join(strat_path, strat_name), 'r') as strat:\n raw_strat = strat.read()\n raw_condition_tuples = re.findall(REGEX_STRING, raw_strat)\n\n limit_dict = {}\n\n for condition in raw_condition_tuples:\n self._populate_limit_dict_from_condition_tuple(condition, limit_dict)\n\n import json\n with open(os.path.join(strat_path, strat_name + '.json'), 'w') as file:\n json.dump({str(k): v for (k,v) in limit_dict.items()}, file)\n\n print(f\"Saved parsed strategy to: {os.path.join(strat_path, strat_name + '.json')}\")\n self.scheduler = limit_dict\n if delete_files:\n os.remove(os.path.join(strat_path, strat_name))\n\n\n def _get_control_loops(self, lines):\n\n # Regex for finding CL names\n cl_regex = (\".* ([A-z0-9]+)\\.c==([A-z0-9]+)\\.c.*\")\n\n for line in lines:\n match_obj = re.match(cl_regex, line)\n if match_obj:\n return match_obj.groups()\n\n return []\n\n def _populate_limit_dict_from_condition_tuple(self, condition_tuple, limit_dict):\n\n # cl_regions = tuple(int(condition_tuple[i]) for i in range(0, len(self.closed_loops)))\n cl_regions = tuple(int(x.loc_dict_inv[int(condition_tuple[i])]) for (i,x) in enumerate(self.control_loops))\n\n cl_names = [f'{cl.name}{cl.index}' for cl in self.control_loops]\n if cl_regions not in limit_dict:\n limit_dict[cl_regions] = []\n\n to_trigger = cl_names.index(condition_tuple[-1])\n\n # main part to decipher\n invariant_string = condition_tuple[-2]\n\n # there could be multiple combinations for the same region\n invariants = invariant_string.replace('(', '').replace(')', '').split('||')\n\n A_eq = []\n b_eq = []\n A_leq = []\n b_leq = []\n\n # each invariant deals with precisely one upper and lower limit per CL\n for invariant in invariants:\n # handle cases one by one\n clock_conditions = invariant.split('&&')\n clock_conditions = [condition.strip()\n for condition in clock_conditions]\n\n # Store clock conditions in the form:\n # A_eq @ c == b_eq\n # A_leq @ c <= b_leq\n # Entry becomes: [(cl1l, cl1h), .., A_eq, b_eq, A_leq, b_leq, to_trigger]\n\n for condition in clock_conditions:\n # 1. Handle case when difference between clocks\n if '-' in condition:\n # the case in which loop1.c - loop2.c == x needn't be handled.\n # It is covered in other cases (where loop clocks have exact\n # valuation matches, e.g. loop1.c == 9)\n if '==' in condition:\n continue\n\n reg_exp_first_loop = '([A-z0-9]+)\\.c-([A-z0-9]+)\\.c.*[<|<=]([+-]?[0-9]+).*'\n match_obj = re.findall(reg_exp_first_loop, condition)\n cl1 = cl_names.index(match_obj[0][0])\n cl2 = cl_names.index(match_obj[0][1])\n\n # Fill matrices\n a_leq_new = [0 for i in range(0, self.ns)]\n a_leq_new[cl1] = 1\n a_leq_new[cl2] = -1\n A_leq.append(a_leq_new)# = np.append(A_leq, a_leq_new)\n b_leq.append([int(match_obj[0][2])/self.common_scale])# = np.append(b_leq,np.array([int(match_obj[0][2])]))\n continue\n\n # 2. Handle cases in which clock are equal\n eq_regex = '([A-z0-9]+)\\.c(<|<=|==)([A-z0-9]+)\\.c'\n match_obj = re.findall(eq_regex, condition)\n if match_obj != []:\n cl1 = cl_names.index(match_obj[0][0])\n cl2 = cl_names.index(match_obj[0][2])\n\n if match_obj[0][1] == '==':\n # Fill matrices\n a_eq_new = [0 for i in range(0, self.ns)]\n a_eq_new[cl1] = 1\n a_eq_new[cl2] = -1\n A_eq.append(a_eq_new)# = np.append(A_eq, a_eq_new)\n b_eq.append([0])# = np.append(b_eq,[np.array([0])])\n elif match_obj[0][1] == '<=':\n a_leq_new = [0 for i in range(0, self.ns)]\n a_leq_new[cl1] = 1\n a_leq_new[cl2] = -1\n A_leq.append(a_leq_new) # = np.append(A_eq, a_eq_new)\n b_leq.append([0]) # = np.append(b_eq,[np.array([0])])\n\n continue\n\n # 3. Handle cases from the individual clock limits\n for cl, name, reg in zip(self.control_loops, cl_names, cl_regions):\n if name in condition:\n # condition = condition.replace(name, '')\n # Check for lower limit\n regex_string_lower_limit = '([0-9]+)([<>]=?|==).*'\n match_obj = re.match(regex_string_lower_limit, condition)\n if match_obj:\n val = int(match_obj.group(1)) / self.common_scale\n oper = match_obj.group(2)\n if oper == \"==\":\n a_eq_new = [0 for i in range(0, self.ns)]\n idx = cl_names.index(name)\n a_eq_new[idx] = 1\n A_eq.append(a_eq_new)# = np.append(A_eq, a_eq_new)\n b_eq.append([val])# = np.append(b_eq, np.array([lims[0]]))\n else:\n if oper == \"<=\":\n lower_lim = val\n elif oper == \"<\":\n lower_lim = val - 0.00001\n\n a_leq_new = [0 for i in range(0, self.ns)]\n idx = cl_names.index(name)\n a_leq_new[idx] = -1\n A_leq.append(a_leq_new) # = np.append(A_leq, a_leq_new1)\n b_leq.append([-lower_lim]) # = np.append(b_leq, np.array([-lims[0]]))\n\n # Check for upper limit:\n regex_string_upper_limit = '.*([<>]=?|==)([0-9]+)'\n match_obj = re.match(regex_string_upper_limit, condition)\n if match_obj:\n oper = match_obj.group(1)\n val = int(match_obj.group(2)) / self.common_scale\n if oper == \"==\":\n a_eq_new = [0 for i in range(0, self.ns)]\n idx = cl_names.index(name)\n a_eq_new[idx] = 1\n A_eq.append(a_eq_new) # = np.append(A_eq, a_eq_new)\n b_eq.append([val]) # = np.append(b_eq, np.array([lims[0]]))\n else:\n if oper == \"<=\":\n upper_lim = val\n elif oper == \"<\":\n upper_lim = val - 0.00001\n\n a_leq_new = [0 for i in range(0, self.ns)]\n idx = cl_names.index(name)\n a_leq_new[idx] = 1\n A_leq.append(a_leq_new) # = np.append(A_leq, a_leq_new1)\n b_leq.append([upper_lim]) # = np.append(b_leq, np.array([-lims[0]]))\n\n expr = tuple((A_eq, b_eq, A_leq, b_leq, to_trigger))\n limit_dict[cl_regions].append(expr)\n\n def _extract_limits(self, condition, cl_low, cl_high):\n\n # NOTE: UPPAAL doesn't have operators of the kind \"> or >=\", but only \"< or <=\"\n\n lower_lim = max(0,cl_low)\n upper_lim = cl_high\n\n # First define the regex strings\n regex_string_lower_limit = '.*([0-9]+)([<>]=?|==).*'\n regex_string_upper_limit = '.*([<>]=?|==)([0-9]+).*'\n\n # Check for upper limit\n match_obj = re.match(regex_string_upper_limit, condition)\n if match_obj:\n oper = match_obj.group(1)\n val = int(match_obj.group(2))/self.common_scale\n\n if oper == \"==\":\n return (val, val)\n elif oper == \"<=\":\n upper_lim = val\n elif oper == \"<\":\n upper_lim = val - 0.001\n\n return (lower_lim, upper_lim)\n\n # Check for lower limit\n match_obj = re.match(regex_string_lower_limit, condition)\n if match_obj:\n val = int(match_obj.group(1))/self.common_scale\n oper = match_obj.group(2)\n\n if oper == \"==\":\n return (val, val)\n elif oper == \"<=\":\n lower_lim = val\n elif oper == \"<\":\n lower_lim = val + 0.001\n\n return (lower_lim, upper_lim)\n\n def simulate(self, Ts: float = 0.01, Tmax: float = 1, x0=None, use_scheduler=True):\n\n if any([type(cl) == abstr.TrafficModelNonlinearETC for cl in self.control_loops]):\n raise NotImplementedError\n\n # Check correct/enough initial conditions\n if x0 is None:\n x0 = [np.random.uniform(low=-4, high=4, size=(cl.abstraction.plant.nx,)) for cl in self.control_loops]\n else:\n if len(x0) != len(self.control_loops):\n print('Supply initial conditions for each control loop.')\n return\n\n for x0i, cl in zip(x0, self.control_loops):\n if len(x0i) != cl.abstraction.plant.nx:\n print(\n f'Initial condition dimension ({len(x0i)}) does not correspond to the expected ({cl.abstraction.plant.nx}).')\n return\n\n x0 = [np.array(x) for x in x0]\n\n # 3D Matrix storing the evolution of the continuous states over time.\n x = [[np.array(x0i)] for x0i in x0]\n xhat = [[np.array(x0i)] for x0i in x0]\n u_hist = [[] for i in range(0, self.ns)] # continuous inputs\n\n\n # Evolution of the traffic model regions over time\n regions = [[cl.abstraction.region_of_state(x0i)] for (x0i, cl) in zip(x0, self.control_loops)]\n\n s = [[str(cl.loc_dict['_'.join([str(i) for i in loc[0]])])] for (loc, cl) in zip(regions, self.control_loops)]\n ss = tuple(q[-1] for q in s)\n print(ss)\n clocks = [[0] for i in range(0, self.ns)]\n\n for i in range(0, self.ns):\n print(f'Controlloop {i} starts in region {regions[i][0]}')\n\n TriggerTimes = [[0] for i in range(0, self.ns)]\n CollisionTimes = {}\n\n N = int(Tmax / Ts) # Number of samples\n\n import scipy\n I = [scipy.integrate.quad_vec(lambda s: scipy.linalg.expm(cl.abstraction.plant.A * s), 0, Ts)[0] for cl in\n self.control_loops]\n\n\n for t in range(0, N):\n # Step 1: Update the continuous states\n utemp = [cl.abstraction.controller.K @ xn[-1] for (cl, xn) in zip(self.control_loops, xhat)]\n xn = [scipy.linalg.expm(cl.abstraction.plant.A * Ts) @ xi[-1] + integral @ cl.abstraction.plant.B @ ui\n for (cl, xi, ui, integral) in zip(self.control_loops, x, utemp, I)]\n\n for i in range(0, self.ns):\n x[i].append(xn[i])\n xhat[i].append(xhat[i][-1])\n u_hist[i].append(utemp[i])\n clocks[i].append(clocks[i][-1] + Ts)\n\n\n ## Step 2: Check triggering conditions\n # If a scheduler is defined use that\n if self.scheduler is not None and use_scheduler:\n to_trigger = -1\n ss = tuple(q[-1] for q in s)\n if ss not in self.scheduler:\n to_trigger = random.randint(0, self.ns-1)\n print(f'State {ss} not in scheduler => Choose to trigger {to_trigger+1}')\n else:\n cc = np.array([[c[-1]] for c in clocks])\n print(f'Current Clock values: {cc}')\n for (Aeq, beq, Aleq, bleq, trig) in self.scheduler[ss]:\n if Aeq != [] and any([i != 0 for i in (Aeq @ cc - beq)]):\n continue\n if Aleq != [] and any([i > 0 for i in Aleq @ cc - bleq]):\n continue\n\n to_trigger = trig\n print(f'Loop {trig+1} should trigger')\n break\n\n for i in range(0, self.ns):\n if to_trigger == i:\n reg = self.control_loops[i].abstraction.region_of_state(x[i][-1])\n si = str(cl.loc_dict['_'.join([str(i) for i in reg])])\n clocks[i][-1] = 0\n s[i].append(si)\n xhat[i][-1] = xn[i]\n regions[i].append(reg)\n TriggerTimes[i].append(t * Ts)\n\n else:\n # reg = self.control_loops[i].abstraction.region_of_state(x[i][-1])\n regions[i].append(regions[i][-1])\n s[i].append(s[i][-1])\n\n else:\n triggers = set()\n for i in range(0, self.ns):\n\n triggered = False\n if type(self.control_loops[i].abstraction) == abstr.TrafficModelLinearPETC:\n xx = np.block([x[i][-1], xhat[i][-1]])\n triggered = xx.T @ self.control_loops[i].abstraction.trigger.Qbar @ xx.T > 0 or \\\n (t * Ts - TriggerTimes[i][-1]) >= self.control_loops[i].tau_max\n else:\n xe = x[i][-1] - xhat[i][-1]\n xdict = {i:j for (i,j) in zip([x, xe], self.control_loops[i].abstraction.original_state)}\n print(xdict)\n sys.exit()\n\n if triggered:\n xhat[i][-1] = xn[i]\n TriggerTimes[i].append(t * Ts)\n triggers.add(i)\n\n reg = self.control_loops[i].abstraction.region_of_state(x[i][-1])\n regions[i].append(reg)\n\n\n if len(triggers) > 1:\n CollisionTimes[t * Ts] = triggers\n\n import matplotlib.pyplot as plt\n\n dur = np.arange(0, Ts * N, Ts)\n for i in range(0, self.ns):\n plt.plot(dur, x[i][0:len(dur)], '--')\n plt.gca().set_prop_cycle(None)\n plt.plot(dur, xhat[i][0:len(dur)])\n plt.title(f'Controlloop {i + 1}: $x(t)$ and $x_e(t)$.')\n plt.show()\n\n for i in range(0, self.ns):\n plt.plot(dur, u_hist[i][0:len(dur)])\n plt.title(f'Controlloop {i + 1}: $u(t)$.')\n plt.show()\n\n for i in range(0, self.ns):\n plt.plot(TriggerTimes[i], i * np.ones(len(TriggerTimes[i])), 'x')\n\n for t, ii in CollisionTimes.items():\n for i in ii:\n plt.plot(t, i, 'dk')\n\n plt.title('Trigger times')\n plt.yticks(range(0, self.ns), [f'Controlloop {i}' for i in range(1, self.ns + 1)])\n plt.show()\n\n for i in range(0, self.ns):\n plt.plot(dur, regions[i][0:len(dur)])\n\n plt.title('Traffic Model Regions')\n plt.legend([f'Controlloop {i}' for i in range(1, self.ns + 1)], loc='upper left')\n plt.show()","repo_name":"sync-lab/ETCetera","sub_path":"ETCetera/Scheduling/NTGA/nta.py","file_name":"nta.py","file_ext":"py","file_size_in_byte":21634,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"25144868287","text":"import scrapy\nfrom scrapy_splash import SplashRequest\nfrom scrapy.exceptions import CloseSpider\n\nimport re\n\nfrom crawlTiki.items import ProductItem\nfrom crawlTiki.helper.TikiHelperHandleDataProduct import TikiHelperHandleDataProduct\nfrom crawlTiki.dto.ObjectDTO import ImageDTO\nfrom crawlTiki.constant.constant import API_DETAIL_URL\nfrom crawlTiki.helper.TikiHelperUtil import TikiHelperUtil\n\n\nclass CrawlCategoriesSpider(scrapy.Spider):\n name = 'dataTiki'\n allowed_domains = ['tiki.vn']\n start_urls = ['https://tiki.vn/']\n\n def __init__(self, *args, **kwargs):\n super(CrawlCategoriesSpider, self).__init__(*args, **kwargs)\n self.limitProduct: int = int(kwargs.get('limitProduct', None))\n self.pathName: str = kwargs.get('pathName', None)\n self.typeImage: str = kwargs.get('typeImage', None)\n self.keyWord: str = kwargs.get('keyWord', None)\n self.counter: int = 1\n self.tikiHelperUtil = TikiHelperUtil()\n\n render_script = \"\"\"\n function main(splash)\n local url = splash.args.url\n assert(splash:go(url))\n assert(splash:wait(5))\n\n return {\n html = splash:html(),\n url = splash:url(),\n }\n end\n \"\"\"\n\n def start_requests(self):\n\n listUrl: list[str] = self.tikiHelperUtil.handle_url(\n self.keyWord, self.limitProduct)\n\n for url in listUrl:\n yield SplashRequest(\n url,\n self.parse_list_product,\n endpoint='render.html',\n args={\n 'wait': 5,\n 'lua_source': self.render_script,\n },\n )\n\n def parse_list_product(self, response):\n if response.status == 404:\n raise CloseSpider('Receive 404 response')\n\n listProducts: list[str] = response.css(\".product-item\")\n if (listProducts == []):\n raise CloseSpider('Error selector list products')\n\n for product in listProducts:\n if self.counter > self.limitProduct:\n break\n\n self.counter += 1\n productLink = product.css(\"a::attr(href)\").extract()\n temp = re.search('\\d{6,10}', str(productLink))\n productID = temp.group()\n\n url = f'{API_DETAIL_URL}{productID}'\n yield response.follow(url=url, callback=self.parse_product_info)\n\n def parse_product_info(self, response):\n if response.status == 404:\n raise CloseSpider('Receive 404 response')\n try:\n item = ProductItem()\n modelOptionsImage: ImageDTO = ImageDTO(\n self.pathName, self.typeImage)\n tikiHelperHandleDataProduct = TikiHelperHandleDataProduct(\n response.text, modelOptionsImage)\n dataProductInfo = tikiHelperHandleDataProduct.handle_Product()\n \n item[\"name\"] = dataProductInfo.name\n item[\"price\"] = dataProductInfo.price\n item[\"originalPrice\"] = dataProductInfo.originalPrice\n item[\"ratingAverageProduct\"] = dataProductInfo.ratingAverageProduct\n item[\"thumbnailUrl\"] = dataProductInfo.thumbnailUrl\n item[\"description\"] = dataProductInfo.description\n item[\"imagesProduct\"] = dataProductInfo.imagesProduct\n item[\"optionsType\"] = dataProductInfo.optionsType\n item[\"optionsProduct\"] = dataProductInfo.optionsProduct\n yield item\n except BaseException as e:\n raise CloseSpider(e)\n","repo_name":"ngocphuphamm/CRAWL-DATA-TIKI","sub_path":"crawlTiki/spiders/crawl.py","file_name":"crawl.py","file_ext":"py","file_size_in_byte":3583,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"4990422574","text":"from django.views import View\nfrom django.shortcuts import render, redirect\nfrom login_signup.models.news import News\nfrom django.contrib.auth.decorators import login_required, permission_required\nfrom django.utils.decorators import method_decorator\n\ndecorators = [\n permission_required('login_signup.can_use_medical_stuff', raise_exception=True),\n login_required(login_url='login')\n]\n\n\n@method_decorator(decorators, name='dispatch')\nclass DeleteNews(View):\n template_name = 'home/news/deleteNews.html'\n\n def get(self, request, *args, **kwargs):\n return render(request, self.template_name)\n\n def post(self, request, *args, **kwargs):\n news = News.objects.get(pk=kwargs['pk'])\n news.delete()\n return redirect('home:newsList')\n","repo_name":"Redshark61/mon-carnet-de-sante","sub_path":"health_book/home/views/news/deleteNews.py","file_name":"deleteNews.py","file_ext":"py","file_size_in_byte":768,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25111945467","text":"import random\r\nimport sys\r\n\r\n\"\"\"\r\n@author: Amit Dhandha\r\n\"\"\"\r\n\"\"\"\r\n DocString:\r\n\r\n A) Introduction:\r\n 'QuizTime has 4 categories to choose from.'\r\n 'Each category has 3 randomly selected questions to choose from. The question are of varying difficulty levels - Easy, Medium & Hard.'\r\n 'Easy questions are scored less and Hard questions are worth more. You need to score 1000 or more to win a trip to Bahamas for 2 people.'\r\n\r\n B) Known Bugs and/or Errors:\r\n Questions repeat within the same category\r\n\"\"\"\r\n\r\ndef initGame():\r\n printline()\r\n print(\"\"\"Welcome to QuizTime! QuizTime has 4 categories to choose from.'\r\n 'Each category has 3 randomly selected questions to choose from. The question are of varying difficulty levels - Easy, Medium & Hard.'\r\n 'Easy questions are scored less and Hard questions are worth more. You need to score 1000 or more to win a trip to Bahamas for 2 people.'\"\"\")\r\n printline()\r\n input('Press Enter to begin the Quiz!\\n')\r\n initInstructions()\r\n\r\ndef initInstructions():\r\n printline()\r\n pointsAcquired = gameStages()\r\n print(\"-----------------------\")\r\n print(f\" PointsAquired={pointsAcquired}\")\r\n print(\"-----------------------\\n\")\r\n if (pointsAcquired >= 1000):\r\n print(\"Congratulations!! You have Won a trip Bahamas!!!\")\r\n else:\r\n print(f\"Insufficient points! Please try again!\")\r\n\r\ndef printCategory(categories):\r\n strCat = 'Choose from the following categories\\n'\r\n for cat in categories:\r\n if cat == 1:\r\n strCat += '\\tCategory 1 : History\\n'\r\n elif cat == 2:\r\n strCat += '\\tCategory 2 : General Knowledge\\n'\r\n elif cat == 3:\r\n strCat += '\\tCategory 3 : Sports\\n'\r\n elif cat == 4:\r\n strCat += '\\tCategory 4 : Movies\\n'\r\n print(strCat);\r\n\r\ndef gameStages():\r\n pointsAcquired = 0\r\n easyPoints, mediumPoints, hardPoints = 50, 75, 150\r\n historyQuestionsList = [['Who was the first president of the United States of America?\\n\\tA. Thomas Jefferson\\n\\tB. George Washington\\n\\tC. Abraham Lincoln\\n\\tD. John Quincy Adams',\r\n 'B', easyPoints],\r\n ['When did French Revolution began?\\n\\tA. 1785\\n\\tB. 1789\\n\\tC. 1795\\n\\tD. 1799', 'B',\r\n mediumPoints], ['When did World War 1 finish?\\n\\tA. 1917\\n\\tB. 1919 \\n\\tC. 1918\\n\\tD. 1920', 'B',\r\n easyPoints],\r\n ['Who was the president of the United States duing Civil War?\\n\\tA. Adam Smith\\n\\tB. Abraham Lincoln \\n\\tC. Andrew Johnson\\n\\tD. James Buchanan',\r\n 'B', mediumPoints],\r\n ['When did India became Independent from British Rule?\\n\\tA. 1934\\n\\tB. 1945\\n\\tC. 1947 \\n\\tD. 1948',\r\n 'C', mediumPoints],\r\n ['Who wrote the book Das Capital?\\n\\tA. Vladimie Lenin\\n\\tB. Adam Smith\\n\\tC. Karl Marx \\n\\tD. Max Mueller',\r\n 'C', hardPoints],\r\n ['When was Germany Unified by Bismarck?\\n\\tA. 1874\\n\\tB. 1871 \\n\\tC. 1878\\n\\tD. 1875',\r\n 'B', hardPoints]]\r\n\r\n moviesQuestionsList = [['Which movie has won the maximum number of Oscars?\\n\\tA. Avataar\\n\\tB. Schindlers List\\n\\tC. Ben-hur \\n\\tD. Jurassic Park',\r\n 'C', hardPoints],\r\n ['Who amoongst the following has the higher number of Oscar Wins?\\n\\tA.Jack Nicholson\\n\\tB.Meryl Streep\\n\\tC.Ingrid Bergmann\\n\\tD.Katharine Hepburn ',\r\n 'D', hardPoints],\r\n ['Who is the director of the movie Jurassic Park?\\n\\tA. Richard Attenborough\\n\\tB. Steven Spielberg \\n\\tC. James Cameron\\n\\tD. Manoj Night Shymalan',\r\n 'B', easyPoints],\r\n ['Where is Hollywood Located?\\n\\tA. New York\\n\\tB. Los Angeles \\n\\tC. Florida\\n\\tD. Washington',\r\n 'B', easyPoints],\r\n ['In which Country the Film Industry is known as Bollywood\\n\\tA. Iran\\n\\tB. Bangladesh\\n\\tC. Sri Lanka\\n\\tD. India ',\r\n 'D', mediumPoints],\r\n ['Who directed the Avataar?\\n\\tA. Michael Caine\\n\\tB. Steven Spielberg\\n\\tC. James Cameron \\n\\tD. Jeff Goldblum',\r\n 'C', mediumPoints]]\r\n\r\n genknowQuestionsList = [['Who is the current CEO of Google\\n\\tA. Sundar Pichai\\n\\tB. Sheryl Sandberg\\n\\tC. Satya Nadella\\n\\tD. Sundar Rajan',\r\n 'A', mediumPoints],\r\n ['Who is the current secretary general of the United Nations?\\n\\tA. Ban Ki Moon\\n\\tB. António Guterres \\n\\tC Ki Ban Mooon\\n\\tD. Kofi Annan',\r\n 'B', hardPoints],\r\n ['In which state is Harvard Business School located?\\n\\tA. MASSACHUSSETS \\n\\tB. CONNECTICUT\\n\\tC. BOSTON\\n\\tD. NEW JERSEY',\r\n 'A', mediumPoints],\r\n ['On which day is American Independence Day Celebrated?\\n\\tA. 4th Aug\\n\\tB. 3rd July\\n\\tC. 4th July \\n\\tD. 3rd Sep',\r\n 'C', easyPoints],\r\n ['Which is most spoken language in the world\\n\\tA. Mandarin \\n\\tC. Arabic\\n\\tB. English\\n\\tD. Spanish',\r\n 'A', easyPoints],\r\n ['Who invented Python language?\\n\\tA. Guido van Rossum \\n\\tB. Dennis Ritchie\\n\\tC. Bill Gates\\n\\tD. Kai Fu',\r\n 'A', easyPoints]]\r\n\r\n sportsQuestionsList = [['Who is the current 100 meters world record holder?\\n\\tA. Usain Bolt\\n\\tB. Christian Coleman\\n\\tC. Tyson Gay\\n\\tD. Asafa Powell',\r\n 'A', easyPoints],\r\n ['Who won the Women Wimbeldon Championship in 2019\\n\\tA. Venus Williams\\n\\tB. Serena William\\n\\tC. Simona Halep\\n\\tD. Ashleigh Barty',\r\n 'C', mediumPoints],\r\n ['Which Country did France defeat in the FIFA World cup semifinal in 2018?\\n\\tA. England\\n\\tB. Belgium\\n\\tC. Croatia\\n\\tD. Brazil',\r\n 'B', hardPoints],\r\n ['Where will the next FIFA world cup in 2022 be held?\\n\\tA. India\\n\\tB. UAE\\n\\tC. Saudi Arabia\\n\\tD. Qatar',\r\n 'D', mediumPoints],\r\n ['Where will the Summer Olympics be held in 2020?\\n\\tA. USA\\n\\tB. Germany\\n\\tC. England\\n\\tD. Japan',\r\n 'D', easyPoints]]\r\n\r\n catsUnSelected = [\"1\", \"2\", \"3\", \"4\"]\r\n countOfCategories, invalidChoiceCount = 1, 0;\r\n\r\n while countOfCategories <= 4:\r\n printCategory(catsUnSelected)\r\n categoryChoice = input(f'Please choose Category {catsUnSelected} > ')\r\n if categoryChoice in catsUnSelected:\r\n catsUnSelected.remove(categoryChoice)\r\n elif invalidChoiceCount == 2:\r\n print(\"\\nYou exceeded maximum amount of wrong inputs, Game Over!!!\")\r\n input(\"Press Enter to exit...\")\r\n sys.exit()\r\n else:\r\n invalidChoiceCount +=1;\r\n print(f\"\\n> Wrong choice, Please select the valid category again! {3 - invalidChoiceCount} try left\\n\")\r\n continue\r\n\r\n countOfCategories += 1\r\n\r\n if categoryChoice == \"1\":\r\n print('\\nYou have chosen General History Category\\n')\r\n questionsCount = 0\r\n while questionsCount < 3:\r\n questionItem = random.choice(historyQuestionsList)\r\n question, answer, ansPoints = questionItem[0], questionItem[1], questionItem[2]\r\n print(question)\r\n userAns = input('Please enter your choice>')\r\n if (userAns.lower() == answer.lower()):\r\n pointsAcquired = pointsAcquired + ansPoints;\r\n print(f\"\\nCorrect Answer!!!\\nTotal Points: {pointsAcquired}\")\r\n questionsCount = questionsCount + 1\r\n\r\n elif categoryChoice == \"2\":\r\n print('\\nYou have chosen General Knowledge Category\\n')\r\n questionsCount = 0\r\n while questionsCount < 3:\r\n questionItem = random.choice(genknowQuestionsList)\r\n question, answer, ansPoints = questionItem[0], questionItem[1], questionItem[2]\r\n print(question)\r\n userAns = input('Please enter your choice>')\r\n if (userAns.lower() == answer.lower()):\r\n pointsAcquired = pointsAcquired + ansPoints;\r\n print(f\"\\nCorrect Answer!!!\\nTotal Points: {pointsAcquired}\")\r\n questionsCount = questionsCount + 1\r\n\r\n elif categoryChoice == \"3\":\r\n print('\\nYou have chosen Sports Category\\n')\r\n questionsCount = 0\r\n while questionsCount < 3:\r\n questionItem = random.choice(sportsQuestionsList)\r\n question, answer, ansPoints = questionItem[0], questionItem[1], questionItem[2]\r\n print(question)\r\n userAns = input('Please enter your choice>')\r\n if (userAns.lower() == answer.lower()):\r\n pointsAcquired = pointsAcquired + ansPoints;\r\n print(f\"\\nCorrect Answer!!!\\nTotal Points: {pointsAcquired}\")\r\n\r\n questionsCount += 1\r\n\r\n elif categoryChoice == \"4\":\r\n print('\\nYou have chosen Movies Category\\n')\r\n questionsCount = 0\r\n while questionsCount < 3:\r\n questionItem = random.choice(moviesQuestionsList)\r\n question, answer, ansPoints = questionItem[0], questionItem[1], questionItem[2]\r\n print(question)\r\n userAns = input('Please enter your choice>')\r\n if (userAns.lower() == answer.lower()):\r\n pointsAcquired = pointsAcquired + ansPoints;\r\n print(f\"\\nCorrect Answer!!!\\nTotal Points: {pointsAcquired}\")\r\n\r\n questionsCount += 1\r\n\r\n if countOfCategories < 4:\r\n print(\"\\nLet's go to next round...\\n\")\r\n\r\n return pointsAcquired;\r\n\r\ndef printline():\r\n print('*' * 100 + \"\\n\")\r\n\r\ninitGame()\r\n","repo_name":"amitdhandha1996/python-programs","sub_path":"Quiz Game.py","file_name":"Quiz Game.py","file_ext":"py","file_size_in_byte":10369,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26615822155","text":"import sys\nimport socket\nfrom PyQt5 import QtCore, QtGui, QtWidgets\n\n\nclass Ui_MainWindow(object):\n def setupUi(self, MainWindow):\n MainWindow.setObjectName(\"MainWindow\")\n MainWindow.resize(612, 277)\n self.centralwidget = QtWidgets.QWidget(MainWindow)\n self.centralwidget.setObjectName(\"centralwidget\")\n self.frame = QtWidgets.QFrame(self.centralwidget)\n self.frame.setGeometry(QtCore.QRect(0, 0, 691, 431))\n self.frame.setStyleSheet(\"background-color: rgb(216, 216, 216);\")\n self.frame.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame.setObjectName(\"frame\")\n self.grafic_button = QtWidgets.QPushButton(self.frame)\n self.grafic_button.setGeometry(QtCore.QRect(10, 140, 75, 23))\n self.grafic_button.setObjectName(\"grafic_button\")\n self.carregar_button = QtWidgets.QPushButton(self.frame)\n self.carregar_button.setGeometry(QtCore.QRect(10, 80, 95, 23))\n self.carregar_button.setObjectName(\"carregar_button\")\n self.gerarIP_button = QtWidgets.QPushButton(self.frame)\n self.gerarIP_button.setGeometry(QtCore.QRect(10, 20, 75, 23))\n self.gerarIP_button.setObjectName(\"gerarIP_button\")\n self.gerarIP_button.clicked.connect(self.gerarIP)\n self.ip_line = QtWidgets.QLineEdit(self.frame)\n self.ip_line.setGeometry(QtCore.QRect(110, 20, 151, 20))\n self.ip_line.setObjectName(\"ip_line\")\n MainWindow.setCentralWidget(self.centralwidget)\n\n self.retranslateUi(MainWindow)\n QtCore.QMetaObject.connectSlotsByName(MainWindow)\n\n self.gerarIP_button.clicked.connect(self.gerarIP)\n\n def retranslateUi(self, MainWindow):\n _translate = QtCore.QCoreApplication.translate\n MainWindow.setWindowTitle(_translate(\"MainWindow\", \"MainWindow\"))\n self.grafic_button.setText(_translate(\"MainWindow\", \"Plotar\"))\n self.carregar_button.setText(_translate(\"MainWindow\", \"Carregar dados\"))\n self.gerarIP_button.setText(_translate(\"MainWindow\", \"Gerar IP\"))\n\n def gerarIP(self):\n ip = socket.gethostbyname(socket.gethostname())\n self.ip_line.setText(ip)\n\n\nif __name__ == \"__main__\":\n app = QtWidgets.QApplication(sys.argv)\n MainWindow = QtWidgets.QMainWindow()\n ui = Ui_MainWindow()\n ui.setupUi(MainWindow)\n MainWindow.show()\n sys.exit(app.exec_())\n","repo_name":"yohanngusso/Projeto_dados_NRZ_RZ","sub_path":"outros arquivos/funções separadas/gerar_ip.py","file_name":"gerar_ip.py","file_ext":"py","file_size_in_byte":2434,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"36318369794","text":"largura = int(input(\"Digite a largura: \"))\r\naltura = int(input(\"Digite a altura: \"))\r\ncont = 1\r\naltura_aux = altura\r\n\r\nwhile altura_aux > 0: \r\n i=largura\r\n while i>0:\r\n if cont==1 or cont ==altura:\r\n print(\"#\", end=\"\")\r\n elif i==1 or i==largura:\r\n print(\"#\", end=\"\")\r\n else:\r\n print(\" \",end=\"\")\r\n i=i-1\r\n altura_aux = altura_aux - 1\r\n cont= cont + 1\r\n print(\"\")\r\n","repo_name":"fgordiano/Python","sub_path":"exerc2_Retangulo2.py","file_name":"exerc2_Retangulo2.py","file_ext":"py","file_size_in_byte":441,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"27291243494","text":"# -*- coding: utf-8 -*-\nfrom urllib import parse, response\nimport json\n\n\nimport scrapy\nfrom scrapy import Request\nimport re\nfrom CnblogsSpider.items import CnblogsspiderItem\nfrom CnblogsSpider.util import common\n\n\nclass CnblogsSpider(scrapy.Spider):\n name = 'cnblogs'\n allowed_domains = ['www.cnblogs.com/news/']\n start_urls = ['http://www.cnblogs.com/news/']\n\n def parse(self, response):\n \"\"\"\n 1.获取列表中的url且将之交给scrapy下载,然后交由detail做解析\n 2.获取下一页的列表数据,将下一页的数据交给parse处理,也就是做1做的事情。\n \"\"\"\n post_nodes = response.css('#post_list .post_item') # [:1]\n for post_node in post_nodes:\n image_url = post_node.css('.post_item_summary a img::attr(src)').extract_first(\"\")\n post_url = post_node.css('h3 a::attr(href)').extract_first(\"\")\n yield Request(url=parse.urljoin(response.url, post_url), meta={\"front_image_url\": image_url},\n callback=self.parse_detail, dont_filter=True)\n\n # 提取出下一页并交给scrapy去做处理\n next_text = response.css('div.pager a:last-child::text').extract_first(\"\")\n if next_text == \"Next >\":\n next_url = response.css('div.pager a:last-child::attr(href)').extract_first(\"\")\n yield Request(url=parse.urljoin(response.url, next_url))\n\n def parse_detail(self, response):\n \"\"\"\n 获取详情信息的函数\n \"\"\"\n article_item = CnblogsspiderItem()\n title = response.css('#news_title a::text').extract_first(\"\")\n publish_time = response.css('#news_info .time::text').extract_first(\"\")\n match_re = re.match(\".*?(\\d+.*)\", publish_time)\n if match_re:\n publish_time = match_re.group(1)\n content = response.css('#news_content').extract()[0]\n tag_list = response.css('.news_tags a::text').extract()\n tags = \",\".join(tag_list)\n\n article_item[\"title\"] = title\n article_item[\"publish_time\"] = publish_time\n article_item[\"content\"] = content\n article_item[\"tags\"] = tags\n article_item[\"url\"] = response.url\n if response.meta.get(\"front_image_url\", \"\"):\n article_item[\"front_img_url\"] = [response.meta.get(\"front_image_url\", \"\")]\n else:\n article_item[\"front_img_url\"] = []\n\n match_re=re.match(\".*?(\\d+)\", response.url)\n if match_re:\n post_id = match_re.group(1)\n\n yield Request(url=parse.urljoin(response.url, \"/NewsAjax/GetAjaxNewsInfo?contentId={}\".format(post_id)),\n meta={\"article_item\": article_item}, callback=self.parse_ajax_data, dont_filter=True)\n\n # 此处代码换成异步的\n # html = requests.get(parse.urljoin(response.url, \"/NewsAjax/GetAjaxNewsInfo?contentId={}\".format(post_id)))\n # j_data = json.load(html.text)\n # like_nums =j_data[\"DiggCount\"] # 点赞数\n # view_nums = j_data[\"TotalView\"] # 阅读数\n # comment_nums = j_data[\"CommentCount\"] # 评论数\n pass\n\n def parse_ajax_data(self, response):\n \"\"\"\n 详情里有几个数据是ajax请求得到的,不是服务器里的数据,这块也优化成异步取数据\n \"\"\"\n j_data = json.loads(response.text)\n like_nums =j_data[\"DiggCount\"] # 点赞数\n view_nums = j_data[\"TotalView\"] # 阅读数\n comment_nums = j_data[\"CommentCount\"] # 评论数\n\n article_item = response.meta.get(\"article_item\")\n article_item[\"like_nums\"] = like_nums\n article_item[\"view_nums\"] = view_nums\n article_item[\"comment_nums\"] = comment_nums\n article_item[\"url_id\"] = common.get_md5(article_item[\"url\"])\n\n yield article_item\n\n","repo_name":"anmutu/CnblogsSpider","sub_path":"CnblogsSpider/spiders/cnblogs.py","file_name":"cnblogs.py","file_ext":"py","file_size_in_byte":3842,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"21336097288","text":"# Implementing a Stack.\n# Last in First out (LIFO)\n\nclass Node:\n def __init__(self,data):\n self.data = data\n self.next = next\n\nclass Stack:\n def __init__(self):\n self.top = None\n \n def pop(self):\n val = None\n if self.top is not None:\n val = self.top.data\n self.top = self.top.next\n return val \n \n def push(self,value):\n newval = Node(value)\n newval.next = self.top\n self.top = newval\n \n\n def peek(self):\n return self.top.data\n\n def isEmpty(self):\n return self.top == None\n\n\n\ns = Stack()\nprint(\"Is Stack Empty? {}\".format(s.isEmpty()))\nprint(\"Pushing Elements in Stack\")\ns.push(10)\ns.push(21)\ns.push(14)\ns.push(63)\ns.push(56)\n\nprint(\"Is Stack Empty? {}\".format(s.isEmpty()))\nprint(\"Checking top of Stack: {}\".format(s.peek()))\nprint(\"Popping the top of the stack value: {}\".format(s.pop()))\nprint(\"Again Checking top of Stack: {}\".format(s.peek()))\n\nprint(\"----------------------------------\")\n# Implementing a Queue\n# First In First Out\n\nclass Queue:\n def __init__(self):\n self.first = None\n self.last = None\n \n def add(self,value):\n newnode = Node(value)\n if self.last is not None:\n self.last.next = newnode\n self.last = newnode\n if self.first is None:\n self.first = self.last\n\n def remove(self):\n dat = None\n if self.first is None:\n self.last = None\n return dat\n dat = self.first.data\n self.first = self.first.next\n return dat\n\n def peek(self):\n if self.first is not None:\n return self.first.data\n\n def isEmpty(self):\n return self.first == None\n\nq = Queue()\nprint(\"Is Queue Empty? {}\".format(q.isEmpty()))\nprint(\"Adding in Queue\")\nq.add(10)\nq.add(20)\nq.add(30)\nq.add(40)\nq.add(50)\nprint(\"Is Queue Empty? {}\".format(q.isEmpty()))\nprint(\"Check the first Element: {}\".format(q.peek()))\nprint(\"Remove the first Element: {}\".format(q.remove()))\nprint(\"Again, Checking the first Element: {}\".format(q.peek()))\nprint(\"Remove the first Element: {}\".format(q.remove()))\nprint(\"Again, Checking the first Element: {}\".format(q.peek()))","repo_name":"raj-shah14/Ctci","sub_path":"StacknQueue/stack.py","file_name":"stack.py","file_ext":"py","file_size_in_byte":2221,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"73002875104","text":"import io\nimport unittest\nfrom pathlib import Path\n\nimport utils\n\n\nclass MyTestCase(unittest.TestCase):\n def test_xml_equals_iet(self):\n self.maxDiff = None\n files = Path('.').glob('./test/*.iet')\n for file in files:\n filepath = str(file.absolute())\n with self.subTest(file=filepath):\n r = file.open(mode='r', encoding='utf-8-sig')\n xml = io.StringIO()\n xml.writelines(\n utils.xmlize(r, filepath=filepath)\n )\n xml.seek(0)\n\n r_iet = io.StringIO()\n\n r_iet.writelines(\n utils.ietlize(xml)\n )\n r_iet.seek(0)\n r.seek(0)\n\n self.assertEqual(r.read(), r_iet.read())\n r.close()\n xml.close()\n r_iet.close()\n\n def test_xml_string_equals_iet_string(self):\n self.maxDiff = None\n files = Path('.').glob('./test/*.iet')\n for file in files:\n filepath = str(file.absolute())\n with self.subTest(file=filepath):\n r = file.open(mode='r', encoding='utf-8-sig')\n restored = ''.join(\n utils.ietlizes(\n utils.xmlizes(r.readlines())\n )\n )\n r.seek(0)\n orig = r.read()\n\n self.assertEqual(orig, restored)\n r.close()\n\n\nif __name__ == '__main__':\n unittest.main()\n","repo_name":"enzanumo/ietInterpreter","sub_path":"ietInterpreter/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":1543,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"5563085862","text":"from datetime import date\n\nfrom django.db import models\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.http import HttpResponse\nfrom django import forms\n\nfrom wagtail.wagtailcore.models import Page, Orderable\nfrom wagtail.wagtailcore.fields import RichTextField, StreamField\nfrom wagtail.wagtailadmin.edit_handlers import FieldPanel, MultiFieldPanel, \\\n InlinePanel, PageChooserPanel, StreamFieldPanel\n\nfrom wagtail.wagtailcore.fields import RichTextField\nfrom wagtail.wagtailadmin.edit_handlers import FieldPanel\nfrom wagtail.wagtailimages.edit_handlers import ImageChooserPanel\nfrom wagtail.wagtaildocs.edit_handlers import DocumentChooserPanel\nfrom wagtail.wagtailsnippets.models import register_snippet\nfrom wagtail.wagtailforms.models import AbstractEmailForm, AbstractFormField\nfrom wagtail.wagtailsearch import index\n\nfrom wagtail.wagtailcore.blocks import TextBlock, StructBlock, StreamBlock, FieldBlock, CharBlock, RichTextBlock, RawHTMLBlock\nfrom wagtail.wagtailimages.blocks import ImageChooserBlock\nfrom wagtail.wagtaildocs.blocks import DocumentChooserBlock\n\n\n\nfrom modelcluster.fields import ParentalKey\nfrom taggit.models import TaggedItemBase\n\n\n\nclass ImageFormatChoiceBlock(FieldBlock):\n field = forms.ChoiceField(choices=(\n ('left', 'Wrap left'), ('right', 'Wrap right'), ('mid', 'Mid width'), ('full', 'Full width'),\n ))\n\n\nclass ImageBlock(StructBlock):\n image = ImageChooserBlock()\n caption = RichTextBlock()\n alignment = ImageFormatChoiceBlock()\n\n\n# A couple of abstract classes that contain commonly used fields\n\nclass LinkFields(models.Model):\n link_external = models.URLField(\"External link\", blank=True)\n link_page = models.ForeignKey(\n 'wagtailcore.Page',\n null=True,\n blank=True,\n related_name='+'\n )\n link_document = models.ForeignKey(\n 'wagtaildocs.Document',\n null=True,\n blank=True,\n related_name='+'\n )\n\n @property\n def link(self):\n if self.link_page:\n return self.link_page.url\n elif self.link_document:\n return self.link_document.url\n else:\n return self.link_external\n\n panels = [\n FieldPanel('link_external'),\n PageChooserPanel('link_page'),\n DocumentChooserPanel('link_document'),\n ]\n\n class Meta:\n abstract = True\n\n\n# Carousel items\n\nclass CarouselItem(LinkFields):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n product_image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n embed_url = models.URLField(\"Embed URL\", blank=True)\n caption = models.CharField(max_length=255, blank=True)\n description = models.CharField(max_length=255, blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n ImageChooserPanel('product_image'),\n FieldPanel('embed_url'),\n FieldPanel('caption'),\n FieldPanel('description'),\n MultiFieldPanel(LinkFields.panels, \"Link\"),\n ]\n\n class Meta:\n abstract = True\n\n\n# Video items\n\nclass VideoItem(LinkFields):\n embed_url = models.URLField(\"Embed URL\", blank=True)\n caption = models.CharField(max_length=255, blank=True)\n description = models.CharField(max_length=255, blank=True)\n\n panels = [\n FieldPanel('embed_url'),\n FieldPanel('caption'),\n FieldPanel('description'),\n MultiFieldPanel(LinkFields.panels, \"Link\"),\n ]\n\n class Meta:\n abstract = True\n\n\n# Related links\n\nclass RelatedLink(LinkFields):\n title = models.CharField(max_length=255, help_text=\"Link title\")\n\n panels = [\n FieldPanel('title'),\n MultiFieldPanel(LinkFields.panels, \"Link\"),\n ]\n\n class Meta:\n abstract = True\n\n\n# Product item\n\nclass ProductItem(models.Model):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n caption = models.CharField(max_length=255, blank=True)\n description = models.CharField(max_length=255, blank=True)\n description = RichTextField()\n video = models.URLField(\"Embed video URL\", blank=True)\n price = models.CharField(max_length=255)\n sale = models.CharField(max_length=255, blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n FieldPanel('caption'),\n FieldPanel('description'),\n FieldPanel('video'),\n FieldPanel('price'),\n FieldPanel('sale'),\n ]\n\n def __unicode__(self):\n return self.image\n\n\n# Advantage item\n\nclass AdvantageItem(models.Model):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n caption = models.CharField(max_length=255, blank=True)\n description = models.CharField(max_length=255, blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n FieldPanel('caption'),\n FieldPanel('description'),\n ]\n\n def __unicode__(self):\n return self.image\n\n\n# Testimonials item\n\nclass TestimonialItem(models.Model):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n caption = models.CharField(max_length=255, blank=True)\n description = RichTextField()\n fb = models.URLField(\"Embed video URL\", blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n FieldPanel('caption'),\n FieldPanel('description'),\n FieldPanel('fb'),\n ]\n\n def __unicode__(self):\n return self.image\n\n\n# Service items\n\nclass ServicelItem(LinkFields):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n\n caption = models.CharField(max_length=255, blank=True)\n description = RichTextField()\n icon = models.CharField(max_length=255, blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n FieldPanel('caption'),\n FieldPanel('description'),\n FieldPanel('icon'),\n ]\n\n class Meta:\n abstract = True\n\n\n# Recipes items\n\nclass RecipeItem(models.Model):\n image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n caption = models.CharField(max_length=255, blank=True)\n description = RichTextField()\n ingredient = RichTextField(blank=True)\n video = models.URLField(\"Embed video URL\", blank=True)\n\n panels = [\n ImageChooserPanel('image'),\n FieldPanel('caption'),\n FieldPanel('description'),\n FieldPanel('ingredient'),\n FieldPanel('video'),\n ]\n\n def __unicode__(self):\n return self.image\n\n\n# Home Page\n\nclass HomePageCarouselItem(Orderable, CarouselItem):\n page = ParentalKey('home.HomePage', related_name='carousel_items')\n\n\nclass HomePageVideoItem(Orderable, VideoItem):\n page = ParentalKey('home.HomePage', related_name='video_items')\n\n\nclass HomePageServiceItem(Orderable, ServicelItem):\n page = ParentalKey('home.HomePage', related_name='service_items')\n\n\nclass HomePageProductItem(Orderable, ProductItem):\n page = ParentalKey('home.HomePage', related_name='product_items')\n\n\nclass HomePageAdvantageItem(Orderable, ProductItem):\n page = ParentalKey('home.HomePage', related_name='advantage_items')\n\n\nclass HomePageTestimonialItem(Orderable, TestimonialItem):\n page = ParentalKey('home.HomePage', related_name='testimonial_items')\n\n\nclass HomePageRecipeItem(Orderable, RecipeItem):\n page = ParentalKey('home.HomePage', related_name='recipe_items')\n\n\nclass HomePageRelatedLink(Orderable, RelatedLink):\n page = ParentalKey('home.HomePage', related_name='related_links')\n\n\nclass HomePage(Page):\n\n class Meta:\n verbose_name = \"Homepage\"\n\nHomePage.content_panels = [\n InlinePanel('carousel_items', label=\"Carousel items\"),\n InlinePanel('video_items', label=\"Video items\"),\n InlinePanel('service_items', label=\"Service items\"),\n InlinePanel('product_items', label=\"Product items\"),\n InlinePanel('advantage_items', label=\"Advantage items\"),\n InlinePanel('testimonial_items', label=\"Testimonial items\"),\n InlinePanel('recipe_items', label=\"Recipe items\"),\n InlinePanel('related_links', label=\"Related links\"),\n]\n\nHomePage.promote_panels = Page.promote_panels\n\n\n\n","repo_name":"wadadaaa/martalp","sub_path":"home/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":8694,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"71443278303","text":"class MafGroupLengthQualityFilter:\n def __init__(self, minLengthIngroup, minLenghtOutgroup, ingroup):\n self.minLengthIngroup = minLengthIngroup\n self.minLengthOutgroup = minLenghtOutgroup\n self.ingroup = ingroup\n \n def accept(self, maf):\n for i in range(maf.count()):\n if maf.name(i) in self.ingroup:\n if maf.length(i) < self.minLengthIngroup:\n return False\n else:\n if maf.length(i) < self.minLengthOutgroup:\n return False \n return True\n \n \n \n \n","repo_name":"kaspermunch/CoalhmmPipeline","sub_path":"CoalhmmPipeline/MafGroupLengthQualityFilter.py","file_name":"MafGroupLengthQualityFilter.py","file_ext":"py","file_size_in_byte":654,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"9960585024","text":"from api.models import Metric\nfrom datetime import datetime, timedelta\nimport pytz\n\n\ndef get_graph_data(metric_name, tags, org, interval=60, since=1):\n\n interval = interval * 60\n\n now = datetime.now(pytz.UTC)\n query_date = now - timedelta(hours=since)\n\n query = {\n 'org': org,\n 'created_on__gte': query_date\n }\n\n for k, v in tags.items():\n query['tags__%s' % k] = v\n\n mt = Metric.objects.filter(**query).order_by('-created_on')\n\n values = [\n {\n 'created_on': m.created_on,\n 'value': m.metrics[metric_name]}\n for m in mt\n ]\n values.reverse()\n\n if values:\n current_interval = values[0]['created_on']\n else:\n current_interval = 0\n\n current_values = []\n\n graph_metrics = []\n\n for v in values:\n diff = v['created_on'] - current_interval\n if diff.total_seconds() >= interval:\n val = 0\n if len(current_values) > 0:\n val = sum(current_values) / len(current_values)\n graph_metrics.append({\n 'date': v['created_on'],\n 'value': val\n })\n\n current_interval = v['created_on']\n current_values = []\n\n current_values.append(v['value'])\n\n if len(current_values) > 0:\n graph_metrics.append({\n 'date': v['created_on'],\n 'value': sum(current_values) / len(current_values)\n })\n\n return graph_metrics\n","repo_name":"dominicplouffe/OEL","sub_path":"api/common/metrics/graph.py","file_name":"graph.py","file_ext":"py","file_size_in_byte":1468,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"73074888862","text":"import requests\nfrom typing import NamedTuple, Dict\nfrom .exceptions import APIException, UserDataException\nfrom django.conf import settings\nimport random\n\n\nclass DomainAndLogin(NamedTuple):\n domain: str\n login: str\n\n\nclass Message(NamedTuple):\n id: str\n fromUserMail: str\n subject: str\n date: str\n body: str\n textBody: str\n htmlBody: str\n\n\nclass ResponseData(NamedTuple):\n status: int\n data: Dict\n user_data: Dict\n\n\ndef main_service(request) -> DomainAndLogin:\n \"\"\"Сервис вывода основной страницы\"\"\"\n data = _get_domain_and_login(request)\n\n if not data.login or not data.domain:\n domain = random.choice(settings.DOMAINS)\n login = ''.join([random.choice(settings.CHARS) for i in range(20)])\n\n request.session['domain'] = domain\n request.session['login'] = login\n return DomainAndLogin(domain=domain, login=login)\n return data\n\n\ndef get_list_messages_service(request) -> ResponseData:\n \"\"\"Сервис по получению сообщений пользователя по сессии\"\"\"\n data = _get_domain_and_login(request)\n\n if not data.login or not data.domain:\n raise UserDataException('Ошибка получение данных пользователя')\n try:\n messages = _get_messages(data)\n except APIException:\n raise UserDataException('Ошибка получения данных из api')\n return ResponseData(status=200, data=messages, user_data=data)\n\n\ndef get_one_message_service(request, id) -> ResponseData:\n \"\"\"Сервис по получению одного сообщения по сессии\"\"\"\n data = _get_domain_and_login(request)\n if not data.login or not data.domain:\n raise UserDataException('Ошибка получения данных пользователя')\n\n try:\n message = _get_one_message(data, id)\n except APIException:\n raise UserDataException('Ошибка получения данных из API')\n return ResponseData(status=200, data=message, user_data=data)\n\n\ndef _get_domain_and_login(request) -> DomainAndLogin:\n \"\"\"Получаем данные пользователя\"\"\"\n domain = request.session.get('domain')\n login = request.session.get('login')\n return DomainAndLogin(domain=domain, login=login)\n\n\ndef _get_messages(user_data: DomainAndLogin) -> list[Message]:\n \"\"\"Получение списка сообщений\"\"\"\n url = f'{settings.API_SEC_MAIL_URL}?action=getMessages&login={user_data.login}&domain={user_data.domain}'\n try:\n messages = requests.get(url).json()\n except Exception:\n raise APIException('Ошибка API')\n return messages\n\n\ndef _get_one_message(user_data: DomainAndLogin, id: int) -> Message:\n \"\"\"Получение одного сообщения\"\"\"\n url = f'{settings.API_SEC_MAIL_URL}?action=readMessage&login={user_data.login}&domain={user_data.domain}&id={id}'\n try:\n message = requests.get(url).json()\n except APIException:\n raise APIException('Ошибка API')\n return Message(id=message['id'],\n fromUserMail=message['from'],\n subject=message['subject'],\n date=message['date'],\n body=message['body'],\n textBody=message['textBody'],\n htmlBody=message['htmlBody'])\n","repo_name":"DmitriiGrekov/temp_mail","sub_path":"temp_mail/core/services.py","file_name":"services.py","file_ext":"py","file_size_in_byte":3428,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"73165994463","text":"#Thanks to @TelescopyBot Owner........\n\n\nimport datetime\nfrom telethon import events\nfrom telethon.errors.rpcerrorlist import YouBlockedUserError\nfrom telethon.tl.functions.account import UpdateNotifySettingsRequest\nfrom userbot.utils import admin_cmd\n\n\n@bot.on(admin_cmd(r\"circle$\"))\nasync def _(event):\n if event.fwd_from:\n return \n if not event.reply_to_msg_id:\n await event.edit(\"Reply to a media msg sur......\")\n return\n reply_message = await event.get_reply_message() \n if not reply_message.media:\n await event.edit(\"Reply to media msg sur......\")\n return\n chat = \"@TelescopyBot\"\n sender = reply_message.sender\n if reply_message.sender.bot:\n await event.edit(\"Yeh kya bawasir hai.... I need actual user.....\")\n return\n kraken = await event.edit(\"Ok I got it🔥🔥\")\n async with event.client.conversation(chat) as conv:\n try: \n await conv.send_message(\"/start\")\n await conv.get_response()\n await conv.send_message(reply_message)\n response = await conv.get_response()\n await event.client.send_read_acknowledge(conv.chat_id)\n except YouBlockedUserError: \n await kraken.edit(\"`Please unblock` @TelescopyBot `and try again`\")\n return\n await kraken.delete()\n if not response.media:\n await event.client.send_message(event.chat_id, response.message)\n if response.media:\n await event.client.send_file(event.chat_id, response)\n","repo_name":"CyberBoyAyush/ABUserbot","sub_path":"userbot/plugins/circle.py","file_name":"circle.py","file_ext":"py","file_size_in_byte":1565,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"7"} +{"seq_id":"34789400607","text":"# Import the necessary libraries for data analysis and visualization\nimport pandas as pd\nfrom sklearn.linear_model import LinearRegression\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import train_test_split\n\n# Load the survey data into a Pandas DataFrame\ndf = pd.read_csv('Mall_Customers.csv')\n\n# Use Pandas to clean and preprocess the data\ndf = df.dropna() # drop rows with missing values\ndf = df[df['Age'] > 0] # drop rows with invalid age values\n\n# Use seaborn to create a line plot\nsns.lineplot(x='Purchase %', y='Age', data=df)\nplt.show()\n\n# Use Pandas to create a scatter plot showing the relationship between Purchase and Income\nsns.scatterplot(x='Purchase %', y='Income', data=df)\n\n# Show the plot\nplt.show()\n\n# Use scikit-learn to build a linear regression model to predict attitude from age, gender, and income\nX = df[['Income']]\ny = df['Purchase %']\nplt.title('Purchasing % Across different Age groups')\nplt.xlabel('Income')\nplt.ylabel('Purchase attitude')\nmodel = LinearRegression()\nmodel.fit(X, y)\n\n# Evaluate the model's performance on the training data\nprint('R-squared:', model.score(X, y))\n\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)\n\nmodel = LinearRegression()\nmodel.fit(X_train, y_train)\n\n# Evaluate the model's performance on the test data\nprint('R-squared:', model.score(X_test, y_test))\n\n# Use the model to make predictions and evaluate its performance\npredictions = model.predict(X)\nprint('R-squared:', model.score(X, y))\n\n# Plot the predictions as a line on top of the data points\nplt.plot(X, predictions, color='red')\n\n# Show the plot\nplt.show()\n\n","repo_name":"MuyiwaNau/Monday02","sub_path":"predict.py","file_name":"predict.py","file_ext":"py","file_size_in_byte":1644,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"32537994344","text":"from django.contrib.auth import get_user_model\nfrom rest_framework.test import APITestCase\nfrom rest_framework.test import APIClient\nfrom rest_framework import status\nfrom django.urls import reverse\nfrom ..models import ToDo\n\nUser = get_user_model()\nclient = APIClient()\n\n\nclass ToDoTests(APITestCase):\n\n def setUp(self) -> None:\n self.user = User.objects.create_user(\n username='casper', email='casper@mail.ru', password='password123')\n self.todo1 = ToDo.objects.create(\n headline='headline1', description='description1')\n self.todo2 = ToDo.objects.create(\n headline='headline2', description='description2')\n return super().setUp()\n\n def test_todo_list(self):\n '''Проверить получение списка задач\n '''\n url = reverse('todo:todolist_create')\n response = client.get(url)\n self.assertEqual(response.status_code, status.HTTP_200_OK)\n self.assertEqual(ToDo.objects.count(), len(response.data))\n\n def test_todo_create(self):\n '''Проверить создание задачи и игнорирование поля execute, должна быть execute=False\n '''\n url = reverse('todo:todolist_create')\n body = {\n \"headline\": \"New headline\",\n \"description\": \"New description\",\n \"deadline\": \"2022-10-30 15:37\",\n \"execute\": \"True\"\n }\n response = client.post(url, body, format='json')\n self.assertEqual(response.status_code, status.HTTP_201_CREATED)\n self.assertEqual(ToDo.objects.count(), 3)\n self.assertEqual(ToDo.objects.get(\n headline='New headline').execute, False)\n\n def test_todo_get(self):\n '''Проверить получение одной задачи\n '''\n url = reverse('todo:todogetupdel', kwargs={'pk': 1})\n client.credentials(HTTP_AUTHORIZATION='Token ' + self.user.token)\n response = client.get(url)\n self.assertEqual(response.status_code, status.HTTP_200_OK)\n self.assertEqual(response.data.get('id'), 1)\n\n def test_todo_update(self):\n '''Проверить редактирование задачи.\n '''\n url = reverse('todo:todogetupdel', kwargs={'pk': 1})\n client.credentials(HTTP_AUTHORIZATION='Token ' + self.user.token)\n body = {\n \"headline\": \"Change headline\",\n \"description\": \"Change description\",\n }\n response = client.patch(url, body, format='json')\n self.assertEqual(response.status_code, status.HTTP_200_OK)\n self.assertEqual(ToDo.objects.get(\n headline='Change headline').description, 'Change description')\n\n def test_todo_update_deadline(self):\n '''Проверить валидацию поля deadline.\n '''\n url = reverse('todo:todogetupdel', kwargs={'pk': 1})\n client.credentials(HTTP_AUTHORIZATION='Token ' + self.user.token)\n body = {\n 'deadline': '2012-10-31 15:20+00:00'\n }\n response = client.patch(url, body, format='json')\n self.assertEqual(response.status_code, status.HTTP_400_BAD_REQUEST)\n\n def test_todo_delete(self):\n '''Проверить удаление задачи.\n '''\n url = reverse('todo:todogetupdel', kwargs={'pk': 1})\n client.credentials(HTTP_AUTHORIZATION='Token ' + self.user.token)\n response = client.delete(url, None, format='json')\n self.assertEqual(response.status_code, status.HTTP_204_NO_CONTENT)\n self.assertEqual(ToDo.objects.count(), 1)\n\n def test_todo_execute_change(self):\n '''Проверить отметить задачу выполненной .\n '''\n url = reverse('todo:todoexecute', kwargs={'pk': 1})\n client.credentials(HTTP_AUTHORIZATION='Token ' + self.user.token)\n response = client.patch(url, None, format='json')\n self.assertEqual(response.status_code, status.HTTP_200_OK)\n self.assertEqual(ToDo.objects.get(id='1').execute, True)\n","repo_name":"Doszhan-M/BuhUchetTest","sub_path":"BuhUchetTest/apps/ToDo/tests/tests_views.py","file_name":"tests_views.py","file_ext":"py","file_size_in_byte":4099,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13958995559","text":"# System\nimport os\n\n# Packages\nimport frontmatter\nimport glob\nimport re\nfrom django.http import Http404, HttpResponseRedirect\nfrom django.template import Context, loader, TemplateDoesNotExist\nfrom django.views.generic.base import TemplateView\nfrom mistune import Markdown, BlockLexer\n\n\ndef _insensitive_glob(pattern, base_dir):\n \"\"\"\n Look for files with glob patterns,\n but case ignoring case\n \"\"\"\n\n def either(c):\n return \"[%s%s]\" % (c.lower(), c.upper()) if c.isalpha() else c\n\n search = \"\".join(map(either, pattern))\n search_path = os.path.join(base_dir, search)\n\n return glob.glob(search_path)\n\n\ndef _template_exists(path):\n \"\"\"\n Check if a template exists\n without raising an exception\n \"\"\"\n\n try:\n loader.get_template(path)\n return True\n except (TemplateDoesNotExist, OSError):\n return False\n\n\ndef _find_template_url(path):\n \"\"\"\n Look for a template by:\n - checking for case-insensitive matches\n - seeing if the URL exactly matches the filename,\n with extension\n \"\"\"\n\n matches = []\n\n first_engine_name = next(iter(loader.engines.templates))\n template_dirs = loader.engines.templates[first_engine_name][\"DIRS\"]\n\n for template_dir in template_dirs:\n for match in (\n _insensitive_glob(path, template_dir)\n + _insensitive_glob(path + \".html\", template_dir)\n + _insensitive_glob(path + \".md\", template_dir)\n ):\n cleaned_match = re.sub(r\"^\" + template_dir, \"\", match)\n cleaned_match = re.sub(r\"\\.(html|md)$\", \"\", cleaned_match)\n matches.append(cleaned_match)\n\n # Only return a found template if we found only one\n if (\n len(matches) == 1\n and matches[0].lower() == \"/\" + path.lower()\n and _get_template(matches[0].lstrip(\"/\"))\n ):\n return matches[0]\n\n\ndef _relative_template_path(path, origin_filepath):\n \"\"\"\n Infer a path to a template from a partial path\n\n - If the path starts with a \"/\",\n simply ask Django to locate the template\n - If the path doesn't start with a \"/\",\n work out the absolute path relative to the origin_filepath\n \"\"\"\n\n if path.startswith(\"/\"):\n # \"absolute\" path, just strip the leading \"/\"\n # so template loader can do its work\n path = path.lstrip(\"/\")\n else:\n # \"relative\" path, use the existing filepath\n path = os.path.relpath(\n os.path.join(os.path.dirname(origin_filepath), path)\n )\n\n return path\n\n\ndef _get_template(url_path):\n \"\"\"\n Given a basic path, find an HTML or Markdown file\n \"\"\"\n\n # Try to match HTML or Markdown files\n if _template_exists(url_path + \".html\"):\n return url_path + \".html\"\n elif _template_exists(os.path.join(url_path, \"index.html\")):\n return os.path.join(url_path, \"index.html\")\n elif _template_exists(url_path + \".md\"):\n return url_path + \".md\"\n elif _template_exists(os.path.join(url_path, \"index.md\")):\n return os.path.join(url_path, \"index.md\")\n\n return None\n\n\nclass WebteamBlockLexer(BlockLexer):\n list_rules = (\n \"newline\",\n \"block_code\",\n \"fences\",\n \"lheading\",\n \"hrule\",\n \"table\",\n \"nptable\",\n \"block_quote\",\n \"list_block\",\n \"block_html\",\n \"text\",\n )\n\n\nclass TemplateFinder(TemplateView):\n parse_markdown = Markdown(\n parse_block_html=True,\n parse_inline_html=True,\n block=WebteamBlockLexer(),\n )\n\n def _parse_markdown_file(self, filepath):\n \"\"\"\n Parse a markdown file into the relevant parts.\n\n - html_content: The parsed HTML from the Markdown content\n - context: Any \"includes\" and custom \"context\" specified in frontmatter\n - template_filepath: An absolute filepath inferred from the frontmatter\n \"\"\"\n\n # Parse frontmatter, and add it to context\n markdown_template = loader.get_template(filepath)\n\n if markdown_template.backend.name == \"django\":\n file_contents = markdown_template.template.render(Context())\n else:\n file_contents = markdown_template.template.render()\n markdown = frontmatter.loads(file_contents)\n\n # Set the template path\n wrapper_template = markdown.metadata.get(\"wrapper_template\")\n context = markdown.metadata.get(\"context\", {})\n\n if not wrapper_template:\n # If no wrapper template specified,\n # this doesn't count as a valid Markdown file\n return None\n\n template_filepath = _relative_template_path(wrapper_template, filepath)\n\n # Parse core HTML content\n context[\"html_content\"] = self.parse_markdown(markdown.content)\n\n # Add any Markdown includes\n for key, path in markdown.metadata.get(\n \"markdown_includes\", {}\n ).items():\n include_path = _relative_template_path(path, filepath)\n include_template = loader.get_template(include_path)\n if include_template.backend.name == \"django\":\n include_content = include_template.template.render(Context())\n else:\n include_content = include_template.template.render()\n context[key] = self.parse_markdown(include_content)\n\n return {\"context\": context, \"template\": template_filepath}\n\n def render_to_response(self, context, **response_kwargs):\n \"\"\"\n Return a response, using the `response_class` for this view, with a\n template rendered with the given context.\n Pass response_kwargs to the constructor of the response class.\n \"\"\"\n\n # Response defaults\n response_kwargs.setdefault(\"content_type\", self.content_type)\n\n # Find .html or .md template files\n path = self.request.path.lstrip(\"/\")\n matching_template = _get_template(path)\n\n # If we couldn't find the template, show 404\n if not matching_template:\n found_template_url = _find_template_url(path)\n\n if found_template_url:\n return HttpResponseRedirect(found_template_url)\n else:\n raise Http404(\"Can't find template for \" + self.request.path)\n\n # If we found a Markdown file, parse it to find its wrapper template\n if matching_template.endswith(\".md\"):\n markdown_data = self._parse_markdown_file(matching_template)\n\n if not markdown_data:\n raise Http404(\n self.request.path + \" not correctly configurated.\"\n )\n\n matching_template = markdown_data[\"template\"]\n context.update(markdown_data[\"context\"])\n\n # Send the response\n return self.response_class(\n request=self.request,\n template=matching_template,\n context=context,\n using=self.template_engine,\n **response_kwargs\n )\n","repo_name":"canonical/canonicalwebteam.django-views","sub_path":"canonicalwebteam/django_views/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":6963,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"10741804221","text":"from typing import List, Dict, Set, Tuple\nfrom enum import Enum\nimport unittest\nfrom copy import deepcopy\n\nINPUT_FILE = \"input.txt\"\nTEST_INPUT_FILE = \"test_input.txt\"\n\nBOOT_PROCESS_CYCLE_COUNT = 6\n\n\nclass CubeStatus(Enum):\n ACTIVE = \"#\"\n INACTIVE = \".\"\n\n\nclass FourDimensionalCoordinates:\n def __init__(self, x, y, z=0, w=0):\n self.x = x\n self.y = y\n self.z = z\n self.w = w\n\n def __key(self):\n return (self.x, self.y, self.z, self.w)\n\n def __hash__(self):\n return hash(self.__key())\n\n def __eq__(self, other):\n if not isinstance(other, FourDimensionalCoordinates):\n return NotImplemented\n\n return self.__key() == other.__key()\n\n\nclass ConwayCubesHandler:\n def __init__(self, activeCubesCoordinates: Set[FourDimensionalCoordinates]):\n self._activeCubesCoordinates = activeCubesCoordinates\n\n def executeBootingCycles(self, numCycles: int) -> None:\n for _ in range(0, numCycles):\n self._executeIteration()\n\n def getActiveCubesNum(self) -> int:\n return len(self._activeCubesCoordinates)\n\n def _executeIteration(self) -> None:\n nextRoundActiveCubes: Set[FourDimensionalCoordinates] = deepcopy(self._activeCubesCoordinates)\n self._deleteNoLongerActiveCubes(nextRoundActiveCubes)\n self._addNewlyActivatedCubes(nextRoundActiveCubes)\n self._activeCubesCoordinates = nextRoundActiveCubes\n\n def _deleteNoLongerActiveCubes(self, nextRoundActiveCubes: Set[FourDimensionalCoordinates]) -> None:\n cubesToDelete: List[Coordinates] = []\n for activeCube in self._activeCubesCoordinates:\n activeNeighborsCount = self._getActiveNeighborsCount(activeCube)\n if activeNeighborsCount != 2 and activeNeighborsCount != 3:\n cubesToDelete.append(activeCube)\n\n for cube in cubesToDelete:\n nextRoundActiveCubes.remove(cube)\n\n def _addNewlyActivatedCubes(self, nextRoundActiveCubes: Set[FourDimensionalCoordinates]) -> None:\n for inactiveCube in self._getNeighborsOfActiveCubes():\n activeNeighborsCount = self._getActiveNeighborsCount(inactiveCube)\n if activeNeighborsCount == 3:\n nextRoundActiveCubes.add(inactiveCube)\n\n def _getActiveNeighborsCount(self, currentCube: FourDimensionalCoordinates) -> int:\n activeNeighborsCount = sum(\n 1 for cube in self._getNeighborCubes(currentCube) if cube in self._activeCubesCoordinates)\n return activeNeighborsCount - 1 if currentCube in self._activeCubesCoordinates else activeNeighborsCount\n\n def _getNeighborsOfActiveCubes(self) -> Set[FourDimensionalCoordinates]:\n cubesInTheArea: Set[Coordinates] = set()\n\n for activeCube in self._activeCubesCoordinates:\n neighborCubes = self._getNeighborCubes(activeCube)\n cubesInTheArea.update(neighborCubes)\n\n cubesInTheArea -= self._activeCubesCoordinates\n return cubesInTheArea\n\n def _getNeighborCubes(self, currentCube: FourDimensionalCoordinates) -> Set[FourDimensionalCoordinates]:\n nearbyCubes: Set[FourDimensionalCoordinates] = set()\n for xOffset in range(-1, 2):\n for yOffset in range(-1, 2):\n for zOffset in range(-1, 2):\n nearbyCubes.add(\n FourDimensionalCoordinates(currentCube.x + xOffset, currentCube.y + yOffset, currentCube.z + zOffset))\n return nearbyCubes\n\n\nclass ConwayCubesHandlerPartTwo(ConwayCubesHandler):\n def _getNeighborCubes(self, currentCube: FourDimensionalCoordinates) -> Set[FourDimensionalCoordinates]:\n nearbyCubes: Set[CoordinatesPartTwo] = set()\n for xOffset in range(-1, 2):\n for yOffset in range(-1, 2):\n for zOffset in range(-1, 2):\n for wOffset in range(-1, 2):\n nearbyCubes.add(FourDimensionalCoordinates(currentCube.x + xOffset, currentCube.y + yOffset,\n currentCube.z + zOffset, currentCube.w + wOffset))\n return nearbyCubes\n\n\ndef getInitialActiveCubesCoordinateNums(inputFile: str) -> Set[FourDimensionalCoordinates]:\n fourDimensionalCoordinatesCollection: Set[FourDimensionalCoordinates] = set()\n\n with open(inputFile, \"r\") as inputFile:\n lines = inputFile.readlines()\n for j in range(len(lines)):\n for i in range(len(lines[j].strip(\"\\n\"))):\n if lines[j][i] == CubeStatus.ACTIVE.value:\n fourDimensionalCoordinatesCollection.add(FourDimensionalCoordinates(i, j))\n\n return fourDimensionalCoordinatesCollection\n\n\ndef main():\n initialActiveCubes: RawCoordinateNumsCollection = getInitialActiveCubesCoordinateNums(INPUT_FILE)\n\n cubeHandler = ConwayCubesHandler(initialActiveCubes)\n cubeHandler.executeBootingCycles(BOOT_PROCESS_CYCLE_COUNT)\n print(cubeHandler.getActiveCubesNum()) # 353\n\n cubeHandler = ConwayCubesHandlerPartTwo(initialActiveCubes)\n cubeHandler.executeBootingCycles(BOOT_PROCESS_CYCLE_COUNT)\n print(cubeHandler.getActiveCubesNum()) # 2472\n\n\nclass ActiveCubesOperationTester(unittest.TestCase):\n def test_getActiveCubesInNextRound_threeDimensionCoordinates_correctActiveCubesNumReturned(self):\n initialActiveCubesCoordinates = getInitialActiveCubesCoordinateNums(TEST_INPUT_FILE)\n cubeHandler = ConwayCubesHandler(initialActiveCubesCoordinates)\n cubeHandler.executeBootingCycles(BOOT_PROCESS_CYCLE_COUNT)\n\n self.assertEqual(112, cubeHandler.getActiveCubesNum())\n\n def test_getActiveCubesInNextRound_fourDimensionCoordinates_correctActiveCubesNumReturned(self):\n initialActiveCubesCoordinates = getInitialActiveCubesCoordinateNums(TEST_INPUT_FILE)\n cubeHandler = ConwayCubesHandlerPartTwo(initialActiveCubesCoordinates)\n cubeHandler.executeBootingCycles(BOOT_PROCESS_CYCLE_COUNT)\n\n self.assertEqual(848, cubeHandler.getActiveCubesNum())\n\n\nif __name__ == '__main__':\n # main()\n unittest.main()\n","repo_name":"zseen/advent-of-code","sub_path":"2020/17/Day17-ConwayCubes.py","file_name":"Day17-ConwayCubes.py","file_ext":"py","file_size_in_byte":6063,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27608085476","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nImport KA JIRA RSS export to GitHub\n\nTo use, create credentials.json: {\"username\": \"yourusername\", \"password\": \"password\"}\nwith your GitHub credentials. Beware API limits!\n\"\"\"\nfrom lxml import etree\nimport re\nimport time\nimport json\nimport argparse\nimport github\nimport argparse\nfrom ansicolor import red, black\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"filename\", help=\"The JIRA XML RSS export to read\")\nparser.add_argument(\"-d\", \"--delay\", type=int, default=10, help=\"How long to delay between creating issue\")\nargs = parser.parse_args()\n\netree = etree.parse(args.filename)\n\nrss = etree.getroot()\nchannel = rss.getchildren()[0]\nitems = channel.xpath(\"item\")\n\ndef itemToDict(item):\n \"Extract important attributes from an item and \"\n key = item.xpath(\"key\")[0].text\n title = item.xpath(\"title\")[0].text\n created = item.xpath(\"created\")[0].text\n reporter = item.xpath(\"reporter\")[0].text\n resolution = item.xpath(\"resolution\")[0].text\n description = item.xpath(\"description\")[0].text\n preview_url = item.xpath(\"customfields/customfield[@id='customfield_10027']//customfieldvalue\")[0].text\n issue_type = item.xpath(\"customfields/customfield[@id='customfield_10300']//customfieldvalue\")[0].text\n framework = item.xpath(\"customfields/customfield[@id='customfield_10029']//customfieldvalue\")[0].text\n try:\n exercise = item.xpath(\"customfields/customfield[@id='customfield_10024']//customfieldvalue\")[0].text\n except IndexError:\n exercise = None\n edit_url = item.xpath(\"customfields/customfield[@id='customfield_10028']//customfieldvalue\")[0].text\n return {\n \"id\": key, \"title\": title, \"created\": created,\n \"reporter\": reporter, \"resolution\": resolution,\n \"preview_url\": preview_url, \"issue_type\": issue_type,\n \"framework\": framework, \"exercise\": exercise,\n \"edit_url\": edit_url, \"description\": description\n }\n\ndef formatItem(item):\n return \"\"\"\nCreated: {0}\nReported by: {1}\nIssue type: {5}\nJIRA status: {2}\nLinks: [Preview]({3}), [Edit]({4})\n\n{6}\n \"\"\".format(item[\"created\"], item[\"reporter\"], item[\"resolution\"],\n item[\"preview_url\"], item[\"edit_url\"], item[\"issue_type\"], item[\"description\"])\n\n\ndef find_issue_keys(repo):\n \"\"\"Generate all issue keys for correctly named GitHub issues\"\"\"\n for issue in repo.get_issues(): # iterate open issues\n # Try to search for key\n m = re.search(r\"\\[([A-Z]{2}-\\d+)\\]\", issue.title)\n if m is None:\n continue # Ignore this issue\n yield m.group(1)\n\ndef createGithubIssue(repo, item):\n \"Create a GitHub issue\"\n repo.create_issue(item[\"title\"], formatItem(item))\n\njiraItems = [itemToDict(item) for item in items]\n\nwith open(\"credentials.json\") as credentialsFile:\n credentials = json.load(credentialsFile)\ngh = github.Github(credentials[\"username\"], credentials[\"password\"])\n\nrepo = gh.get_repo(\"ulikoehler/KADeutschIssues\")\nissues = repo.get_issues()\n\n# Find issues which are in JIRA but not in GitHub\njiraKeys = set([i[\"id\"] for i in jiraItems])\ngithubKeys = set(find_issue_keys(repo))\nmissing_items = jiraKeys.symmetric_difference(githubKeys)\n\nprint(\"Creating {0} new issues on GitHub...\\n\".format(len(missing_items)))\n\n# Iterate every JIRA item, skip items already on github & submit new issue for remaining ones\ntodo = []\nfor item in jiraItems:\n if item[\"id\"] in missing_items:\n todo.append(item)\n\n# Cant use for loop because we might need to retry\nidx = 0\nwhile idx < len(todo):\n item = todo[idx]\n try:\n createGithubIssue(repo, item)\n print(\"Created issue #{0}: {1}\".format(idx + 1, item[\"title\"]))\n time.sleep(float(args.delay))\n idx += 1# Next item only on success\n except github.GithubException as e:\n print(red(\"GitHub error, likely rate limiting (retrying in 60s): {0}\".format(e), bold=True))\n time.sleep(60.)\n # Retry this item\n","repo_name":"ulikoehler/KADeutschIssues","sub_path":"ImportJIRAXML.py","file_name":"ImportJIRAXML.py","file_ext":"py","file_size_in_byte":3971,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"21551754993","text":"import matplotlib.pyplot as plt\r\nimport numpy as np\r\n\r\nclass traceback:\r\n \r\n # Calls function to find tracebacks for each scoring matrix -------------------------------------------------------------------\r\n def __init__(self, sc, seq1, seq2, M1, M2, M3):\r\n self.trc1M1, self.trc2M1 = self.createTracebacks(sc, seq1, seq2, M1)\r\n self.trc1M2, self.trc2M2 = self.createTracebacks(sc, seq1, seq2, M2)\r\n self.trc1M3, self.trc2M3 = self.createTracebacks(sc, seq1, seq2, M3)\r\n \r\n \r\n # Finds the tracebacks for sequence 1 and 2 -----------------------------------------------------------------------------------\r\n def createTracebacks(self, sc, seq1, seq2, M):\r\n trc1, trc2, maxi = '', '', M.max()\r\n \r\n if maxi > 0:\r\n x, y = 0, 0\r\n\r\n # Finds the last max in the case of multiple maxes\r\n for i in range(1, len(seq1) + 1):\r\n for j in range(1, len(seq2) + 1):\r\n if M[i,j] == maxi:\r\n x = i\r\n y = j\r\n\r\n while True:\r\n\r\n if (M[x,y] == M[x-1,y-1] + sc.match or M[x,y] == M[x-1,y-1] + sc.miss) and M[x-1,y-1] == max(M[x-1,y-1], M[x-1,y], M[x, y-1]): \r\n trc1 = seq1[x-1] + trc1\r\n trc2 = seq2[y-1] + trc2\r\n if M[x-1,y-1] == 0:\r\n break\r\n x = x-1\r\n y = y-1\r\n elif M[x,y] == M[x-1,y] + sc.gap or M[x,y] == M[x-1,y] + sc.gapExt:\r\n trc1 = seq1[x-1] + trc1\r\n trc2 = '_' + trc2\r\n x = x-1\r\n elif M[x,y] == M[x, y-1] + sc.gap or M[x,y] == M[x,y-1] + sc.gapExt:\r\n trc1 = '_' + trc1\r\n trc2 = seq2[y-1] + trc2\r\n y = y-1\r\n else:\r\n trc1 = seq1[x-1] + trc1\r\n trc2 = seq2[y-1] + trc2\r\n if M[x-1,y-1] == 0:\r\n break\r\n x = x-1\r\n y = y-1\r\n\r\n return trc1, trc2\r\n\r\n else:\r\n return trc1, trc2\r\n\r\n # Prints tracebacks for each scoring matrix ----------------------------------------------------------------------------------\r\n def printTraces(self):\r\n print(self.trc1M1, ' ', self.trc1M2, ' ', self.trc1M3)\r\n print(self.trc2M1, ' ', self.trc2M2, ' ', self.trc2M3)","repo_name":"zapulam/SmithWatermanProcessor","sub_path":"Traceback.py","file_name":"Traceback.py","file_ext":"py","file_size_in_byte":2478,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18581021052","text":"import json\nimport boto3\nfrom boto3.dynamodb.conditions import Key\n\n\ndynamodb = boto3.resource('dynamodb', region_name='us-east-2', endpoint_url=\"https://dynamodb.us-east-2.amazonaws.com\")\ntable = dynamodb.Table('parata-users')\n\ndef getGroupInfo(group_code):\n response = table.query(\n KeyConditionExpression=Key('group_code').eq(group_code)\n )\n # if len(response['Items']) == 0:\n # return {\n # 'statusCode': 400,\n # 'body': json.dumps('invalid group code')\n # }\n print(response)\n return {\n 'statusCode': 200,\n 'body': json.dumps({\n 'group_code': group_code,\n 'group_members': response['Items']\n })\n }\n\ndef lambda_handler(event, context):\n group_code = event['queryStringParameters']['group_code']\n return getGroupInfo(group_code)\n\n","repo_name":"alicemliu/parade-public","sub_path":"api/get_group_info/lambda_function.py","file_name":"lambda_function.py","file_ext":"py","file_size_in_byte":841,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"42060551545","text":"#!/usr/bin/env python3\n\nfrom __future__ import print_function\nimport argparse\nimport subprocess\nimport shutil\nimport sys\nimport glob\nimport os\nimport csv\n\n# Wrapper for the RP2paths script that takes the same input (results.csv) as the original script but returns\n# the out_paths.csv so as to be compliant with Galaxy\n\n#Debug function\ndef eprint(*args, **kwargs):\n print(*args, file=sys.stderr, **kwargs)\n\n#Although it may seem better to call the RP2path.py script directly, \ndef main(rp_results, timeout, outDir):\n rp2paths_command = ['python', '/src/RP2paths.py', 'all', rp_results, '--outdir', outDir, '--timeout', str(timeout)]\n try:\n #exit_code = subprocess.call(rp2paths_command, stdout=subprocess.DEVNULL, stderr=subprocess.STDOUT)\n #exit_code = subprocess.call(rp2paths_command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n exit_code = subprocess.call(rp2paths_command)\n except OSError as e:\n eprint('ERROR: subprocess detected an error when calling the rp2paths command')\n return 2\n return exit_code\n\n\n#function that takes the .dat input of a file, opens to be read by python and then writes it to a CSV file\ndef readCopyFile(inputFile, outDir):\n #outputFile = outDir+'/'+inputFile.split('/')[-1].replace('.dat', '')+'.csv'\n outputFile = inputFile.split('/')[-1].replace('.dat', '')+'.csv' \n with open(outputFile, 'w') as outF:\n outCSV = csv.writer(outF, delimiter=',', quotechar='\"', quoting=csv.QUOTE_NONNUMERIC)\n with open(inputFile, 'r') as inF:\n inCSV = csv.reader(inF, delimiter=',', quotechar='\"')\n for row in inCSV:\n outCSV.writerow(row)\n return outputFile, inputFile.split('/')[-1].replace('.dat', '')+'.csv'\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser('Python wrapper for the python RP2paths script')\n parser.add_argument('-rp_results', type=str)\n parser.add_argument('-out_paths', type=str)\n parser.add_argument('-timeout', type=int)\n parser.add_argument('-out_compounds', type=str)\n #parser.add_argument('-out_scope_csv', type=str)\n params = parser.parse_args()\n outDir = '/src/results'\n exit_code = main(params.rp_results, params.timeout, outDir)\n shutil.copy2(outDir+'/out_paths.csv', params.out_paths)\n shutil.copy2(outDir+'/compounds.txt', params.out_compounds)\n #print(glob.glob('*_scope.csv'))\n #shutil.copy2(outDir+'/'+glob.glob('*_scope.csv')[0], params.out_scope_csv)\n exit(exit_code)\n","repo_name":"ibisba/DBT_pipeline","sub_path":"galaxy_rp2paths/wrapRP2paths.py","file_name":"wrapRP2paths.py","file_ext":"py","file_size_in_byte":2493,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"14213058321","text":"import argparse\nfrom os import walk, mkdir, listdir\nfrom os.path import splitext, basename, join, exists, isfile\nimport subprocess\n\nimport torch as th\nimport numpy as np\nimport scipy\nfrom scipy.signal import spectrogram\nfrom scipy.io import wavfile\n\nfrom random import shuffle\nfrom math import ceil, floor\n\nfrom typing import Tuple\n\nfrom tqdm import tqdm\n\n\n###############\n# WAV stuff\n###############\n\ndef compute_wav_size(wav_path: str, nb_sec) -> Tuple[int, int]:\n sampling_rate, data = wavfile.read(wav_path)\n split_size = sampling_rate * nb_sec\n nb_split = floor(data.shape[0] / split_size)\n return nb_split, split_size\n\n\ndef open_wav(wav_path: str, nb_sec: int) -> Tuple[int, np.ndarray]:\n assert nb_sec > 0, f\"Split length must be > 0 (actual == {nb_sec}).\"\n\n sampling_rate, data = wavfile.read(wav_path)\n\n split_size = sampling_rate * nb_sec\n nb_split = floor(data.shape[0] / split_size)\n\n splitted_audio = np.asarray(np.split(data[:split_size * nb_split], nb_split))\n\n int_size = splitted_audio.itemsize * 8.\n\n splitted_audio = splitted_audio.astype(np.float16)\n splitted_audio[:, :, 0] = splitted_audio[:, :, 0] / (2. ** int_size) * 2.\n splitted_audio[:, :, 1] = splitted_audio[:, :, 1] / (2. ** int_size) * 2.\n\n return sampling_rate, splitted_audio.mean(axis=2)\n\n\n###############\n# Spectro stuff\n###############\n\ndef spectro_raw_audio(raw_audio_split: np.ndarray, nperseg: int, noverlap: int) -> np.ndarray:\n assert len(raw_audio_split.shape) == 2, \\\n f\"Wrong audio shape len (actual == {len(raw_audio_split.shape)}, needed == 2).\"\n\n assert raw_audio_split.dtype == np.float16, \\\n f\"Wrong ndarray dtype (actual == {raw_audio_split.dtype}, neede == {np.float16}).\"\n\n max_value = raw_audio_split.max()\n min_value = raw_audio_split.min()\n assert max_value <= 1.0 and min_value >= -1., \\\n f\"Raw audio values must be normlized between [-1., 0.] (actual == [{min_value}, {max_value}]).\"\n\n padded_raw_audio_split = np.pad(raw_audio_split, ((0, 0), (0, noverlap)), \"constant\", constant_values=0)\n return np.apply_along_axis(lambda split: spectrogram(split, nperseg=nperseg, noverlap=noverlap)[-1], 1,\n padded_raw_audio_split)\n\n\n###############\n# FFT stuff\n###############\n\ndef compute_fft_size(raw_audio_split_size: Tuple[int, int], nfft: int) -> Tuple[int, int, int]:\n return raw_audio_split_size[0], nfft, raw_audio_split_size[1] // nfft\n\n\ndef fft_raw_audio(raw_audio_split: np.ndarray, nfft: int) -> np.ndarray:\n assert len(raw_audio_split.shape) == 2, \\\n f\"Wrong audio shape len (actual == {len(raw_audio_split.shape)}, needed == 2).\"\n\n assert raw_audio_split.dtype == np.float16, \\\n f\"Wrong ndarray dtype (actual == {raw_audio_split.dtype}, neede == {np.float16}).\"\n\n max_value = raw_audio_split.max()\n min_value = raw_audio_split.min()\n assert max_value <= 1.0 and min_value >= -1., \\\n f\"Raw audio values must be normlized between [-1., 1.] (actual == [{min_value}, {max_value}]).\"\n\n splitted_data = np.stack(np.hsplit(raw_audio_split, raw_audio_split.shape[-1] / nfft), axis=-2)\n return np.apply_along_axis(lambda sub_split: scipy.fft(sub_split), 2, splitted_data)\n\n\ndef ifft_samples(fft_samples: np.ndarray, nfft: int) -> np.ndarray:\n assert len(fft_samples.shape) == 3, \\\n f\"Wrong spectrogram shape len (actual : {len(fft_samples.shape)}, needed : {3})\"\n assert fft_samples.shape[1] == nfft, f\"Only same nfft length for the moment\"\n assert fft_samples.dtype == np.complex128, \\\n f\"Wrong ndarray dtype (actual : {fft_samples.dtype}, needed : {np.complex128})\"\n\n \"\"\"return np.real(np.apply_along_axis(lambda fft_values: scipy.ifft(fft_values, n=nfft), 2, fft_samples)) \\\n .reshape(fft_samples.shape[0], -1)\"\"\"\n fft_samples = fft_samples.transpose((0, 2, 1)).reshape(-1, nfft)\n return np.real(np.apply_along_axis(lambda fft_values: scipy.ifft(fft_values, n=nfft), 1, fft_samples))\n\n\n###############\n# Main\n###############\n\ndef convert_mp3_to_wav(root_dir: str, out_dir: str, limit: int) -> None:\n cpt = 0\n if not exists(out_dir):\n mkdir(out_dir)\n for dirname, dirnames, filenames in tqdm(walk(root_dir)):\n for filename in filenames:\n if splitext(filename)[-1] == \".mp3\":\n subprocess.call([\"ffmpeg\", \"-v\", \"0\", \"-i\", join(dirname, filename),\n \"-y\", \"-ar\", \"44100\", join(out_dir, basename(filename) + \".wav\")])\n cpt += 1\n if cpt >= limit >= 0:\n return\n\n\ndef convert_mp3_to_wav_2(root_dir: str, out_dir: str, limit_per_dir: int = 1) -> None:\n cpt = {}\n for dirname, dirnames, filenames in tqdm(walk(root_dir)):\n for filename in filenames:\n if splitext(filename)[-1] == \".mp3\":\n if dirname not in cpt:\n cpt[dirname] = 0\n if cpt[dirname] < limit_per_dir:\n subprocess.call([\"ffmpeg\", \"-v\", \"0\", \"-i\", join(dirname, filename),\n \"-y\", \"-ar\", \"44100\", join(out_dir, basename(filename) + \".wav\")])\n cpt[dirname] += 1\n\n\ndef copy_mp3_2(root_dir: str, out_dir: str, limit_per_dir: int = 1) -> None:\n cpt = {}\n for dirname, dirnames, filenames in tqdm(walk(root_dir)):\n for filename in filenames:\n if splitext(filename)[-1] == \".mp3\":\n if dirname not in cpt:\n cpt[dirname] = 0\n if cpt[dirname] < limit_per_dir:\n subprocess.call([\"cp\", join(dirname, filename), out_dir])\n cpt[dirname] += 1\n\n\ndef __read_wavs_without_copy(wav_root: str, nb_wav: int, sample_rate: int, n_fft: int, sec: int) -> th.Tensor:\n wav_files = [join(wav_root, f) for f in listdir(wav_root) if splitext(f)[-1] == \".wav\"]\n\n shuffle(wav_files)\n\n if len(wav_files) > nb_wav:\n wav_files = wav_files[:nb_wav]\n\n n_channel = n_fft * 2\n fft_split_size = sample_rate * sec // n_fft\n n_sample = 0\n\n for w in tqdm(wav_files):\n n_sample += compute_wav_size(w, sec)[0]\n\n data = th.zeros(n_sample, n_channel, fft_split_size, dtype=th.float)\n\n curr_split = 0\n\n for w in tqdm(wav_files):\n _, raw_audio = open_wav(w, sec)\n fft_audio = fft_raw_audio(raw_audio, n_fft).transpose((0, 2, 1))\n\n data[curr_split:curr_split + fft_audio.shape[0], :n_fft, :] = \\\n th.tensor(np.real(fft_audio), dtype=th.float)\n data[curr_split:curr_split + fft_audio.shape[0], n_fft:, :] = \\\n th.tensor(np.imag(fft_audio), dtype=th.float)\n\n curr_split += fft_audio.shape[0]\n\n return data\n\n\ndef main() -> None:\n parser = argparse.ArgumentParser(\"Read audio main\")\n\n subparser = parser.add_subparsers(dest=\"mode\")\n subparser.required = True\n\n test_parser = subparser.add_parser(\"test\")\n test_parser.add_argument(\"-i\", \"--input-audio\", type=str, required=True, dest=\"input_audio\")\n test_parser.add_argument(\"--split-length\", type=int, default=10000, dest=\"split_length\")\n\n process_parser = subparser.add_parser(\"process\")\n process_parser.add_argument(\"--mp3-root\", type=str, dest=\"audio_root\", required=True)\n process_parser.add_argument(\"-o\", \"--out-dir\", type=str, dest=\"out_dir\", required=True)\n process_parser.add_argument(\"-l\", \"--limit\", type=int, default=100)\n\n save_parser = subparser.add_parser(\"save\")\n save_parser.add_argument(\"--wav-root\", type=str, dest=\"wav_root\", required=True)\n save_parser.add_argument(\"-n\", \"--nb-wav\", type=int, default=400, dest=\"nb_wav\")\n save_parser.add_argument(\"-o\", \"--out-tensor-file\", type=str, dest=\"out_tensor_file\", required=True)\n save_parser.add_argument(\"--nfft\", type=int, dest=\"n_fft\", required=True)\n save_parser.add_argument(\"--sample-rate\", type=int, default=44100, dest=\"sample_rate\")\n save_parser.add_argument(\"-s\", \"--seconds\", type=int, required=True, dest=\"seconds\")\n\n args = parser.parse_args()\n\n if args.mode == \"test\":\n input_audio = args.input_audio\n\n # in miliseconds\n split_length = args.split_length\n\n if splitext(input_audio)[-1].lower() != \".wav\":\n parser.error(\"input-audio must be a WAV audio file !\")\n exit()\n\n sampling_rate, data = wavfile.read(input_audio)\n\n print(f\"{sampling_rate} {data.shape}\")\n print(f\"{int(data.shape[0] / sampling_rate / 60)}min{int(data.shape[0] / sampling_rate) % 60}s\")\n\n split_size = sampling_rate * split_length / 1000.\n nb_split = ceil(data.shape[0] / split_size) - 1\n\n splitted_audio = np.asarray(np.split(data[:int(split_size * nb_split)], nb_split))\n\n if len(splitted_audio.shape) == 3 and splitted_audio.shape[2] == 2:\n print(\"Stereo WAV\")\n print(f\"actual shape = {splitted_audio.shape}, dtype = {splitted_audio.dtype}\")\n else:\n print(\"Mono WAV - unsuported\")\n exit()\n\n int_size = splitted_audio.itemsize * 8.\n\n splitted_audio = splitted_audio.astype(np.float16)\n\n splitted_audio[:, :, 0] = splitted_audio[:, :, 0] / (2. ** int_size) * 2.\n splitted_audio[:, :, 1] = splitted_audio[:, :, 1] / (2. ** int_size) * 2.\n\n splitted_audio = splitted_audio.mean(axis=2)\n\n print(splitted_audio.shape)\n print(splitted_audio.max())\n print(splitted_audio.min())\n\n nperseg = 128\n\n fft_audio = fft_raw_audio(splitted_audio, nperseg)\n\n print(f\"FFT shape {fft_audio.shape}, max {fft_audio.max()}, min {fft_audio.min()}\")\n\n new_raw_audio = ifft_samples(fft_audio.transpose((0, 2, 1)), nperseg)\n\n wavfile.write(\"test.wav\", sampling_rate, new_raw_audio.reshape(-1))\n\n elif args.mode == \"process\":\n if not exists(args.out_dir):\n mkdir(args.out_dir)\n if exists(args.out_dir) and isfile(args.out_dir):\n print(f\"{args.out_dir} already exists and is a file.\")\n exit()\n\n convert_mp3_to_wav(args.audio_root, args.out_dir, args.limit)\n elif args.mode == \"save\":\n data = __read_wavs_without_copy(args.wav_root, args.nb_wav, args.sample_rate, args.n_fft, args.seconds)\n th.save(data, args.out_tensor_file)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"Ipsedo/MusicAutoEncoder","sub_path":"read_audio.py","file_name":"read_audio.py","file_ext":"py","file_size_in_byte":10303,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"6482735452","text":"def solution(wallpaper):\n answer = []\n row, col = len(wallpaper), len(wallpaper[0])\n \n files = []\n x, y, mx, my = 51, 51, 0, 0\n for i in range(row):\n for j in range(col):\n if wallpaper[i][j] == '#':\n files.append((i, j))\n x = min(x, i)\n y = min(y, j)\n mx = max(mx, i+1)\n my = max(my, j+1)\n # print(files)\n return [x, y, mx, my]","repo_name":"maltepoo/algorithm","sub_path":"프로그래머스/unrated/161990. 바탕화면 정리/바탕화면 정리.py","file_name":"바탕화면 정리.py","file_ext":"py","file_size_in_byte":444,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"41308164021","text":"def isAnagram(s,t):\n size1 = len(s)\n size2 = len(t)\n if size1 != size2:\n return 0\n\n\n s = sorted(s)\n t = sorted(t)\n\n for i in range(0, size1):\n if s[i] != t[i]:\n return 0\n return 1\n\ns = str(input())\nt = str(input())\nif isAnagram(s, t):\n print(\"Yes\")\nelse:\n print(\"No\") \n\n\n\n\n","repo_name":"zhaxylykova/lab2-TSIS3","sub_path":"j.py","file_name":"j.py","file_ext":"py","file_size_in_byte":328,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39526170734","text":"name = 'doron'\nage = 22\nis_male = True\nhobbies = ['Piano', 'Reading']\nprint(name, age)\nprint(hobbies)\nprint(hobbies[1])\n\nuser = {\n 'name': 'doron',\n 'age': 23\n}\nprint(user['name'])\nx = 1\ny = 2\nif x==1:\n print('Im the first one')\n print(x+y)\n print(x!=y)","repo_name":"brun0be/2812-class","sub_path":"Class 1.py","file_name":"Class 1.py","file_ext":"py","file_size_in_byte":268,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"1636331756","text":"import random\n\ndef constant(inputs):\n return 1\n\ndef balanced(inputs):\n return sum(inputs) % 2\n\ndef get_binary(n, l):\n if n == 0:\n return l\n else:\n if len(l) == 0:\n return get_binary(n-1, [\"0\", \"1\"])\n else:\n return get_binary(n-1, [i + \"0\" for i in l] + [i + \"1\" for i in l])\n\ndef get_bits(n):\n strings = get_binary(n, [])\n ret = [[int(i) for i in j] for j in strings]\n return ret\n\ndef deutsch_jozsa_classical(inputs, f):\n init = f(inputs[0])\n for i in range(1, len(inputs)//2+1):\n if f(inputs[i]) != init:\n return 1\n return 0\n\n\nN = 9 # NUMBER OF BITS\ninputs = get_bits(N)\n\n\nif random.random() < 0.5:\n f = constant\n correct = \"CONSTANT\"\nelse:\n f = balanced\n correct = \"BALANCED\"\n\nresult = deutsch_jozsa_classical(inputs, f)\nif result == 1:\n print(\"BALANCED, actual:\", correct)\nelse: \n print(\"CONSTANT, actual:\", correct)\n","repo_name":"lockwo/quantum_computation","sub_path":"classic/classic_deutsch_jozsa.py","file_name":"classic_deutsch_jozsa.py","file_ext":"py","file_size_in_byte":929,"program_lang":"python","lang":"en","doc_type":"code","stars":56,"dataset":"github-code","pt":"79"} +{"seq_id":"9678479498","text":"from flask import Flask, request, render_template, jsonify, redirect, url_for\nimport speech_recognition as sr\nfrom gtts import gTTS\nimport pickle\nimport numpy as np\n# import cv2 # Certifique-se de ter a biblioteca OpenCV instalada\n\napp = Flask(__name__)\n## carregando o modelo para o código\n# with open(\"reconehcimentoVoz.pickle\", \"rb\") as file:\n# modelo_carregado = pickle.load(file)\nmodelo = ''\n\n# Função para pré-processar o espectrograma para o modelo\ndef preprocess_spectrogram_for_model(spectrogram):\n try:\n # Redimensione o espectrograma para o tamanho esperado pelo modelo\n input_size = (model_input_height, model_input_width) # Substitua pelos valores reais\n preprocessed_spectrogram = resize_spectrogram(spectrogram, input_size)\n\n # Realize a normalização, se necessário\n preprocessed_spectrogram = normalize_spectrogram(preprocessed_spectrogram)\n\n return preprocessed_spectrogram\n except Exception as e:\n raise Exception(f\"Erro ao pré-processar o espectrograma para o modelo: {str(e)}\")\n\n# Função para redimensionar o espectrograma para o tamanho esperado\ndef resize_spectrogram(spectrogram, target_size):\n try:\n # Redimensione o espectrograma usando bibliotecas como OpenCV\n # Implemente a lógica de redimensionamento aqui\n # Certifique-se de redimensionar corretamente o espectrograma para o tamanho esperado pelo modelo\n preprocessed_spectrogram = cv2.resize(spectrogram, target_size, interpolation=cv2.INTER_LINEAR)\n\n return preprocessed_spectrogram\n except Exception as e:\n raise Exception(f\"Erro ao redimensionar o espectrograma: {str(e)}\")\n\n# Função para normalizar o espectrograma, se necessário\n# Função para normalizar o espectrograma, se necessário\ndef normalize_spectrogram(spectrogram):\n try:\n # Implemente a lógica de normalização aqui\n # Você pode realizar normalização min-max, z-score, ou outro método apropriado\n # Certifique-se de que o espectrograma esteja normalizado no intervalo esperado pelo modelo\n\n # Exemplo de normalização min-max:\n min_value = np.min(spectrogram)\n max_value = np.max(spectrogram)\n normalized_spectrogram = (spectrogram - min_value) / (max_value - min_value)\n\n return normalized_spectrogram\n except Exception as e:\n raise Exception(f\"Erro ao normalizar o espectrograma: {str(e)}\")\n\ndef preprocess_audio(audio_data):\n try:\n # Converta o áudio em um espectrograma usando a biblioteca librosa\n y, sr = librosa.load(io.BytesIO(audio_data), sr=None)\n spectrogram = librosa.feature.melspectrogram(y, sr=sr)\n # Realize o pré-processamento adicional, como normalização, redimensionamento, etc.\n return spectrogram\n except Exception as e:\n raise Exception(f\"Erro ao pré-processar o áudio: {str(e)}\")\n\n\n\n# Função para fazer previsões com base no modelo carregado\ndef model_predict(spectrogram):\n try:\n # Realize qualquer processamento adicional necessário para alimentar o espectrograma no modelo\n # Suponha que você tenha um modelo carregado em 'modelo_carregado'\n # Certifique-se de que o modelo esperado esteja configurado corretamente\n # Por exemplo, converta o espectrograma para o formato de entrada esperado pelo modelo\n input_data = preprocess_spectrogram_for_model(spectrogram)\n\n # Faça previsões com base no modelo\n predictions = modelo_carregado.predict(input_data)\n\n # Retorne as previsões\n return predictions\n except Exception as e:\n raise Exception(f\"Erro ao fazer previsões: {str(e)}\")\n\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n if request.method == \"POST\":\n if \"audio\" in request.files:\n try:\n # Receba um arquivo de áudio do formulário\n audio_file = request.files[\"audio\"]\n\n # Realize a conversão de áudio para texto usando o modelo\n audio_data = audio_file.read()\n\n # Faça o pré-processamento dos dados de áudio, converta para espectrograma, etc.\n spectrogram = preprocess_audio(audio_data)\n\n # Faça previsões usando o modelo carregado\n predictions = model_predict(spectrogram)\n\n return render_template('result2.html',transcript=predictions)\n except Exception as e:\n return jsonify({\"error\": str(e)})\n\n return render_template('index.html')\n\n@app.route('/result', methods=['POST'])\ndef result():\n if request.form.get('audio'):\n audio_file = request.form['audio']\n transcribed_text = transcribe_audio(audio_file)\n return render_template('result.html',result=transcribed_text)\n\n elif request.form.get('text'):\n text = request.form['text']\n tts = gTTS(text)\n tts.save('static/output.mp4') # Salve o arquivo no diretório 'static'\n return render_template('result.html', result=\"Áudio gerado\")\n\nif __name__ == '__main__':\n app.run(debug=True)\n\n\n# transcribed_text = transcribe_audio(audio_file)\n# def transcribe_audio(audio_file):\n# try:\n# transcript = recognizer.recognize_google(audio_data) # Usando o reconhecimento de fala do Google\n# transcript = model\n\n# return transcript\n# except sr.UnknownValueError:\n# return \"Não foi possível reconhecer a fala\"\n# except sr.RequestError as e:\n# return f\"Erro ao se conectar ao serviço de reconhecimento de fala: {e}\"","repo_name":"SayuryKato/reconhecimento-de-voz","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":5586,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"604291739","text":"n = int(input())\nid = []\nfor i in range(n):\n id.append(int(input()))\nd = [0]+[-1]*(n-1)\np = [0]+[-1]*(n-1)\nin_MST = [False]*n\n\nfor i in range(n):\n u = d.index(max(d))\n d[u] = -1\n in_MST[u] = True\n for v in range(n):\n if (not in_MST[v]) and (id[u]^id[v] > d[v]):\n d[v] = id[u]^id[v]\n p[v] = u\nans = 0\nfor i,v in enumerate(p[1:]):\n ans += id[i+1]^id[v]\nprint(ans)","repo_name":"SupanutT/hardest-courses-i-have-learned","sub_path":"Algorithm/a60a_q3_xor_road/a60a_q3_xor_road.py","file_name":"a60a_q3_xor_road.py","file_ext":"py","file_size_in_byte":408,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"21801908784","text":"from service.dataSerializer.IDataSerializer import IDataSerializer\nimport logging\nimport json\nfrom model.PurchaseOrder import ObjPurchaseOrder\nfrom model.PurchaseOrderDetail import ObjPurchaseOrderDetail\nfrom model.database import db\nfrom sqlalchemy.sql import func\nfrom marshmallow import ValidationError\nfrom sqlalchemy.exc import SQLAlchemyError\nfrom utils.ApiResponse import *\n\nlogger = logging.getLogger(__name__)\n\n\nclass CategoriesAmount(IDataSerializer):\n\n def __init__(self, ft, projectCode):\n self.filter = ft\n self.projectCode = projectCode\n\n def serialize(self):\n \"\"\"\n SELECT \"PURCHASE_ORDER_DETAIL\".category,\n sum(\"PURCHASE_ORDER_DETAIL\".po_line_amount) AS amount,\n count(\"PURCHASE_ORDER\".po_no) AS po_count\n FROM \"PURCHASE_ORDER_DETAIL\"\n JOIN \"PURCHASE_ORDER\" ON \"PURCHASE_ORDER\".po_no = \"PURCHASE_ORDER_DETAIL\".po_no\n GROUP BY \"PURCHASE_ORDER_DETAIL\".category\n \"\"\"\n try:\n result = db.session.query(ObjPurchaseOrderDetail.category,\n func.sum(ObjPurchaseOrderDetail.po_line_amount).label(\"amount\"),\n func.count(ObjPurchaseOrder.po_no).label(\"count\")) \\\n .join(ObjPurchaseOrder, ObjPurchaseOrder.po_no == ObjPurchaseOrderDetail.po_no) \\\n .filter(*self.filter) \\\n .group_by(ObjPurchaseOrderDetail.category)\n\n total_amount = 0\n total_count = 0\n\n for res in result:\n total_amount = total_amount + res.amount\n total_count = total_count + res.count\n\n items = []\n for res in result:\n data = {'category': res.category,\n 'amount': float(res.amount),\n 'amount_percentage': float(res.amount / total_amount),\n 'count': res.count,\n 'count_percentage': float(res.count / total_count)}\n items.append(data)\n except ValidationError as err:\n items = ApiResponse.emitErrorOutput(E_VALIDATION_ERROR, err.messages, \"DatabaseGetHeaderService\")\n except SQLAlchemyError as e:\n db.session.rollback()\n items = ApiResponse.emitErrorOutput(E_SQLALCHEMY_ERROR, str(e), \"DatabaseGetHeaderService\")\n return json.dumps(items)\n","repo_name":"christinelin8102/wimmp-backend","sub_path":"service/dataSerializer/dailyManagement/categories/CategoriesAmount.py","file_name":"CategoriesAmount.py","file_ext":"py","file_size_in_byte":2380,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39324450423","text":"# -*- coding: utf-8 -*-\n#\n\nif False:\n from gluon import *\n import db\n request = current.request\n response = current.response\n session = current.session\n cache = current.cache\n T = current.T\n\nsession.forget(response)\n\nif IS_MOBILE:\n response.top_line = None\nelse:\n response.top_line = DIV(\n TAG.center(T('Здесь Вы можете купить биткоины и оплатить биткоинами разные услуги')),\n _id=\"top_line\")\n\ndef ua(h, url, cls='col-sm-4',\n clsA='ll-blue-bgc'):\n #clsA='gray-bgc'):\n return DIV(A(h, _class='block col-center button btn10 ' + clsA,\n #_style='background-color:%s;' % SKIN['bg-clr-ftr-mnu']\n _href=url,\n )\n , _class=cls)\n\nclass Jammer():\n def read(self,n): return 'x'*n\ndef jam():\n return 'be-'*400\n # return response.stream(Jammer(),4) - that suspend response!\n\ndef consult():\n h = CAT(\n DIV(\n H1('Консультируем бизнес и инвесторов по bitcoin и blockchain'),\n H2('Как использовать биткоин в международных платежах и бизнесе'),\n P('Осуществляем консультации для бизнеса по использованию биткоин платежей и по использованию блокчейн-технологий. Правовое регулирование и применение бухгалтерского учёта для биткоинов. Например:'),\n H3('Как купить биткоин по безналу'),\n H3('Как продать биткоины за безнал'),\n H2('Биткоин-стартапы и блокчейн-стартапы для инвесторов'),\n P('Для инвесторов проведём предварительную экспертизу по биткоин-стартапу или блокчейн-стартапу.',),\n P('Свяжитесь', ' ', A(B('с нами'), _href=URL('default','contacts'))),\n _style='padding:10px 10%;',\n _class='container'),\n )\n return dict(h=h)\n\ndef up():\n h = CAT(\n DIV(\n H1('Глобальный инновационный проект',':'),\n P(\n A('Блокчейн для бизнеса, государства и жизни',\n _href='http://datachains.world', _target='_blank'),\n ' - ', 'приглашаем инвесторов и сторонников'),\n BR(),\n H1('Биткоин-бизнес, стартапы в биткоин экономике:'),\n P('Ищем инвестиции или предпринимателей для развития биткоин стартапов в международный бизнес'),\n UL(\n LI(B(DOMEN), ' - оплата услуг биткоинами, обмен и продажа криптовалют'),\n LI(B('LITE.cash'), ' - платежный шлюз для приёма криптовалют на сайтах'),\n LI(B('WAGERS.win'), ' - служба по созданию своих споров и пари со ставками в биткоинах'),\n ),\n P('Описание смотрите в ',A('бизнес-концепции',\n _href='https://docs.google.com/document/d/1OyDbOruXBc7rpJAhWVf3Jv5-sEVWX5gGR0f1byT1faI/edit?usp=sharing', _target='_blank')),\n _style='padding:10px 10%;',\n _class='container'),\n )\n return dict(h=h)\n\ndef subscribe():\n form = SQLFORM(db.news_descrs, fields = ['email'],\n submit_button = T('Подписаться'),\n labels = {'email': T('Ваш емэйл') },\n formstyle='divs',\n )\n if form.accepts(request.vars, session):\n response.flash = T('Вы подписаны')\n elif form.errors:\n response.flash = T('ОШИБКА!')\n\n return locals()\n\ndef to_many():\n response.title = T(\"Автоматические выплаты на запрлатные счета биткоинов, догикоинов и других криптовалют\")\n response.subtitle = ' '\n h = CAT(DIV(\n H2(T('Зарплатные счета и автовыплаты на них')),\n P(T('Если Вам надо делать выплаты на много различных счетов по заранее заданным множителям или долям, то лучше всего это сделать с помощью “разделяющего счёта” - все поступившие на него платежи будут автоматически разделены на заданные доли и выплачены на соответствующие счета. Такой разделяющий или зарплатный счёт создаётся один раз на сервисе LITE.cash и потом все поступления на него автоматически распределяются между получателям на ихние кошельки криптовалют')),\n P(\n A('LITE.cash divided payouts', _href='http://lite.cash/bs3b/more/divided_payments', _target='_blabk')\n ),\n _class='container'),\n )\n return dict(h=h)\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef crypto_trans():\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n #response.logo2 = IMG(_src=URL('static','images/slide3.png'), _width=256)\n response.title = T(\"Бесплатные переводы СНГ без посредников\")\n response.subtitle = ' '\n response.top_line = DIV(\n TAG.center(T(\"Как перевести деньги в СНГ и другие страны без посредников?\")),\n _id=\"top_line\")\n h = CAT()\n h += DIV(H1(T('ПРИВЕТ')),\n ua(T('Купить'),URL('to_buy', 'index')),\n ua(T('Оплатить'),URL('deal', 'index')),\n ua(T('Обменять'),URL('to_wallet', 'index')),\n ua(T('Начало'),URL('default', 'index'),'col-sm-12'),\n _class='row')\n return dict(h = h)\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef crypto():\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n response.logo2 = IMG(_src=URL('static','images/slide3.png'), _width=256)\n response.title = T(\"Что такое криптовалюта и биткоин?\")\n response.subtitle = ' '\n h = CAT()\n h += DIV(\n ua(T('Купить'),URL('to_buy', 'index')),\n ua(T('Оплатить'),URL('deal', 'index')),\n ua(T('Обменять'),URL('to_wallet', 'index')),\n ua(T('Начало'),URL('default', 'index'),'col-sm-12'),\n _class='row')\n return dict(h = h, ua = ua(T('Да, мне интересно попробовать криптовалюту'), URL('crypto_go'), 'col-sm-10'))\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef crypto_go():\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n response.title = T(\"Как начать добывать биткоин и догикоин?\")\n response.subtitle = ' '\n\n h = CAT()\n h += DIV(\n ua(T('Купить'),URL('to_buy', 'index')),\n ua(T('Оплатить'),URL('deal', 'index')),\n ua(T('Обменять'),URL('to_wallet', 'index')),\n ua(T('Начало'),URL('default', 'index'),'col-sm-12'),\n _class='row')\n return dict(h = h)\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.disk, vars=False, public=False, lang=True)\ndef contacts():\n session.forget(response)\n response.title=T('CONTACTS')\n response.not_show_func = True\n h = CAT(\n H3('ICREATOR'),\n DIV(\n ua(T('GitLab'), 'https://gitlab.com/users/d.ermolaev'),\n ua(T('Source Code'), 'https://github.com/icreator/7pay_in'),\n ua(T('Icreator Site'), 'http://icreator.site/'),\n )\n )\n return dict(h =h)\n\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef index():\n session.forget(response)\n\n if db(db.currs).isempty():\n return \"init DB first: \" + URL(\"tools_dc\", \"init_db_records\")\n\n users = db(db.deal_accs).count()\n\n stats = []\n sum_ = db.currs_stats.count_.sum()\n recs = db(\n ## db.currs.id == db.currs_stats.curr_id).select(sum_, db.currs.ALL, groupby=db.currs_stats.curr_id, orderby=~sum_) ## on mySQL work\n db.currs.id == db.currs_stats.curr_id).select(sum_, db.currs.ALL, groupby=db.currs.id, orderby=~sum_) ## good on PostgreSQL\n\n for r in recs:\n #print r._extra\n #print r._extra['SUM(\"currs_stats\".\"count_\")']\n try:\n uses = r._extra['SUM(currs_stats.count_)']\n except:\n try:\n uses = r._extra['SUM(\"currs_stats\".\"count_\")']\n except:\n uses = r._extra['SUM(`currs_stats`.`count_`)']\n r.currs.update_record(uses = uses)\n stats.append('%s: %s' % (r.currs.name, uses))\n\n h = CAT(\n CAT(DIV(H1('Отличия нашего сервиса', _class='center'), _style='color:steelblue;background-color:gainsboro;',\n _class='row m-0'\n ) if 'short' in request.vars else DIV(\n DIV(\n H1(T('Биткоины купить и потратить'), _class='wb-headline wb-white wb-mt-0 '),\n H1(T('Oплатить биткоинами услуги'), _class='wb-headline wb-white wb-mt-0 '),\n H1(T('Обменять битоины на рубли'), _class='wb-headline wb-white wb-mt-0 '),\n H1(T('Это сервис оплат криптовалютами'), _class='wb-headline- wb-white wb-mt-0 '),\n A(H1('- bitcoin, litecoin, dogecoin...',_class='wb-headline- wb-white wb-mt-0 '),\n _href=URL('seo','index'), _class='lite'),\n _style='bottom:100px;right:10px;',\n _class=IS_MOBILE and 'right col-sm-12' or 'absolute right col-md-10 col-lg-8'), # wb-join- wb-grid-'),\n _style='min-height:350px;background-color:blue;',\n _class='row m-0 relative' + (not IS_LOCAL and ' wb-img' or '')\n )), # if False else '',\n\n DIV(\n DIV(CENTER(H1(T('Наши преимущества'))),\n DIV(\n TAG.center(\n H2(T('Надежность')),\n P(T('Наш сервис работает с 2013 года и зарекомендовал себя с хорошей стороны. Отзывы можно посмотреть на независимом форуме о криптовалютах')),\n A(T('Посмотреть отзывы и обсуждения'), _href='https://bitcointalk.org/index.php?topic=307648.0', _target='_blank',\n _class='button blue-bgc'), # lightblue-bgc\n _style='padding: 30px;',\n _class='col-sm-6'),\n TAG.center(\n H2(T('Открытость')),\n P(T('Вы можете посмотреть свои платежи по адресу криптовалюты, а так же недавние платежи других пользователей нашего сервиса в обезличенном виде')),\n A(T('Посмотреть платежи'), _class='button blue-bgc', _href=URL('where','index'),\n ),\n _style='padding: 30px;',\n _class='col-sm-6'),\n _class='row m-0'),\n DIV(\n P(T('Количество обработанных транзакций за время работы нашей службы') + ': ', ', '.join(stats)),\n H2(T('Нам доверяют уже %s+ человек со всего мира') % (users + 500), _class='center'),\n _class='row m-0'),\n _class='container'),\n _style='background-color:#00188F; color:#ddd;padding-bottom:30px;',\n _class='row m-0'),\n\n DIV(\n DIV(\n H2(T('Защита персональных данных'), _class='blue-c'),\n UL([T('Наш сервис не требует регистраций'), T('Мы не запрашиваем паролей'), T('Ваши средства не хранятся у нас'),\n T('Ваши платежи обезличены и анонимны')]),\n _class='col-sm-6'),\n _style='color:#666;' + (not IS_LOCAL and 'background-image: url(' + URL('static','images/bg-p/lifestyle_girl_tablet_1x.jpg') + ');' or ''),\n _class='row m-0 wb-img-2bbg'),\n DIV(\n DIV(\n H1(\n T('Пресса о нас'),\n _class=\"center\"),\n DIV(\n DIV(\n IMG(_src=URL('static','images/logos/coinTelegram.png'), _width=60),' ',SPAN('The Cointelegraph',\n _style='font-size: 24px; margin: 0; text-transform: uppercase; letter-spacing: 0.12em; color: #ffcd04;font-weight: 700; padding-bottom: 3px;'),\n H3(UL(A('Russians Can Pay Internet and Phone Bills with Bitcoin without Fees',\n _href='http://cointelegraph.com/news/116026/russians-can-pay-internet-and-phone-bills-with-bitcoin-without-fees' ,\n _target='_blank',\n _class='lite'),\n )),\n H3(UL(A('EmerCoin Fever Quickly Making a Name Among Fintech Projects',\n _href='http://cointelegraph.com/news/emercoin-fever-quickly-making-a-name-among-fintech-projects' ,\n _target='_blank',\n _class='lite'),\n )),\n _class='col-sm-6'),\n DIV(\n IMG(_src=URL('static','images/logos/forklog.png'), _width=200),' ',\n H3(UL(A('Bitcoin Conquers Official Moscow',\n _href='http://forklog.net/bitcoin-conquers-official-moscow/' ,\n _target='_blank',\n _class='lite'),\n A('Как биткоин покоряет официальную Москву',\n _href='http://forklog.com/kak-bitkoin-pokoryaet-ofitsialnuyu-moskvu/',\n _target='_blank',\n _class='lite'),\n )),\n _class='col-sm-6'),\n _class='row', _style='margin-top:20px;'),\n DIV(\n DIV(\n IMG(_src=URL('static','images/logos/slon.svg'), _width=200),' ',\n H3(A('Московский Bit: кто зарабатывает на биткоинах в России?',\n _href='https://slon.ru/economics/moskovskiy_bit_kto_zarabatyvaet_na_bitkoinakh_v_rossii-1077043.xhtml' ,\n _target='_blank',\n _class='lite'),\n ' ','Здесь сайт еще со старым интерфейсом, но его можно узнать по логотипу \"7Pay.in\"',\n ),\n _class='col-sm-6'),\n DIV(\n IMG(_src=URL('static','images/logos/tvrain.png'), _width=200),' ',\n H3(A('Московский бит. Краткий путеводитель по криптовалютам в России',\n _href='https://tvrain.ru/teleshow/reportazh/moskovskij_bit_kratkij_putevoditel_po_kriptovaljutam_v_rossii-366325/',\n _target='_blank',\n _class='lite'),\n ),\n _class='col-sm-6'),\n _class='row', _style='margin-top:20px;'),\n DIV(\n DIV(\n _class='col-sm-6'),\n DIV(\n _class='col-sm-6'),\n _class='row', _style='margin-top:20px;'),\n _class='container'),\n _style='background-color:#00188F; color:#ddd;padding-bottom:30px;',\n _class='row m-0 bg-info'),\n DIV(\n DIV(\n H2(T('Удобно платить повторно'), _class='blue-c'),\n P(T('Повторно оплатить услугу можно простым переводом биткоинов тот же адрес, который Вы уже получили на нашем сервисе для оплаты данной услуги. В Вашем кошельке просто присвойте метку для данного адреса чтобы потом по ней в Вашем кошельке найти адрес для оплаты данной услуги или дела. Теперь, не заходя на наш сервис, Вы можете делать повторные оплаты разных дел и услуг из Вашего кошелька криптовалюты.')),\n _class='col-sm-5 pull-right'),\n _style='color:#666;' + (not IS_LOCAL and 'background-image: url(' + URL('static','images/bg-p/lifestyle_guy_computer_1x-2.jpg') + ');' or ''),\n _class='row m-0 wb-img-2bbg'),\n DIV(\n DIV(\n #TAG.center(\n # P('Так же Вы можете купить биткоины лайткоины догикоины на нашем сайте', _class='blue-c'),\n # A(H1(T('Купить биткоины')), _href=URL('to_buy','index'),\n # _class='block button blue-bgc'),\n # _style='padding:10px;',\n # _class='col-sm-6'),\n #TAG.center(\n # P('Быстро оплатить биткоинами услуги ЖКХ, за свет, интернет, игры, телефон', _class='blue-c'),\n # A(H1(T('Оплатить биткоинами')), _href=URL('deal','index'),\n # _class='block button blue-bgc'),\n # _style='padding:10px;',\n # _class='col-sm-6'),\n _class='container'),\n _style='background-color:currentColor;color:#ddd;',\n _class='row m-0 bg-info'),\n DIV(\n DIV(\n #TAG.center(\n # P('Или Вы можете оплатить сотовый телефон и мобильный интернет', _class='blue-c'),\n # A(H1(T('Пополнить сотовый телефон биткоинами')), _href=URL('to_phone','index'),\n # _class='block button blue-bgc'),\n # P('Так же Вы можете получить от нас консультацию об использовании биткоинов в бизнесе', _class='blue-c'),\n # A(H1(T('Консультация для бизнеса и инвесторов о биткоинах')), _href=URL('default','consult'),\n # _class='block button blue-bgc'),\n # _style='padding:10px;',\n # _class='col-sm-12'),\n _class='container'),\n #_style='background-color:currentColor;color:#ddd;',\n _class='row m-0 bg-info'),\n )\n\n return dict(h = h)\n\n\n##@cache.action()\ndef download():\n session.forget(response)\n \"\"\"\n allows downloading of uploaded files\n http://..../[app]/default/download/[filename]\n \"\"\"\n return response.download(request, db)\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef e_gold():\n import rates_lib, db_common\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n title = response.title = T(\"Биткоин - Настоящее цифровое золото...\")\n subtitle = response.subtitle = 'Здесь мы научим тебя как заработать на своё будущее.'\n a_in,x,e = db_common.get_currs_by_abbrev(db, 'BTC')\n a_out,x,e = db_common.get_currs_by_abbrev(db, 'RUB')\n b,s, avg = rates_lib.get_average_rate_bsa(db, a_in.id, a_out.id, None)\n rate2 = rate = avg and int(avg) or '****' #round(float(avg or 0),2)\n #a_in2,x,e = db_common.get_currs_by_abbrev(db, 'SIB')\n #b,s, avg2 = rates_lib.get_average_rate_bsa(db, a_in2.id, a_out.id, None)\n #rate2 = avg2 and round(float(avg2 or 0), 2) or -1\n\n return dict(title=title, subtitle=subtitle,\n rate = rate, rate2 = rate2 )\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef halava():\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n title=response.title = T(\"Как получить немного денег на халяву?\")\n subtitle=response.subtitle = ' '\n return dict(title=title, subtitle=subtitle)\n\n#@cache.action(time_expire=CACHE_EXP_TIME, cache_model=cache.ram, public=True, lang=True)\ndef halava2():\n session.forget(response)\n #ss = common.page_stats(db,response['view'])\n #print response['view'], ss, common.ip()\n\n title=response.title = T(\"Как получить деньги на халяву?\")\n subtitle=response.subtitle = 'от 300 до 3000 рублей в месяц!'\n return dict(title=title, subtitle=subtitle)\n","repo_name":"icreator/7pay_in","sub_path":"controllers/default.py","file_name":"default.py","file_ext":"py","file_size_in_byte":22833,"program_lang":"python","lang":"ru","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"1057104911","text":"def standraziation():\r\n import numpy as np\r\n import xlwt\r\n import xlrd\r\n import math\r\n import csv\r\n import random\r\n \r\n\r\n workbook = xlrd.open_workbook('Dymoladata_summer.xlsx')\r\n #workbook = xlrd.open_workbook('myfile.xls')\r\n sheet1 = workbook.sheet_by_name('Dymoladata_summer')\r\n y = []\r\n mean_val=[]\r\n std_val=[]\r\n for i in range(sheet1.nrows):\r\n y.append(0)\r\n for i in range(sheet1.ncols):\r\n mean_val.append(0)\r\n std_val.append(0)\r\n for index in range(0,sheet1.ncols):\r\n for index1 in range(0,sheet1.nrows):\r\n y[index1]=sheet1.cell_value(index1,index)\r\n mean_val[index]=np.mean(y)\r\n std_val[index]=np.std(y)\r\n\r\n print(mean_val)\r\n print(std_val)\r\n\r\n workbook1 = xlwt.Workbook()\r\n sheet2 = workbook1.add_sheet('standardized_data')\r\n for index1 in range(0,sheet1.nrows):\r\n for index in range(0,sheet1.ncols):\r\n value=(sheet1.cell_value(index1,index)-mean_val[index])/std_val[index]\r\n sheet2.write(index1,index,value)\r\n workbook1.save('standardized_data.xlsx')\r\n\r\n\r\n\r\n workbook2 = xlrd.open_workbook('faulty_data.xlsx')\r\n #workbook = xlrd.open_workbook('myfile.xls')\r\n sheet2 = workbook2.sheet_by_name('faulty_data')\r\n\r\n\r\n workbook3 = xlwt.Workbook()\r\n sheet3 = workbook3.add_sheet('predict_data')\r\n for index1 in range(0,sheet2.nrows):\r\n for index in range(0,sheet2.ncols):\r\n value=(sheet2.cell_value(index1,index)-mean_val[index])/std_val[index]\r\n sheet3.write(index1,index,value)\r\n workbook3.save('predict_data.xls')\r\n\r\n\r\n\r\n","repo_name":"indrasischak/Autoencoder","sub_path":"standardization.py","file_name":"standardization.py","file_ext":"py","file_size_in_byte":1635,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"36829593688","text":"n, m = list(map(int, input().split()))\narray = list(map(int, input().split()))\n\nstart = 0\nend = max(array) # 전체 array를 정렬할 필요는 없다. 최댓값만 알면 됨\n\nresult = 0\n# 이진탐색, start가 end 보다 크면 탐색이 끝났다.\nwhile start <= end:\n total = 0 # 잘린 떡의 길이\n mid = (start + end) // 2\n # 잘린 떡의 길이 구하기\n for x in array:\n if x > mid:\n total += x - mid\n\n # 떡의 양이 부족한 경우 더 많이 자르기 (왼쪽 부분 탐색)\n if total < m:\n end = mid - 1\n # 떡의 양이 충분한 경우 덜 자르기 (오른쪽 부분 탐색)\n else:\n result = mid # 최대한 덜 잘랐을 때가 정답이므로 여기에 result 기록\n start = mid + 1\n print(start, end)\n\nprint(result)","repo_name":"mini-apple/algorithm-level1","sub_path":"5.이진탐색/3-1.떡볶이떡만들기_반복문_모법답안.py","file_name":"3-1.떡볶이떡만들기_반복문_모법답안.py","file_ext":"py","file_size_in_byte":771,"program_lang":"python","lang":"ko","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"24367318650","text":"from typing import Any, Collection, Dict, List, Optional\n\nfrom ase.atoms import Atoms\n\nall_changes: List[str]\n\nclass CalculatorSetupError(RuntimeError): ...\nclass PropertyNotImplementedError(NotImplementedError): ...\n\nclass Calculator:\n results: Dict[str, Any]\n atoms: Optional[Atoms]\n def __init__(self, **kwargs: Any): ...\n def calculate(\n self,\n atoms: Optional[Atoms] = ...,\n properties: Collection[str] = ...,\n system_changes: List[str] = ...,\n ) -> None: ...\n def reset(self) -> None: ...\n","repo_name":"alexjbinnie/narupatools","sub_path":"stubs/ase/calculators/calculator.pyi","file_name":"calculator.pyi","file_ext":"pyi","file_size_in_byte":541,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"33521033735","text":"import argparse\nimport numpy\n\nimport matplotlib.pyplot as plt\n\nfrom scripts.ode.integrator import odeint\n\n\ndef system_autonomous_df(K: numpy.ndarray, Y: numpy.ndarray) -> numpy.ndarray:\n return K.dot(Y)\n\n\ndef kin_successive_df(t: float, Y: numpy.ndarray, k_1: float = .2, k_2: float = .5) -> numpy.ndarray:\n return system_autonomous_df(numpy.array([\n [-k_1, 0, 0],\n [k_1, -k_2, 0],\n [0, k_2, 0]\n ]), Y)\n\n\ndef kin_competitive_df(t: float, Y: numpy.ndarray, k_1: float = .2, k_2: float = .5) -> numpy.ndarray:\n return system_autonomous_df(numpy.array([\n [-k_1-k_2, 0, 0],\n [k_1, 0, 0],\n [k_2, 0, 0]\n ]), Y)\n\n\ndef kin_eq_df(t: float, Y: numpy.ndarray, k_1: float = .2, k_m1: float = .1) -> numpy.ndarray:\n return system_autonomous_df(numpy.array([\n [-k_1, k_m1],\n [k_1, -k_m1]\n ]), Y)\n\n\nY0 = 1\nDX = 0.1\nWIN = (0, 0, 15, 1.1)\n\n\ndef prepare_plot(ax1):\n ax1.set_xlim(WIN[0], WIN[2])\n ax1.set_ylim(WIN[1], WIN[3])\n ax1.set_xlabel('t')\n ax1.set_ylabel('C(t)')\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('-s', '--save', help='save instead of show', type=str)\n\n args = parser.parse_args()\n\n fig_eq = plt.figure(figsize=(8, 5))\n ax1 = fig_eq.add_subplot()\n prepare_plot(ax1)\n\n X = numpy.arange(0, WIN[2], DX)\n Y = odeint(\n kin_eq_df,\n (0, numpy.array([Y0, 0]).transpose()),\n X\n )\n\n ax1.plot(X, Y[:, 0], label='A(t)')\n ax1.plot(X, Y[:, 1], label='B(t)')\n ax1.legend()\n\n fig_eq.show()\n\n # Competitive & successive\n fig_sc = plt.figure(figsize=(9, 5))\n ax1 = fig_sc.add_subplot(1, 2, 1)\n prepare_plot(ax1)\n\n Y = odeint(\n kin_successive_df,\n (0, numpy.array([Y0, 0, 0]).transpose()),\n X\n )\n ax1.plot(X, Y[:, 0], label='A(t)')\n ax1.plot(X, Y[:, 1], label='B(t)')\n ax1.plot(X, Y[:, 2], label='C(t)')\n ax1.legend()\n\n ax2 = fig_sc.add_subplot(1, 2, 2)\n prepare_plot(ax2)\n\n Y = odeint(\n kin_competitive_df,\n (0, numpy.array([Y0, 0, 0]).T),\n X\n )\n\n ax2.plot(X, Y[:, 0], label='A(t)')\n ax2.plot(X, Y[:, 1], label='B(t)')\n ax2.plot(X, Y[:, 2], label='C(t)')\n ax2.legend()\n\n fig_sc.show()\n\n # SAVE\n if args.save:\n fig_eq.savefig('{}_kin_eq.pdf'.format(args.save))\n fig_sc.savefig('{}_kin_sc.pdf'.format(args.save))\n\n","repo_name":"pierre-24/SCHIB209","sub_path":"scripts/eq_kin_system.py","file_name":"eq_kin_system.py","file_ext":"py","file_size_in_byte":2403,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"70966525696","text":"import json\n\nimport numpy as np\nimport pytest\n\nimport xgboost as xgb\nfrom xgboost import testing as tm\n\ntry:\n import matplotlib\n matplotlib.use('Agg')\n from graphviz import Source\n from matplotlib.axes import Axes\nexcept ImportError:\n pass\n\npytestmark = pytest.mark.skipif(**tm.no_multiple(tm.no_matplotlib(),\n tm.no_graphviz()))\n\n\nclass TestPlotting:\n def test_plotting(self):\n m, _ = tm.load_agaricus(__file__)\n booster = xgb.train({'max_depth': 2, 'eta': 1,\n 'objective': 'binary:logistic'}, m,\n num_boost_round=2)\n\n ax = xgb.plot_importance(booster)\n assert isinstance(ax, Axes)\n assert ax.get_title() == 'Feature importance'\n assert ax.get_xlabel() == 'F score'\n assert ax.get_ylabel() == 'Features'\n assert len(ax.patches) == 4\n\n ax = xgb.plot_importance(booster, color='r',\n title='t', xlabel='x', ylabel='y')\n assert isinstance(ax, Axes)\n assert ax.get_title() == 't'\n assert ax.get_xlabel() == 'x'\n assert ax.get_ylabel() == 'y'\n assert len(ax.patches) == 4\n for p in ax.patches:\n assert p.get_facecolor() == (1.0, 0, 0, 1.0) # red\n\n ax = xgb.plot_importance(booster, color=['r', 'r', 'b', 'b'],\n title=None, xlabel=None, ylabel=None)\n assert isinstance(ax, Axes)\n assert ax.get_title() == ''\n assert ax.get_xlabel() == ''\n assert ax.get_ylabel() == ''\n assert len(ax.patches) == 4\n assert ax.patches[0].get_facecolor() == (1.0, 0, 0, 1.0) # red\n assert ax.patches[1].get_facecolor() == (1.0, 0, 0, 1.0) # red\n assert ax.patches[2].get_facecolor() == (0, 0, 1.0, 1.0) # blue\n assert ax.patches[3].get_facecolor() == (0, 0, 1.0, 1.0) # blue\n\n g = xgb.to_graphviz(booster, num_trees=0)\n assert isinstance(g, Source)\n\n ax = xgb.plot_tree(booster, num_trees=0)\n assert isinstance(ax, Axes)\n\n def test_importance_plot_lim(self):\n np.random.seed(1)\n dm = xgb.DMatrix(np.random.randn(100, 100), label=[0, 1] * 50)\n bst = xgb.train({}, dm)\n assert len(bst.get_fscore()) == 71\n ax = xgb.plot_importance(bst)\n assert ax.get_xlim() == (0., 11.)\n assert ax.get_ylim() == (-1., 71.)\n\n ax = xgb.plot_importance(bst, xlim=(0, 5), ylim=(10, 71))\n assert ax.get_xlim() == (0., 5.)\n assert ax.get_ylim() == (10., 71.)\n\n def run_categorical(self, tree_method: str) -> None:\n X, y = tm.make_categorical(1000, 31, 19, onehot=False)\n reg = xgb.XGBRegressor(\n enable_categorical=True, n_estimators=10, tree_method=tree_method\n )\n reg.fit(X, y)\n trees = reg.get_booster().get_dump(dump_format=\"json\")\n for tree in trees:\n j_tree = json.loads(tree)\n assert \"leaf\" in j_tree.keys() or isinstance(\n j_tree[\"split_condition\"], list\n )\n\n graph = xgb.to_graphviz(reg, num_trees=len(j_tree) - 1)\n assert isinstance(graph, Source)\n ax = xgb.plot_tree(reg, num_trees=len(j_tree) - 1)\n assert isinstance(ax, Axes)\n\n @pytest.mark.skipif(**tm.no_pandas())\n def test_categorical(self) -> None:\n self.run_categorical(\"approx\")\n","repo_name":"dmlc/xgboost","sub_path":"tests/python/test_plotting.py","file_name":"test_plotting.py","file_ext":"py","file_size_in_byte":3420,"program_lang":"python","lang":"en","doc_type":"code","stars":25007,"dataset":"github-code","pt":"79"} +{"seq_id":"43302145352","text":"import collections\n\n\ndef make_tree(N):\n ret = {i: {'parent': [0]*LOG, 'depth': None, 'to': []}\n for i in range(1, N+1)}\n S = tuple(map(int, input().split()))\n for i in range(0, len(S), 2):\n ret[S[i+1]]['parent'][0] = S[i]\n ret[S[i]]['to'].append(S[i+1])\n return ret\n\n\ndef setParent(tr):\n for j in range(1, LOG):\n for i in range(1, V+1):\n tmp = tr[i]['parent'][j-1]\n if tmp:\n tr[i]['parent'][j] = tr[tmp]['parent'][j-1]\n\n\ndef cal_depth(tr):\n dq = collections.deque()\n dq.append(ROOT)\n tr[ROOT]['depth'] = 0\n while dq:\n tmp = dq.popleft()\n for node in tr[tmp]['to']:\n dq.append(node)\n tr[node]['depth'] = tr[tmp]['depth'] + 1\n\n\ndef find_LCA(tr, A, B):\n if tr[A]['depth'] < tr[B]['depth']:\n A, B = B, A\n\n for i in range(LOG-1, -1, -1):\n if tr[A]['depth'] - tr[B]['depth'] >= (1 << i):\n A = tr[A]['parent'][i]\n\n if A == B:\n return A\n\n for i in range(LOG-1, -1, -1):\n if tr[A]['parent'][i] != tr[B]['parent'][i]:\n A = tr[A]['parent'][i]\n B = tr[B]['parent'][i]\n\n return tr[A]['parent'][0]\n\n\ndef cal_child_cnt(tr, node):\n ret = 0\n dq = collections.deque()\n dq.append(node)\n while dq:\n tmp = dq.popleft()\n ret += 1\n for node in tr[tmp]['to']:\n dq.append(node)\n return ret\n\n\nLOG = 14\nROOT = 1\nV, E, A, B = map(int, input().split())\ntr = make_tree(V)\nsetParent(tr)\ncal_depth(tr)\nlca = find_LCA(tr, A, B)\nprint(f'{lca} {cal_child_cnt(tr, lca)}')\n","repo_name":"hjyoon/python_ref_for_test","sub_path":"lca.py","file_name":"lca.py","file_ext":"py","file_size_in_byte":1590,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"29304248912","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('hpc', '0001_initial'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='abstract',\n name='level',\n field=models.CharField(default=b'', max_length=1, verbose_name=b'Level', choices=[(b'U', b'Undergraduate'), (b'G', b'Graduate')]),\n ),\n ]\n","repo_name":"enjaz/enjaz","sub_path":"hpc/migrations/0002_abstract_level.py","file_name":"0002_abstract_level.py","file_ext":"py","file_size_in_byte":478,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"73157790015","text":"############################################\n# @Author: Git-123-Hub\n# @Date: 2021/9/1\n# @Description: implementation of replay memory for DQN\n############################################\nimport numpy as np\nimport torch\n\n\nclass replayMemory:\n \"\"\"data structure where we store the agent's experience, and sample from them for the agent to learn\"\"\"\n\n def __init__(self, capacity, batch_size):\n assert int(capacity) > batch_size, 'capacity should be greater than batch size'\n self.capacity = int(capacity)\n self.batch_size = int(batch_size)\n\n self.state = np.zeros(int(capacity), dtype=object)\n self.action = np.zeros(int(capacity), dtype=object)\n self.reward = np.zeros(int(capacity), dtype=float)\n self.next_state = np.zeros(int(capacity), dtype=object)\n self.done = np.zeros(int(capacity), dtype=bool) # type for done\n\n self._index = 0 # current position for adding new experience\n self._size = 0 # record the number of the all the experiences stored\n self.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n def reset(self):\n \"\"\"clear all the experience that has been stored, reset the replay memory\"\"\"\n self.state.fill(0)\n self.action.fill(0)\n self.reward.fill(0)\n self.next_state.fill(0)\n self.done.fill(0)\n\n self._index = 0\n self._size = 0\n\n def add(self, state, action, reward, next_state, done):\n \"\"\" add an experience to the memory \"\"\"\n self.state[self._index] = state\n self.action[self._index] = action\n self.reward[self._index] = reward\n self.next_state[self._index] = next_state\n self.done[self._index] = done\n\n self._index = (self._index + 1) % self.capacity\n if self._size < self.capacity:\n self._size += 1\n\n def sample(self, size=None):\n if size is None: size = self.batch_size\n # sample all the current useful index without duplicate(replace=False)\n indices = np.random.choice(self._size, size=size, replace=False)\n return self.transfer_experience(indices)\n\n def transfer_experience(self, indices):\n \"\"\"transfer the data of `indices` from ndarray to tensor and change the shape of the data if necessary\"\"\"\n # retrieve data using `__getitem__`, Note that the data stored is ndarray\n states, actions, rewards, next_states, dones = self[indices]\n # NOTE that `states`, `actions`, `next_states` will be passed to network(nn.Module),\n # so the first dimension should be `batch_size`\n states = torch.from_numpy(np.vstack(states)).float().to(self.device) # torch.Size([batch_size, state_dim])\n actions = torch.from_numpy(np.vstack(actions)).float().to(self.device) # torch.Size([batch_size, action_dim])\n rewards = torch.from_numpy(rewards).float().to(self.device) # just a tensor with length: batch_size\n next_states = torch.from_numpy(np.vstack(next_states)).float().to(self.device) # Size([batch_size, state_dim])\n dones = torch.from_numpy(dones).float().to(self.device) # just a tensor with length: batch_size\n return states, actions, rewards, next_states, dones\n\n def __len__(self):\n return self._size\n\n def __getitem__(self, index):\n state = self.state[index]\n action = self.action[index]\n reward = self.reward[index]\n next_state = self.next_state[index]\n done = self.done[index]\n return state, action, reward, next_state, done\n","repo_name":"Git-123-Hub/reinforcement-learning-algorithm","sub_path":"utils/replayMemory.py","file_name":"replayMemory.py","file_ext":"py","file_size_in_byte":3546,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"35533656599","text":"\"\"\"Download replay packs via Blizzard Game Data APIs.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport os\nimport sys\nimport shutil\nimport subprocess\nimport argparse\nimport collections\nimport json\nimport requests\n\nimport mpyq\nfrom six import print_ as print # To get access to `flush` in python 2.\n\nAPI_BASE_URL = 'https://us.api.blizzard.com'\nAPI_NAMESPACE = 's2-client-replays'\n\n\ndef mkdirs(path):\n if not os.path.exists(path):\n os.makedirs(path)\n\ndef print_part(*args):\n print(*args, end=\"\", flush=True)\n\n\nclass BnetAPI(object):\n\n def __init__(self, key, secret):\n headers = {\"Content-Type\": \"application/json\"}\n params = {\n \"grant_type\": \"client_credentials\"\n }\n \n response = requests.post(\n \"https://us.battle.net/oauth/token\", headers=headers, params=params, auth=requests.auth.HTTPBasicAuth(key, secret)\n )\n\n if response.status_code != 200:\n raise Exception('Failed to get oauth access token. response={}'.format(response))\n response = json.loads(response.text)\n if 'access_token' in response:\n self._token = response['access_token']\n else:\n raise Exception('Failed to get oauth access token. response={}'.format(response))\n\n def get(self, url, params=None):\n params = params or {}\n params['namespace'] = API_NAMESPACE,\n headers = {\"Authorization\": \"Bearer \" + self._token}\n response = requests.get(url, headers=headers, params=params)\n if response.status_code != 200:\n raise Exception(\"Request to '{}' failed. response={}\".format(url, response))\n response_json = json.loads(response.text)\n if response_json.get('status') == 'nok':\n raise Exception(\"Request to '{}' failed. response={}\".format(url, response_json.get(\"reason\")))\n return response_json\n\n def url(self, path):\n return requests.compat.urljoin(API_BASE_URL, path)\n\n def get_base_url(self):\n return self.get(self.url(\"/data/sc2/archive_url/base_url\"))[\"base_url\"]\n\n def search_by_client_version(self, client_version):\n meta_urls = []\n page = 1\n while True:\n params = {\n 'client_version': client_version,\n '_pageSize': 100,\n '_page': page,\n }\n response = self.get(self.url(\"/data/sc2/search/archive\"), params)\n for result in response['results']:\n assert result['data']['client_version'] == client_version\n meta_urls.append(result['key']['href'])\n if response[\"pageCount\"] <= page:\n break\n page += 1\n return meta_urls\n\n\ndef main():\n args = parse_args()\n\n # Get OAuth token from us region\n api = BnetAPI(args.key, args.secret)\n\n # Get meta file infos for the give client version\n print('Searching replay packs with client version:', args.version)\n meta_file_urls = api.search_by_client_version(args.version)\n if len(meta_file_urls) == 0:\n sys.exit('No matching replay packs found for the client version!')\n\n # Download replay packs.\n download_base_url = api.get_base_url()\n print('Found {} replay packs'.format(len(meta_file_urls)))\n print('Downloading to:', args.download_dir)\n print('Extracting to:', args.replays_dir)\n mkdirs(args.download_dir)\n for i, meta_file_url in enumerate(sorted(meta_file_urls), 1):\n # Construct full url to download replay packs\n meta_file_info = api.get(meta_file_url)\n archive_url = requests.compat.urljoin(download_base_url, meta_file_info['path'])\n\n print_part('{}/{}: {} ... '.format(i, len(meta_file_urls), archive_url))\n\n file_name = archive_url.split('/')[-1]\n file_path = os.path.join(args.download_dir, file_name)\n\n with requests.get(archive_url, stream=True) as response:\n print_part(int(response.headers['Content-Length']) // 1024**2, 'Mb ... ')\n if (not os.path.exists(file_path) or\n os.path.getsize(file_path) != int(response.headers['Content-Length'])):\n with open(file_path, 'wb') as f:\n shutil.copyfileobj(response.raw, f)\n print_part('downloaded')\n else:\n print_part('found')\n\n if args.extract:\n print_part(' ... extracting')\n if os.path.getsize(file_path) <= 22:\n print_part(' ... zip file is empty')\n else:\n subprocess.call(['unzip', '-P', 'iagreetotheeula', '-u', '-o', '-q', '-d', args.replays_dir, file_path])\n if args.remove:\n os.remove(file_path)\n print()\n\n if args.filter_version != 'keep':\n print('Filtering replays.')\n found_versions = collections.defaultdict(int)\n found_str = lambda: ', '.join('%s: %s' % (v, c) for v, c in sorted(found_versions.items()))\n all_replays = [f for f in os.listdir(args.replays_dir) if f.endswith('.SC2Replay')]\n for i, file_name in enumerate(all_replays):\n if i % 100 == 0:\n print_part('\\r%s/%s: %d%%, found: %s' % (i, len(all_replays), 100 * i / len(all_replays), found_str()))\n file_path = os.path.join(args.replays_dir, file_name)\n with open(file_path) as fd:\n try:\n archive = mpyq.MPQArchive(fd).extract()\n except:\n found_versions['corrupt'] += 1\n os.remove(file_path)\n continue\n metadata = json.loads(archive[b'replay.gamemetadata.json'].decode('utf-8'))\n game_version = '.'.join(metadata['GameVersion'].split('.')[:-1])\n found_versions[game_version] += 1\n if args.filter_version == 'sort':\n version_dir = os.path.join(args.replays_dir, game_version)\n if found_versions[game_version] == 1: # First one of this version.\n mkdirs(version_dir)\n os.rename(file_path, os.path.join(version_dir, file_name))\n elif args.filter_version == 'delete':\n if game_version != args.version:\n os.remove(file_path)\n print('\\nFound replays:', found_str())\n\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('--key', required=True, help='Battle.net API key.')\n parser.add_argument('--secret', required=True, help='Battle.net API secret.')\n parser.add_argument('--version', required=True, help='Download all replays from this StarCraft 2 game version, eg: \"4.9.0\".')\n parser.add_argument('--replays_dir', default='./replays', help='Where to save the replays.')\n parser.add_argument('--download_dir', default='./download', help='Where to save the zip files.')\n parser.add_argument('--extract', action='store_true', help='Whether to extract the zip files.')\n parser.add_argument('--remove', action='store_true', help='Whether to delete the zip files after extraction.')\n parser.add_argument('--filter_version', default='keep', choices=['keep', 'delete', 'sort'],\n help=('What to do with replays that don\\'t match the requested version. '\n 'Keep is fast, but does no filtering. Delete deletes any that don\\'t match. '\n 'Sort puts them in sub-directories based on their version.'))\n return parser.parse_args()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"hgkahng/pysc2-replay-parser","sub_path":"download_replays.py","file_name":"download_replays.py","file_ext":"py","file_size_in_byte":7590,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"5796505690","text":"# 실버2\n# N번째 큰 수\nimport sys\nimport heapq\ninput = sys.stdin.readline\nn = int(input())\n#------------메모리 초과---------------#\n# array = [list(map(int, input().split())) for _ in range(n)]\n# result = sorted(array[-1]) # [20, 32, 41, 49, 52]\n# array = list(map(list, zip(*array)))\n# array.sort(reverse=True, key=lambda x : x[n-1])\n# for i in array:\n# for j in range(n-1, -1, -1):\n# for k in range(n-1, -1, -1):\n# if i[j] >= result[k]:\n# result[k] = i[j]\n# break\n# if i[j] <= result[0]:\n# break\n# if i[-1] <= result[0]:\n# break\n# print(result[0])\n\nheap = []\nfor _ in range(n):\n array = list(map(int, input().split()))\n for i in array:\n if len(heap) < n:\n heapq.heappush(heap, i)\n else:\n if i > heap[0]:\n heapq.heappop(heap)\n heapq.heappush(heap, i)\nprint(heap[0]) \n","repo_name":"M1nseokSong/baekjoon-code-review","sub_path":"문제집 8708/2075.py","file_name":"2075.py","file_ext":"py","file_size_in_byte":943,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8476531170","text":"def duplicateNum(arr) :\n new_arr = []\n for num in arr :\n if num in new_arr :\n continue\n else:\n new_arr.append(num)\n return new_arr\n \narray1 = [1,2,3,4,5,6,7,6,5,4,3,2,1]\nprint(duplicateNum(array1))\n\narray2 = [10,2,3,4,5,10,10,10,10,10,2,3]\nprint(duplicateNum(array2))\n\n","repo_name":"wooyong99/SeoulIctWebBase","sub_path":"practice/0731-2.py","file_name":"0731-2.py","file_ext":"py","file_size_in_byte":321,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"73818080574","text":"from django import template\nfrom django.http import request\nfrom product.models import Basket, category, Product, BasketProduct\n\nregister = template.Library()\n\n@register.inclusion_tag('nav/category.html')\ndef navbar(request):\n product_promo = Product.objects.filter(status='promo').order_by('?')[:3]\n categories = category.objects.all()\n user_count = 0\n total_rial = 0\n total_doller = 0\n if request.user.is_authenticated:\n basket = Basket.objects.filter(is_paid = False, user = request.user).first()\n if basket is not None:\n user_count = basket.basketproduct_user.count()\n total_doller = 0\n total_rial = 0\n\n for pro in basket.basketproduct_user.all():\n total_doller = total_doller + pro.product.price_end_doller() * int(pro.qty)\n total_rial = total_rial + pro.product.price_end_rial() * int(pro.qty)\n \n\n context = {\n 'product_promo':product_promo,\n 'categories':categories,\n 'user_count':user_count,\n 'total_doller':total_doller,\n 'total_rial':total_rial\n }\n\n return context","repo_name":"behzad-azadi2693/project_flower_shop_with_django","sub_path":"product/templatetags/navbar_left.py","file_name":"navbar_left.py","file_ext":"py","file_size_in_byte":1131,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28996047472","text":"from dal import autocomplete\nfrom django import forms\nfrom services.forms import ServiceSelect2GenericForeignKeyModelField\nfrom services.models import CompanyServiceAggregation, CompanyServiceCRM\n\nfrom .models import (Company, CompanyAccounting, CompanySubscription,\n CompanyWorkPlace)\n\n\nclass CompanyForm(forms.ModelForm):\n class Meta:\n model = Company\n fields = ('__all__')\n\n\nclass CompanyWorkPlaceInlineForm(forms.ModelForm):\n class Meta:\n model = CompanyWorkPlace\n fields = ('__all__')\n widgets = {\n 'countries': autocomplete.ModelSelect2Multiple(url='country-autocomplete'),\n 'regions': autocomplete.ModelSelect2Multiple(url='region-autocomplete', forward=[\"countries\"]),\n 'cities': autocomplete.ModelSelect2Multiple(url='city-autocomplete', forward=[\"regions\"]),\n 'streets': autocomplete.ModelSelect2Multiple(url='street-autocomplete', forward=[\"cities\"]),\n }\n\n\nclass CompanyAccountingInlineForm(forms.ModelForm):\n class Meta:\n model = CompanyAccounting\n fields = ('__all__')\n\n\nclass CompanySubscriptionInlineForm(autocomplete.FutureModelForm):\n service = ServiceSelect2GenericForeignKeyModelField()\n\n class Meta:\n model = CompanySubscription\n fields = ('__all__')\n\n widgets = {\n 'company': autocomplete.ModelSelect2(url='company-autocomplete'),\n }","repo_name":"VasiliiKletkin/CRM_Otkachaika","sub_path":"crm/companies/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1428,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27325646240","text":"import os\nimport sys\nimport ctypes as ct\nfrom ctypes import CDLL, POINTER\nfrom ctypes import c_size_t, c_int32\nimport json\nimport numpy as np\ncurdir = os.path.dirname(__file__)\n\ndef ensure_contiguous(array):\n return np.ascontiguousarray(array) if not array.flags['C_CONTIGUOUS'] else array\nif __name__ == '__main__':\n prefix_path = f\"{curdir}/runs\"\n num_estimators = 10\n children_left = []\n children_right = []\n threshold = []\n feature = []\n value = []\n for i in range(0, num_estimators):\n with open(f'{prefix_path}/childrenLeft{i}', 'r') as f:\n children_left.append(json.load(f))\n with open(f'{prefix_path}/childrenRight{i}', 'r') as f:\n children_right.append(json.load(f))\n with open(f'{prefix_path}/threshold{i}', 'r') as f:\n threshold.append(json.load(f))\n with open(f'{prefix_path}/feature{i}', 'r') as f:\n feature.append(json.load(f))\n with open(f'{prefix_path}/value{i}', 'r') as f:\n value.append(json.load(f))\n max_children_left_size = max([len(children_left[i]) for i in range(num_estimators)])\n max_children_right_size = max([len(children_right[i]) for i in range(num_estimators)])\n max_feature_size = max([len(feature[i]) for i in range(num_estimators)])\n max_value_size = max([len(value[i]) for i in range(num_estimators)])\n max_threshold_size = max([len(threshold[i]) for i in range(num_estimators)])\n for i in range(num_estimators):\n if len(value[i]) < max_value_size:\n tmp = max_value_size - len(value[i])\n value[i] += [0] * tmp\n if len(children_left[i]) < max_children_left_size:\n tmp = max_children_left_size - len(children_left[i])\n children_left[i] += [-1] * tmp\n if len(children_right[i]) < max_children_right_size:\n tmp = max_children_right_size - len(children_right[i])\n children_right[i] += [-1] * tmp\n if len(threshold[i]) < max_threshold_size:\n tmp = max_threshold_size - len(threshold[i])\n threshold[i] += [-2] * tmp\n if len(feature[i]) < max_feature_size:\n tmp = max_feature_size - len(feature[i])\n feature[i] += [-2] * tmp\n\n children_right = np.array(children_right).ravel()\n children_left = np.array(children_left).ravel()\n threshold = np.array(threshold).ravel()\n feature = np.array(feature).ravel()\n value = np.array(value).ravel()\n\n # value = np.array(value)\n # print(value)\n lib = CDLL(f\"{curdir}/rf_filter.so\")\n\n rf_filter = lib.rf_filter\n seconds = 110\n\n # ND_POINTER_1 = np.ctypeslib.ndpointer(dtype=children_left.dtype, ndim=1, flags=\"C_CONTIGUOUS\")\n children_left_pointer = children_left.ctypes.data_as(POINTER(ct.c_longlong))\n children_right_pointer = children_right.ctypes.data_as(POINTER(ct.c_longlong))\n threshold_pointer = threshold.ctypes.data_as(POINTER(ct.c_longlong))\n feature_pointer = feature.ctypes.data_as(POINTER(ct.c_longlong))\n value_pointer = value.ctypes.data_as(POINTER(ct.c_longlong))\n\n c_uint_p = ct.POINTER(ct.c_uint)\n ret = ensure_contiguous(np.zeros(seconds, dtype=np.uintc))\n _ret = ret.ctypes.data_as(c_uint_p)\n\n rf_filter.argstypes = [c_int32, POINTER(ct.c_longlong), POINTER(ct.c_longlong), POINTER(ct.c_longlong), POINTER(ct.c_longlong), POINTER(ct.c_longlong), c_size_t]\n # rf_filter.argstypes = [ND_POINTER_1, c_size_t]\n rf_filter.restype = None\n\n # print(rf_filter(children_left))\n rf_filter(seconds, children_left_pointer, children_right_pointer, value_pointer, feature_pointer, threshold_pointer, max_children_right_size, _ret)\n # print(rf_filter(children_left, children_left.size))\n\n resdir = sys.argv[2]\n filename = f\"{resdir}/rxpps.log\"\n with open (filename, \"w\") as f:\n for d in np.ctypeslib.as_array(ret, seconds):\n f.write(f\"{d}\\n\")\n","repo_name":"oakeshott/mlids","sub_path":"prog-size/rf-filter/rf_filter_rawsocket_us.py","file_name":"rf_filter_rawsocket_us.py","file_ext":"py","file_size_in_byte":3959,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"2324089441","text":"n = int(input())\nx = 0\nheap = []\nfor _ in range(n):\n inp = input()\n try:\n tmp = int(inp)\n except ValueError:\n tmp, x = map(int, inp.split())\n \n if tmp: #extract\n print(heap[0])\n heap[0] = heap[-1]\n pos = 0\n while 2*pos+2 < len(heap):\n max_kid = 2*pos+1\n if heap[max_kid] < heap[max_kid+1]:\n max_kid += 1\n if heap[pos] < heap[max_kid]:\n heap[pos], heap[max_kid] = heap[max_kid], heap[pos]\n pos = max_kid\n else:\n break\n heap.pop()\n print()\n\n else: #insert\n heap.append(x)\n pos = len(heap)-1\n while(pos > 0)and(heap[pos]>heap[(pos-1)//2]):\n heap[pos], heap[(pos-1)//2] = heap[(pos-1)//2], heap[pos]\n pos = (pos-1)//2\n","repo_name":"diana-bsv/Algorithms-python","sub_path":"19-heap/project19.py","file_name":"project19.py","file_ext":"py","file_size_in_byte":711,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8639733475","text":"#!/usr/bin/env python\n# _*_ coding: utf-8 _*_\n\"\"\"\n# @Time : 3/2/22 6:41 PM\n# @Author : Archer\n# @File : transaction_args.py\n# @desc :\n\"\"\"\nfrom tools.pointsMul import scalar_mult_g\n\n\ndef invoke_create_args(to, t1, t2, domain_name):\n args = {'func_choice': 0,\n 'to': to,\n 't1': t1,\n 't2': t2,\n 'domain_name': domain_name}\n return args\n\n\ndef invoke_commit_args(to, bidding_price):\n args = {'func_choice': 1,\n 'to': to,\n 'bidding_price': bidding_price}\n return args\n\n\ndef invoke_reveal_args(to):\n args = {'func_choice': 2,\n 'to': to}\n return args\n\n\ndef invoke_finalize_args(to):\n args = {'func_choice': 3,\n 'to': to}\n return args\n\n\ndef invoke_update_args(user, to, domain_name, ip):\n signature = user.sign(domain_name)\n args = {'func_choice': 4,\n 'pk': user.pk,\n 'to': to,\n 'domain_name': domain_name,\n 'ip': ip,\n 'signature': signature\n }\n return args\n\n\ndef invoke_transfer_args(user, to, domain_name, pk2, value, end_time):\n enc_fund = scalar_mult_g(value)\n args = {'func_choice': 5,\n 'to': to,\n 'pk1': user.pk,\n 'domain_name': domain_name,\n 'signature': user.sign(domain_name),\n 'pk2': pk2,\n 'funds': enc_fund,\n 'end_time': end_time\n }\n return args\n\n\ndef invoke_receiver_args(to, user, domain_name, value):\n enc_fund = scalar_mult_g(value)\n\n args = {'func_choice': 6,\n 'to': to,\n 'domain_name': domain_name,\n 'pk2': user.pk,\n 'funds': enc_fund,\n 'value': value,\n }\n return args\n\n\ndef invoke_renewal_args(to, pk1, user, domain_name, value):\n enc_fund = scalar_mult_g(value)\n args = {'func_choice': 7,\n 'to': to,\n 'pk1': pk1,\n 'pk2': user.pk,\n 'domain_name': domain_name,\n 'signature': user.sign(domain_name),\n 'funds': enc_fund,\n 'value': value,\n }\n return args\n\n\nif __name__ == '__main__':\n print('hello world')\n","repo_name":"EmmaLu-ux/paperDemo","sub_path":"consortium_3_7_th/transaction_file/transaction_args.py","file_name":"transaction_args.py","file_ext":"py","file_size_in_byte":2170,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"13340537894","text":"import pandas as pd\nprediction_data = pd.read_csv(\"predictions_adversarial_vgg_best.csv\")\nwhite = prediction_data[prediction_data.race == 0]\nnonwhite = prediction_data[prediction_data.race != 0]\n\nfemale = prediction_data[prediction_data.label == 0]\nmale = prediction_data[prediction_data.label == 1]\n\nwhite_female = white[white.label == 0]\nwhite_male = white[white.label == 1]\nnonwhite_female = nonwhite[nonwhite.label == 0]\nnonwhite_male = nonwhite[nonwhite.label == 1]\n\nprint(\"Total Accuracy:\")\nprint(sum(prediction_data.label == prediction_data.pred)/len(prediction_data))\n\n#gender split\ndef confusion(male, female, label1, label2):\n #condition: gender: present (1 is male), absent is female (0)\n tp_men = sum(male.label == male.pred)\n fn_men = sum(male.label != male.pred)\n fp_men = sum(female.label != female.pred)\n tn_men = sum(female.label == female.pred)\n tpr_men = tp_men/(tp_men + fn_men) #sum(male.label == male.pred)/(sum(male.label == male.pred) + sum(female.label != female.pred))\n #false positive rate is 1 - specificity\n fpr_men = 1 - tn_men/(tn_men+fp_men)#sum(male.label != male.pred)/(sum(female.label == female.pred) + sum(male.label != male.pred))\n ppv_men = tp_men/(tp_men + fp_men)\n\n tp_women = sum(female.label == female.pred)\n fn_women = sum(female.label != female.pred)\n fp_women = sum(male.label != male.pred)\n tn_women = sum(male.label == male.pred)\n tpr_women = tp_women/(tp_women + fn_women)\n fpr_women = 1 - tn_women/(tn_women+fp_women)#sum(male.label != male.pred)/(sum(female.label == female.pred) + sum(male.label != male.pred))\n ppv_women = tp_women/(tp_women + fp_women)\n\n print(label1)\n print(\"PPV\", ppv_men)\n print(\"ERR\", 1 - ppv_men)\n print(\"TPR:\", tpr_men)\n print(\"FPR:\", fpr_men)\n\n print(label2)\n print(\"PPV\", ppv_women)\n print(\"ERR\", 1 - ppv_women)\n print(\"TPR:\", tpr_women)\n print(\"FPR:\", fpr_women)\n\n# def confusion_category(data):\n# #0 is female\n# #1 is male\n# correct = sum(data.label == data.pred)\n# acc = correct/len(data)\n#\n# # True Positive (TP): we predict a label of 1 (positive), and the true label is 1.\n# TP = np.sum(np.logical_and(pred_labels == 1, true_labels == 1))\n#\n# # False Positive (FP): we predict a label of 1 (positive), but the true label is 0.\n# FP = np.sum(np.logical_and(pred_labels == 1, true_labels == 0))\n#\n# # False Negative (FN): we predict a label of 0 (negative), but the true label is 1.\n# FN = np.sum(np.logical_and(pred_labels == 0, true_labels == 1))\n# print(\"ACCURACY:\", acc)\n\nconfusion(male, female, \"male\", \"female\")\n\nconfusion(nonwhite_male, nonwhite_female, \"nonwhite_male\", \"nonwhite_female\")\n\nconfusion(white_male, white_female, \"white_male\", \"white_female\")\n","repo_name":"sashahar/CS231N_race_and_facial_recognition","sub_path":"generate_confusion_matrix.py","file_name":"generate_confusion_matrix.py","file_ext":"py","file_size_in_byte":2761,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"42645470616","text":"from fastapi import FastAPI, Response, status, HTTPException, Depends, APIRouter\nfrom sqlalchemy.orm import Session\nfrom typing import List, Optional\nfrom sqlalchemy import func\n# from sqlalchemy.sql.functions import func\nfrom .. import models, schemas, oauth2\nfrom ..database import get_db\n\n\ndef get_invoices(db: Session):\n containers=db.query(models.Invoice).filter(models.Invoice.deleted!=True).all()\n return containers\n\n\nasync def create_invoice(post: schemas.InvoiceCreate,item:List[schemas.InvoiceItem], db: Session = Depends(get_db), current_user: int = Depends(oauth2.get_current_user)):\n \n new_invoice = models.Invoice(invoice_owner_id=current_user.id,payment_due=(post.value_net-post.actual_payment), **post.dict())\n db.add(new_invoice)\n db.commit()\n db.refresh(new_invoice)\n \n new_id=new_invoice.id\n for invoice_item in item:\n prod=invoice_item.product_name\n quant=invoice_item.quantity\n #verify if this product exist\n p= db.query(models.Product).filter(models.Product.product_name==prod).first()\n if not p:\n raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST ,detail=f\"{prod} is not a product\")\n p.quantity_left-=quant\n db.commit()\n new_invoice_item = models.InvoiceItem(invoice_id=new_id,**invoice_item.dict())\n db.add(new_invoice_item)\n db.commit()\n \n return new_invoice\n\n\n\ndef get_invoice(id: int, db: Session , current_user: int = Depends(oauth2.get_current_user)):\n \n\n invoice = db.query(models.Invoice).filter(models.Invoice.id == id,models.Invoice.deleted!=True).first()\n\n if not invoice:\n raise HTTPException(status_code=status.HTTP_404_NOT_FOUND,\n detail=f\"invoice with id: {id} was not found\")\n\n return invoice\n\n\ndef update_invoice(id: int, updated_post: schemas.InvoiceCreate, db: Session = Depends(get_db), current_user: int = Depends(oauth2.get_current_user)):\n\n invoice_query = db.query(models.Invoice).filter(models.Invoice.id == id,models.Invoice.deleted!=True)\n\n invoice = invoice_query.first()\n\n if invoice == None:\n raise HTTPException(status_code=status.HTTP_404_NOT_FOUND,\n detail=f\"invoice with id: {id} does not exist\")\n\n \n invoice_query.update(updated_post.dict(), synchronize_session=False)\n\n db.commit()\n\n return invoice_query.first()\n\ndef delete_invoice(id: int, db: Session = Depends(get_db), current_user: int = Depends(oauth2.get_current_user)):\n\n invoice_query = db.query(models.Invoice).filter(models.Invoice.id == id,models.Invoice.deleted!=True)\n\n invoice = invoice_query.first()\n\n if invoice == None:\n raise HTTPException(status_code=status.HTTP_404_NOT_FOUND,\n detail=f\"invoice with id: {id} does not exist\")\n invoice.deleted = True\n \n db.commit()\n return Response(status_code=status.HTTP_204_NO_CONTENT)\n\n\n\n\n#","repo_name":"thiere18/invoice-generator-api","sub_path":"app/repository/invoice.py","file_name":"invoice.py","file_ext":"py","file_size_in_byte":2934,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"79"} +{"seq_id":"23186979090","text":"\"\"\"\nThe below is class which implements `__next__`as if it was a function using the `yield` keyword:\n\"\"\"\n\n\nclass FirstHundredGenerator(object):\n def __init__(self):\n self.number = 0\n\n def __next__(self):\n if self.number < 100:\n current = self.number\n self.number += 1\n return current\n else:\n raise StopIteration()\n\n\ngen = FirstHundredGenerator()\nnext(gen) # 0\nnext(gen) # 1\n\n\"\"\"\nNotice how the object, with its property, remembers what the value of `self.number` is at all points in time.\n\nThis object is called in Python a generator because every time the next number is available not because it’s in a sequence, but because it is generated from its current state (in this case, by adding 1 to `self.number`).\n\nAll objects that have this `__next__` method are called iterators. All generators are iterators, but not the other way round.\n\nFor example, you could have an iterator on which you can call `next()`, but that doesn’t generate its values. Instead, it could take them from a list or from a database.\n\n*Important*: iterators are objects which have a `__next__` method.\n\nHere’s an example of an iterator which is not a generator:\n\"\"\"\n\n\nclass FirstFiveIterator:\n def __init__(self):\n self.numbers = [1, 2, 3, 4, 5]\n self.i = 0\n\n def __next__(self):\n if self.i < len(self.numbers):\n current = self.numbers[self.i]\n self.i += 1\n return current\n else:\n raise StopIteration()\n\n\n\"\"\"\nAs you can see it’s returning numbers that are not being generated; instead they’re being returned from a list.\n\nIf we run this code though, we will get an error:\n\n\"\"\"\n\nsum(FirstHundredGenerator()) # comment this line out to run the rest of the file.\n\n\"\"\"\nSimilarly if we run this code:\n\"\"\"\n\nfor i in FirstHundredGenerator():\n print(i)\n\n\"\"\"\nAnd that’s because in Python, an `iterator` and an `iterable` are different things. You can iterate over an `iterable`. The iterator is used to get the next value (either from a sequence or generated values).\n\n> You can iterate over iterables, not over iterators.\n\"\"\"\n","repo_name":"tecladocode/complete-python-course","sub_path":"course_contents/9_advanced_built_in_functions/lectures/02_generator_classes_and_iterators/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2160,"program_lang":"python","lang":"en","doc_type":"code","stars":294,"dataset":"github-code","pt":"79"} +{"seq_id":"71690382975","text":"def run():\n if len(inp.split('/')) == 2:\n m, f = inp.split('/')\n obj_module = __import__(m)\n getf(obj_module, f)\n elif len(inp.split('/')) == 3:\n p, m, f = inp.split('/')\n obj_module = __import__(p + '.' + m, fromlist=True)\n getf(obj_module, f)\n\n\ndef getf(m, f):\n if hasattr(m, f):\n func = getattr(m, f)\n func()\n else:\n print('404')\n\nif __name__ == \"__main__\":\n inp = input('url>')\n run()","repo_name":"beicuideyun/my_python_lib","sub_path":"pytestcloud/lib/reflection.py","file_name":"reflection.py","file_ext":"py","file_size_in_byte":469,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"14252899980","text":"from collections import deque\nimport sys\nsys.stdin = open('M.txt')\n\n\ndef dummy(r, c):\n d = ((0, 1), (1, 0), (0, -1), (-1, 0))\n count = idx = 0 # count는 이동 시간, idx는 방향\n snake = deque() # 왼쪽이 뱀의 머리, 오른쪽이 뱀의 꼬리\n snake.append((0, 0))\n start_r, start_c = r, c\n board[start_r][start_c] = -1 # 뱀의 위치 -1\n\n while 1:\n if moves:\n cur = moves[0]\n if int(cur[0]) == count:\n moves.popleft()\n if cur[1] == 'L':\n idx = (idx - 1) % 4\n continue\n elif cur[1] == 'D':\n idx = (idx + 1) % 4\n continue\n\n start_r += d[idx][0]\n start_c += d[idx][1]\n\n if 0 <= start_r < N and 0 <= start_c < N: # 보드 안에 위치해 있을 때, 즉 벽에 안 부딪힐 때\n if board[start_r][start_c] == -1: # 뱀과 부딪힐 때\n return count\n elif board[start_r][start_c] == 1: # 사과가 있을 떄\n board[start_r][start_c] = -1\n snake.appendleft((start_r, start_c))\n else:\n board[start_r][start_c] = -1\n snake.appendleft((start_r, start_c))\n last_r, last_c = snake.pop()\n board[last_r][last_c] = 0\n count += 1\n else:\n return count\n\n return count\n\n\n# 보드의 크기 N\nN = int(sys.stdin.readline())\nboard = [[0] * N for _ in range(N)]\n\n# 사과의 개수 K\nK = int(sys.stdin.readline())\nfor _ in range(K):\n r, c = map(int, sys.stdin.readline().split())\n board[r - 1][c - 1] = 1 # 사과 위치 1\n\n# 뱀의 방향 변환 횟수 L\nL = int(sys.stdin.readline())\nmoves = deque(list(map(str, sys.stdin.readline().split())) for _ in range(L))\n\nprint(dummy(0, 0) + 1)\n","repo_name":"woohree/ALGO2ITHM_STUDY","sub_path":"baekjoon/05월/0530 줄세우기 특정한최단경로 뱀 정수삼각형/g5_3190_뱀/joorii.py","file_name":"joorii.py","file_ext":"py","file_size_in_byte":1907,"program_lang":"python","lang":"ko","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"69851750656","text":"import torch\nfrom tqdm import tqdm\nimport random\n\nclass Data:\n def __init__(self, filename, max_seq_length):\n self._max_seq_len = max_seq_length\n self._max_doc_len = 40\n self._all_context_ids = []\n self._all_docs_ids = []\n self._all_labels = []\n with open(filename, \"r\") as fr:\n lines = fr.readlines()\n for line in tqdm(lines, leave=False):\n line = line.strip().split(\"\\t\")\n context = line[0].split()\n docs = line[1].split(\"[=====]\")\n labels = line[2].split(\"[===]\")\n self._all_context_ids.append([int(x) for x in context])\n doc_ids = []\n for doc in docs:\n doc_ids.append([int(x) for x in doc.split()])\n self._all_docs_ids.append(doc_ids)\n self._all_labels.append([int(x) for x in labels])\n tqdm.write(\"Load data finished...\")\n self._sample_num = len(self._all_context_ids)\n self._all_train_ids = list(range(self._sample_num))\n \n def get_train_next_batch(self, batch_size, pacing_value):\n # train_ids = self._all_train_ids[:pacing_num]\n pacing_num = int(self._sample_num * pacing_value)\n train_ids = self._all_train_ids[:pacing_num]\n batch_idx_list = random.sample(train_ids, batch_size)\n batch_input_ids = []\n batch_token_type_ids = []\n batch_attention_masks = []\n batch_labels = []\n for idx in batch_idx_list:\n context_ids = self._all_context_ids[idx]\n docs_ids = self._all_docs_ids[idx]\n labels = self._all_labels[idx]\n docs_sample_ids = []\n docs_token_type_ids = []\n docs_attention_masks = []\n for doc_ids in docs_ids:\n doc_ids = doc_ids[:self._max_doc_len]\n sample_ids = context_ids + [102] + doc_ids + [102]\n token_type_ids = [0] * (len(context_ids) + 1) + [1] * (len(doc_ids) + 1)\n sample_ids = sample_ids[-(self._max_seq_len - 1):]\n token_type_ids = token_type_ids[-(self._max_seq_len - 1):]\n sample_ids = [101] + sample_ids\n token_type_ids = [0] + token_type_ids\n attention_mask = [1] * len(sample_ids)\n assert len(sample_ids) <= self._max_seq_len\n while len(sample_ids) < self._max_seq_len:\n sample_ids.append(0)\n token_type_ids.append(0)\n attention_mask.append(0)\n assert len(sample_ids) == len(token_type_ids) == len(attention_mask) == self._max_seq_len\n docs_sample_ids.append(sample_ids)\n docs_token_type_ids.append(token_type_ids)\n docs_attention_masks.append(attention_mask)\n batch_input_ids.append(docs_sample_ids)\n batch_token_type_ids.append(docs_token_type_ids)\n batch_attention_masks.append(docs_attention_masks)\n batch_labels.append(labels)\n batch = {\n 'input_ids': torch.LongTensor(batch_input_ids), \n 'token_type_ids': torch.LongTensor(batch_token_type_ids), \n 'attention_mask': torch.LongTensor(batch_attention_masks), \n 'labels': torch.FloatTensor(batch_labels)\n }\n return batch","repo_name":"DaoD/Generation","sub_path":"PromptCAR/MyDataLoader.py","file_name":"MyDataLoader.py","file_ext":"py","file_size_in_byte":3376,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"4316015775","text":"#!/usr/bin/env python\n# coding: utf-8\n\nimport sys\n\n\ndef main():\n print(\"#!/bin/sh\")\n for line in sys.stdin:\n if line.startswith(\"/dev/input\"):\n dev, values = line.split(\":\")\n _, val1, val2, val3 = values.split(\" \")\n print(\n f\"sendevent {dev} {int(val1, 16)} {int(val2, 16)} {int(val3, 16)}\"\n )\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"Konubinix/Devel","sub_path":"bin/adb_getevent_to_sendevent.py","file_name":"adb_getevent_to_sendevent.py","file_ext":"py","file_size_in_byte":406,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"79"} +{"seq_id":"32032005754","text":"def choose_options():\n import random\n options = ('piedra', 'papel', 'tijera')\n user_option = input('Elija , piedra papel o tijera -> ').lower()\n if user_option not in options:\n print('Esa opción no es valida')\n return None, '0'\n #continue\n computer_option = random.choice(options)\n print('User_option --> ', user_option)\n print('Computer option --> ', computer_option)\n return user_option, computer_option\n\n\ndef check_rules(user_option, computer_option, users_wins, computer_wins):\n if user_option == computer_option:\n print('Empate')\n elif user_option == 'piedra':\n if computer_option == 'tijera':\n print('piedra gana a tijera')\n print('user ganó!')\n users_wins += 1\n else:\n print('papel gana a piedra')\n print('computer ganó!')\n computer_wins += 1\n elif user_option == 'papel':\n if computer_option == 'piedra':\n print('papel gana a piedra')\n print('user ganó!')\n users_wins += 1\n else:\n print('tijera gana a papel')\n print('computer ganó')\n computer_wins += 1\n elif user_option == 'tijera':\n if computer_option == 'papel':\n print('tijera gana a papel')\n print('user ganó')\n users_wins += 1\n else:\n print('piedra gana a tijera')\n print('computer ganó!')\n computer_wins += 1\n return users_wins, computer_wins\n\n\ndef check_winner(users_wins, computer_wins):\n if computer_wins == 2:\n print('El ganador es la computadora')\n return True\n if users_wins == 2:\n print('El ganador es el usuario')\n return True\n\n\ndef run_game():\n rounds = 1\n computer_wins = 0\n users_wins = 0\n while True:\n print('*' * 10)\n print('ROUND', rounds)\n print('*' * 10)\n print('computer_wins', computer_wins)\n print('users_wins', users_wins)\n rounds += 1\n user_option, computer_option = choose_options()\n users_wins, computer_wins = check_rules(user_option, computer_option,\n users_wins, computer_wins)\n if check_winner(users_wins, computer_wins):\n break\n\n\nrun_game()\n","repo_name":"AlejoGuerrero17/curso-python-pip","sub_path":"game/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2068,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"34357997643","text":"import uuid\n\nimport graphene\nimport pydantic\nfrom app.db import table\nfrom app.models import EventModel, create_event\nfrom employees.schema import Employee\nfrom graphene_pydantic import PydanticInputObjectType, PydanticObjectType\nfrom starlette.graphql import GraphQLApp\nfrom utils.time import utcnow\n\nfrom expenses.models import ExpenseModel, ResolutionModel\nfrom expenses.view_models import ExpenseViewModel, update_expense_view_model\n\n\nclass ResolutionFormModel(ResolutionModel):\n status: str\n\n\nclass ResolutionForm(PydanticInputObjectType):\n class Meta:\n model = ResolutionFormModel\n exclude_fields = (\"created_at\",)\n\n\nclass Resolution(PydanticObjectType):\n class Meta:\n model = ResolutionModel\n exclude_fields = (\"status\",)\n\n status = graphene.String()\n\n def resolve_status(self, info, **kwargs):\n return self.status\n\n\nclass Expense(PydanticObjectType):\n class Meta:\n model = ExpenseModel\n\n\nclass ExpenseView(PydanticObjectType):\n class Meta:\n model = ExpenseViewModel\n\n\nclass ExpensesQuery(graphene.ObjectType):\n expenses = graphene.List(ExpenseView)\n\n @staticmethod\n def resolve_expenses(info, name):\n expense_view_model_dicts = (\n table(ExpenseViewModel).find({}).sort(\"expense.created_at\", -1)\n )\n expense_view_models = [ExpenseViewModel(**x) for x in expense_view_model_dicts]\n return expense_view_models\n\n\nclass CreateResolutions(graphene.Mutation):\n class Arguments:\n resolution_forms = graphene.List(ResolutionForm)\n\n Output = graphene.List(ExpenseView)\n\n @staticmethod\n def mutate(parent, info, resolution_forms):\n resolutions = []\n for resolution_form in resolution_forms:\n resolution = ResolutionModel(created_at=utcnow(), **resolution_form)\n expense_event = create_event(\"resolution\", resolution)\n table(EventModel).insert_one(expense_event.dict())\n update_expense_view_model(expense_event)\n resolutions.append(resolution)\n\n expense_view_model_dicts = table(ExpenseViewModel).find(\n {\"expense.uuid\": {\"$in\": [x.expense_uuid for x in resolution_forms]}}\n )\n expense_view_models = [ExpenseViewModel(**x) for x in expense_view_model_dicts]\n return expense_view_models\n\n\nclass ExpensesMutation(graphene.ObjectType):\n createResolutions = CreateResolutions.Field()\n","repo_name":"mnieber/code-challenges","sub_path":"cashcog/src/backend/expenses/schema.py","file_name":"schema.py","file_ext":"py","file_size_in_byte":2424,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38612878076","text":"# simple BFS from the entrance\n# until we find ourselves on the edge with \".\" that is not the entrance\nclass Solution:\n def nearestExit(self, maze: List[List[str]], entrance: List[int]) -> int:\n \n m, n = len(maze), len(maze[0])\n q = deque()\n q.append(entrance)\n visited = set()\n steps = 0\n\n while q:\n for _ in range(len(q)):\n x, y = q.popleft()\n\n # if we've already visited this spot\n if (x, y) in visited:\n continue\n visited.add((x, y))\n\n # we're out of bounds or stuck in the wall\n if x < 0 or x >= m or y < 0 or y >= n or maze[x][y] == \"+\":\n continue\n\n # we reach the exit which is != entrance\n if steps != 0 and (x == 0 or x == m - 1 or y == 0 or y == n - 1):\n return steps\n\n # add all neighboring cells\n # (we could also check if they've been visited before)\n for xi, yi in ([-1, 0], [1, 0], [0, -1], [0, 1]):\n q.append([x + xi, y + yi])\n\n steps += 1\n\n return -1\n","repo_name":"yanvoi/leetcode_solutions","sub_path":"Medium/1926NearestExitFromEntranceInMaze.py","file_name":"1926NearestExitFromEntranceInMaze.py","file_ext":"py","file_size_in_byte":1190,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"4815688581","text":"# DO NOT modify this file by hand, changes will be overwritten\nfrom dataclasses import dataclass\n\nfrom cloudformation_cli_python_lib.interface import BaseModel\nfrom cloudformation_cli_python_lib.recast import recast_object\nfrom cloudformation_cli_python_lib.utils import deserialize_list\n\nimport sys\nfrom inspect import getmembers, isclass\nfrom typing import (\n AbstractSet,\n Any,\n Generic,\n Mapping,\n MutableMapping,\n Optional,\n Sequence,\n Type,\n TypeVar,\n)\n\nT = TypeVar(\"T\")\n\n\ndef set_or_none(value: Optional[Sequence[T]]) -> Optional[AbstractSet[T]]:\n if value:\n return set(value)\n return None\n\n\n@dataclass\nclass AwsAppstreamFleet(BaseModel):\n Description: Optional[str]\n ComputeCapacity: Optional[\"_ComputeCapacity\"]\n Platform: Optional[str]\n VpcConfig: Optional[\"_VpcConfig\"]\n FleetType: Optional[str]\n EnableDefaultInternetAccess: Optional[bool]\n DomainJoinInfo: Optional[\"_DomainJoinInfo\"]\n SessionScriptS3Location: Optional[\"_S3Location\"]\n Name: Optional[str]\n ImageName: Optional[str]\n MaxUserDurationInSeconds: Optional[int]\n IdleDisconnectTimeoutInSeconds: Optional[int]\n UsbDeviceFilterStrings: Optional[Sequence[str]]\n DisconnectTimeoutInSeconds: Optional[int]\n DisplayName: Optional[str]\n StreamView: Optional[str]\n IamRoleArn: Optional[str]\n MaxSessionsPerInstance: Optional[int]\n Id: Optional[str]\n InstanceType: Optional[str]\n MaxConcurrentSessions: Optional[int]\n Tags: Optional[Any]\n ImageArn: Optional[str]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_AwsAppstreamFleet\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_AwsAppstreamFleet\"]:\n if not json_data:\n return None\n dataclasses = {n: o for n, o in getmembers(sys.modules[__name__]) if isclass(o)}\n recast_object(cls, json_data, dataclasses)\n return cls(\n Description=json_data.get(\"Description\"),\n ComputeCapacity=ComputeCapacity._deserialize(json_data.get(\"ComputeCapacity\")),\n Platform=json_data.get(\"Platform\"),\n VpcConfig=VpcConfig._deserialize(json_data.get(\"VpcConfig\")),\n FleetType=json_data.get(\"FleetType\"),\n EnableDefaultInternetAccess=json_data.get(\"EnableDefaultInternetAccess\"),\n DomainJoinInfo=DomainJoinInfo._deserialize(json_data.get(\"DomainJoinInfo\")),\n SessionScriptS3Location=S3Location._deserialize(json_data.get(\"SessionScriptS3Location\")),\n Name=json_data.get(\"Name\"),\n ImageName=json_data.get(\"ImageName\"),\n MaxUserDurationInSeconds=json_data.get(\"MaxUserDurationInSeconds\"),\n IdleDisconnectTimeoutInSeconds=json_data.get(\"IdleDisconnectTimeoutInSeconds\"),\n UsbDeviceFilterStrings=json_data.get(\"UsbDeviceFilterStrings\"),\n DisconnectTimeoutInSeconds=json_data.get(\"DisconnectTimeoutInSeconds\"),\n DisplayName=json_data.get(\"DisplayName\"),\n StreamView=json_data.get(\"StreamView\"),\n IamRoleArn=json_data.get(\"IamRoleArn\"),\n MaxSessionsPerInstance=json_data.get(\"MaxSessionsPerInstance\"),\n Id=json_data.get(\"Id\"),\n InstanceType=json_data.get(\"InstanceType\"),\n MaxConcurrentSessions=json_data.get(\"MaxConcurrentSessions\"),\n Tags=json_data.get(\"Tags\"),\n ImageArn=json_data.get(\"ImageArn\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_AwsAppstreamFleet = AwsAppstreamFleet\n\n\n@dataclass\nclass ComputeCapacity(BaseModel):\n DesiredInstances: Optional[int]\n DesiredSessions: Optional[int]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_ComputeCapacity\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_ComputeCapacity\"]:\n if not json_data:\n return None\n return cls(\n DesiredInstances=json_data.get(\"DesiredInstances\"),\n DesiredSessions=json_data.get(\"DesiredSessions\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_ComputeCapacity = ComputeCapacity\n\n\n@dataclass\nclass VpcConfig(BaseModel):\n SubnetIds: Optional[Sequence[str]]\n SecurityGroupIds: Optional[Sequence[str]]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_VpcConfig\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_VpcConfig\"]:\n if not json_data:\n return None\n return cls(\n SubnetIds=json_data.get(\"SubnetIds\"),\n SecurityGroupIds=json_data.get(\"SecurityGroupIds\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_VpcConfig = VpcConfig\n\n\n@dataclass\nclass DomainJoinInfo(BaseModel):\n OrganizationalUnitDistinguishedName: Optional[str]\n DirectoryName: Optional[str]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_DomainJoinInfo\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_DomainJoinInfo\"]:\n if not json_data:\n return None\n return cls(\n OrganizationalUnitDistinguishedName=json_data.get(\"OrganizationalUnitDistinguishedName\"),\n DirectoryName=json_data.get(\"DirectoryName\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_DomainJoinInfo = DomainJoinInfo\n\n\n@dataclass\nclass S3Location(BaseModel):\n S3Bucket: Optional[str]\n S3Key: Optional[str]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_S3Location\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_S3Location\"]:\n if not json_data:\n return None\n return cls(\n S3Bucket=json_data.get(\"S3Bucket\"),\n S3Key=json_data.get(\"S3Key\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_S3Location = S3Location\n\n\n@dataclass\nclass Tag(BaseModel):\n Value: Optional[str]\n Key: Optional[str]\n\n @classmethod\n def _deserialize(\n cls: Type[\"_Tag\"],\n json_data: Optional[Mapping[str, Any]],\n ) -> Optional[\"_Tag\"]:\n if not json_data:\n return None\n return cls(\n Value=json_data.get(\"Value\"),\n Key=json_data.get(\"Key\"),\n )\n\n\n# work around possible type aliasing issues when variable has same name as a model\n_Tag = Tag\n\n\n","repo_name":"aws-cloudformation/aws-cloudformation-samples","sub_path":"hooks/python-hooks/resource-tags/src/awssamples_resourcetags_hook/target_models/aws_appstream_fleet.py","file_name":"aws_appstream_fleet.py","file_ext":"py","file_size_in_byte":6418,"program_lang":"python","lang":"en","doc_type":"code","stars":121,"dataset":"github-code","pt":"79"} +{"seq_id":"38520759115","text":"# looks up an object in a given module and replace it with a Mock\n\n\nimport requests\nfrom datetime import datetime\n\n\ndef is_weekday():\n today = datetime.today()\n # Python's datetime library treats Monday as 0 and Sunday as 6\n return 0 <= today.weekday() < 5\n\n\ndef get_holidays():\n r = requests.get('http://localhost/api/holidays')\n if r.status_code == 200:\n return r.json()\n return None\n\n# as a Decorator :\n# mock for the entire duration of the entire test function\n","repo_name":"FelitaD/Learning-in-Public","sub_path":"Programming Languages/Python/mock/my_calendar.py","file_name":"my_calendar.py","file_ext":"py","file_size_in_byte":490,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"21507383367","text":"import network\nimport random\nimport anet\nimport pickle\n\n\nclass Versus:\n\n def __init__(self, game_kwargs, game, num_matches, player_start, player1=None, player2=None):\n self.game_kwargs = game_kwargs\n self.game_kwargs['player_start'] = player_start\n\n self.game_manager = game\n self.game = game(**game_kwargs)\n\n self.num_matches = num_matches\n\n self.player_start = player_start\n\n self.players = {1: player1, 0: player2}\n\n def match(self, verbose=True):\n if verbose:\n self.game.print_header()\n print('They will play', self.num_matches, 'versus games.')\n if self.player_start == -1:\n print('Starting player is random.')\n else:\n print('Player', self.game.player_to_string(self.player_start), 'makes the first move.')\n\n P1_wins = 0\n\n for i in range(self.num_matches):\n self.game_kwargs['player_start'] = self.choose_starting_player()\n self.game = self.game_manager(**self.game_kwargs)\n\n # Play game\n while not self.game.actual_game_over():\n player = self.players[self.game.get_player(self.game.state)]\n if isinstance(player, anet.Anet):\n move_state = self.game.anet_choose_child(self.game.state, player)\n elif player == 'human':\n move_state = self.game.request_human_move(self.game.state)\n # If no player given, make random moves\n else:\n move_state = self.game.request_random_move(self.game.state)\n # Make the move chosen by player\n self.game.make_actual_move(move_state)\n\n P1_wins += self.game.winner(self.game.state)\n\n if verbose:\n print('P1 wins', P1_wins, 'out of', self.num_matches, 'games (' + str(100*P1_wins/self.num_matches) + ')%')\n\n result = (P1_wins, self.num_matches-P1_wins)\n return result\n\n def choose_starting_player(self):\n if self.player_start == -1:\n n = random.random()\n return 0 if n < 0.5 else 1\n return self.player_start\n","repo_name":"VetleB/AIProg","sub_path":"Project3/versus.py","file_name":"versus.py","file_ext":"py","file_size_in_byte":2176,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"73589855936","text":"#!/usr/bin/python\n# -*- coding: iso-8859-1 -*-\n\nimport os\nimport scipy\nfrom scipy import stats\nimport math\n\nparkinson_class = [\"updrs\"]\nparkinson_experiments = [\"1\", \"2\", \"3\", \"4\", \"5\"]\nn_locutores = 25\nn_grupos = [5, 5, 5, 5, 5]\nversion = \"homoupdrs\"\nscript_bash = [\"baseline_linearregression_arff.sh\", \"baseline_simplelinearregression_arff.sh\", \"baseline_zeror_arff.sh\"]\nalgorithms = [\"LinearRegression\", \"SimpleLinearRegression\", \"ZeroR\"]\nos.system(\"rm print_new_data_algorithms_\"+version+\".dep\")\n\nfor n in range(len(script_bash)):\n\n if os.path.exists(script_bash[n]):\n\n for park_class in range(len(parkinson_class)):\n\n for group in range(len(parkinson_experiments)):\n\n FEATURE = \"Experiment_New_Data_\"+str(group+1)+\"_\"+parkinson_class[park_class]+\"_\"+version\n bash = \"./\"+script_bash[n]+ \" \" + FEATURE\n f = open(\"print_new_data_algorithms_\"+version+\".dep\", \"a\")\n f.write(\"Training \" + bash + \"\\n\")\n f.close()\n os.system(bash)\n\n if os.path.exists('eval/train_devel/'+FEATURE+'.'+algorithms[n]+'.pred'):\n file = open('eval/train_devel/'+FEATURE+'.'+algorithms[n]+'.pred', 'r')\n data = file.readlines();\n valor = []\n pred = []\n err = []\n for x in data:\n line = x.split(\" \");\n line_rel = []\n for l in range(len(line)):\n if line[l] != '':\n line_rel.append(line[l])\n if len(line_rel) == 5:\n valor.append(float(line_rel[1]))\n pred.append(float(line_rel[2]))\n err.append(float(line_rel[3]))\n pred_locutor = []\n valor_locutor = []\n err_locutor_abs = []\n err_locutor_rel = []\n rmse_locutor = []\n for ini in range(n_locutores/n_grupos[n]):\n pred_locutor.append(0)\n valor_locutor.append(0)\n rmse_locutor.append(0)\n for locutor in range((n_locutores/n_grupos[n])):\n valor_locutor[locutor] = valor[(len(pred)/(n_locutores/n_grupos[n]))*locutor]\n for p in range(len(pred)/(n_locutores/n_grupos[n])):\n pred_locutor[locutor] += pred[p+(len(pred)/(n_locutores/n_grupos[n]))*locutor]/(len(pred)/(n_locutores/n_grupos[n]))\n rmse_locutor[locutor] += err[p+(len(pred)/(n_locutores/n_grupos[n]))*locutor]*err[p+(len(pred)/(n_locutores/n_grupos[n]))*locutor]/(len(pred)/(n_locutores/n_grupos[n]))\n err_locutor_abs.append(abs(valor_locutor[locutor]-pred_locutor[locutor]))\n err_locutor_rel.append(abs(valor_locutor[locutor]-pred_locutor[locutor])/valor_locutor[locutor])\n rmse_locutor[locutor] = math.sqrt(rmse_locutor[locutor])\n err_abs = 0\n err_rel = 0\n rmse = 0\n for locutor in range(n_locutores/n_grupos[n]):\n err_abs += err_locutor_abs[locutor]/(n_locutores/n_grupos[n])\n err_rel += err_locutor_rel[locutor]/(n_locutores/n_grupos[n])\n rmse += rmse_locutor[locutor]/(n_locutores/n_grupos[n])\n s = scipy.stats.spearmanr(valor_locutor, pred_locutor)\n if math.isnan(float(s[0])):\n s = [0]\n file.close()\n f = open(\"print_new_data_algorithms_\"+version+\".dep\", \"a\")\n f.write(\"Spearman correlation coefficient: \" + str(s[0]) + \"\\n\")\n f.write(\"REAL: \" + str(valor_locutor) + \"\\n\")\n f.write(\"PREDICTION: \" + str(pred_locutor) + \"\\n\")\n f.write(\"RMSE: \" + str(rmse) + \"\\n\")\n f.write(\"ABSOLUTE ERROR: \" + str(err_abs) + \"\\n\")\n f.write(\"RELATIVE ERROR: \" + str(err_rel) + \"\\n\")\n f.close()\n\n f = open(\"print_new_data_algorithms_\"+version+\".dep\", \"a\")\n f.write(\"-------------------------------------------------------------------\" + \"\\n\" + \"\\n\")\n f.close()\n\n else:\n f = open(\"print_new_data_algorithms_\"+version+\".dep\", \"a\")\n f.write(\"The bash script \"+script_bash[n]+\" has not been created\")\n f.close()\n","repo_name":"clara-jr/Parkinson-Machine-Learning","sub_path":"baseline2/optimize_new_data_updrs_algorithms.py","file_name":"optimize_new_data_updrs_algorithms.py","file_ext":"py","file_size_in_byte":4636,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38128306861","text":"import copy\nimport inspect\nimport os\nimport time\nimport dill\nimport yaml\n\nfrom tqdm import tqdm\nimport torch\n\nfrom cloudtik.runtime.ai.modeling.transfer_learning.common.utils import \\\n verify_directory, validate_model_name\nfrom cloudtik.runtime.ai.modeling.transfer_learning.common.pytorch.model import PyTorchModel\nfrom cloudtik.runtime.ai.modeling.transfer_learning.image_classification.image_classification_dataset import \\\n ImageClassificationDataset\nfrom cloudtik.runtime.ai.modeling.transfer_learning.image_classification.image_classification_model import \\\n ImageClassificationModel\nfrom cloudtik.runtime.ai.modeling.transfer_learning.image_classification.pytorch.image_classification_dataset import \\\n PyTorchImageClassificationDataset\n\n\nclass PyTorchImageClassificationModel(ImageClassificationModel, PyTorchModel):\n \"\"\"\n Class to represent a PyTorch model for image classification\n \"\"\"\n\n def __init__(self, model_name: str, model=None, optimizer=None, loss=None, **kwargs):\n \"\"\"\n Class constructor\n \"\"\"\n # PyTorch models generally do not enforce a fixed input shape\n self._image_size = 'variable'\n\n # extra properties that will become configurable in the future\n self._do_fine_tuning = False\n self._dropout_layer_rate = None\n self._device = 'cpu'\n self._lr_scheduler = None\n self._generate_checkpoints = True\n self._num_classes = None\n\n PyTorchModel.__init__(self, model_name)\n ImageClassificationModel.__init__(self, self._image_size, self._do_fine_tuning, self._dropout_layer_rate,\n self._model_name)\n\n # set up the configurable optimizer and loss functions\n self._check_optimizer_loss(optimizer, loss)\n self._optimizer_class = optimizer if optimizer else torch.optim.Adam\n self._opt_args = {k: v for k, v in kwargs.items() if k in inspect.getfullargspec(self._optimizer_class).args}\n self._loss_class = loss if loss else torch.nn.CrossEntropyLoss\n self._loss_args = {k: v for k, v in kwargs.items() if k in inspect.getfullargspec(self._loss_class).args}\n self._loss = self._loss_class(**self._loss_args)\n\n if model is None:\n self._model = None\n elif isinstance(model, str):\n self.load_from_directory(model)\n layers = list(self._model.children())\n if isinstance(layers[-1], torch.nn.Sequential):\n self._num_classes = layers[-1][-1].out_features\n else:\n self._num_classes = layers[-1].out_features\n elif isinstance(model, torch.nn.Module):\n self._model = model\n layers = list(self._model.children())\n if isinstance(layers[-1], torch.nn.Sequential):\n self._num_classes = layers[-1][-1].out_features\n else:\n self._num_classes = layers[-1].out_features\n else:\n raise TypeError(\"The model input must be a torch.nn.Module, string or\",\n \"None but found a {}\". format(type(model)))\n\n @property\n def num_classes(self):\n \"\"\"\n The number of output neurons in the model; equal to the number of classes in the dataset\n \"\"\"\n return self._num_classes\n\n def _fit(self, output_dir, dataset, epochs, do_eval, early_stopping, lr_decay):\n \"\"\"Main PyTorch training loop\"\"\"\n since = time.time()\n\n device = torch.device(self._device)\n self._model = self._model.to(device)\n\n if dataset.train_subset:\n train_data_loader = dataset.train_loader\n data_length = len(dataset.train_subset)\n else:\n train_data_loader = dataset.data_loader\n data_length = len(dataset.dataset)\n\n if do_eval and dataset.validation_subset:\n validation_data_loader = dataset.validation_loader\n validation_data_length = len(dataset.validation_subset)\n else:\n validation_data_loader = None\n validation_data_length = 0\n\n # For early stopping, if enabled\n patience = 10\n trigger_time = 0\n last_loss = 1.0\n\n if lr_decay:\n self._lr_scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(self._optimizer, factor=0.2, patience=5,\n cooldown=1, min_lr=0.0000000001)\n\n self._history = {}\n self._model.train()\n for epoch in range(epochs):\n print(f'Epoch {epoch + 1}/{epochs}')\n print('-' * 10)\n\n # Training phase\n running_loss = 0.0\n running_corrects = 0\n\n # Iterate over data.\n for inputs, labels in tqdm(train_data_loader, bar_format='{l_bar}{bar:50}{r_bar}{bar:-50b}'):\n inputs = inputs.to(device)\n labels = labels.to(device)\n\n # Zero the parameter gradients\n self._optimizer.zero_grad()\n\n # Forward and backward pass\n with torch.set_grad_enabled(True):\n outputs = self._model(inputs)\n _, preds = torch.max(outputs, 1)\n loss = self._loss(outputs, labels)\n loss.backward()\n self._optimizer.step()\n\n # Statistics\n running_loss += loss.item() * inputs.size(0)\n running_corrects += torch.sum(preds == labels.data)\n\n train_epoch_loss = running_loss / data_length\n train_epoch_acc = float(running_corrects) / data_length\n self._update_history('Loss', train_epoch_loss)\n self._update_history('Acc', train_epoch_acc)\n\n loss_acc_output = f'Loss: {train_epoch_loss:.4f} - Acc: {train_epoch_acc:.4f}'\n\n if do_eval and validation_data_loader is not None:\n self._model.eval()\n running_loss = 0.0\n running_corrects = 0\n\n with torch.no_grad():\n print(\"Performing Evaluation\")\n for inputs, labels in tqdm(validation_data_loader, bar_format='{l_bar}{bar:50}{r_bar}{bar:-50b}'):\n outputs = self._model(inputs)\n _, preds = torch.max(outputs, 1)\n loss = self._loss(outputs, labels)\n\n running_loss += loss.item() * inputs.size(0)\n running_corrects += torch.sum(preds == labels.data)\n\n eval_epoch_loss = running_loss / validation_data_length\n eval_epoch_acc = float(running_corrects) / validation_data_length\n self._update_history('Val Loss', eval_epoch_loss)\n self._update_history('Val Acc', eval_epoch_acc)\n\n loss_acc_output += f' - Val Loss: {eval_epoch_loss:.4f} - Val Acc: {eval_epoch_acc:.4f}'\n\n if lr_decay:\n lr = self._lr_scheduler.optimizer.param_groups[0]['lr']\n self._update_history('LR', lr)\n loss_acc_output += f' - LR: {lr:.4f}'\n self._lr_scheduler.step(eval_epoch_loss)\n\n if early_stopping:\n if eval_epoch_loss >= last_loss:\n trigger_time += 1\n\n if trigger_time >= patience:\n # Stop Early\n print(\"Early stopping has been triggered after \" + str(epoch) + \" epochs.\")\n break\n else:\n trigger_time = 0\n\n last_loss = eval_epoch_loss\n\n print(loss_acc_output)\n\n time_elapsed = time.time() - since\n print(f'Training complete in {time_elapsed // 60:.0f}m {time_elapsed % 60:.0f}s')\n\n if self._generate_checkpoints:\n valid_model_name = validate_model_name(self.model_name)\n checkpoint_dir = os.path.join(output_dir, \"{}_checkpoints\".format(valid_model_name))\n verify_directory(checkpoint_dir)\n try:\n torch.save({\n 'epoch': epochs,\n 'model_state_dict': self._model.state_dict(),\n 'optimizer_state_dict': self._optimizer.state_dict(),\n 'loss': train_epoch_loss,\n }, os.path.join(checkpoint_dir, 'checkpoint.pt'))\n except KeyError:\n # Calling state_dict() on an IPEX optimizer calls into the torch optimizer's __setstate__ method\n # which in PyTorch 1.12 assumes that the first state value will always have a 'step' key\n state_values = list(self._optimizer.state.values())\n if 'step' not in state_values[0].keys():\n state_values[0]['step'] = torch.tensor([])\n torch.save({\n 'epoch': epochs,\n 'model_state_dict': self._model.state_dict(),\n 'optimizer_state_dict': self._optimizer.state_dict(),\n 'loss': train_epoch_loss,\n }, os.path.join(checkpoint_dir, 'checkpoint.pt'))\n\n def _fit_distributed(\n self, nnodes, nproc_per_node, hosts, hostfile,\n epochs, batch_size, ipex_optimize, objects_path):\n self.fit_distributed(\n nnodes, nproc_per_node, hosts, hostfile,\n epochs, batch_size, ipex_optimize,\n objects_path, category=\"image_classification\"\n )\n\n def train(self, dataset: ImageClassificationDataset, output_dir, *,\n epochs=1, initial_checkpoints=None, do_eval=True,\n early_stopping=False, lr_decay=True, seed=None, ipex_optimize=False,\n distributed=False, nnodes=1, nproc_per_node=1, hosts=None, hostfile=None,\n shared_dir=None, temp_dir=None):\n \"\"\"\n Trains the model using the specified image classification dataset. The first time training is called, it\n will get the model from torchvision and add on a fully-connected dense layer with linear activation\n based on the number of classes in the specified dataset. The model and optimizer are defined and trained\n for the specified number of epochs.\n\n Args:\n dataset (ImageClassificationDataset): Dataset to use when training the model\n output_dir (str): Path to a writeable directory for output files\n epochs (int): Number of epochs to train the model (default: 1)\n initial_checkpoints (str): Path to checkpoint weights to load. If the path provided is a directory, the\n latest checkpoint will be used.\n do_eval (bool): If do_eval is True and the dataset has a validation subset, the model will be evaluated\n early_stopping (bool): Enable early stopping if convergence is reached while training\n at the end of each epoch.\n lr_decay (bool): If lr_decay is True and do_eval is True, learning rate decay on the validation loss\n is applied at the end of each epoch.\n seed (int): Optionally set a seed for reproducibility.\n ipex_optimize (bool): Use Intel Extension for PyTorch (IPEX). Defaults to False.\n distributed (bool): Boolean flag to use distributed training. Defaults to False.\n nnodes (int): Number of nodes to use for distributed training. Defaults to 1.\n nproc_per_node (int): Number of processes to spawn per node to use for distributed training. Defaults\n to 1.\n hosts (str): hosts list for distributed training. Defaults to None.\n hostfile (str): Name of the hostfile for distributed training. Defaults to None.\n shared_dir (str): The shared data dir for distributed training.\n temp_dir (str): The temp data dir at local.\n\n Returns:\n Trained PyTorch model object\n \"\"\"\n self._check_train_inputs(\n output_dir, dataset, ImageClassificationDataset,\n epochs, initial_checkpoints)\n\n dataset_num_classes = len(dataset.class_names)\n\n # Check that the number of classes matches the model outputs\n if dataset_num_classes != self.num_classes:\n raise RuntimeError(\"The number of model outputs ({}) differs from the number of dataset classes ({})\".\n format(self.num_classes, dataset_num_classes))\n\n self._set_seed(seed)\n\n self._optimizer = self._optimizer_class(self._model.parameters(), lr=self._learning_rate)\n\n if initial_checkpoints:\n checkpoint = torch.load(initial_checkpoints)\n self._model.load_state_dict(checkpoint['model_state_dict'])\n self._optimizer.load_state_dict(checkpoint['optimizer_state_dict'])\n\n if distributed:\n objects_path = self.save_objects(\n dataset, shared_dir, temp_dir)\n batch_size = dataset._preprocessed['batch_size']\n self._fit_distributed(\n nnodes, nproc_per_node, hosts, hostfile,\n epochs, batch_size, ipex_optimize,\n objects_path)\n else:\n # Call ipex.optimize\n if ipex_optimize:\n import intel_extension_for_pytorch as ipex\n self._model, self._optimizer = ipex.optimize(self._model, optimizer=self._optimizer)\n self._fit(output_dir, dataset, epochs, do_eval, early_stopping, lr_decay)\n\n return self._history\n\n def evaluate(self, dataset: ImageClassificationDataset, use_test_set=False):\n \"\"\"\n Evaluate the accuracy of the model on a dataset.\n If there is a validation set, evaluation will be done on it (by default) or on the test set\n (by setting use_test_set=True). Otherwise, the entire non-partitioned dataset will be\n used for evaluation.\n \"\"\"\n if use_test_set:\n if dataset.test_subset:\n eval_loader = dataset.test_loader\n data_length = len(dataset.test_subset)\n else:\n raise ValueError(\"No test subset is defined\")\n elif dataset.validation_subset:\n eval_loader = dataset.validation_loader\n data_length = len(dataset.validation_subset)\n else:\n eval_loader = dataset.data_loader\n data_length = len(dataset.dataset)\n\n model = self._model\n optimizer = self._optimizer\n\n # Do the evaluation\n device = torch.device(self._device)\n model = model.to(device)\n\n model.eval()\n running_loss = 0.0\n running_corrects = 0\n\n # Iterate over data.\n for inputs, labels in tqdm(eval_loader, bar_format='{l_bar}{bar:50}{r_bar}{bar:-50b}'):\n inputs = inputs.to(device)\n labels = labels.to(device)\n\n # Zero the parameter gradients\n optimizer.zero_grad()\n\n # Forward pass\n with torch.set_grad_enabled(False):\n outputs = model(inputs)\n _, preds = torch.max(outputs, 1)\n loss = self._loss(outputs, labels)\n\n # Statistics\n running_loss += loss.item() * inputs.size(0)\n running_corrects += torch.sum(preds == labels.data)\n\n epoch_loss = running_loss / data_length\n epoch_acc = float(running_corrects) / data_length\n\n print(f'Validation Loss: {epoch_loss:.4f} Acc: {epoch_acc:.4f}')\n\n return [epoch_loss, epoch_acc]\n\n def predict(self, input_samples, return_type='class'):\n \"\"\"\n Perform feed-forward inference and predict the classes of the input_samples.\n\n Args:\n input_samples (tensor): Input tensor with one or more samples to perform inference on\n return_type (str): Using 'class' will return the highest scoring class (default), using 'scores' will\n return the raw output/logits of the last layer of the network, using 'probabilities' will\n return the output vector after applying a softmax function (so results sum to 1)\n\n Returns:\n List of classes, probability vectors, or raw score vectors\n\n Raises:\n ValueError if the return_type is not one of 'class', 'probabilities', or 'scores'\n \"\"\"\n return_types = ['class', 'probabilities', 'scores']\n if not isinstance(return_type, str) or return_type not in return_types:\n raise ValueError('Invalid return_type ({}). Expected one of {}.'.format(return_type, return_types))\n\n self._model.eval()\n with torch.no_grad():\n predictions = self._model(input_samples)\n if return_type == 'class':\n _, predicted_ids = torch.max(predictions, 1)\n return predicted_ids\n elif return_type == 'probabilities':\n return torch.nn.functional.softmax(predictions)\n else:\n return predictions\n\n def export(self, output_dir):\n \"\"\"\n Save a serialized version of the model to the output_dir path\n \"\"\"\n if self._model:\n # Save the model in a format that can be re-loaded for inference\n verify_directory(output_dir)\n valid_model_name = validate_model_name(self.model_name)\n saved_model_dir = os.path.join(output_dir, valid_model_name)\n if os.path.exists(saved_model_dir) and len(os.listdir(saved_model_dir)):\n saved_model_dir = os.path.join(saved_model_dir, \"{}\".format(len(os.listdir(saved_model_dir)) + 1))\n else:\n saved_model_dir = os.path.join(saved_model_dir, \"1\")\n verify_directory(saved_model_dir)\n model_copy = dill.dumps(self._model)\n torch.save(model_copy, os.path.join(saved_model_dir, 'model.pt'))\n print(\"Saved model directory:\", saved_model_dir)\n\n return saved_model_dir\n else:\n raise ValueError(\"Unable to export the model, because it hasn't been trained yet\")\n\n def export_neural_compressor_config(\n self, config_file_path, dataset, batch_size, overwrite=False,\n resize_interpolation='bicubic', accuracy_criterion_relative=0.01, exit_policy_timeout=0,\n exit_policy_max_trials=50, tuning_random_seed=9527,\n tuning_workspace=''):\n \"\"\"\n Writes a neural compressor compatible config file to the specified path\n using args from the specified dataset and parameters.\n\n Args:\n config_file_path (str): Destination path on where to write the .yaml config file.\n dataset (Dataset): A dataset object\n batch_size (int): Batch size to use for quantization and evaluation\n overwrite (bool): Specify whether or not to overwrite the config_file_path, if it already exists\n (default: False)\n resize_interpolation (str): Interpolation type. Select from: 'bilinear', 'nearest', 'bicubic'\n (default: bicubic)\n accuracy_criterion_relative (float): Relative accuracy loss (default: 0.01, which is 1%)\n exit_policy_timeout (int): Tuning timeout in seconds (default: 0). Tuning processing finishes when the\n timeout or max_trials is reached. A tuning timeout of 0 means that the tuning\n phase stops when the accuracy criterion is met.\n exit_policy_max_trials (int): Maximum number of tuning trials (default: 50). Tuning processing finishes when\n the timeout or or max_trials is reached.\n tuning_random_seed (int): Random seed for deterministic tuning (default: 9527).\n tuning_workspace (dir): Path the neural compressor nc_workspace folder.\n If the string is empty and the OUTPUT_DIR env var is set,\n that output directory will be used.\n If the string is empty and the OUTPUT_DIR env var is not set,\n the default neural compressor nc_workspace location will be used.\n Returns:\n None\n Raises:\n FileExistsError if the config file already exists and overwrite is set to False.\n ValueError if the parameters are not within the expected values\n NotImplementedError if the dataset type is not TFCustomImageClassificationDataset.\n \"\"\"\n if os.path.isfile(config_file_path) and not overwrite:\n raise FileExistsError('A file already exists at: {}. Provide a new file path or set overwrite=True',\n config_file_path)\n\n # We can setup the a custom dataset to use the ImageFolder dataset option in neural compressor.\n # They don't have a PyTorch Dataset option, so for now, we only support custom datasets for quantization\n if dataset is not PyTorchImageClassificationDataset \\\n and type(dataset) != PyTorchImageClassificationDataset:\n raise NotImplementedError('quantization has only been implemented for PyTorch image classification models '\n 'with custom datasets')\n\n if batch_size and not isinstance(batch_size, int) or batch_size < 1:\n raise ValueError('Invalid value for batch size ({}). Expected a positive integer.'.format(batch_size))\n\n if resize_interpolation not in ['bilinear', 'nearest', 'bicubic']:\n raise ValueError('Invalid value for resize interpolation ({}). Expected one of the following values: '\n 'bilinear, nearest, bicubic'.format(resize_interpolation))\n\n if accuracy_criterion_relative and not isinstance(accuracy_criterion_relative, float) or \\\n not (0.0 <= accuracy_criterion_relative <= 1.0):\n raise ValueError('Invalid value for the accuracy criterion ({}). Expected a float value between 0.0 '\n 'and 1.0'.format(accuracy_criterion_relative))\n\n if exit_policy_timeout and not isinstance(exit_policy_timeout, int) or exit_policy_timeout < 0:\n raise ValueError('Invalid value for the exit policy timeout ({}). Expected a positive integer or 0.'.\n format(exit_policy_timeout))\n\n if exit_policy_max_trials and not isinstance(exit_policy_max_trials, int) or exit_policy_max_trials < 1:\n raise ValueError('Invalid value for max trials ({}). Expected an integer greater than 0.'.\n format(exit_policy_timeout))\n\n if tuning_random_seed and not isinstance(tuning_random_seed, int) or tuning_random_seed < 0:\n raise ValueError('Invalid value for tuning random seed ({}). Expected a positive integer.'.\n format(tuning_random_seed))\n\n if not isinstance(tuning_workspace, str):\n raise ValueError('Invalid value for the nc_workspace directory. Expected a string.')\n\n # Get the image recognition Neural Compressor template\n config_template = ImageClassificationModel.get_neural_compressor_config_template(self)\n\n # Collect the different data loaders into a list, so that we can update them all the with the data transforms\n dataloader_configs = []\n\n # If tuning_workspace is undefined, use the OUTPUT_DIR, if the env var exists\n if not tuning_workspace:\n output_dir_env_var = os.getenv('OUTPUT_DIR', '')\n\n if output_dir_env_var:\n tuning_workspace = os.path.join(output_dir_env_var, 'nc_workspace')\n\n print(\"tuning_workspace:\", tuning_workspace)\n\n if \"quantization\" in config_template.keys() and \"calibration\" in config_template[\"quantization\"].keys() \\\n and \"dataloader\" in config_template[\"quantization\"][\"calibration\"].keys():\n dataloader_configs.append(config_template[\"quantization\"][\"calibration\"][\"dataloader\"])\n\n if \"evaluation\" in config_template.keys():\n if \"accuracy\" in config_template[\"evaluation\"].keys() and \\\n \"dataloader\" in config_template[\"evaluation\"][\"accuracy\"].keys():\n dataloader_configs.append(config_template[\"evaluation\"][\"accuracy\"][\"dataloader\"])\n if \"performance\" in config_template[\"evaluation\"].keys() and \\\n \"dataloader\" in config_template[\"evaluation\"][\"performance\"].keys():\n dataloader_configs.append(config_template[\"evaluation\"][\"performance\"][\"dataloader\"])\n\n transform_config = {\n \"Resize\": {\n \"size\": self._image_size\n },\n \"CenterCrop\": {\n \"size\": self._image_size\n },\n \"ToTensor\": {},\n \"Normalize\": {\n \"mean\": [0.485, 0.456, 0.406],\n \"std\": [0.229, 0.224, 0.225]\n }\n }\n\n del config_template[\"evaluation\"][\"accuracy\"][\"postprocess\"]\n\n config_template[\"quantization\"][\"approach\"] = \"post_training_dynamic_quant\"\n\n # Update the data loader configs\n for dataloader_config in dataloader_configs:\n # Set the transform configs for resizing and rescaling\n dataloader_config[\"transform\"] = copy.deepcopy(transform_config)\n\n # Update dataset directory for the custom dataset\n if \"dataset\" in dataloader_config.keys() and \"ImageFolder\" in dataloader_config[\"dataset\"].keys():\n dataloader_config[\"dataset\"][\"ImageFolder\"][\"root\"] = dataset.dataset_dir\n\n dataloader_config[\"batch_size\"] = batch_size\n\n if \"tuning\" in config_template.keys():\n config_template[\"tuning\"][\"accuracy_criterion\"][\"relative\"] = accuracy_criterion_relative\n\n if exit_policy_timeout is None:\n config_template[\"tuning\"][\"exit_policy\"].pop('timeout', None)\n else:\n config_template[\"tuning\"][\"exit_policy\"][\"timeout\"] = exit_policy_timeout\n\n if exit_policy_max_trials is None:\n config_template[\"tuning\"][\"exit_policy\"].pop('max_trials', None)\n else:\n config_template[\"tuning\"][\"exit_policy\"][\"max_trials\"] = exit_policy_max_trials\n\n if tuning_random_seed is None:\n config_template[\"tuning\"].pop('random_seed', None)\n else:\n config_template[\"tuning\"][\"random_seed\"] = tuning_random_seed\n\n if tuning_workspace:\n if \"workspace\" not in config_template[\"tuning\"].keys():\n config_template[\"tuning\"][\"workspace\"] = {}\n\n config_template[\"tuning\"][\"workspace\"][\"path\"] = tuning_workspace\n else:\n # No tuning_workspace is defined, so remove it from the config\n if \"workspace\" in config_template[\"tuning\"].keys():\n config_template[\"tuning\"][\"workspace\"].pop(\"path\", None)\n\n if len(config_template[\"tuning\"][\"workspace\"].keys()) == 0:\n config_template[\"tuning\"].pop(\"workspace\", None)\n\n # Create the directory where the file will be written, if it doesn't already exist\n if not os.path.exists(os.path.dirname(config_file_path)):\n os.makedirs(os.path.dirname(config_file_path))\n\n # Write the config file\n with open(config_file_path, \"w\") as config_file:\n yaml.dump(config_template, config_file, sort_keys=False)\n\n def quantize(self, saved_model_dir, output_dir, inc_config_path):\n \"\"\"\n Performs post training quantization using the Neural Compressor on the model from the saved_model_dir\n using the specified config file. The quantized model is written to the output directory.\n\n Args:\n saved_model_dir (str): Source directory for the model to quantize.\n output_dir (str): Writable output directory to save the quantized model\n inc_config_path (str): Path to a neural compressor config file (.yaml)\n\n Returns:\n None\n\n Raises:\n NotADirectoryError if the model is not a directory\n FileNotFoundError if a model.pt is not found in the model or if the inc_config_path file\n is not found.\n FileExistsError if the output_dir already has a model.pt file\n \"\"\"\n # The saved model directory should exist and contain a model.pt file\n if not os.path.isdir(saved_model_dir):\n raise NotADirectoryError(\"The saved model directory ({}) does not exist.\".format(saved_model_dir))\n if not os.path.isfile(os.path.join(saved_model_dir, \"model.pt\")):\n raise FileNotFoundError(\"The saved model directory ({}) should have a model.pt file\".format(\n saved_model_dir))\n\n # Verify that the config file exists\n if not os.path.isfile(inc_config_path):\n raise FileNotFoundError(\"The config file was not found at: {}\".format(inc_config_path))\n\n if not os.path.exists(output_dir):\n os.makedirs(output_dir)\n else:\n # Verify that the output directory doesn't already have a saved_model.pb file\n if os.path.exists(os.path.join(output_dir, \"model.pt\")):\n raise FileExistsError(\"A saved model already exists at:\", os.path.join(output_dir, \"model.pt\"))\n\n from neural_compressor.experimental import Quantization\n quantizer = Quantization(inc_config_path)\n quantizer.model = self._model\n quantized_model = quantizer.fit()\n\n # If quantization was successful, save the model\n if quantized_model:\n quantized_model.save(output_dir)\n import subprocess\n # Change the model filename from best_model.pt to model.pt to match our convention\n p = subprocess.Popen([\"mv\", output_dir + \"/best_model.pt\", output_dir + \"/model.pt\"],\n stdout=subprocess.PIPE)\n stdout, stderr = p.communicate()\n\n def benchmark_neural_compressor(self, saved_model_dir, inc_config_path, mode='performance', model_type='fp32'):\n \"\"\"\n Use neural compressor to benchmark the specified model for performance or accuracy. You must specify whether the\n input model is fp32 or int8. IPEX int8 models are not supported yet.\n\n Args:\n saved_model_dir (str): Path to the directory where the saved model is located\n inc_config_path (str): Path to a neural compressor config file (.yaml)\n mode (str): Performance or accuracy (defaults to performance)\n model_type (str): Floating point (fp32) or quantized integer (int8) model type\n Returns:\n None\n Raises:\n NotADirectoryError if the saved_model_dir is not a directory\n FileNotFoundError if a model.pt is not found in the saved_model_dir or if the inc_config_path file\n is not found.\n ValueError if an unexpected mode is provided\n \"\"\"\n # The saved model directory should exist and contain a model.pt file\n if not os.path.isdir(saved_model_dir):\n raise NotADirectoryError(\"The saved model directory ({}) does not exist.\".format(saved_model_dir))\n if not os.path.isfile(os.path.join(saved_model_dir, \"model.pt\")):\n raise FileNotFoundError(\"The saved model directory ({}) should have a model.pt file\".format(\n saved_model_dir))\n\n # Validate mode\n if mode not in ['performance', 'accuracy']:\n raise ValueError(\"Invalid mode: {}. Expected mode to be 'performance' or 'accuracy'.\".format(mode))\n\n # Verify that the config file exists\n if not os.path.isfile(inc_config_path):\n raise FileNotFoundError(\"The config file was not found at: {}\".format(inc_config_path))\n\n from neural_compressor.experimental import Benchmark, common\n if model_type == \"fp32\":\n evaluator = Benchmark(inc_config_path)\n evaluator.model = self._model\n return evaluator(mode)\n elif model_type == \"int8\":\n try:\n from neural_compressor.utils.pytorch import load\n evaluator = Benchmark(inc_config_path)\n evaluator.model = common.Model(load(os.path.join(saved_model_dir, 'model.pt'), self._model))\n return evaluator(mode)\n except AssertionError:\n raise NotImplementedError(\"This model type is not yet supported by neural compressor benchmarking\")\n","repo_name":"oap-project/cloudtik","sub_path":"python/cloudtik/runtime/ai/modeling/transfer_learning/image_classification/pytorch/image_classification_model.py","file_name":"image_classification_model.py","file_ext":"py","file_size_in_byte":32828,"program_lang":"python","lang":"en","doc_type":"code","stars":23,"dataset":"github-code","pt":"79"} +{"seq_id":"19419316597","text":"import librosa\nimport librosa.display\nfrom python_speech_features import mfcc, logfbank\n\nimport matplotlib.pyplot as plt\n\nimport numpy as np\nimport pandas as pd\nimport os\n\n\ndef data_source(x):\n if x.split('/')[1] == 'tess':\n return 'tess'\n elif x.split('/')[1] == 'savee':\n return 'savee'\n else:\n return 'ravdess'\n \n\ndef eda_data(df):\n df = pd.read_csv('data/data.csv')\n df = df[['filename', 'emotion']]\n df['emotion'] = df['emotion'].str.split('_')\n df['gender'] = df['emotion'].apply(lambda x : x[0])\n df['emotion'] = df['emotion'].apply(lambda x: x[1])\n df['type'] = df['filename'].apply(lambda x:data_source(x))\n \n return df[['filename', 'emotion', 'gender', 'type']]\n\n\ndef distributions(df, col, title):\n count = df.groupby(col).count()['filename']\n fig, ax = plt.subplots(figsize=(6,6))\n ax.pie(count,labels= count.index.str.capitalize(), autopct='%1.1f%%', textprops={'fontsize': 13})\n ax.set_title('Distribution of '+title, fontsize=15)\n\n\ndef calc_fft(y, rate):\n n = len(y)\n freq = np.fft.rfftfreq(n, d=1/rate) \n Y = abs(np.fft.rfft(y)/n) #magnitude -- also /n so that it normalizes/scales the signal\n return (Y, freq)\n\n\ndef frequencies(df):\n signals = {}\n fft = {}\n fbank = {}\n mfccs = {}\n emotion = 'happy sad angry neutral surprised fearful disgust'\n for i in emotion.split(' '):\n wav_file = df[df['emotion'] == i].iloc[0,0]\n signal, rate = librosa.load(wav_file, mono=True, duration=3, offset = .5) #wav file can detect sampling rate\n signals[i] = signal\n fft[i] = calc_fft(signal, rate)\n\n bank = logfbank(signal[:rate], rate, nfilt=26, nfft=1103).T #nfft=sampling_frequency/40, nfilt always 26\n fbank[i] = bank\n mel = mfcc(signal[:rate], rate, numcep=13, nfilt=26, nfft=1103).T #numcep = usually 1/2 of nfilt\n mfccs[i] = mel\n return signals, fft, fbank, mfccs","repo_name":"lbalido/audio_sentiment","sub_path":"src/eda.py","file_name":"eda.py","file_ext":"py","file_size_in_byte":1931,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"29877068197","text":"from flask import Flask, render_template\nimport pymysql\n\n\napp = Flask(__name__,template_folder='templates',static_folder=\"static\")\n\n\n@app.route('/')\ndef index():\n return render_template(\"index.html\") #主页面时\n\n\n@app.route('/index') #分发路由\ndef home():\n return index()\n\n\n@app.route('/movie')\ndef movie():\n datalist = []\n con = pymysql.connect( #连接\n host='localhost',\n port=3306,\n user='root',\n passwd='111111',\n db='movie',\n charset='utf8'\n )\n cur = con.cursor() #新建光标\n sql = \"select * from movies\"\n data = cur.execute(sql)\n result = cur.fetchall() #每执行一次退后一行,取到一行\n for item in result:\n datalist.append(item)\n cur.close()\n con.close()\n print(datalist)\n return render_template(\"movie.html\", movies=datalist) #传给movies\n\n\n@app.route('/score')\ndef score():\n score = []\n num = []\n conn = pymysql.Connect(\n host='localhost',\n port=3306,\n user='root',\n passwd='111111',\n db='movie',\n charset='utf8'\n )\n cur = conn.cursor()\n sql = \"select * from movie_score_num\"\n data = cur.execute(sql)\n result = cur.fetchall()\n for item in result:\n score.append(str(item[0]))\n num.append(item[1])\n\n cur.close()\n conn.close()\n return render_template(\"score.html\", score=score, num=num)\n\n\n@app.route('/country')\ndef country():\n country = []\n num = []\n conn = pymysql.Connect(\n host='localhost',\n port=3306,\n user='root',\n passwd='111111',\n db='movie',\n charset='utf8'\n )\n cur = conn.cursor()\n sql = \"select * from movie_country_num\"\n data = cur.execute(sql)\n result = cur.fetchall()\n for item in result:\n country.append(str(item[0]))\n num.append(item[1])\n\n cur.close()\n conn.close()\n return render_template(\"country.html\", country=country, num=num)\n\n\n\n\n\n@app.route('/year')\ndef year():\n year = []\n num = []\n s = []\n conn = pymysql.Connect(\n host='localhost',\n port=3306,\n user='root',\n passwd='111111',\n db='movie',\n charset='utf8'\n )\n cur = conn.cursor()\n sql = \"select * from movie_year_num\"\n data = cur.execute(sql)\n result = cur.fetchall()\n for item in result:\n s.append(item)\n year.append(str(item[0]))\n num.append(item[1])\n\n cur.close()\n conn.close()\n return render_template(\"year.html\", year=year, num=num)\n\n\n\n\n@app.route('/type')\ndef type():\n type = []\n num = []\n dataset=[]\n conn = pymysql.Connect(\n host='localhost',\n port=3306,\n user='root',\n passwd='111111',\n db='movie',\n charset='utf8'\n )\n cur = conn.cursor()\n sql = \"select * from movie_type_num\"\n data = cur.execute(sql)\n result = cur.fetchall()\n for item in result:\n type.append(str(item[0]))\n num.append(item[1])\n for i in range(len(type)):\n dataset.append({'name':type[i],'value':num[i]})\n\n cur.close()\n conn.close()\n return render_template(\"type.html\", dataset=dataset,type=type,num=num)\n\n\n\n@app.route('/comment') #词云图\ndef word():\n return render_template(\"comment.html\")\n\n\n@app.route('/performance') #展示界面\ndef other():\n return render_template(\"performance.html\")\n\n\n\n\nif __name__ == '__main__':\n app.run()\n app.debug=True","repo_name":"Doechen/Python-Movie-Data-Analysis-Platform","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3444,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6837411792","text":"import requests\nimport json\n\ntoken = \"c3922bf6cf1313e37768687dd70c3b21\"\ntokens = {\"token\": token}\nheader = {'Content-Type': 'application/json'}\nprefixHTTP = 'http://challenge.code2040.org/api/prefix'\n\nchallenge_endpoint = \"http://challenge.code2040.org/api/prefix/validate\"\nresponse1 = requests.post(prefixHTTP, data=json.dumps(tokens),headers=header).json()\n\n#def task_two():\n \n#print \"Response from API: \" + response1.text\nprefix = response1[\"prefix\"]\nmy_array = response1[\"array\"]\nresult_array = []\n\nfor item in my_array:\n\tif not item.startswith(prefix):\n\t\tresult_array.append(str(item))\n\nprefixResult = {\"token\": token,\"array\":result_array}\nresponse1 = requests.post(challenge_endpoint, data=json.dumps(prefixResult),headers=header)\nprint( response1.text)\n","repo_name":"Yeathtsme/code2040","sub_path":"step4Code2040.py","file_name":"step4Code2040.py","file_ext":"py","file_size_in_byte":763,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"16960121293","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n# GNU General Public License v3.0+\n# (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\n\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ibmcloud_python_sdk.resource import resource_instance as sdk\n\n\nANSIBLE_METADATA = {\n 'metadata_version': '1.1',\n 'status': ['preview'],\n 'supported_by': 'community'\n}\n\nDOCUMENTATION = r'''\n---\nmodule: ic_resource_instance_info\nshort_description: Retrieve resource instances on IBM Cloud.\nauthor: James Regis (@jregis)\nversion_added: \"2.9\"\ndescription:\n - This module retrieve information about resource instances.\nnotes:\n - The result contains a list of resource instances.\nrequirements:\n - \"ibmcloud-python-sdk\"\noptions:\n instance:\n description:\n - Restrict results to a resource instance with GUID or name matching.\n'''\n\nEXAMPLES = r'''\n- name: Retrieve resource instance list\n ic_resource_instance_info:\n\n- name: Retrieve specific resource instance\n ic_resource_instance_info\n instance: ibmcloud-ri-baby\n'''\n\n\ndef run_module():\n module_args = dict(\n instance=dict(\n type='str',\n default=None,\n required=False),\n )\n\n module = AnsibleModule(\n argument_spec=module_args,\n supports_check_mode=False\n )\n\n resource_instance = sdk.ResourceInstance()\n\n instance = module.params['instance']\n\n if instance:\n result = resource_instance.get_resource_instance(instance)\n if \"errors\" in result:\n module.fail_json(msg=result)\n else:\n result = resource_instance.get_resource_instances()\n if \"errors\" in result:\n module.fail_json(msg=result)\n\n module.exit_json(**result)\n\n\ndef main():\n run_module()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"goldyfruit/ibmcloud-ansible-collection","sub_path":"plugins/modules/ic_resource_instance_info.py","file_name":"ic_resource_instance_info.py","file_ext":"py","file_size_in_byte":1787,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"71814435456","text":"#addorders.py\n\nimport tkinter\nfrom tkinter import * \nfrom tkinter import messagebox\nfrom functools import partial \nimport sqlite3 \nimport pymysql\nfrom venv import create\nfrom tkinter.ttk import *\nfrom tkcalendar import Calendar, DateEntry\nimport re\nfrom datetime import datetime\nimport pandas as pd\nfrom pandastable import Table, TableModel\n\n# Saving student form data\ndef addOrders(): \n # Checking last input for validating , if not validated , shows error message\n if (len(ord_desc.get()) != 3 or ord_desc.get().isalpha()) == False:\n messagebox.showinfo(\"Save\" , \"Not Validated!\")\n return\n \n try:\n # processing for three date variables : Birth Date , Start Date , End Date\n ordDateObj = datetime.strptime(ord_date.get(), '%d/%m/%Y')\n\n # Connecetion for mysql database\n conn = pymysql.connect(user=\"root\", password=\"\", host=\"localhost\", database=\"sunville\")\n cur = conn.cursor()\n # Excuting insert query \n cur.execute(\"\"\"insert into orders(ORD_NUM ,ORD_AMOUNT,ADVANCE_AMOUNT, ORD_DATE,CUST_CODE,AGENT_CODE,ORD_DESCRIPTION) values(%s,%s,%s,%s,%s,%s,%s)\"\"\" ,\n (ord_num.get(),ord_amt.get(),adv_amt.get(),ordDateObj.strftime(\"%Y-%m-%d\"),cust_code.get(),agent_code.get(),ord_desc.get()))\n conn.close()\n # Show message for successing\n messagebox.showinfo(\"Save\" , \"Success!\")\n b1['state']= 'enabled'\n # Initializing for each input.\n ord_amtEntry.config(state='disabled')\n adv_amtEntry.config(state='disabled')\n cust_codeEntry.config(state='disabled')\n agent_codeEntry.config(state='disabled')\n ord_descEntry.config(state='disabled')\n except Exception as e:\n print(e)\n # If error on saving , shows error message.\n messagebox.showerror(\"Save\" ,\"Failed to save!\")\n# When clicking cancel button , application will be closed.\ndef cancel():\n ord_num.set(\"\")\n ord_amt.set(\"\")\n adv_amt.set(\"\") \n cust_code.set(\"\")\n agent_code.set(\"\")\n ord_desc.set(\"\")\n root.destroy()\n\n# Validating for each input\ndef validate(event , input):\n if( input == \"Order Number\"):\n ord_number = ord_num.get()\n if (len(ord_number) != 6 or ord_number.isdigit()==False):\n messagebox.showerror(\"Invalid!\" ,\"Order number has to be a 6 digit number.\")\n ord_numEntry.focus_set()\n else:\n ord_amtEntry.focus_set()\n ord_amtEntry.config(state='normal')\n elif(input == \"Order Amount\"):\n if (ord_amt.get().isdigit()==True and float(ord_amt.get()) > 0 ):\n adv_amtEntry.focus_set()\n adv_amtEntry.config(state='normal')\n else:\n messagebox.showerror(\"Invalid!\" ,\"Order amount has to be a number greater than 0.\")\n ord_amtEntry.focus_set()\n elif( input == \"Advance Amount\"):\n if (adv_amt.get().isdigit()==True and float(adv_amt.get()) >= 0 and float(adv_amt.get()) <=float(ord_amt.get()) ):\n cust_codeEntry.focus_set()\n cust_codeEntry.config(state='normal')\n else:\n messagebox.showerror(\"Invalid!\" ,\"Advance amount has to be a number greater than 0 and not greater than Order Amount.\")\n adv_amtEntry.focus_set()\n elif( input == \"Customer Code\"):\n if (len(cust_code.get()) == 6 and cust_code.get().isalnum()==True):\n agent_codeEntry.focus_set()\n agent_codeEntry.config(state='normal')\n else:\n messagebox.showerror(\"Invalid!\" ,\"Customer code has to be alphanumberic with length 6.\")\n cust_codeEntry.focus_set()\n elif( input == \"Agent Code\"):\n if (len(agent_code.get()) == 4 and agent_code.get().isalnum()==True):\n ord_descEntry.focus_set()\n ord_descEntry.config(state='normal')\n else:\n messagebox.showerror(\"Invalid!\" ,\"Agent code has to be alphanumberic with length 4.\")\n agent_codeEntry.focus_set() \n elif( input == \"Order Description\"):\n if len(ord_desc.get()) == 3 and ord_desc.get().isalpha():\n pass\n else:\n messagebox.showerror(\"Invalid!\" ,\"Order description has to be only three letters long.\")\n ord_descEntry.focus_set() \n\ndef readdata():\n readwindow = Tk()\n readwindow.title(\"Read Order Data\")\n readwindow.geometry('{}x{}'.format(800, 600))\n \n mainframe1 = Frame(readwindow)\n l = Label(readwindow, text='Here are the Results',font=('times', 20, 'bold'),background = '#154360',foreground = '#FDFEFE')\n l.place(x = 200, y = 10)\n \n df = pd.DataFrame() \n df = TableModel.getSampleData()\n \n \n conn = pymysql.connect(user=\"root\", password=\"\", host=\"localhost\", database=\"sunville\")\n cur = conn.cursor()\n # Excuting insert query \n query = \"select ORD_NUM,ORD_DATE,ORD_DESCRIPTION,CUST_CODE,AGENT_CODE from orders\"\n cur.execute(query) \n df = pd.DataFrame(list(cur.fetchall()),columns =['ORD_NUM','ORD_DATE','ORD_DESCRIPTION','CUST_CODE','AGENT_CODE'])\n #print (df) \n \n table =Table(mainframe1, dataframe=df,showtoolbar=True, showstatusbar=True )\n table.currwidth = 700\n table.currheight = 500\n mainframe1.place(x = 200,y =200,anchor = \"w\")\n try:\n table.show() \n except:\n pass\n conn.close()\n\n\n# Creating main window and setting with width and height \nroot = Tk()\nroot.title(\"Orders\")\nroot.geometry('{}x{}'.format(600, 500))\nmainframe = Frame(root)\nmainframe.pack()\n\n# Setting string variable for 6 input\nord_num = StringVar()\nord_amt = StringVar()\nadv_amt = StringVar()\ncust_code = StringVar()\nagent_code = StringVar()\nord_desc = StringVar()\n# Input for Order Number\nord_numEntry = Entry(mainframe, width=20, textvariable=ord_num)\nord_numEntry.grid(row=0, column=1 ,padx=5, pady=5)\nord_numEntry.bind(\"\", lambda event: validate(event, \"Order Number\"))\nord_numEntry.bind(\"\", lambda event: validate(event, \"Order Number\"))\n# Input for Order Amount\nord_amtEntry = Entry(mainframe, width=20, textvariable=ord_amt)\nord_amtEntry.grid(row=1, column=1 ,padx=5, pady=5)\nord_amtEntry.bind(\"\", lambda event: validate(event, \"Order Amount\"))\nord_amtEntry.bind(\"\", lambda event: validate(event, \"Order Amount\"))\n# Input for Advance Amount\nadv_amtEntry = Entry(mainframe, width=20, textvariable=adv_amt)\nadv_amtEntry.grid(row=2, column=1 ,padx=5, pady=5)\nadv_amtEntry.bind(\"\", lambda event: validate(event, \"Advance Amount\"))\nadv_amtEntry.bind(\"\", lambda event: validate(event, \"Advance Amount\"))\n# Input for Customer Code\ncust_codeEntry = Entry(mainframe, width=20, textvariable=cust_code)\ncust_codeEntry.grid(row=3, column=1 ,padx=5, pady=5)\ncust_codeEntry.bind(\"\", lambda event: validate(event, \"Customer Code\"))\ncust_codeEntry.bind(\"\", lambda event: validate(event, \"Customer Code\"))\n# Input for Agent Code\nagent_codeEntry = Entry(mainframe, width=20, textvariable=agent_code)\nagent_codeEntry.grid(row=4, column=1 ,padx=5, pady=5)\nagent_codeEntry.bind(\"\", lambda event: validate(event, \"Agent Code\"))\nagent_codeEntry.bind(\"\", lambda event: validate(event, \"Agent Code\"))\n# Input for Order Description\nord_descEntry = Entry(mainframe, width=20, textvariable=ord_desc)\nord_descEntry.grid(row=5, column=1 ,padx=5, pady=5)\nord_descEntry.bind(\"\", lambda event: validate(event, \"Order Description\"))\nord_descEntry.bind(\"\", lambda event: validate(event, \"Order Description\"))\n\n# First rest 5 inputs will be disabled for checking validation\n\nord_amtEntry.config(state='disabled')\nadv_amtEntry.config(state='disabled')\ncust_codeEntry.config(state='disabled')\nagent_codeEntry.config(state='disabled')\nord_descEntry.config(state='disabled')\n\n# Date picker for start date , end date\nord_date = StringVar()\nDateEntry(mainframe , textvariable = ord_date , date_pattern='dd/mm/y' ).grid(row=7, column=1, padx=5, pady=5)\n\n# Setting labels for each input\nLabel(mainframe, text='Order Number:*', anchor='w').grid(row=0, column=0 ,padx=5, pady=5, sticky=\"w\")\nLabel(mainframe, text='Order Amount:*', anchor='w').grid(row=1, column=0 ,padx=5, pady=5, sticky=\"w\")\nLabel(mainframe, text='Advance Amount:*', anchor='w').grid(row=2, column=0 ,padx=5, pady=5, sticky=\"w\")\nLabel(mainframe, text='Customer Code:*', anchor='w').grid(row=3, column=0 ,padx=5, pady=5, sticky=\"w\")\nLabel(mainframe, text='Agent Code:', anchor='w').grid(row=4, column=0 ,padx=5, pady=5, sticky=\"w\")\nLabel(mainframe, text='Order Description:', anchor='w').grid(row=5, column=0 ,padx=5, pady=5, sticky=\"w\")\n\nLabel(mainframe, text='Order Date:*', anchor='w').grid(row=7, column=0 ,padx=5, pady=5, sticky=\"w\")\n\n# Buttons for submit and cancel \nbtnFrame = Frame(mainframe)\nButton(btnFrame, text=\"Submit\", command=addOrders).grid(row=0, column=1, padx=5, pady=5)\nButton(btnFrame, text=\"Cancel\", command=cancel).grid(row=0, column=2, padx=5, pady=5)\nb1 = Button(btnFrame, text=\"Read Data\",command= readdata )\n\nb1.grid(row=0, column=3, padx=5, pady=5 )\nbtnFrame.grid(row=10, column=1, padx=5, pady=5)\n\n\nroot.mainloop()\n","repo_name":"d9vp/Python-Internship-Project","sub_path":"addorders.py","file_name":"addorders.py","file_ext":"py","file_size_in_byte":9071,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"41503439187","text":"print (\"\\nCalculando a area ou o perimetro de um retangulo\")\r\nwhile \"Escolha\":\r\n\r\n Escolha = input(\"Voce quer calcular a Area ou o Perimetro? \")\r\n\r\n \r\n if Escolha == \"Area\":\r\n print (\"\\nCalculando a area de um retangulo\\n Area é igual a Base vezes a altura ou a=b.h\\n Caso você queira, use o calculador abaixo\")\r\n base = int(input(\"Qual a base deste retângulo? \"))\r\n altura = int(input(\"Qual a altura deste retângulo? \"))\r\n\r\n area = base * altura\r\n\r\n print (\"\\tSeu retangulo tem área de %d cm²\\n\" % (area))\r\n \r\n\r\n\r\n\r\n elif Escolha == \"Perimetro\":\r\n print (\"\\nCalculando o perimetro de um retangulo\\n O perimetro é a base somado com a altura multiplicado por dois ou (b+h) * 2\\n Caso você queira, use o calculador abaixo\")\r\n base = int(input(\"Qual a base deste retângulo? \"))\r\n altura = int(input(\"Qual a altura deste retângulo? \"))\r\n\r\n perimetro = (base+altura) * 2\r\n\r\n print (\"\\tSeu retangulo tem perimetro de %d cm²\\n\" % (perimetro))\r\n \r\n elif Escolha == \"Sair\":\r\n print (\"Saindo do programa\")\r\n break\r\n \r\n","repo_name":"JonanthaW/Cade-o-Cafe","sub_path":"Scripts pequenos feitos em PYTHON/Calcular Base ou Perimetro de um retângulo.py","file_name":"Calcular Base ou Perimetro de um retângulo.py","file_ext":"py","file_size_in_byte":1206,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28140670582","text":"import tensorflow as tf\n\n\nclass Loss():\n '''\n provides the physics loss function class\n '''\n def __init__(self, pinn, config):\n\n # save neural network (weights are updated during training)\n self.pinn = pinn\n\n def F_residuals(self, t_col):\n '''\n determines physics loss residuals of the differential equation\n at the collocation points\n '''\n # the tf-GradientTape function is used to retreive network derivatives\n with tf.GradientTape() as tape:\n tape.watch(t_col)\n y = self.pinn(t_col)\n y_t = tape.gradient(y, t_col)\n\n return y_t - y + y**3\n","repo_name":"frohrhofer/PINN_TF2","sub_path":"pinn/loss_functions.py","file_name":"loss_functions.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"30417868487","text":"from leezy import Solution, solution\nfrom leezy.assists import TreeContext\n\nclass Q100(Solution):\n @solution\n def isSameTree(self, p, q):\n # 20ms\n if p is None and q is None:\n return True\n if p and q and p.val == q.val:\n return self.isSameTree(p.left, q.left) and self.isSameTree(p.right, q.right)\n return False\n\n @solution\n def is_same_iter(self, p, q):\n # 16ms 81.04%\n stack_p = [p]\n stack_q = [q]\n while stack_p and stack_q:\n p = stack_p.pop()\n q = stack_q.pop()\n if p is None and q is None:\n continue\n if p and q and p.val == q.val:\n stack_p.append(p.left)\n stack_p.append(p.right)\n stack_q.append(q.left)\n stack_q.append(q.right)\n else:\n return False\n return True\n\ndef main():\n q = Q100()\n q.set_context(TreeContext)\n q.add_args([1, 2, 3], [1, 2, 3])\n q.add_args([1, 2, 1], [1, 1, 2])\n q.add_args([1, 2], [1, None, 2])\n q.run()\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"aptend/leetcode-rua","sub_path":"Python/100 - Same Tree/100_same-tree.py","file_name":"100_same-tree.py","file_ext":"py","file_size_in_byte":1133,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"11461572988","text":"# -*- coding: utf-8 -*-\n# 从上到下按层打印二叉树,同一层的节点按从左到右的顺序打印,每一层打印到一行。\n#\n# 例如:\n# 给定二叉树:[3,9,20,null,null,15,7],\n#\n# 3\n# / \\\n# 9 20\n# / \\\n# 15 7\n# 返回其层次遍历结果:\n#\n# [\n# [3],\n# [9,20],\n# [15,7]\n# ]\n#\n# 来源:力扣(LeetCode)\n# 链接:https://leetcode.cn/problems/cong-shang-dao-xia-da-yin-er-cha-shu-ii-lcof\n\nfrom typing import List\n\n\nclass TreeNode:\n def __init__(self, x):\n self.val = x\n self.left = None\n self.right = None\n\n\nclass Solution:\n def levelOrder(self, root: TreeNode) -> List[List[int]]:\n if not root:\n return []\n from collections import deque\n ans = []\n queue = deque([root])\n while queue:\n tmp = []\n length = len(queue)\n for _ in range(length):\n node = queue.popleft()\n tmp.append(node.val)\n if node.left:\n queue.append(node.left)\n if node.right:\n queue.append(node.right)\n ans.append(tmp)\n return ans\n","repo_name":"lisj1211/Algorithm","sub_path":"剑指offer/32 - II. 从上到下打印二叉树 II.py","file_name":"32 - II. 从上到下打印二叉树 II.py","file_ext":"py","file_size_in_byte":1174,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18190244747","text":"import sys\r\nfrom itertools import product\r\n\r\ninput = sys.stdin.readline\r\n\r\ndef dfs(graph: dict, start: tuple[int, int]) -> set[int]:\r\n explored, stack = {start, }, [start]\r\n\r\n while stack:\r\n v = stack.pop()\r\n explored.add(v)\r\n\r\n for adj in graph.get(v):\r\n if adj not in explored:\r\n stack.append(adj)\r\n \r\n return explored\r\n\r\ndelta = list(product([0, 1, -1], repeat=2))\r\n\r\nwhile True:\r\n width, height = map(int, input().split())\r\n \r\n if width == height == 0:\r\n break\r\n\r\n matrix = [list(map(int, input().split())) for _ in range(height)]\r\n d = {(w, h): set() for w in range(width) for h in range(height) if matrix[h][w]}\r\n\r\n for l in d:\r\n for dx, dy in delta:\r\n (x, y) = l\r\n x += dx\r\n y += dy\r\n if x < 0 or y < 0 or x >= width or y >= height:\r\n continue\r\n \r\n if matrix[y][x]:\r\n d.get(l).add((x, y))\r\n\r\n cnt = 0\r\n s = set()\r\n for l in d:\r\n if l not in s:\r\n s = s.union(dfs(d, l))\r\n cnt += 1\r\n\r\n print(cnt)\r\n","repo_name":"thebjko/algorithm","sub_path":"백준/Silver/4963. 섬의 개수/섬의 개수.py","file_name":"섬의 개수.py","file_ext":"py","file_size_in_byte":1131,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"516226683","text":"from __future__ import annotations\nfrom typing import Optional, Callable, Union\n\nimport discord\nfrom discord import ui\n\nfrom .item import Item\nfrom .custom import CustomSelect\n\n\nclass Select(Item):\n def __init__(\n self,\n placeholder: Optional[str] = None,\n min_values: int = 1,\n max_values: int = 1,\n options: Optional[list] = None,\n disabled: bool = False,\n custom_id: Optional[str] = None,\n ) -> None:\n self._placeholder: Optional[str] = placeholder\n self._min_values: int = min_values\n self._max_values: int = max_values\n self._options: list = [] if options is None else options\n self._disabled: bool = disabled\n self._row: Optional[int] = None\n self._custom_id: Optional[str] = custom_id\n\n self.func: Optional[Callable] = None\n self.check_func: Optional[Callable[[discord.Interaction], bool]] = None\n\n def placeholder(self, placeholder: str) -> 'Select':\n self._placeholder = placeholder\n return self\n\n def min_values(self, min_values: int) -> 'Select':\n self._min_values = min_values\n return self\n\n def max_values(self, max_values: int) -> 'Select':\n self._max_values = max_values\n return self\n\n def options(self, options: list[SelectOption]) -> 'Select':\n self._options = [op.to_discord_select_option() for op in options]\n return self\n \n def disabled(self, disabled: bool = False) -> 'Select':\n self._disabled = disabled\n return self\n\n def row(self, row: int) -> 'Select':\n self._row = row\n return self\n\n def on_select(self, func: Callable) -> 'Select':\n self.func = func\n\n return self\n\n def custom_id(self, custom_id: str) -> 'Select':\n self._custom_id = custom_id\n return self\n\n def check(self, func: Callable[[discord.Interaction], bool]) -> 'Select':\n self.check_func = func\n return self\n\n def to_discord_item(self, row: Optional[int]) -> ui.Item:\n return CustomSelect(\n custom_id=self._custom_id,\n placeholder=self._placeholder,\n min_values=self._min_values,\n max_values=self._max_values,\n options=self._options,\n disabled=self._disabled,\n row=row,\n callback=self.func,\n check_func=self.check_func,\n )\n\n\nclass SelectOption:\n def __init__(\n self,\n label: str,\n value: Optional[str] = None,\n description: Optional[str] = None,\n emoji: Optional[Union[str, discord.PartialEmoji]] = None,\n default: bool = False\n ) -> None:\n self._label: str = label\n self._value: Optional[str] = label if value is None else value\n self._description: Optional[str] = description\n\n if isinstance(emoji, str):\n emoji = discord.PartialEmoji.from_str(emoji)\n\n self._emoji: Optional[str] = emoji\n self._default: bool = default\n\n def label(self, label: str) -> 'SelectOption':\n \"\"\"Set the label of the option.\n The label of the option. This is displayed to users.\n Can only be up to 25 characters.\n This function returns the class instance to allow for fluent-style\n chaining.\n Parameters\n -----------\n label: :class:`str`\n The label to set.\n \"\"\"\n self._label = label\n return self\n\n def value(self, value: str) -> 'SelectOption':\n \"\"\"Set the value of the option.\n The value of the option. This is not displayed to users.\n If not provided when constructed then it defaults to the\n label. Can only be up to 100 characters.\n This function returns the class instance to allow for fluent-style\n chaining.\n Parameters\n -----------\n value: :class:`str`\n The value to set.\n \"\"\"\n self._value = value\n return self\n\n def description(self, description: str) -> 'SelectOption':\n \"\"\"Set the description of the option.\n An additional description of the option.\n Can only be up to 50 characters.\n This function returns the class instance to allow for fluent-style\n chaining.\n Parameters\n -----------\n description: :class:`str`\n The description to set.\n \"\"\"\n self._description = description\n return self\n\n def emoji(self, emoji: Optional[Union[str, discord.PartialEmoji]]) -> 'SelectOption':\n \"\"\"Set the emoji of the option.\n This function returns the class instance to allow for fluent-style\n chaining.\n Parameters\n -----------\n emoji: Optional[Union[:class:`str`, :class:`discord.PartialEmoji`]]\n The description to set.\n \"\"\"\n if isinstance(emoji, str):\n emoji = discord.PartialEmoji.from_str(emoji)\n\n self._emoji = emoji\n return self\n\n def default(self, default: bool) -> 'SelectOption':\n \"\"\"Set the default of the option.\n Whether this option is selected by default.\n This function returns the class instance to allow for fluent-style\n chaining.\n Parameters\n -----------\n default: :class:`bool`\n The default value to set.\n \"\"\"\n self._default = default\n return self\n\n def to_discord_select_option(self) -> discord.SelectOption:\n return discord.SelectOption(\n label=self._label,\n value=self._value,\n description=self._description,\n emoji=self._emoji,\n default=self._default\n )\n","repo_name":"sizumita/discord-ext-ui","sub_path":"discord/ext/ui/select.py","file_name":"select.py","file_ext":"py","file_size_in_byte":5719,"program_lang":"python","lang":"en","doc_type":"code","stars":28,"dataset":"github-code","pt":"79"} +{"seq_id":"20195144880","text":"import cv2\nimport numpy as np\nimport os\nfrom matplotlib import pyplot as plt\n\n\n#Instansiasi file cascadenya\nstop_cascade = cv2.CascadeClassifier('cascade_stop_sign.xml')\n\ncap = cv2.VideoCapture(0)\ntes_uji = 1\n#proses segmentasi dan deteksi objek\nwhile (True):\n ret, frame = cap.read()\n gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n stop = stop_cascade.detectMultiScale(gray, 50, 50)\n\n #Membuat gambar grayscale\n #abu_abu = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n # Untuk ngebuat object kotak saat mendeteksi\n for (x,y,w,h) in stop:\n #untuk Membuat objek kotak\n cv2.rectangle(frame,(x,y),(x+w,y+h),(255,255,0),2)\n #Untuk ngebuat teks atau labelling pada object yang terdeteksi\n font = cv2.FONT_HERSHEY_SIMPLEX\n cv2.putText(frame,'STOP',(x-w,y-h), font, 0.5, (11,255,255), 2, cv2.LINE_AA)\n #Untuk ngecek bisa atau tidak perulangannya\n print(\"terdeteksi ke-\"+str(tes_uji))\n tes_uji+=1\n\n cv2.imshow('HASIL CASCADE',frame)\n key = cv2.waitKey(1)\n if key == 27: #tombol keluar\n break\n elif key == ord('x'):\n print (\"Salaah\")\n\ncap.release()\ncv2.destroyAllWindows()\n","repo_name":"otakbeku/PAPCD","sub_path":"cascade_tes_stop.py","file_name":"cascade_tes_stop.py","file_ext":"py","file_size_in_byte":1165,"program_lang":"python","lang":"id","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"19572850945","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport random\n\nfrom crossover import crossover\nfrom individual import create_random_inidividual\nfrom mutate import mutate\nfrom population import plot_population\nfrom selection import select_tournament\n\n\nif __name__ == '__main__':\n # PARAMETERS\n POPULATION_SIZE = 10\n CROSSOVER_PROBABILITY = 0.8\n MUTATION_PROBABILITY = 0.1\n MAX_GENERATIONS = 10\n\n random.seed(29)\n\n # Initial random population\n population = [create_random_inidividual() for _ in range(POPULATION_SIZE)]\n\n for generation_number in range(POPULATION_SIZE):\n\n # SELECTION OPERATION\n selected = select_tournament(population, 3)\n\n # CROSSOVER\n crossed_offspring = []\n for ind1, ind2 in zip(selected[::2], selected[1::2]):\n if random.random() < CROSSOVER_PROBABILITY:\n # Applying crossover to pair of individuals\n children = crossover(ind1, ind2)\n crossed_offspring.append(children[0])\n crossed_offspring.append(children[1])\n else:\n # Passing individuals further without crossover\n crossed_offspring.append(ind1)\n crossed_offspring.append(ind2)\n\n # MUTATION\n mutated = []\n for ind in crossed_offspring:\n if random.random() < MUTATION_PROBABILITY:\n # Applying mutation to an individual\n mutated.append(mutate(ind))\n else:\n # Passing individual further without mutation\n mutated.append(ind)\n\n # Next generation\n population = mutated\n\n plot_population(population)\n\n","repo_name":"RioKKH/LearningGeneticAlgorithmsWithPython","sub_path":"ch2_GAFlow/genetic_algorithm_flow.py","file_name":"genetic_algorithm_flow.py","file_ext":"py","file_size_in_byte":1683,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"29115031625","text":"# This module contains all of the necessary functions for optimizing value-at-risk.\n# Last Updated: 27/08\nimport monteCarloVaR\nimport historicalVaR\nimport parametricVaR\nimport readData\nimport pandas as pd\n\ndef multiVaRMinimum(portfolios, weightsVector, varTechniqueVector, confidenceintervalVector, TVector, repVector):\n \"\"\"This function calculates the minimum portfolio of a set of portfolios, each with their own VaR calculation\n technique. The lookback period is determined by the data itself. The index of each input corresponds to each\n portfolio. Historical Simulations do not require T (period of validity), therefore a 0 is put for their index in\n TVector. If the portfolio is not to be analysed using Monte Carlo repVector[portfolio] = 0 \"\"\"\n vaRList = []\n for index in range(len(portfolios)):\n if varTechniqueVector[index] == \"Parametric\":\n VaR = parametricVaR.paraPortfolioVaR(portfolios[index], weightsVector[index],\n confidenceintervalVector[index], TVector[index] )\n elif varTechniqueVector[index] == \"Historical\":\n VaR = historicalVaR.historicalPortfolioVaR(portfolios[index], confidenceintervalVector[index],\n weightsVector[index])\n else: # varTechniqueVector[index] == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(portfolios[index], [0, TVector[index]], weightsVector[index],\n repVector[index])\n VaR = historicalVaR.historicalSingleVaR(returnsList, confidenceintervalVector[index])\n\n vaRList.append(VaR)\n\n minIndex = vaRList.index(min(vaRList))\n print(\"Min portfolio is Portfolio: \")\n print(minIndex + 1)\n print(\"Min VaR value is: \")\n print(vaRList[minIndex])\n\n\ndef exchangeOptimize(portfolio, assetCodes, weightsVector, selectedAsset, varTechnique, T, reps):\n \"\"\"For a specified derivative, this function searches the WikiFutures list to see if it is listed on another exchange\n which is already being traded on in the portfolio. If it is, then the CVaR of the original portfolio is calculated\n and added, and the CVaR of the new portfolio (where the asset is switched) is calculated and added.\n If the new CVaR is lower, the function returns the exchange with which the new asset is allocated to.\"\"\"\n df = readData.readWikiFutures(\"wikifutures.csv\")\n\n # Get a list of exchanges in portfolio. Construct a dictionary of assets and weights for each exchange\n exchangeList = []\n exchangeAssetDict = {}\n exchangeWeightDict = {}\n exchangeIndex = {}\n for index in range(len(assetCodes)):\n exchange = df.loc[df['Quandl Code'] == assetCodes[index], 'Exchange'].item()\n if exchange not in exchangeList:\n exchangeList.append(exchange)\n exchangeAssetDict[exchange] = [assetCodes[index]]\n exchangeWeightDict[exchange] = [weightsVector[index]]\n exchangeIndex[exchange] = [index]\n else:\n exchangeAssetDict[exchange].append(assetCodes[index])\n exchangeWeightDict[exchange].append(weightsVector[index])\n exchangeIndex[exchange].append(index)\n\n #Calculate the CVaR by summing the CVaR of each exchanges' portfolio\n cVaRList = []\n exchangesPortfoliosDict = {}\n for exchange in exchangeList:\n exchangePortfolio1 = portfolio.drop(portfolio.columns[exchangeIndex[exchange]],axis=1)\n exchangePortfolio2 = exchangePortfolio1.tail(df.loc[df['Quandl Code'] == exchangeAssetDict[exchange][0],\n 'Lookback'].item()) # Cut down to lookback\n exchangesPortfoliosDict[exchange] = exchangePortfolio2.copy(deep = True)\n\n exchangeWeights = exchangeWeightDict[exchange]\n exchangeConfidence = df.loc[df['Quandl Code'] == exchangeAssetDict[exchange][0], 'Confidence'].item()\n\n if varTechnique == \"Parametric\":\n CVaR = parametricVaR.paraPortfolioCVaR(exchangePortfolio2, exchangeWeights, exchangeConfidence, T)\n elif varTechnique == \"Historical\":\n CVaR = historicalVaR.historicalPortfolioVaR(exchangePortfolio2, exchangeConfidence, exchangeWeights)\n else: # varTechnique == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(exchangePortfolio2, reps, [0,T], exchangeWeights)\n CVaR = historicalVaR.historicalSingleVaR(returnsList, exchangeConfidence)\n\n cVaRList.append(CVaR)\n\n cVaRTotal = 0\n for CVaR in cVaRList:\n cVaRTotal += CVaR\n\n print(\"Current CVaR with asset \" + selectedAsset + \" is \" + str(cVaRTotal) + \".\")\n\n newCVaRTotal = cVaRTotal\n newCode = 0\n\n #Get name of asset and calculate the CVaR\n selectedName = df.loc[df['Quandl Code'] == selectedAsset, 'Name'].item()\n for index, row in df.iterrows():\n exchange = row[\"Exchange\"]\n selectedAssetsExchange = df.loc[df['Quandl Code'] == selectedAsset, 'Exchange'].item()\n\n if row[\"Name\"] == selectedName and row[\"Exchange\"] in exchangeList and row[\"Exchange\"] is not selectedAssetsExchange:\n subPortfolioCopy1 = (exchangesPortfoliosDict[exchange]).copy(deep=True)\n subWeightVectorCopy = list(exchangeWeightDict[exchange])\n rowCode = row[\"Quandl Code\"]\n if rowCode not in assetCodes:\n singlePortfoliodirty = readData.readQuandl([str(rowCode) + \".1\"], 1000, True)\n singlePortfolio2 = readData.cleanData(singlePortfoliodirty)\n singlePortfolio = singlePortfolio2.tail(df.loc[df['Quandl Code'] == exchangeAssetDict[exchange][0], 'Lookback'].item())\n\n # MAKE TEMP OF selectAsset's portfolio\n tempSelectedAssetPortfolio = exchangesPortfoliosDict[selectedAssetsExchange].copy(deep = True)\n tempSelectedAssetsWeightsList = list(exchangeWeightDict[selectedAssetsExchange])\n selectedAssetsIndex = 0\n newAssetsIndex = 0\n\n # Get the index of the selected assets exchange, to be used to remove the weight and portfolio values\n for index in range(len(exchangeAssetDict[selectedAssetsExchange])):\n if exchangeAssetDict[selectedAssetsExchange][index] == selectedAsset:\n selectedAssetsIndex = index\n\n # Get the index of the new assets exchange\n for index in range(len(exchangeAssetDict[exchange])):\n if exchangeAssetDict[exchange][index] == rowCode:\n newAssetsIndex = index\n\n tempSelectedAssetPortfolio.drop(tempSelectedAssetPortfolio.columns[selectedAssetsIndex], axis=1, inplace=True)\n tempWeight = tempSelectedAssetsWeightsList.pop(selectedAssetsIndex)\n\n # Add the new asset to the corresponding exchanges dataframe and the weight vector\n subWeightVectorCopy.append(tempWeight)\n subPortfolioCopy = (subPortfolioCopy1.join(singlePortfolio)).copy(deep = True)\n\n # recalculate the CVaR intermediateCVaRTotal\n cVaRRemainder = cVaRTotal - cVaRList[selectedAssetsIndex] - cVaRList[newAssetsIndex]\n\n newExchangeConfidence = df.loc[df['Quandl Code'] == exchangeAssetDict[exchange][0], 'Confidence'].item()\n oldexchangeConfidence = df.loc[df['Quandl Code'] == exchangeAssetDict[selectedAssetsExchange][0], 'Confidence'].item()\n\n # Get CVaR of selected Asset\n\n if varTechnique == \"Parametric\":\n CVaRSelectedAsset = parametricVaR.paraPortfolioCVaR(tempSelectedAssetPortfolio, tempSelectedAssetsWeightsList, oldexchangeConfidence, T)\n elif varTechnique == \"Historical\":\n CVaRSelectedAsset = historicalVaR.historicalPortfolioVaR(tempSelectedAssetPortfolio, oldexchangeConfidence, tempSelectedAssetsWeightsList)\n else: # varTechnique == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(tempSelectedAssetPortfolio, reps, [0, T], tempSelectedAssetsWeightsList)\n CVaRSelectedAsset = historicalVaR.historicalSingleVaR(returnsList, oldexchangeConfidence)\n\n # Get CVaR of new asset\n if varTechnique == \"Parametric\":\n CVaRNewAsset = parametricVaR.paraPortfolioCVaR(subPortfolioCopy,\n subWeightVectorCopy,\n newExchangeConfidence, T)\n elif varTechnique == \"Historical\":\n CVaRNewAsset = historicalVaR.historicalPortfolioVaR(subPortfolioCopy,\n newExchangeConfidence,\n subWeightVectorCopy)\n else: # varTechnique == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(subPortfolioCopy, reps, [0, T],\n subWeightVectorCopy)\n CVaRNewAsset = historicalVaR.historicalSingleVaR(returnsList, newExchangeConfidence)\n\n tempCVaRTotal = CVaRNewAsset + CVaRSelectedAsset + cVaRRemainder\n\n if tempCVaRTotal < cVaRTotal:\n newCVaRTotal = tempCVaRTotal\n newCode = row[\"Quandl Code\"]\n\n if newCode == 0:\n print(\"No matching assets were found in other exchanges.\")\n elif newCode == selectedAsset:\n print(\"CVaR cannot be minimized by switching this asset.\")\n else:\n print(\"CVaR can be minimized by switching with asset \" + newCode + \". New CVaR is \" + str(newCVaRTotal) +\".\")\n\n\ndef optimizeSimilarAssets(portfolio, assetCodes, weightsVector, selectedAsset,\n alternativeAssets, varTechnique, T, reps, confidence):\n \"\"\" Out of a list of alternative assets for a selected asset, this function is able to suggest an alternative asset for\n which the overall portfolio VaR is minimized. For maximum realism, the portfolio must be at a single exchange,\n as all VaR is calculated used the same method, lookback, and confidence.\"\"\"\n\n # Calculate the current VaR\n\n if varTechnique == \"Parametric\":\n currentVaR = parametricVaR.paraPortfolioVaR(portfolio, weightsVector, confidence, T)\n elif varTechnique == \"Historical\":\n currentVaR = historicalVaR.historicalPortfolioVaR(portfolio, confidence, weightsVector)\n else: # varTechniqueVector[index] == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(portfolio, [0, T], weightsVector, reps)\n currentVaR = historicalVaR.historicalSingleVaR(returnsList, confidence)\n\n # Identify index of selected asset in weightsVector\n\n selectedAssetIndex = 0\n for index in range(len(assetCodes)):\n if assetCodes[index] == selectedAsset:\n selectedAssetIndex = index\n\n # Pop the selected asset and weight. Append selected weight to the end of the weight vector\n\n selectedWeight = weightsVector.pop(selectedAssetIndex)\n weightsVector.append(selectedWeight)\n portfolio.drop(portfolio.columns[selectedAssetIndex], axis=1, inplace=True)\n\n # For each alternativeAsset, read the asset data and append selected asset portfolio and weight to the end.\n # Find the length of the original dataset\n\n lengthOriginal = len(portfolio)\n minVaR = currentVaR\n print(\"Original VaR is \" + str(minVaR))\n minAsset = selectedAsset\n\n for alternativeAsset in alternativeAssets:\n singlePortfoliodirty = readData.readQuandl([alternativeAsset], lengthOriginal * 2, True)\n singlePortfolio2 = readData.cleanData(singlePortfoliodirty)\n singlePortfolio = (singlePortfolio2.tail(lengthOriginal)).copy(deep = True)\n\n subPortfolioDirtyCopy = (portfolio.join(singlePortfolio)).copy(deep=True)\n subPortfolioCopy = readData.cleanData(subPortfolioDirtyCopy).copy(deep = True)\n\n print(subPortfolioCopy)\n print(weightsVector)\n\n # Recalculate the VaR, if lower then replace the min VaR with new VaR and new asset\n\n if varTechnique == \"Parametric\":\n newVaR = parametricVaR.paraPortfolioVaR(subPortfolioCopy, weightsVector, confidence, T)\n elif varTechnique == \"Historical\":\n newVaR = historicalVaR.historicalPortfolioVaR(subPortfolioCopy, confidence, weightsVector)\n else: # varTechniqueVector[index] == \"Monte Carlo\"\n returnsList = monteCarloVaR.portfolioMonteCarlo2(subPortfolioCopy, [0, T], weightsVector, reps)\n newVaR = historicalVaR.historicalSingleVaR(returnsList, confidence)\n\n if newVaR < minVaR:\n minVaR = newVaR\n minAsset = alternativeAsset\n\n # Print out results\n if minAsset == selectedAsset:\n print(\"VaR cannot be minimized by switching to an alternative asset.\")\n else:\n print(\"VaR can be minimized by switching with asset \" + minAsset + \". New VaR is \" + str(minVaR) +\".\")\n\n\n","repo_name":"turgon91/Computing-Value-at-risk-for-Collateral-Optimization","sub_path":"optimization.py","file_name":"optimization.py","file_ext":"py","file_size_in_byte":13140,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"23646570973","text":"from unittest import skipIf\n\nfrom decouple import config\nfrom django.test import TestCase, override_settings\nfrom google.api_core.client_options import ClientOptions\nfrom google.auth.credentials import AnonymousCredentials\nfrom google.cloud.exceptions import NotFound\nfrom google.cloud.storage import Blob, Bucket, Client\nfrom physionet.gcs import GCSObject, GCSObjectException\nfrom physionet.settings.base import StorageTypes\n\nTEST_GCS_INTEGRATION = config('TEST_GCS_INTEGRATION', default=True, cast=bool)\nGCS_HOST = config('GCS_HOST', default=None)\n\n\n@skipIf(\n (GCS_HOST is None or not TEST_GCS_INTEGRATION),\n 'Test GCS-backend integration only on dockerized CI/CD pipeline.',\n)\n@override_settings(\n STORAGE_TYPE=StorageTypes.GCP,\n DEFAULT_FILE_STORAGE='physionet.storage.MediaStorage',\n STATICFILES_STORAGE='physionet.storage.StaticStorage',\n GCP_STORAGE_BUCKET_NAME='physionet-media',\n GCP_STATIC_BUCKET_NAME='physionet-static',\n GS_PROJECT_ID='test_project_id',\n GCP_BUCKET_LOCATION='us-west1',\n)\nclass TestGCSObject(TestCase):\n @classmethod\n def setUpTestData(cls):\n cls.gcs_server_endpoint = f'http://{config(\"GCS_HOST\", default=\"gcs\")}:4443'\n cls.bucket_name = 'test'\n cls.path = 'physionet/users/admin/profile.jpg'\n\n def tearDown(self):\n try:\n self._clear_gcs_bucket(self.bucket_name)\n except NotFound:\n pass\n\n def _clear_gcs_bucket(self, name):\n self._get_gcs_client().get_bucket(name).delete(force=True)\n\n def _get_gcs_client(self):\n return Client(\n project=\"test_project_id\",\n credentials=AnonymousCredentials(),\n client_options=ClientOptions(api_endpoint=self.gcs_server_endpoint),\n )\n\n def _monkeypatch_gcsobject(self, gcs_object):\n gcs_object._storage._client = self._get_gcs_client()\n return gcs_object\n\n @override_settings(STORAGE_TYPE=StorageTypes.LOCAL)\n def test_init_raises_exception_when_storage_types_is_local(self):\n self.assertRaises(GCSObjectException, GCSObject, self.path)\n\n @override_settings(STORAGE_TYPE=StorageTypes.GCP)\n def test_init_when_storage_type_is_gcp(self):\n gcs_object = self._monkeypatch_gcsobject(GCSObject(self.path))\n\n self.assertEqual(gcs_object.bucket.name, 'physionet')\n self.assertEqual(gcs_object._object_name, 'users/admin/profile.jpg')\n\n def test_repr(self):\n gcs_object = self._monkeypatch_gcsobject(GCSObject(self.path))\n\n self.assertEqual(\n repr(gcs_object),\n 'GCSObject(Bucket=physionet, Object=\"users/admin/profile.jpg\")',\n )\n\n def test_bucket_property_return_bucket_proper_object(self):\n gcs_object = self._monkeypatch_gcsobject(GCSObject(self.path))\n\n self.assertIsInstance(gcs_object.bucket, Bucket)\n self.assertEqual(gcs_object.bucket.name, 'physionet')\n\n def test_blob_property_return_proper_object(self):\n gcs_object = self._monkeypatch_gcsobject(GCSObject(self.path))\n\n self.assertIsInstance(gcs_object.blob, Blob)\n self.assertEqual(gcs_object.blob.name, 'users/admin/profile.jpg')\n\n def test_mkdir_makes_directories(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir1/dir2/'))\n gcs_object.client.create_bucket('test')\n\n # WHEN\n gcs_object.mkdir()\n\n # THEN\n self.assertTrue(gcs_object.bucket.get_blob('dir1/dir2/'))\n\n def test_mkdir_doesnt_work_when_object_name_is_taken(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir1/dir2/'))\n gcs_object.client.create_bucket('test')\n gcs_object.mkdir()\n\n # WHEN + THEN\n self.assertRaises(GCSObjectException, gcs_object.mkdir)\n\n def test_size_when_object_is_file(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir1/notes.txt'))\n gcs_object.client.create_bucket('test')\n gcs_object.upload_from_string('content')\n\n # WHEN + THEN\n self.assertEqual(gcs_object.size(), len('content'))\n\n def test_size_when_object_is_directory(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir1/'))\n gcs_object_1 = self._monkeypatch_gcsobject(GCSObject('test/dir1/notes1.txt'))\n gcs_object_2 = self._monkeypatch_gcsobject(GCSObject('test/dir1/notes2.txt'))\n\n # create a bucket\n gcs_object.client.create_bucket('test')\n\n # put files into a bucket\n gcs_object_1.upload_from_string('content')\n gcs_object_2.upload_from_string('content')\n\n # WHEN + THEN\n self.assertEqual(gcs_object.size(), len('content') * 2)\n\n def test_rm_deletes_all_files_in_directory_when_object_is_directory(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir1/'))\n gcs_object_1 = self._monkeypatch_gcsobject(GCSObject('test/dir1/notes1.txt'))\n gcs_object_2 = self._monkeypatch_gcsobject(GCSObject('test/dir1/notes2.txt'))\n\n # create a bucket\n gcs_object.client.create_bucket('test')\n\n # put files into a bucket\n gcs_object_1.upload_from_string('content')\n gcs_object_2.upload_from_string('content')\n\n # WHEN\n gcs_object.rm()\n\n # THEN\n self.assertEqual(gcs_object.size(), 0)\n\n def test_rm_removes_file_when_object_is_file(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir/file.jpg'))\n gcs_object.client.create_bucket('test')\n gcs_object.upload_from_string('content')\n\n # WHEN\n gcs_object.rm()\n\n # THEN\n dir_ = self._monkeypatch_gcsobject(self._monkeypatch_gcsobject(GCSObject('test/dir/')))\n self.assertEqual(dir_.size(), 0)\n\n def test_cp_copies_file_to_directory(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir/file.jpg'))\n gcs_object_1 = self._monkeypatch_gcsobject(GCSObject('test/dir/'))\n\n # create a bucket\n gcs_object.client.create_bucket('test')\n\n # put a file into a bucket\n gcs_object.upload_from_string('content')\n\n # WHEN\n gcs_object_1.cp(self._monkeypatch_gcsobject(GCSObject('test/dir_copied/')))\n\n # THEN\n self.assertEqual(gcs_object_1.size(), len('content'))\n self.assertEqual(gcs_object.size(), len('content'))\n\n def test_mv_moves_file_when_object_is_file(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/dir/file.jpg'))\n gcs_object_1 = self._monkeypatch_gcsobject(GCSObject('test/dir/'))\n gcs_object_2 = self._monkeypatch_gcsobject(GCSObject('test/dir_copied/'))\n\n # create a bucket\n gcs_object.client.create_bucket('test')\n\n # put a file into a bucket\n gcs_object.upload_from_string('content')\n\n # WHEN\n gcs_object_1.mv(self._monkeypatch_gcsobject(GCSObject('test/dir_copied/')))\n\n # THEN\n self.assertEqual(gcs_object_2.size(), len('content'))\n self.assertEqual(gcs_object.exists(), False)\n\n def test_rename_file(self):\n # GIVEN\n gcs_object = self._monkeypatch_gcsobject(GCSObject('test/file.jpg'))\n gcs_object.client.create_bucket('test')\n gcs_object.upload_from_string('content')\n\n gcs_object_renamed = self._monkeypatch_gcsobject(GCSObject('test/renamed.jpg'))\n\n # WHEN\n gcs_object.rename(gcs_object_renamed)\n\n # THEN\n self.assertFalse(gcs_object.exists())\n self.assertTrue(gcs_object_renamed.exists())\n self.assertEqual(gcs_object_renamed.size(), len('content'))\n","repo_name":"MIT-LCP/physionet-build","sub_path":"physionet-django/physionet/test_gcs.py","file_name":"test_gcs.py","file_ext":"py","file_size_in_byte":7701,"program_lang":"python","lang":"en","doc_type":"code","stars":50,"dataset":"github-code","pt":"79"} +{"seq_id":"11914090076","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"Tests for `harvdev_utils` package.\"\"\"\nimport pytest\nimport os\nimport time\nimport psycopg2\nimport subprocess\n\n# Minimal prototype test for new proforma parsing software.\n# SQL Alchemy imports\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom sqlalchemy.orm.exc import NoResultFound\n\nfrom harvdev_utils.chado_functions import (\n get_feature_by_uniquename, feature_name_lookup,\n feature_symbol_lookup, feature_synonym_lookup,\n get_organism, CodingError, DataError,\n synonym_name_details, get_cvterm\n)\n\nfrom harvdev_utils.chado_functions.get_or_create import get_or_create\nfrom harvdev_utils.production import (\n Feature, Synonym, FeatureSynonym\n)\nconn2 = False\nsession = None\n\n\ndef setup_module(module):\n \"\"\"Start db.\"\"\"\n global conn2, session\n conn2 = module.conn = startup_db()\n session = get_session()\n\n\ndef teardown_module(module):\n \"\"\"Teardown any state that was previously setup with a setup_module method.\"\"\"\n stop_db(module.conn)\n\n\ndef stop_db(conn):\n \"\"\"Shut down the test database instance.\"\"\"\n subprocess.getoutput('docker rm $(docker stop $(docker ps -a -q --filter ancestor=flybase/proformatestdb --format=\"{{.ID}}\"))')\n if conn:\n conn.close()\n\n\ndef startup_db():\n \"\"\"Start up the test database instance.\"\"\"\n # This first os.system command is a bit hacky, but it'll prevent errors where the database is already running\n # and we attempt to execute 'docker run' again. TODO Revisit this and handle \"docker is already running\" issues better.\n os.system('docker rm $(docker stop $(docker ps -a -q --filter ancestor=flybase/proformatestdb --format=\"{{.ID}}\"))')\n os.system('docker run -p 127.0.0.1:5436:5432 --name harv_util_proformatestdb flybase/proformatestdb &')\n\n conn = None\n trys = 0\n while (not conn and trys < 10):\n trys += 1\n time.sleep(5)\n try:\n conn = psycopg2.connect(host=\"127.0.0.1\", port=\"5436\", database=\"fb_test\", user='tester', password=\"tester\")\n except psycopg2.Error:\n pass\n\n if (not conn):\n print(\"ERROR: Could not connect to test db\")\n stop_db(None)\n exit(-1)\n\n return conn\n\n\ndef get_session():\n \"\"\"Create session.\"\"\"\n # Create our SQL Alchemy engine from our environmental variables.\n engine_var = 'postgresql://' + 'tester' + \":\" + 'tester' + '@' + '127.0.0.1' + ':' + '5436' '/' + 'fb_test'\n\n engine = create_engine(engine_var)\n\n Session = sessionmaker(bind=engine)\n session = Session()\n return session\n\n\nclass TestSomething:\n \"\"\"Test smethings.\"\"\"\n\n def test_unique_lookup_good(self):\n \"\"\"Test uniquename good lookups.\"\"\"\n # standard lookup\n feature = get_feature_by_uniquename(session, \"FBgn0000001\")\n assert feature.name == 'symbol-1'\n\n # lookup with obsolete set.\n feature = get_feature_by_uniquename(session, \"FBgn0000002\", obsolete='f')\n assert feature.name == 'symbol-2'\n\n # check obsolete = 'e' => either.\n feature = get_feature_by_uniquename(session, \"FBgn0000049\", obsolete='e')\n assert feature.name == 'symbol-49'\n\n def test_unique_lookup_bad(self):\n \"\"\"Test uniquename bad lookups.\"\"\"\n # lookup up non allowed obsolete value\n with pytest.raises(CodingError):\n get_feature_by_uniquename(session, \"FBgn0004600\", obsolete='madeup')\n\n # obsolete wrong.\n with pytest.raises(NoResultFound):\n get_feature_by_uniquename(session, \"FBgn0004500\", obsolete='t')\n\n # lookup up non existant uniquename\n with pytest.raises(NoResultFound):\n get_feature_by_uniquename(session, \"Made_Ip\", obsolete='t')\n\n def test_name_lookup_good(self):\n \"\"\"Test name good lookups.\"\"\"\n # check basic lookup\n feature = feature_name_lookup(session, 'symbol-1')\n assert feature.uniquename.startswith('FBgn')\n\n # check type_name and obsolete work\n feature = feature_name_lookup(session, 'symbol-2', type_name='gene', obsolete='f')\n assert feature.name == 'symbol-2'\n\n # check obsolete 'e'\n feature = feature_name_lookup(session, 'symbol-3', obsolete='e')\n assert feature.name == 'symbol-3'\n\n # check organism start in name\n organism = get_organism(session, short='Hsap')\n feature = feature_name_lookup(session, 'Hsap\\\\symbol-1', organism_id=organism.organism_id)\n assert feature.uniquename.startswith('FBgn')\n\n # check converted greek chars are done correctly.\n feature = feature_name_lookup(session, 'gene_with_alpha1')\n feature.name == 'gene_with_alpha1'\n\n def test_name_lookup_bad(self):\n \"\"\"Test name bad lookups.\"\"\"\n # check for bad made up name\n feature = feature_name_lookup(session, 'madeup_name')\n assert feature is None\n\n # Check that obsolete is used and checked\n feature = feature_name_lookup(session, 'symbol-4', type_name='gene', obsolete='t')\n assert feature is None\n\n # check type_name is checked\n feature = feature_name_lookup(session, 'symbol-5', type_name='allele', obsolete='f')\n assert feature is None\n\n def test_symbol_lookup_good(self):\n \"\"\"Test symbol good lookups.\n\n Note symbols are unique, synonyms are NOT, that is the difference.\n \"\"\"\n # Check standard use.\n feature = feature_symbol_lookup(session, 'gene', 'symbol-10')\n assert feature.name == 'symbol-10'\n\n # Add obsolete as 'f'\n feature = feature_symbol_lookup(session, 'gene', 'symbol-11', obsolete='f')\n assert feature.name == 'symbol-11'\n\n # Add obsolete as 'e'\n feature = feature_symbol_lookup(session, 'gene', 'symbol-12', obsolete='e')\n assert feature.name == 'symbol-12'\n\n # lookup diff species.\n organism = get_organism(session, short='Hsap')\n feature = feature_symbol_lookup(session, 'gene', 'Hsap\\\\symbol-2', organism_id=organism.organism_id)\n assert feature.name == 'Hsap\\\\symbol-2'\n assert feature.organism_id == organism.organism_id\n\n # check greek chars are done correctly.\n feature = feature_symbol_lookup(session, 'gene', 'gene_with_&agr;1')\n assert feature.name == 'gene_with_alpha1'\n\n # DO NOT specify a type, look for all\n feature = feature_symbol_lookup(session, None, 'symbol-21')\n assert feature.name == 'symbol-21'\n\n # Test bracket names with no conversion (convert = false)\n feature = feature_symbol_lookup(session, None, 'C9orf72:n.intron14[30GGGGCC]', convert=False)\n assert feature.name == 'C9orf72:n.intron14[30GGGGCC]'\n\n def test_symbol_lookup_bad(self):\n \"\"\"Test symbol bad lookups.\"\"\"\n # Lookup non existent symbol\n with pytest.raises(NoResultFound):\n feature_symbol_lookup(session, 'gene', 'made up')\n\n # Lookup with incorrect obsolete value\n with pytest.raises(NoResultFound):\n feature_symbol_lookup(session, 'gene', 'symbol-30', obsolete='t')\n\n # default org is Dmel so choose Hsap synonym as test\n organism = get_organism(session, short='Hsap')\n with pytest.raises(NoResultFound):\n feature_symbol_lookup(session, 'gene', 'symbol-30', organism_id=organism.organism_id)\n\n # wrong type.\n with pytest.raises(NoResultFound):\n feature_symbol_lookup(session, 'allele', 'genechar-&agr;-[0002]')\n\n # DIVS do not convert but do not specift here so it should fail\n with pytest.raises(NoResultFound):\n feature_symbol_lookup(session, None, 'C9orf72:n.intron14[30GGGGCC]')\n\n def test_synonym_lookup_good(self):\n \"\"\"Test synonym good lookups.\"\"\"\n # return array if cheque_unique not used\n features = feature_synonym_lookup(session, 'gene', 'symbol-1')\n for feature in features:\n assert feature.name in ['symbol-1']\n\n # lookup diff species.\n organism = get_organism(session, short='Hsap')\n feature = feature_synonym_lookup(session, 'gene', 'Hsap\\\\symbol-2', organism_id=organism.organism_id)\n assert feature.name in ['Hsap\\\\symbol-2']\n\n # return feature if check_unique if set\n feature = feature_synonym_lookup(session, 'gene', 'symbol-2', check_unique=True)\n assert feature.name in ['symbol-2']\n\n # return feature if check_unique if set and obsolete is 'f'\n feature = feature_synonym_lookup(session, 'gene', 'symbol-3', check_unique=True, obsolete='f')\n assert feature.name in ['symbol-3']\n\n # return feature if check_unique if set and obsolete is 'e' => either\n feature = feature_synonym_lookup(session, 'gene', 'symbol-4', check_unique=True, obsolete='e')\n assert feature.name in ['symbol-4']\n\n def test_synonym_lookup_bad(self):\n \"\"\"Test synonym bad lookups.\"\"\"\n # gene does not exist\n features = feature_synonym_lookup(session, 'gene', 'made_up')\n assert not features\n\n # gene does not exist make it unique\n with pytest.raises(DataError):\n feature_synonym_lookup(session, 'gene', 'made_up', check_unique=True)\n\n # return feature if check_unique if set and obsolete is 'e' => either\n with pytest.raises(DataError):\n feature_synonym_lookup(session, 'gene', 'symbol-6', check_unique=True, obsolete='t')\n\n # return feature obsolete is 't' NO checkunique\n features = feature_synonym_lookup(session, 'gene', 'symbol-6', obsolete='t')\n assert not features\n\n def test_create_and_lookup(self):\n \"\"\"Create and then lookup.\"\"\"\n name = \"TP{1}Tao[1]\"\n feat_type = get_cvterm(session, 'SO', 'transposable_element_insertion_site')\n organism, plain_name, sgml = synonym_name_details(session, name)\n new_feat, _ = get_or_create(session, Feature, name=plain_name,\n type_id=feat_type.cvterm_id, uniquename=\"FBti:temp_1\",\n organism_id=organism.organism_id)\n\n feature = feature_name_lookup(session, name, type_name='transposable_element_insertion_site')\n assert feature.name == 'TP{1}Tao[1]'\n\n syn_type = get_cvterm(session, 'synonym type', 'symbol')\n\n synonym, _ = get_or_create(session, Synonym, type_id=syn_type.cvterm_id, name=plain_name, synonym_sgml=sgml)\n\n fs, _ = get_or_create(session, FeatureSynonym, feature_id=new_feat.feature_id, synonym_id=synonym.synonym_id,\n pub_id=1)\n fs.is_current = True\n fs.is_internal = False\n feature = feature_symbol_lookup(session, 'transposable_element_insertion_site', name, convert=True)\n assert feature.name == 'TP{1}Tao[1]'\n","repo_name":"FlyBase/harvdev-utils","sub_path":"chado_tests/test_lookups.py","file_name":"test_lookups.py","file_ext":"py","file_size_in_byte":10834,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"25048388984","text":"#/usr/bin/python\n#-*- coding:utf-8 -*-\n\nimport pandas as pd\nimport numpy as np\nfrom pandas import DataFrame,Series\nimport matplotlib.pyplot as plt\n\nfig = plt.figure()\nfig.set(alpha = 0.2)\ndata_train = pd.read_csv(\"/Users/liucong/Desktop/node/Data-anlysis/Titanic/data/train.csv\")\n\nplt.subplot2grid((3,3),(0,0))\ndata_train.Survived.value_counts().plot(kind='bar')\nplt.title(u\"获救情况\")\nplt.ylabel(u\"人数\")\n\nplt.subplot2grid((3,3),(0,1))\ndata_train.Pclass.value_counts().plot(kind = 'bar')\nplt.title(u\"乘客等级分布\")\nplt.ylabel(u\"人数\")\n\nplt.subplot2grid((3,3),(0,2))\ndata_train.Sex.value_counts().plot(kind = 'bar')\nplt.title(u\"乘客性别统计\")\nplt.ylabel(u\"人数\")\n\nplt.subplot2grid((3,3),(1,0))\nplt.scatter( data_train.Age, data_train.Survived)\nplt.xlabel(u\"年龄\")\nplt.title(u\"根据性���看获救情况\")\nplt.grid(b=True,which='major', axis='y')\n\nplt.subplot2grid((3,3),(1,1),colspan=2)\ndata_train.Embarked.value_counts().plot(kind = 'bar')\nplt.title(u\"各登船口统计情况\")\nplt.ylabel(u\"人数\")\n\n\nplt.subplot2grid((3,3),(2,0),colspan=3)\ndata_train.Age[data_train.Embarked == 'S'].plot(kind = 'kde')\ndata_train.Age[data_train.Embarked == 'C'].plot(kind = 'kde')\ndata_train.Age[data_train.Embarked == 'Q'].plot(kind = 'kde')\nplt.xlabel(u\"年龄\")\nplt.ylabel(u\"密度\")\nplt.title(u\"各港口乘客的年龄分布情况\")\nplt.legend((u'S港口',u'C港口',u'Q港口'),loc = 'best')\n\nplt.show()\n\n\n\n","repo_name":"congball/Titanic","sub_path":"attributes_analysis.py","file_name":"attributes_analysis.py","file_ext":"py","file_size_in_byte":1425,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"36924168872","text":"from rest_framework.decorators import api_view, permission_classes\nfrom rest_framework.permissions import IsAuthenticated\nfrom rest_framework.exceptions import ParseError\nfrom api.models import Account, Customer, Service\nfrom django.contrib.auth.hashers import make_password\nfrom rest_framework.response import Response\nfrom rest_framework import status, generics, viewsets\nfrom .serializers import ListCustomerSerializer, EditCustomerSerializer, CustomerSerializer, DeleteCustomerSerializer, CreateCustomerSerializer\nfrom common.token_authentication import TokenAuthentication\nfrom common.token_permission import TokenPermission\nfrom drf_yasg.utils import swagger_auto_schema\nfrom drf_yasg import openapi\nfrom api.permission import CTAccessPermission, SPAAccessPermission\nfrom backend.pagination import CustomPagination\n# from django_filters import rest_framework as filters\nfrom rest_framework import filters\nfrom django_filters.rest_framework import DjangoFilterBackend\nimport requests\nimport json\n__all__ = ['ListCustomerView', 'DetailCustomerView', 'CreateCustomerView']\n\n\n# class ListCustomerFilter(filters.FilterSet):\n# customer_name = filters.CharFilter(lookup_expr='icontains')\n\n# class Meta:\n# model = Customer\n# fields = ('customer_name', 'requester_id')\n\n\nclass ListCustomerView(generics.ListAPIView):\n permission_classes = (SPAAccessPermission,)\n authentication_classes = (TokenAuthentication,)\n # pagination_class = CustomPagination\n # queryset = Customer.objects.all()\n # filterset_class = ListCustomerFilter\n # filter_fields = ('customer_name', 'requester_id')\n filter_backends = [filters.SearchFilter,\n filters.OrderingFilter, DjangoFilterBackend]\n search_fields = ['customer_name'] # field like\n ordering_fields = '__all__' # order by\n ordering = '-created_at'\n filterset_fields = ['customer_name']\n\n @swagger_auto_schema(\n responses={\n 200: ListCustomerSerializer(many=True),\n },\n tags=['admin/customer'],\n operation_id='admin'\n )\n def get(self, request):\n # customers = Customer.objects.filter(on_active=True)\n customers = Customer.objects.all()\n queryset = self.filter_queryset(customers)\n page = self.paginate_queryset(queryset)\n serializer = ListCustomerSerializer(page, many=True)\n result = self.get_paginated_response(serializer.data)\n data = result.data\n return Response(data, status=status.HTTP_200_OK)\n # customers = Customer.objects.all().filter(on_active=True)\n # serializer = ListCustomerSerializer(customers, many=True)\n # return Response(serializer.data, status=status.HTTP_200_OK)\n\n\nclass CreateCustomerView(generics.CreateAPIView):\n\n @swagger_auto_schema(\n\n request_body=CreateCustomerSerializer,\n responses={\n 200: CreateCustomerSerializer,\n },\n tags=['admin/customer'],\n operation_id='Add New Customer'\n )\n def post(self, request, *args, **kwargs):\n return self.create(request, *args, **kwargs)\n\n def create(self, request, *args, **kwargs):\n serializer = CreateCustomerSerializer(data=request.data)\n customer_name = request.data['customer_name']\n email = request.data['main_email']\n if serializer.is_valid():\n url = \"https://usdctechnology.zendesk.com/api/v2/users.json\"\n payload = \"{\\\"user\\\": {\\\"name\\\": \\\"\" + \\\n customer_name+\"\\\", \\\"email\\\": \\\"\"+email+\"\\\"}}\"\n headers = {\n 'Content-Type': 'application/json',\n 'Authorization': 'Bearer eea15b7525bdb980227bb314d2d11b819e4681040c4b80cb33b36892de65b72b',\n }\n response = requests.request(\n \"POST\", url, headers=headers, data=payload.encode('utf-8'))\n try:\n value = response.json()['user']['id']\n self.perform_create(serializer)\n new_customer = Customer.objects.get(\n main_email=email, on_active=True)\n new_customer.set_requester_id(value)\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n except:\n payload: dict = {\n 'status': 'error',\n 'message': 'Email is already being used by another user in zendesk system',\n 'message_code': '400',\n }\n return Response(payload, status=status.HTTP_400_BAD_REQUEST)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\nclass DetailCustomerView(generics.ListAPIView, generics.UpdateAPIView, generics.DestroyAPIView):\n\n permission_classes = (SPAAccessPermission,)\n authentication_classes = (TokenAuthentication,)\n\n @swagger_auto_schema(\n responses={\n 200: CustomerSerializer,\n },\n tags=['admin/customer'],\n operation_id='admin'\n )\n def get(self, request, *args, **kwargs):\n customer_id = kwargs.get('customer_id')\n try:\n customers = Customer.objects.get(id=customer_id)\n except:\n payload: dict = {\n 'message': 'Customer not found',\n 'message_code': '400',\n 'error': {}\n }\n return Response(payload, status=status.HTTP_400_BAD_REQUEST)\n serializer = CustomerSerializer(customers)\n return Response(serializer.data, status=status.HTTP_200_OK)\n\n @swagger_auto_schema(\n request_body=EditCustomerSerializer,\n responses={\n 200: EditCustomerSerializer,\n },\n tags=['admin/customer'],\n operation_id='Edit Customer'\n )\n def put(self, request, *args, **kwargs):\n customer_id = kwargs.get('customer_id')\n try:\n customers = Customer.objects.get(id=customer_id)\n accounts = Account.objects.all().filter(customer_of_id=customer_id)\n except:\n payload: dict = {\n 'message': 'Customer not found',\n 'message_code': '400',\n 'error': {}\n }\n return Response(payload, status=status.HTTP_400_BAD_REQUEST)\n serializer = EditCustomerSerializer(customers, data=request.data)\n if serializer.is_valid():\n if request.data['on_active'] != customers.on_active and request.data['on_active'] is True and accounts:\n for i in accounts:\n i.set_on_active(True)\n elif request.data['on_active'] != customers.on_active and request.data['on_active'] is False and accounts:\n for i in accounts:\n i.set_on_active(False)\n serializer.save()\n response = {\n 'status': 'success',\n 'message_code': status.HTTP_201_CREATED,\n 'message': 'Update success',\n }\n return Response(response, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n @swagger_auto_schema(\n request_body=EditCustomerSerializer,\n responses={\n 200: EditCustomerSerializer,\n },\n tags=['admin/customer'],\n operation_id='Patch Customer'\n )\n def patch(self, request, *args, **kwargs):\n customer_id = kwargs.get('customer_id')\n try:\n customers = Customer.objects.get(id=customer_id)\n except:\n payload: dict = {\n 'message': 'Customer not found',\n 'message_code': '400',\n 'error': {}\n }\n return Response(payload, status=status.HTTP_400_BAD_REQUEST)\n partial = kwargs.pop('partial', False)\n serializer = EditCustomerSerializer(\n customers, data=request.data, partial=partial)\n if serializer.is_valid():\n serializer.save()\n response = {\n 'status': 'success',\n 'message_code': status.HTTP_201_CREATED,\n 'message': 'Update success',\n }\n return Response(response, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n @swagger_auto_schema(\n tags=['admin/customer'],\n operation_id='Delete Customer'\n )\n def delete(self, request, *args, **kwargs):\n customer_id = kwargs.get('customer_id')\n try:\n customer = Customer.objects.get(id=customer_id)\n account = Account.objects.all().filter(customer_of=customer_id, on_active=True)\n service = Service.objects.all().filter(customer_id=customer_id, on_active=True)\n except:\n payload: dict = {\n 'message': 'Customer not found',\n 'message_code': '400',\n 'error': {}\n }\n return Response(payload, status=status.HTTP_400_BAD_REQUEST)\n customer.set_on_active(False)\n for x in range(len(account)):\n account[x].set_on_active(False)\n for x in range(len(service)):\n service[x].set_on_active(False)\n\n payload: dict = {\n 'status': 'success',\n 'message': 'Delete success',\n 'message_code': '200',\n }\n return Response(payload, status=status.HTTP_200_OK)\n","repo_name":"songbao2903/backend_api-v1.0.0","sub_path":"api/supper_admin/customer/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":9381,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"14751262763","text":"#task1:\nlist = [1700, 2020, 1800, 1600, 1900]\ncounter = 0\nfor item in range(len(list)):\n if counter == 3:\n print(list)\n break\n selection = item\n for j in range(item + 1, len(list)):\n if list[selection] > list[j]:\n selection = j\n holder = list[item]\n list[item] = list[selection]\n list[selection] = holder\n counter += 1\nprint(\"Answer to Task 1 is b\")\n\n#task2:\nmy_tuple = (3,6,12,24,48)\n\ndef multiply_it(tuple):\n sum = 1\n try:\n for item in tuple:\n sum = sum * item\n return sum\n except Exception as e:\n print(f\"A {e} error has ocurred\")\n\nprint(f\"The answer to task 2 is {multiply_it(my_tuple)}\")\n\n#task3:\nclass Stack:\n def __init__(self):\n self.items = []\n\n def is_empty(self):\n return self.items == []\n\n def push(self, item):\n self.items.append(item)\n\n def pop(self):\n return self.items.pop()\n\n def peek(self):\n return self.items[len(self.items)-1]\n\n def size(self):\n return len(self.items)\n\ndef reverse_list(list):\n result = []\n stack = Stack()\n for item in list:\n stack.push(item)\n while stack.is_empty() is False:\n result.append(stack.pop())\n print(f\"The list reversed becomes {result}\")\n\nmylist = [1, 2, 3, 4, 5]\nreverse_list(mylist)","repo_name":"Bjolav/INFO135_Assignments","sub_path":"zit010_assignment2.py","file_name":"zit010_assignment2.py","file_ext":"py","file_size_in_byte":1322,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"74591816254","text":"#시각 - 3이 포함된 시각의 모든 경우의 수\nn= int(input()) #n=5\n\nh=0\nm=0 #03분 13,23,, 30~39분,43,53, #15번\ns=0 #3초 30~39초 =>11번\n\ncnt=0\nfor h in range(n+1):\n for m in range(60):\n for s in range(60):\n if str(h).find('3') >=0 or str(m).find('3') >=0 or str(s).find('3') >=0 :\n cnt+=1\nprint(cnt)\n","repo_name":"MDandersonm/this-is-coding-test","sub_path":"04_implementation/4-2.py","file_name":"4-2.py","file_ext":"py","file_size_in_byte":352,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"42803613102","text":"from unittest import mock\n\nfrom django.contrib.auth import get_user_model\nfrom django.test import TestCase\nfrom django.urls import reverse\n\nfrom apps.accounts.models import ReferralRecord\nfrom apps.users import views\n\nUserModel = get_user_model()\n\n\nclass ViewsTestCase(TestCase):\n\n @classmethod\n def setUpTestData(cls):\n cls.user_signup_data = {\n 'wallet': 'http://test.com',\n 'username': 'admin',\n 'email': '79889501003@yandex.ru',\n 'phone': '+79998887766',\n 'first_name': 'test_first_name',\n 'last_name': 'test_last_name',\n 'password1': 'pass0911',\n 'password2': 'pass0911',\n } # user data for signup api\n\n @mock.patch.object(views.SignUpView, 'send_confirmation_email')\n @mock.patch.object(views.SignUpView, 'process_referral')\n def test_signup_successful(self, mocked_method_referral,\n mocked_method_email):\n \"\"\"\n Test signup view handler.\n\n Steps:\n - send user data\n - create user and account\n - send confirmation link to email (only check call)\n - process referral cookies (only check call)\n\n :param mocked_method_referral: views.SignUpView.process_referral mock\n :param mocked_method_email: views.SignUpView.process_referral mock\n :return: None\n \"\"\"\n\n url = reverse('users:signup')\n data = self.user_signup_data.copy()\n response = self.client.post(url, data=data)\n\n self.assertEqual(response.status_code, 302)\n self.assertEqual(response.get('Location'), reverse('core:index'))\n\n self.assertTrue(UserModel.objects.filter(\n account__username=data['username']).exists())\n user = UserModel.objects.get(account__username=data['username'])\n self.assertFalse(user.is_active)\n mocked_method_referral.assert_called_once_with(user)\n mocked_method_email.assert_called_once_with(user)\n\n def test_signup_without_phone(self):\n \"\"\"\n Test signup without Phone Number.\n User shouldn't be created if his phone number is not defined.\n \"\"\"\n url = reverse('users:signup')\n data = dict(self.user_signup_data, phone='')\n response = self.client.post(url, data=data)\n\n self.assertEqual(response.status_code, 200)\n\n with self.assertRaises(UserModel.DoesNotExist):\n UserModel.objects.get(account__username=data['username'])\n\n def test_signup_without_email(self):\n \"\"\"\n Test signup without Email.\n User shouldn't be created if his email is not defined.\n \"\"\"\n url = reverse('users:signup')\n data = dict(self.user_signup_data, email='test')\n response = self.client.post(url, data=data)\n\n self.assertEqual(response.status_code, 200)\n\n with self.assertRaises(UserModel.DoesNotExist):\n UserModel.objects.get(account__username=data['username'])\n\n @mock.patch.object(views.SignUpView, 'send_confirmation_email')\n def test_signup_referrals(self, mocked_confirmation_email):\n \"\"\"\n Test referral records on signup.\n If user has visited referral url from another user,\n then after his registration we should create referral record\n with user relations.\n\n :param mocked_confirmation_email:\n views.SignUpView.mocked_confirmation_email mock\n \"\"\"\n main_user = UserModel.objects.create(email='test@example.com')\n\n # go to referral url\n url = reverse('users:ref', kwargs={'pk': main_user.pk})\n self.client.get(url)\n\n with self.assertRaises(ReferralRecord.DoesNotExist):\n ReferralRecord.objects.get(user=main_user)\n\n # sign up with cookies after referral\n url = reverse('users:signup')\n data = self.user_signup_data.copy()\n self.client.post(url, data=data)\n\n referral = ReferralRecord.objects.get(user=main_user)\n self.assertEqual(referral.status, ReferralRecord.STATUS_CHOICES.WAIT)\n self.assertEqual(referral.referral.account.username, data['username'])\n","repo_name":"clairkkluhnkkk5720/IMBTNG-WEBAPP","sub_path":"src/apps/users/tests/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4158,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"73918225536","text":"import types\nimport sys\n\n\n_stdout_save = sys.stdout\n_stdout_write_save = sys.stdout.write\n\n\ndef _stdout_write_dummy(*args, **kwargs):\n pass\n\n\nclass UserTextIO(types.ModuleType):\n\n _override_quiet = None\n _override_yesno = None\n _yesno_answer_map = {'y': True, 'Y': True, 'yes': True, 'Yes': True,\n 'n': False, 'N': False, 'no': False, 'No':False}\n\n def is_quiet_overridden(self) -> bool:\n \"\"\"returns whether quiet is overridden or not\"\"\"\n return self._override_quiet is not None\n\n @property\n def override_quiet(self):\n \"\"\"Return the actual value of _override_quiet\"\"\"\n return self._override_quiet\n\n @override_quiet.setter\n def override_quiet(self, quiet):\n \"\"\"override quiet\n\n overrides whether functions should be quiet or not, by setting the\n write function of sys.stdout to _stdout_write_dummy\n\n arguments:\n quiet -- what to override quiet with\n \"\"\"\n self._override_quiet = bool(quiet)\n if self._override_quiet:\n sys.stdout.write = _stdout_write_dummy\n else:\n sys.stdout.write = _stdout_write_save\n\n @override_quiet.getter\n def override_quiet(self) -> bool:\n \"\"\"returns the vaule of _override_quiet\n\n exceptions:\n AttributeError -- if override_quiet is not set\n \"\"\"\n if self.is_quiet_overridden():\n return self._override_yesno\n else:\n message = 'Attribute override_quiet is no set'\n raise AttributeError(message)\n\n @override_quiet.deleter\n def override_quiet(self):\n \"\"\"deletes the current value of _override_quiet\n\n sets the write function of sys.stdout back to the actual function,\n which was saved in _stdout_write_save\n \"\"\"\n self._override_quiet = None\n sys.stdout = _stdout_save\n\n def is_yesno_overridden(self) -> bool:\n \"\"\"returns whether yesno is overridden or not\"\"\"\n return self._override_yesno is not None\n\n @property\n def override_yesno(self):\n \"\"\"Return the actual value of _override_yesno\"\"\"\n return self._override_yesno\n\n @override_yesno.setter\n def override_yesno(self, yesno):\n \"\"\"override yesno\n\n overrides the yes/no questions which would normally be asked\n\n arguments:\n yesno -- what to override yesno with\n \"\"\"\n self._override_yesno = bool(yesno)\n\n @override_yesno.getter\n def override_yesno(self) -> bool:\n \"\"\"returns the vaule of _override_yesno\n\n exceptions:\n AttributeError -- if override_yesno is not set\n \"\"\"\n if self.is_yesno_overridden():\n return self._override_yesno\n else:\n message = 'Attribute override_yesno is no set'\n raise AttributeError(message)\n\n @override_yesno.deleter\n def override_yesno(self):\n \"\"\"deletes the current value of _override_yesno\"\"\"\n self._override_yesno = None\n\n def _yesno_choicefield(self, default=False):\n \"\"\"returns a yesno choice field\n\n returns a yesno choice field of the form [y/N], depending on default\n\n arguments:\n default -- which is to be the default, normally No\n \"\"\"\n field_format = '[{}/{}]'\n if default:\n return field_format.format('Y', 'n')\n else:\n return field_format.format('y', 'N')\n\n def ask_yesno_question(self, questiontext, default=False) -> bool:\n \"\"\"Asks the user a question which to answer with yes or no\n\n Warning: this functions behavior depends on the value of override_yesno\n property of this module, if override_yesno is set, no question will be\n asked, instead the value of override_yesno will be returned\n\n arguments:\n questiontext -- the text of the question\n default -- the answer which is to be the default, normally No ('N')\n \"\"\"\n if self.is_yesno_overridden():\n return self.override_yesno\n else:\n text = '{} {}: '.format(questiontext, self._yesno_choicefield(default))\n answer = input(text)\n return self._yesno_answer_map.get(answer, default)\n\n# change out the module for the class, so properties can be used\nif __name__ != '__main__':\n sys.modules[__name__] = UserTextIO(__name__)\n","repo_name":"ZofiaZementa/zztools","sub_path":"zztools/utilities/usertextio.py","file_name":"usertextio.py","file_ext":"py","file_size_in_byte":4360,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18286950410","text":"from django.urls import path\r\nfrom .views import *\r\n\r\nurlpatterns = [\r\n path('books', view_all_books, name='all_books'),\r\n path('books/', view_single_book, name='single_book'),\r\n path('books/year/', view_book_year, name='book_year'),\r\n path('', index, name=\"index\"),\r\n path('books/category/', view_book_category, name='book_category'),\r\n path('books/category//year/', view_book_catyear, name='book_catyear'),\r\n\r\n\r\n]\r\n","repo_name":"capellja/CA298","sub_path":"Lab_2/Library/templates/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":511,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"35796972718","text":"#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n\n'''\nFor pypi\n'''\n\nfrom setuptools import setup, find_packages\n\ndesc = ('Static site generator (SSG) for GIS data publishment as light WebGIS application.')\nsetup(\n name='gislite',\n version='0.0.4',\n keywords=('WebGIS', 'Static site generator'),\n description=desc,\n long_description=''.join(open('ReadMe.rst').readlines()),\n license='MIT License',\n\n url='http://gislite.osgeo.cn',\n author='bukun',\n author_email='bukun@osgeo.cn',\n\n packages=find_packages(\n # include=('torcms',),\n exclude=(\"tester\", \"wcs_imgmap\", 'static', 'docs', 'dist-site', 'deprecated')),\n include_package_data=True,\n\n platforms='any',\n zip_safe=True,\n install_requires=['markdown', 'mapproxy', 'pyyaml', 'openpyxl'],\n\n classifiers=[\n # How mature is this project? Common values are\n # 3 - Alpha\n # 4 - Beta\n # 5 - Production/Stable\n 'Development Status :: 3 - Alpha',\n\n # Indicate who your project is intended for\n 'Intended Audience :: Developers',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n\n # Pick your license as you wish (should match \"license\" above)\n 'License :: OSI Approved :: MIT License',\n\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.7'\n ],\n)\n","repo_name":"bukun/GISLite","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1382,"program_lang":"python","lang":"en","doc_type":"code","stars":45,"dataset":"github-code","pt":"79"} +{"seq_id":"41003456958","text":"from typing import Dict\nfrom typing import Set\n\nfrom tracetools_read import get_field\n\nfrom . import EventHandler\nfrom . import EventMetadata\nfrom . import HandlerMap\nfrom ..data_model.memory_usage import MemoryUsageDataModel\n\n\nclass MemoryUsageHandler(EventHandler):\n \"\"\"Generic handler for memory usage.\"\"\"\n\n def __init__(\n self,\n **kwargs,\n ) -> None:\n if type(self) is MemoryUsageHandler:\n raise RuntimeError('Do not instantiate directly!')\n super().__init__(\n data_model=MemoryUsageDataModel(),\n **kwargs,\n )\n\n @property\n def data(self) -> MemoryUsageDataModel:\n return super().data # type: ignore\n\n def _update(\n self,\n timestamp: int,\n tid: int,\n memory_difference: int,\n ) -> None:\n # Add to data model\n self.data.add_memory_difference(timestamp, tid, memory_difference)\n\n\nclass UserspaceMemoryUsageHandler(MemoryUsageHandler):\n \"\"\"\n Handler that extracts userspace memory usage data.\n\n It uses the following events:\n * lttng_ust_libc:malloc\n * lttng_ust_libc:calloc\n * lttng_ust_libc:realloc\n * lttng_ust_libc:free\n * lttng_ust_libc:memalign\n * lttng_ust_libc:posix_memalign\n\n The above events are generated when LD_PRELOAD-ing liblttng-ust-libc-wrapper.so, see:\n https://lttng.org/docs/v2.10/#doc-liblttng-ust-libc-pthread-wrapper\n\n Implementation inspired by Trace Compass' implementation:\n https://git.eclipse.org/c/tracecompass/org.eclipse.tracecompass.git/tree/lttng/org.eclipse.tracecompass.lttng2.ust.core/src/org/eclipse/tracecompass/internal/lttng2/ust/core/analysis/memory/UstMemoryStateProvider.java#n161\n \"\"\"\n\n def __init__(\n self,\n **kwargs,\n ) -> None:\n # Link event to handling method\n handler_map: HandlerMap = {\n 'lttng_ust_libc:malloc':\n self._handle_malloc,\n 'lttng_ust_libc:calloc':\n self._handle_calloc,\n 'lttng_ust_libc:realloc':\n self._handle_realloc,\n 'lttng_ust_libc:free':\n self._handle_free,\n 'lttng_ust_libc:memalign':\n self._handle_memalign,\n 'lttng_ust_libc:posix_memalign':\n self._handle_posix_memalign,\n }\n super().__init__(\n handler_map=handler_map,\n **kwargs,\n )\n\n # Temporary buffers\n # pointer -> current memory size\n # (used to know keep track of the memory size allocated at a given pointer)\n self._memory: Dict[int, int] = {}\n\n @staticmethod\n def required_events() -> Set[str]:\n return {\n 'lttng_ust_libc:malloc',\n 'lttng_ust_libc:free',\n }\n\n def _handle_malloc(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'ptr')\n if ptr != 0:\n size = get_field(event, 'size')\n self._handle(event, metadata, ptr, size)\n\n def _handle_calloc(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'ptr')\n if ptr != 0:\n nmemb = get_field(event, 'nmemb')\n size = get_field(event, 'size')\n self._handle(event, metadata, ptr, size * nmemb)\n\n def _handle_realloc(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'ptr')\n if ptr != 0:\n new_ptr = get_field(event, 'in_ptr')\n size = get_field(event, 'size')\n self._handle(event, metadata, ptr, 0)\n self._handle(event, metadata, new_ptr, size)\n\n def _handle_free(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'ptr')\n if ptr != 0:\n self._handle(event, metadata, ptr, 0)\n\n def _handle_memalign(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'ptr')\n if ptr != 0:\n size = get_field(event, 'size')\n self._handle(event, metadata, ptr, size)\n\n def _handle_posix_memalign(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n ptr = get_field(event, 'out_ptr')\n if ptr != 0:\n size = get_field(event, 'size')\n self._handle(event, metadata, ptr, size)\n\n def _handle(\n self,\n event: Dict,\n metadata: EventMetadata,\n ptr: int,\n size: int,\n ) -> None:\n timestamp = metadata.timestamp\n tid = metadata.tid\n\n memory_difference = size\n # Store the size allocated for the given pointer\n if memory_difference != 0:\n self._memory[ptr] = memory_difference\n else:\n # Othersize, if size is 0, it means it was deleted\n # Try to fetch the size stored previously\n allocated_memory = self._memory.get(ptr, None)\n if allocated_memory is not None:\n memory_difference = -allocated_memory\n\n self._update(timestamp, tid, memory_difference)\n\n\nclass KernelMemoryUsageHandler(MemoryUsageHandler):\n \"\"\"\n Handler that extracts userspace memory usage data.\n\n It uses the following events:\n * kmem_mm_page_alloc\n * kmem_mm_page_free\n\n Implementation inspired by Trace Compass' implementation:\n https://git.eclipse.org/c/tracecompass/org.eclipse.tracecompass.git/tree/analysis/org.eclipse.tracecompass.analysis.os.linux.core/src/org/eclipse/tracecompass/analysis/os/linux/core/kernelmemoryusage/KernelMemoryStateProvider.java#n84\n \"\"\"\n\n PAGE_SIZE = 4096\n\n def __init__(\n self,\n **kwargs,\n ) -> None:\n # Link event to handling method\n handler_map: HandlerMap = {\n 'kmem_mm_page_alloc':\n self._handle_malloc,\n 'kmem_mm_page_free':\n self._handle_free,\n }\n super().__init__(\n handler_map=handler_map,\n **kwargs,\n )\n\n @staticmethod\n def required_events() -> Set[str]:\n return {\n 'kmem_mm_page_alloc',\n 'kmem_mm_page_free',\n }\n\n def _handle_malloc(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n self._handle(event, metadata, self.PAGE_SIZE)\n\n def _handle_free(\n self, event: Dict, metadata: EventMetadata\n ) -> None:\n self._handle(event, metadata, -self.PAGE_SIZE)\n\n def _handle(\n self,\n event: Dict,\n metadata: EventMetadata,\n inc: int,\n ) -> None:\n order = get_field(event, 'order')\n inc <<= order\n\n timestamp = metadata.timestamp\n tid = metadata.tid\n\n self._update(timestamp, tid, inc)\n","repo_name":"azu-lab/ROS2-E2E-Evaluation","sub_path":"autoware_perf_galactic/tracetools_analysis-galactic_add_tp/tracetools_analysis/tracetools_analysis/processor/memory_usage.py","file_name":"memory_usage.py","file_ext":"py","file_size_in_byte":6789,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"33702869665","text":"import time\n\nimport numpy as np\nfrom numba import jit\n\nprint('CPU: AMD Ryzen 7 5700G')\nprint('CPU Clock: Up to 4.6GHz')\nprint('L2 Cache: 4MiB')\nprint('L3 Cache: 16MiB')\nprint('RAM: 16GB')\nprint('SO: Fedora Linux')\n\nprint('------------------------')\n\nprint('Experimento 01 - Multiplicação de Matrizes')\n\nthreads = [2, 4, 8]\n\na = np.random.rand(500, 500)\nb = np.random.rand(500, 500)\n\n# Começo\nif len(a[0]) != len(b):\n # verifica se o número de colunas de a é igual ao número de linhas de b\n raise ValueError('Error')\n\n# cria uma matriz de zeros com o número de linhas de a e o número de colunas de b\nc = np.zeros((len(a), len(b[0])))\n\n\ndef basic(a, b, c):\n for i in range(len(a)): # percorre as linhas da matriz a\n for j in range(len(b[0])): # percorre as colunas da matriz b\n for k in range(len(b)): # percorre as linhas da matriz b\n # soma o resultado da multiplicação de cada elemento de a com b\n c[i][j] += a[i][k] * b[k][j]\n return i, j, k\n\n\n@jit(nopython=True)\ndef numb(a, b, c):\n for i in range(len(a)):\n for j in range(len(b[0])):\n for k in range(len(b)):\n c[i][j] += a[i][k] * b[k][j]\n return i, j, k\n\n\nstart = time.time()\nbasic(a, b, c)\nend = time.time()\ntempo_sequencial = end - start\nprint('Tempo de execução sequencial: ', tempo_sequencial)\n\nfor i in threads:\n start = time.time()\n numb(a, b, c)\n end = time.time()\n print('Tempo de execução numba com {} threads: '.format(i), end - start)\n print('SpeedUp: ', tempo_sequencial / (end - start))\n\nprint('------------------------')\n\nprint('Experimento 02 - Produto Escalar')\n\nu = np.random.rand(100)\nv = np.random.rand(100)\n\n\ndef prod_escalar(u, v):\n return u * v\n\n\n@jit(nopython=True)\ndef prod_escalar_paralelo(u, v):\n return u * v\n\n\nstart = time.time()\nprod_escalar(u, v)\nend = time.time()\ntempo_sequencial = end - start\nprint('Tempo de execução sequencial: ', tempo_sequencial)\n\nfor i in threads:\n start = time.time()\n prod_escalar_paralelo(u, v)\n end = time.time()\n print('Tempo de execução numba com {} threads: '.format(i), end - start)\n print('SpeedUp: ', tempo_sequencial / (end - start))\n\nprint('------------------------')\n\nprint('Experimento 03 - Análise do Projeto Integrador')\n\n\"\"\" A única parte do projeto integrador que poderia ser otimizado com Numba\ne o paradigma de computação paralela, seria quando formos realizar o treinamento\nde nosso modelo de classificação. Entretanto, o desenvolvimento do projeto\nainda está em andamento, não sendo possível realizar a análise de tempo de\nexecução, já que o modelo ainda não está pronto.\n\nOutras partes do projeto não sofreriam ganhos significativos com a utilização,\ncomo a leitura dos dados, junto a biblioteca Pandas, já que o Numba não o reconhce,\ne também a parte de banco de dados, que pode gerar problemas ao tentar popular.\"\"\"\n\n\"\"\" Por exemplo, o módulo DataParallel do PyTorch permite o uso de paralelismo de dados\ne permite o treinamento simultâneo de um modelo em muitas GPUs. Por meio de bibliotecas\ncomo o PyTorch Lightning, que oferecem uma interface direta para treinamento distribuído\nem várias GPUs ou estações de trabalho, o paralelismo de tarefas pode ser usado.\n\nO PyTorch permite a compilação JIT (Just-In-Time) para otimização de código nativo com a \nbiblioteca TorchScript. Ao fazer isso, o código Python pode ser transformado em uma versão\notimizada para execução para uso em tempo de execução.\"\"\"\n","repo_name":"JonasBrother97/computacao_paralela_em_python","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3532,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"36103430970","text":"import sqlite3\n\nconn = sqlite3.connect('spider.sqlite')\ncur = conn.cursor()\n\n# Finding the ids that send page rank - we are only interested\n# in pages in the SCC that contain in and out links\ncur.execute('''SELECT DISTINCT from_id FROM Links''')\nfrom_ids = list()\nfor row in cur: \n from_ids.append(row[0])\n\n# Finding the ids that receive page rank\nto_ids = list()\nlinks = list()\ncur.execute('''SELECT DISTINCT from_id, to_id FROM Links''')\nfor row in cur:\n from_id = row[0]\n to_id = row[1]\n if from_id == to_id : continue\n if from_id not in from_ids : continue\n if to_id not in from_ids : continue\n links.append(row)\n if to_id not in to_ids : to_ids.append(to_id)\n\n# Retrieving latest page ranks for strongly connected components\nprev_ranks = dict()\nfor node in from_ids:\n cur.execute('''SELECT new_rank FROM Pages WHERE id = ?''', (node, ))\n row = cur.fetchone()\n prev_ranks[node] = row[0]\n\nsval = input('How many iterations:')\nmany = 1\nif ( len(sval) > 0 ) : many = int(sval)\n\n# Sanity check\nif len(prev_ranks) < 1 : \n print(\"Nothing to page rank. Check data.\")\n quit()\n\n# Running page rank in the memory so that it is faster\nfor i in range(many):\n # print prev_ranks.items()[:5]\n next_ranks = dict();\n total = 0.0\n for (node, old_rank) in list(prev_ranks.items()):\n total = total + old_rank\n next_ranks[node] = 0.0\n # printing the total\n\n # Finding the number of outbound links and sending the page rank down each of them\n for (node, old_rank) in list(prev_ranks.items()):\n # printing node, old_rank\n give_ids = list()\n for (from_id, to_id) in links:\n if from_id != node : continue\n # printing ' ',from_id,to_id\n\n if to_id not in to_ids: continue\n give_ids.append(to_id)\n if ( len(give_ids) < 1 ) : continue\n amount = old_rank / len(give_ids)\n # printing node, old_rank,amount, give_ids\n \n for id in give_ids:\n next_ranks[id] = next_ranks[id] + amount\n \n newtot = 0\n for (node, next_rank) in list(next_ranks.items()):\n newtot = newtot + next_rank\n evap = (total - newtot) / len(next_ranks)\n\n # printing newtot, evap\n for node in next_ranks:\n next_ranks[node] = next_ranks[node] + evap\n\n newtot = 0\n for (node, next_rank) in list(next_ranks.items()):\n newtot = newtot + next_rank\n\n # Computing the per-page avg change from old rank to new rank\n totdiff = 0\n for (node, old_rank) in list(prev_ranks.items()):\n new_rank = next_ranks[node]\n diff = abs(old_rank-new_rank)\n totdiff = totdiff + diff\n\n avediff = totdiff / len(prev_ranks)\n print(i+1, avediff)\n\n # rotating\n prev_ranks = next_ranks\n\n# Entering the final ranks back into the database\nprint(list(next_ranks.items())[:5])\ncur.execute('''UPDATE Pages SET old_rank=new_rank''')\nfor (id, new_rank) in list(next_ranks.items()) :\n cur.execute('''UPDATE Pages SET new_rank=? WHERE id=?''', (new_rank, id))\nconn.commit()\ncur.close()\n\n","repo_name":"UtkarshGaurav/PageRanker","sub_path":"sprank.py","file_name":"sprank.py","file_ext":"py","file_size_in_byte":3059,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"70169683456","text":"import sys\nfrom math import ceil\n\ndef solution():\n a, b, v = map(int, sys.stdin.readline().rstrip().split(' '))\n \n d = a - b\n n = ((v - a) / d) + 1\n answer = ceil(n)\n print(answer)\n\nif __name__ == '__main__':\n solution()","repo_name":"kky5738/algorithms_study","sub_path":"baekjoon/level/basic math 1/2869.py","file_name":"2869.py","file_ext":"py","file_size_in_byte":241,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"36313533416","text":"# Source GeeksForGeeks\n\n# Function to sort our a_list in place\ndef insertion_sort(a_list): \n \n # Traverse through 1 to len(arr) \n for i in range(1, len(a_list)): \n \n key = a_list[i] \n \n # Move elements of arr[0..i-1], that are \n # greater than key, to one position ahead \n # of their current position \n j = i-1\n while j >=0 and key < a_list[j] : \n a_list[j+1] = a_list[j] \n j -= 1\n a_list[j+1] = key \n\n # now a_list is sorted! \n\nsample_input = [9,4,2,3,5,6,8,1,7]\ninsertion_sort(sample_input)\nprint(sample_input) # [1,2,3,4,5,6,7,8,9]","repo_name":"Coachable-Dev/coachable-application","sub_path":"2_elementary_sorts/src/insertion_sort.py","file_name":"insertion_sort.py","file_ext":"py","file_size_in_byte":625,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"33568544443","text":"from django.db import models\nfrom customer.models import Customer\nfrom inventory.models import Product\n\n# Create your models here.\nclass Order(models.Model):\n customer=models.ForeignKey(Customer,null=True,on_delete=models.CASCADE)\n product=models.ManyToManyField(Product,null=True)\n order_number = models.CharField(max_length=32, unique=True)\n order_date = models.DateTimeField(auto_now_add=True)\n order_status = models.CharField(max_length=16, choices=[\n ('pending', 'Pending'),\n ('shipped', 'Shipped'),\n ('delivered', 'Delivered'),\n ('canceled', 'CanceLled'),\n ])\n\n","repo_name":"Loice-KaniniMwau/GREENKIOSK_BACKEND","sub_path":"order/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":613,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"19871962803","text":"from django import forms\n\nfrom tables.models import statusTable\n\nclass changeStatusForm(forms.ModelForm):\n class Meta:\n model = statusTable\n fields = ('title',)\n\n title = forms.ModelChoiceField(\n queryset=statusTable.objects.filter(deleted=False),\n label=\"Выберите новый статус\",\n )","repo_name":"CucumentoJolaz/database_project","sub_path":"tables/forms/misc_forms.py","file_name":"misc_forms.py","file_ext":"py","file_size_in_byte":338,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"9754692578","text":"\n\"\"\"@brief Code for profilig of calculating standard deviation\n \n @author IVS-DREAM-TEAM\n \n @file stddev.py\n\"\"\"\n\n\n\nimport cProfile\nimport lib.profiler as pf\nimport sys\n\ndef help():\n \"\"\"\n Function for help if input is uncorrect\n \"\"\" \n print(\"Usage: python stddev.py 2 )\r\ndf = df.loc[df.valid_user].drop('valid_user', axis=1)\r\n\r\n#filter out user200 ~ ...\r\n# df['valid_user2'] = df.userId.map(df.groupby('userId')['userId'].value() > 200)\r\n# df = df.loc[df.valid_user2].drop('valid_user2', axis=1)\r\n\r\n# #randomly sample a couple of them\r\n#sampled_session_id = np.random.choice(df.userId.unique(), 8000, replace=False)\r\nsampled_session_id = np.random.choice(df.userId.unique(), 8000, replace=False)\r\ndf = df.loc[df.userId.isin(sampled_session_id)]\r\n\r\npositive_df = df\r\n\r\n#encder part\r\nuser_encoder = LabelEncoder()\r\nmovie_encoder = LabelEncoder()\r\ndf['userId'] = user_encoder.fit_transform(df.userId )\r\ndf['movieId']= movie_encoder.fit_transform(df.movieId)\r\npositive_df['userId'] = user_encoder.fit_transform(positive_df.userId )\r\npositive_df['movieId']= movie_encoder.fit_transform(positive_df.movieId)\r\n\r\n#rating > 3 filter\r\npositive_df = positive_df.loc[positive_df['rating']>3.0]\r\n\r\n#for answer label\r\n#rating > 3\r\n#dictionary\r\npositive_movie_dict = dict(positive_df.groupby('userId')['movieId'].apply(list))\r\n\r\n\r\n\r\n\r\n\r\nclass YooChooseDataset(InMemoryDataset):\r\n def __init__(self, root, transform=None, pre_transform=None):\r\n super(YooChooseDataset, self).__init__(root, transform, pre_transform)\r\n self.data, self.slices = torch.load(self.processed_paths[0])\r\n\r\n @property\r\n def raw_file_names(self):\r\n return []\r\n\r\n @property\r\n def processed_file_names(self):\r\n return ['../input/moive.dataset']\r\n\r\n def download(self):\r\n pass\r\n\r\n def process(self):\r\n\r\n data_list = []\r\n\r\n # process by session_id\r\n grouped = df.groupby('userId')\r\n for userId, group in tqdm(grouped):\r\n le = LabelEncoder()\r\n user_moive_id = le.fit_transform(group.movieId)\r\n group = group.reset_index(drop=True)\r\n group['user_moive_id'] = user_moive_id\r\n node_features = \\\r\n group.loc[group.userId == userId, ['user_moive_id','userId','movieId']].sort_values(\r\n 'user_moive_id')[['userId','movieId']].drop_duplicates().values\r\n\r\n node_features = torch.LongTensor(node_features).unsqueeze(1)\r\n target_nodes = group.user_moive_id.values[1:]\r\n source_nodes = group.user_moive_id.values[:-1]\r\n edge_index = torch.tensor([source_nodes, target_nodes], dtype=torch.long)\r\n\r\n x = node_features\r\n\r\n\r\n if userId in positive_movie_dict:\r\n positive_indices = le.transform(positive_movie_dict[userId])\r\n label = np.zeros(len(node_features))\r\n label[positive_indices] = 1\r\n else:\r\n label = [0] * len(node_features)\r\n\r\n\r\n y = torch.FloatTensor(label)\r\n\r\n\r\n data = Data(x=x, edge_index=edge_index, y=y)\r\n data_list.append(data)\r\n\r\n data, slices = self.collate(data_list)\r\n torch.save((data, slices), self.processed_paths[0])\r\n\r\n\r\n\r\n\r\n\r\n\r\ndataset = YooChooseDataset('./')\r\n\r\ndataset = dataset.shuffle()\r\none_tenth_length = int(len(dataset) * 0.1)\r\ntrain_dataset = dataset[:one_tenth_length * 8] # 80% of entire dataset\r\nval_dataset = dataset[one_tenth_length*8:one_tenth_length * 9] # 10% of entire dataset\r\ntest_dataset = dataset[one_tenth_length*9:] # 10% of entire dataset\r\n#print(len(train_dataset), len(val_dataset), len(test_dataset))\r\n\r\nprint(\"train_dataset=\", train_dataset)\r\nprint(\"val_dataset=\", val_dataset)\r\nprint(\"test_dataset=\", test_dataset)\r\n\r\nbatch_size= 256 #512\r\ntrain_loader = DataLoader(train_dataset, batch_size=batch_size)\r\nval_loader = DataLoader(val_dataset, batch_size=batch_size)\r\ntest_loader = DataLoader(test_dataset, batch_size=batch_size)\r\n\r\nnum_items = df.userId.max() +1\r\nnum_categories = df.movieId.max()+1\r\n#print(num_items , num_categories)\r\n\r\nembed_dim = 128\r\nfrom torch_geometric.nn import GraphConv, TopKPooling, GatedGraphConv, SAGEConv, SGConv,ARMAConv, GATConv,SAGPooling\r\nfrom torch_geometric.nn import TopKPooling\r\nfrom torch_geometric.nn import global_mean_pool as gap, global_max_pool as gmp\r\nimport torch.nn.functional as F\r\n\r\n\r\nclass Net(torch.nn.Module):\r\n def __init__(self):\r\n super(Net, self).__init__()\r\n\r\n self.conv1 = GraphConv(embed_dim * 2, 128)\r\n self.pool1 = TopKPooling(128, ratio=0.9)\r\n\r\n self.conv2 = GraphConv(128, 128)\r\n self.pool2 = TopKPooling(128, ratio=0.9)\r\n\r\n self.conv3 = GraphConv(128, 128)\r\n self.pool3 = TopKPooling(128, ratio=0.9)\r\n\r\n self.item_embedding = torch.nn.Embedding(num_embeddings=num_items, embedding_dim=embed_dim)\r\n self.category_embedding = torch.nn.Embedding(num_embeddings=num_categories, embedding_dim=embed_dim)\r\n\r\n self.lin1 = torch.nn.Linear(256, 256)\r\n self.lin2 = torch.nn.Linear(256, 128)\r\n self.bn1 = torch.nn.BatchNorm1d(128)\r\n self.bn2 = torch.nn.BatchNorm1d(64)\r\n self.act1 = torch.nn.ReLU()\r\n self.act2 = torch.nn.ReLU()\r\n\r\n def forward(self, data):\r\n x, edge_index, batch = data.x, data.edge_index, data.batch\r\n\r\n # print('x =', x)\r\n # print('====================')\r\n\r\n item_id = x[:, :, 0]\r\n category = x[:, :, 1]\r\n\r\n # print(\"item_id: \", item_id)\r\n # print(\"category: \\n\", category)\r\n # print('====================')\r\n\r\n\r\n emb_item = self.item_embedding(item_id).squeeze(1)\r\n emb_category = self.category_embedding(category).squeeze(1)\r\n\r\n # emb_item = emb_item.squeeze(1)\r\n # emb_cat\r\n x = torch.cat([emb_item, emb_category], dim=1)\r\n # print(x.shape)\r\n x = F.relu(self.conv1(x, edge_index))\r\n # print(x.shape)\r\n x, edge_index, _, batch, _ , _ = self.pool1(x, edge_index, None, batch)\r\n x1 = torch.cat([gmp(x, batch), gap(x, batch)], dim=1)\r\n\r\n x = F.relu(self.conv2(x, edge_index))\r\n\r\n x, edge_index, _, batch, _ , _= self.pool2(x, edge_index, None, batch)\r\n x2 = torch.cat([gmp(x, batch), gap(x, batch)], dim=1)\r\n\r\n x = F.relu(self.conv3(x, edge_index))\r\n\r\n x, edge_index, _, batch, _ , _= self.pool3(x, edge_index, None, batch)\r\n x3 = torch.cat([gmp(x, batch), gap(x, batch)], dim=1)\r\n\r\n x = x1 + x2 + x3\r\n\r\n x = self.lin1(x)\r\n x = self.act1(x)\r\n x = self.lin2(x)\r\n x = F.dropout(x, p=0.5, training=self.training)\r\n x = self.act2(x)\r\n\r\n outputs = []\r\n for i in range(x.size(0)):\r\n output = torch.matmul(emb_item[data.batch == i], x[i, :])\r\n\r\n outputs.append(output)\r\n\r\n x = torch.cat(outputs, dim=0)\r\n x = torch.sigmoid(x)\r\n\r\n # save\r\n # savePath = \"./output/test_model.pth\"\r\n # torch.save(model.state_dict(), savePath)\r\n\r\n\r\n return x\r\n\r\n\r\n\r\n\r\ndevice = torch.device('cuda')\r\nmodel = Net().to(device)\r\noptimizer = torch.optim.Adam(model.parameters(), lr=0.001)\r\ncrit = torch.nn.BCELoss()\r\n\r\n\r\n\r\ndef train():\r\n model.train()\r\n\r\n loss_all = 0\r\n for data in train_loader:\r\n data = data.to(device)\r\n optimizer.zero_grad()\r\n output = model(data)\r\n\r\n label = data.y.to(device)\r\n loss = crit(output, label)\r\n loss.backward()\r\n loss_all += data.num_graphs * loss.item()\r\n optimizer.step()\r\n\r\n\r\n\r\n\r\n\r\n #\r\n # f = open('./output/weights.csv','w',newline='')\r\n # wr = csv.writer(f)\r\n # wr.writerows(model.item_embedding.weight)\r\n # f.close()\r\n # for Aweight in model.item_embedding.weight:\r\n # wr.writerow(Aweight)\r\n #\r\n # f.close()\r\n\r\n return loss_all / len(train_dataset)\r\n\r\n\r\nfrom sklearn.metrics import roc_auc_score\r\n\r\ndef evaluate(loader):\r\n model.eval()\r\n\r\n predictions = []\r\n labels = []\r\n\r\n with torch.no_grad():\r\n for data in loader:\r\n data = data.to(device)\r\n pred = model(data).detach().cpu().numpy()\r\n\r\n label = data.y.detach().cpu().numpy()\r\n predictions.append(pred)\r\n labels.append(label)\r\n\r\n predictions = np.hstack(predictions)\r\n labels = np.hstack(labels)\r\n\r\n return roc_auc_score(labels, predictions)\r\n\r\nfor epoch in range(1, 200):\r\n loss = train()\r\n\r\n if epoch == 10:\r\n # print(model.item_embedding.weight)\r\n with open('./output/weights10.csv', 'w', newline='') as f:\r\n writer = csv.writer((f))\r\n for Aweight in model.item_embedding.weight:\r\n writer.writerow(Aweight.tolist())\r\n\r\n if epoch == 50:\r\n # print(model.item_embedding.weight)\r\n with open('./output/weights50.csv', 'w', newline='') as f:\r\n writer = csv.writer((f))\r\n for Aweight in model.item_embedding.weight:\r\n writer.writerow(Aweight.tolist())\r\n\r\n if epoch == 100:\r\n # print(model.item_embedding.weight)\r\n with open('./output/weights100.csv', 'w', newline='') as f:\r\n writer = csv.writer((f))\r\n for Aweight in model.item_embedding.weight:\r\n writer.writerow(Aweight.tolist())\r\n\r\n #train_acc = evaluate(train_loader)\r\n #val_acc = evaluate(val_loader)\r\n test_acc = evaluate(test_loader)\r\n #print('Epoch: {:03d}, Loss: {:.5f}, Train Auc: {:.5f}, Val Auc: {:.5f}, Test Auc: {:.5f}'.format(epoch, loss, train_acc, val_acc, test_acc))\r\n print('Epoch: {:03d}, Loss: {:.5f}, Test Auc: {:.5f}'.format(epoch, loss,test_acc))\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","repo_name":"dumbdumbdumb/TermProject","sub_path":"URGNN.py","file_name":"URGNN.py","file_ext":"py","file_size_in_byte":10098,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38701484079","text":"def dictionarycon(word):\r\n dictionary = {\"mobile\", \"samsung\", \"sam\", \"sung\", \"man\",\r\n \"mango\", \"icecream\", \"and\", \"go\", \"i\", \"love\", \"ice\", \"cream\"}\r\n return word in dictionary\r\ndef wordbb(s):\r\n wordbbutil(s,len(s),\"\")\r\ndef wordbbutil(s,n,result):\r\n for i in range(1,n+1):\r\n prefix=s[:i]\r\n if dictionarycon(prefix):\r\n if i==n:\r\n result+=prefix\r\n print(result)\r\n return\r\n wordbbutil(s[i:],n-i,result+prefix+\" \")\r\nif __name__ == '__main__':\r\n\r\n wordbb(\"ilovesamsungmobile\")\r\n","repo_name":"Abhishek-IOT/Data_Structures","sub_path":"DATA_STRUCTURES/DSA Questions/Backtracking/Wordbread.py","file_name":"Wordbread.py","file_ext":"py","file_size_in_byte":586,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"3289877609","text":"import pytest\nfrom mock import MagicMock\n\nimport onair.src.reasoning.agent as agent\nfrom onair.src.reasoning.agent import Agent\n\n# __init__ tests\ndef test_Agent__init__sets_vehicle_rep_to_given_vehicle_and_learning_systems_and_mission_status_and_bayesian_status(mocker):\n # Arrange\n arg_vehicle = MagicMock()\n arg_learners_plugin_dict = MagicMock()\n arg_planners_plugin_dict = MagicMock()\n arg_complex_plugin_dict = MagicMock()\n\n fake_headers = MagicMock()\n fake_headers.__len__.return_value = 1 # Fake a header\n fake_learning_systems = MagicMock()\n fake_planning_systems = MagicMock()\n fake_complex_systems = MagicMock()\n fake_mission_status = MagicMock()\n fake_bayesian_status = MagicMock()\n\n mocker.patch.object(arg_vehicle, 'get_headers', return_value=fake_headers)\n mocker.patch.object(arg_vehicle, 'get_status', return_value=fake_mission_status)\n mocker.patch.object(arg_vehicle, 'get_bayesian_status', return_value=fake_bayesian_status)\n mocker.patch(agent.__name__ + '.LearnersInterface', return_value=fake_learning_systems)\n mocker.patch(agent.__name__ + '.PlannersInterface', return_value=fake_planning_systems)\n mocker.patch(agent.__name__ + '.ComplexReasoningInterface', return_value=fake_complex_systems)\n\n cut = Agent.__new__(Agent)\n\n # Act\n cut.__init__(arg_vehicle, arg_learners_plugin_dict, arg_planners_plugin_dict, arg_complex_plugin_dict)\n\n # Assert\n assert cut.vehicle_rep == arg_vehicle\n assert arg_vehicle.get_headers.call_count == 3\n assert arg_vehicle.get_headers.call_args_list[0].args == ()\n assert agent.LearnersInterface.call_count == 1\n assert agent.LearnersInterface.call_args_list[0].args == (fake_headers, arg_learners_plugin_dict)\n assert cut.learning_systems == fake_learning_systems\n assert agent.PlannersInterface.call_count == 1\n assert agent.PlannersInterface.call_args_list[0].args == (fake_headers, arg_planners_plugin_dict)\n assert cut.planning_systems == fake_planning_systems\n assert agent.ComplexReasoningInterface.call_count == 1\n assert agent.ComplexReasoningInterface.call_args_list[0].args == (fake_headers, arg_complex_plugin_dict)\n assert cut.complex_reasoning_systems == fake_complex_systems\n assert arg_vehicle.get_status.call_count == 1\n assert arg_vehicle.get_status.call_args_list[0].args == ()\n assert cut.mission_status == fake_mission_status\n assert arg_vehicle.get_bayesian_status.call_count == 1\n assert arg_vehicle.get_bayesian_status.call_args_list[0].args == ()\n assert cut.bayesian_status == fake_bayesian_status\n\n# render_resoning tests\ndef test_Agent_render_reasoning_returns_call_to_complex_reasoning_systems_render_reasoning(mocker):\n # Arrange\n expected_result = MagicMock()\n fake_complex_resoning_systems = MagicMock()\n\n cut = Agent.__new__(Agent)\n cut.complex_reasoning_systems = fake_complex_resoning_systems\n\n mocker.patch.object(fake_complex_resoning_systems, 'render_reasoning', return_value=expected_result)\n\n # Act\n result = cut.render_reasoning()\n\n # Assert\n assert result == expected_result\n\n# reason tests\ndef test_Agent_reason_updates_vehicle_rep_with_given_frame_and_updates_learning_systems_with_vehicle_curr_data_and_new_mission_status(mocker):\n # Arrange\n arg_frame = MagicMock()\n fake_vehicle_rep = MagicMock()\n fake_vehicle_rep.curr_data = MagicMock()\n\n # Mock and patch\n fake_status = MagicMock()\n fake_PDDL_state = MagicMock()\n fake_state = MagicMock()\n fake_learning_systems = MagicMock()\n fake_planning_systems = MagicMock()\n fake_complex_reasoning_systems = MagicMock()\n fake_learning_systems_reasoning = MagicMock()\n fake_planning_systems_reasoning = MagicMock()\n expected_aggregate_high_level_info = {'vehicle_rep': fake_state, \n 'learning_systems':fake_learning_systems_reasoning,\n 'planning_systems':fake_planning_systems_reasoning}\n expected_result = MagicMock()\n\n cut = Agent.__new__(Agent)\n cut.vehicle_rep = fake_vehicle_rep\n cut.learning_systems = fake_learning_systems\n cut.planning_systems = fake_planning_systems\n cut.complex_reasoning_systems = fake_complex_reasoning_systems\n\n mock_manager = mocker.MagicMock()\n\n mock_manager.attach_mock(mocker.patch.object(fake_vehicle_rep, 'update'), 'cut.vehicle_rep.update')\n mock_manager.attach_mock(mocker.patch.object(fake_learning_systems, 'update'), 'cut.learning_systems.update')\n mock_manager.attach_mock(mocker.patch.object(fake_planning_systems, 'update'), 'cut.planning_systems.update')\n mock_manager.attach_mock(mocker.patch.object(fake_complex_reasoning_systems, 'update'), 'cut.complex_reasoning_systems.update')\n mock_manager.attach_mock(mocker.patch.object(cut, 'render_reasoning', return_value=expected_result), 'cut.render_reasoning')\n\n mocker.patch.object(fake_vehicle_rep, 'get_state_information', side_effect=[fake_status, fake_PDDL_state, fake_state])\n mocker.patch.object(fake_learning_systems, 'render_reasoning', return_value=fake_learning_systems_reasoning)\n mocker.patch.object(fake_planning_systems, 'render_reasoning', return_value=fake_planning_systems_reasoning)\n\n # Act\n result = cut.reason(arg_frame)\n\n # Assert\n result = expected_result\n mock_manager.assert_has_calls([\n mocker.call.cut.vehicle_rep.update(arg_frame),\n mocker.call.cut.learning_systems.update(fake_vehicle_rep.curr_data, fake_status),\n mocker.call.cut.planning_systems.update(fake_PDDL_state),\n mocker.call.cut.complex_reasoning_systems.update(expected_aggregate_high_level_info),\n ], any_order=False)\n assert cut.vehicle_rep.get_state_information.call_count == 3\n assert cut.vehicle_rep.get_state_information.call_args_list[0].args == ()\n assert cut.vehicle_rep.get_state_information.call_args_list[1].args == ()\n assert cut.vehicle_rep.get_state_information.call_args_list[2].args == ()\n assert cut.learning_systems.render_reasoning.call_count == 1\n assert cut.learning_systems.render_reasoning.call_args_list[0].args == ()\n assert cut.planning_systems.render_reasoning.call_count == 1\n assert cut.planning_systems.render_reasoning.call_args_list[0].args == ()\n\n \n# diagnose tests\ndef test_Agent_diagnose_returns_empty_Dict():\n # Arrange\n arg_time_step = MagicMock()\n\n cut = Agent.__new__(Agent)\n cut.learning_systems = MagicMock()\n cut.planning_systems = MagicMock()\n cut.bayesian_status = MagicMock()\n cut.vehicle_rep = MagicMock()\n\n # Act\n result = cut.diagnose(arg_time_step)\n\n # Assert\n assert type(result) == dict\n assert result == {}\n","repo_name":"nasa/OnAIR","sub_path":"test/onair/src/reasoning/test_agent.py","file_name":"test_agent.py","file_ext":"py","file_size_in_byte":6716,"program_lang":"python","lang":"en","doc_type":"code","stars":29,"dataset":"github-code","pt":"79"} +{"seq_id":"72530440895","text":"# -*- coding: utf-8 -*- \n# leetcode time cost : 1824 ms\n# leetcode memory cost : 15.4 MB\n# Time Complexity: O(N*N)\n# Space Complexity: O(1)\n\n# Definition for singly-linked list.\nclass ListNode:\n def __init__(self, x):\n self.val = x\n self.next = None\n\nclass Solution:\n def insertionSortList(self, head: ListNode) -> ListNode:\n newHead = sortPosition = ListNode(-1)\n currentNode = head\n while currentNode:\n # save the next node in unsorted list\n nextNode = currentNode.next\n # loop new link list to find the position of new current node\n while sortPosition.next and sortPosition.next.val < currentNode.val:\n sortPosition = sortPosition.next\n # insert node to new list before the node just bigger than it\n currentNode.next = sortPosition.next\n sortPosition.next = currentNode\n sortPosition = newHead\n # restore the link list loop\n currentNode = nextNode\n return newHead.next","repo_name":"sky-dream/LeetCodeProblemsStudy","sub_path":"[0147][Medium][ Insertion_Sort_List]/Insertion_Sort_List.py","file_name":"Insertion_Sort_List.py","file_ext":"py","file_size_in_byte":1052,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"2873382222","text":"import json\nfrom django.http import JsonResponse\nfrom django.shortcuts import render, redirect, get_object_or_404, HttpResponse\nfrom .scraperfile import MoodleScraper\nfrom django.contrib.auth import views, login, authenticate\nfrom .forms import RegisterForm, MoodleDetailsForm, SignInForm\nfrom .models import Course, User, MoodleDetails\nfrom django.contrib.auth.decorators import login_required\nfrom python_paystack.objects.transactions import Transaction\nfrom python_paystack.managers import TransactionsManager\nfrom django.views.decorators.http import require_http_methods\nfrom django.views.decorators.csrf import csrf_exempt\nfrom datetime import datetime\n\n\n\ndef index(request):\n return render(request, \"scraper/index.html\")\n\nclass SigninView(views.LoginView):\n template_name = \"registration/login.html\"\n authentication_form = SignInForm\n \n\ndef register(request):\n if request.POST:\n form = RegisterForm(request.POST)\n if form.is_valid():\n form.save()\n email = form.cleaned_data[\"email\"] \n password = form.cleaned_data[\"password1\"]\n user = authenticate(request, email=email, password=password)\n login(request, user)\n return redirect(\"student\") \n else:\n form = RegisterForm()\n return render(request, \"registration/register.html\", {\"form\":form})\n\ndef logout(request):\n return views.logout_then_login(request)\n\n@login_required()\ndef student(request):\n user = request.user\n if request.POST:\n form = MoodleDetailsForm(request.POST)\n if form.is_valid():\n username = form.cleaned_data[\"matric_num\"]\n password = form.cleaned_data[\"password\"]\n context = {}\n \n try:\n MoodleDetails.objects.get(user=user)\n except MoodleDetails.DoesNotExist:\n MoodleDetails.objects.create(user=user, username=username, password=password)\n person, error = create_session(request)\n if \"invalid\" in error:\n print(f\"Invalid login : {error}\" )\n context[\"error\"] = error\n return render(request, \"scraper/student.html\", context)\n details = person.get_course_links()\n for index in range(0, len(details), 2):\n Course.objects.create(user=user, course_name=details[index+1], course_url=details[index])\n else:\n form = MoodleDetailsForm()\n user_courses = Course.objects.filter(user=user)\n user_course_one = [course for course in user_courses[::2]]\n user_course_two = [course for course in user_courses[1::2]]\n last_login = user.last_login.strftime(\"%b %d, %Y\")\n context = {\n \"course_one\":user_course_one,\n \"course_two\":user_course_two,\n \"user\": user,\n \"form\":form,\n \"last_login\":last_login,\n }\n return render(request, \"scraper/student.html\", context)\n\ndef create_session(request):\n user = request.user\n detail = MoodleDetails.objects.get(user=user)\n username = detail.username\n password = detail.password\n person = MoodleScraper(username, password)\n error = person.login_student()\n return person, error\n\n@login_required()\ndef get_a_course_page(request, id):\n user = request.user\n user_points = user.userprofile.points\n if user_points != 0:\n course = get_object_or_404(Course, id=id)\n course_name = (course.course_name)[:6]\n person, error = create_session(request)\n if \"invalid\" in error:\n data = {\"error\":error}\n return JsonResponse(data) \n person.get_course_pages(course=course_name)\n user.userprofile.points -= 1\n user.userprofile.save()\n data = {\"success\":f\"Just went to {course.course_name}\"}\n else:\n data = {\"error\":\"No more points, You need to purchase\"}\n return JsonResponse(data)\n\n@login_required()\ndef get_all_courses_page(request): # Make this a form\n if request.method == \"POST\":\n check = request.POST.get(\"check\")\n user = request.user\n user_points = user.userprofile.points\n if user_points != 0:\n person, error = create_session(request)\n if \"invalid\" in error:\n data = {\"error\":error}\n return JsonResponse(data)\n if check !=\"true\":\n data = {\"courses_gone\" : person.get_course_pages()}\n else:\n data = {\"courses_gone\" : person.get_course_pages(True)}\n else:\n data = {\"error\" : \"No more points, You need to purchase\"}\n return JsonResponse(data)\n\n\ndef download_course(request, id):\n course = course = get_object_or_404(Course, id=id)\n course_name = (course.course_name)[:6]\n person, error = create_session(request)\n if \"invalid\" in error:\n data = {\"error\":error}\n return JsonResponse(data)\n person.download_course_file(course_name)\n data = {\"success\":\"Your notes are downloaded\"}\n return JsonResponse(data)\n\n\n@login_required()\ndef make_payment(request, id):\n user = request.user\n email = user.email\n if id == 1:\n transaction = Transaction(50000, email)\n elif id == 2:\n transaction = Transaction(70000, email)\n transaction_manager = TransactionsManager()\n transaction = transaction_manager.initialize_transaction('STANDARD', transaction)\n return redirect(transaction.authorization_url)\n\n\n@csrf_exempt\n@require_http_methods([\"POST\"])\ndef check_payment(request):\n if request.method == \"POST\":\n response = json.loads(request.body)\n event = response[\"event\"]\n status = response[\"data\"][\"status\"]\n sent_email = response[\"data\"][\"customer\"][\"email\"]\n plan = response[\"data\"][\"plan\"][\"name\"]\n amount_paid = response[\"data\"][\"amount\"]\n try:\n user = get_object_or_404(User, email=sent_email)\n if event == \"charge.success\" and status == \"success\":\n print(\"paid\")\n user.userprofile.points = 10000000000\n if plan == \"basic\" and amount_paid == 70000:\n user.userprofile.plan = \"BC\"\n else:\n user.userprofile.plan == \"BB\"\n user.userprofile.paid = True\n user.save()\n except:\n HttpResponse(status_code=400) \n \n return HttpResponse('success')\n\n\n\"\"\"\nMake sure redirecting in Studdent doesn't cause errors-checked\nCreate error channel when logging in create_session, done, remains test\n\"\"\"\n\n\n","repo_name":"NerdPraise/MoodleScraperDjango","sub_path":"scraper/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":6561,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39505132409","text":"from amplification.run import main, parse_args\nimport argparse\n\ndef combos(*xs):\n if xs:\n return [x + combo for x in xs[0] for combo in combos(*xs[1:])]\n else:\n return [()]\n\ndef each(*xs):\n return [y for x in xs for y in x]\n\ndef bind(var, val, descriptor=''):\n return [((var, val, descriptor),)]\n\ndef label(descriptor):\n return bind(None, None, str(descriptor))\n\ndef options(var, opts):\n return each(*[bind(var, val, descriptor) for val, descriptor in opts])\n\ndef repeat(n):\n return each(*[label(i) for i in range(n)])\n\ndef dict_of_dicts_assign(d, ks, v):\n if len(ks) == 0:\n return v\n k = ks[0]\n d[k] = dict_of_dicts_assign(d.get(k, {}), ks[1:], v)\n return d\n\ndef run_experiment(trials, name, dry_run=False):\n for trial in trials:\n descriptors = []\n kwargs = {}\n for k, v, s in trial:\n if k is not None: kwargs[k] = v\n if s is not '': descriptors.append(s)\n if dry_run:\n kwargs[\"train.stub\"] = True\n for k in [\"num_cpu\", \"num_gpu\"]:\n if k in kwargs:\n del kwargs[k]\n main(**parse_args(kwargs))\n else:\n raise NotImplementedError(\"this code path has been removed\")\n\ndef cpus(n): return bind(\"num_cpu\", n)\ndef gpus(n): return bind(\"num_gpu\", n)\n\namplify_opts = options(\"train.amplify\", [(True, \"amp\"), (False, \"sup\")])\ncurriculum_opts = options(\"train.curriculum\", [(True, \"cy\"), (False, \"cn\")])\ndef sizes(*xs): return options(\"task.size\", [(x, str(x)) for x in xs])\ndef tasks(*xs): return options(\"task.name\", [(x, x) for x in xs])\nall_tasks = tasks('graph', 'sum', 'iter', 'eval', 'equals')\n\ntest = combos(cpus(4), gpus(2), bind(\"task.name\", \"evalsum\", \"evalmod\"))\n\nmay16 = combos(\n each(\n combos(cpus(4), gpus(2), bind(\"train.supervised\", False, \"amp\")),\n combos(cpus(2), gpus(1), bind(\"train.supervised\", True, \"sup\"))\n ),\n all_tasks\n)\n\njan30 = combos(\n cpus(4), gpus(2),\n each(\n all_tasks,\n bind(\"task.name\", \"evalsum\", \"evalmod\"),\n combos(\n bind(\"task.name\", \"evalsum\", \"evalsum\"),\n bind(\"task.modulus\", None),\n ),\n combos(\n bind(\"task.name\", \"sum\", \"sumraw\"),\n bind(\"task.modulus\", None),\n )\n )\n)\n\njan30_sup = combos(bind(\"train.supervised\", True, \"sup\"), jan30)\n\njan29_sum = combos(cpus(4), gpus(2), bind(\"task.name\", \"sum\", \"sum\"), bind(\"task.modulus\", None))\n\njan25_sup = combos(cpus(4), gpus(2), bind(\"train.supervised\", True), bind(\"task.name\", \"sum\", \"sum\"), bind(\"task.modulus\", None))\n\njan25_odds = combos(cpus(4), gpus(2),\n each(\n bind(\"task.name\", \"evalsum\", \"evalmod\"),\n combos(\n bind(\"task.name\", \"evalsum\", \"evalsum\"),\n bind(\"task.modulus\", None),\n ),\n combos(\n bind(\"task.name\", \"sum\", \"sumraw\"),\n bind(\"task.modulus\", None),\n ),\n ),\n)\n\njan25 = combos(all_tasks, cpus(4), gpus(2))\n\njan22_fast = combos(all_tasks, bind(\"train.generation_frequency\", 20))\n\njan22 = combos(\n each(\n combos(\n bind(\"task.name\", \"iter\", \"iter\"),\n options(\"task.log_iters\", [(6, \"l6\"), (7, \"l7\")]),\n options(\"model.answerer.depth\", [(6, \"d6\"), (10, \"d10\")])\n ),\n combos(\n bind(\"train.supervised\", True, \"supervised\"),\n all_tasks,\n )\n )\n)\n\njan21_eval = combos(\n bind(\"task.name\", \"eval\"),\n options(\"model.answerer.depth\", [(6, \"d6\"), (10, \"d10\")])\n)\n\njan21 = combos(\n all_tasks, options(\"model.answerer.depth\", [(6, \"d6\"), (10, \"d10\")])\n)\n\nstandard = combos(\n all_tasks, sizes(64),\n)\n\njan15_final = combos(\n each(\n combos(tasks('iter', 'sum'), sizes(16, 32, 64)),\n combos(tasks('eval'), sizes(64)),\n combos(tasks('graph'), sizes(32, 64)),\n ),\n each(\n label(\"prefix\"),\n bind(\"train.random_subset\", True, \"random\"),\n bind(\"train.curriculum\", False, \"none\")\n ),\n)\njan15_variants = combos(\n tasks('iter', 'sum', 'graph'), \n sizes(64),\n each(\n bind('train.loss_threshold', 0.1, \"cautious\"),\n bind(\"train.buffer_size\", 1000, 'small'),\n )\n)\n\njan15_encodings = combos(\n each(\n label(\"prefix\"),\n bind(\"train.random_subset\", True, \"random\"),\n bind(\"train.curriculum\", False, \"none\")\n ),\n bind(\"model.answerer.encoder\", \"concat\", \"concat\"),\n tasks('iter', 'graph', 'sum'), sizes(64),\n)\n\njan15_fast_curriculum = combos(\n each(\n label(\"prefix\"),\n bind(\"train.random_subset\", True, \"random\"),\n bind(\"train.curriculum\", False, \"none\")\n ),\n all_tasks, sizes(64),\n)\n\njan15_curriculum = combos(\n each(\n label(\"prefix\"),\n bind(\"train.random_subset\", True, \"random\"),\n bind(\"train.curriculum\", False, \"none\")\n ),\n tasks(\"iter\", \"sum\", \"graph\", \"eval\", \"equals\"),\n sizes(32, 64),\n)\n\njan15_eval = combos(\n bind(\"train.num_steps\", 200000),\n tasks(\"eval\"),\n sizes(16, 32, 64),\n)\n\njan15_equals = combos(\n bind(\"train.num_steps\", 200000),\n tasks(\"equals\"),\n sizes(32, 64),\n)\n\njan14 = combos(\n bind(\"train.num_steps\", 200000),\n each(\n combos(tasks('iter', 'sum'), sizes(16, 32, 64)),\n combos(tasks('eval'), sizes(64)),\n combos(tasks('graph'), sizes(32, 64)),\n )\n)\n\nharder_tasks = each(\n bind(\"task.name\", 'iter', 'iter'),\n bind(\"task.name\", 'equals', 'equals'),\n combos(\n bind(\"task.name\", 'graph', 'graph'),\n bind(\"task.size\", 20, \"20\")\n ),\n combos(\n bind(\"task.name\", \"sum\", \"sum\"),\n bind(\"task.length\", 6, \"6\"),\n ),\n combos(\n bind(\"task.name\", \"eval\", \"eval\"),\n bind(\"task.size\", 36, \"36\"),\n )\n)\njan11_final = combos(harder_tasks, bind(\"train.num_steps\", 200000))\njan11_answerer = combos(harder_tasks, each(bind(\"train.learn_human_model\", False, \"noasker\"), bind(\"train.supervised\", True, \"sup\")))\njan11_catchup = combos(bind(\"task.name\", 'equals', 'equals'), each(bind(\"train.learn_human_model\", False, \"noasker\"), bind(\"train.supervised\", True, \"sup\")))\n\ndropout = combos(\n options(\"task.name\", [(\"iter\", \"iter\"), (\"sum\", \"sum\"), (\"graph\", \"graph\")]),\n options(\"model.asker.p_drop\", [(0.0, \"00\"), (0.15, \"15\")]),\n options(\"train.asker_data_limit\", [(300, \"300\"), (900, \"900\")]),\n bind(\"train.adjust_drift_epsilon\", False),\n options(\"train.initial_drift_epsilon\", [(1e-2, \"e2\"), (1e-3, \"e3\"), (1e-4, \"e4\")]),\n bind(\"train.num_steps\", 20000),\n bind(\"train.just_asker\", True),\n)\n\n\nsupervised = combos(all_tasks, bind(\"train.just_asker\", True))\njan9 = combos(all_tasks, repeat(2))\njan10 = combos(all_tasks, options(\"train.learn_human_model\", [(False, \"noasker\"), (True, \"full\")]))\njan10_iter = combos(bind(\"task.name\", \"iter\", \"iter\"), options(\"train.learn_human_model\", [(False, \"noasker\"), (True, \"full\")]))\njan10_sup = combos(all_tasks, bind(\"train.supervised\", True, \"sup\"))\n\niterate = combos(bind(\"task.name\", \"iterate\"),\n sizes(8, 40), options(\"task.bit_length\", [(3, \"3\"), (5, \"5\")]),\n amplify_opts, repeat(2), bind(\"train.curriculum\", False))\n\nevals = combos(\n bind(\"task.name\", \"eval\"),\n sizes(20, 100),\n bind(\"train.curriculum\", True),\n repeat(2)\n)\n\ngraph = combos(\n bind(\"task.name\", \"graph\"),\n bind(\"train.curriculum\", True),\n options(\"task.size\", [(20, \"100\"), (8, \"8\")]),\n bind(\"train.amplify\", True),\n bind(\"train.nbatch\", 50),\n bind(\"train.num_steps\", 400000),\n repeat(2),\n)\n\nsums = combos(\n bind(\"task.name\", \"sum\"),\n bind(\"train.curriculum\", False),\n options(\"task.size\", [(3, \"3\"), (4, \"4\"), (5, \"5\")]),\n options(\"train.amplify\", [(True, \"amp\"), (False, \"sup\")]),\n bind(\"train.nbatch\", 50),\n bind(\"train.num_steps\", 300000),\n)\n\nsearch = combos(\n bind(\"task.name\", \"search\"),\n bind(\"train.curriculum\", False),\n options(\"task.size\", [(10, \"10\"), (100, \"100\")]),\n options(\"train.amplify\", [(True, \"amp\"), (False, \"sup\")]),\n bind(\"train.nbatch\", 50),\n bind(\"train.num_steps\", 100000),\n)\n\nequality = combos(\n bind(\"task.name\", \"equals100\"),\n bind(\"train.curriculum\", True, \"cy\"),\n options(\"train.amplify\", [(True, \"amp\"), (False, \"sup\")]),\n bind(\"train.nbatch\", 50),\n bind(\"train.num_steps\", 300000),\n repeat(2),\n)\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=\"run an experiment\")\n parser.add_argument(\"-e\", \"--experiment\")\n parser.add_argument(\"--dry\", default=False, action='store_const', const=True)\n parser.add_argument(\"-n\", \"--name\")\n n = parser.parse_args()\n trials = globals()[n.experiment]\n run_experiment(trials, n.name, dry_run=n.dry)\n","repo_name":"paulfchristiano/amplification","sub_path":"experiment.py","file_name":"experiment.py","file_ext":"py","file_size_in_byte":8875,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"79"} +{"seq_id":"9355711773","text":"#!/usr/bin/env python\n#coding:utf-8\n\nfrom django.shortcuts import render,redirect\nfrom django.http import JsonResponse\nfrom django.http import HttpResponse\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required\n\n\ndef mylogin(request):\n print(request.method)\n if request.method == 'GET':\n return render(request,'login.html')\n elif request.method == 'POST':\n # print(request.POST)\n u = request.POST.get('u')\n p = request.POST.get('p')\n print(u,p)\n user = authenticate(username=u,password=p)\n print(user)\n if user:\n login(request,user)\n return redirect('/index/')\n\n return redirect('/login/')\n\ndef mylogout(request):\n print(request.user)\n logout(request)\n return HttpResponse('logout success')\n\n@login_required\ndef index(request):\n print(request.user)\n return render(request,'index.html')","repo_name":"HiJen/devops7","sub_path":"lesson1/zhangpengfei/myweb/myweb/view.py","file_name":"view.py","file_ext":"py","file_size_in_byte":956,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27135049052","text":"import gzip\nimport sys\nimport numpy as np\n\nHAPMAP_DIR=\"/data/share/HapMap/phasing/r22/2008-12-16\"\n\n\n\ndef read_haplotypes(hap_file):\n \"\"\"Reads haplotypes for the current chromosome into a big matrix\"\"\"\n rows = []\n\n for line in hap_file:\n vals = np.array([int(x) for x in line.rstrip().split()], \n dtype=np.uint8)\n rows.append(vals)\n\n return np.array(rows)\n\n \ndef main():\n if len(sys.argv) != 2:\n sys.stderr.write('usage: %s \\n' % sys.argv[0])\n exit(2)\n \n chrom = sys.argv[1]\n\n legend_path = \"%s/genotypes_%s_YRI_r22_nr.b36_fwd_legend.txt.gz\" % \\\n (HAPMAP_DIR, chrom)\n \n phase_path = \"%s/genotypes_%s_YRI_r22_nr.b36_fwd.phase.gz\" % \\\n (HAPMAP_DIR, chrom)\n\n legend_f = gzip.open(legend_path, \"rb\")\n\n # read haplotypes\n phase_f = gzip.open(phase_path, \"rb\") \n sys.stderr.write(\"reading haplotypes for each individual\\n\")\n hap_array = read_haplotypes(phase_f)\n n_hap = hap_array.shape[1]\n phase_f.close()\n \n header = legend_f.readline()\n\n sys.stderr.write(\"writing haplotypes for each SNP\\n\")\n i = 0\n for line in legend_f:\n legend = line.rstrip().split()\n\n # write 5 header columns containing identifier (twice),\n # position of SNP, and 2 alleles\n sys.stdout.write(\"%s %s %s %s %s \" % (legend[0], legend[0],\n legend[1], legend[2], \n legend[3]))\n\n # write all genotypes for this SNP, one for each haplotype\n sys.stdout.write(\" \".join([\"%d\" % x for x in hap_array[0:n_hap, i]])\n + \"\\n\")\n\n i += 1\n if (i % 100) == 0:\n sys.stderr.write(\".\")\n \n \n legend_f.close()\n\n sys.stderr.write(\"\\ndone\")\n \n\n\nmain()\n","repo_name":"gmcvicker/impute","sub_path":"python/make_phased_hapmap_genotype_file.py","file_name":"make_phased_hapmap_genotype_file.py","file_ext":"py","file_size_in_byte":1839,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"25707818274","text":"from decimal import Decimal\nfrom time import sleep\n\nfrom cryptofeed.defines import TRADES, L2_BOOK, L3_BOOK, TICKER, FUNDING, OPEN_INTEREST\n\nSTATS = 'stats'\nimport requests\n\nfrom cryptostore.custom_work.stats2 import Stats\nfrom cryptostore.data.store import Store\n\n\ndef chunk(iterable, length):\n return (iterable[i: i + length] for i in range(0, len(iterable), length))\n\n\nMAX_RETRIES = 10\n\n\nclass InfluxDB(Store):\n def __init__(self, config: dict):\n self.data = None\n self.host = config.host\n self.db = config.db\n self.addr = f\"{config.host}/write?db={config.db}\"\n if 'create' in config and config.create:\n for i in range(MAX_RETRIES):\n try:\n r = requests.post(f'{config.host}/query', data={'q': f'CREATE DATABASE {config.db}'})\n r.raise_for_status()\n except:\n sleep(5)\n\n def aggregate(self, data):\n self.data = data\n\n def write(self, exchange, data_type, pair, timestamp):\n if not self.data:\n return\n agg = []\n # influx cant handle duplicate data (?!) so we need to\n # incremement timestamps on data that have the same timestamp\n used_ts = set()\n if data_type == TRADES:\n for entry in self.data:\n ts = int(Decimal(entry[\"timestamp\"]) * 1000000000)\n while ts in used_ts:\n ts += 1\n used_ts.add(ts)\n if 'id' in entry:\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]} side=\"{entry[\"side\"]}\",id=\"{entry[\"id\"]}\",amount={entry[\"amount\"]},price={entry[\"price\"]},timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]} {ts}')\n else:\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]} side=\"{entry[\"side\"]}\",amount={entry[\"amount\"]},price={entry[\"price\"]},timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]} {ts}')\n elif data_type == STATS:\n for entry, granularity in self.data:\n typed_entry: Stats = entry\n ts = int(Decimal(typed_entry.ts) * 1000000000)\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]},granularity={granularity} '\n f'open={typed_entry.open},high=\"{typed_entry.high}\",low={typed_entry.low},close={typed_entry.close}'\n f',volume={typed_entry.volume} {ts}'\n )\n\n elif data_type == TICKER:\n for entry in self.data:\n ts = int(Decimal(entry[\"timestamp\"]) * 1000000000)\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]} bid={entry[\"bid\"]},ask={entry[\"ask\"]},timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]} {ts}')\n\n elif data_type == L2_BOOK:\n for entry in self.data:\n ts = int(Decimal(entry[\"timestamp\"]) * 1000000000)\n while ts in used_ts:\n ts += 1\n used_ts.add(ts)\n\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]},delta={entry[\"delta\"]} side=\"{entry[\"side\"]}\",timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]},price={entry[\"price\"]},amount={entry[\"size\"]} {ts}')\n elif data_type == L3_BOOK:\n for entry in self.data:\n ts = int(Decimal(entry[\"timestamp\"]) * 1000000000)\n while ts in used_ts:\n ts += 1\n used_ts.add(ts)\n\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]},delta={entry[\"delta\"]} side=\"{entry[\"side\"]}\",id=\"{entry[\"order_id\"]}\",timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]},price=\"{entry[\"price\"]}\",amount=\"{entry[\"size\"]}\" {ts}')\n ts += 1\n elif data_type == FUNDING:\n for entry in self.data:\n formatted = [f\"{key}={value}\" for key, value in entry.items() if isinstance(value, float)]\n formatted = ','.join(\n formatted + [f'{key}=\"{value}\"' for key, value in entry.items() if not isinstance(value, float)])\n agg.append(f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]} {formatted}')\n elif data_type == OPEN_INTEREST:\n for entry in self.data:\n ts = int(Decimal(entry[\"timestamp\"]) * 1000000000)\n agg.append(\n f'{data_type}-{entry[\"feed\"]},pair={entry[\"pair\"]},exchange={entry[\"feed\"]} open_interest={entry[\"open_interest\"]},timestamp={entry[\"timestamp\"]},receipt_timestamp={entry[\"receipt_timestamp\"]} {ts}')\n\n # https://v2.docs.influxdata.com/v2.0/write-data/best-practices/optimize-writes/\n # Tuning docs indicate 5k is the ideal chunk size for batch writes\n for c in chunk(agg, 5000):\n c = '\\n'.join(c)\n r = requests.post(self.addr, data=c)\n r.raise_for_status()\n self.data = None\n\n def get_start_date(self, exchange: str, data_type: str, pair: str) -> float:\n try:\n r = requests.get(f\"{self.host}/query?db={self.db}\", params={\n 'q': f'SELECT first(timestamp) from \"{data_type}-{exchange}\" where pair=\\'{pair}\\''})\n return r.json()['results'][0]['series'][0]['values'][0][1]\n except Exception:\n return None\n","repo_name":"dsaccon/crypto_arb","sub_path":"cryptostore/data/influx.py","file_name":"influx.py","file_ext":"py","file_size_in_byte":5747,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"34291274813","text":"from typing import Any, Sequence, Optional\nimport sys\n\n\nclass LinkedList:\n class Node:\n \"\"\"\n Внутренний класс, класса LinkedList.\n\n Пользователь напрямую не работает с узлами списка, узлами оперирует список.\n \"\"\"\n\n def __init__(self, value: Any, next_: Optional['Node'] = None):\n \"\"\"\n Создаем новый узел для односвязного списка\n\n :param value: Любое значение, которое помещено в узел\n :param next_: следующий узел, если он есть\n \"\"\"\n self.value = value\n self._next = next_ # Вызывается сеттер\n\n @property\n def next(self):\n \"\"\"Getter возвращает следующий узел связного списка\"\"\"\n return self._next\n\n @next.setter\n def next(self, next_: Optional['Node']):\n \"\"\"Setter проверяет и устанавливает следующий узел связного списка\"\"\"\n self._check_node(next_)\n self._next = next_\n\n def _check_node(self, node):\n if not isinstance(node, self.__class__) and node is not None:\n msg = f\"Устанавливаемое значение должно быть экземпляром класса {self.__class__.__name__} \" \\\n f\"или None, не {node.__class__.__name__}\"\n raise TypeError(msg)\n\n def __repr__(self):\n \"\"\"Метод должен возвращать строку, показывающую, как может быть создан экземпляр.\"\"\"\n return f\"{self.__class__.__name__}({self.value}, {self.next})\"\n\n def __str__(self):\n \"\"\"Вызывается функциями str, print и format. Возвращает строковое представление объекта.\"\"\"\n return f\"{self.value}\"\n\n def __del__(self):\n return f\"The {self.__repr__()} has been removed\"\n\n def __init__(self, data: Sequence = None):\n \"\"\"Конструктор связного списка\"\"\"\n self._len = 0\n self.head = None # Node\n self.tail = None\n\n if self._is_iterable(data): # ToDo Проверить, что объект итерируемый. Метод self.is_iterable\n for value in data:\n self.append(value)\n # print(f\"Element {value} has been added to {self.__class__.__name__}\")\n\n def __str__(self):\n \"\"\"Вызывается функциями str, print и format. Возвращает строковое представление объекта.\"\"\"\n return f\"{[value for value in self]}\"\n\n def __repr__(self):\n \"\"\"Метод должен возвращать строку, показывающую, как может быть создан экземпляр.\"\"\"\n return f\"{self.__class__.__name__}({[value for value in self]})\"\n\n def __len__(self):\n return self._len\n\n def __step_by_step_on_nodes(self, index) -> Node:\n # print(f'Method {self.__step_by_step_on_nodes.__name__} called')\n if not isinstance(index, int):\n raise TypeError(f\"Index must be {int.__name__} not {index.__class__.__name__}\")\n\n if not -self._len <= index < self._len:\n raise IndexError(f'IndexError: {self.__class__.__name__} assignment index out of range')\n\n if index < 0:\n index += self._len\n\n current_node = self.head\n for _ in range(index):\n current_node = current_node.next\n return current_node\n\n def __getitem__(self, key: int) -> Any:\n # print(f'Method {self.__getitem__.__name__} called')\n current_node = self.__step_by_step_on_nodes(key)\n return current_node.value\n\n def __setitem__(self, key, value: Any):\n print(f'Method {self.__setitem__.__name__} called')\n current_node = self.__step_by_step_on_nodes(key)\n current_node.value = value\n\n def __delitem__(self, key):\n print(f'Method {self.__delitem__.__name__} called')\n if key < 0:\n key += self._len\n if 0 < key < self._len - 1:\n prev_node = self.__step_by_step_on_nodes(key - 1)\n delete_node = prev_node.next\n next_node = delete_node.next\n delete_node = None\n self.__linked_nodes(prev_node, next_node)\n self._len -= 1\n if key == 0:\n self.head = self.__step_by_step_on_nodes(key)\n self.head = self.head.next\n self._len -= 1\n if key == self._len - 1:\n prev_node = self.__step_by_step_on_nodes(key - 1)\n delete_node = prev_node.next\n delete_node = None\n self.tail = prev_node\n self._len -= 1\n\n def _value_iterator(self):\n \"\"\"\n Node value generator\n \"\"\"\n print(f'The {self._value_iterator.__name__} method of the {self.__class__.__name__} is called')\n current_node = self.head\n for _ in range(self._len):\n yield current_node.value\n current_node = current_node.next\n\n def __iter__(self):\n print(f'The {self.__iter__.__name__} method of the {self.__class__.__name__} is called')\n return self._value_iterator()\n\n def append(self, value: Any):\n # print(f'The {self.append.__name__} method of the {self.__class__.__name__} is called')\n \"\"\"Добавление элемента в конец связного списка\"\"\"\n append_node = self.Node(value)\n if self.head is None:\n self.head = append_node\n else:\n tail = self.head\n for _ in range(self._len - 1):\n tail = tail.next\n self.__linked_nodes(tail, append_node)\n self._len += 1\n\n @staticmethod\n def __linked_nodes(left: Node, right: Optional[Node]) -> None:\n left.next = right\n\n def to_list(self) -> list:\n return [value for value in self]\n\n def insert(self, index: int, value: Any) -> None:\n print(f'{self.insert.__name__} function called')\n if not isinstance(index, int):\n raise TypeError()\n\n if index == 0:\n insert_node = self.Node(value)\n self.__linked_nodes(insert_node, self.head)\n self.head = insert_node\n self._len += 1\n\n elif 0 < index < self._len:\n prev_node = self.__step_by_step_on_nodes(index - 1)\n current_node = prev_node.next\n insert_node = self.Node(value, next_=current_node)\n self.__linked_nodes(prev_node, insert_node)\n self._len += 1\n\n elif index > self._len - 1:\n self.append(value)\n\n def clear(self) -> None:\n self.head = None\n self._len = 0\n\n def index(self, value: Any) -> int:\n print(f'{self.index.__name__} function called')\n current_node = self.head\n for index in range(self._len):\n if current_node.value == value:\n return index\n else:\n current_node = current_node.next\n raise ValueError(f\"{value} is not in {self.__class__.__name__}\")\n\n def remove(self, value: Any) -> None:\n print(f'{self.remove.__name__} function called')\n index = self.index(value)\n if index == 0:\n self.head = self.head.next\n self._len -= 1\n\n elif 0 < index < self._len - 1:\n prev_node = self.__step_by_step_on_nodes(index - 1)\n remove_node = prev_node.next\n next_node = remove_node.next\n self.__linked_nodes(prev_node, next_node)\n self._len -= 1\n\n elif index == self._len - 1:\n self.tail = self.__step_by_step_on_nodes(index - 1)\n self.tail.next = None\n self._len -= 1\n\n def sort(self) -> None:\n flag = True\n iterations = 0\n while flag:\n flag = False\n for i in range(self._len - iterations - 1):\n current_node = self.__step_by_step_on_nodes(i)\n if not isinstance(current_node.value, type(current_node.next.value)):\n raise TypeError(f'Cannot compare {current_node} and {current_node.next}')\n if current_node.value > current_node.next.value:\n current_node.value, current_node.next.value = current_node.next.value, current_node.value\n flag = True\n iterations += 1\n\n @staticmethod\n def _is_iterable(data):\n \"\"\"Метод для проверки является ли объект итерируемым\"\"\"\n if not hasattr(data, '__iter__'):\n raise AttributeError(f'{data.__class__.__name__} is not iterable')\n else:\n return True\n\n def __contains__(self, item: Any):\n return any(item == value for value in self)\n\n def get_node(self, index):\n return self.__step_by_step_on_nodes(index)\n\n\nif __name__ == '__main__':\n ll = LinkedList([1, 2, 3, 4, 5, 6, 7])\n print(ll[-7])\n # for i in range(len(ll)):\n # print(f'На элемент {ll[i]} ссылаются {sys.getrefcount(ll.get_node(i))} ссылки')\n # node = LinkedList.Node(1, 2)\n # next_node = node.next\n # print(next_node)\n # print(node.__repr__())\n\n","repo_name":"artyomzhibarev/double_linked_list_tmp","sub_path":"linked_list_clear.py","file_name":"linked_list_clear.py","file_ext":"py","file_size_in_byte":9572,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28244426499","text":"\"\"\"\n Problem Statement::\n Write a Python program to store second year percentage of students in array. Write function for sorting array of floating point numbers in ascending order using\n a)\tInsertion sort\n b)\tShell Sort and display top five scores\n\n\"\"\"\nclass sort:\n def __init__(self):\n self.list=[]\n self.length=0\n def get_array(self):\n l = int(input(\"Enter the total number of students \"))\n for i in range(l):\n a=float(input(\"Enter the marks of student\"))\n self.list.append(a)\n self.length=l\n\n def disp_array(self):\n if(self.length==0):\n print(\"Empty\")\n else:\n print(self.list)\n\n\n def insertion_sort(self):\n for i in range(1,self.length):\n\n x = self.list[i]\n j = i - 1\n while j > -1 and self.list[j]>x:\n self.list[j + 1] = self.list[j]\n j -= 1\n self.list[j + 1] = x\n\n def shell_sort(self):\n g=self.length//2\n while(g>0):\n for i in range(g,self.length):\n x=self.list[i]\n j=i\n while(j>=g and self.list[j-g]>x):\n self.list[j]=self.list[j-g]\n j=j-g\n self.list[j]=x\n g=g//2\n\n def top(self):\n if(self.length<5):\n print(\"You have to enter marks for atleast 5 students\")\n else:\n print(\"Top 5 percentages are\")\n cnt=0\n j=self.length-1\n while(cnt!=5):\n print(self.list[j])\n j=j-1\n cnt=cnt+1\n\n\n\n \nb=sort()\nk=0\n\n\nprint(\"MENU\")\nprint(\"---------------------------\")\nprint(\"1: Add an array\")\nprint(\"2: Display an array\")\nprint(\"3: Sort the array\")\n\nprint(\"4 Top 5 percentage\")\nprint(\"-1: To exit\")\nprint(\"---------------------------\")\nwhile(True):\n ch=int(input(\"Enter your choice\"))\n if(ch==1):\n b.get_array()\n k=k+1\n print(\"---------------------------\")\n\n elif(ch==2):\n\n b.disp_array()\n print(\"---------------------------\")\n print()\n elif(ch==3):\n print(\"1:Insertion sort \\n 2:Shell sort\")\n ch=int(input(\"Which sort method is to be used\"))\n if(ch==1):\n if(k==0):\n print(\"No marks are added\")\n else:\n b.insertion_sort()\n print(\"Marks are sorted\")\n b.disp_array()\n elif(ch==2):\n if(k==0):\n print(\"No marks are added\")\n else:\n print(\"Marks are sorted\")\n b.shell_sort()\n b.disp_array()\n else:\n print(\"Wrong choice Please enter valid choice\")\n\n\n print(\"---------------------------\")\n\n\n\n\n\n elif(ch==4):\n b.top()\n print(\"---------------------------\")\n\n elif(ch==-1):\n break\n\n\n else:\n print(\"wrong choice !!\")\n print()\n","repo_name":"vedant-kulk/Second-Year-SPPU-Assignments","sub_path":"Fundamental Of Data Structure(FDS)/Assignment_5 (Shell and Insertion Sort).py","file_name":"Assignment_5 (Shell and Insertion Sort).py","file_ext":"py","file_size_in_byte":2991,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"70810972097","text":"def inputFloat(outputText):\n\tvalid = False\n\twhile valid == False:\n\t\ttry:\n\t\t\treturnValue = float(input(outputText))\n\t\t\tvalid = True\n\t\texcept:\n\t\t\tprint(\"Invalid Input.\")\n\t\t\tvalid = False\n\treturn returnValue\n\ndef calcRepayments(loanAmount2, monthInterest2, numPayments2):\n\treturnResult = monthInterest2 * loanAmount2 / 1 - (1 + monthInterest2) * numPayments2\n\treturn returnResult\n\nwhile __name__ == \"__main__\":\n\tloanAmount = inputFloat(\"Please enter the Loan Amount.\\n\")\n\tmonthInterest = inputFloat(\"Please enter the Monthly rate of interest.\\n\")\n\tnumPayments = inputFloat(\"Please enter the Number of Payments.\\n\")\n\tresult = calcRepayments(loanAmount,monthInterest,numPayments)\n\tprint(str.format(\"The result is:{:.2f}\",result))","repo_name":"TheReal3rd/CollegePython","sub_path":"LoanRepaymentCalculator.py","file_name":"LoanRepaymentCalculator.py","file_ext":"py","file_size_in_byte":724,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"70548916094","text":"# USAGE\n# python detect_mask_webcam.py\n\n# import the necessary packages\nfrom tensorflow.keras.applications.mobilenet_v2 import preprocess_input\n#from tensorflow.keras.preprocessing.image import img_to_array\nfrom tensorflow.keras.models import load_model\nimport numpy as np\n#import argparse\nimport cv2\n#import os\n\ndef mask_webcam():\n # facenet : find face model\n facenet = cv2.dnn.readNet('models/deploy.prototxt', 'models/res10_300x300_ssd_iter_140000.caffemodel')\n # model : model of detection mask\n model = load_model('models/realmodel.h5')\n\n # Reading webcam\n cap = cv2.VideoCapture(0)\n\n # Reading webcam\n #cap = cv2.VideoCapture(0)\n i = 0\n\n #url=\"https://192.168.43.1:8080\" //ipWebcam url\n #cap=cv2.VideoCapture(url+\"/video\")\n\n while cap.isOpened():\n ret, img = cap.read()\n if not ret:\n break\n\n h, w = img.shape[:2]\n\n # 이미지 전처리\n # ref. https://www.pyimagesearch.com/2017/11/06/deep-learning-opencvs-blobfromimage-works/\n blob = cv2.dnn.blobFromImage(img, scalefactor=1., size=(224, 224), mean=(104., 177., 123.))\n\n facenet.setInput(blob)\n dets = facenet.forward()\n\n result_img = img.copy()\n\n for i in range(dets.shape[2]):\n\n confidence = dets[0, 0, i, 2]\n\n if confidence < 0.5:\n continue\n\n x1 = int(dets[0, 0, i, 3] * w)\n y1 = int(dets[0, 0, i, 4] * h)\n x2 = int(dets[0, 0, i, 5] * w)\n y2 = int(dets[0, 0, i, 6] * h)\n\n face = img[y1:y2, x1:x2]\n\n while True:\n try:\n face_input = cv2.resize(face, dsize=(224, 224))\n face_input = cv2.cvtColor(face_input, cv2.COLOR_BGR2RGB)\n face_input = preprocess_input(face_input)\n face_input = np.expand_dims(face_input, axis=0)\n break\n except:\n print(\"resize error\")\n break\n\n (hmask,mask, nomask) = model.predict(face_input).squeeze()\n\n if mask > nomask and mask>hmask:\n color = (0, 255, 0)\n label = 'Mask %d%%' % (mask * 100)\n\n elif hmask>mask and hmask>nomask:\n color = (255, 0, 0)\n label = 'Haf Mask %d%%' % (hmask * 100)\n\n elif nomask>hmask:\n color=(0,0,255)\n label='No Mask %d%%' %(nomask*100)\n\n cv2.rectangle(result_img, pt1=(x1, y1), pt2=(x2, y2), thickness=2, color=color, lineType=cv2.LINE_AA)\n cv2.putText(result_img, text=label, org=(x1, y1 - 10), fontFace=cv2.FONT_HERSHEY_SIMPLEX, fontScale=0.8,color=color, thickness=2, lineType=cv2.LINE_AA)\n\n cv2.imshow('Mask Detection',result_img)\n\n\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n\nif __name__ == \"__main__\":\n\tmask_webcam()\n","repo_name":"hipstermartin/Face-Mask-Detection","sub_path":"detect_mask_webcam.py","file_name":"detect_mask_webcam.py","file_ext":"py","file_size_in_byte":2904,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"21940319948","text":"#! /usr/bin/env python\nimport argparse\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description='Parse flagstat files and print summary')\n parser.add_argument('flagstat_files', type=argparse.FileType('r'), nargs='*', help='Flagstat file')\n arguments = parser.parse_args()\n\n counts = {\n 'total': 0,\n 'mapped': 0,\n 'paired': 0,\n 'properly_paired': 0,\n 'dups': 0,\n 'files': len(arguments.flagstat_files)\n }\n\n for flagstat_file in arguments.flagstat_files:\n print(\"working on {0}...\".format(flagstat_file.name))\n\n for line in flagstat_file:\n line = line.strip()\n print(\"\\t{0}\".format(line))\n\n line_count = int(line.split()[0])\n\n if 'in total' in line:\n counts['total'] += line_count\n\n elif 'paired in sequencing' in line:\n counts['paired'] += line_count\n\n elif 'properly paired' in line:\n counts['properly_paired'] += line_count\n\n elif 'duplicates' in line:\n sample_dups = float(line_count)\n counts['dups'] += line_count\n\n elif 'mapped (' in line:\n sample_mapped = float(line_count)\n counts['mapped'] += line_count\n\n print(\"\\n\\t{0} %duplication\\n\".format(100*sample_dups/sample_mapped))\n\n print(\"Total raw reads: {total:,} reads (Total throughput, 75bp={total_75bp:,} bp, 100bp={total_100bp:,} bp, 150bp={total_150bp:,} bp)\".format(\n total=counts['total'], total_75bp=counts['total']*75, total_100bp=counts['total']*100, total_150bp=counts['total']*150\n ))\n print(\"Total mapped reads: {total:,} reads (Total throughput, 75bp={total_75bp:,} bp, 100bp={total_100bp:,} bp, 150bp={total_150bp:,} bp)\".format(\n total=counts['mapped'], total_75bp=counts['mapped']*75, total_100bp=counts['mapped']*100, total_150bp=counts['mapped']*150\n ))\n print(\"Average mapped per lib: {:,} reads\".format(int(round(float(counts['mapped'])/float(counts['files'])))))\n print(\"Average dups per lib: {:,} reads\".format(int(round(float(counts['dups'])/float(counts['files'])))))\n print(\"Average dups % per lib: {:.2f} %\".format(100*float(counts['dups'])/float(counts['mapped'])))\n","repo_name":"UMCUGenetics/CustomModules","sub_path":"Utils/get_stats_from_flagstat.py","file_name":"get_stats_from_flagstat.py","file_ext":"py","file_size_in_byte":2276,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28977290005","text":"import collections\nfrom typing import List\n\n\ndef findAnagrams(s: str, p: str) -> List[int]:\n need = collections.Counter(p)\n window = collections.defaultdict(int)\n\n left = 0\n right = 0\n valid = 0\n res = []\n\n while right < len(s):\n c = s[right]\n right += 1\n\n if need[c] > 0:\n window[c] += 1\n if window[c] == need[c]:\n valid += 1\n\n while right - left >= len(p):\n if valid == len(need):\n res.append(left)\n pop_c = s[left]\n left += 1\n\n if need[pop_c] > 0:\n if window[pop_c] == need[pop_c]:\n valid -= 1\n window[pop_c] -= 1\n\n return res\n\n\ns = \"cbaebabacd\"\np = \"abc\"\nprint(findAnagrams(s, p))\n\n","repo_name":"hzn666/leetcode","sub_path":"438.py","file_name":"438.py","file_ext":"py","file_size_in_byte":783,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"1379361435","text":"from turtle import Turtle,Screen\r\nimport random\r\nimport time\r\nfrom snake import Snake\r\nimport food\r\nfrom scoreboard import Scoreboard\r\n\r\nt = Turtle()\r\nt.penup()\r\nscreen = Screen()\r\nscreen.setup()\r\nscreen.bgcolor(\"black\")\r\nscreen.title(\"My Snake Game\")\r\nfood = food.Food()\r\n\r\nsnake = Snake()\r\nscreen.listen()\r\nscreen.onkey(snake.up,\"Up\")\r\nscreen.onkey(snake.down,\"Down\")\r\nscreen.onkey(snake.left,\"Left\")\r\nscreen.onkey(snake.right,\"Right\")\r\n\r\nscreen.tracer(0)\r\nis_on = True\r\nscore = Scoreboard()\r\nscore.score()\r\n\r\n\r\nwhile is_on:\r\n time.sleep(0.1)\r\n screen.update()\r\n snake.move()\r\n\r\n if snake.head.distance(food) < 15:\r\n food.refresh()\r\n snake.extend()\r\n score.increase_score()\r\n\r\n if snake.head.xcor()>340 or snake.head.xcor()<-340 or snake.head.ycor()>280 or snake.head.ycor()<-280:\r\n score.reset()\r\n # t.color(\"white\")\r\n # t.write(\"Game Over\",align='center',font=('courier',28,'normal'))\r\n\r\n snake.reset()\r\n for i in snake.new_turtle[1:]:\r\n if snake.head.distance(i) < 10:\r\n score.reset()\r\n # is_on = False\r\n # t.color(\"white\")\r\n # t.write(\"Game Over\", align='center', font=('courier', 28, 'normal'))\r\n\r\n\r\nscreen.exitonclick()\r\n","repo_name":"github-maddy/Snake-game-opp","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1249,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"36829504918","text":"n = int(input())\n\ncount=0\nh, m, s = 0, 0, 0\n\nwhile True:\n if h==n and m==59 and s==59:\n break\n s+=1\n if s==60:\n s-=60\n m+=1\n if m==60:\n m-=60\n h+=1\n if \"3\" in str(h)+str(m)+str(s):\n count+=1\nprint(count)\n ","repo_name":"mini-apple/algorithm-level1","sub_path":"2.구현/2-0.py","file_name":"2-0.py","file_ext":"py","file_size_in_byte":229,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"39701981230","text":"# -*- coding: utf-8 -*-\n# @Time : 2020/8/6 16:45\n# @Author : yang\n# @File : book.py\nimport json\n\nfrom flask import request, render_template, url_for, flash\nfrom flask_login import current_user\n\nfrom app.libs.helper import is_isbn_or_key\nfrom app.spider.yushu_book import YuShuBook\nfrom . import web\nfrom ..forms.book import SearchForm\nfrom ..models.gift import Gift\nfrom ..models.wish import Wish\nfrom ..view_models.book import BookViewModel, BookCollection\nfrom ..view_models.trade import TradeInfo\n\n\n@web.route('/book/search')\ndef search():\n \"\"\"\n :param q: 普通关键字isbn\n :param page: 查询页数\n :return:\n \"\"\"\n form = SearchForm(request.args)\n books = BookCollection()\n\n if form.validate():\n q,page = form.q.data.strip(), form.page.data\n # 判断是否是isbn搜索\n isbn_or_key = is_isbn_or_key(q)\n yushu_book = YuShuBook()\n\n if isbn_or_key == 'isbn':\n yushu_book.search_by_isbn(q)\n else:\n yushu_book.search_by_keyword(q,page=page)\n\n books.fill(yushu_book,q)\n else:\n flash('根据关键字搜索不到,请重新输入')\n return render_template('search_result.html', books=books)\n # return form.errors\n\n# 兩個相同 endpoint 導致 viewfunc map overwrite錯誤\n# @web.route('/book//detail')\n# def book_detail(isbn):\n# yushu_book = YuShuBook()\n# yushu_book.search_by_isbn(isbn)\n# book = BookViewModel(yushu_book.first)\n# return render_template('book_detail.html', book = book, wishes = [], gifts = [])\n# pass\n\n@web.route('/book/htmlstudy')\ndef html():\n d1 = {\n 'age':17,\n 'name':'zhangsan'\n }\n d2 = {\n 'gender':'nan',\n 'phone':'135****8642'\n }\n flash('nihao,8yue')\n flash('here is a trouble', category='error')\n return render_template('test.html', d1=d1, d2=d2)\n\n@web.route('/book//detail')\ndef book_detail(isbn):\n \"\"\"\n 1. 当书籍既不在心愿清单也不在礼物清单时,显示礼物清单\n 2. 当书籍在心愿清单时,显示礼物清单\n 3. 当书籍在礼物清单时,显示心愿清单\n 4. 一本书要防止即在礼物清单,又在赠送清单,这种情况是不符合逻辑的\n\n 这个视图函数不可以直接用cache缓存,因为不同的用户看到的视图不一样\n 优化是一个逐步迭代的过程,建议在优化的初期,只缓存那���和用户无关的“公共数据\"\n \"\"\"\n has_in_gifts = False\n has_in_wishes = False\n # isbn_or_key = is_isbn_or_key(isbn)\n # if isbn_or_key == 'isbn':\n # 获取图书信息\n yushu_book = YuShuBook()\n yushu_book.search_by_isbn(isbn)\n\n if current_user.is_authenticated:\n # 如果未登录,current_user将是一个匿名用户对象\n if Gift.query.filter_by(uid=current_user.id, isbn=isbn,\n launched=False).first():\n has_in_gifts = True\n if Wish.query.filter_by(uid=current_user.id, isbn=isbn,\n launched=False).first():\n has_in_wishes = True\n\n book = BookViewModel(yushu_book.first)\n # if has_in_gifts:\n trade_wishes = Wish.query.filter_by(isbn=isbn, launched=False).all()\n trade_gifts = Gift.query.filter_by(isbn=isbn, launched=False).all()\n trade_wishes_model = TradeInfo(trade_wishes)\n trade_gifts_model = TradeInfo(trade_gifts)\n return render_template('book_detail.html', book=book, has_in_gifts=has_in_gifts,\n has_in_wishes=has_in_wishes,\n wishes=trade_wishes_model,\n gifts=trade_gifts_model)","repo_name":"paststory/fisher","sub_path":"app/web/book.py","file_name":"book.py","file_ext":"py","file_size_in_byte":3673,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"28993803915","text":"from time import sleep\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.support.select import Select\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.chrome.service import Service\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.common.exceptions import NoSuchElementException, TimeoutException, StaleElementReferenceException\nfrom selenium.webdriver.chrome.options import Options\nfrom database import Database\nfrom urllib.parse import unquote\nfrom urllib.request import urlretrieve\nimport re\nimport json\nimport os\nimport getopt\nimport sys\nfrom tqdm import tqdm\n\npbar = None\n\ndef cbk(block_num, block_size, total_size): \n '''''Callback function \n @a:Downloaded data block \n @b:Block size \n @c:Size of the remote file \n ''' \n global pbar\n if pbar is None:\n pbar = tqdm(total=100)\n downloaded = block_num * block_size\n if downloaded < total_size:\n pbar.n = 100.0*downloaded/total_size\n pbar.refresh()\n else:\n pbar.close()\n pbar = None\n\ndef open_link_in_tab(driver, link):\n actions = ActionChains(driver)\n actions.key_down(Keys.CONTROL)\n actions.click(link)\n actions.perform()\n\ndef scrap_page(driver):\n db = Database('newnaratif.db', 'newnaratif')\n if os.path.exists(\"newnaratlif\") is False:\n os.makedirs(\"newnaratlif\")\n\n url = driver.current_url\n match = re.search(r\".com/.*/$\", url)\n filename = match.group()[5:-1]\n\n try:\n date = driver.find_element(By.CSS_SELECTOR, \"time[class='entry-date published']\")\n title = driver.find_element(By.CSS_SELECTOR, \"header[class='entry-header']\").find_element(By.TAG_NAME, 'h1')\n main = driver.find_element(By.ID, \"main\")\n category = main.find_element(By.TAG_NAME, \"header\").find_element(By.TAG_NAME, \"a\")\n if category.text.lower() == \"podcast\":\n iframes = WebDriverWait(driver, 30).until(EC.presence_of_all_elements_located((By.TAG_NAME, 'iframe')))\n iframe = iframes[-1]\n driver.execute_script(\"arguments[0].scrollIntoView();\", iframe)\n driver.switch_to.frame(iframe)\n sleep(1)\n WebDriverWait(driver, 15).until(EC.presence_of_all_elements_located((By.TAG_NAME, 'button')))\n audio = driver.find_element(By.TAG_NAME, 'audio')\n match = re.search(r\"/https.*cloudfront.*mp3$\", audio.get_attribute(\"src\"))\n audio_link = unquote(match.group())[1:]\n driver.switch_to.default_content()\n urlretrieve(audio_link, f\"newnaratlif\\\\{filename}.mp3\", cbk)\n db.save_record(date.text, title.text, url, f\"newnaratlif\\\\{filename}.mp3\")\n else:\n driver.execute_script(\"\"\"\n document.querySelectorAll(\"blockquote\").forEach((element) => {\n element.remove();\n });\n document.querySelector(\"div[class='entry-byline']\")?.remove();\n \"\"\")\n data = main.find_element(By.CSS_SELECTOR, \"div[class='entry-content']\")\n with open(f\"newnaratlif\\\\{filename}.txt\", \"w\") as f:\n f.write(json.dumps({\n 'title': title.text,\n 'date': date.text,\n 'content': data.text,\n }))\n db.save_record(date.text, title.text, url, f\"newnaratlif\\\\{filename}.json\")\n except (StaleElementReferenceException, NoSuchElementException, IndexError, AttributeError) as e:\n with open(\"errors.log\", \"a\", encoding='utf-8') as f:\n f.write(f\"Error: {url} {str(e)}\\n\")\n\ndef scrap_results(driver, url): \n driver.get(url)\n\n articles = driver.find_element(By.ID, \"main\").find_elements(By.TAG_NAME, \"article\")\n for article in articles:\n link = article.find_element(By.CSS_SELECTOR, 'h2 > a')\n driver.execute_script(\"arguments[0].scrollIntoView();\", link)\n url = link.get_attribute(\"href\")\n open_link_in_tab(driver, link)\n sleep(1)\n\n *_, article_window = driver.window_handles\n driver.switch_to.window(article_window)\n sleep(1)\n\n scrap_page(driver)\n\n home, *_ = driver.window_handles\n driver.close()\n driver.switch_to.window(home)\n \ndef main():\n from_page = None\n to_page = None\n page = None\n try:\n opts, args = getopt.getopt(sys.argv[1:], \"f:t:p:\")\n for opt, arg in opts:\n if opt == '-f':\n from_page = int(arg)\n elif opt == '-t':\n to_page = int(arg)\n elif opt == '-p':\n page = arg\n except getopt.GetoptError as err:\n print(err)\n quit()\n\n driver = webdriver.Chrome(service=Service(\"chromedriver.exe\"))\n\n if page is None:\n base_url = \"https://newnaratif.com\"\n for i in range(from_page, to_page+1):\n url = f\"{base_url}/page/{i}/?s=Singapore\"\n scrap_results(driver, url)\n else:\n driver.get(page)\n scrap_page(driver)\n\n driver.close()\n\nif __name__ == \"__main__\":\n main()","repo_name":"ktkhuong/sgparl","sub_path":"src/data-collection/newnaratif.py","file_name":"newnaratif.py","file_ext":"py","file_size_in_byte":5251,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"4562766782","text":"# Scrape the government data from CRS page\nimport json\nimport requests\nimport pandas as pd\nfrom collections import deque\nfrom bs4 import BeautifulSoup as bs\nimport calendar\nfrom collections import deque, defaultdict\n\n# Checking if the website response correctly\n\nabbr_to_num = {name: num for num, name in enumerate(calendar.month_abbr) if num}\ndef run():\n scraped_data = []\n for i in range(1,10):\n url = \"https://www.uiis.ca/category/express-entry/page/\" + str(i) + \"/\"\n print(url)\n with requests.session() as s:\n headers = {'User-Agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.131 Safari/537.36'}\n res = s.get(url, headers = headers)\n if res.status_code == requests.codes.ok: # if response with 200\n soup = bs(res.content,'html.parser')\n table_data = [next.text for next in soup.find_all(['td','th'])]\n meta_data_for_table = [next.text for next in soup.find_all('p')]\n crs_score = [next.text for next in soup.find_all(lambda tag: tag.name in ['h2','p'] and \"CRS score\" in tag.text)]\n invitations = [next.text for next in soup.find_all(lambda tag: tag.name in ['h2','h4','h5','p'] and \"Number of invitation\" in tag.text)]\n # Filtering date and times\n first_filter_meta_data = [i for i in meta_data_for_table if 'Date and time of round:' in i]\n first_filter_meta_data = [i.replace('\\xa0',' ') for i in first_filter_meta_data]\n first_filter_meta_data = [i.replace('Date and time of round: ','') for i in first_filter_meta_data]\n first_filter_meta_data = [i[:-16] if 'UTC' in i else i for i in first_filter_meta_data]\n first_filter_meta_data = [i.strip() for i in first_filter_meta_data]\n # Filtering number of invitations\n invitations = [i.replace('

','').replace('','').replace('

','') for i in invitations]\n invitations = [i.replace('

','').replace('','').replace('

','') for i in invitations]\n invitations = [i.replace('','').replace('\\xa0','') for i in invitations]\n invitations = [i.replace('','') for i in invitations]\n invitations = [i.replace('Number of invitations issued:','').replace(',','').strip() for i in invitations]\n invitations = [i.replace('Footnote*','') for i in invitations]\n \n # Filtering crs scores\n crs_score = [i.replace('CRS score of lowest-ranked candidate invited:\\xa0','').strip() for i in crs_score]\n crs_score = [i.replace('CRS score of lowest-ranked candidate invited:','').strip() for i in crs_score]\n table_data = [i for i in table_data if 'crs score range' not in i.lower()]\n table_data = [i for i in table_data if 'number of candidates' not in i.lower()]\n table_data = [i for i in table_data if 'Classes' not in i]\n table_data = [i for i in table_data if 'Number of Invitations' not in i]\n # table data preprocessing (edge cases) - page 6\n if i == 6:\n edge_cases = [\n '(Federal Skilled Worker)',\n '(Canadian Experience)',\n '(Federal Skilled Trades)',\n '(Provincial nominee)',\n 'Class referred to inparagraph 2(a) of the Express Entry Instructions (Federal Skilled Worker)',\n 'Class referred to inparagraph 2(b) of the Express Entry Instructions (Canadian Experience)',\n 'Class referred to inparagraph 2(c) of the Express Entry Instructions (Federal Skilled Trades)',\n 'Class referred to inparagraph 2(d) of the Express Entry Instructions (Provincial nominee)'\n ]\n idxes = []\n for j in edge_cases:\n idx = table_data.index(j)\n idxes.append(idx)\n idxes.append(idx + 1)\n new_table_data = []\n for x in range(len(table_data)):\n if x not in idxes:\n new_table_data.append(table_data[x])\n table_data = new_table_data\n\n if i == 9:\n idx_not_to_include = [i for i in range(34,84)]\n new_table_data = []\n for x in range(len(table_data)):\n if x not in idx_not_to_include:\n new_table_data.append(table_data[x])\n table_data = new_table_data\n first_filter_meta_data = first_filter_meta_data[:-1]\n invitations = invitations[:-1]\n\n score_dict = defaultdict(list)\n table_queue = deque(table_data)\n while table_queue:\n sc_range = table_queue.popleft()\n score_dict[sc_range].append(table_queue.popleft())\n\n scraped_data.append([first_filter_meta_data, invitations, crs_score, score_dict])\n else:\n print('error 404')\n\n _dates = []\n _invitations = []\n _crs_scores = []\n _score_dist = []\n\n for i in range(0,9):\n _dates.append(scraped_data[i][0])\n _invitations.append(scraped_data[i][1])\n _crs_scores.append(scraped_data[i][2])\n _score_dist.append(scraped_data[i][3])\n\n # _dates = [i for i in _dates for i in i]\n # _invitations = [i for i in _invitations for i in i]\n # _crs_scores = [i for i in _crs_scores for i in i]\n\n return _dates, _invitations, _crs_scores, _score_dist\n \nif __name__ == \"__main__\":\n _dates, _invitations, _crs_scores, _score_dist = run()\n","repo_name":"pkang0831/CRS_main","sub_path":"dist_scraper.py","file_name":"dist_scraper.py","file_ext":"py","file_size_in_byte":5987,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"12402144844","text":"#!/bin/python\nimport matplotlib.pyplot as plt\nimport csv\nimport os\nimport math\nimport argparse\n\ncolor = ['b', 'g', 'r', 'y', 'm', 'c', 'k', 'w']\n\n\ndef export_csv_file(file, filename):\n with open(f\"{filename}_T.csv\", 'w') as f:\n csv.writer(f).writerows(file)\n\ndef read_csv_file(filename):\n file = []\n with open(filename,'r') as csvfile:\n plots = csv.reader(csvfile, delimiter = ',')\n for row in plots:\n file.append(row)\n return file\n\ndef clean_csv_matrix(file):\n max_row_len = len(file[0])\n for i,r in enumerate(file):\n for j,c in enumerate(file[i]):\n if c == '':\n file[i][j] = 0\n try:\n # file[i][j] = round(float(c), 4)\n file[i][j] = float(c)\n except:\n pass\n\n if len(r) < max_row_len:\n file[i].append(0)\n return file\n\ndef transpose_csv_matrix(file):\n zipped_rows = zip(*file)\n return [list(row) for row in zipped_rows]\n\ndef print_csv_matrix(matrix):\n for r in matrix:\n print(*r)\n\ndef plot_to_png(file, filename):\n x = file[args.main][1:len(file[args.main])]\n l_y = []\n yn = []\n y_max = 0\n y_min = 0\n for i, line in enumerate(file):\n if i != args.main and i not in args.ignore:\n if args.plot == [] or i in args.plot:\n if -1 in args.ignore and i == len(file) - 1:\n break\n l_y.append(file[i][0])\n #yn are all lines tha will be ploted on oy\n yn.append(file[i][1:len(file[i])])\n if y_max < max(yn[len(yn)-1]):\n y_max = max(yn[len(yn)-1])\n if y_min < min(yn[len(yn)-1]):\n y_min = min(yn[len(yn)-1])\n\n y_max = math.ceil(y_max)\n y_min = math.floor(y_min)\n\n # output size\n plt.figure(figsize=(args.width, args.height))\n\n # OX, OY axes\n plt.axhline(0, color='#696969')\n plt.axvline(0, color='#696969')\n\n for i, y in enumerate(yn):\n plt.plot(x, y, color = color[i], linestyle = 'solid', marker = 'o',label = l_y[i])\n\n plt.grid()\n plt.xlabel(file[0][0] if args.x_label == 'ox' else args.x_label)\n plt.ylabel(args.y_label)\n plt.title(args.title)\n plt.legend()\n\n if args.verbose:\n plt.show()\n else:\n plt.savefig(f\"{args.output_dir}/{filename}.png\", dpi=100)\n\n\ndef plotting_stuff():\n if args.filename:\n fileNames = [args.filename]\n args.output_dir = '.' if args.output_dir == 'png' else args.output_dir\n args.input_dir = './'\n else:\n fileNames = os.listdir(args.input_dir)\n fileNames = [file for file in fileNames if '.csv' in file and '#' not in file]\n\n if args.output_dir != '.':\n args.output_dir = args.input_dir + '_t' if args.transpose else args.output_dir\n if not args.verbose:\n try:\n os.mkdir(args.output_dir)\n except:\n pass\n\n for filename in fileNames:\n print(filename, \"was loaded\")\n if args.filename:\n file = read_csv_file(args.input_dir + filename)\n else:\n file = read_csv_file(args.input_dir + \"/\" + filename)\n file = clean_csv_matrix(file)\n\n if args.debug:\n print_csv_matrix(file)\n \n try:\n float(file[0][len(file[0])-1])\n except:\n file = transpose_csv_matrix(file)\n if args.debug:\n print(\"Transposed csv file\")\n print_csv_matrix(file)\n print(\"File had to be transposed\")\n\n if args.transpose:\n file = transpose_csv_matrix(file)\n export_csv_file(file, args.output_dir + '/' + filename[0:len(filename)-4])\n continue\n\n if args.filename:\n filename = os.path.basename(args.filename)\n args.filename = filename\n plot_to_png(file, filename[0:len(filename)-4])\n print(filename, \"SUCCESSFULLY plotted to\",\n f\"{args.output_dir}/{filename}.png\")\n\n\n### for terminal arguments\nparser = argparse.ArgumentParser(description=\"\"\"\n Program to plot csv files - made by Catalin\n first row will be used for the OX axis\n and the following rows will be used for the OY axis\n Is important for information to be stored in rows!!!\n Use -t to transpose the file\n To be noted first column will be used as labels\n \"\"\")\n\nparser.add_argument('-v', action='store_true',\n help='verbose', dest='verbose')\n\nparser.add_argument('-d', action='store_true',\n help='debug; output csv file content if something went wrong', dest='debug')\n\nparser.add_argument('--transpose', action='store_true',\n help=\"\"\"transpose csv file, wont plot anything will create a new \n folder with transposed csv content\"\"\", dest='transpose')\n\nparser.add_argument('-f', help='plot only one file',\n dest='filename', default=None)\n\nparser.add_argument(\"-i\", dest=\"input_dir\", default='csv',\n help=\"input directory, default ./csv\", metavar=\"DIR\")\n\nparser.add_argument(\"-o\", dest=\"output_dir\", default='png',\n help=\"output directory, default ./png\", metavar=\"DIR\")\n\nparser.add_argument(\"-x\", dest=\"x_label\", default='ox',\n help=\"x label\", metavar=\"'str'\")\n\nparser.add_argument(\"-y\", dest=\"y_label\", default='oy',\n help=\"y label\", metavar=\"'str'\")\n\nparser.add_argument(\"-t\", dest=\"title\", default='Title',\n help=\"title\", metavar=\"'str'\")\n\nparser.add_argument(\"--main\", type=int, default=0,\n help=\"\"\"select which of the rows or columns should be used as x axis, 0\n will be the first index\"\"\")\n\nparser.add_argument(\"--ignore\", type=int, nargs=\"+\", default=[],\n help=\"\"\"select which of the rows or columns shouldn't be used, 0\n will be the first index, -1 will ignore the last element\"\"\")\n\nparser.add_argument(\"--plot\", dest=\"plot\", nargs=\"+\", type=int, default=[],\n help=\"\"\"choose what columns or rows to use for plotting y axis, 0\n will be the first index\"\"\")\n\nparser.add_argument(\"--height\", type=int, default=6, dest='height',\n help=\"\"\"for aspect ratio default 6\"\"\")\n\nparser.add_argument(\"--width\", type=int, default=6, dest='width',\n help=\"\"\"for aspect ratio default 6\"\"\")\n\nif __name__ == \"__main__\":\n args = parser.parse_args()\n if args.debug:\n print(\"args:\", args)\n plotting_stuff()\n\n","repo_name":"CatalinPlesu/plot_csv","sub_path":"plot_csv.py","file_name":"plot_csv.py","file_ext":"py","file_size_in_byte":6410,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"24410937158","text":"# -*- coding: utf-8 -*-\n\nfrom typing import Dict, Iterable\n\nimport tensorflow as tf\n\nfrom .. import utils\nfrom ...data import Field\nfrom ...layers.text_embedders import TextEmbedder\nfrom ...layers.seq2vec_encoders import CNNEncoder\n\n\ndef TextCNNClassifier(features: Dict[str, Field],\n targets: Dict[str, Field],\n text_embedder: TextEmbedder,\n filters: int = 100,\n kernel_sizes: Iterable[int] = (2, 3, 4, 5),\n conv_layer_activation='relu',\n l1_regularization: float = None,\n l2_regularization: float = None,\n dropout: float = 0.1,\n activation='softmax',\n label_field: str = 'label'):\n inputs = utils.create_inputs(features)\n input_tokens = utils.get_text_inputs(inputs, 'tokens')\n embedded_tokens = text_embedder(input_tokens)\n cnn_encoder = CNNEncoder(filters, kernel_sizes,\n conv_layer_activation,\n l1_regularization,\n l2_regularization)\n encoded_tokens = cnn_encoder(embedded_tokens)\n if dropout:\n encoded_tokens = tf.keras.layers.Dropout(dropout)(encoded_tokens)\n probs = tf.keras.layers.Dense(\n len(targets[label_field].vocab), activation=activation,\n name=label_field)(encoded_tokens)\n return tf.keras.models.Model(inputs=list(inputs.values()),\n outputs=probs,\n name='TextCNNClassifier')\n","repo_name":"linhx13/arc-nlp","sub_path":"arcnlp/tf/models/text_classification/text_cnn.py","file_name":"text_cnn.py","file_ext":"py","file_size_in_byte":1596,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"18717681214","text":"import numpy as np\nimport matplotlib.pyplot as plt\nimport GTM.GTMcore as GTM\nimport GTM.Permittivities as mat\n\n#%%\nc_const = 299792458.0 # m/s\n\n## create a void system: substrate=superstrate=vac\nS = GTM.System()\n# define the prism, airgap and SiC6H layers\nKRS5 = GTM.Layer(thickness = 1e-6, epsilon1=mat.eps_KRS5) # epsilon2 and epsilon3 default to epsilon1\nAirGap = GTM.Layer(thickness = 5.5e-6) # no epsilon default to vacuum\nSiC6H = GTM.Layer(thickness = 3e-6, epsilon1=mat.eps_SiC6Hx, # epsilon2 defaults to epsilon1\n epsilon3 = mat.eps_SiC6Hz)\n\n# start setting up the system\nS.set_superstrate(KRS5)\nS.set_substrate(SiC6H)\nS.add_layer(AirGap)\n\n# angle of incidence\nthetain = np.deg2rad(30.0)\n\nwnplot = np.arange(750.0, 1050.0, dtype=np.float) # wavenumber range of the plot in cm-1\nfplot = wnplot*c_const*1e2 # corresponding frequency\ndz = 100e-9 # spatial resolution\n\n\nRplot = np.zeros(len(wnplot))\n\nEx = []\n\nfor ii, fi in enumerate(fplot):\n S.initialize_sys(fi)\n zeta_sys = np.sin(thetain)*np.sqrt(S.superstrate.epsilon[0,0])\n Sys_Gamma = S.calculate_GammaStar(fi, zeta_sys)\n r, R, t, T = S.calculate_r_t(zeta_sys)\n zplot, E_out, zn_plot = S.calculate_Efield(fi, zeta_sys, dz=dz)\n Rplot[ii] = R[0]\n Ex.append(E_out[0,:])\n\n#%% Plot the results\nExm = np.asarray(Ex) # make a 2D array from the electric field list\n\nwnm, zm = np.meshgrid(wnplot,zplot*1e6)\n\nfig2ab = plt.figure(figsize=(10,4))\naxR = fig2ab.add_subplot(121) \naxR.plot(wnplot,Rplot, '+-')\naxR.set_ylim([0,1.05])\naxR.set_xlabel('wavenumber (cm$^{-1}$)')\naxR.set_ylabel('Reflectivity')\n\naxfield = fig2ab.add_subplot(122)\naxc = axfield.pcolormesh(wnm, zm, np.abs(Exm.T),\n vmin=0, vmax=5, shading='gouraud', cmap=plt.cm.gnuplot2)\nfor ii, zi in enumerate(zn_plot):\n axfield.plot([wnplot.min(),wnplot.max()],\n [zi*1e6, zi*1e6], '--k')\naxfield.invert_yaxis()\naxfield.set_xlabel('wavenumber (cm$^{-1}$)')\naxfield.set_ylabel('z-position ($\\mu$m)')\naxfield.set_ylim([8.5,0])\n\nfig2ab.colorbar(axc)\nfig2ab.tight_layout()\nfig2ab.show()\n\n\n#%% Add the GaN layer\nS2 = GTM.System()\n# define the new airgap and GaN layers\nAirGap2 = GTM.Layer(thickness = 3.5e-6) # no epsilon default to vacuum\nGaN = GTM.Layer(thickness = 2e-6, epsilon1=mat.eps_GaNx,\n epsilon3 = mat.eps_GaNz)\n# start setting up the system\nS2.set_superstrate(KRS5)\nS2.set_substrate(SiC6H)\nS2.add_layer(AirGap2)\nS2.add_layer(GaN)\n\nwnplot2 = np.arange(550, 1050, dtype=np.float) # wavenumber range of the plot in cm-1\nfplot2 = wnplot2*c_const*1e2 # corresponding frequency\n\nRplot2 = np.zeros(len(wnplot2))\nEx2 = []\nfor ii, fi in enumerate(fplot2):\n S2.initialize_sys(fi)\n zeta_sys = np.sin(thetain)*np.sqrt(S.superstrate.epsilon[0,0])\n Sys_Gamma = S2.calculate_GammaStar(fi, zeta_sys)\n r, R, t, T = S2.calculate_r_t(zeta_sys)\n zplot2, E_out, zn_plot2 = S2.calculate_Efield(fi, zeta_sys, dz=dz)\n Rplot2[ii] = R[0]\n Ex2.append(E_out[0,:])\n \n#%% Plot the results\nExm2 = np.asarray(Ex2) # make a 2D array from the electric field list\n\nwnm2, zm2 = np.meshgrid(wnplot2,zplot2*1e6)\n\nfig3ab = plt.figure(figsize=(10,4))\naxR2 = fig3ab.add_subplot(121) \naxR2.plot(wnplot2,Rplot2, '+-')\naxR2.set_ylim([0,1.05])\naxR2.set_xlabel('wavenumber (cm$^{-1}$)')\naxR2.set_ylabel('Reflectivity')\n\naxfield2 = fig3ab.add_subplot(122)\naxc2 = axfield2.pcolormesh(wnm2, zm2, np.abs(Exm2.T),\n vmin=0, vmax=4, \n shading='gouraud', cmap=plt.cm.gnuplot2)\nfor ii, zi in enumerate(zn_plot2):\n axfield2.plot([wnplot2.min(),wnplot2.max()],\n [zi*1e6, zi*1e6], '--k')\naxfield2.invert_yaxis()\naxfield2.set_xlabel('wavenumber (cm$^{-1}$)')\naxfield2.set_ylabel('z-position ($\\mu$m)')\naxfield2.set_ylim([8.5,0])\n\nfig3ab.colorbar(axc2)\nfig3ab.tight_layout()\nfig3ab.show()\n","repo_name":"pyMatJ/pyGTM","sub_path":"examples/TestScript_Passler4x4.py","file_name":"TestScript_Passler4x4.py","file_ext":"py","file_size_in_byte":3839,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"79"} +{"seq_id":"73471230654","text":"from __future__ import print_function\nimport keras\nfrom keras.datasets import cifar10\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Activation, Flatten\nfrom keras.layers import Conv2D, MaxPooling2D\nfrom keras.optimizers import SGD\nfrom keras.utils import print_summary, to_categorical\nimport sys\nimport os\nimport tensorflow as tf\nfrom tqdm import tqdm\n\nbatch_size = 64\nnum_classes = 10\nepochs = 100\n\n(x_train, y_train), (x_test, y_test) = cifar10.load_data()\n\ny_train = to_categorical(y_train, num_classes)\ny_test = to_categorical(y_test, num_classes)\n\nx_train = x_train.astype('float32')\nx_test = x_test.astype('float32')\nx_train /= 255.0\nx_test /= 255.0\n\nX = tf.placeholder(dtype=tf.float32, shape=(None, x_train.shape[1], x_train.shape[2], 3), name='input')\nY = tf.placeholder(dtype=tf.int32, shape=(None), name='ground_truth')\n\nmodel = Sequential()\nmodel.add(Conv2D(32, (3, 3), padding='same', input_shape=x_train.shape[1:]))\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.3))\n\nmodel.add(Conv2D(64, (3, 3), padding='same', input_shape=x_train.shape[1:]))\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.3))\n\nmodel.add(Conv2D(128, (3, 3), padding='same', input_shape=x_train.shape[1:]))\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.4))\n\nmodel.add(Flatten())\nmodel.add(Dense(80))\nmodel.add(Activation('relu'))\nmodel.add(Dropout(0.3))\nmodel.add(Dense(num_classes))\nmodel.add(Activation('softmax'))\n\n\nopt = SGD(lr=0.01, momentum=0.9, decay=0, nesterov=False)\n\nmodel.compile(loss='categorical_crossentropy',\n optimizer=opt,\n metrics=['accuracy'])\n\nmodel.fit(x_train, y_train,\n batch_size=batch_size,\n epochs=epochs,\n validation_split=0.2,\n shuffle=True)\n","repo_name":"ArnaudAbreu/neuralyzer","sub_path":"tests/shity.py","file_name":"shity.py","file_ext":"py","file_size_in_byte":1882,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"3536940590","text":"from __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\n\nANSIBLE_METADATA = {'metadata_version': '1.1',\n 'status': ['preview'],\n 'supported_by': 'community'}\n\n\nDOCUMENTATION = '''\n---\nmodule: azure_rm_appgwlocalnetworkgateway\nversion_added: \"2.5\"\nshort_description: Manage Local Network Gateway instance.\ndescription:\n - Create, update and delete instance of Local Network Gateway.\n\noptions:\n resource_group:\n description:\n - The name of the resource group.\n required: True\n local_network_gateway_name:\n description:\n - The name of the local network gateway.\n required: True\n id:\n description:\n - Resource ID.\n location:\n description:\n - Resource location. If not set, location from the resource group will be used as default.\n local_network_address_space:\n description:\n - Local network site address space.\n suboptions:\n address_prefixes:\n description:\n - A list of address blocks reserved for this virtual network in CIDR notation.\n type: list\n gateway_ip_address:\n description:\n - IP address of local network gateway.\n bgp_settings:\n description:\n - \"Local network gateway's BGP speaker settings.\"\n suboptions:\n asn:\n description:\n - \"The BGP speaker's ASN.\"\n bgp_peering_address:\n description:\n - The BGP peering address and BGP identifier of this BGP speaker.\n peer_weight:\n description:\n - The weight added to routes learned from this BGP speaker.\n resource_guid:\n description:\n - The resource GUID property of the LocalNetworkGateway resource.\n etag:\n description:\n - A unique read-only string that changes whenever the resource is updated.\n state:\n description:\n - Assert the state of the Local Network Gateway.\n - Use 'present' to create or update an Local Network Gateway and 'absent' to delete it.\n default: present\n choices:\n - absent\n - present\n\nextends_documentation_fragment:\n - azure\n\nauthor:\n - \"Zim Kalinowski (@zikalino)\"\n\n'''\n\nEXAMPLES = '''\n - name: Create (or update) Local Network Gateway\n azure_rm_appgwlocalnetworkgateway:\n resource_group: NOT FOUND\n local_network_gateway_name: NOT FOUND\n location: eastus\n'''\n\nRETURN = '''\nid:\n description:\n - Resource ID.\n returned: always\n type: str\n sample: id\n'''\n\nimport time\nfrom ansible.module_utils.azure_rm_common import AzureRMModuleBase\n\ntry:\n from msrestazure.azure_exceptions import CloudError\n from msrestazure.azure_operation import AzureOperationPoller\n from azure.mgmt.network import NetworkManagementClient\n from msrest.serialization import Model\nexcept ImportError:\n # This is handled in azure_rm_common\n pass\n\n\nclass Actions:\n NoAction, Create, Update, Delete = range(4)\n\n\nclass AzureRMLocalNetworkGateways(AzureRMModuleBase):\n \"\"\"Configuration class for an Azure RM Local Network Gateway resource\"\"\"\n\n def __init__(self):\n self.module_arg_spec = dict(\n resource_group=dict(\n type='str',\n required=True\n ),\n local_network_gateway_name=dict(\n type='str',\n required=True\n ),\n id=dict(\n type='str'\n ),\n location=dict(\n type='str'\n ),\n local_network_address_space=dict(\n type='dict'\n ),\n gateway_ip_address=dict(\n type='str'\n ),\n bgp_settings=dict(\n type='dict'\n ),\n resource_guid=dict(\n type='str'\n ),\n etag=dict(\n type='str'\n ),\n state=dict(\n type='str',\n default='present',\n choices=['present', 'absent']\n )\n )\n\n self.resource_group = None\n self.local_network_gateway_name = None\n self.parameters = dict()\n\n self.results = dict(changed=False)\n self.mgmt_client = None\n self.state = None\n self.to_do = Actions.NoAction\n\n super(AzureRMLocalNetworkGateways, self).__init__(derived_arg_spec=self.module_arg_spec,\n supports_check_mode=True,\n supports_tags=False)\n\n def exec_module(self, **kwargs):\n \"\"\"Main module execution method\"\"\"\n\n for key in list(self.module_arg_spec.keys()):\n if hasattr(self, key):\n setattr(self, key, kwargs[key])\n elif kwargs[key] is not None:\n if key == \"id\":\n self.parameters[\"id\"] = kwargs[key]\n elif key == \"location\":\n self.parameters[\"location\"] = kwargs[key]\n elif key == \"local_network_address_space\":\n self.parameters[\"local_network_address_space\"] = kwargs[key]\n elif key == \"gateway_ip_address\":\n self.parameters[\"gateway_ip_address\"] = kwargs[key]\n elif key == \"bgp_settings\":\n self.parameters[\"bgp_settings\"] = kwargs[key]\n elif key == \"resource_guid\":\n self.parameters[\"resource_guid\"] = kwargs[key]\n elif key == \"etag\":\n self.parameters[\"etag\"] = kwargs[key]\n\n old_response = None\n response = None\n\n self.mgmt_client = self.get_mgmt_svc_client(NetworkManagementClient,\n base_url=self._cloud_environment.endpoints.resource_manager)\n\n resource_group = self.get_resource_group(self.resource_group)\n\n if \"location\" not in self.parameters:\n self.parameters[\"location\"] = resource_group.location\n\n old_response = self.get_localnetworkgateway()\n\n if not old_response:\n self.log(\"Local Network Gateway instance doesn't exist\")\n if self.state == 'absent':\n self.log(\"Old instance didn't exist\")\n else:\n self.to_do = Actions.Create\n else:\n self.log(\"Local Network Gateway instance already exists\")\n if self.state == 'absent':\n self.to_do = Actions.Delete\n elif self.state == 'present':\n self.log(\"Need to check if Local Network Gateway instance has to be deleted or may be updated\")\n self.to_do = Actions.Update\n\n if (self.to_do == Actions.Create) or (self.to_do == Actions.Update):\n self.log(\"Need to Create / Update the Local Network Gateway instance\")\n\n if self.check_mode:\n self.results['changed'] = True\n return self.results\n\n response = self.create_update_localnetworkgateway()\n\n if not old_response:\n self.results['changed'] = True\n else:\n self.results['changed'] = old_response.__ne__(response)\n self.log(\"Creation / Update done\")\n elif self.to_do == Actions.Delete:\n self.log(\"Local Network Gateway instance deleted\")\n self.results['changed'] = True\n\n if self.check_mode:\n return self.results\n\n self.delete_localnetworkgateway()\n # make sure instance is actually deleted, for some Azure resources, instance is hanging around\n # for some time after deletion -- this should be really fixed in Azure\n while self.get_localnetworkgateway():\n time.sleep(20)\n else:\n self.log(\"Local Network Gateway instance unchanged\")\n self.results['changed'] = False\n response = old_response\n\n if response:\n self.results[\"id\"] = response[\"id\"]\n\n return self.results\n\n def create_update_localnetworkgateway(self):\n '''\n Creates or updates Local Network Gateway with the specified configuration.\n\n :return: deserialized Local Network Gateway instance state dictionary\n '''\n self.log(\"Creating / Updating the Local Network Gateway instance {0}\".format(self.local_network_gateway_name))\n\n try:\n response = self.mgmt_client.local_network_gateways.create_or_update(resource_group_name=self.resource_group,\n local_network_gateway_name=self.local_network_gateway_name,\n parameters=self.parameters)\n if isinstance(response, AzureOperationPoller):\n response = self.get_poller_result(response)\n\n except CloudError as exc:\n self.log('Error attempting to create the Local Network Gateway instance.')\n self.fail(\"Error creating the Local Network Gateway instance: {0}\".format(str(exc)))\n return response.as_dict()\n\n def delete_localnetworkgateway(self):\n '''\n Deletes specified Local Network Gateway instance in the specified subscription and resource group.\n\n :return: True\n '''\n self.log(\"Deleting the Local Network Gateway instance {0}\".format(self.local_network_gateway_name))\n try:\n response = self.mgmt_client.local_network_gateways.delete(resource_group_name=self.resource_group,\n local_network_gateway_name=self.local_network_gateway_name)\n except CloudError as e:\n self.log('Error attempting to delete the Local Network Gateway instance.')\n self.fail(\"Error deleting the Local Network Gateway instance: {0}\".format(str(e)))\n\n return True\n\n def get_localnetworkgateway(self):\n '''\n Gets the properties of the specified Local Network Gateway.\n\n :return: deserialized Local Network Gateway instance state dictionary\n '''\n self.log(\"Checking if the Local Network Gateway instance {0} is present\".format(self.local_network_gateway_name))\n found = False\n try:\n response = self.mgmt_client.local_network_gateways.get(resource_group_name=self.resource_group,\n local_network_gateway_name=self.local_network_gateway_name)\n found = True\n self.log(\"Response : {0}\".format(response))\n self.log(\"Local Network Gateway instance : {0} found\".format(response.name))\n except CloudError as e:\n self.log('Did not find the Local Network Gateway instance.')\n if found is True:\n return response.as_dict()\n\n return False\n\n\ndef main():\n \"\"\"Main execution\"\"\"\n AzureRMLocalNetworkGateways()\n\nif __name__ == '__main__':\n main()\n","repo_name":"AlexanderYukhanov/ansible-hatchery","sub_path":"library/azure_rm_appgwlocalnetworkgateway.py","file_name":"azure_rm_appgwlocalnetworkgateway.py","file_ext":"py","file_size_in_byte":11218,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"32002219309","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef getProMPBasis(dt, nSteps, n_of_basis, bandwidth):\n nBasis = n_of_basis\n time = np.arange(dt, nSteps * dt, dt)\n assert nSteps == len(time)\n Ts = nSteps * dt - dt\n\n C = np.zeros(nBasis) # Basis function centres\n H = np.zeros(nBasis) # Basis function bandwidths\n\n # for i in range(nBasis):\n # C[i] = -2 * bandwidth + (Ts + 4 * bandwidth) / nBasis * i\n C = np.linspace(0 - 2 * bandwidth, Ts + 2 * bandwidth, nBasis)\n # for i in range(nBasis):\n # H[i] = bandwidth ** 2\n\n Phi = np.zeros((nBasis, nSteps))\n\n for k, time_k in enumerate(time):\n for j in range(nBasis):\n Phi[j, k] = np.exp(-.5 * (time_k - C[j]) ** 2 / bandwidth ** 2) # Basis function activation over time\n for k in range(Phi.shape[1]):\n Phi[:, k] = (Phi[:, k]) / np.sum(Phi[:, k]) # Normalize basis functions and weight by canonical state\n\n return Phi\n\n\nif __name__ == '__main__':\n dt = 0.002\n nBasis = 10\n bandwidth = 0.2\n time = np.arange(-2*bandwidth, 3+2*bandwidth, dt)\n nSteps = len(time)\n Phi = getProMPBasis(dt, nSteps, nBasis, bandwidth)\n\n fig0, ax0 = plt.subplots(1)\n for ii in range(Phi.shape[0]):\n ax0.plot(time, Phi[ii])\n # ax0.plot(time, sum(Phi[i] for i in range(Phi.shape[0])))\n\n plt.xlim(-2*bandwidth, 3+2*bandwidth)\n plt.grid()\n #plt.savefig(\"ProMP_basis_function.pdf\")\n plt.show()\n","repo_name":"vejc94/RL_G169","sub_path":"HW4/daniel/getProMPBasis.py","file_name":"getProMPBasis.py","file_ext":"py","file_size_in_byte":1442,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18939581903","text":"n_esimo = int(input(\"Digite o n-esimo termo:\"))\n\nultimo = 0\nprimeiro = 0\nsegundo = 1\n\nprint(segundo)\nfor i in range(n_esimo): \n ultimo = primeiro + segundo\n primeiro = segundo\n segundo = ultimo\n print(ultimo)","repo_name":"Silovisk/Python_Brasil_Lista_De_Exercicios","sub_path":"03 - Estrutura De Repeticao/ex015 - fibonacci ate n-esimo.py","file_name":"ex015 - fibonacci ate n-esimo.py","file_ext":"py","file_size_in_byte":213,"program_lang":"python","lang":"pt","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"30418315027","text":"from leezy import Solution, solution\nfrom leezy.assists import TreeContext\n\n\nclass Q113(Solution):\n @solution\n def pathSum(self, root, sum):\n # 36ms 73.27% \n current, total = [], []\n if root is None:\n return []\n self.dfs(root, sum, current, total)\n return total\n\n def dfs(self, node, target, current, total):\n if not node.left and not node.right:\n if node.val == target:\n total.append(current+[node.val])\n return\n current.append(node.val)\n if node.left:\n self.dfs(node.left, target-node.val, current, total)\n if node.right:\n self.dfs(node.right, target-node.val, current, total)\n current.pop()\n\n\n\ndef main():\n q = Q113()\n q.set_context(TreeContext)\n q.add_case(q.case([5, 4, 8], 2).assert_equal([]))\n q.add_case(q.case([5, 4, 8, 11, None, 13, 4, 7, 2, None, None, 5, 1], 22)\n .assert_equal([[5, 4, 11, 2], [5, 8, 4, 5]]))\n q.run()\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"aptend/leetcode-rua","sub_path":"Python/113 - Path Sum II/113_path-sum-ii.py","file_name":"113_path-sum-ii.py","file_ext":"py","file_size_in_byte":1050,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"72035538494","text":"# Task 1 Timeit\n# Написать декоратор, который будет считать время работы функции и выводить на экран. Для текущего времени можно использовать модуль time.​ \n\n\n\n# >> long_running()\n# 0.212341254\n# complete\n\nimport time\nfrom functools import wraps\n\ndef print_run_time(func: callable) -> callable:\n #raise NotImplementedError('Implement me!')\n @wraps(func)\n def wrapper(*args, **kwargs):\n start = time.time()\n val = func(*args, **kwargs)\n stop = time.time()\n print(round(stop - start, 9))\n return val\n return wrapper\n \n\n@print_run_time\ndef long_running():\n time.sleep(0.2)\n return 'complete'\n \nlong_running()","repo_name":"graevskiy/epam_training","sub_path":"class_3_functions_advanced/task1_timeit.py","file_name":"task1_timeit.py","file_ext":"py","file_size_in_byte":767,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22318940303","text":"class CoffeeMachine:\n def __init__(self, menu, money, resources):\n self.menu = menu\n self.money = money\n self.resources = resources\n\n\n def __str__(self):\n return f\"The menu is {list(self.menu.keys())}. Money and resources can get accesed via 'report' input.\"\n\n\n def off(self):\n \"\"\"Turns off the coffee machine.\"\"\"\n return \"Machine turning off.\"\n \n\n def report(self):\n \"\"\"Breaks down the report feature to make it more readable\"\"\"\n print(f\"Water: {self.resources['water']}ml\")\n print(f\"Milk: {self.resources['milk']}ml\")\n print(f\"Coffee: {self.resources['coffee']}g\")\n print(f\"Money: ${self.money}\")\n\n\n def resource_check(self, order_ingredients):\n \"\"\"Returns True when order can be made, and False when resources are insufficient\"\"\"\n for item in order_ingredients:\n if order_ingredients[item] > self.resources[item]:\n print(f\"Sorry there is not enough {item}.\")\n return False\n return True\n\n\n def process_coins(self):\n \"\"\"Returns the total calculated from coins inserted.\"\"\"\n print(\"Please insert coins.\")\n total = int(input(\"How many quarters?:\")) * 0.25\n total += int(input(\"How many dimes?:\")) * 0.1\n total += int(input(\"How many nickles?:\")) * 0.05\n total += int(input(\"How many pennies?:\")) * 0.01\n return total\n\n\n def is_transaction_successful(self, money_received, drink_cost):\n \"\"\"Return True when the payment is accepted, or False if money is insufficient.\"\"\"\n if money_received >= drink_cost:\n change = round(money_received - drink_cost, 2)\n print(f\"Here is ${change} in change.\")\n self.money += drink_cost\n return True\n else:\n print(\"Sorry that's not enough money. Money refunded.\")\n return False\n\n\n def make_coffee(self, drink_name, order_ingredients):\n \"\"\"Deduct the required ingredients from the resources.\"\"\"\n for item in order_ingredients:\n self.resources[item] -= order_ingredients[item]\n print(f\"Here is your {drink_name}. Enjoy it!\")\n\n\nmenu = {\n \"espresso\": {\n \"ingredients\": {\n \"water\": 50,\n \"coffee\": 18,\n },\n \"cost\": 1.5,\n },\n \"latte\": {\n \"ingredients\": {\n \"water\": 200,\n \"milk\": 150,\n \"coffee\": 24,\n },\n \"cost\": 2.5,\n },\n \"cappuccino\": {\n \"ingredients\": {\n \"water\": 250,\n \"milk\": 100,\n \"coffee\": 24,\n },\n \"cost\": 3.0,\n }\n }\n\nmoney = 0\n\nresources = {\n \"water\": 300,\n \"milk\": 200,\n \"coffee\": 100,\n}\n\ndef main():\n\n coffee_machine = CoffeeMachine(menu, money, resources)\n\n print(coffee_machine)\n\n is_on = True\n\n while is_on:\n choice = input(\"What would you like? (espresso/latte/cappuccino): \")\n if choice == \"off\":\n print(coffee_machine.off())\n is_on = False\n elif choice == \"report\":\n coffee_machine.report()\n else:\n drink = menu[choice]\n if coffee_machine.resource_check(drink[\"ingredients\"]):\n payment = coffee_machine.process_coins()\n if coffee_machine.is_transaction_successful(payment, drink[\"cost\"]):\n coffee_machine.make_coffee(choice, drink[\"ingredients\"])\n\nif __name__ == \"__main__\":\n main()\n\n\n\n","repo_name":"Sfrancinelli/Python","sub_path":"Udemy/16-day/coffee-machine-POO.py","file_name":"coffee-machine-POO.py","file_ext":"py","file_size_in_byte":3674,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"33995595299","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[25]:\n\n\nimport numpy as np\nimport warnings\nwith warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom matplotlib.patches import FancyArrowPatch\nfrom mpl_toolkits.mplot3d import proj3d\nfrom mpl_toolkits.mplot3d.art3d import Poly3DCollection\n\n\n# # log barrier with newton method \n\n# In[26]:\n\n\ndef phi_calc(x, ineq_constraints):\n log = 0\n d = 0\n hess1 = 0\n hess2 = 0\n h, dh, dh_2 = ineq_constraints()\n \n for h_i, dh_i, dh2_i in zip(h, dh, dh_2):\n h_i_x = h_i(x)\n log += np.log((-1) * h_i_x)\n d += (dh_i / ((-1) * h_i_x))\n hess1 += (np.matmul(dh_i, dh_i.T) / (h_i_x * h_i_x))\n hess2 += (dh2_i / ((-1) * h_i_x))\n\n return (-1) * log, d, hess1+hess2\n\n\ndef log_barrier(x, f, ineq_constraints, t):\n phi, d_phi, h_phi = phi_calc(x, ineq_constraints)\n f_x, grad_x, hess_x = f(x)\n\n log_barr_f = (t * f_x) + phi\n log_barr_g = (t * grad_x) + d_phi\n log_barr_h = (t * hess_x) + h_phi\n\n return log_barr_f, log_barr_g, log_barr_h\n\n\n# In[40]:\n\n\n\ndef line_search(x, f, step, ineq_constraints, t, init_step_len, slope_ratio, back_track_factor):\n alpha = init_step_len\n log_barr_f, log_barr_g_xk, _ = log_barrier(x, f, ineq_constraints, t)\n log_barr_f_next, _, _ = log_barrier((x + (alpha * step)), f, ineq_constraints, t)\n\n while np.isnan(log_barr_f_next) or (log_barr_f_next - log_barr_f) > (slope_ratio * alpha * np.matmul(log_barr_g_xk.T, step)):\n alpha *= back_track_factor\n log_barr_f_next, _, _ = log_barrier((x + (alpha * step)), f, ineq_constraints, t)\n\n return alpha\n\n\ndef newton(x_p,hessian_p):\n I = np.identity(hessian_p.shape[0])\n B = np.linalg.solve(hessian_p, I)\n return (-1) * B.dot(x_p)\n\n\ndef newton_method(x0, f, ineq_constraints, t, n_equ_const, A, epsilon, init_step_len, slope_ratio,back_track_factor, count, max_iter=100):\n outer_obj = []\n obj = []\n success = False\n xk = x0\n xs = [xk]\n \n for k in range(max_iter):\n log_barr_f_xk, log_barr_g_xk, log_barr_h_xk = log_barrier(xk, f, ineq_constraints, t)\n\n if n_equ_const > 0:\n M1 = np.concatenate((log_barr_h_xk, A.T), axis=1)\n M2 = np.concatenate((A, np.zeros((n_equ_const, n_equ_const))), axis=1)\n M = np.concatenate((M1, M2), axis=0)\n v = np.concatenate(((-1) * log_barr_g_xk, np.zeros((n_equ_const, 1))))\n X = np.linalg.solve(M, v)\n step = X[:log_barr_h_xk.shape[0]]\n else:\n step = newton(log_barr_g_xk, log_barr_h_xk)\n\n lambda_x = np.sqrt(np.matmul(np.matmul(step.T, log_barr_h_xk), step))\n if (0.5 * lambda_x * lambda_x) < epsilon:\n success = True\n return xk, success, count,obj,outer_obj,xs\n\n step_len = line_search(xk, f, step, ineq_constraints, t, init_step_len, slope_ratio,back_track_factor)\n xk1 = xk + (step_len * step)\n f_xk, _, _ = f(xk)\n f_xk1, _, _ = f(xk1) \n print(\"Iteration number = {}\\n Location = {} \\n Objective value = {}\\n Step length = {} \\n objective Change = {}\"\n .format(count, xk1, f_xk1,np.linalg.norm(xk - xk1),np.linalg.norm(f_xk1 - f_xk)))\n \n obj.append(f_xk)\n outer_obj.append(np.linalg.norm(f_xk1 - f_xk))\n xs.append(xk1)\n \n xk = xk1\n count += 1\n\n return xk, success, count,obj,outer_obj,xs\n\n\ndef interior(f, ineq_constraints, eq_constraints_mat, eq_constraints_rhs, x0, init_step_len,slope_ratio, back_track_factor, t=1, mu=10, max_iter_outer=100, max_iter_inner=100,epsilon = 1e-4):\n outer_obj_values = []\n obj_values = []\n x_s = []\n A = eq_constraints_mat\n b = eq_constraints_rhs\n n_equ_const = A.shape[0] \n h_list, dh_list, hess_list = ineq_constraints()\n n_ineq_const = len(h_list)\n count = 0\n x_prev = x0\n success_outer = False\n \n for i in range(max_iter_outer):\n x_next, success,count,obj,outer_obj,xs = newton_method(x_prev, f, ineq_constraints, t, n_equ_const, A,epsilon, init_step_len, slope_ratio,back_track_factor, count, max_iter_inner)\n if not success:\n return x_next,obj_values,outer_obj_values,x_s\n\n obj_values += obj\n outer_obj_values += outer_obj\n x_s+=xs\n\n if n_ineq_const/t < epsilon:\n success_outer = True\n return x_next,obj_values,outer_obj_values,x_s\n\n t = t*mu\n x_prev = x_next\n\n return x_next,obj_values,outer_obj_values,x_s\n\n\n# # Quadratic Objective and constraints\n\n# In[41]:\n\n\ndef quadratic_problem(x):\n f_x = (x[0][0] * x[0][0]) + (x[1][0] * x[1][0]) + ((x[2][0]+1) * (x[2][0]+1))\n grad_x = np.array([[2 * x[0][0]], [2 * x[1][0]], [2 * (x[2][0]+1)]])\n hess_x = np.array([[2, 0, 0],\n [0, 2, 0],\n [0, 0, 2]])\n return f_x, grad_x, hess_x\n\n\ndef quadratic_problem_ineq_constraints():\n h = [lambda x: -x[0][0], lambda x: -x[1][0], lambda x: -x[2][0]]\n dh = [np.array([[-1], [0], [0]]), np.array([[0], [-1], [0]]), np.array([[0], [0], [-1]])]\n dh_2 = [np.zeros((3, 3)), np.zeros((3, 3)), np.zeros((3, 3))]\n return h, dh, dh_2\n\n\n# # Linear Objective and constraints\n\n# In[42]:\n\n\ndef linear_problem(x):\n f_x = (-1)*(x[0][0] + x[1][0])\n grad_x = np.array([[-1], [-1]])\n hess_x = np.array([[0, 0], [0, 0]])\n return f_x, grad_x, hess_x\n\n\ndef lp_ineq_constraints():\n h = [lambda x: 1-x[0][0]-x[1][0], lambda x: x[1][0]-1, lambda x: x[0][0]-2, lambda x: -x[1][0]]\n dh = [np.array([[-1], [-1]]), np.array([[0], [1]]), np.array([[1], [0]]), np.array([[0], [-1]])]\n dh_2 = [np.zeros((2, 2)), np.zeros((2, 2)), np.zeros((2, 2)), np.zeros((2, 2))]\n return h, dh, dh_2\n\n\n# # plots\n\n# In[43]:\n\n\ndef plot_iterations(title, obj1, obj2, label1, label2):\n fig, ax = plt.subplots()\n ax.plot(range(len(obj1)), obj1, label=label1)\n ax.plot(range(len(obj2)), obj2, label=label2)\n ax.legend()\n ax.set_title(title)\n ax.set_xlabel(\"Iteration\")\n ax.set_ylabel(\"Objective function value\")\n plt.show()\n\n \ndef plot_feasible_set_3d(path):\n fig = plt.figure()\n ax = Axes3D(fig)\n x = [1, 0, 0]\n y = [0, 1, 0]\n z = [0, 0, 1]\n \n verts = [list(zip(x, y, z))]\n poly_3d_collection = Poly3DCollection(verts, alpha=0.5, edgecolors=\"k\")\n ax.add_collection3d(poly_3d_collection)\n \n x,y,z=[],[],[]\n for i in range(len(path)):\n x.append(path[i][0][0])\n y.append(path[i][1][0])\n z.append(path[i][2][0])\n \n ax.plot(x,y,z,label=\"The algorithm's path\",marker=\".\")\n plt.show()\n \n\n\ndef plot_linear_feasible_region(label,path):\n fig = plt.figure(figsize=(10, 5))\n ax = fig.add_subplot()\n d = np.linspace(-1, 4, 2000)\n x, y = np.meshgrid(d, d)\n plt.imshow(\n ((y >= -x + 1) & (y <= 1) & (x <= 2) & (y >= 0)).astype(int),\n extent=(x.min(), x.max(), y.min(), y.max()),\n origin=\"lower\",\n cmap=\"Greys\",\n alpha=0.3,\n )\n \n x_path,y_path=[],[]\n for i in range(len(path)):\n x_path.append(path[i][0][0])\n y_path.append(path[i][1][0])\n \n x = np.linspace(0, 4, 2000)\n y1 = -x + 1\n y2 = np.ones(x.size)\n y3 = np.zeros(x.size)\n \n plt.plot(x, y1, label=r'$y\\leq1$')\n plt.plot(x, y2, label=r'$y\\geq0$')\n plt.plot(x, y3, label=r'$y\\geq -x + 1$')\n plt.plot(np.ones(x.size) * 2, x ,label=r'$x\\leq2$')\n plt.plot(x_path,y_path,label=\"The algorithm's path\",marker=\".\")\n plt.xlim(0, 3)\n plt.ylim(0, 2)\n plt.xlabel(r'$x$')\n plt.ylabel(r'$y$')\n plt.legend(bbox_to_anchor=(1.05, 1), loc=2, borderaxespad=0.)\n plt.show()\n\n","repo_name":"mori2711/Numerical-Optimization-with-Python-3327","sub_path":"programming_assignment_2.py","file_name":"programming_assignment_2.py","file_ext":"py","file_size_in_byte":7846,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8300260385","text":"#!/usr/bin/env python3\n\n\nimport rclpy\nfrom rclpy.node import Node\n\nfrom std_msgs.msg import String\nfrom geometry_msgs.msg import Twist\nfrom sensor_msgs.msg import LaserScan\n\n\nclass Dummy(Node):\n\n def __init__(self):\n super().__init__('dummy_publisher')\n self.pub_cmd = self.create_publisher(Twist, 'r0/cmd_vel', 10)\n self.sub_laser = self.create_subscription( LaserScan, '/r0/scan_raw', self.callback_laser, 10)\n \n timer_period = 1 # seconds\n self.timer = self.create_timer(timer_period, self.callback_timer)\n \n self.vmax = 0.5\n self.declare_parameter('vmax', self.vmax)\n\n def callback_laser(self, msg : LaserScan):\n nr_of_ranges = len(msg.ranges)\n if nr_of_ranges == 0: return\n cmd = Twist()\n cmd.linear.x = self.vmax\n if msg.ranges[int(nr_of_ranges/2)] < 1:\n cmd.linear.x = 0.1\n if msg.ranges[int(nr_of_ranges/4)] < 2:\n cmd.angular.z = 0.3\n if msg.ranges[int(nr_of_ranges/4*3)] < 2:\n cmd.angular.z = -0.3\n \n self.pub_cmd.publish(cmd)\n \n \n \n def callback_timer(self):\n vmax = self.get_parameter('vmax').get_parameter_value().double_value\n if self.vmax != vmax :\n self.vmax = vmax \n self.get_logger().info('update vmax %f!' % self.vmax)\n\n\n\ndef main(args=None):\n rclpy.init(args=args)\n\n dummy = Dummy()\n\n rclpy.spin(dummy)\n\n # Destroy the node explicitly\n # (optional - otherwise it will be done automatically\n # when the garbage collector destroys the node object)\n dummy.destroy_node()\n rclpy.shutdown()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"tuw-robotics/tuw_gazebo","sub_path":"tuw_gazebo_wanderer/scripts/wanderer_v1.py","file_name":"wanderer_v1.py","file_ext":"py","file_size_in_byte":1696,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"7"} +{"seq_id":"6745900531","text":"from flask import *\nfrom werkzeug import *\nfrom forms import *\nfrom mail_sender import *\nfrom api import *\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef home():\n # 已登录\n if not checklogin():\n return redirect(url_for('homepage'))\n nologin = request.values.get('nologin')\n # 未登录\n if request.remote_addr not in visited_ip:\n visited_ip.append(request.remote_addr)\n topped_posts = posts.query.filter(posts.topped == True).order_by(-posts.update_time).all()\n tenposts = posts.query.filter(posts.head == True).filter(posts.topped == False).filter(\n posts.no_show == False).order_by(-posts.update_time).all()\n return render_template('view/portal.html', topped_posts=topped_posts, tenposts=tenposts, userip=request.remote_addr,\n datetime=datetime.datetime.now(), \\\n loginform=loginform(), signinform=signinform(), nologin=str(nologin))\n\n\n@app.route('/new', methods=['POST'])\ndef newpost():\n if checklogin():\n return redirect(url_for('home', nologin=1))\n if checkkookie():\n return redirect(url_for('homepage', nokookie=1))\n if check_confirmation():\n return redirect(url_for('homepage', no_confirmation=1))\n form = new_post_form()\n if form.validate_on_submit():\n title = form.title.data\n content = form.content.data.replace('\\n', '
')\n section = form.section.data\n show_post = form.no_show.data\n if show_post == '1':\n no_show = False\n if show_post == '0':\n no_show = True\n pic = request.files.get('pic')\n if pic is not None: # 先判断图的后缀是不是要的\n picname = pic.filename.split('.')\n suffix = picname[-1].lower().strip()\n if not (suffix == 'gif' or suffix == 'jpg' or suffix == 'png' or suffix == 'jpeg'):\n # print('weeeeeeeeee')\n pic = None # 图不对就不要\n if pic != None: # 有图\n post = posts(poster=session.get('kookie'), poster_ip=request.remote_addr, head=True, next=0, title=title,\n content=content, section=section, withpic=True, no_show=no_show)\n try:\n db.session.add(post)\n db.session.commit()\n ider = posts.query.filter(posts.ider == md5(content + str(int(time.time())))).first().id\n try:\n # print(ider,'we have pic here')\n suffix = pic.filename.split('.')[-1]\n route = 'static/uploads/' + str(ider) + '.' + suffix\n pic.save(route) # 图片名就是POST ID\n try:\n posts.query.filter(posts.id == ider).update({'pic_route': route})\n db.session.commit()\n except Exception as e:\n print(e)\n return return500('Pic saving error')\n # print('file saved', filename)\n except Exception as e:\n print(e)\n return redirect(url_for('homepage', postok=1))\n except Exception as e:\n db.session.rollback()\n print(e)\n # return redirect(url_for('home'))\n return return500()\n else: # 没图\n # print('没图')\n post = posts(poster=session.get('kookie'), poster_ip=request.remote_addr, head=True, next=0, title=title,\n content=content, section=section, withpic=False, no_show=no_show)\n try:\n db.session.add(post)\n db.session.commit()\n return redirect(url_for('homepage', postok=1))\n except Exception as e:\n db.session.rollback()\n print(e)\n # return redirect(url_for('home'))\n return return500('db error')\n else:\n # print('no validate====================')\n return return500('no validate')\n\n\n@app.route('/comment/', methods=['POST'])\ndef comment(post_id):\n # print(post_id)#id:给【ID】这个串评论\n identifier = md5(str(time.time()))\n if checklogin():\n return redirect(url_for('home', nologin=1))\n if checkkookie():\n return redirect(url_for('homepage', nokookie=1))\n if check_confirmation():\n return redirect(url_for('homepage', no_confirmation=1))\n form = comment_form()\n if form.validate_on_submit():\n content = form.content.data.replace('\\n', '
')\n pic = request.files.get('pic')\n cited_id = parse_cited(content)\n if cited_id != None: # 把引用的id替换成链接\n content = content.replace(str('>>>' + cited_id + '|||'),\n str('' + '>' + cited_id + '|' + ''))\n if pic != None: # 先判断图的后缀是不是要的\n picname = pic.filename.split('.')\n # print(picname)\n suffix = picname[-1].lower().strip()\n # print(suffix)\n if not (suffix == 'gif' or suffix == 'jpg' or suffix == 'png' or suffix == 'jpeg'):\n # print('weeeeeeeeee')\n pic = None # 图不对就不要\n if pic != None:\n # print('we have pic')\n section = posts.query.filter(posts.id == post_id).first().section\n post = posts(poster=session.get('kookie'), poster_ip=request.remote_addr, head=False, next=0,\n title=identifier, content=content, section=section, withpic=True,\n update_time=datetime.datetime.now())\n try:\n db.session.add(post)\n db.session.commit()\n newid = posts.query.filter(posts.title == identifier).first().id\n next_id = posts.query.filter(posts.id == post_id).first().next\n while next_id != 0:\n temp = posts.query.filter(posts.id == next_id).first().next\n if temp == 0:\n break\n else:\n next_id = temp\n if posts.query.filter(posts.id == post_id).first().next == 0:\n next_id = post_id\n posts.query.filter(posts.id == next_id).update({'next': newid})\n replies = posts.query.filter(posts.id == post_id).first().replies + 1 # 增加回复数\n posts.query.filter(posts.id == post_id).update({'replies': replies})\n posts.query.filter(posts.id == post_id).update({'update_time': datetime.datetime.now()}) # 更新时间\n try:\n db.session.commit()\n # print('comment ok')\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n # print('=============================')\n ider = posts.query.filter(posts.ider == md5(content + str(int(time.time())))).first().id\n # print('----------------------',ider)\n try:\n # print(ider,'we have pic here')\n suffix = pic.filename.split('.')[-1]\n route = 'static/uploads/' + str(ider) + '.' + suffix\n pic.save(route) # 图片名就是POST ID\n try:\n posts.query.filter(posts.id == ider).update({'pic_route': route})\n db.session.commit()\n except Exception as e:\n print(e)\n return return500('保存图片路径或保存图时出错')\n # print('file saved', filename)\n except Exception as e:\n print(e)\n return redirect(request.referrer)\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n else:\n section = posts.query.filter(posts.id == post_id).first().section\n post = posts(poster=session.get('kookie'), poster_ip=request.remote_addr, head=False, next=0,\n title=identifier, content=content, section=section, withpic=False,\n update_time=datetime.datetime.now())\n try:\n db.session.add(post)\n db.session.commit()\n newid = posts.query.filter(posts.title == identifier).first().id\n next_id = posts.query.filter(posts.id == post_id).first().next\n while next_id != 0:\n temp = posts.query.filter(posts.id == next_id).first().next\n if temp == 0:\n break\n else:\n next_id = temp\n if posts.query.filter(posts.id == post_id).first().next == 0:\n next_id = post_id\n posts.query.filter(posts.id == next_id).update({'next': newid})\n replies = posts.query.filter(posts.id == post_id).first().replies + 1 # 增加回复数\n posts.query.filter(posts.id == post_id).update({'replies': replies})\n posts.query.filter(posts.id == post_id).update({'update_time': datetime.datetime.now()}) # 更新时间\n try:\n db.session.commit()\n print('comment ok')\n return redirect(request.referrer)\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n else:\n return return500('no validate')\n\n return redirect(request.referrer)\n\n\n@app.route('/viewpost/', methods=['GET', 'POST'])\ndef viewpost(id): # id是headpost的主键\n form = comment_form()\n allposts = []\n next_id = 0\n result = User.query.filter(User.email == session.get('account')).first()\n post = posts.query.filter(posts.id == id).first()\n if post == None:\n return return404()\n posterkookie = post.poster\n allposts.append(post) # 将一楼加至list\n next_id = post.next # 每一楼的id\n while next_id != 0:\n nextpost = posts.query.filter(posts.id == next_id).first()\n allposts.append(nextpost)\n next_id = nextpost.next\n return render_template('view/viewpost.html', form=form, post=post, allposts=allposts,\n len_of_all_posts=len(allposts), result=result, loginform=loginform(),\n signinform=signinform())\n\n\n@app.route('/delpost/', methods=['GET', 'POST'])\ndef delpost(id):\n if checklogin():\n return redirect(url_for('home', nologin=1))\n account = session.get('account')\n this_user = User.query.filter(User.email == account).first()\n this_post = posts.query.filter(posts.id == id).first()\n poster_kookie = this_post.poster\n if this_user.kookies == poster_kookie or this_user.admin == True:\n next_id = this_post.next # 找出指向的下一个postid\n if this_post.head == True: # 是头就把头移除了,等效于整体移除\n db.session.delete(this_post)\n try:\n db.session.commit()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500(str(e))\n return redirect(url_for('homepage', deleteok=1))\n else:\n last_post = posts.query.filter(posts.next == id).first().id # 找出上一个指向该post的id\n db.session.delete(this_post)\n posts.query.filter(posts.id == last_post).update({'next': next_id})\n try:\n db.session.commit()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500(e)\n return redirect(request.referrer)\n else:\n return return500('无权限进行本操作')\n\n\n@app.route('/changeavatar/', methods=['GET', 'POST'])\ndef changeavt(avtid):\n if checklogin():\n return redirect(url_for('home', nologin=1))\n if check_confirmation():\n return redirect(url_for('homepage', no_confirmation=1))\n account = session.get('account')\n kookie = User.query.filter(User.email == account).first().kookies\n User.query.filter(User.email == account).update({'avatar': avtid})\n try:\n db.session.commit()\n try:\n posts.query.filter(posts.poster == kookie).update({'avatar': avtid})\n db.session.commit()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n return redirect(url_for('homepage'))\n\n\n@app.route('/newkookie', methods=['GET', 'POST'])\ndef newkookie():\n if checklogin():\n return redirect(url_for('home', nologin=1))\n if check_confirmation():\n return redirect(url_for('homepage', no_confirmation=1))\n account = session.get('account')\n kookie = cookie(User.query.filter(User.email == account).first().username)\n result = User.query.filter(User.email == account).update({'kookies': kookie})\n try:\n db.session.commit()\n try:\n oldkookie = str(User.query.filter(User.email == account).first().oldkookies)\n oldkookie += (kookie + '-')\n User.query.filter(User.email == account).update({'oldkookies': oldkookie}) # 保存历史kookie\n db.session.commit()\n session['kookie'] = kookie # s设置新kookie\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n except Exception as e:\n db.session.rollback()\n print(e)\n return return500()\n return redirect(request.referrer)\n\n\n@app.route('/home', methods=['GET', 'POST'])\ndef homepage():\n if request.remote_addr not in visited_ip:\n visited_ip.append(request.remote_addr)\n if checklogin():\n return redirect(url_for('home', nologin=1))\n nokookie = None\n email = session.get('account')\n topped_posts = posts.query.filter(posts.topped == True).order_by(-posts.update_time).all()\n result = User.query.filter(User.email == email).first()\n if result == None:\n return redirect(url_for('home', nologin=1))\n session['kookie'] = result.kookies\n if result.kookies == '00000000':\n nokookie = 1\n allposts = posts.query.filter(posts.head == True).filter(posts.topped == False).filter(\n posts.no_show == False).order_by(-posts.post_time).all()\n return render_template('view/home.html', result=result, newpostform=new_post_form(), \\\n allposts=allposts, topped_posts=topped_posts, section='时间线',\n no_confirmation=request.values.get('no_confirmation'), nokookie=nokookie)\n\n\n@app.route('/gg', methods=['GET', 'POST'])\ndef gg():\n return render_template('view/gg.html')\n\n\n@app.route('/section/', methods=['GET', 'POST'])\ndef viewsection(section_name):\n if checklogin():\n return redirect(url_for('home', nologin=1))\n email = session.get('account')\n result = User.query.filter(User.email == email).first()\n if result == None:\n return redirect(url_for('home', nologin=1))\n relposts = posts.query.filter(posts.section == section_name).filter(posts.head == True).order_by(\n -posts.update_time).all()\n return render_template('view/home.html', result=result, newpostform=new_post_form(), \\\n allposts=relposts, section=section_name)\n\n\n@app.route('/logout')\ndef logout():\n session.clear()\n return redirect(url_for('home', logout=1))\n\n\n@app.route('/login', methods=['GET', 'POST'])\ndef login():\n if request.method == 'GET':\n return return500()\n form = loginform()\n if form.validate_on_submit():\n email = form.name.data # 其实是email\n password = form.password.data\n print(email, password)\n result = User.query.filter(User.email == email).first() # 唯一辨识符\n if result == None:\n return redirect(url_for('home', err=1))\n elif result.password == password:\n session['account'] = result.email\n session['kookies'] = result.kookies\n session['avatar'] = result.avatar\n active_ip = request.remote_addr\n User.query.filter(User.email == email).update({'active_ip': active_ip})\n try:\n db.session.commit()\n except Exception as e:\n print(e)\n return redirect(url_for('homepage'))\n else:\n return redirect(url_for('home', err=1))\n else:\n return redirect(url_for('home', novalidation=1))\n\n\n@app.route('/signin', methods=['GET', 'POST'])\ndef signin():\n if request.method == 'GET':\n return return500()\n form = signinform()\n if form.validate_on_submit():\n username = form.name.data\n psw1 = form.password1.data\n psw2 = form.password2.data\n email = form.email.data\n # print(username, psw1, email)\n if psw1 != psw2:\n return return500('密码与确认密码不同')\n if User.query.filter(User.email == email).first():\n return return500('用户邮箱已被注册')\n try:\n newuser = User(username=username, password=psw1, email=email, active_ip=request.remote_addr,\n password_hash=md5(psw1 + str(datetime.datetime.now())), confirmed=False, admin=False)\n db.session.add(newuser)\n db.session.commit()\n session['account'] = email\n session['kookies'] = '00000000'\n psw_hash = User.query.filter(User.email == email).first().password_hash\n if not psw_hash:\n return return500('psw_hash is None')\n try:\n sender(email, 'localhost:5000/api/confirmation?code=' + psw_hash)\n except Exception as e:\n print(e)\n return return500('邮件发送失败')\n return redirect(url_for('homepage'))\n except Exception as e:\n db.session.rollback()\n print(e)\n return redirect(url_for('home', signin_error=1))\n else:\n return redirect(url_for('home', novalidation=1))\n\n\nif __name__ == '__main__':\n # db.drop_all()\n db.create_all()\n app.run(host='localhost', debug=True, port=5000, threaded=True)\n","repo_name":"StuGRua/2020_project","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":18779,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"17663168684","text":"import random\nimport gym\nfrom gym import spaces\nimport numpy as np\nimport colorama\nfrom t_1000.render import GraphGenerator\n\ncolorama.init()\nWINDOW_SIZE = 40\n\nclass TradingEnv(gym.Env):\n visualization = None\n\n def __init__(self, config):\n self.assets = config['assets'],\n self.assets_list = self.assets[0]\n self.currency_list = config['currency'],\n self.currency = self.currency_list[0]\n self.granularity = config['granularity'],\n self.datapoints = config['datapoints']\n self.df_complete = config['df_complete']\n self.df_features = config['df_features']\n self.initial_balance = config['variables']['initial_account_balance']\n self.commission = config['variables']['commission']\n self.variables = config['variables']\n self.shares_held = {}\n self.shares_bought = {}\n self.shares_sold = {}\n self.first_prices = {}\n self.initial_bought = {}\n self.trades = {}\n self.current_price = {}\n self.observation_length = len(self.df_features[self.assets_list[0]].columns) + 4 + 3 # 4,3 = (balance, cost, sales, net_worth) + (shares bought, shares sold, shares held foreach asset)\n\n # action space = buy and sell for each asset, pĺus hold position\n action_space = 1 + len(self.assets_list) * 2\n\n self.action_space = spaces.Box(\n low=np.array([0, 0]),\n high=np.array([action_space, 1]),\n dtype=np.float16)\n\n\n # obs space = (num assets, indicator + (balance, cost, sales, net_worth) + (shares bought, shares sold, shares held foreach asset))\n observation_space = (len(self.assets_list),\n self.observation_length)\n\n self.observation_space = spaces.Box(\n low=-np.finfo(np.float32).max,\n high=np.finfo(np.float32).max,\n shape=observation_space,\n dtype=np.float16)\n\n def reset(self):\n # Reset the state of the environment to an initial state\n self._reset_balance()\n self.current_step = 0\n self._get_first_prices()\n self._compute_initial_bought()\n self._reset_trades()\n\n return self._next_observation()\n\n def step(self, action):\n # Execute one time step within the environment\n self._take_action(action)\n self.current_step += 1\n\n reward = self.compute_reward()\n done = self.net_worth <= 0 or self.balance <= 0 or self.current_step >= len(\n self.df_features[self.assets_list[0]].loc[:, 'open'].values) - 1\n obs = self._next_observation()\n\n return obs, reward, done, {}\n\n def compute_reward(self):\n net_worth_and_buyhold_mean = (self.net_worth + self.buy_and_hold) / 2\n reward = (self.net_worth - self.buy_and_hold) / \\\n net_worth_and_buyhold_mean\n\n # reward = self.net_worth - self.initial_balance\n return reward\n\n def _reset_trades(self):\n for asset in self.assets_list:\n self.trades[asset] = []\n\n def _compute_initial_bought(self):\n \"\"\"spread the initial account balance through all assets\"\"\"\n for asset in self.assets_list:\n self.initial_bought[asset] = 1/len(self.assets_list) * \\\n self.initial_balance / self.first_prices[asset]\n\n def _get_first_prices(self):\n for asset in self.assets_list:\n self.first_prices[asset] = self.df_features[asset]['close'][0]\n\n def _reset_shares_bought_n_sold(self):\n for asset in self.assets_list:\n self.shares_bought[asset] = 0.0\n self.shares_sold[asset] = 0.0\n\n def _reset_cost_n_sales(self):\n self.cost = 0\n self.sales = 0\n\n def _reset_balance(self):\n self._reset_cost_n_sales()\n self.balance = self.initial_balance\n self.net_worth = self.initial_balance\n for asset in self.assets_list:\n self.shares_held[asset] = 0.0\n self.shares_bought[asset] = 0.0\n self.shares_sold[asset] = 0.0\n\n def _next_observation(self):\n observation = np.empty((0, self.observation_length), int)\n for asset in self.assets_list:\n current_step = np.array(self.df_features[asset].values[self.current_step])\n current_step_with_shares = np.array([np.append(current_step, [\n self.balance,\n self.cost,\n self.sales,\n self.net_worth,\n self.shares_bought[asset],\n self.shares_sold[asset],\n self.shares_held[asset]\n ])])\n observation = np.append(observation, current_step_with_shares, axis=0)\n return observation\n\n def _compute_current_price(self):\n for asset in self.assets_list:\n self.current_price[asset] = random.uniform(self.df_features[asset].loc[self.current_step, 'open'],\n self.df_features[asset].loc[self.current_step, 'close'])\n\n def _buy(self, asset, amount):\n self.shares_bought[asset] = self.balance * \\\n amount / self.current_price[asset]\n self.cost = self.shares_bought[asset] * \\\n self.current_price[asset] * (1 + self.commission)\n self.shares_held[asset] += self.shares_bought[asset]\n self.balance -= self.cost\n return True\n\n def _sell(self, asset, amount):\n self.shares_sold[asset] = self.shares_held[asset] * amount\n self.sales = self.shares_sold[asset] * \\\n self.current_price[asset] * (1 - self.commission)\n self.shares_held[asset] -= self.shares_sold[asset]\n self.balance += self.sales\n return True\n\n def _can_buy(self, amount):\n if self.balance >= self.balance * amount * (1 + self.commission):\n return True\n else:\n return False\n\n def _buy_or_sell(self, action_type, amount):\n bought = False\n sold = False\n can_buy = self._can_buy(amount=amount)\n for index, asset in enumerate(self.assets_list*2):\n if action_type < index / 2 + 1 and can_buy and not bought:\n bought = self._buy(asset=asset, amount=amount)\n elif action_type < index + 1 and not sold:\n sold = self._sell(asset=asset, amount=amount)\n\n def _compute_trade(self):\n for asset in self.assets_list:\n if self.shares_sold[asset] > 0 or self.shares_bought[asset] > 0:\n self.trades[asset].append({\n 'price': self.current_price[asset],\n 'step': self.current_step,\n 'amount': self.shares_sold[asset] if self.shares_sold[asset] > 0 else self.shares_bought[asset],\n 'total': self.sales if self.shares_sold[asset] > 0 else self.cost,\n 'type': 'sell' if self.shares_sold[asset] > 0 else 'buy'\n })\n\n def _compute_net_worth(self):\n self.net_worth = self.balance\n for asset in self.assets_list:\n self.net_worth += self.shares_held[asset] * \\\n self.current_price[asset]\n\n def _compute_buy_n_hold_strategy(self):\n buy_and_hold = 0\n for asset in self.assets_list:\n buy_and_hold += self.initial_bought[asset] * \\\n self.current_price[asset]\n self.buy_and_hold = buy_and_hold\n\n def _take_action(self, action):\n self._compute_current_price()\n action_type = action[0]\n amount = action[1]\n # bounds of action_space doesn't seem to work, so this line is necessary to not overflow actions\n if 0 < amount <= 1 and action_type > 0:\n self._reset_shares_bought_n_sold()\n self._reset_cost_n_sales()\n self._buy_or_sell(action_type=action_type, amount=amount)\n self._compute_trade()\n\n self._compute_net_worth()\n self._compute_buy_n_hold_strategy()\n\n def _render_to_file(self, filename='render.txt'):\n pass\n # profit = self.net_worth - INITIAL_ACCOUNT_BALANCE\n\n # file = open(filename, 'a+')\n\n # file.write('Step: {}\\n'.format(self.current_step))\n # file.write('Balance: {}\\n'.format(self.balance))\n # file.write('Shares held: {}\\n'.format(self.shares_held))\n # file.write('Avg cost for held shares: {}\\n'.format(self.cost))\n # file.write('Net worth: {}\\n'.format(self.net_worth))\n # file.write('Buy and hold strategy: {}\\n'.format(self.buy_and_hold))\n # file.write('Profit: {}\\n\\n'.format(profit))\n\n # file.close()\n\n # *--------------------------------------------------------------------\n # * tirar o visualization.render() e trabalhar no rollout.py primeiro?\n # *--------------------------------------------------------------------\n\n def render(self, mode='live', **kwargs):\n\n if mode == 'file':\n self._render_to_file(kwargs.get('filename', 'render.txt'))\n\n elif mode == 'live':\n if self.visualization == None:\n self.visualization = GraphGenerator(assets=self.assets_list, currency=self.currency, granularity=self.granularity[0],\n datapoints=self.datapoints, df_complete=self.df_complete, df_features=self.df_features, variables=self.variables)\n self.visualization.render(current_step=self.current_step, net_worth=self.net_worth, buy_and_hold=self.buy_and_hold,\n trades=self.trades, shares_held=self.shares_held, balance=self.balance, window_size=WINDOW_SIZE)\n\n def close(self):\n if self.visualization != None:\n self.visualization.close()\n self.visualization = None\n","repo_name":"Draichi/T-1000","sub_path":"t_1000/env/trading_env.py","file_name":"trading_env.py","file_ext":"py","file_size_in_byte":9728,"program_lang":"python","lang":"en","doc_type":"code","stars":166,"dataset":"github-code","pt":"7"} +{"seq_id":"6985331895","text":"from django.urls import path\r\nfrom . import views\r\n\r\n\r\napp_name = 'music'\r\n\r\nurlpatterns = [ # Looks for functions in views\r\n # /music/\r\n path('', views.IndexView.as_view(), name='index'), # Using a class, but as_view() treats it as a function\r\n\r\n path('register/', views.UserFormView.as_view(), name='register'),\r\n\r\n # /music/pk/\r\n path('/', views.DetailView.as_view(), name='detail'),\r\n\r\n # /music/album/add/\r\n path('album/add/', views.AlbumCreate.as_view(), name='album-add'),\r\n\r\n # /music/album/pk/\r\n path('album//', views.AlbumUpdate.as_view(), name='album-update'),\r\n\r\n # /music/album/pk/delete/\r\n path('album//delete/', views.AlbumDelete.as_view(), name='album-delete'),\r\n\r\n]\r\n","repo_name":"jchadwick92/first_django","sub_path":"music/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":743,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"2567569112","text":"#Make a program that reads the height and width of a wall in meters, calculate its area and the amount of paint needed\r\n# to paint it, knowing that each liter of paint paints an area of de 2m2.\r\n\r\nlar = float(input('Wall width:'))\r\ncomp = float(input('Wall length:'))\r\n\r\nm2 = lar * comp \r\nli = m2 / 2\r\n\r\nprint('You wall has the dimension of {} x {} and your area is {} m²'.format(lar, comp,m2))\r\nprint('To paint this wall, you will need {}L of ink'.format(li))","repo_name":"NathanSouza27/Python","sub_path":"ING/011 - Painting the wall.py","file_name":"011 - Painting the wall.py","file_ext":"py","file_size_in_byte":461,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"6097142468","text":"from contextlib import contextmanager\nimport time\n\n\n@contextmanager\ndef timer(*args, **kwargs):\n mnt1 = time.perf_counter()\n try:\n yield mnt1\n finally:\n mnt2 = time.perf_counter()\n print(mnt2 - mnt1)\n\n\nsymbols = \"$¢£¥€¤^*&$%#(*))@**(^##JklNJKDBJDBJ&*(&*4848\"\n\nwith timer():\n beyond_ascii = [ord(s) for s in symbols if ord(s) > 127]\n\nprint(f'listcomp:\\n\\t{beyond_ascii}')\n\nwith timer():\n beyond_ascii = list(filter(lambda c: c > 127, map(ord, symbols)))\n\nprint(f'filter & map:\\n\\t{beyond_ascii}')\n","repo_name":"aliakbar1881/Fluent_Python_Tips","sub_path":"chapter2/listcomp_vs_filter&map.py","file_name":"listcomp_vs_filter&map.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7720304520","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, val=0, left=None, right=None):\n# self.val = val\n# self.left = left\n# self.right = right\nclass Solution:\n def sumNumbers(self, root: Optional[TreeNode]) -> int:\n ans = 0\n def solve(root,num=0):\n nonlocal ans\n if not root.left and not root.right: \n # print(num)\n ans+=num*10+root.val\n return\n if root.left:\n solve(root.left,num*10+root.val)\n if root.right:\n solve(root.right,num*10+root.val)\n solve(root)\n return ans\n ","repo_name":"sudo-vaibhav/leetcode-solutions","sub_path":"0129-sum-root-to-leaf-numbers/0129-sum-root-to-leaf-numbers.py","file_name":"0129-sum-root-to-leaf-numbers.py","file_ext":"py","file_size_in_byte":681,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"38030958293","text":"from werkzeug.serving import run_simple\n\nimport json\nfrom backend.ATBackend import ATBackendServer\n\nsettings = json.load(open('config.json'))\n\n\nif __name__ == '__main__':\n backend = ATBackendServer(settings['elasticsearch_host'], settings['elasticsearch_index'],\n settings['elasticsearch_type'],\n ['pretrained_models/at_cnnmodel.model', 'pretrained_models/at_lrmodel.model', \n 'pretrained_models/at_idf.model'])\n run_simple(settings['backend_host'], settings['backend_port'], backend.application)\n \n","repo_name":"emorynlp/qa-demo","sub_path":"run_backend.py","file_name":"run_backend.py","file_ext":"py","file_size_in_byte":596,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"73352974304","text":"import datetime\nfrom f412.models import *\n\ndef dateToString(date):\n try:\n return date.strftime(\"%Y/%m/%d\")\n except:\n date = parseDate(date)\n return date.strftime(\"%Y/%m/%d\")\n\ndef dateToComment(date):\n if date.weekday() == 0:\n date = date - datetime.timedelta(days=3)\n else:\n date = date - datetime.timedelta(days=1)\n return date.strftime(\"%y%m%d\")\n \n#FUncion por si en algun momento es necesaria depuracion\ndef F412ToString(f412):\n toReturn = \"Programa: \" + f412.Programa.name + \"\\n\"\n toReturn += \"Componente: \" + f412.Componente.name + \"\\n\"\n toReturn += \"PN: \" + f412.PN.name + \"\\n\"\n toReturn += \"Area: \" + f412.Area.name + \"\\n\"\n toReturn += \"Defecto: \" + f412.Defecto.name + \"\\n\"\n toReturn += \"Fecha: \" + dateToString(f412.Fecha) + \"\\n\"\n toReturn += \"Estado: \" + f412.Estado.name + \"\\n\"\n toReturn += \"SGM: \" + f412.SGM.number + \"\\n\"\n toReturn += \"Horas: \"+ f412.horas + \"\\n\"\n toReturn += \"Ref: \" + f412.Referencia + \"\\n\"\n toReturn += \"Descripcion: \" + f412.Descripcion + \"\\n\"\n return toReturn\n\ndef parseDate(dateToParse):\n dateToParse = str(dateToParse)\n first = dateToParse.find('-')\n if first == -1:\n first = dateToParse.find('/')\n year = int(dateToParse[:first])\n dateToParse = dateToParse[first+1:]\n \n second = dateToParse.find('-')\n if second == -1:\n second = dateToParse.find('/')\n month = int(dateToParse[:second])\n \n space = dateToParse.find(\" \")\n if space == -1:\n space = len(dateToParse)\n day = int(dateToParse[second+1:space])\n \n if year < 2000:\n year += 2000\n toReturn = datetime.datetime(year, month, day)\n return toReturn\n\ndef toList(df):\n prev = df.values.tolist()\n toReturn = []\n for dfList in prev:\n toReturn = toReturn + dfList\n \n return toReturn\n\ndef toFloat(toConvert):\n try:\n return float(toConvert.replace(',','.'))\n except:\n return 0.0","repo_name":"diazddavid/tfgAerospace","sub_path":"f412/toString.py","file_name":"toString.py","file_ext":"py","file_size_in_byte":1974,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"2630907907","text":"from copy import deepcopy\n\nimport numpy as np\nimport cv2\nfrom matplotlib import pyplot as plt\n\nfrom utils.automation import corrs_automation\n\ndef drawlines(img1src, img2src, lines, pts2src):\n \"\"\"img1 - image on which we draw the epilines for the points in img2\n lines - corresponding epilines\"\"\"\n img1_copy = deepcopy(img1src)\n img2_copy = deepcopy(img2src)\n r, c, _ = img1src.shape\n # Edit: use the same random seed so that two images are comparable!\n np.random.seed(0)\n for r, pt2 in tuple(zip(lines, pts2src))[:10]:\n color = tuple(np.random.randint(0, 255, 3).tolist())\n x0, y0 = map(int, [0, -r[2]/r[1]])\n x1, y1 = map(int, [c, -(r[2]+r[0]*c)/r[1]])\n img1_copy = cv2.line(img1_copy, (x0, y0), (x1, y1), color, 40)\n img2_copy = cv2.circle(img2_copy, tuple(np.int32(pt2[0])), 50, color, -1)\n return img1_copy, img2_copy\n\n\nimg1 = cv2.imread(f\"./data/93.jpg\") #queryimage # left image\nimg2 = cv2.imread(f\"./data/91.jpg\") #trainimage # right image\n\nret, src_pts, dst_pts, _, _ = corrs_automation(img1, img2) # Refer to appendix D\n\n[F, mask] = cv2.findFundamentalMat(src_pts,dst_pts,cv2.RANSAC, 5.0)\n\n# Find epilines corresponding to points in right image (second image) and\n# drawing its lines on left image\nlines1 = cv2.computeCorrespondEpilines(\n dst_pts.reshape(-1, 1, 2), 2, F)\nlines1 = lines1.reshape(-1, 3)\nimg5, img6 = drawlines(img1, img2, lines1, src_pts, dst_pts)\n\n# Find epilines corresponding to points in left image (first image) and\n# drawing its lines on right image\nlines2 = cv2.computeCorrespondEpilines(\n src_pts.reshape(-1, 1, 2), 1, F)\nlines2 = lines2.reshape(-1, 3)\nimg3, img4 = drawlines(img2, img1, lines2, dst_pts, src_pts)\n\n# Convert to RGB to be saved\nimg3 = cv2.cvtColor(img3, cv2.COLOR_BGR2RGB)\nimg4 = cv2.cvtColor(img4, cv2.COLOR_BGR2RGB)\nimg5 = cv2.cvtColor(img5, cv2.COLOR_BGR2RGB)\nimg6 = cv2.cvtColor(img6, cv2.COLOR_BGR2RGB)\n\n# Saving Output Images\nplt.figure()\nplt.subplot(121), plt.axis('off'), plt.imshow(img6)\nplt.subplot(122), plt.axis('off'), plt.imshow(img5)\nplt.suptitle(\"Epipolar Points (Left) and Corresponding Epipolar Lines (Right)\", y = 0.75)\nplt.savefig(\"outputs/A_points_to_B_lines.png\")\n\n\nplt.figure()\nplt.subplot(121), plt.axis('off'), plt.imshow(img4)\nplt.subplot(122), plt.axis('off'), plt.imshow(img3)\nplt.suptitle(\"Epipolar Points (Left) and Corresponding Epipolar Lines (Right)\", y = 0.75)\nplt.savefig(\"outputs/B_points_to_A_lines.png\")\n\nplt.figure()\nplt.subplot(121), plt.axis('off'), plt.imshow(img5)\nplt.subplot(122), plt.axis('off'), plt.imshow(img3)\nplt.suptitle(\"Epilines in both images\", y = 0.75)\nplt.savefig('outputs/epilines.png')","repo_name":"marcus-neo/cvpr","sub_path":"cw1/4b_fundamentalMatrix/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2667,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"28515640125","text":"import serial\nimport time\ntime.sleep(3)\n\n#input_serial = serial.Serial('/dev/ttyACM0', baudrate = 9600, timeout=1)\noutput_serial = serial.Serial('/dev/cu.usbmodem1411', baudrate = 19200, timeout = 1)\n\ndirection = ''\nwhile direction != 'stop':\n\tprint(\"Enter a direction\")\n\tprint(\"[Forward = F, Reverse = B, Left = L, Right = R]\")\n\tdirection = input()\n\toutput_serial.write(direction.encode())\n\tif direction == 'L' or direction == 'R':\n\t\tprint(\"How much turn?\")\n\t\tamount = input()\n\t\toutput_serial.write(amount.encode())\n","repo_name":"dailerob/PetFinder","sub_path":"motor_control_python.py","file_name":"motor_control_python.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"10286774336","text":"# BOJ 7576 Tomato\nimport sys\ninput = sys.stdin.readline\n\ndef bfs(arr, row, col, ripe, empty):\n global visited\n count = -1 # starting from day:0\n new_ripen = 0\n while len(ripen):\n stage = ripen.pop()\n next_stage = []\n count += 1\n while len(stage):\n y, x = stage.pop()\n if y > 0 and not visited[y - 1][x] and arr[y - 1][x] == 0:\n next_stage.append((y - 1, x))\n new_ripen += 1\n arr[y-1][x] = 1\n visited[y-1][x] = 1\n if y < row - 1 and not visited[y + 1][x] and arr[y + 1][x] == 0:\n next_stage.append((y + 1, x))\n new_ripen += 1\n arr[y + 1][x] = 1\n visited[y+1][x] = 1\n if x > 0 and not visited[y][x - 1] and arr[y][x - 1] == 0:\n next_stage.append((y, x - 1))\n new_ripen += 1\n arr[y][x-1] = 1\n visited[y][x-1] = 1\n if x < col - 1 and not visited[y][x + 1] and arr[y][x + 1] == 0:\n next_stage.append((y, x + 1))\n new_ripen += 1\n arr[y][x+1] = 1\n visited[y][x+1] = 1\n if next_stage:\n ripen.append(next_stage)\n if new_ripen!=row*col-ripe-empty:\n return -1\n return count\n\n\nm, n = map(int, input().split())\ntomato = []\nripen = [[]]\nempty = 0\nfor i in range(n):\n row = [*map(int, input().split())]\n tomato.append(row)\n for j in range(len(row)):\n if row[j] == 1:\n ripen[0].append((i, j))\n if row[j] == -1:\n empty += 1\nvisited = [[0] * m for _ in range(n)]\nprint(bfs(tomato, n, m, len(ripen[0]), empty))\n","repo_name":"Qud4300/Baekjoon_Online_Judge","sub_path":"7576 토마토/tomato.py","file_name":"tomato.py","file_ext":"py","file_size_in_byte":1704,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"27662635659","text":"class TaskInfo(object):\n \"\"\"\n This class represents a task in an instance of a workflow.\n \"\"\"\n WAITING = 1\n CANCELLED = 2\n COMPLETED = 4\n LIKELY = 8\n TRIGGERED = 16\n\n def __init__(self, job_id = None, node = None):\n \"\"\"\n Constructor.\n \"\"\"\n self.id = None\n self.job_id = job_id\n self.node_id = None\n self.name = None\n self.status = None\n self.last_change = None\n if node is not None:\n self.node_id = node.id\n self.name = node.spec.name\n self.status = node.state\n","repo_name":"gonicus/clacks","sub_path":"contrib/workflow/SpiffWorkflow/src/SpiffWorkflow/Server/TaskInfo.py","file_name":"TaskInfo.py","file_ext":"py","file_size_in_byte":631,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"15753922361","text":"\"\"\" [연습]\n 함수 정의 : count_words\n 인자 : filename\n\n 인자로 받은 파일명을 open 하여 파일을 읽어서 단어를 수를 출력한다.\n 존재하지 않는 파일명으로 예외가 발생해도 아무런 일을 하지 않는다\n\"\"\"\ndef count_words(filename):\n try:\n with open('./data/'+filename, 'r', encoding='utf-8') as re:\n num = len(re.read().split())\n\n except FileNotFoundError as error:\n print('%s 파일은 존재하지 않습니다. 에러코드 - %s' % (filename,error))\n else:\n print('%s 파일의 단어수는 %d개 입니다' % (filename,num))\n\n\nfilenames = ['sample.xml', 'xxxx.xxx', 'temp.json']\nfor filename in filenames:\n count_words(filename)","repo_name":"zeromook/Python","sub_path":"e_file_class/Ex01_readFile_exam.py","file_name":"Ex01_readFile_exam.py","file_ext":"py","file_size_in_byte":737,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"11897610841","text":"import numpy as np\nimport scipy as sp\nimport matplotlib.pyplot as plt\n\ndef y(t):\n if t < 0: \n return 0\n else:\n return 2/3 * (1 - np.exp(-1.5e6 * t))\n\ndef y_disc(n):\n if n < 0:\n return 0\n else:\n return 2/3 * (1 - (1 - 7.5e5 * 1e-7)**(n*1e7)/(1 + 7.5e5 * 1e-7)**(n*1e7))\n\nyt = sp.vectorize(y)\nyn = sp.vectorize(y_disc)\n\nf = np.loadtxt('4_7_output.dat')\nT = np.linspace(0, 5e-6, 40)\n\nplt.plot(T, yt(T), label='$y(t)$')\nplt.plot(f[:,0], f[:,1], 'o', mfc='g', mec='g', label='y(t) with ngspice')\nplt.plot(T, yn(T), 'o', mfc='orange',label='$y(n)$')\nplt.grid()\nplt.legend()\nplt.savefig('../figs/4_7.pdf')\nplt.show()","repo_name":"TataSaiManoj/EE3900","sub_path":"cktsig/codes/4_7.py","file_name":"4_7.py","file_ext":"py","file_size_in_byte":649,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"23402410022","text":"from abc import abstractmethod\nfrom datetime import date\n\nfrom kingfisher_scrapy.base_spiders import IndexSpider, PeriodicSpider\nfrom kingfisher_scrapy.exceptions import SpiderArgumentError\nfrom kingfisher_scrapy.items import FileError\nfrom kingfisher_scrapy.util import components, handle_http_error\n\n\nclass ChileCompraAPIBase(IndexSpider, PeriodicSpider):\n custom_settings = {\n 'DOWNLOAD_FAIL_ON_DATALOSS': False,\n }\n\n # BaseSpider\n date_format = 'year-month'\n default_from_date = '2009-01'\n dont_truncate = True\n\n # PeriodicSpider\n # The path parameters are {system}/{year}/{month}/{offset}/{limit}.\n pattern = 'http://api.mercadopublico.cl/APISOCDS/OCDS/{0}/{1.year:d}/{1.month:02d}/{2}/{3}'\n formatter = staticmethod(components(-4, -1)) # year-month-offset\n start_requests_callback = 'parse_list'\n\n # IndexSpider\n result_count_pointer = '/pagination/total'\n limit = 100\n parse_list_callback = 'parse_page'\n\n # Local\n available_systems = {\n 'convenio': 'listaOCDSAgnoMesConvenio',\n 'licitacion': 'listaOCDSAgnoMes',\n 'trato-directo': 'listaOCDSAgnoMesTratoDirecto'\n }\n system = None\n\n @classmethod\n def from_crawler(cls, crawler, system=None, *args, **kwargs):\n spider = super().from_crawler(crawler, system=system, *args, **kwargs)\n if system and spider.system not in spider.available_systems:\n raise SpiderArgumentError(f'spider argument `system`: {spider.system!r} not recognized')\n return spider\n\n def build_urls(self, date):\n for system in self.available_systems:\n if self.system and system != self.system:\n continue\n yield self.pattern.format(self.available_systems[system], date, 0, self.limit)\n\n @handle_http_error\n def parse(self, response):\n data = self.parse_list_loader(response)\n if isinstance(data, FileError):\n yield data\n return\n\n # Replace the \"\\u0000\" escape sequence in the JSON string, which is rejected by PostgreSQL.\n # https://www.postgresql.org/docs/current/datatype-json.html\n response = response.replace(body=response.body.replace(b'\\x00', b''))\n yield from super().parse(response)\n\n @handle_http_error\n def parse_page(self, response):\n data = self.parse_list_loader(response)\n if isinstance(data, FileError):\n yield data\n return\n\n for item in data['data']:\n # An item looks like:\n #\n # {\n # \"ocid\": \"ocds-70d2nz-2359-2-LE19\",\n # \"urlTender\": \"https://apis.mercadopublico.cl/OCDS/data/tender/2359-2-LE19\",\n # \"urlAward\": \"https://apis.mercadopublico.cl/OCDS/data/award/2359-2-LE19\",\n # \"urlPlanning\": \"https://apis.mercadopublico.cl/OCDS/data/planning/2359-2-LE19\"\n # }\n yield from self.handle_item(item)\n\n @abstractmethod\n def handle_item(self, item):\n pass\n\n # from IndexSpider\n def parse_list_loader(self, response):\n data = response.json()\n\n # Some files contain invalid packages, e.g.:\n # {\n # \"status\": 500,\n # \"detail\": \"error\"\n # }\n if 'status' in data and data['status'] != 200:\n data['http_code'] = data['status']\n return self.build_file_error_from_response(response, errors=data)\n\n return data\n\n # from IndexSpider\n def url_builder(self, value, data, response):\n # URL looks like http://api.mercadopublico.cl/APISOCDS/OCDS/listaOCDSAgnoMesTratoDirecto/2021/03/31500/100\n system = components(-5, -4)(response.request.url)\n year = int(components(-4, -3)(response.request.url))\n month = int(components(-3, -2)(response.request.url).lstrip('0'))\n\n return self.pattern.format(system, date(year, month, 1), value, self.limit)\n","repo_name":"open-contracting/kingfisher-collect","sub_path":"kingfisher_scrapy/spiders/chile_compra_api_base.py","file_name":"chile_compra_api_base.py","file_ext":"py","file_size_in_byte":3900,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"7"} +{"seq_id":"28977435114","text":"import sqlite3\r\n\r\nconnect = sqlite3.connect('all.db')\r\ncursor = connect.cursor()\r\ncursor.execute(\"\"\"CREATE TABLE IF NOT EXISTS subjects(\r\nmath text,\r\nrussian text,\r\nchemistry text,\r\nobg text,\r\nobshestvo text,\r\ngeography text)\r\n\r\n \"\"\")\r\nconnect.commit()","repo_name":"dsfssfs/vhhvg","sub_path":"bd.py","file_name":"bd.py","file_ext":"py","file_size_in_byte":255,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"28530811730","text":"import os\nimport json\n\n# Get Current Execution Path\ncurrentPath =str(__file__.replace(\"main.py\", \"\"))\n# Get All Files\nfiles = [file for file in os.listdir() if file.endswith(\".txt\")]\n\n# Main Tool Title \"VULNHUNT\"\nprint(\"\"\"\n __ ___ _ _ _ _ _ _ _ _ _ _ _____ \n \\ \\ / / | | | | | \\ | | | | | | | | \\ | |_ _|\n \\ \\ / /| | | | | | \\| | |_| | | | | \\| | | | \n \\ V / | |_| | |___| |\\ | _ | |_| | |\\ | | | \n \\_/ \\___/|_____|_| \\_|_| |_|\\___/|_| \\_| |_| \n \n\"\"\")\n\n# while loop depends on chosen VUL if no chosen (VUL = \"none\") it will keep asking to choose a VUL\nVUL = \"none\" # VUL Variable\nwhile VUL == \"none\":\n print(\"Choose the vulnerability : \\n\")\n # For Loop to Print All vulnerabilities names from all texts in the same directory of main.py script\n for idx in range(len(files)):\n # Get File Path For Each\n filePath = open(currentPath + str(files[idx]))\n # Load as json from text and display section \"Title\"\n print(str(idx) + \".\", json.load(filePath)[\"Title\"])\n # out of the loop scope print last option Exit\n print(\"\\n\" + str(len(files)), \"Exit\\n\")\n # Try Catch expression to validate that user input must be Integer \"رقم\"\n try:\n VUL = int(input(\"Select the number of the vulnerability from the menu: \"))\n # if user choose last option \"Exit\" then tell him Bye Bye and Break Script\n if VUL == len(files):\n print(\"\\n Bye Bye!!\\n\")\n VUL = \"none\"\n break\n # if user choose a number that is not from the VULs list, tell him that it does'nt exist\n # and make him re-Choose\n elif not(VUL in range(0, len(files))):\n print(\"\\nOption Doesn't Exist!\\n\")\n VUL = \"none\"\n # if user wrote anything wrong as if not a number \"Integer\", tell him to input only numbers\n # and make him re-Choose\n except:\n print(\"Input Must be Numerical only\")\n VUL = \"none\"\n\n\n# IF ALL ABOVE CODE WENT SUCCESSFULLY\n# Second Menu will show up\nVUL_SECTION = \"none\"\n# VUL_SECTION describes the examination chosen\n# Second Menu will show up as long as no examination chosen (VUL_SECTION =\"none\") And a VUL is chosen Example (VUL = 0)\n# While loop checks if VUL is Integer and No Examination Chosen\nwhile isinstance(VUL, int) and VUL_SECTION == \"none\":\n # Load Chosen VUL and read text as JSON\n SelectedFileData = json.load(open(currentPath + str(files[VUL])))\n # Print Chosen Vulnerability\n print(\"Chosen Vulnerability: \", SelectedFileData[\"Title\"])\n # Print Available Examination AND Exit\n print(\"\"\"\n Choose the examination method :\n 1.Describe\n 2.Manually\n 3.Bypasses\n \n 4.Exit/Quit\n \n\"\"\")\n # Try Catch expression to validate that examination input must be Integer \"رقم\"\n try:\n VUL_SECTION = int(input(\"Select the number of the method from the menu: \\n\"))\n # if user choose 4 \"Exit/Quit\" then tell him Bye Bye and Break Script\n if VUL_SECTION == 4:\n print(\"\\n Bye Bye!!\\n\")\n VUL_SECTION =\"none\"\n VUL = \"none\"\n break\n # if user choose 1 \"Describe\" then print Describe section of Chosen Vulnerability\n elif VUL_SECTION == 1:\n print(\"\\n\", SelectedFileData[\"Describe\"])\n break\n # if user choose 2 \"Manually\" then print Manually section of Chosen Vulnerability\n elif VUL_SECTION == 2:\n print(\"\\n\", SelectedFileData[\"Manually\"])\n break\n # if user choose 2 \"Bypasses\" then print Bypasses section of Chosen Vulnerability\n elif VUL_SECTION == 3:\n print(\"\\n\", SelectedFileData[\"Bypasses\"])\n break\n # if user wrote anything wrong as if not a number \"Integer\", tell him to input only numbers\n # Or Section File Doesn't Exist in text file\n # and make him re-Choose\n except:\n print(\"Invalid Option Or Not Found Content\")\n VUL_SECTION = \"none\"\n\n","repo_name":"ahmedrowaihi/VULNHUNT-vulnerabilites-menu","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4020,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37063852092","text":"'''\n The logger utility for logging the intermediate results\n'''\n\nimport os\nimport sys\nsys.path.append(\"../\")\nimport utils.gen_utils as gen_utils\n\nclass LJLogger(object):\n def __init__(self,dump_root_path,auto_dump=False,load_history=False):\n pass\n if not os.path.exists(dump_root_path):\n os.makedirs(dump_root_path)\n\n self.dump_file=os.path.join(dump_root_path,\"log.log\")\n if load_history and os.path.isfile(self.dump_file):\n self.dump_dict=gen_utils.read_dict_from_json(self.dump_file)\n else:\n self.dump_dict=dict()\n self.auto_dump=auto_dump\n\n def log(self,time,item_val_dict):\n if not time in self.dump_dict:\n self.dump_dict[time]=dict()\n\n for item,val in item_val_dict.iteritems():\n self.dump_dict[time][item]=val\n\n if self.auto_dump:\n self.dump()\n\n def dump(self):\n gen_utils.write_dict_to_json(self.dump_dict,self.dump_file)\n return\n","repo_name":"grantlj/11731_proj","sub_path":"utils/log_utils.py","file_name":"log_utils.py","file_ext":"py","file_size_in_byte":994,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"73671795424","text":"#!/usr/bin/env python3\n\"\"\"This script demonstrates use of the SIE-Batch API from Python3\"\"\"\n\n# Lint with $ pylint sie_get_py.py (assumes .pylintrc file in dir)\n\nimport calendar\nimport datetime\nfrom datetime import datetime\nfrom io import BytesIO\nimport json\nfrom os import path\nfrom pathlib import Path\nimport re\nimport sys\nimport time\nfrom time import strftime\nimport pycurl\n\nendpoint = 'batch.sie-remote.net'\nuseproxy = False # note: 'false' will not work\n\n# -----------------------------------------------------------------------------\n#\ndef getkeyfromlocalfile():\n \"\"\"Retrieves the SIE-Batch API key\"\"\"\n\n filepath = str(Path.home()) + \"/.sie-get-key.txt\"\n\n if not path.exists(filepath):\n print(\"\\nERROR:\\n\\n No SIE-Batch API keyfile at \"+filepath)\n sys.exit(1)\n\n with open(filepath) as stream:\n myapikey = stream.read().rstrip()\n\n return myapikey\n\n# -----------------------------------------------------------------------------\n#\ndef make_query(url, useproxy, params, outputfilename):\n \"\"\"make query\"\"\"\n\n if outputfilename != '-999':\n try:\n f = open(outputfilename, \"wb\")\n except IOError:\n sys.exit(\"error opening output file for results\")\n else:\n buffer = BytesIO()\n\n c = pycurl.Curl()\n c.setopt(pycurl.URL, url)\n c.setopt(pycurl.HTTPHEADER, ['Content-Type: application/json'])\n c.setopt(pycurl.POST, True)\n c.setopt(pycurl.POSTFIELDS, params)\n c.setopt(pycurl.FOLLOWLOCATION, 1)\n c.setopt(pycurl.CONNECTTIMEOUT, 300)\n c.setopt(pycurl.TIMEOUT, 86400)\n c.setopt(pycurl.USERAGENT, 'sie_get_py/1.0')\n\n # we're going to write the actual data files directly to the outputfile\n # other stuff (apikey check, channel listing, etc.) we're just going\n # to write to a buffer (and then read that))\n if outputfilename == '-999':\n c.setopt(pycurl.WRITEDATA, buffer)\n else:\n c.setopt(pycurl.WRITEDATA, f)\n\n if useproxy:\n c.setopt(pycurl.PROXY, '127.0.0.1')\n c.setopt(pycurl.PROXYPORT, 1080)\n c.setopt(pycurl.PROXYTYPE, pycurl.PROXYTYPE_SOCKS5)\n\n tries = 3\n rc = ''\n\n while tries >= 1:\n c.perform()\n rc = c.getinfo(c.RESPONSE_CODE)\n\n # if writing to a buffer, we need to extract results and change\n # from a bytestring to a string\n if outputfilename == '-999':\n body = buffer.getvalue()\n content = body.decode('iso-8859-1')\n\n # successful transfer? if so, break out of the loop\n # if not, try it again\n #pylint: disable=no-else-break\n if rc == 200:\n break\n else:\n print('Problem in make_query: response code='+str(rc))\n\n #pylint: disable=no-else-return\n if outputfilename == '-999':\n return content\n else:\n sys.exit(0)\n #pylint: enable=no-else-break\n\n# -----------------------------------------------------------------------------\ndef validateapikeyonline(endpoint, useproxy):\n \"\"\" check the API key for validity on the live SIE-Batch API server \"\"\"\n myapikeyval = getkeyfromlocalfile()\n params = {'apikey' : myapikeyval}\n params2 = json.dumps(params)\n queryURL = 'https://' + endpoint + '/siebatchd/v1/validate'\n returned_content = make_query(queryURL, useproxy, params2, '-999')\n returned_content_json_format = json.loads(returned_content)\n status = returned_content_json_format['_status']\n return status\n\n# -----------------------------------------------------------------------------\n#\ndef format_and_printout_the_chan_list(chan_list):\n \"\"\" we have the channel data, now format and print it out in a report \"\"\"\n chan_list_json_format = json.loads(chan_list)\n\n new_hash = {}\n #pylint: disable=unused-variable\n for k, v in chan_list_json_format.items():\n #pylint: enable=unused-variable\n keystring = k.replace('ch', '')\n keystring = keystring.rjust(3, ' ')\n actual_val = chan_list_json_format[k]['description']\n new_hash[keystring] = actual_val\n\n for k, v in sorted(new_hash.items()):\n print(k, new_hash[k])\n\n sys.exit(0)\n\n# -----------------------------------------------------------------------------\n#\ndef list_channels(endpoint, useproxy):\n \"\"\" retrieve a list of channels from the server \"\"\"\n myapikeyval = getkeyfromlocalfile()\n params = {'apikey' : myapikeyval}\n params2 = json.dumps(params)\n queryURL = 'https://' + endpoint + '/siebatchd/v1/validate'\n returned_content = make_query(queryURL, useproxy, params2, '-999')\n returned_content_json_format = json.loads(returned_content)\n extract_bit = returned_content_json_format['profile']['siebatch']\n json_query_object = json.dumps(extract_bit)\n format_and_printout_the_chan_list(json_query_object)\n sys.exit(0)\n\n# -----------------------------------------------------------------------------\n#\ndef format_and_printout_chan_time_limits(chan, earliest_time_string, \\\n latest_time_string, volume):\n \"\"\" print a summary of available channel date range and volume \"\"\"\n\n # take the channel status parameters and print them out in a little report\n chan = chan.rjust(4)\n if int(volume) >= 4:\n volume = '{:,d}'.format(int(volume))\n volume = volume.rjust(16)\n\n # could add a header, but it's pretty self-obvious, right?\n # printf('chan earliest datetime latest datetime octets\\n')\n print(chan+' \"'+earliest_time_string+'\" '+\\\n '\"'+latest_time_string+'\" '+volume)\n sys.exit(0)\n\n# -----------------------------------------------------------------------------\n#\ndef show_intervals(endpoint, useproxy, chan_to_check):\n \"\"\" get the starting and stopping date range and volume \"\"\"\n # with square brackets\n newchan_to_check = '[' + chan_to_check + ']'\n # no brackets, but with ch literal prefix\n chan_with_prefix = 'ch' + chan_to_check\n\n myapikeyval = getkeyfromlocalfile()\n params = {'apikey': myapikeyval, 'channels': newchan_to_check}\n params2 = json.dumps(params)\n\n # {\"channels\":\"[212]\",\"apikey\":\"blah\"} needs to become\n # {\"channels\":[212],\"apikey\":\"blah\"} (e.g., no quotes around [chan])\n params2 = params2.replace('\"[', '[')\n params2 = params2.replace(']\"', ']')\n\n url = 'https://' + endpoint + '/siebatchd/v1/siebatch/chdetails'\n response = make_query(url, useproxy, params2, '-999')\n decoded_results = json.loads(response)\n\n earliest_time_string = \\\n decoded_results['channels'][chan_with_prefix]['earliest']\n latest_time_string = \\\n decoded_results['channels'][chan_with_prefix]['latest']\n size_string = \\\n decoded_results['channels'][chan_with_prefix]['size']\n\n return (earliest_time_string, latest_time_string, size_string)\n\n# ----------------------------------------------------------------------------\n#\ndef fixup_ending_datetime_in_the_future():\n \"\"\" if the ending date is in the future, reel it back in! \"\"\"\n # Replace future times with the current GMT time.\n # The following returns a datetime structure\n epochseconds = time.gmtime()\n enddatetime = time.strftime('%Y-%m-%d %H:%M:%S', epochseconds)\n enddatetime2 = re.sub(r'..$', '00', enddatetime)\n return enddatetime2\n\n# ----------------------------------------------------------------------------\n#\ndef string_fmt_time_to_seconds(string_format_time):\n \"\"\" utility function to convert a string format time to epoch seconds \"\"\"\n dt = datetime.strptime(string_format_time, \"%Y-%m-%d %H:%M:%S\")\n epoch_seconds = calendar.timegm(dt.utctimetuple())\n return epoch_seconds\n\n# -----------------------------------------------------------------------------\n#\ndef check_channel(endpoint, useproxy, chanflagstring, startdatetime, \\\n enddatetime):\n \"\"\" make sure that the channel is available and the dates are in-range \"\"\"\n\n # get the available datetime range for this channel\n #pylint:disable=unused-variable\n (earliest_time_string, latest_time_string, chan_to_check) = \\\n show_intervals(endpoint, useproxy, chanflagstring)\n #pylint:enable=unused-variable\n\n # convert the requested and available start datetimes into Un*x seconds\n requested_start_seconds = string_fmt_time_to_seconds(startdatetime)\n earliest_date_seconds = string_fmt_time_to_seconds(earliest_time_string)\n requested_stop_seconds = string_fmt_time_to_seconds(enddatetime)\n latest_date_seconds = string_fmt_time_to_seconds(latest_time_string)\n\n # start datetime must be earlier than stop date time\n if (requested_stop_seconds - requested_start_seconds) < 0:\n sys.exit('Start datetime must be earlier than stop datetime')\n\n # start datetime may not be earlier than earliest data available\n if (requested_start_seconds - earliest_date_seconds) < 0:\n sys.exit('Start datetime out of range. Must be no earlier than ' + \\\n earliest_time_string)\n\n # end datetime may not be in the future\n if (requested_stop_seconds - latest_date_seconds) > 0:\n enddatetime = fixup_ending_datetime_in_the_future()\n\n return (startdatetime, enddatetime)\n\n# ----------------------------------------------------------------------------\n#\ndef validate_input_time_date_format(mydatetime):\n \"\"\" make sure the user has followed the required datetime format \"\"\"\n # parameter is datetime to format check. if invalid, abort run.\n # if valid, return the validated (but unchanged) datetime (could skip\n # doing this for now, but at some point we might decide to fix up bad\n # string formatting as a convenience to the user, so...)\n\n # check the format with a regex\n if not(re.match(r'/\\A\\d{4}-\\d{2}-\\d{2}\\ \\d{2}:\\d{2}:\\d{2}\\Z/', \\\n mydatetime), mydatetime):\n print(\"bad starting time format -- must be \\\"YYYY-MM-DD HH:MM:SS\\\"\\n\")\n sys.exit(1)\n\n return mydatetime\n\n# -----------------------------------------------------------------------------\n#\ndef zero_unused_seconds(mydatetime):\n \"\"\" if seconds are non-zero in the time stamps, zero them out \"\"\"\n # since SIE-Batch API does not care about seconds, we set them to zero\n mydatetime2 = re.sub(r'..$', '00', mydatetime)\n return mydatetime2\n\n# -----------------------------------------------------------------------------\n#\ndef convert_relative_times_to_real_datetimes(minutesback):\n \"\"\" one option is relative times; if we get one, make it a real time \"\"\"\n # in relative format, the initial \"ending time\" is actually the minutes\n # worth of data we want to retrieve\n # the \"real\" ending datetime will be created from the current GMT time\n # we will be doing math on the epoch seconds\n\n myformat = '%Y-%m-%d %H:%M:%S'\n\n endingtime = time.gmtime()\n epochseconds = calendar.timegm(endingtime)\n\n # now compute the formatted ending date time in standard YYYY-MM-DD HH:MM:SS\n enddatetime = strftime(myformat, endingtime)\n\n # find just the seconds from that string\n extraseconds = int(enddatetime[-2:])\n # subtract the seconds from the full datetime to end up with 00 seconds\n endingtime_seconds = int(epochseconds) - extraseconds\n\n # let's now work on the starting time\n # we compute the \"real\" starting datetime by offsetting backwards\n # our to-be-modified datetime is in epoch seconds, so convert min to seconds\n mysecondsback = int(minutesback) * 60\n startseconds = endingtime_seconds - mysecondsback\n startdatetime = strftime(myformat, time.gmtime(startseconds))\n enddatetime = strftime(myformat, time.gmtime(endingtime_seconds))\n\n return (startdatetime, enddatetime)\n\n# -----------------------------------------------------------------------------\n#\ndef fix_times():\n \"\"\" handles calling the rest of the routines to fix up times \"\"\"\n # arguments come in from the command line so we don't pass them in\n\n # chanflagstring = str(sys.argv[1])\n startdatetime = str(sys.argv[2])\n enddatetime = str(sys.argv[3])\n\n # if relative times, replace the ending time with the current GMT time\n # set the starting time back by the specified number of minutes\n if startdatetime == 'now':\n (startdatetime, enddatetime) = \\\n convert_relative_times_to_real_datetimes(enddatetime)\n else:\n # we have real timedate stamps for starting and ending datetimes\n # process the starting datetime value...\n # correctly written datetime value?\n # also zero the seconds if present (SIE-Batch API doesn't use them)\n validate_input_time_date_format(startdatetime)\n startdatetime = zero_unused_seconds(startdatetime)\n\n # repeat for the ending datetime value...\n validate_input_time_date_format(enddatetime)\n enddatetime = zero_unused_seconds(enddatetime)\n return (startdatetime, enddatetime)\n\n# ----------------------------------------------------------------------------\n#\n# https://stackoverflow.com/questions/1265665/how-can-i-check-if-a-string-represents-an-int-without-using-try-except\n#\ndef isInt_try(v):\n \"\"\" convenience function to see if a string might be integer-ish \"\"\"\n # pylint: disable=unused-variable,multiple-statements,bare-except\n try: i = int(v)\n except: return False\n return True\n\n# -----------------------------------------------------------------------------\n#\ndef build_filename(chanflagstring, startdatetime, enddatetime):\n \"\"\"construct the filename from the command line arguments and return it\"\"\"\n\n string1 = startdatetime.replace(' ', '@')\n string2 = enddatetime.replace(' ', '@')\n\n nmsgchannels = [\"204\", \"206\", \"207\", \"208\", \"221\"]\n if chanflagstring in nmsgchannels:\n filetype = \".nmsg\"\n else:\n filetype = \".jsonl\"\n\n outputfilename = \"sie-ch\" + chanflagstring + \"-{\" + string1 + \\\n \"}-{\" + string2 + \"}\" + filetype\n\n return outputfilename\n\n# -----------------------------------------------------------------------------\n#\ndef print_usage_info():\n \"\"\" deliver a succinct usage summary if needed \"\"\"\n print('''\nUsage:\n\n sie_get_py channel \"now\" minutesBack\n Example: sie_get_py 212 now 15\n\nOR\n\n sie_get_py channel \"startdatetime\" \"enddatetime\"\n Example: sie_get_py 212 \"2020-01-07 00:13:00\" \"2020-01-07 00:28:00\"\n\nConvenience functions:\n\n Check SIE-Batch API key: sie_get_py checkkey\n Get a listing of channels: sie_get_py channels\n Get datetime range and volume for a channel: sie_get_py 212\n\nNotes:\n\n Datetimes are UTC and must be quoted. (Current UTC datetime: $ date -u )\n Zero pad any single digit months, days, hours, minutes or seconds.\n Seconds must be entered as part of the UTC datetimes (but are ignored)\n Ending datetime in the future? It will be clamped to current datetime.\n ''')\n sys.exit(1)\n\n# ----------------------------------------------------------------------------\n#\ndef one_real_arg(endpoint, useproxy, first_arg):\n \"\"\" sometimes we only see one option on the command line; process it \"\"\"\n defined_channels = {'24', '25', '27', '42', '80', '114', '115', \\\n '204', '206', '207', '208', '211', '212', '213', '214', '221'}\n\n if first_arg == 'channels':\n # list channels for the user\n list_channels(endpoint, useproxy)\n sys.exit(0)\n\n elif first_arg == 'checkkey':\n # check the user's key for validity\n status = validateapikeyonline(endpoint, useproxy)\n print(\"API key status is \"+status)\n sys.exit(0)\n\n elif (isInt_try(first_arg) and (first_arg in defined_channels)):\n # list details about the specified channel\n (earliest, latest, datasize) = show_intervals(endpoint, \\\n useproxy, first_arg)\n format_and_printout_chan_time_limits(first_arg, earliest, \\\n latest, datasize)\n sys.exit(0)\n\n elif (not(first_arg in defined_channels) and (isInt_try(first_arg))):\n # the requested channel is not one we offer, so...\n print(\"Channel not offered via this script\")\n sys.exit(0)\n\n else:\n print_usage_info()\n sys.exit(0)\n\n# ----------------------------------------------------------------------------\n#\ndef three_real_args(endpoint, useproxy):\n \"\"\" other times we may see three arguments on the command line... \"\"\"\n chanflagstring = str(sys.argv[1])\n (startdatetime, enddatetime) = fix_times()\n (startdatetime, enddatetime) = check_channel(endpoint, useproxy, \\\n chanflagstring, startdatetime, enddatetime)\n\n outputfilename = build_filename(chanflagstring, startdatetime, enddatetime)\n myapikey = getkeyfromlocalfile()\n params = {\"apikey\": myapikey, \"channel\": int(chanflagstring), \\\n \"start_time\": startdatetime, \"end_time\": enddatetime}\n params2 = json.dumps(params)\n queryURL = \"https://\" + endpoint + \"/siebatchd/v1/siebatch/chfetch\"\n make_query(queryURL, useproxy, params2, outputfilename)\n sys.exit(0)\n\n# ============================================================================\n# main\n\nif len(sys.argv) == 1:\n print_usage_info()\n sys.exit(0)\n\nelif len(sys.argv) >= 2:\n first_arg = sys.argv[1]\n\ncommand_line_arg_count = len(sys.argv)-1\n\nif command_line_arg_count == 1:\n one_real_arg(endpoint, useproxy, first_arg)\n sys.exit(0)\n\nelif command_line_arg_count == 3:\n three_real_args(endpoint, useproxy)\n sys.exit(0)\n\nelif (command_line_arg_count <= 0) or (command_line_arg_count >= 4) or \\\n (command_line_arg_count == 2):\n print_usage_info()\n sys.exit(0)\n","repo_name":"farsightsec/blog-code","sub_path":"sie_get_clients/sie_get_python/sie_get_py.py","file_name":"sie_get_py.py","file_ext":"py","file_size_in_byte":17459,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"7"} +{"seq_id":"29291379640","text":"from Bio import SeqIO\nfrom collections import Counter\nimport pandas as pd\ninputdir = \"/isi/olga/xin/Halophile_project/db_source/nr_halo.faa\"\noutputdir1 = \"/isi/olga/xin/Halophile_project/output/20160401/nr_sgl_seq.fasta\"\noutputdir2 = \"/isi/olga/xin/Halophile_project/output/20160401/nr_dup_seq.fasta\"\nseq_id = []\nseq_id_sgl = []\nseq_id_dup = []\nsgl_seq = []\ndup_seq = []\nhandle = open(inputdir, \"rU\")\nf1 = open(outputdir1, 'w')\nf2 = open(outputdir2, 'w')\nprint(\"open file successfully!\")\nfasta_sequences = SeqIO.parse(handle, \"fasta\")\nfor record in fasta_sequences:\n\tseq_id.append (record.id)\nseq_id_count = Counter(seq_id)\ndf = pd.DataFrame.from_dict(seq_id_count, orient = 'index').reset_index()\ndf = df.rename(columns={'index':'seq_id', 0:'count'})\nhandle.close()\nprint(\"seq id list done!\")\nfor i in range(len(df.iloc[:, 0])):\n\tif df.iloc[i, 1] > 1:\n\t\tseq_id_dup.append(df.iloc[i, 0])\n\tif df.iloc[i, 1] == 1:\n\t\tseq_id_sgl.append(df.iloc[i, 0])\nprint(\"divided seq id to two groups (sigle copy and dupliates)\")\nhandle = open(inputdir, \"rU\")\nfasta_sequences = SeqIO.parse(handle, \"fasta\")\n#print(\"opened\")\n#i = 0\nfor record1 in fasta_sequences:\n\t#print(i)\n\t#i = i + 1\n\tif record1.id in seq_id_dup:\n\t\tdup_seq.append(record1)\n\telse:\n\t\tsgl_seq.append(record1)\nprint(\"alomost there!\")\nSeqIO.write(dup_seq, f2, \"fasta\")\nSeqIO.write(sgl_seq, f1, \"fasta\")\nhandle.close()\nf1.close()\nf2.close()","repo_name":"mascoma/IC_project_scripts","sub_path":"remove_duplicate_seqid.py","file_name":"remove_duplicate_seqid.py","file_ext":"py","file_size_in_byte":1386,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"18630605522","text":"\nfrom matplotlib import pyplot\n\n\nclass Visualizer():\n\n @staticmethod\n def _points_to_two_lists(points):\n xPoints = list(map(lambda point: point[0], points))\n yPoints = list(map(lambda point: point[1], points))\n return xPoints, yPoints\n\n @staticmethod\n def show_line_from_points(points):\n \"\"\"\n show_point_plot([(1, 1), (1.1, 3), (7, 8), (0.9, 4)])\n \"\"\"\n xPoints, yPoints = points_to_two_lists(points)\n pyplot.plot(xPoints, yPoints)\n pyplot.show()\n\n @staticmethod\n def easy_show_scatter_plot(points):\n\n # Set the figure size in inches\n pyplot.figure(figsize=(10,6))\n\n xPoints, yPoints = points_to_two_lists(points)\n pyplot.scatter(xPoints, yPoints, alpha=0.5) # label = \"label_name\"\n\n # Set x and y axes labels\n pyplot.xlabel('X Values')\n pyplot.ylabel('Y Values')\n\n pyplot.title('Scatter Title')\n # pyplot.legend()\n pyplot.show()\n\n @staticmethod\n def scatter_plot(points):\n assert len(points[0]) == 4, \"a point is defined as (x_axis, y_axis, color, alpha)\"\n\n xPoints = list(map(lambda point: point[0], points))\n yPoints = list(map(lambda point: point[1], points))\n colours = list(map(lambda point: point[2], points))\n alphas = list(map(lambda point: point[3], points))\n \n pyplot.figure(figsize=(10,6))\n\n pyplot.scatter(xPoints, yPoints, c = colours, label = \"label_name\", alpha=alphas)\n\n # Set x and y axes labels\n pyplot.xlabel('X Values')\n pyplot.ylabel('Y Values')\n\n pyplot.title('Scatter Title')\n # pyplot.legend()\n pyplot.show()\n\n\n# Test\n# show_scatter_plot([(5, 5), (5, 5), (6, 7), (1, 2), (5.1, 5), (5, 5.1)])\n# show_line_from_points([(1, 1), (1.1, 3), (7, 8), (0.9, 4)])\n# show_scatter_plot([(6, 7), (1, 2), (5.1, 5), (5, 5.1)])\n# multicolor_scatter_plot([(6, 7, \"green\"), (1, 2, \"red\"), (5.1, 5, \"red\"), (5, 5.1, \"blue\")])","repo_name":"Deep-Learning-as-a-Service/EvA","sub_path":"src/utils/Visualizer.py","file_name":"Visualizer.py","file_ext":"py","file_size_in_byte":1985,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"43802288733","text":"from management.models import Account\nfrom .models import Teacher, Exam\nfrom django import forms\nfrom digital_school import settings\n\n\nclass TeacherAccountUpdateForm(forms.ModelForm):\n class Meta:\n model = Account\n fields = ('email',)\n\n\nclass TeacherInfoUpdateForm(forms.ModelForm):\n class Meta:\n model = Teacher\n fields = ('phone_number',)\n\n\nclass ExamCreationForm(forms.ModelForm):\n teacher = forms.ChoiceField(widget=forms.HiddenInput, required=False)\n description = forms.CharField(label='Description',\n widget=forms.Textarea(attrs={'cols': '60', 'rows': '4'}),\n required=False)\n\n exam_creation_date = forms.DateTimeField(input_formats=settings.DATE_INPUT_FORMATS,\n widget=forms.DateTimeInput(attrs={'type': 'datetime-local', 'step': 1}),\n required=True)\n exam_deadline = forms.DateTimeField(input_formats=settings.DATE_INPUT_FORMATS,\n widget=forms.DateTimeInput(attrs={'type': 'datetime-local', 'step': 1}),\n required=True)\n\n '''\n teacher_class_choices = {}\n\n for i in Teacher.grade_class.all():\n teacher_class_choices[i.id] = i.grade_class_name\n TEACHER_CLASS_CHOICES = ((k, v) for k, v in zip(teacher_class_choices.keys(), teacher_class_choices.values()))\n grade_class = forms.ChoiceField(choices=TEACHER_CLASS_CHOICES)'''\n\n class Meta:\n model = Exam\n fields = (\n 'grade_class', 'academic_term', 'academic_term', 'exam_type', 'exam_name', 'description',\n 'exam_creation_date', 'exam_deadline', 'multiplier', 'full_mark',\n )\n","repo_name":"yassinebebba/django-school-platform","sub_path":"teachers/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1777,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"18916025336","text":"from sos.report.plugins import Plugin, UbuntuPlugin, RedHatPlugin\nfrom sos.utilities import is_executable\n\n\nclass Slurm(Plugin, UbuntuPlugin, RedHatPlugin):\n\n short_desc = \"Slurm Workload Manager\"\n\n plugin_name = 'slurm'\n profiles = ('hpc',)\n packages = (\n # Ubuntu\n 'slurm-wlm',\n 'slurmd',\n 'slurmdbd',\n 'slurmctld',\n # EL\n 'slurm',\n 'slurm-slurmctld',\n 'slurm-slurmd',\n 'slurm-slurmdbd',\n )\n services = (\n 'slurmd',\n 'slurmdbd',\n 'slurmctld',\n )\n\n def setup(self):\n \"\"\" Slurm Workload Manager\n \"\"\"\n\n self.add_copy_spec('/etc/slurm/*.conf')\n\n if is_executable('sinfo'):\n self.add_cmd_output([\n 'sinfo --all --list-reasons --long',\n 'sinfo --all --long',\n ])\n\n if is_executable('squeue'):\n self.add_cmd_output([\n 'squeue --all --long',\n ])\n\n scontrol_cmds = [\n 'aliases',\n 'assoc_mgr',\n 'bbstat',\n 'burstBuffer',\n 'config',\n 'daemons',\n 'dwstat',\n 'federation',\n 'frontend',\n 'job',\n 'licenses',\n 'node',\n 'partition',\n 'reservation',\n 'slurmd',\n 'step',\n 'topology',\n ]\n\n if is_executable('scontrol'):\n self.add_cmd_output(\n [f\"scontrol show {i}\" for i in scontrol_cmds]\n )\n\n config_file = '/etc/slurm/slurm.conf'\n\n slurmd_log_file = '/var/log/slurmd.log'\n slurmctld_log_file = '/var/log/slurmctld.log'\n\n try:\n with open(config_file, 'r') as cf:\n for line in cf.read().splitlines():\n if not line:\n continue\n words = line.split('=')\n if words[0].strip() == 'SlurmdLogFile':\n slurmd_log_file = words[1].strip()\n if words[0].strip() == 'SlurmctldLogFile':\n slurmctld_log_file = words[1].strip()\n except IOError as error:\n self._log_error('Could not open conf file %s: %s' %\n (config_file, error))\n\n if not self.get_option(\"all_logs\"):\n self.add_copy_spec([\n slurmd_log_file,\n slurmctld_log_file,\n ])\n else:\n self.add_copy_spec([\n f\"{slurmd_log_file}*\",\n f\"{slurmctld_log_file}*\",\n ])\n\n def postproc(self):\n conf_paths = \"/etc/slurm\"\n\n slurm_keys = [\n 'AccountingStoragePass',\n 'JobCompPass',\n ]\n slurm_keys_regex = r\"(^\\s*(%s)\\s*=\\s*)(.*)\" % \"|\".join(slurm_keys)\n slurmdbd_key_regex = r'(^\\s*(StoragePass)\\s*=\\s*)(.*)'\n\n sub = r'\\1********'\n\n self.do_file_sub(\n f\"{conf_paths}/slurm.conf\",\n slurm_keys_regex, sub\n )\n self.do_file_sub(\n f\"{conf_paths}/slurmdbd.conf\",\n slurmdbd_key_regex, sub\n )\n","repo_name":"sosreport/sos","sub_path":"sos/report/plugins/slurm.py","file_name":"slurm.py","file_ext":"py","file_size_in_byte":3178,"program_lang":"python","lang":"en","doc_type":"code","stars":470,"dataset":"github-code","pt":"7"} +{"seq_id":"120836573","text":"\"\"\"Tests download_pdfs.py.\"\"\"\n\nfrom unittest.mock import MagicMock, PropertyMock, patch\n\nimport pytest\nfrom bs4 import BeautifulSoup\n\nfrom download_pdfs import (\n format_chapter_link,\n format_chapter_name,\n get_chapter_names_from_soup,\n get_chapter_pdf_links_from_soup,\n get_chapters,\n get_filename,\n make_request,\n)\n\n\nclass TestHanserDownload:\n \"\"\"Class to test download_pdfs.py.\"\"\"\n\n def setup_class(self):\n \"\"\"Setup class getting a bs4 for the test html file.\"\"\"\n with open(\"tests/Big Data in der Praxis.html\", encoding=\"utf-8\") as soup_file:\n self.soup = BeautifulSoup(soup_file, \"html.parser\")\n\n def teardown_class(self):\n \"\"\"Set bs4 to None.\"\"\"\n self.soup = None\n\n def test_format_chapter_name(self):\n \"\"\"Tests format_chapter_name.\"\"\"\n tag = BeautifulSoup(\"
Big Data in der Praxis
\", \"html.parser\").h5\n expected_result = \"Big Data in der Praxis\"\n result = format_chapter_name(tag)\n assert isinstance(result, str)\n assert result == expected_result\n\n def test_get_chapter_names_from_soup(self):\n \"\"\"Tests get_chapter_names_from_soup.\"\"\"\n expected_list = [\n \"Big Data in der Praxis\",\n \"Einleitung\",\n \"Big Data\",\n \"Hadoop\",\n \"Das Hadoop-Ecosystem\",\n \"NoSQL und HBase\",\n \"Data Warehousing mit Hive\",\n \"Big-Data-Visualisierung\",\n \"Auf dem Weg zu neuem Wissen – Aufbereiten, Anreichern und Empfehlen\", # noqa: RUF001\n \"Infrastruktur\",\n \"Programmiersprachen\",\n \"Polyglot Persistence\",\n \"Apache Kafka\",\n \"Data Processing Engines\",\n \"Streaming\",\n \"Data Governance\",\n \"Zusammenfassung und Ausblick\",\n \"Häufige Fehler\",\n \"Anleitungen\",\n \"Literaturverzeichnis\",\n \"Index\",\n ]\n result = get_chapter_names_from_soup(self.soup)\n assert isinstance(result, list)\n assert isinstance(result[0], str)\n assert len(result) == 21\n assert result == expected_list\n\n def test_format_chapter_link(self):\n \"\"\"Tests format_chapter_link.\"\"\"\n tag = BeautifulSoup(\n ''\n 'PDF',\n \"html.parser\",\n ).a\n expected_result = (\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.012\"\n )\n result = format_chapter_link(tag)\n assert isinstance(result, str)\n assert result == expected_result\n\n def test_get_chapter_pdf_links_from_soup(self):\n \"\"\"Tests get_chapter_pdf_links_from_soup.\"\"\"\n expected_list = [\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.fm\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.001\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.002\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.003\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.004\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.005\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.006\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.007\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.008\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.009\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.010\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.011\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.012\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.013\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.014\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.015\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.016\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.017\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.018\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.019\",\n \"https://www.hanser-elibrary.com/doi/pdf/10.3139/9783446456013.bm\",\n ]\n result = get_chapter_pdf_links_from_soup(self.soup)\n assert isinstance(result, list)\n assert isinstance(result[0], str)\n assert len(result) == 21\n assert result == expected_list\n\n @patch(\"download_pdfs.get_chapter_names_from_soup\")\n @patch(\"download_pdfs.get_chapter_pdf_links_from_soup\")\n def test_get_chapter(self, link_mock: MagicMock, name_mock: MagicMock):\n \"\"\"Tests get_chapter.\"\"\"\n link_mock.side_effect = [[\"/1\", \"/2\", \"/3\"], [\"/1\", \"/2\", \"/3\"]]\n name_mock.side_effect = [[\"a\", \"b\", \"c\"], [\"/1\", \"/2\"]]\n result1 = get_chapters(self.soup)\n assert isinstance(result1, zip)\n assert list(result1) == [(1, \"a\", \"/1\"), (2, \"b\", \"/2\"), (3, \"c\", \"/3\")]\n result2 = get_chapters(self.soup)\n assert isinstance(result2, zip)\n assert list(result2) == [(1, \"\", \"/1\"), (2, \"\", \"/2\"), (3, \"\", \"/3\")]\n\n def test_get_filename(self):\n \"\"\"Tests get_filename.\"\"\"\n result = get_filename(3, \"test\")\n assert isinstance(result, str)\n assert result == \"003_test.pdf\"\n result = get_filename(23, \"Was tun?\")\n assert isinstance(result, str)\n assert result == \"023_Was tun.pdf\"\n result = get_filename(12, \"\")\n assert isinstance(result, str)\n assert result == \"012.pdf\"\n\n @patch(\"download_pdfs.requests.get\")\n def test_bad_response(self, get_mock: MagicMock):\n \"\"\"Tests treatment of bad response.\"\"\"\n dummy_status_code = 999\n dummy_reason = \"Test failed successfully!\"\n type(get_mock.return_value).ok = PropertyMock(return_value=False)\n type(get_mock.return_value).status_code = PropertyMock(\n return_value=dummy_status_code\n )\n type(get_mock.return_value).reason = PropertyMock(return_value=dummy_reason)\n dummy_url = \"dummy_url\"\n with pytest.raises(\n SystemExit,\n match=f\".*{dummy_url}.*{dummy_status_code}.*{dummy_reason}.*Exiting!\",\n ):\n make_request(dummy_url)\n\n @patch(\"download_pdfs.requests.get\")\n def test_good_response(self, get_mock: MagicMock):\n \"\"\"Tests treatment of good response.\"\"\"\n dummy_url = \"dummy_url\"\n dummy_status_code = 200\n type(get_mock.return_value).ok = PropertyMock(return_value=True)\n type(get_mock.return_value).status_code = PropertyMock(\n return_value=dummy_status_code\n )\n test_response = make_request(dummy_url)\n assert test_response.status_code == dummy_status_code\n assert test_response.ok is True\n","repo_name":"JanEricNitschke/HanserDownloader","sub_path":"tests/test_download_pdfs.py","file_name":"test_download_pdfs.py","file_ext":"py","file_size_in_byte":7215,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"11057133941","text":"import csv\nimport json\nfrom copy import copy\n\n\nclass ItemAndQty:\n\n # This constructor only has three data members.\n\n def __init__(self, item_name, price, quantity):\n self.item_name = str(item_name)\n self.price = float(price)\n self.quantity = int(quantity)\n\n# This method prints the data members.\n\n def __repr__(self):\n s = \"The item name : \" + str(self.item_name) + \".\\n\"\n s += \"The item price : \" + str(self.price) + \".\\n\"\n s += \"The quantity : \" + str(self.quantity) + \".\"\n return s\n\n# The cost method returns the total cost of items by price and quantity multiplication.\n\n def cost(self):\n multiplication_result = self.price * self.quantity\n return multiplication_result\n\n\nclass Shop:\n\n # This constractor has one data member a dictionary\n\n def __init__(self):\n\n self.data_member = {}\n\n\n# This method create Item not exists or increment quantity by quantity supplied if it does.\n\n\n def add_item_and_qty(self, item):\n if item.get(\"item_name\") in self.data_member.keys():\n self.data_member.get(item.get(\"item_name\")\n ).quantity += item.get(\"quantity\")\n else:\n item_and_qty = ItemAndQty(\n item.get(\"item_name\"), item.get(\"price\"), item.get(\"quantity\")\n )\n data_member = {item_and_qty.item_name: item_and_qty}\n self.data_member.update(data_member)\n\n# This method load the stock from csv file\n\n def load_initial_stock(self, csv_file_path):\n\n stock_list = self.csv_to_json(csv_file_path)\n for item in stock_list:\n self.add_item_and_qty(item)\n\n # This method loads the csv in json format\n\n def csv_to_json(self, csv_file_path):\n item_list = []\n with open(csv_file_path, encoding=\"utf-8\") as csvf:\n csv_reader = csv.DictReader(csvf)\n\n for raw in csv_reader:\n item_dict = {}\n item_dict[\"item_name\"] = raw[\"item_name\"]\n item_dict[\"price\"] = float(raw[\"price\"])\n item_dict[\"quantity\"] = int(raw[\"quantity\"])\n item_list.append(item_dict)\n return item_list\n\n# This method returns ItemAndQty object by name or None if not exists\n\n def item_and_qty_by_name(self, item_name):\n item = self.data_member.get(item_name)\n\n if item:\n return item\n return None\n\n# This method returns quantity of the item by name or 0 if not exists\n\n def items_in_stock(self, item_name):\n item = self.data_member.get(item_name)\n if item:\n return self.item_and_qty_by_name(item_name).quantity\n return 0\n\n\nclass ShoppingBasket:\n def __init__(self, shop):\n self.shop = shop\n self.shop.load_initial_stock(\"./stock.csv\")\n self.basket = []\n self.total_cost = 0\n\n # This method add items to the basket if they are in the shop or return 0 if not\n def add_item_and_qty(self, item_name, quantity):\n item = self.shop.item_and_qty_by_name(item_name)\n\n # if the item exist in the shop\n if item:\n # if the item exist in the basket\n if item in self.basket:\n basket_item = self.basket[self.basket.index(item)]\n\n basket_item.quantity += quantity\n\n # if item is not exist in the basket\n else:\n new_item = copy(item)\n new_item.quantity = quantity\n self.basket.append(new_item)\n item.quantity -= quantity\n\n return 0\n\n# This method calculate the total cost of items in the basket\n\n def calculate_total_cost(self):\n for item in self.basket:\n self.total_cost += item.cost()\n return self.total_cost\n\n# This method clears all items in the basket\n def clear_items(self):\n self.basket.clear()\n\n\nif __name__ == \"__main__\":\n\n shop = Shop()\n shop.load_initial_stock(\"./stock.csv\")\n","repo_name":"mohamedlafiteh/strath-uni-problems-and-solution","sub_path":"week-quesions-04/item_and_qt.py","file_name":"item_and_qt.py","file_ext":"py","file_size_in_byte":3981,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"30890649110","text":"#https://www.acmicpc.net/problem/20056\n\nfrom copy import deepcopy\n\nN, M, K = map(int, input().split())\nboard = [[[] for _ in range(N)] for _ in range(N)]\n\n\nfor _ in range(M):\n r, c, m, s, d = map(int, input().split())\n if m != 0:\n board[r - 1][c - 1].append([m, s, d])\n\ndirs = [[-1, 0], [-1, 1], [0, 1], [1, 1], [1, 0], [1, -1], [0, -1], [-1, -1]]\n\nfor _ in range(K):\n n_board = [[[] for _ in range(N)] for _ in range(N)]\n \n # 파이어볼 이동\n for y in range(N):\n for x in range(N):\n if board[y][x] != []:\n for b in range(len(board[y][x])):\n nm, ns, nd = board[y][x][b]\n\n ny, nx = y + dirs[nd][0] * ns, x + dirs[nd][1] * ns\n\n #격자 밖을 벗어나느게 아니라 순환시스템 이라고 생각\n nx = (nx + N) % N\n ny = (ny + N) % N\n n_board[ny][nx].append([nm, ns, nd])\n \n\n # 2개 이상의 파이어볼이 있는 칸을 찾아서 4개의 파이어볼 만들기.\n for y2 in range(N):\n for x2 in range(N):\n if len(n_board[y2][x2]) > 1:\n cm, cs, cd = 0, 0, []\n cnt = len(n_board[y2][x2])\n for c in range(cnt):\n cm += n_board[y2][x2][c][0]\n cs += n_board[y2][x2][c][1]\n cd.append(n_board[y2][x2][c][2] % 2) #홀짝\n cm = int(cm / 5)\n cs = int(cs / cnt)\n n_board[y2][x2] = []\n if cm != 0: # 질량이 0 인 경우 소멸\n if sum(cd) == 0 or sum(cd) == cnt: # 합쳐지는 파이어볼 방향이 모두 홀수거나 짝수인 경우\n for i in range(4):\n n_board[y2][x2].append([cm, cs, i * 2])\n else:\n for j in range(4):\n n_board[y2][x2].append([cm, cs, j * 2 + 1])\n\n board = deepcopy(n_board)\n\n\n# 남아있는 파이어볼 질량의 합 구하기\nsum_m = 0\nfor y in range(N):\n for x in range(N):\n if board[y][x] != []:\n for b in range(len(board[y][x])):\n sum_m += board[y][x][b][0]\nprint(sum_m)\n","repo_name":"comeheredart/algorithm_python","sub_path":"simulation/마법사 상어와 파이어볼.py","file_name":"마법사 상어와 파이어볼.py","file_ext":"py","file_size_in_byte":2239,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26078965405","text":"from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nfrom core.config import config as cfg\nimport cv2\nimport math\nimport numpy as np\n\n\ncv2.ocl.setUseOpenCL(False)\n\n\ndef CHW2HWC(image):\n return image.transpose([1, 2, 0])\n\n\ndef HWC2CHW(image):\n return image.transpose([2, 0, 1])\n\n\n# Image should be in CHW format.\ndef color_normalization(img, mean, stddev):\n assert len(mean) == img.shape[0], \\\n 'channel mean not computed properly'\n assert len(stddev) == img.shape[0], \\\n 'channel stddev not computed properly'\n for idx in range(img.shape[0]):\n img[idx] = img[idx] - mean[idx]\n img[idx] = img[idx] / stddev[idx]\n return img\n\n\ndef pad_image(pad_size, image, order='CHW'):\n if order == 'CHW':\n img = np.pad(\n image, ((0, 0), (pad_size, pad_size), (pad_size, pad_size)),\n mode=str('constant')\n )\n elif order == 'HWC':\n img = np.pad(\n image, ((pad_size, pad_size), (pad_size, pad_size), (0, 0)),\n mode=str('constant')\n )\n return img\n\n\ndef flip_boxes(boxes, im_width):\n boxes_flipped = boxes.copy()\n boxes_flipped[:, 0::4] = im_width - boxes[:, 2::4] - 1\n boxes_flipped[:, 2::4] = im_width - boxes[:, 0::4] - 1\n return boxes_flipped\n\n\ndef clip_boxes_to_image(boxes, height, width):\n \"\"\"Clip an array of boxes to an image with the given height and width.\"\"\"\n boxes[:, [0, 2]] = np.minimum(width - 1., np.maximum(0., boxes[:, [0, 2]]))\n boxes[:, [1, 3]] = np.minimum(height - 1., np.maximum(0., boxes[:, [1, 3]]))\n return boxes\n\n\ndef horizontal_flip_list(prob, images, order='CHW', boxes=None,\n force_flip=False):\n\n _, width, _ = images[0].shape\n if np.random.uniform() < prob or force_flip:\n if boxes is not None:\n boxes = flip_boxes(boxes, width)\n if order == 'CHW':\n out_images = []\n for image in images:\n image = np.asarray(image).swapaxes(2, 0)\n image = image[::-1]\n out_images.append(image.swapaxes(0, 2))\n return out_images, boxes\n elif order == 'HWC':\n return [cv2.flip(image, 1) for image in images], boxes\n return images, boxes\n\n\ndef crop_boxes(boxes, x_offset, y_offset):\n boxes[:, [0, 2]] = boxes[:, [0, 2]] - x_offset\n boxes[:, [1, 3]] = boxes[:, [1, 3]] - y_offset\n return boxes\n\n\ndef random_crop_list(images, size, pad_size=0, order='CHW', boxes=None):\n if pad_size > 0:\n raise NotImplementedError()\n images = [pad_image(pad_size=pad_size, image=image, order=order)\n for image in images]\n\n if order == 'CHW':\n if images[0].shape[1] == size and images[0].shape[2] == size:\n return images, boxes\n height = images[0].shape[1]\n width = images[0].shape[2]\n y_offset = 0\n if height > size:\n y_offset = int(np.random.randint(0, height - size))\n x_offset = 0\n if width > size:\n x_offset = int(np.random.randint(0, width - size))\n cropped = [image[:, y_offset:y_offset + size, x_offset:x_offset + size]\n for image in images]\n assert cropped[0].shape[1] == size, \"Image not cropped properly\"\n assert cropped[0].shape[2] == size, \"Image not cropped properly\"\n elif order == 'HWC':\n if images[0].shape[0] == size and images[0].shape[1] == size:\n return images, boxes\n height = images[0].shape[0]\n width = images[0].shape[1]\n y_offset = 0\n if height > size:\n y_offset = int(np.random.randint(0, height - size))\n x_offset = 0\n if width > size:\n x_offset = int(np.random.randint(0, width - size))\n cropped = [image[y_offset:y_offset + size, x_offset:x_offset + size, :]\n for image in images]\n assert cropped[0].shape[0] == size, \"Image not cropped properly\"\n assert cropped[0].shape[1] == size, \"Image not cropped properly\"\n\n if boxes is not None:\n boxes = crop_boxes(boxes, x_offset, y_offset)\n return cropped, boxes\n\n\ndef center_crop(size, image):\n height = image.shape[0]\n width = image.shape[1]\n y_offset = int(math.ceil((height - size) / 2))\n x_offset = int(math.ceil((width - size) / 2))\n cropped = image[y_offset:y_offset + size, x_offset:x_offset + size, :]\n assert cropped.shape[0] == size, \"Image height not cropped properly\"\n assert cropped.shape[1] == size, \"Image width not cropped properly\"\n return cropped\n\n\ndef spatial_shift_crop_list(size, images, spatial_shift_pos, boxes=None):\n assert spatial_shift_pos in [0, 1, 2]\n\n height = images[0].shape[0]\n width = images[0].shape[1]\n y_offset = int(math.ceil((height - size) / 2))\n x_offset = int(math.ceil((width - size) / 2))\n\n if height > width:\n if spatial_shift_pos == 0:\n y_offset = 0\n elif spatial_shift_pos == 2:\n y_offset = height - size\n else:\n if spatial_shift_pos == 0:\n x_offset = 0\n elif spatial_shift_pos == 2:\n x_offset = width - size\n\n cropped = [image[y_offset:y_offset + size, x_offset:x_offset + size, :]\n for image in images]\n assert cropped[0].shape[0] == size, \"Image height not cropped properly\"\n assert cropped[0].shape[1] == size, \"Image width not cropped properly\"\n\n if boxes is not None:\n boxes[:, [0, 2]] -= x_offset\n boxes[:, [1, 3]] -= y_offset\n\n return cropped, boxes\n\n\ndef scale(size, image):\n height = image.shape[0]\n width = image.shape[1]\n if ((width <= height and width == size)\n or (height <= width and height == size)):\n return image\n new_width = size\n new_height = size\n if width < height:\n new_height = int(math.floor((float(height) / width) * size))\n else:\n new_width = int(math.floor((float(width) / height) * size))\n img = cv2.resize(\n image, (new_width, new_height),\n interpolation=getattr(cv2, cfg.INTERPOLATION))\n return img.astype(np.float32)\n\n\n# Scale the smaller edge of image to size.\ndef scale_boxes(size, boxes, height, width):\n if ((width <= height and width == size)\n or (height <= width and height == size)):\n return boxes\n\n new_width = size\n new_height = size\n if width < height:\n new_height = int(math.floor((float(height) / width) * size))\n boxes *= (float(new_height) / height)\n else:\n new_width = int(math.floor((float(width) / height) * size))\n boxes *= (float(new_width) / width)\n return boxes\n\n\n# Scale the smaller edge of image to a scale from, e.g. 1 / [1/320, 1/256].\n# Image should be in HWC format.\ndef random_short_side_scale_jitter_list(images, min_size,\n max_size, boxes=None):\n\n size = int(round(1.0 / np.random.uniform(1.0 / max_size, 1.0 / min_size)))\n\n height = images[0].shape[0]\n width = images[0].shape[1]\n if ((width <= height and width == size)\n or (height <= width and height == size)):\n return images, boxes\n new_width = size\n new_height = size\n if width < height:\n new_height = int(math.floor((float(height) / width) * size))\n if boxes is not None:\n boxes = boxes * float(new_height) / height\n else:\n new_width = int(math.floor((float(width) / height) * size))\n if boxes is not None:\n boxes = boxes * float(new_width) / width\n return [cv2.resize(image, (new_width, new_height),\n interpolation=getattr(cv2, cfg.INTERPOLATION)\n ).astype(np.float32)\n for image in images], boxes\n\n\n# Image should be in CHW format and BGR channels.\ndef lighting_list(imgs, alphastd, eigval, eigvec, alpha=None):\n if alphastd == 0:\n return imgs\n alpha = np.random.normal(0, alphastd, size=(1, 3))\n eig_vec = np.array(eigvec)\n eig_val = np.reshape(eigval, (1, 3))\n rgb = np.sum(\n eig_vec * np.repeat(alpha, 3, axis=0) * np.repeat(\n eig_val, 3, axis=0),\n axis=1\n )\n out_images = []\n for img in imgs:\n for idx in range(img.shape[0]):\n img[idx] = img[idx] + rgb[2 - idx]\n out_images.append(img)\n return out_images\n\n\ndef blend(image1, image2, alpha):\n return image1 * alpha + image2 * (1 - alpha)\n\n\n# Image should be in CHW format and BGR channels.\ndef grayscale(image):\n img_gray = np.copy(image)\n gray_channel = 0.299 * image[2] + 0.587 * image[1] + 0.114 * image[0]\n img_gray[0] = gray_channel\n img_gray[1] = gray_channel\n img_gray[2] = gray_channel\n return img_gray\n\n\ndef saturation_list(var, images):\n alpha = 1.0 + np.random.uniform(-var, var)\n\n out_images = []\n for image in images:\n img_gray = grayscale(image)\n out_images.append(blend(image, img_gray, alpha))\n return out_images\n\n\ndef brightness_list(var, images):\n alpha = 1.0 + np.random.uniform(-var, var)\n\n out_images = []\n for image in images:\n img_bright = np.zeros(image.shape)\n out_images.append(blend(image, img_bright, alpha))\n return out_images\n\n\ndef contrast_list(var, images):\n alpha = 1.0 + np.random.uniform(-var, var)\n\n out_images = []\n for image in images:\n img_gray = grayscale(image)\n img_gray.fill(np.mean(img_gray[0]))\n out_images.append(blend(image, img_gray, alpha))\n return out_images\n\n\ndef color_jitter_list(images, img_brightness=0, img_contrast=0, img_saturation=0):\n\n jitter = []\n if img_brightness != 0:\n jitter.append('brightness')\n if img_contrast != 0:\n jitter.append('contrast')\n if img_saturation != 0:\n jitter.append('saturation')\n\n if len(jitter) > 0:\n order = np.random.permutation(np.arange(len(jitter)))\n for idx in range(0, len(jitter)):\n if jitter[order[idx]] == 'brightness':\n images = brightness_list(img_brightness, images)\n elif jitter[order[idx]] == 'contrast':\n images = contrast_list(img_contrast, images)\n elif jitter[order[idx]] == 'saturation':\n images = saturation_list(img_saturation, images)\n return images\n","repo_name":"facebookresearch/video-long-term-feature-banks","sub_path":"lib/datasets/image_processor.py","file_name":"image_processor.py","file_ext":"py","file_size_in_byte":10343,"program_lang":"python","lang":"en","doc_type":"code","stars":370,"dataset":"github-code","pt":"7"} +{"seq_id":"10726035284","text":"\"\"\"Game handler\"\"\"\n\nfrom resources.handlers import BaseRequestHandler\n\nfrom server.resources.models import Game, GameListSerializer, GameSerializer\n\n\nclass GameHandler(BaseRequestHandler):\n \"\"\"\n Games request handler.\n Let get games list, and game details.\n \"\"\"\n\n async def get(self, game_name: str = \"\") -> None:\n \"\"\"Get game or all games endpoint\"\"\"\n if not game_name:\n serializer = await GameListSerializer.from_queryset(Game.all())\n else:\n game = await Game.get(name=game_name)\n serializer = await GameSerializer.from_tortoise_orm(game)\n self.write(serializer.json())\n","repo_name":"ar0ne/bg-server","sub_path":"server/handlers/game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"4718207971","text":"import numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy.stats import norm, expon, lognorm, weibull_min, gamma\nimport sys\nimport os\n\n# Get the current script's directory\ncurrent_directory = os.path.dirname(os.path.realpath(__file__))\n\n# Get the parent directory\nparent_directory = os.path.abspath(os.path.join(current_directory, os.pardir))\n\n# Append the parent directory to sys.path\nsys.path.append(parent_directory)\n\n\nfrom models.approximation.generate_summaries import *\nfrom models.approximation.mle import *\nfrom distr_parser.parserfunction import *\n# Function to parse distribution strings\n# Gibbs sampler for two distributions\ndef gibbs_sampler_two_distrbution(iterations, distribution_X, distribution_Y, operator='+'):\n x_samples = distribution_X.rvs(iterations)\n y_samples = distribution_Y.rvs(iterations)\n\n if operator == '+':\n return x_samples + y_samples\n elif operator == '-':\n return x_samples - y_samples\n elif operator == '*':\n return x_samples * y_samples\n elif operator == '/':\n return x_samples / y_samples\n\n# Process the input object\ndef process_input_object(input_object, iterations):\n vector1_str = input_object['vector1']\n vector2_str = input_object['vector2']\n operations = input_object['operations']\n \n vector1_distr = [parse_distribution(d) for d in vector1_str]\n vector2_distr = [parse_distribution(d) for d in vector2_str]\n \n # Initialize an array for the accumulated result\n accumulated_result = np.zeros(iterations)\n \n for v1, v2, op in zip(vector1_distr, vector2_distr, operations):\n result_vector = gibbs_sampler_two_distrbution(iterations, v1, v2, operator=op)\n \n # Accumulate the result component-wise\n accumulated_result += result_vector\n \n return accumulated_result\n\n# Define the input object\ninput_object = {\n 'vector1': ['NORM(20,4)', 'NORM(50,6)', 'NORM(20,5)'],\n 'vector2': ['NORM(30,2)', 'NORM(30,5)', 'NORM(60,6)'],\n 'operations': ['*', '*', '*']\n}\n\niterations = 50000\naccumulated_result = process_input_object(input_object, iterations)\n\n# Plotting the accumulated distribution\nplt.hist(accumulated_result, bins=50, density=True, alpha=0.6, label='Accumulated Result')\nplt.xlabel('Value')\nplt.ylabel('Frequency')\nplt.legend()\nplt.show()","repo_name":"Asad1287/montecarlo_planner","sub_path":"monte_carlo_planner/simulation/monte_carlo_vector.py","file_name":"monte_carlo_vector.py","file_ext":"py","file_size_in_byte":2305,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"19796174447","text":"'''\n@author : Rohithmarktricks\nThis program counts the number of vowels in the given string.\n'''\nSTRT = input(\"Enter the string :\")\nSTRT2 = STRT.lower()\nCOUNT = 0\nfor I in STRT2:\n if(I == 'a' or I == 'e' or I == 'i' or I == 'o' or I == 'u'):\n COUNT = COUNT+1\nprint(\"Number of vowels :\",COUNT)\n","repo_name":"Rohithmarktricks/20186097_CSPP-1","sub_path":"cspp1-practice/m3/vowels.py","file_name":"vowels.py","file_ext":"py","file_size_in_byte":304,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"2750290053","text":"try:\n\timport numpy as np\n\thas_numpy = True\nexcept ImportError:\n\timport math\n\thas_numpy = False\ntry:\n\timport scipy.constants\n\thas_scipy = True\nexcept ImportError:\n\thas_scipy = False\nimport operator as op\nfrom .similar import sim, nsim, gsim, lsim\n\ndef equation_extend(core):\n\tdef product(*args):\n\t\tif len(args) == 1 and has_numpy:\n\t\t\treturn np.prod(args[0])\n\t\telse:\n\t\t\treturn reduce(op.mul,args,1)\n\n\tdef sumargs(*args):\n\t\tif len(args) == 1:\n\t\t\treturn sum(args[0])\n\t\telse:\n\t\t\treturn sum(args)\n\n\tcore.addOp('+',\"({0:s} + {1:s})\",\"\\\\left({0:s} + {1:s}\\\\right)\",False,3,op.add)\n\tcore.addOp('-',\"({0:s} - {1:s})\",\"\\\\left({0:s} - {1:s}\\\\right)\",False,3,op.sub)\n\tcore.addOp('*',\"({0:s} * {1:s})\",\"\\\\left({0:s} \\\\times {1:s}\\\\right)\",False,2,op.mul)\n\tcore.addOp('/',\"({0:s} / {1:s})\",\"\\\\frac{{{0:s}}}{{{1:s}}}\",False,2,op.truediv)\n\tcore.addOp('%',\"({0:s} % {1:s})\",\"\\\\left({0:s} \\\\bmod {1:s}\\\\right)\",False,2,op.mod)\n\tcore.addOp('^',\"({0:s} ^ {1:s})\",\"{0:s}^{{{1:s}}}\",False,1,op.pow)\n\tcore.addOp('**',\"({0:s} ^ {1:s})\",\"{0:s}^{{{1:s}}}\",False,1,op.pow)\n\tcore.addOp('&',\"({0:s} & {1:s})\",\"\\\\left({0:s} \\\\land {1:s}\\\\right)\",False,4,op.and_)\n\tcore.addOp('|',\"({0:s} | {1:s})\",\"\\\\left({0:s} \\\\lor {1:s}\\\\right)\",False,4,op.or_)\n\tcore.addOp('',\"({0:s} {1:s})\",\"\\\\left({0:s} \\\\oplus {1:s}\\\\right)\",False,4,op.xor)\n\tcore.addOp('&|',\"({0:s} {1:s})\",\"\\\\left({0:s} \\\\oplus {1:s}\\\\right)\",False,4,op.xor)\n\tcore.addOp('|&',\"({0:s} {1:s})\",\"\\\\left({0:s} \\\\oplus {1:s}\\\\right)\",False,4,op.xor)\n\tcore.addOp('==',\"({0:s} == {1:s})\",\"\\\\left({0:s} = {1:s}\\\\right)\",False,5,op.eq)\n\tcore.addOp('=',\"({0:s} == {1:s})\",\"\\\\left({0:s} = {1:s}\\\\right)\",False,5,op.eq)\n\tcore.addOp('~',\"({0:s} ~ {1:s})\",\"\\\\left({0:s} \\\\approx {1:s}\\\\right)\",False,5,sim)\n\tcore.addOp('!~',\"({0:s} !~ {1:s})\",\"\\\\left({0:s} \\\\not\\\\approx {1:s}\\\\right)\",False,5,nsim)\n\tcore.addOp('!=',\"({0:s} != {1:s})\",\"\\\\left({0:s} \\\\neg {1:s}\\\\right)\",False,5,op.ne)\n\tcore.addOp('<>',\"({0:s} != {1:s})\",\"\\\\left({0:s} \\\\neg {1:s}\\\\right)\",False,5,op.ne)\n\tcore.addOp('><',\"({0:s} != {1:s})\",\"\\\\left({0:s} \\\\neg {1:s}\\\\right)\",False,5,op.ne)\n\tcore.addOp('<',\"({0:s} < {1:s})\",\"\\\\left({0:s} < {1:s}\\\\right)\",False,5,op.lt)\n\tcore.addOp('>',\"({0:s} > {1:s})\",\"\\\\left({0:s} > {1:s}\\\\right)\",False,5,op.gt)\n\tcore.addOp('<=',\"({0:s} <= {1:s})\",\"\\\\left({0:s} \\\\leq {1:s}\\\\right)\",False,5,op.le)\n\tcore.addOp('>=',\"({0:s} >= {1:s})\",\"\\\\left({0:s} \\\\geq {1:s}\\\\right)\",False,5,op.ge)\n\tcore.addOp('=<',\"({0:s} <= {1:s})\",\"\\\\left({0:s} \\\\leq {1:s}\\\\right)\",False,5,op.le)\n\tcore.addOp('=>',\"({0:s} >= {1:s})\",\"\\\\left({0:s} \\\\geq {1:s}\\\\right)\",False,5,op.ge)\n\tcore.addOp('<~',\"({0:s} <~ {1:s})\",\"\\\\left({0:s} \\lessapprox {1:s}\\\\right)\",False,5,lsim)\n\tcore.addOp('>~',\"({0:s} >~ {1:s})\",\"\\\\left({0:s} \\\\gtrapprox {1:s}\\\\right)\",False,5,gsim)\n\tcore.addOp('~<',\"({0:s} <~ {1:s})\",\"\\\\left({0:s} \\lessapprox {1:s}\\\\right)\",False,5,lsim)\n\tcore.addOp('~>',\"({0:s} >~ {1:s})\",\"\\\\left({0:s} \\\\gtrapprox {1:s}\\\\right)\",False,5,gsim)\n\tcore.addUnaryOp('!',\"(!{0:s})\",\"\\\\neg{0:s}\",op.not_)\n\tcore.addUnaryOp('-',\"-{0:s}\",\"-{0:s}\",op.neg)\n\tcore.addFn('abs',\"abs({0:s})\",\"\\\\left|{0:s}\\\\right|\",1,op.abs)\n\tcore.addFn('sum',\"sum({0:s})\",\"\\\\sum\\\\left({0:s}\\\\right)\",'+',sumargs)\n\tcore.addFn('prod',\"prod({0:s})\",\"\\\\prod\\\\left({0:s}\\\\right)\",'+',product)\n\tif has_numpy:\n\t\tcore.addFn('floor',\"floor({0:s})\",\"\\\\lfloor {0:s} \\\\rfloor\",1,np.floor)\n\t\tcore.addFn('ceil',\"ceil({0:s})\",\"\\\\lceil {0:s} \\\\rceil\",1,np.ceil)\n\t\tcore.addFn('round',\"round({0:s})\",\"\\\\lfloor {0:s} \\\\rceil\",1,np.round)\n\t\tcore.addFn('sin',\"sin({0:s})\",\"\\\\sin\\\\left({0:s}\\\\right)\",1,np.sin)\n\t\tcore.addFn('cos',\"cos({0:s})\",\"\\\\cos\\\\left({0:s}\\\\right)\",1,np.cos)\n\t\tcore.addFn('tan',\"tan({0:s})\",\"\\\\tan\\\\left({0:s}\\\\right)\",1,np.tan)\n\t\tcore.addFn('re',\"re({0:s})\",\"\\\\Re\\\\left({0:s}\\\\right)\",1,np.real)\n\t\tcore.addFn('im',\"re({0:s})\",\"\\\\Im\\\\left({0:s}\\\\right)\",1,np.imag)\n\t\tcore.addFn('sqrt',\"sqrt({0:s})\",\"\\\\sqrt{{{0:s}}}\",1,np.sqrt)\n\t\tcore.addConst(\"pi\",np.pi)\n\t\tcore.addConst(\"e\",np.e)\n\t\tcore.addConst(\"Inf\",np.Inf)\n\t\tcore.addConst(\"NaN\",np.NaN)\n\telse:\n\t\tcore.addFn('floor',\"floor({0:s})\",\"\\\\lfloor {0:s} \\\\rfloor\",1,math.floor)\n\t\tcore.addFn('ceil',\"ceil({0:s})\",\"\\\\lceil {0:s} \\\\rceil\",1,math.ceil)\n\t\tcore.addFn('round',\"round({0:s})\",\"\\\\lfloor {0:s} \\\\rceil\",1,round)\n\t\tcore.addFn('sin',\"sin({0:s})\",\"\\\\sin\\\\left({0:s}\\\\right)\",1,math.sin)\n\t\tcore.addFn('cos',\"cos({0:s})\",\"\\\\cos\\\\left({0:s}\\\\right)\",1,math.cos)\n\t\tcore.addFn('tan',\"tan({0:s})\",\"\\\\tan\\\\left({0:s}\\\\right)\",1,math.tan)\n\t\tcore.addFn('re',\"re({0:s})\",\"\\\\Re\\\\left({0:s}\\\\right)\",1,complex.real)\n\t\tcore.addFn('im',\"re({0:s})\",\"\\\\Im\\\\left({0:s}\\\\right)\",1,complex.imag)\n\t\tcore.addFn('sqrt',\"sqrt({0:s})\",\"\\\\sqrt{{{0:s}}}\",1,math.sqrt)\n\t\tcore.addConst(\"pi\",math.pi)\n\t\tcore.addConst(\"e\",math.e)\n\t\tcore.addConst(\"Inf\",float(\"Inf\"))\n\t\tcore.addConst(\"NaN\",float(\"NaN\"))\n\tif has_scipy:\n\t\tcore.addConst(\"h\",scipy.constants.h)\n\t\tcore.addConst(\"hbar\",scipy.constants.hbar)\n\t\tcore.addConst(\"m_e\",scipy.constants.m_e)\n\t\tcore.addConst(\"m_p\",scipy.constants.m_p)\n\t\tcore.addConst(\"m_n\",scipy.constants.m_n)\n\t\tcore.addConst(\"c\",scipy.constants.c)\n\t\tcore.addConst(\"N_A\",scipy.constants.N_A)\n\t\tcore.addConst(\"mu_0\",scipy.constants.mu_0)\n\t\tcore.addConst(\"eps_0\",scipy.constants.epsilon_0)\n\t\tcore.addConst(\"k\",scipy.constants.k)\n\t\tcore.addConst(\"G\",scipy.constants.G)\n\t\tcore.addConst(\"g\",scipy.constants.g)\n\t\tcore.addConst(\"q\",scipy.constants.e)\n\t\tcore.addConst(\"R\",scipy.constants.R)\n\t\tcore.addConst(\"sigma\",scipy.constants.e)\n\t\tcore.addConst(\"Rb\",scipy.constants.Rydberg)","repo_name":"titanscouting/tra-analysis","sub_path":"analysis-master/tra_analysis/equation/parser/Hybrid_Utils/equation_base.py","file_name":"equation_base.py","file_ext":"py","file_size_in_byte":5491,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"18240195211","text":"import csv\nimport dateutil.parser as dtp\nimport django\nimport glob\nimport os\nimport sys\n\n\ncurr_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\nsys.path.append(curr_dir)\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"UnBreakiFi.settings\")\n\n\nif 'setup' in dir(django):\n django.setup()\n\nfrom SystemApp.models import Individual, Vendor, Transaction, Features\nimport SystemApp.manual_features as F\n\n\ndef parse():\n my_dir = os.path.dirname(os.path.realpath(__file__))\n file_lst = glob.glob(my_dir+'/transaction-data/*.csv')\n if file_lst:\n save_to_db(file_lst)\n\n\ndef save_to_db(file_lst, uploaded=False):\n i = 0\n l = len(file_lst)\n users = set()\n for path in file_lst:\n if uploaded:\n reader = csv.reader(path)\n user = read_and_save(reader)\n else:\n with open(path) as csvfile:\n reader = csv.reader(csvfile)\n user = read_and_save(reader)\n users.add(user)\n i += 1\n print_progress_bar(i, l, prefix='Progress:', suffix='Complete',\n length=50)\n return users\n\n\ndef update_features(user):\n if user:\n feature = Features(user=user)\n feature.student = F.is_student(user.auth_id)\n feature.has_kids = F.has_kids(user.auth_id)\n feature.student_loan = F.has_been_paying_student_loans(user.auth_id)\n feature.pets = F.has_pets(user.auth_id)\n feature.an_artist = F.is_an_artist(user.auth_id)\n feature.moved = F.is_moving(user.auth_id)\n feature.peaceful = F.likes_peace(user.auth_id)\n feature.proposing = F.is_proposing(user.auth_id)\n feature.athletic = F.is_athletic(user.auth_id)\n feature.divorced = F.is_divorced(user.auth_id)\n feature.outgoing = F.is_outgoing(user.auth_id)\n feature.figurine_stuffs = F.is_into_stuffs(user.auth_id)\n feature.student = F.is_student(user.auth_id)\n feature.save()\n\n\ndef read_and_save(reader):\n user = None\n next(reader, None)\n for row in reader:\n # the_date = None\n try:\n the_date = dtp.parse(row[1])\n except:\n given_date = row[1].split('/')\n the_date = dtp.parse(given_date[0]+'/31/'+given_date[2])\n if the_date:\n amount = row[3]\n location = row[4]\n user = Individual.objects.filter(auth_id=int(row[0]))\n if user:\n user = user[0]\n else:\n user = Individual.objects.create(auth_id=int(row[0]))\n\n vendor_name, transaction_name = parse_vendor(row[2])\n\n the_vendor = Vendor.objects.filter(store_name=vendor_name)\n if not the_vendor:\n the_vendor = Vendor.objects.create(store_name=vendor_name)\n else:\n the_vendor = the_vendor[0]\n\n Transaction.objects.create(\n user=user, date=the_date, amount=float(amount),\n location=location, name=transaction_name, vendor=the_vendor)\n update_features(user=user)\n return user.auth_id\n\n\ndef parse_vendor(name):\n name = name.strip()\n split_name = name.split('-')\n\n if len(split_name) == 1:\n return split_name[0], split_name[0]\n return (i.strip() for i in split_name)\n\n\n# Print iterations progress\ndef print_progress_bar(iteration, total, prefix='',\n suffix='', decimals=1, length=100, fill='█'):\n \"\"\"\n Call in a loop to create terminal progress bar\n @params:\n iteration - Required : current iteration (Int)\n total - Required : total iterations (Int)\n prefix - Optional : prefix string (Str)\n suffix - Optional : suffix string (Str)\n decimals - Optional : positive number of decimals in percent complete (Int)\n length - Optional : character length of bar (Int)\n fill - Optional : bar fill character (Str)\n \"\"\"\n percent = (\"{0:.\" + str(decimals) + \"f}\").format(100 * (iteration / float(total)))\n filled_length = int(length * iteration // total)\n bar = fill * filled_length + '-' * (length - filled_length)\n st = '\\r%s |%s| %s%% %s' % (prefix, bar, percent, suffix) + '\\r'\n print(st)\n # Print New Line on Complete\n if iteration == total:\n print()\n\n\ndef manual_flush():\n print(\"Deleting Individuals...\")\n Individual.objects.all().delete()\n print(\"Done!\\nDeleting Vendors\")\n Vendor.objects.all().delete()\n print(\"Done!\\nDatabase Emtpy now!!\")\n\n\nif __name__ == '__main__':\n if len(sys.argv) == 2:\n if sys.argv[1] == '--parse':\n parse()\n elif sys.argv[1] == '--delete':\n sure = input('Are you sure you want to delete?\\n[Y]es or [N]o: ')\n if sure.lower() == 'y':\n manual_flush()\n else:\n print('Not deleted')\n else:\n print('wrong argument given: ', sys.argv[1], '\\nUsage: \\n\\t--parse : parse csv files \\n\\t--delete : delete all the data from the database\\n')\n else:\n print('Must provide 1 argument')\n","repo_name":"hkdahal/Un-BreakiFi","sub_path":"interact.py","file_name":"interact.py","file_ext":"py","file_size_in_byte":5096,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"14717457192","text":"r\"\"\"Evaluation script.\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport os.path as osp\nimport pdb\nimport sys\nimport traceback\n\nfrom absl import flags\nfrom absl import logging\nfrom monty.collections import AttrDict\nimport sklearn.cluster\nimport tensorflow as tf\n\nfrom stacked_capsule_autoencoders.capsules.configs import data_config\nfrom stacked_capsule_autoencoders.capsules.configs import model_config\nfrom stacked_capsule_autoencoders.capsules.eval import cluster_classify\nfrom stacked_capsule_autoencoders.capsules.eval import collect_results\nfrom stacked_capsule_autoencoders.capsules.plot import make_tsne_plot\nfrom stacked_capsule_autoencoders.capsules.train import tools\nfrom stacked_capsule_autoencoders.capsules import capsule as _capsule\n\nflags.DEFINE_string('snapshot', '', 'Checkpoint file.')\nflags.DEFINE_string(\n 'tsne_figure_name', 'tsne.png', 'Filename for the TSNE '\n 'figure. It will be saved in the checkpoint folder.')\n\n# These two flags are necessary for model loading. Don't change them!\nflags.DEFINE_string('dataset', 'mnist', 'Don\\'t change!')\nflags.DEFINE_string('model', 'scae', 'Don\\'t change!.')\n\n\ndef _collect_results(sess, tensors, dataset, n_batches):\n \"\"\"Collects some tensors from many batches.\"\"\"\n\n for i in range(n_batches):\n print('\\rCollecting: {}/{}'.format(i + 1, n_batches), end='')\n if i == 20:\n sess.run(tensors)\n\n\ndef main(_=None):\n FLAGS = flags.FLAGS # pylint: disable=invalid-name,redefined-outer-name\n config = FLAGS\n FLAGS.__dict__['config'] = config\n\n # Build the graph\n with tf.Graph().as_default():\n\n model_dict = model_config.get(FLAGS)\n data_dict = data_config.get(FLAGS)\n\n model = model_dict.model\n trainset = data_dict.trainset\n validset = data_dict.validset\n\n # Optimisation target\n validset = tools.maybe_convert_dataset(validset)\n trainset = tools.maybe_convert_dataset(trainset)\n\n t1 = model(trainset)\n t2 = model(validset)\n\n sess = tf.Session()\n saver = tf.train.Saver()\n saver.restore(sess, FLAGS.snapshot)\n\n if config.dataset == 'mnist':\n _collect_results(sess, _capsule.save_op, validset,\n 10000 // FLAGS.batch_size)\n _collect_results(sess, _capsule.save_op, trainset,\n 60000 // FLAGS.batch_size)\n elif config.dataset == 'svhn':\n _collect_results(sess, _capsule.save_op, validset,\n 26032 // FLAGS.batch_size)\n\n _collect_results(sess, _capsule.save_op, trainset,\n 73257 // FLAGS.batch_size)\n elif config.dataset == 'cifar10':\n _collect_results(sess, _capsule.save_op, validset,\n 10000 // FLAGS.batch_size)\n\n _collect_results(sess, _capsule.save_op, trainset,\n 50000 // FLAGS.batch_size)\n\n\nif __name__ == '__main__':\n try:\n logging.set_verbosity(logging.INFO)\n tf.app.run()\n except Exception as err: # pylint: disable=broad-except\n FLAGS = flags.FLAGS\n\n last_traceback = sys.exc_info()[2]\n traceback.print_tb(last_traceback)\n print(err)\n pdb.post_mortem(last_traceback)\n","repo_name":"YuZiHanorz/stacked_capsule_autoencoders","sub_path":"test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":3297,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"36704150621","text":"\"\"\"\n Plug and Play ADMM for Poisson denoising\n Authors: XXX\n XXX\n Jialin Liu (danny19921123@gmail.com)\n\"\"\"\n\nimport os\nimport numpy as np\nfrom PIL import Image\nimport torch\nimport cv2\nimport glob\nfrom skimage.measure import compare_psnr\nfrom numpy import linalg as LA\nimport scipy.io as sio\nfrom utils.utils import load_model\nfrom utils.utils import psnr\nfrom utils.config import analyze_parse\n\ndef pnp_admm_poisson_denoising(noisy, denoiser, clean, **opts):\n \"\"\"\n Parameters:\n :noisy - the noisy observation.\n :denoiser - the Gaussian denoiser used in Plug-and-Play ADMM.\n :clean - the clean image used to monitor PSNR.\n :opts - the kwargs for hyperparameters in Plug-and-Play ADMM.\n :lam - the value of 1/alpha.\n :beta - the prior weight parameter.\n :maxitr - the max number of iterations.\n :verbose - a flag that enables/disables info printing.\n :peak - the peak value of the original clean image before scaling.\n :maxval - the max pixel value of the original clean image.\n - NOTE: if `peak` and `M` options exist in `opts`, then the `clean` image is the scaled version of the original image.\n \"\"\"\n\n \"\"\" Process parameters. \"\"\"\n\n lam = opts.get('lam', 10.0)\n beta = opts.get('beta', 1.0)\n maxitr = opts.get('maxitr', 50)\n peak = opts.get('peak', None)\n maxval = opts.get('maxval', None)\n data_range = opts.get('data_range', 1.0)\n verbose = opts.get('verbose', 1)\n outdir = opts.get('outdir', '.')\n\n\n \"\"\" Initialization. \"\"\"\n\n m, n = noisy.shape\n\n\n noisy_flat = np.reshape(noisy, -1)\n x = np.copy(noisy_flat)\n v = np.zeros_like(noisy_flat, dtype=np.float64)\n u = np.zeros_like(noisy_flat, dtype=np.float64)\n\n\n \"\"\" Main loop. \"\"\"\n\n for i in range(maxitr):\n\n # record the variables in the current iteration\n x_old = np.copy(x)\n v_old = np.copy(v)\n u_old = np.copy(u)\n\n \"\"\" proximal step. \"\"\"\n\n vtilde = np.copy((lam * (x + u) - 1.0) / lam)\n v = np.copy((vtilde + np.sqrt(np.square(vtilde) + 4.0*noisy_flat/lam)) / 2.0)\n\n \"\"\" denoising step. \"\"\"\n\n xtilde = np.copy(2*v - x_old - u_old)\n\n # scale xtilde to be in range of [0,1]\n mintmp = np.min(xtilde)\n maxtmp = np.max(xtilde)\n xtilde = (xtilde - mintmp) / (maxtmp - mintmp)\n\n # load to torch\n xtilde_torch = np.reshape(xtilde, (1,1,m,n))\n xtilde_torch = torch.from_numpy(xtilde_torch).type(torch.cuda.FloatTensor)\n\n # denoise \n r = denoiser(xtilde_torch).cpu().numpy()\n r = np.reshape(r, -1)\n x = xtilde - r\n\n # rescale the denoised v back to original scale\n x = x * (maxtmp - mintmp) + mintmp\n\n \"\"\" dual update \"\"\"\n\n u = np.copy(u_old + x_old - v)\n\n \"\"\" Monitors \"\"\"\n\n fpr = np.sqrt(np.sum(np.square(x + u - x_old - u_old)))\n\n if peak is not None and maxval is not None:\n psnrs = compare_psnr(im_true=clean,\n im_test=np.reshape(x/peak*maxval, (m, n)),\n data_range=data_range)\n if verbose:\n print(\"i = {},\\t psnr = {},\\t fpr = {}\".format(i+1, psnrs, fpr))\n\n\n \"\"\" Get restored image. \"\"\"\n x = np.reshape((x) , (m, n))\n return x\n\n# ---- input arguments ----\nopt = analyze_parse(40, 0.1, 100) # the arguments are default sigma, default alpha and default max iteration.\n\n# ---- load the model ---- \nmodel = load_model(opt.model_type, opt.sigma)\n\nwith torch.no_grad():\n # ---- load the problem ---- \n mat = sio.loadmat('Demo_mat/poisson_demo.mat')\n clean_scaled = mat.get('clean_scaled').astype(np.float64)\n clean = mat.get('clean').astype(np.float64)\n noisy = mat.get('noisy').astype(np.float64)\n peak = float(mat.get('peak'))\n maxval = float(mat.get('maxval'))\n\n # ---- options ---- \n opts = dict(lam=1/opt.alpha, beta=1.0, maxitr=opt.maxitr,\n peak=peak, maxval=maxval, data_range=255.0, verbose=opt.verbose)\n \n # ---- plug and play !!!! ----\n out = pnp_admm_poisson_denoising(noisy, model, clean, **opts)\n\n\n","repo_name":"uclaopt/Provable_Plug_and_Play","sub_path":"pnp_admm_poisson_denoise.py","file_name":"pnp_admm_poisson_denoise.py","file_ext":"py","file_size_in_byte":4216,"program_lang":"python","lang":"en","doc_type":"code","stars":63,"dataset":"github-code","pt":"7"} +{"seq_id":"24214787339","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[8]:\n\n\ndef quadratic(equation):\n e=equation.split()\n a=int(e[0])\n b=int(e[2])\n c=int(e[4])\n s=(-b+((b**2)-(4*a*c))**0.5)/(2*a)\n h=(-b-((b**2)-(4*a*c))**0.5)/(2*a)\n d=[s,h]\n print(d)\n\nequation=(input(\"enter the equation in the form a_*x^2_b_*x_c incorporating signs of a,b,c :\"))\nquadratic(equation)\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n","repo_name":"piyush-daigavhane/ERC-ECS-Python-Question","sub_path":"question_2.py","file_name":"question_2.py","file_ext":"py","file_size_in_byte":445,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13456796861","text":"import os\nimport requests\nimport zipfile\nimport cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport sys\nfrom nnObjs import *\n\ndef loadMNISTDataset(path):\n\n labels = os.listdir(path)\n X = []\n y = []\n print(\"Loading images...\")\n\n numFiles = 0\n for label in labels:\n files = os.listdir(path + \"/\" + label)\n numFiles += len(files)\n\n cumNum = 1\n for label in labels:\n files = os.listdir(path + \"/\" + label)\n for file in files:\n print(f\"Loading file {file} in label {label} in {path} PROGRESS: %{100*cumNum/numFiles:.3f}\")\n image = cv2.imread(path + \"/\" + label + \"/\" + file, cv2.IMREAD_UNCHANGED)\n X.append(image)\n y.append(label)\n cumNum += 1\n\n return np.array(X), np.array(y).astype(\"uint8\")\n\nURL = 'https://nnfs.io/datasets/fashion_mnist_images.zip'\nFILE = 'fashion_mnist_images.zip'\nFOLDER = \"fashionMNISTImages\"\n\n\n# prepare data\nif not os.path.isfile(FILE):\n\n print(f'Downloading {URL} and saving as {FILE}...')\n \n r = requests.get(URL)\n with open(FILE, \"wb\") as zipFile:\n zipFile.write(r.content)\n\nif not os.path.isdir(FOLDER):\n print(f'Unzipping to folder {FOLDER}...')\n\n with zipfile.ZipFile(FILE) as zipFile:\n zipFile.extractall(FOLDER)\n\n# display example data\nimageData = cv2.imread(FOLDER + \"/train/4/0011.png\", cv2.IMREAD_UNCHANGED)\nnp.set_printoptions(linewidth=200)\nplt. imshow (imageData, cmap=\"gray\")\nplt. show (block=False)\nplt. pause (3)\nplt. close ()\n\n#load train and test data\nX, y = loadMNISTDataset ( FOLDER + \"/train\" )\nXTest, yTest = loadMNISTDataset ( FOLDER + \"/test\" )\n\n#scale image data (0, 255) to (-1, 1)\nX = (X .astype(np.float32) - 255/2) / (255/2)\nXTest = (XTest.astype(np.float32) - 255/2) / (255/2)\n\n#reshape\nX = X. reshape(X. shape[0], -1)\nXTest = XTest. reshape(XTest. shape[0], -1)\nX = np.transpose(X)\nXTest = np.transpose(XTest)\n\n#shuffle\nindexes = np.array(range(X.shape[1]))\nnp.random.shuffle(indexes)\nX = X[:, indexes]\ny = y[indexes]\n\n# create model\nmodel1 = model()\n\nmodel1.addLayer(layerDense(64, X.shape[0]))\nmodel1.addLayer(actReLu())\nmodel1.addLayer(layerDense(64, 64))\nmodel1.addLayer(actReLu())\nmodel1.addLayer(layerDense(10, 64))\nmodel1.addLayer(actSoftmax())\n\nmodel1.set( loss = lossCatCrossEnt(), \n optimizer = optimizerAdam(decay=5e-5),\n accuracy=accuracyCategorical()\n )\n\nmodel1.establish()\n\nmodel1.train(X, y, numEpoch=5, batchSize=128, printEvery=100)\n\nmodel1.validate(XVal=XTest, yVal=yTest, batchSize=128)\n\nmodel1.plotEpoch()\n\nparams = model1.getParams()\n\nmodel1.saveParams(\"fashionMNIST.params\")\n\nmodel1.saveModel(\"fashionMNIST.model\")\n\n# CREATE NEW MODEL FROM TRAINED MODEL PARAMETERS\nmodel2 = model()\n\nmodel2.addLayer(layerDense(64, X.shape[0]))\nmodel2.addLayer(actReLu())\nmodel2.addLayer(layerDense(64, 64))\nmodel2.addLayer(actReLu())\nmodel2.addLayer(layerDense(10, 64)) \nmodel2.addLayer(actSoftmax())\n\nmodel2.set(loss=lossCatCrossEnt(), accuracy=accuracyCategorical())\n\nmodel2.establish()\n\nmodel2.loadParams(\"fashionMNIST.params\")\n\nmodel2.validate(XTest, yTest)\n\n# LOAD SAVED MODEL\nmodel3 = model.load(\"fashionMNIST.model\")\n\nmodel3.validate(XTest, yTest)\n\nconfidences = model3.predict(XTest, batchSize=2500)\npredictions = model3.outputLayerActivation.prediction(confidences)\n\n\n# USE MODEL3\nmodel3 = model.load(\"fashionMNIST.model\")\n\nfashionMNISTLabels = {\n 0: 'T-shirt/top',\n 1: 'Trouser',\n 2: 'Pullover',\n 3: 'Dress',\n 4: 'Coat',\n 5: 'Sandal',\n 6: 'Shirt',\n 7: 'Sneaker',\n 8: 'Bag',\n 9: 'Ankle boot'\n }\n\nimageData = cv2.imread(FOLDER + \"/test/4/0011.png\", cv2.IMREAD_GRAYSCALE)\nplt.figure()\nplt.imshow (imageData, cmap=\"gray\")\n\n#imageData = 255 - imageData\n#plt.figure()\n#plt.imshow (imageData, cmap=\"gray\")\n\nimageData = cv2.resize(imageData, (28, 28))\nplt.figure()\nplt.imshow (imageData, cmap=\"gray\")\n\nimageData = (imageData .astype(np.float32) - 255/2) / (255/2)\nimageData = imageData.reshape(1, -1)\nimageData = np.transpose(imageData)\n\n\nconfidences = model3.predict(imageData)\npredictions = model3.outputLayerActivation.prediction(confidences)\n\n#print(predictions)\n#print(predictions.shape)\n#print(predictions[0, 0])\n\nprediction = fashionMNISTLabels[predictions[0, 0]]\n\nprint(prediction)\n\nplt.show()","repo_name":"omerozy/appMLP","sub_path":"fashionMINST.py","file_name":"fashionMINST.py","file_ext":"py","file_size_in_byte":4653,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"42673717978","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n# How much previous days count\nDECAY = 0.8\n\n# How much success in a higher grade counts for lower grades\n# and how much failure in lower grades counts for higher grades\nOTHER_GRADE_WEIGHT = 0.5\n\n# How much a partially completed route counts\nPARTIAL_ROUTE_WEIGHT = 0.5\n\n# Should we translate color names to English?\nENGLISH = False\n\n","repo_name":"samhocevar/climbing","sub_path":"src/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":380,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"20397359916","text":"'''\nFile: solve.py\nAuthor: Yutong Dai (rothdyt@gmail.com)\nFile Created: 2020-02-23 10:51\nLast Modified: 2021-01-28 00:05\n--------------------------------------------\nDescription:\n'''\nimport sys\nimport os\nimport time\nimport numpy as np\nsys.path.append(\"../\")\nsys.path.append(\"../..\")\n\nimport src.solvers.FaRSAGroup.utils as utils\nfrom src.solvers.FaRSAGroup.params import *\nfrom src.solvers.FaRSAGroup.Algorithm import AlgoBase, set_prox_stepsize\nimport src.solvers.FaRSAGroup.printUtils as printUtils\n\n\ndef solve(f, r, X_initial=None, proxStepsize=None, method='gradient',\n update_proxStepsize='const', params=params, print_group=True, print_second_level=False,\n kappa_1=1e-2, kappa_2=1e-2, fraction=0.8, print_time=False, cg_backtrack_strategy=\"cutfrac\"):\n \"\"\"\n Solve the group l2 regualrized problem.\n\n Args:\n f: loss function object\n r: regularizer object\n X_intial(np.array): starting point\n proxStepsize(np.float64/np.array/None): \n If it is set to None, then one should call `set_prox_stepsize` to initialize the proxStepsize\n method: parameters for `set_prox_stepsize`.\n update_proxStepsize:\n 'const': use constant params['zeta']\n 'single': use only one proxStpesize for all groups; estimated by Lipschtiz constant\n 'block': each group has its own proxStepsize; at each iteration, working groups are updated with the same Lipschtiz constant estimation\n 'group': each group has its own proxStepsize; at each iteration, proxStepsize for each working groups are updated separately\n 'groupv': varinat of group and only allows to decrease proxstepsize\n update_kappa:\n 'chicg': update min(kappa, 1/2) as chicg suggests\n 'hack': update kappa as hack suggests\n params:\n a dictionary containing parameters for linesearch methods.\n force_cg: force one to try newton step with the unit stepsize\n \"\"\"\n # print out some key parameters\n if update_proxStepsize in ('single', 'const'):\n alpha_type = 'singlealpha'\n else:\n alpha_type = 'groupalpha'\n if proxStepsize is None:\n # proxStepsize = set_prox_stepsize(f, r, alpha_type, method)\n proxStepsize = np.min([set_prox_stepsize(f, r, alpha_type, method), 1])\n print('proxStepsize is [None]; Set up by {}'.format(method))\n else:\n print('proxStepsize is [{}]'.format(type(proxStepsize)))\n print(\"Update proxStepsize using {}\".format(update_proxStepsize))\n print(\"Print time per iteration: {}\".format(print_time))\n print(\"Termination tol:cg:{} | pg:{}\".format(\n params['tol_cg'], params['tol_pg']))\n print(\"maxtime:{} | maxiter:{}\".format(\n params['max_time'], params['max_iter']))\n print(\n f\"kappa1_max:{params['kappa1_max']:3.3e} | kappa2_max:{params['kappa2_max']:3.3e}\")\n print(\n f\"kappa1_min:{params['kappa1_min']:3.3e} | kappa2_min:{params['kappa2_min']:3.3e}\")\n print(\n f\"kappa_increase:{params['kappa_increase']} | kappa_decrease:{params['kappa_decrease']}\")\n print(f\"kappaStrategy: dynamic | count freq:{params['kappa_freq_count']}\")\n print(f\"Equipped with early termination\")\n # set up algorithms\n Ndata = f.n\n p = f.p\n K = r.K\n G_i_starts = r.starts\n G_i_ends = r.ends\n unique_groups = r.unique_groups\n problem_attribute = f.__str__()\n problem_attribute += \"Regularizer:{:.>44}\\n\".format(r.__str__())\n problem_attribute += \"Penalty Parameter:{:.>30}lambda={:3.4f}\\n\".format(\n '', r.penalty)\n problem_attribute += \"Number of groups:{:.>32}\\n\".format(K)\n problem_attribute += \"Update proxStepsize:{:.>35}\\n\".format(\n update_proxStepsize)\n algo = AlgoBase(f, r, proxStepsize, params, kappa_1, kappa_2)\n iteration_start = time.time()\n if X_initial is None:\n X = np.zeros([p, 1])\n else:\n X = X_initial\n fevals = gevals = HvProds = subitsT = 0\n info = {}\n if params['save_log']:\n outID = '{}'.format(params['tag'])\n else:\n outID = f.datasetName\n time_so_far = print_cost = 0\n if params['printlevel'] >= 1:\n utils.print_problem(problem_attribute, outID)\n # normX = utils.l2_norm(X)\n normX = np.sqrt(np.dot(X.T, X))[0][0]\n algo.fval = algo.f.evaluate_function_value(X)\n rval = algo.r.func(X)\n algo.F = algo.fval + rval\n fevals += 1\n gevals += 1\n iteration = 0\n time_update_proxStepszie = 0\n if Ndata < p:\n print('n=p, no fraction')\n prox_time = 0\n newton_time = 0\n pg_ls_time = 0\n cg_ls_time = 0\n pg_time = 0\n cg_time = 0\n consequtive_pg = 0\n if params[\"kappa_freq_count\"]:\n kappa_increase, kappa_decrease = 0, 0\n kappa_1_max, kappa_1_min = kappa_1, kappa_1\n kappa_2_max, kappa_2_min = kappa_2, kappa_2\n\n while True:\n # print('=============================')\n # print(\"Iteration:{}\".format(iteration))\n # if iteration == 9:\n # print(1)\n prox_time_iter = time.time()\n algo.proximal_step(X)\n # print(\"||s||:{}\".format(utils.l2_norm(algo.proximal)))\n prox_time_iter = time.time() - prox_time_iter\n prox_time += prox_time_iter\n # call set_cg first because the internal dependence issues.\n algo.set_cg()\n algo.set_pg()\n # print(f\"kappa_1:{algo.kappa_1:2.3e} | kappa_2:{algo.kappa_2:2.3e}\")\n gI_cg, nI_cg, gI_pg, nI_pg = len(algo.I_cg_group), np.sum(\n algo.I_cg_index), algo.K - len(algo.I_cg_group), np.sum(algo.I_pg_index)\n # if iteration in [95, 96]: # debug madelon to remove\n # # print(\"Iter:{} | old: {} | new: {}\".format(iteration, chicg_last_iteration, algo.chi_cg))\n # print(iteration)\n # print(utils.get_classification(algo.zeroGroup, algo.nonzeroGroup, algo.zeroProxGroup, algo.nonzeroProxGroup))\n iteration_end = time.time() - iteration_start - print_cost\n time_so_far += iteration_end\n\n if params['printlevel'] == 2:\n if iteration % params['printevery'] == 0:\n # utils.print_header(outID, print_time)\n printUtils.print_header(outID, print_time)\n res = utils.get_classification(\n algo.zeroGroup, algo.nonZeroGroup, algo.zeroProxGroup, algo.nonZeroProxGroup)\n nn, nz, zn, zz = len(\n res['NZ-NZ']), len(res['NZ-Z']), len(res['Z-NZ']), len(res['Z-Z'])\n # utils.print_iteration(iteration, algo.fval, normX, algo.F, algo.proxStepsize, algo.chi_cg, algo.chi_pg,\n # gI_cg, nI_cg, gI_pg, nI_pg, nn, nz, zn, zz, outID)\n printUtils.print_iteration(iteration, algo.fval, normX, algo.F, algo.proxStepsize, algo.kappa_1, algo.chi_cg, algo.chi_pg,\n gI_cg, nI_cg, gI_pg, nI_pg, nn, nz, zn, zz, outID)\n if iteration == 0:\n chi_cg_0 = algo.chi_cg\n chi_pg_0 = algo.chi_pg\n chi_cg_termination = params['tol_cg'] * max(1, chi_cg_0)\n chi_pg_termination = params['tol_pg'] * max(1, chi_pg_0)\n if (algo.chi_cg <= chi_cg_termination) and (algo.chi_pg <= chi_pg_termination):\n info['status'] = 0\n break\n if iteration >= params['max_iter']:\n info['status'] = 1\n break\n if time_so_far > params['max_time']:\n info['status'] = 2\n break\n iteration_start = time.time()\n\n if algo.chi_pg <= params['Gamma'] * algo.chi_cg:\n # choose the working groups\n cg_time_iter = time.time()\n if Ndata < p:\n algo.select_cg_frac(fraction)\n else:\n algo.select_cg()\n y = algo.cg_step(X, cg_backtrack_strategy)\n cg_time_iter = time.time() - cg_time_iter\n cg_time += cg_time_iter\n cg_ls_time += algo.ls_time_iter\n newton_time += algo.newton_time_iter\n HvProds += algo.subits\n fevals += algo.cg_feval\n if algo.status in [-1, 4]:\n info['status'] = algo.status\n if algo.status == -1:\n print(\"maxback cg:\", algo.cg_backtrack)\n utils.print_cg_step(algo.typeofIteration, algo.nI_cgs, algo.gradF_Icgs_norm, algo.subprobFlag, algo.subits,\n algo.res, algo.res_target, algo.normd, algo.cg_type, algo.newZB, algo.dirder,\n algo.projection_attempts, algo.cg_backtrack, algo.cg_stepsize,\n prox_time_iter, algo.newton_time_iter, algo.ls_time_iter, cg_time_iter,\n outID, print_time)\n break\n else:\n pg_time_iter = time.time()\n algo.select_pg()\n y = algo.pg_step(X)\n pg_time_iter = time.time() - pg_time_iter\n pg_time += pg_time_iter\n pg_ls_time += algo.ls_time_iter\n fevals += algo.pg_feval\n if algo.pg_flag == False:\n if update_proxStepsize == 'const':\n algo.proxStepsize *= algo.params['zeta']\n if algo.status == -1:\n info['status'] = algo.status\n print(\"maxback pg:\", algo.pg_backtrack)\n utils.print_pg_step(algo.typeofIteration, algo.nI_pgs, algo.subits, algo.normd,\n algo.pg_backtrack, algo.pg_stepsize,\n prox_time_iter, algo.ls_time_iter, pg_time_iter,\n outID, print_time)\n break\n if update_proxStepsize != 'const':\n time_update_proxStepszie_begin = time.time()\n if algo.typeofIteration == 'cg':\n algo.get_proxStepsize(\n X, y, algo.f_old, algo.fval, algo.d_use, algo.stepSize_use)\n else:\n algo.get_proxStepsize(\n X, y, algo.f_old, algo.fval, algo.d_use, algo.stepSize_use)\n algo.proxStepsize = algo.newProxStepsize\n time_update_proxStepszie_end = time.time()\n time_update_proxStepszie += time_update_proxStepszie_end - \\\n time_update_proxStepszie_begin\n\n subitsT += algo.subits\n temp = time.time()\n if params['printlevel'] == 2:\n if algo.typeofIteration == 'cg':\n utils.print_cg_step(algo.typeofIteration, algo.nI_cgs, algo.gradF_Icgs_norm, algo.subprobFlag, algo.subits,\n algo.res, algo.res_target, algo.normd, algo.cg_type, algo.newZB, algo.dirder,\n algo.projection_attempts, algo.cg_backtrack, algo.cg_stepsize,\n prox_time_iter, algo.newton_time_iter, algo.ls_time_iter, cg_time_iter,\n outID, print_time)\n else:\n utils.print_pg_step(algo.typeofIteration, algo.nI_pgs, algo.subits, algo.normd,\n algo.pg_backtrack, algo.pg_stepsize,\n prox_time_iter, algo.ls_time_iter, pg_time_iter,\n outID, print_time)\n if print_second_level:\n if len(algo.bar_I_cg_group) != 0:\n utils.print_more(algo.bar_I_cg_index, algo.bar_I_cg_group, algo.bar_chi_cg, algo.I_cg_index, algo.chi_cg,\n algo.norm_gradF_bar_I_cg, algo.group_X_norm, algo.group_gradF_norm,\n algo.small_radius_lst, algo.outter_radius_lst, algo.inner_radius_lst,\n algo.kappa_1, algo.kappa_2, algo.kappa_3, outID)\n else:\n utils.print_empty_bar_Icg(outID)\n print_cost = time.time() - temp\n X = y\n # this is need if update_proxStepsize is not const or single\n # algo.fval = algo.f.evaluate_function_value(X)\n # algo.F = algo.fval + algo.r.evaluate_function_value(X)\n # normX = utils.l2_norm(X)\n normX = np.sqrt(np.dot(X.T, X))[0][0]\n gevals += 1 # about to call gradient method in the next iteration\n iteration += 1\n if Ndata < p:\n if (iteration % 5 == 1) and (algo.typeofIteration == 'cg'):\n F_seq_switch.pop(0)\n F_seq_switch.append(algo.F)\n if np.abs(F_seq_switch[0] - F_seq_switch[1]) <= 1e-3:\n fraction = 1\n else:\n fraction = fraction_init\n if algo.typeofIteration == 'cg':\n consequtive_pg = 0\n # 6 - 1 = 5 total backtracks\n if algo.cg_backtrack + algo.projection_attempts > 6:\n # default: 1e3 make 10 also controlable # factor to increase and decrease (10, 1/10)\n algo.kappa_1 = min(\n algo.kappa_1 * params['kappa_increase'], params['kappa1_max'])\n algo.kappa_2 = min(\n algo.kappa_2 * params['kappa_increase'], params['kappa2_max'])\n if params[\"kappa_freq_count\"]:\n kappa_increase += 1\n kappa_1_max = max(algo.kappa_1, kappa_1_max)\n kappa_2_max = max(algo.kappa_2, kappa_2_max)\n else:\n consequtive_pg += 1\n if consequtive_pg > 5:\n # default: 1e-5\n algo.kappa_1 = max(\n algo.kappa_1 * params['kappa_decrease'], params['kappa1_min'])\n algo.kappa_2 = max(\n algo.kappa_2 * params['kappa_decrease'], params['kappa2_min'])\n if params[\"kappa_freq_count\"]:\n kappa_decrease += 1\n kappa_1_min = min(kappa_1_min, algo.kappa_1)\n kappa_2_min = min(kappa_2_min, algo.kappa_2)\n\n if params['printlevel'] == 2:\n utils.print_exit(info['status'], outID)\n nnz = utils.get_group_structure(\n X, K, unique_groups, G_i_starts, G_i_ends, epsilon=1e-8)\n info['n'] = algo.f.n\n info['p'] = algo.f.p\n info['Lambda'] = algo.r.penalty\n info['num_groups'] = algo.r.K\n info['nnz'] = nnz\n info['nz'] = algo.K - nnz\n info['F'] = algo.F\n info['normX'] = normX\n info['f'] = algo.fval\n info['chipg'] = algo.chi_pg\n info['chicg'] = algo.chi_cg\n info['fevals'] = fevals\n info['HvProds'] = HvProds\n info['time'] = time_so_far\n info['iteration'] = iteration\n info['num_pg_steps'] = algo.num_pg_steps\n info['num_cg0_stpes'] = algo.num_cg0_steps\n info['num_cgdesc_steps'] = algo.num_cgdesc_steps\n info['gevals'] = gevals\n info['subits'] = subitsT\n info['time_update_stepsize'] = time_update_proxStepszie\n info['X'] = X\n info['zeroGroup'] = algo.zeroGroup\n info['nonZeroGroup'] = algo.nonZeroGroup\n info['proxStepsize'] = algo.proxStepsize\n info['cg_time'] = cg_time\n info['newton_time'] = newton_time\n info['cg_ls_time'] = cg_ls_time\n info['pg_time'] = pg_time\n info['pg_ls_time'] = pg_ls_time\n info['prox_time'] = prox_time\n if params[\"kappa_freq_count\"]:\n info['kappa_increase'] = kappa_increase\n info['kappa_decrease'] = kappa_decrease\n info['kappa1_max'] = kappa_1_max\n info['kappa2_max'] = kappa_2_max\n info['kappa1_min'] = kappa_1_min\n info['kappa2_min'] = kappa_2_min\n if params['printlevel'] == 2 and info['status'] != -1 and print_group:\n utils.print_group_sparsity(\n X, K, unique_groups, G_i_starts, G_i_ends, outID, epsilon=1e-8)\n if params['printlevel'] == 2:\n utils.print_result(info, outID)\n return info\n","repo_name":"Yutong-Dai/S-PStorm","sub_path":"src/solvers/FaRSAGroup/solve.py","file_name":"solve.py","file_ext":"py","file_size_in_byte":15871,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"30833876600","text":"import numpy as np\nimport tensorflow as tf\nimport os\nfrom PIL import Image\n\nfrom networks import dcgan_utils\n\n\ndef build_model():\n \"\"\"\n Builds the generator part of the eye-glass generating DCGAN model.\n\n :return: the built generator model as tf.keras Sequential object (not compiled yet)\n \"\"\"\n inp = tf.keras.layers.InputLayer((25,)) # input layer\n fc = tf.keras.layers.Dense(7040) # fully connected layer\n reshape = tf.keras.layers.Reshape(target_shape=(4, 11, 160)) # reshape tensor\n # \"deconvolutional\" layers\n deconv1 = tf.keras.layers.Conv2DTranspose(80, (5, 5), strides=(2, 2), padding='same')\n deconv2 = tf.keras.layers.Conv2DTranspose(40, (5, 5), strides=(2, 2), padding='same')\n deconv3 = tf.keras.layers.Conv2DTranspose(20, (5, 5), strides=(2, 2), padding='same')\n deconv4 = tf.keras.layers.Conv2DTranspose(3, (5, 5), strides=(2, 2), padding='same', activation='tanh')\n\n model = tf.keras.models.Sequential(\n [\n inp,\n fc,\n tf.keras.layers.BatchNormalization(),\n tf.keras.layers.ReLU(),\n reshape,\n deconv1,\n tf.keras.layers.BatchNormalization(),\n tf.keras.layers.ReLU(),\n deconv2,\n tf.keras.layers.BatchNormalization(),\n tf.keras.layers.ReLU(),\n deconv3,\n tf.keras.layers.BatchNormalization(),\n tf.keras.layers.ReLU(),\n deconv4\n ],\n name='Generator'\n )\n\n model.summary()\n\n return model\n\n\n# NOTE: start fresh training instead\n@DeprecationWarning\ndef load_gen_weights(gmodel):\n from deprecated import model_importer\n npas = model_importer.load_dcgan_mat_model_weights('../matlab-models/gen.mat')\n gmodel.layers[0].set_weights([npas[0], dcgan_utils.get_xavier_initialization((7040,))])\n gmodel.layers[4].set_weights([np.reshape(npas[3], (5, 5, 80, 160)), dcgan_utils.get_xavier_initialization((80,))])\n gmodel.layers[7].set_weights([np.reshape(npas[6], (5, 5, 40, 80)), dcgan_utils.get_xavier_initialization((40,))])\n gmodel.layers[10].set_weights([np.reshape(npas[9], (5, 5, 20, 40)), dcgan_utils.get_xavier_initialization((20,))])\n gmodel.layers[13].set_weights([np.reshape(npas[12], (5, 5, 3, 20)), dcgan_utils.get_xavier_initialization((3,))])\n\n return gmodel\n\n\ndef scale_gen_output(prediction):\n \"\"\"\n Scales the values of a NumPy array with original range [-1, 1] to range [0, 255].\n\n :param prediction: a numpy array with values ranging between -1 and 1\n :return: a numpy array with integer values between 0 and 255\n \"\"\"\n prediction += 1 # shift to range [0, 2]\n prediction *= 127.5 # scale to range [0, 255]\n prediction = np.round(prediction, 0)\n prediction = prediction.astype(int)\n\n return prediction\n\n\ndef save_gen_output_to_file(matrix):\n print(f'Saving image matrix of size {np.shape(matrix)}')\n matrix = np.asarray(matrix, dtype=np.uint8)\n img = Image.fromarray(matrix, 'RGB')\n\n # img.show()\n if not os.path.exists('../../out'):\n os.makedirs('../../out')\n img.save('../../out/generated_glass.png', 'PNG')\n","repo_name":"PhilK-7/agns-port","sub_path":"agns-py/src/networks/eyeglass_generator.py","file_name":"eyeglass_generator.py","file_ext":"py","file_size_in_byte":3142,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"14444199013","text":"from PIL import ImageTk, Image, ImageDraw\nimport PIL\nfrom tkinter import *\nimport matplotlib.pyplot as plt\nimport numpy as np\n\n\nclass DoodleInput():\n def __init__(self) -> None:\n width = 300 # canvas width\n height = 300 # canvas height\n center = height//2\n white = (255, 255, 255) # canvas back\n\n self.drawing_coords = []\n\n master = Tk()\n\n # create a tkinter canvas to draw on\n self.canvas = Canvas(master, width=width, height=height, bg='white')\n self.canvas.pack()\n\n # create an empty PIL image and draw object to draw on\n self.output_image = PIL.Image.new(\"RGB\", (width, height), white)\n self.draw = ImageDraw.Draw(self.output_image)\n self.canvas.pack(expand=YES, fill=BOTH)\n self.canvas.bind(\"\", self.paint)\n\n # add a button to save the image\n button=Button(text=\"save\",command=self.save)\n button.pack()\n\n master.mainloop()\n\n def save(self):\n coords = np.asarray(self.drawing_coords.copy())\n coords[:,1] = -coords[:,1] + 300\n np.savetxt('draw_coords.csv', coords, delimiter=\",\")\n self.plot(coords)\n # self.output_image.save(filename)\n\n def paint(self, event):\n x1, y1 = (event.x - 1), (event.y - 1)\n x2, y2 = (event.x + 1), (event.y + 1)\n self.drawing_coords.append((x1, y2))\n self.canvas.create_oval(x1, y1, x2, y2, fill=\"black\",width=5)\n self.draw.line([x1, y1, x2, y2],fill=\"black\",width=5)\n\n def plot(self, coords):\n plt.plot(coords[:,0], coords[:,1])\n plt.show()\n\n\nif __name__ == \"__main__\":\n di = DoodleInput()","repo_name":"edbatk/RobotSystems","sub_path":"ArmPi/Week 8/doodleInput.py","file_name":"doodleInput.py","file_ext":"py","file_size_in_byte":1657,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25721962054","text":"import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.animation import FuncAnimation\nfrom IPython.display import display\nimport math\n\npd.options.mode.chained_assignment = None\n\n##################################\n# FUNCOES #\n##################################\n\n \ndef integrate(x, y, y0 = 0):\n yi = [y0]\n \n for i in range(1, min(len(x), len(y))):\n dy = ((y[i-1] + y[i]) / 2) * (x[i] - x[i-1])\n yi.append(yi[i-1] + dy)\n \n return yi\n\ndef derivate(x, y, y0 = 0):\n '''\n array, array, object -> array\n \n Recebe dois vetores com dados e retorna um vetor de mesmo tamanho, que seria o resultado da derivada de y em x.\n y0 é o valor inicial da derivada.\n '''\n yi = [y0]\n \n for i in range(1, min(len(x), len(y))):\n dy = y[i] - y[i-1]\n dx = x[i] - x[i-1]\n if dx == 0:\n yi.append(dy)\n else:\n yi.append(dy / dx)\n \n return yi\n\ndef frange(start, stop, step=1.0):\n \n i = start\n while i < stop:\n yield i\n i += step\n \ndef erro(df, dft1, dft2, x):\n e1 = []\n e2 = []\n \n for i in range(len(df)):\n j = math.floor(df['time'][i] * 1000)\n e1.append(df[x][i] - dft1[x][j])\n e2.append(df[x][i] - dft2[x][j])\n \n df[x + '_e1'] = e1\n df[x + '_e2'] = e2\n \ndef plot_experimentos(df, x, labels = '', units = '', exp = ''):\n h = len(x)\n fig, plots = plt.subplots(h, 5, sharex = 'col', sharey = 'row')\n fig.set_figwidth(20)\n fig.set_figheight(4 * h)\n fig.subplots_adjust(hspace = 0.05, wspace = 0, top = 0.93)\n fig.suptitle('Valores experimentais para os experimentos de ' + exp)\n fig.set_facecolor('#FFFFFF')\n for i in range(5):\n j = 0\n plots[j, i].scatter(df[i]['time'], df[i][x[j]], s = 1 if h == 3 else 10)\n plots[j, i].set_title('Experimento ' + str(i+1))\n \n if h == 3:\n j += 1\n plots[j, i].scatter(df[i]['time'], df[i][x[j]], s = 1)\n \n j += 1 \n plots[j, i].scatter(df[i]['time'], df[i][x[j]], s = 1 if h == 3 else 10)\n plots[j, i].set_xlabel('Tempo (s)')\n \n j = 0\n plots[j, 0].set_ylabel(labels[j] + ' (' + units[j] + ')')\n \n if h == 3:\n j += 1\n plots[j, 0].set_ylabel(labels[j] + ' (' + units[j] + ')')\n \n j += 1\n plots[j, 0].set_ylabel(labels[j] + ' (' + units[j] + ')')\n \n plt.show()\n \ndef plot_modelos(df, dft1, dft2, x, labels = '', units = '', exp = ''):\n def plot_modelo(plots, df, dft, x, i = 0, m = 1):\n i *= 2\n m -= 1\n plots[i, m].scatter(df['time'], df[x], s = 1)\n plots[i, m].plot(dft['time'], dft[x], color = 'C1')\n plots[i, m].set_xticklabels([])\n \n i += 1\n plots[i, m].set_xlabel('Tempo (s)')\n plots[i, m].scatter(df['time'], df[x + '_e' + str(m+1)], s = 1)\n plots[i, m].plot([0, max(df['time'])], [0, 0], '--', color = 'C1')\n \n h = len(x)\n \n fig, plots = plt.subplots(2 * h, 2, sharey = 'row')\n fig.set_figwidth(10)\n fig.set_figheight(8 * h)\n fig.subplots_adjust(wspace = 0, top = 0.93)\n fig.suptitle('Comparação dos resultados do método de Euler com o método\\nde Euler-Richardson para os experimentos de ' + exp)\n fig.set_facecolor('#FFFFFF')\n \n plots[0, 0].set_title('Método de Euler')\n plots[0, 1].set_title('Método de Euler-Richardson')\n for i in range(h):\n plot_modelo(plots, df, dft1, x[i], i = i, m = 1)\n plot_modelo(plots, df, dft2, x[i], i = i, m = 2)\n plots[2*i, 0].set_ylabel(labels[i] + ' (' + units[i] + ')')\n plots[2*i+1, 0].set_ylabel('Erro de ' + labels[i] + ' (' + units[i] + ')')\n \n \n \n\n##################################\n# DADOS #\n##################################\n#\n## 'CHUTE' = valores que esquecemos (Vitor, principalmente) de coletar, então estão estimados e podem precisar de ajustes\n#\n#\n## A plotagem dos gráficos não é definitiva e está aí mais pra visualizarmos e fazermos testes\n\n############ Constantes ###########\n\ng = 9.8 # aceleração da gravidade (m/s^2)\npar = 1.2 # densidade do ar (kg/m^3)\n\n########## Bloco em rampa #########\nthetar = 6 * math.pi / 180 # inclinacao (rad)\nD = 3.64 # distância percorrida (m)\nmr = 0.34 # massa do bloco (g)\nAr = 0.008 # área frontal do bloco (m^2)\nr = [] # lista para armazenar os dados do acelerometro\n\nfor i in range(5):\n # lendo dados do acelerometro\n r.append(pd.read_csv('r' + str(i+1) + '.csv'))\n \n\n############### MCU ###############\nR = 2.1 # raio (m)\nt = pd.read_csv('mcu.csv') # tabela com os dados dos tempos das voltas\nc = [] # lista para armazenar os tempos medidos pra cada experimento\n\nfor i in range(5):\n c.append(t.loc[t['rep'] == i+1])\n c[i] = c[i].reset_index()\n c[i] = c[i].filter(items=['time', 'theta'])\n\n############# Pêndulo #############\nl = 0.823 # comprimento das cordas (m)\na = 0.051 # largura do cesto (m)\nb = 0.144 # comprimento do cesto (m)\nL = math.sqrt(l**2 - (a/2)**2 - (b/2)**2) # comprimento da corda imaginária do pêndulo (m)\nmp = 0.23 # massa do pendulo (kg)\nAp = 0.0059 # área frontal do pêndulo (m^2)\np = [] # lista para armazenar os dados do giroscópio\ntmin_p = [7.05, 5.15, 5.2, 5.15, 5.3] # tempos iniciais (s)\ntheta0 = [] # inclinação inicial\n\nfor i in range(5):\n # lendo dados do giroscópio\n p.append(pd.read_csv('p' + str(i+1) + '.csv'))\n \n \n##################################\n# LIMPEZA DOS DADOS #\n##################################\n \n########## Bloco em rampa #########\n \ntmin_r = [4.7, 2.9, 6.2, 4.6, 4.5] # tempos iniciais estimados (s)\n\nfor i in range(5):\n # removendo variáveis que não serão utilizadas\n r[i] = r[i].filter(items=['time', 'gFx'])\n r[i].columns = ['time', 'a']\n # removendo dados antes do tempo inicial estimado\n r[i] = r[i].loc[r[i]['time'] >= tmin_r[i]]\n r[i] = r[i].reset_index(drop = True)\n # convertendo valores para m/s^2\n r[i].update(pd.Series(r[i]['a'] * g, name = 'a')) \n # 'zerando' o tempo\n r[i].update(pd.Series(r[i]['time'] - tmin_r[i], name = 'time'))\n # calculando as velocidades experimentais\n r[i]['v'] = integrate(r[i]['time'], r[i]['a'])\n # calculando o deslocamento experimental\n r[i]['d'] = integrate(r[i]['time'], r[i]['v'])\n # filtrando dados com deslocamento maior que o tamanho total da rampa\n r[i] = r[i].loc[r[i]['d'] <= D]\n \n############### MCU ###############\n \nfor i in range(5):\n c[i]['theta'][0] = 1*math.pi\n c[i]['theta'][1] = 2*math.pi\n for j in [2, 4]:\n c[i]['theta'][j] = (j+1)*math.pi\n c[i]['theta'][j+1] = (j+2)*math.pi\n c[i]['time'][j] += c[i]['time'][j-1]\n c[i]['time'][j+1] += c[i]['time'][j-1]\n c[i]['w'] = derivate(c[i]['time'], c[i]['theta'], y0 = c[i]['theta'][0] / c[i]['time'][0])\n \n############# Pêndulo #############\n \ntmin_p = [7.05, 5.15, 5.2, 5.15, 5.3] # tempos iniciais estimados (s)\n\nfor i in range(5):\n # removendo variáveis que não serão utilizadas\n p[i] = p[i].filter(items=['time', 'wx'])\n p[i].columns = ['time', 'w']\n # removendo dados fora do intervalo alvo\n p[i] = p[i].loc[(p[i]['time'] >= tmin_p[i]) & (p[i]['time'] <= tmin_p[i] + 30)]\n p[i] = p[i].reset_index(drop = True)\n # 'zerando' o tempo\n p[i].update(pd.Series(p[i]['time'] - tmin_p[i], name = 'time'))\n # calculando a inclinação experimental\n p[i]['theta'] = integrate(p[i]['time'], p[i]['w'])\n # estimando a inclinação inicial\n theta0.append(-max(p[i]['theta']) / 2)\n # recalculando a inclinação experimental\n p[i].update(pd.Series(integrate(p[i]['time'], p[i]['w'], y0 = theta0[i]), name = 'theta'))\n # calculando a aceleração angular\n p[i]['a'] = derivate(p[i]['time'], p[i]['w'], y0 = g * math.sin(theta0[i]) / L)\n\n##################################\n# GRAFICOS EXPERIMENTAIS #\n##################################\n\n########## Bloco em rampa ######### \n\nplot_experimentos(r, ['d', 'v', 'a'], labels = ['Deslocamento Linear', 'Velocidade Linear', 'Aceleração Linear'], units = ['m', 'm/s', r'm/s$^2$'], exp = 'Bloco em Rampa')\n\n############### MCU ###############\n\nplot_experimentos(c, ['theta', 'w'], labels = ['Deslocamento Angular', 'Velocidade Angular'], units = ['rad', 'rad/s'], exp = 'Movimento Circular Uniforme')\n\n############# Pêndulo ############# \n\nplot_experimentos(p, ['theta', 'w', 'a'], labels = ['Deslocamento Angular', 'Velocidade Angular', 'Aceleração Angular'], units = ['rad', 'rad/s', r'rad/s$^2$'], exp = 'Movimento Pendular')\n\n##################################\n# MODELO #\n##################################\n\nre = r[0]\nce = c[0]\npe = p[0]\n\nfor i in range(1,5):\n re = re.append(r[i], ignore_index = True)\n ce = ce.append(c[i], ignore_index = True)\n pe = pe.append(p[i], ignore_index = True)\n\n########## Bloco em rampa #########\n\nBr = par * Ar / mr * 4\nAr = g * math.sin(thetar)\n\ntmax = max(re['time'])\n\ndt = 0.001\n\neur1 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'd' : 0.0,\n 'v' : 0.0,\n 'a' : 0.0})\n\neur1['d'][0] = 0.0\neur1['v'][0] = 0.0\neur1['a'][0] = Ar\n \nfor i in range(1, len(eur1)):\n eur1['d'][i] = eur1['d'][i-1] + eur1['v'][i-1] * dt\n eur1['v'][i] = eur1['v'][i-1] + eur1['a'][i-1] * dt\n eur1['a'][i] = Ar - Br * eur1['v'][i] ** 2\n\neur2 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'd' : 0.0,\n 'v' : 0.0,\n 'a' : 0.0})\n\neur2['d'][0] = 0.0\neur2['v'][0] = 0.0\neur2['a'][0] = Ar\n \nfor i in range(1, len(eur2)):\n vmid = eur2['v'][i-1] + eur2['a'][i-1] * dt / 2\n amid = Ar - Br * vmid ** 2\n eur2['v'][i] = eur2['v'][i-1] + amid * dt\n eur2['d'][i] = eur2['d'][i-1] + vmid * dt\n eur2['a'][i] = Ar - Br * eur2['v'][i] ** 2\n \nerro(re, eur1, eur2, 'd')\nerro(re, eur1, eur2, 'v')\nerro(re, eur1, eur2, 'a')\n\n\n############### MCU ###############\n\nwc = np.mean(ce['w'])\n\ntmax = max(ce['time'])\n \neuc1 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'theta' : 0.0})\n\neuc1['theta'][0] = 0.0\n \nfor i in range(1, len(euc1)):\n euc1['theta'][i] = euc1['theta'][i-1] + wc * dt\neuc1['w'] = wc\n\neuc2 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'theta' : 0.0})\n\neuc2['theta'][0] = 0.0\n \nfor i in range(1, len(euc2)):\n euc2['theta'][i] = euc2['theta'][i-1] + wc * dt\neuc2['w'] = wc\n \nerro(ce, euc1, euc2, 'theta')\nerro(ce, euc1, euc2, 'w')\n\n############# Pêndulo #############\n\nBp = par * Ap * L / mp\nthetap0 = np.mean(theta0)\nAp = g / L\n\ntmax = max(pe['time'])\n\neup1 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'theta' : 0.0,\n 'w' : 0.0,\n 'a' : 0.0})\n\neup1['theta'][0] = thetap0\neup1['w'][0] = 0.0\neup1['a'][0] = - Ap * math.sin(thetap0)\n \nfor i in range(1, len(eup1)):\n eup1['theta'][i] = eup1['theta'][i-1] + eup1['w'][i-1] * dt\n eup1['w'][i] = eup1['w'][i-1] + eup1['a'][i-1] * dt\n eup1['a'][i] = -Ap * math.sin(eup1['theta'][i]) - Bp * eup1['w'][i] * abs(eup1['w'][i])\n\neup2 = pd.DataFrame({'time' : list(frange(0.0, tmax, dt)),\n 'theta' : 0.0,\n 'w' : 0.0,\n 'a' : 0.0})\n\neup2['theta'][0] = thetap0\neup2['w'][0] = 0.0\neup2['a'][0] = - Ap * math.sin(thetap0)\n \nfor i in range(1, len(eup2)):\n thetamid = eup2['theta'][i-1] + eup2['w'][i-1] * dt / 2\n wmid = eup2['w'][i-1] + eup2['a'][i-1] * dt / 2\n amid = -Ap * math.sin(thetamid) - Bp * wmid * abs(wmid)\n eup2['theta'][i] = eup2['theta'][i-1] + wmid * dt\n eup2['w'][i] = eup2['w'][i-1] + amid * dt\n eup2['a'][i] = -Ap * math.sin(eup2['theta'][i]) - Bp * eup2['w'][i] * abs(eup2['w'][i])\n \nerro(pe, eup1, eup2, 'theta')\nerro(pe, eup1, eup2, 'w')\nerro(pe, eup1, eup2, 'a')\n \n \n##################################\n# GRAFICOS #\n##################################\n\n########## Bloco em rampa #########\n\nplot_modelos(re, eur1, eur2, ['d', 'v', 'a'], labels = ['Deslocamento Linear', 'Velocidade Linear', 'Aceleração Linear'], units = ['m', 'm/s', r'm/s$^2$'], exp = 'Bloco em Rampa')\n\n############### MCU ###############\n\nplot_modelos(ce, euc1, euc2, ['theta', 'w'], labels = ['Deslocamento Angular', 'Velocidade Angular'], units = ['rad', 'rad/s'], exp = 'Movimento Circular Uniforme')\n\n############# Pêndulo ############# \n\nplot_modelos(pe, eup1, eup2, ['theta', 'w', 'a'], labels = ['Deslocamento Angular', 'Velocidade Angular', 'Aceleração Angular'], units = ['rad', 'rad/s', r'rad/s$^2$'], exp = 'Movimento Pendular')\n","repo_name":"SisiBakaMuyo/mac0209","sub_path":"EP2/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":13259,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"34726769339","text":"alien_0 = {'color':'green','point':5}\nalien_1 = {'color':'red','point':15}\nalien_2 = {'color':'yellow','point':10}\naliens = [alien_0,alien_1,alien_2]\nprint('# 列表中存字典')\nfor alien in aliens:\n print(alien)\n\nprint('\\n# 10个外星人')\naliens=[]\nfor i in range(10):\n newAlien = {'speed':'slow','color':'green','point':5}\n aliens.append(newAlien)\n\nfor alien in aliens:\n print(alien)\nprint(len(aliens))\n\nprint('\\n# 改变前3个外星人的属性')\nfor alien in aliens[:3]:\n alien['color']='yellow'\n alien['speed']='medium'\n alien['point']=10\n\nfor alien in aliens:\n print(alien)\n\nprint('\\n# 字典中存列表````')\nprint(\"name_language = {\\n'me':['js','python'],\\n'fay':['java','js']\\n}\")\n\nname_language = {\n 'me':['js','python'],\n 'fay':['java','js'],\n 'lool':['java']\n}\nfor name,language in name_language.items():\n print(name)\n for l in language:\n print('\\t' + l)\n\nprint('\\n# 字典中存字典')\nusers ={\n 'Messie':{\n 'first':'Zhang',\n 'last':'Messie',\n 'location':'China'\n },\n\n 'Kaka':{\n 'first':'Sheng',\n 'last':'kaka',\n 'location':'Brazil'\n },\n}\nusers['C_ronarido'] = {\n 'first':'Stiyano',\n 'last':'C_ronarido',\n 'location':'Potora',\n}\nfor name,info in users.items():\n print('\\nPlayer name is ' + name)\n print('Player full name is ' + info['first'] + info['last'])\n print('Player is from ' + info['location'])\n","repo_name":"zd6515843/super_chen","sub_path":"PythonSelfStudy/Lession01-09/lession_06_嵌套.py","file_name":"lession_06_嵌套.py","file_ext":"py","file_size_in_byte":1441,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"32148721843","text":"import requests\nimport logging\nimport os\n\n\ndef post_to_server(video_title, accuracy, warning):\n # http://183.81.35.24:5010/api/msg_warning/Update?title=Em giờ nơi ấy có ổn không ? Em ừ đi anh thấy yên lòng ~ huy Vạc cover hay nhất 2019&accuracy=10&timer=12\n headers = {'content-type': 'application/json'}\n url = 'http://183.81.35.24:5010/api/msg_warning/Update'\n params = {'title': video_title, 'accuracy': accuracy, 'timer': '0', 'warning': warning}\n print(params)\n # req = requests.post(url, params=params, headers=headers)\n\n\nloggers_dict = {}\n# Logging\nLOGLEVEL = 20 # logging.INFO\nLOG_FORMAT = \"%(name)-20s %(levelname)-8s %(message)s\"\nFILE_LOGLEVEL = 10 # logging.DEBUG\nFILE_LOG_FORMAT = \"%(asctime)s %(name)-20s %(levelname)-8s %(message)s\"\n\n\n# --- Logging utilities --------------------------------------------------------\ndef add_file_handler(logger, log_filepath, loglevel=FILE_LOGLEVEL,\n log_format=FILE_LOG_FORMAT):\n \"\"\"Add a file handler to the logger.\"\"\"\n file_handler = logging.FileHandler(log_filepath)\n file_handler.setLevel(loglevel)\n file_handler.setFormatter(logging.Formatter(log_format))\n logger.addHandler(file_handler)\n\n\ndef setup_loggers(log_dir, loglevel=FILE_LOGLEVEL, log_format=FILE_LOG_FORMAT):\n \"\"\"Setup the loggers with file handlers.\"\"\"\n for name in logging.Logger.manager.loggerDict.keys():\n if name.startswith('music'):\n add_file_handler(\n logging.getLogger(name), os.path.join(log_dir, name + '.log'),\n loglevel, log_format)\n","repo_name":"Minh112/Check_copyright_video","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1589,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"4142736270","text":"import math\r\nnum1= int(input (\"numero por favor \"))\r\nnum2= int(input (\"numero por favor \"))\r\n\r\ntop=num2+1\r\n\r\nfor x in range(num1, top):\r\n verificador=0\r\n num=(x)\r\n raiz= int(math.sqrt(num)+1)\r\n for i in range (2,raiz):\r\n divisor= num%i\r\n if divisor==0:\r\n verificador+=1\r\n \r\n if verificador==0:\r\n print (\"El numero \" + format(num) + \" es primo y tiene \" + format(verificador) +\" divisores\")\r\n\r\n else:\r\n print (\"El numero \" + format(num) + \" no es primo\")\r\n","repo_name":"bibliofilica/NumerosPrimos","sub_path":"primo rango.py","file_name":"primo rango.py","file_ext":"py","file_size_in_byte":523,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"72225623902","text":"#!/usr/bin/python3\n# -*- encoding: utf-8 -*-\n# Graficas de los pulsos en los canales adquiridos.\n# Los archivos a analizar pueden tener la extensión .dat o .bz2\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport os\n\nN=5000 # Número de puntos a graficar\n\ndata_dir = 'data/' # Directorio donde se encuentran los datos\nplot_dir = 'plot/' # Directorio donde se guardaran las graficas\nfilename = input('Ingrese el nombre del archivo (*.dat o *.bz2) a procesar : ')\n#filename= 'spnk_nogps_2021_07_23_03h00.dat.bz2' # Nombre del archivo a graficar\n\nch1, ch2 = np.loadtxt(os.path.join(data_dir, filename), unpack=1, dtype=int)\n\nx=np.linspace(0,N,N-1)\nfig,ax = plt.subplots(nrows=1, ncols=1, figsize=(11,7))\n\nax.step(x,ch1[:N-1], 'r-.',lw=1, label='CH1')\nax.step(x,ch2[:N-1], 'b.-',lw=1, label='CH2')\n\n#ax.plot(ch1, 'r-o',lw=1, label='CH1')\n#ax.plot(ch2, 'b.-',lw=1, label='CH2')\n\nax.legend(fontsize=11)\nax.grid()\nax.set_xlim(0,1500)\nax.set_ylabel('Amplitude (ADC)',fontsize=14)\nax.set_xlabel('Time (ADC.bin)',fontsize=14)\n\nplt.savefig(os.path.join(plot_dir,'pulses_{}.png'.format(filename)))\nplt.show()\n\n","repo_name":"lharnaldi/acq_scripts","sub_path":"python/0_channels.py","file_name":"0_channels.py","file_ext":"py","file_size_in_byte":1106,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"3136789307","text":"import flask.json\nfrom flask import Flask, jsonify, render_template, request\nfrom flask_sqlalchemy import SQLAlchemy # pip install Flask-SQLAlchemy\nimport random\nimport json\nimport os\n\n# The API documentation is viewable at https://documenter.getpostman.com/view/15623189/TzRNEV41\n\n\nAPI_KEY = \"MySecretAPIKey\"\nFILE_URL = 'sqlite:///database/cafes.db'\napp = Flask(__name__)\n\n\"\"\"\n DEFAULT FLASK APP CONFIGURATION\n ===============================\n default_config = {\n 'APPLICATION_ROOT': '/',\n 'DEBUG': None,\n 'ENV': None,\n 'EXPLAIN_TEMPLATE_LOADING': False,\n 'JSONIFY_MIMETYPE': 'application/json',\n 'JSONIFY_PRETTYPRINT_REGULAR': False,\n 'JSON_AS_ASCII': True,\n 'JSON_SORT_KEYS': True,\n 'MAX_CONTENT_LENGTH': None,\n 'MAX_COOKIE_SIZE': 4093,\n 'PERMANENT_SESSION_LIFETIME': datetime.timedelta(days = 31),\n 'PREFERRED_URL_SCHEME': 'http',\n 'PRESERVE_CONTEXT_ON_EXCEPTION': None,\n 'PROPAGATE_EXCEPTIONS': None,\n 'SECRET_KEY': None,\n 'SEND_FILE_MAX_AGE_DEFAULT': None,\n 'SERVER_NAME': None,\n 'SESSION_COOKIE_DOMAIN': None,\n 'SESSION_COOKIE_HTTPONLY': True,\n 'SESSION_COOKIE_NAME': 'session',\n 'SESSION_COOKIE_PATH': None,\n 'SESSION_COOKIE_SAMESITE': None,\n 'SESSION_COOKIE_SECURE': False,\n 'SESSION_REFRESH_EACH_REQUEST': True,\n 'TEMPLATES_AUTO_RELOAD': None,\n 'TESTING': False,\n 'TRAP_BAD_REQUEST_ERRORS': None,\n 'TRAP_HTTP_EXCEPTIONS': False,\n 'USE_X_SENDFILE': False\n }\n\"\"\"\n\n# Connect to Database\napp.config['SQLALCHEMY_DATABASE_URI'] = FILE_URL\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(app)\n\n\n# Cafe TABLE Configuration\nclass Cafe(db.Model):\n __tablename__ = \"Cafes\"\n\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String(250), unique=True, nullable=False)\n map_url = db.Column(db.String(500), nullable=False)\n img_url = db.Column(db.String(500), nullable=False)\n location = db.Column(db.String(250), nullable=False)\n seats = db.Column(db.String(250), nullable=False)\n has_toilet = db.Column(db.Boolean, nullable=False)\n has_wifi = db.Column(db.Boolean, nullable=False)\n has_sockets = db.Column(db.Boolean, nullable=False)\n can_take_calls = db.Column(db.Boolean, nullable=False)\n coffee_price = db.Column(db.String(250), nullable=True)\n\n # Angela's method: Convert database record to a dictionary\n def to_dict(self):\n # # Method 1.\n # dictionary = {}\n # # Loop through each column in the data record\n # for column in self.__table__.columns:\n # # Create a new dictionary entry;\n # # where the key is the name of the column\n # # and the value is the value of the column\n # dictionary[column.name] = getattr(self, column.name)\n # return dictionary\n\n # Method 2. Alternatively use Dictionary Comprehension to do the same thing.\n return {column.name: getattr(self, column.name) for column in self.__table__.columns}\n\n\n# Create the database file and tables\nif not os.path.isfile(FILE_URL):\n db.create_all()\n\n\n# HOME PAGE\n\n@app.route(\"/\")\ndef home():\n return render_template(\"index.html\")\n \n\n# HTTP GET - Read Records\n\n@app.route(\"/random\", methods=[\"GET\"])\ndef get_random_cafe():\n # get a random cafe from the database\n # Select all results from the search \n all_cafes = db.session.query(Cafe).all()\n random_cafe = random.choice(all_cafes)\n # Turn the random cafe SQLAlchemy Object into a JSON Response object\n\n # # The original manual dictionary method\n # cafe = {\n # 'id': random_cafe.id,\n # 'name': random_cafe.name,\n # 'map_url': random_cafe.map_url,\n # 'img_url': random_cafe.img_url,\n # 'location': random_cafe.location,\n # 'seats': random_cafe.seats,\n # 'has_toilet': random_cafe.has_toilet,\n # 'has_wifi': random_cafe.has_wifi,\n # 'has_sockets': random_cafe.has_sockets,\n # 'can_take_calls': random_cafe.can_take_calls,\n # 'coffee_price': random_cafe.coffee_price,\n # }\n # return jsonify(cafe=cafe)\n\n # # This method saves me having to manually type out the dictionary\n # cafe = jsonify(\n # # jsonify the dictionary\n # cafe=jsonify(\n # # jsonify the cafe data\n # id=random_cafe.id,\n # name=random_cafe.name,\n # map_url=random_cafe.map_url,\n # img_url=random_cafe.img_url,\n # location=random_cafe.location,\n # seats=random_cafe.seats,\n # has_toilet=random_cafe.has_toilet,\n # has_wifi=random_cafe.has_wifi,\n # has_sockets=random_cafe.has_sockets,\n # can_take_calls=random_cafe.can_take_calls,\n # coffee_price=random_cafe.coffee_price,\n # ).json # convert the Response object to a dictionary\n # )\n # return cafe\n\n # Even better solution from Angela: add to_dict() function to the class\n # Simply convert the random_cafe data record to a dictionary of key-value pairs.\n # 200 \tOK \tAction completed successfully\n return jsonify(cafes=random_cafe.to_dict()), 200\n\n\n@app.route(\"/all\", methods=[\"GET\"])\ndef get_all_cafes():\n # get all cafes from the database\n # Select all results from the search \n all_cafes = db.session.query(Cafe).all()\n # combine into a list of dictionaries\n all_cafes_dict = [cafe.to_dict() for cafe in all_cafes]\n \"\"\"\n {\n \"all_cafes\": [\n {\n \"can_take_calls\": true, \n \"coffee_price\": \"\\u00a32.40\", \n \"has_sockets\": true, \n \"has_toilet\": true, \n \"has_wifi\": false, \n \"id\": 1, \n \"img_url\": \"https://atlondonbridge.com/wp-content/uploads/2019/02/Pano_9758_9761-Edit-190918_LTS_Science_Gallery-Medium-Crop-V2.jpg\", \n \"location\": \"London Bridge\", \n \"map_url\": \"https://g.page/scigallerylon?share\", \n \"name\": \"Science Gallery London\", \n \"seats\": \"50+\"\n }, \n ...\n ]\n }\n \"\"\"\n # 200 \tOK \tAction completed successfully\n return jsonify(cafes=all_cafes_dict), 200\n\n\n@app.route(\"/search\", methods=[\"GET\"])\ndef find_cafes():\n # Get value from URL query string e.g. http://127.0.0.1:5006/search?loc=Peckham\n # https://flask.palletsprojects.com/en/1.1.x/api/#flask.Request.args\n location = request.args.get('loc')\n # Select all results from the search \n found_cafes = db.session.query(Cafe).filter_by(location=location).all()\n if found_cafes:\n # combine into a list of dictionaries\n # 200 \tOK \tAction completed successfully\n return jsonify(cafes=[cafe.to_dict() for cafe in found_cafes]), 200\n else:\n # 404 \tNot Found \tRequested file was not found\n return jsonify(error={\"Not Found\": \"Sorry, we don't have a cafe at that location.\"}), 404\n\n\n# HTTP POST - Create Record\n\n@app.route(\"/add\", methods=[\"POST\"])\ndef add_cafe():\n def check_bool(value):\n return 1 if value == '1' or value.lower() == 'true' else 0\n\n # Get field values from request body and create a new Cafe object\n # request.form returns an immutable dictionary (ImmutableMultiDict)\n # the boolean values must be blank, 0 or 1\n data = request.form\n new_cafe = Cafe(\n name=data['name'],\n map_url=data['map_url'],\n img_url=data['img_url'],\n location=data['location'],\n seats=data['seats'],\n has_toilet=check_bool(data['has_toilet']),\n has_wifi=check_bool(data['has_wifi']),\n has_sockets=check_bool(data['has_sockets']),\n can_take_calls=check_bool(data['can_take_calls']),\n coffee_price=data['coffee_price'],\n )\n # Check if cafe is already in the database\n # Select all results from the search \n search_cafe = db.session.query(Cafe).filter_by(\n name=new_cafe.name,\n location=new_cafe.location\n ).all()\n if search_cafe:\n # 400 \tBad Request\n # Request had bad syntax or was impossible to fulfill\n return jsonify(error={\"exists\": \"Cafe already exists.\"}), 400\n else:\n # Add cafe to database\n db.session.add(new_cafe)\n db.session.commit()\n # 200 \tOK \tAction completed successfully\n return jsonify(response={\"success\": \"Successfully added the new cafe.\"}), 200\n\n\n# HTTP PUT/PATCH - Update Record\n\n@app.route(\"/update-price/\", methods=[\"PATCH\"])\ndef update_price(cafe_id):\n # Get value from URL query string e.g. http://127.0.0.1:5006/search?loc=Peckham\n # https://flask.palletsprojects.com/en/1.1.x/api/#flask.Request.args\n new_price = request.args.get('new-price')\n # Select the first match from the search: \n cafe = db.session.query(Cafe).filter_by(\n id=int(cafe_id),\n ).first()\n # Angela used:\n # cafe = db.session.query(Cafe).get(cafe_id)\n if cafe:\n # print(\"cafe =\", cafe, type(cafe)) # >>> cafe = \n cafe.coffee_price = new_price\n db.session.commit()\n # 200 \tOK \tAction completed successfully\n return jsonify(response={\"success\": f\"Price updated to {cafe.coffee_price} for {cafe.name}.\"}), 200\n else:\n # 404 \tNot Found \tRequested file was not found\n return jsonify(error={\"Not Found\": f\"A cafe with ID={cafe_id} was not found.\"}), 404\n\n\n# HTTP DELETE - Delete Record\n\n@app.route(\"/report-closed/\", methods=[\"DELETE\"])\ndef delete_cafe(cafe_id):\n key_data = request.form['api-key']\n if key_data == API_KEY:\n # delete the entry\n cafe = db.session.query(Cafe).get(cafe_id)\n if cafe:\n db.session.delete(cafe)\n db.session.commit()\n # 200 \tOK \tAction completed successfully\n return jsonify(response={\"success\": f\"Cafe {cafe.name} removed from database.\"}), 200\n else:\n # 404 \tNot Found \tRequested file was not found\n return jsonify(error={\"Not Found\": f\"A cafe with ID={cafe_id} was not found.\"}), 404\n else:\n # 403 \tForbidden\n # Request does not specify the file name, or the directory\n # or the file does not have the permission that allows the pages to be viewed from the web\n return jsonify(error={\"Forbidden\": \"Not Authorized to delete a cafe.\"}), 403\n pass\n\n\nif __name__ == '__main__':\n app.run(debug=True, port=5006)\n","repo_name":"SadSack963/day-66_REST_API","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":10628,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"42373131995","text":"import sys\nimport os\nimport glob\nimport time\nimport json\nimport firebase_admin\n\nfrom datetime import datetime\nfrom firebase_admin import credentials\nfrom google.cloud import firestore\n\nos.system('modprobe w1-gpio')\nos.system('modprobe w1-therm')\n\nbase_dir = '/sys/bus/w1/devices/'\ndevice_folder = glob.glob(base_dir + '28*')[0]\ndevice_file = device_folder + '/w1_slave'\n\nproject_id = \"smart-water-168ca\"\n\nos.environ[\"GOOGLE_APPLICATION_CREDENTIALS\"] = \"/home/pi/Desktop/smart-water-168ca-d9f996738e27.json\"\ncred = credentials.ApplicationDefault()\nfirebase_admin.initialize_app(cred, {\n 'projectId': project_id,\n})\ndb = firestore.Client()\n \ndef generateID():\n data = datetime.now()\n sensorBase_ref = db.collection(u'sensorBase').document()\n sensorBase_ref.set({\n \"registration\": data,\n \"location\": \"\"\n })\n return sensorBase_ref.id\n \ndef getSensorBase_id():\n overwrite = False\n id_sensorBase = 0\n \n try:\n with open('sensorBase.json', 'r') as f:\n \n jsonfile = f.read()\n sensorBase_json = json.loads(jsonfile)\n\n if sensorBase_json['id_sensorBase'] == False:\n id_sensorBase = generateID()\n overwrite = True\n else:\n id_sensorBase = sensorBase_json['id_sensorBase']\n \n except:\n id_sensorBase = generateID()\n overwrite = True\n \n if overwrite == True:\n with open('sensorBase.json', 'w+') as f: \n datastore = {\"id_sensorBase\": id_sensorBase}\n json.dump(datastore, f)\n \n return id_sensorBase\n\nidSensorBase= getSensorBase_id()\n\ndef read_temp_raw():\n f = open(device_file, 'r')\n lines = f.readlines()\n f.close()\n return lines\n \ndef read_temp():\n lines = read_temp_raw()\n while lines[0].strip()[-3:] != 'YES':\n time.sleep(0.2)\n lines = read_temp_raw()\n equals_pos = lines[1].find('t=')\n if equals_pos != -1:\n temp_string = lines[1][equals_pos+2:]\n temp_c = float(temp_string) / 1000.0\n temp_f = temp_c * 9.0 / 5.0 + 32.0\n \n data = {\n \"celsius\": temp_c,\n }\n \n return data\n\nwhile True:\n\n jsonTemp = read_temp()\n data = datetime.now()\n jsonTemp['id_sensorBase'] =idSensorBase\n jsonTemp['datatime'] =data\n\n doc_base = db.collection(u'sensorBase/'+idSensorBase+\"/sensorData\")\n doc_base.add(jsonTemp)\n print(\"Deu Certo!\")\n time.sleep(10)\n\n\n","repo_name":"maathh/Water-IOT","sub_path":"Raspberry/base_monitoramento.py","file_name":"base_monitoramento.py","file_ext":"py","file_size_in_byte":2487,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"24559822967","text":"from typing import Any, Dict, List, Optional, cast\n\nimport requests\nfrom langchain_core.callbacks import CallbackManagerForLLMRun\nfrom langchain_core.language_models.llms import LLM\nfrom langchain_core.pydantic_v1 import BaseModel, Extra, SecretStr, root_validator\nfrom langchain_core.utils import convert_to_secret_str, get_from_dict_or_env\n\n\nclass AI21PenaltyData(BaseModel):\n \"\"\"Parameters for AI21 penalty data.\"\"\"\n\n scale: int = 0\n applyToWhitespaces: bool = True\n applyToPunctuations: bool = True\n applyToNumbers: bool = True\n applyToStopwords: bool = True\n applyToEmojis: bool = True\n\n\nclass AI21(LLM):\n \"\"\"AI21 large language models.\n\n To use, you should have the environment variable ``AI21_API_KEY``\n set with your API key or pass it as a named parameter to the constructor.\n\n Example:\n .. code-block:: python\n\n from langchain_community.llms import AI21\n ai21 = AI21(ai21_api_key=\"my-api-key\", model=\"j2-jumbo-instruct\")\n \"\"\"\n\n model: str = \"j2-jumbo-instruct\"\n \"\"\"Model name to use.\"\"\"\n\n temperature: float = 0.7\n \"\"\"What sampling temperature to use.\"\"\"\n\n maxTokens: int = 256\n \"\"\"The maximum number of tokens to generate in the completion.\"\"\"\n\n minTokens: int = 0\n \"\"\"The minimum number of tokens to generate in the completion.\"\"\"\n\n topP: float = 1.0\n \"\"\"Total probability mass of tokens to consider at each step.\"\"\"\n\n presencePenalty: AI21PenaltyData = AI21PenaltyData()\n \"\"\"Penalizes repeated tokens.\"\"\"\n\n countPenalty: AI21PenaltyData = AI21PenaltyData()\n \"\"\"Penalizes repeated tokens according to count.\"\"\"\n\n frequencyPenalty: AI21PenaltyData = AI21PenaltyData()\n \"\"\"Penalizes repeated tokens according to frequency.\"\"\"\n\n numResults: int = 1\n \"\"\"How many completions to generate for each prompt.\"\"\"\n\n logitBias: Optional[Dict[str, float]] = None\n \"\"\"Adjust the probability of specific tokens being generated.\"\"\"\n\n ai21_api_key: Optional[SecretStr] = None\n\n stop: Optional[List[str]] = None\n\n base_url: Optional[str] = None\n \"\"\"Base url to use, if None decides based on model name.\"\"\"\n\n class Config:\n \"\"\"Configuration for this pydantic object.\"\"\"\n\n extra = Extra.forbid\n\n @root_validator()\n def validate_environment(cls, values: Dict) -> Dict:\n \"\"\"Validate that api key exists in environment.\"\"\"\n ai21_api_key = convert_to_secret_str(\n get_from_dict_or_env(values, \"ai21_api_key\", \"AI21_API_KEY\")\n )\n values[\"ai21_api_key\"] = ai21_api_key\n return values\n\n @property\n def _default_params(self) -> Dict[str, Any]:\n \"\"\"Get the default parameters for calling AI21 API.\"\"\"\n return {\n \"temperature\": self.temperature,\n \"maxTokens\": self.maxTokens,\n \"minTokens\": self.minTokens,\n \"topP\": self.topP,\n \"presencePenalty\": self.presencePenalty.dict(),\n \"countPenalty\": self.countPenalty.dict(),\n \"frequencyPenalty\": self.frequencyPenalty.dict(),\n \"numResults\": self.numResults,\n \"logitBias\": self.logitBias,\n }\n\n @property\n def _identifying_params(self) -> Dict[str, Any]:\n \"\"\"Get the identifying parameters.\"\"\"\n return {**{\"model\": self.model}, **self._default_params}\n\n @property\n def _llm_type(self) -> str:\n \"\"\"Return type of llm.\"\"\"\n return \"ai21\"\n\n def _call(\n self,\n prompt: str,\n stop: Optional[List[str]] = None,\n run_manager: Optional[CallbackManagerForLLMRun] = None,\n **kwargs: Any,\n ) -> str:\n \"\"\"Call out to AI21's complete endpoint.\n\n Args:\n prompt: The prompt to pass into the model.\n stop: Optional list of stop words to use when generating.\n\n Returns:\n The string generated by the model.\n\n Example:\n .. code-block:: python\n\n response = ai21(\"Tell me a joke.\")\n \"\"\"\n if self.stop is not None and stop is not None:\n raise ValueError(\"`stop` found in both the input and default params.\")\n elif self.stop is not None:\n stop = self.stop\n elif stop is None:\n stop = []\n if self.base_url is not None:\n base_url = self.base_url\n else:\n if self.model in (\"j1-grande-instruct\",):\n base_url = \"https://api.ai21.com/studio/v1/experimental\"\n else:\n base_url = \"https://api.ai21.com/studio/v1\"\n params = {**self._default_params, **kwargs}\n self.ai21_api_key = cast(SecretStr, self.ai21_api_key)\n response = requests.post(\n url=f\"{base_url}/{self.model}/complete\",\n headers={\"Authorization\": f\"Bearer {self.ai21_api_key.get_secret_value()}\"},\n json={\"prompt\": prompt, \"stopSequences\": stop, **params},\n )\n if response.status_code != 200:\n optional_detail = response.json().get(\"error\")\n raise ValueError(\n f\"AI21 /complete call failed with status code {response.status_code}.\"\n f\" Details: {optional_detail}\"\n )\n response_json = response.json()\n return response_json[\"completions\"][0][\"data\"][\"text\"]\n","repo_name":"langchain-ai/langchain","sub_path":"libs/community/langchain_community/llms/ai21.py","file_name":"ai21.py","file_ext":"py","file_size_in_byte":5292,"program_lang":"python","lang":"en","doc_type":"code","stars":68990,"dataset":"github-code","pt":"7"} +{"seq_id":"6666604706","text":"# -*- coding:utf-8 -*-\r\n\"\"\" Provide log related functions. You need to Initialize the logger and use the logger to make logs.\r\n\r\nExample:\r\n\r\n>>> logger = Initialize()\r\n\r\nUse logger.level(\\*msg) to log like:\r\n\r\n>>> logger.error(\"Pickle data writing Failed.\")\r\n\r\n>>> logger.info(\"Pickle data of \", foo, \" written successfully.\")\r\n\r\nThe log will be stored into LogFile.log by default.\r\n\"\"\"\r\n\r\n__author__ = \"Wang Hewen\"\r\n\r\nimport sys\r\nimport logging\r\n\r\nlogging.currentframe = lambda: sys._getframe(5)\r\nclass Logger(logging.Logger):\r\n def debug(self, *args, **kwargs):\r\n super().log(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def info(self, *args, **kwargs):\r\n super().info(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def warning(self, *args, **kwargs):\r\n super().warning(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def warn(self, *args, **kwargs):\r\n super().warn(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def error(self, *args, **kwargs):\r\n super().error(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def exception(self, *args, exc_info=True, **kwargs):\r\n super().exception(\"\".join([str(arg) for arg in args]), exc_info = exc_info, **kwargs)\r\n\r\n def critical(self, *args, **kwargs):\r\n super().critical(\"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def log(self, level, *args, **kwargs):\r\n super().log(level, \"\".join([str(arg) for arg in args]), **kwargs)\r\n\r\n def _log(self, level, msg, args, exc_info=None, extra=None, stack_info=False):\r\n super()._log(level, msg, args, exc_info=None, extra=None, stack_info=False)\r\n\r\n\r\ndef Initialize(FileName = \"LogFile.log\", LogLevel = \"INFO\", WriteToStream = False):\r\n '''\r\nInitialize loggers for logging. A logger will be returned.\r\n\r\n:param String FileName: Path of the log file\r\n:param String LogLevel: LogLevel of the logger, which can be \"DEBUG\", \"INFO\", \"ERROR\"\r\n:param Boolean WriteToStream: Whether to write to stdout\r\n:return: logger: The logger used for logging\r\n:rtype: logging.loggger\r\n '''\r\n if LogLevel not in [\"DEBUG\", \"INFO\", \"ERROR\"]:\r\n raise ValueError(\"LogLevel is not correctly set.\")\r\n logging.Logger.manager.setLoggerClass(Logger)\r\n logger = logging.getLogger(__name__) #__name__ == CommonModules.Log\r\n handlers = logger.handlers[:]\r\n for handler in handlers:\r\n handler.close()\r\n logger.removeHandler(handler)\r\n fileHandler = logging.FileHandler(FileName)\r\n fileHandler.setFormatter(logging.Formatter('%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s: %(message)s', datefmt = '%Y/%m/%d %H:%M:%S'))\r\n if LogLevel == \"DEBUG\":\r\n streamHandler = logging.StreamHandler(stream = sys.stdout)\r\n streamHandler.setLevel(logging.DEBUG)\r\n fileHandler.setLevel(logging.DEBUG)\r\n logger.setLevel(logging.DEBUG)\r\n if LogLevel == \"INFO\":\r\n streamHandler = logging.StreamHandler(stream = sys.stdout)\r\n streamHandler.setLevel(logging.INFO)\r\n fileHandler.setLevel(logging.INFO)\r\n logger.setLevel(logging.INFO)\r\n if LogLevel == \"ERROR\":\r\n streamHandler = logging.StreamHandler(stream = sys.stderr)\r\n streamHandler.setLevel(logging.ERROR)\r\n fileHandler.setLevel(logging.ERROR)\r\n logger.setLevel(logging.ERROR)\r\n\r\n streamHandler.setFormatter(logging.Formatter('%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s: %(message)s', datefmt = '%Y/%m/%d %H:%M:%S'))\r\n if WriteToStream:\r\n logger.addHandler(streamHandler) \r\n logger.addHandler(fileHandler)\r\n return logger\r\n","repo_name":"wanghewen/CommonModules","sub_path":"CommonModules/Log.py","file_name":"Log.py","file_ext":"py","file_size_in_byte":3611,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"7"} +{"seq_id":"14840831355","text":"import numpy as np\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.feature_selection import RFE\n\n\ndef rfe(X: np.array, y: np.array, n_features_to_select: int = 3):\n \"\"\"\n Performs recursive feature elimination\n\n Args:\n - X: Input numpy array (train_data)\n - y: Input numpy array (train_data labels)\n - n_features_to_select: number of features to keep\n\n Returns:\n - X_transformed: Output numpy array (with columns filtered)\n - y: numpy array (train_data labels)\n - rfe: fitted RFE object\n \"\"\"\n model = LinearRegression()\n rfe = RFE(\n estimator=model,\n n_features_to_select=n_features_to_select\n )\n\n rfe.fit(X, y)\n X_transformed = rfe.transform(X)\n return X_transformed, y, rfe","repo_name":"alanchn31/ML-ToolBox","sub_path":"supervised/feature_selection/rfe.py","file_name":"rfe.py","file_ext":"py","file_size_in_byte":778,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"6540510959","text":"from django.contrib import admin\nfrom .models import factura, codigoFinanciero\nfrom import_export.admin import ImportExportModelAdmin\n\n\nadmin.site.site_header = \"Gestion Pilar\"\nadmin.site.site_title = \"Gestion Pilar\"\n\n\n\n\n\n@admin.register(codigoFinanciero)\nclass codigoFinancieroAdmin(ImportExportModelAdmin):\n list_display= ('codigo',)\n\n@admin.register(factura)\nclass facturaAdmin(ImportExportModelAdmin):\n list_display=('emision', 'nroFactura', 'proveedor', 'total', 'objeto')\n\n\n def get_queryset(self, request):\n\n # Obtener el usuario actualmente autenticado\n user = request.user\n\n # Obtener el nombre del grupo al que pertenece el usuario\n user_group_name = user.groups.first().name if user.groups.exists() else None\n\n # Inicializar el queryset con todas las facturas\n queryset = super().get_queryset(request)\n\n # Filtrar las facturas por el nombre del grupo del usuario\n if user_group_name:\n queryset = queryset.filter(codigo__codigo=user_group_name)\n\n return queryset\n","repo_name":"Excel-ente/facturas","sub_path":"facturas/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1056,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70078597985","text":"# -*- coding: utf-8 -*-\nimport errno\n\nfrom .poll import POLL_READ, POLL_WRITE, POLL_ERROR, POLL_DISCONNECT\nfrom .error import BrokenPipeError, ConnectionError\nfrom ..future import FutureSourcePair, FutureCanceled, RaisedFuture, CompletedFuture\n\n__all__ = ('PollAwaiter',)\n#------------------------------------------------------------------------------#\n# Poll Awaiter #\n#------------------------------------------------------------------------------#\nclass PollAwaiter (object):\n \"\"\"File await object\n \"\"\"\n __slots__ = ('fd', 'poller', 'mask', 'entries',)\n\n def __init__ (self, fd, poller):\n self.fd = fd\n self.poller = poller\n\n # state\n self.mask = 0\n self.entries = []\n\n #--------------------------------------------------------------------------#\n # Await #\n #--------------------------------------------------------------------------#\n def Await (self, mask, cancel = None):\n \"\"\"Await event specified by mask argument\n \"\"\"\n if mask is None:\n self.Dispose (BrokenPipeError (errno.EPIPE, 'Detached from core'))\n return CompletedFuture (None)\n elif not mask:\n return RaisedFuture (ValueError ('Empty event mask'))\n elif mask & self.mask:\n return RaisedFuture (ValueError ('Intersecting event mask: {}'.format (self)))\n\n # source\n future, source = FutureSourcePair ()\n if cancel:\n def cancel_cont (result, error):\n self.dispatch (mask)\n source.TrySetCanceled ()\n cancel.Await ().OnCompleted (cancel_cont)\n\n # register\n if self.mask:\n self.poller.Modify (self.fd, self.mask | mask)\n else:\n self.poller.Register (self.fd, mask)\n\n # update state\n self.mask |= mask\n self.entries.append ((mask, source))\n\n return future\n\n #--------------------------------------------------------------------------#\n # Resolve #\n #--------------------------------------------------------------------------#\n def Resolve (self, event):\n \"\"\"Resolve pending events effected by specified event mask\n \"\"\"\n if event & ~POLL_ERROR:\n for source in self.dispatch (event):\n source.TrySetResult (event)\n\n else:\n error = BrokenPipeError (errno.EPIPE, 'Broken pipe') if event & POLL_DISCONNECT else \\\n ConnectionError ()\n for source in self.dispatch (self.mask):\n source.TrySetException (error)\n\n #--------------------------------------------------------------------------#\n # Private #\n #--------------------------------------------------------------------------#\n def dispatch (self, event):\n \"\"\"Dispatch sources effected by specified event mask\n \"\"\"\n entries, effected = [], []\n\n # find effected\n for mask, source in self.entries:\n if mask & event:\n effected.append (source)\n else:\n entries.append ((mask, source))\n\n # update state\n self.mask &= ~event\n self.entries = entries\n\n if self.mask:\n self.poller.Modify (self.fd, self.mask)\n else:\n self.poller.Unregister (self.fd)\n\n return effected\n\n def __str__ (self):\n \"\"\"String representation\n \"\"\"\n events = []\n self.mask & POLL_READ and events.append ('read')\n self.mask & POLL_WRITE and events.append ('write')\n self.mask & POLL_ERROR and events.append ('error')\n return ''.format (self.fd, ','.join (events), id (self))\n __repr__ = __str__\n\n #--------------------------------------------------------------------------#\n # Disposable #\n #--------------------------------------------------------------------------#\n def Dispose (self, error = None):\n \"\"\"Dispose file and resolve all pending events with specified error\n \"\"\"\n error = error or FutureCanceled ('File await object has been disposed')\n\n for source in self.dispatch (self.mask):\n source.TrySetException (error)\n\n def __enter__ (self):\n return self\n\n def __exit__ (self, et, eo, tb):\n self.Dispose ()\n return False\n\n# vim: nu ft=python columns=120 :\n","repo_name":"aslpavel/async","sub_path":"core/poll_await.py","file_name":"poll_await.py","file_ext":"py","file_size_in_byte":4690,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"14527919740","text":"class Solution:\n def removeElement(self, nums: List[int], val: int) -> int:\n \"\"\"\n Given a list of integers and an integer, remove any\n list entries which are equal to that integer in-place\n and return the number of remaining entries as an int.\n \"\"\"\n\n index = 0 \n while index < len(nums):\n # if same as value, remove\n if nums[index] == val:\n nums.pop(index)\n # if different, move onto next\n else:\n index += 1\n \n return index","repo_name":"Foggalong/leetcode","sub_path":"problems/remove_element/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":566,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37153552840","text":"# 다음 데이터는 동일한 상품의 포장지 색상에 따른 매출액에 대한 자료이다.\n# 포장지 색상에 따른 제품의 매출액에 차이가 존재하는지 two-sample t-검정을 하시오.\n'''\nimport numpy as np\nimport scipy.stats as stats\n\nblue = [70, 68, 82, 78, 72, 68, 67, 68, 88, 60, 80]\nred = [60, 65, 55, 58, 67, 59, 61, 68, 77, 66, 66]\n\n# 귀무: 포장지 색상에 따른 제품의 매출액에 차이가 없다.\n# 대립: 포장지 색상에 따른 제품의 매출액에 차이가 있다.\n\nprint(np.mean(blue)) # 72.8\nprint(np.mean(red)) # 63.8\n\ntwo_sample = stats.ttest_ind(blue, red)\nprint(two_sample)\n# Ttest_indResult(statistic=2.9280203225212174, pvalue=0.008316545714784403)\n# pvalue=0.008316 < 0.05 이므로 귀무가설 기각!\n# 포장지 색상에 따른 제품의 매출액에 차이가 있다.\n'''\n\n# 남아 신생아 몸무게의 평균 검정을 수행하려고 한다.\n# 파일명 : babyboom.csv (testdata 폴더에 있음) # 1:여아, 2:남아 \n\n# 남아 신생아의 몸무게는 평균이 3000(g)으로 알려져 왔으나 이것이 틀렸다는 주장이 나왔다.\n# 표본으로 남아를 뽑아 체중을 측정하였다고 할 때 새로운 주장이 맞는지 검정하시오.\n\n# 귀무 : 남아 신생아의 몸무게는 평균이 3000(g) 이다.\n# 대립 : 남아 신생아의 몸무게는 평균이 3000(g)이 아니다.\n\"\"\"\nimport numpy as np\nimport scipy.stats as stats\nimport pandas as pd\n\ndata = pd.read_csv('../testdata/babyboom.csv')\nprint(data.head(3),len(data)) # 44\nprint(data.isnull().sum()) # 결측치 0 개\nbdata = data[data['gender'] == 2]\nprint(bdata.head(3), len(bdata)) # 남아의 수는 26\nprint(np.mean(bdata.weight)) # 3375.30\n\nprint(stats.shapiro(bdata.weight)) # pvalue=0.2022 정규성 만족 \n\nprint(stats.ttest_1samp(bdata.weight, popmean= 3000))\n# Ttest_1sampResult(statistic=4.47078356044109, pvalue=0.00014690296107439875)\n# pvalue=0.0001469 < 0.05 이므로 귀무가설 기각\n\"\"\"\n\n# 에이콘 주식회사에서 영업사원들의 '지각횟수'와 '판매횟수' 간에 관계가 있는지 알아보려고 한다.\n# 영업사원 5명을 대상으로 한 달 동안 '지각횟수'와 '판매횟수'를 조사했더니 아래와 같은 결과를 얻었다.\n# 둘 사이의 상관계수를 출력하고 상관관계가 있는지 설명하시오.\n\"\"\"\nimport pandas as pd\nimport numpy as np\n\n# 지각횟수(x) = 1,2,3,4,5\n# 판매횟수(y) = 8,7,6,4,5\n\ndf = pd.DataFrame({'late':(1,2,3,4,5),'sale':(8,7,6,4,5)})\nprint(df.corr(method='pearson'))\nprint(np.corrcoef(df.late, df.sale))\n\"\"\"\n\n# 소득 수준에 따른 외식 성향을 나타내고 있다. 주말 저녁에 외식을 하면 1, 외식을 하지 않으면 0으로 처리되었다.\n# 'eat_out.txt' 데이터에 대하여 소득 수준이 외식에 영향을 미치는지 로지스틱 회귀분석을 실시한다.\n# ① 소스 코드와 모델의 분류정확도를 출력하시오.\n# ② 키보드로 소득 수준(양의 정수)을 입력하면 외식 여부 분류 결과 출력하시오.\n\n# 조건1 : 모델 생성은 glm 함수를 사용하도록 한다.\n# 조건2 : 키보드로 입력할 소득 수준 값은 45로 한다.\n\nimport pandas as pd\nimport statsmodels.api as sm\nimport statsmodels.formula.api as smf\nimport numpy as np\nfrom sklearn.metrics import accuracy_score\n\ndata = pd.read_csv('../testdata/last_test.csv')\nprint(data)\neatout = data[(data.요일 == '토')| (data.요일 == '일')]\nprint(eatout)\n\nformula = '외식유무 ~ 소득수준'\nresult = smf.glm(formula = formula, data = eatout, family = sm.families.Binomial()).fit() # binomial 을 넣기 때문에 이항분포가 된다\nprint(result)\nprint(result.summary())\n\npred = result.predict(eatout)\nprint('정확도: ', accuracy_score(eatout['외식유무'], np.around(pred)))\n\nkey = int(input('소득 수준 입력 : 45를 눌러주세요...'))\nnewdf = pd.DataFrame({'소득수준':[key]})\npred2 = result.predict(newdf)\nprint(np.rint(pred2.values))\n","repo_name":"tyler-0331/python","sub_path":"pypro3/anal5/11.py","file_name":"11.py","file_ext":"py","file_size_in_byte":3981,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"368841172","text":"import logging\nimport sys\nimport warnings\nimport numpy as np\n\nglobal logger\nlogger = logging.getLogger('simnibs')\nsh = logging.StreamHandler()\nformatter = logging.Formatter('[ %(name)s ]%(levelname)s: %(message)s')\nsh.setFormatter(formatter)\nsh.setLevel(logging.INFO)\nlogger.addHandler(sh)\nlogger.setLevel(logging.DEBUG)\nlogging.addLevelName(25, 'SUMMARY')\n\n\ndef log_warnings(message, category, filename, lineno, file=None, line=None):\n logger.warn(warnings.formatwarning(message, category, filename, lineno))\n\n\n# This is causing errors in pytest\n#warnings.showwarning = log_warnings\n\n\ndef register_excepthook(logger):\n def log_excep(exc_type, exc_value, exc_traceback):\n if issubclass(exc_type, KeyboardInterrupt):\n sys.__excepthook__(exc_type, exc_value, exc_traceback)\n return\n logger.debug(\n \"Traceback\",\n exc_info=(exc_type, exc_value, exc_traceback)\n )\n logger.critical(\n \"Uncaught exception\",\n exc_info=(exc_type, exc_value, None)\n )\n sys.excepthook = log_excep\n\n\ndef unregister_excepthook():\n sys.excepthook = sys.__excepthook__\n\n\ndef format_time(running_time):\n \"\"\"Format time in seconds as hours:minutes:seconds.\n \n PARAMETERS\n ----------\n running_time : float\n Time in seconds.\n \n RETURNS\n ----------\n running_time : str\n The time formatted as hours:minutes:seconds.\n \"\"\"\n hrs = np.uint16(np.floor(running_time/(60.**2)))\n mts = np.uint16(np.floor(running_time/60.-hrs*60))\n sec = np.uint16(np.round(running_time-hrs*60.**2-mts*60.))\n\n return \"{:02d}:{:02d}:{:02d}\".format(hrs,mts,sec)","repo_name":"simnibs/simnibs","sub_path":"simnibs/utils/simnibs_logger.py","file_name":"simnibs_logger.py","file_ext":"py","file_size_in_byte":1665,"program_lang":"python","lang":"en","doc_type":"code","stars":87,"dataset":"github-code","pt":"7"} +{"seq_id":"31785380706","text":"import matplotlib.pyplot as plt\nimport numpy as np\nimport tensorflow as tf\nfrom dataset import WindowGenerator\n\n\nclass Baseline(tf.keras.Model):\n def __init__(self, label_index=None):\n super().__init__()\n self.label_index = label_index\n\n def call(self, inputs):\n if self.label_index is None:\n return inputs\n result = inputs[:, :, self.label_index]\n return result[:, :, tf.newaxis]\n\n\nclass MultiStepLastBaseline(tf.keras.Model):\n def call(self, inputs):\n return tf.tile(inputs[:, -1:, :], [1, OUT_STEPS, 1])\n\n\nclass RepeatBaseline(tf.keras.Model):\n def call(self, inputs):\n return inputs\n\n\nclass ResidualWrapper(tf.keras.Model):\n def __init__(self, model):\n super().__init__()\n self.model = model\n\n def call(self, inputs, *args, **kwargs):\n delta = self.model(inputs, *args, **kwargs)\n # The prediction for each timestep is the input\n # from the previous time step plus the delta\n # calculated by the model.\n return inputs + delta\n\n\nclass FeedBack(tf.keras.Model):\n def __init__(self, units, out_steps):\n super().__init__()\n self.out_steps = out_steps\n self.units = units\n self.lstm_cell = tf.keras.layers.LSTMCell(units)\n # Also wrap the LSTMCell in an RNN to simplify the `warmup` method.\n self.lstm_rnn = tf.keras.layers.RNN(self.lstm_cell, return_state=True)\n self.dense = tf.keras.layers.Dense(num_features)\n\n def warmup(self, inputs):\n # inputs.shape => (batch, time, features)\n # x.shape => (batch, lstm_units)\n x, *state = self.lstm_rnn(inputs)\n # predictions.shape => (batch, features)\n prediction = self.dense(x)\n return prediction, state\n\n def call(self, inputs, training=None):\n # Use a TensorArray to capture dynamically unrolled outputs.\n predictions = []\n # Initialize the lstm state\n prediction, state = self.warmup(inputs)\n # Insert the first prediction\n predictions.append(prediction)\n\n # Run the rest of the prediction steps\n for n in range(1, self.out_steps):\n # Use the last prediction as input.\n x = prediction\n # Execute one lstm step.\n x, state = self.lstm_cell(x, states=state, training=training)\n # Convert the lstm output to a prediction.\n prediction = self.dense(x)\n # Add the prediction to the output\n predictions.append(prediction)\n\n # predictions.shape => (time, batch, features)\n predictions = tf.stack(predictions)\n # predictions.shape => (batch, time, features)\n predictions = tf.transpose(predictions, [1, 0, 2])\n return predictions\n\n\ndef compile_and_fit(model, window, patience=8):\n early_stopping = tf.keras.callbacks.EarlyStopping(monitor='val_loss',\n patience=patience,\n mode='min')\n model.compile(loss=tf.losses.MeanSquaredError(),\n optimizer=tf.optimizers.Adam(),\n metrics=[tf.metrics.MeanAbsoluteError()])\n history = model.fit(window.train, epochs=MAX_EPOCHS,\n validation_data=window.val,\n callbacks=[early_stopping])\n model.save(filepath='tmp/model/')\n return history\n\n\ndef plot_experiments(metric_idx, val_loss, train_loss, train_accs, val_accs, test_accs):\n x = np.arange(len(train_loss))\n val_mae = [v[metric_idx] for v in val_loss.values()]\n test_mae = [v[metric_idx] for v in train_loss.values()]\n\n plt.ylabel('mean_absolute_error [DIJA, normalized]')\n plt.bar(x - 0.17, val_mae, 0.3, label='Validation')\n plt.bar(x + 0.17, test_mae, 0.3, label='Test')\n plt.xticks(ticks=x, labels=train_loss.keys(), rotation=45)\n _ = plt.legend()\n plt.show()\n\n train_acc = [v for v in train_accs.values()]\n val_acc = [v for v in val_accs.values()]\n test_acc = [v for v in test_accs.values()]\n plt.ylabel('binary accuracy')\n plt.bar(x - 0.23, train_acc, 0.2, label='Train')\n plt.bar(x, val_acc, 0.2, label='Validation')\n plt.bar(x + 0.23, test_acc, 0.2, label='Test')\n plt.xticks(ticks=x, labels=train_loss.keys(), rotation=45)\n _ = plt.legend()\n plt.show()\n\n\ndef train_and_evaluate(model, window, plot_window, name):\n if name == 'Baseline' or name == 'Last' or name == 'Repeat':\n model.compile(loss=tf.losses.MeanSquaredError(), metrics=[tf.metrics.MeanAbsoluteError()])\n else:\n compile_and_fit(model=model, window=window)\n val_performance[name] = model.evaluate(window.val)\n performance[name] = model.evaluate(window.test, verbose=0)\n train_acc[name] = window.binary_accuracy(model=model, ds='train')\n val_acc[name] = window.binary_accuracy(model=model, ds='val')\n test_acc[name] = window.binary_accuracy(model=model, ds='test')\n plot_window.plot(name=name, model=model, acc=test_acc[name])\n\n\ndef run_single_step_models():\n baseline = Baseline(label_index=0)\n train_and_evaluate(baseline, single_step_window, wide_window, 'Baseline')\n\n linear = tf.keras.Sequential([tf.keras.layers.Dense(units=1)])\n train_and_evaluate(linear, single_step_window, wide_window, 'Linear')\n\n\n # dense = tf.keras.Sequential([\n # tf.keras.layers.Dense(units=64, activation='relu'),\n # tf.keras.layers.Dense(units=1)\n # ])\n # train_and_evaluate(dense, single_step_window, wide_window, 'Dense')\n\n # multi_step_dense = tf.keras.Sequential([\n # # Shape: (time, features) => (time*features)\n # tf.keras.layers.Flatten(),\n # tf.keras.layers.Dense(units=32, activation='relu'),\n # tf.keras.layers.Dense(units=1),\n # # Add back the time dimension.\n # # Shape: (outputs) => (1, outputs)\n # tf.keras.layers.Reshape([1, -1]),\n # ])\n # train_and_evaluate(multi_step_dense, conv_window, conv_window, 'Multi-step dense')\n #\n # conv_model = tf.keras.Sequential([\n # tf.keras.layers.Conv1D(filters=16,\n # kernel_size=(CONV_WIDTH,),\n # activation='relu'),\n # tf.keras.layers.Dense(units=16, activation='relu'),\n # tf.keras.layers.Dense(units=1),\n # ])\n # train_and_evaluate(conv_model, conv_window, wide_conv_window, 'Conv')\n #\n # lstm_model = tf.keras.models.Sequential([\n # # Shape [batch, time, features] => [batch, time, lstm_units]\n # tf.keras.layers.LSTM(32, return_sequences=True),\n # # Shape => [batch, time, features]\n # tf.keras.layers.Dense(units=1)\n # ])\n # train_and_evaluate(lstm_model, wide_window, wide_window, 'LSTM')\n\n metric_idx = baseline.metrics_names.index('mean_absolute_error')\n return metric_idx\n\n\ndef run_multi_output_models():\n baseline = Baseline(label_index=0)\n train_and_evaluate(baseline, single_step_window, wide_window, 'Baseline Multi-out')\n\n dense = tf.keras.Sequential([\n tf.keras.layers.Dense(units=64, activation='relu'),\n tf.keras.layers.Dense(units=64, activation='relu'),\n tf.keras.layers.Dense(units=num_features)\n ])\n train_and_evaluate(dense, single_step_window, wide_window, 'Dense Multi-out')\n\n lstm_model = tf.keras.models.Sequential([\n # Shape [batch, time, features] => [batch, time, lstm_units]\n tf.keras.layers.LSTM(32, return_sequences=True),\n # Shape => [batch, time, features]\n tf.keras.layers.Dense(units=num_features)\n ])\n train_and_evaluate(lstm_model, wide_window_multi_out, wide_window, 'LSTM Multi-out')\n\n residual_lstm = ResidualWrapper(\n tf.keras.Sequential([\n tf.keras.layers.LSTM(32, return_sequences=True),\n tf.keras.layers.Dense(\n num_features,\n # The predicted deltas should start small\n # So initialize the output layer with zeros\n kernel_initializer=tf.initializers.zeros())\n ]))\n train_and_evaluate(residual_lstm, wide_window_multi_out, wide_window, 'Residual LSTM Multi-out')\n\n metric_idx = baseline.metrics_names.index('mean_absolute_error')\n return metric_idx\n\n\ndef run_multi_step_models():\n last_baseline = MultiStepLastBaseline()\n train_and_evaluate(last_baseline, multi_window, multi_window, 'Last')\n\n repeat_baseline = RepeatBaseline()\n train_and_evaluate(repeat_baseline, multi_window, multi_window, 'Repeat')\n\n multi_linear_model = tf.keras.Sequential([\n # Take the last time-step.\n # Shape [batch, time, features] => [batch, 1, features]\n tf.keras.layers.Lambda(lambda x: x[:, -1:, :]),\n # Shape => [batch, 1, out_steps*features]\n tf.keras.layers.Dense(OUT_STEPS * num_features, kernel_initializer=tf.initializers.zeros()),\n # Shape => [batch, out_steps, features]\n tf.keras.layers.Reshape([OUT_STEPS, num_features])\n ])\n train_and_evaluate(multi_linear_model, multi_window, multi_window, 'Linear Multi')\n\n # multi_dense_model = tf.keras.Sequential([\n # # Take the last time step.\n # # Shape [batch, time, features] => [batch, 1, features]\n # tf.keras.layers.Lambda(lambda x: x[:, -1:, :]),\n # # Shape => [batch, 1, dense_units]\n # tf.keras.layers.Dense(512, activation='relu'),\n # # Shape => [batch, out_steps*features]\n # tf.keras.layers.Dense(OUT_STEPS * num_features, kernel_initializer=tf.initializers.zeros()),\n # # Shape => [batch, out_steps, features]\n # tf.keras.layers.Reshape([OUT_STEPS, num_features])\n # ])\n # train_and_evaluate(multi_dense_model, multi_window, multi_window, 'Dense Multi')\n #\n # multi_conv_model = tf.keras.Sequential([\n # # Shape [batch, time, features] => [batch, CONV_WIDTH, features]\n # tf.keras.layers.Lambda(lambda x: x[:, -CONV_WIDTH:, :]),\n # # Shape => [batch, 1, conv_units]\n # tf.keras.layers.Conv1D(1024, activation='relu', kernel_size=CONV_WIDTH),\n # # Shape => [batch, 1, out_steps*features]\n # tf.keras.layers.Dense(OUT_STEPS * num_features, kernel_initializer=tf.initializers.zeros()),\n # # Shape => [batch, out_steps, features]\n # tf.keras.layers.Reshape([OUT_STEPS, num_features])\n # ])\n # train_and_evaluate(multi_conv_model, multi_window, multi_window, 'Conv Multi')\n #\n # multi_lstm_model = tf.keras.Sequential([\n # # Shape [batch, time, features] => [batch, lstm_units]\n # # Adding more `lstm_units` just overfits more quickly.\n # tf.keras.layers.LSTM(32, return_sequences=False),\n # # Shape => [batch, out_steps*features]\n # tf.keras.layers.Dense(OUT_STEPS * num_features, kernel_initializer=tf.initializers.zeros()),\n # # Shape => [batch, out_steps, features]\n # tf.keras.layers.Reshape([OUT_STEPS, num_features])\n # ])\n # train_and_evaluate(multi_lstm_model, multi_window, multi_window, 'LSTM Multi')\n #\n # feedback_model = FeedBack(units=32, out_steps=OUT_STEPS)\n # train_and_evaluate(feedback_model, multi_window, multi_window, 'AR LSTM')\n\n metric_idx = repeat_baseline.metrics_names.index('mean_absolute_error')\n return metric_idx\n\n\nif __name__ == \"__main__\":\n val_performance, performance, train_acc, val_acc, test_acc = {}, {}, {}, {}, {}\n num_features = 118\n MAX_EPOCHS = 200\n CONV_WIDTH = 5\n LABEL_WIDTH = 24\n INPUT_WIDTH = LABEL_WIDTH + (CONV_WIDTH - 1)\n OUT_STEPS = 24\n single_step_window = WindowGenerator(\n input_width=1, label_width=1, shift=1, label_columns=['DIJA'])\n wide_window = WindowGenerator(\n input_width=24, label_width=24, shift=1, label_columns=['DIJA'])\n wide_window_multi_out = WindowGenerator(\n input_width=24, label_width=24, shift=1)\n conv_window = WindowGenerator(\n input_width=CONV_WIDTH, label_width=1, shift=1, label_columns=['DIJA'])\n wide_conv_window = WindowGenerator(\n input_width=INPUT_WIDTH, label_width=LABEL_WIDTH, shift=1, label_columns=['DIJA'])\n multi_window = WindowGenerator(\n input_width=24, label_width=OUT_STEPS, shift=OUT_STEPS)\n\n metric_index = run_single_step_models()\n # _ = run_multi_output_models()\n # _ = run_multi_step_models()\n plot_experiments(metric_index, val_performance, performance, train_acc, val_acc, test_acc)\n","repo_name":"ricky-ma/NLPStonks","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":12401,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"31680989939","text":"from lcapy import s, sympify, j, sqrt\n\nzeta = sympify('zeta_', real=True)\nomega0 = sympify('omega0', real=True)\n\np1a = -zeta * omega0 + j * omega0 * sqrt(1 - zeta**2)\np1b = -zeta * omega0 - j * omega0 * sqrt(1 - zeta**2)\n\nalpha1 = sympify('alpha_1', real=True)\nomega1 = sympify('omega_1', real=True)\n\np1a = -alpha1 + j * omega1\np1b = -alpha1 - j * omega1\n\nH = p1a * p1b / ((s - p1a) * (s - p1b))\n\nG = H / s\n\ng = G.partfrac().inverse_laplace(causal=True)\nh = H.partfrac().inverse_laplace(causal=True)\n\ndef topy(expr):\n\n s = str(expr)\n s = s.replace('**', '^').replace('_', '').replace('*', ' * ').replace('^', '**').replace('Heaviside(t)', '(t >= 0)').replace('/', ' / ') \n return s\n\nprint(topy(g))\n\nprint(topy(h))\n","repo_name":"mph-/dsp-notebooks","sub_path":"intro/responses/pzlp2ud.py","file_name":"pzlp2ud.py","file_ext":"py","file_size_in_byte":726,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"7607537115","text":"import time\n\n\nclass NotInteger(Exception):\n pass\n\n\nprint(\"First task\")\n\ntry:\n input_number = int(input(\"Enter a number or an Error will be thrown: \"))\n if type(input_number) != int:\n raise TypeError\nexcept (TypeError, ValueError):\n print(f'You entered an invalid number and an Error was thrown!')\n\n\ntime.sleep(2)\nprint(\"Second Task\")\n\ntry:\n second_task_number = int(input(\"Enter a number for the second task: \"))\n if type(second_task_number) != int:\n raise TypeError\nexcept (TypeError, ValueError):\n print(f'The number that you entered for the second task was invalid and an Error was thrown!')\nfinally:\n time.sleep(1)\n print(f'The entering phase ended.')\n\n\ntime.sleep(2)\nprint('Third task')\n\ntry:\n third_task_number = int(input(\"Enter a number for the third task: \"))\n if type(third_task_number) != int:\n raise NotInteger\nexcept (NotInteger, ValueError):\n time.sleep(1)\n print(\"The input data is not an integer!\")\n\ntime.sleep(2)\nprint(f'Fourth task')\ntime.sleep(1)\n\nclass GymPRs:\n def __init__(self, name, squat, deadlift, bench):\n self.name = name\n self.squat = squat\n self.deadlift = deadlift\n self.bench = bench\n\n def flex(self):\n print(f'{self.name} says: \"Yoo bro, my PRs are - deadlift: {self.deadlift}kg, benchpress: {self.bench}kg, squat: {self.squat}kg\"')\n\n def oneThousandPound(self):\n total_weight = self.squat + self.deadlift + self.bench\n total_weight_kg = total_weight * 0.45359237\n if total_weight_kg >= 453.592:\n print(\"Bro is in the 1000 pound club!\")\n else:\n remaining_weight_lb = round(1000 - (total_weight * 2.20462), 1)\n print(f\"Worthy for the 1000 pound club?\\nYou have to train more! You need {remaining_weight_lb}lbs more to reach the club.\")\n\n\nmy_Prs = GymPRs(\"Peter\", 110, 202, 100)\nmy_Prs.flex()\nmy_Prs.oneThousandPound()\n\nthe_data_is_valid = True\ntry:\n trainee_name = input(\"Enter your name: \")\n if type(trainee_name) != str:\n raise ValueError\nexcept ValueError:\n print(f'The name that you entered is not a string, and the program rose an Error!')\n the_data_is_valid = False\n\ntry:\n trainee_squat = int(input(\"Enter your squat PR: \"))\n if type(trainee_squat) != int:\n raise ValueError\nexcept ValueError:\n print(f'The squat PR that you entered is not valid and the program rose an Error!')\n the_data_is_valid = False\n\ntry:\n trainee_deadlift = int(input(\"Enter your deadlift PR: \"))\n if type(trainee_deadlift) != int:\n raise ValueError\nexcept ValueError:\n print(f'The deadlift PR that you entered is not valid and the program rose an Error!')\n the_data_is_valid = False\n\ntry:\n trainee_benchpress = int(input(\"Enter your benchpress PR: \"))\n if type(trainee_benchpress) != int:\n raise ValueError\nexcept ValueError:\n print(f'The benchpress PR that you entered is not valid and the program rose an Error!')\n the_data_is_valid = False\n\nif the_data_is_valid:\n time.sleep(1)\n trainee_Prs = GymPRs(trainee_name, trainee_squat, trainee_deadlift, trainee_benchpress)\n time.sleep(1)\n trainee_Prs.flex()\n time.sleep(1)\n trainee_Prs.oneThousandPound()\nelse:\n print(f'The class could not be called because you entered invalid data. Sorry.')\n","repo_name":"PowerCell46/Python-Programming-Basics","sub_path":"University Python Tasks/Lecture 8 Tasks.py","file_name":"Lecture 8 Tasks.py","file_ext":"py","file_size_in_byte":3316,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"412252395","text":"#!/usr/bin/env python3\n\nwith open('input1.txt', 'r') as f:\n lines = f.read().splitlines()\n\nclass Pack(object):\n def __init__(self):\n self.packs = list()\n self.total = 0\n\n def add(self, val):\n self.packs.append(val)\n self.total += val\n\npacks = [Pack()]\n\nfor line in lines:\n line = line.strip()\n if line == '':\n packs.append(Pack())\n continue\n val = int(line)\n packs[-1].add(val)\n\nbig = None\n\nfor pack in packs:\n if big is None or big.total < pack.total:\n big = pack\n\nprint(big.total)\n\n\n","repo_name":"mgeorg/advent_of_code2022","sub_path":"day01/day1p1.py","file_name":"day1p1.py","file_ext":"py","file_size_in_byte":514,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"71570508703","text":"from flask import Blueprint, jsonify, current_app\n\nfrom app.config import QueueNames\nfrom app.delivery import send_to_providers\nfrom app.models import EMAIL_TYPE\nfrom app.celery import provider_tasks\nfrom app.dao import notifications_dao\nfrom app.errors import register_errors\n\ndelivery_blueprint = Blueprint('delivery', __name__)\n\n\nregister_errors(delivery_blueprint)\n\n\n@delivery_blueprint.route('/deliver/notification/', methods=['POST'])\ndef send_notification_to_provider(notification_id):\n notification = notifications_dao.get_notification_by_id(notification_id)\n if not notification:\n return jsonify({\"result\": \"error\", \"message\": \"No result found\"}), 404\n\n if notification.notification_type == EMAIL_TYPE:\n send_response(\n send_to_providers.send_email_to_provider,\n provider_tasks.deliver_email,\n notification,\n QueueNames.SEND_EMAIL\n )\n else:\n send_response(\n send_to_providers.send_sms_to_provider,\n provider_tasks.deliver_sms,\n notification,\n QueueNames.SEND_SMS\n )\n return jsonify({}), 204\n\n\ndef send_response(send_call, task_call, notification, queue):\n try:\n send_call(notification)\n except Exception as e:\n current_app.logger.exception(\n \"Failed to send notification, retrying in celery. ID {} type {}\".format(\n notification.id,\n notification.notification_type),\n e)\n task_call.apply_async((str(notification.id)), queue=queue)\n","repo_name":"govau/notify","sub_path":"api/app/delivery/rest.py","file_name":"rest.py","file_ext":"py","file_size_in_byte":1579,"program_lang":"python","lang":"en","doc_type":"code","stars":17,"dataset":"github-code","pt":"7"} +{"seq_id":"17381584777","text":"def sol(arr):\r\n n=len(arr)\r\n re1=['0']*n\r\n re2=['0']*n\r\n\r\n for i in range (n):\r\n if i%2==0:\r\n re1[i]='1'\r\n re2[i]='0'\r\n else:\r\n re1[i]='0'\r\n re2[i] = '1'\r\n t1=0\r\n t2=0\r\n for i in range(n):\r\n if arr[i]!=re1[i]:\r\n t1+=1\r\n if arr[i]!=re2[i]:\r\n t2+=1\r\n print(t1,t2)\r\n return min(t1,t2)\r\n\r\n\r\nif __name__ == '__main__':\r\n st='0001010111'\r\n str=list(st)\r\n print(sol(str))\r\n","repo_name":"Samundar9525/data_structure_using_Python","sub_path":"string/binary alternating.py","file_name":"binary alternating.py","file_ext":"py","file_size_in_byte":496,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"27846238863","text":"\"\"\"\nOperations to be shared across multiple tests\n\"\"\"\nimport pytest\nimport torch\nimport torch.nn as nn\nfrom torch.utils.data import DataLoader\n\nfrom two_step_zoo.datasets import Sphere, SupervisedDataset\nfrom two_step_zoo.networks import MLP\n\n\ndef pytest_configure():\n pytest.latent_dim = 2\n pytest.hidden_dims = [5, 5]\n pytest.batch_size = 5\n pytest.data_dim = 4\n pytest.data_len = 64\n pytest.image_data_shape = [3, 32, 32] # For Multiscale Flow\n pytest.noise_dim = 1 # For AVB\n pytest.activation = nn.ReLU\n pytest.optim_cfg = {\n \"optimizer\": \"sgd\",\n \"lr\": 1e-3,\n \"disc_lr\": 1e-3, # For AVB\n \"nll_lr\": 1e-3, # For AVB\n \"ge_lr\": 1e-3, # For BiGAN\n \"rec_lr\": 1e-3, # For WAE\n }\n\n\n@pytest.fixture\ndef dataloader():\n manifold_dim = 2\n\n dataset = Sphere(\"sphere\", \"train\", manifold_dim, pytest.data_dim, pytest.data_len)\n dataloader = DataLoader(dataset, pytest.batch_size)\n return dataloader\n\n\n@pytest.fixture\ndef imagelike_dataloader():\n dataset = SupervisedDataset(\n name=\"Test Image Dataset\",\n role=\"train\",\n x=torch.rand((pytest.data_len, *pytest.image_data_shape))\n )\n dataloader = DataLoader(dataset, pytest.batch_size)\n return dataloader\n\n\n@pytest.fixture\ndef batch(dataloader):\n for batch in dataloader:\n return batch\n\n\n@pytest.fixture\ndef basic_encoder():\n return MLP(\n input_dim=pytest.data_dim,\n hidden_dims=pytest.hidden_dims,\n output_dim=pytest.latent_dim,\n activation=pytest.activation\n )\n\n\n@pytest.fixture\ndef basic_decoder():\n return MLP(\n input_dim=pytest.latent_dim,\n hidden_dims=pytest.hidden_dims,\n output_dim=pytest.data_dim,\n activation=pytest.activation\n )\n\n\n@pytest.fixture\ndef multihead_encoder():\n return MLP(\n input_dim=pytest.data_dim,\n hidden_dims=pytest.hidden_dims,\n output_dim=2*pytest.latent_dim,\n activation=pytest.activation,\n output_split_sizes=[pytest.latent_dim, pytest.latent_dim]\n )\n\n\n@pytest.fixture\ndef multihead_decoder():\n return MLP(\n input_dim=pytest.latent_dim,\n hidden_dims=pytest.hidden_dims,\n output_dim=2*pytest.data_dim,\n activation=pytest.activation,\n output_split_sizes=[pytest.data_dim, pytest.data_dim]\n )\n\n\n@pytest.fixture\ndef basic_discriminator():\n return MLP(\n input_dim=pytest.data_dim+pytest.latent_dim,\n hidden_dims=pytest.hidden_dims,\n output_dim=1,\n activation=pytest.activation\n )\n","repo_name":"layer6ai-labs/two_step_zoo","sub_path":"tests/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":2555,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"7"} +{"seq_id":"74733427742","text":"# code_#1 image segmentation\n\n'''从文件夹中读取所有图片,并从每张图片的一系列坐标开始,以100*100/33*33的大小的分割框对图像进行分割;每个点之间相距30个像素点\n,每次向右移动60/30像素,将图片分割成小图片。最终以原图片名+行名(A-H)+列名(1-12)的形式保存在一个以原图片名为名的文件夹中'''\n\nimport cv2\nimport os\nfile_pathname=\"\" #文件路径\n\ndef read_path(file_pathname):\n for root, dirs, files in os.walk(file_pathname):\n #os.walk \n for file in files: \n ##print(file_pathname+'/'+file) #test code\n if file.endswith('.png'): #selcet picture\n #Load pictures\n img = cv2.imread(file_pathname+'/'+file)\n file_name, file_extend = os.path.splitext(file) \n # Define the coordinates坐标 and size of the crop box\n\n coords = [[275,225], [275,255], [275, 285], [275,315], [275, 345], [275,375], [275, 405], \n [275,435], [275, 465], [275,495], [275, 525], [275,555], [275, 585], [275,615],[275, 645],[275,675]]\n width, height = 33, 33\n move_x = 30\n #当分割框包含1个衣藻时选择该组参数\n\n\n '''\n coords = [[275,225], [275, 285],[275, 345], [275, 405], [275, 465], [275, 525], [275, 585], [275, 645]]\n width, height = 100, 100\n move_x = 60\n #当分割框包含9个衣藻时选择该组参数\n '''\n\n\n # Loop through all the coordinates and crop the image\n for i, coord in enumerate(coords):\n x, y = coord\n for j in range(1,25): \n crop_img = img[y:y+height, x:x+width]\n cv2.imwrite(f\"D:\\\\desktop\\\\Fredrik\\\\Supplemental_Data_Set_1\\\\Image_segmentation\\\\B_TAP_photos_segmentation_in_one\\\\{file_name}_{chr(i+65)}_{j}.jpg\", crop_img)\n x += move_x\n\nread_path(file_pathname)\n\n# code_#2 color_standard_generate\nfrom PIL import Image\nimport openpyxl\n\n# 定义起始颜色和结束颜色\nstart_color = (0, 30, 0 ) # 墨绿色\nend_color = (240, 240, 50) # 黄绿色\n\n# 定义Excel文件名和工作表名\nfilename = \"\"\nsheetname = \"Color standards\"\n\n# 定义Excel中起始单元格的位置\nstart_row = 1\nstart_col = 1\n\n# 定义颜色图的尺寸和颜色数量\nwidth = 1\nheight = 300\nnum_colors = height\n\n# 生成颜色图\ngradient = Image.new('RGB', (width, height))\nfor i in range(num_colors):\n r = start_color[0] + int((i / num_colors) * (end_color[0] - start_color[0]))\n g = start_color[1] + int((i / num_colors) * (end_color[1] - start_color[1]))\n b = start_color[2] + int((i / num_colors) * (end_color[2] - start_color[2]))\n gradient.putpixel((0, i), (r, g, b))\n\n# 将颜色输出到Excel中\nwb = openpyxl.Workbook()\nws = wb.active\nfor i in range(num_colors):\n color = gradient.getpixel((0, i))\n ws.cell(row=start_row+i, column=start_col).value = f\"RGB({color[0]}, {color[1]}, {color[2]})\"\n ws.cell(row=start_row+i, column=start_col+1).fill = openpyxl.styles.PatternFill(start_color='FF'+\"\".join([hex(c)[2:].rjust(2, '0').upper() for c in color]), fill_type='solid')\n \n# 保存Excel文件\nwb.save(filename)\n","repo_name":"Fredrik-Qi/Phenotypic-Quantification-and-Clustering-Analysis-of-Chlamydomonas-reinhardtii-Mutants_code","sub_path":"Supplemental_Data_Set_1.py","file_name":"Supplemental_Data_Set_1.py","file_ext":"py","file_size_in_byte":3341,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25541331245","text":"test_case = int(input())\r\nres = []\r\nnum = []\r\nfnum = ''\r\nwithdraw_amount = []\r\nfor i in range(test_case):\r\n n,bal = input().split(' ')\r\n n = int(n)\r\n bal = int(bal)\r\n withdraw_amount = list(map(int,input().split()))\r\n for k in withdraw_amount:\r\n avlbl = bal - k\r\n if avlbl < 0:\r\n num.append('0')\r\n else:\r\n num.append('1')\r\n bal -= k\r\n for l in num:\r\n fnum += l\r\n res.append(fnum)\r\n fnum = ''\r\n num.clear()\r\nfor i in res:\r\n print(i)\r\n\r\n\r\n\r\n","repo_name":"SanjayKotabagi/CODECHEF_SOLUTIONS","sub_path":"BEGINNER_LEVEL/ATM2.py","file_name":"ATM2.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"34217888584","text":"# -*- coding: utf-8 -*-\r\nimport pyxel\r\nfrom module.character import HumanGenerator\r\nfrom module.fieldStates.baseFieldState import BaseFieldState\r\nfrom module.map.cemetery import cemetery\r\nfrom module.params.monster import monsterParams\r\nfrom overrides import overrides\r\n\r\n\r\nclass StateCemetery(BaseFieldState):\r\n '''\r\n 墓地の地下のクラス\\n\r\n BaseFieldStateを継承。\r\n '''\r\n # State名\r\n stateName = \"CEMETERY\"\r\n\r\n # マップ\r\n _map = cemetery.map\r\n\r\n # 出現するモンスターリスト\r\n enemy_set = (\r\n HumanGenerator.generate(1),\r\n HumanGenerator.generate(2),\r\n monsterParams[\"WOLF_LV1\"],\r\n monsterParams[\"BAT_LV1\"],\r\n monsterParams[\"ZOMBIE_LV1\"],\r\n monsterParams[\"SKELETON_LV1\"],\r\n )\r\n\r\n def __init__(self, **kwargs):\r\n '''\r\n クラス初期化\r\n '''\r\n super().__init__(**kwargs)\r\n\r\n @overrides\r\n def onEnter(self):\r\n '''\r\n 状態開始時の処理\r\n '''\r\n super().onEnter()\r\n\r\n # 壁の色を初期化する\r\n self.set_wall_color(pyxel.COLOR_RED, pyxel.COLOR_PURPLE)\r\n\r\n @overrides\r\n def onExit(self):\r\n '''\r\n 状態終了時の処理\r\n '''\r\n super().onExit()\r\n","repo_name":"aburi6800/Python-OnyxOfBlack","sub_path":"source/module/fieldStates/stateCemetery.py","file_name":"stateCemetery.py","file_ext":"py","file_size_in_byte":1270,"program_lang":"python","lang":"ja","doc_type":"code","stars":4,"dataset":"github-code","pt":"7"} +{"seq_id":"6084254680","text":"#!/usr/bin/env python\n\nimport sys\nimport os\nimport time\nimport RPi.GPIO as io\nimport signal\nimport atexit\n\n# GPIO konfigurieren\nio.setmode(io.BCM)\n\n#Variablen deklarieren\nPIR_PIN = 4\nLED_PIN = 27\n\n# Funktionen\ndef signal_term_handler(signal, frame):\n print('got SIGTERM')\n io.cleanup()\n sys.exit(0)\n\ndef goodbye():\n print('Goodbye')\n io.cleanup()\n\ndef main():\n global turned_off\n global motion_count\n\n print('start - Warten auf erstes Event', flush=True)\n\n while True:\n if io.input(PIR_PIN):\n io.output(LED_PIN, io.HIGH)\n print(time.strftime('Bewegung erkannt %H:%M:%S '), flush=True)\n time.sleep(1)\n io.output(LED_PIN, io.LOW)\n time.sleep(.5)\n\n# --------------------------------------------------------------------\n# los gehts \n#---------------------------------------------------------------------\n\n# Abfangen von SIGTERM\nsignal.signal(signal.SIGTERM, signal_term_handler)\n\n# Abfangen von KeyboardInterrupt CRTL-C\natexit.register(goodbye)\n\n# IO setzen für LED und PIR\nio.setup(PIR_PIN, io.IN, pull_up_down=io.PUD_DOWN)\nio.setup(LED_PIN, io.OUT)\n\n# und los gehts\nmain()\n","repo_name":"spitzlbergerj/CaravanPi","sub_path":"pir/pir-testgeraet.py","file_name":"pir-testgeraet.py","file_ext":"py","file_size_in_byte":1157,"program_lang":"python","lang":"en","doc_type":"code","stars":23,"dataset":"github-code","pt":"7"} +{"seq_id":"22712766748","text":"import numpy as np\nfrom tensorflow.python.keras.models import Sequential\nfrom tensorflow.python.keras.layers.core import Dense\nfrom sklearn import datasets\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.model_selection import train_test_split\nfrom tensorflow.python.keras.optimizers import Adam\nfrom sklearn.model_selection import cross_validation\n\ndataset = datasets.load_iris()\n\nfeatures = dataset.data\ny= dataset.target.reshape(-1,1)\n\nencoder=OneHotEncoder()\ntarget=encoder.fit_transform(y)\n\ntrain_feature,test_feature,train_target,test_target=train_test_split(features,target,test_size=0.3)\n\nmodel= Sequential()\n\nmodel.add(Dense(10,input_dim=4, activation='relu'))\nmodel.add(Dense(10,input_dim=10, activation='relu'))\nmodel.add(Dense(10,input_dim=10, activation='relu'))\nmodel.add(Dense(10,input_dim=10, activation='relu'))\nmodel.add(Dense(10,input_dim=10, activation='relu'))\nmodel.add(Dense(10,input_dim=10, activation='relu'))\nmodel.add(Dense(3, activation='sigmoid'))\n\noptimizers=Adam(lr=0.05)\n\nmodel.compile(loss='categorical_crossentropy',\n optimizer=optimizers,\n metrics=['accuracy'])\n\nmodel.fit(train_feature,train_target,batch_size=20,epochs=1000, verbose=2)\n\nresult=model.evaluate(test_feature,test_target)\n\nprint(\"Resultant Error %.2f\",result[0])\nprint(\"Resultant accuracy %.2f\",result[1])\n","repo_name":"varunsly/Neural_Network_Iris-dataset_Prediction","sub_path":"Iris_keras.py","file_name":"Iris_keras.py","file_ext":"py","file_size_in_byte":1345,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"31584609927","text":"from starlette.responses import JSONResponse\nfrom uuid import uuid4\nfrom src.extensions import create_virtual_account\nfrom src.extensions import create_contact\n\nfrom src.utils.http import session\n\nfrom src.models.request.razorpay import VirtualAccoutRequest\nfrom src.models.request.razorpayx import CreateContactRequest\n\nfrom src.database.models.users import User\nfrom src.database.models.account import Account\nfrom src.database.models.upi import UPI\n\nfrom src.database.database import async_db_session\n\n\n\nasync def create_account(request):\n json = await request.json()\n user_id = json.get(\"user_id\")\n\n if not user_id:\n return JSONResponse({\"message\": \"user_id is required\"}, status_code=400)\n\n else:\n try:\n user: User = await User.get(user_id)\n print(user)\n contact_request = CreateContactRequest(\n name=user.full_name,\n contact=json.get(\"contact\"),\n email=json.get(\"email\"),\n type=json.get(\"type\"),\n reference_id=json.get(\"reference_id\"),\n notes=[],\n )\n contact = await create_contact(session=session, data=contact_request)\n except Exception as e:\n print(str(e))\n return JSONResponse(\n {\"message\": str(e), \"error\": \"Could not create Contact\"},\n status_code=400,\n )\n\n # receivers = {\n # \"types\": [\"vpa\"],\n # \"vpa\": {\n # \"descriptor\": user.full_name.replace(\" \", \"-\"),\n # },\n # }\n try:\n # account_request = VirtualAccoutRequest(\n # receivers=receivers,\n # close_by=None,\n # notes=json.get(\"notes\"),\n # description=json.get(\"description\"),\n # customer_id=contact.id,\n # )\n # virtual_account = await create_virtual_account(\n # session=session, data=account_request\n # )\n # MOCK ACCOUNT CREATION\n upi_id = ''\n \n upi_id = await UPI.get_by_user_id(user_id)\n if upi_id:\n account = await Account.get_by_user_id(user_id)\n return JSONResponse({\n \"user_id\": user_id,\n \"account_id\": account.id,\n \"balance\": float(account.balance),\n \"upi_id\": upi_id.id,\n })\n else:\n account_id = str(uuid4())\n upi_id = f\"{user.full_name.lower().replace(' ', '-')}@okicici\"\n \n await Account.create(id=account_id, balance=0, contact_id=contact.id, user_id=user_id)\n await UPI.create(id=upi_id, user_id=user_id)\n \n except Exception as e:\n print('2')\n await async_db_session.rollback()\n return JSONResponse(\n {\"message\": str(e), \"error\": \"Could not create Account\"},\n status_code=400,\n )\n\n return JSONResponse({\n \"user_id\": user_id,\n \"account_id\": account_id,\n \"balance\": 0,\n \"upi_id\": upi_id,\n })\n","repo_name":"SudodevsHQ/betsushi","sub_path":"src/routes/create_virtual_account.py","file_name":"create_virtual_account.py","file_ext":"py","file_size_in_byte":3214,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"70376907422","text":"# USSD program to withdraw,check balance,send money,buy airtime,\nuser={\n \"Name\":\"Mike\",\n \"Agent_no\":'4463',\n \"Pin\":'2342',\n \"Balance\":1000\n}\ndef menu():\n print(f\"Hello {user['Name']}, \\n Welcome to Safaricom\")\n print(\"What do you want to do?\")\n print(\"1. Withdraw Money\")\n print(\"2. Check Balance\")\n print(\"3. Send Money\")\n print(\"4. Buy Airtime\")\n \ndef Withdraw(withdraw_amount):\n transaction_fee=30\n menu()\n\n option=input(\"Select your Option: \")\n if option ==\"1\":\n # check if the balance is more than the withdraw_amount\n # if user['Balance'] <= withdraw_amount:\n # print(f\"Sorry {user['Name']} \\n Insufficient balance!!! \")\n if user['Balance'] < (withdraw_amount + transaction_fee):\n print(\"Failed \\n You must have the transaction fee to complete this request\")\n else:\n # prompt user to input agent number\n agent_no=input(\"Enter Agent number: \")\n if agent_no != user['Agent_no']:\n print(\"Wrong/ Invalid agent number\")\n else:\n pin=input(\"Enter your pin: \")\n if pin != user['Pin']:\n print(\"Wrong pin. \\n Try again later\")\n # withdraw_amount(amount)\n else:\n new_balance=user['Balance'] -(withdraw_amount + transaction_fee)\n print(f\"Withdrawal Successful You have withdrawn {withdraw_amount} from agent number {user['Agent_no']} \\n Your new balance is {new_balance} \\n Thank you being Our esteemed Customer\") \n else:\n print(\"Sorry!!1 Service On Progress\") \n \n# Withdraw(900)\n \n \n \n ","repo_name":"maxmusau/Python","sub_path":"Lesson7d.py","file_name":"Lesson7d.py","file_ext":"py","file_size_in_byte":1704,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"36282834745","text":"from paddleocr import PaddleOCR\nimport sys\n\nsys.path.append(\"/mnt/Software/200-Apps/imageFlow\")\nfrom config import fm_config\n\nfname = \"/mnt/Software/200-Apps/test/IN/download.jpg\"\n\n#\n# ocr = PaddleOCR(\n# det_model_dir=fm_config.DET_MODEL_DIR,\n# rec_model_dir=fm_config.REC_MODEL_DIR,\n# # rec_char_dict_path=\"\",\n# cls_model_dir=fm_config.CLS_MODEL_DIR,\n# use_angle_cls=fm_config.USE_ANCLE_CLS,\n# lang=fm_config.IMG_OCR_LANG,\n# )\n\nocr = PaddleOCR(\n det_model_dir=\"/mnt/Software/200-Apps/filemaster_old/models/whl/det/en/en_PP-OCRv3_det_infer/\",\n rec_model_dir=\"/mnt/Software/200-Apps/filemaster_old/models/whl/rec/en/en_PP-OCRv3_rec_infer/\",\n # rec_char_dict_path=\"\",\n cls_model_dir=\"/mnt/Software/200-Apps/filemaster_old/models/whl/cls/ch_ppocr_mobile_v2.0_cls_infer/\",\n use_angle_cls=True,\n lang=\"en\",\n)\ntext = \"\"\ntry:\n result = ocr.ocr(fname, det=True, cls=True)\n for idx in range(len(result)):\n res = result[idx]\n for line in res:\n text = text + \" \" + line[1][0]\n if len(text):\n print(text)\n\n else:\n print(\"NO text found\")\nexcept Exception as e:\n print(e)\n","repo_name":"Cap-n-Proud/imageflow","sub_path":"units/ocr.py","file_name":"ocr.py","file_ext":"py","file_size_in_byte":1154,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"40508004298","text":"\"\"\"Module defining the domain model entities.\"\"\"\nfrom __future__ import annotations\n\nimport logging\nimport os.path\nimport re\nfrom collections.abc import Generator\nfrom dataclasses import asdict, dataclass\nfrom datetime import datetime, timedelta\nfrom enum import Enum\nfrom pathlib import Path\nfrom typing import Any\n\nimport cv2 as cv\nimport numpy as np\n\nlogger = logging.getLogger(__name__)\n\nVIDEO_DEFAULT_HEIGHT: int = 360\nVIDEO_DEFAULT_WIDTH: int = 640\n\n\nclass TimestampNotFoundError(Exception):\n \"\"\"Exception to raise if no timestamp are found.\"\"\"\n\n\nclass Status(str, Enum):\n \"\"\"Enum for job progress.\"\"\"\n\n PENDING = \"Pending\"\n RUNNING = \"Running\"\n PAUSED = \"Paused\"\n DONE = \"Done\"\n QUEUED = \"Queued\"\n ERROR = \"Error\"\n\n\nclass Video:\n \"\"\"Video class.\n\n This class provides various functions to retrieve data from a video file.\n\n Parameters\n ----------\n path : str\n Path to this video file as a string\n frame_count : int\n Number of frames in the video\n fps : int\n Frames per second in the video\n width : int\n Width in pixels\n height : int\n Height in pixels\n timestamp : datetime\n timestamp for when the video starts\n output_width : int\n Frame output width. Default to `VIDEO_DEFAULT_WIDTH`\n constant.\n output_height : int\n Frame output height. Default to\n `VIDEO_DEFAULT_HEIGHT` constant.\n\n Attribute\n ---------\n _path : str\n Path to the video file associated with the video.\n id : Optional[int]\n Video id from repository(database).\n frames : List[Frame]\n List of data frames containing detections associated with video.\n\n Methods\n -------\n vidcap_release()\n Release OpenCV videocapture on associated video file.\n exists()\n Checks if the path is valid, by checking if its a file on the disk.\n from_path(path: str)\n Named constructor that creates and populates a video object with\n metadata read from the file. Raises FileNotFoundError if the\n file could not be read, or is not a video file.\n timestamp_at(idx: int)\n Return timestamp at index in video as a `datetime` object.\n add_detection_frame(frame Frame)\n Add a single data-frame containing detections to this video.\n is_processed()\n Checks if the video has been fully processed by comparing with\n self.frames.\n\n Examples\n --------\n >>> video = Video.from_path(\"test.mp4\")\n >>> one_frame = video[5]\n >>> print(one_frame.shape)\n (720, 1280, 3)\n >>> many_frames = video[5,10]\n >>> print(many_frames.shape)\n (5, 720, 1280, 3)\n >>> len(video)\n 20\n >>> many_frames = video[10:]\n >>> print(many_frames.shape)\n (10, 720, 1280, 3)\n\n Raises\n ------\n FileNotFoundError\n If error reading video file when creating `from_path()`.\n TimestampNotFound\n If no timestamp where found when creating `from_path()`.\n \"\"\"\n\n def __init__(\n self,\n path: str,\n frame_count: int,\n fps: int,\n width: int,\n height: int,\n timestamp: datetime,\n output_width: int = VIDEO_DEFAULT_WIDTH,\n output_height: int = VIDEO_DEFAULT_HEIGHT,\n ) -> None:\n self.id: int | None = None\n self._path: str = path\n self.frame_count: int = frame_count\n self.fps: int = fps\n self.width: int = width\n self.height: int = height\n self.output_width: int = output_width\n self.output_height: int = output_height\n self.timestamp: datetime = timestamp\n self._current_frame = 0\n self._video_capture: cv.VideoCapture = cv.VideoCapture(self._path) # type: ignore\n self.frames: list[Frame] = []\n\n if output_height <= 0 or output_width <= 0:\n raise ValueError(\n \"Output width and height must be positive, not %s, %s\",\n output_width,\n output_height,\n )\n\n def _scale_convert(self, img: np.ndarray) -> np.ndarray:\n \"\"\"Convert and scale image using OpenCV.\n\n Converts image from BGR to RGB, and scales down to\n `self.output_{height,width}`\n\n Parameter\n ---------\n img : np.ndarray\n image to convert and scale\n\n Return:\n ------\n ndarray:\n Scaled and converted image\n \"\"\"\n new_img = cv.cvtColor(img, cv.COLOR_BGR2RGB) # type: ignore\n\n new_img = cv.resize( # type: ignore\n new_img,\n (self.output_width, self.output_height),\n interpolation=cv.INTER_AREA, # type: ignore\n )\n return new_img\n\n def vidcap_release(self) -> None:\n \"\"\"Release Video Capture.\"\"\"\n self._video_capture.release()\n\n def __iter__(self) -> Video:\n \"\"\"Class iterator.\n\n This never releases the VideoCapture. Not sure if it's kept alive, and\n if that's the case, this could cause a memory leak. To make sure this\n gets released, run `self.vidcap_release()`.\n\n See Also\n --------\n Video.vidcap_release()\n\n \"\"\"\n self._video_capture = cv.VideoCapture(self._path) # type: ignore\n self._video_capture.set(cv.CAP_PROP_POS_MSEC, 0) # type: ignore\n return self\n\n def __next__(self) -> np.ndarray:\n \"\"\"Get next item from iterator.\n\n Return:\n ------\n np.ndarray\n One frame of video as `ndarray`.\n\n \"\"\"\n err, img = self._video_capture.read()\n if not err:\n self.vidcap_release()\n raise StopIteration\n return self._scale_convert(img)\n\n def __get__(self, key: int, owner: object | None = None) -> np.ndarray:\n \"\"\"Get one frame of video.\n\n Used by `__getitem__` when only one key is given.\n\n Returns\n -------\n numpy.ndarray\n One frame of video as `ndarray`.\n\n Raise\n -----\n RuntimeError :\n if OpenCV fails to either read or set properties.\n \"\"\"\n if key < 0:\n raise IndexError\n\n if key >= self.frame_count:\n raise IndexError\n\n self._video_capture = cv.VideoCapture(self._path) # type: ignore\n retval = self._video_capture.set(cv.CAP_PROP_POS_FRAMES, key) # type: ignore\n\n if not retval:\n raise RuntimeError( # pragma: no cover\n f\"Unexpected error when setting catpure property, {retval}\",\n )\n\n retval, img = self._video_capture.read()\n\n if not retval:\n raise RuntimeError(\n f\"Unexpected error when reading frame at {key}\",\n ) # pragma: no cover\n\n self._video_capture.release()\n\n return self._scale_convert(img)\n\n def __getitem__(self, interval: slice | int) -> np.ndarray:\n \"\"\"Get a slice of video.\n\n Get a interval of frames from video, `variable[start:stop:step].\n Note `step` is not implemented and will raise a `exception`.\n\n Examples\n --------\n >>> video = Video.from_path(\"test.mp4\")\n >>> one_frame = video[5]\n >>> print(one_frame.shape)\n (720, 1280, 3)\n >>> many_frames = video[5,10]\n >>> print(many_frames.shape)\n (5, 720, 1280, 3)\n\n Returns\n -------\n numpy.ndarray\n Interval of frames returned in format\n `(frame, height, width, channels)`\n\n See Also\n --------\n __get__ : Used when only start in slice given.\n\n Raise\n -----\n RuntimeError :\n if OpenCV fails to either read or set properties.\n \"\"\"\n # If only one key is given\n if isinstance(interval, int):\n return self.__get__(interval)\n\n if not isinstance(interval, slice):\n raise TypeError(\"%s is not %s\", type(interval), type(slice))\n\n if isinstance(interval.stop, int) and interval.stop >= self.frame_count:\n raise IndexError(\n \"Index for stop in slice more then frame count %s\",\n self.frame_count,\n )\n\n if interval.start < 0:\n raise IndexError(\"Index for start in slice is less then 0\")\n\n # Slice stepping is not implemented.\n if interval.step is not None:\n raise NotImplementedError(\"Step in slicing is not implemented\")\n\n # If slicing with `video[0:] or video[0:-1]` all frames from start to\n # end or end-1 of video is wanted.\n if interval.stop is None:\n stop = self.frame_count\n elif interval.stop < 0:\n stop = self.frame_count + interval.stop\n print(stop)\n else:\n stop = interval.stop\n\n numbers = stop - interval.start\n\n self._video_capture = cv.VideoCapture(self._path) # type: ignore\n retval = self._video_capture.set(\n cv.CAP_PROP_POS_FRAMES,\n interval.start, # type: ignore\n )\n\n if not retval:\n raise RuntimeError(\"Unexpected error\") # pragma: no cover\n\n frames = []\n\n for _ in range(numbers):\n retval, img = self._video_capture.read()\n\n if not retval:\n raise RuntimeError(\"Unexpected error\") # pragma: no cover\n frames.append(self._scale_convert(img))\n\n self._video_capture.release()\n return np.array(frames)\n\n def iter_from(self, start: int) -> Generator[np.ndarray, None, None]:\n \"\"\"Iterate from start to the end of the video.\n\n Parameter\n ---------\n start : int\n The frame to start at\n\n Yields\n ------\n np.ndarray :\n A single scaled frame.\n\n Raises\n ------\n RuntimeError :\n Unexpected errors occur with OpenCV\n \"\"\"\n if start >= self.frame_count or start < 0:\n raise IndexError(\n f\"Start is out of bounds for buffer of size {self.frame_count}, got {start}\",\n )\n self._video_capture = cv.VideoCapture(self._path) # type: ignore\n retval = self._video_capture.set(cv.CAP_PROP_POS_FRAMES, start) # type: ignore\n\n if not retval:\n raise RuntimeError(\"Unexpected error\") # pragma: no cover\n\n numbers = self.frame_count - start\n\n for _ in range(numbers):\n retval, img = self._video_capture.read()\n\n if not retval:\n raise RuntimeError(\"Unexpected error\") # pragma: no cover\n yield self._scale_convert(img)\n\n self._video_capture.release()\n\n def __len__(self) -> int:\n \"\"\"Get length of video in frames.\"\"\"\n return self.frame_count\n\n def exists(self) -> bool:\n \"\"\"Check if the file path is a valid file.\"\"\"\n return os.path.isfile(self._path)\n\n @classmethod\n def from_path(\n cls,\n path: str,\n output_width: int = VIDEO_DEFAULT_WIDTH,\n output_height: int = VIDEO_DEFAULT_HEIGHT,\n ) -> Video:\n \"\"\"Named constructor to create a `Video` from path.\n\n Examples\n --------\n >>> video = Video.from_path(\"video.mp4\")\n >>> type(video)\n \n\n Raises\n ------\n FileNotFoundError\n If `path` to file do not exists.\n \"\"\"\n if not Path(path).exists():\n raise FileNotFoundError(\"Video file %s not found.\", path)\n\n timestamp = parse_str_to_date(Path(path).name)\n if timestamp is None:\n raise TimestampNotFoundError(f\"No timestamp found for file {path}\")\n\n height, width, fps, frame_numbers = _get_video_metadata(path)\n\n return cls(\n path=path,\n frame_count=frame_numbers,\n fps=fps,\n width=width,\n height=height,\n timestamp=timestamp,\n output_width=output_width,\n output_height=output_height,\n )\n\n def timestamp_at(self, idx: int) -> datetime:\n \"\"\"Return timestamp at index in video.\n\n Parameter\n ---------\n idx : int\n Index in video.\n\n Return:\n ------\n datetime :\n Timestamp for the frame at index.\n\n \"\"\"\n if idx > self.frame_count:\n raise IndexError\n if idx < 0:\n raise IndexError\n\n return self.timestamp + (timedelta(seconds=int(idx / self.fps)))\n\n def add_detection_frame(self, frame: Frame) -> None:\n \"\"\"Update detected data associated with this video.\n\n Parameters\n ----------\n frames : List[Frame]\n List of data-frames to add to this video. Which contains detections for a given frame.\n force_update : bool\n Will overwrite already stored data-frames of same index with new one in frames parameter.\n Default False.\n\n Raises\n ------\n RuntimeError\n When any frame index goes past total frames in video.\n\n Return\n ------\n bool :\n True when data-frames were successfully updated. False when inputted frames have overlap\n with existing data within video.\n \"\"\"\n if frame in self.frames:\n raise RuntimeError(\n f\"Frame with index {frame.idx} is already added to this video.\",\n )\n if frame.idx > self.frame_count:\n raise IndexError(\n f\"Frame of index {frame.idx} is beyond total frames in video.\",\n )\n\n self.frames.append(frame)\n\n def is_processed(self) -> bool:\n \"\"\"Check if this video has been fully processed.\n\n Return:\n ------\n bool :\n True if the entire video has been processed. The entire detection_frames dict must be fully\n mapped with data-frames for all frames in video.\n \"\"\"\n if not len(self.frames) == self.frame_count:\n logger.info(\n f\"Video {self._path} is not fully processed. {len(self.frames)}/{self.frame_count}\",\n )\n return False\n\n # Check continious index\n for i in range(self.frame_count):\n if self.frames[i].idx != i:\n logger.warning(\n \"Frame index {self.detection_frames[i].idx} does not match videos index {i}\",\n )\n return False\n\n logger.info(\"Video {self._path} is processed.\")\n return True\n\n\ndef parse_str_to_date(string: str, offset_min: int = 30) -> datetime | None:\n \"\"\"Parse string to date.\n\n Input can either be a string with a date, or a string with a date and\n offset. If an offset is found, `offset_min` will be multiplied with the\n offset and the result will be to the returned date.\n\n Parameter\n ---------\n string: str\n string to parse to date on the format:\n `[yyyy-mm-dd_hh-mm-ss]` or `[yyyy-mm-dd_hh-mm-ss]-xxx`\n offset_min: int\n Minutes to offset for each increment\n\n\n Return:\n ------\n datetime :\n parsed datetime object, or None if unsuccessful\n\n Example:\n -------\n >>> parse_str_to_date(\"test-[2020-03-28_12-30-10]-000.mp4\")\n datetime.datetime(2020, 3, 28, 12, 30, 10)\n >>> parse_str_to_date(\"test-[2020-03-28_12-30-10]-001.mp4\")\n datetime.datetime(2020, 3, 28, 13, 0, 10)\n >>> parse_str_to_date(\"test-[2020-03-28_12-30-10].mp4\")\n datetime.datetime(2020, 3, 28, 12, 30, 10)\n >>> parse_str_to_date(\"test.mp4\")\n None\n \"\"\"\n match = re.compile(\n r\"\\[\\d{4}(-\\d{2}){2}_(\\d{2}-){2}\\d{2}\\](-\\d{3})?\",\n ).search(string)\n\n if not match:\n logger.warning(f\"no date found in str, {string}\")\n return None\n\n try:\n # Offset is optional in the regex, (-\\d{3})?. This tries to split on\n # \"]-\", which means there exist an offset, []-. If it\n # fails it means there are no offset.\n timestamp, offset = match[0].split(\"]-\")\n\n # timestamp still has a \"[\" at the start. This strips it.\n timestamp = timestamp[1:]\n offset_int: int = int(offset)\n except ValueError:\n # No offset found, so only grab what's inside the brackets, and set\n # offset to zero.\n timestamp = match[0][1:-1]\n offset_int = 0\n\n date = \"-\".join(timestamp.split(\"_\"))\n\n year, month, day, hour, minute, second = (int(x) for x in date.split(\"-\"))\n\n try:\n return datetime(year, month, day, hour, minute, second) + timedelta(\n minutes=offset_min * offset_int,\n )\n except ValueError:\n return None\n\n\ndef _get_video_metadata(path: str) -> tuple[int, ...]:\n \"\"\"Get metadata from video using `opencv`.\n\n Parameter\n ---------\n path : str\n path to file to get metadata from.\n\n Return:\n ------\n Typle[int, int, int, int] :\n A tuple with the metadata:\n (height, width, FPS, frame_count)\n\n Raises\n ------\n FileNotFoundError:\n If the file can't be opened it will throw FileNotFoundError\n RuntimeError:\n If there are problems getting any metadata.\n \"\"\"\n video = cv.VideoCapture(path) # type: ignore\n\n if not video.isOpened():\n raise FileNotFoundError(f\"Could not open {path}\")\n\n metadata = (\n int(video.get(cv.CAP_PROP_FRAME_HEIGHT)), # type: ignore\n int(video.get(cv.CAP_PROP_FRAME_WIDTH)), # type: ignore\n int(video.get(cv.CAP_PROP_FPS)), # type: ignore\n int(video.get(cv.CAP_PROP_FRAME_COUNT)), # type: ignore\n )\n\n video.release()\n\n # Frame count becomes \"-9223372036854775808\" when testing with png. Opencv\n # should return 0 if it fails, but apparently not in this case...\n for meta in metadata:\n if meta < 1:\n raise RuntimeError(f\"Could not get metadata for file {path}\")\n\n return metadata\n\n\n@dataclass\nclass Frame:\n \"\"\"Simple dataclass representing frame.\"\"\"\n\n idx: int\n detections: list[Detection]\n timestamp: datetime | None = None\n video_id: int | None = None\n\n def __eq__(self, other: object) -> bool:\n \"\"\"Check if two Frames are the same.\"\"\"\n return (\n isinstance(other, Frame)\n and self.idx == other.idx\n and self.video_id == other.video_id\n )\n\n def to_json(self) -> dict[str, Any]:\n \"\"\"Convert frame to json.\n\n Return:\n ------\n Dict[str, Any] :\n Object as json:\n {\n \"idx\": int,\n \"detections\": List[Detection],\n \"timestamp\": None|str\n }\n\n \"\"\"\n timestamp_tmp: str | None = None\n if self.timestamp:\n timestamp_tmp = self.timestamp.isoformat()\n else:\n timestamp_tmp = None\n\n return {\n \"idx\": self.idx,\n \"detections\": [det.to_json() for det in self.detections if det],\n \"timestamp\": timestamp_tmp,\n \"video_id\": self.video_id,\n }\n\n\n@dataclass\nclass BBox:\n \"\"\"Class representing a Bounding box.\"\"\"\n\n x1: float\n y1: float\n x2: float\n y2: float\n\n\n@dataclass\nclass Detection:\n \"\"\"Class representing a Detection.\n\n Parameter\n ---------\n bbox: BBox\n A bounding box\n probaility: float\n Probability from detection\n label: int\n Class label from detection\n frame: int\n Which absolute frame it belongs to\n frame_id: int\n The relative frame in a video\n video_id: int\n ID of the video this detection is found.\n\n Example:\n -------\n >>> bbox = BBox(10,20,30,40)\n >>> detection = (bbox, 0.8, 1, 4)\n >>> print(detection)\n (BBox(x1=10, y1=20, x2=30, y2=40), 0.8, 1, 4)\n \"\"\"\n\n bbox: BBox\n probability: float\n label: int\n frame: int\n frame_id: int | None = None\n video_id: int | None = None\n\n def to_json(self) -> dict[str, Any]:\n \"\"\"Convert detection to json.\n\n Return:\n ------\n Dict[str, Any] :\n Detection as json,\n {\n \"bbox\": BBox,\n \"probability\": float,\n \"label\": int,\n \"frame\": int,\n }\n\n \"\"\"\n return {\n \"bbox\": asdict(self.bbox),\n \"probability\": self.probability,\n \"label\": self.label,\n \"frame\": self.frame,\n \"frame_id\": self.frame_id,\n \"video_id\": self.video_id,\n }\n\n def set_frame(\n self,\n frame: int,\n frame_id: int,\n video_id: int | None,\n ) -> Detection:\n \"\"\"Update frame nr.\n\n Returns itself so it can be used in list comprehensions.\n\n Parameter\n ---------\n frame : int\n The frame number the detection is found in.\n\n Return:\n ------\n Detection :\n Returns self.\n \"\"\"\n if not isinstance(frame_id, int):\n raise RuntimeError(f\"Frame: expected int, got {type(frame_id)}\")\n\n if not isinstance(video_id, int):\n raise RuntimeError(f\"Video: expected int, got {type(video_id)}\")\n\n self.frame = frame\n self.frame_id = frame_id\n self.video_id = video_id\n\n return self\n\n @classmethod\n def from_api(\n cls,\n bbox: dict[str, Any],\n probability: float,\n label: int,\n frame: int,\n frame_id: int,\n video_id: int,\n ) -> Detection:\n \"\"\"Create Detection class from tracker.\n\n Parameter\n ---------\n bbox: Dict[Any, str]\n Dict representation of BBox\n probaility: float\n Probability from detection\n label: int\n Class label from detection\n frame: int\n Which frame it belongs to\n\n Return:\n ------\n Detection :\n A detection object\n \"\"\"\n return cls(BBox(**bbox), probability, label, frame, frame_id, video_id)\n\n\nclass Object:\n \"\"\"Class representation of an object that has been detected and tracked.\"\"\"\n\n def __init__(\n self,\n label: int,\n detections: list[Detection] | None = None,\n track_id: int | None = None,\n ) -> None:\n \"\"\"Create an Object.\n\n Parameters\n ----------\n label : int\n The label given to it by the tracker and detector\n detections : List[Detection]\n List of detections accociated with this object. Default=[]\n track_id : int\n Tracking ID for this object. Default=None\n \"\"\"\n self.id: int | None\n self.label: int = label\n self.probability: float = 0.0\n self._detections = detections if detections is not None else []\n self.track_id: int | None = track_id\n self.time_in: datetime | None = None\n self.time_out: datetime | None = None\n self._calc_label()\n\n def to_api(self) -> dict[str, Any]:\n \"\"\"Convert relevant member data for use in api.\n\n Returns\n -------\n Dict[str, Any]\n Returns mapping between variable name and value.\n \"\"\"\n return {\n \"id\": self.id,\n \"label\": self.label,\n \"probability\": self.probability,\n \"_detections\": self._detections,\n \"time_in\": self.time_in,\n \"time_out\": self.time_out,\n \"video_ids\": self.video_ids,\n }\n\n def _calc_label(self) -> None:\n \"\"\"Calculate label.\"\"\"\n if len(self._detections) == 0:\n return\n\n self.label = int(\n np.bincount([detect.label for detect in self._detections]).argmax(),\n )\n\n self.probability = sum(\n detect.probability\n for detect in self._detections\n if detect.label == self.label\n ) / len(self._detections)\n\n @classmethod\n def from_api(\n cls,\n track_id: int,\n detections: list[dict[str, Any]],\n label: int,\n ) -> Object:\n \"\"\"Create Object class from tracker.\n\n Parameter\n ---------\n track_id : int\n track_id from tracker\n detections : List[Dict[Any,str]]\n List of detections associated.\n label : int\n Class label\n\n Return:\n ------\n Object :\n Fully featured Object.\n \"\"\"\n dets = [Detection.from_api(**detect) for detect in detections]\n return cls(label, dets, track_id)\n\n def get_results(self) -> dict[str, Any]:\n \"\"\"Return information on this object.\n\n Return:\n ------\n Dict[str, Any] :\n\n \"\"\"\n self._calc_label()\n\n return {\n \"track_id\": self.track_id,\n \"label\": self.label,\n \"probability\": self.probability,\n \"time_in\": self.time_in,\n \"time_out\": self.time_out,\n }\n\n def __eq__(self, o: object) -> bool:\n \"\"\"Check if two Objects are same.\n\n Currently, this doesn't do much, but in the future, it should implement to\n check detections as well.\n \"\"\"\n return (\n isinstance(o, Object)\n and self.label == o.label\n and self.probability == o.probability\n and len(self._detections) == len(o._detections)\n )\n\n def add_detection(self, detection: Detection) -> None:\n \"\"\"Add a detection to the object.\n\n Parameter\n ---------\n detection : Detection\n \"\"\"\n self._detections.append(detection)\n self._calc_label()\n\n def number_of_detections(self) -> int:\n \"\"\"Return the number of detections.\n\n Return:\n ------\n int :\n Number of detections\n \"\"\"\n return len(self._detections)\n\n def get_detection(self, idx: int) -> Detection | None:\n \"\"\"Return the detection at index idx.\n\n Parameter\n ---------\n idx: int\n Index\n\n Return:\n ------\n Optional[Detection] :\n Detection at index idx or None if none found.\n \"\"\"\n try:\n return self._detections[idx]\n except IndexError:\n return None\n\n def get_frames(self) -> list[tuple[int | None, int | None, BBox]]:\n \"\"\"Return which frame and which video this object is in.\n\n frame_id tells what frame in the video with video_id contains a\n detection associated with this object.\n\n Return:\n ------\n List[Tuple[Optional[int], Optional[int], BBox]] :\n [(frame_id, video_id), (frame_id, video_id)]\n \"\"\"\n return [\n (det.frame_id, det.video_id, det.bbox) for det in self._detections\n ]\n\n @property\n def video_ids(self) -> list[int]:\n \"\"\"Derive all video the object is part of.\n\n Return:\n ------\n List[int]\n List of video id's.\n \"\"\"\n video_id: set[int] = set()\n for det in self._detections:\n if det.video_id is not None:\n video_id.add(det.video_id)\n return list(video_id)\n\n\nclass Job:\n \"\"\"Class representation of a job.\"\"\"\n\n def __init__(\n self,\n name: str,\n description: str,\n location: str,\n status: Status = Status.PENDING,\n progress: int = 0,\n ) -> None:\n self.id: int | None = None\n self.name: str = name\n self.description: str = description\n self._status: Status = status\n self._objects: list[Object] = []\n self.videos: list[Video] = []\n self.location: str = location\n self.next_batch: int = 0\n self.progress: int = progress\n\n @property\n def stats(self) -> dict[str, Any]:\n \"\"\"Return statistics for a job.\n\n Return:\n ------\n Dict[str, any] :\n {\n total_objects : int\n total_labels: int\n labels : {\n int : int\n int : int\n ...\n }\n }\n \"\"\"\n dct: dict[str, Any] = {\n \"total_labels\": 0,\n \"total_objects\": 0,\n \"labels\": {},\n }\n\n labels: dict[int, int] = {}\n\n for o in self._objects:\n if o.label not in labels:\n labels[o.label] = 0\n labels[o.label] += 1\n\n dct[\"labels\"] = labels\n dct[\"total_labels\"] = len(dct[\"labels\"])\n dct[\"total_objects\"] = len(self._objects)\n\n return dct\n\n def __hash__(self) -> int:\n \"\"\"Hash of object used in eg. `set()` to avoid duplicate.\"\"\"\n return hash(\n (type(self),)\n + (self.name, self.description, self.id, self.location),\n )\n\n def __eq__(self, other: object) -> bool:\n \"\"\"Check if job is equal to another object.\"\"\"\n if not isinstance(other, Job):\n return False\n # Note: Will not be able to check equality if jobs do not have `id`,\n # as this is the only unique parameter. A job without `id` is not\n # seen by `repository` before, so it is a new `job`.\n if self.id and other.id:\n return self.id == other.id\n return False\n\n def __repr__(self) -> str:\n \"\"\"Override of default __repr__. Gives object representation as a string.\"\"\"\n return str(self.__class__) + \": \" + str(self.__dict__)\n\n def add_object(self, obj: Object) -> None:\n \"\"\"Add an object to a job.\n\n Parameter\n ---------\n obj : Object\n An object to add\n \"\"\"\n self._objects.append(obj)\n\n def number_of_objects(self) -> int:\n \"\"\"Return number of objects.\n\n Return:\n ------\n int :\n Number of objects\n \"\"\"\n return len(self._objects)\n\n def get_object(self, idx: int) -> Object | None:\n \"\"\"Return object at index.\n\n Paramter\n --------\n idx : int\n Index in the list\n\n Return:\n ------\n Optional[Object] :\n Object at index idx. If none are found, returns None.\n \"\"\"\n try:\n return self._objects[idx]\n except IndexError:\n return None\n\n def get_result(self) -> list[dict[str, Any]]:\n \"\"\"Return result from all objects.\n\n Return:\n ------\n List[Dict[str, Any]] :\n\n \"\"\"\n return [obj.get_results() for obj in self._objects]\n\n def add_video(self, video: Video) -> bool:\n \"\"\"Add a video to this job in order to be processed.\n\n Parameter\n ---------\n video : Video\n Video to add to this job. Must have a valid timestamp.\n\n Return:\n ------\n bool :\n True if video has a set timestamp, and is not already in the\n videos list. False otherwise.\n\n \"\"\"\n if video.timestamp is None:\n logger.warning(\"Videos added to job must have set timestamp.\")\n return False\n\n if video in self.videos:\n logger.warning(\"Attempted to add an existing video to a job.\")\n return False\n\n self.videos.append(video)\n self.videos.sort(key=lambda x: x.timestamp.timestamp())\n return True\n\n def add_videos(self, videos: list[Video]) -> bool:\n \"\"\"Add a list of videos to this job in order to be processed.\n\n Parameter\n ---------\n videos : List[Video]\n List of videos to add. All must have a valid timestamp.\n\n Return:\n ------\n bool :\n True if all videos in the list has a timestamp, false otherwise.\n No videos gets added if False is returned.\n \"\"\"\n for video in videos:\n if video in self.videos:\n logger.warning(\"Video has already been added to the job.\")\n return False\n\n if video.timestamp in [v.timestamp for v in videos if v != video]:\n logger.warning(\"Duplicate timestamp.\")\n return False\n\n for video in videos:\n self.videos.append(video)\n\n self.videos.sort(key=lambda x: x.timestamp.timestamp())\n return True\n\n def remove_video(self, video: Video) -> bool:\n \"\"\"Remove an existing video from this job.\n\n Parameter\n ---------\n video : Video\n video to remove from the job.\n\n Return:\n ------\n bool :\n True if the video was removed from the job. False otherwise.\n \"\"\"\n if video in self.videos:\n self.videos.remove(video)\n return True\n else:\n return False\n\n def total_frames(self) -> int:\n \"\"\"Get the total frames in all videos for this job.\n\n Return:\n ------\n int :\n Ammount of frames in total over all video objects in this job.\n \"\"\"\n return sum(v.frame_count for v in self.videos)\n\n def status(self) -> Status:\n \"\"\"Get the job status for this job.\"\"\"\n return self._status\n\n def start(self) -> None:\n \"\"\"Mark the job as started.\"\"\"\n if self._status is Status.DONE or self._status is Status.RUNNING:\n raise JobStatusException(\n \"A running or completed job can not be started.\",\n )\n logger.debug(\"Job '%s' starting\", self.name)\n self._status = Status.RUNNING\n\n def pause(self) -> None:\n \"\"\"Mark the job as paused.\"\"\"\n if self._status not in [Status.RUNNING, Status.QUEUED]:\n raise JobStatusException(\"Only a running job can be paused.\")\n logger.debug(\"Job '%s' paused\", self.name)\n self._status = Status.PAUSED\n\n def complete(self) -> None:\n \"\"\"Mark the job as completed.\"\"\"\n if self._status is not Status.RUNNING:\n raise JobStatusException(\"Only a running job can be completed.\")\n logger.debug(\"Job '%s' marked as completed\", self.name)\n self._status = Status.DONE\n\n def queue(self) -> None:\n \"\"\"Mark the job as queued.\"\"\"\n if self._status not in [Status.PENDING, Status.PAUSED]:\n raise JobStatusException(\n \"Only a pending or paused job can be queued.\",\n )\n logger.debug(\"Job '%s' marked as queued\", self.name)\n self._status = Status.QUEUED\n\n def mark_as_error(self) -> None:\n \"\"\"Mark the job as in a error state.\"\"\"\n if self._status is not Status.RUNNING:\n raise JobStatusException(\"Only a running job can error\")\n logger.debug(\"Job '%s' has status as error\", self.name)\n self._status = Status.ERROR\n\n\nclass JobStatusException(Exception):\n \"\"\"Exception when job attempt to change into invalid state.\"\"\"\n\n\nclass Project:\n \"\"\"Project class.\n\n Top level abstraction for organisation of jobs connected to specified\n projects.\n\n Parameters\n ----------\n id : int\n Project internal id number\n name : str\n Project name\n number : str\n A unique project number. This number is a reference to external\n reference number used by the user.\n location : str\n Optional string representing the location for this project.\n description : str\n Project description.\n\n Attribute\n ---------\n number_of_jobs : int\n Number of jobs associated with project.\n\n Methods\n -------\n add_job(job: Job)\n Adds a new job to project. No duplicates allowed.\n remove_job(number: str)\n Removes job from project.\n list_jobs()\n Returns a list of associated _jobs_.\n \"\"\"\n\n def __init__(\n self,\n name: str,\n number: str,\n description: str,\n location: str | None = None,\n ) -> None:\n self.id: int\n self.name: str = name\n self.number: str = number\n self.description: str = description\n self.location: str | None = location\n self.jobs: list[Job] = []\n\n def __str__(self) -> str:\n \"\"\"Print class members.\"\"\"\n return f\"Name: {self.name}, Description: {self.description}\"\n\n def __eq__(self, other: object) -> bool:\n \"\"\"Check equality between objects.\n\n Operator used in tests to check if objects from DB is correct.\n \"\"\"\n if not isinstance(other, Project):\n return False\n return (\n other.id == self.id\n and other.name == self.name\n and other.description == self.description\n and other.number == self.number\n and other.location == self.location\n )\n\n def __hash__(self) -> int:\n \"\"\"Hash of object used in eg. `dict()` or `set()` to avoid duplicate.\"\"\"\n return hash((type(self),) + tuple(self.__dict__))\n\n def __repr__(self) -> str:\n \"\"\"Override of default __repr__. Gives object representation as a string.\"\"\"\n return (\n str(self.__class__) + \": \" + str(self.__dict__)\n ) # pragma: no cover\n\n @classmethod\n def from_dict(cls, project_data: dict) -> Project:\n \"\"\"Only an example method of a \"named constructor\".\"\"\"\n return cls(\n name=project_data[\"name\"],\n number=project_data[\"number\"],\n description=project_data[\"description\"],\n location=project_data[\"location\"],\n )\n\n @property\n def number_of_jobs(self) -> int:\n \"\"\"Get number of jobs associated with project.\n\n Returns\n -------\n int\n Number of jobs in project\n \"\"\"\n return len(self.jobs)\n\n def add_job(self, job: Job) -> Project:\n \"\"\"Add job to project.\n\n Parameter\n ---------\n job : Job\n Job to be added to project.\n \"\"\"\n if job in self.jobs:\n logger.debug(\n \"Attempted to add existing job '%s' to a project\",\n job.name,\n )\n else:\n logger.debug(\"Added job '%s' to project\", job.name)\n self.jobs.append(job)\n return self\n\n def get_jobs(self) -> list[Job]:\n \"\"\"Retrieve all jobs from the project.\n\n Returns\n -------\n : List[Job]\n List containing all jobs within the project\n \"\"\"\n return self.jobs\n\n def get_job(self, job_id: int) -> Job | None:\n \"\"\"Retrive a single job from the project.\n\n Parameters\n ----------\n job_id : int\n Index of the job we seek. 0 is not valid.\n\n Returns\n -------\n Job\n The job object if found.\n \"\"\"\n for job in self.jobs:\n if job.id == job_id:\n return job\n\n return None\n\n def remove_job(self, job: Job) -> bool:\n \"\"\"Remove job from project.\n\n Parameters\n ----------\n job : Job\n Job to be removed.\n\n Returns\n -------\n bool\n True if the job was successfully removed\n \"\"\"\n if job in self.jobs:\n self.jobs.remove(job)\n logger.debug(\"Removed job with name '%s' from a project\", job.name)\n return True\n else:\n logger.debug(\n \"Could not find job with name '%s' to remove in project\",\n job.name,\n )\n return False\n","repo_name":"tomrtk/fish-code","sub_path":"src/core/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":39728,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"7"} +{"seq_id":"72864906782","text":"import pytest\nimport torch\n\nfrom mmseg.models.backbones.vit import VisionTransformer\nfrom .utils import check_norm_state\n\n\ndef test_vit_backbone():\n with pytest.raises(TypeError):\n # pretrained must be a string path\n model = VisionTransformer()\n model.init_weights(pretrained=0)\n\n with pytest.raises(TypeError):\n # img_size must be int or tuple\n model = VisionTransformer(img_size=512.0)\n\n with pytest.raises(TypeError):\n # out_indices must be int ,list or tuple\n model = VisionTransformer(out_indices=1.)\n\n with pytest.raises(TypeError):\n # test upsample_pos_embed function\n x = torch.randn(1, 196)\n VisionTransformer.resize_pos_embed(x, 512, 512, 224, 224, 'bilinear')\n\n with pytest.raises(IndexError):\n # forward inputs must be [N, C, H, W]\n x = torch.randn(3, 30, 30)\n model = VisionTransformer()\n model(x)\n\n with pytest.raises(AssertionError):\n # The length of img_size tuple must be lower than 3.\n VisionTransformer(img_size=(224, 224, 224))\n\n with pytest.raises(TypeError):\n # Pretrained must be None or Str.\n VisionTransformer(pretrained=123)\n\n with pytest.raises(AssertionError):\n # with_cls_token must be True when output_cls_token == True\n VisionTransformer(with_cls_token=False, output_cls_token=True)\n\n # Test img_size isinstance tuple\n imgs = torch.randn(1, 3, 224, 224)\n model = VisionTransformer(img_size=(224, ))\n model.init_weights()\n model(imgs)\n\n # Test img_size isinstance tuple\n imgs = torch.randn(1, 3, 224, 224)\n model = VisionTransformer(img_size=(224, 224))\n model(imgs)\n\n # Test norm_eval = True\n model = VisionTransformer(norm_eval=True)\n model.train()\n\n # Test ViT backbone with input size of 224 and patch size of 16\n model = VisionTransformer()\n model.init_weights()\n model.train()\n\n assert check_norm_state(model.modules(), True)\n\n # Test normal size input image\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test large size input image\n imgs = torch.randn(1, 3, 256, 256)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 16, 16)\n\n # Test small size input image\n imgs = torch.randn(1, 3, 32, 32)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 2, 2)\n\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test unbalanced size input image\n imgs = torch.randn(1, 3, 112, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 7, 14)\n\n # Test irregular input image\n imgs = torch.randn(1, 3, 234, 345)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 15, 22)\n\n # Test with_cp=True\n model = VisionTransformer(with_cp=True)\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test with_cls_token=False\n model = VisionTransformer(with_cls_token=False)\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test final norm\n model = VisionTransformer(final_norm=True)\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test patch norm\n model = VisionTransformer(patch_norm=True)\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[-1].shape == (1, 768, 14, 14)\n\n # Test output_cls_token\n model = VisionTransformer(with_cls_token=True, output_cls_token=True)\n imgs = torch.randn(1, 3, 224, 224)\n feat = model(imgs)\n assert feat[0][0].shape == (1, 768, 14, 14)\n assert feat[0][1].shape == (1, 768)\n","repo_name":"Ascend/ModelZoo-PyTorch","sub_path":"PyTorch/contrib/cv/semantic_segmentation/DeeplabV3_for_Pytorch/tests/test_models/test_backbones/test_vit.py","file_name":"test_vit.py","file_ext":"py","file_size_in_byte":3780,"program_lang":"python","lang":"en","doc_type":"code","stars":31,"dataset":"github-code","pt":"7"} +{"seq_id":"72685358624","text":"\"\"\"\n# Definition for a Node.\nclass Node:\n def __init__(self, val, children):\n self.val = val\n self.children = children\n\"\"\"\n\nfrom collections import deque\n\nclass Solution:\n def maxDepth(self, root: 'Node') -> int:\n if root is None:\n return 0\n\n self.max_depth = 0\n self._helper(root, 1)\n\n return self.max_depth\n \n def _helper(self, node, depth: int):\n self.max_depth = max(depth, self.max_depth)\n for child in node.children:\n if child is not None:\n self._helper(child, depth+1)\n","repo_name":"daviddwlee84/LeetCode","sub_path":"Python3/BinaryTree/MaximumDepthOfNaryTree/Naive559.py","file_name":"Naive559.py","file_ext":"py","file_size_in_byte":581,"program_lang":"python","lang":"en","doc_type":"code","stars":15,"dataset":"github-code","pt":"7"} +{"seq_id":"22961876795","text":"# Zadatak 1.15Napisati program koji imitira rad bafera. Maksimalni broj elemenata u baferuje5. Korisnik sa standardnog ulaza unosi podatke do unosa rečiquit. Program ih smešta u bafer,posto se bafer napuni unosi se ispisuju na standarni izlaz i bafer se prazn\n\nbufer = [] \n\ni = 0 \n\nwhile (True):\n try:\n bufer.append(int(input()))\n i += 1\n \n if i == 5:\n i = 0\n print(bufer)\n bufer = []\n print(\"bufer se ispraznio\")\n continue\n\n except:\n print(\"FAILED\")\n break","repo_name":"fongsd/pp","sub_path":"pp/python/1_15.py","file_name":"1_15.py","file_ext":"py","file_size_in_byte":563,"program_lang":"python","lang":"sh","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25872478405","text":"import pandas as pd\nimport numpy as np\nimport scipy.sparse as sp\nimport time, copy, os, pickle, re\nfrom lightfm import LightFM\nfrom fuzzywuzzy import fuzz, process\n\nfrom utils import (\n DatasetFaster,\n fetch_user_ratings_dataset,\n fetch_user_ratings_goodreads,\n predict_list,\n load_model,\n book_map,\n fancy_title,\n fancy_list\n)\n\nfrom config import *\n\nfrom telegram import Update, ReplyKeyboardMarkup\nfrom telegram.ext import (\n Updater,\n CommandHandler,\n MessageHandler,\n Filters,\n ConversationHandler,\n CallbackQueryHandler,\n CallbackContext,\n)\n\n\nCHOOSING_SCENARIO, SELECTING_ENGINE = map(chr, range(2))\nCHOOSING_USER_ID, CHOOSING_GR_ID, CHOOSING_CUSTOM = map(chr, range(2, 5))\nTYPING_USER, TYPING_GOODREADS, TYPING_BOOK, SELECTING_RATING = map(chr, range(5, 9))\n\n\ndef start(update, context):\n context.user_data['user_ratings'] = None\n context.user_data['rated_dict'] = {}\n context.user_data['selected_book'] = None\n \n update.message.reply_text(text=start_text, disable_web_page_preview=True)\n \n buttons = [[\n 'Dataset Id',\n 'GoodReads Id',\n 'Custom Setup'\n ]]\n keyboard = ReplyKeyboardMarkup(buttons, one_time_keyboard=True)\n update.message.reply_text(text=scenario_text, reply_markup=keyboard)\n\n return CHOOSING_SCENARIO\n\n\ndef recommend(update, context):\n buttons = [[\n 'LightFM',\n #'Hybrid LightFM'\n ]]\n keyboard = ReplyKeyboardMarkup(buttons, one_time_keyboard=True)\n update.message.reply_text(text=recommend_text, reply_markup=keyboard)\n\n return SELECTING_ENGINE\n\n\ndef ask_user_id(update, context):\n update.message.reply_text(text=id_text)\n \n return TYPING_USER\n\n\ndef ask_user_id_again(update, context):\n update.message.reply_text(text=wrongid_text)\n \n return TYPING_USER\n\n\ndef save_user_id(update, context):\n try:\n user_id = int(update.message.text)\n context.user_data['user_ratings'] = fetch_user_ratings_dataset(user_id)\n return recommend(update, context)\n except ValueError:\n return ask_user_id_again(update, context)\n\n\ndef ask_goodreads_id(update, context):\n update.message.reply_text(text=gr_text)\n \n return TYPING_GOODREADS\n\n\ndef ask_goodreads_id_again(update, context):\n update.message.reply_text(text=nouser_text)\n \n return TYPING_GOODREADS\n\n\ndef save_goodreads_id(update, context):\n try:\n goodreads_id = int(re.search('[0-9]+', update.message.text).group())\n context.user_data['user_ratings'] = fetch_user_ratings_goodreads(goodreads_id)\n return recommend(update, context)\n except:\n return ask_goodreads_id_again(update, context)\n\n \ndef inform_books_rating(update, context):\n update.message.reply_text(text=rating_text)\n \n return ask_book_rating(update, context)\n\n\ndef ask_book_rating(update, context):\n buttons = [['Finish']]\n keyboard = ReplyKeyboardMarkup(buttons, one_time_keyboard=True)\n update.message.reply_text(text=askbook_text, reply_markup=keyboard)\n \n return TYPING_BOOK\n\n\ndef save_selected_book(update, context):\n book_name = update.message.text\n selected_book = process.extract(book_name, book_map['title'].values, scorer=fuzz.ratio)[0][0]\n context.user_data['selected_book'] = selected_book\n\n return show_selected_book(update, context)\n\n\ndef show_selected_book(update, context):\n buttons = [['1', '2', '3', '4', '5'],\n ['Cancel', 'Finish']]\n keyboard = ReplyKeyboardMarkup(buttons, one_time_keyboard=True)\n update.message.reply_text(text=fancy_title(context.user_data['selected_book']),\n parse_mode='HTML',\n disable_web_page_preview=True)\n update.message.reply_text(text=nowrate_text, reply_markup=keyboard)\n\n return SELECTING_RATING\n\n\ndef save_book_rating(update, context):\n book_rating = int(update.message.text)\n context.user_data['rated_dict'][context.user_data['selected_book']] = book_rating\n\n return ask_book_rating(update, context)\n\n\ndef rating_finished(update, context):\n books, rates = zip(*context.user_data['rated_dict'].items())\n book_ids = book_map['id'][book_map['title'].isin(books)].values\n\n context.user_data['user_ratings'] = pd.DataFrame(\n {'user_id': np.repeat(0, len(book_ids)), 'book_id': book_ids, 'rating': rates})\n\n return recommend(update, context)\n\n\ndef rec_lightfm(update, context):\n lightfm = load_model('lightfm1.pickle')\n warn = update.message.reply_text(text=wait_text)\n \n ratings = pd.read_csv('data/ratings.csv')\n dataset = DatasetFaster()\n dataset.fit(ratings.user_id.nunique(), 10000)\n interactions_new, weights_new = dataset.build_interactions(ratings, context.user_data['user_ratings'])\n\n lightfm.fit_partial(\n interactions=interactions_new,\n sample_weight=weights_new\n )\n \n reclist = predict_list(lightfm, context.user_data['user_ratings'])\n # warn.edit_text(text=fancy_list(reclist), parse_mode='HTML', disable_web_page_preview=True)\n rec = update.message.reply_text(text=fancy_list(reclist), parse_mode='HTML', disable_web_page_preview=True)\n\n\nconvhandler = ConversationHandler(\n entry_points=[CommandHandler('start', start)],\n states={\n CHOOSING_SCENARIO: [\n MessageHandler(Filters.regex('^Dataset Id$'), ask_user_id),\n MessageHandler(Filters.regex('^GoodReads Id$'), ask_goodreads_id),\n MessageHandler(Filters.regex('^Custom Setup$'), inform_books_rating)\n ],\n TYPING_USER: [\n MessageHandler(Filters.text & ~Filters.command, save_user_id)\n ],\n TYPING_GOODREADS: [\n MessageHandler(Filters.text & ~Filters.command, save_goodreads_id)\n ],\n TYPING_BOOK: [\n MessageHandler(Filters.text & ~Filters.command & ~Filters.regex('^Finish$'), save_selected_book),\n MessageHandler(Filters.regex('^Finish$'), rating_finished)\n ],\n SELECTING_RATING: [\n MessageHandler(Filters.regex('^[1-5]$'), save_book_rating),\n MessageHandler(Filters.regex('^Finish$'), rating_finished),\n MessageHandler(Filters.regex('^Cancel$'), ask_book_rating)\n ],\n SELECTING_ENGINE: [\n MessageHandler(Filters.regex('^LightFM$'), rec_lightfm),\n # MessageHandler(Filters.regex('^Hybrid LightFM$'), rec_lightfm_hybrid)\n ]\n },\n fallbacks=[CommandHandler('start', start)]\n)\n\n\ndef main():\n TOKEN = os.environ['TOKEN']\n PORT = int(os.environ.get('PORT', '8443'))\n \n updater = Updater(token=TOKEN, use_context=True)\n dispatcher = updater.dispatcher\n dispatcher.add_handler(convhandler)\n updater.start_webhook(\n listen='0.0.0.0',\n port=PORT,\n url_path=TOKEN,\n webhook_url='https://goodbooks-bot.herokuapp.com/' + TOKEN\n )\n updater.idle()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"yuasosnin/goodbooks-recommender","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":6924,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"1449174706","text":"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\nPREDICTION/VIEWS.py\n\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\n\nimport os\nimport json\nfrom random import randint\n\nfrom django.shortcuts import render\nfrom django.http import JsonResponse\nfrom django.utils.safestring import mark_safe\nfrom django.conf import settings\n\nimport common.utility as CU\nimport football.models.football as FM\nimport members.models.members as MM\nimport prediction.models.universal as PU\n\nimport logging\nprog_lg = logging.getLogger('progress')\nexcp_lg = logging.getLogger('exception')\n\n\n\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\nUNIVERSAL\n\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\n\n\ndef univ_preds(request):\n bk = FM.TimeMachine.GetTodaysBracket()\n prog_lg.debug(bk)\n context = {\n 'season': bk['season'],\n 'round': bk['round'],\n 'now_dt': mark_safe(json.dumps(MM.Profile_Reporter.GetUserNow(request.user))),\n 'spectrumOpts': [\"Home Win\", \"Away Win\", \"Tie\", \"Abstain\"],\n 'players': mark_safe(json.dumps(FM.Reports_General.GetPlayersBySeason(bk['season']))),\n 'abilsOwned': mark_safe(json.dumps(PU.Reporter.GetAbilitiesOwned(request.user, bk['season']))),\n 'abilsUsedRound': mark_safe(json.dumps(PU.Reporter.GetAbilsUsedByRound(request.user, bk['season'], bk['round']))),\n 'fixture': mark_safe(json.dumps(PU.Reporter_Common.GetFixtureLocalized(bk['season'], bk['round'], request.user))),\n 'predictions': mark_safe(json.dumps(PU.Editor.GetOrCreatePreds(bk['season'], bk['round'], request.user))),\n }\n return render(request, 'pu_predictions.html', context)\n\n\ndef univ_headq(request):\n season = FM.TimeMachine.GetTodaysBracket()['season']\n \n context = {\n 'roster': PU.Reporter.GetUserRoster(request.user, season),\n 'abilsOwned': PU.Reporter.GetAbilitiesOwned(request.user, season),\n 'store': PU.Reporter_Store.GetStoreData(),\n 'storeAvailable': PU.Reporter_Store.GetStoreAvailable(request.user, season),\n }\n return render(request, 'pu_headquarters.html', context)\n\n\ndef univ_ranks(request): \n seasons = FM.Reports_General.GetSeasons()\n #seasons = ['IPL2015']\n season = FM.TimeMachine.GetTodaysBracket()['season']\n roundList = FM.Reports_General.GetRounds(season, \"lastData\")\n roundv = roundList[0] if roundList else None\n \n puRecord = PU.Reporter.GetRecord(request.user, season, roundv) \n rankData = PU.Reporter_Ranks.RunRankData(season, roundv, \"Friends\", request.user)\n \n context = {\n 'seasons': seasons,\n 'roundList': roundList,\n 'modes': [\"Friends\", \"Yours\", \"Top Users\"],\n \n # logged user's performance\n \n 'status': mark_safe(json.dumps(puRecord['status'])),\n 'recordAT': mark_safe(json.dumps(puRecord['recordAT'])),\n 'recordRD': mark_safe(json.dumps(puRecord['recordRD'])),\n 'thresholds': mark_safe(json.dumps(puRecord['thresholds'])),\n \n # points distribution\n \n 'userCntAT': rankData['userCntAT'],\n 'binCntAT': rankData['binCntAT'],\n 'histAT': mark_safe(json.dumps(rankData['histAT'])),\n 'gradeAT': mark_safe(json.dumps(rankData['gradeAT'])),\n 'highlightAT': mark_safe(json.dumps(rankData['highlightAT'])),\n \n 'userCntRD': rankData['userCntRD'],\n 'binCntRD': rankData['binCntRD'],\n 'histRD': mark_safe(json.dumps(rankData['histRD'])),\n 'gradeRD': mark_safe(json.dumps(rankData['gradeRD'])),\n 'highlightRD': mark_safe(json.dumps(rankData['highlightRD'])),\n \n # leaderboards\n \n 'ranksAT': mark_safe(json.dumps(rankData['ranksAT'])),\n 'ranksRD': mark_safe(json.dumps(rankData['ranksRD'])),\n }\n return render(request, 'pu_standings.html', context)\n\n\ndef univ_rules(request):\n context = {\n }\n return render(request, 'pu_rules.html', context)\n\n\ndef universal_jx(request, command):\n \n prog_lg.info(\"ajax edit command: \" + command)\n \n \n if command == 'get_fixtSummary':\n season = request.GET.get('season')\n hret = PU.Reporter.GetFixturesSummary(season) \n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n elif command == 'delete_events': \n hret = PU.Editor_Admin.DeleteEvents()\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n \n elif command == 'update_accumulator': \n hret = PU.Editor.PopTokenAccumulator(request.user)\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n elif command == 'update_buyUpgrade': \n upgradeType = request.POST.get('upgradeType')\n upgradeLevel = request.POST.get('upgradeLevel')\n hret = PU.Editor.UpgradeBuy(request.user, upgradeType, upgradeLevel)\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n \n elif command == 'update_prediction': \n pred_st = request.POST.get('pred_st')\n pred_dx = json.loads(pred_st)\n hret = PU.Editor.SavePrediction(request.user, pred_dx)\n \n bk = FM.TimeMachine.GetTodaysBracket()\n hret.results = {\n 'saveRes': hret.results,\n 'now_dt': MM.Profile_Reporter.GetUserNow(request.user),\n 'abilsOwned': PU.Reporter.GetAbilitiesOwned(request.user, bk['season']),\n 'abilsUsedRound': PU.Reporter.GetAbilsUsedByRound(request.user, bk['season'], bk['round']),\n 'predictions': PU.Editor.GetOrCreatePreds(bk['season'], bk['round'], request.user),\n }\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n elif command == 'get_predictions': \n season = FM.TimeMachine.GetTodaysBracket()['season']\n roundv = request.GET.get('round')\n \n hret = CU.HttpReturn()\n hret.status = 201\n hret.results = {\n 'now_dt': MM.Profile_Reporter.GetUserNow(request.user),\n 'abilsOwned': PU.Reporter.GetAbilitiesOwned(request.user, season),\n 'abilsUsedRound': PU.Reporter.GetAbilsUsedByRound(request.user, season, roundv),\n 'fixture': PU.Reporter_Common.GetFixtureLocalized(season, roundv, request.user),\n 'predictions': PU.Editor.GetOrCreatePreds(season, roundv, request.user),\n }\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n \n elif command == 'get_ranksRnd':\n mode = request.GET.get('mode')\n season = request.GET.get('season')\n \n roundList = FM.Reports_General.GetRounds(season, 'lastData')\n if not roundList:\n roundList = ['01']\n lastRound = roundList[0] # list is sorted descending\n \n results = PU.Reporter_Ranks.RunRankData(season, lastRound, mode, request.user)\n puRecord = PU.Reporter.GetRecord(request.user, season, lastRound)\n \n results = dict(results, **puRecord);\n results['roundList'] = roundList\n \n hret = CU.HttpReturn()\n hret.results = results\n hret.status = 201\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n elif command == 'get_ranks':\n mode = request.GET.get('mode')\n season = request.GET.get('season')\n roundv = request.GET.get('roundv')\n \n results = PU.Reporter_Ranks.RunRankData(season, roundv, mode, request.user)\n puRecord = PU.Reporter.GetRecord(request.user, season, roundv) \n results = dict(results, **puRecord);\n \n hret = CU.HttpReturn()\n hret.results = results\n hret.status = 201\n return JsonResponse(hret.results, safe=False, status=hret.status)\n \n \n else:\n msg = \"command invalid: \" + command\n excp_lg.error(msg)\n return JsonResponse(msg, safe=False, status = 404)\n\n\n\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\nEND OF FILE\n\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"","repo_name":"PhilBusko/PersianLeague","sub_path":"prediction/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":8127,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"8845612311","text":"import time\nimport numpy as np\nimport pyautogui\n\npyautogui.moveTo(1520, 800)\npyautogui.click()\n\n\ndef capture():\n position = pyautogui.position()\n screenshot = pyautogui.screenshot(region=(position[0] - 1, position[1] - 1, 1, 1))\n screenshot = np.array(screenshot)\n return screenshot[0, 0]\n\n\nclick_counter = 0\nmax_click = 100\nis_cast = True\nwhile True:\n color_code = capture()\n print(color_code, click_counter)\n if is_cast and color_code[0] > 100 and color_code[1] > 110 and color_code[2] > 110:\n time.sleep(0.5)\n pyautogui.moveTo(1520, 800)\n pyautogui.click()\n click_counter += 1\n is_cast = False\n if color_code[0] > 130 and color_code[1] > 175 and color_code[2] > 110:\n pyautogui.moveTo(1520, 800)\n pyautogui.click()\n click_counter += 1\n is_cast = True\n time.sleep(3)\n if click_counter >= max_click:\n break\n","repo_name":"Bug-Too/clicky","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":912,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"4241413541","text":"import atexit\n\nimport numpy as np\n\nimport pycuda.driver as cuda\nimport pycuda.gpuarray as gpuarray\nfrom pycuda.compiler import SourceModule\nfrom pycuda.driver import ctx_flags\n\ncuda.init()\ncudadevice = cuda.Device(0)\ncudacontext = cudadevice.make_context(flags=ctx_flags.SCHED_YIELD)\natexit.register(cudacontext.pop)\n\nimport skcuda\nfrom skcuda.misc import _get_minmax_kernel\nskcuda.misc._global_cublas_allocator = cuda.mem_alloc\n\nfrom .pattern import Pattern\n\n\nclass PatternCUDA(Pattern):\n\n mod = SourceModule(\"\"\"\n __global__ void correlate(float *input, float *patterns, float *result, int range_low, int range_high)\n {\n int x = (threadIdx.x + (blockDim.x*blockIdx.x));\n int y = (threadIdx.y + (blockDim.y*blockIdx.y));\n int iidx = x * 8;\n int ridx = (x * blockDim.y * gridDim.y) + y;\n int pidx = y * 24;\n \n float d;\n result[ridx] = 0;\n \n for (int i=range_low;i d[most]:\n most = key\n\nprint(most, d[most])\n","repo_name":"demidovakatya/notebooks","sub_path":"old_python_code/most_messages.py","file_name":"most_messages.py","file_ext":"py","file_size_in_byte":388,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"17764492765","text":"from subprocess import Popen, PIPE, STDOUT\n\n\ndef test_nbi_jupyterhub():\n p = Popen(['jupyterhub', '--debug'], stdin=PIPE, stdout=PIPE, stderr=STDOUT)\n running = False\n for line in p.stdout:\n if 'JupyterHub is now running at' in str(line):\n running = True\n break\n\n assert running\n","repo_name":"ucphhpc/docker-jupyterhub","sub_path":"hub/tests/test_jupyterhub.py","file_name":"test_jupyterhub.py","file_ext":"py","file_size_in_byte":320,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"72157435424","text":"import pandas as pd\nfrom datetime import datetime\nimport datetime as dt\nimport csv\nimport os\n\n\nVIDNAME = 0\nQ_START = 5\nQ_END = 6\nDB_START = 3\nDB_END = 4\nSIM = 2\nLENGTH = 7\n\nWHITE = '\\u001b[37m'\nBLUE = '\\033[94m'\nGREEN = '\\033[92m'\nYELLOW = '\\033[93m'\nRED = '\\033[91m'\nENDC = '\\033[0m'\nGREY = '\\u001b[38;5;244m'\nBOLD = '\\u001b[1m'\nBOLDOFF = '\\u001b[21m'\nUNDERLINE = '\\u001b[4m'\n\ndef join_from_path(logpath1, logpath2, destdir):\n log1 = pd.read_csv(logpath1, index_col=None)\n log2 = pd.read_csv(logpath2, index_col=None)\n\n result = log1\n\n log1 = log1.reset_index()\n log1.columns.values[0] = 'id'\n log1['id'] = log1.index\n\n nonzero1 = log1.loc[log1['Confidence'] > 0]\n nonzero2 = log2.loc[log2['Confidence'] > 0]\n\n for index1, row1 in nonzero1.iterrows():\n for index2, row2 in nonzero2.iterrows():\n if ((row1['Database Video'] == row2['Database Video']) and ((row2['Start Time'] <= row1['Start Time'] < row2['End Time']) or (row1['Start Time'] <= row2['Start Time'] < row1['End Time']))):\n l1 = row1['End Time'] - row1['Start Time']\n l2 = row2['End Time'] - row2['Start Time']\n p1 = l1 / (l1 + l2)\n p2 = l2 / (l1 + l2)\n newscore = row1['Confidence']*p1 + row2['Confidence']*p2\n result.iloc[index1, result.columns.get_loc('Confidence')] = newscore\n result.iloc[index1, result.columns.get_loc('Start Time')] = min(row1['Start Time'], row2['Start Time'])\n result.iloc[index1, result.columns.get_loc('Query Start Time')] = min(row1['Query Start Time'], row2['Query Start Time'])\n result.iloc[index1, result.columns.get_loc('End Time')] = max(row1['End Time'], row2['End Time'])\n result.iloc[index1, result.columns.get_loc('Query End Time')] = max(row1['Query End Time'], row2['Query End Time'])\n\n # print(result)\n outpath = os.path.join(destdir, \"combined.mp4.csv\")\n result.to_csv(outpath, index=False)\n return outpath\n\ndef read_logfile(logpath, shortestmatch):\n logfile = pd.read_csv(logpath, index_col=None)\n\n df = logfile.groupby('Database Video').sum().reset_index()\n threshold = df['Confidence'].mean() + 1.5*df['Confidence'].std()\n df = df.loc[df['Confidence'] > threshold]\n df['Length'] = (df['End Time'] - df['Start Time']) / 1000.\n df = df.sort_values(['Confidence', 'Length'])\n video_names = df['Database Video'].tolist()\n\n # df['Length'] = df.apply(lambda row: float(row['End Time'] - row['Start Time']) / 1000, axis=1)\n df = df.loc[df['Length'] > shortestmatch]\n logfile = logfile[logfile['Database Video'].isin(df['Database Video']).tolist()]\n logfile['Length'] = (logfile['End Time'] - logfile['Start Time']) / 1000.\n\n with open(\"./results/resultcache.txt\", \"w\") as file:\n for item in video_names:\n file.write(f\"{item}\\n\")\n\n return logfile.to_numpy()\n\ndef str_timestamp(num_ms):\n n_sec = int(dt.timedelta(milliseconds=num_ms).total_seconds())\n\n n_hrs = n_sec // 3600\n n_sec = n_sec % 3600\n\n n_min = n_sec // 60\n n_sec = n_sec % 60\n\n return \"{:02d}:{:02d}:{:02d}\".format(n_hrs, n_min, n_sec)\n\n\ndef print_log(logfile):\n if len(logfile) < 1:\n print(\"NO MATCHES FOUND\")\n return\n print(f\"\\n{BOLD}MATCH(ES) FOUND:{ENDC}\\n\")\n print(\"{}:{:>30}{:>16}{:>20}{:>24}{:>24}\".format(\n \"#\", \"Name of Matching Video\",\"Similarity\",\"Length (sec.)\", \"Range in DB\", \"Range in Query\"))\n for i, row in enumerate(logfile):\n score = row[SIM] * 100\n score_color = RED\n if score > 90:\n score_color = GREEN\n elif score > 75:\n score_color = YELLOW\n\n db_range = \"{} - {}\".format(\n str_timestamp(row[DB_START]),\n str_timestamp(row[DB_END]))\n query_range = \"{} - {}\".format(\n str_timestamp(row[Q_START]),\n str_timestamp(row[Q_END]))\n print(\"{}:\\033[94m{:>30}\\033[0m{}{:>15.3f}%\\033[0m\\u001b[38;5;244m{:>20}{:>24}{:>24}\\033[0m\".format(\n i, row[VIDNAME],score_color,score,row[LENGTH],db_range, query_range))\n print(\"\\n\")\n","repo_name":"cnagda/id-pirated-vid","sub_path":"python/logfile.py","file_name":"logfile.py","file_ext":"py","file_size_in_byte":4162,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"69804165982","text":"# 给你一个字符串 s,请你将 s 分割成一些子串,使每个子串都是 回文串 。返回 s 所有可能的分割方案。 \n# \n# 回文串 是正着读和反着读都一样的字符串。 \n# \n# \n# \n# 示例 1: \n# \n# \n# 输入:s = \"aab\"\n# 输出:[[\"a\",\"a\",\"b\"],[\"aa\",\"b\"]]\n# \n# \n# 示例 2: \n# \n# \n# 输入:s = \"a\"\n# 输出:[[\"a\"]]\n# \n# \n# \n# \n# 提示: \n# \n# \n# 1 <= s.length <= 16 \n# s 仅由小写英文字母组成 \n# \n# Related Topics 字符串 动态规划 回溯 👍 808 👎 0\n\n\n# leetcode submit region begin(Prohibit modification and deletion)\nclass Solution:\n def partition(self, s: str) -> List[List[str]]: # 回溯+DP\n length = len(s)\n judge = [[False] * length for _ in range(length)]\n sublist = []\n container = []\n for i in range(length):\n judge[i][i] = True\n for start in range(length - 1, 0, -1):\n for end in range(start, length):\n judge[start - 1][end] = (start >= end - 1 or judge[start][end - 1]) \\\n and s[start - 1] == s[end]\n\n def backtrack(start=0):\n if start == length:\n container.append(sublist[:])\n for end in range(start, length):\n if judge[start][end]:\n sublist.append(s[start: end + 1])\n backtrack(end + 1)\n sublist.pop()\n\n backtrack()\n return container\n\n# leetcode submit region end(Prohibit modification and deletion)\n","repo_name":"xxsddm/Algorithm-Beginner","sub_path":"leetcode/0101-0200/[131]分割回文串.py","file_name":"[131]分割回文串.py","file_ext":"py","file_size_in_byte":1549,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7609939139","text":"#!/bin/python3\n\nimport sys\n\nN = int(input())\narr = list(map(int,input().split()))\n\nif all([x % 2 == 0 for x in arr]):\n print(0)\n\nelif all([x % 2 != 0 for x in arr]):\n if N % 2 == 0:\n print(N)\n else:\n print('NO')\n\nelse:\n count = 0\n\n if arr[0] % 2 != 0 and arr[1] % 2 == 0:\n arr[0] += 1\n arr[1] += 1\n count += 2\n\n for i in range(1, N):\n if arr[i] % 2 != 0 and arr[i - 1] % 2 != 0:\n arr[i] += 1\n arr[i - 1] += 1\n count += 2\n if arr[i] % 2 != 0 and arr[i - 1] % 2 == 0:\n if i + 1 != N:\n arr[i] += 1\n arr[i + 1] += 1\n count += 2\n\n if arr[-1] % 2 != 0:\n print('NO')\n else:\n print(count)\n","repo_name":"H-Shen/Collection_of_my_coding_practice","sub_path":"Hackerrank/Algorithms/Fair_Rations.py3","file_name":"Fair_Rations.py3","file_ext":"py3","file_size_in_byte":757,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"36895552246","text":"from app import app, db\nfrom faker import Faker\nfrom models import User, Product, Brand, Category, Invoice, InvoiceProducts, Invoice, Role\nfrom datetime import datetime\nfrom sqlalchemy.orm import sessionmaker\nimport random\n\nfake = Faker()\n\nwith app.app_context():\n db.create_all()\n def delete_data():\n # this deletes existing db data in columns \n print(\"🦸 Delete_data...\")\n User.query.delete()\n Product.query.delete()\n Category.query.delete()\n Brand.query.delete()\n Invoice.query.delete()\n InvoiceProducts.query.delete()\n Role.query.delete()\n \n def seed_data():\n print(\"🦸‍♀️ Seeding User Roles...\")\n admin_role = Role(name='Admin')\n db.session.add(admin_role)\n\n client_role = Role(name='Client')\n db.session.add(client_role)\n\n print(\"🦸‍♀️ Seeding Users with Faker...\")\n roles = [1] + [random.choice([2, 3, 4]) for _ in range(49)]\n\n for i in range(50): # Generate 50 fake users\n user = User(\n first_name=fake.first_name(),\n last_name=fake.last_name(),\n username=fake.user_name(),\n email=fake.email(),\n ph_address=fake.address(),\n password=fake.password(),\n telephone=fake.phone_number(),\n city_town=fake.city(),\n role=roles[i] \n )\n db.session.add(user)\n\n db.session.commit()\n\n print(\"🦸‍♀️ Seeding Brands with Faker...\")\n for _ in range(50): # Generate 2 fake brands\n brand = Brand(\n brand_name=fake.company(),\n brand_logo=fake.url()\n )\n db.session.add(brand)\n\n categories = (\"skin\", \"face\", \"nails\", \"eyes\", \"hair\")\n print(\"🦸‍♀️ Seeding Categories with Faker...\")\n for _ in range(5): # Generate 4 fake categories\n category_name = fake.unique.random_element(elements=categories)\n category = Category(cat_name=category_name)\n db.session.add(category)\n\n print(\"🦸‍♀️ Seeding Products with Faker...\")\n product_1 = Product(image=\"https://images.pexels.com/photos/3373739/pexels-photo-3373739.jpeg?auto=compress&cs=tinysrgb&w=1260&h=750&dpr=1\", p_name=\"Huddah Lipstick - Red Velvet\", description=\"High-quality red lipstick that provides a smooth and long-lasting finish. Perfect for any occasion.\", price=200, quantity=30, category=1, brand=1)\n product_2 = Product(image=\"https://images.pexels.com/photos/5403543/pexels-photo-5403543.jpeg?auto=compress&cs=tinysrgb&w=1260&h=750&dpr=1\", p_name=\"Rihanna Mascara - Volume Boost\", description=\"Achieve voluminous lashes with Rihanna's mascara. This mascara lifts, separates, and adds volume for a dramatic look.\", price=100, category=1,quantity=50, brand=2)\n product_3 = Product(image=\"https://images.pexels.com/photos/3685523/pexels-photo-3685523.jpeg?auto=compress&cs=tinysrgb&w=1260&h=750&dpr=1\", p_name=\"Urban Chic Eyeshadow Palette\", description=\"Explore a variety of shades with this eyeshadow palette. From subtle neutrals to bold colors, create endless eye looks.\", price=250, category=3, quantity=200 ,brand=1)\n product_4 = Product(image=\"https://images.pexels.com/photos/2661256/pexels-photo-2661256.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Luxury Makeup Brush Set\", description=\"Upgrade your makeup routine with this luxurious brush set. The soft bristles and ergonomic handles ensure a flawless application.\", price=150, category=4,quantity=34, brand=2)\n product_5 = Product(image=\"https://images.pexels.com/photos/6417915/pexels-photo-6417915.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Maybelline Lip Gloss - Pink Delight\", description=\"Shiny and moisturizing lip gloss by Maybelline. Add a pop of color and shine to your lips with this Pink Delight shade.\", price=120, category=1,quantity=70, brand=3)\n product_6 = Product(image=\"https://images.pexels.com/photos/5849420/pexels-photo-5849420.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"MAC Blush Brush\", description=\"Sculpt and define your cheeks with this MAC blush brush. Soft and angled bristles make application easy and precise.\", price=50, category=4,quantity=390 , brand=4)\n product_7 = Product(image=\"https://images.pexels.com/photos/7797740/pexels-photo-7797740.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Maybelline Lipstick - Coral Crush\", description=\"Vibrant coral lipstick by Maybelline. Provides a creamy texture and bold color payoff.\", price=180, category=1, quantity=100 , brand=3)\n product_8 = Product(image=\"https://images.pexels.com/photos/6476122/pexels-photo-6476122.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"MAC Foundation - Natural Glow\", description=\"Lightweight foundation for a natural glow. Blends seamlessly and provides all-day coverage.\", price=280, category=2,quantity=94, brand=4)\n product_9 = Product(image=\"https://images.pexels.com/photos/3997378/pexels-photo-3997378.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Rihanna Face Cream - Hydrating Moisture\", description=\"Hydrating face cream by Rihanna. Infused with moisturizing ingredients for soft and supple skin.\", price=220, category=2,quantity=60, brand=2)\n product_10 = Product(image=\"https://images.pexels.com/photos/5871834/pexels-photo-5871834.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"MAC Concealer - Full Coverage\", description=\"Full coverage concealer by MAC. Conceals imperfections and brightens the under-eye area.\", price=150, category=2, quantity=20, brand=4)\n product_11 = Product(image=\"https://images.pexels.com/photos/3115708/pexels-photo-3115708.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Maybelline Eyeshadow Palette - Bold Hues\", description=\"Dive into a world of bold hues with this Maybelline eyeshadow palette. Create vibrant and daring eye looks with a mix of matte and shimmer shades.\", price=280, category=3,quantity=150, brand=3)\n product_12 = Product(image=\"https://images.pexels.com/photos/279480/pexels-photo-279480.jpeg?auto=compress&cs=tinysrgb&w=1600\", p_name=\"Huddah Makeup Brush Set\", description=\"High-quality makeup brush set by Huddah Cosmetics. Includes brushes for eyes, face, and lips.\", price=200, category=4,quantity=120, brand=1)\n\n # Add products to the session and commit\n db.session.add_all([\n product_1, product_2, product_3, product_4, product_5, product_6, product_7, product_8, product_9, product_10, product_11, product_12\n ])\n db.session.commit()\n\n print(\"🦸‍♀️ Seeding Invoices with Faker...\")\n for _ in range(50): # Generate 3 fake invoices\n user = User.query.order_by(User.id).first()\n product = Product.query.order_by(Product.id).first()\n invoice = Invoice(\n users=user,\n products=product,\n quantity=fake.random_int(min=1, max=20),\n cost=fake.random_int(min=10, max=5000)\n )\n db.session.add(invoice)\n\n print(\"🦸‍♀️ Seeding Invoice_Products with Faker...\")\n products = Product.query.all()\n invoices = Invoice.query.all()\n for _ in range(50): # Generate 6 fake invoice products\n product = fake.random_element(products)\n invoice = fake.random_element(invoices)\n invoice_product = InvoiceProducts(\n product_rl=product,\n invoice_rl=invoice\n )\n db.session.add(invoice_product)\n\n db.session.commit()\n\nif __name__ == \"__main__\":\n with app.app_context():\n db.app = app # Bind the app to the current SQLAlchemy instance\n delete_data()\n db.session.commit()\n seed_data()\n db.session.commit()\n \n print(\"🦸‍♀️ Done seeding!\")\n","repo_name":"Bii-teki/bloom-beauty-backend","sub_path":"seed.py","file_name":"seed.py","file_ext":"py","file_size_in_byte":7966,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"19547927799","text":"def solution(n, wires):\n answer = n\n for wire in wires:\n tmp = wires[:]\n tmp.remove(wire)\n num = count_num(tmp, wire[0]) + 1\n if answer > abs(n - 2 * num):\n answer = abs(n - 2 * num)\n return answer\n\ndef count_num(arr, num):\n tmp = []\n left = []\n answer = 0\n for i in arr:\n if i[0] == num:\n left.append(i[1])\n elif i[1] == num:\n left.append(i[0])\n else:\n tmp.append(i)\n if left == []:\n return 0\n else:\n for j in left:\n answer += count_num(tmp, j)\n return answer + len(left)\n","repo_name":"MinHoon-LEE/Ps_Sql","sub_path":"Programmers/Algorithm/Python/86971/86971.py","file_name":"86971.py","file_ext":"py","file_size_in_byte":625,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"35319890071","text":"\"\"\"DPP Retriever\"\"\"\n\nfrom openicl import DatasetReader\nfrom openicl.icl_retriever.icl_topk_retriever import TopkRetriever\nfrom openicl.utils.logging import get_logger\nfrom typing import Optional\nimport tqdm\nimport numpy as np\nimport math\nfrom accelerate import Accelerator\n\nlogger = get_logger(__name__)\n\n\nclass DPPRetriever(TopkRetriever):\n \"\"\"DPP In-context Learning Retriever Class\n Class of DPP Retriever.\n Two-stage DPP is used, where first stage is to get results of TopK to reduce candidate sets\n chechout https://arxiv.org/abs/2302.05698 for details.\n \n Attributes:\n dataset_reader (:obj:`DatasetReader`): An instance of the :obj:`DatasetReader` class.\n ice_separator (:obj:`str`, optional): A string that separates each in-context example.\n ice_eos_token (:obj:`str`, optional): A string that is added to the end of in-context examples.\n prompt_eos_token (:obj:`str`, optional): A string that is added to the end of the prompt.\n ice_num (:obj:`int`, optional): The number of data in the in-context examples.\n index_split (:obj:`str`, optional): A string for the index dataset name. The index dataset is used to select data for in-context examples. Defaults to ``train``.\n test_split (:obj:`str`, optional): A string for the generation dataset name. The test dataset is used to generate prompts for each data. Defaults to ``test``.\n index_ds (:obj:`Dataset`): The index dataset. Used to select data for in-context examples.\n test_ds (:obj:`Dataset`): The test dataset. Used to generate prompts for each data.\n accelerator (:obj:`Accelerator`, optional): An instance of the :obj:`Accelerator` class, used for multiprocessing.\n batch_size (:obj:`int`, optional): Batch size for the :obj:`DataLoader`. \n model (:obj:`SentenceTransformer`): An instance of :obj:`SentenceTransformer` class, used to calculate embeddings.\n tokenizer (:obj:`AutoTokenizer`): Tokenizer for :obj:`model`.\n index (:obj:`IndexIDMap`): Index generated with FAISS.\n seed (:obj:`int`, optional): Seed for the random number generator. (:obj:`random_state` in :obj:`sample_exact_k_dpp` method)\n scale_factor (:obj:`float`, optional): A factor when gets the kernel.\n \"\"\"\n model = None\n\n def __init__(self,\n dataset_reader: DatasetReader,\n ice_separator: Optional[str] = '\\n',\n ice_eos_token: Optional[str] = '\\n',\n prompt_eos_token: Optional[str] = '',\n sentence_transformers_model_name: Optional[str] = 'all-mpnet-base-v2',\n ice_num: Optional[int] = 1,\n candidate_num: Optional[int] = 1,\n index_split: Optional[str] = 'train',\n test_split: Optional[str] = 'test',\n tokenizer_name: Optional[str] = 'gpt2-xl',\n batch_size: Optional[int] = 1,\n accelerator: Optional[Accelerator] = None,\n seed: Optional[int] = 1,\n scale_factor: Optional[float] = 0.1\n ) -> None:\n super().__init__(dataset_reader, ice_separator, ice_eos_token, prompt_eos_token,\n sentence_transformers_model_name, ice_num, index_split, test_split, tokenizer_name, batch_size,\n accelerator)\n self.candidate_num = candidate_num\n self.seed = seed\n self.scale_factor = scale_factor\n\n def dpp_search(self):\n res_list = self.forward(self.dataloader, process_bar=True, information=\"Embedding test set...\")\n rtr_idx_list = [[] for _ in range(len(res_list))]\n logger.info(\"Retrieving data for test set...\")\n for entry in tqdm.tqdm(res_list, disable=not self.is_main_process):\n idx = entry['metadata']['id']\n\n # get TopK results\n embed = np.expand_dims(entry['embed'], axis=0)\n near_ids = np.array(self.index.search(embed, self.candidate_num)[1][0].tolist())\n\n # DPP stage\n near_reps, rel_scores, kernel_matrix = self.get_kernel(embed, near_ids.tolist())\n\n # MAP inference\n samples_ids = fast_map_dpp(kernel_matrix, self.ice_num)\n\n # ordered by relevance score\n samples_scores = np.array([rel_scores[i] for i in samples_ids])\n samples_ids = samples_ids[(-samples_scores).argsort()].tolist()\n rtr_sub_list = [int(near_ids[i]) for i in samples_ids]\n\n rtr_idx_list[idx] = rtr_sub_list\n\n return rtr_idx_list\n\n def retrieve(self):\n return self.dpp_search()\n\n def get_kernel(self, embed, candidates):\n near_reps = np.stack([self.index.index.reconstruct(i) for i in candidates], axis=0)\n # normalize first\n embed = embed / np.linalg.norm(embed)\n near_reps = near_reps / np.linalg.norm(near_reps, keepdims=True, axis=1)\n\n # to make kernel-matrix non-negative\n rel_scores = np.matmul(embed, near_reps.T)[0]\n rel_scores = (rel_scores + 1) / 2\n\n # to prevent overflow error\n rel_scores -= rel_scores.max()\n\n # to balance relevance and diversity\n rel_scores = np.exp(rel_scores / (2 * self.scale_factor))\n\n # to make kernel-matrix non-negative\n sim_matrix = np.matmul(near_reps, near_reps.T)\n sim_matrix = (sim_matrix + 1) / 2\n\n kernel_matrix = rel_scores[None] * sim_matrix * rel_scores[:, None]\n return near_reps, rel_scores, kernel_matrix\n\n\ndef fast_map_dpp(kernel_matrix, max_length):\n \"\"\"\n fast implementation of the greedy algorithm\n reference: https://github.com/laming-chen/fast-map-dpp/blob/master/dpp_test.py\n paper: Fast Greedy MAP Inference for Determinantal Point Process to Improve Recommendation Diversity\n \"\"\"\n item_size = kernel_matrix.shape[0]\n cis = np.zeros((max_length, item_size))\n di2s = np.copy(np.diag(kernel_matrix))\n selected_items = list()\n selected_item = np.argmax(di2s)\n selected_items.append(int(selected_item))\n while len(selected_items) < max_length:\n k = len(selected_items) - 1\n ci_optimal = cis[:k, selected_item]\n di_optimal = math.sqrt(di2s[selected_item])\n elements = kernel_matrix[selected_item, :]\n eis = (elements - np.dot(ci_optimal, cis[:k, :])) / di_optimal\n cis[k, :] = eis\n di2s -= np.square(eis)\n selected_item = np.argmax(di2s)\n selected_items.append(int(selected_item))\n return selected_items\n","repo_name":"Shark-NLP/OpenICL","sub_path":"openicl/icl_retriever/icl_dpp_retriever.py","file_name":"icl_dpp_retriever.py","file_ext":"py","file_size_in_byte":6538,"program_lang":"python","lang":"en","doc_type":"code","stars":460,"dataset":"github-code","pt":"7"} +{"seq_id":"60682937","text":"from sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.model_selection import train_test_split\nimport pandas as pd\nimport pickle\n\ndf = pd.read_csv('dataset/imdb.txt', names=['sentence', 'label'], sep='\\t')\nsentences=df.sentence\ny = df.label\nsentences_train, sentences_test, y_train, y_test = train_test_split(sentences, y, test_size=0.25, random_state=1000)\nvectorizer = CountVectorizer()\nvectorizer.fit(sentences_train)\nX_train = vectorizer.transform(sentences_train)\nX_test = vectorizer.transform(sentences_test)\nmodel = LogisticRegression()\nmodel.fit(X_train, y_train)\nscore = model.score(X_test, y_test)\nprint(\"Accuracy:{}\".format(score) )\n\n#save model\npickle.dump(vectorizer, open('vectorizer','wb'))\npickle.dump(model, open('model', 'wb'))\n\n#predict sample\nsample=[\"Movie was awesome\",\"Movie was bad\"]\ntest=vectorizer.transform(sample)\nres=model.predict(test)\nfor i in res:\n if(abs(i)>abs(1-i)):\n print(\"pos\")\n else:\n print(\"neg\")\n","repo_name":"AMAN2202/Movie_Review_Classification_using_Neural_Network_and_API","sub_path":"nlp/modelling/nlp/logistic.py","file_name":"logistic.py","file_ext":"py","file_size_in_byte":1028,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"16786408822","text":"import numpy as np\n\nmaxID = 0\nids = np.arange(1033)\nseats = []\nwith open(\"data.txt\", \"r\") as file:\n for line in file.read().split(\"\\n\"):\n seat = sum([int(i) * 2 ** j for j, i in enumerate(line.replace(\"L\", \"0\").replace(\"R\", \"1\").replace(\"F\", \"0\").replace(\"B\", \"1\")[::-1])])\n seats.append(seat)\n if seat > maxID:\n maxID = seat\nprint(maxID)\nfree = np.asarray([i for i in ids if i not in seats])\nyour = free[np.where(np.gradient(free) != 1)[0][1]]\nprint(your)","repo_name":"Haakooto/AdventOfCode","sub_path":"2020/05/binary_boarding.py","file_name":"binary_boarding.py","file_ext":"py","file_size_in_byte":491,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"32961591276","text":"# Dependencies\nimport datetime as dt\nimport numpy as np\nimport pandas as pd\n\nimport sqlalchemy\nfrom sqlalchemy.ext.automap import automap_base\nfrom sqlalchemy.orm import Session\nfrom sqlalchemy import create_engine, func\n\nfrom flask import Flask, jsonify\n\n# Set up database engine\nengine = create_engine(\"sqlite:///hawaii.sqlite\")\n\n# Reflect database into classes\nBase = automap_base()\nBase.prepare(engine, reflect=True)\n\n# Create a variable for each class\nMeasurement = Base.classes.measurement\nStation = Base.classes.station\n\n# Create session link from Python to database\nsession = Session(engine)\n\n# Define flask app\napp = Flask(__name__)\n\n# Define welcome route\n@app.route(\"/\")\n\n# Create function for routing information\ndef welcome():\n return(\n '''\n Welcome to the Climate Analysis API!\n Available Routes:\n /api/v1.0/precipitation\n /api/v1.0/stations\n /api/v1.0/tobs\n /api/v1.0/temp/start/end\n ''')\n\n# Define precipitation route\n@app.route(\"/api/v1.0/precipitation\")\n\n# Create precipitation function\ndef precipitation():\n # Calculate date one year ago from date in database\n prev_year = dt.date(2017, 8, 23) - dt.timedelta(days=365)\n # Query for date/precipitation for previous year\n precipitation = session.query(Measurement.date, Measurement.prcp).filter(Measurement.date >= prev_year).all()\n # Use jsonify() to romat results into JSON structured file (dictionary)\n precip = {date: prcp for date, prcp in precipitation}\n return jsonify(precip)\n\n# Define station route\n@app.route(\"/api/v1.0/stations\")\n\n# Create station function\ndef stations():\n # Collect all stations\n results = session.query(Station.station).all()\n # Unravel results into one-dimensional array, then a list, then jsonify\n stations = list(np.ravel(results))\n return jsonify(stations=stations)\n\n# Define temperature route\n@app.route(\"/api/v1.0/tobs\")\n\n# Create temp function\ndef temp_monthly():\n # Calculate date one year ago from date in database\n prev_year = dt.date(2017, 8, 23) - dt.timedelta(days=365)\n # Query primary station for all temps from previous year\n results = session.query(Measurement.tobs).filter(Measurement.station == 'USC00519281').filter(Measurement.date >= prev_year).all()\n # Unravel results into one-dimensional array, then a list\n temps = list(np.ravel(results))\n # jsonify results\n return jsonify(temps=temps)\n\n# Define status route\n@app.route(\"/api.v1.0/temp/\")\n@app.route(\"/api.v1.0/temp//\")\n\n# Create stats function including start and end parameters\ndef stats(start=None, end=None):\n sel = [func.min(Measurement.tobs), func.avg(Measurement.tobs), func.max(Measurement.tobs)]\n\n if not end:\n results = session.query(*sel).\\\n filter(Measurement.date >= start).all()\n temps = list(np.ravel(results))\n return jsonify(temps)\n\n results = session.query(*sel).\\\n filter(Measurement.date >= start).\\\n filter(Measurement.date <= end).all()\n temps = list(np.ravel(results))\n return jsonify(temps)","repo_name":"HollyC13/surfs_up","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3055,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"28482076905","text":"import os\nimport zipfile\nfrom pathlib import Path\nfrom unittest import TestCase\nfrom unittest.mock import patch, MagicMock\n\nfrom drover.io import ArchiveMapping, get_digest, get_relative_file_names, write_archive\n\n\nclass TestGetDigest(TestCase):\n def test_zero_file_digest_is_none(self):\n assert get_digest(tuple()) is None\n\n\nclass TestGetRelativeFileNames(TestCase):\n def test_without_excludes(self):\n expected_source_path = '/'\n expected_relative_file_names = set((\n Path('file_0'),\n Path('path_a/file_a_0'),\n Path('path_a/file_a_1'),\n Path('path_b/file_b_0')))\n expected_walk = (\n ('/', ('path_a', 'path_b',), ('file_0',)),\n ('/path_a', (), ('file_a_0', 'file_a_1')),\n ('/path_b', (), ('file_b_0',)),\n )\n with patch.object(os, 'walk', return_value=expected_walk) as mock_walk:\n names = set(get_relative_file_names(expected_source_path))\n mock_walk.assert_called_once_with(str(expected_source_path))\n assert names == expected_relative_file_names\n\n\nclass TestWriteArchive(TestCase):\n def test_write_empty_archive(self):\n expected_archive_file_name = Path('archive.zip')\n mock_zip_file = MagicMock(spec=zipfile.ZipFile)\n with patch.object(zipfile, 'ZipFile') as mock_zip_file_cls:\n mock_zip_file_cls.return_value.__enter__.return_value = mock_zip_file\n write_archive(expected_archive_file_name, [])\n mock_zip_file_cls.assert_called_once_with(\n expected_archive_file_name, 'w',\n compression=zipfile.ZIP_DEFLATED,\n compresslevel=9)\n mock_zip_file.write.assert_not_called()\n\n def test_write_non_empty_archive(self):\n expected_archive_file_name = Path('archive.zip')\n expected_archive_mappings = [\n ArchiveMapping(source_file_name=Path('source/a'), archive_file_name=Path('archive/a')),\n ArchiveMapping(source_file_name=Path('source/b'), archive_file_name=Path('archive/b')),\n ]\n mock_zip_file = MagicMock(spec=zipfile.ZipFile)\n with patch.object(zipfile, 'ZipFile') as mock_zip_file_cls:\n mock_zip_file_cls.return_value.__enter__.return_value = mock_zip_file\n write_archive(expected_archive_file_name, expected_archive_mappings)\n mock_zip_file_cls.assert_called_once_with(\n expected_archive_file_name, 'w',\n compression=zipfile.ZIP_DEFLATED,\n compresslevel=9)\n for expected_archive_mapping in expected_archive_mappings:\n mock_zip_file.write.assert_any_call(\n filename=expected_archive_mapping.source_file_name,\n arcname=expected_archive_mapping.archive_file_name)\n","repo_name":"jwilges/drover","sub_path":"tests/test_io.py","file_name":"test_io.py","file_ext":"py","file_size_in_byte":2818,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"7"} +{"seq_id":"43661620568","text":"import os\n\nfrom corcym.settings import PATH_P\n\npath = PATH_P\nfiles = os.listdir(path)\n\nif files:\n for f in files:\n rm_path = PATH_P + f\"{f}\"\n os.remove(rm_path)\n","repo_name":"talhaumer/corcym","sub_path":"seeding-scripts/removefiles.py","file_name":"removefiles.py","file_ext":"py","file_size_in_byte":178,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"24129543870","text":"import re\nimport unicodedata\n\n\ndef _format_size(value, digits, unit):\n if digits > 0:\n return \"{{:.{}f}}{}\".format(digits, unit).format(value)\n else:\n return \"{{:d}}{}\".format(unit).format(value)\n\n\ndef format_size(bytes_, digits=1):\n if bytes_ < 1024:\n return _format_size(bytes_, digits, \"B\")\n\n kilo = bytes_ / 1024\n if kilo < 1024:\n return _format_size(kilo, digits, \"kB\")\n\n mega = kilo / 1024\n if mega < 1024:\n return _format_size(mega, digits, \"MB\")\n\n return _format_size(mega / 1024, digits, \"GB\")\n\n\ndef format_duration(total_seconds):\n total_seconds = int(total_seconds)\n hours = total_seconds // 3600\n remainder = total_seconds % 3600\n minutes = remainder // 60\n seconds = total_seconds % 60\n\n if hours:\n return \"{} h {} min\".format(hours, minutes)\n\n if minutes:\n return \"{} min {} sec\".format(minutes, seconds)\n\n return \"{} sec\".format(seconds)\n\n\ndef read_int(msg, min, max, default):\n msg = msg + \" [default {}]: \".format(default)\n\n while True:\n try:\n val = input(msg)\n if not val:\n return default\n if min <= int(val) <= max:\n return int(val)\n except ValueError:\n pass\n\n\ndef slugify(value):\n re_pattern = re.compile(r'[^\\w\\s-]', flags=re.U)\n re_spaces = re.compile(r'[-\\s]+', flags=re.U)\n value = str(value)\n value = unicodedata.normalize('NFKD', value).encode('ascii', 'ignore').decode('ascii')\n value = re_pattern.sub('', value).strip().lower()\n return re_spaces.sub('-', value)\n","repo_name":"GaspardIV/twitchautocut","sub_path":"twitchdl/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1599,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"21356390645","text":"from django import template\nfrom django.template.defaulttags import register\nfrom todo.models import Task\n\n\nclass TaskPriorityNode(template.Node):\n def __init__(self, priority_level):\n self.priority_level = template.Variable(priority_level)\n\n def render(self, context):\n try:\n priority_level_value = self.priority_level.resolve(context)\n\n return \"\".format(Task.priority_label(priority_level_value))\n except template.VariableDoesNotExist:\n return ''\n\n@register.tag(name=\"task_priority\")\ndef do_task_priority(parser, token):\n try:\n tag_name, priority_level = token.split_contents()\n except ValueError:\n raise template.TemplateSyntaxError(\n \"%r tag requires a single argument\" % token.contents.split()[0]\n )\n\n return TaskPriorityNode(priority_level)","repo_name":"dwolosowicz/django-todo-app","sub_path":"project/todo/templatetags/task_priority.py","file_name":"task_priority.py","file_ext":"py","file_size_in_byte":893,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"1797467836","text":"import qgis, qgis.utils, os, datetime, PIL, platform\n\nif platform.system() == \"Windows\":\n from PIL import ImageGrab\nfrom PIL.ImageQt import ImageQt\n\nfrom qgis.PyQt import uic, QtCore, QtGui\nfrom qgis.PyQt.QtWidgets import QWidget, QMainWindow, QSpinBox, QAction, QDialog, QFrame\n\nfrom Lamia.qgisiface.iface.qgiswidget.tools.lamia_abstractformtool import (\n AbstractLamiaFormTool,\n)\nfrom .lamia_form_pictureviewer import PictureViewer\n\nbase3 = QtCore.QObject()\n\n\nclass BaseSketchTool(AbstractLamiaFormTool):\n\n PREPROTOOLNAME = \"sketch\"\n DBASETABLENAME = \"media\"\n LOADFIRST = False\n\n tooltreewidgetCAT = QtCore.QCoreApplication.translate(\"base3\", \"Resources\")\n tooltreewidgetSUBCAT = QtCore.QCoreApplication.translate(\"base3\", \"Sketches\")\n\n tooltreewidgetICONPATH = os.path.join(\n os.path.dirname(__file__), \"lamia_form_sketch_icon.png\"\n )\n\n tempparentjoin = {}\n linkdict = {\n \"colparent\": \"id_object\",\n \"colthistable\": \"id_resource\",\n \"tctable\": \"tcobjectresource\",\n \"tctablecolparent\": \"lid_object\",\n \"tctablecolthistable\": \"lid_resource\",\n }\n for tablename in [\n \"deficiency\",\n \"observation\",\n \"node\",\n \"edge\",\n \"surface\",\n \"equipment\",\n \"facility\",\n ]:\n tempparentjoin[tablename] = linkdict\n PARENTJOIN = tempparentjoin\n\n TABLEFILTERFIELD = {\"typemedia\": \"CRO\"}\n GEOMETRYSKIP = True\n\n def __init__(self, **kwargs):\n super(BaseSketchTool, self).__init__(**kwargs)\n\n def initMainToolWidget(self):\n # ****************************************************************************************\n # userui\n self.toolwidgetmain = UserUI()\n self.formtoolwidgetconfdictmain = {\n \"media\": {\"linkfield\": \"id_media\", \"widgets\": {}},\n \"object\": {\"linkfield\": \"id_object\", \"widgets\": {}},\n \"resource\": {\"linkfield\": \"id_ressource\", \"widgets\": {}},\n }\n\n # self.groupBox_geom.setParent(None)\n # self.frame_editing.setVisible(False)\n self.toolwidgetmain.stackedWidget.setCurrentIndex(1)\n\n self.toolwidgetmain.pushButton_open.clicked.connect(self.openFile)\n self.toolwidgetmain.pushButton_edit.clicked.connect(self.editPhoto)\n self.toolwidgetmain.pushButton_getfromclipboard.clicked.connect(self.pasteImage)\n self.editorwindow = ScribbleMainWindow(parentwdg=self)\n self.photowdg = PictureViewer()\n self.toolwidgetmain.frame_cr.layout().addWidget(self.photowdg)\n\n def postSelectFeature(self):\n\n if self.currentFeaturePK is None: # first time\n datecreation = str(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"))\n self.formutils.applyResultDict(\n {\"resource\": {\"datetimeresource\": datecreation}}\n )\n self.editorwindow.reinitSize()\n self.editorwindow.clear()\n self.photowdg.clear()\n\n else:\n sql = \"SELECT file FROM media_qgis WHERE pk_media = \" + str(\n self.currentFeaturePK\n )\n file = self.dbase.query(sql)[0][0]\n if (\n file is not None\n and file != \"\"\n and os.path.isfile(self.dbase.completePathOfFile(file))\n ):\n self.editorwindow.openImage(self.dbase.completePathOfFile(file))\n self.formutils.showImageinLabelWidget(\n self.photowdg, self.dbase.completePathOfFile(file)\n )\n else:\n self.editorwindow.clear()\n self.photowdg.clear()\n\n def pasteImage(self):\n if platform.system() == \"Windows\":\n pilimage = PIL.ImageGrab.grabclipboard()\n if pilimage is not None:\n im = ImageQt(pilimage)\n self.editorwindow.setImage(im)\n self.photowdg.clear()\n self.photowdg.setPixmap(im)\n self.editPhoto()\n\n def editPhoto(self):\n if qgis.utils.iface is not None:\n self.editorwindow.show()\n else:\n self.editorwindow.setWindowModality(QtCore.Qt.ApplicationModal)\n self.editorwindow.show()\n\n def openFile(self):\n if self.currentFeaturePK is not None:\n sql = \"SELECT file FROM media_qgis WHERE pk_media = \" + str(\n self.currentFeaturePK\n )\n query = self.dbase.query(sql)\n # result = [row[0] for row in query]\n resultfile = query[0][0]\n if os.path.isfile(self.dbase.completePathOfFile(resultfile)):\n filepath = self.dbase.completePathOfFile(resultfile)\n os.startfile(filepath)\n\n \"\"\"\n def createParentFeature(self):\n pkobjet = self.dbase.createNewObjet()\n\n if False:\n # lastrevision = self.dbase.maxrevision\n datetimecreation = str(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"))\n lastobjetid = self.dbase.getLastId('Objet') + 1\n sql = \"INSERT INTO Objet (id_objet, lpk_revision_begin, datetimecreation ) \"\n sql += \"VALUES(\" + str(lastobjetid ) + \",\" + str(self.dbase.maxrevision) + \",'\" + datetimecreation + \"');\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n pkobjet = self.dbase.getLastRowId('Objet')\n\n lastressourceid = self.dbase.getLastId('Ressource') + 1\n sql = \"INSERT INTO Ressource (id_ressource, lpk_objet) \"\n sql += \"VALUES(\" + str(lastressourceid) + \",\" + str(pkobjet) + \");\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n pkres = self.dbase.getLastRowId('Ressource')\n\n\n\n\n\n pkphoto = self.currentFeaturePK\n lastidphoto = self.dbase.getLastId('Photo') + 1\n datecreation = datetime.datetime.now().strftime(\"%Y-%m-%d\")\n datetimecreation = str(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"))\n\n\n fileimage = os.path.join('.', self.dbasetablename, ''.join(datecreation.split('-')),\n str(lastidphoto) + '_croquis.png')\n if not os.path.exists(os.path.dirname(self.dbase.completePathOfFile(fileimage))):\n os.makedirs(os.path.dirname(self.dbase.completePathOfFile(fileimage)))\n self.editorwindow.saveImage(self.dbase.completePathOfFile(fileimage))\n\n\n sql = \"UPDATE Photo SET id_photo = \" + str(lastidphoto) + \",\"\n sql += \"lpk_ressource = \" + str(pkres)\n sql += \", typephoto = 'CRO' \"\n sql += \" WHERE pk_photo = \" + str(pkphoto) + \";\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n\n\n sql = \"UPDATE Ressource SET file = '\" + fileimage + \"', datetimeressource = '\" + datetimecreation + \"'\"\n sql += \" WHERE pk_ressource = \" + str(pkres) + \";\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n\n if self.parentWidget is not None and self.parentWidget.currentFeature is not None:\n # self.linkagespec = {'Tcobjetressource'\n if 'Tcobjetressource' in self.linkagespec.keys():\n #get parent id_objet\n sql = \" SELECT id_objet FROM \" + self.parentWidget.dbasetablename.lower() + \"_qgis\"\n sql += \" WHERE pk_\" + self.parentWidget.dbasetablename.lower() + \" = \" + str(self.parentWidget.currentFeaturePK)\n currentparentlinkfield = self.dbase.query(sql)[0][0]\n\n #currentparentlinkfield = self.parentWidget.currentFeature['id_objet']\n sql = \"INSERT INTO Tcobjetressource(lpk_revision_begin, lid_objet, lid_ressource) \"\n sql += \" VALUES(\" + str(self.dbase.maxrevision) + \",\" + str(currentparentlinkfield) + ',' + str(lastressourceid) + \")\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n\n\n if False:\n #lastrevision = self.dbase.getLastPk('Revision')\n lastrevision = self.dbase.maxrevision\n datecreation = QtCore.QDate.fromString(str(datetime.date.today()), 'yyyy-MM-dd').toString('yyyy-MM-dd')\n lastobjetid = self.dbase.getLastId('Objet') + 1\n sql = \"INSERT INTO Objet (id_objet, id_revisionbegin, datecreation ) \"\n sql += \"VALUES(\" + str(lastobjetid ) + \",\" + str(lastrevision) + \",'\" + datecreation + \"');\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n #idobjet = self.dbase.getLastRowId('Objet')\n\n\n lastressourceid = self.dbase.getLastId('Ressource') + 1\n sql = \"INSERT INTO Ressource (id_ressource, id_revisionbegin, id_objet) \"\n sql += \"VALUES(\" + str(lastressourceid) + \",\" + str(lastrevision) + \",\" + str(lastobjetid) + \");\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n lastressourcepk = self.dbase.getLastRowId('Ressource')\n\n\n pkcroquis = self.currentFeature.id()\n lastidcroquis = self.dbase.getLastId('Photo') + 1\n\n fileimage = os.path.join('.', self.dbasetablename, ''.join(datecreation.split('-')),\n str(lastidcroquis) + '_croquis.png')\n if not os.path.exists(os.path.dirname(self.dbase.completePathOfFile(fileimage))):\n os.makedirs(os.path.dirname(self.dbase.completePathOfFile(fileimage)))\n self.editorwindow.saveImage(self.dbase.completePathOfFile(fileimage))\n\n\n\n sql = \"UPDATE Photo SET id_objet = \" + str(lastobjetid) + \",\"\n sql += \"id_ressource = \" + str(lastressourceid) + \",\"\n sql += \"id_photo = \" + str(lastidcroquis) + \",\"\n sql += \"id_revisionbegin = \" + str(lastrevision) + \",\"\n sql += \"typephoto = 'CRO' \"\n sql += \" WHERE pk_photo = \" + str(pkcroquis) + \";\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n\n sql = \"UPDATE Ressource SET file = '\" + fileimage + \"', dateressource = '\" + datecreation + \"'\"\n sql += \" WHERE pk_ressource = \" + str( lastressourcepk) + \";\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n\n\n\n if self.parentWidget is not None and self.parentWidget.currentFeature is not None:\n currentparentlinkfield = self.parentWidget.currentFeature['id_objet']\n sql = \"INSERT INTO Tcobjetressource(id_tcobjet, id_tcressource,id_revisionbegin) \"\n sql += \" VALUES(\" + str(currentparentlinkfield) + \", \" + str(lastressourceid) + \",\" + str(lastrevision) + \");\"\n query = self.dbase.query(sql)\n self.dbase.commit()\n \"\"\"\n\n def widgetClicked(self, **kwargs):\n super().widgetClicked(**kwargs)\n if self.currentFeaturePK is None:\n self.photowdg.clear()\n\n def postSaveFeature(self, savedfeaturepk=None):\n\n if self.currentFeaturePK is None: # first creation\n idphoto, pkres = self.dbase.getValuesFromPk(\n self.DBASETABLENAME.lower() + \"_qgis\",\n [\"id_\" + self.DBASETABLENAME.lower(), \"pk_resource\"],\n savedfeaturepk,\n )\n datecreation = datetime.datetime.now().strftime(\"%Y-%m-%d\")\n datetimecreation = str(\n datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n )\n\n # sql = \"UPDATE media SET mediatype = 'CRO' WHERE pk_photo = {}\".format(str(savedfeaturepk))\n # query = self.dbase.query(sql)\n\n fileimage = os.path.join(\n \".\",\n self.DBASETABLENAME,\n \"\".join(datecreation.split(\"-\")),\n str(idphoto) + \"_sketch.png\",\n )\n if not os.path.exists(\n os.path.dirname(self.dbase.completePathOfFile(fileimage))\n ):\n os.makedirs(os.path.dirname(self.dbase.completePathOfFile(fileimage)))\n\n sql = (\n \"UPDATE resource SET file = '\"\n + fileimage\n + \"', datetimeresource = '\"\n + datetimecreation\n + \"'\"\n )\n sql += \" WHERE pk_resource = \" + str(pkres)\n query = self.dbase.query(sql)\n\n else:\n fileimage = self.dbase.getValuesFromPk(\n self.DBASETABLENAME + \"_qgis\", \"file\", self.currentFeaturePK\n )\n\n self.editorwindow.saveImage(self.dbase.completePathOfFile(fileimage))\n\n\nclass UserUI(QWidget):\n def __init__(self, parent=None):\n super(UserUI, self).__init__(parent=parent)\n uipath = os.path.join(os.path.dirname(__file__), \"lamia_form_media_ui.ui\")\n uic.loadUi(uipath, self)\n\n\nclass ScribbleMainWindow(QMainWindow):\n def __init__(self, parentwdg=None, parent=None):\n super(ScribbleMainWindow, self).__init__(parent=parent)\n uipath = os.path.join(\n os.path.dirname(__file__), \"lamia_form_sketch_drawingwdg_ui.ui\"\n )\n uic.loadUi(uipath, self)\n self.parentwdg = parentwdg\n self.scribbleArea = ScribbleArea(parent=self)\n self.scribbleArea.clearImage()\n self.scribbleArea.mainWindow = self # maybe not using this?\n\n self.scrollArea.setWidget(self.scribbleArea)\n\n self.toolbar = self.addToolBar(\"Color\")\n self.colorwdg = qgis.gui.QgsColorButton()\n self.colorwdg.setColor(QtCore.Qt.black)\n self.toolbar.addWidget(self.colorwdg)\n self.spinb = QSpinBox()\n self.spinb.setValue(3)\n self.toolbar.addWidget(self.spinb)\n\n self.spinb.valueChanged.connect(self.scribbleArea.setPenWidth)\n self.colorwdg.colorChanged.connect(self.scribbleArea.setPenColor)\n\n self.clearAction = QAction(\"Clear\", self)\n self.clearAction.triggered.connect(self.scribbleArea.clearImage)\n self.toolbar.addAction(self.clearAction)\n\n self.reinitAction = QAction(\"reinit\", self)\n self.reinitAction.triggered.connect(self.scribbleArea.reinit)\n self.toolbar.addAction(self.reinitAction)\n\n def reinitSize(self):\n self.scribbleArea.reinitSize()\n\n def saveImage(self, file):\n self.scribbleArea.saveImage(file, \"png\")\n\n def clear(self):\n self.scribbleArea.clearImage()\n\n def openImage(self, file):\n self.scribbleArea.openImage(file)\n\n def setImage(self, qimage):\n self.scribbleArea.setImage(qimage)\n\n def closeEvent(self, event):\n self.parentwdg.photowdg.clear()\n self.parentwdg.photowdg.setPixmap(self.scribbleArea.image)\n event.accept()\n\n\nclass ScribbleArea(QWidget):\n \"\"\"\n this scales the image but it's not good, too many refreshes really mess it up!!!\n \"\"\"\n\n def __init__(self, larg=500, haut=500, parent=None):\n super(ScribbleArea, self).__init__(parent)\n # self.setAttribute(QtCore.Qt.WA_StaticContents)\n self.setFixedSize(larg, haut)\n self.modified = False\n self.scribbling = False\n self.myPenWidth = 3\n self.myPenColor = QtCore.Qt.black\n imageSize = QtCore.QSize(larg, haut)\n # self.image = QtGui.QImage()\n self.image = QtGui.QImage(imageSize, QtGui.QImage.Format_RGB32)\n self.lastPoint = QtCore.QPoint()\n\n self.currentfilename = None\n\n def openImage(self, fileName=None):\n loadedImage = QtGui.QImage()\n if not loadedImage.load(fileName):\n return False\n self.currentfilename = fileName\n return self.setImage(loadedImage)\n\n def reinit(self):\n if self.currentfilename is not None:\n self.openImage(self.currentfilename)\n\n def reinitSize(self):\n self.setFixedSize(500, 500)\n self.mainWindow.resize(500, 500)\n imageSize = QtCore.QSize(500, 500)\n self.image = QtGui.QImage(imageSize, QtGui.QImage.Format_RGB32)\n\n def setImage(self, loadedImage):\n w = loadedImage.width()\n h = loadedImage.height()\n self.setFixedSize(w, h)\n self.mainWindow.resize(w, h)\n\n # newSize = loadedImage.size().expandedTo(self.size())\n # self.resizeImage(loadedImage, newSize)\n self.image = loadedImage\n self.modified = False\n self.update()\n return True\n\n def saveImage(self, fileName, fileFormat):\n if self.image.save(fileName, fileFormat):\n self.modified = False\n return True\n else:\n return False\n\n def setPenColor(self, newColor):\n self.myPenColor = newColor\n\n def setPenWidth(self, newWidth):\n self.myPenWidth = newWidth\n\n def clearImage(self):\n # print('celar')\n self.image.fill(QtGui.qRgb(255, 255, 255))\n self.modified = True\n self.update()\n\n def mousePressEvent(self, event):\n # print \"self.image.width() = %d\" % self.image.width()\n # print \"self.image.height() = %d\" % self.image.height()\n # print \"self.image.size() = %s\" % self.image.size()\n # print \"self.size() = %s\" % self.size()\n # print \"event.pos() = %s\" % event.pos()\n if event.button() == QtCore.Qt.LeftButton:\n self.lastPoint = event.pos()\n self.scribbling = True\n\n def mouseMoveEvent(self, event):\n if (event.buttons() & QtCore.Qt.LeftButton) and self.scribbling:\n self.drawLineTo(event.pos())\n\n def mouseReleaseEvent(self, event):\n if event.button() == QtCore.Qt.LeftButton and self.scribbling:\n self.drawLineTo(event.pos())\n self.scribbling = False\n\n def paintEvent(self, event):\n painter = QtGui.QPainter(self)\n painter.drawImage(event.rect(), self.image, event.rect())\n\n def drawLineTo(self, endPoint):\n painter = QtGui.QPainter(self.image)\n painter.setPen(\n QtGui.QPen(\n self.myPenColor,\n self.myPenWidth,\n QtCore.Qt.SolidLine,\n QtCore.Qt.RoundCap,\n QtCore.Qt.RoundJoin,\n )\n )\n painter.drawLine(self.lastPoint, endPoint)\n self.modified = True\n\n # rad = self.myPenWidth / 2 + 2\n # self.update(QtCore.QRect(self.lastPoint, endPoint).normalized().adjusted(-rad, -rad, +rad, +rad))\n self.update()\n self.lastPoint = QtCore.QPoint(endPoint)\n\n def print_(self):\n printer = QtGui.QPrinter(QtGui.QPrinter.HighResolution)\n\n printDialog = QtGui.QPrintDialog(printer, self)\n if printDialog.exec_() == QDialog.Accepted:\n painter = QtGui.QPainter(printer)\n rect = painter.viewport()\n size = self.image.size()\n size.scale(rect.size(), QtCore.Qt.KeepAspectRatio)\n painter.setViewport(rect.x(), rect.y(), size.width(), size.height())\n painter.setWindow(self.image.rect())\n painter.drawImage(0, 0, self.image)\n painter.end()\n\n def isModified(self):\n return self.modified\n\n def penColor(self):\n return self.myPenColor\n\n def penWidth(self):\n return self.myPenWidth\n","repo_name":"Artelia/Lamia","sub_path":"config/base3/qgswidgets/lamia_form_sketch.py","file_name":"lamia_form_sketch.py","file_ext":"py","file_size_in_byte":19056,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"7"} +{"seq_id":"25863504416","text":"from pdfminer.pdfdocument import PDFDocument\r\nfrom pdfminer.pdfpage import PDFPage\r\nfrom pdfminer.pdfparser import PDFParser\r\nfrom pdfminer.pdfinterp import PDFResourceManager, PDFPageInterpreter\r\nfrom pdfminer.converter import PDFPageAggregator\r\nfrom pdfminer.layout import LAParams, LTTextBox, LTTextLine, LTFigure\r\nimport os\r\nimport json\r\nfrom multiprocessing import Pool\r\n\r\ndef parse_text(layout):\r\n \"\"\"Function to recursively parse the layout tree.\"\"\"\r\n result = []\r\n if not hasattr(layout, '__iter__'):\r\n return result\r\n for lt_obj in layout:\r\n if isinstance(lt_obj, LTTextLine):\r\n bbox = lt_obj.bbox\r\n text = lt_obj.get_text().strip()\r\n if text != '':\r\n result += [(bbox, text)]\r\n else:\r\n result += parse_text(lt_obj)\r\n return result\r\n\r\n\r\ndef parse_case(case_path):\r\n \"\"\"Parse all the pdf files in the folder.\"\"\"\r\n try:\r\n result = {\r\n 'id': case_path.split('/')[-2], \r\n 'docs': {}\r\n }\r\n\r\n for name in os.listdir(case_path):\r\n if name[0] == '.' or name[-4:] != '.pdf':\r\n continue\r\n doc_id = name.split('.')[0]\r\n result['docs'][doc_id] = {'pages': {}}\r\n doc_obj = result['docs'][doc_id]\r\n\r\n path = case_path + name\r\n fp = open(path, 'rb')\r\n parser = PDFParser(fp)\r\n doc = PDFDocument(parser)\r\n rsrcmgr = PDFResourceManager()\r\n laparams = LAParams(detect_vertical=True, all_texts=True)\r\n device = PDFPageAggregator(rsrcmgr, laparams=laparams)\r\n interpreter = PDFPageInterpreter(rsrcmgr, device)\r\n\r\n for page in PDFPage.create_pages(doc):\r\n interpreter.process_page(page)\r\n layout = device.get_result()\r\n doc_obj['pages'][layout.pageid] = {\r\n 'size': (layout.width, layout.height),\r\n 'text': parse_text(layout)\r\n }\r\n # print(layout.width, layout.height)\r\n\r\n output = open(case_path + 'parsed.json', 'w')\r\n json.dump(result, output, indent=None)\r\n except:\r\n print(\"Error \" + case_path)\r\n\r\n return None\r\n\r\n\r\n\r\ndef main(base_path):\r\n case_list = []\r\n for direc in os.listdir(base_path):\r\n path = base_path + direc + '/'\r\n if not os.path.isdir(path):\r\n continue\r\n case_list.append(path)\r\n # Multiprocessing, for speed up\r\n pool = Pool(processes=8)\r\n output = pool.map(parse_case, case_list)\r\n\r\n\r\nif __name__ == '__main__':\r\n main('Content/')\r\n","repo_name":"thomas0809/GraphIE","sub_path":"sentence-level/scripts-for-visual-ie/parse_pdf.py","file_name":"parse_pdf.py","file_ext":"py","file_size_in_byte":2634,"program_lang":"python","lang":"en","doc_type":"code","stars":107,"dataset":"github-code","pt":"7"} +{"seq_id":"31404847930","text":"from PyQt5.QtCore import QObject, pyqtSignal, pyqtSlot\nfrom PyQt5.QtWidgets import QDialog\nfrom ..view.Ui_progress import Ui_progressDialog\n\n\nclass ProgressDialog(QDialog, Ui_progressDialog):\n def __init__(self, parent=None):\n super(ProgressDialog, self).__init__(parent)\n self.setupUi(self)\n self.setWindowTitle('下载')\n self.listWidget.itemClicked.connect(self.item_clicked)\n\n # 下载切换\n def item_clicked(self, item):\n # 获取当前选中的item\n item = self.listWidget.selectedItems()[0]\n if item.text() == '正在下载':\n self.stackedWidget.setCurrentIndex(1)\n if item.text() == '已完成':\n self.stackedWidget.setCurrentIndex(0)\n\n\nif __name__ == '__main__':\n import sys\n from PyQt5.QtWidgets import QApplication\n\n app = QApplication(sys.argv)\n progressDialog = ProgressDialog()\n progressDialog.show()\n sys.exit(app.exec_())\n","repo_name":"jlx001/title_","sub_path":"app/function/progress.py","file_name":"progress.py","file_ext":"py","file_size_in_byte":944,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37626403816","text":"import numpy as np\nimport pandas as pd\nimport pymorphy2\nfrom nltk import RegexpTokenizer\nimport pickle\n\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn import metrics\nfrom sklearn.metrics import accuracy_score\n\nall_feeds_file = \"all_train.csv\"\nmy_feeds_file = \"my.csv\"\n\nmorph = pymorphy2.MorphAnalyzer()\n\nall_feeds = pd.read_csv(\n all_feeds_file,\n encoding=\"utf-8\",\n usecols=[\"text\"]\n).values\n\nall_labels = pd.read_csv(\n all_feeds_file,\n encoding=\"utf-8\",\n usecols=[\"label\"]\n).values\n\nmy_feeds = pd.read_csv(\n my_feeds_file,\n encoding=\"utf-8\",\n usecols=[\"text\"]\n).values\n\nmy_labels = pd.read_csv(\n my_feeds_file,\n encoding=\"utf-8\",\n usecols=[\"label\"]\n).values\n\nY_train = []\nfor label in all_labels:\n Y_train.append(int(label[0]))\n\ny_true = []\nfor label in my_labels:\n y_true.append(int(label[0]))\n\n\ndef get_normal_form(one_word):\n return morph.parse(one_word)[0].normal_form\n\n\ntokenizer = RegexpTokenizer(r'\\w+')\n\n\n# массив всех слов в нормальной форме\ndef get_normal_form_words(feeds):\n normal_form_words = []\n for feed in feeds:\n words = tokenizer.tokenize(feed[0].lower())\n normal_form_words += list(map(get_normal_form, words))\n\n return normal_form_words\n\n\n# количество уникальных слов в массиве\ndef get_count_of_unique_words(words_array):\n return len(np.unique(words_array))\n\n\n# мешок слов\ndef get_words_bag(feeds, all_unique):\n bag = []\n for feed in feeds:\n feed_all_words = get_normal_form_words([feed])\n feed_bag = []\n for word in all_unique:\n count = feed_all_words.count(word)\n feed_bag.append(count)\n\n bag.append(feed_bag)\n\n return bag\n\n\ndef save_array_to_file(array, text):\n with open(text, 'wb') as f:\n pickle.dump(array, f)\n\n\ndef read_array_from_file(text):\n with open(text, 'rb') as f:\n return pickle.load(f)\n\n\nall_unique_words = np.unique(get_normal_form_words(all_feeds))\n\n# words_bag = get_words_bag(all_feeds, all_unique_words)\n# my_words_bag = get_words_bag(my_feeds, all_unique_words)\n\n# save_array_to_file(words_bag, \"all_bag.txt\")\n# save_array_to_file(my_words_bag, \"my_bag.txt\")\n\nwords_bag = read_array_from_file(\"all_bag.txt\")\nmy_words_bag = read_array_from_file(\"my_bag.txt\")\n\nreg_model = LogisticRegression(max_iter=100000)\nreg_model.fit(words_bag, Y_train)\n\ny_pred = reg_model.predict(my_words_bag)\n\n# метрики\nprint(accuracy_score(y_true, y_pred))\nprint(metrics.classification_report(y_true, y_pred, digits=3))\n\npositive_weights = dict(zip(all_unique_words, reg_model.coef_[2]))\nneutral_weights = dict(zip(all_unique_words, reg_model.coef_[1]))\nnegative_weights = dict(zip(all_unique_words, reg_model.coef_[0]))\n\n\ndef get_first_last_words(dict_weights):\n sorted_list = list({k: v for k, v in sorted(dict_weights.items(), key=lambda item: item[1], reverse=True)})\n first_10 = sorted_list[0:10]\n reversed_sorted_list = sorted_list[::-1]\n last_10 = reversed_sorted_list[0:10]\n print(first_10, last_10)\n\n\n# первые 10 и последние 10 слов для каждого класса\nget_first_last_words(positive_weights)\nget_first_last_words(negative_weights)\nget_first_last_words(negative_weights)\n","repo_name":"iapolya/NLP_ITIS_3","sub_path":"FilmReviews/regression.py","file_name":"regression.py","file_ext":"py","file_size_in_byte":3293,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"22442279033","text":"\"\"\"\n Son una secuencia de números enteros los cuales siguen una regla sencilla:\n\nel primer elemento de la secuencia es igual a uno (Fib1 = 1)\nel segundo elemento también es igual a uno (Fib2 = 1)\ncada número después de ellos son la suman de los dos números anteriores (Fibi = Fibi-1 + Fibi-2)\nAquí están algunos de los primeros números en la serie Fibonacci:\n\nfib_1 = 1 fib_2 = 1 fib_3 = 1 + 1 = 2 fib_4 = 1 + 2 = 3 fib_5 = 2 + 3 = 5 fib_6 = 3 + 5 = 8 fib_7 = 5 + 8 = 13\n \"\"\"\n \ndef fib(n):\n if n < 1:\n return None\n if n < 3:\n return 1\n\n elem_1 = elem_2 = 1\n the_sum = 0\n for i in range(3, n + 1):\n the_sum = elem_1 + elem_2\n elem_1, elem_2 = elem_2, the_sum\n return the_sum\n\n\nfor n in range(1, 10): # probando\n print(n, \"->\", fib(n))\n\ndef factorial_function(n):\n if n < 0:\n return None\n if n < 2:\n return 1\n return n * factorial_function(n - 1)","repo_name":"Betelgeusep/python_apuntes","sub_path":"funciones/ejemplos_funciones/fibonacci.py","file_name":"fibonacci.py","file_ext":"py","file_size_in_byte":934,"program_lang":"python","lang":"es","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"3932746951","text":"import urllib.request \r\nimport json\r\nimport ssl\r\n##The program will prompt for a URL, read the JSON data from that URL using\r\n##urllib and then parse and extract the comment counts from the JSON data,\r\n##compute the sum of the numbers in the file and enter the sum below:\r\n##Sample data: http://py4e-data.dr-chuck.net/comments_42.json (Sum=2553)\r\n##Actual data: http://py4e-data.dr-chuck.net/comments_792632.json (Sum ends with 12)\r\n\r\n# Ignore SSL certificate errors\r\nctx = ssl.create_default_context()\r\nctx.check_hostname = False\r\nctx.verify_mode = ssl.CERT_NONE\r\n\r\nurl = input('Enter - ')\r\nprint('Retriving...')\r\nconnection = urllib.request.urlopen(url)\r\ndata=connection.read().decode()\r\njs= json.loads(data)\r\n\r\n\r\n\r\ncount = 0\r\n\r\nfor item in js['comments']:\r\n count = count + int(item['count'])\r\n\r\nprint('count:', count)\r\n","repo_name":"mahta-khoobi/PythonSpecialization_Coursera","sub_path":"3- Accessing Web Data/5- JSON.py","file_name":"5- JSON.py","file_ext":"py","file_size_in_byte":826,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"39563042272","text":"from django.core.management.base import BaseCommand\nfrom counselor.models import *\n\nimport requests\n\nfrom ecocloud.tools import load_csv\n\nCONTINENTS = [\"asia\", \"north america\", \"europe\", \"south america\", \"africa\"]\n\n\nclass Command(BaseCommand):\n\n def handle(self, *args, **options):\n load_csv()\n\n regions = requests.get(\"https://api.aiven.io/v1/clouds\").json()\n\n for region in regions[\"clouds\"]:\n region_object = Region.objects.filter(name=region[\"cloud_name\"])\n if not region_object.exists():\n if \"do\" not in region[\"cloud_name\"] and \"upcloud\" not in region[\"cloud_name\"]:\n country = region[\"cloud_description\"].lower().split(',')[0]\n if country in CONTINENTS:\n country = region[\"cloud_description\"].lower().split(',')[1][1:].split('-')[0][:-1]\n Region.objects.create(name=region[\"cloud_name\"], continent=region[\"geo_region\"], country=country)\n\n services = requests.get(\"https://api.aiven.io/v1/service_types\").json()\n\n for service_name in services[\"service_types\"]:\n service_object = Service.objects.create(name=service_name)\n for service_plan in services[\"service_types\"][service_name][\"service_plans\"]:\n service_plan_name = service_plan[\"service_plan\"]\n for region_name in service_plan[\"regions\"]:\n region = Region.objects.filter(name=region_name).first()\n ServiceRegionRelation.objects.create(\n region=region, service=service_object,\n service_plan=service_plan_name,\n price=float(service_plan[\"regions\"][region_name][\"price_usd\"]) * 730\n )\n","repo_name":"sepehrjavid/ecocloud","sub_path":"counselor/management/commands/fetchdata.py","file_name":"fetchdata.py","file_ext":"py","file_size_in_byte":1765,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"34661219357","text":"from random import randint, random\nfrom package.sga.organism import Organism\nfrom package.constants import CROSSOVER_OPERATOR, CROSSOVER_RATE, ELITISM_AMOUNT, GENERATION_THRESHOLD, ITEMS, MUTATION_RATE, POPULATION_SIZE, SELECTION_OPERATOR, TOURNAMENT_K, AsBit, Bit\n\n\nclass Population:\n generation_number: int\n organisms: list[Organism] # always sorted in order of fitness\n best_organism_ever: Organism # always updated\n\n def __init__(self):\n self.generation_number = 0\n self.organisms = [Population.gen_random_org()\n for _ in range(POPULATION_SIZE)]\n self.organisms.sort()\n self.best_organism_ever = self.get_best_organism()\n\n def perform_generation(self) -> None:\n self.generation_number += 1\n\n mutated_children: list[Organism] = []\n\n elites = self.__get_elites()\n for elite in elites:\n mutated_children.append(elite)\n\n parents: list[Organism] = self.__select_parents()\n while len(mutated_children) < POPULATION_SIZE:\n parent1 = parents[randint(1, len(parents)) - 1]\n parent2 = parents[randint(1, len(parents)) - 1]\n if random() < CROSSOVER_RATE:\n child1, child2 = Population.__crossover(parent1, parent2)\n mutated_children.append(child1)\n mutated_children.append(child2)\n else:\n mutated_children.append(parent1)\n mutated_children.append(parent2)\n\n for child in mutated_children:\n if random() < MUTATION_RATE:\n child.mutate()\n\n self.organisms = mutated_children\n self.organisms.sort()\n\n if self.get_best_organism().get_total_value() > self.best_organism_ever.get_total_value():\n self.best_organism_ever = self.get_best_organism()\n\n def is_finished(self):\n best_fitness = self.get_best_fitness()\n worst_fitness = self.get_worst_fitness()\n difference = abs(best_fitness - worst_fitness)\n less_than_1percent_difference = difference < .01 * \\\n min(abs(best_fitness), abs(worst_fitness))\n if GENERATION_THRESHOLD == -1:\n return less_than_1percent_difference\n return less_than_1percent_difference or self.generation_number >= GENERATION_THRESHOLD\n\n def get_best_organism(self) -> Organism:\n return self.organisms[-1]\n\n def get_best_fitness(self) -> int:\n return self.organisms[-1].fitness\n\n def get_worst_fitness(self) -> int:\n return self.organisms[0].fitness\n\n @staticmethod\n def __crossover(parent1: Organism, parent2: Organism) -> tuple[Organism, Organism]:\n def uniform(parent1: Organism, parent2: Organism):\n mask = Population.__gen_random_bitstring(len(parent1.chromosome))\n return crossover_from_mask(parent1, parent2, mask)\n\n def single_point(parent1: Organism, parent2: Organism):\n point = randint(1, len(parent1.chromosome)-1)\n mask = [0 if i < point else 1 for i in range(\n len(parent1.chromosome))]\n return crossover_from_mask(parent1, parent2, mask) # type:ignore\n\n def double_point(parent1: Organism, parent2: Organism):\n point1, point2 = sorted(\n [randint(1, len(parent1.chromosome)-1) for _ in range(2)])\n mask = [1 if i >= point1 and i <\n point2 else 0 for i in range(len(parent1.chromosome))]\n return crossover_from_mask(parent1, parent2, mask) # type:ignore\n\n def crossover_from_mask(parent1: Organism, parent2: Organism, mask: list[Bit]) -> tuple[Organism, Organism]:\n chromosome1: list[Bit] = []\n chromosome2: list[Bit] = []\n for i in range(len(mask)):\n if mask[i] == 0:\n chromosome1.append(parent1.chromosome[i])\n chromosome2.append(parent2.chromosome[i])\n else:\n chromosome1.append(parent2.chromosome[i])\n chromosome2.append(parent1.chromosome[i])\n return Organism(chromosome1), Organism(chromosome2)\n\n crossover_operators = {'uniform': uniform,\n 'single point': single_point, 'double point': double_point}\n return crossover_operators[CROSSOVER_OPERATOR](parent1, parent2)\n\n def __get_elites(self) -> list[Organism]:\n elites: list[Organism] = []\n for i in range(1, ELITISM_AMOUNT+1):\n elites.append(self.organisms[-i])\n return elites\n\n def __select_parents(self) -> list[Organism]:\n def roulette(self: Population) -> list[Organism]:\n # add constant to all fitness values so they are all positive\n lowest_fitness = self.get_worst_fitness()\n shifter: int = 0\n if lowest_fitness <= 0:\n shifter = abs(lowest_fitness) + 1\n shifted_fitnesses = list(\n map(lambda x: x + shifter, [org.fitness for org in self.organisms]))\n total_fitness = sum(shifted_fitnesses)\n\n parents: list[Organism] = []\n for _ in range(POPULATION_SIZE - ELITISM_AMOUNT):\n spin = randint(1, total_fitness)\n for i in range(len(shifted_fitnesses)):\n spin -= shifted_fitnesses[i]\n if spin <= 0:\n parents.append(self.organisms[i])\n break\n return parents\n\n def tournament(self: Population) -> list[Organism]:\n parents: list[Organism] = []\n for _ in range(POPULATION_SIZE - ELITISM_AMOUNT):\n fighter1 = self.organisms[randint(1, len(self.organisms)) - 1]\n fighter2 = self.organisms[randint(1, len(self.organisms)) - 1]\n if random() < TOURNAMENT_K:\n parents.append(max(fighter1, fighter2))\n else:\n parents.append(min(fighter1, fighter2))\n return parents\n\n selection_operators = {\n 'roulette': roulette, 'tournament': tournament} # type: ignore\n return selection_operators[SELECTION_OPERATOR](self)\n\n @staticmethod\n def gen_random_org() -> Organism:\n return Organism(Population.__gen_random_bitstring(len(ITEMS)))\n\n @staticmethod\n def __gen_random_bitstring(length: int) -> list[Bit]:\n return [AsBit(randint(0, 1)) for _ in range(length)]\n","repo_name":"parkerbedlan/knapsack-evolutionary-computation","sub_path":"package/sga/population.py","file_name":"population.py","file_ext":"py","file_size_in_byte":6463,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7158199367","text":"\"\"\"\nDefines CNN architectures\n\"\"\"\n\nimport tensorflow as tf\nfrom tfutils import model_tool\n\ndef alexnet(images, train=True, norm=True, seed=0, **kwargs):\n \"\"\"\n Alexnet\n \"\"\"\n m = model_tool.ConvNet(seed=seed)\n\n conv_kwargs = {\"add_bn\": False, \"init\": \"xavier\", \"weight_decay\": 0.0001}\n pool_kwargs = {\"pool_type\": \"maxpool\"}\n fc_kwargs = {\"init\": \"trunc_norm\", \"weight_decay\": 0.0001, \"stddev\": 0.01}\n\n dropout = 0.5 if train else None\n\n m.conv(96, 11, 4, padding=\"VALID\", layer=\"conv1\",\n in_layer=images, **conv_kwargs)\n if norm:\n m.lrn(depth_radius=5, bias=1, alpha=0.0001, beta=0.75, layer=\"lrn1\")\n m.pool(3, 2, layer=\"pool1\", **pool_kwargs)\n\n m.conv(256, 5, 1, layer=\"conv2\", **conv_kwargs)\n if norm:\n m.lrn(depth_radius=5, bias=1, alpha=0.0001, beta=0.75, layer=\"lrn2\")\n m.pool(3, 2, layer=\"pool2\", **pool_kwargs)\n\n m.conv(384, 3, 1, layer=\"conv3\", **conv_kwargs)\n m.conv(384, 3, 1, layer=\"conv4\", **conv_kwargs)\n\n m.conv(256, 3, 1, layer=\"conv5\", **conv_kwargs)\n m.pool(3, 2, layer=\"pool5\", **pool_kwargs)\n\n m.fc(4096, dropout=dropout, bias=0.1, layer=\"fc6\", **fc_kwargs)\n m.fc(4096, dropout=dropout, bias=0.1, layer=\"fc7\", **fc_kwargs)\n m.fc(1000, activation=None, dropout=None, bias=0, layer=\"fc8\", **fc_kwargs)\n\n return m\n\n","repo_name":"dawnfinzi/CS431","sub_path":"alexnet/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1320,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"14196780885","text":"# from pudb import set_trace; set_trace()\nfrom typing import List\nimport math\n\n\nclass Solution1:\n def champagneTower(self, poured: int, query_row: int, query_glass: int) -> float:\n \"\"\"LeetCode 799\n\n Terrible performance. I solved this one two years ago, but couldn't\n solve it today. Granted, I was tight on time, but I was not on the\n right path. The way I was trying to simulate was not correct, because\n I tried to find out when the parent cup gets filled and analyze the\n filling situation per cup for child afterwards. This is vastly\n complicated. I had to see the solution, which uses another way to\n simulate. We keep track of the amount of fluid going through each cup.\n And since the amount spilled is deterministic based on the amount that\n goes through a cup, we can easily compute the amount the children will\n get.\n\n O(N^2), N is the number of glasses.\n \"\"\"\n glasses = [[0] * c for c in range(1, 102)]\n glasses[0][0] = poured\n for r in range(query_row + 1):\n for j, g in enumerate(glasses[r]):\n if g > 1:\n glasses[r + 1][j] += (g - 1) / 2\n glasses[r + 1][j + 1] += (g - 1) / 2\n glasses[r][j] = 1\n return glasses[query_row][query_glass]\n\n\nclass Solution2:\n def champagneTower(self, poured: int, query_row: int, query_glass: int) -> float:\n \"\"\"1D DP to save our face.\n\n 108 ms, 88% ranking\n \"\"\"\n glasses = [0] * 100\n glasses[0] = poured\n for r in range(query_row):\n temp = 0\n for j in range(r + 1):\n spill = max((glasses[j] - 1) / 2, 0)\n glasses[j] = spill + temp\n temp = spill\n glasses[r + 1] = temp\n return min(glasses[query_glass], 1.0)\n\n\n\nclass Solution3:\n def champagneTower(self, poured: int, query_row: int, query_glass: int) -> float:\n \"\"\"Smarter 1D DP\n \"\"\"\n glasses = [0] * (query_row + 1)\n glasses[0] = poured\n for r in range(1, query_row + 1):\n for j in range(r, -1, -1):\n glasses[j] = max((glasses[j] - 1) / 2, 0) + max((glasses[j - 1] - 1) / 2, 0)\n return min(glasses[query_glass], 1.0)\n\n\nsol = Solution3()\ntests = [\n (1, 1, 1, 0.0),\n (2, 1, 1, 0.5),\n (100000009, 33, 17, 1.0),\n (1, 1, 0, 0),\n]\n\nfor i, (poured, query_row, query_glass, ans) in enumerate(tests):\n res = sol.champagneTower(poured, query_row, query_glass)\n if math.isclose(res, ans):\n print(f'Test {i}: PASS')\n else:\n print(f'Test {i}; Fail. Ans: {ans}, Res: {res}')\n","repo_name":"FanchenBao/leetcode","sub_path":"2022_03_challenge/03_04_2022.py","file_name":"03_04_2022.py","file_ext":"py","file_size_in_byte":2696,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"26749168475","text":"class Archivo:\n def __init__(self, nombre, extension, delimitador):\n self.nombre = nombre\n self.extension = extension\n self.delimitador = delimitador\n self.texto = \"\"\n\n def get_nombre(self):\n return (self.nombre)\n\n def get_extension(self):\n return (self.extension)\n\n def abrir_Archivo(self):\n filas = open(self.nombre + \".\" + self.extension, \"r\")\n self.texto = filas.readline()\n texto_delimitado = self.texto.replace(\" \", delimitador)\n print(texto_delimitado)\n\n\nnombre_archivo = str(input(\"escriba el nombre del archivo: \"))\next = str(input(\"ingrese la extension: \"))\ndelimitador = str(input(\"ingrese un delimitador: \"))\narchivo = Archivo(nombre_archivo, ext, delimitador)\narchivo.abrir_Archivo()\n","repo_name":"pipo508/Trabajos-Practicos-Computaci-n","sub_path":"tp3/ejercicio 8/ejercicio8.py","file_name":"ejercicio8.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"74756294335","text":"from odoo import models, fields, _\nfrom odoo.exceptions import RedirectWarning, UserError, ValidationError, AccessError\nfrom odoo.tools.misc import formatLang, format_date, get_lang\nfrom odoo.tools import float_compare, date_utils, email_split, email_re\nfrom collections import defaultdict\n\n\nclass AccountMove(models.Model):\n _inherit = 'account.move'\n\n # original def _post, copiado y cambiado el nombre y modificado para que no postee la factura de pos.order\n def _post_global_invoices(self, soft=True):\n \"\"\"Post/Validate the documents.\n\n Posting the documents will give it a number, and check that the document is\n complete (some fields might not be required if not posted but are required\n otherwise).\n If the journal is locked with a hash table, it will be impossible to change\n some fields afterwards.\n\n :param soft (bool): if True, future documents are not immediately posted,\n but are set to be auto posted automatically at the set accounting date.\n Nothing will be performed on those documents before the accounting date.\n :return Model: the documents that have been posted\n \"\"\"\n if soft:\n future_moves = self.filtered(lambda move: move.date > fields.Date.context_today(self))\n future_moves.auto_post = True\n for move in future_moves:\n msg = _('This move will be posted at the accounting date: %(date)s', date=format_date(self.env, move.date))\n move.message_post(body=msg)\n to_post = self - future_moves\n else:\n to_post = self\n\n # `user_has_group` won't be bypassed by `sudo()` since it doesn't change the user anymore.\n if not self.env.su and not self.env.user.has_group('account.group_account_invoice'):\n raise AccessError(_(\"You don't have the access rights to post an invoice.\"))\n for move in to_post:\n if move.partner_bank_id and not move.partner_bank_id.active:\n raise UserError(_(\"The recipient bank account link to this invoice is archived.\\nSo you cannot confirm the invoice.\"))\n if move.state == 'posted':\n raise UserError(_('The entry %s (id %s) is already posted.') % (move.name, move.id))\n if not move.line_ids.filtered(lambda line: not line.display_type):\n raise UserError(_('You need to add a line before posting.'))\n if move.auto_post and move.date > fields.Date.context_today(self):\n date_msg = move.date.strftime(get_lang(self.env).date_format)\n raise UserError(_(\"This move is configured to be auto-posted on %s\", date_msg))\n\n if not move.partner_id:\n if move.is_sale_document():\n raise UserError(_(\"The field 'Customer' is required, please complete it to validate the Customer Invoice.\"))\n elif move.is_purchase_document():\n raise UserError(_(\"The field 'Vendor' is required, please complete it to validate the Vendor Bill.\"))\n\n if move.is_invoice(include_receipts=True) and float_compare(move.amount_total, 0.0, precision_rounding=move.currency_id.rounding) < 0:\n raise UserError(_(\"You cannot validate an invoice with a negative total amount. You should create a credit note instead. Use the action menu to transform it into a credit note or refund.\"))\n\n # Handle case when the invoice_date is not set. In that case, the invoice_date is set at today and then,\n # lines are recomputed accordingly.\n # /!\\ 'check_move_validity' must be there since the dynamic lines will be recomputed outside the 'onchange'\n # environment.\n if not move.invoice_date:\n if move.is_sale_document(include_receipts=True):\n move.invoice_date = fields.Date.context_today(self)\n move.with_context(check_move_validity=False)._onchange_invoice_date()\n elif move.is_purchase_document(include_receipts=True):\n raise UserError(_(\"The Bill/Refund date is required to validate this document.\"))\n\n # When the accounting date is prior to the tax lock date, move it automatically to the next available date.\n # /!\\ 'check_move_validity' must be there since the dynamic lines will be recomputed outside the 'onchange'\n # environment.\n if (move.company_id.tax_lock_date and move.date <= move.company_id.tax_lock_date) and (move.line_ids.tax_ids or move.line_ids.tax_tag_ids):\n move.date = move._get_accounting_date(move.invoice_date or move.date, True)\n move.with_context(check_move_validity=False)._onchange_currency()\n\n # Create the analytic lines in batch is faster as it leads to less cache invalidation.\n to_post.mapped('line_ids').create_analytic_lines()\n\n for move in to_post:\n move.message_subscribe([p.id for p in [move.partner_id] if p not in move.sudo().message_partner_ids])\n\n # Compute 'ref' for 'out_invoice'.\n if move._auto_compute_invoice_reference():\n to_write = {\n 'payment_reference': move._get_invoice_computed_reference(),\n 'line_ids': []\n }\n for line in move.line_ids.filtered(lambda line: line.account_id.user_type_id.type in ('receivable', 'payable')):\n to_write['line_ids'].append((1, line.id, {'name': to_write['payment_reference']}))\n move.write(to_write)\n\n for move in to_post:\n if move.is_sale_document() \\\n and move.journal_id.sale_activity_type_id \\\n and (move.journal_id.sale_activity_user_id or move.invoice_user_id).id not in (self.env.ref('base.user_root').id, False):\n move.activity_schedule(\n date_deadline=min((date for date in move.line_ids.mapped('date_maturity') if date), default=move.date),\n activity_type_id=move.journal_id.sale_activity_type_id.id,\n summary=move.journal_id.sale_activity_note,\n user_id=move.journal_id.sale_activity_user_id.id or move.invoice_user_id.id,\n )\n\n customer_count, supplier_count = defaultdict(int), defaultdict(int)\n for move in to_post:\n if move.is_sale_document():\n customer_count[move.partner_id] += 1\n elif move.is_purchase_document():\n supplier_count[move.partner_id] += 1\n for partner, count in customer_count.items():\n (partner | partner.commercial_partner_id)._increase_rank('customer_rank', count)\n for partner, count in supplier_count.items():\n (partner | partner.commercial_partner_id)._increase_rank('supplier_rank', count)\n\n # Trigger action for paid invoices in amount is zero\n to_post.filtered(\n lambda m: m.is_invoice(include_receipts=True) and m.currency_id.is_zero(m.amount_total)\n ).action_invoice_paid()\n\n # Force balance check since nothing prevents another module to create an incorrect entry.\n # This is performed at the very end to avoid flushing fields before the whole processing.\n to_post._check_balanced()\n return to_post\n","repo_name":"farmaciasaguascalientes/fciags","sub_path":"modulos/zublime_invoicing_pos_global/models/account_move.py","file_name":"account_move.py","file_ext":"py","file_size_in_byte":7391,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"29723907581","text":"import tensorflow as tf\n# 数据集相关\nPATH2RECORD = '/home/uryuo/db/ntire/tfRecord/' # 将数据集转 tfrecord 位置\n# 原始图片 glob 路径\nIMG_HAZE_GT_PATH = '/home/uryuo/db/ntire/image/haze_1/indoor/trainGT/*.jpg'\nIMG_HAZE_NOISE_PATH = '/home/uryuo/db/ntire/image/haze_1/indoor/trainHaze/*.jpg'\nIMG_HAZE_VAILD_PATH = '/home/uryuo/db/ntire/image/haze_1/indoor/validateHaze/*.png'\n# tfrecord 命名\nRECORD_TRAIN_NAME = 'haze_1_train.tfrecord'\nRECORD_VAILD_NAME = 'haze_1_vaild.tfrecord'\n# record验证图片存储LUJING\nRECORD_ALIABLE_PATH = 'resource/'\n# trainning\nTRAIN_BATCH = 64\n# network\nIMG_CROP_SIZE = [512, 512, 3]\nINPUT_SIZE = [512, 512, 3]\n\n\n# limit gpu mem useage\ndef gpu_option():\n gpu_options = tf.GPUOptions(allow_growth=True)\n return tf.ConfigProto(gpu_options=gpu_options)\n","repo_name":"okingjerryo/deep_model_experiment","sub_path":"dehazeExperment/args.py","file_name":"args.py","file_ext":"py","file_size_in_byte":816,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"42404164171","text":"import asyncio\nfrom queue import PriorityQueue\nfrom queue import Queue\nfrom time import time\n\n\nclass ReporterManager:\n def __append(self, reporter):\n index = len(self.__reporters)\n self.__reporters.append(reporter)\n return index\n\n def listen_to(self, reporter):\n index = self.__append(reporter)\n report_at = int(time()) + reporter.interval()\n self.__put(report_at, index)\n\n def add_nop(self, reporter):\n index = self.__append(reporter)\n self.__put_nop(index)\n\n def __reporter(self, index):\n return self.__reporters[index]\n\n def __put(self, report_at, index):\n self.__reporter_queue.put((report_at, index))\n\n def __get(self):\n return self.__reporter_queue.get()\n\n def __empty(self):\n return self.__reporter_queue.empty()\n\n def __put_nop(self, index):\n self.__reporter_nop_queue.put_nowait(index)\n\n def __get_nop(self):\n return self.__reporter_nop_queue.get_nowait()\n\n def __empty_nop(self):\n return self.__reporter_nop_queue.empty()\n\n def __size_nop(self):\n return self.__reporter_nop_queue.qsize()\n\n def __check_nop_queue(self):\n if self.__empty_nop():\n return\n\n nop_size = self.__size_nop()\n for i in range(nop_size):\n index = self.__get_nop()\n reporter = self.__reporter(index)\n if reporter.interval() > 0:\n report_at = int(time()) + reporter.interval()\n self.__put(report_at, index)\n else:\n self.__put_nop(index)\n\n async def __async_loop(self, timeout=None):\n if timeout is not None:\n started_at = int(time())\n\n self.__curr_task = asyncio.current_task()\n while not self.__quit_requested:\n self.__check_nop_queue()\n if self.__empty():\n await asyncio.sleep(Const.SLEEP_NO_REPORT_SEC)\n continue\n\n now = int(time())\n (report_at, index) = self.__get()\n reporter = self.__reporter(index)\n\n if (timeout is not None\n and max(now, report_at) - started_at >= timeout):\n break\n\n if reporter.interval() <= 0:\n self.__put_nop(index)\n if self.__empty():\n await asyncio.sleep(Const.SLEEP_NO_REPORT_SEC)\n continue\n\n if report_at >= now:\n await asyncio.sleep(report_at - now)\n [report, alarm] = reporter.report()\n self.__report_queue.push(report)\n if alarm is not None:\n self.__alarm_queue.push(alarm)\n\n self.__put(report_at + reporter.interval(), index)\n\n def start_loop(self, timeout=None):\n loop = asyncio.new_event_loop()\n asyncio.set_event_loop(loop)\n try:\n loop.run_until_complete(self.__async_loop(timeout))\n except:\n pass\n\n def request_stop(self):\n self.__quit_requested = True\n self.__curr_task.cancel()\n\n def __init__(self, report_queue, alarm_queue):\n self.__reporter_queue = PriorityQueue()\n self.__reporter_nop_queue = Queue()\n self.__report_queue = report_queue\n self.__reporters = []\n self.__alarm_queue = alarm_queue\n self.__quit_requested = False\n","repo_name":"atmark-techno/Azure-IoT-samples","sub_path":"Armadillo-IoT_GW/modules/lib/reporter_manager.py","file_name":"reporter_manager.py","file_ext":"py","file_size_in_byte":3375,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"4477525989","text":"# written by aditya sharma\n# taken from https://www.datacamp.com/community/tutorials/cnn-tensorflow-python\n\n\n# Import libraries\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\n#matplotlib inline\nimport os\nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"0\" #for training on gpu\n\n#loading training data\ndata = input_data.read_data_sets('data/MNIST/',one_hot=True)\n\n# reshaping images\ntrain_X = data.train.images.reshape(-1, 28, 28, 1)\ntest_X = data.test.images.reshape(-1, 28, 28, 1)\n\ntrain_y = data.train.labels\ntest_y = data.test.labels\n\n#defining hyper parameters\ntraining_iters = 1\nlearning_rate = 0.001\nbatch_size = 32\n\n#MNIST data input (img shape: 28*28)\nn_input = 28\n\n#MNIST total classes (0-9 digits)\nn_classes = 10\n\n#defining input and output placeholders\nx = tf.placeholder(\"float\", [None, 28,28,1])\ny = tf.placeholder(\"float\", [None, n_classes])\n\ndef conv2d(x, W, b, strides=1):\n # Conv2D wrapper, with bias and relu activation\n x = tf.nn.conv2d(x, W, strides=[1, strides, strides, 1], padding='SAME')\n x = tf.nn.bias_add(x, b)\n return tf.nn.relu(x)\n\ndef maxpool2d(x, k=2):\n # Max filter\n return tf.nn.max_pool(x, ksize=[1, k, k, 1], strides=[1, k, k, 1],padding='SAME')\n\n#defining Lorentzian activation functions\ndef activationLorentz(x, x0, lambd):\n return (0.5*lambd)/(np.pi * (np.square(x - x0) + np.square(0.5*lambd)))\n\ndef activationDerivLorentz(x, x0, lambd):\n return (-16*lambd*(x - x0))/(np.pi * np.square(4*np.square(x - x0) + lambd**2)) \n\n\n#defining weights\nweights = {\n 'wc1': tf.get_variable('W0', shape=(3,3,1,32), initializer=tf.contrib.layers.xavier_initializer()), \n 'wc2': tf.get_variable('W1', shape=(3,3,32,64), initializer=tf.contrib.layers.xavier_initializer()), \n 'wc3': tf.get_variable('W2', shape=(3,3,64,128), initializer=tf.contrib.layers.xavier_initializer()), \n 'wd1': tf.get_variable('W3', shape=(4*4*128,128), initializer=tf.contrib.layers.xavier_initializer()), \n 'out': tf.get_variable('W6', shape=(128,n_classes), initializer=tf.contrib.layers.xavier_initializer()), \n}\nbiases = {\n 'bc1': tf.get_variable('B0', shape=(32), initializer=tf.contrib.layers.xavier_initializer()),\n 'bc2': tf.get_variable('B1', shape=(64), initializer=tf.contrib.layers.xavier_initializer()),\n 'bc3': tf.get_variable('B2', shape=(128), initializer=tf.contrib.layers.xavier_initializer()),\n 'bd1': tf.get_variable('B3', shape=(128), initializer=tf.contrib.layers.xavier_initializer()),\n 'out': tf.get_variable('B4', shape=(10), initializer=tf.contrib.layers.xavier_initializer()),\n}\n\ndef conv_net(x, weights, biases): \n\n # here we call the conv2d function we had defined above and pass the input image x, weights wc1 and bias bc1.\n conv1 = conv2d(x, weights['wc1'], biases['bc1'])\n # Max Pooling (down-sampling), this chooses the max value from a 2*2 matrix window and outputs a 14*14 matrix.\n conv1 = maxpool2d(conv1, k=2)\n\n # Convolution Layer\n # here we call the conv2d function we had defined above and pass the input image x, weights wc2 and bias bc2.\n conv2 = conv2d(conv1, weights['wc2'], biases['bc2'])\n # Max Pooling (down-sampling), this chooses the max value from a 2*2 matrix window and outputs a 7*7 matrix.\n conv2 = maxpool2d(conv2, k=2)\n\n conv3 = conv2d(conv2, weights['wc3'], biases['bc3'])\n # Max Pooling (down-sampling), this chooses the max value from a 2*2 matrix window and outputs a 4*4.\n conv3 = maxpool2d(conv3, k=2)\n\n # Fully connected layer\n # Reshape conv2 output to fit fully connected layer input\n fc1 = tf.reshape(conv3, [-1, weights['wd1'].get_shape().as_list()[0]])\n fc1 = tf.add(tf.matmul(fc1, weights['wd1']), biases['bd1'])\n fc1 = tf.nn.relu(fc1)\n # Output, class prediction\n # finally we multiply the fully connected layer with the weights and add a bias term. \n out = tf.add(tf.matmul(fc1, weights['out']), biases['out'])\n return out\n\n# passing input through network\npred = conv_net(x, weights, biases)\n# computing cost function\ncost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=pred, labels=y))\n# using adam optimizer to update weights and biases\noptimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)\n\n#Here you check whether the index of the maximum value of the predicted image is equal to the actual labelled image. and both will be a column vector.\ncorrect_prediction = tf.equal(tf.argmax(pred, 1), tf.argmax(y, 1))\n\n#calculate accuracy across all the given images and average them out. \naccuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n\n# Initializing the variables\ninit = tf.global_variables_initializer()\n\nwith tf.Session() as sess:\n sess.run(init) \n train_loss = []\n test_loss = []\n train_accuracy = []\n test_accuracy = []\n # training\n for i in range(training_iters):\n for batch in range(len(train_X)//batch_size):\n batch_x = train_X[batch*batch_size:min((batch+1)*batch_size,len(train_X))]\n batch_y = train_y[batch*batch_size:min((batch+1)*batch_size,len(train_y))] \n # Run optimization op (backprop).\n # Calculate batch loss and accuracy\n opt = sess.run(optimizer, feed_dict={x: batch_x,\n y: batch_y})\n loss, acc = sess.run([cost, accuracy], feed_dict={x: batch_x,\n y: batch_y})\n\n # Calculate accuracy for all 10000 mnist test images\n test_acc,valid_loss = sess.run([accuracy,cost], feed_dict={x: test_X,\n y : test_y})\n train_loss.append(loss)\n test_loss.append(valid_loss)\n train_accuracy.append(acc)\n test_accuracy.append(test_acc)\n print(\"Validation Loss: \" + str(valid_loss))","repo_name":"bendavidsteel/final-year-project","sub_path":"Misc/convneuralnetwork.py","file_name":"convneuralnetwork.py","file_ext":"py","file_size_in_byte":5927,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27818196072","text":"import psycopg2\nimport sys\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\n\ndef main():\n\n\tmovieIndexQuery= \"\"\"CREATE INDEX employeeNames ON Employee(name);\"\"\"\n\n\t#connect to database\n\tpassw = \"Cinema2078\"\n\t#print(\"please enter the password for cs421g78:\")\n\t#pw= str(input())\n\tconnection = psycopg2.connect(user=\"cs421g78\",\n\t\tpassword=passw,\n\t\thost=\"comp421.cs.mcgill.ca\",\n\t\tport=\"5432\",\n\t\tdatabase=\"cs421\")\n\tprint(\"Welcome to the CinemaClub Databse Interace\")\n\tprint(\"\")\n\tdisplay_menu()\n\n\twhile(1): #infinite loop for user inputs \n\t\tprint(\"Input:\", end=\" \")\n\t\tuse_in= str(input())\n\t\tif(use_in==\"\"):\n\t\t\tprint(\"Invalid Input, please try again\")\n\t\t\tcontinue\n\t\telif(int(use_in)==0): #prompted menu to be displayed \n\t\t\tdisplay_menu()\n\t\telif(int(use_in)==10): # prompted exit \n\t\t\texit_prog(connection)\n\n\t\telif(int(use_in)==1):\n\t\t\tprocessOne(connection)\n\t\t\tprint('Do you wish to continue? (Y/N)')\n\t\t\texitq=input()\n\t\t\tif exitq=='N':\n\t\t\t\texit_prog(connection)\n\t\t\telse:\n\t\t\t\tdisplay_menu()\n\n\t\telif(int(use_in)==2):\n\t\t\tprocessTwo(connection)\n\t\t\tprint('Do you wish to continue? (Y/N)')\n\t\t\texitq=input()\n\t\t\tif exitq=='N':\n\t\t\t\texit_prog(connection)\n\t\t\telse:\n\t\t\t\tdisplay_menu()\n\n\t\telif(int(use_in)==3):\n\t\t\tprocessThree(connection)\n\t\t\tprint('Do you wish to continue? (Y/N)')\n\t\t\texitq=input()\n\t\t\tif exitq=='N':\n\t\t\t\texit_prog(connection)\n\t\t\telse:\n\t\t\t\tdisplay_menu()\n\n\t\telif(int(use_in)==4):\n\t\t\tprocessFour(connection)\n\t\t\tprint('Do you wish to continue? (Y/N)')\n\t\t\texitq=input()\n\t\t\tif exitq=='N':\n\t\t\t\texit_prog(connection)\n\t\t\telse:\n\t\t\t\tdisplay_menu()\n\n\t\telif(int(use_in)==5):\n\t\t\tprocessFive(connection)\n\t\t\tprint('Do you wish to continue? (Y/N)')\n\t\t\texitq=input()\n\t\t\tif exitq=='N':\n\t\t\t\texit_prog(connection)\n\t\t\telse:\n\t\t\t\tdisplay_menu()\n\t\telse:\n\t\t\tprint(\"Invalid Input, please try again\")\n\n\ndef processOne(connection): #Make a New Employee\n\tprint(\"What is their name?\")\n\tnew_name= str(input())\n\t#new_name= \"Aleks\"\n\tprint(\"What is their address?\")\n\tnew_add= str(input())\n\t#new_add= \"3434 St. Famille\" \n\tprint(\"What is their new email?\")\n\tnew_email= str(input())\n\t#new_email= \"avm@avm.ca\"\n\tprint(\"Where do they work? please enter a cid\")\n\tnew_cid= str(input())\n\t#new_cid= \"1461\"\n\tprint(\"What is new employeeID?\")\n\tnew_eid= str(input())\n\t#new_eid= \"11111\"\n\tprint(\"What is their starting salary?\")\n\tnew_sal= str(input())\n\t#new_sal= \"500000\"\n\tQuery =\"INSERT INTO employee(eid, cid, name, email, salary, address) VALUES('\"\n\tQuery_with_val = Query+new_eid+\"', '\"+new_cid+\"', '\"+new_name+\"', '\"+new_email+\"','\"+new_sal+\"', '\"+new_add+\"') RETURNING eid;\"\n\t#dat1 = pd.read_sql_query(Query_fin, connection)\n\tdat1 = pd.read_sql_query(Query_with_val, connection)\n\tdat1.set_index(['eid'])\n\tprint(dat1)\n\tprint(Query_with_val)\n\n\n#confirmed works\ndef processTwo(connection): #What movie titles are screening on this date\n\tprint(\"What is the date you would like to select\")\n\tin_date =str(input())\n\t#in_date =\"2121-02-01\"\n\tquery = \"SELECT name, date from screening s, movie m where s.movieid = m.movieid and date ='2121-02-01'\"\n\tq_input = \"SELECT name from screening s, movie m where s.movieid = m.movieid and date =\"\n\tq_fin = q_input+\"'\"+in_date+\"';\"\n\n\tdat1 = pd.read_sql_query(q_fin, connection)\n\tdat1.set_index(['name'])\n\tprint(\"Movie Titles:\\n\")\n\tprint(dat1)\n\tprint(q_fin)\n\n\n#confirmed works\ndef processThree(connection): #How many cinemas are have screened a movie\n\tprint(\"What movie are you looking for?\")\n\tmovie =str(input())\n\t#movie =\"quis urna. Nunc quis arcu\"\n\tquery = \"SELECT count(cid) from screening s, movie m where m.name like '%\"+movie+\"%' and m.movieid=s.movieid\"\n\n\tdat1 = pd.read_sql_query(query, connection)\n\tprint(dat1)\n\tprint(\"Number of cinemas that screened the movie:\\n\")\n\tprint(dat1['count'][0])\n\tprint(query)\n\n\ndef processFour(connection): # How many Items has a customer purchased the\n\tprint(\"Email of the customer?\")\n\temail =str(input())\n\t#email='Quisque@sodales.co.uk'\n\tquery=\"SELECT SUM(quantity) from orders o, customer c where o.custid=c.custid and c.custemail='\"+email+\"'\"\n\tdat1 = pd.read_sql_query(query, connection)\n\tprint('Total items bought')\n\tprint(dat1['sum'][0])\n\tprint(query)\n\ndef processFive(connection): #Show which cinemas have a room larger than 260 seats\n\n\tprint(\"What size of room are you looking for?\")\n\tcap=input()\n\t#cap=280\n\tquery= \"\"\"select address, nbr as room_nb, capacity from cinema c, room r\n\t\t\twhere c.cid=r.cid \n\t\t\tand r.capacity > \"\"\"+str(cap)\n\n\tdat1 = pd.read_sql_query(query, connection)\n\tprint(dat1.keys())\n\tdat1.set_index(['address', 'room_nb', 'capacity'])\n\tprint(\"Cinemas:\\n\")\n\tprint(dat1)\n\tprint(query)\n\n\n\ndef exit_prog(connection):\n\t#close connection to database\n\tconnection = None\n\tprint(\"Exiting Program\")\n\tsys.exit()\n\n\ndef display_menu():\n\tprint(\"-----------------Program Menu-----------------\")\n\tprint(\"Enter the associted number to begin a process:\")\t\n\tprint(\"0: Redisplay Menu\")\n\tprint(\"1: Create a new Employee\")\n\tprint(\"2: What movie screenings are on this date\")\n\tprint(\"3: How many cinemas are have screened a certain movie\")\n\tprint(\"4: How many Items has a customer purchased \")\n\tprint(\"5: Show which cinemas have a room larger than x seats\")\n\tprint(\"10: Exit Program\")\n\nmain()\n","repo_name":"AleksasMurauskas/COMP421Project3","sub_path":"project.py","file_name":"project.py","file_ext":"py","file_size_in_byte":5120,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18757883262","text":"from django.conf.urls import patterns, include, url\n#from django.views.generic.base import TemplateView\nfrom django.http import HttpResponse\nfrom settings import MEDIA_ROOT\n\n# Uncomment the next two lines to enable the admin:\n# from django.contrib import admin\n# admin.autodiscover()\n\nurlpatterns = patterns('',\n # Examples:\n #url(r'^pic/(?P.*)$', 'util.web_helpers.pic_csrf.pic_csrf', name='pic_csrf'),\n #url(r'^$', 'apps.cc.views.index', name='index'),\n url(r'^$', 'apps.cc.views.index', name='index'),\n url(r'^api$', 'apps.cc.views.api', name='api'),\n url(r'^url/(?P\\d+)/delete$', 'apps.cc.views.delete', name='delete'),\n\n # Uncomment the admin/doc line below to enable admin documentation:\n # url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n # Uncomment the next line to enable the admin:\n # url(r'^admin/', include(admin.site.urls)),\n url(r'^500/$', 'util.web_helpers.error_page.server_error_500', name='server_error_500'),\n url(r'^404/$', 'util.web_helpers.error_page.server_error_404', name='server_error_404'),\n url(r'^static/(?P.*)$', 'django.views.static.serve', {'document_root': MEDIA_ROOT }),\n url(r'^robots\\.txt$', lambda r: HttpResponse(\"User-agent: *\\nDisallow: /\", mimetype=\"text/plain\")),\n)\n\nhandler404 = 'util.web_helpers.error_page.server_error_404'\nhandler404 = 'util.web_helpers.error_page.server_error_500'","repo_name":"giveme168/zhiqu","sub_path":"urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1401,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"43600054775","text":"import vk_api\nfrom random import randint\nimport keyboards_main\nimport time\nimport methods\nimport sql_connect\n\nstart_time = time.time()\ntoken = \"vk1.a.-1Nd_H3CRuKOQfBtctmLINsMx6CWFAD-WrdhgaZ0Hryr9DXUP7RXFxlspYtt-J05daqNQnKtEoR15u2IEQcmOVPH2gKe8YnFwbzIKtjmKHNyt6sE0Bi_TMet1MvpN82aF6NbiFuI3Iw2xhAXaXMT_aEKvt76egb0jZM20frQhkVgwD85xhk0n9mWJWLv9-c45D_aXHz_n7xYRAGW6kXCxg\"\nvk = vk_api.VkApi(token=token)\nvk._auth_token()\nb = 0\n\n\ndef writing(print_text, us_id, k_b):\n global b\n a = (vk.method(\"messages.send\", {\"user_id\": us_id, \"message\": print_text, \"random_id\": randint(1, 1000),\n \"keyboard\": k_b.get_keyboard()}))\n while a < b:\n print('Ошибка отправки')\n a = (vk.method(\"messages.send\", {\"user_id\": us_id, \"message\": print_text, \"random_id\": randint(1, 1000),\n \"keyboard\": k_b.get_keyboard()}))\n b = a\n\n\ndef writing_only_text(print_text, us_id):\n global b\n a = (vk.method(\"messages.send\", {\"user_id\": us_id, \"message\": print_text, \"random_id\": randint(1, 1000)}))\n while a < b:\n print('Ошибка отправки')\n a = (vk.method(\"messages.send\", {\"user_id\": us_id, \"message\": print_text, \"random_id\": randint(1, 1000)}))\n b = a\n\n\ndef searching_in_db(vk_id):\n return sql_connect.db_connection_select(f'SELECT * FROM users WHERE vk_id = {vk_id} LIMIT 1;')\n\n\ndef return_ammunition_text(us_id, vk_id):\n generated_string = ''\n for work_line in sql_connect.db_connection_select_tuple(f'SELECT durability_left, (SELECT deffence FROM ammunition '\n f'WHERE id = ammunition_id), (SELECT attack FROM '\n f'ammunition WHERE id = ammunition_id), '\n f'(SELECT durability FROM ammunition WHERE id = '\n f'ammunition_id), (SELECT name FROM ammunition WHERE id = '\n f'ammunition_id) FROM ammunition_users WHERE user_id = '\n f'{us_id};'):\n var_string = f'{work_line[4]}: \\nатака = {work_line[2]}\\nзащита = {work_line[1]}\\nПрочность = {work_line[0]}/' \\\n f'{work_line[3]}\\n\\n'\n generated_string += var_string\n writing('В инвентаре у игрока лежат следующие артефакты:\\n\\n' + generated_string, vk_id, keyboards_main.keyboard_4)\n\n\ndef shop(us_id, vk_id):\n if sql_connect.db_connection_select(f'SELECT id FROM shop WHERE location_id = (SELECT location_id FROM users '\n f'WHERE id = {us_id})') is not None:\n player.part = 'shop'\n generated_string = ''\n kb_list = ['Меню']\n for line_1 in shop_request(us_id):\n var_string = f'{line_1[0]}: \\nатака = {line_1[1]} \\nзащита = {line_1[2]}\\nПрочность: {line_1[3]}\\nЦена: ' \\\n f'{line_1[4]}\\n\\n'\n generated_string += var_string\n kb_list.append(line_1[0])\n writing(\"Содержимое первой страницы магазина:\\n\\n\" + generated_string,\n vk_id, keyboards_main.new_keyboard(kb_list))\n\n else:\n writing('Некорректное значение, выбери пункт меню.', user_id,\n keyboards_main.keyboard_3)\n\n\ndef buying_ammunition(us_id, work_text, vk_id):\n working_price = sql_connect.db_connection_select(f\"SELECT price FROM ammunition WHERE name = '{work_text}';\")\n if player.money >= working_price[0]:\n sql_connect.db_connection_insert(f\"INSERT INTO ammunition_users (user_id, ammunition_id, durability_left) \"\n f\"VALUES ({us_id}, (SELECT id FROM ammunition WHERE name = '{work_text}'), \"\n f\"(SELECT durability FROM ammunition WHERE id = (SELECT id FROM ammunition \"\n f\"WHERE name = '{work_text}')));\")\n player.money -= working_price[0]\n writing_only_text(f'Герой приобрёл {work_text}', vk_id)\n else:\n writing_only_text('К сожалению у тебя недостаточно средств на приобретение данного артефакта(', vk_id)\n\n\ndef shop_request(us_id):\n return sql_connect.db_connection_select_tuple(f'SELECT name, attack, deffence, durability, price FROM ammunition '\n f'WHERE id IN (SELECT ammunition_id FROM ammunition_shop WHERE '\n f'shop_id = (SELECT id FROM shop WHERE location_id = (SELECT '\n f'location_id FROM users WHERE id = {us_id})));')\n\n\nclass Player:\n room = 1\n part = 'main'\n\n def __init__(self, name, vk_id, player_id, country, level, location_id, race, born, experience, money, mem_time):\n self.name = name\n self.vk_id = vk_id\n self.id = player_id\n self.country = country\n self.level = level\n self.location_id = location_id\n self.race = race\n self.born = born\n self.experience = experience\n self.money = money\n self.mem_time = mem_time\n\n\nclass PlayerInRegistration:\n name = ''\n country = ''\n location = ''\n race = ''\n room = 1\n\n def __init__(self, vk_id, mem_time):\n self.vk_id = vk_id\n self.mem_time = mem_time\n\n\nplayers = []\nlist_of_players_id = []\n\nwhile True:\n try:\n messages = vk.method(\"messages.getConversations\", {\"offset\": 0, \"count\": 20, \"filter\": \"unanswered\"})\n print(\"%s секунд на цикл\" % (time.time() - start_time))\n start_time = time.time()\n if messages[\"count\"] >= 1:\n for message in range(messages[\"count\"]):\n text = messages['items'][message]['last_message']['text']\n user_id = messages['items'][message]['last_message']['from_id']\n if user_id not in list_of_players_id:\n user_now = searching_in_db(int(f'{user_id}'))\n if user_now is None:\n if text == \"Регистрация\":\n print('регистрация игрока в базе данных')\n players.append(PlayerInRegistration(user_id, time.time()))\n list_of_players_id.append(user_id)\n writing_only_text('Введи своё игровое имя', user_id)\n else:\n writing_only_text(\"Введи 'Регистрация', начнём твою регистрацию в игре\", user_id)\n else:\n list_of_players_id.append(user_now[2])\n print(list_of_players_id)\n players.append(\n Player(user_now[1], user_now[2], user_now[0], user_now[3], user_now[6], user_now[7],\n user_now[5], user_now[4], user_now[8], user_now[9], time.time()))\n writing(\"Основная страница героя, меню\", user_id, keyboards_main.keyboard_3)\n elif user_id in list_of_players_id:\n for player in players:\n if player.vk_id == user_id:\n player.mem_time = time.time()\n if type(player) == PlayerInRegistration:\n if player.room == 1:\n player.name = text\n player.room = 2\n writing('Выбери страну', user_id, keyboards_main.keyboard_0)\n elif player.room == 2:\n if text == 'Империя':\n player.country = 'Империя'\n player.location = 1\n player.room = 3\n writing('Выбери расу', user_id, keyboards_main.keyboard_1)\n elif text == 'Королевство':\n player.country = 'Королевство'\n player.location = 2\n player.room = 3\n writing('Выбери расу', user_id, keyboards_main.keyboard_1)\n else:\n writing('Некорректное значение, выбери страну.', user_id,\n keyboards_main.keyboard_0)\n elif player.room == 3:\n if text == 'Эльф':\n player.race = 'Эльф'\n player.room = 4\n writing('переходим в основную игру', user_id, keyboards_main.keyboard_2)\n elif text == 'Гном':\n player.race = 'Гном'\n player.room = 4\n writing('переходим в основную игру', user_id, keyboards_main.keyboard_2)\n else:\n writing('Некорректное значение, выбери расу.', user_id,\n keyboards_main.keyboard_1)\n elif player.room == 4:\n if text == 'Продолжить':\n sql_connect.db_connection_insert(f'INSERT users (name, vk_id, country, race, '\n f'location_id) VALUES (\"{player.name}\", '\n f'{player.vk_id}, \"{player.country}\",'\n f'\"{player.race}\", {player.location});')\n for i in range(len(list_of_players_id)):\n if list_of_players_id[i] == user_id:\n list_of_players_id.pop(i)\n for i in range(len(players)):\n if players[i].vk_id == user_id:\n players.pop(i)\n else:\n if player.part == 'main':\n if player.room == 1:\n if text == 'Карта':\n player.part = 'map'\n writing(\"Герой зашёл в карту\", user_id, keyboards_main.keyboard_5)\n elif text == 'Инвентарь':\n player.part = 'inventory'\n writing(\"Герой зашёл в инвентарь\", user_id, keyboards_main.keyboard_4)\n elif text == \"Магазин\":\n writing_only_text('Герой зашёл в магазин. \\n', user_id)\n shop(player.id, user_id)\n else:\n writing('Некорректное значение, выбери пункт меню.', user_id,\n keyboards_main.keyboard_3)\n elif player.part == 'shop':\n if player.room == 1:\n if text == 'Меню':\n writing(\"Основная страница героя, меню\", user_id, keyboards_main.keyboard_3)\n player.part = 'main'\n else:\n for line in shop_request(player.id):\n if text == line[0]:\n buying_ammunition(player.id, text, user_id)\n\n shop(player.id, user_id)\n elif player.part == 'inventory':\n if player.room == 1:\n if text == 'Меню':\n writing(\"Основная страница героя, меню\", user_id, keyboards_main.keyboard_3)\n player.part = 'main'\n elif text == 'Оружие':\n return_ammunition_text(player.id, user_id)\n else:\n writing('Некорректное значение, выбери пункт меню.', user_id,\n keyboards_main.keyboard_4)\n elif player.part == 'map':\n if player.room == 1:\n if text == 'Меню':\n writing(\"Основная страница героя, меню\", user_id, keyboards_main.keyboard_3)\n player.part = 'main'\n elif text == 'Квест':\n ls, w_str = methods.quests_list(user_id, player.id)\n writing(w_str, user_id, keyboards_main.new_keyboard(ls))\n player.part = 'quest'\n else:\n writing('Некорректное значение, выбери пункт меню.', user_id,\n keyboards_main.keyboard_5)\n elif player.part == 'quest':\n if player.room == 1:\n if text == '':\n pass\n else:\n writing('Некорректное значение, выбери пункт меню.', user_id,\n keyboards_main.keyboard_5)\n else:\n var_1 = 0\n for i in range(len(players)):\n if time.time() - players[i - var_1].mem_time >= 300:\n sql_connect.db_connection_insert(f'UPDATE users SET level_ = {players[i - var_1].level}, '\n f'location_id = {players[i - var_1].location_id}, experience = '\n f'{players[i - var_1].experience}, money = '\n f'{players[i - var_1].money} WHERE id = {players[i - var_1].id};')\n print(f'Удаляем игрока {players[i - var_1].name} из оперативной памяти')\n list_of_players_id.remove(players[i - var_1].vk_id)\n players.pop(i - var_1)\n print(players)\n print(list_of_players_id)\n var_1 += 1\n except vk_api.exceptions.ApiError or ConnectionError or vk_api.exceptions.ApiHttpError:\n pass\n","repo_name":"3ovutVanja/ChatBotGameFantasy","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":16300,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"23973795946","text":"# 3. Узнайте у пользователя число n.\n# Найдите сумму чисел n + nn + nnn.\n# Например, пользователь ввёл число 3. Считаем 3 + 33 + 333 = 369.\n\nnumber = input(\"Введите цифру от 0 до 9: \")\ntry:\n if int(number) in range(10):\n print(int(number) + int(number * 2) + int(number * 3))\n else:\n print(\"Введенная цифра больше вне диапазона от 0 до 9. Попробуйте еще раз\")\nexcept:\n print(\"Вы ввели не цифру. Попробуйте еще раз\")\n","repo_name":"MA32021/gbPythonBasics","sub_path":"Lesson_1_3.py","file_name":"Lesson_1_3.py","file_ext":"py","file_size_in_byte":615,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"16355145574","text":"# -*- coding: utf-8 -*-\n# @Time : 2022/3/14 17:05\n# @Author : kl\n# @File : test_register.py\nimport os\nimport ddt\nimport unittest\nimport requests\nfrom tools.HandleExcel import HandleExcel\nfrom tools.handle_path import testdatas_dir\nfrom tools.handle_requests import HandleRequests\nfrom tools.hand_log import HandleLog\n\nexcel_path = os.path.join(testdatas_dir, \"cases.xlsx\")\nhe = HandleExcel(excel_path,\"注册\")\ndatas = he.read_all_rows_data()\nprint(datas)\nlogger = HandleLog()\n\n@ddt.ddt\nclass TestRegister(unittest.TestCase):\n name = \"注册\"\n\n @classmethod\n def setUpClass(cls) -> None:\n logger.info(f\"============= {cls.name} 接口测试开始! ==============\")\n\n @classmethod\n def tearDownClass(cls) -> None:\n logger.info(f\"============= {cls.name} 接口测试结束! ==============\")\n\n @ddt.data(*datas)\n def test_register_success(self, case):\n logger.info(f\"********* {case.get('title')} 用例 ********\")\n hr = HandleRequests()\n resp = hr.send_req(case[\"method\"], case[\"url\"], eval(case['req_data']))\n res_dict = resp.json()\n print(res_dict)\n\nif __name__ == '__main__':\n unittest.main()","repo_name":"kuanglei1986/lemon","sub_path":"0308/testcases/test_register.py","file_name":"test_register.py","file_ext":"py","file_size_in_byte":1178,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"69929232576","text":"from flask import Flask,request,render_template\nfrom graduation_project.web_app.form import Start_Form\nfrom graduation_project.web_app.redis_queue import Queue\napp = Flask(__name__)\napp.secret_key='renjian'\nqueue=Queue()\n\n# @app.route('/')\n# def index():\n# return render_template('start_crawl.htm')\n\n\n\n@app.route('/hello')\ndef hello():\n return 'Hello World'\n\n\n# @app.route('/login', methods=['POST', 'GET'])\n# def login():\n# error = None\n# if request.method == 'POST':\n# if valid_login(request.form['username'],\n# request.form['password']):\n# return log_the_user_in(request.form['username'])\n# else:\n# error = 'Invalid username/password'\n# # the code below is executed if the request method\n# # was GET or the credentials were invalid\n# return render_template('login.html', error=error)\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef login():\n form = Start_Form()\n if form.validate_on_submit():\n start_data=form.data['start_data']\n if form.data['start_data']:\n queue.push(start_data)\n return 'sucess,your data is %s'%start_data\n else:\n return 'not data'\n return render_template('start_crawl.htm', form=form)\n\n\n\n@app.route('/user/')\ndef show_user_profile(username):\n # show the user profile for that user\n return 'User %s' % username\n\n\n# @app.route('/login', methods=['GET', 'POST'])\n# def login():\n# if request.method == 'POST':\n# do_the_login()\n# else:\n# show_the_login_form()\n\n\n\nif __name__ == '__main__':\n app.run(debug=True)","repo_name":"cdrenjian/graduation_project","sub_path":"web_app/crawl_app.py","file_name":"crawl_app.py","file_ext":"py","file_size_in_byte":1624,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39896201919","text":"# Function creates a new NTE, notes segment and moves the\r\n# information in TXA 13 to field three of the NTE segment\r\n\r\ndef update_nte_3(json_in):\r\n\r\n for (s, value) in json_in.items():\r\n if s.startswith('TXA')and len(json_in[s]['13']) > 0:\r\n nte = {'1': \"\", '2': \"\", '3': json_in[s]['13'], '4': \"\"}\r\n json_in.update({'NTE.'+str(len(json_in)): nte})\r\n break\r\n\r\n return json_in\r\n","repo_name":"jsdavis02/mft_hl7_oasis_components","sub_path":"oasis_fun/update_nte_3.py","file_name":"update_nte_3.py","file_ext":"py","file_size_in_byte":424,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39507015249","text":"# 213. House Robber II\n# https://leetcode.com/problems/house-robber-ii/description/\n\n\nclass Solution:\n def rob(self, nums: List[int]) -> int:\n\n return max(nums[0], self.dynamic(nums[:-1]), self.dynamic(nums[1:]))\n\n\n def dynamic(self, nums):\n rob1, rob2 = 0, 0\n\n\n for n in nums:\n temp = max(rob1 + n, rob2)\n rob1 = rob2\n rob2 = temp\n\n return rob2\n \n \n \n# Example 1:\n\n# Input: nums = [2,3,2]\n# Output: 3\n# Explanation: You cannot rob house 1 (money = 2) and then rob house 3 (money = 2), because they are adjacent houses.\n \n \n# Example 2:\n\n# Input: nums = [1,2,3,1]\n# Output: 4\n# Explanation: Rob house 1 (money = 1) and then rob house 3 (money = 3).\n# Total amount you can rob = 1 + 3 = 4.\n","repo_name":"anoopanni/leetcode","sub_path":"HourseRobber2.py","file_name":"HourseRobber2.py","file_ext":"py","file_size_in_byte":768,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"21202350823","text":"# Code for whole NA domain using Kao's dataset\nimport os \nimport netCDF4 as nc\nimport numpy as np\nfrom itertools import cycle\nfrom pprint import pprint\nfrom time import process_time \n\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\ninput_path= '/home/7xw/data/GWSP3_DayMet/2014NA'\nfile_name = '/home/7xw/data/GWSP3_DayMet/2014NA/clmforc.Daymet4.1km.FSDS.2014-01.nc'\n\nvis =0 \ndebug = 1\nsave_memory = 1\n\nnumber_of_subdomains = 4200 # 4200 for 700 Summit nodes\ni_timesteps = 248 # 248 for 31 days\n\nstart = process_time()\n# Open a new NetCDF file to write the data to. For format, you can choose from\n# 'NETCDF3_CLASSIC', 'NETCDF3_64BIT', 'NETCDF4_CLASSIC', and 'NETCDF4'\nr_nc_fid = nc.Dataset(file_name, 'r', format='NETCDF4')\n\ntotal_rows = r_nc_fid.dimensions['x'].size\ntotal_cols = r_nc_fid.dimensions['y'].size\ntotal_timesteps = r_nc_fid.dimensions['time'].size\n\n\nFSDS = r_nc_fid['FSDS'][0:i_timesteps, :, :] # read (timestep, y, x) format\n#FSDS = FSDS.transpose(0,2,1) # change to (time, x,y) format\nend = process_time()\nprint(\"Reading FSDS takes {}\".format(end-start))\n\nif vis :\n x = np.linspace(1, total_rows, total_rows, dtype=int)\n y = np.linspace(1, total_cols, total_cols, dtype=int)\n fig, ax = plt.subplots(subplot_kw={'projection': '3d'})\n X,Y = np.meshgrid(x, y)\n ax.view_init(90,0)\n ax.plot_wireframe(Y, X, FSDS[0])\n plt.show()\n\n# Create a global ID for ALL the gridcells\n\nstart = process_time()\ntotal_gridcells = total_rows * total_cols\ngrid_ids = np.linspace(0, total_gridcells-1, total_gridcells, dtype=int)\n\n# create a mask for land grid_ids (1)\nmask = FSDS[0] # FSDS is in (time, Y, X) format\nmask = np.where(~np.isnan(mask), 1, 0)\n\n# create an flattened list of land gridID and reduce the size of gridIDs array\ngrid_ids = grid_ids.reshape(total_cols,total_rows)\ngrid_ids = np.multiply(mask,grid_ids)\ngrid_ids = grid_ids[grid_ids != 0]\n\nend = process_time()\nprint(\"Generate Grid_id takes {}\".format(end-start))\n\n# use the size of land gridcells to resize the FSDS matrix\nstart = process_time()\nlandcells = len(grid_ids)\nif debug:\n print('number of land cells is '+str(landcells))\n \nFSDS=FSDS[~np.isnan(FSDS)]\nFSDS = np.reshape(FSDS,(i_timesteps,landcells))\n\nend = process_time()\nprint(\"Creating dense FSDS takes {}\".format(end-start))\n\n\nstart = process_time()\n# partition landcells into subdomains\n# number_of_subdomains = 4200 # 4200 for 700 Summit nodes\n\n# cyclic (round-robin) partition\ndomains = [[] for _ in range(number_of_subdomains)]\nfor element, domain in zip(grid_ids, cycle(domains)):\n domain.append(element)\n\n#for i in range(number_of_subdomains):\n# # convert local gridID-list into array\n# grid_id_arr[i] = np.array(domains[i]) \ngrid_id_domains = domains.copy()\n\n# partition the FSDS over landcells\n# landcell_idx is alse the column_idx of FSDS\nlandcell_idx = np.linspace(0, landcells-1, landcells, dtype=int)\n\ndomains = [[] for _ in range(number_of_subdomains)]\nfor element, domain in zip(landcell_idx, cycle(domains)):\n domain.append(element)\n \n# save the boundaries of each subdomain (for array_split)\nsize_of_subdomains = [ len(domain) for domain in domains]\n\n# partitioned landcells_idx in subdomains \narranged_grid_idx = np.concatenate(domains).ravel()\nprint(arranged_grid_idx)\n# find the original index of landcells for column swap\nnp.sort(arranged_grid_idx)\ngrid_swap_idx = (np.argsort(arranged_grid_idx))\n\n# create swap index and arrange FSDS\nidx = np.empty_like(grid_swap_idx)\nidx[grid_swap_idx] = np.arange(len(grid_swap_idx))\nFSDS = FSDS[:,idx]\n\nend = process_time()\nprint(\"Partitioning FSDS/GridID takes {}\".format(end-start))\n","repo_name":"daliwang/DaymetDataUtilities","sub_path":"partition_3rd_swap.py","file_name":"partition_3rd_swap.py","file_ext":"py","file_size_in_byte":3655,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28606680639","text":"from django.contrib.auth import login, logout\nfrom django.contrib.auth.views import LoginView\nfrom django.http import HttpResponseRedirect\nfrom django.shortcuts import render, redirect\nfrom django.urls import reverse_lazy\nfrom django.views.generic import ListView, DetailView, CreateView\nfrom django.views.generic.base import View\nfrom django.core.paginator import Paginator\n\nfrom django.contrib import messages\nfrom transliterate import slugify\n\nfrom .models import *\nfrom .forms import *\n\ndef index(request):\n if request.user.is_authenticated:\n return redirect(f'/rec_article/')\n return redirect(f'/all_articles/')\n\n\nclass AllArticles(ListView):\n # Представление главной страницы со всеми статьями\n paginate_by = 10\n model = Article\n template_name = 'Blog/index.html'\n context_object_name = 'articles'\n queryset = Article.objects.all().order_by('-pk')\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['title'] = 'Главная страница'\n context['head'] = 'Все статьи'\n\n return context\n\n\nclass RecArticle(View):\n # Представление главной страницы с рекомендациями\n def get(self, request):\n # Поиск рекомендованных статей\n subscribe_article = MarkedArticle.objects.filter(subscribe_to_post=True, showing_to_user=request.user)\n recommended_articles = Article.objects.none()\n\n if subscribe_article.exists():\n for i in subscribe_article:\n recommended_articles = recommended_articles.union(\n Article.objects.filter(author=i.article.author, time_create__gte=i.time_subscribe_to_post)\n )\n recommended_articles = recommended_articles.order_by('-time_create')\n head = 'Ваши рекомендации'\n paginator = Paginator(recommended_articles, 10)\n page_number = request.GET.get('page')\n page_obj = paginator.get_page(page_number)\n\n else:\n head = 'Вы не подписались не на одну статью'\n page_obj = None\n\n context = {\n 'page_obj': page_obj,\n 'title': 'Рекомендации',\n 'head': head\n }\n return render(request, 'Blog/index.html', context=context)\n\n\nclass ReadArticle(View):\n # Представление прочитанных статей на главной странице\n\n def get(self, request):\n # Поиск прочитанных статей\n articles_read = ReadTheArticle.objects.filter(showing_to_user=self.request.user)\n articles_read_user = Article.objects.none()\n if articles_read.exists():\n for i in articles_read:\n articles_read_user = articles_read_user.union(Article.objects.filter(pk=i.article.pk))\n\n if articles_read_user.exists():\n head = 'Прочитанные статьи'\n paginator = Paginator(articles_read_user, 10)\n page_number = request.GET.get('page')\n page_obj = paginator.get_page(page_number)\n else:\n head = 'Вы не прочитали не одной статьи'\n page_obj = None\n\n context = {\n 'page_obj': page_obj,\n 'title': 'Прочитанные статьи',\n 'head': head\n }\n return render(request, 'Blog/index.html', context=context)\n\n\nclass Author(ListView):\n # Представление страницы пользователей с количествами постов\n model = CountArticleUser\n template_name = 'Blog/authors.html'\n context_object_name = 'count_article_user'\n queryset = CountArticleUser.objects.all().order_by('author')\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['sort_form'] = SortAutorForm()\n context['title'] = 'Пользователи'\n return context\n\n\nclass SortListUser(View):\n # Представление отсортированной страницы пользователей(по количеству постов, по авторам)\n def post(self, request):\n sort_form = SortAutorForm(request.POST)\n count_article_user = CountArticleUser.objects.all()\n if sort_form.is_valid():\n needed_sort = sort_form.cleaned_data.get(\"sort_autor_form\")\n if needed_sort == \"количество\":\n count_article_user = count_article_user.order_by(\"-count\")\n elif needed_sort == \"автор\":\n count_article_user = count_article_user.order_by(\"author\")\n context = {\n 'sort_form': sort_form,\n 'count_article_user': count_article_user,\n 'title': 'Пользователи'\n }\n return render(request, 'Blog/authors.html', context=context)\n\n\nclass ShowArticleAuthor(DetailView):\n # Представление страницы пользователя с написанными им статьями\n model = User\n template_name = 'Blog/author_detail.html'\n context_object_name = 'user'\n pk_url_kwarg = 'user_id'\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['title'] = f'Посты пользователя {self.request.user}'\n context['article'] = Article.objects.filter(author=kwargs['object']).order_by('-time_create')\n\n return context\n\n\nclass ShowArticle(DetailView):\n # Представление для конкретной статьи\n model = Article\n template_name = 'Blog/article_detail.html'\n context_object_name = 'article'\n slug_url_kwarg = 'article_slug'\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['title'] = f'Пост {context[\"article\"]}'\n print(context[\"article\"].slug)\n a = Article.objects.get(slug=context[\"article\"].slug)\n try:\n context['marked_article'] = MarkedArticle.objects.get(showing_to_user=self.request.user, article=a)\n except MarkedArticle.DoesNotExist:\n context['marked_article'] = None\n\n try:\n context['read_the_article'] = ReadTheArticle.objects.get(showing_to_user=self.request.user, article=a)\n except ReadTheArticle.DoesNotExist:\n context['read_the_article'] = None\n\n return context\n\n\ndef sub_from_post(request, article_slug):\n subscribe_to_post = MarkedArticle.objects.create(\n article=Article.objects.get(slug=article_slug), showing_to_user=request.user, subscribe_to_post=True\n )\n\n return redirect(f'/article/{article_slug}/')\n\n\ndef unsub_from_post(request, article_slug):\n a = Article.objects.get(slug=article_slug)\n unsub = MarkedArticle.objects.get(showing_to_user=request.user, article=a)\n unsub.delete()\n\n return redirect(f'/article/{article_slug}/')\n\n\ndef mark_as_read(request, article_slug):\n marked_article = ReadTheArticle.objects.create(\n article=Article.objects.get(slug=article_slug), showing_to_user=request.user\n )\n\n return redirect(f'/article/{article_slug}/')\n\n\ndef remove_the_mark(request, article_slug):\n a = Article.objects.get(slug=article_slug)\n remove_mark = ReadTheArticle.objects.get(showing_to_user=request.user, article=a)\n remove_mark.delete()\n\n return redirect(f'/article/{article_slug}/')\n\n\nclass AddArticlePage(View):\n def get(self, request):\n form = AddArticleForm(request.POST)\n context = {\n 'title': 'Добавление статьи',\n 'form': form\n }\n return render(request, 'Blog/add_article.html', context)\n\n\nclass AddArticle(View):\n def post(self, request):\n form = AddArticleForm(request.POST)\n if form.is_valid():\n add_article = form.save(commit=False)\n add_article.title = form.cleaned_data['title']\n add_article.content = form.cleaned_data['content']\n add_article.author = request.user\n add_article.slug = slugify(form.cleaned_data['title'])\n add_article.save()\n\n cau = CountArticleUser.objects.get(author=request.user)\n cau.count += 1\n cau.save()\n\n messages.add_message(request, messages.INFO, \"Статья успешно добавлена!\")\n return HttpResponseRedirect('/')\n\n\nclass RegisterUser(CreateView):\n form_class = RegisterUserForm\n template_name = 'Blog/register.html'\n success_url = reverse_lazy('authors')\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['title'] = \"Регистрация\"\n messages.add_message(self.request, messages.INFO, \"Статья успешно добавлена!\")\n\n return context\n\n def form_valid(self, form):\n user = form.save()\n login(self.request, user)\n CountArticleUser.objects.create(\n author=self.request.user, count=0\n )\n return redirect('index')\n\n\nclass LoginUser(LoginView):\n form_class = LoginUserForm\n template_name = 'Blog/login.html'\n\n def get_context_data(self, *, object_list=None, **kwargs):\n context = super().get_context_data(**kwargs)\n context['title'] = 'Авторизация'\n return context\n\n def get_success_url(self):\n return reverse_lazy('index')\n\n\ndef logout_user(request):\n logout(request)\n print('hello')\n return redirect('login')\n\n\n","repo_name":"TriplG/WB-Tech","sub_path":"Blog/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":9818,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6125463030","text":"# 57-WAP to swap the values of two variable that are defined as global variables\nfn,sn=5,18\ndef swap():\n global fn,sn\n temp=fn\n fn=sn\n sn=temp\n print(\"After Swapping:\\n First Num={} and Second Num={}\".format(fn,sn))\n\nswap()","repo_name":"faiz1675/Python-Programs","sub_path":"LabFile/57-funSwap.py","file_name":"57-funSwap.py","file_ext":"py","file_size_in_byte":238,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"34597808291","text":"import difflib\nimport html.parser\nimport itertools\nimport operator\nimport sys\n\nclass Parser(html.parser.HTMLParser):\n def __init__(self, *args, **kargs):\n html.parser.HTMLParser.__init__(self, *args, **kargs)\n self.avoid_collisions_nodes = []\n\n def handle_starttag(self, tag, attrs):\n attrs = \" \".join([\"=\".join([a, '\"%s\"' % b]) for a, b in attrs])\n node = (\"<%s %s>\" % (tag, attrs)) if attrs else (\"<%s>\" % tag)\n self.avoid_collisions_nodes.append(node)\n\n def handle_endtag(self, tag):\n node = \"\" % tag\n self.avoid_collisions_nodes.append(node)\n\n def handle_data(self, data):\n if self.avoid_collisions_nodes:\n if not self.avoid_collisions_nodes[-1].startswith(\"<\"):\n self.avoid_collisions_nodes[-1] += data\n return\n self.avoid_collisions_nodes.append(data)\n\n def handle_entityref(self, name):\n self.handle_data(\"&%s;\" % name)\n\n def handle_charref(self, name):\n self.handle_data(\"&%s;\" % name)\n\ndef character_data(node):\n return not node.startswith(\"<\")\n\ndef content(line):\n return line[2:]\n\ndef out(text):\n # http://stackoverflow.com/questions/4601912\n sys.stdout.flush()\n sys.stdout.buffer.write(text.encode(\"utf-8\"))\n\ndef lines(old, new):\n delta = difflib.ndiff(old, new)\n\n print(\"\")\n print(\"\")\n print(\"\")\n first = operator.itemgetter(0)\n for item, group in itertools.groupby(delta, first):\n if item == \"?\":\n continue\n\n out(\"\" % item)\n group = [content(line) for line in group]\n cdata = any(map(character_data, group))\n\n if cdata:\n if item == \"+\":\n out(\"\")\n elif item == \"-\":\n out(\"\")\n\n if cdata or (item != \"-\"):\n for line in group:\n out(line)\n\n if cdata:\n if item == \"+\":\n out(\"\")\n elif item == \"-\":\n out(\"\")\n\ndef files(a, b):\n parser_a = Parser()\n with open(a, \"rb\") as f:\n for line in f:\n try:\n line = line.decode(\"utf-8\")\n except:\n continue\n parser_a.feed(line)\n old = parser_a.avoid_collisions_nodes\n\n parser_b = Parser()\n with open(b, \"rb\") as f:\n for line in f:\n try:\n line = line.decode(\"utf-8\")\n except:\n continue\n parser_b.feed(line)\n new = parser_b.avoid_collisions_nodes\n\n lines(old, new)\n\n","repo_name":"sbp/orinoco","sub_path":"compare.py","file_name":"compare.py","file_ext":"py","file_size_in_byte":2737,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"37175753741","text":"from django.core.exceptions import ValidationError\nfrom django.core.validators import MaxValueValidator, MinValueValidator\nfrom django.db import models\nfrom .customer import Customer, validate_balance\n\n\nclass BalanceWriteOff(models.Model):\n \"\"\"\n Model related Many-To-Many with Customer model.\n Contains information about the customer balance write-offs.\n It includes:\n customer: foreign key to Customer model, show who exactly\n WAS written-off ucoins.\n related_name - incoming_balance_write_offs_set\n from_customer:\n foreign key to Customer model, show who exactly\n written-off ucoins.\n related_name - outgoing_balance_write_offs_set\n count: positive integer field, contains the count of ucoins\n for which the balance was written-off.\n comment: string integer field with max_length = 250.\n date: datetime field.\n \"\"\"\n customer = models.ForeignKey(Customer, on_delete=models.CASCADE,\n related_name=\"incoming_balance_write_offs_set\")\n from_customer = models.ForeignKey(Customer, on_delete=models.SET_NULL, null=True,\n help_text=\"Покупатель, который списал юкойны с баланса.\",\n related_name=\"outgoing_balance_write_offs_set\")\n # count = models.PositiveIntegerField(default=0, null=False)\n count = models.FloatField(\n default=0.0,\n validators=[\n validate_balance,\n MaxValueValidator(9999.9),\n MinValueValidator(0.0)\n ],\n null=False\n )\n header = models.CharField(max_length=100, null=False, blank=False, default=\"Списание\")\n comment = models.CharField(max_length=250, null=False, blank=True, default=\"\")\n date = models.DateTimeField(auto_now_add=True)\n\n type = \"write_off\"\n\n class Meta:\n verbose_name = \"Списание баланса\"\n verbose_name_plural = \"Списания балансов\"\n ordering = [\"-date\"]\n\n def __str__(self):\n return f\"Списание со счёта пользователя {self.customer.name()} на {self.count}\"\n\n def customer_name(self):\n return self.customer.name()\n\n def from_customer_name(self):\n if self.from_customer:\n return self.from_customer.name()\n return None\n\n def save(self, *args, **kwargs):\n try:\n self.customer.decrease_balance(self.count)\n except ValidationError as err:\n raise err\n else:\n super(BalanceWriteOff, self).save(*args, **kwargs)\n","repo_name":"DevCrusader/UStoreAPI","sub_path":"project/customer/models/balance_write_off.py","file_name":"balance_write_off.py","file_ext":"py","file_size_in_byte":2739,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"30643734343","text":"import RPi.GPIO as GPIO\n\nclass Motor:\n def __init__(self, en, in1, in2):\n self.EN = en\n self.IN1 = in1\n self.IN2 = in2\n\n GPIO.setup([self.EN, self.IN1, self.IN2], GPIO.OUT, initial=GPIO.LOW)\n\n def stop(self):\n GPIO.output([self.EN, self.IN1, self.IN2], GPIO.LOW)\n\n def forward(self):\n GPIO.output([self.EN, self.IN1], GPIO.HIGH)\n GPIO.output(self.IN2, GPIO.LOW)\n\n def backward(self):\n GPIO.output([self.EN, self.IN2], GPIO.HIGH)\n GPIO.output(self.IN1, GPIO.LOW)\n\nclass Robot:\n def __init__(self):\n GPIO.setwarnings(False)\n GPIO.setmode(GPIO.BOARD)\n\n self.left_motor = Motor(32, 38, 40)\n self.right_motor = Motor(33, 37, 35)\n\n self.steering = \"fw\"\n self.direction = \"stopped\"\n\n def __del__(self):\n self.stop()\n try:\n GPIO.cleanup()\n except Exception as e:\n print(\"Cleanup failed\")\n print(e)\n\n def move(self):\n if self.direction == \"fw\":\n if self.steering == \"fw\":\n self.left_motor.forward()\n self.right_motor.forward()\n elif self.steering == \"left\":\n self.right_motor.forward()\n self.left_motor.stop()\n else:\n self.left_motor.forward()\n self.right_motor.stop()\n elif self.direction == \"bw\":\n if self.steering == \"fw\":\n self.left_motor.backward()\n self.right_motor.backward()\n elif self.steering == \"left\":\n self.left_motor.backward()\n self.right_motor.stop()\n else:\n self.right_motor.backward()\n self.left_motor.stop()\n else:\n self.left_motor.stop()\n self.right_motor.stop()\n\n def no_steering(self):\n self.steering = \"fw\"\n self.move()\n\n def stop(self):\n self.direction = \"stopped\"\n self.move()\n\n def forward(self):\n self.direction = \"fw\"\n self.move()\n\n def backward(self):\n self.direction = \"bw\"\n self.move()\n\n def right(self):\n self.steering = \"right\"\n self.move()\n\n def left(self):\n self.steering = \"left\"\n self.move()\n","repo_name":"Jostyck9/remote-robot-surveillance","sub_path":"robot/robot.py","file_name":"robot.py","file_ext":"py","file_size_in_byte":2291,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28208022805","text":"#!/usr/bin/env python3\n\nimport os\nimport sys\nimport matplotlib.pyplot as plt\n\nresultfile = sys.argv[1]\n\nx = []\ny = []\n\ndef get_garph():\n fp = open(resultfile,'r')\n data = fp.readlines()\n\n for i in data:\n tempx = (i.split())[0]\n tempy = (i.split())[1]\n x.append(tempx)\n y.append(int(tempy))\n\n fp.close()\n plt.ylim(1, 100)\n plt.title(\"Exam Score-2020\", fontsize=24)\n plt.xlabel(\"Name\", fontsize=14)\n plt.ylabel(\"Score\", fontsize=14)\n plt.plot(x, y)\n plt.show()\n return\n\nif __name__ == '__main__':\n get_garph()\n","repo_name":"allenshi86/Python","sub_path":"Graphical_data/Graphical_data_matplotlib.py","file_name":"Graphical_data_matplotlib.py","file_ext":"py","file_size_in_byte":572,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31695596875","text":"'''\nDescription: 逻辑回归\n视频:https://www.bilibili.com/video/BV1Y7411d7Ys?p=6\n博客: https://blog.csdn.net/bit452/article/details/109680909\nAuthor: HCQ\nCompany(School): UCAS\nEmail: 1756260160@qq.com\nDate: 2020-12-05 16:53:13\nLastEditTime: 2020-12-05 17:12:56\nFilePath: /pytorch/PyTorch深度学习实践/06逻辑回归.py\n'''\n\n\nimport torch\n# import torch.nn.functional as F\n \n# 1 prepare dataset\nx_data = torch.Tensor([[1.0], [2.0], [3.0]])\ny_data = torch.Tensor([[0], [0], [1]]) # =============================[0], [0], [1]]=========================\n \n# 2 design model using class\nclass LogisticRegressionModel(torch.nn.Module):\n def __init__(self):\n super(LogisticRegressionModel, self).__init__()\n self.linear = torch.nn.Linear(1,1)\n \n def forward(self, x):\n # y_pred = F.sigmoid(self.linear(x))\n y_pred = torch.sigmoid(self.linear(x)) # ========================sigmoid=============================\n return y_pred\nmodel = LogisticRegressionModel()\n \n# 3 construct loss and optimizer\n# 默认情况下,loss会基于element平均,如果size_average=False的话,loss会被累加。\ncriterion = torch.nn.BCELoss(size_average = False) # ==================BCELoss===========================\noptimizer = torch.optim.SGD(model.parameters(), lr = 0.01)\n \n# 4 training cycle forward, backward, update\nfor epoch in range(1000):\n y_pred = model(x_data)\n loss = criterion(y_pred, y_data)\n print(epoch, loss.item())\n \n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n \nprint('w = ', model.linear.weight.item())\nprint('b = ', model.linear.bias.item())\n \nx_test = torch.Tensor([[4.0]])\ny_test = model(x_test)\nprint('y_pred = ', y_test.data)\n\n\n\n# 可视化\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nx = np.linspace(0, 10, 200)\nx_t = torch.Tensor(x).view((200, 1))\ny_t = model(x_t) # 使用训练好的模型\ny = y_t.data.numpy()\nplt.plot(x, y)\nplt.plot([0, 10], [0.5, 0.5], c='r')\nplt.xlabel('Hours')\nplt.ylabel('Probability of Pass')\nplt.grid()\nplt.show()\n\n\n\n","repo_name":"HuangCongQing/pytorch","sub_path":"PyTorch深度学习实践/06逻辑回归.py","file_name":"06逻辑回归.py","file_ext":"py","file_size_in_byte":2045,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"79"} +{"seq_id":"42384295901","text":"def decrypt(library, message):\r\n \"\"\"\r\n return String for parameters\r\n library a list of Strings\r\n and message a string\r\n \"\"\"\r\n\r\n # Make lists of keys and values\r\n keys = []\r\n values = []\r\n for entry in library:\r\n entry = entry.split()\r\n keys.append(entry[1])\r\n values.append(entry[0])\r\n\r\n # Decode the message\r\n decoded = ''\r\n message = message.split()\r\n for m in range(len(message)):\r\n for k in range(len(keys)):\r\n if message[m] == keys[k]:\r\n decoded += values[k]\r\n if message[m] not in keys:\r\n decoded += '?'\r\n return decoded","repo_name":"Lihong062/CS101","sub_path":"APTs/APT-4/MorseLikeCode.py","file_name":"MorseLikeCode.py","file_ext":"py","file_size_in_byte":641,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27713867039","text":"import csv\nclass Product:\n\n def __init__(self, product_id, category):\n self.product_id = product_id\n self.category = category\n\n if any(v is None for v in [self.product_id, self.category]):\n raise ValueError('Arguments missing')\n\n def __str__(self):\n return \"Product: {}, Category: {}\".format(\n self.product_id,\n self.category\n )\n\n\ndef load_catalog_from_file(filename):\n catalog = {}\n with open(filename) as f:\n csvReader = csv.reader(f)\n for line in csvReader:\n product = Product(product_id=line[0], category=line[5])\n catalog[product.product_id] = product.category\n return catalog\n","repo_name":"AneliyaPPetkova/Programming","sub_path":"Python/7.SupermarketsDataAnalysis/product.py","file_name":"product.py","file_ext":"py","file_size_in_byte":705,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"73289684736","text":"# -*- coding: utf-8 -*-\n\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom img import img\nimport sys\nimport pytube as pt\nimport os\nimport moviepy.editor as mp\nimport subprocess\n\n\n\nclass Ui_Dialog(object):\n def setupUi(self, Dialog):\n Dialog.setObjectName(\"Dialog\")\n Dialog.resize(400, 286)\n Dialog.setMinimumSize(QtCore.QSize(400, 286))\n Dialog.setMaximumSize(QtCore.QSize(400, 286))\n self.label = QtWidgets.QLabel(Dialog)\n self.label.setGeometry(QtCore.QRect(0, -8, 401, 301))\n self.label.setStyleSheet(\"background-color: rgb(66, 66, 66);\")\n self.label.setText(\"\")\n self.label.setObjectName(\"label\")\n self.label_2 = QtWidgets.QLabel(Dialog)\n self.label_2.setGeometry(QtCore.QRect(0, 0, 401, 61))\n self.label_2.setStyleSheet(\"image: url(:/img/Settings2_00000.png);\")\n self.label_2.setText(\"\")\n self.label_2.setObjectName(\"label_2\")\n self.label_3 = QtWidgets.QLabel(Dialog)\n self.label_3.setGeometry(QtCore.QRect(10, 90, 141, 61))\n self.label_3.setStyleSheet(\"image: url(:/img/AdMusic_00000.png);\")\n self.label_3.setText(\"\")\n self.label_3.setObjectName(\"label_3\")\n self.label_4 = QtWidgets.QLabel(Dialog)\n self.label_4.setGeometry(QtCore.QRect(0, -40, 441, 371))\n self.label_4.setStyleSheet(\"image: url(:/img/settingsFundo_00000.png);\")\n self.label_4.setText(\"\")\n self.label_4.setObjectName(\"label_4\")\n self.lineEdit = QtWidgets.QLineEdit(Dialog)\n self.lineEdit.setGeometry(QtCore.QRect(140, 110, 211, 20))\n self.lineEdit.setStyleSheet(\"QLineEdit { \\n\"\n\" border-color: rgb(85, 0, 255);\\n\"\n\" background-color: rgb(222, 222, 222);\\n\"\n\" color: rgb(0, 0, 0);\\n\"\n\"}\\n\"\n\"QLineEdit:hover {\\n\"\n\" border:2px solid rgb(45, 45, 45);\\n\"\n\"}\\n\"\n\"QLineEdit:focus {\\n\"\n\" border:2px solid rgb(0, 0, 0);\\n\"\n\"\\n\"\n\"}\\n\"\n\"\")\n self.lineEdit.setObjectName(\"lineEdit\")\n self.pushButton = QtWidgets.QPushButton(Dialog)\n self.pushButton.setGeometry(QtCore.QRect(290, 140, 61, 31))\n self.pushButton.setStyleSheet(\"QPushButton {\\n\"\n\" image: url(:/img/Add_00000.png);\\n\"\n\" background-color: rgba(205, 205, 205, 0);\\n\"\n\" boder-radius: 5px;\\n\"\n\"}\\n\"\n\"QPushButton:hover{\\n\"\n\" \\n\"\n\" image: url(:/img/AddHigh_00000.png);\\n\"\n\"}\\n\"\n\"QPushButton:pressed{\\n\"\n\" image: url(:/img/Add_00000.png);\\n\"\n\"}\")\n self.pushButton.setText(\"\")\n self.pushButton.setObjectName(\"pushButton\")\n self.pushButton_2 = QtWidgets.QPushButton(Dialog)\n self.pushButton_2.setGeometry(QtCore.QRect(140, 210, 141, 71))\n self.pushButton_2.setStyleSheet(\"QPushButton {\\n\"\n\" \\n\"\n\" image: url(:/img/viewmusic_00000.png);\\n\"\n\" background-color: rgba(205, 205, 205, 0);\\n\"\n\" boder-radius: 5px;\\n\"\n\"}\\n\"\n\"QPushButton:hover{\\n\"\n\" image: url(:/img/viewmusic2_00000.png);\\n\"\n\"}\\n\"\n\"QPushButton:pressed{\\n\"\n\" image: url(:/img/viewmusic_00000.png);\\n\"\n\"}\")\n self.pushButton_2.setText(\"\")\n self.pushButton_2.setObjectName(\"pushButton_2\")\n\n self.retranslateUi(Dialog)\n QtCore.QMetaObject.connectSlotsByName(Dialog)\n\n self.pushButton.clicked.connect(self.add_music)\n self.pushButton_2.clicked.connect(self.view_music)\n\n\n def retranslateUi(self, Dialog):\n _translate = QtCore.QCoreApplication.translate\n Dialog.setWindowTitle(_translate(\"Dialog\", \"Dialog\"))\n self.lineEdit.setPlaceholderText(_translate(\"Dialog\", \"enter the URL of the song \"))\n\n def view_music(self):\n path = './musics'\n path = os.path.realpath(path)\n os.startfile(path)\n\n def add_music(self):\n try:\n # Download for Youtube\n try:\n os.mkdir('./musics')\n except:\n pass\n url = self.lineEdit.text()\n stream = pt.YouTube(url = url).streams.get_audio_only()\n stream.download()\n title = str(stream.title)\n\n # Converter of mp4 to mp3\n clip = mp.AudioFileClip(title + '.mp4')\n clip.write_audiofile(f'musics/'+ title + '.mp3')\n\n # Remove mp4\n os.remove(title + '.mp4')\n except:\n error = QtWidgets.QMessageBox()\n error.setWindowTitle(\"Error\")\n error.setIcon(QtWidgets.QMessageBox.Critical)\n error.setText(\"Unable to download, error found\")\n error.exec()\n else:\n msg = QtWidgets.QMessageBox()\n msg.setWindowTitle(\"Done\")\n msg.setIcon(QtWidgets.QMessageBox.Information)\n msg.setText(\"Your music was successfully downloaded\")\n msg.exec()\n\n def retranslateUi(self, Dialog):\n Dialog.setWindowIcon(QtGui.QIcon('img/settings_00000'))\n \n \n\n\n\n\nif __name__ == \"__main__\":\n import sys\n app = QtWidgets.QApplication(sys.argv)\n Dialog = QtWidgets.QDialog()\n ui = Ui_Dialog()\n ui.setupUi(Dialog)\n Dialog.show()\n app.exec_()","repo_name":"eumaninho54/MusicPlayer","sub_path":"settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":5159,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"20507478087","text":"#!/usr/bin/env python3\r\n\r\nfrom priority_queue_112 import PQ\r\nimport sys\r\nm = int(sys.argv[1])\r\n\r\ndef desc(t):\r\n return t[1]\r\n\r\ndef main():\r\n queue = PQ()\r\n # read first m elements into the queue\r\n i = 0\r\n while i < m:\r\n num = int(sys.stdin.readline().rstrip())\r\n queue.insert(num)\r\n\r\n i += 1\r\n\r\n # get the minimum M numbers from stdin\r\n for element in sys.stdin:\r\n element = int(element.rstrip())\r\n if element < queue.getMax():\r\n queue.insert(element)\r\n queue.delMax()\r\n\r\n while not(queue.is_empty()):\r\n # print(queue.d, queue.N)\r\n print(queue.delMax())\r\n\r\n\r\nif __name__ == '__main__':\r\n main()\r\n","repo_name":"ClassicRevive/labwork","sub_path":"ca117/lab20/minelts_112.py","file_name":"minelts_112.py","file_ext":"py","file_size_in_byte":694,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"72603076735","text":"from microquake.core.stream import Trace, Stream\nfrom microquake.core.settings import settings\nimport numpy as np\nfrom obspy.core import UTCDateTime\nfrom loguru import logger\nfrom microquake.db.models.alchemy import ContinuousData\nfrom microquake.db.connectors import connect_timescale\nfrom datetime import datetime\nfrom sqlalchemy import desc\nfrom pytz import utc\n\n\ndef get_continuous_data(start_time, end_time, sensor_id=None):\n\n db_lag = get_db_lag()\n logger.info(f'the database lag is {db_lag} seconds')\n\n if type(start_time) is datetime:\n start_time = UTCDateTime(start_time)\n\n if type(end_time) is datetime:\n end_time = UTCDateTime(end_time)\n\n session, engine = connect_timescale()\n inventory = settings.inventory\n\n e_time = end_time.datetime\n s_time = start_time.datetime\n\n network_code = inventory.networks[0].code\n\n t = ContinuousData.time\n et = ContinuousData.end_time\n sid = ContinuousData.sensor_id\n\n if sensor_id is None:\n logger.info('requesting data for all sensors')\n cds = session.query(ContinuousData).filter(t <= e_time,\n et > s_time)\n else:\n if inventory.select(sensor_id) is None:\n logger.error(f'the sensor {sensor_id} is not in the inventory')\n\n return\n logger.info(f'requesting data for sensor {sensor_id}')\n cds = session.query(ContinuousData).filter(t <= e_time,\n et > s_time,\n sid == sensor_id)\n\n traces = []\n for cd in cds:\n x = np.array(cd.x)\n y = np.array(cd.y)\n z = np.array(cd.z)\n tr_x = Trace(data=x)\n tr_x.stats.starttime = UTCDateTime(cd.time)\n tr_x.stats.sampling_rate = cd.sample_rate\n tr_x.stats.channel = 'X'\n tr_x.stats.station = str(cd.sensor_id)\n tr_x.stats.network = network_code\n traces.append(tr_x)\n tr_y = Trace(data=y)\n tr_y.stats.starttime = UTCDateTime(cd.time)\n tr_y.stats.sampling_rate = cd.sample_rate\n tr_y.stats.channel = 'Y'\n tr_y.stats.station = str(cd.sensor_id)\n tr_y.stats.network = network_code\n traces.append(tr_y)\n tr_z = Trace(data=z)\n tr_z.stats.starttime = UTCDateTime(cd.time)\n tr_z.stats.sampling_rate = cd.sample_rate\n tr_z.stats.channel = 'Z'\n tr_z.stats.station = str(cd.sensor_id)\n tr_z.stats.network = network_code\n traces.append(tr_z)\n\n time_now = UTCDateTime.now()\n delay = time_now - end_time\n\n st = Stream(traces=traces).trim(starttime=start_time, endtime=end_time)\n\n if st is None:\n logger.warning(f'no data recovered from the database! '\n f'the current database lag is {db_lag}')\n return None\n\n duration = (end_time - start_time)\n\n trs = []\n # st = st.merge(fill_value=np.nan)\n for i, tr in enumerate(st):\n expected_number_sample = tr.stats.sampling_rate * duration\n if np.all(tr.data == 0):\n logger.warning(f'data from sensor {tr.stats.station} contains '\n f'only zero. The trace will not be kept')\n continue\n elif np.any(np.isnan(tr.data)):\n logger.warning(f'data from sensor {tr.stats.station} contains '\n f'some NaN. The trace will not be kept')\n continue\n elif len(tr) < 0.9 * expected_number_sample:\n logger.warning(f'data from sensor {tr.stats.station} contains '\n f'too little data. The trace will not be kepy')\n continue\n\n trs.append(tr)\n\n if not trs:\n return None\n\n st = Stream(traces=trs).trim(starttime=start_time,\n endtime=end_time)\n\n session.close()\n engine.dispose()\n\n return st.detrend('demean')\n\n\ndef get_db_lag(percentile=75):\n \"\"\"\n returns lag in seconds\n :param percentile: percentile of data to use to determine the delay.\n For instance, if percentile = 75, 75 percent of the trace will have a lag\n of less than the returned value.\n :return: lag in second\n \"\"\"\n\n session, engine = connect_timescale()\n\n inventory = settings.inventory\n t = ContinuousData.time\n sensor_id = ContinuousData.sensor_id\n\n times = []\n for sensor in inventory.stations():\n\n records = session.query(t, sensor_id).filter(\n sensor_id == sensor.code).order_by(desc(t)).limit(1)\n\n for record in records:\n times.append(record.time.timestamp())\n\n if not times:\n return None\n\n time = datetime.utcfromtimestamp(np.percentile(times, percentile))\n\n lag = datetime.utcnow().replace(tzinfo=utc) - time.replace(tzinfo=utc)\n\n session.close()\n engine.dispose()\n\n return lag.total_seconds()\n\n\n","repo_name":"jeanphilippemercier/microquake","sub_path":"microquake/core/helpers/timescale_db.py","file_name":"timescale_db.py","file_ext":"py","file_size_in_byte":4880,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"10638349506","text":"#!/usr/bin/ennv python\n\nimport json\n\n\nwith open('../data/categories.json', 'r') as f:\n categories = json.load(f)\n\ntop_level_categories = [c for c in categories\n if (c['parents'] == []) and ('US' in c.get('country_whitelist', ['US']))]\n","repo_name":"philpot/yelp91030","sub_path":"src/categories.py","file_name":"categories.py","file_ext":"py","file_size_in_byte":261,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"2151501726","text":"#load model\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.keras.preprocessing.text import Tokenizer\nfrom tensorflow.keras.preprocessing.sequence import pad_sequences\nimport model as m\n\n#load model\nmodel = tf. keras.models.load_model('saved_model/TGA_model')\n\n# Check its architecture\nmodel.summary()\n\n\n# make predictions\ndef inputPrediction(inputText):\n\n # import tokenizer from previous model.py\n _, tokenizer = m.Tokenization()\n pred = ''\n confidence = 0.0\n inputList = [inputText]\n for i in range(0, len(inputList)):\n max_length = 100\n trunc_type = 'post'\n sequence = tokenizer.texts_to_sequences(inputList)\n padded = pad_sequences(sequence, maxlen = max_length, truncating = trunc_type)\n output = model.predict(padded)\n if (output[i][0] <= 0.5):\n print(output[i])\n pred = 'negative'\n confidence += output[i][0] \n else:\n print(output[i])\n pred = 'positive'\n confidence += output[i][0]\n\n print('Review: ' + inputList[i] + '\\n' + 'Sentiment: ' + pred + ' ' + str(output[i][0]) + '\\n' + '\\n')\n return pred, confidence \n\nsample = \"happy good wonderful\"\n\ninputPrediction(sample)","repo_name":"mshah016/app","sub_path":"appTEST/prediction.py","file_name":"prediction.py","file_ext":"py","file_size_in_byte":1243,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"10271301860","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 8 21:29:47 2018\n\n@author: Chinmay\n\"\"\"\n\nimport numpy as np\nimport gdal, copy\nfrom gdalconst import * \nimport matplotlib.pyplot as plt\n\n# input path\ninputpath = 'D:/Chinmay/Pune/Analysis_06_06_2018/NITK_RSGIS_20180824_154257/Outputs/LT05_L1TP_147047_19900318_20170131_01_T1.tar/LT5[147_047](1990-03-18_04-48)'\n\n\n# read spectral indices into numpy arrays\nndvidata = gdal.Open(inputpath+'NDVI.TIF', GA_ReadOnly)\nNDVI = np.array(ndvidata.GetRasterBand(1).ReadAsArray())\nndwidata = gdal.Open(inputpath+'NDWI.TIF', GA_ReadOnly)\nNDWI = np.array(ndwidata.GetRasterBand(1).ReadAsArray())\nndbidata = gdal.Open(inputpath+'NDBI_01.TIF', GA_ReadOnly)\nNDBI = np.array(ndbidata.GetRasterBand(1).ReadAsArray())\nndbaidata = gdal.Open(inputpath+'NDBaI_01.TIF', GA_ReadOnly)\nNDBAI = np.array(ndbaidata.GetRasterBand(1).ReadAsArray())\nmndwidata = gdal.Open(inputpath+'MNDWI_01.TIF', GA_ReadOnly)\nMNDWI = np.array(mndwidata.GetRasterBand(1).ReadAsArray())\n\n\n# create dictionary of thhresholds to be used for each spectral index to be classified \\\n#into a thematic raster class\n\nthresholds = {\"Wat_MNDWI\": 0.3,\n \"Veg_NDVI\" : 0.3,\n \"Veg_NDWI1\": 0.0,\n \"Veg_NDWI2\": 0.3,\n \"Veg_NDBI1\" : -0.02,\n \"Veg_NDBI2\" : -0.2,\n \"Veg_NDBaI1\": -0.3,\n \"Veg_NDBaI2\": -0.6,\n \"Sed_NDBaI1\": -0.6,\n \"Sed_NDBaI2\": 0.75,\n \"Built_NDVI\": 0.02,\n \"Built_NDWI\": 0.2,\n \"Built_NDBI\": 0.15,\n \"Built_NDBaI\": -0.25,\n \"Fallow_NDVI\": 0.2,\n \"Fallow_NDWI1\": -0.1,\n \"Fallow_NDWI2\": -0.25,\n \"Fallow_NDBI1\": 0.1,\n \"Fallow_NDBI2\": 0.3,\n \"Fallow_NDBaI\": -0.3\n }\n\n \n\ndef lulc_classify(NDVI, NDBI, NDBAI, NDWI, MNDWI):\n \"\"\"\n \n \"\"\"\n #create empty raster of same size as others\n Classified_raster = np.zeros(NDVI.shape)\n #Water\n #Classified_raster[np.where(MNDWI>=thresholds[\"Wat_MNDWI\"])] = 1\n \n \n # Sediment/river bed\n Classified_raster[np.where(np.logical_and(NDBAI>= thresholds[\"Sed_NDBaI2\"], \\\n NDBAI <= thresholds[\"Sed_NDBaI1\"]))] = 2 \n \n #Builtup area\n Classified_raster[np.where(np.logical_and(NDVI < thresholds[\"Built_NDVI\"] , \\\n NDWI < thresholds[\"Built_NDWI\"]) & \\\n np.logical_and(NDBI>=thresholds[\"Built_NDBI\"], NDBAI < thresholds[\"Built_NDBaI\"]))]=3\n \n return Classified_raster\n\n \n# # Vegetation\n# \n# Classified_raster[np.where(np.logical_and(NDWI>thresholds[\"Veg_NDWI1\"], NDWI < thresholds[\"Veg_NDWI2\"]) & \\\n# NDVI> thresholds[\"Veg_NDVI\"] & \\\n# (np.logical_and(NDBI > [\"Veg_NDBI2\"], NDBI < [\"Veg_NDBI1\"])) & \\\n# (np.logical_and(NDBAI > [\"Veg_NDBaI2\"], NDBI < [\"Veg_NDBaI1\"])))] = 3\n\n\n Classified_raster[np.where(NDVI < thresholds[\"Built_NDVI\"] & \\\n NDWI < thresholds[\"Built_NDWI\"] & \\\n NDBI >= thresholds[\"Built_NDBI\"] & \\\n NDBAI < thresholds[\"Built_NDBaI\"])] = 4\n \n# #Fallow land\n Classified_raster[np.where(np.logical_and(NDBI>thresholds[\"Fallow_NDBI1\"], NDBI <= thresholds[\"Fallow_NDBI2\"]) &\\\n NDVI < thresholds[\"Fallow_NDVI\"] & \\\n np.logical_and(NDWI >= thresholds[\"Fallow_NDWI2\"], NDWI < thresholds[\"Fallow_NDBI1\"]) & \\\n NDBAI < thresholds[\"Fallow_NDBaI\"])] = 5 \n# \n \n \n \n return Classified_raster\n\n\n ","repo_name":"devalc/Dtree_Lulc","sub_path":"classify_spectral_indices_into_lulc.py","file_name":"classify_spectral_indices_into_lulc.py","file_ext":"py","file_size_in_byte":3775,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"13452920805","text":"from core.item import ItemEnum\n\n\nclass Room:\n \"\"\"\n A class representing a room in the map.\n\n Args:\n name (str): The name of the room.\n description (str): The description of the room.\n \"\"\"\n\n def __init__(self, name: str, description: str):\n self.name = name\n self.description = description\n\n\nclass Map:\n \"\"\"\n A class representing the game map.\n \"\"\"\n\n map = {\n \"Starting Room\": {\n \"north\": \"North Hallway\",\n \"south\": None,\n \"east\": None,\n \"west\": None,\n },\n \"North Hallway\": {\n \"north\": None,\n \"south\": \"Starting Room\",\n \"east\": \"North East Wing\",\n \"west\": \"North West Wing\",\n },\n \"North West Wing\": {\n \"north\": None,\n \"south\": None,\n \"east\": \"North Hallway\",\n \"west\": None,\n },\n \"North East Wing\": {\n \"north\": None,\n \"south\": \"Hidden Room\",\n \"east\": None,\n \"west\": \"North Hallway\",\n },\n \"Hidden Room\": {\n \"north\": \"North East Wing\",\n \"south\": None,\n \"east\": None,\n \"west\": None,\n },\n }\n\n danger_levels = {\n \"Starting Room\": 0, # No danger\n \"North Hallway\": 0, # No danger\n \"North West Wing\": 20, # Moderate danger\n \"North East Wing\": 50, # High danger\n \"Hidden Room\": 80, # Singnificantly High danger\n }\n\n room_danger_message = {\n \"Starting Room\": \"There is no danger here.\",\n \"North Hallway\": \"There is no danger here.\",\n \"North West Wing\": \"\\nThere is poisonous gas here, and you inhaled it. Now your health is significantly reduced.\", # Moderate danger\n \"North East Wing\": \"\\nA zombie appeared out of nowhere! He's attacking you, do something!\", # High danger\n \"Hidden Room\": \"\\nA witch is here! She cast a spell on you, which is making your skin burn!!\", # Singnificantly High danger\n }\n\n room_enemy = {\n \"Starting Room\": None,\n \"North Hallway\": None,\n \"North West Wing\": \"Poisonous gas\",\n \"North East Wing\": \"Zombie\",\n \"Hidden Room\": \"Witch\",\n }\n\n items = {\n \"North Hallway\": {\"Medicine\": ItemEnum.MEDICINE},\n \"North West Wing\": {\"Sword\": ItemEnum.SWORD},\n \"North East Wing\": {\"Wand\": ItemEnum.WAND},\n \"Hidden Room\": {\"Key\": ItemEnum.KEY},\n }\n\n load_room = {\n \"Starting Room\": Room(\n \"Starting Room\", \"You find yourself in a dimly lit room.\"\n ),\n \"North Hallway\": Room(\n \"North Hallway\",\n \"You are in a long hallway with doors to the west and east.\",\n ),\n \"North East Wing\": Room(\n \"North East Wing\", \"You sense something mysterious in this room.\"\n ),\n \"Hidden Room\": Room(\n \"Hidden Room\", \"Congratulations! You found the hidden treasure room!\"\n ),\n \"North West Wing\": Room(\n \"North West Wing\", \"This room seems to be abandoned and dusty.\"\n ),\n }\n","repo_name":"akshitverma4/Mystic-Maze","sub_path":"core/map.py","file_name":"map.py","file_ext":"py","file_size_in_byte":3106,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"20561262020","text":"import pandas as pd\nfrom sklearn import preprocessing\nimport numpy as np\nnp.warnings.filterwarnings('ignore', category=np.VisibleDeprecationWarning)\nhistory_points = 50\nfrom db import Db\nimport sys\n\n\ndef get_raw_data(symbol,print_sample=True):\n \n model = Db()\n records = model.get_data(symbol)\n \n data = pd.DataFrame(data = records, \n columns = ['open_val','high_val','low_val','close_val','volume'])\n data = data.drop(0, axis=0)\n if print_sample:\n print(data.head(5))\n print(data.count())\n return data.to_numpy()\n\n\ndef create_dataset(symbol):\n data = get_raw_data(symbol)\n data_normaliser = preprocessing.MinMaxScaler()\n data_normalised = data_normaliser.fit_transform(data)\n ohlcv_histories_normalised = get_ohlcv_histories_normalised(data_normalised)\n\n next_day_open_values_normalised = np.array(\n [data_normalised[:, 0][i + history_points].copy() for i in range(len(data_normalised) - history_points)])\n next_day_open_values_normalised = np.expand_dims(next_day_open_values_normalised, -1)\n\n next_day_open_values = get_next_day_open_values(data)\n\n y_normaliser = get_y_normaliser(next_day_open_values)\n\n technical_indicators_normalised = get_technical_indicators(ohlcv_histories_normalised)\n\n assert ohlcv_histories_normalised.shape[0] == next_day_open_values_normalised.shape[0] == \\\n technical_indicators_normalised.shape[0]\n return ohlcv_histories_normalised, technical_indicators_normalised, next_day_open_values_normalised, \\\n next_day_open_values, y_normaliser\n\ndef get_y_normaliser(data):\n y_normaliser = preprocessing.MinMaxScaler()\n y_normaliser.fit(data)\n return y_normaliser\n\ndef get_next_day_open_values(data):\n next_day_open_values = np.array([data[:, 0][i + history_points].copy() for i in range(len(data) - history_points)])\n next_day_open_values = np.expand_dims(next_day_open_values, -1)\n return next_day_open_values\n\ndef get_ohlcv_histories_normalised(data_normalised, last=0):\n # using the last {history_points} open close high low volume data points, predict the next open value\n if (last == 0):\n rng = range(len(data_normalised) - history_points)\n else:\n rng = range(len(data_normalised) - history_points + 1 - last, len(data_normalised) - history_points + 2)\n return np.array(\n [data_normalised[i:i + history_points].copy() for i in rng])\n\ndef get_next_day_open_values_normalised(data_normalised, last=0):\n # using the last {history_points} open close high low volume data points, predict the next open value\n if (last == 0):\n rng = range(len(data_normalised) - history_points)\n else:\n rng = range(len(data_normalised) - history_points + 1 - last, len(data_normalised) - history_points + 2)\n return np.array(\n [data_normalised[i:i + history_points].copy() for i in rng])\n\n\ndef get_technical_indicators(ohlcv_histories_normalised):\n technical_indicators = []\n for his in ohlcv_histories_normalised:\n # note since we are using his[3] we are taking the SMA of the closing price\n sma = np.mean(his[:, 3])\n macd = calc_ema(his, 12) - calc_ema(his, 26)\n technical_indicators.append(np.array([sma]))\n # technical_indicators.append(np.array([sma,macd,]))\n\n technical_indicators = np.array(technical_indicators)\n\n tech_ind_scaler = preprocessing.MinMaxScaler()\n return tech_ind_scaler.fit_transform(technical_indicators)\n\n\ndef calc_ema(values, time_period):\n sma = np.mean(values[:, 3])\n ema_values = [sma]\n k = 2 / (1 + time_period)\n for i in range(len(values) - time_period, len(values)):\n close = values[i][3]\n ema_values.append(close * k + ema_values[-1] * (1 - k))\n return ema_values[-1]\n\"\"\"\n\ndef multiple_csv_to_dataset(test_set_name):\n import os\n ohlcv_histories = 0\n technical_indicators = 0\n next_day_open_values = 0\n for csv_file_path in list(filter(lambda x: x.endswith('daily.csv'), os.listdir('./'))):\n if not csv_file_path == test_set_name:\n print(csv_file_path)\n if type(ohlcv_histories) == int:\n ohlcv_histories, technical_indicators, next_day_open_values, _, _ = csv_to_dataset(csv_file_path)\n else:\n a, b, c, _, _ = csv_to_dataset(csv_file_path)\n ohlcv_histories = np.concatenate((ohlcv_histories, a), 0)\n technical_indicators = np.concatenate((technical_indicators, b), 0)\n next_day_open_values = np.concatenate((next_day_open_values, c), 0)\n\n ohlcv_train = ohlcv_histories\n tech_ind_train = technical_indicators\n y_train = next_day_open_values\n\n ohlcv_test, tech_ind_test, y_test, unscaled_y_test, y_normaliser = csv_to_dataset(test_set_name)\n\n return ohlcv_train, tech_ind_train, y_train, ohlcv_test, tech_ind_test, y_test, unscaled_y_test, y_normaliser\n\"\"\"","repo_name":"shlomitub28/cod-stock-demo","sub_path":"util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":4921,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"73487787456","text":"import numpy as np\nimport pandas as pd\nfrom scipy import stats\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.decomposition import PCA\n\ndef moving_avg(df, n, step):\n \"\"\"\n moving avg with step size\n (Did not use pandas.rolling since step size not supported.)\n Input:\n df: Input dataframe. First column (index 0) is time stamp.\n n: window length\n step: step size of the moving average\n Return:\n data: 2D array with features, has less number of rows than input df. \n \"\"\"\n T = df.shape[0]\n new_t = range(0, T, step)\n data = np.zeros((len(new_t), df.shape[1]))\n\n for i, t in enumerate(new_t):\n data[i, :] = np.nanmean(df.iloc[t: t + n, :].values, axis=0) # ignore nan when computing mean.\n # discard the last window\n return data[:-1, :]\n\ndef moving_vote_majority(label_df, n, step):\n \"\"\"\n Take majority vote of labels in the moving window.\n Input:\n label_df: dataframe with one column denoting the label stream.\n n: window size.\n step: step size of the moving window\n Return:\n labels: 1D array with labels.\n \"\"\"\n T = label_df.shape[0]\n new_t = range(0, T, step)\n labels = np.zeros((len(new_t)))\n for i, t in enumerate(new_t):\n labels[i] = stats.mode(label_df.iloc[t: t + n].values)[0][0]\n return labels[:-1]\n\ndef fill_missing(training, test, col_threshold, replace):\n \"\"\"\n training: 2D array\n test: 2D array or None. If test is None, only process training data.\n col_threshold: if nan values in one column is greater than col_threshold, the column is ignored.\n replace: True: replace nan value in each row. False: delete the row if it contains nan.\n \"\"\"\n # delete nan columns\n training = training[:, np.sum(np.isnan(training), axis=0) < col_threshold * training.shape[0]]\n if test is not None:\n test = test[:, np.sum(np.isnan(test), axis=0) < col_threshold * test.shape[0]]\n\n # process nans in each row\n if not replace:\n training = training[np.logical_not(np.any(np.isnan(training), axis=1)), :]\n if test is not None:\n test = test[np.logical_not(np.any(np.isnan(test), axis=1)), :]\n else:\n # replace with previous values\n for i in range(0, training.shape[0]):\n if i == 0:\n training[i, np.isnan(training[i, :])] = 0.0\n else:\n training[i, np.isnan(training[i, :])] = training[i - 1, np.isnan(training[i, :])]\n # remove the column with zero variance\n var_train = np.var(training, axis=0)\n logical_mask = var_train > 0\n training = training[:, logical_mask]\n \n if test is not None:\n test = test[:, logical_mask]\n mean_train = np.mean(training, axis=0)\n for i in range(0, test.shape[0]):\n if i == 0:\n # fill in the mean of training data\n test[i, np.isnan(test[i, :])] = mean_train[np.isnan(test[i, :])]\n else:\n test[i, np.isnan(test[i, :])] = test[i - 1, np.isnan(test[i, :])]\n \n return training, test\n\ndef extract_feature_per_person(person):\n \"\"\"\n For each person in the dataset, smooth data, fill missing values and perform dimension reduction.\n Use ADL1~3 for training, ADL4~5 for testing.\n Return:\n train_reduced: [n_train, n_features]. Sequence of training feature vectors.\n test_reduced: [n_test, n_features]. Sequence of testing feature vectors.\n train_labels: [n_train, ]. Sequence of training activity labels in {0, 101, 102, 103, 104, 105}\n test_labels: [n_test, ]. Sequence of testing activity labels in {0, 101, 102, 103, 104, 105}\n train_len: length of subsequences in training data. sum(train_len) = n_train.\n test_len: length of subsequences in testing data. sum(test_len) = n_test.\n \"\"\"\n # Load data\n sadl_n = []\n for n in range(1, 6):\n sadl_n.append(pd.read_table('data/S%d-ADL%d.dat' % (person, n), sep='\\s+', header=None, dtype=float))\n\n # Smooth data, time: col 0, features: col 1~36, labels: col 244 \n winsize = 15\n stepsize = 8\n # train data\n train_sample = np.empty((0, 36))\n train_labels = np.empty((0))\n train_len = []\n for i in range(0, 3):\n features = moving_avg(sadl_n[i].iloc[:, 1:37], winsize, stepsize)\n labels = moving_vote_majority(sadl_n[i].iloc[:, 244], winsize, stepsize)\n train_sample = np.concatenate((train_sample, features), axis=0)\n train_len.append(features.shape[0])\n train_labels = np.concatenate( (train_labels, labels) )\n train_len = np.array(train_len)\n # test data\n test_sample = np.empty((0, 36))\n test_labels = np.empty((0))\n test_len = []\n for i in range(3, 5):\n features = moving_avg(sadl_n[i].iloc[:, 1:37], winsize, stepsize)\n labels = moving_vote_majority(sadl_n[i].iloc[:, 244], winsize, stepsize)\n test_sample = np.concatenate((test_sample, features), axis=0)\n test_len.append(features.shape[0])\n test_labels = np.concatenate( (test_labels, labels) )\n test_len = np.array(test_len)\n\n # Fill missing values\n col_threshold = 0.5\n train, test = fill_missing(train_sample, test_sample, col_threshold, True)\n\n # Normalize features\n scalar = StandardScaler() # center to mean and normalize to unit variance\n train_normalized = scalar.fit_transform(train)\n test_normalized = scalar.fit_transform(test)\n\n # Dimension reduction\n pca = PCA()\n pca.fit(train_normalized)\n var_thres = 0.95 # keep components to up to 95% total variance\n n_comp = (pca.explained_variance_ratio_.cumsum() < var_thres).sum() + 1\n\n pca_train = PCA(n_components=n_comp)\n train_reduced = pca_train.fit_transform(train_normalized)\n test_reduced = pca_train.transform(test_normalized)\n\n return train_reduced, test_reduced, train_labels, test_labels, train_len, test_len\n \n","repo_name":"flian2/human-activity-recognition-hmm","sub_path":"clean_data.py","file_name":"clean_data.py","file_ext":"py","file_size_in_byte":5916,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"79"} +{"seq_id":"15105360905","text":"from choice import *\r\nfrom graphs import graph\r\nfrom algorithms import *\r\n\r\nif __name__ == '__main__':\r\n workflow = workflow()\r\n f = fun()\r\n d, e = int, float\r\n if workflow == 1:\r\n d = polynomial_degree()\r\n else:\r\n e = approximation_error()\r\n n = nodes_number()\r\n start, end = interval()\r\n coefficient = []\r\n weights, roots = laguerre(n)\r\n\r\n if workflow == 1:\r\n for i in range(0, d + 1):\r\n coefficient.append(lam(f, n, i, weights, roots))\r\n print(\"Blad aproksymacji wynosi: \" + str(error(roots, weights, f, coefficient, n, d)))\r\n\r\n else:\r\n d = 1\r\n while True:\r\n coefficient = []\r\n for i in range(0, d + 1):\r\n coefficient.append(lam(f, n, i, weights, roots))\r\n if error(roots, weights, f, coefficient, n, d) <= e:\r\n print(f\"Oczekiwany blad udalo sie uzyskac dla wielomiana stopnia: {d}\")\r\n break\r\n else:\r\n d += 1\r\n\r\n calculated_x, calculated_y, real_x, real_y = [], [], [], []\r\n h = (end - start) / 100\r\n while start <= end:\r\n result = 0.0\r\n for j in range(0, d + 1):\r\n result += coefficient[j] * laguerre_polynomial(j, start)\r\n calculated_x.append(start)\r\n calculated_y.append(result)\r\n real_x.append(start)\r\n real_y.append(f(start))\r\n start += h\r\n\r\n graph(calculated_x, calculated_y, real_x, real_y)\r\n","repo_name":"stepniaczky/NumericalMethods","sub_path":"laguerre_polynomials_approximation/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1464,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"20228097454","text":"import numpy as np\nimport flopy\nimport numpy as np\nimport flopy.utils.binaryfile as bf\nimport matplotlib.pyplot as plt\n\n\nLx = 2.0\nLz = 1.0\nnlay = 50\nnrow = 1\nncol = 100\ndelr = Lx / ncol\ndelc = 1.0\ndelv = Lz / nlay\nhenry_top = 1.0\nhenry_botm = np.linspace(henry_top - delv, 0.0, nlay)\nqinflow = 5.702 # m3/day\ndmcoef = 0.57024 # m2/day Could also try 1.62925 as another case of the Henry problem\nhk = 864.0 # m/day\n\nmodelname = \"henry\"\nswt = flopy.seawat.Seawat(modelname, exe_name=r\"C:\\Users\\ccl124\\bin\\swt_v4x64.exe\")\nprint(swt.namefile)\n\nipakcb = 53\n\ndis = flopy.modflow.ModflowDis(\n swt,\n nlay,\n nrow,\n ncol,\n nper=1,\n delr=delr,\n delc=delc,\n laycbd=0,\n top=henry_top,\n botm=henry_botm,\n perlen=1.5,\n nstp=15,\n)\n\n# Variables for the BAS package\nibound = np.ones((nlay, nrow, ncol), dtype=np.int32)\nibound[:, :, -1] = -1\n\nbas = flopy.modflow.ModflowBas(swt, ibound, 0)\n\nlpf = flopy.modflow.ModflowLpf(swt, hk=hk, vka=hk, ipakcb=ipakcb)\n\npcg = flopy.modflow.ModflowPcg(swt, hclose=1.0e-8)\n\noc = flopy.modflow.ModflowOc(\n swt,\n stress_period_data={(0, 0): [\"save head\", \"save budget\"]},\n compact=True,\n)\n\nitype = flopy.mt3d.Mt3dSsm.itype_dict()\nwel_data = {}\nssm_data = {}\nwel_sp1 = []\nssm_sp1 = []\nfor k in range(nlay):\n wel_sp1.append([k, 0, 0, qinflow / nlay])\n ssm_sp1.append([k, 0, 0, 0.0, itype[\"WEL\"]])\n ssm_sp1.append([k, 0, ncol - 1, 35.0, itype[\"BAS6\"]])\nwel_data[0] = wel_sp1\nssm_data[0] = ssm_sp1\nwel = flopy.modflow.ModflowWel(swt, stress_period_data=wel_data, ipakcb=ipakcb)\n\nbtn = flopy.mt3d.Mt3dBtn(\n swt,\n nprs=-5,\n prsity=0.35,\n sconc=35.0,\n ifmtcn=0,\n chkmas=False,\n nprobs=10,\n nprmas=10,\n dt0=0.001,\n\n)\nadv = flopy.mt3d.Mt3dAdv(swt, mixelm=0)\ndsp = flopy.mt3d.Mt3dDsp(swt, al=0.0, trpt=1.0, trpv=1.0, dmcoef=dmcoef)\ngcg = flopy.mt3d.Mt3dGcg(swt, iter1=500, mxiter=1, isolve=1, cclose=1e-7)\nssm = flopy.mt3d.Mt3dSsm(swt, stress_period_data=ssm_data)\n\nvdf = flopy.seawat.SeawatVdf(\n swt,\n iwtable=0,\n densemin=0,\n densemax=0,\n denseref=1000.0,\n denseslp=0.7143,\n firstdt=1e-3,\n)\n\nswt.write_input()\n\nsuccess, buff = swt.run_model(silent=True, report=True)\nif not success:\n raise Exception(\"SEAWAT did not terminate normally.\")\n\nucnobj = bf.UcnFile(\"MT3D001.UCN\", model=swt)\ntimes = ucnobj.get_times()\nconcentration = ucnobj.get_data(totim=times[-1])\n\ncbbobj = bf.CellBudgetFile(\"henry.cbc\")\ntimes = cbbobj.get_times()\nqx = cbbobj.get_data(text=\"flow right face\", totim=times[-1])[0]\nqy = np.zeros((nlay, nrow, ncol), dtype=float)\nqz = cbbobj.get_data(text=\"flow lower face\", totim=times[-1])[0]\n\nfig = plt.figure(figsize=(12,9))\nax = fig.add_subplot(1, 1, 1, aspect=\"equal\")\npmv = flopy.plot.PlotCrossSection(model=swt, ax=ax, line={\"row\": 0})\narr = pmv.plot_array(concentration)\npmv.plot_vector(qx, qy, -qz, color=\"white\", kstep=3, hstep=3)\nplt.colorbar(arr, shrink=0.5, ax=ax)\nax.set_title(\"Simulated Concentrations\");\nplt.show()\n\nheadobj = bf.HeadFile(\"henry.hds\")\ntimes = headobj.get_times()\nhead = headobj.get_data(totim=times[-1])\n\nfig = plt.figure(figsize=(12, 9))\nax = fig.add_subplot(1, 1, 1, aspect=\"equal\")\npmv = flopy.plot.PlotCrossSection(model=swt, ax=ax, line={\"row\": 0})\narr = pmv.plot_array(head)\ncontours = pmv.contour_array(head, colors=\"white\")\nax.clabel(contours, fmt=\"%2.2f\")\nplt.colorbar(arr, shrink=0.5, ax=ax)\nax.set_title(\"Simulated Heads\");\nplt.show()","repo_name":"connorcleary/code_cdrive","sub_path":"swgw/scripts/henry_seawat.py","file_name":"henry_seawat.py","file_ext":"py","file_size_in_byte":3405,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"13480875285","text":"from .Describer import Describer\n\n\nclass GeneralDescriber(Describer):\n TEMP_SKEW_THRESHOLD = 0.4\n\n def get_section_header(self):\n return \"General frame info\"\n\n def describe(self, df):\n messages = [f\"Shape: {df.shape}\"]\n\n # NULLS\n columns_with_nulls = df.columns[df.isna().any()].array.to_numpy()\n if len(columns_with_nulls) > 0:\n messages.append(\n f\"Columns with one or more null values: {columns_with_nulls}\"\n )\n\n return messages\n","repo_name":"jammerware/pandashape","sub_path":"src/pandashape/describers/GeneralDescriber.py","file_name":"GeneralDescriber.py","file_ext":"py","file_size_in_byte":519,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"12401111979","text":"from functions1 import sigmoid\nimport numpy as np\nimport pandas_ex01 as pd\nimport matplotlib.pyplot as plt\n\ndata = [[2, 0], [4, 0], [6, 0], [8, 1], [10, 1], [12, 1], [14, 1]]\n\nx_data = [i[0] for i in data]\ny_data = [i[1] for i in data]\n\nplt.scatter(x_data, y_data)\nplt.xlim(0, 15)\nplt.ylim(-0.1, 1.1)\n# plt.show()\n\na, b = 0, 0\nlr = 0.05\n\nfor i in range(2001):\n for x_data, y_data in data:\n a_diff = x_data*(sigmoid(a * x_data + b) - y_data)\n b_diff = sigmoid(a * x_data + b) - y_data\n\n a = a - (lr * a_diff)\n b = b - (lr * b_diff)\n\n if i % 100 == 0:\n print(f'epoch={i}, 기울기={a:.04f}, 절편={b:.04f}')\n\nplt.scatter(x_data, y_data)\nplt.xlim(0, 15)\nplt.ylim(-0.1, 1.1)\nx_range = (np.arange(0, 15, 0.1))\nplt.plot(x_range, np.array([sigmoid(a * x + b) for x in x_range]))\nplt.show()\n","repo_name":"kimhyeongju/coding_practice","sub_path":"DeepLearning/logisticRegression.py","file_name":"logisticRegression.py","file_ext":"py","file_size_in_byte":826,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"15195198838","text":"from transaction import Transactions\n\n\nif __name__ == \"__main__\":\n transaction = Transactions()\n\n print(f\"Hello User {transaction.id}\")\n print(\"Please add some items\")\n\n item_list = []\n item = {}\n while True:\n item_name = input(\"Item Name :\")\n try:\n if int(item_name):\n print(\"You cannot fill the item beginning with digit\")\n continue\n except:\n pass\n else:\n pass\n\n count = input(\"Count :\")\n try:\n count = int(count)\n except:\n print(\"You cannot fill count with alphabets\")\n continue\n \n price = input(\"Price :\")\n try:\n price = int(price)\n except:\n print(\"You cannot fill price with alphabets\")\n continue\n \n \n item[item_name] = {\n \"count\":count,\n \"price\":price\n }\n\n item_list.append(item)\n print(f\"Item ={item_name}\")\n print(f\"Count ={count}\")\n print(f\"Price ={price}\")\n \n print(item_list)\n\n if input(\"add again ?(y/n)\") != \"y\":\n transaction.add_items(items=item_list)\n break\n else:\n continue\n \n while True:\n next_step = int(input(\n \"\"\"\n what you want to do next ?\n [1] edit item\n [2] delete item\n [3] reset chart \n [4] finish\n \"\"\"\n ))\n\n if next_step == 1:\n transaction.show_cart()\n name = input(\"what item you want to edit ?\")\n if transaction.check_list_item(item_name=name):\n item_name = input(\"Item Name :\")\n count = int(input(\"Count :\"))\n price = int(input(\"Price :\"))\n \n print(f\"Item ={item_name}\")\n print(f\"Count ={count}\")\n print(f\"Price ={price}\")\n\n transaction.edit_item(name=name, item=item_name, count=count, price=price)\n \n if next_step == 2:\n dele = input(\"What item you want to delete ?\")\n transaction.delete_item(item=dele)\n\n if next_step == 3:\n print(\"Reset Chart\")\n transaction.reset_cart()\n\n if next_step == 4:\n print(\"Finish Process\")\n transaction.show_cart()\n break","repo_name":"AlifvianM/Pacmann-SuperStoreApp","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2400,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"5223541074","text":"\"\"\"\nWrite a function that takes 2 non-empty arrays of integers and returns an array of 2 numbers (one from each array) whose absolute difference is the smallest. The number in the first array should appear first in the return array.\nTC: O(nlog(n) + mlog(m)), where n is the length of array1 and m is the length of array2\nSC: O(1)\n\"\"\"\ndef smallest_difference(array1, array2):\n array1.sort(), array2.sort()\n\n smallest_diff, pair = float('inf'), []\n p1 = p2 = 0\n\n while p1 < len(array1) and p2 < len(array2):\n diff = abs(array1[p1] - array2[p2])\n if diff < smallest_diff:\n smallest_diff, pair = diff, [array1[p1], array2[p2]]\n\n if p1 + 1 == len(array1):\n p2 += 1\n elif p2 + 1 == len(array2):\n p1 += 1\n elif abs(array1[p1 + 1] - array2[p2]) > abs(array1[p1] - array2[p2 + 1]):\n p2 += 1\n elif p2 + 1 == len(array2) or abs(array1[p1 + 1] - array2[p2]) < abs(array1[p1] - array2[p2 + 1]):\n p1 += 1\n else:\n p1, p2 = p1 + 1, p2 + 1\n\n return pair\n\n\nif __name__ == \"__main__\":\n # print(smallest_difference(\n # [-1, 5, 10, 20, 28, 3],\n # [26, 134, 135, 15, 17]\n # ))\n\n # print(smallest_difference(\n # [-1, 5, 10, 20, 3],\n # [26, 134, 135, 15, 17]\n # ))\n\n # print(smallest_difference(\n # [240, 124, 86, 111, 2, 84, 954, 27, 89],\n # [1, 3, 954, 19, 8]\n # ))\n\n print(smallest_difference(\n [10, 1000, 9124, 2142, 59, 24, 596, 591, 124, -123, 530],\n [-1441, -124, -25, 1014, 1500, 660, 410, 245, 530]\n ))\n","repo_name":"mattdepillis/python_dsa","sub_path":"ae_questions/arrays/medium/smallest_difference/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":1476,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22681412253","text":"#KHỞI TẠO\r\nimport numpy as np\r\nimport plotly.express as px\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\nimport tensorflow as tf\r\nfrom sklearn.metrics import accuracy_score, classification_report\r\n\r\ntrain_dir = r'C:\\Users\\ADMIN\\Downloads\\XLA\\Data\\dataset2-master\\dataset2-master\\images\\TRAIN'\r\ntest_dir = r'C:\\Users\\ADMIN\\Downloads\\XLA\\Data\\dataset2-master\\dataset2-master\\images\\TEST'\r\n\r\n#LOAD DATA ẢNH\r\n#KHỞI TẠO GENERATORS\r\ntrain_gen = tf.keras.preprocessing.image.ImageDataGenerator(\r\n preprocessing_function=tf.keras.applications.mobilenet_v2.preprocess_input,\r\n validation_split=0.2\r\n)\r\ntest_gen = tf.keras.preprocessing.image.ImageDataGenerator(\r\n preprocessing_function=tf.keras.applications.mobilenet_v2.preprocess_input\r\n)\r\n\r\n#FLOW DATA ẢNH VÀO CÁC DIRECTORY\r\ntrain_images = train_gen.flow_from_directory( #80% of directory go to here\r\n directory=train_dir,\r\n target_size=(224, 224),\r\n color_mode='rgb',\r\n class_mode='categorical',\r\n batch_size=32,\r\n shuffle=True,\r\n seed=42,\r\n subset='training'\r\n)\r\nval_images = train_gen.flow_from_directory( #another 20% go to here\r\n directory=train_dir,\r\n target_size=(224, 224),\r\n color_mode='rgb',\r\n class_mode='categorical',\r\n batch_size=32,\r\n shuffle=False,\r\n seed=42,\r\n subset='validation'\r\n)\r\ntest_images = test_gen.flow_from_directory( \r\n directory=test_dir,\r\n target_size=(224, 224),\r\n color_mode='rgb',\r\n class_mode='categorical',\r\n batch_size=32,\r\n shuffle=False,\r\n seed=42\r\n)\r\n\r\ntrain_images.next()[1]\r\n\r\n#Build Pretrained Model\r\npretrained_model = tf.keras.applications.MobileNetV2(\r\n input_shape=(224, 224, 3),\r\n include_top=False,\r\n weights='imagenet',\r\n pooling='avg'\r\n)\r\npretrained_model.trainable = False\r\n\r\n#Build Classification Model\r\ninputs = pretrained_model.input\r\nx = tf.keras.layers.Dense(128, activation='relu')(pretrained_model.output)\r\noutputs = tf.keras.layers.Dense(4, activation='softmax')(x)\r\n\r\nmodel = tf.keras.Model(inputs=inputs, outputs=outputs)\r\n\r\nmodel.compile(\r\n optimizer='adam',\r\n loss='categorical_crossentropy',\r\n metrics=['accuracy']\r\n)\r\nprint(model.summary())\r\n\r\n#Training\r\nhistory = model.fit(\r\n train_images,\r\n validation_data=val_images,\r\n epochs=30,\r\n callbacks=[\r\n tf.keras.callbacks.EarlyStopping(\r\n monitor='val_loss',\r\n patience=3,\r\n restore_best_weights=True\r\n )\r\n ]\r\n)\r\n\r\nfig = px.line(\r\n history.history,\r\n y=['loss','val_loss'],\r\n labels={'index':\"Epoch\", \"value\": \"Loss\"},\r\n title=\"Training and Validation Loss Over Time\"\r\n)\r\nfig.show()\r\n\r\n#KẾT QUẢ\r\nCLASS_NAMES = list(train_images.class_indices.keys())\r\nCLASS_NAMES\r\n\r\npredictions = np.argmax(model.predict(val_images), axis=1)\r\n\r\nacc = accuracy_score(val_images.labels, predictions)\r\ncm = tf.math.confusion_matrix(val_images.labels, predictions)\r\nclr = classification_report(val_images.labels, predictions, target_names=CLASS_NAMES)\r\n\r\nprint(\"Validation Accuracy: {:.3f}%\".format(acc * 100))\r\n\r\nplt.figure(figsize=(8, 8))\r\nsns.heatmap(cm, annot=True, fmt='g', vmin=0, cmap='Blues', cbar=False)\r\nplt.xticks(ticks= np.arange(4) + 0.5, labels=CLASS_NAMES)\r\nplt.yticks(ticks= np.arange(4) + 0.5, labels=CLASS_NAMES)\r\nplt.xlabel(\"Dự đoán\")\r\nplt.ylabel(\"Thực tế\")\r\nplt.title(\"Ma trận lỗi\")\r\nplt.show()\r\n\r\nprint(\"Classification Report:\\n----------------------\\n\", clr)\r\n\r\nmodel.save('model.h5')","repo_name":"HuynhQuocDan/Code_Final_Report_AI","sub_path":"Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":3461,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"72364285056","text":"\n# heuristic: tim ra ket noi tieu hao nang luong lon nhat\n# GA: tim ra ca the co tieu hao nang luong nho nhat\n\nimport random, time, os, sys\nlib_path = os.path.abspath(os.path.join('..'))\nsys.path.append(lib_path)\n\nfrom .heuristic import *\n\n\nGEN = 100\nCP = 0.8\nMP = 0.1\nNUM_OF_INDIVIDUALS = 100\nTERMINATE = 30\nalpha = 0.5\n\ndef random_init_individual(num_relay):\n \"Initial individual with any num of relay\"\n indi = []\n Y = random.randint(1, num_relay)\n xs = Y/num_relay\n count_relay = 0\n for i in range(0, num_relay):\n xx = random.random()\n if xx < xs:\n indi.append(1)\n else:\n indi.append(0)\n return indi\n\ndef count_current_relay(individual):\n sum = 0\n for g in individual:\n if g == 1:\n sum += 1\n return sum\n\n# mom and dad instead of parent1 and parent2 =)\ndef cross(mom, dad):\n num_relay = len(mom)\n mid = random.randint(0, num_relay-1)\n child1 = mom[:mid] + dad[mid:]\n child2 = dad[:mid] + mom[mid:]\n return child1, child2\n\n\ndef mutate(original):\n fake = original[:]\n ll = len(fake)\n count1 = 0\n count2 = 0\n id1 = random.randint(0, ll-1)\n while fake[id1] == 0:\n count1 += 1\n id1 = random.randint(0, ll-1)\n if count1 >= 2*ll:\n break\n id2 = random.randint(0, ll-1)\n while fake[id2] == 1:\n count2 += 1\n id2 = random.randint(0, ll-1)\n if count2 >= 2*ll:\n break\n fake[id1], fake[id2] = fake[id2], fake[id1]\n return fake\n\ndef normalize_loss(indi):\n if indi[1].loss(alpha) < 0:\n return float(\"inf\")\n else:\n return 10000*indi[1].loss(alpha) + indi[1].total_tranmission_loss()\n\n# def sort(individuals):\n# ll = len(individuals)\n# new_indis = individuals[:]\n# for i in range(len(individuals)):\n# for j in range(i+1, len(individuals)):\n# if new_indis[i][1].loss(alpha) > new_indis[j][1].loss(alpha):\n# new_indis[i], new_indis[j] = new_indis[j], new_indis[i]\n# elif new_indis[i][1].loss(alpha) == new_indis[j][1].loss(alpha):\n# if new_indis[i][1].total_tranmission_loss() > new_indis[j][1].total_tranmission_loss():\n# new_indis[i], new_indis[j] = new_indis[j], new_indis[i]\n# return new_indis\n\ndef GA(inp: WusnInput) -> int:\n # Khoi tao quan the\n individuals = []\n\n # Cac ca the da duoc tinh toan\n calculated = {}\n\n for i in range (0, NUM_OF_INDIVIDUALS):\n indi = random_init_individual(inp.num_of_relays)\n out = heuristic(inp, indi)\n \n calculated[str(indi)] = out\n individuals.append([indi, out])\n\n print(individuals[0])\n \n count_stable = 0\n max_c = individuals[0][1].loss(alpha)\n prev_max = individuals[0][1].loss(alpha)\n\n # Iterate through generations\n for it in range(0, GEN):\n start = time.time()\n none = 0\n not_none = 0\n # Crossover and mutation\n for id1 in range(0, NUM_OF_INDIVIDUALS):\n id2 = 0\n xx = random.random()\n if xx < CP:\n id2 = random.randint(0, NUM_OF_INDIVIDUALS-1)\n while id2 == id1:\n id2 = random.randint(0, NUM_OF_INDIVIDUALS-1)\n son, daughter = cross(individuals[id1][0], individuals[id2][0])\n\n if str(son) in calculated:\n out1 = calculated[str(son)]\n else:\n s = time.time()\n out1 = heuristic(inp, son)\n t = time.time()\n \n if str(daughter) in calculated:\n out2 = calculated[str(daughter)]\n else:\n # s = time.time()\n out2 = heuristic(inp, daughter)\n # t = time.time()\n\n if out1.mapping == {}:\n none += 1\n else: \n not_none += 1\n if out2.mapping == {}:\n none += 1\n else:\n not_none += 1\n\n individuals.append([son, out1])\n individuals.append([daughter, out2])\n\n xx2 = random.random()\n if xx2 < MP:\n grand_child1 = mutate(son)\n grand_child2 = mutate(daughter)\n m_out1 = heuristic(inp, grand_child1)\n m_out2 = heuristic(inp, grand_child2)\n\n if m_out1.mapping == {}:\n none += 1\n else:\n not_none += 1\n if m_out2.mapping == {}:\n none += 1\n else: \n not_none += 1\n\n individuals.append([grand_child1, m_out1])\n individuals.append([grand_child2, m_out2])\n\n individuals2 = sorted(individuals, key=normalize_loss)\n # individuals2 = sort(individuals)\n individuals = individuals2[:NUM_OF_INDIVIDUALS-1] \n individuals.append(individuals2[-1])\n if individuals[0][1].loss(alpha) < max_c:\n max_c = individuals[0][1].loss(alpha)\n if individuals[0][1].loss(alpha) == prev_max:\n count_stable += 1\n else:\n count_stable = 0\n if count_stable == TERMINATE:\n print(\"TERMINATE\")\n break\n prev_max = individuals[0][1].loss(alpha)\n end = time.time()\n print(\"none: %d, not_none: %d\" % (none, not_none))\n print(\"Gen: %d, time: %fs, min: %f %f %f\" % (it, end - start, len(individuals[0][1].used_relays), individuals[0][1].loss(alpha), individuals[NUM_OF_INDIVIDUALS-1][1].loss(alpha)))\n # print(max_c)\n return individuals[0]\n\n# Bo sung dieu kien SN va RN co ket noi duoc voi nhau hay khong bang cach them ban kinh sn va rn\n","repo_name":"duymanh96w/WusnNewModel","sub_path":"GAwHeuristic/GA.py","file_name":"GA.py","file_ext":"py","file_size_in_byte":5873,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"2549869304","text":"#!/usr/bin/env python3\r\n\"\"\"\r\nThis script contains the class for studying oscillating clusters on videos in 2D\r\n\"\"\"\r\n\r\nfrom cv2 import (\r\n connectedComponents, connectedComponentsWithStats, MORPH_CROSS,\r\n getStructuringElement, CV_16U, erode, dilate, morphologyEx, MORPH_OPEN,\r\n MORPH_CLOSE, MORPH_GRADIENT, BORDER_CONSTANT, resize, imshow, waitKey,\r\n FONT_HERSHEY_SIMPLEX, putText)\r\nfrom numpy import (\r\n append, float32, sum, mean, zeros, empty, array, nonzero, unique,\r\n isin, logical_or, logical_not, greater, uint8,\r\n uint32, min, any)\r\nfrom cellects.image_analysis.morphological_operations import get_minimal_distance_between_2_shapes\r\n\r\n\r\nclass ClusterFluxStudy:\r\n def __init__(self, dims):\r\n self.dims = dims\r\n\r\n self.pixels_data = empty((4, 0), dtype=uint32)\r\n self.clusters_id = zeros(self.dims[1:], dtype=uint32)\r\n # self.alive_clusters_in_flux = empty(0, dtype=uint32)#list()\r\n self.cluster_total_number = 0\r\n self.cross_33 = getStructuringElement(MORPH_CROSS, (3, 3))\r\n\r\n def update_flux(self, t, contours, current_flux, period_tracking, clusters_final_data):\r\n # flux_dir_changed = logical_xor(current_flux, self.clusters_id)\r\n # Save the data from pixels that are not anymore in efflux\r\n lost = greater(self.clusters_id > 0, current_flux > 0)\r\n # lost = logical_not(equal(current_flux > 0, self.clusters_id > 0))\r\n # lost = flux_dir_changed * (self.clusters_id > 0)\r\n # lost_coord = nonzero(lost)\r\n # if any(lost):\r\n # Some pixels of that cluster faded, save their data\r\n lost_data = nonzero(lost)\r\n lost_data = array((period_tracking[lost], # lost_coord[0], lost_coord[1],\r\n self.clusters_id[lost], lost_data[0], lost_data[1]), dtype=uint32)\r\n # Add this to the array containing the data of each cluster that are still alive\r\n self.pixels_data = append(self.pixels_data, lost_data, axis=1)\r\n # Stop considering these pixels in period_tracking because they switched\r\n period_tracking[lost] = 0\r\n current_period_tracking = zeros(self.dims[1:], dtype=bool)\r\n for curr_clust_id in unique(current_flux)[1:]:\r\n # Get all pixels that were in the same flux previously\r\n curr_clust = current_flux == curr_clust_id\r\n already = self.clusters_id * curr_clust\r\n new = greater(curr_clust, self.clusters_id > 0)\r\n # new = flux_dir_changed * (current_flux == curr_clust_id)\r\n\r\n if not any(already):\r\n # It is an entirely new cluster:\r\n cluster_pixels = new\r\n self.cluster_total_number += 1\r\n cluster_name = self.cluster_total_number\r\n # self.alive_clusters_in_flux = append(self.alive_clusters_in_flux, cluster_name)\r\n else:\r\n # Check whether parts of that cluster correspond to several clusters in clusters_id\r\n cluster_names = unique(already)[1:]\r\n # keep only one cluster name to gather clusters that just became connected\r\n cluster_name = min(cluster_names)\r\n # Put the same cluster name for new ones and every pixels that were\r\n # a part of a cluster touching the current cluster\r\n cluster_pixels = logical_or(isin(self.clusters_id, cluster_names), new)\r\n # new = self.clusters_id == cluster_names\r\n # If they are more than one,\r\n if len(cluster_names) > 1:\r\n # Update these cluster names in pixels_data\r\n self.pixels_data[1, isin(self.pixels_data[1, :], cluster_names)] = cluster_name\r\n # self.pixels_data[self.pixels_data[1, :] == cluster_names] = cluster_name\r\n # Update these cluster names in alive_clusters_in_flux: remove names that are not used anymore\r\n # self.alive_clusters_in_flux = delete(self.alive_clusters_in_flux, isin(self.alive_clusters_in_flux, cluster_names[cluster_names != cluster_name]))\r\n # cluster_names_to_remove = cluster_names.copy()\r\n # cluster_names_to_remove = delete(cluster_names_to_remove,\r\n # nonzero(cluster_names_to_remove == cluster_name))\r\n # # Remove the deleted clusters from the alive cluster list\r\n # [self.alive_clusters_in_flux.remove(i) for i in cluster_names_to_remove if i in self.alive_clusters_in_flux]\r\n # Update clusters_id\r\n self.clusters_id[cluster_pixels] = cluster_name\r\n # Update period_tracking\r\n current_period_tracking[curr_clust] = True\r\n\r\n period_tracking[current_period_tracking] += 1\r\n # Remove lost pixels from clusters_id\r\n self.clusters_id[lost] = 0\r\n # self.alive_clusters_in_flux = self.alive_clusters_in_flux[isin(self.alive_clusters_in_flux, unique(self.clusters_id))]\r\n\r\n # Find out which clusters are still alive or not\r\n # still_alive_clusters = isin(self.pixels_data[1, :], self.alive_clusters_in_flux)\r\n still_alive_clusters = isin(self.pixels_data[1, :], unique(self.clusters_id))\r\n clusters_to_archive = unique(self.pixels_data[1, logical_not(still_alive_clusters)])\r\n # store their data in clusters_final_data\r\n for cluster in clusters_to_archive:\r\n cluster_bool = self.pixels_data[1, :] == cluster\r\n cluster_size = sum(cluster_bool)\r\n cluster_img = zeros(self.dims[1:], dtype=uint8)\r\n cluster_img[self.pixels_data[2, cluster_bool], self.pixels_data[3, cluster_bool]] = 1\r\n if any(dilate(cluster_img, kernel=self.cross_33, borderType=BORDER_CONSTANT, borderValue=0) * contours):\r\n minimal_distance = 1\r\n else:\r\n if cluster_size > 200:\r\n cluster_img = nonzero(morphologyEx(cluster_img, MORPH_GRADIENT, self.cross_33))\r\n contours[cluster_img] = 2\r\n else:\r\n contours[self.pixels_data[2, cluster_bool], self.pixels_data[3, cluster_bool]] = 2\r\n # Get the minimal distance between the border of the cell(s) (noted 1 in contours)\r\n # and the border of the cluster in the cell(s) (now noted 2 in contours)\r\n minimal_distance = get_minimal_distance_between_2_shapes(contours)\r\n data_to_save = array([[mean(self.pixels_data[0, cluster_bool]), t,\r\n cluster_size, minimal_distance]], dtype=float32)\r\n clusters_final_data = append(clusters_final_data, data_to_save,\r\n axis=0) # [\"mean_pixel_period\", \"total_size\", \"death_time\"]\r\n # and remove their data from pixels_data\r\n self.pixels_data = self.pixels_data[:, still_alive_clusters]\r\n\r\n return period_tracking, clusters_final_data\r\n\r\n\r\n","repo_name":"Aurele-B/Cellects","sub_path":"src/cellects/image_analysis/cluster_flux_study.py","file_name":"cluster_flux_study.py","file_ext":"py","file_size_in_byte":7030,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"74634275454","text":"\nfrom django.views.decorators.csrf import csrf_exempt\nfrom rest_framework import status, serializers\nfrom django.contrib.auth import authenticate, login\nfrom django.contrib.auth.models import User\n\nfrom .models import Event, Category, Tipology\nfrom django.http import JsonResponse\nfrom event.serializers import CategorySerializer, EventSerializer, TipologySerializer\n\nimport json\n\n# Create your views here.\n\n@csrf_exempt\ndef add_user_view(request):\n message = \"\"\n if request.method == 'POST':\n jsonUser = json.loads(request.body.decode('utf-8'))\n password = jsonUser['password']\n email = jsonUser['email']\n\n user_model = User.objects.create_user(username=email, password=password)\n user_model.email = email\n user_model.save()\n message = 'Usuario Registrado'\n\n return JsonResponse({'message':message}, status=status.HTTP_200_OK)\n\n@csrf_exempt\ndef login_user_view(request):\n message = \"\"\n status_code = status.HTTP_200_OK\n if request.method == 'POST':\n jsonUser = json.loads(request.body.decode('utf-8'))\n username = jsonUser['email']\n password = jsonUser['password']\n user = authenticate(username=username, password=password)\n if user is not None:\n login(request, user)\n message = 'usuario logeado: ' + str(user.pk)\n return JsonResponse({'message': message, 'id':user.pk}, status=status_code)\n else:\n message = 'Usuario o contrasenia incorrectos.'\n status_code = status.HTTP_401_UNAUTHORIZED\n return JsonResponse({'message': message}, status=status_code)\n\n \n\n\n@csrf_exempt\ndef all_events_view(request):\n if request.method == 'GET':\n user = request.GET.get('userId')\n event_list = Event.objects.filter(creation_user=user).order_by('-creation_date')\n serializer = EventSerializer(event_list, many=True)\n return JsonResponse(serializer.data, safe=False)\n\n\n@csrf_exempt\ndef add_event_view(request):\n if request.method == 'POST':\n jsonEvent = json.loads(request.body.decode('utf-8'))\n name = jsonEvent['name']\n detail = jsonEvent['detail']\n place = jsonEvent['place']\n address = jsonEvent['address']\n start_date = jsonEvent['start_date']\n end_date = jsonEvent['end_date']\n category_type = jsonEvent['category_type']\n event_type = jsonEvent['event_type']\n creation_user = jsonEvent['creation_user']\n\n category_model = Category.objects.get(pk=category_type)\n tipology_model = Tipology.objects.get(pk=event_type)\n user_model = User.objects.get(pk=creation_user)\n\n event_model = Event()\n event_model.name = name\n event_model.detail = detail\n event_model.place = place\n event_model.address = address\n event_model.start_date = start_date\n event_model.end_date = end_date\n event_model.category_type = category_model\n event_model.event_type = tipology_model\n event_model.creation_user = user_model\n event_model.save()\n message = 'Evento Registrado'\n\n return JsonResponse({'message': message})\n\n\n@csrf_exempt\ndef edit_event_view(request):\n message = \"\"\n if request.method == 'POST':\n jsonEvent = json.loads(request.body.decode('utf-8'))\n id= jsonEvent['id']\n name = jsonEvent['name']\n detail = jsonEvent['detail']\n place = jsonEvent['place']\n address = jsonEvent['address']\n start_date = jsonEvent['start_date']\n end_date = jsonEvent['end_date']\n category_type = jsonEvent['category_type']\n event_type = jsonEvent['event_type']\n creation_user = jsonEvent['creation_user']\n\n category_model = Category.objects.get(pk=category_type)\n tipology_model = Tipology.objects.get(pk=event_type)\n user_model = User.objects.get(pk=creation_user)\n\n event_model = Event.objects.get(pk=id)\n event_model.name = name\n event_model.detail = detail\n event_model.place = place\n event_model.address = address\n event_model.start_date = start_date\n event_model.end_date = end_date\n event_model.category_type = category_model\n event_model.event_type = tipology_model\n event_model.creation_user = user_model\n event_model.save()\n message = 'Evento: ' + str(event_model.pk) + ' actualizado'\n\n return JsonResponse({'message': message})\n\n\n@csrf_exempt\ndef delete_event_view(request):\n message = \"\"\n if request.method == 'POST':\n jsonEvent = json.loads(request.body.decode('utf-8'))\n id = jsonEvent['id']\n event_model = Event.objects.get(pk=id).delete()\n message = 'Evento:' + id + ' eliminado'\n\n return JsonResponse({'message': message})\n\n\n@csrf_exempt\ndef category_event_view(request):\n if request.method == 'GET':\n category_model = Category.objects.all()\n serializer = CategorySerializer(category_model, many=True)\n\n return JsonResponse(serializer.data, safe=False)\n\n\n@csrf_exempt\ndef tipology_event_view(request):\n if request.method == 'GET':\n typology_model = Category.objects.all()\n serializer = TipologySerializer(typology_model, many=True)\n\n return JsonResponse(serializer.data, safe=False)\n\n\n","repo_name":"emantilla/bdb-test-backend","sub_path":"event/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5308,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"73815683136","text":"from llama_index import SimpleDirectoryReader, GPTListIndex, readers, GPTSimpleVectorIndex, LLMPredictor, PromptHelper, ServiceContext\nfrom langchain import OpenAI\nimport os\n\nos.environ[\"OPENAI_API_KEY\"] = 'YOUR_API_KEY_HERE'\n\ndef construct_index(directory_path):\n # set maximum input size\n max_input_size = 4096\n # set number of output tokens\n num_outputs = 2000\n # set maximum chunk overlap\n max_chunk_overlap = 20\n # set chunk size limit\n chunk_size_limit = 600 \n\n # define prompt helper\n prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)\n\n # define LLM\n llm_predictor = LLMPredictor(llm=OpenAI(temperature=0.5, model_name=\"text-davinci-003\", max_tokens=num_outputs))\n \n documents = SimpleDirectoryReader(directory_path).load_data()\n \n service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)\n index = GPTSimpleVectorIndex.from_documents(documents, service_context=service_context)\n\n return index\n\n# Save the index to disk outside of construct_index\nif not os.path.isfile('index.json'):\n index = construct_index('context_data/iotium')\n index.save_to_disk('index.json')\n\ndef ask_ai():\n index = GPTSimpleVectorIndex.load_from_disk('index.json')\n while True: \n query = input(\"What do you want to ask? \")\n response = index.query(query)\n print(f\"Response: {response.response}\")\n\nask_ai()\n","repo_name":"Gokuljokul/CustomGPT","sub_path":"trainDataset.py","file_name":"trainDataset.py","file_ext":"py","file_size_in_byte":1480,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"15486816267","text":"from datetime import datetime, timedelta\nfrom data_manager import DataManager\nfrom flight_search import FlightSearch\nfrom notification_manager import NotificationManager\n\ndata_manager = DataManager()\nsheet_data = data_manager.get_destination_data()\nflight_search = FlightSearch()\nnotification_manager = NotificationManager()\n\nif sheet_data[0][\"iataCode\"] == \"\":\n for row in sheet_data:\n row[\"iataCode\"] = flight_search.get_destination_code(row[\"city\"])\n data_manager.destination_data = sheet_data\n data_manager.update_destination_codes()\n\ntomorrow = datetime.now() + timedelta(days=1)\nsix_month_from_today = datetime.now() + timedelta(days=(6 * 30))\n\nfor destination in sheet_data:\n flight = flight_search.check_flights(\n destination[\"home\"],\n destination[\"iataCode\"],\n from_time=tomorrow,\n to_time=six_month_from_today,\n nights_in_dst_from=destination[\"minNrOfDays\"],\n nights_in_dst_to=destination[\"maxNrOfDays\"],\n nrOfPersons=destination[\"nrOfPersons\"]\n )\n if flight is None:\n continue\n if flight.price < destination[\"lowestPricePerPerson\"]:\n msg = f\"Low price alert! Only {flight.price} EUR per person to fly from {flight.origin_city}-{flight.origin_airport}\\\n to {flight.destination_city}-{flight.destination_airport} and back, from {flight.out_date} to {flight.return_date}.\"\n if flight.via_city:\n msg += f\"\\nThe flight has a stop-over via {flight.via_city}.\"\n for route in flight.route:\n airline = route[\"airline\"]\n city_code_from = route[\"cityCodeFrom\"]\n city_code_to = route[\"cityCodeTo\"]\n flight_no = route[\"flight_no\"]\n departure = route[\"local_departure\"][:16].replace('T', '_')\n arrival = route[\"local_arrival\"][:16].replace('T', '_')\n msg += f\"\\n {city_code_from}-{city_code_to} {airline}{flight_no} {departure} {arrival}\"\n msg += f\"\\nbooking link: https://www.google.co.uk/flights?hl=en#flt={destination['home']}.{destination['iataCode']}.{flight.out_date}*{destination['iataCode']}.{destination['home']}.{flight.return_date}\"\n\n notification_manager.send_msg(message=msg, lst_emails=destination[\"email\"])","repo_name":"MarcBruyland/flight-deal","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2227,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"185811219","text":"import streamlit as st\r\nimport zipfile\r\nfrom Bio import SeqIO\r\nfrom io import StringIO\r\nfrom io import BytesIO\r\nimport random\r\n\r\ndef preprocessing():\r\n st.markdown(\r\n \"\"\"\r\n \r\n \"\"\",\r\n unsafe_allow_html=True,\r\n )\r\n\r\n with open(\"quotes.txt\", \"r\", encoding=\"utf-8\") as f:\r\n file_text = f.read()\r\n quotes = file_text.split(\"\\n\\n\")\r\n\r\n st.title(\"Welcome to the Preprocessing Page!\")\r\n st.subheader(\"Upload your FASTA file\")\r\n uploaded_file = st.file_uploader(\"Upload\", type=[\"fasta\",\"fa\"])\r\n #user_length = st.text_input(\"Enter the legth of the Sequence: \")\r\n col1, col2 = st.columns([1,10])\r\n with col1:\r\n show_submit = st.button(\"Submit\")\r\n with col2:\r\n show_example = st.button(\"Load Example\")\r\n\r\n def load_example_file():\r\n with open(\"example.fasta\",\"rb\") as f:\r\n return BytesIO(f.read())\r\n\r\n def preprocess_sequences(records):\r\n #remove sequences with \"X\"\r\n print(\"Removing the duplicate sequences...\")\r\n records = [r for r in records if \"X\" not in r.seq]\r\n\r\n #remove duplicate sequences\r\n sequences = []\r\n unique_records = []\r\n for record in records:\r\n sequence = str(record.seq)\r\n if sequence not in sequences:\r\n unique_records.append(record)\r\n sequences.append(sequence)\r\n st.write(\"Total number of sequences after removing duplicate sequences is:: \", len(unique_records))\r\n return unique_records\r\n \r\n def filter_sequences(sequences):#, target_length):\r\n print(\"Removing unequal length sequences...\")\r\n filtered_sequences = []\r\n for record in sequences:\r\n #if(len(record.seq) == target_length):\r\n filtered_sequences.append(record)\r\n st.write(\"Total number of sequences after making sequences of equal length is:: \", len(filtered_sequences))\r\n return filtered_sequences\r\n\r\n # Initialize the current index\r\n if \"current_index\" not in st.session_state:\r\n st.session_state.current_index = 0\r\n\r\n st.write(\"\"\"***\"\"\")\r\n\r\n if show_submit:\r\n with st.spinner(\"SAMOSA is cleaning your dataset. Please wait....\"):\r\n with zipfile.ZipFile(\"preprocessed_data.zip\", \"w\") as output_zip:\r\n if(uploaded_file is not None):\r\n stringio = StringIO(uploaded_file.getvalue().decode(\"utf-8\"))\r\n records = list(SeqIO.parse(stringio, \"fasta\"))\r\n st.write(\"Total number of sequences before preprocessing is:: \",len(records))\r\n\r\n #preprocess the sequences\r\n processed_records = preprocess_sequences(records)\r\n sequence_filtered = filter_sequences(processed_records)#,int(user_length))\r\n\r\n #download the preprocessed file\r\n if(len(sequence_filtered)>0):\r\n with StringIO() as output:\r\n SeqIO.write(sequence_filtered, output, \"fasta\")\r\n processed_file = output.getvalue().encode()\r\n\r\n st.download_button(label=\"Download Preprocessed File\", data=processed_file, file_name=\"preprocessed.fasta\", mime=\"application/octet-stream\")\r\n quote = random.choice(quotes)\r\n st.write(quote)\r\n\r\n elif show_example:\r\n with st.spinner(\"SAMOSA is cleaning the example dataset. Please Wait.....\"):\r\n with zipfile.ZipFile(\"example_preprocessed_data.zip\", \"w\") as output_zip:\r\n if(uploaded_file is None):\r\n file_contents = load_example_file()\r\n stringio = StringIO(file_contents.getvalue().decode(\"utf-8\"))\r\n records = list(SeqIO.parse(stringio, \"fasta\"))\r\n st.write(\"Total number of sequences before preprocessing is:: \",len(records))\r\n\r\n processed_records = preprocess_sequences(records)\r\n\r\n if(len(processed_records)>0):\r\n with StringIO() as output:\r\n SeqIO.write(processed_records, output, \"fasta\")\r\n processed_file = output.getvalue().encode()\r\n\r\n st.download_button(label=\"Download Preprocessed Example File\", data=processed_file, file_name=\"preprocessed_example_file.fasta\", mime=\"application/octet-stream\")\r\n\r\n st.write(\"\")\r\n st.write(\"\"\"***\"\"\")\r\n quote = random.choice(quotes)\r\n st.write(quote)\r\n st.write(\"\"\"***\"\"\")\r\n\r\n\r\n","repo_name":"psychedelic2007/biohub_webserver","sub_path":"util/pages/Preprocessing.py","file_name":"Preprocessing.py","file_ext":"py","file_size_in_byte":4703,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"12724953013","text":"# \n#\n#\n#\n#\n# This is the script for the technical task for the process in Motor AI.\n# Author: Andrés Prada\n\nimport os, glob, cv2, time\nfrom pathlib import Path\nimport numpy as np\nfrom keras.applications.resnet import ResNet152\nfrom sklearn.cluster import AgglomerativeClustering\nfrom sklearn.utils.multiclass import unique_labels\nfrom sklearn import manifold\nfrom sklearn.metrics import confusion_matrix, normalized_mutual_info_score\nimport matplotlib.pyplot as plt\n\n# Importts from other models:\n#\n# from keras import Model\n# from cifar100vgg import cifar100vgg\n# from keras.applications.inception_resnet_v2 import InceptionResNetV2\n# from keras.applications.nasnet import NASNetLarge\n# from keras.applications.vgg16 import VGG16\n# from keras.applications.xception import Xception\n#\n\n\ndef visualize_data(Z, labels, num_clusters):\n\t'''\n\t\tThis function helps to visualize the data performing a dimensionality reduction with TSNE.\n\t'''\n\ttsne = manifold.TSNE(n_components=2, init='pca', random_state=0).fit_transform(Z)\n\tfig = plt.figure()\n\tplt.scatter(tsne[:, 0], tsne[:, 1], s=2, c=labels, cmap=plt.cm.get_cmap(\"jet\", num_clusters))\n\tplt.colorbar(ticks=range(num_clusters))\n\tplt.show()\n\ndef plot_matrix(y_true, y_pred, classes, title=None, cmap=plt.cm.Blues):\n\t\"\"\"\n\t This function plots the matrix true label vs cluster label.\n\t It is edited from https://scikit-learn.org/stable/auto_examples/model_selection/plot_confusion_matrix.html\n\t \"\"\"\n\n\t# Compute confusion matrix\n\tcm = confusion_matrix(y_true, y_pred)\n\n\t# Only use the labels that appear in the data\n\tclasses = classes[unique_labels(y_true, y_pred)]\n\n\tfig, ax = plt.subplots()\n\tim = ax.imshow(cm, interpolation='nearest', cmap=cmap)\n\tax.figure.colorbar(im, ax=ax)\n\t# We want to show all ticks...\n\tax.set(xticks=np.arange(cm.shape[1]), yticks=np.arange(cm.shape[0]), xticklabels=classes, yticklabels=classes, title=title, ylabel='True label', xlabel='Cluster label')\n\t \n\t# Rotate the tick labels and set their alignment.\n\tplt.setp(ax.get_xticklabels(), fontsize=6, rotation=45, ha=\"right\", rotation_mode=\"anchor\")\n\tplt.setp(ax.get_yticklabels(), fontsize=6)\n\n\t# Loop over data dimensions and create text annotations.\n\tthresh = cm.max() / 2.\n\tfor i in range(cm.shape[0]):\n\t\tfor j in range(cm.shape[1]):\n\t\t\tif cm[i,j] > 0:\n\t\t\t\tax.text(j, i, cm[i, j], fontsize=6, ha=\"center\", va=\"center\", color=\"white\" if cm[i, j] > thresh else \"black\")\n\tfig.tight_layout()\n\tplt.show()\n\treturn ax\n\nif __name__ == '__main__':\n\n\tstart_time = time.time()\n\t# Open images folder\n\timage_path = \"tsrd-train/\"\n\timage_paths = [str(p)for p in Path(image_path).glob('*.png')]\n\n\t# Extracct labels from image name\n\tlabels = [int(str(p).split(\"/\")[1].split(\"_\")[0][-2:]) for p in image_paths]\n\n\n\t# Commented models. For testing, please check the README to adjust image dimensions and feature vector size\n\n\t#model = Xception(include_top=False, weights='imagenet', pooling='avg')\n\t#model = NASNetLarge(include_top=False, weights='imagenet', pooling='avg')\n\t#model = InceptionResNetV2(include_top=False, weights='imagenet', pooling='avg')\n\t#model = VGG16(include_top=False, weights='imagenet', pooling='avg')\n\t#model = cifar100vgg(train=False)\n\t#my_layer = model.model.layers[56]\n\t#model = Model(model.model.input, outputs=my_layer.output)\n\n\t# Define the model\n\tmodel = ResNet152(include_top=False, weights='imagenet', pooling='avg')\n\tn_clusters = 58\n\tmodel.layers[0].trainable = False\n\tdims = [224,224]\n\tvect_len = 2048\n\n\t# Define list to store vector values\n\tfeature_vects = np.zeros((len(image_paths), vect_len), dtype=float)\n\n\t# Extract vectors\n\tfor idx, img in enumerate(image_paths):\n\t\t# Print\n\t\tprint(\"Extracting vector features for image: \"+str(idx))\n\n\t\t# Load and reshape the image to input to the network\n\t\timg = cv2.resize(cv2.imread(img), (dims[0], dims[1]))\n\n\t\t# Add the 4th dim (1, 224, 224, 3)\n\t\timg = np.expand_dims(img.copy(), axis=0)\n\n\t\t# Predict and store value\n\t\tfeature_vects[idx, :] = model.predict(img).flatten()\n\n\t# Cluster the vectors\n\tclusters = AgglomerativeClustering(n_clusters=n_clusters).fit(feature_vects)\n\n\t# Check running time\n\tprint(\"--- %s seconds ---\" % (time.time() - start_time))\n\n\t#Evaluate similarity normalized_mutual_info_score\n\tnmi = normalized_mutual_info_score(labels, clusters.labels_, average_method='warn')\n\tprint('Evaluation of similarity with normalized mutual score: ' + str(nmi))\n\n\t# Print true label vs cluster\n\tplot_matrix(labels, clusters.labels_, np.unique(labels), title='Labeled images')\n\n\t# Finally, visualize data\n\tvisualize_data(feature_vects, labels, n_clusters)\n\n\n\n\n\n\n\n\n\n","repo_name":"AndresPrada/clustering-tsrd","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4569,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"29199607744","text":"import argparse, sys\nctecka = argparse.ArgumentParser()\n\nctecka.add_argument(\"files\", default=(sys.stdin,), nargs=\"*\", type=argparse.FileType())\nctecka.add_argument(\"--min\", \"-m\", default=-1, type=int)\nctecka.add_argument(\"--ignore-case\", \"-i\", action=\"store_true\")\nctecka.add_argument(\"--alphabetical\", \"-a\", action=\"store_true\")\nctecka.add_argument(\"--output\", \"-o\", default=sys.stdout, type=argparse.FileType(\"w\"))\n\nargs = ctecka.parse_args()\n\ndef spocti_slova(soubory, ignore_case):\n frekvence = {}\n for soubor in soubory:\n with soubor:\n obsah_souboru = \"\".join(soubor.readlines()).split()\n for slovo in obsah_souboru:\n if ignore_case:\n slovo = slovo.lower()\n if slovo not in frekvence:\n frekvence[slovo] = 1\n else:\n frekvence[slovo] += 1\n return frekvence\n\nfrekvence_slov = spocti_slova(args.files, args.ignore_case)\nmaximalni_frekvence = len(str(max(frekvence_slov.values())))\n\nif not args.alphabetical:\n poradi_slov = sorted(frekvence_slov.items(), key=lambda x: (-x[1], x[0]))\nelse:\n poradi_slov = sorted(frekvence_slov.items())\n\nwith args.output as soubor:\n for slovo, pocet in poradi_slov:\n if pocet != -1 and pocet >= args.min:\n soubor.writelines(f\"{str(pocet).rjust(maximalni_frekvence)} {slovo}\\n\")\n","repo_name":"MaximPolak/Python","sub_path":"2021-06-24/Frekvencni_distribuce.py","file_name":"Frekvencni_distribuce.py","file_ext":"py","file_size_in_byte":1352,"program_lang":"python","lang":"cs","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"44776267798","text":"#\n# https://github.com/dusty-nv/jetson-inference/blob/master/docs/detectnet-console-2.md\n#\n# sudo systemctl status track_detect.service\n# sudo systemctl restart track_detect.service\n# sudo systemctl disable track_detect.service\n# sudo systemctl enable track_detect.service\n# sudo systemctl stop track_detect.service\n# sudo systemctl start track_detect.service\n# tail -f /tmp/jetson.log\n#\n# Importing all the necessary modules\nfrom typing import Counter\nimport jetson.inference\nimport jetson.utils\nimport time\nimport cv2\nimport numpy as np \nfrom datetime import datetime\nfrom tracker import *\nfrom support_functions import *\nimport pandas as pd\nimport paho.mqtt.client as mqtt\nimport sys;\nimport logging;\nimport json;\n\ndef on_connect(client, userdata, flags, rc):\n if rc==0:\n client.connected_flag=True #set flag\n logging.debug(\"paho mqtt client connected ok\")\n elif rc==5:\n logging.debug(\"paho mqtt client not connected, authentication failure\")\n client.bad_connection_flag=True\n else:\n logging.debug(\"paho mqtt client not connected, returned code=%d\",rc)\n client.bad_connection_flag=True\n\nlogging.basicConfig(filename='/tmp/jetson.log', level=logging.DEBUG)\n\nclient_name='Jetson'\nclient = mqtt.Client(client_name)\nhost='130.191.161.21' # broker address\nclient.connected_flag=False\nclient.bad_connection_flag=False\nclient.on_connect=on_connect # bind callback function\nclient.username_pw_set(username=\"starlab\",password=\"starlab!\")\nclient.connect(host, port=1883, keepalive=60, bind_address=\"\")\n\nclient.loop_start() #Start loop\n\nwhile not client.connected_flag and client.bad_connection_flag: #wait in loop\n logging.debug(\"In wait loop\")\n time.sleep(1)\n\nlogging.debug('client.bad_connection_flag: %r',client.bad_connection_flag)\nlogging.debug('client.connected_flag: %r\\n\\n',client.connected_flag)\n\nmsg = f\"started\"\ntopic = f\"pelco/jetson\"\nresult = client.publish(topic, msg)\nstatus = result[0]\nif status == 0:\n logging.debug(f\"Send `{msg}` to topic `{topic}`\")\nelse:\n logging.debug(f\"Failed to send message to topic {topic}\")\n sys.exit()\n\ntracker = EuclideanDistTracker()\n\n# For FPS text need time\ntimeStamp=time.time()\nfpsFilt=0\n\n# Model \n#net = jetson.inference.detectNet(argv=[\"--model=/media/jetson/UGUR_USB_C/models/epoch_max/ssd-mobilenet.onnx\", \"--labels=/media/jetson/UGUR_USB_C/models/april_model/labels.txt\", \"--input-blob=input_0\", \"--output-cvg=scores\", \"--output-bbox=boxes\"], threshold=0.5)\nnet = jetson.inference.detectNet(argv=[\"--model=/home/iot/jetson-inference/python/training/detection/ssd/models/last_model/ssd-mobilenet.onnx\", \"--labels=/home/iot/jetson-inference/python/training/detection/ssd/models/last_model/labels.txt\", \"--input-blob=input_0\", \"--output-cvg=scores\", \"--output-bbox=boxes\"], threshold=0.5)\n\n# Picture Size for display\ndispW=1280\ndispH=720\nflip=2\nfont=cv2.FONT_HERSHEY_SIMPLEX # Font for the texts\n\n# Video Settings\n#cap=cv2.VideoCapture('file:///home/jetson/Desktop/Sample_Video/suv_truck.mp4') #can be changed with any other video source or a file\n\n\ncap=cv2.VideoCapture('rtsp://ued:uU8xwmin@sunray.sdsu.edu/stream2')\n\ncap.set(cv2.CAP_PROP_FRAME_WIDTH, dispW)\ncap.set(cv2.CAP_PROP_FRAME_HEIGHT, dispH)\ncap.set(cv2.CAP_PROP_FPS, int(3))\n\nresult = cv2.VideoWriter('output.mp4', cv2.VideoWriter_fourcc(*'MP4V'),10, (1280,720))\ndetected_objects = []\nwhile True:\n\n ret, img = cap.read()\n\n if ret == False:\n condition = False\n break\n \n # Image Info to OpenCV\n height=img.shape[0]\n width=img.shape[1]\n\n # Changing the color to Needed format for detection for Nvidia Jetson\n frame=cv2.cvtColor(img,cv2.COLOR_BGR2RGBA).astype(np.float32)\n # Converting the image from Numpy to CUDA\n frame=jetson.utils.cudaFromNumpy(frame)\n\n # Getting Detections for each Frame\n detections=net.Detect(frame, width, height)\n\n \n detection_list = []\n \n for detect in detections:\n # TODO: Uncomment this for all the info for each detection\n print(detect)\n \n # Get all the information from the detection class\n \n ID=detect.ClassID\n top=detect.Top\n left=detect.Left\n bottom=detect.Bottom\n right=detect.Right\n width_d = detect.Width\n height_d = detect.Height\n item=net.GetClassDesc(ID)\n center = detect.Center\n center_x = int(center[0])\n center_y = int(center[1])\n confidence = detect.Confidence\n \n # Getting the center locations in integer format in a tuple\n center = (center_x,center_y)\n \n # Rendering the image\n # * Boundary Box\n img = cv2.rectangle(img,(int(left),int(top)),(int(right),int(bottom)),(0,255,0),2)\n \n # * Center Circles\n img = cv2.circle(img, center, 1, (255,255,255), 10)\n # * Class ID and Confidence Text\n cv2.putText(img,str(item)+\" \"+str(round(detect.Confidence,2)),(int(left)+75,int(top)-15),font,1,(255,255,0),2)\n\n \n detection_list.append([int(left),int(top),int(width_d),int(height_d),item])\n \n currentTime = datetime.now()\n \n print(item)\n\n with open(\"/home/iot/Desktop/app/Detections.txt\", \"a\") as f:\n f.write(\"The current timestamp is: \" + str(datetime.now()))\n f.write(\"\\n\")\n f.write(\"The detection Details are: \" + str(detect))\n f.write(\"\\n\")\n f.write(\"Detected is: \" + str(item))\n f.write(\"\\n\")\n f.close()\n \n data = {\n \"id\": format(detect.ClassID),\n \"top\": \"{:.2f}\".format(detect.Top),\n \"left\": \"{:.2f}\".format(detect.Left),\n \"bottom\": \"{:.2f}\".format(detect.Bottom),\n \"right\": \"{:.2f}\".format(detect.Right),\n \"width_d\": \"{:.2f}\".format(detect.Width),\n \"height_d\": \"{:.2f}\".format(detect.Height),\n \"class\": format(net.GetClassDesc(ID)),\n #\"center\": format(detect.Center),\n \"center_x\": format(int(center[0])),\n \"center_y\": format(int(center[1])),\n \"confidence\": \"{:.2f}\".format(detect.Confidence),\n #\"class\": format(category_index[classes+1]['name']),\n #\"box\": [format(x_min_disp), format(y_min_disp), format(x_max_disp), format(y_max_disp)],\n #\"date\": format(t.month) + '/' + format(t.day) + '/' + format(t.year),\n \"time\": format(currentTime.hour) + ':' + format(currentTime.minute) + ':' + format(currentTime.second)\n #\"frame\": ['height:'+format(height), 'width:'+format(width)],\n #\"score\": \"{:.2f}\".format(score),\n #\"inference_time\": \"{:.4f}\".format(inference_time)\n }\n msg = json.dumps(data)\n topic = f\"pelco/jetson\"\n result = client.publish(topic, msg)\n status = result[0]\n if status == 0:\n logging.debug(f\"Send `{msg}` to topic `{topic}`\")\n else:\n logging.debug(f\"Failed to send message to topic {topic}\")\n\n \n boxes_ids = tracker.update(detection_list)\n \n for box_id in boxes_ids:\n x, y, w, h, id, vehicle_type = box_id\n \n temp = next((obj for obj in detected_objects if obj.id == id), None)\n \n if temp == None:\n detected_objects.append(Detected(x, y, w, h, id, vehicle_type))\n\n else:\n prev_cx, prev_cy = temp.cx, temp.cy\n\n print(prev_cx,prev_cy)\n \n temp.update(x, y, w, h)\n\n print(temp.cx,temp.cy)\n\n img = cv2.circle(img, (prev_cx,prev_cy), 1, (255,0,0), 10)\n img = cv2.arrowedLine(img,(prev_cx,prev_cy),(temp.cx,temp.cy),(0,255,0),2)\n #cv2.putText(img,\"Motion Vector\",(prev_cx+10,prev_cy),font,1,(255,255,0),2)\n temp.get_ingress(datetime.now().strftime(\"%H:%M:%S:%f\"))\n temp.get_exgress(datetime.now().strftime(\"%H:%M:%S:%f\"))\n \n \n cv2.putText(img,\"ID:\"+str(id),(x,y -15),font,1,(2550,0),2)\n\n exgress_times = {}\n ingress_times = {}\n pet_calc = {}\n if len(boxes_ids)>1:\n for box_id in boxes_ids:\n _,_,_,_,id,_ = box_id\n temp = next((obj for obj in detected_objects if obj.id == id), None)\n if temp != None:\n exgress_times[id] = temp.exgress\n ingress_times[id] = temp.ingress\n \n # ! Just Showing some stuff will delete later\n print(\"This is exgress times = \")\n print(exgress_times)\n print(\"This is ingress times = \")\n print(ingress_times)\n \n print(datetime.now())\n tp2 = 650\n for (k1,v1) in exgress_times.items():\n for (k2,v2) in ingress_times.items():\n if k1 != k2:\n if v1 != None and v2 != None:\n pet_calc[k2] = v2 - v1\n pet_text = f\"PET: ID{k2}->ID{k1}={abs(round((v2-v1).total_seconds(),2))}s\"\n print(pet_text)\n cv2.putText(img,pet_text,(900,tp2),font,1,(0,0,255),2)\n tp2 += 30\n\n dt=time.time()-timeStamp\n timeStamp=time.time()\n fps=1/dt\n fpsFilt=.9*fpsFilt + .1*fps\n\n cv2.putText(img,str(round(fpsFilt,1))+' fps',(0,30),font,1,(255,255,0),2)\n\n # Yellow Box\n overlay = img.copy()\n cv2.rectangle(overlay,(150,110),(900,680),(0,255,255),-1)\n alpha = 0.4\n img = image_new = cv2.addWeighted(overlay, alpha, img, 1 - alpha, 0)\n\n # Uncomment for saving the output file \n #result.write(img)\n #cv2.imshow(\"Frame\",img)\n\n #if cv2.waitKey(1) & 0xFF == ord('q'):\n # break\n \ncap.release()\ncv2.destroyAllWindows()\nclient.loop_stop()\n\ndf = pd.DataFrame(columns=['id','vehicle_type','ingress_time','egress_time'])\n\nfor x in detected_objects:\n if x.ingress != None or x.exgress != None:\n\n new_row = {'id':x.id,'vehicle_type':x.vehicle_type, 'ingress_time':x.ingress.strftime(\"%H:%M:%S:%f\"),'egress_time':x.exgress.strftime(\"%H:%M:%S:%f\")}\n df = df.append(new_row,ignore_index=True)\n #print(f\"Object ID {x.id}, Object Type = {x.vehicle_type}, Ingress = {x.ingress.strftime(\"%H:%M:%S:%f\")}, Egress = {x.exgress.strftime(\"%H:%M:%S:%f\")}\")\n\ndf.to_csv(\"results.csv\",index=False)\nprint(df.set_index('id').dropna())\n","repo_name":"cpaolini/itmc","sub_path":"jetson/track_detect.py","file_name":"track_detect.py","file_ext":"py","file_size_in_byte":10271,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37617704269","text":"from django.conf import settings\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.views import View\nfrom django.views.generic.edit import UpdateView\nfrom django.views.generic.edit import DeleteView\nfrom django.views.generic.detail import DetailView\nfrom django.views.generic import TemplateView\nfrom django.views.generic.base import TemplateResponseMixin\n\nfrom django.db.models import Sum\nfrom django.db.models import Count\nfrom django.db.models import Case, When, Value, IntegerField\nfrom django.urls import reverse_lazy\nfrom django.utils import timezone\nfrom django.shortcuts import render, redirect\n\nfrom aps.models import Person\nfrom aps.models import Circle\nfrom aps.models import Role\nfrom aps.models import RoleFiller\n\n\n\n# Helper methods ########################################################################################\n\ndef getObjectsNotInGlassfrog(modelObject, orderBy='name'):\n # get newest import date in db table (as last date of import)\n last_import_date = getLastImportedDate(modelObject)\n object_list = modelObject.objects.filter(last_imported__lt=last_import_date).order_by(orderBy).annotate(is_deleted_in_glassfrog=Value(1, IntegerField()))\n return object_list\n\n\ndef getLastImportedDate(modelObject):\n last_imported_objects = modelObject.objects.order_by('-last_imported')\n if last_imported_objects.exists():\n last_imported_object = last_imported_objects[0]\n last_import_date = last_imported_object.last_imported\n else:\n last_import_date = timezone.now()\n return last_import_date\n\n\ndef sumField(object_list, field_to_sum):\n return round(object_list.aggregate(Sum(field_to_sum))[field_to_sum + '__sum'], 1)\n\n\ndef prepareObjectListContext(object_list, field_to_sum):\n total = 0.0\n if object_list.exists():\n total = sumField(object_list, field_to_sum)\n context = {\n 'total': total,\n 'object_list': object_list\n }\n return context\n\n\ndef preparePersonListContext(object_list):\n # get generic additional context\n context = prepareObjectListContext(object_list, 'contract_fte')\n # add person specific additional context\n try: # can only count/sum when the object_list has related items (i.e. rolefillers) in the query result set. That is not always the case, depending on the query done.\n context['total_roles'] = sumField(object_list, 'rolefiller__count')\n context['total_attention_points'] = sumField(object_list, 'rolefiller__attention_points__sum')\n context['total_utilisation'] = Person.calculate_expected_utilisation_percentage(\n context['total_attention_points'], context['total'])\n except:\n pass # dont worry, if we don't have the totals, we dont show them\n return context\n\n\ndef getAnnotatedPersons():\n return Person.objects.annotate(\n Count('rolefiller'),\n # todo: add Round() function to get rid of rounding errors in the sum function\n # docs: https://docs.djangoproject.com/es/1.10/ref/models/expressions/#func-expressions\n Sum('rolefiller__attention_points'),\n Count('rolefiller__role__circle', distinct=True)\n # only count each circle once, when someone has more than one role in a circle\n )\n\ndef getPersonDetailAdditionalContext(person_object):\n additional_context = {}\n # get data related to this person\n annotated_person_object = getAnnotatedPersons().get(id=person_object.id)\n circles = Circle.objects.filter(roles__rolefiller__person__id=person_object.id).distinct().order_by('name')\n leadlink_roles = RoleFiller.objects.all().filter(role__name=\"Lead Link\").filter(person__id=person_object.id)\n leadlink_circles = []\n\n for circle in circles:\n # get all roles this person has in this circle\n circle.person_roles = RoleFiller.objects.filter(role__circle__id=circle.id, person__id=person_object.id).order_by('role')\n circle.total_attention_points = 0\n for rolefill in circle.person_roles:\n circle.total_attention_points += rolefill.attention_points\n\n circle.total_attention_points = round(circle.total_attention_points, 2) # python can be weird with floats\n\n # if this person is Lead Link of this circle, store this circle in his leadlink_circles list\n if leadlink_roles.filter(role__circle=circle):\n circle.additional_context = prepareSubCircleListContextRecursive(circle_object=circle)\n leadlink_circles.append(circle)\n\n # add data to the additional_context, before it is passed to the template\n additional_context['circle_count'] = annotated_person_object.rolefiller__role__circle__count\n additional_context['rolefiller_count'] = annotated_person_object.rolefiller__count\n additional_context['attention_points_sum'] = annotated_person_object.rolefiller__attention_points__sum\n additional_context['circles'] = circles\n additional_context['circles_leadlink'] = leadlink_circles\n\n return additional_context\n\ndef getCircleDetailAdditionalContext(circle__object):\n last_import_date_rolefiller = getLastImportedDate(RoleFiller)\n additional_context = {}\n # add a breadcrumb list of parent circles\n additional_context['breadcrumbs'] = getBreadCrumbList(circle__object)\n # add data to the additional_context, before it is passed to the template\n # store sub-circles, and their total granted and assigned attention points\n additional_context['circles'] = prepareSubCircleListContextRecursive(circle__object)\n # store rolefillers and their total assigned attention points\n additional_context['rolefillers'] = prepareObjectListContext(\n RoleFiller.objects.filter(role__circle__pk=circle__object.pk).order_by(\"person__name\", \"role__name\").annotate(\n is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_rolefiller,then=Value(1)),default=Value(0),output_field=IntegerField())),\n field_to_sum='attention_points')\n # store total granted, assigned attention points\n additional_context['attention_points_granted'] = circle__object.attention_points # just an alias of the objects property value\n additional_context['attention_points_assigned'] = round(additional_context['circles']['attention_points_assigned'] + additional_context['rolefillers']['total'], 1)\n additional_context['attention_points_balance'] = round(additional_context['attention_points_granted'] - additional_context['attention_points_assigned'], 1)\n # add Unassigned Roles to additional_context\n additional_context['unassigned_roles'] = Role.objects.filter(circle__pk=circle__object.pk, rolefiller__isnull=True)\n return additional_context\n\n\ndef prepareSubCircleListContextRecursive(circle_object=None, circle_list=None, field_to_sum='attention_points', subcircle_list=None):\n last_import_date_circle = getLastImportedDate(Circle)\n if subcircle_list != None:\n sub_circle_list = subcircle_list\n elif circle_object == None:\n if circle_list == None:\n circle_object = getRootCircle() # needed to get attention points\n sub_circle_list = Circle.objects.all().annotate(is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_circle,then=Value(1)),default=Value(0),output_field=IntegerField()))\n else:\n sub_circle_list = circle_list\n else:\n sub_circle_list = circle_object.sub_circles.all().annotate(is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_circle,then=Value(1)),default=Value(0),output_field=IntegerField()))\n\n context = {\n 'total': 0,\n 'object_list': sub_circle_list\n }\n # get total assigned attention points within subcircles (recursive)\n total_assigned = 0.0\n for sub_circle in sub_circle_list:\n assigned_in_sub_circle = 0.0\n # add this sub_circles rolefiller assigned points\n circle_rolefiller_list = RoleFiller.objects.filter(role__circle__pk=sub_circle.pk)\n if circle_rolefiller_list.exists():\n assigned_in_sub_circle += sumField(circle_rolefiller_list, field_to_sum)\n else:\n assigned_in_sub_circle += 0\n # add this circles sub-circles assigned points (recursive)\n list_context = prepareSubCircleListContextRecursive(sub_circle, None, field_to_sum)\n assigned_in_sub_circle += list_context['attention_points_assigned']\n total_assigned += assigned_in_sub_circle\n # store values in the sub_circle object, to be retrieved in the template\n sub_circle.attention_points_granted = round(sub_circle.attention_points, 1) # alias\n sub_circle.attention_points_assigned = round(assigned_in_sub_circle, 1)\n sub_circle.attention_points_balance = round(sub_circle.attention_points_granted - sub_circle.attention_points_assigned, 1)\n # Add additional circle list specific data to the circle object, to retrieve in the template\n context['attention_points_granted'] = round(context['total'], 1) # alias for conveneance\n context['attention_points_assigned'] = round(total_assigned, 1)\n context['attention_points_balance'] = round(context['total'] - total_assigned, 1)\n # context['total_assigned'] = total_assigned\n # context['total_balance'] = circle_object.attention_points - total_assigned\n return context\n\n\ndef getRootCircle():\n # query all circles without a super-circle. These could also be deleted circles\n # therefore we inverse order by glassfrog ID, as the root circle has PROBABLY the lowest glassfrog ID.\n root_circle = Circle.objects.filter(super_circle__isnull=True).order_by('glassfrog_id').first()\n return root_circle\n\n\ndef getBreadCrumbList(circle_object, bread_crumb_list=None):\n if not bread_crumb_list:\n bread_crumb_list = []\n parent = circle_object.super_circle\n if parent:\n bread_crumb_list.append(parent)\n bread_crumb_list = getBreadCrumbList(parent, bread_crumb_list)\n return bread_crumb_list\n\n\ndef debug(ctx):\n try:\n fail\n except:\n raise Exception(ctx)\n pass\n\ndef getLeadLinkDetails(self, filter_to_use, include_GCC = False):\n # Check if the current user is a leadlink of any circles, and pass the details of the sub-circles to the view\n person = Person.objects.get(user=self.request.user.id)\n leadlink_role_of_circles = RoleFiller.objects.filter(person=person).filter(role__name=\"Lead Link\")\n if leadlink_role_of_circles.exists():\n circles = []\n for rolefiller in leadlink_role_of_circles:\n # Find all the sub-circles where they are a leadlink\n if filter_to_use == \"super_circle\":\n filter = Circle.objects.filter(super_circle__name=rolefiller.role.circle)\n elif filter_to_use == \"rolefiller\":\n filter = Circle.objects.filter(name=rolefiller.role.circle)\n else:\n filter = None\n\n circles_to_add = filter\n for circle in circles_to_add:\n circles.append(circle)\n\n if include_GCC:\n # Special case for GCC - GCC LL should be able to set GCC points\n if rolefiller.role.circle.name == \"General Company Circle\":\n circles.append(Circle.objects.get(name=rolefiller.role.circle))\n\n return circles\n else:\n return None\n\ndef getLeadlinkCircleDetails(self):\n return getLeadLinkDetails(self, \"super_circle\", True)\n\ndef getLeadLinkRolefillerDetails(self):\n return getLeadLinkDetails(self, \"rolefiller\")\n\n# VIEW CLASSES #########################################################################################################\n\n# List View Classes #############################\nclass BullfrogListView(LoginRequiredMixin, View):\n context = {}\n login_url = reverse_lazy('login')\n template_name = '' # should override in subclasses\n\n def get_template_names(self):\n # get and validate requested output format\n output = self.request.GET.get('output') \\\n if self.request.GET.get('output') in ['html','json'] \\\n else 'html'\n # change default template to requested output version\n template_name = self.template_name.rsplit( \".\", 1 )[ 0 ] + \".\" + output\n return [template_name]\n\n def render(self, request):\n if RoleFiller.objects.filter(role__name=\"Lead Link\", person__user=self.request.user.id):\n self.context['is_leadlink'] = True\n else:\n self.context['is_leadlink'] = False\n\n return render(request, self.get_template_names()[0], self.context)\n\n\nclass Index(BullfrogListView):\n template_name = 'index.html'\n\n def get(self, request):\n\n # If the user still has the default password he has to change it\n if request.user.check_password(settings.DEFAULT_PASSWORD):\n return redirect('password_change')\n\n\n # select PEOPLE that have no fte set (usually because the were recently imported from Glassfrog)\n object_list = Person.objects.filter(contract_fte=0).order_by('name')\n self.context['persons_without_fte'] = preparePersonListContext(object_list)\n\n # select CIRCLES that have no attention points set (usually because they were recently imported from Glassfrog)\n object_list = Circle.objects.filter(attention_points=0).order_by('name')\n # self.context['circles_without_attention_points'] = prepareCircleListContext(object_list, field_to_sum='attention_points')\n self.context['circles_without_attention_points'] = prepareSubCircleListContextRecursive(None, object_list)\n\n # select ROLEFILLERS that have no attention points set (usually because they were recently imported from Glassfrog)\n object_list = RoleFiller.objects.filter(attention_points=0).order_by('role')\n self.context['rolefillers_without_attention_points'] = prepareObjectListContext(object_list, field_to_sum='attention_points')\n\n # select PEOPLE that are no longer in Glassfrog\n object_list = getObjectsNotInGlassfrog(Person)\n self.context['persons_not_in_glassfrog'] = preparePersonListContext(object_list)\n\n # select CIRCLES that are no longer in Glassfrog\n object_list = getObjectsNotInGlassfrog(Circle)\n self.context['circles_not_in_glassfrog'] = prepareObjectListContext(object_list,\n field_to_sum='attention_points')\n\n # select ROLEFILLER that are no longer in Glassfrog\n object_list = getObjectsNotInGlassfrog(RoleFiller, orderBy='role')\n self.context['rolefillers_not_in_glassfrog'] = prepareObjectListContext(object_list,\n field_to_sum='attention_points')\n\n # select ROLES that are no longer in Glassfrog\n object_list = getObjectsNotInGlassfrog(Role)\n self.context['roles_not_in_glassfrog'] = prepareObjectListContext(object_list,\n field_to_sum='glassfrog_id') # summing up glassfrog_id is a 'hack'. it wont be used, but makes it possible to reuse the generic method\n\n # get information relevant for lead links\n self.context['leadlink_subcircles'] = getLeadlinkCircleDetails(self)\n self.context['leadlink_circle_rolefillers'] = getLeadLinkRolefillerDetails(self)\n\n return self.render(request)\n\n\nclass PeopleView(BullfrogListView):\n template_name = 'people.html'\n def get(self, request):\n last_import_date_person = getLastImportedDate(Person)\n object_list = getAnnotatedPersons().order_by('name').annotate(\n is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_person,then=Value(1)),default=Value(0),output_field=IntegerField()))\n self.context['persons'] = preparePersonListContext(object_list)\n\n return self.render(request)\n\n\nclass CirclesView(BullfrogListView):\n template_name = 'circles.html'\n def get(self, request):\n last_import_date_circle = getLastImportedDate(Circle)\n object_list = Circle.objects.order_by('name').annotate(\n is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_circle,then=Value(1)),default=Value(0),output_field=IntegerField()))\n self.context['circles'] = prepareSubCircleListContextRecursive(None, object_list)\n self.context['leadlink_subcircles'] = getLeadlinkCircleDetails(self)\n\n return self.render(request)\n\n\n\nclass RoleFillersView(BullfrogListView):\n template_name = 'rolefillers.html'\n def get(self, request):\n last_import_date_rolefiller = getLastImportedDate(RoleFiller)\n object_list = RoleFiller.objects.order_by('role__name', 'person__name').annotate(\n is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_rolefiller,then=Value(1)),default=Value(0),output_field=IntegerField()))\n self.context['rolefillers'] = prepareObjectListContext(object_list, 'attention_points')\n self.context['leadlink_circle_rolefillers'] = getLeadLinkRolefillerDetails(self)\n\n return self.render(request)\n\n\nclass RolesView(BullfrogListView):\n template_name = 'roles.html'\n def get(self, request):\n last_import_date_role = getLastImportedDate(Role)\n object_list = Role.objects.order_by('circle', 'name').annotate(\n is_deleted_in_glassfrog=Case(When(last_imported__lt=last_import_date_role,then=Value(1)),default=Value(0),output_field=IntegerField()))\n self.context['roles'] = prepareObjectListContext(object_list, 'glassfrog_id') # hack: will create an unused sum of glassfrog_id's, to be able to reuse the generic method.\n return self.render(request)\n\n\n# Update View Classes ###################################\nclass BullfrogUpdateView(LoginRequiredMixin, UpdateView):\n login_url = reverse_lazy('login')\n template_name = 'generic_update_form.html'\n fields = ['attention_points'] # default, should be overridden by subclass if you want something else\n delete_url = '_delete' # placeholder, should be overridden by subclass with something like 'person_delete'\n success_url = reverse_lazy(\n 'index') # where to return after update, can be overridden by subclass, or by passing a ?next= url parameter\n\n def get_success_url(self):\n # look in the url for the ?next= paramater, and use that to redirect on success, or default to whatever the base class returns\n return self.request.GET.get('next', super(UpdateView, self).get_success_url())\n\n\nclass PersonUpdate(BullfrogUpdateView):\n model = Person\n delete_url = 'person_delete'\n fields = ['contract_fte']\n\n\nclass CircleUpdate(BullfrogUpdateView):\n model = Circle\n delete_url = 'circle_delete'\n\n\nclass RoleFillerUpdate(BullfrogUpdateView):\n model = RoleFiller\n delete_url = 'rolefiller_delete'\n\n\nclass RoleUpdate(BullfrogUpdateView):\n model = Role\n delete_url = 'role_delete'\n success_url = reverse_lazy('roles')\n fields = [] # overriding the default\n\n\n# Delete View Classes ###################################\nclass BullfrogDeleteView(LoginRequiredMixin, DeleteView):\n login_url = reverse_lazy('login')\n template_name = 'generic_delete_form.html'\n success_url = reverse_lazy('index')\n\n\nclass PersonDelete(BullfrogDeleteView):\n model = Person\n\n\nclass CircleDelete(BullfrogDeleteView):\n model = Circle\n\n\nclass RoleFillerDelete(BullfrogDeleteView):\n model = RoleFiller\n\n\nclass RoleDelete(BullfrogDeleteView):\n model = Role\n\n\n# Detail View Classes #############################\nclass BullfrogDetailView(LoginRequiredMixin, DetailView):\n login_url = reverse_lazy('login')\n template_name = '' # should override in subclasses\n\n def get_template_names(self):\n # get and validate requested output format\n output = self.request.GET.get('output') \\\n if self.request.GET.get('output') in ['html','json'] \\\n else 'html'\n # change default template to requested output version\n template_name = self.template_name.rsplit( \".\", 1 )[ 0 ] + \".\" + output\n return [template_name]\n\n\nclass PersonDetailView(BullfrogDetailView):\n model = Person\n template_name = \"person.html\"\n\n def get_context_data(self, **kwargs):\n # get the context for this object that the super class auto generates (object details)\n context = super(PersonDetailView, self).get_context_data(**kwargs)\n # add additional data to the context, before it is passed to the template\n context['additional_context'] = getPersonDetailAdditionalContext(self.object)\n\n return context\n\n\nclass CircleDetailView(BullfrogDetailView):\n model = Circle\n template_name = \"circle.html\"\n\n def get_context_data(self, **kwargs):\n # get the context for this object that the super class auto generates (object details)\n context = super(CircleDetailView, self).get_context_data(**kwargs)\n # add additional data to the context, before it is passed to the template\n context['additional_context'] = getCircleDetailAdditionalContext(self.object)\n context['leadlink_subcircles'] = getLeadlinkCircleDetails(self)\n context['leadlink_circle_rolefillers'] = getLeadLinkRolefillerDetails(self)\n\n\n rolefillers = context['additional_context']['rolefillers']['object_list']\n personRolesList = []\n people = []\n for obj in rolefillers:\n if not obj.person.pk in people:\n people.append(obj.person.pk)\n personRoles = rolefillers.filter(person__pk=obj.person.pk)\n personRolesList.append(prepareObjectListContext(personRoles,'attention_points'))\n\n context['additional_context']['roles_per_person'] = personRolesList\n\n return context\n\n\nclass MeView(PersonDetailView):\n template_name = \"me.html\"\n\n def get(self, request):\n # If the user still has the default password he has to change it\n if request.user.check_password(settings.DEFAULT_PASSWORD):\n return redirect('password_change')\n return super(MeView, self).get(request)\n\n # Get the Person object belonging to the logged in user as the object for this view (instead of passing an object via url, like is the default for a detail view)\n def get_object(self):\n return Person.objects.get(user=self.request.user.id)\n\n def get_context_data(self, **kwargs):\n # get the context for this object that the super class auto generates (object details)\n context = super(MeView, self).get_context_data(**kwargs)\n # add additional data to the context, before it is passed to the template\n context['additional_context'] = getPersonDetailAdditionalContext(self.object)\n\n # get information relevant for lead links\n if RoleFiller.objects.filter(role__name=\"Lead Link\", person__user=self.request.user.id):\n context['is_leadlink'] = True\n leadlink_of_circles = context['additional_context']['circles_leadlink']\n\n def get_object_from_leadlinks_circles(object, role_or_circle):\n relevant_objects = []\n\n for circle in leadlink_of_circles:\n try:\n if role_or_circle == \"role\":\n circle_objects = object.filter(role__circle=circle)\n elif role_or_circle == \"supercircle\":\n circle_objects = object.filter(super_circle=circle)\n else:\n circle_objects = object.filter(circle=circle)\n\n for circle_object in circle_objects:\n relevant_objects.append(circle_object)\n except:\n pass\n\n return relevant_objects\n\n # show roles not in glassfrog\n roles_not_in_gf = getObjectsNotInGlassfrog(Role)\n context['roles_not_in_glassfrog'] = get_object_from_leadlinks_circles(roles_not_in_gf, \"circle\")\n\n # show rolefillers that have no attention points\n roles_with_no_ap = RoleFiller.objects.filter(attention_points=0).order_by('role')\n context['rolefillers_without_attention_points'] = get_object_from_leadlinks_circles(roles_with_no_ap, \"role\")\n\n # select rolefillers that are no longer in Glassfrog\n rolefillers_not_in_gf = getObjectsNotInGlassfrog(RoleFiller, orderBy='role')\n context['rolefillers_not_in_glassfrog'] = get_object_from_leadlinks_circles(rolefillers_not_in_gf, \"role\")\n\n # show circles that are no longer in Glassfrog\n circles_not_in_gf = getObjectsNotInGlassfrog(Circle)\n context['circles_not_in_glassfrog'] = get_object_from_leadlinks_circles(circles_not_in_gf, \"circle\")\n\n # show circles that have no attention points\n circles_with_no_ap = Circle.objects.filter(attention_points=0).order_by('name')\n relevant_circles_with_no_ap = get_object_from_leadlinks_circles(circles_with_no_ap, \"supercircle\")\n circle_list_prepared = prepareSubCircleListContextRecursive(subcircle_list=relevant_circles_with_no_ap)\n context['circles_without_attention_points'] = circle_list_prepared\n\n # get details of the circles that they are lead link\n leadlink_circles_list = prepareSubCircleListContextRecursive(subcircle_list=leadlink_of_circles)\n context['leadlink_circles'] = leadlink_circles_list\n\n context['leadlink_subcircles'] = getLeadlinkCircleDetails(self)\n context['leadlink_circle_rolefillers'] = getLeadLinkRolefillerDetails(self)\n\n return context\n\n\n# IMPORT VIEW ##########################################################################################################\n\nfrom aps.utils.GlassfrogImporter import GlassfrogImporter\n\nclass DoImport(LoginRequiredMixin, View):\n login_url = reverse_lazy('login')\n\n def get(self, request):\n imp = GlassfrogImporter()\n imp.doImport()\n\n numPeopleInGlassfrog = len(imp.people)\n numPeopleInBullfrog = len(Person.objects.all())\n successPeople = True if numPeopleInGlassfrog <= numPeopleInBullfrog else False\n\n numCirclesInGlassfrog = len(imp.circles)\n numCirclesInBullfrog = len(Circle.objects.all())\n successCircles = True if numCirclesInGlassfrog <= numCirclesInBullfrog else False\n\n numRolesInGlassfrog = len(imp.roles)\n numRolesInBullfrog = len(Role.objects.all())\n successRoles = True if numRolesInGlassfrog <= numRolesInBullfrog else False\n\n numRoleFillersInGlassfrog = len(imp.rolefillers)\n numRoleFillersInBullfrog = len(RoleFiller.objects.all())\n successRoleFillers = True if numRoleFillersInGlassfrog <= numRoleFillersInBullfrog else False\n\n messages = [\n 'People Import successful:' + str(successPeople),\n 'Number of people in Glassfrog:' + str(numPeopleInGlassfrog),\n 'Number of people in Bullfrog:' + str(numPeopleInBullfrog),\n '',\n 'Circles Import successful:' + str(successCircles),\n 'Number of circles in Glassfrog:' + str(numCirclesInGlassfrog),\n 'Number of circles in Bullfrog:' + str(numCirclesInBullfrog),\n '',\n 'Roles Import successful:' + str(successRoles),\n 'Number of roles in Glassfrog:' + str(numRolesInGlassfrog),\n 'Number of roles in Bullfrog:' + str(numRolesInBullfrog),\n '',\n 'Role Fillers Import successful:' + str(successRoleFillers),\n 'Number of role fillers in Glassfrog:' + str(numRoleFillersInGlassfrog),\n 'Number of role fillers in Bullfrog:' + str(numRoleFillersInBullfrog),\n ]\n\n return render(request, 'import.html', {'messages': messages})\n\n# ABOUT VIEW ##########################################################################################################\n\nclass AboutView(TemplateView):\n context = {}\n template_name = \"about.html\"\n","repo_name":"themobilecompany/Bullfrog","sub_path":"aps/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":28081,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"23542919577","text":"from datetime import datetime\n\nfrom todo_detail_view import TodoDetailView\nfrom constants import *\nfrom model import Repeat\nfrom todo_view_model import TodoViewModel\n\nDAY_STR = ('월', '화', '수', '목', '금', '토', '일')\n\n\nclass TodoListView:\n\n def __init__(self):\n self.viewmodel = TodoViewModel()\n\n def list(self):\n self.printlist()\n\n def search(self):\n print('검색할 항목을 선택하세요')\n print('1. 제목')\n print('2. 완료 여부')\n print('3. 진행도')\n print('4. 중요도')\n print('5. 내용')\n print('6. 날짜')\n choice = int(input('>>> '))\n\n if choice == 1:\n name = input('제목: ')\n self.printlist(name=name)\n\n if choice == 2:\n done = input('완료 여부(True/False): ')\n self.printlist(done=(done == 'True'))\n\n if choice == 3:\n try:\n progress = int(input('진행도: '))\n self.printlist(progress=progress)\n except ValueError:\n print(\"정수를 입력하세요.\")\n\n if choice == 4:\n try:\n importance = int(input('중요도: '))\n self.printlist(importance=importance)\n except ValueError:\n print(\"정수를 입력하세요.\")\n\n if choice == 5:\n content = input('내용: ')\n self.printlist(content=content)\n\n if choice == 6:\n try:\n date = datetime.strptime(input('날짜 (yyyy-mm-dd): '), '%Y-%m-%d').date()\n self.printlist(date=date)\n except ValueError:\n print('잘못된 형식입니다.')\n\n def printlist(self, **kwargs):\n for obj in self.viewmodel.get_list(**kwargs):\n print(f'날짜: {obj.date.strftime(\"%Y-%m-%d\")}')\n print(f'제목: {obj.name}')\n print(f'완료 여부: {obj.done}')\n print(f'진행도: {obj.progress}')\n print(f'중요도: {obj.importance}')\n if obj.repeat:\n print(f'반복 날짜: {\",\".join(day for i, day in enumerate(DAY_STR) if 1 << i & obj.repeat.day)}')\n print(f'반복 주기(주): {obj.repeat.week_interval}')\n print(f'만료 날짜: {obj.repeat.due.strftime(\"%Y-%m-%d\")}')\n print(f'내용: {obj.content}')\n print()\n\n while True:\n print('TODO 상세보기: {id}')\n print('종료: 엔터')\n s = input('>>> ')\n if not s:\n break\n id = int(s)\n TodoDetailView(id).load()\n print()\n\n def create(self):\n try:\n date = datetime.strptime(input('날짜 (yyyy-mm-dd): '), '%Y-%m-%d').date()\n except ValueError:\n print('잘못된 형식입니다.')\n return\n name = input('이름: ')\n content = input('내용: ')\n has_repeat = input('반복 여부(True/False): ')\n repeat = None\n if has_repeat == 'True':\n try:\n day = sum(1 << DAY_STR.index(d) for d in input('요일 (콤마로 구분): ').split(','))\n except ValueError:\n print('잘못된 형식입니다.')\n return\n try:\n week_interval = int(input('반복 간격 (주): '))\n except ValueError:\n print(\"정수를 입력하세요.\")\n return\n try:\n due = datetime.strptime(input('만료 날짜 (yyyy-mm-dd): '), '%Y-%m-%d').date()\n except ValueError:\n print('잘못된 형식입니다.')\n return\n repeat = Repeat(day=day, week_interval=week_interval, due=due)\n\n try:\n progress = int(input('진행도: '))\n except ValueError:\n print(\"정수를 입력하세요.\")\n return\n try:\n importance = int(input('중요도: '))\n except ValueError:\n print(\"정수를 입력하세요.\")\n return\n v = self.viewmodel.create(date, name, content, repeat=repeat,\n done=False, progress=progress, importance=importance)\n if v == CODE_INVALID_DATE:\n print('날짜가 잘못되었습니다.')\n elif v == CODE_INVALID_PROGRESS:\n print('진행도가 잘못되었습니다.')\n elif v == CODE_INVALID_IMPORTANCE:\n print('중요도가 잘못되었습니다.')\n\n def load(self):\n\n while True:\n print('1. 목록 출력')\n print('2. 검색')\n print('3. 추가')\n print('4. 종료')\n choice = int(input('>>> '))\n\n if choice == 1:\n self.list()\n if choice == 2:\n self.search()\n if choice == 3:\n self.create()\n if choice == 4:\n break\n\n\nif __name__ == '__main__':\n todo = TodoListView()\n todo.load()\n","repo_name":"Over-the-SNU/Time-Prototype","sub_path":"todo_view.py","file_name":"todo_view.py","file_ext":"py","file_size_in_byte":5049,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"30941554612","text":"from testsutes.base_tecase import BaseTestCase\nfrom framework.Logger import Logger\nfrom memorandum.homepage import HomePage\nimport unittest\nimport appium\nlogger=Logger(\"register\").getlog()\nclass testRegister(BaseTestCase):\n def test_Register(self):\n self.hp=HomePage(self.driver)\n self.hp.register(\"LOL\",\"LOL@qq.com\",\"7654321\")\n try:\n self.assertEqual(\"智能备忘录\",self.driver.find_element(*self.hp.assertmemo).text)\n logger.info(\"注册成功\")\n print(\"jjjjjjjjjjjj\")\n except Exception as e:\n logger.error(\"注册失败\")\n print(\"hhhhhhhhhhhhhhhh\")\n self.hp.exit_login()\nif __name__==\"__main__\":\n unittest.main(verbosity=2)\n\n","repo_name":"15935622817/appium","sub_path":"testsutes/test_register.py","file_name":"test_register.py","file_ext":"py","file_size_in_byte":728,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"30209933790","text":"from os import path\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfrom yieldenv.settings import PROJECT_ROOT\n\nRB_FACTOR = 25\nRS_FACTOR = 50\n# EXPONENT = 0.7\n\n\ndef borrow_lend_rates(\n util_rate: float,\n rb_factor: float = RB_FACTOR,\n rs_factor: float = RS_FACTOR,\n) -> tuple[float, float]:\n \"\"\"\n calculate borrow and supply rates based on utilization ratio\n with an arbitrarily-set shape\n \"\"\"\n\n assert (\n 0 <= util_rate < 1\n ), f\"utilization ratio must lie in [0,1), but got {util_rate}\"\n\n borrow_rate = util_rate / (rb_factor * (1 - util_rate))\n # initial_borrow_rate / (1 - util_rate) ** EXPONENT\n supply_rate = util_rate / (rs_factor * (1 - util_rate))\n # initial_supply_rate / (1 - util_rate) ** EXPONENT\n return borrow_rate, supply_rate\n\n\nif __name__ == \"__main__\":\n # exploratory plot\n util_rates = np.concatenate(\n [np.arange(0, 0.83, step=0.02), np.arange(0.835, 0.9999, step=0.0001)]\n )\n\n borrow_rates = []\n lend_rates = []\n\n for u in util_rates:\n r1, r2 = borrow_lend_rates(u)\n borrow_rates.append(r1)\n lend_rates.append(r2)\n\n plt.rcParams.update({\"font.size\": 15})\n\n plt.plot(\n util_rates,\n borrow_rates,\n label=f\"borrow interest rate $r_b=\\\\frac{{u}}{{{RB_FACTOR} \\\\times (1-u)}}$\",\n )\n plt.plot(\n util_rates,\n lend_rates,\n label=f\"supply interest rate $r_s=\\\\frac{{u}}{{{RS_FACTOR} \\\\times (1-u)}}$\",\n )\n plt.xlabel(\"utilization ratio $u$\")\n plt.ylabel(\"interest rate per annum $r$\")\n plt.xlim(0, 1)\n plt.ylim(0, 1.6)\n plt.legend(frameon=False)\n plt.tight_layout()\n fig_path = path.join(PROJECT_ROOT, f\"assets/interest_model.pdf\")\n plt.savefig(fig_path)\n plt.show()\n plt.close()\n","repo_name":"xujiahuayz/yieldAggregators","sub_path":"yieldenv/interest_rate.py","file_name":"interest_rate.py","file_ext":"py","file_size_in_byte":1780,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"7"} +{"seq_id":"1802013995","text":"\"\"\"\nМодель таблицы.\n\"\"\"\n\nfrom sqlalchemy import CheckConstraint, Column, Date, Integer, String\nfrom sqlalchemy.engine import Engine\nfrom sqlalchemy.ext.declarative import declarative_base\n\nBase = declarative_base()\n\n\nclass Data(Base):\n \"\"\"\n Класс для таблицы, где мы будем хранить данные.\n \"\"\"\n\n __tablename__ = \"data\"\n id = Column(\"id\", Integer, primary_key=True)\n company = Column(\"company\", String, nullable=False)\n metric = Column(\"metric\", String)\n date = Column(\"date\", Date)\n fact = Column(\"fact\", Integer)\n forecast = Column(\"forecast\", Integer)\n check = CheckConstraint(\"metric = 'Qliq' or metric = 'Qoil'\")\n\n def __repr__(self) -> str:\n return f\"\"\"\n {self.date} {self.company} получила {self.fact} {self.metric}\n Прогноз: {self.forecast}\n \"\"\"\n\n\ndef create_tables(engine: Engine) -> None:\n \"\"\"\n Создаёт таблицу (предварительно удалив, если она уже существует).\n\n Args:\n engine: двигатель SQLAlchemy\n \"\"\"\n Base.metadata.drop_all(engine)\n Base.metadata.create_all(engine)\n","repo_name":"vyhuholl/test_backend_developer","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1197,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37999374858","text":"import pandas as pd\nimport xgboost as xgb\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import accuracy_score, confusion_matrix\n\nglh_df = pd.read_csv(r'C:\\Users\\LENOVO\\CV\\textile-defect-inspection\\dataset\\glh_features.csv')\nglcm_df = pd.read_csv(r'C:\\Users\\LENOVO\\CV\\textile-defect-inspection\\dataset\\glcm_features.csv')\n\ndf = pd.concat([glcm_df,glh_df],axis=1)\n\nX = df.iloc[:,:-1]\ny = df['label']\n\nX_train, X_test, y_train, y_test = train_test_split(X,y, test_size=0.3, random_state=None)\n\nscaler = MinMaxScaler()\nX_train = scaler.fit_transform(X_train)\nX_test = scaler.transform(X_test)\n\nxgb_classifier = xgb.XGBClassifier()\nxgb_classifier.fit(X_train,y_train)\n\ny_pred = xgb_classifier.predict(X_test)\n\nprint(accuracy_score(y_test, y_pred))\nprint(confusion_matrix(y_test, y_pred))","repo_name":"soucs/textile-defect-inspection","sub_path":"xgboost.py","file_name":"xgboost.py","file_ext":"py","file_size_in_byte":857,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"69957641505","text":"import pygame\npygame.init()\nclass button:\n def __init__(self,\n x = None,\n y = None,\n w = None,\n h = None,\n c = None):\n self.x = x\n self.y = y\n self.h = h\n self.w = w\n self.c = c\n self.rect = pygame.Rect(self.x,self.y,self.w,self.h)\n def draw(self,pygame_window):\n pygame.draw.rect(pygame_window,self.c,self.rect)\nbut1 = button(x=0,y=0,w=600,h=300,c=(0,0,255))","repo_name":"dimasribnyj14/HomeWorksDimaSribnyj","sub_path":"Practice/Second Course Logika/gimn/modules/button.py","file_name":"button.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"27844302117","text":"\"\"\"Tests for reconfiguring mcp.\"\"\"\nimport os\nimport tempfile\n\nimport pytest\n\nfrom testifycompat import assert_equal\nfrom testifycompat import run\nfrom testifycompat import setup\nfrom testifycompat import suite\nfrom testifycompat import teardown\nfrom testifycompat import TestCase\nfrom tests.assertions import assert_length\nfrom tron import mcp\nfrom tron.config import config_parse\nfrom tron.config import schema\nfrom tron.serialize import filehandler\n\n\nclass TestMCPReconfigure(TestCase):\n\n os.environ[\"SSH_AUTH_SOCK\"] = \"test-socket\"\n pre_config = dict(\n ssh_options=dict(agent=True, identities=[\"tests/test_id_rsa\"],),\n nodes=[dict(name=\"node0\", hostname=\"batch0\"), dict(name=\"node1\", hostname=\"batch1\"),],\n node_pools=[dict(name=\"nodePool\", nodes=[\"node0\", \"node1\"])],\n command_context={\"thischanges\": \"froma\",},\n jobs=[\n dict(\n name=\"test_unchanged\",\n node=\"node0\",\n schedule=\"daily\",\n actions=[dict(name=\"action_unchanged\", command=\"command_unchanged\",),],\n ),\n dict(\n name=\"test_remove\",\n node=\"node1\",\n schedule={\"type\": \"cron\", \"value\": \"* * * * *\"},\n actions=[dict(name=\"action_remove\", command=\"command_remove\",),],\n cleanup_action=dict(name=\"cleanup\", command=\"doit\"),\n ),\n dict(\n name=\"test_change\",\n node=\"nodePool\",\n schedule={\"type\": \"cron\", \"value\": \"* * * * *\"},\n actions=[\n dict(name=\"action_change\", command=\"command_change\",),\n dict(name=\"action_remove2\", command=\"command_remove2\", requires=[\"action_change\"],),\n ],\n ),\n dict(\n name=\"test_daily_change\",\n node=\"node0\",\n schedule=\"daily\",\n actions=[dict(name=\"action_daily_change\", command=\"command\",),],\n ),\n dict(\n name=\"test_action_added\",\n node=\"node0\",\n schedule={\"type\": \"cron\", \"value\": \"* * * * *\"},\n actions=[dict(name=\"action_first\", command=\"command_do_it\"),],\n ),\n ],\n )\n\n post_config = dict(\n ssh_options=dict(agent=True, identities=[\"tests/test_id_rsa\"],),\n nodes=[dict(name=\"node0\", hostname=\"batch0\"), dict(name=\"node1\", hostname=\"batch1\"),],\n node_pools=[dict(name=\"nodePool\", nodes=[\"node0\", \"node1\"])],\n command_context={\"a_variable\": \"is_constant\", \"thischanges\": \"tob\",},\n jobs=[\n dict(\n name=\"test_unchanged\",\n node=\"node0\",\n schedule=\"daily\",\n actions=[dict(name=\"action_unchanged\", command=\"command_unchanged\",),],\n ),\n dict(\n name=\"test_change\",\n node=\"nodePool\",\n schedule=\"daily\",\n actions=[dict(name=\"action_change\", command=\"command_changed\",),],\n ),\n dict(\n name=\"test_daily_change\",\n node=\"node0\",\n schedule=\"daily\",\n actions=[dict(name=\"action_daily_change\", command=\"command_changed\",),],\n ),\n dict(\n name=\"test_new\",\n node=\"nodePool\",\n schedule={\"type\": \"cron\", \"value\": \"* * * * *\"},\n actions=[dict(name=\"action_new\", command=\"command_new\",),],\n ),\n dict(\n name=\"test_action_added\",\n node=\"node0\",\n schedule={\"type\": \"cron\", \"value\": \"* * * * *\"},\n actions=[\n dict(name=\"action_first\", command=\"command_do_it\"),\n dict(name=\"action_second\", command=\"command_ok\"),\n ],\n ),\n ],\n )\n\n def _get_config(self, idx, output_dir):\n config = dict(self.post_config if idx else self.pre_config)\n config[\"output_stream_dir\"] = output_dir\n return config\n\n def _get_runs_to_schedule(self, sched):\n last_run = sched.job.runs.get_newest(include_manual=False)\n last_run_time = last_run.run_time if last_run else None\n return sched.get_runs_to_schedule(last_run_time)\n\n @setup\n def setup_mcp(self):\n self.test_dir = tempfile.mkdtemp()\n self.mcp = mcp.MasterControlProgram(self.test_dir, \"config\")\n config = {schema.MASTER_NAMESPACE: self._get_config(0, self.test_dir)}\n container = config_parse.ConfigContainer.create(config)\n self.mcp.apply_config(container)\n\n @teardown\n def teardown_mcp(self):\n filehandler.OutputPath(self.test_dir).delete()\n filehandler.FileHandleManager.reset()\n\n def reconfigure(self):\n config = {schema.MASTER_NAMESPACE: self._get_config(1, self.test_dir)}\n container = config_parse.ConfigContainer.create(config)\n self.mcp.apply_config(container, reconfigure=True)\n\n @suite(\"integration\")\n def test_job_list(self):\n count = len(self.pre_config[\"jobs\"])\n assert_equal(len(self.mcp.jobs.get_names()), count)\n self.reconfigure()\n assert_equal(len(self.mcp.jobs.get_names()), count)\n\n @pytest.mark.skip(reason=\"This test doesn't currently as run1 is not scheduled.\",)\n @suite(\"integration\")\n def test_job_unchanged(self):\n assert \"MASTER.test_unchanged\" in self.mcp.jobs\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_unchanged\")\n orig_job = job_sched.job\n run0 = next(self._get_runs_to_schedule(job_sched))\n run0.start()\n run1 = next(self._get_runs_to_schedule(job_sched))\n\n assert_equal(job_sched.job.name, \"MASTER.test_unchanged\")\n action_map = job_sched.job.action_graph.action_map\n assert_equal(len(action_map), 1)\n assert_equal(action_map[\"action_unchanged\"].name, \"action_unchanged\")\n assert_equal(str(job_sched.job.scheduler), \"daily 00:00:00 \")\n\n self.reconfigure()\n assert job_sched is self.mcp.jobs.get_by_name(\"MASTER.test_unchanged\")\n assert job_sched.job is orig_job\n\n assert_equal(len(job_sched.job.runs.runs), 2)\n assert_equal(job_sched.job.runs.runs[1], run0)\n assert_equal(job_sched.job.runs.runs[0], run1)\n assert run1.is_scheduled\n assert_equal(job_sched.job.context[\"a_variable\"], \"is_constant\")\n assert_equal(job_sched.job.context[\"thischanges\"], \"tob\")\n\n @suite(\"integration\")\n def test_job_unchanged_disabled(self):\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_unchanged\")\n orig_job = job_sched.job\n next(self._get_runs_to_schedule(job_sched))\n job_sched.disable()\n\n self.reconfigure()\n assert job_sched is self.mcp.jobs.get_by_name(\"MASTER.test_unchanged\")\n assert job_sched.job is orig_job\n assert not job_sched.job.enabled\n\n @suite(\"integration\")\n def test_job_removed(self):\n assert \"MASTER.test_remove\" in self.mcp.jobs\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_remove\")\n run0 = next(self._get_runs_to_schedule(job_sched))\n run0.start()\n run1 = next(self._get_runs_to_schedule(job_sched))\n\n assert_equal(job_sched.job.name, \"MASTER.test_remove\")\n action_map = job_sched.job.action_graph.action_map\n assert_equal(len(action_map), 2)\n assert_equal(action_map[\"action_remove\"].name, \"action_remove\")\n\n self.reconfigure()\n assert \"test_remove\" not in self.mcp.jobs\n assert not job_sched.job.enabled\n assert not run1.is_scheduled\n\n @suite(\"integration\")\n def test_job_changed(self):\n assert \"MASTER.test_change\" in self.mcp.jobs\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_change\")\n run0 = next(self._get_runs_to_schedule(job_sched))\n run0.start()\n next(self._get_runs_to_schedule(job_sched))\n assert_equal(len(job_sched.job.runs.runs), 2)\n\n assert_equal(job_sched.job.name, \"MASTER.test_change\")\n action_map = job_sched.job.action_graph.action_map\n assert_equal(len(action_map), 2)\n\n self.reconfigure()\n new_job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_change\")\n assert new_job_sched is job_sched\n assert new_job_sched.job is job_sched.job\n\n assert_equal(new_job_sched.job.name, \"MASTER.test_change\")\n action_map = job_sched.job.action_graph.action_map\n assert_equal(len(action_map), 1)\n\n assert_equal(len(new_job_sched.job.runs.runs), 2)\n assert new_job_sched.job.runs.runs[1].is_starting\n assert new_job_sched.job.runs.runs[0].is_scheduled\n assert_equal(job_sched.job.context[\"a_variable\"], \"is_constant\")\n assert new_job_sched.job.context.base.job is new_job_sched.job\n\n @suite(\"integration\")\n def test_job_changed_disabled(self):\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_change\")\n job_sched.disable()\n assert not job_sched.job.enabled\n\n self.reconfigure()\n new_job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_change\")\n assert not new_job_sched.job.enabled\n\n @suite(\"integration\")\n def test_job_new(self):\n assert \"test_new\" not in self.mcp.jobs\n self.reconfigure()\n\n assert \"MASTER.test_new\" in self.mcp.jobs\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_new\")\n\n assert_equal(job_sched.job.name, \"MASTER.test_new\")\n action_map = job_sched.job.action_graph.action_map\n assert_equal(len(action_map), 1)\n assert_equal(action_map[\"action_new\"].name, \"action_new\")\n assert_equal(action_map[\"action_new\"].command, \"command_new\")\n assert_equal(len(job_sched.job.runs.runs), 1)\n assert job_sched.job.runs.runs[0].is_scheduled\n\n @suite(\"integration\")\n def test_daily_reschedule(self):\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_daily_change\")\n\n next(self._get_runs_to_schedule(job_sched))\n\n assert_equal(len(job_sched.job.runs.runs), 1)\n run = job_sched.job.runs.runs[0]\n assert run.is_scheduled\n\n action_runs = run.action_runs\n self.reconfigure()\n assert action_runs.is_cancelled\n\n assert_equal(len(job_sched.job.runs.runs), 1)\n new_run = job_sched.job.runs.runs[0]\n assert new_run is not run\n assert new_run.is_scheduled\n assert_equal(run.run_time, new_run.run_time)\n\n @suite(\"integration\")\n def test_action_added(self):\n self.reconfigure()\n job_sched = self.mcp.jobs.get_by_name(\"MASTER.test_action_added\")\n assert_length(job_sched.job.action_graph.action_map, 2)\n\n\nif __name__ == \"__main__\":\n run()\n","repo_name":"Yelp/Tron","sub_path":"tests/mcp_reconfigure_test.py","file_name":"mcp_reconfigure_test.py","file_ext":"py","file_size_in_byte":10819,"program_lang":"python","lang":"en","doc_type":"code","stars":340,"dataset":"github-code","pt":"7"} +{"seq_id":"2879428938","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport json\nimport re\nfrom typing import Optional\n\nfrom praw.models import Submission\n\nfrom bdfr.exceptions import SiteDownloaderError\nfrom bdfr.resource import Resource\nfrom bdfr.site_authenticator import SiteAuthenticator\nfrom bdfr.site_downloaders.base_downloader import BaseDownloader\n\n\nclass Imgur(BaseDownloader):\n def __init__(self, post: Submission):\n super().__init__(post)\n self.raw_data = {}\n\n def find_resources(self, authenticator: Optional[SiteAuthenticator] = None) -> list[Resource]:\n self.raw_data = self._get_data(self.post.url)\n\n out = []\n if \"is_album\" in self.raw_data:\n for image in self.raw_data[\"images\"]:\n if \"mp4\" in image:\n out.append(Resource(self.post, image[\"mp4\"], Resource.retry_download(image[\"mp4\"])))\n else:\n out.append(Resource(self.post, image[\"link\"], Resource.retry_download(image[\"link\"])))\n else:\n if \"mp4\" in self.raw_data:\n out.append(Resource(self.post, self.raw_data[\"mp4\"], Resource.retry_download(self.raw_data[\"mp4\"])))\n else:\n out.append(Resource(self.post, self.raw_data[\"link\"], Resource.retry_download(self.raw_data[\"link\"])))\n return out\n\n @staticmethod\n def _get_data(link: str) -> dict:\n try:\n if link.endswith(\"/\"):\n link = link.removesuffix(\"/\")\n if re.search(r\".*/(.*?)(gallery/|a/)\", link):\n imgur_id = re.match(r\".*/(?:gallery/|a/)(.*?)(?:/.*)?$\", link).group(1)\n link = f\"https://api.imgur.com/3/album/{imgur_id}\"\n else:\n imgur_id = re.match(r\".*/(.*?)(?:_d)?(?:\\..{0,})?$\", link).group(1)\n link = f\"https://api.imgur.com/3/image/{imgur_id}\"\n except AttributeError:\n raise SiteDownloaderError(f\"Could not extract Imgur ID from {link}\")\n\n headers = {\n \"referer\": \"https://imgur.com/\",\n \"origin\": \"https://imgur.com\",\n \"content-type\": \"application/json\",\n \"Authorization\": \"Client-ID 546c25a59c58ad7\",\n }\n res = Imgur.retrieve_url(link, headers=headers)\n\n try:\n image_dict = json.loads(res.text)\n except json.JSONDecodeError as e:\n raise SiteDownloaderError(f\"Could not parse received response as JSON: {e}\")\n\n return image_dict[\"data\"]\n","repo_name":"aliparlakci/bulk-downloader-for-reddit","sub_path":"bdfr/site_downloaders/imgur.py","file_name":"imgur.py","file_ext":"py","file_size_in_byte":2477,"program_lang":"python","lang":"en","doc_type":"code","stars":2113,"dataset":"github-code","pt":"7"} +{"seq_id":"2516714806","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport requests\nfrom bs4 import BeautifulSoup\n\nURL = 'https://au.indeed.com/jobs?q=data+science+internship&l=Sydney+NSW'\npage = requests.get(URL)\n\nsoup = BeautifulSoup(page.content, 'html.parser')\n\n\n# In[2]:\n\n\nresults = soup.find(id='resultsCol')\n\nprint(results.prettify())\n\n\n# In[3]:\n\n\njob_elems = results.find_all('div', class_='jobsearch-SerpJobCard')\nprint(job_elems)\n\n\n# In[4]:\n\n\nfor job_elem in job_elems:\n print(job_elem, end='\\n'*2)\n\n\n# In[5]:\n\n\nfor job_elem in job_elems:\n title_elem = job_elem.find('h2', class_='title')\n company_elem = job_elem.find('div', class_='sjcl')\n location_elem = job_elem.find('div', class_='summary')\n if None in (title_elem, company_elem, location_elem):\n continue\n print(title_elem.text.strip())\n print(company_elem.text.strip())\n print(location_elem.text.strip())\nprint()\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n","repo_name":"vidyasagarverma/webscrapping","sub_path":"webscrapping from indeed.py","file_name":"webscrapping from indeed.py","file_ext":"py","file_size_in_byte":923,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"18733465886","text":"\nfw=\"aaa\"\nsw=\"cba\"\ntw=\"cdb\"\ntemp1=\"\"\ntemp2=\"\"\ntemp3=\"\"\nf1w=fw.lstrip(\"a\")\ns2w=sw.lstrip(\"a\")\nt2w=tw.lstrip(\"a\")\nl1=list(f1w)\nl2= list(s2w)\nl3= list(t2w)\nfor i in l1:\n temp1+=str(ord(i)-ord(\"a\"))\n\nfor i in l2:\n temp2+=str(ord(i)-ord(\"a\"))\nfor i in l3:\n temp3+=str(ord(i)-ord(\"a\"))\nif(temp1==\"\"):\n temp1=int(0)\nif(temp2==\"\"):\n temp2=int(0)\n\nif(int(temp1)+int(temp2)==int(temp3)):\n print(\"true\")\n\n","repo_name":"sathvikchandu/my-programs-","sub_path":"leetcode/leetcode_weekly_30.05.2021.py","file_name":"leetcode_weekly_30.05.2021.py","file_ext":"py","file_size_in_byte":412,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70602460703","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy.integrate import solve_ivp\n\n# pop_vv_vec=[1,0]\n\n\n# dim=2\nd_matrix = np.zeros((2, 2))\n# print(d_matrix)\n\npvc_before_pulse = 0\npcv_before_pulse = 0\npvv_before_pulse = 1 # Ground state\npcc_before_pulse = 0 # excited state\n\nmatrix_before_pulse = [[pvc_before_pulse, pvv_before_pulse], [pcv_before_pulse, pcc_before_pulse]]\n# print(matrix_before_pulse)\n\n\npvc_after_pulse = [1, 0]\npcv_after_pulse = np.transpose(pvc_after_pulse)\n\npvv_after_pulse = 0\npcc_after_pulse = 1 # excited state\n\n# ------------------ time scale -------------------\nti = 0\ntf = 20E-12 # s\nN = 5000\ntev = np.linspace(ti, tf, N)\n# ---- constants and parameters of the system -----\nhbar = 6.582119569E-16 # eV*s\nihbar = 1j * hbar\nE_lvl_spacing = 1 # eV\ntrans_dipole_mom = 1\ndephase_E = 1E-3 # eV\npop_cc_ = 0\npop_vv_i = 1\n\n# -------- parameters of optical pulse -------------\nAmp = 1\nt0 = 10E-12 # seconds\npulse_width = 1000E-16 # seconds\n# pulse_width=(pulse_width_fs)*(10**-15)\nom = E_lvl_spacing / hbar # Hz\n\n# -------------- initial conditions ----------------\np0 = 0 + 0 * 1j\n\n\n# function computing the EM pulse\ndef e_pulse(t, t0, om, Amp, pulse_width):\n return Amp * np.exp((-(t - t0) ** 2) / ((2 * pulse_width) ** 2)) * np.exp(1j * om * t)\n\n\n# RHS of the equation (version 1)\ndef f(t, p, E_lvl_spacing, ihbar, trans_dipole_mom, dephase_E, t0, om, Amp, pulse_width):\n dp = (1 / ihbar) * ((-1j * dephase_E - E_lvl_spacing) * p) + (1 / ihbar) * trans_dipole_mom * e_pulse(t, t0, om, Amp, pulse_width)\n # print(dp)\n return dp\n\n\n# RHS of the equation (version 2)\ndef f1(t, p, E_lvl_spacing, ihbar, trans_dipole_mom, dephase_E, t0, om, Amp, pulse_width):\n dp = (1 / ihbar) * ((-1j * dephase_E - E_lvl_spacing + om * hbar) * p) +\\\n (1 / ihbar) * trans_dipole_mom * e_pulse(t, t0, om*0, Amp, pulse_width)\n # print(dp)\n return dp\n\n# --------------------- solver ---------------------\nsol = solve_ivp(f, (ti, tf), [p0], t_eval=tev, first_step=tf/N,\n args=(E_lvl_spacing, ihbar, trans_dipole_mom, dephase_E, t0, om, Amp, pulse_width))\n\nt = sol.t\np = sol.y\nnorm = np.max(np.abs(p[0]))\n\nplt.plot(t/1e-12, e_pulse(t, t0, om, Amp, pulse_width))\nplt.plot(t/1e-12, np.real(p[0])/norm)\nplt.plot(t/1e-12, np.imag(p[0])/norm)\nplt.legend(['EM pump pulse', 'Polarization, real part', 'Polarization, imaginary part'])\nplt.xlabel(\"Time (ps)\")\nplt.ylabel(\"Polarization (a.u.)\")\nplt.show()","repo_name":"freude/pySBE","sub_path":"scripts/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":2449,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"7"} +{"seq_id":"39484940808","text":"#!/usr/bin/python\n# -*- coding: utf8 -*-\n\n# Ce fichier définit des valeurs utilisées dans plusieurs scripts\n# du projet Baxter de Sara, Eder, Alain\n\nDEPLACEMENT = 0.5; # en CM, incrément de chaque déplacement\nDEPLACEMENT = DEPLACEMENT * 0.01\n\nBAXTER_USED_LIMB = 'right' # or 'left'\n","repo_name":"alainsanguinetti/projet_baxter","sub_path":"scripts/valeurs.py","file_name":"valeurs.py","file_ext":"py","file_size_in_byte":286,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7275225126","text":"import numpy as np\n\ndef calcular_pi(n):\n \"\"\"\n n: la cantidad de terminos a emplear en la sumatoria\n return: la estimacion de pi/8\n \"\"\"\n if(isinstance(n, int)):\n #se inicializan los parametros auxiliares\n #resultado\n pi_octavos = 0\n #parametro auxiliar\n i = 1\n #llamado de la funcion recursiva\n return calcular_pi_aux(n, i, pi_octavos)\n else:\n raise ValueError(\"Tipo de datos incorrecto\")\n \ndef calcular_pi_aux(n, i, pi_octavos):\n \"\"\"\n n: cantidad de terminos en la sumatoria para aproximar pi/8\n i: variable que se incrementa en cada llamado recursivo\n pi_octavos: resultado a retornar\n \"\"\"\n if(n == 0):\n #condicion de parada: se retorna la variable con el resultado acumulado\n return pi_octavos\n else: \n #termino de la sumatoria de leibniz\n pi_octavos += 1 / (i * (i + 2))\n i += 4\n n -= 1\n #se hace el llamado recursivo\n return calcular_pi_aux(n, i, pi_octavos)\n \n \ndef test_1(n): \n pi_estimado = 8 * calcular_pi(n) \n error = np.pi - pi_estimado\n print(\"Estimacion de pi: \", pi_estimado )\n print(\"Error \", error)\n \nn = 1000\ntest_1(n)","repo_name":"BAMDH/Cosas_Uni","sub_path":"Intro programación/Prácticas/Notebooks_jupyter/Recursividad_cola/calcular_pi.py","file_name":"calcular_pi.py","file_ext":"py","file_size_in_byte":1224,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"29768954169","text":"#!/usr/bin/env python3\n\nimport csv\nimport os\nimport re\nimport sys\nimport valve\n\nfrom argparse import ArgumentParser\n\n\ndef get_cl_id(cell_names, name):\n name = re.sub(r\"(DC|B|M|NK|T): \", \"\", name)\n if \" & \" in name:\n name = name.split(\" & \")[0]\n return cell_names.get(name, None)\n\n\ndef get_clean_marker(name):\n if name.endswith(\"++\"):\n return name[:-2], \"high\"\n elif name.endswith(\"+-\"):\n return name[:-2], \"low\"\n elif name.endswith(\"+\"):\n return name[:-1], \"positive\"\n elif name.endswith(\"-\"):\n return name[:-1], \"negative\"\n return name, None\n\n\ndef get_pr_id(gate_names, name):\n return gate_names.get(name, None)\n\n\ndef validate(cell_names_file, cell_levels_file, gate_names_file, cell_gate_file):\n \"\"\"\n :param cell_names_file: path to file containing Cell Ontology IDs & Labels\n :param cell_levels_file: path to file containing Cell Ontology IDs & membrane parts\n :param gate_names_file: path to file containing Protein Ontology IDs & Labels\n :param cell_gate_file: path to file to validate\n \"\"\"\n errors = []\n cell_names = {}\n with open(cell_names_file, \"r\") as f:\n reader = csv.reader(f, delimiter=\"\\t\")\n next(reader)\n for row in reader:\n cell_names[row[1]] = row[0]\n\n gate_names = {}\n with open(gate_names_file, \"r\") as f:\n reader = csv.reader(f, delimiter=\"\\t\")\n next(reader)\n for row in reader:\n gate_names[row[1]] = row[0]\n\n cell_levels = {}\n with open(cell_levels_file, \"r\") as f:\n reader = csv.reader(f, delimiter=\"\\t\")\n next(reader)\n for row in reader:\n curie = row[0]\n has_part = row[1]\n lacks_part = row[2]\n high_amount = row[3]\n low_amount = row[4]\n valid_gates = {}\n if has_part:\n for p in has_part.split(\"|\"):\n valid_gates[p] = \"positive\"\n if lacks_part:\n for p in lacks_part.split(\"|\"):\n valid_gates[p] = \"negative\"\n if high_amount:\n for p in high_amount.split(\"|\"):\n valid_gates[p] = \"high\"\n if low_amount:\n for p in low_amount.split(\"|\"):\n valid_gates[p] = \"low\"\n cell_levels[curie] = valid_gates\n\n table = os.path.splitext(os.path.basename(cell_gate_file))[0]\n with open(cell_gate_file, \"r\") as f:\n reader = csv.reader(f, delimiter=\"\\t\")\n next(reader)\n idx = 1\n for line in reader:\n idx += 1\n cell_pop_name = line[0].strip()\n cell = get_cl_id(cell_names, cell_pop_name)\n if not cell:\n errors.append(\n {\n \"table\": table,\n \"cell\": valve.idx_to_a1(idx, 1),\n \"level\": \"ERROR\",\n \"message\": f\"'{cell_pop_name}' must be a name or synonym from Cell Ontology\",\n }\n )\n continue\n valid_gates = cell_levels.get(cell, {})\n\n # Compare to provided gates\n for marker_name in [x.strip() for x in line[1].split(\",\")]:\n marker_name_clean, level = get_clean_marker(marker_name)\n marker = get_pr_id(gate_names, marker_name_clean)\n if not marker:\n errors.append(\n {\n \"table\": table,\n \"cell\": valve.idx_to_a1(idx, 2),\n \"level\": \"ERROR\",\n \"message\": f\"'{marker_name}' must be a name or synonym from Protein Ontology\",\n }\n )\n continue\n if marker not in valid_gates:\n # TODO - is this OK?\n continue\n expected_level = valid_gates[marker]\n if level == \"positive\" and expected_level in [\"high\", \"low\"]:\n errors.append(\n {\n \"table\": table,\n \"cell\": valve.idx_to_a1(idx, 2),\n \"level\": \"INFO\",\n \"message\": f\"For this cell population, {marker_name_clean} has {expected_level} expression\",\n }\n )\n elif level in [\"high\", \"low\"] and expected_level == \"positive\":\n errors.append(\n {\n \"table\": table,\n \"cell\": valve.idx_to_a1(idx, 2),\n \"level\": \"INFO\",\n \"message\": f\"For this cell population, {marker_name_clean} is positive, but not {level}\",\n }\n )\n elif level != expected_level:\n errors.append(\n {\n \"table\": table,\n \"cell\": valve.idx_to_a1(idx, 2),\n \"level\": \"ERROR\",\n \"message\": f\"For this cell population, {marker_name_clean} must be {expected_level}\",\n }\n )\n return errors\n\n\ndef main():\n parser = ArgumentParser()\n parser.add_argument(\"cell_names\", help=\"Cell Ontology IDs & labels/synonyms\")\n parser.add_argument(\"cell_levels\", help=\"Cell Ontology IDs & membrane parts\")\n parser.add_argument(\"gate_names\", help=\"Protein Ontology IDs & labels/synonyms\")\n parser.add_argument(\"input\", help=\"File to validate (cell populations & gates)\")\n args = parser.parse_args()\n errors = validate(args.cell_names, args.cell_levels, args.gate_names, args.input)\n if errors:\n writer = csv.DictWriter(\n sys.stdout,\n delimiter=\"\\t\",\n lineterminator=\"\\n\",\n fieldnames=[\"table\", \"cell\", \"level\", \"message\"],\n )\n writer.writeheader()\n writer.writerows(errors)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"jamesaoverton/cell-name-and-marker-validator","sub_path":"src/validate.py","file_name":"validate.py","file_ext":"py","file_size_in_byte":6136,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"8392495256","text":"#!/usr/bin/python3\n\nfrom datetime import datetime\nfrom models import storage\nimport uuid\n\n\nclass BaseModel:\n def __init__(self, *args, **kwargs) -> None:\n if len(kwargs) != 0:\n date_format = '%Y-%m-%dT%H:%M:%S.%f'\n for key, value in kwargs.items():\n if key == 'created_at' or key == 'updated_at':\n time = datetime.strptime(value, date_format)\n setattr(self, key, time)\n elif key != '__class__':\n setattr(self, key, value)\n else:\n self.id = str(uuid.uuid4())\n self.created_at = datetime.now()\n self.updated_at = datetime.now()\n storage.new(self)\n\n def __str__(self) -> str:\n return (f'[{self.__class__.__name__}] ({self.id}) {self.__dict__}')\n\n def save(self):\n self.updated_at = datetime.now()\n storage.save()\n\n def to_dict(self):\n upd_dict = {\n 'created_at': self.created_at.strftime('%Y-%m-%dT%H:%M:%S.%f'),\n 'updated_at': self.updated_at.strftime('%Y-%m-%dT%H:%M:%S.%f'),\n '__class__': self.__class__.__name__,\n }\n return {**self.__dict__, **upd_dict}\n","repo_name":"dkokonkwo/alu-AirBnB_clone","sub_path":"models/base_model.py","file_name":"base_model.py","file_ext":"py","file_size_in_byte":1209,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"17150177467","text":"import MapReduce\nimport sys\n\n\"\"\"\nMultiply Example in the Simple Python MapReduce Framework\n\"\"\"\n\nmr = MapReduce.MapReduce()\n\n# =============================\n# Do not modify above this line\n\n\ndef mapper(record):\n if record[0] == 'a':\n for k in range(5):\n key = (record[1], k)\n value = (record[2], record[3])\n mr.emit_intermediate(key, value)\n else:\n for i in range(5):\n key = (i, record[2])\n value = (record[1], record[3])\n mr.emit_intermediate(key, value)\n\n\ndef reducer(key, value):\n count = 0\n for i in range(5):\n prod = []\n for item in value:\n if item[0] == i:\n prod.append(item[1])\n if len(prod) == 2:\n count += prod[0] * prod[1]\n mr.emit((key[0], key[1], count))\n\n\n# Do not modify below this line\n# =============================\nif __name__ == '__main__':\n inputdata = open(sys.argv[1])\n mr.execute(inputdata, mapper, reducer)\n","repo_name":"bhujyo/Data_science","sub_path":"MapReduce/multiply.py","file_name":"multiply.py","file_ext":"py","file_size_in_byte":989,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"27781167819","text":"from googlecloudsdk.api_lib.dns import export_util\nfrom googlecloudsdk.api_lib.dns import util\nfrom googlecloudsdk.calliope import base\nfrom googlecloudsdk.calliope import exceptions\nfrom googlecloudsdk.core import log\nfrom googlecloudsdk.core import properties\nfrom googlecloudsdk.core.util import files\nfrom googlecloudsdk.third_party.apitools.base import py as apitools_base\n\n\nclass Export(base.Command):\n \"\"\"Export your record-sets into a file.\n\n This command exports the record-sets contained within the specified\n managed-zone into a file.\n \"\"\"\n\n detailed_help = {\n 'DESCRIPTION': '{description}',\n 'EXAMPLES': \"\"\"\\\n To export record-sets into a yaml file, run:\n\n $ {command} YAML_RECORDS_FILE -z MANAGED_ZONE\n\n To import record-sets into a zone file, run:\n\n $ {command} ZONE_FILE --zone-file-format -z MANAGED_ZONE\n \"\"\",\n }\n\n @staticmethod\n def Args(parser):\n parser.add_argument('records_file',\n help='File to which record-sets should be exported.')\n parser.add_argument(\n '--zone-file-format',\n required=False,\n action='store_true',\n help='Indicates that records-file should be in the zone file format.')\n\n @util.HandleHttpError\n def Run(self, args):\n dns = self.context['dns_client']\n messages = self.context['dns_messages']\n resources = self.context['dns_resources']\n project_id = properties.VALUES.core.project.Get(required=True)\n\n # Get the managed-zone.\n zone_ref = resources.Parse(args.zone, collection='dns.managedZones')\n try:\n zone = dns.managedZones.Get(zone_ref.Request())\n except apitools_base.HttpError as error:\n raise exceptions.HttpException(util.GetErrorMessage(error))\n\n # Get all the record-sets.\n record_sets = []\n for record_set in apitools_base.YieldFromList(\n dns.resourceRecordSets,\n messages.DnsResourceRecordSetsListRequest(project=project_id,\n managedZone=zone_ref.Name()),\n field='rrsets'):\n record_sets.append(record_set)\n\n # Export the record-sets.\n try:\n with files.Context(open(args.records_file, 'w')) as export_file:\n if args.zone_file_format:\n export_util.WriteToZoneFile(export_file, record_sets, zone.dnsName)\n else:\n export_util.WriteToYamlFile(export_file, record_sets)\n except Exception as exp:\n msg = 'unable to export record-sets to file [{0}]: {1}'.format(\n args.records_file, exp)\n raise exceptions.ToolException(msg)\n\n log.status.Print('Exported record-sets to [{0}].'.format(args.records_file))\n","repo_name":"twistedpair/google-cloud-sdk","sub_path":"google-cloud-sdk/lib/googlecloudsdk/surface/dns/record_sets/export.py","file_name":"export.py","file_ext":"py","file_size_in_byte":2662,"program_lang":"python","lang":"en","doc_type":"code","stars":55,"dataset":"github-code","pt":"7"} +{"seq_id":"38118999305","text":"import json\n\nimport pytest\nfrom django.core.exceptions import ObjectDoesNotExist\n\nfrom techtest.authors.models import Author\n\npytestmark = pytest.mark.django_db\n\n\ndef test_authors(client):\n # Test create / retrieve\n data = {\n \"first_name\": \"test_author_fn\",\n \"last_name\": \"test_author_ln\"\n }\n res = client.post('/authors/', data=data, content_type=\"application/json\")\n json_data = json.loads(res.content)\n author_id = json_data.get('id')\n assert res.status_code == 201\n assert json_data.get('first_name') == data['first_name']\n assert json_data.get('last_name') == data['last_name']\n assert Author.objects.get(id=json_data.get('id'))\n # Test retrieve\n res = client.get(f'/authors/{author_id}/')\n assert res.status_code == 200\n assert json_data.get('first_name') == data['first_name']\n assert json_data.get('last_name') == data['last_name']\n assert Author.objects.get(id=author_id)\n\n # Test list\n res = client.get('/authors/')\n json_data = json.loads(res.content)\n assert len(json_data) == 1\n assert json_data[0].get('first_name') == data['first_name']\n assert json_data[0].get('last_name') == data['last_name']\n\n # Test update\n updated_data = {\n \"first_name\": \"test_author_fn_update\",\n \"last_name\": \"test_author_ln_update\"\n }\n res = client.put(f'/authors/{author_id}/', data=updated_data, content_type='application/json')\n json_data = json.loads(res.content)\n assert res.status_code == 200\n assert json_data.get('first_name') == updated_data['first_name']\n assert json_data.get('last_name') == updated_data['last_name']\n\n # Test delete\n res = client.delete(f'/authors/{author_id}/')\n assert res.status_code == 200\n # Test if it was removed correctly\n with pytest.raises(ObjectDoesNotExist):\n Author.objects.get(id=author_id)\n\n # Test retrieve object that does not exist\n res = client.get('/authors/2/')\n assert res.status_code == 404\n\n\n\n\n\n\n","repo_name":"belingarb/djangotechtestmain","sub_path":"techtest/authors/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1989,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"10377596805","text":"import os, subprocess\nimport typing as T\n\nfrom . import mesonlib\nfrom .mesonlib import EnvironmentException, MachineChoice, PerMachine, split_args\nfrom . import mlog\n\n_T = T.TypeVar('_T')\n\n\n# These classes contains all the data pulled from configuration files (native\n# and cross file currently), and also assists with the reading environment\n# variables.\n#\n# At this time there isn't an ironclad difference between this an other sources\n# of state like `coredata`. But one rough guide is much what is in `coredata` is\n# the *output* of the configuration process: the final decisions after tests.\n# This, on the other hand has *inputs*. The config files are parsed, but\n# otherwise minimally transformed. When more complex fallbacks (environment\n# detection) exist, they are defined elsewhere as functions that construct\n# instances of these classes.\n\n\nknown_cpu_families = (\n 'aarch64',\n 'alpha',\n 'arc',\n 'arm',\n 'avr',\n 'c2000',\n 'dspic',\n 'e2k',\n 'ia64',\n 'm68k',\n 'microblaze',\n 'mips',\n 'mips64',\n 'parisc',\n 'pic24',\n 'ppc',\n 'ppc64',\n 'riscv32',\n 'riscv64',\n 'rl78',\n 'rx',\n 's390',\n 's390x',\n 'sh4',\n 'sparc',\n 'sparc64',\n 'wasm32',\n 'wasm64',\n 'x86',\n 'x86_64',\n)\n\n# It would feel more natural to call this \"64_BIT_CPU_FAMILES\", but\n# python identifiers cannot start with numbers\nCPU_FAMILES_64_BIT = [\n 'aarch64',\n 'alpha',\n 'ia64',\n 'mips64',\n 'ppc64',\n 'riscv64',\n 's390x',\n 'sparc64',\n 'wasm64',\n 'x86_64',\n]\n\ndef get_env_var_pair(for_machine: MachineChoice,\n is_cross: bool,\n var_name: str) -> T.Tuple[T.Optional[str], T.Optional[str]]:\n \"\"\"\n Returns the exact env var and the value.\n \"\"\"\n candidates = PerMachine(\n # The prefixed build version takes priority, but if we are native\n # compiling we fall back on the unprefixed host version. This\n # allows native builds to never need to worry about the 'BUILD_*'\n # ones.\n ([var_name + '_FOR_BUILD'] if is_cross else [var_name]),\n # Always just the unprefixed host verions\n [var_name]\n )[for_machine]\n for var in candidates:\n value = os.environ.get(var)\n if value is not None:\n break\n else:\n formatted = ', '.join(['{!r}'.format(var) for var in candidates])\n mlog.debug('None of {} are defined in the environment, not changing global flags.'.format(formatted))\n return None\n mlog.log('Using {!r} from environment with value: {!r}'.format(var, value))\n return var, value\n\ndef get_env_var(for_machine: MachineChoice,\n is_cross: bool,\n var_name: str) -> T.Tuple[T.Optional[str], T.Optional[str]]:\n ret = get_env_var_pair(for_machine, is_cross, var_name)\n if ret is None:\n return None\n else:\n var, value = ret\n return value\n\nclass Properties:\n def __init__(\n self,\n properties: T.Optional[T.Dict[str, T.Union[str, T.List[str]]]] = None,\n ):\n self.properties = properties or {} # type: T.Dict[str, T.Union[str, T.List[str]]]\n\n def has_stdlib(self, language: str) -> bool:\n return language + '_stdlib' in self.properties\n\n # Some of get_stdlib, get_root, get_sys_root are wider than is actually\n # true, but without heterogenious dict annotations it's not practical to\n # narrow them\n def get_stdlib(self, language: str) -> T.Union[str, T.List[str]]:\n return self.properties[language + '_stdlib']\n\n def get_root(self) -> T.Optional[T.Union[str, T.List[str]]]:\n return self.properties.get('root', None)\n\n def get_sys_root(self) -> T.Optional[T.Union[str, T.List[str]]]:\n return self.properties.get('sys_root', None)\n\n def get_pkg_config_libdir(self) -> T.Optional[T.List[str]]:\n p = self.properties.get('pkg_config_libdir', None)\n if p is None:\n return p\n return mesonlib.listify(p)\n\n def __eq__(self, other: T.Any) -> 'T.Union[bool, NotImplemented]':\n if isinstance(other, type(self)):\n return self.properties == other.properties\n return NotImplemented\n\n # TODO consider removing so Properties is less freeform\n def __getitem__(self, key: str) -> T.Any:\n return self.properties[key]\n\n # TODO consider removing so Properties is less freeform\n def __contains__(self, item: T.Any) -> bool:\n return item in self.properties\n\n # TODO consider removing, for same reasons as above\n def get(self, key: str, default: T.Any = None) -> T.Any:\n return self.properties.get(key, default)\n\nclass MachineInfo:\n def __init__(self, system: str, cpu_family: str, cpu: str, endian: str):\n self.system = system\n self.cpu_family = cpu_family\n self.cpu = cpu\n self.endian = endian\n self.is_64_bit = cpu_family in CPU_FAMILES_64_BIT # type: bool\n\n def __eq__(self, other: T.Any) -> 'T.Union[bool, NotImplemented]':\n if self.__class__ is not other.__class__:\n return NotImplemented\n return \\\n self.system == other.system and \\\n self.cpu_family == other.cpu_family and \\\n self.cpu == other.cpu and \\\n self.endian == other.endian\n\n def __ne__(self, other: T.Any) -> 'T.Union[bool, NotImplemented]':\n if self.__class__ is not other.__class__:\n return NotImplemented\n return not self.__eq__(other)\n\n def __repr__(self) -> str:\n return ''.format(self.system, self.cpu_family, self.cpu)\n\n @classmethod\n def from_literal(cls, literal: T.Dict[str, str]) -> 'MachineInfo':\n minimum_literal = {'cpu', 'cpu_family', 'endian', 'system'}\n if set(literal) < minimum_literal:\n raise EnvironmentException(\n 'Machine info is currently {}\\n'.format(literal) +\n 'but is missing {}.'.format(minimum_literal - set(literal)))\n\n cpu_family = literal['cpu_family']\n if cpu_family not in known_cpu_families:\n mlog.warning('Unknown CPU family {}, please report this at https://github.com/mesonbuild/meson/issues/new'.format(cpu_family))\n\n endian = literal['endian']\n if endian not in ('little', 'big'):\n mlog.warning('Unknown endian {}'.format(endian))\n\n return cls(literal['system'], cpu_family, literal['cpu'], endian)\n\n def is_windows(self) -> bool:\n \"\"\"\n Machine is windows?\n \"\"\"\n return self.system == 'windows' or 'mingw' in self.system\n\n def is_cygwin(self) -> bool:\n \"\"\"\n Machine is cygwin?\n \"\"\"\n return self.system.startswith('cygwin')\n\n def is_linux(self) -> bool:\n \"\"\"\n Machine is linux?\n \"\"\"\n return self.system == 'linux'\n\n def is_darwin(self) -> bool:\n \"\"\"\n Machine is Darwin (iOS/tvOS/OS X)?\n \"\"\"\n return self.system in {'darwin', 'ios', 'tvos'}\n\n def is_android(self) -> bool:\n \"\"\"\n Machine is Android?\n \"\"\"\n return self.system == 'android'\n\n def is_haiku(self) -> bool:\n \"\"\"\n Machine is Haiku?\n \"\"\"\n return self.system == 'haiku'\n\n def is_netbsd(self) -> bool:\n \"\"\"\n Machine is NetBSD?\n \"\"\"\n return self.system == 'netbsd'\n\n def is_openbsd(self) -> bool:\n \"\"\"\n Machine is OpenBSD?\n \"\"\"\n return self.system == 'openbsd'\n\n def is_dragonflybsd(self) -> bool:\n \"\"\"Machine is DragonflyBSD?\"\"\"\n return self.system == 'dragonfly'\n\n def is_freebsd(self) -> bool:\n \"\"\"Machine is FreeBSD?\"\"\"\n return self.system == 'freebsd'\n\n def is_sunos(self) -> bool:\n \"\"\"Machine is illumos or Solaris?\"\"\"\n return self.system == 'sunos'\n\n def is_hurd(self) -> bool:\n \"\"\"\n Machine is GNU/Hurd?\n \"\"\"\n return self.system == 'gnu'\n\n def is_irix(self) -> bool:\n \"\"\"Machine is IRIX?\"\"\"\n return self.system.startswith('irix')\n\n # Various prefixes and suffixes for import libraries, shared libraries,\n # static libraries, and executables.\n # Versioning is added to these names in the backends as-needed.\n def get_exe_suffix(self) -> str:\n if self.is_windows() or self.is_cygwin():\n return 'exe'\n else:\n return ''\n\n def get_object_suffix(self) -> str:\n if self.is_windows():\n return 'obj'\n else:\n return 'o'\n\n def libdir_layout_is_win(self) -> bool:\n return self.is_windows() or self.is_cygwin()\n\nclass BinaryTable:\n def __init__(\n self,\n binaries: T.Optional[T.Dict[str, T.Union[str, T.List[str]]]] = None,\n ):\n self.binaries = binaries or {} # type: T.Dict[str, T.Union[str, T.List[str]]]\n for name, command in self.binaries.items():\n if not isinstance(command, (list, str)):\n # TODO generalize message\n raise mesonlib.MesonException(\n 'Invalid type {!r} for binary {!r} in cross file'\n ''.format(command, name))\n\n # Map from language identifiers to environment variables.\n evarMap = {\n # Compilers\n 'c': 'CC',\n 'cpp': 'CXX',\n 'cs': 'CSC',\n 'd': 'DC',\n 'fortran': 'FC',\n 'objc': 'OBJC',\n 'objcpp': 'OBJCXX',\n 'rust': 'RUSTC',\n 'vala': 'VALAC',\n\n # Linkers\n 'c_ld': 'CC_LD',\n 'cpp_ld': 'CXX_LD',\n 'd_ld': 'DC_LD',\n 'fortran_ld': 'FC_LD',\n 'objc_ld': 'OBJC_LD',\n 'objcpp_ld': 'OBJCXX_LD',\n 'rust_ld': 'RUSTC_LD',\n\n # Binutils\n 'strip': 'STRIP',\n 'ar': 'AR',\n 'windres': 'WINDRES',\n\n # Other tools\n 'cmake': 'CMAKE',\n 'qmake': 'QMAKE',\n 'pkgconfig': 'PKG_CONFIG',\n } # type: T.Dict[str, str]\n\n # Deprecated environment variables mapped from the new variable to the old one\n # Deprecated in 0.54.0\n DEPRECATION_MAP = {\n 'DC_LD': 'D_LD',\n 'FC_LD': 'F_LD',\n 'RUSTC_LD': 'RUST_LD',\n 'OBJCXX_LD': 'OBJCPP_LD',\n } # type: T.Dict[str, str]\n\n @staticmethod\n def detect_ccache() -> T.List[str]:\n try:\n subprocess.check_call(['ccache', '--version'], stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n except (OSError, subprocess.CalledProcessError):\n return []\n return ['ccache']\n\n @classmethod\n def parse_entry(cls, entry: T.Union[str, T.List[str]]) -> T.Tuple[T.List[str], T.List[str]]:\n compiler = mesonlib.stringlistify(entry)\n # Ensure ccache exists and remove it if it doesn't\n if compiler[0] == 'ccache':\n compiler = compiler[1:]\n ccache = cls.detect_ccache()\n else:\n ccache = []\n # Return value has to be a list of compiler 'choices'\n return compiler, ccache\n\n def lookup_entry(self,\n for_machine: MachineChoice,\n is_cross: bool,\n name: str) -> T.Optional[T.List[str]]:\n \"\"\"Lookup binary in cross/native file and fallback to environment.\n\n Returns command with args as list if found, Returns `None` if nothing is\n found.\n \"\"\"\n # Try explicit map, don't fall back on env var\n # Try explict map, then env vars\n for _ in [()]: # a trick to get `break`\n raw_command = self.binaries.get(name)\n if raw_command is not None:\n command = mesonlib.stringlistify(raw_command)\n break # found\n evar = self.evarMap.get(name)\n if evar is not None:\n raw_command = get_env_var(for_machine, is_cross, evar)\n if raw_command is None:\n deprecated = self.DEPRECATION_MAP.get(evar)\n if deprecated is not None:\n raw_command = get_env_var(for_machine, is_cross, deprecated)\n if raw_command is not None:\n mlog.deprecation(\n 'The', deprecated, 'environment variable is deprecated in favor of',\n evar, once=True)\n if raw_command is not None:\n command = split_args(raw_command)\n break # found\n command = None\n\n\n # Do not return empty or blank string entries\n if command is not None and (len(command) == 0 or len(command[0].strip()) == 0):\n command = None\n return command\n\nclass Directories:\n\n \"\"\"Data class that holds information about directories for native and cross\n builds.\n \"\"\"\n\n def __init__(self, bindir: T.Optional[str] = None, datadir: T.Optional[str] = None,\n includedir: T.Optional[str] = None, infodir: T.Optional[str] = None,\n libdir: T.Optional[str] = None, libexecdir: T.Optional[str] = None,\n localedir: T.Optional[str] = None, localstatedir: T.Optional[str] = None,\n mandir: T.Optional[str] = None, prefix: T.Optional[str] = None,\n sbindir: T.Optional[str] = None, sharedstatedir: T.Optional[str] = None,\n sysconfdir: T.Optional[str] = None):\n self.bindir = bindir\n self.datadir = datadir\n self.includedir = includedir\n self.infodir = infodir\n self.libdir = libdir\n self.libexecdir = libexecdir\n self.localedir = localedir\n self.localstatedir = localstatedir\n self.mandir = mandir\n self.prefix = prefix\n self.sbindir = sbindir\n self.sharedstatedir = sharedstatedir\n self.sysconfdir = sysconfdir\n\n def __contains__(self, key: str) -> bool:\n return hasattr(self, key)\n\n def __getitem__(self, key: str) -> T.Optional[str]:\n # Mypy can't figure out what to do with getattr here, so we'll case for it\n return T.cast(T.Optional[str], getattr(self, key))\n\n def __setitem__(self, key: str, value: T.Optional[str]) -> None:\n setattr(self, key, value)\n\n def __iter__(self) -> T.Iterator[T.Tuple[str, str]]:\n return iter(self.__dict__.items())\n","repo_name":"SoftwareGuy/stealthChamp-qemu","sub_path":"meson/mesonbuild/envconfig.py","file_name":"envconfig.py","file_ext":"py","file_size_in_byte":14287,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"7"} +{"seq_id":"13155075701","text":"\"\"\"最终版2018/11/27(修改了只能爬最新的 不能爬'昨天'的评论内容 测试表示微博限制单个用户访问频次较严重 本次添加了五个Cookie)\"\"\"\r\nimport requests\r\nimport json\r\nimport re\r\nimport ceshi.cookies as ck\r\nimport random\r\nimport aiohttp\r\nimport asyncio\r\nclass Wb_Comment:\r\n def __init__(self,url):\r\n self.headers = {\r\n 'User-Agent':'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.67 Mobile Safari/537.36',\r\n 'Cookie':random.choice(ck.cookies)\r\n }\r\n self.url = url\r\n self.comment_list = []\r\n p1 = re.compile(r'.*/(.*)', re.S)\r\n cuturl = re.findall(p1, self.url)\r\n self.mid = cuturl[0]\r\n def get_frist(self):\r\n url = 'https://m.weibo.cn/comments/hotflow?id={0}&mid={0}&max_id_type=0'.format(self.mid)\r\n response = requests.get(url,headers=self.headers).text\r\n html = json.loads(response)\r\n items = html['data']['data']\r\n for item in items:\r\n ID = item['id']\r\n Name = item['user']['screen_name']\r\n Content = item['text']\r\n Agree = item['like_count']\r\n Time = item['created_at']\r\n self.comment_list.append({'ID': ID, 'Name': Name, 'Content': Content,'Agree':Agree, 'Time': Time})\r\n next_id = html['data']['max_id']\r\n return next_id\r\n async def get_next(self):\r\n maxid = self.get_frist()\r\n while 1:\r\n url = 'https://m.weibo.cn/comments/hotflow?id={0}&mid={0}&max_id={1}&max_id_type=0'.format(self.mid,maxid)\r\n async with aiohttp.ClientSession() as session:\r\n async with session.get(url,headers=self.headers) as pagehtml:\r\n response = await pagehtml.text(\"utf-8\", \"ignore\")\r\n html = json.loads(response)\r\n items = html['data']['data']\r\n for item in items:\r\n ID = item['id']\r\n Name = item['user']['screen_name']\r\n Content = item['text']\r\n Agree = item['like_count']\r\n Time = item['created_at']\r\n timeArray = datetime.datetime.strptime(Time, '%a %b %d %H:%M:%S +0800 %Y')\r\n otherStyleTime = timeArray.strftime(\r\n '%Y/%m/%d %H:%M:%S') # 评论时间 Tue Nov 20 12:39:24 +0800 2018 转为 2018/11/20 12:39:24 格式\r\n self.comment_list.append({'ID': ID, 'Name': Name, 'Content': Content, 'Agree': Agree, 'Time': otherStyleTime})\r\n maxid = html['data']['max_id']\r\n if maxid == 0:\r\n break\r\n def main(self):\r\n asyncio.set_event_loop(asyncio.new_event_loop())\r\n loop = asyncio.get_event_loop()\r\n loop.run_until_complete(self.get_next())\r\n loop.close()\r\n # wb_commment_dict = {'最新评论': len(self.comment_list)}\r\n wb_commment_dict = {'最新评论': self.comment_list,'type': 'weibo'}\r\n # print(wb_commment_dict)\r\n return wb_commment_dict\r\n\r\n\r\n","repo_name":"mango941231/Python-webspider","sub_path":"Crawl_WB_Commentasyncio.py","file_name":"Crawl_WB_Commentasyncio.py","file_ext":"py","file_size_in_byte":3195,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"16377991600","text":"class Node:\n def __init__(self, value, left, right):\n self.value = value\n self.left = left\n self.right = right\n\n\nclass BinaryTree:\n def __init__(self, array):\n node_list = [Node(value, None, None) for value in array]\n for ind, node in enumerate(node_list):\n left = 2 * ind + 1\n right = 2 * ind + 2\n if left < len(node_list):\n node.left = node_list[left]\n if right < len(node_list):\n node.right = node_list[right]\n\n self.root = node_list[0]\n\n def preorder(self):\n s = ''\n def recursive(node):\n nonlocal s\n s += str(node.value) + ' '\n if node.left:\n recursive(node.left)\n if node.right:\n recursive(node.right)\n\n s = '['\n recursive(self.root)\n s += ']'\n print(s)\n \n def inorder(self):\n s = ''\n def recursive(node):\n nonlocal s\n if node.left:\n recursive(node.left)\n s += str(node.value) + ' '\n if node.right:\n recursive(node.right)\n\n s = '['\n recursive(self.root)\n s += ']'\n print(s)\n \n def postorder(self):\n s = ''\n def recursive(node):\n nonlocal s\n if node.left:\n recursive(node.left)\n if node.right:\n recursive(node.right)\n s += str(node.value) + ' '\n\n s = '['\n recursive(self.root)\n s += ']'\n print(s)\n\n def bfs(self, value):\n queue = []\n isFound = False\n queue.append(self.root)\n\n while queue:\n node = queue[0]\n del queue[0]\n if node.value == value:\n isFound = True\n return isFound\n if node.left:\n queue.append(node.left)\n if node.right:\n queue.append(node.right)\n return isFound\n \n def dfs(self, value):\n isFound = False\n def recursive(node):\n nonlocal isFound\n if node.value == value:\n isFound = True\n return\n if isFound is True:\n return\n if node.left is not None:\n recursive(node.left)\n if node.right is not None:\n recursive(node.right)\n \n recursive(self.root)\n return isFound\n\n\ntree = BinaryTree([i for i in range(13)])\ntree.preorder()\ntree.inorder()\ntree.postorder()\n\nprint(tree.dfs(4))\nprint(tree.dfs(11))\n\nprint(tree.bfs(6))\nprint(tree.bfs(17))","repo_name":"hodduck-v/algorithm_nklcb1","sub_path":"0326/binary_tree_node.py","file_name":"binary_tree_node.py","file_ext":"py","file_size_in_byte":2661,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"12460095273","text":"from PIL import Image\nfrom cStringIO import StringIO\nimport os\n\nos.chdir(os.path.dirname(__file__))\n\ndef index(req,image,**opts):\n\tdir = './'\n\n\ts = StringIO()\n\ttypes = ['gif','jpg','png']\n\n\t# Open image\n\timg = Image.open(dir+image);\n\n\t# Default size \n\tsize = list(img.size)\n\n\tif opts.has_key('width'):\n\t\tsize[0] = int(opts['width'])\n\tif opts.has_key('height'):\n\t\tsize[1] = int(opts['height'])\n\n\tsize = tuple(size)\n\t# Scale image \n\tif size != img.size:\n\t\timg.thumbnail(size,Image.ANTIALIAS)\n\n\tformat = img.format\n\tif opts.has_key('format') and opts['format'] in types:\n\t\tformat = opts['format']\n\n\tif format == 'jpg':\n\t\tformat = 'jpeg'\n\n\timg.save(s,format)\n\n\treq.content_type = 'image/'+format\n\timg = s.getvalue()\n\ts.close()\n\treturn img\n","repo_name":"Joshua2504/regnum-maps-old","sub_path":"beta/images/show.py","file_name":"show.py","file_ext":"py","file_size_in_byte":735,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"40201893059","text":"# -*- coding: utf-8 -*-\n\nimport collections\n\nfrom copy import deepcopy\n\n\ndef _get_deep_attr(obj, attr):\n for path in attr.split(\".\"):\n try:\n obj = getattr(obj, path)\n except AttributeError: \n return None\n return obj\n\ndef json_transform(obj, attrs_tree, extend=None):\n result = {}\n for attr in attrs_tree:\n _attr = attr.split(\"-\")[0]\n val = attrs_tree[attr]\n \n if isinstance(val, dict): # val is a map\n objarray = getattr(obj, _attr)\n val = deepcopy(val)\n sort_key = val.pop(\"$_sort_key_$\") if \"$_sort_key_$\" in val else None\n filter_funct = val.pop(\"$filter$\") if \"$filter$\" in val else (lambda e: True)\n result[attr] = list()\n if sort_key: \n objarray = sorted(objarray, key=sort_key) \n for item in objarray:\n if filter_funct(item): result[attr].append(json_transform(item, attrs_tree=val))\n \n elif isinstance(val, (list, tuple)): # val is list\n columns = val\n objarray = getattr(obj, _attr)\n \n result[attr] = {}\n result[attr][\"$RECORDS$\"] = []\n \n functs, formats, list_attrs, _columns = [], [], [], []\n for col in columns: \n list_attrs.append(col[\"$attr\"])\n formats.append(col[\"$format\"] if \"$format\" in col else None)\n functs.append(col[\"$funct\"] if \"$funct\" in col else None)\n _columns.append(dict([ (k, col[k]) for k in col if k != '$funct']))\n \n for item in objarray:\n row = []\n for idx, col in enumerate(columns):\n l_attr = list_attrs[idx]\n funct = functs[idx]\n fmat = formats[idx]\n raw_value = _get_deep_attr(item, l_attr)\n value = funct(raw_value) if (funct is not None) else (fmat % (raw_value,) if raw_value is not None else \"\") if (fmat is not None) else raw_value\n row.append(value)\n result[attr][\"$RECORDS$\"].append(row)\n \n result[attr][\"$COLUMNS$\"] = _columns\n elif callable(val):\n funct = val\n field_value = funct( _get_deep_attr(obj, _attr) )\n result[attr] = field_value \n else:\n fmat = attrs_tree[attr]\n field_value = _get_deep_attr(obj, _attr)\n result[attr] = field_value if fmat is None else (fmat %(field_value, ) if (field_value is not None) else None)\n if extend:\n result.update(extend)\n return result\n","repo_name":"rotoudjimaye/web-analytics","sub_path":"src/main/webapp/WEB-INF/jython-app/webanalytics/json_transforms.py","file_name":"json_transforms.py","file_ext":"py","file_size_in_byte":2720,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"7582189099","text":"\"\"\"\nImport json data from JSON file to Datababse\n\"\"\"\nimport os\nimport json\nfrom gogoedu.models import Catagory,Lesson,Word\nfrom django.core.management import BaseCommand\nfrom elearning.settings import BASE_DIR\n\n\nclass Command(BaseCommand):\n def import_vocab_from_file(self):\n data_folder = os.path.join(BASE_DIR, 'gogoedu', 'static/json_file/vocab')\n for data_file in os.listdir(data_folder):\n with open(os.path.join(data_folder, data_file), encoding='utf-8') as data_file:\n data = json.loads(data_file.read())\n catagory,createdCatagory=Catagory.objects.get_or_create(name=\"Kanji Genki\")\n for data_object in data:\n kana = data_object.get('Kana', None)\n kanji = data_object.get('Kanji', None)\n lesson = data_object.get('Lesson', None)\n mean = data_object.get('Meaning', None)\n try:\n lesson, created_lesson = Lesson.objects.get_or_create(\n name = lesson,\n catagory = catagory,\n )\n if created_lesson:\n lesson.save()\n display_format = \"\\nLesson, {}, has been saved.\"\n print(display_format.format(lesson))\n word, created_word = Word.objects.get_or_create(\n word=kana,\n mean=mean,\n kanji=kanji,\n catagory=catagory,\n )\n word.lesson.add(lesson)\n if created_word:\n word.save()\n display_format = \"\\nWord, {}, has been saved.\"\n print(display_format.format(word))\n except Exception as ex:\n print(str(ex))\n msg = \"\\n\\nSomething went wrong saving this Word: {}\\n{}\".format(kana, str(ex))\n print(msg)\n \n\n\n def handle(self, *args, **options):\n \"\"\"\n Call the function to import data\n \"\"\"\n self.import_vocab_from_file()","repo_name":"tuandang98/gogoedu","sub_path":"gogoedu/management/commands/import_vocab_from_json_file.py","file_name":"import_vocab_from_json_file.py","file_ext":"py","file_size_in_byte":2292,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"11850646185","text":"import decimal\nfrom datetime import datetime\n\nfrom django.db.models import Sum, F, Value, ExpressionWrapper, Q, Subquery, OuterRef, FloatField\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\nfrom django.views.generic import UpdateView, ListView, View\nfrom directory.models import PaymentAccount, CurrenciesRates, Counterparties, InitialDebts, Currencies\nfrom payments.models import PaymentDocuments\n\nfrom registers.forms import AccountSettingsSet, AccountBalancesFilter, DashboardFilter\nfrom registers.models import AccountSettings\nfrom registers.views_reports import AccountBalancesView\n\n\ndef htmx_projects(request):\n form = DashboardFilter(request.GET)\n return HttpResponse(form[\"project\"])\n\n\nclass DashboardView(View):\n def get(self, request):\n main_currency = AccountSettings.load().currency()\n form = DashboardFilter(request.GET)\n accounts = PaymentAccount.objects.all()\n\n paydocs = PaymentDocuments.objects.all().annotate(amount_convert=F('inflow_amount')+F('outflow_amount'))\n #paydocs = []\n rates = {main_currency.id: decimal.Decimal(1)}\n for doc in paydocs:\n if doc.currency_id not in rates:\n rates[doc.currency_id] = (\n CurrenciesRates.objects.filter(\n accounting_currency=main_currency,\n currency=F('currency__id'),\n date__lte=datetime.now(),\n ).order_by('-date')[:1].first().rate\n )\n # doc.amount_convert=(doc.inflow_amount + doc.outflow_amount) / rates[doc.currency_id]\n doc.amount_convert /= rates[doc.currency_id]\n\n print(doc.date, doc.inflow_amount, doc.outflow_amount, doc.currency, doc.amount_convert)\n\n paydocs_before = paydocs\n\n if form.is_valid():\n if form.cleaned_data['organization']:\n accounts = accounts.filter(organization=form.cleaned_data['organization'])\n paydocs = paydocs.filter(organization=form.cleaned_data['organization'])\n paydocs_before = paydocs_before.filter(organization=form.cleaned_data['organization'])\n if form.cleaned_data['date_start']:\n paydocs = paydocs.filter(date__gte=form.cleaned_data['date_start'])\n if form.cleaned_data['date_end']:\n paydocs = paydocs.filter(date__lte=form.cleaned_data['date_end'])\n paydocs_before = paydocs_before.filter(date__lte=form.cleaned_data['date_start'])\n\n cf_table, cf_bar = self.get_cf_table(paydocs)\n\n context = {\n 'form': form,\n 'today': datetime.today(),\n 'main_currency': main_currency,\n 'account_balances': self.get_balances(accounts, paydocs_before),\n 'cf_table': cf_table,\n 'cf_bar': cf_bar,\n 'cf_dynamics': self.get_cf_dynamics(paydocs),\n }\n\n return render(request, 'registers/dashboard.html', context=context)\n\n # chart 1: account balances\n @staticmethod\n def get_balances(accounts, paydocs_before):\n account_balances = {}\n accounts = accounts.values_list('account', flat=True).distinct()\n\n for acc in accounts:\n currency = PaymentAccount.objects.filter(account=acc).values_list('currency__code', flat=True)[0]\n\n receipts = paydocs_before.filter(account__account=acc)\n receipts_sum = receipts.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_sum is None:\n receipts_sum = 0\n\n payments = paydocs_before.filter(account__account=acc)\n payments_sum = payments.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_sum is None:\n payments_sum = 0\n\n final_balance = receipts_sum - payments_sum\n account_balances[acc] = [int(final_balance), currency]\n\n account_balances = [[k, *v] for k, v in account_balances.items()]\n\n return account_balances\n\n @staticmethod\n def get_amount_sum(paydocs):\n amount_sum = paydocs.aggregate(Sum(\"amount_convert\")).get('amount_convert__sum', 0.00)\n if amount_sum is None:\n amount_sum = 0\n return amount_sum\n\n # chart 2, 3 cf table and cf bar\n def get_cf_table(self, paydocs):\n cf_table = {}\n\n receipts_total = paydocs.filter(flow='Receipts')\n payments_total = paydocs.filter(flow='Payments')\n receipts_oper = paydocs.filter(item__activity='operating', flow='Receipts')\n payments_oper = paydocs.filter(item__activity='operating', flow='Payments')\n receipts_invest = paydocs.filter(item__activity='investing', flow='Receipts')\n payments_invest = paydocs.filter(item__activity='investing', flow='Payments')\n receipts_fin = paydocs.filter(item__activity='financing', flow='Receipts')\n payments_fin = paydocs.filter(item__activity='financing', flow='Payments')\n\n receipts_sum = self.get_amount_sum(receipts_total)\n payments_sum = self.get_amount_sum(payments_total)\n receipts_oper_sum = self.get_amount_sum(receipts_oper)\n payments_oper_sum = self.get_amount_sum(payments_oper)\n receipts_invest_sum = self.get_amount_sum(receipts_invest)\n payments_invest_sum = self.get_amount_sum(payments_invest)\n receipts_fin_sum = self.get_amount_sum(receipts_fin)\n payments_fin_sum = self.get_amount_sum(payments_fin)\n\n cf = receipts_sum - payments_sum\n cf_oper = receipts_oper_sum - payments_oper_sum\n cf_invest = receipts_invest_sum - payments_invest_sum\n cf_fin = receipts_fin_sum - payments_fin_sum\n\n cf_table['receipts'] = [int(receipts_oper_sum), int(receipts_invest_sum), int(receipts_fin_sum),\n int(receipts_sum)]\n cf_table['payments'] = [int(payments_oper_sum), int(payments_invest_sum), int(payments_fin_sum),\n int(payments_sum)]\n cf_table['cash flow'] = [int(cf_oper), int(cf_invest), int(cf_fin), int(cf)]\n\n cf_table = [[k, *v] for k, v in cf_table.items()]\n\n cf_total = cf_table[2][1] + cf_table[2][2] + cf_table[2][3]\n\n cf_bar = [['Operating', cf_table[2][1]], ['Investment', cf_table[2][2]],\n ['Financing', cf_table[2][3]], ['Total', cf_total]]\n\n return cf_table, cf_bar\n\n # function for getting full data\n @staticmethod\n def get_dynamics(paydocs):\n dynamics = {}\n for paydoc in paydocs:\n month = str(paydoc.date.month).rjust(2, '0')\n period = f'{str(paydoc.date.year)}/{month}'\n amount = float(paydoc.amount_convert)\n if period not in dynamics:\n dynamics[period] = amount\n else:\n dynamics[period] += amount\n\n return dynamics\n\n # get dynamics of receipts and payments\n def get_cf_dynamics(self, paydocs):\n payments = paydocs.filter(flow='Payments')\n receipts = paydocs.filter(flow='Receipts')\n receipts_dynamics = self.get_dynamics(receipts)\n for k, v in receipts_dynamics.items():\n receipts_dynamics[k] = int(v)\n payments_dynamics = self.get_dynamics(payments)\n for k, v in payments_dynamics.items():\n payments_dynamics[k] = int(v * (-1))\n\n total_cf = {k: [receipts_dynamics.get(k, 0), payments_dynamics.get(k, 0)]\n for k in set(receipts_dynamics) | set(payments_dynamics)}\n total_cf = sorted(total_cf.items(), key=lambda x: x[0])\n total_cf = dict(total_cf)\n\n cf_dynamics = []\n for k, v in total_cf.items():\n cf_dynamics.append([k, *v, v[0] + v[1]])\n\n return cf_dynamics\n\n\nclass ChartsOperView(View):\n def get(self, request):\n main_currency = AccountSettings.load().currency()\n form = DashboardFilter(request.GET)\n receipts = PaymentDocuments.objects.filter(flow='Receipts', item__activity='operating')\n payments = PaymentDocuments.objects.filter(flow='Payments', item__activity='operating')\n\n if form.is_valid():\n if form.cleaned_data['organization']:\n receipts = receipts.filter(organization=form.cleaned_data['organization'])\n payments = payments.filter(organization=form.cleaned_data['organization'])\n if form.cleaned_data['project']:\n receipts = receipts.filter(project=form.cleaned_data['project'])\n payments = payments.filter(project=form.cleaned_data['project'])\n if form.cleaned_data['date_start']:\n receipts = receipts.filter(date__gte=form.cleaned_data['date_start'])\n payments = payments.filter(date__gte=form.cleaned_data['date_start'])\n if form.cleaned_data['date_end']:\n receipts = receipts.filter(date__lte=form.cleaned_data['date_end'])\n payments = payments.filter(date__lte=form.cleaned_data['date_end'])\n\n rp_dynamics = self.get_rp_dynamics(receipts, payments)\n\n context = {\n 'form': form,\n 'today': datetime.today(),\n 'main_currency': main_currency,\n 'payments_bar': self.get_bar_payments(payments),\n 'receipts_structure': self.get_structure(receipts),\n 'payments_structure': self.get_structure(payments),\n 'rp_dynamics': rp_dynamics,\n 'top_customers': self.get_bar_top10(receipts, 'Receipts'),\n 'top_suppliers': self.get_bar_top10(payments, 'Payments'),\n }\n\n return render(request, 'registers/charts_oper.html', context=context)\n\n # charts 1, 2, 4\n def get_structure(self, doc):\n main_currency = AccountSettings.load().currency()\n data = {}\n for i in doc:\n rate = float(AccountBalancesView.get_rate(i.currency, main_currency))\n item = str(i.item)\n amount = float(i.inflow_amount) / rate if i.inflow_amount != 0 else float(i.outflow_amount) / rate\n\n if item not in data:\n data[item] = int(amount)\n else:\n data[item] += int(amount)\n\n data_sort = sorted(data.items(), key=lambda x: x[1], reverse=True)\n data_sort = dict(data_sort)\n structure = [[k, v] for k, v in data_sort.items()]\n\n return structure\n\n # function for data payments and payments\n @staticmethod\n def get_dynamics(doc):\n main_currency = AccountSettings.load().currency()\n dynamics = {}\n for i in doc:\n rate = float(AccountBalancesView.get_rate(i.currency, main_currency))\n month = str(i.date.month).rjust(2, '0')\n period = f'{str(i.date.year)}/{month}'\n amount = float(i.inflow_amount) / rate if i.inflow_amount != 0 else float(i.outflow_amount) / rate\n if period not in dynamics:\n dynamics[period] = int(amount)\n else:\n dynamics[period] += int(amount)\n\n return dynamics\n\n # get data payments and payments for chart 3\n def get_total_cf(self, receipts, payments):\n payments_dynamics = self.get_dynamics(payments)\n receipts_dynamics = self.get_dynamics(receipts)\n\n total_cf = {k: [receipts_dynamics.get(k, 0), payments_dynamics.get(k, 0)]\n for k in set(receipts_dynamics) | set(payments_dynamics)}\n total_cf = sorted(total_cf.items(), key=lambda x: x[0])\n total_cf = dict(total_cf)\n\n return total_cf\n\n # chart 3 Dynamics of payments and payments\n def get_rp_dynamics(self, receipts, payments):\n total_cf = self.get_total_cf(receipts, payments)\n rp_dynamics = []\n for k, v in total_cf.items():\n rp_dynamics.append([k, *v])\n\n return rp_dynamics\n\n # chart 5 Payments bar by group\n @staticmethod\n def get_bar_payments(paydocs):\n main_currency = AccountSettings.load().currency()\n data = {}\n for i in paydocs:\n rate = float(AccountBalancesView.get_rate(i.currency, main_currency))\n items_group = str(i.item.group)\n amount = float(i.inflow_amount) / rate if i.inflow_amount != 0 else float(i.outflow_amount) / rate\n if items_group not in data:\n data[items_group] = amount\n else:\n data[items_group] += amount\n data_sort = sorted(data.items(), key=lambda x: x[1], reverse=True)\n structure = dict(data_sort)\n\n return [[k, v] for k, v in structure.items()]\n\n # chart 6, 7 TOP-10 counterparty\n @staticmethod\n def get_bar_top10(paydocs, flow):\n main_currency = AccountSettings.load().currency()\n if flow == 'Receipts':\n counterparties = Counterparties.objects.filter(customer=True).values_list('id', flat=True)\n else:\n counterparties = Counterparties.objects.filter(suppliers=True).values_list('id', flat=True)\n data = {}\n paydocs_sum = (\n paydocs\n .filter(counterparty__in=counterparties)\n .annotate(amount_sum=Sum(\"inflow_amount\") + Sum(\"outflow_amount\"))\n .order_by('counterparty', 'currency', 'amount_sum')\n .values('counterparty__counterparty', 'currency', 'amount_sum')\n )\n counterparty = None\n for doc_sum in paydocs_sum:\n cp = doc_sum['counterparty__counterparty']\n if counterparty != cp:\n counterparty = cp\n data[cp] = 0.0\n rate = AccountBalancesView.get_rate(doc_sum['currency'], main_currency)\n data[cp] += int(float(doc_sum['amount_sum']) / (float(rate) or 1.0))\n\n data_sort = sorted(data.items(), key=lambda x: x[1], reverse=True)\n top10 = dict(data_sort)\n\n return [[k, v] for k, v in top10.items()][:10]\n\n\nclass ChartsFinView(View):\n main_currency = AccountSettings.load().currency()\n\n def get(self, request):\n form = DashboardFilter(request.GET)\n currencies = Currencies.objects.all()\n agents = Counterparties.objects.all()\n initial_debts = InitialDebts.objects.all()\n paydocs = PaymentDocuments.objects.filter(item__activity='financing') \\\n .annotate(amount=F('inflow_amount') + F('outflow_amount'),\n rate=Value(AccountBalancesView.get_rate(F('currecy__id'), self.main_currency)),\n amount_convert=(F('amount') / F('rate')))\n\n # for i in paydocs:\n # print(i.date, i.currency, i.rate, i.amount_convert)\n\n paydocs_before = paydocs\n\n if form.is_valid():\n if form.cleaned_data['organization']:\n paydocs = paydocs.filter(organization=form.cleaned_data['organization'])\n paydocs_before = paydocs_before.filter(organization=form.cleaned_data['organization'])\n initial_debts = initial_debts.filter(organization=form.cleaned_data['organization'])\n if form.cleaned_data['date_start']:\n paydocs = paydocs.filter(date__gte=form.cleaned_data['date_start'])\n paydocs_before = paydocs_before.filter(date__gte=form.cleaned_data['date_start'])\n else:\n paydocs_before = PaymentDocuments.objects.none()\n if form.cleaned_data['date_end']:\n paydocs = paydocs.filter(date__lte=form.cleaned_data['date_end'])\n paydocs_before = paydocs_before.filter(date__lte=form.cleaned_data['date_end'])\n\n # credit_portfolio = self.get_loan_portfolio(agents, paydocs, initial_debts)\n # debit_portfolio = self.get_loan_portfolio(agents, paydocs, initial_debts)\n loans_table = self.get_loan_tables(agents, currencies, paydocs, paydocs_before, initial_debts)\n\n context = {\n 'form': form,\n 'today': datetime.today(),\n 'loans_table': loans_table,\n # 'credit_portfolio': credit_portfolio,\n # 'debit_portfolio': debit_portfolio,\n # 'cf_fin_dynamics': self.get_cf_dynamics(paydocs),\n }\n\n return render(request, 'registers/charts_fin.html', context=context)\n\n # chart loans_table\n def get_loan_tables(self, agents, currencies, paydocs, paydocs_before, initial_debts, ):\n loans_table = []\n print(initial_debts)\n for agent in agents:\n agent_name = str(agent)\n initial_debts = initial_debts.filter(counterparty=agent)\n paydocs = paydocs.filter(counterparty=agent)\n paydocs_before = paydocs_before.filter(counterparty=agent)\n for currency in currencies:\n currency_name = str(currency)\n initial_debt = initial_debts.filter(currency=currency)\n debit = initial_debt.aggregate(Sum(\"debit\")).get('debit__sum', 0.00)\n if debit is None:\n debit = 0.0\n credit = initial_debt.aggregate(Sum(\"credit\")).get('credit__sum', 0.00)\n if credit is None:\n credit = 0.0\n initial_debt = debit - credit\n\n # print(f'initial - {agent.counterparty}: {initial_debt}')\n\n receipts = paydocs.filter(flow='Receipts', currency=currency)\n receipts_sum = receipts.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_sum is None:\n receipts_sum = 0\n\n receipts_before = paydocs_before.filter(flow='Receipts', currency=currency)\n receipts_before_sum = receipts_before.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_before_sum is None:\n receipts_before_sum = 0\n\n payments = paydocs.filter(flow='Payments', currency=currency)\n payments_sum = payments.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_sum is None:\n payments_sum = 0\n\n payments_before = paydocs_before.filter(flow='Payments', currency=currency)\n payments_before_sum = payments_before.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_before_sum is None:\n payments_before_sum = 0\n\n start_balance = int(initial_debt) + int(receipts_before_sum) - int(payments_before_sum)\n start_debit = abs(start_balance) if start_balance > 0 else 0\n start_credit = abs(start_balance) if start_balance < 0 else 0\n\n final_balance = int(initial_debt) - int(payments_sum) + int(receipts_sum)\n final_debit = abs(final_balance) if final_balance > 0 else 0\n final_credit = abs(final_balance) if final_balance < 0 else 0\n\n loans_table.append([agent_name, int(start_debit), int(start_credit), int(receipts_sum),\n int(payments_sum), int(final_debit), int(final_credit), currency_name])\n\n print(f'table - {loans_table}')\n return loans_table\n\n # charts portfolios\n def get_loan_portfolio(self, agents, paydocs, initial_debts):\n portfolio = []\n for agent in agents:\n initial_debts = initial_debts.filter(counterparty=agent)\n print(f' the {initial_debts}')\n if initial_debts is None:\n initial_debts.credit = 0\n initial_debts.debit = 0\n # if initial_debts.debit is None:\n\n receipts = paydocs.filter(counterparty=agent)\n receipts_sum = receipts.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_sum is None:\n receipts_sum = 0\n payments = paydocs.filter(counterparty=agent)\n payments_sum = payments.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_sum is None:\n payments_sum = 0\n\n open_balance = 0 # agent.credit - agent.debit\n final_balance = abs(open_balance + receipts_sum - payments_sum)\n agent = str(agent)\n\n portfolio.append([agent, int(final_balance)])\n\n return portfolio\n\n # function for getting full data\n @staticmethod\n def get_dynamics(paydocs):\n dynamics = {}\n for paydoc in paydocs:\n month = str(paydoc.date.month).rjust(2, '0')\n period = f'{str(paydoc.date.year)}/{month}'\n amount = float(paydoc.inflow_amount) if paydoc.inflow_amount != 0 else float(paydoc.outflow_amount)\n if period not in dynamics:\n dynamics[period] = amount\n else:\n dynamics[period] += amount\n\n return dynamics\n\n # get data payments and payments\n def get_total_cf(self, paydocs):\n payments_dynamics = self.get_dynamics(paydocs)\n receipts_dynamics = self.get_dynamics(paydocs)\n\n total_cf = {k: [receipts_dynamics.get(k, 0), payments_dynamics.get(k, 0)]\n for k in set(receipts_dynamics) | set(payments_dynamics)}\n total_cf = sorted(total_cf.items(), key=lambda x: x[0])\n total_cf = dict(total_cf)\n\n return total_cf\n\n # chart 4 total cash flow\n def get_cf_dynamics(self, paydocs):\n total_cf = self.get_total_cf(paydocs)\n cf_dynamics = []\n for k, v in total_cf.items():\n cf_dynamics.append([k, *v, v[0] - v[1]])\n\n return cf_dynamics\n\n\ndef ChartsInvestView(request):\n pass\n\n\nclass ChartsFinView1(View):\n def get(self, request):\n form = DashboardFilter(request.GET)\n\n lenders = InitialDebts.objects.filter(type_debt='Lender')\n borrowers = InitialDebts.objects.filter(type_debt='Borrower')\n\n receipts = PaymentDocuments.objects.filter(flow='Receipts', item__activity='financing')\n payments = PaymentDocuments.objects.filter(flow='Payments', item__activity='financing')\n receipts_before = receipts\n payments_before = payments\n\n if form.is_valid():\n if form.cleaned_data['organization']:\n receipts = receipts.filter(organization=form.cleaned_data['organization'])\n payments = payments.filter(organization=form.cleaned_data['organization'])\n lenders = lenders.filter(organization=form.cleaned_data['organization'])\n borrowers = borrowers.filter(organization=form.cleaned_data['organization'])\n\n if form.cleaned_data['date_start']:\n receipts = receipts.filter(date__gte=form.cleaned_data['date_start'])\n payments = payments.filter(date__gte=form.cleaned_data['date_start'])\n receipts_before = receipts_before.filter(date__lte=form.cleaned_data['date_start'])\n payments_before = payments_before.filter(date__lte=form.cleaned_data['date_start'])\n else:\n receipts_before = PaymentDocuments.objects.none()\n payments_before = PaymentDocuments.objects.none()\n if form.cleaned_data['date_end']:\n receipts = receipts.filter(date__lte=form.cleaned_data['date_end'])\n payments = payments.filter(date__lte=form.cleaned_data['date_end'])\n\n loan_portfolio = self.get_loan_portfolio(lenders, receipts, payments)\n debit_portfolio = self.get_loan_portfolio(borrowers, receipts, payments)\n lenders_table = self.get_tables(lenders, receipts, payments, receipts_before, payments_before)\n borrowers_table = self.get_tables(borrowers, receipts, payments, receipts_before, payments_before)\n\n context = {\n 'form': form,\n 'today': datetime.today(),\n 'loan_portfolio': loan_portfolio,\n 'debit_portfolio': debit_portfolio,\n 'cf_fin_dynamics': self.get_cf_dynamics(receipts, payments),\n 'lenders_table': lenders_table,\n 'borrowers_table': borrowers_table,\n }\n\n return render(request, 'registers/charts_fin.html', context=context)\n\n # charts 1, 2 portfolios\n\n def get_loan_portfolio(self, agents, receipts, payments):\n portfolio = []\n for agent in agents:\n if agent.credit is None:\n agent.credit = 0\n if agent.debit is None:\n agent.debit = 0\n receipts = receipts.filter(counterparty=agent.counterparty)\n receipts_sum = receipts.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_sum is None:\n receipts_sum = 0\n payments = payments.filter(counterparty=agent.counterparty)\n payments_sum = payments.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_sum is None:\n payments_sum = 0\n\n open_balance = agent.credit - agent.debit\n final_balance = abs(open_balance + receipts_sum - payments_sum)\n agent = str(agent.counterparty)\n\n portfolio.append([agent, int(final_balance)])\n print(agents)\n print(portfolio)\n return portfolio\n\n # chart 4, 5: agents_table\n def get_tables(self, agents, receipts, payments, receipts_before, payments_before):\n agents_table = []\n\n for agent in agents:\n receipts = receipts.filter(counterparty=agent.counterparty)\n receipts_sum = receipts.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_sum == None: receipts_sum = 0\n\n receipts_before = receipts_before.filter(counterparty=agent.counterparty)\n receipts_before_sum = receipts_before.aggregate(Sum(\"inflow_amount\")).get('inflow_amount__sum', 0.00)\n if receipts_before_sum == None: receipts_before_sum = 0\n\n payments = payments.filter(counterparty=agent.counterparty)\n payments_sum = payments.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_sum == None: payments_sum = 0\n\n payments_before = payments_before.filter(counterparty=agent.counterparty)\n payments_before_sum = payments_before.aggregate(Sum(\"outflow_amount\")).get('outflow_amount__sum', 0.00)\n if payments_before_sum == None: payments_before_sum = 0\n\n start_balance = agent.debit - agent.credit + receipts_before_sum - payments_before_sum\n start_debit = abs(start_balance) if start_balance > 0 else 0\n start_credit = abs(start_balance) if start_balance < 0 else 0\n\n final_balance = agent.debit - agent.credit - receipts_sum + payments_sum\n final_debit = abs(final_balance) if final_balance > 0 else 0\n final_credit = abs(final_balance) if final_balance < 0 else 0\n agent = str(agent.counterparty)\n\n agents_table.append([agent, int(start_debit), int(start_credit), int(receipts_sum),\n int(payments_sum), int(final_debit), int(final_credit)])\n\n print(agents_table)\n return agents_table\n\n # function for getting full data\n @staticmethod\n def get_dynamics(flow):\n dynamics = {}\n for i in flow:\n month = str(i.date.month).rjust(2, '0')\n period = f'{str(i.date.year)}/{month}'\n amount = float(i.inflow_amount) if i.inflow_amount != 0 else float(i.outflow_amount)\n if period not in dynamics:\n dynamics[period] = amount\n else:\n dynamics[period] += amount\n\n return dynamics\n\n # get data payments and payments\n def get_total_cf(self, payments, receipts):\n receipts_dynamics = self.get_dynamics(receipts)\n payments_dynamics = self.get_dynamics(payments)\n\n total_cf = {k: [receipts_dynamics.get(k, 0), payments_dynamics.get(k, 0)]\n for k in set(receipts_dynamics) | set(payments_dynamics)}\n total_cf = sorted(total_cf.items(), key=lambda x: x[0])\n total_cf = dict(total_cf)\n\n return total_cf\n\n # chart 4 total cash flow\n def get_cf_dynamics(self, payments, receipts):\n total_cf = self.get_total_cf(payments, receipts)\n cf_dynamics = []\n for k, v in total_cf.items():\n cf_dynamics.append([k, *v, v[0] - v[1]])\n\n return cf_dynamics\n\n\n\n","repo_name":"Alata9/EnterpriseFinance","sub_path":"registers/views_charts.py","file_name":"views_charts.py","file_ext":"py","file_size_in_byte":28392,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"45382424178","text":"# -*- coding: utf-8 -*-\n\nimport telebot # Librería de la API del bot.\nfrom telebot import types # Tipos para la API del bot.\nimport time # Librería para hacer que el programa que controla el bot no se acabe.\nimport wikipedia\nimport picamera\nfrom time import sleep\nTOKEN = 'Aquí va el tokken' # Nuestro tokken del bot (el que @BotFather nos dió).\n\nbot = telebot.TeleBot(TOKEN) # Creamos el objeto de nuestro bot.\n@bot.message_handler(commands=['hora']) # Indicamos que lo siguiente va a controlar el comando '/hora'.\ndef command_hora(m): # Definimos una función que resuelva lo que necesitemos.\n cid = m.chat.id # Guardamos el ID de la conversación para poder responder.\n bot.send_message( cid, time.strftime(\"%H:%M:%S\")) # Con la función 'send_message()' del bot, enviamos al ID almacenado el texto que queremos.\n@bot.message_handler(commands=['prueba']) # Indicamos que lo siguiente va a controlar el comando '/prueba'\ndef command_prueba(m): # Definimos una función que resuleva lo que necesitemos.\n cid = m.chat.id # Guardamos el ID de la conversación para poder responder.\n bot.send_message( cid, 'probando') # Con la función 'send_message()' del bot, enviamos al ID almacenado el texto que queremos.\n@bot.message_handler(commands=['wiki'])\ndef command_wiki(m):\n cid = m.chat.id\n msg = m.text[6:]\n if msg =='':\n bot.send_message(cid, \"Escribe algo despues de /wiki\")\n else :\n try:\n bot.send_message(cid,wikipedia.summary(msg, sentences=6))\n except wikipedia.exceptions.DisambiguationError as e:\n bot.send_message(cid, e)\n@bot.message_handler(commands=['picamera'])\ndef command_picamera(m):\n cid = m.chat.id\n msg = m.text[6:]\n try:\n with picamera.PiCamera() as camera:\n# camera.hflip = True\n# camera.vflip = True\n camera.capture(\"imagen.jpg\")\n sleep(1)\n photo = open('imagen.jpg', 'rb')\n bot.send_photo(cid, photo)\n except :\n bot.send_message(cid, \"Fallo al arrancar la camara, prueba mas tarde\")\n\nbot.polling(none_stop=True) # Con esto, le decimos al bot que siga funcionando incluso si encuentra algún fallo.\n","repo_name":"inakidml/PiCameraToTelegramBot","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":2170,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"3012645754","text":"with open(\"input\") as file:lines = file.read().splitlines()\n\nwhen=int(lines[0])\nbuses= list(map(int,list(filter(lambda b: b!=\"x\", lines[1].split(\",\")))))\n\nwait=0\nfound=False\nwhile not(found):\n for b in buses:\n if (when%b==0):\n found=True\n break\n else:\n wait+=1\n when+=1\n\nprint(\"part 1: \",wait*b)\nprint()\n\n#part 2\n\n#copy paste from somewhere\ndef gcd(a,b):\n \"\"\"Compute the greatest common divisor of a and b\"\"\"\n while b > 0:\n a, b = b, a % b\n return a\n \ndef lcm(a, b):\n \"\"\"Compute the lowest common multiple of a and b\"\"\"\n return a * b / gcd(a, b)\n##\n\nbusesAll = lines[1].split(\",\")\n\nbig = max(buses)\nwhen=0\n\nbigIndex=busesAll.index(str(big))\n\nbusesWithIndexDiffs={}\nindexDiffFromFirstAndBig=\"\"\n\nfor i in range(len(busesAll)):\n if (busesAll[i]!=\"x\"):\n if(int(busesAll[i])!=big):\n busesWithIndexDiffs[int(busesAll[i])] = -(bigIndex-i)\n if indexDiffFromFirstAndBig==\"\":\n indexDiffFromFirstAndBig=-(bigIndex-i)\n\n#print(busesWithIndexDiffs)\n\nfound=set([big])\nincr=big\nwhile len(found)= 2:\n wfs = wfs[:, 0]\n return wfs\n\n\ndefault_mall = dict(\n # Factory Input Stores\n sound_output={k: v for k, v in upload_files_store.getter_items()},\n step_factories=dict(\n # ML\n chunker=FuncFactory(simple_chunker),\n featurizer=FuncFactory(simple_featurizer),\n ),\n # Output Store\n data=dict(),\n steps={k: v for k, v in pipeline_step_store.getter_items()},\n pipelines=dict(),\n exec_outputs=dict(),\n learned_models=dict(),\n models_scores=dict(),\n source=None,\n)\n\nmall = get_mall(defaults=default_mall)\ncrudifier = partial(Crudifier, mall=mall)\n\n\n@add_to_upload_files_store\n@crudifier(output_store='sound_output')\ndef upload_sound(train_audio: List[WaveForm], tag: str):\n return train_audio, tag\n\n\ndef get_step_name(step):\n return [k for k, v in get_mall()['steps'].items() if v == step][0]\n\n\ndef _save_name_getter(args, kwargs, function=None, return_value=None):\n return kwargs['save_name']\n\n\n@Persist.function_call(key_getter=_save_name_getter, store=pipeline_step_store)\n@crudifier(param_to_mall_map=dict(step_factory='step_factories'), output_store='steps')\ndef mk_step(step_factory: Callable, kwargs: dict):\n kwargs = clean_dict(kwargs)\n step = partial(step_factory, **kwargs)()\n\n return step\n\n\ndef get_selected_step_factory_sig():\n selected_step_factory = get_mall()['step_factories'].get(\n b.selected_step_factory.get()\n )\n if selected_step_factory:\n return Sig(selected_step_factory)\n\n\n@crudifier(output_store='pipelines',)\n@resolve_item_getter_args\ndef mk_pipeline(steps: Iterable[Callable]):\n return LineParametrized(*steps)\n\n\n@crudifier(\n param_to_mall_map=dict(tagged_data='sound_output', preprocess_pipeline='pipelines'),\n output_store='learned_models',\n)\n@UploadFilesStore.resolve_item_getter_args\ndef learn_outlier_model(tagged_data, preprocess_pipeline, n_centroids=5):\n\n sound, tag = tagged_sounds_to_single_array(*tagged_data)\n wfs = np.array(sound)\n\n wfs = assert_dims(wfs)\n\n fvs = preprocess_pipeline(wfs)\n model = Stroll(n_centroids=n_centroids)\n model.fit(X=fvs)\n\n return model\n\n\n@crudifier(\n param_to_mall_map=dict(\n tagged_data='sound_output',\n preprocess_pipeline='pipelines',\n fitted_model='learned_models',\n ),\n output_store='models_scores',\n)\n@UploadFilesStore.resolve_item_getter_args\ndef apply_fitted_model(tagged_data, preprocess_pipeline, fitted_model):\n try:\n dill_files['tagged_data'] = tagged_data\n except Exception as e:\n print('tagged_data', e)\n try:\n dill_files['preprocess_pipeline'] = preprocess_pipeline\n except Exception as e:\n print('preprocess_pipeline', e)\n try:\n dill_files['fitted_model'] = fitted_model\n except Exception as e:\n print('fitted_model', e)\n sound, tag = tagged_sounds_to_single_array(*tagged_data)\n wfs = np.array(sound)\n wfs = assert_dims(wfs)\n\n fvs = preprocess_pipeline(wfs)\n scores = fitted_model.score_samples(X=fvs)\n return scores\n\n\n@crudifier(param_to_mall_map=dict(pipeline='pipelines'),)\ndef visualize_pipeline(pipeline: LineParametrized):\n\n return pipeline\n\n\n@crudifier(param_to_mall_map=dict(scores='models_scores'),)\ndef visualize_scores(scores, threshold=80, num_segs=3):\n\n intervals = scores_to_intervals(scores, threshold, num_segs)\n\n return scores, intervals\n\n\n@crudifier(\n param_to_mall_map=dict(\n preprocess_pipeline='pipelines', fitted_model='learned_models',\n ),\n output_store='source',\n)\ndef live_apply_fitted_model(\n preprocess_pipeline,\n fitted_model,\n input_device=None,\n rate=44100,\n width=2,\n channels=1,\n frames_per_buffer=44100,\n seconds_to_keep_in_stream_buffer=60,\n graph_types='volume',\n):\n stop_stream()\n audio_store_rootdir = Path.cwd() / 'audio_store'\n audio_store_rootdir.mkdir(parents=True, exist_ok=True)\n audio_store = WavFileStore(rootdir=str(audio_store_rootdir))\n\n source = mk_live_graph_data_buffer(\n input_device,\n rate,\n width,\n channels,\n frames_per_buffer,\n seconds_to_keep_in_stream_buffer,\n graph_types,\n audio_store=audio_store,\n **si_apply_fitted_model(preprocess_pipeline, fitted_model),\n )\n source.start()\n return source\n","repo_name":"otosense/plunk","sub_path":"plunk/ap/app4_drill_but_make_it_slabsiter/components.py","file_name":"components.py","file_ext":"py","file_size_in_byte":5999,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"16735667301","text":"import numpy as np\nfrom data import get_in_out_simple_predictive_NN\nfrom sklearn.model_selection import train_test_split\nfrom tensorflow.keras import Sequential, layers\nimport matplotlib.pyplot as plt\n\n# Parameters\nlearning_rate = 1.59e-7\nmomentum = 0.95\nnb_epochs = 10\nbatch_size = 128\ntest_ratio = 0.3\nnb_neurone_layer_1 = 256\nnb_neurone_layer_2 = 256\nnb_neurone_layer_3 = 64\n\n\ndef generate_predictive_NN():\n \"\"\"\n Generate the predictive NN which predict the next angles of the arm based on the current ones\n\n :return: None\n \"\"\"\n # Load data\n inputs, outputs = get_in_out_simple_predictive_NN()\n n, p = inputs.shape\n nb_inputs = 12 # 17\n nb_outputs = 5\n\n X_train, X_test, y_train, y_test = train_test_split(inputs, outputs, test_size=test_ratio)\n\n mod = Sequential()\n mod.add(layers.Dense(nb_neurone_layer_1, activation='relu', input_shape=(nb_inputs,)))\n mod.add(layers.Dense(nb_neurone_layer_2, activation='relu'))\n mod.add(layers.Dropout(0.5))\n mod.add(layers.Dense(nb_neurone_layer_3, activation='relu'))\n mod.add(layers.Dense(nb_outputs, activation='linear'))\n mod.compile(loss='mse')\n print(\"Model created !\")\n\n print(\"Training of the model ...\")\n hist = mod.fit(X_train, y_train, batch_size=batch_size, epochs=nb_epochs, validation_data=(X_test, y_test))\n print(\"Model trained !\")\n mod.save('models/predictive_NN')\n print(\"Model saved !\")\n\n loss_values = hist.history['loss']\n val_loss_values = hist.history['val_loss']\n epochs = range(1, len(loss_values) + 1)\n\n plt.plot(epochs, loss_values, label='Training Loss')\n plt.plot(epochs, val_loss_values, label='Validation Loss')\n plt.xlabel('Epoch')\n plt.ylabel('Loss')\n plt.legend()\n\n plt.show()\n","repo_name":"MarcDcls/AI_Prosthesis","sub_path":"predictive_NN.py","file_name":"predictive_NN.py","file_ext":"py","file_size_in_byte":1750,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25244091624","text":"'使用threading 模块创建线程,直接从threading.Thread继承'\n\nimport threading\nimport time\n\n\nexitFlag = 0\n\n\ndef print_time(threadName, delay, counter):\n while counter:\n if exitFlag:\n threading.Thread.exit()\n time.sleep(delay)\n print('{0}: {1}'.format(threadName, time.ctime(time.time())))\n counter -= 1\n\n\nclass MyThread(threading.Thread):\n '''\n 继承threaing.Thread类,重写__init__()和run()方法\n '''\n\n def __init__(self, threadID, name, counter):\n '''\n 重写__init__()\n :param threadID:\n :param name: 线程名字\n :param counter:\n '''\n threading.Thread.__init__(self)\n self.threadID = threadID\n self.name = name\n self.counter = counter\n\n def run(self):\n '''\n 将要执行的代码写入run()函数里,线程在创建后会直接运行run()函数\n :return:\n '''\n print('starting ', self.name)\n print_time(self.name, self.counter, 5)\n print('Exiting ', self.name)\n\n\n\ndef main():\n # 创建线程\n thread1 = MyThread(1, 'Thread-1', 1)\n thread2 = MyThread(2, 'Thread-2', 2)\n\n # 启动线程\n thread1.start()\n thread2.start()\n\n print('Exiting Main Thread')\n\nif __name__ == '__main__':\n main()","repo_name":"haalo/learning-Python3","sub_path":"day01/threadings_test/threading_01.py","file_name":"threading_01.py","file_ext":"py","file_size_in_byte":1306,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"10442111016","text":"from typing import Dict, List, Tuple\nimport argparse\nimport json\nimport re\nfrom unidecode import unidecode\nfrom pathlib import Path\nfrom redminelib import Redmine\nfrom redminelib.resources import Issue as RedmineIssue\nfrom redminelib.managers import ResourceManager as RedmineResource\n\nurl = 'https://redmine.apidb.org'\ndefault_fields = dict(\n status_name='Data Processing (EBI)',\n cf_17=\"Data Processing (EBI)\",\n)\ninsdc_pattern = r'^GC[AF]_\\d{9}(\\.\\d+)?$'\naccession_api_url = \"https://www.ebi.ac.uk/ena/browser/api/xml/%s\"\nveupathdb_id = 1976\n\n\ndef load_abbrevs(path: str) -> List[str]:\n \"\"\"\n Load a list of organism abbrevs from a file. Expected to be one per line.\n\n Args:\n path: Path to the organism abbrevs file.\n\n Returns:\n A list of all organism_abbrevs.\n\n \"\"\"\n if not path:\n print(\"Warning: I don't have a list of older abbrevs to compare with.\")\n return []\n abbrevs = []\n with open(path, \"r\") as abbr_file:\n for line in abbr_file:\n line = line.rstrip()\n if line:\n fields = line.split(\"\\t\")\n if len(fields) == 1:\n abbrevs.append(line)\n else:\n raise Exception(\n \"Can't load current abbrevs from a multicolumn string\")\n return abbrevs\n\n\ndef retrieve_rnaseq_datasets(redmine: Redmine, output_dir_path: str, build: int = None,\n abbrevs_file: str = None) -> None:\n \"\"\"\n Get RNA-Seq metadata from Redmine, store them in json files.\n Each issue/dataset is stored as one file in the output dir.\n\n Args:\n redmine: A connected Redmine object.\n output_dir_path: Directory where the dataset files are to be stored.\n build: BRC build number.\n abbrevs_file: Path to a list of organism_abbrevs that are already in use.\n \"\"\"\n\n all_abbrevs = load_abbrevs(abbrevs_file)\n\n issues = get_issues(redmine, \"RNA-seq\", build)\n if not issues:\n print(\"No files to create\")\n return\n \n # Create the output dir\n output_dir = Path(output_dir_path)\n output_dir.mkdir(exist_ok=True)\n \n # Write all datasets in files\n all_datasets = []\n used_names = []\n\n problems = []\n ok_datasets = []\n warn_abbrevs = []\n \n for issue in issues:\n dataset, problem = parse_dataset(issue)\n \n if problem:\n problems.append({\"issue\": issue, \"desc\": problem})\n continue\n\n try:\n component = dataset[\"component\"]\n organism = dataset[\"species\"]\n dataset_name = dataset[\"name\"]\n \n if dataset_name in used_names:\n problems.append(\n {\"issue\": issue, \"desc\": f\"Dataset name already used: {dataset_name}\"}\n )\n continue\n else:\n used_names.append(dataset_name)\n \n if abbrevs_file and organism not in all_abbrevs:\n warn_abbrevs.append({\"issue\": issue, \"desc\": organism})\n \n ok_datasets.append({\"issue\": issue, \"desc\": organism})\n \n # Create directory\n dataset_dir = output_dir / component\n dataset_dir.mkdir(exist_ok=True)\n \n # Create file\n file_name = organism + \"_\" + dataset_name + \".json\"\n dataset_file = dataset_dir / file_name\n with open(dataset_file, \"w\") as f:\n json.dump([dataset], f, indent=True)\n except Exception as error:\n problems.append({\"issue\": issue, \"desc\": str(error)})\n pass\n all_datasets.append(dataset)\n\n print(\"%d issues total\" % len(issues))\n print_summaries(problems, \"issues with problems\")\n print_summaries(\n warn_abbrevs,\n \"issues using unknown organism_abbrevs (maybe new ones). Those are still imported\"\n )\n print_summaries(ok_datasets, \"datasets imported correctly\")\n\n # Create a single merged file as well\n merged_file = Path(output_dir) / \"all.json\"\n with open(merged_file, \"w\") as f:\n json.dump(all_datasets, f, indent=True)\n\n \ndef print_summaries(summaries: Dict, description: str) -> None:\n \"\"\"Print a summary of various counts.\n\n This will print one line for each issue in the dict, with its description, the issue id\n and the issue subject.\n\n Args:\n summaries: Dict with 2 keys:\n issue: A Redmine Issue object.\n desc: A description for that issue.\n \"\"\"\n desc_length = 64\n \n if summaries:\n print()\n print(f\"{len(summaries)} {description}:\")\n for summary in summaries:\n desc = summary[\"desc\"]\n issue = summary[\"issue\"]\n print(f\"\\t{desc:{desc_length}}\\t{issue.id}\\t({issue.subject})\")\n \n \ndef parse_dataset(issue: RedmineIssue) -> Tuple[Dict, str]:\n \"\"\"\n Extract RNA-Seq dataset metadata from a Redmine issue.\n\n Args:\n issue: A Redmine issue.\n\n Returns:\n A tuple of 2 objects:\n datasets: A dict representing a dataset, with the following keys:\n component: String for the BRC component DB.\n species: String for the organism abbrev.\n name: String for the internal dataset name.\n problem: A string description if there was a parsing problem\n (empty string otherwise).\n \"\"\"\n customs = get_custom_fields(issue)\n dataset = {\n \"component\": \"\",\n \"species\": \"\",\n \"name\": \"\",\n \"runs\": [],\n }\n problem = \"\"\n\n dataset[\"component\"] = get_custom_value(customs, \"Component DB\")\n dataset[\"species\"] = get_custom_value(customs, \"Organism Abbreviation\").strip()\n dataset[\"name\"] = get_custom_value(customs, \"Internal dataset name\").strip()\n\n if not dataset[\"species\"]:\n problem = \"Missing Organism Abbreviation\"\n elif not check_organism_abbrev(dataset[\"species\"]):\n problem = f\"Wrong Organism Abbreviation format: '{dataset['species']}'\"\n elif not dataset[\"name\"]:\n problem = \"Missing Internal dataset name\"\n else:\n dataset[\"name\"] = normalize_name(dataset[\"name\"])\n \n # Get samples/runs\n samples_str = get_custom_value(customs, \"Sample Names\")\n try:\n samples = parse_samples(samples_str)\n \n if not samples:\n problem = \"Missing Samples\"\n \n dataset[\"runs\"] = samples\n except Exception as e:\n problem = str(e)\n \n return dataset, problem\n\n\ndef check_organism_abbrev(name: str) -> bool:\n \"\"\"Check the organism_abbrevs string format to avoid special characters.\n\n Args:\n name: organism_abbrev to check.\n \n Returns:\n True if the organism_abbrev format is correct.\n False otherwise.\n \"\"\"\n return not re.search(r'[ \\/\\(\\)#\\[\\]:]', name)\n\n\ndef normalize_name(old_name: str) -> str:\n \"\"\"Remove special characters from an organism_abbrev, keep ascii only.\n\n Args:\n old_name: the organism_abbrev to format.\n \n Returns:\n The formatted organism_abbrev.\n \"\"\"\n \n # Remove any diacritics\n name = old_name.strip()\n name = unidecode(name)\n name = re.sub(r\"[ /]\", \"_\", name)\n name = re.sub(r\"[;:.,()\\[\\]{}]\", \"\", name)\n name = re.sub(r\"\\+\", \"_plus_\", name)\n name = re.sub(r\"\\*\", \"_star_\", name)\n name = re.sub(r\"%\", \"pc_\", name)\n name = re.sub(r\"_+\", \"_\", name)\n if re.search(r\"[^A-Za-z0-9_.-]\", name):\n print(\"WARNING: name contains special characters: %s (%s)\" % (old_name, name))\n return\n \n return name\n\n\ndef parse_samples(sample_str: str) -> List[Dict]:\n \"\"\"Parse a list of samples from a Redmine task.\n\n Args:\n sample_str: The value of the field 'Sample Names' from an RNA-Seq Redmine task.\n \n Returns:\n A list of samples dicts, with the following keys:\n name: the name of the sample.\n accessions: a list of string representing the SRA accessions for that sample.\n \"\"\"\n samples = []\n \n # Parse each line\n lines = sample_str.split(\"\\n\")\n\n sample_names = dict()\n for line in lines:\n line = line.strip()\n if line == \"\":\n continue\n\n # Get sample_name -> accessions\n parts = line.split(\":\")\n if len(parts) > 2:\n end = parts[-1]\n start = \":\".join(parts[:-1])\n parts = [start, end]\n \n if len(parts) == 2:\n sample_name = parts[0].strip()\n \n if sample_name in sample_names:\n raise Exception(\"Several samples have the same name '%s'\" % sample_name)\n else:\n sample_names[sample_name] = True\n \n accessions_str = parts[1].strip()\n accessions = [x.strip() for x in accessions_str.split(\",\")]\n \n if not validate_accessions(accessions):\n if validate_accessions(sample_name.split(\",\")):\n raise Exception(\"Sample name and accessions are switched?\")\n else:\n raise Exception(f\"Invalid accession among '{accessions}'\")\n \n sample = {\n \"name\": normalize_name(sample_name),\n \"accessions\": accessions\n }\n samples.append(sample)\n else:\n raise Exception(\"Sample line doesn't have 2 parts: '%s'\" % line)\n \n return samples\n\n\ndef validate_accessions(accessions: str) -> bool:\n \"\"\"Check SRA accessions format, to make sure we get proper ones.\n\n Args:\n accessions: a list of strings to check\n\n Return:\n True if all strings are proper SRA accessions.\n False if at least one is not a proper SRA accession.\n \"\"\"\n if \"\" in accessions:\n return False\n for acc in accessions:\n if not re.search(r'^[SE]R[RSXP]\\d+$', acc):\n return False\n return True\n\n\ndef get_custom_fields(issue: RedmineIssue) -> Dict:\n \"\"\"Put all Redmine custom fields in a dict instead of an array.\n\n Args:\n issue: A Redmine issue.\n \n Returns:\n A dict where each key is a custom field.\n \"\"\"\n \n cfs = {}\n for c in issue.custom_fields:\n cfs[c[\"name\"]] = c\n return cfs\n\n\ndef get_custom_value(customs: Dict, key: str) -> str:\n \"\"\"Retrieve a custom value from a customs dict.\n\n Args:\n customs: Dict of customs values gotten from get_custom_fields.\n key: Key to extract the value from the custom dict.\n \n Returns:\n A single value.\n Throws an exception if there are more than 1 value.\n If there is no such key in the dict, return an empty string.\n \"\"\"\n \n try:\n value = customs[key][\"value\"]\n if isinstance(value, list):\n if len(value) == 1:\n value = value[0]\n elif len(value) > 1:\n raise Exception(\"More than 1 values for key %s\" % (key))\n return value\n except KeyError:\n print(\"No field %s\" % (key))\n return \"\"\n \n\ndef get_issues(redmine: Redmine, datatype: str, build: int = None) -> List[RedmineIssue]:\n \"\"\"Retrieve all issue for new genomes, be they with or without gene sets.\n\n Args:\n redmine: A Redmine connected object.\n datatype: What datatype to use to filter the issues.\n build: The BRC build to use to filter.\n\n Returns:\n A list of Redmine issues.\n \"\"\"\n \n other_fields = {\"cf_94\": datatype}\n if build:\n version_id = get_version_id(redmine, build)\n other_fields[\"fixed_version_id\"] = version_id\n\n return list(get_ebi_issues(redmine, other_fields))\n\n\ndef get_version_id(redmine: Redmine, build: int) -> int:\n \"\"\"Given a build number, get the Redmine version id for it.\n\n Args:\n redmine: A Redmine connected object.\n build: The BRC build to use to filter.\n\n Returns:\n The version id from Redmine for that build.\n \"\"\"\n versions = redmine.version.filter(project_id=veupathdb_id)\n version_name = \"Build \" + str(build)\n version_id = [version.id for version in versions if version.name == version_name]\n return version_id\n\n \ndef get_ebi_issues(redmine, other_fields=dict()) -> RedmineResource:\n \"\"\"Get EBI issues from Redmine, add other fields if provided.\n\n Args:\n redmine: A Redmine connected object.\n other_fields: A dict of fields to provide to filter the issues.\n\n Returns:\n A Redmine resource set.\n \"\"\"\n # Other fields replace the keys that already exist in default_fields\n search_fields = {**default_fields, **other_fields}\n \n return redmine.issue.filter(**search_fields)\n \n\ndef main():\n parser = argparse.ArgumentParser(description='Retrieve metadata from Redmine')\n \n parser.add_argument('--key', type=str, required=True,\n help='Redmine authentification key')\n parser.add_argument('--output_dir', type=str, required=True,\n help='Output_dir')\n # Choice\n parser.add_argument('--get', choices=['rnaseq', 'dnaseq'], required=True,\n help='Get rnaseq, or dnaseq issues')\n # Optional\n parser.add_argument('--build', type=int,\n help='Restrict to a given build')\n parser.add_argument('--current_abbrevs', type=str,\n help='File that contains the list of current organism_abbrevs')\n args = parser.parse_args()\n \n # Start Redmine API\n redmine = Redmine(url, key=args.key)\n \n # Choose which data to retrieve\n if args.get == 'rnaseq':\n retrieve_rnaseq_datasets(redmine, args.output_dir, args.build, args.current_abbrevs)\n elif args.get == 'dnaseq':\n # TODO\n # retrieve_dnaseq_datasets(redmine, args.output_dir, args.build, args.current_abbrevs)\n print(\"Not yet implemented\")\n else:\n print(\"Need to say what data you want to --get: rnaseq? dnaseq?\")\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"Ensembl/ensembl-production-imported","sub_path":"scripts/brc4/get_rnaseq_from_redmine.py","file_name":"get_rnaseq_from_redmine.py","file_ext":"py","file_size_in_byte":14026,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"32670336966","text":"import os\nfrom discord.ext import commands\nfrom database import session, Prefixes\nfrom discord import Intents\n\ndef get_prefix(_bot, message):\n try:\n return session.query(Prefixes).filter_by(server_id=message.guild.id).one().prefix\n except:\n return \"!!\"\n\ndef main():\n intents: Intents = Intents.default()\n\n Bot = commands.Bot(command_prefix=get_prefix, intents=intents)\n Bot.remove_command(\"help\")\n\n for f in os.listdir('./cogs'):\n if f.endswith('.py'):\n Bot.load_extension(f'cogs.{f[:-3]}')\n\n token = os.environ['TOKEN']\n\n Bot.run(token)\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"Darth-Kylokun/prayer-bot","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"38828804081","text":"f = open('text_of_program.txt', 'r', encoding='utf-8')\n\narray = [row.strip() for row in f]\nfor i in array:\n if(i == ''):\n array.remove(i) #удалить пробелы\n\nnew_array = []\nfor i in array:\n new_array.append(list(i)) #разбить строки на символы\n\nfor i in range(len(new_array) - 1): #удаление всех string в строках\n start_symbol = False\n for j in range(len(new_array[i]) - 1):\n if (start_symbol == False and new_array[i][j] == '\"'):\n start_symbol = True\n elif (start_symbol == True and new_array[i][j] == '\"'):\n start_symbol = False\n new_array[i][j] = False\n elif (start_symbol == True):\n new_array[i][j] = False\n\ntext_without_string = \"\"\n\nfor i in range(len(new_array)):\n for j in range(len(new_array[i])):\n text_without_string += str(new_array[i][j])\n text_without_string += '\\n'\n\nlines_text_without_string = text_without_string.split('\\n')\n\nnodes_count = 0\nfor i in lines_text_without_string:\n if (len(i) > 1 and i.count('{') == 0 and i.count('}') == 0 and i.count('};') == 0):\n nodes_count += 1\n\nprint('Количество узлов = ', nodes_count)\n\nedge_count = 0\nfor i in lines_text_without_string:\n if (len(i) > 1 and i.count('{') == 0 and i.count('}') == 0 and i.count('};') == 0):\n edge_count += 1\n if (i.count('if') > 0):\n edge_count += 2\n if (i.count('else') > 0):\n edge_count += 1\n if (i.count('for') > 0):\n edge_count += 1\n if (i.count('while') > 0):\n edge_count += 1\n if (i.count('switch') > 0):\n edge_count += 1\n if (i.count('case') > 0):\n edge_count += 1\n\nprint('Количество рёбер = ', edge_count)\n\nV = edge_count - nodes_count + 2\nprint('Метрика Мак-Кейба = ', V)\n\nf.close()","repo_name":"VitasVital/Bachelor_course","sub_path":"4Cours1Sem/Technologies_and_development_standards/Gusev.V.E.09-832_lab_2.py","file_name":"Gusev.V.E.09-832_lab_2.py","file_ext":"py","file_size_in_byte":1852,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"43554768237","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Aug 12 14:50:39 2021\n\nTODO:\n multidimensional;\n Duchinsky rotation;\n\n@author: Bing\n\"\"\"\n\nimport math\n\nfrom math import factorial\nfrom math import sqrt, exp\nfrom scipy.special import hermite, binom\n# import numba\n\ndef dfactortial(n):\n return math.prod(range(n, 0, -2))\n\n# @numba.jit\ndef FranckCondon(Ln, Lm, d):\n '''\n Analytical formula for the Franck-Condon factors from\n\n Chang, J.-L. Journal of Molecular Spectroscopy 232, 102–104 (2005).\n\n Parameters\n ----------\n Ln : TYPE\n DESCRIPTION.\n Lm : TYPE\n DESCRIPTION.\n d : TYPE\n DESCRIPTION.\n\n Returns\n -------\n float\n Franck-Condon overlap.\n\n '''\n\n # If the excited state frequency (Ln[1]) is greater than the ground state\n # frequency (Lm[1]) then we must swap Ln and Lm for the program, but then\n # take the absolute value of the result.\n if (Ln[1] > Lm[1]):\n Ln, Lm = Lm, Ln\n\n n = Ln[0]\n m = Lm[0]\n wn_wavenumbers = Ln[1]\n wm_wavenumbers = Lm[1]\n\n wn = wn_wavenumbers/8065.5/27.2116\n wm = wm_wavenumbers/8065.5/27.2116\n # f = float(wn)/wm\n # w = wm\n\n # The formula is used for (x+d)^2 whereas I use (x-d)^2 for\n # the excited-state surface\n d = -d\n # F is the (massless) force constant for the mode. But which w?\n # F = w ** 2\n\n #convertedQSquared = deltaQ**2/(6.02214*(10**23) * 9.1094*(10**-28))\n # convertedQSquared = deltaQ**2\n\n\n # X is defined as such in Siders, Marcus 1981 Average frequency?\n # X = convertedQSquared / 2\n A = 2. * sqrt(wn * wm)/(wn + wm)\n S = d**2 * wn*wm/(wn + wm)\n\n p = sqrt(A*exp(-S)/(factorial(n) * factorial(m))) / 2.**((n+m)/2)\n\n def I(i,j):\n if (i+j) % 2 == 0:\n K = (i+j)//2\n return dfactortial(i+j-1)/(wn + wm)**K\n else:\n return 0\n\n fc = 0\n for i in range(n+1):\n F = hermite(n-i)\n bn = - wm * sqrt(wn)* d/(wn + wm)\n\n for j in range(m+1):\n\n G = hermite(m-j)\n bm = wn * sqrt(wm) * d/(wn + wm)\n\n fc += binom(n, i) * binom(m, j) * F(bn) * G(bm) * (2*sqrt(wn))**i *\\\n (2*sqrt(wm))**j * I(i, j)\n\n return fc * p\n\n\n\n\n# def genIntensities( deltaE, deltaQ, w_wavenumbers, wprime_wavenumbers):\n# \"\"\" wprime must be greater than w\"\"\"\n# wprime = wprime_wavenumbers/8065.5/27.2116\n# w = w_wavenumbers/8065.5/27.2116\n# intensityFunction = lambda n: (diffFreqOverlap([n, wprime_wavenumbers], [0, w_wavenumbers], deltaQ))**2\n# intensities = map(intensityFunction, range(0,11))\n# return intensities\n\n# def genEnergies(deltaE, w_wavenumbers, wprime_wavenumbers):\n# wprime = wprime_wavenumbers/8065.5/27.2116\n# w = w_wavenumbers/8065.5/27.2116\n# energyFunction = lambda n: (deltaE + (n+0.5)*(wprime) - 0.5*w)\n# energies = map(energyFunction, range(0, 11))\n# return energies\n\n\nif __name__ == '__main__':\n\n fc = FranckCondon([2, 500], [2, 500], 0)\n print(fc)","repo_name":"binggu56/pyqed","sub_path":"pyqed/FranckCondon.py","file_name":"FranckCondon.py","file_ext":"py","file_size_in_byte":2994,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"7"} +{"seq_id":"72685403104","text":"class RecentCounter:\n\n def __init__(self):\n self.count = 0\n self.requests = []\n\n def ping(self, t: int) -> int:\n in_range_count = 1\n i = 0\n for req in self.requests:\n if t - 3000 <= req <= t:\n in_range_count += 1\n elif req > t:\n break\n i += 1\n self.requests.insert(i, t)\n return in_range_count\n\n\n# Your RecentCounter object will be instantiated and called as such:\n# obj = RecentCounter()\n# param_1 = obj.ping(t)\n\n\n# TLE\n# https://leetcode.com/submissions/detail/403134818/testcase/\n","repo_name":"daviddwlee84/LeetCode","sub_path":"Python3/Design/NumberOfRecentCells/Naive933.py","file_name":"Naive933.py","file_ext":"py","file_size_in_byte":601,"program_lang":"python","lang":"en","doc_type":"code","stars":15,"dataset":"github-code","pt":"7"} +{"seq_id":"7152033944","text":"import random\nimport time\n\nlottery = []\n\nwhile len(lottery) != 6:\n lottery.append(random.randint(1, 45))\n lottery = list(set(lottery))\n\nfor key, value in enumerate(lottery):\n print(\"%d번 숫자\" % (key+1), end=\" \")\n time.sleep(1)\n print(\"%d입니다.\" % value)\n time.sleep(0.5)\n","repo_name":"Meengkko/bigdata_python2019","sub_path":"01_jump_to_python/5_APP/exer2/q13.py","file_name":"q13.py","file_ext":"py","file_size_in_byte":297,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"35018432676","text":"# Configuration file for the Sphinx documentation builder.\n#\n# For the full list of built-in configuration values, see the documentation:\n# https://www.sphinx-doc.org/en/master/usage/configuration.html\n\n# -- Project information -----------------------------------------------------\n# https://www.sphinx-doc.org/en/master/usage/configuration.html#project-information\n\nproject = 'Dummy'\ncopyright = '2022, Ben'\nauthor = 'Ben'\nrelease = '1.0'\n\n# -- General configuration ---------------------------------------------------\n# https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration\n\nextensions = [\n\t\"autoapi.extension\",\n\t\"sphinx_rtd_theme\",\n \"myst_parser\",\n]\n\ntemplates_path = ['_templates']\nexclude_patterns = []\n\nautoapi_type = 'python'\nautoapi_dirs = ['../../src']\n\nsource_suffix = {\n \".rst\": \"restructuredtext\",\n \".md\": \"markdown\",\n}\n\nmyst_enable_extensions = [\n \"linkify\", # trun URLs into links automatically\n \"dollarmath\", # support for $..$ and $$..$$ math environments\n \"substitution\", # enable jina2 style substitutions\n \"deflist\", # enable markup for Pandoc style definitions \n \"html_image\", # enable support for HTML image tags\n \"html_admonition\", # enable support for HTML info/warning/tip boxes\n \"colon_fence\", # enable support for colon fence environment, e.g. for figure-md\n]\nmyst_dmath_double_inline = True # enable support for inline $$-blocks \nmyst_heading_anchors = 4 # enable automatic anchor generation down to n-th level headings\n\n# -- Options for HTML output -------------------------------------------------\n# https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output\n\nhtml_theme = \"sphinx_rtd_theme\"\nhtml_static_path = ['_static']\n","repo_name":"bwulff/dummy_for_myst_parser","sub_path":"doc/source/conf.py","file_name":"conf.py","file_ext":"py","file_size_in_byte":1846,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"4301276804","text":"# This script converts all jpg file in current dir to thumbnail size and save the files in 'thumbnails' folder\n\nfrom PIL import Image\nimport glob\nimport os\n\nsize = 128, 128\n\nif not os.path.exists('thumbnails'):\n os.makedirs('thumbnails')\n\nfor infile in glob.glob(\"*.jpg\"):\n file, ext = os.path.splitext(infile)\n im = Image.open(infile)\n im.thumbnail(size)\n os.chdir('thumbnails')\n im.save(file + \"_thumbnail\" + \".jpg\")\n os.chdir(os.pardir)\n","repo_name":"kamalhisyam/rally-workshop","sub_path":"run3/src/to_thumbnails.py","file_name":"to_thumbnails.py","file_ext":"py","file_size_in_byte":461,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7973048844","text":"\n\n #2- search\nN =int(input())\nA = input().split()\nM = int(input())\n\ni=0\nfound = False\nfor i in range(len(A)):\n if(int(A[i])==M):\n print(i)\n found = True\n break\nif(found == False):\n print(-1)\n","repo_name":"HusseinAlswasy/Sheet3","sub_path":"Search.py","file_name":"Search.py","file_ext":"py","file_size_in_byte":262,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"42081783568","text":"import sys\r\nimport os\r\nimport zipfile\r\nimport shutil\r\nimport nibabel as nib\r\nimport tarfile\r\nimport os\r\nfrom PIL import Image\r\nimport gzip\r\nimport dicom2nifti\r\nimport meshlib.mrmeshpy as mr\r\nimport meshlib.mrmeshnumpy as mrn\r\nimport cv2\r\nimport numpy as np\r\nimport scipy.signal\r\nimport torch\r\nimport monai\r\nfrom PyQt5 import uic, QtWidgets\r\nfrom PyQt5.QtWidgets import QFileDialog,QMainWindow\r\nfrom PyQt5.QtCore import Qt, QTimer\r\nfrom PyQt5 import QtGui\r\nfrom interfaz_acsi_choose import Ui_segunda_ventana\r\nfrom interfaz_acsi_info import Ui_tercera_ventana\r\nfrom PyQt5.uic import loadUi\r\n\r\nfrom datetime import datetime\r\nimport tempfile\r\nfrom glob import glob\r\nfrom torch.utils.data import random_split, DataLoader\r\nimport torchio as tio\r\nimport pytorch_lightning as pl\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\n\r\n# !pip install -q monai==1.1.0\r\n# !pip install -q torch==1.10.2 torchtext torchvision\r\n# !pip install -q torchio==0.18.73\r\n# !pip install -q pytorch-lightning==1.5.10\r\n# !pip install -q pandas==1.1.5 seaborn==0.11.1\r\n\r\n# Cargar la interfaz de usuario desde el archivo .ui\r\nqtCreatorFile = \"interfaz_acsi.ui\"\r\nUi_MainWindow, QtBaseClass = uic.loadUiType(qtCreatorFile)\r\n\r\nclass Model(pl.LightningModule):\r\n def __init__(self, net, criterion, learning_rate, optimizer_class):\r\n super().__init__()\r\n self.lr = learning_rate\r\n self.net = net\r\n self.criterion = criterion\r\n self.optimizer_class = optimizer_class\r\n \r\n def configure_optimizers(self):\r\n optimizer = self.optimizer_class(self.parameters(), lr=self.lr)\r\n return optimizer\r\n \r\n def prepare_batch(self, batch):\r\n return batch['image'][tio.DATA], batch['label'][tio.DATA]\r\n \r\n def infer_batch(self, batch):\r\n x, y = self.prepare_batch(batch)\r\n y_hat = self.net(x)\r\n return y_hat, y\r\n\r\n def training_step(self, batch, batch_idx):\r\n y_hat, y = self.infer_batch(batch)\r\n loss = self.criterion(y_hat, y)\r\n self.log('train_loss', loss, prog_bar=True)\r\n return loss\r\n \r\n def validation_step(self, batch, batch_idx):\r\n y_hat, y = self.infer_batch(batch)\r\n loss = self.criterion(y_hat, y)\r\n self.log('val_loss', loss)\r\n return loss\r\n \r\n def forward(self, x):\r\n # Pasar la entrada a través de la red\r\n return self.net(x)\r\n \r\nunet = monai.networks.nets.UNet(\r\n dimensions=3,\r\n in_channels=1,\r\n out_channels=3,\r\n channels=(8, 16, 32, 64),\r\n strides=(2, 2, 2),\r\n)\r\n\r\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\r\n\r\n# Define el modelo (asegúrate de que sea la misma arquitectura que se usó para entrenar)\r\nmodel = Model(\r\n net=unet,\r\n criterion=monai.losses.DiceCELoss(softmax=True),\r\n learning_rate=1e-2,\r\n optimizer_class=torch.optim.AdamW,\r\n)\r\n\r\n# Cargar los pesos del modelo\r\nmodel_path = 'model_weights.pth'\r\nmodel.load_state_dict(torch.load(model_path, map_location=torch.device('cpu')))\r\n\r\n# Poner el modelo en modo de evaluación\r\nmodel.eval()\r\n\r\n\r\nclass MyApp(QMainWindow, Ui_MainWindow):\r\n def __init__(self):\r\n super(MyApp, self).__init__()\r\n self.setupUi(self)\r\n self.setWindowTitle(\"Proyecto ACSI\")\r\n self.logo_upch.setPixmap(QtGui.QPixmap(\"upch_logo.png\"))\r\n self.logo_pucp.setPixmap(QtGui.QPixmap(\"pucp_logo.png\"))\r\n icon = QtGui.QIcon()\r\n icon.addPixmap(QtGui.QPixmap(\"upch_logo.png\"), QtGui.QIcon.Normal, QtGui.QIcon.Off)\r\n self.setFixedSize(705, 591)\r\n self.ventana_sec = Ventana_dos(parent=self)\r\n self.ventana_third = Ventana_tres(parent=self)\r\n # Conectar los botones a las funciones correspondientes\r\n self.load_file.activated.connect(self.agregar_carpeta)\r\n self.start_op.clicked.connect(self.inicio_tarea)\r\n self.save_file.clicked.connect(self.guardar_carpeta)\r\n self.timer = QTimer(self)\r\n self.timer.timeout.connect(self.play_next_image)\r\n \r\n self.next_both.clicked.connect(self.next_both_graphs)\r\n self.back_both.clicked.connect(self.back_both_graphs)\r\n self.stop_both.clicked.connect(self.stop_both_graphs)\r\n self.play_both.clicked.connect(self.play_both_graphs)\r\n self.choose.clicked.connect(self.choose_graph)\r\n self.button_info.clicked.connect(self.show_info)\r\n\r\n def show_info(self):\r\n try:\r\n self.ventana_third.show() \r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n\r\n def choose_graph(self):\r\n try:\r\n self.ventana_sec.show()\r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n\r\n def recibir_new_frame(self, new_i_frame):\r\n try:\r\n self.index_orig=new_i_frame\r\n self.graph_1.setPixmap(QtGui.QPixmap(str(self.rutas_frame_orig[new_i_frame])))\r\n self.graph_2.setPixmap(QtGui.QPixmap(str(self.rutas_frame_seg[new_i_frame])))\r\n self.t_frame.setText(str(new_i_frame))\r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n \r\n def play_both_graphs(self):\r\n if not self.rutas_frame_orig == [] and not self.rutas_frame_seg == []:\r\n self.index_orig = 0\r\n self.index_orig = 0\r\n self.timer.start(60) # Iniciar el QTimer para las imágenes originales\r\n\r\n def next_both_graphs(self):\r\n if not self.rutas_frame_orig == [] and not self.rutas_frame_seg == []:\r\n if self.index_orig < len(self.rutas_frame_orig) - 1:\r\n self.index_orig += 1\r\n self.graph_1.setPixmap(QtGui.QPixmap(str(self.rutas_frame_orig[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n if self.index_orig < len(self.rutas_frame_seg) - 1:\r\n self.index_orig += 1\r\n self.graph_2.setPixmap(QtGui.QPixmap(str(self.rutas_frame_seg[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n\r\n def back_both_graphs(self):\r\n if not self.rutas_frame_orig == [] and not self.rutas_frame_seg == []:\r\n if self.index_orig > 0:\r\n self.index_orig -= 1\r\n self.graph_1.setPixmap(QtGui.QPixmap(str(self.rutas_frame_orig[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n if self.index_orig > 0:\r\n self.index_orig -= 1\r\n self.graph_2.setPixmap(QtGui.QPixmap(str(self.rutas_frame_seg[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n\r\n\r\n def stop_both_graphs(self):\r\n try:\r\n self.timer.stop() # Detener el QTimer\r\n self.t_frame.setText(str(self.index_orig))\r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n\r\n def play_next_image(self):\r\n try:\r\n # Actualizar frames de la imagen original\r\n if self.index_orig < len(self.rutas_frame_orig):\r\n self.graph_1.setPixmap(QtGui.QPixmap(str(self.rutas_frame_orig[self.index_orig])))\r\n self.index_orig += 1\r\n else:\r\n self.index_orig = 0 # Reiniciar el índice si ha llegado al final\r\n\r\n # Actualizar frames de la imagen segmentada\r\n if self.index_orig < len(self.rutas_frame_seg):\r\n self.graph_2.setPixmap(QtGui.QPixmap(str(self.rutas_frame_seg[self.index_orig])))\r\n self.index_orig += 1\r\n else:\r\n self.index_orig = 0 # Reiniciar el índice si ha llegado al final\r\n\r\n self.t_frame.setText(str(self.index_orig)) # Actualizar el texto del frame\r\n\r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n\r\n \r\n def agregar_carpeta(self):\r\n try:\r\n texto=self.load_file.currentText()\r\n escritorio = os.path.expanduser(\"~/Desktop\")\r\n self.contenido_carpetas={}\r\n \r\n if texto == \"Cargar NIfTI\":\r\n archivo_nii_gz, _ = QFileDialog.getOpenFileName(self, \"Seleccionar Archivo NIfTI\", escritorio, filter=\"Archivos NIfTI (*.nii.gz)\")\r\n if not archivo_nii_gz:\r\n return # El usuario canceló la selección o no eligió un archivo NIfTI\r\n\r\n self.ruta_completa = archivo_nii_gz\r\n self.ruta_salida_nift = os.path.dirname(archivo_nii_gz)\r\n print(f\"Archivo seleccionado: {archivo_nii_gz}\")\r\n self.ruta_completa = archivo_nii_gz\r\n imagen_nii = nib.load(self.ruta_completa)\r\n self.datos = imagen_nii.get_fdata()\r\n print(self.datos.shape)\r\n self.flag_zip = 0\r\n self.mostrar_frames_graph_1(self.datos, self.flag_zip)\r\n\r\n # elif texto == \"Cargar archivo ZIP\":\r\n # archivo_zip, _ = QFileDialog.getOpenFileName(self, \"Seleccionar Archivo ZIP\", escritorio, filter=\"Archivos ZIP (*.zip)\")\r\n # self.flag_zip=1\r\n # if not archivo_zip:\r\n # return # El usuario canceló la selección o no eligió un archivo ZIP\r\n # # Abre el archivo ZIP en modo de lectura\r\n # self.directorio_destino = os.path.dirname(archivo_zip)\r\n # self.nombre_del_archivo = os.path.basename(archivo_zip)\r\n # new_archivo_zip, extension = os.path.splitext(archivo_zip)\r\n # if os.path.exists(new_archivo_zip):\r\n # # Si la carpeta existe, elimínala junto con su contenido\r\n # shutil.rmtree(new_archivo_zip)\r\n # os.mkdir(new_archivo_zip)\r\n # self.nombre_aux=os.path.basename(new_archivo_zip)\r\n # print(new_archivo_zip)\r\n # with zipfile.ZipFile(archivo_zip, 'r') as zip_ref:\r\n # zip_ref.extractall(new_archivo_zip)\r\n # print(f\"Se seleccionó una carpeta: {new_archivo_zip}\")\r\n # lista_de_archivos = zip_ref.namelist()\r\n # ruta_carpeta = os.path.join(new_archivo_zip, lista_de_archivos[0])\r\n # ruta_carpeta = ruta_carpeta.replace('\\\\', '/')\r\n # ruta_carpeta = ruta_carpeta[:-1]\r\n # print(f\"ruta:{ruta_carpeta}\")\r\n # self.nombre_aux=os.path.basename(ruta_carpeta)\r\n # print(f\"nombre_aux:{self.nombre_aux}\")\r\n # lista_de_archivos.pop(0)\r\n # print(f\"nombre: {lista_de_archivos}\")\r\n # self.ruta_salida_nift_1 = os.path.dirname(ruta_carpeta)\r\n # print(f\"ruta_salida_nift_1: {self.ruta_salida_nift_1}\") \r\n # self.ruta_salida_nift = os.path.join(self.ruta_salida_nift_1, str(self.nombre_aux)+str(\"_nii_gz\"))\r\n # self.ruta_salida_nift=self.ruta_salida_nift.replace('\\\\', '/')\r\n # print(self.ruta_salida_nift)\r\n # os.makedirs(self.ruta_salida_nift, exist_ok=True)\r\n # dicom2nifti.convert_directory(ruta_carpeta, self.ruta_salida_nift, compression=True, reorient=True)\r\n\r\n # nombre_archivo_generado = os.listdir(self.ruta_salida_nift)[0] # Suponiendo que solo haya un archivo\r\n # # Renombra el archivo generado a tu nombre personalizado\r\n # nombre_personalizado = str(self.nombre_aux)+str(\"_nii_gz.nii.gz\")\r\n # print(f\"nombre_f:{nombre_personalizado}\")\r\n # nueva_ruta_archivo = os.path.join(self.ruta_salida_nift, nombre_personalizado)\r\n # os.rename(os.path.join(self.ruta_salida_nift, nombre_archivo_generado), nueva_ruta_archivo)\r\n \r\n # archivo_ni_gz = os.listdir(self.ruta_salida_nift)\r\n # ruta_completa = os.path.join(self.ruta_salida_nift, archivo_ni_gz[0])\r\n # self.ruta_completa = ruta_completa.replace('\\\\', '/')\r\n # print(f\"Ruta_nift:{self.ruta_completa}\")\r\n # # Cargar el archivo NIfTI\r\n # imagen_nii = nib.load(self.ruta_completa)\r\n # # Acceder a los datos de la imagen (esto puede variar según tus necesidades)\r\n # self.datos = imagen_nii.get_fdata()\r\n # print(self.datos.shape)\r\n # self.mostrar_frames_graph_1(self.datos,self.flag_zip)\r\n\r\n elif texto == \"Cargar DICOM\":\r\n archivo_tar, _ = QFileDialog.getOpenFileName(self, \"Seleccionar Archivo TAR\", escritorio, filter=\"Archivos TAR (*.tar.gz *.tar)\")\r\n self.flag_zip=1\r\n if not archivo_tar:\r\n return # El usuario canceló la selección o no eligió un archivo ZIP\r\n # Abre el archivo TAR en modo de lectura\r\n self.directorio_destino = os.path.dirname(archivo_tar)\r\n self.nombre_del_archivo = os.path.basename(archivo_tar) \r\n archivo_tar, extension = os.path.splitext(archivo_tar)\r\n if os.path.exists(archivo_tar):\r\n # Si la carpeta existe, elimínala junto con su contenido\r\n shutil.rmtree(archivo_tar)\r\n os.mkdir(archivo_tar)\r\n self.nombre_aux=os.path.basename(archivo_tar)\r\n new_archivo_tar=os.path.join(self.directorio_destino, self.nombre_del_archivo)\r\n new_archivo_tar = new_archivo_tar.replace('\\\\', '/')\r\n \r\n shutil.unpack_archive(new_archivo_tar, extract_dir=archivo_tar)\r\n with tarfile.open(new_archivo_tar, 'r') as tar:\r\n # Obtiene la lista de nombres de archivos en el archivo tar\r\n lista_de_archivos = tar.getnames()\r\n print(f\"Se seleccionó una carpeta: {new_archivo_tar}\")\r\n ruta_carpeta = os.path.join(archivo_tar, lista_de_archivos[0])\r\n ruta_carpeta = ruta_carpeta.replace('\\\\', '/')\r\n print(f\"Ruta: {ruta_carpeta}\")\r\n lista_de_archivos.pop(0)\r\n print(f\"nombre: {lista_de_archivos}\")\r\n self.nombre_aux=os.path.basename(ruta_carpeta)\r\n self.ruta_salida_nift_1 = os.path.dirname(ruta_carpeta)\r\n self.ruta_salida_nift = os.path.join(self.ruta_salida_nift_1, str(self.nombre_aux)+str(\"_nii_gz\"))\r\n self.ruta_salida_nift=self.ruta_salida_nift.replace('\\\\', '/')\r\n print(self.ruta_salida_nift)\r\n os.makedirs(self.ruta_salida_nift, exist_ok=True)\r\n dicom2nifti.convert_directory(ruta_carpeta, self.ruta_salida_nift, compression=True, reorient=True)\r\n\r\n nombre_archivo_generado = os.listdir(self.ruta_salida_nift)[0] # Suponiendo que solo haya un archivo\r\n # Renombra el archivo generado a tu nombre personalizado\r\n nombre_personalizado = str(self.nombre_aux)+str(\"_nii_gz.nii.gz\")\r\n nueva_ruta_archivo = os.path.join(self.ruta_salida_nift, nombre_personalizado)\r\n os.rename(os.path.join(self.ruta_salida_nift, nombre_archivo_generado), nueva_ruta_archivo)\r\n\r\n archivo_ni_gz = os.listdir(self.ruta_salida_nift)\r\n ruta_completa = os.path.join(self.ruta_salida_nift, archivo_ni_gz[0])\r\n self.ruta_completa = ruta_completa.replace('\\\\', '/')\r\n print(f\"Ruta_nift:{self.ruta_completa}\")\r\n # Cargar el archivo NIfTI\r\n imagen_nii = nib.load(self.ruta_completa)\r\n # Acceder a los datos de la imagen (esto puede variar según tus necesidades)\r\n self.datos = imagen_nii.get_fdata()\r\n print(self.datos.shape)\r\n self.mostrar_frames_graph_1(self.datos,self.flag_zip)\r\n except Exception as e:\r\n print(f\"Error al mostrar las imágenes: {e}\") \r\n \r\n def mostrar_frames_graph_1(self, contenido_carpetas, flag_zip):\r\n try:\r\n ########Colocar el procesamiento para obtener frames en 2D########\r\n self.images_orig=contenido_carpetas\r\n # Crear una lista para almacenar las rutas\r\n self.rutas_frame_orig = []\r\n self.index_orig=0\r\n self.ruta_slices = os.path.join(os.path.dirname(self.ruta_salida_nift), \"aux_slices\")\r\n self.ruta_slices = self.ruta_slices.replace('\\\\', '/')\r\n print(f\"dir_ruta_slices:{self.ruta_slices}\")\r\n os.makedirs(self.ruta_slices, exist_ok=True)\r\n for i, imagen_2d in enumerate(contenido_carpetas[1,1,:]):\r\n # Convierte la imagen 2D en formato CV_8U (8 bits sin signo) para OpenCV\r\n imagen_2d = np.uint8(contenido_carpetas[:,:,i])\r\n # Define el nombre del archivo de imagen\r\n nombre_archivo = f\"imagen_{i}.png\"\r\n # Ruta completa del archivo de imagen\r\n ruta_archivo = os.path.join(self.ruta_slices, nombre_archivo)\r\n # Guarda la imagen como archivo\r\n cv2.imwrite(ruta_archivo, imagen_2d)\r\n # Agrega la ruta del archivo a la lista\r\n self.rutas_frame_orig.append(ruta_archivo)\r\n print(i)\r\n print(f\"Se han guardado {len(self.rutas_frame_orig)} imágenes en {self.ruta_slices}\")\r\n self.graph_1.setPixmap(QtGui.QPixmap(str(self.rutas_frame_orig[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n except Exception as e:\r\n print(f\"Error al mostrar las imágenes: {e}\")\r\n\r\n def inicio_tarea(self):\r\n try:\r\n print(\"Iniciando segmentación...\")\r\n self.segmentacion()\r\n self.mostrar_frames_graph_2()\r\n except Exception as e:\r\n print(f\"Error en inicio_tarea: {e}\")\r\n\r\n \r\n def get_preprocessing_transform(self):\r\n return tio.Compose([\r\n tio.RescaleIntensity((-1, 1)), # Reemplaza target_size con el tamaño deseado\r\n tio.EnsureShapeMultiple(8), # Para U-Net\r\n tio.OneHot(),\r\n ])\r\n def convertir_a_stl(self, mask_array, predictions_folder):\r\n # Debes asegurarte de que 'mask_array' es una matriz 3D que representa tu segmentación\r\n # Si 'mask_array' no es 3D, necesitarás ajustarlo\r\n\r\n # Convertir la matriz 3D a SimpleVolume\r\n simpleVolume = mrn.simpleVolumeFrom3Darray(mask_array.astype(float))\r\n\r\n # Convertir SimpleVolume a FloatGrid\r\n floatGrid = mr.simpleVolumeToDenseGrid(simpleVolume)\r\n\r\n # Crear malla usando un iso-valor y el tamaño de voxel\r\n mesh = mr.gridToMesh(floatGrid, mr.Vector3f(0.1, 0.1, 0.1), 0.5)\r\n\r\n # Guardar la malla como STL\r\n stl_path = os.path.join(predictions_folder, 'mesh.stl')\r\n mr.saveMesh(mesh, stl_path)\r\n \r\n def segmentacion(self):\r\n try:\r\n subject = tio.Subject(image=tio.ScalarImage(self.ruta_completa))\r\n preprocess = self.get_preprocessing_transform()\r\n preprocessed_subject = preprocess(subject)\r\n\r\n predictions_folder = 'predictions'\r\n if not os.path.exists(predictions_folder):\r\n os.makedirs(predictions_folder)\r\n\r\n with torch.no_grad():\r\n input_tensor = preprocessed_subject['image'][tio.DATA][None]\r\n input_tensor = input_tensor.to(device)\r\n output = model(input_tensor)\r\n output_label = output.argmax(dim=1, keepdim=True).cpu()\r\n output_label_squeezed = output_label.squeeze(0)\r\n\r\n mask_array = output_label.numpy().squeeze()\r\n mask_array = (mask_array * 255).astype(np.uint8)\r\n\r\n # Guardar cada corte del volumen como PNG y la lista de rutas\r\n self.mascaras_obtenidas = []\r\n for i, slice in enumerate(mask_array):\r\n slice_image = Image.fromarray(slice)\r\n slice_path = os.path.join(predictions_folder, f'mask_slice_{i}.png')\r\n slice_image.save(slice_path)\r\n self.mascaras_obtenidas.append(slice_path)\r\n\r\n # Guardar la segmentación completa como archivo NIfTI\r\n nifti_path = os.path.join(predictions_folder, 'segmentacion_completa.nii.gz')\r\n segmented_image = tio.ScalarImage(tensor=output_label_squeezed)\r\n segmented_image.save(nifti_path)\r\n\r\n # Calcular el volumen\r\n # Asumiendo que cada voxel es 1x1x1 mm^3, pero verifica esto con tus datos\r\n volumen = np.sum(segmented_image.numpy() > 0) # Cuenta los voxels en la ROI\r\n self.t_vol_h.setText(str(volumen)+\"mm3\") # Muestra el volumen calculado\r\n self.convertir_a_stl(mask_array, predictions_folder)\r\n\r\n print(\"Segmentación completada y máscaras guardadas.\")\r\n except Exception as e:\r\n print(f\"Error en la segmentación: {e}\")\r\n\r\n \r\n def mostrar_frames_graph_2(self):\r\n try:\r\n self.rutas_frame_seg = self.mascaras_obtenidas\r\n self.index_orig = 0\r\n\r\n if self.rutas_frame_seg:\r\n self.graph_2.setPixmap(QtGui.QPixmap(str(self.rutas_frame_seg[self.index_orig])))\r\n self.t_frame.setText(str(self.index_orig))\r\n else:\r\n print(\"No hay máscaras para mostrar.\")\r\n except Exception as e:\r\n print(f\"Error al mostrar las imágenes segmentadas: {e}\")\r\n\r\n\r\n def guardar_carpeta(self):\r\n options = QFileDialog.Options()\r\n options |= QFileDialog.DontUseNativeDialog\r\n escritorio = os.path.expanduser(\"~/Desktop\")\r\n carpeta_seleccionada = QFileDialog.getExistingDirectory(self, \"Guardar Carpeta...\", escritorio, options=options)\r\n\r\n if carpeta_seleccionada:\r\n try:\r\n carpeta_origen = self.ruta_nueva\r\n carpeta_destino = os.path.join(carpeta_seleccionada, self.nombre_aux+\"_result\")\r\n\r\n if os.path.exists(carpeta_destino):\r\n contador = 1\r\n while True:\r\n nuevo_nombre = f\"{self.nombre_aux} ({contador})\"+\"_result\"\r\n carpeta_destino = os.path.join(carpeta_seleccionada, nuevo_nombre)\r\n if not os.path.exists(carpeta_destino):\r\n break\r\n contador += 1\r\n\r\n shutil.copytree(carpeta_origen, carpeta_destino)\r\n print(\"Carpeta de destino:\", carpeta_destino)\r\n print(\"Guardado con éxito\")\r\n except FileNotFoundError as e:\r\n print(\"Error: La carpeta de origen no se encuentra.\")\r\n except PermissionError as e:\r\n print(\"Error: No tienes permisos para copiar la carpeta.\")\r\n except Exception as e:\r\n print(f\"Error al copiar la carpeta: {e}\")\r\n\r\n def closeEvent(self, event):\r\n confirmacion = QtWidgets.QMessageBox.question(self, \"Confirmar Salida\", \"¿Estás seguro de que deseas salir?\",\r\n QtWidgets.QMessageBox.Yes | QtWidgets.QMessageBox.No)\r\n if confirmacion == QtWidgets.QMessageBox.Yes:\r\n event.accept()\r\n else:\r\n event.ignore()\r\n\r\nclass Ventana_dos(QMainWindow):\r\n def __init__(self, parent=None):\r\n super(Ventana_dos, self).__init__(parent)\r\n loadUi(\"interfaz_acsi_choose.ui\", self)\r\n self.acept_f.clicked.connect(self.close_segunda_ventana)\r\n self.setWindowTitle(\"Proyecto ACSI\")\r\n self.setFixedSize(234, 84)\r\n\r\n def close_segunda_ventana(self):\r\n try:\r\n new_i_frame = int(self.c_frame.toPlainText())\r\n print(\"Valor de c_frame:\", new_i_frame)\r\n self.parent().recibir_new_frame(new_i_frame)\r\n self.close()\r\n except Exception as e:\r\n print(f\"Error: {e}\")\r\n\r\nclass Ventana_tres(QMainWindow):\r\n def __init__(self, parent=None):\r\n super(Ventana_tres, self).__init__(parent)\r\n loadUi(\"interfaz_acsi_info.ui\", self)\r\n self.setWindowTitle(\"Proyecto ACSI\")\r\n self.setFixedSize(527, 306)\r\n\r\nif __name__ == \"__main__\":\r\n app = QtWidgets.QApplication(sys.argv)\r\n window = MyApp()\r\n window.show()\r\n sys.exit(app.exec_())\r\n","repo_name":"cpariona/Hippocampus-3D-Segmentation","sub_path":"codigo_fuente/code_function_acsi.py","file_name":"code_function_acsi.py","file_ext":"py","file_size_in_byte":24945,"program_lang":"python","lang":"es","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"10285886636","text":"# BOJ 10989 Sort 3 -Counting Sort\nimport sys\ninput = sys.stdin.readline\n\nn = int(input())\ndic = dict()\nfor _ in range(n):\n x = int(input())\n if x in dic:\n dic[x] += 1\n else :\n dic[x] = 1\nfor key,count in sorted(dic.items()):\n for _ in range(count):\n print(key)","repo_name":"Qud4300/Baekjoon_Online_Judge","sub_path":"10989 수 정렬하기 3/sort3.py","file_name":"sort3.py","file_ext":"py","file_size_in_byte":293,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"31900442828","text":"#your usual imports and then this code\r\n\r\n@client.command()\r\nasync def user(ctx, member:discord.Member=None):\r\n if member is None:\r\n member = ctx.message.author\r\n pronoun = \"Your\"\r\n else:\r\n pronoun = str(member)\r\n name = f\"{member.name}#{member.discriminator}\"\r\n status = ctx.author.status.name\r\n created_on = member.created_at.__format__('%A, %d. %B %Y @ %H:%M:%S')\r\n userAvatarUrl = member.avatar_url\r\n join = member.joined_at.__format__('%A, %d. %B %Y @ %H:%M:%S')\r\n statoos = member.activity\r\n house = member.top_role\r\n permissions = member.permissions_in(ctx.message.channel)\r\n #userhighest role\r\n await ctx.send(\"``` ```\")\r\n await ctx.send(f\"\"\"**`Here's Some Dirt On:` {member.mention}!:\r\nUsername is: `{str(member.name)}`\r\nUserTag is: `{str(member.discriminator)}`\r\nUser ID is: `{str(member.id)}`\r\nUser Presence is: `{str(status)}`\r\nUser Is Playing: `{str(statoos)}`\r\nUser Highest Role: `{str(house)}`\r\nUser Created On: `{str(created_on)}`\r\nUser Joined On: `{str(join)}`\r\nUser Permissions: `{str(permissions)}`**\"\"\")\r\n await ctx.send(userAvatarUrl)\r\n await ctx.send(\"``` ```\")","repo_name":"Javi-beep/discord-bot-commands","sub_path":"info kinda commands/userstats.py","file_name":"userstats.py","file_ext":"py","file_size_in_byte":1149,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"10599408914","text":"import sys\nimport heapq as hq\ninput = sys.stdin.readline\n\ndef dijkstra(start):\n q = []\n distance = [INF] * (n+1)\n \n hq.heappush(q, (0, start))\n distance[start] = 0\n while q:\n dist, now = hq.heappop(q)\n if distance[now] < dist: continue # 이미 갱신됨\n for cost, next in graph[now]:\n tmp = dist + cost\n if distance[next] > tmp:\n distance[next] = tmp\n hq.heappush(q, (tmp, next))\n \n return distance\n\ndef calculate_infection():\n count, time = 0, 0\n distance = dijkstra(c)\n for d in distance:\n if d < INF:\n count +=1\n time = max(time, d)\n return count, time \n\nif __name__ == \"__main__\":\n INF = int(1e9)\n t = int(input())\n for _ in range(t):\n n, d, c = map(int, input().split())\n\n # 의존성\n graph = [[] for _ in range(n+1)]\n for _ in range(d):\n a, b, s = map(int, input().split())\n graph[b].append((s, a))\n \n print(\" \".join(map(str, calculate_infection())))\n ","repo_name":"Altu-Bitu-2/Altu-Bitu-Algorithms-Study","sub_path":"[최단 경로] 5월 3일/B10282.py","file_name":"B10282.py","file_ext":"py","file_size_in_byte":1090,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"14642763966","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('place', '0006_auto_20171231_1554'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Hotel',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('country_id', models.IntegerField(default=-1, verbose_name='\\u6240\\u5c5e\\u56fd\\u5bb6id')),\n ('province_id', models.IntegerField(default=-1, verbose_name='\\u6240\\u5c5e\\u7701\\u4efdid')),\n ('city_id', models.IntegerField(default=-1, verbose_name='\\u6240\\u5c5e\\u57ce\\u5e02id')),\n ('district_id', models.IntegerField(default=-1, verbose_name='\\u6240\\u5c5e\\u884c\\u653f\\u533aid')),\n ('name', models.CharField(default=b'', max_length=200, verbose_name='\\u9152\\u5e97\\u540d\\u79f0')),\n ('en_name', models.CharField(default=b'', max_length=200, verbose_name='\\u9152\\u5e97\\u82f1\\u6587\\u540d\\u79f0')),\n ('addr', models.CharField(default=b'', max_length=200, verbose_name='\\u9152\\u5e97\\u5730\\u5740')),\n ('price', models.DecimalField(default=0.0, max_digits=16, decimal_places=8)),\n ('lng', models.DecimalField(default=0.0, max_digits=16, decimal_places=8)),\n ('lat', models.DecimalField(default=0.0, max_digits=16, decimal_places=8)),\n ('cmt_num', models.IntegerField(default=-1, verbose_name='\\u8bc4\\u8bba\\u6570\\u76ee')),\n ('pts', models.DecimalField(default=0.0, verbose_name='\\u8bc4\\u5206', max_digits=16, decimal_places=8)),\n ('pts_level', models.CharField(default=b'', max_length=200, verbose_name='\\u8bc4\\u5206\\u7b49\\u7ea7')),\n ('note_num', models.IntegerField(default=-1, verbose_name='\\u6e38\\u8bb0\\u6570\\u76ee')),\n ('desc', models.CharField(default=b'', max_length=200, verbose_name='\\u63cf\\u8ff0')),\n ('main_pic', models.CharField(default=b'', max_length=200, verbose_name='\\u9152\\u5e97\\u4e3b\\u56fe')),\n ('url', models.CharField(default=b'', max_length=200, null=True, verbose_name='url\\u94fe\\u63a5')),\n ('status', models.IntegerField(default=-1, verbose_name='\\u9152\\u5e97\\u72b6\\u6001')),\n ('created', models.DateTimeField(auto_now_add=True, null=True)),\n ('updated', models.DateTimeField(auto_now=True, null=True)),\n ],\n ),\n ]\n","repo_name":"Cribbee/gofree130","sub_path":"place/migrations/0007_hotel.py","file_name":"0007_hotel.py","file_ext":"py","file_size_in_byte":2581,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"14493584633","text":"#!/usr/bin/env python3\r\n\r\nimport socket\r\nimport os, sys, json\r\nimport signal\r\nfrom subprocess import Popen, PIPE\r\n\r\nPORT = 65432 # Port to listen on (non-privileged ports are > 1023)\r\nHOST = '127.0.0.1' # Standard loopback interface address (localhost)\r\nHOST = socket.gethostname()\r\n\r\n#To run and survive closing terminal: nohup python socket-server.py\r\n#To run periodically - chrontab -e: \r\n# 00 * * * * python /var/www/html/projects/asly/socket-server.py\r\n\r\n\r\ndef kill_port_proc(port_num):\r\n process = Popen([\"lsof\", \"-i\", \":{0}\".format(port_num)], stdout=PIPE, stderr=PIPE)\r\n stdout, stderr = process.communicate()\r\n for process in str(stdout.decode(\"utf-8\")).split(\"\\n\")[1:]: \r\n data = [x for x in process.split(\" \") if x != '']\r\n if (len(data) <= 1):\r\n continue\r\n\r\n os.kill(int(data[1]), signal.SIGKILL)\r\n print(\"Killed process, please run the script again\")\r\n\r\n#Now for the application specific part, functions, libraries, and data to load\r\n#Application specific libraries\r\n# sys.path.append(\"/var/www/html/code_utils\")\r\n# from parsing_lib import *\r\n\r\n#shelve_fpath=\"parsing/verbs.shelve\"\r\n\r\n#identify the data processing function for each application\r\ndef process_data(data0):\r\n data0_decoded=data.decode(\"utf-8\")\r\n data0_decoded='Test: %s'%data0_decoded\r\n output0=data0_decoded.encode(\"utf-8\")\r\n return output0\r\n\r\n\r\n\r\ntry:\r\n while True:\r\n with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:\r\n s.bind((HOST, PORT))\r\n s.listen()\r\n conn, addr = s.accept()\r\n with conn:\r\n print('Connected by', addr)\r\n while True:\r\n data = conn.recv(1024)\r\n \r\n output=process_data(data)\r\n if not data:\r\n break\r\n #conn.sendall(data)\r\n conn.sendall(output)\r\nexcept: #if the port is already used and we want to restart it \r\n pass\r\n kill_port_proc(PORT)\r\n print(\"Killed the port process, run the script again\")\r\n\r\n","repo_name":"hmghaly/word_align","sub_path":"socket-server.py","file_name":"socket-server.py","file_ext":"py","file_size_in_byte":2086,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"71807567584","text":"\"\"\"\n\n Extract the data contained in a buffer and save to file\n\n Usefull to create a input file for afl-unicorn\n\n 2018 Giulio Ginesi\n\n\"\"\"\n\nimport gdb\nimport zlib\nimport datetime\nimport time\n\nBUFFER_ADDR = 0x76c6a008\nBUFFER_LEN = 0x200ee\n\ntry:\n buffer = read_memory(BUFFER_ADDR, BUFFER_LEN)\n print(\"Dumping buffer from 0x{0:x} of length 0x{1:x}\".format(BUFFER_ADDR, BUFFER_LEN))\nexcept:\n print(\"Error reading memory region!\")\n\n#compressed_buffer = zlib.compress(buffer)\n\ntimestamp = datetime.datetime.fromtimestamp(time.time()).strftime('%Y%m%d_%H%M%S')\ntry:\n out_name = 'input_'+timestamp+'.bin'\n out_file = open(out_name, 'wb')\n out_file.write(buffer)\n out_file.close()\n print(\"Saved file with name {}\".format(out_name))\nexcept:\n print(\"Error saving file!\")\n","repo_name":"blazef104/avionics-fuzzing","sub_path":"fuzzing/afl-unicorn/tools/extract_from_memory.py","file_name":"extract_from_memory.py","file_ext":"py","file_size_in_byte":795,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"7"} +{"seq_id":"27858809526","text":"import scrapy\nimport logging\nimport pytz\n\nfrom datetime import datetime\nfrom salescanner.crawling.spiders.utils.utils import Utils\nfrom salescanner.crawling.items import SalescannerItem\nfrom salescanner.crawling.spiders.utils.spider_indexor import SpiderIndexor\n\n\n@SpiderIndexor('olx')\nclass OLXAdsSpider(scrapy.Spider):\n\n MAX_NUMBER_OF_PAGES = 25\n name = 'olx_sales'\n\n def __init__(self, **kwargs):\n self.allowed_domains = ['olx.bg']\n self.start_urls = ['https://www.olx.bg/ads/']\n self.pages_processed = 0\n\n super().__init__(**kwargs)\n logging.getLogger('scrapy').setLevel(logging.WARNING)\n\n def parse(self, response):\n print(f'OLX LIST PAGE: {response.url}')\n offers_response = response.css('.offers')[1].css('.detailsLinkPromoted::attr(href), .detailsLink::attr(href)')\n offers_urls = set(offers_response.getall())\n\n for offer_url in offers_urls:\n split_url = offer_url.split('/')\n if 'job' in split_url or 'ad' not in split_url:\n continue\n\n if 'd' in split_url:\n split_url.remove('d')\n offer_url = '/'.join(split_url)\n\n yield scrapy.Request(offer_url, callback=self.parse_details_page)\n self.pages_processed += 1\n\n next_page_url = response.css('.next > a.pageNextPrev::attr(href)').get()\n if next_page_url is not None and self.pages_processed < OLXAdsSpider.MAX_NUMBER_OF_PAGES:\n yield scrapy.Request(next_page_url, callback=self.parse, dont_filter=True)\n \n\n def parse_details_page(self, response):\n split_url = response.url.split('/')\n if 'job' in split_url or 'ad' not in split_url:\n return\n\n image_url = response.css('.descgallery__image img.bigImage::attr(src)').get()\n title = response.css('.offer-titlebox > h1::text').get()\n price = response.css('.offer-titlebox__price > .pricelabel > strong::text').get()\n description = response.css('.descriptioncontent > #textContent *::text').getall()\n description = ' '.join([line.strip() for line in description])\n upload_datetime = response.css('.offer-bottombar__items .offer-bottombar__item em strong::text').get()\n \n ad_item = SalescannerItem()\n ad_item['url'] = response.url\n ad_item['title'] = title.strip() if title else title\n ad_item['price'] = price.strip() if price else price\n ad_item['image_url'] = image_url\n ad_item['description'] = description\n ad_item['upload_time'] = self.parse_upload_datetime(upload_datetime)\n yield ad_item\n\n def parse_upload_datetime(self, datetime_str):\n if datetime_str is None:\n return None\n\n datetime_str = datetime_str.strip()\n datetime_str = datetime_str[2:].split(',')\n time_portion = datetime_str[0].split(':')\n date_portion = datetime_str[1].strip().split(' ')\n\n datetime_obj = datetime(\n int(date_portion[2]),\n Utils.month_to_number(date_portion[1]),\n int(date_portion[0]),\n int(time_portion[0]),\n int(time_portion[1]))\n return self._datetime_to_timestamp(datetime_obj)\n \n def _datetime_to_timestamp(self, datetime_obj):\n timezone = pytz.timezone('Europe/Sofia')\n datetime_obj = timezone.localize(datetime_obj)\n\n return int(datetime_obj.timestamp() * 1000)\n","repo_name":"ZdravkoHvarlingov/sale-scanner","sub_path":"salescanner/crawling/spiders/olx_ads_spider.py","file_name":"olx_ads_spider.py","file_ext":"py","file_size_in_byte":3439,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"30420574537","text":"from __future__ import annotations\n\nimport asyncio\nimport hikari\nimport lightbulb\nimport logging\nimport miru\nimport config\nimport solis.helper.riddle\nfrom solis.views.view import ButtonView\nfrom __init__ import __version__\nfrom aiohttp import ClientSession\nfrom pytz import utc\nfrom apscheduler.schedulers.asyncio import AsyncIOScheduler\n\nis_playing_riddle = False\n\n\ndef setup() -> None:\n logging.info(\"Running bot setup...\")\n\n\nbot = lightbulb.BotApp(\n token=config.token,\n default_enabled_guilds=config.DEFAULT_GUILD_ID,\n owner_ids=config.OWNERS_ID,\n help_slash_command=True,\n case_insensitive_prefix_commands=True,\n prefix=\"!\",\n intents=hikari.Intents.ALL\n\n)\nbot.d.scheduler = AsyncIOScheduler()\nbot.d.scheduler.configure(timezome=utc)\nbot.load_extensions_from(\"../solis/extensions\")\nmiru.load(bot)\n\n\n@bot.listen(hikari.StartingEvent)\nasync def on_starting(event: hikari.StartingEvent) -> None:\n bot.d.scheduler.start()\n bot.d.session = ClientSession(trust_env=True)\n logging.info(\"AIOHTTP session started\")\n\n\n@bot.listen(hikari.StartedEvent)\nasync def on_started(event: hikari.StartedEvent) -> None:\n await bot.rest.create_message(\n config.TEST_CHANNEL_ID,\n f\"Solis is now online! (Version {__version__})\"\n )\n\n\n@bot.listen(hikari.StoppingEvent)\nasync def on_stopping(event: hikari.StoppingEvent) -> None:\n await bot.d.session.close()\n logging.info(\"AIOHTTP session closed\")\n bot.d.scheduler.shutdown()\n await bot.rest.create_message(\n config.TEST_CHANNEL_ID,\n f\"Solis is shutting down. (Version {__version__})\"\n )\n\n\n@bot.listen(hikari.DMMessageCreateEvent)\nasync def on_pm_message_create(event: hikari.DMMessageCreateEvent) -> None:\n if event.message.author.is_bot:\n return\n\n await event.message.respond(\n f\"You need to DM <@{config.BOT_OWNER}> to send a message to moderators.\"\n )\n\n\n@bot.command\n@lightbulb.command('ping', 'say pong!')\n@lightbulb.implements(lightbulb.SlashCommand)\nasync def ping(ctx):\n await ctx.respond(\n f\"Pong! DWSP latency: {ctx.bot.heartbeat_latency * 1_000:,.0f} ms.\")\n\n\n@bot.command\n@lightbulb.command('button', 'button test')\n@lightbulb.implements(lightbulb.SlashCommand)\nasync def button(ctx: lightbulb.SlashContext):\n view = ButtonView(timeout=60)\n message = await ctx.respond(\"Button test\", components=view.build())\n message = await message\n view.start(message)\n await view.wait()\n print(\"All done.\")\n\n\ndef run() -> None:\n setup()\n bot.run(\n activity=hikari.Activity(\n name=f\"/help | Version {__version__}\",\n type=hikari.ActivityType.WATCHING\n )\n )\n\n\n# Bot is listening to the specified guilds\n@bot.listen(hikari.GuildMessageCreateEvent)\nasync def on_guild_message_event(event: hikari.GuildMessageCreateEvent) -> None:\n event_author = event.message.author\n message = event.message\n\n global is_playing_riddle\n if event_author.is_bot or not message.content:\n return\n\n if any(e in message.content.lower().split(\" \") or i in message.content.lower()\n for e in {\"gm\", \"gm!\", \"goodmorning\"}\n for i in {\"good morning\", \"morning everyone\"}):\n await message.respond(\"Good morning! \" + event_author.mention,\n user_mentions=True,\n mentions_reply=True)\n\n if any(e in message.content.lower().split(\" \") or i in message.content.lower()\n for e in {\"gn\", \"gn!\", \"goodnight\"}\n for i in {\"good night\", \"night everyone\"}):\n await message.respond(\"Good night! \" + event_author.mention,\n user_mentions=True,\n mentions_reply=True)\n\n if any(e in message.content.lower().split(\" \") for e in\n {\"riddle\", \"teaser\", \"riddles\", \"teasers\"}) and is_playing_riddle is False \\\n and message.content != \"!fun riddle\" and message.content != \"/fun riddle\":\n resp = await message.respond(\"Did someone mention a riddle? :eyes: \" + event_author.mention,\n user_mentions=True,\n mentions_reply=True)\n await resp.add_reaction(\"❌\")\n await resp.add_reaction(\"✅\")\n is_playing_riddle = True\n try:\n reaction = await bot.wait_for(\n hikari.ReactionAddEvent,\n timeout=10,\n predicate=lambda new_event:\n isinstance(new_event, hikari.ReactionEvent)\n and new_event.user_id == event_author.id\n and str(new_event.emoji_name) in {\"❌\", \"✅\"}\n )\n if reaction.emoji_name == \"✅\":\n await resp.remove_reaction(emoji=\"❌\")\n await solis.helper.riddle.on_riddle(message)\n is_playing_riddle = False\n else:\n await resp.remove_reaction(emoji=\"✅\")\n await resp.edit(\"Okay maybe next time!\")\n is_playing_riddle = False\n\n except asyncio.TimeoutError:\n is_playing_riddle = False\n await message.respond(\"The riddle timed out :c\")\n\n\n@bot.listen(lightbulb.CommandErrorEvent)\nasync def on_command_error(event: lightbulb.CommandErrorEvent) -> None:\n exc = getattr(event.exception, \"__cause__\", event.exception)\n\n if isinstance(exc, lightbulb.NotOwner):\n await event.context.respond(\"You need to be an owner to do that.\")\n return\n raise event.exception\n\n\n","repo_name":"Memo-Aldu/solis","sub_path":"solis/bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":5498,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"20412268303","text":"\"\"\"\nAssumptions:\n None\n\nAlgorithm:\n 1. divisor = 10^(number of digits - 1)\n 2. While n != 0\n a. first = (int) n / divisor\n b. last = (int) n % 10\n c. if head != tail:\n return False\n d. remove first and last element from n\n e. divide divisor by 100\n 3. return True\n\nTime complexity : O(log(n))\n\nSpace complexity : O(1)\n\"\"\"\n\nfrom math import log10\n\n\ndef is_palindrome(n: int):\n \"\"\"\n To check whether a integer is a palindrome or not\n :param n: integer\n :return: Boolean\n \"\"\"\n digits = int(log10(n)) # to get number of digits - 1\n divisor = 10 ** digits\n\n while n is not 0:\n\n # to get the first element\n head = n // divisor\n\n # to get the last element\n tail = n % 10\n\n if head is not tail:\n return False\n\n # removing head element from n\n n %= divisor\n\n # removing tail element from n\n n //= 10\n\n # reducing divisor by 100 as two elements have been removed\n divisor //= 100\n\n return True\n","repo_name":"Pranjulcr7/Boodl-Tech_solutions","sub_path":"Problem2.py","file_name":"Problem2.py","file_ext":"py","file_size_in_byte":1058,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"26731667338","text":"# https://www.youtube.com/watch?v=AShHJdSIxkY&ab_channel=MarkJay\n\nimport pyaudio\nimport struct\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nCHUNK = 1024 * 250\nFORMAT = pyaudio.paInt16\nCHANNELS = 2\nRATE = 44100\n\np = pyaudio.PyAudio()\n\nstream = p.open(format=FORMAT,\n channels=CHANNELS,\n rate=RATE,\n input=True,\n output=True,\n frames_per_buffer=CHUNK)\n\nraw_data = stream.read(CHUNK)\ndata = np.frombuffer(raw_data, dtype=np.int16)\nplt.plot(data)\nplt.show()\n","repo_name":"animesh-chouhan/ledstrip-sound-reactive","sub_path":"capture_test.py","file_name":"capture_test.py","file_ext":"py","file_size_in_byte":536,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"36082785128","text":"# coding: utf-8\nfrom datetime import datetime\n\nfrom scrapy import Spider, Request\nfrom dateutil import parser\nfrom pytz import timezone\nfrom w3lib.url import add_or_replace_parameter\n\nfrom pollution_app.pollution import Kind\nfrom pollution_app.items import AppItem\nfrom pollution_app.settings import SCRAPER_TIMEZONE\n\n\nclass EnglandSpider(Spider):\n name = u\"gb_england\"\n tz = u\"GMT\"\n source = u\"http://www.airqualityengland.co.uk\"\n\n def start_requests(self):\n codes = (u\"AY1\", u\"BAR6\", u\"BAR9\", u\"HB010\", u\"HB011\", u\"BAI2\", u\"WIL1\", u\"BUR2\", u\"BUR1\", u\"WIL8\", u\"WIL5\",\n u\"NEW2\", u\"CAM3\", u\"CAM5\", u\"CAM4\", u\"CAM1\", u\"CRL2\", u\"HB013\", u\"WIL3\", u\"HB012\", u\"EWE2\", u\"FAR2\",\n u\"GA1\", u\"GA2\", u\"GA3\", u\"GIRT\", u\"FAR1\", u\"T55\", u\"LHR2\", u\"T54\", u\"HEN\", u\"HB008\", u\"HB009\", u\"HI1\",\n u\"SIPS\", u\"HB002\", u\"HB003\", u\"HS5\", u\"HS4\", u\"HS2\", u\"HS9\", u\"HS8\", u\"HS7\", u\"HS6\", u\"BN2\", u\"HIL1\",\n u\"HIL4\", u\"HIL5\", u\"HI3\", u\"HB006\", u\"HB007\", u\"MAN1\", u\"MAN7\", u\"MAHG\", u\"WIL7\", u\"NUL1\", u\"OX6\",\n u\"OX3\", u\"REA2\", u\"REA4\", u\"RED3\", u\"IMP\", u\"ORCH\", u\"M60\", u\"WIL4\", u\"CW\", u\"SLH7\", u\"SLH3\", u\"SLH6\",\n u\"SLH5\", u\"SLH8\", u\"SLH9\", u\"SLH4\", u\"GX\", u\"SHOL\", u\"MONK\", u\"HB005\", u\"STK7\", u\"STK5\", u\"SUN2\",\n u\"SUN4\", u\"BN1\", u\"TAM1\", u\"TAME\", u\"GOS1\", u\"TRAF\", u\"TRF2\", u\"WD1\", u\"WL4\", u\"WL1\", u\"WL5\", u\"HB004\",\n u\"WAT\", u\"HB001\", u\"WID2\", u\"WID1\", u\"WIG7\", u\"NEW3\", u\"WYA4\", u\"WSTO\", u\"YK10\", u\"YK11\", u\"YK16\",\n u\"YK7\", u\"YK13\", u\"YK8\", u\"YK9\", u\"YK15\", u\"YK018\", u\"BAR3\", u\"BPLE\", u\"BATH\", u\"BIL\", u\"BBRD\",\n u\"BIRR\", u\"AGRN\", u\"BIR1\", u\"BLAR\", u\"BLC2\", u\"BORN\", u\"BDMA\", u\"BRT3\", u\"BRS8\", u\"BURW\", u\"CAM\",\n u\"CANK\", u\"CANT\", u\"CARL\", u\"MACK\", u\"CHAT\", u\"CHLG\", u\"CHS7\", u\"CHBO\", u\"CHBR\", u\"COAL\", u\"DCST\",\n u\"EB\", u\"EX\", u\"GLAZ\", u\"HM\", u\"HONI\", u\"HUL2\", u\"HULR\", u\"LB\", u\"LEAM\", u\"LEAR\", u\"LEED\", u\"LED6\",\n u\"LEIR\", u\"LECU\", u\"LEOM\", u\"LIN3\", u\"LVP\", u\"LH\", u\"LUTR\", u\"MAN3\", u\"MKTH\", u\"MID\", u\"NEWC\", u\"NCA3\",\n u\"NTN3\", u\"NO12\", u\"NOTT\", u\"NWBV\", u\"BOLD\", u\"OX\", u\"OX8\", u\"PLYM\", u\"PMTH\", u\"PRES\", u\"REA5\",\n u\"ROCH\", u\"ECCL\", u\"SASH\", u\"SDY\", u\"SCN2\", u\"SHBR\", u\"SHDG\", u\"SHE\", u\"SIB\", u\"SA33\", u\"SOUT\",\n u\"SEND\", u\"SHLW\", u\"OSY\", u\"SOTR\", u\"EAGL\", u\"STKR\", u\"STOK\", u\"STOR\", u\"SUNR\", u\"WAL4\", u\"WAR\",\n u\"WEYB\", u\"WFEN\", u\"WSMR\", u\"WIG5\", u\"TRAN\", u\"WTHG\", u\"YW\")\n\n # codes = (u\"LEIR\",)\n url = u\"http://www.airqualityengland.co.uk/site/latest\"\n for code_value in codes:\n url = add_or_replace_parameter(url, u\"site_id\", code_value)\n\n yield Request(\n url=url,\n callback=self.parse,\n meta={u\"code\": code_value}\n )\n\n def get_station_data(self, resp):\n data_time = resp.xpath(u'//*[@id=\"pageSubArea\"]/div/p[1]/text()').re(u\"(\\d\\d\\/\\d\\d\\/\\d\\d\\d\\d\\s\\d\\d:\\d\\d)\")\n data_time = parser.parse(data_time[0]).replace(tzinfo=timezone(self.tz)) if data_time else None\n\n table = resp.xpath(u'//*[@id=\"pageSubArea\"]/div/table/tr')[1:]\n\n station_data = dict()\n for row in table:\n pollutant_name = row.xpath(u'td[1]/text()').extract_first().split(u\" (\")[0]\n pollutant_name_ind = row.xpath(u'td[1]/sub/text()').extract_first() if row.xpath(u'td[1]/sub/text()').extract_first() != None else u\"\"\n pollutant_name_time = row.xpath(u\"td[last()]/text()\").extract_first()\n\n pollutant_name = (\n u\" \".join((pollutant_name, pollutant_name_ind, pollutant_name_time))\n ).replace(u\" \", u\" \")\n\n pollutant_value = row.xpath(u\"td[last() - 1]/text()\").extract_first()\n\n if pollutant_value is not None:\n if u\"\\xa0\" in pollutant_value:\n pollutant_value = pollutant_value.split(u\"\\xa0\")[0]\n else:\n pollutant_value = pollutant_value.split(u\" \")[0]\n\n pollutant_value = pollutant_value if pollutant_value != u\"No\" else None\n\n pollutant = Kind(self.name).get_dict(r_key=pollutant_name, r_val=pollutant_value)\n if pollutant:\n station_data[pollutant[u\"key\"]] = pollutant[u\"val\"]\n\n if station_data:\n items = AppItem()\n items[u\"scrap_time\"] = datetime.now(tz=timezone(SCRAPER_TIMEZONE))\n items[u\"data_time\"] = data_time\n items[u\"data_value\"] = station_data\n items[u\"source\"] = self.source\n items[u\"source_id\"] = resp.meta[u\"code\"]\n\n yield items\n\n def parse(self, response):\n for el in self.get_station_data(response):\n yield el\n\n","repo_name":"tillroy/AmbienceData","sub_path":"pollution_app_root/pollution_app/spiders/gb_england.py","file_name":"gb_england.py","file_ext":"py","file_size_in_byte":4800,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"5586559264","text":"'''\nAlarm Clock GUI Client\n\nRequirements:\nvlc: pip install vlc\nmutagen: pip install mutagen\npython-crontab: pip install python-crontab\n'''\n\nimport sys\nimport string\nimport vlc\nimport os\nimport getpass\nfrom time import sleep\nfrom mutagen.mp3 import MP3\nfrom crontab import CronTab\nfrom tkinter import *\n\n\nusername = getpass.getuser()\nscript_path = os.path.abspath(sys.argv[0])\n\n\nclass Window(Frame):\n\t\n def __init__(self, master=None):\n Frame.__init__(self, master)\n self.master=master\n self.init_window()\n\n def init_window(self):\n self.master.title('Alarm')\n self.pack(fill=BOTH, expand=1)\n\n self.columnconfigure(1, weight=1)\n self.columnconfigure(3, pad=7)\n self.rowconfigure(3, weight=1)\n self.rowconfigure(5, pad=7)\n \n head_label = Label(self, text=\"Alarm Clock\")\n head_label.grid(padx=5, pady=5)\n\n scheduled_tasks_list_box = Listbox(self)\n scheduled_tasks_list_box.grid(row=1, column=0, columnspan=2, rowspan=4, padx=5, sticky=E+W+S+N)\n self.show_scheduled_alarms(scheduled_tasks_list_box)\n scheduled_tasks_list_box.bind(\"<>\", self.onScheduledAlarmSelect)\n\n #scrollbar_scheduled_tasks_list_box = Scrollbar(self.master, orient=\"horizontal\")\n #scrollbar_scheduled_tasks_list_box.config(command=scheduled_tasks_list_box.xview)\n #scrollbar_scheduled_tasks_list_box.pack(side=\"bottom\", fill=\"x\")\n #scheduled_tasks_list_box.config(xscrollcommand=Scrollbar.set)\n\n self.var=StringVar()\n\n\n ringnow_button = Button(self, text=\"Ring Now\", command=self.ringnow)\n ringnow_button.grid(row=1, column=3)\n \n delete_scheduled_alarm_button = Button(self, text=\"Delete Scheduled Alarm\", command= lambda: self.delete_scheduled_alarm(scheduled_tasks_list_box))\n delete_scheduled_alarm_button.grid(row=2, column=3)\n\n def show_scheduled_alarms(self, scheduled_tasks_list_box):\n scheduled_tasks_list_box.delete(0, END)\n cron_task = CronTab(user=username)\n for job in cron_task:\n if job.comment.startswith('alarm-clock-'):\n scheduled_tasks_list_box.insert(END, job)\n\n def onScheduledAlarmSelect(self, val):\n sender=val.widget\n idx=sender.curselection()\n value=sender.get(idx) \n self.var.set(value)\n\n def ringnow(self):\n alarm_file = vlc.MediaPlayer(os.path.join(os.path.dirname(script_path), \"alarm.mp3\"))\n alarm_file_length = MP3(os.path.join(os.path.dirname(script_path), \"alarm.mp3\")).info.length\n alarm_file.play()\n\n def delete_scheduled_alarm(self, scheduled_tasks_list_box):\n #import pdb\n #pdb.set_trace()\n cron_task = CronTab(user=username)\n idno = self.var.get()\n print(idno)\n for job in cron_task:\n if job.comment=='alarm-clock-'+idno:\n cron_task.remove(job)\n cron_task.write()\n self.show_scheduled_alarms(scheduled_tasks_list_box)\n\nroot = Tk()\nroot.geometry(\"400x300\")\napp = Window(root)\nroot.mainloop()","repo_name":"siddhantkhandelwal/Python-Scripts","sub_path":"alarm-clock/alarm-clock-gui.py","file_name":"alarm-clock-gui.py","file_ext":"py","file_size_in_byte":3122,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"37409393595","text":"#-*- coding:utf8 -*-\n\nimport os\n\n# Build paths inside the project like this: os.path.join(BASE_DIR, ...)\nBASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/1.9/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = '_*b844))&omdj+7)zx#1$%c4a$*cv61a)d3t=a!mr0^+ylxh2&'\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = True\n\nfrom constant import svr_config as Conf\n\n\n#========================Celery Settings==============\nCELERY_CONF = Conf['CELERY']\nBROKER_URL = CELERY_CONF['broker_url']\n#CELERY_RESULT_BACKEND = \"redis://:foo@localhost:6379/15\"\nCELERY_RESULT_BACKEND = CELERY_CONF['celery_result_backend']\n\nALLOWED_HOSTS = ['*']\n\n\n# Application definition\n\nDEFAULT_APPS = [\n # 'django.contrib.admin',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n]\n\nCUSTOM_APPS = [\n 'demo',\n 'accounts',\n 'utils',\n 'blog',\n 'middleware',\n 'cache',\n 'base'\n]\n\nTHIRD_APPS = [\n 'rest_framework',\n 'corsheaders',\n 'guardian'\n]\n\nINSTALLED_APPS = DEFAULT_APPS + CUSTOM_APPS + THIRD_APPS\n\nMIDDLEWARE_CLASSES = [\n 'django.middleware.security.SecurityMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'corsheaders.middleware.CorsMiddleware', #跨域\n 'django.middleware.common.CommonMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.auth.middleware.SessionAuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'django.middleware.clickjacking.XFrameOptionsMiddleware',\n]\n\nROOT_URLCONF = 'devsysproj.urls'\n\nTEMPLATES = [\n {\n 'BACKEND': 'django.template.backends.django.DjangoTemplates',\n 'DIRS': [],\n 'APP_DIRS': True,\n 'OPTIONS': {\n 'context_processors': [\n 'django.template.context_processors.debug',\n 'django.template.context_processors.request',\n 'django.contrib.auth.context_processors.auth',\n 'django.contrib.messages.context_processors.messages',\n ],\n },\n },\n]\n\nWSGI_APPLICATION = 'devsysproj.wsgi.application'\n\n\n# Database\n# https://docs.djangoproject.com/en/1.9/ref/settings/#databases\nif Conf['DB']['db_type'] == 'sqlite':\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),\n }\n }\nelse:\n DB_CONF = Conf['DB']\n DATABASES = {\n 'default': {\n 'ENGINE': DB_CONF['engine'],\n 'NAME': DB_CONF['dbname'],\n 'USER': DB_CONF['user'],\n 'PASSWORD': DB_CONF['password'],\n 'HOST': DB_CONF['host'],\n 'PORT': DB_CONF['port']\n }\n }\n\n# Password validation\n# https://docs.djangoproject.com/en/1.9/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',\n },\n]\n\n\n# Internationalization\n# https://docs.djangoproject.com/en/1.9/topics/i18n/\n\nLANGUAGE_CODE = 'en-us'\n\nTIME_ZONE = 'UTC'\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/1.9/howto/static-files/\n\nSTATIC_URL = '/static/'\n\n#=================REST_FRAME_WORK SETTINGS=============\nREST_FRAMEWORK = {\n\n 'DEFAULT_RENDERER_CLASSES': (\n 'rest_framework.renderers.JSONRenderer',\n ),\n\n 'DEFAULT_AUTHENTICATION_CLASSES':(\n 'rest_framework.authentication.SessionAuthentication',\n 'rest_framework.authentication.BasicAuthentication'\n )\n}\n\n\n#==============CORS SETTINGS===================\nCORS_ORIGIN_ALLOW_ALL = False\n\nCORS_ALLOW_HEADERS = (\n 'accept',\n 'accept-encoding',\n 'authorization',\n 'content-type',\n 'dnt',\n 'origin',\n 'user-agent',\n 'x-csrftoken',\n 'x-requested-with',\n 'Access-Control-Allow-Origin'\n)\n\nCORS_ORIGIN_WHITELIST = (\n 'google.com',\n 'hostname.example.com',\n 'localhost:8000',\n '127.0.0.1:8080',\n '127.0.0.1:8000',\n '0.0.0.0:8080',\n 'localhost:8080',\n '192.168.1.217:8080'\n)\n\nCORS_ALLOW_CREDENTIALS = True\n\nCORS_ALLOW_METHODS = (\n 'DELETE',\n 'GET',\n 'OPTIONS',\n 'PATCH',\n 'POST',\n 'PUT',\n)\n\nCUR_PATH = os.getcwd()\n\n#=================Logging Settings===============\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': False,\n 'handlers': {\n 'file': {\n 'level': 'INFO',\n 'class': 'logging.FileHandler',\n 'filename': '{cur_path}/log/debug.log'.format(cur_path=CUR_PATH),\n },\n },\n 'loggers': {\n 'django': {\n 'handlers': ['file'],\n 'level': 'INFO',\n 'propagate': True,\n },\n },\n}\n\n#=========Email Settings==========\nEMAIL_CONF = Conf['EMAIL']\nEMAIL_HOST = EMAIL_CONF['email_host']\nEMAIL_HOST_USER = EMAIL_CONF['email_host_user']# 用户\nEMAIL_HOST_PASSWORD = EMAIL_CONF['email_host_password'] # 密码\nEMAIL_SUBJECT_PREFIX = u'[重置密码]' # ���邮件Subject-line前缀,默认是'[django]'\nEMAIL_USE_TLS = False #与SMTP服务器通信时,是否启动TLS链接(安全链接)。默认是false\n\n\n#============Memcached Settings=======\nMEMACHED_CONF = Conf['MEMCACHED']\nMEMACHED_CONF_HOST = MEMACHED_CONF['location']\nCACHES = {\n 'default': {\n 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',\n 'LOCATION': MEMACHED_CONF_HOST,\n }\n}\n\nAUTHENTICATION_BACKENDS = [\n 'django.contrib.auth.backends.ModelBackend', # default\n 'guardian.backends.ObjectPermissionBackend',\n]\n","repo_name":"EroticMango/devsys","sub_path":"devsysproj/devsysproj/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":6168,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"2237097531","text":"def DruhaNeboPata(a,b,c,d,e,f):\n \n # v nejhorsim pripade 7\n ####################\n # Použijeme algoritmus na určení druheho nejtezsiho:\n # nalezneme nejtezsiho pomoci peti porovnani\n # na druheho nejtezsiho mame dva kandidaty, bud u a e/f, nebo z a x/y a a/b/c/d\n # celkem 6 nebo 7 porovnani, v nejhoršim případě 7, když je u težší než z.\n # p\n # / \\\n # u \\\n # / \\ \\\n # / \\ \\\n # x y z\n # / \\ / \\ / \\\n # a b c d e f\n # Využití možnosti vrátit i druhý nejlehčí jsem zatím nenalezl...\n\n # JeTezsi = lambda x, y: x > y\n\n vratTezsi = lambda i, j: i if JeTezsi(i ,j) else j\n\n x = vratTezsi(a, b)\n y = vratTezsi(c, d) \n z = vratTezsi(e, f)\n u = vratTezsi(x, y)\n if vratTezsi(u, z) == u:\n if u == x:\n return vratTezsi(y, z)\n else: # u == y\n return vratTezsi(x, z)\n else:\n if z == e:\n return vratTezsi(f, u)\n else:\n return vratTezsi(u, e)\n\n\n# print(DruhaNeboPata(*map(int, input().split())))","repo_name":"zdenecek/mff_stuff","sub_path":"programovani_1/cviceni_holan/secondmax.py","file_name":"secondmax.py","file_ext":"py","file_size_in_byte":1069,"program_lang":"python","lang":"cs","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"1090062787","text":"# create a calculater that determines the number of times the letters \n# t, r, u, e, l, o, v, e occur in 2 names\n\n# print() greeting\n\nprint(\"Welcome to Love Calculator!\")\n\n# Ask user for both names\n\nname1 = input(\"What is your name? \\n\") \nname2 = input(\"What is thier name? \\n\")\n\n# Use lower() to make all letters lower case and easier to count\n\nname1 = name1.lower()\nname2 = name2.lower()\n\n# Create variables to store scores\n\ntotal1 = 0\ntotal2 = 0\n\n# Add together the scores\n\ntotal1 += (name1 + name2).count(\"t\")\ntotal1 += (name1 + name2).count(\"r\")\ntotal1 += (name1 + name2).count(\"u\")\ntotal1 += (name1 + name2).count(\"e\")\ntotal2 += (name1 + name2).count(\"l\")\ntotal2 += (name1 + name2).count(\"o\")\ntotal2 += (name1 + name2).count(\"v\")\ntotal2 += (name1 + name2).count(\"e\")\n\n# Concatenate the 2 int for your final score\n\ntotal = int(str(total1) + str(total2))\n\n# print() scores along with message for user\nif total < 10 or total > 90:\n print(f\"Your score is {total}, you go together like coke and mentos.\")\nelif total > 40 and total < 50:\n print(f\"Your score is {total}, you are alright together.\")\nelse:\n print(f\"Your score is {total}.\")\n ","repo_name":"JHSpaz988/-100daysofcodingpythonchallenge","sub_path":"day3/projects/love_calculator.py","file_name":"love_calculator.py","file_ext":"py","file_size_in_byte":1151,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"30791049574","text":"\"\"\"\n(c) 2022 Shoshi (Sharon) Cooper. No duplication is permitted for commercial use. Any significant changes made must be\nstated explicitly and the original source code, if used, must be available and credited to Shoshi (Sharon) Cooper.\n\nFor stock-related classes\n\"\"\"\nfrom TaxAlgorithms.dependencies_for_programs.aggregated_list_class import Aggregated\n\n\nclass CorporateItems(object):\n def __init__(self, fmv, ab=None, **kwargs):\n super().__init__(**kwargs)\n self.fmv = fmv\n # Adjusted basis will be calculated once it's given to the shareholder\n self.ab = ab\n\n\nclass Stock(CorporateItems):\n def __init__(self, fmv, shares, ab=None, shareholder=None, par=None, corp=None, **kwargs):\n super().__init__(fmv=fmv, ab=ab, **kwargs)\n self.shareholder = shareholder\n self.par = par\n self.shares = shares\n self.corp = corp\n\n def sell(self, shareholder, num_shares, amount):\n shareholder.shares = type(self)(amount, num_shares)\n shareholder.shares.ab = amount\n proportion = num_shares / self.shares\n gain = amount - proportion * self.ab\n self.shares -= num_shares\n self.fmv = (amount / num_shares) * self.shares\n self.ab -= proportion * self.ab\n\n self.gain_on_sale = gain\n\n # Must update amount if I have it.\n if hasattr(self, '_amount'):\n self._amount = self.ab\n\n @property\n def amount(self):\n try:\n return self._amount\n except AttributeError:\n return self.ab\n\n @amount.setter\n def amount(self, value):\n self._amount = value\n\n\n def set_holding_period(self, properties_contributed):\n \"\"\"\n We're looking at the holding period in terms of the % of a single share. If you only sold 1 share of stock\n and you had a 40% to 60% breakdown, you'd still treat 40% as short term and 60% as long term.\n \"\"\"\n self.long_term_percent = 0\n self.short_term_percent = 0\n\n if all([prop.does_holding_period_tack for prop in properties_contributed]):\n # Then set holding period to \"long term\" or longer than 1 year\n self.long_term_percent = 1\n elif all([not prop.does_holding_period_tack for prop in properties_contributed]):\n self.short_term_percent = 1\n else:\n # When you have a mixed bag of holding periods like this, we must allocate based per share based on FMV\n agg_fmv = sum([prop.fmv for prop in properties_contributed])\n for prop in properties_contributed:\n # (property fmv / aggregate fmv)\n if prop.does_holding_period_tack:\n self.long_term_percent += (prop.fmv / agg_fmv)\n else:\n self.short_term_percent += (prop.fmv / agg_fmv)\n\n def __str__(self):\n return f\"<{self.shares} in {self.corp}>\"\n\n def __repr__(self):\n return str(self)\n\n\nclass Boot(CorporateItems):\n def __init__(self, fmv):\n super().__init__(fmv)\n # The adjusted basis of boot received by the shareholder is the FMV on date of transfer\n self.ab = self.fmv\n # holding period of boot received is always a fresh start, so starts at 0\n self.holding_period = 0\n\n\nclass Bonds(Boot):\n \"\"\"Bonds and corporate debt do not count as stock for section 351\"\"\"\n pass\n\n\n\nclass VotingStock(Stock):\n pass\n\nclass NonVotingStock(Stock):\n pass\n\n\n# Partial shares\n\nclass PartialShares(object):\n\n def __init__(self, stock_object, proportion):\n self.stock_object = stock_object\n self.proportion = proportion\n\n def __getattr__(self, item):\n if item == 'stock_object' or item == 'proportion':\n return super().__getattribute__(item)\n attribute = getattr(self.stock_object, item)\n if isinstance(attribute, (int, float)):\n return attribute * self.proportion\n return attribute\n\n\n\nclass MultiplePurchases(Aggregated):\n \"\"\"Multiple Purchases of the same stock\"\"\"\n def __init__(self, iterable=(), corp_name=None):\n super().__init__(iterable)\n self.corp = corp_name\n\n\n\n\nclass CommonStock(VotingStock):\n\n def stock_dividend(self, other_stock_object):\n # Adding stock in a stock dividend or a stock split\n if not isinstance(other_stock_object, type(self)):\n # if it's a different class of stock (So preferred or class B common or something)\n total_fmv = self.fmv + other_stock_object.fmv\n basis_of_self = (self.fmv / total_fmv) * self.ab\n basis_of_other = (other_stock_object.fmv / total_fmv) * self.ab\n self.ab = basis_of_self\n other_stock_object.ab = basis_of_other\n return self, other_stock_object\n # if it's the same class of stock, then the number of shares changes so the basis per share changes, but AB\n # does not change\n self.shares += other_stock_object.shares\n return self\n\n\nclass StockRights(object):\n def __init__(self, class_of_stock, price_guaranteed, num_rights, mv_rights_pr, curr_stock):\n \"\"\"\n Class of stock -- which class of stock it is. Should be an actual Python class (in this document or otherwise)\n Price guaranteed -- guaranteed purchase price\n Num_rights -- the number of rights you received\n mv_rights_pr -- market value of rights per right\n curr_stock -- if you already own this class of stock, this is where you pass that info in.\n \"\"\"\n self.class_of_stock = class_of_stock\n self.price_guaranteed = price_guaranteed\n self.num_rights = num_rights\n self.curr_stock = curr_stock\n\n self._mv_stock_ps = self.curr_stock.fmv / self.curr_stock.shares\n self._mv_rights_pr = mv_rights_pr\n\n self.ab = 0\n\n value_of_rights = self.num_rights * self.market_value_rights_per_right\n # Must allocate from original stock if value of rights >= .15 * value of stock\n if value_of_rights >= .15 * self.curr_stock.fmv:\n self.allocate()\n\n\n def allocate(self):\n stock_value = self.market_value_stock_per_share * self.curr_stock.shares\n rights_value = self.market_value_rights_per_right * self.num_rights\n cost_of_stock = self.curr_stock.ab\n self.curr_stock.ab = round((stock_value / (stock_value + rights_value)) * cost_of_stock, 2)\n self.ab = round((rights_value / (stock_value + rights_value)) * cost_of_stock, 2)\n\n @property\n def market_value_stock_per_share(self):\n return self._mv_stock_ps\n\n @property\n def market_value_rights_per_right(self):\n return self._mv_rights_pr\n\n def exercise(self, num_rights):\n # FMV is set to None because it does not matter and is not relevant at the moment\n new_shares = self.class_of_stock(fmv=None, shares=num_rights,\n ab=(self.ab * num_rights / self.num_rights) +\n num_rights * self.price_guaranteed)\n self.ab = self.ab * (self.num_rights - num_rights) / self.num_rights\n self.num_rights -= num_rights\n return new_shares\n\n def sell(self, num_rights, price_per_right):\n capital_gain = num_rights * price_per_right - self.ab\n self.num_rights -= num_rights\n return capital_gain\n\n\ndef is_stock_nonqualified_preferred(does_holder_have_right_to_require_issuer_to_redeem_or_buy_stock: bool,\n is_issuer_required_to_redeem_buy_stock: bool, does_issuer_have_right_to_redeem_buy_stock: bool,\n likelihood_of_exercising_that_right_on_issue_date: float,\n does_dividend_rate_on_stock_vary_with_reference_to_interest_rates_commodities_etc):\n if does_holder_have_right_to_require_issuer_to_redeem_or_buy_stock:\n return True\n if is_issuer_required_to_redeem_buy_stock:\n return True\n if does_issuer_have_right_to_redeem_buy_stock:\n if likelihood_of_exercising_that_right_on_issue_date > .5:\n return True\n if does_dividend_rate_on_stock_vary_with_reference_to_interest_rates_commodities_etc:\n return True\n return False\n\n\n\n\nclass PrefStock(NonVotingStock):\n\n def __init__(self, fmv, shares, corp=None, ab=None, par=None, **kwargs):\n super().__init__(fmv=fmv, shares=shares, corp=corp, ab=ab, par=par, **kwargs)\n self._info = kwargs\n\n\nclass PreferredStock(object):\n def __new__(cls, fmv, shares, ab=None, corp=None, par=None, flat_dividend=0, dividend_rate=0,\n dividend_rate_reference_index=None, redeemable_for=(), manditorily_redeemable=False,\n holder_has_right_to_require_redemption=False, right_of_issuer_to_redeem_stock=None,\n requirement_of_issuer_to_redeem_stock=False,\n **kwargs):\n information = {'flat_dividend': flat_dividend,\n 'dividend_rate':dividend_rate,\n 'dividend_rate_reference_index': dividend_rate_reference_index,\n 'redeemable_for': redeemable_for,\n 'manditorily_redeemable': manditorily_redeemable,\n 'right_of_issuer_to_redeem_stock': right_of_issuer_to_redeem_stock,\n 'requirement_of_issuer_to_redeem_stock': requirement_of_issuer_to_redeem_stock,\n 'holder_has_right_to_require_redemption': holder_has_right_to_require_redemption\n }\n information.update(kwargs)\n\n type_of_stock = QualifiedPreferredStock\n # Check if it's nonqualified preferred stock based on the criteria below:\n counts_as_nonqualified_preferred = [\n dividend_rate_reference_index is not None,\n len(redeemable_for) > 0 or manditorily_redeemable,\n requirement_of_issuer_to_redeem_stock or holder_has_right_to_require_redemption,\n right_of_issuer_to_redeem_stock is not None\n ]\n for is_nonqualified in counts_as_nonqualified_preferred:\n if is_nonqualified:\n type_of_stock = NonQualifiedPreferredStock\n break\n\n return type_of_stock(fmv=fmv, shares=shares, ab=ab, corp=corp, par=par, **information)\n\n\n\nclass QualifiedPreferredStock(PrefStock):\n pass\n\nclass NonQualifiedPreferredStock(PrefStock, Boot):\n pass\n\n\n\n\nclass ClassesStock(object):\n \"\"\"A corporation's collection of stock classes\"\"\"\n\n class DefStck(object):\n def __init__(self, stnding):\n self.issued = stnding\n self.outstanding = stnding\n self.treasury = []\n self.par = None\n self.value = 0\n self.apic = 0\n self.shares_issued = Aggregated()\n\n def __init__(self, classes_stock_to_outstanding_shares:dict = None):\n self._information = {}\n\n if classes_stock_to_outstanding_shares is not None:\n for stck_class, outstanding in classes_stock_to_outstanding_shares.items():\n self.add_stock_class(stck_class, outstanding)\n\n def add_stock_class(self, stock_class, outstanding_shares, issued=None, treasury=0, par=None):\n \"\"\"Adds a class of stock\"\"\"\n self._information[stock_class] = self.DefStck(outstanding_shares)\n issued = issued if issued is not None else outstanding_shares\n for attr_name, attr_val in [('issued', issued), ('treasury', treasury), ('par', par)]:\n self._information[stock_class][attr_name] = attr_val\n\n def issue(self, stock_class, shares_issued, issue_price, issue_costs):\n \"\"\"Issues new stock\"\"\"\n stock_info = self[stock_class]\n if stock_info.par is None:\n stock_info['value'] += issue_price - issue_costs\n else:\n stock_info['value'] += shares_issued * stock_info.par\n stock_info['apic'] += (issue_price - (shares_issued * stock_info.par)) - issue_costs\n\n for attr_name in ['issued', 'outstanding']:\n stock_info[attr_name] += shares_issued\n\n shares = stock_class(fmv=issue_price, shares=shares_issued)\n stock_info['shares_issued'].append(shares)\n return shares\n\n def treasury_resale(self, stock_class, num_shares, total_sold_for, costs):\n \"\"\"Resells treasury stock\"\"\"\n # TODO: Do this\n raise NotImplementedError(\"Didn't do this one yet\")\n\n def buy_back(self, shares):\n \"\"\"Buys back shares as treasury stock\"\"\"\n info = self[type(shares)]\n if info.par is None:\n info.value -= shares.fmv\n else:\n info.value -= shares.shares * info.par\n info.apic -= shares.fmv - shares.shares * info.par\n\n info.outstanding -= shares.shares\n info.treasury.append(shares)\n\n\n def __getitem__(self, stock_class):\n return self._information[stock_class]\n\n\n def _add_attr(self, attr_name, start=0):\n \"\"\"Compiles an attribute across all stock classes\"\"\"\n total = start\n for stck_class, information in self._information.items():\n total += information[attr_name]\n return total\n\n @property\n def outstanding(self):\n return self._add_attr('outstanding')\n\n @property\n def fmv(self):\n return self._add_attr('value') + self._add_attr('apic')\n\n @property\n def ab(self):\n shrs = Aggregated(self._add_attr('shares_issued', []))\n return shrs.ab\n\n\n\n\n\n\n\n\n\n\n\n","repo_name":"shoshicooper/TaxAlgorithms","sub_path":"TaxAlgorithms/dependencies_for_programs/classes_stock.py","file_name":"classes_stock.py","file_ext":"py","file_size_in_byte":13419,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"5907244760","text":"from tkinter import *\n\nclass Application(Frame):\n \n def __init__(self, master=None):\n Frame.__init__(self, master)\n self.pack()\n self.createWidgets()\n \n def createWidgets(self):\n top_frame = Frame(self)\n self.entry1 = Entry(top_frame)\n self.entry1.pack(side=LEFT)\n self.entry2 = Entry(top_frame)\n self.entry2.pack(side=LEFT)\n top_frame.pack()\n bottom_frame = Frame()\n self.resultlabel = Label(bottom_frame,text='Product:')\n self.resultlabel.pack()\n self.button = Button(bottom_frame,text='Multiply',command=self.handle)\n self.button.pack()\n bottom_frame.pack()\n \n def handle(self):\n entry1 = self.entry1.get()\n entry2 = self.entry2.get()\n try:\n entry1 = float(entry1)\n except ValueError:\n entry1 = False\n try:\n entry2 = float(entry2)\n except ValueError:\n entry2 = False\n if entry1 and entry2:\n product = entry1 * entry2\n else:\n product = '***ERROR***'\n self.resultlabel.config(text='Product:{0}'.format(product))\n \nroot = Tk()\napp = Application(master=root)\napp.mainloop()\n","repo_name":"aborgo/Certification_Work","sub_path":"tkinter_multiplier.py","file_name":"tkinter_multiplier.py","file_ext":"py","file_size_in_byte":1232,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"28608224787","text":"'''Hacer un programa que llene obligatoriamente un arreglo de numeros \r\nde 5 posiciones validados , al final debe mostrar cuantos pares e impares se encuentran en el arreglo '''\r\n\r\ndef validarNumero(n):\r\n if n.isdigit():\r\n return int(n)\r\n else:\r\n print('No es un numero')\r\n return 0\r\n\r\na = [0,0,0,0,0]\r\ncp = 0\r\nci = 0\r\nfor i in range (0,5):\r\n b = validarNumero(input ('Escribe un valor\\n'))\r\n if not b == 0:\r\n a[i] = b\r\n \r\nfor i in a:\r\n if i % 2 == 0:\r\n cp += 1\r\n else:\r\n ci += 1\r\nfor i in a:\r\n print(i)\r\n\r\nprint('Total de pares ', cp, ' y de impares' ,ci )","repo_name":"jis0017/Parcial_1_python_jesus","sub_path":"PRACTIA11.py","file_name":"PRACTIA11.py","file_ext":"py","file_size_in_byte":623,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"2215098874","text":"import math\nimport numpy as np\nimport scipy as sp\n\n\n\nclass vector_td(np.matrix):\n\n def __new__(cls, input_array):\n obj = np.matrix(input_array, float).view(cls)\n \n # transposes to the matrix arithmetic order of 1 column 3 rows \n return obj.T\n\n def __str__(self):\n return 'vector_td containing elements: '+str(self)\n\n\n\nclass transformation_matrix(np.matrix):\n\n def __new__(cls, rotang=None, rotax=None, travec=None):\n obj = np.matrix(np.diagflat(np.ones(4)), float).view(cls) \n \n if travec != None or rotang or rotax != None:\n\n if travec != None:\n obj[:3,3:4] = travec\n \n if rotang or rotax != None: \n\n if rotang:\n obj.rotang = rotang\n\n if rotax != None:\n obj.rotax = rotax\n\n else:\n obj.rotax = vector_td([0,0,0])\n\n obj.matrix_rotate()\n \n else:\n print('neigther vectortd nor angle given')\n \n return obj\n\n\n def __array_finalize__(self, obj):\n self.rotang = getattr(obj, 'rotang', None)\n self.rotax = getattr(obj, 'rotax' , None)\n self.travec = getattr(obj, 'travec', None)\n\n\n def matrix_rotate(self):\n c = math.cos(math.pi/180*self.rotang)\n s = math.sin(math.pi/180*self.rotang)\n\n x = self.rotax[0]\n y = self.rotax[1]\n z = self.rotax[2]\n\n self[0,0] = c+(1-c)*(x)**2\n self[0,1] = (1-c)*x*y-s*z\n self[0,2] = (1-c)*x*z+s*y\n\n self[1,0] = (1-c)*x*y+s*z\n self[1,1] = c+(1-c)*y**2\n self[1,2] = (1-c)*y*z-s*x\n\n self[2,0] = (1-c)*x*z-s*y\n self[2,1] = (1-c)*y*z+s*x\n self[2,2] = c+(1-c)*z**2\n\n return self\n\n def __str__(self):\n return 'Transformation Matrix mit der Matrix: '+self.name+'und vom typ type(self) '\n\n\n\nclass vector_td_position(vector_td):\n\n def __new__(cls, coordinates):\n obj = vector_td(coordinates).view(cls)\n obj.matrix = transformation_matrix()\n\n return obj\n\n\n def __array_finalize__(self, obj):\n self.matrix = getattr(obj, 'matrix', None)\n\n\n def rotate(self, rotang, rotax):\n self.matrix = transformation_matrix(rotang=rotang, rotax=vector_td(rotax))[:3,:3]\n result = self.matrix * self\n\n for i in range(len(result)):\n self[i] = result[i]\n\n return self \n\n\n def transit(self, travec):\n self.matrix = transformation_matrix(travec=vector_td(travec))[:3,3:4]\n result = self.matrix + np.asmatrix(self)\n\n for i in range(len(result)):\n self[i] = result[i]\n\n return self\n \n\n def translocate(self, rotax, rotang=0, travec=np.zeros(3)):\n self.matrix = transformation_matrix(rotang=0, rotax=vector_td(0,0,0), travec=vector_td(0,0,0))\n vector = np.resize(self, (4,1))\n vector[3] = 1\n result = self.matrix * vector\n\n for i in range(3):\n self[i] = result[i]\n\n return self\n\n\n def __str__(self):\n return 'vector_td_position: '+type(self)\n\n\n\nclass td_object(list):\n\n def __new__(cls, vertex_list):\n print ('cls')\n obj = vertex_list\n obj.orient = transformation_matrix()[:3,:3]\n obj.position = vector_td_position([0,0,0])\n \n def __array_finalize__(obj, self):\n self.orient = getattr(obj, 'orient', None)\n self.position = getattr(obj, 'position', None)\n","repo_name":"JoergReinhardt/python_kinematic","sub_path":"kinematic.py","file_name":"kinematic.py","file_ext":"py","file_size_in_byte":3545,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"11152521582","text":"import frappe, json\n\n@frappe.whitelist()\ndef check_order_tracking(name):\n order_tracking = frappe.db.sql(\"\"\" SELECT COUNT(*) as count FROM `tabOrder Tracking` WHERE purchase_order_ref=%s \"\"\",name, as_dict=1 )\n\n return order_tracking[0].count > 0\n\ndef validate_po(doc, method):\n if len(doc.orders) > 0:\n for i in doc.items:\n if i.qty != i.final_moq:\n frappe.throw(\"Final MOQ (\" + str(i.final_moq) + \") for item \" + i.item_name + \" is not equal to Order Qty (\" + str(i.qty) + \")\")\n\n@frappe.whitelist()\ndef create_order_tracking(doc):\n data = json.loads(doc)\n print(data['orders'])\n\n obj = {\n \"doctype\": \"Order Tracking\",\n \"supplier\": data['supplier'],\n \"purchase_order_ref\": data['name'],\n \"order_tracking_items\": get_order_tracking_items(data),\n \"purchase_order_date\": data['transaction_date'],\n }\n ot = frappe.get_doc(obj).insert()\n return ot.name\n\ndef get_order_tracking_items(doc):\n items = []\n for i in doc['orders']:\n items.append({\n \"order\": i['order'],\n })\n return items\ndef get_order_tracking_location():\n status = [{\n \"status\": \"Waiting\",\n }]\n return status\n\n\ndef on_trash_po(doc, method):\n for i in doc.orders:\n frappe.db.sql(\"\"\" UPDATE `tabOrder` SET purchase_order='' WHERE name=%s \"\"\", (i.order))\n frappe.db.commit()","repo_name":"exvas/grand","sub_path":"grand/doc_events/purchase_order.py","file_name":"purchase_order.py","file_ext":"py","file_size_in_byte":1390,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38614674994","text":"\"\"\"\nQuestion: Anagrams\n\nGiven two strings s and t , write a function to determine if t is an anagram of s.\n\nSample Input 1\ns = \"anagram\", t = \"nagaram\"\nSample Output 1\nTrue\n\nSample Input 2\ns = \"rat\", t = \"car\"\nSample Output 2\nFalse\n\"\"\"\n\n# Time Complexity --> nlogn + mlogm\ndef check_anagram(sentence_1,sentence_2):\n #write your code here\n sentence_1 = (sentence_1.lower()).replace(\" \", \"\")\n sentence_2 = (sentence_2.lower()).replace(\" \", \"\")\n\n if sorted(sentence_1) == sorted(sentence_2):\n return True\n else:\n return False\n \nif __name__ == \"__main__\":\n sentence_1 = \"aabc\"\n sentence_2 = \"abcd\"\n print(check_anagram(sentence_1,sentence_2))\n","repo_name":"Mananjot/100-Day-of-Code","sub_path":"day25.py","file_name":"day25.py","file_ext":"py","file_size_in_byte":656,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"3598966684","text":"from sorter_inceptionV3 import Sorter\nimport argparse\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"train_dir\")\n parser.add_argument(\"validation_dir\")\n parser.add_argument(\"save_weights_path\")\n parser.add_argument(\"--finetuning\", type=str, default=\"\")\n parser.add_argument(\"--gpus\", type=int, default=1)\n args = parser.parse_args()\n classes = [\"hyper\", \"non-hyper\"]\n train_dir = args.train_dir\n validation_dir = args.validation_dir\n\n sorter = Sorter(\n classes=classes,\n train_dir=train_dir,\n validation_dir=validation_dir,\n save_weights_path=args.save_weights_path,\n finetuning_weights_path=args.finetuning,\n img_size=(300, 300),\n n_gpus=args.gpus,\n color_randomize_options={\n 'h': 0.05,\n 's': 0.1,\n 'v': 20,\n },\n ealry_stopping_options={\n 'patience': 10,\n }\n )\n\n # train\n sorter.train()\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"niigata-bioinfo/hypermutation-ai-code","sub_path":"detect_hypermutation/train_hyper.py","file_name":"train_hyper.py","file_ext":"py","file_size_in_byte":1015,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"11627163921","text":"import setuptools\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetuptools.setup(\n name=\"neo4j_for_django\",\n version=\"0.0.2\",\n author=\"Lila Rest\",\n author_email=\"mail@lila.rest\",\n description=\"This Python 3.X package provides Neo4j support for the Django framework.\",\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n url=\"https://github.com/LilaRest/neo4j_for_django\",\n packages=setuptools.find_packages(),\n install_requires=['neo4j'],\n extras_require={\"bcrypt\": [\"bcrypt\"], },\n classifiers=[\n 'Intended Audience :: Developers',\n 'Programming Language :: Python',\n \"Programming Language :: Python :: 3\",\n 'Programming Language :: Python :: 3 :: Only',\n \"License :: OSI Approved :: GNU General Public License v3 (GPLv3)\",\n \"Operating System :: OS Independent\",\n \"Framework :: Django\",\n \"Framework :: Django :: 2.2\",\n \"Natural Language :: English\",\n \"Topic :: Database\",\n ],\n)","repo_name":"LilaRest/neo4j_for_django","sub_path":"src/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1043,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"71577357374","text":"#!/usr/bin/env python\n\n\"\"\"compare two tractor catalogues that should have same objects\n\"\"\"\n\nfrom __future__ import division, print_function\n\nimport matplotlib\nmatplotlib.use('Agg') #display backend\nimport os\nimport sys\nimport logging\nimport argparse\nimport numpy as np\nfrom scipy import stats as sp_stats\n#import seaborn as sns\n\nimport matplotlib.pyplot as plt\n\nfrom astropy.io import fits\nfrom astropy.table import vstack, Table\nfrom astrometry.libkd.spherematch import match_radec\n\n#import thesis_code.targets as targets\nfrom legacyanalysis import targets \nfrom legacyanalysis.pathnames import get_outdir\n\n\nclass Matched_Cats():\n def __init__(self):\n self.data={}\n def initialize(self,data_1,data_2,m1,m2,m1_unm,m2_unm,d12, deg2_decam,deg2_bokmos):\n self.d12= d12 #deg separations between matches objects\n self.deg2_decam= deg2_decam \n self.deg2_bokmos= deg2_bokmos \n self.data['m_decam']= targets.data_extract(data_1,m1) \n self.data['m_bokmos']= targets.data_extract(data_2,m2)\n self.data['u_decam']= targets.data_extract(data_1,m1_unm)\n self.data['u_bokmos']= targets.data_extract(data_2,m2_unm)\n def add_d12(self,d12):\n '''concatenate new d12 with existing matched deg separation array'''\n self.d12= np.concatenate([self.d12, d12])\n def add_dict(self,match_type,new_data):\n '''match_type -- m_decam,m_bokmos,u_decam, etc\n new data -- data returend from read_from..() to be concatenated with existing m_decam, etc'''\n for key in self.data[match_type].keys(): \n self.data[match_type][key]= np.concatenate([self.data[match_type][key],new_data[key]])\n\ndef deg2_lower_limit(data):\n '''deg2 spanned by objects in each data set, lower limit'''\n ra= data['ra'].max()-data['ra'].min()\n assert(ra > 0.)\n dec= abs(data['dec'].max()-data['dec'].min())\n return ra*dec\n\ndef match_it(cat1,cat2):\n '''cat1,2 are tractor catalogue to match objects between'''\n #match cats\n data_1= targets.read_from_tractor_cat(cat1)\n data_2= targets.read_from_tractor_cat(cat2)\n #deg2 spanned by objects in each data set\n deg2_decam= deg2_lower_limit(data_1)\n deg2_bokmos= deg2_lower_limit(data_2)\n #all the 'all1' objects that have match in 'all2' \n m1, m2, d12 = match_radec(data_1['ra'],data_1['dec'],data_2['ra'],data_2['dec'],\\\n 1.0/3600.0,nearest=True)\n m1_unm = np.delete(np.arange(len(data_1['ra'])),m1,axis=0)\n m2_unm = np.delete(np.arange(len(data_2['ra'])),m2,axis=0)\n return data_1,data_2,m1,m2,m1_unm,m2_unm,d12, deg2_decam,deg2_bokmos\n\ndef read_lines(fn):\n fin=open(fn,'r')\n lines=fin.readlines()\n fin.close()\n return list(np.char.strip(lines))\n\n#plotting vars\nlaba=dict(fontweight='bold',fontsize='medium')\nkwargs_axtext=dict(fontweight='bold',fontsize='large',va='top',ha='left')\nleg_args=dict(frameon=True,fontsize='x-small')\n\ndef plot_nobs(b):\n '''make histograms of nobs so can compare depths of g,r,z between the two catalogues''' \n hi=0 \n for cam in ['m_decam','m_bokmos']:\n for band in 'grz':\n hi= np.max((hi, b[cam].data[band+'_nobs'].max()))\n bins= hi\n for cam in ['m_decam','m_bokmos']:\n for band in 'grz':\n junk=plt.hist(b[cam].data[band+'_nobs'],bins=bins,normed=True,cumulative=True,align='mid')\n xlab=plt.xlabel('nobs %s' % band, **laba)\n ylab=plt.ylabel('CDF', **laba)\n plt.savefig(os.path.join(get_outdir('bmd'),'hist_nobs_%s_%s.png' % (band,cam[2:])), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\n#def plot_nobs_2(b):\n# '''improved version of plot_nobs'''\n# for cam in ['m_decam','m_bokmos']:\n# for band in 'grz':\n# junk=plt.hist(b[cam].data[band+'_nobs'],bins=10,normed=True,cumulative=True,align='mid')\n# xlab=plt.xlabel('nobs %s' % band, **laba)\n# ylab=plt.xlabel('CDF', **laba)\n# plt.savefig(os.path.join(get_outdir('bmd'),'hist_nobs_%s_%s.png' % (band,cam[2:])), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n# plt.close()\n#\n#\n# c1= 'b' \n# c2= 'r'\n# ###\n# decam_max= [b['m_decam'].data[b+'_nobs'].max() for b in 'grz']\n# bokmos_max= [b['m_bokmos'].data[b+'_nobs'].max() for b in 'grz']\n# types= np.arange(1, np.max((decam_max,bokmos_max)) +1)\n# ind = types.copy() # the x locations for the groups\n# width = 1 # the width of the bars\n# ###\n# ht_decam, ht_bokmos= np.zeros(5,dtype=int),np.zeros(5,dtype=int)\n# for cnt,typ in enumerate(types):\n# ht_decam[cnt]= np.where(obj[m_types[0]].data['type'] == typ)[0].shape[0] / float(obj['deg2_decam'])\n# ht_bokmos[cnt]= np.where(obj[m_types[1]].data['type'] == typ)[0].shape[0] / float(obj['deg2_bokmos'])\n# ###\n# fig, ax = plt.subplots()\n# rects1 = ax.bar(ind, ht_decam, width, color=c1)\n# rects2 = ax.bar(ind + width, ht_bokmos, width, color=c2)\n# ylab= ax.set_ylabel(\"counts/deg2\")\n# if matched: ti= ax.set_title('Matched')\n# else: ti= ax.set_title('Unmatched')\n# ax.set_xticks(ind + width)\n# ax.set_xticklabels(types)\n# ax.legend((rects1[0], rects2[0]), ('decam', 'bokmos'),**leg_args)\n# #save\n# if matched: name='hist_types_Matched.png'\n# else: name='hist_types_Unmatched.png'\n# plt.savefig(os.path.join(get_outdir('bmd'),name), bbox_extra_artists=[ylab,ti], bbox_inches='tight',dpi=150)\n# plt.close()\n#\n\ndef plot_radec(obj, addname=''): \n '''obj[m_types] -- DECaLS() objects with matched OR unmatched indices'''\n #set seaborn panel styles\n #sns.set_style('ticks',{\"axes.facecolor\": \".97\"})\n #sns.set_palette('colorblind')\n #setup plot\n fig,ax=plt.subplots(1,2,figsize=(9,6),sharey=True,sharex=True)\n plt.subplots_adjust(wspace=0.25)\n #plt.subplots_adjust(wspace=0.5)\n #plot\n ax[0].scatter(obj['m_decam'].data['ra'], obj['m_decam'].data['dec'], \\\n edgecolor='b',c='none',lw=1.)\n ax[1].scatter(obj['u_decam'].data['ra'], obj['u_decam'].data['dec'], \\\n edgecolor='b',c='none',lw=1.,label='DECaLS')\n ax[1].scatter(obj['u_bokmos'].data['ra'], obj['u_bokmos'].data['dec'], \\\n edgecolor='g',c='none',lw=1.,label='BASS/MzLS')\n for cnt,ti in zip(range(2),['Matched','Unmatched']):\n ti=ax[cnt].set_title(ti,**laba)\n xlab=ax[cnt].set_xlabel('RA', **laba)\n ylab=ax[0].set_ylabel('DEC', **laba)\n ax[0].legend(loc='upper left',**leg_args)\n #save\n #sns.despine()\n plt.savefig(os.path.join(get_outdir('bmd'),'radec%s.png' % addname), bbox_extra_artists=[xlab,ylab,ti], bbox_inches='tight',dpi=150)\n plt.close()\n\n\ndef plot_HistTypes(obj,m_types=['m_decam','m_bokmos'], addname=''):\n '''decam,bokmos -- DECaLS() objects with matched OR unmatched indices'''\n #matched or unmatched objects\n if m_types[0].startswith('m_') and m_types[1].startswith('m_'): matched=True\n elif m_types[0].startswith('u_') and m_types[1].startswith('u_'): matched=False \n else: raise ValueError\n #sns.set_style(\"whitegrid\")\n #sns.set_palette('colorblind')\n #c1=sns.color_palette()[2] \n #c2=sns.color_palette()[0] #'b'\n c1= 'b' \n c2= 'r'\n ###\n types= ['PSF','SIMP','EXP','DEV','COMP']\n ind = np.arange(len(types)) # the x locations for the groups\n width = 0.35 # the width of the bars\n ###\n ht_decam, ht_bokmos= np.zeros(5,dtype=int),np.zeros(5,dtype=int)\n for cnt,typ in enumerate(types):\n ht_decam[cnt]= np.where(obj[m_types[0]].data['type'] == typ)[0].shape[0] / float(obj['deg2_decam'])\n ht_bokmos[cnt]= np.where(obj[m_types[1]].data['type'] == typ)[0].shape[0] / float(obj['deg2_bokmos'])\n ###\n fig, ax = plt.subplots()\n rects1 = ax.bar(ind, ht_decam, width, color=c1)\n rects2 = ax.bar(ind + width, ht_bokmos, width, color=c2)\n ylab= ax.set_ylabel(\"counts/deg2\")\n if matched: ti= ax.set_title('Matched')\n else: ti= ax.set_title('Unmatched')\n ax.set_xticks(ind + width)\n ax.set_xticklabels(types)\n ax.legend((rects1[0], rects2[0]), ('DECaLS', 'BASS/MzLS'),**leg_args)\n #save\n if matched: name='hist_types_Matched%s.png' % addname\n else: name='hist_types_Unmatched%s.png' % addname\n plt.savefig(os.path.join(get_outdir('bmd'),name), bbox_extra_artists=[ylab,ti], bbox_inches='tight',dpi=150)\n plt.close()\n\n\ndef bin_up(data_bin_by,data_for_percentile, bin_edges=np.arange(20.,26.,0.25)):\n '''finds indices for 0.25 bins, returns bin centers and q25,50,75 percentiles of data_percentile in each bin\n bin_edges: compute percentiles for each sample between bin_edges\n '''\n count= np.zeros(len(bin_edges)-1)+np.nan\n q25,q50,q75= count.copy(),count.copy(),count.copy()\n for i,low,hi in zip(range(len(count)), bin_edges[:-1],bin_edges[1:]):\n ind= np.all((low <= data_bin_by,data_bin_by < hi),axis=0)\n if np.where(ind)[0].size > 0:\n count[i]= np.where(ind)[0].size\n q25[i]= np.percentile(data_for_percentile[ind],q=25)\n q50[i]= np.percentile(data_for_percentile[ind],q=50)\n q75[i]= np.percentile(data_for_percentile[ind],q=75)\n else:\n pass #given qs nan, which they already have\n return (bin_edges[1:]+bin_edges[:-1])/2.,count,q25,q50,q75\n\ndef indices_for_type(obj,inst='m_decam',type='all'):\n '''return mask for selecting type == all,psf,lrg\n data -- obj['m_decam'].data\n lrg mask -- obje['m_decam'].lrg'''\n if type == 'all': \n return np.ones(obj[inst].data['type'].size, dtype=bool) #1 = True\n elif type == 'psf': \n return obj[inst].data['type'] == 'PSF'\n elif type == 'lrg': \n return obj[inst].lrg\n else: raise ValueError\n\n\ndef plot_SN_vs_mag(obj, found_by='matched',type='all', addname=''):\n '''obj['m_decam'] is DECaLS() object\n found_by -- 'matched' or 'unmatched' \n type -- all,psf,lrg'''\n #indices for type == all,psf, or lrg\n assert(found_by == 'matched' or found_by == 'unmatched')\n prefix= found_by[0]+'_' # m_ or u_\n index={}\n for key in ['decam','bokmos']:\n index[key]= indices_for_type(obj,inst=prefix+key,type=type)\n #bin up SN values\n min,max= 18.,25.\n bin_SN=dict(decam={},bokmos={})\n for key in bin_SN.keys():\n for band in ['g','r','z']:\n bin_SN[key][band]={}\n i= index[key]\n bin_edges= np.linspace(min,max,num=30)\n bin_SN[key][band]['binc'],count,bin_SN[key][band]['q25'],bin_SN[key][band]['q50'],bin_SN[key][band]['q75']=\\\n bin_up(obj[prefix+key].data[band+'mag'][i], \\\n obj[prefix+key].data[band+'flux'][i]*np.sqrt(obj[prefix+key].data[band+'flux_ivar'][i]),\\\n bin_edges=bin_edges)\n #setup plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n #plot SN\n for cnt,band in zip(range(3),['g','r','z']):\n #horiz line at SN = 5\n ax[cnt].plot([1,40],[5,5],'k--',lw=2)\n #data\n for inst,color,lab in zip(['decam','bokmos'],['b','g'],['DECaLS','BASS/MzLS']):\n ax[cnt].plot(bin_SN[inst][band]['binc'], bin_SN[inst][band]['q50'],c=color,ls='-',lw=2,label=lab)\n ax[cnt].fill_between(bin_SN[inst][band]['binc'],bin_SN[inst][band]['q25'],bin_SN[inst][band]['q75'],color=color,alpha=0.25)\n #labels\n ax[2].legend(loc=1,**leg_args)\n for cnt,band in zip(range(3),['g','r','z']):\n ax[cnt].set_yscale('log')\n xlab=ax[cnt].set_xlabel('%s' % band, **laba)\n ax[cnt].set_ylim(1,100)\n ax[cnt].set_xlim(20.,26.)\n ylab=ax[0].set_ylabel('S/N', **laba)\n text_args= dict(verticalalignment='bottom',horizontalalignment='right',fontsize=10)\n ax[2].text(26,5,'S/N = 5 ',**text_args)\n plt.savefig(os.path.join(get_outdir('bmd'),'sn_%s_%s%s.png' % (found_by,type,addname)), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\ndef plot_matched_dmag_vs_psf_fwhm(obj, type='psf'):\n '''using matched sample, plot diff in mags vs. DECAM psf_fwhm in bins \n obj['m_decam'] is DECaLS() object'''\n #indices\n index= np.all((indices_for_type(b,inst='m_decam',type=type),\\\n indices_for_type(b,inst='m_bokmos',type=type)), axis=0) #both bokmos and decam of same type\n #bin up by DECAM psf_fwhm\n bin_edges= np.linspace(0,3,num=6)\n vals={}\n for band in ['g','r','z']:\n vals[band]={}\n vals[band]['binc'],count,vals[band]['q25'],vals[band]['q50'],vals[band]['q75']=\\\n bin_up(obj['m_decam'].data[band+'_psf_fwhm'][index], \\\n obj['m_bokmos'].data[band+'mag'][index]- obj['m_decam'].data[band+'mag'][index], \\\n bin_edges=bin_edges)\n#setup plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n text_args= dict(verticalalignment='center',horizontalalignment='left',fontsize=10)\n #plot\n for cnt,band in zip(range(3),['g','r','z']):\n ax[cnt].plot(vals[band]['binc'], vals[band]['q50'],c='b',ls='-',lw=2)\n ax[cnt].fill_between(vals[band]['binc'],vals[band]['q25'],vals[band]['q75'],color='b',alpha=0.25)\n ax[cnt].text(0.05,0.95,band,transform=ax[cnt].transAxes,**text_args)\n #finish\n xlab=ax[1].set_xlabel('decam PSF_FWHM', **laba)\n ylab=ax[0].set_ylabel(r'Median $\\Delta \\, m$ (decam - bokmos)', **laba)\n ti= plt.suptitle('%s Objects, Matched' % type.upper())\n plt.savefig(os.path.join(get_outdir('bmd'),'dmag_vs_psf_fwhm_%s.png' % type), bbox_extra_artists=[ti,xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\ndef plot_matched_decam_vs_bokmos_psf_fwhm(obj, type='psf'):\n '''using matched sample, plot decam psf_fwhm vs. bokmos psf_fwhm \n obj['m_decam'] is DECaLS() object'''\n #indices\n index= np.all((indices_for_type(b,inst='m_decam',type=type),\\\n indices_for_type(b,inst='m_bokmos',type=type)), axis=0) #both bokmos and decam of same type\n #setup plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n text_args= dict(verticalalignment='center',horizontalalignment='left',fontsize=10)\n #plot\n for cnt,band in zip(range(3),['g','r','z']):\n ax[cnt].scatter(obj['m_bokmos'].data[band+'_psf_fwhm'][index], obj['m_decam'].data[band+'_psf_fwhm'][index],\\\n edgecolor='b',c='none',lw=1.)\n ax[cnt].text(0.05,0.95,band,transform=ax[cnt].transAxes,**text_args)\n #finish\n for cnt,band in zip(range(3),['g','r','z']):\n ax[cnt].set_xlim(0,3)\n ax[cnt].set_ylim(0,3)\n xlab=ax[1].set_xlabel('PSF_FWHM (bokmos)', **laba)\n ylab=ax[0].set_ylabel('PSF_FWHM (decam)', **laba)\n ti= plt.suptitle('%s Objects, Matched' % type.upper())\n plt.savefig(os.path.join(get_outdir('bmd'),'decam_vs_bokmos_psf_fwhm_%s.png' % type), bbox_extra_artists=[ti,xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\n\n\ndef plot_confusion_matrix(cm,ticknames, addname=''):\n '''cm -- NxN array containing the Confusion Matrix values\n ticknames -- list of strings of length == N, column and row names for cm plot'''\n plt.imshow(cm, interpolation='nearest', cmap=plt.cm.Blues,vmin=0,vmax=1)\n cbar=plt.colorbar()\n plt.xticks(range(len(ticknames)), ticknames)\n plt.yticks(range(len(ticknames)), ticknames)\n ylab=plt.ylabel('True (DECaLS)')\n xlab=plt.xlabel('Predicted (BASS/MzLS)')\n for row in range(len(ticknames)):\n for col in range(len(ticknames)):\n if np.isnan(cm[row,col]):\n plt.text(col,row,'n/a',va='center',ha='center')\n elif cm[row,col] > 0.5:\n plt.text(col,row,'%.2f' % cm[row,col],va='center',ha='center',color='yellow')\n else:\n plt.text(col,row,'%.2f' % cm[row,col],va='center',ha='center',color='black')\n plt.savefig(os.path.join(get_outdir('bmd'),'confusion_matrix%s.png' % addname), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\ndef create_confusion_matrix(obj):\n '''compares MATCHED decam (truth) to bokmos (prediction)\n return 5x5 confusion matrix and colum/row names\n obj[m_decam'] is DECaLS object'''\n cm=np.zeros((5,5))-1\n types=['PSF','SIMP','EXP','DEV','COMP']\n for i_dec,dec_type in enumerate(types):\n ind= np.where(obj['m_decam'].data['type'] == dec_type)[0]\n for i_bass,bass_type in enumerate(types):\n n_bass= np.where(obj['m_bokmos'].data['type'][ind] == bass_type)[0].size\n if ind.size > 0: cm[i_dec,i_bass]= float(n_bass)/ind.size #ind.size is constant for each loop over bass_types\n else: cm[i_dec,i_bass]= np.nan\n return cm,types\n\ndef plot_matched_separation_hist(d12):\n '''d12 is array of distances in degress between matched objects'''\n #pixscale to convert d12 into N pixels\n pixscale=dict(decam=0.25,bokmos=0.45)\n #sns.set_style('ticks',{\"axes.facecolor\": \".97\"})\n #sns.set_palette('colorblind')\n #setup plot\n fig,ax=plt.subplots()\n #plot\n ax.hist(d12*3600,bins=50,color='b',align='mid')\n ax2 = ax.twiny()\n ax2.hist(d12*3600./pixscale['bokmos'],bins=50,color='g',align='mid',visible=False)\n xlab= ax.set_xlabel(\"arcsec\")\n xlab= ax2.set_xlabel(\"pixels [BASS]\")\n ylab= ax.set_ylabel(\"Matched\")\n #save\n #sns.despine()\n plt.savefig(os.path.join(get_outdir('bmd'),\"separation_hist.png\"), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\ndef plot_psf_hists(decam,bokmos, zoom=False):\n '''decam,bokmos are DECaLS() objects matched to decam ra,dec'''\n #divide into samples of 0.25 mag bins, store q50 of each\n width=0.25 #in mag\n low_vals= np.arange(20.,26.,width)\n med={}\n for b in ['g','r','z']: med[b]=np.zeros(low_vals.size)-100\n for i,low in enumerate(low_vals):\n for band in ['g','r','z']:\n ind= np.all((low <= decam[band+'mag'],decam[band+'mag'] < low+width),axis=0)\n if np.where(ind)[0].size > 0:\n med[band][i]= np.percentile(bokmos[band+'mag'][ind] - decam[band+'mag'][ind],q=50)\n else: \n med[band][i]= np.nan\n #make plot\n #set seaborn panel styles\n #sns.set_style('ticks',{\"axes.facecolor\": \".97\"})\n #sns.set_palette('colorblind')\n #setup plot\n fig,ax=plt.subplots(1,3,figsize=(9,3)) #,sharey=True)\n plt.subplots_adjust(wspace=0.5)\n #plot\n for cnt,band in zip(range(3),['r','g','z']):\n ax[cnt].scatter(low_vals, med[band],\\\n edgecolor='b',c='none',lw=2.) #,label=m_type.split('_')[-1])\n xlab=ax[cnt].set_xlabel('bins of %s (decam)' % band, **laba)\n ylab=ax[cnt].set_ylabel('q50[%s bokmos - decam]' % band, **laba)\n if zoom: ax[cnt].set_ylim(-0.25,0.25)\n # sup=plt.suptitle('decam with matching bokmos',**laba)\n #save\n #sns.despine()\n if zoom: name=\"median_color_diff_zoom.png\"\n else: name=\"median_color_diff.png\"\n plt.savefig(os.path.join(get_outdir('bmd'),name), bbox_extra_artists=[xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\n##########\n#funcs for flux diff / sqrt(inv var + inv var)\ndef n_gt_3_sigma(sample, low=-8.,hi=8.):\n '''for a sample that should be distributed as N(mean=0,stddev=1), returns mask for the N that are greater 3 sigma\n low,hi -- minimum and maximum sample values that will be considered'''\n i_left= np.all((sample >= low,sample <= -3.),axis=0)\n i_right= np.all((sample <= hi,sample>=3),axis=0)\n #assert i_left and i_right are mutually exclusive\n false_arr= np.all((i_left,i_right),axis=0) #should be array of Falses\n assert( np.all(false_arr == False) ) #should be np.all([True,True,...]) which evaluates to True\n return np.any((i_left,i_right),axis=0)\n\ndef gauss_stats(n_samples=10000):\n '''returns mean,std,q25, frac outliers > 3 sigma for n_samples drawn from unit gaussian N(0,1)'''\n G= sp_stats.norm(0,1)\n mean=std=q25=perc_out=0.\n for i in range(10): #draw 10 times, take avg of the 10 measurements of each statistic\n draws= G.rvs(n_samples) \n mean+= np.mean(draws)\n std+= np.std(draws)\n q25+= np.percentile(draws,q=25)\n perc_out+= 2*G.cdf(-3)*100 #HACH same number ea time\n mean/= 10.\n std/= 10.\n q25/= 10.\n perc_out/= 10.\n tol=1e-1\n assert(abs(mean) <= tol)\n assert(abs(std-1.) <= tol)\n return mean,std,q25,perc_out\n\ndef sample_gauss_stats(sample, low=-20,hi=20):\n '''return dictionary of stats about the data and stats for a sample that is unit gaussian distributed\n low,hi -- minimum and maximum sample values that will be considered'''\n a=dict(sample={},gauss={})\n #vals for unit gaussian distributed data\n a['gauss']['mean'],a['gauss']['std'],a['gauss']['q25'],a['gauss']['perc_out']= gauss_stats(n_samples=sample.size)\n #vals for actual sample\n a['sample']['mean'],a['sample']['std'],a['sample']['q25'],a['sample']['q75']= \\\n np.mean(sample),np.std(sample),np.percentile(sample,q=25),np.percentile(sample,q=75) \n i_outliers= n_gt_3_sigma(sample, low=low,hi=hi)\n a['sample']['perc_out']= sample[i_outliers].size/float(sample.size)*100.\n return a\n\n\ntext_args= dict(verticalalignment='center',fontsize=8)\ndef plot_dflux_chisq(b,type='psf', low=-8.,hi=8.,addname=''):\n #join indices b/c matched\n i_type= np.all((indices_for_type(b, inst='m_decam',type=type),\\\n indices_for_type(b, inst='m_bokmos',type=type)), axis=0) #both bokmos and decam of same type\n #get flux diff for each band\n hist= dict(g=0,r=0,z=0)\n binc= dict(g=0,r=0,z=0)\n stats=dict(g=0,r=0,z=0)\n #chi \n sample,mag={},{}\n for band in ['g','r','z']:\n sample[band]= (b['m_decam'].data[band+'flux'][i_type]-b['m_bokmos'].data[band+'flux'][i_type])/np.sqrt(\\\n np.power(b['m_decam'].data[band+'flux_ivar'][i_type],-1)+np.power(b['m_bokmos'].data[band+'flux_ivar'][i_type],-1))\n mag[band]= 22.5-2.5*np.log10(b['m_decam'].data[band+'flux'][i_type])\n #loop over mag bins, one 3 panel for each mag bin\n for b_low,b_hi in zip([18,19,20,21,22,23],[19,20,21,22,23,24]):\n #plot each filter\n for band in ['g','r','z']:\n imag= np.all((b_low <= mag[band],mag[band] < b_hi),axis=0)\n #print(\"len(imag)=\",len(imag),\"len(sample)=\",len(sample),\"len(sample[imag])=\",len(sample[imag]))\n hist[band],bins,junk= plt.hist(sample[band][imag],range=(low,hi),bins=50,normed=True)\n db= (bins[1:]-bins[:-1])/2\n binc[band]= bins[:-1]+db\n plt.close() #b/c plt.hist above\n #for drawing unit gaussian N(0,1)\n G= sp_stats.norm(0,1)\n xvals= np.linspace(low,hi)\n #plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n for cnt,band in zip(range(3),['g','r','z']):\n ax[cnt].step(binc[band],hist[band], where='mid',c='b',lw=2)\n ax[cnt].plot(xvals,G.pdf(xvals))\n #labels\n for cnt,band in zip(range(3),['g','r','z']):\n if band == 'r': xlab=ax[cnt].set_xlabel(r'%s $(F_{d}-F_{bm})/\\sqrt{\\sigma^2_{d}+\\sigma^2_{bm}}$' % band, **laba)\n else: xlab=ax[cnt].set_xlabel('%s' % band, **laba)\n #xlab=ax[cnt].set_xlabel('%s' % band, **laba)\n ax[cnt].set_ylim(0,0.6)\n ax[cnt].set_xlim(low,hi)\n ylab=ax[0].set_ylabel('PDF', **laba)\n ti=ax[1].set_title(\"%s (%.1f <= %s < %.1f)\" % (type,b_low,band,b_hi),**laba)\n #put stats in suptitle\n plt.savefig(os.path.join(get_outdir('bmd'),'dflux_chisq_%s_%.1f-%s-%.1f%s.png' % (type,b_low,band,b_hi,addname)), bbox_extra_artists=[ti,xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n################\n\ndef plot_magRatio_vs_mag(b,type='psf',addname=''):\n #join indices b/c matched\n i_type= np.all((indices_for_type(b, inst='m_decam',type=type),\\\n indices_for_type(b, inst='m_bokmos',type=type)), axis=0) #both bokmos and decam of same type\n #plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n for cnt,band in zip(range(3),['g','r','z']):\n magRatio= np.log10(b['m_bokmos'].data[band+'flux'][i_type])/np.log10(b['m_decam'].data[band+'flux'][i_type]) -1.\n mag= 22.5-2.5*np.log10(b['m_decam'].data[band+'flux'][i_type])\n ax[cnt].scatter(mag,magRatio, c='b',edgecolor='b',s=5) #,c='none',lw=2.)\n #labels\n for cnt,band in zip(range(3),['g','r','z']):\n xlab=ax[cnt].set_xlabel('%s AB' % band, **laba)\n ax[cnt].set_ylim(-0.5,0.5)\n ax[cnt].set_xlim(18,26)\n ylab=ax[0].set_ylabel(r'$m_{bm}/m_d - 1$', **laba)\n ti=ax[1].set_title(\"%s\" % type,**laba)\n #put stats in suptitle\n plt.savefig(os.path.join(get_outdir('bmd'),'magRatio_vs_mag_%s%s.png' % (type,addname)), bbox_extra_artists=[ti,xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n################\n\n\n\ntext_args= dict(verticalalignment='center',fontsize=8)\ndef plot_N_per_deg2(obj,type='all',req_mags=[24.,23.4,22.5],addname=''):\n '''image requirements grz<=24,23.4,22.5\n compute number density in each bin for each band mag [18,requirement]'''\n #indices for type for matched and unmatched samples\n index={}\n for inst in ['m_decam','u_decam','m_bokmos','u_bokmos']:\n index[inst]= indices_for_type(obj, inst=inst,type=type) \n bin_nd=dict(decam={},bokmos={})\n for inst in ['decam','bokmos']:\n bin_nd[inst]={}\n for band,req in zip(['g','r','z'],req_mags):\n bin_nd[inst][band]={}\n bin_edges= np.linspace(18.,req,num=15)\n i_m,i_u= index['m_'+inst], index['u_'+inst] #need m+u\n #join m_decam,u_decam OR m_bokmos,u_bokmos and only with correct all,psf,lrg index\n sample= np.ma.concatenate((obj['m_'+inst].data[band+'mag'][i_m], obj['u_'+inst].data[band+'mag'][i_u]),axis=0)\n bin_nd[inst][band]['binc'],bin_nd[inst][band]['cnt'],q25,q50,q75=\\\n bin_up(sample,sample,bin_edges=bin_edges)\n #plot\n fig,ax=plt.subplots(1,3,figsize=(9,3),sharey=True)\n plt.subplots_adjust(wspace=0.25)\n for cnt,band in zip(range(3),['g','r','z']):\n for inst,color,lab in zip(['decam','bokmos'],['b','g'],['DECaLS','BASS/MzLS']):\n ax[cnt].step(bin_nd[inst][band]['binc'],bin_nd[inst][band]['cnt']/obj['deg2_'+inst], where='mid',c=color,lw=2,label=lab)\n #labels\n for cnt,band in zip(range(3),['g','r','z']):\n xlab=ax[cnt].set_xlabel('%s' % band) #, **laba)\n #ax[cnt].set_ylim(0,0.6)\n #ax[cnt].set_xlim(maglow,maghi)\n ax[0].legend(loc='upper left', **leg_args)\n ylab=ax[0].set_ylabel('counts/deg2') #, **laba)\n ti=plt.suptitle(\"%ss\" % type.upper(),**laba)\n # Make space for and rotate the x-axis tick labels\n fig.autofmt_xdate()\n #put stats in suptitle\n plt.savefig(os.path.join(get_outdir('bmd'),'n_per_deg2_%s%s.png' % (type,addname)), bbox_extra_artists=[ti,xlab,ylab], bbox_inches='tight',dpi=150)\n plt.close()\n\n\nparser=argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n description='DECaLS simulations.')\nparser.add_argument('-fn1', type=str, help='process this brick (required input)',required=True)\nparser.add_argument('-fn2', type=str, help='object type (STAR, ELG, LRG, BGS)',required=True) \nargs = parser.parse_args()\n\n# Set the debugging level\nif args.verbose:\n lvl = logging.DEBUG\nelse:\n lvl = logging.INFO\nlogging.basicConfig(format='%(message)s', level=lvl, stream=sys.stdout)\nlog = logging.getLogger('__name__')\n\n#get lists of tractor cats to compare\nfns_1= read_lines(args.fn1) \nlog.info('Combining tractor catalogues: ',fns_1) \n#if fns_1.size == 1: fns_1,fns_2= [fns_1],[fns_2]\n#object to store concatenated matched tractor cats\na=Matched_Cats()\nfor cnt,cat1,cat2 in zip(range(len(fns_1)),fns_1,fns_2):\n data_1,data_2,m1,m2,m1_unm,m2_unm,d12, deg2_decam,deg2_bokmos= match_it(cat1,cat2)\n if cnt == 0:\n a.initialize(data_1,data_2,m1,m2,m1_unm,m2_unm,d12, deg2_decam,deg2_bokmos)\n else: \n a.add_d12(d12)\n a.deg2_decam+= deg2_decam\n a.deg2_bokmos+= deg2_bokmos\n a.add_dict('m_decam', targets.data_extract(data_1,m1) )\n a.add_dict('m_bokmos', targets.data_extract(data_2,m2))\n a.add_dict('u_decam', targets.data_extract(data_1,m1_unm))\n a.add_dict('u_bokmos', targets.data_extract(data_2,m2_unm))\n#each key a.data[key] becomes DECaLS() object with grz mags,i_lrg, etc\nb={}\nb['d12']= a.d12\nb['deg2_decam']= a.deg2_decam\nb['deg2_bokmos']= a.deg2_bokmos\nfor match_type in a.data.keys(): b[match_type]= targets.DECaLS(a.data[match_type], w1=True)\n#store N matched objects not masked before join decam,bokmos masks\nm_decam_not_masked,m_bokmos_not_masked= b['m_decam'].count_not_masked(),b['m_bokmos'].count_not_masked()\n#update masks for matched objects to be the join of decam and bokmos masks\nmask= np.any((b['m_decam'].mask, b['m_bokmos'].mask),axis=0)\nb['m_decam'].update_masks_for_everything(mask=np.any((b['m_decam'].mask, b['m_bokmos'].mask),axis=0),\\\n mask_wise=np.any((b['m_decam'].mask_wise, b['m_bokmos'].mask_wise),axis=0) )\nb['m_bokmos'].update_masks_for_everything(mask=np.any((b['m_decam'].mask, b['m_bokmos'].mask),axis=0),\\\n mask_wise=np.any((b['m_decam'].mask_wise, b['m_bokmos'].mask_wise),axis=0) )\n\n#plots\n#plot_radec(b)\n#plot_matched_separation_hist(b['d12'])\n# Depths are very different so develop a cut to make fair comparison\n#plot_SN_vs_mag(b, found_by='matched',type='psf')\n# mask=True where BASS SN g < 5 or BASS SN r < 5\nsn_crit=5.\nmask= np.any((b['m_bokmos'].data['gflux']*np.sqrt(b['m_bokmos'].data['gflux_ivar']) < sn_crit,\\\n b['m_bokmos'].data['rflux']*np.sqrt(b['m_bokmos'].data['rflux_ivar']) < sn_crit),\\\n axis=0)\nb['m_decam'].update_masks_for_everything(mask=mask, mask_wise=mask)\nb['m_bokmos'].update_masks_for_everything(mask=mask, mask_wise=mask)\n# contintue with fairer comparison\n#plot_radec(b,addname='snGe5')\n#plot_HistTypes(b,m_types=['m_decam','m_bokmos'],addname='snGe5')\n#plot_SN_vs_mag(b, found_by='matched',type='psf',addname='snGe5')\n#plot_SN_vs_mag(b, found_by='matched',type='all')\n#plot_SN_vs_mag(b, found_by='matched',type='lrg')\n#plot_SN_vs_mag(b, found_by='unmatched',type='all')\n#plot_SN_vs_mag(b, found_by='unmatched',type='psf')\n#plot_SN_vs_mag(b, found_by='unmatched',type='lrg')\n#cm,names= create_confusion_matrix(b)\n#plot_confusion_matrix(cm,names,addname='snGe5')\nplot_dflux_chisq(b,type='psf',addname='snGe5')\n#plot_dflux_chisq(b,type='all',addname='snGe5')\n# Number density cutting to requirement mags: grz<=24,23.4,22.5\nprint('square deg covered by decam=',b['deg2_decam'],'and by bokmos=',b['deg2_bokmos'])\n#plot_N_per_deg2(b,type='psf',addname='snGe5')\n#plot_N_per_deg2(b,type='lrg',addname='snGe5')\nplot_magRatio_vs_mag(b,type='psf',addname='snGe5')\n\n\n\n\nprint('exiting early')\nsys.exit()\n\nplot_matched_dmag_vs_psf_fwhm(b, type='psf')\nplot_matched_decam_vs_bokmos_psf_fwhm(b, type='psf')\n\nprint('finished comparison: bass-mosaic-decals')\n#sys.exit()\n#\n#\n##REVISE THIS BELOW\n##print stats of total objects, each group, # masked, etc\n#print(\"---- DECAM ----\")\n#print(\"N not masked due to grz= %d, N total= %d\" % \\\n# (m_decam_not_masked+b['u_decam'].count_not_masked(), b['m_decam'].count_total()+b['u_decam'].count_total()))\n#print(\"-- Matched --\")\n#print(\"N not masked before join bokmos mask= %d, N not masked after= %d\" % \\\n# (m_decam_not_masked, b['m_decam'].count_not_masked()))\n#print(\"-- Unmatched -- \")\n#print(\"N masked before join bokmos mask = N masked after = %d\" % \\\n# (b['u_decam'].count_total()- b['u_decam'].count_not_masked()))\n####bokmos\n#print(\"---- BOKMOS ----\")\n#print(\"N not masked due to grz= %d, N total= %d\" % \\\n# (m_bokmos_not_masked+b['u_bokmos'].count_not_masked(), b['m_bokmos'].count_total()+b['u_bokmos'].count_total()))\n#print(\"-- Matched --\")\n#print(\"N not masked before join decam mask= %d, N not masked after= %d\" % \\\n# (m_bokmos_not_masked, b['m_bokmos'].count_not_masked()))\n#print(\"-- Unmatched -- \")\n#print(\"N masked before join decam mask = N masked after = %d\" % \\\n# (b['u_bokmos'].count_total()- b['u_bokmos'].count_not_masked()))\n#print('done')\n\n\n","repo_name":"DriftingPig/Obi-Metallica","sub_path":"legacypipe/py/legacyanalysis/compare_tractor_cats.py","file_name":"compare_tractor_cats.py","file_ext":"py","file_size_in_byte":32364,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"26174447291","text":"# pylint: disable=missing-module-docstring,missing-function-docstring,eval-used\nimport sys\n\nimport operator\n\ndef main():\n \"\"\"Implement the calculator\"\"\"\n\n\n ops = {\n '+': operator.add,\n '-': operator.sub,\n '*': operator.mul,\n '/': operator.truediv, # use operator.div for Python 2\n '%': operator.mod,\n '^': operator.xor,\n }\n # print(ops[sys.argv[2]])\n # result = sys.argv[1] + ops[sys.argv[2]] + sys.argv[3]\n result = ops[sys.argv[2]](int(sys.argv[1]), int(sys.argv[3]))\n return result\n\nif __name__ == \"__main__\":\n print(main())\n \n","repo_name":"chachine/data-challenges","sub_path":"01-Python/01-Programming-Basics/05-System-Parameters/calc.py","file_name":"calc.py","file_ext":"py","file_size_in_byte":600,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"11859652716","text":"def binary_search(a, i):\r\n MOD = 1000000007\r\n lo = 0\r\n hi = i-1\r\n ans = 0\r\n while lo < hi:\r\n if a[lo]+a[hi] > a[i]:\r\n ans = (ans+hi-lo)%MOD\r\n hi = hi-1\r\n else:\r\n lo = lo+1\r\n return ans\r\n\r\nclass Solution:\r\n # @param A : list of integers\r\n # @return an integer\r\n def nTriang(self, A):\r\n MOD = 1000000007\r\n n = len(A)\r\n A.sort()\r\n ans = 0\r\n for i in range(n-1, 1, -1):\r\n ans =(ans+binary_search(A, i))%MOD\r\n return ans\r\n","repo_name":"PrinceSinghhub/InterviewBit-Two-Pointers","sub_path":"Interview Bit Two Pointer/Counting Triangles.py","file_name":"Counting Triangles.py","file_ext":"py","file_size_in_byte":548,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"17774572585","text":"# COSC 61, Professor Palmer \n# Authors: Abby Owen, Annie Revers\n# author_operations.py - SQL commands for author operations\n\nfrom mysql.connector import MySQLConnection, Error, errorcode, FieldType\nfrom dbconfig import read_db_config\nimport getpass\nfrom ManUser import *\n\n\n# TODO: File BLOBs for insert manuscript\n\n########### register_author ###########\ndef register_author(mycursor, words):\n\n insert_user = \"INSERT INTO SysUser (UserType) VALUES (%s)\"\n user_type = (\"author\", )\n \n \n try: \n mycursor.execute(insert_user, user_type)\n user_id = mycursor.lastrowid\n\n q1 = \"INSERT INTO Author (AuthorId, AuthorFirstName, AuthorLastName, AuthorEmail, AuthorAffiliation) VALUES (%s, %s, %s, %s, %s)\"\n val = (user_id, words[2], words[3], words[4], words[5])\n \n mycursor.execute(q1, val)\n \n print(f\"Thank you for registering. Your author ID is {user_id}\")\n return user_id\n except Error as err:\n print(f\"Error registering author: {err}\")\n return None\n\n\n########### login_author ###########\ndef login_author(mycursor, id):\n \n q = \"SELECT * FROM Author WHERE AuthorId = (%s)\"\n val = int(id)\n try:\n mycursor.execute(q, (val,))\n res = mycursor.fetchone()\n row = dict(zip(mycursor.column_names, res))\n \n print(f\"WELCOME AUTHOR {row['AuthorFirstName']} {row['AuthorLastName']}\".format(row))\n return row\n \n except Error as err:\n print(f\"Error logging in author, no author found with this ID: {err}\")\n return None\n\n\n########### check_author ###########\ndef check_author(mycursor, fname, lname, order):\n check_exist_sql = \"SELECT EXISTS(SELECT 1 FROM Author WHERE AuthorFirstName = %s AND AuthorLastName = %s)\"\n\n vals = (fname, lname)\n try: \n mycursor.execute(check_exist_sql, vals)\n res = mycursor.fetchone()\n if res[0] == 0:\n return None\n else:\n get_id_sql = \"SELECT AuthorId FROM Author WHERE AuthorFirstName = %s AND AuthorLastName = %s\"\n vals = (fname, lname)\n try: \n mycursor.execute(get_id_sql, vals)\n id = mycursor.fetchone()[0]\n return id\n except Error as err:\n print(f\"Error getting co author ID: {err}\")\n return None\n \n except Error as err:\n print(f\"Error accessing Author database: {err}\")\n\n\n########### submit_response ###########\ndef submit_response(man_id, mycursor):\n get_status_sql = \"SELECT ManStatus, DateUpdated FROM Manuscript WHERE ManuscriptId = %s\"\n val = (man_id, )\n try:\n mycursor.execute(get_status_sql, val)\n res = mycursor.fetchone()\n print(f\"Recieved manuscript with unique ID: {man_id}\")\n print(f\"Manuscript {man_id} Status: {res[0]}\")\n print(f\"Manuscript {man_id} recieved at: {res[1]}\")\n except Error as err:\n print(f\"Error getting manuscript information: {err}\")\n\n\n\n########### submit_manuscript ###########\n# We must: \n# INSERT INTO Manuscript (Title, ICode) VALUES (%s, %s)\n# INSERT INTO AuthorGroup (ManuscriptId, AuthorId, OrderNum) VALUES (%s, %s, %s)\n# Check if other authors are in the Author database\n# If they are not, add them.\n# FOR EACH AUTHOR: \n# INSERT INTO AuthorGroup (ManuscriptId, AuthorId, OrderNum) VALUES (%s, %s, %s)\ndef submit_manuscript(user, mycursor, title, icode, authors, filename):\n # Check permissions of user\n if user.get_id() == None:\n print(\"You do not have the proper permissions for this action. Please log in with you Author ID to submit a manuscript.\")\n return None\n \n \n # Insert the manuscript\n insert_man_sql = \"INSERT INTO Manuscript (Title, ICodeId) VALUES (%s, %s)\"\n vals = (title, icode)\n man_id = None\n try:\n mycursor.execute(insert_man_sql, vals)\n man_id = mycursor.lastrowid\n\n except Error as err:\n print(f\"Error inserting manuscript: {err}\")\n return None\n\n # If the manuscript was inserted, add primary author to author group\n if man_id != None:\n insert_primary_sql = \"INSERT INTO AuthorGroup (ManuscriptId, AuthorId, OrderNum) VALUES (%s, %s, %s)\"\n vals = (man_id, user.get_id(), 1)\n try: \n mycursor.execute(insert_primary_sql, vals)\n except Error as err:\n print(f\"Error inserting primary author: {err}\")\n return None\n \n # Check for additional authors\n if len(authors) != 0: \n for i in range(len(authors)):\n a = authors[i]\n fname, lname = a.split(\" \")\n # check if the author is already in the database\n co_id = check_author(mycursor, fname, lname, i)\n\n # if not, insert the author to the database\n if co_id == None: \n try: \n insert_user = \"INSERT INTO SysUser (UserType) VALUES (%s)\"\n u = (\"author\", )\n mycursor.execute(insert_user, u)\n co_id = mycursor.lastrowid\n\n insert_co_sql = \"INSERT INTO Author (AuthorId, AuthorFirstName, AuthorLastName) VALUES (%s, %s, %s)\"\n vals = (co_id, fname, lname)\n \n mycursor.execute(insert_co_sql, vals)\n \n except Error as err:\n print(f\"Error inserting co-author: {err}\")\n # delete manuscript?\n return None\n \n # if the author was in the database\n else:\n insert_group = \"INSERT INTO AuthorGroup (ManuscriptId, AuthorId, OrderNum) VALUES (%s, %s, %s)\"\n print(f\"Author Group ID: {co_id}\")\n vals = (man_id, co_id, i)\n try: \n mycursor.execute(insert_group, vals)\n except Error as err:\n print(f\"Error inserting to Author Group: {err}\")\n \n submit_response(man_id, mycursor)\n return man_id\n\n\n########### status ###########\ndef author_status(mycursor, user):\n try: \n print(\"######### MANUSCRIPT STATUSES #########\")\n statuses = [\"Recieved\", \"Under Review\", \"Rejected\", \"Accepted\", \"Typesetting\", \"Ready\", \"Scheduled\", \"Published\"]\n counts = {}\n for status in statuses:\n get_count = \"SELECT COUNT(ManStatus) FROM LeadAuthorManuscripts WHERE AuthorId = %s AND ManStatus = %s\"\n vals = (user.get_id(), status)\n mycursor.execute(get_count, vals)\n res = mycursor.fetchone()[0]\n counts[status] = res\n \n for c in counts: \n print(f\"{c}: {counts[c]}\")\n # mycursor.execute(status_sql, val)\n # res = mycursor.fetchall()\n \n # output = \"Manuscript Statuses \\n############## \\n Recieved \\n ############## \\n\"\n # for x in res:\n # print(x)\n except Error as err:\n print(f\"Error in getting author manuscripts: {err}\")\n\n \n\n\n\n\n\n\n \n \n\n \n\n \n\n\n\n\n \n\n \n\n\n","repo_name":"abbyowen/lab2cosc61_abby_annie","sub_path":"author_operations.py","file_name":"author_operations.py","file_ext":"py","file_size_in_byte":7097,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"15842679221","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Jun 4 11:56:22 2020\n\nThis script assesses what ranking metric is best at ranking\nstructures according to MCS or FP similarity to the true structure.\n\nThis should give an independent assessment from the identity case.\nIf the results differ markedly, we'll have to wonder why.\n\n@author: stravsm\n\"\"\"\n\n\nimport os\nimport sys\nsys.path.append(os.environ['MSNOVELIST_BASE'])\n\nimport numpy as np\nimport pandas as pd\nfrom matplotlib import pyplot as plt\nimport infrastructure.score as msc\nimport infrastructure.utils as utils\nimport fp_management.database as db\n\nfrom fp_management import fingerprint_map as fpm\nimport smiles_config as sc\n\nimport time\nimport pickle\nfrom tqdm import tqdm\n# Setup logger\nimport logging\n\nfrom rdkit import Chem\n\nimport h5py\n\nremove_perfect_match = False\nranking_score = \"score_mod_platt\"\nf1_cutoff = 0.5\n\nlogging.basicConfig(format='%(asctime)s - %(message)s', \n datefmt='%d-%b-%y %H:%M:%S')\nlogger = logging.getLogger(\"MSNovelist\")\nlogger.setLevel(logging.INFO)\nlogger.info(\"evaluation_summary startup\")\n\npicklepath = sc.config.get(\"evaluation_picklepath\", {})\n\nfp_map = fpm.FingerprintMap(sc.config[\"fp_map\"])\n\neval_id = str(int(time.time()))\neval_counter = 0\npickle_id = eval_id\nif sc.config['eval_id'] != '':\n eval_id = sc.config['eval_id']\nif sc.config['eval_counter'] != '':\n eval_counter = sc.config['eval_counter']\n pickle_id = sc.config['eval_id'] + \"-\" + sc.config['eval_counter']\n picklepath = {pickle_id: sc.config[\"eval_folder\"] + \"eval_\" + pickle_id + \".pkl\"}\nelse:\n sc.config[\"model_tag\"] = list(picklepath.keys())[0]\n \nsc.config.setdefault('cv_fold', 0)\ncv_fold = sc.config[\"cv_fold\"]\nevaluation_set_ = sc.config['evaluation_set']\nevaluation_set = f\"fold{cv_fold}-{evaluation_set_}\"\n\npicklepath_coverage_ref = sc.config[\"coverage_baseline\"]\n\nevaluation_logger = utils.EvaluationLogger(\"coverage\", sc.config, \n eval_id, eval_counter, pickle_id)\n\ndef check_dict(v):\n if isinstance(v, dict):\n return v[evaluation_set]\n else:\n return v\n\n\nresults_complete = {k: pickle.load(open(pp, 'rb')) for k, pp in picklepath.items()}\nresults_complete = {k: check_dict(v) for k, v in results_complete.items()}\nresults_complete = pd.concat([r[[\"nn\", \"mol\", \"mol_ref\", \n \"fingerprint\", \"fingerprint_ref\", \"fingerprint_ref_true\", \n \"inchikey1\", \"inchikey1_ref\",\n \"score\"]].assign(source = k)\n for k, r in results_complete.items()])\nn_total_ = len(set(results_complete[\"nn\"]))\n\nnn_in_results = set(results_complete[\"nn\"])\n\ncoverage_ref_ = pickle.load(open(picklepath_coverage_ref, 'rb'))\ncoverage_ref = coverage_ref_[evaluation_set]\ncoverage_ref = coverage_ref.loc[lambda row: row[\"mf\"] == row[\"mf_ref\"]]\ncoverage_ref = coverage_ref.loc[coverage_ref[\"nn\"].isin(nn_in_results)]\ncoverage_ref = msc.compute_candidate_scores(coverage_ref, fp_map, \n additive_smoothing_n = n_total_,\n f1_cutoff = f1_cutoff)\n\n\ncoverage_max = coverage_ref.groupby(\"nn\")[ranking_score].agg(max_score_cov = max)\n\n\nresults_ok = results_complete.loc[results_complete[\"fingerprint\"].notna()].copy()\n\n\nif remove_perfect_match:\n results_ok = results_ok.loc[\n results_ok[\"inchikey1\"] != results_ok[\"inchikey1_ref\"]]\n\n\nresults_ok = msc.compute_candidate_scores(results_ok, fp_map, \n additive_smoothing_n = n_total_,\n f1_cutoff = f1_cutoff)\nresults_ok = results_ok.merge(coverage_max, on = \"nn\", how = \"left\")\nresults_ok[\"overcount\"] = results_ok[ranking_score] > results_ok[\"max_score_cov\"]\n\nn_results_ok = len(results_ok)\nlogger.info(f\"Computing coverage for {n_results_ok} results with correct MF\")\n\nresults_coverage_ = results_ok.merge(\n coverage_ref[[\"nn\", \"inchikey1\", \"score\"]],\n how='right',\n left_on = [\"nn\", \"inchikey1\"],\n right_on = [\"nn\", \"inchikey1\"],\n suffixes = ['', '_cov']\n )\nresults_coverage = results_coverage_.groupby(\"nn\").agg(\n coverage = (\"source\", lambda ser: sum(ser.notna())),\n ratio = (\"source\", lambda ser: sum(ser.notna() / len(ser))),\n total = (\"source\", len),\n overcount = (\"overcount\", lambda ser: 0. + sum(ser)))\n\n\ncoverage_summary = coverage_ref.groupby(\"nn\").first()[[\"mol_ref\", \"fingerprint_ref_true\", \"fingerprint_ref\"]]\ncoverage_summary = msc.compute_fp_quality_mw(coverage_summary, fp_map) \ncoverage_summary = coverage_summary.join(results_coverage)\ncoverage_summary[\"rank\"] = coverage_summary[\"ratio\"].rank(ascending=False, method='first')\ncoverage_summary[\"value\"] = coverage_summary[\"ratio\"]\ncoverage_summary[\"eval_score\"] = \"coverage\"\ncoverage_summary[\"eval_metric\"] = \"coverage\"\nevaluation_logger.append_csv(\"rank\", coverage_summary)\n\n\ncoverage_summary.sort_values(\"predicted_fp_quality\", ascending=False, inplace=True)\ncoverage_summary[\"index\"] = np.arange(len(coverage_summary))\ncoverage_summary[\"coverage_sum\"] = np.cumsum(coverage_summary[\"coverage\"])\ncoverage_summary[\"total_sum\"] = np.cumsum(coverage_summary[\"total\"])\ncoverage_summary[\"running_coverage\"] = coverage_summary.apply(lambda row:\n row[\"coverage_sum\"] / row[\"total_sum\"],\n axis=1)\n \n\n\n\nplt.scatter(coverage_summary[\"predicted_fp_quality\"], coverage_summary[\"ratio\"])\n\nfig, (ax1, ax2) = plt.subplots(2, 1)\nax1.invert_xaxis()\nax1.plot(\n coverage_summary[\"predicted_fp_quality\"],\n np.cumsum(coverage_summary[\"total\"])\n )\nax1.plot(\n coverage_summary[\"predicted_fp_quality\"],\n np.cumsum(coverage_summary[\"coverage\"])\n )\nax2.invert_xaxis()\nax2.scatter(\n coverage_summary[\"predicted_fp_quality\"],\n coverage_summary[\"ratio\"],\n )\nax2.plot(\n coverage_summary[\"predicted_fp_quality\"],\n coverage_summary[\"running_coverage\"],\n )\nax2.plot(\n coverage_summary.rolling(window=10)[\"predicted_fp_quality\"].mean(),\n coverage_summary.rolling(window=10)[\"ratio\"].mean())\n\n\ncoverage_summary.sort_values(\"mol_weight\", ascending=True, inplace=True)\ncoverage_summary[\"index\"] = np.arange(len(coverage_summary))\ncoverage_summary[\"coverage_sum\"] = np.cumsum(coverage_summary[\"coverage\"])\ncoverage_summary[\"total_sum\"] = np.cumsum(coverage_summary[\"total\"])\ncoverage_summary[\"running_coverage\"] = coverage_summary.apply(lambda row:\n row[\"coverage_sum\"] / row[\"total_sum\"],\n axis=1)\n\n\n\nfig, (ax1, ax2) = plt.subplots(2, 1)\nax1.plot(\n coverage_summary[\"mol_weight\"],\n np.cumsum(coverage_summary[\"total\"])\n )\nax1.plot(\n coverage_summary[\"mol_weight\"],\n np.cumsum(coverage_summary[\"coverage\"])\n )\nax2.scatter(\n coverage_summary[\"mol_weight\"],\n coverage_summary[\"ratio\"],\n )\nax2.plot(\n coverage_summary[\"mol_weight\"],\n coverage_summary[\"running_coverage\"],\n )\nax2.plot(\n coverage_summary.rolling(window=10)[\"mol_weight\"].mean(),\n coverage_summary.rolling(window=10)[\"ratio\"].mean())\n\n\n\n\n","repo_name":"meowcat/MSNovelist","sub_path":"evaluation/top_rediscovery.py","file_name":"top_rediscovery.py","file_ext":"py","file_size_in_byte":7250,"program_lang":"python","lang":"en","doc_type":"code","stars":51,"dataset":"github-code","pt":"79"} +{"seq_id":"13026036709","text":"#!/usr/bin/python -tt\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport matplotlib.cm as cm\n\ntestdata = pd.read_csv('mnist-test-labeled.csv')\ntraindata = pd.read_csv('train.csv')\n\n#Print the average digit images for the two datasets\ndef extract_digit(n,dset):\n dic = {}\n for index, row in dset.iterrows():\n key = row['label']\n if key not in dic:\n dic[key] = []\n dic[key].append(row)\n img = pd.DataFrame(dic[n])\n mean = img.describe().loc['mean'].values\n return mean\n\ndef transfer(digit_pixel):\n img_matrix = np.zeros((28,28))\n for i in range(0,27):\n for j in range (0,27):\n index = i * 28 + j\n img_matrix[i][j] =digit_pixel[index+1]\n return img_matrix\n\n\ndef displayTrain(digit,dset):\n mean = extract_digit(digit,dset)\n img = transfer(mean)\n plt.imshow(img,cmap=cm.binary)\n fname = 'trainAvg' + str(digit)+'.png'\n plt.savefig(fname)\n\ndef displayTest(digit,dset):\n mean = extract_digit(digit,dset)\n img = transfer(mean)\n plt.imshow(img,cmap=cm.binary)\n fname = 'testAvg' + str(digit)+'.png'\n plt.savefig(fname)\n \nfor i in range(10):\n displayTrain(i,traindata)\n displayTest(i,testdata)","repo_name":"ujkharel/Kaggle-Digit-Recognizer-Competition","sub_path":"AveragePic.py","file_name":"AveragePic.py","file_ext":"py","file_size_in_byte":1231,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6781024010","text":"def reverseParentheses(strr, lenn):\n st = []\n for i in range(lenn):\n \n # Push the index of the current\n # opening bracket\n if (strr[i] == '('):\n st.append(i)\n \n # Reverse the substring starting\n # after the last encountered opening\n # bracket till the current character\n elif (strr[i] == ')'):\n temp = strr[st[-1]:i + 1]\n strr = strr[:st[-1]] + temp[::-1] + \\\n strr[i + 1:]\n del st[-1]\n \n # To store the modified string\n res = \"\"\n for i in range(lenn):\n if (strr[i] != ')' and strr[i] != '('):\n res += strr[i]\n return res\n \n \n\ndef solution(inputString):\n lenn = len(inputString)\n \n return reverseParentheses(inputString, lenn)\n\nstring = 'foo(bar(baz))blim'\nprint(solution(string))\n","repo_name":"aoamusat/codes","sub_path":"py/reverse_in_parentheses.py","file_name":"reverse_in_parentheses.py","file_ext":"py","file_size_in_byte":852,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"1768182203","text":"import os\nimport urllib.parse\n\nfrom jobserv_runner.handlers.simple import HandlerError, SimpleHandler\n\n\nclass GitPoller(SimpleHandler):\n def _get_http_clone_token(self, clone_url):\n secrets = self.rundef.get('secrets', {})\n if clone_url.startswith('https://github.com'):\n tok = secrets.get('githubtok')\n if tok:\n return tok\n\n # we can't determine by URL if its a gitlab repo, so just assume\n # the rundef/secrets are done sanely by the user\n env = self.rundef['env']\n user = env.get('gitlabuser') or secrets.get('gitlabuser')\n if user:\n token = self.rundef['secrets']['gitlabtok']\n return user + ':' + token\n\n def _clone(self, log, dst):\n clone_url = self.rundef['env']['GIT_URL']\n log.info('Clone_url: %s', clone_url)\n\n token = self._get_http_clone_token(clone_url)\n if token:\n log.info('Using an HTTP token for cloning')\n p = urllib.parse.urlsplit(clone_url)\n clone_url = p.scheme + '://' + token + '@' + p.netloc + p.path\n\n if not log.exec(['git', 'clone', clone_url, dst]):\n raise HandlerError(\n 'Unable to clone: ' + self.rundef['env']['GIT_URL'])\n\n sha = self.rundef['env'].get('GIT_SHA')\n if sha:\n log.info('Checking out: %s', sha)\n if not log.exec(['git', 'branch', 'jobserv-run', sha], cwd=dst):\n raise HandlerError('Unable to branch: ' + sha)\n if not log.exec(['git', 'checkout', 'jobserv-run'], cwd=dst):\n raise HandlerError('Unable to checkout: ' + sha)\n\n def prepare_mounts(self):\n mounts = super().prepare_mounts()\n\n repo_dir = os.path.join(self.run_dir, 'repo')\n with self.log_context('Cloning git repository') as log:\n self._clone(log, repo_dir)\n mounts.append((repo_dir, '/repo'))\n self.container_cwd = '/repo'\n return mounts\n\n\nhandler = GitPoller\n","repo_name":"ursinha/jobserv","sub_path":"runner/jobserv_runner/handlers/git_poller.py","file_name":"git_poller.py","file_ext":"py","file_size_in_byte":2006,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"27518250309","text":"from django.http import HttpResponse, HttpResponseNotFound\nfrom django.views.decorators.csrf import csrf_exempt\nfrom server.database import get_poi_coords, get_poi_types\nfrom server.config import route_poi\nimport json,logging,ast\nfrom collections import defaultdict\nimport cPickle as pic\n\n@csrf_exempt\ndef get_coords(request):\n with open ('server/poi_coord_route.py') as f:\n lines = f.read().splitlines()\n \n route_poi = map(int, lines)\n tags = request.POST.getlist(\"tags\")\n available_tags = get_poi_types()\n\n for tag in tags:\n if tag not in available_tags:\n return HttpResponseNotFound(\"One of your tags does not exist.\")\n\n return HttpResponse(json.dumps({\"coords\": get_poi_coords(tags,route_poi)}))\n\n\n@csrf_exempt\ndef get_types(request):\n return HttpResponse(json.dumps({\"types\": get_poi_types()}))\n\n@csrf_exempt\ndef save_coords(request):\n try:\n ele = request.POST.getlist(\"elements\")\n name = str(request.POST.get(\"name\"))\n if name.isspace():\n logging.error('saving POis failed since there is no name or there is space in the name',exc_info=True)\n return HttpResponse(\"Json not Saved\")\n\n else:\n filename = \"/home/kthiruko/newbackend_rust/data/poisets/%s.json\" %name\n \n ## I have no idea whyy we get 2 square brackets while writing to a json file so did some stuff to make it right\n json_data = {\"elements\" : map (ast.literal_eval , ele) , \"name\": name}\n struct_element = map (ast.literal_eval , ele)\n d = defaultdict(list)\n d[\"elements\"] = struct_element\n d[\"name\"]=name\n json_d = json.dumps(d)\n data = json.loads(json_d)\n temp = data[\"elements\"]\n data[\"elements\"] = temp[0]\n # what1 = json.dumps(what)\n # print what1\n\n with open(filename, 'w') as outfile:\n json.dump(data,outfile)\n \n return HttpResponse(\"Json Saved\")\n \n except:\n logging.error('saving POis failed',exc_info=True)\n return HttpResponse(\"Json not Saved\")","repo_name":"oSoc17/lopeningent_backend","sub_path":"server/server/interface/pois.py","file_name":"pois.py","file_ext":"py","file_size_in_byte":2050,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"12106839337","text":"from tkinter import *\r\nfrom PIL import ImageTk,Image\r\nfrom random import randint\r\nimport random\r\n\r\nroot = Tk()\r\nroot.title(\"flashcards\")\r\nroot.iconbitmap(\"icons/superman.ico\")\r\nroot.geometry(\"600x600\")\r\n\r\n#creating a menubar\r\nmy_menu = Menu(root)\r\nroot.config(menu=my_menu)\r\n\r\nresult=0\r\n\r\n#create randpm chnage function\r\ndef change():\r\n \r\n global list1\r\n global rando\r\n #creating a list\r\n list1 = ['ntr_np','ntr_tem','ntr_jlk','prabhas','mahesh_sri',\r\n 'mahesh_ban','mahesh_maha']\r\n \r\n #creating a random int\r\n rando = randint(0,len(list1)-1)\r\n states = \"angel/\" + list1[rando] + \".png\"\r\n \r\n #inserting our images\r\n global my_label\r\n global my_pic\r\n my_pic = ImageTk.PhotoImage(Image.open(states))\r\n my_label.config(image=my_pic,bg=\"white\")\r\n \r\n \r\n#calling submitt button\r\n\r\ndef submit():\r\n global answer\r\n answer = e1.get().lower()\r\n if answer == list1[rando]:\r\n \r\n result = \"correct it is\" +\" \" + list1[rando]\r\n else:\r\n result = \"incorrect it is not correct answer\"\r\n my_label1 = Label(hero_frame,text=result,font=\"arial,25\")\r\n my_label1.pack(pady=15)\r\n e1.delete(0,END)\r\n#calling functions\r\ndef heroes():\r\n global my_label\r\n removeall()\r\n hero_frame.pack(fill=BOTH,expand=1)\r\n my_label = Label(hero_frame)\r\n my_label.pack(pady=15)\r\n change()\r\n \r\n #creating a entrybox\r\n global e1\r\n e1 = Entry(hero_frame,width=30)\r\n e1.pack()\r\n #creating a next button\r\n button1 = Button(hero_frame,text=\"next\",command=heroes)\r\n button1.pack(pady=15)\r\n \r\n #button to show mwssage\r\n button2 = Button(hero_frame,text=\"submit\",command=submit)\r\n button2.pack()\r\n \r\n#calling select function\r\ndef select():\r\n if radio_cap.get()==names[answer]:\r\n result=\"correct\"\r\n answerlabel.config(text=result)\r\n\r\ndef movies():\r\n removeall()\r\n state_frame_movies.pack(fill=BOTH,expand=1)\r\n my_label = Label(state_frame_movies)\r\n my_label.pack()\r\n global list1\r\n global rando\r\n global names\r\n global answer\r\n list1 = ['ntr_np','ntr_tem','ntr_jlk','prabhas','mahesh_sri',\r\n 'mahesh_ban','mahesh_maha']\r\n \r\n names = {\"ntr_np\":'nanakuprematho',\"ntr_tem\":'temper',\"ntr_jlk\":'jailavakusa',\r\n \"prabhas\":'saaho',\"mahesh_sri\":'srimanthudu',\"mahesh_ban\":'bharath ane nenu',\"mahesh_maha\":'maharshi'}\r\n \r\n #creating a random int\r\n \r\n count=1\r\n answer_list = []\r\n while count<4:\r\n \r\n rando = randint(0,len(list1)-1)\r\n \r\n if count == 1:\r\n \r\n \r\n answer = list1[rando]\r\n global my_pic\r\n states = \"angel/\" + list1[rando] + \".png\"\r\n my_pic = ImageTk.PhotoImage(Image.open(states))\r\n my_label.config(image=my_pic)\r\n \r\n \r\n answer_list.append(list1[rando])\r\n list1.remove(list1[rando])\r\n random.shuffle(list1)\r\n count+=1\r\n \r\n global radio_cap\r\n radio_cap=StringVar()\r\n radio_cap.set(names[answer_list[0]])\r\n \r\n radio_button1 = Radiobutton(state_frame_movies,text=names[answer_list[0]],variable=radio_cap,value=names[answer_list[0]]).pack()\r\n radio_button2 = Radiobutton(state_frame_movies,text=names[answer_list[1]],variable=radio_cap,value=names[answer_list[1]]).pack()\r\n radio_button3 = Radiobutton(state_frame_movies,text=names[answer_list[2]],variable=radio_cap,value=names[answer_list[2]]).pack()\r\n \r\n #creating a next button\r\n \r\n nxt_button = Button(state_frame_movies,text=\"next\",command=movies)\r\n nxt_button.pack()\r\n \r\n #creating a select button\r\n \r\n btselect = Button(state_frame_movies,text=\"select\",font=\"arial,20\",command=select)\r\n btselect.pack(pady=15)\r\n \r\n #creating answer label\r\n global answerlabel\r\n answerlabel = Label(state_frame_movies,text=\" \",font=\"arial,20\")\r\n answerlabel.pack(pady=15)\r\n \r\n\r\n#removing framses and deleting childrens \r\ndef removeall():\r\n #looping through for deleting frame children\r\n for widget in hero_frame.winfo_children():\r\n widget.destroy()\r\n for widget in state_frame_movies.winfo_children():\r\n widget.destroy()\r\n \r\n hero_frame.pack_forget()\r\n state_frame_movies.pack_forget()\r\n \r\n#creating submenu titles\r\ntollywood_menu = Menu(my_menu)\r\nmy_menu.add_separator()\r\nmy_menu.add_cascade(label=\"tollywood\",menu=tollywood_menu)\r\ntollywood_menu.add_command(label=\"heroes\",command=heroes)\r\ntollywood_menu.add_command(label=\"movies\",command=movies)\r\ntollywood_menu.add_command(label=\"exit\",command=root.destroy)\r\n\r\nhero_frame = Frame(root,width=500,height=500,bg=\"white\")\r\nstate_frame_movies = Frame(root,width=500,height=500,bg=\"yellow\")\r\n\r\n\r\n\r\nroot.mainloop()\r\n\r\n\r\n","repo_name":"vishnuvardhan1807/tkinter-projects","sub_path":"flashcaeds.py","file_name":"flashcaeds.py","file_ext":"py","file_size_in_byte":4777,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"20964646350","text":"import os\n#os.environ[\"CUDA_VISIBLE_DEVICES\"]=\"-1\"\nimport tensorflow as tf\nimport numpy as np\n#from PIL import Image\n#import skimage\n#import tensorflow_datasets as tfds\n#import matplotlib.pyplot as plt\nfrom tensorflow.python.keras import backend as K\nfrom tqdm import tqdm\nimport pathlib\n\nBATCH_SIZE = 32\nIMG_HEIGHT = 448\nIMG_WIDTH = 448\nNUM_CLASSES = 200\n\n\nclass DataSet:\n\n def __init__(self, path_root):\n self.data_dir = pathlib.Path(path_root + '/CUB_200_2011/CUB_200_2011/images')\n self.image_path = path_root + \"/CUB_200_2011/CUB_200_2011/images/\"\n self.image_name_path = path_root + \"/CUB_200_2011/CUB_200_2011/images.txt\"\n self.semantics_path2 = path_root + \"/CUB_200_2011/CUB_200_2011/attributes/image_attribute_labels.txt\"\n self.semantics_path1 = path_root + \"/CUB_200_2011/attributes.txt\"\n self.split_path = path_root + \"/CUB_200_2011/CUB_200_2011/train_test_split.txt\"\n self.class_path = path_root + \"/CUB_200_2011/CUB_200_2011/classes.txt\"\n self.label_path = path_root + \"/CUB_200_2011/CUB_200_2011/image_class_labels.txt\"\n self.AUTOTUNE = tf.data.experimental.AUTOTUNE\n\n def load(self, GPU=True, train=True, batch_size=32):#discard\n index = self.get_split()\n if GPU:\n n = len(index)\n else:\n n = 50\n if train:\n #phi = self.get_phi(index)# Φ, semantic matrix, 28*200\n labels = self.get_label(n, index, set=0)\n images = self.get_image(n, index, set=0)\n else:\n labels = self.get_label(n, index, set=1)\n images = self.get_image(n, index, set=1)\n #phi = self.get_semantic(n, index, set=1) # φ, semantic features 28, n\n\n ds = tf.data.Dataset.from_tensor_slices((images, np.asarray(labels))).cache().shuffle(50).batch(batch_size).prefetch(tf.data.experimental.AUTOTUNE)\n\n return ds\n\n def prepare_for_training(self, ds, batch_size=32, cache=True):\n # This is a small dataset, only load it once, and keep it in memory.\n # use `.cache(filename)` to cache preprocessing work for datasets that don't\n # fit in memory.\n \"\"\"\n if cache:\n if isinstance(cache, str):\n ds = ds.cache(cache)\n else:\n ds = ds.cache()\n \"\"\"\n cache_dir = os.path.join(os.getcwd(), 'cache_dir')\n try:\n os.makedirs(cache_dir)\n except OSError:\n print('Cache directory already exists')\n cached = ds.cache(os.path.join(cache_dir, 'cache.temp'))\n ds = ds.shuffle(50).repeat().batch(batch_size).prefetch(tf.data.experimental.AUTOTUNE)\n\n return ds\n\n def get_label(self, n, index, set=0):\n file = open(self.label_path, \"r\")\n labels = file.readlines()\n label_new = []\n for i in range(n):\n if index[i] == set:\n label_new.append(int(labels[i].split(' ')[1].split('\\n')[0]) - 1)# start from 0\n\n return label_new\n\n def decode_img(self, img):\n # convert the compressed string to a 3D uint8 tensor\n img = tf.image.decode_jpeg(img, channels=3)\n # Use `convert_image_dtype` to convert to floats in the [0,1] range.\n img = tf.image.convert_image_dtype(img, tf.float32)\n # resize the image to the desired size.\n return tf.image.resize(img, [IMG_WIDTH, IMG_HEIGHT])\n\n def get_image(self, n, index, set=0):# discard\n images_names = open(self.image_name_path, \"r\")\n images = images_names.readlines()\n print(\"loading images...\")\n image_new = []\n for i in tqdm(range(n)):\n if index[i] == set:\n im_path = self.image_path + images[i].split(' ')[1].split('\\n')[0]\n #img = np.asarray(Image.open(im_path).resize((IMG_WIDTH, IMG_HEIGHT)), dtype=np.float32)\n img = tf.io.read_file(im_path)\n img = self.decode_img(img)\n # image = tf.keras.preprocessing.image.load_img(im_path, target_size=(IMG_WIDTH, IMG_HEIGHT))\n image_new.append(img)\n else:\n pass\n\n return image_new\n\n def get_attribute(self):\n file = open(self.semantics_path1, \"r\")\n lines = file.readlines()\n attributes = {}\n print(\"loading attributes...\")\n for line in lines:\n id = line.split(\" \")[0] # No. of attribute, 28 categories, 312 in total\n info = line.split(\" \")[1].split(\"::\")\n if info[0] in attributes.keys():\n attributes[info[0]] += [int(id)]\n else:\n attributes[info[0]] = [int(id)]\n\n return attributes\n\n def get_semantic(self, n, index, set=0, file_path=None):\n attributes = self.get_attribute()\n n_att = len(attributes.keys()) # 28\n birds_at = {}\n print(\"loading semantics...\")\n file = open(self.semantics_path2, \"r\")\n lines = file.readlines()\n for line in lines:\n id_bird = line.split(\" \")[0]\n if id_bird not in birds_at.keys():\n birds_at[id_bird] = np.zeros(n_att)\n\n id_att = int(line.split(\" \")[1])\n present = int(line.split(\" \")[2])\n if present:\n for i, key in enumerate(attributes.keys()):\n if id_att in attributes[key]:\n birds_at[id_bird][i] += np.where(np.array(attributes[key]) == id_att)[0][0]\n\n birds_semantics = [] # 11788*28 list\n for i, key in enumerate(birds_at.keys()):\n if i < n:\n if index[i] == set:\n birds_semantics.append(birds_at[key])\n else:\n pass\n else:\n break\n print(\"Finished!\")\n\n return np.asarray(birds_semantics)\n\n def get_split(self, index=True):\n file = open(self.split_path, \"r\")\n ids = file.readlines()\n if index:\n for i in range(len(ids)):#len(set)):\n ids[i] = int(ids[i].split(' ')[1].split('\\n')[0])\n return ids\n else:\n images_names = open(self.image_name_path, \"r\")\n images = images_names.readlines()\n print(\"splitting...\")\n train_list = []\n test_list = []\n for i in range(len(ids)):#len(set)):\n set = int(ids[i].split(' ')[1].split('\\n')[0])\n if set == 0:\n train_list.append(self.image_path + images[i].split(' ')[1].split('\\n')[0])\n else:\n test_list.append(self.image_path + images[i].split(' ')[1].split('\\n')[0])\n\n return tf.data.Dataset.from_tensor_slices(train_list), tf.data.Dataset.from_tensor_slices(test_list)#.cache()\n\n def get_phi(self):\n index = self.get_split(index=True)\n labels = self.get_label(len(index), index, set=0)\n semantics = self.get_semantic(len(index), index, set=0)\n phi = np.zeros((semantics[0].shape[0], max(labels)+1))\n lcount = {x:labels.count(x) for x in labels}\n for i in range(len(semantics)):\n phi[:, labels[i]] += semantics[i]\n for j in range(phi.shape[0]):\n phi[:, j] = phi[:, j] / lcount[j]\n\n return tf.convert_to_tensor(phi, dtype=tf.float32)\n\n def process_path(self, file_path):\n parts = tf.strings.split(file_path, '/')\n # The second to last is the class-directory\n label = int(tf.strings.split(parts[-2], '.')[0])# == self.CLASS_NAMES\n # load the raw data from the file as a string\n img = tf.io.read_file(file_path)\n img = self.decode_img(img)\n\n return img, label\n\n def load_gpu(self, batch_size=32):#autotune=4\n # Set `num_parallel_calls` so multiple images are loaded/processed in parallel.\n self.CLASS_NAMES = np.unique(\n np.array([item.name for item in self.data_dir.glob('[!.]*') if item.name != \"LICENSE.txt\"]))\n train_list_ds, test_list_ds = self.get_split(index=False)\n #dataset = train_list_ds.interleave(tf.data.TFRecordDataset, cycle_length=FLAGS.num_parallel_reads, num_parallel_calls=tf.data.experimental.AUTOTUNE)\n train_ds = train_list_ds.map(self.process_path, num_parallel_calls=self.AUTOTUNE)\n test_ds = test_list_ds.map(self.process_path, num_parallel_calls=self.AUTOTUNE)\n train = self.prepare_for_training(train_ds, batch_size)\n test = self.prepare_for_training(test_ds, batch_size)\n for image, label in train.take(1):\n print(\"Image shape: \", image.numpy().shape)\n print(\"Label: \", label.numpy())\n\n return train, test\n\n def loadtfds(self, dataset_name, batch_size=32): #not working\n # Load data from tensorflow_datasets\n raw_train, raw_test = tfds.load(name=dataset_name, split=[\"train\", \"test\"], batch_size=32)\n train = raw_train.map(lambda x: tf.image.resize(x['image'], (IMG_WIDTH, IMG_HEIGHT)))\n test = raw_test.map(lambda x: tf.image.resize(x['image'], (IMG_WIDTH, IMG_HEIGHT)))\n ds_train = train.shuffle(1000).repeat().batch(batch_size).prefetch(tf.data.experimental.AUTOTUNE)\n ds_test = test.shuffle(1000).repeat().batch(batch_size).prefetch(tf.data.experimental.AUTOTUNE)\n # for batch in ds_train:\n # ...\n return ds_train, ds_test\n\n\nif __name__ == '__main__':\n print(\"Num GPUs Available: \", len(tf.config.experimental.list_physical_devices('GPU')))\n path_root = os.path.abspath(os.path.dirname(__file__)) # '/content/gdrive/My Drive/data'\n bird_data = DataSet(path_root)\n #train_ds = bird_data.load(GPU=True, train=True, batch_size=32)\n #ds_train, ds_test = bird_data.loadtfds('caltech_birds2011')\n ds_train, ds_test = bird_data.load_gpu(batch_size=4)\n \"\"\"\n filename1 = 'train_ds.tfrecord'\n writer1 = tf.data.experimental.TFRecordWriter(filename1)\n writer1.write(train_ds)\n #read\n #raw_dataset = tf.data.TFRecordDataset(filenames)\n \"\"\"\n #image_batch, label_batch = next(iter(ds_train))\n\n","repo_name":"LindsayXX/DD2412_project","sub_path":"final_model/dataloaderGPUs.py","file_name":"dataloaderGPUs.py","file_ext":"py","file_size_in_byte":10064,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"13775607899","text":"import random\nimport string\n\nfrom util.base_augmentation_script import read_original_dataset, output_to_file\n\n\ndef replace_n_chars(n):\n ori_data = read_original_dataset()\n new_dataset = {}\n for command, intent in ori_data.items():\n length = len(command)\n for _ in range(n):\n pos = random.randint(0, length)\n change_to = random.choice(string.ascii_lowercase)\n command = command[:pos] + change_to + command[pos + 1:]\n new_dataset[command] = intent\n return new_dataset\n\n\nfor k in range(1, 10):\n data = replace_n_chars(k)\n output_to_file(data, f'./typo_dataset/data_with_{k}_typo.json')\n","repo_name":"SongyuWang-UO/VA_project","sub_path":"typo_dataset/argumentation_script.py","file_name":"argumentation_script.py","file_ext":"py","file_size_in_byte":655,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"71863355776","text":"#Variables\nkernel_file='Luxur'\nenv=Environment()\nenv['CFLAGS']=' -Wall -fstrength-reduce -fomit-frame-pointer\\\n -finline-functions -nostdinc -fno-builtin\\\n -fno-stack-protector'\nenv['CCCOMSTR']='[C] $SOURCES'\nenv['ASCOMSTR']='[as] $SOURCES'\nenv['LINKCOM']='ld -T link.ld -o $TARGET $SOURCES'\nenv['LINKCOMSTR']='[LD] $SOURCES > $TARGET'\n\n\ndebug = ARGUMENTS.get('debug', 0)\nif int(debug):\n\tenv.Append(CCFLAGS = ' -g')\n\tenv['CCCOMSTR']='[C] [DBG] $SOURCES'\n\tenv['ASCOMSTR']='[AS][DBG] $SOURCES'\n\nlnkd = Builder(action = 'ld -T link.ld -o $TARGET $SOURCES')\nenv.Append(BUILDERS = {'Linkit' : lnkd})\n\nExport('env')\nobjs=(SConscript(['boot/SConscript',\n 'kernel/SConscript',\n\t\t 'lib/SConscript'],'env'))\n\n\n#linkeamos los Objetos en el kernel\nenv.Program(kernel_file,objs)\n","repo_name":"M4c0t/IndiaOS","sub_path":"LICENSE.md/SConstruct","file_name":"SConstruct","file_ext":"","file_size_in_byte":786,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"79"} +{"seq_id":"9724489256","text":"import math\nimport pandas as pd\nimport simplepbr\nfrom direct.showbase.ShowBase import ShowBase\nfrom panda3d.core import *\n\nIMAGE_WIDTH = 640\nIMAGE_HEIGHT = 480\nCAMERA_FOV = (math.pi/180)*60\n\n\n\n\n\nclass MyApp(ShowBase):\n\t\n\tdef __init__(self):\n\t\t\n\t\tShowBase.__init__(self)\n\t\t#simplepbr.init()\n\t\t#myMaterial = Material()\n\t\t#myMaterial.setAmbient((1, 1, 1, 1)) # Make this material blue\n\t\talight = AmbientLight('alight')\n\t\talight.setColor((1.0, 1.0, 1.0, 1))\n\t\t\n\t\t\n\t\t# Load the environment model.\n\n\t\tself.scene = self.loader.loadModel(\"part.bam\")\n\t\tself.scene.setColor(1.0, 1.0, 1.0, 1.0)\n\t\talnp = self.scene.attachNewNode(alight)\n\t\tself.scene.setLight(alnp)\n\t\t#self.scene.setMaterial(myMaterial)\n\t\t\n\t\t# Reparent the model to render.\n\n\t\tself.scene.reparentTo(self.render)\n\n\t\t# Apply scale and position transforms on the model.\n\n\t\tself.scene.setScale(0.25, 0.25, 0.25)\n\n\t\tself.scene.setPos(-8, 42, 0)\n\n\n'''Returns a Pandas DataFrame with predefined fields'''\ndef init_df():\n\tdata = pd.DataFrame(\n\t\t\t{'filename': pd.Series(dtype='str'), 'class': pd.Series(dtype='str'), 'id': pd.Series(dtype='int'), 'cam_pos': pd.Series(dtype='str'),\n\t\t\t 'cam_axis': pd.Series(dtype='str'), 'part_pos': pd.Series(dtype='str')} )\n\treturn data\n\t\t\n\nif __name__ == \"__main__\":\n\tdata = init_df()\n\tdata.loc[len(data.index)] = ['test', 'part', 0, '[3;3;-3]', '[-3;-3;3]', '[0;0;0]']\n\t\n\tapp = MyApp()\n\tapp.run()\n","repo_name":"OzzyP97/KONE.533-Automated-Data-Generation-Setup","sub_path":"demo.py","file_name":"demo.py","file_ext":"py","file_size_in_byte":1383,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"20532494993","text":"import logging\nfrom sqlalchemy.orm import Session\n\nfrom app.models import models\nfrom app import schemas\n\nfrom datetime import datetime\n\nlogging.basicConfig(\n level=logging.DEBUG,\n format=\"%(asctime)s %(message)s\",\n handlers=[\n logging.FileHandler(\"/tmp/VideoConfLog.log\"),\n logging.StreamHandler()\n ]\n)\nlogger = logging.getLogger(\"VideoConfLog\")\n\n\ndef create_conference(db: Session):\n db_item = models.VideoConference()\n db.add(db_item)\n db.commit()\n\n return db_item\n","repo_name":"adamteale/video_conference_example","sub_path":"backend/app/app/crud.py","file_name":"crud.py","file_ext":"py","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"79"} +{"seq_id":"34966962370","text":"from collections import deque\n\nn, m, start = map(int, input().split())\nvertexList = [ [] for _ in range(n+1) ]\ncheck = [False]*(n+1)\n\nfor _ in range(m):\n u, v = map(int, input().split())\n vertexList[u].append(v)\n vertexList[v].append(u)\nfor i in range(1, n+1):\n vertexList[i].sort() \n\ndef dfs(x):\n global check\n check[x] = True\n print(x, end=' ')\n for adjacentVertex in vertexList[x]:\n if check[adjacentVertex] == False:\n dfs(adjacentVertex)\n\ndef bfs(x):\n check = [False]*(n+1)\n q = deque()\n q.append(x)\n check[x] = True\n while q:\n next_vertex = q.popleft()\n print(next_vertex, end=' ')\n for adjacentVertex in vertexList[next_vertex]:\n if check[adjacentVertex] == False:\n check[adjacentVertex] = True\n q.append(adjacentVertex)\n\ndfs(start)\nprint()\nbfs(start)","repo_name":"kizarrd/algorithm_with_python","sub_path":"dfs_and_bfs/acmicpc1260.py","file_name":"acmicpc1260.py","file_ext":"py","file_size_in_byte":877,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"41923353643","text":"from flask import Flask, request, send_from_directory, send_file, make_response\nfrom flask_cors import CORS\nfrom time import perf_counter\n\nfrom compress.lib import compressImage\n\napp = Flask(__name__,\n static_folder=None)\n\ncors = CORS(app,\n resources={\n r'/api/*' : {\n 'origins' : '*',\n 'expose_headers': ['Compress-Time']\n }})\n\n@app.route('/api/compress', methods=['POST'])\ndef compress_route():\n try:\n file = request.files['file']\n ratio = int(request.form['rate'])\n\n startTime = perf_counter()\n result = compressImage(file, ratio)\n endTime = perf_counter()\n\n response = make_response(send_file(result, mimetype=file.mimetype))\n response.headers['Compress-Time'] = endTime - startTime\n return response\n except:\n return '', 501\n\n@app.route('/', defaults={ 'path': 'index.html' })\n@app.route('/')\ndef catch_all(path):\n return send_from_directory('./client/dist', path)\n","repo_name":"hilyafadhilah/Algeo02-20017","sub_path":"src/run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":1038,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8874161521","text":"import unittest\n\nimport numpy as np\nimport torch\nfrom torch.nn import Conv2d, Module, ReLU, MaxPool2d, Linear, LogSoftmax\n\nfrom torch_pp import StandardScaler\nfrom torch_pp.minmaxscaler import MinMaxScaler\n\n\nclass Cnn(Module):\n def __int__(self, num_channels: int, classes: int):\n super(self).__init__()\n self.scaler = StandardScaler()\n self.conv1 = Conv2d(in_channels=num_channels, out_channels=20, kernel_size=(5, 5))\n self.relu1 = ReLU()\n self.maxpool1 = MaxPool2d(kernel_size=(2, 2), stride=(2, 2))\n self.fc2 = Linear(in_features=500, out_features=classes)\n self.logSoftmax = LogSoftmax(dim=1)\n\n def forward(self, x):\n x = self.scaler.fit_transform(x)\n x = self.conv1(x)\n x = self.relu1(x)\n x = self.maxpool1(x)\n x = self.fc2(x)\n output = self.logSoftmax(x)\n return output\n\n\nclass TestScalers(unittest.TestCase):\n def test_standard_scaler_transform(self):\n input_x = torch.from_numpy(np.array([[20., 1.], [-3., 700.], [-11., 3.]])).to(dtype=torch.double)\n scaler = StandardScaler()\n transformed_x = scaler.fit_transform(input_x)\n transformed_back_x = scaler.inverse_transform(transformed_x)\n torch.testing.assert_close(input_x, transformed_back_x)\n\n def test_minmax_scaler_transform(self):\n input_x = torch.from_numpy(np.array([[20., 1.], [-3., 700.], [-11., 3.]])).to(dtype=torch.float64)\n scaler = MinMaxScaler()\n transformed_x = scaler.fit_transform(input_x)\n transformed_back_x = scaler.inverse_transform(transformed_x)\n torch.testing.assert_close(input_x, transformed_back_x)\n\n\nif __name__ == '__main__':\n unittest.main()\n","repo_name":"Pravuk/torch-preprocessing","sub_path":"torch_pp/test_.py","file_name":"test_.py","file_ext":"py","file_size_in_byte":1714,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39488664230","text":"#!/usr/bin/env python\n\"\"\"\nNAME:\n\n hashjoin: join a small lookup table with unsorted targets\n\nDESCRIPTION:\n\nI wanted a quick way to annotate lines with additional metadata\nfrom a lookup file. I had a set of UUIDS and human readable\nlabels, and a set of target files that would be decorated by combining\nthe input line with the lookup value when the lookup key was present\nin the line.\n\nsample lookup: key value\n\nid1 Item 1\nid2 Another item\n\ntarget:\n\nThis is id1 and other stuff\nAnd here id2 is here\n\n\nThe join linux join command requires sorted input.\nHowever, given a small lookup table hashjoin allows for\njoining of unsorted lookup and target table(s).\n\nhashjoin.py uses the first field of the lookup as the key,\nand then \n\nAUTHOR:\n\n Jud Dagnall \n\nEXAMPLES:\n\n # common usage:\n hashjoin \n\n\"\"\"\n\nfrom __future__ import print_function\n\nimport argparse\nimport fileinput\nimport json\nimport logging\nimport re\nimport sys\n\nTIMESTAMP_FORMAT='%(asctime)s %(levelname)s - %(message)s'\n\ndef parse_args(args=None):\n desc=\"\"\n p = argparse.ArgumentParser(description=desc)\n #p.add_argument('', help=\"default: %(default)s\", default='')\n p.add_argument('-d', '--delimiter', \n help=\"regex delimiter for lookup table\",\n default='\\s+')\n p.add_argument('-D', '--debug', action='store_true',\n help='enable debugging')\n p.add_argument('-o', '--only', action='store_true', \n help=\"print only lines with matches\")\n p.add_argument('-O', '--output-delimiter', \n help=\"output delimiter. default = space\",\n default=\" \")\n p.add_argument('-T', '--tab-output', action='store_true',\n help=\"use tab as the output delimiter\")\n p.add_argument('lookup', help='whitespace delimited lookup')\n p.add_argument('targets', nargs='*', \n help='targets for lookup')\n\n \n # accept arguments as a param, so we\n # can import and run this module with a commandline-like\n # syntax.\n if args is None: \n args = sys.argv[1:]\n return p.parse_args(args)\n\ndef run(opts):\n logging.debug(\"starting\")\n lookups = {}\n if opts.tab_output:\n opts.output_delimiter = \"\\t\"\n\n delimiter = re.compile(opts.delimiter)\n\n for line in open(opts.lookup):\n k, v = re.split(delimiter, line.rstrip(\"\\n\"), 1)\n lookups[k] = v\n logging.debug('\"%s\" -> \"%s\"', k, v)\n\n logging.debug(\"lookups: %s\", lookups)\n for line in fileinput.input(opts.targets):\n line = line.rstrip(\"\\n\")\n matched = False\n for pattern, extra in lookups.items():\n if pattern in line:\n line = line + opts.output_delimiter + extra\n matched = True\n logging.debug('matched %s', pattern)\n break\n if matched or not opts.only:\n print(line)\n\nif __name__ == '__main__':\n opts = parse_args(sys.argv[1:])\n debug_level = logging.DEBUG if opts.debug else logging.INFO\n logging.basicConfig(level=debug_level ,format=TIMESTAMP_FORMAT)\n run(opts)\n","repo_name":"thejud/scripts","sub_path":"bin/hashjoin.py","file_name":"hashjoin.py","file_ext":"py","file_size_in_byte":3073,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"13816834773","text":"#!/usr/bin/env python3\n\nimport os\nimport numpy as np\nfrom functools import reduce\n\nfile = \"input.txt\"\n\ndata = np.array([line.rstrip('\\n') for line in open(file)])\nbit_length = len(data[0])\ndata_length = len(data)\ncommon_array = [0 for _ in range(bit_length)]\nprint(common_array)\nfor line in data:\n for i, bit in enumerate(line):\n if bit == '0':\n common_array[i] -= 1\n elif bit == '1':\n common_array[i] += 1\nprint(common_array)\n\ngamma = 0\nfor co in common_array:\n gamma = gamma << 1\n if co > 0:\n gamma += 1\n elif co == 0:\n print('Something fishy')\n\nepsilon = (gamma^(2**(bit_length)-1))\nprint(f\"Gamma {gamma:>08b}\")\nprint(f\"Epilson {epsilon:>08b}\")\nprint(f\"Power {gamma*epsilon:>08b}\")\nprint(f\"Power {gamma*epsilon}\")\n\n\n# second\nmost_co_data = np.array([[int(i) for i in line] for line in data])\nfor bit_i in range(bit_length):\n if len(most_co_data) == 1:\n break\n most_common = -((len(most_co_data)+1)//2) + sum(most_co_data[:,bit_i])\n if most_common >= 0:\n most_common = 1\n else:\n most_common = 0\n most_co_data = np.array([dat for dat in most_co_data if dat[bit_i]==most_common])\noxy = reduce(lambda a, b: ((a<<1) + b), most_co_data[0])\nprint(f\"Oxy:\\t\\t{oxy}\")\nprint(f\"Oxy:\\t\\t{oxy:>08b}\")\n\nlest_co_data = np.array([[int(i) for i in line] for line in data])\nfor bit_i in range(bit_length):\n if len(lest_co_data) == 1:\n break\n lest_common = -(len(lest_co_data+1))//2 + sum(lest_co_data[:,bit_i])\n if lest_common >= 0:\n lest_common = 0\n else:\n lest_common = 1\n lest_co_data = np.array([dat for dat in lest_co_data if dat[bit_i]==lest_common])\nco2 = reduce(lambda a, b: ((a<<1) + b), lest_co_data[0])\nprint(f\"CO2:\\t\\t{co2}\")\nprint(f\"CO2:\\t\\t{co2:>08b}\")\n\nprint(f\"Life support:\\t{co2*oxy}\")\n","repo_name":"fridewald/aco2021","sub_path":"03/01.py","file_name":"01.py","file_ext":"py","file_size_in_byte":1823,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"34744483391","text":"\"\"\"--------------------------------------------------------------------------\n Business | Asesores y Consultores en Tecnología S.A. de C.V.\n Programmer | Dyanko Cisneros Mendoza\n Customer | Human Quality\n Project | Meeting Room\n Version | 0.1 --------------------------------------------------------- \"\"\"\n\n## CONTROL SCRIPT IMPORT -------------------------------------------------------\nfrom gui import TLP, BTN, BTNPAGE, BTNGROUP, BTNSTATE, LBL, LVL, POPUP, PAGE\nfrom extronlib import event, Version\nfrom extronlib.device import eBUSDevice, ProcessorDevice, UIDevice\nfrom extronlib.interface import (ContactInterface, DigitalIOInterface, \\\n EthernetClientInterface, EthernetServerInterfaceEx, FlexIOInterface, \\\n IRInterface, RelayInterface, SerialInterface, SWPowerInterface, \\\n VolumeInterface)\nfrom extronlib.ui import Button, Knob, Label, Level\nfrom extronlib.system import Clock, MESet, Wait\n\n## MODULE IMPORT ---------------------------------------------------------------\n## IP:\nimport extr_matrix_DXPHD4k_Series_v1_1_1_0 as DeviceA\nimport extr_other_MediaPort200_v1_1_0_0 as DeviceB\nimport csco_vtc_SX_Series_TC73_v1_3_0_0 as DeviceC\nimport biam_dsp_TesiraSeries_v1_5_20_0 as DeviceD\nimport lutr_lc_CasetaWirelessSmartBridgePro_v1_0_2_0 as DeviceE\n## RS-232:\nimport smfy_controller_RS485_RTS_Transmitter_v1_0_0_0 as DeviceF\n## IR/Serial\n\nprint(Version())\n\n## PROCESOR DEFINITION ---------------------------------------------------------\nIPCP = ProcessorDevice('IPlink')\n\n## IP:\nMATRIX = DeviceA.EthernetClass('10.10.10.10', 23, Model='DXP 88 HD 4k')\nBRIDGE = DeviceB.EthernetClass('10.10.10.11', 23, Model='MediaPort 200')\nCISCO = DeviceC.EthernetClass('10.10.10.12', 23, Model='SX20 TC7.3.X')\nBIAMP = DeviceD.EthernetClass('192.168.10.150', 23, Model='TesiraFORTE CI')\nLUTRON = DeviceE.EthernetClass('192.168.10.15', 23, Model='Caseta Wireless Smart Bridge Pro')\n## RS-232:\nSOMFY = DeviceF.SerialClass(IPCP, 'COM1', Baud=9600, Model='RS485 RTS Transmitter')\n\n## INITIALIZATE ----------------------------------------------------------------\ndef initialize():\n \"\"\"This is the last function that loads when starting the system \"\"\"\n ## OPEN CONNECTION SOCKETS\n ## IP\n MATRIX.Connect()\n BRIDGE.Connect()\n CISCO.Connect()\n BIAMP.Connect()\n LUTRON.Connect()\n ## RS-232\n SOMFY.Initialize()\n\n ## RECURSIVE FUNCTIONS\n update_loop_matrix()\n update_loop_bridge()\n update_loop_cisco()\n update_loop_biamp()\n\n ## POWER COUNTER VARIABLE\n global PWRCOUNT\n PWRCOUNT = 4 #Color Pwr Button Feedback 4=Too Much Red Button, 3=Red, 2=Slow Red, 1=Gray\n\n ## DATA INITIALIZE\n ## Cisco Camera PAGE\n CISCO_DATA['PresetMode'] = 'Recall'\n CISCO_DATA['Camera'] = 'Local'\n BTNGROUP['VCPTZ'].SetCurrent(BTN['VCRecall'])\n BTNGROUP['VCCam'].SetCurrent(BTN['VCLocal'])\n\n ## Cisco Dial PAGE\n global dialerVC ## To access the Dial String variable in all program\n dialerVC = '' ## Clean the Dial String Variable\n CISCO_DATA['Dial'] = '' ## Clean the Dial Data in Dictionary\n LBL['VCDial'].SetText('') ## Clean the Dial Data in GUI\n\n ## VoIP Dial PAGE\n global dialerVI ## To access the Dial String variable in all program\n dialerVI = '' ## Clean the Dial String Variable\n VOIP_DATA['Dial'] = '' ## Clean the Dial Data in Dictionary\n LBL['Dial'].SetText('') ## Clean the Dial Data in gui\n\n ## TOUCH PANEL FUNCTIONS\n TLP.HideAllPopups()\n TLP.ShowPage(PAGE['Index'])\n BTNGROUP['Main'].SetCurrent(None) ##Turn Off all feedback button in GUI Main PAGE\n LBL['CountAll'].SetText('')\n\n ## NOTIFY TO CONSOLE\n print('System Inicializate')\n pass\n\n## SUBSCRIBE FUNCTIONS ---------------------------------------------------------\ndef subscribe_matrix():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n MATRIX.SubscribeStatus('ConnectionStatus', None, matrix_parsing)\n MATRIX.SubscribeStatus('OutputTieStatus', {'Output':'1', 'Tie Type':'Video'}, matrix_parsing)\n MATRIX.SubscribeStatus('OutputTieStatus', {'Output':'2', 'Tie Type':'Video'}, matrix_parsing)\n MATRIX.SubscribeStatus('OutputTieStatus', {'Output':'3', 'Tie Type':'Video'}, matrix_parsing)\n MATRIX.SubscribeStatus('OutputTieStatus', {'Output':'4', 'Tie Type':'Video'}, matrix_parsing)\n MATRIX.SubscribeStatus('OutputTieStatus', {'Output':'1', 'Tie Type':'Audio'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '1'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '2'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '3'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '4'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '5'}, matrix_parsing)\n MATRIX.SubscribeStatus('SignalStatus', {'Input' : '6'}, matrix_parsing)\n pass\n\ndef subscribe_bridge():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n BRIDGE.SubscribeStatus('ConnectionStatus', None, bridge_parsing)\n BRIDGE.SubscribeStatus('HDMIInputEDID', None, bridge_parsing)\n BRIDGE.SubscribeStatus('USBHostStatus', None, bridge_parsing)\n BRIDGE.SubscribeStatus('USBTerminalType', None, bridge_parsing)\n BRIDGE.SubscribeStatus('VideoSendStatus', None, bridge_parsing)\n BRIDGE.SubscribeStatus('VideoSignalPresence', None, bridge_parsing)\n pass\n\ndef subscribe_cisco():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n CISCO.SubscribeStatus('ConnectionStatus', None, cisco_parsing)\n CISCO.SubscribeStatus('CallStatus', {'Call':'1'}, cisco_parsing)\n CISCO.SubscribeStatus('PresentationMode', None, cisco_parsing)\n CISCO.SubscribeStatus('Standby', None, cisco_parsing)\n CISCO.SubscribeStatus('AutoAnswer', None, cisco_parsing)\n CISCO.SubscribeStatus('Volume', None, cisco_parsing)\n pass\n\ndef subscribe_biamp():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n BIAMP.SubscribeStatus('ConnectionStatus', None, biamp_parsing)\n BIAMP.SubscribeStatus('MuteControl', {'Instance Tag':'lvl_spk', 'Channel':'1'}, biamp_parsing)\n BIAMP.SubscribeStatus('MuteControl', {'Instance Tag':'lvl_vcrx', 'Channel':'1'}, biamp_parsing)\n BIAMP.SubscribeStatus('MuteControl', {'Instance Tag':'mute_mix', 'Channel':'1'}, biamp_parsing)\n BIAMP.SubscribeStatus('LevelControl', {'Instance Tag':'lvl_spk', 'Channel':'1'}, biamp_parsing)\n pass\n\ndef subscribe_lutron():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n LUTRON.SubscribeStatus('ConnectionStatus', None, lutron_parsing)\n pass\n\ndef subscribe_somfy():\n \"\"\"This send Subscribe Commands to Device \"\"\"\n SOMFY.SubscribeStatus('ConnectionStatus', None, somfy_parsing)\n pass\n\n## UPDATE FUNCTIONS ------------------------------------------------------------\ndef update_matrix():\n \"\"\"This send Update Commands to Device\"\"\"\n MATRIX.Update('OutputTieStatus', {'Output':'1', 'Tie Type':'Video'})\n MATRIX.Update('OutputTieStatus', {'Output':'2', 'Tie Type':'Video'})\n MATRIX.Update('OutputTieStatus', {'Output':'3', 'Tie Type':'Video'})\n MATRIX.Update('OutputTieStatus', {'Output':'4', 'Tie Type':'Video'})\n MATRIX.Update('OutputTieStatus', {'Output':'1', 'Tie Type':'Audio'})\n MATRIX.Update('SignalStatus', {'Input' : '1'})\n MATRIX.Update('SignalStatus', {'Input' : '2'})\n MATRIX.Update('SignalStatus', {'Input' : '3'})\n MATRIX.Update('SignalStatus', {'Input' : '4'})\n MATRIX.Update('SignalStatus', {'Input' : '5'})\n MATRIX.Update('SignalStatus', {'Input' : '6'})\n pass\n\ndef update_bridge():\n \"\"\"This send Update Commands to Device\"\"\"\n BRIDGE.Update('ConnectionStatus')\n BRIDGE.Update('HDMIInputEDID')\n BRIDGE.Update('USBHostStatus')\n BRIDGE.Update('USBTerminalType')\n BRIDGE.Update('VideoSendStatus')\n BRIDGE.Update('VideoSignalPresence')\n pass\n\ndef update_cisco():\n \"\"\"This send Update Commands to Device\"\"\"\n CISCO.Update('CallStatus', {'Call':'1'})\n CISCO.Update('PresentationMode')\n CISCO.Update('Standby')\n CISCO.Update('AutoAnswer')\n CISCO.Update('Volume')\n pass\n\ndef update_biamp():\n \"\"\"This send Update Commands to Device\"\"\"\n BIAMP.Update('MuteControl', {'Instance Tag':'lvl_spk', 'Channel':'1'})\n BIAMP.Update('MuteControl', {'Instance Tag':'lvl_vcrx', 'Channel':'1'})\n BIAMP.Update('MuteControl', {'Instance Tag':'mute_mix', 'Channel':'1'})\n BIAMP.Update('LevelControl', {'Instance Tag':'lvl_spk', 'Channel':'1'})\n pass\n\ndef update_lutron():\n \"\"\"This send Update Commands to Device\"\"\"\n LUTRON.Update('OutputLevel', {'Integration ID':'2'})\n pass\n\n## DATA PARSING FUNCTIONS ------------------------------------------------------\n## These functions receive the data of the devices in real time\n## Each function stores the parsed data in dictionaries and activate feedback\n## Each function works with the subscription methods of the Python modules\ndef matrix_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Matrix Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n MATRIX_DATA['ConexModule'] = True\n BTN['LANMatrix'].SetState(1)\n else:\n MATRIX_DATA['ConexModule'] = False\n BTN['LANMatrix'].SetState(0)\n ## Disconnect the IP Socket\n MATRIX.Disconnect()\n\n elif command == 'OutputTieStatus':\n if qualifier['Output'] == '1': ## Left Display\n if qualifier['Tie Type'] == 'Video':\n if value == '1':\n BTNGROUP['LCD1_S'].SetCurrent(BTN['LHDMI'])\n elif value == '2':\n BTNGROUP['LCD1_S'].SetCurrent(BTN['LVGA'])\n elif value == '3':\n BTNGROUP['LCD1_S'].SetCurrent(BTN['LPTZ'])\n elif value == '4':\n BTNGROUP['LCD1_S'].SetCurrent(BTN['LShare'])\n\n elif qualifier['Output'] == '2': ## Right Display\n if qualifier['Tie Type'] == 'Video':\n if value == '1':\n BTNGROUP['LCD2_S'].SetCurrent(BTN['RHDMI'])\n elif value == '2':\n BTNGROUP['LCD2_S'].SetCurrent(BTN['RVGA'])\n elif value == '3':\n BTNGROUP['LCD2_S'].SetCurrent(BTN['RPTZ'])\n elif value == '4':\n BTNGROUP['LCD2_S'].SetCurrent(BTN['RShare'])\n\n elif qualifier['Output'] == '3': ## VC Content Input\n if qualifier['Tie Type'] == 'Video':\n if value == '1':\n BTNGROUP['VCPC_S'].SetCurrent(BTN['VCHDMI'])\n elif value == '2':\n BTNGROUP['VCPC_S'].SetCurrent(BTN['VCVGA'])\n elif value == '3':\n BTNGROUP['VCPC_S'].SetCurrent(BTN['VCPTZ'])\n elif value == '4':\n BTNGROUP['VCPC_S'].SetCurrent(BTN['VCShare'])\n\n elif qualifier['Output'] == '4': ## Webex Input\n if qualifier['Tie Type'] == 'Video':\n if value == '1':\n BTNGROUP['Webex'].SetCurrent(BTN['WHDMI'])\n elif value == '2':\n BTNGROUP['Webex'].SetCurrent(BTN['WVGA'])\n elif value == '3':\n BTNGROUP['Webex'].SetCurrent(BTN['WPTZ'])\n elif value == '4':\n BTNGROUP['Webex'].SetCurrent(BTN['WShare'])\n elif value == '5':\n BTNGROUP['Webex'].SetCurrent(BTN['WCisco1'])\n elif value == '6':\n BTNGROUP['Webex'].SetCurrent(BTN['WCisco2'])\n\n elif qualifier['Output'] == '1': ## Audio HDMI Matrix Dembedder\n if qualifier['Tie Type'] == 'Audio':\n if value == '1':\n BTNGROUP['Audio'].SetCurrent(BTN['XHDMI'])\n elif value == '2':\n BTNGROUP['Audio'].SetCurrent(BTN['XVGA'])\n elif value == '4':\n BTNGROUP['Audio'].SetCurrent(BTN['XShare'])\n\n elif command == 'SignalStatus':\n if qualifier['Input'] == '1':\n if value == 'Signal Detected':\n BTN['Signal1'].SetState(1)\n else:\n BTN['Signal1'].SetState(0)\n elif qualifier['Input'] == '2':\n if value == 'Signal Detected':\n BTN['Signal2'].SetState(1)\n else:\n BTN['Signal2'].SetState(0)\n elif qualifier['Input'] == '3':\n if value == 'Signal Detected':\n BTN['Signal3'].SetState(1)\n else:\n BTN['Signal3'].SetState(0)\n elif qualifier['Input'] == '4':\n if value == 'Signal Detected':\n BTN['Signal4'].SetState(1)\n else:\n BTN['Signal4'].SetState(0)\n elif qualifier['Input'] == '5':\n if value == 'Signal Detected':\n BTN['Signal5'].SetState(1)\n else:\n BTN['Signal5'].SetState(0)\n elif qualifier['Input'] == '6':\n if value == 'Signal Detected':\n BTN['Signal6'].SetState(1)\n else:\n BTN['Signal6'].SetState(0)\n pass\n\ndef bridge_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Bridge Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n BRIDGE_DATA['ConexModule'] = True\n BTN['LanBridge'].SetState(1)\n else:\n BRIDGE_DATA['ConexModule'] = False\n BTN['LanBridge'].SetState(0)\n ## Disconnect the IP Socket\n BRIDGE.Disconnect()\n\n elif command == 'HDMIInputEDID':\n print(value)\n BRIDGE_DATA['InputEDID'] = value\n\n elif command == 'USBHostStatus':\n print(value)\n BRIDGE_DATA['USBHost'] = value\n\n elif command == 'USBTerminalType':\n print(value)\n BRIDGE_DATA['USBTerminal'] = value\n\n elif command == 'VideoSendStatus':\n print(value)\n if value == 'On':\n BRIDGE_DATA['VideoSend'] = True\n else:\n BRIDGE_DATA['VideoSend'] = False\n\n elif command == 'VideoSignalPresence':\n print(value)\n if value == 'Signal':\n BRIDGE_DATA['VideoSignal'] = True\n else:\n BRIDGE_DATA['VideoSignal'] = False\n pass\n\ndef cisco_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Cisco Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n CISCO_DATA['ConexModule'] = True\n BTN['LANCisco'].SetState(1)\n else:\n CISCO_DATA['ConexModule'] = False\n BTN['LANCisco'].SetState(0)\n ## Disconnect the IP Socket\n CISCO.Disconnect()\n\n elif command == 'CallStatus':\n print(qualifier + value)\n CISCO_DATA['Call'] = value\n\n elif command == 'PresentationMode':\n print(value)\n CISCO_DATA['Content'] = value\n\n elif command == 'Standby':\n print(value)\n if value == 'Activate':\n CISCO_DATA['Power'] = True\n else:\n CISCO_DATA['Power'] = False\n\n elif command == 'AutoAnswer':\n if value == 'On':\n CISCO_DATA['AutoAnswer'] = True\n BTN['VCAutoAn'].SetState(1)\n elif value == 'Off':\n CISCO_DATA['AutoAnswer'] = False\n BTN['VCAutoAn'].SetState(0)\n\n elif command == 'Volume':\n print(value)\n LVL['VC'].SetLevel(value) ## Send volume value to Level Bar\n CISCO_DATA['Volume'] = value ## Store volume value in dictionary\n pass\n\ndef biamp_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Biamp Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n BIAMP_DATA['ConexModule'] = True\n BTN['LANBiamp'].SetState(1)\n else:\n BIAMP_DATA['ConexModule'] = False\n BTN['LANBiamp'].SetState(0)\n ## Turn Off feedback Buttons\n LVL['Spk'].SetLevel(-100)\n ## Disconnect the IP Socket\n BIAMP.Disconnect()\n\n elif command == 'MuteControl':\n print(str(qualifier) + ' ' + str(value))\n\n if qualifier['Instance Tag'] == 'lvl_spk':\n if value == 'On':\n BIAMP_DATA['MuteSpk'] = True\n BTN['XSpk'].SetState(1)\n elif value == 'Off':\n BIAMP_DATA['MuteSpk'] = False\n BTN['XSpk'].SetState(0)\n\n elif qualifier['Instance Tag'] == 'lvl_vcrx':\n if value == 'On':\n BIAMP_DATA['MuteVCRx'] = True\n BTN['XVC'].SetState(1)\n elif value == 'Off':\n BIAMP_DATA['MuteVCRx'] = False\n BTN['XVC'].SetState(0)\n\n elif qualifier['Instance Tag'] == 'mute_mix':\n if value == 'On':\n BIAMP_DATA['Mute_Mics'] = True\n BTN['XMics'].SetState(1)\n elif value == 'Off':\n BIAMP_DATA['Mute_Mics'] = False\n BTN['XMics'].SetState(0)\n\n elif command == 'LevelControl':\n print(str(qualifier) + ' ' + str(value))\n value = int(value) ## Convert reported volume to Integer\n LVL['Spk'].SetLevel(value) ## Send volume value to Level Bar\n BIAMP_DATA['lvl_spk'] = value ## Store volume value in dictionary\n pass\n\ndef lutron_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Lutron Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n LUTRON_DATA['ConexModule'] = True\n BTN['LANLutron'].SetState(1)\n else:\n LUTRON_DATA['ConexModule'] = False\n BTN['LANLutron'].SetState(0)\n ## Disconnect the IP Socket\n LUTRON.Disconnect()\n pass\n\ndef somfy_parsing(command, value, qualifier):\n \"\"\"Retrieve the Real Information of the Device \"\"\"\n if command == 'ConnectionStatus':\n print('Somfy Module Conex status: {}'.format(value))\n\n if value == 'Connected':\n SOMFY_DATA['ConexModule'] = True\n BTN['232Somfy'].SetState(1)\n else:\n SOMFY_DATA['ConexModule'] = False\n BTN['232Somfy'].SetState(0)\n\n pass\n## EVENT FUNCTIONS ----------------------------------------------------------------\n## This functions report a 'Online' / 'Offline' status after to send a Connect()\n## CAUTION: If you never make a Connect(), the Module never work with Subscriptions\n@event(MATRIX, 'Connected')\n@event(MATRIX, 'Disconnected')\ndef matrix_conex_event(interface, state):\n \"\"\"MATRIX CONNECT() STATUS \"\"\"\n print('Matrix Conex Event: ' + state)\n if state == 'Connected':\n BTN['LANMatrix'].SetState(1)\n MATRIX_DATA['ConexEvent'] = True\n ## Send & Query Information\n subscribe_matrix()\n update_matrix()\n if state == 'Disconnected':\n BTN['LANMatrix'].SetState(0)\n MATRIX_DATA['ConexEvent'] = False\n trying_matrix()\n pass\n\n@event(BRIDGE, 'Connected')\n@event(BRIDGE, 'Disconnected')\ndef bridge_conex_event(interface, state):\n \"\"\"BRIDGE CONNECT() STATUS \"\"\"\n print('Bridge Conex Event: ' + state)\n if state == 'Connected':\n BTN['LanBridge'].SetState(1)\n BRIDGE_DATA['ConexEvent'] = True\n ## Send & Query Information\n subscribe_bridge()\n update_bridge()\n if state == 'Disconnected':\n BTN['LanBridge'].SetState(0)\n BRIDGE_DATA['ConexEvent'] = False\n trying_bridge()\n pass\n\n@event(CISCO, 'Connected')\n@event(CISCO, 'Disconnected')\ndef cisco_conex_event(interface, state):\n \"\"\"CISCO CONNECT() STATUS \"\"\"\n print('Cisco Conex Event: ' + state)\n if state == 'Connected':\n BTN['LANCisco'].SetState(1)\n CISCO_DATA['ConexEvent'] = True\n ## Send & Query Information\n subscribe_cisco()\n update_cisco()\n if state == 'Disconnected':\n BTN['LANCisco'].SetState(0)\n CISCO_DATA['ConexEvent'] = False\n trying_cisco()\n pass\n\n@event(BIAMP, 'Connected')\n@event(BIAMP, 'Disconnected')\ndef biamp_conex_event(interface, state):\n \"\"\"DEVICE CONNECT() STATUS \"\"\"\n print('Biamp Conex Event: ' + state)\n if state == 'Connected':\n BTN['LANBiamp'].SetState(1)\n BIAMP_DATA['ConexEvent'] = True\n ## Send & Query Information\n subscribe_biamp()\n update_biamp()\n if state == 'Disconnected':\n BTN['LANBiamp'].SetState(0)\n BIAMP_DATA['ConexEvent'] = False\n trying_biamp()\n pass\n\n@event(LUTRON, 'Connected')\n@event(LUTRON, 'Disconnected')\ndef lutron_conex_event(interface, state):\n \"\"\"DEVICE CONNECT() STATUS \"\"\"\n print('Lutron Conex Event: ' + state)\n if state == 'Connected':\n BTN['LANLutron'].SetState(1)\n LUTRON_DATA['ConexEvent'] = True\n ## Send & Query Information\n subscribe_lutron()\n update_lutron()\n if state == 'Disconnected':\n BTN['LANLutron'].SetState(0)\n LUTRON_DATA['ConexEvent'] = False\n trying_lutron()\n pass\n\n## RECURSIVE FUNCTIONS ------------------------------------------------------------\n## Help´s when the device was Off in the first Connect() method when the code starts\ndef trying_matrix():\n \"\"\"Try to make a Connect() to device\"\"\"\n if MATRIX_DATA['ConexEvent'] == False:\n print('Tryng to make a Connect() in Matrix')\n MATRIX.Connect(4) ## Have 4 seconds to try to connect\n pass\nloop_trying_matrix = Wait(5, trying_matrix)\n\ndef trying_bridge():\n \"\"\"Try to make a Connect() to device\"\"\"\n if BRIDGE_DATA['ConexEvent'] == False:\n print('Tryng to make a Connect() in Bridge')\n BRIDGE.Connect(4) ## Have 4 seconds to try to connect\n pass\nloop_trying_bridge = Wait(5, trying_bridge)\n\ndef trying_cisco():\n \"\"\"Try to make a Connect() to device\"\"\"\n if CISCO_DATA['ConexEvent'] == False:\n print('Tryng to make a Connect() in Cisco')\n CISCO.Connect(4) ## Have 4 seconds to try to connect\n pass\nloop_trying_cisco = Wait(5, trying_cisco)\n\ndef trying_biamp():\n \"\"\"Try to make a Connect() to device\"\"\"\n if BIAMP_DATA['ConexEvent'] == False:\n print('Tryng to make a Connect() in Biamp')\n BIAMP.Connect(4) ## Have 4 seconds to try to connect\n pass\nloop_trying_biamp = Wait(5, trying_biamp)\n\ndef trying_lutron():\n \"\"\"Try to make a Connect() to device\"\"\"\n if LUTRON_DATA['ConexEvent'] == False:\n print('Tryng to make a Connect() in Lutron')\n LUTRON.Connect(4) ## Have 4 seconds to try to connect\n pass\nloop_trying_lutron = Wait(5, trying_lutron)\n\n## RECURSIVE LOOP FUNCTIONS -----------------------------------------------------------\n## This not affect any device\n## This return True / False when no response is received from Module\n## If in 5 times the data is not reported (connectionCounter = 5) from the Update Command\n## Generate 'Connected' / 'Disconnected'\n\ndef update_loop_matrix():\n \"\"\"Continuos Update Commands to produce Module Connected / Disconnected\"\"\"\n MATRIX.Update('SignalStatus', {'Input':'1'})\n loop_update_matrix.Restart()\nloop_update_matrix = Wait(12, update_loop_matrix)\n\ndef update_loop_bridge():\n \"\"\"Continuos Update Commands to produce Module Connected / Disconnected\"\"\"\n BRIDGE.Update('AutoImage')\n loop_update_bridge.Restart()\nloop_update_bridge = Wait(12, update_loop_bridge)\n\ndef update_loop_cisco():\n \"\"\"Continuos Update Commands to produce Module Connected / Disconnected\"\"\"\n CISCO.Update('AutoAnswer')\n loop_update_cisco.Restart()\nloop_update_cisco = Wait(12, update_loop_cisco)\n\ndef update_loop_biamp():\n \"\"\"Continuos Update Commands to produce Module Connected / Disconnected\"\"\"\n BIAMP.Update('VerboseMode')\n loop_update_biamp.Restart()\nloop_update_biamp = Wait(12, update_loop_biamp)\n\ndef update_loop_lutron():\n \"\"\"Continuos Update Commands to produce Module Connected / Disconnected\"\"\"\n LUTRON.Update('OutputLevel', {'Integration ID':'2'})\n loop_update_lutron.Restart()\nloop_update_lutron = Wait(12, update_loop_lutron)\n\n## DATA DICTIONARIES -----------------------------------------------------------\n## Each dictionary store the real time information of room devices\n## IP\nMATRIX_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n}\n\nBRIDGE_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n ##\n 'InputEDID' : '',\n 'USBHost' : '',\n 'USBTerminal': '',\n 'VideoSend' : None,\n 'VideoSignal': None,\n}\n\nCISCO_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n ##\n 'AutoAnswer' : None,\n 'CallStatus' : '',\n 'Camera' : '',\n 'Dial' : '',\n 'Power' : None,\n 'PresetMode' : '',\n 'Volume' : None,\n}\n\nBIAMP_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n ##\n 'MuteSpk' : None,\n 'MuteVCRx' : None,\n 'Mute_Mics' : None,\n 'lvl_spk' : None,\n}\n\nVOIP_DATA = {\n 'Dial' : '',\n 'DTMF' : False\n}\n\nLUTRON_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n}\n\n## RS-232\nSOMFY_DATA = {\n 'ConexModule': None,\n 'ConexEvent' : None,\n}\n\n## IR-Serial\nLCD1_DATA = {\n 'Input' : '',\n 'Power' : '',\n}\n\nLCD2_DATA = {\n 'Input' : '',\n 'Power' : '',\n}\n## PAGE USER EVENTS ------------------------------------------------------------\n## PAGE Index ------------------------------------------------------------------\n@event(BTN['Index'], 'Pressed')\ndef index_events(button, state):\n \"\"\"User Actions: Touch Index Page\"\"\"\n TLP.ShowPage(PAGE['Main'])\n TLP.ShowPopup(POPUP['Hi'])\n print('Touch Mode: %s' % 'Index')\n pass\n\n## PAGE Main -------------------------------------------------------------------\n@event(BTNPAGE['Main'], BTNSTATE['List'])\ndef main_events(button, state):\n \"\"\"User Actions: Touch Main Page\"\"\"\n if button is BTN['Video'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['Video'])\n LBL['Master'].SetText('Seleccionar Display')\n print('Touch Mode: %s' % 'Video')\n\n elif button is BTN['VC'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['VC'])\n LBL['Master'].SetText('Control de Videoconferencia')\n print('Touch Mode: %s' % 'VC')\n\n elif button is BTN['Webex'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['Webex'])\n LBL['Master'].SetText('Control de Webconferencia')\n print('Touch Mode: %s' % 'Webex')\n\n elif button is BTN['VoIP'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['VoIP'])\n LBL['Master'].SetText('Telefonía IP')\n print('Touch Mode: %s' % 'VoIP')\n\n elif button is BTN['Lights'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['Lights'])\n LBL['Master'].SetText('Control de Iluminación')\n print('Touch Mode: %s' % 'Lights')\n\n elif button is BTN['Audio'] and state == 'Pressed':\n ## Query Data from Biamp\n update_biamp()\n TLP.ShowPopup(POPUP['Audio'])\n LBL['Master'].SetText('Control de Audio')\n print('Touch Mode: %s' % 'Audio')\n\n elif button is BTN['Status'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['Status'])\n LBL['Master'].SetText('Información de Dispositivos')\n print('Touch Mode: %s' % 'Status')\n\n elif button is BTN['PwrOff'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['Power'])\n LBL['Master'].SetText('¿Deseas Apagar el Sistema?')\n print('Touch Mode: %s' % 'PowerOff')\n\n ##Turn On the feedbak of last pressed button\n BTNGROUP['Main'].SetCurrent(button)\n pass\n\n## PAGE Video ------------------------------------------------------------------\n@event(BTNPAGE['Video'], BTNSTATE['List'])\ndef video_events(button, state):\n \"\"\"User Actions: Touch Video Page\"\"\"\n if button is BTN['DisplayL'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['LCD1'])\n LBL['Master'].SetText('Control de Pantalla Izquierda')\n print('Video Mode: %s' % 'Display L')\n\n elif button is BTN['DisplayR'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['LCD2'])\n LBL['Master'].SetText('Control de Pantalla Derecha')\n print('Video Mode: %s' % 'Display R')\n pass\n\n## PAGE Display L --------------------------------------------------------------\n@event(BTNPAGE['LCD1'], BTNSTATE['List'])\ndef display_l_events(button, state):\n \"\"\"User Actions: Touch LCD-L Page\"\"\"\n if button is BTN['LHDMI'] and state == 'Pressed':\n ## HDMI to Display Left - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'1', 'Output':'1', 'Tie Type':'Video'})\n print('Button Pressed - LCD L: %s' % 'HDMI')\n\n elif button is BTN['LVGA'] and state == 'Pressed':\n ## VGA to Display Left - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'2', 'Output':'1', 'Tie Type':'Video'})\n print('Button Pressed - LCD L: %s' % 'VGA')\n\n elif button is BTN['LPTZ'] and state == 'Pressed':\n ## PTZ to Display Left - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'3', 'Output':'1', 'Tie Type':'Video'})\n print('Button Pressed - LCD L: %s' % 'PTZ')\n\n elif button is BTN['LShare'] and state == 'Pressed':\n ## ShareLink to Display Left - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'4', 'Output':'1', 'Tie Type':'Video'})\n print('Button Pressed - LCD L: %s' % 'ShareLink')\n\n elif button is BTN['LPwrOn'] and state == 'Pressed':\n ## The system send the command action directly\n print('Button Pressed - LCD L: %s' % 'PowerOn')\n\n elif button is BTN['LPwrOff'] and state == 'Pressed':\n ## The system send the command action directly\n print('Button Pressed - LCD L: %s' % 'PowerOff')\n\n elif button is BTN['LBack'] and state == 'Pressed':\n ## Hide actual POPUP and show the Display Selection POPUP\n TLP.ShowPopup(POPUP['Video'])\n print('Button Pressed - LCD L: %s' % 'Back')\n pass\n\n## PAGE Display R --------------------------------------------------------------\n@event(BTNPAGE['LCD2'], BTNSTATE['List'])\ndef display_r_events(button, state):\n \"\"\"User Actions: Touch LCD-R Page\"\"\"\n if button is BTN['RHDMI'] and state == 'Pressed':\n ## HDMI to Display Right - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'1', 'Output':'2', 'Tie Type':'Video'})\n print('Button Pressed - LCD R: %s' % 'HDMI')\n\n elif button is BTN['RVGA'] and state == 'Pressed':\n ## VGA to Display Right - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'2', 'Output':'2', 'Tie Type':'Video'})\n print('Button Pressed - LCD R: %s' % 'VGA')\n\n elif button is BTN['RPTZ'] and state == 'Pressed':\n ## PTZ to Display Right - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'3', 'Output':'2', 'Tie Type':'Video'})\n print('Button Pressed - LCD R: %s' % 'PTZ')\n\n elif button is BTN['RShare'] and state == 'Pressed':\n ## ShareLink to Display Right - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'4', 'Output':'2', 'Tie Type':'Video'})\n print('Button Pressed - LCD R: %s' % 'ShareLink')\n\n elif button is BTN['RPwrOn'] and state == 'Pressed':\n ## The system send the command action directly\n print('Button Pressed - LCD R: %s' % 'PowerOn')\n\n elif button is BTN['RPwrOff'] and state == 'Pressed':\n ## The system send the command action directly\n print('Button Pressed - LCD R: %s' % 'PowerOff')\n\n elif button is BTN['RBack'] and state == 'Pressed':\n ## Hide actual POPUP and show the Display Selection POPUP\n TLP.ShowPopup(POPUP['Video'])\n print('Button Pressed - LCD R: %s' % 'Back')\n pass\n\n## PAGE VC ---------------------------------------------------------------------\n@event(BTNPAGE['VCCall'], BTNSTATE['List'])\ndef vc_call_events(button, state):\n \"\"\"User Actions: Touch VC Page\"\"\"\n if button is BTN['VCCall'] and state == 'Pressed':\n ##--This button dial the number typed on the touch panel (Cisco VC)\n CISCO.Set('Hook', 'Dial', {'Protocol':'H323', 'Number': CISCO_DATA['Dial']})\n print('Button Pressed - VC: %s' % 'Call')\n\n elif button is BTN['VCHangup'] and state == 'Pressed':\n ##--This button hangs up all active calls (Cisco VC)\n CISCO.Set('Hook', 'Disconnect All', {'Protocol':'H323'})\n print('Button Pressed - VC: %s' % 'Hangup')\n pass\n\n## This function is called when the user press a Dial Button\n## This function add or remove data from the panel Dial Number\ndef dialer_vc(btn_name):\n \"\"\"User Actions: Touch VC Page\"\"\"\n global dialerVC\n\n if btn_name == 'Delete': #If the user push 'Delete' button\n dialerVC = dialerVC[:-1] #Remove the last char of the string\n CISCO_DATA['Dial'] = dialerVC #Asign the string to the data dictionary\n LBL['VCDial'].SetText(dialerVC) #Send the string to GUI Label\n\n else: #If the user push a [*#0-9] button\n number = str(btn_name[4]) #Extract the valid character of BTN name\n dialerVC += number #Append the last char to the string\n CISCO_DATA['Dial'] = dialerVC #Asign the string to the data dictionary\n LBL['VCDial'].SetText(dialerVC) #Send the string to GUI Label\n pass\n\n@event(BTNPAGE['VCDial'], BTNSTATE['List'])\ndef vc_dial_events(button, state):\n \"\"\"User Actions: Touch VC Page\"\"\"\n ## All the VC Dial Buttons pressed come in button variable\n if state == 'Pressed' or state == 'Repeated':\n print('Button Pressed - VC: %s' % button.Name)\n dialer_vc(button.Name) #Recall a validation function\n button.SetState(1)\n else:\n button.SetState(0)\n pass\n\n@event(BTNPAGE['VCOpt'], BTNSTATE['List'])\ndef vc_opt_events(button, state):\n \"\"\"User Actions: Touch VC Page\"\"\"\n\n ## VC Options: Content Control\n if button is BTN['VCEnviar'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['VC_PC'])\n BTN['VCEnviar'].SetState(1)\n print('Button Pressed - VC: %s' % 'Content')\n else:\n BTN['VCEnviar'].SetState(0)\n\n ## VC Options: Camera Control\n if button is BTN['VCCamara'] and state == 'Pressed':\n TLP.ShowPopup(POPUP['VC_Cam'])\n BTN['VCCamara'].SetState(1)\n print('Button Pressed - VC: %s' % 'Camera')\n else:\n BTN['VCCamara'].SetState(0)\n\n ## VC Options: AutoAnswer\n if button is BTN['VCAutoAn'] and state == 'Pressed':\n #\n if CISCO_DATA['AutoAnswer'] == True:\n CISCO.Set('AutoAnswer', 'Off')\n\n elif CISCO_DATA['AutoAnswer'] == False:\n CISCO.Set('AutoAnswer', 'On')\n print('Button Pressed - VC: %s' % 'AutoAnswer')\n pass\n\n## PAGE VC Content -------------------------------------------------------------\n@event(BTNPAGE['VCPC'], BTNSTATE['List'])\ndef vc_cam_content_sources(button, state):\n \"\"\"User Actions: Touch VC Content Page\"\"\"\n\n if button is BTN['VCHDMI'] and state == 'Pressed':\n ## HDMI to Cisco Computer Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'1', 'Output':'3', 'Tie Type':'Video'})\n print('Button Pressed - VC Share: %s' % 'HDMI')\n\n elif button is BTN['VCVGA'] and state == 'Pressed':\n ## VGA to Cisco Computer Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'2', 'Output':'3', 'Tie Type':'Video'})\n print('Button Pressed - VC Share: %s' % 'VGA')\n\n elif button is BTN['VCPTZ'] and state == 'Pressed':\n ## PTZ to Cisco Computer Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'3', 'Output':'3', 'Tie Type':'Video'})\n print('Button Pressed - VC Share: %s' % 'PTZ')\n\n elif button is BTN['VCShare'] and state == 'Pressed':\n ## ShareLink to Cisco Computer Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'4', 'Output':'3', 'Tie Type':'Video'})\n print('Button Pressed - VC Share: %s' % 'ClickShare')\n\n elif button is BTN['VCBack2'] and state == 'Pressed':\n ## Hide VC Content POPUP and Show the main VC POPUP\n TLP.ShowPopup(POPUP['VC'])\n print('Button Pressed - VC Share: %s' % 'Back')\n\n elif button is BTN['VCSend'] and state == 'Pressed':\n ## Play - Share graphics presentation\n CISCO.Set('Presentation', '1')\n print('Button Pressed - VC Share: %s' % 'Send')\n\n elif button is BTN['VCStop'] and state == 'Pressed':\n ## Stop - Sharing graphics\n CISCO.Set('Presentation', 'Stop')\n print('Button Pressed - VC Share: %s' % 'Stop')\n pass\n\n## PAGE VC Camera --------------------------------------------------------------\n@event(BTNPAGE['VCCam'], BTNSTATE['List'])\ndef vc_nav_events(button, state):\n \"\"\"User Actions: Touch VC Camera Page\"\"\"\n if button is BTN['VCUp']:\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraTiltSX20', 'Up', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Up')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Up')\n print('Cam Remota - Cisco: %s' % 'Cam Up')\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraTiltSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n #--\n elif button is BTN['VCLeft']:\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraPanSX20', 'Left', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Left')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Left')\n print('Cam Remota - Cisco: %s' % 'Cam Left')\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraPanSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n #--\n elif button is BTN['VCDown']:\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraTiltSX20', 'Down', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Down')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Down')\n print('Cam Remota - Cisco: %s' % 'Cam Down')\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraTiltSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n #--\n elif button is BTN['VCRight']:\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraPanSX20', 'Right', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Right')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Right')\n print('Cam Remota - Cisco: %s' % 'Cam Right')\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraPanSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraPan/Tilt', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n #--\n elif button is BTN['VCZoom1']: #+\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraZoomSX20', 'In', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Zoom+')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraZoom', 'In')\n print('Cam Remota - Cisco: %s' % 'Cam Zoom+')\n BTN['VCZoom1'].SetState(1)\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraZoomSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraZoom', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n BTN['VCZoom1'].SetState(0)\n #--\n elif button is BTN['VCZoom2']: #-\n if state == 'Pressed' or state == 'Repeated':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraZoomSX20', 'Out', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Zoom-')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraZoom', 'Out')\n print('Cam Remota - Cisco: %s' % 'Cam Zoom-')\n BTN['VCZoom2'].SetState(1)\n #--\n elif state == 'Released':\n if CISCO_DATA['Camera'] == 'Local':\n CISCO.Set('CameraZoomSX20', 'Stop', {'Speed':7})\n print('Cam Local - Cisco: %s' % 'Cam Stop')\n elif CISCO_DATA['Camera'] == 'Remote':\n CISCO.Set('FarEndCameraZoom', 'Stop')\n print('Cam Remota - Cisco: %s' % 'Cam Stop')\n BTN['VCZoom2'].SetState(0)\n #--\n if button is BTN['VCLocal'] and state == 'Pressed':\n CISCO_DATA['Camera'] = 'Local'\n BTNGROUP['VCCam'].SetCurrent(BTN['VCLocal'])\n print('Button Pressed - Cisco: %s' % 'Cam Local')\n #--\n elif button is BTN['VCRemote'] and state == 'Pressed':\n CISCO_DATA['Camera'] = 'Remote'\n BTNGROUP['VCCam'].SetCurrent(BTN['VCRemote'])\n print('Button Pressed - Cisco: %s' % 'Cam Remote')\n pass\n\n\n@event(BTNPAGE['VCPre'], BTNSTATE['List'])\ndef vc_cam_events(button, state):\n \"\"\"User Actions: Touch VC Camera Page\"\"\"\n if button is BTN['VCP1'] and state == 'Pressed':\n if CISCO_DATA['Camera'] == 'Local':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('CameraPresetPositionRecallSX20', '1')\n print('Recall Local Preset Cisco: %s' % '1')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('CameraPresetSaveSX20', '1')\n print('Save Local Preset Cisco: %s' % '1')\n #--\n elif CISCO_DATA['Camera'] == 'Remote':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('FarEndCameraPresetRecall', '1')\n print('Recall Remote Preset Cisco: %s' % '1')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('FarEndCameraPresetSave', '1')\n print('Save Remote Preset Cisco: %s' % '1')\n #--\n elif button is BTN['VCP2'] and state == 'Pressed':\n if CISCO_DATA['Camera'] == 'Local':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('CameraPresetPositionRecallSX20', '2')\n print('Recall Local Preset Cisco: %s' % '2')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('CameraPresetSaveSX20', '2')\n print('Save Local Preset Cisco: %s' % '2')\n #--\n elif CISCO_DATA['Camera'] == 'Remote':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('FarEndCameraPresetRecall', '2')\n print('Recall Remote Preset Cisco: %s' % '2')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('FarEndCameraPresetSave', '2')\n print('Save Remote Preset Cisco: %s' % '2')\n #--\n elif button is BTN['VCP3'] and state == 'Pressed':\n if CISCO_DATA['Camera'] == 'Local':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('CameraPresetPositionRecallSX20', '3')\n print('Recall Local Preset Cisco: %s' % '3')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('CameraPresetSaveSX20', '3')\n print('Save Local Preset Cisco: %s' % '3')\n #--\n elif CISCO_DATA['Camera'] == 'Remote':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('FarEndCameraPresetRecall', '3')\n print('Recall Remote Preset Cisco: %s' % '3')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('FarEndCameraPresetSave', '3')\n print('Save Remote Preset Cisco: %s' % '3')\n #--\n elif button is BTN['VCP4'] and state == 'Pressed':\n if CISCO_DATA['Camera'] == 'Local':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('CameraPresetPositionRecallSX20', '4')\n print('Recall Local Preset Cisco: %s' % '4')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('CameraPresetSaveSX20', '4')\n print('Save Local Preset Cisco: %s' % '4')\n #--\n elif CISCO_DATA['Camera'] == 'Remote':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('FarEndCameraPresetRecall', '4')\n print('Recall Remote Preset Cisco: %s' % '4')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('FarEndCameraPresetSave', '4')\n print('Save Remote Preset Cisco: %s' % '4')\n #--\n elif button is BTN['VCP5'] and state == 'Pressed':\n if CISCO_DATA['Camera'] == 'Local':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('CameraPresetPositionRecallSX20', '5')\n print('Recall Local Preset Cisco: %s' % '5')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('CameraPresetSaveSX20', '5')\n print('Save Local Preset Cisco: %s' % '5')\n #--\n elif CISCO_DATA['Camera'] == 'Remote':\n if CISCO_DATA['PresetMode'] == 'Recall':\n CISCO.Set('FarEndCameraPresetRecall', '5')\n print('Recall Remote Preset Cisco: %s' % '5')\n elif CISCO_DATA['PresetMode'] == 'Save':\n CISCO.Set('FarEndCameraPresetSave', '5')\n print('Save Remote Preset Cisco: %s' % '5')\n #--\n elif button is BTN['VCRecall'] and state == 'Pressed':\n CISCO_DATA['PresetMode'] = 'Recall'\n BTNGROUP['VCPTZ'].SetCurrent(BTN['VCRecall'])\n print('Button Pressed - Cisco: %s' % 'Recall')\n #--\n elif button is BTN['VCSave'] and state == 'Pressed':\n CISCO_DATA['PresetMode'] = 'Save'\n BTNGROUP['VCPTZ'].SetCurrent(BTN['VCSave'])\n print('Button Pressed - Cisco: %s' % 'Save')\n pass\n\n## PAGE Webex ------------------------------------------------------------------\n@event(BTNPAGE['Webex'], BTNSTATE['List'])\ndef webex_events(button, state):\n \"\"\"User Actions: Touch Webex Page\"\"\"\n if button is BTN['WHDMI'] and state == 'Pressed':\n ## HDMI to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'1', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'HDMI')\n\n elif button is BTN['WVGA'] and state == 'Pressed':\n ## VGA to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'2', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'VGA')\n\n elif button is BTN['WPTZ'] and state == 'Pressed':\n ## PTZ to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'3', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'PTZ')\n\n elif button is BTN['WShare'] and state == 'Pressed':\n ## ShareLink to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'4', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'ShareLink')\n\n elif button is BTN['WCisco1'] and state == 'Pressed':\n ## Cisco 1 to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'5', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'Cisco 1')\n\n elif button is BTN['WCisco2'] and state == 'Pressed':\n ## Cisco 2 to MediaPort200 Input - Video\n MATRIX.Set('MatrixTieCommand', None, {'Input':'6', 'Output':'5', 'Tie Type':'Video'})\n print('Button Pressed - Webex: %s' % 'Cisco 2')\n pass\n\n## PAGE VoIP -------------------------------------------------------------------\n@event(BTNPAGE['TelCall'], BTNSTATE['List'])\ndef vi_call_events(button, state):\n \"\"\"User Actions: Touch VoIP Page\"\"\"\n if button is BTN['Call'] and state == 'Pressed':\n ##--This button dial the number typed on the touch panel (Biamp VoIP)\n BIAMP.Set('VoIPHook', 'Dial',\n {'Instance Tag':'Dialer', 'Line':'1', 'Call Appearance':'1', \\\n 'Number':VOIP_DATA['Dial']})\n print('Button Pressed - VoIP: %s' % 'Call')\n #--\n elif button is BTN['Hangup'] and state == 'Pressed':\n ##--This button hangs up all active calls (Biamp VoIP)\n BIAMP.Set('VoIPHook', 'End',\n {'Instance Tag':'Dialer', 'Line':'1', 'Call Appearance':'1'})\n print('Button Pressed - VoIP: %s' % 'Hangup')\n pass\n\n## This function is called when the user press a Dial Button\n## This function add or remove data from the panel Dial Number\ndef dialer_voip(btn_name):\n \"\"\"User Actions: Touch VoIP Page\"\"\"\n global dialerVI\n\n if btn_name == 'Delete': #If the user push 'Delete' button\n dialerVI = dialerVI[:-1] #Remove the last char of the string\n VOIP_DATA['Dial'] = dialerVI #Asign the string to the data dictionary\n LBL['Dial'].SetText(dialerVI) #Send the string to GUI Label\n\n else: #If the user push a [*#0-9] button\n number = str(btn_name[4]) #Extract the valid character of BTN name\n if VOIP_DATA['DTMF'] == False: #If the DTMF is off\n dialerVI += number #Append the last char to the string\n VOIP_DATA['Dial'] = dialerVI #Asign the string to the data dictionary\n LBL['Dial'].SetText(dialerVI) #Send the string to GUI Label\n elif VOIP_DATA['DTMF'] == True: #If DTMF is On\n BIAMP.Set('DTMF', number, {'Instance Tag':'Dialer', 'Line':'1'})\n pass\n\n@event(BTNPAGE['TelDial'], BTNSTATE['List'])\ndef vi_dial_events(button, state):\n \"\"\"User Actions: Touch VoIP Page\"\"\"\n ## All the VoIP Dial Buttons pressed come in button variable\n if state == 'Pressed' or state == 'Repeated':\n print('Button Pressed - VoIP: %s' % button.Name)\n dialer_voip(button.Name) #Recall a validation function\n button.SetState(1)\n else:\n button.SetState(0)\n pass\n\n@event(BTNPAGE['TelOpt'], BTNSTATE['List'])\ndef vi_opt_events(button, state):\n \"\"\"User Actions: Touch VoIP Page\"\"\"\n ## VoIP Redial Control\n if button is BTN['Redial'] and state == 'Pressed':\n BIAMP.Set('VoIPHook', 'Redial', {'Instance Tag':'Dialer', \\\n 'Line':'1', 'Call Appearance':'1'})\n print('Button Pressed - VoIP: %s' % 'Redial')\n\n ## VoIP DTMF Control\n elif button is BTN['DTMF'] and state == 'Pressed':\n if VOIP_DATA['DTMF'] == False:\n VOIP_DATA['DTMF'] = True\n BTN['DTMF'].SetState(1)\n print('Button Pressed - VoIP: %s' % 'DTMF On')\n #--\n elif VOIP_DATA['DTMF'] == True:\n VOIP_DATA['DTMF'] = False\n BTN['DTMF'].SetState(0)\n print('Button Pressed - VoIP: %s' % 'DTMF Off')\n print('Button Pressed - VoIP: %s' % 'DTMF')\n\n ## Hold / Resume Control\n elif button is BTN['Hold'] and state == 'Pressed':\n print('Button Pressed - VoIP: %s' % 'Hold/Resume')\n pass\n\n## PAGE Audio ------------------------------------------------------------------\n@event(BTNPAGE['Audio1'], BTNSTATE['List'])\ndef audio_source_events(button, state):\n \"\"\"User Actions: Touch Audio Page\"\"\"\n\n if button is BTN['XHDMI'] and state == 'Pressed':\n ## HDMI to HDMI Audio Dembedder Input - Audio\n MATRIX.Set('MatrixTieCommand', None, {'Input':'1', 'Output':'1', 'Tie Type':'Audio'})\n print('Button Pressed - Audio: %s' % 'HDMI')\n\n elif button is BTN['XVGA'] and state == 'Pressed':\n ## VGA to HDMI Audio Dembedder Input - Audio\n MATRIX.Set('MatrixTieCommand', None, {'Input':'2', 'Output':'1', 'Tie Type':'Audio'})\n print('Button Pressed - Audio: %s' % 'VGA')\n\n elif button is BTN['XShare'] and state == 'Pressed':\n ## ShareLink to HDMI Audio Dembedder Input - Audio\n MATRIX.Set('MatrixTieCommand', None, {'Input':'4', 'Output':'1', 'Tie Type':'Audio'})\n print('Button Pressed - Audio: %s' % 'ShareLink')\n pass\n\n@event(BTNPAGE['Audio2'], BTNSTATE['List'])\ndef audio_vol_events(button, state):\n \"\"\"User Actions: Touch Audio Page\"\"\"\n\n ## Data of current Biamp Block Gain\n global CURRENTLVL1\n global CURRENTLVL2\n CURRENTLVL1 = BIAMP_DATA['lvl_spk']\n CURRENTLVL2 = CISCO_DATA['Volume']\n\n ## Audio Speaker: Vol -\n if button is BTN['XSpkLess']:\n if state == 'Pressed' or state == 'Repeated':\n CURRENTLVL1 -= 5 ## Decrease 5 dB\n if CURRENTLVL1 < -100:\n print('Biamp Minimun gain')\n else:\n BIAMP.Set('LevelControl', CURRENTLVL1, {'Instance Tag':'lvl_spk', 'Channel':'1'})\n LVL['Spk'].SetLevel(CURRENTLVL1)\n BTN['XSpkLess'].SetState(1)\n else:\n BTN['XSpkLess'].SetState(0)\n print('Button Pressed - Audio: %s' % 'Spk-')\n\n ## Audio Speaker: Vol +\n elif button is BTN['XSpkPlus']:\n if state == 'Pressed' or state == 'Repeated':\n CURRENTLVL1 += 5 ## Increase 5 dB\n if CURRENTLVL1 > 12:\n print('Biamp Maximun gain')\n else:\n BIAMP.Set('LevelControl', CURRENTLVL1, {'Instance Tag':'lvl_spk', 'Channel':'1'})\n LVL['Spk'].SetLevel(CURRENTLVL1)\n BTN['XSpkPlus'].SetState(1)\n else:\n BTN['XSpkPlus'].SetState(0)\n print('Button Pressed - Audio: %s' % 'Spk+')\n\n ## Audio VC Remote: Vol -\n if button is BTN['XVCLess']:\n if state == 'Pressed' or state == 'Repeated':\n CURRENTLVL2 -= 5 ## Decrease 5 dB\n if CURRENTLVL2 < 0:\n print('VC Minimun gain')\n else:\n CISCO.Set('Volume', CURRENTLVL2)\n LVL['VC'].SetLevel(CURRENTLVL2)\n BTN['XVCLess'].SetState(1)\n else:\n BTN['XVCLess'].SetState(0)\n print('Button Pressed - Audio: %s' % 'VC-')\n\n ## Audio VC Remote: Vol +\n if button is BTN['XVCPlus']:\n if state == 'Pressed' or state == 'Repeated':\n CURRENTLVL2 += 5 ## Increase 5 dB\n if CURRENTLVL2 < 0:\n print('VC Maximun gain')\n else:\n CISCO.Set('Volume', CURRENTLVL2)\n LVL['VC'].SetLevel(CURRENTLVL2)\n BTN['XVCPlus'].SetState(1)\n else:\n BTN['XVCPlus'].SetState(0)\n print('Button Pressed - Audio: %s' % 'VC+')\n pass\n\n@event(BTNPAGE['Audio3'], BTNSTATE['List'])\ndef audio_mute_events(button, state):\n \"\"\"User Actions: Touch Audio Page\"\"\"\n\n ## Mute Speaker Audio Control\n if button is BTN['XSpk'] and state == 'Pressed':\n if BIAMP_DATA['MuteSpk'] == True:\n BIAMP.Set('MuteControl', 'Off', {'Instance Tag':'lvl_spk', 'Channel':'1'})\n elif BIAMP_DATA['MuteSpk'] == False:\n BIAMP.Set('MuteControl', 'On', {'Instance Tag':'lvl_spk', 'Channel':'1'})\n print('Button Pressed - Audio: %s' % 'Mute Spk')\n\n ## Mute VC Remote Audio Control\n elif button is BTN['XVC'] and state == 'Pressed':\n if BIAMP_DATA['MuteVCRx'] == True:\n BIAMP.Set('MuteControl', 'Off', {'Instance Tag':'lvl_vcrx', 'Channel':'1'})\n elif BIAMP_DATA['MuteVCRx'] == False:\n BIAMP.Set('MuteControl', 'On', {'Instance Tag':'lvl_vcrx', 'Channel':'1'})\n print('Button Pressed - Audio: %s' % 'Mute VC')\n\n ## Mute All Mics Audio Control\n elif button is BTN['XMics'] and state == 'Pressed':\n if BIAMP_DATA['Mute_Mics'] == True:\n BIAMP.Set('MuteControl', 'Off', {'Instance Tag':'mute_mix', 'Channel':'1'})\n elif BIAMP_DATA['Mute_Mics'] == False:\n BIAMP.Set('MuteControl', 'On', {'Instance Tag':'mute_mix', 'Channel':'1'})\n print('Button Pressed - Audio: %s' % 'Mute Mics')\n pass\n\n## Lights PAGE -----------------------------------------------------------------\n@event(BTNPAGE['Lights'], BTNSTATE['List'])\ndef lights_events(button, state):\n \"\"\"User Actions: Touch Lights Page\"\"\"\n\n if button is BTN['Escene1'] and state == 'Pressed':\n ## All Lights Off\n LUTRON.Set('4ButtonPicoControls', 'Press', {'Integration ID':'2', 'Button':'4'})\n print('Button Pressed - Lights: %s' % 'Escene 1')\n\n elif button is BTN['Escene2'] and state == 'Pressed':\n ## Black soft Lights\n LUTRON.Set('4ButtonPicoControls', 'Press', {'Integration ID':'2', 'Button':'3/Lower'})\n print('Button Pressed - Lights: %s' % 'Escene 2')\n\n elif button is BTN['Escene3'] and state == 'Pressed':\n ## White soft Lights\n LUTRON.Set('4ButtonPicoControls', 'Press', {'Integration ID':'2', 'Button':'2/Raise'})\n print('Button Pressed - Lights: %s' % 'Escene 3')\n\n elif button is BTN['Escene4'] and state == 'Pressed':\n ## All Lights On\n LUTRON.Set('4ButtonPicoControls', 'Press', {'Integration ID':'2', 'Button':'1'})\n print('Button Pressed - Lights: %s' % 'Escene 4')\n\n ## Mutually Exclusive\n BTNGROUP['Lights'].SetCurrent(button)\n pass\n\n## Blinds PAGE -----------------------------------------------------------------\n@event(BTNPAGE['Blinds'], BTNSTATE['List'])\ndef lights_events(button, state):\n \"\"\"User Actions: Touch Blinds Page\"\"\"\n\n if button is BTN['BlindsUp'] and state == 'Pressed' or state == 'Repeated':\n ## Blinds Up\n #SOMFY.Set('Tilt', 'Up', {'Channel':'1', 'Amplitude':1})\n BTNGROUP['Blinds'].SetCurrent(button)\n print('Button Pressed - Lights: %s' % 'Blinds Up')\n\n elif button is BTN['BlindsSt'] and state == 'Pressed':\n ## Blinds Stop\n #SOMFY.Set('Position', 'Stop', {'Channel':'1'})\n BTNGROUP['Blinds'].SetCurrent(button)\n print('Button Pressed - Lights: %s' % 'Blinds Stop')\n\n elif button is BTN['BlindsDw'] and state == 'Pressed' or state == 'Repeated':\n ## Blinds Down\n #SOMFY.Set('Tilt', 'Down', {'Channel':'1', 'Amplitude':1})\n BTNGROUP['Blinds'].SetCurrent(button)\n print('Button Pressed - Lights: %s' % 'Blinds Down')\n \n ## Mutually Exclusive\n BTNGROUP['Blinds'].SetCurrent(button)\n pass\n\n## Status PAGE -----------------------------------------------------------------\n\n## Power PAGE ------------------------------------------------------------------\n@event(BTN['PowerAll'], BTNSTATE['List'])\ndef power_events(button, state):\n \"\"\"User Actions: Touch PowerOff Page\"\"\"\n\n global PWRCOUNT\n ## If the user press the Power Button:\n ## Only Turn On the first state of button - Does not do any action\n if state == 'Pressed':\n BTN['PowerAll'].SetState(4)\n print('Button Pressed: %s' % 'PowerAll')\n\n ## If the user holds down the button:\n ## A variable is Decremented from 4 to 0 seconds\n ## In each new value, Turn On each visual state of the Power Button\n ## Whne the value is equal to 0, ShutDown all devices in the System\n elif state == 'Repeated':\n PWRCOUNT = PWRCOUNT - 1\n BTN['PowerAll'].SetState(PWRCOUNT)\n LBL['CountAll'].SetText(str(PWRCOUNT))\n print('Button Repeated: %s' % 'PowerAll')\n ## SHUTDOWN ALL DEVICES\n if PWRCOUNT == 0:\n TLP.ShowPage(PAGE['Index'])\n\n ## If the user release the Button:\n ## Clean the counter power data in GUI and delete the visual feedback\n elif state == 'Released':\n PWRCOUNT = 4\n BTN['PowerAll'].SetState(0)\n LBL['CountAll'].SetText('')\n print('Button Released: %s' % 'PowerAll')\n pass\n\n## End Events Definitions-------------------------------------------------------\ninitialize()\n","repo_name":"dyanko14/ACT_HQ_MeetingRoom","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":60999,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"25751998840","text":"import collections, math, bisect, heapq, random, functools, itertools, copy, typing\nimport platform; LOCAL = (platform.uname().node == 'AMO')\n\n\nimport sys; input = lambda: sys.stdin.readline().rstrip(\"\\r\\n\")\ninp = lambda f=int: list(map(f, input().split()))\n\ndef make_arr(*args):\n def func(x):\n if len(args) == 1: return [x() for _ in range(args[0])]\n return [make_arr(*args[1:])(x) for _ in range(args[0])]\n return func\n\ndef debug(*args):\n if LOCAL:\n print('\\033[92m', end='')\n printf(*args)\n print('\\033[0m', end='')\n\ndef printf(*args):\n if LOCAL:\n print('>>>: ', end='')\n for arg in args:\n if isinstance(arg, typing.Iterable) and \\\n not isinstance(arg, str) and \\\n not isinstance(arg, dict):\n print(' '.join(map(str, arg)), end=' ')\n else:\n print(arg, end=' ')\n print()\n\n# avaliable on Google, AtCoder\n# sys.setrecursionlimit(10**6)\n# import numpy as np\n# import scipy\n\n# d4 = [(1,0),(0,1),(-1,0),(0,-1)]\n# d8 = [(1,0),(1,1),(0,1),(-1,1),(-1,0),(-1,-1),(0,-1),(1,-1)]\n# d6 = [(2,0),(1,1),(-1,1),(-2,0),(-1,-1),(1,-1)] # hexagonal layout\n\ndef solve(cas):\n n, m = inp()\n G = [[] for _ in range(n+1)]\n for _ in range(m):\n a, b, x, y = inp()\n G[a].append((b, x, y))\n G[b].append((a, -x, -y))\n \n pos = [None] * (n+1)\n pos[1] = (0, 0)\n def bfs():\n q = collections.deque([1])\n while q:\n u = q.popleft()\n for v, x, y in G[u]:\n if pos[v] == None:\n pos[v] = (pos[u][0] + x, pos[u][1] + y)\n q.append(v)\n else:\n if pos[v] != 'undecidable':\n if pos[v][0] != pos[u][0] + x or pos[v][1] != pos[u][1] + y:\n pos[v] = 'undecidable'\n \n \n bfs()\n for i in range(1, n+1):\n if pos[i] is None:\n print('undecidable')\n else:\n printf(pos[i])\n \n\ncas = 1\nfor _ in range(cas):\n solve(_)\n\n","repo_name":"amomorning/online-challenges","sub_path":"atcoder/abc320/d.py","file_name":"d.py","file_ext":"py","file_size_in_byte":2069,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"34804790504","text":"\"\"\"Data models for Flask Cafe\"\"\"\n\n\nfrom flask_bcrypt import Bcrypt\nfrom flask_sqlalchemy import SQLAlchemy\n\n\nbcrypt = Bcrypt()\ndb = SQLAlchemy()\n\nDEFAULT_IMG_URL = \"/static/images/default-pic.png\"\n\n\nclass User(db.Model):\n \"\"\"A user.\"\"\"\n\n __tablename__ = \"users\"\n\n id = db.Column(db.Integer, primary_key=True, autoincrement=True)\n\n username = db.Column(db.String(25), nullable=False, unique=True)\n\n admin = db.Column(db.Boolean, nullable=False, default=False)\n\n email = db.Column(\n db.String(50),\n nullable=False,\n )\n\n first_name = db.Column(\n db.String(25),\n nullable=False,\n )\n\n last_name = db.Column(\n db.String(25),\n nullable=False,\n )\n\n description = db.Column(\n db.Text,\n nullable=False,\n )\n\n image_url = db.Column(\n db.Text,\n nullable=False,\n default=DEFAULT_IMG_URL,\n )\n\n password = db.Column(db.Text, nullable=False)\n\n def get_full_name(self):\n return f\"{self.first_name} {self.last_name}\"\n\n @classmethod\n def register(\n cls,\n username,\n password,\n description,\n first_name,\n last_name,\n email,\n image_url=DEFAULT_IMG_URL,\n admin=False,\n ):\n \"\"\"Sign up user.\n\n Hashes password and adds user to system.\n \"\"\"\n\n hashed_pwd = bcrypt.generate_password_hash(password).decode(\"UTF-8\")\n user = User(\n username=username,\n password=hashed_pwd,\n first_name=first_name,\n last_name=last_name,\n email=email,\n image_url=image_url,\n description=description,\n admin=admin,\n )\n db.session.add(user)\n return user\n\n @classmethod\n def authenticate(cls, username, password):\n \"\"\"Find user with `username` and `password`.\n\n This is a class method (call it on the class, not an individual user.)\n It searches for a user whose password hash matches this password\n and, if it finds such a user, returns that user object.\n\n If this can't find matching user (or if password is wrong), returns\n False.\n \"\"\"\n\n user = cls.query.filter_by(username=username).first()\n\n if user:\n is_auth = bcrypt.check_password_hash(user.password, password)\n if is_auth:\n return user\n\n return False\n\n\nclass City(db.Model):\n \"\"\"Cities for cafes.\"\"\"\n\n __tablename__ = \"cities\"\n\n code = db.Column(\n db.Text,\n primary_key=True,\n )\n\n name = db.Column(\n db.Text,\n nullable=False,\n )\n\n state = db.Column(\n db.String(2),\n nullable=False,\n )\n\n\nclass Cafe(db.Model):\n \"\"\"Cafe information.\"\"\"\n\n __tablename__ = \"cafes\"\n\n id = db.Column(\n db.Integer,\n primary_key=True,\n )\n\n name = db.Column(\n db.Text,\n nullable=False,\n )\n\n description = db.Column(\n db.Text,\n nullable=False,\n )\n\n url = db.Column(\n db.Text,\n nullable=False,\n )\n\n address = db.Column(\n db.Text,\n nullable=False,\n )\n\n city_code = db.Column(\n db.Text,\n db.ForeignKey(\"cities.code\"),\n nullable=False,\n )\n\n image_url = db.Column(\n db.Text,\n nullable=False,\n default=\"/static/images/default-cafe.jpg\",\n )\n\n city = db.relationship(\"City\", backref=\"cafes\")\n\n def __repr__(self):\n return f''\n\n def get_city_state(self):\n \"\"\"Return 'city, state' for cafe.\"\"\"\n\n city = self.city\n return f\"{city.name}, {city.state}\"\n\n\ndef connect_db(app):\n \"\"\"Connect this database to provided Flask app.\n\n You should call this in your Flask app.\n \"\"\"\n\n app.app_context().push()\n db.app = app\n db.init_app(app)\n","repo_name":"danielzeljko/flask-cafe","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3880,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"23577319535","text":"import json\n\n\ndef hello(event, context):\n data = json.loads(event['body'])\n\n print(type(data))\n print(data)\n\n body = {\n \"type\": type(data).__name__,\n \"data\": data\n }\n\n response = {\n \"statusCode\": 200,\n \"body\": json.dumps(body)\n }\n\n return response\n","repo_name":"fbaltor/principia-challenge","sub_path":"api/handler.py","file_name":"handler.py","file_ext":"py","file_size_in_byte":300,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6344345547","text":"import tkinter as tk\r\nfrom tkinter import messagebox\r\nimport sqlite3\r\n\r\ndef criar_tabela():\r\n conn = sqlite3.connect('escola.db')\r\n c = conn.cursor()\r\n\r\n c.execute('''CREATE TABLE IF NOT EXISTS alunos (\r\n id INTEGER PRIMARY KEY,\r\n nome TEXT,\r\n presencas INTEGER,\r\n faltas INTEGER,\r\n nota1 REAL,\r\n nota2 REAL,\r\n media REAL,\r\n frequencia REAL\r\n )''')\r\n\r\n conn.commit()\r\n conn.close()\r\n\r\ndef calcular_situacao(media, frequencia):\r\n if media >= 6.0 and frequencia >= 75:\r\n return \"Aprovado\"\r\n else:\r\n return \"Reprovado\"\r\n\r\ndef adicionar_aluno():\r\n nome = nome_entry.get()\r\n presencas = int(presencas_entry.get())\r\n faltas = int(faltas_entry.get())\r\n nota1 = float(nota1_entry.get())\r\n nota2 = float(nota2_entry.get())\r\n\r\n conn = sqlite3.connect('escola.db')\r\n c = conn.cursor()\r\n\r\n media = (nota1 + nota2) / 2\r\n total_aulas = presencas + faltas\r\n frequencia = (presencas / total_aulas) * 100\r\n\r\n situacao = calcular_situacao(media, frequencia)\r\n\r\n c.execute(\"INSERT INTO alunos (nome, presencas, faltas, nota1, nota2, media, frequencia) VALUES (?, ?, ?, ?, ?, ?, ?)\",\r\n (nome, presencas, faltas, nota1, nota2, media, frequencia))\r\n\r\n conn.commit()\r\n conn.close()\r\n\r\n messagebox.showinfo(\"Sucesso\", f\"Aluno {nome} adicionado com sucesso. Situação: {situacao}\")\r\n\r\n# Cria a janela principal\r\nroot = tk.Tk()\r\nroot.title(\"Controle de Alunos\")\r\n\r\n# Cria os widgets da interface\r\nnome_label = tk.Label(root, text=\"Nome:\")\r\nnome_label.pack()\r\nnome_entry = tk.Entry(root)\r\nnome_entry.pack()\r\n\r\npresencas_label = tk.Label(root, text=\"Presenças:\")\r\npresencas_label.pack()\r\npresencas_entry = tk.Entry(root)\r\npresencas_entry.pack()\r\n\r\nfaltas_label = tk.Label(root, text=\"Faltas:\")\r\nfaltas_label.pack()\r\nfaltas_entry = tk.Entry(root)\r\nfaltas_entry.pack()\r\n\r\nnota1_label = tk.Label(root, text=\"Nota 1:\")\r\nnota1_label.pack()\r\nnota1_entry = tk.Entry(root)\r\nnota1_entry.pack()\r\n\r\nnota2_label = tk.Label(root, text=\"Nota 2:\")\r\nnota2_label.pack()\r\nnota2_entry = tk.Entry(root)\r\nnota2_entry.pack()\r\n\r\nadicionar_button = tk.Button(root, text=\"Adicionar Aluno\", command=adicionar_aluno)\r\nadicionar_button.pack()\r\n\r\nif __name__ == \"__main__\":\r\n criar_tabela()\r\n root.mainloop()\r\n","repo_name":"dbcfilho/average_database","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":2422,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18378421100","text":"import numpy as np\nimport math\nimport tensorflow.compat.v1 as tf\n\nclass CosineAnnealer:\n\n def __init__(self, start, end, steps):\n self.start = start\n self.end = end\n self.steps = steps\n self.n = 0\n\n def step(self):\n self.n += 1\n cos = tf.math.cos(np.pi * (self.n / self.steps)) + 1\n return self.end + (self.start - self.end) / 2. * cos\n\n def getval(self,step):\n cos = tf.math.cos(np.pi * (step / self.steps)) + 1\n return self.end + (self.start - self.end) / 2. * cos\n\nclass OneCycleScheduler():\n \"\"\"\n From https://www.avanwyk.com/tensorflow-2-super-convergence-with-the-1cycle-policy/, by Andrich van Wyk modified from\n fastai lib. Modified again to apply for TPU code.\n \"\"\"\n\n def __init__(self, lr_max, steps, mom_min=0.85, mom_max=0.95, phase_1_pct=0.45, div_factor=25.):\n super(OneCycleScheduler, self).__init__()\n lr_min = lr_max / div_factor\n final_lr = lr_max / (div_factor * 1e4)\n phase_1_steps = steps * phase_1_pct\n phase_2_steps = steps - phase_1_steps\n\n self.phase_1_steps = phase_1_steps\n self.phase_2_steps = phase_2_steps\n self.phase = 0\n self.step = 0\n\n self.phases = [[CosineAnnealer(lr_min, lr_max, phase_1_steps), CosineAnnealer(mom_max, mom_min, phase_1_steps)],\n [CosineAnnealer(lr_max, final_lr, phase_2_steps),\n CosineAnnealer(mom_min, mom_max, phase_2_steps)]]\n\n self.lrs = []\n self.moms = []\n\n @tf.function\n def getlrmom(self,step):\n if step= {0} - 22d AND time < {0} GROUP BY time(1h)'.format(date)) # Query written in InfluxQL\n points = results.get_points()\n values = results.raw[\"series\"][0][\"values\"]\n columns = results.raw[\"series\"][0][\"columns\"]\n df = pd.DataFrame(values, columns=columns)\n df.index = [parser.parse(d) for d in df[\"time\"].values]\n split_time = df.index[-1]-td(days=1)\n\n train = df.loc[:split_time]\n test = df.loc[split_time+td(seconds=1):]\n train_x, train_y = split_labels(train)\n test_x, test_y = split_labels(test)\n return train_x, train_y, test_x, test_y\n\n\ndef split_labels(df):\n x = df[[\"time\"]].rename(columns={\"time\": \"Time\"})\n y = df[[\"mean\"]].rename(columns={\"mean\": \"Demand\"})\n x.index = list(range(len(x)))\n y.index = list(range(len(y)))\n return x,y\n\n\nif __name__ == \"__main__\":\n train_x, train_y, test_x, test_y = get_train_test(date=\"2020-03-01\")\n print(train_x,train_y) \n\n","repo_name":"NielsOerbaek/caiso-mlflow","sub_path":"prepros.py","file_name":"prepros.py","file_ext":"py","file_size_in_byte":1554,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"3709630522","text":"import hashlib\nimport ipaddress\nimport json\nimport socket\nimport sys\nimport threading\nimport time\nimport uuid\n\nfrom . import crypto_funcs as cf\n\nmsg_del_time = 30\nPORT = 65432\n\n\nclass NodeConnection(threading.Thread):\n def __init__(self, main_node, sock, id, host, port):\n\n super(NodeConnection, self).__init__()\n\n self.host = host\n self.port = port\n self.main_node = main_node\n self.sock = sock\n self.terminate_flag = threading.Event()\n self.last_ping = time.time()\n # Variable for parsing the incoming json messages\n self.buffer = \"\"\n\n # The id of the connected node\n self.public_key = cf.load_key(id)\n self.id = id\n\n self.main_node.debug_print(\"Connection \" + self.host + \":\" + str(self.port))\n\n def send(self, data):\n try:\n self.sock.sendall(data.encode(\"utf-8\"))\n\n except Exception as e:\n self.main_node.debug_print(\"Exception: \" + str(e))\n self.terminate_flag.set()\n\n def stop(self):\n self.terminate_flag.set()\n\n def run(self):\n self.sock.settimeout(10.0)\n\n while not self.terminate_flag.is_set():\n if time.time() - self.last_ping > self.main_node.dead_time:\n self.terminate_flag.set()\n print(\"node\" + self.id + \" is dead\")\n\n try:\n message = self.sock.recv(4096)\n if message == \"ping\":\n self.last_ping = time.time()\n else:\n self.main_node.node_message(self, message)\n\n except socket.timeout:\n pass\n\n except Exception as e:\n self.terminate_flag.set()\n self.main_node.debug_print(e)\n\n time.sleep(0.01)\n\n self.main_node.node_disconnected(self)\n self.sock.settimeout(None)\n self.sock.close()\n del self.main_node.nodes_connected[self.main_node.nodes_connected.index(self)]\n time.sleep(1)\n\n\nclass Node(threading.Thread):\n def __init__(self, host=\"\", port=65432):\n super(Node, self).__init__()\n\n self.terminate_flag = threading.Event()\n self.pinger = Pinger(self) # start pinger\n self.debug = True\n\n self.dead_time = 45 # time to disconect from node if not pinged, nodes ping after 20s\n\n self.host = host\n self.ip = host # own ip, will be changed by connection later\n self.port = port\n\n self.nodes_connected = []\n\n self.msgs = {} # hashes of recieved messages\n self.peers = []\n\n self.id = uuid.uuid4()\n\n self.max_peers = 10\n\n hostname = socket.gethostname()\n\n self.local_ip = socket.gethostbyname(hostname)\n\n self.banned = []\n\n self.sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n self.debug_print(\"Initialisation of the Node on port: \" + str(self.port))\n self.sock.bind((self.host, self.port))\n self.sock.settimeout(10.0)\n self.sock.listen(1)\n\n def debug_print(self, msg):\n if self.debug:\n print(\"[debug] \" + str(msg))\n\n def network_send(self, message, exc=[]):\n for i in self.nodes_connected:\n if i.host not in exc:\n i.send(json.dumps(message))\n\n def connect_to(self, host, port=PORT):\n\n if not self.check_ip_to_connect(host):\n self.debug_print(\"connect_to: Cannot connect!!\")\n return False\n\n if len(self.nodes_connected) >= self.max_peers:\n self.debug_print(\"Peers limit reached.\")\n return True\n\n for node in self.nodes_connected:\n if node.host == host:\n print(\"[connect_to]: Already connected with this node.\")\n return True\n\n try:\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n sock.connect((host, port))\n\n sock.send(self.id.encode(\"utf-8\"))\n connected_node_id = sock.recv(1024).decode(\"utf-8\")\n\n if self.id == connected_node_id:\n self.debug_print(\"Possible own ip: \" + host)\n if ipaddress.ip_address(host).is_private:\n self.local_ip = host\n else:\n self.ip = host\n self.banned.append(host)\n sock.close()\n return False\n\n thread_client = self.create_new_connection(sock, connected_node_id, host, port)\n thread_client.start()\n self.nodes_connected.append(thread_client)\n self.node_connected(thread_client)\n\n except Exception as e:\n self.debug_print(\"connect_to: Could not connect with node. (\" + str(e) + \")\")\n\n def create_new_connection(self, connection, id, host, port):\n return NodeConnection(self, connection, id, host, port)\n\n def stop(self):\n self.terminate_flag.set()\n\n def run(self):\n self.pinger.start()\n while not self.terminate_flag.is_set():\n try:\n connection, client_address = self.sock.accept()\n\n connected_node_id = connection.recv(2048).decode(\"utf-8\")\n connection.send(self.id.encode(\"utf-8\"))\n\n if self.id != connected_node_id:\n thread_client = self.create_new_connection(\n connection,\n connected_node_id,\n client_address[0],\n client_address[1],\n )\n thread_client.start()\n\n self.nodes_connected.append(thread_client)\n\n self.node_connected(thread_client)\n\n else:\n connection.close()\n\n except socket.timeout:\n pass\n\n except Exception as e:\n raise e\n\n time.sleep(0.01)\n\n self.pinger.stop()\n for t in self.nodes_connected:\n t.stop()\n\n self.sock.close()\n print(\"Node stopped\")\n\n def ConnectToNodes(self):\n for i in self.peers:\n if not self.connect_to(i, PORT):\n # delete wrong / own ip from peers\n del self.peers[self.peers.index(i)]\n\n def message(self, type, data, ex=[]):\n # time that the message was sent\n dict = {\"type\": type, \"data\": data}\n if \"time\" not in dict:\n dict[\"time\"] = str(time.time())\n\n self.network_send(dict, ex)\n\n def send_peers(self):\n self.message(\"peers\", self.peers)\n\n def check_validity(self, msg):\n if not (\"time\" in msg and \"type\" in msg and \"data\" in msg):\n return False\n\n return True\n\n def check_expired(self, dta):\n sth = str(dta)\n msghash = hashlib.md5(sth.encode(\"utf-8\")).hexdigest().decode()\n\n if float(time.time()) - float(dta[\"time\"]) < float(msg_del_time):\n if msghash not in self.msgs:\n self.msgs[msghash] = time.time()\n return False\n else:\n # if message is expired\n self.debug_print(\"expired:\" + dta[\"msg\"])\n return True\n\n def announce(self, dta, n):\n self.message(dta[\"type\"], dta[\"data\"], dta, ex=n)\n if len(self.msgs) > len(self.peers) * 20:\n for i in self.msgs.copy():\n if time.time() - self.msgs[i] > msg_del_time:\n del self.msgs[i]\n\n def data_handler(self, dta, n):\n if self.check_expired(dta):\n return False\n else:\n self.announce(dta, n)\n\n type = dta[\"type\"]\n data = dta[\"data\"]\n\n if type == \"peers\":\n # peers handling\n for i in data:\n if self.check_ip_to_connect(i):\n self.peers.append(i)\n\n self.debug_print(\"Known Peers: \" + str(self.peers))\n self.ConnectToNodes() # cpnnect to new nodes\n return True\n\n else:\n self.on_message(dta)\n\n def check_ip_to_connect(self, ip):\n if (\n ip not in self.peers\n and ip != \"\"\n and ip != self.ip\n and ip != self.local_ip\n and ip not in self.banned\n ):\n return True\n else:\n return False\n\n def on_message(self, data):\n self.debug_print(\"Incomig Message: \" + data)\n\n def on_connect(self, n):\n pass\n\n def loadstate(self, file=\"state.json\"):\n with open(file, \"r\") as f:\n peers = json.load(f)\n for i in peers:\n self.connect_to(i)\n\n def savestate(self, file=\"state.json\"):\n with open(file, \"w+\") as f:\n json.dump(self.peers, f)\n\n def node_connected(self, node):\n self.debug_print(\"node_connected: \" + node.id)\n if node.host not in self.peers:\n self.peers.append(node.host)\n self.send_peers()\n self.on_connect(node)\n\n def node_disconnected(self, node):\n self.debug_print(\"node_disconnected: \" + node.id)\n if node.host in self.peers:\n self.peers.remove(node.host)\n\n def node_message(self, node, data):\n try:\n json.loads(data)\n except json.decoder.JSONDecodeError:\n self.debug_print(f\"Error loading message from {node.id}\")\n return\n self.data_handler(json.loads(data), [node.host, self.ip])\n\n\nclass Pinger(threading.Thread):\n def __init__(self, parent):\n self.terminate_flag = threading.Event()\n super(Pinger, self).__init__()\n self.parent = parent\n self.dead_time = 30 # time to disconect from node if not pinged\n\n def stop(self):\n self.terminate_flag.set()\n\n def run(self):\n print(\"Pinger Started\")\n while not self.terminate_flag.is_set(): # Check whether the thread needs to be closed\n for i in self.parent.nodes_connected:\n i.send(\"ping\")\n time.sleep(20)\n print(\"Pinger stopped\")\n","repo_name":"GianisTsol/cryptocoin","sub_path":"cryptocoin/p2pbase.py","file_name":"p2pbase.py","file_ext":"py","file_size_in_byte":10066,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"2285574018","text":"#!/usr/bin/env python\n\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\nfrom matplotlib.colors import LogNorm, ListedColormap,BoundaryNorm\n\nimport numpy as np\nimport datetime as dt\nimport os, pickle\nfrom scipy.ndimage.filters import gaussian_filter\nfrom scipy.stats import pearsonr\nimport pandas as pd\nfrom mpl_toolkits.basemap import *\nfrom sklearn.calibration import CalibratedClassifierCV, calibration_curve\nfrom sklearn import metrics\nfrom keras.models import Model, model_from_json, save_model, load_model\nfrom keras.layers import Dense, Activation, Conv2D, Input, AveragePooling2D, Flatten, LeakyReLU\nfrom keras.layers import Dropout, BatchNormalization\nfrom keras.regularizers import l2\nfrom keras.optimizers import SGD, Adam\nimport keras.backend as K\nimport tensorflow as tf\n\ndef computeSTP(row):\n lclterm = ((2000.0-row['MLLCL-potential_mean'])/1000.0)\n lclterm = np.where(row['MLLCL-potential_mean']<1000, 1.0, lclterm)\n lclterm = np.where(row['MLLCL-potential_mean']>2000, 0.0, lclterm)\n\n shrterm = (row['shr06']/20.0)\n shrterm = np.where(row['shr06'] > 30, 1.5, shrterm)\n shrterm = np.where(row['shr06'] < 12.5, 0.0, shrterm)\n\n stp = (row['SBCAPE-potential_mean']/1500.0) * lclterm * (row['SRH01-potential_mean']/150.0) * shrterm\n return stp\n\ndef read_csv_files():\n # read in all CSV files for 1km forecasts\n tdate = sdate\n all_files = []\n while tdate <= edate:\n yyyymmdd = tdate.strftime('%Y%m%d')\n csv_file = '/glade/work/sobash/NSC_objects/grid_data_ncarstorm_3km_csv_preprocessed/grid_data_NCARSTORM_d01_%s-0000.csv.gz'%(yyyymmdd)\n\n if os.path.exists(csv_file): all_files.append(csv_file)\n tdate += dateinc\n print('Reading %s files'%(len(all_files)))\n\n df = pd.concat((pd.read_csv(f, compression='gzip') for f in all_files))\n\n #if model == 'NSC': df['stp'] = df.apply(computeSTP, axis=1) \n\n #if model == 'NSC': df['datetime'] = pd.to_datetime(df['Valid_Date'])\n #if model == 'NCAR':\n df['datetime'] = pd.to_datetime(df['Date'])\n #df['Run_Date'] = pd.to_datetime(df['Date']) - pd.to_timedelta(df['fhr'])\n df['year'] = df['datetime'].dt.year\n df['month'] = df['datetime'].dt.month\n df['hour'] = df['datetime'].dt.hour\n df['dayofyear'] = df['datetime'].dt.dayofyear\n return df, len(all_files)\n\ndef brier_score_keras(obs, preds):\n return K.mean((preds - obs) ** 2)\n\ndef brier_skill_score_keras(obs, preds):\n climo = K.mean((obs - K.mean(obs)) ** 2)\n bs = brier_score_keras(obs, preds)\n ratio = (bs / climo)\n return climo\n\ndef auc(obs, preds):\n auc = tf.metrics.auc(obs, preds)[1]\n K.get_session().run(tf.local_variables_initializer())\n return auc\n \ndef normalize_multivariate_data(data, scaling_values=None):\n \"\"\"\n Normalize each channel in the 4 dimensional data matrix independently.\n\n Args:\n data: 4-dimensional array with dimensions (example, y, x, channel/variable)\n scaling_values: pandas dataframe containing mean and std columns\n\n Returns:\n normalized data array, scaling_values\n \"\"\"\n print(data.shape, data.dtype)\n normed_data = np.zeros(data.shape, dtype=data.dtype)\n scale_cols = [\"mean\", \"std\"]\n if scaling_values is None:\n scaling_values = pd.DataFrame(np.zeros((data.shape[-1], len(scale_cols)), dtype=np.float32),\n columns=scale_cols)\n for i in range(data.shape[-1]): scaling_values.loc[i, [\"mean\", \"std\"]] = [data[:, i].mean(), data[:, i].std()]\n\n for i in range(data.shape[-1]):\n normed_data[:, i] = (data[:, i] - scaling_values.loc[i, \"mean\"]) / scaling_values.loc[i, \"std\"]\n\n return normed_data, scaling_values\n\ndef plot_forecast(predictions, prefix=\"\"):\n #test = readNCLcm('MPL_Greys')[25::] + [[1,1,1]] + readNCLcm('MPL_Reds')[10::]\n #test = readNCLcm('perc2_9lev')[1::]\n #cmap = ListedColormap(test)\n cmap = plt.get_cmap('RdGy_r')\n norm = BoundaryNorm(np.arange(0,1.1,0.1), ncolors=cmap.N, clip=True)\n\n print(predictions)\n\n #awips = Basemap(projection='lcc', llcrnrlon=-133.459, llcrnrlat=12.19, urcrnrlon=-49.38641, urcrnrlat=57.2894, lat_1=25.0, lat_2=25.0, lon_0=-95, resolution='l', area_thresh=10000.)\n\n #fig, axes, m = pickle.load(open('/glade/u/home/sobash/NSC_scripts/ch_pk_files/rt2015_ch_CONUS.pk', 'r'))\n #fig, axes, m = pickle.load(open('/glade/u/home/sobash/NSC_scripts/dav_pk_files/rt2015_ch_CONUS.pk', 'rb'))\n fig, axes, m = pickle.load(open('rt2015_ch_CONUS.pk', 'rb')) \n\n lats, lons = predictions['lat'].values, predictions['lon'].values\n x, y = m(lons, lats)\n\n # do something convoluted here to only plot each point once\n probmax = {}\n for i,p in enumerate(predictions['predict_proba'].values):\n thiskey = '%f%f'%(lats[i],lons[i])\n if thiskey in probmax:\n if p > probmax[thiskey]:\n probmax[thiskey] = p\n else:\n probmax[thiskey] = p\n\n for i,p in enumerate(predictions['predict_proba'].values):\n thiskey = '%f%f'%(lats[i],lons[i])\n thisvalue = probmax[thiskey]\n\n color = cmap(norm([thisvalue])[0])\n probmax[thiskey] = -999\n if thisvalue >= 0.05:\n a = plt.text(x[i], y[i], int(round(thisvalue*100)), fontsize=10, ha='center', va='center', family='monospace', color=color, fontweight='bold')\n #a = m.scatter(x, y, s=50, c=predictions['predict_proba'].values, lw=0.5, edgecolors='k', cmap=cmap, norm=norm)\n\n # ADD COLORBAR\n #cax = fig.add_axes([0.02,0.1,0.02,0.3])\n #cb = plt.colorbar(a, cax=cax, orientation='vertical', extendfrac=0.0)\n #cb.outline.set_linewidth(0.5)\n #cb.ax.tick_params(labelsize=10)\n\n plt.savefig('forecast%s.png'%prefix)\n\nsdate = dt.datetime(2012,6,29,0,0,0)\nedate = dt.datetime(2012,6,29,0,0,0)\ndateinc = dt.timedelta(days=1)\ndf, numfcsts = read_csv_files()\n\nprint('Training random forest classifier')\n\nfeatures = ['fhr', 'dayofyear', 'lat', 'lon', 'UP_HELI_MAX', 'UP_HELI_MAX03', 'UP_HELI_MAX01', 'W_UP_MAX', 'W_DN_MAX', 'WSPD10MAX', 'MUCAPE', 'SHR06', 'MLCINH', 'MLLCL', 'SHR01', 'SRH01', 'SRH03', 'T2', 'TD2', 'PSFC','CAPESHEAR', 'STP', 'LR75']\nlarge_scale_features = ['U925','U850','U700','U500','V925','V850','V700','V500','T925','T850','T700','T500','TD925','TD850','TD700','TD500']\nneighbor_features = [ f+'-%s1'%n for f in large_scale_features for n in ['E','S','N','W'] ]\nneighbor_time_features = [ f+'-%s'%n for f in ['STP', 'CAPESHEAR', 'MUCAPE', 'SBCINH', 'MLLCL', 'SHR06', 'SHR01'] for n in ['TP1', 'TM1'] ]\nfeatures = features + large_scale_features + neighbor_features + neighbor_time_features\n\n# normalize data we want to use\nscaling_values = pickle.load(open('scaling_values.pk', 'rb'))\n\nnorm_in_data, scaling_values = normalize_multivariate_data(df[features].values, scaling_values=scaling_values)\n\n# load combined architecture and weights\ndense_model = load_model('neural_network.h5', custom_objects={'brier_score_keras': brier_score_keras, 'brier_skill_score_keras':brier_skill_score_keras, 'auc':auc })\n\npredictions_proba = dense_model.predict(norm_in_data)\nprint(predictions_proba.max())\nprint(predictions_proba) \n\n#labels: all, wind, hailone, torn\ndf['predict_proba'] = predictions_proba[:,1]\nforecast_mask = (df['fhr'] > 12)\nplot_forecast(df[forecast_mask])\n","repo_name":"ahijevyc/machine-learning","sub_path":"neural_network_forecast_gridded.py","file_name":"neural_network_forecast_gridded.py","file_ext":"py","file_size_in_byte":7303,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"79"} +{"seq_id":"9224681880","text":"import sys\n\ndef main():\n\tif len(sys.argv) > 1:\n\t\tfor i in range(1,len(sys.argv)):\n\t\t\tfilename = sys.argv[i]\n\t\t\tfile = open (filename, \"r\")\n\t\t\tcontent = file.read()\n\t\t\tprint (content)\n\t\t\tfile.close()\n\telse:\n\t\treturn \"Please enter a file\"\n\nif __name__ == '__main__':\n\tprint(main())\n","repo_name":"nsapundzhiev/HackBG","sub_path":"week2/ 2-File-System-Problems/cat2.py","file_name":"cat2.py","file_ext":"py","file_size_in_byte":280,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"29781769565","text":"from django.urls import path, include\nfrom django.conf import settings\nfrom django.contrib.auth.views import LoginView, LogoutView\nfrom rest_framework import routers\n\nfrom .views import (index, admin, avatar, login_api, user_data_api, csrf_api, basic_medium_metadata,\n responsible_ticket_metadata, responsible_metadata, basic_medium_without_certificate_metadata,\n roles_metadata, generate_fixtures)\nfrom .api import (UserViewSet, BasicMediumExpedientViewSet, RequestTicketViewSet,\n MovementTicketViewSet, ResponsibilityCertificateViewSet)\n\n\n# Incluyo el enrutador de los ViewSet declarados para el API\nrouter = routers.SimpleRouter()\nrouter.register(r'users', UserViewSet)\nrouter.register(r'basic_medium', BasicMediumExpedientViewSet)\nrouter.register(r'request_ticket', RequestTicketViewSet)\nrouter.register(r'movement_ticket', MovementTicketViewSet)\nrouter.register(r'responsibility_certificate', ResponsibilityCertificateViewSet)\n\nurlpatterns = [\n # Default System urls\n path('', index, name='index'),\n path('admin/', admin, name='dashboard'),\n\n # Authentication urls\n path('login/', LoginView.as_view(), name='login'),\n path('logout/', LogoutView.as_view(), name='logout'),\n\n # API urls\n path('api/login/', login_api, name='api_login'),\n path('api/avatar/', avatar, name=\"avatar\"),\n path('api/csrf/', csrf_api, name=\"csrf_api\"),\n path('api/user/', user_data_api, name=\"user_data\"),\n path('api/responsible/', responsible_metadata, name='responsible_list'),\n path('api/responsible_ticket/', responsible_ticket_metadata, name='responsible_ticket_list'),\n path('api/roles/', roles_metadata, name='roles_list'),\n path('api/mediums/', basic_medium_metadata, name='basic_medium_list'),\n path('api/mediums_certificate/', basic_medium_without_certificate_metadata, name='basic_medium_certificate_list'),\n path('api/', include(router.urls)),\n]\n\n# This is required for static files while in development mode. (DEBUG=TRUE)\nif settings.DEBUG:\n urlpatterns += [path('fixtures/', generate_fixtures)]\n","repo_name":"ISW-P5/Gestion-Medios-Basicos","sub_path":"system/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":2097,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"16298642126","text":"from fastapi import APIRouter, HTTPException\nfrom automart.models import VehicleMake, VehicleModel\nfrom automart.schemas import VehicleMakeView, VehicleModelView\nfrom automart.lib.nhtsa import fetch_make_models\n\nrouter = APIRouter(\n prefix=\"/vehicle-make\",\n tags=[\"Vehicle Makes\"],\n)\n\n\n@router.get(\"/\", response_model=list[VehicleMakeView])\nasync def get_vehicle_makes(page: int = 0, perPage: int = 25):\n query = VehicleMake.select().limit(perPage).offset(page * perPage)\n vehicle_makes = [VehicleMakeView.from_orm(vehicle_makes) for vehicle_makes in query]\n return vehicle_makes\n\n\n@router.get(\"/{make_id}\", response_model=None)\nasync def get_vehicle_make(vehicle_make_id: int):\n vehicle_make = VehicleMake.get_by_id(vehicle_make_id)\n\n if not vehicle_make:\n raise HTTPException(status_code=404)\n\n return VehicleMakeView.from_orm(vehicle_make)\n\n\n@router.get(\"/{make_id}/models\", response_model=list[VehicleModelView])\nasync def get_vehicle_models_by_make(vehicle_make_id: int):\n query = (\n VehicleModel.select().where(VehicleMake.id == vehicle_make_id).join(VehicleMake)\n )\n vehicle_makes = [\n VehicleModelView.from_orm(vehicle_makes) for vehicle_makes in query\n ]\n return vehicle_makes\n","repo_name":"jloiola/automart","sub_path":"automart/routes/vehicle_make.py","file_name":"vehicle_make.py","file_ext":"py","file_size_in_byte":1246,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"8359123435","text":"print(\"\\nCalculate the Greatest Common Divisor of two Integers: \\n\")\na = int(input(\"Enter first integer: \"))\nb = int(input(\"Enter second integer: \"))\nx, y = a, b\nif a == 0:\n gcd = 0 #I didn't know that 0 could be considered as a divisor,\n #the logic behind this if statement from the pseudo-algorithm is unclear\nelse:\n while b != 0:\n if a > b:\n a = a - b\n else:\n b = b - a\n gcd = a\n\nprint(\"\\ngcd(\"+ str(x)+ \", \"+ str(y)+\") = \" +str(gcd))\n\n","repo_name":"dankodak/Programmierkurs","sub_path":"Abgaben/Blatt 2/Aufgabe 2/Team 42164/Ben Romdhane_Houssem_HoussemBenRomdhane_755653/Ueb2_Auf2_HoussemBenRomdhane.py","file_name":"Ueb2_Auf2_HoussemBenRomdhane.py","file_ext":"py","file_size_in_byte":495,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"18756686206","text":"import numpy as np\r\nfrom mpl_toolkits.mplot3d import Axes3D\r\nfrom mpl_toolkits.mplot3d import proj3d\r\nfrom matplotlib.lines import Line2D\r\nfrom matplotlib.widgets import TextBox,RadioButtons\r\nfrom matplotlib.widgets import Button as widgets_Button\r\nfrom matplotlib.offsetbox import AnchoredText\r\nfrom matplotlib.backends.backend_tkagg import (FigureCanvasTkAgg, NavigationToolbar2Tk)\r\nfrom matplotlib.figure import Figure\r\nfrom matplotlib.patches import Circle, Wedge\r\nimport mpl_toolkits.mplot3d.art3d as art3d\r\n\r\nimport matplotlib.animation as animation\r\nfrom tkinter import *\r\nfrom tkinter import ttk\r\nfrom tkinter.filedialog import asksaveasfilename\r\nfrom tkinter import messagebox \r\n\r\nfrom calc_pl import *\r\nfrom help_windowAstron3D import *\r\nfrom func_tools import *\r\n\r\n#import keyboard\r\n\r\n\r\n\r\nclass c:\r\n planets={\"Sun\" : \"☉\", \"Moon\" : \"☽\", \"Mercury\" : \"☿\", \"Venus\" : \"♀\", \"Mars\" : \"♂\", \"Jupiter\" : \"♃\", \"Saturn\" : \"♄\", \"Uranus\" : \"♅\", \"Neptune\" : \"♆\", \"Pluto\" : \"♇\", \"Node_N\" : \"☊\", \"Node_S\" : \"☋\"}\r\n planets2=[\"Sun\", \"Moon\", \"Mercury\", \"Venus\", \"Mars\", \"Jupiter\", \"Saturn\", \"Uranus\", \"Neptune\", \"Pluto\", \"Node_N\", \"Node_S\", \"Asc\", \"MC\"]\r\n\r\nhouses_names=[\"\",\"I\", \"II\", \"III\", \"IV\", \"V\", \"VI\", \"VII\", \"VIII\", \"IX\", \"X\", \"XI\", \"XII\"]\r\nzodiac = [\"\",\"Aries\", \"Taurus\", \"Gemini\", \"Cancer\", \"Leo\", \"Virgo\", \"Libra\", \"Scorpio\", \"Sagittarius\", \"Capricorn\", \"Aquarius\", \"Pisces\"] \r\nzodiac2 = [\"\", \"♈\", \"♉\", \"♊\", \"♋\", \"♌\", \"♍\", \"♎\", \"♏\", \"♐\", \"♑\", \"♒\", \"♓\"]\r\n\r\ncolor_ecl=\"#A9A54B\"\r\ncolor_ecl_zod=\"#767109\"\r\ncolor_oran=\"#F38600\"\r\ncolor_yel=\"#EFC300\"\r\npl_colors={\"Sun\" : \"#F2C500\", \"Moon\" : \"#FF8D00\", \"Mercury\" : \"#008CD2\", \"Venus\" : \"#098100\", \"Mars\" : \"#AA0000\", \"Jupiter\" : \"#6E8CA5\", \"Saturn\" : \"#874400\", \"Uranus\" : \"#1900D6\", \"Neptune\" : \"#006695\", \"Pluto\" : \"#6D00A0\", \"Node_N\" : \"#333\", \"Node_S\" : \"#333\"} \r\n\r\nmain_circles_clickable = True # ecliptic, equator, horizon, prime vertical\r\nmain_circles_picker = 2 if main_circles_clickable == True else 0\r\n\r\ndata_pts=50\r\n\r\nclass astro3D():\r\n def __init__(self, parent,frame, planets_data, geo_latitude, data={}): #timestamp=0\r\n self.parent=parent\r\n bgr_col=\"#E0EFF0\"\r\n self.fig = Figure(figsize=(11,8), dpi=100, facecolor = bgr_col)\r\n self.ax = self.fig.add_subplot(111, projection='3d', facecolor = bgr_col)\r\n\r\n at = AnchoredText(\"Astronomia 3D by Popiel\", loc=\"lower right\",bbox_to_anchor=(0.9, 0.195), frameon=False,borderpad=0, prop=dict(alpha=0.2,size=13,color=\"#78A1A4\",fontfamily=\"Lucida Handwriting\"),bbox_transform=self.ax.transAxes)\r\n at.set_zorder(0)\r\n self.ax.add_artist(at)\r\n\r\n #================ \r\n\r\n size900=size760=False\r\n screenHeight = parent.winfo_screenheight()\r\n if screenHeight<880:size900=True\r\n if screenHeight<780:size760=True;size900=False\r\n\r\n f_w=11;f_h=8\r\n if size900==True: dd=0.93;self.fig.set_size_inches(f_w*dd, f_h*dd);\r\n elif size760==True: dd=0.80;self.fig.set_size_inches(f_w*dd, f_h*dd);\r\n\r\n #================ \r\n\r\n self.ids={}\r\n self.equat_ids=[]\r\n self.ecl_ids=[]\r\n self.pl_ids=[]\r\n self.leg_obj={}\r\n self.leg_items_ids=[]\r\n self.ecliptic_scale={}\r\n self.equator_scale={}\r\n\r\n\r\n #========== fix distorted circles (shown as ovals) ======\r\n limits = self.ax.get_w_lims()\r\n self.ax.set_box_aspect((limits[1]-limits[0],limits[3]-limits[2],limits[5]-limits[4]))\r\n #===============\r\n\r\n self.fig.subplots_adjust(left=-0.2, bottom=-0.3, right=0.95, top=1.25, wspace=None, hspace=None) #margins\r\n\r\n self.planets_data=planets_data\r\n self.geo_latitude=geo_latitude\r\n self.data=data\r\n self.timestampIni = self.data[\"timestamp\"]\r\n\r\n self.ε=23.44\r\n self.ε=self.data[\"obliquity\"]\r\n self.trueNode=data[\"trueNode\"]\r\n\r\n\r\n self.txt1=[0.87,0.206]\r\n self.prop_txt1={\"color\":'#333', \"fontsize\":8}\r\n self.id_text = self.ax.text2D(*self.txt1, \"\", **self.prop_txt1, transform=self.ax.transAxes)\r\n\r\n self.txt_name=[0.2,0.818]\r\n self.txt_time=[0.913,0.29]\r\n self.txt_time2=[0.911,0.26]\r\n self.txt1=[0.18,0.206]\r\n self.txt2=[0.41,0.206]\r\n self.txt3=[0.18,0.230]\r\n\r\n if size760==True:\r\n self.txt_time=[0.917+0.09,0.28]\r\n self.txt_time2=[0.915+0.09,0.25]\r\n elif size900==True:\r\n self.txt_time=[0.917+0.02,0.29]\r\n self.txt_time2=[0.915+0.02,0.26]\r\n\r\n bgr_col_bb1=\"#2D6266\"\r\n bgr_col_bb3=\"#70ADB2\"\r\n bgr_col_bb2=\"#EFE9DD\" \r\n PAGE_BG1 = \"#134752\"\r\n bgr_col_bb1=PAGE_BG1\r\n COL_2=\"#AA0000\"\r\n COL_3=\"#AB1D00\"\r\n bgr_col_bb3=\"#C4D4D5\" #blue\r\n\r\n\r\n self.prop_txt1={\"color\":'#333', \"fontsize\":8}\r\n self.prop_txt2={\"bbox\": dict(boxstyle=\"round\", facecolor=bgr_col_bb2, ec=\"#CEEFF1\", pad=0.3, alpha=0.8), \"color\":\"#333\", \"fontsize\":9 }\r\n col_4=\"#001D4B\"\r\n self.prop_txt3={\"bbox\": dict(boxstyle=\"round\", facecolor=bgr_col_bb3, ec=\"#CAE2E3\", pad=0.3, alpha=0.84), \"color\":col_4, \"fontsize\":10, \"fontweight\":400, \"fontfamily\":\"Tahoma\"}\r\n\r\n self.prop_txt_name={\"bbox\": dict(boxstyle=\"round\", facecolor=bgr_col_bb1, ec=\"#CEEFF1\", pad=0.4, alpha=0.9), \"color\":\"#EEEEEE\", \"fontsize\":10, \"fontweight\":800, \"alpha\":1}\r\n self.prop_txt_time={\"bbox\": dict(boxstyle=\"round\", facecolor=bgr_col_bb3, ec=\"#CAE2E3\", pad=0.5, alpha=0.84), \"color\":COL_3, \"fontsize\":10, \"fontweight\":600, \"fontfamily\":\"Consolas\"}\r\n self.prop_txt_time2={\"bbox\": dict(boxstyle=\"round\", facecolor=bgr_col_bb3, ec=\"#CAE2E3\", pad=0.4, alpha=0.74), \"color\":\"#333\", \"fontsize\":8, \"fontweight\":600, \"fontfamily\":\"Consolas\"}\r\n\r\n self.id_text = self.ax.text2D(*self.txt1, \"\", **self.prop_txt1, transform=self.ax.transAxes)\r\n self.id_text_2 = self.ax.text2D(*self.txt2, \"\", **self.prop_txt2, transform=self.ax.transAxes)\r\n self.id_text_3 = self.ax.text2D(*self.txt3, \"\", **self.prop_txt3, transform=self.ax.transAxes)\r\n\r\n help_x=1.036; help_y=0.2065\r\n if size760==True:help_x=1.036+0.126\r\n elif size900==True:help_x=1.036+0.035\r\n self.id_help_ico = self.ax.text2D(help_x, help_y, \"?\", c = \"#F6F6F6\", size = 10, fontweight = 800, picker=5, bbox = dict(boxstyle = \"circle\", edgecolor = \"#990000\",facecolor = \"#800000\",pad = 0.2), ha = \"left\", va = \"center\",alpha = 0.7, transform = self.ax.transAxes)\r\n\r\n #=============\r\n\r\n\r\n alpha_main=0.7\r\n self.alpha_main=alpha_main\r\n self.sw_cir_alpha=False\r\n\r\n self.ψ2 = np.linspace(-np.pi/2, np.pi/2, 100)\r\n self.w = np.array([0, -1, 0])\r\n\r\n #========== Horizon\r\n color_hor=\"#001440\"\r\n color_hor=\"#333333\"\r\n self.ψ = np.linspace(0, 2 * np.pi, 100)\r\n φ=self.ψ\r\n r=1\r\n x1=r * np.cos(φ)\r\n y1=r * np.sin(φ)\r\n z1=np.zeros(np.size(x1))\r\n \r\n '''\r\n #solid\r\n id_,=self.ax.plot(x1,y1,z1,color=color_hor,linewidth=2 ,picker=2)\r\n self.hor_circl=id_\r\n #id_=id(id_)\r\n self.ids[id(id_)] =[[x1[3],y1[3],z1[0]],\"Horizon\"]\r\n '''\r\n\r\n φ_1 = np.linspace(-np.pi/2, np.pi/2, 50)\r\n x1=r * np.cos(φ_1); y1=r * np.sin(φ_1); z1=np.zeros(np.size(x1));\r\n id_1, = self.ax.plot(x1,y1,z1 ,color=color_hor,picker=main_circles_picker, linewidth=1, linestyle = \"solid\", alpha=alpha_main)\r\n\r\n self.hor_circl_1=id_1\r\n self.ids[id(id_1)] =[[x1[20],y1[20],z1[0]],\"Horizon\"]\r\n\r\n φ_2 = np.linspace(np.pi/2,np.pi*3/2, 50)\r\n x1=r * np.cos(φ_2); y1=r * np.sin(φ_2); z1=np.zeros(np.size(x1));\r\n id_2, = self.ax.plot(x1,y1,z1 ,color=color_hor,picker=main_circles_picker, linewidth=1, linestyle = \"dashed\", alpha=alpha_main) \r\n\r\n self.hor_circl_2=id_2\r\n self.ids[id(id_2)] =[self.ids[id(id_1)][0],\"Horizon\"]\r\n\r\n\r\n #======== Prime vertical\r\n \r\n color_vert=\"#050835\"\r\n color_vert=color_hor\r\n θ=self.ψ\r\n\r\n '''\r\n #solid \r\n z2= r * np.cos(θ) \r\n x2= r * np.sin(θ)\r\n y2=np.zeros(np.size(x2))\r\n id_,=self.ax.plot(x2,y2,z2,color=color_vertlinewidth=1,picker=2 ) #\"#0009BC\"\r\n self.ids[id(id_)] =[[x2[20],y2[20],z2[20]],\"Prime Vertical\"] #\"v_rot+π/2\"\r\n self.prime_vert=id_\r\n '''\r\n φ_1 = np.linspace(0, np.pi, 50)\r\n x2=r * np.sin(φ_1); z2=r * np.cos(φ_1); y2=np.zeros(np.size(x2));\r\n id_1, = self.ax.plot(x2,y2,z2 ,color=color_vert, linewidth=1, linestyle = \"solid\", alpha=alpha_main,picker=main_circles_picker)\r\n\r\n self.prime_vert_1=id_1\r\n self.ids[id(id_1)] =[[x2[13],y2[13],z2[13]],\"Prime Vertical\"]\r\n\r\n φ_2 = np.linspace(np.pi,2*np.pi, 50)\r\n x2=r * np.sin(φ_2); z2=r * np.cos(φ_2); y2=np.zeros(np.size(x1));\r\n id_2, = self.ax.plot(x2,y2,z2 ,color=color_vert, linewidth=1, linestyle = \"dashed\", alpha=alpha_main,picker=main_circles_picker) \r\n\r\n self.prime_vert_2=id_2\r\n self.ids[id(id_2)] =[self.ids[id(id_1)][0], \"Prime Vertical\"]\r\n\r\n X,Y,Z =self.draw_axes()\r\n\r\n id_=self.ax.scatter(1, 0, 0, marker=\"o\",c=\"#265678\", s=9, zorder=0); \"#3B6D91\"\r\n self.ids[id(id_)] =[[1,0,0],\" φ=0, θ=0\"]\r\n self.start_pt=id_\r\n\r\n self.plot_circle_scale(\"Horizon\");\r\n self.plot_circle_scale(\"Prime vertical\")\r\n\r\n\r\n #================\r\n\r\n self.date_utc = self.data[\"d_utc\"]\r\n self.time_utc = self.data[\"t_utc\"]\r\n self.timestamp = self.data[\"timestamp\"]\r\n self.geo_longitude=float(self.data[\"lon\"])\r\n\r\n name = \" {} {} \".format(self.data[\"n\"], self.data[\"ln\"])\r\n time_utc = \" UTC: {: >10}, {: >8} \".format(self.data[\"d_utc\"], self.data[\"t_utc\"])\r\n d_loc=\"{}-{}-{}\".format(self.data[\"d\"][2], self.data[\"d\"][1], self.data[\"d\"][0])\r\n t_loc=\"{: >2}:{:0>2}:{}\".format(self.data[\"t\"][0], self.data[\"t\"][1], self.data[\"t\"][2])\r\n time_loc = \"{: >10}, {: >8}\".format(d_loc, t_loc)\r\n time_loc= \"{: ^21}\".format(time_loc)\r\n name=\"\"\r\n self.id_text_name= self.ax.text2D(*self.txt_name, name, **self.prop_txt_name, transform=self.ax.transAxes)\r\n self.id_text_time= self.ax.text2D(*self.txt_time, time_loc, **self.prop_txt_time, transform=self.ax.transAxes)\r\n self.id_text_time2= self.ax.text2D(*self.txt_time2, time_utc, **self.prop_txt_time2, transform=self.ax.transAxes)\r\n\r\n #================\r\n\r\n self.plot_Equator_Ecliptic() \r\n self.draw_sphere()\r\n self.sphere_.set_visible(False) \r\n self.draw_surface()\r\n self.plot_meridian()\r\n \r\n self.planets_obj={}\r\n self.houses_obj={}\r\n self.zodiac_obj={}\r\n self.planets_ids={}\r\n self.houses_ids={}\r\n self.zodiac_ids={}\r\n self.plot_planets(self.planets_data)\r\n self.plot_zodiac()\r\n\r\n style_=(0, (1, 2))\r\n id_axes,=self.ax.plot(X,Y,Z ,color='#999999',linestyle = style_,label='Axes', alpha=alpha_main,picker=0)\r\n self.leg_items_ids.append(id_axes) ;self.axes_id=id_axes\r\n\r\n self.azim0=20\r\n self.elev0=15\r\n self.ax.view_init(elev = self.elev0, azim = self.azim0)\r\n\r\n \r\n col_=\"#900000\"\r\n self.annot = self.ax.annotate(\"click\", xy=(0,0), fontsize=9, c=col_, xytext=(-20,20),textcoords=\"offset points\",bbox=dict(boxstyle=\"round,pad=0.5\", fc=\"#FCFF90\", alpha = 0.7),arrowprops=dict(arrowstyle=\"->\"), zorder=20) #\"offset pixels\"\r\n self.annot.set_visible(False)\r\n\r\n self.click_tip_list=[\"North Pole\", \"South Pole\", \"Ecliptic North Pole\", \"Ecliptic South Pole\"]\r\n\r\n\r\n #========== Legend ========\r\n leg_colPM=\"#333333\"\r\n leg_colHor=color_hor\r\n leg_colPM=color_vert \r\n add_item0 = Line2D([0], [0], marker=\"o\", color=\"w\", label=\"Sphere\", markerfacecolor=\"#D0C7E8\", markersize=10,linewidth=10)\r\n add_item1 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Equator\", markerfacecolor=\"#700000\", markersize=10,linewidth=10)\r\n add_item2 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Ecliptic\", markerfacecolor=\"#A9A54B\", markersize=10)#\r\n add_item3 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Horizon\", markerfacecolor=leg_colHor, markersize=10)\r\n add_item4 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Prime Vertical\", markerfacecolor=leg_colPM, markersize=10)\r\n add_item5 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Meridian\",markerfacecolor=\"#333333\", markersize=10)#\r\n add_item6 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Proj. Horizons\",markerfacecolor=\"#2D305C\", markersize=10)\r\n add_item7 = Line2D([0], [0], marker=\"s\", color=\"w\", label=\"Parallels\",markerfacecolor=\"#333333\", markersize=10)\r\n add_item8 = Line2D([0], [0], marker=\"*\", color=\"w\", label=\"Extra off\",markerfacecolor='#543E17', markersize=10)#\r\n\r\n add_item9 = Line2D([0], [0], marker='o', color='#DCDCDC', label=\"Show half\",markerfacecolor='#333333', markersize=1,linestyle=\"dashed\")\r\n\r\n id_scale= Line2D([0], [0], marker='o', color='#DCDCDC', label=\"Scale\",markerfacecolor='#333333', markersize=1,linestyle=\"dashed\")\r\n #add_item24 = Line2D([0], [0], marker=\"o\", color=\"w\", label=\"Show grid\", markerfacecolor='#111111', markersize=10)\r\n\r\n self.leg_mer_circle=add_item5\r\n\r\n for id_ in [add_item0, add_item1, add_item2, add_item3, add_item4, add_item5,add_item8, add_item9,id_scale]:\r\n self.leg_items_ids.append(id_)\r\n\r\n leg_x=0.99;leg_y=0.99\r\n self.legend = self.fig.legend(handles=self.leg_items_ids,loc=\"upper right\",fancybox=True, shadow=True,borderpad=0.6,bbox_to_anchor=(leg_x, leg_y), prop=dict(size=10))\r\n\r\n for legline, line, text in zip(self.legend.get_lines(), self.leg_items_ids, self.legend.get_texts()):\r\n legline.set_picker(True)\r\n legline.set_pickradius(6)\r\n txt=text.get_text()\r\n self.leg_obj[legline] = (txt, line)\r\n\r\n\r\n #===============\r\n x001 = 0.92\r\n y001 = 0.58\r\n\r\n #Texts \"View:\"\r\n resize = 0.58\r\n f_size = 9\r\n f_col = \"#333\"\r\n b_col = \"#D3D3D3\"\r\n e_col = \"#999999\"\r\n b_col2 = \"#EEEEEE\"\r\n font_f = \"Courier New\" #\"Tahoma\"\r\n font_w = \"bold\"\r\n alpha_t = 0.9\r\n\r\n x00 = 0.916\r\n y00 = 0.58\r\n\r\n if size760==True:resize = 0.7; dd=0.93;x00=x00+0.1\r\n elif size900==True: resize = 0.625; dd=0.93;x00=x00+0.03\r\n\r\n x0 = x00; y0 = y00;dx = 0.0368; dx1 = 0.04 \r\n dx = resize * dx; x01= 0.009\r\n\r\n self.id_view=self.ax.text2D(x0, y0, \"View:\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w, bbox = dict(boxstyle = \"round\", edgecolor = b_col2, facecolor = b_col2,pad = 0.3), picker = 3, ha = \"left\", va = \"center\",alpha = 1, transform = self.ax.transAxes)\r\n self.id_E = self.ax.text2D(x0 + x01 + 2*dx, y0, \"E\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"round\", edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"center\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_W = self.ax.text2D(x0 + x01 + 3*dx, y0, \"W\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"round\", edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"center\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_N = self.ax.text2D(x0 + x01 + 4*dx, y0, \"N\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"round\", edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"center\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes) \r\n self.id_S = self.ax.text2D(x0 + x01 + 5*dx, y0, \"S\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"round\", edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"center\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n\r\n #Texts \"Azim:\"\r\n x0 = x00 \r\n dy00 = 0.04\r\n dy00 = resize * dy00\r\n y0 = y00 - dy00;\r\n dy = 0.035;\r\n dx = 0.0449 ; dx1 = 0.024\r\n dx = resize * dx\r\n dy = resize * dy\r\n\r\n self.id_azim=self.ax.text2D(x0, y0 , \"Azim:\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w, bbox = dict(boxstyle = \"round\", edgecolor = b_col2, facecolor = b_col2, pad = 0.3), picker = 3, ha = \"left\", va = \"center\",alpha = 1, transform = self.ax.transAxes) \r\n\r\n self.id_A0 = self.ax.text2D(x0 + dx1 + dx, y0 , \" 0\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\", alpha = alpha_t, transform = self.ax.transAxes) \r\n self.id_A90 = self.ax.text2D(x0 + dx1 + 2*dx, y0 , \"90\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_A180 = self.ax.text2D(x0 + dx1 + 3*dx, y0 , \"180\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n\r\n self.id_elev=self.ax.text2D(x0, y0 - dy, \"Elev:\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , bbox = dict(boxstyle = \"round\",edgecolor = b_col2, facecolor = b_col2,pad = 0.3), picker = 3, ha = \"left\", va = \"center\",alpha = 1, transform = self.ax.transAxes) \r\n\r\n self.id_E0 = self.ax.text2D(x0 + dx1 + dx, y0 - dy, \" 0\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes) \r\n self.id_E90 = self.ax.text2D(x0 + dx1 + 2*dx, y0 - dy, \"90\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_E180 = self.ax.text2D(x0 + dx1 + 3*dx, y0 - dy, \"180\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes) \r\n \r\n #Texts \"Init \"\r\n\r\n dy = 0.05; dx = 0.04\r\n x0 = x00\r\n dy00 = 0.07\r\n dy00 = resize * dy00\r\n y0 = y00 - dy00 \r\n dy = 0.056; dx = 0.077\r\n dx = resize * dx\r\n dy = resize * dy\r\n\r\n self.id_v_start = self.ax.text2D(x0, y0 - dy, \"Init \", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\", alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_v_chart = self.ax.text2D(x0 + dx, y0 - dy, \"Chart\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\", alpha = alpha_t, transform = self.ax.transAxes) \r\n self.id_v_Eq = self.ax.text2D(x0 + 2*dx, y0 - dy, \"Equat\", c = f_col, size = f_size, fontfamily = font_f, fontweight = font_w , picker = 5,bbox = dict(boxstyle = \"round\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\", alpha = alpha_t, transform = self.ax.transAxes) \r\n\r\n x0 = x00; \r\n dx00 = 0.082\r\n dx00 = resize * dx00\r\n x0 = x0 + dx00\r\n dy00 = 0.15\r\n dy00 = resize * dy00\r\n y0 = y00 - dy00\r\n dx1 = 0.02; dx2=0.024\r\n dy1 = 0.08; dy2 = 0.14; dy3 = 0.11;\r\n dy1 = resize * dy1; dy2 = resize * dy2; dy3 = resize * dy3\r\n\r\n dx2=0.028;dx2b=0.024\r\n f_size2=7\r\n self.id_v_up = self.ax.text2D(x0 + resize * dx1, y0 - dy1, \" \", c = f_col, size = f_size2, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"rarrow\",edgecolor = e_col,facecolor = b_col,pad = 0.3), rotation=90, ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_v_down = self.ax.text2D(x0 + resize * dx1, y0 - dy2, \" \", c = f_col, size = f_size2, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"larrow\", edgecolor = e_col,facecolor = b_col,pad = 0.3), rotation=90, ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes) \r\n self.id_v_left = self.ax.text2D(x0 + resize * dx1 - resize * dx2, y0 - dy3, \" \", c = f_col, size = f_size2, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"larrow\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n self.id_v_right = self.ax.text2D(x0 + resize * dx1 + resize * dx2b, y0 - dy3, \" \", c = f_col, size = f_size2, fontfamily = font_f, fontweight = font_w , picker = 5, bbox = dict(boxstyle = \"rarrow\",edgecolor = e_col,facecolor = b_col,pad = 0.3), ha = \"left\", va = \"center\",alpha = alpha_t, transform = self.ax.transAxes)\r\n\r\n\r\n self.id_v_anim=None\r\n self.id_v_prev=self.id_v_next=None\r\n self.id_v_test=None\r\n\r\n\r\n self.half=0\r\n self.view_East=[self.hor_circl_1, self.prime_vert_1, self.eq_circl_1, self.ecl_circl_1]\r\n self.view_West=[self.hor_circl_2, self.prime_vert_2, self.eq_circl_2, self.ecl_circl_2]\r\n\r\n self.circles_alpha()\r\n\r\n self.ax.set_axis_off() \r\n self.ax.axison=False\r\n\r\n self.ax.set_xlabel('x - axis'); self.ax.set_ylabel('y - axis'); self.ax.set_zlabel('z - axis')\r\n arr=[-1,0,1]\r\n self.ax.set_xticks(arr); self.ax.set_yticks(arr); self.ax.set_zticks(arr);\r\n\r\n self.xlim = self.ax.get_xlim()\r\n self.ylim = self.ax.get_ylim()\r\n self.zlim = self.ax.get_zlim()\r\n\r\n\r\n #============= Tkinter =============\r\n self.canvas = FigureCanvasTkAgg(self.fig, master = frame)\r\n self.canvas.draw() \r\n self.canvas.get_tk_widget().pack(side=TOP, anchor=NW, fill=BOTH, expand=True)\r\n\r\n self.canvas.mpl_connect('button_press_event', self.on_click)\r\n self.canvas.mpl_connect('pick_event', self.onpick) \r\n self.canvas.mpl_connect('key_release_event', self.key_)\r\n #self.canvas.mpl_connect('scroll_event', self.scroll_zoom)\r\n \r\n #=========== window size \r\n\r\n mins_x=1000;mins_y=700\r\n self.parent.minsize(mins_x,mins_y)\r\n w,h = self.fig.get_size_inches()*self.fig.dpi\r\n h = h + 100 #figure plus parent window\r\n w = w + 50\r\n window=self.parent\r\n screenHeight = window.winfo_screenheight()\r\n screenWidth = window.winfo_screenwidth()\r\n x = int(screenWidth/2 - w/2)\r\n y = int(screenHeight/2 - h/2)\r\n window.wm_geometry(\"+{}+{}\".format(x, y))\r\n window.resizable(False, False) \r\n\r\n def draw_axes(self): \r\n z4 = np.linspace(-1.01,1.01,50)\r\n y4 = np.zeros(np.size(z4))\r\n x4 = np.zeros(np.size(y4))\r\n\r\n x5 = np.linspace(-1.01,1.01,50)\r\n y5 = np.zeros(np.size(z4))\r\n z5 = np.zeros(np.size(y4))\r\n\r\n y6 = np.linspace(-1.01,1.01,50)\r\n x6 = np.zeros(np.size(z4))\r\n z6 = np.zeros(np.size(y4))\r\n\r\n #========joined data for legend - one item only to swicth off\r\n X=np.append(x4,np.NaN);X=np.append(X,x5) ;X=np.append(X,np.NaN); X=np.append(X,x6)\r\n Y=np.append(y4,np.NaN);Y=np.append(Y,y5) ;Y=np.append(Y,np.NaN); Y=np.append(Y,y6)\r\n Z=np.append(z4,np.NaN);Z=np.append(Z,z5) ;Z=np.append(Z,np.NaN); Z=np.append(Z,z6)\r\n\r\n return [X,Y,Z] \r\n\r\n def circles_alpha(self):\r\n if self.sw_cir_alpha==False:\r\n self.prime_vert_1.set_alpha(0.4)\r\n self.prime_vert_2.set_alpha(0.2)\r\n self.hor_circl_2.set_alpha(0.2) \r\n self.mer_circle.set_alpha(0.2)\r\n self.axes_id.set_alpha(0.2)\r\n\r\n for cir in [self.eq_circl_2, self.ecl_circl_2]:\r\n cir.set_alpha(0.2) \r\n\r\n for el in self.ecliptic_scale:\r\n self.ecliptic_scale[el][0].set_alpha(0.5) \r\n self.ecliptic_scale[el][1].set_alpha(0.5) \r\n for el in self.equator_scale:\r\n self.equator_scale[el][0].set_alpha(0.5) \r\n self.equator_scale[el][1].set_alpha(0.5)\r\n\r\n self.sw_cir_alpha=True\r\n\r\n elif self.sw_cir_alpha==True:\r\n self.prime_vert_1.set_alpha(self.alpha_main)\r\n self.prime_vert_2.set_alpha(self.alpha_main)\r\n self.hor_circl_2.set_alpha(self.alpha_main) \r\n self.mer_circle.set_alpha(self.alpha_main)\r\n self.axes_id.set_alpha(self.alpha_main)\r\n\r\n for cir in [self.eq_circl_1, self.ecl_circl_1]:\r\n cir.set_alpha(self.alpha_main_2)\r\n for cir in [self.eq_circl_2, self.ecl_circl_2]:\r\n cir.set_alpha(self.alpha_main_2) \r\n\r\n for el in self.ecliptic_scale:\r\n self.ecliptic_scale[el][0].set_alpha(self.alpha_main) \r\n self.ecliptic_scale[el][1].set_alpha(self.alpha_main) \r\n for el in self.equator_scale:\r\n self.equator_scale[el][0].set_alpha(self.alpha_main) \r\n self.equator_scale[el][1].set_alpha(self.alpha_main) \r\n\r\n self.sw_cir_alpha=False\r\n\r\n\r\n def plot_Equator_Ecliptic(self):\r\n ARMC=self.planets_data[\"MC\"][\"eq\"][0] \r\n ARMC=float(ARMC)\r\n ARMC=norm_(ARMC)\r\n self.ARMC=ARMC\r\n φ = self.ψ\r\n self.k = k = np.array([1, 0, 0])\r\n self.v = v = np.array([0, 1, 0])\r\n\r\n geo_latitude=self.geo_latitude\r\n s1 = \"S\" if self.geo_latitude<0 else \"N\"\r\n s2 = \"W\" if self.geo_longitude<0 else \"E\"\r\n txt=\"ARMC={:.0f}°, lat={:.0f}°{}, lon={:.0f}°{}\".format(ARMC,self.geo_latitude,s1,self.geo_longitude,s2)\r\n self.id_text.set_text(txt)\r\n self.south = True if self.geo_latitude<0 else False\r\n\r\n alpha_main_2=0.8\r\n self.alpha_main_2=alpha_main_2\r\n \r\n #===== Celestial Equator\r\n \r\n self.geo_latitude = geo_latitude\r\n self.geo_lat = np.radians(self.geo_latitude)\r\n geo_lat = self.geo_lat\r\n\r\n rotation = np.pi/2 + geo_lat\r\n self.rotation=rotation\r\n self.eq_rot_ang=rotation\r\n\r\n v_rot = self.Rodrigues_rotation(v,k,rotation)\r\n self.equator_rot=v_rot\r\n\r\n x,y,z=self.equator_rot\r\n x2= np.sin(φ) + x * np.cos(φ)\r\n y2= y * np.cos(φ)\r\n z2= z * np.cos(φ)\r\n '''\r\n #==== solid\r\n id_,=self.ax.plot(x2,y2,z2 ,color=\"#700000\", linestyle = \"solid\" ,picker=2) #'dashed'\r\n #id_=id(id_)\r\n self.ids[id(id_)] =[[x2[18],y2[18],z2[18]],\"Celestial Equator\"] #\"v_rot+π/2\"\r\n self.eq_circl=id_\r\n '''\r\n color_equat=\"#700000\"\r\n id_1=id_2=0 \r\n\r\n φ_1 = np.linspace(0, np.pi, data_pts)\r\n x2, y2, z2 = self.circl_vect(self.equator_rot,k,φ_1)\r\n eq_1 = [x2[0],y2[0],z2[0]]\r\n self.eq_circl_1, = self.ax.plot(x2, y2, z2, color=color_equat, picker=main_circles_picker, linewidth=1, linestyle = \"solid\", alpha=self.alpha_main_2)\r\n\r\n φ_2 = np.linspace(np.pi,2*np.pi, data_pts)\r\n x2b, y2b, z2b = self.circl_vect(self.equator_rot,k,φ_2)\r\n eq_2 = [x2b[0],y2b[0],z2b[0]]\r\n self.eq_circl_2, = self.ax.plot(x2b, y2b, z2b, color=color_equat, picker=main_circles_picker, linewidth=1, linestyle = \"dashed\", alpha=self.alpha_main_2)\r\n\r\n\r\n self.ids[id(self.eq_circl_1)] =[[x2[13],y2[13],z2[13]],\"Celestial Equator\"]\r\n self.ids[id(self.eq_circl_2)] =[self.ids[id(self.eq_circl_1)][0],\"Celestial Equator\"] \r\n\r\n self.equat_ids.append(self.eq_circl_1); self.equat_ids.append(self.eq_circl_2)\r\n\r\n #====== ARMC\r\n ARMC_= eq_1\r\n ARMC_= np.multiply(1.05,ARMC_)\r\n self.armc_id, = self.ax.plot(*ARMC_,c=\"#555555\",marker=\"$ARMC$\",markersize=12,picker=4) #point\r\n self.equat_ids.append(self.armc_id)\r\n self.ids[id(self.armc_id)] =[eq_1,\"ARMC\"]\r\n self.armc_id.set_visible(False)\r\n ARMC=np.radians(ARMC)\r\n self.v_equinox_ang=ARMC\r\n\r\n #------- 2nd point on rotated circle oblique\r\n self.v_equinox_ang=ARMC\r\n φ2=-ARMC\r\n x12, y12, z12 = self.circl_vect(self.equator_rot,k,φ2)\r\n self.v_equinox=[x12,y12,z12] # vernal equinox\r\n self.a_equinox=[-x12,-y12,-z12]\r\n self.equinox_id, = self.ax.plot(*self.v_equinox, c=\"#555\", marker=\"o\", markersize=2, zorder=3, alpha=1);\r\n self.ecl_ids.append(self.equinox_id)\r\n\r\n #====== poleN\r\n rot=self.rotation\r\n v_rot=self.equator_rot\r\n color_4=\"#444444\"\r\n rotation = self.eq_rot_ang + np.pi/2\r\n\r\n v_rot1 = self.Rodrigues_rotation(v,k,rotation)\r\n self.poleS=v_rot1\r\n v_rot2=-v_rot1\r\n self.poleN=v_rot2\r\n\r\n self.poleN_id, =self.ax.plot(*self.poleN,c=color_4,marker=\"$N$\",picker=5, zorder=2) #point\r\n self.ids[id(self.poleN_id)] =[self.poleN,\"North Pole\"] #\"v_rot-π/2\"\r\n self.equat_ids.append(self.poleN_id)\r\n\r\n self.poleS_id, =self.ax.plot(*self.poleS,c=color_4, marker=\"$S$\",picker=5, zorder=2) #point\r\n self.ids[id(self.poleS_id)] =[self.poleS,\"South Pole\"] #\"v_rot+π/2\"\r\n self.equat_ids.append(self.poleS_id)\r\n\r\n poleN_θ_hor=np.arccos(self.poleN[2])\r\n self.poleN_elev=np.degrees(poleN_θ_hor) \r\n \r\n\r\n #====== Ecliptic\r\n φ1 = -ARMC - np.pi/2\r\n x13, y13, z13 = self.circl_vect(self.equator_rot,k,φ1)\r\n self.v_rot4=[x13,y13,z13]\r\n\r\n #ε=23.43 \r\n k1=np.array(self.v_equinox) \r\n v1=np.array(self.v_rot4)\r\n ε=np.radians(self.ε)\r\n rotation = ε\r\n v_rot3 = self.Rodrigues_rotation(v1,k1,rotation)\r\n self.ecliptic_rot=v_rot3\r\n v2= -self.ecliptic_rot\r\n\r\n φ4=np.arctan(-k1[2]/v2[2])\r\n x3t= v2[0] * np.sin(φ4) + k1[0] * np.cos(φ4)\r\n if x3t<0:\r\n φ4 = φ4 + np.pi\r\n\r\n φ4=norm_r(φ4)\r\n self.asc_lon=φ4\r\n\r\n φ3 = np.arctan(-v2[0]/k1[0]) if k1[0] != 0 else np.pi/2\r\n\r\n if norm_r(ARMC)< np.pi:\r\n φ3=np.pi/2 - φ3\r\n ecl_st=0 + φ3 #ecl_start\r\n else:\r\n φ3= - np.pi/2 - φ3\r\n ecl_st=0 + φ3\r\n \r\n self.ecl_start=np.degrees(φ3)\r\n φ_1 = np.linspace(ecl_st , ecl_st + np.pi, data_pts)\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ_1)\r\n self.ecl_circl_1, = self.ax.plot(x3,y3,z3 ,color=color_ecl, picker=main_circles_picker, linestyle = \"solid\", alpha=alpha_main_2)\r\n self.ids[id(self.ecl_circl_1)] =[[x3[15],y3[15],z3[15]],\"Ecliptic\"]\r\n \r\n\r\n φ_2 = np.linspace(ecl_st + np.pi, ecl_st + 2*np.pi, data_pts)\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ_2)\r\n self.ecl_circl_2, = self.ax.plot(x3,y3,z3 ,color=color_ecl, picker=main_circles_picker, linestyle = \"dashed\", alpha=alpha_main_2)\r\n self.ids[id(self.ecl_circl_2)] =[self.ids[id(self.ecl_circl_1)][0],\"Ecliptic\"]\r\n \r\n r1=1.07\r\n x3p, y3p, z3p = self.circl_vect(k1, v2, φ3, r1)\r\n self.v_MC=self.circl_vect(k1, v2, φ3)\r\n\r\n self.mc_id,= self.ax.plot(x3p,y3p,z3p,c=\"#444444\",marker=\"$MC$\",markersize=10,picker=4, zorder=2); #point\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ3)\r\n lon_=f'Medium Coeli: lon={round(self.planets_data[\"MC\"][\"ecl\"][0],2)}°'\r\n self.ids[id(self.mc_id)] =[[x3, y3, z3],lon_]\r\n self.ecl_ids.append(self.mc_id)\r\n self.mc_id.set_visible(False)\r\n\r\n\r\n #====== Asc ========\r\n x13, y13, z13 = self.circl_vect(k1, v2, φ4)\r\n self.asc_id2, =self.ax.plot(x13,y13,z13, c=\"#333\", marker=\"o\", markersize=3, zorder=2, alpha=0.6);#point\r\n self.ecl_ids.append(self.asc_id2)\r\n asc=[x13,y13,z13]\r\n self.v_asc=asc\r\n\r\n r1=1.07\r\n x3p, y3p, z3p = self.circl_vect(k1, v2, φ4, r1)\r\n asc_t=[x3p,y3p,z3p-0.03]\r\n self.asc_id,=self.ax.plot(*asc_t, c=\"#555555\",marker=\"$Asc$\",markersize=10,picker=5, zorder=2) ;#point\r\n\r\n self.asc_id2.set_visible(False)\r\n self.asc_id.set_visible(False)\r\n self.ecl_ids.append(self.asc_id)\r\n\r\n\r\n self.Asc_φ_hor=np.arccos(self.v_asc[0])\r\n if self.v_asc[1]<0: self.Asc_φ_hor = -abs(self.Asc_φ_hor)\r\n self.Asc_φ_hor = norm_r(self.Asc_φ_hor)\r\n\r\n\r\n '''\r\n #=== solid\r\n x3= k1[0] * np.sin(φ) + self.ecliptic_rot[0] * np.cos(φ)\r\n y3= k1[1] * np.sin(φ) + self.ecliptic_rot[1] * np.cos(φ)\r\n z3= k1[2] * np.sin(φ) + self.ecliptic_rot[2] * np.cos(φ)\r\n id_,=self.ax.plot(x3,y3,z3 ,color = color_ecl, linestyle = \"solid\" ,picker=2) #'dashed'\r\n self.ids[id(id_)] =[[x2[18],y2[18],z2[18]],\"Ecliptic\"] \r\n self.ecl_circle=id_\r\n '''\r\n\r\n #============= Ecliptic pole \r\n\r\n rotation = -np.pi/2\r\n k1=np.array(self.v_equinox) \r\n v1=np.array(self.ecliptic_rot)\r\n\r\n v_rot6 = self.Rodrigues_rotation(v1,k1,rotation)\r\n self.poleN_ecl = v_rot6\r\n self.poleN_ecl_id, =self.ax.plot(*self.poleN_ecl,c=color_ecl, marker=\"$n$\",markersize=5,picker=5, zorder=2, alpha=0.8)\r\n self.poleS_ecl=-self.poleN_ecl; \r\n self.poleS_ecl_id, =self.ax.plot(*self.poleS_ecl,c=color_ecl, marker=\"$s$\",markersize=5,picker=5, zorder=2, alpha=0.8)\r\n\r\n self.ids[id(self.poleN_ecl_id)] =[self.poleN_ecl,\"Ecliptic North Pole\"]\r\n self.ecl_ids.append(self.poleN_ecl_id)\r\n self.ids[id(self.poleS_ecl_id)] =[self.poleS_ecl,\"Ecliptic South Pole\"]\r\n self.ecl_ids.append(self.poleS_ecl_id)\r\n\r\n\r\n poleN_ecl_θ_hor=np.arccos(self.poleN_ecl[2])\r\n pole_ecl_elev= np.pi/2 - poleN_ecl_θ_hor\r\n self.pole_ecl_elev=np.degrees(pole_ecl_elev)\r\n\r\n self.plot_ecliptic_scale()\r\n self.plot_equator_scale()\r\n\r\n\r\n def plot_Equator_Ecliptic_next(self):\r\n ARMC=self.planets_data[\"MC\"][\"eq\"][0]\r\n ARMC=float(ARMC)\r\n ARMC=norm_(ARMC)\r\n self.ARMC=ARMC\r\n \r\n φ = self.ψ\r\n self.k = k = np.array([1, 0, 0])\r\n self.v = v = np.array([0, 1, 0])\r\n geo_latitude=self.geo_latitude\r\n\r\n s1 = \"S\" if self.geo_latitude<0 else \"N\"\r\n s2 = \"W\" if self.geo_longitude<0 else \"E\"\r\n txt=\"ARMC={:.0f}°, lat={:.0f}°{}, lon={:.0f}°{}\".format(ARMC,self.geo_latitude,s1,self.geo_longitude,s2)\r\n self.id_text.set_text(txt)\r\n self.south = True if self.geo_latitude<0 else False\r\n \r\n #===== Celestial Equator\r\n \r\n self.geo_latitude = geo_latitude\r\n self.geo_lat = np.radians(self.geo_latitude)\r\n geo_lat = self.geo_lat\r\n\r\n rotation = np.pi/2 + geo_lat\r\n self.rotation=rotation\r\n self.eq_rot_ang=rotation\r\n\r\n v_rot = self.Rodrigues_rotation(v,k,rotation)\r\n self.equator_rot=v_rot\r\n\r\n x,y,z=self.equator_rot\r\n x2= np.sin(φ) + x * np.cos(φ)\r\n y2= y * np.cos(φ)\r\n z2= z * np.cos(φ)\r\n\r\n color_equat=\"#700000\"\r\n id_1=id_2=0 \r\n\r\n φ_1 = np.linspace(0, np.pi, 50)\r\n x2, y2, z2 = self.circl_vect(self.equator_rot,k,φ_1)\r\n eq_1 = [x2[0],y2[0],z2[0]]\r\n self.eq_circl_1.set_data_3d(x2, y2, z2)\r\n\r\n φ_2 = np.linspace(np.pi,2*np.pi, 50)\r\n x2b, y2b, z2b = self.circl_vect(self.equator_rot,k,φ_2)\r\n eq_2 = [x2b[0],y2b[0],z2b[0]]\r\n \r\n self.eq_circl_2.set_data_3d(x2b, y2b, z2b)\r\n\r\n #====== ARMC\r\n ARMC_= eq_1\r\n ARMC_= np.multiply(1.05,ARMC_)\r\n self.armc_id.set_data_3d(*ARMC_)\r\n ARMC=np.radians(ARMC)\r\n self.v_equinox_ang=ARMC\r\n\r\n self.v_equinox_ang=ARMC\r\n φ2=-ARMC\r\n x12, y12, z12 = self.circl_vect(self.equator_rot,k,φ2)\r\n self.v_equinox=[x12,y12,z12]\r\n self.a_equinox=[-x12,-y12,-z12]\r\n self.equinox_id.set_data_3d(x12, y12, z12)\r\n\r\n #====== poleN\r\n rot=self.rotation\r\n v_rot=self.equator_rot\r\n color_4=\"#444444\"\r\n\r\n rotation = self.eq_rot_ang + np.pi/2\r\n v_rot1 = self.Rodrigues_rotation(v,k,rotation)\r\n self.poleS=v_rot1\r\n v_rot2=-v_rot1\r\n self.poleN=v_rot2\r\n\r\n self.poleN_id.set_data_3d(*self.poleN)\r\n self.ids[id(self.poleN_id)] =[self.poleN,\"North Pole\"]\r\n\r\n self.poleS_id.set_data_3d(*self.poleS)\r\n self.ids[id(self.poleS_id)] =[self.poleS,\"South Pole\"]\r\n poleN_θ_hor=np.arccos(self.poleN[2])\r\n self.poleN_elev=np.degrees(poleN_θ_hor) \r\n\r\n #====== Ecliptic \r\n\r\n φ1 = -ARMC - np.pi/2\r\n x13, y13, z13 = self.circl_vect(self.equator_rot,k,φ1)\r\n self.v_rot4=[x13,y13,z13]\r\n \r\n k1=np.array(self.v_equinox) \r\n v1=np.array(self.v_rot4)\r\n ε=np.radians(self.ε)\r\n rotation = ε\r\n v_rot3 = self.Rodrigues_rotation(v1,k1,rotation)\r\n self.ecliptic_rot=v_rot3\r\n v2= -self.ecliptic_rot\r\n\r\n φ4=np.arctan(-k1[2]/v2[2])\r\n x3t= v2[0] * np.sin(φ4) + k1[0] * np.cos(φ4)\r\n if x3t<0:\r\n φ4 = φ4 + np.pi\r\n\r\n φ4=norm_r(φ4)\r\n self.asc_lon=φ4\r\n\r\n φ3 = np.arctan(-v2[0]/k1[0]) if k1[0] != 0 else np.pi/2\r\n\r\n if norm_r(ARMC)< np.pi:\r\n φ3=np.pi/2 - φ3\r\n ecl_st=0 + φ3 #ecl_start\r\n else:\r\n φ3= - np.pi/2 - φ3\r\n ecl_st=0 + φ3\r\n\r\n self.ecl_start=np.degrees(φ3)\r\n φ_1 = np.linspace(ecl_st , ecl_st + np.pi, 50)\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ_1)\r\n self.ecl_circl_1.set_data_3d(x3, y3, z3)\r\n φ_2 = np.linspace(ecl_st + np.pi, ecl_st + 2*np.pi, 50)\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ_2)\r\n self.ecl_circl_2.set_data_3d(x3, y3, z3)\r\n\r\n\r\n #====== MC ========\r\n r1=1.07\r\n\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ3, r1)\r\n self.v_MC=self.circl_vect(k1, v2, φ3)\r\n self.mc_id.set_data_3d(x3, y3, z3)\r\n\r\n x3, y3, z3 = self.circl_vect(k1, v2, φ3)\r\n lon_=f'Medium Coeli: lon={round(self.planets_data[\"MC\"][\"ecl\"][0],2)}°'\r\n self.ids[id(self.mc_id)] =[[x3, y3, z3],lon_]\r\n\r\n #====== Asc ========\r\n r1=1.07\r\n\r\n x3p, y3p, z3p = self.circl_vect(k1, v2, φ4, r1)\r\n asc_t=[x3p,y3p,z3p-0.03]\r\n self.asc_id.set_data_3d(*asc_t)\r\n\r\n\r\n x13, y13, z13 = self.circl_vect(k1, v2, φ4)\r\n asc=[x13,y13,z13]\r\n self.v_asc=asc\r\n self.asc_id2.set_data_3d(x13, y13, z13)\r\n\r\n self.Asc_φ_hor=np.arccos(self.v_asc[0])\r\n if self.v_asc[1]<0: self.Asc_φ_hor = -abs(self.Asc_φ_hor)\r\n self.Asc_φ_hor = norm_r(self.Asc_φ_hor)\r\n \r\n #============= Ecliptic pole \r\n rotation = -np.pi/2\r\n k1=np.array(self.v_equinox) \r\n v1=np.array(self.ecliptic_rot)\r\n v_rot6 = self.Rodrigues_rotation(v1,k1,rotation)\r\n self.poleN_ecl = v_rot6\r\n self.poleS_ecl=-self.poleN_ecl; \r\n self.poleN_ecl_id.set_data_3d(*self.poleN_ecl)\r\n self.poleS_ecl_id.set_data_3d(*self.poleS_ecl)\r\n\r\n poleN_ecl_θ_hor=np.arccos(self.poleN_ecl[2])\r\n pole_ecl_elev= np.pi/2 - poleN_ecl_θ_hor\r\n self.pole_ecl_elev=np.degrees(pole_ecl_elev)\r\n\r\n self.ids[id(self.poleN_ecl_id)] =[self.poleN_ecl,\"Ecliptic North Pole\"]\r\n self.ids[id(self.poleS_ecl_id)] =[self.poleS_ecl,\"Ecliptic South Pole\"]\r\n \r\n self.plot_ecliptic_scale()\r\n self.plot_equator_scale()\r\n\r\n\r\n def plot_meridian(self): \r\n color_=\"#666666\" \r\n alpha_main=self.alpha_main\r\n θ=self.ψ\r\n y2=np.sin(θ)\r\n z2=np.cos(θ)\r\n x2=np.zeros(np.size(y2))\r\n id_,=self.ax.plot(x2,y2,z2 ,color=color_, linestyle = (0, (1, 2,1,3)), alpha=alpha_main, picker=main_circles_picker, zorder=0)\r\n self.ids[id(id_)] =[[x2[6],y2[6],z2[6]],\"Meridian\"]\r\n self.mer_circle=id_\r\n\r\n def plot_next_prev(self, seconds=0, minutes=0, hours=0):\r\n self.clear_ann()\r\n self.id_text_2.set_text(\"\")\r\n tz_=self.data[\"tz\"]\r\n t_obj=add_days(self.timestamp, seconds=seconds, minutes=minutes, hours=hours,tz_=tz_)\r\n date_t= t_obj[\"date_utc\"]; time_t= t_obj[\"time_utc\"]\r\n date_t_loc= t_obj[\"date_loc\"]; time_t_loc= t_obj[\"time_loc\"] \r\n self.timestamp=t_obj[\"timestamp\"]\r\n\r\n self.date_utc = date_t\r\n self.time_utc = time_t\r\n\r\n latitude = float(self.data[\"lat\"])\r\n longitude = float(self.data[\"lon\"])\r\n self.geo_latitude=latitude\r\n self.planets_data = calc_.get_planets_data(self.date_utc, self.time_utc, latitude, longitude)\r\n\r\n self.plot_Equator_Ecliptic_next()\r\n self.plot_zodiac_next()\r\n self.plot_planets_next(self.planets_data)\r\n\r\n time_utc = \" UTC: {}, {} \".format(self.date_utc, self.time_utc)\r\n time_loc = \"{: >10}, {: >8}\".format(date_t_loc, time_t_loc)\r\n time_loc= \"{: ^21}\".format(time_loc)\r\n\r\n self.id_text_time.set_text(time_loc)\r\n self.id_text_time2.set_text(time_utc)\r\n\r\n self.legend_updt()\r\n\r\n\r\n def plot_planets(self, planets_data):\r\n for i, pl in enumerate(planets_data):\r\n if pl not in c.planets.keys():continue\r\n pl_RA=planets_data[pl]['eq'][0]\r\n pl_decl=planets_data[pl]['eq'][1]\r\n lon=planets_data[pl]['ecl'][0]\r\n lat=planets_data[pl]['ecl'][1]\r\n sym = \"$\" + c.planets[pl] + \"$\"\r\n txt=\"{}: RA={:.2f}°, decl={:.2f}°, lon={:.1f}°, lat={:.1f}°\".format(pl, pl_RA, pl_decl, lon, lat)\r\n self.plot_planet_equat(pl_RA ,pl_decl, lon, lat ,rad=False,meridian=False, name=pl,sym=sym, text=txt)\r\n\r\n def plot_planet_equat(self, RA, decl, lon, lat, rad=False, meridian=False, parallel=False, name=\"\",sym =\"\",text=\"\"):\r\n pl_RA=np.radians(RA) if rad==False else RA\r\n pl_decl=np.radians(decl) if rad==False else decl\r\n rotation=pl_RA\r\n φ=self.ψ\r\n θ=self.ψ\r\n k=self.k\r\n k2 = self.poleN;\r\n v2 = self.equator_rot; \r\n \r\n color_pl2=pl_colors[name]\r\n color_pl1=pl_colors[name]\r\n color_dot=\"#1C0367\"\r\n color_dot=color_pl1\r\n\r\n rotation=rotation - self.v_equinox_ang\r\n v_rot5 = self.Rodrigues_rotation(v2,k2,rotation)\r\n\r\n v_mer = self.circl_vect(v_rot5,k2,φ)\r\n style_=(0, (1,4,1,6)) \r\n color_oran=\"#F38600\"\r\n color_yel=\"#FFCD00\"\r\n color_mer_par=color_pl1\r\n style_= (0, (1,3,)) if name in [\"Sun\",\"Moon\",\"Jupiter\"] else style_\r\n alpha_= 0.8 if name in [\"Sun\",\"Moon\",\"Jupiter\"] else 0.6 #fix light color contrast\r\n id_Mer, = self.ax.plot(*v_mer ,color=color_mer_par, linestyle=style_, linewidth=1,alpha=alpha_, picker=2)\r\n if meridian==False:\r\n id_Mer.set_visible(False)\r\n\r\n id_Mer_nat, = self.ax.plot(*v_mer ,color=color_mer_par, linestyle=(0, (2,4,3,4)), linewidth=1,alpha=alpha_, picker=2)\r\n id_Mer_nat.set_visible(False)\r\n \r\n shift=np.sin(pl_decl)\r\n data=self.plot_parallel(shift,color_=color_mer_par, style_=style_, alpha=alpha_) \r\n id_Par=data[0]\r\n if parallel==False:\r\n id_Par.set_visible(False)\r\n\r\n φ2 = pl_RA - self.v_equinox_ang\r\n φ2=norm_r(φ2)\r\n pl_φ_eq=φ2\r\n v_RA = self.circl_vect(self.equator_rot,k,φ2)\r\n\r\n id_pt_Eq, = self.ax.plot(*v_RA, c=color_dot, marker=\"o\", markersize=2, picker=4, zorder=3)\r\n id_pt_Eq.set_visible(False)\r\n\r\n pl_declination=pl_decl\r\n θ2=pl_decl\r\n v_pl = self.circl_vect(v_RA,k2,θ2)\r\n pl_θ_hor=np.arccos(v_pl[2])\r\n\r\n #-- planet dot \r\n if \"Node\" in name:\r\n id_pl, =self.ax.plot(*v_pl, c=color_pl1, marker=\"o\", picker=5, markersize=2,zorder=6) #onpick\r\n else:\r\n id_pl, =self.ax.plot(*v_pl, c=color_pl1, marker=\"o\", picker=5, markersize=4,zorder=6) #onpick\r\n\r\n arr=[-0.05,0.05]\r\n shift=0.07\r\n v_pl_t=np.add(v_pl,shift)\r\n\r\n if \"Node\" in name:\r\n id_t =self.ax.text(*v_pl_t,s=sym,fontsize=11,c=color_pl2, label=name, fontweight=800, picker=5,bbox=dict(boxstyle=\"round\",edgecolor=(\"#FFF\"),facecolor=(\"#FFF\"),pad=0.0, alpha=0.0), alpha=1.0, zorder=2)\r\n elif name in [\"Sun\",\"Moon\"]:\r\n id_t =self.ax.text(*v_pl_t,s=sym,fontsize=17,c=color_pl2, label=name, fontweight=800, picker=5,bbox=dict(boxstyle=\"round\",edgecolor=(\"#FFF\"),facecolor=(\"#FFF\"),pad=0.0, alpha=0.0), alpha=1.0, zorder=4)\r\n else:\r\n id_t =self.ax.text(*v_pl_t,s=sym,fontsize=16,c=color_pl2, label=name, fontweight=800, picker=5,bbox=dict(boxstyle=\"round\",edgecolor=(\"#FFF\"),facecolor=(\"#FFF\"),pad=0.0, alpha=0.0), alpha=1.0, zorder=4)\r\n\r\n v_hor=[v_pl[0],v_pl[1],0]\r\n v_hor=v_hor/np.linalg.norm(v_hor) \r\n id_pt_Hor, = self.ax.plot(*v_hor,c=color_dot,marker=\"o\", markersize=2, picker=4, zorder=7) \r\n id_pt_Hor.set_visible(False)\r\n\r\n pl_φ_hor = np.arccos(v_hor[0]); \r\n if v_hor[1] < 0:\r\n pl_φ_hor = 2*np.pi - pl_φ_hor\r\n\r\n azimuth = 2*np.pi - pl_φ_hor + np.pi/2\r\n azimuth = norm_r(azimuth)\r\n\r\n pl_θ_hor= np.arccos(v_pl[2])\r\n pl_alt_hor= np.pi/2 - pl_θ_hor\r\n #pl_alt_hor = rd(pl_alt_hor)\r\n\r\n #------ planet's proportionate horizon\r\n #tan(a)/tan(A)=sin(b) # Napier\r\n #np.tan(np.pi/2 - θ_pl)/np.tan(α)=np.sin(np.pi/2 - φ_pl) -> α= np.arctan(np.tan(np.pi/2 - θ_pl)/np.sin(np.pi/2 - φ_pl))\r\n #np.tan(np.pi/2 - θ_proj_hor)/np.tan(α)=np.sin(np.pi/2)\r\n #np.tan(np.pi/2 - θ_proj_hor)=np.tan(α) -> np.pi/2 - θ_proj_hor =α -> θ_proj_hor =np.pi/2 - α\r\n\r\n α = np.arctan(np.tan(np.pi/2 - pl_θ_hor)/np.sin(np.pi/2 - pl_φ_hor)) # α = proportionate horizon's inclination\r\n k=self.k\r\n w=self.w\r\n if v_pl[0]<0: α = α + np.pi\r\n v_rot8 = self.Rodrigues_rotation(k, w, α) \r\n φ2 = self.ψ2\r\n x2,y2,z2 = self.circl_vect(v_rot8, w, φ2)\r\n id_proj_hor,=self.ax.plot(x2,y2,z2 ,color=color_mer_par, linestyle = style_, linewidth=1, alpha=alpha_, picker=1, zorder=1)\r\n id_proj_hor.set_visible(False)\r\n\r\n id_proj_hor_nat,=self.ax.plot(x2,y2,z2 ,color=color_mer_par, linestyle = (0, (2,4,3,4)), linewidth=1, alpha=alpha_, picker=1, zorder=1)\r\n id_proj_hor_nat.set_visible(False)\r\n \r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot \r\n φ2=np.radians(lon) \r\n v_ecl = self.circl_vect(k2,v2,φ2)\r\n id_pt_Ecl, = self.ax.plot(*v_ecl,c=color_dot,marker=\"o\", markersize=2, picker=4, zorder=5);\r\n id_pt_Ecl.set_visible(False)\r\n pl_φ_ecl=lon\r\n\r\n #----- help_lines\r\n color_lin=\"#0D5022\"\r\n\r\n ve_=self.line_vect(v_RA, v_pl)\r\n id_l_eq,=self.ax.plot(*ve_ ,color=color_lin, linestyle = (0, (1,2,1,3)), lw=1, alpha=0.6, picker=1)\r\n id_l_eq.set_visible(False)\r\n \r\n ve_=self.line_vect(v_hor, v_pl)\r\n id_l_ho,=self.ax.plot(*ve_ ,color=color_lin, linestyle = (0, (1,2,1,3)), lw=1, alpha=0.6, picker=1)\r\n id_l_ho.set_visible(False)\r\n\r\n ve_=self.line_vect(v_ecl, v_pl)\r\n id_l_ecl,=self.ax.plot(*ve_ ,color=color_lin, linestyle = (0, (1,2,1,3)), lw=1, alpha=0.6, picker=1)\r\n id_l_ecl.set_visible(False)\r\n\r\n self.planets_obj[id_t]={} \r\n self.planets_obj[id_t][\"id_pl\"]=id_pl\r\n self.planets_obj[id_t][\"id_Mer\"]=id_Mer \r\n self.planets_obj[id_t][\"id_Mer_nat\"]=id_Mer_nat\r\n self.planets_obj[id_t][\"id_Par\"]=id_Par \r\n self.planets_obj[id_t][\"id_proj_hor\"]=id_proj_hor \r\n self.planets_obj[id_t][\"id_proj_hor_nat\"]=id_proj_hor_nat \r\n self.planets_obj[id_t][\"name\"]=name\r\n self.planets_obj[id_t][\"txt\"]=text\r\n\r\n self.planets_obj[id_t][\"id_pt_Eq\"] = (id_pt_Eq, rd(norm_r(pl_φ_eq + self.v_equinox_ang)), v_RA, rd(pl_φ_eq)) \r\n self.planets_obj[id_t][\"id_pt_Hor\"] = (id_pt_Hor, rd(pl_φ_hor), v_hor, rd(azimuth), rd(pl_alt_hor))\r\n self.planets_obj[id_t][\"id_pt_Ecl\"] = (id_pt_Ecl, round(pl_φ_ecl,1), v_ecl)\r\n\r\n self.planets_obj[id_t][\"id_l_eq\"]=id_l_eq\r\n self.planets_obj[id_t][\"id_l_ho\"]=id_l_ho\r\n self.planets_obj[id_t][\"id_l_ecl\"]=id_l_ecl\r\n\r\n self.planets_obj[id_t][\"show_mer_par\"]=0\r\n\r\n \r\n self.planets_ids[name]={}\r\n self.planets_ids[name][\"id_pl\"]=id_pl\r\n self.planets_ids[name][\"id_t\"]=id_t\r\n self.planets_ids[name][\"id_Mer\"]=id_Mer\r\n self.planets_ids[name][\"id_Mer_nat\"]=id_Mer_nat\r\n self.planets_ids[name][\"id_Par\"]=id_Par\r\n self.planets_ids[name][\"id_proj_hor\"]=id_proj_hor \r\n self.planets_ids[name][\"id_proj_hor_nat\"]=id_proj_hor_nat\r\n self.planets_ids[name][\"id_pt_Eq\"]=id_pt_Eq\r\n self.planets_ids[name][\"id_pt_Hor\"]=id_pt_Hor\r\n self.planets_ids[name][\"id_pt_Ecl\"]=id_pt_Ecl\r\n self.planets_ids[name][\"id_l_eq\"]=id_l_eq\r\n self.planets_ids[name][\"id_l_ho\"]=id_l_ho\r\n self.planets_ids[name][\"id_l_ecl\"]=id_l_ecl\r\n\r\n\r\n def plot_planets_next(self, planets_data):\r\n for i, pl in enumerate(planets_data):\r\n if pl not in c.planets.keys():continue\r\n pl_RA=planets_data[pl]['eq'][0]\r\n pl_decl=planets_data[pl]['eq'][1]\r\n lon=planets_data[pl]['ecl'][0]\r\n lat=planets_data[pl]['ecl'][1]\r\n txt=\"{}: RA={:.2f}°, decl={:.2f}°, lon={:.1f}°, lat={:.1f}°\".format(pl, pl_RA, pl_decl, lon, lat)\r\n self.plot_planet_equat_next(pl_RA ,pl_decl, lon, lat ,rad=False,meridian=False, name=pl, text=txt) \r\n\r\n def plot_planet_equat_next(self, RA, decl, lon, lat, rad=False, meridian=False, parallel=False, name=\"\", text=\"\"):\r\n pl_RA=np.radians(RA) if rad==False else RA\r\n pl_decl=np.radians(decl) if rad==False else decl\r\n rotation=pl_RA\r\n φ=self.ψ\r\n θ=self.ψ\r\n k=self.k\r\n j=np.array([0,0,1])\r\n k2 = self.poleN;\r\n v2 = self.equator_rot; \r\n\r\n color_pl2=pl_colors[name]\r\n color_pl1=pl_colors[name]\r\n\r\n rotation=rotation - self.v_equinox_ang\r\n v_rot5 = self.Rodrigues_rotation(v2,k2,rotation)\r\n \r\n v_mer = self.circl_vect(v_rot5,k2,φ)\r\n style_=\"dashed\"\r\n color_=\"#543E17\" \r\n self.planets_ids[name][\"id_Mer\"].set_data_3d(*v_mer)\r\n\r\n shift=np.sin(pl_decl)\r\n data=self.plot_parallel(shift, redraw=False) \r\n id_Par=data[0]\r\n v_par=data[1]\r\n self.planets_ids[name][\"id_Par\"].set_data_3d(*v_par) \r\n\r\n φ2 = pl_RA - self.v_equinox_ang\r\n φ2=norm_r(φ2)\r\n pl_φ_eq=φ2\r\n\r\n v_RA = self.circl_vect(self.equator_rot,k,φ2)\r\n\r\n self.planets_ids[name][\"id_pt_Eq\"].set_data_3d(*v_RA)\r\n \r\n pl_declination=pl_decl\r\n θ2=pl_decl\r\n v_pl = self.circl_vect(v_RA,k2,θ2)\r\n\r\n pl_θ_hor=np.arccos(v_pl[2])\r\n\r\n self.planets_ids[name][\"id_pl\"].set_data_3d(*v_pl) \r\n arr=[-0.05,0.05]\r\n shift=0.07\r\n v_vert=np.cross(v_pl,j)\r\n v_vert=v_vert/np.linalg.norm(v_vert)\r\n v_pl_t=np.add(v_pl,shift)\r\n\r\n self.planets_ids[name][\"id_t\"].set_position_3d(v_pl_t);\r\n\r\n\r\n v_hor=[v_pl[0],v_pl[1],0]\r\n v_hor=v_hor/np.linalg.norm(v_hor) \r\n self.planets_ids[name][\"id_pt_Hor\"].set_data_3d(*v_hor)\r\n\r\n pl_φ_hor= np.arccos(v_hor[0]); \r\n\r\n if v_hor[1] < 0:\r\n pl_φ_hor = 2*np.pi - pl_φ_hor\r\n\r\n azimuth = 2*np.pi - pl_φ_hor + np.pi/2\r\n azimuth = norm_r(azimuth)\r\n\r\n pl_θ_hor= np.arccos(v_pl[2])\r\n pl_alt_hor= np.pi/2 - pl_θ_hor\r\n\r\n α = np.arctan(np.tan(np.pi/2 - pl_θ_hor)/np.sin(np.pi/2 - pl_φ_hor)) # α = proportionate horizon's inclination\r\n k=self.k\r\n w=self.w\r\n v_rot8 = self.Rodrigues_rotation(k, w, α)\r\n φ2 = self.ψ2\r\n x2,y2,z2 = self.circl_vect(v_rot8, w, φ2)\r\n self.planets_ids[name][\"id_proj_hor\"].set_data_3d(x2,y2,z2)\r\n\r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot \r\n φ2=np.radians(lon) \r\n v_ecl = self.circl_vect(k2,v2,φ2)\r\n self.planets_ids[name][\"id_pt_Ecl\"].set_data_3d(*v_ecl)\r\n pl_φ_ecl=lon\r\n\r\n #----- help_lines\r\n ve_=self.line_vect(v_RA, v_pl)\r\n self.planets_ids[name][\"id_l_eq\"].set_data_3d(*ve_)\r\n\r\n ve_=self.line_vect(v_hor, v_pl)\r\n self.planets_ids[name][\"id_l_ho\"].set_data_3d(*ve_)\r\n \r\n ve_=self.line_vect(v_ecl, v_pl)\r\n self.planets_ids[name][\"id_l_ecl\"].set_data_3d(*ve_)\r\n\r\n id_l_ecl=0\r\n id_t=self.planets_ids[name][\"id_t\"]\r\n id_pt_Eq=self.planets_obj[id_t][\"id_pt_Eq\"][0]\r\n id_pt_Hor=self.planets_obj[id_t][\"id_pt_Hor\"][0]\r\n id_pt_Ecl=self.planets_obj[id_t][\"id_pt_Ecl\"][0]\r\n self.planets_obj[id_t][\"id_pt_Eq\"] = (id_pt_Eq, rd(norm_r(pl_φ_eq + self.v_equinox_ang)), v_RA, rd(pl_φ_eq)) \r\n self.planets_obj[id_t][\"id_pt_Hor\"] = (id_pt_Hor, rd(pl_φ_hor), v_hor, rd(azimuth), rd(pl_alt_hor))\r\n self.planets_obj[id_t][\"id_pt_Ecl\"] = (id_pt_Ecl, round(pl_φ_ecl,1), v_ecl)\r\n self.planets_obj[id_t][\"txt\"]=text\r\n\r\n def plot_zodiac(self): \r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot\r\n ψ=0;i=1\r\n while ψ < 2*np.pi:\r\n v_ = self.circl_vect(k2,v2,ψ)\r\n v_1=np.multiply(v_, 1.01)\r\n v_2=np.multiply(v_, 0.99)\r\n data=np.array([v_1,v_2])\r\n id_1, = self.ax.plot(*v_, c=color_ecl_zod, marker=\"D\", markersize=2, picker=3, zorder=2)\r\n\r\n name=zodiac[i]\r\n sym =zodiac2[i]\r\n self.zodiac_obj[id_1] = (name, sym, v_)\r\n\r\n sym = \"$\" + zodiac2[i] + \"$\"\r\n v_s = self.circl_vect(k2, v2, ψ + np.pi/12)\r\n v_s=np.multiply(v_s, 1.05)\r\n id_2, = self.ax.plot(*v_s, c=\"#4E4C25\", marker=sym, markersize=10, picker=3, zorder=0, alpha=0.2)\r\n self.ecl_ids.append(id_2)\r\n\r\n self.zodiac_ids[i]={}\r\n self.zodiac_ids[i][\"dot\"]=id_1\r\n self.zodiac_ids[i][\"text\"]=id_2\r\n\r\n ψ+=np.pi/6\r\n i+=1\r\n\r\n def plot_zodiac_next(self):\r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot\r\n ψ=0;i=1\r\n while ψ < 2*np.pi:\r\n v_ = self.circl_vect(k2,v2,ψ)\r\n v_1=np.multiply(v_, 1.01)\r\n v_2=np.multiply(v_, 0.99)\r\n data=np.array([v_1,v_2])\r\n self.zodiac_ids[i][\"dot\"].set_data_3d(*v_) \r\n tup=self.zodiac_obj[self.zodiac_ids[i][\"dot\"]] \r\n self.zodiac_obj[self.zodiac_ids[i][\"dot\"]]=(tup[1],tup[0],v_)\r\n v_s = self.circl_vect(k2, v2, ψ + np.pi/12)\r\n v_s=np.multiply(v_s, 1.05)\r\n self.zodiac_ids[i][\"text\"].set_data_3d(*v_s)\r\n \r\n ψ+=np.pi/6\r\n i+=1\r\n\r\n def plot_ecliptic_scale(self): \r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot\r\n ψ=0;i=1\r\n first=True if len(self.ecliptic_scale)==0 else False\r\n while ψ < 360:\r\n v_ = self.circl_vect(k2,v2,np.radians(ψ))\r\n if first:\r\n id_, = self.ax.plot(*v_, c=\"#333\", marker=\"o\", markersize=1)#,zorder=1\r\n id_t=self.ax.text(*v_,s=str(ψ),fontsize=6,c=\"#333\", fontweight=400, picker=5, alpha=self.alpha_main, zorder=0)\r\n self.ecliptic_scale[i]=(id_,id_t)\r\n id_.set_visible(False)\r\n id_t.set_visible(False)\r\n else:\r\n self.ecliptic_scale[i][0].set_data_3d(*v_)\r\n self.ecliptic_scale[i][1].set_position_3d(v_)\r\n ψ+=10\r\n i+=1\r\n\r\n def plot_equator_scale(self): \r\n k2=np.array(self.v_equinox) \r\n v2= -self.ecliptic_rot\r\n ψ=0;i=1\r\n first=True if len(self.equator_scale)==0 else False\r\n while ψ < 360:\r\n φ2 = np.radians(ψ) - self.v_equinox_ang\r\n φ2=norm_r(φ2)\r\n v_ = self.circl_vect(self.equator_rot,self.k,φ2) \r\n if first:\r\n id_, = self.ax.plot(*v_, c=\"#333\", marker=\"o\", markersize=1)#,zorder=1\r\n id_t=self.ax.text(*v_,s=str(ψ),fontsize=6,c=\"#333\", fontweight=400, picker=5, alpha=self.alpha_main, zorder=0)\r\n self.equator_scale[i]=(id_,id_t)\r\n id_.set_visible(False)\r\n id_t.set_visible(False)\r\n else:\r\n self.equator_scale[i][0].set_data_3d(*v_)\r\n self.equator_scale[i][1].set_position_3d(v_)\r\n ψ+=10\r\n i+=1\r\n\r\n\r\n def plot_circle_scale(self, circle):\r\n k2=np.array([1,0,0])\r\n v2=np.array([0,1,0])\r\n if circle==\"Horizon\":\r\n obj_scale=self.Hor_scale={}\r\n k2=np.array([1,0,0])\r\n v2=np.array([0,1,0])\r\n elif circle==\"Prime vertical\":\r\n obj_scale=self.PV_scale={}\r\n k2=np.array([1,0,0])\r\n v2=np.array([0,0,1])\r\n\r\n ψ=0;i=1\r\n while ψ < 360:\r\n v_ = self.circl_vect(k2,v2,np.radians(ψ)) \r\n id_, = self.ax.plot(*v_, c=\"#333\", marker=\"o\", markersize=1)\r\n id_t=self.ax.text(*v_,s=str(ψ),fontsize=6,c=\"#333\", fontweight=400, picker=5, alpha=self.alpha_main, zorder=0)\r\n obj_scale[i]=(id_,id_t)\r\n id_.set_visible(False)\r\n id_t.set_visible(False)\r\n ψ+=10\r\n i+=1\r\n\r\n def draw_projected_horizon(self,rotation=np.pi/6, color_=\"#2D305C\", linestyle_=\"dashed\"):\r\n k=self.k\r\n w=np.array([0, -1, 0])\r\n v_rot6 = self.Rodrigues_rotation(k,w,rotation)\r\n φ2 = self.ψ2\r\n x2= w[0] * np.sin(φ2) + v_rot6[0] * np.cos(φ2)\r\n y2= w[1] * np.sin(φ2) + v_rot6[1] * np.cos(φ2)\r\n z2= w[2] * np.sin(φ2) + v_rot6[2] * np.cos(φ2)\r\n id_,=self.ax.plot(x2,y2,z2 ,color=color_, linestyle = linestyle_, linewidth=1)\r\n self.ids[id(id_)] =[[x2[3],y2[3],z2[0]],\"proportionate horizon\"] \r\n self.proj_horizons.append(id_)\r\n return id_\r\n\r\n def plot_parallel(self,shift,color_=\"#CABF6B\", alpha=1,style_=(0, (1, 3)), redraw=True):\r\n frac_poleN=np.multiply(shift,self.poleN) \r\n o_=[0,0,0]; o_shift=np.add(o_,frac_poleN)\r\n c=o_shift\r\n r=shiftx=np.sqrt(1-np.square(shift))\r\n v_rot_shift=self.equator_rot; k_shift=self.k\r\n v_rot_shift= np.multiply(r,v_rot_shift); k_shift= np.multiply(r,k_shift); \r\n v = k_shift ; w = v_rot_shift\r\n φ = self.ψ\r\n x =c[0] + v[0]*np.cos(φ) + w[0]*np.sin(φ)\r\n y =c[1] + v[1]*np.cos(φ) + w[1]*np.sin(φ)\r\n z =c[2] + v[2]*np.cos(φ) + w[2]*np.sin(φ)\r\n if redraw==True:\r\n circle_id, =self.ax.plot(x,y,z ,color=color_, linestyle = style_ , alpha=alpha, picker=2, lw=1)\r\n return (circle_id,[x,y,z]) \r\n else:\r\n return (None,[x,y,z])\r\n\r\n def draw_sphere(self):\r\n φ = np.linspace(0, 2 * np.pi, 100)\r\n θ = np.linspace(0, np.pi, 100)\r\n x = 1 * np.outer(np.cos(φ), np.sin(θ))\r\n y = 1 * np.outer(np.sin(φ), np.sin(θ))\r\n z = 1 * np.outer(np.ones(np.size(φ)), np.cos(θ))\r\n col_viol=\"#EFE9FF\"\r\n col_yel=\"#E9E4D1\"\r\n self.sph_alf=0.3\r\n id_ = self.ax.plot_surface(x, y, z, rstride=4, cstride=4, color=col_yel, linewidth=0, alpha=self.sph_alf)\r\n self.sphere_ = id_\r\n\r\n def draw_surface1(self):\r\n col_viol=\"#EFE9FF\"\r\n col_yel=\"#E9E4D1\"\r\n col_='#333'\r\n self.surface = Circle((0., 0.), 1, color=col_yel,alpha=0.3)\r\n self.ax.add_patch( self.surface)\r\n art3d.pathpatch_2d_to_3d( self.surface, z=0, zdir=\"z\")\r\n\r\n def draw_surface(self):\r\n col_yel1=\"#E9E4D1\"\r\n col_yel2=\"#D0CAB2\"\r\n \r\n angle=90\r\n theta1, theta2 = angle, angle + 180\r\n\r\n self.surface1 = Wedge((0., 0.), 1, theta1, theta2, fc=col_yel2,alpha=0.3)\r\n self.surface2 = Wedge((0., 0.), 1, theta2, theta1, fc=col_yel1,alpha=0.3)\r\n self.ax.add_patch(self.surface1)\r\n self.ax.add_patch(self.surface2)\r\n art3d.pathpatch_2d_to_3d(self.surface1, z=0, zdir=\"z\")\r\n art3d.pathpatch_2d_to_3d(self.surface2, z=0, zdir=\"z\")\r\n\r\n\r\n def new_now(self, planets_data, geo_latitude, data):\r\n self.planets_data=planets_data\r\n self.geo_latitude=geo_latitude\r\n self.data=data\r\n self.timestampIni = self.data[\"timestamp\"]\r\n\r\n self.plot_Equator_Ecliptic_next()\r\n self.plot_zodiac_next()\r\n self.plot_planets_next(self.planets_data)\r\n\r\n ARMC=self.planets_data[\"MC\"][\"eq\"][0]\r\n self.date_utc = self.data[\"d_utc\"]\r\n self.time_utc = self.data[\"t_utc\"]\r\n self.timestamp = self.data[\"timestamp\"]\r\n self.geo_longitude=float(self.data[\"lon\"])\r\n\r\n s1 = \"S\" if self.geo_latitude<0 else \"N\"\r\n s2 = \"W\" if self.geo_longitude<0 else \"E\"\r\n txt=\"ARMC={:.0f}°, lat={:.0f}°{}, lon={:.0f}°{}\".format(ARMC,self.geo_latitude,s1,self.geo_longitude,s2)\r\n self.id_text.set_text(txt)\r\n\r\n name = \" {} {} \".format(self.data[\"n\"], self.data[\"ln\"])\r\n time_loc_ = \" {}, {}\".format(self.data[\"date_loc\"],self.data[\"time_loc\"])\r\n name = \" {} {} \".format(self.data[\"n\"], self.data[\"ln\"])\r\n time_utc_ = \" UTC: {}, {} \".format(self.date_utc, self.time_utc)\r\n self.id_text_name.set_text(name)\r\n self.id_text_time.set_text(time_loc_)\r\n self.id_text_time2.set_text(time_utc_)\r\n\r\n self.legend_updt()\r\n\r\n self.canvas.draw() \r\n\r\n\r\n def Rodrigues_rotation(self, v, k, rotation): #Rodrigues' rotation formula #rotation about k; v= vector to rotate\r\n v_rot = v * np.cos(rotation) + np.cross(k, v) * np.sin(rotation) + k * np.dot(k, v) * (1 - np.cos(rotation))\r\n return v_rot\r\n\r\n def circl_vect(self,k,v,φ2,r=1): \r\n x = r * v[0] * np.sin(φ2) + r * k[0] * np.cos(φ2)\r\n y = r * v[1] * np.sin(φ2) + r * k[1] * np.cos(φ2)\r\n z = r * v[2] * np.sin(φ2) + r * k[2] * np.cos(φ2)\r\n return [x, y, z]\r\n\r\n def line_vect(self, k, v, n=10, col=\"#0D5022\"):\r\n x = np.linspace(k[0],v[0],n)\r\n y = np.linspace(k[1],v[1],n)\r\n z = np.linspace(k[2],v[2],n)\r\n return [x, y, z]\r\n \r\n def legend_updt(self):\r\n for leg in self.leg_obj:\r\n lin=self.leg_obj[leg][1]\r\n if lin==None:continue\r\n if isinstance(lin,tuple):lin=lin[0]\r\n isVisible = lin.get_visible()\r\n leg.set_alpha(1.0 if isVisible else 0.2)\r\n self.canvas.draw_idle()\r\n\r\n def show_annot(self,v3d,txt, interval_=5000):\r\n x2, y2, _ = proj3d.proj_transform(*v3d, self.ax.get_proj())\r\n self.annot.xy = x2, y2\r\n self.annot.set_text(txt)\r\n self.annot.set_visible(True)\r\n try:self.timer.stop()\r\n except:pass\r\n self.timer = self.canvas.new_timer(interval=interval_)\r\n self.timer.add_callback(self.clear_ann)\r\n self.timer.start()\r\n\r\n def toggle_ecliptic_scale(self, hide=False):\r\n isVisible= self.ecliptic_scale[1][1].get_visible()\r\n if hide==True:isVisible=True\r\n if self.ecl_circl_1.get_visible()==False:isVisible=True\r\n for i in self.ecliptic_scale:\r\n self.ecliptic_scale[i][0].set_visible(not isVisible)\r\n self.ecliptic_scale[i][1].set_visible(not isVisible)\r\n\r\n def toggle_equator_scale(self, hide=False):\r\n isVisible= self.equator_scale[1][1].get_visible()\r\n if hide==True:isVisible=True\r\n if self.eq_circl_1.get_visible()==False:isVisible=True\r\n for i in self.equator_scale:\r\n self.equator_scale[i][0].set_visible(not isVisible)\r\n self.equator_scale[i][1].set_visible(not isVisible) \r\n\r\n def toggle_horizon_scale(self, hide=False):\r\n isVisible= self.Hor_scale[1][1].get_visible()\r\n if hide==True:isVisible=True\r\n if self.hor_circl_1.get_visible()==False:isVisible=True\r\n for i in self.Hor_scale:\r\n self.Hor_scale[i][0].set_visible(not isVisible)\r\n self.Hor_scale[i][1].set_visible(not isVisible) \r\n\r\n def toggle_prime_vert_scale(self, hide=False):\r\n isVisible= self.PV_scale[1][1].get_visible()\r\n if hide==True:isVisible=True\r\n if self.prime_vert_1.get_visible()==False:isVisible=True\r\n for i in self.PV_scale:\r\n self.PV_scale[i][0].set_visible(not isVisible)\r\n self.PV_scale[i][1].set_visible(not isVisible) \r\n \r\n\r\n def key_(self,event):\r\n if event.key==\"escape\":\r\n pass\r\n #exit()\r\n\r\n elif event.key==\"left\":\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n azim_=azim_+1\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n self.canvas.draw_idle()\r\n\r\n elif event.key==\"right\":\r\n azim_, elev_ = self.ax.azim, self.ax.elev \r\n azim_=azim_-1\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n self.canvas.draw_idle()\r\n\r\n elif event.key==\"up\":\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n elev_+=1\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n self.canvas.draw_idle()\r\n\r\n elif event.key==\"down\":\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n elev_-=1\r\n self.ax.view_init(azim = azim_, elev = elev_)\r\n self.canvas.draw_idle()\r\n\r\n if event.key==\"j\":\r\n self.ax.view_init(azim = 0, elev = 0) \r\n self.canvas.draw_idle()\r\n if event.key==\"h\":\r\n self.canvas.draw_idle()\r\n elif event.key==\"ctrl+s\":\r\n filename=r\"3D_astro\"\r\n filename=asksaveasfilename(parent=self.parent,title=\"Save file\",initialdir=\"C:\\\\\",initialfile = filename,filetypes=[('image, .png', '*.png'),('All Files', '*.*')])\r\n if filename==\"\":return\r\n filename=f\"{filename}.png\"\r\n self.fig.savefig(filename)\r\n\r\n def onpick(self,event):\r\n if event.mouseevent.button in [2,\"up\",\"down\"]:return\r\n legend = event.artist\r\n if legend in self.legend.get_lines():\r\n txt=self.leg_obj[legend]\r\n txt=self.leg_obj[legend][0]\r\n if txt==\"Axes\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.axes_id.get_alpha() != self.alpha_main else False\r\n alfa2= self.alpha_main if dim else 0.2\r\n self.axes_id.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.axes_id.get_visible()\r\n self.axes_id.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n\r\n elif txt==\"Sphere\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n for id_ in [self.surface1, self.surface2]:#self.surface\r\n isVisible = id_.get_visible()\r\n id_.set_visible(not isVisible)\r\n return \r\n isVisible = self.sphere_.get_visible()\r\n self.sphere_.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2) \r\n\r\n elif txt==\"Ecliptic\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.ecl_circl_2.get_alpha() != self.alpha_main_2 else False\r\n alfa2= self.alpha_main_2 if dim else 0.3\r\n for cir in [self.ecl_circl_1, self.ecl_circl_2]:\r\n cir.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.ecl_circl_1.get_visible()\r\n self.ecl_circl_1.set_visible(not isVisible)\r\n self.ecl_circl_2.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n for id_ in self.ecl_ids:\r\n id_.set_visible(not isVisible)\r\n for id_ in self.houses_obj:\r\n id_.set_visible(not isVisible)\r\n for id_ in self.zodiac_obj:\r\n id_.set_visible(not isVisible)\r\n\r\n self.toggle_ecliptic_scale(hide= True)\r\n\r\n elif txt==\"Equator\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.eq_circl_2.get_alpha() != self.alpha_main_2 else False\r\n alfa2= self.alpha_main_2 if dim else 0.3\r\n for cir in [self.eq_circl_1, self.eq_circl_2]:\r\n cir.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.eq_circl_1.get_visible()\r\n self.eq_circl_1.set_visible(not isVisible)\r\n self.eq_circl_2.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n for id_ in self.equat_ids:\r\n id_.set_visible(not isVisible)\r\n\r\n self.toggle_equator_scale(hide= True) \r\n\r\n elif txt==\"Horizon\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.hor_circl_2.get_alpha() != self.alpha_main else False\r\n alfa2= self.alpha_main if dim else 0.2\r\n for cir in [self.hor_circl_1, self.hor_circl_2]:\r\n cir.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.hor_circl_1.get_visible()\r\n self.hor_circl_1.set_visible(not isVisible)\r\n self.hor_circl_2.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n\r\n self.toggle_horizon_scale(hide= True) \r\n\r\n elif txt==\"Prime Vertical\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.prime_vert_2.get_alpha() != self.alpha_main else False\r\n alfa2= self.alpha_main if dim else 0.2\r\n for cir in [self.prime_vert_1, self.prime_vert_2]:\r\n cir.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.prime_vert_1.get_visible()\r\n self.prime_vert_1.set_visible(not isVisible)\r\n self.prime_vert_2.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n\r\n self.toggle_prime_vert_scale(hide= True)\r\n\r\n elif txt==\"Meridian\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n dim = True if self.mer_circle.get_alpha() != self.alpha_main else False\r\n alfa2= self.alpha_main if dim else 0.2\r\n self.mer_circle.set_alpha(alfa2)\r\n self.canvas.draw_idle()\r\n return\r\n isVisible = self.mer_circle.get_visible()\r\n self.mer_circle.set_visible(not isVisible)\r\n legend.set_alpha(1.0 if not isVisible else 0.2)\r\n\r\n elif txt==\"Show half\":\r\n self.annot.set_visible(False)\r\n if event.mouseevent.button==3:\r\n self.circles_alpha()\r\n self.canvas.draw_idle()\r\n return\r\n if self.half==0:\r\n self.half=1\r\n for id_ in self.view_West:\r\n id_.set_visible(False)\r\n for id_ in self.view_East:\r\n id_.set_visible(True) \r\n elif self.half==1:\r\n self.half=2\r\n for id_ in self.view_West:\r\n id_.set_visible(True)\r\n for id_ in self.view_East:\r\n id_.set_visible(False)\r\n elif self.half==2:\r\n self.half=0\r\n for id_ in self.view_West:\r\n id_.set_visible(True)\r\n for id_ in self.view_East:\r\n id_.set_visible(True)\r\n\r\n elif txt==\"Extra off\":\r\n for id_t in self.planets_obj:\r\n obj = self.planets_obj[id_t] \r\n for el in [\"id_Mer\", \"id_Mer_nat\", \"id_Par\", \"id_proj_hor\", \"id_proj_hor_nat\"]:\r\n obj[el].set_visible(False)\r\n for el in [\"id_pt_Eq\", \"id_pt_Hor\", \"id_pt_Ecl\"]:\r\n obj[el][0].set_visible(False)\r\n for el in [\"id_l_eq\", \"id_l_ecl\",\"id_l_ho\"]:\r\n obj[el].set_visible(False)\r\n self.annot.set_visible(False)\r\n\r\n self.toggle_ecliptic_scale(hide= True) \r\n self.toggle_equator_scale(hide= True) \r\n self.toggle_horizon_scale(hide= True) \r\n self.toggle_prime_vert_scale(hide= True)\r\n\r\n elif txt==\"Scale\":\r\n if event.mouseevent.button==3:\r\n self.toggle_equator_scale() \r\n else:\r\n self.toggle_ecliptic_scale()\r\n self.toggle_equator_scale()\r\n self.canvas.draw_idle()\r\n return\r\n\r\n if event.artist==self.id_v_start:\r\n if event.mouseevent.button==1:\r\n self.timestamp=self.timestampIni\r\n self.plot_next_prev(minutes=0)\r\n self.extra_off()\r\n self.ax.set_xlim(self.xlim)\r\n self.ax.set_ylim(self.ylim)\r\n self.ax.set_zlim(self.zlim)\r\n self.ax.view_init(azim = self.azim0, elev = self.elev0) \r\n self.mer_circle.set_visible(True)\r\n self.prime_vert_1.set_visible(True)\r\n self.prime_vert_2.set_visible(True) \r\n\r\n elif event.artist==self.id_v_chart:\r\n if event.mouseevent.button==3:\r\n azim_= 90 + np.degrees(self.Asc_φ_hor)\r\n elev_=self.pole_ecl_elev+90\r\n if self.south==True:elev_=self.pole_ecl_elev-90\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n else:\r\n azim_=90 + np.degrees(self.Asc_φ_hor)\r\n a1('azim_ 90 + np.degrees(self.Asc_φ_hor)',azim_,\"elev =\", self.pole_ecl_elev); #71,24\r\n self.ax.view_init(azim = azim_, elev = self.pole_ecl_elev) \r\n self.mer_circle.set_visible(False)\r\n self.prime_vert_1.set_visible(False)\r\n self.prime_vert_2.set_visible(False) \r\n if event.mouseevent.button==3:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n self.ax.view_init(azim = 30, elev = elev_)\r\n\r\n \r\n elif event.artist==self.id_v_Eq:\r\n if event.mouseevent.button == 3:\r\n self.ax.view_init(azim = 90, elev = self.poleN_elev-90)\r\n else:\r\n self.ax.view_init(azim = 90, elev = self.poleN_elev)\r\n\r\n elif event.artist==self.id_v_anim:\r\n s=0; m=0; h=0\r\n val=self.radio.value_selected\r\n Δt=int(self.text_box.text)\r\n if val==\"sec\":\r\n s=Δt\r\n elif val==\"min\":\r\n m=Δt\r\n elif val==\"hour\":\r\n h=Δt\r\n \r\n if event.mouseevent.button==3:\r\n s=-s ; m=-m ; h=-h\r\n for i in range(10):\r\n self.parent.after(i*500, lambda s=s, m=m, h=h: self.plot_next_prev(seconds=s, minutes=m, hours=h))\r\n self.canvas.draw_idle() \r\n\r\n elif event.artist==self.id_v_prev:\r\n s=self.radio.value_selected\r\n Δt=int(self.text_box.text)\r\n if s==\"sec\":\r\n self.plot_next_prev(seconds=-Δt)\r\n elif s==\"min\":\r\n self.plot_next_prev(minutes=-Δt) \r\n elif s==\"hour\":\r\n self.plot_next_prev(hours=-Δt)\r\n\r\n val=self.radio.value_selected\r\n Δt=int(self.text_box.text)\r\n if val==\"sec\":\r\n s=Δt\r\n elif val==\"min\":\r\n m=Δt\r\n elif val==\"hour\":\r\n h=Δt\r\n\r\n elif event.artist==self.id_v_next:\r\n s=self.radio.value_selected\r\n Δt=int(self.text_box.text)\r\n if s==\"sec\":\r\n self.plot_next_prev(seconds=Δt)\r\n elif s==\"min\":\r\n\r\n self.plot_next_prev(minutes=Δt) \r\n elif s==\"hour\":\r\n self.plot_next_prev(hours=Δt)\r\n\r\n elif event.artist==self.id_E:\r\n self.ax.view_init(azim = 30,elev = 15) \r\n\r\n elif event.artist==self.id_W:\r\n self.ax.view_init(azim = 200, elev = 15)\r\n\r\n elif event.artist==self.id_N:\r\n self.ax.view_init(azim = 180, elev = -90) \r\n \r\n elif event.artist==self.id_S:\r\n self.ax.view_init(azim = 180, elev = 90) \r\n\r\n elif event.artist==self.id_A0:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = 0, elev = elev_) \r\n else:\r\n self.ax.view_init(azim = 0, elev = elev_) \r\n\r\n elif event.artist==self.id_E0:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = azim_, elev = 0) \r\n else:\r\n self.ax.view_init(azim = azim_, elev = 0) \r\n\r\n elif event.artist==self.id_A90:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = -90, elev = elev_) \r\n else:\r\n self.ax.view_init(azim = 90, elev = elev_) \r\n\r\n elif event.artist==self.id_A180:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = 180, elev = elev_) \r\n else:\r\n azim_=azim_+10\r\n self.ax.view_init(azim = 180, elev = elev_) \r\n\r\n elif event.artist==self.id_E90:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = azim_, elev = -90) \r\n else:\r\n self.ax.view_init(azim = azim_, elev = 90) \r\n\r\n elif event.artist==self.id_E180:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n self.ax.view_init(azim = azim_, elev = 180) \r\n else:\r\n self.ax.view_init(azim = azim_, elev = 180) \r\n\r\n elif event.artist==self.id_v_left:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n azim_-=1\r\n else:\r\n azim_-=10\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n\r\n elif event.artist==self.id_v_right:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n azim_+=1\r\n else: \r\n azim_+=10\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n\r\n elif event.artist==self.id_v_up:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n elev_+=1\r\n else: \r\n elev_=elev_+10\r\n self.ax.view_init(azim = azim_, elev = elev_) \r\n\r\n elif event.artist==self.id_v_down:\r\n azim_, elev_ = self.ax.azim, self.ax.elev\r\n if event.mouseevent.button==3:\r\n elev_-=1\r\n else: \r\n elev_=elev_-10\r\n self.ax.view_init(azim = azim_, elev = elev_)\r\n\r\n elif event.artist==self.id_v_test:\r\n pass\r\n\r\n elif event.artist==self.id_help_ico:\r\n self.helpDialog = helpWindow_3D(self.parent,title=\"Help\")\r\n\r\n if event.artist in [self.id_v_start, self.id_v_up, self.id_v_down, self.id_v_left, self.id_v_right, self.id_E, self.id_W, self.id_N, self.id_S, self.id_E0, self.id_A0, self.id_E90, self.id_E180, self.id_A90, self.id_A180,self.id_v_Eq, self.id_v_chart, self.id_v_test,self.id_view, self.id_azim, self.id_elev]:\r\n azim, elev = self.ax.azim, self.ax.elev\r\n txt=\" azim={:.1f}°, elev={:.1f}° \".format(azim, elev)\r\n self.id_text_2.set_text(txt)\r\n self.canvas.draw_idle()\r\n self.timer_azim_elev(15000) \r\n return\r\n\r\n if event.artist in [self.eq_circl_1, self.eq_circl_2] and self.eq_circl_1.get_visible():\r\n el=event.artist\r\n el=event.artist\r\n ind = event.ind\r\n xx, yy, zz = el.get_data_3d()\r\n i=ind[0]\r\n x,y,z=[xx[i],yy[i],zz[i]]\r\n φ2=np.arccos(x);\r\n φ2a=np.degrees(φ2)\r\n if y>=0:\r\n φ2 = φ2 + np.pi/2\r\n elif y<0:\r\n φ2 = np.pi/2 - φ2\r\n if φ2<0: φ2 = φ2 + 2*np.pi \r\n if self.south==True: φ2=np.pi-φ2\r\n φ2=norm_r(φ2) \r\n φ2a=np.degrees(φ2)\r\n RA= norm_(φ2a + self.ARMC)\r\n self.show_txt_tip(RA,\"Equator\")\r\n\r\n elif event.artist in [self.ecl_circl_1, self.ecl_circl_2] and self.ecl_circl_1.get_visible():\r\n el=event.artist\r\n ind = event.ind\r\n xx, yy, zz = el.get_data_3d()\r\n i=ind[0]\r\n x,y,z=[xx[i],yy[i],zz[i]]\r\n d=np.dot(self.v_equinox,[x,y,z]);\r\n φ2=np.arccos(d)\r\n lon=np.degrees(φ2)\r\n \r\n if self.south==True: φ2=np.pi-φ2\r\n φ2=norm_r(φ2) \r\n φ2a=np.degrees(φ2)\r\n v=np.cross(self.v_equinox,[x,y,z])\r\n\r\n if self.v_MC[1]<=0 and v[2]<0:\r\n lon=norm_(360-lon)\r\n if self.v_MC[1]>0 and v[2]>0:\r\n lon=norm_(360-lon)\r\n\r\n self.show_txt_tip(lon,\"Ecliptic\")\r\n\r\n\r\n if event.mouseevent.button==3:\r\n for id_t in self.planets_obj.keys(): # click circles to hide\r\n obj = self.planets_obj[id_t]\r\n for key_ in [\"id_Mer\", \"id_Par\", \"id_proj_hor\", \"id_Mer_nat\", \"id_proj_hor_nat\"]:\r\n id_=obj[key_]\r\n if event.artist==id_:\r\n id_.set_visible(False)\r\n return\r\n\r\n\r\n for id_t in self.planets_obj.keys(): # click planet text\r\n obj = self.planets_obj[id_t]\r\n if event.artist==id_t:\r\n if event.mouseevent.button==3:\r\n pass\r\n '''\r\n isVisible = obj[\"id_Mer_nat\"].get_visible()\r\n obj[\"id_Mer_nat\"].set_visible(not isVisible)\r\n obj[\"id_proj_hor_nat\"].set_visible(not isVisible)\r\n '''\r\n else:\r\n pass\r\n \r\n φ_hor = obj[\"id_pt_Hor\"][1]\r\n azimuth = obj[\"id_pt_Hor\"][3]\r\n altitude = obj[\"id_pt_Hor\"][4]\r\n txt=\" {}, azim={:.1f}°(E), {:.1f}°(N), alt={:.1f}°\".format(obj[\"txt\"], φ_hor, azimuth, altitude)\r\n txt=txt.replace(\"Node_\",\"Node \")\r\n self.id_text_2.set_text(txt)\r\n self.timer_azim_elev(15000) \r\n self.canvas.draw_idle()\r\n return\r\n\r\n for id_t in self.planets_obj.keys(): # click planet dot\r\n obj = self.planets_obj[id_t]\r\n id_pl = obj[\"id_pl\"]\r\n if event.artist==id_pl:\r\n if id_pl.get_visible()==False:return\r\n if event.mouseevent.button==1:\r\n if self.planets_obj[id_t][\"show_mer_par\"]==0:\r\n obj[\"id_Mer\"].set_visible(True)\r\n obj[\"id_Par\"].set_visible(True)\r\n #obj[\"id_proj_hor\"].set_visible(True) #planet's proportionate horizon\r\n self.planets_obj[id_t][\"show_mer_par\"]=1\r\n elif self.planets_obj[id_t][\"show_mer_par\"]==1:\r\n obj[\"id_Mer\"].set_visible(True)\r\n obj[\"id_Par\"].set_visible(False)\r\n #obj[\"id_proj_hor\"].set_visible(False)\r\n self.planets_obj[id_t][\"show_mer_par\"]=2\r\n elif self.planets_obj[id_t][\"show_mer_par\"]==2:\r\n obj[\"id_Mer\"].set_visible(False)\r\n obj[\"id_Par\"].set_visible(True)\r\n #obj[\"id_proj_hor\"].set_visible(False)\r\n self.planets_obj[id_t][\"show_mer_par\"]=3\r\n elif self.planets_obj[id_t][\"show_mer_par\"]==3:\r\n obj[\"id_Mer\"].set_visible(False)\r\n obj[\"id_Par\"].set_visible(False)\r\n #obj[\"id_proj_hor\"].set_visible(False)\r\n self.planets_obj[id_t][\"show_mer_par\"]=0\r\n\r\n elif event.mouseevent.button==3:\r\n for el in [\"id_pt_Eq\", \"id_pt_Hor\", \"id_pt_Ecl\"]:\r\n isVisible = obj[el][0].get_visible()\r\n obj[el][0].set_visible(not isVisible)\r\n for el in [\"id_l_eq\", \"id_l_ecl\", \"id_l_ho\"]:\r\n isVisible = obj[el].get_visible()\r\n obj[el].set_visible(not isVisible)\r\n\r\n φ_hor = obj[\"id_pt_Hor\"][1]\r\n azimuth = obj[\"id_pt_Hor\"][3]\r\n altitude = obj[\"id_pt_Hor\"][4]\r\n txt=\" {}, azim={:.1f}°(E), {:.1f}°(N), alt={:.1f}°\".format(obj[\"txt\"], φ_hor, azimuth, altitude)\r\n txt=txt.replace(\"Node_\",\"Node \")\r\n self.id_text_2.set_text(txt)\r\n self.timer_azim_elev(15000) \r\n return\r\n\r\n for id_t in self.planets_obj.keys(): # click projection dot\r\n obj = self.planets_obj[id_t]\r\n id_pl = obj[\"id_pl\"]\r\n id_dot_Eq = obj[\"id_pt_Eq\"][0]\r\n id_dot_Ec = obj[\"id_pt_Ecl\"][0]\r\n id_dot_H = obj[\"id_pt_Hor\"][0]\r\n name = obj[\"name\"]\r\n\r\n if event.artist in [id_dot_Eq, id_dot_Ec ,id_dot_H]:\r\n if event.mouseevent.button == 3:\r\n self.clear_ann(event)\r\n else:\r\n if event.artist ==id_dot_Eq:\r\n if id_dot_Eq.get_visible()==False:return\r\n txt=\" {} RA={:.0f}°\".format(name, obj[\"id_pt_Eq\"][1])\r\n xyz=obj[\"id_pt_Eq\"][2]\r\n elif event.artist ==id_dot_H:\r\n if id_dot_H.get_visible()==False:return\r\n txt=\" {} azim={:.0f}°(from E) \".format(name,obj[\"id_pt_Hor\"][1])# φ hor.\r\n xyz=obj[\"id_pt_Hor\"][2] \r\n elif event.artist ==id_dot_Ec:\r\n if id_dot_Ec.get_visible()==False:return \r\n txt=\" {} ecl. long={:.0f}° \".format(name, obj[\"id_pt_Ecl\"][1])\r\n xyz=obj[\"id_pt_Ecl\"][2] \r\n self.show_annot(xyz,txt, interval_=2000)\r\n self.canvas.draw_idle()\r\n return\r\n\r\n for id_ in self.houses_obj.keys():\r\n if event.artist==id_:\r\n obj = self.houses_obj[id_]\r\n txt=\" {}, {} \".format(obj[0], obj[1])\r\n self.show_annot(obj[2],txt)\r\n\r\n if event.mouseevent.button == 3:\r\n self.clear_ann(event)\r\n else:\r\n self.canvas.draw_idle()\r\n return\r\n\r\n for id_ in self.zodiac_obj.keys(): \r\n if event.artist==id_:\r\n obj = self.zodiac_obj[id_]\r\n txt=\" {}, {} \".format(obj[0], obj[1])\r\n self.show_annot(obj[2],txt)\r\n x2, y2, _ = proj3d.proj_transform(*obj[2], self.ax.get_proj())\r\n if event.mouseevent.button == 3:\r\n self.clear_ann(event)\r\n else:\r\n self.canvas.draw_idle()\r\n return\r\n\r\n\r\n if id(event.artist) in self.ids.keys():\r\n xyz=self.ids[id(event.artist)][0]\r\n txt=self.ids[id(event.artist)][1]\r\n for i in range(len(self.click_tip_list)):\r\n if self.click_tip_list[i] in txt:\r\n s=True\r\n x2, y2, _ = proj3d.proj_transform(xyz[0],xyz[1],xyz[2], self.ax.get_proj())\r\n self.annot.xy = x2, y2\r\n self.annot.set_text(txt)\r\n self.annot.set_visible(True)\r\n\r\n try:self.timer.stop()\r\n except:pass\r\n self.timer = self.canvas.new_timer(interval=5000)\r\n self.timer.add_callback(self.clear_ann)\r\n self.timer.start()\r\n self.canvas.draw_idle()\r\n\r\n def on_click(self,event):\r\n if event.button==3:\r\n self.clear_ann()\r\n\r\n def show_txt_tip(self,dig,which=\"\"):\r\n if which==\"Ecliptic\":\r\n txt = \" Click on Ecliptic: lon={:.0f}° \".format(dig)\r\n elif which==\"Equator\":\r\n txt = \" Click on Equator: RA={:.0f}° \".format(dig)\r\n else:\r\n txt = \" {} \".format(dig)\r\n self.id_text_3.set_text(txt)\r\n self.canvas.draw_idle()\r\n try:self.timer3.stop()\r\n except:pass\r\n self.timer3 = self.canvas.new_timer(interval=15000)\r\n self.timer3.add_callback(self.hide_txt_tip)\r\n self.timer3.start()\r\n\r\n def hide_txt_tip(self):\r\n self.id_text_3.set_text(\"\")\r\n self.canvas.draw_idle()\r\n try:self.timer3.stop()\r\n except:pass\r\n\r\n def zoom_(self, event, r):\r\n l_x1, l_x2 = self.ax.get_xlim()\r\n l_y1, l_y2 = self.ax.get_ylim()\r\n xdata = event.xdata\r\n ydata = event.ydata\r\n width = l_x2 - l_x1\r\n height = l_y2 - l_y1\r\n w = r*width ; h = r*height\r\n dx1 = xdata - (xdata - l_x1) * r\r\n dx2 = xdata + (l_x2 - xdata) * r\r\n dy1 = ydata - (ydata - l_y1) * r\r\n dy2 = ydata + (l_y2 - ydata) * r\r\n self.ax.set_xlim(dx1, dx2)\r\n self.ax.set_ylim(dy1, dy2)\r\n self.ax.figure.canvas.draw()\r\n\r\n def scroll_zoom(self, event):\r\n r=0.20\r\n r=1-r\r\n if event.button == \"up\":\r\n self.zoom_(event,r) \r\n elif event.button == \"down\":\r\n self.zoom_(event,1/r)\r\n\r\n\r\n def extra_off(self):\r\n for id_t in self.planets_obj:\r\n obj = self.planets_obj[id_t] \r\n for el in [\"id_Mer\", \"id_Mer_nat\", \"id_Par\", \"id_proj_hor\", \"id_proj_hor_nat\"]:\r\n obj[el].set_visible(False)\r\n for el in [\"id_pt_Eq\", \"id_pt_Hor\", \"id_pt_Ecl\"]:\r\n obj[el][0].set_visible(False)\r\n for el in [\"id_l_eq\", \"id_l_ecl\",\"id_l_ho\"]:\r\n obj[el].set_visible(False)\r\n self.annot.set_visible(False)\r\n #self.hide_h_line()\r\n\r\n #for el in [\"id_asc_Eq\", \"id_Mer\", \"id_l_eq\",\"id_Mer_nat\"]: \r\n # self.ASC_inf[el].set_visible(False)\r\n\r\n self.toggle_ecliptic_scale(hide= True) \r\n self.toggle_equator_scale(hide= True) \r\n self.toggle_horizon_scale(hide= True) \r\n self.toggle_prime_vert_scale(hide= True)\r\n\r\n def timer_azim_elev(self, interval=15000, event=None):\r\n try:self.timer_az.stop()\r\n except:pass\r\n self.timer_az = self.canvas.new_timer(interval=interval)\r\n self.timer_az.add_callback(self.clear_azim_elev)\r\n self.timer_az.start()\r\n\r\n def clear_azim_elev(self, event=None):\r\n try:self.timer_az.stop()\r\n except:pass\r\n self.id_text_2.set_text(\"\")\r\n self.canvas.draw_idle()\r\n print(\"clear azim_elev\") \r\n \r\n def clear_ann(self, event=None):\r\n self.annot.set_text(\"\")\r\n self.annot.set_visible(False)\r\n self.canvas.draw_idle()\r\n try:self.timer.stop()\r\n except:pass\r\n\r\n def save_animation(self, frames=40, minutes=0, hours=0, format=\"gif\"):\r\n def animate(frame_n):\r\n self.plot_next_prev(minutes=minutes, hours=hours)\r\n return \r\n\r\n resp = messagebox.askquestion(\"Sample animation\", f\"Save sample animation?\\nframes={frames}\\nminutes={minutes}, hours={hours}\") \r\n if resp != \"yes\":\r\n return\r\n anim = animation.FuncAnimation(self.fig, animate, frames=frames, interval=200, repeat=False, blit=False)\r\n video_ = animation.FFMpegWriter(fps=3)\r\n if format==\"gif\":\r\n filename=\"Astronomia3D_animation\"\r\n filename=asksaveasfilename(parent=self.parent,title=\"Save file\",initialdir=\"C:\\\\\",initialfile = filename,filetypes=[('image, .gif', '*.gif'),('All Files', '*.*')])\r\n if filename==\"\":return\r\n filename=f\"{filename}.gif\"\r\n elif format==\"mp4\":\r\n filename=\"Astronomia3D_animation\"\r\n filename=asksaveasfilename(parent=self.parent,title=\"Save file\",initialdir=\"C:\\\\\",initialfile = filename,filetypes=[('video, .mp4', '*.mp4'),('All Files', '*.*')])\r\n if filename==\"\":return\r\n filename=f\"{filename}.mp4\"; \r\n\r\n anim.save(filename, writer=video_)\r\n\r\n#=====================\r\n\r\nPAGE_BG1 = \"#134752\"\r\nBUTT_BG = \"#095161\"\r\nPAGE_BG_2=\"#0F5274\"\r\nFONT_BT2 = (\"Segoe UI\", 10, \"bold\")\r\nFONT_TIT = (\"Tahoma\", 10, \"bold\")\r\nCOLOR_TIT = \"#E1E1E1\"\r\nFONT_SYM = (\"Consolas\", 15, \"normal\")\r\nFONT_N = (\"Tahoma\", 10, \"bold\")\r\n\r\nclass GUI_astro3D:\r\n def __init__(self, parent, planets_data, geo_latitude, data, title_=\"Astronomia 3D - Popiel\"):\r\n self.parent = parent\r\n self.parent.title(title_) \r\n self.parent.bind(\"\", self.keypressed)\r\n self.parent.bind(\"\", self.keypressed)\r\n self.parent.bind(\"1\", self.keypressed)\r\n self.parent.bind(\"2\", self.keypressed)\r\n self.parent.bind(\"\", self.keypressed)\r\n self.parent.bind(\"\", self.keypressed)\r\n \r\n self.page_plot = Frame(self.parent,bg=PAGE_BG1,borderwidth=1, relief=\"ridge\")\r\n self.page_plot.pack(side=\"top\",fill=\"both\",expand=True,anchor=\"sw\",ipadx=0, ipady=0) \r\n\r\n self.page0 = Frame(self.parent,bg=PAGE_BG1,borderwidth=1, relief=\"ridge\")\r\n self.page0.pack(side=\"left\",fill=\"both\",expand=True,anchor=\"nw\",ipadx=0, ipady=0)\r\n self.page1 = Frame(self.page0,bg=PAGE_BG1,borderwidth=0, relief=\"ridge\")\r\n self.page1.pack(side=\"top\",fill=\"both\",expand=True,anchor=\"nw\",ipadx=0, ipady=0,pady=(0,0))\r\n self.page2 = Frame(self.page0,bg=PAGE_BG1,borderwidth=0, relief=\"ridge\") #\r\n self.page2.pack(side=\"bottom\",fill=\"both\",expand=True,anchor=\"nw\",ipadx=0, ipady=0)\r\n\r\n self.page3 = Frame(self.parent,bg=PAGE_BG1,borderwidth=1, relief=\"ridge\")\r\n self.page3.pack(side=\"right\",fill=\"both\",expand=False,anchor=\"ne\",ipadx=0, ipady=0) \r\n self.page4 = Frame(self.parent,bg=PAGE_BG1,borderwidth=1, relief=\"ridge\")\r\n self.page4.pack(side=\"right\",fill=\"both\",expand=False,anchor=\"ne\",ipadx=0, ipady=0) \r\n self.page5 = Frame(self.parent,bg=PAGE_BG1,borderwidth=1, relief=\"ridge\")\r\n self.page5.pack(side=\"right\",fill=\"both\",expand=False,anchor=\"ne\",ipadx=0, ipady=0) \r\n\r\n self.chkvars=[]\r\n chkboxes={}\r\n for i, el in enumerate(c.planets):\r\n if el==\"Node_S\":continue\r\n chkvar = IntVar(); chkvar.set(1) \r\n self.chkvars.append(chkvar)\r\n name=el\r\n sym=c.planets[el]\r\n chkboxes[name]=chkvar\r\n font_f = \"Lucida Console\"\r\n font_ = (font_f, 13, \"normal\")\r\n padx_=1\r\n pady_=(2,0)\r\n if \"Node_\" in name:font_= (font_f, 11, \"normal\"); pady_=(3,0);\r\n if name==\"Sun\":padx_=0\r\n id_=Checkbutton(self.page1, text=sym, variable=self.chkvars[i], command = lambda name=name, chkvar=chkvar: self.planets_toggle(name,chkvar))\r\n id_.config(bg=PAGE_BG1, fg=\"#F0F0F0\", offvalue = 0, onvalue=1, font=font_, activebackground=PAGE_BG_2, activeforeground=\"#F0F0F0\", selectcolor=PAGE_BG_2, justify=\"center\", borderwidth=3)\r\n id_.grid(row=0, column=i+2,sticky='nw', padx=padx_, pady=pady_)\r\n\r\n lbl_empt = Label(self.page1, text=\" \", width=2, justify=\"center\", bg=PAGE_BG1, font=(\"Tahoma\", 8, \"bold\"))\r\n lbl_empt.grid(row=0, column=15,sticky=\"nw\", padx=0, pady=(2,0))\r\n\r\n self.chkvar_all = IntVar() ; self.chkvar_all.set(1)\r\n self.chk_all = Checkbutton(self.page1, text=\"All\", variable=self.chkvar_all, command = lambda name=\"All\", chkvar=self.chkvar_all: self.planets_toggle(name,chkvar))\r\n self.chk_all.config(offvalue = 0, onvalue=1, bg=PAGE_BG1, fg=\"#F0F0F0\", font=(\"Arial\", 8, \"bold\"), activebackground=PAGE_BG_2, activeforeground=\"#F0F0F0\", selectcolor=PAGE_BG_2,justify=\"center\",borderwidth=3)\r\n self.chk_all.grid(row=0, column=1,sticky='nw', padx=(4,4), pady=(3,0))\r\n\r\n sel_col= \"#015E70\" \r\n COL_1 = \"#E3E3E3\"\r\n style_ = ttk.Style()\r\n try:\r\n style_.theme_use(\"new_style\") \r\n except:\r\n style_ = ttk.Style()\r\n configure_ = dict(foreground = COL_1, selectbackground = sel_col,fieldbackground = sel_col,background = sel_col, arrowcolor = COL_1)\r\n style_.theme_create(\"new_style\", parent=\"alt\", settings = {\"TCombobox\":{\"configure\":configure_}} ) \r\n style_.theme_use(\"new_style\") \r\n\r\n self.combo_ = ttk.Combobox(self.page2, width = 24) \r\n self.combo_.grid(row=0, column=1,sticky=\"w\", padx=10, pady=(0,5)) \r\n self.combo_['values'] = (\"Extra off\", \"Planet projection points\", \"Planet meridian\", \"Planet parallel\", \r\n \"Planet prop. horizon\", \"All ecliptic points\", \"Zodiac ecliptic points\",\"Zodiac symbols\",\r\n \"Horizon projections off\",\r\n \"Ecliptic scale\", \"Equator scale\", \"Horizon scale\", \"Prime vertical scale\")\r\n self.combo_.set(\"Hide/show\")\r\n self.combo_[\"state\"] = \"readonly\"\r\n \r\n self.combo_.bind('<>',self.combobox_callback) \r\n self.combo_.bind('',self.combobox_callback)\r\n\r\n #SpinBox input range Validation\r\n self.vc = (self.parent.register(self.input_validate), \"%P\", \"%d\", \"%W\")\r\n\r\n self.time_shift_buttons()\r\n self.anim_buttons()\r\n self.geo_buttons()\r\n\r\n self.plot = astro3D(self.parent,self.page_plot, planets_data, geo_latitude, data=data) \r\n\r\n def time_shift_buttons(self): \r\n sel_col= \"#015E70\"\r\n COL=PAGE_BG1\r\n COL_F=\"#F6F6F6\"\r\n FONT_ = (\"Tahoma\", 8, \"bold\")\r\n FONT_2 = (\"Arial\", 9, \"bold\")\r\n COL_1 = \"#F6F6F6\"\r\n\r\n f_args_p={\"side\":\"top\",\"fill\":\"both\",\"expand\":False,\"anchor\":\"nw\"}\r\n\r\n self.fr = Frame(self.page3,bg=PAGE_BG1,borderwidth=0)\r\n\r\n self.f0 = Frame(self.fr,bg=PAGE_BG1,borderwidth=4)\r\n self.f0.pack(side=\"left\",fill=\"both\",expand=False,anchor=\"nw\", padx=0, pady=0) \r\n self.f1 = Frame(self.fr,bg=PAGE_BG1,borderwidth=4, padx=0, pady=0)\r\n self.f1.pack(**f_args_p)\r\n self.f2 = Frame(self.fr,bg=PAGE_BG1,borderwidth=4, padx=0, pady=0)\r\n self.f2.pack(**f_args_p)\r\n\r\n self.rad_var = StringVar(None, \"Days\")\r\n \r\n w=6; indic=0;\r\n ipx=4; ipy=1; py=0;\r\n args={\"width\":w,\"bg\":COL,\"fg\":COL_F,\"font\":FONT_,\"command\":None,\"variable\":self.rad_var,\"indicator\":indic, \"selectcolor\":sel_col}\r\n args_p={\"fill\":None, \"ipadx\":ipx, \"ipady\":ipy, \"pady\":py,\"side\":\"left\",\"anchor\":\"nw\"}\r\n self.rad_h=Radiobutton(self.f1, text = \"Hours\", value=\"Hours\", **args); self.rad_h.pack(**args_p) \r\n self.rad_m=Radiobutton(self.f1, text = \"Minutes\", value=\"Minutes\", **args); self.rad_m.pack(**args_p) \r\n self.rad_s=Radiobutton(self.f1, text = \"Seconds\", value=\"Seconds\", **args); self.rad_s.pack(**args_p)\r\n\r\n self.spbox_var = IntVar()\r\n self.sp = Spinbox(self.f2, from_=1, to=60,increment=1,textvariable=self.spbox_var,bg=sel_col,width=7,fg=COL_1,font=(\"Tahoma\", 11, \"normal\"),justify=\"center\",buttonbackground=COL)\r\n self.sp.pack(side=\"left\",anchor=\"nw\" ,padx=(0,7), pady=(2,0)) \r\n #SpinBox range Validation\r\n self.sp.config(validate =\"key\", validatecommand = self.vc )\r\n\r\n unit=\"Minutes\"\r\n self.spbox_var.set(10)\r\n self.rad_var.set(unit)\r\n\r\n F_arr=(\"Helvetica\", 14, \"bold\")\r\n ipx=4;ipy=0;py=(0,0);px=3;\r\n ipy=0\r\n arr_p={\"fill\":None, \"ipadx\":ipx, \"ipady\":ipy, \"padx\":px, \"pady\":py,\"side\":\"left\",\"anchor\":\"center\"} \r\n self.bt_arrowL = Button(self.f2,text=\"◄\",command=lambda e=None, Δt=None:self.prev_tk(e, Δt),width=4,bg=COL,fg=COL_F,font=FONT_2)\r\n self.bt_arrowR = Button(self.f2,text=\"►\",command=lambda e=None, Δt=None:self.next_tk(e, Δt),width=4,bg=COL,fg=COL_F,font=FONT_2)\r\n self.bt_arrowL.pack(**arr_p)\r\n self.bt_arrowR.pack(**arr_p)\r\n\r\n self.fr.grid(row=0, column=18,sticky=\"w\", padx=10, pady=(0,0),)\r\n self.bt_arrowL.bind(\"\", lambda e=None, Δt=1:self.prev_tk(e, r_click=Δt))\r\n self.bt_arrowR.bind(\"\", lambda e=None, Δt=1:self.next_tk(e, r_click=Δt)) \r\n\r\n self.rad_h.bind(\"\", lambda e=None, s=\"Hours\":self.radio_(e, s))\r\n self.rad_m.bind(\"\", lambda e=None, s=\"Minutes\":self.radio_(e, s))\r\n self.rad_s.bind(\"\", lambda e=None, s=\"Seconds\":self.radio_(e, s))\r\n\r\n def anim_buttons(self):\r\n sel_col= \"#015E70\"\r\n COL=PAGE_BG1\r\n COL_F=\"#F6F6F6\"\r\n FONT_ = (\"Tahoma\", 8, \"bold\")\r\n FONT_2 = (\"Arial\", 8, \"bold\")\r\n COL_1 = \"#F6F6F6\"\r\n\r\n self.bt_anim = Button(self.page4,text=\"Animation\",command=self.anim_tk,width=12,bg=COL,fg=COL_F,font=FONT_2)#\r\n self.bt_anim.grid(row=0, column=0, columnspan=2, sticky=\"w\", padx=(11, 8), pady=(3,5),ipady=0) \r\n self.spbox_var_st = IntVar()\r\n self.sp_st = Spinbox(self.page4, from_=1, to=30,increment=1,textvariable=self.spbox_var_st, bg=sel_col,width=4,fg=COL_1,font=(\"Tahoma\", 11, \"normal\"),justify=\"center\",buttonbackground=COL)\r\n self.sp_st.grid(row=1, column=0,sticky=\"w\", padx=(11, 0), pady=(6,5)) \r\n self.spbox_var_st.set(10)\r\n\r\n self.lbl2 = Label(self.page4, text=\"steps\", justify=\"center\", bg=PAGE_BG1, fg=\"#F0F0F0\", font=(\"Arial\",8, \"bold\"))\r\n self.lbl2.grid(row=1, column=1, sticky=\"w\", padx=(0,0), pady=4)\r\n\r\n #SpinBox range Validation\r\n self.sp_st.config(validate =\"key\", validatecommand = self.vc)\r\n self.bt_anim.bind(\"\", self.anim_back_tk)\r\n\r\n def geo_buttons(self):\r\n sel_col= \"#015E70\"\r\n COL=PAGE_BG1\r\n COL_F=\"#F6F6F6\"\r\n FONT_ = (\"Tahoma\", 8, \"bold\")\r\n FONT_2 = (\"Arial\", 8, \"bold\")\r\n COL_1 = \"#F6F6F6\"\r\n FONT_E = (\"Arial\", 10, \"normal\")\r\n\r\n self.bt_ = Button(self.page5,text=\"New for current time\",command=self.new_,width=21,bg=COL,fg=COL_F,font=FONT_2)#\r\n self.bt_.grid(row=0, column=0, columnspan=4, sticky=\"w\", padx=(10, 8), pady=(3,5),ipady=0) \r\n\r\n self.long_v = StringVar() ; self.lat_v = StringVar() \r\n self.lat_v.set(\"50.07\"); self.long_v.set(\"19.90\")\r\n #self.lat_v.set(\"40.67\") ; self.long_v.set(\"-73.95\")\r\n\r\n self.inp_lat = Entry(self.page5, textvar = self.lat_v, width=6, font = FONT_E, justify=\"center\", bg=sel_col, fg=COL_1, relief=\"sunken\")\r\n self.inp_lat.grid(row=1, column=0, columnspan=1, sticky=\"w\", padx=(10, 0), pady=(7,2),ipady=0) \r\n self.inp_long = Entry(self.page5, textvar = self.long_v, width=7,font = FONT_E, justify=\"center\", bg=sel_col, fg=COL_1, relief=\"sunken\")\r\n self.inp_long.grid(row=1, column=2, columnspan=1, sticky=\"w\", padx=(5, 0), pady=(7,2),ipady=0)\r\n\r\n lbl3 = Label(self.page5, text=\"lat\", justify=\"left\", width=1, bg=PAGE_BG1, fg=\"#F0F0F0\", font=(\"Arial\",8, \"bold\"))\r\n lbl3.grid(row=1, column=1, sticky=\"w\", padx=(2,0), pady=(7,2))\r\n lbl4 = Label(self.page5, text=\"lon\", justify=\"left\", width=2, bg=PAGE_BG1, fg=\"#F0F0F0\", font=(\"Arial\",8, \"bold\"))\r\n lbl4.grid(row=1, column=3, sticky=\"w\", padx=(0,6), pady=(7,2))\r\n\r\n self.inp_lat.config(validate =\"key\", validatecommand = self.vc)\r\n self.inp_long.config(validate =\"key\", validatecommand = self.vc)\r\n\r\n self.bt_.bind(\"\", self.reset_lat)\r\n\r\n\r\n\r\n def input_validate(self, input_, action, name):\r\n if input_:\r\n if input_==\"-\":\r\n return True\r\n elif input_==\".\" and \"entry\" in name:\r\n return True\r\n elif input_==\".\" and \"entry\" not in name:\r\n return False\r\n if \"entry2\" in name:\r\n try:\r\n input_=float(input_)\r\n if not self.test_float(input_, 2):\r\n return False\r\n return True\r\n except ValueError:\r\n return False\r\n elif \"entry\" in name:\r\n try:\r\n input_=float(input_)\r\n if not self.test_float(input_, 1):\r\n return False\r\n return True\r\n except ValueError:\r\n return False\r\n else: #Spinbox \r\n try:\r\n int(input_)\r\n if len(str(input_))>6:\r\n return False\r\n return True\r\n except ValueError:\r\n return False\r\n elif action==\"0\": #validatecommand + backspace\r\n return True\r\n else:\r\n return False\r\n\r\n def test_float(self, f, inp):\r\n s=str(f)\r\n s=s.replace(\"-\",\"\")\r\n if \".\" in s:\r\n arr=s.split(\".\")\r\n if len(arr)>2: \r\n return False #only 1 dot\r\n if len(arr[1])>2: \r\n return False # only 2 decimal digits\r\n\r\n lim = 2 if inp==1 else 3 \r\n if len(arr[0])>lim:\r\n return False\r\n else:\r\n lim2= 90 if inp==1 else 180 \r\n if abs(int(arr[0]))>lim2:\r\n return False\r\n return True \r\n else:\r\n lim = 2 if inp==1 else 3 \r\n if len(s)>lim:\r\n return False\r\n else:\r\n lim2= 90 if inp==1 else 180 \r\n if abs(int(arr[0]))>lim2:\r\n return False\r\n return True\r\n return True\r\n\r\n\r\n def input_validate0(self, input_, action):\r\n if input_:\r\n if input_==\"-\":\r\n return True\r\n\r\n try:\r\n int(input_)\r\n if len(input_)>4:\r\n return False\r\n return True\r\n except ValueError:\r\n return False\r\n elif action==\"0\": #validatecommand + backspace\r\n return True\r\n else:\r\n return False\r\n\r\n def radio_(self, event=None, s=None):\r\n val =self.rad_var.get()\r\n a1('',val , s)\r\n if s==\"Hours\":\r\n self.spbox_var.set(1)\r\n elif s==\"Minutes\":\r\n self.spbox_var.set(10)\r\n elif s==\"Seconds\":\r\n self.spbox_var.set(30)\r\n self.rad_var.set(s)\r\n\r\n def planets_toggle(self,name,id_):\r\n if name!=\"All\":\r\n obj=self.plot.planets_ids[name]\r\n if id_.get()==1:\r\n obj[\"id_pl\"].set_visible(True)\r\n obj[\"id_t\"].set_visible(True)\r\n if name==\"Node_N\":\r\n self.plot.planets_ids[\"Node_S\"][\"id_pl\"].set_visible(True)\r\n self.plot.planets_ids[\"Node_S\"][\"id_t\"].set_visible(True)\r\n elif id_.get()==0:\r\n for el_ in obj:\r\n obj[el_].set_visible(False)\r\n if name==\"Node_N\":\r\n for el_ in self.plot.planets_ids[\"Node_S\"]:\r\n self.plot.planets_ids[\"Node_S\"][el_].set_visible(False)\r\n self.chkvar_all.set(0)\r\n\r\n elif name==\"All\":\r\n if id_.get()==1:\r\n for id__ in self.chkvars:\r\n id__.set(1)\r\n for pl in self.plot.planets_ids:\r\n obj=self.plot.planets_ids[pl]\r\n obj[\"id_pl\"].set_visible(True)\r\n obj[\"id_t\"].set_visible(True)\r\n \r\n elif id_.get()==0:\r\n for id__ in self.chkvars:\r\n id__.set(0)\r\n for pl in self.plot.planets_ids:\r\n obj=self.plot.planets_ids[pl]\r\n for el_ in obj:\r\n obj[el_].set_visible(False)\r\n self.plot.canvas.draw_idle()\r\n\r\n def combobox_callback(self,event):\r\n str_ = self.combo_.get()\r\n if str_==\"Planet projection points\": \r\n for name in c.planets2:\r\n if name in [ \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=self.plot.planets_ids[name][\"id_pl\"].get_visible() \r\n if not isVisible:continue\r\n for el in [\"id_pt_Eq\", \"id_pt_Hor\", \"id_pt_Ecl\"]:\r\n obj[el].set_visible(isVisible)\r\n \r\n elif str_==\"Planet meridian\":\r\n for name in c.planets2:\r\n if name in [ \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=obj[\"id_pl\"].get_visible()\r\n if not isVisible:continue\r\n obj[\"id_Mer\"].set_visible(isVisible)\r\n\r\n elif str_==\"Planet parallel\":\r\n for name in c.planets2:\r\n if name in [\"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=obj[\"id_pl\"].get_visible()\r\n if not isVisible:continue\r\n obj[\"id_Par\"].set_visible(isVisible)\r\n\r\n elif str_==\"Planet prop. horizon\":\r\n for name in c.planets2:\r\n if name in [ \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=obj[\"id_pl\"].get_visible()\r\n if not isVisible:continue\r\n obj[\"id_proj_hor\"].set_visible(isVisible)\r\n\r\n elif str_==\"Planet meridian natal\":\r\n for name in c.planets2:\r\n if name in [\"Node_N\", \"Node_S\", \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=obj[\"id_pl\"].get_visible()\r\n if not isVisible:continue\r\n obj[\"id_Mer_nat\"].set_visible(isVisible)\r\n\r\n elif str_==\"Planet prop. horizon natal\":\r\n for name in c.planets2:\r\n if name in [\"Node_N\", \"Node_S\", \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n isVisible=obj[\"id_pl\"].get_visible()\r\n if not isVisible:continue \r\n obj[\"id_proj_hor_nat\"].set_visible(isVisible)\r\n\r\n elif str_==\"Houses ecliptic points\": \r\n self.toggle_houses_p()\r\n\r\n elif str_==\"Zodiac ecliptic points\": \r\n self.toggle_ecl_p()\r\n\r\n elif str_==\"Zodiac symbols\":\r\n self.toggle_ecl_sym()\r\n\r\n elif str_==\"All ecliptic points\":\r\n self.toggle_ecl_p()\r\n self.toggle_ecl_sym()\r\n self.toggle_houses_p()\r\n\r\n elif str_==\"Celestial equator points\":\r\n isVisible=self.plot.armc_id.get_visible() \r\n self.plot.armc_id.set_visible(not isVisible)\r\n\r\n elif str_==\"Asc, MC, ARMC...\":\r\n isVisible=self.plot.asc_id.get_visible()\r\n for id_ in [self.plot.armc_id, self.plot.mc_id, self.plot.asc_id, self.plot.asc_id2, self.plot.poleN_id, self.plot.poleS_id]:\r\n id_.set_visible(not isVisible)\r\n\r\n elif str_==\"Horizon projections off\":\r\n for name in self.plot.planets_ids:\r\n if name in [\"Node_N\", \"Node_S\", \"Asc\", \"MC\"]:continue\r\n obj=self.plot.planets_ids[name]\r\n for el in [\"id_l_ho\", \"id_pt_Hor\"]:\r\n obj[el].set_visible(False)\r\n\r\n\r\n elif str_==\"Ecliptic scale\":\r\n self.plot.toggle_ecliptic_scale()\r\n\r\n elif str_==\"Equator scale\":\r\n self.plot.toggle_equator_scale()\r\n\r\n elif str_==\"Horizon scale\":\r\n self.plot.toggle_horizon_scale()\r\n\r\n elif str_==\"Prime vertical scale\":\r\n self.plot.toggle_prime_vert_scale()\r\n\r\n elif str_==\"Extra off\":\r\n self.plot.extra_off()\r\n\r\n self.plot.canvas.draw_idle()\r\n\r\n\r\n def toggle_houses_p(self):\r\n isVisible=self.plot.houses_ids[1][\"dot\"].get_visible()\r\n for i in range(1,13):\r\n self.plot.houses_ids[i][\"dot\"].set_visible(not isVisible)\r\n\r\n def toggle_ecl_p(self):\r\n isVisible=self.plot.zodiac_ids[1][\"dot\"].get_visible()\r\n for i in range(1,13):\r\n self.plot.zodiac_ids[i][\"dot\"].set_visible(not isVisible)\r\n\r\n def toggle_ecl_sym(self):\r\n isVisible=self.plot.zodiac_ids[1][\"text\"].get_visible()\r\n for i in range(1,13):\r\n self.plot.zodiac_ids[i][\"text\"].set_visible(not isVisible)\r\n\r\n def prev_tk(self, event, r_click=None):\r\n Δt = self.spbox_var.get()\r\n val = self.rad_var.get()\r\n if r_click!=None:\r\n Δt=r_click\r\n if val ==\"Seconds\":\r\n self.plot.plot_next_prev(seconds=-Δt)\r\n elif val ==\"Minutes\":\r\n self.plot.plot_next_prev(minutes=-Δt)\r\n elif val ==\"Hours\":\r\n self.plot.plot_next_prev(hours=-Δt)\r\n\r\n def next_tk(self, event, r_click=None):\r\n Δt = self.spbox_var.get()\r\n val = self.rad_var.get()\r\n if r_click!=None:\r\n Δt=r_click\r\n if val ==\"Seconds\":\r\n self.plot.plot_next_prev(seconds=Δt)\r\n elif val ==\"Minutes\":\r\n self.plot.plot_next_prev(minutes=Δt)\r\n elif val ==\"Hours\":\r\n self.plot.plot_next_prev(hours=Δt)\r\n\r\n def anim_tk(self):\r\n s=0; m=0; h=0\r\n Δt=self.spbox_var.get()\r\n val=self.rad_var.get()\r\n if val==\"Seconds\":\r\n s=Δt\r\n elif val==\"Minutes\":\r\n m=Δt\r\n elif val==\"Hours\":\r\n h=Δt\r\n steps=self.spbox_var_st.get()\r\n for i in range(steps):\r\n self.parent.after(i*300, lambda s=s, m=m, h=h: self.plot.plot_next_prev(seconds=s, minutes=m, hours=h))\r\n self.plot.canvas.draw_idle() \r\n return\r\n \r\n def anim_back_tk(self, event):\r\n s=0; m=0; h=0\r\n Δt=self.spbox_var.get()\r\n val=self.rad_var.get()\r\n Δt=-Δt\r\n if val==\"Seconds\":\r\n s=Δt\r\n elif val==\"Minutes\":\r\n m=Δt\r\n elif val==\"Hours\":\r\n h=Δt\r\n steps=self.spbox_var_st.get()\r\n for i in range(steps):\r\n self.parent.after(i*300, lambda s=s, m=m, h=h: self.plot.plot_next_prev(seconds=s, minutes=m, hours=h))\r\n self.plot.canvas.draw_idle() \r\n return\r\n\r\n def reset_lat(self, time_):\r\n self.lat_v.set(\"0\")\r\n self.long_v.set(\"0\")\r\n\r\n def new_(self):\r\n Δt=self.spbox_var.get()\r\n lat_ = self.lat_v.get()\r\n lon_ = self.long_v.get()\r\n if lat_==\"\" or lon_==\"\":return\r\n lat_ = float(lat_)\r\n lon_ = float(lon_)\r\n if abs(lat_)>90 or abs(lon_)>180:\r\n messagebox.showwarning(\"Wrong value\", \"Values must be within intervals:\\nlatitude:\\n[0, 90] for North and [0, -90] for South\\nlongitude:\\n[0, 180] for East and [0, -180] for West\")\r\n return\r\n self.plot.extra_off()\r\n\r\n tz_=None\r\n t_obj = get_time_now(tz_=tz_)\r\n date_utc=t_obj[\"date_utc\"]\r\n time_utc=t_obj[\"time_utc\"]\r\n dataNow={'n': 'Planets', 'ln': 'positions'}\r\n dataNow[\"d_utc\"]=date_utc\r\n dataNow[\"t_utc\"]=time_utc\r\n dataNow[\"timestamp\"]=t_obj[\"timestamp\"]\r\n dataNow[\"lat\"]=lat_\r\n dataNow[\"lon\"]=lon_\r\n dataNow[\"tz\"]=tz_\r\n dataNow[\"date_loc\"]=t_obj[\"date_loc\"]\r\n dataNow[\"time_loc\"]=t_obj[\"time_loc\"]\r\n planets_data = calc_.get_planets_data(date_utc,time_utc,lat_,lon_,trueNode=True)\r\n self.plot.new_now(planets_data, lat_, data=dataNow)\r\n\r\n\r\n def keypressed(self,event):\r\n if event.keysym==\"F1\":\r\n self.prev_tk(event)\r\n elif event.keysym==\"F2\":\r\n self.next_tk(event)\r\n elif event.keysym==\"1\":\r\n self.prev_tk(event)\r\n elif event.keysym==\"2\":\r\n self.next_tk(event)\r\n elif (event.state==12 and event.keysym==\"a\") or (event.state==12 and event.keysym==\"d\"): # 12=ctrl\r\n time_obj={\"minutes\":10}\r\n frames=self.spbox_var_st.get()\r\n Δt=self.spbox_var.get()\r\n val=self.rad_var.get()\r\n if val==\"Hours\":\r\n time_obj={\"hours\":Δt}\r\n elif val==\"Minutes\":\r\n time_obj={\"minutes\":Δt}\r\n if event.keysym==\"a\":\r\n time_obj.update({\"format\":\"gif\"})\r\n elif event.keysym==\"d\":\r\n time_obj.update({\"format\":\"mp4\"})\r\n self.plot.save_animation(frames=frames, **time_obj)\r\n\r\n\r\n def close_dialog(self):\r\n self.parent.wm_attributes(\"-disabled\", False)\r\n self.parent.deiconify() \r\n self.parent.destroy()\r\n\r\n def Exit(self):\r\n self.parent.destroy() \r\n\r\n\r\n\r\ndef draw_chart3D_now(data=None):\r\n if data==None:\r\n latitude = 40.673\r\n longitude=-73.945\r\n tz_ = \"America/New_York\"\r\n date_=get_time_now(seconds=None) \r\n date_utc = date_[\"date_utc\"]\r\n time_utc = date_[\"time_utc\"]\r\n dataNow={'n': 'Planets', 'ln': 'positions'}\r\n dataNow[\"d_utc\"]=date_[\"date_utc\"]\r\n dataNow[\"t_utc\"]=date_[\"time_utc\"]\r\n dataNow[\"timestamp\"]=date_[\"timestamp\"]\r\n dataNow[\"lat\"]=latitude\r\n dataNow[\"lon\"]=longitude\r\n dataNow[\"tz\"]=tz_\r\n d= date_[\"date_loc\"].split(\"-\")\r\n d=list(reversed(d))\r\n dataNow[\"d\"]=d \r\n t= date_[\"time_loc\"].split(\":\") \r\n dataNow[\"t\"]=t\r\n trueNode=True\r\n geo_latitude=latitude\r\n planets_data = calc_.get_planets_data(date_utc,time_utc,latitude,longitude,trueNode=trueNode)\r\n ε=calc_.ε\r\n dataNow[\"obliquity\"]=ε \r\n dataNow[\"trueNode\"]=trueNode\r\n else:\r\n date_utc = data[\"d_utc\"]\r\n time_utc = data[\"t_utc\"]\r\n timestamp = data[\"timestamp\"]\r\n latitude = float(data[\"lat\"])\r\n longitude = float(data[\"lon\"])\r\n geo_latitude=latitude\r\n trueNode=data[\"trueNode\"]\r\n planets_data = calc_.get_planets_data(date_utc,time_utc,latitude,longitude,trueNode=trueNode)\r\n dataNow=data\r\n GUI = GUI_astro3D(root, planets_data, geo_latitude, data=dataNow)\r\n return GUI\r\n\r\n\r\n\r\nif __name__ == \"__main__\":\r\n trueNode=True\r\n #print('sys.argv',len(sys.argv),'\\n',sys.argv);\r\n if len(sys.argv)>1:\r\n if len(sys.argv)>2 and sys.argv[2]==\"standalone\": #standalone command line\r\n data=eval(sys.argv[1])\r\n cmd_obj={\"date_utc\":data[\"d_utc\"], \"time_utc\": data[\"t_utc\"], \"lat\" : data[\"lat\"], \"lon\": data[\"lon\"],\"data\":data, \"trueNode\" : data[\"trueNode\"],\"timestamp\":data[\"timestamp\"]}\r\n else:\r\n json_object=sys.argv[1]\r\n cmd_obj= json.loads(json_object) \r\n\r\n date_utc=cmd_obj[\"date_utc\"]\r\n time_utc=cmd_obj[\"time_utc\"]\r\n latitude=float(cmd_obj[\"lat\"])\r\n longitude=float(cmd_obj[\"lon\"])\r\n trueNode=cmd_obj[\"trueNode\"] \r\n data=cmd_obj[\"data\"] \r\n print(\"command line\") \r\n else:\r\n data=None \r\n\r\n root = Tk()\r\n calc_=calc_for_3D()\r\n GUI = draw_chart3D_now(data)\r\n #keyboard.add_hotkey('esc', GUI.Exit)\r\n mainloop()\r\n\r\n\r\n","repo_name":"Drozdman-1/Astronomia3D","sub_path":"astronomia3D.pyw","file_name":"astronomia3D.pyw","file_ext":"pyw","file_size_in_byte":121480,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31367333581","text":"from vispy import app, scene\r\nimport numpy as np\r\nimport math\r\nfrom collections import deque\r\n\r\nfrom PySide6.QtCore import Qt, Slot, QSize\r\nfrom PySide6.QtGui import QPainter, QFont, QScreen\r\nfrom PySide6.QtWidgets import QWidget, QFrame, QVBoxLayout, QPushButton\r\n# from PySide6.QtOpenGLWidgets import QOpenGLWidget\r\nfrom PySide6.QtUiTools import QUiLoader\r\nfrom PySide6.QtWidgets import QApplication\r\nfrom PySide6.QtDesigner import QPyDesignerCustomWidgetCollection\r\n\r\nclass MyPlotLabel(QWidget):\r\n def __init__(self, p_label, parent =None):\r\n super().__init__(parent)\r\n self.p_label = p_label\r\n \r\n def paintEvent(self, event):\r\n painter = QPainter(self)\r\n \r\n # cur_viewport = painter.viewport()\r\n # painter.setViewport(0,0,35,cur_viewport.height())\r\n\r\n # color = QColor(222, 0, 0)\r\n # painter.fillRect(painter.viewport(), color)\r\n\r\n # color = QColor(0, 0, 0)\r\n # color.setNamedColor('#d4d4d4')\r\n # painter.setPen(color)\r\n # painter.setBrush(QColor(0, 200, 0))\r\n # painter.drawRoundedRect(0, 0, 35, painter.viewport().height(), 5, 5)\r\n\r\n # clip = QRect(2, 2, painter.viewport().width() - 4, painter.viewport().height() - 4);\r\n \r\n painter.setPen(Qt.white)\r\n # painter.translate(10, 55)\r\n painter.translate(painter.viewport().width() - 4, painter.viewport().height() - 4)\r\n painter.rotate(-90)\r\n # painter.drawText(clip, self.p_label)\r\n bold = QFont();\r\n bold.setBold(True);\r\n painter.setFont(bold);\r\n painter.drawText(0, 0, self.p_label.upper())\r\n painter.end()\r\n\r\n# class PlotWidget(QOpenGLWidget):\r\nclass PlotWidget(QWidget):\r\n def __init__(self, controller, comp_name, odr, time_window, n_curves=1, parent=None, p_id = 0):\r\n super().__init__(parent)\r\n self.parent = parent\r\n self.controller = controller\r\n self.controller.sig_logging.connect(self.s_is_logging)\r\n \r\n self.is_docked = True\r\n self.app_qt = self.controller.qt_app\r\n\r\n self.p_id = p_id\r\n self.comp_name = comp_name\r\n \r\n self.canvas = scene.SceneCanvas(bgcolor=\"#1b1d23\", keys='interactive', vsync=True)\r\n self.canvas.native.setMinimumSize(QSize(300, 150))\r\n \r\n QPyDesignerCustomWidgetCollection.registerCustomWidget(PlotWidget, module=\"PlotWidget\")\r\n loader = QUiLoader()\r\n plot_widget = loader.load(\"HSD_GUI\\GUI\\plot_widget.ui\", parent)\r\n title_frame = plot_widget.frame_plot.findChild(QFrame,\"frame_title\")\r\n contents_frame = plot_widget.frame_plot.findChild(QFrame,\"frame_contents\")\r\n pushButton_pop_out = title_frame.findChild(QPushButton, \"pushButton_pop_out\")\r\n pushButton_pop_out.clicked.connect(self.clicked_pop_out_button)\r\n \r\n #Main layout\r\n main_layout = QVBoxLayout()\r\n self.setLayout(main_layout)\r\n main_layout.addWidget(plot_widget)\r\n\r\n title_frame.layout().addWidget(MyPlotLabel(\"{}\".format(self.comp_name)))\r\n \r\n contents_frame.layout().addWidget(self.canvas.native)\r\n \r\n grid = self.canvas.central_widget.add_grid(spacing=0)\r\n self.viewbox = grid.add_view(row=0, col=1, camera='panzoom')\r\n\r\n # add the axes\r\n x_axis = scene.AxisWidget(orientation='bottom')\r\n x_axis.stretch = (1, 0.1)\r\n x_axis.axis.axis_width = x_axis.axis.tick_width = 1\r\n x_axis.axis.text_color = '#1b1d23' #same as background (hide text)\r\n x_axis.axis.axis_color = '#d2d2d2'\r\n x_axis.axis.tick_color = '#666666'\r\n grid.add_widget(x_axis, row=1, col=1)\r\n x_axis.link_view(self.viewbox)\r\n y_axis = scene.AxisWidget(orientation='left')\r\n y_axis.stretch = (0.1, 1)\r\n y_axis.axis.axis_width = y_axis.axis.tick_width = 1\r\n y_axis.axis.text_color = '#d2d2d2'\r\n y_axis.axis.axis_color = '#d2d2d2'\r\n y_axis.axis.tick_color = '#666666'\r\n grid.add_widget(y_axis, row=0, col=0)\r\n y_axis.link_view(self.viewbox)\r\n\r\n self.positions = dict()\r\n self._data = dict() # dict of queues\r\n\r\n self.update_plots_ui(odr, time_window, n_curves)\r\n\r\n self.updateTimer = app.Timer() # Setting up the timer to update UI\r\n self.updateTimer.connect(self.update_plot)\r\n\r\n def update_plots_ui(self, odr, time_window, n_curves): \r\n \r\n self.positions = dict()\r\n self._data = dict() # dict of queues\r\n self.curves = dict()\r\n\r\n self.n_curves = n_curves\r\n\r\n self.downsampling_factor = 1\r\n self.samples_cnt = int(odr * time_window)\r\n digits = int(math.log10(self.samples_cnt))+1\r\n \r\n #Downsampling based on sensor ODR and time_window (Max plottable point in a time_window = 9999)\r\n if digits > 4:\r\n self.downsampling_factor = 10**(digits-4)\r\n self.N = int(self.samples_cnt/self.downsampling_factor)\r\n\r\n self.color = dict()\r\n # color arrays\r\n for i in range(n_curves):\r\n self.color[i] = np.ones((self.N, 4), dtype=np.float32)\r\n self.color[i][:, 0] = np.linspace((i/self.n_curves), ((i+1)/self.n_curves), self.N)\r\n self.color[i][:, 1] = self.color[i][::-1, 0]\r\n\r\n # add a line plot inside the viewbox\r\n for i in range(self.n_curves):\r\n self._data[i] = deque(maxlen=self.N)\r\n self.positions[i] = np.zeros((self.N, 2), dtype='f')#self._dtype_from_data_type(data_type))\r\n x_lim = [0., self.N]\r\n self.positions[i][:, 0] = np.linspace(x_lim[0], x_lim[1], self.N)\r\n\r\n self.curves[i] = scene.Line(self.positions[i], self.color[i], method='gl', parent=self.viewbox.scene)\r\n #TODO TEST\r\n self.app_qt.processEvents()\r\n\r\n #TODO use FS\r\n self.min_d = 0\r\n self.max_d = 10 \r\n self.viewbox.camera.set_range(x = (0,int(self.N)), y =(self.min_d,self.max_d))\r\n\r\n self.update()\r\n\r\n @Slot()\r\n def clicked_pop_out_button(self):\r\n if self.is_docked:\r\n self.pop_out_widget()\r\n self.is_docked = False\r\n else:\r\n self.pop_in_widget()\r\n self.is_docked = True\r\n\r\n @Slot(bool)\r\n def s_is_logging(self, state: bool):\r\n print(\"sensor\", self.comp_name, \"is_logging\", state)\r\n if state:\r\n self.updateTimer.start(0)\r\n else:\r\n self.updateTimer.stop()\r\n\r\n def reset(self):\r\n pass\r\n # self._data = deque(maxlen=self._window_size)\r\n\r\n # for c in self._curves:\r\n # self._plot.removeItem(c)\r\n\r\n # self._curves = []\r\n\r\n # for i in range(self.n_curves):\r\n # self._curves.append(self._plot.plot(pen=pg.mkPen(self.colors[i % len(self.colors)], width=1),connect='auto'))\r\n # self._curves[i].setDownsampling(auto=True, method='peak')\r\n # self._curves[i].setClipToView(True)\r\n\r\n def update_plot(self, ev): \r\n if len(self._data[0]) > 0:\r\n max_d_now = []\r\n min_d_now = []\r\n for i in range(self.n_curves):\r\n if len(self._data[i]) < self.N:\r\n a = self.positions[i][:(self.N - len(self._data[i])),1]\r\n self.positions[i][:, 1] = np.append(a,np.array(self._data[i]))\r\n else: \r\n self.positions[i][:, 1] = np.array(self._data[i])\r\n\r\n self.curves[i].set_data(pos=self.positions[i], color=self.color[i])\r\n \r\n #TODO to be improved\r\n min_d_now.append(np.amin(self.positions[i][:, 1]))\r\n max_d_now.append(np.amax(self.positions[i][:, 1]))\r\n \r\n max_data = np.amax(max_d_now)\r\n min_data = np.amin(min_d_now)\r\n if min_data < self.min_d:\r\n self.viewbox.camera.set_range(x=(0,self.N),y=(min_data, self.max_d))\r\n self.min_d = min_data\r\n if max_data > self.max_d:\r\n self.viewbox.camera.set_range(x=(0,self.N),y=(self.min_d, max_data))\r\n self.max_d = max_data\r\n \r\n self.app_qt.processEvents()\r\n\r\n def add_data(self, data):\r\n # if len(data) > 0:\r\n for i in range(self.n_curves):\r\n # Downsampling for PLOTs\r\n self._data[i].extend(data[i][0:data[i].size:self.downsampling_factor])\r\n\r\n def closeEvent(self, event):\r\n event.accept()\r\n\r\n def closeEvent(self, event):\r\n self.pop_in_widget()\r\n self.is_docked = True\r\n\r\n def pop_out_widget(self):\r\n self.setWindowFlags(Qt.Dialog)\r\n center = QScreen.availableGeometry(QApplication.primaryScreen()).center()\r\n geo = self.frameGeometry()\r\n geo.moveCenter(center)\r\n self.move(geo.topLeft())\r\n self.show()\r\n\r\n def pop_in_widget(self):\r\n self.setWindowFlags(Qt.Widget)\r\n self.parent.layout().insertWidget(self.p_id, self)","repo_name":"dmelpi/eai-polimi","sub_path":"contrib/DATA_Analysis/Vespucci/HSD_GUI/PlotWidget.py","file_name":"PlotWidget.py","file_ext":"py","file_size_in_byte":9019,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"79"} +{"seq_id":"8081493841","text":"from lm import LanguageModel\nfrom text_preprocessing import TextPrepocessing\n\nimport re\nimport os\n\n# Necessary for unit-tests\ndef get_input(text):\n return input(text)\n\n# Necessary for unit-tests\ndef main():\n pass\n\nclass Main():\n program_is_over = False\n\n @property\n def language_model(self):\n return self.__language_model\n @language_model.setter\n def language_model(self, language_model):\n self.__language_model = language_model\n\n def __init__(self):\n print(\"Main instance created\")\n\n # Check entered int value\n def enter_int_number(self):\n while True:\n user_input = get_input(\"Please, enter an intager number:\\n\")\n try:\n value = int(user_input)\n if (value > 0):\n return value\n else:\n print(\"Entered number is negative\")\n except ValueError:\n print(\"Entered incorrect input\")\n\n # Present information about functionality of the program\n def present_options(self):\n print(\"\\nEnter the number of row for desired operation.\")\n print(\"For generating text, first of all you should create language model. \\n\")\n print(\"1. Create a new language model\")\n print(\"2. Generate a text from the language model, and print it to the screen\")\n print(\"3. Generate a user-specified number of texts from the language model, and write them to a file\")\n print(\"4. Create a new language model with smoothing\")\n print(\"5. Exit the program \\n\")\n\n # \"Switch\" imitation\n def choose_option(self):\n entered_option_number = self.enter_int_number()\n switcher = {\n 1: self.create_language_model,\n 2: self.generate_text,\n 3: self.generate_text_and_save,\n 4: self.create_language_model_with_smoothing,\n 5: self.exit_the_program\n }\n func = switcher.get(entered_option_number, self.option_does_not_exist)\n \n return func()\n \n def create_language_model(self, smoothing = False):\n print(\"\\nCreating a language model\")\n print(\"Enter the number for N-parameter\")\n n_parameter = self.enter_int_number()\n\n if(n_parameter > 10 or n_parameter <= 1):\n print(\"n-parameter is invalid. Please, enter the value less than 10 and more than 1.\")\n return\n\n self.language_model = LanguageModel(n_parameter)\n if smoothing:\n self.language_model.turn_on_smoothing() \n\n valid_file_path = self.find_file()\n\n with open(valid_file_path, \"r\") as openedFile:\n full_text = openedFile.read()\n text_preprocessing = TextPrepocessing(full_text)\n tokens = text_preprocessing.start_preprocessing() \n self.language_model.train(tokens)\n\n def find_file(self):\n try:\n print(\"\\nEnter the file-path with text for training the language model.\")\n print(\"Or enter \\\"NONE\\\" or \\\"none\\\" (without \\\"\\\") or just press \\\"Enter\\\" key for reading train_shakespeare.txt from the same directory with main.py\")\n entered_file_path = input()\n if (entered_file_path == \"NONE\" or entered_file_path == \"none\" or entered_file_path == \"\"): \n # Read the default file from the same directory, if NONE entered \n __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__)))\n entered_file_path = __location__.replace('\\\\', '/') + \"/train_shakespeare.txt\" \n else:\n # Read file from the entered file-path\n entered_file_path = entered_file_path.replace('\\\\', '/') \n\n with open(entered_file_path, \"r\") as openedFile:\n print(\"\\nFile succesfully found \\n\")\n\n return entered_file_path \n except FileNotFoundError as ex:\n print(\"File not found. Please, try one more time\")\n self.find_file()\n\n def generate_text(self, intered_text = None):\n try:\n if (self.language_model == None):\n raise AttributeError()\n print(\"Enter desired begining of the text\")\n print(\"Or enter \\\"NONE\\\" or \\\"none\\\" (without \\\"\\\") or just press \\\"Enter\\\" key for generating random text\")\n entered_begining = input()\n generated_text = \"\"\n if (entered_begining != \"NONE\" or entered_begining != \"none\" or entered_begining != \"\"):\n generated_text = self.language_model.generate(entered_begining.split()) \n else:\n generated_text = self.language_model.generate()\n\n if generated_text != None:\n print(\"\\nGenerated text:\")\n print(generated_text)\n else:\n print(\"\\nUnfortunately with this beginning nothing was found\")\n \n except AttributeError:\n print(\"\\nIt is necessary first of all create a language model (option 1)\")\n self.create_language_model()\n\n def generate_text_and_save(self):\n print(\"\\nEnter desired number of texts\")\n entered_number_of_texts = self.enter_int_number()\n if (entered_number_of_texts > 0 and entered_number_of_texts < 1000):\n print(\"Writing texts to a file has started\")\n __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__)))\n file_in_root_folder = __location__.replace('\\\\', '/') + \"/new_shakespeare.txt\" \n with open(file_in_root_folder, \"w\") as created_file:\n i = 1\n while i in range(0, entered_number_of_texts + 1):\n generated_text = self.language_model.generate()\n created_file.write(\"{0}. {1}{2}\".format(i, generated_text, \"\\n\"))\n i += 1\n print(\"Writing texts to a file has ended\")\n else:\n print(\"Entered inappropriate number\")\n self.generate_text_and_save()\n\n def create_language_model_with_smoothing(self):\n self.create_language_model(smoothing = True)\n\n def exit_the_program(self):\n print(\"\\nProgram is over.\")\n Main.program_is_over = True\n\n # If entered the not existed number of operation\n def option_does_not_exist(self):\n print(\"\\nThere is no such option. Repeat entering option number again.\\n\")\n self.choose_option() \n\n def start(self):\n while True:\n # Close program if this option picked\n if Main.program_is_over:\n break\n self.present_options()\n self.choose_option()\n\nif __name__ == \"__main__\":\n # stuff only to run when not called via 'import' here\n main = Main()\n main.start()","repo_name":"FoxGriVer/PythonLabs","sub_path":"n-gramms/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":6830,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"22844423566","text":"print(\"function s()\")\ndef s():\n # local\n a = 11\n b = 22\n c = 33\n print('local:', a, b, c) # 11, 22, 33\n\n# global\na = 100\nb = 200\nc = 300\n\ns()\nprint('global:', a, b, c) # 100, 200, 300\nprint()\n\n\nprint(\"\\nfunction f()\")\ndef f():\n # local\n aa = 11\n bb = 22\n print('local:', aa, bb, cc) # 11, 22, 300 (c is global variable)\n\n# global\naa = 100\nbb = 200\ncc = 300\n\nf()\nprint('global:', aa, bb, cc) # 100, 200, 300\n\n\nprint(\"\\nfunction ss()\")\ndef ss(aa, bb, cc):\n # local\n print('local:', aa, bb, cc)\n\n# global\naa = 100\nbb = 200\ncc = 300\n\nss(aa, bb, cc) # параметри, що передаються у ф-ію, автоматично стають локальними змінними всередині ф-ії\nprint('global:', aa, bb, cc)\n\n\n\nprint(\"\\nfunction sss()\")\ndef sss(aa, bb, cc):\n # local\n print(id(aa)) # aa - is global\n aa = 20\n print(id(aa)) # aa - is local now\n print('local:', aa, bb, cc) # local: 20 200 300\n\n# global\naa = 100\nbb = 200\ncc = 300\n\nsss(aa, bb, cc) # параметри, що передаються у ф-ію, автоматично стають локальними змінними всередині ф-ії\nprint('global:', aa, bb, cc) # global: 100 200 300\nprint(id(aa))\n","repo_name":"SvitlanaPY/Python_tasks","sub_path":"Func/Func_Rules/Obl_Vydym_local_global.py","file_name":"Obl_Vydym_local_global.py","file_ext":"py","file_size_in_byte":1279,"program_lang":"python","lang":"uk","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"19680241218","text":"#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n# vim:fenc=utf-8\n\n\"\"\"\nDefine file paths\n\"\"\"\n\nfrom __future__ import annotations\n\nfrom pathlib import Path\nimport os, platform, sys, json\nfrom meerschaum.utils.typing import Union\nfrom meerschaum.config.static import STATIC_CONFIG\n\nDOT_CONFIG_DIR_PATH = Path(\n os.environ.get('XDG_CONFIG_HOME', Path.home() / '.config')\n if platform.system() != 'Windows'\n else os.environ.get('AppData', Path.home() / 'AppData' / 'Roaming')\n)\n\nDEFAULT_ROOT_DIR_PATH = (\n (DOT_CONFIG_DIR_PATH / 'meerschaum')\n if platform.system() != 'Windows'\n else (DOT_CONFIG_DIR_PATH / 'Meerschaum')\n)\n\n\nENVIRONMENT_ROOT_DIR = STATIC_CONFIG['environment']['root']\nif ENVIRONMENT_ROOT_DIR in os.environ:\n _ROOT_DIR_PATH = Path(os.environ[ENVIRONMENT_ROOT_DIR]).resolve()\n if not _ROOT_DIR_PATH.exists():\n print(\n f\"Invalid root directory '{str(_ROOT_DIR_PATH)}' set for \" +\n f\"environment variable '{ENVIRONMENT_ROOT_DIR}'.\\n\" +\n f\"Please enter a valid path for {ENVIRONMENT_ROOT_DIR}.\",\n file = sys.stderr,\n )\n sys.exit(1)\nelse:\n _ROOT_DIR_PATH = DEFAULT_ROOT_DIR_PATH\n\nENVIRONMENT_PLUGINS_DIR = STATIC_CONFIG['environment']['plugins']\nif ENVIRONMENT_PLUGINS_DIR in os.environ:\n try:\n PLUGINS_DIR_PATHS = (\n [\n Path(path).resolve()\n for path in json.loads(os.environ[ENVIRONMENT_PLUGINS_DIR])\n ] if os.environ[ENVIRONMENT_PLUGINS_DIR].lstrip().startswith('[')\n else [\n Path(path_str).resolve()\n for path_str in os.environ[ENVIRONMENT_PLUGINS_DIR].split(':')\n if path_str\n ]\n )\n except Exception as e:\n PLUGINS_DIR_PATHS = []\n\n if not PLUGINS_DIR_PATHS:\n print(\n \"Invalid plugins directories set for \"\n f\"environment variable '{ENVIRONMENT_PLUGINS_DIR}'.\\n\\n\"\n f\"Set this to a colon-separated path string:\\n\\n\"\n f\"`export {ENVIRONMENT_PLUGINS_DIR}=./plugins:/another/path/to/plugins`\\n\\n\"\n \"or a JSON-encoded path list:\\n\\n\"\n f\"`export {ENVIRONMENT_PLUGINS_DIR}=\" + \"'[\\\"./plugins\\\", \\\"/another/path/to/plugins\\\"]'`\"\n f\"\",\n )\n sys.exit(1)\nelse:\n PLUGINS_DIR_PATHS = [_ROOT_DIR_PATH / 'plugins']\n\n### Remove duplicate plugins paths.\n_seen_plugins_paths, _plugins_paths_to_remove = set(), set()\nfor _plugin_path in PLUGINS_DIR_PATHS:\n if _plugin_path in _seen_plugins_paths:\n _plugins_paths_to_remove.add(_plugin_path)\n _seen_plugins_paths.add(_plugin_path)\nfor _plugin_path in _plugins_paths_to_remove:\n PLUGINS_DIR_PATHS.remove(_plugin_path)\n\n\npaths = {\n 'PACKAGE_ROOT_PATH' : str(Path(__file__).parent.parent.resolve()),\n 'ROOT_DIR_PATH' : str(_ROOT_DIR_PATH),\n 'VIRTENV_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'venvs'),\n 'CONFIG_DIR_PATH' : ('{ROOT_DIR_PATH}', 'config'),\n 'DEFAULT_CONFIG_DIR_PATH' : ('{ROOT_DIR_PATH}', 'default_config'),\n 'PATCH_DIR_PATH' : ('{ROOT_DIR_PATH}', 'patch_config'),\n 'PERMANENT_PATCH_DIR_PATH' : ('{ROOT_DIR_PATH}', 'permanent_patch_config'),\n 'INTERNAL_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', '.internal'),\n\n 'STACK_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'stack'),\n 'STACK_COMPOSE_FILENAME' : 'docker-compose.yaml',\n 'STACK_COMPOSE_PATH' : ('{STACK_RESOURCES_PATH}', '{STACK_COMPOSE_FILENAME}'),\n 'STACK_ENV_FILENAME' : '.env',\n 'STACK_ENV_PATH' : ('{STACK_RESOURCES_PATH}', '{STACK_ENV_FILENAME}'),\n\n 'SHELL_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', ),\n 'SHELL_HISTORY_PATH' : ('{SHELL_RESOURCES_PATH}', '.mrsm_history'),\n\n 'API_RESOURCES_PATH' : ('{PACKAGE_ROOT_PATH}', 'api', 'resources'),\n 'API_STATIC_PATH' : ('{API_RESOURCES_PATH}', 'static'),\n 'API_TEMPLATES_PATH' : ('{API_RESOURCES_PATH}', 'templates'),\n 'API_CONFIG_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'api'),\n 'API_SECRET_KEY_PATH' : ('{API_CONFIG_RESOURCES_PATH}', '.api_secret_key'),\n 'API_UVICORN_RESOURCES_PATH' : ('{API_CONFIG_RESOURCES_PATH}', 'uvicorn'),\n 'API_UVICORN_CONFIG_PATH' : ('{API_UVICORN_RESOURCES_PATH}', '.thread_config.json'),\n\n 'CACHE_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', '.cache'),\n 'PIPES_CACHE_RESOURCES_PATH' : ('{CACHE_RESOURCES_PATH}', 'pipes'),\n 'USERS_CACHE_RESOURCES_PATH' : ('{CACHE_RESOURCES_PATH}', 'users'),\n\n 'PLUGINS_RESOURCES_PATH' : ('{INTERNAL_RESOURCES_PATH}', 'plugins'),\n 'PLUGINS_INTERNAL_LOCK_PATH' : ('{INTERNAL_RESOURCES_PATH}', 'plugins.lock'),\n 'PLUGINS_ARCHIVES_RESOURCES_PATH': ('{PLUGINS_RESOURCES_PATH}', '.archives'),\n 'PLUGINS_TEMP_RESOURCES_PATH' : ('{PLUGINS_RESOURCES_PATH}', '.tmp'),\n 'PLUGINS_INIT_PATH' : ('{PLUGINS_RESOURCES_PATH}', '__init__.py'),\n\n 'SQLITE_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'sqlite'),\n 'SQLITE_DB_PATH' : ('{SQLITE_RESOURCES_PATH}', 'mrsm_local.db'),\n\n 'DUCKDB_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'duckdb'),\n 'DUCKDB_PATH' : ('{DUCKDB_RESOURCES_PATH}', 'duck.db'),\n\n 'GRAFANA_RESOURCES_PATH' : ('{STACK_RESOURCES_PATH}', 'grafana', 'resources'),\n 'GRAFANA_DATASOURCE_PATH' : (\n '{GRAFANA_RESOURCES_PATH}', 'provisioning', 'datasources', 'datasource.yaml'\n ),\n 'GRAFANA_DASHBOARD_PATH' : (\n '{GRAFANA_RESOURCES_PATH}', 'provisioning', 'dashboards', 'dashboard.yaml'\n ),\n 'MOSQUITTO_RESOURCES_PATH' : ('{STACK_RESOURCES_PATH}', 'mosquitto', 'resources'),\n 'MOSQUITTO_CONFIG_PATH' : ('{MOSQUITTO_RESOURCES_PATH}', 'mosquitto.conf'),\n\n 'PORTABLE_CHECK_READLINE_PATH' : ('{SHELL_RESOURCES_PATH}', '.readline_attempted_install'),\n\n 'DAEMON_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'jobs'),\n 'LOGS_RESOURCES_PATH' : ('{ROOT_DIR_PATH}', 'logs'),\n}\n\ndef set_root(root: Union[Path, str]):\n \"\"\"Modify the value of `ROOT_DIR_PATH`.\"\"\"\n paths['ROOT_DIR_PATH'] = Path(root).resolve()\n for path_name, path_parts in paths.items():\n if isinstance(path_parts, tuple) and path_parts[0] == '{ROOT_DIR_PATH}':\n globals()[path_name] = __getattr__(path_name)\n\ndef __getattr__(name: str) -> Path:\n if name not in paths:\n if name not in globals():\n raise AttributeError(f\"Could not import '{name}'.\")\n return globals()[name]\n\n if isinstance(paths[name], (list, tuple)) and len(paths[name]) > 0:\n ### recurse through paths to create resource directories.\n parts = []\n for p in paths[name]:\n if str(p).startswith('{') and str(p).endswith('}'):\n parts.append(__getattr__(p[1:-1]))\n else:\n parts.append(p)\n path = Path(os.path.join(*parts))\n else:\n path = Path(paths[name])\n\n ### Create directories or touch files.\n if name.endswith('RESOURCES_PATH') or name == 'CONFIG_DIR_PATH':\n path.mkdir(parents=True, exist_ok=True)\n elif 'FILENAME' in name:\n path = str(path)\n\n return path\n\n","repo_name":"bmeares/Meerschaum","sub_path":"meerschaum/config/_paths.py","file_name":"_paths.py","file_ext":"py","file_size_in_byte":7311,"program_lang":"python","lang":"en","doc_type":"code","stars":110,"dataset":"github-code","pt":"79"} +{"seq_id":"9542583549","text":"#!/usr/bin/env python\n# encoding: utf-8\n# PYTHON_ARGCOMPLETE_OK\n\n# from __future__ imports must occur at the beginning of the file\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\nfrom __future__ import division\n\nimport sys\nimport time\n\n# unify Python 2 and 3\nif sys.version_info[0] == 3:\n\traw_input = input\n\nfrom . import const\nfrom . import gvar\nfrom .printer_util import (\n\tiswindows, human_speed, human_size, human_time_short)\n\ndef colorstr(msg, fg, bg):\n\tCSI = '\\x1b['\n\tfgs = ''\n\tbgs = ''\n\tif fg >=0 and fg <= 7:\n\t\tfgs = str(fg + 30)\n\n\tif bg >= 0 and bg <=7:\n\t\tbgs = str(bg + 40)\n\n\tcs = ';'.join([fgs, bgs]).strip(';')\n\tif cs:\n\t\treturn CSI + cs + 'm' + msg + CSI + '0m'\n\telse:\n\t\treturn msg\n\ndef pr(msg):\n\tprint(msg)\n\t# we need to flush the output periodically to see the latest status\n\tnow = time.time()\n\tif now - gvar.last_stdout_flush >= const.PrintFlushPeriodInSec:\n\t\tsys.stdout.flush()\n\t\tgvar.last_stdout_flush = now\n\ndef prcolor(msg, fg, bg):\n\tif sys.stdout.isatty() and not iswindows():\n\t\tpr(colorstr(msg, fg, bg))\n\telse:\n\t\tpr(msg)\n\ndef ask(msg, enter = True):\n\tpr(msg)\n\tif enter:\n\t\tpr('Press [Enter] when you are done')\n\treturn raw_input()\n\n# print progress\n# https://stackoverflow.com/questions/3173320/text-progress-bar-in-the-console\ndef pprgr(finish, total, start_time = None, existing = 0,\n\t\tprefix = '', suffix = '', seg = 20):\n\t# we don't want this goes to the log, so we use stderr\n\tif total > 0:\n\t\tsegth = seg * finish // total\n\t\tpercent = 100 * finish // total\n\t\tcurrent_batch_percent = 100 * (finish - existing) // total\n\telse:\n\t\tsegth = seg\n\t\tpercent = 100\n\t\tcurrent_batch_percent = 0\n\n\teta = ''\n\tnow = time.time()\n\tif start_time is not None and current_batch_percent > 5 and finish > 0:\n\t\tfinishf = float(finish) - float(existing)\n\t\ttotalf = float(total)\n\t\tremainf = totalf - float(finish)\n\t\telapsed = now - start_time\n\t\tspeed = human_speed(finishf / elapsed)\n\t\teta = 'ETA: ' + human_time_short(elapsed * remainf / finishf) + \\\n\t\t\t\t' (' + speed + ', ' + \\\n\t\t\t\thuman_time_short(elapsed) + ' gone)'\n\tmsg = '\\r' + prefix + '[' + segth * '=' + (seg - segth) * '_' + ']' + \\\n\t\t\" {}% ({}/{})\".format(percent, human_size(finish, 1), human_size(total, 1)) + \\\n\t\t' ' + eta + suffix\n\t#msg = '\\r' + prefix + '[' + segth * '=' + (seg - segth) * '_' + ']' + \\\n\t#\t\" {}% ({}/{})\".format(percent, human_size(finish), human_size(total)) + \\\n\t#\t' ' + eta + suffix\n\tsys.stderr.write(msg + ' ') # space is used as a clearer\n\tsys.stderr.flush()\n\n","repo_name":"houtianze/bypy","sub_path":"bypy/printer_console.py","file_name":"printer_console.py","file_ext":"py","file_size_in_byte":2479,"program_lang":"python","lang":"en","doc_type":"code","stars":7130,"dataset":"github-code","pt":"79"} +{"seq_id":"18949580729","text":"import falcon\n\nfrom pomoccore import db\nfrom pomoccore.models import User\nfrom pomoccore.models import Admin\nfrom pomoccore.utils import validators\nfrom pomoccore.utils import response\nfrom pomoccore.utils.errors import APIUnprocessableEntityError\n\n\nclass AdminByUsernameController(object):\n @falcon.before(validators.admin.username_exists)\n def on_get(self, req, resp):\n admin = db.Session.query(User).filter_by(username=req.get_json('username')).one()\n\n data = dict()\n data['admin'] = dict()\n for scope in req.scope:\n try:\n data['admin'][scope] = getattr(admin, scope)\n except AttributeError:\n raise APIUnprocessableEntityError('Invalid scope \\'{0}\\''.format(scope),\n 'Scope is not part of the teacher.')\n\n response.set_successful_response(\n resp, falcon.HTTP_200, 'Ignacio! Where is the damn internal code?',\n 'Successful admin data retrieval', 'Admin data successfully gathered.', data\n )\n\n\nclass AdminController(object):\n @falcon.before(validators.admin.exists)\n def on_get(self, req, resp):\n data = dict()\n data['admin'] = dict()\n if req.get_json('admin_id') == '__all__':\n admins = db.Session.query(User).filter_by(user_type='admin').order_by(User.last_name.asc(),\n User.first_name.asc(),\n User.middle_name.asc(),\n User.id_number.asc()).all()\n\n row_ctr = 0\n for admin in admins:\n data['admin'][row_ctr] = dict()\n for scope in req.scope:\n try:\n if scope == 'birth_date':\n data['admin'][row_ctr][scope] = getattr(admin, scope).strftime('%B %d, %Y')\n else:\n data['admin'][row_ctr][scope] = getattr(admin, scope)\n except AttributeError:\n raise APIUnprocessableEntityError('Invalid scope \\'{0}\\''.format(scope),\n 'Scope is not part of the admin.')\n\n row_ctr += 1\n else:\n admin = db.Session.query(Admin).filter_by(admin_id=req.get_json('admin_id')).one()\n\n data['admin'] = dict()\n for scope in req.scope:\n try:\n data['admin'][scope] = getattr(admin, scope)\n except AttributeError:\n raise APIUnprocessableEntityError('Invalid scope \\'{0}\\''.format(scope),\n 'Scope is not part of the admin.')\n\n response.set_successful_response(\n resp, falcon.HTTP_200, 'Ignacio! Where is the damn internal code?',\n 'Successful admin data retrieval', 'Admin data successfully gathered.', data\n )\n","repo_name":"Pisay127/pomoc-core","sub_path":"pomoccore/controllers/admin_controller.py","file_name":"admin_controller.py","file_ext":"py","file_size_in_byte":3112,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39486011895","text":"# Lucien Lo \nfrom BEvents import BEvents\nfrom BranchBuilder import BranchBuilder\n\nclass MultiBEvents(object):\n def __init__(self, files, trees, maxEvents = -1):\n self.files = files\n self.trees = trees\n total_number_evt = sum([tree.GetEntries() for tree in self.trees])\n self.nEvents = min(total_number_evt, maxEvents) if (maxEvents > -1) else total_number_evt\n\n self.EvtClassDict = {}\n total_sum_entries = 0\n for tree in self.trees:\n total_sum_entries += tree.GetEntries()\n self.EvtClassDict[total_sum_entries] = BEvents(tree)\n self.iEvent = -1\n \n def __getitem__(self, i):\n if i >= self.nEvents:\n self.iEvent = -1\n raise IndexError(\"the index is out of range: \" + str(i))\n self.iEvent = i\n for entries in self.EvtClassDict:\n if self.iEvent < entries: break\n return self.EvtClassDict[entries]\n\n def __iter__(self):\n for self.iEvent in xrange(self.nEvents):\n for entries in self.EvtClassDict:\n if self.iEvent < entries: break\n yield self.EvtClassDict[entries]\n self.iEvent = -1\n","repo_name":"lucien1011/PyNTupleAPI","sub_path":"Events/MultiBEvents.py","file_name":"MultiBEvents.py","file_ext":"py","file_size_in_byte":1198,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"42229125945","text":"print(\"======== VENDING MACHINE BERAS ========\\n\")\npil = False\nwhile pil is False:\n print(\"======== Pilih Jenis Beras :\")\n print(\"1. Pandan Wangi (Rp 9.500/kg) \")\n print(\"2. Raja Lele (Rp.10.000/kg)\")\n print(\"3. Setra Ramos (Rp 9.000/kg)\")\n print(\"4. IR 42 (Rp 8.600/kg) \")\n print(\"5. Solok (Rp 9.200/kg) \")\n beras = int(input(\"(1-5): \"))\n if beras == 1:\n bname = \"Pandan Wangi\"\n price = 9500\n pil = True\n elif beras == 2:\n bname = \"Raja Lele\"\n price = 10000\n pil = True\n elif beras == 3:\n bname = \"Setra Ramos\"\n price = 9000\n pil = True\n elif beras == 4:\n bname = \"IR 42\"\n price = 8600\n pil = True\n elif beras == 5:\n bname = \"Solok\"\n price = 9200\n pil = True\n else : \n print(\"Nilai yang anda masukkan salah!\")\n print(\"Silahkan pilih kembali (1-5): \")\n pil = False\nprint(\"Anda memilih beras : \" + bname +\"\\n\")\namt = False\nwhile amt == False:\n print(\"======== Paket beras : \")\n print(\"1. 1 KG\")\n print(\"2. 5 KG\")\n print(\"3.10 KG\")\n amtCat = int(input(\"Paket beras : \")) \n if amtCat == 1:\n tot = price\n pak = \"1Kg\"\n amt = True\n elif amtCat == 2:\n tot = price * 5\n pak = \"5Kg\"\n amt = True\n elif amtCat == 3:\n tot = price * 10\n pak = \"10Kg\"\n amt = True\n else : \n print(\"Pilihan yang anda masukkan salah.\")\n print(\"Coba lagi!\")\n amt = False\nfinalState = False\nwhile finalState == False:\n print(\"\\nRincian belanja: \")\n print(\"Beras : \" + bname )\n print(\"Paket : \" + pak)\n print(\"Total : Rp\" + str(tot)+\"\\n\")\n print (\"Pilih metode pembayaran : \")\n print(\"1. Tunai\")\n print(\"2. E-money \")\n mp = int(input(\"(1/2) : \"))\n if mp == 1:\n kembali = 0\n cash = 0\n while cash < tot:\n cash = int(input(\"Masukkan jumlah uang: Rp.\"))\n kembali = cash - tot \n if cash < tot:\n print(\"Mohon maaf, jumlah uang anda tidak mencukupi\")\n else : finalState = True\n eMoney = False\n elif mp == 2:\n cash = input(\"Masukkan nomor e-money: \")\n eMoney = True\n finalState = True\n else :\n print(\"Pilihan anda salah. Coba lagi!\")\n finalState = False\n\nprint(\"Transaksi berhasil!\")\nprint(\"Rincian belanja: \")\nprint(\"Beras : \" + bname )\nprint(\"Paket : \" + pak)\nprint(\"Total : Rp\" + str(tot)+\"\\n\")\nif eMoney == True:\n sald = 2504000 - tot\n print (\"Sisa saldo anda : Rp.\" + str(sald))\nelse :\n print(\"Bayar : Rp\" + str(cash))\n print(\"Kembali : Rp \" + str(kembali))\n","repo_name":"Rizqirazkafi/kuliah-y2","sub_path":"TBO/vmBeras.py","file_name":"vmBeras.py","file_ext":"py","file_size_in_byte":2699,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"43494956383","text":"import requests\nfrom bs4 import BeautifulSoup\n\nurl = 'https://hogangnono.com/search?q=%EA%B2%BD%EA%B8%B0%EB%8F%84%20%EA%B9%80%ED%8F%AC%EC%8B%9C%20%ED%92%8D%EB%AC%B4%EB%8F%99'\nres = requests.get(url)\nres.raise_for_status()\n\nsoup = BeautifulSoup(res.text, 'lxml')\n\n# 해당지역 단지 목록 가져오기\narea_list = soup.find_all('li', attrs={'class', 'apt'})\nprint(area_list)","repo_name":"simbyungki/web_scraping","sub_path":"practice01/bs4_pungmu_hogangnono.py","file_name":"bs4_pungmu_hogangnono.py","file_ext":"py","file_size_in_byte":377,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"17042922181","text":"from datetime import date\nano = int(input('Em que ano você nasceu? '))\nhoje = date.today().year\nidade = hoje-ano-1\n\nprint('Você tem {} anos e sua categoria é '.format(idade),end=\"\")\n\nif idade <= 9:\n print('\\033[36mMirim')\nelif idade <= 14:\n print('\\033[32mInfantil')\nelif idade <= 19:\n print('\\033[34mJunior')\nelif idade <= 25:\n print('\\033[035mSênior')\nelif idade > 25:\n print('\\033[31mMaster')\n","repo_name":"Joaoespindola1/Python-aprendizado","sub_path":"Desafios/ex041.py","file_name":"ex041.py","file_ext":"py","file_size_in_byte":415,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"71376751935","text":"import pandas as pd\n\n'''\nObject to preprocess data\n'''\n\ndef read_in_dataset(dataset, data_folder='raw', data_type='csv', verbose=False):\n '''\n Read in dataset (csv format) to pandas dataframe\n\n Keyword Arguments:\n ------------------\n * dataset - string with dataset filename\n * data_folder - string with either raw or processed\n * verbose - True will print intormation about the dataset\n\n Returns:\n --------\n a pandas dataframe\n '''\n if data_type == 'csv':\n df = pd.read_csv('../data/{}/{}'.format(data_folder, dataset))\n elif data_type == 'excel':\n df = pd.read_excel('../data/{}/{}'.format(data_folder, dataset))\n\n if verbose:\n print('\\n{0:-^80}'.format(' Reading in the following dataset: {0}'.format(dataset)))\n print(\"\\n Shape: {0} rows and {1} columns\".format(*df.shape))\n print('\\n{0:-^80}\\n'.format(' It has the following columns '))\n print(df.columns)\n print('\\n{0:-^80}\\n'.format(' The first 5 rows look like this '))\n print(df.head())\n\n return df\n\n\ndef get_num_of_levels(series, sep='>', verbose=False):\n '''\n Get maximum number of category levels\n '''\n max_num_levels = series.str.split(sep).apply(len).max()\n\n if verbose:\n print(\"Max Number of Category Levels: {}\".format(max_num_levels))\n\n return max_num_levels\n\n\ndef flatten_categories(category_series, df=None, drop_col=None, sep='>'):\n '''\n Take in Series with categories in string format and flatten into columns\n\n Keyword Arguments:\n ------------------\n * category_series - series with string of categories\n * df - pandas dataframe\n * drop_col - name of column with nested categories (string)\n * sep - puncuation that separates categories\n\n\n Returns:\n --------\n a pandas dataframe\n '''\n\n num_levels = get_num_of_levels(category_series, sep=sep)\n\n col_labels = ['L' + str(i) for i in range(1, num_levels+1)]\n\n\n category_levels = pd.DataFrame(category_series.str.split(sep).values.tolist(), columns=col_labels)\n category_levels.fillna(value=pd.np.nan, inplace=True)\n\n if df is not None:\n merged_df = pd.merge(df,category_levels, left_index=True, right_index=True).drop(drop_col, axis=1)\n\n return merged_df\n else:\n return category_levels\n\n\ndef search_cons_status(cons_searches, raw_searches):\n '''\n Get status on number of unique search terms\n '''\n\n num_searches = len(raw_searches)\n new_num_searches = len(set(cons_searches))\n\n num_cons = num_searches - new_num_searches\n\n per_reduction = round((num_cons/num_searches)*100,1)\n\n print(f'New Number of Searches: {new_num_searches}')\n print(f'Number of Consolidated Searches: {num_cons}')\n print(f'Percent Reduction: {per_reduction}%')\n","repo_name":"jaime-garvey/hd-decor-case","sub_path":"modules/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":2782,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"14585690640","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n\"\"\"Unit tests for meaning.py\"\"\"\n\nimport meaning\nimport unittest\n\nclass KnownValues(unittest.TestCase):\n\n knownParserValues = (\n (\"*German: [[wichtig]]\",\n [('de','wichtig','',1,False,'')]\n ),\n (\"*[[Esperanto]]: [[grava]]\",\n [('eo','grava','',1,False,'')]\n ),\n (\"*{{fr}}: [[importante]] {{f}}\",\n [('fr','importante','f',1,False,'')]\n ),\n (\"*Dutch: [[voorbeelden]] ''n, pl'', [[instructies]] {{f}}, {{p}}\",\n [('nl','voorbeelden','n',2,False,''),\n ('nl','instructies', 'f',2,False,'')]\n ),\n (\"*Russian: [[шесток]] ''m'' (shestok)\",\n [('ru','шесток','m',1,False,'shestok')]\n ),\n (\"*Kazakh: сәлем, салам, сәлеметсіздер(respectable)\",\n [('ka','сәлем','',1,False,''),\n ('ka','салам','',1,False,''),\n ('ka','сәлеметсіздер','',1,False,'respectable')]\n ),\n (\"*Chinese(Mandarin):[[你好]](ni3 hao3), [[您好]](''formal'' nin2 hao3)\",\n [('zh','你好','',1,False,'ni3 hao3'),\n ('zh','您好','',1,False,\"''formal'' nin2 hao3\")]\n ),\n (\"*German: [[Lamm]] ''n'' [[:de:Lamm|(de)]]\",\n [('de','Lamm','n',1,False,'')]\n ),\n (\"*Italian: [[pronto#Italian|pronto]]\",\n [('it','pronto','',1,False,'')]\n ),\n )\n\n def testParser(self):\n '''self.term, self.gender, self.number, self.diminutive and remark parsed correctly from Wiki format'''\n for wikiline, results in self.knownParserValues:\n ameaning = meaning.Meaning('en', 'dummy')\n ameaning.parseTranslations(wikiline)\n i=0\n for termlang, thisterm, termgender, termnumber, termisadiminutive, remark in results:\n resultterm = ameaning.translations[termlang]['alltrans'][i]['trans']\n self.assertEqual(resultterm.getTerm(), thisterm)\n self.assertEqual(resultterm.getGender(), termgender)\n self.assertEqual(resultterm.getNumber(), termnumber)\n# self.assertEqual(resultterm.getIsDiminutive(), termisadiminutive)\n self.assertEqual(ameaning.translations[termlang]['alltrans'][i]['remark'], remark)\n i+=1\n\nif __name__ == \"__main__\":\n unittest.main()\n \n","repo_name":"dantman/pywikia","sub_path":"wiktionary/meaningtest.py","file_name":"meaningtest.py","file_ext":"py","file_size_in_byte":2692,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"79"} +{"seq_id":"72659735615","text":"from Stack import *\n\nclass MyQueue:\n\tdef __init__(self):\n\t\tself.first = None\n\t\tself.last = None\n\t\tself.size = 0\n\t\tself.addItemStack = MyStack()\n\t\tself.removeItemStack = MyStack()\n\t\n\tdef add(self, item):\n\t\tself.addItemStack.push(item)\n\t\tself.size += 1\n\t\tself.last = item\n\n\t\tif self.size == 1:\n\t\t\tself.first = self.last\n\n\n\tdef remove(self):\n\t\t#This is when the first value in queue will change after remove action\n\t\ttopItem = self.peek()\n\t\tself.removeItemStack.pop()\n\t\tself.size -= 1\n\n\t\t#update self.first\n\t\tself.first = self.peek()\n\t\tif self.size <= 1:\n\t\t\tself.last = self.first\n\n\t\treturn topItem\n\n\tdef isEmpty(self):\n\t\treturn True if self.size == 0 else False\n\t\n\tdef peek(self):\n\t\tif self.removeItemStack.isEmpty():\n\t\t\tif self.addItemStack.isEmpty():\n\t\t\t\treturn None\n\t\t\telse:\n\t\t\t\t#pop all the items from addItemStack, push all the items to removeItemStack\n\t\t\t\twhile not self.addItemStack.isEmpty():\n\t\t\t\t\tpopItem = self.addItemStack.pop()\n\t\t\t\t\tself.removeItemStack.push(popItem)\n\n\t\t\t\ttopItem = self.removeItemStack.top\n\t\t\t\treturn topItem\n\t\telse:\n\t\t\ttopItem = self.removeItemStack.top\n\t\t\treturn topItem\n\nif __name__ == \"__main__\":\n\tqstack = MyQueue()\n\tfor i in range(1, 6):\n\t\tqstack.add(i)\n\t\n\tprint(\"first: \", qstack.first)\n\tprint(\"last: \", qstack.last)\n\tprint(\"peek: \", qstack.peek())\n\t\n\tprint(\"remove item: \", qstack.remove())\n\n\tprint(\"first: \", qstack.first)\n\tprint(\"peek: \", qstack.peek())\n\tprint(\"last: \", qstack.last)\n\n\twhile not qstack.isEmpty():\n\t\tprint(qstack.remove())\n\n\tprint(\"after queue is empty: \")\n\tprint(\"first: \", qstack.first)\n\tprint(\"last: \", qstack.last)\n\n\n","repo_name":"becca6223/Tech_Prep","sub_path":"stack_and_queues/QueueViaStacks.py","file_name":"QueueViaStacks.py","file_ext":"py","file_size_in_byte":1575,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"38581980904","text":"#!/usr/bin/env python\n\n\"\"\"Interpolate missing values, preserving some amount of variance.\n\nUsage: {PROG} -f field -x field [ input ]\n\n-x field - x axis column (required)\n-f field - y axis column (field to be interpolated - required)\n\nIf given, the input file is used, otherwise sys.stdin. Output is always to\nsys.stdout.\n\nCurrently, all this does is linear interpolation. Variance-preserving\ninterpolation is TBD.\n\n\"\"\"\n\nimport getopt\nimport os\nimport sys\n\nimport numpy as np\nimport pandas as pd\n\nPROG = os.path.split(sys.argv[0])[1]\n\ndef usage(msg=None):\n \"help user\"\n if msg is not None:\n printe(msg)\n printe()\n printe(__doc__.strip().format(**globals()))\n\ndef main():\n \"see __doc__\"\n opts, args = getopt.getopt(sys.argv[1:], \"hf:x:\",\n [\"help\", \"field=\", \"xaxis=\",])\n x_axis = None\n field = None\n for opt, arg in opts:\n if opt in (\"--help\", \"-h\"):\n usage()\n return 0\n if opt in (\"-f\", \"--field\"):\n field = arg\n elif opt in (\"-x\", \"--xaxis\"):\n x_axis = arg\n if field is None or x_axis is None:\n usage(\"-x and -f are both required.\")\n return 1\n\n infile = open(args[0]) if args else sys.stdin\n header = next(infile).strip().split(\",\")\n dtype = {\n field: np.float,\n }\n for col in header:\n if col not in (x_axis, field):\n dtype[col] = str\n frame = pd.read_csv(infile, dtype=dtype, names=header,\n parse_dates=[x_axis])\n frame.index = frame[x_axis]\n field_data = frame[[field]]\n field_data = field_data.resample(\"D\").mean().interpolate()\n del frame[field]\n frame = field_data.join(frame, how=\"outer\")\n del frame[x_axis]\n frame = frame.reset_index()\n frame = frame.replace(np.nan, \"\")\n frame.to_csv(sys.stdout, index=False)\n return 0\n\ndef printe(*args, file=sys.stderr, **kwds):\n \"print, defaultint to stderr for output\"\n return print(*args, file=file, **kwds)\n\nif __name__ == \"__main__\":\n sys.exit(main())\n","repo_name":"smontanaro/csvprogs","sub_path":"csvprogs/src/interp.py","file_name":"interp.py","file_ext":"py","file_size_in_byte":2055,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"5195632322","text":"quizes = [\n ['Question 1', ['Ans 1', 'Ans 2', 'Ans 3', 'Ans 4']],\n ['Question 2', ['Ans 1', 'Ans 2', 'Ans 3', 'Ans 4']],\n ['Question 3', ['Ans 1', 'Ans 2', 'Ans 3', 'Ans 4']],\n ['Question 4', ['Ans 1', 'Ans 2', 'Ans 3', 'Ans 4']],\n ['Question 5', ['Ans 1', 'Ans 2', 'Ans 3', 'Ans 4']],\n]\n\nanswers = [1, 2, 2, 3, 4]\n\ndef calculate_score(user_ans):\n score = 0\n for i in range(len(user_ans)):\n if user_ans[i] == answers[i]:\n score += 1\n return score\n\ndef prompt_quiz(quiz):\n print('-'*40)\n print(quiz[0])\n print()\n for i in range(4):\n print('[' + str(i + 1) + '] ' + quiz[1][i]) \n print()\n print('-'*40)\n\ndef play():\n user_ans = [] \n for i in range(len(quizes)):\n prompt_quiz(quizes[i])\n ans = 0\n while ans < 1 or ans > 4:\n try:\n ans = int(input('Your answer: '))\n except ValueError as err:\n pass\n user_ans.append(ans)\n score = calculate_score(user_ans)\n print('Your score is ' + str(score))\n\ndef main():\n while True:\n play()\n answer = input('Play again? (y/n): ')\n if answer != 'y':\n break\nmain()","repo_name":"sorasora46/oreo-piethon","sub_path":"mini_projects/quiz_game/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1188,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6950905578","text":"import sys\ninput = sys.stdin.readline\n\nN, M = map(int, input().split())\ngrid = []\n\ndx = [-1, 0, 1, 0]\ndy = [0, -1, 0, 1]\n\nfor i in range(N):\n temp = list(input().rstrip(\"\\n\"))\n a = []\n for j in temp:\n a.append(int(j))\n grid.append(a)\nprint(grid)\n\n\ndef DFS(x, y, grid, visited, N, M):\n for i in range(4):\n nx = x + dx[i]\n ny = y + dy[i]\n\n if 0 <= nx < N and 0 <= ny < M and not visited[nx][ny] and grid[nx][ny] == 0:\n visited[nx][ny] = True\n DFS(nx, ny, grid, visited, N, M)\n\n\nvisited = [[False for _ in range(M)] for _ in range(N)]\nanswer = 0\nfor i in range(N):\n for j in range(M):\n if grid[i][j] == 0 and not visited[i][j]:\n answer += 1\n DFS(i, j, grid, visited, N, M)\n\nprint(answer)\n","repo_name":"TypingOK/algorithm_study","sub_path":"이것이 취업을 위한 코딩테스트다/음료수 얼려 먹기.py","file_name":"음료수 얼려 먹기.py","file_ext":"py","file_size_in_byte":782,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"34477896247","text":"from shutil import copyfile\nimport os\n\n\ndef create_directory(dir_name):\n try:\n os.makedirs(dir_name)\n except OSError:\n if not os.path.isdir(dir_name):\n raise\n\n\ndef read_compile(lines, file):\n with open(file) as problem1_file:\n append_lines = [line for line in problem1_file]\n append_lines = append_lines[2:]\n\n # Generate submission file for Problem1.py\n with open('submission/%s' % file, 'w') as combined_file:\n [combined_file.write(line) for line in lines]\n [combined_file.write(append_line) for append_line in append_lines]\n\n\nif __name__ == '__main__':\n create_directory('submission')\n\n # Combine decryptRC4 with Problem1 and Problem2\n with open('decryptRC4.py') as main_file:\n lines = [line for line in main_file]\n\n # remove lines after \"\"if __name__ == '__main__':\"\"\n lines = lines[:177]\n\n read_compile(lines, 'Problem1.py')\n read_compile(lines, 'Problem2.py')\n\n # Copy the generated results\n copyfile(\"Problem1.txt\", \"submission/Problem1.txt\")\n copyfile(\"Problem2.txt\", \"submission/Problem2.txt\")\n","repo_name":"Andyccs/CZ4024AttackRC4","sub_path":"generate_submission.py","file_name":"generate_submission.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"71856164416","text":"import random\r\nimport pygame.display\r\nimport pygame.font\r\nimport pygame.image\r\nimport pygame.mixer\r\nimport pygame.sprite\r\nfrom sprites import *\r\n\r\n\r\nclass Game:\r\n def __init__(self):\r\n # initialize game window, etc.\r\n pyg.init()\r\n pyg.mixer.init()\r\n self.screen = pyg.display.set_mode((WIDTH, HEIGHT))\r\n pyg.display.set_caption(TITTLE)\r\n self.clock = pyg.time.Clock()\r\n self.running = True\r\n self.font_name_1 = pyg.font.match_font(\"verdana\")\r\n self.font_name_2 = pyg.font.match_font(\"georgia\")\r\n self.load_data()\r\n\r\n def load_data(self):\r\n self.dir = path.dirname(__file__)\r\n self.img_dir = path.join(self.dir, \"Jumper_Data\")\r\n with open(path.join(self.dir, HS_FILE), \"r\") as f:\r\n try:\r\n self.highscore = int(f.read())\r\n except:\r\n self.highscore = 0\r\n\r\n # load spritesheet image\r\n self.spritesheet = Spritesheet(path.join(self.img_dir, SPRITESHEET))\r\n\r\n # Cloud images\r\n self.cloud_images = []\r\n for i in range(1, 4):\r\n self.cloud_images.append(pyg.image.load(path.join(self.img_dir, f\"cloud{i}.png\".format(i))).convert())\r\n\r\n # load sound\r\n self.snd_dir = path.join(self.dir, \"Game_Sound\")\r\n self.jump_sound = pygame.mixer.Sound(path.join(self.snd_dir, 'Jump10.wav'))\r\n self.boost_sound = pygame.mixer.Sound(path.join(self.snd_dir, 'Powerup7.wav'))\r\n\r\n\r\n\r\n def new(self):\r\n # Starts a new game\r\n self.score = 0\r\n self.all_sprites = pyg.sprite.Group()\r\n self.platforms = pyg.sprite.Group()\r\n self.powerups = pyg.sprite.Group()\r\n self.mobs = pyg.sprite.Group()\r\n self.clouds = pyg.sprite.Group()\r\n self.mob = Mob(self)\r\n self.player = Player(self)\r\n for plat in PLAYER_LIST:\r\n Platform(self, *plat)\r\n self.mob_timer = 0\r\n pyg.mixer.music.load(path.join(self.snd_dir, 'Happy Tune.wav'))\r\n self.run()\r\n\r\n def run(self):\r\n # Game loop\r\n pyg.mixer.music.play(loops = -1)\r\n self.playing = True\r\n while self.playing:\r\n # this confirms that the loop have completed under 0.03 second\r\n # And if it had completed before 0.3 second then it will wait till the times up and then run next loop\r\n # so that it will run at the same speed\r\n self.clock.tick(FPS)\r\n # Processing input (events)\r\n self.events()\r\n # To update the sprites after meeting a certain condition\r\n self.update()\r\n # Draw / Render\r\n self.draw()\r\n pyg.mixer.music.fadeout(500)\r\n\r\n def update(self):\r\n # Game loop - update\r\n # To update the sprites after meeting a certain condition in the next frame\r\n self.all_sprites.update()\r\n\r\n # spqwn a mob\r\n now = pyg.time.get_ticks()\r\n if now - self.mob_timer > 10000 + random.choice([-1000, -500, 0, 500, 1000]):\r\n self.mob_timer = now\r\n Mob(self)\r\n\r\n # hit mobs?\r\n # If hit mobs game will end\r\n # For pixel perfect collision we are using pygame_mask function\r\n mob_hits = pyg.sprite.spritecollide(self.player, self.mobs, False, pyg.sprite.collide_mask)\r\n if mob_hits:\r\n self.playing = False\r\n\r\n # check if player hits a platform - only if falling\r\n if self.player.vel.y > 0:\r\n hits = pyg.sprite.spritecollide(self.player, self.platforms, False)\r\n if hits:\r\n lowest_plat = hits[0]\r\n # Collecting the lowest platform in the collision.\r\n # If collision is happening with more than one platform\r\n for hit in hits:\r\n if hit.rect.bottom > lowest_plat.rect.bottom:\r\n lowest_plat = hit\r\n\r\n # Checking whether the player are out of the platform or not\r\n if self.player.pos.x < lowest_plat.rect.right + 7 and \\\r\n self.player.pos.x > lowest_plat.rect.left - 7:\r\n # If the player feet get on the platform ony then he can land on it\r\n if self.player.pos.y < lowest_plat.rect.centery:\r\n self.player.pos.y = hits[0].rect.top\r\n self.player.vel.y = 0\r\n self.player.jumping = False\r\n\r\n # if player reaches top 1/4 of the screen\r\n if self.player.rect.top <= HEIGHT / 4:\r\n if random.randrange(100)<15:\r\n Cloud(self)\r\n self.player.pos.y += max(abs(self.player.vel.y), 2)\r\n for cloud in self.clouds:\r\n cloud.rect.y += max(abs(self.player.vel.y/2), 2)\r\n for mob in self.mobs:\r\n mob.rect.y += max(abs(self.player.vel.y), 2)\r\n for plat in self.platforms:\r\n plat.rect.y += max(abs(self.player.vel.y), 2)\r\n if plat.rect.top >= HEIGHT:\r\n plat.kill()\r\n self.score += 10\r\n\r\n # If a player hits a powerup\r\n pow_hits = pyg.sprite.spritecollide(self.player, self.powerups, True)\r\n for powerup in pow_hits:\r\n if powerup.type == 'boost':\r\n self.boost_sound.play()\r\n self.player.vel.y = - BOOST_POWER\r\n self.player.jumping = False\r\n\r\n\r\n # If player hits bottom of the screen (Die!)\r\n if self.player.rect.bottom > HEIGHT:\r\n for sprite in self.all_sprites:\r\n sprite.rect.y -= max(self.player.vel.y, 10)\r\n if sprite.rect.bottom < 0:\r\n sprite.kill()\r\n if len(self.platforms) == 0:\r\n self.playing = False\r\n\r\n\r\n # Spawn new platform to keep same average number\r\n while len(self.platforms) < 6:\r\n width = random.randint(50, 150)\r\n x = random.randint(0, WIDTH-width)\r\n y = random.randint(-75, -30)\r\n p = Platform(self, x, y)\r\n self.platforms.add(p)\r\n self.all_sprites.add(p)\r\n\r\n\r\n def events(self):\r\n # Game loop - events\r\n # Processing input (events)\r\n for event in pyg.event.get():\r\n # Check for closing window\r\n if event.type == pyg.QUIT:\r\n if self.playing:\r\n self.playing = False\r\n self.running = False\r\n if event.type == pyg.KEYDOWN:\r\n if event.key == pyg.K_SPACE:\r\n self.player.jump()\r\n\r\n if event.type == pyg.KEYUP:\r\n if event.key == pyg.K_SPACE:\r\n self.player.jump_cut()\r\n\r\n\r\n def draw(self):\r\n # Game loop - draw\r\n # To drawing updated frame on the screen\r\n self.screen.fill(BGCOLOR)\r\n # To draw all the updated sprites\r\n self.all_sprites.draw(self.screen)\r\n self.screen.blit(self.player.image, self.player.rect)\r\n self.screen.blit(self.mob.image, self.mob.rect)\r\n self.draw_text((f\"Score: {self.score}\"), self.font_name_2, 23, color[\"white\"], 60, 15)\r\n if self.score > self.highscore:\r\n self.draw_text(f\"High Score: {self.score}\", self.font_name_2, 23, color[\"white\"], WIDTH - 90, 15)\r\n else :\r\n self.draw_text(f\"High Score: {self.highscore}\", self.font_name_2, 23, color[\"white\"], WIDTH-90, 15)\r\n # After drawing everything flip the display\r\n pyg.display.flip()\r\n\r\n def show_start_screen(self):\r\n # Game splash / Start screen\r\n self.background(\"splash.jpeg\")\r\n\r\n # Initializing game start music\r\n pyg.mixer.music.load(path.join(self.snd_dir, 'Yippee.wav'))\r\n pyg.mixer.music.play(loops = -1)\r\n\r\n #self.screen.fill(BGCOLOR)\r\n self.draw_text(TITTLE, self.font_name_2, 55, color[\"white\"], WIDTH/2, HEIGHT/4)\r\n self.draw_text(\"Arrows to move Space to jump\", self.font_name_2, 25, color[\"white\"], WIDTH/2, HEIGHT/2)\r\n self.draw_text(\"Press any key to play\", self.font_name_2, 25, color[\"black\"], WIDTH/2, HEIGHT*3/4+30)\r\n self.draw_text(f\"High Score: {self.highscore}\", self.font_name_2, 25, color[\"white\"], WIDTH/2, 15)\r\n pyg.display.flip()\r\n self.wait_for_key()\r\n pyg.mixer_music.fadeout(500)\r\n\r\n def background(self, image):\r\n # Changing start/end screen background\r\n self.image = path.join(self.img_dir, image)\r\n bg = pygame.image.load(self.image)\r\n bg = pyg.transform.scale(bg, (WIDTH, HEIGHT))\r\n self.screen.blit(bg, (0, 0))\r\n\r\n def show_go_screen(self):\r\n\r\n # Game over / Continue screen\r\n\r\n\r\n if not self.running:\r\n return\r\n\r\n # Initializing game ending music\r\n pyg.mixer.music.load(path.join(self.snd_dir, 'prologue.mp3'))\r\n pyg.mixer.music.play(loops=-1)\r\n self.background(\"end.jpeg\")\r\n #self.screen.fill(BGCOLOR)\r\n self.draw_text(\"GAME OVER\", self.font_name_2, 60, color[\"white\"], WIDTH / 2, HEIGHT / 4)\r\n self.draw_text(f\"Your Score: {self.score}\", self.font_name_2, 25, color[\"white\"], WIDTH / 2, HEIGHT / 2)\r\n self.draw_text(\"Press any key to play again\", self.font_name_2, 25, color[\"white\"], WIDTH / 2, HEIGHT * 3 / 4+30)\r\n if self.score > self.highscore:\r\n self.highscore = self.score\r\n self.draw_text(\"NEW HIGH SCORE!\", self.font_name_2, 25, color[\"red\"], WIDTH / 2, HEIGHT / 2 + 40)\r\n with open(path.join(self.dir, HS_FILE), \"w\") as f:\r\n f.write(str(self.score))\r\n else:\r\n self.draw_text(f\"High Score: {self.highscore} \", self.font_name_2, 25, color[\"red\"], WIDTH / 2, HEIGHT / 2 + 40)\r\n\r\n pyg.display.flip()\r\n self.wait_for_key()\r\n pyg.mixer_music.fadeout(500)\r\n\r\n\r\n\r\n def wait_for_key(self):\r\n waiting = True\r\n while waiting:\r\n self.clock.tick(FPS)\r\n for event in pyg.event.get():\r\n if event.type == pyg.QUIT:\r\n waiting = False\r\n self.running = False\r\n if event.type == pyg.KEYUP:\r\n waiting = False\r\n\r\n def draw_text(self, text, font_style, size, color, x, y):\r\n font = pygame.font.Font(font_style, size)\r\n text_surface = font.render(text, True, color)\r\n text_rect = text_surface.get_rect()\r\n text_rect.midtop = (x, y)\r\n self.screen.blit(text_surface, text_rect)\r\n\r\nif __name__ == \"__main__\":\r\n g = Game()\r\n g.show_start_screen()\r\n while g.running:\r\n g.new()\r\n g.show_go_screen()\r\n\r\n pyg.quit()\r\n","repo_name":"Ronak-kumar/Jumper-Game","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":10691,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"6700695229","text":"import sys\nfrom collections import defaultdict\nimport re\n\ndef initialise():\n with open(sys.argv[1]) as file:\n data = file.readlines()\n\n cuboids = []\n\n for line in data:\n op = line.split(' ')[0]\n ranges = ''.join(line.split(' ')[1:])\n values = [int(x) for x in re.findall(r'-{0,1}\\d{1,10}', ranges)]\n cuboids.append((op, tuple(values)))\n\n return cuboids\n\n\n\ndef apply_cuboids_to_point(coords, cuboids):\n x,y,z = coords\n\n result = False\n\n for cuboid in cuboids:\n op = cuboid[0]\n xmin, xmax, ymin, ymax, zmin, zmax = cuboid[1]\n\n if xmin <= x <= xmax and ymin <= y <= ymax and zmin <= z <= zmax:\n if op == 'on':\n result = True\n else:\n result = False\n \n return result\n\ndef iterate_over_points(cuboids):\n count=0\n \n bigcube = range(-50,51)\n\n for x in bigcube:\n for y in bigcube:\n for z in bigcube:\n if apply_cuboids_to_point((x,y,z), cuboids):\n count += 1\n\n return count\n\ncuboids = initialise()\n\nprint(cuboids)\n\nanswer = iterate_over_points(cuboids)\n\nprint(answer)\n\n\n","repo_name":"DavidBartram/advent-of-code","sub_path":"2021/day22-1.py","file_name":"day22-1.py","file_ext":"py","file_size_in_byte":1162,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"43876909484","text":"#!/usr/bin/python\n'''defines the MalConv architecture.\nAdapted from https://arxiv.org/pdf/1710.09435.pdf\nThings different about our implementation and that of the original paper:\n * The paper uses batch_size = 256 and SGD(lr=0.01, momentum=0.9, decay=UNDISCLOSED, nesterov=True )\n * The paper didn't have a special EOF symbol\n * The paper allowed for up to 2MB malware sizes, we use 1.0MB because of memory on a Titan X\n '''\n\ndef main(): \n from keras.layers import Dense, Conv1D, Activation, GlobalMaxPooling1D, Input, Embedding, Multiply\n from keras.models import Model\n from keras import backend as K\n from keras import metrics\n import multi_gpu\n import os\n import math\n import random\n import argparse\n import os\n import numpy as np\n import requests\n\n batch_size = 100\n input_dim = 257 # every byte plus a special padding symbol\n padding_char = 256\n\n parser = argparse.ArgumentParser()\n parser.add_argument('--gpus', help='number of GPUs', default=1)\n\n args = parser.parse_args()\n ngpus = int(args.gpus)\n\n if os.path.exists('malconv.h5'):\n print(\"restoring malconv.h5 from disk for continuation training...\")\n from keras.models import load_model\n basemodel = load_model('malconv.h5')\n _, maxlen, embedding_size = basemodel.layers[1].output_shape\n input_dim\n else:\n maxlen = 2**20 # 1MB\n embedding_size = 8 \n\n # define model structure\n inp = Input( shape=(maxlen,))\n emb = Embedding( input_dim, embedding_size )( inp )\n filt = Conv1D( filters=128, kernel_size=500, strides=500, use_bias=True, activation='relu', padding='valid' )(emb)\n attn = Conv1D( filters=128, kernel_size=500, strides=500, use_bias=True, activation='sigmoid', padding='valid')(emb)\n gated = Multiply()([filt,attn])\n feat = GlobalMaxPooling1D()( gated )\n dense = Dense(128, activation='relu')(feat)\n outp = Dense(1, activation='sigmoid')(dense)\n\n basemodel = Model( inp, outp )\n\n basemodel.summary() \n\n print(\"Using %i GPUs\" %ngpus)\n\n if ngpus > 1:\n model = multi_gpu.make_parallel(basemodel,ngpus)\n else:\n model = basemodel\n\n from keras.optimizers import SGD\n model.compile( loss='binary_crossentropy', optimizer=SGD(lr=0.01,momentum=0.9,nesterov=True,decay=1e-3), metrics=[metrics.binary_accuracy] )\n\n def bytez_to_numpy(bytez,maxlen):\n b = np.ones( (maxlen,), dtype=np.uint16 )*padding_char\n bytez = np.frombuffer( bytez[:maxlen], dtype=np.uint8 )\n b[:len(bytez)] = bytez\n return b\n\n def getfile_service(sha256,url=None,maxlen=maxlen):\n if url is None:\n raise NotImplementedError(\"You must provide your own url for getting file bytez by sha256\")\n r = requests.get( url, params={'sha256':sha256} )\n if not r.ok:\n return None\n return bytez_to_numpy( r.content, maxlen ) \n\n def generator( hashes, labels, batch_size, shuffle=True ):\n X = []\n y = []\n zipped = list(zip(hashes, labels))\n while True:\n if shuffle:\n random.shuffle( zipped )\n for sha256,l in zipped:\n x = getfile_service(sha256)\n if x is None:\n continue\n X.append( x )\n y.append( l )\n if len(X) == batch_size:\n yield np.asarray(X,dtype=np.uint16), np.asarray(y)\n X = []\n y = []\n\n import pandas as pd\n train_labels = pd.read_csv('ember_training.csv.gz')\n train_labels = train_labels[ train_labels['y'] != -1 ] # get only labeled samples\n labels = train_labels['y'].tolist()\n hashes = train_labels['sha256'].tolist()\n\n from sklearn.model_selection import train_test_split\n hashes_train, hashes_val, labels_train, labels_val = train_test_split( hashes, labels, test_size=200 )\n\n train_gen = generator( hashes_train, labels_train, batch_size )\n val_gen = generator( hashes_val, labels_val, batch_size )\n\n from keras.callbacks import LearningRateScheduler\n\n base = K.get_value( model.optimizer.lr )\n def schedule(epoch):\n return base / 10.0**(epoch//2)\n\n model.fit_generator(\n train_gen,\n steps_per_epoch=len(hashes_train)//batch_size,\n epochs=10,\n validation_data=val_gen,\n callbacks=[ LearningRateScheduler( schedule ) ],\n validation_steps=int(math.ceil(len(hashes_val)/batch_size)),\n )\n\n basemodel.save('malconv.h5')\n\n test_labels = pd.read_csv('ember_test.csv.gz')\n labels_test = test_labels['y'].tolist()\n hashes_test = test_labels['sha256'].tolist()\n\n test_generator = generator(hashes_test,labels_test,batch_size=1,shuffle=False)\n test_p = basemodel.predict_generator( test_generator, steps=len(test_labels), verbose=1 )\n\n\nif __name__ == '__main__':\n print('*'*80)\n print('''\nThis is nonfunctional demonstration code that is provided for convenience. It shows\n- The MalConv structure used in our paper\n- Training procedure used in the paper\n- How to load the weights for the MalConv model that we used.\n\nIt may be made functional by modifying the code to retrieve file contents by sha256\nfrom a user-defined URL.\n\nYou may use the provided weights under the Ember AGPL-3.0 license included in the parent directory.\nWe also ask that you cite the original MalConv paper and refer to the Ember paper as the implementation.\n\n(1) E. Raff, J. Barker, J. Sylvester, R. Brandon, B. Catanzaro, C. Nicholas, \"Malware Detection by Eating a Whole EXE\", in ArXiv e-prints. Oct. 2017.\n\n@ARTICLE{raff2017malware,\n title={Malware detection by eating a whole exe},\n author={Raff, Edward and Barker, Jon and Sylvester, Jared and Brandon, Robert and Catanzaro, Bryan and Nicholas, Charles},\n journal={arXiv preprint arXiv:1710.09435},\n year={2017}\n}\n\n(2) H. Anderson and P. Roth, \"EMBER: An Open Dataset for Training Static PE Malware Machine Learning Models”, in ArXiv e-prints. Apr. 2018.\n\n@ARTICLE{2018arXiv180404637A,\n author = {{Anderson}, H.~S. and {Roth}, P.},\n title = \"{EMBER: An Open Dataset for Training Static PE Malware Machine Learning Models}\",\n journal = {ArXiv e-prints},\n archivePrefix = \"arXiv\",\n eprint = {1804.04637},\n primaryClass = \"cs.CR\",\n keywords = {Computer Science - Cryptography and Security},\n year = 2018,\n month = apr,\n adsurl = {http://adsabs.harvard.edu/abs/2018arXiv180404637A},\n}\n''')\n print('*'*80)\n\n #main() # uncomment this line after fixing the URL NotImplementedError above","repo_name":"elastic/ember","sub_path":"malconv/malconv.py","file_name":"malconv.py","file_ext":"py","file_size_in_byte":6591,"program_lang":"python","lang":"en","doc_type":"code","stars":850,"dataset":"github-code","pt":"79"} +{"seq_id":"74221843775","text":"from collections import deque\n\n\nclass TreeNode:\n\n def __init__(self, value):\n self.value = value\n self.children = []\n\n def __repr__(self):\n return self.value\n\n\ndef print_tree(root):\n\n stack = deque()\n stack.append([root, 0])\n level_str = \"\\n\"\n\n while len(stack) > 0:\n node, level = stack.pop()\n if level > 0 and len(stack) > 0 and level <= stack[-1][1]:\n level_str += \" \"*(level-1) + \"├─\"\n elif level > 0:\n level_str += \" \"*(level-1) + \"└─\"\n level_str += str(node.value)\n level_str += \"\\n\"\n level += 1\n for child in node.children:\n stack.append([child, level])\n\n print(level_str)\n\n\nsample_root_node = TreeNode(\"A\")\ntwo = TreeNode(\"B\")\nthree = TreeNode(\"C\")\nsample_root_node.children = [three, two]\nfour = TreeNode(\"D\")\nfive = TreeNode(\"E\")\nsix = TreeNode(\"F\")\nseven = TreeNode(\"G\")\ntwo.children = [five, four]\nthree.children = [seven, six]\n\n\nprint_tree(sample_root_node)\n","repo_name":"dom-wright/data_structures_algorithms","sub_path":"algorithms/search/tree/tree_print_path.py","file_name":"tree_print_path.py","file_ext":"py","file_size_in_byte":1007,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"70417724095","text":"\"\"\"модуль для управления состоянием профиля пользователя\"\"\"\r\nimport broker\r\nimport db_requests\r\nfrom decimal import Decimal\r\n\r\nfrom sqlite3 import IntegrityError\r\n\r\nimport exchange\r\nfrom custom_types import AssetPrice, Order\r\nimport exceptions\r\n\r\n\r\ndef onboarding(user_id: int) -> bool:\r\n \"\"\"Онбординг\"\"\"\r\n try:\r\n db_requests.insert(\"users\", {\"user_id\": user_id})\r\n return True\r\n except IntegrityError:\r\n return False\r\n\r\n\r\ndef get_balance(user_id: int) -> Decimal:\r\n \"\"\"Получить сведения о балансе\"\"\"\r\n try:\r\n balance_dict = db_requests.select(\"users\", [\"balance\"], user_id)\r\n balance = Decimal(balance_dict['balance'])\r\n return balance\r\n except KeyError:\r\n db_requests.insert(\"users\", {\"user_id\": user_id})\r\n balance = Decimal(0)\r\n return balance\r\n\r\n\r\ndef get_lots(user_id: int, asset_name: str) -> int:\r\n try:\r\n pcs = db_requests.select(\"users\", [asset_name], user_id)[asset_name]\r\n lots = pcs / AssetPrice.lot_weight.value\r\n return lots\r\n except KeyError:\r\n db_requests.insert(\"users\", {\"user_id\": user_id})\r\n return 0\r\n\r\n\r\ndef change_balance(user_id: int, money: Decimal) -> Decimal:\r\n \"\"\"Изменить баланс\"\"\"\r\n try:\r\n balance_dict = db_requests.select(\"users\", [\"balance\"], user_id)\r\n balance = Decimal(balance_dict['balance'])\r\n new_balance = balance + money\r\n db_requests.update_balance(user_id, str(new_balance))\r\n return new_balance\r\n except KeyError:\r\n db_requests.insert(\"users\", {\"user_id\": user_id})\r\n db_requests.update_balance(user_id, str(money))\r\n return money\r\n\r\n\r\ndef change_asset_pcs(order: Order, difference: int) -> None:\r\n \"\"\"Изменить кол-во акций\"\"\"\r\n try:\r\n asset_dict = db_requests.select(\"users\", [order.asset_name], order.user_id)\r\n asset_pcs = asset_dict[order.asset_name]\r\n new_pcs = asset_pcs + difference\r\n db_requests.update_asset_pcs(order.user_id, order.asset_name, new_pcs)\r\n db_requests.update_status(order.order_id)\r\n except KeyError:\r\n db_requests.insert(\"users\", {\"user_id\": order.user_id})\r\n db_requests.update_asset_pcs(order.user_id, order.asset_name, difference)\r\n\r\n\r\ndef get_portfolio(user_id: int) -> dict:\r\n \"\"\"Получить сведения о портфеле\"\"\"\r\n asset_list = AssetPrice.ASSETS.value\r\n portfolio = db_requests.select(\"users\", asset_list, user_id)\r\n if len(portfolio) != 0:\r\n value = Decimal(0)\r\n for asset_name in portfolio.keys():\r\n value += AssetPrice[asset_name].value * portfolio[asset_name]\r\n portfolio[\"value\"] = value\r\n return portfolio\r\n else:\r\n db_requests.insert(\"users\", {\"user_id\": user_id})\r\n portfolio = db_requests.select(\"users\", asset_list, user_id)\r\n portfolio[\"value\"] = Decimal(0)\r\n return portfolio\r\n\r\n\r\ndef parse_order(order: str) -> dict:\r\n \"\"\" Распарсить запрос из телеграмма\"\"\"\r\n parsed_order = {}\r\n parsed_request = order.split()\r\n if len(parsed_request) == 2:\r\n if parsed_request[0] in AssetPrice.ASSETS.value:\r\n parsed_order['asset'] = parsed_request[0]\r\n else:\r\n raise exceptions.NotCorrectMessage(f\"Неверный формат. Проверьте написание названия акции.\\n\"\r\n f\"Доустпный список акций: {' '.join(AssetPrice.ASSETS.value)}\"\r\n )\r\n try:\r\n parsed_order['lots'] = int(parsed_request[1])\r\n except ValueError:\r\n raise exceptions.NotCorrectMessage(\"Неверный формат. Проверьте написание кол-ва лотов. \")\r\n if parsed_order['lots'] <= 0:\r\n raise exceptions.NotCorrectMessage(f\"Неверный формат. Проверьте написание кол-ва лотов.\\n\"\r\n f\"Число должно быть больше, чем 0\"\r\n )\r\n return parsed_order\r\n else:\r\n raise exceptions.NotCorrectMessage(f\"Неверный формат. Проверьте написание команды.\\n\"\r\n f\"Команды:\\n\"\r\n f\"/buy asset lots\\n\"\r\n f\"/sell asset lots\\n\"\r\n f\"Где, asset -- название актива, lots -- кол-во лотов (лот = 10 шт.)\"\r\n )\r\n\r\n\r\ndef sell_assets(order: Order) -> bool:\r\n \"\"\"Создать заявку на покупку акций\"\"\"\r\n lots = get_lots(order.user_id, order.asset_name)\r\n if broker.check_assets(order, lots):\r\n broker.create_sell_order(order)\r\n exchange.close_the_deal(order)\r\n else:\r\n raise exceptions.BrokerException(f\"Не хватает акций.\\n\"\r\n f\"Купите пакеты акций: /buy {order.asset_name} {order.lots}\")\r\n return True\r\n\r\n\r\ndef buy_assets(order: Order) -> bool:\r\n \"\"\"Создать заявку на покупку акций\"\"\"\r\n balance = get_balance(order.user_id)\r\n if broker.check_balance(order, balance):\r\n broker.create_buy_order(order)\r\n exchange.close_the_deal(order)\r\n else:\r\n raise exceptions.BrokerException(f\"Не хватает средств.\\n\"\r\n f\"Пополните баланс командой: /change_balance {order.price}\")\r\n return True\r\n\r\n\r\ndef get_my_orders(user_id: int) -> str:\r\n \"\"\"получить все заявки от профиля по id\"\"\"\r\n my_orders = db_requests.fetchall_by_id(\"reg_orders\", [\"type\", \"asset\", \"price\", \"status\", \"date\"],\r\n user_id)\r\n message = ''\r\n if len(my_orders) == 0:\r\n message = 'У профиля нет заявок'\r\n return message\r\n else:\r\n for row in my_orders:\r\n for key in row.keys():\r\n message += str(row[key]) + ' '\r\n message += '\\n'\r\n return message\r\n","repo_name":"yaroslavcherkaev/sbercontest","sub_path":"profile.py","file_name":"profile.py","file_ext":"py","file_size_in_byte":6470,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"50450438273","text":"import httplib2\nfrom apiclient import discovery\nfrom oauth2client.service_account import ServiceAccountCredentials\nimport settings\n\nhttpAuth = settings.credentials.authorize(httplib2.Http())\nservice = discovery.build('sheets', 'v4', http=httpAuth)\n\nres = service.spreadsheets().values().batchUpdate(\n spreadsheetId=settings.spreadsheetId,\n body={\n 'valueInputOption': 'USER_ENTERED',\n 'data': [\n {\n 'range': 'First list!B2:D5',\n 'majorDimension': 'ROWS',\n 'values': [\n ['Test B', 'Test C', 'Test D'],\n ['24', '5', '=B3/C3']\n ]\n }\n ]\n }\n).execute()\n","repo_name":"TRTSS/SteamParserV1","sub_path":"editSheet.py","file_name":"editSheet.py","file_ext":"py","file_size_in_byte":698,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"70965751616","text":"import string\nfrom nltk.corpus import stopwords\nfrom collections import Counter\n\ndef strip_punct(s):\n while True:\n if s:\n if s[-1] in string.punctuation: s = s[:-1]\n elif s[0] in string.punctuation: s = s[1:]\n else: return s\n else: return ''\n\n\ndef add_space(s):\n if not s: return ''\n result = s[0]\n prev_c = s[0]\n for c in s[1:]:\n if prev_c in string.punctuation:\n result += ' '\n result += c\n else:\n if c in string.punctuation:\n result += ' '\n result += c\n else:\n result += c\n prev_c = c\n result = ' '.join(result.split())\n return result\n\n\ndef normalize_phrase(phrase, subq_config, ignore=[]):\n # Rule 1: split composite mentions based on 'and'\n if subq_config['split_composite_mention']:\n phrases = [p.strip() for p in phrase.split(' and ')]\n else:\n phrases = [phrase.strip()]\n\n # Rule 2: remove punctuations\n if subq_config['remove_punctuation']:\n phrases = [strip_punct(p).strip() for p in phrases if strip_punct(p).strip() != '']\n \n # Rule 3: remove lowercase phrases\n if subq_config['remove_lowercase_phrase']:\n phrases = [p for p in phrases if p != p.lower()]\n\n # Rule 4: remove the def 'the'\n if subq_config['remove_the']:\n phrases = [' '.join(p.split()[1:]) if p.split()[0].lower() == 'the' else p for p in phrases]\n\n # Rule 5: remove short phrase\n phrases = [p for p in phrases if len(p) >= subq_config['min_phrase_length']]\n \n # Rule 6: remove stopword\n if subq_config['remove_stopword']:\n phrases = [p for p in phrases if p.lower() not in stopwords.words('english')]\n \n # Rule 7: remove subtype\n if subq_config['remove_subtype']:\n phrases = [p for p in phrases if p.lower() not in ignore]\n \n # find the start positions of phrases\n if len(phrases) > 0:\n starts = [phrase.find(p) for p in phrases]\n else:\n starts = None\n\n return phrases, starts\n\n\ndef postproc_retrieved_data(retrieved_data, nlp, subq_config, ignore=[]):\n # This code normalizes retrieved phrases, detects abbreviations, and merges duplicate sentences\n \n processed_data = {}\n\n for json_example in retrieved_data:\n sent = json_example['sentence']\n pos = json_example['pos']\n phrase = json_example['phrase']\n\n if sent not in processed_data.keys():\n processed_data[sent] = {\n 'sentence': sent,\n 'poss': [pos],\n 'phrases': [phrase]\n }\n else :\n processed_data[sent]['poss'].append(pos)\n processed_data[sent]['phrases'].append(phrase)\n\n for sent in processed_data.keys():\n json_example = processed_data[sent]\n \n abbreviations = {}\n doc = nlp(json_example['sentence'])\n for abrv in doc._.abbreviations:\n if abrv.text.lower() not in ignore :\n abbreviations[abrv._.long_form.text] = [abrv.text, (abrv.start_char, abrv.end_char)]\n\n processed_data[sent]['abbreviations'] = abbreviations\n\n phrases = []\n pos_mark = [0 for _ in range(len(json_example['sentence']))]\n for phrase, pos in zip(json_example['phrases'], json_example['poss']):\n normalized_names, starts = normalize_phrase(phrase, subq_config, ignore)\n\n if normalized_names == []:\n pass\n else:\n # If there are conflicts between spans within duplicate sentences, the higher-ranked span is selected.\n for nn, st in zip(normalized_names, starts):\n start_idx = pos[0] + st\n\n if nn != json_example['sentence'][start_idx:start_idx+len(nn)]:\n continue\n\n if sum(pos_mark[start_idx:start_idx+len(nn)]) == 0:\n phrases.append({\n 'name': nn,\n 'start': start_idx,\n 'end': start_idx+len(nn)\n })\n\n for i in range(start_idx, start_idx+len(nn)):\n pos_mark[i] = 1\n \n processed_data[sent]['phrases'] = phrases\n del processed_data[sent]['poss']\n \n return processed_data\n\n\ndef preproc_dictionary(dictionary, min_phrase_frequency=0, lowercase_matching=True):\n # This code pre-processes a raw pseudo-dictionary for dictionary matching.\n type_counter = {}\n\n phrases = [p.split(\"\\t\")[0].strip() for p in dictionary]\n\n # add spaces before and after punctuations\n phrases = [add_space(p) for p in phrases]\n\n if lowercase_matching:\n phrases = [p.lower() for p in phrases]\n\n # type of dictionary\n t = dictionary[0].split('\\t')[1].strip()\n for p in phrases:\n if p not in type_counter.keys(): type_counter[p] = Counter()\n type_counter[p][t] += 1\n\n # Remove duplicates\n phrases = list(set(phrases))\n\n processed_phrases = []\n\n cnt = 0\n for p in phrases:\n if max(type_counter[p].values()) > min_phrase_frequency:\n processed_phrases.append(p)\n cnt += 1\n\n return processed_phrases, type_counter\n\n\n","repo_name":"dmis-lab/GeNER","sub_path":"utils/norm_utils.py","file_name":"norm_utils.py","file_ext":"py","file_size_in_byte":5294,"program_lang":"python","lang":"en","doc_type":"code","stars":72,"dataset":"github-code","pt":"79"} +{"seq_id":"16370360314","text":"\"\"\"In Chapter 4 there was a problem that asked you to write a program that lets the user play\nRock-Paper-Scissors against the computer. In that program there were exactly five rounds.\nRewrite the program so that it is a best 3 out of 5. That is, the first player to win three times is\nthe winner.\n\"\"\"\n\nlist=['scissors','rock','paper']\nimport random\n\nyou_win=0\ncomputer_win=0\ni=0\n\nwhile i<5:\n if you_win==3 or computer_win==3:\n \n if you_win==3:\n print(\"you win :)\")\n else:\n print(\"you lose :(\")\n break\n\n computer=random.choice(list)\n print(\"round \",i+1,\":\")\n yourturn=input(\"let's play: \")\n \n while(yourturn not in list):\n yourturn=input(\"oops! play again: \")\n \n else:\n if yourturn==computer:\n print(\"computer choice:\",computer)\n print(\"noone win this round\")\n\n elif((yourturn==list[0] and computer==list[2]) or (yourturn==list[1] and computer==list[0])or(yourturn==list[2] and computer==list[1])): \n print(\"computer choice:\",computer)\n you_win+=1\n print(\"you win this round\")\n \n else:\n print(\"computer choice:\",computer)\n computer_win+=1\n print(\"you lose this round\")\n\n i+=1 \n \nelse:\n print(\"Final:\")\n if you_win>computer_win:\n print(\"you WIN :)\")\n elif computer_win>you_win:\n print(\"you LOSE :(\")\n else:\n print(\"Noone win\")\n","repo_name":"monkeybuzinis/Python","sub_path":"9.while loop/12.py","file_name":"12.py","file_ext":"py","file_size_in_byte":1498,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"79"} +{"seq_id":"6294152925","text":"# Criar uma tabela de acordo com as informações do ranking dos players da NBA stats.nba.com\n# Utilizar libs como requests2, pandas, beautifulsoup4 e selenium\n# Aprender webscraping básico.\n\n# Bibliotécas\nimport time\nimport requests\nimport pandas as pd\nfrom bs4 import BeautifulSoup\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.common.keys import Keys\nfrom webdriver_manager.chrome import ChromeDriverManager\nfrom selenium.common.exceptions import NoSuchElementException \nimport json\n\n# navegar até o site\nurl = \"https://stats.nba.com/players/traditional/?SeasonType=Playoffs&sort=PLAYER_NAME&dir=1\"\ndriver = webdriver.Chrome(ChromeDriverManager().install())\ndriver.get(url)\ntime.sleep(3)\n\n#POPUP do site\npopup = driver.find_element_by_xpath(\"//button[@id='onetrust-accept-btn-handler']\")\nif popup:\n popup.click()\n# ordenar por categorias\ntop10ranking = {}\nrankings = {\n '3points': {'field': 'FG3M', 'label': '3PM'},\n 'points': {'field': 'PTS', 'label': 'PTS'},\n 'assistants': {'field': 'AST', 'label': 'AST'},\n 'rebounds': {'field': 'REB', 'label': 'REB'},\n 'steals': {'field': 'STL', 'label': 'STL'},\n 'blocks': {'field': 'BLK', 'label': 'BLK'},\n}\n\n#Função da tabela\ndef buildrank(type):\n field = rankings[type]['field']\n label = rankings[type]['label']\n \n driver.find_element_by_xpath(f\"//div[@class='nba-stat-table']//table//thead//tr//th[@data-field='{field}']\").click()\n element = driver.find_element_by_xpath(\"//div[@class='nba-stat-table']//table\")\n html_content = element.get_attribute('outerHTML')\n # Parsear o conteúdo com BeautifulSoup\n soup = BeautifulSoup(html_content, 'html.parser')\n table = soup.find(name='table')\n # Estruturar com o Panda\n df_full = pd.read_html(str(table))[0].head(10)\n df = df_full[[\"Unnamed: 0\", \"PLAYER\", \"TEAM\", label]]\n df.columns = ['pos', 'player', 'team', 'total']\n # Transformar os Dados em um dicionário de dados próprio\n return df.to_dict('records')\n\ntop10ranking['points'] = buildrank('points')\n\nfor i in rankings:\n top10ranking[i] = buildrank(i)\n\n#Fechar bot\ndriver.quit()\n\n# Converter e salvar em um arquivo JSON\njs = json.dumps(top10ranking)\nfp = open('ranking.json', 'w')\nfp.write(js)\nfp.close()","repo_name":"RaphaelR89/Webscrapping","sub_path":"NBAranking/nbaranking.py","file_name":"nbaranking.py","file_ext":"py","file_size_in_byte":2283,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"24113606082","text":"from tkinter import *\n\nmain_window = Tk()\nmain_window.title('Food Calculator')\nmain_window.geometry('300x200')\nmain_window.resizable(width=FALSE, height=FALSE)\nmain_window_bg_colour = 'gray77' # Set this as a variable that can be used across the code\nmain_window.configure(bg=main_window_bg_colour)\n\n\nmain_label = Label(main_window, text=\"Choose your fruit\")\nmain_label.place(relx=0.5, y=15, anchor=CENTER)\nmain_label.configure(bg=main_window_bg_colour)\n\nradio_button_variable = IntVar() # Define variable as integer\nprice_result = IntVar() # Define variable as integer\n\n\ndef calculate_price():\n radio_button_value = int(radio_button_variable.get()) # Make sure this variable is an integer\n input_weight_value = int(input_weight_entry.get()) # Make sure this variable is an integer\n price = radio_button_value * input_weight_value # Make the calculations and assign the value to a variable\n price_result.set(price) # Set the \"price_result\" variable as the value of the \"price\" variable\n\n # price_result.set(int(radio_button_variable.get()) * int(input_weight_entry.get())) # Same as the above in one line\n\n\nradio_button_apple = Radiobutton(main_window, text=\"Apple\", variable=radio_button_variable, value=10)\nradio_button_apple.place(x=5, y=40)\nradio_button_apple.configure(bg=main_window_bg_colour)\n\nradio_button_banana = Radiobutton(main_window, text=\"Banana\", variable=radio_button_variable, value=11)\nradio_button_banana.place(x=5, y=60)\nradio_button_banana.configure(bg=main_window_bg_colour)\n\nradio_button_orange = Radiobutton(main_window, text=\"Grapes\", variable=radio_button_variable, value=15)\nradio_button_orange.place(x=5, y=80)\nradio_button_orange.configure(bg=main_window_bg_colour)\n\nradio_button_apple = Radiobutton(main_window, text=\"Kiwi\", variable=radio_button_variable, value=12)\nradio_button_apple.place(x=100, y=40)\nradio_button_apple.configure(bg=main_window_bg_colour)\n\nradio_button_banana = Radiobutton(main_window, text=\"Pear\", variable=radio_button_variable, value=8)\nradio_button_banana.place(x=100, y=60)\nradio_button_banana.configure(bg=main_window_bg_colour)\n\nradio_button_orange = Radiobutton(main_window, text=\"Orange\", variable=radio_button_variable, value=5)\nradio_button_orange.place(x=100, y=80)\nradio_button_orange.configure(bg=main_window_bg_colour)\n\n\ninput_weight_entry = Entry(main_window, width=10) # Set this variable as an \"entry\" type one in the window called \"main_window\"\ninput_weight_entry.place(x=100, y=130)\n\ninput_weight_entry_label = Label(main_window, text=\"Weight (kg): \")\ninput_weight_entry_label.place(x=5, y=132)\ninput_weight_entry_label.configure(bg=main_window_bg_colour)\n\ncalculate_price_entry = Entry(main_window, width=10, textvariable=price_result) # Output the \"price_result\" variable value\ncalculate_price_entry.place(x=100, y=160)\n\ncalculate_price_button = Button(main_window, text='Calc', command= lambda : calculate_price()) # When clicked run the \"calculate_price\" function\ncalculate_price_button.place(x=50, y=162)\ncalculate_price_button.configure(highlightbackground=main_window_bg_colour)\n\nmain_window.mainloop()","repo_name":"TheFREAK69/Udemy-Master_Tkinter_By_Building_5_Fully_Functioning_Apps","sub_path":"Project 1 - Food Calculator.py","file_name":"Project 1 - Food Calculator.py","file_ext":"py","file_size_in_byte":3097,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"39711700128","text":"# -*- coding: utf-8 -*-\nimport time\n\nfrom selenium import webdriver\n\nfrom share import unittest123321\n\n\nclass WebdriverDemo(unittest123321.TestCase):\n def setUp(self):\n self.driver = webdriver.Firefox()\n self.driver.implicitly_wait(30)\n self.base_url = \"http://test.paymax.cc\"\n self.verificationErrors = []\n self.accept_next_alert = True\n\n def test_webdriver_demo(self):\n driver = self.driver\n driver.maximize_window()\n\n '''访问地址'''\n driver.get(self.base_url + \"/\")\n #通过CSS路径定位:\n # driver.find_element_by_css_selector(\"a[class='button blue cta']\").click()\n #通过XPATH路径定位\n # driver.find_element_by_xpath(\".//*[@id='bgstylec']/div[2]/div/p/a\").click()\n # driver.find_element_by_css_selector(\"button[class='btn btn-primary none']\").click() #登录\n driver.find_element_by_css_selector(\"button[class='btn btn-success']\").click()\n\n time.sleep(3)\n\n def close_alert_and_get_its_text(self):\n try:\n alert = self.driver.switch_to_alert()\n alert_text = alert.text\n if self.accept_next_alert:\n alert.accept()\n else:\n alert.dismiss()\n return alert_text\n finally:\n self.accept_next_alert = True\n\n def tearDown(self):\n self.driver.quit()\n self.assertEqual([], self.verificationErrors)\n\n\nif __name__ == \"__main__\":\n unittest123321.main()","repo_name":"hivencent/paymax-test","sub_path":"demo/share_2_front_end.py","file_name":"share_2_front_end.py","file_ext":"py","file_size_in_byte":1504,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"31568591310","text":"from django.shortcuts import render, redirect\nfrom orders.forms import PizzaForm, SubForm, PastaForm, SaladForm, DinnerPlatterForm\nfrom orders.models import Pizza, Sub, Pasta, Salad, DinnerPlatter\nfrom cart.models import CartItem, OrderItem\nfrom django.db.models import Sum\nfrom decimal import Decimal\n\n\n# Create your views here.\ndef cart(request):\n if request.method == \"POST\":\n if request.POST['menu_item'] == 'Pizza':\n form = PizzaForm(request.POST)\n\n if form.is_valid():\n form = form.cleaned_data\n # Get the form data to build a Pizza and CartItem\n menu = \"Pizza\"\n style = form['style']\n size = form['size']\n is_special = False\n if 'is_special' in request.POST:\n is_special = True\n num_toppings = form['num_toppings']\n topping_list = []\n for topping in form['toppings']:\n topping_list.append(str(topping))\n toppings = \", \".join(topping_list)\n\n # Lookup price\n if is_special:\n # numtoppings doesn't matter\n menu_pizza = Pizza.objects.get(style=style, size=size, is_special=is_special)\n else:\n menu_pizza = Pizza.objects.get(style=style, size=size, num_toppings=num_toppings, is_special=is_special)\n price = menu_pizza.price\n print(price)\n\n # Create the Cart Item\n order = CartItem(menu=menu, size=size, style=style, additional=toppings, is_special=is_special, user_id=request.user.id, price=price)\n order.save()\n\n elif request.POST['menu_item'] == 'Sub':\n form = SubForm(request.POST)\n if form.is_valid():\n form = form.cleaned_data\n menu = 'Sub'\n ingredients = form['ingredients']\n size = form['size']\n extras_list = []\n added_cost = Decimal(0.00)\n for extra in form['extras']:\n extras_list.append(str(extra))\n added_cost += extra.added_cost\n\n extras = \", \".join(extras_list)\n\n # Lookup price\n menu_sub = Sub.objects.get(ingredients=ingredients, size=size)\n price = menu_sub.price + Decimal(added_cost)\n\n # Create the Cart Item\n order = CartItem(menu=menu, size=size, style=ingredients, additional=extras, is_special=False, user_id=request.user.id, price=price)\n order.save()\n\n redirect('index')\n elif request.POST['menu_item'] == 'Pasta':\n form = PastaForm(request.POST)\n if form.is_valid():\n form = form.cleaned_data\n menu = 'Pasta'\n style = form['style']\n\n # Lookup price\n menu_pasta = Pasta.objects.get(style=style)\n price = menu_pasta.price\n\n # Create the Cart Item\n order = CartItem(menu=menu, style=style, user_id=request.user.id, price=price)\n order.save()\n elif request.POST['menu_item'] == 'Salad':\n form = SaladForm(request.POST)\n if form.is_valid():\n form = form.cleaned_data\n menu = 'Salad'\n style = form['style']\n\n # Lookup price\n menu_salad = Salad.objects.get(style=style)\n price = menu_salad.price\n\n # Create the Cart Item\n order = CartItem(menu=menu, style=style, user_id=request.user.id, price=price)\n order.save()\n elif request.POST['menu_item'] == 'DinnerPlatter':\n form = DinnerPlatterForm(request.POST)\n if form.is_valid():\n form = form.cleaned_data\n menu = 'DinnerPlatter'\n style = form['style']\n size = form['size']\n\n # Lookup price\n menu_dinner_platter = DinnerPlatter.objects.get(style=style, size=size)\n price = menu_dinner_platter.price\n\n # Create the Cart Item\n order = CartItem(menu=menu, style=style, size=size, user_id=request.user.id, price=price)\n order.save()\n else:\n print(\"Post Error\")\n redirect('index')\n\n cart_items = get_cart_items(request)\n if cart_items:\n total_cost = Decimal(cart_items.aggregate(Sum('price'))['price__sum'])\n else:\n total_cost = 0\n\n context = {\n 'cart_items': cart_items,\n 'num_cart_items': cart_items.count(),\n 'total_cost': total_cost\n }\n\n return render(request, 'cart/cart.html', context)\n\n else:\n\n # Display Shopping Cart for GET request\n cart_items = get_cart_items(request)\n if cart_items:\n total_cost = Decimal(cart_items.aggregate(Sum('price'))['price__sum'])\n else:\n total_cost = 0\n\n context = {\n 'cart_items': cart_items,\n 'num_cart_items': cart_items.count(),\n 'total_cost': total_cost\n }\n return render(request, 'cart/cart.html', context)\n\n\ndef checkout(request):\n if request.method == \"POST\":\n\n cart_items = get_cart_items(request)\n\n for item in cart_items:\n menu = item.menu\n size = item.size\n style = item.style\n additional = item.additional\n is_special = item.is_special\n price = item.price\n user_id = item.user_id\n order = OrderItem(menu=menu, size=size, style=style, additional=additional, is_special=is_special, price=price, user_id=user_id)\n order.save()\n item.delete()\n\n all_order_items = OrderItem.objects.order_by('is_complete', 'created_at')\n\n user_order_items = all_order_items.filter(user_id=request.user.id)\n cart_items = get_cart_items(request)\n\n context = {\n 'cart_items': cart_items,\n 'all_order_items': all_order_items,\n 'user_order_items': user_order_items,\n 'num_cart_items': cart_items.count(),\n }\n\n return render(request, 'cart/checkout.html', context)\n else:\n all_order_items = OrderItem.objects.order_by('is_complete', 'created_at')\n\n user_order_items = all_order_items.filter(user_id=request.user.id)\n cart_items = get_cart_items(request)\n\n context = {\n 'cart_items': cart_items,\n 'all_order_items': all_order_items,\n 'user_order_items': user_order_items,\n 'num_cart_items': cart_items.count(),\n }\n\n return render(request, 'cart/checkout.html', context)\n\n\ndef remove(request, cart_item_id):\n cart_item = CartItem.objects.filter(id=cart_item_id)\n cart_item.delete()\n return redirect('cart')\n\n\ndef mark_complete(request, order_item_id):\n order_item = OrderItem.objects.filter(id=order_item_id)\n order_item.update(is_complete=True)\n return redirect('checkout')\n\n\n# Helper Funcitons\ndef get_cart_items(request):\n cart_items = CartItem.objects.filter(user_id=request.user.id)\n return cart_items\n","repo_name":"learyjk/pizza-cs50","sub_path":"cart/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":7359,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"50115573623","text":"#!/usr/bin/env python3\nimport requests\nimport json\n\nSEARCH_PLAYLIST_ENDPOINT ='https://api.spotify.com/v1/search?type=track'\nPLAYBACK_ENDPOINT = 'https://api.spotify.com/v1/me/player/queue'\n\ndef searchForTrack(query, token): #returns a dictionary of responses\n searchParams = {'type': 'track'}\n searchParams['q'] = query\n searchParams['market'] = 'US'\n searchParams['limit'] = 50\n myheaders = {\"Accept\":\"application/json\"}\n myheaders[\"Content-Type\"] = \"application/json\"\n myheaders[\"Authorization\"] = \"Bearer \" + token\n resp = requests.get(SEARCH_PLAYLIST_ENDPOINT, params=searchParams, headers=myheaders)\n #writeToFile(json.dumps(resp.json()))\n return json.dumps(resp.json())\n\ndef doTheThing(uri, token):\n myheaders = {\"Accept\":\"application/json\"}\n myheaders[\"Content-Type\"] = \"application/json\"\n myheaders[\"Authorization\"] = \"Bearer \" + token\n searchParams = {'uri':uri}\n resp = requests.post(PLAYBACK_ENDPOINT, params=searchParams, headers=myheaders)\n return resp\n\ndef testQueue():\n doTheThing('spotify:track:3okk47CKOqAm1TXmVPzNYf')\n\ndef testSearch():\n testQ = \"Something About Us Daft Punk\"\n tok = 'BQArMtuw7RqCYEOG4ZIO2EnsQJV9_l8gpWl9CTWeQTB9ClYc5VDasCf7Kag_kODpCK3s1OoXXIePzJ15WjyVyPmKLlQtuy_vRHt2dpObg35AijY-N77yj71nuP7Y3eDTvLpAy20vA8zCzDMAOYR2LHF6EAOWygx_GBwpGMJRUD7N1zDM3wSXIpAGmP4'\n log = open('testLog.txt', 'w')\n response = searchForTrack(testQ, tok)\n log.close()\n print(response)\n\ndef writeToFile(dict):\n f = open('httpResponse.json', 'w')\n f.write(dict)\n f.close()","repo_name":"scaboodles/dj_hotkey","sub_path":"query.py","file_name":"query.py","file_ext":"py","file_size_in_byte":1557,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"74346679294","text":"from tkinter import *\nfrom pojeto.banco.conexao import Conexao\nfrom pojeto.controler.cadastrarVC import CadastrarVC\nfrom functools import partial\n\n\n\n\nclass Vendas:\n def abrirVendas(self):\n janela = Tk() # Cria a tela principal\n janela.resizable(0, 0)\n quadro = Frame(janela, bd=2, relief=\"raised\", pady=10, padx=10)\n quadro.place(x=20, y=20) # posiciona o quadro na posição xy\n fonte = ('Arial', '16', 'bold') # Define o padrão de fonte\n\n\n def btnVendas():\n cadastrar = CadastrarVC(edProduto.get(), edVender.get(), edInvisivel.get(), edFranquia.get())\n cadastrado = cadastrar.CadastrarVenda()\n return cadastrado\n\n\n#FORMULARIO\n\n\n lbProduto = Label(quadro, text=\"Produto:\", font=fonte, pady=10, padx=10)\n lbVender = Label(quadro, text=\"Vender:\", font=fonte, pady=10, padx=10)\n lbFranquia = Label(quadro, text=\"Franquia:\", font=fonte, pady=10, padx=10)\n edProduto = Entry(quadro, font=fonte, width=25)\n edInvisivel = Entry(quadro, font=fonte, width=25, text='7')\n edVender = Entry(quadro, font=fonte, width=25)\n edFranquia = Entry(quadro, font=fonte, width=25)\n\n\n btVenda = Button(quadro, font=fonte, text=\"Vender\", fg=\"green\",\n activebackground=\"#A9A9A9\", activeforeground=\"white\")\n btVenda['command'] = partial(btnVendas)\n\n\n #EXIBINDO OS ITENS\n\n\n lbProduto.grid(row=0, column=0)\n edProduto.grid(row=0, column=1)\n lbVender.grid(row=2, column=0)\n edVender.grid(row=2, column=1)\n lbFranquia.grid(row=3, column=0)\n edFranquia.grid(row=3, column=1)\n\n\n\n btVenda.grid(row=5, column=1)\n\n\n\n\n #photoLupa = PhotoImage(file=\"img/lupa2.png\")\n #logoLupa = photoLupa.subsample(15, 15)\n btBusca = Button(quadro, bd=1) #image=logoLupa\n btBusca.grid(row=0, column=3)\n\n # criando lb mensagem\n lbMensagem = Label(janela, text=\"\", font=fonte)\n lbMensagem.place(x=20, y=280)\n\n\n janela.geometry(\"640x320+200+200\") # Larg x Alt + DistaciaEsq + DistandiaTop\n janela.title(\"Consulta, Cadastro e Alteração\") # Define o titulo da janela\n #janela.iconbitmap(\"img/icone.ico\")\n janela.mainloop() # Exibe a janela\n\n","repo_name":"LuanKennedy/Python","sub_path":"pojeto/view/venda.py","file_name":"venda.py","file_ext":"py","file_size_in_byte":2274,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"29465815620","text":"import pygame\nfrom tank import Tank\n\nclass Player(Tank):\n\tdef __init__(self, game, level, position=None, direction=None, filename=None):\n\t\tsuper().__init__(game, level, side=0, position=None, direction=None, filename=None)\n\t\tself.start_position = position\n\t\tself.start_direction = direction\n\t\tself.lives = 3\n\t\tself.score = 0 # total score\n\n\t\t# store how many bonuses in this stage this player has collected 计分板\n\t\tself.trophies = {\n\t\t\t\"bonus\" : 0,\n\t\t\t\"enemy0\" : 0,\n\t\t\t\"enemy1\" : 0,\n\t\t\t\"enemy2\" : 0,\n\t\t\t\"enemy3\" : 0\n\t\t}\n\t\t\n\t\tif filename == None:\n\t\t\tfilename = (0, 0, 16*2, 16*2)\n\t\tself.image = self.game.sprites.subsurface(filename)\n\t\tself.image_up = self.image\n\t\tself.image_left = pygame.transform.rotate(self.image, 90)\n\t\tself.image_down = pygame.transform.rotate(self.image, 180)\n\t\tself.image_right = pygame.transform.rotate(self.image, 270)\n\n\t\tif direction == None:\n\t\t\tself.rotate(self.DIR_UP, False)\n\t\telse:\n\t\t\tself.rotate(direction, False)\n\n\tdef move(self, direction):\n\t\t\"\"\" move player if possible \"\"\"\n\t\tif self.state == self.STATE_EXPLODING:\n\t\t\tif not self.explosion.active:\n\t\t\t\tself.state = self.STATE_DEAD\n\t\t\t\tdel self.explosion\n\n\t\tif self.state != self.STATE_ALIVE:\n\t\t\treturn\n\n\t\tif self.direction != direction: # rotate player\n\t\t\tself.rotate(direction)\n\n\t\tif self.paralised: # 瘫痪\n\t\t\treturn\n\n\t\t# speed: px per move, default = 2\n\t\tif direction == self.DIR_UP: # move player\n\t\t\tnew_position = [self.rect.left, self.rect.top - self.speed]\n\t\t\tif new_position[1] < 0:\n\t\t\t\treturn\n\t\telif direction == self.DIR_RIGHT:\n\t\t\tnew_position = [self.rect.left + self.speed, self.rect.top]\n\t\t\tif new_position[0] > (416 - 26):\n\t\t\t\treturn\n\t\telif direction == self.DIR_DOWN:\n\t\t\tnew_position = [self.rect.left, self.rect.top + self.speed]\n\t\t\tif new_position[1] > (416 - 26):\n\t\t\t\treturn\n\t\telif direction == self.DIR_LEFT:\n\t\t\tnew_position = [self.rect.left - self.speed, self.rect.top]\n\t\t\tif new_position[0] < 0:\n\t\t\t\treturn\n\n\t\tplayer_rect = pygame.Rect(new_position, [26, 26])\n\n\t\t# collisions with tiles\n\t\tif player_rect.collidelist(self.level.obstacle_rects) != -1:\n\t\t\treturn # ignore update\n\n\t\t# collisions with other players\n\t\tfor player in self.game.players:\n\t\t\tif player != self and player.state == player.STATE_ALIVE and player_rect.colliderect(player.rect) == True:\n\t\t\t\treturn # ignore update\n\n\t\t# collisions with enemies\n\t\tfor enemy in self.game.enemies:\n\t\t\tif player_rect.colliderect(enemy.rect) == True:\n\t\t\t\treturn # ignore update\n\n\t\t# collisions with bonuses\n\t\tfor bonus in self.game.bonuses:\n\t\t\tif player_rect.colliderect(bonus.rect) == True:\n\t\t\t\tself.bonus = bonus\n\n\t\t#if no collision, move player\n\t\tself.rect.topleft = (new_position[0], new_position[1])\n\n\tdef reset(self):\n\t\t\"\"\" reset player \"\"\"\n\t\tself.rotate(self.start_direction, False)\n\t\tself.rect.topleft = self.start_position\n\t\tself.superpowers = 0\n\t\tself.max_active_bullets = 1\n\t\tself.health = 100\n\t\tself.paralised = False\n\t\tself.paused = False\n\t\tself.pressed = [False] * 4\n\t\tself.state = self.STATE_ALIVE","repo_name":"Movelocity/battle-city","sub_path":"player.py","file_name":"player.py","file_ext":"py","file_size_in_byte":2981,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"79"} +{"seq_id":"279148752","text":"import customtkinter as ctk\nfrom .settings import *\nfrom PIL import Image\nimport os\nfrom .components import TitleLabel, SmallLabelButton\nfrom typing import Callable\n\n\ndef change_theme():\n \"\"\"\n This function changes the theme of the app, is called by the button on the title bar\n \"\"\"\n\n if ctk.get_appearance_mode() == \"Dark\":\n ctk.set_appearance_mode(\"light\")\n else:\n ctk.set_appearance_mode(\"dark\")\n\n\nclass TitleBar(ctk.CTkFrame):\n \"\"\"\n The title bar that is present in every frame, the text in the title can be changed by a given method\n \"\"\"\n def __init__(self, master,\n title_text,\n refresh_command=None,\n back_command: Callable[..., None] | None = None):\n super().__init__(master,\n fg_color=\"transparent\")\n\n self.back_command = back_command\n\n # The label containing the text that explains what the current scene is showing\n self.title_label = TitleLabel(self,\n title_text)\n\n self.title_label.grid(row=0, column=0, sticky=\"nw\", pady=(DEFAULT_PAD, 0), padx=DEFAULT_PAD)\n\n if self.back_command is not None:\n self.back_button = SmallLabelButton(self,\n text=\"Back\",\n width=0,\n height=0,\n border_spacing=0,\n command=self.back_command)\n self.back_button.grid(row=1, column=0, sticky=\"w\", padx=DEFAULT_PAD)\n self.title_label.grid_configure(pady=(0, 0))\n\n # The image within the theme change button\n self.theme_image = ctk.CTkImage(\n light_image=Image.open(os.path.join(os.getcwd(), \"images/dark-theme.png\")),\n dark_image=Image.open(os.path.join(os.getcwd(), \"images/light-theme.png\")),\n size=(30, 30)\n )\n\n # The button that changes the theme when clicked\n self.theme_button = ctk.CTkButton(\n self,\n fg_color=\"transparent\",\n image=self.theme_image,\n command=change_theme,\n text=\"\",\n width=45,\n height=45,\n hover_color=BUTTON_HOVER_COLOR,\n corner_radius=8\n )\n self.theme_button.grid(row=0, rowspan=2, column=2, sticky=\"e\", pady=DEFAULT_PAD, padx=(0, DEFAULT_PAD))\n\n self.command = refresh_command\n\n if self.command is not None:\n self.refresh_image = ctk.CTkImage(\n light_image=Image.open(os.path.join(os.getcwd(), \"images/light-refresh.png\")),\n dark_image=Image.open(os.path.join(os.getcwd(), \"images/dark-refresh.png\")),\n size=(30, 30)\n )\n\n self.refresh_button = ctk.CTkButton(\n self,\n fg_color=\"transparent\",\n image=self.refresh_image,\n command=self.command,\n text=\"\",\n width=45,\n height=45,\n hover_color=BUTTON_HOVER_COLOR,\n corner_radius=8\n )\n\n self.refresh_button.grid(row=0, rowspan=2, column=1, sticky=\"e\", **DEFAULT_PAD_COMPLETE)\n\n self.rowconfigure(0, weight=1)\n self.rowconfigure(1, weight=0)\n self.columnconfigure(0, weight=1)\n self.columnconfigure([1, 2], weight=0)\n","repo_name":"bilboderbyshire/school-report-writer","sub_path":"scripts/title_bar.py","file_name":"title_bar.py","file_ext":"py","file_size_in_byte":3489,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"79"} +{"seq_id":"41222589494","text":"import os\nimport unittest\n\nfrom numpy.testing import assert_array_equal\n\nfrom tests.base import FIXTURES_DIR, SWORD_PNG_PATH\n\nfrom pixels2svg.utils import pixel\n\nIMAGE_FORMATS_DIR = os.path.join(FIXTURES_DIR, 'formats')\n\n\nclass TestUtilsPixel(unittest.TestCase):\n\n def test_id_to_from_rgba(self):\n\n self.assertEqual(\n (184, 0, 255, 54),\n pixel.id_to_rgba(pixel.rgba_to_id((184, 0, 255, 54)))\n )\n\n def test_id_to_from_rgba_array(self):\n\n img = pixel.read_image(SWORD_PNG_PATH)\n\n assert_array_equal(\n img,\n pixel.id_array_to_rgba_array(pixel.rgba_array_to_id_array(img))\n )\n\n def test_read_image(self):\n image_arrays = []\n for filename in os.listdir(IMAGE_FORMATS_DIR):\n image_path = os.path.join(IMAGE_FORMATS_DIR, filename)\n image_arrays.append(pixel.read_image(image_path))\n array_base = image_arrays[0]\n for array_comp in image_arrays[1:]:\n assert_array_equal(array_base, array_comp)\n","repo_name":"ValentinFrancois/pixels2svg","sub_path":"tests/test_utils_pixel.py","file_name":"test_utils_pixel.py","file_ext":"py","file_size_in_byte":1035,"program_lang":"python","lang":"en","doc_type":"code","stars":14,"dataset":"github-code","pt":"7"} +{"seq_id":"12293428351","text":"import os\nimport logging\nimport config\nimport googleapiclient.discovery\nimport googleapiclient.errors\nimport tweepy\nimport warnings\nwarnings.simplefilter(action='ignore', category=FutureWarning)\nimport pandas as pd\nimport dfgui\nimport csv\nfrom tkinter import *\nimport tkinter as tk\nfrom tkinter import filedialog as fd\nfrom tkinter.messagebox import showinfo\nfrom functools import partial\n\n \ndef twitter(search):\n global errorlabel\n if errorlabel:\n errorlabel.destroy()\n search = search.get()\n with open('searchlog.txt', 'a+') as f:\n f.write(search)\n f.write('\\n')\n f.close()\n #api keys provided by twitter\n consumer_key = config.consumer_key\n consumer_secret = config.consumer_secret\n access_key= config.access_key\n access_secret = config.access_secret\n tweet_num = 8\n tweets = []\n data = []\n auth = tweepy.OAuthHandler(consumer_key, consumer_secret)\n auth.set_access_token(access_key, access_secret)\n api = tweepy.API(auth,wait_on_rate_limit=True)\n #getting results from search \n results = tweepy.Cursor(api.search_tweets,q=search+ \" -filter:retweets\",\n lang=\"en\").items(tweet_num)\n #filling up data into lists\n for status in results:\n if status.lang == 'en':\n data = {'User': status.user.name,\n 'Account name': '@'+status.user.screen_name, \n 'Tweet': status.text, \n 'Time': status.created_at,\n 'Nr of retweets': status.retweet_count,\n 'Nr of favorited': status.favorite_count}\n tweets.append(data)\n if len(data):\n #coverting that data into a csv and gui\n twitinfo = pd.DataFrame(tweets) \n twitinfo.head()\n twitinfo.to_csv('twitter.csv', mode='a+',encoding='utf-8')\n print (twitinfo)\n dfgui.show(twitinfo)\n \n else:\n errorlabel = Label(tk, text=\"No data available for search, please retry\", bg='#ffffff')\n errorlabel.place(x=65, y=300)\n with open('errorlogs.txt', 'a+') as f:\n f.write(\"No data available for \")\n f.write(search)\n f.write('\\n')\n f.close()\ndef youtube(search):\n global errorlabel\n if errorlabel:\n errorlabel.destroy()\n search = search.get()\n with open('searchlog.txt', 'a+') as f:\n f.write(search)\n f.write('\\n')\n f.close()\n rows = []\n videoID = []\n SCOPES = 'https://www.googleapis.com/auth/youtube.force-ssl'\n api_service_name = \"youtube\"\n api_version = \"v3\"\n API_KEY = config.API_KEY\n youtube = googleapiclient.discovery.build(\n api_service_name, api_version, developerKey = API_KEY)\n #searching for video\n request = youtube.search().list(\n part=\"id,snippet\",\n type='video',\n q=search,\n videoDuration='any',\n videoDefinition='any',\n maxResults=1,\n fields=\"nextPageToken,items(id(videoId),snippet(publishedAt,channelId,channelTitle,title,description))\"\n )\n response = request.execute()\n #for each video found looks through comments\n for items in response['items']:\n videoID.append(items['id']['videoId'])\n try:\n for item in videoID:\n video_response=youtube.commentThreads().list(\n part='id,snippet,replies',\n videoId=item,\n maxResults=8\n ).execute()\n #grabs the comments and places them in list\n for item in video_response['items']:\n comment = str(item['snippet']['topLevelComment']['snippet']['textDisplay'])\n rows.append([\n items['snippet']['channelTitle'],\n items['snippet']['title'],\n items['snippet']['description'],comment])\n except:\n with open('errorlogs.txt', 'a+') as f:\n f.write(search)\n f.write(\" ,video has comments disabled\")\n f.write('\\n')\n f.close()\n print(search,\" video has comments disabled\")\n \n if len(rows):\n #coverts into an csv and gui\n ytinfo = pd.DataFrame(rows, columns = [\"Channel Name\", \"Title\", \"Description\", \"Comment\" ])\n ytinfo.to_csv('youtube.csv',mode='a+', encoding='utf-8')\n print(ytinfo)\n dfgui.show(ytinfo)\n else:\n errorlabel = Label(tk, text=\"No data available for search, please retry\", bg='#ffffff')\n errorlabel.place(x=65, y=300)\n with open('errorlogs.txt', 'a+') as f:\n f.write(\"No data available for \")\n f.write(search)\n f.write('\\n')\n f.close()\ndef csv():\n filetypes = (\n ('csv files', '*.csv'),\n ('All files', '*.csv*')\n )\n\n filename = fd.askopenfilename(\n title='Open a file',\n initialdir='/',\n filetypes=filetypes)\n if filename:\n openedCSV = pd.read_csv(filename)\n if (\"twitter.csv\" in str(filename)):\n new = openedCSV[~openedCSV['Nr of retweets'].isin(['Nr of retweets'])]\n dfgui.show(new)\n elif(\"youtube.csv\" in str(filename)):\n new = openedCSV[~openedCSV['Channel Name'].isin(['Channel Name'])]\n dfgui.show(new)\n else:\n dfgui.show(openedCSV)\ndef viewHistory():\n tk = Tk()\n tk.geometry(\"600x500\")\n tk.title(\"History file\")\n txtarea = Text(tk, width=600, height=500, bg='#f1f3f7')\n txtarea.place(x=0, y=0)\n try:\n tf = open(\"searchlog.txt\", \"r\")\n txtarea.insert(END, tf.read())\n tf.close()\n except:\n print(\"searchlog file doesn't exist yet\")\ndef clearHistory():\n with open(\"searchlog.txt\", \"w\") as tf:\n tf.write(\"\")\n tf.close()\ndef on_click(event):\n if searchEntry.cget('fg') == 'grey':\n searchEntry.delete(0, \"end\")\n searchEntry.insert(0, '')\n\n \n#building gui\ntk = Tk()\ntk.geometry(\"720x405\")\ntk.title(\"BuzzHunt\")\ntk['bg']='#ffffff'\n\ntry:\n photo = PhotoImage(file = \"images/buzz.png\")\n logo1 = PhotoImage(file = \"images/logo1.png\")\n logo2 = PhotoImage(file = \"images/logo2.png\")\n chooseOne = PhotoImage(file = \"images/chooseOne.png\")\n ytButton = PhotoImage(file = \"images/ytbutton.png\")\n twButton = PhotoImage(file = \"images/twbutton.png\")\n csvButton = PhotoImage(file = \"images/csvbutton.png\")\n cHistory = PhotoImage(file = \"images/clearHistory.png\")\n vHistory = PhotoImage(file = \"images/viewHistory.png\")\nexcept:\n print(\"Images not found\")\ntk.iconphoto(False, photo)\n\nlabel1 = Label(image = logo1, bg='#ffffff').place(x=300, y=0) \nlabel2 = Label(image = logo2, bg='#ffffff').place(x=0, y=0)\nlabel3 = Label(image = chooseOne, bg='#ffffff').place(x=65, y=210)\nsearch = StringVar()\nerrorlabel = None\nsearchEntry = Entry(tk, textvariable=search, bg='#f9f9f9', width = 25)\nsearchEntry.place(x=66, y=180)\nsearchEntry.insert(0, 'Search..')\nsearchEntry.bind('', on_click)\nsearchEntry.config(fg = 'grey')\n\n\nyoutube = partial(youtube, search)\ntwitter = partial(twitter, search)\nb1 = Button(tk, text=\"On Twitter\", command=twitter, bg='#00acee', image=twButton,highlightthickness = 0, bd = 0).place(x=160, y=240)\nb2 = Button(tk, text=\"On YouTube\", command=youtube, bg='#c4302b', image=ytButton,highlightthickness = 0, bd = 0).place(x=65, y=240)\nb3 = Button(tk, text=\"Open CSV File\", command=csv, bg='#7289DA', image=csvButton,highlightthickness = 0, bd = 0).place(x=255, y=240)\nb4 = Button(tk, text=\"View History\", command=viewHistory, bg='#ffffff',image=vHistory, highlightthickness=0, bd=0).place(x=65, y=360)\nb5 = Button(tk, text=\"Clear History\", command=clearHistory, bg='#ffffff', image=cHistory, highlightthickness=0, bd=0).place(x=171, y=360)\ntk.mainloop() \n","repo_name":"Saad-iit/OISNT_Project","sub_path":"FinalOistproject.py","file_name":"FinalOistproject.py","file_ext":"py","file_size_in_byte":7805,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"28893135899","text":"import cv2\nimport argparse\nimport numpy as np\n\ndef adjacent_video(videoPath1, videoPath2, outputPath):\n \"\"\"\n Name: adjacent_video\n Input: path to left video, path to right video, path to output\n Return: none\n What it does: puts left and right video sidebyside\n \"\"\"\n # read existing video in folder\n cap = cv2.VideoCapture(videoPath1)\n cap2 = cv2.VideoCapture(videoPath2)\n\n # output file\n fourcc = 0x00000021\n out = cv2.VideoWriter(outputPath, fourcc, 20.0, (1280,480))\n\n while(True):\n ret, frame = cap.read()\n ret2, frame2 = cap2.read()\n if not ret and not ret2:\n break;\n if ret:\n lframe = frame\n if ret:\n lframe2 = frame2\n #merge images side by side \n combined_frame = np.concatenate([lframe, lframe2], axis = 1)\n out.write(combined_frame)\n\n cap.release()\n cap2.release()\n cv2.destroyAllWindows()\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"inputPath\", help = \"path to video with *'XVID' codec\")\n parser.add_argument(\"inputPath2\", help = \"path to video with *'XVID' codec\")\n parser.add_argument(\"outputPath\", help = \"path to output video with faces and eyes detected\")\n args = parser.parse_args()\n adjacent_video(args.inputPath, args.inputPath2, args.outputPath)\n","repo_name":"tuftsceeo/EDL2020","sub_path":"student_notebook_root_folder/archive/image_processing/python_scripts/adjacent_video.py","file_name":"adjacent_video.py","file_ext":"py","file_size_in_byte":1380,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"43056671752","text":"import math\n\n# Calculate the perpendicular distance from a point to a line segment\ndef point_to_line_distance(point, start, end):\n line_length = math.dist(start, end)\n if line_length == 0:\n return math.dist(point, start)\n \n t = ((point[0] - start[0]) * (end[0] - start[0]) + (point[1] - start[1]) * (end[1] - start[1])) / (line_length ** 2)\n t = max(0, min(1, t))\n \n projection = (start[0] + t * (end[0] - start[0]), start[1] + t * (end[1] - start[1]))\n return math.dist(point, projection)\n\n# Recursive Douglas-Peucker simplification\ndef douglas_peucker_simplify(path, epsilon):\n dmax = 0\n index = 0\n \n for i in range(1, len(path) - 1):\n d = point_to_line_distance(path[i], path[0], path[-1])\n if d > dmax:\n index = i\n dmax = d\n \n if dmax > epsilon:\n recursive_result1 = douglas_peucker_simplify(path[:index + 1], epsilon)\n recursive_result2 = douglas_peucker_simplify(path[index:], epsilon)\n simplified_path = recursive_result1[:-1] + recursive_result2\n else:\n simplified_path = [path[0], path[-1]]\n \n return simplified_path\n\n\ndef smooth_path(path):\n \n if len(path) <= 2:\n return path # path already smoothed\n\n smoothed_path = path\n # current_dir = [path[1][0] - path[0][0], path[1][1] - path[0][1]]\n # smoothed_path = [path[0]]\n # for i in range(2, len(path)):\n # if [path[i][0] - path[i-1][0], path[i][1] - path[i-1][1]] != current_dir:\n # smoothed_path.append(path[i-1])\n # current_dir = [path[i][0] - path[i-1][0], path[i][1] - path[i-1][1]]\n # smoothed_path.append(path[-1])\n\n\n smoothed_path = douglas_peucker_simplify(smoothed_path, epsilon=1)\n\n return smoothed_path\n","repo_name":"jjah0001/ECE4191G11","sub_path":"SCR/path_planning_search_based/path_smoothing.py","file_name":"path_smoothing.py","file_ext":"py","file_size_in_byte":1759,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"40739193803","text":"\"\"\"\nDjango settings for project.\n\"\"\"\n\n# Build paths inside the project like this: os.path.join(BASE_DIR, ...)\nimport os.path\n\n# Debug\n#DEBUG = False\nDEBUG = True\nTEMPLATE_DEBUG = DEBUG\nADMINS = (\n ('', ''),\n)\nMANAGERS = ADMINS\n\nSECRET_KEY = ''\nALLOWED_HOSTS = []\n\nLANGUAGE_CODE = 'en-us'\nTIME_ZONE = 'America/Chicago'\nSITE_ID = 1\nUSE_I18N = False\nUSE_L10N = False\nUSE_TZ = False\nDEFAULT_CHARSET = 'utf-8'\nFILE_CHARSET = 'utf-8'\n\nSERVER_URL = \"\"\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\nROOT_DIR = os.path.dirname(__file__)\nROOT_URL = \"/djskeletor/\"\nROOT_URLCONF = 'djskeletor.urls'\nWSGI_APPLICATION = 'djskeletor.wsgi.application'\nMEDIA_ROOT = ''\nADMIN_MEDIA_PREFIX = '/static/admin/'\nSTATIC_ROOT = ''\nSTATIC_URL = \"/static/\"\nSTATICFILES_DIRS = ()\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n# 'django.contrib.staticfiles.finders.DefaultStorageFinder',\n)\n\nDATABASES = {\n 'default': {\n 'HOST': '127.0.0.1',\n 'PORT': '3306',\n 'NAME': 'djskeletor',\n 'ENGINE': 'django.db.backends.mysql',\n 'USER': '',\n 'PASSWORD': ''\n },\n}\n\nINSTALLED_APPS = (\n 'django.contrib.admin',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.humanize',\n 'django.contrib.messages',\n 'django.contrib.sessions',\n 'django.contrib.sites',\n 'django.contrib.staticfiles',\n 'djskeletor',\n 'djskeletor.core',\n 'djskeletor.myapp',\n 'djtools',\n)\n\nMIDDLEWARE_CLASSES = (\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.common.CommonMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'django.middleware.clickjacking.XFrameOptionsMiddleware',\n)\n# template stuff\nTEMPLATES = [\n {\n 'BACKEND': 'django.template.backends.django.DjangoTemplates',\n 'DIRS': [\n os.path.join(os.path.dirname(__file__), 'templates'),\n \"/data2/django_templates/djkorra/\",\n \"/data2/django_templates/djcher/\",\n \"/data2/django_templates/\",\n ],\n 'APP_DIRS': True,\n 'OPTIONS': {\n 'context_processors': [\n \"djtools.context_processors.sitevars\",\n 'django.contrib.auth.context_processors.auth',\n 'django.template.context_processors.debug',\n 'django.template.context_processors.media',\n 'django.core.context_processors.request',\n 'django.template.context_processors.static',\n 'django.contrib.messages.context_processors.messages',\n ],\n #'loaders': [\n # # insert your TEMPLATE_LOADERS here\n #]\n },\n },\n]\n# caching\nCACHES = {\n 'default': {\n 'BACKEND': 'django.core.cache.backends.dummy.DummyCache',\n #'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',\n #'LOCATION': '127.0.0.1:11211',\n #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache',\n #'LOCATION': '/var/tmp/django_directory_cache',\n #'TIMEOUT': 60*20,\n #'KEY_PREFIX': \"DIRECTORY_\",\n #'OPTIONS': {\n # 'MAX_ENTRIES': 80000,\n #}\n }\n}\nCACHE_MIDDLEWARE_ANONYMOUS_ONLY = True\n\n# LDAP Constants\nLDAP_SERVER = ''\nLDAP_PORT = '636'\nLDAP_PROTOCOL = \"ldaps\"\nLDAP_BASE = \"\"\nLDAP_USER = \"\"\nLDAP_PASS = \"\"\nLDAP_EMAIL_DOMAIN = \"\"\n# auth backends\nAUTHENTICATION_BACKENDS = (\n 'djauth.ldapBackend.LDAPBackend',\n 'django.contrib.auth.backends.ModelBackend',\n)\nLOGIN_URL = '/djskeletor/accounts/login/'\nLOGIN_REDIRECT_URL = '/djskeletor/'\nUSE_X_FORWARDED_HOST = True\n#SESSION_ENGINE = \"django.contrib.sessions.backends.cache\"\nSESSION_EXPIRE_AT_BROWSER_CLOSE = False\nSESSION_COOKIE_DOMAIN=\".carthage.edu\"\nSESSION_COOKIE_NAME ='django_carthage_cookie'\nSESSION_COOKIE_AGE = 86400\n\n# logging\nLOG_FILEPATH = os.path.join(os.path.dirname(__file__), \"logs/\")\nLOG_FILENAME = LOG_FILEPATH + \"debug.log\"\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': True,\n 'formatters': {\n 'standard': {\n 'format' : \"[%(asctime)s] %(levelname)s [%(name)s:%(lineno)s] %(message)s\",\n 'datefmt' : \"%Y/%b/%d %H:%M:%S\"\n },\n 'verbose': {\n 'format': '%(levelname)s %(asctime)s %(module)s %(process)d %(thread)d %(message)s',\n 'datefmt' : \"%Y/%b/%d %H:%M:%S\"\n },\n 'simple': {\n 'format': '%(levelname)s %(message)s'\n },\n },\n 'handlers': {\n 'logfile': {\n 'level':'DEBUG',\n 'class':'logging.handlers.RotatingFileHandler',\n 'filename': LOG_FILENAME,\n 'maxBytes': 50000,\n 'backupCount': 2,\n 'formatter': 'standard',\n },\n 'console':{\n 'level':'INFO',\n 'class':'logging.StreamHandler',\n 'formatter': 'standard'\n },\n 'mail_admins': {\n 'level': 'ERROR',\n 'class': 'django.utils.log.AdminEmailHandler'\n }\n },\n 'loggers': {\n 'mugshots.upload': {\n 'handlers':['logfile'],\n 'propagate': True,\n 'level':'DEBUG',\n },\n 'core': {\n 'handlers':['logfile'],\n 'propagate': True,\n 'level':'DEBUG',\n },\n 'django': {\n 'handlers':['console'],\n 'propagate': True,\n 'level':'WARN',\n },\n 'django.db.backends': {\n 'handlers': ['console'],\n 'level': 'DEBUG',\n 'propagate': False,\n },\n 'django.request': {\n 'handlers': ['mail_admins'],\n 'level': 'ERROR',\n 'propagate': True,\n },\n }\n}\n","repo_name":"carthage-college/django_1.11.x-djparking","sub_path":"djparking/settings_default.py","file_name":"settings_default.py","file_ext":"py","file_size_in_byte":5876,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37721363356","text":"import typing\n\nclass DispatchRepr:\n def __repr__(self) -> str:\n _repr = f\"\"\"\n Dispatch(\\n\n vals={self._vals},\n children={self._children}\n )\n \"\"\"\n return _repr\n\nclass a_HrefRepr:\n def __repr__(self) -> str:\n _repr = f\"\"\"\n A_Href(\n href={self.tree.tree_attrs[\"href\"]},\n content={self.struct}\n\n )\n \"\"\"\n return _repr\n\nclass ImgRepr:\n def __repr__(self) -> str:\n _repr = f\"\"\"\n Img(\n src={self.tree.tree_attrs[\"src\"]},\n content={self.struct}\n\n )\n \"\"\"\n return _repr\n\nclass StructuredRecordRepr:\n def __repr__(self) -> str:\n _repr = \"\"\"\n StructuredRecord(\n records = {},\n content = {}\n )\n \"\"\".format(\",\\n\".join(map(repr, self.records)), self.struct)\n return _repr\n\nclass MatchedRunRepr:\n def __repr__(self) -> str:\n _repr = \"\"\"\n MatchedRun(\n runs = {},\n \n )\n \"\"\".format(\",\\n\".join(map(repr, self.run_objs)))\n return _repr\n\nclass PatternMatcherRepr:\n def __repr__(self) -> str:\n _repr = \"\"\"\n PatternMatcher(\n runs = {},\n struct = {}\n )\n \"\"\".format(\",\\n\".join(map(repr, self.runs)), self.struct)\n return _repr\n \n\n","repo_name":"Ajax12345/scrapio","sub_path":"scrapio/parsed_reprs.py","file_name":"parsed_reprs.py","file_ext":"py","file_size_in_byte":1366,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"11029728130","text":"from django.db import models\nfrom django.conf import settings\nfrom django.urls import reverse\nfrom django.utils.translation import gettext as _\nfrom django.contrib.auth.models import User\nfrom model_utils import Choices\n# Create your models here.\n\n\nclass Akcesoria(models.Model):\n\n CO_CHOICES = Choices(\n ('plecki', _('Plecki')),\n ('book', _('Książka')),\n ('szklo', _('Szkło')),\n ('folia', _('Folia')),\n ('inne', _('Inne')),\n )\n\n data = models.DateField(auto_now_add=True)\n kto = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE, default=User)\n co = models.CharField(max_length=15, choices=CO_CHOICES, default=CO_CHOICES.plecki)\n kwota = models.DecimalField(max_digits=5, decimal_places=2)\n model = models.CharField(max_length=128, null=True, blank=True)\n\n def __str__(self):\n return self.co\n\n def get_absolute_url(self):\n return reverse(\"dashboard:dashboard-home\")\n","repo_name":"kmoskal/playbb","sub_path":"akcesoria/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":965,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"35642687440","text":"# Raikage The Third\n\nfrom django.urls import path\n\nfrom .views import *\n\nurlpatterns =[\n path('', homepage, name='home'),\n path('demande-de-devis/', DevisPageView.as_view(), name='devis'),\n path('marketing/', marketing, name='marketing'),\n path('mobile-app/', mobile, name='mobileapp'),\n path('software-development/', softwares, name='softwaredev'),\n path('website-development/', website, name='websitedev'),\n\n]","repo_name":"LeonardoRubuz/startech","sub_path":"visiteurs/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":429,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26590732465","text":"#8. Faça um Programa que peça a idade e a altura de 5 pessoas, armazene cada informação no seu respectivo vetor.\n# Imprima a idade e a altura na ordem inversa a ordem lida.\nlistaIdades = []\nlistaAlturas = []\nfor i in range(1,6):\n idade = int(input(f'Informe a idade da pessoa {i}/5: '))\n altura = float(input(f'Informe a altura da pessoa {i}/5: '))\n listaIdades.append(idade)\n listaAlturas.append(altura)\nlistaAlturas.reverse()\nlistaIdades.reverse()\nprint(listaIdades)\nprint(listaAlturas)\n","repo_name":"MdotSouza/WikiPython","sub_path":"04ExerciciosListas/l4ex8.py","file_name":"l4ex8.py","file_ext":"py","file_size_in_byte":505,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"27606105307","text":"# coding: utf-8\n\nfrom apis.commons.apis_commons import Apis\n\n\nclass ApisUtils(Apis):\n\n def __init__(self):\n super(ApisUtils, self).__init__()\n\n def get_user_message(self):\n \"\"\"\n 获取用户通知\n \"\"\"\n filter_params = '{\"logic\":\"and\",\"filters\":[{\"field\":\"isDeleted\",\"operator\":\"eq\",\"value\":False},' \\\n '{\"field\":\"userId\",\"operator\":\"eq\",\"value\":\"39fe8e4c-5292-d5da-a219-caa9fafa3913\"}]}'\n\n params = {\n \"skipCount\": 0,\n \"sorting\": \"creationTime desc\",\n \"filter\": filter_params\n }","repo_name":"zj1995-09-09/supercare_api_test","sub_path":"testcase/commons/commons_steps.py","file_name":"commons_steps.py","file_ext":"py","file_size_in_byte":590,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"22557292671","text":"import unittest\n\"\"\"\nTests for stop finder, stop getter, and route selector\n\"\"\"\n\nclass test(unittest.TestCase):\n \n\n def test_model_coverage(self):\n \"\"\"\n Tests that there are models for every stop link\n So we have all of the models\n \"\"\"\n print('testing model coverage')\n \n import os\n import json\n fails=0\n total = 0\n routes = json.loads(open('/home/student/db/resources/trimmed_routes.json','r').read())\n route = routes['15']\n \n \n for route in routes:\n for v_num,v in enumerate(routes[route]):\n failed_on_variation = False\n for i in range(1, len(v) -1 ):\n total += 1\n if not os.path.exists('/data/neural_models3/'+str(v[i])+'_'+str(v[i+1])+'.bin'):\n fails+=1\n failed_on_variation = True\n\n if failed_on_variation:\n print(route,v_num)\n f=open('failed_routes_log.log','a')\n f.write(str(route)+'_'+str(v_num)+'\\n')\n f.close()\n\n print(total,fails)\n input()\n self.assertEqual(fails==0,True) \n\n def test_build_network(self):\n \"\"\"\n Tests that the network object can be built from scratch. Takes forever.\n \"\"\"\n\n from dbanalysis.network import simple_network4\n import pickle\n import time\n t1 = time.time()\n n = simple_network4.simple_network()\n\n n.prepare_dijkstra()\n n.properly_add_foot_links()\n n.generate_time_tables()\n for node in n.nodes:\n\n n.nodes[node].timetable.concat_and_sort()\n with open('/data/done2.bin','wb') as handle:\n pickle.dump(n,handle,protocol=pickle.HIGHEST_PROTOCOL)\n for node in n.nodes:\n print(node)\n self.assertEqual(len(n.nodes[node].timetable.data) > 0, True)\n del(n)\n end_time = time.time() - t1\n import datetime\n dt = datetime.datetime.now()\n f=open('report.txt','a')\n f.write('*******Test report for ' + str(dt) + '*******\\n')\n f.write('Network object constructed and all timetables generated in ' + str(end_time) + 'seconds\\n')\n f.close()\ndef main():\n unittest.main()\n\nif __name__ == \"__main__\":\n main()\n \n","repo_name":"Manjunathsk92/dbanalysis","sub_path":"dbanalysis/tests/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":2395,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"19079867853","text":"import os\n\nimport numpy as np\n\n\ndef triangle_subdivision(n_div):\n faces = []\n for i in range(1, n_div + 1)[::-1]:\n for j in range(n_div + 1 - i):\n k = n_div - i - j\n f = [(i, j, k), (i - 1, j + 1, k), (i - 1, j, k + 1)]\n faces.append(f)\n for i in range(1, n_div + 1)[::-1]:\n for j in range(1, n_div + 1 - i):\n k = n_div - i - j\n f = [(i, j, k), (i - 1, j, k + 1), (i, j - 1, k + 1)]\n faces.append(f)\n return faces\n\n\ndef subdivide_edges(coords, faces, n_div):\n n_edges = faces.shape[0] * 3 // 2\n nv_new = n_edges * (n_div - 1)\n new_coords = np.zeros((nv_new, 3), dtype=coords.dtype)\n\n nv = coords.shape[0]\n count = 0\n edges = set()\n e_mapping = {}\n for f in faces:\n for a, b in [[f[0], f[1]], [f[0], f[2]], [f[1], f[2]]]:\n e = (a, b) if a < b else (b, a)\n if e in edges:\n continue\n edges.add(e)\n for i in range(1, n_div):\n c = (coords[a] * i + coords[b] * (n_div - i)) / n_div\n # c /= np.linalg.norm(c)\n # new_coords.append(c)\n new_coords[count] = c\n e_mapping[(e[0], e[1], i)] = count + nv\n count += 1\n\n return new_coords, e_mapping\n\n\ndef subdivide_inside(coords, faces, e_mapping, n_div, count_base):\n # new_coords, new_faces = [], []\n nv_new = (n_div - 1) * (n_div - 2) // 2 * len(faces)\n new_coords = np.zeros((nv_new, 3), dtype=coords.dtype)\n count = 0\n new_faces = []\n nf_base = triangle_subdivision(n_div)\n\n for f in faces:\n mapping = {\n (n_div, 0, 0): f[0],\n (0, n_div, 0): f[1],\n (0, 0, n_div): f[2],\n }\n\n for ii, jj in [[0, 1], [0, 2], [1, 2]]:\n a, b = f[[ii, jj]]\n aa, bb = (a, b) if a < b else (b, a)\n for step in range(1, n_div):\n val = e_mapping[aa, bb, step]\n key = [0, 0, 0]\n key[ii] = step\n key[jj] = n_div - step\n mapping[tuple(key)] = val\n\n for i in range(n_div)[::-1]:\n for j in range(n_div + 1 - i):\n k = n_div - i - j\n if (i, j, k) not in mapping:\n mapping[(i, j, k)] = count + count_base\n c = np.sum(\n coords[f] * (np.array([i, j, k])[:, np.newaxis] / n_div), axis=0\n )\n # c /= np.linalg.norm(c)\n # new_coords.append(c)\n new_coords[count] = c\n count += 1\n\n nf = [[mapping[v] for v in f] for f in nf_base]\n new_faces += nf\n\n # new_coords = np.array(new_coords)\n new_faces = np.array(new_faces)\n return new_coords, new_faces\n\n\ndef surface_subdivision(coords, faces, n_div):\n edge_coords, e_mapping = subdivide_edges(coords, faces, n_div)\n count = coords.shape[0] + edge_coords.shape[0]\n inside_coords, new_faces = subdivide_inside(coords, faces, e_mapping, n_div, count)\n print(coords.shape, edge_coords.shape, inside_coords.shape)\n new_coords = np.concatenate([coords, edge_coords, inside_coords], axis=0)\n return new_coords, new_faces\n\n\n# new_coords, new_faces = [], []\n\n# # count = coords.shape[0]\n\n# new_coords_edges, e_mapping = subdivide_edges(coords, faces, n_div)\n\n\n# new_coords = np.array(new_coords)\n# new_faces = np.array(new_faces)\n# coords = np.concatenate([coords, new_coords], axis=0)\n\n# return coords, new_faces\n","repo_name":"neuroboros/neuroboros","sub_path":"src/neuroboros/surface/subdivision.py","file_name":"subdivision.py","file_ext":"py","file_size_in_byte":3579,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"24763824353","text":"from vyper.compiler import (\n compile_code,\n compile_codes,\n)\nfrom vyper.exceptions import (\n StructureException,\n)\nfrom vyper.interfaces import (\n ERC20,\n ERC721,\n)\nfrom vyper.signatures.interface import (\n extract_file_interface_imports,\n extract_sigs,\n)\n\n\ndef test_basic_extract_interface():\n code = \"\"\"\n# Events\n\nTransfer: event({_from: address, _to: address, _value: uint256})\n\n# Functions\n\n@constant\n@public\ndef allowance(_owner: address, _spender: address) -> (uint256, uint256):\n return 1, 2\n \"\"\"\n\n out = compile_code(code, ['interface'])\n out = out['interface']\n code_pass = '\\n'.join(code.split('\\n')[:-2] + [' pass']) # replace with a pass statement.\n\n assert code_pass.strip() == out.strip()\n\n\ndef test_basic_extract_external_interface():\n code = \"\"\"\n@constant\n@public\ndef allowance(_owner: address, _spender: address) -> (uint256, uint256):\n return 1, 2\n\n@public\ndef test(_owner: address):\n pass\n\n@constant\n@private\ndef _prive(_owner: address, _spender: address) -> (uint256, uint256):\n return 1, 2\n \"\"\"\n\n interface = \"\"\"\n# External Contracts\ncontract One:\n def allowance(_owner: address, _spender: address) -> (uint256, uint256): constant\n def test(_owner: address): modifying\n \"\"\"\n\n out = compile_codes({'one.vy': code}, ['external_interface'])[0]\n out = out['external_interface']\n\n assert interface.strip() == out.strip()\n\n\ndef test_basic_interface_implements(assert_compile_failed):\n code = \"\"\"\nfrom vyper.interfaces import ERC20\n\nimplements: ERC20\n\n\n@public\ndef test() -> bool:\n return True\n \"\"\"\n\n assert_compile_failed(\n lambda: compile_codes({'one.vy': code}),\n StructureException\n )\n\n\ndef test_builtin_interfaces_parse():\n assert len(extract_sigs({'type': 'vyper', 'code': ERC20.interface_code})) == 8\n assert len(extract_sigs({'type': 'vyper', 'code': ERC721.interface_code})) == 13\n\n\ndef test_external_interface_parsing(assert_compile_failed):\n interface_code = \"\"\"\n@public\ndef foo() -> uint256:\n pass\n\n@public\ndef bar() -> uint256:\n pass\n \"\"\"\n\n interface_codes = {\n 'FooBarInterface': {\n 'type': 'vyper',\n 'code': interface_code\n }\n }\n\n code = \"\"\"\nimport a as FooBarInterface\n\nimplements: FooBarInterface\n\n@public\ndef foo() -> uint256:\n return 1\n\n@public\ndef bar() -> uint256:\n return 2\n \"\"\"\n\n assert compile_codes({'one.vy': code}, interface_codes=interface_codes)[0]\n\n not_implemented_code = \"\"\"\nimport a as FooBarInterface\n\nimplements: FooBarInterface\n\n@public\ndef foo() -> uint256:\n return 1\n\n \"\"\"\n\n assert_compile_failed(\n lambda: compile_codes({'one.vy': not_implemented_code}, interface_codes=interface_codes)[0],\n StructureException\n )\n\n\ndef test_extract_file_interface_imports(assert_compile_failed):\n code = \"\"\"\nimport a as FooBarInterface\n \"\"\"\n\n assert extract_file_interface_imports(code) == {'FooBarInterface': 'a'}\n\n invalid_no_alias_code = \"\"\"\nimport a\n \"\"\"\n assert_compile_failed(\n lambda: extract_file_interface_imports(invalid_no_alias_code), StructureException\n )\n\n invalid_interfac_already_exists_code = \"\"\"\nimport a as A\nimport a as A\n \"\"\"\n assert_compile_failed(lambda: extract_file_interface_imports(invalid_interfac_already_exists_code), StructureException) # noqa: E501\n\n\ndef test_external_call_to_interface(w3, get_contract):\n token_code = \"\"\"\nbalanceOf: public(map(address, uint256))\n\n@public\ndef transfer(to: address, value: uint256):\n self.balanceOf[to] += value\n \"\"\"\n\n code = \"\"\"\nimport one as TokenCode\n\ncontract EPI:\n def test() -> uint256: constant\n\n\ntoken_address: TokenCode\n\n\n@public\ndef __init__(_token_address: address):\n self.token_address = TokenCode(_token_address)\n\n\n@public\ndef test():\n self.token_address.transfer(msg.sender, 1000)\n \"\"\"\n\n erc20 = get_contract(token_code)\n test_c = get_contract(code, *[erc20.address], interface_codes={\n 'TokenCode': {'type': 'vyper', 'code': token_code}\n })\n\n sender = w3.eth.accounts[0]\n assert erc20.balanceOf(sender) == 0\n\n test_c.test(transact={})\n assert erc20.balanceOf(sender) == 1000\n\n\ndef test_external_call_to_builtin_interface(w3, get_contract):\n token_code = \"\"\"\nbalanceOf: public(map(address, uint256))\n\n@public\ndef transfer(to: address, value: uint256):\n self.balanceOf[to] += value\n \"\"\"\n\n code = \"\"\"\nfrom vyper.interfaces import ERC20\n\n\ntoken_address: ERC20\n\n\n@public\ndef __init__(_token_address: address):\n self.token_address = ERC20(_token_address)\n\n\n@public\ndef test():\n self.token_address.transfer(msg.sender, 1000)\n \"\"\"\n\n erc20 = get_contract(token_code)\n test_c = get_contract(code, *[erc20.address], interface_codes={\n 'TokenCode': {\n 'type': 'vyper',\n 'code': token_code\n }\n })\n\n sender = w3.eth.accounts[0]\n assert erc20.balanceOf(sender) == 0\n\n test_c.test(transact={})\n assert erc20.balanceOf(sender) == 1000\n\n\ndef test_json_interface(get_contract):\n code = \"\"\"\nimport folding as Folding\n\nimplements: Folding\n\n@public\ndef test(a: uint256) -> uint256:\n return 1 + a\n\n\n@public\ndef test2(a: uint256):\n pass\n \"\"\"\n\n interface_codes = {\n 'Folding': {\n 'type': 'json',\n 'code': [\n {\n \"name\": \"test\",\n \"outputs\": [{\n \"type\": \"uint256\",\n \"name\": \"out\"\n }],\n \"inputs\": [{\n \"type\": \"uint256\",\n \"name\": \"s\"\n }],\n \"constant\": False,\n \"payable\": False,\n \"type\": \"function\",\n },\n {\n \"name\": \"test2\",\n \"outputs\": [],\n \"inputs\": [{\n \"type\": \"uint256\",\n \"name\": \"s\"\n }],\n \"constant\": False,\n \"payable\": False,\n \"type\": \"function\",\n }\n ]\n }\n }\n\n c = get_contract(code, interface_codes=interface_codes)\n\n assert c.test(2) == 3\n\n\ndef test_units_interface(w3, get_contract):\n code = \"\"\"\nimport balanceof as BalanceOf\n\nimplements: BalanceOf\n\n@public\n@constant\ndef balanceOf(owner: address) -> wei_value:\n return as_wei_value(1, \"ether\")\n \"\"\"\n interface_code = \"\"\"\n@public\n@constant\ndef balanceOf(owner: address) -> uint256:\n pass\n \"\"\"\n interface_codes = {\n \"BalanceOf\": {\n 'type': 'vyper',\n 'code': interface_code\n }\n }\n c = get_contract(code, interface_codes=interface_codes)\n\n assert c.balanceOf(w3.eth.accounts[0]) == w3.toWei(1, \"ether\")\n","repo_name":"ssteiger/Vyper-Contract-GUI-Depreciated","sub_path":"vyper/tests/parser/functions/test_interfaces.py","file_name":"test_interfaces.py","file_ext":"py","file_size_in_byte":6850,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"21663292742","text":"# -*- coding: utf-8 -*-\n\nclass Wrapper:\n def __init__(self):\n self.data = [\n {'city': 'Marseilles'},\n {'city': 'Paris'}, \n {'city': 'Lyon'}, \n ]\n self.website = 'https://www.mazda.fr/forms-v2/dealer-locatorfrance/'\n self.category = 'car dealer'\n self.http_method = 'POST'\n self.response_format = 'JSON'\n self.notes = 'Uses JSON for payload'\n\n def run(self, browser, inputs):\n browser.get(self.website)\n browser.wait_load('input[name=\"bylocation\"]')#div.main-search')\n browser.keys('input[name=\"bylocation\"]', inputs['city'] + '', False)#True)\n browser.click('div.main-search > button', True)\n browser.wait_load('div.dealer-name a')\n return {\n 'name': [e.partition('. ')[-1] for e in browser.text('div.dealer-name a')],\n 'address': browser.text('li[ng-if=\"address.FirstLine\"]'),\n 'city': browser.text('li[ng-if=\"address.TownCity\"]'),\n 'postcode': browser.text('li[ng-if=\"address.PostCode\"]'),\n }\n","repo_name":"richardpenman/minwrap","sub_path":"wrappers/mazda.py","file_name":"mazda.py","file_ext":"py","file_size_in_byte":1076,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"20672358613","text":"from sqlalchemy import create_engine, MetaData, Table, Integer, Column, JSON, Text, DateTime\nfrom sqlalchemy.engine import Engine\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker, Session\n\nBase = declarative_base()\n\n\ndef connect(db_URI: str) -> Session:\n database_URI = db_URI\n created_engine = create_engine(database_URI, echo=True)\n session = sessionmaker(bind=created_engine, )\n session = session()\n _create_table_if_needed('news_article', created_engine)\n return session\n\n\ndef _create_table_if_needed(table_name: str, created_engine: Engine) -> None:\n metadata = MetaData()\n new_table = Table(table_name, metadata,\n Column('id', Integer, primary_key=True),\n Column('author', JSON),\n Column('body', Text),\n Column('categories', JSON),\n Column('characters_count', Integer),\n Column('entities', JSON),\n Column('hashtags', Text),\n Column('keywords', Text),\n Column('language', Text),\n Column('links', JSON),\n Column('media', JSON),\n Column('paragraphs_count', Integer),\n Column('published_at', DateTime),\n Column('sentences_count', Integer),\n Column('sentiment', JSON),\n Column('social_shares_count', JSON),\n Column('source', JSON),\n Column('summary', JSON),\n Column('title', Text),\n Column('words_count', Integer),\n mysql_charset='latin1')\n metadata.create_all(created_engine, tables=[new_table], checkfirst=True)\n","repo_name":"venomouscyanide/SQL-Alchemy-JSON-Import","sub_path":"connection.py","file_name":"connection.py","file_ext":"py","file_size_in_byte":1830,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13495422478","text":"import numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.distributions import Categorical\n\nfrom common import obesrvation_shape, device, Agent\nfrom utils import get_cumulative_rewards, generate_session, generate_session_batch, get_total_rewards\n\n\ndef reinforce_loss(logprobas, actions, rewards):\n # select log-probabilities for chosen actions, log pi(a_i|s_i)\n logprobas_for_actions = torch.sum(logprobas * to_one_hot(actions).to(torch.float), dim=1)\n return torch.mean(-logprobas_for_actions * rewards)\n\n\nclass ActorCritic(Agent):\n def __init__(self, observation_space, action_space, args):\n self.model = MLPPolicy(observation_space, action_space).to(device)\n self.optimizer = torch.optim.Adam(self.model.parameters(), args.lr)\n self.gamma = args.gamma\n self.value_loss_coef = args.value_loss_coef\n self.entropy_coef = args.entropy_coef\n\n def save(self, path):\n torch.save({\n 'state_dict': self.model.state_dict()\n }, path)\n\n def load(self, path):\n checkpoint = torch.load(path)\n state_dict = checkpoint.get('state_dict', checkpoint)\n self.model.load_state_dict(state_dict)\n\n def update(self, states, actions, cumulative_rewards, dones=None):\n states = torch.tensor(states).to(device, torch.float)\n actions = torch.tensor(actions).to(device, torch.float)\n cumulative_returns = torch.tensor(cumulative_rewards).to(device, torch.float)\n\n # predict logits, probas and log-probas using an agent.\n logits, values = self.model(states)\n probas = F.softmax(logits, dim=1)\n logprobas = F.log_softmax(logits, dim=1)\n\n # REINFORCE objective function\n self.optimizer.zero_grad()\n rewards_with_baseline = cumulative_returns - values.squeeze(1)\n policy_loss = reinforce_loss(logprobas, actions, rewards_with_baseline)\n value_loss = F.smooth_l1_loss(values, cumulative_returns.unsqueeze(1))\n entropy = torch.mean(-torch.sum(logprobas * probas, dim=1))\n loss = policy_loss + self.value_loss_coef * value_loss + self.entropy_coef * entropy\n loss.backward()\n self.optimizer.step()\n\n return loss, policy_loss, value_loss, entropy\n\n def get_action(self, state):\n with torch.no_grad():\n logits, _ = self.model(torch.tensor(state).to(device, torch.float))\n dist = Categorical(F.softmax(logits, dim=1))\n sample = dist.sample()\n return sample.detach().cpu().numpy()\n\n def get_value(self, state):\n with torch.no_grad():\n _, values = self.model(torch.tensor(state).to(device, torch.float))\n return values.detach().cpu().numpy()\n\n\nclass MLPPolicy(nn.Module):\n def __init__(self, observation_space, action_space):\n super(MLPPolicy, self).__init__()\n\n self.logits = nn.Sequential(\n nn.Linear(*obesrvation_shape(observation_space), 64),\n nn.ReLU(),\n nn.Linear(64, 64),\n nn.ReLU(),\n nn.Linear(64, action_space.n)\n )\n\n self.state_values = nn.Sequential(\n nn.Linear(*obesrvation_shape(observation_space), 64),\n nn.ReLU(),\n nn.Linear(64, 64),\n nn.ReLU(),\n nn.Linear(64, 1)\n )\n\n def forward(self, x):\n x = x.view(x.size(0), -1)\n return self.logits(x), self.state_values(x)\n\n\ndef to_one_hot(y, n_dims=None):\n \"\"\" Take an integer vector (tensor of variable) and convert it to 1-hot matrix. \"\"\"\n y_tensor = y.to(torch.long).view(-1, 1)\n n_dims = n_dims if n_dims is not None else int(torch.max(y_tensor)) + 1\n y_one_hot = y_tensor.new_zeros(y_tensor.size(0), n_dims).scatter_(1, y_tensor, 1)\n return y_one_hot\n\n\ndef train(agent, env, args, max_reward=200):\n total_rewars = []\n for i in range(1, args.num_steps + 1):\n states, actions, rewards, dones, lastvalues = generate_session_batch(agent, env, args.episode_length)\n\n # get cumulative rewards\n # use fictive reward for handling unfinished sessions\n rewards_ = np.c_[rewards, lastvalues]\n dones_ = np.c_[dones, np.zeros(dones.shape[0])]\n cumulative_rewards = get_cumulative_rewards(rewards_, agent.gamma, dones=dones_)[:, :-1]\n\n # reshape sessions to form a batch\n batch_size = env.num_envs * args.episode_length\n states = states.reshape((batch_size,) + env.observation_space.shape)\n actions = actions.reshape((batch_size,))\n cumulative_rewards = cumulative_rewards.reshape((batch_size,))\n dones = dones.reshape((batch_size,))\n\n # update policy on a batch\n losses = agent.update(states, actions, cumulative_rewards, dones)\n # note that last session reward is truncated\n total_rewars.extend(get_total_rewards(rewards, dones))\n\n current_mean_reward = np.mean(total_rewars[-100:])\n if i % 100 == 0:\n print(\"Iteration: %i, Mean reward:%.3f\" % (i, current_mean_reward))\n if current_mean_reward > max_reward:\n return\n","repo_name":"vadimadr/reinforcement_learning","sub_path":"a2c/actor_critic.py","file_name":"actor_critic.py","file_ext":"py","file_size_in_byte":5122,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"39952013727","text":"from textwrap import dedent\n\nfrom weaveio.data import Data\nfrom weaveio.hierarchy import Hierarchy, Multiple\n\n\nclass CASU(Hierarchy):\n idname = 'id'\n\nclass Run(Hierarchy):\n idname = 'id'\n\nclass RawSpectrum(Hierarchy):\n parents = [CASU, Run]\n identifier_builder = ['casu', 'run']\n\n\n\ndata = Data(dbname='lowleveltest')\ndata.hierarchies = {CASU, Run, RawSpectrum}\nwith data.write:\n # data.drop_all_constraints(indexes=True)\n # data.apply_constraints() # needed here because we're doing it ourselves\n with data.write_cypher('ignore') as query:\n casu = CASU(id=1)\n run = Run(id=1)\n raw = RawSpectrum(casu=casu, run=run)\n\n cypher, params = query.render_query()\n print(dedent(cypher))\n # results = data.graph.execute(cypher, **params)","repo_name":"philastrophist/weave-io","sub_path":"merge_test.py","file_name":"merge_test.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"21221710357","text":"# coding: utf-8\n\nimport os\nimport codecs\n\n\ndef load_csv(file_path):\n _csv_list = []\n _file = codecs.open(file_path)\n\n try:\n _data = _file.read()\n if _data[:3] == codecs.BOM_UTF8:\n _data = _data[3:]\n _lines = _data.splitlines()\n _merge_line = \"\"\n _line_index = 0\n for i in range(len(_lines)):\n _merge_line += _lines[i] + \"\\n\"\n if _merge_line.count('\"') % 2 == 0:\n _column_list = []\n _columns = _merge_line.strip(\"\\n\").split(\",\")\n _merge_column = \"\"\n for _col in _columns:\n _merge_column += _col + \",\"\n if _merge_column.count('\"') % 2 == 0:\n _column_list.append(_merge_column.strip(\",\").strip('\"'))\n _merge_column = \"\"\n _csv_list.append(_column_list)\n # _csv_list[_line_index].append()\n _merge_line = \"\"\n _line_index += 1\n return _csv_list\n\n finally:\n if _file is not None:\n _file.close()\n\n\ndef get_files(directory, extension=\".json\"):\n \"\"\"\n Get files from the directory\n :param directory: the directory\n :param extension: the file extension\n :return: the files\n \"\"\"\n if not os.path.isdir(directory):\n raise Exception(\"The param 'directory' should be a directory\")\n\n _file_list = []\n for _item in os.listdir(directory):\n _item_full_path = os.path.join(directory, _item)\n if os.path.isdir(_item_full_path):\n for _file in get_files(os.path.join(directory, _item), extension):\n _file_list.append(_file)\n else:\n _name, _extension = os.path.splitext(_item_full_path)\n if _extension.lower() == extension.lower():\n _file_list.append(_item_full_path)\n\n return _file_list\n\n\ndef have_file(directory, file_name, extension=\".json\"):\n \"\"\"\n Check if the directory contains specified file\n :param directory: the directory\n :param file_name: the file name\n :param extension: the file extension\n :return: True or False\n \"\"\"\n for _file in get_files(directory, extension):\n if os.path.split(_file)[1].lower() == file_name.lower():\n return True\n\n return False\n","repo_name":"supor/label_UI","sub_path":"Framework/Test-Automation/framework/taf_utility.py","file_name":"taf_utility.py","file_ext":"py","file_size_in_byte":2320,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"10349007430","text":"#!/usr/bin/python\n\nimport sys\nimport re\n\nlines = []\nmatch = 0\n\nfor line in sys.stdin:\n\tlines.append(line)\n\nfor line in lines:\n\tif \"{{title}}\" in line:\n\t\ttitle = \"Some Page :)\"\n\t\tfor c_line in lines:\n\t\t #DOC: finds first h2 and uses text inside as title\n\t\t\tif \"

\" in c_line and match == 0:\n\t\t\t\ttitle = re.search(r\"

[^<]+<\\/h2>\", c_line).group()[4:-5]\n\t\t\t\tmatch = 1\n\t\tline = line.replace(\"{{title}}\", title)\n\tsys.stdout.write(line)\n\n","repo_name":"flber/mx","sub_path":"plugins/title.py","file_name":"title.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"7"} +{"seq_id":"2829236314","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport datetime\nfrom django.utils.timezone import utc\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('goblin', '0002_projectlink'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='projectlink',\n name='type',\n field=models.CharField(default=datetime.datetime(2015, 4, 29, 21, 22, 3, 951741, tzinfo=utc), max_length=100),\n preserve_default=False,\n ),\n migrations.AddField(\n model_name='projectlink',\n name='url',\n field=models.URLField(default='about:blank'),\n preserve_default=False,\n ),\n ]\n","repo_name":"src-r-r/django-project-goblin","sub_path":"goblin/migrations/0003_auto_20150429_1422.py","file_name":"0003_auto_20150429_1422.py","file_ext":"py","file_size_in_byte":753,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"29318223057","text":"import sys\nimport tempfile\n\nfrom envprobe.community_descriptions import downloader, local_data\nfrom envprobe.vartypes import EnvVarExtendedInformation\n\n\nname = 'descriptions'\ndescription = \\\n \"\"\"\n The 'descriptions' subcommand allows access to the \"Envprobe Variable\n Descriptions Knowledge Base\" project.\n \"\"\"\nhelp = \"Access facilities related to the \\\"variable desciptions\\\" project\"\n\nupdate_name = 'update'\nupdate_description = \\\n \"\"\"\n Check if the descriptions project has a newer version available.\n If there is, the data will be downloaded and extracted and subsequent calls\n to Envprobe will behave according to the new information, if no\n user-specific settings exist.\n\n This command requires Internet access.\n \"\"\"\nupdate_help = \"Download varaible descriptions from the Internet.\"\n\nepilogue = \\\n \"\"\"\n The canonical repository for this sister project is avaiable at:\n http://github.com/whisperity/Envprobe-Descriptions\n \"\"\"\n\n\ndef update_command(args):\n print(\"Checking for latest version of the Envprobe Variable Descriptions \"\n \"Knowledge Base project.\")\n storage_cfg = local_data.get_storage_configuration(read_only=False)\n new_version = downloader.fetch_latest_version_information()\n\n if new_version == storage_cfg.version:\n # Right now, we use a simple equality check, because the versions are\n # pure commit IDs.\n print(\"Nothing to update - the latest data is already available.\")\n return\n\n # Open all the variable information managers for the saved data. This is\n # needed so we can gather which keys to delete.\n variables_to_clear = set()\n for manager in local_data.generate_variable_information_managers():\n variables_to_clear.update(manager.keys())\n\n with tempfile.TemporaryDirectory(prefix=\"envprobe-community-kb-\") as tempd:\n sources = downloader.download_latest_data(tempd)\n for source in sources:\n print(\"Extracting '{}'...\".format(source.name))\n try:\n source.parse()\n except Exception as e:\n print(\"[WARN] Failed to parse '{}':\\t{}\"\n .format(source.name, str(e)),\n file=sys.stderr)\n continue\n\n storage_cfg.set_comment_for(source.name, source.comment)\n set_vars = 0\n for variable in source:\n information = EnvVarExtendedInformation()\n information.apply(source[variable])\n\n try:\n manager = local_data.get_variable_information_manager(\n variable, read_only=False)\n manager.set(variable, information, source.name)\n set_vars += 1\n\n try:\n variables_to_clear.remove(variable)\n except KeyError:\n pass\n except Exception as e:\n print(\"[WARN] Failed to update configuration for \"\n \"'{}':\\t{}\".format(variable, str(e)),\n file=sys.stderr)\n continue\n print(\"\\textracted {} variables.\".format(set_vars))\n\n print(\"Cleaning up old information...\")\n set_vars = 0\n for variable in variables_to_clear:\n try:\n manager = local_data.get_variable_information_manager(\n variable, read_only=False)\n del manager[variable]\n set_vars += 1\n except Exception as e:\n print(\"[WARN] Failed to clean up '{}':\\t{}\".format(variable,\n str(e)),\n file=sys.stderr)\n print(\"\\tcleaned up {} records.\".format(set_vars))\n\n storage_cfg.version = new_version\n\n\ndef register_update(argparser):\n parser = argparser.add_parser(\n name=update_name,\n description=update_description,\n help=update_help,\n epilog=epilogue\n )\n\n parser.set_defaults(func=update_command)\n\n\ndef register(argparser, shell):\n parser = argparser.add_parser(\n name=name,\n description=description,\n help=help,\n epilog=epilogue\n )\n subparsers = parser.add_subparsers(\n title=\"available_commands\")\n\n register_update(subparsers)\n","repo_name":"whisperity/envprobe","sub_path":"src/envprobe/commands/descriptions.py","file_name":"descriptions.py","file_ext":"py","file_size_in_byte":4354,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"7"} +{"seq_id":"33670279844","text":"\"\"\"\nBase settings to build other settings files upon.\n\"\"\"\nimport os\n\nfrom pathlib import Path\n\nimport environ\n\nfrom django.utils.translation import gettext_lazy as _\n\n\nROOT_DIR = Path(__file__).resolve(strict=True).parent.parent\nAPPS_DIR = ROOT_DIR / \"ghibil\"\nenv = environ.Env()\n\nREAD_DOT_ENV_FILE = env.bool(\"DJANGO_READ_DOT_ENV_FILE\", default=True)\nif READ_DOT_ENV_FILE:\n # OS environment variables take precedence over variables from .env\n env.read_env(str(ROOT_DIR / \".env\"))\n\n# GENERAL\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#debug\nDEBUG = env.bool(\"DJANGO_DEBUG\", False)\n# Local time zone. Choices are\n# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name\n# though not all of them may be available with every OS.\n# In Windows, this must be set to your system time zone.\nTIME_ZONE = \"UTC\"\n# https://docs.djangoproject.com/en/dev/ref/settings/#language-code\nLANGUAGE_CODE = \"en-us\"\n# https://docs.djangoproject.com/en/dev/ref/settings/#site-id\nSITE_ID = 1\n# https://docs.djangoproject.com/en/dev/ref/settings/#use-i18n\nUSE_I18N = True\n# https://docs.djangoproject.com/en/dev/ref/settings/#use-l10n\nUSE_L10N = True\n# https://docs.djangoproject.com/en/dev/ref/settings/#use-tz\nUSE_TZ = True\n# https://docs.djangoproject.com/en/dev/ref/settings/#languages\nLANGUAGES = [(\"en-us\", _(\"English\")), (\"de-de\", _(\"German\"))]\n# https://docs.djangoproject.com/en/dev/ref/settings/#locale-paths\nLOCALE_PATHS = [str(ROOT_DIR / \"locale\")]\n\n# DATABASES\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#databases\n\nDATABASES = {\n \"default\": {\n \"ENGINE\": \"django.db.backends.sqlite3\",\n \"NAME\": os.path.join(ROOT_DIR, \"db.sqlite3\"),\n },\n}\n# URLS\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#root-urlconf\nROOT_URLCONF = \"config.urls\"\n\n# APPS\n# -----------------------------------------------------------------------------\nDJANGO_APPS = [\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"django.contrib.staticfiles\",\n \"django.contrib.admin\",\n]\nTHIRD_PARTY_APPS = [\n \"rest_framework\",\n]\n\nLOCAL_APPS = [\n \"ghibil.movies.apps.MoviesAppConfig\",\n]\n# https://docs.djangoproject.com/en/dev/ref/settings/#installed-apps\nINSTALLED_APPS = DJANGO_APPS + THIRD_PARTY_APPS + LOCAL_APPS\n\n# MIDDLEWARE\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#middleware\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\n# TEMPLATES\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#templates\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ],\n },\n },\n]\n\nSECRET_KEY = env(\n \"DJANGO_SECRET_KEY\",\n default=\"LJvMIEyCZ4lM9onk5EyhaeioR6R9CHJ3VaMiCsBepsA9ER6CQp1lNPIBRnX6Ax3h\",\n)\n# https://docs.djangoproject.com/en/dev/ref/settings/#allowed-hosts\nALLOWED_HOSTS = [\"localhost\", \"0.0.0.0\", \"127.0.0.1\"]\n\n# CACHES\n# -----------------------------------------------------------------------------\n# https://docs.djangoproject.com/en/dev/ref/settings/#caches\nCACHES = {\n \"default\": {\n \"BACKEND\": \"django_redis.cache.RedisCache\",\n \"LOCATION\": env.get_value(\n var=\"REDIS_URL\", default=\"redis://127.0.0.1:6379/0\",\n ),\n \"OPTIONS\": {\n \"CLIENT_CLASS\": \"django_redis.client.DefaultClient\",\n },\n },\n}\nSTATIC_URL = \"/static/\"\n","repo_name":"abtinmo/ghibil","sub_path":"config/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":4474,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"74863115741","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Aug 15 15:33:25 2017\n\n@author: oreilly\n\"\"\"\n\nfrom .modelingParameter import getParameterTypeNameFromID\n\n# Equivalence rules. These are define such that\n# parameterEquivalenceRules[idFrom][idTo] is giving a rule to convert\n# values from an entity idFrom to an entity idTo. This rule is specified as \n# a lambda function that that has to be applies to the values of the \n# given parameters.\nparameterEquivalenceRules = {}\n\n#\"BBP-131005\":volume_brain_region\n#\"BBP-131006\":volume_unilateral_brain_region\nparameterEquivalenceRules[\"BBP-131005\"] = {\"BBP-131006\":lambda x: x/2}\nparameterEquivalenceRules[\"BBP-131006\"] = {\"BBP-131005\":lambda x: x*2}\n\n\n\nclass EquivalenceFinder:\n \n def __init__(self, condition):\n self.condition = condition\n\n\n def run(self):\n # Apply rules for parameter equivalences\n return self.applyParameterEquivalence(self.condition)\n \n \n def applyParameterEquivalence(self, condition):\n \n for idFrom in parameterEquivalenceRules:\n for idTo in parameterEquivalenceRules[idFrom]:\n condition.addEquivalences(\"Parameter type ID\", idFrom, idTo, parameterEquivalenceRules[idFrom][idTo])\n condition.addEquivalences(\"Parameter name\", \n getParameterTypeNameFromID(idFrom), \n getParameterTypeNameFromID(idTo), \n parameterEquivalenceRules[idFrom][idTo])\n \n return condition","repo_name":"BlueBrain/nat","sub_path":"nat/equivalenceFinder.py","file_name":"equivalenceFinder.py","file_ext":"py","file_size_in_byte":1553,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"7"} +{"seq_id":"27696572637","text":"### Load the data ###\n\nlabels = ['normal','abnormal']\nimage_size = 224\n\n# Borrowed \"get_data\" function from\n# https://www.analyticsvidhya.com/blog/2020/10/create-image-classification-model-python-keras/\n\nimport os \nimport cv2 \nimport numpy as np\n\ndef get_data(data_dir):\n data = []\n for label in labels:\n path = os.path.join(data_dir,label)\n class_num = labels.index(label)\n for img in os.listdir(path):\n try:\n img_arr = cv2.imread(os.path.join(path,img))[...,::-1] # Convert BGR to RGB format\n resized_arr = cv2.resize(img_arr, (image_size, image_size)) # Reshape images to preferred size\n data.append([resized_arr, class_num])\n except Exception as e:\n print(e)\n return np.array(data)\n\ntrain = get_data('../two_class_post_weld/train') ## define current path to the train folder\nval = get_data('../two_class_post_weld/test') ## define current path to the test folder\n\n### Compare the number of the images in both cases ###\n\nimport seaborn as sns\n\ncount_number = []\nfor i in train:\n if(i[1]==0):\n count_number.append(\"normal\")\n else:\n count_number.append(\"abnormal\")\n\nsns.set_style('whitegrid')\nsns.countplot(count_number)\n\n### Visualize a random image from both classes ###\n\nimport matplotlib.pyplot as plt\n\n# visualize a normal welding image\nplt.figure(figsize = (5,5))\nplt.imshow(train[1][0])\nplt.title(labels[train[0][1]])\n\n# visualize a abnormal welding image\nplt.figure(figsize = (5,5))\nplt.imshow(train[-1][0])\nplt.title(labels[train[-1][1]])\n\n### Data Preprocessing ###\n\n# split the data into feature and label\n\nx_train = []\ny_train = []\nx_val = []\ny_val = []\n\nfor feature, label in train:\n x_train.append(feature)\n y_train.append(label)\n\nfor feature, label in val:\n x_val.append(feature)\n y_val.append(label)\n\n# normalize the data\n\nimport numpy as np\n\nx_train = np.array(x_train) / 255\nx_val = np.array(x_val) / 255\n\nx_train.reshape(-1, image_size, image_size, 1)\ny_train = np.array(y_train)\n\nx_val.reshape(-1, image_size, image_size,1)\ny_val = np.array(y_val)\n\n# data augmentation \n\nfrom keras.preprocessing.image import ImageDataGenerator\n\ndata_generator = ImageDataGenerator(\n featurewise_center=False, # False to input mean zero\n samplewise_center=False, # False tosample mean zero\n featurewise_std_normalization=False, # False to divide input by its standard deviation\n samplewise_std_normalization=False, # False to divide sample by its standard deviation\n zca_whitening=False, # False to apply ZCA whitening\n rotation_range=25, # rotate images in 25 degree\n zoom_range=0.3, # zoom images 0.3 times\n width_shift_range=0.1, # shift images 0.1 horizontally (fraction of total width)\n height_shift_range=0.1, # shift images 0.1 vertically (fraction of total height)\n horizontal_flip=True, # flip images horizontally\n vertical_flip=True) # flip images vertically\n\ndata_generator.fit(x_train)\n\n### Build the model ###\n\nimport keras\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, MaxPool2D, Flatten, Dropout\n\nmodel = Sequential()\nmodel.add(Conv2D(8,3,padding=\"same\",activation=\"relu\",input_shape=(224,224,3)))\nmodel.add(MaxPool2D())\nmodel.add(Conv2D(16,3,padding=\"same\",activation=\"relu\"))\nmodel.add(MaxPool2D())\nmodel.add(Conv2D(32,3,padding=\"same\",activation=\"relu\"))\nmodel.add(MaxPool2D())\nmodel.add(Conv2D(64,3,padding=\"same\",activation=\"relu\"))\nmodel.add(MaxPool2D())\nmodel.add(Dropout(0.4))\n\nmodel.add(Flatten())\nmodel.add(Dense(128,activation=\"relu\"))\nmodel.add(Dense(2,activation=\"softmax\"))\n\nmodel.summary()\n\n### Evaluate the results ###\n\nimport tensorflow as tf\nfrom keras.optimizers import SGD, Adam\n\nopt = Adam(lr=0.000001)\nmodel.compile(optimizer=opt, loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True), metrics=['accuracy'])\n\nhistory = model.fit(x_train, y_train, epochs=500, validation_data=(x_val,y_val))\n\nacc= history.history['accuracy']\nval_acc = history.history['val_accuracy']\nloss = history.history['loss']\nval_loss = history.history['val_loss']\n\nepochs_range = range(500)\n\nplt.figure(figsize=(15,15))\nplt.subplot(2,2,1)\nplt.plot(epochs_range, acc, label='Training Accuracy')\nplt.plot(epochs_range, val_acc, label='Validation Accuracy')\nplt.legend(loc='lower right')\nplt.title('Training and Validation Accuracy')\n\nplt.subplot(2,2,2)\nplt.plot(epochs_range, loss, label='Training Loss')\nplt.plot(epochs_range, val_loss, label='Validation Loss')\nplt.legend(loc='upper right')\nplt.title('Training and Validation Loss')\nplt.show()\n\nfrom sklearn.metrics import classification_report\n\npredictions = model.predict_classes(x_val)\npredictions = predictions.reshape(1,-1)[0]\nprint(classification_report(y_val, predictions, target_names = ['normal welding (Class 0)','abnormal welding(Class 1)']))\n","repo_name":"ykim836/ECE","sub_path":"post_welding_code.py","file_name":"post_welding_code.py","file_ext":"py","file_size_in_byte":4839,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13695741484","text":"import argparse\nimport binascii\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"addr\", type=str)\nparser.add_argument(\"pad\", type=int)\n\ndef write_pld(addr, pad):\n\tfw = open('pld', 'wb')\n\t\n\t# Write shellcode\n\tfw.write(b'\\x31\\xc0')\n\tfw.write(b'\\x50')\n\tfw.write(b'\\x68\\x6e\\x2f\\x73\\x68')\n\tfw.write(b'\\x68\\x2f\\x2f\\x62\\x69')\n\tfw.write(b'\\x89\\xe3')\n\tfw.write(b'\\x99')\n\tfw.write(b'\\x52')\n\tfw.write(b'\\x56')\n\tfw.write(b'\\x89\\xe1')\n\tfw.write(b'\\xb0\\x0b')\n\tfw.write(b'\\xcd\\x80')\n\n\t# Write pad\n\tfor i in range(pad):\n\t\tfw.write(b'\\x90')\n\n\t# Convert and write addr\n\t# 0xabcdefgh -> \\xgh\\xef\\xcd\\xab\n\tformat_str = \"{gh}{ef}{cd}{ab}\"\n\taddr_formatted = format_str.format(ab=addr[:2],\n\t\t\t\t\t cd=addr[2:4],\n\t\t\t\t\t ef=addr[4:6],\n\t\t\t\t\t gh=addr[6:])\n\tprint(addr_formatted)\n\taddr_bytes = binascii.a2b_hex(addr_formatted)\n\tfw.write(addr_bytes)\n\n\tfw.close()\n\n\nif __name__ == '__main__':\n\targs = parser.parse_args()\n\taddr = args.addr\n\twhile len(addr) < 8:\n\t\taddr = \"0\" + addr\n\twrite_pld(addr, args.pad)\n\n","repo_name":"JaredDobry/brute_shellcode","sub_path":"shellcode.py","file_name":"shellcode.py","file_ext":"py","file_size_in_byte":990,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"72685341664","text":"from typing import Optional, List\nfrom collections import deque\nfrom BinaryTree.TreeNodeModule import TreeNode\n\n# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, val=0, left=None, right=None):\n# self.val = val\n# self.left = left\n# self.right = right\n\nclass Solution:\n def zigzagLevelOrder(self, root: Optional[TreeNode]) -> List[List[int]]:\n if not root:\n return []\n\n # node and depth\n queue = deque([(root, 1)])\n answer = []\n while queue:\n node, depth = queue.pop()\n if node.left:\n queue.appendleft((node.left, depth + 1))\n if node.right:\n queue.appendleft((node.right, depth + 1))\n \n if depth > len(answer):\n answer.append([])\n answer[-1].append(node.val)\n \n return [level if depth % 2 == 0 else list(reversed(level)) for depth, level in enumerate(answer)]\n","repo_name":"daviddwlee84/LeetCode","sub_path":"Python3/BinaryTree/BinaryTreeZigzagLevelOrderTraversal/BFS103.py","file_name":"BFS103.py","file_ext":"py","file_size_in_byte":987,"program_lang":"python","lang":"en","doc_type":"code","stars":15,"dataset":"github-code","pt":"7"} +{"seq_id":"94528515","text":"\"\"\" Python 3 compatibility tools. \"\"\"\nfrom __future__ import division, print_function\nimport itertools\nimport sys\nimport os\nfrom io import BytesIO, IOBase\n\n\nif sys.version_info[0] < 3:\n input = raw_input\n range = xrange\n\n filter = itertools.ifilter\n map = itertools.imap\n zip = itertools.izip\n\n\ndef is_it_local():\n script_dir = str(os.getcwd()).split('/')\n username = \"dipta007\"\n return username in script_dir\n\n\ndef READ(fileName):\n if is_it_local():\n sys.stdin = open(f'./{fileName}', 'r')\n\n# region fastio\nBUFSIZE = 8192\n\nclass FastIO(IOBase):\n newlines = 0\n\n def __init__(self, file):\n self._fd = file.fileno()\n self.buffer = BytesIO()\n self.writable = \"x\" in file.mode or \"r\" not in file.mode\n self.write = self.buffer.write if self.writable else None\n\n def read(self):\n while True:\n b = os.read(self._fd, max(os.fstat(self._fd).st_size, BUFSIZE))\n if not b:\n break\n ptr = self.buffer.tell()\n self.buffer.seek(0, 2), self.buffer.write(b), self.buffer.seek(ptr)\n self.newlines = 0\n return self.buffer.read()\n\n def readline(self):\n while self.newlines == 0:\n b = os.read(self._fd, max(os.fstat(self._fd).st_size, BUFSIZE))\n self.newlines = b.count(b\"\\n\") + (not b)\n ptr = self.buffer.tell()\n self.buffer.seek(0, 2), self.buffer.write(b), self.buffer.seek(ptr)\n self.newlines -= 1\n return self.buffer.readline()\n\n def flush(self):\n if self.writable:\n os.write(self._fd, self.buffer.getvalue())\n self.buffer.truncate(0), self.buffer.seek(0)\n\n\nclass IOWrapper(IOBase):\n def __init__(self, file):\n self.buffer = FastIO(file)\n self.flush = self.buffer.flush\n self.writable = self.buffer.writable\n self.write = lambda s: self.buffer.write(s.encode(\"ascii\"))\n self.read = lambda: self.buffer.read().decode(\"ascii\")\n self.readline = lambda: self.buffer.readline().decode(\"ascii\")\n\nif not is_it_local():\n sys.stdin, sys.stdout = IOWrapper(sys.stdin), IOWrapper(sys.stdout)\n input = lambda: sys.stdin.readline().rstrip(\"\\r\\n\")\n\n# endregion\n\n\ndef input1(type=int):\n return type(input())\n\n\ndef input2(type=int):\n [a, b] = list(map(type, input().split()))\n return a, b\n\n\ndef input3(type=int):\n [a, b, c] = list(map(type, input().split()))\n return a, b, c\n\n\ndef input_array(type=int):\n return list(map(type, input().split()))\n\n\ndef input_string():\n s = input()\n return list(s)\n\n##############################################################\n\ndef is_it_okay(st01, st10):\n return abs(st01 - st10) <= 1\n\ndef is_it_double(st, mp):\n if mp.get(st[::-1], 0) == 1:\n return True\n return False\n\ndef main():\n t = input1()\n for ci in range(t):\n n = input1()\n\n mat = []\n st01, st10 = 0, 0\n mp = {}\n for i in range(n):\n st = input()\n now = [st]\n mp[st] = 1\n\n if st[0] == '0' and st[-1] == '1':\n st01 += 1\n now.append(0)\n now.append(1)\n elif st[0] == '1' and st[-1] == '0':\n st10 += 1\n now.append(1)\n now.append(0)\n else:\n now.append(ord(st[0]) - 48)\n now.append(ord(st[-1]) - 48)\n \n mat.append(now)\n\n swapped = []\n # print(st01, st10)\n\n # No 0...1 or 1...0\n if st01 == 0 and st10 == 0:\n flg = 0\n nw = mat[0]\n for [st, a, b] in mat:\n if nw[1] != a or nw[2] != b:\n print(-1)\n flg = 1\n break\n if not flg:\n print(0)\n print(\"\")\n elif is_it_okay(st01, st10):\n print(0)\n print(\"\")\n else:\n for i, [s, a, b] in enumerate(mat):\n if st01 > st10 + 1 and a == 0 and b == 1 and not is_it_double(s, mp):\n swapped.append(i)\n mp[s] -= 1\n mp[s[::-1]] = 1\n st01 -= 1\n st10 += 1\n elif st10 > st01 + 1 and a == 1 and b == 0 and not is_it_double(s, mp):\n swapped.append(i)\n mp[s] -= 1\n mp[s[::-1]] = 1\n st10 -= 1\n st01 += 1\n\n if is_it_okay(st01, st10):\n print(len(swapped))\n print(\" \".join([str(x+1) for x in swapped]))\n else:\n print(-1)\n\n\n pass\n\nif __name__ == '__main__':\n # READ('in.txt')\n main()","repo_name":"dipta007/Competitive-Programming","sub_path":"Codeforces/Practice/1277D.py","file_name":"1277D.py","file_ext":"py","file_size_in_byte":4149,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"7"} +{"seq_id":"28152643454","text":"#!/usr/bin/env python3\r\n\"\"\"Take the code from wait_n and alter it into a new function task_wait_n\"\"\"\r\nfrom typing import List\r\n\r\ntask_wait_random = __import__('3-tasks').task_wait_random\r\n\r\n\r\nasync def task_wait_n(n: int, max_delay: int) -> List[float]:\r\n \"\"\"task_wait_n\"\"\"\r\n wList = []\r\n for i in range(n):\r\n wList.append(await task_wait_random(max_delay))\r\n return sorted(wList)\r\n ","repo_name":"maycolroa/holbertonschool-web_back_end","sub_path":"0x01-python_async_function/4-tasks.py","file_name":"4-tasks.py","file_ext":"py","file_size_in_byte":405,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"31237061872","text":"import json\n\nfrom http.server import BaseHTTPRequestHandler\nfrom package.mongoServer import MongoServer\n\nDOCUMENT_MAIN_KEY = \"document\"\n\nclass Server(BaseHTTPRequestHandler):\n def __init__(self, request, client_address, server):\n print(\"testing\")\n self._mongo_server = MongoServer()\n super().__init__(request, client_address, server)\n\n def do_GET(self):\n self._set_response()\n\n def do_POST(self):\n document_post: dict = self._get_json_content_from_request()\n if document_post.get(DOCUMENT_MAIN_KEY) != None:\n self._mongo_server.write_document_to_database(document_post.get(DOCUMENT_MAIN_KEY))\n self._set_response()\n else:\n self._set_error_response(\"The json request does not have an document key!\")\n\n\n def _get_json_content_from_request(self) -> object:\n content_type = self.headers.get(\"content-type\")\n if content_type != \"application/json\":\n self._set_error_response(\"Bad formatted request! It is not an application/json.\")\n else:\n content_length = int(self.headers.get(\"content-length\"))\n return json.loads(self.rfile.read(content_length))\n\n def _set_response(self):\n self.send_response(200)\n self.end_headers()\n\n def _set_error_response(self, message: str):\n self.send_response(400, message)\n self.end_headers()","repo_name":"GoldenCodeRam/Project-Olive","sub_path":"database-backup-server/src/package/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":1293,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"24913178583","text":"from math import log2\nfrom typing import Union\n\nimport torch\nfrom numpy import ndarray\nfrom torch import Tensor\n\nfrom .utils import my_clamp, my_round\n\n\ndef _integer_quantize(\n x: Union[Tensor, ndarray], width: int, frac_width: int, is_signed: bool = True\n):\n \"\"\"\n - Do linear quantization to input according to a scale and number of bits\n - Note that `bias` can be negative or larger than `bits`\n\n ---\n - forward: convert IEEE FP32/64 to fixed-point\n - backward: STE\n\n ---\n width: the bit width of the fixed-point number\n frac_width: the number of fractional bits. Note that `bias` can be negative or larger than `bits`\n\n ---\n For example: 0b101 . 00111, bits = 8, bias = 5\n\n \"\"\"\n if is_signed:\n int_min = -(2 ** (width - 1))\n int_max = 2 ** (width - 1) - 1\n else:\n int_min = 0\n int_max = 2**width - 1\n # thresh = 2 ** (width - 1)\n scale = 2**frac_width\n\n if isinstance(x, (Tensor, ndarray)):\n return my_clamp(my_round(x.mul(scale)), int_min, int_max).div(scale)\n elif isinstance(x, int):\n return x\n else:\n return my_clamp(my_round(x * scale), int_min, int_max) / scale\n\n\nclass IntegerQuantize(torch.autograd.Function):\n @staticmethod\n def forward(ctx, x, width, frac_width, is_signed):\n return _integer_quantize(\n x, width=width, frac_width=frac_width, is_signed=is_signed\n )\n\n @staticmethod\n def backward(ctx, grad_output):\n \"\"\"\n STE shortcut for saving GPU memory\n \"\"\"\n return grad_output, None, None, None\n\n\ndef integer_quantizer(\n x: Union[Tensor, ndarray], width: int, frac_width: int, is_signed: bool = True\n):\n \"\"\"\n - Do linear quantization to input according to a scale and number of bits\n - Note that `bias` can be negative or larger than `bits`\n\n ---\n - forward: convert IEEE FP32/64 to fixed-point\n - backward: STE\n\n ---\n width: the bit width of the fixed-point number\n frac_width: the number of fractional bits. Note that `bias` can be negative or larger than `bits`\n\n ---\n For example: 0b101 . 00111, bits = 8, bias = 5\n\n \"\"\"\n return IntegerQuantize.apply(x, width, frac_width, is_signed)\n\n\ndef integer_fraction(\n width: int, frac_choices: list, min_value: float, max_value: float\n):\n max_half_range = max(abs(min_value), abs(max_value))\n int_width = int(log2(max(0.5, max_half_range))) + 2\n frac_width = max(0, width - int_width)\n frac_width = max(filter(lambda x: x <= frac_width, frac_choices))\n return frac_width\n","repo_name":"ChengZhang-98/llm-mixed-q","sub_path":"src/llm_mixed_q/models/quantize/quantizers/integer.py","file_name":"integer.py","file_ext":"py","file_size_in_byte":2566,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"71352540384","text":"from kubernetes import client, config\nfrom kubernetes.client import ApiClient\nfrom kubernetes.client.rest import ApiException\nimport yaml\nimport json\n\nconfig.load_kube_config()\nv1 = client.CoreV1Api()\n\ndef __format_data_for_create_configmap(client_output):\n temp_dict={}\n temp_list=[]\n json_data=ApiClient().sanitize_for_serialization(client_output)\n \n if type(json_data) is str:\n print(\"FORMAT_DATA :{}\".format(type(json_data)))\n json_data = json.loads(json_data)\n temp_list.append(json_data)\n return temp_list\n\ndef create_configmap(yaml_body=None,namespace=\"default\"):\n try:\n yaml_data=open(\"config.yaml\", \"rb\").read().decode('utf-8')\n yaml_body=yaml.safe_load(yaml_data)\n resp = v1.create_namespaced_config_map(\n body=yaml_body, namespace=\"{}\".format(namespace))\n\n data=__format_data_for_create_configmap(resp)\n print (data) \n except ApiException as e:\n print(\"ERROR IN create_configmap:\\n{}\".format(e.body))\n print(\"TYPE :{}\".format(type(e)))\n return __format_data_for_create_configmap(e.body)\n\ndef update_configmap(k8s_object_name=None,yaml_body=None,namespace=\"default\"):\n try:\n yaml_data=open(\"change_config.yaml\", \"rb\").read().decode('utf-8')\n yaml_body=yaml.safe_load(yaml_data)\n resp = v1.patch_namespaced_config_map(\n name=k8s_object_name,\n body=yaml_body, \n namespace=\"{}\".format(namespace))\n\n data=__format_data_for_create_configmap(resp)\n return data\n except ApiException as e:\n print(\"ERROR IN create_deployment:\\n{}\".format(e.body))\n print(\"TYPE :{}\".format(type(e)))\n return __format_data_for_create_configmap(e.body)\n\ndef delete_configmap(k8s_object_name=None,namespace=\"default\"):\n try:\n resp = v1.delete_namespaced_config_map(\n name=k8s_object_name,\n namespace=\"{}\".format(namespace),\n body=v1.delete_namespaced_config_map(k8s_object_name, namespace)\n )\n except ApiException as e:\n print(\"ERROR IN create_deployment:\\n{}\".format(e.body))\n print(\"TYPE :{}\".format(type(e)))\n return __format_data_for_create_configmap(e.body)\n\n# export secret=$(kubectl get serviceaccount default -o json | jq -r '.secrets[].name')\n# kubectl get secret $secret -o yaml | grep \"token:\" | awk {'print $2'} | base64 -d > token\n# APISERVER=$(kubectl config view | grep server | cut -f 2- -d \":\" | tr -d \" \")\nif __name__ == '__main__':\n create_configmap(\"default\")\n #update_configmap(k8s_object_name=\"munish\")\n #delete_configmap(k8s_object_name=\"munish\")","repo_name":"mumehta/python-k8s","sub_path":"test_config.py","file_name":"test_config.py","file_ext":"py","file_size_in_byte":2692,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"21691335937","text":"from django.urls import path\nfrom rest_framework_simplejwt.views import (\n TokenObtainPairView,\n TokenRefreshView,\n)\n\nfrom . import views\n\nurlpatterns = [\n path('', views.posts),\n path('create/', views.create_post),\n path('my-posts/', views.my_posts),\n path('/', views.post),\n path('/delete/', views.delete_post),\n path('/edit/', views.edit_post),\n]\n","repo_name":"safwantaliparamba/Blog-App","sub_path":"Backend/api/v1/posts/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":398,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"6962660575","text":"from openerp import api, models\nfrom odoo.exceptions import UserError\n\n\nclass PaymentCnabReport(models.AbstractModel):\n _name = 'report.br_payment_invoice.report_cnab_payment_invoice'\n\n @api.model\n def get_report_values(self, docids, data=None):\n report = self.env['ir.actions.report']._get_report_from_name(\n 'br_payment_invoice.report_cnab_payment_invoice')\n lines = []\n docs = []\n for docid in docids:\n doc = self.env['account.invoice'].browse(docid)\n docs.append(doc)\n lines.append(doc.get_order_line())\n docargs = {\n 'doc_ids': docids,\n 'doc_model': report.model,\n 'docs': self,\n 'lines': lines,\n }\n if not lines:\n raise UserError(\n \"Este documento ainda não possui um comprovante de pagamento.\")\n return docargs\n","repo_name":"mamcode/wbsfoosbr","sub_path":"br_payment_invoice/reports/reports.py","file_name":"reports.py","file_ext":"py","file_size_in_byte":898,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"24333903315","text":"import pygame\r\n\r\npygame.init()\r\n\r\n#display dimensions\r\ndisplay_width = 1200\r\ndisplay_height = 640\r\n\r\n#colour\r\nblack = (0, 0, 0)\r\nwhite = (255, 255, 255)\r\nred = (230, 0, 0)\r\nbright_red = (255, 0, 0)\r\nlight_brown = (139, 69, 19)\r\nmidnight_blue = (25, 25, 112)\r\ngrey = (128, 128, 128)\r\nlight_grey = (192, 192, 192)\r\ncardboard_brown = (165, 136, 85)\r\nmint_green = (152, 255, 152)\r\ntan = (196,144,124)\r\npeach = (255,224,189)\r\n\r\n#player\r\nplx = 500\r\nply = 250\r\nxmov = 0\r\nymov = 0\r\nms = 2\r\n\r\n\r\ndisplayScreen = pygame.display.set_mode((display_width,display_height))\r\npxa = pygame.PixelArray(displayScreen)\r\npygame.display.set_caption('Test')\r\nclock = pygame.time.Clock()\r\n\r\nkeys = pygame.key.get_pressed()\r\n\r\ndef character():\r\n global plx, ply\r\n pygame.draw.rect(displayScreen, tan, (plx+8, ply-10, 20, 10))\r\n head = pygame.draw.ellipse(displayScreen, peach, (plx-7, ply-55, 50, 50))\r\n pygame.draw.rect(displayScreen, black, (plx+8, ply-32, 4, 8))\r\n pygame.draw.rect(displayScreen, black, (plx+24, ply-32, 4, 8))\r\n pygame.draw.rect(displayScreen, black, (plx, ply, 36, 40))\r\n pygame.draw.rect(displayScreen, red, (plx+10, ply, 16, 40))\r\n pygame.draw.polygon(displayScreen, black, ((plx,ply),(plx,ply+10),(plx-22,ply+20)))\r\n pygame.draw.polygon(displayScreen, black, ((plx+36,ply),(plx+36,ply+10),(plx+54,ply+20)))\r\n pygame.draw.rect(displayScreen, black, (plx+3, ply+40, 10, 30))\r\n pygame.draw.rect(displayScreen, black, (plx+22, ply+40, 10, 30))\r\n #pxa[10, 20] = black\r\n\r\n\r\n \r\nwhile True:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n pygame.quit()\r\n quit()\r\n\r\n xmov = 0\r\n ymov = 0\r\n ms = 1\r\n\r\n\r\n keys = pygame.key.get_pressed()\r\n if keys[pygame.K_a]:\r\n xmov = -1\r\n if keys[pygame.K_d]:\r\n xmov = 1\r\n if keys[pygame.K_w]:\r\n ymov = -1\r\n if keys[pygame.K_s]:\r\n ymov = 1\r\n\r\n\r\n plx += xmov * ms\r\n ply += ymov * ms\r\n\r\n displayScreen.fill(white)\r\n character()\r\n pygame.display.update()\r\n\r\n","repo_name":"josephz03/CPT-Game","sub_path":"Development/Movement.py","file_name":"Movement.py","file_ext":"py","file_size_in_byte":2042,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"33390912641","text":"from lxml import etree as et\nimport nltk\nimport xml.etree.cElementTree as ET\nimport csv\nimport ast\nimport time\nfrom nltk import word_tokenize\n\n\n# necessary for named entity recognition. The xml corpus is reformed into plain text format\ndef create_text():\n parser = et.XMLParser(encoding='iso-8859-5', recover=True)\n tree = et.parse(\"../corpora/ukwac1_fixed.xml\", parser)\n\n root = tree.getroot()\n\n with open(\"../xml_as_text_for_ner.txt\", \"w\") as txt_out:\n\n for txt in root.findall('text'):\n for sentence in txt.findall('s'):\n sent = \"\"\n\n is_single_open = False\n\n line = sentence.text.split('\\n')\n\n for word, next_word in zip(line, line[1:] + [line[0]]):\n if word == '':\n continue\n\n word = word.split('\\t')\n next_word = next_word.split('\\t')\n\n if next_word[0] in [\",\", \".\", \"'\", \")\"]:\n sent += word[0]\n elif next_word[0] == \"n't\":\n sent += word[0]\n elif next_word[0] == \"'re\":\n sent += word[0]\n elif next_word[0] == \"'s\":\n sent += word[0]\n elif next_word[0] == \"'ll\":\n sent += word[0]\n elif next_word[0] == \":\":\n sent += word[0]\n elif word[0] == \"(\":\n sent += word[0]\n elif word[0] == \"'\":\n if is_single_open:\n sent += word[0] + \" \"\n is_single_open = False\n else:\n sent += word[0]\n is_single_open = True\n else:\n sent += word[0] + \" \"\n txt_out.write(sent+\"\\n\")\n\n\n# creates a vocabulary xml file for easier processing\ndef create_vocabulary_xml():\n with open(\"../O2_match_vocabulary_to_sentences/matched_vocabulary.csv\", 'r') as voc_in:\n sent_reader = csv.reader(voc_in, delimiter=';')\n next(sent_reader)\n\n voc_root = ET.Element(\"root\")\n pos_root = ET.Element(\"root\")\n\n pos_list = []\n for v_row in sent_reader:\n\n voc_doc = ET.SubElement(voc_root, \"vocable\", name=v_row[1].replace(\"\\“\", \""\"))\n ET.SubElement(voc_doc, \"lemma\", name=v_row[5].replace(\"\\“\", \""\"))\n ET.SubElement(voc_doc, \"chapter\", name=v_row[15])\n ET.SubElement(voc_doc, \"book\", name=v_row[16])\n ET.SubElement(voc_doc, \"pos\", name=(str(v_row[7]).replace(\"\\“\", \""\")))\n\n tuple_pair = ast.literal_eval(v_row[7])\n #print(tuple_pair)\n for tp in tuple_pair:\n if isinstance(tp, list):\n print(tp)\n for t in tp:\n #print(t)\n pos_list.append((str(t[0]).replace(\"\\“\", \""\"),\n str(t[1]).replace(\"\\“\", \""\"),\n v_row[15],\n v_row[16]))\n else:\n #print(tp)\n pos_list.append((str(tp[0]).replace(\"\\“\", \""\"),\n str(tp[1]).replace(\"\\“\", \""\"),\n v_row[15],\n v_row[16]))\n\n pos_list_set = list(set(pos_list))\n for pos_tuple in pos_list_set:\n pos_doc = ET.SubElement(pos_root, \"lemma\", name=pos_tuple[0])\n ET.SubElement(pos_doc, \"pos\", name=pos_tuple[1])\n ET.SubElement(pos_doc, \"chapter\", name=pos_tuple[2])\n ET.SubElement(pos_doc, \"book\", name=pos_tuple[3])\n tree_voc = ET.ElementTree(voc_root)\n tree_pos = ET.ElementTree(pos_root)\n tree_voc.write(\"./input/voc.xml\")\n tree_pos.write(\"./input/lemma_chap_book_pos.xml\")\n voc_in.close()\n\nif __name__ == \"__main__\":\n # create_text()\n create_vocabulary_xml()\n","repo_name":"Mells/Preprocess03","sub_path":"O3_extract_sentences_from_corpus/00_preprocess.py","file_name":"00_preprocess.py","file_ext":"py","file_size_in_byte":4138,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"40514565210","text":"import numpy as np\n\ndef selectThreshold(yval, pval):\n #SELECTTHRESHOLD Find the best threshold (epsilon) to use for selecting\n #outliers\n # [bestEpsilon bestF1] = SELECTTHRESHOLD(yval, pval) finds the best\n # threshold to use for selecting outliers based on the results from a\n # validation set (pval) and the ground truth (yval).\n #\n \n bestEpsilon = 0\n bestF1 = 0\n F1 = 0\n \n stepsize = (max(pval) - min(pval)) / 1000\n for epsilon in np.arange(min(pval),max(pval),stepsize):\n # ====================== YOUR CODE HERE ======================\n # Instructions: Compute the F1 score of choosing epsilon as the\n # threshold and place the value in F1. The code at the\n # end of the loop will compare the F1 score for this\n # choice of epsilon and set it to be the best epsilon if\n # it is better than the current choice of epsilon.\n # \n # Note: You can use predictions = (pval < epsilon) to get a binary vector\n # of 0's and 1's of the outlier predictions\n \n tp, fp, fn = 0,0,0\n prec, rec = 0,0\n \n predictions = pval < epsilon\n predictions = np.reshape(predictions, (yval.shape))\n\n # predict anomaly & actual anomaly\n tp = sum((predictions == 1) & (yval == 1))\n \n # predict anomaly & actual nonanomaly\n fp = sum((predictions == 1) & (yval == 0))\n \n # predict nonanomaly & actual anomaly\n fn = sum((predictions == 0) & (yval == 1))\n\n # RuntimeWarning: invalid value encountered in true_divide\n # 0으로 나누는 것을 방지하기 위해 dummy 를 분모에 더해준다.\n dummy = 1e-7\n\n prec = tp / (tp + fp + dummy)\n rec = tp / (tp + fn + dummy)\n F1 = 2 * prec * rec / (prec + rec + dummy)\n # =============================================================\n \n if F1 > bestF1:\n bestF1 = F1\n bestEpsilon = epsilon\n \n return bestEpsilon, bestF1","repo_name":"jo1jun/Machine-Learning-Python","sub_path":"ex8/selectThreshold.py","file_name":"selectThreshold.py","file_ext":"py","file_size_in_byte":2113,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"71089736223","text":"# Example of how to work with model predictions...\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nresults_csv = '../data/test_results_transformer_moves.csv'\n# results_csv = '../data/test_results_stockfish_score.csv'\ndf = pd.read_csv(results_csv)\n\np_correct = df[df['move_played'] == True]['p_model']\navg_nll = np.average(-np.log(p_correct))\nprint('NLL:', avg_nll)\n\nplt.hist(p_correct, color='k', bins=np.geomspace(1e-5, 1e0, 21), rwidth=0.8, zorder=10)\nplt.xscale('log')\nplt.grid()\nplt.show()\n","repo_name":"gregdeon/cpsc-540-ml","sub_path":"project/analysis/example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":519,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"12649474057","text":"#!/usr/bin/python3\n\"\"\"\nClass BaseModel module\n\"\"\"\nimport uuid\nfrom datetime import datetime\nimport models\n\n\nclass BaseModel:\n \"\"\"\n Defines all common attributes/methods for other classes\n \"\"\"\n def __init__(self, *args, **kwargs):\n \"\"\"\n Class contractor\n\n Args:\n args - positional arguments\n kwargs - key value pair arguments\n \"\"\"\n \n if kwargs:\n format_str = '%Y-%m-%dT%H:%M:%S.%f'\n for key, value in kwargs.items():\n if key == '__class__':\n continue\n elif key == 'created_at':\n self.created_at = datetime.strptime(\n kwargs['created_at'], format_str)\n elif key == 'updated_at':\n self.updated_at = datetime.strptime(\n kwargs['updated_at'], format_str)\n else:\n setattr(self, key, value)\n else:\n self.id = str(uuid.uuid4())\n self.created_at = datetime.now()\n self.updated_at = datetime.now()\n models.storage.new(self) \n\n def __str__(self):\n \"\"\"Returns string representation of BaseModel instances\"\"\"\n\n clsName = self.__class__.__name__\n return \"[{}] ({}) {}\".format(clsName, self.id, self.__dict__)\n\n def save(self):\n \"\"\"Updates the attribute updated_at with the current datetime\"\"\"\n self.updated_at = datetime.now()\n models.storage.save()\n\n def to_dict(self):\n \"\"\"Returns a dictionary that contains all\n keys values pair of the instance\"\"\"\n my_dict = self.__dict__.copy()\n my_dict['updated_at'] = self.updated_at.isoformat()\n my_dict['created_at'] = self.created_at.isoformat()\n my_dict['__class__'] = self.__class__.__name__\n return my_dict\n","repo_name":"DOREENKDAVID/AirBnB_clone","sub_path":"models/base_model.py","file_name":"base_model.py","file_ext":"py","file_size_in_byte":1867,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26839861583","text":"from rest_framework.permissions import BasePermission, SAFE_METHODS\nfrom rest_framework.request import Request\nfrom rest_framework.views import APIView\nfrom rest_framework.viewsets import GenericViewSet\n\nfrom api.v1.comparisons.models import ComparisonGroup, Comparison\n\n\nclass IsComparisonGroupAuthorOrReadOnly(BasePermission):\n def has_object_permission(\n self, request: Request, view: APIView, comparison_group: ComparisonGroup\n ) -> bool:\n return bool(\n request.method in SAFE_METHODS or request.user == comparison_group.author\n )\n\n\nclass IsComparisonCreatorOrReadOnly(BasePermission):\n def has_object_permission(\n self, request: Request, view: APIView, comparison: Comparison\n ) -> bool:\n return bool(\n request.method in SAFE_METHODS or request.user == comparison.creator\n )\n\n\nclass IsAuthorOfComparisonPassedInBodyIfExists(BasePermission):\n def has_permission(self, request: Request, view: GenericViewSet) -> bool:\n try:\n comparison_group = ComparisonGroup.objects.get(\n id=request.data[\"comparison_group_id\"]\n )\n return bool(\n request.method in SAFE_METHODS\n or request.user == comparison_group.author\n )\n except (KeyError, ComparisonGroup.DoesNotExist):\n return True\n","repo_name":"FCTL3314/HealthNutrition-Backend","sub_path":"api/v1/comparisons/permissions.py","file_name":"permissions.py","file_ext":"py","file_size_in_byte":1368,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"7720390600","text":"class Solution:\n def minMeetingRooms(self, intervals: List[List[int]]) -> int:\n hp = []\n ans = 0\n intervals.sort(key=lambda x:x[0])\n \n for s,e in intervals:\n if hp and s>=hp[0][0]:\n heappop(hp)\n heappush(hp,(e,(s,e)))\n ans = max(ans,len(hp))\n return ans","repo_name":"sudo-vaibhav/leetcode-solutions","sub_path":"0253-meeting-rooms-ii/0253-meeting-rooms-ii.py","file_name":"0253-meeting-rooms-ii.py","file_ext":"py","file_size_in_byte":346,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"12514021794","text":"import os\nfrom tkinter.tix import MAX\nimport cv2\nimport imageio\nimport numpy as np\nfrom PIL import Image\nfrom tqdm import tqdm\nfrom glob import glob\nfrom utils.argmaxMeanIOU import ArgmaxMeanIOU\nfrom tensorflow import keras, argmax\nfrom utils.dataset import CATEGORIES_COLORS\n\nIMG_SIZE = (720, 480)\nOUTPUT_SIZE = (450, 300)\n\nVIDEO_PATH = r\"F:\\ROAD_VIDEO\\CLIP\\*\"\nMODEL_PATH = r\"./trained_models/AttentionResUNet-F16_MultiDataset_384-384_epoch-60_loss-0.31_miou_0.54.h5\"\n\nSHOW_FRAMES = False\nEXPORT_GIF = True\nMAX_60SEC = True\nMAX_BATCH_SIZE = 25\n\nGIF_DURATION = 40\n\nif __name__ == \"__main__\":\n\n for model_path in glob(MODEL_PATH):\n\n segmentation_model = keras.models.load_model(model_path, custom_objects={'ArgmaxMeanIOU': ArgmaxMeanIOU})\n segmentation_model_size = segmentation_model.get_layer(index=0).input_shape[0][1:-1][::-1]\n\n # video_path = r\"F:\\ROAD_VIDEO\\CLIP\\ombre complexe + croisement de route.mp4\"\n for video_path in glob(VIDEO_PATH):\n\n video_name = os.path.basename(video_path)\n model_name = os.path.basename(model_path)\n\n video_capture = cv2.VideoCapture(video_path)\n\n frame_count = int(video_capture.get(cv2.CAP_PROP_FRAME_COUNT))\n\n image_lst = []\n\n i = 0\n\n img_for_process = []\n\n with tqdm(total=frame_count, desc=\"Video : \" + video_name) as pbar:\n\n while video_capture.isOpened():\n ret, frame = video_capture.read()\n\n i += 1\n\n if not ret or (MAX_60SEC and i > 120 * 10):\n print(\"Error while reading the video.\")\n break\n\n img_resized = cv2.resize(frame, segmentation_model_size, interpolation=cv2.INTER_AREA)\n img_resized = cv2.cvtColor(img_resized, cv2.COLOR_RGB2BGR)\n\n img_for_process.append(img_resized)\n\n if len(img_for_process) == MAX_BATCH_SIZE:\n\n img_for_process = np.array(img_for_process)\n\n result_batch = segmentation_model.predict(img_for_process / 255. )\n\n for i in range(MAX_BATCH_SIZE):\n\n img_resized = img_for_process[i]\n result_segmentation = result_batch[i]\n\n # Argmax\n result_segmentation = argmax(result_segmentation, axis=-1)\n segmentation = np.zeros(result_segmentation.shape + (3,), dtype=np.uint8)\n for categorie in CATEGORIES_COLORS.keys():\n segmentation[result_segmentation == categorie] = CATEGORIES_COLORS[categorie][\"color\"]\n\n if segmentation_model_size != OUTPUT_SIZE:\n img_resized = cv2.resize(img_resized, OUTPUT_SIZE, interpolation=cv2.INTER_AREA)\n segmentation = cv2.resize(segmentation, OUTPUT_SIZE, interpolation=cv2.INTER_AREA)\n\n overlay_segmentation = cv2.addWeighted(img_resized, 0.7, segmentation, 0.7, 0.52)\n output_image = cv2.hconcat([img_resized, segmentation, overlay_segmentation])\n\n image_lst.append(output_image)\n\n if SHOW_FRAMES:\n cv2.imshow(\"self.EVT_SEGMENTATION_IMAGE\", cv2.cvtColor(segmentation, cv2.COLOR_RGB2BGR))\n cv2.imshow(\"self.EVT_ROAD_IMAGE\", cv2.cvtColor(img_resized, cv2.COLOR_RGB2BGR))\n cv2.imshow(\"output_image\", cv2.cvtColor(output_image, cv2.COLOR_RGB2BGR))\n\n if cv2.waitKey(1) == ord('q'):\n break\n\n pbar.update(1)\n\n img_for_process = []\n\n pbar.close()\n\n if EXPORT_GIF:\n print(\"Printing GIF\")\n imageio.mimsave('./image/' + video_name + \"---\" + model_name + '.gif', image_lst, fps=40, subrectangles=True)\n print(\"GIF saved\")","repo_name":"n-rocher/RoadSegmentation","sub_path":"tools/to_gif.py","file_name":"to_gif.py","file_ext":"py","file_size_in_byte":4158,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"9049325862","text":"'''\nHelper functions\n'''\nimport numpy as np\nimport cv2\n\n\ndef sliding_window(img, window_size, step_size):\n \"\"\"slide a window across the image and yield windows data iteratively\"\"\"\n for y in xrange(0, img.shape[0], step_size[1]):\n for x in xrange(0, img.shape[1], step_size[0]):\n # yield current window\n # NOTE: this window may be smaller than expected window_size\n yield (x, y, img[y:y + window_size[1], x:x + window_size[0]])\n\n\ndef sliding_window_faster(img, window_size, step_size):\n \"\"\"silde a window across the image and\n return the standard windows tensor with xy\"\"\"\n standard_windows = []\n x_vec, y_vec = [], []\n for y in xrange(0, img.shape[0], step_size[1]):\n for x in xrange(0, img.shape[1], step_size[0]):\n window = img[y:y + window_size[1], x:x + window_size[0]]\n if window.shape == (window_size[1], window_size[0]):\n x_vec.append(x)\n y_vec.append(y)\n standard_windows.append(window)\n\n return x_vec, y_vec, standard_windows\n\n\ndef pyramid(img, downscale=1.5, min_size=(30, 30)):\n \"\"\"compute image pyramid through down sampling\"\"\"\n # min_size: (w, h)\n yield img\n\n while True:\n h = int(img.shape[0] / downscale)\n w = int(img.shape[1] / downscale)\n img = cv2.resize(img, (w, h))\n\n if img.shape[0] < min_size[1] or img.shape[1] < min_size[0]:\n break\n\n yield img\n\n\ndef draw_detections(img, detections):\n for det in detections:\n cv2.rectangle(img, (det[0], det[1]), (det[2], det[3]),\n color=(0, 0, 0), thickness=2)\n\n cv2.imshow('detection', img)\n cv2.waitKey(0)\n\n\ndef find_biggest_window(detections):\n \"\"\"find the biggest bounding window from detections matrix\"\"\"\n x1 = detections[:, 0]\n y1 = detections[:, 1]\n x2 = detections[:, 2]\n y2 = detections[:, 3]\n area = (x2 - x1 + 1) * (y2 - y1 + 1)\n idx = np.argmax(area)\n\n return detections[idx]\n\n\ndef get_overlap(box1, box2):\n \"\"\"compute the overlap of box1 and box2 comparing to box1 area\"\"\"\n area = (box1[2] - box1[0] + 1.0) * (box1[3] - box1[1] + 1.0)\n xx1 = max(box1[0], box2[0])\n yy1 = max(box1[1], box2[1])\n xx2 = min(box1[2], box2[2])\n yy2 = min(box1[3], box2[3])\n w = max(0, xx2 - xx1)\n h = max(0, yy2 - yy1)\n\n return w * h / area\n","repo_name":"xueeinstein/it-shoes","sub_path":"helper.py","file_name":"helper.py","file_ext":"py","file_size_in_byte":2374,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"27538282334","text":"SCOPETYPE = 'OPENADC'\nPLATFORM = 'CWLITEXMEGA'\nCRYPTO_TARGET='TINYAES128C'\nSS_VER='SS_VER_1_1'\n\nexec(open(\"basic_setup.py\").read())\n\ncw.program_target(scope, prog, \"../hardware/victims/firmware/simpleserial-aes/simpleserial-aes-CWLITEXMEGA.hex\".format(PLATFORM))\n\n\nimport numpy as np\nimport time\n\nktp = cw.ktp.Basic()\ntrace_array = []\ntextin_array = []\n\nkey, text = ktp.next()\n\ntarget.set_key(key)\n\nN = 100\nfor i in range(N):\n scope.arm()\n if text[0] & 0x01:\n text[0] = 0x0F\n else:\n text[0] = 0x00\n target.simpleserial_write('p', text)\n \n ret = scope.capture()\n if ret:\n print(\"Target timed out!\")\n continue\n \n response = target.simpleserial_read('r', 16)\n \n trace_array.append(scope.get_last_trace())\n textin_array.append(text)\n \n key, text = ktp.next()\n\n\n#MAIN\n\nassert len(trace_array) == 100\nprint(\"✔️ OK to continue!\")\n\none_list = []\nzero_list = []\nfor i in range(len(trace_array)):\n if textin_array[i][0] == 0x00:\n one_list.append(trace_array[i])\n else:\n zero_list.append(trace_array[i])\n\none_avg = np.mean(one_list, axis = 0)\nzero_avg = np.mean(zero_list, axis = 0)\n\ndiff_avg = one_avg - zero_avg\n\nimport matplotlib.pyplot as plt\nplt.plot(diff_avg)\nplt.show()","repo_name":"VladTalmaciu/CW_tests","sub_path":"scripts/lab_3_1.py","file_name":"lab_3_1.py","file_ext":"py","file_size_in_byte":1259,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"25948586580","text":"from operator import attrgetter\n\nfrom flask import abort, request\n\nfrom ...services.brand import service as brand_service\nfrom ...services.orga import service as orga_service\nfrom ...services.orga import birthday_service as orga_birthday_service\nfrom ...services.user import service as user_service\nfrom ...util.export import serialize_to_csv\nfrom ...util.framework.blueprint import create_blueprint\nfrom ...util.framework.flash import flash_success\nfrom ...util.templating import templated\nfrom ...util.views import redirect_to, respond_no_content, textified\n\nfrom ..authorization.decorators import permission_required\nfrom ..authorization.registry import permission_registry\nfrom ..orga_team_admin.authorization import OrgaTeamPermission\n\nfrom .authorization import OrgaBirthdayPermission, OrgaDetailPermission\nfrom .forms import OrgaFlagCreateForm\n\n\nblueprint = create_blueprint('orga_admin', __name__)\n\n\npermission_registry.register_enum(OrgaBirthdayPermission)\npermission_registry.register_enum(OrgaDetailPermission)\npermission_registry.register_enum(OrgaTeamPermission)\n\n\n@blueprint.route('/persons')\n@permission_required(OrgaDetailPermission.view)\n@templated\ndef persons():\n \"\"\"List brands to choose from.\"\"\"\n brands_with_person_counts = orga_service.get_brands_with_person_counts()\n\n return {\n 'brands_with_person_counts': brands_with_person_counts,\n }\n\n\n@blueprint.route('/persons/')\n@permission_required(OrgaDetailPermission.view)\n@templated\ndef persons_for_brand(brand_id):\n \"\"\"List organizers for the brand with details.\"\"\"\n brand = _get_brand_or_404(brand_id)\n\n orgas = orga_service.get_orgas_for_brand(brand.id)\n\n return {\n 'brand': brand,\n 'orgas': orgas,\n }\n\n\n@blueprint.route('/persons//create')\n@permission_required(OrgaTeamPermission.administrate_memberships)\n@templated\ndef create_orgaflag_form(brand_id):\n \"\"\"Show form to give the organizer flag to a user.\"\"\"\n brand = _get_brand_or_404(brand_id)\n\n form = OrgaFlagCreateForm()\n\n return {\n 'brand': brand,\n 'form': form,\n }\n\n\n@blueprint.route('/persons/', methods=['POST'])\n@permission_required(OrgaTeamPermission.administrate_memberships)\ndef create_orgaflag(brand_id):\n \"\"\"Give the organizer flag to a user.\"\"\"\n brand = _get_brand_or_404(brand_id)\n\n form = OrgaFlagCreateForm(request.form)\n\n user_id = form.user_id.data.strip()\n user = _get_user_or_404(user_id)\n\n orga_flag = orga_service.create_orga_flag(brand.id, user.id)\n\n flash_success('{} wurde das Orga-Flag für die Marke {} gegeben.',\n orga_flag.user.screen_name, orga_flag.brand.title)\n return redirect_to('.persons_for_brand', brand_id=orga_flag.brand.id)\n\n\n@blueprint.route('/persons//', methods=['DELETE'])\n@permission_required(OrgaTeamPermission.administrate_memberships)\n@respond_no_content\ndef remove_orgaflag(brand_id, user_id):\n \"\"\"Remove the organizer flag for a brand from a person.\"\"\"\n orga_flag = orga_service.find_orga_flag(brand_id, user_id)\n\n if orga_flag is None:\n abort(404)\n\n brand = orga_flag.brand\n user = orga_flag.user\n\n orga_service.delete_orga_flag(orga_flag)\n\n flash_success('{} wurde das Orga-Flag für die Marke {} entzogen.',\n user.screen_name, brand.title)\n\n\n@blueprint.route('/persons//export')\n@permission_required(OrgaDetailPermission.view)\n@textified\ndef export_persons(brand_id):\n \"\"\"Export the list of organizers for the brand as a CSV document in\n Microsoft Excel dialect.\n \"\"\"\n brand = _get_brand_or_404(brand_id)\n\n field_names = [\n 'Benutzername',\n 'Vorname',\n 'Nachname',\n 'Geburtstag',\n 'Straße',\n 'PLZ',\n 'Ort',\n 'Land',\n 'E-Mail-Adresse',\n 'Telefonnummer',\n ]\n\n def to_dict(user):\n date_of_birth = user.detail.date_of_birth.strftime('%d.%m.%Y') \\\n if user.detail.date_of_birth else None\n\n return {\n 'Benutzername': user.screen_name,\n 'Vorname': user.detail.first_names,\n 'Nachname': user.detail.last_name,\n 'Geburtstag': date_of_birth,\n 'Straße': user.detail.street,\n 'PLZ': user.detail.zip_code,\n 'Ort': user.detail.city,\n 'Land': user.detail.country,\n 'E-Mail-Adresse': user.email_address,\n 'Telefonnummer': user.detail.phone_number,\n }\n\n orgas = orga_service.get_orgas_for_brand(brand.id)\n orgas.sort(key=attrgetter('screen_name'))\n rows = map(to_dict, orgas)\n return serialize_to_csv(field_names, rows)\n\n\n@blueprint.route('/birthdays')\n@permission_required(OrgaBirthdayPermission.list)\n@templated\ndef birthdays():\n orgas = list(\n orga_birthday_service.collect_orgas_with_next_birthdays(limit=5))\n\n return {\n 'orgas': orgas,\n }\n\n\ndef _get_brand_or_404(brand_id):\n brand = brand_service.find_brand(brand_id)\n\n if brand is None:\n abort(404)\n\n return brand\n\n\ndef _get_user_or_404(user_id):\n user = user_service.find_user(user_id)\n\n if user is None:\n abort(404)\n\n return user\n","repo_name":"agreements/byceps","sub_path":"byceps/blueprints/orga_admin/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"7"} +{"seq_id":"38874396888","text":"import sqlalchemy\n\n\n\ndef create_DB(connection):\n connection.execute(\n \"\"\"\n create table if not exists Execut (\n Idexec serial primary key,\n Name varchar(40) not null\n );\n\n create table if not exists Albums (\n Idalbum serial primary key,\n Name varchar(40) not null,\n Year integer not null\n );\n\n create table if not exists Song (\n Idsong serial primary key,\n Name varchar(40) not null,\n Length integer not null,\n Idalbum integer not null,\n FOREIGN KEY (Idalbum) REFERENCES Albums (Idalbum)\n );\n\n create table if not exists Collection (\n Idcollection serial primary key,\n Name varchar(40) not null,\n Year integer not null\n );\n\n create table if not exists Styles (\n Idstyle serial primary key,\n Name varchar(40) not null\n );\n \"\"\")\n\ndef create_link(connection):\n connection.execute(\n \"\"\"\n create table if not exists CollectionSong (\n Idcollection integer references Collection(Idcollection),\n Idsong integer references Song(Idsong),\n constraint CollectionSong_pk primary key (Idcollection, Idsong)\n );\n\n create table if not exists AlbumsExec (\n Idalbum integer references Albums(Idalbum),\n Idexec integer references Execut(Idexec),\n constraint AlbumsExec_pk primary key (Idalbum, Idexec)\n );\n\n create table if not exists StylesExec (\n Idstyle integer references Styles(Idstyle),\n Idexec integer references Execut(Idexec),\n constraint StylesExec_pk primary key (Idstyle, Idexec)\n );\n \"\"\")\n\ndef add_execut_DB(connection):\n table = 'Execut'\n column = 'Name'\n executors = ['Исполнитель один',\n 'Второй',\n 'Третий',\n 'Четвертый',\n 'Пятый исполнитель',\n 'Шестой',\n 'Седьмой',\n 'Восьмой'] # xD\n for executor in executors:\n value = f\"\"\"insert into {table}({column}) values('{executor}');\"\"\"\n connection.execute(value)\n\ndef add_albums_DB(connection):\n table = 'albums'\n column1 = 'name'\n column2 = 'year'\n albums = [['One',2020],\n ['Two',2021],\n ['Three',2019],\n ['Four',1999],\n ['Five',2000],\n ['Six',2018],\n ['Seven',1998],\n ['Eigth',2012]]\n for album in albums:\n value = f\"\"\"insert into {table}({column1},{column2}) values('{album[0]}',{album[1]});\"\"\"\n connection.execute(value)\n\ndef add_styles_DB(connection):\n table = 'styles'\n column = 'name'\n styles = ['One',\n 'Two',\n 'Three',\n 'Four',\n 'Five',\n 'Six',\n 'Seven',\n 'Eigth']\n for style in styles:\n value = f\"\"\"insert into {table}({column}) values('{style}');\"\"\"\n connection.execute(value)\n\ndef add_collections_DB(connection):\n table = 'collection'\n column1 = 'name'\n column2 = 'year'\n collections = [['Сборник один',2020],\n ['Сборник два',2021],\n ['Сборник три',2019],\n ['Сборник четыре',2017],\n ['Сборник пять',2016],\n ['Сборник шесть',1976],\n ['Сборник семь',2018],\n ['Сборник восемь',2012]]\n for collection in collections:\n value = f\"\"\"insert into {table}({column1},{column2}) values('{collection[0]}',{collection[1]});\"\"\"\n connection.execute(value)\n\ndef add_songs_DB(connection):\n table = 'song'\n column1 = 'name'\n column2 = 'length'\n column3 = 'idalbum'\n songs = [['Трек 1',125,2],\n ['Трек 2',102,5],\n ['Трек 3',402,5],\n ['Трек 4',356,5],\n ['Трек my',296,2],\n ['Трек 6',159,1],\n ['Трек 7',180,1],\n ['Трек 8',195,3],\n ['Трек 9',235,4],\n ['Трек 10',160,6],\n ['Трек 11',162,7],\n ['Трек мой',188,8],\n ['Трек 13',210,3],\n ['Мой трек',295,8],\n ['Трек 13',212,6],\n ['Трек 14',265,5],\n ['Трек 15',333,3]]\n for song in songs:\n value = f\"\"\"insert into {table}({column1},{column2},{column3}) values('{song[0]}',{song[1]},{song[2]});\"\"\"\n connection.execute(value)\n\ndef add_albumexec_DB(connection):\n table = 'albumsexec'\n column1 = 'idalbum'\n column2 = 'idexec'\n albumexecs = [[1,2],\n [2,5],\n [3,2],\n [6,1],\n [8,1],\n [8,3],\n [7,4],\n [6,5],\n [3,7],\n [2,3],\n [1,3],\n [2,8],\n [4,6],\n [5,3],\n [4,3]]\n for albumexec in albumexecs:\n value = f\"\"\"insert into {table}({column1},{column2}) values({albumexec[0]},{albumexec[1]});\"\"\"\n connection.execute(value)\n\ndef add_collectionsong_DB(connection):\n table = 'collectionsong'\n column1 = 'idcollection'\n column2 = 'idsong'\n collectionsongs = [[1,1],\n [2,2],\n [3,3],\n [4,4],\n [5,5],\n [6,6],\n [7,7],\n [8,8],\n [1,9],\n [2,10],\n [3,11],\n [4,12],\n [5,13],\n [6,14],\n [7,15]]\n for collectionsong in collectionsongs:\n value = f\"\"\"insert into {table}({column1},{column2}) values({collectionsong[0]},{collectionsong[1]});\"\"\"\n connection.execute(value)\n\n\ndef add_stylesexec_DB(connection):\n table = 'stylesexec'\n column1 = 'idstyle'\n column2 = 'idexec'\n stylesexecs = [[1,1],\n [2,2],\n [3,3],\n [4,4],\n [5,5],\n [1,6],\n [2,7],\n [3,8],\n [4,7],\n [5,6],\n [1,5],\n [2,4],\n [3,5],\n [4,2],\n [5,1]]\n for stylesexec in stylesexecs:\n value = f\"\"\"insert into {table}({column1},{column2}) values({stylesexec[0]},{stylesexec[1]});\"\"\"\n connection.execute(value)\n\n\n\ndef select_albums_2018_DB(connection):\n table = 'albums'\n column1 = 'name'\n column2 = 'year'\n value = f\"\"\"select {column1},{column2} from {table} where {column2}=2018;\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nназвание и год выхода альбомов, вышедших в 2018 году')\n print(data)\n\ndef select_long_song_DB(connection):\n table = 'song'\n column1 = 'name'\n column2 = 'length'\n value = f\"\"\"select {column1},{column2} from {table} ORDER BY {column2} DESC LIMIT 1;\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nназвание и продолжительность самого длительного трека')\n print(data)\n\ndef select_song_3_5_DB(connection):\n table = 'song'\n column1 = 'name'\n column2 = 'length'\n min_length = int(3.5*60)\n value = f\"\"\"select {column1} from {table} where {column2} >= {min_length};\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nназвание треков, продолжительность которых не менее 3,5 минуты')\n print(data)\n\ndef select_collections_2018_2020_DB(connection):\n table = 'collection'\n column1 = 'name'\n column2 = 'year'\n value = f\"\"\"select {column1},{column2} from {table} where {column2} BETWEEN 2018 AND 2020;\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nназвания сборников, вышедших в период с 2018 по 2020 год включительно')\n print(data)\n\ndef select_exec_1_slovo_DB(connection):\n table = 'execut'\n column = 'name'\n value = f\"\"\"select {column} from {table} where {column} not like '%% %%';\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nисполнители, чье имя состоит из 1 слова')\n print(data)\n\ndef select_song_my_DB(connection):\n table = 'song'\n column = 'name'\n value = f\"\"\"select {column} from {table} where lower ({column}) like lower ('%%my%%') OR lower ({column}) like lower ('%%мой%%');\"\"\"\n data = connection.execute(value).fetchall()\n print('\\nназвание треко��, которые содержат слово \"мой\"/\"my\"')\n print(data)\n\n\n\ndef create_tables(connection):\n '''Блок создания таблиц'''\n create_DB(connection) # создание таблиц\n create_link(connection) # создание таблиц-связей\n\ndef add_data_in_tables(connection):\n '''Блок добавления данных\n SQL 1\n Заполните базу данных из предыдущего домашнего задания. В ней должно быть:\n\n не менее 8 исполнителей;\n не менее 5 жанров;\n не менее 8 альбомов;\n не менее 15 треков;\n не менее 8 сборников.\n Внимание! Должны быть заполнены все поля каждой таблицы,\n в т.ч. таблицы связей\n (исполнителей с жанрами, исполнителей с альбомами, сборников с треками).\n '''\n add_execut_DB(connection)\n add_albums_DB(connection)\n add_styles_DB(connection)\n add_collections_DB(connection)\n add_songs_DB(connection)\n add_albumexec_DB(connection)\n add_collectionsong_DB(connection)\n add_stylesexec_DB(connection)\n\ndef select_tables(connection):\n '''Блок SELECT запросов\n SQL 2\n Написать SELECT-запросы, которые выведут информацию согласно инструкциям ниже.\n Внимание! Результаты запросов не должны быть пустыми (учтите при заполнении таблиц).\n\n название и год выхода альбомов, вышедших в 2018 году;\n название и продолжительность самого длительного трека;\n название треков, продолжительность которых не менее 3,5 минуты;\n названия сборников, вышедших в период с 2018 по 2020 год включительно;\n исполнители, чье имя состоит из 1 слова;\n название треков, которые содержат слово \"мой\"/\"my\".\n Результатом работы будет 3 файла (с INSERT, SELECT запросами и CREATE запросами из предыдущего задания)\n в формате .sql (или .py/.ipynb, если вы будете писать запросы с использованием SQLAlchemy).\n '''\n select_albums_2018_DB(connection)\n select_long_song_DB(connection)\n select_song_3_5_DB(connection)\n select_collections_2018_2020_DB(connection)\n select_exec_1_slovo_DB(connection)\n select_song_my_DB(connection)\n\nengine = sqlalchemy.create_engine('postgresql://postgres:pass@localhost:5432/postgres')\nconnection = engine.connect()\nnum = ''\nwhile num != '4':\n num = input('Введите одно из нижеперечисленных чисел: \\n1. для создания БД\\n2. для добавления данных в таблицы\\n3. для производства SELECT запросов\\n4. для выхода из программы\\n')\n if num == '1':\n create_tables(connection)\n print('Таблицы созданы')\n elif num == '2':\n try:\n add_data_in_tables(connection)\n except:\n print('Возможно данные были внесены ранее')\n print('Данные внесены в таблицы')\n elif num == '3':\n select_tables(connection)\n print('\\nSELECT запросы выполнены')\n elif num == '4':\n print('ПРОГРАММА ЗАВЕРШИЛА СВОЮ РАБОТУ')\n else:\n print('Вами введено некорректный символ. Введите одно из нижеперечисленных чисел: \\n1. для создания БД\\n2. для добавления данных в таблицы\\n3. для производства SELECT запросово')\n","repo_name":"pavzzz2909/Homeworks-Netology","sub_path":"3 Базы данных для Python разработчиков/3 Select-запросы, выборки из одной таблицы/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":12965,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"39826635427","text":"from rest_framework import serializers\nfrom ..models.location import Location\nfrom .address_serializer import AddressSerializer\n\n\nclass LocationSerializer(serializers.ModelSerializer):\n \"\"\"\n Serializer for Location model.\n \"\"\"\n address = AddressSerializer()\n\n class Meta:\n model = Location\n fields = '__all__'\n read_only_fields = ['id']\n\n def create(self, validated_data):\n \"\"\"\n Create and return a new Location instance.\n \"\"\"\n address_data = validated_data.pop('address')\n address_serializer = AddressSerializer(data=address_data)\n address_serializer.is_valid(raise_exception=True)\n address = address_serializer.save()\n return Location.objects.create(address=address, **validated_data)\n\n def update(self, instance, validated_data):\n \"\"\"\n Update and return an existing Location instance.\n \"\"\"\n address_data = validated_data.pop('address', None)\n if address_data:\n address_serializer = AddressSerializer(instance.address, data=address_data)\n address_serializer.is_valid(raise_exception=True)\n address = address_serializer.save()\n validated_data['address'] = address\n instance.location_name = validated_data.get('location_name', instance.location_name)\n instance.description = validated_data.get('description', instance.description)\n instance.photo = validated_data.get('photo', instance.photo)\n instance.save()\n return instance\n","repo_name":"azaz0/search-trip","sub_path":"trip_management/serializers/location_serializer.py","file_name":"location_serializer.py","file_ext":"py","file_size_in_byte":1539,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"1087046964","text":"import time\nfrom contextlib import contextmanager\n\nimport numpy as np\n\nfrom imitation.helpers.misc_util import zipsame, prettify_time, var_shape, numel\n\n\ndef cell(x, width):\n \"\"\"Format a tabular cell to the specified width\"\"\"\n if isinstance(x, np.ndarray):\n assert x.ndim == 0\n x = x.item()\n rep = \"{:G}\".format(x) if isinstance(x, float) else str(x)\n return rep + (' ' * (width - len(rep)))\n\n\ndef columnize(names, tuples, widths, indent=2):\n \"\"\"Generate and return the content of table\n (w/o logging or printing anything)\n\n Args:\n width (int): Width of each cell in the table\n indent (int): Indentation spacing prepended to every row in the table\n \"\"\"\n indent_space = indent * ' '\n # Add row containing the names\n table = indent_space + \" | \".join(cell(name, width) for name, width in zipsame(names, widths))\n table_width = len(table)\n # Add header hline\n table += '\\n' + indent_space + ('-' * table_width)\n for tuple_ in tuples:\n # Add a new row\n table += '\\n' + indent_space\n table += \" | \".join(cell(value, width) for value, width in zipsame(tuple_, widths))\n # Add closing hline\n table += '\\n' + indent_space + ('-' * table_width)\n return table\n\n\ndef colorize(string, color, bold=False, highlight=False):\n color2num = {'gray': 30, 'red': 31, 'green': 32, 'yellow': 33, 'blue': 34,\n 'magenta': 35, 'cyan': 36, 'white': 37, 'crimson': 38}\n attr = []\n num = color2num[color]\n if highlight:\n num += 10\n attr.append(str(num))\n if bold:\n attr.append('1')\n return '\\x1b[%sm%s\\x1b[0m' % (';'.join(attr), string)\n\n\ndef log_module_info(logger, name, *components):\n assert len(components) > 0, \"components list is empty\"\n for component in components:\n logger.info(\"logging {}/{} specs\".format(name, component.name))\n names = [var.name for var in component.trainable_vars]\n shapes = [var_shape(var) for var in component.trainable_vars]\n num_paramss = [numel(var) for var in component.trainable_vars]\n zipped_info = zipsame(names, shapes, num_paramss)\n logger.info(columnize(names=['name', 'shape', 'num_params'],\n tuples=zipped_info,\n widths=[40, 16, 10]))\n logger.info(\" total num params: {}\".format(sum(num_paramss)))\n\n\ndef timed_cm_wrapper(comm=None, logger=None, color_message='magenta', color_elapsed_time='cyan'):\n \"\"\"Wraps a context manager that records the time taken by encapsulated ops\"\"\"\n @contextmanager\n def _timed(message):\n \"\"\"Display the time it took for the mpi master\n to perform the task within the context manager\n \"\"\"\n if comm is None or comm.Get_rank() == 0:\n logger.info(colorize(message, color=color_message))\n tstart = time.time()\n yield\n logger.info(colorize(\" [done in {:.3f} seconds]\".format(time.time() - tstart),\n color=color_elapsed_time))\n else:\n yield\n return _timed\n\n\ndef pretty_iter(logger, i):\n \"\"\"Display the current iteration with a colored decorator\"\"\"\n logger.info(colorize(\"I T E R A T I O N {}\".format(i), color='blue'))\n\n\ndef pretty_elapsed(logger, tstart):\n \"\"\"Display the elapsed time with a colored decorator\"\"\"\n elapsed = prettify_time(time.time() - tstart)\n # logger.info('')\n logger.info(colorize(\"E L A P S E D {}\".format(elapsed), color='green'))\n","repo_name":"lionelblonde/sam-tf","sub_path":"imitation/helpers/console_util.py","file_name":"console_util.py","file_ext":"py","file_size_in_byte":3514,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"7"} +{"seq_id":"4905289454","text":"__author__ = 'Joao'\nfrom SolTw import _Utils as _Utils\n\n\nclass Hashtags:\n def __init__(self, dictionary=dict()):\n dictionary= _Utils.CastToDictionary(dictionary)\n dictionary= _Utils.removeEmptyFields(dictionary)\n self.indices=\"\"\n self.text=\"\"\n if (\"indices\" in dictionary):\n self.indices=dictionary[\"indices\"]\n if (\"text\" in dictionary):\n self.text=dictionary[\"text\"]\n\n\n def __str__(self):\n dic=self.__dict__\n lista=list()\n for key in dic:\n lista.append(key)\n for key in lista:\n if dic[key]==None or dic[key]==\"\":\n del dic[key]\n return \"HASHTAGS: \"+str(dic)\n\n def __repr__(self):\n dic=self.__dict__\n lista=list()\n for key in dic:\n lista.append(key)\n for key in lista:\n if dic[key]==None or dic[key]==\"\":\n del dic[key]\n return \"HASHTAGS: \"+str(dic)","repo_name":"JoaoMachadoJr/SOL","sub_path":"Projeto/SolTw/_Hashtags.py","file_name":"_Hashtags.py","file_ext":"py","file_size_in_byte":991,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"35415368981","text":"# -*- coding: utf-8 -*-\n'''\nCreated on 27/05/2014\n\n@author: aferreira\n'''\n\nimport wx\nfrom py.una.pol.tava.presenter.ptoolbar import ToolBarPresenter\nfrom wx import GetTranslation as _\nimport py.una.pol.tava.view.vi18n as C\nimport py.una.pol.tava.view.vimages as I\n\n\nclass MainToolBar(wx.ToolBar):\n '''\n Clase que representa al ToolBar Principal desplegando algunas opciones\n de trabajo que pueden utilizarse dentro del área de trabajo.\n '''\n\n def __init__(self, parent):\n '''\n Constructor para la clase MainToolBar\n :param parent: referencia a la clase padre de MainToolBar.\n '''\n super(MainToolBar, self).__init__(parent, wx.TB_HORIZONTAL)\n\n # Creacion del Presenter\n self.presenter = ToolBarPresenter(self)\n\n # Creacion de los ids de los diferentes tools\n self.SetIdReferences()\n\n self.SetToolBitmapSize((16, 16))\n\n # Nuevo Proyecto\n new_bmp = wx.ArtProvider.GetBitmap(wx.ART_NEW)\n self.AddLabelTool(self.ID_NEW_PRO, '', new_bmp)\n self.EnableTool(self.ID_NEW_PRO, True)\n self.Bind(wx.EVT_TOOL, self.OnProjectNew, id=self.ID_NEW_PRO)\n\n # Abrir Proyecto\n open_bmp = wx.ArtProvider.GetBitmap(wx.ART_FILE_OPEN)\n self.AddLabelTool(self.ID_OPEN_PRO, '', open_bmp)\n self.EnableTool(self.ID_OPEN_PRO, False)\n self.Bind(wx.EVT_TOOL, self.OnProjectOpen, id=self.ID_OPEN_PRO)\n\n # Cerrar Proyecto\n self.AddLabelTool(self.ID_CLOSE_PRO, '', I.close_png)\n self.EnableTool(self.ID_CLOSE_PRO, False)\n self.Bind(wx.EVT_TOOL, self.OnProjectClose, id=self.ID_CLOSE_PRO)\n\n # Eliminar Proyecto\n self.AddLabelTool(self.ID_DEL_PRO, '', I.delete_png)\n self.EnableTool(self.ID_DEL_PRO, False)\n self.Bind(wx.EVT_TOOL, self.OnProjectDelete, id=self.ID_DEL_PRO)\n\n # Blog de Proyecto\n self.AddLabelTool(self.ID_BLOG_PRO, '', I.blog_png)\n self.EnableTool(self.ID_BLOG_PRO, False)\n\n self.AddSeparator()\n\n # Desocultar Proyecto\n self.AddLabelTool(self.ID_UNHIDE_PRO, '', I.hide_left_png)\n self.EnableTool(self.ID_UNHIDE_PRO, True)\n self.Bind(wx.EVT_TOOL, self.OnProjectUnHide, id=self.ID_UNHIDE_PRO)\n\n # Ocultar Proyecto\n self.AddLabelTool(self.ID_HIDE_PRO, '', I.hide_right_png)\n self.EnableTool(self.ID_HIDE_PRO, False)\n self.Bind(wx.EVT_TOOL, self.OnProjectHide, id=self.ID_HIDE_PRO)\n\n self.AddSeparator()\n\n # Salir Aplicacion\n exit_bmp = wx.ArtProvider.GetBitmap(wx.ART_QUIT)\n self.AddLabelTool(self.ID_EXIT_PRO, '', exit_bmp)\n self.Bind(wx.EVT_TOOL, parent.OnApplicationExit, id=self.ID_EXIT_PRO)\n\n # Establecemos los labels\n self.SetLabels()\n\n # Finalizando la creacion del toolbar\n self.Realize()\n\n def SetIdReferences(self):\n self.ID_NEW_PRO = wx.NewId()\n self.ID_OPEN_PRO = wx.NewId()\n self.ID_CLOSE_PRO = wx.NewId()\n self.ID_DEL_PRO = wx.NewId()\n self.ID_BLOG_PRO = wx.NewId()\n self.ID_EXIT_PRO = wx.NewId()\n self.ID_HIDE_PRO = wx.NewId()\n self.ID_UNHIDE_PRO = wx.NewId()\n\n def EnableDisableOpenProject(self):\n self.EnableTool(self.ID_DEL_PRO, True)\n self.EnableTool(self.ID_CLOSE_PRO, True)\n self.EnableTool(self.ID_BLOG_PRO, True)\n self.EnableTool(self.ID_OPEN_PRO, False)\n self.EnableTool(self.ID_HIDE_PRO, False)\n\n def EnableDisableCloseProject(self):\n self.EnableTool(self.ID_DEL_PRO, True)\n self.EnableTool(self.ID_OPEN_PRO, True)\n self.EnableTool(self.ID_CLOSE_PRO, False)\n self.EnableTool(self.ID_BLOG_PRO, False)\n self.EnableTool(self.ID_HIDE_PRO, True)\n\n def DisableAllProject(self):\n self.EnableTool(self.ID_DEL_PRO, False)\n self.EnableTool(self.ID_CLOSE_PRO, False)\n self.EnableTool(self.ID_BLOG_PRO, False)\n self.EnableTool(self.ID_OPEN_PRO, False)\n self.EnableTool(self.ID_HIDE_PRO, False)\n\n def OnProjectClose(self, event):\n self.presenter.CloseProject()\n\n def OnProjectOpen(self, event):\n self.presenter.OpenProject()\n\n def OnProjectDelete(self, event):\n self.presenter.DeleteProject()\n\n def OnProjectNew(self, event):\n self.presenter.NewProject()\n\n def OnProjectHide(self, event):\n self.presenter.HideProject()\n\n def OnProjectUnHide(self, event):\n self.presenter.UnHideProject()\n\n def SetLabels(self):\n self.SetToolShortHelp(self.ID_NEW_PRO, _(C.MTB_NP))\n self.SetToolShortHelp(self.ID_OPEN_PRO, _(C.MTB_OP))\n self.SetToolShortHelp(self.ID_CLOSE_PRO, _(C.MTB_CP))\n self.SetToolShortHelp(self.ID_DEL_PRO, _(C.MTB_DP))\n self.SetToolShortHelp(self.ID_BLOG_PRO, _(C.MTB_BP))\n self.SetToolShortHelp(self.ID_EXIT_PRO, _(C.MTB_EX))\n self.SetToolShortHelp(self.ID_HIDE_PRO, _(C.MTB_HP))\n self.SetToolShortHelp(self.ID_UNHIDE_PRO, _(C.MTB_UHP))\n","repo_name":"abrahanfretes/tava_antes","sub_path":"py/una/pol/tava/view/vtoolbar.py","file_name":"vtoolbar.py","file_ext":"py","file_size_in_byte":4983,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"33213138704","text":"\"\"\"\n2.5D DC inversion of with Topography\n====================================\n\nThis is an example for 2.5D DC inversion. Earth includes a topography,\nand below the topography conductive and resistive cylinders are embedded.\nSensitivity weighting is used for the inversion.\nApproximate depth of investigation is computed by selecting\n1 percent of max(sqrt(diag(JtJ))), and regions having smaller sensitivity\nthan this is blanked.\nUser is promoted to try different suvey_type such as 'pole-dipole',\n'dipole-pole', and 'pole-pole'.\n\"\"\"\n\nfrom SimPEG.electromagnetics.static import resistivity as DC\nfrom SimPEG.electromagnetics.static.utils import generate_dcip_survey, genTopography\nfrom SimPEG import (\n maps,\n utils,\n data_misfit,\n regularization,\n optimization,\n inversion,\n inverse_problem,\n directives,\n)\nfrom discretize.utils import active_from_xyz\nimport matplotlib.pyplot as plt\nfrom matplotlib import colors\nimport numpy as np\nfrom pylab import hist\n\ntry:\n from pymatsolver import Pardiso as Solver\nexcept ImportError:\n from SimPEG import SolverLU as Solver\n\n\ndef run(plotIt=True, survey_type=\"dipole-dipole\"):\n np.random.seed(1)\n # Initiate I/O class for DC\n IO = DC.IO()\n # Obtain ABMN locations\n\n xmin, xmax = 0.0, 200.0\n ymin, ymax = 0.0, 0.0\n zmin, zmax = 0, 0\n endl = np.array([[xmin, ymin, zmin], [xmax, ymax, zmax]])\n # Generate DC survey object\n survey = generate_dcip_survey(\n endl, survey_type=survey_type, dim=2, a=10, b=10, n=10\n )\n survey = IO.from_abmn_locations_to_survey(\n survey.locations_a,\n survey.locations_b,\n survey.locations_m,\n survey.locations_n,\n survey_type,\n data_dc_type=\"volt\",\n )\n\n # Obtain 2D TensorMesh\n mesh, actind = IO.set_mesh()\n topo, mesh1D = genTopography(mesh, -10, 0, its=100)\n actind = active_from_xyz(mesh, np.c_[mesh1D.cell_centers_x, topo])\n survey.drape_electrodes_on_topography(mesh, actind, option=\"top\")\n\n # Build a conductivity model\n blk_inds_c = utils.model_builder.getIndicesSphere(\n np.r_[60.0, -25.0], 12.5, mesh.gridCC\n )\n blk_inds_r = utils.model_builder.getIndicesSphere(\n np.r_[140.0, -25.0], 12.5, mesh.gridCC\n )\n sigma = np.ones(mesh.nC) * 1.0 / 100.0\n sigma[blk_inds_c] = 1.0 / 10.0\n sigma[blk_inds_r] = 1.0 / 1000.0\n sigma[~actind] = 1.0 / 1e8\n rho = 1.0 / sigma\n\n # Show the true conductivity model\n if plotIt:\n fig = plt.figure(figsize=(12, 3))\n ax = plt.subplot(111)\n temp = rho.copy()\n temp[~actind] = np.nan\n out = mesh.plot_image(\n temp,\n grid=True,\n ax=ax,\n grid_opts={\"alpha\": 0.2},\n clim=(10, 1000),\n pcolor_opts={\"cmap\": \"viridis\", \"norm\": colors.LogNorm()},\n )\n ax.plot(\n survey.electrode_locations[:, 0], survey.electrode_locations[:, 1], \"k.\"\n )\n ax.set_xlim(IO.grids[:, 0].min(), IO.grids[:, 0].max())\n ax.set_ylim(-IO.grids[:, 1].max(), IO.grids[:, 1].min())\n cb = plt.colorbar(out[0])\n cb.set_label(\"Resistivity (ohm-m)\")\n ax.set_aspect(\"equal\")\n plt.show()\n\n # Use Exponential Map: m = log(rho)\n actmap = maps.InjectActiveCells(mesh, indActive=actind, valInactive=np.log(1e8))\n mapping = maps.ExpMap(mesh) * actmap\n\n # Generate mtrue\n mtrue = np.log(rho[actind])\n\n # Generate 2.5D DC problem\n # \"N\" means potential is defined at nodes\n prb = DC.Simulation2DNodal(\n mesh, survey=survey, rhoMap=mapping, storeJ=True, Solver=Solver, verbose=True\n )\n\n # Make synthetic DC data with 5% Gaussian noise\n data = prb.make_synthetic_data(mtrue, relative_error=0.05, add_noise=True)\n\n IO.data_dc = data.dobs\n # Show apparent resisitivty pseudo-section\n if plotIt:\n IO.plotPseudoSection(data=data.dobs, data_type=\"apparent_resistivity\")\n\n # Show apparent resisitivty histogram\n if plotIt:\n fig = plt.figure()\n out = hist(data.dobs, bins=20)\n plt.xlabel(r\"Apparent Resisitivty ($\\Omega$m)\")\n plt.show()\n\n # Set initial model based upon histogram\n m0 = np.ones(actmap.nP) * np.log(100.0)\n\n # Set standard_deviation\n # floor (10 ohm-m)\n eps = 1.0\n # percentage\n relative = 0.05\n dmisfit = data_misfit.L2DataMisfit(simulation=prb, data=data)\n uncert = abs(data.dobs) * relative + eps\n dmisfit.standard_deviation = uncert\n\n # Map for a regularization\n regmap = maps.IdentityMap(nP=int(actind.sum()))\n\n # Related to inversion\n reg = regularization.Sparse(mesh, indActive=actind, mapping=regmap)\n opt = optimization.InexactGaussNewton(maxIter=15)\n invProb = inverse_problem.BaseInvProblem(dmisfit, reg, opt)\n beta = directives.BetaSchedule(coolingFactor=5, coolingRate=2)\n betaest = directives.BetaEstimate_ByEig(beta0_ratio=1e0)\n target = directives.TargetMisfit()\n updateSensW = directives.UpdateSensitivityWeights()\n update_Jacobi = directives.UpdatePreconditioner()\n inv = inversion.BaseInversion(\n invProb, directiveList=[beta, target, updateSensW, betaest, update_Jacobi]\n )\n prb.counter = opt.counter = utils.Counter()\n opt.LSshorten = 0.5\n opt.remember(\"xc\")\n\n # Run inversion\n mopt = inv.run(m0)\n\n # Get diag(JtJ)\n mask_inds = np.ones(mesh.nC, dtype=bool)\n jtj = np.sqrt(updateSensW.JtJdiag[0])\n jtj /= jtj.max()\n temp = np.ones_like(jtj, dtype=bool)\n temp[jtj > 0.005] = False\n mask_inds[actind] = temp\n actind_final = np.logical_and(actind, ~mask_inds)\n jtj_cc = np.ones(mesh.nC) * np.nan\n jtj_cc[actind] = jtj\n\n # Show the sensitivity\n if plotIt:\n fig = plt.figure(figsize=(12, 3))\n ax = plt.subplot(111)\n temp = rho.copy()\n temp[~actind] = np.nan\n out = mesh.plot_image(\n jtj_cc,\n grid=True,\n ax=ax,\n grid_opts={\"alpha\": 0.2},\n clim=(0.005, 0.5),\n pcolor_opts={\"cmap\": \"viridis\", \"norm\": colors.LogNorm()},\n )\n ax.plot(\n survey.electrode_locations[:, 0], survey.electrode_locations[:, 1], \"k.\"\n )\n ax.set_xlim(IO.grids[:, 0].min(), IO.grids[:, 0].max())\n ax.set_ylim(-IO.grids[:, 1].max(), IO.grids[:, 1].min())\n cb = plt.colorbar(out[0])\n cb.set_label(\"Sensitivity\")\n ax.set_aspect(\"equal\")\n plt.show()\n\n # Convert obtained inversion model to resistivity\n # rho = M(m), where M(.) is a mapping\n\n rho_est = mapping * mopt\n rho_est[~actind_final] = np.nan\n rho_true = rho.copy()\n rho_true[~actind_final] = np.nan\n\n # show recovered conductivity\n if plotIt:\n fig, ax = plt.subplots(2, 1, figsize=(20, 6))\n out1 = mesh.plot_image(\n rho_true,\n clim=(10, 1000),\n pcolor_opts={\"cmap\": \"viridis\", \"norm\": colors.LogNorm()},\n ax=ax[0],\n )\n out2 = mesh.plot_image(\n rho_est,\n clim=(10, 1000),\n pcolor_opts={\"cmap\": \"viridis\", \"norm\": colors.LogNorm()},\n ax=ax[1],\n )\n out = [out1, out2]\n for i in range(2):\n ax[i].plot(\n survey.electrode_locations[:, 0], survey.electrode_locations[:, 1], \"kv\"\n )\n ax[i].set_xlim(IO.grids[:, 0].min(), IO.grids[:, 0].max())\n ax[i].set_ylim(-IO.grids[:, 1].max(), IO.grids[:, 1].min())\n cb = plt.colorbar(out[i][0], ax=ax[i])\n cb.set_label(r\"Resistivity ($\\Omega$m)\")\n ax[i].set_xlabel(\"Northing (m)\")\n ax[i].set_ylabel(\"Elevation (m)\")\n ax[i].set_aspect(\"equal\")\n plt.tight_layout()\n plt.show()\n\n\nif __name__ == \"__main__\":\n run()\n","repo_name":"simpeg/simpeg","sub_path":"examples/_archived/plot_inv_dcip_dipoledipole_2_5Dinversion.py","file_name":"plot_inv_dcip_dipoledipole_2_5Dinversion.py","file_ext":"py","file_size_in_byte":7779,"program_lang":"python","lang":"en","doc_type":"code","stars":438,"dataset":"github-code","pt":"7"} +{"seq_id":"25072093028","text":"from ast import AST, Store\nfrom typing import Iterator, List, Optional\n\nfrom .WalkAST import NodeStack\nfrom ..primitives import Primitive\n\nclass Reference:\n\tdef __init__(self, node: AST, stack: NodeStack):\n\t\tself.node = node\n\t\tself.stack = stack.copy()\n\t\tself.is_store: bool = hasattr(node, 'ctx') and type(node.ctx) == Store\n\t\n\tdef __eq__(self, other):\n\t\tif isinstance(other, AST):\n\t\t\treturn self.node == other\n\t\treturn super().__eq__(other)\n\nclass ProgramData:\n\n\tclass Element:\n\n\t\tdef __init__(\n\t\t\tself,\n\t\t\ttype: str,\n\t\t\tscope: int,\n\t\t\tid: str,\n\t\t\tdeclaration: Optional[AST],\n\t\t\tstack: Optional[NodeStack],\n\t\t) -> None:\n\t\t\tself.type = type\n\t\t\tself.scope = scope\n\t\t\tself.id = id\n\t\t\tself.declaration: Optional[Reference] = None\n\t\t\tself.container_scope: Optional[int] = None\n\t\t\tself.references: List[Reference] = []\n\t\t\tself.values: List[Primitive] = []\n\t\t\tself.evaluations: List[int] = []\n\t\t\tif declaration:\n\t\t\t\tself.declaration = Reference(declaration, stack or [])\n\n\t\tdef set_container_scope(self, scope: int) -> None:\n\t\t\tself.container_scope = scope\n\n\t\tdef reference(self, node: AST, stack: NodeStack) -> None:\n\t\t\tself.references.append(Reference(node, stack))\n\n\t\tdef value(self, value: Primitive) -> None:\n\t\t\tself.values.append(value)\n\n\t\tdef evaluation(self, branch: int) -> None:\n\t\t\tself.evaluations.append(branch)\n\n\t\tdef has_reference(self, node: AST) -> bool:\n\t\t\treturn node == self.declaration or node in self.references\n\n\t\tdef __repr__(self) -> str:\n\t\t\tcscope = f' -> S{self.container_scope}' if self.container_scope else ''\n\t\t\tinfo = [f'{self.type}: (S{self.scope}{cscope}) {self.id}']\n\t\t\tif not self.declaration is None:\n\t\t\t\tinfo.append(f' declared on {self.declaration.node.lineno}')\n\t\t\tif len(self.references) > 0:\n\t\t\t\tinfo.append(f' referenced on {\", \".join(str(r.node.lineno) for r in self.references)}')\n\t\t\tif len(self.values) > 0:\n\t\t\t\tinfo.append(f' assigned values {\", \".join(str(v) for v in self.values)}')\n\t\t\tif len(self.evaluations) > 0:\n\t\t\t\tinfo.append(f' evaluated {len(self.evaluations)} times')\n\t\t\treturn '\\n'.join(info)\n\n\tdef __init__(self):\n\t\tself.elements: List[self.Element] = []\n\n\tdef declare(\n\t\tself,\n\t\ttype: str,\n\t\tscope: int,\n\t\tid: str,\n\t\tdeclaration: Optional[AST] = None,\n\t\tstack: Optional[NodeStack] = None\n\t) -> Element:\n\t\tel = self.Element(type, scope, id, declaration, stack)\n\t\tself.elements.append(el)\n\t\treturn el\n\n\tdef element_n(self, n: int) -> Element:\n\t\treturn self.elements[n]\n\n\tdef element_for_node(self, node: AST, types: List[str]) -> Element:\n\t\tfor i, el in enumerate(self.elements):\n\t\t\tif el.type in types and el.has_reference(node):\n\t\t\t\treturn i, el\n\t\treturn None, None\n\t\n\tdef elements_for_types(self, types: List[str]) -> Iterator[Element]:\n\t\tfor el in self.elements:\n\t\t\tif el.type in types:\n\t\t\t\tyield el\n\n\tdef element_for_id(self, id: str) -> Element:\n\t\tfor el in self.elements:\n\t\t\tif el.id == id:\n\t\t\t\treturn el\n\t\treturn None\n\t\n\tdef element_for_scope(self, scope: int) -> Element:\n\t\tfor el in self.elements:\n\t\t\tif el.container_scope == scope:\n\t\t\t\treturn el\n\t\treturn None\n\t\n\tdef elements_in_scope(self, scope: int, types: List[str] = None) -> Iterator[Element]:\n\t\tfor el in self.elements:\n\t\t\tif el.scope == scope and (types is None or el.type in types):\n\t\t\t\tyield el\n\n\tdef __repr__(self) -> str:\n\t\treturn '\\n'.join(str(el) for el in self.elements)\n","repo_name":"teemulehtinen/qlcpy","sub_path":"qlcpy/instrument/ProgramData.py","file_name":"ProgramData.py","file_ext":"py","file_size_in_byte":3303,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"45989110637","text":"import dotenv\nimport openai\nimport re\nimport pandas as pd\nimport streamlit as st\nfrom streamlit_lottie import st_lottie\nfrom utils import load_lottie_url\n\nst.set_page_config(page_title=\"TestCaseGPT\", page_icon=\"🤖\", layout=\"wide\")\nconfig = dotenv.dotenv_values(\".env\")\nopenai.api_key = config['OPENAI_API_KEY']\n\nwith st.sidebar:\n lottie_image1 = load_lottie_url('https://assets1.lottiefiles.com/packages/lf20_ofa3xwo7.json')\n st_lottie(lottie_image1)\n\nst.markdown(\n \"\"\"\n

TestCaseGPT,测试流程加速器 🚀️

\n

将用户故事转化为测试用例,只需一步 !

\n \"\"\",\n unsafe_allow_html=True\n)\n\n\ndef local_css(file_name):\n with open(file_name) as f:\n st.markdown(f\"\", unsafe_allow_html=True)\n\n\nlocal_css(\"style.css\")\nuser_story = st.text_input(label=\"📖 用户故事\", label_visibility=\"hidden\",\n placeholder=\"【用户故事描述】:作为___,我希望___,以便___。\", key=\"input\")\n\nprompt_userstory = \"我希望你作为一个软件产品经理,负责生成验收标准,用来验证软件是否符合用户故事中指定的功能要求。验收标准应该是具体的、可衡量的、可实现的、相关的。此外,你应该确保验收标准涵盖所有可能的情况和边缘案例。通过定义清晰而全面的验收标准,你可以帮助确保软件符合必要的标准,并确保用户的需求得到满足。按照描述的格式,就下面的主题写出10条专业而详细的验收标准。请尽你最大的努力。用中文回答。只返回验收标准的内容。不要返回其他内容。\" \\\n \"\\n主题: \" + user_story\n\nprompt_testcase = \"您是软件测试和软件质量保证方面的专家,专门从事功能测试,您帮助我之前的许多人生成了满足特定要求的功能测试用例。\\n\" \\\n \"您生成的测试用例能涵盖正常场景、异常场景、边界场景。\\n\" \\\n \"您生成的测试用例优先级包括 P0、P1、P2,P0为最高优先级,P2代表最低优先级。\\n\" \\\n \"以所述测试用例格式,至少编写五条关于以下主题的专业和详细测试用例。尽你最大的努力。请使用中文回答, 请勿返回除测试用例内容以外的其他内容。不要用引号包装响应。\\n\" \\\n \"测试用例格式:\\n\" \\\n \"用例编号:\\n\" \\\n \"用例名称:\\n\" \\\n \"用例类型:\\n\" \\\n \"优先级:\\n\" \\\n \"前置条件:\\n\" \\\n \"步骤描述:\\n\" \\\n \"预期结果:\\n\" \\\n \"主题: \"\n\n\ndef clean_criteria(text):\n # 先对文本按照换行符进行分割\n lines = text.split(\"\\n\")\n new_lines = []\n for line in lines:\n # 如果是空行,空字符串,单个字符串,直接丢弃\n if len(line) < 2:\n continue\n new_lines.append(re.sub(r'^(\\d+).', '📝 ', line))\n\n text = \"\\n\".join(new_lines)\n return text\n\n\ndef clean_testcase(text):\n # 先对文本按照换行符进行分割,如果有连续的换行符,那么分割出来的元素会是空字符串\n lines = text.split(\"\\n\")\n new_lines = \"\"\n for line in lines:\n # 如果不是空字符串,那么就是正常的文本,需要进行处理\n if line != \"\":\n # 先把正常文本开头的数字和点给去掉\n line = re.sub(r'^(\\d+)\\. ', '', line)\n # 继续去除末尾的空字符串\n line = line.rstrip()\n # 把处理好的文本拼接起来\n new_lines += line + \" \"\n # 如果是空字符串,那么就是连续的换行符\n else:\n # 先把新字符串末尾的\n new_lines.rstrip(\" \")\n # 再添加一个换行符,用来分隔不同的测试用例\n new_lines += \"\\n\"\n\n return new_lines\n\n\ndef output_criteria(prompt):\n # 输出验收标准\n generate_criteria = \"\"\n openai_resp = []\n for resp in openai.Completion.create(model=\"text-davinci-003\", prompt=prompt, stream=True,\n max_tokens=1024,\n temperature=0.7\n ):\n openai_resp.append(resp.choices[0].text)\n generate_criteria = \"\".join(openai_resp).strip()\n formatted_text = f\"

{clean_criteria(generate_criteria)}

\"\n criteria_box.markdown(formatted_text, unsafe_allow_html=True)\n return generate_criteria\n\n\ndef output_testcase(case_title):\n formatted_text = \"\"\n openai_resp = []\n prompt = prompt_testcase + case_title\n for resp in openai.Completion.create(model=\"text-davinci-003\", prompt=prompt, stream=True,\n max_tokens=3072,\n temperature=0.7):\n openai_resp.append(resp.choices[0].text)\n generate_testcase = \"\".join(openai_resp).strip()\n formatted_text = clean_testcase(generate_testcase)\n markdown_text = f\"

{formatted_text}

\"\n case_box.markdown(markdown_text, unsafe_allow_html=True)\n return formatted_text\n\n\ndef export_testcase(InputCase):\n # 定义正则表达式\n regex = r\"用例编号(.+) 用例名称(.+) 用例类型(.+) 优先级(.+) 前置条件(.+) 步骤描述(.+) 预期结果(.+)\"\n TestCaseLines = re.findall(regex, InputCase)\n CaseIds = []\n Names = []\n CaseTypes = []\n Priorities = []\n Preconditions = []\n Steps = []\n ExpectedResults = []\n for line in TestCaseLines:\n # 删除开头的中英文冒号和空格\n CaseId = line[0]\n CaseId = re.sub(r'^:', '', CaseId)\n CaseId = re.sub(r'^:', '', CaseId).strip()\n CaseIds.append(CaseId)\n\n # 删除开头的中英文冒号和空格\n Name = line[1]\n Name = re.sub(r'^:', '', Name)\n Name = re.sub(r'^:', '', Name).strip()\n Names.append(Name)\n\n # 删除开头的中英文冒号和空格\n CaseType = line[2]\n CaseType = re.sub(r'^:', '', CaseType)\n CaseType = re.sub(r'^:', '', CaseType).strip()\n CaseTypes.append(CaseType)\n\n # 删除开头的中英文冒号和空格\n Priority = line[3]\n Priority = re.sub(r'^:', '', Priority)\n Priority = re.sub(r'^:', '', Priority).strip()\n Priorities.append(Priority)\n\n # 删除开头的中英文冒号和空格\n Precondition = line[4]\n Precondition = re.sub(r'^:', '', Precondition)\n Precondition = re.sub(r'^:', '', Precondition).strip()\n Preconditions.append(Precondition)\n\n # 删除开头的中英文冒号和空格\n Step = line[5]\n Step = re.sub(r'^:', '', Step)\n Step = re.sub(r'^:', '', Step).strip()\n Steps.append(Step)\n\n # 删除开头的中英文冒号和空格\n ExpectedResult = line[6]\n ExpectedResult = re.sub(r'^:', '', ExpectedResult)\n ExpectedResult = re.sub(r'^:', '', ExpectedResult).strip()\n ExpectedResults.append(ExpectedResult)\n\n test_case_data = {'用例名称': Names, '用例类型': CaseTypes, '优先级': Priorities,\n '前置条件': Preconditions,\n '步骤描述': Steps, '预期结果': ExpectedResults}\n data = pd.DataFrame(test_case_data)\n st.dataframe(data)\n\n # 将DataFrame写入Excel文件\n excel_data = {'用例名称': Names, '用例类型': CaseTypes, '优先级': Priorities,\n '前置条件': Preconditions,\n '步骤描述': Steps, '预期结果': ExpectedResults, '维护人': \"marsdev\"}\n data = pd.DataFrame(excel_data)\n data.to_excel('测试用例.xlsx', index=False)\n\n # 将Excel文件读取为字节流\n\n if st.download_button:\n st.error('想体验更多付费功能,请升级到 Testcaseplus💫', icon=\"🚨\")\n with open('测试用例.xlsx', 'rb') as f:\n excel_bytes = f.read()\n # 将Excel文件作为字节流提供给用户进行下载\n st.download_button(label='导出测试用例', data=excel_bytes, file_name='测试用例.xlsx',\n mime='application/vnd.openxmlformats-officedocument.spreadsheetml.sheet')\n\n\nif st.button(\"一键生成测试用例\", type=\"primary\"):\n user_story.replace(\"\\n\", \"\")\n if not user_story:\n st.warning('请输入用户故事')\n st.stop()\n\n criteria_box = st.expander(label=\"测试点拆分\", expanded=True)\n with criteria_box:\n criteria_box = st.empty()\n criteria = output_criteria(prompt_userstory)\n testcase_box = st.expander(label=\"测试用例生成\", expanded=True)\n with testcase_box:\n case_box = st.empty()\n all_case = re.split(r\"\\n\", criteria)\n case_list = []\n for case in all_case:\n print(\"before sub\")\n print(case)\n case = re.sub(r'^(\\d+).', '', case)\n case = case.lstrip()\n print(\"after sub\")\n print(case)\n if len(case) < 2:\n continue\n case_list.append(case)\n print(case_list)\n testcase = output_testcase(case_list[0])\n export_testcase(testcase)\n\n","repo_name":"sunwandi04/testcase","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":9512,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"75083079264","text":"import os\n\ncook_book = dict()\n\nwith open('recipes.txt', 'r', encoding='utf-8') as f:\n lines = f.readlines()\n list = []\n result = []\n n = '\\n'\n for l in lines:\n if l == '\\n':\n list += l\n result.append(list)\n list = []\n else:\n list.append(l)\n result.append(list)\n\nfor cook_list in result:\n ingredient_list = []\n\n for list_ in cook_list:\n list = list_.replace('\\n', '')\n if \"|\" in list:\n ing_list = list.split(' | ')\n ingredient_list.append(ing_list)\n ing_list = []\n\n for ing_list_2 in ingredient_list:\n ing_dict = dict()\n if ing_list_2:\n ing_dict['ingredient_name'] = ing_list_2[0]\n ing_dict['quantity'] = ing_list_2[1]\n ing_dict['measure'] = ing_list_2[2]\n ing_list.append(ing_dict)\n cook_book[cook_list[0].replace('\\n', '')] = ing_list\n\n\ndef get_shop_list_by_dishes(dishes: list, person_count: int) -> dict:\n \"\"\"Функция обрабатывает список блюд и количество персон.\n\n :param dishes: Список блюд.\n :param person_count: Количество человек.\n :return: Выводит на экран список ингредиентов\n для приготовления всех блюд.\n \"\"\"\n food_basket = {}\n for dish in dishes:\n if dish not in cook_book:\n raise TypeError(f'{dish} отсутствует в кулинарной книге')\n for ingredient in cook_book[dish]:\n if ingredient['ingredient_name'] not in food_basket:\n food_basket[ingredient['ingredient_name']] \\\n = {'measure': ingredient['measure'], 'quantity': \\\n int(ingredient['quantity']) * person_count}\n else:\n food_basket[ingredient['ingredient_name']]['quantity'] \\\n += int(ingredient['quantity']) * person_count\n return food_basket\n\n\ndef get_dirs_and_files():\n \"\"\" Функция создает ссылку \"dirs\" на заданную директорию и\n создает список файлов в ней.\n\n :return: Вызывает функцию get_len_files() и передает\n подготовленные данные.\"\"\"\n path = os.path.join(os.getcwd() + '\\sorted/')\n result = os.walk(path)\n for dirs, folder, files in result:\n dirs_ = dirs\n files_ = files\n return get_len_files(dirs_, files_)\n\n\ndict_lens_and_files = dict()\n\n\ndef get_len_files(dirs, files):\n \"\"\" Функция создает словарь. Ключами словаря являются количкество\n строк в файлах, а значениями имена файлов соответственно.\n\n :param dirs: Ссылка на заданную директорию.\n :param files: Список файлов в заданной папке.\n :return: Вызывает функцию sorted_files_dict().\"\"\"\n for file in files:\n file_path = dirs + file\n with open(file_path, 'r', encoding='utf-8') as f:\n lines = f.readlines()\n file_len = len(lines)\n dict_lens_and_files[file_len] = file\n return sorted_files_dict()\n\n\nsorted_dict = dict()\n\n\ndef sorted_files_dict():\n \"\"\" Функция сортирует словарь по ключу.\n :return: Вызывает функцию save_sorted_files().\"\"\"\n sorted_dict.update(dict(sorted(dict_lens_and_files.items())))\n return save_sorted_files()\n\n\ndef save_sorted_files():\n \"\"\" Функция записывает информацию о всех файлах и текст из самих файлов\n из заданной папки в новый файл.\n Каждый раз файл удаляется и создается заново.\n :return: Результат всей операции выводится на экран.\"\"\"\n file_link = os.path.join(os.getcwd(), 'file.txt')\n os.remove(file_link)\n for file_num in sorted_dict:\n file_for_link = sorted_dict.get(file_num)\n file_link_ = os.path.join(os.getcwd() + '\\sorted/', file_for_link)\n h1 = sorted_dict.get(file_num)\n with open(file_link_, 'rt', encoding='utf-8') as fl:\n lines = fl.readlines()\n with open(file_link, 'a+') as f:\n f.write(f'{h1}\\n')\n f.write(f'{file_num}\\n')\n for line in lines:\n f.write(line.strip())\n f.write(f'\\n')\n f.flush()\n with open('file.txt') as f1:\n result = f1.read()\n return result\n\n\nprint()\nprint()\n\n# Вызываем функцию по получению списка продуктов из кулинарной\n# книги согласно списку блюд\nprint(get_shop_list_by_dishes(['Запеченный картофель', 'Омлет', 'Фахитос'], 2))\n\nprint()\nprint()\n\n# Вызываем функцию для последующей сортировки и объединения\n# текстовых файлов в один\nprint(get_dirs_and_files())\n","repo_name":"VolshVs/OOP_2","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5222,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"73856997022","text":"from concurrent.futures import ThreadPoolExecutor, as_completed\nfrom typing import Union\n\nfrom gpt_generator.modules.architector import gpt_architector\nfrom gpt_generator.modules.worker import gpt_worker\nfrom helpers.architecture_extractor import structure_2_dict\nfrom helpers.code_extractor import extract_code\nfrom models.file_interface import File_Collection, File_to_write\nfrom models.gpt_responses_interface import ArchitectorResponse\n\nimport logging\n\nlogging.basicConfig(level=logging.ERROR)\n\n\ndef gpt_main(prompt: str) -> File_Collection:\n try:\n # Generate initial project and response\n project, response = gpt_architector(prompt)\n\n # Assert project is of the correct type\n assert isinstance(project, ArchitectorResponse)\n\n # Initialize File_Collection\n file_collection = File_Collection(project.name)\n\n # Add initial response to collection\n file_collection.collections[\"raw_responses\"].append(\n File_to_write(\"architector.txt\", str(response))\n )\n\n structure_dictionary: dict[str, str] = structure_2_dict(project.structure)\n\n # Create a ThreadPool to execute gpt_worker on each file\n with ThreadPoolExecutor() as executor:\n future_to_file = {executor.submit(\n gpt_worker, project, file.file_name, file.description): file for file in project.files}\n\n for future in as_completed(future_to_file):\n try:\n # Obtain the completed future's corresponding file\n file = future_to_file[future]\n\n # Get the result from the future\n result = future.result()\n\n # Add results to the collection\n file_collection.collections[\"raw_responses\"].append(\n File_to_write(f\"raw_{file.file_name}.txt\", result, \"\")\n )\n file_collection.collections[\"formatted_responses\"].append(\n File_to_write(file.file_name, extract_code(result),\n structure_dictionary[file.file_name])\n )\n except Exception as e:\n # Log any exceptions that occur within the thread\n logging.error(f\"An error occurred in thread: {e}\")\n\n return file_collection\n\n except Exception as e:\n # Log any other exceptions\n logging.error(f\"An error occurred: {e}\")\n return File_Collection(\"\")\n","repo_name":"bukomp/mAIcroservicer","sub_path":"src/gpt_generator/gpt_main.py","file_name":"gpt_main.py","file_ext":"py","file_size_in_byte":2270,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"14526738976","text":"#!/usr/bin/env python\n\nimport subprocess\n\nimport inspect, os, sys\n# From http://stackoverflow.com/questions/279237/python-import-a-module-from-a-folder\ncmd_subfolder = os.path.realpath(os.path.abspath(os.path.join(os.path.split(inspect.getfile( inspect.currentframe() ))[0],\"..\")))\nif cmd_subfolder not in sys.path:\n sys.path.insert(0, cmd_subfolder)\n\nimport mosq_test\n\ndef write_config(filename, port1, port2):\n with open(filename, 'w') as f:\n f.write(\"port %d\\n\" % (port1))\n f.write(\"\\n\")\n f.write(\"listener %d\\n\" % (port2))\n f.write(\"mount_point mount/\\n\")\n f.write(\"\\n\")\n f.write(\"log_type debug\\n\")\n\n(port1, port2) = mosq_test.get_port(2)\nconf_file = os.path.basename(__file__).replace('.py', '.conf')\nwrite_config(conf_file, port1, port2)\n\nrc = 1\nkeepalive = 60\nconnect_packet = mosq_test.gen_connect(\"test2\", keepalive=keepalive)\nconnack_packet = mosq_test.gen_connack(rc=0)\n\nmid = 1\nsubscribe_packet = mosq_test.gen_subscribe(mid, \"#\", 0)\nsuback_packet = mosq_test.gen_suback(mid, 0)\n\npublish_packet = mosq_test.gen_publish(\"mount/test\", qos=0, payload=\"mount point\")\n\nbroker = mosq_test.start_broker(filename=os.path.basename(__file__), use_conf=True, port=port1)\n\ntry:\n sock = mosq_test.do_client_connect(connect_packet, connack_packet, timeout=20, port=port1)\n sock.send(subscribe_packet)\n\n if mosq_test.expect_packet(sock, \"suback\", suback_packet):\n pub = subprocess.Popen(['./10-listener-mount-point-helper.py', str(port2)], stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n pub.wait()\n (stdo, stde) = pub.communicate()\n # Should have now received a publish command\n\n if mosq_test.expect_packet(sock, \"publish\", publish_packet):\n rc = 0\n\n sock.close()\nfinally:\n os.remove(conf_file)\n broker.terminate()\n broker.wait()\n (stdo, stde) = broker.communicate()\n if rc:\n print(stde)\n\nexit(rc)\n\n","repo_name":"hui6075/mosquitto-cluster","sub_path":"test/broker/10-listener-mount-point.py","file_name":"10-listener-mount-point.py","file_ext":"py","file_size_in_byte":1927,"program_lang":"python","lang":"en","doc_type":"code","stars":316,"dataset":"github-code","pt":"7"} +{"seq_id":"2372667998","text":"import copy\nimport random\n\n\nclass AI:\n\n def __init__(self, player=2):\n self.player = player\n\n def random(self, board):\n empty_squares = board.get_empty_squares()\n random_index = random.randrange(0, len(empty_squares))\n return empty_squares[random_index]\n\n def minimax(self, board, maximizing):\n\n case = board.check_victory_conditions()\n\n if case == 1:\n return 1, None\n\n if case == 2:\n return -1, None\n\n elif board.isfull():\n return 0, None\n\n if maximizing:\n max_evaluation = -10\n best_move = None\n empty_squares = board.get_empty_squares()\n\n for (row, col) in empty_squares:\n temp_board = copy.deepcopy(board)\n temp_board.mark_square(row, col, 1)\n evaluation = self.minimax(temp_board, False)[0]\n if evaluation > max_evaluation:\n max_evaluation = evaluation\n best_move = (row, col)\n\n return max_evaluation, best_move\n\n elif not maximizing:\n min_evaluation = 10\n best_move = None\n empty_squares = board.get_empty_squares()\n\n for (row, col) in empty_squares:\n temp_board = copy.deepcopy(board)\n temp_board.mark_square(row, col, self.player)\n evaluation = self.minimax(temp_board, True)[0]\n if evaluation < min_evaluation:\n min_evaluation = evaluation\n best_move = (row, col)\n\n return min_evaluation, best_move\n\n def evaluate_impossible_difficulty(self, main_board):\n evaluation, move = self.minimax(main_board, False)\n return move\n\n def evaluate_medium_difficulty(self, main_board):\n move = self.random(main_board)\n i = random.randint(1, 6)\n if i in range(1, 5):\n evaluation, move = self.minimax(main_board, False)\n return move\n\n def evaluate_easy_difficulty(self, main_board):\n move = self.random(main_board)\n return move\n","repo_name":"Tox2401/TicTacToe","sub_path":"AI.py","file_name":"AI.py","file_ext":"py","file_size_in_byte":2115,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13629392545","text":"# Get propane data from Otodata API\n# The time-based series are stored in InfluxDB.\n\n# Requires username/password used with the Nee-vo mobile app (or similar app)\n\nimport json\nimport requests\nimport datetime\nfrom pytz import timezone\nfrom influxdb import InfluxDBClient\n\n# Set parameters\nNEEVO_AUTH = ''\nTIMEZONE = 'US/Eastern'\nPROVIDER_NAME = ''\nMETRIC_NAME = 'propane'\n\nINFLUXDB_HOST = ''\nINFLUXDB_PORT = ''\nINFLUXDB_USERNAME = ''\nINFLUXDB_PASSWORD = ''\nINFLUXDB_DATABASE = ''\n\n# Instantiate database\ninfluxClient = InfluxDBClient(\n host=INFLUXDB_HOST,\n port=INFLUXDB_PORT,\n username=INFLUXDB_USERNAME,\n password=INFLUXDB_PASSWORD,\n database=INFLUXDB_DATABASE\n)\n\n\ndef getDateTimeByZone(tz):\n t = timezone(tz)\n return datetime.datetime.now(t)\n\n\ndef printme(str):\n t = getDateTimeByZone(TIMEZONE)\n print(t.strftime(\"%Y-%m-%d %H:%M:%S\"), str)\n return\n\n\ndef processChartData(tank_size, tank_level):\n try:\n remaining = (tank_level/100) * tank_size\n json_body = []\n json_body.append({\n \"measurement\": METRIC_NAME,\n \"tags\": {\n \"source\": PROVIDER_NAME\n },\n \"time\": getDateTimeByZone(TIMEZONE),\n \"fields\": {\n \"litres\": remaining,\n \"level\": tank_level\n }\n })\n influxClient.write_points(json_body)\n\n except Exception as e:\n if hasattr(e, 'message'):\n printme(e.message)\n else:\n printme(e)\n exit(0)\n\ndef main():\n\n session = requests.Session()\n\n printme(\"Getting levels\")\n r = session.get(\n \"https://telematics.otodatanetwork.com:4432/v1.5/DataService.svc/GetAllDisplayPropaneDevices\",\n headers={\n \"Accept\": \"*/*\",\n \"Accept-Language\": \"en-US;q=1, fa-US;q=0.9\",\n \"User-Agent\": \"Nee-Vo/2.5 (iPhone; iOS 16.0.2; Scale/3.00)\",\n \"Host\": \"telematics.otodatanetwork.com:4432\",\n \"Authorization\": NEEVO_AUTH\n }\n )\n\n j = json.loads(r.text)\n processChartData(j[0]['TankCapacity'], j[0]['Level'])\n\n printme('Finished')\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"CeeBeeEh/Otodata-HASS","sub_path":"otodata-hass.py","file_name":"otodata-hass.py","file_ext":"py","file_size_in_byte":2224,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"20295782830","text":"from typing import Iterable\nfrom meshed import DAG\nfrom front import APP_KEY, RENDERING_KEY, ELEMENT_KEY, NAME_KEY\nfrom collections.abc import Callable\nfrom front.crude import prepare_for_crude_dispatch\nfrom streamlitfront.elements import TextInput, SelectBox\n\nfrom streamlitfront.base import mk_app\nfrom streamlitfront.examples.util import Graph\nfrom streamlitfront.elements import (\n AudioRecorder,\n FileUploader,\n MultiSourceInput,\n)\nimport streamlit as st\n\nparam_to_mall_maps = dict(train_audio='train_audio', tag='tag_store')\n\nif 'mall' not in st.session_state:\n st.session_state['mall'] = dict(\n # train_audio={},\n # tag={},\n # unused_store={\"to\": \"illustrate\"},\n tag_sound_output={}\n )\n\nmall = st.session_state['mall']\n# mall = dict(\n# # train_audio={},\n# # tag={},\n# # unused_store={\"to\": \"illustrate\"},\n# tag_sound_output={}\n# )\n\n\ndef crudify(funcs):\n for func in funcs:\n yield prepare_for_crude_dispatch(\n func, param_to_mall_map=param_to_mall_map, mall=mall\n )\n\n\nWaveForm = Iterable[int]\n\n\n# @code_to_dag\n@prepare_for_crude_dispatch(mall=mall, output_store='tag_sound_output')\ndef tag_sound(train_audio: WaveForm, tag: str):\n # mall[\"tag_store\"] = tag\n return (train_audio, tag)\n\n\n@prepare_for_crude_dispatch(mall=mall, param_to_mall_map={'result': 'tag_sound_output'})\ndef display_tag_sound(result):\n return result\n\n\n# crudified_tag_sound = prepare_for_crude_dispatch(\n# tag_sound, mall=mall, output_store=\"tag_sound_output\"\n# )\n# crudified_display_tag_sound = prepare_for_crude_dispatch(\n# display_tag_sound, mall=mall, param_to_mall_map={\"result\": \"tag_sound_output\"}\n# )\n\nprint(type(tag_sound))\nfrom i2 import Sig\n\nprint(Sig(display_tag_sound))\n\nconfig_ = {\n APP_KEY: {'title': 'Simple Real Audio ML'},\n # OBJ_KEY: {\"trans\": crudify},\n RENDERING_KEY: {\n 'tag_sound': {\n # \"description\": {\"content\": \"A very simple learn model example.\"},\n 'execution': {\n 'inputs': {\n 'train_audio': {\n ELEMENT_KEY: MultiSourceInput,\n 'From a file': {ELEMENT_KEY: FileUploader, 'type': 'wav',},\n 'From the microphone': {ELEMENT_KEY: AudioRecorder},\n },\n # \"tag\": {\n # ELEMENT_KEY: TextInput,\n # },\n },\n }\n },\n 'display_tag_sound': {\n 'execution': {\n 'inputs': {\n 'result': {\n ELEMENT_KEY: SelectBox,\n 'options': mall['tag_sound_output'],\n },\n }\n },\n },\n DAG: {'graph': {ELEMENT_KEY: Graph, NAME_KEY: 'Flow',},},\n Callable: {\n 'execution': {\n 'inputs': {\n 'save_name': {ELEMENT_KEY: TextInput, NAME_KEY: 'Save output as',},\n }\n },\n },\n },\n}\n\napp = mk_app([tag_sound, display_tag_sound], config=config_)\napp()\nst.write(mall)\n","repo_name":"otosense/plunk","sub_path":"plunk/sb/front_experiments/edge_interface/sound_tagger.py","file_name":"sound_tagger.py","file_ext":"py","file_size_in_byte":3124,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"8360088824","text":"'''\nGiven a string if all the chars are unique \n\treturn True\nelse \n\treturn False\n'''\n\n# Time complexity is O(n)\n\ndef IsAllCharsUnique(string):\n\tstring_dict = {}\n\n\tfor i in string:\n\t\tif i in string_dict: return False\n\t\telse: string_dict[i] = 1 \n\n\treturn True\n\nprint(IsAllCharsUnique(\"aebcde\"))","repo_name":"Manoji97/Data-Structures-Algorithms-Complete","sub_path":"InterviewQuestions/Python/1_Arrays/7_UniqueCharsInString.py","file_name":"7_UniqueCharsInString.py","file_ext":"py","file_size_in_byte":293,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"29466616996","text":"import phaseDiversity3PSFs as PD\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport fs\nimport os\nimport pyfits\n#import seaborn as sns\n#sns.set()\n\npupilRadius = 1.6e-3\nlbda = 0.6375e-6\nF = 80e-3\npxsize = 5.3e-6\nN = 400\ndxp = lbda*F/(N*pxsize)\ndeltaZ = 3.19e-3\njmin = 4\njmax = 30\n\n#retrieve ajsIDL and ajsTrue\n\najsTrueFolderPath = 'C:\\\\Users\\\\Jojo\\\\Desktop\\\\PdM-HEIG\\\\Science\\\\data\\\\devPD\\\\PSFforIDLtreatment\\\\ajsTrue'\najsTrueFile = os.listdir(ajsTrueFolderPath)\najsIDLFolderPath = 'C:\\\\Users\\\\Jojo\\\\Desktop\\\\PdM-HEIG\\\\Science\\\\data\\\\devPD\\\\PSFforIDLtreatment\\\\IDLajs'\najsIDLFile = os.listdir(ajsIDLFolderPath)\n\njsTrue = np.zeros([len(ajsIDLFile),jmax-jmin+1])\najsTrue = np.zeros([len(ajsIDLFile),jmax-jmin+1])\nrmsWFeTrue = np.zeros(len(ajsIDLFile))\n\njsIDL = np.zeros([len(ajsIDLFile),jmax-jmin+1])\najsIDLmodal = np.zeros([len(ajsIDLFile),jmax-jmin+1])\nrmsWFeIDLmodalretrieved = np.zeros(len(ajsIDLFile))\nrmsemodal = np.zeros(len(ajsIDLFile))\n\najsIDLzonal = np.zeros([len(ajsIDLFile),jmax-jmin+1])\nrmsWFeIDLzonalretrieved = np.zeros(len(ajsIDLFile))\nrmsezonal = np.zeros(len(ajsIDLFile))\n\nrmsWFeIDLtrue = np.zeros(len(ajsIDLFile))\n\nfor i in np.arange(len(ajsIDLFile)):\n rmsWFeTrue[i] = (((ajsTrueFile[i]).replace('.','_')).split('_'))[-2]\n rmsWFeIDLtrue[i] = (((ajsIDLFile[i]).replace('.','_')).split('_'))[-2]\n \n jsajsTrue = np.loadtxt(ajsTrueFolderPath+'\\\\'+ajsTrueFile[i])\n jsTrue[i,:] = jsajsTrue[0,:]\n ajsTrue[i,:] = jsajsTrue[1,:]*1e9*lbda/2/np.pi\n \n jsajsIDL = np.loadtxt(ajsIDLFolderPath+'\\\\'+ajsIDLFile[i],delimiter=',',skiprows=1)\n jsIDL[i,:] = jsajsIDL[:,0]\n ajsIDLmodal[i,:] = jsajsIDL[:,1]*1000\n ajsIDLzonal[i,:] = jsajsIDL[:,2]*1000\n \n rmsWFeIDLmodalretrieved[i] = fs.RMSwavefrontError(jsIDL[i,:],ajsIDLmodal[i,:])\n rmsemodal[i] = fs.RMSE(ajsIDLmodal[i,:],ajsTrue[i,:])\n rmsWFeIDLzonalretrieved[i] = fs.RMSwavefrontError(jsIDL[i,:],ajsIDLzonal[i,:])\n rmsezonal[i] = fs.RMSE(ajsIDLzonal[i,:],ajsTrue[i,:])\n\n\nrmsWFefolderPath = 'C:\\\\Users\\\\Jojo\\\\Desktop\\\\PdM-HEIG\\\\Science\\\\data\\\\devPD\\\\PSFforIDLtreatment\\\\PSFs'\nrmsWFerrorFolderPaths = os.listdir(rmsWFefolderPath)\n\nNrmsWFe = len(rmsWFerrorFolderPaths)\nrmsePy = np.zeros(NrmsWFe)\nrmsWFerrorsPyRetrieved = np.zeros(NrmsWFe)\nrmsWFeTruePy = np.zeros(NrmsWFe)\njsretrieved = np.zeros([NrmsWFe,jmax-jmin+1])\najsretrieved = np.zeros([NrmsWFe,jmax-jmin+1])\n\nfor irms,rmsWFdir in enumerate(rmsWFerrorFolderPaths):\n PSFfolderPaths = rmsWFefolderPath+'\\\\' + rmsWFdir\n sPSFfiles = os.listdir(PSFfolderPaths)\n rmsWFeTruePy[irms] = (rmsWFdir.split('_'))[-1]\n \n PSFs = np.zeros([3,400,400])\n deltaZs = np.array([])\n for ipsf,sPSFfile in enumerate(sPSFfiles):\n PSFfilePath = rmsWFefolderPath+ '\\\\' + rmsWFdir + '\\\\' + sPSFfile\n hdulist = pyfits.open(PSFfilePath)\n PSFs[ipsf,:,:] = hdulist[0].data\n deltaZs = np.append(deltaZs,int(((sPSFfile.replace('.','_')).split('_'))[-2]))\n \n IxdeltaZ =np.argsort(deltaZs)\n \n phaseDiv = PD.phaseDiversity3PSFs(PSFs[IxdeltaZ[1],:,:],PSFs[IxdeltaZ[2],:,:],PSFs[IxdeltaZ[0],:,:],deltaZ,lbda,pxsize,F,pupilRadius,jmin,jmax)\n \n jsretrieved[irms,:] = phaseDiv.result['js']\n ajsretrieved[irms,:] = phaseDiv.result['ajs']*1e9*lbda/2/np.pi\n \n a = np.where(rmsWFeTrue==rmsWFeTruePy[irms])\n \n rmsePy[irms] = fs.RMSE(ajsretrieved[irms,:],ajsTrue[a,:])\n rmsWFerrorsPyRetrieved[irms] = fs.RMSwavefrontError(jsretrieved[irms,:],ajsretrieved[irms,:])\n\n\n\nrmsWFerrorMax = np.max(np.append(rmsWFerrorsPyRetrieved,rmsWFeTrue))\nrmsWFerrorMin = np.min(np.append(rmsWFerrorsPyRetrieved,rmsWFeTrue))\n\nfnamerms = '../../../fig/PDDev/compDiversity/rmsWFerrorsretrieved_rmsWFeWthIDL%s'\nfnamermse = '../../../fig/PDDev/compDiversity/rmse_rmsWFeWthIDL%s'\nIxsortIDl = np.argsort(rmsWFeIDLtrue)\nIxsortPy = np.argsort(rmsWFeTruePy)\nfig = plt.figure()\nplt.hold(True)\nplt.plot(rmsWFeTruePy[IxsortPy],rmsWFerrorsPyRetrieved[IxsortPy],label='Analytical Algo')\nplt.plot(rmsWFeIDLtrue[IxsortIDl],rmsWFeIDLmodalretrieved[IxsortIDl],label='ONERA modal')\nplt.plot(rmsWFeIDLtrue[IxsortIDl],rmsWFeIDLzonalretrieved[IxsortIDl],label='ONERA zonal')\nplt.plot([rmsWFerrorMin,rmsWFerrorMax],[rmsWFerrorMin,rmsWFerrorMax],linewidth=2,c='grey')\nplt.xlim([rmsWFerrorMin,rmsWFerrorMax])\nplt.ylim([rmsWFerrorMin,rmsWFerrorMax])\nplt.xlabel('$\\sigma_{WF,rms}$ true [nm]')\nplt.ylabel('$\\sigma_{WF,rms}$ retrieved [nm]')\nplt.legend(loc='best')\nplt.grid()\nplt.savefig(fnamerms % ('.png'), dpi=300)\nplt.savefig(fnamerms % ('.pdf'), dpi=300)\n#plt.close(fig)\n\nfig = plt.figure()\nplt.plot(rmsWFeTruePy[IxsortPy],rmsePy[IxsortPy],label='Analytical Algo')\nplt.plot(rmsWFeIDLtrue[IxsortIDl],rmsemodal[IxsortIDl],label='ONERA modal')\nplt.plot(rmsWFeIDLtrue[IxsortIDl],rmsezonal[IxsortIDl],label='ONERA zonal')\nplt.xlim([rmsWFerrorMin,rmsWFerrorMax])\nplt.xlabel('$\\sigma_{WF,rms}$ true [nm]')\nplt.ylabel('RMSE [nm]')\nplt.legend(loc='best')\nplt.grid()\nplt.savefig(fnamermse % ('.png'), dpi=300)\nplt.savefig(fnamermse % ('.pdf'), dpi=300)\n#plt.close(fig)\n\n\nfor irms,rmsWFe in enumerate(rmsWFeTrue):\n filename = '../../../fig/PDDev/compDiversity/js_ajs_Python_IDL_rmsWfe_%d%s'\n a = np.where(rmsWFeTruePy==rmsWFe)\n plt.figure() \n plt.title('$\\sigma_{WF,rms}$ = %4.1f' % (rmsWFe))\n plt.hold(True)\n plt.plot(jsTrue[irms,:],ajsTrue[irms,:],linewidth = 3,label='True')\n plt.plot(jsretrieved[irms,:],ajsretrieved[irms,:],linewidth=2,label='Python retrieved')\n plt.plot(jsIDL[irms,:],ajsIDLmodal[irms,:],linewidth=2,label='IDL modal retrieved')\n plt.plot(jsIDL[irms,:],ajsIDLzonal[irms,:],linewidth=2,label='IDL zonal retrieved')\n plt.xlabel('j')\n plt.ylabel('aj [nm]')\n plt.xlim([jsTrue[irms,0],jsTrue[irms,-1]])\n plt.legend(loc='best')\n plt.grid()\n plt.savefig(filename % (rmsWFe,'.png'), dpi=300)\n plt.savefig(filename % (rmsWFe,'.pdf'), dpi=300)","repo_name":"jordanvoirin/PdM","sub_path":"Python/phaseDiversity/comparisonWthDiversity.py","file_name":"comparisonWthDiversity.py","file_ext":"py","file_size_in_byte":5842,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"7734609883","text":"import argparse\nimport sys\n\n# import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport torch\n# import transformersm\nfrom fairseq.data import Dictionary\nfrom fairseq.data.encoders.fastbpe import fastBPE\n# from keras.preprocessing.sequence import pad_sequences\nfrom keras_preprocessing.sequence import pad_sequences\nfrom sklearn.metrics import f1_score\nfrom sklearn.model_selection import train_test_split\nfrom torch import nn\nfrom torch.nn import CrossEntropyLoss\nfrom torch.utils.data import (DataLoader, RandomSampler, SequentialSampler,\n TensorDataset)\nfrom tqdm.notebook import trange\nfrom transformers import *\nfrom transformers import (AdamW, RobertaConfig, RobertaForTokenClassification,\n get_linear_schedule_with_warmup)\n\n\nclass Ner(BertPreTrainedModel):\n config_class = RobertaConfig\n pretrained_model_archive_map = ROBERTA_PRETRAINED_MODEL_ARCHIVE_MAP\n base_model_prefix = \"roberta\"\n\n def __init__(self, config):\n super().__init__(config)\n self.num_labels = config.num_labels\n self.roberta = RobertaModel(config)\n self.dropout = nn.Dropout(config.hidden_dropout_prob)\n self.classifier = nn.Linear(\n config.hidden_size*4, config.num_labels)\n self.init_weights()\n\n def forward(\n self,\n input_ids=None,\n attention_mask=None,\n token_type_ids=None,\n position_ids=None,\n head_mask=None,\n inputs_embeds=None,\n labels=None,\n ):\n\n outputs = self.roberta(\n input_ids,\n attention_mask=attention_mask,\n token_type_ids=token_type_ids,\n position_ids=position_ids,\n head_mask=head_mask,\n inputs_embeds=inputs_embeds,\n )\n\n output = torch.cat(\n (outputs[2][-1], outputs[2][-2], outputs[2][-3], outputs[2][-4]), dim=-1)\n sequence_output = self.dropout(output)\n logits = self.classifier(sequence_output)\n outputs = logits\n\n if labels is not None:\n loss_fct = CrossEntropyLoss()\n active_loss = attention_mask.view(-1) == 1\n active_logits = logits.view(-1, self.num_labels)\n active_labels = torch.where(\n active_loss, labels.view(-1), torch.tensor(\n loss_fct.ignore_index).type_as(labels)\n )\n loss = loss_fct(active_logits, active_labels)\n outputs = (loss, logits)\n return outputs\n\n\nmodel2 = torch.load(\"model/best_ner_syllable.pt\", map_location='cpu')\n\ndf_test = []\ni = 1\nfor line in open(\"data/test_word.conll\", \"r\").readlines():\n if len(line.split()) < 2:\n i += 1\n else:\n tmp = line.split()\n tmp.append('s'+str(i))\n df_test.append(tmp)\n\ndata_test = pd.DataFrame(df_test, columns=['Word', 'Tag', 'Sentence#'])\n\n\ndef concatWord(data):\n def tuple_func(f): return [(w, t)\n for w, t in zip(f['Word'].values, f['Tag'].values)]\n sentences_with_tag = data.groupby('Sentence#').apply(tuple_func)\n # print(sentences_with_tag)\n sentences_with_tag = [sent for sent in sentences_with_tag]\n return sentences_with_tag\n\n\nsentences_with_tag_test = concatWord(data_test)\n\nsentences_test = [' '.join([word[0] for word in sent])\n for sent in sentences_with_tag_test]\nlabels_test = [[word[1] for word in sent] for sent in sentences_with_tag_test]\n\n\ndf_train = []\ni = 1\nfor line in open(\"data/train_syllable.conll\", \"r\").readlines():\n if len(line.split()) < 2:\n i += 1\n else:\n tmp = line.split()\n tmp.append('s'+str(i))\n df_train.append(tmp)\ndata_train = pd.DataFrame(df_train, columns=['Word', 'Tag', 'Sentence#'])\nlabel2idx = {k: v for v, k in enumerate(data_train.Tag.unique())}\nids_to_labels = {v: k for v, k in enumerate(data_train.Tag.unique())}\nlabel2idx['PAD'] = 20\nlabel2idx['[CLS]'] = 21\nlabel2idx['[SEP]'] = 22\nlabel2idx['X'] = 23\nids_to_labels[20] = 'PAD'\nids_to_labels[21] = '[CLS]'\nids_to_labels[22] = '[SEP]'\nids_to_labels[23] = 'X'\n\nlabels_value = ['PAD', '[CLS]', '[SEP]', 'X']+data_train.Tag.unique().tolist()\nparser = argparse.ArgumentParser()\nparser.add_argument('--bpe-codes',\n default=\"phobert-base/bpe.codes\",\n required=False,\n type=str,\n help='path to fastBPE BPE'\n )\nargs, unknown = parser.parse_known_args()\n\nbpe = fastBPE(args)\nvocab = Dictionary()\nvocab.add_from_file(\"phobert-base/vocab.txt\")\n\n\ndef text2output(text, vocab):\n sentences_test\n subwords_test = [' '+bpe.encode(text)+' ']\n input_ids_test = pad_sequences([vocab.encode_line(sent, append_eos=False, add_if_not_exist=False).long().tolist() for sent in subwords_test],\n truncating='post', padding='post', maxlen=90, value=1.0, dtype='long')\n attenion_mask_test = [[float(val != 1) for val in sent]\n for sent in input_ids_test]\n\n X_test = input_ids_test\n test_mask = attenion_mask_test\n\n X_test = torch.tensor(X_test)\n test_mask = torch.tensor(test_mask)\n pred_labels_ids = []\n with torch.no_grad():\n output_test = model2(X_test[0:1], test_mask[0:1])\n\n logit = output_test.detach().cpu().numpy()\n pred_labels_ids.extend([list(pred_label)\n for pred_label in np.argmax(logit, axis=2)])\n pred = [ids_to_labels[pred_labels_ids[j][i]] for j in range(\n len(pred_labels_ids)) for i in range(len(pred_labels_ids[j]))]\n return list(zip(subwords_test[0].split(), pred))\n # return subwords_test[0].split(), pred\n\nbang_nguyen_am = [['a', 'à', 'á', 'ả', 'ã', 'ạ', 'a'],\n ['ă', 'ằ', 'ắ', 'ẳ', 'ẵ', 'ặ', 'aw'],\n ['â', 'ầ', 'ấ', 'ẩ', 'ẫ', 'ậ', 'aa'],\n ['e', 'è', 'é', 'ẻ', 'ẽ', 'ẹ', 'e'],\n ['ê', 'ề', 'ế', 'ể', 'ễ', 'ệ', 'ee'],\n ['i', 'ì', 'í', 'ỉ', 'ĩ', 'ị', 'i'],\n ['o', 'ò', 'ó', 'ỏ', 'õ', 'ọ', 'o'],\n ['ô', 'ồ', 'ố', 'ổ', 'ỗ', 'ộ', 'oo'],\n ['ơ', 'ờ', 'ớ', 'ở', 'ỡ', 'ợ', 'ow'],\n ['u', 'ù', 'ú', 'ủ', 'ũ', 'ụ', 'u'],\n ['ư', 'ừ', 'ứ', 'ử', 'ữ', 'ự', 'uw'],\n ['y', 'ỳ', 'ý', 'ỷ', 'ỹ', 'ỵ', 'y']]\nbang_ky_tu_dau = ['', 'f', 's', 'r', 'x', 'j']\n\nnguyen_am_to_ids = {}\n\nfor i in range(len(bang_nguyen_am)):\n for j in range(len(bang_nguyen_am[i]) - 1):\n nguyen_am_to_ids[bang_nguyen_am[i][j]] = (i, j)\n\n\ndef chuan_hoa_dau_tu_tieng_viet(word):\n if not is_valid_vietnam_word(word):\n return word\n\n chars = list(word)\n dau_cau = 0\n nguyen_am_index = []\n qu_or_gi = False\n for index, char in enumerate(chars):\n x, y = nguyen_am_to_ids.get(char, (-1, -1))\n if x == -1:\n continue\n elif x == 9: # check qu\n if index != 0 and chars[index - 1] == 'q':\n chars[index] = 'u'\n qu_or_gi = True\n elif x == 5: # check gi\n if index != 0 and chars[index - 1] == 'g':\n chars[index] = 'i'\n qu_or_gi = True\n if y != 0:\n dau_cau = y\n chars[index] = bang_nguyen_am[x][0]\n if not qu_or_gi or index != 1:\n nguyen_am_index.append(index)\n if len(nguyen_am_index) < 2:\n if qu_or_gi:\n if len(chars) == 2:\n x, y = nguyen_am_to_ids.get(chars[1])\n chars[1] = bang_nguyen_am[x][dau_cau]\n else:\n x, y = nguyen_am_to_ids.get(chars[2], (-1, -1))\n if x != -1:\n chars[2] = bang_nguyen_am[x][dau_cau]\n else:\n chars[1] = bang_nguyen_am[5][dau_cau] if chars[1] == 'i' else bang_nguyen_am[9][dau_cau]\n return ''.join(chars)\n return word\n\n for index in nguyen_am_index:\n x, y = nguyen_am_to_ids[chars[index]]\n if x == 4 or x == 8: # ê, ơ\n chars[index] = bang_nguyen_am[x][dau_cau]\n # for index2 in nguyen_am_index:\n # if index2 != index:\n # x, y = nguyen_am_to_ids[chars[index]]\n # chars[index2] = bang_nguyen_am[x][0]\n return ''.join(chars)\n\n if len(nguyen_am_index) == 2:\n if nguyen_am_index[-1] == len(chars) - 1:\n x, y = nguyen_am_to_ids[chars[nguyen_am_index[0]]]\n chars[nguyen_am_index[0]] = bang_nguyen_am[x][dau_cau]\n # x, y = nguyen_am_to_ids[chars[nguyen_am_index[1]]]\n # chars[nguyen_am_index[1]] = bang_nguyen_am[x][0]\n else:\n # x, y = nguyen_am_to_ids[chars[nguyen_am_index[0]]]\n # chars[nguyen_am_index[0]] = bang_nguyen_am[x][0]\n x, y = nguyen_am_to_ids[chars[nguyen_am_index[1]]]\n chars[nguyen_am_index[1]] = bang_nguyen_am[x][dau_cau]\n else:\n # x, y = nguyen_am_to_ids[chars[nguyen_am_index[0]]]\n # chars[nguyen_am_index[0]] = bang_nguyen_am[x][0]\n x, y = nguyen_am_to_ids[chars[nguyen_am_index[1]]]\n chars[nguyen_am_index[1]] = bang_nguyen_am[x][dau_cau]\n # x, y = nguyen_am_to_ids[chars[nguyen_am_index[2]]]\n # chars[nguyen_am_index[2]] = bang_nguyen_am[x][0]\n return ''.join(chars)\n\n\ndef is_valid_vietnam_word(word):\n chars = list(word)\n nguyen_am_index = -1\n for index, char in enumerate(chars):\n x, y = nguyen_am_to_ids.get(char, (-1, -1))\n if x != -1:\n if nguyen_am_index == -1:\n nguyen_am_index = index\n else:\n if index - nguyen_am_index != 1:\n return False\n nguyen_am_index = index\n return True\n\n\ndef chuan_hoa_dau_cau_tieng_viet(sentence):\n \"\"\"\n Chuyển câu tiếng việt về chuẩn gõ dấu kiểu cũ.\n :param sentence:\n :return:\n \"\"\"\n sentence = sentence.lower()\n words = sentence.split()\n for index, word in enumerate(words):\n cw = re.sub(r'(^\\p{P}*)([p{L}.]*\\p{L}+)(\\p{P}*$)',\n r'\\1/\\2/\\3', word).split('/')\n # print(cw)\n if len(cw) == 3:\n cw[1] = chuan_hoa_dau_tu_tieng_viet(cw[1])\n words[index] = ''.join(cw)\n return ' '.join(words)\n\n\ndef chuan_hoa_chuam_cau(text):\n tmp = []\n for t in text.split():\n if t[-1] in ',.:!?;*#$%^*()!~':\n t = t[:-1]+' '+t[-1]\n tmp.append(t)\n return ' '.join(tmp)\n\nimport unicodedata\ntext = chuan_hoa_dau_tu_tieng_viet(sys.argv[1])\ntext = chuan_hoa_chuam_cau(text)\ntext = unicodedata.normalize('NFC', text)\nfinal_result = text2output(text, vocab)\n# print(sys.argv[1])\n# print(text)\n# print(final_result)\nf = open(\"data/first_text.txt\", \"w\")\nf.write(sys.argv[1])\nf = open(\"data/final_result.txt\", \"w\")\nf.write(' '.join([str(item) for item in final_result]))\nf.close()\n\n","repo_name":"AndrewBui-AI-CS/NLP_CovidNer","sub_path":"api/helper.py","file_name":"helper.py","file_ext":"py","file_size_in_byte":11022,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26376759162","text":"\"\"\"\r\n\"Assignment 4 - SK Learn Introduction\" - Part 1\r\nBoston Dataset\r\n\"\"\"\r\n\r\n\r\nfrom sklearn.linear_model import LinearRegression\r\nfrom sklearn.datasets import load_boston\r\nimport pandas as pd\r\nimport seaborn as sns\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\ndef plot_graph(boston_get_data, feature_name, target_name):\r\n\r\n boston_plot_data_df = pd.DataFrame(\r\n boston_get_data.data, columns=boston_get_data.feature_names\r\n )\r\n boston_plot_data_df[target_name] = boston_get_data.target\r\n sns.pairplot(\r\n boston_plot_data_df,\r\n x_vars=[feature_name],\r\n y_vars=target_name,\r\n height=7,\r\n aspect=0.7,\r\n kind=\"reg\",\r\n )\r\n plt.title(\"Slope of the factor with the largest effect\")\r\n plt.show()\r\n\r\n\r\ndef make_regression_boston_data_and_plot():\r\n\r\n boston_get_data = load_boston()\r\n x_boston_dataframe = pd.DataFrame(\r\n boston_get_data.data, columns=boston_get_data.feature_names\r\n )\r\n y_boston_dataframe = pd.DataFrame(boston_get_data.target, columns=[\"MEDV\"])\r\n linear_regr = LinearRegression()\r\n linear_regr.fit(x_boston_dataframe, y_boston_dataframe)\r\n coeff_df = pd.DataFrame(\r\n linear_regr.coef_.T, x_boston_dataframe.columns, columns=[\"Coefficient\"]\r\n )\r\n print(coeff_df)\r\n max_coeff_index = coeff_df.abs()[\"Coefficient\"].idxmax()\r\n max_abs_coeff_value = coeff_df.abs().loc[max_coeff_index, \"Coefficient\"]\r\n print(\r\n \"The factor which has the largest effect on the price of housing in Boston is:\"\r\n + max_coeff_index\r\n )\r\n print(\r\n \"The absolute value of the coefficient of the factor which has the largest effect is: {}\".format(\r\n max_abs_coeff_value\r\n )\r\n )\r\n plot_graph(boston_get_data, max_coeff_index, \"MEDV\")\r\n\r\n\r\nif __name__ == \"__main__\":\r\n\r\n make_regression_boston_data_and_plot()\r\n\r\n\r\n\"\"\"\r\n\r\nComment\r\n\r\nIn order to solve the first part of \"Assignment 4 - SK Learn\r\nIntroduction\" (the part regarding the Boston dataset) first I\r\nimport \"LinearRegression\" from \"sklearn.linear_model\", \"load_boston\"\r\nfrom \"sklearn.datasets\", \"pandas\" as \"pd\", \"seaborn\" as \"sns\" and\r\n\"matplotlib.pyplot\" as \"plt\".\r\nThen I define the first function, named \"plot_graph\", and I include\r\nin the brackets \"boston_get_data\", \"feature_name\" and \"target_name\".\r\nMoreover, I set \"boston_plot_data_df\" equal to \"pd.DataFrame()\" and I\r\nalso include in the brackets \"boston_get_data.data\" and I set\r\n\"columns\" equal to \"boston_get_data.feature_names\". In addition to\r\nthis, I also set \"boston_plot_data_df[target_name]\" equal to\r\n\"boston_get_data.target\".\r\nFurthermore, I also include \"sns.pairplot\" and inside the brackets\r\nI include \"boston_plot_data_df\", \"x_vars=[feature_name]\",\r\n\"y_vars=target_name\" and I also set \"height\" equal to \"7\", \"aspect\"\r\nequal to \"0.7\" and \"kind\" equal to \"reg\".\r\nMoreover, I also include \"plt.title()\" and inside the brackets I also\r\ninclude the title of the plot, which is \"Slope of the factor with the\r\nlargest effect\". Lastly, I also include \"plt.show()\".\r\nFurthermore, I also define \"make_regression_boston_data_and_plot()\". To\r\nthis regard, first I set \"boston_get_data\" equal to \"load_boston()\" in\r\norder to take into consideration the data of the Boston dataset. Moreover,\r\nI set \"x_boston_dataframe\" equal to \"pd.DataFrame()\" and inside the\r\nbrackets I include \"boston_get_data.data\" and I also set \"columns\" equal\r\nto \"boston_get_data.feature_names\". In addition to this, I also set\r\n\"y_boston_dataframe\" equal to \"pd.DataFrame()\" and inside the brackets I\r\ninclude \"boston_get_data.target\" and I also set \"columns\" equal to\r\n\"[\"MEDV\"]\". I then set \"linear_regr\" equal to \"LinearRegression()\" and I\r\nalso include \"linear_regr.fit()\" and inside the brackets I include\r\n\"x_boston_dataframe\" and \"y_boston_dataframe\".\r\nMoreover, I set \"coeff_df\" equal to \"pd.DataFrame()\" and inside the brackets\r\nI include \"linear_regr.coef_.T\", \"x_boston_dataframe.columns\" and I also\r\nset \"columns\" equal to \"[\"Coefficient\"]\". I then use \"print()\" to print\r\nthe coefficients.\r\nFurthermore, I set \"max_coeff_index\" equal to\r\n\"coeff_df.abs()[\"Coefficient\"].idxmax()\" and also \"max_abs_coeff_value\"\r\nequal to \"coeff_df.abs().loc[max_coeff_index, \"Coefficient\"]\".\r\nMoreover, I use \"print()\" to print the factor which has the largest effect\r\non the price of housing in Boston and then also to print the absolute\r\nvalue of the coefficient of the factor which has the largest effect on the\r\nprice of housing in Boston. I then also include \"plot_graph()\" and inside\r\nthe brackets I include \"boston_get_data\", \"max_coeff_index\" and \"\"MEDV\"\".\r\nFinally, I also include \"if __name__ == \"__main__\":\" and\r\n\"make_regression_boston_data_and_plot()\".\r\nLastly, I type \"black\" in the \"Terminal\" followed by the path of the file\r\nin \".py format\" in order to format the whole code contained in the file\r\ntaken into consideration (basically the code of the file you are reading\r\nand so the code of the file in \".py format\" named\r\n\"Assignment.4.SK.Learn.Introduction.Part.1\").\r\n\r\n\r\nFrom the results obtained by running this code it is possible to state\r\nthat the factor that has the largest effect on the price of housing in\r\nBoston is \"NOX\" and also that the absolute value of its coefficient is\r\n17.766611228299986.\r\n\r\n\"\"\"\r\n","repo_name":"federico-mecchia/Assignment.4.SK.Learn.Introduction","sub_path":"Assignment.4.SK.Learn.Introduction.Part.1.py","file_name":"Assignment.4.SK.Learn.Introduction.Part.1.py","file_ext":"py","file_size_in_byte":5273,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"17381423377","text":"def sol(arr):\r\n #negative no store karo\r\n neg=0\r\n largneg=-99999999999999\r\n flag=0\r\n for i in arr:\r\n if i <0:\r\n largneg=max(largneg,i)\r\n neg+=1\r\n if i==0:\r\n flag=1\r\n # print(largneg)\r\n if (neg)==0:\r\n return min(arr)\r\n\r\n if (neg)%2==0 and largneg!=-99999999999999:\r\n arr.remove(largneg)\r\n res=1\r\n for i in arr:\r\n if i ==0:\r\n continue\r\n res*=i\r\n return res\r\n\r\nif __name__ == '__main__':\r\n arr=[ 2,5,-1,-2,-3,-4]\r\n # arr=[1,2,3,4,5,6,0]\r\n print(sol(arr))\r\n","repo_name":"Samundar9525/data_structure_using_Python","sub_path":"greedy/minimum product subaray.py","file_name":"minimum product subaray.py","file_ext":"py","file_size_in_byte":581,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"17934463164","text":"import pandas as pd\r\nfrom datetime import datetime\r\n\r\ndf = pd.read_csv('python hands-on - dataset.csv')\r\nprint(df)\r\n\r\ndef isObsolete(benchmark_date,date):\r\n #date format must be 'yyyy-mm-dd'\r\n try:\r\n bm_date = datetime.strptime(benchmark_date,'%Y-%m-%d')\r\n date_to_check = datetime.strptime(date,'%Y-%m-%d')\r\n if date_to_check < bm_date:\r\n return 'TRUE'\r\n else:\r\n return 'FALSE'\r\n except TypeError as TE:\r\n print('Date must be a string in the format \"yyyy-mm-dd\"')\r\n except Exception as e:\r\n print('Some Other Error Occured: '+str(e)) \r\n\r\n \r\nbenchmark_date = '2021-01-01'\r\n\r\ndf['obsolete'] = df['date'].apply(lambda x: isObsolete(benchmark_date,x))\r\nprint(df)\r\n\r\nexportname = 'python hands-on - dataset.json'\r\ntry:\r\n df.to_json(exportname,orient='columns')\r\n print('File saved to current directory as '+ exportname)\r\nexcept Exception as e:\r\n print('Error saving json file: '+str(e))","repo_name":"jo5hxxvii/Data2BotsPythonHandsOnTest","sub_path":"Data2Bots Python Hands on Test.py","file_name":"Data2Bots Python Hands on Test.py","file_ext":"py","file_size_in_byte":977,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70492683422","text":"from functools import reduce\nfrom .util import *\n\nclass Subleq:\n NEXT = object()\n AGAIN = object()\n\n @staticmethod\n def to_string(a):\n if a == Subleq.NEXT:\n return \"NEXT\"\n elif a == Subleq.AGAIN:\n return \"AGAIN\"\n elif a == None:\n return \"NONE\"\n else:\n return a\n\nclass Program:\n def __init__(self):\n self.memory = []\n self.labels = dict()\n self.idx = None\n\n def index(self):\n return self.idx or len(self.memory)\n\n def seek(self, index):\n self.idx = index\n\n def set(self, data, line):\n if self.idx == len(self.memory):\n self.idx = None\n\n if self.idx != None:\n if self.idx < len(self.memory):\n if self.memory[self.idx] == None:\n self.memory[self.idx] = data\n self.idx += 1\n else:\n Error(\"trying to overwrite already written address 0x{:04x} in line {}\".format(self.idx, self.line))\n elif data != None:\n while self.idx >= len(self.memory):\n self.memory.append(None)\n self.memory[self.idx] = data\n self.idx += 1\n\n else:\n self.memory.append(data)\n\nclass Token:\n def is_seek(self, token):\n return token[0] == \"@\"\n\n def is_label(self, token):\n return token[-1] == \":\"\n\n def is_number(self, token):\n return token[0] in \"-0123456789\" or token == \"NONE\"\n\n def is_symbol(self, token):\n return not (self.is_seek(token) or self.is_label(token) or self.is_number(token))\n\nclass Assembler:\n def __init__(self, token = None, program = None, **kwargs):\n if program == None:\n program = Program()\n if token == None:\n token = Token()\n\n self.program = program\n self.token = token\n\n def next_token(self):\n return self.tokens.pop(0)\n\n def parse_seek(self, line, token):\n token = token[1:]\n try:\n token = parse_int(token)\n except:\n Error(\"could not parse {} as integer in line {}\".format(token, line))\n self.program.seek(token)\n\n def parse_label(self, line, token):\n token = token[:-1]\n if not self.token.is_symbol(token) or token == \"NEXT\" or token == \"AGAIN\":\n Error(\"definition of reserved label {} in line {}\".format(token, line))\n if token in self.program.labels:\n Error(\"redefinition of label {} in line {}\".format(token, line))\n self.program.labels[token] = self.program.index()\n\n def parse_number(self, line, token):\n if token == \"NONE\":\n token = None\n else:\n try:\n token = parse_int(token)\n except:\n Error(\"could not parse {} as integer in line {}\".format(token, line))\n\n self.program.set(token, line)\n\n def parse_symbol(self, line, token):\n if token == \"NEXT\":\n token = Subleq.NEXT\n elif token == \"AGAIN\":\n token = Subleq.AGAIN\n self.program.set(token, line)\n\n def parse_token(self, line, token):\n if self.token.is_seek(token):\n self.parse_seek(line, token)\n elif self.token.is_label(token):\n self.parse_label(line, token)\n elif self.token.is_number(token):\n self.parse_number(line, token)\n else:\n self.parse_symbol(line, token)\n\n def parse_finish(self):\n pass\n\n def write_start(self):\n pass\n\n def write_word(self, word):\n pass\n\n def write_finish(self):\n pass\n\n def write_output(self):\n self.write_start()\n\n for i, word in enumerate(self.program.memory):\n if word == None:\n word = 0\n elif type(word) == str:\n try:\n word = self.program.labels[word]\n except:\n Error(\"undefined label {} at address 0x{:04x}\".format(word, i))\n elif word == Subleq.NEXT:\n word = i + 1\n elif word == Subleq.AGAIN:\n word = i - 2\n word = neg2twos(word)\n\n self.write_word(word)\n\n self.write_finish()\n\n def assemble(self):\n data = self.read_input()\n\n lines = enumerate(data.split('\\n'), 1)\n tokenlists = map(lambda e: list(map(lambda tok: (e[0], tok), e[1].split())), lines)\n\n self.tokens = reduce(lambda a, b: a + b, tokenlists, [])\n\n while len(self.tokens) != 0:\n line, token = self.next_token()\n self.parse_token(line, token)\n\n self.parse_finish()\n self.write_output()\n","repo_name":"Ferdi265/hl-subleq-tools","sub_path":"hlsubleq/asm.py","file_name":"asm.py","file_ext":"py","file_size_in_byte":4707,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"26450639962","text":"from flask_wtf import FlaskForm\nfrom wtforms import StringField, SubmitField, PasswordField, BooleanField\nfrom wtforms.validators import Required, Length, Email, EqualTo, ValidationError\nfrom application.models import Users, Posts\nfrom flask_login import current_user\n\ndef Unique_title(): #attempt to colaspe to singel unique function message = 'value must be unique'\n def _Unique_title(form, feild):\n if str(Posts.query.filter_by(title = feild.data).all()) != '[]':\n raise ValidationError(\"Value entered not unique.\")\n print('failed validation test for uniqueness')\n return _Unique_title\n\n\ndef Unique_content():\n message = 'value must be unique'\n def _Unique_content(form, feild):\n if str(Posts.query.filter_by(content = feild.data).all()) != '[]':\n raise ValidationError(\"Value entered not unique.\")\n print('failed validation test for uniqueness')\n return _Unique_content\n\n\n\n\n\n\n\n\n\nclass PostForm(FlaskForm):\n title = StringField('Title',\n validators = [\n Required(),\n Unique_title(),\n Length(min=2, max=100)\n ]\n )\n content = StringField('Content',\n validators = [\n Required(),\n Unique_content(),\n Length(min=2, max=1000)\n ]\n )\n submit = SubmitField('Post!') \n\nclass RegistrationForm(FlaskForm):\n\n first_name = StringField('First Name',\n validators = [\n Required(),\n Length(min=2, max=30)\n ]\n )\n last_name = StringField('last Name',\n validators = [\n Required(),\n Length(min=2, max=30)\n ]\n )\n\n email = StringField('Email',\n validators = [\n Required(),\n Email()\n ]\n )\n password = PasswordField('Password',\n validators = [\n Required(),\n ]\n )\n confirm_password = PasswordField('Confirm Password',\n validators = [\n Required(),\n EqualTo('password')\n ]\n )\n submit = SubmitField('Sign Up')\n\n def validate_email(self, email):\n user = Users.query.filter_by(email=email.data).first()\n\n if user:\n raise ValidationError('Email already in use')\n\nclass LoginForm(FlaskForm):\n email = StringField('Email',\n validators=[\n Required(),\n Email()\n ]\n )\n\n password = PasswordField('Password',\n validators=[\n Required()\n ]\n )\n\n remember = BooleanField('Remember Me')\n submit = SubmitField('Login')\n\nclass UpdateAccountForm(FlaskForm):\n first_name = StringField('First Name',\n validators=[\n Required(),\n Length(min=4, max=30)\n ])\n last_name = StringField('Last Name',\n validators=[\n Required(),\n Length(min=4, max=30)\n ])\n email = StringField('Email',\n validators=[\n Required(),\n Email()\n ])\n submit = SubmitField('Update')\n\n def validate_email(self,email):\n if email.data != current_user.email:\n user = Users.query.filter_by(email=email.data).first()\n if user:\n raise ValidationError('Email already in use')\n\n\n\n\n\n\n\n\n\n\n\n\n","repo_name":"MattCrutchley/flask-app","sub_path":"application/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":3261,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"21224107204","text":"import requests\nfrom splinter import Browser\nfrom bs4 import BeautifulSoup\nimport time\n\ndef init_browser():\n executable_path = {\"executable_path\": \"/usr/local/bin/chromedriver\"}\n return Browser(\"chrome\", **executable_path, headless=False)\n\ndef scrape_all():\n browser = init_browser()\n # visit NASA News\n url = \"https://mars.nasa.gov/news/?page=0&per_page=40&order=publish_date+desc%2Ccreated_at+desc&search=&category=19%2C165%2C184%2C204&blank_scope=Latest\"\n browser.visit(url)\n\n time.sleep(2)\n html = browser.html\n soup = BeautifulSoup(html, 'lxml')\n results = soup.find('div', class_=\"list_text\")\n news_title=results.a.text\n news_p=results.find(\"div\",class_='article_teaser_body').text\n browser.quit()\n\n\n browser = init_browser()\n # visit twitter to get Mars weather\n weather_url=\"https://twitter.com/marswxreport?lang=en\"\n browser.visit(weather_url)\n\n time.sleep(1)\n\n response = requests.get(weather_url)\n soup = BeautifulSoup(response.text, 'html.parser')\n results=soup.find_all('div',class_=\"js-tweet-text-container\")\n all_tweet=[result.text for result in results]\n weather=[]\n for x in all_tweet:\n if \"InSight\" in x:\n data={}\n data=x\n weather.append(data)\n weather=weather[0]\n new_weather=weather.replace('\\n','')\n mars_weather=new_weather.rsplit('pic', 1)[0] \n browser.quit()\n\n browser = init_browser()\n # visit the space site to get the latest pic of Mars\n space_url='https://www.jpl.nasa.gov/spaceimages/?search=&category=Mars'\n browser.visit(space_url)\n \n time.sleep(1)\n\n html = browser.html\n soup = BeautifulSoup(html, 'html.parser')\n\n browser.click_link_by_partial_text(\"FULL IMAGE\")\n browser.find_by_css('a.fancybox-next').click() \n browser.click_link_by_partial_href('/spaceimages/detail')\n browser.find_by_css('img.main_image').click()\n featured_image_url=browser.url\n \n browser.quit()\n\n\n browser = init_browser()\n # Mars Hemispheres\n hem_url='https://astrogeology.usgs.gov/search/results?q=hemisphere+enhanced&k1=target&v1=Mars'\n browser.visit(hem_url)\n\n time.sleep(1)\n\n html=browser.html\n soup = BeautifulSoup(html, 'html.parser')\n img_url = []\n\n for each_products in range(0,4):\n products = {} \n browser.find_by_css(\"a.product-item h3\")[each_products].click() \n products[\"title\"] = browser.find_by_css(\"h2.title\").text \n button = browser.find_link_by_text(\"Sample\")\n products[\"img_url\"] = button[\"href\"]\n img_url.append(products)\n browser.back() \n\n browser.quit()\n\n\n # put all variables into dict \n all={\n \"news_title\":news_title,\n \"news_p\":news_p,\n \"mars_weather\":mars_weather,\n \"featured_image_url\":featured_image_url,\n \"first_img\":img_url[0]['img_url'],\n \"first_name\":img_url[0]['title'],\n \"second_img\":img_url[1]['img_url'],\n \"second_name\":img_url[1]['title'],\n \"third_img\":img_url[2]['img_url'],\n \"third_name\":img_url[2]['title'],\n \"fourth_img\":img_url[3]['img_url'],\n \"fourth_name\":img_url[3]['title']\n}\n\n return all\n\n\n\n\n","repo_name":"busy0312/W12_Web-scraping-challenge","sub_path":"Missions_to_Mars/scrape_mars.py","file_name":"scrape_mars.py","file_ext":"py","file_size_in_byte":3143,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"21787535778","text":"from bcad.binterpreter.rqq import *\nfrom bcad.binterpreter.offscreen_display import offscreenViewer3d\nfrom bcad.binterpreter.glfw_display import glfwViewer3d\n\nfrom bcad.binterpreter.events import EVEnum, EventProcessor, ee, ep\nfrom bcad.binterpreter.singleton import Singleton\n\nimport imgui\nfrom imgui.integrations.glfw import GlfwRenderer\n\nimport json\nimport time\n\nclass MainWindow():\n def __init__(self, gui=True, pipe=None, img=None):\n self.use_imgui = gui\n self.use_occt = not gui\n self.please_stop = False\n self.rqq = rqQueue()\n self.texture_updated = False\n self.prev_pos = [0, 0]\n self.offscreen_view_size = [0,0]\n\n if self.use_imgui:\n print(\"Creating IMGUI context\")\n imgui.create_context()\n else:\n print(\"IMGUI disabled\")\n # Create a windowed mode window and its OpenGL context\n\n if self.use_imgui:\n self.canva = glfwViewer3d()\n self.canva.set_pipe(pipe)\n self.canva.set_img(img)\n self.canva.init_driver()\n self.impl = GlfwRenderer(self.canva.window)\n self.objtree = None\n self.show_objtree = True\n self.show_views = True\n else:\n self.canva = offscreenViewer3d()\n self.canva.set_pipe(pipe)\n self.canva.set_img(img)\n\n def parse_reply(self):\n rp = self.canva.pipe.recv()\n jdata = json.loads(rp)\n if jdata['rp'] == replies[RP_IMAGE_DATA]:\n if (self.offscreen_view_size[0] == self.canva.view_size[0]) and (self.offscreen_view_size[1] == self.canva.view_size[1]):\n self.canva.set_image(self.canva.view_size[0], self.canva.view_size[1], data=self.canva.img.buf)\n else:\n self.canva.set_image_black()\n self.canva.reply_received()\n elif jdata['rp'] == replies[RP_ACK]:\n if (self.offscreen_view_size[0] == self.canva.view_size[0]) and (self.offscreen_view_size[1] == self.canva.view_size[1]):\n self.canva.set_image(self.canva.view_size[0], self.canva.view_size[1], data=self.canva.img.buf)\n else:\n self.canva.set_image_black()\n self.canva.reply_received()\n elif jdata['rp'] == replies[RP_ACK_SET_SIZE]:\n self.offscreen_view_size = [jdata['args'][0], jdata['args'][1]]\n if (self.offscreen_view_size[0] == self.canva.view_size[0]) and (self.offscreen_view_size[1] == self.canva.view_size[1]):\n self.canva.set_image(self.canva.view_size[0], self.canva.view_size[1], data=self.canva.img.buf)\n else:\n self.canva.set_image_black()\n self.canva.reply_received()\n elif jdata['rp'] == replies[RP_ACK_GET_OBJECT_TREE]:\n objtree = jdata['args']\n self.objtree = objtree\n self.canva.reply_received()\n elif jdata['rp'] == replies[RP_NOP]:\n self.canva.reply_received()\n\n def object_node(self, node):\n if node:\n if node['children']:\n if (imgui.tree_node(node['name'])):\n for c in node['children']:\n self.object_node(c)\n imgui.tree_pop()\n else:\n imgui.text(node['name'])\n\n def object_tree(self, first_frame, x, y, w, h):\n if first_frame:\n imgui.core.set_next_window_position(x, y, imgui.ALWAYS)\n imgui.begin(\"Objects\")\n\n if self.objtree:\n self.object_node(self.objtree)\n\n hovered = imgui.core.is_window_hovered()\n wh = imgui.core.get_window_size()\n\n imgui.end()\n return hovered, wh\n\n def views_list(self, first_frame, x, y, w, h):\n if first_frame:\n imgui.core.set_next_window_position(x, y, imgui.ALWAYS)\n imgui.begin(\"Views\")\n if imgui.button('Left'):\n self.rqq.rq_set_view('left')\n elif imgui.button('Right'):\n self.rqq.rq_set_view('right')\n elif imgui.button('Top'):\n self.rqq.rq_set_view('top')\n elif imgui.button('Bottom'):\n self.rqq.rq_set_view('bottom')\n elif imgui.button('Front'):\n self.rqq.rq_set_view('front')\n elif imgui.button('Rear'):\n self.rqq.rq_set_view('rear')\n elif imgui.button('Iso1'):\n self.rqq.rq_set_view('iso1')\n elif imgui.button('Iso2'):\n self.rqq.rq_set_view('iso2')\n elif imgui.button('Iso3'):\n self.rqq.rq_set_view('iso3')\n elif imgui.button('Iso4'):\n self.rqq.rq_set_view('iso4')\n elif imgui.button('Iso5'):\n self.rqq.rq_set_view('iso5')\n elif imgui.button('Iso6'):\n self.rqq.rq_set_view('iso6')\n elif imgui.button('Iso7'):\n self.rqq.rq_set_view('iso7')\n elif imgui.button('Iso8'):\n self.rqq.rq_set_view('iso8')\n hovered = imgui.core.is_window_hovered()\n wh = imgui.core.get_window_size()\n\n imgui.end()\n return hovered, wh\n\n def mainloop(self):\n if self.use_imgui:\n self.canva.init_shader()\n self.canva.create_objects()\n \n self.rqq.rq_set_size(self.canva.view_size[0], self.canva.view_size[1])\n print(\"Waiting set size reply\")\n self.rqq.process(self.canva)\n self.parse_reply()\n \n self.rqq.rq_get_object_tree()\n print(\"Waiting get object tree reply\")\n self.rqq.process(self.canva)\n self.parse_reply()\n \n self.rqq.rq_load_image()\n last = time.time()\n first_frame = True\n while (not self.canva.should_close() and (not self.please_stop)):\n current = time.time()\n menu_bar_w_h = (0,0)\n objtree_w_h = (0,0)\n views_w_h = (0,0)\n self.canva.proc()\n if self.canva.get_need_resize():\n self.canva.set_image_black()\n self.rqq.rq_set_size(self.canva.view_size[0], self.canva.view_size[1])\n self.canva.start_frame()\n self.canva.swap_buffers()\n self.canva.poll_events()\n continue\n qlen = self.rqq.process(self.canva)\n if (current-last)>1:\n last = current\n self.rqq.rq_check_redraw()\n if self.canva.pipe.poll() == True:\n self.parse_reply()\n self.impl.process_inputs()\n imgui.new_frame()\n\n if imgui.begin_main_menu_bar():\n if imgui.begin_menu(\"File\", True):\n clicked_quit, selected_quit = imgui.menu_item(\"Quit\", 'Ctrl+Q', False, True)\n if clicked_quit:\n self.please_stop = True\n imgui.end_menu()\n if imgui.begin_menu(\"Render\", True):\n imgui.menu_item(\"Save\", None, False, True)\n imgui.end_menu()\n if imgui.begin_menu(\"View\", True):\n clicked_view_objtree, selected_view_objtree = imgui.menu_item(\"Show object tree\", None, False, True)\n if clicked_view_objtree:\n self.show_objtree = not(self.show_objtree)\n clicked_view_views, selected_view_views = imgui.menu_item(\"Show views\", None, False, True)\n if clicked_view_views:\n self.show_views = not(self.show_views)\n imgui.end_menu()\n wh = imgui.core.get_window_size()\n menu_bar_w_h = wh\n imgui.end_main_menu_bar()\n\n hovered = False\n if self.show_objtree:\n hovered, objtree_w_h = self.object_tree(first_frame, 0, menu_bar_w_h[1], 0, 0)\n\n if self.show_views:\n hovered, views_w_h = self.views_list(first_frame, objtree_w_h[0], menu_bar_w_h[1], 0, 0)\n\n if not hovered:\n # right button rotation\n if imgui.is_mouse_down(1):\n if self.canva.drag_start == None:\n pos = imgui.get_io().mouse_pos\n self.canva.drag_start = [pos[0], pos[1]]\n self.rqq.rq_start_rotation(self.canva.drag_start[0], self.canva.drag_start[1])\n else:\n pos = imgui.get_io().mouse_pos\n self.pt = [pos[0], pos[1]]\n self.rqq.rq_rotate(self.pt[0], self.pt[1])\n # left button panning\n elif imgui.is_mouse_down(2):\n if self.canva.drag_start == None:\n pos = imgui.get_io().mouse_pos\n self.canva.drag_start = [pos[0], pos[1]]\n self.prev_pos = pos\n else:\n pos = imgui.get_io().mouse_pos\n if not self.prev_pos == pos:\n self.rqq.rq_pan(pos[0]-self.prev_pos[0], pos[1]-self.prev_pos[1])\n self.prev_pos = pos\n # wheel button scrolling\n else:\n self.canva.drag_start = None\n mw = imgui.get_io().mouse_wheel\n pos = imgui.get_io().mouse_pos\n if mw != 0:\n self.rqq.rq_scroll(mw)\n else:\n if not self.prev_pos == pos:\n self.rqq.rq_move(pos[0], pos[1])\n self.prev_pos = pos\n\n first_frame = False\n self.canva.start_frame()\n imgui.render()\n draw_data = imgui.get_draw_data()\n self.impl.render(draw_data)\n self.canva.swap_buffers()\n self.canva.poll_events()\n self.canva.rq_stop()\n self.canva.pipe.close()\n print(\"GUI stopped\")\n else:\n while True:\n ep.process()\n if self.canva.pipe.poll() == True:\n rq = self.canva.pipe.recv()\n jdata = json.loads(rq)\n #print(\"rq:\", rq, \"jdata:\", type(jdata))\n if jdata['rq'] == requests[RQ_LOAD_IMAGE]:\n tex = self.canva.call_load_image()\n elif jdata['rq'] == requests[RQ_START_ROTATION]:\n self.canva.call_start_rotation(jdata['args'][0], jdata['args'][1])\n elif jdata['rq'] == requests[RQ_ROTATE]:\n self.canva.call_rotate(jdata['args'][0], jdata['args'][1])\n elif jdata['rq'] == requests[RQ_SET_SIZE]:\n self.canva.call_set_size(jdata['args'][0], jdata['args'][1])\n elif jdata['rq'] == requests[RQ_SCROLL]:\n self.canva.call_scroll(jdata['args'])\n elif jdata['rq'] == requests[RQ_MOVE]:\n self.canva.call_move(jdata['args'][0], jdata['args'][1])\n elif jdata['rq'] == requests[RQ_PAN]:\n self.canva.call_pan(jdata['args'][0], jdata['args'][1])\n elif jdata['rq'] == requests[RQ_CHECK_REDRAW]:\n self.canva.call_check_redraw()\n elif jdata['rq'] == requests[RQ_GET_OBJECT_TREE]:\n self.canva.call_get_object_tree()\n elif jdata['rq'] == requests[RQ_SET_VIEW]:\n self.canva.call_set_view(jdata['args'])\n elif jdata['rq'] == requests[RQ_STOP]:\n break\n time.sleep(0.01)\n self.canva.img.unlink()\n self.canva.pipe.close()\n print(\"OCCT stopped\")\n","repo_name":"snegovick/bcad","sub_path":"bcad/binterpreter/main_window.py","file_name":"main_window.py","file_ext":"py","file_size_in_byte":12202,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"73778174624","text":"from datetime import datetime\nfrom random import randint\nfrom flask import Flask, render_template\nfrom configparser import ConfigParser \n\ncp = ConfigParser()\ncp.read('config.cfg')\nlower_m_delta = cp.getint('general','lower_m_delta')\nupper_m_delta = cp.getint('general','upper_m_delta')\nport = cp.getint('general','port')\n\napp = Flask(__name__)\n\n@app.route('/')\ndef root():\n\n sides = ['to', 'past']\n d = datetime.now()\n min_delta = randint(lower_m_delta,upper_m_delta)\n\n side_idx = randint(0,1)\n side = sides[side_idx]\n\n hour = d.hour\n minute = d.minute\n\n if side_idx == 0:\n minute = minute + min_delta\n if minute > 59:\n hour = (hour + (minute / 60)) % 24\n minute = minute % 60\n else:\n minute = minute - min_delta\n if minute < 0:\n hour = (hour + (minute / 60)) % 24\n minute = minute % 60\n broken_time = \"The time is {} minute{} {} {:02d}:{:02d} hrs\".format(min_delta, 's' if min_delta > 1 else '', side, int(hour), int(minute))\n actual_time = \"The actual time is {} hrs\".format(d.strftime('%H:%m:%S'))\n return render_template('home.html', broken_time=broken_time,actual_time=actual_time)\n\napp.run('0.0.0.0', port)\n","repo_name":"desonalleyne/broken_timebox","sub_path":"root.py","file_name":"root.py","file_ext":"py","file_size_in_byte":1221,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"43208896170","text":"import argparse\nimport socket\nimport shlex\nimport subprocess\nimport sys\nimport textwrap\nimport threading\n\ndef execute(cmd):\n cmd = cmd.strip()\n if not cmd:\n return\n output = subprocess.check_output(shlex.split(cmd),stderr = subprocess.STDOUT)\n\n return output.decode()\n\nclass NetCat:\n def __init__(self, args, buffer=None):\n self.args = args\n self.buffer = buffer\n self.socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n\n def run(self):\n if self.args.listen:\n self.listen()\n else:\n self.send()\n \n def send(self):\n self.socket.connect((self.args.target, self.args.port))\n if self.buffer:\n self.socket.send(self.buffer)\n \n try:\n while True:\n recv_length = 1\n res = ''\n while recv_length:\n data = self.socket.recv(4096)\n recv_length = len(data)\n res += data.decode()\n if recv_length < 4096:\n break\n if res:\n print(res)\n buffer = input('> ')\n buffer += '\\n'\n self.socket.send(buffer.encode())\n except KeyboardInterrupt:\n print(\"User terminated!\")\n self.socket.close()\n sys.exit()\n\n def listen(self):\n self.socket.bind((self.args.target, self.args.port))\n self.socket.listen(5)\n\n while True:\n client_socket, _ = self.socket.accept()\n client_thread = threading.Thread(target=self.handle, args=(client_socket,))\n client_thread.start()\n \n def handle(self, client_socket):\n if self.args.execute:\n output = execute(self.args.execute)\n client_socket.send(output.encode())\n \n elif self.args.upload:\n file_buffer = b''\n while True:\n data = client_socket.recv(4096)\n if data:\n file_buffer += data\n else:\n break\n \n with open(self.args.upload, 'wb') as f:\n f.write(file_buffer)\n \n message = f'Saved file {self.args.uploads}'\n client_socket.send(message.encode())\n \n elif self.args.command:\n cmd_buffer = b''\n while True:\n try:\n client_socket.send(b'l3vi4th4n@send #> ')\n while '\\n' not in cmd_buffer.decode():\n cmd_buffer += client_socket.recv(64)\n \n res = execute(cmd_buffer.decode())\n if res:\n client_socket.send(res.encode())\n \n cmd_buffer = b''\n except Exception as e:\n print(f'server kill {e}')\n self.socket.close()\n sys.exit()\n \n\n\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='l3vi4th4n Net Tool', \n formatter_class=argparse.RawDescriptionHelpFormatter, \n epilog = textwrap.dedent('''[Example]:\n Netcat.py -t [HOST] -p [PORT] -l -c # command shell\n Netcat.py -t [HOST] -p [PORT] -l -u=[FILENAME] # upload the file\n Netcat.py -t [HOST] -p [PORT] -l -e [COMMAND] # execute specified command\n Netcat.py -t [HOST] -p [PORT] # connect the HOST\n echo 'ABC' | ./Netcat.py -t [HOST] -p [PORT] # send the 'ABC' to HOST PORT\n '''))\n \n parser.add_argument('-c', '--command', action='store_true', help='command shell')\n parser.add_argument('-e', '--execute', help='execute specified command')\n parser.add_argument('-l', '--listen', action='store_true', help='listen')\n parser.add_argument('-p', '--port', type=int, default=5555, help='spcified port')\n parser.add_argument('-t', '--target', help='specified host')\n parser.add_argument('-u', '--upload',help='upload file')\n args = parser.parse_args()\n\n if args.listen:\n buffer = ''\n else:\n buffer = sys.stdin.read()\n\n nc = NetCat(args, buffer.encode())\n nc.run()\n\n","repo_name":"k3lpi3b4nsh33/Python-Hacking","sub_path":"1. Basic Networking Tools/Netcat.py","file_name":"Netcat.py","file_ext":"py","file_size_in_byte":4331,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"10719732620","text":"import requests\nimport time\nimport downloader\nfrom auth import SPOTIFY_GET_CURRENT_TRACK_URL,SPOTIFY_GET_USER_QUEUE_URL\nfrom pprint import pprint\nfrom tkinter import *\n\nclass Spotify:\n def __init__(self):\n self.__url_gct = SPOTIFY_GET_CURRENT_TRACK_URL\n self.__url_gcq = SPOTIFY_GET_USER_QUEUE_URL\n self.__menu()\n\n def main_gct(self):\n current_track_info = self.get_current_track()\n n = current_track_info['track_name']\n downloader.Downloader(n).down()\n\n def main_gcq(self):\n\n current_track_info = self.get_current_queue()\n ls1 = list((current_track_info.values()))\n for i in ls1:\n downloader.Downloader(i).down()\n\n def get_current_queue(self):\n\n response = requests.get(\n self.__url_gcq,\n headers={\n \"Authorization\": f\"Bearer {self.__access_token_gcq}\"\n }\n )\n json_new = response.json()\n\n ls_id = []\n ls_name = []\n for i in json_new[\"queue\"]:\n if i[\"id\"] not in ls_id:\n ls_id.append(i[\"id\"])\n if i[\"name\"] not in ls_name:\n ls_name.append(i[\"name\"])\n dict_fn = dict(zip(ls_name, ls_id))\n return (dict_fn)\n\n def get_current_track(self):\n\n response = requests.get(\n self.__url_gct,\n headers={\n \"Authorization\": f\"Bearer {self.__access_token_gct}\"\n }\n )\n json_resp = response.json()\n\n track_id = json_resp['item']['id']\n track_name = json_resp['item']['name']\n artists = [artist for artist in json_resp['item']['artists']]\n link = json_resp['item']['external_urls']['spotify']\n artist_names = ', '.join([artist['name'] for artist in artists])\n\n current_track_info = {\n \"id\": track_id,\n \"track_name\": track_name,\n \"artists\": artist_names,\n \"link\": link\n }\n return current_track_info\n def __menu(self):\n self.root =Tk()\n self.root.geometry(\"1580x380\")\n\n filename = PhotoImage(file=\"light.png\")\n background_label = Label(self.root, image=filename)\n background_label.place(x=0, y=0, relwidth=1, relheight=1)\n\n img = PhotoImage(file=\"icon.png\")\n self.root.iconphoto(False, img)\n self.root.title(\"Spotify_Audio_Downloader\")\n\n self.e1 = Entry(self.root, width=40,borderwidth=10, bg=\"green\", fg=\"white\")\n self.e1.grid(row=5, column=0, columnspan=2)\n self.e2 = Entry(self.root, width=40,borderwidth=10, bg=\"green\", fg=\"white\")\n self.e2.grid(row=5, column=2, columnspan=2)\n\n self.myLabel1 = Label(self.root,text=\"Developed By Author\").grid(row=0,column=5)\n self.myLabel2 = Label(self.root,text=\"Date Created on 23/10/2022\").grid(row=1,column=5)\n self.myLabel2 = Label(self.root, text=\" \").grid(row=2, column=0)\n\n self.myButton1 = Button(self.root,text=\"Enter auth code below and press to download currently playing track\",borderwidth=5,command=self.click1,padx=20,pady=20).grid(row=3,column=0,columnspan=2)\n self.myButton2 = Button(self.root, text=\"Enter auth code below and press to download top 2 queued playlist tracks\",borderwidth=5,command=self.click2,padx=20,pady=20).grid(row=3,column=2,columnspan=2)\n self.myButton3 = Button(self.root, text=\"Press to exit the window\",borderwidth=5,command=self.root.quit,padx=20,pady=20).grid(row=3,column=4,columnspan=2)\n self.myButton4=Button(self.root, text=\"Press to clear text\",borderwidth=5,command=self.clear1,padx=20,pady=20).grid(row=7,column=0,columnspan=2)\n self.myButton5=Button(self.root, text=\"Press to clear text\",borderwidth=5,command=self.clear2,padx=20,pady=20).grid(row=7,column=2,columnspan=2)\n\n self.myLabel3 = Label(self.root, text=\" \").grid(row=4, column=0)\n self.myLabel0 = Label(self.root, text=\" \").grid(row=6, column=0)\n self.myLabel6 = Label(self.root, text=\" \").grid(row=8, column=0)\n\n self.root.mainloop()\n\n def click1(self):\n self.__access_token_gct=self.e1.get()\n self.main_gct()\n\n def click2(self):\n self.__access_token_gcq = self.e2.get()\n self.main_gcq()\n\n def clear1(self):\n self.e1.delete(0,END)\n\n def clear2(self):\n self.e2.delete(0,END)\n\n\nif __name__ == '__main__':\n Spotify()\n","repo_name":"hrshankar2002/Spotify-audio-scraper","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4487,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"28172304580","text":"import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom scipy.integrate import ode\nimport scipy.optimize\n\nimport seaborn as sns\n\nimport matplotlib\ncolors = matplotlib.rcParams['axes.prop_cycle'].by_key()['color']\nblack = matplotlib.rcParams['axes.labelcolor']\n\ntcellcolor = '#0E1E97'\ntcellcoloralt = '#0e7b97'\npmhccolor = colors[3]\ncolors = [tcellcolor, tcellcoloralt, pmhccolor]\n\nimport sys\nsys.path.append('..')\nimport plotting\nfrom lib import *\n \nplt.style.use('../paper.mplstyle')\n\ndef T7(T0, K):\n ts = [0.0, 3.5, 7.0]\n xs = odeint(fcompfull, [T0, C], ts, args=(alpha, mu, K, delta))\n return xs[-1, 0]\n\n# parameters as in Fig.3D (slowly increasing antigen availability)\nalpha = 1.2\nmu = -0.5\ndelta = 0.0\nC = 10\nT0s = np.logspace(0, 3, 50)\nKs = np.logspace(0, 3, 50)\n\n# alternative parameters as in Fig.1\n#alpha = 1.5\n#mu = 1.2\n#delta = 0.22\n#C = 10**6.7\n#T0s = np.logspace(0, 6.5, 50)\n#Ks = np.logspace(0, 6.5, 50)\n\n\nfoldexpansions = np.zeros((len(T0s), len(Ks)))\nfor i, T0 in enumerate(T0s):\n for j, K in enumerate(Ks):\n foldexpansions[i, j] = T7(T0, K)/T0\n\nfig, ax = plt.subplots(figsize=(2.75, 2.25))\nX, Y = np.meshgrid(Ks, T0s)\nplt.pcolor(X, Y, np.log10(foldexpansions), cmap='viridis')\ncbar = plt.colorbar()\ncbar.set_label('log$_{10}$ fold expansion')\nCS = ax.contour(X, Y, foldexpansions, colors='w', levels=[3, 10, 100, 400])\nax.set_xscale('log')\nax.set_yscale('log')\nplt.clabel(CS, CS.levels, inline=False, inline_spacing=2, fmt='%g')\nax.set_xlabel('$K$')\nax.set_ylabel('$T(0)$')\nfig.tight_layout()\nplt.show()\nfig.savefig('figS1.svg')\nfig.savefig('figS1.png', dpi=300)\n","repo_name":"andim/paper-tcellexp","sub_path":"figS1/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":1627,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"36520735809","text":"import pandas as pd\nimport requests as reql\nimport pickle\nfrom flask import Flask\nimport doc_processing\nfrom unclassified_grouping import find_group\nfrom document_similarity import *\nfrom data_extractor import *\n# from discovery_connector import call_classification\nfrom loguru import logger\nimport re\nimport os\nimport nltk\nimport configparser\nimport classifier_transactions\nfrom loguru import logger\nimport error_updation\nfrom error_updation import *\nimport datetime\nfrom datetime import timedelta\nimport os\nimport os.path\nimport shutil\nfrom dbprocess import dbprocess\nfrom build_models import unclassified_doc_update\nfrom build_models import classified_doc_update\nfrom elasticsearch import Elasticsearch\nfrom elasticsearch.connection import create_ssl_context\nimport ssl\nfrom elastic_update import update_group_es\n\n\nconfig = configparser.ConfigParser()\nconfig.read('config.ini')\napp = Flask(__name__)\n\n#logger.add(loginfo_filename, format=\"{time:YYYY-MM-DD at HH:mm:ss} | {level} | {message}\", backtrace=True,\n# level='DEBUG', rotation=\"1 week\", compression=\"zip\", enqueue=True)\n\npath_separator = config['CLASSIFIER']['PATH_SEP']\ntemp_directory = config['CLASSIFIER']['TEMP_DIR_LOC']\nclassification_model_accuracy = config['CLASSIFIER']['CLASSIFICATION_FINAL_ACCURACY']\nsimiliarity_accuracy = config['CLASSIFIER']['DOC_SIMILIARITY_ACCURACY']\nroot_dir = config['CLASSIFIER']['TRAINING_DATASET_LOCATION']\nextraction_loc = config['CLASSIFIER']['CLASSIFIED_METADATA']\nmax_file_size = int(config['CLASSIFIER']['SOURCE_FILE_SIZE_MAXBYTE'])\nmin_file_size= int(config['CLASSIFIER']['SOURCE_FILE_SIZE_MINBYTE'])\nmodel_location = config['CLASSIFIER']['MODEL_SAVE_LOCATION']\nclass_predicted_loc = config['CLASSIFIER']['CLASSIFIED_RESULT']\nclass_error_loc = config['CLASSIFIER']['CLASSIFIED_ERROR']\nsim_template_loc = config['CLASSIFIER']['SIM_TEMPLATE_GROUP']\n\nvector_model_pkl = config['CLASSIFIER']['VEC_MODEL_PKL_NAME']\nclass_model_pkl = config['CLASSIFIER']['CLASS_PKL_NAME']\nis_model_enabled = False\nlog_level = config['CLASSIFIER']['LOG_LEVEL']\nlog_location = config['CLASSIFIER']['LOGGER_LOC']\nlog_name = os.path.normpath(os.path.join(log_location, \"classifier_log.log\"))\nloginfo_filename = config['CLASSIFIER']['LOGINFO_FILENAME']\nlogdebug_filename = config['CLASSIFIER']['LOGDEBUG_FILENAME']\nerror_code=config['CLASSIFIER']['CLASSIFIER_ERROR_CODE']\nwrite_file_system = config['CLASSIFIER']['FILE_OUT']\nocr_store_full_text = int(config['CLASSIFIER']['OCR_STORE_FULL_TEXT'])\n\nmodel_path = os.path.normpath(os.path.join(model_location, class_model_pkl))\nvec_model_path = os.path.normpath(os.path.join(model_location, vector_model_pkl))\nes_host = config['CLASSIFIER']['ES_HOST']\nes_port = config['CLASSIFIER']['ES_PORT']\nindex = config['CLASSIFIER']['INDEX']\n\n\nssl_context = create_ssl_context()\nssl_context.check_hostname = False\nssl_context.verify_mode = ssl.CERT_NONE\nes = Elasticsearch([{'host': es_host, 'port': es_port}],scheme=\"https\",\n # to ensure that it does not use the default value `True`\n verify_certs=False,\n ssl_context= ssl_context,\n http_auth=(\"admin\", \"admin\"))\n\ndt = str(datetime.datetime.now()).replace(\":\",\"_\")\nnewname = 'classifier_loginfor.log'+'.zip'+dt+'.zip'\nnewname_debug = 'classifier_logdebug.log'+'.zip' +dt+ '.zip'\nif os.path.exists('classifier_loginfor.log' +'.zip'):\n os.rename('classifier_loginfor.log' +'.zip', newname)\n shutil.move(newname,log_location )\nif os.path.exists('classifier_logdebug.log' +'.zip'):\n os.rename('classifier_logdebug.log' +'.zip', newname_debug)\n shutil.move(newname_debug,log_location )\nlogger.add('classifier_loginfor.log' , format=\"{time} {message} | {level} | {message}\", backtrace=True, level='INFO', rotation=\"1 day\", enqueue=True, compression=\"zip\")\nlogger.add('classifier_logdebug.log' , format=\"{time} {message} | {level} | {message}\", backtrace=True, level='DEBUG', rotation=\"1 day\", enqueue=True, compression=\"zip\")\n\n\ndef remove_stop_words(text):\n import spacy\n from spacy.lang.en import English\n from spacy.lang.en.stop_words import STOP_WORDS\n nlp = English()\n # \"nlp\" Object is used to create documents with linguistic annotations.\n document = nlp(text)\n # Create list of word tokens\n token_list = []\n for token in document:\n token_list.append(token.text)\n\n # Create list of word tokens after removing stopwords\n filtered_sentence = []\n for word in token_list:\n lexeme = nlp.vocab[word]\n if lexeme.is_stop == False:\n filtered_sentence.append(word)\n return ' '.join(filtered_sentence)\n\ntemplates = []\nfor name in os.listdir(root_dir):\n templates.append(name)\nlogger.info(\"\\n\\n templates : {}\", templates)\n\n\nfrom sklearn import preprocessing\nle = preprocessing.LabelEncoder()\nle.fit(templates)\ntemplate_id = le.transform(le.classes_)\ncategorydf = pd.DataFrame({'template_id': template_id, 'templates': templates})\ncategorydf = categorydf.sort_values('template_id')\nid_to_category = dict(categorydf[['template_id', 'templates']].values)\nlogger.info(\" \\n\\n id_to_category : {}\", id_to_category)\n\n\nunclassified_dict = {}\nprediction_json = {}\n\n\ndef tokenize_stem(text):\n from nltk.stem.snowball import SnowballStemmer\n stemmer = SnowballStemmer(\"english\")\n logger.info(\"\\n\\n tokenize_stem : {}\")\n tokens = [word for sent in nltk.sent_tokenize(text) for word in nltk.word_tokenize(sent)]\n filtered_tokens = []\n # filter out any tokens not containing letters (e.g., numeric tokens, raw punctuation)\n for token in tokens:\n if re.search('[a-zA-Z]', token):\n filtered_tokens.append(token)\n stems = [stemmer.stem(t) for t in filtered_tokens]\n return stems\n\n\ndef file_validation(file_name,file_size):\n is_valid = False\n logger.info(\" file_validation : {}\", file_name)\n print(\" file_name :\",file_name)\n print(\" File Size :\",file_size)\n if os.path.exists(file_name) and os.access(file_name,os.R_OK) and (min_file_size <= file_size ) and (max_file_size >= file_size) :\n is_valid = True\n logger.info(\" file_validation : {}\",file_validation)\n print(\"file_validation ->\",is_valid)\n return is_valid\n\n\ndef get_file_name(file_path):\n filename = ''\n if len(file_path) > 1:\n filename = os.path.splitext(os.path.basename(file_path))[0]\n return str(filename)\n\n\ndef add_unclassified_docs(group_details,disc_inbound_id,org_filename,is_training,auth_key):\n unclassified_dict = {}\n try:\n if group_details['error_code'] == 0 and group_details['group_no'] is not None and int(group_details['group_no']) > 0 :\n prediction_json[\"group_no\"] = group_details['group_no']\n prediction_json[\"unclassified_file_name\"] = group_details['unclassified_file_name']\n unclassified_dict['DiscoveryInBoundId'] = disc_inbound_id\n unclassified_dict['ClassificationGroupId'] = int(group_details['group_no'])\n unclassified_dict['FileLocation'] = group_details['file_name']\n logger.info(\"\\n\\n\\n **************** class :group_details : {} \", group_details['file_name'])\n unclassified_dict['Name'] = org_filename\n unclassified_dict['DisplayName'] = org_filename\n unclassified_dict['IsTemplate'] = group_details['new_group']\n unclassified_dict['IsTrainingDocument'] = is_training\n #classifier_transactions.add_unclassified_doc_info(unclassified_dict,auth_key)\n else:\n prediction_json[\"error_msg\"] = group_details['error_msg']\n prediction_json[\"error_code\"] = group_details['error_code']\n #classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n except Exception as error:\n print(error)\n #classifier_transactions.update_inbound_status(disc_inbound_id)\n #error_updation.exception_log(error, \" Error occurred when adding unclassified details \", str(disc_inbound_id))\n\n\n@app.route('/'+config['CLASSIFIER']['CLASSIFIER_URL_NAME'], methods=['POST'])\ndef classifier():\n from flask import jsonify,request\n dfx_data = request.get_json(force=True)\n logger.info(\" classifier JSON: {}\",type(dfx_data))\n # filename = data['file_name']\n # disc_inbound_id = data['inbound_id']\n # is_training_source = data['IsTrainingSource']\n # is_unclassified = data['is_unclassified']\n details = classification(dfx_data)\n logger.info(\"\\n classifier :{}\", details)\n return jsonify(details)\n\n\n# def classification(dfx_data):\n# import gc\n# try:\n# prediction_json['error_code'] = 0\n# prediction_json[\"error_msg\"] = ''\n# filename = dfx_data['file_name']\n# prediction_json['disc_inbound_id'] = dfx_data['inbound_id']\n# auth_key = dfx_data['Authorization']\n# is_training_source = 0\n# if dfx_data['FileLength'] is not None and type(dfx_data['FileLength']) is not int:\n# file_size = int(dfx_data['FileLength'])\n# else:\n# file_size = dfx_data['FileLength']\n# disc_inbound_id = dfx_data['inbound_id']\n# if dfx_data['IsTrainingSource'] is not None and type(dfx_data['IsTrainingSource']) is not int:\n# is_training_source = int(dfx_data['IsTrainingSource'])\n# is_unclassified = dfx_data['is_unclassified']\n# data_processing = {}\n# if file_validation(filename,file_size) and int(disc_inbound_id) > 0 :\n# if int(is_unclassified) == 1:\n# data_processing['text_file_name'] = filename\n# data_processing['error_code'] = 0\n# else:\n# data_processing = doc_processing.filter_text_from_file(filename,disc_inbound_id,auth_key)\n# logger.info(\"\\n\\n\\n data_processing: {}\", data_processing)\n# if int(data_processing['error_code']) is not 2 and data_processing['text_file_name'] is not None and data_processing['text_file_name'].strip() is not '' and int(data_processing['error_code']) == 0 and len(data_processing['text_file_name'].strip()) > 1 :\n# txt_filename = data_processing['text_file_name']\n# text_extraction = ''\n# logger.info(\" \\n Classification: txt_filename {}\",txt_filename)\n# with open(txt_filename, 'rb') as text_file:\n# #text_extraction = text_file.read()\n# text_raw = ''+(text_file.read()).decode()\n# # logger.info(text_raw)\n# text_extraction = remove_stop_words(text_raw)\n# # logger.info(text_extraction)\n# if os.path.exists(vec_model_path) and os.path.exists(model_path):\n# global is_model_enabled\n# is_model_enabled = True\n# logger.info(\"\\n\\n is_model_enabled : {}\",is_model_enabled)\n# if len(filename.strip()) > 1 and text_extraction is not None and len(text_extraction.strip()) > 0 and is_model_enabled :\n# logger.info(\" Prediction Section : ******************** \\n\")\n# features_list=vectorizer_model.transform([text_extraction]) # .toarray()\n# prediction=classifier_model.predict(features_list)\n# prediction_id = int(prediction)\n# logger.debug(\"\\n\\n prediction : {}\", prediction_id)\n# proba_pred=classifier_model.predict_proba(features_list)\n# new_doc_classifier=proba_pred[:, prediction_id]\n# logger.info(\"\\n\\n new_doc_classifier: {}\", new_doc_classifier, float(classification_model_accuracy), float(new_doc_classifier))\n# if is_training_source == 0 and float(new_doc_classifier) >= float(classification_model_accuracy):\n# class_dict = {}\n# class_id = 0\n# logger.info(\"Prediction : {}\", prediction_id)\n# logger.info(\"Prediction Probability :{}\", proba_pred[:, prediction_id])\n# prediction_json[\"file_name\"] = filename\n# prediction_json[\"predition_type\"] = new_doc_classifier\n# class_name = id_to_category.get(int(prediction_id))\n# org_subclass_id = classifier_transactions.find_subclassification_id(class_name,auth_key)\n# logger.info(\"\\n\\n class_name : {}\", class_name)\n# prediction_json['predicted_category'] = class_name\n# class_dict[\"ClassificationTemplateId\"] = classifier_transactions.find_classification_id(org_subclass_id,auth_key)\n# class_dict[\"SubClassificationTemplateId\"] = org_subclass_id\n# class_dict[\"Type\"] = class_name\n# class_dict[\"DiscoveryInBoundId\"] = int(disc_inbound_id)\n# # predicted_dir_name = os.path.normpath(os.path.join(class_predicted_loc,class_name))\n# # if not os.path.exists(predicted_dir_name):\n# # os.makedirs(predicted_dir_name, mode=0o777, exist_ok=False)\n# # logger.info(\"predicted_file_name:{}\",predicted_dir_name)\n# class_id = classifier_transactions.save_classified_result(class_dict,txt_filename,auth_key)\n# logger.info(\"\\n\\n **************** class_id : {}\", class_id)\n# if is_unclassified is not None and int(is_unclassified) == 1:\n# classifier_transactions.update_unclassified_status(dfx_data['unclass_id'],auth_key)\n# if os.path.exists(txt_filename):\n# os.remove(txt_filename)\n# dir_name = os.path.split(txt_filename)[0]\n# temp_file_name = os.path.split(txt_filename)[1]\n# template_loc = os.path.normpath(os.path.join(sim_template_loc ,os.path.split(dir_name)[1]))\n# print( \"\\n dir_name:\",dir_name,\"\\n temp_file_name: \",temp_file_name,\"\\n template_loc\" , template_loc )\n# if os.path.exists(template_loc) and os.path.isdir(template_loc) and len(\n# os.listdir(dir_name)) == 0:\n# # os.remove(os.path.normpath(os.path.join(template_loc, temp_file_name)))\n# os.system(\"rm -rf \"+template_loc)\n# import shutil\n# shutil.rmtree(dir_name)\n# import shutil\n# if os.path.exists(txt_filename):\n# os.remove(txt_filename)\n# elif is_training_source == 1 and float(new_doc_classifier) >= float(classification_model_accuracy):\n# if os.path.exists(txt_filename):\n# os.remove(txt_filename)\n# prediction_json[\"error_msg\"] = \" The document is marked for training \"\n# prediction_json['error_code'] = 1\n# classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n# error_updation.custom_error_update_log(\" Duplicate training document \", \" Duplicate training document \",\n# str(disc_inbound_id))\n# elif int(is_unclassified) is not 1:\n# group_details = find_group(txt_filename,auth_key)\n# add_unclassified_docs(group_details, disc_inbound_id,get_file_name(filename), is_training_source,auth_key)\n# elif len(filename.strip()) > 1 and text_extraction is not None and len(text_extraction.strip()) > 1 and not is_model_enabled:\n# group_details = find_group(txt_filename,auth_key)\n# add_unclassified_docs(group_details, disc_inbound_id, get_file_name(filename), is_training_source,auth_key)\n# else:\n# prediction_json[\"error_msg\"] = data_processing['error_msg']\n# prediction_json['error_code'] = int(data_processing['error_code'])\n# classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n# logger.info(\" prediction_json: {}\", prediction_json)\n# error_updation.custom_error_update_log(data_processing['error_msg'],\n# data_processing['error_msg'],\n# str(disc_inbound_id))\n# else:\n# prediction_json[\"error_msg\"] = \" Kindly check the document's path/read permission/size \"\n# prediction_json['error_code'] = 1\n# classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n# error_updation.custom_error_update_log(\" Kindly check the document's path/permission/size \", \" Kindly check the document's path/permission/size \", str(disc_inbound_id))\n# logger.debug(prediction_json)\n# logger.info(\"\\n\\n prediction_json {}\", prediction_json)\n# except Exception as exception:\n# classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n# error_updation.exception_log(exception, prediction_json, str(disc_inbound_id))\n# # prediction_json['error_code'] = 1\n# logger.debug(\"\\n\\n exception: {}\", exception)\n# prediction_json[\"error_msg\"] = \" Document Exception : \"+str(exception)\n# return json.dumps(str(prediction_json))\n# # tagging the text files\n\ndef eliminating_classified():\n return {\n \"size\": 5000,\n \"query\": {\n \"bool\": {\n \"filter\": {\n \"wildcard\": {\n \"content\": \"*\"\n }\n },\n\n \"must_not\": [\n {\"exists\": {\"field\": \"classified\"}},\n {\"exists\": {\"field\": \"group_no\"}}\n ]\n }\n }\n}\n\n\n\n\n\ndef classification(auth_key):\n import gc\n #try:\n data_processing = {}\n #disc_inbound_id = dfx_data['inbound_id']\n filter_classified = es.search(index=index, body=eliminating_classified())\n raw_data = filter_classified['hits']['hits']\n for data in raw_data:\n es_id = data['_id']#urllib.parse.quote(data['_id'], safe='')\n print(es_id)\n inbound_id=data['_source']['resourceName']\n content = data['_source']['content']\n uri= data['_source']['uri']\n data_processing = doc_processing.filter_text_from_file(inbound_id,content,'auth_key')\n logger.info(\"\\n\\n\\n data_processing: {}\", data_processing)\n txt_filename = data_processing['text_file_name']\n text_extraction = ''\n logger.info(\" \\n Classification: txt_filename {}\",txt_filename)\n\n if os.path.exists(vec_model_path) and os.path.exists(model_path):\n global is_model_enabled\n is_model_enabled = True\n logger.info(\"\\n\\n is_model_enabled : {}\",is_model_enabled)\n with open(txt_filename, 'rb+') as text_file:\n # text_extraction = text_file.read()\n text_raw = '' + (text_file.read()).decode()\n # logger.info(text_raw)\n text_extraction = remove_stop_words(text_raw)\n # logger.info(text_extraction)\n if text_extraction is not None and len(text_extraction.strip()) > 0 and is_model_enabled :\n logger.info(\" Prediction Section : ******************** \\n\")\n features_list=vectorizer_model.transform([text_extraction]) # .toarray()\n prediction=classifier_model.predict(features_list)\n prediction_id = int(prediction)\n logger.debug(\"\\n\\n prediction : {}\", prediction_id)\n proba_pred=classifier_model.predict_proba(features_list)\n new_doc_classifier=proba_pred[:, prediction_id]\n logger.info(\"\\n\\n new_doc_classifier: {}\", new_doc_classifier, float(classification_model_accuracy), float(new_doc_classifier))\n if float(new_doc_classifier) >= float(classification_model_accuracy):\n class_dict = {}\n class_id = 0\n logger.info(\"Prediction : {}\", prediction_id)\n logger.info(\"Prediction Probability :{}\", proba_pred[:, prediction_id])\n prediction_json[\"file_name\"] = inbound_id\n prediction_json[\"predition_type\"] = new_doc_classifier\n class_name = id_to_category.get(int(prediction_id))\n #org_subclass_id = classifier_transactions.find_subclassification_id(class_name,auth_key)\n logger.info(\"\\n\\n class_name : {}\", class_name)\n prediction_json['predicted_category'] = class_name\n classified_doc_update(class_name,prediction_json[\"predition_type\"][0],content,uri)\n\n # class_dict[\"ClassificationTemplateId\"] = classifier_transactions.find_classification_id(org_subclass_id,auth_key)\n # class_dict[\"SubClassificationTemplateId\"] = org_subclass_id\n # class_dict[\"Type\"] = class_name\n # class_dict[\"DiscoveryInBoundId\"] = int(inbound_id)\n\n #--class_id = classifier_transactions.save_classified_result(class_dict,txt_filename,auth_key)\n #--logger.info(\"\\n\\n **************** class_id : {}\", class_id)\n# if is_unclassified is not None and int(is_unclassified) == 1:\n# classifier_transactions.update_unclassified_status(dfx_data['unclass_id'],auth_key)\n else:\n group_details = find_group(txt_filename, auth_key)\n\n unclassified_doc_update(uri, str(group_details['accuracy_rate']),content,str(group_details['group_no']),str(group_details['group_file_path']),str(group_details['doc_name']))\n elastic_update = update_group_es(es_id, str(group_details['group_no']))\n\n\n\n else:\n group_details = find_group(txt_filename,auth_key)\n print(group_details)\n unclassified_doc_update(uri, str(group_details['accuracy_rate']), content, str(group_details['group_no']),str(group_details['group_file_path']),str(group_details['doc_name']))\n elastic_update = update_group_es(es_id, str(group_details['group_no']))\n\n #print(unclassified_doc_update(uri,))\n #add_unclassified_docs(group_details, inbound_id, get_file_name(txt_filename), \"0\",auth_key)\n\n \n \n # except Exception as exception:\n # #classifier_transactions.update_inbound_status(disc_inbound_id,auth_key)\n # #error_updation.exception_log(exception, prediction_json, str(disc_inbound_id))\n # # prediction_json['error_code'] = 1\n # logger.debug(\"\\n\\n exception: {}\", exception)\n #prediction_json[\"error_msg\"] = \" Document Exception : \"+str(exception)\n return json.dumps(str(prediction_json))\n\n\nif __name__ == '__main__':\n if os.path.exists(vec_model_path) and os.path.exists(model_path):\n logger.info(\"{} \\n\\n\\n\")\n from pyfiglet import Figlet\n f = Figlet(font='slant')\n logger.info(\"{}\", f.renderText('DFX CLASSIFIER ') + \"v 0.1\")\n with open(model_path, 'rb') as model_file:\n classifier_model = pickle.load(model_file)\n with open(vec_model_path, 'rb') as vec_model:\n vectorizer_model = pickle.load(vec_model)\n is_model_enabled = True\n logger.info(\" \\n\\n classification :-> is_model_enabled : {}\", int(is_model_enabled))\n #classifier_api = os.path.normpath(os.path.join(config['CLASSIFIER']['CLASSIFIER_HOST'], config['CLASSIFIER']['CLASSIFIER_URL_NAME']))\n # print(classification(r'\\\\192.168.0.14\\Users\\giri\\Desktop\\Credit_Card_Agreements_2018_Q3\\1st Financial Bank USA\\MasterCard or Visa Credit Card Agreement.pdf',9))\n # payload = {}\n # payload['file_name'] = r'C:\\Users\\baskaran\\Desktop\\Milton.pdf'\n # payload['unclass_id'] = 0\n # payload['inbound_id'] = 354\n # payload['IsTrainingSource'] = 0\n # payload['is_unclassified'] = 0\n # payload['FileLength'] = 24546\n # payload['Authorization'] = 'bearer PJZ45ervujGYDWxe0lxRzDPG0fORLHfbaSYHEgDnpUBKF-ueDNlKIWZoNpHUWzNLHVeR5pjcVyRuQpAOn4Ae759UQxT5sH1eshjjR13REREOdr7TfuOXCQhUTFhNUqf8iJChuJjZbXdNYqEnqFX6osHv5YBJbyer5IoiquNOkejk-TQplGJd0gL1_AoT19k_TDpqijaCsZtoHrYRt2ORNaWebl0cdd9pi7jWdYLX6vHnA3c1nxjEQsVjHAxkGIDPGKKYVQPaoXtsDcw6Uk0ypN4siyzdcPj2hWOX7TUWNsI'\n # print(classification(payload))\n # {\"file_name\": \"\\\\\\\\WIN-EN47KDFIU2V\\\\CRMFiles\\\\fsfdfsdfK.docx\", \"unclass_id\": 0, \"inbound_id\": 265,\n # \"IsTrainingSource\": 0, \"is_unclassified\": 0,\n # \"Authorization\": \"bearer 6aAHmqDL4XmXLXAX6lbqhQSTVGR15n8pSaMFIcHn9V0GRhGHHnX4NOf6oSZRHdL9hkjR3U21O6RD8fzRafZIuQpvayXq_hq3qkYrCprUxgbNSj-b9_atmSVzmQlMCqkARfmmxabZqMCP6qKaWM5HyncQvps5SDi9yExt4jGE5qZaRT3DicWlPq1LEdaHKrfAUDX5yNL-PYJyiTmHFiAw-8wUgk0GxvoAGRnhPMTrZo3ZE2nMlDT5SDaJDc40YIMZ9EvgzziV0nj_kWYn5eMezKgcWCbgjfxNpTUEXpZo8lA\",\n # \"FileLength\": null}\n # {\"file_name\": \"\\\\\\\\WIN-EN47KDFIU2V\\\\CRMFiles\\\\9PYMPC_tic.pdf\", \"unclass_id\": 0, \"inbound_id\": 207,\n # \"IsTrainingSource\": 0, \"is_unclassified\": 0,\n # \"Authorization\": \"bearer PJZ45ervujGYDWxe0lxRzDPG0fORLHfbaSYHEgDnpUBKF-ueDNlKIWZoNpHUWzNLHVeR5pjcVyRuQpAOn4Ae759UQxT5sH1eshjjR13REREOdr7TfuOXCQhUTFhNUqf8iJChuJjZbXdNYqEnqFX6osHv5YBJbyer5IoiquNOkejk-TQplGJd0gL1_AoT19k_TDpqijaCsZtoHrYRt2ORNaWebl0cdd9pi7jWdYLX6vHnA3c1nxjEQsVjHAxkGIDPGKKYVQPaoXtsDcw6Uk0ypN4siyzdcPj2hWOX7TUWNsI\"}\n # {\"file_name\": \"\\\\\\\\192.168.0.14\\\\dfx\\\\crm_tmp\\\\2~~d87a4763-8b07-ea11-b806-00155d00090f~~8M1XMA_tic.pdf\",\n # \"unclass_id\": 0, \"inbound_id\": 1, \"IsTrainingSource\": 0, \"is_unclassified\": 0,\n # \"Authorization\": \"bearer 9jYF-wEagkECw_DX75Y80W69TT3m6YQGqQNST4XHhaxb3uiUYTlAYxJnrddmLVB3YXl4p-zPK_KecjmVX3LAZEd0gvbpcILM4ItbYyCwe8dB3Huj8qScXsFpH_ccUAOZydigKLxKR_Px4OICZKGsbxQYQOEhHdDGuDMPxybNDksFNp1L7PrBSWAzs18AmfUDE7RDiHCRDmdkBihJHLPU_AdqAwfJy5Oj2dahFGZVhfcoIb7u-7dr0ok-u5dymJlziEc40M3A9hTYHS4bj9wkAfCuJX7iDx550RW20hh2dZI\"}\n #app.run(host=config['CLASSIFIER']['CLASSIFIER_HOST'], port=config['CLASSIFIER']['API_PORT_NUMBER'], debug=False, threaded = True)\n classification(\"auth_key\")\n\n","repo_name":"giridharanbabu/DFXClassification_V1","sub_path":"classification.py","file_name":"classification.py","file_ext":"py","file_size_in_byte":26535,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"46978979527","text":"class TreeNode:\r\n def __init__(self, x):\r\n self.val = x\r\n self.left = None\r\n self.right = None\r\n\r\ndef Serialize(root):\r\n if not root: return ''\r\n que = [root]\r\n t = [root.val]\r\n while que:\r\n n = len(que)\r\n h_level = []\r\n for _ in range(n):\r\n node = que.pop(0)\r\n if not node.left and not node.right: break\r\n if node.left:\r\n que.append(node.left)\r\n h_level.append(node.left.val)\r\n else: \r\n h_level.append('#')\r\n if node.right:\r\n que.append(node.right)\r\n h_level.append(node.right.val)\r\n else: \r\n h_level.append('#')\r\n t += h_level\r\n t.append('!')\r\n t = map(str, t)\r\n return ','.join(t)\r\n\r\n\r\ndef Deserialize(s):\r\n if not s: return None\r\n s = s.split(',')\r\n root = TreeNode(int(s[0]))\r\n q = [root]\r\n i = 1\r\n while i < len(s)-1:\r\n node = q.pop(0)\r\n if s[i] == '!':\r\n break\r\n if s[i] == '#':\r\n node.left = None\r\n else:\r\n node.left = TreeNode(int(s[i]))\r\n q.append(node.left)\r\n if s[i+1] == '!':\r\n break\r\n if s[i+1] == '#':\r\n node.right = None\r\n else:\r\n node.right = TreeNode(int(s[i+1]))\r\n q.append(node.right)\r\n i += 2\r\n return root\r\n\r\n\r\n\r\n\r\n\r\n\r\n# root = TreeNode(5)\r\n# root.left = TreeNode(4)\r\n# root.left.left = TreeNode(3)\r\n# root.left.left.left = TreeNode(2)\r\n\r\nroot = TreeNode(8)\r\nroot.left = TreeNode(6)\r\nroot.right = TreeNode(10)\r\nroot.left.left = TreeNode(5)\r\nroot.left.right = TreeNode(7)\r\nroot.right.left = TreeNode(9)\r\nroot.right.right = TreeNode(11)\r\n\r\na = Serialize(root)\r\nprint(a)\r\nb = Deserialize(a)\r\nprint(Serialize(b))\r\n\r\ns = []\r\ndef dfs(root):\r\n if not root: return \r\n s.append(root.val)\r\n dfs(root.left)\r\n dfs(root.right)\r\n\r\ndfs(root)\r\nprint(s)\r\n\r\n\r\n\r\n\r\n'''\r\n链接:https://www.nowcoder.com/questionTerminal/cf7e25aa97c04cc1a68c8f040e71fb84?f=discussion\r\n来源:牛客网\r\n\r\ntypedef TreeNode node;\r\ntypedef TreeNode* pnode;\r\ntypedef int* pint;\r\nclass Solution {\r\n vector buf;\r\n void dfs(pnode p){\r\n if(!p) buf.push_back(0x23333);\r\n else{\r\n buf.push_back(p -> val);\r\n dfs(p -> left);\r\n dfs(p -> right);\r\n }\r\n }\r\n pnode dfs2(pint& p){\r\n if(*p == 0x23333){\r\n ++p;\r\n return NULL;\r\n }\r\n pnode res = new node(*p);\r\n ++p;\r\n res -> left = dfs2(p);\r\n res -> right = dfs2(p);\r\n return res;\r\n }\r\npublic:\r\n char* Serialize(TreeNode *p) {\r\n buf.clear();\r\n dfs(p);\r\n int *res = new int[buf.size()];\r\n for(unsigned int i = 0; i < buf.size(); ++i) res[i] = buf[i];\r\n return (char*)res;\r\n }\r\n TreeNode* Deserialize(char *str) {\r\n int *p = (int*)str;\r\n return dfs2(p);\r\n }\r\n};\r\n\r\n'''","repo_name":"asd55667/Jianzhi_offer","sub_path":"61Serialize_Deserialize.py","file_name":"61Serialize_Deserialize.py","file_ext":"py","file_size_in_byte":2995,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"23368408427","text":"import argparse\nfrom voronoi_game import VoronoiGame\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--spawn\", type=int, default=5, help=\"Number of days after which a new unit spawns at the \"\n \"homebase\")\n parser.add_argument(\"--last\", type=int, default=100, help=\"Total number of days the game goes on for\")\n parser.add_argument(\"--seed\", \"-s\", type=int, default=2, help=\"Seed used by random number generator, specify 0 to \"\n \"use no seed and have different random behavior on \"\n \"each launch\")\n parser.add_argument(\"--port\", type=int, default=8080, help=\"Port to start, specify -1 to auto-assign\")\n parser.add_argument(\"--address\", \"-a\", type=str, default=\"127.0.0.1\", help=\"Address\")\n parser.add_argument(\"--no_browser\", \"-nb\", action=\"store_true\", help=\"Disable browser launching in GUI mode\")\n parser.add_argument(\"--no_gui\", \"-ng\", action=\"store_true\", help=\"Disable GUI\")\n parser.add_argument(\"--log_path\", default=\"log\", help=\"Directory path to dump log files, filepath if \"\n \"disable_logging is false\")\n parser.add_argument(\"--disable_logging\", action=\"store_true\", help=\"Disable Logging, log_path becomes path to file\")\n parser.add_argument(\"--disable_timeout\", action=\"store_true\", help=\"Disable timeouts for player code\")\n parser.add_argument(\"--player1\", \"-p1\", default=\"d\", help=\"Specifying player 1 out of 4\")\n parser.add_argument(\"--player2\", \"-p2\", default=\"d\", help=\"Specifying player 2 out of 4\")\n parser.add_argument(\"--player3\", \"-p3\", default=\"d\", help=\"Specifying player 3 out of 4\")\n parser.add_argument(\"--player4\", \"-p4\", default=\"d\", help=\"Specifying player 4 out of 4\")\n parser.add_argument(\"--dump_state\", action=\"store_true\", help=\"Dump game.pkl for rendering\")\n args = parser.parse_args()\n player_list = tuple([args.player1, args.player2, args.player3, args.player4])\n del args.player1\n del args.player2\n del args.player3\n del args.player4\n\n if args.disable_logging:\n if args.log_path == \"log\":\n args.log_path = \"results.log\"\n \n voronoi_game = VoronoiGame(player_list, args)\n","repo_name":"rvg-18227/Voronoi","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2384,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"10081169512","text":"import sys\nsys.stdin = open(\"input.txt\")\nfrom collections import deque\n\ndef knight():\n global cnt\n dx = [1, 1, -1, -1, 2, 2, -2, -2]\n dy = [2, -2, 2, -2, 1, -1, 1, -1]\n # 시작점\n q = deque([[sx,sy]])\n while q:\n x,y = q.popleft()\n # 이동하려는 칸 도착시 현재칸 까지 이동횟수 리턴\n if x == lx and y == ly:\n cnt = graph[x][y]\n\n return graph[x][y]\n # 나이트 이동 방향 탐색\n for k in range(8):\n nx = x + dx[k]\n ny = y + dy[k]\n # 범위 내 있고 탐색 X 탐색\n if 0 <= nx < I and 0 <= ny < I:\n if not graph[nx][ny]:\n q.append([nx, ny])\n graph[nx][ny] = graph[x][y] + 1 # 이동횟수 초기화\n\n\nT = int(input())\nfor tc in range(1, T+1):\n I = int(input())\n graph = [[0 for _ in range(I)] for _ in range(I)]\n sx, sy = map(int, input().split())\n lx, ly = map(int, input().split())\n cnt = 0\n knight()\n print(cnt)\n # print(graph)","repo_name":"MalangCowFarm/Algo_JinHwan","sub_path":"week 9/나이트의 이동_7562(BOJ)/sol.py","file_name":"sol.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"70283989982","text":"import numpy\nimport random\nimport string\nimport nltk\nimport nltk.corpus\nimport nltk.tokenize.punkt\nimport string\nfrom nltk.stem.lancaster import LancasterStemmer\nimport json\nimport pickle\nimport requests\nimport re\nfrom collections import Counter\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.metrics.pairwise import cosine_similarity\nfrom framework.constants import log as log\n\nnltk.download('punkt')\nnltk.download('stopwords')\n\n# Create tokenizer\nstemmer = LancasterStemmer()\nstopwords = nltk.corpus.stopwords.words('english')\nstopwords.extend(string.punctuation)\nstopwords.append('')\ntokenizer = nltk.tokenize.WordPunctTokenizer()\n\ndef train(inp):\n with open(\"intents.json\") as file:\n data = json.load(file)\n try:\n with open(\"data.pickle\", \"rb\") as f:\n words, labels, training, output = pickle.load(f)\n except:\n words = []\n labels = []\n docs_x = []\n docs_y = []\n\n for intent in data[\"intents\"]:\n for pattern in intent[\"patterns\"]:\n wrds = nltk.word_tokenize(pattern)\n words.extend(wrds)\n docs_x.append(wrds)\n docs_y.append(intent[\"tag\"])\n\n if intent[\"tag\"] not in labels:\n labels.append(intent[\"tag\"])\n\n words = [stemmer.stem(w.lower()) for w in words if w != \"?\"]\n words = sorted(list(set(words)))\n\n labels = sorted(labels)\n\n training = []\n output = []\n\n out_empty = [0 for _ in range(len(labels))]\n\n for x, doc in enumerate(docs_x):\n bag = []\n\n wrds = [stemmer.stem(w.lower()) for w in doc]\n\n for w in words:\n if w in wrds:\n bag.append(1)\n else:\n bag.append(0)\n\n output_row = out_empty[:]\n output_row[labels.index(docs_y[x])] = 1\n\n training.append(bag)\n output.append(output_row)\n\n\n training = numpy.array(training)\n output = numpy.array(output)\n with open(\"data.pickle\", \"wb\") as f:\n pickle.dump((words, labels, training, output), f)\n\ndef eventhandle(inp):\n message = inp\n with open(\"nlp/intents.json\",'rb') as f:\n data = json.load(f)\n maxi = 0\n response_item = None\n for items in data[\"intents\"]:\n # regex = ''\n for b in items[\"patterns\"]:\n match = get_jaccard_sim(b,message)\n log.debug(str(match))\n # log.info(\"maxi : \", maxi)\n if(match > maxi):\n maxi = match\n response_item = items\n # log.info(\"maxi changes : \", maxi)\n if response_item is None:\n return \"Sorry I do not understand\"\n else: \n responses = response_item['responses']\n return random.choice(responses)\n\n\ndef get_cosine_sim(*strs): \n vectors = [t for t in get_vectors(*strs)]\n return cosine_similarity(vectors)\n\n\ndef get_vectors(*strs):\n text = [t for t in strs]\n log.debug(text)\n vectorizer = CountVectorizer(text)\n log.debug(vectorizer)\n vectorizer.fit(text)\n return vectorizer.transform(text).toarray()\n\n\ndef get_jaccard_sim(str1, str2): \n a = set(str1.lower().split()) \n # log.debug(\"intent: \", a)\n b = set(str2.lower().split())\n # log.debug(\"input :\", b)\n c = a.intersection(b)\n return float(len(c)) / (len(a) + len(b) - len(c))\n\ndef is_ci_token_stopword_match(a, b):\n \"\"\"Check if a and b are matches.\"\"\"\n tokens_a = [token.lower().strip(string.punctuation) for token in tokenizer.tokenize(a) \\\n if token.lower().strip(string.punctuation) not in stopwords]\n log.debug(\"sentence1: \", tokens_a)\n tokens_b = [token.lower().strip(string.punctuation) for token in tokenizer.tokenize(b) \\\n if token.lower().strip(string.punctuation) not in stopwords]\n log.debug(\"sentence2: \", tokens_b)\n return (tokens_a == tokens_b)\n\ndef bag_of_words(s,words):\n bag = [0 for _ in range(len(words))]\n s_words = nltk.word_tokenize(s)\n log.debug(\"before: \",s_words)\n s_words1 = [stemmer.stem(word.lower()) for word in s_words if word!=\"?\"]\n log.debug(\"final: \", s_words1)\n\n","repo_name":"akankshanb/Bot-Engine","sub_path":"plotbot/nlp/analyze.py","file_name":"analyze.py","file_ext":"py","file_size_in_byte":4206,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"9890989195","text":"# coding: utf-8\n\"\"\"\nAdd NOT NULL constraint for data unless fed_id and component_id are not null.\n\"\"\"\n\nimport os\n\nimport flask_sqlalchemy\n\nfrom .actions_add_objects_readable_by_all_users_by_default import MIGRATION_INDEX as PREVIOUS_MIGRATION_INDEX\n\nMIGRATION_INDEX = PREVIOUS_MIGRATION_INDEX + 1\nMIGRATION_NAME, _ = os.path.splitext(os.path.basename(__file__))\n\n\ndef run(db: flask_sqlalchemy.SQLAlchemy) -> bool:\n constraints = db.session.execute(db.text(\"\"\"\n SELECT conname\n FROM pg_catalog.pg_constraint\n WHERE conname = 'files_not_null_check_data'\n \"\"\")).fetchall()\n\n if len(constraints) > 0:\n return False\n\n db.session.execute(db.text(\"\"\"\n UPDATE files SET data = '{\"storage\": \"local\", \"original_file_name\": \"\"}'::json WHERE data IS NULL AND (fed_id IS NULL OR component_id IS NULL)\n \"\"\"))\n\n # Perform migration\n db.session.execute(db.text(\"\"\"\n ALTER TABLE files\n ADD CONSTRAINT files_not_null_check_data\n CHECK (\n (\n fed_id IS NOT NULL AND\n component_id IS NOT NULL\n ) OR data IS NOT NULL\n )\n \"\"\"))\n return True\n","repo_name":"sciapp/sampledb","sub_path":"sampledb/models/migrations/files_add_data_not_null_constraint.py","file_name":"files_add_data_not_null_constraint.py","file_ext":"py","file_size_in_byte":1199,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"7"} +{"seq_id":"70497383265","text":"#!/usr/bin/env python\n#\n\n# Python modules to load\nimport chimera\nimport os\t\t\t # For running OS commands\nimport subprocess \t\t# For invoking bash scripts inside this python script\nimport glob\nimport fnmatch # For gettin numbers of files\nimport shutil # For deleting existing analysis folders\nfrom shutil import copyfile\n\nfrom chimera import runCommand as rc # use 'rc' as shorthand for rc\nfrom chimera import replyobj # for emitting status messages\nfrom chimera.tkgui import saveReplyLog, clearReplyLog\n\n# Definition for ignoring hidden files, note use of sorted\ndef listdir_nohidden(path):\n return sorted(glob.glob(os.path.join(path, '*/')))\n\n#####################################################################################\n# REQUIRED VARIABLES - edit these to make point the script to PDB's\n#####################################################################################\n\n# Required variables\nthreshold = 3 # Volume threshold, in sigma\norigin = 'originIndex 0' # Insert an volume origin command here if desired\nmove = '0,0,0' # If the PDB's need moving\n#name = '28mini' # For file naming\n#res = '9.07' # Resolution for map model cross correlation\n\n#####################################################################################\n# Get to work\n#####################################################################################\n\n# Current working directory is set\ncwd = os.getcwd()\nprint('Current working diretory set:')\nprint(cwd)\n\n# Get directories inside PDB_built_final_geometry\ngeom = listdir_nohidden(cwd+\"/PDB_built_final_geometry\")\n\n# Get map info, name and resolution\nfor line in open(\"map/.map_info\"):\n if \"Map\" in line:\n print(line)\n fields = line.strip().split()\n # Array indices start at 0 unlike AWK\n print(fields[1])\n name = fields[1]\n\nfor line in open(\"map/.map_info\"):\n if \"Resolution\" in line:\n print(line)\n fields = line.strip().split()\n # Array indices start at 0 unlike AWK\n print(fields[1])\n res = fields[1]\n\n# Gather name of map found in the /map folder\nfor file in os.listdir(cwd+\"/map\"):\n if file.endswith(\".mrc\"):\n map=file\n\nshutil.rmtree(str(cwd)+'/PDB_built_final_geometry_all', ignore_errors=True)\nos.mkdir(str(cwd)+'/PDB_built_final_geometry_all')\nos.mkdir(str(cwd)+'/PDB_built_final_geometry_all/images')\nos.mkdir(str(cwd)+'/PDB_built_final_geometry_all/movies_legs')\nos.mkdir(str(cwd)+'/PDB_built_final_geometry_all/movies_cage')\nos.mkdir(str(cwd)+'/PDB_built_final_geometry_all/PDB')\n\n# Since we now make movies etc, set window size\nrc('windowsize 1000 1000')\n\n#####################################################################################\n# Open reference model for alignment against\n#####################################################################################\n\n# Open ref PDB into #0\nrc('open #0 '+str(cwd)+'/scripts/ref_PDB.pdb')\n\n#####################################################################################\n# First loop through geometry classes, then then PDB's within that geometry\n#####################################################################################\n\n# Loop through directories inside /PDB_built_final_geometry\nfor dir in geom:\n # Get information about geometry\n signature = os.path.basename(os.path.dirname(dir))\n print(dir)\n print (signature)\n\n # Gather name of PDB found in the current geometry directory\n # Note sorted is very important to make list ordered\n pdblist = [fn for fn in sorted(os.listdir(dir)) if fn.endswith(\".pdb\")]\n print(pdblist)\n # No of PDB's\n pdbno = len(fnmatch.filter(os.listdir(dir), '*.pdb'))\n\n # Create directory for RMSD, alignbment matrices and CC saveReplyLog\n shutil.rmtree(str(dir)+'/RMSD', ignore_errors=True)\n os.mkdir(str(dir)+'/RMSD')\n shutil.rmtree(str(dir)+'/mapCC', ignore_errors=True)\n os.mkdir(str(dir)+'/mapCC')\n os.mkdir(str(dir)+'/mapCC/matrix')\n os.mkdir(str(dir)+'/mapCC/images')\n\n # Make more appropriate view for figures, use refPDB in #0 as starting orientation\n # Save to internal view accessible by 'reset p1'\n rc ('modeldisplay #0')\n rc('reset; focus; turn y -120; turn z -15; turn x 70; focus; savepos p1')\n\n #Open structures\n for i in range(0,pdbno):\n file=str(pdblist[i])\n rc('open '+str(dir)+'/'+str(file))\n\n # Color structures because you should\n rc('color #ADD8E6 #:1-225; color #fb9a99 #:331-838; color #6a3d9a #:839-1074; color #6a3d9a #:1075-1198; color #1F78B4 #:1199-1576; color #da0048 #:1577-1675')\n rc('color white #0')\n rc('~modeldisplay #0')\n\n #####################################################################################\n # Measure angles\n #####################################################################################\n\n #\n # Currently this is implemented in 2_TS_measure_angles.py\n #\n\n #####################################################################################\n # Calulate RMSD of aligned models\n #####################################################################################\n\n # Do structural alignment\n for i in range(1,pdbno+1):\n rc('match #'+str(i)+':1500-1600@ca #0:1500-1600@ca')\n\n # Save matrix file for restoring structural alignmemnt\n rc('matrixget '+dir+'/RMSD/RMSD_alignment_matrix')\n\n # Clear reply log ready for gathering RMSD\n clearReplyLog()\n\n # Do RMSD measurements in a matrix\n for i in range(1,pdbno+1):\n for j in range(1,pdbno+1):\n rc('rmsd #'+str(i)+' #'+str(j))\n\n # Save reply log ready for calculated RMSD values\n saveReplyLog(dir+'/RMSD/RMSD.log')\n\n # Save session\n rc('~modeldisplay #; modeldisplay #0; focus; ~modeldisplay #; modeldisplay #1-#'+str(pdbno))\n rc('save '+dir+'/RMSD/RMSD_'+str(signature)+'_session.py')\n\n #####################################################################################\n # Make images of fits and set up directory with PDB's and images with geometry signature prefix\n #####################################################################################\n\n # Open map for cross-correlation and making images, into #999\n rc('open #999 '+str(cwd)+'/map/'+str(map))\n rc('volume #999 sdLevel 4 color grey transparency 0.8')\n #Save default cage and model orientation\n rc('savepos p2')\n\n # Split matrices from RMSD alignment to get map trans/rots for making images\n os.chdir(str(dir)+\"RMSD\")\n #Call shell script for parsing matrix files\n subprocess.call(str(cwd)+'/scripts/matrix/extract_alignment_matrices.sh', shell=True)\n # Return to working directory\n os.chdir(\"..\")\n\n # Gather names of map alignment matrices\n # Note sorted is very important to make list ordered\n matlist = [fn for fn in sorted(os.listdir(dir+'/mapCC/matrix')) if fn.endswith(\".mat\")]\n print(matlist)\n # No of PDB's\n matno = len(fnmatch.filter(os.listdir(dir+'/mapCC/matrix'), '*.mat'))\n\n # Loop through matrices, apply to map keeping structure in place to inspect the alignment\n rc('2dlabels create label1 color black size 40 xpos 0.75 ypos 0.93 text \"\"')\n rc('2dlabels create label2 color black size 40 xpos 0.93 ypos 0.88 text \"\"')\n for j in range(0,matno):\n #Set view\n rc('reset p2')\n file=str(matlist[j])\n rc('matrixset '+str(dir)+'/mapCC/matrix/'+str(file))\n rc('modeldisplay #0')\n rc('turn y 60; turn x 30; focus #0; scale 0.9; clip hither 30; clip yon -80')\n rc('~modeldisplay #; modeldisplay #999; modeldisplay #'+str(j+1))\n #Make label\n rc('2dlabels delete label1; 2dlabels create label1 color black size 40 xpos 0.75 ypos 0.93 text \"'+str(signature)+'\"')\n rc('2dlabels delete label2; 2dlabels create label2 color black size 40 xpos 0.93 ypos 0.88 text \"'+str(j+1)+'\"')\n #Save image\n file=str(pdblist[j])\n rc('copy file '+str(dir)+'/mapCC/images/'+str(file)+'.png')\n rc('copy file '+str(dir)+'/../../PDB_built_final_geometry_all/images/'+str(signature)+'_'+str(file)+'.png')\n #Save image sequence for movie encoding\n rc('copy file '+str(dir)+'/../../PDB_built_final_geometry_all/movies_legs/'+str(signature)+'_image_'+str(j+1).zfill(3)+'.png')\n #Save PDB into directory with geometry signature prefix\n rc('write #'+str(j+1)+' '+str(dir)+'/../../PDB_built_final_geometry_all/PDB/'+str(signature)+'_'+str(file)+'.pdb')\n #Save this view so it can be recalled from mapCC session file, using: reset leg1, leg2, leg3 etc\n rc('savepos leg'+str(j+1))\n #Save an image of the cage with these geometries displayed\n rc('reset')\n rc('2dlabels delete label1; 2dlabels create label1 color black size 40 xpos 0.05 ypos 0.93 text \"'+str(signature)+'\"')\n rc('2dlabels delete label2; 2dlabels create label2 color black size 40 xpos 0.05 ypos 0.88 text \"'+str(name)+'\"')\n rc('~modeldisplay #; modeldisplay #1-'+str(pdbno)+'; modeldisplay #999')\n rc('copy file '+str(dir)+'/../../PDB_built_final_geometry_all/movies_cage/'+str(signature)+'_image_'+str(j+1).zfill(3)+'.png')\n\n rc('2dlabels delete label1; 2dlabels delete label2')\n\n #Call shell script encoding movie\n #subprocess.call(str(cwd)+'/scripts/matrix/extract_alignment_matrices.sh', shell=True)\n\n #####################################################################################\n # Calulate model-map cross correlation\n #####################################################################################\n\n # Reset positions to bring triskelia back onto original cage positions, save an image\n rc('reset')\n rc('~modeldisplay #; modeldisplay #1-'+str(pdbno)+'; modeldisplay #999')\n rc('copy file '+str(dir)+'/mapCC/map_CC_'+str(signature)+'.png')\n\n # Clear reply log ready for gathering model cross correlation\n clearReplyLog()\n\n # Do model-map cross correlation\n for i in range(1,pdbno+1):\n rc('molmap #'+str(i)+' '+str(res)+' modelId #1000')\n rc('measure correlation #999 #1000')\n rc('close #1000')\n\n # Save reply log ready for measured model cross correlation\n saveReplyLog(dir+'/mapCC/map_CC.log')\n\n # Save session\n rc('save '+dir+'/mapCC/map_CC_'+str(signature)+'_session.py')\n\n #####################################################################################\n # Close models ready for next geometry class\n #####################################################################################\n\n # Close structures\n rc('close #1-#'+str(pdbno))\n\n #####################################################################################\n # Create leg path session for figures\n #####################################################################################\n\n # Gather name of leg bilds found in the current geometry directory\n # Note sorted is very important to make list ordered\n leglist = [fn for fn in sorted(os.listdir(dir+'/PDB_angles_bild')) if fn.endswith(\"leg.bild\")]\n print(leglist)\n # No of PDB's\n legno = len(fnmatch.filter(os.listdir(dir+'/PDB_angles_bild'), '*leg.bild'))\n\n #Open leg paths\n for i in range(0,legno):\n file=str(leglist[i])\n rc('open '+str(dir)+'PDB_angles_bild/'+str(file))\n\n #Apply alignment matrix\n rc('matrixset '+dir+'/RMSD/RMSD_alignment_matrix')\n\n # Save session\n rc('~modeldisplay #; modeldisplay #0; focus; ~modeldisplay #; modeldisplay #1-#'+str(legno))\n rc('save '+dir+'/RMSD/legs_'+str(signature)+'_session.py')\n\n # Close legs\n rc('close #1-#'+str(pdbno))\n\n #####################################################################################\n # Create principal component session for figures\n #####################################################################################\n\n # Gather name of leg bilds found in the current geometry directory\n # Note sorted is very important to make list ordered\n paxlist = [fn for fn in sorted(os.listdir(dir+'/PDB_angles_bild')) if fn.endswith(\"pax.bild\")]\n print(paxlist)\n # No of PDB's\n paxno = len(fnmatch.filter(os.listdir(dir+'/PDB_angles_bild'), '*pax.bild'))\n\n #Open leg paths\n for i in range(0,paxno):\n file=str(paxlist[i])\n rc('open '+str(dir)+'PDB_angles_bild/'+str(file))\n\n #Apply alignment matrix\n rc('matrixset '+dir+'/RMSD/RMSD_alignment_matrix')\n\n # Save session\n rc('~modeldisplay #; modeldisplay #0; focus; ~modeldisplay #; modeldisplay #1-#'+str(paxno))\n rc('save '+dir+'/RMSD/pax_'+str(signature)+'_session.py')\n\n # Close pax\n rc('close #1-#'+str(paxno))\n\n#####################################################################################\n# Tidy up\n#####################################################################################\n\n#Call shell script encoding movie\nos.chdir(cwd)\nsubprocess.call(str(cwd)+'/scripts/movie/movie_encode.sh', shell=True)\n\n# Close chimera when finished\nrc('stop')\n\n\n\n\n#### Depreciated movie code\n\n # Loop through matrices as above but make a movie of this\n #rc('movie record')\n #rc('2dlabels create label1 color black size 40 xpos 0.75 ypos 0.93 text \"\"')\n #rc('2dlabels create label2 color black size 40 xpos 0.93 ypos 0.88 text \"\"')\n #for k in range(0,matno):\n #rc('reset p2')\n #file=str(matlist[i])\n #rc('matrixset '+str(dir)+'/mapCC/matrix/'+str(file))\n #rc('modeldisplay #0')\n #rc('turn y 60; turn x 30; focus #0; scale 0.9; clip hither 30; clip yon -80')\n #rc('~modeldisplay #; modeldisplay #999; modeldisplay #'+str(k+1))\n #rc('2dlabels delete label1; 2dlabels create label1 color black size 40 xpos 0.75 ypos 0.93 text \"'+str(signature)+'\"')\n #rc('2dlabels delete label2; 2dlabels create label2 color black size 40 xpos 0.93 ypos 0.88 text \"'+str(k)+'\"')\n #rc('wait 10')\n #rc('movie encode output '+str(dir)+'/../../PDB_built_final_geometry_all/movies/'+str(signature)+'_movie.mp4 quality highest')\n #rc('2dlabels delete label1; 2dlabels delete label2')\n","repo_name":"kylelmorris/Morris_2019_NSMB","sub_path":"cage_modelling/36barrel/scripts/6_build_final_geometry_RMSD_mapCC.py","file_name":"6_build_final_geometry_RMSD_mapCC.py","file_ext":"py","file_size_in_byte":14031,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"74366793182","text":"import argparse, time\nfrom reconhece_frente import reconhece_frente\nfrom reconhece_verso import reconhece_verso\nfrom reconhece_texto import reconhece_texto\nfrom apply_filter import apply_filter\n\n\ndef main():\n name = str(time.time()).split('.')[0]\n \n parser = argparse.ArgumentParser()\n\n parser.add_argument('-s', '--support')\n parser.add_argument('-f', '--frente')\n parser.add_argument('-v', '--verso')\n\n args = parser.parse_args()\n\n for currentArgument in args.__dict__:\n if args.__dict__[currentArgument] != None:\n if currentArgument == 'frente':\n doc_frente = args.__dict__[currentArgument]\n\n elif currentArgument == 'verso':\n doc_verso = args.__dict__[currentArgument]\n\n elif currentArgument == 'support':\n print('Help')\n\n else:\n print(f'{currentArgument} não é um argumento válido')\n\n if doc_frente and doc_verso:\n reconhece_frente(doc_frente, name)\n reconhece_verso(doc_verso, name)\n apply_filter(name)\n reconhece_texto(name)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"anthonyreis/reconhecimento_doc","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1141,"program_lang":"python","lang":"pt","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"75158807263","text":"import pandas as pd\n\ndf = pd.read_csv(\"executions_fx_btc.tsv\", sep=\"\\t\")\n\ndf[\"base_price\"] = df.iloc[0][\"price\"]\ndf[\"price\"] = df[\"price\"] - df[\"base_price\"]\n# print(df)\ndf[\"date\"] = pd.to_datetime(df[\"date\"], format=\"%Y-%m-%dT%H:%M:%S.%fZ\")\ndf = df.set_index(\"date\")\n\n\ndf = df[\"price\"].resample(\"S\").mean()\ndf = df.resample(\"S\").interpolate()\n\ndfs = []\nfor i in range(1, 101):\n _df = df.rolling(i).mean()\n _df.name = str(i)\n dfs.append(_df)\n\ndf = pd.concat(dfs, axis=1)\ndf = df.reset_index()\ndf = df.drop(\"date\", axis=1)\n# print(df)\n\nfor index, row in df.iterrows():\n for i, value in enumerate(row.values):\n if not pd.isna(value):\n x = index / 10\n y = value / 1_000\n z = i / 10\n print(x, y, z, sep=\"\\t\")\n\n if index == 300:\n break","repo_name":"ShunsukeMzk/PlotterServer","sub_path":"tools/editor.py","file_name":"editor.py","file_ext":"py","file_size_in_byte":803,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"13213682731","text":"\nimport pandas as pd # Pandas 资料处理套件\nimport mplfinance as mpf # 绘制 K 线套件\n \n# 读取每日股价资料\n\n\nnames = ['Date', 'Adj_Close', 'Close', 'High', 'Low', 'Open', 'Volume']\ndf = pd.read_csv(r'stock2317.csv', # 每日股价资料 CSV 档\n header=None, \n names=names,\n skiprows=3, # 跳过开头三行档头标记\n index_col=None, \n delimiter=',')\ndf['Date'] = pd.to_datetime(df['Date']) # 字串转为 datetime 资料型态\ndf.set_index('Date', inplace=True) # 指定索引排序字段为 Date\ndf_adj = df.iloc[:,1:5] # 'Close', 'High', 'Low', 'Open'\n\nfrom pylab import mpl\nmpl.rcParams['font.sans-serif'] = ['SimHei']\nmpl.rcParams['axes.unicode_minus'] = False\n\n# 设定绘图颜色\nmc = mpf.make_marketcolors(\n\tup='red', # 上涨 K 棒颜色\n\tdown='green', # 下跌 K 棒颜色 \n\tedge='i', # K线线柱边缘颜色(i代表继承自up和down的颜色)\n\twick='i', # 上下影线颜色\n\tvolume='in', # 成交量长条图的颜色\n\tinherit=True) # 是否继承颜色设定\n\ns = mpf.make_mpf_style(\n\tgridaxis='both', # 格线位置\n\tgridstyle='-.', # 格线线型\n\ty_on_right=False, # y轴位置是否在右\n rc={'font.size':12, \n 'font.sans-serif':'SimHei', # 指定中文字型\n 'axes.titlesize':18,\n 'axes.labelsize':16,\n 'xtick.labelsize':12,\n 'ytick.labelsize':12},\n\tmarketcolors=mc)\n\nmpf.plot(df_adj,\n type='candle', # 指定 K 棒绘图符号格式\n title=' 鸿海2020年股价',\n ylabel='股价',\n figratio=(15, 10), # 图形大小\n figscale=1,\n xrotation=20, # 日期显示旋转角度\n style=s, # 设定绘图风格\n show_nontrading=False) # 是否显示无交易日\n","repo_name":"pony0613/Binance-Trading-Bot","sub_path":"kline.py","file_name":"kline.py","file_ext":"py","file_size_in_byte":2081,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7226872378","text":"import numpy as np\nfrom numba import njit, guvectorize\n\n\ndef set_parameters(r=0.05, beta=0.94, sigma=2.0, sigma_eps=0.5, N_eps=7, N_a=200, amax=1000):\n \"\"\"\n Defines the main parameters of the model.\n \"\"\"\n params = dict()\n\n # INTEREST RATE\n params['r'] = r\n\n # PREFERENCES\n params['sigma'] = sigma # Inverse elasticity of substitution\n params['beta'] = beta # Subjective discount factor\n\n # IDIOSYNCRATIC PRODUCTIVITY\n params['rho_eps'] = 0.5\n params['sigma_eps'] = sigma_eps\n params['N_eps'] = N_eps # Number of idiosyncatic states (epsilon)\n params['y_eps'], params['pi_eps'], params['Pi_eps'] = markov_incomes(\n rho=params['rho_eps'], sigma_y=params['sigma_eps'], N=params['N_eps'])\n\n # ASSET GRID\n params['N_a'] = N_a # Number of grid points\n params['amax'] = amax # Maximum value of the grid\n params['a'] = agrid(amin=0.0, amax=params['amax'], N=params['N_a']) # Borrowing limit and asset grid\n\n return params\n\n\n@guvectorize(['void(float64[:], float64[:], float64[:], float64[:])'], '(n),(nq),(n)->(nq)')\ndef interpolate_y(x, xq, y, yq):\n \"\"\"\n Efficient linear interpolation exploiting monotonicity.\n\n Complexity O(n+nq), so most efficient when x and xq have comparable number of points.\n Extrapolates linearly when xq out of domain of x.\n\n Parameters\n ----------\n x: array\n ascending data points\n xq: array\n ascending query points\n y: array\n data points\n yq: array\n empty to be filled with interpolated points\n \"\"\"\n nxq, nx = xq.shape[0], x.shape[0]\n\n xi = 0\n x_low = x[0]\n x_high = x[1]\n for xqi_cur in range(nxq):\n xq_cur = xq[xqi_cur]\n while xi < nx - 2:\n if x_high >= xq_cur:\n break\n xi += 1\n x_low = x_high\n x_high = x[xi + 1]\n\n xqpi_cur = (x_high - xq_cur) / (x_high - x_low)\n yq[xqi_cur] = xqpi_cur * y[xi] + (1 - xqpi_cur) * y[xi + 1]\n\n\n@guvectorize(['void(float64[:], float64[:], uint32[:], float64[:])'], '(n),(nq)->(nq),(nq)')\ndef interpolate_coord(x, xq, xqi, xqpi):\n \"\"\"\n Efficient linear interpolation exploiting monotonicity. xq = xqpi * x[xqi] + (1-xqpi) * x[xqi+1]\n\n Parameters\n ----------\n x: array\n ascending data points\n xq: array\n ascending query points\n xq: array\n empty to be filled with indices of lower bracketing gridpoints\n xqpi: array\n empty to be filled with weights on lower bracketing gridpoints\n\n \"\"\"\n nxq, nx = xq.shape[0], x.shape[0]\n\n xi = 0\n x_low = x[0]\n x_high = x[1]\n for xqi_cur in range(nxq):\n xq_cur = xq[xqi_cur]\n while xi < nx - 2:\n if x_high >= xq_cur:\n break\n xi += 1\n x_low = x_high\n x_high = x[xi + 1]\n\n xqpi[xqi_cur] = (x_high - xq_cur) / (x_high - x_low)\n xqi[xqi_cur] = xi\n\n\n@njit(fastmath=True)\ndef forward_step(D, Pi_T, a_pol_i, a_pol_pi):\n \"\"\"\n Single forward step to update distribution using an arbitrary asset policy.\n\n Efficient implementation of D_t = Lam_{t-1}' @ D_{t-1} using sparsity of Lam_{t-1}.\n\n Parameters\n ----------\n D: np.ndarray\n Beginning-of-period distribution over s_t, a_(t-1)\n Pi_T: np.ndarray\n Transpose Markov matrix that maps s_t to s_(t+1)\n a_pol_i: np.ndarray\n Left gridpoint of asset policy\n a_pol_pi: np.ndarray\n Weight on left gridpoint of asset policy\n\n Returns\n ----------\n Dnew : np.ndarray\n Beginning-of-next-period dist s_(t+1), a_t\n\n \"\"\"\n # first create Dnew from updating asset state\n Dnew = np.zeros((D.shape[0], D.shape[1]))\n for s in range(D.shape[0]):\n for i in range(D.shape[1]):\n apol = a_pol_i[s, i]\n api = a_pol_pi[s, i]\n d = D[s, i]\n Dnew[s, apol] += d * api\n Dnew[s, apol + 1] += d * (1 - api)\n\n # then use transpose Markov matrix to update income state\n Dnew = Pi_T @ Dnew\n\n return Dnew\n\n\ndef markov_rouwenhorst(rho, sigma, N=7):\n \"\"\"\n Rouwenhorst method to discretize an AR(1) process\n \"\"\"\n # parametrize Rouwenhorst for n=2\n p = (1 + rho) / 2\n Pi = np.array([[p, 1 - p], [1 - p, p]])\n\n # implement recursion to build from n=3 to n=N\n for n in range(3, N + 1):\n P1, P2, P3, P4 = (np.zeros((n, n)) for _ in range(4))\n P1[:-1, :-1] = p * Pi\n P2[:-1, 1:] = (1 - p) * Pi\n P3[1:, :-1] = (1 - p) * Pi\n P4[1:, 1:] = p * Pi\n Pi = P1 + P2 + P3 + P4\n Pi[1:-1] /= 2\n\n # invariant distribution and scaling\n pi = stationary(Pi)\n s = np.linspace(-1, 1, N)\n s *= (sigma / np.sqrt(var(s, pi)))\n\n return s, pi, Pi\n\n\ndef markov_incomes(rho, sigma_y, N=11):\n \"\"\"\n Simple helper method that assumes AR(1) process in logs for incomes and scales aggregate income\n to 1, also that takes in sdy as the *cross-sectional* sd of log incomes\n \"\"\"\n sigma = sigma_y * np.sqrt(1 - rho ** 2)\n s, pi, Pi = markov_rouwenhorst(rho, sigma, N)\n y = np.exp(s) / np.sum(pi * np.exp(s))\n return y, pi, Pi\n\n\ndef mean(x, pr):\n pr = pr / np.sum(pr)\n return np.sum(pr * x)\n\n\ndef cov(x, y, pr):\n pr = pr / np.sum(pr)\n return np.sum(pr * (x - mean(x, pr)) * (y - mean(y, pr)))\n\n\ndef var(x, pr):\n pr = pr / np.sum(pr)\n return cov(x, x, pr)\n\n\ndef ineq(ss, pop):\n \"\"\"\n Inequality statistics.\n \"\"\"\n T, Neps, Na = ss['a'].shape\n a_flat = ss['a'].reshape(T, 1, Neps * Na).squeeze() # reshape multi-dimensional policies\n Dst_flat = ss['D'].reshape(T, 1, Neps * Na).squeeze() # flatten out the joint distribution\n\n # Lorenz curves\n a = np.einsum('js,js->s', pop.pi[:, np.newaxis], a_flat)\n p = np.einsum('js,js->s', pop.pi[:, np.newaxis], Dst_flat)\n p = p / np.sum(p) # Make sure sums to one\n a_sorted = np.sort(a) # Sort vectors from lowest to highest\n a_sorted_i = np.argsort(a)\n p_a_sorted = p[a_sorted_i] # Recover associated probabilities\n lorenz_a_pctl, lorenz_a = lorenz(a_sorted, p_a_sorted) # Get Lorenz curves\n\n return lorenz_a_pctl, lorenz_a\n\n\ndef lorenz(x, pr):\n \"\"\"\n Returns Lorenz curve.\n \"\"\"\n # first do percentiles of the total population\n pctl = np.concatenate(([0], pr.cumsum() - pr / 2, [1]))\n # now do percentiles of total wealth (returns only zeros if sum(pr*x) = 0)\n wealthshare = (x * pr / np.sum(x * pr) if np.sum(x * pr) != 0 else np.zeros_like(x))\n wealthpctl = np.concatenate(([0], wealthshare.cumsum() - wealthshare / 2, [1]))\n return pctl, wealthpctl\n\n\ndef find_nearest(array, value):\n array = np.asarray(array)\n return (np.abs(array - value)).argmin()\n\n\ndef agrid(amax, N, amin=0):\n \"\"\"\n Grid a+pivot evenly log-spaced between amin+pivot and amax+pivot\n \"\"\"\n pivot = np.abs(amin) + 0.25\n a = np.geomspace(amin + pivot, amax + pivot, N) - pivot\n a[0] = amin # make sure *exactly* equal to amin\n\n return a\n\n\n@njit\ndef within_tolerance(x1, x2, tol):\n \"\"\"\n Efficiently test max(abs(x1-x2)) <= tol for arrays of same dimensions x1, x2.\n \"\"\"\n y1 = x1.ravel()\n y2 = x2.ravel()\n for i in range(y1.shape[0]):\n if np.abs(y1[i] - y2[i]) > tol:\n return False\n return True\n\n\ndef stationary(Pi, pi_seed=None, tol=1E-11, maxit=10_000):\n \"\"\"\n Find invariant distribution of a Markov chain by iteration.\n \"\"\"\n if pi_seed is None:\n pi = np.ones(Pi.shape[0]) / Pi.shape[0]\n else:\n pi = pi_seed\n\n for it in range(maxit):\n pi_new = pi @ Pi\n if within_tolerance(pi_new, pi, tol):\n break\n pi = pi_new\n else:\n raise ValueError(f'No convergence after {maxit} forward iterations!')\n pi = pi_new\n\n return pi\n\n\ndef make_path(x, T):\n \"\"\"\n Takes in x as either a number, a vector or a matrix, turning it into a path.\n \"\"\"\n x = np.asarray(x)\n if x.ndim <= 1:\n return np.tile(x, (T, 1))\n\n elif x.ndim == 2:\n return np.tile(x, (T, 1, 1))\n\n\ndef make_full_path(x, T):\n \"\"\"\n Takes a path x (vector/matrix), and repeats the last line until x has T lines.\n \"\"\"\n if x.ndim == 1:\n raise ValueError('x must be a column vector')\n\n if T < x.shape[0]:\n raise ValueError('T must be greater than the number of lines in x')\n\n return np.vstack((x, make_path(x[-1], T - x.shape[0])))\n\n\ndef pack_jacobians(jacdict, inputs, outputs, T):\n \"\"\"\n If we have T*T jacobians from nI inputs to nO outputs in jacdict, combine into (nO*T)*(nI*T) jacobian matrix.\n \"\"\"\n nI, nO = len(inputs), len(outputs)\n\n outjac = np.empty((nO * T, nI * T))\n for iO in range(nO):\n subdict = jacdict.get(outputs[iO], {})\n for iI in range(nI):\n outjac[(T * iO):(T * (iO + 1)), (T * iI):(T * (iI + 1))] = make_matrix(\n subdict.get(inputs[iI], np.zeros((T, T))), T)\n return outjac\n\n\ndef unpack_jacobians(bigjac, inputs, outputs, T):\n \"\"\"\n If we have an (nO*T)*(nI*T) jacobian and provide names of nO outputs and nI inputs, output nested dictionary\n \"\"\"\n nI, nO = len(inputs), len(outputs)\n\n jacdict = {}\n for iO in range(nO):\n jacdict[outputs[iO]] = {}\n for iI in range(nI):\n jacdict[outputs[iO]][inputs[iI]] = bigjac[(T * iO):(T * (iO + 1)), (T * iI):(T * (iI + 1))]\n return jacdict\n\n\ndef make_matrix(A, T):\n \"\"\"\n If A is not an outright ndarray, e.g. it is SimpleSparse, call its .matrix(T) method\n to convert it to T*T array.\n \"\"\"\n if not isinstance(A, np.ndarray):\n return A.matrix(T)\n else:\n return A\n\n\ndef pack_vectors(vs, names, T):\n v = np.zeros(len(names)*T)\n for i, name in enumerate(names):\n if name in vs:\n v[i*T:(i+1)*T] = vs[name]\n return v\n\n\ndef unpack_vectors(v, names, T):\n vs = {}\n for i, name in enumerate(names):\n vs[name] = v[i*T:(i+1)*T]\n return vs\n","repo_name":"FredericMartenet/OLGHA","sub_path":"model/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":9920,"program_lang":"python","lang":"en","doc_type":"code","stars":16,"dataset":"github-code","pt":"7"} +{"seq_id":"8927581840","text":"from opendc.models.cpu import CPU\nfrom opendc.util import exceptions\nfrom opendc.util.rest import Response\n\n\ndef GET(request):\n \"\"\"Get the specs of a CPU.\"\"\"\n\n # Make sure required parameters are there\n\n try:\n request.check_required_parameters(\n path={\n 'id': 'int'\n }\n )\n\n except exceptions.ParameterError as e:\n return Response(400, e.message)\n\n # Instantiate a CPU and make sure it exists\n\n cpu = CPU.from_primary_key((request.params_path['id'],))\n\n if not cpu.exists():\n return Response(404, '{} not found.'.format(cpu))\n\n # Return this CPU\n\n return Response(\n 200,\n 'Successfully retrieved {}.'.format(cpu),\n cpu.to_JSON()\n )\n","repo_name":"atlarge-research/opendc-web-server","sub_path":"opendc/api/v1/specifications/cpus/id/endpoint.py","file_name":"endpoint.py","file_ext":"py","file_size_in_byte":747,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"7"} +{"seq_id":"33534647505","text":"\"\"\"\nGiven n as input, print the following pattern.\nInput: n=4\nOutput:\ny1 \ny1y2 \ny1y2y3 \ny1y2y3y4\n\"\"\"\n\nn = 4\npt = \"y\"\nst = \"\"\nfor i in range (1, n+1):\n st = st + pt + str(i)\n print(st)\nprint(\"---\")\n\n","repo_name":"enigmatic-cipher/JPR-M1-Task-1","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":200,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"44024870100","text":"import inspect\nfrom dataclasses import dataclass\nfrom functools import partial\nfrom typing import Any, Callable, Dict, List, Optional, Sequence, Tuple\n\nfrom pytorch_lightning.utilities import rank_zero_warn\nfrom pytorch_lightning.utilities.enums import LightningEnum\n\nfrom flash.core.data.callback import ControlFlow\nfrom flash.core.data.utilities.collate import default_collate\nfrom flash.core.data.utils import _STAGES_PREFIX\nfrom flash.core.utilities.stages import RunningStage\nfrom flash.core.utilities.types import INPUT_TRANSFORM_TYPE\n\n\nclass InputTransformPlacement(LightningEnum):\n PER_SAMPLE_TRANSFORM = \"per_sample_transform\"\n PER_BATCH_TRANSFORM = \"per_batch_transform\"\n COLLATE = \"collate\"\n PER_SAMPLE_TRANSFORM_ON_DEVICE = \"per_sample_transform_on_device\"\n PER_BATCH_TRANSFORM_ON_DEVICE = \"per_batch_transform_on_device\"\n\n\nINVALID_STAGES_FOR_INPUT_TRANSFORMS = [RunningStage.SANITY_CHECKING, RunningStage.TUNING]\n\n\n@dataclass\nclass _InputTransformPerStage:\n collate_in_worker: bool\n transforms: Optional[Dict[str, Callable]] = None\n\n\n@dataclass\nclass InputTransform:\n def __post_init__(self):\n self.callbacks: Optional[List] = None\n\n # used to keep track of provided transforms\n self._transform: Dict[RunningStage, _InputTransformPerStage] = {}\n\n # For all the stages possible, set/load the transforms.\n for stage in RunningStage:\n if stage not in INVALID_STAGES_FOR_INPUT_TRANSFORMS:\n self._populate_transforms_for_stage(stage)\n\n def current_transform(self, stage: RunningStage, current_fn: str) -> Callable:\n return self._transform[stage].transforms.get(current_fn, self._identity)\n\n ########################\n # PER SAMPLE TRANSFORM #\n ########################\n\n def per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for all stages stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n pass\n\n def train_per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for the training stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_sample_transform()\n\n def val_per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for the validating stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform()\n\n def test_per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for the testing stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_sample_transform()\n\n def predict_per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for the predicting stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform()\n\n def serve_per_sample_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on cpu for the serving stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform()\n\n ##################################\n # PER SAMPLE TRANSFORM ON DEVICE #\n ##################################\n\n def per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for all stages stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n pass\n\n def train_per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for the training stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_sample_transform_on_device()\n\n def val_per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for the validating stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform_on_device()\n\n def test_per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for the testing stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_sample_transform_on_device()\n\n def predict_per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for the predicting stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_sample_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform_on_device()\n\n def serve_per_sample_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a single sample on device for the serving stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def serve_per_sample_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_sample_transform_on_device()\n\n #######################\n # PER BATCH TRANSFORM #\n #######################\n\n def per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for all stages stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n pass\n\n def train_per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for the training stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_batch_transform()\n\n def val_per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for the validating stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform()\n\n def test_per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for the testing stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_batch_transform()\n\n def predict_per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for the predicting stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform()\n\n def serve_per_batch_transform(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on cpu for the serving stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform()\n\n #################################\n # PER BATCH TRANSFORM ON DEVICE #\n #################################\n\n def per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for all stages stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n pass\n\n def train_per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for the training stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_batch_transform_on_device()\n\n def val_per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for the validating stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform_on_device()\n\n def test_per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for the testing stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n \"\"\"\n return self.per_batch_transform_on_device()\n\n def predict_per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for the predicting stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def per_batch_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform_on_device()\n\n def serve_per_batch_transform_on_device(self) -> Callable:\n \"\"\"Defines the transform to be applied on a batch of data on device for the serving stage.\n\n The input data of the transform would have the following form::\n\n {\n DataKeys.INPUT: ...,\n DataKeys.TARGET: ...,\n DataKeys.METADATA: ...,\n }\n\n You would need to use :class:`flash.core.data.transforms.ApplyToKeys` as follows:\n\n .. code-block:: python\n\n from flash.core.data.transforms import ApplyToKeys\n\n\n class MyInputTransform(InputTransform):\n def serve_per_batch_transform_on_device(self) -> Callable:\n return ApplyToKeys(\"input\", my_func)\n\n \"\"\"\n return self.per_batch_transform_on_device()\n\n ###########\n # COLLATE #\n ###########\n\n def collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of sample to create a batch for all stages.\"\"\"\n return default_collate\n\n def train_collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of training sample to create a training batch.\"\"\"\n return self.collate()\n\n def val_collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of validating sample to create a validating batch.\"\"\"\n return self.collate()\n\n def test_collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of testing sample to create a testing batch.\"\"\"\n return self.collate()\n\n def predict_collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of predicting sample to create a predicting batch.\"\"\"\n return self.collate()\n\n def serve_collate(self) -> Callable:\n \"\"\"Defines the transform to be applied on a list of serving sample to create a serving batch.\"\"\"\n return self.collate()\n\n ########################################\n # HOOKS CALLED INTERNALLY WITHIN FLASH #\n ########################################\n\n def _per_sample_transform(self, sample: Any, stage: RunningStage) -> Any:\n fn = self.current_transform(stage=stage, current_fn=\"per_sample_transform\")\n if isinstance(sample, list):\n return [fn(s) for s in sample]\n return fn(sample)\n\n def _per_batch_transform(self, batch: Any, stage: RunningStage) -> Any:\n \"\"\"Transforms to apply to a whole batch (if possible use this for efficiency).\n\n .. note:: This option is mutually exclusive with :meth:`per_sample_transform_on_device`, since if both are\n specified, uncollation has to be applied.\n\n \"\"\"\n return self.current_transform(stage=stage, current_fn=\"per_batch_transform\")(batch)\n\n def _collate(self, samples: Sequence, stage: RunningStage) -> Any:\n \"\"\"Transform to convert a sequence of samples to a collated batch.\"\"\"\n return self.current_transform(stage=stage, current_fn=\"collate\")(samples)\n\n def _per_sample_transform_on_device(self, sample: Any, stage: RunningStage) -> Any:\n \"\"\"Transforms to apply to the data before the collation (per-sample basis).\n\n .. note:: This option is mutually exclusive with :meth:`per_batch_transform`, since if both are\n specified, uncollation has to be applied. .. note:: This function won't be called within the dataloader\n workers, since to make that happen each of the workers would have to create it's own CUDA-context which\n would pollute GPU memory (if on GPU).\n\n \"\"\"\n fn = self.current_transform(stage=stage, current_fn=\"per_sample_transform_on_device\")\n if isinstance(sample, list):\n return [fn(s) for s in sample]\n return fn(sample)\n\n def _per_batch_transform_on_device(self, batch: Any, stage: RunningStage) -> Any:\n \"\"\"Transforms to apply to a whole batch (if possible use this for efficiency).\n\n .. note:: This function won't be called within the dataloader workers, since to make that happen each of\n the workers would have to create it's own CUDA-context which would pollute GPU memory (if on GPU).\n\n \"\"\"\n return self.current_transform(stage=stage, current_fn=\"per_batch_transform_on_device\")(batch)\n\n #############\n # UTILITIES #\n #############\n\n def inject_collate_fn(self, collate_fn: Callable):\n # For all the stages possible, set collate function\n if collate_fn is not default_collate:\n for stage in RunningStage:\n if stage not in [RunningStage.SANITY_CHECKING, RunningStage.TUNING]:\n self._transform[stage].transforms[InputTransformPlacement.COLLATE.value] = collate_fn\n\n def _populate_transforms_for_stage(self, running_stage: RunningStage):\n transform, collate_in_worker = self.__check_transforms(\n transform=self.__resolve_transforms(running_stage),\n )\n\n self._transform[running_stage] = _InputTransformPerStage(\n collate_in_worker=collate_in_worker,\n transforms=transform,\n )\n\n def __resolve_transforms(self, running_stage: RunningStage) -> Optional[Dict[str, Callable]]:\n transforms = {}\n stage = _STAGES_PREFIX[running_stage]\n\n # iterate over all transforms hook name\n for transform_name in InputTransformPlacement:\n transform_name = transform_name.value\n\n method_name = f\"{stage}_{transform_name}\"\n\n # get associated transform\n try:\n fn = getattr(self, method_name)()\n except AttributeError as e:\n raise AttributeError(\n str(e) + \". Make sure you include a call to super().__init__(...) in your __init__ after setting \"\n \"all attributes.\"\n )\n\n if fn is None:\n continue\n\n if not callable(fn):\n raise TypeError(f\"The hook {method_name} should return a callable.\")\n\n transforms[transform_name] = fn\n\n return transforms\n\n def __check_transforms(self, transform: Dict[str, Callable]) -> Tuple[Dict[str, Callable], Optional[bool]]:\n is_per_batch_transform_in = \"per_batch_transform\" in transform\n is_per_sample_transform_on_device_in = \"per_sample_transform_on_device\" in transform\n\n if is_per_batch_transform_in and is_per_sample_transform_on_device_in:\n raise TypeError(\n f\"{transform}: `per_batch_transform` and `per_sample_transform_on_device` are mutually exclusive.\"\n )\n\n collate_in_worker: Optional[bool] = not is_per_sample_transform_on_device_in\n\n return transform, collate_in_worker\n\n @staticmethod\n def _identity(x: Any) -> Any:\n return x\n\n def __str__(self) -> str:\n return f\"{self.__class__.__name__}(\" + f\"transform={self._transform})\"\n\n\ndef create_or_configure_input_transform(\n transform: INPUT_TRANSFORM_TYPE,\n transform_kwargs: Optional[Dict] = None,\n) -> Optional[InputTransform]:\n if not transform_kwargs:\n transform_kwargs = {}\n\n if isinstance(transform, InputTransform):\n return transform\n\n if inspect.isclass(transform) and issubclass(transform, InputTransform):\n # Deprecation Warning\n rank_zero_warn(\n \"Please pass an instantiated object of the `InputTransform` class. Passing the Class and keyword arguments\"\n \" separately has been deprecated since v0.8.0 and will be removed in v0.9.0.\",\n stacklevel=8,\n category=FutureWarning,\n )\n return transform(**transform_kwargs)\n\n if isinstance(transform, partial):\n return transform(**transform_kwargs)\n\n if not transform:\n return None\n\n raise ValueError(f\"The format for the transform isn't correct. Found {transform}\")\n\n\nclass _InputTransformProcessor:\n \"\"\"\n This class is used to encapsulate the following functions of an `InputTransform` Object:\n Inside a worker:\n per_sample_transform: Function to transform an individual sample\n collate: Function to merge sample into a batch\n per_batch_transform: Function to transform an individual batch\n\n Inside main process:\n per_sample_transform_on_device: Function to transform an individual sample\n collate: Function to merge sample into a batch\n per_batch_transform_on_device: Function to transform an individual batch\n \"\"\"\n\n def __init__(\n self,\n input_transform: InputTransform,\n collate_fn: Callable,\n per_sample_transform: Callable,\n per_batch_transform: Callable,\n stage: RunningStage,\n apply_per_sample_transform: bool = True,\n on_device: bool = False,\n ):\n super().__init__()\n self.input_transform = input_transform\n self.callback = ControlFlow(self.input_transform.callbacks or [])\n self.collate_fn = collate_fn\n self.per_sample_transform = per_sample_transform\n self.per_batch_transform = per_batch_transform\n self.apply_per_sample_transform = apply_per_sample_transform\n self.stage = stage\n self.on_device = on_device\n\n def __call__(self, samples: Sequence[Any]) -> Any:\n if not self.on_device:\n for sample in samples:\n self.callback.on_load_sample(sample, self.stage)\n\n if self.apply_per_sample_transform:\n list_samples = [samples] if not isinstance(samples, list) else samples\n\n transformed_samples = [self.per_sample_transform(sample, self.stage) for sample in list_samples]\n\n for sample in transformed_samples:\n if self.on_device:\n self.callback.on_per_sample_transform_on_device(sample, self.stage)\n else:\n self.callback.on_per_sample_transform(sample, self.stage)\n\n collated_samples = self.collate_fn(transformed_samples, self.stage)\n self.callback.on_collate(collated_samples, self.stage)\n else:\n collated_samples = samples\n\n transformed_collated_samples = self.per_batch_transform(collated_samples, self.stage)\n if self.on_device:\n self.callback.on_per_batch_transform_on_device(transformed_collated_samples, self.stage)\n else:\n self.callback.on_per_batch_transform(transformed_collated_samples, self.stage)\n return transformed_collated_samples\n\n def __str__(self) -> str:\n # todo: define repr function which would take object and string attributes to be shown\n return (\n \"_InputTransformProcessor:\\n\"\n f\"\\t(per_sample_transform): {str(self.per_sample_transform)}\\n\"\n f\"\\t(collate_fn): {str(self.collate_fn)}\\n\"\n f\"\\t(per_batch_transform): {str(self.per_batch_transform)}\\n\"\n f\"\\t(apply_per_sample_transform): {str(self.apply_per_sample_transform)}\\n\"\n f\"\\t(on_device): {str(self.on_device)}\\n\"\n f\"\\t(stage): {str(self.stage)}\"\n )\n\n\ndef __make_collates(input_transform: InputTransform, on_device: bool, collate: Callable) -> Tuple[Callable, Callable]:\n \"\"\"Returns the appropriate collate functions based on whether the transforms happen in a DataLoader worker or on the\n device (main process).\"\"\"\n if on_device:\n return input_transform._identity, collate\n return collate, input_transform._identity\n\n\ndef __configure_worker_and_device_collate_fn(\n running_stage: RunningStage, input_transform: InputTransform\n) -> Tuple[Callable, Callable]:\n transform_for_stage: _InputTransformPerStage = input_transform._transform[running_stage]\n\n worker_collate_fn, device_collate_fn = __make_collates(\n input_transform, not transform_for_stage.collate_in_worker, input_transform._collate\n )\n\n return worker_collate_fn, device_collate_fn\n\n\ndef create_worker_input_transform_processor(\n running_stage: RunningStage, input_transform: InputTransform\n) -> _InputTransformProcessor:\n \"\"\"This utility is used to create the 2 `_InputTransformProcessor` objects which contain the transforms used as the\n DataLoader `collate_fn`.\"\"\"\n worker_collate_fn, _ = __configure_worker_and_device_collate_fn(\n running_stage=running_stage, input_transform=input_transform\n )\n return _InputTransformProcessor(\n input_transform,\n worker_collate_fn,\n input_transform._per_sample_transform,\n input_transform._per_batch_transform,\n running_stage,\n )\n\n\ndef create_device_input_transform_processor(\n running_stage: RunningStage, input_transform: InputTransform\n) -> _InputTransformProcessor:\n \"\"\"This utility is used to create a `_InputTransformProcessor` object which contain the transforms used as the\n DataModule `on_after_batch_transfer` hook.\"\"\"\n _, device_collate_fn = __configure_worker_and_device_collate_fn(\n running_stage=running_stage, input_transform=input_transform\n )\n return _InputTransformProcessor(\n input_transform,\n device_collate_fn,\n input_transform._per_sample_transform_on_device,\n input_transform._per_batch_transform_on_device,\n running_stage,\n apply_per_sample_transform=device_collate_fn != input_transform._identity,\n on_device=True,\n )\n","repo_name":"Lightning-Universe/lightning-flash","sub_path":"src/flash/core/data/io/input_transform.py","file_name":"input_transform.py","file_ext":"py","file_size_in_byte":30739,"program_lang":"python","lang":"en","doc_type":"code","stars":1717,"dataset":"github-code","pt":"7"} +{"seq_id":"26906075940","text":"import nafuma.auxillary as aux\nimport os\n\ndef test_swap_values():\n\n\n test_dict = {'test1': 1, 'test2': 2}\n key1 = 'test1'\n key2 = 'test2'\n\n oldval1 = test_dict[key1]\n oldval2 = test_dict[key2]\n\n new_dict = aux.swap_values(options=test_dict, key1=key1, key2=key2) \n\n assert (test_dict[key1] == oldval2) and (test_dict[key2] == oldval1)\n\n\ndef test_ceil() -> None:\n\n assert aux.ceil(1.05, 0.5) == 1.5\n assert aux.ceil(1.05, 1) == 2.0\n assert aux.ceil(1.1, 0.2) == 1.2\n\n\ndef test_floor() -> None:\n\n assert aux.floor(2.02, 1) == 2.0\n assert aux.floor(2.02, 0.01) == 2.02\n assert aux.floor(2.013, 0.01) == 2.01\n\n\n\ndef test_options() -> None:\n\n\n options = {}\n\n default_options = {\n 'test1': 1,\n 'test2': 2, \n 'test3': 3,\n 'test4': 4,\n 'test5': 5,\n }\n\n\n options = aux.update_options(options=options, default_options=default_options)\n\n assert options['test1'] == default_options['test1']\n\n\ndef test_save_options() -> None:\n \n options = {'test1': 1, 'test2': 2}\n path = 'tmp.dat'\n\n aux.save_options(options, path)\n\n assert os.path.isfile(path)\n\n os.remove(path)\n\n\ndef test_load_options() -> None:\n \n options = {'test1': 1, 'test2': 2}\n path = 'tmp.dat'\n\n aux.save_options(options, path)\n\n loaded_options = aux.load_options(path)\n\n assert (loaded_options['test1'] == 1) and (loaded_options['test2'] == 2)\n\n os.remove(path)\n","repo_name":"rasmusthog/nafuma","sub_path":"nafuma/test/test_auxillary.py","file_name":"test_auxillary.py","file_ext":"py","file_size_in_byte":1440,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"6319021647","text":"import math\nimport os\n\nimport pandas as pd\nimport numpy as np\nfrom tqdm import tqdm\nimport tensorflow as tf\nfrom tensorflow import keras\nimport keras.backend as K\n\nfrom RecSys_Course_AT_PoliMi.Pipeline.data_extraction import get_dataframes\n\ntqdm.pandas()\n\n\ndef compute_mean(x, y):\n if y != 0:\n return x / y\n else:\n return -1\n\n\ndef compute_quartet_entropy(wi, sp, su, au):\n total = wi + sp + su + au\n if total == 0:\n return -1\n if wi == 0:\n wi_coeff = 0\n else:\n wi_coeff = - (wi / total) * math.log2(wi / total)\n if sp == 0:\n sp_coeff = 0\n else:\n sp_coeff = - (sp / total) * math.log2(sp / total)\n if su == 0:\n su_coeff = 0\n else:\n su_coeff = - (su / total) * math.log2(su / total)\n if au == 0:\n au_coeff = 0\n else:\n au_coeff = - (au / total) * math.log2(au / total)\n return 1 - ((wi_coeff + sp_coeff + su_coeff + au_coeff) / 2)\n\n\ndef compute_season(x, record):\n month = x.date.month\n elems = record[x.item_id]\n if month in [12, 1, 2]:\n elems[0] += 1\n elif month in [3, 4, 5]:\n elems[1] += 1\n elif month in [6, 7, 8]:\n elems[2] += 1\n else:\n elems[3] += 1\n\n\ndef compute_seasonality_tendency(df, attributes):\n print('Computing Seasonality Tendency...')\n assert len(attributes) == 4\n return df.progress_apply(lambda x:\n compute_quartet_entropy(x[attributes[0]], x[attributes[1]],\n x[attributes[2]], x[attributes[3]]),\n axis=1)\n\n\ndef compute_season_tendency(df, target, columns):\n assert target in columns\n print('Computing {} Tendency...'.format(target))\n return df.progress_apply(lambda x:\n compute_mean(x[target],\n x[columns[0]] + x[columns[1]] +\n x[columns[2]] + x[columns[3]]),\n axis=1)\n\n\ndef extract_season(sessions, item_features, columns, attr_type='views'):\n assert type(attr_type) == str\n assert len(columns) == 4\n sessions.date = pd.to_datetime(sessions.date)\n record = {item: [0, 0, 0, 0] for item in item_features.item_id.unique()}\n\n print('Building Statistics...')\n sessions.progress_apply(lambda x: compute_season(x, record), axis=1)\n season_df = pd.DataFrame.from_dict(record, orient='index', columns=columns)\n season_df['seasonality_' + attr_type + '_tendency'] = compute_seasonality_tendency(season_df, columns)\n season_df[columns[0] + '_tendency'] = compute_season_tendency(season_df, columns[0], columns)\n season_df[columns[1] + '_tendency'] = compute_season_tendency(season_df, columns[1], columns)\n season_df[columns[2] + '_tendency'] = compute_season_tendency(season_df, columns[2], columns)\n season_df[columns[3] + '_tendency'] = compute_season_tendency(season_df, columns[3], columns)\n return season_df\n\n\ndef get_item_attributes(dataset_path, path, init_date='2020-01-01', end_date='2021-05-31', use_base_features=False):\n \"\"\"\n init_date is inclusive\n end_date is exclusive\n dataset_path is the relative path to the Dataset directory to pass to get_dataframes\n path is the name of the file to load/create (if not present it will create a file\n with name 'path_init_date_end_date.csv')\n When use_base_features is True creates simil-one-hot-encoding based on original dataset features\n \"\"\"\n path = path.split('.csv')[0] + '_' + init_date.replace('-', '_') + '_' + end_date.replace('-', '_') + '.csv'\n if os.path.exists(path):\n print('Attributes already computed, reloading...')\n return pd.read_csv(path)\n print('Attributes not computed, creating...')\n item_features_df, train_sessions_df, train_purchases_df, test_sessions_df, candidate_items_df = get_dataframes(\n dataset_path)\n train_sessions_df = train_sessions_df[(train_sessions_df.date >= init_date) & (train_sessions_df.date < end_date)]\n train_purchases_df = train_purchases_df[\n (train_purchases_df.date >= init_date) & (train_purchases_df.date <= end_date)]\n\n columns_views = ['winter_views', 'spring_views', 'summer_views', 'autumn_views']\n season_views_df = extract_season(train_sessions_df, item_features_df, columns_views, 'views')\n columns_purchases = ['winter_purchases', 'spring_purchases', 'summer_purchases', 'autumn_purchases']\n season_purchases_df = extract_season(train_purchases_df, item_features_df, columns_purchases, 'purchases')\n\n season_df = season_views_df.merge(right=season_purchases_df, left_index=True, right_index=True)\n season_df.insert(0, 'item_id', season_df.index)\n\n season_df['total_views'] = season_df.apply(lambda x: x[columns_views[0]] + x[columns_views[1]] +\n x[columns_views[2]] + x[columns_views[3]], axis=1)\n season_df['total_purchases'] = season_df.apply(lambda x: x[columns_purchases[0]] + x[columns_purchases[1]] +\n x[columns_purchases[2]] + x[columns_purchases[3]], axis=1)\n\n if use_base_features:\n item_features_unstack = simil_one_hot_mapping(item_features_df, 'feature_category_id',\n 'feature_value_id', 'item_id')\n season_df = item_features_unstack.merge(right=season_df, left_on='item_id', right_on='item_id')\n season_df.reset_index(inplace=True)\n season_df.drop(columns=['index'], inplace=True)\n\n season_df.to_csv(path, index=False)\n return season_df\n\n\nclass VariationalAutoEncoder(keras.Model):\n def __init__(self, inputShape, batchSize, latentSize):\n super(VariationalAutoEncoder, self).__init__()\n self.inputShape = inputShape\n self.batchSize = batchSize\n self.latentSize = latentSize\n self.input_layer = keras.Input(shape=self.inputShape)\n self.e2 = keras.layers.Dense(units=self.latentSize * 4)(self.input_layer)\n self.b2 = tf.keras.layers.BatchNormalization()(self.e2)\n self.r2 = tf.keras.layers.LeakyReLU(alpha=0.3)(self.b2)\n self.d2 = tf.keras.layers.Dropout(0.2, seed=42)(self.r2)\n self.z_mean = keras.layers.Dense(self.latentSize)(self.d2)\n self.z_log_sigma = keras.layers.Dense(self.latentSize)(self.d2)\n self.z = keras.layers.Lambda(self.sampling)([self.z_mean, self.z_log_sigma])\n self.encoder = keras.Model(self.input_layer, [self.z_mean, self.z_log_sigma, self.z], name='encoder')\n self.latent_inputs = keras.Input(shape=(self.latentSize,), name='z_sampling')\n self.e5 = keras.layers.Dense(units=self.latentSize * 4)(self.latent_inputs)\n self.b5 = tf.keras.layers.BatchNormalization()(self.e5)\n self.r5 = tf.keras.layers.LeakyReLU(alpha=0.3)(self.b5)\n self.d5 = tf.keras.layers.Dropout(0.2, seed=42)(self.r5)\n self.output_layer = keras.layers.Dense(self.inputShape[0], activation='sigmoid')(self.d5)\n self.decoder = keras.Model(self.latent_inputs, self.output_layer, name='decoder')\n self.output_layer = self.decoder(self.encoder(self.input_layer)[2])\n self.vae = keras.Model(self.input_layer, self.output_layer, name='vae_mlp')\n\n def sampling(self, args):\n z_mean, z_log_sigma = args\n epsilon = K.random_normal(shape=(K.shape(z_mean)[0], self.latentSize),\n mean=0., stddev=0.1)\n return z_mean + K.exp(z_log_sigma) * epsilon\n\n def call(self, vector):\n z_mean, z_log_sigma, z = self.encoder(vector)\n reconstructed = self.decoder(z)\n kl_loss = -0.5 * tf.reduce_mean(\n z_log_sigma - tf.square(z_mean) - tf.exp(z_log_sigma) + 1\n )\n self.add_loss(kl_loss)\n self.add_metric(kl_loss, name='kl_loss', aggregation='mean')\n return reconstructed\n\n def createEmbedding(self, vector):\n return self.encoder(vector)\n\n\ndef simil_one_hot_mapping(df, column_1, column_2, column_index):\n assert type(column_1) == str and type(column_2) == str and type(column_index) == str\n assert column_1 in df.columns and column_2 in df.columns and column_index in df.columns\n cp = df.copy(deep=True)\n cp['mapping'] = cp[column_1].astype(str) + '-' + cp[column_2].astype(str)\n keys = list(cp['mapping'].unique())\n values = [v for v in range(len(keys))]\n mapping_dict = dict(zip(keys, values))\n cp['mapping'] = cp['mapping'].map(mapping_dict)\n cp.drop([column_1, column_2], axis=1, inplace=True)\n cp['value'] = 1\n cp = cp.pivot(index=column_index, columns='mapping', values='value')\n cp.fillna(value=0, inplace=True)\n cp.reset_index(inplace=True)\n return cp\n\n\ndef get_embeddings(dataset_df, epochs, batch_size, learning_rate, validation_split, latent_dim,\n patience_early, patience_reduce, path, one_hot=False):\n path = path.split('.csv')[0] + '_' + str(latent_dim) + '.csv'\n if os.path.exists(path):\n print('Embeddings already computed, reloading...')\n return pd.read_csv(path)\n print('Embeddings not computed, creating...')\n assert dataset_df.columns[0] == 'item_id'\n if not one_hot:\n dataset_df = simil_one_hot_mapping(dataset_df, 'feature_category_id', 'feature_value_id', 'item_id')\n dataset_tensor = tf.convert_to_tensor(dataset_df.copy(deep=True)[dataset_df.columns[1:]].values)\n auto_encoder = VariationalAutoEncoder(inputShape=(dataset_tensor.shape[1],),\n batchSize=batch_size,\n latentSize=latent_dim)\n auto_encoder.compile(optimizer=tf.keras.optimizers.Adam(learning_rate=learning_rate),\n loss=keras.losses.MeanSquaredError())\n callbacks = [\n tf.keras.callbacks.EarlyStopping(monitor='val_loss', patience=patience_early, restore_best_weights=True),\n tf.keras.callbacks.ReduceLROnPlateau(monitor='val_loss', factor=0.1, patience=patience_reduce, min_lr=1e-7,\n verbose=1, cooldown=0)]\n auto_encoder.fit(dataset_tensor, dataset_tensor,\n epochs=epochs,\n batch_size=batch_size,\n validation_split=validation_split,\n callbacks=[callbacks],\n verbose=1)\n encoded_df = pd.DataFrame(auto_encoder.createEmbedding(dataset_tensor)[2])\n encoded_df.insert(0, 'item_id', dataset_df.item_id)\n encoded_df.to_csv(path, index=False)\n return encoded_df\n\n\ndef compute_aggregation_score(item_list, aggregator, values):\n rows = values.loc[values.item_id.isin(item_list)]\n if aggregator == 'sum':\n return np.sum(np.array(rows[rows.columns[1:]]), axis=0)\n if aggregator == 'mean':\n return np.mean(np.array(rows[rows.columns[1:]]), axis=0)\n if aggregator == 'prod':\n return np.prod(np.array(rows[rows.columns[1:]]), axis=0)\n\n\ndef get_session_views_embeddings(sessions, dataset_path, latent_dim, aggregator, path):\n assert aggregator in ['sum', 'mean', 'prod']\n path = path.split('.csv')[0] + '_' + str(latent_dim) + '_' + str(aggregator) + '.csv'\n if os.path.exists(path):\n print('Embedding already computed, reloading...')\n return pd.read_csv(path)\n print('Embeddings not computed, creating...')\n item_features_df = pd.read_csv(dataset_path + 'Dataset/item_features.csv', sep=',')\n item_attributes = simil_one_hot_mapping(item_features_df, 'feature_category_id', 'feature_value_id', 'item_id')\n item_embeddings = get_embeddings(dataset_df=item_attributes[item_attributes.columns[:905]], epochs=200, batch_size=32,\n learning_rate=1e-3, validation_split=0.2, latent_dim=latent_dim, patience_early=10,\n patience_reduce=5, path='../../Dataset/item_embeddings.csv', one_hot=True)\n session_dict = sessions.groupby(['session_id'])['item_id'].progress_apply(lambda items:\n list(items.value_counts().index)).to_dict()\n session_scores_dict = {key: compute_aggregation_score(value, aggregator, item_embeddings)\n for key, value in tqdm(session_dict.items())}\n\n session_scores_df = pd.DataFrame.from_dict(session_scores_dict, orient='index')\n session_scores_df.reset_index(inplace=True)\n session_scores_df.rename(columns={session_scores_df.columns[0]: 'session_id'}, inplace=True)\n session_scores_df.to_csv(path, index=False)\n return session_scores_df\n","repo_name":"recsyspolimi/recsys-challenge-2022-dressipi","sub_path":"RecSys_Course_AT_PoliMi/Pipeline/attributes_utils.py","file_name":"attributes_utils.py","file_ext":"py","file_size_in_byte":12609,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"36987870360","text":"#%%\n# import libraries\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\n#%%\n# read data\npath = 'ex1data1.txt'\ndata = pd.read_csv(path, header = None, names= ['Population', 'Profit'])\n\n#%%\n# show imported data details\n# print('data = \\n', data.head(10))\n# print('**************************************')\n# print('data.describe = \\n', data.describe())\n# print('**************************************')\n\n#%%\n# draw data\nfig, ax = plt.subplots(figsize= (6,6))\nplt.scatter(data['Population'], data['Profit'], marker= 'o', s= 12, label= 'Training data')\nax.set(title= 'Best Fit Line', xlabel= 'Population', ylabel= 'Profit')\n\n#%%\n# adding a new column called ones before the data\ndata.insert(0, 'Ones', 1)\n# print('new data = \\n', data.head(10))\n# print('**************************************')\n\n#%%\n# separate X (training data) from y (target variable)\ncols = data.shape[1]\nX = data.iloc[:, 0 : cols-1]\ny = data.iloc[:, cols-1 : cols]\n\n# print('X data = \\n', X.head(10))\n# print('**************************************')\n# print('y data = \\n', y.head(10))\n# print('**************************************')\n\n#%%\n# Convert data from data frames to numpy matrices\nX = np.matrix(X.values)\ny = np.matrix(y.values)\ntheta = np.matrix(np.array(np.zeros((cols-1, 1))))\n\n# print('X \\n', X)\nprint('X.shape = ', X.shape)\nprint('**************************************')\n# print('y \\n', y)\nprint('y.shape = ', y.shape)\nprint('**************************************')\n# print('theta \\n', theta)\nprint('theta.shape = ', theta.shape)\nprint('**************************************')\n\n#%%\n# Cost function\ndef computeCost(X, y, theta):\n J = np.sum(np.power(((X * theta) - y), 2)) / (2 * len(y))\n return float(J)\n\nprint('With theta = [0 : 0]\\nCost computed =', computeCost(X, y, theta))\nprint('Expected cost value (approx) 32.07')\nprint('**************************************')\nprint('With theta = [-1 : 2]\\nCost computed =', computeCost(X, y, np.array([[-1],[2]])))\nprint('Expected cost value (approx) 54.24')\nprint('**************************************')\n\n#%%\n# Running Gradient Descent\ndef gradientDescent(X, y, theta, alpha, num_iters):\n temp = np.matrix(np.zeros(theta.shape))\n J_history = np.zeros((num_iters, 1))\n \n for i in range(num_iters):\n for j in range(len(theta)):\n temp[j, 0] = theta[j, 0] - ((alpha / len(y)) * np.sum(np.multiply(((X * theta) - y), X[:, j])))\n \n theta = temp.copy()\n J_history[i] = computeCost(X, y, theta)\n \n return theta, J_history\n\niterations = 1500\nalpha = 0.01\n\ntheta, J_history = gradientDescent(X, y, theta, alpha, iterations)\n\nprint('Theta found by gradient descent:\\n', theta[0,0], '\\n ', theta[1,0])\nprint('Expected theta values (approx)\\n -3.6303\\n 1.1664')\nprint('**************************************')\n# print('J_history\\n', J_history)\n# print('**************************************')\n\n#%%\n# Plot the linear fit\nplt.plot(X[:, 1], X*theta, 'r-', linewidth= 1.5, label= 'Linear regression')\nplt.legend()\n\n#%%\n# Predict values for population sizes of 35,000 and 70,000\npredict1 = [1., 3.5] * theta\nprint('For population = 35,000, we predict a profit of %f' %(predict1 * 10000))\npredict2 = [1., 7.] * theta\nprint('For population = 70,000, we predict a profit of %f' %(predict2 * 10000))\nprint('**************************************')\n\n#%%\n# draw error graph\nfig, ax = plt.subplots(figsize=(6,6))\nax.plot(np.arange(1, iterations+1), J_history, 'r', linewidth=2)\nax.set(xlabel= 'No. of Iterations', ylabel= 'Cost', title= 'Error vs. Training Epoch')\n\n#%%\n# Visualizing J(theta_0, theta_1)\n\n# Grid over which we will calculate J\ntheta0_vals = np.linspace(-10, 10, 100)\ntheta1_vals = np.linspace(-1, 4, 100)\n\n# initialize J_vals to a matrix of 0's\nJ_vals = np.zeros((len(theta0_vals), len(theta1_vals)))\n\n# Fill out J_vals\nfor i in range(len(theta0_vals)):\n for j in range(len(theta1_vals)):\n t = np.array([[theta0_vals[i]], [theta1_vals[j]]])\n J_vals[i,j] = computeCost(X, y, t)\n \nJ_vals = J_vals.T\ntheta0_vals, theta1_vals = np.meshgrid(theta0_vals, theta1_vals)\nfig, ax = plt.subplots(figsize= (7,7), subplot_kw={'projection' : '3d'})\nax.plot_surface(theta0_vals, theta1_vals, J_vals, cmap= 'jet')\n# ax.view_init(0,90)\nax.set(xlabel= '$\\mathrm{\\\\theta_{0}}$', ylabel= '$\\mathrm{\\\\theta_{1}}$', zlabel= 'Cost Function $\\mathrm{J(\\\\theta)}$')\nplt.plot(theta[0,0], theta[1,0], 'rx', markersize=10, linewidth=2)","repo_name":"xAbdalla/Machine_Learning_Exercises-Stanford_University","sub_path":"Python/ex1/ex1.py","file_name":"ex1.py","file_ext":"py","file_size_in_byte":4438,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"16396095209","text":"'''\nm = n* sum(x*y) - sum(x)-sum(y) / n*sum(x**2) - sum(x) **2\n\nb = sum(y) - m*sum(x) / n\n\nn : representa el valor de la cantidad de datos que nos dan\n\nm : pendiente de nuestra ecuacion lineal\n\nb : valor de origen de coordenadas\n\n\n\n'''\n\n\n\n\n\nimport pandas as pd\nimport numpy as np\nimport sympy \n\nx = sympy.Symbol('x')\n\n# datos: lista de valores (x, y)\ndef min_cuadrado(lista_x, lista_y):\n\n if len(lista_x) != len(lista_y):\n print(\"Las longitudes de la listas no son iguales\")\n else:\n n = len(lista_x)\n #convierto a tipo Numpy las listas dadas por parametro, ya que es mas\n #facil de laburar\n values_x = np.array(lista_x)\n values_y = np.array(lista_y)\n\n\n #calculo las sumatorias\n sum_x = np.sum(values_x)\n sum_y = np.sum(values_y)\n sum_xy = np.sum(values_x*values_y)\n sum_x_x = np.sum(values_x*values_x)\n\n\n #utilizo las formulas para obtener 'm' y 'b'\n #m = n* sum(x*y) - sum(x)-sum(y) / n*sum(x**2) - sum(x) **2\n \n m = ((n * sum_xy) - (sum_x * sum_y)) / ((n * sum_x_x) - (sum_x**2))\n \n #b = sum(y) - m*sum(x) / n\n\n b = (sum_y - m * sum_x) / n\n\n\n return(m * x) + b\n\n\n\nlista_x = [1, 2, 3, 4]\nlista_y = [1.4, 1.1, 0.7, 0.1]\n\n\necuacion = min_cuadrado (lista_x, lista_y)\nprint(ecuacion)\n","repo_name":"T0p2/Practica-de-lenguajes","sub_path":"Practica_nueva/Metodo_min_cuadrado.py","file_name":"Metodo_min_cuadrado.py","file_ext":"py","file_size_in_byte":1302,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"33853568379","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\n ©\n Author: Karim Makki\n\"\"\"\n\nimport trimesh\nimport numpy as np\nimport nibabel as nib\nimport os\nfrom scipy.ndimage.filters import gaussian_filter\nimport argparse\nimport skfmm\nfrom skimage import measure\nimport timeit\nimport fast_Gaussian_curvature_3D as g3D\n\n\n### Distance calculation limited to narrow band\n\ndef phi_narrow(mask, band=5):\n\n tmp = np.ones(mask.shape)\n tmp[mask!=0]= -1\n sgd = np.array(skfmm.distance(tmp, narrow=band), float)\n R = np.where(sgd != 0)\n sgd[sgd == 0] = 1\n\n return sgd, R\n\ndef local_gaussian_filter(scalar_function, sigma=2):\n\n mask = np.zeros(scalar_function.shape)\n mask[scalar_function!=0] = 1\n smooth_scalar_function = gaussian_filter(scalar_function*mask, sigma=sigma)\n\n return smooth_scalar_function\n\ndef hessian_adjoint_narrowband(hessian,R):\n\n Ha = np.zeros(hessian.shape)\n Ha[0,0,R[0],R[1],R[2]] = hessian[1,1,R[0],R[1],R[2]]*hessian[2,2,R[0],R[1],R[2]] - hessian[1,2,R[0],R[1],R[2]]*hessian[2,1,R[0],R[1],R[2]]\n Ha[0,1,R[0],R[1],R[2]] = hessian[1,2,R[0],R[1],R[2]]*hessian[2,0,R[0],R[1],R[2]] - hessian[1,0,R[0],R[1],R[2]]*hessian[2,2,R[0],R[1],R[2]]\n Ha[0,2,R[0],R[1],R[2]] = hessian[1,0,R[0],R[1],R[2]]*hessian[2,1,R[0],R[1],R[2]] - hessian[1,1,R[0],R[1],R[2]]*hessian[2,0,R[0],R[1],R[2]]\n\n Ha[1,0,R[0],R[1],R[2]] = hessian[0,2,R[0],R[1],R[2]]*hessian[2,1,R[0],R[1],R[2]] - hessian[0,1,R[0],R[1],R[2]]*hessian[2,2,R[0],R[1],R[2]]\n Ha[1,1,R[0],R[1],R[2]] = hessian[0,0,R[0],R[1],R[2]]*hessian[2,2,R[0],R[1],R[2]] - hessian[0,2,R[0],R[1],R[2]]*hessian[2,0,R[0],R[1],R[2]]\n Ha[1,2,R[0],R[1],R[2]] = hessian[0,1,R[0],R[1],R[2]]*hessian[2,0,R[0],R[1],R[2]] - hessian[0,0,R[0],R[1],R[2]]*hessian[2,1,R[0],R[1],R[2]]\n\n Ha[2,0,R[0],R[1],R[2]] = hessian[0,1,R[0],R[1],R[2]]*hessian[1,2,R[0],R[1],R[2]] - hessian[0,2,R[0],R[1],R[2]]*hessian[1,1,R[0],R[1],R[2]]\n Ha[2,1,R[0],R[1],R[2]] = hessian[1,0,R[0],R[1],R[2]]*hessian[0,2,R[0],R[1],R[2]] - hessian[0,0,R[0],R[1],R[2]]*hessian[1,2,R[0],R[1],R[2]]\n Ha[2,2,R[0],R[1],R[2]] = hessian[0,0,R[0],R[1],R[2]]*hessian[1,1,R[0],R[1],R[2]] - hessian[0,1,R[0],R[1],R[2]]*hessian[1,0,R[0],R[1],R[2]]\n\n return Ha\n\n\ndef L2_norm_grad_narrowband(gx,gy,gz,R):\n\n norm_grad = np.zeros(gx.shape)\n norm_grad[R] = np.sqrt(gx[R]**2 + gy[R]**2 + gz[R]**2)\n norm_grad = local_gaussian_filter(norm_grad, sigma=1)\n norm_grad[np.where(norm_grad==0)]=1 # just to avoid dividing by zero\n\n return norm_grad\n\ndef hessian_trace_narrowband(hessian,R):\n\n return hessian[0,0,R[0],R[1],R[2]] + hessian[1,1,R[0],R[1],R[2]] + hessian[2,2,R[0],R[1],R[2]]\n\n\ndef curvatures_narrowband(phi_grad,Ha,hessian,R):\n\n gx, gy, gz = phi_grad\n norm = L2_norm_grad_narrowband(gx,gy,gz,R)\n gx /= norm\n gy /= norm\n gz /= norm\n\n gaussian_curv = np.zeros(gx.shape)\n mean_curv = np.zeros(gx.shape)\n\n gaussian_curv[R] = gx[R] * (gx[R]*Ha[0,0,R[0],R[1],R[2]]+gy[R]*Ha[1,0,R[0],R[1],R[2]]+\\\n gz[R]*Ha[2,0,R[0],R[1],R[2]]) + gy[R] *(gx[R]*Ha[0,1,R[0],R[1],R[2]]+gy[R]*\\\n Ha[1,1,R[0],R[1],R[2]]+gz[R]*Ha[2,1,R[0],R[1],R[2]])+ gz[R] * (gx[R]*Ha[0,2,R[0],R[1],R[2]]\\\n +gy[R]*Ha[1,2,R[0],R[1],R[2]]+gz[R]*Ha[2,2,R[0],R[1],R[2]])\n\n gaussian_curv[R] /= L2_norm_grad_narrowband(gx,gy,gz,R)[R]**4\n\n mean_curv[R] = (gx[R] * (gx[R]*hessian[0,0,R[0],R[1],R[2]]+gy[R]*hessian[1,0,R[0],R[1],R[2]]+gz[R]*hessian[2,0,R[0],R[1],R[2]]) + \\\n gy[R] * (gx[R]*hessian[0,1,R[0],R[1],R[2]]+gy[R]*hessian[1,1,R[0],R[1],R[2]]+gz[R]*hessian[2,1,R[0],R[1],R[2]])\\\n + gz[R] * (gx[R]*hessian[0,2,R[0],R[1],R[2]]+gy[R]*hessian[1,2,R[0],R[1],R[2]]+gz[R]*hessian[2,2,R[0],R[1],R[2]])) \\\n - (L2_norm_grad_narrowband(gx,gy,gz,R)[R]**2 * hessian_trace_narrowband(hessian,R))\n\n mean_curv[R] /= -2*L2_norm_grad_narrowband(gx,gy,gz,R)[R]**3\n\n return gaussian_curv, mean_curv\n\ndef principal_curvatures(K_M, K_G):\n\n tmp = np.sqrt(np.absolute(K_M**2- K_G))\n k1 = K_M - tmp\n k2 = K_M + tmp\n\n return k1, k2\n\n\ndef save_result(verts,curv,save_path):\n\n res = np.append(verts,curv[...,None],axis=1)\n np.save(save_path, res)\n\n\nif __name__ == '__main__':\n\n parser = argparse.ArgumentParser()\n parser.add_argument('-in', '--mask', help='3D shape binary mask, as NIFTI file', type=str, required = True)\n parser.add_argument('-o', '--output', help='output directory', type=str, default = './Gaussian_curvature_results3D')\n\n args = parser.parse_args()\n\n # Example of use : python3 curvatures_narrowband.py -in ./3D_data/stanford_bunny_binary.nii.gz -o /home/karim/Bureau/Courbure/narrow_band\n\n output_path = args.output\n\n if not os.path.exists(output_path):\n os.makedirs(output_path)\n\n shape = nib.load(args.mask).get_data()\n\n start_time = timeit.default_timer()\n\n shape, dx, dy, dz = g3D.bbox_3D(shape,5)\n\n phi, R = phi_narrow(shape,5) ## distance calculation limited to narrow band\n phi = local_gaussian_filter(phi, sigma=2) ## smoothing of the level set signed distance function on a narrow band\n\n#################### Computation of Gaussian and mean curvatures ###############################\n phi_grad, hessian = g3D.hessian(phi) ### To do in narrowband\n Ha = hessian_adjoint_narrowband(hessian,R)\n Gaussian_curvature, mean_curvature = curvatures_narrowband(phi_grad,Ha,hessian,R)\n#################################################################################################\n#################### Computation of principal curvatures #######################################\n K1, K2 = principal_curvatures(mean_curvature, Gaussian_curvature)\n#################################################################################################\n # extract explicitly the implicit surface mesh using the scikit-image toolbox\n\n verts, faces, normals, values = measure.marching_cubes_lewiner(phi, 0.0)\n print(verts.shape)\n\n ### Affect per-vertex curvature values, with a nearest neighbour interpolation of vertices on the grid\n gaussian_curv = g3D.texture_spline_interpolation3D(verts, Gaussian_curvature)\n mean_curv = g3D.texture_spline_interpolation3D(verts, mean_curvature)\n k1 = g3D.texture_spline_interpolation3D(verts, K1)\n k2 = g3D.texture_spline_interpolation3D(verts, K2)\n\n elapsed = timeit.default_timer() - start_time\n print(\"The proposed method takes (in seconds):\\n\")\n print(elapsed)\n\n verts = g3D.align_origin_back(verts,dx,dy,dz)\n m = trimesh.Trimesh(vertices=verts, faces=faces)\n m.export(os.path.join(output_path, \"surface_mesh.obj\"))\n\n #### Save results as numpy array arrays\n\n save_result(verts,gaussian_curv,os.path.join(output_path,\"gaussian_curv.npy\"))\n save_result(verts,mean_curv,os.path.join(output_path,\"mean_curv.npy\"))\n save_result(verts,k1,os.path.join(output_path,\"min_curv.npy\"))\n save_result(verts,k2,os.path.join(output_path,\"max_curv.npy\"))\n save_result(verts,2*gaussian_curv,os.path.join(output_path,\"Ricci_scalar.npy\"))\n\n ## Display results\n\n g3D.display_mesh(verts, faces, normals, gaussian_curv, os.path.join(output_path, \"Gaussian_curvature_Makki.png\"))\n g3D.display_mesh(verts, faces, normals, 2*gaussian_curv, os.path.join(output_path, \"Ricci_scalar_Makki.png\"))\n g3D.display_mesh(verts, faces, normals, mean_curv, os.path.join(output_path, \"mean_curvature_Makki.png\"))\n g3D.display_mesh(verts, faces, normals, k1, os.path.join(output_path, \"Minimum_curvature_Makki.png\"))\n g3D.display_mesh(verts, faces, normals, k2, os.path.join(output_path, \"Maximum_curvature_Makki.png\"))\n","repo_name":"k16makki/Medima_tools","sub_path":"curvatures_narrowband.py","file_name":"curvatures_narrowband.py","file_ext":"py","file_size_in_byte":7523,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"72235041503","text":"from lib.loggers import Logger\n\nlogger = Logger()\n\n\nclass Cache(dict):\n DEBUG = False\n\n def __str__(self):\n representation = (\n \"{ \" + \", \".join([f\"'{key}': ...\" for key in self.keys()]) + \"}\"\n )\n\n return representation\n\n def delete(self, key: str) -> None:\n if self.DEBUG:\n logger.info(f\"Deleting key '{key}' from cache.\")\n\n del self[key]\n","repo_name":"asynched/search-engine-api","sub_path":"src/lib/cache/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":406,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"21309595496","text":"import sys\nimport configparser\nimport logging\nimport os\nimport requests\nimport smtplib\nimport picamera\nfrom time import sleep\nfrom email.mime.text import MIMEText\nimport RPi.GPIO as GPIO\nimport time\nfrom datetime import datetime, time\nimport subprocess\nimport socket\nimport threading\nfrom threading import Thread\nimport shutil\n\nconfig = configparser.ConfigParser()\nconfig.read('webcam.ini')\n\nlogger = logging.getLogger(__name__)\nhandler = logging.FileHandler(config['DEFAULT']['log.file'])\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nhandler.setFormatter(formatter)\nlogger.addHandler(handler)\nlogger.addHandler(logging.StreamHandler())\nlogger.setLevel(logging.ERROR)\nsnapshot_interval = int(config['DEFAULT']['camera.snapshot.interval'])\n\nclass Webcam:\n\n def get_file_ip_address(self):\n file_name = 'ipaddress'\n if os.path.isfile(file_name):\n file = open('ipaddress', 'r')\n ip_address = file.read()\n file.close()\n logger.debug('file ip address is %s' % ip_address)\n return ip_address\n\n def set_file_ip_address(self, ip_address):\n file_name = 'ipaddress'\n if os.path.isfile(file_name):\n file = open('ipaddress', 'w')\n file.write(ip_address)\n file.close()\n logger.info('changed file ip address to %s' % ip_address)\n\n def get_router_ip_address(self):\n request = requests.get('http://ipecho.net/plain')\n if request.status_code == 200:\n ip_address = request.text\n logger.debug('router ip address is %s' % ip_address)\n return ip_address\n\n def send_email(self, subject, message):\n fromaddr = config['DEFAULT']['mail.from']\n toaddrs = config['DEFAULT']['mail.to']\n\n msg = MIMEText(message)\n msg['Subject'] = subject\n msg['From'] = fromaddr\n msg['To'] = toaddrs\n\n username = config['DEFAULT']['smtp.username']\n password = config['DEFAULT']['smtp.password']\n\n server = smtplib.SMTP(\"smtp.gmail.com\", 587)\n server.starttls()\n server.login(username, password)\n server.sendmail(fromaddr, toaddrs, msg.as_string())\n server.quit()\n\n def update_noip(self, ip_address):\n url = 'https://dynupdate.no-ip.com/nic/update?hostname=%s&myip=%s' % (config['DEFAULT']['noip.hostname'], ip_address)\n headers = {'Authorization': 'Basic %s' % config['DEFAULT']['noip.authorization'], 'User-Agent': '%s' % config['DEFAULT']['noip.useragent']}\n request = requests.get(url, headers=headers)\n if request.status_code == 200:\n logger.info('changed no-ip address to %s' % ip_address)\n return True\n else:\n logger.error('problem updating no-ip address: %s' % request.text)\n return False\n\n def start_camera(self):\n camera = picamera.PiCamera()\n camera.resolution = (1024, 768)\n camera.exposure_mode = 'sports'\n camera.vflip = True\n camera.exposure_mode = 'auto'\n camera.metering = 'average'\n sleep(3)\n\n snapshot = Snapshot(camera)\n snapshot.start()\n\n stream_server = StreamServer(camera)\n stream_server.start()\n\nclass Snapshot(Thread):\n\n def __init__(self, camera):\n Thread.__init__(self)\n self.camera = camera\n\n def run(self):\n\n while True:\n now = datetime.now()\n now_time = now.time()\n if now_time >= time(6,00) and now_time <= time(21,30):\n timestamp = now.strftime('%m/%d %H:%M:%S')\n timestamp += \" (\" + self.get_temp() + \")\"\n self.camera.annotate_text = timestamp\n\n try:\n if self.camera.recording:\n logger.debug('take snapshot using video port')\n self.camera.capture('/var/www/html/camera.jpg', use_video_port=True)\n else:\n logger.debug('take snapshot')\n self.camera.capture('/var/www/html/camera.jpg')\n\n shutil.copy('/var/www/html/camera.jpg', '/var/www/html/timelapse/camera-%s.jpg' % now.strftime('%Y-%m-%d_%H:%M:%S'))\n except:\n logger.error('unexpected snapshot error: ', sys.exc_info()[0])\n\n sleep(snapshot_interval)\n\n def get_temp(self):\n try:\n output = subprocess.check_output([\"/opt/vc/bin/vcgencmd\", \"measure_temp\"])\n text = output.decode('utf-8')\n return text[5:-3]\n except:\n return \"\"\n\nclass StreamServer(Thread):\n\n def __init__(self, camera):\n Thread.__init__(self)\n self.camera = camera\n\n def run(self):\n logger.info('stream server is started')\n\n server_socket = socket.socket()\n server_socket.bind(('0.0.0.0', 8000))\n server_socket.listen(0)\n\n try:\n while True:\n logger.debug('waiting for connection')\n connection = server_socket.accept()[0]\n\n logger.debug('create stream')\n stream = Stream(self.camera, connection)\n stream.start()\n finally:\n logger.info('stream server is stopped')\n server_socket.close()\n\nclass Stream(Thread):\n\n def __init__(self, camera, connection):\n Thread.__init__(self)\n self.camera = camera\n self.connection = connection\n\n def run(self):\n logger.debug('start recording')\n file = self.connection.makefile('wb')\n self.camera.start_recording(file, format='h264')\n\n try:\n while True:\n self.camera.wait_recording(60)\n except (ConnectionResetError):\n logger.warn('connection reset error')\n self.camera.stop_recording()\n except:\n logger.error('unexpected streaming error: ', sys.exc_info()[0])\n self.camera.stop_recording()\n self.connection.close()\n file.close()\n","repo_name":"jeffjohnston/webcam","sub_path":"webcam.py","file_name":"webcam.py","file_ext":"py","file_size_in_byte":6049,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"30098600219","text":"\"\"\" 210616 16:00 항해 37번 by 신영\"\"\"\n\"\"\"\n계단 오르기 게임은 계단 아래 시작점부터 계단 꼭대기에 위치한 도착점까지 가는 게임이다. 각각의 계단에는 일정한 점수가 쓰여 있는데 계단을 밟으면 그 계단에 쓰여 있는 점수를 얻게 된다.\n계단 오르는 데는 다음과 같은 규칙이 있다.\n계단은 한 번에 한 계단씩 또는 두 계단씩 오를 수 있다. 즉, 한 계단을 밟으면서 이어서 다음 계단이나, 다음 다음 계단으로 오를 수 있다.\n연속된 세 개의 계단을 모두 밟아서는 안 된다. 단, 시작점은 계단에 포함되지 않는다.\n마지막 도착 계단은 반드시 밟아야 한다.\n이 게임에서 얻을 수 있는 총 점수의 최댓값을 구하는 프로그램을 작성하시오.\n입력\n입력의 첫째 줄에 계단의 개수가 주어진다.\n둘째 줄부터 한 줄에 하나씩 제일 아래에 놓인 계단부터 순서대로 각 계단에 쓰여 있는 점수가 주어진다. 계단의 개수는 300이하의 자연수이고, 계단에 쓰여 있는 점수는 10, 000이하의 자연수이다.\n출력\n첫째 줄에 계단 오르기 게임에서 얻을 수 있는 총 점수의 최댓값을 출력한다.\n입력 \n6\n10\n20\n15\n25\n10\n20\n출력 75\n\"\"\"\n# from sys import stdin\n# stairs = int(stdin.readline())\n\n# for _ in range(stairs):\n # score = int(stdin.readline())\n # scoreLst.append(score)\n\nstairs = 6\nscoreLst = [10, 20, 15, 25, 10, 20]\nscoreLst.reverse()\n# scoreLst.append(0)\n\nv1Score = 20\nv2Score = 20\nindexV1 = 1\nindexV2 = 2\n\nwhile indexV1 <= len(scoreLst):\n v1Score += scoreLst[indexV1]\n indexV1 += 2\n if indexV1 >= len(scoreLst):\n break\n v1Score += scoreLst[indexV1]\n indexV1 += 1\n\nwhile indexV2 <= len(scoreLst):\n v2Score += scoreLst[indexV2]\n indexV2 += 1\n if indexV1 >= len(scoreLst):\n break\n v2Score += scoreLst[indexV2]\n indexV2 += 2\n\nprint(max(v1Score, v2Score))\n","repo_name":"FrancisJeon/Baekjoon-python","sub_path":"ing/BOJ_2579.py","file_name":"BOJ_2579.py","file_ext":"py","file_size_in_byte":1975,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"20517576017","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Nov 25 10:59:17 2020\r\n\r\n@author: sandhya chettiar\r\n\"\"\"\r\n\r\nfrom skimage import data\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\n\r\n#Image reading\r\n\r\nimg = data.text()\r\nplt.gray()\r\n\r\n\r\nm, n = img.shape\r\n\r\nmask = [[1,2,1],\r\n [2,4,2],\r\n [1,2,2]]\r\nmask = np.array(mask)\r\n\r\nmask = mask/16\r\n\r\n\r\nimg_new = np.zeros([m, n])\r\n\r\nfor i in range(1,m-1):\r\n for j in range(1,n-1):\r\n temp = img[i-1,j-1]*mask[2-0,2-0] + img[i-1,j]*mask[2-0,2-1] + img[i-1,j+1]*mask[2-0,2-2] + img[i,j-1]*mask[2-1,2-0] + img[i,j]*mask[2-1,2-1] + img[i,j+1]*mask[2-1,2-2] + img[i+1,j-1]*mask[2-2,2-0] + img[i+1,j]*mask[2-2,2-1] +img[i+1,j+1]*mask[2-2,2-2]\r\n\r\n img_new[i,j] = temp\r\n \r\n\r\n\r\nimg_new = img_new.astype(np.uint8)\r\nplt.figure(1)\r\nplt.subplot(1,2,1)\r\nplt.imshow(img)\r\n\r\nplt.subplot(1,2,2)\r\nplt.imshow(img_new)\r\n\r\nplt.show()\r\n","repo_name":"Sandhya18Chettiar/Digital-Image-Processing-Lab","sub_path":"gaussian.py","file_name":"gaussian.py","file_ext":"py","file_size_in_byte":895,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"2229732009","text":"#coding: utf-8\nfrom __future__ import unicode_literals\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import division\n\nimport os\n\nfrom .settings_base import *\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.postgresql_psycopg2',\n 'NAME': os.environ['DATABASE_NAME'],\n 'USER': os.environ['DATABASE_USERNAME'],\n 'PASSWORD': os.environ['DATABASE_PASSWORD'],\n 'HOST': os.environ['DATABASE_HOST'],\n }\n}\n\nEMAIL_HOST = 'smtp.yandex.ru'\nEMAIL_HOST_USER = 'noreply@erofeimarkov.ru'\nEMAIL_HOST_PASSWORD = os.environ['SMTP_PASSWORD']\nEMAIL_PORT = 587\nEMAIL_USE_TLS = True\n\nDEBUG = False\n\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': True,\n 'formatters': {\n 'verbose': {\n 'format': '%(asctime)s PID#%(process)d [%(levelname)s] %(name)s: %(message)s at %(pathname)s line %(lineno)d'\n },\n },\n 'handlers': {\n 'default': {\n 'level': 'INFO',\n 'class': 'logging.FileHandler',\n 'filename': '/var/log/uwsgi/django.log',\n 'formatter': 'verbose',\n },\n },\n 'loggers': {\n 'django': {\n 'handlers': ['default'],\n 'level': 'DEBUG',\n 'propagate': False,\n },\n '': {\n 'handlers': ['default'],\n 'level': 'DEBUG',\n },\n },\n}\n","repo_name":"migunow/erofeimarkov","sub_path":"app/erofeimarkov/settings_prod.py","file_name":"settings_prod.py","file_ext":"py","file_size_in_byte":1392,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"1491214083","text":"from unittest import mock\n\nimport pandas as pd\nimport pytest\n\nfrom metadata.core import Metadata\nfrom pipeline_dataops.extract.core import from_api\n\n\n@pytest.fixture\ndef mocked_requests():\n with mock.patch(\"requests.get\") as mock_get:\n yield mock_get\n\n\n@pytest.mark.skip(\n reason=\"\"\"\n Not implemented yet.\n This test, 'test_from_api', is an integration test because it tests\n the interaction between the 'from_api' function and the actual API endpoint.\n It checks if the function correctly handles responses from the API and\n properly processes the raw data. Therefore, it involves the integration\n of multiple components of the system, and not just the behavior of\n individual units in isolation.\n \"\"\"\n)\ndef test_from_api(sample_raw_df):\n assert True\n","repo_name":"gao-hongnan/TheBareOps","sub_path":"pipeline-dataops/tests/integration/extract/test_extract.py","file_name":"test_extract.py","file_ext":"py","file_size_in_byte":792,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"1783654694","text":"from flask import Flask, request, jsonify, abort, redirect, url_for, render_template, send_file, flash\nfrom bs4 import BeautifulSoup\nimport requests, statistics\nimport openpyxl\nfrom flask_wtf import FlaskForm\nfrom wtforms import StringField, FileField\nfrom wtforms.validators import DataRequired\nimport os\nfrom werkzeug.utils import secure_filename\nimport auto\nimport json\n\napp = Flask(__name__, static_url_path='/static')\n\napp.config.update(dict(\n SECRET_KEY=\"powerful secretkey\",\n WTF_CSRF_SECRET_KEY=\"a csrf secret key\"\n))\n\n \n@app.route('/')\ndef redir_submit():\n return redirect(url_for('index'))\n\nclass MyForm(FlaskForm):\n # name = 'name'\n file = FileField(validators=[DataRequired()]\n )\n# @app.route('/index', methods=['GET', 'POST'])\n# def form():\n\n# #Причем в начале проверяем наличие авторизации, если флага нет, то кидаем обработку 401 ошибки и не даем работать с прогой\n# # if not session.get('logged_in'):\n# # abort(401)\n\n# \"\"\"\n# Тут делаем что-то полезное в случае успешной авторизации\n# \"\"\"\n\n# return render_template('index.html')\n\n\n@app.route('/submit', methods=('GET', 'POST'))\ndef submit():\n form = MyForm()\n \n if form.validate_on_submit():\n f = form.file.data\n filename = 'analogi.xlsx'\n f.save(os.path.join(filename))\n auto.autoru_appraiser(filename)\n \n\n return send_file(filename,\n mimetype='xlsx',\n # attachment_filename=filename,\n as_attachment=True)\n \n return render_template('submit.html', form=form)\n\n\n@app.errorhandler(500)\ndef page_not_found(e):\n error = 'Произошла ошибка при работе скрипта, вероятно auto.ru опять показывает капчу. Сообщите об этом Алексею прямо сейчас.'\n return render_template('index.html', error=error), 500\n\n\ndef get_dropdown_values():\n\n \"\"\"\n dummy function, replace with e.g. database call. If data not change, this function is not needed but dictionary\n could be defined globally\n \"\"\"\n with open('marks_and_models.json', 'r', encoding='utf-8') as fh: #открываем файл на чтение\n class_entry_relations = json.load(fh)\n\n return class_entry_relations\n\n\n@app.route('/_update_dropdown')\ndef update_dropdown():\n\n # the value of the first dropdown (selected by the user)\n selected_class = request.args.get('selected_class', type=str)\n\n # get values for the second dropdown\n updated_values = get_dropdown_values()[selected_class]\n\n # create the value sin the dropdown as a html string\n html_string_selected = ''\n for entry in updated_values:\n html_string_selected += ''.format(entry, entry)\n\n return jsonify(html_string_selected=html_string_selected)\n\n\n@app.route('/_process_data')\ndef process_data():\n selected_class = request.args.get('selected_class', type=str)\n selected_entry = request.args.get('selected_entry', type=str)\n\n # process the two selected values here and return the response; here we just create a dummy string\n\n return jsonify(random_text=\"you selected {} and {}\".format(selected_class, selected_entry))\n\n\n@app.route('/index', methods=('GET', 'POST'))\ndef index():\n\n \"\"\"\n Initialize the dropdown menues\n \"\"\"\n form = MyForm()\n \n if request.method == 'POST':\n filename = 'shablon.xlsx'\n f = openpyxl.load_workbook(filename)\n sheetobject = f['Объекты оценки']\n sheetobject.cell(row=2, column=1).value = request.form['all_classes']\n sheetobject.cell(row=2, column=2).value = request.form['all_entries']\n sheetobject.cell(row=2, column=3).value = request.form['obj_year']\n sheetobject.cell(row=2, column=4).value = request.form['obj_engvol']\n sheetobject.cell(row=2, column=5).value = request.form['obj_hp']\n sheetobject.cell(row=2, column=9).value = request.form['obj_kpp']\n sheetobject.cell(row=2, column=7).value = request.form['obj_gear']\n sheetobject.cell(row=2, column=6).value = request.form['obj_eng']\n sheetobject.cell(row=2, column=8).value = request.form['obj_tob']\n sheetobject.cell(row=2, column=11).value = request.form['obj_mileage']\n filename = 'analogi.xlsx'\n \n f.save(os.path.join(filename))\n \n auto.autoru_appraiser(filename)\n \n\n return send_file(filename,\n mimetype='xlsx',\n # attachment_filename=filename,\n as_attachment=True)\n\n class_entry_relations = get_dropdown_values()\n\n default_classes = sorted(class_entry_relations.keys())\n default_values = class_entry_relations[default_classes[0]]\n\n return render_template('index.html',\n all_classes=default_classes,\n all_entries=default_values)\n\nif __name__ == \"__main__\":\n app.run(debug=True)","repo_name":"DemalexS/appreser","sub_path":"hello.py","file_name":"hello.py","file_ext":"py","file_size_in_byte":5164,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"4615108400","text":"import re\nimport subprocess\n\n\nbucket_path = \"s3://zapata-zmachine/projects/bp-combustion-project/phase-2/mpea-runtime/data/publication-data/cisd_rdms/\"\naws_profile = \"--profile zapata-zmachine\"\nls_cmd = \"aws s3 ls \"\ncp_cmd = \"aws s3 cp \"\nuntar_cmd = \"tar zxvf \"\ntar_cmd = \"tar zcvf \"\n\n\ndef compress_and_upload(target, result, path=None):\n \"\"\"Little function to compress a file with tar zcvf and\n upload it to amazon with path appended to the bucket_path\n Args:\n target: the path of the target file to compress and upload\n result: the path where to store the compressed file\n path: Optionally, the subfolder where to upload the file in the\n S3 bucket\n \"\"\"\n\n compress_cmd = tar_cmd + result + \" \" + target\n print(compress_cmd)\n subprocess.check_output(compress_cmd, shell=True)\n\n upload_cmd = cp_cmd + result + \" \" + bucket_path + path + \" \" + aws_profile\n print(upload_cmd)\n subprocess.check_output(upload_cmd, shell=True)\n\n\ndef download_and_uncompress(target):\n \"\"\"Little function to download a file from AWS S3\n and uncompress it with tar zxvf.\n Args:\n target: The path to append to the bucket path to get the file\n \"\"\"\n\n download_cmd = cp_cmd + bucket_path + target + \" . \" + aws_profile\n print(download_cmd)\n subprocess.check_output(download_cmd, shell=True)\n\n filename = re.sub(r\".*/\", \"\", target)\n\n uncompress_cmd = untar_cmd + filename\n print(uncompress_cmd)\n subprocess.check_output(uncompress_cmd, shell=True)\n\n\n# @profile\ndef main():\n\n mol_search = re.compile(\".*molecule: '([A-Z0-9]*)'\")\n act_search = re.compile(\".*nactiveorbs: '([A-Z0-9]*)'\")\n type_search = re.compile(\".*orb-type: '([ A-Za-z0-9]*)'\")\n path_search = re.compile(\".*hamiltonian: '([^']*)'\")\n missing_rdms = set()\n found_rdms = set()\n n_rdms = 0\n with open(\"hamiltonian_data\", \"r\") as f:\n for line in f:\n match_mol = mol_search.search(line)\n match_act = act_search.search(line)\n match_type = type_search.search(line)\n if match_type:\n n_rdms += 1\n molname = match_mol.group(1)\n nactiveorbs = match_act.group(1)\n orbtype = match_type.group(1)\n\n if orbtype == \"MP2 NOs\":\n orbtype = \"NOs\"\n newname = molname + \"-\" + nactiveorbs + \"-\" + orbtype + \"-cisd-rdm.tgz\"\n\n ham_path = path_search.search(line).group(1)\n print(ham_path)\n folder_path = re.sub(\"hamiltonian.tgz\", \"\", ham_path)\n ls_folder = subprocess.check_output(\n ls_cmd + \"s3://zapata-zmachine/\" + folder_path + \" \" + aws_profile,\n shell=True,\n )\n if \"cisd-rdm\" not in str(ls_folder):\n print(f\"CISD RDM missing for {molname}, {nactiveorbs}, {orbtype}.\")\n missing_rdms.add(newname)\n else:\n found_rdms.add(newname)\n cisd_cp_cmd = (\n \"aws s3api copy-object --copy-source \"\n + \"zapata-zmachine/\"\n + folder_path\n + \"cisd-rdms.tgz\"\n + \" --bucket zapata-zmachine --key projects/bp-combustion-project/phase-2/mpea-runtime/data/publication-data/cisd_rdms/\"\n + newname\n + \" \"\n + aws_profile\n )\n cp_output = subprocess.check_output(cisd_cp_cmd, shell=True)\n\n assert len(missing_rdms) + len(found_rdms) == n_rdms\n print(\"The missing RDMs are:\")\n print(missing_rdms)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"yannnbingz/madtequila-benchmark","sub_path":"grouping_data/Gather_CISD_RDMs.py","file_name":"Gather_CISD_RDMs.py","file_ext":"py","file_size_in_byte":3761,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"19469030160","text":"#This method goes through the string and returns the index of the nextOperand.\n#If there isn't another operand None is returned\ndef findNextOperand(string):\n for i in range (0, len(string)):\n if isOperand(string[i]):\n return i\n#This method finds the end of a segment. A segment is defined as one number, fraction, or mixed number\n#so it just finds the index of the first segment breaker (anything in the string terminators). If one \n#is not found the length of the string is returned.\ndef findSegmentEnd(string):\n terminators = '+-/*^()'\n for i in range (0, len(string)):\n if string[i] in terminators:\n return i\n return len(string)\n#This finds the end of segment in a similar manor to findSegmentEnd, but it is used before spaces are removed\n#so spaces are considered a segment breaker. However, in a mixed number a space is mandatory so it determines\n#if the space is just part of a mixed number to prevent false positives. If there is no segment end in the string\n#the length of the string is returned.\ndef findSegmentEndWithSpaces(string):\n terminators = '+-/*^() '\n for i in range (0, len(string)):\n if string[i] in terminators:\n if(string[i] == \" \"):\n if followedByFraction(string[i+1:]):\n return i\n else:\n return i\n\n return len(string)\n\n#This method determines if before the next operator or parentheses there is a \\ indicating a fraction. \n#This is used to ensure spaces are not removed that belong to a mixed number.\ndef followedByFraction(string):\n terminators = '+-/*^() '\n for i in range (0, len(string)):\n if string[i] in terminators:\n return False\n elif string[i] == \"\\\\\":\n return True\n return False\n\n#This method goes through the string and converts all negatives into (0-negative) to make the string parsing\n#of converting into custom classes simpler.\ndef fixNegatives(string):\n newString = \"\"\n forValue = 0\n if string[0] == \"-\":\n newString += \"(0-\"\n indexOfNumberEnd = findSegmentEndWithSpaces(string[1:]) + 1\n portion = string[1:indexOfNumberEnd]\n newString += portion\n newString += \")\"\n forValue = indexOfNumberEnd\n while forValue < len(string) :\n if (string[forValue] == \" \" or string[forValue] == \"(\") and string[forValue+1] == \"-\" and (isNumber(string[forValue + 2]) or string[forValue+2] == \".\"):\n if(string[forValue] == \"(\"):\n newString += \"(\"\n indexOfTermEnd = findSegmentEndWithSpaces(string[forValue+2:]) + forValue + 2\n term = string[forValue+2:indexOfTermEnd]\n newString = newString + \"(0-\" + term + \")\"\n forValue = indexOfTermEnd\n else:\n newString += string[forValue]\n forValue += 1\n return newString\n\n\n#This method goes through the string and removes any spaces unless it is followed by a fraction and that spaces is necessary.\n#This is to make the string parsing of converting terms into a custom class simpler.\ndef removeSpaces(string):\n newString = \"\"\n for i in range (0, len(string)):\n if string[i] != \" \":\n newString += string[i]\n else:\n if (isOperand(string[i-1]) == False and followedByFraction(string[i+1:])):\n newString += string[i]\n return newString\n\n#This method goes through the entire expression and converts it into the form \"fracList[0] operator fracList[1]...\". This is so\n#eval can be called on the expression even though it has fractions and mixed numbers. It does this by building a new String keeping\n#operators, parentheses, and converting the terms into a custom class and appending it onto fracList.\ndef fractionalize(expression):\n lastIndex = 0\n alreadyConverted = \"\"\n segmentEnd = 0\n fracList = []\n #Append first\n indexOfFirstOperand = findNextOperand(expression)\n subExpression = expression[:indexOfFirstOperand]\n for i in range (0, len(subExpression)):\n if subExpression[i] == \"(\" or subExpression[i] == \")\":\n alreadyConverted += subExpression[i]\n else:\n segmentEnd = findSegmentEnd(expression[i:]) + i\n firstTerm = subExpression[i:segmentEnd]\n fractionVersion = convertToFraction(firstTerm)\n fracList.append(fractionVersion)\n alreadyConverted += \"fracList[\" + str(len(fracList) - 1) + \"]\"\n break\n\n #Continue on with the rest of the expression\n for x in range (segmentEnd, len(expression)):\n if expression[x] == \")\":\n alreadyConverted += expression[x]\n elif isOperand(expression[x]):\n if (expression[x] == \"^\"):\n alreadyConverted += \"**\"\n else:\n alreadyConverted += expression[x]\n while ( expression[x+1] == \"(\" or expression[x+1] == \")\" ):\n alreadyConverted += expression[x+1]\n x+=1\n x+=1\n segmentEnd = findSegmentEnd(expression[x + 1:])\n term = expression[x:segmentEnd + x + 1]\n fractionVersion = convertToFraction(term)\n fracList.append(fractionVersion)\n alreadyConverted += \"fracList[\" + str(len(fracList) - 1) + \"]\"\n return alreadyConverted, fracList\n\n\n#This method determines if something is a number. If it is, True is returned. If not, False is returned. This is used in ConvertToFraction\ndef isNumber(string):\n try:\n float(string)\n return True\n except ValueError :\n return False\n\n\n#This method takes a String and converts it into either a decimal or fraction. If the term is a mixed number\n#this method will convert that into an impromper fraction.\ndef convertToFraction(string):\n if isNumber(string):\n \treturn decimal(float(string))\n elif \" \" in string:\n separator = string.find(\" \")\n numberPart = float(string[:separator])\n fractionPart = string[separator + 1 :]\n separator = fractionPart.find(\"\\\\\")\n numerator = float(fractionPart[:separator])\n denominator = float(fractionPart[separator + 1:])\n return fraction(numerator + numberPart * denominator, denominator)\n else:\n separator = string.find(\"\\\\\")\n numerator = float(string[:separator])\n denominator = float(string[separator + 1:])\n return fraction(numerator, denominator)\n\n#This method returns true if a character is an operator and false if it isn't\ndef isOperand(char):\n operands = '+-/*^'\n return char in operands\n\n#This method uses Euclids algorithim to find the GCD of two numbers.\ndef gcd (a , b) :\n a = float(a)\n b = float(b)\n if a == 0:\n return b\n else:\n return gcd(b % a, a)\n\n#This method finds the least common multiple by multipling the two numbers together (guaranteed to be a multiple)\n#and dividing that product by the gcd.\ndef lcm(a, b):\n return a * b / gcd(a,b)\n\n'''\nThe fraction class is used to hold both mixed numbers and fractions. Mixed numbers are stored as impromper fractions.\nThis class overides the standard operators, the toString method, the float method, and has a the helper methods simplifyToMixedNumber, \nsimplify, and reciprocal. \n'''\nclass fraction :\n\t#The constructor just takes the given parameters (numerator and denominator) and\n\t#sets the proper values within the class\n def __init__(self , Numerator=0, Denominator=1) :\n self.numerator = Numerator\n self.denominator = Denominator\n #This method returns a string representation of the fraction. It first simpifies the fraction (part of simplifytoMixedNumber)\n #and if possible converts it into a mixed number. It then returns the fraction in the form A B\\C, but if A is 0 it just returns B\\C\n #and if B is 0 it just returns A.\n def __str__(self) :\n if(self.denominator == 0):\n return \"Please Do Not Divide By Zero\"\n simplified = self.simplifyToMixedNumber()\n frac = simplified[1]\n if simplified[0] == 0:\n return str(simplified[1].numerator) + \"\\\\\" + str(simplified[1].denominator)\n elif simplified[1].numerator == 0 :\n return str(simplified[0])\n else:\n return str(simplified[0]) + \" \" + str(simplified[1].numerator) + \"\\\\\" + str(simplified[1].denominator)\n #This method returns a fraction of the product of the two given fractions.\n def __mul__(self , other) :\n product = fraction(self.numerator*other.numerator , self.denominator*other.denominator)\n return product\n #This method gets a float aproximation of the fraction by dividing the numerator by the denominator.\n def __float__(self) :\n return float(self.numerator)/self.denominator\n #This method gets the reciporcal by simplify flipping the numerator and denominator\n def reciprocal(self) :\n inverse = fraction(self.denominator , self.numerator)\n return inverse\n \t#This method subtracts two fractions by first putting them over a common denominator then subtracting. It returns\n \t#a fraction.\n def __sub__(self, other):\n leastCommonMultiple = lcm(self.denominator, other.denominator)\n fracOne = fraction((self.numerator) * (leastCommonMultiple / (self.denominator)) , leastCommonMultiple)\n fracTwo = fraction((other.numerator) * (leastCommonMultiple / other.denominator), leastCommonMultiple)\n return fraction(fracOne.numerator - fracTwo.numerator, leastCommonMultiple)\n #This method adds two fractions by first putting them over a common denominator then adding. It returns\n \t#a fraction.\n def __add__ (self, other):\n leastCommonMultiple = lcm(self.denominator, other.denominator)\n fracOne = fraction((self.numerator) * (leastCommonMultiple / (self.denominator)) , leastCommonMultiple)\n fracTwo = fraction((other.numerator) * (leastCommonMultiple / other.denominator), leastCommonMultiple)\n return fraction(fracOne.numerator + fracTwo.numerator , leastCommonMultiple)\n #This method divides two fractions by returning the product of the first fraction and the reciprocal of the second\n def __truediv__(self, other):\n recip = fraction(other.denominator, other.numerator)\n return self * recip\n #This method handles fractional exponents. It first determines if the given radical is invalid throwing an invalidRadical exception.\n #If not then it solves the fractional exponent. If the output is a decimal value it returns a decimal to make things clear for the user.\n def __pow__ (self, power):\n if float(self) < 0 and float(power) % 2 == 0:\n raise invalidRadical\n if float(power) >= 0:\n partOne = fraction(self.numerator ** (1/power.denominator), self.denominator ** (1/power.denominator))\n partTwo = fraction(partOne.numerator ** power.numerator, partOne.denominator ** power.numerator)\n if(float.is_integer(partTwo.numerator) and float.is_integer(partTwo.numerator)):\n return partTwo\n else:\n return decimal(float(partTwo))\n else:\n power = abs(power)\n partOne = fraction(self.numerator ** power.demoninator, self.denominator ** power.demoninator)\n partTwo = fraction(partOne.numerator ** (1/power.numerator), partOne.demoniator ** (1/power.numerator))\n return fraction.reciprocal(partTwo)\n #This method simplifies a fraction to the lowest possible numerator and denominator.\n def simplify(self):\n if(self.numerator == 0):\n return fraction(0, 1)\n else:\n greatestCommonDemoniator = gcd(self.numerator, self.denominator)\n return fraction(self.numerator / greatestCommonDemoniator, self.denominator / greatestCommonDemoniator)\n #This method converts a fraction into a mixed number with the fraction portion in simplest terms. It returns\n #a tuple of the preceeding number and the fraction.\n def simplifyToMixedNumber(self):\n if self.numerator >= self.denominator:\n numberPart = int(self.numerator / self.denominator)\n remainder = self - fraction(numberPart, 1)\n return numberPart , remainder.simplify()\n return 0, self.simplify()\n\n#This method takes in a number and finds the length of its mantessa by determining the length\n#of its string version after the period.\ndef lenMantessa (number):\n stringVersion = str(number)\n indexOfDecimal = stringVersion.find(\".\")\n return len(stringVersion[indexOfDecimal + 1:])\n'''\nThe custom class decimal inherits from fraction and is used for non-fraction/mixed number entries.\nThe purpose of this is such that all decimals are compatible with fractions (so an error isn't called\nwhen doing an operation on a fraction and a decimal). As well because a custom class is used custome exceptions\ncan be thrown such as the invalidRadical. \n'''\nclass decimal (fraction):\n\t#The constructor in essence converts each decimal into a fraction that has a numerator and denominator. \n\t#However, by using a custom class instead of just making every decimal fraction the toString can output\n\t#a more appropriate result. This method works by finding the length of the mantissa, multipling the decimal\n\t# times 10 ^ (length of mantessa) and making the denominator 10 ^ length of the mantissa. The fraction is then simplified.\n def __init__(self, numerator=0):\n length = lenMantessa(numerator)\n newNumerator = numerator * 10 ** length\n denominator = 10 ** length\n greatestCommonDemoniator = abs(gcd(newNumerator, denominator))\n fraction.__init__(self, newNumerator / greatestCommonDemoniator, denominator / greatestCommonDemoniator)\n #toString method which just returns a string of the float value of each decimal.\n def __str__(self):\n decimalValue = self.numerator / self.denominator\n return str(decimalValue)\n #If one of the items is a fraction it calls the fraction class's mulitplication method to correct typing errors and \n #provide the proper output, if not the float values of each decimal are multiplied together\n def __mul__(self, other):\n if type(other) == type(fraction()):\n return fraction.__mul__(self, other)\n else:\n fractionOne = float(self)\n fractionTwo = float(other)\n return decimal(fractionOne * fractionTwo)\n #If one of the items is a fraction it calls the fraction class's subtractaction method to correct typing errors and \n #provide the proper output, if not the float values of each decimal are subtracted together\n def __sub__(self, other):\n if type(other) == type(fraction()):\n return fraction.__sub__(self, other)\n else:\n fractionOne = float(self)\n fractionTwo = float(other)\n return decimal(fractionOne - fractionTwo)\n #If one of the items is a fraction it calls the fraction class's addition method to correct typing errors and \n #provide the proper output, if not the float values of each decimal are added together\n def __add__(self, other):\n if type(other) == type(fraction()):\n return fraction.__add__(self, other)\n else:\n fractionOne = float(self)\n fractionTwo = float(other)\n return decimal(fractionOne + fractionTwo)\n #If one of the items is a fraction it calls the fraction class's division method to correct typing errors and \n #provide the proper output, if not the float values of each decimal are divided.\n def __truediv__(self, other):\n if type(other) == type(fraction()):\n return fraction.__truediv__(self, other)\n else:\n fractionOne = float(self)\n fractionTwo = float(other)\n return decimal(fractionOne / fractionTwo)\n #If one of the items is a fraction it calls the fraction class's division method to correct typing errors and \n #provide the proper output, if not the method checks to see if there is an invalid radical and if there isn't\n #takes float value of the number to power of the float value of the other. If there is an invalid radical\n #an exception is thrown.\n def __pow__(self, power):\n if type(power) == type(fraction()):\n return fraction.__pow__(self, power)\n else:\n fractionOne = float(self)\n powerDecimal = float(power)\n if decimal(power).denominator %2 == 0 and fractionOne < 0:\n raise invalidRadical\n else:\n return decimal(fractionOne ** powerDecimal)\n\n\n#If a radical is invalid such as -2 ^ 1\\2 then this exception is thrown.\n#A custom exception is used to properly give an error message to the user.\nclass invalidRadical (Exception):\n pass\n\n'''\nThis method handles the entire programs control flow. It runs a loop that takes in inputs, readies them\nfor evaluation, evaluates, then repeats. If an error arises then the method prints out an error messgae\ninforming the user as to avoid any crashes. The main loop only concludes when the user enters\nthe keyword 'done' (case insensitive)\n'''\ndef main():\n print(\"At any point to end the session please enter \\'done\\'\")\n while True:\n expression = input(\"Please Enter An Expression: \")\n if expression != None and expression != \"\":\n if(expression.lower() == \"done\"):\n print (\"Session Concluded\")\n break\n else:\n try:\n expression = fixNegatives(expression)\n expression = removeSpaces(expression)\n fractionalized = fractionalize(expression)\n expressionAsString = fractionalized[0]\n fracList = fractionalized[1]\n except:\n print (\"The syntax behind your expression is invalid. Please refer to the Readme file.\")\n continue\n try:\n evaluate = eval(expressionAsString)\n print(evaluate)\n except ZeroDivisionError:\n print (\"Please Don't Divide by Zero\")\n except invalidRadical:\n print(\"Invalid Radical\")\n except:\n print(\"Invalid Expression\")\n else:\n print (\"You didn't enter anything\")\n\n\n\nmain() #Start program\n\n","repo_name":"jasonscharff/ATCS-PythonCalculator","sub_path":"mainDoc.py","file_name":"mainDoc.py","file_ext":"py","file_size_in_byte":18463,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"31446717923","text":"import json\nimport logging\nimport os\nimport time\n\nfrom instagram_selenium_crawler import Client, InstagramCommonCrawler, InstagramHashtagCrawler\n\nlogger = logging.getLogger()\nlogger.setLevel(logging.INFO)\n\ntarget_hashtag = '###target_hashtag_without_sharp###'\nlogin_account_id = '###instagram_login_id###'\nlogin_account_pw = '###instagram_login_pw###'\n\ndirname = f'hashtag/{target_hashtag}/'\nfilename_prefix = f'{int(time.time())}_{target_hashtag}'\n\nos.makedirs(dirname, exist_ok=True)\n\nclient = Client(headless=False)\n\nnext_max_id = None\nnext_page = None\n\n# next_max_id=\"QVFBM21Bam5wdHNjM1V4R1dvMnNxYURONzRmbEdHTm5ucVlrcDc0QnEtS2RMRVN6UlkxTE5TcGtnbDNnN1JqXzlaWVh2RmNzQ1dHUWpLdzg4WnlFVk1mbg==\"\n# next_media_ids= [\n# \"2714711203075369621\",\n# \"2714740669070384594\"\n# ]\n# next_page=1672\n\n\ndef get_hashtag_posts(target_hashtag, next_max_id=None, next_media_ids=None, next_page=None):\n hashtag_json_str = hashtag.get_hashtag_posts(target_hashtag, next_max_id, next_media_ids, next_page)\n filename = f\"{dirname}{filename_prefix}_{next_page if next_page else 0}.json\"\n print(filename)\n\n with open(filename, 'w', newline='') as f:\n f.write(hashtag_json_str)\n\n hashtag_json = json.loads(hashtag_json_str)\n\n if next_max_id is None:\n next_max_id = hashtag_json['data']['recent']['next_max_id']\n next_page = hashtag_json['data']['recent']['next_page']\n next_media_ids = hashtag_json['data']['recent']['next_media_ids']\n else:\n next_max_id = hashtag_json['next_max_id']\n next_page = hashtag_json['next_page']\n next_media_ids = hashtag_json['next_media_ids']\n\n return hashtag_json, next_max_id, next_media_ids, next_page\n\n\ntry:\n login = InstagramCommonCrawler(client=client, logger=logger).login(login_account_id, login_account_pw)\n\n if login:\n hashtag = InstagramHashtagCrawler(client)\n\n if next_max_id is None:\n hashtag_json, next_max_id, next_media_ids, next_page = get_hashtag_posts(target_hashtag)\n\n while next_max_id is not None:\n time.sleep(2)\n\n hashtag_json, next_max_id, next_media_ids, next_page = get_hashtag_posts(target_hashtag, next_max_id,\n next_media_ids, next_page)\n\n logger.info(f\"next_max_id: {next_max_id}\")\n\n client.driver_quit()\n\nexcept Exception as e:\n logger.exception(f\"Failed to function {e}\")\n client.driver_quit()\n","repo_name":"kory-/instagram_selenium_crawler","sub_path":"sample/hashtag.py","file_name":"hashtag.py","file_ext":"py","file_size_in_byte":2488,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"6982870700","text":"#!/usr/bin/python -u\n# -*- coding: latin-1 -*-\n# \n# Fibonacci (bidirectional) in Z3\n# \n# This Z3 model was written by Hakan Kjellerstrand (hakank@gmail.com)\n# See also my Z3 page: http://hakank.org/z3/\n# \n# \nfrom z3 import *\n\n# From https://rise4fun.com/Z3/0pld:\n# \"\"\"\n# (declare-fun fib (Int) Int)\n# (assert (= 1 (fib 0)))\n# (assert (= 1 (fib 1)))\n# (assert (forall (x Int) (=> (>= x 2) (= (fib x) (+ (fib (- x 1)) (fib (- x 2)))))))\n# (assert (= 2 (fib 2)))\n# \"\"\"\n\nsol = Solver()\n\nmax_n = 31\n\n#\n# Note: One have to set a max limit on fib\n#\n# https://stackoverflow.com/questions/6915227/can-z3-check-the-satisfiability-of-formulas-that-contain-recursive-functions\n# Leonardo de Moura:\n# \"\"\"\n# The models produced by Z3 assign an interpretation for each uninterpreted function symbol. The models can\n# be viewed as functional programs. The current version does not produce recursive definitions.\n# The first example [Fibonacci] is satisfiable, but Z3 fails to produce an interpretation for fib because\n# it does not support recursive definitions. We have plans to extend Z3 in this direction.\n# \"\"\"\nfib = Function(\"fib\", IntSort(), IntSort())\nx = Int(\"x\")\n# sol.add(fib(0) == 1)\n# sol.add(fib(1) == 1)\n# sol.add(ForAll(x, Implies(And(x >= 2, x <= max_n), fib(x) == fib(x-1) + fib(x-2))))\n# Simpler:\nsol.add(ForAll(x, If(And(x >= 2, x <= max_n), fib(x) == fib(x-1) + fib(x-2), fib(x) == 1)))\n\n# sol.add(x == fib(2))\ny = Int(\"y\")\nz = Int(\"z\")\nsol.add(y>0, y <= max_n, z >0, z <= max_n)\n\nsol.add(10946 == fib(y))\nsol.add(2178309 == fib(z))\n\nprint(sol)\nif sol.check()==sat:\n mod = sol.model()\n # print(\"x:\", mod.eval(x))\n print(\"z:\", mod.eval(z), \"y:\", mod.eval(y))\n sol.add(z != mod.eval(z),y != mod.eval(y))\n","repo_name":"hakank/hakank","sub_path":"z3/fibonacci.py","file_name":"fibonacci.py","file_ext":"py","file_size_in_byte":1718,"program_lang":"python","lang":"en","doc_type":"code","stars":339,"dataset":"github-code","pt":"7"} +{"seq_id":"13777544241","text":"# Nama: Evry Nazyli Ciptanto\n# NIM: 0110220045\n# Kelas: TI 08\n\nclass Node:\n def __init__(self, data = None, next = None):\n self.data = data\n self.next = next\n \nclass LinkedList:\n def __init__(self, head = None):\n self.head = head\n \n def add_last(self, new_data):\n if self.head is None:\n self.head = Node(new_data)\n else:\n current = self.head\n while current.next is not None:\n current = current.next\n current.next = Node(new_data)\n \n def cetak(self):\n if self.head is None:\n print('List kosong')\n else:\n current = self.head\n while current is not None:\n print(current.data, end=' ')\n current = current.next\n print()\n\n def sum_odd(self):\n # Tuliskan implementasi fungsi sum_odd() di bawah ini\n # Hapus pass jika implementasi sudah dibuat\n # membuat variabel awal x = 0\n x = 0\n # inisialisasi element pertama\n current_node = self.head\n # kondisi perulangan selama element tidak none\n while current_node is not None:\n # kondisi untuk mencari bilangan ganjil\n if(current_node.data % 2 != 0):\n # menambahkan data ke element x\n x += current_node.data\n # incremenet atau pindah ke element berikutnya\n current_node = current_node.next\n # mengembalikan hasil\n return x\n \n def get_max(self):\n # Tuliskan implementasi fungsi get_max() di bawah ini\n # Hapus pass jika implementasi sudah dibuat\n # inisialisasi element pertama\n current_node = self.head\n # jika element pertama none\n if current_node == None:\n # kembalikan nilai none\n return\n # kondisi yang lain\n else :\n # membuat variabel awal x = 0\n x = 0\n # kondisi perulangan selama element tidak none\n while current_node is not None:\n # kondisi untuk mencari bilangan max\n if x < current_node.data:\n # jika kondisi terpenuhi set x dengan nilai element\n x = current_node.data\n # incremenet atau pindah ke element berikutnya\n current_node = current_node.next\n # mengembalikan bilangan terbesar\n return x\n\n\n# Mulai baris ini hingga baris paling bawah\n# digunakan untuk mengetes fungsi yang telah dibuat.\n# Tidak perlu mengubah bagian ini.\n# Ketika dijalankan, program akan menampilkan contoh\n# pemanggilan fungsi dan solusi yang seharusnya.\n# Cocokkan hasil pemanggilan fungsi dengan solusi \n# yang seharusnya.\ndef test():\n list1 = LinkedList()\n list1.add_last(1)\n list1.add_last(2)\n list1.add_last(3)\n list1.add_last(4)\n list1.add_last(5)\n print('list1 : ', end='')\n list1.cetak()\n r1 = list1.sum_odd()\n print(f\"list1.sum_odd() = {r1} \\t(solusi: 9)\")\n r2 = list1.get_max()\n print(f\"list1.get_max() = {r2} \\t(solusi: 5)\")\n print()\n\n list2 = LinkedList()\n list2.add_last(9)\n list2.add_last(9)\n list2.add_last(9)\n print('list2 : ', end='')\n list2.cetak()\n r1 = list2.sum_odd()\n print(f\"list2.sum_odd() = {r1} \\t(solusi: 27)\")\n r2 = list2.get_max()\n print(f\"list2.get_max() = {r2} \\t(solusi: 9)\")\n print()\n\n list3 = LinkedList()\n list3.add_last(6)\n list3.add_last(2)\n list3.add_last(8)\n list3.add_last(4)\n print('list3 : ', end='')\n list3.cetak()\n r1 = list3.sum_odd()\n print(f\"list3.sum_odd() = {r1} \\t(solusi: 0)\")\n r2 = list3.get_max()\n print(f\"list3.get_max() = {r2} \\t(solusi: 8)\")\n print()\n\n list4 = LinkedList()\n print('list4 : ', end='')\n list4.cetak()\n r1 = list4.sum_odd()\n print(f\"list4.sum_odd() = {r1} \\t(solusi: 0)\")\n r2 = list4.get_max()\n print(f\"list4.get_max() = {r2} \\t(solusi: None)\")\n print()\n\nif __name__ == '__main__':\n test()","repo_name":"Nazyli/DDP-TI08-Lab-10","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3595,"program_lang":"python","lang":"id","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"39947074798","text":"import math\n\n\ndef main(x, y):\n storage = 0.\n n = len(x)\n for i in range(1, n):\n storage += math.sin(y[n - math.ceil(i / 3)]**2 + 58*x[i])**7\n return 13*storage\n\n\nprint(main([-0.22, -0.76, -0.89], [-0.14, -0.14, 0.54]))\n","repo_name":"ucantjugglikeme/python_exam","sub_path":"Task5v5.py","file_name":"Task5v5.py","file_ext":"py","file_size_in_byte":238,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"74165464544","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Nov 9 10:46:14 2018\n\n@author: alxgr\n\"\"\"\n\n###############################################################################\nimport numpy as np\nfrom scipy.interpolate import interp1d\n###############################################################################\nfrom pyro.control import controller\nfrom pyro.dynamic import mechanical\n###############################################################################\n\n\n\n\n###############################################################################\n# Computed Torque\n###############################################################################\n \nclass ComputedTorqueController( controller.StaticController ) :\n \"\"\" \n Inverse dynamic controller for mechanical system\n\n \"\"\" \n \n ############################\n def __init__(self, model = mechanical.MechanicalSystem() , traj = None ):\n \"\"\"\n \n ---------------------------------------\n r : reference signal vector k x 1\n y : sensor signal vector p x 1\n u : control inputs vector m x 1\n t : time 1 x 1\n ---------------------------------------\n u = c( y , r , t )\n \n \"\"\"\n \n self.model = model\n \n # Dimensions\n self.k = model.dof \n self.m = model.m\n self.p = model.p\n \n super().__init__(self.k, self.m, self.p)\n \n # Label\n self.name = 'Computed Torque Controller'\n \n # Params\n self.w0 = 1\n self.zeta = 0.7 \n \n # Mode\n if traj == None:\n self.c = self.c_fixed_goal\n else:\n self.load_trajectory( traj )\n self.mode = 'interpol'\n self.c = self.c_trajectory_following\n \n \n #############################\n def c_fixed_goal( self , y , r , t = 0 ):\n \"\"\" \n Feedback static computation u = c(y,r,t)\n \n INPUTS\n y : sensor signal vector p x 1\n r : reference signal vector k x 1\n t : time 1 x 1\n \n OUPUTS\n u : control inputs vector m x 1\n \n \"\"\"\n \n x = y \n q_d = r\n \n u = self.fixed_goal_ctl( x , q_d , t )\n \n return u\n \n \n \n ############################\n def fixed_goal_ctl( self , x , q_d , t = 0 ):\n \"\"\" \n \n Given desired fixed goal state and actual state, compute torques\n \n \"\"\"\n [ q , dq ] = self.model.x2q( x ) \n \n ddq_d = np.zeros( self.model.dof )\n dq_d = np.zeros( self.model.dof )\n\n ddq_r = self.compute_ddq_r( ddq_d , dq_d , q_d , dq , q )\n \n u = self.model.actuator_forces( q , dq , ddq_r )\n \n return u\n \n \n ############################\n def compute_ddq_r( self , ddq_d , dq_d , q_d , dq , q ):\n \"\"\" \n \n Given desired trajectory and actual state, compute ddq_r\n \n \"\"\"\n \n q_e = q - q_d\n dq_e = dq - dq_d\n \n ddq_r = ddq_d - 2 * self.zeta * self.w0 * dq_e - self.w0 ** 2 * q_e\n \n return ddq_r\n \n \n ############################\n def load_trajectory( self , traj ):\n \"\"\" \n \n Load Open-Loop trajectory solution to use as reference trajectory\n \n \"\"\"\n \n self.trajectory = traj\n \n q = traj.x[ :, 0 : self.model.dof ]\n dq = traj.x[ :, self.model.dof : 2 * self.model.dof ]\n ddq = traj.dx[:, self.model.dof : 2 * self.model.dof ]\n t = traj.t\n \n # Create interpol functions\n self.q = interp1d(t,q.T)\n self.dq = interp1d(t,dq.T)\n self.ddq = interp1d(t,ddq.T)\n \n \n ############################\n def get_traj( self , t ):\n \"\"\" \n \n Find closest point on the trajectory\n \n \"\"\"\n \n if t < self.trajectory.time_final :\n\n # Load trajectory\n q = self.q( t )\n dq = self.dq( t )\n ddq = self.ddq( t ) \n\n else:\n \n q = self.rbar\n dq = np.zeros( self.model.dof )\n ddq = np.zeros( self.model.dof )\n \n return ddq , dq , q\n \n \n ############################\n def traj_following_ctl( self , x , t = 0 ):\n \"\"\" \n \n Given desired loaded trajectory and actual state, compute torques\n \n \"\"\"\n \n [ q , dq ] = self.model.x2q( x ) \n \n ddq_d , dq_d , q_d = self.get_traj( t )\n\n ddq_r = self.compute_ddq_r( ddq_d , dq_d , q_d , dq , q )\n \n u = self.model.actuator_forces( q , dq , ddq_r )\n \n return u\n \n \n #############################\n def c_trajectory_following( self , y , r , t ):\n \"\"\" \n Feedback static computation u = c(y,r,t)\n \n INPUTS\n y : sensor signal vector p x 1\n r : reference signal vector k x 1\n t : time 1 x 1\n \n OUPUTS\n u : control inputs vector m x 1\n \n \"\"\"\n \n x = y \n \n u = self.traj_following_ctl( x , t )\n \n \n return u\n \n\n\n##############################################################################\n \nclass SlidingModeController( ComputedTorqueController ):\n \"\"\" \n Sliding Mode Controller for fully actuated mechanical systems\n \"\"\"\n \n \n ############################\n def __init__( self , model , traj = None ):\n \"\"\" \"\"\"\n \n super().__init__( model , traj )\n \n # Params\n self.lam = 1 # Sliding surface slope\n self.gain = 1 # Discontinuous gain\n self.nab = 0.1 # Min convergence rate\n \n \n ############################\n def compute_sliding_variables( self , ddq_d , dq_d , q_d , dq , q ):\n \"\"\" \n \n Given desired trajectory and actual state\n \n \"\"\" \n q_e = q - q_d\n dq_e = dq - dq_d\n \n s = dq_e + self.lam * q_e\n dq_r = dq_d - self.lam * q_e\n ddq_r = ddq_d - self.lam * dq_e\n \n return [ s , dq_r , ddq_r ]\n \n \n ############################\n def K( self , q , t ):\n \"\"\" Discontinuous gain matrix \"\"\"\n \n dist_max = np.diag( np.ones( self.model.dof ) ) * self.gain\n conv_min = np.diag( np.ones( self.model.dof ) ) * self.nab\n \n K = dist_max + np.dot( self.model.H( q ) , conv_min ) \n \n return K\n \n \n ############################\n def sliding_torque( self , ddq_r , s , dq , q , t ):\n \"\"\" \n \n Given actual state, compute torque necessarly to guarantee convergence\n \n \"\"\"\n \n u_computed = self.model.actuator_forces( q , dq , ddq_r )\n \n u_discontinuous = np.dot( self.K( q , t ) , np.sign( s ) )\n \n u_tot = u_computed - u_discontinuous\n \n return u_tot\n \n \n ############################\n def traj_following_ctl( self , x , t = 0 ):\n \"\"\" \n \n Given desired loaded trajectory and actual state, compute torques\n \n \"\"\"\n \n [ q , dq ] = self.model.x2q( x ) \n \n ddq_d , dq_d , q_d = self.get_traj( t )\n\n [ s , dq_r , ddq_r ] = self.compute_sliding_variables( ddq_d , dq_d , \n q_d , dq , q )\n \n u = self.sliding_torque( ddq_r , s , dq , q , t )\n \n return u\n \n \n ############################\n def fixed_goal_ctl( self , x , q_d , t = 0 ):\n \"\"\" \n \n Given desired fixed goal state and actual state, compute torques\n \n \"\"\"\n \n [ q , dq ] = self.model.x2q( x ) \n \n ddq_d = np.zeros( self.model.dof )\n dq_d = np.zeros( self.model.dof )\n\n [ s , dq_r , ddq_r ] = self.compute_sliding_variables( ddq_d , dq_d , \n q_d , dq , q )\n \n u = self.sliding_torque( ddq_r , s , dq , q , t )\n \n return u\n\n \n'''\n#################################################################\n################## Main ########\n#################################################################\n'''\n\n\nif __name__ == \"__main__\": \n \"\"\" MAIN TEST \"\"\"\n \n from pyro.dynamic import pendulum\n\n\n sys = pendulum.DoublePendulum()\n ctl = ComputedTorqueController( sys )\n \n # New cl-dynamic\n cl_sys = ctl + sys\n \n cl_sys.x0 = np.array([2,1,0,0])\n cl_sys.compute_trajectory()\n cl_sys.plot_phase_plane_trajectory(0,2)\n cl_sys.animate_simulation()\n \n","repo_name":"alx87grd/AlexRobotics","sub_path":"pyro/control/nonlinear.py","file_name":"nonlinear.py","file_ext":"py","file_size_in_byte":9249,"program_lang":"python","lang":"en","doc_type":"code","stars":15,"dataset":"github-code","pt":"7"} +{"seq_id":"25925928618","text":"from amiibo import AmiiboDump\nfrom amiibo.crypto import AmiiboBaseError\nimport copy\n\nclass InvalidAmiiboDump(AmiiboBaseError):\n pass\n\nclass IncorrectGameDataIdException(Exception):\n pass\n\n\nclass InvalidSsbuChecksum(Exception):\n pass\n\nclass SsbuAmiiboDump(AmiiboDump):\n \"\"\"\n Class that's a thin wrapper around AmiiboDump.\n Checks the amiibo has the super smash bros game id in the game data section on unlock\n Writes the HMAC for the game data before locking\n \"\"\"\n def __init__(self, master_keys, dump, is_locked=True):\n super().__init__(master_keys, dump, is_locked)\n self.dumpcopy = copy.deepcopy(self)\n if is_locked == True:\n self.dumpcopy.unlock()\n\n def unlock(self, verify=True):\n super().unlock(verify=verify)\n\n # Checks if the amiibo's game is Super Smash Bros. Ultimate, and if not, we initialize it.\n if bytes(self.data[266:270]).hex() != \"34f80200\":\n self.data[0x14] = self.data[0x14] | (1 << 5)\n self.data[266:270] = bytes.fromhex(\"34f80200\")\n self.data[0x100:0x108] = bytes.fromhex('01006A803016E000')\n self.data[0x130:0x208] = bytes.fromhex(\"00\" * 0xD8)\n self.data[304:308] = self._calculate_crc32(self.data[308:520]).to_bytes(4, \"little\")\n\n if self.data[304:308].hex() != self._calculate_crc32(self.data[308:520]).to_bytes(4, \"little\").hex():\n raise InvalidSsbuChecksum(f'The checksum for this game data is not correct. Please use an untampered amiibo')\n\n def lock(self):\n if self.data[444:502] != self.dumpcopy.data[444:502]:\n self.data[311] = self.data[311] | 1\n if self.amiibo_nickname[-1] != '□':\n if len(self.amiibo_nickname) == 10:\n self.amiibo_nickname = self.amiibo_nickname[:-1] + '□'\n else:\n self.amiibo_nickname = self.amiibo_nickname + '□'\n elif self.dumpcopy.amiibo_nickname[-1] == '□' and self.amiibo_nickname[-1] != '□':\n if len(self.amiibo_nickname) == 10:\n self.amiibo_nickname = self.amiibo_nickname[:-1] + '□'\n else:\n self.amiibo_nickname = self.amiibo_nickname + '□'\n checksum = self._calculate_crc32(self.data[308:520])\n mii_checksum = str(hex(self.crc16_ccitt_wii(self.data[0xA0:0xFE]))).lstrip('0x')\n while len(mii_checksum) < 4:\n mii_checksum = '0' + mii_checksum\n self.data[304:308] = checksum.to_bytes(4, \"little\")\n self.data[0xFE:0x100] = bytes.fromhex(mii_checksum)\n super().lock()\n\n @staticmethod\n def _calculate_crc32(input):\n # Setup CRC 32 table. Translated from js to python from amiibox codebase\n # (should move this out so it sets up once, but it's quick enough as is)\n p0 = 0xEDB88320 | 0x80000000\n p0 = p0 >> 0\n\n u0 = [0] * 0x100\n i = 1\n while (i & 0xFF):\n t0 = i\n for _ in range(8):\n b = (t0 & 0x1) >> 0\n t0 = (t0 >> 0x1) >> 0\n if b:\n t0 = (t0 ^ p0) >> 0\n u0[i] = t0 >> 0\n i += 1\n\n # Calculate CRC32 from table\n t = 0x0\n for k in input:\n t = ((t >> 0x8) ^ u0[(k ^ t) & 0xFF]) >> 0\n return (t ^ 0xFFFFFFFF) >> 0\n\n def crc16_ccitt_wii(self, data):\n crc = 0\n\n for byte in data:\n byte = int.from_bytes([byte], 'big')\n\n crc = crc ^ (byte << 8)\n\n for _ in range(8):\n crc = crc << 1\n\n if (crc & 0x10000) > 0:\n crc ^= 0x1021\n\n return (crc & 0xFFFF)\n\n\n @property\n def amiibo_nickname(self):\n # TODO: why is the Amiibo nickname big endian,\n # but the Mii nickname litle endian?\n return self.data[0x020:0x034].decode('utf-16-be').rstrip('\\x00')\n\n @amiibo_nickname.setter\n def amiibo_nickname(self, name):\n utf16 = name.encode('utf-16-be')\n if len(utf16) > 20:\n raise ValueError\n self.data[0x020:0x034] = utf16.ljust(20, b'\\x00')\n","repo_name":"jozz024/smash-amiibo-editor","sub_path":"utils/ssbu_amiibo.py","file_name":"ssbu_amiibo.py","file_ext":"py","file_size_in_byte":4124,"program_lang":"python","lang":"en","doc_type":"code","stars":17,"dataset":"github-code","pt":"7"} +{"seq_id":"9718059918","text":"from ddsp import core\nfrom ddsp import losses\nimport numpy as np\nimport tensorflow as tf\n\n\nclass LossGroupTest(tf.test.TestCase):\n\n def setUp(self):\n \"\"\"Create some dummy input data for the chain.\"\"\"\n super().setUp()\n\n # Create a network output dictionary.\n self.nn_outputs = {\n 'audio': tf.ones((3, 8000), dtype=tf.float32),\n 'audio_synth': tf.ones((3, 8000), dtype=tf.float32),\n 'magnitudes': tf.ones((3, 200, 2), dtype=tf.float32),\n 'f0_hz': 200 + tf.ones((3, 200, 1), dtype=tf.float32),\n }\n\n # Create Processors.\n spectral_loss = losses.SpectralLoss()\n crepe_loss = losses.PretrainedCREPEEmbeddingLoss(name='crepe_loss')\n\n # Create DAG for testing.\n self.dag = [\n (spectral_loss, ['audio', 'audio_synth']),\n (crepe_loss, ['audio', 'audio_synth']),\n ]\n self.expected_outputs = [\n 'spectral_loss',\n 'crepe_loss'\n ]\n\n def _check_tensor_outputs(self, strings_to_check, outputs):\n for tensor_string in strings_to_check:\n tensor = core.nested_lookup(tensor_string, outputs)\n self.assertIsInstance(tensor, (np.ndarray, tf.Tensor))\n\n def test_dag_construction(self):\n \"\"\"Tests if DAG is built properly and runs.\n \"\"\"\n loss_group = losses.LossGroup(dag=self.dag)\n print('!!!!!!!!!!!', loss_group.dag, loss_group.loss_names, self.dag)\n loss_outputs = loss_group(self.nn_outputs)\n self.assertIsInstance(loss_outputs, dict)\n self._check_tensor_outputs(self.expected_outputs, loss_outputs)\n\n\nclass SpectralLossTest(tf.test.TestCase):\n\n def test_output_shape_is_correct(self):\n \"\"\"Test correct shape with all losses active.\"\"\"\n loss_obj = losses.SpectralLoss(\n mag_weight=1.0,\n delta_time_weight=1.0,\n delta_freq_weight=1.0,\n cumsum_freq_weight=1.0,\n logmag_weight=1.0,\n loudness_weight=1.0,\n )\n\n input_audio = tf.ones((3, 8000), dtype=tf.float32)\n target_audio = tf.ones((3, 8000), dtype=tf.float32)\n\n loss = loss_obj(input_audio, target_audio)\n\n self.assertListEqual([], loss.shape.as_list())\n self.assertTrue(np.isfinite(loss))\n\n\n\n\nclass PretrainedCREPEEmbeddingLossTest(tf.test.TestCase):\n\n def test_output_shape_is_correct(self):\n loss_obj = losses.PretrainedCREPEEmbeddingLoss()\n\n input_audio = tf.ones((3, 16000), dtype=tf.float32)\n target_audio = tf.ones((3, 16000), dtype=tf.float32)\n\n loss = loss_obj(input_audio, target_audio)\n\n self.assertListEqual([], loss.shape.as_list())\n self.assertTrue(np.isfinite(loss))\n\n\nif __name__ == '__main__':\n tf.test.main()\n","repo_name":"magenta/ddsp","sub_path":"ddsp/losses_test.py","file_name":"losses_test.py","file_ext":"py","file_size_in_byte":2577,"program_lang":"python","lang":"en","doc_type":"code","stars":2656,"dataset":"github-code","pt":"7"} +{"seq_id":"24403063213","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"Validator script for CS 2204 Homework submissions\"\"\"\nimport sys\nimport os\nimport traceback\nfrom contextlib import redirect_stdout, redirect_stderr\nfrom importlib.util import spec_from_file_location, module_from_spec\nfrom io import StringIO\n\nimport pycodestyle\n\n\n#########################################################################\n# Test infrastructure\n\ndef points(val):\n \"\"\"Simple decorator to add a __points__ property to a function\n Usage: @points(10)\n \"\"\"\n def decorator(func):\n func.__points__ = val\n return func\n return decorator\n\n\ndef abort(func):\n \"\"\"Simple decorator to add a __abort__ property to a function\"\"\"\n func.__abort__ = True\n return func\n\n\ndef import_file(filename, module_name=None):\n \"\"\"Import a file with a given module name. Returns the module object\"\"\"\n if module_name is None:\n module_name, _ = os.path.splitext(os.path.basename(filename))\n spec = spec_from_file_location(module_name, filename)\n module = module_from_spec(spec)\n spec.loader.exec_module(module)\n return module\n\n\ndef run_file(filename):\n \"\"\"Runs the file at the top-level. Returns standard output, error tuple\"\"\"\n with StringIO() as stdout, StringIO() as stderr:\n with redirect_stdout(stdout), redirect_stderr(stderr):\n import_file(filename, \"__main__\")\n out = stdout.getvalue()\n err = stderr.getvalue()\n\n return out, err\n\n\ndef check_pep8_style(filename):\n \"\"\"Checking PEP8 style for filename. Returns n_errors, messages tuple\"\"\"\n\n with StringIO() as out, \\\n redirect_stdout(out), \\\n redirect_stderr(sys.stdout):\n\n pep8_checker = pycodestyle.Checker(filename, show_source=True)\n pep8_errors = pep8_checker.check_all()\n pep8_msgs = out.getvalue()\n\n return pep8_errors, pep8_msgs\n\n\ndef validate(*tests):\n \"\"\"Validation function\"\"\"\n total_score = 0\n divider = \"-\" * 45\n\n for test in tests:\n max_points = getattr(test, \"__points__\", 0)\n abort = False\n\n print(divider)\n prefix = f\"{test.__doc__}:\"\n print(f\"{prefix:<{len(divider) - 10}}\", end=\"\")\n try:\n with StringIO() as out, \\\n redirect_stdout(out), \\\n redirect_stderr(sys.stdout):\n\n partial = test()\n msg = out.getvalue()\n\n except: # noqa\n with StringIO() as tb_out:\n traceback.print_exc(limit=-1, file=tb_out)\n msg = tb_out.getvalue()\n\n points = min(max_points, 0)\n abort = getattr(test, \"__abort__\", False)\n\n else:\n if partial is not None:\n points = partial\n else:\n points = max(max_points, 0)\n\n suffix = \"ok\"\n if max_points <= 0:\n if points != 0:\n suffix = f\"{points:+d} pts\"\n else:\n suffix = f\"{points}/{max_points} pts\"\n print(f\"{suffix:>10}\")\n\n if msg:\n print(f\"\\n\\n{msg}\")\n\n total_score += points\n\n if abort:\n print(\"Aborting future tests. Fix this one, first!\")\n break\n\n total_score = max(total_score, 0)\n print(\"=\" * len(divider))\n print(f\"Final score (estimated): {total_score:.0f} pts\")\n\n\n#########################################################################\n# Assignment-specific tests\n\n# Notes:\n# You can set the max positive points for each test with @points()\n# You can also set a negative value with @points, it is interpreted as penalty.\n# If the test returns with\n# - None: max positive points are added, no penalties are applied\n# - Number: a given points are added, if negative, given penalties are applied\n# - Exception: no positive points are added, max penalties applied\n\n\n@points(10)\ndef test_codon_constructor():\n \"\"\"Testing Codon class initializer\"\"\"\n module = import_file(\"dna.py\")\n Codon = module.Codon\n\n # valid patterns\n for pattern in \"act\", \"Act\", \"CTG\":\n _ = Codon(pattern)\n\n # invalid patterns\n for pattern in None, 7, \"\", \"actg\", \"acx\":\n try:\n _ = Codon(pattern)\n except: # noqa\n pass\n else:\n print(f\"Initializer should refuse this parameter: {pattern!r}\")\n return 0\n\n\n@points(10)\ndef test_codon_str():\n \"\"\"Testing Codon string conversion\"\"\"\n module = import_file(\"dna.py\")\n Codon = module.Codon\n assert str(Codon(\"ACT\")) == \"[ACT]\"\n assert str(Codon(\"act\")) == \"[ACT]\"\n assert str(Codon(\"ggG\")) == \"[GGG]\"\n\n\n@points(10)\ndef test_codon_eq():\n \"\"\"Testing Codon equality\"\"\"\n module = import_file(\"dna.py\")\n Codon = module.Codon\n assert Codon(\"ACT\") == Codon(\"ACT\")\n assert Codon(\"ACT\") == Codon(\"act\")\n assert Codon(\"act\") != Codon(\"acg\")\n\n\n@points(10)\ndef test_codon_transcribe():\n \"\"\"Testing Codon transcription\"\"\"\n module = import_file(\"dna.py\")\n Codon = module.Codon\n assert Codon(\"ACT\").transcribe() == \"\"\n assert Codon(\"act\").transcribe() == \"\"\n assert Codon(\"acg\").transcribe() == \"\"\n\n\n@points(10)\ndef test_gene_constructor():\n \"\"\"Testing Gene class initializer\"\"\"\n module = import_file(\"dna.py\")\n Gene = module.Gene\n\n # valid patterns\n for pattern in \"\", \"ac\", \"actg\", \"actgg\", \"actggc\":\n _ = Gene(pattern)\n\n # invalid patterns\n for pattern in None, 7, \"xkcd\":\n try:\n _ = Gene(pattern)\n except: # noqa\n pass\n else:\n print(f\"Initializer should refuse this parameter: {pattern!r}\")\n return 0\n\n\n@points(10)\ndef test_gene_str():\n \"\"\"Testing Gene string conversion\"\"\"\n module = import_file(\"dna.py\")\n Gene = module.Gene\n assert str(Gene(\"\")) == \"\"\n assert str(Gene(\"ac\")) == \"\"\n assert str(Gene(\"act\")) == \"[ACT]\"\n assert str(Gene(\"actgg\")) == \"[ACT]\"\n assert str(Gene(\"actggc\")) == \"[ACT][GGC]\"\n assert str(Gene(\"actggctt\")) == \"[ACT][GGC]\"\n\n\n@points(10)\ndef test_gene_transcribe():\n \"\"\"Testing Gene transcription\"\"\"\n module = import_file(\"dna.py\")\n Gene = module.Gene\n assert Gene(\"\").transcribe() == \"\"\n assert Gene(\"ac\").transcribe() == \"\"\n assert Gene(\"act\").transcribe() == \"\"\n assert Gene(\"actgg\").transcribe() == \"\"\n assert Gene(\"actggctt\").transcribe() == \"\"\n\n\n@points(15)\ndef test_gene_contains():\n \"\"\"Testing Codon search in Gene\"\"\"\n module = import_file(\"dna.py\")\n Codon, Gene = module.Codon, module.Gene\n gene = Gene(\"actgggctgacctgttaaactgtc\")\n assert Codon(\"act\") in gene\n assert Codon(\"ggg\") in gene\n assert Codon(\"gtt\") not in gene\n assert Codon(\"ccc\") not in gene\n assert Codon(\"acg\") not in gene\n\n\n@points(15)\ndef test_gc_content():\n \"\"\"Testing GC content calculation\"\"\"\n from math import isclose\n from random import choices\n module = import_file(\"dna.py\")\n Gene = module.Gene\n for k in 10, 100, 1000:\n gene_str = \"\".join(choices(\"actg\", k=k))\n gene_str_cut = gene_str[:k - (k % 3)]\n gc_cnt = gene_str_cut.count(\"g\") + gene_str_cut.count(\"c\")\n gc_content = gc_cnt / len(gene_str_cut)\n if not isclose(gc_content, Gene(gene_str).gc_content()):\n print(f\"GC content should be {gc_content} for: \\n{gene_str}\")\n return 0\n\n\n@points(-10)\ndef test_author():\n \"\"\"Checking author information\"\"\"\n module = import_file(\"dna.py\")\n author = getattr(module, \"__author__\", None)\n assert isinstance(author, str) and author, \\\n \"__author__ variable is not set properly\"\n\n\ndef test_pep8():\n \"\"\"Checking PEP8 style\"\"\"\n\n n_errors, msgs = check_pep8_style(\"dna.py\")\n if n_errors:\n print(msgs)\n return -n_errors\n\n\nif __name__ == \"__main__\":\n validate(\n test_codon_constructor,\n test_codon_str,\n test_codon_eq,\n test_codon_transcribe,\n test_gene_constructor,\n test_gene_str,\n test_gene_transcribe,\n test_gene_contains,\n test_gc_content,\n test_author,\n test_pep8,\n )\n","repo_name":"kdmalc/data-structs-algos","sub_path":"CS2204/CS_HW3/validator.py","file_name":"validator.py","file_ext":"py","file_size_in_byte":8116,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"74665788062","text":"import os\nimport time\nimport yaml\n\nimport numpy as np\nimport batoid\nimport danish\nfrom danish import DonutFactory\n\n\ndef time_image():\n obsc = yaml.safe_load(open(os.path.join(danish.datadir, 'RubinObsc.yaml')))\n factory = DonutFactory(\n obsc_radii=obsc['radii'],\n obsc_centers=obsc['centers'],\n obsc_th_mins=obsc['th_mins'],\n )\n\n telescope = batoid.Optic.fromYaml(\"LSST_r.yaml\")\n telescope = telescope.withGloballyShiftedOptic(\"Detector\", (0,0,0.0015))\n zref = batoid.zernikeTA(\n telescope, np.deg2rad(1.67), 0.0, 620e-9,\n jmax=66, nrad=20, naz=120, reference='chief', eps=0.61\n )\n\n N = 200\n np.random.seed(123)\n t0 = time.time()\n for _ in range(N):\n aberrations = np.array(zref)\n aberrations[4] += np.random.uniform(-0.1, 0.1)\n aberrations[5:23] += np.random.uniform(-0.1, 0.1, size=18)\n aberrations *= 620e-9\n img = factory.image(\n aberrations=aberrations, thx=np.deg2rad(1.67), thy=0.0\n )\n t1 = time.time()\n print(f\"Time for factory.image(): {(t1-t0)/N*1e3:.2f} ms\")\n\n\nif __name__ == \"__main__\":\n time_image()\n","repo_name":"jmeyers314/danish","sub_path":"devel/time_image.py","file_name":"time_image.py","file_ext":"py","file_size_in_byte":1143,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"37781861514","text":"#!/usr/bin/env python3\n'''\nProduce CSV files easily readable by Postgres' /copy\n'''\nimport gzip\nimport json\nimport tensorflow_hub as hub\nimport pandas as pd\nimport sys\nfrom tqdm import tqdm\nfrom nltk.tokenize import sent_tokenize\nimport tensorflow as tf\n\n\ndef load_abstracts(filename):\n with open(filename) as file:\n data = json.load(file)\n paper_ids = [key for key in data.keys()]\n abstracts = [data[key] for key in paper_ids]\n paper_ids = [int(paper_id) for paper_id in paper_ids]\n del data\n return paper_ids, abstracts\n\n\ndef embed_abstracts(abstracts, paper_ids, prefix):\n gpus = tf.config.experimental.list_physical_devices('GPU')\n if gpus:\n for gpu in gpus:\n tf.config.experimental.set_memory_growth(gpu, True)\n\n print('Loading USE')\n embed = hub.load(\"https://tfhub.dev/google/universal-sentence-encoder-large/5\")\n\n def embed_abstract(abstract, paper_id):\n sentences = sent_tokenize(abstract)\n sentence_embeddings = embed(sentences).numpy()\n return str(paper_id) + '\\t' + '{' + ','.join([str(v) for v in sentence_embeddings.mean(axis=0)]) + '}\\n'\n\n def write_embeddings(embeddings, filename):\n print(f'writing to {filename}')\n with open(filename, 'w') as file:\n file.writelines(embeddings)\n\n print('Computing embeddings')\n embeddings = []\n suffix = 0\n for (paper_id, abstract) in tqdm(zip(paper_ids, abstracts)):\n if len(embeddings) >= 250000:\n write_embeddings(embeddings, f'{prefix}_{suffix}.tsv')\n suffix += 1\n embeddings = []\n embeddings.append(embed_abstract(abstract, paper_id))\n suffix += 1\n write_embeddings(embeddings, f'{prefix}_{suffix}.tsv')\n\n\ndef main():\n infilename = sys.argv[1]\n outfilename = sys.argv[2]\n print(f'Loading abstracts from {infilename}')\n paper_ids, abstracts = load_abstracts(infilename)\n embed_abstracts(abstracts, paper_ids, outfilename)\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"APEX-WSU/mag_50k","sub_path":"scripts/python/abstract_use_embeddings_postgres.py","file_name":"abstract_use_embeddings_postgres.py","file_ext":"py","file_size_in_byte":2020,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"6228048624","text":"'''\r\nRaising an exception[raising exception means raising error description from our end when we get error]\r\n====================\r\n --------- python\r\n When there is fault in runtime---exception is raised--handles by--|\r\n by system --------try.excet(user defined)\r\n\r\nTill now exception is raised internally by system,\r\nif there is need to raised an exception,\r\nthen raise keyword to raise an exception.\r\n\r\nexception is raised with recepct to certain condition.\r\nSyntax to raise exception\r\n=========================\r\n\r\n raise ExceptionName('Message')\r\n\r\n'''\r\n\r\n\r\n\r\nx=int(input(\"enter numerator:\"))#x=9|x=9\r\ny=int(input(\"enter denominator:\"))#y=2|y=0\r\n\r\nif y==0: #2==0F|0==0T\r\n\r\n raise ZeroDivisionError('Denominator cannot be Zero!!')\r\n\r\n\r\nelse:\r\n d=x/y #9/2=>4.5 |9/0=> Exception is rasied|ZeroDivisionError\r\n print(\"Division is:\",d)\r\n\r\n'''\r\noutput:-\r\n\r\nenter numerator:9\r\nenter denominator:2\r\nDivision is: 4.5\r\n\r\n======== RESTART: C:/8.exception handling in python/raising_exception.py =======\r\nenter numerator:9\r\nenter denominator:0\r\nTraceback (most recent call last):\r\n File \"C:/8.exception handling in python/raising_exception.py\", line 27, in \r\n raise ZeroDivisionError('Denominator cannot be Zero!!')\r\nZeroDivisionError: Denominator cannot be Zero!!\r\n\r\n'''\r\n\r\n'''\r\nhome work:-\r\n\r\ncheck weither it is digit or not then if it is not digit then raise a exception for above code\r\n\r\n'''","repo_name":"CODERAJIT/Python-files","sub_path":"exception_handling in python/rasing_exception.py","file_name":"rasing_exception.py","file_ext":"py","file_size_in_byte":1571,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"19186464576","text":"from appium import webdriver\nfrom time import sleep\nfrom appium.webdriver.common.touch_action import TouchAction\n\ndesired_capabilities ={\n\"platformName\": \"Android\",\n\"platformVersion\": \"5.1.1\",\n\"deviceName\": \"62001\"\n}\n\ndriver = webdriver.Remote('http://127.0.0.1:4723/wd/hub',desired_capabilities=desired_capabilities)\nsleep(3)\n\n#定位元素\nmtxx = driver.find_element_by_accessibility_id(('美图秀秀'))\n#长按元素 可以一系列操作 执行\npressed = TouchAction(driver).long_press(mtxx).perform()\n#拖动并释放元素\npressed.move_to(x=125,y=84).release().perform()\n#点击确认\ndriver.tap([(591,732)])\nsleep((5))\n\n#用命令行工具实现\n#百度ANDRIOD KEYCODE\n#CMD:adb shell input keyevent KEYCODE_HOME\n#CMD:adb shell input keyevent 3","repo_name":"bai345767318/python-java","sub_path":"python/appium/demo08.py","file_name":"demo08.py","file_ext":"py","file_size_in_byte":765,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"35386725864","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\n\nx = np.linspace(0, 5, 1000)\n\nf = 2\ny = np.sin(f * x * 2 * np.pi)\ny += 0.1 * np.sin(2 * f * x * 2 * np.pi)\ny += 0.4 * np.sin(3 * f * x * 2 * np.pi)\n\n\nif __name__ == \"__main__\":\n\n df = pd.DataFrame({\"x\": x, \"y\": y})\n df.to_csv(\"sinus.csv\", float_format='%.5f', index=False)\n\n plt.plot(x, y)\n plt.show()\n","repo_name":"MaxiHartmann/pyqt_learning","sub_path":"004_matplot_selection/data/generate.py","file_name":"generate.py","file_ext":"py","file_size_in_byte":386,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"74998440221","text":"def search_quadruplets(arr,target):\n arr.sort()\n quadruplets = []\n\n for i in range(0,len(arr)-3):\n if i>0 and arr[i] == arr[i-1]:\n continue\n for j in range(i+1,len(arr)-2):\n if j > i+1 and arr[j] == arr[j-1]:\n continue\n search_pairs(arr,target,i,j,quadruplets)\n return quadruplets\n\ndef search_pairs(arr,target_sum, first, second, quadruplets):\n left = second + 1\n right = len(arr) - 1\n\n while (left < right):\n suma = arr[first] + arr[second] + arr[left] + arr[right]\n if suma == target_sum:\n quadruplets.append([arr[first],arr[second],arr[left],arr[right]])\n left += 1\n right -= 1\n while (left < right and arr[left]==arr[left-1]):\n left += 1\n while (left < right and arr[right]==arr[right+1]):\n right -= 1\n elif suma < target_sum:\n left += 1\n else:\n right -= 1\n\nprint(search_quadruplets([4,1,2,-1,1,-3],1)) #[[-3, -1, 1, 4], [-3, 1, 1, 2]]\nprint(search_quadruplets([4,1,2,-1,1,-3],1)) #[[-2, 0, 2, 2], [-1, 0, 1, 2]]\n ","repo_name":"alimalim77/Python-Practice-Track","sub_path":"Warm Up/four_sum.py","file_name":"four_sum.py","file_ext":"py","file_size_in_byte":1143,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"72598515743","text":"#!env python3\n\nimport os\nimport sys\n# import pandas as pd\nfrom PyQt5 import QtGui, QtCore, QtWidgets\nfrom PyQt5.QtGui import QIcon, QColor\nfrom PyQt5.QtCore import Qt\nfrom PyQt5.QtWidgets import QApplication, QWidget, QDialog, QMainWindow, QDesktopWidget\nsys.path.append(os.path.dirname(__file__)) \ntry:\n from . import mainwindow\nexcept (ImportError, SystemError):\n try: \n import mainwindow\n except ModuleNotFoundError:\n pass\n\nfrom collections import defaultdict\n\n\ndef find_files(dirname, match=None):\n result = []\n for root, dirs, files in os.walk(dirname):\n for file_ in files:\n if match is None or match in file_:\n result.append(file_)\n return result\n\n\nclass LensRater(QMainWindow, mainwindow.Ui_MainWindow):\n \"\"\" A hacked-together gui for rating lens images.\n TODO:\n Side-by-side for subtractions or single band images\n Expand image - zoom out\n Zoom in further\n Review choices made\n Produce output montages\n Access images remotely\n \"\"\"\n\n colour_codes = [Qt.black, QColor(27, 76, 198), QColor(234, 224, 23), QColor(198, 13, 13), Qt.gray]\n categories = [\"\", \"Maybe\", \"Probably\", \"Definitely\"]\n\n def __init__(self, image_dir):\n super(LensRater, self).__init__()\n self.setupUi(self)\n\n self.next_button.clicked.connect(self.nextImage)\n self.prev_button.clicked.connect(self.prevImage)\n\n self.image_dir = image_dir\n self.image_files = sorted(find_files(image_dir, match=\".png\"))\n if len(self.image_files) == 0:\n print(\"No image files found.\")\n sys.exit(0)\n self.image_index = 0\n self.current_image = None\n self.progress_bar.setMaximum(len(self.image_files))\n self.scorefile = self.image_dir + \"/scores.csv\"\n self.scores = defaultdict(lambda: -1)\n self.setChildrenFocusPolicy(QtCore.Qt.NoFocus)\n QtWidgets.qApp.installEventFilter(self)\n\n self.radios = [self.radio_0, self.radio_1, self.radio_2, self.radio_3]\n for i in range(4):\n self.radios[i].toggled.connect(self.radio_score)\n # self.radios[i].setFocusPolicy(QtCore.Qt.NoFocus)\n \n self.set_username_button.clicked.connect(self.update_username)\n self.username_edit.returnPressed.connect(self.set_username)\n self.jump_box.returnPressed.connect(self.jumped_to)\n self.actionSave_and_Quit.triggered.connect(self.close)\n self.actionQuit_without_saving.triggered.connect(self.quit)\n self.actionReset_scores.triggered.connect(self.reset)\n\n p = self.palette()\n p.setColor(self.backgroundRole(), Qt.white)\n self.setPalette(p)\n\n qtRectangle = self.frameGeometry()\n centerPoint = QDesktopWidget().availableGeometry().center()\n qtRectangle.moveCenter(centerPoint)\n self.move(qtRectangle.topLeft())\n\n self.set_label.setText(os.path.abspath(image_dir).split(\"/\")[-1])\n self.username = os.environ.get(\"USER\", \"nobody\")\n self.load()\n self.username_edit.setText(self.username)\n self.go_to_last()\n # self.goto_image(0)\n\n def set_colour_code(self, colour):\n p = self.colour_label.palette()\n p.setColor(self.backgroundRole(), colour)\n self.colour_label.setPalette(p)\n\n def setChildrenFocusPolicy(self, policy):\n def recursiveSetChildFocusPolicy(parentQWidget):\n for childQWidget in parentQWidget.findChildren(QWidget):\n if childQWidget == self.username_edit or childQWidget == self.jump_box:\n continue\n childQWidget.setFocusPolicy(policy)\n recursiveSetChildFocusPolicy(childQWidget)\n\n recursiveSetChildFocusPolicy(self)\n\n def eventFilter(self, source, event):\n if event.type() == QtCore.QEvent.KeyPress:\n if event.key() == QtCore.Qt.Key_Left or event.key(\n ) == QtCore.Qt.Key_Right:\n self.keyPressEvent(event)\n return True\n return super(LensRater, self).eventFilter(source, event)\n\n def keyPressEvent(self, e):\n if e.key() == QtCore.Qt.Key_Escape:\n self.close()\n elif e.key() == QtCore.Qt.Key_Right or e.key() == QtCore.Qt.Key_J:\n self.nextImage()\n elif e.key() == QtCore.Qt.Key_Left or e.key() == QtCore.Qt.Key_K:\n self.prevImage()\n elif e.key() == QtCore.Qt.Key_End:\n self.goto_image(len(self.image_files) - 1)\n elif e.key() == QtCore.Qt.Key_Home:\n self.goto_image(0)\n elif e.key() == QtCore.Qt.Key_PageUp:\n self.goto_image(max(self.image_index - 10, 0))\n elif e.key() == QtCore.Qt.Key_PageDown:\n self.goto_image(\n min(self.image_index + 10, len(self.image_files) - 1))\n elif e.key() == QtCore.Qt.Key_Space:\n # self.up_score()\n self.nextImage()\n elif e.key() == QtCore.Qt.Key_0 or e.key() == QtCore.Qt.Key_QuoteLeft:\n self.score_image(0)\n elif e.key() == QtCore.Qt.Key_1:\n self.score_image(1)\n elif e.key() == QtCore.Qt.Key_2:\n self.score_image(2)\n elif e.key() == QtCore.Qt.Key_3:\n self.score_image(3)\n # elif e.key() == QtCore.Qt.Key_Right:\n # pass\n # elif e.key() == QtCore.Qt.Key_Right:\n # pass\n # elif e.key() == QtCore.Qt.Key_Right:\n # pass\n elif e.key() == QtCore.Qt.Key_Return:\n pass\n # self.close()\n\n def nextImage(self):\n if self.image_index < len(self.image_files) - 1:\n self.image_index += 1\n self.goto_image(self.image_index)\n\n def prevImage(self):\n if self.image_index > 0:\n self.image_index -= 1\n self.goto_image(self.image_index)\n\n def goto_image(self, index):\n self.image_index = index\n self.current_image = self.image_files[index]\n self.set_display_image(self.image_files[index])\n self.position_label.setText(str(index) +\": \" + self.current_image)\n self.progress_bar.setValue(index + 1)\n current_score = self.scores[self.current_image]\n self.set_colour_code(LensRater.colour_codes[current_score])\n self.toggle_radio(current_score)\n\n if current_score < 0: # Implicitly score when seen\n self.score_image(0)\n\n def toggle_radio(self, score):\n for i in range(4):\n self.radios[i].setChecked(i == score)\n\n def resizeEvent(self, event):\n self.goto_image(self.image_index)\n\n def set_display_image(self, impath):\n min_dim = min(self.image_label.width(), self.image_label.height())\n image_profile = QtGui.QImage(self.image_dir + \"/\" + impath)\n image_profile = image_profile.scaled(min_dim, min_dim, \\\n aspectRatioMode=QtCore.Qt.KeepAspectRatio, \\\n transformMode=QtCore.Qt.SmoothTransformation)\n # self.image_label.setScaledContents(True)\n self.image_label.setPixmap(QtGui.QPixmap.fromImage(image_profile))\n\n def up_score(self):\n current_score = self.scores[self.current_image]\n newscore = current_score + 1\n if newscore > 3:\n newscore = 0\n self.score_image(newscore)\n\n def go_to_last(self):\n for i, f in enumerate(self.image_files):\n if self.scores[f] < 0:\n self.goto_image(i)\n break\n \n def radio_score(self):\n for i in range(4):\n if self.radios[i].isChecked():\n self.score_image(i)\n break\n\n def score_image(self, score):\n self.scores[self.current_image] = score\n self.set_colour_code(LensRater.colour_codes[score])\n self.toggle_radio(score)\n self.colour_label.setText(LensRater.categories[score])\n\n def load(self):\n if os.path.isfile(self.scorefile):\n with open(self.scorefile, \"r\") as f:\n lines = [s.split(\",\") for s in f.readlines()]\n self.username = lines[1][0]\n for line in lines[1:]:\n self.scores[line[1]] = int(line[2])\n\n def jumped_to(self):\n # self.jump_box.setEnabled(False)\n self.setFocus()\n goto = self.jump_box.text()\n try:\n goto = int(goto)\n self.goto_image(goto)\n except ValueError:\n pass\n self.jump_box.setText(\"\")\n\n def update_username(self):\n self.username_edit.setEnabled(True)\n self.username_edit.setText(\"\")\n self.username_edit.setFocus()\n\n def set_username(self):\n self.set_username_(self.username_edit.text())\n self.username_edit.setEnabled(False)\n self.username_edit.setFocus()\n\n def set_username_(self, username):\n self.username = username\n\n def review_scores(self):\n pass\n\n def save(self):\n with open(self.scorefile, \"w\") as f:\n f.write(\"username,image,score\\n\")\n for file_ in self.image_files:\n f.write(\"%s,%s,%d\\n\" % (self.username, file_.split(\"/\")[-1],\\\n self.scores[file_]))\n\n def quit(self):\n sys.exit(0)\n\n def reset(self):\n for key in self.scores:\n self.scores[key] = -1\n\n def close(self):\n self.save()\n super(LensRater, self).close()\n\n\ndef main():\n app = QApplication(sys.argv)\n iconloc = \"/\".join(__file__.split(\"/\")[0:-1]) + '/icon.png'\n if not os.path.isfile(iconloc):\n iconloc = \"../icon.png\"\n if not os.path.isfile(iconloc):\n iconloc = \"/\".join(__file__.split(\"/\")[0:-1]) + \"/../icon.png\"\n app.setWindowIcon(QIcon(iconloc))\n imgdir = \".\"\n if len(sys.argv) > 1:\n imgdir = sys.argv[1]\n window = LensRater(imgdir)\n app.aboutToQuit.connect(window.save)\n window.show()\n sys.exit(app.exec_())\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"coljac/lensrater","sub_path":"lensrater/lensrater.py","file_name":"lensrater.py","file_ext":"py","file_size_in_byte":9947,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"11358456981","text":"import unittest\n#\n# Design and implement an algorithm that can preprocess a\n# graph and then answer the question \"is x connected to y in the\n# graph\" for any x and y in constant time Theta(1).\n#\n\n#\n# `process_graph` will be called only once on each graph. If you want,\n# you can store whatever information you need for `is_connected` in\n# global variables\n#\ndef mark_component(G, node):\n marked = {}\n open_list = [node]\n while len(open_list) > 0:\n current_node = open_list.pop()\n marked[current_node] = True\n for neighbor in G[current_node]:\n if neighbor not in marked:\n open_list.append(neighbor)\n\n return marked\n\nmarked_dict = {}\ndef process_graph(G):\n for node in G.keys():\n marked = mark_component(G, node)\n marked_dict[node] = marked\n pass\n\n#\n# When being graded, `is_connected` will be called\n# many times so this routine needs to be quick\n#\ndef is_connected(i, j):\n return j in marked_dict[i]\n\n#######\n# Testing\n#\nclass test_preprocess(unittest.TestCase):\n def test(self):\n G = {'a':{'b':1},\n 'b':{'a':1},\n 'c':{'d':1},\n 'd':{'c':1},\n 'e':{}}\n process_graph(G)\n self.assertTrue(is_connected('a', 'b'))\n self.assertFalse(is_connected('a', 'c'))\n\n G = {'a':{'b':1, 'c':1},\n 'b':{'a':1},\n 'c':{'d':1, 'a':1},\n 'd':{'c':1},\n 'e':{}}\n process_graph(G)\n self.assertTrue(is_connected('a', 'b'))\n self.assertTrue(is_connected('a', 'c'))\n self.assertFalse(is_connected('a', 'e'))\n\n\n","repo_name":"gajo357/IntroToAlgorithms","sub_path":"Week7/PreprocessGraphModule.py","file_name":"PreprocessGraphModule.py","file_ext":"py","file_size_in_byte":1625,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"38415886448","text":"\"\"\"\n作者:萌新源\n时间:2022/3/30\n操作系统:debian for raspberry pi\n修改请保留本插件的版权\n本插件版权属于萌新源\n要发布请注明出处\n\n\"\"\"\nfrom nonebot import on_keyword\nfrom nonebot.typing import T_State\nfrom nonebot.adapters.onebot.v11 import GroupMessageEvent,Bot,Message,MessageSegment,Event\nimport requests\nimport json\n\nxingzuo = on_keyword({'#星座'})\n\n@xingzuo.handle()\nasync def xz(bot: Bot, event: Event, state: T_State):\n anses = str(event.get_message()).strip()\n ansek = anses.strip('#星座')\n url = f'http://hm.suol.cc/API/xzys.php?msg={ansek}'\n x = requests.get(url)\n ansx = x.text\n b = '{br}'\n n = '\\n'\n# print(hua)\n if b in ansx:\n ansu = ansx.replace(b,n)\n await xingzuo.finish(Message(f'{ansu}'))\n else:\n await xingzuo.finish(Message(f'{ansx}'))\n \n \n\n","repo_name":"Akira-TL/my_bot1","sub_path":"src/plugins/Genshin_Constellation_Score/xzim.py","file_name":"xzim.py","file_ext":"py","file_size_in_byte":854,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"42131596840","text":"from flask import request, render_template, flash, redirect, url_for, jsonify, Flask\nfrom flask import current_app as app\nfrom flask_sqlalchemy import SQLAlchemy # 데이터베이스\nfrom flask_migrate import Migrate\n\nimport os\nimport requests\nimport json\n\n# 모듈화 파일\n\n\n# 데이터 베이스\nfrom models import db, migrate, User, guestBook\n\n# 알림 기능\nfrom informationPart import m_corona\nfrom informationPart import m_movie\nfrom informationPart import m_musicChart\nfrom informationPart import m_welcome\nfrom informationPart import m_weather\n\n# 시스템\nfrom systemPart import myPage\nfrom systemPart import signUp\nfrom systemPart import signIn\nfrom systemPart import blockId\nfrom systemPart import inventory\nfrom systemPart import shop\nfrom systemPart import ranking\nfrom systemPart import get_kakaoKey\n\n# 게임\nfrom gamePart import gamePartInfo\nfrom gamePart import beefUp\nfrom gamePart import mine\nfrom gamePart import fishing\nfrom gamePart import sneeze\nfrom gamePart import hatching\nfrom gamePart import growing\nfrom gamePart import growing_eat\nfrom gamePart import growing_play\n\nfrom minzy import minzy\n\napp = Flask(__name__)\n\n# 데이터베이스 초기화\nBASE_DIR = os.path.dirname(__file__)\ndbfile = os.path.join(BASE_DIR, 'db.sqlite')\n\napp.config['SQLALCHEMY_DATABASE_URI']= 'sqlite:///'+dbfile\napp.config['SQLALCHEMY_COMMIT_ON_TEARDOWN'] = True\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\napp.config['SECRET_KEY'] = 'hijoker'\n\ndb.init_app(app)\nmigrate.init_app(app, db, render_as_batch=True)\ndb.app = app\ndb.create_all()\n\n@app.route('/', methods=['GET','POST']) # 인덱스 페이지\ndef index():\n if request.method == 'POST':\n if not request.form['writer'] or \\\n not request.form['text']:\n flash('입력하지 않은 내용이 있습니다', 'error')\n else:\n gb = guestBook(request.form['writer'], request.form['text'])\n db.session.add(gb)\n db.session.commit()\n flash('글이 성공적으로 작성되었습니다')\n return render_template('index.html', guestBooks = guestBook.query.all())\n\n@app.route('/welcome', methods=['POST']) # 웰컴블록\ndef call_welcome_skill():\n res = m_welcome.welcome()\n return jsonify(res)\n\n@app.route('/func_quick', methods=['POST']) # 기능부분 바로가기메뉴\ndef call_quickmenu():\n res = m_welcome.quick_menu()\n return jsonify(res)\n\n@app.route('/gamePartInfo', methods=['POST']) # 기능부분 바로가기메뉴\ndef call_gamePartInfo():\n res = gamePartInfo.gamePartInfo()\n return jsonify(res)\n\n@app.route('/weather', methods=['POST']) # 날씨 알림\ndef call_weather():\n res = m_weather.weather(request.get_json())\n return jsonify(res)\n\n@app.route('/musicChart',methods=['POST']) # 멜론차트 알림\ndef call_music():\n res = m_musicChart.musicChart()\n return jsonify(res)\n \n@app.route('/movies', methods=['POST']) # 영화 알림\ndef call_movie():\n res = m_movie.movie(request.get_json())\n return jsonify(res)\n\n@app.route('/corona', methods=['POST']) #코로나 알림\ndef call_corona():\n res = m_corona.corona()\n return jsonify(res)\n \n@app.route('/myPage', methods=['POST']) # 마이페이지\ndef call_myPage():\n res = myPage.myPage(request.get_json())\n return jsonify(res)\n\n@app.route('/makeNickname', methods=['POST']) # 닉네임 있는 유저인지 확인\ndef call_makeNickname():\n res = get_kakaoKey.makeNickname(request.get_json())\n return jsonify(res)\n \n@app.route('/signUp', methods=['POST']) # 회원가입\ndef call_signUp():\n res = signUp.signUp(request.get_json())\n return jsonify(res)\n\n@app.route('/signUp_yes', methods=['POST']) # 회원가입2\ndef call_signUp_yes():\n res = signUp.signUp_yes(request.get_json())\n return jsonify(res)\n \n@app.route('/signIn', methods=['POST']) # 로그인\ndef call_signIn():\n res = signIn.signIn(request.get_json())\n return jsonify(res)\n\n@app.route('/attendance', methods=['POST']) # 출석체크\ndef call_attendance():\n res = attendance.attendance(request.get_json())\n return jsonify(res)\n \n@app.route('/inventory', methods=['POST']) # 유저 인벤토리 출력\ndef call_inventory():\n res = inventory.inventory(request.get_json())\n return jsonify(res)\n\n@app.route('/fish_inven', methods=['POST']) # 물고기 인벤토리 출력\ndef call_fish_inven():\n res = inventory.fish_inven(request.get_json())\n return jsonify(res)\n \n@app.route('/sellItem', methods=['POST']) # 아이템 판매하기\ndef call_sellItem():\n res = inventory.sellItem(request.get_json())\n return jsonify(res)\n\n@app.route('/sellItem_yes', methods=['POST']) # 아이템 판매하기 확정\ndef call_sellItem_yes():\n res = inventory.sellItem_yes(request.get_json())\n return jsonify(res)\n \n@app.route('/viewItemDescript', methods=['POST']) # 아이템 설명보기\ndef call_viewItemDescript():\n res = inventory.viewItemDescript(request.get_json())\n return jsonify(res)\n \n@app.route('/itemLock', methods=['POST']) # 아이템 설명보기\ndef call_itemLock():\n res = inventory.itemLock(request.get_json())\n return jsonify(res)\n \n@app.route('/buyAnEquipment', methods=['POST']) # 상점에서 구입 버튼 클릭시\ndef call_buyAnEquipment():\n res = shop.buyAnEquipment(request.get_json())\n return jsonify(res)\n \n@app.route('/shop', methods=['POST']) # 상점\ndef call_shop():\n res = shop.shop()\n return jsonify(res)\n \n@app.route('/shop_equipment', methods=['POST']) # 장비 상점\ndef call_shop_equipment():\n res = shop.shop_equipment()\n return jsonify(res)\n \n@app.route('/shop_pet', methods=['POST']) # 장비 상점\ndef call_shop_pet():\n res = shop.shop_pet()\n return jsonify(res)\n \n@app.route('/ranking', methods=['POST']) # 랭킹\ndef call_ranking():\n res = ranking.ranking(request.get_json())\n return jsonify(res)\n \n@app.route('/beefUp_select', methods=['POST']) # 강화 게임 장비 선택\ndef call_beefUp_select():\n res = beefUp.beefUp_select(request.get_json())\n return jsonify(res)\n \n@app.route('/beefUp', methods=['POST']) # 강화 게임\ndef call_beefUp():\n res = beefUp.beefUp(request.get_json())\n return jsonify(res)\n \n@app.route('/beefUp_try', methods=['POST']) # 강화 게임 결과\ndef call_beefUp_try():\n res = beefUp.beefUp_try(request.get_json())\n return jsonify(res)\n\n@app.route('/mine', methods=['POST']) # 채굴\ndef call_mine():\n res = mine.mine(request.get_json())\n return jsonify(res)\n \n@app.route('/fishing', methods=['POST']) # 낚시\ndef call_fishing():\n res = fishing.fishing(request.get_json())\n return jsonify(res)\n \n@app.route('/hatching', methods=['POST']) # 부화기\ndef call_hatching():\n res = hatching.hatching(request.get_json())\n return jsonify(res)\n \n@app.route('/growing', methods=['POST']) # 훈련센터\ndef call_growing():\n res = growing.growing(request.get_json())\n return jsonify(res)\n \n@app.route('/growing_eat', methods=['POST']) # 훈련센터 먹이주기\ndef call_growing_eat():\n res = growing_eat.growing_eat(request.get_json())\n return jsonify(res)\n \n@app.route('/growing_play', methods=['POST']) # 훈련센터 놀기\ndef call_growing_play():\n res = growing_play.growing_play(request.get_json())\n return jsonify(res)\n \n@app.route('/sneeze_index', methods=['POST']) # 기침게임 인덱스\ndef call_sneeze_index():\n res = sneeze.sneeze_index(request.get_json())\n return jsonify(res)\n \n@app.route('/sneeze_onclick', methods=['POST']) # 기침게임 인덱스\ndef call_sneeze_onclick():\n res = sneeze.sneeze_onclick(request.get_json())\n return jsonify(res)\n \n@app.route('/blockId', methods=['POST']) # 블록아이디 확인용\ndef call_blockId():\n res = blockId.blockId(request.get_json())\n return jsonify(res)\n \n\nif __name__ ==\"__main__\":\n app.run(debug=True,host=\"0.0.0.0\", port=1234)\n\n\n \n \n\n","repo_name":"kuyang95/1319_KakaoChatBot","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":8188,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"41722991366","text":"# NAME:\n# FILENAME:\n# SUMMARY:\n\nimport pygame\n\n\nclass Texture:\n def __init__(self, img_path, dim):\n \"\"\"\n A (perhaps overkill?) class for storing/computing/running textures in pygame\n As-is, this only repeats a SINGLE texture\n :param img_path: path to image for texture\n :param dim: width of the base image within the texture\n \"\"\"\n self.image = pygame.image.load(img_path)\n self.dim = dim # dimension of each square of the texture (e.g., 28x28, so a big image would have several)\n self.image = pygame.transform.scale(self.image, (self.dim, self.dim))\n self.image.set_alpha(85)\n\n def tiled(self, max_width, max_height):\n tiled_surf = pygame.Surface((max_width, max_height))\n for x in range(0, max_width + 1, self.dim):\n for y in range(0, max_height + 1, self.dim):\n tiled_surf.blit(self.image, (x, y)) # these makes a 1 surface made of several tiles of self.image\n return tiled_surf\n","repo_name":"Sown-Dev/CMPT330","sub_path":"HW5/textures.py","file_name":"textures.py","file_ext":"py","file_size_in_byte":1006,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"71563551264","text":"from Bio import SeqIO\nfrom Bio.Seq import Seq\nfrom Bio.Align import PairwiseAligner\nfrom tqdm import tqdm\n\nnon_amb_nucs = [\"A\",\"G\",\"C\",\"T\"]\n\nreference = next(SeqIO.parse(open(snakemake.input[0],\"r\"),\"fasta\")).seq\nsequences = list(SeqIO.parse(open(snakemake.input[1],\"r\"),\"fasta\"))\n\naligner = PairwiseAligner()\naligner.mode = \"global\"\naligner.match_score = 1\naligner.mismatch_score = -2\naligner.gap_score = -2\n\nsequence_output = []\n\nfor seq_ref in tqdm(sequences):\n\tseq = seq_ref.seq\n\tseq_str = str(seq_ref.seq).upper()\n\n\tnot_good = False\n\tfor nuc in seq_str:\n\t\tif nuc not in non_amb_nucs:\n\t\t\tnot_good = True\n\t\t\tbreak\n\tif not_good:\n\t\tcontinue\n\n\tseq_ref.seq = Seq(seq_str)\n\n\tif len(seq) != len(reference):\n\t\tcontinue\n\talignment = aligner.align(reference, seq)[0]\n\n\tif \"-\" in alignment.target or \"-\" in alignment.query:\n\t\tcontinue\n\n\tsequence_output.append(seq_ref)\n\nSeqIO.write(sequence_output,open(snakemake.output[0],\"w\"),\"fasta\")\n\n","repo_name":"ArmaanAhmed22/NonSpecificLAMPDesign","sub_path":"preprocess/sequence_filter.py","file_name":"sequence_filter.py","file_ext":"py","file_size_in_byte":931,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"74861600864","text":"class Vertex:\n def __init__(self, name):\n self.name = name\n\n\ngraph = {\n # 有向无环图\n # 'A': ['B', 'C'],\n # 'B': ['E'],\n # 'C': ['D'],\n # 'D': ['B'],\n # 'E': [],\n\n # 有向有环图\n # 'A': ['B'],\n # 'B': ['D', 'E'],\n # 'C': ['A'],\n # 'D': ['C'],\n # 'E': [],\n\n 'A': ['B', 'C', 'D'],\n 'B': ['D', 'E'],\n 'C': ['F'],\n 'D': ['C', 'F', 'G'],\n 'E': ['D', 'G'],\n 'F': [],\n 'G': ['F']\n}\n\n\ndef top_sort():\n '''\n 简单的拓扑排序\n :return: 打印节点名和顺序\n '''\n for counter in range(0, len(graph)):\n # 查找入度为0的顶点\n v = findNewVertexOfIndegreeZero()\n # 如果查找不到,说明存在环\n if v is None:\n raise Exception('存在环')\n print('{}的顺序是{}'.format(v, counter))\n # 删除该顶点以及其边\n del graph[v]\n\n\ndef findNewVertexOfIndegreeZero():\n '''\n 查找入度为0的节点\n :return:\n '''\n # 外层循环,遍历所有的节点\n for v in graph.keys():\n # 内层循环,遍历每个节点的邻接节点列表\n for l in graph.values():\n # 如果发现当前节点出现在某节点的邻接列表里,说明它的入度不为0,跳出循环\n if v in l:\n break\n # 如果未出现在所有的邻接列表里,则说明入度为0,return它\n else:\n return v\n\n\ndef top_sort2():\n # 将所有的节点入度初始化为0\n in_degrees = dict((u, 0) for u in graph.keys())\n # 计算所有节点的入度,访问所有顶点的邻接表,出现的次数就是入度\n for v in graph.keys():\n for w in graph[v]:\n in_degrees[w] += 1\n # 筛选入度为0的顶点\n q = [u for u in in_degrees if in_degrees[u] == 0]\n seq = []\n #使用列表的append和pop实现栈,保存入度为0的顶点\n while q:\n u = q.pop()\n seq.append(u)\n # 获取与u邻接的顶点并将其入度-1\n for v in graph[u]:\n in_degrees[v] -= 1\n if in_degrees[v] == 0:\n q.append(v)\n #如果排序后的顶点数和图中的顶点数相同\n if len(seq) == len(in_degrees):\n print(seq)\n return seq\n else:\n return None\n\n\n\nif __name__ == '__main__':\n # top_sort()\n # top_sort2()\n q = []\n q.append(1)\n q.append(2)\n q.append(3)\n print(q)\n a = q.pop()\n print(a)\n print(q)","repo_name":"zjxht62/LearnPython","sub_path":"DataStructureAndAlgorithm/graph/graph_demo.py","file_name":"graph_demo.py","file_ext":"py","file_size_in_byte":2476,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"16232745050","text":"# -*- coding: utf-8 -*-\nimport os\n\n# Project absolute path\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\n\n# Scrapy settings for hopwork project\nBOT_NAME = 'hopwork'\n\nSPIDER_MODULES = ['hopwork.spiders']\nNEWSPIDER_MODULE = 'hopwork.spiders'\n\n# Obey robots.txt rules\nROBOTSTXT_OBEY = False\n\n# Configure maximum concurrent requests performed by Scrapy\nCONCURRENT_REQUESTS = 16\n\n# Configure a delay for requests for the same website\nDOWNLOAD_DELAY = 0\n\n# Enable http cookies\nCOOKIES_ENABLED = True\n\n# Override the default request headers:\nDEFAULT_REQUEST_HEADERS = {\n 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',\n 'Accept-Language': 'en-US,en;q=0.8,ru;q=0.6,uk;q=0.4,und;q=0.2',\n 'Referer': 'https://www.hopwork.fr',\n 'User-Agent': 'Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)',\n}\n\n# Enable or disable downloader middlewares\n# See http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html\nDOWNLOADER_MIDDLEWARES = {\n # 'scrapy.contrib.downloadermiddleware.useragent.UserAgentMiddleware': None,\n # 'random_useragent.RandomUserAgentMiddleware': 400\n}\n\n# Configure item pipelines\n# See http://scrapy.readthedocs.org/en/latest/topics/item-pipeline.html\nITEM_PIPELINES = {\n 'hopwork.pipelines.HopworkPipeline': 300,\n}\n\n# Show skipped requests\nDUPEFILTER_DEBUG = True","repo_name":"lisneifild/hopwork","sub_path":"hopwork/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":1363,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"36636626297","text":"import json, yaml\nfrom flask import Flask, request\nfrom prometheus_flask_exporter import PrometheusMetrics\nfrom os import path\nimport time\nfrom kubernetes import client, config\n\nconfig.load_kube_config()\napp = Flask(__name__)\nmetrics = PrometheusMetrics(app)\n\n\n@app.route('/regist', methods=['POST'])\ndef robot_registration():\n response = {'success': False}\n\n # parameters = {'name': ''}, request.post\n paramaters = json.loads(request.get_data())\n name = paramaters['name']\n response['success'] = True\n k8s_apps_v1 = client.AppsV1Api()\n\n # statepublisher.yaml\n with open(path.join(path.dirname(__file__), \"rb-function/statepublisher.yaml\")) as file:\n dep = yaml.safe_load(file)\n # paramater change\n dep['spec']['template']['spec']['containers'][0]['env'][1]['value'] ='{}'.format(name)\n dep['metadata']['name'] = 'ros-state-publisher-{}'.format(name.replace('_', '-'))\n resp = k8s_apps_v1.create_namespaced_deployment(body=dep, namespace=\"default\")\n print(\"Deployment created. status='%s'\" % resp.metadata.name)\n\n time.sleep(5)\n\n # Deployment create, rb_name = name\n # slamgmapping.yaml\n with open(path.join(path.dirname(__file__), \"rb-function/slamgmapping.yaml\")) as file:\n dep3 = yaml.safe_load(file)\n # paramater change\n dep3['spec']['template']['spec']['containers'][0]['env'][1]['value'] ='{}'.format(name)\n dep3['metadata']['name'] = 'ros-slam-gmapping-{}'.format(name.replace('_', '-'))\n k8s_apps_v1 = client.AppsV1Api()\n resp = k8s_apps_v1.create_namespaced_deployment(body=dep3, namespace=\"default\")\n print(\"Deployment created. status='%s'\" % resp.metadata.name)\n\n time.sleep(5)\n\n # mapmerge.yaml\n try:\n k8s_apps_v1.delete_namespaced_deployment(name='ros-map-merge', namespace='default')\n except:\n print('map merge is not existe')\n with open(path.join(path.dirname(__file__), \"rb-function/mapmerge.yaml\")) as file:\n dep6 = yaml.safe_load(file)\n # paramater change\n resp = k8s_apps_v1.create_namespaced_deployment(body=dep6, namespace=\"default\")\n print(\"Deployment created. status='%s'\" % resp.metadata.name)\n return response\n\n\n@app.route('/')\ndef get_check():\n return \"checking Running Server\"\n\n\nif __name__=='__main__':\n app.run(host='114.70.21.161', port='8082')\n\n","repo_name":"ChanghyunRyu/R2C-Robot_to_Cloud","sub_path":"cloud_api/regist-api-server.py","file_name":"regist-api-server.py","file_ext":"py","file_size_in_byte":2381,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"35717547533","text":"from sqlalchemy.orm import Session\n\nimport models, schemas\n\n\n\ndef get_items(db: Session, skip: int = 0, limit: int = 100):\n return db.query(models.Item).offset(skip).limit(limit).all()\n\n\ndef get_item_id(db: Session, item_id: int):\n return db.query(models.Item).filter(models.Item.id == item_id).first()\n\ndef get_item_title(db: Session, title: str):\n return db.query(models.Item).filter(models.Item.title == title).first()\n\n\ndef create_item(db: Session, item: schemas.ItemCreate):\n db_item = models.Item(**item.dict())\n db.add(db_item)\n db.commit()\n db.refresh(db_item)\n return db_item\n#crud.mod_item(db, item_id=item_id,item = item\ndef mod_item(db: Session, item: schemas.Item):\n db_item = db.query(models.Item).filter(models.Item.id == item.id).first()\n if not db_item:\n raise HTTPException(status_code=404, detail=\"Item not found\")\n item_data = item.dict(exclude_unset=True)\n for key, value in item_data.items():\n setattr(db_item, key, value)\n db.add(db_item)\n db.commit()\n db.refresh(db_item)\n \n return db_item\n\ndef del_item(db: Session, item_id: int):\n db_item =db.query(models.Item).filter(models.Item.id == item_id).first()\n db.delete(db_item)\n db.commit()\n #db.refresh(db_item)\n return db_item \n","repo_name":"Antonioarf/megadados_p1","sub_path":"crud.py","file_name":"crud.py","file_ext":"py","file_size_in_byte":1286,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"43754667518","text":"from mtda.console.serial import SerialConsole\nfrom mtda.support.usb import Composite\n\n\nclass UsbFunctionConsole(SerialConsole):\n\n def __init__(self, mtda):\n super().__init__(mtda)\n self.hotplug = True\n self.port = None\n self.rate = 9600\n Composite.mtda = mtda\n\n def configure(self, conf, role='console'):\n self.mtda.debug(3, \"console.usbf.configure()\")\n\n super().configure(conf)\n if self.port is None:\n self.port = \"/dev/ttyGS0\" if role == \"console\" else \"/dev/ttyGS1\"\n result = Composite.configure(role, conf)\n\n self.mtda.debug(3, \"console.usbf.configure(): {}\".format(result))\n return result\n\n def configure_systemd(self, dir):\n return None\n\n\ndef instantiate(mtda):\n return UsbFunctionConsole(mtda)\n","repo_name":"siemens/mtda","sub_path":"mtda/console/usbf.py","file_name":"usbf.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","stars":31,"dataset":"github-code","pt":"7"} +{"seq_id":"14198841425","text":"# from pudb import set_trace; set_trace()\nfrom typing import List\nimport math\nfrom collections import defaultdict\n\n\nclass DSU:\n def __init__(self, N: int):\n self.par = list(range(N))\n self.rnk = [0] * N\n # zero must be the ultimate parent, such that whoever unions with zero,\n # its parent will be zero, regardless of when the union happens\n self.rnk[0] = math.inf\n\n def find(self, x: int) -> int:\n if self.par[x] != x:\n self.par[x] = self.find(self.par[x])\n return self.par[x]\n\n def union(self, x: int, y: int) -> bool:\n x_par, y_par = self.find(x), self.find(y)\n if x_par == y_par:\n return False\n if self.rnk[x_par] > self.rnk[y_par]:\n self.par[y_par] = x_par\n elif self.rnk[x_par] < self.rnk[y_par]:\n self.par[x_par] = y_par\n else:\n self.rnk[x_par] += 1\n self.par[y_par] = x_par\n return True\n\n def detach(self, x: int) -> None:\n \"\"\"Detach x from any group it might have been associated with.\"\"\"\n self.par[x] = x\n self.rnk[x] = 0\n\n\nclass Solution:\n def findAllPeople(self, n: int, meetings: List[List[int]], firstPerson: int) -> List[int]:\n \"\"\"This is a good problem. It is apparently union-find, but there are\n two twists.\n\n First, we need to deal with the situation like this\n\n [[3,1,3],[1,2,2],[0,3,3]] with firstPerson = 3\n\n The first meeting is between 1 and 2; neither has secret. The second\n meeting is between 3 and 1, so now 1 has secret. However, the meeting\n between 1 and 2 happens earlier. Thus, 2 still does not have secret,\n despite 1 having secret. If we use a naive union-find, where 1 and 2\n are unioned. Then when 1 and 3 are unioned, 2 would've been unioned\n with 3 as well, making him share secret. We must break such tie. Hence\n during iteration of the meeting (in ascending time order), whenever all\n the meetings of the same time have ended, we check to see if any of the\n people involved does not share secret. For such outsider, we must cut\n its tie to any of the union before.\n\n Second, we want anyone that union with someone else who has zero as\n parent to also have zero as parent. However, since the union is decided\n by the rank of the parent, it is likely the rank of zero is lower than\n the rank of another parent, if the other parent is unioned ahead of\n time. Therefore, we must arbitrarily set the rank of zero maximum,\n thus guaranteeing that whoever unions with some with parent as zero\n also gets zero as parent.\n\n O(MlogM + (M + N)alpha(N)), where M = len(meetings), and alpha(N) is\n the inverse function of Ackermann function, which describes the time\n complexity of union find with path compression.\n\n 5625 ms, faster than 12.50%\n\n Time complexity inspired by: https://leetcode.com/problems/find-all-people-with-secret/discuss/1599815/C%2B%2B-Union-Find\n \"\"\"\n meetings.sort(key=lambda tup: tup[2])\n dsu = DSU(n)\n dsu.union(0, firstPerson)\n pre_t = 0\n outsider = set()\n for x, y, t in meetings:\n if t != pre_t:\n for p in outsider:\n if dsu.find(p) != 0:\n # at the end of a specific time, if a person is not\n # part of the secret, he must be removed of any\n # connection to any other outsider who might become an\n # insider later on. This is to avoid such other outsider\n # becoming an insider pulls the current person into the\n # secret as well\n dsu.detach(p)\n outsider = set()\n dsu.union(x, y)\n pre_t = t\n if dsu.find(x):\n outsider.add(x)\n if dsu.find(y):\n outsider.add(y)\n return [i for i in range(n) if dsu.find(i) == 0]\n\n\nsol = Solution()\ntests = [\n (6, [[1,2,5],[2,3,8],[1,5,10]], 1, [0,1,2,3,5]),\n (4, [[3,1,3],[1,2,2],[0,3,3]], 3, [0,1,3]),\n (5, [[3,4,2],[1,2,1],[2,3,1]], 1, [0,1,2,3,4]),\n (6, [[0,2,1],[1,3,1],[4,5,1]], 1, [0,1,2,3]),\n (6, [[0,2,1],[1,3,1],[4,5,1],[1,4,1]], 1, [0,1,2,3,4,5]),\n (5, [[1,4,3],[0,4,3]], 3, [0,1,3,4]),\n (4, [[1,2,1],[0,3,1],[2,0,1]], 3, [0,1,2,3])\n]\n\nfor i, (n, meetings, firstPerson, ans) in enumerate(tests):\n res = sol.findAllPeople(n, meetings, firstPerson)\n res.sort()\n ans.sort()\n if res == ans:\n print(f'Test {i}: PASS')\n else:\n print(f'Test {i}; Fail. Ans: {ans}, Res: {res}')\n","repo_name":"FanchenBao/leetcode","sub_path":"LeetCode_2092.py","file_name":"LeetCode_2092.py","file_ext":"py","file_size_in_byte":4743,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"4748626636","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Jan 27 15:09:30 2015\n\n@author: Pete\n\nThis module is a test processing module for data created with TurbineDAQ.\n\n\"\"\"\n\nfrom __future__ import division, print_function\nimport pandas as pd\n\nprint(\"Imported processing module\")\n\ndef process_run(section, nrun):\n print(\"Processing\", section, \"run\", nrun)\n summary = pd.Series()\n summary[\"mean_cp\"] = 0.3\n return summary\n\ndef process_latest_run(section):\n print(\"Processing latest run of\", section)","repo_name":"petebachant/TurbineDAQ","sub_path":"test/Modules/processing.py","file_name":"processing.py","file_ext":"py","file_size_in_byte":498,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"} +{"seq_id":"5153067986","text":"class GameTickPacket:\n def __init__(self, bots):\n self.bots = bots\n self.rockets = [rockets()] * len(self.bots)\n self.bounds = bounds()\n self.latest_tick = 0\n self.attributes = [None, None]\n\n def update(self, index):\n self.rockets[index].verticalSpeed.append(self.bots[index].verticalSpeed[0])\n self.rockets[index].horizontalSpeed.append(self.bots[index].horizontalSpeed[0])\n self.rockets[index].x.append(self.bots[index].x)\n self.rockets[index].y.append(self.bots[index].y)\n self.rockets[index].isDead.append(self.bots[index].isDead)\n self.rockets[index].thrust.append(self.bots[index].thrust_power)\n self.attributes[0] = self.bounds.attributes\n self.attributes[1] = self.rockets[index].attributes\n # print(self.attributes)\n self.latest_tick += 1\n\n def reset(self):\n self.__init__(self.bots)\n\n\nclass rockets:\n def __init__(self):\n self.verticalSpeed = []\n self.x = []\n self.y = []\n self.horizontalSpeed = []\n self.thrust = []\n self.isDead = []\n self.attributes = [self.verticalSpeed, self.x, self.y, self.horizontalSpeed, self.thrust, self.isDead]\n\n\nclass bounds:\n def __init__(self):\n self.x = 640\n self.y = 480\n self.attributes = (self.x, self.y)\n","repo_name":"iamPres/self-landing-rockets","sub_path":"GameTickPacket.py","file_name":"GameTickPacket.py","file_ext":"py","file_size_in_byte":1347,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7552410951","text":"from libdesklets.controls import Control\nfrom IArrayBuffer import IArrayBuffer\n\n\n#----------------------------------------------------------------------------------#\n#\n# ArrayBuffer Class\n#\n#\nclass ArrayBuffer(Control, IArrayBuffer):\n\n #\n # ArrayBuffer Constructor\n #\n def __init__(self):\n\n self.__buffer = []\n self.__size = 0\n self.__cursor = 0\n self.__window_pos = 0\n self.__window_size = 0\n self.__filltype = None\n\n Control.__init__(self)\n\n\n #\n # Get the current object at cursor position\n #\n def __read(self):\n\n return self.__buffer[self.__window_pos:self.__window_pos + \\\n self.__window_size]\n\n\n #\n # Get the entire buffer\n #\n def __read_all(self):\n\n return self.__buffer\n\n\n #\n # Deletes a single line from the buffer\n #\n def __delete(self, pos):\n\n self.__buffer = [self.__filltype] + self.__buffer[:pos] + \\\n self.__buffer[pos + 1:]\n\n\n #\n # Fill the buffer with an object\n #\n def __fill(self, obj):\n\n for i in range(self.__size):\n self.__buffer[i] = obj\n\n\n #\n # Write object to current cursor position within buffer\n #\n def __write(self, objArr):\n\n n = len(objArr)\n\n if self.__cursor + n > self.__size:\n if n >= self.__size:\n tmp = n - self.__size\n self.__buffer = objArr[tmp:]\n\n else:\n tmp = self.__size - n\n self.__buffer = self.__buffer[self.__cursor - \\\n tmp:self.__cursor]\n self.__buffer += objArr\n\n self.__cursor = self.__size\n\n else:\n self.__buffer[self.__cursor:] = objArr + \\\n self.__buffer[self.__cursor + n:]\n self.__cursor += n\n\n\n #\n # Get the current viewable window position\n #\n def __get_window_pos(self):\n\n return self.__window_pos\n\n\n #\n # Get the filltype\n #\n def __get_filltype(self):\n\n return self.__filltype\n\n\n #\n # Sets the filltype used when resizing the buffer\n #\n def __set_filltype(self, obj):\n\n self.__filltype = obj\n\n\n #\n # Set the viewable window position\n #\n def __set_window_pos(self, pos):\n\n if pos < 0:\n pos = 0\n\n if pos + self.__window_size > self.__size:\n pos = self.__size - self.__window_size\n\n self.__window_pos = pos\n\n\n #\n # Get the size of the viewable window\n #\n def __get_window_size(self):\n\n return self.__window_size\n\n\n #\n # Set the size of the viewable window\n #\n def __set_window_size(self, size):\n\n if size > self.__size:\n size = self.__size\n\n self.__window_size = size\n\n\n #\n # Get the size of the buffer\n #\n def __get_size(self):\n\n return self.__size\n\n\n #\n # Set the size of the buffer\n #\n def __set_size(self, size):\n\n if size > self.__size:\n\n tmp = [ self.__filltype for i in range(size - self.__size) ]\n\n self.__buffer = tmp + self.__buffer\n\n else:\n self.__buffer = self.__buffer[self.__size - size:]\n\n self.__size = size\n\n\n #\n # Get the current buffer cursor position\n #\n def __get_cursor(self):\n\n return self.__cursor\n\n\n #\n # Set the position of the buffer cursor\n #\n def __set_cursor(self, pos):\n\n if pos < 0:\n pos = 0\n\n if pos > self.__size:\n pos = self.__size\n\n self.__cursor = pos\n\n\n #\n # Interface\n #\n read = property(fget = __read,\n doc = \"Return the viewable window array\")\n read_all = property(fget = __read_all,\n doc = \"Return the entire buffer array\")\n delete = property(fset = __delete,\n doc = \"Delete line from buffer\")\n write = property(fset = __write,\n doc = \"Write line at position cursor\")\n fill = property(fset = __fill,\n doc = \"Fill the buffer with an object\")\n filltype = property(fget = __get_filltype, fset = __set_filltype,\n doc = \"Get/Set the empty space fill type\")\n window_pos = property(fget = __get_window_pos, fset = __set_window_pos,\n doc = \"Get/Set viewable window position\")\n window_size = property(fget = __get_window_size, fset = __set_window_size,\n doc = \"Get/Set viewable window size\")\n size = property(fget = __get_size, fset = __set_size,\n doc = \"Get/Set buffer size\")\n cursor = property(fget = __get_cursor, fset = __set_cursor,\n doc = \"Get/Set cursor position\")\n\n\n#----------------------------------------------------------------------------------#\n\ndef get_class(): return ArrayBuffer\n\n","repo_name":"RaumZeit/gdesklets-core","sub_path":"Controls/ArrayBuffer/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":5945,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"17285871882","text":"from flask_sqlalchemy import SQLAlchemy\nfrom datetime import datetime\n\ndb = SQLAlchemy()\n\n\nclass Task(db.Model):\n id = db.Column(db.Integer , primary_key=True)\n name = db.Column(db.String( 250), nullable=False)\n description = db.Column(db.Text)\n created_at = db.Column(db.DateTime)\n completed = db.Column(db.Boolean , default=False)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.created_at = datetime.now() ","repo_name":"Kholoud731/Flask-API","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"2690658187","text":"import pandas as pd\nimport numpy as np\nimport sqlite3 as sql\nfrom sklearn.ensemble import RandomForestClassifier\nfrom pprint import pprint\nimport itertools\nfrom matplotlib import pyplot as plt\nfrom sklearn.metrics import confusion_matrix\n\npd.options.display.expand_frame_repr = False\n\n\ndef plot_confusion_matrix(cm, fig, ax, classes,\n normalize=False,\n cmap=plt.cm.Blues):\n \"\"\"\n This function prints and plots the confusion matrix.\n Normalization can be applied by setting `normalize=True`.\n \"\"\"\n im = ax.imshow(cm, interpolation='nearest', cmap=cmap)\n tick_marks = range(len(classes))\n ax.set_xticks(tick_marks)\n ax.set_yticks(tick_marks)\n ax.set_xticklabels(classes)\n ax.set_yticklabels(classes)\n\n if normalize:\n cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]\n\n thresh = cm.max() / 2.\n for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])):\n ax.text(j, i, cm[i, j],\n horizontalalignment=\"center\",\n color=\"white\" if cm[i, j] > thresh else \"black\")\n ax.set_ylabel('True label')\n ax.set_xlabel('Predicted label')\n ax.invert_yaxis()\n\nconn = sql.connect('Data/database.sqlite')\n\n# query = 'SELECT * FROM Match INNER JOIN League on League.id = Match.league_id'\n\nquery = 'SELECT * FROM Match WHERE league_id == 1729'\n\n# fields = ['home_team_api_id', 'away_team_api_id', 'date', 'home_team_goal', 'away_team_goal']\n# more_fields = ['goal', 'shoton', 'shotoff', 'possession', 'cross', 'corner', 'foulcommit']\n# odds = ['B365H', 'B365D', 'B365A', 'BWH', 'BWD', 'BWA', 'IWH', 'IWD', 'IWA', 'LBH', 'LBD', 'LBA', 'PSH', 'PSD']\n\ndf = pd.read_sql_query(query, conn, index_col='id')\n\ndf['GD'] = df['home_team_goal'] - df['away_team_goal']\ndf = df.loc[df['GD'] != 0]\ndf['result'] = np.where(df['GD'] > 0, 1, 0)\n\n# CHOOSE FEATURES\n# --------------------------------------------------------------------------------------------------\nFEATURES = ['B365H', 'BWH', 'LBH']\ndf = df.loc[:, ['result'] + FEATURES]\n# --------------------------------------------------------------------------------------------------\n\ndf = df.sample(frac=1) # shuffle\ndf = df.dropna() # drop NaNs\n\nsplit = 0.80\nsplit_idx = int(split * len(df))\ntrain_df = df.iloc[:split_idx, :]\ntest_df = df.iloc[split_idx:, :]\n\n\nX_train = train_df.loc[:, FEATURES]\nT_train = train_df.loc[:, 'result']\n\n\nRF = RandomForestClassifier()\nRF.fit(X_train, T_train)\ntraining_results = RF.predict(X_train)\n\nY = training_results\nT = T_train\ncm = confusion_matrix(y_true=T, y_pred=Y)\nfig, (ax1, ax2) = plt.subplots(1, 2)\nax1.set_title('Train results')\nplot_confusion_matrix(cm, fig, ax1, ['loss', 'win'])\n\n# TEST\nX_test = test_df.loc[:, FEATURES]\nT_test = test_df.loc[:, 'result']\n\ntesting_results = RF.predict(X_test)\n\nY = testing_results\nT = T_test\ncm = confusion_matrix(y_true=T, y_pred=Y)\nplot_confusion_matrix(cm, fig, ax2, ['loss', 'win'])\nax2.set_title('Test results')\nplt.tight_layout()\nplt.show()\n","repo_name":"AvraSaslow/Soccer-Events-Prediction-Model","sub_path":"democlassifier.py","file_name":"democlassifier.py","file_ext":"py","file_size_in_byte":3006,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"72844830942","text":"from __future__ import print_function\nimport argparse\nimport torch\n#from utils.nms.py_cpu_nms import py_cpu_nms\nfrom models.faceboxes import FaceBoxes\n\nparser = argparse.ArgumentParser(description='FaceBoxes')\n\nparser.add_argument('-m', '--trained_model', default='weights/FaceBoxesProd.pth',\n type=str, help='Trained state_dict file path to open')\nparser.add_argument('--save_folder', default='onnx/FaceBoxes.onnx', type=str, help='Dir to save results')\nparser.add_argument('--cpu', action=\"store_true\", default=True, help='Use cpu inference')\nparser.add_argument('--dataset', default='PASCAL', type=str, choices=['AFW', 'PASCAL', 'FDDB'], help='dataset')\nparser.add_argument('--confidence_threshold', default=0.05, type=float, help='confidence_threshold')\nparser.add_argument('--top_k', default=5000, type=int, help='top_k')\nparser.add_argument('--nms_threshold', default=0.3, type=float, help='nms_threshold')\nparser.add_argument('--keep_top_k', default=750, type=int, help='keep_top_k')\nparser.add_argument('-s', '--show_image', action=\"store_true\", default=False, help='show detection results')\nparser.add_argument('--vis_thres', default=0.5, type=float, help='visualization_threshold')\nargs = parser.parse_args()\n\n\ndef check_keys(model, pretrained_state_dict):\n ckpt_keys = set(pretrained_state_dict.keys())\n model_keys = set(model.state_dict().keys())\n used_pretrained_keys = model_keys & ckpt_keys\n unused_pretrained_keys = ckpt_keys - model_keys\n missing_keys = model_keys - ckpt_keys\n print('Missing keys:{}'.format(len(missing_keys)))\n print('Unused checkpoint keys:{}'.format(len(unused_pretrained_keys)))\n print('Used keys:{}'.format(len(used_pretrained_keys)))\n assert len(used_pretrained_keys) > 0, 'load NONE from pretrained checkpoint'\n return True\n\n\ndef remove_prefix(state_dict, prefix):\n ''' Old style model is stored with all names of parameters sharing common prefix 'module.' '''\n print('remove prefix \\'{}\\''.format(prefix))\n f = lambda x: x.split(prefix, 1)[-1] if x.startswith(prefix) else x\n return {f(key): value for key, value in state_dict.items()}\n\n\ndef load_model(model, pretrained_path, load_to_cpu):\n print('Loading pretrained model from {}'.format(pretrained_path))\n if load_to_cpu:\n pretrained_dict = torch.load(pretrained_path, map_location=lambda storage, loc: storage)\n else:\n device = torch.cuda.current_device()\n pretrained_dict = torch.load(pretrained_path, map_location=lambda storage, loc: storage.cuda(device))\n if \"state_dict\" in pretrained_dict.keys():\n pretrained_dict = remove_prefix(pretrained_dict['state_dict'], 'module.')\n else:\n pretrained_dict = remove_prefix(pretrained_dict, 'module.')\n check_keys(model, pretrained_dict)\n model.load_state_dict(pretrained_dict, strict=False)\n return model\n\n\nif __name__ == '__main__':\n \n torch.set_grad_enabled(False)\n # net and model\n net = FaceBoxes(phase='test', size=None, num_classes=2) # initialize detector\n net = load_model(net, args.trained_model, args.cpu)\n net.eval()\n input_names = [\"image\"]\n output_names = [\"class\",\"loc\"]\n dynamic_axes = {'image': {0: '-1'}, 'class': {0: '-1'}, 'loc': {0: '-1'}}\n dummy_input = torch.randn(1, 3, 1024, 1024)\n torch.onnx.export(net, dummy_input, args.save_folder, input_names = input_names, dynamic_axes = dynamic_axes, output_names = output_names, opset_version=11, verbose=True)\n\n\n\n \n","repo_name":"Ascend/ModelZoo-PyTorch","sub_path":"ACL_PyTorch/contrib/cv/face/FaceBoxes/faceboxes_pth2onnx.py","file_name":"faceboxes_pth2onnx.py","file_ext":"py","file_size_in_byte":3483,"program_lang":"python","lang":"en","doc_type":"code","stars":31,"dataset":"github-code","pt":"7"} +{"seq_id":"23758773151","text":"def merge(*args):\n fin = args[0] # на вывод\n my_keys = fin.keys() #мои ключи\n length = len(args)\n for i in range(1, length):\n temp_keys = args[i].keys() #проверк\n for k in temp_keys:\n if k in my_keys:\n if type(fin[k]) is list:\n temp = fin[k]\n else:\n temp = [fin[k]]\n temp.append(args[i][k])\n fin[k] = temp\n my_keys = fin.keys()\n else:\n fin[k] = args[i][k]\n return fin\n\na = {1:2, 3:4}\nb = {1:10, 2:5, 7:10}\nc = {1:7, 2:10}\nprint(merge(a, b, c))","repo_name":"paladinchik56/for-ITIS-by-Grisha","sub_path":"1 curse/test3/deb.py","file_name":"deb.py","file_ext":"py","file_size_in_byte":645,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"40635628806","text":"##10072014c.py\r\n##John brook\r\n##class notes from 10-07-2014\r\n\r\n\r\n\r\n##============================================================##\r\n###Base class\r\n##\r\n##class Critter(object):\r\n## \"\"\"A virtual pet\"\"\"\r\n## def talk(self): #class methoid #callable function\r\n## print(\"Hi. I'm an instance of class Critter.\")\r\n##\r\n##def main():\r\n## crit = Critter() #shortens class\r\n## crit.talk() #class w/ fun\r\n##\r\n##main()\r\n\r\n\r\n##============================================================##\r\n\r\n##class Critter(object):\r\n## \"\"\"A virtual pet\"\"\"\r\n## def talk(self): #class methoid #callable function\r\n## print(\"Hi. I'm an instance of class Critter.\")\r\n## def __init__(self, legs, color, toxic, name):\r\n## self.name = name\r\n## self.legs = legs\r\n## self.color = color\r\n## self.toxic = toxic\r\n## def __str__(self):\r\n## rep = \"Critter object\\n\"\r\n## rep += \"name: \" + self.name + \"\\n\"\r\n## rep += \"legs: \" + self.legs + \"\\n\"\r\n## rep += \"color: \" + self.color + \"\\n\"\r\n## rep += \"toxic: \" + self.toxic + \"\\n\"\r\n## return rep\r\n##\r\n##\r\n##def main():\r\n## crit = Critter(\"6\",\"blue\",\"yes\",\"frank\") #shortens class\r\n## crit.talk() #class w/ fun\r\n## print(crit)\r\n##\r\n## crit1 = Critter(\"6\",\"purple\",\"no\",\"bob\") #shortens class\r\n## crit.talk() #class w/ fun\r\n## print(crit1)\r\n##\r\n##main()\r\n##============================================================##\r\n\r\nclass playerClass(object):\r\n \"\"\"A virtual pet\"\"\"\r\n def talk(self): #class methoid #callable function\r\n print(\"Hi. I'm an instance of class Critter.\")\r\n def __init__(self,name, HP, MP, ATK, DEF):\r\n self.name = name\r\n self.HP = HP\r\n self.MP = MP\r\n self.ATK = ATK\r\n self.DEF = DEF\r\n \r\n def __str__(self):\r\n rep = \"Critter object\\n\"\r\n rep += \"Name: \" + self.name + \"\\n\"\r\n rep += \"HP: \" + self.HP + \"\\n\"\r\n rep += \"MP: \" + self.MP + \"\\n\"\r\n rep += \"ATK: \" + self.ATK + \"\\n\"\r\n rep += \"DEF: \" + self.DEF + \"\\n\"\r\n return rep\r\n def update_health(self, delta):\r\n self.health = delta\r\n\r\n\r\ndef main():\r\n crit = playerClass(\"bob\",\"6\",\"21\",\"22\",\"22\") #shortens class\r\n crit.talk() #class w/ fun\r\n crit.update_health(\"+5\")\r\n print(crit)\r\n\r\n\r\n\r\nmain()\r\n##============================================================##\r\n","repo_name":"ChocolaKuma/MCC_Learning_Python","sub_path":"10072014c.py","file_name":"10072014c.py","file_ext":"py","file_size_in_byte":2372,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"7681407057","text":"\"\"\"RegEx finditer function\"\"\"\n\n\nimport re\nimport requests\n\n\n# re.finditer(pattern, string, flags=0)\nhtml = requests.get(\"https://docs.python.org/2/library/re.html\").text\npattern = r\"\\b(the\\s+\\w+)\\s+\"\nregex = re.compile(pattern, re.IGNORECASE)\nfor match in regex.finditer(html):\n print(f\"{match.start()}: {match.group(1)}\")\n","repo_name":"Maulik5041/Everyday-coding","sub_path":"Regex/finditer.py","file_name":"finditer.py","file_ext":"py","file_size_in_byte":326,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"1738516638","text":"from django.shortcuts import render\nfrom .models import Annonce\nfrom .forms import Form_Annonce, Form_Add\nfrom django.db.models import Avg\nimport requests\nimport json\n\n# Create your views here.\n\ndef accueil(request):\n ann_form = Form_Annonce()\n add_form = Form_Add()\n\n result_moy, quartile10, quartile90, err = None, None, None, None\n message = None\n\n if request.method == \"POST\":\n form_a = Form_Annonce(request.POST)\n form_l = Form_Add(request.POST)\n if form_a.is_valid():\n data = form_a.cleaned_data\n if data['code_ville']:\n result_moy = Annonce.objects.filter(code_ville = data['code_ville']).aggregate(moyenne = Avg('prix'))['moyenne']\n liste_prix = Annonce.objects.filter(code_ville = data['code_ville']).order_by('prix').values_list('prix', flat=True)\n if len(liste_prix) != 0:\n quartile10 = liste_prix[len(liste_prix) // 10]\n quartile90 = liste_prix[len(liste_prix) * 9 // 10]\n else: err = True\n elif data['ville']:\n data['ville'] = data['ville'].lower()\n result_moy = Annonce.objects.filter(ville = data['ville']).aggregate(moyenne = Avg('prix'))['moyenne']\n liste_prix = Annonce.objects.filter(ville = data['ville']).order_by('prix').values_list('prix', flat=True)\n if len(liste_prix) != 0:\n quartile10 = liste_prix[len(liste_prix) // 10]\n quartile90 = liste_prix[len(liste_prix) * 9 // 10]\n else: err = True\n elif data['departement']:\n result_moy = Annonce.objects.filter(departement = data['departement']).aggregate(moyenne = Avg('prix'))['moyenne']\n liste_prix = Annonce.objects.filter(departement = data['departement']).order_by('prix').values_list('prix', flat=True)\n if len(liste_prix) != 0:\n quartile10 = liste_prix[len(liste_prix) // 10]\n quartile90 = liste_prix[len(liste_prix) * 9 // 10]\n else: err = True\n if result_moy: result_moy = int(result_moy)\n if form_l.is_valid():\n # Find the id, ask API and return useful data\n lien = str(form_l.cleaned_data['lien'])\n \n id_bien = lien[lien.find(\"/ag\") + 3: lien.find(\"?q\")]\n\n url = \"https://www.bienici.com/realEstateAd.json?id=ag\" + str(id_bien)\n response = requests.get(url)\n if response.ok:\n json_data = json.loads(response.content)\n code_postal, ville, prix = json_data['postalCode'], json_data['city'], json_data['priceWithoutFees']\n try:\n annonce = Annonce(prix = prix, departement = code_postal[:-3], ville = ville, code_ville = code_postal)\n annonce.save()\n message = \"L'annonce \" + id_bien + \" a bien été enregistrée dans la base de données\"\n except:\n message = \"Erreur lors de l'ajout de l'annonce\"\n pass\n\n return render(request, \"app_data/home.html\", context={\"ann_form\": ann_form, \"add_form\": add_form, \"result_moy\": result_moy, \"quartile10\": quartile10, \"quartile90\": quartile90, \"err\": err, \"message\": message})\n \n","repo_name":"maelc49/test_technique","sub_path":"1_Dev/exercice_dev/app_data/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":3341,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"19279729422","text":"\"\"\"\n This module computes the outranking flows if basic(PROMETHEE I) or\n profile-based style.\n\n Implementation and naming of conventions are taken from\n :cite:p:'BransMareschal2005'\n\"\"\"\nimport pandas as pd\nfrom core.enums import FlowType\nfrom typing import Tuple, Union\n\n__all__ = [\"calculate_promethee_outranking_flows\"]\n\nfrom core.input_validation import basic_outranking_flows_validation, \\\n profile_based_outranking_flows_validation, \\\n check_outranking_flows_type\n\n\ndef _calculate_flow(preferences: Union[Tuple[pd.DataFrame, pd.DataFrame],\n pd.DataFrame],\n positive: bool = True) -> pd.Series:\n \"\"\"\n Calculate positive or negative outranking flow in basic(PROMETHEE I)\n style.\n\n :param preferences: pd.DataFrame with alternatives as index and\n alternatives as columns or Tuple of pd.DataFrame with alternatives as\n index and profiles as columns and pd.DataFrame with profiles as index\n and alternatives as columns.\n :param positive: bool, if True function returns positive outranking flow\n else returns negative outranking flow.\n :return: pd.Series with alternatives as index and positive or\n negative flows are values.\n \"\"\"\n if isinstance(preferences, tuple):\n if positive:\n flows = preferences[0].mean(axis=1)\n else:\n flows = preferences[1].mean(axis=0)\n return flows\n else:\n # Current alternative is not took into account\n # (btw. its inner preference is 0)\n axis = 1 if positive else 0\n aggregated_preferences = \\\n preferences.sum(axis=axis) / (preferences.shape[0] - 1)\n\n return aggregated_preferences\n\n\ndef _calculate_profile_based_flow(preferences: Tuple[pd.DataFrame,\n pd.DataFrame],\n profiles_preferences: pd.DataFrame,\n positive: bool = True) -> pd.Series:\n \"\"\"\n Calculate positive or negative outranking profile-based flows.\n\n :param preferences: Tuple of pd.DataFrame with alternatives as index\n and profiles as columns and pd.DataFrame with profiles as index\n and alternatives as columns.\n :param profiles_preferences: pd.DataFrame with profiles as index and\n profiles as columns.\n bool, if True function returns positive outranking flow\n else returns negative outranking flow.\n :return: pd.Series with\n MultiIndex(\"R\" + alternatives, profiles + alternative) as index and\n positive or negative flows are values.\n \"\"\"\n n_profiles = len(profiles_preferences)\n alternatives_groups_flows = []\n alternatives_groups_names = []\n axis = 1 if positive else 0\n\n # Iterate over alternatives\n for alternative, alternative_preferences in preferences[0].iterrows():\n # Create subset of profiles + current alternative preferences\n\n # Copy profiles preferences\n alternative_group_preferences = profiles_preferences.copy()\n # Add current alternative preferences to the end of\n # the subset (row and column)\n alternative_group_preferences.loc[alternative] = \\\n alternative_preferences\n alternative_group_preferences[alternative] = \\\n preferences[1][alternative]\n\n # Calculate flows for current group\n alternatives_groups_flows.append(\n alternative_group_preferences.sum(axis=axis) / n_profiles)\n alternatives_groups_names.append(f\"R{alternative}\")\n\n # Combine all groups flows\n return pd.concat(\n alternatives_groups_flows, keys=alternatives_groups_names)\n\n\ndef calculate_promethee_outranking_flows(\n preferences: Union[Tuple[pd.DataFrame, pd.DataFrame], pd.DataFrame],\n flow_type: FlowType,\n profiles_preferences: pd.DataFrame = None) -> pd.DataFrame:\n \"\"\"\n Calculate outranking flows in basic(PROMETHEE I) or profile-based style.\n Basic(PROMETHEE I) flows are calculated as mean of subtractions of\n preferences where current alternative is preferred to\n profiles/alternatives and preferences where profiles/alternatives\n is preferred to current alternative.\n Profile-based flows are calculated by creating\n subsets: profiles + current alternative and calculating flows in that\n set as in basic style. Because of modularity of this project\n preferences for that flows are obtained in different way (needs\n alternatives vs profiles and profiles vs profiles preferences).\n\n :param preferences: pd.DataFrame with alternatives as index and\n alternatives as columns or Tuple of pd.DataFrame with alternatives as\n index and alternatives as columns and pd.DataFrame with profiles as index\n and alternatives as columns.\n :param flow_type: FlowType enum with type of outranking\n flows (BASIC OR PROFILE_BASED).\n :param profiles_preferences: pd.DataFrame with profiles as index and\n profiles as columns.\n :return: pd.DataFrame with alternatives as index and 'positive' and\n 'negative' columns if flow_type is BASIC or pd.DataFrame with\n MultiIndex(\"R\" + alternatives, profiles+alternative) as index and\n 'positive' and 'negative' columns if flow_type is PROFILE_BASED.\n \"\"\"\n\n # flow_type validation\n check_outranking_flows_type(flow_type)\n\n if flow_type == FlowType.BASIC:\n # Input validation for basic(PROMETHEE I) style\n basic_outranking_flows_validation(preferences)\n\n # Get alternatives as index\n index = preferences[0].index if isinstance(preferences, tuple) \\\n else preferences.index\n return pd.DataFrame({'positive': _calculate_flow(preferences),\n 'negative':\n _calculate_flow(preferences, positive=False)\n }, index=index)\n\n elif flow_type == FlowType.PROFILE_BASED:\n # Input validation for profile-based style\n profile_based_outranking_flows_validation(preferences,\n profiles_preferences)\n return pd.DataFrame({'positive': _calculate_profile_based_flow(\n preferences, profiles_preferences),\n 'negative': _calculate_profile_based_flow(\n preferences, profiles_preferences,\n positive=False)})\n","repo_name":"WAndraszyk/Construct-your-own-PROMETHEE-with-Python-for-MCDA","sub_path":"modular_parts/flows/M8_PrometheeOutrankingFlows.py","file_name":"M8_PrometheeOutrankingFlows.py","file_ext":"py","file_size_in_byte":6350,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"23576089815","text":"import cv2\nimport numpy as np\n\nimg1 = cv2.imread('robot.jpg')\nimg2 = cv2.imread('opencv-logo2.png')\n#cv2.imshow('img1',img1)\n#cv2.imshow('img2',img2)\n\nrows, cols, channels = img2.shape\n#print(img2.shape)\nroi = img1[0:rows, 0:cols]\n#cv2.imshow('roi',roi)\nimg2gray = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY)\n#cv2.imshow('img2gray',img2gray)\nret, mask = cv2.threshold(img2gray, 10,255, cv2.THRESH_BINARY)\n#cv2.imshow('mask',mask)\nmask_inv = cv2.bitwise_not(mask)\n#cv2.imshow('mask_inv',mask_inv)\nimg1_bg = cv2.bitwise_and(roi, roi, mask= mask_inv)\nimg1_fg = cv2.bitwise_and(img2,img2, mask=mask)\n#cv2.imshow('img1_bg',img1_bg)\n#cv2.imshow('img1_fg',img1_fg)\ndst = cv2.add(img1_bg,img1_fg)\n#cv2.imshow('dst',dst)\nimg1[0:rows, 0:cols] = dst\ncv2.imshow('img1',img1)\ncv2.waitKey()\n \n","repo_name":"siddhesh13/raspberryPi","sub_path":"OpenCV/addingTwoImages.py","file_name":"addingTwoImages.py","file_ext":"py","file_size_in_byte":774,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"33232054788","text":"import pickle\nimport queue\nfrom functools import total_ordering\nfrom typing import Set\n\n\n@total_ordering\nclass RpniNode:\n __slots__ = ['output', 'children', 'prefix', \"type\"]\n\n def __init__(self, output=None, children=None, automaton_type='moore'):\n if output is None and automaton_type == 'mealy':\n output = dict()\n if children is None:\n children = dict()\n self.output = output\n self.children = children\n self.prefix = ()\n self.type = automaton_type\n\n def shallow_copy(self):\n output = self.output if self.type != 'mealy' else dict(self.output)\n return RpniNode(output, dict(self.children), self.type)\n\n def copy(self):\n return pickle.loads(pickle.dumps(self, -1))\n\n def __lt__(self, other):\n return (len(self.prefix), self.prefix) < (len(other.prefix), other.prefix)\n\n def __eq__(self, other):\n return self.prefix == other.prefix\n\n def __hash__(self):\n return id(self) # TODO This is a hack\n\n def get_all_nodes(self) -> Set['RpniNode']:\n qu = queue.Queue()\n qu.put(self)\n nodes = set()\n while not qu.empty():\n state = qu.get()\n nodes.add(state)\n for child in state.children.values():\n if child not in nodes:\n qu.put(child)\n return nodes\n\n def to_automaton(self):\n nodes = self.get_all_nodes()\n nodes.remove(self) # dunno whether order is preserved?\n nodes = [self] + list(nodes)\n return to_automaton(nodes, self.type)\n\n def compatible_outputs(self, other):\n so, oo = [self.output, other.output]\n cmp = lambda x, y: x is None or y is None or x == y\n if self.type == 'moore':\n return cmp(so, oo)\n else:\n return all(cmp(so[key], oo[key]) for key in filter(lambda k: k in oo, so))\n\n def get_child_by_prefix(self, prefix):\n node = self\n for symbol in prefix:\n node = node.children[symbol]\n return node\n\n\nclass StateMerging:\n def __init__(self, data, automaton_type, print_info=True):\n self.data = data\n self.automaton_type = automaton_type\n self.print_info = print_info\n\n self.root = createPTA(data, automaton_type)\n self.merges = []\n\n def merge(self, red_node, lex_min_blue, copy_nodes=False):\n \"\"\"\n Merge two states and return the root node of resulting model.\n \"\"\"\n\n if self.automaton_type == 'mealy':\n raise NotImplementedError()\n\n if not copy_nodes:\n self.merges.append((red_node, lex_min_blue))\n\n root_node = self.root.copy() if copy_nodes else self.root\n lex_min_blue = lex_min_blue.copy() if copy_nodes else lex_min_blue\n\n red_node_in_tree = root_node\n for p in red_node.prefix:\n red_node_in_tree = red_node_in_tree.children[p]\n\n to_update = root_node\n for p in lex_min_blue.prefix[:-1]:\n to_update = to_update.children[p]\n\n to_update.children[lex_min_blue.prefix[-1]] = red_node_in_tree\n\n if not self._fold(red_node_in_tree, lex_min_blue, not copy_nodes):\n return None\n\n return root_node\n\n def _fold(self, red_node, blue_node, report):\n # Change the output of red only to concrete output, ignore None\n if report and not RpniNode.compatible_outputs(red_node, blue_node):\n print(f\"conflict {red_node.prefix} ({red_node.output}) {blue_node.prefix} ({blue_node.output})\")\n return False\n red_node.output = blue_node.output if blue_node.output is not None else red_node.output\n\n for i in blue_node.children.keys():\n if i in red_node.children.keys():\n self._fold(red_node.children[i], blue_node.children[i], report)\n else:\n red_node.children[i] = blue_node.children[i]\n return True\n\n def to_automaton(self):\n return self.root.to_automaton()\n\n def replay_log(self, commands: list):\n for command, args in commands:\n if command == \"merge\":\n self.merge(self.root.get_child_by_prefix(args[0]), self.root.get_child_by_prefix(args[1]))\n elif command == \"promote\":\n pass\n\n @staticmethod\n def replay_log_on_pta(data, commands: list, automaton_type):\n sm = StateMerging(data, automaton_type)\n sm.replay_log(commands)\n return sm.to_automaton()\n\n\ndef check_sequence(root_node, seq, automaton_type):\n \"\"\"\n Checks whether each sequence in the dataset is valid in the current automaton.\n \"\"\"\n curr_node = root_node\n for i, o in seq:\n if automaton_type == 'mealy':\n input_outputs = {i: o for i, o in curr_node.children.keys()}\n if i[0] not in input_outputs.keys() or o is not None and input_outputs[i[0]] != o:\n return False\n curr_node = curr_node.children[(i[0], input_outputs[i[0]])]\n else:\n # For dfa and moore, check if outputs are the same, iff output in test data is concrete (not None)\n curr_node = curr_node.children[i]\n if o is not None and curr_node.output != o:\n return False\n return True\n\n\ndef createPTA(data, automaton_type):\n data.sort(key=lambda x: len(x[0]))\n\n root_node = RpniNode(automaton_type=automaton_type)\n for seq, label in data:\n curr_node = root_node\n for idx, symbol in enumerate(seq):\n if symbol not in curr_node.children.keys():\n node = RpniNode(automaton_type=automaton_type)\n node.prefix = curr_node.prefix + (symbol,)\n curr_node.children[symbol] = node\n\n if automaton_type == 'mealy' and idx == len(seq) - 1:\n if symbol not in curr_node.output:\n curr_node.output[symbol] = label\n if curr_node.output[symbol] != label:\n return None\n curr_node = curr_node.children[symbol]\n if automaton_type == 'moore' or automaton_type == 'dfa':\n if curr_node.output is None:\n curr_node.output = label\n if curr_node.output != label:\n return None\n\n return root_node\n\n\ndef extract_unique_sequences(root_node):\n def get_leaf_nodes(root):\n leaves = []\n\n def _get_leaf_nodes(node):\n if node is not None:\n if len(node.children.keys()) == 0:\n leaves.append(node)\n for n in node.children.values():\n _get_leaf_nodes(n)\n\n _get_leaf_nodes(root)\n return leaves\n\n leaf_nodes = get_leaf_nodes(root_node)\n paths = []\n for node in leaf_nodes:\n seq = []\n curr_node = root_node\n for i in node.prefix:\n curr_node = curr_node.children[i]\n seq.append((i, curr_node.output))\n paths.append(seq)\n\n return paths\n\n\ndef to_automaton(red, automaton_type):\n from aalpy.automata import DfaState, Dfa, MooreMachine, MooreState, MealyMachine, MealyState\n\n if automaton_type == 'dfa':\n state, automaton = DfaState, Dfa\n elif automaton_type == 'moore':\n state, automaton = MooreState, MooreMachine\n else:\n state, automaton = MealyState, MealyMachine\n\n initial_state = None\n prefix_state_map = {}\n for i, r in enumerate(red):\n if automaton_type == 'moore' or automaton_type == 'dfa':\n prefix_state_map[r.prefix] = state(f's{i}', r.output)\n else:\n prefix_state_map[r.prefix] = state(f's{i}')\n if i == 0:\n initial_state = prefix_state_map[r.prefix]\n\n for r in red:\n for i, c in r.children.items():\n if automaton_type == 'moore' or automaton_type == 'dfa':\n prefix_state_map[r.prefix].transitions[i] = prefix_state_map[c.prefix]\n else:\n prefix_state_map[r.prefix].transitions[i] = prefix_state_map[c.prefix]\n prefix_state_map[r.prefix].output_fun[i] = r.output[i] if i in r.output else None\n\n return automaton(initial_state, list(prefix_state_map.values()))\n\n\ndef visualize_pta(root_node, path='pta.pdf'):\n from pydot import Dot, Node, Edge\n graph = Dot('fpta', graph_type='digraph')\n\n graph.add_node(Node(str(root_node.prefix), label=f'{root_node.output}'))\n\n queue = [root_node]\n visited = set()\n visited.add(root_node.prefix)\n while queue:\n curr = queue.pop(0)\n for i, c in curr.children.items():\n if c.prefix not in visited:\n graph.add_node(Node(str(c.prefix), label=f'{c.output}'))\n graph.add_edge(Edge(str(curr.prefix), str(c.prefix), label=f'{i}'))\n if c.prefix not in visited:\n queue.append(c)\n visited.add(c.prefix)\n\n graph.add_node(Node('__start0', shape='none', label=''))\n graph.add_edge(Edge('__start0', str(root_node.prefix), label=''))\n\n graph.write(path=path, format='pdf')\n","repo_name":"DES-Lab/AALpy","sub_path":"aalpy/learning_algs/deterministic_passive/rpni_helper_functions.py","file_name":"rpni_helper_functions.py","file_ext":"py","file_size_in_byte":9033,"program_lang":"python","lang":"en","doc_type":"code","stars":121,"dataset":"github-code","pt":"7"} +{"seq_id":"33390920711","text":"import csv\nimport ast\n\nfreq_most_common = []\nfreq_common = []\n# punctuation_pos = [\"SENT\", \"#\", \"$\", \"\\\"\", \"''\", \"'\", \"(\", \")\", \",\", \":\"]\n\n# ------------------------------------- GDEX POINTS -----------------------------------------------------\n\n\n# Sentence length: a sentence between 10 and 20 words long was preferred, with and shorter ones penalized.\n# No sentence is longer than 20 words.\n\ndef sentence_length(m_lemma_tag):\n # points = 20\n sent_length = len(m_lemma_tag)\n\n # sentence is greater than 10\n if sent_length >= 10:\n points = 20\n # sentence is smaller than 10\n else:\n points = 20 * (sent_length/10)\n\n return points\n\n\n# Word frequencies: a sentence was penalized for each word that was not amongst the commonest 17,000 words in the\n# language, with a further penalty applied for rare words.\n# max Points = 3\ndef common_words(m_lemma_tag):\n points = 0\n x = 100 / len(m_lemma_tag)\n # print(len(m_sentence[1]))\n\n # one free 'Punct' for end of sentence every other is negative\n punctuations = 0\n for m_word in m_lemma_tag:\n if m_word[1] == 'Punct':\n if punctuations == 0:\n points += x\n punctuations += 1\n else:\n # if you ignore proper nouns the resulting 'good' sentences are really bad\n if m_word[0] in freq_most_common:\n points += x\n elif m_word[0] in freq_common:\n points += x/2\n\n # for m_word in m_sentence[1]:\n # if m_word[0] in freq_most_common:\n # points += x\n # elif m_word[0] in freq_common:\n # points += x / 2\n\n points = 50 * (points/100)\n return points\n\n\n# Sentences containing pronouns and anaphors like this that it or one often fail to present a self-contained piece of\n# language which makes sense without further context, so sentences containing these words were penalized.\n# max Points = 1\ndef contain_pronouns_anaphora(m_lemma_tag):\n x = len(m_lemma_tag)\n found = x\n for m_word in m_lemma_tag:\n if m_word[1] == \"Pron\":\n found -= 1\n\n points = 20 * (found/x)\n return points\n\n\n# Whole sentence—identified as beginning with a capital letter and ending with a full step, exclamation mark, or\n# question mark, were preferred.\n# max Points = 2\ndef whole_sentence(m_sentence):\n points = 10\n punctuation = [\"!\", \".\", \"?\"]\n #quote = ['\\'', '\"']\n first_letter = m_sentence[:1]\n last_letter = m_sentence[-1]\n\n if first_letter.isupper(): # or first_letter in quote:\n if last_letter in punctuation:\n pass # is upper and has punctuation\n else:\n points -= 5 # is upper and has no punctuation\n else:\n if last_letter in punctuation:\n points -= 5 # is not upper and has punctuation\n else:\n points -= 10 # is not upper and has no punctuation\n\n return points\n\n\n# ------------------------------------- LOAD FREQUENCIES --------------------------------------------------\n\ndef load_frequencies():\n with open('lemma.num.17000.txt', 'r') as freq_in:\n f = csv.reader(freq_in, delimiter=' ')\n row_number = 1\n for row in f:\n if row_number <= 7000:\n freq_most_common.append(row[2])\n else:\n freq_common.append(row[2])\n row_number += 1\n freq_in.close()\n\n\ndef compute_points(m_sentence, m_lemma_tag):\n\n points_len = sentence_length(m_lemma_tag) # max 20p\n # print(\"len\", points_len, \"/\", 20)\n\n points_common = common_words(m_lemma_tag) # max 50p\n # print(\"com\", points_common, \"/\", 50)\n\n points_pronoun = contain_pronouns_anaphora(m_lemma_tag) # max 20p\n # print(\"pro\", points_pronoun, \"/\", 20)\n\n points_whole_sent = whole_sentence(m_sentence) # max 10p\n # print(\"sent\", points_whole_sent, \"/\", 10)\n\n m_gdex_points = points_len + points_common + points_pronoun + points_whole_sent\n return round(m_gdex_points, 1)\n\n\ndef main():\n if __name__ == \"__main__\":\n print(\"#### starting loading sentence file ####\")\n\n with open(\"./output/03_1_sentences_from_corpus.txt\", 'r') as read_sent, \\\n open(\"./output/03_2_calculate_GDEX.csv\", \"w\") as s_out:\n index = 1\n writer = csv.writer(s_out, delimiter=';')\n good_sentence = 0\n bad_sentence = 0\n for line in read_sent:\n if index % 100000 == 0:\n print(\"{:,}\".format(index), \"/\", \"{:,}\".format(15500000))\n index += 1\n\n row_as_list = ast.literal_eval(line)\n\n # 0 | 1 | 2 | 3 | 4 | 5\n # vocable | chapter | book | sentence | lemmatag | lemmavocdict\n\n sentence = row_as_list[3]\n lemma_tag = row_as_list[4]\n\n gdex_number = compute_points(sentence, lemma_tag)\n row_as_list.insert(3, gdex_number)\n\n if gdex_number > 60:\n good_sentence += 1\n writer.writerow([row_as_list[0]] + [row_as_list[1]] + [row_as_list[2]] + [row_as_list[3]]\n + [row_as_list[4]] + [row_as_list[5]] + [row_as_list[6]])\n else:\n bad_sentence += 1\n\n s_out.close()\n read_sent.close()\n print(\"#### finished loading sentence file ####\")\n print(\"good:\", good_sentence)\n print(\"bad:\", bad_sentence)\n\n\nif __name__ == \"__main__\":\n print(\"#### starting loading sentence file ####\")\n average = 0\n index = 1\n highest = 0\n lowest = 100\n with open(\"./output/03_1_sentences_from_corpus.txt\", 'r') as read_sent, \\\n open(\"./output/03_2_calculate_GDEX.csv\", \"w\") as s_out:\n\n writer = csv.writer(s_out, delimiter=';')\n good_sentence = 0\n bad_sentence = 0\n for line in read_sent:\n if index % 100000 == 0:\n print(\"{:,}\".format(index), \"/\", \"{:,}\".format(15500000))\n index += 1\n\n row_as_list = ast.literal_eval(line)\n\n # 0 | 1 | 2 | 3 | 4 | 5\n # vocable | chapter | book | sentence | lemmatag | lemmavocdict\n\n sentence = row_as_list[3]\n lemma_tag = row_as_list[4]\n\n gdex_number = compute_points(sentence, lemma_tag)\n row_as_list.insert(3, gdex_number)\n\n average += gdex_number\n if gdex_number > highest:\n highest = gdex_number\n if gdex_number < lowest:\n lowest = gdex_number\n\n if gdex_number > 50:\n #print(sentence)\n #if \"I can smell\" in sentence:\n # print(\"gdex:\", gdex_number, sentence)\n good_sentence += 1\n writer.writerow([row_as_list[0]] + [row_as_list[1]] + [row_as_list[2]] + [row_as_list[3]]\n + [row_as_list[4]] + [row_as_list[5]] + [row_as_list[6]])\n else:\n bad_sentence += 1\n\n s_out.close()\n read_sent.close()\n print(\"#### finished loading sentence file ####\")\n print(\"good:\\t\", good_sentence)\n print(\"bad:\\t\", bad_sentence)\n print(\"avg:\\t\", gdex_number / index)\n print(\"high:\\t\", highest)\n print(\"low:\\t\", lowest)\n","repo_name":"Mells/Preprocess03","sub_path":"O3_extract_sentences_from_corpus/O3_2_calculate_GDEX.py","file_name":"O3_2_calculate_GDEX.py","file_ext":"py","file_size_in_byte":7411,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"36763785377","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\n#from ARAS datasets Selected House A\n#Anomaly state is checked for resident 1 in House A\n#We divided the datasets into two equal parts\n#We used all the second 15 days of the datasets in to check for anomalies\n#You can change this division\n\n# DATASETS :\n# Change to CSV format \n# Add a row (first row) for the sensor name and resident number (In short)\n# Add a column to show the seconds.\n\n\n#The first 15 days (two weeks) are used for the initial learning of the algorithm.\n#from the sixteenth day, NNC begins to be calculated.\n\n# Pay Attention:\n# To use this code, the reading part of the datasets must be changed\n\n###............................................. Main Part ............................................\n\nimport random\nimport statistics\nimport numpy as np\nimport pandas as pd\nfrom sklearn import tree\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.metrics import classification_report\nfrom sklearn.metrics import precision_recall_fscore_support as score\n\n#........................................... Functions ..................................... :\n\ndef randomcheck(testday ,resident, classcode):\n subNNC = 0\n y = testday['r'+str(resident)]\n if classcode == 10 or classcode == 20:\n testday.drop(['r1'], axis=1 , inplace = True)\n testday.drop(['r2'], axis=1 , inplace = True)\n else:\n testday.drop(['r'+str(resident)], axis=1 , inplace = True)\n T = 0\n for i in range (3600):\n rt = random.randint(1,86400)\n yt = y[rt-1:rt]\n xt = testday[rt-1:rt]\n yp = globals()[\"dt\"+str(classcode)].predict(xt)\n if int(yt) == yp:\n T = T + 1\n if T < 900:\n subNNC = 20\n if T > 900 and T < 1400:\n subNNC = 14 \n if T > 1400 and T < 1900:\n subNNC = 12\n if T > 1900 and T < 2400:\n subNNC = 5\n if T > 2400 and T < 2900:\n subNNC = 3\n if T > 2900 and T < 3400:\n subNNC = 1\n return subNNC\n\n#.............////////////////////..............\n\ndef TimePart(day,resident):\n for g in range(27):\n globals()[\"p\"+str(g+1)] = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n for n in range(24):\n for t in range(27):\n globals()[\"h\"+str(t+1)] = 0\n for s in range((n)*3600,((n+1)*3600)-1):\n o = day['r'+str(resident)][s]\n globals()[\"h\"+str(o)] = 1 + globals()[\"h\"+str(o)]\n if globals()[\"h\"+str(o)] > 60:\n globals()[\"p\"+str(o)][n] = 1\n out = [p1, p2, p3, p4, p5, p6, p7, p8, p9, p10, p11, p12, p13 , p14,\n p15, p16, p17, p18, p19, p20, p21, p22, p23, p24, p25, p26, p27]\n return out\n\n#.............////////////////////..............\n\ndef predNorm (x, classcode):\n \n for b in range(27):\n globals()[\"p\"+str(b+1)] = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n for n in range(24):\n for b in range(27):\n globals()[\"y_pred\"+str(b+1)] = 0\n for s in range((n)*3600,((n+1)*3600)-1):\n pred = int(globals()[\"dt\"+str(classcode)].predict(x[s:s+1]))\n globals()[\"y_pred\"+str(pred)] = 1 + globals()[\"y_pred\"+str(pred)] \n if globals()[\"y_pred\"+str(pred)] > 60:\n globals()[\"p\"+str(pred)][n] = 1\n ex = [p1, p2, p3, p4, p5, p6, p7, p8, p9, p10, p11, p12, p13 , p14,\n p15, p16, p17, p18, p19, p20, p21, p22, p23, p24, p25, p26, p27] \n return ex\n\ndef checkpred (tday, resident, classcode):\n subNNC = 0\n real = TimePart(tday,resident)\n if classcode == 10 or classcode == 20:\n tday.drop(['r1'], axis=1 , inplace = True)\n tday.drop(['r2'], axis=1 , inplace = True)\n else:\n tday.drop(['r'+str(resident)], axis=1 , inplace = True)\n x_test = tday\n pred = predNorm(x_test, classcode)\n for d in range(27):\n if real[d] != pred[d]:\n subNNC = 1 + subNNC\n return subNNC\n\n#.............////////////////////..............\n\ndef norm (tday,resident):\n subNNC = 0\n for g in range(27):\n globals()[\"k\"+str(g+1)] = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n\n#In this section, the datasets of the days should be entered in order\n for i in range(15):\n d = pd.read_csv('c://DAY_0'+str(i+1)+'.csv' , encoding = 'ansi')\n y = TimePart(d,resident) \n for f in range(27):\n for p in range(24):\n globals()[\"k\"+str(f+1)][p] = globals()[\"k\"+str(f+1)][p] + y[f][p]\n\n j = TimePart(tday,resident)\n for f in range(27):\n for k in range(24):\n timechart = globals()[\"k\"+str(f+1)][k]\n tcharttest = j[f][k]\n if timechart < 3 and tcharttest == 1:\n subNNC = subNNC + 1\n if timechart > 12 and tcharttest == 0:\n subNNC = subNNC + 1\n return subNNC\n\n#.............////////////////////..............\n\ndef sectable (dy, resident):\n\n for g in range(27):\n globals()[\"sec\"+str(g+1)] = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0] \n for n in range(24):\n for x in range((n)*3600,((n+1)*3600)-1):\n o = dy['r'+str(resident)][x]\n globals()[\"sec\"+str(o)][n] = 1 + globals()[\"sec\"+str(o)][n]\n secout = [sec1,sec2,sec3,sec4,sec5,sec6,sec7,sec8,sec9,sec10,sec11,sec12,sec13,sec14,\n sec15,sec16,sec17,sec18,sec19,sec20,sec21,sec22,sec23,sec24,sec25,sec26,sec27]\n return secout\n\ndef MaxMin (resident):\n for g in range(27):\n globals()[\"ma\"+str(g+1)] = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n globals()[\"mi\"+str(g+1)] = [4000,4000,4000,4000,4000,4000,4000,4000,4000,4000,4000,4000,\n 4000,4000,4000,4000,4000,4000,4000,4000,4000,4000,4000,4000]\n \n#In this section, the datasets of the days should be entered in order\n for i in range(15):\n day = pd.read_csv('c://DAY_0'+str(i+1)+'.csv' , encoding = 'ansi')\n ts = sectable(day,resident)\n for w in range(27):\n for q in range(24):\n if ts[w][q] > globals()[\"ma\"+str(w+1)][q]:\n globals()[\"ma\"+str(w+1)][q] = ts[w][q]\n if ts[w][q] < globals()[\"mi\"+str(w+1)][q]:\n globals()[\"mi\"+str(w+1)][q] = ts[w][q] \n mintabl = [mi1,mi2,mi3,mi4,mi5,mi6,mi7,mi8,mi9,mi10,mi11,mi12,mi13,mi14,\n mi15,mi16,mi17,mi18,mi19,mi20,mi21,mi22,mi23,mi24,mi25,mi26,mi27]\n maxtabl = [ma1,ma2,ma3,ma4,ma5,ma6,ma7,ma8,ma9,ma10,ma11,ma12,ma13,ma14,\n ma15,ma16,ma17,ma18,ma19,ma20,ma21,ma22,ma23,ma24,ma25,ma26,ma27] \n return (mintabl , maxtabl)\n\ndef MaxMincheck(testday, resident):\n subNNC = 0\n st = sectable(testday, resident)\n mi , ma = MaxMin(resident)\n for p in range(27):\n for w in range(24):\n if st[p][w] < mi[p][w]:\n subNNC = 4 + subNNC\n if st[p][w] > ma[p][w]:\n subNNC = 2 + subNNC\n return subNNC\n\n#.............////////////////////..............\n\ndef TimeJob (daay, resident):\n \n for v in range(27):\n globals()[\"ac\"+str(v+1)] = [] \n for l in range(27):\n globals()[\"st\"+str(l+1)] = -1 \n t = 0 \n while t < 86400:\n z = t\n o = daay['r'+str(resident)][t]\n globals()[\"st\"+str(o)] = 1 + globals()[\"st\"+str(o)]\n globals()[\"ac\"+str(o)].append([z,1])\n globals()[\"ac\"+str(o)][globals()[\"st\"+str(o)]][1] = 1 + globals()[\"ac\"+str(o)][globals()[\"st\"+str(o)]][1]\n for a in range(z+1,86400):\n if daay['r'+str(resident)][a] == o:\n globals()[\"ac\"+str(o)][globals()[\"st\"+str(o)]][1] = 1 + globals()[\"ac\"+str(o)][globals()[\"st\"+str(o)]][1]\n if a == 86399:\n t = 86400\n else:\n t = a\n break\n t = t + 1\n\n acout = [len(ac1),len(ac2),len(ac3),len(ac4),len(ac5),len(ac6),len(ac7),len(ac8),len(ac9),len(ac10),\n len(ac11),len(ac12),len(ac13),len(ac14),len(ac15),len(ac16),len(ac17),len(ac18),len(ac19),\n len(ac20),len(ac21),len(ac22),len(ac23),len(ac24),len(ac25),len(ac26),len(ac27)]\n \n for m in range(27):\n globals()[\"tts\"+str(m+1)] = 0\n \n for g in range(27):\n h = globals()[\"ac\"+str(g+1)]\n for s in range(len(h)):\n globals()[\"tts\"+str(g+1)] = globals()[\"ac\"+str(g+1)][s][1] + globals()[\"tts\"+str(g+1)]\n \n ttsout = [tts1, tts2, tts3, tts4, tts5, tts6, tts7, tts8, tts9, tts10, tts11, tts12, tts13, tts14,\n tts15, tts16, tts17, tts18, tts19, tts20, tts21, tts22, tts23, tts24, tts25, tts26, tts27]\n \n return acout , ttsout\n \ndef numericalCheck (day, resident):\n subNNC = 0\n for v in range(27):\n globals()[\"c\"+str(v+1)] = 0 \n for f in range(27):\n globals()[\"s\"+str(f+1)] = 0\n for k in range(27):\n globals()[\"av\"+str(k+1)] = 0\n for d in range(27):\n globals()[\"mc\"+str(d+1)] = 0\n for s in range(27):\n globals()[\"mt\"+str(s+1)] = 0\n for s in range(27):\n globals()[\"mi\"+str(s+1)] = 90000\n#In this section, the datasets of the days should be entered in order\n for i in range(15):\n daaay = pd.read_csv('c://DAY_0'+str(i+1)+'.csv' , encoding = 'ansi')\n count , total = TimeJob(daaay,resident)\n for j in range(27):\n if count[j] > globals()[\"mc\"+str(j+1)]:\n globals()[\"mc\"+str(j+1)] = count[j]\n if total[j] > globals()[\"mt\"+str(j+1)]:\n globals()[\"mt\"+str(j+1)] = total[j]\n if total[j] > 0:\n if total[j] < globals()[\"mi\"+str(j+1)]:\n globals()[\"mi\"+str(j+1)] = total[j]\n if count[j] > 0:\n globals()[\"c\"+str(j+1)] = globals()[\"c\"+str(j+1)] +1\n globals()[\"s\"+str(j+1)] = globals()[\"s\"+str(j+1)] +total[j]\n for g in range(27):\n if globals()[\"c\"+str(g+1)] > 0 :\n globals()[\"av\"+str(g+1)] = globals()[\"s\"+str(g+1)]/globals()[\"c\"+str(g+1)]\n \n cnt , ttl = TimeJob(day,resident)\n for r in range(27):\n if ttl[r] > 0 and ttl[r] < globals()[\"mi\"+str(r+1)]:\n subNNC = subNNC + 1\n if cnt[r] > globals()[\"mc\"+str(r+1)]:\n subNNC = subNNC + 1\n if ttl[r] > globals()[\"mt\"+str(r+1)]:\n subNNC = subNNC + 1\n if ttl[r] > ((globals()[\"av\"+str(r+1)])*1.5):\n subNNC = subNNC + 1\n if ttl[r] < ((globals()[\"av\"+str(r+1)])*0.5):\n subNNC = subNNC + 1\n if (globals()[\"c\"+str(r+1)]) > 12 and cnt[r] < 1:\n subNNC = subNNC + 1\n if (globals()[\"c\"+str(r+1)]) < 3 and cnt[r] > 0:\n subNNC = subNNC + 1\n return subNNC\n#.............////////////////////..............\n\n#In this section, the datasets of the days should be entered in order\n\nfor i in range(15):\n globals()[\"df0\" + str(i+1)] = pd.read_csv('c://DAY_0'+str(i+1)+'.csv' , encoding = 'ansi')\n\nw0 = df01.append(df02, ignore_index=True)\nfor j in range(12):\n globals()[\"w\"+str(j+1)] = globals()[\"w\"+str(j)].append(globals()[\"df0\" + str(j+3)] , ignore_index=True)\n if j==11:\n dataR10 = globals()[\"w\"+str(j+1)].append(globals()[\"df0\" + str(j+4)] , ignore_index=True)\n dataR11 = globals()[\"w\"+str(j+1)].append(globals()[\"df0\" + str(j+4)] , ignore_index=True)\n dataR20 = globals()[\"w\"+str(j+1)].append(globals()[\"df0\" + str(j+4)] , ignore_index=True)\n dataR21 = globals()[\"w\"+str(j+1)].append(globals()[\"df0\" + str(j+4)] , ignore_index=True)\n\nprint(\"\\nWELCOME\\n\")\n\n## ------------------------------------------------------ Resident 1 --------------------------------------------------\n\ndt11 = tree.DecisionTreeClassifier(criterion='gini') \ny11 = dataR11.r1\ndataR11.drop(['r1'], axis=1 , inplace = True)\nx11 = dataR11\ndt11.fit(x11,y11)\n\ndt10 = tree.DecisionTreeClassifier(criterion='gini') \ny10 = dataR10.r1\ndataR10.drop(['r1'], axis=1 , inplace = True)\ndataR10.drop(['r2'], axis=1 , inplace = True)\nx10 = dataR10\ndt10.fit(x10,y10)\n\nsubNNC11 = []\nsubNNC21 = []\nsubNNC10 = []\nsubNNC20 = []\nsubNNC3 = []\nsubNNC4 = []\nsubNNC5 = []\n\n#In this section, the 16th to the 30th days are called for review\n\nfor daycounter in range(15):\n \n#sub-NNC for Resident 1, Considering the behavior of the resident 2 in the last 15 days:\n dataSetsNNC11 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC11 = randomcheck(dataSetsNNC11,1,11)\n subNNC11.append(sbNNC11)\n dataSetsNNC21 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC21 = checkpred(dataSetsNNC21,1,11)\n subNNC21.append(sbNNC21)\n \n#sub-NNC for Resident 1, Regardless of resident behavior 2 in the last 15 days:\n dataSetsNNC10 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC10 = randomcheck(dataSetsNNC10,1,10)\n subNNC10.append(sbNNC10)\n dataSetsNNC20 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC20 = checkpred(dataSetsNNC20,1,10)\n subNNC20.append(sbNNC20)\n\n dataSetsNNC3 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC3 = numericalCheck(dataSetsNNC3,1)\n subNNC3.append(sbNNC3)\n dataSetsNNC4 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC4 = MaxMincheck(dataSetsNNC4,1)\n subNNC4.append(sbNNC4)\n dataSetsNNC5 = pd.read_csv('c://DAY_0'+str(daycounter+16)+'.csv' , encoding = 'ansi')\n sbNNC5 = norm(dataSetsNNC5,1)\n subNNC5.append(sbNNC5)\n\n#************************************************ FINAL NNC FOR RESIDENT 1 ************************************************ \n\nDailyNNC1 = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\nfor i in range(15):\n DailyNNC1[i] = subNNC11[i]+subNNC21[i]+subNNC10[i]+subNNC20[i]+subNNC3[i]+subNNC4[i]+subNNC5[i]\nNNC1 = np.mean(DailyNNC1)\nfor i in range(15):\n if DailyNNC1[i] > (NNC1*1.3):\n print(\"on\",i+16,\"day Abnormal state was diagnosed for the Resident 1\")\n\n\nprint(DailyNNC1)\nprint(\"Final Average NNC for Resident 1 in the last 15 days : \",NNC1)\n\n#*************************************************************************************************************************\n\n\n# In[ ]:\n\n\n\n\n","repo_name":"seyedkazemi/NNC","sub_path":"NNC-semicode-v1.py","file_name":"NNC-semicode-v1.py","file_ext":"py","file_size_in_byte":14275,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"34681160389","text":"\"\"\"\nThe functions that adapts a queryset in JSONs to use by frontend\n\ncontains:\n user_lister\n requested_user_lister\n event_lister\n category_lister\n subcategory_lister\n product_lister\n transaction_lister\n donateIn_lister\n needRequest_lister\n\"\"\"\n\n\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom App1.Components.item_functions import *\n\n\ndef user_lister(user_queryset):\n user_json = {}\n for user in user_queryset:\n user_json[user.id] = user_item(user)\n\n return user_json\n\n\ndef requested_user_lister(needy_queryset, donator_queryset, pagination_params=None):\n if pagination_params:\n return error(\"TODO\", {\"message\": \"Have no pagination yet; coming soon\"})\n\n needy_json = user_lister(needy_queryset)\n donator_json = user_lister(donator_queryset)\n\n empty_needy = [0 if len(needy_json) else 1]\n empty_donator = [0 if len(donator_json) else 1]\n\n final_json = {\"success\": \"1\",\n \"empty_needy\": empty_needy[0],\n \"empty_donator\": empty_donator[0],\n \"pagination_params\": pagination_params,\n \"needy_set\": needy_json,\n \"donator_set\": donator_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef event_lister(event_queryset, pagination_params=None):\n \"\"\"\n it passes pagination params to front if exists\n \"\"\"\n # Create a json for an event:\n event_json = {}\n for event in event_queryset:\n event_json[event.id] = event_item(event)\n\n empty = [0 if len(event_json) else 1]\n final_json = {\"success\": \"1\",\n \"empty\": empty[0],\n \"pagination_params\": pagination_params,\n \"event_set\": event_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef category_lister(category_queryset):\n category_json = {}\n for category in category_queryset:\n category_json[category.id] = category_item(category)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(category_json) else 1][0],\n \"count\": len(category_queryset),\n # \"pagination_params\": pagination_params,\n \"category_set\": category_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef subcategory_lister(subcategory_queryset):\n subcategory_json = {}\n for subcategory in subcategory_queryset:\n subcategory_json[subcategory.id] = subcategory_item(subcategory)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(subcategory_json) else 1][0],\n \"count\": len(subcategory_queryset),\n # \"pagination_params\": pagination_params,\n \"subcategory_set\": subcategory_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef product_lister(product_queryset):\n product_json = {}\n for product in product_queryset:\n product_json[product.id] = product_item(product)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(product_json) else 1][0],\n \"count\": len(product_queryset),\n # \"pagination_params\": pagination_params,\n \"product_set\": product_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef transaction_lister(transaction_queryset):\n transaction_json = {}\n for transaction in transaction_queryset:\n transaction_json[transaction.id] = transaction_item(transaction)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(transaction_json) else 1][0],\n \"count\": len(transaction_queryset),\n # \"pagination_params\": pagination_params,\n \"transaction_set\": transaction_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef donateIn_lister(donates_queryset):\n donate_json = {}\n for donate in donates_queryset:\n donate_json[donate.id] = donateIn_item(donate)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(donate_json) else 1][0],\n \"count\": len(donates_queryset),\n # \"pagination_params\": pagination_params,\n \"donate_set\": donate_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n\n\ndef needRequest_lister(needRequest_queryset):\n needRequest_json = {}\n for needRequest in needRequest_queryset:\n needRequest_json[needRequest.id] = needRequest_item(needRequest)\n\n final_json = {\"success\": \"1\",\n \"empty\": [0 if len(needRequest_json) else 1][0],\n \"count\": len(needRequest_queryset),\n # \"pagination_params\": pagination_params,\n \"needRequest_set\": needRequest_json}\n\n return Response(final_json,\n status=status.HTTP_200_OK)\n","repo_name":"ehsankarbasian/django_charity","sub_path":"Backend/App1/Components/lister_functions.py","file_name":"lister_functions.py","file_ext":"py","file_size_in_byte":4970,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"39181479223","text":"import matplotlib.pyplot as plt\nimport numpy as np\nimport random\nimport time\nfrom numpy import loadtxt\n \n# check error Pocket \ndef check_errorP(w, x ,label):\n if int(np.sign(w.T.dot(x))) != label:\n return True\n else :\n return False\n \n# get num of total error in dataset \ndef sum_error(w, dataset):\n errors = 0\n for x, label in dataset:\n if check_errorP(w, x, label):\n errors += 1\n return errors\n\n# Pocket演算法實作\ndef pocket(dataset):\n w = np.zeros(3) \n iterp = 0\n premisclassified = sum_error(w,dataset)\n bestw = 0\n \n for iterp in range(0,100000):\n misclassified = None\n while True:\n x, label = random.choice(dataset) # random choose samples\n if check_errorP(w,x,label):\n x = np.array(x)\n w = w + label * x\n misclassified = sum_error(w,dataset)\n break\n \n if misclassified < premisclassified: # misclassified smaller than premisclassified\n bestw = w\n premisclassified = misclassified\n \n iterp = iterp + 1 \n #w = bestw\n if premisclassified == 0:\n break\n print(\"iter: %d , misclassified: %d , best misclassified: %d\" % (iterp,misclassified , premisclassified))\n if premisclassified == 0 :\n print(\"!!halt!! iter: %d , misclassified: %d , best misclassified: %d\" % (iterp,misclassified , premisclassified))\n else :\n print(\"!!iter done!! iter: %d , misclassified: %d , best misclassified: %d\" % (iterp,misclassified , premisclassified))\n print(\"Accuracy :\" , ((samples-premisclassified)/samples)*100, \"%\")\n return bestw , iterp \n\nif __name__ == '__main__': \n m, b = 3, 10 # set the value of m and b\n print(\"the num of total samples : \")\n samples = int(input()) # the num of samples\n x = np.arange(samples)\n y = m * x + b\n plt.plot(x, y,'-y',label = 'Original')\n half = int(samples / 2)\n data = np.load('dataset.npz')\n dataset = list(zip(data['a'],data['b'],data['c'])) \n dataset = list(zip(dataset,data['z']))\n x0 = data['a']\n xdata = data['b']\n ydata = data['c']\n label = data['z'] \n print(\"Pocket:\") \n start1 = time.time() \n wbest , iterp = pocket(dataset)\n end1 = time.time()\n print(\"num of iterations: \",iterp)\n print(\"執行時間:%f 秒\"% float(end1-start1))\n l = np.arange(samples)\n ap,bp = -wbest[1]/wbest[2], -wbest[0]/wbest[2]\n print(\"m: %f, b: %f\" % (ap,bp))\n plt.plot(l, ap*l + bp, 'g-',label = 'Pocket')\n plt.plot(xdata[:half], ydata[:half], '.', color='blue',label='positive')\n plt.plot(xdata[half:], ydata[half:], 'x', color='red',label='negative')\n plt.legend(loc='upper left');\n plt.show()\n","repo_name":"nianyinwu/Perceptron","sub_path":"Pocket.py","file_name":"Pocket.py","file_ext":"py","file_size_in_byte":2778,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"12481332825","text":"import cv2\nimport skvideo.io\nimport pandas as pd\nimport numpy as np\nimport tqdm\n\n\nspecies_list = ['species_fourspot',\n 'species_grey sole',\n 'species_other',\n 'species_plaice',\n 'species_summer',\n 'species_windowpane',\n 'species_winter']\ndf = pd.read_csv('../../data/training.csv')\n\ndef process_row(row,frame,class_name, display=False, bounding_pad=2):\n x1, x2, y1, y2 = map(int, row[['x1', 'x2', 'y1', 'y2']].as_matrix())\n length = row['length']\n if length < 100:\n delta = 50\n elif length >= 100 and length < 150:\n delta = 75\n elif length >= 150 and length < 200:\n delta = 100\n elif length >= 200 and length < 250:\n delta = 125\n elif length >= 250 and length < 300:\n delta = 150\n elif length >= 300 and length < 350:\n delta = 175\n else:\n delta = int(length//2)\n \n delta += bounding_pad\n mid_x = int(x1 + x2)//2\n mid_y = int(y1 + y2)//2\n max_y, max_x, ch = frame.shape\n start_x = max(0, mid_x-delta)\n start_y = max(0, mid_y-delta)\n end_x = min(mid_x+delta, max_x)\n end_y = min(mid_y+delta, max_y)\n \n image = np.copy(frame)\n cropped_image = np.copy(image)[start_y:end_y,start_x:end_x]\n \n if display:\n pass\n else:\n return cropped_image\n\ndef process_video(df, video_index,skip=None,display_frames=True):\n if type(video_index)==str:\n x = df[df['video_id']==video_index].dropna()\n else:\n videos = df['video_id'].unique()\n x = df[df['video_id']==videos[video_index]].dropna()\n m = x['video_id'].iloc[0]\n base_path = '../../data/train_videos/'\n x['species'] = x.apply(lambda row: row[species_list].argmax(), axis=1)\n max_frame = x['frame'].max()\n vid_generator = skvideo.io.FFmpegReader(base_path + m + '.mp4')\n counter = -1\n outer_counter = 0\n images = []\n for i, row in x.iterrows():\n target_frame = row['frame']\n for f in vid_generator.nextFrame():\n frame = f\n counter +=1\n if counter == target_frame:\n break\n if skip:\n if outer_counter % skip == 0:\n ret = process_row(row,frame,row['species'], display=display_frames, bounding_pad=20)\n else:\n ret = process_row(row,frame,row['species'], display=display_frames, bounding_pad=20)\n if display_frames == False:\n images.append(ret)\n outer_counter += 1\n return images\n\n\ntotal_counter = 0\nall_videos = videos = df['video_id'].unique()\nbase_im_folder = '../../data/binary_classification/fish/'\n\nfor video in tqdm.tqdm(all_videos):\n try:\n u = process_video(df, video, None, False)\n for img in u:\n fname = base_im_folder + '{}.png'.format(total_counter)\n total_counter+=1\n cv2.imwrite(fname, cv2.cvtColor(img, cv2.COLOR_RGB2BGR))\n except Exception as e:\n print(\"cant process {}\".format(video))\n pass\n\n\n\n","repo_name":"aditbiswas1/fish-detection","sub_path":"preprocess_data/generate_fish_for_binary.py","file_name":"generate_fish_for_binary.py","file_ext":"py","file_size_in_byte":2941,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"7"} +{"seq_id":"819045045","text":"import matplotlib\nmatplotlib.use('Agg')\nimport math\nimport dash\nimport dash_bootstrap_components as dbc\nfrom dash import dcc, html\nfrom dash.dependencies import Input, Output\nimport pandas as pd\nimport networkx as nx\nimport matplotlib.pyplot as plt\nimport io\nimport base64\n\n# Entity Class\nclass Entity:\n def __init__(self, name, type):\n self.name = name\n self.type = type\n self.relations = {}\n\n def add_relation(self, relation, target):\n if relation not in self.relations:\n self.relations[relation] = []\n if target not in self.relations[relation]:\n self.relations[relation].append(target)\n\n# Load CSVs\nrelationships = pd.read_csv('relationships.csv')\nparticipants = pd.read_csv('participants.csv')\n\n# Initialize entities from participants.csv\nentities = {row['Name']: Entity(row['Name'], row['Type']) for _, row in participants.iterrows()}\n\n# Populate relationships from relationships.csv\nfor _, row in relationships.iterrows():\n subject_type, relationship, object_type = row['SubjectType'], row['Relationship'], row['ObjectType']\n for entity_name, entity in entities.items():\n if entity.type == subject_type:\n for target_name, target in entities.items():\n if target.type == object_type:\n entity.add_relation(relationship, target_name)\n\n# Dash App\napp = dash.Dash(__name__, external_stylesheets=[dbc.themes.BOOTSTRAP])\n\nparticipants_options = [{'label': entity.name, 'value': entity.name} for entity in entities.values()]\nrelationship_options = list(set([rel for entity in entities.values() for rel in entity.relations]))\nrelation_types = relationships['RelationType'].unique()\n\napp.layout = dbc.Container([\n html.H1(\"Relationship Viewer\"),\n dbc.Row([\n dbc.Col([\n html.Label(\"Select Participants:\"),\n dcc.Dropdown(id='participants-dropdown', options=participants_options, multi=True, value=list(entities.keys())),\n ], width=4),\n dbc.Col([\n html.Label(\"Select type of relation:\"),\n dcc.Dropdown(id='relation-type-dropdown', options=[{'label': rel_type, 'value': rel_type} for rel_type in relation_types], multi=True),\n ], width=4),\n dbc.Col([\n html.Label(\"Select Relationships:\"),\n dcc.Dropdown(id='relationships-dropdown', multi=True, value=relationship_options),\n ], width=4),\n ]),\n dbc.Row([\n dbc.Col([\n dbc.Button(\"Run\", id=\"run-btn\", color=\"primary\", className=\"mt-2\")\n ], width=2)\n ]),\n dbc.Row([\n dbc.Col([\n html.Img(id='network-graph', className=\"mt-4\")\n ])\n ]),\n], fluid=True)\n\ndef generate_layout(G):\n pos = nx.spring_layout(G)\n for node, data in G.nodes(data=True):\n if data['type'] == 'Human':\n r = 2\n angle = hash(node) % 360\n pos[node] = (r * math.cos(math.radians(angle)), r * math.sin(math.radians(angle)))\n return pos\n\n@app.callback(\n Output('relationships-dropdown', 'options'),\n Output('relationships-dropdown', 'value'),\n Input('relation-type-dropdown', 'value')\n)\ndef filter_relationships_by_type(selected_relation_type):\n if not selected_relation_type:\n options = [{'label': rel, 'value': rel} for rel in relationship_options]\n return options, []\n filtered_relationships = relationships[relationships['RelationType'].isin(selected_relation_type)]\n available_relations = filtered_relationships['Relationship'].unique()\n options = [{'label': rel, 'value': rel} for rel in available_relations]\n return options, list(available_relations)\n\n@app.callback(\n Output('network-graph', 'src'),\n Input('run-btn', 'n_clicks'),\n Input('participants-dropdown', 'value'),\n Input('relationships-dropdown', 'value')\n)\ndef update_output(n_clicks, selected_participants, selected_relationships):\n if n_clicks is None:\n return dash.no_update\n\n G = nx.DiGraph()\n color_map = {\n \"Human\": \"red\",\n \"Food\": \"yellow\",\n \"Animal\": \"blue\",\n \"Pet\": \"orange\",\n \"Toy\": \"green\",\n \"Vehicle\": \"purple\"\n }\n\n for entity_name in selected_participants:\n entity = entities[entity_name]\n G.add_node(entity.name, type=entity.type)\n\n for entity_name in selected_participants:\n entity = entities[entity_name]\n for relation, targets in entity.relations.items():\n if relation not in selected_relationships:\n continue\n for target in targets:\n if target in selected_participants:\n G.add_edge(entity.name, target, relation=relation)\n\n pos = nx.shell_layout(G) # or any other layout you prefer\n node_colors = [color_map.get(data['type'], \"gray\") for node, data in G.nodes(data=True)]\n labels = nx.get_edge_attributes(G, 'relation')\n \n non_self_edges = [(u, v) for u, v in G.edges() if u != v]\n\n # 1. Adjust label position for curved edges.\n # This creates an offset for edge labels. The values can be fine-tuned for best visual result.\n label_pos = {}\n for u, v, data in G.edges(data=True):\n if u != v: # Only for non-self-edges\n x0, y0 = pos[u]\n x1, y1 = pos[v]\n label_pos[(u, v)] = ((x0 + x1) / 2 + (y1 - y0) * 0.2, (y0 + y1) / 2 + (x0 - x1) * 0.2)\n\n # 2. Color the labels based on the target node (object) of the edge.\n label_colors = {(u, v): color_map[G.nodes[v]['type']] for u, v in non_self_edges}\n\n edge_colors = [color_map[G.nodes[v]['type']] for u, v in G.edges() if u != v]\n\n plt.figure(figsize=(10, 6))\n \n # Draw the nodes and non-self-edges with the specified edge colors and curved arrows\n nx.draw(G, pos, with_labels=True, node_size=2000, node_color=node_colors, font_size=10, width=2, alpha=0.6, edge_color=edge_colors, arrowsize=20, edgelist=non_self_edges, connectionstyle=\"arc3,rad=0.2\")\n\n # Filter edge labels\n filtered_labels = {(u, v): d for (u, v), d in labels.items() if u != v and u in pos and v in pos}\n \n # Filter the label_colors to match the filtered_labels\n filtered_label_colors = [label_colors[edge] for edge in filtered_labels.keys()]\n\n for edge, color in zip(filtered_labels.keys(), filtered_label_colors):\n label = filtered_labels[edge]\n x, y = pos[edge[0]]\n x2, y2 = pos[edge[1]]\n x_avg, y_avg = (x + x2) / 2, (y + y2) / 2 # Midpoint of the edge\n\n plt.text(x_avg, y_avg, label, color=color)\n\n buf = io.BytesIO()\n plt.savefig(buf, format=\"png\")\n buf.seek(0)\n base64_image = base64.b64encode(buf.getvalue()).decode('utf-8')\n\n return \"data:image/png;base64,{}\".format(base64_image)\n\nif __name__ == \"__main__\":\n app.run_server(debug=True, port=8050)\n\nif __name__ == \"__main__\":\n app.run_server(debug=True, port=8050)","repo_name":"PedroNVSRamos/relationships-visualization-rug","sub_path":"relationships.py","file_name":"relationships.py","file_ext":"py","file_size_in_byte":6848,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"16439041057","text":"\nfrom robodk import robolink \nfrom robodk import robomath \nimport numpy as np\nimport cv2 as cv\nimport os\nimport pandas as pd\n\n\nif __name__ == '__main__':\n RDK = robolink.Robolink()\n CAM_NAME = 'My Camera'\n CAM_PARAMS = 'SIZE=640x480'\n WINDOW_NAME = 'My Camera Feed'\n BASE_PATH = '/Users/akshitshishodia/tracker/roboDK /test2/test'\n data = pd.DataFrame(columns=['name','joint_pose'])\n cnt = 0 \n\n\n inventory = RDK.Item('inventory')\n cam_item = RDK.Item(CAM_NAME, robolink.ITEM_TYPE_CAMERA)\n if not cam_item.Valid():\n cam_item = RDK.Cam2D_Add(RDK.AddFrame(CAM_NAME + ' Frame'), CAM_PARAMS)\n cam_item.setName(CAM_NAME)\n cam_item.setParam('Open', 1)\n\n while cam_item.setParam('isOpen') == '1':\n\n img_socket = None\n bytes_img = RDK.Cam2D_Snapshot('', cam_item)\n joint_pose = inventory.Joints()\n\n if isinstance(bytes_img, bytes) and bytes_img != b'':\n nparr = np.frombuffer(bytes_img, np.uint8)\n img_socket = cv.imdecode(nparr, cv.IMREAD_COLOR)\n if img_socket is None:\n break\n name = str(cnt)+ \".png\"\n path = os.path.join(BASE_PATH,name)\n cv.imwrite(path,img_socket)\n data = data.append({'name':name,'joint_pose':joint_pose[0][0]},ignore_index = True)\n data.to_csv(\"name_and_pose2.csv\")\n cnt+=1\n cv.imshow(WINDOW_NAME, img_socket)\n\n\n\n cv.destroyAllWindows()\n RDK.Cam2D_Close(cam_item)\n","repo_name":"Akshit0601/Anomaly-Detection","sub_path":"roboDK /main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1451,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"18982530294","text":"import gym\nimport misc.controller as ctrl\nfrom misc.agents import learning, linear\nimport matplotlib.pyplot as plt\n\nenv = gym.make('CartPole-v0')\n\np = 0.5\nn = 1000\nn_average = 100\nrender = False\n\nagent_a = learning.LearningAgent(\n learning_type=learning.LearningType.MONTECARLO)\nagent_b = learning.LearningAgent(learning_type=learning.LearningType.QLEARNING)\nagent_c = linear.LearningLinearAgent()\na = ctrl.Controller(env, agent_a)\nb = ctrl.Controller(env, agent_b)\nc = ctrl.Controller(env, agent_c)\nc.run_episodes(n, render=render, p_noise=p)\na.run_episodes(n, render=render, p_noise=p)\nb.run_episodes(n, render=render, p_noise=p)\n\nya = a.stats.getRollingAverage(n_average)\nyb = b.stats.getRollingAverage(n_average)\nyc = c.stats.getRollingAverage(n_average)\nx = range(len(ya))\nplt.plot(x, ya, x, yb, x, yc)\nplt.title(\"p = 0.5\")\nplt.show()\n\n# env.close()\n","repo_name":"tdardinier/InterGP","sub_path":"src/temp/simulation.py","file_name":"simulation.py","file_ext":"py","file_size_in_byte":858,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"37686218512","text":"import re\n\ndef is_hipercard(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 19:\n if re.match('^606282|3841\\d{2}',''.join(n[:6])):\n return True\n return False\n\ndef is_dankort(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 19:\n if re.match('^(5019)\\d+$',n):\n return True\n return False\n\ndef is_bcglobal(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 19:\n if re.match('^(6541|6556)[0-9]{12}$',n):\n return True\n return False\n\ndef is_koreancard(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 19:\n if re.match('^9[0-9]{15}$',n):\n return True\n return False\n\ndef is_carteblanche(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 19:\n if re.match('^389[0-9]{11}$',n):\n return True\n return False\n\ndef is_instapayment(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 16 and length <= 19:\n if re.match('^(6360)\\d+$',n) or re.match('^63[7-9][0-9]{13}$',n):\n return True\n return False\n\ndef is_laser(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['6706','6709','6771']\n if length >= 16 and length <= 19:\n if ''.join(n[:4]) in form or re.match('^(6304|6706|6709|6771)[0-9]{12,15}$',n):\n return True\n return False\n\n\ndef is_bcmc(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['6703']\n if length >= 13 and length <= 19:\n if ''.join(n[:4]) in form:\n return True\n return False\n\ndef is_solo(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['6334','6767']\n if length >= 16 and length <= 19:\n if ''.join(n[:4]) in form or re.match('^(6334|6767)[0-9]{12}|(6334|6767)[0-9]{14}|(6334|6767)[0-9]{15}$',n):\n return True\n return False\n\ndef is_switch(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['633110','633312','633304','633303','633301','633300']\n if length >= 16 and length <= 19:\n if ''.join(n[:6]) in form or re.match('^(4903|4905|4911|4936|6333|6759)[0-9]{12}|(4903|4905|4911|4936|6333|6759)[0-9]{14}|(4903|4905|4911|4936|6333|6759)[0-9]{15}|564182[0-9]{10}|564182[0-9]{12}|564182[0-9]{13}|633110[0-9]{10}|633110[0-9]{12}|633110[0-9]{13}$',n):\n return True\n return False\n\ndef is_jcb(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length == 16:\n if ''.join(n[:4]) in strings_between(3528, 3589)or re.match('^(?:2131|1800|35\\d{3})\\d{11}$',n):\n return True\n return False\n\ndef is_unionpay(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 12 and length <= 19:\n if re.match('^(62|88)\\d+$',n) or re.match('^(62[0-9]{14,17})$',n):\n return True\n return False\n\ndef is_visa(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 13 and length <= 16:\n if n[0] == '4':\n if(re.match('^4[0-9]{12}(?:[0-9]{3})?$',n)):\n return True\n return False\n\ndef is_dinersclub(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['30','36']\n if length >= 13 and length <= 19:\n if ''.join(n[:2]) in form or (re.match('^3(?:0[0-5]|[68][0-9])[0-9]{11}$',n)):\n return True\n return False\n\ndef is_cartebancaire(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['4035','4360']\n if length >= 13 and length <= 19:\n if ''.join(n[:4]) in form:\n return True\n return False\n\ndef is_vpay(n):\n \"\"\"Checks if credit card number fits the visa format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['4370','482']\n if length >= 13 and length <= 19:\n if ''.join(n[:4]) in form or ''.join(n[:3]) in form:\n return True\n return False\n\ndef is_visa_electron(n):\n \"\"\"Checks if credit card number fits the visa electron format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['026', '508', '844', '913', '917','405']\n\n if length == 16:\n if n[0] == '4':\n if ''.join(n[1:4]) in form or ''.join(n[1:6]) == '17500':\n return True\n return False\n\n\ndef is_mastercard(n):\n \"\"\"Checks if credit card number fits the mastercard format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length >= 16 and length <= 19:\n if ''.join(n[:2]) in strings_between(50, 56) or (re.match('^5[1-5][0-9]{14}$',n)):\n return True\n return False\n\ndef is_rupay(n):\n \"\"\"Checks if credit card number fits the mastercard format.\"\"\"\n n, length = str(n), len(str(n))\n if length >= 13 and length <= 19:\n if ''.join(n[:6]) in strings_between(508500, 508999) or ''.join(n[:6]) in strings_between(606985, 607984) or ''.join(n[:6]) in strings_between(608001, 608500) or ''.join(n[:6]) in strings_between(652150, 653149):\n return True\n return False\n\n\ndef is_elo(n):\n \"\"\"Checks if credit card number fits the mastercard format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['5066']\n if length >= 16 and length <= 19:\n if ''.join(n[:4]) in form:\n return True\n return False\n\ndef is_amex(n):\n \"\"\"Checks if credit card number fits the american express format.\"\"\"\n n, length = str(n), len(str(n))\n if length == 15:\n if n[0] == '3' and (n[1] == '4' or n[1] == '7'):\n if(re.match('^3[47][0-9]{13}$',n)):\n return True\n else:\n return False\n return False\n\n\ndef is_maestro(n):\n \"\"\"Checks if credit card number fits the maestro format.\"\"\"\n n, length = str(n), len(str(n))\n form = ['5018', '5020', '5038', '5893', '6304',\n '6759', '6761', '6762', '6763','6731',\n '06','6779','677','678','679']\n\n if length >= 12 and length <= 19:\n if ''.join(n[:4]) in form:\n return True\n elif ''.join(n[:3]) in form:\n return True;\n elif ''.join(n[:2]) in form:\n return True;\n return False\n\n\ndef is_discover(n):\n \"\"\"Checks if credit card number fits the discover card format.\"\"\"\n n, length = str(n), len(str(n))\n\n if length == 16:\n if n[0] == '6':\n if ''.join(n[1:4]) == '011' or n[1] == '5':\n return True\n elif n[1] == '4' and n[2] in strings_between(4, 10):\n return True\n elif ''.join(n[1:6]) in strings_between(22126, 22926):\n return True\n return False\n\n\ndef get_format(n):\n \"\"\"Gets a list of the formats a credit card number fits.\"\"\"\n formats = []\n\n if is_visa(n):\n formats.append('visa')\n if is_visa_electron(n):\n formats.append('visa electron')\n if is_mastercard(n):\n formats.append('mastercard')\n if is_amex(n):\n formats.append('amex')\n if is_maestro(n):\n formats.append('maestro')\n if is_discover(n):\n formats.append('discover')\n if is_rupay(n):\n formats.append('rupay')\n if is_hipercard(n):\n formats.append('hipercard')\n if is_dankort(n):\n formats.append('dankort')\n if is_instapayment(n):\n formats.append('instapayment')\n if is_laser(n):\n formats.append('laser')\n if is_bcmc(n):\n formats.append('bcmc')\n if is_jcb(n):\n formats.append('jcb')\n if is_unionpay(n):\n formats.append('unionpay')\n if is_solo(n):\n formats.append('solo')\n if is_dinersclub(n):\n formats.append('dinersclub')\n if is_cartebancaire(n):\n formats.append('cartebancaire')\n if is_elo(n):\n formats.append('elo')\n if is_vpay(n):\n formats.append('vpay')\n if is_switch(n):\n formats.append('switch')\n if is_carteblanche(n):\n formats.append('carteblanche')\n if is_bcglobal(n):\n formats.append('bcglobal')\n if is_koreancard(n):\n formats.append('koreancard')\n return formats\n\n\ndef strings_between(a, b):\n \"\"\"Generates a list of strings between a and b.\"\"\"\n return list(map(str, range(a, b)))\n","repo_name":"chetan92/CardValidator","sub_path":"CardValidator-master/cardvalidator/formatter.py","file_name":"formatter.py","file_ext":"py","file_size_in_byte":8817,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"7"} +{"seq_id":"9931105230","text":"# -*- coding: utf-8 -*-\n# @Time : 2020/3/9 11:54\n# @Author : 潘师傅\n# @File : XfpApi.py\nimport requests\nimport json\nfrom GlobalMap import GlobalMap\nfrom Config.Config import *\nimport unittest\nimport time\nimport random\nimport datetime\nimport calendar\n\n\nclass appletApi:\n\n def __init__(self):\n self.appletText = GlobalMap()\n\n def RandomText(self, textArr):\n \"\"\"指定字符串随机取值\"\"\"\n # ['你好啊','阿米里!','扣你七娃','你好','hello']\n length = len(textArr)\n if length < 1:\n return ''\n if length == 1:\n return str(textArr[0])\n randomNumber = random.randint(0, length - 1)\n return str(textArr[randomNumber])\n\n def Merge(self, dict1, dict2):\n return (dict2.update(dict1))\n\n def PostRequest(self, url, data, header=None, Status=1, files=None, saasCode=XfpsaasCode):\n \"\"\"post请求\"\"\"\n if header is not None:\n r = requests.post(url=(ApiXfpUrl + url),\n data=json.dumps(data, ensure_ascii=False),\n headers={\n 'Content-Type': 'application/json'\n\n })\n else:\n data1 = {\"page\": {\n 'size': '100',\n 'current': '1'\n },\n \"saasCode\": saasCode,\n \"saasCodeSys\": saasCode\n }\n self.Merge(data1, data)\n time.sleep(0.2)\n r = requests.post(url=(ApiXfpUrl + url),\n data=(json.dumps(data,\n ensure_ascii=False).encode(\"UTF-8\")),\n headers={\n 'Content-Type': 'application/json',\n 'Authorization': 'Bearer' + ' ' + self.appletText.get(\"user_token\")\n },\n files=files)\n r.raise_for_status()\n self.appletText.set_map('URL', ApiXfpUrl + url)\n globals()['XfpText'] = globals()['r.text'] = json.loads(r.text)\n self.appletText.set_map('XfpText', globals()['r.text'])\n self.appletText.set_map('ApiXfpUrl', url)\n self.appletText.set_map('msg', globals()['XfpText']['msg'])\n self.appletText.set_map('code', globals()['XfpText']['code'])\n self.appletText.set_map('data', globals()['XfpText']['data'])\n time.sleep(0.2)\n if Status == 1:\n try:\n assert \"成功\", globals()['r.text']['msg']\n except BaseException as e:\n print(\"断言错误,错误原因:%s\" % e)\n raise RuntimeError(self.appletText.get('URL'))\n if globals()['r.text']['code'] == 500:\n raise RuntimeError(self.appletText.get('ApiXfpUrl'))\n\n if r.elapsed.total_seconds() > 5:\n print('接口请求过慢')\n print(self.appletText.get('ApiXfpUrl'))\n if r.elapsed.total_seconds() > 10:\n print('接口请求过慢大于10秒')\n print(self.appletText.get('ApiXfpUrl'))\n\n def sendCodeWeiXin(self, userName='19859080323'):\n \"\"\"获取验证码\"\"\"\n self.PostRequest(url='/api/auth/sendCodeWeiXin',\n data={\n 'senderSource': '微信小程序',\n 'senderDevice': 'ces',\n 'userName': userName\n })\n\n def Login(self, userName=XfpUser, code='12345678'):\n \"\"\"登录\"\"\"\n self.PostRequest(url='/api/auth/loginByCodeWeiXin',\n data={\n \"senderSource\": \"微信小程序\",\n # \"senderDevice\": \"ces\",\n \"userName\": userName,\n \"code\": code})\n if self.appletText.get('msg') == '成功':\n if (globals()['XfpText']['data']['userDetail']) is not 'None':\n self.appletText.set_map('user_token', globals()['XfpText']['data']['token'])\n\n def LogIn(self, userName=XfpUser, password=XfpPwd, saasCode=XfpsaasCode, authCode=None, device=None):\n \"\"\"登录\"\"\"\n if device is None:\n device = deviceId\n if authCode is None:\n self.PostRequest(url='/api/auth/login',\n data={\"userName\": userName,\n 'saasCode': saasCode,\n 'deviceId': device,\n # 'deviceId': deviceId,\n \"password\": password},\n header=1)\n else:\n self.PostRequest(url='/api/auth/login',\n data={\"userName\": userName,\n 'saasCode': saasCode,\n 'authCode': authCode,\n \"password\": password},\n header=1)\n\n if self.appletText.get('msg') == '成功':\n if (globals()['XfpText']['data']['userDetail']) is not 'None':\n if authCode is None:\n self.appletText.set_map('user_token', globals()['XfpText']['data']['token'])\n\n else:\n self.appletText.set_map('user_token', globals()['XfpText']['data']['token'])\n try:\n self.appletText.set_map('resultStr', globals()['r.text']['data']['resultStr'])\n except:\n pass\n\n else:\n self.appletText.set_map('userId', globals()['XfpText']['data']['userDetail']['id'])\n\n else:\n self.appletText.set_map('data', globals()['XfpText']['data'])\n\n\nif __name__ == '__main__':\n a = appletApi()\n\n\n","repo_name":"yebenxiaozhang/Projects_xfj","sub_path":"XFP/PubilcAPI/appletApi.py","file_name":"appletApi.py","file_ext":"py","file_size_in_byte":5836,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"24481982341","text":"\"\"\"empty message\n\nRevision ID: 102d609bce75\nRevises: 12c9567eaed5\nCreate Date: 2015-01-25 22:28:39.076329\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '102d609bce75'\ndown_revision = '12c9567eaed5'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('wufoo_textarea_sentiment', sa.Column('survey_id', sa.Integer(), nullable=True))\n op.create_foreign_key(None, 'wufoo_textarea_sentiment', 'wufoo_survey', ['survey_id'], ['id'])\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.drop_constraint(None, 'wufoo_textarea_sentiment', type_='foreignkey')\n op.drop_column('wufoo_textarea_sentiment', 'survey_id')\n ### end Alembic commands ###\n","repo_name":"wigginslab/lean-workbench","sub_path":"migrations/versions/102d609bce75_.py","file_name":"102d609bce75_.py","file_ext":"py","file_size_in_byte":823,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"7"} +{"seq_id":"9878348943","text":"import sys\n\nn, m = map(int, sys.stdin.readline().split())\ntrees = list(map(int, sys.stdin.readline().split()))\n\nstart, end = 1, max(trees)\n\ndef get_woods(meter):\n woods = 0\n for tree in trees:\n woods += tree - meter if tree > meter else 0\n return woods\n\nwhile start <= end:\n mid = (start + end) // 2\n woods = get_woods(mid)\n \n if woods >= m:\n start = mid + 1\n else:\n end = mid - 1\n\nprint(end)","repo_name":"mchu7797/coding-tests-algorithm","sub_path":"2805.py","file_name":"2805.py","file_ext":"py","file_size_in_byte":437,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"24453659340","text":"import json\nfrom django.shortcuts import render,redirect\nfrom django.contrib.auth.decorators import login_required\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.contrib import messages\nfrom django.http import JsonResponse\nfrom datetime import datetime\nfrom .models import Task, Assignment\nfrom .forms import Task_creation_form\nfrom company.models import Company,Notification\n\n\n@login_required(login_url='account:login')\ndef get_task_info_based_on_status(request, status):\n user = request.user\n status = status\n task_query = Task.objects.filter(assignee=user, status=status)\n\n task_list = []\n for t in task_query:\n task = {\n 'finishing_expected': t.finishing_expected.strftime(\"%b %d, %Y, %I:%M %p\"),\n 'title':t.title,\n 'task_id':t.id,\n 'priority':t.priority,\n 'company':t.company.name,\n 'company_id':t.company.id\n }\n task_list.append(task)\n\n return JsonResponse({'data':task_list}, safe=False)\n\n\n@login_required(login_url='account:login')\ndef get_task_info_for_task_modal(request, id):\n task = Task.objects.get(pk=id)\n\n assignee_list = []\n for a in task.assignee.all():\n assignee = {\n 'assignee_id':a.id,\n 'assignee_name':a.username,\n 'assignee_img_url':a.profile_picture.url,\n }\n assignee_list.append(assignee)\n\n if task.finishing_date:\n finishing_date = task.finishing_date.strftime(\"%b %d, %Y, %I:%M %p\"),\n else:\n finishing_date = None,\n\n data = [{\n 'title':task.title,\n 'status':task.status,\n 'drescription':task.drescription,\n 'priority':task.priority,\n 'checklist':task.checklist,\n 'company':task.company.name,\n 'company_id':task.company.id,\n 'company_logo_url':task.company.company_logo.url,\n 'assignees':assignee_list,\n 'finishing_expected':task.finishing_expected.strftime(\"%b %d, %Y, %I:%M %p\"),\n 'date_created':task.date_created.strftime(\"%b %d, %Y, %I:%M %p\"),\n 'finishing_date':finishing_date,\n },]\n return JsonResponse({'data':data}, safe=False )\n\n# cancelling their implementation to shortend dev time,\n# use if to check null or not, then use below code, or response won't be ok\n# 'last_updated':task.last_updated.strftime(\"%b %d, %Y, %I:%M %p\"),\n# 'finishing_date':task.finishing_date.strftime(\"%b %d, %Y, %I:%M %p\"),\n\ndef user_workspace(request):\n tasks = Task.objects.filter(assignee=request.user, status='ongoing')\n context = {\n 'tasks':tasks,\n }\n return render(request, 'task/user-workspace.html', context)\n\n\n@login_required(login_url='account:login')\ndef show_company_list(request):\n user = request.user\n companies = Company.objects.filter(members=user)\n users_companies = Company.objects.filter(creator=user)\n\n context = {\n 'companies':companies,\n 'user_companies':users_companies,\n }\n return render(request, 'task/show-company-list.html', context)\n\n\n@login_required(login_url='account:login')\ndef create_task(request, c_id):\n company_id = c_id\n form = Task_creation_form(user_id=request.user.id, company_id=company_id)\n\n if request.method == 'POST':\n form = Task_creation_form(user_id=request.user.id,company_id=company_id,data=request.POST)\n if form.is_valid():\n task = form.save(commit=False)\n task.creator = request.user\n\n if task.status == 'finished':\n task.finishing_date = datetime.today()\n\n task.save()\n\n company = form.cleaned_data['company']\n for user in form.cleaned_data['assignee']:\n Assignment.objects.create(user=user, task=task)\n title = 'Task Assignment'\n subject = f\"{company.name} assigned you new Task.\"\n notification = Notification.objects.create(\n type = \"assignment\",\n title = title,\n subject = subject,\n company_sender = company,\n )\n notification.receivers.set([user])\n\n messages.success(request, 'Task Created Successfully')\n return redirect('task:user-workspace')\n\n context = {\n 'form':form\n }\n return render(request, 'task/create-task-form.html', context)\n\n\n@login_required(login_url='account:login')\ndef edit_task(request, id):\n task = Task.objects.get(pk=id)\n user_id = request.user.id\n company_id =task.company.id\n form = Task_creation_form(user_id=user_id,company_id=company_id,instance=task)\n\n if request.method == \"POST\":\n form = Task_creation_form(user_id=user_id,company_id=company_id,data=request.POST, instance=task)\n\n if form.is_valid():\n task = form.save(commit=False)\n\n if task.status == 'finished':\n task.finishing_date = datetime.today()\n\n task.save()\n company = form.cleaned_data['company']\n for user in form.cleaned_data['assignee']:\n try:\n assignment = Assignment.objects.get(user=user, task=task)\n except:\n assignment = None\n\n if assignment is None:\n Assignment.objects.create(user=user, task=task)\n title = 'Task Assignment'\n subject = f\"{company.name} assigned you new Task.\"\n notification = Notification.objects.create(\n type = \"assignment\",\n title = title,\n subject = subject,\n company_sender = company,\n )\n notification.receivers.set([user])\n\n messages.success(request, 'Task Updated')\n return redirect('task:user-workspace')\n\n context = {\n 'form':form,\n }\n return render(request, 'task/create-task-form.html', context)\n\n\n@login_required(login_url='account:login')\ndef delete_task(request, id):\n task = Task.objects.get(pk=id)\n\n context = {\n 'task':task,\n }\n return render(request, 'task/delete-task.html', context)\n\n@login_required(login_url='account:login')\ndef confirm_delete_task(request, id):\n task= Task.objects.get(pk=id)\n task.delete()\n messages.warning(request, 'Task deleted')\n return redirect(request.GET['next'] if 'next' in request.GET else 'task:user-workspace')\n\n\n","repo_name":"Yasir-GitAc/task-management","sub_path":"task/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5857,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"7"} +{"seq_id":"32151320389","text":"def zakres(pierwszy=0, ostatni=100, krok=1):\n liczba = pierwszy\n while liczba < ostatni:\n yield liczba\n liczba += krok\n\nzakres1 = zakres(1,50)\n#zakres2 = zakres(1,50)\n\nfor a in zakres1:\n print(a)\n\nfor b in zakres1:\n print(b)","repo_name":"marcingastol/WSB_Python_2022-23_GR3","sub_path":"T3/Niedziela/2.py","file_name":"2.py","file_ext":"py","file_size_in_byte":250,"program_lang":"python","lang":"pl","doc_type":"code","stars":3,"dataset":"github-code","pt":"7"}