태그 id 값을 넣어줌\n#
\n#print(detected_value.text)\n\n\n# 페이지 이동\nurl = \"https://play.google.com/store/movies/top\"\nbrowser.get(url)\n\nimport time\ninterval = 2 # 4초에 한번씩 스크롤 내림\n\n# 현재 문서 높이를 가져와서 저장\nprev_height = browser.execute_script(\"return document.body.scrollheght\")\n\n# 반복 수행\nwhile True:\n # 스크롤을 가장 아래로 내림\n browser.execute_script(\"window.scrollTo(0, document.body.scrollHeight)\")\n\n # 페이지 로딩 대기\n time.sleep(interval)\n\n # 현재 문서 높이를 가져와서 저장\n curr_height = browser.execute_script(\"return document.body.scrollheght\")\n if curr_height == prev_height:\n break\n\n prev_height = curr_height\n\nprint(\"스크롤 완료\")\nbrowser.get_screenshot_as_file(\"google_movie.png\")\n \n\nimport requests\nfrom bs4 import BeautifulSoup\n\n\nsoup = BeautifulSoup(browser.page_source, \"lxml\")\n\n#movies = soup.find_all(\"div\", attrs={\"class\":[\"ImZGtf mpg5gc\", \"Vpfmgd\"]})\nmovies = soup.find_all(\"div\", attrs={\"class\":\"Vpfmgd\"})\nprint(len(movies))\n\n\nfor movie in movies:\n title = movie.find(\"div\", attrs={\"class\":\"WsMG1c nnK0zc\"}).get_text()\n print(title)\n # 활인 전 가격\n original_price = movie.find(\"span\", attrs={\"class\":\"SUZt4c djCuy\"})\n if original_price:\n original_price = original_price.get_text()\n else:\n # print(title, \"활인되지 않은 영화 제외\")\n continue\n\n # 활인된 가격 /
₩7,000\n price = movie.find(\"span\", attrs={\"class\", \"VfPpfd ZdBevf i5DZme\"}).get_text()\n\n # 링크 /
\n link = movie.find(\"a\", attrs={\"class\":\"JC71ub\"})[\"href\"] #모든 영화 포스터 링크들이 클래스가 \"JC71ub\"이다\n # \"https://play.google.com\" + link\n\n print(f\"제목 : {title}\")\n print(f\"할인 전 가격 : {original_price}\")\n print(f\"할인 후 가격 : {price}\")\n print(\"링크 : \" + \"https://play.google.com\" + link)\n print(\"-\" * 100)\n\n browser.quit()\n\n\n''' \n#실행결과\n\n'''","repo_name":"SEONGJAE-YOO/python-project.","sub_path":"headless_google_movie3.py","file_name":"headless_google_movie3.py","file_ext":"py","file_size_in_byte":3248,"program_lang":"python","lang":"ko","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"73018229065","text":"import argparse\nimport sys\nimport cPickle\nfrom trapeza import *\nfrom trapeza.match import *\n\n\ndef main():\n parser = argparse.ArgumentParser(description=\"Manipulate and combine tabular data files. Use this utility to \"\n \"process a master data set for use with trapeza-match.\")\n parser.add_argument(\"-o\", \n \"--output\", \n type=argparse.FileType('wb'), \n default=sys.stdout, \n help=\"Specify an output file (default standard output)\")\n parser.add_argument(\"-i\", \n \"--input-format\", \n choices=formats.available_input_formats(),\n default=\"csv\",\n help=\"Treat input read from stdin and from files whose type cannot be inferred as being in \"\n \"the specified format. Default is CSV.\")\n parser.add_argument(\"--input-encoding\",\n default=\"utf-8\",\n help=\"Treat input data as the specified encoding (for input formats that support Unicode). \"\n \"Column names specified on the command line will be treated as the same encoding.\")\n parser.add_argument(\"-p\", \n \"--profile\", \n type=argparse.FileType('rb'),\n help=\"Specify the profile spreadsheet\")\n parser.add_argument(\"-m\", \n \"--master\", \n type=argparse.FileType('rb'),\n help=\"Specify the master spreadsheet\")\n parser.add_argument(\"--primary-key\", \n help=\"Set the column name in the master sheet where unique identifiers are stored.\")\n\n args = parser.parse_args()\n \n if args.profile is None or args.master is None or args.primary_key is None:\n sys.stderr.write(\"{}: you must specify a master and profile sheet and a primary key column.\\n\"\n .format(sys.argv[0]))\n exit(1)\n try:\n profile = Profile(source=load_source(args.profile, get_format(args.profile.name, args.input_format),\n args.input_encoding))\n master = load_source(args.master, get_format(args.master.name, args.input_format), args.input_encoding)\n except Exception:\n sys.stderr.write(\"{}: an error occured while loading input files.\\n\".format(sys.argv[0]))\n return 1\n \n master.set_primary_key(args.primary_key.decode(args.input_encoding))\n\n pm = ProcessedSource(master, True, profile)\n pm.process()\n\n try:\n cPickle.dump(pm, args.output, protocol=cPickle.HIGHEST_PROTOCOL)\n except Exception as e:\n sys.stderr.write(\"{}: an error occured while writing output: {}\\n\".format(sys.argv[0], e))\n return 1\n \n return 0\n\nif __name__ == '__main__':\n exit(main())\n","repo_name":"davidmreed/trapeza","sub_path":"trapeza-process.py","file_name":"trapeza-process.py","file_ext":"py","file_size_in_byte":2935,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"44501654027","text":"from Course import Course\nfrom DataHandler import DataHandler as dataHandl\nfrom ErrorHandler import ErrorHandler as errorHandl\nfrom FileOperations import FileOperations as fileOps\nfrom UserInterface import UserInterface as UI\n\nLINE_STARTER = \">\"\n\n\ndef main():\n main_proc_enable = True\n data_handler = None\n input_filename = input(\"Input file: \")\n\n file_handler = fileOps(input_filename)\n\n if (file_handler.status):\n data_handler = dataHandl(file_handler.data)\n else:\n main_proc_enable = False\n\n user_if = UI(LINE_STARTER)\n\n while(main_proc_enable):\n has_valid_command = user_if.get_command()\n\n if (has_valid_command and not user_if.is_quit()):\n command = user_if.pass_command()\n main_proc_enable = data_handler.process_command(command)\n else:\n main_proc_enable = False\n\n\nif __name__ == \"__main__\":\n main()","repo_name":"Ihsara/Education-center","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":900,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"24423413330","text":"import csv\nimport os\n\ninsert_query = \"INSERT INTO md5 (video_name, md5_val) VALUES ('%s', '%s');\"\n\n# Reads from the csv file.\nqueries = []\nwith open('md5.csv') as file:\n csv_file = csv.reader(file, delimiter=',')\n\n # Reads line by line.\n for row in csv_file:\n video_name = os.path.splitext(row[0])[0]\n md5_val = row[1]\n query = insert_query % (video_name, md5_val)\n queries.append(query)\n\n# Writes the result to the given file.\nwith open('md5.sql', 'a') as file:\n for query in queries:\n file.write(query + \"\\n\")\n","repo_name":"yunpengn/AudioDup","sub_path":"collect_md5.py","file_name":"collect_md5.py","file_ext":"py","file_size_in_byte":559,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"30247658674","text":"import requests\nimport pandas as pd\nfrom sqlalchemy import create_engine\nfrom library import user\n\npd.set_option('display.unicode.east_asian_width', True) # 对齐\npd.set_option('display.max_columns', 5) # 设置最大列数\n\n\ndef cloud_warehouse(js_code, category, star=4):\n \"\"\"\n 根据json语句查询云仓信息\n :param js_code: 查询语句\n :param category: 执行的操作类别\n :parameter star: 输出DataFrame 从第几列开始,默认第4列\n :return: DataFrame\n \"\"\"\n cloud = user.CloudWarehouse\n url = cloud.url # 接口网址\n json_data = {\n \"funcType\": \"select\", # 方法类别\n \"callCategory\": category, # 操作类别\n \"id\": cloud.id,\n \"pwd\": cloud.password,\n # 字段内容\n \"jsondata\": js_code}\n\n res = requests.post(url=url, json=json_data) # 用Post方法调用接口\n js_data = res.json() # 解析获取到的数据,这里是json,解���后的数据类型为dict\n\n if js_data['Flag'] is False:\n print(\"警告:\", js_data['ErrInfo'])\n else:\n if js_data[\"MsgInfo\"]:\n js_msg_info = js_data[\"MsgInfo\"] # 截取dict中自己要的值\n # key = list(js_MsgInfo[0].keys()) # 获取dict中的key值,并放在list中\n basket = list() # 一篮子计划,你懂的\n for dict_i in js_msg_info:\n # 用dict生成dataframe,必须指定index,这里的index为dict中的key值\n df_i = pd.DataFrame.from_dict(dict_i, orient='index')\n basket.append(df_i.T) # 因为index 为key值,所以要做矩阵转置\n df = pd.concat(basket, ignore_index=True) # 按列明组合,并重新定义index\n df = df.iloc[:, star:] # 去掉前面无用数值\n return df\n else:\n return js_data\n\n\ndef to_sql(data_frame, table_name, database='my_data', exists='fail'):\n \"\"\"\n 将DataFrame插入至数据库\n :param data_frame: DataFrame格式数据\n :param table_name: 表名称\n :parameter database: 数据库名称,默认为:my_data\n :parameter exists: 如果表名称存在则报错,可改为:replace替换、append追加\n :return:\n \"\"\"\n sql = user.SqlServer\n engine = create_engine(\n f\"mssql+pymssql://{sql.yaoud_user}:{sql.yaoud_password}@{sql.yaoud_IP}/{database}?charset=utf8\", echo=False)\n data_frame.to_sql(table_name, engine, index=False, if_exists=exists) # 有replace替换、append追加\n print(f'{table_name} 表已添加至 {database}')\n\n\nif __name__ == '__main__':\n all_id = user.CloudWarehouse\n data = {\n \"Operator_Id\": all_id.Operator_Id,\n \"Con_Id\": all_id.Con_Id,\n \"Ldc_Id\": all_id.Ldc_Id,\n \"Businessbill_No\": \"\",\n \"Goods_No\": \"\",\n \"Start_Date\": \"2022-07-07\",\n \"End_Date\": \"2022-07-07\"\n }\n\n operation = \"kcztbh\"\n\n js = cloud_warehouse(data, operation)\n print(js)\n","repo_name":"mbalucard/Company_project","sub_path":"library/json_data.py","file_name":"json_data.py","file_ext":"py","file_size_in_byte":2936,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"73662300745","text":"from Crypto.PublicKey import RSA\nimport Crypto.Random\nfrom Crypto.Signature import PKCS1_v1_5\nfrom Crypto.Hash import SHA256\nimport binascii\n\n\nclass Wallet:\n def __init__(self, node_id):\n self.private_key = None\n self.public_key = None\n self.node_id = str(node_id)\n\n @property\n def public_key(self):\n return self.__public_key\n\n @public_key.setter\n def public_key(self, value):\n self.__public_key = value\n\n def save_keys_in_file(self):\n if self.private_key is not None and self.public_key is not None:\n try:\n with open(f'wallet-{self.node_id}', mode='w') as output_file:\n output_file.write(self.private_key)\n output_file.write('\\n')\n output_file.write(self.public_key)\n except (IOError, IndexError):\n raise Exception('Saving wallet failed.')\n\n def create_keys(self):\n private_key, public_key = self.generate_keys()\n self.private_key = private_key\n self.public_key = public_key\n\n def load_keys(self):\n try:\n with open(f'wallet-{self.node_id}', mode='r') as input_file:\n file_content = input_file.readlines()\n self.private_key = file_content[0].strip('\\n')\n self.public_key = file_content[1]\n except IOError:\n raise Exception('Loading wallet failed.')\n\n def generate_keys(self):\n private_key = RSA.generate(2048, Crypto.Random.new().read)\n public_key = private_key.publickey()\n return (binascii.hexlify(private_key\n .exportKey(format='DER')).decode('ascii'),\n binascii.hexlify(public_key\n .exportKey(format='DER')).decode('ascii'))\n\n def sign_transaction(self, sender, recipient, amount):\n if self.public_key is None:\n raise Exception('Wallet is not set up.')\n signer = PKCS1_v1_5.new(RSA.importKey(binascii\n .unhexlify(self.private_key)))\n payload_hash = SHA256.new((str(sender) + str(recipient) + str(amount))\n .encode('utf8'))\n signature = signer.sign(payload_hash)\n return binascii.hexlify(signature).decode('ascii')\n\n @staticmethod\n def verify_transaction(transaction):\n public_key = RSA.importKey(binascii.unhexlify(transaction.sender))\n verifier = PKCS1_v1_5.new(public_key)\n payload_hash = SHA256.new((str(transaction.sender) + str(\n transaction.recipient) + str(\n transaction.amount))\n .encode('utf8'))\n return verifier.verify(payload_hash,\n binascii.unhexlify(transaction.signature))\n","repo_name":"jaqb8/blockchain-app","sub_path":"node/wallet.py","file_name":"wallet.py","file_ext":"py","file_size_in_byte":2795,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"17851275187","text":"import sys\nimport time\n\nfrom chatting_client import ChattingClient\nfrom chatting_server import ChattingServer\nfrom msg_parser import MsgParser\n\nPORT_NUM = 60860\n\n\nclass P2pChatting(object):\n\n def __init__(self, name, receivedMsgHandler):\n self.my_name = name\n self.is_chairman = False\n self.members = {}\n self.msgHandler = receivedMsgHandler\n self.client = None\n self.server = None\n self.addr = None\n self.listen_port = PORT_NUM\n\n def setUserName(self, name):\n self.my_name = name\n\n def setListenPort(self, port):\n self.listen_port = port\n\n def setAddr(self, addr):\n self.addr = addr\n if addr == 'localhost' or addr == '127.0.0.1':\n self.is_chairman = True\n else:\n self.is_chairman = False\n\n def start(self):\n self.stop()\n self.server = ChattingServer(self, self.addr, self.listen_port)\n self.server.start()\n\n if not self.is_chairman:\n self.client = ChattingClient(self.my_name, self, self.addr, PORT_NUM)\n self.client.start()\n\n def stop(self):\n if self.server:\n self.server.stop()\n if self.client:\n self.client.stop()\n\n def broadcastMsg(self, msg):\n self.msgHandler(self.my_name + ' : ' + msg)\n msg = MsgParser.buildChatMsg(self.my_name, msg)\n if self.client:\n self.client.sendMsg(msg)\n if self.server:\n self.server.sendMsg(msg)\n '''\n def onNewConnection(self, addr, name):\n if self.is_chairman:\n redirect, newAddr = self._allocParentNode(addr)\n if redirect:\n self.server.redirect(addr, name, newAddr)\n '''\n\n def removeMember(self, addr):\n ip, port = addr\n for name, info in self.members.items():\n if info['ip'] == ip and info['port'] == port:\n del self.members[name]\n print('Member {} removed'.format(name))\n return\n\n def sendToUpperlayer(self, msg):\n self.msgHandler(msg['name'] + ' : ' + msg['chat'])\n if self.is_chairman:\n return\n self.client.sendMsg(msg)\n\n def sendToUnderlayer(self, msg):\n self.msgHandler(msg['name'] + ' : ' + msg['chat'])\n self.server.sendMsg(msg)\n\n def onCtrlMsg(self, msg, addr):\n if msg['type'] == 'join':\n self.onJoinMsg(msg, addr)\n elif msg['type'] == 'redirect':\n self.onRedirectMsg(msg, addr)\n elif msg['type'] == 'leave':\n self.onLeaveMsg(msg, addr)\n else:\n print(msg, addr)\n\n def onLeaveMsg(self, msg, addr):\n name = msg['name']\n self.msgHandler(name + ' is leaving, bye!')\n if name in self.members:\n del self.members[name]\n\n def onRedirectMsg(self, msg, addr):\n self.addr = msg['parent_node']\n print('stopping')\n self.client.setServer(self.addr, PORT_NUM)\n self.client.connectServer()\n print('Client reconnect to {}'.format(self.addr))\n\n def onJoinMsg(self, msg, addr):\n if not self.is_chairman:\n return\n ip, port = addr\n welcome = msg['name'] + ' has joined the chatting, welcome!'\n self.broadcastMsg(welcome)\n parentName, parentIP = self._allocParentNode(ip)\n self.members[msg['name']] = {'ip': ip, 'port': port}\n if parentIP:\n print('Redirect', msg['name'], 'to', parentName)\n time.sleep(2)\n self.server.redirect(addr, msg['name'], parentIP)\n\n def _allocParentNode(self, ip):\n ip = ip.split('.')\n for name, info in self.members.items():\n addr = info['ip'].split('.')\n if ip[0] == addr[0] and ip[1] == addr[1] and ip[2] == addr[2]:\n return name, info['ip']\n return None, None\n\n def leave(self):\n if self.is_chairman:\n return\n\n if self.client:\n leaveMsg = MsgParser.buildLeaveMsg(self.my_name)\n self.client.sendMsg(leaveMsg)\n time.sleep(1)\n\n if self.server:\n redirectMsg = MsgParser.buildRedirectMsg(self.addr)\n self.server.sendMsg(redirectMsg)\n time.sleep(3)\n\nclass TestMsgHandler:\n def OnRecvedMsg(self, msg):\n print('Received Msg:', msg)\n\n\nif __name__ == '__main__':\n if len(sys.argv) < 3:\n print(\"Correct usage: script, IP address, port number, name\")\n exit()\n IP_address = str(sys.argv[1])\n name = str(sys.argv[2])\n testHandler = TestMsgHandler()\n chatting = P2pChatting(name, testHandler.OnRecvedMsg)\n chatting.setAddr(IP_address)\n if len(sys.argv) == 4:\n portNum = int(sys.argv[3])\n chatting.setListenPort(portNum)\n chatting.start()\n print('Input your msg:')\n while True:\n data = sys.stdin.readline().strip()\n if data == 'exit' or data == 'quit':\n break\n elif data == 'leave' or data == 'Leave':\n chatting.leave()\n break\n else:\n chatting.broadcastMsg(data)\n print('Ending!')\n chatting.stop()\n","repo_name":"FrankDuan/MultipartyChatting","sub_path":"p2p_chatting.py","file_name":"p2p_chatting.py","file_ext":"py","file_size_in_byte":5128,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72580349385","text":"from aiogram import types\nfrom order import order\n#from order_full import order_full\n#from states import states\nfrom aiogram.types import ReplyKeyboardMarkup, ReplyKeyboardRemove, InlineKeyboardButton, InlineKeyboardMarkup, KeyboardButton\nfrom lo import bot, dp\n\nb1 = KeyboardButton('kz🇰🇿')\nb2 = KeyboardButton('ru🇷🇺')\nb3 = KeyboardButton('eng🇺🇸')\n\nmarkup_language = ReplyKeyboardMarkup().row(\n b1, b2, b3\n)\n\ninline_set = InlineKeyboardMarkup(row_width=3)\nset_diplom = InlineKeyboardButton('Дипломат🍣', callback_data='diplomat')\nset_posi = InlineKeyboardButton('Позитивчик🍣', callback_data='posi')\nset_bar = InlineKeyboardButton('Бармаглот🍣', callback_data='barmaglog')\nset_he_she = InlineKeyboardButton('Он и Она🍣', callback_data='He_She')\ninline_set.insert(set_diplom)\ninline_set.insert(set_posi)\ninline_set.insert(set_bar)\ninline_set.insert(set_he_she)\n\nText = \"Тілді таңдаңыз, пажалуиста выберите язык, please choose language\"\ntext_kz = \"сушиға🍣 тапсырыс беру үшін /sushi_kz басыңыз\"\ntext_ru = \"чтобы заказат суши🍣 нажмите кнопку /sushi_ru\"\ntext_eng = \"For order sushi🍣 press /suhi_eng\"\ntext_else = \"нажмите только кнопки\"\ntext_else_kz = \"тек қана кнопкаларды басыңыз\"\n\n@dp.message_handler(commands=['start'])\nasync def start(message: types.Message):\n await bot.send_message(message.from_user.id, text=Text, reply_markup=markup_language)\n\n@dp.message_handler(commands=['sushi_kz'])\nasync def send_message(message: types.Message):\n await bot.send_message(message.from_user.id, text='Сетті таңдаңыз', reply_markup=inline_set)\n\n@dp.message_handler(content_types=['text'])\nasync def answer(message: types.Message):\n m = message.text\n if m == \"kz🇰🇿\":\n await bot.send_message(message.from_user.id, text=text_kz)\n #await bot.send_message(chat_id='@yerekerun', text=\"админ сука\")\n elif m == \"ru🇷🇺\":\n await bot.send_message(message.from_user.id, text=text_ru)\n elif m == \"eng🇺🇸\":\n await bot.send_message(message.from_user.id, text=text_eng)\n else:\n await bot.send_message(message.from_user.id, text=text_else)\n\n@dp.callback_query_handler(text='diplomat')\n@dp.callback_query_handler(text='posi')\n@dp.callback_query_handler(text='barmaglog')\n@dp.callback_query_handler(text='He_She')\nasync def procces(call: types.CallbackQuery):\n answer_data = call.data\n if answer_data == 'diplomat':\n text = \"✳Сет для студентов❗ \\n Сет называется «𝐃𝐈𝐏𝐋𝐎𝐌𝐀𝐍𝐓 \\nПримечание: \\n-Покажите студенческий билет и получите сет со скидкой 4900тг \\nВнимание: ТОЛЬКО НА ДОСТАВКУ \\n——————————————————————\\n60шт суши и одна пицца🔺\\n🍣Филадельфия 10шт\\n🍣Эби кинг с креветкой 10шт\\n🍣Цезарь 10шт\\n🍣Сяке темпура 10шт\\n🍣Бонита 10шт\\n🍣Ёжик в тумане 10шт\\n🍕Пепперони пицца\\n🍱Цена со скидкой = 4900тг🔔\\n——————————————————————\\n Тапсырыс беру үшін /order 📲 басыңыз\"\n elif answer_data == 'posi':\n text = \"\"\n elif answer_data == 'barmaglog':\n text = \"\"\n elif answer_data == 'He_She':\n text = \"\"\n else:\n text = \"Please choose in Inline buttons\"\n await bot.send_message(call.from_user.id, text=text)\n\n\n\n","repo_name":"Yerek-Athlete/Chat_bot","sub_path":"venv/sushi.py","file_name":"sushi.py","file_ext":"py","file_size_in_byte":3723,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"35749400000","text":"import pandas as pd\nimport streamlit as st\n#import code\n\n\n@st.cache\ndef load():\n data = pd.read_csv(\"https://raw.githubusercontent.com/datasets/population/master/data/population.csv\")\n data.drop(columns = 'Country Code', inplace=True)\n data.replace('United States','US', inplace=True)\n max_year_indexes = data.groupby(['Country Name']).apply(lambda x: pd.Series({'value': x['Year'].idxmax()}))\n return data.iloc[max_year_indexes['value']].reset_index(drop=True).drop(columns=['Year']).rename(columns={'Country Name': 'country', 'Value': 'country_population'})\n \n#code.interact(local=locals())\n","repo_name":"jesperfj/streamlit-covid","sub_path":"worldpopulation.py","file_name":"worldpopulation.py","file_ext":"py","file_size_in_byte":611,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5042929042","text":"import json as j\nimport tkinter as tk\nfrom tkinter import scrolledtext\n\n\nclass EntryWithPlaceholder(tk.Entry):\n def __init__(self, master=None, placeholder=\"PLACEHOLDER\", color='grey'):\n super().__init__(master, width=20, command=None)\n\n self.placeholder = placeholder\n self.placeholder_color = color\n self.default_fg_color = self['fg']\n\n self.bind(\"
\", self.foc_in)\n self.bind(\"\", self.foc_out)\n\n self.put_placeholder()\n\n def put_placeholder(self):\n self.insert(0, self.placeholder)\n self['fg'] = self.placeholder_color\n\n def foc_in(self, *args):\n if self['fg'] == self.placeholder_color:\n self.delete('0', 'end')\n self['fg'] = self.default_fg_color\n\n def foc_out(self, *args):\n if not self.get():\n self.put_placeholder()\n\n\nk = str()\n\n\ndef uid():\n msg = tk.Tk()\n msg.title(\"Enter Unique ID\")\n msg.geometry('300x100')\n\n lbl = tk.Label(msg, text=\"UID:\", font=(\"Arial Bold\", 14))\n lbl.grid(column=0, row=0, padx=20, pady=20)\n id_ = EntryWithPlaceholder(msg, placeholder=\"Enter Unique ID\")\n id_.grid(column=1, row=0, padx=20, pady=20)\n id_.focus()\n\n def destroy():\n global k\n k = id_.get()\n msg.quit()\n msg.destroy()\n\n btn = tk.Button(msg, text=\"OK\", command=destroy)\n btn.grid(column=2, row=0, padx=10, pady=20)\n\n msg.mainloop()\n return int(k)\n\n\ndef recDisp(ls):\n ls = j.dumps(ls, indent=4)\n n_w = tk.Tk()\n n_w.title(\"Record\")\n n_w.geometry(\"500x250\")\n sc_txt = scrolledtext.ScrolledText(n_w, width=50, height=20)\n sc_txt.grid(column=0, row=3)\n sc_txt.insert(tk.INSERT, ls)\n n_w.mainloop()\n\n\nif __name__ == '__main__':\n print(uid())\n","repo_name":"rajprakharpatel/HospitalManagement","sub_path":"GUI_elements.py","file_name":"GUI_elements.py","file_ext":"py","file_size_in_byte":1765,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"36025340378","text":"import time\nimport grpc\nfrom concurrent import futures\nimport crud_pb2_grpc\n\n_ONE_DAY_IN_SECONDS = 60 * 60 * 24\n\n\nclass RemoteServer():\n server = None\n\n def __init__(self, cmd_map, grpc_port):\n RemoteServer.setup(cmd_map, grpc_port)\n\n @classmethod\n def setup(cls, cmd_map, grpc_port):\n cls.server = grpc.server(futures.ThreadPoolExecutor(max_workers=10))\n crud_pb2_grpc.add_MapServicer_to_server(cmd_map, cls.server)\n port = f'[::]:{grpc_port}'\n cls.server.add_insecure_port(port)\n\n def serve(self):\n RemoteServer.server.start()\n try:\n while True:\n time.sleep(_ONE_DAY_IN_SECONDS)\n except KeyboardInterrupt:\n RemoteServer.server.stop(0)\n","repo_name":"jchami/gbc074","sub_path":"Projeto2/grpcserver.py","file_name":"grpcserver.py","file_ext":"py","file_size_in_byte":745,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"13792853357","text":"\"\"\"\nInference module for window-based strategy.\n\"\"\"\n\nimport math\nimport os\nimport os.path as osp\nfrom itertools import product\nfrom pathlib import Path\n\nimport numpy as np\nimport torch\nimport torchvision.transforms.functional as TF\n\nimport fire\nfrom tqdm import tqdm\nfrom PIL import Image\nfrom skimage.io import imread\n\nfrom models import initialize_trainer\n\n\ndef _get_top_left_coordinates(height, width, patch_size):\n \"\"\"Calculate coordinates of top-left corners for patches.\"\"\"\n\n n_h = math.ceil(height / patch_size)\n n_w = math.ceil(width / patch_size)\n tops = np.linspace(0, height - patch_size, n_h, dtype=int)\n lefts = np.linspace(0, width - patch_size, n_w, dtype=int)\n\n return product(tops, lefts)\n\n\ndef divide_image_to_patches(img, patch_size):\n \"\"\"\n Divide a large image (mask) to patches with (possibly overlapping) tile strategy.\n\n Args:\n img: input image of shape (H, W, 3)\n patch_size: target size of patches\n\n Returns:\n patches: patches of shape (N, patch_size, patch_size, 3)\n \"\"\"\n\n assert len(img.shape) == 3 and img.shape[-1] == 3\n\n height, width, _ = img.shape\n coordinates = _get_top_left_coordinates(height, width, patch_size)\n\n patches = []\n\n for top, left in coordinates:\n patches.append(img[top:top + patch_size, left:left + patch_size])\n\n return np.array(patches).astype('uint8')\n\n\ndef combine_patches_to_image(patches, target_height, target_width):\n \"\"\"Combine patches back to a single image (mask).\n\n Args:\n patches: predicted patches of shape (N, H, W, C) or (N, H, W)\n target_height: target height of combined image\n target_width: target width of combined image\n\n Returns:\n combined: combined output of shape (H, W, C) or (H, W)\n \"\"\"\n\n counter = 0\n patch_size = patches.shape[1]\n coordinates = _get_top_left_coordinates(\n target_height, target_width, patch_size)\n\n if len(patches.shape) == 3: # channel dimension is missing\n patches = np.expand_dims(patches, -1)\n\n # The last channel is the number of overlapping patches for a given pixel,\n # used for averaging predictions from multiple windows.\n combined = np.zeros((target_height, target_width, patches.shape[-1] + 1))\n\n for top, left in coordinates:\n patch = combined[top:top + patch_size, left:left + patch_size, :-1]\n overlaps = combined[top:top + patch_size, left:left + patch_size, -1:]\n patch = (patch * overlaps + patches[counter]) / (overlaps + 1)\n combined[top:top + patch_size, left:left + patch_size, :-1] = patch\n overlaps += 1.\n counter += 1\n\n return np.squeeze(combined[..., :-1])\n\n\ndef predict(trainer, img_path, patch_size, device='cpu'):\n \"\"\"Predict on a single input image.\n\n Arguments:\n trainer: trainer for inference\n img_path: instance of `torch.utils.data.Dataset`\n patch_size: patch size when feeding into network\n device: target device\n\n Returns:\n predictions: list of model predictions of size (H, W)\n \"\"\"\n\n img = imread(img_path)\n patches = divide_image_to_patches(img, patch_size)\n predictions = []\n\n for patch in patches:\n input_ = TF.to_tensor(Image.fromarray(patch)).to(device).unsqueeze(0)\n input_, _ = trainer.preprocess(input_)\n prediction = trainer.postprocess(trainer.model(input_))\n prediction = prediction.detach().cpu().numpy()\n predictions.append(prediction[..., np.newaxis])\n\n predictions = np.concatenate(predictions)\n\n return combine_patches_to_image(predictions, img.shape[0], img.shape[1])\n\n\ndef save_predictions(predictions, img_paths, output_dir='predictions'):\n \"\"\"Save predictions to disk.\n\n Args:\n predictions: model predictions of size (N, H, W)\n img_paths: list of paths to input images\n output_dir: path to output directory\n \"\"\"\n\n print(f'\\nSaving prediction to {output_dir} ...')\n\n if not osp.exists(output_dir):\n os.mkdir(output_dir)\n\n for pred, img_path in tqdm(zip(predictions, img_paths), total=len(predictions)):\n img_name = osp.basename(img_path)\n pred = pred.astype('uint8')\n Image.fromarray(pred * 255).save(osp.join(output_dir, img_name))\n\n\ndef infer(trainer, data_dir, patch_size, output_dir=None, device='cpu'):\n \"\"\"Making inference on a directory of images with given model checkpoint.\"\"\"\n\n if output_dir is not None and not osp.exists(output_dir):\n os.mkdir(output_dir)\n\n data_dir = Path(data_dir).expanduser()\n img_paths = list((data_dir / 'images').iterdir())\n\n print(f'Predicting {len(img_paths)} images from {data_dir} ...')\n predictions = [\n predict(trainer, img_path, patch_size, device=device)\n for img_path in tqdm(img_paths)\n ]\n\n if output_dir is not None:\n save_predictions(predictions, img_paths, output_dir)\n\n return predictions\n\n\ndef main(data_dir, model_type='mild', patch_size=464, checkpoint=None,\n output_dir=None, device=None):\n\n if output_dir is None and checkpoint is not None:\n checkpoint = Path(checkpoint).expanduser()\n output_dir = checkpoint.parent.parent / 'results'\n if not output_dir.exists():\n output_dir.mkdir()\n\n device = device or ('cuda' if torch.cuda.is_available() else 'cpu')\n trainer = initialize_trainer(model_type, device=device)\n if checkpoint is not None:\n trainer.load_checkpoint(checkpoint)\n\n infer(trainer, data_dir, patch_size, output_dir, device=device)\n\n\nif __name__ == '__main__':\n fire.Fire(main)\n","repo_name":"mrcfps/WESUP","sub_path":"infer_tile.py","file_name":"infer_tile.py","file_ext":"py","file_size_in_byte":5584,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"81"}
+{"seq_id":"1375132958","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n# Se cargan módulos\nimport cv2\nimport argparse\n\n# Se configura los argumentos de la línea de comandos\np = argparse.ArgumentParser(\"Extrae una region de una imagen\")\np.add_argument(\"x1\",default=None,type=int,\n action=\"store\", help=\"Cordenada x de primera esquina\")\np.add_argument(\"y1\",default=None,type=int,\n action=\"store\", help=\"Cordenada y de primera esquina\")\np.add_argument(\"x2\",default=None,type=int,\n action=\"store\", help=\"Cordenada x de segunda esquina\")\np.add_argument(\"y2\",default=None,type=int,\n action=\"store\", help=\"Cordenada y de segunda esquina\")\np.add_argument(\"archivo\",default=None,\n action=\"store\", help=\"Nombre de archivo\")\nopts = p.parse_args()\n\n\n# Se checa que las cordenadas sean válidas\nif opts.x1>=opts.x2:\n p.error(\"Error coordenadas x1 ({0}) y x2 ({1}) incorrectas\".format(opts.x1,opts.x2))\n\nif opts.y1>=opts.y2:\n p.error(\"Error coordenadas x1 ({0}) y x2 ({1}) incorrectas\".format(opts.x1,opts.x2))\n\n# Se lee la imagen\nimg = cv2.imread(opts.archivo)\n\n# Se extraé las propiedades de la imagen\nheight, width, depth = img.shape\n\n# Se checa que las cordenadas sean válidas para la imagen\nif opts.x1>width or opts.x2>width:\n p.error(\"Cordenadas más grande que el ancho de la imagen ({0})\".format(width))\n p.error()\n\nif opts.y1>height or opts.y2>height:\n p.error(\"Cordenadas más grande que el ancho de la imagen ({0})\".format(width))\n\n# Se corta la región de interes\ncrop_img = img[opts.y1:opts.y2, opts.x1:opts.x2]\n\n# Se guarda la imagen\ncv2.imwrite(\"imagen_recortada.png\", crop_img)\n\n","repo_name":"ivanvladimir/cursoML","sub_path":"sesion4/cortandoimagen.py","file_name":"cortandoimagen.py","file_ext":"py","file_size_in_byte":1627,"program_lang":"python","lang":"es","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"37936727775","text":"from pytube import YouTube, Playlist\n\n\ndef get_title(id, type):\n if type == 'video':\n url = f'https://www.youtube.com/watch?v={id}'\n return YouTube(url).title\n elif type == 'playlist':\n url = f'https://www.youtube.com/playlist?list={id}'\n return YouTube(Playlist(url).video_urls[0]).title","repo_name":"disasstor/Youtube-Grasper-Bot","sub_path":"bot/utils/extractor_title.py","file_name":"extractor_title.py","file_ext":"py","file_size_in_byte":322,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"30995742292","text":"class Layer(object):\n \"\"\"\n This class represents a dense layer in a neural network.\n\n This layer utilises tensors (https://en.wikipedia.org/wiki/Tensor) - the construct of choice in neural networks.\n\n Attributes:\n in_size: size of the input tensor.\n out_size: size of the output tensor.\n t_in: the input tensor.\n t_out: the output tensor.\n weights: the weights tensor.\n bias: the bias tensor.\n activator: the activator function which is ran after activation.\n \"\"\"\n\n # Optional way for Python to optimize variable lookups\n __slots__ = \"t_in\", \"t_out\", \"weights\", \"bias\", \"activator\"\n\n def __init__(self, in_size, out_size, activator=lambda x: x):\n \"\"\"\n Method to initialize the neural network layer with the given in tensor size and out tensor size.\n If activator is empty, this layer will be linear.\n\n Args:\n in_size: the size of the input tensor.\n out_size: the size of the output tensor.\n activator: optional; the activator function. If not supplied, this will be a linear activator.\n \"\"\"\n\n # Set output tensor (input will be supplied during inference)\n self.t_in = None\n self.t_out = [0] * out_size\n\n # Set weight tensor\n self.weights = [0] * (in_size * out_size)\n # Set bias tensor\n self.bias = [0] * out_size\n # Set the activation method\n self.activator = activator\n\n def activate(self):\n \"\"\"\n Method to activate this layer, runs activator and sets output tensor.\n \"\"\"\n\n # Run for each neuron (output)\n out_size = len(self.t_out)\n for n in range(out_size):\n # Sum input of all input neurons (inputs), multiplied by corresponding weight\n input_sum = 0\n for i in range(len(self.t_in)):\n input_sum += self.t_in[i] * self.weights[i * out_size + n]\n input_sum = input_sum + self.bias[n]\n # Run activator on input\n self.t_out[n] = self.activator(input_sum)\n\n def read_file(self, fh):\n \"\"\"\n Method to import weights and biases from an input file.\n\n Args:\n fh: input stream from which this layer's weights and biases are read.\n \"\"\"\n\n # Import weight data\n for i in range(len(self.weights)):\n self.weights[i] = eval(fh.readline())\n # Import bias data\n for i in range(len(self.bias)):\n self.bias[i] = eval(fh.readline())\n","repo_name":"TUDSSL/BFree","sub_path":"software/applications-and-benchmarks/app-micro-nn/neural_net_src/layer.py","file_name":"layer.py","file_ext":"py","file_size_in_byte":2538,"program_lang":"python","lang":"en","doc_type":"code","stars":268,"dataset":"github-code","pt":"81"}
+{"seq_id":"12069061004","text":"src = [2, 2, 2, 7, 23, 1, 44, 44, 3, 2, 10, 7, 4, 11]\n\n\ndef count_val(lst):\n tmp_dict = {}\n for el in lst:\n if el in tmp_dict:\n tmp_dict[el] += 1\n else:\n tmp_dict[el] = 0\n return tmp_dict\n\n\n# 1 способ\ndict_src = count_val(src)\nresult = [num for num, count in dict_src.items() if count == 0]\nprint(result)\n\n# 2 способ\nsrc = [2, 2, 2, 7, 23, 1, 44, 44, 3, 2, 10, 7, 4, 11]\nprint([x for x in src if src.count(x) == 1])\n\n","repo_name":"arseniy-smirnovvv/lessons_geekbrains","sub_path":"arseniy_smirnov_dz_5/task_5_5.py","file_name":"task_5_5.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"33487927844","text":"'''\nRemove all elements from a linked list of integers that have value val.\n\nExample:\n\nInput: 1->2->6->3->4->5->6, val = 6\nOutput: 1->2->3->4->5\n'''\n# Definition for singly-linked list.\n# class ListNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution(object):\n def removeElements(self, head, val):\n \"\"\"\n :type head: ListNode\n :type val: int\n :rtype: ListNode\n \"\"\"\n first = ListNode(None)\n first.next = head\n pre = first\n while head:\n if head.val == val:\n pre.next = head.next\n head.next = None\n head = pre.next\n else:\n pre = head\n head = head.next\n return first.next\nclass Solution(object):\n def removeElements(self, head, val):\n \"\"\"\n :type head: ListNode\n :type val: int\n :rtype: ListNode\n \"\"\"\n dummpy = ListNode(None)\n dummpy.next = head\n next = dummpy\n while next != None and next.next != None:\n if next.next.val == val:\n next.next = next.next.next\n else:\n next = next.next\n return dummpy.next\n","repo_name":"XiongQiuQiu/leetcode-slove","sub_path":"Algorithms/203-Remove-Linked-List-Elements.py","file_name":"203-Remove-Linked-List-Elements.py","file_ext":"py","file_size_in_byte":1249,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"19398119419","text":"\"\"\"\nClasse mère des classe comportant des modèles\n\nElle gere la sauvegarde automatique des modèle\nAuthor : bsanchez@starclay.fr\ndate : 06/08/2020\n\"\"\"\n\nfrom abc import ABC\nfrom abc import ABCMeta\nfrom abc import abstractmethod\nfrom abc import ABC\nimport tempfile\nimport pickle\nimport logging\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.keras.callbacks import EarlyStopping\nfrom tensorflow.keras.callbacks import ModelCheckpoint\n\nimport os, sys\nfrom .similarity_model import SimilarityModel\n\n\nclass TrainingModel(ABC):\n \n def __init__(self, name, path_run,\n nomenclature_distance, \n load_path=None, **kwargs):\n \"\"\"\n Superclasse pour les modèles à entrainer\n\n :param name: nom du modèle\n :param path_run: chemin de sauvegarde\n :param nomenclature_distance: NomenclatureDistance à utiliser\n :param load_path: chemin à charger\n :param kwargs: dict d'arguments pour la construction du modèle (passé à build_model, puis define_layers)\n \"\"\"\n self.model_name = name\n self.path_run = path_run\n self.logger = logging.getLogger(name)\n \n self.nomenclature_distance = nomenclature_distance\n if load_path is not None:\n self.model = self.load_model(load_path)\n else:\n self.model = self.build_model(**kwargs)\n self.logger.info(self.model.summary())\n \n @abstractmethod\n def define_layers(self, **kwargs):\n \"\"\"\n Methode qui définit les couches\n \n :params kwargs: dict d'arguments pour la construction des couches\n \n :returns: input (tf.Tensor ou list de tf.tensor), \n output (tf.Tensor ou list de tf.tensor)\n \"\"\"\n raise NotImplementedError()\n \n def build_model(self, **kwargs):\n \"\"\"\n Method qui crée les couches via define_layers, puis encapsule dans un SimilarityModel\n \n :params kwargs: dict d'arguments pour la construction des couches\n :returns: keras.Model\n \"\"\"\n inputs, output = self.define_layers(**kwargs)\n\n model = SimilarityModel(inputs=inputs, \n outputs=output, \n nomenclature_distance=self.nomenclature_distance)\n return model\n \n def train_model(self, training_pairs_batcher, validation_pairs_batcher, nb_epochs):\n \"\"\"\n Entraine le modèle:\n - compile le modèle\n - ajoute des callbacks (ModelCheckpoint, EarlyStopping)\n - lance le training\n\n :param training_pairs_batcher: AnchorPositivePairsBatch contenant les données de training\n :param validation_pairs_batcher: AnchorPositivePairsBatch contenant les données de validation\n :param nb_epochs: nb d'époques de training\n :returns: training History\n \"\"\"\n self.model.compile(\n optimizer=keras.optimizers.Adam(learning_rate=1e-3),\n loss=keras.losses.MeanSquaredError(), \n# run_eagerly=tf.executing_eagerly()\n )\n self.logger.info(f'Running eagerly : {tf.executing_eagerly()}')\n \n train_save_path = os.path.join(self.path_run, 'train_weights')\n mc = ModelCheckpoint(os.path.join(train_save_path, 'best_model_{epoch:02d}-{val_loss:.5f}'), \n save_best_only=True, save_weights_only=False, \n mode='auto', period=1, verbose=1)\n es = EarlyStopping(monitor='val_loss', mode='min', verbose=1, patience=10)\n if not os.path.exists(train_save_path):\n os.makedirs(train_save_path)\n\n self.logger.info('training')\n history = self.model.fit(training_pairs_batcher, \n epochs=nb_epochs,\n validation_data=validation_pairs_batcher,\n callbacks = [es, mc],\n verbose = True)\n return history\n\n def run_model_siamese(self, testing_pairs_batcher):\n return\n \n def run_model_single_side(self, formatted_data):\n \"\"\"\n prediction pour formatted_data (formatté via AnchorPositivePairsBatch.format_input)\n\n :param formatted_data: tf.tensor\n :returns: tf.tensor\n \"\"\"\n return self.model(formatted_data, training=False)\n \n def load_model(self,path):\n \"\"\"\n Charge un modele\n IMPORTANT : le modèle chargé est de type keras.Model, pas SimilarityModel\n \n :params path: chemin du model à être charger\n :returns: void\n \"\"\"\n return keras.models.load_model(path)\n","repo_name":"etalab-ia/ami-ia-insee-aiee2","sub_path":"nomenclatures/training_classes/training_model.py","file_name":"training_model.py","file_ext":"py","file_size_in_byte":4725,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41661162069","text":"from transformers import DetrFeatureExtractor, DetrForSegmentation\nimport numpy as np\nimport torch\nimport cv2\nfrom matplotlib import pyplot as plt\nimport itertools\nimport seaborn as sns\nfrom copy import deepcopy\nfrom clustering import filter_clusters\nimport motionnet\nfrom diffusers import StableDiffusionInpaintPipeline\nfrom PIL import Image\nimport os\n#from clip_interrogator import Config, Interrogator\n\n\nparts = ['body', 'half', 'face']\n\nfeature_extractor = DetrFeatureExtractor.from_pretrained('facebook/detr-resnet-50-panoptic')\nmodel = DetrForSegmentation.from_pretrained('facebook/detr-resnet-50-panoptic')\n\n\n# point in rectangle\ndef point_in_rect(x, y, rect):\n return rect[0] <= x <= rect[2] and rect[1] <= y <= rect[3]\n\ndef get_cluster(x, y, l):\n if x < 0 or y < 0 or x >= l.shape[0] or y >= l.shape[1]:\n return []\n\n r = []\n if l[x,y] == 1:\n r = [(x, y)]\n l[x, y] = 0\n r.extend(get_cluster(x-1, y, l))\n r.extend(get_cluster(x+1, y, l))\n r.extend(get_cluster(x, y-1, l))\n r.extend(get_cluster(x, y+1, l))\n\n return r\n\n\ndef get_clusters(l):\n clusters = []\n while True:\n x, y = np.where(l == 1)\n if len(x) == 0:\n break\n clusters.append(get_cluster(x[0], y[0], l))\n return clusters\n\n\ndef predict_animal_mask(im,\n gr_slider_confidence):\n image = Image.fromarray(im) # im: numpy array 3d: 480, 640, 3: to PIL Image\n image = image.resize((200,200)) # PIL image # could I upsample output instead? better?\n #result_image = np.array(image.convert('RGB'))\n result_image = im\n\n # encoding is a dict with pixel_values and pixel_mask\n encoding = feature_extractor(images=image, return_tensors=\"pt\") #pt=Pytorch, tf=TensorFlow\n outputs = model(**encoding) # odict with keys: ['logits', 'pred_boxes', 'pred_masks', 'last_hidden_state', 'encoder_last_hidden_state']\n logits = outputs.logits # torch.Size([1, 100, 251]); class logits? but why 251?\n bboxes = outputs.pred_boxes\n masks = outputs.pred_masks # torch.Size([1, 100, 200, 200]); mask logits? for every pixel, score in each of the 100 classes? there is a mask per class\n\n # keep only the masks with high confidence?--------------------------------\n # compute the prob per mask (i.e., class), excluding the \"no-object\" class (the last one)\n prob_per_query = outputs.logits.softmax(-1)[..., :-1].max(-1)[0] # why logits last dim 251?\n # threshold the confidence\n keep = prob_per_query > gr_slider_confidence/100.0\n\n # postprocess the mask (numpy arrays)\n label_per_pixel = torch.argmax(masks[keep].squeeze(),dim=0).detach().numpy() # from the masks per class, select the highest per pixel\n\n processed_sizes = torch.as_tensor(encoding['pixel_values'].shape[-2:]).unsqueeze(0)\n result = feature_extractor.post_process_panoptic(outputs, processed_sizes)[0]\n\n # We extract the segments info and the panoptic result from DETR's prediction\n segments_info = deepcopy(result[\"segments_info\"])\n\n result = []\n category = 1 # 1 is the category id for the \"person\" class\n #color_mask = np.zeros(image.size+(3,))\n color_mask = np.zeros(result_image.shape)\n print('colormask', color_mask.shape) # colormask (200, 200, 3)\n palette = itertools.cycle(sns.color_palette())\n keypoints = motionnet.get_keypoints(result_image)\n\n\n w, h, c = result_image.shape\n noses = np.array([(k[0][1]*w, k[0][0]*h, k[0][2], i) for i, k in enumerate(keypoints)])\n noses = noses[noses[:, 2] > 0.1]\n\n for lbl, cat in zip(np.unique(label_per_pixel), segments_info): #enumerate(palette()):\n if cat['category_id'] == category:\n mask = filter_clusters(label_per_pixel == lbl)\n mask = cv2.resize(mask, dsize=(result_image.shape[1], result_image.shape[0]), interpolation=cv2.INTER_LINEAR)\n y, x = np.where(mask != 0)\n nose = None\n rect = (x.min(), y.min(), x.max(), y.max())\n for key in keypoints:\n if point_in_rect(key[0][1] * h, key[0][0] * w, rect):\n nose = key\n break\n\n if nose is not None:\n cv2.rectangle(color_mask, (int(np.min(x)), int(np.min(y))), (int(np.max(x)), int(np.max(y))), (255,0,0), 1)\n color_mask[mask == 1, :] = np.asarray(next(palette))*255 #color\n result.append((mask, nose, (x.min(), y.min(), x.max(), y.max())))\n\n\n # Show image + mask\n pred_img = np.array(result_image)*0.25 + color_mask*0.75\n\n\n motionnet.loop_through_people(pred_img, keypoints)\n\n pred_img = pred_img.astype(np.uint8)\n\n return pred_img, result\n\n\n\ndef possible_parts(image, detection):\n result = []\n mask, keypoints, rect = detection\n img_w, img_h, _ = image.shape\n x_min, y_min, x_max, y_max = rect\n\n if keypoints[motionnet.LABELS['leftShoulder']][2] > motionnet.THRESHOLD and keypoints[motionnet.LABELS['rightShoulder']][2]:\n leftShoulder_y, leftShoulder_x, _ = keypoints[motionnet.LABELS['leftShoulder']]\n rightShoulder_y, rightShoulder_x, _ = keypoints[motionnet.LABELS['rightShoulder']]\n\n base_line_shoulder = keypoints[motionnet.LABELS['leftShoulder']][0]\n result.append(['face', (rightShoulder_x, base_line_shoulder, leftShoulder_x, y_min)])\n else:\n nose = keypoints[motionnet.LABELS['nose']]\n leftEar = keypoints[motionnet.LABELS['leftEar']]\n rightEar = keypoints[motionnet.LABELS['rightEar']]\n leftEye = keypoints[motionnet.LABELS['leftEye']]\n rightEye = keypoints[motionnet.LABELS['rightEye']]\n\n width = abs(leftEye[1] - rightEye[1])\n\n result.append(['face', (rightEar[1] - width, nose[0] + 3 * width, leftEar[1] + width, rightEye[0] - width)])\n\n if keypoints[motionnet.LABELS['leftHip']][2] > motionnet.THRESHOLD and keypoints[motionnet.LABELS['rightHip']][2]:\n leftHip_y, leftHip_x, _ = keypoints[motionnet.LABELS['leftHip']]\n rightHip_y, rightHip_x, _ = keypoints[motionnet.LABELS['rightHip']]\n base_line_hip = keypoints[motionnet.LABELS['leftHip']][1]\n\n hip_size = leftHip_x - rightHip_x\n\n result.append(['half', (rightHip_x - hip_size, base_line_hip, leftHip_x + hip_size, y_min)])\n\n if keypoints[motionnet.LABELS['leftAnkle']][2] > motionnet.THRESHOLD and keypoints[motionnet.LABELS['rightAnkle']][2]:\n leftAnkle_y, leftAnkle_x, _ = keypoints[motionnet.LABELS['leftAnkle']]\n rightAnkle_y, rightAnkle_x, _ = keypoints[motionnet.LABELS['rightAnkle']]\n\n leftHip_y, leftHip_x, _ = keypoints[motionnet.LABELS['leftHip']]\n rightHip_y, rightHip_x, _ = keypoints[motionnet.LABELS['rightHip']]\n hip_size = leftHip_x - rightHip_x\n\n leftEye = keypoints[motionnet.LABELS['leftEye']]\n rightEye = keypoints[motionnet.LABELS['rightEye']]\n\n width = abs(leftEye[1] - rightEye[1])\n\n result.append(['body', (rightAnkle_x - hip_size, leftAnkle_y + width, leftAnkle_x + hip_size, y_min)])\n\n result_filter = []\n for part in result:\n rect = part[1]\n x_min, y_min, x_max, y_max = rect\n if x_min < 0:\n x_min = 0\n if y_min < 0:\n y_min = 0\n if x_max > image.shape[1]:\n x_max = image.shape[1]\n if y_max > image.shape[0]:\n y_max = image.shape[0]\n\n result_filter.append([part[0], (int(x_min * img_w), int(y_min * img_h), int(x_max * img_w), int(y_max))])\n\n return result_filter\n\n\ndef process_image(filename, output_filename=None):\n image = cv2.imread(filename)[:,:,::-1]\n org_image = image.copy()\n\n _, detection = predict_animal_mask(image, 85)\n\n detection_max = None\n area_max = 0\n for part in detection:\n mask, keypoints, rect = part\n x_min, y_min, x_max, y_max = rect\n area = (x_max - x_min) * (y_max - y_min)\n if area > area_max:\n area_max = area\n detection_max = part\n\n mask, nose, rect = detection_max\n plt.imshow(mask)\n plt.show()\n\n x1, y1, x2, y2 = rect\n w = abs(x2 - x1)\n h = abs(y2 - y1)\n sub = org_image[y1:y2, x1:x2]\n scale_percent = 512 / max([w, h])\n\n width = int(sub.shape[1] * scale_percent)\n height = int(sub.shape[0] * scale_percent)\n dim = (width, height)\n print(dim)\n resized = cv2.resize(sub, dim, interpolation=cv2.INTER_AREA)\n print('resized.shape', resized.shape)\n # put resized image in center of final_image\n yoff = round((512-height)/2)\n xoff = round((512-width)/2)\n\n final_image = np.zeros((512, 512, 3), dtype=np.uint8)\n final_image[yoff:yoff+height, xoff:xoff+width, :] = resized\n mask_image = np.zeros((512, 512, 3), dtype=np.uint8)\n mask_image.fill(255)\n mask_image[yoff:yoff+height, xoff:xoff+width, :] = 0\n\n if output_filename is not None:\n rr = Image.fromarray(final_image)\n rr.save(output_filename)\n\n '''\n pil_image = Image.fromarray(np.uint8(org_image)).convert('RGB')\n ci = Interrogator(Config(clip_model_name=\"ViT-L-14/openai\"))\n ci.config.blip_num_beams = 64\n ci.config.chunk_size = 2048\n ci.config.flavor_intermediate_count = 2048\n\n prompt = ci.interrogate(pil_image)\n\n # python load auth_token from environment variable\n # os.environ\n \n auth_token = os.environ['HF_AUTH_TOKEN']\n\n device = \"cuda\" if torch.cuda.is_available() else \"cpu\"\n\n pipe = StableDiffusionInpaintPipeline.from_pretrained(\n \"runwayml/stable-diffusion-inpainting\",\n revision=\"fp16\",\n torch_dtype=torch.float16,\n use_auth_token=auth_token\n ).to(device)\n\n output = pipe(prompt=prompt, image=Image.fromarray(final_image), mask_image=Image.fromarray(mask_image)).images[0]\n result = np.array(output)\n result[yoff:yoff+height, xoff:xoff+width, :] = resized\n\n if output_filename is not None:\n result = Image.fromarray(result)\n result.save(output_filename)\n '''\n\n return rr\n\n\ndef test_image(image_path, gr_slider_confidence):\n gr_image_input = cv2.imread(image_path)\n print(gr_image_input.shape)\n\n pred_img, result = predict_animal_mask(gr_image_input, gr_slider_confidence)\n\n\n detection_max = None\n area_max = 0\n for part in result:\n mask, keypoints, rect = part\n x_min, y_min, x_max, y_max = rect\n area = (x_max - x_min) * (y_max - y_min)\n if area > area_max:\n area_max = area\n detection_max = part\n\n mask, nose, rect = detection_max\n plt.imshow(mask)\n plt.show()\n\n\n\n\ntest_image('example_image_3.jpg', 40)\n#test_image('cuerpo1.jpg', 85)\n\n#image_format.process_image('example_image_1.jpg', 'o_2.png')\n#image_format.process_image('example_image_2.jpg', 'o_3.png')\n#image_format.process_image('example_image_3.jpg', 'o_4.png')\n#image_format.process_image('example_image_2.jpeg', 'o_5.png')\n#image_format.process_image('face1.jpg', 'o_6.png')\n#image_format.process_image('face2.jpg', 'o_7.png')\n#image_format.process_image('example_image_3.jpg', 'o_4.png')","repo_name":"carlgira/automatic-image-processing","sub_path":"image_format.py","file_name":"image_format.py","file_ext":"py","file_size_in_byte":10992,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23268351541","text":"import sending\nimport queries_to_bd\nimport keyboards\nimport telebot\nimport common_methods\n\n#основной метод проверки входящих сообщений\ndef main(bot, message):\n\n text_data = None\n audio_data = None\n document_data = None\n current_result_text = None\n current_reply_markup = telebot.types.InlineKeyboardMarkup()\n current_parsemod = None\n photo_data = None\n poll_data = None\n invoice_data = None\n sticker_data = None\n\n #проверяем, задал ли бот вопрос пользователю, на который он обязательно должен ответить текстом в чате\n flg_need_response = queries_to_bd.check_need_response_flg(message.chat.id)\n\n #если бот ждет ответа от пользователя в виде текстового сообщения, то\n if flg_need_response == 1:\n\n #проверяем о чем вообще шла речь, для этого заберем ID последней нажатой кнопки и заодно msg_id, в рамках которого будем редачить сообщение\n what_is_current_context = queries_to_bd.get_last_pressed_button(message.chat.id)\n\n #проверка, что пользователь прислал текст\n if message.content_type == 'text':\n\n #если последняя нажатая кнопка относится к меню с ID = 4, то это напоминалки. В напоминалках только в 1 месте требуется, чтобы пользователь дал ответ текстом - это при вводе информации, что же нужно напомнить\n if what_is_current_context.startswith(\"4\"):\n\n #создает в таблице запись о напоминалке\n queries_to_bd.create_new_notification(message.chat.id, message.text)\n\n #забирает ID созданной записи (напоминалки)\n notification_id = queries_to_bd.get_last_notification_id(message.chat.id)\n\n #забираем клавиатуру для редактирования напоминалки\n (current_result_text, current_reply_markup) = keyboards.notification_edit(notification_id)\n\n #если последняя нажатая кнопка относится к меню с ID = 5, то это шифрование/дешифрование\n elif what_is_current_context.startswith(\"5\"):\n\n #вытаскиваем тип операции\n operation_type = (what_is_current_context.split('/'))[-1]\n\n #вытаскиваем язык\n lang_code = (what_is_current_context.split('/'))[-2]\n\n #вытаскиваем ключ\n key = (what_is_current_context.split('/'))[-3]\n\n #отправляем на шифровку/дешифровку\n (current_result_text, current_reply_markup) = keyboards.crypting_result(operation_type, lang_code, key, message.text)\n\n current_parsemod = 'MarkdownV2'\n\n #текстовые данные ожидаются от пользователя по ветке 8 только в случае преобразования текста в войс\n elif what_is_current_context in [\"8/1/1\", \"8/2/1\"]:\n\n #вытаскиваем язык. 1 - русский, 2 - английский\n lang_code = (what_is_current_context.split('/'))[-2]\n\n #получаем путь, по которому лежит аудиофайл\n audio_data = common_methods.convert_text_to_speech(str(message.chat.id), message.text, lang_code)\n\n #получает последнее меню по ветке распознавания войса в текст\n (current_result_text, current_reply_markup) = keyboards.text_speech_result_voice(message.text, lang_code)\n\n #текстовые данные ожидаются от пользователя по ветке 9/2 только в случае получения названия города для погоды\n elif what_is_current_context == \"9/2\":\n\n #получает последнее меню по ветке распознавания войса в текст\n (current_result_text, current_reply_markup) = keyboards.weather_last_menu(message.text)\n\n #текстовые данные ожидаются от пользователя по ветке 9/3 только в случае получения текста, который будем помещен в QR-код\n elif what_is_current_context == \"9/3\":\n\n #получает последнее меню по ветке распознавания войса в текст\n (current_result_text, current_reply_markup) = keyboards.qr_code_result(message.text)\n\n document_data = common_methods.create_qr_code(message.text, str(message.chat.id))\n\n #текстовые данные ожидаются от пользователя по ветке 9/4 только в случае получения от пользователя тегов, по которым будем искать пикчу на реакторе\n elif what_is_current_context == \"9/4\":\n\n #получает последнее меню по ветке получения пикчи по тегу\n (current_result_text, current_reply_markup) = keyboards.get_pic_by_teg_result()\n\n #получает адрес изображения, спойлер и подпись к нему\n photo_data = common_methods.get_pic_by_teg(message.text)\n\n #если тип контента не подходит ни к одному из вариантов, которые ждет бот - сообщаем юзеру об этом\n else:\n current_result_text = 'Ты прислал не тот тип контента который нужен. Поздравляю тебя, начинай заново\\n/menu'\n\n #проверка, что пользователь прислал войс\n elif message.content_type == 'voice':\n\n #если последняя нажатая кнопка относится к меню с ID = 8, то это это преобразование войса в текст. Либо на русском языке, либо на английском другого не дано\n if what_is_current_context in [\"8/1/2\", \"8/2/2\"]:\n\n #вытаскиваем язык. 1 - русский, 2 - английский\n lang_code = (what_is_current_context.split('/'))[-2]\n\n #получаем инфо о войсе\n file_info = bot.get_file(message.voice.file_id)\n\n #скачиваем войс\n downloaded_file = bot.download_file(file_info.file_path)\n\n file_path = 'assets/temp/convert_speech_to_text/' + str(message.chat.id) +'.ogg'\n\n #записываем войс на диск\n with open(file_path, 'wb') as new_file:\n new_file.write(downloaded_file)\n\n #отправляем на преобразование в текст и получаем результат\n text = common_methods.convert_speech_to_text(lang_code, file_path)\n\n #получает последнее меню по ветке распознавания войса в текст\n (current_result_text, current_reply_markup) = keyboards.text_speech_result_text(text, lang_code)\n\n #если тип контента не подходит ни к одному из вариантов, которые ждет бот - сообщаем юзеру об этом\n else:\n current_result_text = 'Ты прислал не тот тип контента который нужен. Поздравляю тебя умник, начинай заново\\n/menu'\n\n #в противном случае, бот не ждет ответа от пользователя в виде сообщения и поэтому отдает по умолчанию основную клавиатуру\n else:\n\n #получает данные основной клавиатуры\n (current_result_text, current_reply_markup) = keyboards.main_menu(message.chat.id)\n\n\n #собираем текстовые данные\n text_data = (current_result_text, current_reply_markup, current_parsemod, 0, 1)\n\n #отправляем все в единый метод отправки\n sending.main(bot, message.chat.id, text_data, photo_data, poll_data, audio_data, invoice_data, sticker_data, document_data)","repo_name":"Nevillested/Ararararagi_bot","sub_path":"simple_message_cases.py","file_name":"simple_message_cases.py","file_ext":"py","file_size_in_byte":9187,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4192555394","text":"from selenium import webdriver\nfrom selenium.webdriver import ActionChains\nfrom selenium.webdriver.chrome.service import Service\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom webdriver_manager.chrome import ChromeDriverManager\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom time import sleep\n\ns = Service(ChromeDriverManager().install())\ndriver = webdriver.Chrome(service=s)\n\ndriver.maximize_window()\n\ndriver.get('https://emag.ro')\nsleep(2)\ndriver.find_element(By.XPATH, '//button[text()=\"Accept\"]').click()\nsleep(2)\ndriver.find_element(By.XPATH, '(//i[@class=\"em em-close\"]/parent::button)[3]').click()\n# driver.find_element(By.XPATH, f'//button[@class=\"js-dismiss-login-notice-btn dismiss-btn btn btn-link pad-sep-none pad-hrz-none\"]').click()\nsleep(1)\n\n# scriu in campul de cautare \"biscuiti\"\ndriver.find_element(By.ID, \"searchboxTrigger\").send_keys('biscuiti')\nsleep(2)\n# fac click pe randul doi\ndriver.find_element(By.XPATH, f'(//a[@class=\"searchbox-suggestion-result searchbox-active-item\"])[2]').click()\nsleep(2)\n# driver.find_element(By.XPATH, f'(//i[@class=\"em em-close\"]/parent::button)[3]').click()\n\n\n# fac click pe inima de la produs\n\nbiscuit1 = driver.find_element(By.XPATH, f'//a[contains(text(), \"Biscuiti cu ciocolata si fulgi de cocos Bounty Cookies, 180g\")]/parent::div/parent::div/parent::div/parent::div//i[@class=\"em em-fav em-fav-bold\"]')\nbiscuit1.click()\n\nsleep(2)\nbiscuit2 = driver.find_element(By.XPATH, f'//a[contains(text(), \"2 x Biscuiti cu unt Leibniz, 200 gr.\")]/parent::div/parent::div/parent::div//i[@class=\"em em-fav em-fav-bold\"]')\nbiscuit2.click()\nsleep(2)\ndriver.find_element(By.ID, \"my_wishlist\").click()\nsleep(2)\ndriver.find_element(By.XPATH, '//span[contains(text(), \"Biscuiti cu ciocolata si fulgi de cocos Bounty Cookies, 180g\" )]/parent::a/parent::h2/parent::div/parent::div//span[contains(text(), \"Sterge\")]').click()\nsleep(2)\n# driver.find_element(By.XPATH, f\"//a[contains(text(), \"Sampon L'Oreal Paris Elseve Dream Long reparator pentru par lung, deteriorat, 400 ml\")]/parent::div/parent::div/parent::div/parent::div//i[@class=\"em em-fav em-fav-bold\"]\")\n# (//span[contains(text(), \"Sampon de par\")])[1]","repo_name":"TeodoraBirle03/Emag_Project","sub_path":"test_scripts/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":2231,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"7505632199","text":"from .base_attack import BaseAttack\nfrom .genetic import Genetic\nimport torch\nimport dgl\nimport pandas as pd\nimport numpy as np\nfrom bayesopt.bayesopt.predictors import GPWL, BayesianLinearRegression, NullSurrogate\nfrom .utils import correct_predictions, random_sample_flip, random_sample_rewire_swap, population_graphs, extrapolate_breakeven\nfrom copy import deepcopy\n\n\nclass BayesOptAttack(BaseAttack):\n def __init__(self, classifier: torch.nn.Module, loss_fn: torch.nn.Module,\n batch_size: int = 1, n_init: int = 10,\n edit_per_stage=None,\n surrogate: str = 'bayeslinregress',\n mode: str = 'flip',\n target_class: int = None,\n surrogate_settings: dict = None,\n acq_settings: dict = None,\n verbose: bool = True,\n terminate_after_n_fail: int = None,\n n_hop_constraint: int = None,\n preserve_disconnected_components: bool = False,):\n \"\"\"\n Attacking classifier via Bayesian optimisation with GP/Bayesian Linear regression surrogate with Weisfeiler-\n Lehman kernels.\n :param classifier: see BaseAttack\n :param loss_fn: see BaseAttack\n :param batch_size: the number of possible adversarial samples to propose at each BO iteration. Larger batch\n size will lead to faster performance, but correspondingly the performance might decrease\n :param edit_per_stage: int or float. the number of edits amortised to each stage. A smaller edit_per_stage leads\n to more stages which is more greedy, a larger edit_per_stage is less greedy but leads to a larger search\n space.\n :param surrogate: the choice of surrogate.\n :param n_init: the number of initial perturbations to be sampled randomly from the search space\n :param mode: str: 'flip', 'add', 'remove' or 'rewire': allowed edit operations on the edges.\n :param surrogate_settings: dict: any parameters to be passed to the surrogates. See bayesopt/gp_predictor.py\n :param acq_settings: dict: any parameters to be passed to the acquisition function.\n :param verbose: whether to enable diagnostic information.\n :param terminate_after_n_fail: the tolerance when the BO agent fails to push the attack loss. If this is not None (\n a positive int), after this number of successive failures in increasing attack loss the attack will be aborted.\n :param n_hop_constraint: int. If not None (a positive int), and edge perturbation (either rewire or flip) must\n be constrained within the n_hop distance of the first node.\n\n \"\"\"\n super().__init__(classifier, loss_fn)\n self.target_class = target_class\n if acq_settings is None:\n acq_settings = {}\n if 'acq_type' not in acq_settings.keys(): acq_settings['acq_type'] = 'ei'\n if 'acq_optimiser' not in acq_settings.keys(): acq_settings['acq_optimiser'] = 'mutation'\n if 'acq_max_step' not in acq_settings.keys(): acq_settings['acq_max_step'] = 400\n if 'random_frac' not in acq_settings.keys(): acq_settings['random_frac'] = 0.5\n self.acq_settings = acq_settings\n self.batch_size = batch_size\n self.n_init = n_init\n self.edit_per_stage = edit_per_stage\n if surrogate_settings is None:\n surrogate_settings = {}\n\n if surrogate == 'gpwl': self.surrogate = GPWL(**surrogate_settings)\n elif surrogate == 'bayeslinregress': self.surrogate = BayesianLinearRegression(**surrogate_settings)\n elif surrogate == 'null': self.surrogate = NullSurrogate()\n else: raise ValueError(f'Unrecognised surrogate choice {surrogate}')\n\n self.verbose = verbose\n assert mode in ['flip', 'add', 'remove', 'rewire'], f'mode {mode} is not recognised!'\n self.mode = mode\n # save a record of previous query history\n self.query_history = []\n self.loss_history = []\n self.terminate_after_n_fail = terminate_after_n_fail if terminate_after_n_fail is not None and terminate_after_n_fail > 0 else None\n self.n_hop_constraint = n_hop_constraint if n_hop_constraint is not None and n_hop_constraint > 0 else None\n self.preserve_disconnected_components = preserve_disconnected_components\n\n def attack(self, graph: dgl.DGLGraph, label: torch.tensor, budget, max_queries: int):\n \"\"\"\n The main attack loop.\n - For BO, at each iteration, we only modify one edge. If we have budget > 1, we use a greedy approach to\n partition the total max_queries into int(max_queries/budget) stages. At each stage, we attack on the\n *base graph*: in the first stage, it is the original graph (i.e. graph passed as an argument here);\n in the subsequent stages, it is the best perturbed graph of the previous stage that led to the largest\n classifier loss.\n - The optimisation terminates once it detects a successful attack.\n\n For the rest, see documentation for Genetic and BaseAttack\n \"\"\"\n if isinstance(budget, float):\n assert 0 < budget < 1., f'if a float is supplied, this number must be within 0 and 1 but got {budget}'\n budget = np.round(budget * graph.num_edges()).astype(np.int)\n if isinstance(self.edit_per_stage, float):\n self.edit_per_stage = np.round(self.edit_per_stage * graph.num_edges()).astype(np.int)\n stages, edits_per_stage = self.get_stage_statistics(max_queries, budget)\n if self.verbose:\n print(f'Total number of {max_queries} of queries is divided into {stages}')\n print(f'Edits per stage is {edits_per_stage}')\n self.query_history = []\n self.loss_history = []\n dfs = []\n self.committed_edits = []\n base_graph = graph\n i = 0\n adv_example = None\n is_edge_weighted = 'weight' in graph.edata.keys()\n\n best_loss = -np.inf\n n_fail = 0\n while i < max_queries:\n curr_stage = np.digitize(i, stages) - 1\n prev_stage = np.digitize(max(0, i - self.batch_size), stages) - 1\n edit_allowed_this_stage = edits_per_stage[curr_stage]\n if curr_stage != prev_stage or i == 0:\n if i > 0:\n best_idx = torch.argmax(self.surrogate.y)\n base_graph = deepcopy(self.surrogate.X[best_idx])\n # update the list of prohibited edges\n if len(self.query_history) > 0:\n self.committed_edits += self.query_history[-self.surrogate.y.shape[0] + int(best_idx)]\n\n if self.verbose:\n print(f'Entering Stage {curr_stage}. ')\n print(f'Committed edge edits={self.committed_edits}')\n # sample randomly at the start of each stage\n n_init = min(self.n_init, stages[curr_stage + 1] - stages[curr_stage])\n if self.mode == 'rewire':\n samples = [random_sample_rewire_swap(base_graph, edit_allowed_this_stage, rewire_only=not is_edge_weighted, n_hop=self.n_hop_constraint,\n preserve_disconnected_components=self.preserve_disconnected_components\n ) for _\n in range(n_init)]\n else:\n samples = [\n random_sample_flip(base_graph, edit_allowed_this_stage, remove_edge_only=self.mode == 'remove',\n add_edge_only=self.mode == 'add', n_hop=self.n_hop_constraint,\n committed_edges=self.committed_edits,\n preserve_disconnected_components=self.preserve_disconnected_components,)\n for _ in range(n_init)]\n if not len(samples):\n print('Patience reached. Terminating the current run')\n break\n\n perturbed_graphs = population_graphs(base_graph, samples, self.mode)\n self.query_history += samples\n i += n_init\n else:\n perturbed_graphs = self.suggest(base_graph, edit_allowed_this_stage, )\n i += self.batch_size\n\n with torch.no_grad():\n try:\n preds = self.classifier(dgl.batch(perturbed_graphs))\n except:\n preds = torch.cat([self.classifier(g) for g in perturbed_graphs])\n if preds.ndimension() == 1:\n preds.reshape(-1, 1)\n\n # dgl.batch and dgl.unbatch create lots of problems. use this as a fallback option\n # see reference in github issue:\n # https://github.com/dmlc/dgl/issues/2409\n if preds.shape[0] != len(perturbed_graphs):\n preds = self.classifier(perturbed_graphs)\n\n if len(perturbed_graphs) == 1 and preds.shape[1] == 1:\n labels = label[0].reshape(1)\n else:\n labels = torch.repeat_interleave(label, len(perturbed_graphs))\n losses = self.loss_fn(preds, labels, reduction='none')\n if losses.ndimension() == 0:\n losses = losses.reshape(1)\n self.loss_history += losses.detach().numpy().tolist()\n\n if self.verbose:\n print(f'Iteration {i}. Loss: {losses.detach().numpy()}.')\n\n dfs.append(self.construct_dataframe(losses, preds, label.squeeze(), i + 1))\n\n if len(self.loss_history) > 200 and extrapolate_breakeven(self.loss_history) > 1e5:\n print(f'Predicted breakeven point {extrapolate_breakeven(self.loss_history)} and run terminated')\n break\n\n if (self.target_class is None and np.sum(correct_predictions(preds.numpy(), labels.numpy())) < len(perturbed_graphs)) \\\n or (self.target_class is not None and (np.argmax(preds.numpy(), axis=1) == self.target_class).any()):\n print('Attack succeeded!')\n if self.target_class is None:\n comps = correct_predictions(preds.numpy(), labels.numpy())\n for i, comp in enumerate(comps):\n if not comp:\n adv_example = perturbed_graphs[i]\n break\n else:\n for i, pred in enumerate(preds):\n if np.argmax(pred.numpy()) == self.target_class:\n adv_example = perturbed_graphs[i]\n break\n break\n reset_surrogate = False\n self.observe(perturbed_graphs, losses, reset_surrogate=reset_surrogate)\n\n if np.max(losses.numpy()) > best_loss:\n n_fail = 0\n best_loss = torch.max(losses).detach().numpy()\n else:\n n_fail += len(perturbed_graphs)\n if self.terminate_after_n_fail is not None and n_fail > self.terminate_after_n_fail:\n print('Patience reached. Terminating the current run')\n break\n\n return pd.concat(dfs), adv_example\n\n def suggest(self, base_graph: dgl.DGLGraph, n_edit: int, prohibited_edges: list = None):\n \"\"\"\n The BO function to suggest perturbations to be queried from self.classifier\n :param base_graph: the graph on which we perform perturbations\n :param n_edit: number of edge edit allowed per perturbation\n :param prohibited_edges: list of edge edits that are not allowed.\n :return: a list of dgl graphs of shape self.batch_size\n \"\"\"\n is_edge_weighted = 'weight' in base_graph.edata.keys()\n candidate_graphs = None\n\n n_samples = self.acq_settings['acq_max_step']\n if self.acq_settings['acq_optimiser'] == 'random':\n if self.mode == 'rewire':\n candidate_samples = [random_sample_rewire_swap(base_graph,\n n_edit,\n rewire_only=not is_edge_weighted,\n n_hop=self.n_hop_constraint,\n preserve_disconnected_components=self.preserve_disconnected_components,\n ) for _\n in range(n_samples)]\n else:\n candidate_samples = [random_sample_flip(base_graph, n_edit, remove_edge_only=self.mode == 'remove',\n add_edge_only=self.mode == 'add', n_hop=self.n_hop_constraint,\n committed_edges=self.committed_edits,\n preserve_disconnected_components=self.preserve_disconnected_components,\n\n )\n for _ in range(n_samples)]\n\n elif self.acq_settings['acq_optimiser'] in ['genetic', 'mutation']:\n n_round = 10\n top_k = 3\n pop_size = max(n_samples // n_round, 100)\n # optionally set the fraction of randomly generated samples\n n_rand = np.round(pop_size * self.acq_settings['random_frac']).astype(np.int)\n n_mutate = pop_size - n_rand\n\n genetic_optimiser = Genetic(classifier=lambda x_: 0, loss_fn=lambda x_: 0,\n population_size=pop_size,\n mutation_rate=1., mode=self.mode)\n if self.mode == 'rewire':\n candidate_samples = [\n random_sample_rewire_swap(base_graph, n_edit, rewire_only=not is_edge_weighted,\n n_hop=self.n_hop_constraint,\n preserve_disconnected_components=self.preserve_disconnected_components,\n\n ) for _ in\n range(n_rand)]\n else:\n candidate_samples = [\n random_sample_flip(base_graph, n_edit, remove_edge_only=self.mode == 'remove',\n add_edge_only=self.mode == 'add',\n n_hop=self.n_hop_constraint,\n committed_edges=self.committed_edits,\n preserve_disconnected_components=self.preserve_disconnected_components,\n\n ) for _ in range(n_rand)]\n\n self.query_history += candidate_samples\n topk_indices = torch.topk(self.surrogate.y, min(self.surrogate.y.shape[0], top_k))[1]\n while len(candidate_samples) < pop_size:\n selected_index = topk_indices[np.random.randint(len(topk_indices))]\n candidate_samples.append(\n genetic_optimiser.mutate_sample(base_graph,\n self.query_history[-len(self.surrogate.y) + selected_index],\n )\n )\n candidate_graphs = population_graphs(base_graph, candidate_samples, self.mode)\n acq_values = self.surrogate.acquisition(candidate_graphs, acq_func=self.acq_settings['acq_type'], bias=None)\n\n # for each mutation round, alternate between optimising the topology (A) with features (X)\n for r in range(n_round):\n topk_indices = torch.topk(acq_values, min(len(candidate_graphs), top_k))[1]\n while len(candidate_samples) < pop_size:\n selected_sample = candidate_samples[np.random.randint(len(topk_indices))]\n candidate_samples.append(\n genetic_optimiser.mutate_sample(base_graph, selected_sample,))\n candidate_samples = candidate_samples[n_mutate:]\n\n candidate_graphs = population_graphs(base_graph, candidate_samples, self.mode)\n acq_values = self.surrogate.acquisition(candidate_graphs, acq_func=self.acq_settings['acq_type'], bias=None)\n\n else:\n raise NotImplementedError(f'Unable to parse the acq_optimiser {self.acq_settings[\"acq_optimiser\"]}')\n\n if candidate_graphs is None:\n candidate_graphs = population_graphs(base_graph, candidate_samples, self.mode)\n\n acq_values = self.surrogate.acquisition(candidate_graphs, acq_func=self.acq_settings['acq_type'])\n\n acq_values_np = acq_values.detach().numpy().flatten()\n acq_values_np_, unique_idx = np.unique(acq_values_np, return_index=True)\n i = np.argpartition(acq_values_np_, -min(acq_values_np_.shape[0], self.batch_size))[\n -min(acq_values_np_.shape[0], self.batch_size):]\n indices = np.array([unique_idx[j] for j in i])\n suggested = [candidate_graphs[j] for j in indices]\n self.query_history += [candidate_samples[j] for j in indices]\n return suggested\n\n def observe(self, X, y, reset_surrogate=False):\n \"\"\"\n Update the BO with new sample-target pair(s) we obtained from quering the classifer\n :param X: a list of dgl graphs. The list of dgl graphs we queried from the classifier\n :param y: a Tensor of shape[0] = len(X). The tensor of the classifier loss\n :param reset_surrogate: whether to reset the surrogate (clearing all previous fitted (X, y)).\n \"\"\"\n nan_idx = (y != y).nonzero().view(-1)\n if nan_idx.shape[0] > 0:\n for i in nan_idx:\n X.pop(i)\n y = y[y == y]\n if self.surrogate.X is None or reset_surrogate:\n self.surrogate.fit(X, y)\n else:\n self.surrogate.update(X, y)\n\n @staticmethod\n def construct_dataframe(losses: np.array, predictions: torch.tensor, label: torch.tensor, queries: int) \\\n -> pd.DataFrame:\n \"\"\"Construct a pandas dataframe consistent with the base class. This dataframe is for all samples evaluated\n after exactly `queries` queries.\"\"\"\n labels = np.tile(label, len(predictions))\n df = pd.DataFrame({'losses': losses,\n 'correct_prediction': correct_predictions(predictions.numpy(), labels),\n 'queries': queries})\n return df\n\n def get_stage_statistics(self, max_queries: int, budget: int):\n if self.edit_per_stage is None:\n self.edit_per_stage = budget\n if budget % self.edit_per_stage:\n num_stages = budget // self.edit_per_stage + 1\n else:\n num_stages = budget // self.edit_per_stage\n query_per_edit = max_queries // budget\n stage_length = self.edit_per_stage * query_per_edit\n stages = []\n edits_per_stages = []\n for i in range(num_stages):\n stages.append(min(max_queries, i * stage_length))\n if sum(edits_per_stages) + self.edit_per_stage < budget:\n edits_per_stages.append(self.edit_per_stage)\n else:\n edits_per_stages.append(budget - sum(edits_per_stages))\n stages.append(max_queries)\n return np.array(stages), np.array(edits_per_stages)\n","repo_name":"xingchenwan/grabnel","sub_path":"src/attack/bayesopt_attack.py","file_name":"bayesopt_attack.py","file_ext":"py","file_size_in_byte":19657,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"81"}
+{"seq_id":"13113257132","text":"def calcular_edad(an):\n e = 2023 - an\n return e\n\ndef mayor_de_edad(e):\n m = False\n if e >= 18:\n m = \"Mayor de edad\"\n else:\n m = \"menor de edad\"\n return m\ndef generar_correo(n,a,p):\n for x in a:\n p += remplazar_caracteres(x)\n a = p\n\n\n n = n.strip()\n a = a.strip()\n correo = remplazar_caracteres(n[0].lower())\n x = a.split()\n #primer apellido\n correo += x[0].lower()\n #primera letra del segundo apellido\n correo += remplazar_caracteres(x[1][0].lower())\n correo += \"@unemi.edu.ec\"\n return correo\n\n\ndef remplazar_caracteres(l):\n r = l\n if l == \"ñ\":\n r = \"n\"\n elif l ==\"à\":\n r = \"a\"\n elif l == \"è\":\n r = \"e\"\n elif l == \"ì\":\n r = \"i\"\n elif l == \"ò\":\n r = \"o\"\n elif l == \"ù\":\n r = \"u\"\n return r\n\n#def generar_contrasena():\n #pass\ndef run():\n nombres = input(\"ingrese sus nombres: \")\n apellidos = input(\"ingrese sus apellidos : \")\n anio_nacimiento = int(input(\"Imgrese su año de nacimiento\"))\n edad = calcular_edad(anio_nacimiento)\n tipo_edad = mayor_de_edad(edad)\n correo = generar_correo(nombres,apellidos)\n #contrasena = generar_contrasena()\n\n\n print(f\"La edad de {nombres} es {edad}, es una persona {tipo_edad} su cuenta de correo generado es: {correo}\")\n\n\n\n#if_name== \"__main_\":\n\nrun()\n","repo_name":"theoriginalex/S2-TALLER_2.py","sub_path":"trabajo en clases.py","file_name":"trabajo en clases.py","file_ext":"py","file_size_in_byte":1364,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"35187286841","text":"#\n# @lc app=leetcode.cn id=547 lang=python3\n#\n# [547] 省份数量\n#\n\n# @lc code=start\nfrom typing import List\nclass bxjj:\n def __init__(self, l: int) -> None:\n self.data, self.l = [-1]*l, l\n def father(self, x: int) -> int:\n if self.data[x]<0: return x\n self.data[x] = self.father(self.data[x])\n return self.data[x]\n def merge(self, a: int, b: int) -> None:\n a, b = self.father(a), self.father(b)\n if a == b: return\n if self.data[a] < self.data[b]:\n self.data[a] += self.data[b]\n self.data[b] = a\n else:\n self.data[b] += self.data[a]\n self.data[a] = b\n def query(self, a: int, b: int) -> bool:\n return self.father(a) == self.father(b)\nclass Solution:\n def findCircleNum(self, isConnected: List[List[int]]) -> int:\n n = len(isConnected)\n a = bxjj(n)\n for i in range(n):\n for j in range(n):\n if isConnected[i][j]:\n a.merge(i, j)\n ret = 0\n for i in a.data:\n if i < 0: ret += 1\n return ret\n# @lc code=end\n\n","repo_name":"HellOwhatAs/Leetcode","sub_path":"547.省份数量.py","file_name":"547.省份数量.py","file_ext":"py","file_size_in_byte":1123,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"29369715306","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Jul 4 00:00:17 2021\r\n\r\n@author: LONG QIANG\r\n\r\n值迭代算求解GridWorld问题\r\n基于状态值函数\r\n\"\"\"\r\n\r\nimport numpy as np\r\n\r\n'''\r\n创建一个随机确定性策略\r\n'''\r\ndef create_random_greedy_policy(env):\r\n random_greedy_policy = {} # 用字典表示策略\r\n for state in env.get_state_space(): # 遍历每一个状态\r\n random_greedy_policy[state] = np.zeros(env.action_space_size)\r\n # 随机选择一个动作,设置其概率为1\r\n random_greedy_policy[state][np.random.choice(range(env.action_space_size))] = 1.0\r\n \r\n return random_greedy_policy # 返回策略\r\n\r\n'''\r\n迭代更新状态值函数\r\n'''\r\ndef statevalue_update(env,V):\r\n V_new = np.zeros_like(V) # 初始化新的状态值函数\r\n Psa = env.Psa() # 获取状态转移概率矩阵\r\n delta = 0 # 值函数更新前后最大绝对差值\r\n epsilon = 0.001 # 更新容忍系数\r\n no_value_change = True # 是否更新指示器\r\n \r\n # 对每一个状态进行循环\r\n for s_i,s in enumerate(env.get_state_space()):\r\n action_values = np.zeros(env.action_space_size)\r\n for a_i,a in enumerate(env.get_action_space()):\r\n for ns_i,ns in enumerate(env.get_state_space()):\r\n reward = env.Rsa(s,a,ns) # (s,a)转移到ns的即时奖励\r\n prob = Psa[s_i,a_i,ns_i] # (s,a)转移到ns的概率\r\n action_values[a_i] += prob*(reward+env.gamma*V[ns_i])\r\n V_new[s_i] = np.max(action_values)\r\n \r\n # 维持最大的增量\r\n delta = max(delta,np.abs(V_new[s_i]-V[s_i]))\r\n \r\n # 检查是否满足终止条件\r\n if delta >= epsilon:\r\n no_value_change = False\r\n \r\n return V_new, no_value_change\r\n\r\n'''\r\n策略改进函数,用贪婪法求解最优策略\r\n'''\r\ndef policy_update(env,V):\r\n Psa = env.Psa() # 获取状态转移概率矩阵\r\n policy = create_random_greedy_policy(env) # 初始化策略\r\n\r\n # 求解最优策略 \r\n for s_i,s in enumerate(env.get_state_space()): # 对每一个状态进行循环 \r\n action_values = np.zeros(env.action_space_size)\r\n for a_i,a in enumerate(env.get_action_space()):\r\n for ns_i,ns in enumerate(env.get_state_space()):\r\n reward = env.Rsa(s,a,ns) # (s,a)转移到ns的即时奖励\r\n prob = Psa[s_i,a_i,ns_i] # (s,a)转移到ns的概率\r\n action_values[a_i] += prob*(reward+env.gamma*V[ns_i])\r\n \r\n # 求解贪婪策略\r\n best_action = np.argmax(action_values)\r\n policy[s] = np.eye(env.action_space_size)[best_action]\r\n \r\n return policy\r\n\r\n'''\r\n将policy表示成矩阵形式\r\n'''\r\ndef policy_express(env,policy):\r\n policy_mat = np.zeros((env.grid_height,env.grid_width))\r\n for s in env.get_state_space():\r\n policy_mat[s[0]][s[1]] = np.argmax(policy[s])\r\n \r\n return policy_mat\r\n \r\n'''\r\n值迭代主程序,该函数是算法2-4的具体实现\r\n'''\r\ndef value_iteration(env,episode_limit=100):\r\n V = np.zeros(env.state_space_size)\r\n \r\n # 迭代法求解最优状态值\r\n for i in range(episode_limit):\r\n print('第{}次迭代'.format(i)) \r\n V,no_value_change = statevalue_update(env,V)\r\n print('V=',V)\r\n if no_value_change:\r\n print('Iteration terminate with stable state value.')\r\n break\r\n \r\n # 计算最优策略\r\n policy = policy_update(env,V)\r\n\r\n # 将决策表示成矩阵形式\r\n policy_mat = policy_express(env,policy)\r\n \r\n # 返回最优策略和对应状态值\r\n return policy,policy_mat, V\r\n\r\n'''\r\n主程序\r\n'''\r\nif __name__ == '__main__':\r\n import GridWorld\r\n env = GridWorld.GridWorldEnv()\r\n policy_opt,policy_mat,V_opt = value_iteration(env,episode_limit=100)\r\n print(policy_mat)\r\n print(V_opt)","repo_name":"QiangLong2017/Deep-Reiforcement-Learning","sub_path":"code/code2-2 值迭代法求解例1-1.py","file_name":"code2-2 值迭代法求解例1-1.py","file_ext":"py","file_size_in_byte":4113,"program_lang":"python","lang":"en","doc_type":"code","stars":32,"dataset":"github-code","pt":"81"}
+{"seq_id":"17378037111","text":"from django.shortcuts import render\n\n# Create your views here.\nfrom tweets.models import Tweet\n\n\ndef tweet_view(request):\n tweet_content = Tweet.objects.all()\n context = {\n \"tweet_content\": tweet_content\n }\n return render(request, \"index.html\", context)\n","repo_name":"Oviep/twitter_clone","sub_path":"tweets/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":273,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"37619611824","text":"import pytest\nimport pandas as pd\nimport datatest as dt\n\nd = {\n'cars' : ['BMW','Toyato','Audi'],\n'year' : [2000,2001,2002],\n}\n\nmyvar = pd.DataFrame(d)\n\ndef test_columns():\n dt.validate(\n myvar.columns,\n {'cars', 'year'}\n )\n","repo_name":"MuraliMR259/Python-code","sub_path":"pytest_python.py","file_name":"pytest_python.py","file_ext":"py","file_size_in_byte":243,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"37876692355","text":"#!/usr/bin/env python3\n\nimport sys\n\nfor line in sys.stdin:\n a = line.strip().split()\n if a[1] != \"T\":\n continue\n if not a[2].startswith(\"furi\"):\n continue\n print(\"%s = 0x%s;\" % (a[2], a[0]))\n","repo_name":"Disasm/flipper-sandbox","sub_path":"utils/nm2ld.py","file_name":"nm2ld.py","file_ext":"py","file_size_in_byte":217,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"15784112575","text":"# encode(): str -> bytes\n# decode(): bytes -> str\n\n# import socket module\nfrom socket import *\nserverSocket = socket(AF_INET, SOCK_STREAM)\n# Prepare a sever socket\n# 将套接字绑定到本地地址,地址是一个元组\n# 对于IP套接字,地址是一对(主机、端口)\n# 主机必须引用本地主机\nserverSocket.bind((\"localhost\", 6789))\n# 使服务器能够接受连接\nserverSocket.listen(1)\n\nwhile True:\n # Establish the connection\n print('Ready to serve...')\n # 返回一个表示连接的新套接字和客户端地址\n connectionSocket, addr = serverSocket.accept()\n try:\n # 从套接字接收最多缓冲大小的字节,返回的是 bytes 类型\n message = connectionSocket.recv(1024)\n # 分割 bytes,到各个 list,其中 index = 1 是请求文件地址\n filename = message.split()[1]\n # filename[1:] 去除了斜杠: b'/HelloWorld.html' -> b'HelloWorld.html'\n f = open(filename[1:])\n outputdata = f.read()\n # Send one HTTP header line into socket\n # 注意: 最后一定要两个\\n\n # Connection 一定要设置为 close\n header = 'HTTP/1.1 200 OK\\nConnection:close\\nContent-type:text/html\\nContent-length:%d\\n\\n' % len(outputdata)\n # 向套接字发送一个数据字符串\n connectionSocket.send(header.encode())\n # Send the content of the requested file to the client\n for i in range(0, len(outputdata)):\n connectionSocket.send(outputdata[i].encode())\n f.close()\n connectionSocket.close()\n except IOError:\n # Send response message for file not found\n header = 'HTTP/1.1 404 Not Found'\n # encode(): str -> bytes\n connectionSocket.send(header.encode())\n # Close client socket\n # 不能使用 serverSocket.close() 因为会直接关闭 socket\n connectionSocket.close()\n\nserverSocket.close()\n","repo_name":"chenyuxiang0425/Computer-Networking-A-Top-Down-Approach","sub_path":"lab01/WebServer.py","file_name":"WebServer.py","file_ext":"py","file_size_in_byte":1920,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"81"}
+{"seq_id":"12461330428","text":"import libtcodpy as libtcod\n\n\ndark_wall = libtcod.Color(0, 0, 100)\nlight_wall = libtcod.Color(130, 110, 50)\ndark_ground = libtcod.Color(50, 50, 150)\nlight_ground = libtcod.Color(200, 180, 50)\n\ndarker_red = libtcod.darker_red\n# render_all missing health in health bar\ndarker_green = libtcod.darker_green\n# place_objects troll\ndarker_orange = libtcod.darker_orange\n# place_objects shield\ndark_red = libtcod.dark_red\n# new_game~player_death dead player, place_objects~monster_death corpse\nwhite = libtcod.white\n# settings~player, settings~stairs, message default message color,\n# menu default foreground, Object.clear, render_all floor, render_all wall,\n# render_all default foreground, render_all~render_bar default foreground\nyellow = libtcod.yellow\n# Equipment.dequip message, Item.drop dropped item message,\n# player_game.check_level_up level up message\nred = libtcod.red\n# handle_keys~next_level descending message,\n# Item.pick_up full inventory message, new_game welcome message,\n# spells~cast_heal full health message, spells~cast_lightning no target message\ngreen = libtcod.green\n# Item.pick_up item pick up message\nsky = libtcod.sky\n# new_game dagger, place_objects sword\nviolet = libtcod.violet\n# place_objects healing potion\norange = libtcod.orange\n# place_objects~monster_death death message,\n# spells~cast_fireball area message, spells~cast_fireball hit message\nblack = libtcod.black\n# render_all default background\nlight_green = libtcod.light_green\n# Equipment.equip message, spells~cast_confuse success message\nlight_violet = libtcod.light_violet\n# handle_keys~next_level heal message, spells~cast_heal heal message\nlight_yellow = libtcod.light_yellow\n# place_objects scroll of lightning bolt, place_objects scroll of fireball,\n# place_objects scroll of confusion, tutmut default foreground\nlight_red = libtcod.light_red\n# render_all health total in health bar\nlight_gray = libtcod.light_gray\n# render_all default foreground\nlight_blue = libtcod.light_blue\n# spells~cast_lightning hit message\nlight_cyan = libtcod.light_cyan\n# spells~cast_fireball targeting message,\n# spells~cast_confuse targeting message\ndesaturated_green = libtcod.desaturated_green\n# place_objects orc\n","repo_name":"Akhier/Py-TutMut","sub_path":"color.py","file_name":"color.py","file_ext":"py","file_size_in_byte":2185,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"81"}
+{"seq_id":"3428808081","text":"from itertools import count\nfrom peewee import *\n\ndb = SqliteDatabase('database.sqlite')\n\nclass BaseModel(Model):\n class Meta:\n database = db\n\n\nclass User(BaseModel):\n id = IntegerField(primary_key=True)\n status = CharField()\n name = CharField()\n username = CharField()\n\nclass Room(BaseModel):\n id = AutoField()\n name = CharField()\n price = IntegerField()\n max_people_count = IntegerField()\n\nclass Order(BaseModel):\n id = AutoField()\n status = CharField(default = \"В обработке\")\n user = ForeignKeyField(User, backref='order', on_delete='CASCADE')\n room = CharField()\n count_people = IntegerField()\n phone = CharField()\n\n\n\ndb.create_tables([User, Room, Order])","repo_name":"CoOsmoze/AntiCafeBot","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":721,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"15654745011","text":"from tkinter import Tk, Label, Entry, Button, IntVar, Checkbutton, END\nimport youtube_dl\n\n\nclass MainWindow:\n def __init__(self, window):\n self.window = window\n window.title(\"youtube-dl GUI\")\n Label(window, text=\"URL# 1\").grid(row=1, column=1)\n Label(window, text=\"URL# 2\").grid(row=2, column=1)\n Label(window, text=\"URL# 3\").grid(row=3, column=1)\n Label(window, text=\"URL# 4\").grid(row=4, column=1)\n Label(window, text=\"URL# 5\").grid(row=5, column=1)\n self.url1 = Entry(window, width=50)\n self.url2 = Entry(window, width=50)\n self.url3 = Entry(window, width=50)\n self.url4 = Entry(window, width=50)\n self.url5 = Entry(window, width=50)\n self.url1.grid(row=1, column=2, padx=5, pady=10)\n self.url2.grid(row=2, column=2, padx=5, pady=10)\n self.url3.grid(row=3, column=2, padx=5, pady=10)\n self.url4.grid(row=4, column=2, padx=5, pady=10)\n self.url5.grid(row=5, column=2, padx=5, pady=10)\n Button(window, text=\"Done\", command=window.quit).grid(row=7, column=1, pady=10)\n self.checked1 = IntVar()\n self.checked2 = IntVar()\n self.checked3 = IntVar()\n self.checked4 = IntVar()\n self.checked5 = IntVar()\n Checkbutton(window, text=\"Audio Only\", variable=self.checked1).grid(row=1, column=3)\n Checkbutton(window, text=\"Audio Only\", variable=self.checked2).grid(row=2, column=3)\n Checkbutton(window, text=\"Audio Only\", variable=self.checked3).grid(row=3, column=3)\n Checkbutton(window, text=\"Audio Only\", variable=self.checked4).grid(row=4, column=3)\n Checkbutton(window, text=\"Audio Only\", variable=self.checked5).grid(row=5, column=3)\n Button(window, text=\"Download\", command=self.download).grid(row=7, column=3, pady=10)\n Button(window, text=\"Clear\", command=self.clear_inputs).grid(row=7, column=2)\n\n def download(self): # test link http://www.youtube.com/watch?v=BaW_jenozKc\n def downloader(link, opts):\n with youtube_dl.YoutubeDL(opts) as ydl:\n ydl.download([link])\n\n def downloader_opts(audio_check):\n if audio_check != 0: # 6\n ydl_opts = {\n 'format': 'bestaudio/best',\n 'postprocessors': [{\n 'key': 'FFmpegExtractAudio',\n 'preferredcodec': 'mp3',\n 'preferredquality': '256',\n }],\n }\n else:\n ydl_opts = {}\n return ydl_opts\n link1 = self.url1.get()\n link2 = self.url2.get()\n link3 = self.url3.get()\n link4 = self.url4.get()\n link5 = self.url5.get()\n if link1 != \"\":\n audio1 = self.checked1.get()\n downloader(link1, downloader_opts(audio1))\n print(\"Done\")\n else:\n print(\"URL# 1 is empty\")\n if link2 != \"\":\n audio2 = self.checked2.get()\n downloader(link2, downloader_opts(audio2))\n print(\"Done\")\n else:\n print(\"URL# 2 is empty\")\n if link3 != \"\":\n audio3 = self.checked3.get()\n downloader(link3, downloader_opts(audio3))\n print(\"Done\")\n else:\n print(\"URL# 3 is empty\")\n if link4 != \"\":\n audio4 = self.checked4.get()\n downloader(link4, downloader_opts(audio4))\n print(\"Done\")\n else:\n print(\"URL# 4 is empty\")\n if link5 != \"\":\n audio5 = self.checked5.get()\n downloader(link5, downloader_opts(audio5))\n print(\"Done\")\n else:\n print(\"URL# 5 is empty\")\n print(\"Done\")\n\n def clear_inputs(self): # clearing the entry boxes\n self.url1.delete(0, END) # deleting from first to the end\n self.url2.delete(0, END)\n self.url3.delete(0, END)\n self.url4.delete(0, END)\n self.url5.delete(0, END)\n self.checked1.set(0)\n self.checked2.set(0)\n self.checked3.set(0)\n self.checked4.set(0)\n self.checked5.set(0)\n\nif __name__ == '__main__':\n root = Tk()\n my_gui = MainWindow(root)\n root.mainloop()\n","repo_name":"catdog13/youtube_downloader","sub_path":"youtube_downloader_GUI.py","file_name":"youtube_downloader_GUI.py","file_ext":"py","file_size_in_byte":4245,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"29930308783","text":"from ..units import Entity\nfrom ...board import Board\nfrom ...utils.geom import get_hypotenuse_length, get_polygon_radius\n\n\ndef resize_unit(unit: Entity, board: Board) -> None:\n \"\"\"\n Resize a unit to fit a tile of the given board\n\n Args:\n unit: The unit to resize\n board: The board on which the tile of the unit is located\n \"\"\"\n if board.graphics is not None and unit.sprite is not None and unit.sprite.rect is not None:\n multiply_ratio = board.graphics.sideLength / max(unit.sprite.rect.height, unit.sprite.rect.width)\n hypotenuse = get_hypotenuse_length(unit.sprite.rect.height * multiply_ratio,\n unit.sprite.rect.width * multiply_ratio)\n tile_diameter = get_polygon_radius(board.graphics.nbrOfSides, board.graphics.sideLength) * 2\n while hypotenuse > tile_diameter:\n multiply_ratio *= 0.99\n hypotenuse = get_hypotenuse_length(unit.sprite.rect.height * multiply_ratio,\n unit.sprite.rect.width * multiply_ratio)\n unit.sprite.size(int(round(unit.sprite.rect.width * multiply_ratio)),\n int(round(unit.sprite.rect.height * multiply_ratio)))\n","repo_name":"Angeall/pyTGF","sub_path":"pytgf/characters/utils/units.py","file_name":"units.py","file_ext":"py","file_size_in_byte":1237,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"8688914691","text":"# coding=utf-8\n\nfrom utils.gensim_util import tokenization, similarity\nfrom utils.inverted_index import get_inverted_index\nfrom utils.question_util import load_questions\n\n\ndef search(question):\n # 分词\n words = tokenization(content=question)\n # 获取倒排索引\n inverted_index = get_inverted_index('inverted_index.pkl')\n # 从倒排索引中过滤出words中数据\n real_questions = []\n for word in words:\n if word in inverted_index:\n for q in inverted_index[word].keys():\n real_questions.append(q)\n # 去重\n real_questions = list(set(real_questions))\n # 计算相似度\n question_similarities = similarity(question, real_questions)\n question_similarities = list(map(lambda m: {\n 'sentence': m['sentence'],\n 'similarity': str(m['similarity'])\n }, question_similarities))\n # # 找到真正的问答详情\n # question_details = load_questions()\n return question_similarities\n\n\ndef question_search(question):\n questions = load_questions()\n if question in questions:\n return {\n 'question': question,\n 'detail': questions[question]\n }\n return {\n 'question': question,\n 'detail': '没有找到'\n }\n\n\n# if __name__ == '__main__':\n # print(search('结婚一年提出离婚,男方能否要求返还彩礼?'))\n # print(question_search('结婚一年提出离婚,男方能否要求返还彩礼?'))\n","repo_name":"supermareo/chinacourt","sub_path":"service/question_service.py","file_name":"question_service.py","file_ext":"py","file_size_in_byte":1467,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5913138744","text":"def f(x):\n return x*x\nd = {}\nn = int(input())\nfor i in range (1, n+1):\n x = int(input())\n if x not in d:\n d[x] = f(x)\n print(f(x))\n else:\n print(d.get(x))\n\n\n\n\n\n\n\n","repo_name":"deada11/PyLessons","sub_path":"FirstCourseOfPython/dicts-3.py","file_name":"dicts-3.py","file_ext":"py","file_size_in_byte":195,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"34082939799","text":"class Solution:\n def isValid(self, s: str) -> bool:\n dic = {'(':')','{':'}','[':']'}\n left = {'(','{','['}\n length = len(s)\n if length%2 == 1:\n return False\n tempStack = []\n for i in range(length):\n if s[i] in left:\n tempStack.append(s[i])\n else:\n if tempStack == []:\n return False\n top = tempStack.pop()\n if dic[top] == s[i]:\n continue\n else: \n return False\n if len(tempStack) != 0:\n return False\n return True\n","repo_name":"zjsdcae/LeetCode","sub_path":"Stack/020.py","file_name":"020.py","file_ext":"py","file_size_in_byte":646,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"27775607856","text":"import pandas as pd\r\n\r\nask_state = True\r\nmain_list = []\r\n\r\nwhile ask_state == True:\r\n ask_subject = input(\"subject (box kain, props, game items, stationery, indofest, cooking supplies): \")\r\n ask_item_name = input(\"Enter the name of the item: \")\r\n ask_invent_qty = input('Enter the storage quantity of item: ')\r\n\r\n ask_state = input('continue y/n: ')\r\n if ask_state == \"n\":\r\n ask_state = False\r\n else:\r\n ask_state = True\r\n\r\nsub_list = [ask_subject, ask_item_name, ask_invent_qty]\r\nmain_list.append(sub_list)\r\n\r\ncolumns_list = ['Subject', 'Item', 'Quantity']\r\n\r\ndf = pd.DataFrame(main_list, columns=columns_list)\r\n\r\nprint(df.to_excel(\"excel.xlsx\"))\r\n\r\n\r\n\r\n \r\n ","repo_name":"chritzadz/coding_problems","sub_path":"input_invent_v1.py","file_name":"input_invent_v1.py","file_ext":"py","file_size_in_byte":704,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"39508465183","text":"n=int(input())\narr=[]\nfor i in range(n):\n x,y=map(int,input().split())\n arr.append([x,y])\ndist=10**18\nax,ay=arr[0]\nfor i in range(n):\n tDist=-(10**18)\n for j in range(n):\n if i==j: continue\n tDist=max(tDist,(arr[i][0]-arr[j][0])**2+(arr[i][1]-arr[j][1])**2)\n if tDist int:\n EXPS = set((\"+\", \"-\", \"*\", \"/\"))\n stack = []\n ans = int(tokens[0])\n\n for token in tokens:\n if token not in EXPS:\n stack.append(token)\n continue\n \n num2 = int(stack.pop())\n num1 = int(stack.pop())\n\n ans = self.calc(\n num1,\n num2,\n token\n )\n stack.append(ans)\n\n return ans\n \n \n def calc(self, a, b, exp):\n if exp == \"+\": return a + b\n if exp == '-': return a - b\n if exp == \"*\": return a * b\n if exp == '/': return int(a / b)\n\n\n\"\"\"\nYou are given an array of strings tokens that represents an arithmetic expression in a Reverse Polish Notation.\n\nEvaluate the expression. Return an integer that represents the value of the expression.\n\nNote that:\n\nThe valid operators are '+', '-', '*', and '/'.\nEach operand may be an integer or another expression.\nThe division between two integers always truncates toward zero.\nThere will not be any division by zero.\nThe input represents a valid arithmetic expression in a reverse polish notation.\nThe answer and all the intermediate calculations can be represented in a 32-bit integer.\n \n\nExample 1:\n\nInput: tokens = [\"2\",\"1\",\"+\",\"3\",\"*\"]\nOutput: 9\nExplanation: ((2 + 1) * 3) = 9\nExample 2:\n\nInput: tokens = [\"4\",\"13\",\"5\",\"/\",\"+\"]\nOutput: 6\nExplanation: (4 + (13 / 5)) = 6\n\nExample 3:\n\nInput: tokens = [\"10\",\"6\",\"9\",\"3\",\"+\",\"-11\",\"*\",\"/\",\"*\",\"17\",\"+\",\"5\",\"+\"]\nOutput: 22\nExplanation: ((10 * (6 / ((9 + 3) * -11))) + 17) + 5\n= ((10 * (6 / (12 * -11))) + 17) + 5\n= ((10 * (6 / -132)) + 17) + 5\n= ((10 * 0) + 17) + 5\n= (0 + 17) + 5\n= 17 + 5\n= 22\n \n\nConstraints:\n\n1 <= tokens.length <= 104\ntokens[i] is either an operator: \"+\", \"-\", \"*\", or \"/\", or an integer in the range [-200, 200].\n\"\"\"","repo_name":"lauvsong/leetcode","sub_path":"medium/150.py","file_name":"150.py","file_ext":"py","file_size_in_byte":1913,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31174907225","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\n***************************************************************************\r\n* Copyright (c) 2022 *\r\n* Shai Seger *\r\n* *\r\n* This file is a supplement to the FreeCAD CAx development system. *\r\n* *\r\n* This program is free software; you can redistribute it and/or modify *\r\n* it under the terms of the GNU Lesser General Public License (LGPL) *\r\n* as published by the Free Software Foundation; either version 2 of *\r\n* the License, or (at your option) any later version. *\r\n* for detail see the LICENCE text file. *\r\n* *\r\n* This software is distributed in the hope that it will be useful, *\r\n* but WITHOUT ANY WARRANTY; without even the implied warranty of *\r\n* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *\r\n* GNU Library General Public License for more details. *\r\n* *\r\n* You should have received a copy of the GNU Library General Public *\r\n* License along with this macro; if not, write to the Free Software *\r\n* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *\r\n* USA *\r\n* *\r\n***************************************************************************\r\n\"\"\"\r\nfrom screw_maker import *\r\nimport FastenerBase\r\n\r\n# PCB spacers / Wurth standard WA-SSTII\r\n\r\ncos30 = math.cos(math.radians(30))\r\n\r\n\r\ndef pspMakeFace(m, sw, l, id, thl):\r\n id2 = id / 2.0\r\n sw2 = sw / 2.0\r\n m2 = m / 2.0\r\n d2 = 0.95 * sw2 / cos30\r\n l1 = l - (d2 - sw2) / 2.0\r\n dd = m2 - id2\r\n thl1 = thl - id2\r\n\r\n fm = FastenerBase.FSFaceMaker()\r\n fm.AddPoints(\r\n (id2, l - dd),\r\n (id2 + dd, l),\r\n (sw2, l),\r\n (d2, l1),\r\n (d2, dd),\r\n (sw2, 0),\r\n (id2 + dd, 0),\r\n (id2, dd),\r\n )\r\n if thl > 0:\r\n # separate holes\r\n fm.AddPoints((id2, thl1), (0, thl), (0, l - thl), (id2, l - thl1))\r\n return fm.GetFace()\r\n\r\n\r\ndef makePCBSpacer(self, fa):\r\n diam = fa.calc_diam\r\n width = fa.width\r\n flen = fa.calc_len\r\n\r\n FreeCAD.Console.PrintLog(\r\n \"Making PCB spacer\" + diam + \"x\" + str(flen) + \"x\" + str(width) + \"\\n\"\r\n )\r\n\r\n th, _ = fa.dimTable\r\n dia = self.getDia(fa.calc_diam, True)\r\n P = FsData[\"MetricPitchTable\"][fa.diameter][0]\r\n id = self.GetInnerThreadMinDiameter(dia, P)\r\n w = float(width)\r\n l = float(flen)\r\n if l > th:\r\n # separate thread holes on both sides\r\n thl = 10\r\n if thl + 0.5 > l / 2.0:\r\n thl = l / 2.0 - 0.5\r\n else:\r\n thl = 0\r\n\r\n f = pspMakeFace(dia * 1.05, w, l, id, thl)\r\n p = self.RevolveZ(f)\r\n htool = self.makeHexPrism(w, l)\r\n htool.translate(Base.Vector(0.0, 0.0, - 0.1))\r\n fSolid = p.common(htool)\r\n if fa.thread:\r\n if thl > 0: # blind & threaded from both sides\r\n threadCutter = self.CreateInnerThreadCutter(dia, P, thl - dia / 2)\r\n fSolid = fSolid.cut(threadCutter)\r\n threadCutter.rotate(\r\n Base.Vector(0.0, 0.0, l / 2),\r\n Base.Vector(1.0, 0.0, 0.0),\r\n 180\r\n )\r\n else: # has through hole, fully threaded\r\n threadCutter = self.CreateInnerThreadCutter(dia, P, l + P)\r\n fSolid = fSolid.cut(threadCutter)\r\n return fSolid\r\n","repo_name":"shaise/FreeCAD_FastenersWB","sub_path":"FsFunctions/FSmakePCBSpacer.py","file_name":"FSmakePCBSpacer.py","file_ext":"py","file_size_in_byte":3870,"program_lang":"python","lang":"en","doc_type":"code","stars":231,"dataset":"github-code","pt":"81"}
+{"seq_id":"37741904608","text":"from flask import Blueprint, jsonify, request\nfrom flask_login import login_required\nfrom app.models import db, User\n\nfollows_routes = Blueprint('follows', __name__)\n\n#GET /api/follows/\n## PEOPLE YOU FOLLOW\n@follows_routes.route('/')\n@login_required\ndef get_follows(id):\n user = User.query.get(id)\n follows = user.following.all()\n return {'follows': [follow.to_dict() for follow in follows]}\n\n#GET /api/follows/\n## PEOPLE FOLLOWING **YOU**\n@follows_routes.route('//followers')\n@login_required\ndef get_your_followers(id):\n user = User.query.get(id)\n followers = user.followers.all()\n return {'followers': [follower.to_dict() for follower in followers]}\n\n#POST\n@follows_routes.route('//new-follow', methods=[\"POST\"])\n@login_required\ndef post_follow(id):\n \n user = User.query.get(id)\n followerId = request.json['followerId']\n followedId = request.json['followedId']\n\n followers = user.followers.all()\n newFollower = User.query.get(followerId)\n\n if newFollower in user.followers:\n user.followers.remove(newFollower)\n db.session.commit()\n return newFollower.to_dict()\n else:\n user.followers.append(newFollower)\n db.session.commit()\n return newFollower.to_dict()","repo_name":"alex-pober/Instagram-Clone","sub_path":"app/api/follows_routes.py","file_name":"follows_routes.py","file_ext":"py","file_size_in_byte":1280,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"24031555646","text":"# https://www.codewars.com/kata/5567e7d0adb11174c50000a7/train/python\n\n# This kata is about singly linked list. A linked list is an ordered set of data elements, each containing a link to its successor (and sometimes its predecessor, known as a double linked list). You are you to implement an algorithm to find the kth to last element.\n\n# For example given a linked list of:\n\n# a -> b -> c -> d\n\n# if k is the number one then d should be returned\n# if k is the number two then c should be returned\n# if k is the number three then b should be returned\n# if k is the number four then a should be returned\n# if k exceeds the size of the list then None returned\n# Special Note --> Node classes contain two fields; data and next. And to access the head of the list, use head. e.g. linked_list.head\n\n\ndef get_length(l):\n count, node = 0, l.head\n while node != None:\n node = node.next\n count += 1\n return count\n\ndef search_k_from_end(linked_list, k):\n length = get_length(linked_list)\n count, node = 0, linked_list.head\n while count < length - k:\n node = node.next\n count += 1\n return node.data if k <= length else None","repo_name":"phil-huynh/Problem-Sets","sub_path":"python/CodeWars/6kyu/node_mania.py","file_name":"node_mania.py","file_ext":"py","file_size_in_byte":1162,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"35563927190","text":"import os,logging,subprocess, time\nimport tensorflow as tf\nimport PIL\nimport numpy as np\nfrom scipy import misc\nfrom collections import defaultdict\nfrom .facenet import facenet\nfrom .facenet.align import detect_face\nimport random\nfrom time import sleep\n\n\ndef _parse_function(filename):\n image_string = tf.read_file(filename)\n image_decoded = tf.image.decode_image(image_string,channels=3)\n return tf.expand_dims(image_decoded, 0), filename\n\n\n\ndef pil_to_array(pilImage):\n \"\"\"\n Load a PIL image and return it as a numpy array. For grayscale\n images, the return array is MxN. For RGB images, the return value\n is MxNx3. For RGBA images the return value is MxNx4\n \"\"\"\n def toarray(im, dtype=np.uint8):\n \"\"\"Return a 1D array of dtype.\"\"\"\n # Pillow wants us to use \"tobytes\"\n if hasattr(im, 'tobytes'):\n x_str = im.tobytes('raw', im.mode)\n else:\n x_str = im.tostring('raw', im.mode)\n x = np.fromstring(x_str, dtype)\n return x\n\n if pilImage.mode in ('RGBA', 'RGBX'):\n im = pilImage # no need to convert images\n elif pilImage.mode == 'L':\n im = pilImage # no need to luminance images\n # return MxN luminance array\n x = toarray(im)\n x.shape = im.size[1], im.size[0]\n return x\n elif pilImage.mode == 'RGB':\n # return MxNx3 RGB array\n im = pilImage # no need to RGB images\n x = toarray(im)\n x.shape = im.size[1], im.size[0], 3\n return x\n elif pilImage.mode.startswith('I;16'):\n # return MxN luminance array of uint16\n im = pilImage\n if im.mode.endswith('B'):\n x = toarray(im, '>u2')\n else:\n x = toarray(im, ' min_score:\n top,left = (int(boxes[0][i][0] * shape[0]), int(boxes[0][i][1] * shape[1]))\n bot,right = (int(boxes[0][i][2] * shape[0]), int(boxes[0][i][3] * shape[1]))\n detections.append({\n 'x': left,\n 'y':top,\n 'w':right-left,\n 'h':bot-top,\n 'score': scores[0][i],\n 'object_name': self.class_index_to_string[int(classes[0][i])]\n })\n return detections\n\n def load(self):\n self.detection_graph = tf.Graph()\n with self.detection_graph.as_default():\n self.filenames_placeholder = tf.placeholder(\"string\")\n dataset = tf.contrib.data.Dataset.from_tensor_slices(self.filenames_placeholder)\n dataset = dataset.map(_parse_function)\n self.iterator = dataset.make_initializable_iterator()\n self.od_graph_def = tf.GraphDef()\n with tf.gfile.GFile(self.model_path, 'rb') as fid:\n serialized_graph = fid.read()\n self.od_graph_def.ParseFromString(serialized_graph)\n self.image, self.fname = self.iterator.get_next()\n tf.import_graph_def(self.od_graph_def, name='',input_map={'image_tensor': self.image})\n config = tf.ConfigProto()\n config.gpu_options.per_process_gpu_memory_fraction = 0.15\n self.session = tf.Session(graph=self.detection_graph,config=config)\n self.boxes = self.detection_graph.get_tensor_by_name('detection_boxes:0')\n self.scores = self.detection_graph.get_tensor_by_name('detection_scores:0')\n self.classes = self.detection_graph.get_tensor_by_name('detection_classes:0')\n self.num_detections = self.detection_graph.get_tensor_by_name('num_detections:0')\n\n\nclass FaceDetector():\n\n def __init__(self,session=None):\n self.image_size = 182\n self.margin = 44\n self.gpu_memory_fraction = 0.1\n self.session = session\n self.minsize = 20\n self.threshold = [0.6, 0.7, 0.7]\n self.factor = 0.709\n\n def load(self):\n logging.info('Creating networks and loading parameters')\n with tf.Graph().as_default():\n gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=self.gpu_memory_fraction)\n self.session = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options, log_device_placement=False))\n with self.session.as_default():\n self.pnet, self.rnet, self.onet = detect_face.create_mtcnn(self.session, None)\n\n def detect(self,image_path):\n aligned = []\n try:\n img = misc.imread(image_path)\n except (IOError, ValueError, IndexError) as e:\n errorMessage = '{}: {}'.format(image_path, e)\n logging.info(errorMessage)\n else:\n if img.ndim < 2:\n logging.info('Unable to align \"%s\"' % image_path)\n return []\n if img.ndim == 2:\n img = facenet.to_rgb(img)\n img = img[:, :, 0:3]\n bounding_boxes, _ = detect_face.detect_face(img, self.minsize, self.pnet, self.rnet, self.onet, self.threshold, self.factor)\n nrof_faces = bounding_boxes.shape[0]\n if nrof_faces > 0:\n det_all = bounding_boxes[:, 0:4]\n img_size = np.asarray(img.shape)[0:2]\n for boxindex in range(nrof_faces):\n det = np.squeeze(det_all[boxindex, :])\n bb = np.zeros(4, dtype=np.int32)\n bb[0] = np.maximum(det[0] - self.margin / 2, 0)\n bb[1] = np.maximum(det[1] - self.margin / 2, 0)\n bb[2] = np.minimum(det[2] + self.margin / 2, img_size[1])\n bb[3] = np.minimum(det[3] + self.margin / 2, img_size[0])\n cropped = img[bb[1]:bb[3], bb[0]:bb[2], :]\n scaled = misc.imresize(cropped, (self.image_size, self.image_size), interp='bilinear')\n left, top, right, bottom = bb[0], bb[1], bb[2], bb[3]\n aligned.append({\n 'scaled':scaled,\n 'x': left,\n 'y':top,\n 'w':right-left,\n 'h':bottom-top,\n })\n return aligned","repo_name":"simmoncn/DeepVideoAnalytics","sub_path":"dvalib/detector.py","file_name":"detector.py","file_ext":"py","file_size_in_byte":7586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"81"}
+{"seq_id":"70389278346","text":"## This version works, without the extra blank line\n##\n## Write a program that asks the user for a positive number 'n' as input.\n## Assume that the user enters a number greater than or equal to 3 and\n## print a triangle as described below. For example if the user enters 6 then the output should be:\n## *\n## **\n## ***\n## ****\n## *****\n## ******\n## *****\n## ****\n## ***\n## **\n## *\n\nn = int(input ('Enter a positive number great than or equal to 3:'))\ni = 1\nk = 1\nwhile i <= n and i > 0 and k < 2*n:\n print ('*'*i)\n if k < n:\n i = i+1\n else:\n i = i-1\n k = k+1\n \n \n","repo_name":"lipingzhu/PythonEdX","sub_path":"quiz3_part3_1.py","file_name":"quiz3_part3_1.py","file_ext":"py","file_size_in_byte":598,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"11143962020","text":"import cv2\nface_cascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml') \nsmile_cascade = cv2.CascadeClassifier('haarcascade_smile.xml') \ndef detect(gray, frame): \n faces = face_cascade.detectMultiScale(gray, 1.3, 5) \n for (x, y, w, h) in faces: \n cv2.rectangle(frame, (x, y), ((x + w), (y + h)), (255, 0, 0), 2) \n roi_gray = gray[y:y + h, x:x + w] \n roi_color = frame[y:y + h, x:x + w] \n smiles = smile_cascade.detectMultiScale(roi_gray, 1.8, 20) \n \n for (sx, sy, sw, sh) in smiles: \n cv2.rectangle(roi_color, (sx, sy), ((sx + sw), (sy + sh)), (0, 0, 255), 2) \n return frame \nvideo = cv2.VideoCapture(0,cv2.CAP_DSHOW) \nwhile True: \n a, frame = video.read() # used a as a throwaway variable\n \n gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY) \n \n feed = detect(gray, frame) \n \n cv2.imshow('Video', feed) \n \n if cv2.waitKey(1) & 0xff == ord('e'): \n break\nvideo.release() \ncv2.destroyAllWindows() \n","repo_name":"vulbsti/Smile_detection","sub_path":"smile_windows.py","file_name":"smile_windows.py","file_ext":"py","file_size_in_byte":1110,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4820554723","text":"import yaml\nimport subprocess\nimport base64\nimport json\n\n\ndef create_contract_query_cmd(address, q):\n prefix = [\"junod\", \"q\", \"wasm\", \"contract-state\", \"smart\"]\n\n return prefix + [address, q]\n\n\ndef contract_query(query):\n q = json.dumps(query[\"content\"]).replace(\"null\", f'{{}}')\n\n subprocess.run(create_contract_query_cmd(query[\"address\"], q), check=True)\n\n\ndef create_tx_cmd(cmd):\n prefix = [\"junod\", \"tx\", \"wasm\"]\n suffix = [\n \"--from\", \"user1\", \"--chain-id\", \"testing\", \"--yes\", \"-b\", \"block\",\n \"--gas\", \"auto\", \"--gas-adjustment\", \"1.5\"\n ]\n\n return prefix + cmd + suffix\n\n\ndef create_store_cmd(contract_file):\n return create_tx_cmd([\"store\", contract_file])\n\n\ndef create_instantiate_cmd(name, code_id, init_msg):\n return create_tx_cmd(\n [\"instantiate\", code_id, init_msg, \"--label\", name, \"--no-admin\"])\n\n\ndef get_code_id_from_response(res):\n # parse response and get code id\n for event in res[\"events\"]:\n if event[\"type\"] == \"store_code\":\n for attr in event[\"attributes\"]:\n if base64.b64decode(attr[\"key\"]).decode('utf-8') == \"code_id\":\n return base64.b64decode(attr[\"value\"]).decode('utf-8')\n\n raise Exception(\"No code id found in response\")\n\n\ndef store(contract_file):\n p = subprocess.run(create_store_cmd(contract_file),\n check=True,\n stdout=subprocess.PIPE)\n res = yaml.safe_load(p.stdout)\n\n return get_code_id_from_response(res)\n\n\ndef get_contract_addresses_from_response(res, contract_name_list):\n addresses = {}\n\n events = res[\"logs\"][0][\"events\"]\n\n # parse response and get contract addresses from the list of names\n for event in events:\n if event[\"type\"] == \"wasm\":\n for attr in event[\"attributes\"]:\n if attr[\"key\"] in contract_name_list:\n addresses[attr[\"key\"]] = attr[\"value\"]\n\n # check in instantiate event type if nothing found\n if len(addresses) == 0:\n for event in events:\n if event[\"type\"] == \"instantiate\":\n for attr in event[\"attributes\"]:\n if attr[\"key\"] in contract_name_list:\n addresses[attr[\"key\"]] = attr[\"value\"]\n\n return addresses\n\n\ndef format_msg(msg):\n return msg.replace(\"null\", f'{{}}')\n\n\ndef instantiate_contract(instance):\n init_msg = json.dumps(instance[\"msg\"])\n\n init_msg = format_msg(init_msg)\n\n p = subprocess.run(create_instantiate_cmd(instance[\"name\"],\n str(instance[\"code\"]), init_msg),\n check=True,\n stdout=subprocess.PIPE)\n res = yaml.safe_load(p.stdout)\n\n # print(json.dumps(res, indent=4))\n # exit(1)\n\n return get_contract_addresses_from_response(res, instance[\"address_list\"])\n\n\ndef compile_wasm():\n subprocess.run([\"cargo\", \"wasm\"],\n check=True,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL)\n\n\ndef contracts():\n # read contracts.yaml file\n with open(\"contracts.yaml\") as file:\n contracts = yaml.safe_load(file)\n return contracts\n\n\ndef queries():\n # read queries.yaml file\n with open(\"queries.yaml\") as file:\n contracts = yaml.safe_load(file)\n return contracts\n\n\ndef dump_code_ids(codes):\n with open(\"codes.yaml\", \"w\") as file:\n yaml.dump(codes, file)\n\n\ndef dump_contract_addresses(addresses_from_serie, serie):\n with open(\"addresses.yaml\") as file:\n addresses = yaml.safe_load(file)\n addresses[serie] = addresses_from_serie\n with open(\"addresses.yaml\", \"w\") as addr_file:\n yaml.dump(addresses, addr_file)","repo_name":"lumtis/wasmlab","sub_path":"scripts/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":3732,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"36876757656","text":"from codecs import open\nfrom os.path import abspath, dirname, join, expanduser\n\nfrom setuptools import Command, find_packages, setup\n\nthis_dir = abspath(dirname(__file__))\n\nwith open(join(this_dir, 'README.md'), encoding='utf-8') as file:\n long_description = file.read()\n\nsetup(\n name='bhimupipy',\n version=\"0.0.2\",\n description=\"A package to find upi details details\",\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n url=\"https://github.com/shrikantbache/bhimupipy.git\",\n author=\"BACHE SHRIKANT GANGADHAR\",\n author_email='bacheshrikant@gmail.com',\n maintainer=\"BACHE SHRIKANT GANGADHAR\",\n maintainer_email=\"bacheshrikant@gmail.com\",\n license='MIT',\n license_file=\"LICENSE\",\n platforms=\"any\",\n\n classifiers=[\n 'Topic :: Utilities',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n\n ],\n keywords=['search', 'bhimupijs', 'brute', 'address', 'vpa', 'upi', 'spy'],\n include_package_data=True,\n install_requires=['requests'],\n entry_points={\n 'console_scripts': [\n 'bhimupipy=bhimupipy.__init__:ExecuteBhimupiPy',\n ],\n },\n\n)\n","repo_name":"shrikantbache/bhimupipy","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1189,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"37971119596","text":"from .Message import Message\nfrom typing import List, Dict, Any\n\n\nclass Mailer:\n\n def __init__(self):\n\n self.mailbox: Dict[int, List[Message]] = dict()\n self.delivered: List[Message] = []\n\n def get_messages(self, recipient: int) -> List[Message]:\n \"\"\"\n Method to get all the messages of a given recipient\n :param recipient: recipient id\n :return: list with all the messages\n \"\"\"\n\n return self.mailbox.get(recipient, [])\n\n def deliver_message(self, sender: int, recipient: int, content: Any, title: str):\n\n \"\"\"\n The method create a message and append it to delivered messages list.\n :param title: title of the message\n :param recipient: id of the recipient\n :param sender: id of the sender\n :param content: an object with the content of the message\n :return: none\n \"\"\"\n\n msg = Message(sender, recipient, content, title)\n\n self.delivered.append(msg)\n\n def assign_messages(self):\n\n \"\"\"\n Get all messages from the delivered list and assign them according to the message recipient.\n :return:\n \"\"\"\n\n # Empty the current mailbox\n self.mailbox = dict()\n\n for msg in self.delivered:\n\n recipient = msg.recipient\n\n # If this recipient already have messages\n try:\n self.mailbox[recipient].append(msg)\n\n # Except if we don't have any messages for this recipient yet\n except KeyError:\n\n self.mailbox[recipient] = [msg]\n\n # Empty delivered messages\n self.delivered = []\n\n\n\n\n\n\n\n","repo_name":"evyatarluv/Artificial-Intelligence","sub_path":"DCOP/Mailer.py","file_name":"Mailer.py","file_ext":"py","file_size_in_byte":1648,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"29462281063","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.contrib import admin\n\n# Register your models here.\nfrom .models import Note, NoteContent\n\n\nclass NoteContentInline(admin.TabularInline):\n model = NoteContent\n extra = 1\n\n\nclass NoteAdmin(admin.ModelAdmin):\n list_display = ('title', 'description', 'date_created', 'date_modified')\n inlines = [NoteContentInline]\n\n\nadmin.site.register(Note, NoteAdmin)\n","repo_name":"ladeoshodi/gnotes","sub_path":"api/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":439,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"11553573784","text":"\"\"\"\nID: warreng1\nLANG: PYTHON3\nTASK: dualpal\n\"\"\"\n\nwith open('dualpal.in', 'r') as fin:\n firstnumbers, start = fin.readline().split()\n firstnumbers = int(firstnumbers)\n start = int(start)\n\ndef converttobase(num, base):\n newstr = ''\n copynum = num\n while copynum:\n mod = copynum % base\n copynum = copynum // base\n newstr = chr(48+mod+7*(mod>=10)) + newstr\n return newstr\n\nanswers = []\n\nwhile firstnumbers:\n #Look for matching numbers:\n amountofpalin = 0\n for i in range(2, 11):\n convertnum = converttobase(start+1, i)\n if convertnum == convertnum[::-1]:\n amountofpalin += 1\n if amountofpalin == 2:\n answers.append(start+1)\n firstnumbers -= 1\n break\n start += 1\n\nwith open('dualpal.out', 'w') as fout:\n for i in answers:\n #print(i)\n print(i, file=fout)","repo_name":"waguo/USACO","sub_path":"training/Chapter_1/dualpal/dualpal.py","file_name":"dualpal.py","file_ext":"py","file_size_in_byte":885,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"38662516848","text":"from pprint import pprint\nfrom browsermobproxy import Server\nfrom selenium import webdriver\nfrom time import sleep\n\nclass ProxyManager:\n\n __BMP = \"/home/gitcha/BMP/bin/browsermob-proxy\"\n bmp_port={'port':8090}\n\n def __init__(self): \n self.__server = Server(ProxyManager.__BMP, options=ProxyManager.bmp_port)\n self.__client = None\n \n\n def start_server(self):\n self.__server.start()\n return self.__server\n\n \n def start_client(self):\n self.__client = self.__server.create_proxy(params={'trustAllServers':'true'})\n return self.__client\n \n\n @property\n def client(self):\n return self.__client\n\n \n @property\n def server(self):\n return self.__server\n\n\n \nif __name__ == \"__main__\":\n\n proxy = ProxyManager()\n server = proxy.start_server()\n client = proxy.start_client()\n \n client.new_har(\"ya.ru\")\n print(client.proxy)\n\n options = webdriver.ChromeOptions()\n options.add_argument(\"--proxy-server={}\".format(client.proxy))\n options.add_argument('ignore-certificate-errors')\n \n driver = webdriver.Chrome(options=options)\n driver.get(\"https://www.ya.ru\")\n sleep(3)\n\n pprint(client.har)\n \n driver.close()\n server.stop()\n driver.quit\n\n \n\n \n\n \n\n","repo_name":"vobla13/Appium-Pytest-Android","sub_path":"proxy/bmp_proxy.py","file_name":"bmp_proxy.py","file_ext":"py","file_size_in_byte":1297,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"2541998376","text":"L1 = [2, 7, 5, 6, 7, 1, 6, 2, 1, 7, 6]\n\nx = 0\ncpt = 0\nsave = 0\nsbr = 0\nfor i in range(len(L1)):\n x = L1[i]\n for y in range(len(L1)):\n if x == L1[y]:\n cpt += 1\n\n if cpt > save:\n save = cpt\n sbr = L1[i]\n cpt = 0\n\n\nprint(\"Le nombre le plus frequent dans la liste est le :\",sbr,\"(\",save,\"x)\")","repo_name":"Ethan68000/TP-r1.07","sub_path":"tp4/Exo4 TP4 (part 1).py","file_name":"Exo4 TP4 (part 1).py","file_ext":"py","file_size_in_byte":332,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"36369076471","text":"\"\"\"crmproj URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/1.11/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.conf.urls import url, include\n 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url\nfrom crm import views\n\nurlpatterns = [\n\n url(r'^depart/list/',views.depart_list,name=\"departlist\" ),\n url(r'^depart/add/',views.depart_add,name=\"departadd\" ),\n url(r'^depart/edit/(\\d+)/',views.depart_edit,name=\"departedit\" ),\n url(r'^depart/del/(\\d+)/',views.depart_del,name=\"departdel\" ),\n url(r'^user/list/',views.user_list,name=\"userlist\" ),\n url(r'^user/add/',views.user_add,name=\"useradd\" ),\n url(r'^user/edit/(\\d+)/',views.user_edit,name=\"useredit\" ),\n url(r'^user/del/(\\d+)/',views.user_del,name=\"userdel\" ),\n url(r'^class/list/',views.class_list,name=\"classlist\" ),\n url(r'^class/add/',views.class_change,name=\"classadd\" ),\n url(r'^class/edit/(\\d+)/',views.class_change,name=\"classedit\" ),\n\n]\n","repo_name":"zhouyu37/crmproj","sub_path":"crm/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1408,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1646438118","text":"import argparse\nimport os\nimport joblib\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.feature_extraction.text import HashingVectorizer\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.experimental import enable_halving_search_cv\nfrom sklearn.model_selection import HalvingGridSearchCV\nfrom sklearn.feature_extraction import text\nimport nltk\nfrom nltk.stem.snowball import SnowballStemmer\nimport re\nimport numpy as np\nfrom nltk.tokenize import word_tokenize\nfrom nltk.probability import FreqDist\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom nltk.stem import WordNetLemmatizer\nfrom pprint import pprint\nimport pandas as pd\n\n\n\nclass LemmaTokenizer:\n def __init__(self):\n self.wnl = WordNetLemmatizer()\n def __call__(self, doc):\n return [self.wnl.lemmatize(t) for t in word_tokenize(doc)]\n\ndef stemmed_words(doc):\n return (stemmer.stem(w) for w in analyzer(doc))\n\n\"\"\"\nmodel_fn\n model_dir: (sting) specifies location of saved model\n\nThis function is used by AWS Sagemaker to load the model for deployment.\nIt does this by simply loading the model that was saved at the end of the\n__main__ training block above and returning it to be used by the predict_fn\nfunction below.\n\"\"\"\ndef model_fn(model_dir):\n model = joblib.load(os.path.join(model_dir, \"model.joblib\"))\n return model\n\n\"\"\"\ninput_fn\n request_body: the body of the request sent to the model. The type can vary.\n request_content_type: (string) specifies the format/variable type of the request\n\nThis function is used by AWS Sagemaker to format a request body that is sent to \nthe deployed model.\nIn order to do this, we must transform the request body into a numpy array and\nreturn that array to be used by the predict_fn function below.\n\nNote: Oftentimes, you will have multiple cases in order to\nhandle various request_content_types. Howver, in this simple case, we are \nonly going to accept text/csv and raise an error for all other formats.\n\"\"\"\ndef input_fn(request_body, request_content_type):\n if request_content_type == 'text/csv':\n samples = []\n for r in request_body.split('|'):\n samples.append(list(map(float,r.split(','))))\n return np.array(samples)\n else:\n raise ValueError(\"Thie model only supports text/csv input\")\n\n\"\"\"\npredict_fn\n input_data: (numpy array) returned array from input_fn above \n model (sklearn model) returned model loaded from model_fn above\n\nThis function is used by AWS Sagemaker to make the prediction on the data\nformatted by the input_fn above using the trained model.\n\"\"\"\ndef predict_fn(input_data, model):\n return model.predict(input_data)\n\n\"\"\"\noutput_fn\n prediction: the returned value from predict_fn above\n content_type: (string) the content type the endpoint expects to be returned\n\nThis function reformats the predictions returned from predict_fn to the final\nformat that will be returned as the API call response.\n\nNote: While we don't use content_type in this example, oftentimes you will use\nthat argument to handle different expected return types.\n\"\"\"\ndef output_fn(prediction, content_type):\n return '|'.join([t for t in prediction])\n\nif __name__ =='__main__':\n # Create a parser object to collect the environment variables that are in the\n # default AWS Scikit-learn Docker container.\n parser = argparse.ArgumentParser()\n\n parser.add_argument('--output-data-dir', type=str, default=os.environ.get('SM_OUTPUT_DATA_DIR'))\n parser.add_argument('--model-dir', type=str, default=os.environ.get('SM_MODEL_DIR'))\n parser.add_argument('--train', type=str, default=os.environ.get('SM_CHANNEL_TRAIN'))\n parser.add_argument('--test', type=str, default=os.environ.get('SM_CHANNEL_TEST'))\n\n args = parser.parse_args()\n\n # Load data from the location specified by args.train (In this case, an S3 bucket).\n nltk.download('wordnet')\n nltk.download('punkt')\n\n stemmer = SnowballStemmer('english')\n analyzer = HashingVectorizer().build_analyzer()\n\n #df = pd.read_csv('./archive/modified.csv', encoding=\"ISO-8859-1\")\n\n df = pd.read_csv(os.path.join(args.train, 'modified.csv'), index_col=0, engine=\"python\")\n\n df.dropna(subset = ['abstract','headline'], inplace=True)\n\n X = df['abstract']\n X_headline = df['headline']\n X_section = df['section']\n Y = df['clickbait_category_4']\n\n documents = []\n\n # text preprocessing\n\n Y_modified = pd.DataFrame()\n\n\n for sen in range(0, len(X)):\n # Remove all the special characters\n\n headline = X_headline.get(sen)\n abstract = X.get(sen)\n section = X_section.get(sen)\n\n if not(headline is None) and not(abstract is None) and not(section is None):\n\n doc = section + \" : \" + headline + \" : \" + abstract\n document = re.sub(r'\\W', ' ', str(doc))\n\n # remove all single characters\n document = re.sub(r'\\s+[a-zA-Z]\\s+', ' ', document)\n\n # Remove single characters from the start\n document = re.sub(r'\\^[a-zA-Z]\\s+', ' ', document)\n\n # Substituting multiple spaces with single space\n document = re.sub(r'\\s+', ' ', document, flags=re.I)\n\n # Removing prefixed 'b'\n document = re.sub(r'^b\\s+', '', document)\n\n # Converting to Lowercase\n document = document.lower()\n\n # Lemmatization\n document = document.split()\n\n # document = [stemmer.lemmatize(word) for word in document]\n document = ' '.join(document)\n\n documents.append(document)\n\n #print(\"document: \" + document)\n\n Y_modified = Y_modified.append({'clickbait_category_4': Y[sen]}, ignore_index=True)\n\n\n # frequency filtering\n\n\n tokens = word_tokenize(\"\\n\".join(X.values))\n freq = FreqDist(tokens)\n frequent_words = []\n\n for key, value in freq.items():\n if value >= 200:\n frequent_words.append(key.lower())\n\n stop_words = text.ENGLISH_STOP_WORDS\n\n vectorizer = TfidfVectorizer(tokenizer=LemmaTokenizer(), max_features=2500, analyzer=stemmed_words , stop_words=stop_words)\n\n # create the parameter grid:\n\n # Number of trees in random forest\n n_estimators = [int(x) for x in np.linspace(start = 200, stop = 2000, num = 10)]\n # Number of features to consider at every split\n max_features = ['auto', 'sqrt']\n # Maximum number of levels in tree\n max_depth = [int(x) for x in np.linspace(10, 110, num = 11)]\n max_depth.append(None)\n # Minimum number of samples required to split a node\n min_samples_split = [2, 5, 10]\n # Minimum number of samples required at each leaf node\n min_samples_leaf = [1, 2, 4]\n # Method of selecting samples for training each tree\n bootstrap = [True, False]\n\n # Create the random grid\n # 'n_estimators': n_estimators,\n\n random_grid = {\n 'max_features': max_features,\n 'max_depth': max_depth,\n 'min_samples_split': min_samples_split,\n 'min_samples_leaf': min_samples_leaf,\n 'bootstrap': bootstrap}\n\n pprint(random_grid)\n\n\n processed_features = vectorizer.fit_transform(documents)\n\n processed_features.shape\n\n X_dense = processed_features.todense()\n\n X_dense.shape\n\n x_train, x_test, y_train, y_test = train_test_split(X_dense, Y_modified['clickbait_category_4'], test_size = 0.2)\n\n x_train.shape, x_test.shape\n\n y_train.shape, y_test.shape\n\n clf = RandomForestClassifier(n_estimators = 100)\n\n model = HalvingGridSearchCV(estimator=clf, param_grid = random_grid, cv=3, factor=2,\n resource='n_estimators',max_resources=30).fit(x_train, y_train)\n\n #Save the model to the location specified by args.model_dir\n joblib.dump(model, os.path.join(args.model_dir, \"model.joblib\"))","repo_name":"abhinavGirish/newsy","sub_path":"aws_training.py","file_name":"aws_training.py","file_ext":"py","file_size_in_byte":7880,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"25967820336","text":"from database import DATABASE\nimport myparser\nimport freqlist\nfrom dict import DICT\nfrom math import log, fabs, exp\n\nMAX = 12000\nSTEP = 1000\nEPS = 0.01\n\n\ndef get_pr_func(db=DATABASE, max_=MAX, step=STEP):\n '''\n Zwraca funkcję, która przyporządokwuje ang. słowu szacowane prawdopodobieństwo znania tego słowa.\n\n :param db: baza danych zawierająca informacje o słownictwie użytkownika\n :param max_: maks. pozycja na liście frekwencyjnej słów branych pod uwagę\n :param step: dł. przedziału, na jakim jest szacowane prawdopodobieństwo\n :return: funkcja, która przyporządokwuje ang. słowu szacowane prawdopodobieństwo znania tego słowa\n '''\n rank_prob = [(-step // 2, 1.0)]\n rank = 0\n while rank + step <= max_:\n pr = db.known_words(rank, rank + step)\n mean_rank = rank + step // 2\n if pr is None:\n (r, p) = rank_prob[-1]\n pr = abs(r) / mean_rank * p\n rank_prob.append((mean_rank, pr))\n rank += step\n\n def pr_func(word):\n x = freqlist.FREQLIST.rank(word)\n if x is None:\n return EPS\n i = 0\n while i+1 < len(rank_prob):\n if rank_prob[i][0] <= x < rank_prob[i+1][0]:\n break\n i += 1\n if i+1 == len(rank_prob):\n (r, p) = rank_prob[-1]\n pr = abs(r) / x * p\n return pr\n (x1, y1) = rank_prob[i]\n (x2, y2) = rank_prob[i+1]\n return y1 + (y2 - y1) / (x2 - x1) * (x - x1)\n\n return pr_func\n\n\ndef get_utility_func(db=DATABASE, fl=freqlist.FREQLIST):\n '''\n Zwraca funkcję użyteczności, która przyporządkowuje słowu jego wartość.\n\n :param db: baza danych zawierająca informacje o słownictwie użytkownika\n :param fl: lista frekwencyjna\n :return: funkcja użyteczności, która przyporządkowuje słowu jego wartość\n '''\n pr = get_pr_func(db=db)\n C = 11.2\n\n def utility_func(word):\n f = fl.freq(word)\n p = 1.0 - pr(word)\n # print((word, f, p))\n return (log(C+f) - log(C)) * exp(C*p)\n\n return utility_func\n\n\ndef select(stream, db=DATABASE):\n '''\n Zwraca listę słów posortowanych według wartości funkcji użyteczności dla podanego strumienia znakowego.\n\n :param stream: strumień znakowy\n :param db: baza danych zawierająca informacje o słownictwie użytkownika\n :return: lista słów posortowanych według wartości funkcji użyteczności\n '''\n words = myparser.parse(stream)\n temp_fl = freqlist.FreqList()\n temp_fl.load(stream=stream)\n fl = freqlist.DynMixedFreqList([(freqlist.FREQLIST, 0.5), (temp_fl, 0.5)])\n utility_fun = get_utility_func(db=db, fl=fl)\n pr = get_pr_func(db=db)\n mean_word = []\n for word in words:\n if DICT.correct(word) is not None and DATABASE.known_now(word) == False:\n mean_word.append((utility_fun(word), word, pr(word), ))\n mean_word.sort(reverse=True)\n # print(mean_word)\n return [word for (_, word, _) in mean_word]\n\nif __name__ == '__main__':\n pass\n","repo_name":"kocichy/Words","sub_path":"selector.py","file_name":"selector.py","file_ext":"py","file_size_in_byte":3057,"program_lang":"python","lang":"pl","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41740810192","text":"import cv2\nimport json\nimport requests\nimport time\nfrom requests.structures import CaseInsensitiveDict\n\nurl = \"https://demo.thingsboard.io/api/v1/Yn5VHwYZhjyl8zpcHF4T/telemetry\"\nheaders = CaseInsensitiveDict()\nheaders[\"Content-Type\"] = \"application/json\"\n\n# Initialize the cv2 QRCode detector\ndetector = cv2.QRCodeDetector()\n\n# Open the webcam\ncap = cv2.VideoCapture(0)\n\n# Set the scanner window dimensions\nscan_width = 300\nscan_height = 300\n\n# Set the position of the scanner window\nscan_x = int((cap.get(cv2.CAP_PROP_FRAME_WIDTH) - scan_width) / 2)\nscan_y = int((cap.get(cv2.CAP_PROP_FRAME_HEIGHT) - scan_height) / 2)\n\n# Set the delay time in seconds\ndelay_time = 5\n\n# Set the start time to the current time\nstart_time = time.time()\n\nwhile True:\n # Read the image from the webcam\n ret, img = cap.read()\n\n # Create a gray version of the image\n gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n\n # Detect and decode the QR code in the image\n data, bbox, _ = detector.detectAndDecode(img)\n\n # If a QR code is detected\n if data and (time.time() - start_time) >= delay_time:\n # Print the decoded data to the console\n print(\"Data:\", data)\n\n # Convert the data string to a dictionary\n data_dict = eval(data)\n\n # Loading the json data\n json_data = json.loads(data)\n\n # Extracting the keys and values from the json data\n keys = []\n values = []\n for key, value in json_data.items():\n keys.append(key)\n values.append(value)\n\n # Constructing the data string to be sent\n data = {}\n for i in range(len(keys)):\n data[keys[i]] = values[i]\n\n # Sending the data to Thingsboard\n response = requests.post(url, headers=headers, json=data)\n print(response.status_code)\n \n # Set the start time to the current time\n start_time = time.time()\n\n # Draw the scanner window\n cv2.rectangle(img, (scan_x, scan_y), (scan_x+scan_width, scan_y+scan_height), (255, 0, 0), 2)\n\n # Show the image in a window\n cv2.imshow(\"QR Code Scanner\", img)\n\n # Exit the loop if the \"s\" key is pressed\n if cv2.waitKey(1) & 0xFF == ord(\"s\"):\n break\n# Release the webcam and close the window\ncap.release()\ncv2.destroyAllWindows()\n","repo_name":"Pentagon-Deerhack/HEMOve","sub_path":"QRcode/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2290,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"8121633951","text":"#!/usr/bin/env python\n\n#Coding ini detect bola merah dan bveeta mini follow bola\n#Bizbot Technology\n#Programmer: Ts.Khairul\n\nimport cv2\nimport serial\nimport time\n\n# Connect to the serial port to send commands to the robot\n#ser = serial.Serial('/dev/ttyUSB1', 57600) # Replace with the correct port and baud rate for your robot\n#time.sleep(2)\n\n# Load the image or video\ncap = cv2.VideoCapture(0) # Use 0 for webcam or specify file path for video\n\n# Define the lower and upper bounds of the red color in HSV color space\nred_lower = (0, 120, 70)\nred_upper = (10, 255, 255)\n\nser = serial.Serial('/dev/ttyUSB1', 57600, timeout=0.050)\ntime.sleep(2) #must wait for 2 seconds for connections established\n\nwhile True:\n # Capture frame-by-frame\n ret, frame = cap.read()\n if not ret:\n break\n\n # Convert the frame to HSV color space\n hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV)\n\n # Threshold the HSV image to get only red colors\n mask = cv2.inRange(hsv, red_lower, red_upper)\n\n # Find contours of the red color in the mask image\n contours, hierarchy = cv2.findContours(mask, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n\n # Draw a rectangle around the ball if it is detected\n if len(contours) > 0:\n # Find the largest contour (assuming it is the ball)\n largest_contour = max(contours, key=cv2.contourArea)\n (x, y, w, h) = cv2.boundingRect(largest_contour)\n cv2.rectangle(frame, (x, y), (x + w, y + h), (0, 255, 0), 2)\n\n # Compute the center of the ball\n cx = int(x + w/2)\n cy = int(y + h/2)\n\n # Send commands to the robot to follow the ball\n if cx < 320:\n #ser.write(b'l') # Turn left\n ser.write(b\"m 0 -30 \\r\\n\")\n elif cx > 320:\n #ser.write(b'r') # Turn right\n ser.write(b\"m -30 0 \\r\\n\")\n else:\n #ser.write(b'f') # Move forward\n ser.write(b\"m -80 -80 \\r\\n\")\n else:\n # Stop the robot if the ball is not detected\n #ser.write(b's') # Stop moving\n ser.write(b\"m 0 0 \\r\\n\")\n\n # Display the resulting image\n cv2.imshow('frame', frame)\n\n # Exit the loop if 'q' is pressed\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n\n# Release the capture and destroy all windows\ncap.release()\ncv2.destroyAllWindows()\n","repo_name":"skj84/bveeta-R007","sub_path":"Tutorial_6_red_ball_follow.py","file_name":"Tutorial_6_red_ball_follow.py","file_ext":"py","file_size_in_byte":2315,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"81"}
+{"seq_id":"10280544652","text":"import unittest\n\nfrom title_case import title_case\n\nclass Test(unittest.TestCase):\n def test_title_case(self):\n string='the quick brown fox jumped over the lazy dog'\n result=title_case(string)\n self.assertEqual(result, 'The Quick Brown Fox Jumped Over The Lazy Dog')\n\nif __name__=='__main__':\n unittest.main()","repo_name":"kendayao/algorithms-practice","sub_path":"12-title-case/test_title_case.py","file_name":"test_title_case.py","file_ext":"py","file_size_in_byte":336,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"71817466492","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\" \n@Author:ling \n@Date: 2017/11/13 \n\"\"\"\n\n\ndef triangles(maxline):\n a, li = 0, [1]\n while a < maxline:\n yield li\n li = [1] + [li[x] + li[x+1] for x in range(a)] + [1]\n a += 1\n\nn = 0\nfor t in triangles(10):\n print(t)\n n += 1\n if n == 10:\n break\n","repo_name":"leif-sh/PythonDemo","sub_path":"ling/triangles.py","file_name":"triangles.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"37389433712","text":"from pygdbmi.constants import DEFAULT_GDB_TIMEOUT_SEC\nfrom pygdbmi.gdbcontroller import GdbController\n\nfrom playground.gdb_wrapper.gdb_wrapper import gdb_wrapper\n\n\nclass gdb_wrapper_x86_64(gdb_wrapper):\n def __init__(self, port: int = None, file: str = None):\n self.gdb_ctrl = GdbController([\"gdb-multiarch\", \"-q\", \"--interpreter=mi\"])\n self.gdb_ctrl.write(\"set architecture i386:x86_64\")\n self._registers = {'r{}'.format(i) for i in range(8, 16)}\n self._registers.update({'rax', 'rdi', 'rsi', 'rdx', 'rcx', 'rbx', 'rsp', 'rbp', 'rip', 'eflags'})\n # self._registers.update({'eax', 'edi', 'esi', 'edx', 'ecx', 'ebx', 'esp', 'ebp', 'eip'})\n self._flags_name = 'eflags'\n self._flag_to_pos = {'CF': [0], 'PF': [2], 'AF': [4], 'ZF': [6], 'SF': [7], 'TF': [8], 'IF': [9], 'DF': [10],\n 'OF': [11], 'IOPL': [11, 12], 'NT': [14], 'RF': [16], 'VM': [17], 'AC': [18], 'VIF': [19],\n 'VIP': [20], 'ID': [21]}\n\n super().__init__(port, file)\n\n @gdb_wrapper.no_response()\n def get_flags(self, timeout_sec: int = DEFAULT_GDB_TIMEOUT_SEC):\n log = self.gdb_ctrl.write(\"print ${}\".format(self._flags_name), timeout_sec)\n _, _, values = self._parse_log(log, 'console')['payload'].partition(' = ')\n result = {}\n all_flags = values.rstrip('\\\\n').strip('][').split()\n for flag_value in all_flags:\n flag_name, _, value = flag_value.partition('=')\n if value == '':\n value = 1\n result[flag_name] = value\n for flag_name in self._flag_to_pos:\n if flag_name not in result:\n result[flag_name] = 0\n return result\n","repo_name":"OSLL/asm_web_debug","sub_path":"playground/gdb_wrapper/gdb_wrapper_x86.py","file_name":"gdb_wrapper_x86.py","file_ext":"py","file_size_in_byte":1725,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"42517843389","text":"from __future__ import absolute_import, division, print_function\n\n__metaclass__ = type\n\nANSIBLE_METADATA = {\n \"metadata_version\": \"1.1\",\n \"status\": [\"preview\"],\n \"supported_by\": \"community\",\n}\n\nDOCUMENTATION = \"\"\"\n---\nmodule: oci_opsi_operations_insights_warehouse_resource_usage_summary_facts\nshort_description: Fetches details about a OperationsInsightsWarehouseResourceUsageSummary resource in Oracle Cloud Infrastructure\ndescription:\n - Fetches details about a OperationsInsightsWarehouseResourceUsageSummary resource in Oracle Cloud Infrastructure\n - Gets the details of resources used by an Operations Insights Warehouse.\n There is only expected to be 1 warehouse per tenant. The warehouse is expected to be in the root compartment.\nversion_added: \"2.9.0\"\nauthor: Oracle (@oracle)\noptions:\n operations_insights_warehouse_id:\n description:\n - Unique Operations Insights Warehouse identifier\n type: str\n aliases: [\"id\"]\n required: true\nextends_documentation_fragment: [ oracle.oci.oracle ]\n\"\"\"\n\nEXAMPLES = \"\"\"\n- name: Get a specific operations_insights_warehouse_resource_usage_summary\n oci_opsi_operations_insights_warehouse_resource_usage_summary_facts:\n # required\n operations_insights_warehouse_id: \"ocid1.operationsinsightswarehouse.oc1..xxxxxxEXAMPLExxxxxx\"\n\n\"\"\"\n\nRETURN = \"\"\"\noperations_insights_warehouse_resource_usage_summary:\n description:\n - OperationsInsightsWarehouseResourceUsageSummary resource\n returned: on success\n type: complex\n contains:\n id:\n description:\n - OPSI Warehouse OCID\n returned: on success\n type: str\n sample: \"ocid1.resource.oc1..xxxxxxEXAMPLExxxxxx\"\n cpu_used:\n description:\n - Number of OCPUs used by OPSI Warehouse ADW. Can be fractional.\n returned: on success\n type: float\n sample: 1.2\n storage_used_in_gbs:\n description:\n - Storage by OPSI Warehouse ADW in GB.\n returned: on success\n type: float\n sample: 1.2\n lifecycle_state:\n description:\n - Possible lifecycle states\n returned: on success\n type: str\n sample: CREATING\n sample: {\n \"id\": \"ocid1.resource.oc1..xxxxxxEXAMPLExxxxxx\",\n \"cpu_used\": 1.2,\n \"storage_used_in_gbs\": 1.2,\n \"lifecycle_state\": \"CREATING\"\n }\n\"\"\"\n\nfrom ansible_collections.oracle.oci.plugins.module_utils import oci_common_utils\nfrom ansible_collections.oracle.oci.plugins.module_utils.oci_resource_utils import (\n OCIResourceFactsHelperBase,\n get_custom_class,\n OCIAnsibleModule,\n)\n\ntry:\n from oci.opsi import OperationsInsightsClient\n\n HAS_OCI_PY_SDK = True\nexcept ImportError:\n HAS_OCI_PY_SDK = False\n\n\nclass OperationsInsightsWarehouseResourceUsageSummaryFactsHelperGen(\n OCIResourceFactsHelperBase\n):\n \"\"\"Supported operations: get\"\"\"\n\n def get_required_params_for_get(self):\n return [\n \"operations_insights_warehouse_id\",\n ]\n\n def get_resource(self):\n return oci_common_utils.call_with_backoff(\n self.client.summarize_operations_insights_warehouse_resource_usage,\n operations_insights_warehouse_id=self.module.params.get(\n \"operations_insights_warehouse_id\"\n ),\n )\n\n\nOperationsInsightsWarehouseResourceUsageSummaryFactsHelperCustom = get_custom_class(\n \"OperationsInsightsWarehouseResourceUsageSummaryFactsHelperCustom\"\n)\n\n\nclass ResourceFactsHelper(\n OperationsInsightsWarehouseResourceUsageSummaryFactsHelperCustom,\n OperationsInsightsWarehouseResourceUsageSummaryFactsHelperGen,\n):\n pass\n\n\ndef main():\n module_args = oci_common_utils.get_common_arg_spec()\n module_args.update(\n dict(\n operations_insights_warehouse_id=dict(\n aliases=[\"id\"], type=\"str\", required=True\n ),\n )\n )\n\n module = OCIAnsibleModule(argument_spec=module_args)\n\n if not HAS_OCI_PY_SDK:\n module.fail_json(msg=\"oci python sdk required for this module.\")\n\n resource_facts_helper = ResourceFactsHelper(\n module=module,\n resource_type=\"operations_insights_warehouse_resource_usage_summary\",\n service_client_class=OperationsInsightsClient,\n namespace=\"opsi\",\n )\n\n result = []\n\n if resource_facts_helper.is_get():\n result = resource_facts_helper.get()\n else:\n resource_facts_helper.fail()\n\n module.exit_json(operations_insights_warehouse_resource_usage_summary=result)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"oracle/oci-ansible-collection","sub_path":"plugins/modules/oci_opsi_operations_insights_warehouse_resource_usage_summary_facts.py","file_name":"oci_opsi_operations_insights_warehouse_resource_usage_summary_facts.py","file_ext":"py","file_size_in_byte":4692,"program_lang":"python","lang":"en","doc_type":"code","stars":151,"dataset":"github-code","pt":"78"}
+{"seq_id":"12922759289","text":"import pytest\nimport threading\nimport time\n\nfrom psycopg2_mq import MQSource, MQWorker\n\n\ndef test_integration(model, dbsession, worker_proxy):\n worker_proxy.start(\n queues={\n 'dummy': DummyQueue(),\n }\n )\n\n source = MQSource(dbsession=dbsession, model=model)\n with dbsession.begin():\n job_id = source.call('dummy', 'echo', {'message': 'hello world'})\n\n while True:\n with dbsession.begin():\n job = source.find_job(job_id)\n if job.state in {\n model.JobStates.COMPLETED,\n model.JobStates.FAILED,\n model.JobStates.LOST,\n }:\n break\n time.sleep(0.1)\n\n job = source.find_job(job_id)\n assert job.state == model.JobStates.COMPLETED\n assert job.result == {\n 'queue': 'dummy',\n 'method': 'echo',\n 'args': {'message': 'hello world'},\n }\n assert job.start_time is not None\n assert job.end_time is not None\n\n\nclass DummyQueue:\n def execute_job(self, job):\n return {\n 'queue': job.queue,\n 'method': job.method,\n 'args': job.args,\n }\n\n\nclass WorkerProxy:\n worker = None\n thread = None\n\n def __init__(self, model, dbengine):\n self.model = model\n self.dbengine = dbengine\n\n def start(self, **kw):\n self.worker = MQWorker(\n engine=self.dbengine,\n model=self.model,\n capture_signals=False,\n **kw,\n )\n self.thread = threading.Thread(target=self.worker.run)\n self.thread.daemon = True\n self.thread.start()\n return self.worker\n\n def stop(self):\n self.worker.shutdown_gracefully()\n self.thread.join()\n\n\n@pytest.fixture\ndef worker_proxy(model, dbengine):\n proxy = WorkerProxy(model, dbengine)\n try:\n yield proxy\n finally:\n proxy.stop()\n","repo_name":"mmerickel/psycopg2_mq","sub_path":"tests/test_integration.py","file_name":"test_integration.py","file_ext":"py","file_size_in_byte":1909,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"73411504252","text":"\nwith open(\"saved.txt\", \"r\") as f:\n bit_string = f.readlines()[0]\nnumber_of_bits = len(bit_string)\n\npointer = 0\nn = 0\n\ndef read_n_bits(num):\n global pointer\n global number_of_bits\n if num == 0: return 0\n res = int(bit_string[pointer:pointer+num], 2)\n pointer += num\n return res\n\nrows = read_n_bits(16)\ncols = read_n_bits(16)\ncolor_dif_bits = read_n_bits(8)\nn = rows * cols\n\nv = [[False]*cols for i in range(rows)]\n\npath = []\npath_to_index = {}\nk = 0\nfor i in range(rows):\n for j in range(cols):\n path.append((i,j))\n path_to_index[(i,j)] = k\n k += 1\n\nv = [False]*n\nr = v[:]\ng = v[:]\nb = v[:]\n\n\ndef neighbor_cells(r, c, visited):\n neighbors = [\n (r + rd, c + cd)\n for rd, cd in\n [(1,1),(1,0),(1,-1),(0,1),(0,-1),(-1,1),(-1,0),(-1,-1)]\n ]\n neighbors = [(r, c) for r, c in neighbors\n if (r >= 0 and r < rows and c >= 0 and c < cols\n and not visited[path_to_index[(r, c)]]\n )]\n\n return neighbors\n\n\n\nfor i in range(n):\n if v[i]: continue\n v[i] = True\n\n r[i], g[i], b[i] = [read_n_bits(8) for _ in range(3)]\n\n queue = [i]\n\n while queue:\n cur = queue.pop(0)\n cur_r, cur_c = path[cur]\n\n neighbors = neighbor_cells(cur_r, cur_c, v)\n bitset = read_n_bits(len(neighbors))\n # print(len(neighbors))\n # if len(neighbors) == 4: exit()\n if bitset == 0: continue\n\n for index, nei in enumerate(neighbors):\n if not (bitset & (1 << index)): continue\n nei_r, nei_c = nei\n nei_index = path_to_index[(nei_r, nei_c)]\n v[nei_index] = True\n\n __ = [None]*3\n for _ in range(3):\n if read_n_bits(1) == 1:\n __[_] = read_n_bits(color_dif_bits - 1)\n else:\n __[_] = -read_n_bits(color_dif_bits - 1)\n\n r[nei_index] = r[cur] + __[0]\n g[nei_index] = g[cur] + __[1]\n b[nei_index] = b[cur] + __[2]\n\n queue.append(nei_index)\n\nprint(pointer, number_of_bits)\nprint((number_of_bits + 7999) // 8000, \"kb\")\n\ncolor_string = f\"{rows} {cols}\\n\"\nfor i in range(n):\n color_string += f\"{r[i]} {g[i]} {b[i]}\\n\"\n\nwith open(\"decompressed_img.txt\", \"w\") as f:\n f.write(color_string)\n\n# for i in range(n):\n# print(r[i], g[i], b[i])\n","repo_name":"ollelapidus/Gymnasiearbete---Complete-Bucket-Tiling","sub_path":"cbt_decompression.py","file_name":"cbt_decompression.py","file_ext":"py","file_size_in_byte":2361,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"832110504","text":"from test_framework.test_framework import IoPTestFramework\nfrom test_framework.util import *\n\n# Create one-input, one-output, no-fee transaction:\nclass MempoolSpendCoinbaseTest(IoPTestFramework):\n\n def __init__(self):\n super().__init__()\n self.num_nodes = 1\n self.setup_clean_chain = False\n\n def setup_network(self):\n # Just need one node for this test\n args = [\"-checkmempool\", \"-debug=mempool\"]\n self.nodes = []\n self.nodes.append(start_node(0, self.options.tmpdir, args))\n self.is_network_split = False\n\n def run_test(self):\n chain_height = self.nodes[0].getblockcount()\n assert_equal(chain_height, 200)\n node0_address = self.nodes[0].getnewaddress()\n\n # Coinbase at height chain_height-100+1 ok in mempool, should\n # get mined. Coinbase at height chain_height-100+2 is\n # is too immature to spend.\n b = [ self.nodes[0].getblockhash(n) for n in range(101, 103) ]\n coinbase_txids = [ self.nodes[0].getblock(h)['tx'][0] for h in b ]\n spends_raw = [ create_tx(self.nodes[0], txid, node0_address, 49.99) for txid in coinbase_txids ]\n\n spend_101_id = self.nodes[0].sendrawtransaction(spends_raw[0])\n\n # coinbase at height 102 should be too immature to spend\n assert_raises(JSONRPCException, self.nodes[0].sendrawtransaction, spends_raw[1])\n\n # mempool should have just spend_101:\n assert_equal(self.nodes[0].getrawmempool(), [ spend_101_id ])\n\n # mine a block, spend_101 should get confirmed\n self.nodes[0].generate(1)\n assert_equal(set(self.nodes[0].getrawmempool()), set())\n\n # ... and now height 102 can be spent:\n spend_102_id = self.nodes[0].sendrawtransaction(spends_raw[1])\n assert_equal(self.nodes[0].getrawmempool(), [ spend_102_id ])\n\nif __name__ == '__main__':\n MempoolSpendCoinbaseTest().main()\n","repo_name":"Internet-of-People/iop-blockchain","sub_path":"qa/rpc-tests/mempool_spendcoinbase.py","file_name":"mempool_spendcoinbase.py","file_ext":"py","file_size_in_byte":1913,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"78"}
+{"seq_id":"18913343628","text":"import requests\nimport os\nfrom twilio.rest import Client\n\nBERLIN_LAT = \"52.509079\"\nBERLIN_LONG = \"13.298040\"\nTHALASSERY_LAT = \"11.753000\"\nTHALASSERY_LONG = \"75.493400\"\n\n#\n\nowa_endpoint = os.environ['OPEN_WEATHER_ENDPOINT']\nowa_token = os.environ['OPEN_WEATHER_API_TOKEN']\ntwilio_account_sid = os.environ['TWILIO_ACCOUNT_SID']\ntwilio_auth_token = os.environ['TWILIO_AUTH_TOKEN']\ntwilio_phone_no = os.environ['TWILIO_PHONE_NO']\nclien_phone_no = os.environ['CLIENT_PHONE_NO']\n\n\nparameters = {\n \"lat\":THALASSERY_LAT,\n \"lon\":THALASSERY_LONG,\n \"exclude\":\"current,minutely,daily\",\n \"appid\":owa_token,\n}\n\nresponse = requests.get(url=owa_endpoint, params=parameters)\nresponse.raise_for_status()\ndata = response.json()[\"hourly\"]\nweather_ids = []\nfor hour in range(12):\n weather_ids.append(data[hour][\"weather\"][0][\"id\"])\n\nwill_rain = False\nfor id in weather_ids:\n if id < 700:\n will_rain = True\nif will_rain:\n client = Client(twilio_account_sid, twilio_auth_token)\n message = client.messages \\\n .create(\n body=\"It's going to rain today. Carry an umbrella\",\n from_=twilio_phone_no,\n to=clien_phone_no\n )\n print(message.status)","repo_name":"rohitrav33ndran/100daysofcode","sub_path":"day35/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1180,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"2997119417","text":"'''\nAugust 25, 2020\nSudoku Solver using Backtracking Algorithm\nAngela Zhang\n\nbacktracking algorithm logic\n1. pick an empty box\n2. try all numbers\n3. find one that works\n4. repeat\n5. as soon as solution cannot be completed, backtrack\n - erase current, go back to previous one and try others\n - if not, keep going back and try others until it does\n'''\n\n# empty slots denoted by 0\nboard = [\n\n # valid board\n [5, 3, 0, 0, 7, 0, 0, 0, 0],\n [6, 0, 0, 1, 9, 5, 0, 0, 0],\n [0, 9, 8, 0, 0, 0, 0, 6, 0],\n [8, 0, 0, 0, 6, 0, 0, 0, 3],\n [4, 0, 0, 8, 0, 3, 0, 0, 1],\n [7, 0, 0, 0, 2, 0, 0, 0, 6],\n [0, 6, 0, 0, 0, 0, 2, 8, 0],\n [0, 0, 0, 4, 1, 9, 0, 0, 5],\n [0, 0, 0, 0, 8, 0, 0, 7, 9]\n\n # invalid board\n # [7,8,0,4,0,0,1,2,0],\n # [6,0,0,0,7,5,0,0,9],\n # [0,0,0,6,0,1,0,7,8],\n # [0,0,7,0,4,0,2,6,0],\n # [0,8,1,0,5,0,9,3,0],\n # [9,0,4,0,6,0,0,0,5],\n # [0,7,0,3,0,0,0,1,2],\n # [1,2,0,0,0,7,4,0,0],\n # [0,4,9,2,0,6,0,0,7]\n\n]\n\ndef solve_board(b):\n # base case\n found = find_empty(b)\n if not found:\n return True # board is solved\n\n row = found[0]\n col = found[1]\n\n for i in range(1, 10): # tries numbers 1-9\n if is_valid(b, i, found):\n b[row][col] = i # update board if valid number\n\n if solve_board(b): # recursive calls solve_board() on new board\n return True\n\n b[row][col] = 0 # backtrack if not viable\n\n return False # cannot solve\n\n\ndef print_board(b): # prints out board for visualization\n n = len(b) # dimension of board\n\n for row in range(n):\n print_row = \"| \" # row to be printed, reset every row\n if row % 3 == 0 and row != 0:\n print(\"-------------------------------------\")\n for column in range(n):\n print_row += (str(b[row][column]) + \" \")\n if (column + 1) % 3 == 0:\n print_row += \"| \"\n print(print_row)\n\ndef is_valid(b, value, pos):\n n = len(b) # dimension of board\n r = pos[0] # row ind of value\n c = pos[1] # column ind of value\n\n # check row\n for i in range(n):\n if b[r][i] == value and i != c:\n return False\n\n # check column\n for i in range(n):\n if b[i][c] == value and i != r:\n return False\n\n # check 3 x 3\n box_x = c // 3\n box_y = r // 3 # gives us x,y coordinates of 3x3 box of value\n for row in range(box_y*3, box_y*3 + 3):\n for column in range(box_x*3, box_x*3 + 3):\n if b[row][column] == value and (row, column) != pos:\n return False\n\n return True\n\n\ndef find_empty(b): # helper function that returns tuple with coordinates of empty slot, false if no empty\n n = len(b) # dimension of board\n\n for row in range(n):\n for column in range(n):\n if b[row][column] == 0:\n return (row, column)\n\n return False\n\n\nprint(\"original board:\")\nprint_board(board)\nsolve_board(board)\nprint(\"\\n\" + \"solution:\")\nif not find_empty(board):\n print_board(board)\nelse:\n print(\"invalid board. unable to solve\")","repo_name":"azhang4216/sudoku-solver","sub_path":"sudoku_solver.py","file_name":"sudoku_solver.py","file_ext":"py","file_size_in_byte":3077,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"21106535532","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nSetup for the SixGill.\n\nSource:: https://github.com/ampledata/sixgill\n\"\"\"\n\n\n__title__ = 'sixgill'\n__version__ = '0.0.1b1'\n__author__ = 'Greg Albrecht '\n__license__ = 'Apache License, Version 2.0'\n__copyright__ = 'Copyright 2016 Orion Labs, Inc.'\n\n\nimport os\nimport setuptools\nimport sys\n\n\ndef publish():\n \"\"\"Function for publishing package to pypi.\"\"\"\n if sys.argv[-1] == 'publish':\n os.system('python setup.py sdist upload')\n sys.exit()\n\n\npublish()\n\n\nsetuptools.setup(\n name='sixgill',\n version=__version__,\n description='SixGill',\n author='Greg Albrecht',\n author_email='gba@orionlabs.io',\n packages=['sixgill'],\n package_data={'': ['LICENSE']},\n license=open('LICENSE').read(),\n long_description=open('README.rst').read(),\n url='https://github.com/ampledata/sixGill',\n setup_requires=[\n 'coverage >= 3.7.1',\n 'httpretty >= 0.8.10',\n 'nose >= 1.3.7'\n ],\n install_requires=[\n 'pynmea2 >= 1.4.2',\n 'pyserial == 2.7',\n 'requests >= 2.7.0'\n ],\n package_dir={'sixgill': 'sixgill'},\n zip_safe=False,\n include_package_data=True,\n entry_points={'console_scripts': ['sixgill = sixgill.cmd:cli']}\n)\n","repo_name":"ampledata/sixgill","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1270,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"10815289548","text":"import json\n\nimport jsonpath as jsonpath\nimport requests\n\nfrom config.config import get_env, collection_get\n\n\nclass WeatherUtil:\n env = get_env()\n weather_config_key = 'QWEATHER_' + env\n location_path = collection_get(weather_config_key, 'LOCATION_PATH')\n weather_path = collection_get(weather_config_key, 'WEATHER_PATH')\n air_path = collection_get(weather_config_key, 'AIR_PATH')\n key = collection_get(weather_config_key, 'KEY')\n\n def get_weather_now(self, location_id: int):\n response = requests.get(self.weather_path, params={'location': location_id, 'key': self.key})\n if response.status_code == 200:\n data = response.text\n jsondata = json.loads(data)\n weather = jsondata['daily']\n return weather\n return None\n\n def get_air_now(self, location_id: int):\n response = requests.get(self.air_path, params={'location': location_id, 'key': self.key})\n if response.status_code == 200:\n data = response.text\n jsondata = json.loads(data)\n air = jsondata['daily']\n return air\n return None\n\n def create_weather_str(self, city_name: str, air, weather, day: str):\n weatherStr = f\"\"\n if day == '今天':\n weatherStr = \"今日 \"\n weather = weather[0]\n air = air[0]\n elif day == '明天':\n weatherStr = \"明天 \"\n weather = weather[1]\n air = air[1]\n elif day == '后天':\n weatherStr = \"后天 \"\n weather = weather[2]\n air = air[2]\n if city_name is not None:\n weatherStr = weatherStr + f\"{city_name} 天气:{weather['textDay']}\\n\" \\\n f\" ⬆️ 最高气温:{weather['tempMax']}°C\\n\" \\\n f\" ⬇️ 最低气温:{weather['tempMin']}°C\\n\" \\\n f\" 🌞 紫外线等级:{weather['uvIndex']}\\n\" \\\n f\" 🌪 最高风速:{weather['windScaleDay']}级\\n\" \\\n f\" 🌫 空气质量:{air['category']}\\n\" \\\n f\" 🔢 空气指数:{air['aqi']}\"\n return weatherStr\n\n def get_weather_str(self, city_name: str, day: str):\n location_id = self.getLocationId(city_name)\n weather = self.get_weather_now(location_id)\n air = self.get_air_now(location_id)\n\n return self.create_weather_str(weather=weather, air=air, city_name=city_name, day=day)\n\n def getLocationId(self, city):\n response = requests.get(self.location_path, params={'location': city, 'key': self.key})\n if response.status_code == 200:\n return json.loads(response.text)['location'][0]['id']\n","repo_name":"JasonSTong/chatgpt-empower-wechat","sub_path":"util/http_utils.py","file_name":"http_utils.py","file_ext":"py","file_size_in_byte":2806,"program_lang":"python","lang":"en","doc_type":"code","stars":55,"dataset":"github-code","pt":"78"}
+{"seq_id":"26141093798","text":"import speech_recognition as sr\nimport tts\n\ndef listen():\n global recognizer\n try:\n with sr.Microphone() as source:\n tts.speak('You may now speak')\n audio = recognizer.record(source,duration=5)\n text = recognizer.recognize_google(audio)\n except Exception as e:\n print(e)\n return \"stop\"\n return text\n\nrecognizer = sr.Recognizer()","repo_name":"samip-gyawali/gpt-assistant","sub_path":"stt.py","file_name":"stt.py","file_ext":"py","file_size_in_byte":394,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"14068354970","text":"import numpy as np\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom Logistic_Regression import *\r\nfrom metrics import *\r\nfrom sklearn.linear_model import LogisticRegression as SKLR\r\nfrom sklearn.datasets import load_breast_cancer\r\n\r\nnp.random.seed(42)\r\n\r\nN = 30\r\nP = 2\r\nX = pd.DataFrame(np.random.randn(N, P))\r\ny = pd.Series(np.random.randint(2, size=N)) #integer because y and y_pred must have same class\r\nX = (X - X.min()) / (X.max() - X.min()) #normalized\r\n\r\nfit_intercept = True\r\n\r\n#L1 type\r\nprint(\"\\nRegularized Logistic Regression (L1 type): \")\r\nLR = LogisticRegression(fit_intercept=fit_intercept)\r\nLR.fit_autograd(X, y, n_iter = 200, batch_size = 5, lr = 1.5, reg_type = \"l1_reg\") \r\ny_hat = LR.predict(X)\r\nprint('theta = ', LR.coef_)\r\nprint('Accuracy = ', accuracy(y_hat, y))\r\nLR.plot_decision_boundary(X,y)\r\n\r\n#L2 type\r\nprint(\"\\nRegularized Logistic Regression (L2 type): \")\r\nLR = LogisticRegression(fit_intercept=fit_intercept)\r\nLR.fit_autograd(X, y, n_iter = 200, batch_size = 5, lr = 1.5, reg_type = \"l2_reg\") \r\ny_hat = LR.predict(X)\r\nprint('theta = ', LR.coef_)\r\nprint('Accuracy = ', accuracy(y_hat, y))\r\nLR.plot_decision_boundary(X,y)","repo_name":"DhruvinPatel31/ML_assignment_3","sub_path":"ML-Assignment-3/q2_reg_lr.py","file_name":"q2_reg_lr.py","file_ext":"py","file_size_in_byte":1167,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"30874528156","text":"import os\nimport cv2\nfrom glob import glob\nfrom tqdm import tqdm\nfrom joblib import Parallel,delayed\n\nfrom engine import utils\nfrom constants import NIH_DIR,CXRAY_SZ\n\n\ndef resize_nih_images():\n \n def resize_save(src):\n name = src.split('/')[-1].replace('.png','.jpg')\n dst = '%snih_images_320/%s'%(NIH_DIR,name) \n if os.path.exists(dst):\n return True\n utils.create_folder(dst)\n try:\n img = cv2.imread(src,0)\n img = cv2.resize(img,(CXRAY_SZ,CXRAY_SZ))\n status = cv2.imwrite(dst,img)\n return status\n except Exception as e:\n print(src,e)\n return False\n\n paths = glob('%simages/*.png'%NIH_DIR)\n utils.create_folder('%snih_images_320/'%(NIH_DIR))\n results = Parallel(n_jobs=-1,prefer='threads')(delayed(resize_save)(path) for path in tqdm(paths,desc='Preparing NIH Data'))\n utils.unique(results)\n\ndef main():\n resize_nih_images()","repo_name":"mcintoshML/Data-Bias-Analysis","sub_path":"data_prep/prep_nih_data.py","file_name":"prep_nih_data.py","file_ext":"py","file_size_in_byte":964,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"42864775379","text":"import numpy as np\nimport torch\nfrom pytorch_tabnet.pretraining import TabNetPretrainer\nfrom sklearn.model_selection import StratifiedKFold, KFold\nfrom torch import nn\nimport wandb\nfrom pytorch_tabnet.callbacks import Callback\n\nfrom augmentations.augmentations import Augmentator\nfrom pipelines.common_pipeline import CommonPipeline\nfrom utils.datasets import CommonDataset\n\n\nclass WandbTabNetCallback(Callback):\n def __init__(self, wandb_obj, params):\n super().__init__()\n self.wandb = wandb_obj\n wandb.config.update({\n 'parameters': params\n })\n\n def set_params(self, params):\n super().set_params(params)\n\n def set_trainer(self, model):\n super().set_trainer(model)\n\n def on_epoch_begin(self, epoch, logs=None):\n super().on_epoch_begin(epoch, logs)\n if logs is not None:\n wandb.log(logs)\n\n def on_epoch_end(self, epoch, logs=None):\n super().on_epoch_end(epoch, logs)\n if logs is not None:\n wandb.log(logs)\n\n def on_batch_begin(self, batch, logs=None):\n super().on_batch_begin(batch, logs)\n\n def on_batch_end(self, batch, logs=None):\n super().on_batch_end(batch, logs)\n\n def on_train_begin(self, logs=None):\n super().on_train_begin(logs)\n\n def on_train_end(self, logs=None):\n super().on_train_end(logs)\n if logs is not None:\n wandb.log(logs)\n\n\nclass TabNetClassificationPipeline(CommonPipeline):\n\n def __init__(self, model, dataset: CommonDataset, wandb_obj, wandb_project_name, BEST_MODELS_FOLDER, **kwargs):\n super().__init__(model, dataset, wandb_obj, wandb_project_name, BEST_MODELS_FOLDER, **kwargs)\n\n def launch_full_cv(self,\n augmentation_types_with_proportions: dict[str, float],\n augmentators: dict[str, Augmentator],\n metric_names_mapping: dict[str, (str, str)],\n n_splits=5,\n seed=7575):\n kf = StratifiedKFold(n_splits=n_splits, random_state=seed, shuffle=True)\n results = self._launch_cv(kf, augmentation_types_with_proportions, augmentators)\n self._log_summary(augmentation_types_with_proportions, results, metric_names_mapping)\n return results\n\n def _train_model(self, train_idx, test_idx):\n wandb_callback = WandbTabNetCallback(self.wandb_obj, self.model.get_params())\n self.model_copy.fit(\n self.dataset[train_idx][0], self.dataset[train_idx][1],\n eval_set=[(self.dataset[test_idx][0], self.dataset[test_idx][1])],\n loss_fn=nn.CrossEntropyLoss(),\n callbacks=[wandb_callback],\n **self.kwargs\n )\n return self.model_copy.history\n\n def _save_model_copy(self, path):\n self.model_copy.save_model(path)\n\n\nclass TabNetRegressionPipeline(CommonPipeline):\n\n def __init__(self, model, dataset: CommonDataset, wandb_obj, wandb_project_name, BEST_MODELS_FOLDER, **kwargs):\n super().__init__(model, dataset, wandb_obj, wandb_project_name, BEST_MODELS_FOLDER, **kwargs)\n\n def launch_full_cv(self,\n augmentation_types_with_proportions: dict[str, float],\n augmentators: dict[str, Augmentator],\n metric_names_mapping: dict[str, str],\n n_splits=5,\n seed=7575):\n kf = KFold(n_splits=n_splits, random_state=seed, shuffle=True)\n results = self._launch_cv(kf, augmentation_types_with_proportions, augmentators)\n self._log_summary(augmentation_types_with_proportions, results, metric_names_mapping)\n return results\n\n def _train_model(self, train_idx, test_idx):\n wandb_callback = WandbTabNetCallback(self.wandb_obj, self.model.get_params())\n self.model_copy.fit(\n self.dataset[train_idx][0], self.dataset[train_idx][1],\n eval_set=[(self.dataset[test_idx][0], self.dataset[test_idx][1])],\n loss_fn=nn.MSELoss(),\n callbacks=[wandb_callback],\n **self.kwargs\n )\n return self.model_copy.history\n\n def _save_model_copy(self, path):\n self.model_copy.save_model(path)\n\n\nclass TabNetPretrainingPipeline:\n def __init__(self, dataset: CommonDataset, wandb_obj,\n wandb_project_name, model_save_path, **kwargs):\n self.model = None\n self.model_save_path = model_save_path\n self.wandb_project_name = wandb_project_name\n self.wandb_obj = wandb_obj\n self.dataset = dataset\n self.kwargs = kwargs\n\n def launch_pretraining(self, pretraining_ratio=0.8, SEED=7575):\n self.wandb_obj.init(self.wandb_project_name, 'fdcf')\n np.random.seed(SEED)\n length = len(self.dataset.X)\n N_valid = int(length * 0.2)\n valid_idx = np.random.choice(length, N_valid, replace=False)\n X_train = np.delete(self.dataset.X, valid_idx, axis=0)\n X_valid = self.dataset.X[valid_idx]\n self.model = TabNetPretrainer(\n optimizer_fn=torch.optim.Adam,\n optimizer_params=dict(lr=2e-2),\n mask_type=\"entmax\"\n )\n wandb_callback = WandbTabNetCallback(self.wandb_obj, self.model.get_params())\n self.model.fit(\n X_train=X_train,\n eval_set=[X_valid],\n pretraining_ratio=pretraining_ratio,\n callbacks=[wandb_callback],\n **self.kwargs\n )\n self.model.save_model(self.model_save_path)\n\n","repo_name":"Zhekuson/TabnetResearch","sub_path":"code/pipelines/tabnet_pipelines.py","file_name":"tabnet_pipelines.py","file_ext":"py","file_size_in_byte":5524,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"78"}
+{"seq_id":"21145161240","text":"#!/usr/bin/python\n#-*- encoding: UTF-8 -*-\n\nfrom ctypes import *\nckipws = None\n\ndef initial(main_dll, py_dll, ini):\n global ckipws\n c_main_dll = c_wchar_p(main_dll)\n c_ini = c_wchar_p(ini)\n ckipws = CDLL(py_dll)\n ckipws.Initial(c_main_dll,c_ini)\n\ndef segment(inputStr, mode = 0):\n global ckipws\n Result = ''\n try:\n CResult = ckipws.Segment(inputStr)\n CResult = cast(CResult,c_wchar_p)\n Result = CResult.value\n except:\n pass\n finally:\n if mode == 0:\n WSResult = []\n Result = Result.split()\n for res in Result:\n re = res.strip()\n re = res[0:len(res)-1]\n temp = re.split(u'(')\n word = temp[0]\n pos = temp[1]\n WSResult.append((word,pos))\n #[('蔡英文', 'Nb'), ('是', 'SHI'), ...]\n return WSResult\n else:\n #蔡英文(Nb) 是(SHI) 中華民國(Nc)...\n return Result\n\ndef segList(corpus):\n # #指定 CKIPWS 統系統檔, 請勿修改\n main_dll = 'CKIPWS.dll'\n py_dll = 'PY_CKIPWS.dll'\n # 指定 CKIPWS 的設定檔\n ini = 'ws.ini'\n\t# 進行 CKIPWS 初始化的動作\n initial(main_dll, py_dll, ini)\n \n ans = []\n for text in corpus:\n ans.append(segment(text))\n # 結果在 Result 中\n return ans\n","repo_name":"666XD/simpleParser","sub_path":"CKIPWS.py","file_name":"CKIPWS.py","file_ext":"py","file_size_in_byte":1374,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"23927982516","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Client',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('client_name', models.CharField(max_length=100)),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='Product',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('product_name', models.CharField(max_length=200)),\n ('price', models.IntegerField(default=0)),\n ('delivery_date', models.DateField(verbose_name=b'date delivered')),\n ('client', models.ForeignKey(to='sales.Client')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n ]\n","repo_name":"sprzedwojski/django-erp","sub_path":"sales/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":1133,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"78"}
+{"seq_id":"31161943966","text":"##chrome diansour game##\nimport pygame\nimport random\nfrom pygame.locals import *\n\n\npygame.init()\n\nclock = pygame.time.Clock()\nfps = 100\n\nscreenwidth = 1000\nscreenhieght = 936\ntrack_scroll = 0\nmove_speed = 4\nscreen = pygame.display.set_mode((screenwidth, screenhieght))\npygame.display.set_caption('chrome dino')\n\nrun =True\nwhile run :\n\t\n\tclock.tick(fps)\n\t\n\t\n\tbg = pygame.image.load('Bg.png')\n\tscreen.blit (bg,(0,0))\n\ttk = pygame.image.load('Track.png')\n\tscreen.blit (tk,(track_scroll,1227))\n\ttrack_scroll -= move_speed\n\tif abs(track_scroll) > 500 :\n\t\t\ttrack_scroll = 0\n\t\n\t\n\t\n\tpygame.display.update()","repo_name":"Legend891/chrome-trex","sub_path":"chrome diansaur.py","file_name":"chrome diansaur.py","file_ext":"py","file_size_in_byte":598,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"4229704183","text":"from flask import Flask, render_template, request, redirect, url_for, flash, Markup, jsonify\nimport datetime\nimport RPi.GPIO as GPIO\nimport Adafruit_ADS1x15\n\nstate = True\n\nadc = Adafruit_ADS1x15.ADS1115()\napp = Flask(__name__)\n\n\n@app.route('/')\ndef index():\n print(\"start webpage\")\n return render_template('index.html')\n\n\n@app.route('/updateTime')\ndef updateTime():\n print('update time and button change ads val to perc')\n\n now = datetime.datetime.now()\n timeString = \"TIME: \" + now.strftime(\"%H:%M:%S\")\n global state\n OutputType = \"value\" if state else \"percent\"\n OutputValue = adc_val(state)\n UpdateTimeOnweb = {\n 'type': OutputType,\n 'value': OutputValue,\n 'timeupdate': timeString\n }\n return jsonify(**UpdateTimeOnweb)\n\n\n@app.route('/changeState')\ndef changeState():\n global state\n state = not state\n print(state)\n return jsonify(state=state)\n\n\ndef GPIO_init(GPIO):\n global state\n global max_val\n max_val = 1\n state = True\n GPIO.setmode(GPIO.BOARD)\n GPIO.setwarnings(False)\n # GPIO.setup(P_OUT, GPIO.OUT)\ndef adc_val(state):\n val = max(adc.read_adc(0),0)\n global max_val \n max_val = max(val,max_val )\n\n return val if state else int(val/max_val*100)\n \n\nif __name__ == '__main__':\n\n GPIO_init(GPIO)\n app.run(debug=True, host='0.0.0.0', port=80)\n","repo_name":"etnk125/Raspi","sub_path":"asg09_lab-webserver/lab/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1355,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"18497243671","text":"import mysql.connector\r\nfrom henryInterfaceClasses import Author, Book, Branch, Price, Category, Cat_Book, Cat_Price, Publisher, Pub_Book, Pub_Price\r\n\r\nclass HenryDAO():\r\n def __init__(self):\r\n self.mydb = mysql.connector.connect(\r\n user='root',\r\n passwd='root',\r\n database='comp3421',\r\n host='127.0.0.1')\r\n\r\n self.mycur = self.mydb.cursor()\r\n\r\n def getAuthorData(self):\r\n # Perform the query\r\n sql = \"SELECT distinct(A.author_num), author_first, author_last FROM henry_author A LEFT JOIN henry_wrote W ON A.AUTHOR_NUM = W.AUTHOR_NUM WHERE W.AUTHOR_NUM IS NOT NULL ORDER BY w.author_num\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n # Display the results\r\n for row in self.mycur:\r\n data.append(Author(row[0], row[1], row[2]))\r\n return data\r\n\r\n def getBookData(self, auth_number):\r\n sql = f\"SELECT b.book_code, title, price FROM HENRY_BOOK b JOIN HENRY_WROTE w ON b.book_code = w.book_code JOIN HENRY_AUTHOR a ON w.AUTHOR_NUM = a.AUTHOR_NUM WHERE (w.BOOK_CODE = b.BOOK_CODE) AND (w.AUTHOR_NUM like '{auth_number}')\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Book(row[0], row[1], row[2]))\r\n return data\r\n\r\n def getBranchData(self, book_code):\r\n sql = f\"select B.branch_name, I.on_hand from HENRY_BRANCH B JOIN HENRY_INVENTORY I ON B.BRANCH_NUM = I.BRANCH_NUM where B.BRANCH_NUM=I.BRANCH_NUM and I.BOOK_CODE='{book_code}'\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Branch(row[0], row[1]))\r\n return data\r\n\r\n def getPriceData(self, BOOK_CODE, AUTHOR_NUM):\r\n sql = f\"SELECT price from HENRY_BOOK B JOIN HENRY_WROTE W ON W.BOOK_CODE = B.BOOK_CODE WHERE (W.BOOK_CODE = '{BOOK_CODE}') AND (W.AUTHOR_NUM = '{AUTHOR_NUM}')\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Price(row[0]))\r\n return data\r\n\r\n def getCategory(self):\r\n sql = \"SELECT DISTINCT(type) From HENRY_BOOK\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Category(row[0]))\r\n return data\r\n\r\n def getCategory_Book(self, TYPE):\r\n sql = f\"SELECT title, book_code, price FROM HENRY_BOOK WHERE (type = '{TYPE}')\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Cat_Book(row[0], row[1], row[2]))\r\n return data\r\n\r\n def cat_getPriceData(self, BOOK_CODE):\r\n sql = f\"SELECT price from HENRY_BOOK WHERE BOOK_CODE = '{BOOK_CODE}'\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Cat_Price(row[0]))\r\n return data\r\n\r\n def getPublisher(self):\r\n sql = \"SELECT P.PUBLISHER_NAME, P.PUBLISHER_CODE FROM HENRY_publisher P LEFT OUTER JOIN henry_book B ON P.publisher_code = B.publisher_code GROUP BY B.PUBLISHER_CODE HAVING B.publisher_code is not null\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Publisher(row[0], row[1]))\r\n return data\r\n\r\n def getPublisher_Book(self, PUBLISHER_CODE):\r\n sql = f\"SELECT title, book_code, price FROM HENRY_BOOK WHERE (PUBLISHER_CODE = '{PUBLISHER_CODE}')\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Pub_Book(row[0], row[1], row[2]))\r\n return data\r\n\r\n def pub_getPriceData(self, BOOK_CODE, PUBLISHER_CODE):\r\n sql = f\"SELECT price from HENRY_BOOK WHERE (BOOK_CODE = '{BOOK_CODE}') AND (PUBLISHER_CODE = '{PUBLISHER_CODE}')\"\r\n self.mycur.execute(sql)\r\n\r\n data = []\r\n for row in self.mycur:\r\n data.append(Pub_Price(row[0]))\r\n return data\r\n\r\n def close(self):\r\n self.mydb.commit()\r\n self.mydb.close()\r\nobj = HenryDAO()","repo_name":"KAPILDESAI007/Academic-Projects","sub_path":"henryDAO.py","file_name":"henryDAO.py","file_ext":"py","file_size_in_byte":4031,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"26682633908","text":"#-*- coding = utf-8 -*-\n#@Time: 17:43\n#@Author : floyd\n#@File :.py\n#@Software :PyCharm\nimport requests\nimport json\nfrom fake_useragent import UserAgent\n\nua=UserAgent()\ndef fanyi(word):\n data={\n 'from':'en',\n 'to':'zh',\n 'query':word,\n 'transtype':'translang',\n 'simple_means_flag':'3',\n 'sign':'704513.926512',\n 'token':'2b03bb02439027e78bb0e65809fdd5c5',\n 'domain':'common',\n 'raw_trans':word\n }\n proxies={\n 'HTTP' : '113.194.28.125:9999'\n }\n header={\n\n 'user-agent':str(ua.random),\n 'x-requested-with':'XMLHttpRequest'\n }\n url=\"https://fanyi.baidu.com/multitransapi\"\n res=requests.post(url,data=data,headers=header,proxies=proxies).text\n res=json.loads(res)\n result=res['data']['cands'][0]\n return result\n\nif __name__==\"__main__\":\n word=''\n while word != 'q':\n word = input ('请输入:')\n if word != 'q':\n print(fanyi(word))\n\n\n","repo_name":"saintkobe/bigbug","sub_path":"翻译/百度翻译.py","file_name":"百度翻译.py","file_ext":"py","file_size_in_byte":935,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"24122068639","text":"from fastapi import APIRouter, Depends, Response, status\nfrom psycopg import Connection\n\nfrom app.database.connection import get_db_connection\nfrom app.schemas.appointment import AppointmentCreate\nfrom app.usecases.appointment import AppointmentUseCases\n\nrouter = APIRouter(tags=[\"Appointments\"], prefix=\"/appointments\")\n\n\n@router.get('')\ndef list_all(\n db_connection: Connection = Depends(get_db_connection)\n):\n uc = AppointmentUseCases(db_connection)\n appointments = uc.get_all_appointments()\n return appointments\n \n@router.post('')\ndef create_appointment(\n appointment: AppointmentCreate,\n db_connection: Connection = Depends(get_db_connection)\n):\n uc = AppointmentUseCases(db_connection)\n uc.create_appointment(appointment)\n return Response(status_code=status.HTTP_201_CREATED)\n\n@router.get('/{appointment_id}')\ndef get_appointment(\n appointment_id: int,\n db_connection: Connection = Depends(get_db_connection)\n):\n uc = AppointmentUseCases(db_connection)\n appointment = uc.get_appointment(appointment_id)\n return appointment\n\n@router.put('/{appointment_id}')\ndef update_appointment(\n appointment_id: int,\n appointment: AppointmentCreate,\n db_connection: Connection = Depends(get_db_connection)\n):\n uc = AppointmentUseCases(db_connection)\n uc.update_appointment(appointment_id, appointment)\n return Response(status_code=status.HTTP_204_NO_CONTENT)\n\n@router.delete('/{appointment_id}')\ndef delete_appointment(\n appointment_id: int,\n db_connection: Connection = Depends(get_db_connection)\n):\n uc = AppointmentUseCases(db_connection)\n uc.delete_appointment(appointment_id)\n return Response(status_code=status.HTTP_204_NO_CONTENT)\n","repo_name":"Jackson-Vieira/MindMeet-PADB","sub_path":"app/routes/appointment.py","file_name":"appointment.py","file_ext":"py","file_size_in_byte":1708,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"3774531967","text":"from scipy import stats\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport requests\nimport json\n\nASE_DEFS = {\n '11820001': 'ASE_NL01',\n '11820002': 'ASE_NL02',\n '11820003': 'ASE_NL03',\n '11820004': 'ASE_NL04',\n '11820005': 'ASE_NL05',\n}\n\nGAS_DEFS = {\n 'co': 'carbon_monoxide__air_',\n 'no': 'nitrogen_monoxide__air_',\n 'no2': 'nitrogen_dioxide__air_',\n 'o3': 'ozone__air_',\n}\n\nSITE_DEFS = {\n 'breuk-sw': 'breukelen_snelweg',\n 'nijm-ruy': 'nijmegen_ruyterstraat',\n}\n\n# InfluxDB creds\nINFLUX_DB_URL = 'https://test.smartemission.nl/influxdb/query?pretty=true'\nINFLUX_DB = 'smartemission'\nINFLUX_DB_READ_USER = 'smart'\nINFLUX_DB_READ_PW = 'smart'\nMEAS_RIVM = 'rivm'\nMEAS_ASE = 'joserefined'\n\n\n# Get calibrated or reference data from InfluxDB\ndef get_data(db, measurement, component, station, time_start, time_end):\n q = \"SELECT time,component,station,value from %s WHERE time >= '%s' AND time <= '%s' AND station = '%s' AND component = '%s'\" % (measurement, time_start, time_end, station, component)\n payload = {'q': q, 'db': db}\n auth = (INFLUX_DB_READ_USER, INFLUX_DB_READ_PW)\n\n # Remote Influx DB query\n r = requests.get(INFLUX_DB_URL, auth=auth, params=payload)\n json_data = json.loads(r.text)\n\n values_arr = None\n try:\n values_arr = json_data['results'][0]['series'][0]['values']\n except:\n pass\n\n # We need values hashtable indexed by datetime for later skipping\n values_dict = {}\n for value in values_arr:\n values_dict[value[0]] = value\n return values_arr, values_dict\n\n\n# Make regression plots for specific date-period, a site, ASE stations, and gasses\ndef make_plots(date_start, date_end, site, ase_stations, gasses):\n\n # Go through all gasses and then for each gas RIVM and all AirSensEUR stations\n for gas in gasses:\n\n # Time formatting: complete ISO Datetime\n time_start = '%sT%sZ' % (date_start, '00:00:00')\n time_end = '%sT%sZ' % (date_end, '00:00:00')\n\n # Get RIVM (reference) ASE data from InfluxDB\n rivm_val_arr, rivm_val_dict = get_data(INFLUX_DB, MEAS_RIVM, GAS_DEFS[gas], SITE_DEFS[site], time_start, time_end)\n\n for ase in ase_stations:\n # Get calibrated (calculated) ASE data from InfluxDB\n se_val_arr, se_val_dict = get_data(INFLUX_DB, MEAS_ASE, gas, ase, time_start, time_end)\n\n x, y = [], []\n for se_val in se_val_arr:\n try:\n # only use entries where for both hour-values exist, otherwise skip\n rivm_val = rivm_val_dict[se_val[0]]\n except:\n # No RIVM value for this hour, skip\n continue\n\n y.append(se_val[3])\n rivm_val_g = rivm_val[3]\n if gas == 'co':\n # RIVM has CO in mg/m3 so convert to ug/m3\n rivm_val_g *= 1000.0\n x.append(rivm_val_g)\n\n # Make a Pandas Dataframe from the x and y single-dim arrays\n d = {'x': x, 'y': y}\n df = pd.DataFrame(data=d)\n\n # R-square calculator from the two x (Ref) and y (ASE) arrays\n def linregress(x, y):\n # Alternative R2 calc, leads to same value\n slope, intercept, r_value, p_value, std_err = stats.linregress(x, y)\n r2 = r_value ** 2.0\n # r2 = stats.pearsonr(x, y)[0] ** 2.0\n return r2, slope\n\n # R-square calculator from the two x (Ref) and y (ASE) arrays\n def r_squared(x, y):\n r2, slope = linregress(x, y)\n return r2\n\n # Plot styling\n sns.set(rc={'axes.grid': True, 'grid.linestyle': '-', \"axes.titlesize\": 9, \"axes.labelsize\": 9, 'grid.color': '.8', 'patch.edgecolor': 'w', \"xtick.major.size\": 4, \"ytick.major.size\": 4})\n sns.set(font_scale=0.8)\n\n # Draw the plot\n # sns.lmplot(x='x', y='y', data=df)\n # sns.scatterplot(x=\"x\", y=\"y\", data=df)\n g = sns.PairGrid(df, y_vars=[\"y\"], x_vars=[\"x\"], height=5)\n g.map(sns.scatterplot, s=50)\n # g.set(ylim=(0,120), yticks=[0, 10, 20, 30, 40, 50, 60, 70, 80, 90,100,110,120])\n\n # More styling\n r2, slope = linregress(x, y)\n\n plt.title(\"%s - %s - %s - %s to %s - R2=%.3f m=%.2f\" % (gas.upper(), site, ASE_DEFS[ase], date_start, date_end, r2, slope))\n plt.xlabel('RIVM Ref (SOS) - 1h avg - ug/m3')\n plt.ylabel('ASE Calc - 1h avg - ug/m3')\n # plt.show()\n\n # Save to png file\n plt.savefig('airsenseur/%s-%s-%s-%s-%s.png' % (site, gas, ASE_DEFS[ase], date_start, date_end))\n plt.close()\n\n\nif __name__ == '__main__':\n make_plots('2018-09-10', '2018-10-09', 'breuk-sw', ASE_DEFS.keys(), ['no', 'no2', 'o3'])\n make_plots('2018-12-25', '2019-01-24', 'nijm-ruy', ['11820001'], ['co', 'no', 'no2', 'o3'])\n","repo_name":"smartemission/smartemission","sub_path":"etl/calibration/airsenseurplots.py","file_name":"airsenseurplots.py","file_ext":"py","file_size_in_byte":5001,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"78"}
+{"seq_id":"43323479564","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Nov 27 16:51:25 2017\n\n@author: tobias\n\"\"\"\nimport os\nimport re\nimport random\nimport subprocess\nfrom Bio import SeqIO\nfrom Bio import AlignIO\nfrom Bio.SeqRecord import SeqRecord\nfrom Bio.Align import MultipleSeqAlignment\nfrom Bio.Alphabet import generic_dna\n\ndef read_fasta(fasta):\n name, seq = None, []\n for line in fasta:\n line = line.rstrip()\n if line.startswith(\">\"):\n if name: yield (name, ''.join(seq))\n name, seq = line, []\n else:\n seq.append(line)\n if name: yield (name, ''.join(seq))\n\ndef get_seq_dict(dictionary, fasta_path):\n new_dict = {}\n for locus in dictionary.keys():\n new_dict.setdefault(locus,[])\n alignment = SeqIO.parse(open(fasta_path),'fasta')\n for seq in alignment:\n ref_loc = str(seq.id).split('_')[0]\n new_dict[ref_loc].append(seq)\n seq.name=''\n seq.description=''\n return new_dict\n\n\ndef create_reference_fasta(out_dir,path_to_alignments):\n # Create a list of fasta files from the input directory\n file_list = [fn for fn in os.listdir(path_to_alignments) if fn.endswith(\".fasta\")]\n reference_list = []\n for fasta_alignment in file_list:\n sequence_name = re.sub(\".fasta\",\"\",fasta_alignment)\n orig_aln = os.path.join(path_to_alignments,fasta_alignment)\n sep_reference = \"%s/%s\" %(out_dir,fasta_alignment)\n reference_list.append(sep_reference)\n cons_cmd = \"cons -sequence %s -outseq %s -name %s -plurality 0.1 -setcase 0.1\" %(orig_aln,sep_reference,sequence_name)\n os.system(cons_cmd)\n reference = os.path.join(out_dir,\"joined_fasta_library.fasta\")\n join_fastas = \"cat %s/*.fasta > %s\" %(out_dir,reference)\n os.system(join_fastas)\n return reference\n\n\n\nfasta_file = '/Users/tobias/Desktop/cos2.fasta'\n\nlocus_bait_dict = {}\nwith open(fasta_file) as f:\n for name, seq in read_fasta(f):\n locus_name = re.sub('>','',name.split('_')[0])\n locus_bait_dict.setdefault(locus_name,[])\n locus_bait_dict[locus_name].append(seq)\n\nlocus_fasta_dict = get_seq_dict(locus_bait_dict,fasta_file)\n\n \n\nout_path = '/Users/tobias/Desktop/merging_probes/sequence_files'\nfor locus in locus_fasta_dict:\n filename = '%s_sequences.fasta' %locus\n with open(os.path.join(out_path,filename), \"w\") as out_file:\n seq_list = locus_fasta_dict[locus]\n index = 0\n for sequence in seq_list:\n sequence.id = '%s_%i' %(locus,index)\n sequence.name=''\n sequence.description='' \n index += 1\n out_file.write(sequence.format('fasta'))\n\n# align the sequence fasta files\naln_path = '/Users/tobias/Desktop/merging_probes/alignments'\nfor fasta in os.listdir(out_path):\n fasta_file = os.path.join(out_path,fasta)\n new_file_name = re.sub('_sequences.fasta','_sequence_alignment.fasta',fasta)\n aln = os.path.join(aln_path,new_file_name)\n aln_stdout = open(aln, 'w')\n # run MAFFT on the temp file\n cmd = [\"mafft\",\"--maxiterate\", \"1000\", fasta_file]\n # just pass all ENV params\n proc = subprocess.Popen(cmd,stderr=subprocess.PIPE,stdout=aln_stdout)\n stderr = proc.communicate()\n aln_stdout.close()\n\n\noutput = '/Users/tobias/Desktop/merging_probes/new_reference'\ncreate_reference_fasta(output,aln_path)\n\n","repo_name":"AntonelliLab/seqcap_processor","sub_path":"src/merge_baits_for_each_locus.py","file_name":"merge_baits_for_each_locus.py","file_ext":"py","file_size_in_byte":3374,"program_lang":"python","lang":"en","doc_type":"code","stars":20,"dataset":"github-code","pt":"78"}
+{"seq_id":"22241933551","text":"from PIL import Image\nfrom threading import Thread\nfrom django.conf import settings\nimport pandas as pd\nimport os\nfrom urllib.request import urlopen\nimport os\nimport numpy as np\nimport traceback\nimport logging\n\n\ncomp_logger = logging.getLogger(__name__)\n\n\nclass ResultImageDownloaderThread(Thread):\n\n\tdef __init__(self,\n\t\t\t\t thread_name,\n\t\t\t\t input_df,\n\t\t\t\t image_download_dir):\n\n\t\tThread.__init__(self)\n\t\tself.thread_name = thread_name\n\t\tself.input_df = input_df\n\t\tself.image_download_dir = image_download_dir\n\n\n\tdef run(self):\n\t\tself.input_df.apply(self.process, axis=1)\n\n\n\tdef downloadImage(self,url,file_path):\n\t\tif not os.path.exists(file_path):\n\t\t\t# try:\n\t\t\tf = open(file_path,'wb')\n\t\t\tf.write(urlopen(url).read())\n\t\t\tf.close()\n\t\t\t# except:\n\t\t\t# \treturn \n\t\t\timage = Image.open(file_path)\n\t\t\timage.convert('RGB').save(file_path)\n\n\n\tdef process(self, row):\n\t\tproduct_image_url = row['r_image_url'].split(settings.IMAGE_LIST_SEPARATOR)[0].strip()\n\t\t# r_title = str(row['r_item_name'])\n\t\t# image_name = r_title.replace(' ','_').replace('.','_').replace('/','_').replace('\"','_').replace('\\'','_')\n\t\t# image_name = '{}.jpg'.format(image_name)\n\t\timage_name = row['result_image']\n\n\t\timage_path = os.path.join(self.image_download_dir,image_name)\n\t\ttry:\n\t\t\tif not os.path.exists(image_path):\n\t\t\t\tself.downloadImage(product_image_url,image_path)\n\t\texcept:\n\t\t\ttraceback_error = traceback.format_exc()\n\t\t\tcomp_logger.info(traceback_error)\n\n\n\n\ndef getResultImageMapperFileDF(client_df,\n\t\t\t\t\t\t\t result_image_mapper_file_path):\n\t\"\"\"\n\tCreates a result image mapper file that holds result item name to image file mapping\n\treturn the mapping data frame\n\t\"\"\"\n\t\n\tif os.path.exists(result_image_mapper_file_path):\n\t\tresult_df = pd.read_csv(result_image_mapper_file_path,sep='\\t', encoding='ISO-8859-1')\n\telse:\t\n\t\t# Getting unique product images(single)\n\t\tresult_df = client_df[['r_item_name']].copy()\n\t\tresult_df = result_df.drop_duplicates(subset=['r_item_name'], keep='first')\n\t\ttotal_rows_in_client_file = len(result_df)\n\t\tresult_df['result_image'] = list(map(lambda x: settings.PROJECT_RESULT_IMAGE_PATTERN.format(x),range(1,total_rows_in_client_file+1)))\n\t\tresult_df.to_csv(result_image_mapper_file_path, index=False,sep='\\t', encoding='ISO-8859-1')\n\treturn result_df\n\n\n\n\ndef getResultMultiImageMapperFile(client_df, \n\t\t\t\t\t\t\t\t result_image_mapper_file_path):\n\t\"\"\"\n\tCreates a result image mapper file that holds result item name to image file mapping\n\treturn the mapping data frame\n\t\"\"\"\n\t\n\tif os.path.exists(result_image_mapper_file_path):\n\t\tresult_df = pd.read_csv(result_image_mapper_file_path,sep='\\t', encoding='ISO-8859-1')\n\n\telse:\n\t\t# split single image list to multiple image listing\n\t\tr_image_stack_df = pd.DataFrame(client_df['r_image_url'].str.split(settings.IMAGE_LIST_SEPARATOR).tolist(), \n\t\t\t\t\t\t\t\tindex=[client_df['r_item_name'],client_df['s_sku']]).stack()\n\n\t\tr_image_stack_df = r_image_stack_df.reset_index()\n\n\t\tdel r_image_stack_df['level_2']\n\n\t\tr_image_stack_df.rename(columns={0:'r_image_url'}, inplace=True)\n\t\tr_image_stack_df['r_image_url'] = r_image_stack_df['r_image_url'].str.strip()\n\n\t\t# unique r_image_url listing and integer name mapping\n\t\tr_image_uniq_df = r_image_stack_df[['r_image_url']]\n\t\tr_image_uniq_df = r_image_uniq_df.drop_duplicates(subset=['r_image_url'], keep='first')\n\t\ttotal_rows_in_client_file = len(r_image_uniq_df)\n\t\t# r_image_uniq_df['result_image'] = list(map(lambda x: '{}.jpg'.format(x),range(1,total_rows_in_client_file+1)))\n\t\tr_image_uniq_df['result_image'] = list(map(lambda x: settings.PROJECT_RESULT_IMAGE_PATTERN.format(x),range(1,total_rows_in_client_file+1)))\n\t\t\n\t\t\"\"\"\n\t\tmerging r_image stacked df with unique integer name mapping\n\t\tExploded r_image_url df with integer name mapping\n\t\t\"\"\"\n\n\t\tresult_df = pd.merge(r_image_stack_df, r_image_uniq_df, how='inner',on=['r_image_url'])\n\t\tresult_df.drop_duplicates(subset=['s_sku','r_item_name','r_image_url'], keep='first', inplace=True)\n\n\t\t# There is no redundant product combination(redundancy may occur due to SERP KEY used) in this df persistance\n\t\tresult_df.to_csv(result_image_mapper_file_path, index=False,sep='\\t', encoding='ISO-8859-1')\n\t\t\n\treturn result_df\n\n\n\n\ndef main(client_input_file_path, \n\t\t image_download_dir,\n\t\t result_image_mapper_file_path):\n\n\tclient_df = pd.read_csv(client_input_file_path, sep='\\t', encoding='ISO-8859-1')\n\tclient_df = client_df[client_df['r_image_url'].notnull()]\n\tclient_df = client_df.fillna(value='')\n\n\n\t# result_df = getResultImageMapperFileDF(client_df, result_image_mapper_file_path)\n\t# client_df = pd.merge(client_df, result_df, how='inner', on=['r_item_name'])\n\n\tresult_df = getResultMultiImageMapperFile(client_df=client_df,\n\t\t\t\t\t\t\t\t\t\t\t result_image_mapper_file_path=result_image_mapper_file_path)\n\n\n\tNUMBER_OF_THREAD = settings.PROJECT_IMAGES_DOWNLOAD_THREADS\n\tdf_list = np.array_split(result_df, NUMBER_OF_THREAD)\n\tthread_list = []\n\n\t# Threaded image downloader\n\tfor item in range(0, NUMBER_OF_THREAD):\n\t\tdf_subset = df_list[item]\n\t\tthread_name = 'Thread{}'.format(item)\n\t\tthread = ResultImageDownloaderThread(thread_name=thread_name,\n\t\t\t\t\t\t\t\t\t\t\t input_df=df_subset,\n\t\t\t\t\t\t\t\t\t\t\t image_download_dir=image_download_dir)\n\t\tthread_list.append(thread)\t\t\n\t\tthread.start()\n\n\tclient_df = client_df.iloc[0:0]\n\n\tfor thread in thread_list:\n\t\tthread.join()\n","repo_name":"rohanneps/django","sub_path":"Django-Rest-Framework/rest_app/core/image_download_helper/downloadResultImageHelper.py","file_name":"downloadResultImageHelper.py","file_ext":"py","file_size_in_byte":5292,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"27690856730","text":"# -*- coding: utf-8 -*-\r\nimport pandas as pd \r\nimport numpy as np\r\n\r\nfirst = pd.read_pickle(\"customers_more_than_one_product.pkl\")\r\ndata = pd.read_pickle(\"customers_more_than_one_product.pkl\")\r\n\r\ndata['reviewerID'] = data[\"reviewerID\"].rank(method='dense').astype(int)\r\ndata['asin'] = data[\"asin\"].rank(method='dense').astype(int)\r\ndata['brand'] = data[\"brand\"].rank(method='dense').astype(int)\r\ndata = data.drop([\"reviewTime\",\"reviewText\",\"title\"],1)\r\n\r\n#tahminleme için örneklem sayısını 15000 belirledik\r\nX= data[\"asin\"].values[:15000]\r\nx1=data[\"overall\"].values[:15000]\r\nx2=data[\"subjectivity\"].values[:15000]\r\nx3=data[\"polarity\"].values[:15000]\r\nX=np.vstack((X, x1,x2,x3)).T\r\n\r\ny= data[\"brand\"].values[:15000]\r\n\r\n#Ürünün bilgilerini öğretip kim almıştır gibi bir çıkarımda bulunmaya çalıştık\r\n#error rate 0,61 çıktı.\r\n\r\n#Ürün bilgilerini öğretip hangi brand olduğunu tahminlemeye çalıştık.\r\n#error rate 0.25 çıktı.\r\n\r\n#to split into train test sets\r\nfrom sklearn.model_selection import train_test_split\r\nX_train,X_test , y_train , y_test = train_test_split(X, y, test_size=0.3, random_state=1)\r\n\r\n\r\nfrom sklearn.neural_network import MLPClassifier\r\nimport time\r\n\r\n#Hidden layer size'ı da değiştirerek denedik. En başarılı sonuç alıdğımız bu oldu.\r\nclf= MLPClassifier(hidden_layer_sizes=(16,16), max_iter=1000)\r\n\r\nstart = time.time()\r\nclf.fit(X_train, y_train)\r\nstop = time.time()\r\n\r\naccuracy=clf.score(X_test, y_test)\r\ny_pred = clf.predict(X_test)\r\nerror_rate = 1 - accuracy\r\n\r\ntraining_time = (stop - start)\r\n\r\nprint(\"\\nTraining Time (in ms): \",training_time)\r\nprint(\"Error (cost): \", error_rate)\r\n","repo_name":"almiragurkan/Data-Mining","sub_path":"07-MLPClassifier.py","file_name":"07-MLPClassifier.py","file_ext":"py","file_size_in_byte":1650,"program_lang":"python","lang":"tr","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"9770940093","text":"import unittest2\nfrom sparktestingbase.sqltestcase import SQLTestCase\n\nfrom datetime import datetime\nfrom pyspark.sql import Row\nfrom pyspark.sql.types import StructType\n\n\nclass SimpleSQLTest(SQLTestCase):\n \"\"\"A simple test.\"\"\"\n\n def test_empty_expected_equal(self):\n allTypes = self.sc.parallelize([])\n df = self.sqlCtx.createDataFrame(allTypes, StructType([]))\n self.assertDataFrameEqual(df, df)\n\n def test_simple_expected_equal(self):\n allTypes = self.sc.parallelize(\n [\n Row(\n i=1,\n s=\"string\",\n d=1.0,\n lng=1,\n b=True,\n list=[1, 2, 3],\n dict={\"s\": 0},\n row=Row(a=1),\n time=datetime(2014, 8, 1, 14, 1, 5),\n )\n ]\n )\n df = allTypes.toDF()\n self.assertDataFrameEqual(df, df)\n\n @unittest2.expectedFailure\n def test_dif_schemas_unequal(self):\n allTypes1 = self.sc.parallelize([Row(d=1.0)])\n allTypes2 = self.sc.parallelize([Row(d=\"1.0\")])\n self.assertDataFrameEqual(allTypes1.toDF(), allTypes2.toDF(), 0.0001)\n\n\nif __name__ == \"__main__\":\n unittest2.main()\n","repo_name":"alexott/spark-playground","sub_path":"testing/src/python/test-spark-testing-base.py","file_name":"test-spark-testing-base.py","file_ext":"py","file_size_in_byte":1267,"program_lang":"python","lang":"en","doc_type":"code","stars":20,"dataset":"github-code","pt":"78"}
+{"seq_id":"29750331493","text":"import matplotlib.pyplot as plt\nfrom matplotlib import colors as mcolors\nimport numpy as np\nimport pandas as pd\nimport math\ncolors = dict(mcolors.BASE_COLORS, **mcolors.CSS4_COLORS)\n\ndef MontarGraficoCurva(paises, dados, ano1, ano2, cores,titulo,eixo_Y): \n #Ele recebe o nome dos paises, e os dados para a criação do grafico em curva\n print(\"Criou o Grafico dos paises \" + str(paises))\n anos = []\n \n #gera o intervalo de anos apartir da entrada do usuario para o eixo x\n for ano in range(ano1, ano2 + 1):\n \tanos.append(ano)\n #gera as curvas do grafico\n for i in range(0, len(paises)):\n plt.plot(anos, dados[i], label = str(paises[i]), color = cores[i])\n\n plt.title(titulo)\n plt.ylabel(eixo_Y)\n plt.xlabel(\"Anos\")\n plt.legend()\n plt.show()\n return\n\n\ndef MontarGraficoBarra(paises, dados, ano1, ano2, cores, titulo,eixo_Y): \n #Ele recebe o nome dos paises, e os dados para a criação do grafico em curva\n print(\"Criou o Grafico dos paises \" + str(paises))\n anos = []\n #gera o intervalo de anos apartir da entrada do usuario para o eixo x\n for ano in range(ano1, ano2 + 1):\n \tanos.append(ano)\n #chama a função com as variações de grafico e em seguida mostra ao usuario\n ax = plt.subplot()\n ax = DefinirAx(ax, np.array(anos), dados, cores, paises)\n\n plt.title(titulo)\n plt.ylabel(eixo_Y)\n plt.xlabel(\"anos\")\n plt.legend(loc = \"best\")\n plt.show()\n return\n\ndef DefinirAx(ax, a, dados, cores, paises): #Função para a formatação do grafico de barras\n width = 0.7 #de Acordo com a quantidade de dados escolhidos pelo usuario\n\n if(len(dados) == 1):\n ax.bar(a,dados[0], width/2, align=\"center\", color=cores[0], label = str(paises[0]))\n\n if(len(dados) == 2):\n ax.bar(a-width/2, dados[0], width/2, align=\"center\", color=cores[0], label = str(paises[0]))\n ax.bar(a,dados[1], width/2, align=\"center\", color=cores[1], label = str(paises[1]))\n\n if(len(dados) == 3):\n ax.bar(a-width/3, dados[0], width/3, align=\"center\", color=cores[0], label = str(paises[0]))\n ax.bar(a,dados[1], width/3, align=\"center\", color=cores[1], label = str(paises[1]))\n ax.bar(a+width/3, dados[2], width/3, align=\"center\", color=cores[2], label = str(paises[2]))\n\n if(len(dados) == 4):\n ax.bar(a-width/4, dados[0], width/4, align=\"center\", color=cores[0], label = str(paises[0]))\n ax.bar(a-width/2, dados[1], width/4, align=\"center\", color=cores[1], label = str(paises[1]))\n ax.bar(a+width/2, dados[2], width/4, align=\"center\", color=cores[2], label = str(paises[2]))\n ax.bar(a+width/4, dados[3], width/4, align=\"center\", color=cores[3], label = str(paises[3]))\n\n if(len(dados) == 5):\n ax.bar(a-width/5, dados[0], width/5, align=\"center\", color=cores[0], label = str(paises[0]))\n ax.bar(a-width/2.5, dados[1], width/5, align=\"center\", color=cores[1], label = str(paises[1]))\n ax.bar(a, dados[2], width/5, align=\"center\", color=cores[2], label = str(paises[2]))\n ax.bar(a+width/2.5, dados[3], width/5, align=\"center\", color=cores[3], label = str(paises[3]))\n ax.bar(a+width/5, dados[4], width/5, align=\"center\", color=cores[4], label = str(paises[4]))\n return ax\n\n","repo_name":"eriklemy/Projeto_CAE","sub_path":"util/MontarGrafico.py","file_name":"MontarGrafico.py","file_ext":"py","file_size_in_byte":3302,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74125039932","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport numpy as np\n\n\n# In[22]:\n\n\ndef f(x):\n\tval=x**4-2*x+1\n\treturn val\n\n#setting endpoints and number of values\nN=1000000\na=0.\nb=5.\n\ndef riemann_dan(f,a,b,N):\n#setting bin width\n\tdelta_x=(b-a)/N\n\n#Create empty list\n\txvals=[]\n\tyvals=[]\n\ts=0\n\n#Loop over all y values\n\tfor i in range(0,N,1):\n#set x value at left edge of bin\n\t\tx=a+i*delta_x\n#set function value of bin\n\t\ty=f(x)\n#append x and y to list\n\t\txvals.append(x)\n\t\tyvals.append(y)\n#increment sum\t\n\t\ts+=y\n\ts*=delta_x\n\treturn s\n\n\ndef trapezoid(f,a,b,N):\n h = (b-a)/N\n fa = f(a)\n fb = f(b)\n y = 0\n for i in range(1,N,2):\n y += (f(a+i*h)+f(a+(i+1)*h))\n s = h*(.5*fa + .5*fb + y)\n return s\n \n \n \ndef simpson(f,a,b,N):\n h = (b-a)/N\n fa = f(a)\n fb = f(b)\n y1 = 0\n y2 = 0\n for i in range(1,N,2):\n y1+= f(a+i*h)\n for i in range(2,N-1,2):\n y2 += f(a+i*h)\n s = 1/3*h*(fa + fb + 4*y1 + 2*y2)\n return s\n \n \n \n\n\n# In[23]:\n\n\nprint(simpson(f,0,2,1000))\n\n\n# In[ ]:\n\n\n\n\n","repo_name":"dgrin1/inclass_hw3_2020","sub_path":"row2ints.py","file_name":"row2ints.py","file_ext":"py","file_size_in_byte":1054,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"19737949612","text":"import requests\nfrom http.cookiejar import LWPCookieJar\nfrom bs4 import BeautifulSoup\nimport pprint\n\nflag = ['typeflag Manhwa', 'typeflag Manga', 'typeflag Manhwa']\n\nclass MangaScrapper:\n def __init__(self, link = 'https://kiryuu.id'):\n self.link = link\n self._requestText = None\n self.session = self._build_session()\n\n def _request(self, link:str):\n self._requestText = self.session.get(link).text\n\n def _build_session(self) -> requests.Session:\n \"\"\"\n Buat session baru\n \"\"\"\n\n session = requests.Session()\n session.headers[\"User-Agent\"] = \"Mozilla/5.0 (Linux; Android 7.0; 5060 Build/NRD90M; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/58.0.3029.83 Mobile Safari/537.36\"\n session.cookies = LWPCookieJar()\n return session\n\n def getHome(self, page = None):\n if page is None:\n self._request(self.link)\n else:\n self._request(self.link+\"/page/{}/\".format(str(page)))\n\n soup = BeautifulSoup(self._requestText, \"html.parser\")\n content = soup.find('div', id=\"content\")\n popular_today = content.find('div', class_='hotslid')\n post_body = content.find('div', class_='postbody')\n # sidebar = content.find('div', id=\"sidebar\")\n\n result = {}\n\n # postbody\n for i, post in enumerate(post_body.find_all('div', class_=\"bixbox\")):\n if not post.find('div', class_=\"series-gen\"):\n releases = post.find('div', class_=\"releases\").find(\"h2\")\n # print(releases.text)\n list_upd = post.find('div', class_=\"listupd\").find_all(\"div\", class_=\"utao\")\n\n result[releases.text] = []\n for upd in list_upd: \n upd_item = {\n \"title\": upd.find('div', class_=\"luf\").a.h4.text,\n \"cover\": upd.find('div', class_=\"imgu\").img['src'],\n \"link\": upd.find('div', class_=\"imgu\").a['href']\n }\n\n if upd.find('div', class_=\"luf\").ul is not None:\n upd_item['type'] = upd.find('div', class_=\"luf\").ul['class'][0]\n else:\n upd_item['type'] = 'Belum Rilis'\n\n result[releases.text].append(upd_item)\n\n elif post.find('div', class_=\"series-gen\"):\n releases = post.find('div', class_=\"releases\").find(\"h2\")\n series_gen = post.find('div', class_=\"series-gen\")\n\n result[releases.text] = {}\n \n # list_upd = series_gen.find(\"div\", class_=\"listupd\").find_all('div')\n \n # print(pprint.pformat(list_upd))\n\n for i, rekomendasi in enumerate(series_gen.find_all('li')):\n try:\n list_upd = series_gen.find(\"div\", class_=\"listupd\").find_all('div', class_=\"tab-pane\")[i]\n except IndexError as e:\n continue\n\n result[releases.text][rekomendasi.a.text] = []\n for bs in list_upd.find_all('div', class_=\"bs\"):\n bsx = bs.find('div', class_=\"bsx\")\n\n if bsx:\n upd_item = {\n \"title\": bsx.a['title'],\n \"cover\": bsx.find('div', class_=\"limit\").img['src'],\n \"link\": bsx.a['href'],\n \"type\": bsx.find('div', class_=\"limit\").find(\"span\", class_=\"type\")['class'][1]\n }\n\n result[releases.text][rekomendasi.a.text].append(upd_item)\n\n # print(\"jumlah dari \" + releases.text, len(result[releases.text]))\n\n # hotlist\n for i, post in enumerate(popular_today):\n releases = post.find('div', class_=\"releases\").find(\"h2\")\n # print(releases.text)\n list_upd = post.find('div', class_=\"listupd\").find_all(\"div\", class_=\"bs\")\n\n result[releases.text] = []\n for upd in list_upd: \n bsx = upd.find('div', class_=\"bsx\")\n\n if bsx:\n upd_item = {\n \"title\": bsx.a['title'],\n \"cover\": bsx.find('div', class_=\"limit\").img['src'],\n \"link\": bsx.a['href'],\n \"type\": bsx.find('div', class_=\"limit\").find(\"span\", class_=\"type\")['class'][1]\n }\n\n result[releases.text].append(upd_item)\n\n # serial populer\n # incomming\n\n print(pprint.pformat(result))\n\n return result\n\n def mangaInfo(self, manga: str):\n manga = manga.replace(f\"{self.link}\", '')\n self._request(self.link + manga)\n\n soup = BeautifulSoup(self._requestText, \"html.parser\")\n post_body = soup.find('div', class_=\"postbody\")\n\n # information\n information = post_body.find(\"div\", class_=\"seriestucon\")\n genres = information.find('div', class_='seriestucontentr').find('div', class_=\"seriestugenre\").find_all('a')\n information_tables = information.find('div', class_='seriestucontentr').find(\"table\", class_=\"infotable\").find('tbody').find_all('tr')\n sinopsis = information.find('div', class_='seriestucontentr').find('div', class_='entry-content').find_all('p')\n\n # episodes\n chapters = post_body.find(\"div\", class_=\"epcheck\").find('ul', class_='clstyle').find_all('li')\n\n # \"\\n\".join(.find('div', class_='entry-content').find_all('p'))\n result = {\n \"title\": information.find(\"div\", class_=\"seriestuheader\").h1.text,\n \"title_alternatif\": information.find(\"div\", class_=\"seriestuheader\").div.text.strip(),\n \"sinopsis\": '\\n'.join([x.text for x in sinopsis]),\n \"information\": {},\n \"chapters\": [],\n \"genres\": []\n }\n\n # chapters loop\n for chapter in chapters:\n chapter_item = {\n \"title\": chapter.find(\"div\", class_=\"eph-num\").a.find(\"span\", class_=\"chapternum\").text,\n \"link\": chapter.find(\"div\", class_=\"eph-num\").a['href'],\n \"updateat\": chapter.find(\"div\", class_=\"eph-num\").a.find(\"span\", class_=\"chapterdate\").text\n }\n\n result['chapters'].append(chapter_item)\n\n # genre loop\n for genre in genres:\n genre_item = {\n \"title\": genre.text,\n \"link\": genre['href']\n }\n result['genres'].append(genre_item)\n\n # information loop\n for informa in information_tables:\n td = informa.find_all('td')\n result['information'][td[0].text.lower().replace(' ', '')] = td[1].text\n\n return result\n\n def getManga(self, manga: str):\n manga = manga.replace(f\"{self.link}\", '')\n self._request(self.link + manga)\n \n soup = BeautifulSoup(self._requestText, \"html.parser\")\n post_area = soup.find('div', class_=\"postarea\")\n\n result = []\n\n for manga_img in post_area.find(\"div\", id=\"readerarea\").find_all('img'):\n result.append(manga_img['src'])\n\n \n return result\n\n def searchManga(self, query: str, page: str = None):\n query = query.replace(f\"{self.link}\", '')\n if page is not None:\n self._request(\"%s/page/%s/?s=%s\" % (self.link, page, query))\n else:\n self._request(\"%s/?s=%s\" % (self.link, query))\n\n soup = BeautifulSoup(self._requestText, \"html.parser\")\n post_body = soup.find(\"div\", class_=\"postbody\")\n \n result = {\n \"manga_result\": []\n }\n\n # scrap for manga\n for manga_search in post_body.find(\"div\", class_=\"listupd\").find_all(\"div\", class_=\"bs\"):\n bsx = manga_search.find(\"div\", class_=\"bsx\")\n\n manga_item = {\n \"title\": bsx.a['title'],\n \"cover\": bsx.find('div', class_=\"limit\").img['src'],\n \"link\": bsx.a['href'],\n \"type\": bsx.find('div', class_=\"limit\").find(\"span\", class_=\"type\")['class'][1],\n \"chapter\": bsx.find('div', class_=\"epxs\").text\n }\n\n result['manga_result'].append(manga_item)\n\n # scrap for pagenation\n pagination = post_body.find(\"div\", class_=\"pagination\")\n max_page = pagination.find_all(\"a\", class_=\"page-numbers\")\n\n if max_page:\n if len(max_page) >= 4:\n max_page = max_page[2]\n result['max_page'] = int(max_page.text)\n elif len(max_page) == 3:\n max_page = max_page[1]\n result['max_page'] = int(max_page.text)\n elif len(max_page) == 2:\n max_page = max_page[0]\n result['max_page'] = int(max_page.text)\n \n return result\n\n def getMangaList(self, option):\n pass\n\n def _isLink(self, link: str):\n print(link)\n return link.startswith('http') or link.startswith('https')\n\nif __name__ == \"__main__\":\n manga = MangaScrapper().getHome()","repo_name":"kevinoctavian/myApp-api","sub_path":"scrapper/mangascrapper.py","file_name":"mangascrapper.py","file_ext":"py","file_size_in_byte":9262,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74571467452","text":"# _*_ coding:utf-8 _*_\n\n'''\n使用传统的机器学习的方法进行文本情感分析\n'''\n\nimport codecs\nimport jieba\nimport numpy as np\n\n# from gensim.models.doc2vec import Doc2Vec\nfrom gensim.models.word2vec import Word2Vec\nfrom sklearn.externals import joblib\nfrom sklearn.svm import SVC\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.ensemble import ExtraTreesClassifier\nfrom sklearn.ensemble import GradientBoostingClassifier\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn import neighbors\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.linear_model import SGDClassifier\n\n\n# 对每个句子的所有词向量取均值\n# text 需要是切完词的 词列表\n# size 一般是词向量的维度\n# word_vector_model: 训练好的词向量模型 (一般使用 gensim 中的 WordVector 进行词向量训练)或者是直接加载训练好的模型\ndef buildWordVector(text, size, word_vector_model):\n vec = np.zeros(size).reshape((1, size))\n count = 0.\n for word in text:\n try:\n vec += word_vector_model[word].reshape((1, size))\n count += 1.\n except KeyError:\n continue\n if count != 0:\n vec /= count\n return vec\n\n\n# 计算词向量\ndef get_train_vecs(x_train, x_test, n_dim):\n '''\n x_train: 训练集\n x_test: 测试集\n n_dim: 训练词向量的维度\n '''\n n_dim = n_dim\n # 初始化模型和生成词汇表\n all_text = x_train + x_test\n text_w2v = Word2Vec(size=n_dim, min_count=5, workers=1)\n text_w2v.build_vocab(all_text)\n text_w2v.train(all_text, total_examples=text_w2v.corpus_count, epochs=5)\n\n # 分别得到训练集和测试集文本的词向量合集,这个数据集就很大了\n train_vecs = np.concatenate([buildWordVector(text, n_dim, text_w2v) for text in x_train])\n # 将训练集的词向量进行保存\n np.save(storedpaths + \"train_vecs.npy\", train_vecs)\n print(\"训练集数据的词向量维度:{}\".format(train_vecs.shape))\n\n test_vecs = np.concatenate([buildWordVector(text, n_dim, text_w2v) for text in x_test])\n # 将测试集的词向量进行保存\n np.save(storedpaths + \"test_vecs.npy\", test_vecs)\n print(\"测试集数据的词向量维度:{}\".format(test_vecs.shape))\n\n # 保存词向量模型,供后面使用预训练的词向量模型预测使用\n text_w2v.save(storedpaths + \"w2v_model.pkl\")\n\n\n# 加载向量化的文本和标签(利用get_train_vecs函数里保存的训练集词向量和测试集词向量)\ndef get_data():\n train_vecs = np.load(storedpaths + 'train_vecs.npy')\n y_train = np.load(storedpaths + 'y_train.npy')\n test_vecs = np.load(storedpaths + 'test_vecs.npy')\n y_test = np.load(storedpaths + 'y_test.npy')\n return train_vecs, y_train, test_vecs, y_test\n\n\n# 训练svm模型并保存模型\ndef svm_train(train_vecs, y_train, test_vecs, y_test):\n clf = SVC(kernel='rbf', verbose=True)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练朴素贝叶斯模型并保存模型\ndef NB_train(train_vecs, y_train, test_vecs, y_test):\n gnb = GaussianNB()\n gnb.fit(train_vecs, y_train)\n joblib.dump(gnb, storedpaths + 'model_gnb.pkl')\n test_scores = gnb.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练决策树模型并保存模型\ndef decision_tree(train_vecs, y_train, test_vecs, y_test):\n clf = DecisionTreeClassifier(max_depth=10, min_samples_split=2, random_state=0)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_dtree.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练随机森林算法并保存模型\ndef random_forest(train_vecs, y_train, test_vecs, y_test):\n clf = RandomForestClassifier(n_estimators=10, max_depth=10, min_samples_split=2, n_jobs=1, random_state=0)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_randomforest.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练 ExtraTreesClassifier 分类算法并保存模型\ndef extract_tree(train_vecs, y_train, test_vecs, y_test):\n clf = ExtraTreesClassifier(n_estimators=10, max_depth=10, min_samples_split=2, n_jobs=1, random_state=0)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_extracttree.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练 GBDT 分类算法并保存模型\ndef gbdt_classifier(train_vecs, y_train, test_vecs, y_test):\n clf = GradientBoostingClassifier(n_estimators=100, learning_rate=1.0, max_depth=10, random_state=0)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_gbdt.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练近邻分类算法并保存模型\ndef nn_classifier(n_neighbors, train_vecs, y_train, test_vecs, y_test):\n clf = neighbors.KNeighborsClassifier(n_neighbors, weights='uniform')\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_nn.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练 LogisticRegression 分类算法并保存模型\ndef LR_classifier(train_vecs, y_train, test_vecs, y_test):\n clf = LogisticRegression(C=50. / len(y_train), multi_class='multinomial',penalty='l1', solver='saga', tol=0.1)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_lr.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练 随机梯度下降 分类算法并保存模型\ndef SGD_classifier(train_vecs, y_train, test_vecs, y_test):\n clf = SGDClassifier(alpha=0.001, max_iter=100)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_sgd.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 训练多层感知机分类算法并保存模型\ndef MP_classifier(train_vecs, y_train, test_vecs, y_test):\n clf = MLPClassifier(solver='lbfgs', alpha=1e-5, hidden_layer_sizes=(5, 2), random_state=1)\n clf.fit(train_vecs, y_train)\n joblib.dump(clf, storedpaths + 'model_mp.pkl')\n test_scores = clf.score(test_vecs, y_test)\n return test_scores\n\n\n# 得到待预测单个句子的词向量\n# 预先进行分词操作\ndef get_predict_vecs(string, n_dim, w2v_model_path):\n '''\n string: 输入的句子\n n_dim: 词向量维度\n w2v_model_path: 預训练词向量的模型路径\n '''\n n_dim = n_dim\n text_w2v = Word2Vec.load(w2v_model_path)\n words = [i for i in jieba.cut(string, cut_all=False)]\n train_vecs = buildWordVector(words, n_dim, text_w2v)\n\n return train_vecs\n\n\n# 调用训练模型进行预测\ndef svm_predict(string, trainmodelpath):\n words_vecs = get_predict_vecs(string, n_dim, w2v_model_path)\n clf = joblib.load(trainmodelpath)\n result = clf.predict(words_vecs)\n\n return result\n\nif __name__ == \"__main__\":\n \"\"\"\n 先训练模型、保存模型再预测模型\n \"\"\"\n # datapaths = \"./data/\"\n # storedpaths = \"./store/\"\n #\n # positive_data = []\n # y_positive = []\n #\n # neutral_data = []\n # y_neutral = []\n #\n # negative_data = []\n # y_negative = []\n #\n # print(\"#------------------------------------------------------#\")\n # print(\"加载数据集\")\n # with codecs.open(datapaths + \"pos.csv\", \"r\", \"utf-8\") as f1, \\\n # codecs.open(datapaths + \"neutral.csv\", \"r\", \"utf-8\") as f2, \\\n # codecs.open(datapaths + \"neg.csv\", \"r\", \"utf-8\") as f3:\n # for line in f1:\n # positive_data.append(\" \".join(i for i in jieba.lcut(line.strip(), cut_all=False)))\n # # y_positive.append([1,0,0])\n # y_positive.append([0])\n # for line in f2:\n # neutral_data.append(\" \".join(i for i in jieba.lcut(line.strip(), cut_all=False)))\n # # y_neutral.append([0,1,0])\n # y_neutral.append([1])\n # for line in f3:\n # negative_data.append(\" \".join(i for i in jieba.lcut(line.strip(), cut_all=False)))\n # # y_negative.append([0,0,1])\n # y_negative.append([2])\n #\n # print(\"positive data:{}\".format(len(positive_data)))\n # print(\"neutral data:{}\".format(len(neutral_data)))\n # print(\"negative data:{}\".format(len(negative_data)))\n #\n # x_text = positive_data + neutral_data + negative_data\n # y_label = y_positive + y_neutral + y_negative\n # print(\"#------------------------------------------------------#\")\n # print(\"\\n\")\n #\n # # 数据集混洗\n # shuffle_indices = np.random.permutation(np.arange(len(y_label)))\n # train_test_percent = 0.2\n #\n # x_train = []\n # x_test = []\n #\n # y_train = []\n # y_test = []\n #\n # for i in shuffle_indices[:-(int(len(shuffle_indices) * train_test_percent))]:\n # x_train.append(x_text[i])\n # y_train.append(y_label[i])\n #\n # for i in shuffle_indices[-(int(len(shuffle_indices) * train_test_percent)):]:\n # x_test.append(x_text[i])\n # y_test.append(y_label[i])\n #\n # x_train_pos = 0\n # x_train_neu = 0\n # x_train_neg = 0\n #\n # x_test_pos = 0\n # x_test_neu = 0\n # x_test_neg = 0\n #\n # for i in y_train:\n # if i[0] == 0:\n # x_train_pos += 1\n # elif i[0] == 1:\n # x_train_neu += 1\n # else:\n # x_train_neg += 1\n #\n # for i in y_test:\n # if i[0] == 0:\n # x_test_pos += 1\n # elif i[0] == 1:\n # x_test_neu += 1\n # else:\n # x_test_neg += 1\n #\n # print(\"#------------------------------------------------------#\")\n #\n # print(\"保存标签数据\") # 这是经过分词处理好后的数据,可以直接使用\n # np.save(storedpaths + \"y_train.npy\", np.array(y_train))\n # np.save(storedpaths + \"y_test.npy\", np.array(y_test))\n #\n # print(\"训练集总数:{}\".format(len(x_train)))\n # print(\"训练集正样本:{}\".format(x_train_pos))\n # print(\"训练集中性样本:{}\".format(x_train_neu))\n # print(\"训练集负样本:{}\".format(x_train_neg))\n #\n # print(\"测试集总数:{}\".format(len(x_test)))\n # print(\"测试集正样本:{}\".format(x_test_pos))\n # print(\"测试集中性样本:{}\".format(x_test_neu))\n # print(\"测试集负样本:{}\".format(x_test_neg))\n #\n # print(\"#------------------------------------------------------#\")\n # print(\"\\n\")\n #\n # # Train model\n # n_dim = 100 # 默认是100维\n # n_neighbors = 10\n #\n # # 获得词向量模型\n # get_train_vecs(x_train, x_test, n_dim)\n #\n # # 获得向量化的文本和标签\n # train_vecs, y_train, test_vecs, y_test = get_data()\n\n # 训练模型并在测试集上验证分数\n # test_scores = svm_train(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"SVM测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n\n # test_scores = NB_train(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"NB测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = nn_classifier(n_neighbors, train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"NN测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = LR_classifier(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"LR测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = SGD_classifier(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"SGD测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = decision_tree(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"TREE测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = random_forest(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"Random_Forest测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = extract_tree(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"Extract_Tree测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = gbdt_classifier(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"GBDT_Tree测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n #\n # test_scores = MP_classifier(train_vecs, y_train, test_vecs, y_test)\n # print(\"#----------------------------------------#\")\n # print(\"MP测试集测试得分:{}\".format(test_scores))\n # print(\"#----------------------------------------#\")\n\n # 预测单个句子\n # string = '最大的困扰是穴位有点把不准,不过我牢记中里先生离穴不离经的教诲。'\n # trainmodelpath = './store/model.pkl'\n # w2v_model_path = './store/w2v_model.pkl'\n # n_dim = 300\n # # n_dim = 100\n # print(svm_predict(string, trainmodelpath))\n\n # 预测\n trainmodelpath = './store/model.pkl'\n w2v_model_path = './store/w2v_model.pkl'\n n_dim = 100\n with open('./data/predict_test.txt',mode='r',encoding='utf-8') as fr:\n with open('./out/predict_out_svm.txt',mode='w',encoding='utf-8') as fw:\n for line in fr.readlines():\n fw.write(str(svm_predict(line,trainmodelpath)[0]) + '\\n')","repo_name":"7125messi/sentiment_analysis_from_raw_corpus","sub_path":"05_sentiment_analysis_ml.py","file_name":"05_sentiment_analysis_ml.py","file_ext":"py","file_size_in_byte":14273,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"78"}
+{"seq_id":"28774800454","text":"N, M, C = [int(i) for i in input().split()]\nBList = [int(i) for i in input().split()]\nALists = [[int(i) for i in input().split()] for _ in range(N)]\n\nans = 0\nfor AList in ALists:\n point = C\n for i, A in enumerate(AList):\n point += A * BList[i]\n if point > 0:\n ans += 1\n\nprint(ans)\n\n","repo_name":"yasunariston/atCoder","sub_path":"ABC/121/B.py","file_name":"B.py","file_ext":"py","file_size_in_byte":305,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"71664336585","text":"n=int(input())\r\nc=0\r\nl1=list(map(int,input().split()))\r\nl2=list(map(int,input().split()))\r\nl1.pop(0)\r\nl2.pop(0)\r\nd={}\r\nfor i in l1:\r\n if i not in d:\r\n d[i]=1\r\n c+=1\r\nfor i in l2:\r\n if i not in d:\r\n d[i]=1\r\n c+=1\r\n\r\nf=0\r\nfor i in range(1,n+1):\r\n if i not in d:\r\n f=1\r\n break\r\nif f==1:\r\n print('Oh, my keyboard!')\r\nelse:\r\n print('I become the guy.')\r\n","repo_name":"sandeep414-boo/codeforces","sub_path":"I Wanna Be the Guy.py","file_name":"I Wanna Be the Guy.py","file_ext":"py","file_size_in_byte":406,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4268429622","text":"import pytest\n\nfrom django import db\nfrom django.core.management import call_command\nfrom django.db.migrations.executor import MigrationExecutor\nfrom unittest import mock\n\nfrom submission import constants\n\n\n@pytest.fixture\ndef erp_zendesk_payload():\n return {\n 'meta': {\n 'action_name': constants.ACTION_NAME_ZENDESK,\n 'email_address': 'erp+testform+testform@gmail.com',\n 'full_name': 'test test',\n 'funnel_steps': [],\n 'ingress_url': 'https://erp.service.dev.uktrade.io/triage/',\n 'sender': {'country_code': None,\n 'email_address': 'erp+testform@jhgk.com'},\n 'service_name': 'erp',\n 'spam_control': {},\n 'subject': 'ERP form was submitted'\n },\n 'data': {\n 'commodities': '190300 - Other',\n 'company_name': 'TASTE OF FOOD',\n 'company_number': '2387892',\n 'company_type': 'LIMITED',\n 'email': 'erp+testform@gmail.com',\n 'employees': '1-10',\n 'employment_regions': ['NORTH_EAST'],\n 'family_name': 'test',\n 'given_name': 'test',\n 'has_market_price_changed': False,\n 'has_market_size_changed': False,\n 'has_other_changes': False,\n 'has_price_changed': False,\n 'has_volume_changed': False,\n 'market_size_known': False,\n 'other_information': 'test',\n 'quarter_four_2018': '0',\n 'quarter_one_2019': '0',\n 'quarter_three_2019': '0',\n 'quarter_two_2019': '0',\n 'sales_volume_unit': 'KILOGRAM',\n 'sector': 'AEROSPACE',\n 'tariff_quota': 'N/A',\n 'tariff_rate': 'N/A',\n 'turnover': '0-25k'\n },\n }\n\n\n@pytest.fixture\ndef email_action_payload():\n return {\n 'data': {\n 'text_body': 'hello there',\n 'html_body': 'Hello there',\n },\n 'meta': {\n 'action_name': constants.ACTION_NAME_EMAIL,\n 'recipients': ['foo@bar.com', 'foo2@bar.com'],\n 'subject': 'Hello',\n 'reply_to': ['email-user@example.com'],\n }\n }\n\n\n@pytest.fixture\ndef zendesk_action_payload():\n return {\n 'data': {\n 'title': 'hello',\n },\n 'meta': {\n 'action_name': constants.ACTION_NAME_ZENDESK,\n 'subject': 'Hello',\n 'full_name': 'Jim Example',\n 'email_address': 'zendesk-user@example.com',\n 'service_name': 'Market Access',\n 'form_url': '/some/form/',\n 'ingress_url': 'https://www.example.com',\n 'sort_fields_alphabetically': True\n },\n }\n\n\n# This payload is to support backward compatibility with gov-notify action\n# which is replaced by gov-notify-email eventually this can be removed.\n@pytest.fixture\ndef gov_notify_action_payload_old():\n return {\n 'data': {\n 'title': 'hello',\n },\n 'meta': {\n 'action_name': 'gov-notify',\n 'template_id': '213123',\n 'email_address': 'notify-user@example.com',\n }\n }\n\n\n@pytest.fixture\ndef gov_notify_email_action_payload():\n return {\n 'data': {\n 'title': 'hello',\n },\n 'meta': {\n 'action_name': constants.ACTION_NAME_GOV_NOTIFY_EMAIL,\n 'template_id': '213123',\n 'email_address': 'notify-user@example.com',\n 'sender': {\n 'country_code': None,\n 'email_address': 'erp+testform@jhgk.com',\n 'ip_address': '252.252.928.233'\n },\n }\n }\n\n\n@pytest.fixture\ndef gov_notify_letter_action_payload():\n return {\n 'data': {\n 'address_line_1': 'The Occupier',\n 'address_line_2': '123 High Street',\n 'postcode': 'SW14 6BF',\n 'name': 'John Smith',\n },\n 'meta': {\n 'action_name': constants.ACTION_NAME_GOV_NOTIFY_LETTER,\n 'template_id': '21312345',\n }\n }\n\n\n@pytest.fixture\ndef pardot_action_payload():\n return {\n 'data': {\n 'title': 'hello',\n },\n 'meta': {\n 'action_name': constants.ACTION_NAME_PARDOT,\n 'pardot_url': 'http://www.example.com/some/submission/path/',\n }\n }\n\n\n@pytest.fixture()\ndef migration(transactional_db):\n \"\"\"\n This fixture returns a helper object to test Django data migrations.\n The fixture returns an object with two methods;\n - `before` to initialize db to the state before the migration under test\n - `after` to execute the migration and bring db to the state after the\n migration. The methods return `old_apps` and `new_apps` respectively; these\n can be used to initiate the ORM models as in the migrations themselves.\n For example:\n def test_foo_set_to_bar(migration):\n old_apps = migration.before([('my_app', '0001_inital')])\n Foo = old_apps.get_model('my_app', 'foo')\n Foo.objects.create(bar=False)\n assert Foo.objects.count() == 1\n assert Foo.objects.filter(bar=False).count() == Foo.objects.count()\n # executing migration\n new_apps = migration.apply('my_app', '0002_set_foo_bar')\n Foo = new_apps.get_model('my_app', 'foo')\n assert Foo.objects.filter(bar=False).count() == 0\n assert Foo.objects.filter(bar=True).count() == Foo.objects.count()\n From: https://gist.github.com/asfaltboy/b3e6f9b5d95af8ba2cc46f2ba6eae5e2\n \"\"\"\n class Migrator:\n def before(self, migrate_from):\n \"\"\" Specify app and starting migration name as in:\n before(['app', '0001_before']) => app/migrations/0001_before.py\n \"\"\"\n\n self.migrate_from = migrate_from\n self.executor = MigrationExecutor(db.connection)\n self.executor.migrate(self.migrate_from)\n self._old_apps = self.executor.loader.project_state(\n self.migrate_from).apps\n return self._old_apps\n\n def apply(self, app, migrate_to):\n \"\"\" Migrate forwards to the \"migrate_to\" migration \"\"\"\n self.migrate_to = [(app, migrate_to)]\n self.executor.loader.build_graph() # reload.\n self.executor.migrate(self.migrate_to)\n self._new_apps = self.executor.loader.project_state(\n self.migrate_to).apps\n return self._new_apps\n\n yield Migrator()\n call_command('migrate')\n\n\n@pytest.fixture(autouse=False)\ndef mock_middleware_test_sig():\n yield mock.patch(\n 'client.helpers.RequestSignatureChecker.test_signature',\n return_value=True,\n ).start()\n","repo_name":"uktrade/directory-forms-api","sub_path":"conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":6927,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41178775038","text":"#Display settings\n\nDEFAULT_IMAGE_SIZE=(300,300)\nFPS=144\nHEIGHT=1000\nWIDTH=1600\nSTART_X,START_Y=0,-300\nX_OFFSET,Y_OFFSET=20,0\n\nBG_IMAGE_PATH='graphics/0/bg.png'\nGAME_INDICES=[1,2,3] #there are 0 and 4 outside of the playing area\nSYMBOLS_PATH='graphics/0/symbols'\n\nFONT_STYLE='graphics/font/RobotoCondensed-LightItalic.ttf'\nFONT_SIZE=32\nBIG_FONT_SIZE=48\n#Slot symbols\nsymbols={\n 'diamond':f\"{SYMBOLS_PATH}/0_diamond.png\",\n 'floppy':f\"{SYMBOLS_PATH}/0_floppy.png\",\n 'hourglass':f\"{SYMBOLS_PATH}/0_hourglass.png\",\n 'seven':f\"{SYMBOLS_PATH}/0_seven.png\",\n 'telephone':f\"{SYMBOLS_PATH}/0_telephone.png\",\n}","repo_name":"destrike00/casino-is-life","sub_path":"settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":614,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"71325910665","text":"from itertools import cycle\n\nfrom individual_circuit import *\n\nclass Population:\n \n def __init__(self):\n self.pop_size = 0\n self.individuals = []\n self.fittest = None\n self.least_fittest = None\n \n \n def fill_population(self, pop_size, target, num_one_gates, num_two_gates, max_moments, n_qubits):\n for i in range(pop_size):\n unique = True\n random_tensor = generate_random_tensor(n_qubits, num_one_gates, num_two_gates, max_moments)\n for j in range(self.pop_size):\n if random_tensor == self.individuals[j].tensor:\n unique = False\n break\n if unique:\n curr_circuit = IndividualCircuit(random_tensor, target, num_one_gates, num_two_gates, n_qubits)\n self.pop_size += 1\n self.individuals.append(curr_circuit)\n else:\n i -= 1\n self.update()\n \n \n \n def add_circuit(self, circuit, increase_size=True):\n self.individuals.append(circuit)\n if increase_size:\n self.pop_size += 1\n self.update()\n \n \n def add_circuits(self, circuits, increase_size=True):\n self.individuals.extend(circuits.individuals)\n if increase_size:\n self.pop_size += len(circuits)\n self.update()\n \n \n def get_fittest_fitness(self):\n self.update()\n return self.fittest.get_fitness()\n \n \n def remove_circuit(self, circuit):\n self.individuals.remove(circuit)\n self.pop_size -= 1\n self.update()\n \n \n def update(self):\n self.individuals.sort(key=lambda circuit:circuit.fitness, reverse=True)\n self.fittest = self.individuals[0]\n self.least_fittest = self.individuals[-1]\n self.pop_size = len(self.individuals)\n \n \n def train_all(self):\n start_time = time.time()\n upper_limit = 500\n iterations = int(self.pop_size/upper_limit)\n new_individuals = []\n \n for j in range(iterations):\n args = []\n gpu_device = cycle(range(8))\n for i in range(upper_limit):\n args.append((self.individuals[i], str(next(gpu_device))))\n\n with mp.get_context(\"spawn\").Pool(processes=upper_limit) as pool:\n results = pool.starmap(circuit_fitness, args)\n\n new_individuals += results\n \n remainder = self.pop_size % upper_limit\n\n args = []\n gpu_device = cycle(range(8))\n for i in range(iterations * upper_limit, iterations * upper_limit + remainder):\n args.append((self.individuals[i], str(next(gpu_device))))\n \n if remainder > 0:\n with mp.get_context(\"spawn\").Pool(processes=remainder) as pool:\n results = pool.starmap(circuit_fitness, args)\n\n new_individuals += results\n \n self.individuals = new_individuals\n \n\n end_time = time.time()\n exec_time = end_time - start_time\n print(\"Execution time multiprocessing {}\".format(exec_time))\n self.update()\n \n \n \n def get_fittests(self, num_fittests):\n \"\"\"\n RETURN: copies of IndividualCircuit Objects\n \"\"\"\n self.fittest = self.individuals[0]\n fittests = self.individuals[:num_fittests]\n return copy.deepcopy(fittests)","repo_name":"yiyiccai/Quantum-Machine-Learning-for-Ansatz-Search","sub_path":"final_code/population.py","file_name":"population.py","file_ext":"py","file_size_in_byte":3468,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"32123826247","text":"#프로그래머스 게임맵최단거리 (level2)\nfrom collections import deque\ndef solution(maps):\n\n x_len=len(maps)\n\n y_len=len(maps[0])\n\n dx=[1,-1,0,0]\n dy=[0,0,1,-1]\n maps[0][0]=0\n queue=deque()\n queue.append((0,0,1))\n while queue:\n x,y,answer=queue.popleft()\n\n for i in range(4):\n next_x = x + dx[i]\n next_y = y + dy[i]\n if 0 <= next_x < x_len and 0 <= next_y < y_len:\n if maps[next_x][next_y] == 1:\n maps[next_x][next_y] = answer+1\n queue.append((next_x,next_y,maps[next_x][next_y]))\n\n if maps[x_len-1][y_len-1]==1:\n return -1\n else:\n return maps[x_len-1][y_len-1]\n","repo_name":"imdduoming/hitalgor","sub_path":"최단거리/게임맵최단거리.py","file_name":"게임맵최단거리.py","file_ext":"py","file_size_in_byte":718,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"8904163026","text":"#!/usr/bin/env python\r\nimport logging\r\nimport os\r\nimport psutil\r\nimport time\r\nimport datetime\r\nimport urllib\r\nimport urllib2\r\nfrom email.mime.text import MIMEText\r\nfrom Keys import Keys as limskeys\r\nimport smtplib\r\nimport traceback\r\nimport subprocess\r\nimport sys\r\nimport glob\r\nimport shutil\r\nimport xml.etree.ElementTree as ET\r\n\r\n# UNIVERSAL\r\nSILENCE_EMAIL = False\r\nTROVAEMON_ID = \"trovaemon\"\r\nTROVAEMON_PWD = \"trovaemon123\"\r\nTROVABASE_JAR = '/mnt/data/Bioinformatics/trovapipe/trovabase/v13/upload.jar'\r\nTROVAPIPE_SCRIPT_LOCATION = '/home/palak/isilon/BI-125-trovapipe-structure-modification/DataAnalysisWrapper_1604v1_0_0.py'\r\nMAX_LOG_SIZE = 500000000 # (500 Mb) \r\n\r\n# TESTING\r\nDATABASE = 'trovabase_dev' # test\r\nUNIFLOW_URL = 'https://trovagene_dev.uniconnect.com:8100/uniflow' # test\r\nREVIEW_CACHE = '/mnt/rnd/Bioinformatics/trovapipe_DEV/review' # test\r\n\r\n# PRODUCTION\r\n#UNIFLOW_URL = 'https://trovagene.uniconnect.com/uniflow' # production\r\n#DATABASE = 'trovabase' # production\r\n#REVIEW_CACHE = '/mnt/prd/review/' # production\r\n\r\n''' Returns a list of paths to MiSeq repositories\r\n'''\r\ndef Repositories() :\r\n\t\r\n\trepositories_file = os.path.join( \"Utilities\" , \"repositories.xml\" )\r\n\t#msg = \"Repo file used \" + repositories_file\r\n\t#logging.info( msg )\r\n\t#msg = \"Repo file used \" + os.path.abspath( repositories_file )\r\n\t#logging.info( msg )\r\n\t\r\n\tif not os.path.exists( repositories_file ) :\r\n\t\tlogging.critical( \"The Utilities/repositories.xml file is missing. Analysis cannot proceed...\" )\r\n\t\tlogging.critical( os.getcwd() )\r\n\t\tsys.exit()\r\n\r\n\ttree = ET.parse( repositories_file )\r\n\troot = tree.getroot()\r\n\t\r\n\trepositories = list()\r\n\t\r\n\tfor child in root:\r\n\t\tif child.tag == \"repository\" :\r\n\t\t\trepositories.append( child.text )\r\n\t\t\t#msg = \"Repo identified \" + child.text\r\n\t\t\t#logging.info( msg )\r\n\t\t\t\r\n\treturn repositories\r\n\t\t\r\n''' Deletes the lock file upon normal exit\r\n'''\r\ndef CleanUp( LOCK_FILE , APP_NAME ) :\r\n\tLogAndEmail('Stopping ' + APP_NAME + ': ' + datetime.datetime.fromtimestamp( time.time() ).strftime( '%Y-%m-%d %H:%M:%S' ) , 0 )\r\n\tif os.path.exists( LOCK_FILE ) : os.remove( LOCK_FILE )\r\n\tlogging.info( 'Complete.\\n')\r\n\r\n''' Central routing for errors and messages in general\r\n Priorities:\r\n\t0: Lowest, just log the message as info\r\n\t1: Low, log the message as info and email bioinformatics\r\n\t2: Medium, log the message as error, email bioinformatics\r\n\t>=3: High, log the message as critical, email bioinformatics & executive level\r\n'''\r\ndef LogAndEmail( message , priority ) :\r\n\r\n\tif priority == 0 :\r\n\t\tlogging.info( message )\r\n\telif priority > 0 :\r\n\t\tlogging.error( message )\r\n\t\tEmail( message , priority )\r\n\t\t\r\n''' Email message to bioinfo\r\n'''\r\ndef Email( message , priority ) :\r\n\r\n\tif SILENCE_EMAIL : return\r\n\t\r\n\tmsg = MIMEText( message )\r\n\r\n\tdistribution_list = os.path.join( \"Utilities\" , \"email.xml\" )\r\n\t\r\n\tpriority_1 = list() # not currently used, mostly for warnings\r\n\tpriority_2 = list() # basic error\r\n\tpriority_3 = list() # moderate error, not used\r\n\tpriority_4 = list() # critical error\r\n\t\r\n\tif not os.path.exists( distribution_list ) :\r\n\t\tlogging.critical( \"Email distribution list is not accessible. The message below was not delivered: \" )\r\n\t\tlogging.critical( message )\r\n\t\tsys.exit()\r\n\t\t\r\n\ttree = ET.parse( distribution_list )\r\n\troot = tree.getroot()\r\n\t\t\r\n\tfor child in root:\r\n\t\tif child.tag == \"Priority-1\" :\r\n\t\t\tfor grandchild in child :\r\n\t\t\t\tif grandchild.tag == \"address\" :\r\n\t\t\t\t\tpriority_1.append( grandchild.text )\r\n\t\telif child.tag == \"Priority-2\" :\r\n\t\t\tfor grandchild in child :\r\n\t\t\t\tif grandchild.tag == \"address\" :\r\n\t\t\t\t\tpriority_2.append( grandchild.text )\r\n\t\telif child.tag == \"Priority-3\" :\r\n\t\t\tfor grandchild in child :\r\n\t\t\t\tif grandchild.tag == \"address\" :\r\n\t\t\t\t\tpriority_3.append( grandchild.text )\r\n\t\telif child.tag == \"Priority-4\" :\r\n\t\t\tfor grandchild in child :\r\n\t\t\t\tif grandchild.tag == \"address\" :\r\n\t\t\t\t\tpriority_4.append( grandchild.text )\r\n\t\r\n\tsend_list = \"\"\r\n\t\r\n\tif priority == 1 :\r\n\t\tsend_list = \",\".join( priority_1 )\r\n\telif priority == 2 :\r\n\t\tsend_list = \",\".join( priority_2 )\r\n\telif priority == 3 :\r\n\t\tsend_list = \",\".join( priority_3 )\r\n\telif priority == 4 :\r\n\t\tsend_list = \",\".join( priority_4 )\r\n\t\t\r\n\tme = \"trovaemon@trovagene.com\"\r\n\r\n\t# use this later to automatically generate tickets?\r\n\t#cc = \"support@trovagene.atlassian.net\"\r\n\t\r\n\tme = \"trovaemon@trovagene.com\"\r\n\tmsg['Subject'] = 'AUTOMATED TROVAPIPE MESSAGE: Error'\r\n\tmsg['From'] = me\r\n\tmsg['To'] = send_list\r\n\t\r\n\ts = smtplib.SMTP('localhost') \r\n\t\r\n\t# s.set_debuglevel(1) # debug output 2 screen\r\n\t\r\n\t# if 'Cc' in msg :\r\n\t\t# s.sendmail( msg[ 'From' ] , msg[ 'To' ].split( ',' ) + msg[ 'Cc' ].split( ',' ), msg.as_string() )\r\n\t# else :\r\n\t\r\n\ts.sendmail( msg[ 'From' ] , msg[ 'To' ].split( ',' ) , msg.as_string() )\r\n\ts.quit()\r\n\t\r\n''' JSON comes back as a complex string, this\r\n converts to dictionary\r\n\t\r\n\tALERT: Exceptions not handled, you must catch\r\n\t- work with uniconnect to dev better schema\r\n'''\r\ndef ParseJSONOneRun( JSON ) :\r\n\t\r\n\ttokens1 = JSON.split( '{' )\r\n\ttokens2 = tokens1[ 1 ].split( '}' )\r\n\ttemp = tokens2[ 0 ].replace( '\\n' , '' )\r\n\tresults = dict()\r\n\tcomma_tokens = temp.split( ',' )\r\n\tfor token in comma_tokens :\r\n\t\tbits = token.replace( '\"' , '' ).split( ':' )\r\n\t\tresults[ bits[ 0 ].strip() ] = bits[ 1 ].strip()\r\n\t\r\n\treturn results\r\n\t\r\n''' JSON comes back as a complex string, this\r\n converts to dictionary of dictionaries:\r\n\tone per run \r\n\t\r\n\tALERT: Exceptions not handled, you must catch\r\n\t- work with uniconnect to dev better schema\r\n'''\r\ndef ParseJSONMultipleRuns( JSON ) :\r\n\tresults = dict()\r\n\ttemp = dict()\r\n\tleft_tokens = JSON.split( '{' )\r\n\tfor left_token in left_tokens :\r\n\t\tif left_token.startswith( '\"' ) :\r\n\t\t\tright_tokens = left_token.split( '}' )\r\n\t\t\tfor right_token in right_tokens :\r\n\t\t\t\tgroups = right_token.replace( '\\n,' , '' ).replace( '])' , '' ).strip().replace( '\\n' , '' ).strip().split( ',' )\r\n\t\t\t\ttemp = {}\r\n\t\t\t\tfor group in groups :\r\n\t\t\t\t\titems = group.split( '\":\"' )\r\n\t\t\t\t\tif len( items ) == 2 :\r\n\t\t\t\t\t\ttemp[ items[ 0 ].replace( '\"' , '' ).strip() ] = items[ 1 ].replace( '\"' , '' ).strip()\r\n\t\t\t\tif 'flowcellID' in temp :\r\n\t\t\t\t\tresults[ temp[ 'flowcellID' ] ] = temp\r\n\treturn results\r\n\t\r\n''' Uploads CSV results file to uniflow\r\n'''\r\ndef UploadRunResultsToUniflow( results_cache , flowcellID , run_folder_name ) :\r\n\r\n\ttry :\r\n\t\r\n\t\t# Verify all results files exist \r\n\t\t\r\n\t\twide_results = \"\" # Homo sapiens readable\r\n\t\traw_results = \"\"\r\n\t\tsample_sheet = \"\"\r\n\t\tstats_results = \"\"\r\n\t\tdb_results = \"\" # database readable\r\n\t\trunqc_results = \"\"\r\n\t\tlog_file = \"\"\r\n\t\t\r\n\t\toutfiles_dir = os.path.join( results_cache, run_folder_name )\r\n\t\treview_cache = os.path.join( REVIEW_CACHE , run_folder_name )\r\n\t\tif os.path.exists( review_cache ) : shutil.rmtree( review_cache )\r\n\t\tos.mkdir( review_cache )\r\n\t\t\r\n\t\t#results_files = glob.glob( os.path.join( results_cache , '*.csv' ) )\r\n\t\t#log_files = glob.glob( os.path.join( results_cache , '*.txt' ) )\r\n\t\t\r\n\t\tqc_files = glob.glob( os.path.join( results_cache , '*.csv' ) )\r\n\t\tfor file in qc_files :\r\n\t\t\tif 'RunQuality' in file :\r\n\t\t\t\trunqc_results = file\r\n\t\t\t\t#shutil.copy( os.path.join( results_cache , file ) , review_cache )\r\n\t\t\t\tshutil.move( os.path.join( results_cache , file ) , outfiles_dir )\r\n\t\t\r\n\t\tresults_files = glob.glob( os.path.join( outfiles_dir , '*.csv' ) )\r\n\t\tfor file in results_files :\r\n\t\t\tif file.endswith( '_raw.csv' ) :\r\n\t\t\t\traw_results = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\telif file.endswith( '_sheet.csv' ) :\r\n\t\t\t\tsample_sheet = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\telif file.endswith( '_stats.csv' ) :\r\n\t\t\t\tstats_results = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\telif file.endswith( '_summary.csv' ) :\r\n\t\t\t\tdb_results = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\telif file.endswith( '_wide.csv' ) :\r\n\t\t\t\twide_results = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\telif 'RunQuality' in file :\r\n\t\t\t\trunqc_results = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\r\n\t\tlog_files = glob.glob( os.path.join( outfiles_dir , '*.txt' ) )\r\n\t\tfor file in log_files :\r\n\t\t\t#if file.startswith( 'log' ) :\r\n\t\t\tif file.endswith( '.txt' ) :\r\n\t\t\t\tlog_file = file\r\n\t\t\t\tshutil.copy( os.path.join( outfiles_dir , file ) , review_cache )\r\n\t\t\t\t\r\n\t\t#wide_results = db_results.replace( '_summary', '' ) # wide results has no '_' extension\r\n\t\t#wide_results = db_results.replace( '_summary', '_wide' ) # wide results has no '_' extension\r\n\t\t\r\n\t\t#if not os.path.exists( wide_results ) or raw_results == \"\" or sample_sheet == \"\" or stats_results == \"\" or db_results == \"\" or runqc_results == \"\" or log_file == \"\" :\r\n\t\tif wide_results == \"\" or raw_results == \"\" or sample_sheet == \"\" or stats_results == \"\" or db_results == \"\" or runqc_results == \"\" or log_file == \"\" :\r\n\t\t\tLogAndEmail( 'Trovapipe Results missing from results cache: ' + outfiles_dir + '\\nwide_results: ' + wide_results + '\\nraw_results: ' + raw_results + '\\nsample_sheet: ' + sample_sheet + '\\nstats_results: ' + stats_results + '\\ndb_results: ' + db_results + '\\nrunqc_results: ' + runqc_results + '\\nlog_file: ' + log_file, 3 )\r\n\t\t\tsys.exit()\r\n\t\t#else : \r\n\t\t#\tshutil.copy( os.path.join( results_cache , wide_results ) , review_cache )\r\n\t\r\n\t\tLogAndEmail( 'Uploading trovapipe results to uniflow: ' + outfiles_dir , 0 )\r\n\t\r\n\t\tbatchID = 'RB' + flowcellID\r\n\t\t\r\n\t\tproc = subprocess.Popen( [ 'curl' , \\\r\n\t\t\t'-F' , 'userId=' + TROVAEMON_ID , \\\r\n\t\t\t'-F' , 'password=' + TROVAEMON_PWD , \\\r\n\t\t\t'-F' , 'stepName=API Result Upload' , \\\r\n\t\t\t'-F' , 'batchId=' + batchID , \\\r\n\t\t\t'-F' , 'flowCellID=' + flowcellID , \\\r\n\t\t\t'-F' , 'status=success' , \\\r\n\t\t\t'-F' , 'formNumber=0' , \\\r\n\t\t\t'-F' , 'Submit=true' , \\\r\n\t\t\t'-F' , 'accountId=Trovagene' , \\\r\n\t\t '-F' , 'csvResultsFile=@' + db_results , \\\r\n\t\t\tUNIFLOW_URL ] , stdout=subprocess.PIPE , stderr=subprocess.PIPE )\r\n\t\t\t\t \r\n\t\tout, err = proc.communicate()\r\n\t\tboth = out + err\r\n\r\n\t\tif \"SYSTEM EXCEPTION\" in both or \"*1\" in both or \"couldn't open file\" in both :\r\n\t\t\tmessage = \"An error occurred while trying to upload a csv results file to UniFlow using curl: \" + db_results + '\\n' + both\r\n\t\t\tLogAndEmail( message , 4 )\r\n\t\t\tsys.exit()\t\r\n\t\t\t\r\n\texcept Exception as detail : # Catch ALL exceptions \r\n\t\tmessage = \"An error occurred while trying to upload a csv results file to UniFlow using curl: \" + db_results + '\\n' + traceback.format_exc()\r\n\t\tLogAndEmail( message , 4 )\r\n\t\r\n''' Gets the state of the run from UNIFlow\r\n by way of a RESTful API\r\n\t- if parsing error return None\r\n\t- if exists return state, if not create and set/return\r\n'''\t\r\ndef GetUNIFlowState( flowcellID ) :\r\n \r\n\tvalues = { \r\n\t'userId' : TROVAEMON_ID, \r\n\t'password' : TROVAEMON_PWD, \r\n\t'stepName' : 'Query Run Status', \r\n\t'Submit' : 'true', \r\n\t'flowcellID' : flowcellID \r\n\t}\r\n\t\r\n\tdata = urllib.urlencode( values )\r\n\trequest = urllib2.Request( UNIFLOW_URL , data )\r\n\tresponse = urllib2.urlopen( request )\r\n\r\n\tJSON = response.read()\r\n\t\t\r\n\tresults = dict()\r\n\t\r\n\ttry :\r\n\t\tresults = ParseJSONOneRun( JSON )\r\n\texcept Exception as detail : # Catch ALL exceptions\r\n\t\tmessage = \"UNIFlow Error - Unable to parse JSON when querying state of flowcellID: \" + flowcellID + \" \" + str( detail ) + \" \" + traceback.format_exc()\r\n\t\tLogAndEmail( message , 3 )\r\n\t\treturn None\r\n\t\t\r\n\tif 'status' not in results or 'flowcellID' not in results :\r\n\t\tmessage = \"UNIFlow Error - Unable to find status or flowcellID results when querying state of run: \" + flowcellID\r\n\t\tLogAndEmail( message , 3 )\r\n\t\treturn None\r\n\t\r\n\telse :\r\n\t\r\n\t\tif results[ 'status' ] == 'NONE' and results[ 'flowcellID' ] == 'NGS_RUN_DOES_NOT_EXIST' :\r\n\t\t\tresponse_code = SetUNIFlowState( flowcellID , limskeys.NGS_RUN_PENDING )\r\n\t\t\tif response_code == 200 : return limskeys.NGS_RUN_PENDING\r\n\t\t\telse :\r\n\t\t\t\tmessage = \"UNIFlow Error - response code was not 200 (not okay) when setting state of run: \" + \\\r\n\t\t\t\tflowcellID + \" to \" + limskeys.NGS_RUN_PENDING\r\n\t\t\t\tLogAndEmail( message , 3 )\t\t\t\t\r\n\t\t\t\treturn None\r\n\t\telse :\r\n\t\t\treturn results[ 'status' ]\r\n\t\r\n''' Returns flowcellID and metadata for runs\r\n with the specified status\r\n'''\r\ndef QueryRunsByStatus( status ) :\r\n\t\r\n\tvalues = {\r\n\t'userId' : TROVAEMON_ID,\r\n\t'password': TROVAEMON_PWD,\r\n\t'stepName': 'Query Runs By Status',\r\n\t'Submit' : 'true', \r\n\t'status' : status\r\n\t}\r\n\t\r\n\turl_encoded_values = urllib.urlencode( values )\r\n\trequest = urllib2.Request( UNIFLOW_URL , url_encoded_values )\r\n\tresponse = urllib2.urlopen( request )\r\n\tJSON = response.read()\r\n\t\r\n\tif \"SYSTEM EXCEPTION\" in JSON :\r\n\t\tmessage = \"Uniflow Error - Error when attempting to query UNIFlow stepName: Query Runs By Status.\\nUNIFlow output: \" + JSON\r\n\t\tLogAndEmail( message , 4 )\r\n\t\treturn None\r\n\t\t\r\n\ttry :\r\n\t\tresults = ParseJSONMultipleRuns( JSON )\r\n\t\treturn results\r\n\texcept Exception as detail : # Catch ALL exceptions\r\n\t\tmessage = \"Uniflow Error - Unable to retrieve run dictionary when parsing JSON \" + JSON + '\\n' + traceback.format_exc()\r\n\t\tLogAndEmail( message , 4 )\r\n\t\treturn None\t\r\n\t\r\n''' Sets the state of an run. If run \r\n does not exist it will be created.\r\n'''\r\ndef SetUNIFlowState( flowcellID , state ) :\r\n\t\r\n\tLogAndEmail( \"Setting \" + flowcellID + \" to \" + state , 0 )\r\n\t\r\n\tvalues = { \r\n\t'userId' : TROVAEMON_ID, \r\n\t'password' : TROVAEMON_PWD, \r\n\t'stepName' : 'Update Run Status', \r\n\t'Submit' : 'true', \r\n\t'flowcellID' : flowcellID , \r\n\t'status' : state \r\n\t}\r\n\t\r\n\turl_encoded_values = urllib.urlencode( values )\r\n\trequest = urllib2.Request( UNIFLOW_URL , url_encoded_values )\r\n\tresponse = urllib2.urlopen( request )\r\n\tcode = response.getcode()\r\n\t\r\n\tif code != 200 :\r\n\t\tmessage = \"TrovapipeUtils~UNIFlow Error - response code was not 200 (not okay) when setting state of run: \" + \\\r\n\t\trunID + \" to \" + state + \" response: \" + str( response.getcode() )\r\n\t\tLogAndEmail( message , 3 )\r\n\t\tsys.exit()\r\n\t\t\r\n\treturn code\r\n\t\r\n''' Determines if this instance is a clone\r\n if so kills the process\r\n'''\r\ndef NoClone( LOCK_FILE ) :\r\n\tpid = str( os.getpid() )\r\n\tpidfile = LOCK_FILE\r\n\t\t\r\n\tif os.path.isfile( pidfile ) :\r\n\t\tif IsDead( pidfile ) :\r\n\t\t\t file( pidfile , 'w' ).write( pid )\r\n\t\t\t logging.warning( \"A lock.pid file existed with a dead process...\" )\r\n\t\telse :\r\n\t\t\tlogging.info( \"Terminating analysis, a process already exists...\" )\r\n\t\t\tsys.exit()\r\n\telse:\r\n\t\tfile( pidfile , 'w' ).write( pid )\t\r\n\t\t\r\n''' Process ID file still exists suggesting this\r\n code is already running. Verify that's true.\t\t\r\n'''\r\ndef IsDead( pid_file ) :\r\n\twith open( pid_file , 'rb' ) as input :\r\n\t\tpid = input.readline().strip()\r\n\t\tif psutil.pid_exists( int( pid ) ) : return False\r\n\t\telse : return True\r\n\t\r\n''' Archives logs when they get too large\r\n'''\r\ndef ArchiveLog( APP_LOG ) :\r\n\t\r\n\tif not os.path.exists( APP_LOG ) : open( APP_LOG , 'w' ).close()\r\n\t\t\r\n\tif os.path.getsize( APP_LOG ) > MAX_LOG_SIZE : \r\n\t\r\n\t\tbase = os.path.basename( APP_LOG ).split('.')[0]\r\n\t\tarchive_dir = os.path.join( os.path.dirname( APP_LOG ) , 'archive' )\r\n\t\t\r\n\t\tindices = [ 0 ]\r\n\t\tfiles = glob.glob( os.path.join( archive_dir , base + \"*\" ) )\r\n\t\tfor file in files :\r\n\t\t\tindices.append( int( file.split('.')[ 1 ] ) )\r\n\t\t\t\r\n\t\tnew_index = max( indices ) + 1\r\n\t\tnew_archive_file = os.path.join( archive_dir , base + '.' + str( new_index ) + '.log' )\r\n\t\tif os.path.exists( APP_LOG ) : shutil.move( APP_LOG , new_archive_file )\r\n","repo_name":"palakpsheth/Trovagene","sub_path":"scripts/RunQC/Utilities/TrovapipeUtils.py","file_name":"TrovapipeUtils.py","file_ext":"py","file_size_in_byte":15558,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12378698834","text":"# ergasia 4\r\ndef plus(b):\r\n\treturn '+'+str(b)\r\n\r\ndef minus(b):\r\n\treturn '-'+str(b)\r\n\r\ndef times(b):\r\n\treturn '*'+str(b)\r\n\r\n\r\n\r\ndef zero(a):\r\n\tif a[0]== '+':\r\n\t\treturn int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn -int(a[1])\r\n\telse : \r\n\t\treturn 0\r\n\r\ndef one(a):\r\n\tif a[0]== '+':\r\n\t\treturn 1+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 1-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 1*int(a[1])\r\n\telse:\r\n\t\treturn 1\r\n\r\ndef tow(a):\r\n\tif a[0]== '+':\r\n\t\treturn 2+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 2-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 2*int(a[1])\r\n\telse:\r\n\t\treturn 2\r\n\r\ndef three(a):\r\n\tif a[0]== '+':\r\n\t\treturn 3+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 3-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 3*int(a[1])\r\n\telse:\r\n\t\treturn 3\r\n\r\ndef four(a):\r\n\tif a[0]== '+':\r\n\t\treturn 4+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 4-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 4*int(a[1])\r\n\telse:\r\n\t\treturn 4\r\n\r\ndef five(a):\r\n\tif a[0]== '+':\r\n\t\treturn 5+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 5-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 5*int(a[1])\r\n\telse:\r\n\t\treturn 5\r\n\r\ndef six(a):\r\n\tif a[0]== '+':\r\n\t\treturn 6+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 6-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 6*int(a[1])\r\n\telse:\r\n\t\treturn 6\r\ndef seven(a):\r\n\tif a[0]== '+':\r\n\t\treturn 7+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 7-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 7*int(a[1])\r\n\telse:\r\n\t\treturn 7\r\n\r\ndef eight(a):\r\n\tif a[0]== '+':\r\n\t\treturn 8+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 8-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 8*int(a[1])\r\n\telse:\r\n\t\treturn 8\r\n\r\ndef nine(a):\r\n\tif a[0]== '+':\r\n\t\treturn 9+int(a[1])\r\n\telif a[0]== '-':\r\n\t\treturn 9-int(a[1])\r\n\telif a[0]==\"*\" : \r\n\t\treturn 9*int(a[1])\r\n\telse:\r\n\t\treturn 9\r\nanswer='1'\r\nwhile answer == '1':\r\n\t\r\n\tc=input('kaleste thn sinaryisi edo : ')\r\n\tleftbrackets=[]\r\n\trightbrackets=[]\r\n\t# στις 2 προηγουμενες λιστες θα αποθηκευονται οι θεσεις των δεξιων και αριστερων παρενθεσεων της μεταβλητης c\r\n\tfor i in range(0,len(c)):\r\n\t\tif c[i] == '(':\r\n\t\t\tleftbrackets.append(i)\r\n\t\telif c[i] == ')' :\r\n\t\t\trightbrackets.append(i)\r\n\r\n\tflag1=0\r\n\tflag2=0\r\n\t# με τις flag1 και flag2 στην συνεχεια θα ελενχετε αν η συναρτηση εχει κληθει σωστα απο τον χρηστη\r\n\tch=''\r\n\r\n\tif len(leftbrackets) ==1 and len(rightbrackets) ==1:\r\n\t\tif c[0:leftbrackets[0]] =='zero':\r\n\t\t\tprint(zero(\" \"))\r\n\t\t\tflag1=1\r\n\t\telif c[0:leftbrackets[0]] =='one':\r\n\t\t\tprint(one(\" \"))\r\n\t\t\tflag1=1\r\n\t\telif c[0:leftbrackets[0]] =='tow':\r\n\t\t\tflag1=1\r\n\t\t\tprint(tow(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='three':\r\n\t\t\tflag1=1\r\n\t\t\tprint(three(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='four':\r\n\t\t\tflag1=1\r\n\t\t\tprint(four(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='five':\r\n\t\t\tflag1=1\r\n\t\t\tprint(five(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='six':\r\n\t\t\tflag1=1\r\n\t\t\tprint(six(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='seven':\r\n\t\t\tflag1=1\r\n\t\t\tprint(seven(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='eight':\r\n\t\t\tflag1=1\r\n\t\t\tprint(eight(\" \"))\r\n\t\telif c[0:leftbrackets[0]] =='nine':\r\n\t\t\tflag1=1\r\n\t\t\tprint(nine(\" \"))\r\n\t\tflag2=1\r\n\r\n\tif len(leftbrackets) ==3 and len(rightbrackets) ==3:\r\n\t\tif c[leftbrackets[0]+1:leftbrackets[1]]=='plus':\r\n\t\t\tch='+'\r\n\t\telif c[leftbrackets[0]+1:leftbrackets[1]]=='minus':\r\n\t\t\tch='-'\r\n\t\telif c[leftbrackets[0]+1:leftbrackets[1]]=='times':\r\n\t\t\tch='*'\r\n\r\n\t\tif c[leftbrackets[1]+1:leftbrackets[2]]=='zero':\r\n\t\t\tch = ch+'0'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='one':\r\n\t\t\tch = ch+'1'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='tow':\r\n\t\t\tch = ch+'2'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='three':\r\n\t\t\tch = ch+'3'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='four':\r\n\t\t\tch = ch+'4'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='five':\r\n\t\t\tch = ch+'5'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='six':\r\n\t\t\tch = ch+'6'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='seven':\r\n\t\t\tch = ch+'7'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='eight':\r\n\t\t\tch = ch+'8'\r\n\t\telif c[leftbrackets[1]+1:leftbrackets[2]]=='nine':\r\n\t\t\tch = ch+'9'\r\n\r\n\t\tif len(ch)==2:\r\n\t\t\tif (ch[0]=='+' or ch[0]=='-' or ch[0]=='*') and int(ch[1])<=9 and int(ch[1])>=0:\r\n\t\t\t\tflag1=1\r\n\r\n\t\tif flag1==1:\r\n\t\t\tif c[0:leftbrackets[0]] =='zero':\r\n\t\t\t\tprint(zero(ch))\r\n\t\t\t\tflag2=1\r\n\t\t\telif c[0:leftbrackets[0]] =='one':\r\n\t\t\t\tprint(one(ch))\r\n\t\t\t\tflag2=1\r\n\t\t\telif c[0:leftbrackets[0]] =='tow':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(tow(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='three':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(three(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='four':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(four(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='five':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(five(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='six':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(six(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='seven':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(seven(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='eight':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(eight(ch))\r\n\t\t\telif c[0:leftbrackets[0]] =='nine':\r\n\t\t\t\tflag2=1\r\n\t\t\t\tprint(nine(ch))\t\r\n\r\n\r\n\tif flag1 == 0 | flag2 == 0:\r\n\t\tprint('h sinartisi den klithike sosta')\r\n\tanswer=input(\"an thelete na epanalavete tin diadikasia patiste to 1 allios patiste kati allo\")\r\n\tpass\r\naa=input(\"press enter to finish\")\t\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","repo_name":"stathisp18056/-1-","sub_path":"ergasia4.py","file_name":"ergasia4.py","file_ext":"py","file_size_in_byte":5190,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5108263672","text":"import textwrap\n\nimport synthtool as s\nfrom synthtool import gcp\nfrom synthtool.languages import python\n\ncommon = gcp.CommonTemplates()\n\ndefault_version = \"v2\"\n\nfor library in s.get_staging_dirs(default_version):\n # Do not expose ModelServiceClient and ModelServiceAsyncClient, as there\n # is no public API endpoint for the models service.\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/__init__.py\",\n r\"from \\.services\\.model_service import ModelServiceClient\",\n \"\",\n )\n\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/__init__.py\",\n r\"from \\.services\\.model_service import ModelServiceAsyncClient\",\n \"\",\n )\n\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/__init__.py\",\n r\"\"\"[\"']ModelServiceClient[\"'],\"\"\",\n \"\",\n )\n\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/__init__.py\",\n r\"\"\"[\"']ModelServiceAsyncClient[\"'],\"\"\",\n \"\",\n )\n\n # Adjust Model docstring so that Sphinx does not think that \"predicted_\" is\n # a reference to something, issuing a false warning.\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/types/model.py\",\n r'will have a \"predicted_\"',\n \"will have a `predicted_`\",\n )\n\n # Avoid breaking change due to change in field renames.\n # https://github.com/googleapis/python-bigquery/issues/319\n s.replace(\n library / f\"google/cloud/bigquery_{library.name}/types/standard_sql.py\",\n r\"type_ \",\n \"type \",\n )\n\n s.move(\n library,\n excludes=[\n \"*.tar.gz\",\n \".coveragerc\",\n \"docs/index.rst\",\n f\"docs/bigquery_{library.name}/*_service.rst\",\n f\"docs/bigquery_{library.name}/services.rst\",\n \"README.rst\",\n \"noxfile.py\",\n \"setup.py\",\n f\"scripts/fixup_bigquery_{library.name}_keywords.py\",\n \"google/cloud/bigquery/__init__.py\",\n \"google/cloud/bigquery/py.typed\",\n # There are no public API endpoints for the generated ModelServiceClient,\n # thus there's no point in generating it and its tests.\n f\"google/cloud/bigquery_{library.name}/services/**\",\n f\"tests/unit/gapic/bigquery_{library.name}/**\",\n ],\n )\n\ns.remove_staging_dirs()\n\n# ----------------------------------------------------------------------------\n# Add templated files\n# ----------------------------------------------------------------------------\ntemplated_files = common.py_library(\n cov_level=100,\n samples=True,\n microgenerator=True,\n split_system_tests=True,\n intersphinx_dependencies={\n \"pandas\": \"http://pandas.pydata.org/pandas-docs/dev\",\n \"geopandas\": \"https://geopandas.org/\",\n },\n)\n\n# BigQuery has a custom multiprocessing note\ns.move(\n templated_files,\n excludes=[\n \"noxfile.py\",\n \"docs/multiprocessing.rst\",\n \".coveragerc\",\n # Include custom SNIPPETS_TESTS job for performance.\n # https://github.com/googleapis/python-bigquery/issues/191\n \".kokoro/presubmit/presubmit.cfg\",\n # Group all renovate PRs together. If this works well, remove this and\n # update the shared templates (possibly with configuration option to\n # py_library.)\n \"renovate.json\",\n ],\n)\n\n# ----------------------------------------------------------------------------\n# Samples templates\n# ----------------------------------------------------------------------------\n\npython.py_samples()\n\ns.replace(\n \"docs/conf.py\",\n r'\\{\"members\": True\\}',\n '{\"members\": True, \"inherited-members\": True}',\n)\n\n# Tell Sphinx to ingore autogenerated docs files.\ns.replace(\n \"docs/conf.py\",\n r'\"samples/snippets/README\\.rst\",',\n '\\\\g<0>\\n \"bigquery_v2/services.rst\", # generated by the code generator',\n)\n\n# ----------------------------------------------------------------------------\n# pytype-related changes\n# ----------------------------------------------------------------------------\n\n# Add .pytype to .gitignore\ns.replace(\".gitignore\", r\"\\.pytest_cache\", \"\\\\g<0>\\n.pytype\")\n\n# Add pytype config to setup.cfg\ns.replace(\n \"setup.cfg\",\n r\"universal = 1\",\n textwrap.dedent(\n \"\"\" \\\\g<0>\n\n [pytype]\n python_version = 3.8\n inputs =\n google/cloud/\n exclude =\n tests/\n google/cloud/bigquery_v2/\n output = .pytype/\n disable =\n # There's some issue with finding some pyi files, thus disabling.\n # The issue https://github.com/google/pytype/issues/150 is closed, but the\n # error still occurs for some reason.\n pyi-error\"\"\"\n ),\n)\n\ns.shell.run([\"nox\", \"-s\", \"blacken\"], hide_output=False)\n","repo_name":"suryaprakashreddy034/python_bq","sub_path":"owlbot.py","file_name":"owlbot.py","file_ext":"py","file_size_in_byte":4773,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"13012516805","text":"import discord\nimport gspread\nfrom oauth2client.service_account import ServiceAccountCredentials\nimport json\n\nclient = discord.Client()\njson_key = 'Gnor Bot json_key.json'\nscope = ['https://spreadsheets.google.com/feeds',\n 'https://www.googleapis.com/auth/drive']\nglobal all_sheet_values\n\ncredentials = ServiceAccountCredentials.from_json_keyfile_name(json_key, scope)\ngc = gspread.authorize(credentials)\nsheet = gc.open_by_url(\"https://docs.google.com/spreadsheets/d/1KrRiRQ_J9laMaZyf9orBZvedbepJXI3LEbaUylMZ1TM\").sheet1\nall_sheet_values = sheet.get_all_values()\n\n@client.event\nasync def on_ready():\n print('Logged in as')\n print(client.user.name)\n print(client.user.id)\n print('------')\n\n@client.event\nasync def on_member_join(member):\n defaultchannel = discord.utils.get(member.server.channels, name = settings['custom']['default_channel'])\n defaultrole = discord.utils.get(member.server.roles, name = settings['custom']['default_role'])\n if defaultchannel is not None and defaultrole is not None:\n await client.add_roles(member, defaultrole)\n await client.send_message(defaultchannel, member.name + ' has joined the server')\n else:\n print('Default channel or default role not found for join notification')\n\n@client.event\nasync def on_member_leave(member):\n defaultchannel = discord.utils.get(member.server.channels, name = settings['custom']['default_channel'])\n if defaultchannel is not None:\n await client.send_message(defaultchannel, member.name + ' has left the server')\n else:\n print('Default channel not found for leave notification')\n\n@client.event\nasync def on_message(message):\n if message.content.startswith('uwu'):\n await client.add_reaction(message, 'gnaruwu:389219155288653837')\n\n elif message.content.startswith('?role '):\n newrole = discord.utils.get(message.author.server.roles, name = message.content[6:])\n if newrole is None:\n await client.send_message(message.channel, 'No such role')\n elif getrolelayer(newrole) > 0:\n await client.send_message(message.channel, 'You do not have the permit to join/leave this role')\n elif(not newrole in message.author.roles):\n await client.add_roles(message.author, newrole)\n await client.send_message(message.channel, 'You joined ' + newrole.name)\n else:\n await client.remove_roles(message.author, newrole)\n await client.send_message(message.channel, 'You left ' + newrole.name)\n \n elif message.content.startswith('?matchup '):\n i = find_sheet_row_by_matchup(message.content[9:])\n if i is not None:\n message_string = ('```Matchup: Gnar vs. ' + all_sheet_values[i][0] +\n '\\nDifficulty: ' + all_sheet_values[i][1] +\n '\\nStat Shards: ' + all_sheet_values[i][8] +\n '\\nStarting Items: ' + all_sheet_values[i][9] +\n '\\nCore Items: ' + all_sheet_values[i][10] +\n '\\n\\nInformation: ' + all_sheet_values[i][11] + '```')\n await client.send_message(message.channel, message_string)\n else:\n await client.send_message(message.channel, 'Could not find matchup info for ' + message.content[9:])\n\n elif message.content.startswith('?refresh matchups') and \\\n 'Mega Gnar' in [role.name for role in message.author.roles]:\n all_sheet_values = sheet.get_all_values()\n await client.send_message(message.channel, 'Updated matchups!')\n\n elif message.content.startswith('?ranks'):\n roles = message.author.server.roles\n member_count = []\n for index in roles:\n member_count.append(0)\n for index1 in message.author.server.members:\n i = 0\n for index2 in roles:\n if index2 in index1.roles:\n member_count[i]+=1\n i+=1\n string = 'Server ranks:\\n'\n i = 0\n for index in roles:\n if not index.is_everyone and getrolelayer(index) == 0:\n string += index.name + \": \" + str(member_count[i]) + \" members\\n\"\n i+=1\n await client.send_message(message.channel, string)\n\n elif message.content.startswith('?assignrole '):\n if getperm(message.author) < 3:\n await client.send_message(message.channel, 'You do not have the permit to access this command')\n return\n arguments = message.content.split(' ', 2)\n role = discord.utils.get(message.author.server.roles, name = arguments[2])\n if role is None:\n await client.send_message(message.channel, 'No such role')\n elif not message.mentions:\n await client.send_message(message.channel, 'No user mentioned')\n elif getperm(message.author) < getrolelayer(role):\n await client.send_message(message.channel, 'You do not have the permit to assign this role')\n elif getperm(message.channel.server.me) <= getrolelayer(role):\n await client.send_message(message.channel, 'The bot does not have the permit to assign this role')\n elif role not in message.mentions[0].roles:\n await client.add_roles(message.mentions[0], role)\n await client.send_message(message.channel, message.mentions[0].name + ' has joined ' + role.name)\n else:\n await client.remove_roles(message.mentions[0], role)\n await client.send_message(message.channel, message.mentions[0].name + ' has left ' + role.name)\n\n else:\n arguments = message.content.split(' ')\n output = ''\n for customcommand in settings['custom_commands']:\n if arguments[0] == customcommand:\n output = settings['custom_commands'][customcommand]\n output = output.replace('{user}', message.author.name)\n output = output.replace('{server}', message.author.server.name)\n output = output.replace('{channel}', message.channel.name)\n if output != '':\n await client.send_message(message.channel, output)\n\n \n \ndef find_sheet_row_by_matchup(champion):\n for index, sublist in enumerate(all_sheet_values, start=3):\n if sublist[0].lower().replace('\\'', '').replace(' ', '') \\\n == champion.lower().replace('\\'', '').replace(' ', ''):\n return index - 3\n\ndef getperm(member):\n maxprio = 0\n j = 1\n while j <= len(settings['perms']):\n for index in member.roles:\n if index.name in settings['perms']['layer' + str(j)]:\n maxprio = j\n j+=1\n return maxprio\n\ndef getrolelayer(role):\n layer = 0\n j = 1\n while j <= len(settings['perms']):\n if role.name in settings['perms']['layer' + str(j)]:\n layer = j\n j+=1\n return layer\n\nwith open('bot_info.txt') as file:\n lines = file.readlines()\n bot = lines[1].strip()\n\nwith open('settings.json') as set:\n settings = json.load(set)\n\nclient.run(bot)\n","repo_name":"nabibyte427/GnorBot","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":7060,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"33042673756","text":"\n\"\"\"\"\n# Created on Tue Mar 21 15:46:39 2017\n\n# @author: keriabermudez\n\n# This script is an example of how to use the cell tracking \n\n\"\"\"\n\nimport cell_segmentation as cellseg\nfrom skimage import io\nimport numpy as np\n\n\n# %%\n#Reading Image Stack\n\n\npath= '/Users/keriabermudez/Dropbox/Projects/Gregory_Brittingham/ERK_KTR/'\n\nimage_dapi = 'ERK_DAPI2 (Converted).mov'\nimage_erk = 'ERK_KTR1 (Converted).mov'\n\npath_results = path +'Results/'\n\n\n#%%\n#crop\ndapi = cellseg.video_to_tif(path+image_dapi,0)\nerk = cellseg.video_to_tif(path+image_erk,1)\ndapi = dapi[:,500:,500:]\nerk = erk[:,500:,500:]\n\n\n# %% \n# Creating ztsack of intesitiy image and color zstack\n\ndapi_color = np.zeros((dapi.shape[0], dapi.shape[1], dapi.shape[2],3), dtype = np.uint8)\ndapi_color[:,:,:,2 ] = dapi.copy()\n\nerk_color = np.zeros((erk.shape[0], erk.shape[1], erk.shape[2],3), dtype = np.uint8)\nerk_color[:,:,:,1 ] = erk.copy()\n#%%\n# 2) Try paremeters in one slice or zlevel\n\nz = 0\nzlevel_image_dapi = dapi[z].copy()\nzlevel_image_erk = erk[z].copy()\n\nzlevel_color_erk = erk_color[z].copy()\nzlevel_color_dapi = dapi_color[z].copy()\n\n#%%\n# Enhance, Blur, and Segment \n\ncl1, gaussian_blur_cl1, segmented_zlevel, centers = cellseg.enhance_blur_segment(zlevel_image_dapi,enhance = False, blur = False, kernel = 7, n_intensities = 2)\nio.imshow(cl1)\nio.imshow(gaussian_blur_cl1)\nio.imshow(segmented_zlevel)\ncenters\n \n#%%\n# Test Region Detection parameters\n\nlabeled = cellseg.watershedsegment(segmented_zlevel,smooth_distance = True,kernel = 3) ##increase kernel size if it is oversegmenting\n\n\n#%% Draw Contours \n\n\nzlevel_image_color_regions_d = cellseg.draw_contours(labeled,zlevel_color_dapi.copy(), with_labels = True, color = (255,0,0),width = 2 )\nzlevel_image_color_regions_e = cellseg.draw_contours(labeled,zlevel_color_erk.copy(), with_labels = True, color = (255,0,0),width = 2 )\n\nio.imshow(zlevel_image_color_regions_d)\nio.imshow(zlevel_image_color_regions_e)\n\nio.imsave(path_results+'Z-level_Regions_'+str(z)+'dapi.tif',zlevel_image_color_regions_d)\nio.imsave(path_results+'Z-level_Regions_'+str(z)+'erk.tif',zlevel_image_color_regions_e)\n\n# Get Measurements\npositions_regions = cellseg.regions_measure(labeled,zlevel_image_erk)\n\n#%%\n# Test Blob Detection parameters\n\nzlevel_image_color_marked = cellseg.draw_blob_log(cl1,zlevel_color_dapi.copy(), with_labels = True,max_sigma=30, min_sigma=20,num_sigma=10,threshold=.01,overlap=0.6,color_blobs = (255,0,0),width =3)\nio.imshow(zlevel_image_color_marked)\nio.imsave(path_results+'Z-level_Blobs_'+str(z)+'.tif',zlevel_image_color_marked)\n\n #Get Measurements\npositions_blobs = cellseg.blob_log_measure(cl1,zlevel_image_erk, max_sigma=30, min_sigma=20,num_sigma=10,threshold=.01,overlap=0.6 )\n\n\n#%%\n# 2) Cell Tracking With Blob Detection, Note: This can take a long time to run\n\nct = cellseg.cell_tracking(erk,dapi,erk_color)\n\n# Setting Segmentation Parameters\nct.set_segment_param(enhance = False, blur = False, n_intensities = 2)\n\n# Setting Blob Parameters\nct.set_blob_param(max_sigma=30,min_sigma=20,num_sigma=10,threshold=.01,overlap=0.6)\n\n# Track with Blob\nct.track_with_blob(min_slices =1,color_blobs= (255,0,0))\nct.draw_trajectories(color_trajectory= (255,255,0) )\n\ntable_positions = ct.positions_table\nio.imsave(path_results+'Blobs.tif',ct.zstack_color)\n\n#table_positions.to_csv(path_results+image[:-4]+'Positions_Table.csv')\n#%%\n# 3) Cell Tracking With Regions\n\nct2 = cellseg.cell_tracking(erk,dapi,dapi_color.copy())\n\n# Setting Segmentation Parameters\nct2.set_segment_param(enhance = False, blur = False, n_intensities = 2)\nct2.set_watershed_param(smooth_distance = True, kernel = 3)\n\n# Track with Blob\nct2.track_with_regions(min_slices =1, color_contours= (255,0,0))\nct2.draw_trajectories(color_trajectory= (255,255,0) )\n\n\n#table_positions.to_csv(path_results+image[:-4]+'Positions_Table.csv')\nio.imsave(path_results+'Regions.tif',ct2.zstack_color)\n\n\n","repo_name":"FenyoLab/cell-segmentation-analysis","sub_path":"video_tracking_greg.py","file_name":"video_tracking_greg.py","file_ext":"py","file_size_in_byte":3886,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"43027905336","text":"import misc.helpers as helpers\n\n# A 2d list that simply stores True if a tile has been seen, or False if it has not\nclass FieldOfView(list):\n def __init__(self, width, height):\n self.width = width\n self.height = height\n for _ in range(width):\n col = []\n for _ in range(height):\n col.append(False)\n self.append(col)\n \n # What each player can currently see, to darken tiles outside of FOV\n self.current = []\n self.reset_current()\n\n def reset_current(self):\n self.current = []\n for _ in range(self.width):\n col = []\n for _ in range(self.height):\n col.append(False)\n self.current.append(col)\n\n def contains(self, x, y):\n return self[x][y]\n \n def can_see(self, x, y):\n return self.current[x][y]\n\n # Add new tiles to the remembered FOV, and then recalculate the current tiles each player can see\n def update(self, world, creature):\n self.update_total(world, creature)\n self.refresh_current(world.players)\n \n def update_all(self, world, players):\n for p in players:\n self.update_total(world, p)\n self.refresh_current(world.players)\n\n def update_total(self, world, creature):\n for x in range(-creature.vision_radius, creature.vision_radius + 1):\n for y in range(-creature.vision_radius, creature.vision_radius + 1):\n to_x, to_y = creature.x + x, creature.y + y\n if to_x < 0 or to_y < 0 or to_x >= self.width or to_y >= self.height:\n continue\n\n if self[to_x][to_y]:\n continue\n\n # If a player can see a floor they can see the walls adjacent to it, so we dont get weird visual errors of \"floating\" walls\n if creature.can_see(to_x, to_y):\n if world.is_floor(to_x, to_y):\n self[to_x][to_y] = True\n if world.is_wall(to_x - 1, to_y):\n self[to_x-1][to_y] = True\n if world.is_wall(to_x, to_y - 1):\n self[to_x][to_y-1] = True\n if world.is_wall(to_x - 1, to_y - 1):\n self[to_x-1][to_y-1] = True\n\n def refresh_current(self, players):\n self.reset_current()\n for p in players:\n self.update_current(p)\n \n def update_current(self, creature):\n for x in range(-creature.vision_radius, creature.vision_radius + 1):\n for y in range(-creature.vision_radius, creature.vision_radius + 1):\n to_x, to_y = creature.x + x, creature.y + y\n if to_x < 0 or to_y < 0 or to_x >= self.width or to_y >= self.height:\n continue\n\n if self.current[to_x][to_y]:\n continue\n \n if creature.can_see(to_x, to_y):\n self.current[to_x][to_y] = True\n\n # If you see a non-active creature, activate all enemies in the room\n c = creature.world.get_creature_at_location(to_x, to_y)\n if c and c.can_be_activated() and not c.is_active():\n if c.home_room:\n c.world.activate_room_enemies(c.home_room, creature)\n else:\n c.activate(creature)\n s = \"You see a \" + c.name\n w = c.equipment.slot(\"Main\")\n if w:\n s += \" wielding a \" + w.name\n creature.notify(s)\n \n def print(self):\n for y in range(self.height):\n for x in range(self.width):\n if self[x][y] == True:\n c = '0'\n else:\n c = '1'\n print(c, end='')\n print()\n\ndef can_see(world, sx, sy, dx, dy, radius):\n if (sx - dx) * (sx - dx) + (sy - dy) * (sy - dy) > radius * radius:\n return False\n\n l = helpers.get_line(sx, sy, dx, dy)\n for p_x, p_y in l:\n if not world.block_sight(p_x, p_y) or (p_x == dx and p_y == dy):\n continue\n return False\n return True","repo_name":"alexs2112/isometric-demo","sub_path":"world/fov.py","file_name":"fov.py","file_ext":"py","file_size_in_byte":3703,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31899189730","text":"from __future__ import annotations\n\nimport json\nfrom typing import TYPE_CHECKING\n\nfrom django.core.exceptions import ObjectDoesNotExist\n\nif TYPE_CHECKING:\n from ..models import SiteProfile\n\n\ndef get_or_create_site_profile_obj(single_site, site_obj, apps) -> SiteProfile | None:\n site_profile_model_cls = apps.get_model(\"edc_sites\", \"SiteProfile\")\n opts = dict(\n title=single_site.description,\n country=single_site.country,\n country_code=single_site.country_code,\n languages=json.dumps(single_site.languages) if single_site.languages else None,\n )\n try:\n site_profile = site_profile_model_cls.objects.get(site=site_obj)\n except ObjectDoesNotExist:\n site_profile = site_profile_model_cls.objects.create(site=site_obj, **opts)\n else:\n for k, v in opts.items():\n setattr(site_profile, k, v)\n site_profile.save()\n return site_profile\n","repo_name":"clinicedc/edc-sites","sub_path":"edc_sites/utils/get_or_create_site_profile_obj.py","file_name":"get_or_create_site_profile_obj.py","file_ext":"py","file_size_in_byte":920,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"36986824302","text":"from FlickrCaptor import FlickrCaptor\nfrom FileManager import FileManager\nfrom datetime import datetime, timedelta\nfrom time import sleep\nimport json\nfrom LoggerBuilder import LoggerBuilder\nfrom ExitLogger import ExitLogger\nimport logging\nimport os\nimport sys\nimport requests\n\nsource = 'flickr'\nlogger = LoggerBuilder(source, logging.WARNING, logging.INFO).get_logger()\n# Limit for Flickr API calls: 3600 queries / h (http://www.flickr.com/services/developer/api/)\n# Limit for flickr.photos.search: Flickr will return at most the first 4,000 results for any given search query.\n\n# min_date should be written using the YYYY-MM-DD format\nmin_date_format = '%Y-%m-%d'\nmin_date_file = \"min_date\"\n\nclass InitError(Exception):\n\tpass\n\ndef main():\t\n\tlast_date = datetime.now() - timedelta(days=2)\n\n\tcaptor = FlickrCaptor(logger)\n\tfile_manager = FileManager()\n\t\n\ttry:\n\t\ttry:\n\t\t\twoe_ids = file_manager.get_locations(source)\n\t\texcept IOError:\n\t\t\traise InitError(\"File %s is missing\"% file_manager.get_locations_path(source))\n\t\texcept ValueError:\n\t\t\traise InitError(\"The %s file does not contain any correct JSON object\"% file_manager.get_locations_path(source))\n\t\t\n\t\tmin_date_file_path = file_manager.get_path(source, None, min_date_file)\n\t\ttry:\n\t\t\tmin_date_json = file_manager.read_json(source, min_date_file)\n\t\t\tmin_date = datetime.strptime(min_date_json['min_date'], min_date_format)\n\t\texcept IOError:\n\t\t\traise InitError(\"File %s is missing. You should create this file and set {'min_date':YYYY-MM-DD} in it.\"% min_date_file_path)\n\t\texcept (ValueError, KeyError):\n\t\t\traise InitError(\"You need to set {\\\"min_date\\\":\\\"YYYY-MM-DD\\\"} in file %s\"% min_date_file_path)\n\t\t\n\t\tzero_day = timedelta(days=0) \n\t\tone_day = timedelta(days=1)\n\n\t\tif (last_date - min_date) < zero_day:\n\t\t\traise InitError(\"The date set as min_date in %s is after 2 days ago.\"% min_date_file_path)\n\t\t\n\t\twhile (last_date - min_date) >= zero_day:\n\t\t\tlogger.warning(\"---- Capting for %s\" % min_date.strftime('%y-%m-%d'))\n\t\t\tfor city, woe_id in woe_ids.iteritems():\n\t\t\t\tmax_date = min_date + one_day\n\t\t\t\tlogger.warning(\"Capting for %s\" % city)\n\t\t\t\tloaded_json = captor.get_data(min_date, max_date, woe_id)\n\t\t\t\t\n\t\t\t\ttry:\n\t\t\t\t\tfile_manager.write_json(loaded_json, source, city, min_date.strftime('%y-%m-%d'))\n\t\t\t\texcept IOError:\n\t\t\t\t\traise InitError(\"Folder %s is missing\"% file_manager.get_folder_path(source, city))\n\n\t\t\t\tlogger.warning(\"New JSON written for %s \" % city)\n\t\t\t\n\t\t\tmin_date = max_date\n\t\t\tf = open(min_date_file_path, 'w+')\n\t\t\tjson.dump({'min_date':min_date.strftime(min_date_format)}, f)\n\t\t\tf.close()\n\n\t\t\tlogger.warning(\"... Sleeping for 5 s\")\n\t\t\tsleep(5)\n\t\treturn 0\n\t\t\n\texcept InitError as e:\n\t\tlogger.critical(\"%s: %s\"% (type(e).__name__, e))\n\t\treturn 1\t\n\texcept requests.exceptions.RequestException as e:\n\t\tlogger.critical(\"%s: %s\"% (type(e).__name__, e))\n\t\treturn 2\n\texcept FlickrCaptor.FlickrApiError as e:\n\t\tlogger.critical(\"%s: %s\"% (type(e).__name__, e))\n\t\treturn 3\n\texcept Exception as e:\n\t\tlogger.critical(e, exc_info=True)\n\t\treturn 4\n\nif __name__ == \"__main__\":\n\tstart = datetime.now().strftime('%y-%m-%d %H:%M:%S')\n\tlogger.critical(\"START\")\n\tres = main()\n\tlogger.critical(\"STOP (%d)\"% res)\n\tstop = datetime.now().strftime('%y-%m-%d %H:%M:%S')\n\tExitLogger(source).log(start, stop, res)\n\tsys.exit(res)","repo_name":"ComplexCity/capture","sub_path":"FlickrCapture.py","file_name":"FlickrCapture.py","file_ext":"py","file_size_in_byte":3277,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"41197328053","text":"from django import template\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.urls import reverse\n\nregister = template.Library()\n\n\n@register.filter\ndef rating_score(obj, user):\n \"\"\"\n Returns the score a user has given an object\n \"\"\"\n if not user.is_authenticated or not hasattr(obj, \"_ratings_field\"):\n return False\n\n ratings_descriptor = getattr(obj, obj._ratings_field)\n try:\n rating = ratings_descriptor.get(user=user).score\n except ratings_descriptor.model.DoesNotExist:\n rating = None\n\n return rating\n\n\n@register.filter\ndef has_rated(user, obj):\n \"\"\"\n Returns whether or not the user has rated the given object\n \"\"\"\n return rating_score(obj, user) is not None\n\n\n@register.filter\ndef rate_url(obj, score=1):\n \"\"\"\n Generates a link to \"rate\" the given object with the provided score - this\n can be used as a form target or for POSTing via Ajax.\n \"\"\"\n return reverse(\n \"ratings_rate_object\",\n args=(\n ContentType.objects.get_for_model(obj).pk,\n obj.pk,\n score,\n ),\n )\n\n\n@register.filter\ndef unrate_url(obj):\n \"\"\"\n Generates a link to \"un-rate\" the given object - this\n can be used as a form target or for POSTing via Ajax.\n \"\"\"\n return reverse(\n \"ratings_unrate_object\",\n args=(\n ContentType.objects.get_for_model(obj).pk,\n obj.pk,\n ),\n )\n","repo_name":"django/djangosnippets.org","sub_path":"ratings/templatetags/ratings_tags.py","file_name":"ratings_tags.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","stars":408,"dataset":"github-code","pt":"81"}
+{"seq_id":"18124485031","text":"import logging\nimport math\nimport os\n\nimport paddle\nfrom paddle.amp import GradScaler, auto_cast\nfrom paddle.optimizer import AdamW\nfrom paddlenlp.transformers import (\n BertForQuestionAnswering,\n BertTokenizer,\n ErnieForQuestionAnswering,\n ErnieTokenizer,\n MPNetForQuestionAnswering,\n MPNetTokenizer,\n)\nfrom tqdm import tqdm\n\nfrom args import parse_args\nfrom data import get_dev_dataloader, get_train_dataloader\nfrom metric import compute_prediction, squad_evaluate\nfrom utils import (\n CrossEntropyLossForSQuAD,\n get_scheduler,\n get_writer,\n save_json,\n set_seed,\n)\n\nlogger = logging.getLogger(__name__)\n\n\nMODEL_CLASSES = {\n \"bert\": (BertForQuestionAnswering, BertTokenizer, True),\n \"ernie\": (ErnieForQuestionAnswering, ErnieTokenizer, True),\n \"mpnet\": (MPNetForQuestionAnswering, MPNetTokenizer, False),\n}\n\n\n@paddle.no_grad()\ndef evaluate(model, data_loader, args, output_dir=\"./\"):\n model.eval()\n all_start_logits = []\n all_end_logits = []\n\n for batch in data_loader:\n input_ids, token_type_ids = batch\n start_logits_tensor, end_logits_tensor = (\n model(input_ids, token_type_ids=token_type_ids)\n if args.need_token_type_ids\n else model(input_ids)\n )\n all_start_logits.extend(start_logits_tensor.numpy().tolist())\n all_end_logits.extend(end_logits_tensor.numpy().tolist())\n\n all_predictions, all_nbest_json, scores_diff_json = compute_prediction(\n data_loader.dataset.data,\n data_loader.dataset.new_data,\n (all_start_logits, all_end_logits),\n args.version_2_with_negative,\n args.n_best_size,\n args.max_answer_length,\n args.null_score_diff_threshold,\n )\n\n save_json(all_predictions, os.path.join(output_dir, \"all_predictions.json\"))\n if args.save_nbest_json:\n save_json(all_nbest_json, os.path.join(output_dir, \"all_nbest_json.json\"))\n\n eval_results = squad_evaluate(\n examples=data_loader.dataset.data,\n preds=all_predictions,\n na_probs=scores_diff_json,\n )\n return eval_results\n\n\ndef main(args):\n logging.basicConfig(\n format=\"%(asctime)s - %(levelname)s - %(name)s - %(message)s\",\n datefmt=\"%m/%d/%Y %H:%M:%S\",\n level=logging.INFO,\n handlers=[\n logging.FileHandler(\n os.path.join(os.path.dirname(args.output_dir), \"run.log\"),\n mode=\"w\",\n encoding=\"utf-8\",\n )\n ],\n )\n logger.info(\"********** Configuration Arguments **********\")\n for arg, value in sorted(vars(args).items()):\n logger.info(f\"{arg}: {value}\")\n logger.info(\"**************************************************\")\n paddle.set_device(args.device)\n set_seed(args)\n writer = get_writer(args)\n\n # get model and tokenizer\n model_class, tokenizer_class, args.need_token_type_ids = MODEL_CLASSES[\n args.model_type\n ]\n model = model_class.from_pretrained(args.model_name_or_path)\n if args.use_huggingface_tokenizer and args.model_type == \"mpnet\":\n from transformers import MPNetTokenizerFast\n\n tokenizer = MPNetTokenizerFast.from_pretrained(\"./\")\n else:\n tokenizer = tokenizer_class.from_pretrained(args.model_name_or_path)\n\n # get dataloader\n train_dataloader = get_train_dataloader(tokenizer, args)\n dev_dataloader = get_dev_dataloader(tokenizer, args)\n\n num_update_steps_per_epoch = math.ceil(\n len(train_dataloader) / args.gradient_accumulation_steps\n )\n if args.max_train_steps > 0:\n args.num_train_epochs = math.ceil(\n args.max_train_steps / num_update_steps_per_epoch\n )\n else:\n args.max_train_steps = args.num_train_epochs * num_update_steps_per_epoch\n\n # get lr_scheduler\n lr_scheduler = get_scheduler(\n learning_rate=args.learning_rate,\n scheduler_type=args.scheduler_type,\n num_warmup_steps=args.warmup_steps\n if args.warmup_steps > 0\n else args.warmup_radio,\n num_training_steps=args.max_train_steps,\n )\n\n total_batch_size = args.train_batch_size * args.gradient_accumulation_steps\n\n decay_params = [\n p.name\n for n, p in model.named_parameters()\n if not any(nd in n for nd in [\"bias\", \"norm\"])\n ]\n\n optimizer = AdamW(\n learning_rate=lr_scheduler,\n beta1=0.9,\n beta2=0.98,\n epsilon=args.adam_epsilon,\n parameters=model.parameters(),\n weight_decay=args.weight_decay,\n apply_decay_param_fun=lambda x: x in decay_params,\n )\n\n loss_fn = CrossEntropyLossForSQuAD()\n\n if args.use_amp:\n scaler = GradScaler(init_loss_scaling=args.scale_loss)\n\n logger.info(\"********** Running training **********\")\n logger.info(f\" Num examples = {len(train_dataloader.dataset)}\")\n logger.info(f\" Num Epochs = {args.num_train_epochs}\")\n logger.info(f\" Instantaneous train batch size = {args.train_batch_size}\")\n logger.info(f\" Instantaneous eval batch size = {args.eval_batch_size}\")\n logger.info(f\" Total train batch size (w. accumulation) = {total_batch_size}\")\n logger.info(f\" Gradient Accumulation steps = {args.gradient_accumulation_steps}\")\n logger.info(f\" Total optimization steps = {args.max_train_steps}\")\n\n save_json(vars(args), os.path.join(args.output_dir, \"args.json\"))\n progress_bar = tqdm(range(args.max_train_steps))\n\n global_steps = 0\n tr_loss, logging_loss = 0.0, 0.0\n\n for _ in range(args.num_train_epochs):\n for step, batch in enumerate(train_dataloader):\n model.train()\n with auto_cast(\n args.use_amp, custom_white_list=[\"layer_norm\", \"softmax\", \"gelu\"]\n ):\n input_ids, token_type_ids, start_positions, end_positions = batch\n logits = (\n model(input_ids, token_type_ids=token_type_ids)\n if args.need_token_type_ids\n else model(input_ids)\n )\n loss = (\n loss_fn(logits, (start_positions, end_positions))\n / args.gradient_accumulation_steps\n )\n tr_loss += loss.item()\n\n if args.use_amp:\n scaler.scale(loss).backward()\n else:\n loss.backward()\n\n if (\n step % args.gradient_accumulation_steps == 0\n or step == len(train_dataloader) - 1\n ):\n if args.use_amp:\n scaler.minimize(optimizer, loss)\n else:\n optimizer.step()\n\n lr_scheduler.step()\n optimizer.clear_grad()\n progress_bar.update(1)\n global_steps += 1\n\n if args.logging_steps > 0 and global_steps % args.logging_steps == 0:\n writer.add_scalar(\"lr\", lr_scheduler.get_lr(), global_steps)\n writer.add_scalar(\n \"loss\",\n (tr_loss - logging_loss) / args.logging_steps,\n global_steps,\n )\n logger.info(\n \"global_steps {} - lr: {:.10f} loss: {:.8f}\".format(\n global_steps,\n lr_scheduler.get_lr(),\n (tr_loss - logging_loss) / args.logging_steps,\n )\n )\n logging_loss = tr_loss\n\n if args.save_steps > 0 and global_steps % args.save_steps == 0 and global_steps>=11000:\n logger.info(\"********** Running evaluating **********\")\n logger.info(f\"********** Step {global_steps} **********\")\n output_dir = os.path.join(args.output_dir, f\"step-{global_steps}\")\n os.makedirs(output_dir, exist_ok=True)\n eval_results = evaluate(model, dev_dataloader, args, output_dir)\n for k, v in eval_results.items():\n if \"exact\" in k or \"f1\" in k:\n writer.add_scalar(f\"eval/{k}\", v, global_steps)\n logger.info(f\" {k} = {v}\")\n model.save_pretrained(output_dir)\n tokenizer.save_pretrained(output_dir)\n logger.info(\"********** Evaluating Done **********\")\n\n if global_steps >= args.max_train_steps:\n logger.info(\"********** Training Done **********\")\n return\n\n\nif __name__ == \"__main__\":\n args = parse_args()\n main(args)\n","repo_name":"JunnYu/paddle-mpnet","sub_path":"clean_squad代码/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":8656,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"72196344906","text":"from tkinter import Tk # from tkinter import Tk for Python 3.x\nfrom tkinter.filedialog import askopenfilename\n\nimport numpy as np\nimport pandas as pd\nimport scipy.io as sio\nimport plot_extra as pe\n\nimport matplotlib.pyplot as plt\nplt.rcParams.update(plt.rcParamsDefault)\nplt.rcParams['mathtext.fontset'] = 'cm' # 'cm' Computer modern # 'dejavuserif', 'dejavusans'\nplt.rcParams['font.family'] = 'serif'\n# plt.rcParams['font.serif'] = 'cmr10' # 'https://matplotlib.org/3.1.1/gallery/text_labels_and_annotations/font_file.html\nplt.rc('axes', unicode_minus=False)\n#https://stackoverflow.com/questions/29188757/matplotlib-specify-format-of-floats-for-tick-labels\n\n#https://matplotlib.org/stable/tutorials/text/usetex.html\n# Matplotlib's LaTeX support requires a working LaTeX installation\n# Text handling through LaTeX is slower than Matplotlib's very capable mathtext,\n# but is more flexible, since different LaTeX packages (font packages, math packages, etc.) can be used.\nplt.rcParams['text.usetex'] = 'True'\n\n\n#############\n# solves a warning with a previous syntax\n#https://stackoverflow.com/questions/65645194/warning-set-it-to-a-single-string-instead\nplt.rcParams['text.latex.preamble'] = r'\\usepackage{amsmath} \\usepackage{crimson} \\usepackage{siunitx}'\n# from matplotlib.ticker import FormatStrFormatter\n# from matplotlib.offsetbox import AnchoredText\n\n\ndef plot_charts(figure_type='.pdf'):\n print(\"#####################\")\n print(\"Function name: \", plot_charts.__name__)\n\n ###############################################################\n # CASES - file names - chart limits\n ###############################################################\n csv_full_path = '../Modelos/Intro/SimpleMass_RawData.csv'\n figure_full_path = '../Modelos/Intro/PhaseFreqControl' + figure_type\n\n ###############################################################\n # Opening csv file\n ###############################################################\n print('###########################')\n print(' CSV ')\n print('Chosen file: ', csv_full_path)\n\n # https://pandas.pydata.org/docs/reference/api/pandas.read_csv.html\n csv_df = pd.read_csv(csv_full_path, header=0)\n\n ###############################################################\n # size of the figure (in inches), and linestyles\n ###############################################################\n figsizex = 5\n figsizey = 5\n fig, axes = plt.subplots(3, 1, sharex=True,\n figsize=(figsizex, figsizey),\n num='Charts')\n\n axes[0].plot(csv_df['time'],\n csv_df['pload'],\n color='black',\n linewidth=1,\n linestyle='dashed',\n label=r'Electric load (consumption)')\n\n axes[0].plot(csv_df['time'],\n csv_df['pmec'],\n color=pe.cor_dalt['gray'],\n linewidth=1,\n linestyle='solid',\n label=r'Mechanical power (generation)')\n\n axes[1].plot(csv_df['time'],\n csv_df['f'],\n color='black',\n linewidth=1,\n linestyle='solid',\n label=r'Frequency')\n\n axes[2].plot(csv_df['time'],\n csv_df['pinert'],\n color=pe.cor_dalt['red'],\n linewidth=1,\n linestyle='solid',\n label=r'Inertial')\n\n axes[2].plot(csv_df['time'],\n csv_df['pprim'],\n color=pe.cor_dalt['blue'],\n linewidth=1,\n linestyle='solid',\n label=r'Primary')\n\n axes[2].plot(csv_df['time'],\n csv_df['psec'],\n color=pe.cor_dalt['green'],\n linewidth=1,\n linestyle='solid',\n label=r'Secondary')\n\n # axes[2].plot(csv_df['time'],\n # csv_df['pinert'] + csv_df['pprim'] + csv_df['psec'],\n # color=pe.cor_dalt['gray'],\n # linewidth=1,\n # linestyle='dotted',\n # label=r'Total reserve')\n\n ##########################################################################\n # axis limits\n ##########################################################################\n axes[2].set_xticks(np.arange(0, 100, 10))\n axes[2].set_xlim([0, 80])\n\n axes[0].set_yticks(np.arange(0.50, 0.60, 0.01))\n axes[0].set_ylim([0.499, 0.56])\n\n axes[1].set_yticks(np.arange(0.98, 1.01, 0.002))\n axes[1].set_ylim([0.9875, 1.0005])\n\n axes[2].set_yticks(np.arange(0, 0.08, 0.01))\n axes[2].set_ylim([-0.01, 0.06])\n\n ##########################################################################\n # axis names\n ##########################################################################\n axes[2].set_xlabel(r'Time (\\si{\\second})')\n\n axes[0].set_ylabel(r'Total power (\\si{pu})')\n axes[1].set_ylabel(r'Frequency (\\si{pu})')\n axes[2].set_ylabel(r'Reserve power (\\si{pu})')\n\n ##########################################################################\n # chart identification - legend - abcdefghi\n ##########################################################################\n # https://matplotlib.org/stable/gallery/color/named_colors.html\n # colors lightgray gray aliceblue whitesmoke\n corlegenda = 'whitesmoke'\n #\n axes[0].annotate(r'a', xy=(0.05, 0.25), xycoords='axes fraction',\n bbox=dict(boxstyle='circle', fc=corlegenda))\n\n axes[1].annotate(r'b', xy=(0.05, 0.25), xycoords='axes fraction',\n bbox=dict(boxstyle='circle', fc=corlegenda))\n\n axes[2].annotate(r'c', xy=(0.05, 0.25), xycoords='axes fraction',\n bbox=dict(boxstyle='circle', fc=corlegenda))\n\n ##########################################################################\n # annotations\n ##########################################################################\n axes[0].annotate(r'Nadir happens when powers equalize',\n xy=(12.02, 0.55), xycoords='data',\n xytext=(15, 0.53), textcoords='data',\n ha='left',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n axes[1].annotate(r'RoCoF',\n xy=(10.0, 0.99888), xycoords='data',\n xytext=(12.0, 0.995), textcoords='data',\n arrowprops=dict(arrowstyle=\"->\",connectionstyle=\"arc3\"))\n\n axes[1].annotate(r'Nadir',\n xy=(12.02, 0.9888), xycoords='data',\n xytext=(20, 0.9885), textcoords='data',\n va='center',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n axes[1].annotate(r'Level defined by primary response',\n xy=(20.0, 0.99), xycoords='data',\n xytext=(15.0, 0.999), textcoords='data',\n ha='left',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n axes[1].annotate(r'Secondary response',\n xy=(40.0, 0.995255), xycoords='data',\n xytext=(50.0, 0.9950), textcoords='data',\n ha='left',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n axes[2].annotate(r'Proportional response',\n xy=(26.0, 0.05), xycoords='data',\n xytext=(12.0, 0.0275), textcoords='data',\n ha='left',\n va='center',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n axes[2].annotate(r'Integral response',\n xy=(45.0, 0.032), xycoords='data',\n xytext=(60.0, 0.05), textcoords='data',\n ha='right',\n va='center',\n arrowprops=dict(arrowstyle=\"->\", connectionstyle=\"arc3\"))\n\n ##########################################################################\n # axis legends\n ##########################################################################\n axes[0].legend(loc='best', frameon=False, prop={'size': 10})\n axes[1].legend(loc='best', frameon=False, prop={'size': 10})\n axes[2].legend(loc='center right', frameon=False, prop={'size': 10})\n\n ##########################################################################\n # align, tighten, shown and save\n ##########################################################################\n fig.align_ylabels(axes[:])\n fig.tight_layout()\n # fig.show()\n\n if figure_type == '.pdf':\n plt.savefig(figure_full_path, format=\"pdf\", bbox_inches=\"tight\")\n elif figure_type == '.eps':\n plt.savefig(figure_full_path, format='eps')\n\n plt.show()\n\n\ndef main():\n print(\"#####################\")\n print(\"Function name: \", main.__name__)\n\n figure_type = '.pdf'\n # figure_type = ''\n\n plot_charts(figure_type=figure_type)\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"santosmota/PhD_Daniel_Mota_Data","sub_path":"PythonCharts/Intro_FreqRespPhases.py","file_name":"Intro_FreqRespPhases.py","file_ext":"py","file_size_in_byte":9004,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"10356989069","text":"import json\nimport os\n\nimport boto3\n\nREGION = os.environ.get('REGION', 'us-east-1')\nDOMAIN_NAME = os.environ.get(\"DOMAIN_NAME\", \"*\")\nAWS_ENVIRONMENT = os.environ.get('AWSENV', 'AWS')\n\ndef check_request(message,allowed_method):\n \"\"\"\n Returns 400 if message does not validate with correct HTTP Method\n \"\"\"\n if ('pathParameters' not in message or\n message['httpMethod'] != allowed_method):\n \n return jsonresponse('msg\\': \\'Bad Request',400) \n \ndef get_dynamodb_table(table_name):\n if AWS_ENVIRONMENT == 'AWS_SAM_LOCAL':\n dynamodb_table = boto3.resource(\n 'dynamodb',\n endpoint_url='http://dynamodb:8000'\n )\n else:\n dynamodb_table = boto3.resource(\n 'dynamodb',\n region_name=REGION\n )\n\n return dynamodb_table.Table(table_name)\n\ndef jsonresponse(message, status_code=200):\n \"\"\"\n Return a properly formatted response for AWS Lambda proxy integration\n including CORS\n \"\"\"\n return {\n \"statusCode\": status_code,\n \"headers\": {\n \"Access-Control-Allow-Headers\": \"Content-Type,X-Amz-Date,Authorization,X-Api-Key,x-requested-with\",\n \"Access-Control-Allow-Origin\": \"{}\".format(DOMAIN_NAME),\n \"Access-Control-Allow-Methods\": \"POST,GET,OPTIONS\"\n },\n \"body\": json.dumps(message)\n }\n","repo_name":"chrishollinworth/awschallenge","sub_path":"challenge1/three-tier-example/backend/common/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":1368,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23041286035","text":"details = {\n \"name\" : \"Torikus sadik\",\n \"age\" : 22,\n \"favs\": [\"c\",\"c++\",\"js\",\"python\"]\n}\n\n\n# get method\nname = details.get(\"name\")\n# print(name)\n\n# copy method\ncopy = details.copy()\n# print(copy)\n\n\n# items method\nitems = copy.items()\n# print(items)\n\n# key method\nkeys = copy.keys()\nprint(keys)\n\n# values method\nvalues = copy.values()\nprint(values)\n\n\n# pop method\npoped = details.pop(\"age\")\n# print(poped)\n\n\n# popItem\npop = details.popitem()\nprint(pop)","repo_name":"shakil-babu/Python-Basic-Intermediate","sub_path":"Tutorial/12.Dictionaries/08.methods.py","file_name":"08.methods.py","file_ext":"py","file_size_in_byte":460,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"19250785344","text":"import os\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.patches\n\nimport skimage\nimport skimage.measure\nimport skimage.color\nimport skimage.restoration\nimport skimage.io\nimport skimage.filters\nimport skimage.morphology\nimport skimage.segmentation\nfrom skimage.transform import resize\nimport copy \nimport math\nfrom skimage.morphology import erosion\nfrom skimage.morphology import square\n\nfrom nn import *\nfrom q4 import *\n# do not include any more libraries here!\n# no opencv, no sklearn, etc!\nimport warnings\nwarnings.simplefilter(action='ignore', category=FutureWarning)\nwarnings.simplefilter(action='ignore', category=UserWarning)\n\nfor img in os.listdir('../images'):\n \n im1 = skimage.img_as_float(skimage.io.imread(os.path.join('../images',img)))\n \n bboxes, bw = findLetters(im1)\n \n plt.imshow(bw, cmap='gray')\n\n for bbox in bboxes:\n minr, minc, maxr, maxc = bbox\n rect = matplotlib.patches.Rectangle((minc, minr), maxc - minc, maxr - minr,\n fill=False, edgecolor='red', linewidth=2)\n plt.gca().add_patch(rect)\n plt.show()\n \n # sort rows (the same rows are nearby)\n bboxes.sort(key=lambda x:x[0])\n\n grey_image = skimage.color.rgb2gray(im1)\n\n # find the rows using..RANSAC, counting, clustering, etc.\n y_1, x_1, y_2, x_2 = bboxes[0]\n origin_row = (y_1 + y_2) // 2\n err = abs(origin_row - y_1) \n sort_temp = []\n sorted_bboxes = []\n for bbox in bboxes:\n minr, minc, maxr, maxc = bbox\n row = (minr + maxr) // 2\n if abs(origin_row - row) < err:\n sort_temp.append(bbox)\n else:\n sort_temp.sort(key=lambda x:x[1])\n sorted_bboxes.append(sort_temp)\n sort_temp = []\n sort_temp.append(bbox)\n origin_row = row\n sort_temp.sort(key=lambda x:x[1])\n sorted_bboxes.append(sort_temp)\n data_x_temp = []\n data_x = []\n # crop the bounding boxes\n # note.. before you flatten, transpose the image (that's how the dataset is!)\n # consider doing a square crop, and even using np.pad() to get your images looking more like the dataset\n for same_line in sorted_bboxes:\n for i, sort_loc in enumerate(same_line):\n minr, minc, maxr, maxc = sort_loc\n patch_image = grey_image[minr : maxr + 1, minc : maxc + 1]\n patch_image = erosion(patch_image, selem = square(8))\n r_diff = maxr - minr\n c_diff = maxc - minc\n pad_r = r_diff // 8\n pad_c = c_diff // 8\n patch_image = np.pad(patch_image, ((pad_r,pad_r),(pad_c,pad_c)), mode='maximum')\n patch_image = resize(patch_image, (32, 32))\n patch_image_T = patch_image.T\n data_x_temp.append(patch_image_T.flatten())\n data_x.append(np.array(data_x_temp))\n data_x_temp = []\n \n # load the weights\n # run the crops through your neural network and print them out\n \n import pickle\n import string\n letters = np.array([_ for _ in string.ascii_uppercase[:26]] + [str(_) for _ in range(10)])\n params = pickle.load(open('q3_weights.pickle','rb'))\n\n for group in data_x:\n h_1 = forward(group, params, 'layer1')\n predict_y = forward(h_1, params, 'output',softmax)\n max_output = np.amax(predict_y, axis = 1, keepdims = True)\n pred_class = np.ones(predict_y.shape) * (predict_y//max_output == 1)\n pred = np.nonzero(pred_class)[1]\n print(letters[pred])\n\n\n\n","repo_name":"ChangNeinei/Text_Extracting_NN","sub_path":"run_q4.py","file_name":"run_q4.py","file_ext":"py","file_size_in_byte":3509,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"17347747706","text":"from transformers import AutoModelForCausalLM, AutoTokenizer\nfrom transformers.generation import GenerationConfig\nfrom datetime import datetime\n# 可选的模型包括: \"Qwen/Qwen-7B-Chat\", \"Qwen/Qwen-14B-Chat\"\nmodel_path=\"/opt/data/private/ytw/LLM/Qwen-14B-Chat\"\ntokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)\n\n# 打开bf16精度,A100、H100、RTX3060、RTX3070等显卡建议启用以节省显存\n# model = AutoModelForCausalLM.from_pretrained(\"Qwen/Qwen-7B-Chat\", device_map=\"auto\", trust_remote_code=True, bf16=True).eval()\n# 打开fp16精度,V100、P100、T4等显卡建议启用以节省显存\n# model = AutoModelForCausalLM.from_pretrained(\"Qwen/Qwen-7B-Chat\", device_map=\"auto\", trust_remote_code=True, fp16=True).eval()\n# 使用CPU进行推理,需要约32GB内存\n# model = AutoModelForCausalLM.from_pretrained(\"Qwen/Qwen-7B-Chat\", device_map=\"cpu\", trust_remote_code=True).eval()\n# 默认使用自动模式,根据设备自动选择精度\nmodel = AutoModelForCausalLM.from_pretrained(model_path, device_map=\"auto\", trust_remote_code=True).eval()\n\n# 可指定不同的生成长度、top_p等相关超参\nmodel.generation_config = GenerationConfig.from_pretrained(model_path, trust_remote_code=True)\n\n# 第一轮对话\ns_time =datetime.now()\nfor i in range(10):\n \n response, history = model.chat(tokenizer, \"给我讲一个年轻人奋斗创业最终取得成功的故事。\", history=None)\n print(response)\n \ne_time =datetime.now()\nall_time = e_time-s_time\nprint(\"All time cost {}, per conversation cost {}\".format(all_time,all_time/100))\n# 你好!很高兴为你提供帮助。\n\n# 第二轮对话\n# response, history = model.chat(tokenizer, \"给我讲一个年轻人奋斗创业最终取得成功的故事。\", history=history)\n# print(response)\n# 这是一个关于一个年轻人奋斗创业最终取得成功的故事。\n# 故事的主人公叫李明,他来自一个普通的家庭,父母都是普通的工人。从小,李明就立下了一个目标:要成为一名成功的企业家。\n# 为了实现这个目标,李明勤奋学习,考上了大学。在大学期间,他积极参加各种创业比赛,获得了不少奖项。他还利用课余时间去实习,积累了宝贵的经验。\n# 毕业后,李明决定开始自己的创业之路。他开始寻找投资机会,但多次都被拒绝了。然而,他并没有放弃。他继续努力,不断改进自己的创业计划,并寻找新的投资机会。\n# 最终,李明成功地获得了一笔投资,开始了自己的创业之路。他成立了一家科技公司,专注于开发新型软件。在他的领导下,公司迅速发展起来,成为了一家成功的科技企业。\n# 李明的成功并不是偶然的。他勤奋、坚韧、勇于冒险,不断学习和改进自己。他的成功也证明了,只要努力奋斗,任何人都有可能取得成功。\n\n# 第三轮对话\n# response, history = model.chat(tokenizer, \"给这个故事起一个标题\", history=history)\n# print(response)s\n# 《奋斗创业:一个年轻人的成功之路》","repo_name":"augusyan/LLM_Eval","sub_path":"demo_qwen_chat.py","file_name":"demo_qwen_chat.py","file_ext":"py","file_size_in_byte":3107,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23042878430","text":"from sklearn.feature_extraction.text import TfidfVectorizer\nfrom tqdm import tqdm, trange\nimport torch\nfrom sklearn.metrics.pairwise import cosine_similarity\nimport numpy as np\nimport itertools\nimport argparse\nimport os\nimport torch.nn as nn\nfrom transformers import BertConfig, BertForSequenceClassification, BertTokenizer\nfrom bm25_custom import get_bm25_weights\nimport nltk\nnltk.download('stopwords')\nfrom nltk import word_tokenize\nfrom nltk.corpus import stopwords\nimport string\nimport re\n\n\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = \"1\"\n\n\ndef get_values(file):\n \"\"\"\n get label context and response.\n :param file: filel name\n :param get_c_d:\n :return:\n \"\"\"\n data = open(file, 'r').readlines()\n data = [sent.split('\\n')[0].split('\\t') for sent in data]\n cr = [\"\\n\".join([sen for sen in a[1:]]) for a in data]\n return cr\n\n\ndef makedocs():\n sample_to_doc = []\n all_docs_str = []\n all_docs_list = []\n doc = []\n corpus_lines = 0\n doc_cnt = 0\n with open(\"../tf-idf/post_train_ver2.txt\", \"r\", encoding='utf-8') as f:\n for line in tqdm(f, desc=\"Loading Dataset\", total=corpus_lines):\n line = line.strip()\n if line == \"\":\n if (len(doc) != 0):\n all_docs_str.append(\"\\n\".join(doc))\n all_docs_list.append(doc)\n doc_cnt = 0\n doc = []\n # remove last added sample because there won't be a subsequent line anymore in the doc\n sample_to_doc.pop()\n else:\n p = re.compile(\"[^0-9.]\")\n line=line.split()\n newline=[]\n for word in line:\n if \"\".join(p.findall(word)):\n newline.append(word)\n line=\" \".join(newline)\n # store as one sample\n if doc_cnt >= 10:\n continue\n sample = {\"doc_id\": len(all_docs_list),\n \"line\": len(doc)}\n sample_to_doc.append(sample)\n if line==\"\":\n continue\n doc.append(line.lower())\n doc_cnt += 1\n corpus_lines = corpus_lines + 1\n\n # if last row in file is not empty\n if all_docs_list[-1] != doc:\n all_docs_list.append(doc)\n all_docs_str.append(\"\\n\".join(doc))\n sample_to_doc.pop()\n\n for doc in all_docs_list:\n if len(doc) == 0:\n print(doc)\n num_docs = len(all_docs_list)\n return all_docs_str, all_docs_list\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--batch_size\",\n default=0,\n type=int,\n help=\"The batch size.\")\n parser.add_argument(\"--model_type\", default=\"bert\", type=str,\n help=\"Model type selected in the list: bert\")\n parser.add_argument(\"--model_name_or_path\", default=\"bert-base-uncased\", type=str,\n help=\"bert file location\")\n parser.add_argument(\"--config_name\", default=\"\", type=str,\n help=\"Pretrained config name or path if not the same as model_name\")\n parser.add_argument(\"--tokenizer_name\", default=\"\", type=str,\n help=\"Pretrained tokenizer name or path if not the same as model_name\")\n parser.add_argument(\"--cache_dir\", default=\"bert_cache\", type=str,\n help=\"Where do you want to store the pre-trained models downloaded from s3\")\n parser.add_argument(\"--do_lower_case\", action='store_true', default=True,\n help=\"Set this flag if you are using an uncased model.\")\n parser.add_argument(\"--load\", default=False, type=bool)\n parser.add_argument(\"--save_path\",\n default=\"../tf-idf/checkpoint/ubuntu.bert.pt\",\n type=str,\n help=\"The path to save model.\")\n\n args = parser.parse_args()\n # load bert\n print(args)\n\n # load the vocab file\n targetlist = get_values('../tf-idf/valid.txt')\n l = 50000\n targetlist = targetlist[:l]\n all_docs_str, all_docs_list = makedocs()\n\n wholestring = \"\"\n\n #tiv = TfidfVectorizer(stop_words=\"english\").fit(all_docs_str)\n #tokenizer = tiv.build_tokenizer()\n #all_docs_numpy = tiv.transform(all_docs_str)\n #all_docs_text_numpy = np.array(all_docs_list)\n # way to find relative docs\n stop = stopwords.words('english') + list(string.punctuation)\n corpus=[]\n strcorpus=[]\n for doc in all_docs_str:\n for sent in doc.split('\\n'):\n strcorpus.append(sent)\n corpus.append(list(set([i for i in word_tokenize(sent.lower()) if i not in stop])))\n target = []\n target_part=[]\n for i,doc in enumerate(tqdm(targetlist)):\n if i % 1000==0 and target_part:\n target.append(target_part)\n target_part=[]\n target_part.append(list(set([i for i in word_tokenize(doc.split(\"\\n\")[-1].lower()) if i not in stop])))\n if target_part:\n target.append(target_part)\n\n strcorpus=np.array(strcorpus)\n cnt=0\n\n for target_part in tqdm(target):\n relevant_str_idx_part = get_bm25_weights(corpus,target_part, n_jobs=-1)\n\n for relevant_str_idx_single in (relevant_str_idx_part):\n with torch.no_grad():\n top10doc=strcorpus[relevant_str_idx_single]\n prev = \"\"\n sencount=0\n for i, sen in enumerate(top10doc):\n if sencount==10:\n break\n\n if prev != sen and len(sen.split())>2:\n wholestring += sen + '\\n'\n prev = sen\n sencount+=1\n\n if cnt == 440:\n print(\"0\")\n if prev==\"\":\n wholestring+='[empty]\\n'\n print(\"empty\")\n wholestring += '\\n'\n cnt+=1\n\n file = open(\"bm25_only_dev1_r.txt\", \"w\", encoding='utf-8')\n file.write(wholestring)\n\n # pickle.dump(dataset, open('ubuntu_data/dataset_1M.pkl', 'wb'))","repo_name":"hanjanghoon/NLP_Dialogue_BERT-Knowledge-IR","sub_path":"IR/bm25/onlybm25.py","file_name":"onlybm25.py","file_ext":"py","file_size_in_byte":6179,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"37647127930","text":"#Upwind for convection term. CD for diffusion term\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef tdma(a,b,c,d,T):\n \"\"\" a: ap(main diagonal), b: -ae, c: -aw, d: Su, T: variable vector \"\"\"\n N=len(a)\n P=np.ones(N)\n Q=np.ones(N)\n P[0]=-b[0]/a[0]\n Q[0]= d[0]/a[0]\n for i in range(1,N):\n denom=1/(a[i]+c[i]*P[i-1])\n P[i] = -b[i]*denom\n Q[i] = (d[i]-c[i]*Q[i-1])*denom\n T[N-1]=Q[N-1]\n for i in range(N-2,-1,-1):\n T[i]=P[i]*T[i+1]+Q[i]\n\n#Case\ncase=\"case1\"\n\n#geometric\nlength = 1 #m\n\n#fluid properties\nrho=1\nk=0.1\n\nif(case==\"case1\"):\n nx=5\n velocity=-0.1\nelif(case==\"case2\"):\n nx=5\n velocity=2.5\nelse:#case3 or junk case :)\n nx=20\n velocity=2.5\n\ndx = length/nx\n\n#bc and initial conditions\n\nTa=1\nTb=0\n\nF = rho*velocity\nD = k/dx\n\n#declaring and initialising arrays\nu = np.ones(nx)*velocity\nT=np.ones(nx)\naw=np.ones(nx)*(D+F)\nae=np.ones(nx)*D\nSp=np.zeros(nx)\nSu=np.zeros(nx)\n\n#leftmost node\naw[0]=0\nSp[0]=-(2*D+F)\nSu[0]=(2*D+F)*Ta\n#rightmost node\nae[nx-1]=0\nSp[nx-1]=-2*D\nSu[nx-1]=2*D*Tb\nap=aw+ae-Sp\n\n#solve\ntdma(ap,-ae,-aw,Su,T)\n\n#exact solution\nx=np.linspace(0,length,nx+1)+dx/2\nx=x[:-1]\nTexact = Ta+(Tb-Ta)* (np.exp(rho*velocity*x/k) -1) / (np.exp(rho*velocity*length/k) -1)\n\nprint(T)\nprint(Texact)\n\n#plot\nplt.plot(x,T,linestyle='--' ,marker='o',color='b',label='Numerical Solution [Upwind]')\nplt.plot(x,Texact,'-gD',label='Exact Solution')\n\nplt.title(\"Comparison of distribution of property\")\nplt.xlabel(\"Position\")\nplt.ylabel(\"Property\")\nplt.legend()\nplt.grid()\nfigname=\"comparison_5.2_\"+case\nplt.savefig(figname+\".png\")\nplt.show()\n","repo_name":"nipinl/codingVersteegMalalasekera","sub_path":"07_1dConvectionDifffusion_upwind_5.2.py","file_name":"07_1dConvectionDifffusion_upwind_5.2.py","file_ext":"py","file_size_in_byte":1610,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"25782686384","text":"from selenium import webdriver\nimport time\nimport pandas as pd\nimport os\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nimport psycopg2\n\nDIR_PATH = os.path.abspath(os.path.dirname(__file__))\nfolder_name = \"_Output\"\nfile_name='Jodi_data.csv'\n\ndef download_file(chromedriver_url):\n path = DIR_PATH\n options = webdriver.ChromeOptions();\n prefs = {\"download.default_directory\": path};\n options.add_experimental_option(\"prefs\", prefs);\n driver = webdriver.Chrome(service_log_path=chromedriver_url, options=options)\n driver.get(\"http://www.jodidb.org/TableViewer/tableView.aspx?ReportId=93906\")\n driver.maximize_window()\n driver.find_element(By.XPATH,\"/html/body/table/tbody/tr[1]/td[2]/table[2]/tbody/tr[1]/td/div/table/tbody/tr/td[5]/a/img\").click()\n driver.find_element(By.XPATH,\"/html/body/table/tbody/tr[1]/td[2]/table[2]/tbody/tr[1]/td/div/table/tbody/tr/td[6]/div/div/p[3]/nobr/a\").click()\n driver.switch_to.window(driver.window_handles[1])\n WebDriverWait(driver, 10).until(EC.presence_of_element_located((By.XPATH, '/html/body/form/table/tbody/tr[1]/td[2]/input[1]')))\n download = driver.find_element(By.XPATH, \"/html/body/form/table/tbody/tr[1]/td[2]/input[1]\").click()\n time.sleep(5)\n driver.close()\n\ndef process_file():\n path=os.path.abspath(os.path.dirname(__file__))\n list=os.listdir(path)\n time_sorted_list = sorted(list, key=os.path.getmtime)\n file_name = time_sorted_list[len(time_sorted_list) - 1]\n df = pd.read_csv(file_name, header=3)\n df.drop(0,axis=0, inplace=True)\n melted_df = df.melt(id_vars=['Time'], value_vars=df.loc['Algeria':'Total Africa NG'], var_name='Country',value_name='Value')\n melted_df.rename(columns = {'Time':'Country','Country':'Time'}, inplace = True)\n path = DIR_PATH + folder_name\n try:\n os.mkdir(path)\n except OSError as error:\n print(error)\n date(melted_df,path)\n\ndef date(melted_df,path):\n df=melted_df\n df['Time'] = pd.to_datetime(df['Time'], format='mixed', dayfirst=True).dt.strftime('%d-%m-%Y')\n df[\"Time\"] = pd.to_datetime(df[\"Time\"])\n print(df['Time'])\n df.to_csv(path+\"/Jodi_data.csv\",header=df.columns,index=False,encoding='utf-8')\n print(df)\n\ndef table():\n conn=psycopg2.connect(host='localhost',dbname='JodiDB',user='postgres',password='1234',port=5432)\n cur = conn.cursor()\n path=os.path.join(DIR_PATH,folder_name, file_name)\n\n #create table\n cur.execute('''CREATE TABLE IF NOT EXISTS jodi(\n country VARCHAR(40),\n Time DATE NOT NULL,\n value FlOAT)''')\n\n print(\"table created\")\n #copy data from csv\n sql2 = '''COPY jodi(country,Time,value)\n FROM 'E:/jodidb/venv/_Output/Jodi_data.csv'\n DELIMITER ','\n CSV HEADER;'''\n\n cur.execute(sql2)\n conn.commit()\n cur.close()\n conn.close()\n\n\ndownload_file(\"E:/selenium/chromedriver.exe\")\nprocess_file()\ntable()\n","repo_name":"KHemanth2001/jodidb","sub_path":"jodi_postgres.py","file_name":"jodi_postgres.py","file_ext":"py","file_size_in_byte":2996,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14969532073","text":"import os\nimport pandas as pd\nimport cv2\nimport numpy as np\n\n\ndef image_label(number, label): \n y = 0 # Unknown\n for index, row in label.iterrows():\n start = row['Start']\n end = row['End']\n if number >= start and number <= end:\n y = row[\"Label\"]\n return y\n return y\n\ndef create_training_csv_multiclass(data_path, label_path, sheet_name, writer):\n # Policy: Discard frames that don't meet the 12 sliding window requirement \n df = pd.read_excel(label_path, sheet_name=sheet_name)\n subjects = list(df['Subject'].unique())\n print(\"Subjects:\", subjects)\n \n for subject in subjects:\n print(\"Processing Subject\", subject)\n\n subject_data_path = os.path.join(data_path, subject) \n subject_df = df[df['Subject'] == subject]\n\n numbers = sorted([int(f[3:-4]) for f in os.listdir(subject_data_path)]) \n list_filename = []\n current_label = -1\n past_label = -1\n for number in numbers: \n filename = \"img\" + str(number).zfill(5) + \".jpg\"\n path_filename = os.path.join(subject_data_path, filename)\n \n # Label\n current_label = image_label(number, subject_df) \n\n # Discard if the image is in excepted image\n # RISK: if we allow the model to train on blank image, it could disrupt training process\n img = cv2.imread(path_filename)\n all_zeros = not np.any(img)\n if all_zeros:\n current_label = 0\n\n # Write to CSV\n if current_label == past_label:\n list_filename.append(path_filename)\n if len(list_filename) == 12: \n for filename in list_filename:\n writer.write(filename)\n writer.write(',')\n writer.write(str(past_label))\n writer.write('\\n')\n list_filename = []\n else:\n # POLICY DISCARD\n list_filename = []\n list_filename.append(path_filename)\n past_label = current_label \n \n\nif __name__ == \"__main__\": \n # Training\n filewriter = open('training_pubspeak_25042023_face_detection.csv', 'w')\n filewriter.write(\"1,2,3,4,5,6,7,8,9,10,11,12,Label\\n\")\n create_training_csv_multiclass( \"D:\\Dataset Skripsi Batch Final Image Face Detection\",\n \"D:\\CodeProject2\\SKRIPSI_FINAL\\pubspeak_label_25042023.xlsx\",\n \"Training\",\n filewriter)\n filewriter.close()\n\n # Testing\n filewriter = open('testing_pubspeak_25042023_face_detection.csv', 'w')\n filewriter.write(\"1,2,3,4,5,6,7,8,9,10,11,12,Label\\n\")\n create_training_csv_multiclass( \"D:\\Dataset Skripsi Batch Final Image Face Detection\",\n \"D:\\CodeProject2\\SKRIPSI_FINAL\\pubspeak_label_25042023.xlsx\",\n \"Testing\",\n filewriter)\n filewriter.close()\n \n\n","repo_name":"vincentdar/SKRIPSI_FINAL","sub_path":"training/create_training_csv_binary.py","file_name":"create_training_csv_binary.py","file_ext":"py","file_size_in_byte":3223,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41155428446","text":"class Solution:\n def subdomainVisits(self, cpdomains: List[str]) -> List[str]:\n \n hmap = collections.defaultdict(int)\n \n for d in cpdomains:\n count, domain = d.split(\" \")\n sub_domain = domain.split(\".\")\n for i in range(len(sub_domain)):\n hmap[\".\".join(sub_domain[i:])] += int(count)\n \n return [str(v)+ \" \" + k for k, v in hmap.items()]\n \n \n \n \n \n \n ","repo_name":"tejeshreddy/competitive-programming","sub_path":"811-subdomain-visit-count/811-subdomain-visit-count.py","file_name":"811-subdomain-visit-count.py","file_ext":"py","file_size_in_byte":503,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"7825680324","text":"from flask import redirect, url_for\n\nimport db\nfrom models import Usuario, Proveedor, Producto\nclass ValidarLogin:\n#Clase ValidarLogin\n#Recibe un correo y una contrasena\n#Tiene un método llamado valida_usuario_login\n def __init__(self, correo, contrasena):\n self.correo = correo\n self.contrasena = contrasena\n\n\n def valida_usuario_login(self):\n #Método que valida que el inicio de sesión de un usuario, retorna el tipo de acceso si coinciden los datos\n usuario = db.session.query(Usuario).filter_by(correo_electronico=self.correo).first()\n if self.contrasena == \"\":\n return False\n elif usuario is None:\n return False\n elif usuario.contrasena == self.contrasena:\n return usuario\n else:\n return False\n\n\n @classmethod\n def get_by_id(self, id):\n usuario = db.session.query(Usuario).filter_by(id_usuario=id).first()\n if usuario != None:\n return usuario\n else:\n return None\n\n def __str__(self):\n return \"{} -> {}\".format(self.correo, self.contrasena)\n\nclass ValidarUsuario:\n#Clase validarUsuario valida los formularios de ingreso de usuarios y modificacion de usuario\n#args:\n#nombre hace referencia al nombre del usuario\n#apellido ahce referencia al apellido del usuario\n#correo_electronico hace referencia al correo del usuario (no se puede repetir)\n#contrasena hace referencia a la contraseña del usuario\n#tipo_acceso hace referencia al tipo de acceso del usuario (ADMINISTRADOR, CLIENTE, PROVEEDOR)\n def __init__(self, nombre, correo_electronico, contrasena, tipo_acceso, apellido=None):\n self.nombre = nombre\n self.apellido = apellido\n self.correo_electronico = correo_electronico\n self.contrasena = contrasena\n self.tipo_acceso = tipo_acceso\n\n\n def validar_formulario_nUsuario(self):\n #Método que valida el formulario de ingreso de un nuevo usuario\n\n #Se consulta en la base de datos si existe un usuario con el correo electronico del nuevo usuario\n validacion_correo = db.session.query(Usuario).filter_by(correo_electronico=self.correo_electronico).first()\n\n #Se validan que los campos obligatorios tengan informacion\n if self.nombre != \"\" and self.contrasena != \"\" and self.tipo_acceso != \"\" and self.correo_electronico != \"\":\n if validacion_correo is not None:#Si la variable contiene informacion de un usuario, es porque el correo electronico del nuevo usuario ya existe\n # print(\"error correo\")\n return \"errorCorreo\"\n else:\n return True\n else:#Si no se cumplen las condiciones anteriores es porque uno o mas campos han quedado vacíos\n return \"camposVacios\"\n\n def modifica_usuario(self, id_usuario):\n #Funcion que modifica los datos de un usuario\n\n #Se obtiene el usuario a modificar\n usuario = db.session.query(Usuario).filter_by(id_usuario=id_usuario).first()\n\n if self.nombre == \"\" and self.apellido == \"\" and self.correo_electronico == \"\" and self.contrasena == \"\":\n #Si se deja en blanco el formulario\n return False\n elif self.nombre != \"\" and self.apellido != \"\" and self.correo_electronico != \"\" and self.contrasena != \"\" and self.tipo_acceso != \"\":\n if usuario.correo_electronico == self.correo_electronico:\n #Se valida que no ingrese un correo repetido\n return False\n else:\n #Si se rellenan todos los campos\n usuario.nombre = self.nombre\n usuario.apellido = self.apellido\n usuario.correo_electronico = self.correo_electronico\n usuario.contrasena = self.contrasena\n usuario.tipo_acceso = self.tipo_acceso\n\n db.session.commit()\n return True\n else:\n #Si se llena solo algunos campos del formulario\n if self.nombre != \"\":\n # Si el nombre no esta vacio\n usuario.nombre = self.nombre\n if self.apellido != \"\":\n # Si el apellido no esta vacio\n usuario.apellido = self.apellido\n if self.correo_electronico != \"\":\n # Si el correo electronico no esta vacio\n usuario.correo_electronico = self.correo_electronico\n if self.contrasena != \"\":\n # Si la contraseña no esta vacia\n usuario.contrasena = self.contrasena\n if self.tipo_acceso != \"\":\n # Si el tipo de acceso no esta vacio\n usuario.tipo_acceso = self.tipo_acceso\n #Se guardan los cambios en la bd\n db.session.commit()\n return True\n\nclass ValidaProveedor:\n#Función que valida los formularios de ingreso de un nuevo proveedor y el formulario de modificaciones de proveedores\n def __init__(self, nombre_empresa, cif, correo_electronico, telefono, direccion, facturacion, descuento, iva):\n self.nombre_empresa = nombre_empresa\n self.cif = cif\n self.correo_electronico = correo_electronico\n self.telefono = telefono\n self.dirreccion = direccion\n self.facturacion = facturacion\n self.descuento = descuento\n self.iva = iva\n\n def __str__(self):\n return \"{} -> {} -> {} -> {} -> {} -> {} -> {} -> {}\".format(self.nombre_empresa, self.cif, self.telefono, self.direccion, self.correo_electronico,\n self.facturacion, self.descuento, self.iva)\n def validar_proveedor(self):\n #valida el formulario de ingreso de un nuevo proveedor\n\n #Se consulta a la base de datos si existe algun proveedor con el cif ingresado por el usuario\n proveedor = db.session.query(Proveedor).filter_by(cif=self.cif).first()\n\n #Se validan que los campos obligatorios no queden vacíos\n if self.nombre_empresa == \"\" or self.cif == \"\" or self.telefono == \"\" or self.dirreccion == \"\" or self.facturacion == \"\" or self.iva == \"\":\n return False\n elif proveedor != None:\n if proveedor.cif == self.cif: #Se valida que no exista un proveedor con el cif del nuevo proveedor\n return \"cifRepetido\"\n else:\n return True\n def modifica_proveedor(self):\n #Funcion que modifica los datos de un producto\n p = db.session.query(Proveedor).filter_by(cif=self.cif).first()\n\n if self.nombre_empresa == \"\" and self.correo_electronico == \"\" and self.telefono == \"\" and self.dirreccion == \"\" and self.facturacion == \"\" and self.descuento == \"\" and self.iva == \"\":\n #Si se deja en blanco el formulario\n return False\n elif self.nombre_empresa != \"\" and self.correo_electronico != \"\" and self.telefono != \"\" and self.dirreccion != \"\" and self.facturacion != \"\" and self.descuento != \"\" and self.iva != \"\":\n #Si se rellenan todos los campos\n p.nombre_empresa = self.nombre_empresa\n p.correo_electronico = self.correo_electronico\n p.direccion = self.dirreccion\n p.telefono = self.telefono\n p.facturacion = self.facturacion\n p.descuento = self.descuento\n p.iva = self.iva\n\n db.session.commit()\n return True\n else:\n #Si se llena solo algunos campos del formulario\n if self.nombre_empresa != \"\":\n # Si el nombre no esta vacio\n p.nombre_empresa = self.nombre_empresa\n if self.correo_electronico != \"\":\n # Si la descripcion no esta vacia\n p.correo_electronico = self.correo_electronico\n if self.dirreccion != \"\":\n # Si la marca no esta vacia\n p.direccion = self.dirreccion\n if self.telefono != \"\":\n # Si el stock no esta vacio\n p.telefono = self.telefono\n if self.facturacion != \"\":\n # Si el precio no esta vacio\n p.facturacion = self.facturacion\n if self.descuento != \"\":\n # Si la ubicacion no esta vacia\n p.descuento = self.descuento\n if self.iva != \"\":\n # Si la ubicacion no esta vacia\n p.iva = self.iva\n\n db.session.commit()\n return True\n\n def __str__(self):\n return \"{} -> {} -> {} -> {} -> {}\".format(self.nombre, self.apellido, self.correo_electronico, self.contrasena, self.tipo_usuario)\n\n\nclass ValidaProducto:\n#Clase que valida el ingreso de un nuevo producto o la modificación de uno existente\n\n def __init__(self, numero_referencia, nombre_producto, descripcion, marca, stock, precio, ubicacion):\n self.numero_referencia = numero_referencia\n self.nombre_producto = nombre_producto\n self.descripcion = descripcion\n self.marca = marca\n self.stock = stock\n self.precio = precio\n self.ubicacion = ubicacion\n\n def validar_producto(self):\n # Valida producto es una funcion que valida si los campos obligatorios del formulario crear producto han quedado vacios o no\n\n # Con el número de referencia se busca en la base de datos si hay algún otro producto con el mismo número\n producto = db.session.query(Producto).filter_by(numero_referencia=self.numero_referencia).first()\n\n if self.numero_referencia == \"\" or self.nombre_producto == \"\" or self.descripcion == \"\" or self.stock == \"\" or self.precio == \"\" or self.ubicacion == \"\":\n # Si alguno de los campos obligatorios queda vacío retorna una respuesta\n return False\n elif producto != None:\n # Si al consultar a la base de datos por ese numero de referencia retorna un valor, se retorna una respuesta de numero repetido\n if producto.numero_referencia == self.numero_referencia:\n return \"referencia_repetida\"\n else:\n # En caso contrario, está correcto\n return True\n\n def modifica_producto(self, nombre_imagen):\n #Funcion que modifica los datos de un producto\n p = db.session.query(Producto).filter_by(numero_referencia=self.numero_referencia).first()\n\n if self.nombre_producto == \"\" and self.stock == \"\" and self.marca == \"\" and self.precio == \"\" and self.ubicacion == \"\":\n #Si se deja en blanco el formulario\n return False\n elif self.nombre_producto != \"\" and self.descripcion != \"\" and self.marca != \"\" and self.stock != \"\" and self.precio != \"\" and self.ubicacion != \"\" :\n #Si se rellenan todos los campos\n p.nombre_producto = self.nombre_producto\n p.descripcion = self.descripcion\n p.marca = self.marca\n p.stock = self.stock\n p.precio = self.precio\n p.ubicacion = self.ubicacion\n p.imagen = nombre_imagen\n db.session.commit()\n return True\n else:\n #Si se llena solo algunos campos del formulario\n if self.nombre_producto != \"\":\n # Si el nombre no esta vacio\n p.nombre_producto = self.nombre_producto\n if self.descripcion != \"\":\n # Si la descripcion no esta vacia\n p.descripcion = self.descripcion\n if self.marca != \"\":\n # Si la marca no esta vacia\n p.marca = self.marca\n if self.stock != \"\":\n # Si el stock no esta vacio\n p.stock = self.stock\n if self.precio != \"\":\n # Si el precio no esta vacio\n p.precio = self.precio\n if self.ubicacion != \"\":\n # Si la ubicacion no esta vacia\n p.ubicacion = self.ubicacion\n\n if nombre_imagen != \"\":\n # Si el nombre de la imagen no está vacio\n p.imagen = nombre_imagen\n\n db.session.commit()\n return True\n\ndef __str__(self):\n return \"{} -> {} -> {} -> {} -> {} -> {} -> {} -> {}\".format(self.nombre_empresa, self.cif, self.correo_electronico, self.telefono, self.dirreccion, self.facturacion,\n self.descuento, self.iva)","repo_name":"jptamayoa/suministrosInformaticos","sub_path":"validaciones.py","file_name":"validaciones.py","file_ext":"py","file_size_in_byte":12378,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"42669602628","text":"import glob\nimport math\nimport numpy as np\nimport os\nimport pathlib\n\nimport pandas as pd\nimport tensorflow as tf\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\n\nPATHBASE = '/home/xandao/Imagens'\nPATCHES = [3]\n\n\ndef next_patch(spec, n):\n step = math.floor(spec.shape[0] / n)\n for i in range(n):\n yield spec[i * step:(i + 1) * step, :, :]\n\n\ndef get_model(model, **kwargs):\n if model == 'vgg16':\n return tf.keras.applications.vgg16.VGG16(**kwargs), tf.keras.applications.vgg16.preprocess_input\n if model == 'resnet50v2':\n return tf.keras.applications.resnet_v2.ResNet50V2(**kwargs), tf.keras.applications.resnet_v2.preprocess_input\n if model == 'mobilenetv2':\n return tf.keras.applications.mobilenet_v2.MobileNetV2(\n **kwargs), tf.keras.applications.mobilenet_v2.preprocess_input\n\n raise ValueError\n\n\n\ndef extract_features(cnn, color, dataset, gpuid, folds, image_size, input_path, level, minimum_image, output_path, patches, region):\n os.environ[\"CUDA_VISIBLE_DEVICES\"] = str(gpuid)\n spec_height = image_size[0]\n spec_width = image_size[1]\n input_path_proto = os.path.join(input_path, 'f%d', '*.jpeg')\n\n for n_patches in patches:\n print(\"Slicing images into %d non-overlapping patches...\" % (n_patches))\n tf.keras.backend.clear_session()\n\n input_shape = (math.floor(spec_height / n_patches), spec_width, 3)\n\n model, preprocess_input = get_model(cnn, weights='imagenet', include_top=False,\n input_shape=input_shape, pooling='avg')\n total_samples=0\n imgs_sliced = []\n for fold in folds:\n print(\"Extracting features for fold %d...\" % (fold))\n if len(glob.glob(input_path_proto % (fold))) == 0:\n raise RuntimeError(\"No files found in: %s\" % (input_path_proto % (fold)))\n\n features = []\n for fname in sorted(glob.glob(input_path_proto % (fold))):\n print('fname: %s' % fname)\n img = tf.keras.preprocessing.image.load_img(fname)\n spec = tf.keras.preprocessing.image.img_to_array(img)\n for p in next_patch(spec, n_patches):\n p = preprocess_input(p)\n imgs_sliced.append(tf.keras.preprocessing.image.array_to_img(p))\n p = np.expand_dims(p, axis=0)\n features.append(model.predict(p))\n\n features = np.concatenate(features)\n save_file('npy', features, fold, n_patches, output_path)\n save_file('npz', features, fold, n_patches, output_path)\n n_samples, n_features = features.shape\n total_samples+=n_samples\n print(total_samples)\n save_information(color, cnn, dataset, image_size, input_path, level, minimum_image, n_features, output_path, n_patches, region, total_samples)\n\n\ndef save_file(extension, features, fold, n_patches, output_path):\n output_path = os.path.join(output_path, extension)\n\n if not os.path.exists(output_path):\n os.makedirs(output_path)\n\n output_path = os.path.join(output_path, 'fold-%d_patches-%d.' + extension)\n output_filename = output_path % (fold, n_patches)\n if extension == 'npy':\n np.save(output_filename, features, allow_pickle=True)\n else:\n np.savez_compressed(output_filename, x=features, y=np.repeat(fold, features.shape[0]))\n\n\ndef save_information(color, cnn, dataset, image_size, input_path, level, minimum_image, n_features, output_path, patch, region, total_samples):\n height = image_size[0]\n width = image_size[1]\n index = ['color', 'cnn', 'dataset', 'height', 'width', 'level', 'minimum_image', 'input_path', 'output_path',\n 'patch', 'n_features', 'total_samples']\n data = [cnn, color, dataset, height, width, level, minimum_image, input_path, output_path, patch, n_features, total_samples]\n\n if region:\n index.append('region')\n data.append(region)\n\n df = pd.DataFrame(data, index=index)\n filename = os.path.join(output_path, 'info.csv')\n df.to_csv(filename, sep=';', index=index, header=None, lineterminator='\\n', doublequote=True)\n\n\ndef prepare(cnn, color, dataset, image_size, level, minimum_image, input_path, region=None):\n if not os.path.exists(input_path):\n raise SystemError('path (%s) not exists' % input_path)\n\n list_folders = [f for f in pathlib.Path(input_path).glob('*') if f.is_dir()]\n folds = len(list_folders)\n folds = list(range(1, folds + 1))\n image_size = (int(image_size), int(image_size))\n gpuid = 0\n patches = PATCHES\n patches = list(patches)\n\n features_folder = dataset + '_features'\n if region:\n output_path = os.path.join(PATHBASE, features_folder, color, str(image_size[0]), level, region, minimum_image, cnn)\n else:\n output_path = os.path.join(PATHBASE, features_folder, color, str(image_size[0]), level, minimum_image, cnn)\n\n if not os.path.exists(output_path):\n os.makedirs(output_path)\n\n print(\"Feature Extraction Parameters\")\n print(\"Pre-trained model: %s\" % cnn)\n print(\"Non-overlapping patches per image: %s\" % str(patches))\n print(\"Folds: %s\" % str(folds))\n print(\"Image Dimensions h=%s, w=%s \" % (image_size, image_size))\n print(\"Format string for input: %s \" % input_path)\n print(\"Format string for output: %s \" % output_path)\n print(\"GPU ID: %d\" % gpuid)\n\n extract_features(cnn, color, dataset, gpuid, folds, image_size, input_path, level, minimum_image, output_path, patches, region)\n\n\ndef main():\n list_color = ['RGB', 'GRAYSCALE']\n list_size = ['256', '400', '512']\n list_minimum_image = ['20', '10', '5']\n list_cnn = ['vgg16', 'resnet50v2', 'mobilenetv2']\n list_dataset = ['pr_dataset', 'br_dataset', 'region_dataset']\n list_level = ['specific_epithet_trusted']\n list_region = ['Norte', 'Nordeste', 'Sul', 'Sudeste', 'Centro-Oeste']\n for cnn in list_cnn:\n for color in list_color:\n for image_size in list_size:\n for minimum_image in list_minimum_image:\n for level in list_level:\n for dataset in list_dataset:\n print('cnn: %s color: %s dataset: %s image_size: %s level: %s minimum_image: %s '\n % (cnn, color, dataset, image_size, level, minimum_image))\n if 'region_dataset' == dataset:\n for region in list_region:\n path = os.path.join(PATHBASE, dataset, color, level, region, image_size,\n minimum_image)\n prepare(cnn, color, dataset, image_size, level, minimum_image, path, region=region)\n else:\n path = os.path.join(PATHBASE, dataset, color, level, image_size, minimum_image)\n prepare(cnn, color, dataset, image_size, level, minimum_image, path)\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"xaaaandao/deep_feature","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":7088,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21108101229","text":"import numbers\nimport typing\nfrom typing import Any, Callable, Iterable, List, Optional, Tuple, Union\n\nfrom pyglove.core import geno\nfrom pyglove.core import object_utils\nfrom pyglove.core import symbolic\nfrom pyglove.core import typing as pg_typing\nfrom pyglove.core.hyper import base\nfrom pyglove.core.hyper import object_template\n\n\n@symbolic.members([\n ('num_choices', pg_typing.Int(min_value=0).noneable(),\n ('Number of choices to make. If set to None, any number of choices is '\n 'acceptable.')),\n ('candidates', pg_typing.List(pg_typing.Any()),\n ('Candidate values, which may contain nested hyper values.'\n 'Candidate can customize its display value (literal) by implementing the '\n '`pg.Formattable` interface.')),\n ('choices_distinct', pg_typing.Bool(True), 'Whether choices are distinct.'),\n ('choices_sorted', pg_typing.Bool(False), 'Whether choices are sorted.'),\n ('where', pg_typing.Callable([pg_typing.Object(base.HyperPrimitive)],\n returns=pg_typing.Bool()).noneable(),\n ('Callable object to filter nested hyper values. If None, all nested '\n 'hyper value will be included in the encoding/decoding process. '\n 'Otherwise only the hyper values on which `where` returns True will be '\n 'included. `where` can be useful to partition a search space into '\n 'separate optimization processes. '\n 'Please see `ObjectTemplate` docstr for details.'))\n])\nclass Choices(base.HyperPrimitive):\n \"\"\"Categorical choices from a list of candidates.\n\n Example::\n\n # A single categorical choice:\n v = pg.oneof([1, 2, 3])\n\n # A multiple categorical choice as a list:\n vs = pg.manyof(2, [1, 2, 3])\n\n # A hierarchical categorical choice:\n v2 = pg.oneof([\n 'foo',\n 'bar',\n pg.manyof(2, [1, 2, 3])\n ])\n\n See also:\n\n * :class:`pyglove.hyper.OneOf`\n * :class:`pyglove.hyper.ManyOf`\n * :func:`pyglove.oneof`\n * :func:`pyglove.manyof`\n * :func:`pyglove.permutate`\n \"\"\"\n\n def _on_bound(self):\n \"\"\"On members are bound.\"\"\"\n super()._on_bound()\n if self.num_choices > len(self.candidates) and self.choices_distinct:\n raise ValueError(\n f'{len(self.candidates)} candidates cannot produce '\n f'{self.num_choices} distinct choices.')\n self._candidate_templates = [\n object_template.ObjectTemplate(c, where=self.where)\n for c in self.candidates]\n # ValueSpec for candidate.\n self._value_spec = None\n\n def _update_children_paths(\n self, old_path: object_utils.KeyPath, new_path: object_utils.KeyPath):\n \"\"\"Customized logic to update children paths.\"\"\"\n super()._update_children_paths(old_path, new_path)\n for t in self._candidate_templates:\n t.root_path = self.sym_path\n\n @property\n def candidate_templates(self):\n \"\"\"Returns candidate templates.\"\"\"\n return self._candidate_templates\n\n @property\n def is_leaf(self) -> bool:\n \"\"\"Returns whether this is a leaf node.\"\"\"\n for t in self._candidate_templates:\n if not t.is_constant:\n return False\n return True\n\n def dna_spec(self,\n location: Optional[object_utils.KeyPath] = None) -> geno.Choices:\n \"\"\"Returns corresponding DNASpec.\"\"\"\n return geno.Choices(\n num_choices=self.num_choices,\n candidates=[ct.dna_spec() for ct in self._candidate_templates],\n literal_values=[self._literal_value(c)\n for i, c in enumerate(self.candidates)],\n distinct=self.choices_distinct,\n sorted=self.choices_sorted,\n hints=self.hints,\n name=self.name,\n location=location or object_utils.KeyPath())\n\n def _literal_value(\n self, candidate: Any, max_len: int = 120) -> Union[int, float, str]:\n \"\"\"Returns literal value for candidate.\"\"\"\n if isinstance(candidate, numbers.Number):\n return candidate\n\n literal = object_utils.format(candidate, compact=True,\n hide_default_values=True,\n hide_missing_values=True,\n strip_object_id=True)\n if len(literal) > max_len:\n literal = literal[:max_len - 3] + '...'\n return literal\n\n def _decode(self) -> List[Any]:\n \"\"\"Decode a DNA into a list of object.\"\"\"\n dna = self._dna\n if self.num_choices == 1:\n # Single choice.\n if not isinstance(dna.value, int):\n raise ValueError(\n object_utils.message_on_path(\n f'Did you forget to specify values for conditional choices?\\n'\n f'Expect integer for {self.__class__.__name__}. '\n f'Encountered: {dna!r}.', self.sym_path))\n if dna.value >= len(self.candidates):\n raise ValueError(\n object_utils.message_on_path(\n f'Choice out of range. Value: {dna.value!r}, '\n f'Candidates: {len(self.candidates)}.', self.sym_path))\n choices = [self._candidate_templates[dna.value].decode(\n geno.DNA(None, dna.children))]\n else:\n # Multi choices.\n if len(dna.children) != self.num_choices:\n raise ValueError(\n object_utils.message_on_path(\n f'Number of DNA child values does not match the number of '\n f'choices. Child values: {dna.children!r}, '\n f'Choices: {self.num_choices}.', self.sym_path))\n if self.choices_distinct or self.choices_sorted:\n sub_dna_values = [s.value for s in dna]\n if (self.choices_distinct\n and len(set(sub_dna_values)) != len(dna.children)):\n raise ValueError(\n object_utils.message_on_path(\n f'DNA child values should be distinct. '\n f'Encountered: {sub_dna_values}.', self.sym_path))\n if self.choices_sorted and sorted(sub_dna_values) != sub_dna_values:\n raise ValueError(\n object_utils.message_on_path(\n f'DNA child values should be sorted. '\n f'Encountered: {sub_dna_values}.', self.sym_path))\n choices = []\n for i, sub_dna in enumerate(dna):\n if not isinstance(sub_dna.value, int):\n raise ValueError(\n object_utils.message_on_path(\n f'Choice value should be int. '\n f'Encountered: {sub_dna.value}.',\n object_utils.KeyPath(i, self.sym_path)))\n if sub_dna.value >= len(self.candidates):\n raise ValueError(\n object_utils.message_on_path(\n f'Choice out of range. Value: {sub_dna.value}, '\n f'Candidates: {len(self.candidates)}.',\n object_utils.KeyPath(i, self.sym_path)))\n choices.append(self._candidate_templates[sub_dna.value].decode(\n geno.DNA(None, sub_dna.children)))\n return choices\n\n def encode(self, value: List[Any]) -> geno.DNA:\n \"\"\"Encode a list of values into DNA.\n\n Example::\n\n # DNA of an object containing a single OneOf.\n # {'a': 1} => DNA(0)\n {\n 'a': one_of([1, 2])\n }\n\n\n # DNA of an object containing multiple OneOfs.\n # {'b': 1, 'c': bar} => DNA([0, 1])\n {\n 'b': pg.oneof([1, 2]),\n 'c': pg.oneof(['foo', 'bar'])\n }\n\n # DNA of an object containing conditional space.\n # {'a': {'b': 1} => DNA(0, 0, 0)])\n # {'a': {'b': [4, 7]} => DNA(1, [(0, 1), 2])\n # {'a': {'b': 'bar'} => DNA(2)\n {\n 'a': {\n 'b': pg.oneof([\n pg.oneof([\n pg.oneof([1, 2]),\n pg.oneof(3, 4)]),\n pg.manyof(2, [\n pg.oneof([4, 5]),\n 6,\n 7\n ]),\n ]),\n 'bar',\n ])\n }\n }\n\n Args:\n value: A list of value that can match choice candidates.\n\n Returns:\n Encoded DNA.\n\n Raises:\n ValueError if value cannot be encoded.\n \"\"\"\n if not isinstance(value, list):\n raise ValueError(\n object_utils.message_on_path(\n f'Cannot encode value: value should be a list type. '\n f'Encountered: {value!r}.', self.sym_path))\n choices = []\n if self.num_choices is not None and len(value) != self.num_choices:\n raise ValueError(\n object_utils.message_on_path(\n f'Length of input list is different from the number of choices '\n f'({self.num_choices}). Encountered: {value}.', self.sym_path))\n for v in value:\n choice_id = None\n child_dna = None\n for i, b in enumerate(self._candidate_templates):\n succeeded, child_dna = b.try_encode(v)\n if succeeded:\n choice_id = i\n break\n if child_dna is None:\n raise ValueError(\n object_utils.message_on_path(\n f'Cannot encode value: no candidates matches with '\n f'the value. Value: {v!r}, Candidates: {self.candidates}.',\n self.sym_path))\n choices.append(geno.DNA(choice_id, [child_dna]))\n return geno.DNA(None, choices)\n\n\n@symbolic.members(\n [],\n init_arg_list=[\n 'num_choices', 'candidates', 'choices_distinct',\n 'choices_sorted', 'hints'\n ],\n # TODO(daiyip): Change to 'ManyOf' once existing code migrates to ManyOf.\n serialization_key='hyper.ManyOf',\n additional_keys=['pyglove.generators.genetic.ChoiceList']\n)\nclass ManyOf(Choices):\n \"\"\"N Choose K.\n\n Example::\n\n # Chooses 2 distinct candidates.\n v = pg.manyof(2, [1, 2, 3])\n\n # Chooses 2 non-distinct candidates.\n v = pg.manyof(2, [1, 2, 3], distinct=False)\n\n # Chooses 2 distinct candidates sorted by their indices.\n v = pg.manyof(2, [1, 2, 3], sorted=True)\n\n # Permutates the candidates.\n v = pg.permutate([1, 2, 3])\n\n # A hierarchical categorical choice:\n v2 = pg.manyof(2, [\n 'foo',\n 'bar',\n pg.oneof([1, 2, 3])\n ])\n\n See also:\n\n * :func:`pyglove.manyof`\n * :func:`pyglove.permutate`\n * :class:`pyglove.hyper.Choices`\n * :class:`pyglove.hyper.OneOf`\n * :class:`pyglove.hyper.Float`\n * :class:`pyglove.hyper.CustomHyper`\n \"\"\"\n\n def custom_apply(\n self,\n path: object_utils.KeyPath,\n value_spec: pg_typing.ValueSpec,\n allow_partial: bool,\n child_transform: Optional[Callable[\n [object_utils.KeyPath, pg_typing.Field, Any], Any]] = None\n ) -> Tuple[bool, 'Choices']:\n \"\"\"Validate candidates during value_spec binding time.\"\"\"\n # Check if value_spec directly accepts `self`.\n if value_spec.value_type and isinstance(self, value_spec.value_type):\n return (False, self)\n\n if self._value_spec:\n src_spec = self._value_spec\n dest_spec = value_spec\n if not dest_spec.is_compatible(src_spec):\n raise TypeError(\n object_utils.message_on_path(\n f'Cannot bind an incompatible value spec {dest_spec} '\n f'to {self.__class__.__name__} with bound spec {src_spec}.',\n path))\n return (False, self)\n\n list_spec = typing.cast(\n pg_typing.List,\n pg_typing.ensure_value_spec(\n value_spec, pg_typing.List(pg_typing.Any()), path))\n if list_spec:\n for i, c in enumerate(self.candidates):\n list_spec.element.value.apply(\n c,\n self._allow_partial,\n root_path=path + f'candidates[{i}]')\n self._value_spec = list_spec\n return (False, self)\n\n\n@symbolic.members(\n [\n ('num_choices', 1)\n ],\n init_arg_list=['candidates', 'hints', 'where'],\n serialization_key='hyper.OneOf',\n additional_keys=['pyglove.generators.genetic.ChoiceValue']\n)\nclass OneOf(Choices):\n \"\"\"N Choose 1.\n\n Example::\n\n # A single categorical choice:\n v = pg.oneof([1, 2, 3])\n\n # A hierarchical categorical choice:\n v2 = pg.oneof([\n 'foo',\n 'bar',\n pg.oneof([1, 2, 3])\n ])\n\n See also:\n\n * :func:`pyglove.oneof`\n * :class:`pyglove.hyper.Choices`\n * :class:`pyglove.hyper.ManyOf`\n * :class:`pyglove.hyper.Float`\n * :class:`pyglove.hyper.CustomHyper`\n \"\"\"\n\n def _on_bound(self):\n \"\"\"Event triggered when members are bound.\"\"\"\n super()._on_bound()\n assert self.num_choices == 1\n\n def _decode(self) -> Any:\n \"\"\"Decode a DNA into an object.\"\"\"\n return super()._decode()[0]\n\n def encode(self, value: Any) -> geno.DNA:\n \"\"\"Encode a value into a DNA.\"\"\"\n # NOTE(daiyip): Single choice DNA will automatically be pulled\n # up from children to current node. Thus we simply returns\n # encoded DNA from parent node.\n return super().encode([value])\n\n def custom_apply(\n self,\n path: object_utils.KeyPath,\n value_spec: pg_typing.ValueSpec,\n allow_partial: bool,\n child_transform: Optional[Callable[\n [object_utils.KeyPath, pg_typing.Field, Any], Any]] = None\n ) -> Tuple[bool, 'OneOf']:\n \"\"\"Validate candidates during value_spec binding time.\"\"\"\n # Check if value_spec directly accepts `self`.\n if value_spec.value_type and isinstance(self, value_spec.value_type):\n return (False, self)\n\n if self._value_spec:\n if not value_spec.is_compatible(self._value_spec):\n raise TypeError(\n object_utils.message_on_path(\n f'Cannot bind an incompatible value spec {value_spec} '\n f'to {self.__class__.__name__} with bound '\n f'spec {self._value_spec}.', path))\n return (False, self)\n\n for i, c in enumerate(self.candidates):\n value_spec.apply(\n c,\n self._allow_partial,\n root_path=path + f'candidates[{i}]')\n self._value_spec = value_spec\n return (False, self)\n\n#\n# Helper methods for creating hyper values.\n#\n\n\ndef oneof(candidates: Iterable[Any],\n *,\n name: Optional[str] = None,\n hints: Optional[Any] = None) -> Any:\n \"\"\"N choose 1.\n\n Example::\n\n @pg.members([\n ('x', pg.typing.Int())\n ])\n class A(pg.Object):\n pass\n\n # A single categorical choice:\n v = pg.oneof([1, 2, 3])\n\n # A complex type as candidate.\n v1 = pg.oneof(['a', {'x': 1}, A(1)])\n\n # A hierarchical categorical choice:\n v2 = pg.oneof([\n 'foo',\n 'bar',\n A(pg.oneof([1, 2, 3]))\n ])\n\n See also:\n\n * :class:`pyglove.hyper.OneOf`\n * :func:`pyglove.manyof`\n * :func:`pyglove.floatv`\n * :func:`pyglove.permutate`\n * :func:`pyglove.evolve`\n\n .. note::\n\n Under symbolic mode (by default), `pg.oneof` returns a ``pg.hyper.OneOf``\n object. Under dynamic evaluation mode, which is called under the context of\n :meth:`pyglove.hyper.DynamicEvaluationContext.collect` or\n :meth:`pyglove.hyper.DynamicEvaluationContext.apply`, it evaluates to\n a concrete candidate value.\n\n To use conditional search space in dynamic evaluation mode, the candidate\n should be wrapped with a `lambda` function, which is not necessary under\n symbolic mode. For example::\n\n pg.oneof([lambda: pg.oneof([0, 1], name='sub'), 2], name='root')\n\n Args:\n candidates: Candidates to select from. Items of candidate can be any type,\n therefore it can have nested hyper primitives, which forms a hierarchical\n search space.\n name: A name that can be used to identify a decision point in the search\n space. This is needed when the code to instantiate the same hyper\n primitive may be called multiple times under a\n `pg.DynamicEvaluationContext.collect` context or under a\n `pg.DynamicEvaluationContext.apply` context.\n hints: An optional value which acts as a hint for the controller.\n\n Returns:\n In symbolic mode, this function returns a `ChoiceValue`.\n In dynamic evaluation mode, this function returns one of the items in\n `candidates`.\n If evaluated under a `pg.DynamicEvaluationContext.apply` scope,\n this function will return the selected candidate.\n If evaluated under a `pg.DynamicEvaluationContext.collect`\n scope, it will return the first candidate.\n \"\"\"\n return OneOf(candidates=list(candidates), name=name, hints=hints)\n\n\ndef manyof(k: int,\n candidates: Iterable[Any],\n distinct: bool = True,\n sorted: bool = False, # pylint: disable=redefined-builtin\n *,\n name: Optional[str] = None,\n hints: Optional[Any] = None,\n **kwargs) -> Any:\n \"\"\"N choose K.\n\n Example::\n\n @pg.members([\n ('x', pg.typing.Int())\n ])\n class A(pg.Object):\n pass\n\n # Chooses 2 distinct candidates.\n v = pg.manyof(2, [1, 2, 3])\n\n # Chooses 2 non-distinct candidates.\n v = pg.manyof(2, [1, 2, 3], distinct=False)\n\n # Chooses 2 distinct candidates sorted by their indices.\n v = pg.manyof(2, [1, 2, 3], sorted=True)\n\n # A complex type as candidate.\n v1 = pg.manyof(2, ['a', {'x': 1}, A(1)])\n\n # A hierarchical categorical choice:\n v2 = pg.manyof(2, [\n 'foo',\n 'bar',\n A(pg.oneof([1, 2, 3]))\n ])\n\n .. note::\n\n Under symbolic mode (by default), `pg.manyof` returns a ``pg.hyper.ManyOf``\n object. Under dynamic evaluation mode, which is called under the context of\n :meth:`pyglove.hyper.DynamicEvaluationContext.collect` or\n :meth:`pyglove.hyper.DynamicEvaluationContext.apply`, it evaluates to\n a concrete candidate value.\n\n To use conditional search space in dynamic evaluate mode, the candidate\n should be wrapped with a `lambda` function, which is not necessary under\n symbolic mode. For example::\n\n pg.manyof(2, [\n lambda: pg.oneof([0, 1], name='sub_a'),\n lambda: pg.floatv(0.0, 1.0, name='sub_b'),\n lambda: pg.manyof(2, ['a', 'b', 'c'], name='sub_c')\n ], name='root')\n\n See also:\n\n * :class:`pyglove.hyper.ManyOf`\n * :func:`pyglove.manyof`\n * :func:`pyglove.floatv`\n * :func:`pyglove.permutate`\n * :func:`pyglove.evolve`\n\n Args:\n k: number of choices to make. Should be no larger than the length of\n `candidates` unless `choice_distinct` is set to False,\n candidates: Candidates to select from. Items of candidate can be any type,\n therefore it can have nested hyper primitives, which forms a hierarchical\n search space.\n distinct: If True, each choice needs to be unique.\n sorted: If True, choices are sorted by their indices in the\n candidates.\n name: A name that can be used to identify a decision point in the search\n space. This is needed when the code to instantiate the same hyper\n primitive may be called multiple times under a\n `pg.DynamicEvaluationContext.collect` context or a\n `pg.DynamicEvaluationContext.apply` context.\n hints: An optional value which acts as a hint for the controller.\n **kwargs: Keyword arguments for backward compatibility.\n `choices_distinct`: Old name for `distinct`.\n `choices_sorted`: Old name for `sorted`.\n\n Returns:\n In symbolic mode, this function returns a `Choices`.\n In dynamic evaluate mode, this function returns a list of items in\n `candidates`.\n If evaluated under a `pg.DynamicEvaluationContext.apply` scope,\n this function will return a list of selected candidates.\n If evaluated under a `pg.DynamicEvaluationContext.collect`\n scope, it will return a list of the first valid combination from the\n `candidates`. For example::\n\n # Evaluates to [0, 1, 2].\n manyof(3, range(5))\n\n # Evaluates to [0, 0, 0].\n manyof(3, range(5), distinct=False)\n \"\"\"\n choices_distinct = kwargs.pop('choices_distinct', distinct)\n choices_sorted = kwargs.pop('choices_sorted', sorted)\n return ManyOf(\n num_choices=k,\n candidates=list(candidates),\n choices_distinct=choices_distinct,\n choices_sorted=choices_sorted,\n name=name,\n hints=hints)\n\n\ndef permutate(candidates: Iterable[Any],\n name: Optional[str] = None,\n hints: Optional[Any] = None) -> Any:\n \"\"\"Permuatation of candidates.\n\n Example::\n\n @pg.members([\n ('x', pg.typing.Int())\n ])\n class A(pg.Object):\n pass\n\n # Permutates the candidates.\n v = pg.permutate([1, 2, 3])\n\n # A complex type as candidate.\n v1 = pg.permutate(['a', {'x': 1}, A(1)])\n\n # A hierarchical categorical choice:\n v2 = pg.permutate([\n 'foo',\n 'bar',\n A(pg.oneof([1, 2, 3]))\n ])\n\n .. note::\n\n Under symbolic mode (by default), `pg.manyof` returns a ``pg.hyper.ManyOf``\n object. Under dynamic evaluate mode, which is called under the context of\n :meth:`pyglove.hyper.DynamicEvaluationContext.collect` or\n :meth:`pyglove.hyper.DynamicEvaluationContext.apply`, it evaluates to\n a concrete candidate value.\n\n To use conditional search space in dynamic evaluate mode, the candidate\n should be wrapped with a `lambda` function, which is not necessary under\n symbolic mode. For example::\n\n pg.permutate([\n lambda: pg.oneof([0, 1], name='sub_a'),\n lambda: pg.floatv(0.0, 1.0, name='sub_b'),\n lambda: pg.manyof(2, ['a', 'b', 'c'], name='sub_c')\n ], name='root')\n\n See also:\n\n * :class:`pyglove.hyper.ManyOf`\n * :func:`pyglove.oneof`\n * :func:`pyglove.manyof`\n * :func:`pyglove.floatv`\n * :func:`pyglove.evolve`\n\n Args:\n candidates: Candidates to select from. Items of candidate can be any type,\n therefore it can have nested hyper primitives, which forms a hierarchical\n search space.\n name: A name that can be used to identify a decision point in the search\n space. This is needed when the code to instantiate the same hyper\n primitive may be called multiple times under a\n `pg.DynamicEvaluationContext.collect` context or a\n `pg.DynamicEvaluationContext.apply` context.\n hints: An optional value which acts as a hint for the controller.\n\n Returns:\n In symbolic mode, this function returns a `Choices`.\n In dynamic evaluate mode, this function returns a permutation from\n `candidates`.\n If evaluated under an `pg.DynamicEvaluationContext.apply` scope,\n this function will return a permutation of candidates based on controller\n decisions.\n If evaluated under a `pg.DynamicEvaluationContext.collect`\n scope, it will return the first valid permutation.\n For example::\n\n # Evaluates to [0, 1, 2, 3, 4].\n permutate(range(5), name='numbers')\n \"\"\"\n candidates = list(candidates)\n return manyof(\n len(candidates), candidates,\n choices_distinct=True, choices_sorted=False, name=name, hints=hints)\n","repo_name":"google/pyglove","sub_path":"pyglove/core/hyper/categorical.py","file_name":"categorical.py","file_ext":"py","file_size_in_byte":22595,"program_lang":"python","lang":"en","doc_type":"code","stars":306,"dataset":"github-code","pt":"81"}
+{"seq_id":"11139197940","text":"\nimport torch.nn as nn\n\nimport math\nimport torch.utils.model_zoo as model_zoo\nfrom torchvision.ops import nms\nfrom tortoise.models.retinanet.utils import BasicBlock, Bottleneck, BBoxTransform, ClipBoxes\nfrom tortoise.models.retinanet.anchors import Anchors\nfrom tortoise.models.retinanet import losses\nfrom tortoise.models.retinanet.dataloader import CocoDataset, Resizer, AspectRatioBasedSampler, Normalizer, Augmenter, collater\nfrom tortoise.utils import timetaken, download_file\n\nfrom torch.utils.data import DataLoader\nfrom torchvision import transforms\nimport torch\n\nfrom pycocotools.cocoeval import COCOeval\nimport os, json, time, copy, yaml\nfrom tqdm import tqdm\n\n# https://github.com/Delgan/loguru\nfrom loguru import logger\n# logger.info(\"If you're using Python {}, prefer {feature} of course!\", 3.6, feature=\"f-strings\")\n\nclass PyramidFeatures(nn.Module):\n def __init__(self, C3_size, C4_size, C5_size, feature_size=256):\n super(PyramidFeatures, self).__init__()\n\n # upsample C5 to get P5 from the FPN paper\n self.P5_1 = nn.Conv2d(C5_size, feature_size, kernel_size=1, stride=1, padding=0)\n self.P5_upsampled = nn.Upsample(scale_factor=2, mode='nearest')\n self.P5_2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, stride=1, padding=1)\n\n # add P5 elementwise to C4\n self.P4_1 = nn.Conv2d(C4_size, feature_size, kernel_size=1, stride=1, padding=0)\n self.P4_upsampled = nn.Upsample(scale_factor=2, mode='nearest')\n self.P4_2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, stride=1, padding=1)\n\n # add P4 elementwise to C3\n self.P3_1 = nn.Conv2d(C3_size, feature_size, kernel_size=1, stride=1, padding=0)\n self.P3_2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, stride=1, padding=1)\n\n # \"P6 is obtained via a 3x3 stride-2 conv on C5\"\n self.P6 = nn.Conv2d(C5_size, feature_size, kernel_size=3, stride=2, padding=1)\n\n # \"P7 is computed by applying ReLU followed by a 3x3 stride-2 conv on P6\"\n self.P7_1 = nn.ReLU()\n self.P7_2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, stride=2, padding=1)\n\n def forward(self, inputs):\n C3, C4, C5 = inputs\n\n P5_x = self.P5_1(C5)\n P5_upsampled_x = self.P5_upsampled(P5_x)\n P5_x = self.P5_2(P5_x)\n\n P4_x = self.P4_1(C4)\n P4_x = P5_upsampled_x + P4_x\n P4_upsampled_x = self.P4_upsampled(P4_x)\n P4_x = self.P4_2(P4_x)\n\n P3_x = self.P3_1(C3)\n P3_x = P3_x + P4_upsampled_x\n P3_x = self.P3_2(P3_x)\n\n P6_x = self.P6(C5)\n\n P7_x = self.P7_1(P6_x)\n P7_x = self.P7_2(P7_x)\n\n return [P3_x, P4_x, P5_x, P6_x, P7_x]\nclass RegressionModel(nn.Module):\n def __init__(self, num_features_in, num_anchors=9, feature_size=256):\n super(RegressionModel, self).__init__()\n\n self.conv1 = nn.Conv2d(num_features_in, feature_size, kernel_size=3, padding=1)\n self.act1 = nn.ReLU()\n\n self.conv2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act2 = nn.ReLU()\n\n self.conv3 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act3 = nn.ReLU()\n\n self.conv4 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act4 = nn.ReLU()\n\n self.output = nn.Conv2d(feature_size, num_anchors * 4, kernel_size=3, padding=1)\n\n def forward(self, x):\n out = self.conv1(x)\n out = self.act1(out)\n\n out = self.conv2(out)\n out = self.act2(out)\n\n out = self.conv3(out)\n out = self.act3(out)\n\n out = self.conv4(out)\n out = self.act4(out)\n\n out = self.output(out)\n\n # out is B x C x W x H, with C = 4*num_anchors\n out = out.permute(0, 2, 3, 1)\n\n return out.contiguous().view(out.shape[0], -1, 4)\nclass ClassificationModel(nn.Module):\n def __init__(self, num_features_in, num_anchors=9, num_classes=80, prior=0.01, feature_size=256):\n super(ClassificationModel, self).__init__()\n\n self.num_classes = num_classes\n self.num_anchors = num_anchors\n\n self.conv1 = nn.Conv2d(num_features_in, feature_size, kernel_size=3, padding=1)\n self.act1 = nn.ReLU()\n\n self.conv2 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act2 = nn.ReLU()\n\n self.conv3 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act3 = nn.ReLU()\n\n self.conv4 = nn.Conv2d(feature_size, feature_size, kernel_size=3, padding=1)\n self.act4 = nn.ReLU()\n\n self.output = nn.Conv2d(feature_size, num_anchors * num_classes, kernel_size=3, padding=1)\n self.output_act = nn.Sigmoid()\n\n def forward(self, x):\n out = self.conv1(x)\n out = self.act1(out)\n\n out = self.conv2(out)\n out = self.act2(out)\n\n out = self.conv3(out)\n out = self.act3(out)\n\n out = self.conv4(out)\n out = self.act4(out)\n\n out = self.output(out)\n out = self.output_act(out)\n\n # out is B x C x W x H, with C = n_classes + n_anchors\n out1 = out.permute(0, 2, 3, 1)\n\n batch_size, width, height, channels = out1.shape\n\n out2 = out1.view(batch_size, width, height, self.num_anchors, self.num_classes)\n\n return out2.contiguous().view(x.shape[0], -1, self.num_classes)\nclass ResNet(nn.Module):\n\n def __init__(self, num_classes, block, layers):\n self.inplanes = 64\n super(ResNet, self).__init__()\n self.conv1 = nn.Conv2d(3, 64, kernel_size=7, stride=2, padding=3, bias=False)\n self.bn1 = nn.BatchNorm2d(64)\n self.relu = nn.ReLU(inplace=True)\n self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n self.layer1 = self._make_layer(block, 64, layers[0])\n self.layer2 = self._make_layer(block, 128, layers[1], stride=2)\n self.layer3 = self._make_layer(block, 256, layers[2], stride=2)\n self.layer4 = self._make_layer(block, 512, layers[3], stride=2)\n\n if block == BasicBlock:\n fpn_sizes = [self.layer2[layers[1] - 1].conv2.out_channels, self.layer3[layers[2] - 1].conv2.out_channels,\n self.layer4[layers[3] - 1].conv2.out_channels]\n elif block == Bottleneck:\n fpn_sizes = [self.layer2[layers[1] - 1].conv3.out_channels, self.layer3[layers[2] - 1].conv3.out_channels,\n self.layer4[layers[3] - 1].conv3.out_channels]\n else:\n raise ValueError(f\"Block type {block} not understood\")\n\n self.fpn = PyramidFeatures(fpn_sizes[0], fpn_sizes[1], fpn_sizes[2])\n self.regressionModel = RegressionModel(256)\n self.classificationModel = ClassificationModel(256, num_classes=num_classes)\n\n for m in self.modules():\n if isinstance(m, nn.Conv2d):\n n = m.kernel_size[0] * m.kernel_size[1] * m.out_channels\n m.weight.data.normal_(0, math.sqrt(2. / n))\n elif isinstance(m, nn.BatchNorm2d):\n m.weight.data.fill_(1)\n m.bias.data.zero_()\n\n prior = 0.01\n self.classificationModel.output.weight.data.fill_(0)\n self.classificationModel.output.bias.data.fill_(-math.log((1.0 - prior) / prior))\n self.regressionModel.output.weight.data.fill_(0)\n self.regressionModel.output.bias.data.fill_(0)\n self.freeze_bn()\n\n def _make_layer(self, block, planes, blocks, stride=1):\n downsample = None\n if stride != 1 or self.inplanes != planes * block.expansion:\n downsample = nn.Sequential(\n nn.Conv2d(self.inplanes, planes * block.expansion, kernel_size=1, stride=stride, bias=False),\n nn.BatchNorm2d(planes * block.expansion)\n )\n layers = [block(self.inplanes, planes, stride, downsample)]\n self.inplanes = planes * block.expansion\n for i in range(1, blocks):\n layers.append(block(self.inplanes, planes))\n\n return nn.Sequential(*layers)\n\n def freeze_bn(self):\n '''Freeze BatchNorm layers.'''\n for layer in self.modules():\n if isinstance(layer, nn.BatchNorm2d):\n layer.eval()\n\n def forward(self, inputs):\n img_batch = inputs\n x = self.conv1(img_batch)\n x = self.bn1(x)\n x = self.relu(x)\n x = self.maxpool(x)\n\n x1 = self.layer1(x)\n x2 = self.layer2(x1)\n x3 = self.layer3(x2)\n x4 = self.layer4(x3)\n\n features = self.fpn([x2, x3, x4])\n regression = torch.cat([self.regressionModel(feature) for feature in features], dim=1)\n classification = torch.cat([self.classificationModel(feature) for feature in features], dim=1)\n return regression, classification\n\nclass RetinanetModel:\n def __init__(self):\n self.model_dir = \"weights\"\n self.exp_name = \"experiment\"\n self.model_config = {\n \"resnet18\": {\n \"block\": BasicBlock,\n \"layer\": [2, 2, 2, 2],\n \"resnet_url\": \"https://download.pytorch.org/models/resnet18-5c106cde.pth\",\n \"retinanet_url\": \"/media/nk/PortableSSD1/RnD/pytorch-multimodels/weights/resnet18/experiment_7/best.pth\",\n \"config\": \"/media/nk/PortableSSD1/RnD/pytorch-multimodels/data/hyps/retinanet/hyp.scratch-low.yaml\",\n \"sha256\": \"\"\n },\n \"resnet34\": {\n \"block\": BasicBlock,\n \"layer\": [3, 4, 6, 3],\n \"resnet_url\": \"https://download.pytorch.org/models/resnet34-333f7ec4.pth\",\n \"retinanet_url\": \"https://download.pytorch.org/models/resnet34-333f7ec4.pth\",\n \"sha256\": \"\"\n },\n \"resnet50\": {\n \"block\": Bottleneck,\n \"layer\": [3, 4, 6, 3],\n \"resnet_url\": \"https://download.pytorch.org/models/resnet50-19c8e357.pth\",\n \"retinanet_url\": \"https://download.pytorch.org/models/resnet50-19c8e357.pth\",\n \"sha256\": \"\"\n },\n \"resnet101\": {\n \"block\": Bottleneck,\n \"layer\": [3, 4, 23, 3],\n \"resnet_url\": \"https://download.pytorch.org/models/resnet101-5d3b4d8f.pth\",\n \"retinanet_url\": \"https://download.pytorch.org/models/resnet101-5d3b4d8f.pth\",\n \"sha256\": \"\"\n },\n \"resnet152\": {\n \"block\": Bottleneck,\n \"layer\": [3, 8, 36, 3],\n \"resnet_url\": \"https://download.pytorch.org/models/resnet152-b121ed2d.pth\",\n \"retinanet_url\": \"https://download.pytorch.org/models/resnet152-b121ed2d.pth\",\n \"sha256\": \"\"\n }\n }\n \n @classmethod\n def load_config(self, backbone, filename, model_dir):\n from munch import DefaultMunch\n if not os.path.exists(f\"{model_dir}/{filename}\"):\n download_file(self.model_config[backbone][\"config\"], filename, model_dir)\n with open(f\"{model_dir}/{filename}\", \"r\") as stream:\n try: config = yaml.safe_load(stream)\n except yaml.YAMLError as exc:\n print(exc); exit()\n # https://stackoverflow.com/a/24852544\n return DefaultMunch.fromDict(config) \n \n def get_info(self, backbone): return self.model_config[backbone][\"retinanet_url\"]\n @classmethod\n def from_scratch(self, backbone, model_dir=None, device=\"cuda\"):\n self.__init__(self)\n if model_dir: self.model_dir = model_dir \n if backbone not in self.model_config.keys(): \n logger.debug(f\"Backbone not Found: {backbone}\")\n assert False, \"backbone Not found\" \n \n self.args = self.load_config(backbone, \"config.yaml\", os.path.join(self.model_dir, backbone))\n self.args.device = device\n os.makedirs(os.path.join(self.model_dir, backbone), exist_ok=True)\n self.args.backbone = backbone\n self.model = ResNet(self.args.num_classes, self.model_config[backbone][\"block\"], self.model_config[backbone][\"layer\"])\n self.model.load_state_dict(model_zoo.load_url(self.model_config[backbone]['resnet_url'], model_dir=os.path.join(self.model_dir, backbone)), strict=False)\n self.args.exp_name = f\"{os.path.join(self.model_dir, backbone)}/{self.exp_name}_{len(os.listdir(os.path.join(self.model_dir, backbone)))}\"\n os.makedirs(f\"{self.args.exp_name }/\", exist_ok=True)\n logger.info(f\"Export directory: {self.args.exp_name}\")\n return self()\n \n @classmethod\n def from_pretrained(self, backbone, model_dir=None, exp_name=None, filename=\"best.pth\", device=\"cuda\"):\n self.__init__(self)\n if model_dir: self.model_dir = model_dir \n if backbone not in self.model_config.keys(): \n logger.debug(f\"Backbone not Found: {backbone}\")\n assert False, \"backbone Not found\"\n\n self.args = self.load_config(backbone, \"config.yaml\", os.path.join(self.model_dir, backbone))\n self.args.device = device\n os.makedirs(os.path.join(self.model_dir, backbone), exist_ok=True)\n self.args.backbone = backbone\n self.fun_exp_name = exp_name\n self.model = ResNet(self.args.num_classes, self.model_config[backbone][\"block\"], self.model_config[backbone][\"layer\"])\n if exp_name:\n self.model.load_state_dict(torch.load(os.path.join(self.model_dir, backbone, exp_name, filename))[\"state_dict\"])\n else:\n # self.model.load_state_dict(model_zoo.load_url(self.model_config[backbone]['retinanet_url'])[\"state_dict\"], model_dir=os.path.join(self.model_dir, backbone))\n self.model.load_state_dict(torch.load(self.model_config[backbone]['retinanet_url'])[\"state_dict\"])\n self.args.exp_name = f\"{os.path.join(self.model_dir, backbone)}/{exp_name}\" if exp_name else \\\n f\"{os.path.join(self.model_dir, backbone)}/{self.exp_name}_{len(os.listdir(os.path.join(self.model_dir, backbone)))}\"\n os.makedirs(f\"{self.args.exp_name }/\", exist_ok=True)\n logger.info(f\"Export directory: {self.args.exp_name}\")\n return self()\n \n @timetaken\n def run_oneepoch(self, dataloader_train, scaler, epoch):\n epoch_loss = torch.zeros(len(dataloader_train))\n pbar = tqdm(enumerate(dataloader_train), total=len(dataloader_train), desc=\"Epoch: {0}\".format(epoch))\n self.focalLoss = losses.FocalLoss()\n self.anchors = Anchors()\n for iteration, data in pbar:\n self.optimizer.zero_grad()\n with torch.cuda.amp.autocast():\n regressions, classifications = self.model(data['img'].to(self.args.device).float())\n classification_loss, regression_loss = self.focalLoss(classifications, regressions, self.anchors(data['img'].to(self.args.device).float()), data['annot']) \n classification_loss, regression_loss = classification_loss.mean(), regression_loss.mean()\n loss = classification_loss + regression_loss\n loss = loss * 2\n torch.nn.utils.clip_grad_norm_(self.model.parameters(), 0.1)\n scaler.scale(loss).backward()\n scaler.step(self.optimizer)\n scaler.update()\n\n epoch_loss[iteration] = loss.cpu().item()\n tb_lr = [x['lr'] for x in self.optimizer.param_groups][0]\n pbar.set_postfix_str(\"Loss: {:.6f} lr: {:.10f}\".format(epoch_loss.sum()/iteration, tb_lr))\n return epoch_loss.mean().item()\n \n @timetaken\n def train(self, best_acc=0.0):\n dataset_train = CocoDataset(self.args.coco_path, set_name='train2017',transform=transforms.Compose([Normalizer(), Augmenter(), Resizer()]))\n sampler = AspectRatioBasedSampler(dataset_train, batch_size=self.args.train_batchsize, drop_last=False)\n dataloader_train = DataLoader(dataset_train, num_workers=8, collate_fn=collater, batch_sampler=sampler)\n self.optimizer = torch.optim.AdamW(self.model.parameters(), lr=self.args.lr, weight_decay=self.args.weight_decay, betas=(0.9, 0.95))\n if self.fun_exp_name: self.optimizer.load_state_dict(torch.load(os.path.join(self.args.exp_name, \"best.pth\"))[\"optimizer\"])\n self.scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(self.optimizer, patience=3, verbose=True)\n scaler = torch.cuda.amp.GradScaler()\n self.model = torch.nn.DataParallel(self.model); self.training = True\n for epoch in range(self.args.start_epoch, self.args.total_epoch):\n self.model.train()\n self.model.module.freeze_bn()\n epoch_loss = self.run_oneepoch(dataloader_train, scaler, epoch)\n ap50_95, ap50 = self.validation()\n self.scheduler.step()\n if ap50_95 > best_acc: \n torch.save({\n \"iterations\": epoch * len(dataloader_train),\n \"state_dict\": self.model.module.state_dict(),\n \"optimizer\": self.optimizer.state_dict(),\n \"acc\": [ap50_95, ap50]\n }, f\"{self.args.exp_name}/best.pth\")\n best_acc = copy.deepcopy(ap50_95)\n torch.save({\n \"iterations\": epoch * len(dataloader_train),\n \"state_dict\": self.model.module.state_dict(),\n \"optimizer\": self.optimizer.state_dict(),\n \"acc\": [ap50_95, ap50]\n }, f\"{self.args.exp_name}/last.pth\")\n time.sleep(300)\n\n @torch.no_grad()\n def validation(self, model=None, threshold=0.05):\n self.model = model if isinstance(model, torch.nn.Module) else self.model\n self.model.eval().to(self.args.device)\n dataset = CocoDataset(self.args.coco_path, set_name='val2017', transform=transforms.Compose([Normalizer(), Resizer()]))\n results, image_ids = [], []\n self.anchors = Anchors()\n self.regressBoxes = BBoxTransform()\n self.clipBoxes = ClipBoxes()\n for index in tqdm(range(len(dataset)), desc=\"Validation \"):\n if index == 500 and not threshold: break\n data = dataset[index]\n scale = data['scale']\n img_batch = data['img'].permute(2, 0, 1).to(self.args.device).float().unsqueeze(dim=0)\n anchors = self.anchors(img_batch)\n regression, classification = self.model(img_batch)\n transformed_anchors = self.regressBoxes(anchors, regression)\n transformed_anchors = self.clipBoxes(transformed_anchors, img_batch)\n\n finalResult = [[], [], []]\n finalScores = torch.Tensor([])\n finalAnchorBoxesIndexes = torch.Tensor([]).long()\n finalAnchorBoxesCoordinates = torch.Tensor([])\n if torch.cuda.is_available():\n finalScores = finalScores.cuda()\n finalAnchorBoxesIndexes = finalAnchorBoxesIndexes.cuda()\n finalAnchorBoxesCoordinates = finalAnchorBoxesCoordinates.cuda()\n\n for i in range(classification.shape[2]):\n scores = torch.squeeze(classification[:, :, i])\n scores_over_thresh = (scores > 0.05)\n if scores_over_thresh.sum() == 0:\n # no boxes to NMS, just continue\n continue\n\n scores = scores[scores_over_thresh]\n anchorBoxes = torch.squeeze(transformed_anchors)\n anchorBoxes = anchorBoxes[scores_over_thresh]\n anchors_nms_idx = nms(anchorBoxes, scores, 0.5)\n\n finalResult[0].extend(scores[anchors_nms_idx])\n finalResult[1].extend(torch.tensor([i] * anchors_nms_idx.shape[0]))\n finalResult[2].extend(anchorBoxes[anchors_nms_idx])\n\n finalScores = torch.cat((finalScores, scores[anchors_nms_idx]))\n finalAnchorBoxesIndexesValue = torch.tensor([i] * anchors_nms_idx.shape[0])\n if torch.cuda.is_available():\n finalAnchorBoxesIndexesValue = finalAnchorBoxesIndexesValue.cuda()\n finalAnchorBoxesIndexes = torch.cat((finalAnchorBoxesIndexes, finalAnchorBoxesIndexesValue))\n finalAnchorBoxesCoordinates = torch.cat((finalAnchorBoxesCoordinates, anchorBoxes[anchors_nms_idx]))\n scores, labels, boxes = [finalScores, finalAnchorBoxesIndexes, finalAnchorBoxesCoordinates]\n scores, labels, boxes = scores.cpu(), labels.cpu(), boxes.cpu()\n boxes /= scale\n if boxes.shape[0] > 0:\n boxes[:, 2] -= boxes[:, 0]\n boxes[:, 3] -= boxes[:, 1]\n for box_id in range(boxes.shape[0]):\n score = float(scores[box_id])\n label = int(labels[box_id])\n box = boxes[box_id, :]\n if score < 0.05: break\n image_result = {\n 'image_id' : dataset.image_ids[index],\n 'category_id' : dataset.label_to_coco_label(label),\n 'score' : float(score),\n 'bbox' : box.tolist(),\n }\n results.append(image_result)\n image_ids.append(dataset.image_ids[index])\n if not len(results):\n logger.debug(\"No Object is detected...!!!\"); return 0.0, 0.0\n json.dump(results, open(f'{self.args.exp_name}/{dataset.set_name}_bbox_results.json', 'w'), indent=4)\n coco_true = dataset.coco\n coco_pred = coco_true.loadRes(f'{self.args.exp_name}/{dataset.set_name}_bbox_results.json')\n coco_eval = COCOeval(coco_true, coco_pred, 'bbox')\n coco_eval.params.imgIds = image_ids\n coco_eval.evaluate()\n coco_eval.accumulate()\n coco_eval.summarize()\n \n # ap50_95, ap50\n return coco_eval.stats[0], coco_eval.stats[1]\n\n @timetaken\n def export(self, dummy_input, file_path):\n self.model.eval()\n torch.onnx.export(self.model, dummy_input, file_path, opset_version=12)\n\n # @timetaken\n def quantize(self, dummy_input: torch.nn.Module):\n from aimet_torch.model_preparer import prepare_model\n from aimet_common.defs import QuantScheme\n from aimet_torch.quantsim import QuantizationSimModel\n device = dummy_input.device\n # dummy_input = torch.randn(dummy_input)\n model = prepare_model(self.model).to(device)\n download_file(\n \"https://raw.githubusercontent.com/quic/aimet/develop/TrainingExtensions/common/src/python/aimet_common/quantsim_config/default_config_per_channel.json\",\n \"pcq_config.json\", os.path.join(self.model_dir, self.args.backbone)\n )\n quant_sim = QuantizationSimModel(model, dummy_input=dummy_input,\n quant_scheme=QuantScheme.post_training_tf_enhanced,\n default_param_bw=8, default_output_bw=8,\n config_file=os.path.join(self.model_dir, self.args.backbone, \"pcq_config.json\"))\n quant_sim.compute_encodings(self.validation, forward_pass_callback_args=(None))\n self.validation(quant_sim.model)\n os.makedirs(os.path.join(self.args.exp_name, \"quant\"), exist_ok=True)\n quant_sim.model.cpu()\n quant_sim.export(path=os.path.join(self.args.exp_name, \"quant\"), filename_prefix=f'quantized_{self.args.backbone}', dummy_input=dummy_input.cpu())\n \n \n @timetaken\n def predict(img, save_image=True):\n return []\n \n ","repo_name":"nkdatascientist/tortoise-models","sub_path":"tortoise/models/retinanet/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":23706,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"3774007533","text":"import datetime as dt\nimport plotly.graph_objects as go\nimport pandas as pd\n\n\nclass PriceIndependentIndicator:\n\n def __init__(self, df: pd.DataFrame, base_indicator: str):\n self.df_plot = df.copy()\n self.fig = go.Figure()\n self.base_indicator = base_indicator\n\n self.add_timestring()\n self.create_line_fig(base_indicator)\n\n\n def add_timestring(self):\n \"\"\"Hack for removing weekend candles\"\"\"\n self.df_plot['sTime'] = [dt.datetime.strftime(x, \"s%y-%m-%d %H:%M\") for x in self.df_plot.time]\n\n\n def create_line_fig(self, base_indicator):\n self.fig.add_trace(go.Scatter(\n x=self.df_plot.sTime,\n y=self.df_plot[f'{base_indicator}'],\n mode='lines',\n name=f'{base_indicator}'\n ))\n\n\n def add_secondary_line_based_indicators(self, indicators: list):\n for indicator in indicators:\n self.fig.add_trace(go.Scatter(\n x=self.df_plot.sTime,\n y=self.df_plot[f\"{indicator}\"],\n mode='lines',\n name=f'{indicator}'\n ))\n\n def add_bar_based_indicators(self, indicators: list):\n for indicator in indicators:\n self.fig.add_trace(go.Bar(\n x=self.df_plot.sTime,\n y=self.df_plot[f\"{indicator}\"],\n name=f'{indicator}'\n ))\n\n\n # Show and Customise plot #\n\n\n def update_layout(self, width, height, ticks):\n self.fig.update_xaxes(\n gridcolor=\"#1f292f\",\n nticks=ticks,\n )\n\n self.fig.update_yaxes(\n gridcolor=\"#1f292f\"\n )\n\n self.fig.update_layout(\n width=width,\n height=height,\n paper_bgcolor=\"#2c303c\",\n plot_bgcolor=\"#2c303c\",\n margin=dict(l=10, r=10, b=10, t=10),\n font=dict(size=8, color=\"#e1e1e1\")\n )\n\n def show_plot(self, width=1600, height=900, ticks=5):\n self.update_layout(width, height, ticks)\n self.fig.show()","repo_name":"LMBroadhurst/algobot_statistical_finance_dashboard","sub_path":"oanda_trading_bot/charting/price_independent_indicator.py","file_name":"price_independent_indicator.py","file_ext":"py","file_size_in_byte":2035,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"69910862984","text":"import pandas as pd\n\nimport math\n\n\nnumber_of_members = 0\nsm =['Sheet1','Sheet3','Sheet4','Sheet5'] # make sure to add the names of all the sheets here under sheetname variable sm.\ncn = ['HQ','P BTY','Q BTY','R BTY'] # make sure to add all the names of the companies.\ncount=0 \nm=0\n\nwriter = pd.ExcelWriter('final_attendence.xlsx', engine='xlsxwriter') # instead of 'final.xlsx', enter path of the new file to be created (This is where the data has to be saved)\nfor i in sm:\n \n status=[]\n df = pd.read_excel('test_final_1.xlsx', sheet_name= i) # instead of 'attendence.xlsx', enter path of the attendance sheet from which data has to be taken.\n \n\n #convert each column to list, remove NAN values \n name = df.columns[3]\n filtered_column_names = df.loc[9:,name].loc[~df[name].isin(['PARBAT ALI','Name'])].tolist()\n ecode = df.columns[2]\n filtered_column_code = df.loc[9:,ecode].loc[~df[ecode].isin(['E. Code'])].tolist()\n TD_time = df.columns[10] #TD Time is same as A in time\n filtered_column_TD = df.loc[9:,TD_time].loc[~df[TD_time].isin(['A. InTime'])].fillna(-1).tolist()\n SE_time = df.columns[8] #SE time is same as S out time\n filtered_column_SE = df.loc[9:,SE_time].loc[~df[SE_time].isin(['S. OutTime'])].fillna(-1).tolist()\n Sin = df.columns[6]\n filtered_column_Sin = df.iloc[9:df[Sin].last_valid_index() + 1, df.columns.get_loc(Sin)].fillna(-1).tolist()\n\n\n\n size = len(filtered_column_names)\n members =[]\n x=0\n while x = 0 else len(items_copy) + index\n indices.append(normalized_index)\n indices.sort(reverse=True)\n\n if self.whitelist:\n index = len(items)-1\n while index >= 0:\n if index not in indices:\n del items_copy[index]\n index -= 1\n else:\n for index in indices:\n if index < len(items_copy):\n del items_copy[index]\n\n return items_copy\n","repo_name":"narranoid/Blender-Texture-Grapher","sub_path":"stringfiddle/format/filter/index_filter.py","file_name":"index_filter.py","file_ext":"py","file_size_in_byte":849,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"18661656687","text":"import unittest\nfrom unittest.mock import patch\nfrom io import StringIO\nfrom gradescope_utils.autograder_utils.decorators import weight, visibility\ntry:\n import hw11 as hw11\nexcept Exception:\n raise Exception(f'Could not process your file, remember to code under the main() function, not call the main() function, and set the file name correctly.')\nimport solution.hw11 as solution\nimport random\n\nclass TestHW(unittest.TestCase):\n def setUpAns(self):\n solution.main()\n ansFile = open(\"[Solution]StudentSubmissions.txt\", \"r\")\n ansData = ansFile.readlines()\n return ansData\n \n def setUpSub(self):\n hw11.main()\n subFile = open(\"StudentSubmissions.txt\", \"r\")\n subData = subFile.readlines()\n return subData\n \n # @visibility(\"hidden\")\n @weight(9) \n def test_hw(self):\n ans = self.setUpAns()\n submission = self.setUpSub()\n self.assertEqual(ans, submission)\n \n","repo_name":"pnngocdoan/autograder_SUA","sub_path":"Week 4/hw11/tests/test_hw.py","file_name":"test_hw.py","file_ext":"py","file_size_in_byte":961,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14905076895","text":"from tensorflow.keras.models import Model\nfrom tensorflow.keras.layers import Conv2D, Dropout, BatchNormalization, Input, SeparableConv2D, GlobalAveragePooling2D, Dense, MaxPooling2D, Activation, Flatten\nfrom tensorflow.keras.layers import add as add_concat\nfrom keras import backend as K\nfrom constants import strategy, bnmomemtum, SHAPE\nwith strategy.scope():\n def fire(x, filters, kernel_size):\n if not isinstance(filters, list):\n filters = [filters, filters]\n x = SeparableConv2D(filters[0], kernel_size, padding='same', use_bias=False)(x)\n x = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(x)\n x = Activation('relu')(x)\n x = SeparableConv2D(filters[1], kernel_size, padding='same', use_bias=False)(x)\n return BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(x)\n\n def fire_module_separable_conv(filters, kernel_size=(3, 3)):\n return lambda x: fire(x, filters, kernel_size)\n\n channel_axis = 1 if K.image_data_format() == 'channels_first' else -1\n img_input = Input(shape=SHAPE)\n\n x = Conv2D(32, (3, 3), strides=(2, 2), use_bias=False)(img_input)\n x = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(x)\n x = Activation('relu')(x)\n\n x = Conv2D(64, (3, 3), use_bias=False, name='block1_conv2')(x)\n x = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(x)\n x = Activation('relu')(x)\n\n residual = Conv2D(128, (1, 1), strides=(2, 2), padding='same', use_bias=False)(x)\n residual = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(residual)\n\n x = fire_module_separable_conv(128)(x)\n\n x = MaxPooling2D((3, 3), strides=(2, 2), padding='same')(x)\n x = add_concat([x, residual])\n\n residual = Conv2D(256, (1, 1), strides=(2, 2), padding='same', use_bias=False)(x)\n residual = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(residual)\n\n x = Activation('relu')(x)\n x = fire_module_separable_conv(256)(x)\n\n x = MaxPooling2D((3, 3), strides=(2, 2), padding='same')(x)\n x = add_concat([x, residual])\n\n for i in range(4):\n residual = x\n\n x = Activation('relu')(x)\n x = SeparableConv2D(256, (3, 3), padding='same', use_bias=False)(x)\n x = BatchNormalization(axis=channel_axis, center=True, scale=False, momentum=bnmomemtum)(x)\n x = Activation('relu')(x)\n x = fire_module_separable_conv(256)(x)\n\n x = add_concat([x, residual])\n\n\n x = fire_module_separable_conv([728, 1024])(x)\n x = Activation('relu')(x)\n y = GlobalAveragePooling2D()(x)\n\n y = Dense(3096)(y)\n y = Activation('relu')(y)\n y = Dropout(0.3)(y)\n\n y = Dense(1548)(y)\n y = Activation('relu')(y)\n y = Dropout(0.3)(y)\n\n head_root = Dense(168, activation = 'softmax', name='head_root')(y)\n head_vowel = Dense(11, activation = 'softmax', name='head_vowel')(y)\n head_consonant = Dense(7, activation = 'softmax', name='head_consonant')(y)\n\n model = Model(inputs=img_input, outputs=[head_root, head_vowel, head_consonant])\n\nmodel.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n","repo_name":"MaximDrobchak/ML-Bengali_Handwritten_Grapheme_Classification","sub_path":"model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":3280,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"20252199710","text":"from typing import Sequence\n\nimport attrs\n\nfrom apischema import deserialize, serialize, settings\nfrom apischema.json_schema import deserialization_schema\nfrom apischema.objects import ObjectField\n\nprev_default_object_fields = settings.default_object_fields\n\n\ndef attrs_fields(cls: type) -> Sequence[ObjectField] | None:\n if hasattr(cls, \"__attrs_attrs__\"):\n return [\n ObjectField(\n a.name, a.type, required=a.default == attrs.NOTHING, default=a.default\n )\n for a in getattr(cls, \"__attrs_attrs__\")\n ]\n else:\n return prev_default_object_fields(cls)\n\n\nsettings.default_object_fields = attrs_fields\n\n\n@attrs.define\nclass Foo:\n bar: int\n\n\nassert deserialize(Foo, {\"bar\": 0}) == Foo(0)\nassert serialize(Foo, Foo(0)) == {\"bar\": 0}\nassert deserialization_schema(Foo) == {\n \"$schema\": \"http://json-schema.org/draft/2020-12/schema#\",\n \"type\": \"object\",\n \"properties\": {\"bar\": {\"type\": \"integer\"}},\n \"required\": [\"bar\"],\n \"additionalProperties\": False,\n}\n","repo_name":"wyfo/apischema","sub_path":"examples/examples/attrs_support.py","file_name":"attrs_support.py","file_ext":"py","file_size_in_byte":1039,"program_lang":"python","lang":"en","doc_type":"code","stars":206,"dataset":"github-code","pt":"81"}
+{"seq_id":"15881092364","text":"from tornado.ioloop import IOLoop, PeriodicCallback\nfrom tornado import gen\nfrom tornado.websocket import websocket_connect\n\nimport numpy as np\nimport h5py\n\nimport json\nimport random\nimport string\n\nimport sys\n\nimport random\n\nclass Client(object):\n def __init__(self, url, timeout):\n self.f = h5py.File('pollen_scraped.h5', 'a')\n self.day = np.zeros((1, 24), dtype=np.int32)\n self.dataset = np.zeros((0, 24), dtype=np.int32)\n\n self.stations = ['albacete', 'alcazar', 'alicante', 'almeria', 'avila', 'badajoz',\n 'barcelona', 'barcelona-uab', 'bejar', 'bilbao', 'burgos', 'burjassot', 'caceres',\n 'cadiz', 'cartagena', 'castellon-de-la-plana', 'ciudad-real', 'cordoba', 'coruña',\n 'cuenca', 'elche', 'gerona', 'granada', 'gijon', 'guadalajara', 'huelva', 'huesca',\n 'jaen-hospital', 'jaen', 'jativa', 'las-palmas', 'leon', 'lerida', 'logroño',\n 'madrid-subiza', 'madrid-hospital', 'malaga', 'murcia', 'oviedo', 'palencia',\n 'palma-mallorca', 'pamplona', 'ponferrada', 'pontevedra', 'salamanca', 'san-sebastian',\n 'santa-cruz-tenerife', 'santander', 'santiago-compostela', 'segovia', 'sevilla-macarena',\n 'sevilla-tomillar', 'soria', 'talavera', 'tarragona', 'teruel', 'toledo', 'torrelavega',\n 'tudela', 'valencia', 'valladolid', 'vitoria', 'zamora', 'zaragoza']\n\n\n self.ids = [1, 2, 3, 4, 5, 7, 6, 50, 8, 9, 11, 10, 12, 58, 16, 13, 15, 62, 14, 17, 18,\n 19, 65, 20, 21, 60, 59, 23, 22, 24, 28, 25, 26, 27, 29, 57, 30, 31, 32, 33, 35, 34, 36,\n 37, 38, 43, 45, 39, 64, 40, 41, 47, 42, 61, 44, 46, 48, 49, 63, 51, 54, 53, 55, 56]\n\n self.s = 5\n self.counter = 0\n\n self.url = url\n self.timeout = timeout\n self.ioloop = IOLoop.instance()\n self.ws = None\n self.connect()\n self.ioloop.start()\n\n @gen.coroutine\n def connect(self):\n print(\"trying to connect\")\n try:\n self.ws = yield websocket_connect(self.url)\n except Exception(e):\n print(\"connection error\")\n else:\n print(\"connected\")\n self.run()\n\n @gen.coroutine\n def run(self):\n while True:\n payload = yield self.ws.read_message()\n if payload is None:\n print(\"connection closed\")\n self.ws = None\n break\n\n if payload[0] == \"o\":\n msg = json.dumps(['{\"msg\":\"connect\", \"version\":\"1\", \"support\":[\"1\",\"pre2\",\"pre1\"]}'])\n self.ws.write_message(msg.encode('utf8'))\n elif payload[0] == 'a':\n frame = json.loads(payload[1:])\n body = json.loads(frame[0])\n\n if 'msg' in body:\n if body['msg'] == 'connected':\n self.day = np.zeros((1, 24), dtype=np.int32)\n self.dataset = np.zeros((0, 24), dtype=np.int32)\n self.counter = 0\n\n print('scraping from', self.stations[self.s])\n\n stringmsg = '{{\"msg\":\"sub\",\"id\":\"die7LBp7mEbzW3ffQ\",\"name\":\"polenes2015\",\"features\":[{{\"selector\":{{\"$and\":[{{\"fecha\":{{\"$gte\":19700101,\"$lte\":20201231}}}},{{\"idEstacion\":{}}}]}},\"options\":{{\"sort\":{{\"fecha\":1}}}},\"jump\":1}}]}}'.format(self.ids[self.s]);\n #print(stringmsg)\n msg = json.dumps([stringmsg])\n self.ws.write_message(msg.encode('utf8'));\n elif body['msg'] == 'ping':\n print('pong')\n msg = json.dumps(['{\"msg\": \"pong\"}'])\n self.ws.write_message(msg.encode('utf8'));\n elif body['msg'] == 'ready':\n print(self.dataset.shape)\n\n if self.dataset.shape[0] > 0:\n print(self.dataset[0, 0], '-', self.dataset[-1, 0])\n\n if self.stations[self.s] in self.f:\n del self.f[self.stations[self.s]]\n self.f.create_dataset(self.stations[self.s], data=self.dataset)\n print('------------')\n\n self.s += 1\n if self.s < len(self.stations):\n self.day = np.zeros((1, 24), dtype=np.int32)\n self.dataset = np.zeros((0, 24), dtype=np.int32)\n self.counter = 0\n\n stringmsg = \"{\\\"msg\\\":\\\"unsub\\\",\\\"id\\\":\\\"die7LBp7mEbzW3ffQ\\\"}\"\n msg = json.dumps([stringmsg])\n self.ws.write_message(msg.encode('utf8'));\n\n print('scraping from', self.stations[self.s])\n\n stringmsg = '{{\"msg\":\"sub\",\"id\":\"die7LBp7mEbzW3ffQ\",\"name\":\"polenes2015\",\"features\":[{{\"selector\":{{\"$and\":[{{\"fecha\":{{\"$gte\":19700101,\"$lte\":20201231}}}},{{\"idEstacion\":{}}}]}},\"options\":{{\"sort\":{{\"fecha\":1}}}},\"jump\":1}}]}}'.format(self.ids[self.s]);\n #print(stringmsg)\n msg = json.dumps([stringmsg])\n self.ws.write_message(msg.encode('utf8'));\n\n else:\n sys.exit()\n elif body['msg'] == 'added':\n self.day[0][0] = body['fields']['fecha'];\n\n self.counter += 1\n if self.counter % 365 == 0:\n print(self.counter//365, 'years scraped')\n\n for i in range(1, 23):\n if str(i) in body['fields'] and isinstance(body['fields'][str(i)], (int, float)):\n self.day[0][i] = body['fields'][str(i)]\n else:\n if str(i) in body['fields']:\n print('type error:', body['fields'][str(i)])\n self.day[0][i] = 0;\n self.dataset = np.append(self.dataset, self.day, axis = 0);\n\n\nif __name__ == \"__main__\":\n client = Client(\"wss://www.polenes.com/sockjs/490/j6ieo28e/websocket\", 5)\n","repo_name":"bolito2/pollen-forecast","sub_path":"tornado_scraper.py","file_name":"tornado_scraper.py","file_ext":"py","file_size_in_byte":6256,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"27872611963","text":"import os, sys\nimport numpy as np\nimport argparse\nfrom time import time\nfrom meta_model import *\nfrom utils import *\nimport pdb\nfrom pydoc import locate\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--settings_path', type=str)\nFLAGS = parser.parse_args()\n\n\n# main function\ndef main():\n settings = get_settings(settings_path=FLAGS.settings_path)\n model = IDK(settings=settings)\n\n # if the model has already been trained, load it.\n # else, train it\n new_run = 1\n if model.retrain:\n print('Begin training!')\n model.train()\n model.loadModel() # need to do this because last validation run needs to be cleared\n else:\n try:\n model.loadModel()\n print('Skip training!')\n new_run = 0\n except:\n print('Begin training!')\n model.train()\n model.loadModel() # need to do this because last validation run needs to be cleared\n\n # test the saved model\n if new_run or model.retest:\n print('Begin testing!')\n model.test()\n\n\ndef get_settings(settings_path):\n # read in settings\n settings = file_to_dict(settings_path)\n saving_path = os.path.dirname(settings_path)\n\n if \"experimentType\" not in settings:\n settings[\"experimentType\"] = 'pathak'\n\n # add settings\n settings[\"saving_path\"] = saving_path\n settings[\"train_data_path\"] = settings[\"data_pathname\"]\n settings[\"test_data_path\"] = settings[\"data_pathname\"]\n settings[\"delta_t\"] = settings[\"dt\"]\n ODE = locate('odelibrary.{}'.format(settings[\"f0_name\"]))\n if settings['f0_name']=='L96M':\n if settings[\"usef0\"]:\n physics = ODE(slow_only=settings['slowOnly'])\n settings[\"f0\"] = lambda t, y: physics.rhs(y, t)\n if settings[\"diff\"]==\"TrueDeriv\":\n raise ValueError(\"True Derivatives not yet setup for L96MS case\")\n else:\n if settings[\"usef0\"]:\n if settings[\"experimentType\"]=='pathak':\n physics1 = ODE()\n eps = settings['f0eps']\n physics1.b = physics1.b*(1+eps)\n elif settings[\"experimentType\"]=='GPerror':\n physics1 = ODE(random_closure=True, epsGP=settings['f0eps'])\n elif settings[\"experimentType\"]=='wrongModel':\n physics1 = ODE()\n settings[\"f0\"] = lambda t, y: physics1.rhs(y, t)\n\n if settings[\"diff\"]==\"TrueDeriv\":\n physics2 = ODE()\n settings[\"fTRUE\"] = lambda t, y: physics2.rhs(y, t)\n return settings\n\nif __name__ == '__main__':\n start_time = time()\n main()\n total_time = time() - start_time\n print(\"Total run time is {:2.2f} minutes\".format(total_time/60))\n","repo_name":"mattlevine22/ode_model_error","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2707,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"30146011534","text":"from django.contrib.auth import models, backends\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.conf import settings\n#from user import models as models_user\nfrom django.db.models.query_utils import Q\n\n\nclass FacebookBackend(backends.ModelBackend):\n def authenticate(self, facebook_id=None, facebook_email=None):\n '''\n Authenticate the facebook user by id OR facebook_email\n '''\n filter_clause = False\n if facebook_id:\n filter_clause = Q(facebook_id=facebook_id)\n\n if facebook_email:\n email_filter = Q(user__email=facebook_email)\n if filter_clause:\n filter_clause |= email_filter\n else:\n filter_clause = email_filter\n\n if filter_clause:\n try:\n profile_string = settings.AUTH_PROFILE_MODULE\n except AttributeError:\n profile_string = None\n if profile_string:\n profile_model = profile_string.split('.')[-1]\n profile_class = ContentType.objects.get(model=profile_model.lower())\n profile = profile_class.get_object_for_this_type(filter_clause)\n #profiles = profile_class.objects.filter(filter_clause).order_by('user')[:1]\n if profile:\n user = profile.user\n return user\n else:\n raise KeyError\n\n\n\n","repo_name":"MyaThandarKyaw/Django-facebook","sub_path":"django_facebook/auth_backends.py","file_name":"auth_backends.py","file_ext":"py","file_size_in_byte":1441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"23218478869","text":"import argparse\nimport os\nimport torch\nimport torch.nn as nn\nimport pandas as pd\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data.sampler import Sampler\nimport matplotlib.pyplot as plt\nfrom predict import *\nfrom tqdm import tqdm\nimport numpy as np\n\nfrom dataset import ECGDataset\nfrom resnet import resnet34\nfrom utils import cal_f1s, cal_aucs, split_data, accuracy_score\n\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('--data-dir', type=str, default=r'D:\\ECG_omri_maor\\Data\\WFDB', help='Directory for data dir')\n parser.add_argument('--leads', type=str, default='all', help='ECG leads to use')\n parser.add_argument('--seed', type=int, default=42, help='Seed to split data')\n parser.add_argument('--num-classes', type=int, default=int, help='Num of diagnostic classes')\n parser.add_argument('--lr', '--learning-rate', type=float, default=0.0001, help='Learning rate')\n parser.add_argument('--batch-size', type=int, default=32, help='Batch size')\n parser.add_argument('--num-workers', type=int, default=4, help='Num of workers to load data')\n parser.add_argument('--phase', type=str, default='train', help='Phase: train or test')\n parser.add_argument('--epochs', type=int, default=28, help='Training epochs')\n parser.add_argument('--resume', default=False, action='store_true', help='Resume')\n parser.add_argument('--use-gpu', default=True, action='store_true', help='Use GPU')\n parser.add_argument('--model-path', type=str, default='', help='Path to saved model')\n return parser.parse_args()\n\n\ndef train(dataloader, net, args, criterion, epoch, scheduler, optimizer, device):\n print('Training epoch %d:' % epoch)\n net.train()\n running_loss = 0\n output_list, labels_list = [], []\n for _, (data, labels) in enumerate(tqdm(dataloader)):\n data, labels = data.to(device), labels.to(device)\n output = net(data)\n loss = criterion(output, labels)\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n running_loss += loss.item()\n output_list.append(output.data.cpu().numpy())\n labels_list.append(labels.data.cpu().numpy())\n # scheduler.step()\n print(f'Avg Loss per epoch:{running_loss / (len(tqdm(dataloader)))}')\n y_trues = np.vstack(labels_list)\n y_scores = np.vstack(output_list)\n f1s = cal_f1s(y_trues, y_scores)\n avg_f1 = np.mean(f1s)\n # y_train_accuracy = y_scores > 0.5 # Ask Ariel if OK\n # accuracies = accuracy_score(y_trues, y_train_accuracy)\n # avg_accuracy = np.mean(accuracies)\n output_loss = running_loss / len(output_list)\n return output_loss, avg_f1 # , avg_accuracy\n\n\ndef evaluate(dataloader, net, args, criterion, device):\n print('Validating...')\n net.eval()\n running_loss = 0\n output_list, labels_list = [], []\n for _, (data, labels) in enumerate(tqdm(dataloader)):\n data, labels = data.to(device), labels.to(device)\n output = net(data)\n loss = criterion(output, labels)\n running_loss += loss.item()\n output = torch.sigmoid(output)\n output_list.append(output.data.cpu().numpy())\n labels_list.append(labels.data.cpu().numpy())\n print(f'Avg Loss per epoch:{running_loss / (len(output_list))}')\n y_trues = np.vstack(labels_list)\n y_scores = np.vstack(output_list)\n f1s = cal_f1s(y_trues, y_scores)\n avg_f1 = np.mean(f1s)\n output_loss = running_loss / len(output_list)\n print('F1s:', f1s)\n print('Avg F1: %.4f' % avg_f1)\n if args.phase == 'train' and avg_f1 > args.best_metric:\n args.best_metric = avg_f1\n torch.save(net.state_dict(), args.model_path)\n else:\n aucs = cal_aucs(y_trues, y_scores)\n avg_auc = np.mean(aucs)\n print('AUCs:', aucs)\n print('Avg AUC: %.4f' % avg_auc)\n return output_loss, avg_f1\n\n\ndef calc_weights(df, dataloader, lambda_param=0.01):\n df = df.labels.iloc[:, 1:8]\n label_counts = df.sum()\n label_weights = np.array(1 / label_counts)\n sample_weights = [0] * len(tqdm(dataloader))\n for idx, (data, label) in enumerate(tqdm(dataloader)):\n sample_weights[idx] = np.dot(label_weights, label)\n # regularization_term = lambda_param * sum(sample_weights ** 2)\n # sample_weights = [1/label_counts[i] for i in range(len(df.columns))]\n return sample_weights\n\n\ndef plot_metrics(all_metrics_dict):\n epoch_vec = np.arange(args.epochs)\n\n # F1\n fig, ax = plt.subplots(nrows=1, ncols=1) # create figure & 1 axis\n ax.plot(epoch_vec, all_metrics_dict[\"train_f1\"], epoch_vec, all_metrics_dict[\"val_f1\"])\n ax.set_xlabel(\"epoch\")\n ax.set_ylabel(\"F1 score\")\n ax.set_title('F1 as a function of epochs')\n ax.legend(['train', 'val'])\n fig.savefig(\"plots/f1.png\")\n\n # Loss\n fig, ax = plt.subplots(nrows=1, ncols=1) # create figure & 1 axis\n ax.plot(epoch_vec, all_metrics_dict[\"train_loss\"], epoch_vec, all_metrics_dict[\"val_loss\"])\n ax.set_xlabel(\"epoch\")\n ax.set_ylabel(\"Loss\")\n ax.legend(['train', 'val'])\n ax.set_title('Loss as a function of epochs')\n fig.savefig(\"plots/loss.png\")\n\n\nif __name__ == \"__main__\":\n args = parse_args()\n args.best_metric = 0\n data_dir = os.path.normpath(args.data_dir)\n database = os.path.basename(data_dir)\n\n if not args.model_path:\n args.model_path = f'models/resnet34_{database}_{args.leads}_{args.seed}.pth'\n\n if args.use_gpu and torch.cuda.is_available():\n device = torch.device('cuda:0')\n else:\n device = 'cpu'\n\n if args.leads == 'all':\n leads = 'all'\n nleads = 12\n else:\n leads = args.leads.split(',')\n nleads = len(leads)\n label_csv = os.path.join(data_dir, 'labels.csv')\n train_folds, val_folds, test_folds = split_data(seed=args.seed)\n folds_df = pd.DataFrame({'train': [train_folds, val_folds, test_folds]})\n folds_df.to_csv('folds_val.csv')\n train_dataset = ECGDataset('train', data_dir, label_csv, train_folds, leads)\n # train_weights = calc_weights(train_dataset, train_dataset)\n # train_sampler = torch.utils.data.sampler.WeightedRandomSampler(train_weights, len(train_weights), replacement=True)\n train_loader = DataLoader(train_dataset, batch_size=args.batch_size, num_workers=args.num_workers,\n pin_memory=True, shuffle=True) # , sampler=train_sampler)\n val_dataset = ECGDataset('val', data_dir, label_csv, val_folds, leads)\n val_loader = DataLoader(val_dataset, batch_size=args.batch_size, shuffle=False, num_workers=args.num_workers,\n pin_memory=True)\n test_dataset = ECGDataset('test', data_dir, label_csv, test_folds, leads)\n test_loader = DataLoader(test_dataset, batch_size=args.batch_size, shuffle=False, num_workers=args.num_workers,\n pin_memory=True)\n net = resnet34(input_channels=nleads).to(device)\n optimizer = torch.optim.Adam(net.parameters(), lr=args.lr)\n scheduler = torch.optim.lr_scheduler.StepLR(optimizer, 10, gamma=0.1)\n\n criterion = nn.BCEWithLogitsLoss()\n # train_accuracy = []\n train_loss = []\n val_loss = []\n train_f1 = []\n val_f1 = []\n if args.phase == 'train':\n if args.resume:\n net.load_state_dict(torch.load(args.model_path, map_location=device))\n for epoch in range(args.epochs):\n curr_train_loss, curr_train_f1 = train(train_loader, net, args, criterion, epoch,\n scheduler, optimizer, device)\n curr_val_loss, curr_val_f1 = evaluate(val_loader, net, args, criterion, device)\n train_loss.append(curr_train_loss)\n val_loss.append(curr_val_loss)\n # train_accuracy.append(curr_train_accuracy)\n train_f1.append(curr_train_f1)\n val_f1.append(curr_val_f1)\n\n # Testing and documentation phase\n args.threshold_path = f'models/{database}-threshold.pkl'\n thresholds = get_thresholds(val_loader, net, device, threshold_path=args.threshold_path)\n print('Thresholds:', thresholds)\n\n print('Results on validation data:')\n apply_thresholds(val_loader, net, device, thresholds, 'val')\n\n print('Results on test data:')\n apply_thresholds(test_loader, net, device, thresholds, 'test')\n net.load_state_dict(torch.load(args.model_path, map_location=device))\n test_loss, test_f1 = evaluate(test_loader, net, args, criterion, device)\n all_metrics_dict = {'train_f1': train_f1,\n 'val_f1': val_f1, 'train_loss': train_loss, 'val_loss': val_loss}\n plot_metrics(all_metrics_dict)\n print(f'Test results: Loss:{test_loss}, F1:{test_f1}')\n","repo_name":"OmriDan/ecg-diagnosis","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":8660,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"40341157390","text":"# -*- coding: utf-8 -*-\n\"\"\"\nPDU generic module for commands in state PDU.\n\"\"\"\n\nimport six\nimport abc\nfrom moler.cmd.commandtextualgeneric import CommandTextualGeneric\n\n__author__ = 'Marcin Usielski'\n__copyright__ = 'Copyright (C) 2020, Nokia'\n__email__ = 'marcin.usielski@nokia.com'\n\n\n@six.add_metaclass(abc.ABCMeta)\nclass GenericPduAten(CommandTextualGeneric):\n def __init__(self, connection, prompt=None, newline_chars=None, runner=None):\n \"\"\"\n Base class for Aten PDU commands in all states.\n\n :param connection: connection to device.\n :param prompt: expected prompt sending by device after command execution. Maybe String or compiled re.\n :param runner: runner to run command.\n \"\"\"\n\n super(GenericPduAten, self).__init__(connection=connection, prompt=prompt, newline_chars=newline_chars,\n runner=runner)\n","repo_name":"nokia/moler","sub_path":"moler/cmd/pdu_aten/generic_pdu_aten.py","file_name":"generic_pdu_aten.py","file_ext":"py","file_size_in_byte":904,"program_lang":"python","lang":"en","doc_type":"code","stars":57,"dataset":"github-code","pt":"78"}
+{"seq_id":"14782098118","text":"# %%\nfrom aocd.models import Puzzle \npuzzle = Puzzle(year = 2016, day=21)\n# %%\nX = [x.split() for x in puzzle.input_data.split('\\n')]\n\ndef swap_pos(pwd, instr):\n x = int(instr[2])\n y = int(instr[5])\n tmp = pwd[x]\n pwd[x] = pwd[y]\n pwd[y] = tmp\n return pwd\n\ndef swap_let(pwd, instr):\n x = pwd.index(instr[2])\n y = pwd.index(instr[5])\n tmp = pwd[x]\n pwd[x] = pwd[y]\n pwd[y] = tmp\n return pwd\n\ndef swap(pwd, instr):\n return swap_pos(pwd, instr) if instr[1] == 'position' else swap_let(pwd, instr)\n\ndef rotate_lr(pwd, d, s):\n if d == 'right':\n s = len(pwd) - (s % len(pwd))\n return pwd[s%len(pwd):] + pwd[:s%len(pwd)]\n\ndef rotate(pwd, instr):\n if instr[1] == 'based':\n n = pwd.index(instr[6])\n n += (1 if n < 4 else 2)\n return rotate_lr(pwd, 'right', n)\n else:\n return rotate_lr(pwd, instr[1], int(instr[2]))\n\ndef reverse(pwd, instr):\n x = int(instr[2])\n y = int(instr[4])\n head = pwd[:x]\n mid = pwd[x:y+1]\n tail = pwd[y+1:]\n return head + mid[::-1] + tail\n\ndef move(pwd, instr):\n x = int(instr[2])\n y = int(instr[5])\n tmp = pwd.pop(x)\n pwd.insert(y, tmp)\n return pwd\n\nfmap = {'swap': swap, 'rotate': rotate, 'reverse': reverse, 'move': move}\n\n\n# %%\npwd = list('abcdefgh')\nfor instr in X:\n res = fmap[instr[0]](pwd, instr)\n if len(res) != len(pwd):\n print(\"error\")\n print(pwd)\n print(res)\n print(instr)\n pwd = res\nres = ''.join(pwd)\nres\n# %%\npuzzle.answer_a = res\n# %%\n\ndef r_rotate_lr(pwd, d, s):\n if d == 'left':\n s = len(pwd) - (s % len(pwd))\n return pwd[s%len(pwd):] + pwd[:s%len(pwd)]\n\ndef r_rotate(pwd, instr):\n if instr[1] == 'based':\n n = pwd.index(instr[6])\n n += (1 if n < 4 else 2)\n for i in range(1, len(pwd)+5):\n attempt = rotate_lr(pwd.copy(), 'left', i)\n if rotate(attempt, instr) == pwd:\n return attempt\n else:\n return r_rotate_lr(pwd, instr[1], int(instr[2]))\n\ndef r_move(pwd, instr):\n x = int(instr[5])\n y = int(instr[2])\n tmp = pwd.pop(x)\n pwd.insert(y, tmp)\n return pwd\n\nfmap = {'swap': swap, 'rotate': r_rotate, 'reverse': reverse, 'move': r_move}\n\npwd = list('fbgdceah')\nfor instr in reversed(X):\n pwd = fmap[instr[0]](pwd, instr)\nres = ''.join(pwd)\nres\n# %%\npuzzle.answer_b = res\n# %%\n","repo_name":"elibaldwin/aoc2020","sub_path":"practice/aoc2016day21.py","file_name":"aoc2016day21.py","file_ext":"py","file_size_in_byte":2204,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"423133729","text":"#!/usr/bin/python3\n# coding=utf-8\n\n\nimport threading\nimport time\n\ndef sing():\n for i in range(3):\n print(\"singing...%d\" % i)\n time.sleep(1)\n\ndef dance():\n for i in range(3):\n print(\"dancing...%d\" % i)\n time.sleep(1)\n\nif __name__ == '__main__':\n \n sing_thread = threading.Thread(target = sing)\n dance_thread = threading.Thread(target = dance)\n\n sing_thread.start()\n dance_thread.start()\n\n","repo_name":"zhangshengasdf/python-notes","sub_path":"Multitask/thread.py","file_name":"thread.py","file_ext":"py","file_size_in_byte":436,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"21885546152","text":"#Keep entering pizza toppings until active is False.\n\nprompt = \"\\nWhat sort of toppings would you like?\"\nprompt += \"\\nEnter 'quit' when you have added all of your toppings! \"\n\nactive = True\n\nwhile active == True:\n toppings = input(prompt)\n \n if toppings == 'quit':\n active = False\n else:\n print(toppings.title()+\"? Nice! I'll throw it on your pizza.\")\n\n","repo_name":"jaoist/pycc","sub_path":"Inputs_and_Whiles/7-6active.py","file_name":"7-6active.py","file_ext":"py","file_size_in_byte":379,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"70858383612","text":"import os\nimport discord\nfrom commands import *\nimport random\nfrom replit import db\nfrom keep_alive import keep_alive\nimport utils\nfrom random import seed\nfrom random import randint\nfrom vector2 import Vector2\nfrom game import Game\n\nseed(1)\n\nlast_game_state_id = 0\n\nclient = discord.Client()\ngames = {}\n\n@client.event\nasync def on_ready():\n print(\"We have logged in as {0.user}\".format(client))\n\n@client.event\nasync def on_message(message):\n msg = message.content\n if message.author == client.user:\n if msg.endswith('move?'):\n await controls(message, client)\n elif msg.startswith('Guess'):\n await guesses(message, client)\n else:\n\n if msg.startswith('!play'):\n global last_game_state_id\n last_game_state_id = message.id\n games[message.author] = Game(Vector2(0,0), Vector2(0,0), Vector2(0,0))\n games[message.author].new_level()\n await play(message, client, games[message.author].get_board())\n\n if msg.startswith('!help'):\n await help(message, client)\n\n if msg.startswith('!guess'):\n await guess(message, client)\n\n@client.event\nasync def on_reaction_add(reaction, user):\n if user == client.user:\n #print(\"Failed because user is bot\")\n return\n\n if reaction.message.author != client.user:\n #print(\"Failed because the message wasnt the message created by this bot\")\n return\n\n #if reaction.message.id == last_game_state_id:\n #print(\"Failed because the message reacted to wasnt the game state message\")\n #return\n\n if reaction.message.content.endswith('move?'):\n user_input = 0\n if reaction.emoji == utils.up_arrow_emoji:\n user_input = 0\n elif reaction.emoji == utils.down_arrow_emoji:\n user_input = 2\n elif reaction.emoji == utils.left_arrow_emoji:\n user_input = 3\n elif reaction.emoji == utils.right_arrow_emoji:\n user_input = 1\n\n users = await reaction.users().flatten()\n users = users[0]\n if games[user].update_state(user_input):\n await update_game(reaction.message.channel, user.name, games[user].get_board())\n else:\n await reaction.message.channel.send('Cannot move in that direction.')\n await update_game(reaction.message.channel, user.name, games[user].get_board())\n\n elif reaction.message.content.startswith('Guess'):\n guess = 0\n answer = randint(0, 10)\n print(answer)\n if reaction.emoji == utils.one_emoji:\n guess = 1\n elif reaction.emoji == utils.two_emoji:\n guess = 2\n elif reaction.emoji == utils.three_emoji:\n guess = 3\n elif reaction.emoji == utils.four_emoji:\n guess = 4\n elif reaction.emoji == utils.five_emoji:\n guess = 5\n\n if guess == answer:\n await reaction.message.channel.send(\"Ding ding ding correct it was {0}.\".format(answer))\n else:\n await reaction.message.channel.send(\"Incorrect, the correct answer was {0}.\".format(answer))\n\n\nkeep_alive()\n\nmy_secret = os.environ['TOKEN']\nclient.run(my_secret)","repo_name":"CedarBishop/GameBot","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2927,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"11460252085","text":"# 6.0001/6.00 Problem Set 5 - RSS Feed Filter\n# Name:\n# Collaborators:\n# Time:\n\nimport feedparser\nimport string\nimport time\nimport threading\nfrom project_util import translate_html\nfrom datetime import datetime, timezone\nimport pytz\n\n\n#-----------------------------------------------------------------------\n\n#======================\n# Code for retrieving and parsing\n# Google and Yahoo News feeds\n# Do not change this code\n#======================\n\ndef process(url):\n \"\"\"\n Fetches news items from the rss url and parses them.\n Returns a list of NewsStory-s.\n \"\"\"\n feed = feedparser.parse(url)\n entries = feed.entries\n ret = []\n\n for entry in entries:\n guid = entry.guid\n title = translate_html(entry.title)\n link = entry.link\n description = \"\"\n try:\n description = translate_html(entry.description)\n except:\n description = translate_html(entry.title_detail)\n pubdate = translate_html(entry.published)\n\n try:\n pubdate = datetime.strptime(pubdate, \"%a, %d %b %Y %H:%M:%S %Z\")\n pubdate.replace(tzinfo=pytz.timezone(\"GMT\"))\n # pubdate = pubdate.astimezone(pytz.timezone('EST'))\n # pubdate.replace(tzinfo=None)\n except ValueError:\n pubdate = datetime.strptime(pubdate, \"%a, %d %b %Y %H:%M:%S %z\")\n\n #newsStory = NewsStory(guid, title, description, link, pubdate)\n newsStory = NewsStory(guid, title, description, link, pubdate)\n ret.append(newsStory)\n return ret\n\n#======================\n# Data structure design\n#======================\n\n# Problem 1\n\nclass NewsStory(object):\n\n # deffining the data attribute for class NewsStory\n def __init__(self, guid, title, description, link, pubdate):\n self.guid = guid\n self.title = title\n self.description = description\n self.link = link\n self.pubdate = pubdate\n\n # creating different ways to manipulate the class NewsStory\n def get_guide(self):\n return self.guid\n\n def get_title(self):\n return self.title\n\n def get_description(self):\n return self.description\n\n def get_link(self):\n return self.link\n\n def get_pubdate(self):\n return self.pubdate\n#======================\n# Triggers\n#======================\n\nclass Trigger(object):\n def evaluate(self, story):\n \"\"\"\n Returns True if an alert should be generated\n for the given news item, or False otherwise.\n \"\"\"\n # DO NOT CHANGE THIS!\n raise NotImplementedError\n\n# PHRASE TRIGGERS\n\n# Problem 2\n# TODO: PhraseTrigger\nclass PhraseTrigger(Trigger):\n\n def __init__(self, phrase):\n self.phrase = phrase.upper()\n\n # returning true if the phrase exist in the string argument text, false otherwise\n def is_phrase_in(self, string_argument):\n\n # converting string_argument to upper for casse sensitive string_argument text. then splitting the word in the phrase with the space\n self.string_argument = string_argument.upper()\n phrase_split = self.phrase.split()\n word_list = \"\"\n\n # iterating through the string_argument text and if the char is withing the list of string.pronunciatio(!\"#$%&'()*+,-./:;<=>?@[\\]^_`{|}~),\n # replacing that with space and if not in the list, just adding the char in the word_list\n # for char in string_argument:\n for char in self.string_argument:\n if char in string.punctuation:\n word_list += \" \"\n else:\n word_list += char\n\n # splitting the word_list with the space and iterating over word_list to remove the '' if any in the word_list\n word_list = word_list.split()\n while \"\" in word_list:\n word_list.remove(\"\")\n\n # iterating over phrase_split and if words in the word_list and the phrase mathes, returning true else returning false\n for word in phrase_split:\n word_list_join = \" \".join(word_list)\n if word in word_list and self.phrase in word_list_join:\n result = True\n else:\n result = False\n\n return result\n\n\n# Problem 3\n# TODO: TitleTrigger\n# returning true if the title of the story contains a valid phrase, false otherwise\nclass TitleTrigger(PhraseTrigger):\n\n def __init__(self, phrase):\n PhraseTrigger.__init__(self, phrase)\n\n def evaluate(self, story):\n return self.is_phrase_in(story.get_title())\n\n\n# Problem 4\n# TODO: DescriptionTrigger\n# returning true if the description of the story contains a valid phrase, false otherwise\nclass DescriptionTrigger(PhraseTrigger):\n\n def __init__(self, phrase):\n PhraseTrigger.__init__(self, phrase)\n\n def evaluate(self, story):\n return self.is_phrase_in(story.get_description())\n\n\n\n# TIME TRIGGERS\n\n# Problem 5\n# TODO: TimeTrigger\n# Constructor:\n# Input: Time has to be in EST and in the format of \"%d %b %Y %H:%M:%S\".\n# Convert time from string to a datetime before saving it as an attribute.\nclass TimeTrigger(Trigger):\n\n def __init__(self, time):\n time = datetime.strptime(time, \"%d %b %Y %H:%M:%S\")\n time = time.replace(tzinfo=pytz.timezone(\"EST\"))\n self.time = time\n\n# Problem 6\n# TODO: BeforeTrigger and AfterTrigger\n# BeforeTrigger fires when a story is published strictly before the trigger’s time,\nclass BeforeTrigger(TimeTrigger):\n\n def evaluate(self, story):\n pub_date = story.get_pubdate().replace(tzinfo=pytz.timezone(\"EST\"))\n return self.time > pub_date\n\n# AfterTrigger fires when a story is published strictly after the trigger’s time\nclass AfterTrigger(TimeTrigger):\n\n def evaluate(self, story):\n pub_date = story.get_pubdate().replace(tzinfo=pytz.timezone(\"EST\"))\n return self.time < pub_date\n\n\n# COMPOSITE TRIGGERS\n\n# Problem 7\n# TODO: NotTrigger\nclass NotTrigger(Trigger):\n\n # trigger should take this other trigger as an argument to its constructor\n def __init__(self, trigger):\n self.trigger = trigger\n\n #This trigger should produce its output by inverting the output of another trigger.\n def evaluate(self, story):\n return not self.trigger.evaluate(story)\n\n# Problem 8\n# TODO: AndTrigger\nclass AndTrigger(Trigger):\n\n #This trigger should take two triggers as arguments to its constructor\n def __init__(self, trigger_1, trigger_2):\n self.trigger1 = trigger_1\n self.trigger2 = trigger_2\n\n #should fire if either one (or both) of its inputted triggers would fire on that item.\n def evaluate(self, story):\n return self.trigger1.evaluate(story) and self.trigger2.evaluate(story)\n\n# Problem 9\n# TODO: OrTrigger\nclass OrTrigger(Trigger):\n\n #This trigger should take two triggers as arguments to its constructor\n def __init__(self, trigger_1, trigger_2):\n self.trigger1 = trigger_1\n self.trigger2 = trigger_2\n\n #should fire if either one (or both) of its inputted triggers would fire on that item\n def evaluate(self, story):\n return self.trigger1.evaluate(story) or self.trigger2.evaluate(story)\n\n#======================\n# Filtering\n#======================\n\n# Problem 10\n#returning a list of only the stories for which a trigger fires.\ndef filter_stories(stories, triggerlist):\n \"\"\"\n Takes in a list of NewsStory instances.\n\n Returns: a list of only the stories for which a trigger in triggerlist fires.\n \"\"\"\n # TODO: Problem 10\n # This is a placeholder\n # (we're just returning all the stories, with no filtering)\n triggered_stories = []\n for story in stories:\n for trigger in triggerlist:\n if trigger.evaluate(story):\n if story not in triggered_stories:\n triggered_stories.append(story)\n return triggered_stories\n\n\n\n#======================\n# User-Specified Triggers\n#======================\n# Problem 11\ndef read_trigger_config(filename):\n \"\"\"\n filename: the name of a trigger configuration file\n\n Returns: a list of trigger objects specified by the trigger configuration\n file.\n \"\"\"\n # We give you the code to read in the file and eliminate blank lines and\n # comments. You don't need to know how it works for now!\n trigger_file = open(filename, 'r')\n lines = []\n for line in trigger_file:\n line = line.rstrip()\n if not (len(line) == 0 or line.startswith('//')):\n lines.append(line)\n\n # TODO: Problem 11\n # line is the list of lines that you need to parse and for which you need\n # to build triggers\n\n print(lines) # for now, print it so you see what it contains!\n\n\n\nSLEEPTIME = 120 #seconds -- how often we poll\n\n\nif __name__ == '__main__':\n # A sample trigger list - you might need to change the phrases to correspond\n # to what is currently in the news\n try:\n #t1 = TitleTrigger(\"election\")\n #t2 = DescriptionTrigger(\"Trump\")\n #t3 = DescriptionTrigger(\"Biden\")\n #t4 = AndTrigger(t2, t3)\n #triggerlist = [t1, t4]\n\n # Problem 11\n # TODO: After implementing read_trigger_config, uncomment this line\n #triggerlist = read_trigger_config('triggers.txt')s\n\n # HELPER CODE - you don't need to understand this!\n # Reads and writes Newsstories to stories.txt in specified format\n # Retrieves and filters the stories from the RSS feeds\n guidShown = []\n def get_cont(newstory):\n if newstory.get_guid() not in guidShown:\n guidShown.append(newstory.get_guid())\n\n while True:\n\n\n print(\"Polling . . .\", end=' ')\n # Get stories from BBC's Top Stories RSS news feed\n stories = process(\"http://feeds.bbci.co.uk/news/rss.xml\")\n\n # Get stories from Yahoo's Top Stories RSS news feed\n stories.extend(process(\"http://news.yahoo.com/rss/topstories\"))\n\n stories = filter_stories(stories, triggerlist)\n\n\n #@ISMAMA\n file = open('stories.txt', 'w')\n for s in stories:\n file.write(s.title.strip())\n file.write(\"\\n\")\n for i in range(len(s.title)):\n file.write(\"-\")\n file.write(\"\\n\")\n file.write(s.description.strip())\n file.write(\"\\n\")\n file.write(s.link.strip())\n file.write(\"\\n\")\n file.write(\"_\"*60)\n for s in range(2):\n file.write(\"\\n\")\n file.close()\n\n #Do not uncomment these lines\n #Dlist(map(get_cont, stories))\n #Dscrollbar.config(command=cont.yview)\n\n\n print(\"Sleeping...\")\n time.sleep(SLEEPTIME)\n\n except Exception as e:\n print(e)","repo_name":"thapaSujit/CS50","sub_path":"OOP/ps5.py","file_name":"ps5.py","file_ext":"py","file_size_in_byte":10780,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74724104252","text":"#infinite loop\r\n\r\n\r\n\r\n#i=0\r\n#while i<=10:\r\n # print(\"hello world\",i)\r\n # i=i+1\r\n #\r\n#while True:\r\n # print(\"deepanshu\")\r\n # # for loop\r\n# name=\"deepanshu\"\r\n# for i in range( 1, 10):\r\n# print(f\"your name is {name}\",i) \r\n\r\n# from playsound import playsound\r\n# # playsound('audio.mp3') \r\n\r\nnum=input(\"Enter the number\")\r\nnum1=int(num)\r\nfor i in range(1,11):\r\n print(num1, 'X' ,i, '=' , num1*i)","repo_name":"deepanshu136/Python-Tutorials","sub_path":"python tutorial/tutorial4/tutorial6/basic5.py","file_name":"basic5.py","file_ext":"py","file_size_in_byte":405,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"5270836403","text":"from PIL import ImageFile\nfrom copy import deepcopy\nimport collections\nfrom datasets import load_from_disk, set_caching_enabled\nfrom utils import data_utils, utils\nfrom utils.args_helper import (\n DataArguments,\n ModelArguments,\n TrainingArguments\n)\nfrom tqdm import tqdm\nfrom torchvision.transforms import (\n CenterCrop,\n ColorJitter,\n Compose,\n Normalize,\n RandomHorizontalFlip,\n RandomVerticalFlip,\n RandomResizedCrop,\n RandomRotation,\n Resize,\n ToTensor,\n)\nfrom trainer.detr_trainer import DetrTrainer\nfrom transformers import HfArgumentParser\nfrom transformers.trainer_utils import get_last_checkpoint, is_main_process\nfrom typing import Dict, Union, Any, Optional, List, Tuple\n\nimport datasets\nimport json\nimport logging\nimport numpy as np\nimport os\nimport pandas as pd\nimport sys\nimport torch\nimport torch.nn as nn\nimport transformers\n\nfrom simmc2.model.utils import ambiguous_candidates_evaluation as eval_utils\nfrom trainer.detr_trainer import DetrTrainer \nfrom tqdm import tqdm\n\nfrom torch.utils.data import DataLoader\n \nset_caching_enabled(True)\nlogger = logging.getLogger(__name__)\n\n#####\n# Main Functions\n#####\ndef run(model_args, data_args, training_args):\n training_args.output_dir=\"{}/{}_{}_lr{}_bs{}\".format(\n training_args.output_dir,\n model_args.model_name_or_path.replace(\"/\", \"_\"),\n training_args.lr_scheduler_type,\n training_args.learning_rate,\n training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps\n )\n os.makedirs(training_args.output_dir, exist_ok=True)\n cache_dir_path = \"{}/{}_{}_lr{}_bs{}\".format(\n data_args.cache_dir_name,\n model_args.model_name_or_path.replace(\"/\", \"_\").replace('.',''),\n training_args.lr_scheduler_type,\n training_args.learning_rate,\n training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps\n )\n os.makedirs(cache_dir_path, exist_ok=True)\n\n # Data loading\n eval_dset, meta_dset, gold_data = data_utils.load_image_text_eval_dataset(data_path=data_args.devtest_dataset_path)\n # eval_dset = eval_dset.train_test_split(0.05)['test']\n \n if (data_args.prediction_path is None or not os.path.exists(data_args.prediction_path)):\n eval_dset = eval_dset.map(\n data_utils.convert_dialogue_to_caption,\n num_proc=data_args.preprocessing_num_workers,\n desc=\"convert object attributes to caption\",\n load_from_cache_file=True,\n cache_file_name=os.path.join(cache_dir_path, \"ds_converted.arrow\"),\n fn_kwargs={\"num_utterances\": data_args.num_utterances},\n remove_columns=[\"dialogue\"]\n )\n \n # Preprocessing\n tokenizer = transformers.AutoTokenizer.from_pretrained(model_args.model_name_or_path)\n if data_args.additional_special_token_path is not None and os.path.isfile(data_args.additional_special_token_path):\n with open(data_args.additional_special_token_path, \"rb\") as handle:\n special_tokens_dict = json.load(handle)\n num_added_toks = tokenizer.add_special_tokens(special_tokens_dict)\n logger.info(f\"Added {num_added_toks} tokens\")\n logger.info(f\"All special tokens: {tokenizer.all_special_tokens}\")\n feature_extractor = transformers.AutoFeatureExtractor.from_pretrained(model_args.model_name_or_path)\n processor = transformers.CLIPProcessor(feature_extractor, tokenizer)\n\n eval_dset = eval_dset.map(\n data_utils.tokenize_captions,\n num_proc=data_args.preprocessing_num_workers,\n desc=\"tokenize captions\",\n fn_kwargs={\n \"tokenizer\": tokenizer,\n \"max_seq_length\": data_args.max_seq_length,\n }\n )\n \n normalize = Normalize(mean=feature_extractor.image_mean, std=feature_extractor.image_std)\n eval_transforms = Compose(\n [\n Resize(feature_extractor.size),\n # CenterCrop(feature_extractor.size),\n ToTensor(),\n # normalize,\n ]\n )\n\n def eval_image_preprocess(example_batch): \n images = [\n eval_transforms(\n image.convert(\"RGB\").crop((\n bbox[0], bbox[1], bbox[0]+max(5, bbox[3]), bbox[1]+max(5, bbox[2])\n ))\n )\n for image, bbox in zip(example_batch[\"image\"], example_batch[\"bbox\"])\n ]\n captions = [caption for caption in example_batch[\"caption\"]]\n example_batch[\"pixel_values\"] = feature_extractor(\n images=images, text=captions, return_tensors=\"pt\")[\"pixel_values\"]\n return example_batch\n\n eval_dset = eval_dset.with_transform(eval_image_preprocess)\n\n # Training and evaluation\n model = transformers.CLIPModel.from_pretrained(model_args.model_name_or_path)\n\n def collate_fn(examples):\n pixel_values = torch.stack([example[\"pixel_values\"] for example in examples])\n input_ids = torch.tensor([example[\"input_ids\"] for example in examples], dtype=torch.long)\n attention_mask = torch.tensor([example[\"attention_mask\"] for example in examples], dtype=torch.long)\n return {\n \"pixel_values\": pixel_values,\n \"input_ids\": input_ids,\n \"attention_mask\": attention_mask,\n \"return_loss\": True,\n }\n\n trainer = DetrTrainer(\n model=model,\n args=training_args,\n data_collator=collate_fn,\n train_dataset=None,\n eval_dataset=None,\n tokenizer=processor\n )\n\n # Evaluation\n # predictions = trainer.predict(eval_dset)\n\n dataloader = DataLoader(\n eval_dset, shuffle=False,\n batch_size=training_args.per_device_train_batch_size, \n num_workers=training_args.dataloader_num_workers,\n collate_fn=collate_fn\n )\n\n print('Performing inference on test data...')\n model = model.cuda()\n logits_batch = []\n for batch in tqdm(dataloader):\n batch[\"pixel_values\"] = batch[\"pixel_values\"].cuda()\n batch[\"input_ids\"] = batch[\"input_ids\"].cuda()\n batch[\"attention_mask\"] = batch[\"attention_mask\"].cuda()\n outputs = model(**batch)\n logits_batch.append(outputs.logits_per_image.diagonal().cpu().detach().numpy())\n logits = np.concatenate(logits_batch)\n\n data_args.prediction_path = f'{cache_dir_path}/prediction_logits.pt'\n torch.save(logits, open(data_args.prediction_path, 'wb'))\n else:\n logits = torch.load(open(data_args.prediction_path, 'rb'))\n\n # Compute Metrics\n def compute_metrics(logits):\n \"\"\"Aggregate predictions & compute evaluation metric per utterance\"\"\"\n\n print('Collecting metadata for predictions...')\n pred_dict = {'dialog_id': [], 'turn_id': [], 'object_id': [], 'logit': [], 'num_labels': []}\n for row, logit in tqdm(zip(meta_dset, logits)):\n pred_dict['dialog_id'].append(row['dialog_id'])\n pred_dict['turn_id'].append(row['turn_id'])\n pred_dict['object_id'].append(row['object_id'])\n pred_dict['num_labels'].append(len(row['labels']))\n pred_dict['logit'].append(logit)\n\n print('Aggregating predictions...')\n df = pd.DataFrame(pred_dict)\n agg_preds = df.groupby(['dialog_id','turn_id','num_labels']).agg({'object_id': list, 'logit': list})\n agg_preds = agg_preds.reset_index().to_dict(orient='records')\n\n print('Filtering per utterance predictions...')\n results = collections.defaultdict(list)\n for agg_pred in agg_preds:\n dialog_id, turn_id, num_labels = agg_pred['dialog_id'], agg_pred['turn_id'], agg_pred['num_labels']\n object_ids, logits = np.array(agg_pred['object_id']), np.array(agg_pred['logit'])\n\n # ALL\n # indexes = range(len(logits))\n\n # ORACLE\n indexes = np.argpartition(logits, -num_labels)[-num_labels:] if num_labels != 0 else []\n\n # Top-k\n # indexes = np.argpartition(logits, -min(len(logits), 15))[-min(len(logits), 15):]\n\n # THRESHOLD\n # indexes = np.where(logits > np.mean(logits))[0]\n # indexes = np.where(logits > np.min(logits))[0]\n # indexes = np.where(logits > np.median(logits))[0]\n # print(logits)\n # logits = torch.sigmoid(torch.from_numpy(logits))\n # indexes = np.where(logits >= 0.5)[0]\n acc_object_ids = object_ids[indexes].tolist()\n\n new_instance = {\n \"turn_id\": turn_id,\n \"disambiguation_candidates\": acc_object_ids\n }\n results[dialog_id].append(new_instance)\n\n # Restructure results JSON and save.\n print('Comparing predictions with ground truths...')\n results = [{\n \"dialog_id\": dialog_id,\n \"predictions\": predictions,\n } for dialog_id, predictions in results.items()]\n\n # print(\"results\", results[0])\n # print()\n # print(\"gold_data\", gold_data[\"dialogue_data\"][0])\n # print()\n if \"coref_candidates\" in data_args.devtest_dataset_path:\n metrics = eval_utils.evaluate_ambiguous_candidates(gold_data, results, is_actually_coref=True)\n else:\n metrics = eval_utils.evaluate_ambiguous_candidates(gold_data, results, is_actually_coref=False)\n\n print('== Eval Metrics ==')\n print('Recall: ', metrics[\"recall\"])\n print('Precision: ', metrics[\"precision\"])\n print('F1-Score: ', metrics[\"f1\"])\n\n return metrics\n\n print('Calculating evaluation metrics...')\n metrics = compute_metrics(logits)\n \n # Report Metrics\n trainer.log_metrics(\"test\", metrics)\n trainer.save_metrics(\"test\", metrics) \n\ndef main():\n ###\n # Parsing & Initialization\n ###\n # Parse argument\n parser = HfArgumentParser((ModelArguments, DataArguments, TrainingArguments))\n if len(sys.argv) == 2 and sys.argv[1].endswith(\".json\"):\n model_args, data_args, training_args = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1]))\n else:\n model_args, data_args, training_args = parser.parse_args_into_dataclasses()\n\n # Set random seed\n utils.init_env(training_args.seed)\n \n # Detect last checkpoint\n last_checkpoint = None\n if os.path.isdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir:\n last_checkpoint = get_last_checkpoint(training_args.output_dir)\n if last_checkpoint is None and len(os.listdir(training_args.output_dir)) > 0:\n raise ValueError(\n f\"Output directory ({training_args.output_dir}) already exists and is not empty.\"\n \"Use --overwrite_output_dir to overcome.\"\n )\n elif last_checkpoint is not None:\n logger.info(\n f\"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change \"\n \"the `--output_dir` or add `--overwrite_output_dir` to train from scratch.\"\n )\n\n ###\n # Prepare logger\n ###\n # Init logging\n os.makedirs(\"./log\", exist_ok=True)\n \n logging.basicConfig(\n format=\"%(asctime)s - %(levelname)s - %(name)s - %(message)s\",\n datefmt=\"%m/%d/%Y %H:%M:%S\",\n handlers=[logging.StreamHandler(sys.stdout), logging.FileHandler(\n \"./log/log_{}_{}_lr{}_bs{}\".format(\n model_args.model_name_or_path.replace(\"/\", \"_\"),\n training_args.lr_scheduler_type,\n training_args.learning_rate,\n training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps\n ), mode=\"w\")],\n )\n logger.setLevel(logging.INFO if is_main_process(training_args.local_rank) else logging.WARN)\n\n # Log on each process the small summary:\n logger.warning(\n f\"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}\"\n f\"distributed training: {bool(training_args.local_rank != -1)}, 16-bits training: {training_args.fp16}\"\n )\n # Set the verbosity to warn of the Transformers logger (on main process only):\n if is_main_process(training_args.local_rank):\n transformers.utils.logging.set_verbosity(transformers.logging.WARNING)\n logger.info(\"Training/evaluation parameters %s\", training_args)\n \n ###\n # RUN RUN RUN!!!\n ###\n run(model_args, data_args, training_args)\n \nif __name__ == '__main__':\n main()","repo_name":"holylovenia/multimodal-object-identification","sub_path":"text_image_alignment_prediction.py","file_name":"text_image_alignment_prediction.py","file_ext":"py","file_size_in_byte":12847,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"29474073874","text":"import numpy as np\nimport cv2\n\n\ndef binary_to_multi_object_mask(binary_masks):\n\n \"\"\"\n Encode multiple 2d binary masks into a single 2d multi-object segmentation mask\n\n Parameters\n ----------\n binary_masks: numpy.ndarray of shape (n_objects, height, width)\n 2d binary masks\n\n Returns\n -------\n multi_object_mask: numpy.ndarray of shape (height, width)\n 2d multi-object mask\n \"\"\"\n\n multi_object_mask = np.zeros((binary_masks.shape[1], binary_masks.shape[2]))\n for i, binary_mask in enumerate(binary_masks):\n non_zero_idx = binary_mask == 1\n multi_object_mask[non_zero_idx] = i + 1\n\n return multi_object_mask\n\n\ndef polygon_to_mask(polygon, shape):\n\n \"\"\"\n Create binary segmentation mask from polygon\n\n Parameters\n ----------\n polygon: list of shape (n_polygons, n_points, 2)\n List of polygons\n\n shape: tuple of shape (2)\n Height and width of the mask\n\n Returns\n -------\n mask: numpy.ndarray of shape (height, width)\n 2d segmentation mask\n \"\"\"\n\n mask = np.zeros(shape)\n # Convert list of points to tuple pairs of X and Y coordinates\n points = np.array(polygon).reshape(-1, 2)\n # Draw mask from the polygon\n cv2.fillPoly(mask, [points], 1, lineType=cv2.LINE_8, shift=0)\n mask = np.array(mask).astype(np.uint8)\n\n return mask\n\n\ndef mask_to_bounding_box(mask):\n\n \"\"\"\n Get bounding box from a binary segmentation mask\n\n Parameters\n ----------\n mask: numpy.ndarray of shape (height, width)\n 2d binary mask\n\n Returns\n -------\n bounding_box: list of shape (4)\n Bounding box\n \"\"\"\n\n non_zero_idx = np.where(mask == 1)\n bounding_box = [\n int(np.min(non_zero_idx[1])),\n int(np.min(non_zero_idx[0])),\n int(np.max(non_zero_idx[1])),\n int(np.max(non_zero_idx[0]))\n ]\n\n return bounding_box\n\n\ndef coco_to_voc_bounding_box(bounding_box):\n\n \"\"\"\n Convert bounding box annotation from VOC to COCO format\n\n Parameters\n ----------\n bounding_box: list of shape (4)\n Bounding box with x1, y1, width, height values\n\n Returns\n -------\n bounding_box: list of shape (4)\n Bounding box with x1, y1, x2, y2 values\n \"\"\"\n\n x1 = bounding_box[0]\n y1 = bounding_box[1]\n x2 = x1 + bounding_box[2]\n y2 = y1 + bounding_box[3]\n\n return x1, y1, x2, y2\n\n\ndef coco_to_yolo_bounding_box(bounding_box):\n\n \"\"\"\n Convert bounding box annotation from COCO to YOLO format\n\n Parameters\n ----------\n bounding_box: list of shape (4)\n Bounding box with x1, y1, width, height values\n\n Returns\n -------\n bounding_box: list of shape (4)\n Bounding box with x_center, y_center, width, height values\n \"\"\"\n\n x1 = bounding_box[0]\n y1 = bounding_box[1]\n width = bounding_box[2]\n height = bounding_box[3]\n x_center = x1 + (width // 2)\n y_center = y1 + (height // 2)\n\n return x_center, y_center, width, height\n","repo_name":"gunesevitan/rsna-2023-abdominal-trauma-detection","sub_path":"src/annotation_utilities.py","file_name":"annotation_utilities.py","file_ext":"py","file_size_in_byte":2975,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"17598197179","text":"import random\n\nfrom die import Die\nfrom player import Player\n\nMAX_PLAYERS = 4\nMIN_PLAYERS = 2\n\nscreens = {\n 'START': 1,\n 'TURNS': 2,\n 'BOARD': 3,\n 'PROPS': 4,\n 'CARDS': 5\n}\n\n\nclass Game:\n def __init__(self, id):\n self.game_id = id\n self.players = []\n self.player_turn = 0 # The index of the players array that holds the player whose turn it is\n self.dice_values = [1, 1]\n self.is_playing = False\n self.current_screen = screens.get('START')\n self.available_icons = [True, True, True, True]\n self.turn_rolls = []\n\n def get_id(self):\n return self.game_id\n\n def get_players(self):\n return self.players\n\n def get_num_players(self):\n return len(self.players)\n\n def roll(self):\n self.dice_values = [random.randint(0, 5), random.randint(0, 5)]\n\n def done_roll(self, last_roll):\n if self.current_screen == screens.get('TURNS'):\n self.turn_rolls.append(last_roll)\n if len(self.turn_rolls) == len(self.players):\n unset_players = self.players.copy()\n for j in range(0, len(unset_players)):\n largest = self.turn_rolls.index(max(self.turn_rolls))\n self.players[j] = unset_players[largest] # appends correct player to empty list\n self.turn_rolls[largest] = -1 # get rid of the largest element in list\n self.current_screen = screens.get('BOARD')\n self.players[self.player_turn].last_roll = last_roll\n\n def next_player(self):\n self.player_turn = (self.player_turn + 1) % self.get_num_players()\n if self.get_curr_player().bankrupt:\n self.next_player()\n\n def set_screen(self, screen_name):\n self.current_screen = screens.get(screen_name)\n\n def screen_is(self, screen_name):\n return self.current_screen == screens.get(screen_name)\n\n def get_curr_player(self):\n return self.players[self.player_turn]\n\n def add_player(self):\n icon_num = 0\n while not self.available_icons[icon_num]:\n icon_num = (icon_num + 1) % len(self.available_icons)\n self.players.append(Player(False, icon_num, 'Player ' + str(self.get_num_players() + 1), 1, []))\n self.available_icons[icon_num] = False\n\n def set_icon(self, player_num, icon_num):\n \"\"\"\n Changes the given player's icon\n :param player_num:\n :param icon_num:\n :return:\n \"\"\"\n player = self.players[player_num]\n\n # Make old icon available for other players\n self.available_icons[player.icon_num] = True\n player.icon_num = icon_num\n # Make new icon unavailable\n self.available_icons[icon_num] = False\n","repo_name":"eobrie17/ski-monopoly","sub_path":"game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":2729,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"26362443205","text":"# -*- coding: utf-8 -*-\n\nimport numpy as np\nimport pandas as pd\n\nlist_county_a = [['Ahrensfelde', '0'],\n ['Altlandsberg', '1'],\n ['Am Mellensee', '0'],\n ['Altdöbern', '0'],\n ['Alt Tucheband', '0'],\n ['Angermünde', '0'],\n ['Althüttendorf', '0']\n ['Alt Zauche-Wußwerk', '0']\n ]\n\n","repo_name":"hyacinth0213/spider_project","sub_path":"plot_spider/myPlotSpider/landinfo/brandenburg_county_list.py","file_name":"brandenburg_county_list.py","file_ext":"py","file_size_in_byte":409,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"25627244688","text":"import sys\nimport hashlib\nimport argparse\nimport os\n\nPATH_TO_CACHE = \"/tmp/nginx_cache\"\n\n\ndef get_request_uri(command_args):\n parser = argparse.ArgumentParser()\n parser.add_argument('--request_uri', help='request uri')\n args = parser.parse_args()\n request_uri = args.request_uri\n return request_uri\n\n\ndef format_uri(uri):\n uri_parts = filter(lambda uri_item: uri_item != 'purge', uri.split('/'))\n formatted_uri = '/'.join(uri_parts)\n return formatted_uri\n\n\ndef hash_uri(uri):\n hashed_uri = hashlib.md5(uri.encode('utf-8')).hexdigest()\n return hashed_uri\n\n\ndef get_all_cache_files():\n return os.listdir(PATH_TO_CACHE)\n\n\ndef get_cache_by_request_uri(uri):\n formatted_uri = format_uri(uri)\n hashed_uri = hash_uri(formatted_uri)\n all_files = get_all_cache_files()\n cache_file = next((file for file in all_files if file == hashed_uri), None)\n return cache_file\n\n\ndef purge_cache(cache_file):\n path = os.path.join(PATH_TO_CACHE, cache_file)\n if os.path.isfile(path):\n os.remove(path)\n else:\n print(\"Error: %s file not found\" % path)\n\n\ndef main(command_args):\n request_uri = get_request_uri(command_args)\n print('Request cache for purge %s' % request_uri)\n cache = get_cache_by_request_uri(request_uri)\n if cache:\n purge_cache(cache)\n print('Cache successful purged for uri %s' % request_uri)\n\n\n\nif __name__ == '__main__':\n print('Python %s on %s' % (sys.version, sys.platform))\n main(sys.argv)","repo_name":"T1dehunter/home-work","sub_path":"lesson-7-nginx-customization/nginx-cache/cache_cleaner.py","file_name":"cache_cleaner.py","file_ext":"py","file_size_in_byte":1490,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74376130812","text":"\"\"\"Module to init the application\"\"\"\nfrom flask import Flask, request, jsonify\nfrom flask_sqlalchemy import SQLAlchemy\nfrom extract.extraction import Extract\nfrom models.engine import db, ma\nfrom models.models import Data, DataSchema\nfrom dotenv import load_dotenv\nimport os\n\nload_dotenv()\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\napp.config['SQLALCHEMY_DATABASE_URI'] = os.getenv(\"DATABASE_URL\")\n\ndb.init_app(app)\nma.init_app(app)\nwith app.app_context():\n db.create_all()\n\n\n@app.route('/extract', methods=['POST'])\ndef add_extraction():\n \"\"\"Add extract information from path document to database\"\"\"\n pdf_path = request.args.get('doc_path')\n doc_data = Extract()\n try:\n image_path = doc_data.convert_PDF_to_image(pdf_path)\n except:\n return jsonify({\"error_msg\": \"Couldn't find path to the pdf file\"})\n image = doc_data.cv_image(image_path)\n text = doc_data.ocr_reading(image, image_path)\n data = doc_data.get_data_from_text(text)\n data['doc_path'] = pdf_path\n try:\n save_data = Data(\n vendor_name=data['Vendor name'],\n fiscal_number=data['Fiscal number'], contract=data['Contract number'],\n comments=data['Comments'],\n start_date=data['Start date'],\n end_date=data['End date'],\n doc_path=data['doc_path']\n )\n db.session.add(save_data)\n db.session.commit()\n result = [[True, save_data.id], data]\n except:\n return jsonify({\"error_msg\": \"Couldn't store data in database\"})\n return jsonify(result)\n\n\n@app.route('/db_data/', methods=['GET'])\ndef get_extraction_data():\n \"\"\"Query all data from database\"\"\"\n request_data = request.args.get('table_name')\n if request_data == Data.__tablename__:\n data = Data.query.order_by(Data.id.desc()).all()\n data_schema = DataSchema(many=True)\n result = data_schema.dump(data)\n return jsonify(result)\n else:\n return jsonify({\"error_msg\": \"Table name '{}\\' doesn't exist\".format(request_data)})\n\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n","repo_name":"el-dani-cortes/PRGX_technical_challenge","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2120,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"11526396562","text":"# Задать список из n чисел последовательности (1+1/n)**n и вывести на экран их сумму\n\namount = int(input('Введите количество элементов последовательности: '))\nsumma = 0\n\n# ----------- Вариант 1 -----------------\n# for i in range(1, amount): \n# summa = summa + ((1 + 1 / i) ** i)\n# print(summa)\n\n\n# ----------- Вариант 2 -----------------\nsummas = [1 + (1 / i) ** i for i in range(1, amount)]\nprint(sum(summas))\n","repo_name":"netonblpb/PythonTasks","sub_path":"Task16.py","file_name":"Task16.py","file_ext":"py","file_size_in_byte":537,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17163926267","text":"n = int(input())\ncard = list(map(int, input().split()))\ncard.sort()\nm= int(input())\nnum = list(map(int, input().split()))\n\ndef binary(target):\n start =0\n end = n-1\n while start <= end:\n mid = (start + end) //2\n if target == card[mid]:\n return 1\n elif target < card[mid]:\n end = mid-1\n else:\n start = mid+1\n return 0\nfor i in num:\n print(binary(i),end = \" \")\nprint()","repo_name":"twodf78/coding_test","sub_path":"SearchAlgorithm/binary/10815.py","file_name":"10815.py","file_ext":"py","file_size_in_byte":441,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"34580457770","text":"import numpy as np\nfrom PIL import Image\n\n\nPATH = \"data/5\"\nRESOLUTION = 75\n\nVALUE_MATRIX = np.zeros((RESOLUTION, RESOLUTION, RESOLUTION), dtype=int)\n\n\n\n\n# read matrix from file\n\nfile = open(PATH+\"/value_matrix\", \"r\")\ninput = file.read()\nslices = input.split(\"]], [[\")\nslices[0] = slices[0][3:]\nslices[-1] = slices[-1][:-3]\nfor x in range(RESOLUTION):\n print(x)\n rows = slices[x].split(\"], [\")\n for y in range(RESOLUTION):\n row = rows[y]\n row = row.split(\",\")\n for z in range(RESOLUTION):\n VALUE_MATRIX[x][y][z] = int(row[z])\n z += 1\n y += 1\n\n\n\n\n\n\nfor i in range(RESOLUTION):\n img = Image.fromarray(np.uint8(VALUE_MATRIX[i]), 'L')\n img.save(f\"{PATH}/vals/image{i}.png\")","repo_name":"cle4rly/extreme-image-segmentation_rpml","sub_path":"src/show_values.py","file_name":"show_values.py","file_ext":"py","file_size_in_byte":733,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20668851079","text":"import gc\r\nfrom multiprocessing import reduction\r\n#exercisey\r\nclass Node:\r\n def __init__(self, data):\r\n self.data = data\r\n self.next = None\r\n\r\n\r\nclass CircularLinkedList:\r\n def __init__(self):\r\n self.last = None\r\n\r\n def addToEmpty(self, data):\r\n\r\n if self.last != None:\r\n return self.last\r\n\r\n newNode = Node(data)\r\n\r\n self.last = newNode\r\n\r\n self.last.next = self.last\r\n return self.last\r\n\r\n def addFront(self, data):\r\n\r\n if self.last == None:\r\n return self.addToEmpty(data)\r\n\r\n newNode = Node(data)\r\n\r\n newNode.next = self.last.next\r\n\r\n self.last.next = newNode\r\n\r\n return self.last\r\n\r\n def addEnd(self, data):\r\n if self.last == None:\r\n return self.addToEmpty(data)\r\n\r\n newNode = Node(data)\r\n\r\n newNode.next = self.last.next\r\n self.last.next = newNode\r\n\r\n self.last = newNode\r\n\r\n return self.last\r\n\r\n def addAfter(self, data, item):\r\n\r\n if self.last == None:\r\n return None\r\n\r\n newNode = Node(data)\r\n p = self.last.next\r\n while p:\r\n\r\n if p.data == item:\r\n\r\n newNode.next = p.next\r\n p.next = newNode\r\n\r\n if p == self.last:\r\n self.last = newNode\r\n return self.last\r\n else:\r\n return self.last\r\n p = p.next\r\n if p == self.last.next:\r\n print(item, \"The given node is not present in the list\")\r\n break\r\n\r\n def deleteNode(self, last, key):\r\n\r\n if last == None:\r\n return\r\n\r\n if (last).data == key and (last).next == last:\r\n\r\n last = None\r\n\r\n temp = last\r\n d = None\r\n\r\n if (last).data == key:\r\n while temp.next != last:\r\n temp = temp.next\r\n\r\n temp.next = (last).next\r\n last = temp.next\r\n\r\n while temp.next != last and temp.next.data != key:\r\n temp = temp.next\r\n\r\n if temp.next.data == key:\r\n d = temp.next\r\n temp.next = d.next\r\n\r\n return last\r\n\r\n def traverse(self):\r\n if self.last == None:\r\n print(\"The list is empty\")\r\n return\r\n\r\n newNode = self.last.next\r\n while newNode:\r\n print(newNode.data, end=\" \")\r\n newNode = newNode.next\r\n if newNode == self.last.next:\r\n break\r\n\r\n\t\r\n\r\n#exercise function\r\n\r\n\r\ndef exercise(bmi,bmr):\r\n \r\n print(\"\\n\\n\\t\\t\\t\\t\\tJOGGING or CYCLING or SKIPPING\")\r\n print(\"\\nBased on BMI\")\r\n if bmi<18:\r\n print(\"under weight\")\r\n elif bmi>=18 and bmi<=24:\r\n print(\"Healthy\")\r\n else:\r\n print(\"Over weight\")\r\n print(\"\\nMaintanence Factor : \",1800)\r\n ex=input(\"\\nENTER 1 for joggin 2 for cycling 3 for skipping\")\r\n exer=abs((bmr-1700)+(bmr-2000))\r\n val=0\r\n if ex=='1':\r\n joggin=75\r\n val=exer//joggin\r\n print(\"Jog for atleast \",val,\"km today\")\r\n\r\n elif ex=='2':\r\n cycle=100\r\n val=exer//cycle\r\n print(\"Do cycling for atleast \",val,\"km today\")\r\n\r\n else:\r\n # per min\r\n skip=20\r\n val=exer//skip\r\n print(\"Skipping for atleast \",val,\"mins today\")\r\n \r\n return val\r\n#food\r\ndef food(bmi,bmr):\r\n # node creation\r\n\r\n class Node:\r\n\r\n def __init__(self, data,calorie):\r\n self.data = data\r\n self.calorie=calorie\r\n self.next = None\r\n self.prev = None\r\n\r\n\r\n class DoublyLinkedList:\r\n\r\n def __init__(self):\r\n self.head = None\r\n def insert_front(self, data,cal):\r\n new_node = Node(data,cal)\r\n new_node.next = self.head\r\n if self.head is not None:\r\n self.head.prev = new_node\r\n self.head = new_node\r\n def insert_after(self, prev_node, data,cal):\r\n if prev_node is None:\r\n print(\"previous node cannot be null\")\r\n return\r\n new_node = Node(data,cal)\r\n new_node.next = prev_node.next\r\n prev_node.next = new_node\r\n new_node.prev = prev_node\r\n if new_node.next:\r\n new_node.next.prev = new_node\r\n def insert_end(self, data,cal):\r\n new_node = Node(data,cal)\r\n if self.head is None:\r\n self.head = new_node\r\n return\r\n temp = self.head\r\n while temp.next:\r\n temp = temp.next\r\n temp.next = new_node\r\n new_node.prev = temp\r\n\r\n return\r\n def deleteNode(self, dele):\r\n if self.head is None or dele is None:\r\n return\r\n if self.head == dele:\r\n self.head = dele.next\r\n if dele.next is not None:\r\n dele.next.prev = dele.prev\r\n if dele.prev is not None:\r\n dele.prev.next = dele.next\r\n gc.collect()\r\n def display_list(self, node):\r\n\r\n while node:\r\n print(node.data,node.calorie)\r\n last = node\r\n node = node.next\r\n #search\r\n def search(self,head,data):\r\n temp=head\r\n while temp:\r\n if temp.data==data:\r\n return temp.calorie\r\n \r\n temp = temp.next\r\n if temp==None:\r\n print(\"The given data doesnt exist:\")\r\n return 0\r\n \r\n\r\n # initialize an empty node\r\n d_linked_list = DoublyLinkedList()\r\n m_linked_list = DoublyLinkedList()\r\n l_linked_list = DoublyLinkedList()\r\n\r\n d_linked_list.insert_end(\"Chaapati\",71)\r\n d_linked_list.insert_end(\"Dosai\",100)\r\n d_linked_list.insert_end(\"Idli\",39)\r\n d_linked_list.insert_end(\"bread\",70)\r\n d_linked_list.insert_end(\"dhal\",120)\r\n d_linked_list.insert_end(\"chicken\",250)\r\n d_linked_list.insert_end(\"mutton\",154)\r\n d_linked_list.insert_end(\"chenna\",719)\r\n d_linked_list.insert_end(\"soya\",300)\r\n d_linked_list.insert_end(\"egg\",78)\r\n\r\n\r\n #lunch\r\n l_linked_list.insert_end(\"chicken\",250)\r\n l_linked_list.insert_end(\"mutton\",154)\r\n l_linked_list.insert_end(\"fish\",200)\r\n l_linked_list.insert_end(\"meals\",150)\r\n l_linked_list.insert_end(\"veg_side_dish\",120)\r\n l_linked_list.insert_end(\"egg\",78)\r\n\r\n\r\n\r\n # morning\r\n m_linked_list.insert_end(\"chenna\",719)\r\n m_linked_list.insert_end(\"chease\",402)\r\n m_linked_list.insert_end(\"chease\",402)\r\n m_linked_list.insert_end(\"Chaapati\",71)\r\n m_linked_list.insert_end(\"Dosai\",100)\r\n m_linked_list.insert_end(\"Idli\",39)\r\n m_linked_list.insert_end(\"oats\",300)\r\n m_linked_list.insert_end(\"dhal\",120)\r\n\r\n\r\n\r\n print()\r\n food=input(\"\\t\\t\\t\\t\\t DIET PLAN\\n\\n\\t\\t\\t\\t\\tBreakfast(1)\\n\\t\\t\\t\\t\\t Lunch(2)\\n\\t\\t\\t\\t\\t dinner(3)\\nEnter options(1,2,3):\")\r\n \r\n\r\n print(\"Calories are described for 100grms\")\r\n\r\n print(\"ENTER CALORIE INTAKE:\",2000)\r\n q=0\r\n print()\r\n tt='y'\r\n items=[]\r\n ee='n'\r\n #2300d_linked_list.display_list(d_linked_list.head)\r\n print(\"ENTER THE FOODS U WANT IN AND THE COUNT for a day :\")\r\n while q<2000 and tt==\"y\" and ee=='n':\r\n \r\n if food=='1':\r\n print(\"Breakfast\")\r\n m_linked_list.display_list(m_linked_list.head) \r\n elif food==\"2\":\r\n print(\"Lunch\")\r\n l_linked_list.display_list(l_linked_list.head)\r\n else:\r\n print(\"Dinner\")\r\n d_linked_list.display_list(d_linked_list.head) \r\n v=input()\r\n qty=int(input(\"How many:=\"))\r\n \r\n if food==1:\r\n s=m_linked_list.search(m_linked_list.head,v)\r\n elif food==2:\r\n s=l_linked_list.search(l_linked_list.head,v)\r\n else:\r\n s=d_linked_list.search(d_linked_list.head,v)\r\n temp=q+(s*qty)\r\n if temp>bmr:\r\n t=input(\"Remove an food from list(Y/N): \")\r\n if t=='y':\r\n k=input()\r\n if food=='1':\r\n s=m_linked_list.search(k)\r\n elif food==\"2\":\r\n s=l_linked_list.search(k)\r\n else:\r\n s=d_linked_list.search(k)\r\n print()\r\n ty=input(\"How much you wana remove: \")\r\n q-=s*ty\r\n else:\r\n tt=input(\"if Insuffiecency in nutrient(Y/N): \")\r\n \r\n else:\r\n q+=s*qty\r\n items.append(v)\r\n print(\"calories=\",[q,qty])\r\n ee=input(\"Wanna stop (y/n):\")\r\n food=input(\"\\nbreakfast lunch dinner(1,2,3):\")\r\n print(\"TOTAL CALORIE OF FOOD:=\",q)\r\n print(\"food list:- \",items) \r\nclass TreeNode:\r\n def __init__(self, data,password,bmr):\r\n self.data = data\r\n self.password=password\r\n self.bmr=bmr\r\n self.parent = None\r\n self.left = None\r\n self.right = None\r\n\r\n\r\nclass SplayTree:\r\n def __init__(self):\r\n self.root = None\r\n\r\n def leftRotate(self, x):\r\n y = x.right\r\n x.right = y.left\r\n if y.left != None:\r\n y.left.parent = x\r\n\r\n y.parent = x.parent\r\n # x is root\r\n if x.parent == None:\r\n self.root = y\r\n # x is left child\r\n elif x == x.parent.left:\r\n x.parent.left = y\r\n # x is right child\r\n else:\r\n x.parent.right = y\r\n\r\n y.left = x\r\n x.parent = y\r\n\r\n def rightRotate(self, x):\r\n y = x.left\r\n x.left = y.right\r\n if y.right != None:\r\n y.right.parent = x\r\n\r\n y.parent = x.parent\r\n # x is root\r\n if x.parent == None:\r\n self.root = y\r\n # x is right child\r\n elif x == x.parent.right:\r\n x.parent.right = y\r\n # x is left child\r\n else:\r\n x.parent.left = y\r\n\r\n y.right = x\r\n x.parent = y\r\n\r\n def splay(self, n):\r\n # node is not root\r\n while n.parent != None:\r\n # node is child of root, one rotation\r\n if n.parent == self.root:\r\n if n == n.parent.left:\r\n self.rightRotate(n.parent)\r\n else:\r\n self.leftRotate(n.parent)\r\n\r\n else:\r\n p = n.parent\r\n g = p.parent # grandparent\r\n\r\n if n.parent.left == n and p.parent.left == p: # both are left children\r\n self.rightRotate(g)\r\n self.rightRotate(p)\r\n\r\n elif n.parent.right == n and p.parent.right == p: # both are right children\r\n self.leftRotate(g)\r\n self.leftRotate(p)\r\n\r\n elif n.parent.right == n and p.parent.left == p:\r\n self.leftRotate(p)\r\n self.rightRotate(g)\r\n\r\n elif n.parent.left == n and p.parent.right == p:\r\n self.rightRotate(p)\r\n self.leftRotate(g)\r\n\r\n def insert(self, n):\r\n y = None\r\n temp = self.root\r\n while temp != None:\r\n y = temp\r\n if n.data < temp.data:\r\n temp = temp.left\r\n else:\r\n temp = temp.right\r\n\r\n n.parent = y\r\n\r\n if y == None: # newly added node is root\r\n self.root = n\r\n elif n.data < y.data:\r\n y.left = n\r\n else:\r\n y.right = n\r\n\r\n self.splay(n)\r\n\r\n def bstSearch(self, n, x, y):\r\n if x == n.data and y==n.password:\r\n self.splay(n)\r\n print(\"Logged In\")\r\n return n. bmr\r\n elif x < n.data:\r\n return self.bstSearch(n.left, x)\r\n elif x > n.data:\r\n return self.bstSearch(n.right, x)\r\n else:\r\n return None\r\n\r\n def preOrder(self, n):\r\n if n != None:\r\n print(n.data,n.password,n.bmr)\r\n self.preOrder(n.left)\r\n self.preOrder(n.right)\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n tree = SplayTree()\r\n cll = CircularLinkedList()\r\n\r\n u=\"yes\"\r\n print(\"\\t\\t\\t\\tWELCOME TO SSN MINI FITNESS APP\")\r\n while(u==\"yes\"):\r\n new=input(\"IF NEW USER (Y/N): \")\r\n if new=='y':\r\n user_name=input(\"ENTER NAME: \")\r\n password=input(\"ENTER PASSWORD: \")\r\n recheck=input(\"ENTER PASSWORD AGAIN: \")\r\n while password!=recheck:\r\n print(\"WRONG PASSWORD\")\r\n password=input(\"ENTER PASSWORD: \")\r\n recheck=input(\"ENTER PASSWORD AGAIN: \")\r\n heigh=int(input(\"ENTER HEGHT(cm): \"))\r\n weight=int(input(\"ENTER WEIGHT(kg): \"))\r\n age=int(input(\"ENTER AGE: \"))\r\n bmi=weight//(heigh*heigh*0.0001)\r\n sex=input(\"ENTER SEX (MALE/FEMLAE): \")\r\n print(\"YOUR BMI IS \",bmi)\r\n pound_weight=int(weight*2.20462)\r\n\r\n if sex==\"female\":\r\n bmr=int(655+(9.563*weight )+(1.850*heigh)-(4.676*age))\r\n else:\r\n bmr=int(66.47+(13.75*weight) + (5.003*heigh)-(6.755*age))\r\n print(\"YOUR BMR: \",bmr)\r\n a = TreeNode(user_name,password,bmr)\r\n tree.insert(a)\r\n f=input(\"Diet plan(d) / Exercise(e)\")\r\n if f=='d':\r\n food(bmi,bmr)\r\n else:\r\n goma=exercise(bmi,bmr)\r\n if goma!=0:\r\n cll.addFront(goma)\r\n vk=input(\"Wanna trace your routine:(y/n):\")\r\n if vk=='y':\r\n cll.traverse()\r\n\r\n else:\r\n \r\n user_name=input(\"ENTER NAME: \")\r\n password=input(\"ENTER PASSWORD: \")\r\n bmr=tree.bstSearch(tree.root, user_name,password)\r\n if bmr==None: \r\n print(\"Wrong password or wrong user name\")\r\n user_name=input(\"ENTER NAME: \")\r\n password=input(\"ENTER PASSWORD: \")\r\n bmr=tree.bstSearch(tree.root, user_name,password)\r\n f=input(\"Diet plan(d) / Exercise(e)\")\r\n if f=='d':\r\n food(bmi,bmr)\r\n else:\r\n goma=exercise(bmi,bmr)\r\n if goma!=0:\r\n cll.addFront(goma)\r\n vk=input(\"Wanna trace your routine:(y/n):\")\r\n if vk=='y':\r\n cll.traverse()\r\n u=input(\"\\n\\n\\t\\t\\t\\tif you want to contniue (yes/no):\")\r\n\r\n","repo_name":"JDeepak45/ADS-Mini-project","sub_path":"exercise.py","file_name":"exercise.py","file_ext":"py","file_size_in_byte":14606,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"71850451453","text":"###########################################\n# Author : Bastien Girardet, Deborah De Wolff\n# Date : 13.05.2018\n# Course : Applications in Object-oriented Programming and Databases\n# Teachers : Binswanger Johannes, Zürcher Ruben\n# Project : Bibliotek\n# Name : book.py SQL Alchemy Model\n# #########################################\n\nfrom db import db\n\nclass BookModel(db.Model):\n \"\"\"SQLAlchemy Book Model\"\"\"\n\n # We assign the correct table\n __tablename__ = 'books'\n \n # Table columns\n bookId = db.Column(db.Integer, primary_key=True)\n title = db.Column(db.String(300))\n publisher = db.Column(db.String(150))\n published_date = db.Column(db.String(15))\n description = db.Column(db.Text())\n isbn10 = db.Column(db.String(20))\n isbn13 = db.Column(db.String(20))\n booktype = db.Column(db.String(30))\n language = db.Column(db.String(30))\n thumbnail = db.Column(db.String(200))\n page_count = db.Column(db.Integer)\n md5 = db.Column(db.String(32))\n url_info = db.Column(db.String(200))\n dl_link1 = db.Column(db.String(200))\n dl_link2 = db.Column(db.String(200))\n chosen_url = db.Column(db.String(200))\n filepath = db.Column(db.String(200))\n\n # Foreign key a book belongs to a category and an author\n categoryId = db.Column('categoryId', db.Integer, db.ForeignKey('categories.categoryId'))\n authorId = db.Column('authorId', db.Integer, db.ForeignKey('authors.authorId'))\n\n # We reference the parents\n category = db.relationship('CategoryModel',cascade=\"save-update\")\n author = db.relationship('AuthorModel',cascade=\"save-update\")\n \n # A book might many portfolio book relations\n portfolios_books = db.relationship(\"PortfolioBookModel\", cascade=\"save-update, merge, delete\")\n\n def __init__(self, title, authorId, categoryId,publisher,published_date,description,isbn10,isbn13,booktype,language,thumbnail,page_count,md5,url_info,dl_link1,dl_link2,chosen_url,filepath):\n \"\"\"[summary]\n \n Arguments:\n title {string} -- title of the book\n authorId {int} -- id of the author of the book\n categoryId {int} -- id of the category of the book\n publisher {string} -- name of the publisher of the book\n published_date {string} -- publishing date\n description {string} -- description of the book\n isbn10 {string} -- isbn10 of the book\n isbn13 {string} -- isbn13 of the book\n booktype {string} -- type of the book (extension)\n language {string} -- Language of the book\n thumbnail {string} -- Url of the book's thumbnail\n page_count {int} -- Page count of the book\n md5 {string} -- libgen.io identifier\n url_info {string} -- google book api book info url\n dl_link1 {string} -- download link 1\n dl_link2 {string} -- download link 2\n chosen_url {string} -- Chosen url for the download (Whether dl_link1 or dl_link2)\n filepath {string} -- file path where the book is digitally contained\n \"\"\"\n # Instance variables\n self.title = title\n self.authorId = authorId\n self.categoryId = categoryId\n self.publisher = publisher\n self.published_date = published_date\n self.description = description\n self.isbn10 = isbn10\n self.isbn13 = isbn13\n self.booktype = booktype\n self.language = language\n self.thumbnail = thumbnail\n self.page_count = page_count\n self.md5 = md5\n self.url_info = url_info\n self.dl_link1 = dl_link1\n self.dl_link2 = dl_link2\n self.filepath = filepath\n self.chosen_url = chosen_url\n\n def json(self):\n \"\"\"Return a JSON data of the instance variables\"\"\"\n\n return {\n 'bookId' : self.bookId,\n 'title' : self.title,\n 'authorId' : self.authorId,\n 'categoryId' : self.categoryId,\n 'publisher' : self.publisher,\n 'published_date' : self.published_date,\n 'description' : self.description,\n 'isbn10' : self.isbn10,\n 'isbn13' : self.isbn13,\n 'booktype' : self.booktype,\n 'language' : self.language,\n 'thumbnail' : self.thumbnail,\n 'page_count' : self.page_count,\n 'md5' : self.md5,\n 'url_info' : self.url_info,\n 'dl_link1' : self.dl_link1,\n 'dl_link2' : self.dl_link2,\n 'filepath' : self.filepath,\n 'chosen_url' : self.chosen_url}\n\n # Important methods used to retrieve data through SQL Alchemy\n @classmethod\n def find_by_title(cls, title):\n \"\"\"Retrieve the book provided its title\"\"\"\n \n return cls.query.filter_by(title=title).first()\n\n @classmethod\n def find_by_id(cls, bookId):\n \"\"\"Retrieve the book provided its bookId\"\"\"\n\n return cls.query.filter_by(bookId=bookId).first()\n\n @classmethod\n def find_by_author_id(cls, authorId):\n \"\"\"Retrieve the book provided its authorId\"\"\"\n\n return cls.query.filter_by(authorId=authorId).all()\n \n @classmethod\n def find_by_category_id(cls, categoryId):\n \"\"\"Retrieve the book provided its categoryId\"\"\"\n\n return cls.query.filter_by(categoryId=categoryId).all()\n\n def save_to_db(self):\n \"\"\"Methods used to push and commit to the database\"\"\"\n\n db.session.add(self)\n db.session.commit()\n\n def delete_from_db(self):\n \"\"\"Methods used to delete and commit to the database\"\"\" \n\n db.session.delete(self)\n db.session.commit()\n","repo_name":"basgir/bibliotek","sub_path":"models/book.py","file_name":"book.py","file_ext":"py","file_size_in_byte":5730,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"10876847576","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom RedditUtils import RedditUtils\nfrom subprocess import call\nimport datetime\nimport random\n\ndef daysSinceEpoch():\n epoch = datetime.datetime.utcfromtimestamp(0)\n today = datetime.datetime.today()\n d = today - epoch\n return str(d.days)\n\n\n\ndef runMain(utilsInstance, subReddit):\n call([\"./reset.sh\"], shell=True)\n\n submissionsList = utilsInstance.returnTop()\n #postNos = input(\"enter post numbers (separated by a space): \")\n numTopComments = 14 #int(input(\"how many top comments: \"))\n globalCount = 0\n \n top = [0,1,2,3,4]\n #for post in postNos.split(\" \"):\n for post in top:\n globalCount += 1\n title = submissionsList[int(post)].title\n postHtml = utilsInstance.generatePostTitle(submissionsList[int(post)])\n utilsInstance.genVideoClip(postHtml, submissionsList[int(post)].title, globalCount)\n\n topComments = utilsInstance.getTopComments(numTopComments, submissionsList[int(post)].comments.list()) \n # commentsList = submissionsList[int(post)].comments.list()\n for commentList in topComments:\n \n # first comment, create image & audioFile\n globalCount += 1\n html = utilsInstance.generateOneComment(commentList[0])\n utilsInstance.genVideoClip(html, commentList[0].body, globalCount)\n \n # create image and audiofile for second comment if one exists\n if(len(commentList) == 2):\n globalCount += 1\n html = utilsInstance.generateTwoComments(commentList[0], commentList[1])\n utilsInstance.genVideoClip(html, commentList[1].body, globalCount)\n fileName = subReddit + \"_\" + daysSinceEpoch() + \"_\" + str(random.randint(1,1000000))\n # combine the individual video files into one file and put it in /output\n call([\"./cleanCombine.sh \" + fileName], shell=True)\n # create a thumbnail\n try:\n utilsInstance.createThumbnail(title, fileName)\n youtubeUpload(title, submissionsList[int(post)].url, subReddit, fileName)\n except:\n pass\n\ndef youtubeUpload(videoTitle, videoDesc, board, fileName):\n #upload video to youtube\n videoParams = '--title \"' + videoTitle + '\" --client-secrets=\"./yt_creds_file.json\" --category=Entertainment --description=\"' + videoDesc + '\" --tags=\"reddit,' + board + ',trending\" --default-language=\"en\" --privacy=\"public\" --default-audio-language=\"en\" --embeddable=True --thumbnail ./output/' + fileName + '.jpg ./output/' + fileName + '_vid.mp4'\n print(\"calling...\")\n print(\"youtube-upload \")# + videoParams)\n call([\"youtube-upload \" + videoParams], shell=True)\n","repo_name":"forbesjon2/redditYT","sub_path":"Run.py","file_name":"Run.py","file_ext":"py","file_size_in_byte":2730,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20221165067","text":"import threading\n\ndef first_func():\n global x\n\n while(x < 300):\n x += 1\n\n print(threading.current_thread())\n if x == 300:\n print(x)\n \ndef second_func():\n global x\n \n x = 450\n while(x < 600):\n x += 1\n \n print(threading.current_thread())\n print(x)\n\ndef main():\n global x\n x = 0\n\n first_thread = threading.Thread(target = first_func, name = 'first thread function')\n first_thread.start()\n #first_thread.join()\n ''' mian thread will wait for the child threads to get completed before completing its execution.\n Its like, where ever threrad's join will be invoked, the control flow of execution wont execute its parent thread, and parent thread will wait for that child thread to get completed.\n Then only, the next command/instruction of parent thread will execute.'''\n\n second_thread = threading.Thread(target = second_func, name = 'second thread function')\n second_thread.start()\n\n print(threading.enumerate())\n print(threading.current_thread())\n\n\nif (__name__ == '__main__'):\n main()","repo_name":"SharanJaiswal/py","sub_path":"multithreading_multiprocessing/02_mt.py","file_name":"02_mt.py","file_ext":"py","file_size_in_byte":1078,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33660162967","text":"from m5.SimObject import *\nfrom m5.params import *\nfrom m5.proxy import *\n\nfrom m5.objects.ClockedObject import ClockedObject\nfrom m5.objects.IndexingPolicies import *\nfrom m5.objects.ReplacementPolicies import *\n\n\nclass HWPProbeEvent(object):\n def __init__(self, prefetcher, obj, *listOfNames):\n self.obj = obj\n self.prefetcher = prefetcher\n self.names = listOfNames\n\n def register(self):\n if self.obj:\n for name in self.names:\n self.prefetcher.getCCObject().addEventProbe(\n self.obj.getCCObject(), name\n )\n\n\nclass BasePrefetcher(ClockedObject):\n type = \"BasePrefetcher\"\n abstract = True\n cxx_class = \"gem5::prefetch::Base\"\n cxx_header = \"mem/cache/prefetch/base.hh\"\n cxx_exports = [PyBindMethod(\"addEventProbe\"), PyBindMethod(\"addMMU\")]\n sys = Param.System(Parent.any, \"System this prefetcher belongs to\")\n\n # Get the block size from the parent (system)\n block_size = Param.Int(Parent.cache_line_size, \"Block size in bytes\")\n\n on_miss = Param.Bool(False, \"Only notify prefetcher on misses\")\n on_read = Param.Bool(True, \"Notify prefetcher on reads\")\n on_write = Param.Bool(True, \"Notify prefetcher on writes\")\n on_data = Param.Bool(True, \"Notify prefetcher on data accesses\")\n on_inst = Param.Bool(True, \"Notify prefetcher on instruction accesses\")\n prefetch_on_access = Param.Bool(\n Parent.prefetch_on_access,\n \"Notify the hardware prefetcher on every access (not just misses)\",\n )\n prefetch_on_pf_hit = Param.Bool(\n Parent.prefetch_on_pf_hit,\n \"Notify the hardware prefetcher on hit on prefetched lines\",\n )\n use_virtual_addresses = Param.Bool(\n False, \"Use virtual addresses for prefetching\"\n )\n page_bytes = Param.MemorySize(\n \"4KiB\", \"Size of pages for virtual addresses\"\n )\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self._events = []\n self._mmus = []\n\n def addEvent(self, newObject):\n self._events.append(newObject)\n\n # Override the normal SimObject::regProbeListeners method and\n # register deferred event handlers.\n def regProbeListeners(self):\n for mmu in self._mmus:\n self.getCCObject().addMMU(mmu.getCCObject())\n for event in self._events:\n event.register()\n self.getCCObject().regProbeListeners()\n\n def listenFromProbe(self, simObj, *probeNames):\n if not isinstance(simObj, SimObject):\n raise TypeError(\"argument must be of SimObject type\")\n if len(probeNames) <= 0:\n raise TypeError(\"probeNames must have at least one element\")\n self.addEvent(HWPProbeEvent(self, simObj, *probeNames))\n\n def registerMMU(self, simObj):\n if not isinstance(simObj, SimObject):\n raise TypeError(\"argument must be a SimObject type\")\n self._mmus.append(simObj)\n\n\nclass MultiPrefetcher(BasePrefetcher):\n type = \"MultiPrefetcher\"\n cxx_class = \"gem5::prefetch::Multi\"\n cxx_header = \"mem/cache/prefetch/multi.hh\"\n\n prefetchers = VectorParam.BasePrefetcher([], \"Array of prefetchers\")\n\n\nclass QueuedPrefetcher(BasePrefetcher):\n type = \"QueuedPrefetcher\"\n abstract = True\n cxx_class = \"gem5::prefetch::Queued\"\n cxx_header = \"mem/cache/prefetch/queued.hh\"\n latency = Param.Int(1, \"Latency for generated prefetches\")\n queue_size = Param.Int(32, \"Maximum number of queued prefetches\")\n max_prefetch_requests_with_pending_translation = Param.Int(\n 32,\n \"Maximum number of queued prefetches that have a missing translation\",\n )\n queue_squash = Param.Bool(True, \"Squash queued prefetch on demand access\")\n queue_filter = Param.Bool(True, \"Don't queue redundant prefetches\")\n cache_snoop = Param.Bool(\n False, \"Snoop cache to eliminate redundant request\"\n )\n\n tag_prefetch = Param.Bool(\n True, \"Tag prefetch with PC of generating access\"\n )\n\n # The throttle_control_percentage controls how many of the candidate\n # addresses generated by the prefetcher will be finally turned into\n # prefetch requests\n # - If set to 100, all candidates can be discarded (one request\n # will always be allowed to be generated)\n # - Setting it to 0 will disable the throttle control, so requests are\n # created for all candidates\n # - If set to 60, 40% of candidates will generate a request, and the\n # remaining 60% will be generated depending on the current accuracy\n throttle_control_percentage = Param.Percent(\n 0,\n \"Percentage of requests \\\n that can be throttled depending on the accuracy of the prefetcher.\",\n )\n\n\nclass StridePrefetcherHashedSetAssociative(SetAssociative):\n type = \"StridePrefetcherHashedSetAssociative\"\n cxx_class = \"gem5::prefetch::StridePrefetcherHashedSetAssociative\"\n cxx_header = \"mem/cache/prefetch/stride.hh\"\n\n\nclass StridePrefetcher(QueuedPrefetcher):\n type = \"StridePrefetcher\"\n cxx_class = \"gem5::prefetch::Stride\"\n cxx_header = \"mem/cache/prefetch/stride.hh\"\n\n # Do not consult stride prefetcher on instruction accesses\n on_inst = False\n\n confidence_counter_bits = Param.Unsigned(\n 3, \"Number of bits of the confidence counter\"\n )\n initial_confidence = Param.Unsigned(\n 4, \"Starting confidence of new entries\"\n )\n confidence_threshold = Param.Percent(\n 50, \"Prefetch generation confidence threshold\"\n )\n\n use_requestor_id = Param.Bool(True, \"Use requestor id based history\")\n\n degree = Param.Int(4, \"Number of prefetches to generate\")\n\n table_assoc = Param.Int(4, \"Associativity of the PC table\")\n table_entries = Param.MemorySize(\"64\", \"Number of entries of the PC table\")\n table_indexing_policy = Param.BaseIndexingPolicy(\n StridePrefetcherHashedSetAssociative(\n entry_size=1, assoc=Parent.table_assoc, size=Parent.table_entries\n ),\n \"Indexing policy of the PC table\",\n )\n table_replacement_policy = Param.BaseReplacementPolicy(\n RandomRP(), \"Replacement policy of the PC table\"\n )\n\n\nclass TaggedPrefetcher(QueuedPrefetcher):\n type = \"TaggedPrefetcher\"\n cxx_class = \"gem5::prefetch::Tagged\"\n cxx_header = \"mem/cache/prefetch/tagged.hh\"\n\n degree = Param.Int(2, \"Number of prefetches to generate\")\n\n\nclass IndirectMemoryPrefetcher(QueuedPrefetcher):\n type = \"IndirectMemoryPrefetcher\"\n cxx_class = \"gem5::prefetch::IndirectMemory\"\n cxx_header = \"mem/cache/prefetch/indirect_memory.hh\"\n pt_table_entries = Param.MemorySize(\n \"16\", \"Number of entries of the Prefetch Table\"\n )\n pt_table_assoc = Param.Unsigned(16, \"Associativity of the Prefetch Table\")\n pt_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.pt_table_assoc,\n size=Parent.pt_table_entries,\n ),\n \"Indexing policy of the pattern table\",\n )\n pt_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the pattern table\"\n )\n max_prefetch_distance = Param.Unsigned(16, \"Maximum prefetch distance\")\n num_indirect_counter_bits = Param.Unsigned(\n 3, \"Number of bits of the indirect counter\"\n )\n ipd_table_entries = Param.MemorySize(\n \"4\", \"Number of entries of the Indirect Pattern Detector\"\n )\n ipd_table_assoc = Param.Unsigned(\n 4, \"Associativity of the Indirect Pattern Detector\"\n )\n ipd_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.ipd_table_assoc,\n size=Parent.ipd_table_entries,\n ),\n \"Indexing policy of the Indirect Pattern Detector\",\n )\n ipd_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the Indirect Pattern Detector\"\n )\n shift_values = VectorParam.Int([2, 3, 4, -3], \"Shift values to evaluate\")\n addr_array_len = Param.Unsigned(4, \"Number of misses tracked\")\n prefetch_threshold = Param.Unsigned(\n 2, \"Counter threshold to start the indirect prefetching\"\n )\n stream_counter_threshold = Param.Unsigned(\n 4, \"Counter threshold to enable the stream prefetcher\"\n )\n streaming_distance = Param.Unsigned(\n 4, \"Number of prefetches to generate when using the stream prefetcher\"\n )\n\n\nclass SignaturePathPrefetcher(QueuedPrefetcher):\n type = \"SignaturePathPrefetcher\"\n cxx_class = \"gem5::prefetch::SignaturePath\"\n cxx_header = \"mem/cache/prefetch/signature_path.hh\"\n\n signature_shift = Param.UInt8(\n 3, \"Number of bits to shift when calculating a new signature\"\n )\n signature_bits = Param.UInt16(12, \"Size of the signature, in bits\")\n signature_table_entries = Param.MemorySize(\n \"1024\", \"Number of entries of the signature table\"\n )\n signature_table_assoc = Param.Unsigned(\n 2, \"Associativity of the signature table\"\n )\n signature_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.signature_table_assoc,\n size=Parent.signature_table_entries,\n ),\n \"Indexing policy of the signature table\",\n )\n signature_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the signature table\"\n )\n\n num_counter_bits = Param.UInt8(\n 3, \"Number of bits of the saturating counters\"\n )\n pattern_table_entries = Param.MemorySize(\n \"4096\", \"Number of entries of the pattern table\"\n )\n pattern_table_assoc = Param.Unsigned(\n 1, \"Associativity of the pattern table\"\n )\n strides_per_pattern_entry = Param.Unsigned(\n 4, \"Number of strides stored in each pattern entry\"\n )\n pattern_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.pattern_table_assoc,\n size=Parent.pattern_table_entries,\n ),\n \"Indexing policy of the pattern table\",\n )\n pattern_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the pattern table\"\n )\n\n prefetch_confidence_threshold = Param.Float(\n 0.5, \"Minimum confidence to issue prefetches\"\n )\n lookahead_confidence_threshold = Param.Float(\n 0.75, \"Minimum confidence to continue exploring lookahead entries\"\n )\n\n\nclass SignaturePathPrefetcherV2(SignaturePathPrefetcher):\n type = \"SignaturePathPrefetcherV2\"\n cxx_class = \"gem5::prefetch::SignaturePathV2\"\n cxx_header = \"mem/cache/prefetch/signature_path_v2.hh\"\n\n signature_table_entries = \"256\"\n signature_table_assoc = 1\n pattern_table_entries = \"512\"\n pattern_table_assoc = 1\n num_counter_bits = 4\n prefetch_confidence_threshold = 0.25\n lookahead_confidence_threshold = 0.25\n\n global_history_register_entries = Param.MemorySize(\n \"8\", \"Number of entries of global history register\"\n )\n global_history_register_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.global_history_register_entries,\n size=Parent.global_history_register_entries,\n ),\n \"Indexing policy of the global history register\",\n )\n global_history_register_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the global history register\"\n )\n\n\nclass AccessMapPatternMatching(ClockedObject):\n type = \"AccessMapPatternMatching\"\n cxx_class = \"gem5::prefetch::AccessMapPatternMatching\"\n cxx_header = \"mem/cache/prefetch/access_map_pattern_matching.hh\"\n\n block_size = Param.Unsigned(\n Parent.block_size,\n \"Cacheline size used by the prefetcher using this object\",\n )\n\n limit_stride = Param.Unsigned(\n 0, \"Limit the strides checked up to -X/X, if 0, disable the limit\"\n )\n start_degree = Param.Unsigned(\n 4, \"Initial degree (Maximum number of prefetches generated\"\n )\n hot_zone_size = Param.MemorySize(\"2KiB\", \"Memory covered by a hot zone\")\n access_map_table_entries = Param.MemorySize(\n \"256\", \"Number of entries in the access map table\"\n )\n access_map_table_assoc = Param.Unsigned(\n 8, \"Associativity of the access map table\"\n )\n access_map_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.access_map_table_assoc,\n size=Parent.access_map_table_entries,\n ),\n \"Indexing policy of the access map table\",\n )\n access_map_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the access map table\"\n )\n high_coverage_threshold = Param.Float(\n 0.25, \"A prefetch coverage factor bigger than this is considered high\"\n )\n low_coverage_threshold = Param.Float(\n 0.125, \"A prefetch coverage factor smaller than this is considered low\"\n )\n high_accuracy_threshold = Param.Float(\n 0.5, \"A prefetch accuracy factor bigger than this is considered high\"\n )\n low_accuracy_threshold = Param.Float(\n 0.25, \"A prefetch accuracy factor smaller than this is considered low\"\n )\n high_cache_hit_threshold = Param.Float(\n 0.875, \"A cache hit ratio bigger than this is considered high\"\n )\n low_cache_hit_threshold = Param.Float(\n 0.75, \"A cache hit ratio smaller than this is considered low\"\n )\n epoch_cycles = Param.Cycles(256000, \"Cycles in an epoch period\")\n offchip_memory_latency = Param.Latency(\n \"30ns\", \"Memory latency used to compute the required memory bandwidth\"\n )\n\n\nclass AMPMPrefetcher(QueuedPrefetcher):\n type = \"AMPMPrefetcher\"\n cxx_class = \"gem5::prefetch::AMPM\"\n cxx_header = \"mem/cache/prefetch/access_map_pattern_matching.hh\"\n ampm = Param.AccessMapPatternMatching(\n AccessMapPatternMatching(), \"Access Map Pattern Matching object\"\n )\n\n\nclass DeltaCorrelatingPredictionTables(SimObject):\n type = \"DeltaCorrelatingPredictionTables\"\n cxx_class = \"gem5::prefetch::DeltaCorrelatingPredictionTables\"\n cxx_header = \"mem/cache/prefetch/delta_correlating_prediction_tables.hh\"\n deltas_per_entry = Param.Unsigned(\n 20, \"Number of deltas stored in each table entry\"\n )\n delta_bits = Param.Unsigned(12, \"Bits per delta\")\n delta_mask_bits = Param.Unsigned(\n 8, \"Lower bits to mask when comparing deltas\"\n )\n table_entries = Param.MemorySize(\"128\", \"Number of entries in the table\")\n table_assoc = Param.Unsigned(128, \"Associativity of the table\")\n table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1, assoc=Parent.table_assoc, size=Parent.table_entries\n ),\n \"Indexing policy of the table\",\n )\n table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the table\"\n )\n\n\nclass DCPTPrefetcher(QueuedPrefetcher):\n type = \"DCPTPrefetcher\"\n cxx_class = \"gem5::prefetch::DCPT\"\n cxx_header = \"mem/cache/prefetch/delta_correlating_prediction_tables.hh\"\n dcpt = Param.DeltaCorrelatingPredictionTables(\n DeltaCorrelatingPredictionTables(),\n \"Delta Correlating Prediction Tables object\",\n )\n\n\nclass IrregularStreamBufferPrefetcher(QueuedPrefetcher):\n type = \"IrregularStreamBufferPrefetcher\"\n cxx_class = \"gem5::prefetch::IrregularStreamBuffer\"\n cxx_header = \"mem/cache/prefetch/irregular_stream_buffer.hh\"\n\n num_counter_bits = Param.Unsigned(\n 2, \"Number of bits of the confidence counter\"\n )\n chunk_size = Param.Unsigned(\n 256, \"Maximum number of addresses in a temporal stream\"\n )\n degree = Param.Unsigned(4, \"Number of prefetches to generate\")\n training_unit_assoc = Param.Unsigned(\n 128, \"Associativity of the training unit\"\n )\n training_unit_entries = Param.MemorySize(\n \"128\", \"Number of entries of the training unit\"\n )\n training_unit_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.training_unit_assoc,\n size=Parent.training_unit_entries,\n ),\n \"Indexing policy of the training unit\",\n )\n training_unit_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the training unit\"\n )\n\n prefetch_candidates_per_entry = Param.Unsigned(\n 16, \"Number of prefetch candidates stored in a SP-AMC entry\"\n )\n address_map_cache_assoc = Param.Unsigned(\n 128, \"Associativity of the PS/SP AMCs\"\n )\n address_map_cache_entries = Param.MemorySize(\n \"128\", \"Number of entries of the PS/SP AMCs\"\n )\n ps_address_map_cache_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.address_map_cache_assoc,\n size=Parent.address_map_cache_entries,\n ),\n \"Indexing policy of the Physical-to-Structural Address Map Cache\",\n )\n ps_address_map_cache_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(),\n \"Replacement policy of the Physical-to-Structural Address Map Cache\",\n )\n sp_address_map_cache_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.address_map_cache_assoc,\n size=Parent.address_map_cache_entries,\n ),\n \"Indexing policy of the Structural-to-Physical Address Mao Cache\",\n )\n sp_address_map_cache_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(),\n \"Replacement policy of the Structural-to-Physical Address Map Cache\",\n )\n\n\nclass SlimAccessMapPatternMatching(AccessMapPatternMatching):\n start_degree = 2\n limit_stride = 4\n\n\nclass SlimDeltaCorrelatingPredictionTables(DeltaCorrelatingPredictionTables):\n table_entries = \"256\"\n table_assoc = 256\n deltas_per_entry = 9\n\n\nclass SlimAMPMPrefetcher(QueuedPrefetcher):\n type = \"SlimAMPMPrefetcher\"\n cxx_class = \"gem5::prefetch::SlimAMPM\"\n cxx_header = \"mem/cache/prefetch/slim_ampm.hh\"\n\n ampm = Param.AccessMapPatternMatching(\n SlimAccessMapPatternMatching(), \"Access Map Pattern Matching object\"\n )\n dcpt = Param.DeltaCorrelatingPredictionTables(\n SlimDeltaCorrelatingPredictionTables(),\n \"Delta Correlating Prediction Tables object\",\n )\n\n\nclass BOPPrefetcher(QueuedPrefetcher):\n type = \"BOPPrefetcher\"\n cxx_class = \"gem5::prefetch::BOP\"\n cxx_header = \"mem/cache/prefetch/bop.hh\"\n score_max = Param.Unsigned(31, \"Max. score to update the best offset\")\n round_max = Param.Unsigned(100, \"Max. round to update the best offset\")\n bad_score = Param.Unsigned(10, \"Score at which the HWP is disabled\")\n rr_size = Param.Unsigned(64, \"Number of entries of each RR bank\")\n tag_bits = Param.Unsigned(12, \"Bits used to store the tag\")\n offset_list_size = Param.Unsigned(\n 46, \"Number of entries in the offsets list\"\n )\n negative_offsets_enable = Param.Bool(\n True,\n \"Initialize the offsets list also with negative values \\\n (i.e. the table will have half of the entries with positive \\\n offsets and the other half with negative ones)\",\n )\n delay_queue_enable = Param.Bool(True, \"Enable the delay queue\")\n delay_queue_size = Param.Unsigned(\n 15, \"Number of entries in the delay queue\"\n )\n delay_queue_cycles = Param.Cycles(\n 60,\n \"Cycles to delay a write in the left RR table from the delay \\\n queue\",\n )\n\n\nclass SBOOEPrefetcher(QueuedPrefetcher):\n type = \"SBOOEPrefetcher\"\n cxx_class = \"gem5::prefetch::SBOOE\"\n cxx_header = \"mem/cache/prefetch/sbooe.hh\"\n latency_buffer_size = Param.Int(32, \"Entries in the latency buffer\")\n sequential_prefetchers = Param.Int(9, \"Number of sequential prefetchers\")\n sandbox_entries = Param.Int(1024, \"Size of the address buffer\")\n score_threshold_pct = Param.Percent(\n 25,\n \"Min. threshold to issue a \\\n prefetch. The value is the percentage of sandbox entries to use\",\n )\n\n\nclass STeMSPrefetcher(QueuedPrefetcher):\n type = \"STeMSPrefetcher\"\n cxx_class = \"gem5::prefetch::STeMS\"\n cxx_header = \"mem/cache/prefetch/spatio_temporal_memory_streaming.hh\"\n\n spatial_region_size = Param.MemorySize(\n \"2KiB\", \"Memory covered by a hot zone\"\n )\n active_generation_table_entries = Param.MemorySize(\n \"64\", \"Number of entries in the active generation table\"\n )\n active_generation_table_assoc = Param.Unsigned(\n 64, \"Associativity of the active generation table\"\n )\n active_generation_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.active_generation_table_assoc,\n size=Parent.active_generation_table_entries,\n ),\n \"Indexing policy of the active generation table\",\n )\n active_generation_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the active generation table\"\n )\n\n pattern_sequence_table_entries = Param.MemorySize(\n \"16384\", \"Number of entries in the pattern sequence table\"\n )\n pattern_sequence_table_assoc = Param.Unsigned(\n 16384, \"Associativity of the pattern sequence table\"\n )\n pattern_sequence_table_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1,\n assoc=Parent.pattern_sequence_table_assoc,\n size=Parent.pattern_sequence_table_entries,\n ),\n \"Indexing policy of the pattern sequence table\",\n )\n pattern_sequence_table_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the pattern sequence table\"\n )\n\n region_miss_order_buffer_entries = Param.Unsigned(\n 131072, \"Number of entries of the Region Miss Order Buffer\"\n )\n add_duplicate_entries_to_rmob = Param.Bool(\n True, \"Add duplicate entries to RMOB\"\n )\n reconstruction_entries = Param.Unsigned(\n 256, \"Number of reconstruction entries\"\n )\n\n\nclass HWPProbeEventRetiredInsts(HWPProbeEvent):\n def register(self):\n if self.obj:\n for name in self.names:\n self.prefetcher.getCCObject().addEventProbeRetiredInsts(\n self.obj.getCCObject(), name\n )\n\n\nclass PIFPrefetcher(QueuedPrefetcher):\n type = \"PIFPrefetcher\"\n cxx_class = \"gem5::prefetch::PIF\"\n cxx_header = \"mem/cache/prefetch/pif.hh\"\n cxx_exports = [PyBindMethod(\"addEventProbeRetiredInsts\")]\n\n prec_spatial_region_bits = Param.Unsigned(\n 2, \"Number of preceding addresses in the spatial region\"\n )\n succ_spatial_region_bits = Param.Unsigned(\n 8, \"Number of subsequent addresses in the spatial region\"\n )\n compactor_entries = Param.Unsigned(2, \"Entries in the temp. compactor\")\n stream_address_buffer_entries = Param.Unsigned(7, \"Entries in the SAB\")\n history_buffer_size = Param.Unsigned(16, \"Entries in the history buffer\")\n\n index_entries = Param.MemorySize(\"64\", \"Number of entries in the index\")\n index_assoc = Param.Unsigned(64, \"Associativity of the index\")\n index_indexing_policy = Param.BaseIndexingPolicy(\n SetAssociative(\n entry_size=1, assoc=Parent.index_assoc, size=Parent.index_entries\n ),\n \"Indexing policy of the index\",\n )\n index_replacement_policy = Param.BaseReplacementPolicy(\n LRURP(), \"Replacement policy of the index\"\n )\n\n def listenFromProbeRetiredInstructions(self, simObj):\n if not isinstance(simObj, SimObject):\n raise TypeError(\"argument must be of SimObject type\")\n self.addEvent(\n HWPProbeEventRetiredInsts(self, simObj, \"RetiredInstsPC\")\n )\n","repo_name":"gem5/gem5","sub_path":"src/mem/cache/prefetch/Prefetcher.py","file_name":"Prefetcher.py","file_ext":"py","file_size_in_byte":23971,"program_lang":"python","lang":"en","doc_type":"code","stars":1196,"dataset":"github-code","pt":"78"}
+{"seq_id":"41759965483","text":"from django.urls import path\nfrom single_pages import views\n\nurlpatterns = [\n path('', views.home), # 홈페이지: 서버IP/\n path('best/', views.best), # 베스트 상품 페이지: 서버IP/best/\n path('mypage/', views.mypage), # 마이페이지: 서버IP/mypage/\n path('company/', views.company), # 회사소개 페이지: 서버IP/company/\n path('cart/', views.cart), # 장바구니 페이지: 서버IP/cart/\n path('check_order/', views.check_order), # 주문 재고 확인: 서버IP/check_order/\n path('order/', views.OrderCreate.as_view()), # 주문하기 페이지: 서버IP/order/\n]\n","repo_name":"soohyxn/S-Market","sub_path":"single_pages/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":613,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"5470106291","text":"import sys\nfrom PyQt5.QtWidgets import *\nfrom PyQt5.QtCore import *\nfrom PyQt5 import QtWidgets,QtCore\n\n\nclass TestForm(QMainWindow): #PyQt5.QtWidgets에서 상속됨\n #생성자\n def __init__(self):\n super().__init__()#부모의 생성자\n self.setupUI()#함수 선언\n def setupUI(self):\n self.setWindowTitle(\"파이큐티 테스트\")\n self.setGeometry(800,400,500,500) # 창크기 앞쪽 두개는 윈도우 위치 뒤쪽 두개는 윈도우창 크기\n\n label1=QLabel(\"입력 테스트\",self)\n label1.move(20,20)\n\n label2=QLabel(\"출력 테스트\",self)\n label2.move(20,60)\n\n self.lineEdit=QLineEdit(\"\",self) #default값\n self.plainEdit=QtWidgets.QPlainTextEdit(self)\n self.plainEdit.setReadOnly(True) #쓰기방지\n\n self.lineEdit.move(100,20)\n self.plainEdit.setGeometry(QtCore.QRect(20,90,260,200))\n\n self.lineEdit.textChanged.connect(self.lineEditChanged)\n self.lineEdit.returnPressed.connect(self.lineEditEnter)\n\n #상태바\n self.statusBar=QStatusBar(self)\n self.setStatusBar(self.statusBar)\n\n def lineEditChanged(self):\n self.statusBar.showMessage(self.lineEdit.text())\n\n def lineEditEnter(self):\n self.plainEdit.appendPlainText(self.lineEdit.text())\n self.lineEdit.clear() # 메모리 해제\n\n\nif __name__ == \"__main__\":\n app=QApplication(sys.argv)\n window=TestForm()\n window.show()\n app.exec_()\n","repo_name":"dreamer9107/MINIPROJECT2","sub_path":"example/1-3.py","file_name":"1-3.py","file_ext":"py","file_size_in_byte":1483,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"37872735384","text":"import base64, datetime, html, json, os\n\nimport requests\n\nfrom commands.CommandTemplate import CommandTemplate\nimport Constants, GlobalStore, PermissionLevel\nfrom util import DateTimeUtil, IrcFormattingUtil, StringUtil\nfrom IrcMessage import IrcMessage\nfrom CustomExceptions import CommandInputException, WebRequestException\nfrom StringWithSuffix import StringWithSuffix\n\n\nclass Command(CommandTemplate):\n\ttriggers = ['twitterwatcher', 'twitterwatch']\n\thelptext = \"Automatically says new tweets of watched accounts. Use parameter 'add' to add an account to watch and 'remove' to stop watching an account. 'latest' shows latest tweet. \" \\\n\t\t\t \"Use 'setname' and 'removename' to set and remove a display name. These parameters need to be followed by a Twitter name. 'list' lists all accounts being watched\"\n\tscheduledFunctionTime = 300.0 #Check every 5 minutes\n\trunInThread = True\n\n\twatchData = {} #keys are Twitter usernames, contains fields with highest ID and which channel(s) to report new tweets to, and a display name if specified\n\tMAX_TWEETS_TO_MENTION = 3\n\tSECONDS_AGE_FOR_FULL_DATE = 604800 # After 7 days, don't list a tweet as '6 days, 7 hours ago', but as the full date\n\n\tdef onLoad(self):\n\t\tGlobalStore.commandhandler.addCommandFunction(__file__, 'getTweetDescription', self.getTweetDescription)\n\n\t\t#First retrieve which Twitter accounts we should follow, if that file exists\n\t\twatchedFilepath = os.path.join(GlobalStore.scriptfolder, 'data', 'WatchedTwitterAccounts.json')\n\t\tif os.path.exists(watchedFilepath):\n\t\t\twith open(watchedFilepath, 'r', encoding='utf-8') as watchedFile:\n\t\t\t\tself.watchData = json.load(watchedFile)\n\t\t#If we can't identify to Twitter, disable the automatic check for new messages\n\t\tif not GlobalStore.commandhandler.getApiKey('key', 'twitter') or not GlobalStore.commandhandler.getApiKey('secret', 'twitter'):\n\t\t\tself.logWarning(\"[TwitterWatcher] Twitter API credentials not found!\")\n\t\t\tself.scheduledFunctionTime = None\n\n\tdef executeScheduledFunction(self):\n\t\tself.checkForNewTweets()\n\n\tdef execute(self, message):\n\t\t\"\"\"\n\t\t:type message: IrcMessage\n\t\t\"\"\"\n\t\tif message.messagePartsLength == 0:\n\t\t\tmessage.reply(self.helptext)\n\t\t\treturn\n\n\t\tparameter = message.messageParts[0].lower()\n\t\tserverChannelPair = [message.bot.serverfolder, message.source] #List not tuple, because JSON can't save tuples and converts them to a list\n\n\t\t#Start with the commands that don't need a username parameter\n\t\tif parameter == 'help':\n\t\t\tmessage.reply(self.helptext)\n\t\t\treturn\n\t\tif parameter == 'list':\n\t\t\t#List all the accounts we're watching for this channel\n\t\t\twatchlist = []\n\t\t\tfor username, usernameData in self.watchData.items():\n\t\t\t\tif serverChannelPair in usernameData['targets']:\n\t\t\t\t\twatchlist.append(self.getDisplayName(username))\n\t\t\twatchlistLength = len(watchlist)\n\t\t\tif watchlistLength == 0:\n\t\t\t\treply = \"I'm not watching any Twitter users for this channel\"\n\t\t\telif watchlistLength == 1:\n\t\t\t\treply = \"I just watch {} for the people here\".format(watchlist[0])\n\t\t\telse:\n\t\t\t\twatchlist.sort()\n\t\t\t\treply = \"I watch {:,} Twitter users for this channel: {}\".format(watchlistLength, \"; \".join(watchlist))\n\t\t\tmessage.reply(reply)\n\t\t\treturn\n\t\t#'update' forces an update check, but it's only available to admins. Also doesn't need a username\n\t\tif parameter == 'update':\n\t\t\tif not message.doesSenderHavePermission(PermissionLevel.BOT):\n\t\t\t\treply = \"Only my bot admin(s) can force an update, sorry!\"\n\t\t\telif self.scheduledFunctionIsExecuting:\n\t\t\t\treply = \"I was updating already! Lucky you, now it'll be done quicker\"\n\t\t\telse:\n\t\t\t\tself.checkForNewTweets()\n\t\t\t\tself.resetScheduledFunctionGreenlet()\n\t\t\t\treply = \"Finished forced TwitterWatcher update check\"\n\t\t\tmessage.reply(reply)\n\t\t\treturn\n\n\t\t#All the other parameters need an account name, so check for that now\n\t\tif message.messagePartsLength == 1:\n\t\t\tmessage.reply(\"Please add a Twitter account name too, so I know where to look\")\n\t\t\treturn\n\n\t\taccountName = message.messageParts[1]\n\t\taccountNameLowered = accountName.lower()\n\t\tisUserBeingWatchedHere = accountNameLowered in self.watchData and serverChannelPair in self.watchData[accountNameLowered]['targets']\n\n\t\tif parameter == 'add':\n\t\t\tif isUserBeingWatchedHere:\n\t\t\t\treply = \"I'm already keeping a close eye on {}. On their tweets, I mean\".format(self.getDisplayName(accountNameLowered, accountName))\n\t\t\telif not message.doesSenderHavePermission(PermissionLevel.CHANNEL):\n\t\t\t\traise CommandInputException(\"Sorry, only my channel admins are allowed to add people to my watch list\")\n\t\t\telse:\n\t\t\t\t#New account\n\t\t\t\tif accountNameLowered not in self.watchData:\n\t\t\t\t\tself.watchData[accountNameLowered] = {'targets': [serverChannelPair]}\n\t\t\t\t#Existing account\n\t\t\t\telse:\n\t\t\t\t\tself.watchData[accountNameLowered]['targets'].append(serverChannelPair)\n\t\t\t\t#If a display name was provided, add that too\n\t\t\t\tif message.messagePartsLength > 2:\n\t\t\t\t\tself.watchData[accountNameLowered]['displayname'] = \" \".join(message.messageParts[2:])\n\t\t\t\telif accountName != accountNameLowered:\n\t\t\t\t\tself.watchData[accountNameLowered]['displayname'] = accountName\n\t\t\t\t#Save the whole thing\n\t\t\t\tself.saveWatchData()\n\t\t\t\tself.checkForNewTweets([accountNameLowered], False)\n\t\t\t\treply = \"Ok, I'll keep you informed about any new tweets {}... makes? Tweets? What's the verb here?\".format(self.getDisplayName(accountNameLowered))\n\t\telif parameter == 'remove':\n\t\t\tif not isUserBeingWatchedHere:\n\t\t\t\treply = \"I already wasn't watching {}! Not even secretly\".format(accountName)\n\t\t\telif not message.doesSenderHavePermission(PermissionLevel.CHANNEL):\n\t\t\t\traise CommandInputException(\"Only my channel admins are allowed to remove people from my watch list, sorry\")\n\t\t\telse:\n\t\t\t\tself.watchData[accountNameLowered]['targets'].remove(serverChannelPair)\n\t\t\t\t#If this channel was the only place we were reporting this user's tweets to, remove it all together\n\t\t\t\tif len(self.watchData[accountNameLowered]['targets']) == 0:\n\t\t\t\t\tdel self.watchData[accountNameLowered]\n\t\t\t\tself.saveWatchData()\n\t\t\t\treply = \"Ok, I won't keep you updated on whatever {} posts. Tweets. Messages? I don't know the proper verb\".format(accountName)\n\t\telif parameter == 'latest':\n\t\t\t#Download the latest tweet for the provided username\n\t\t\ttry:\n\t\t\t\ttweets = self.downloadTweets(accountNameLowered, 1)\n\t\t\texcept WebRequestException as wre:\n\t\t\t\tself.logError(\"[TwitterWatcher] Error occured while downloading single tweet for user {}: {}\".format(accountName, wre))\n\t\t\t\treply = \"Woops, something went wrong there. Tell my owner(s), maybe it's something they can fix. Or maybe it's Twitter's fault, in which case all we can do is wait\"\n\t\t\telse:\n\t\t\t\tif not tweets:\n\t\t\t\t\treply = \"Sorry, I couldn't find any tweets by {}. Maybe they haven't tweeted yet, or maybe you made a typo?\".format(accountName)\n\t\t\t\telse:\n\t\t\t\t\treply = self.formatNewTweetText(accountName, tweets[0], addTweetAge=True)\n\t\telif parameter == 'setname':\n\t\t\t#Allow users to set a display name\n\t\t\tif not isUserBeingWatchedHere:\n\t\t\t\treply = \"I'm not watching {}, so I can't change the display name. Add them with the 'add' parameter first\".format(accountName)\n\t\t\telif not message.doesSenderHavePermission(PermissionLevel.CHANNEL):\n\t\t\t\traise CommandInputException(\"Sorry, only my channel admins are allowed to set nicknames for the people on my watch list\")\n\t\t\telif message.messagePartsLength < 2:\n\t\t\t\treply = \"Please add a display name for '{}' too. You don't want me thinking up nicknames for people\".format(accountName)\n\t\t\telse:\n\t\t\t\tself.watchData[accountNameLowered]['displayname'] = \" \".join(message.messageParts[2:])\n\t\t\t\tself.saveWatchData()\n\t\t\t\treply = \"Ok, I will call {} '{}' from now on\".format(accountName, self.watchData[accountNameLowered]['displayname'])\n\t\telif parameter == 'removename':\n\t\t\tif not isUserBeingWatchedHere:\n\t\t\t\treply = \"I wasn't calling them anything anyway, since I'm not following {}\".format(accountName)\n\t\t\telif 'displayname' not in self.watchData[accountNameLowered]:\n\t\t\t\treply = \"I didn't have a nickname listed for {} anyway, so I guess I did what you asked?\".format(accountNameLowered)\n\t\t\telif not message.doesSenderHavePermission(PermissionLevel.CHANNEL):\n\t\t\t\traise CommandInputException(\"Only my channel admins are allowed to remove nicknames for the people on my watch list, sorry\")\n\t\t\telse:\n\t\t\t\tdel self.watchData[accountNameLowered]['displayname']\n\t\t\t\tself.saveWatchData()\n\t\t\t\treply = \"Ok, I will just call them by their account name, {}\".format(accountName)\n\t\telse:\n\t\t\treply = \"I don't know what to do with the parameter '{}', sorry. Try rereading the help text?\".format(parameter)\n\n\t\tmessage.replyWithLengthLimit(reply)\n\n\tdef updateTwitterToken(self):\n\t\tapiKey = GlobalStore.commandhandler.getApiKey('key', 'twitter')\n\t\tapiSecret = GlobalStore.commandhandler.getApiKey('secret', 'twitter')\n\t\tif not apiKey or not apiSecret:\n\t\t\tself.logError(\"[TwitterWatcher] No Twitter API key and/or secret found\")\n\t\t\treturn False\n\n\t\tcredentials = base64.b64encode(\"{}:{}\".format(apiKey, apiSecret))\n\t\theaders = {\"Authorization\": \"Basic {}\".format(credentials), \"Content-Type\": \"application/x-www-form-urlencoded;charset=UTF-8\"}\n\t\tdata = \"grant_type=client_credentials\"\n\n\t\treq = requests.post(\"https://api.twitter.com/oauth2/token\", data=data, headers=headers)\n\t\treply = json.loads(req.text)\n\t\tif 'access_token' not in reply:\n\t\t\tself.logError(\"[TwitterWatcher] An error occurred while retrieving Twitter token: \" + json.dumps(reply))\n\t\t\treturn False\n\n\t\tGlobalStore.commandhandler.apikeys['twitter']['token'] = reply['access_token']\n\t\tGlobalStore.commandhandler.apikeys['twitter']['tokentype'] = reply['token_type']\n\t\tGlobalStore.commandhandler.saveApiKeys()\n\t\treturn True\n\n\tdef downloadTweets(self, username, maxTweetCount=200, downloadNewerThanId=None, downloadOlderThanId=None, includeReplies=False, includeRetweets=False):\n\t\t# First check if we can even connect to the Twitter API\n\t\tapiToken = GlobalStore.commandhandler.getApiKey('token', 'twitter')\n\t\tapiTokenType = GlobalStore.commandhandler.getApiKey('tokentype', 'twitter')\n\t\tif not apiToken or not apiTokenType:\n\t\t\tself.logInfo(\"[TwitterWatcher] No twitter token found, retrieving a new one\")\n\t\t\ttokenUpdateSuccess = self.updateTwitterToken()\n\t\t\tif not tokenUpdateSuccess:\n\t\t\t\tself.logError(\"Unable to retrieve a new Twitter token!\")\n\t\t\t\traise WebRequestException(\"Unable to retrieve Twitter authentication token!\")\n\n\t\t# Now download tweets!\n\t\theaders = {'Authorization': \"{} {}\".format(apiTokenType, apiToken)}\n\t\tparams = {'screen_name': username, 'count': min(200, maxTweetCount), 'trim_user': 'true', 'tweet_mode': 'extended',\n\t\t\t\t 'exclude_replies': 'false' if includeReplies else 'true',\n\t\t\t\t 'include_rts': True} # Always get retweets, remove them later if necessary. Needed because 'count' always includes retweets, even if you don't want them\n\t\tif downloadOlderThanId:\n\t\t\tparams['max_id'] = downloadOlderThanId\n\n\t\ttweets = []\n\t\tif downloadNewerThanId:\n\t\t\tparams['since_id'] = downloadNewerThanId\n\t\treq = None\n\t\twhile len(tweets) < maxTweetCount:\n\t\t\tparams['count'] = maxTweetCount - len(tweets) # Get as many tweets as we still need\n\t\t\ttry:\n\t\t\t\treq = requests.get(\"https://api.twitter.com/1.1/statuses/user_timeline.json\", headers=headers, params=params, timeout=20.0)\n\t\t\t\tapireply = json.loads(req.text)\n\t\t\texcept requests.exceptions.Timeout:\n\t\t\t\tself.logError(\"[TwitterWatcher] Twitter API reply took too long to arrive\")\n\t\t\t\traise WebRequestException(\"Twitter took too long to respond\")\n\t\t\texcept ValueError:\n\t\t\t\tself.logError(\"[TwitterWatcher] Didn't get parsable JSON return from Twitter API: {}\".format(StringUtil.removeNewlines(req.text, '|') if req else \"[no response retrieved]\"))\n\t\t\t\traise WebRequestException(\"Twitter API returned unexpected data\")\n\t\t\texcept Exception as e:\n\t\t\t\tself.logError(\"[TwitterWatcher] Tweet download threw an unexpected error of type '{}': {}\".format(type(e), str(e)))\n\t\t\t\traise WebRequestException(\"Unknown error occurred while retrieving Twitter API data\")\n\n\t\t\tif len(apireply) == 0:\n\t\t\t\t# No more tweets to parse!\n\t\t\t\tbreak\n\t\t\t# Check for errors\n\t\t\tif req.status_code == 404:\n\t\t\t\traise WebRequestException(\"The provided tweet(s) couldn't be found, sorry\")\n\t\t\tif isinstance(apireply, dict) and 'errors' in apireply:\n\t\t\t\terrorMessages = '; '.join(e['message'] for e in apireply['errors'])\n\t\t\t\tself.logError(\"[TwitterWatcher] Error occurred while retrieving tweets for {}. Parameters: {}; apireply: {}; errors: {}\".format(username, params, apireply, errorMessages))\n\t\t\t\traise WebRequestException(\"The Twitter API reply contained errors\")\n\t\t\t# Sometimes the API does not return a list of tweets for some reason. Catch that\n\t\t\tif not isinstance(apireply, list):\n\t\t\t\tself.logError(\"[TwitterWatcher] Unexpected reply from Twitter API. Expected tweet list, got {}: {}\".format(type(apireply), apireply))\n\t\t\t\traise WebRequestException(\"The Twitter API reply contained unexpected data\")\n\t\t\t# Tweets are sorted reverse-chronologically, so we can get the highest ID from the first tweet\n\t\t\tparams['since_id'] = apireply[0]['id']\n\t\t\t# Remove retweets if necessary (done manually to make the 'count' variable be accurate)\n\t\t\tif not includeRetweets:\n\t\t\t\tapireply = [t for t in apireply if 'retweeted_status' not in t]\n\t\t\t# There are tweets, store those\n\t\t\ttweets.extend(apireply)\n\t\treturn tweets\n\n\tdef checkForNewTweets(self, usernamesToCheck=None, reportNewTweets=True):\n\t\tif not usernamesToCheck:\n\t\t\tusernamesToCheck = self.watchData #Don't copy the username list\n\t\tif not usernamesToCheck:\n\t\t\treturn\n\n\t\tnow = datetime.datetime.utcnow()\n\t\twatchDataChanged = False\n\t\ttweetAgeCutoff = self.scheduledFunctionTime * 1.1 #Give tweet age a little grace period, so tweets can't fall between checks\n\t\t#Retrieve the latest tweets for every account.\n\t\tfor username in usernamesToCheck:\n\t\t\tif username not in self.watchData:\n\t\t\t\tself.logWarning(\"[TwitterWatcher] Asked to check account '{}' for new tweets, but it is not in the watchlist\".format(username))\n\t\t\t\tcontinue\n\t\t\ttry:\n\t\t\t\ttweets = self.downloadTweets(username, maxTweetCount=10, downloadNewerThanId=self.watchData[username].get('highestId', None), includeRetweets=False)\n\t\t\texcept WebRequestException as wre:\n\t\t\t\tself.logError(\"[TwitterWatcher] Couldn't retrieve tweets for '{}': {}\".format(username, wre))\n\t\t\t\tcontinue\n\t\t\t#If there aren't any new tweets, move on\n\t\t\tif len(tweets) == 0:\n\t\t\t\tcontinue\n\t\t\t#Always store the highest ID, so we don't encounter the same tweet twice\n\t\t\twatchDataChanged = True\n\t\t\tself.watchData[username]['highestId'] = tweets[0]['id']\n\t\t\t#If we don't have to actually report the tweets, then we have nothing left to do\n\t\t\tif not reportNewTweets:\n\t\t\t\tcontinue\n\n\t\t\t#Go through the tweets to check if they're not too old to report\n\t\t\tfirstOldTweetIndex = -1\n\t\t\tfor index, tweet in enumerate(tweets):\n\t\t\t\tif (now - self.getTweetPostTime(tweet['created_at'])).total_seconds() > tweetAgeCutoff:\n\t\t\t\t\tfirstOldTweetIndex = index\n\t\t\t\t\tbreak\n\t\t\t#If all tweets are old, stop here\n\t\t\tif firstOldTweetIndex == 0:\n\t\t\t\tcontinue\n\t\t\t#Otherwise remove the old tweet and every older tweet\n\t\t\telif firstOldTweetIndex > -1:\n\t\t\t\ttweets = tweets[:firstOldTweetIndex]\n\n\t\t\t#To prevent spam, only mention the latest few tweets, in case of somebody posting a LOT in a short timespan\n\t\t\tif len(tweets) > self.MAX_TWEETS_TO_MENTION:\n\t\t\t\ttweetsSkipped = len(tweets) - self.MAX_TWEETS_TO_MENTION\n\t\t\t\ttweets = tweets[-self.MAX_TWEETS_TO_MENTION:]\n\t\t\telse:\n\t\t\t\ttweetsSkipped = 0\n\n\t\t\t#Reverse the tweets so we get them old to new, instead of new to old\n\t\t\ttweets.reverse()\n\t\t\t#New recent tweets! Shout about it (if we're in the place where we should shout)\n\t\t\tfor target in self.watchData[username]['targets']:\n\t\t\t\t#'target' is a tuple with the server name at [0] and the channel name at [1]\n\t\t\t\t#Just ignore it if we're either not on the server or not in the channel\n\t\t\t\tif target[0] not in GlobalStore.bothandler.bots:\n\t\t\t\t\tcontinue\n\t\t\t\ttargetbot = GlobalStore.bothandler.bots[target[0]]\n\t\t\t\tif target[1] not in targetbot.channelsUserList:\n\t\t\t\t\tcontinue\n\t\t\t\ttargetchannel = target[1]\n\t\t\t\t#Now go tell that channel all about the tweets\n\t\t\t\tfor tweet in tweets:\n\t\t\t\t\tformattedTweet = self.formatNewTweetText(username, tweet)\n\t\t\t\t\ttargetbot.sendLengthLimitedMessage(targetchannel, formattedTweet.mainString, suffix=formattedTweet.suffix)\n\t\t\t\t#If we skipped a few tweets, make a mention of that too\n\t\t\t\tif tweetsSkipped > 0:\n\t\t\t\t\ttargetbot.sendMessage(targetchannel, \"(skipped {:,} of {}'s tweets)\".format(tweetsSkipped, self.getDisplayName(username)))\n\t\tif watchDataChanged:\n\t\t\tself.saveWatchData()\n\n\tdef formatNewTweetText(self, username, tweetData, addTweetAge=False, addTweetUrl=True):\n\t\t# Fix special characters (convert '&' to '&' for instance), and remove newlines\n\t\tformattedTweetText = StringUtil.removeNewlines(html.unescape(tweetData['full_text']), Constants.GREY_SEPARATOR)\n\t\tsuffixes = []\n\t\t#Remove the link to the photo at the end, but mention that there is one\n\t\tif 'media' in tweetData['entities']:\n\t\t\tmediaTypes = []\n\t\t\tfor mediaItem in tweetData['entities']['media']:\n\t\t\t\tformattedTweetText = formattedTweetText.replace(mediaItem['url'], '')\n\t\t\t\tif mediaItem['type'] not in mediaTypes:\n\t\t\t\t\tmediaTypes.append(mediaItem['type'])\n\t\t\tsuffixes.append(\"(has {})\".format(\", \".join(mediaTypes)))\n\t\t# Finalize the return text\n\t\tformattedTweetText = \"{}: {}\".format(IrcFormattingUtil.makeTextBold(self.getDisplayName(username)), formattedTweetText)\n\t\tif addTweetAge:\n\t\t\tpostDateTime = self.getTweetPostTime(tweetData['created_at'])\n\t\t\ttweetAge = datetime.datetime.utcnow() - postDateTime\n\t\t\t# For older tweets, list the post date, otherwise list how old it is\n\t\t\ttweetAgeString = \" | \" # Use a normal pipe here instead of GREY_SEPARATOR, because we're going to color the whole age string grey\n\t\t\tif tweetAge.total_seconds() > self.SECONDS_AGE_FOR_FULL_DATE:\n\t\t\t\ttweetAgeString += postDateTime.strftime('%Y-%m-%d')\n\t\t\telif tweetAge.total_seconds() <= 60:\n\t\t\t\ttweetAgeString += \"posted just now\"\n\t\t\telse:\n\t\t\t\ttweetAgeString += f\"{DateTimeUtil.durationSecondsToText(tweetAge.total_seconds(), precision=DateTimeUtil.MINUTES)} ago\"\n\t\t\tsuffixes.append(IrcFormattingUtil.makeTextColoured(tweetAgeString, IrcFormattingUtil.Colours.GREY))\n\t\tif addTweetUrl:\n\t\t\tsuffixes.append(\" | https://twitter.com/_/status/\") #Use _ instead of username to save some characters\n\t\t\tsuffixes.append(tweetData['id_str'])\n\t\treturn StringWithSuffix(formattedTweetText, suffixes)\n\n\tdef getTweetDescription(self, twitterUsername, tweetId, addTweetUrl=True):\n\t\t\"\"\"\n\t\tGet a display string describing the tweet from the provided ID\n\t\t:param twitterUsername: The username of the person that made the tweet\n\t\t:param tweetId: The tweet ID to get a description of\n\t\t:param addTweetUrl: If True (the default), the URL to the tweet will be added to the end of the display string\n\t\t:return: A StringWithSuffix describing the tweet, or None if the tweet couldn't be retrieved\n\t\t\"\"\"\n\t\tif not isinstance(tweetId, int):\n\t\t\ttweetId = int(tweetId, 10)\n\t\ttweetList = self.downloadTweets(username=twitterUsername, downloadNewerThanId=tweetId-1, downloadOlderThanId=tweetId+1, maxTweetCount=1)\n\t\tif not tweetList:\n\t\t\treturn None\n\t\treturn self.formatNewTweetText(twitterUsername, tweetList[0], addTweetAge=True, addTweetUrl=addTweetUrl)\n\n\t@staticmethod\n\tdef getTweetPostTime(createdAt):\n\t\treturn datetime.datetime.strptime(createdAt, \"%a %b %d %H:%M:%S +0000 %Y\")\n\n\tdef getDisplayName(self, username, alternativeName=None):\n\t\tif username not in self.watchData:\n\t\t\treturn username\n\t\treturn self.watchData[username].get('displayname', alternativeName if alternativeName else username)\n\n\tdef saveWatchData(self):\n\t\twatchDataFilePath = os.path.join(GlobalStore.scriptfolder, 'data', 'WatchedTwitterAccounts.json')\n\t\twith open(watchDataFilePath, 'w', encoding='utf-8') as watchDataFile:\n\t\t\twatchDataFile.write(json.dumps(self.watchData))\n","repo_name":"Didero/DideRobot","sub_path":"commands/TwitterWatcher.py","file_name":"TwitterWatcher.py","file_ext":"py","file_size_in_byte":19778,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"78"}
+{"seq_id":"4763575162","text":"\r\ndef sentenceMaker(phrase):\r\n capitalized = phrase.capitalize()\r\n interrogatives = \"how\", \"what\", \"why\"\r\n if phrase.startswith(interrogatives):\r\n return (capitalized + '?')\r\n else:\r\n return (capitalized + '.')\r\n\r\ninputs = []\r\nwhile True:\r\n user_input = input(\"Say something: \")\r\n if user_input == '/end':\r\n break\r\n else:\r\n inputs.append(sentenceMaker(user_input))\r\n\r\nprint(\" \".join(inputs))","repo_name":"xraycodes/python_practice","sub_path":"text.py","file_name":"text.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"22540251734","text":"\"\"\"Policy for MAGE using PyTorch.\"\"\"\nfrom typing import List, Tuple\n\nfrom nnrl.nn.critic import HardValue\nfrom nnrl.optim import build_optimizer\n\nfrom raylab.agents.sop import SOPTorchPolicy\nfrom raylab.options import configure, option\nfrom raylab.policy import EnvFnMixin\nfrom raylab.policy.action_dist import WrapDeterministicPolicy\nfrom raylab.policy.losses import MAGE, MaximumLikelihood\nfrom raylab.policy.model_based.lightning import LightningModelTrainer, TrainingSpec\nfrom raylab.policy.model_based.policy import MBPolicyMixin\nfrom raylab.utils.types import StatDict\n\n\ndef default_model_training() -> dict:\n \"\"\"Model training routine used by MAGE paper.\"\"\"\n spec = TrainingSpec()\n spec.datamodule.holdout_ratio = 0.0\n spec.datamodule.max_holdout = 0\n spec.datamodule.batch_size = 256\n spec.datamodule.shuffle = True\n spec.datamodule.num_workers = 0\n spec.training.max_epochs = None\n spec.training.max_steps = 120\n spec.training.patience = None\n spec.warmup = spec.training\n return spec.to_dict()\n\n\n@configure\n@option(\"model_training\", default=default_model_training())\n@option(\"model_update_interval\", default=25)\n@option(\"improvement_steps\", default=10, override=True)\n@option(\"policy_delay\", 2, override=True)\n@option(\"batch_size\", default=1024, override=True)\n@option(\"lambda\", default=0.05, help=\"TD error regularization for MAGE loss\")\n@option(\"module/type\", \"MAGE\", override=True)\n@option(\"optimizer/models\", default={\"type\": \"Adam\", \"lr\": 1e-4, \"weight_decay\": 1e-4})\n@option(\"optimizer/actor\", default={\"type\": \"Adam\", \"lr\": 1e-4}, override=True)\n@option(\"optimizer/critics\", default={\"type\": \"Adam\", \"lr\": 1e-4}, override=True)\n@option(\"exploration_config/pure_exploration_steps\", 1000, override=True)\nclass MAGETorchPolicy(MBPolicyMixin, EnvFnMixin, SOPTorchPolicy):\n \"\"\"MAGE policy in PyTorch to use with RLlib.\n\n Attributes:\n loss_model: maximum likelihood loss for model ensemble\n loss_actor: deterministic policy gradient loss\n loss_critic: model-based action-value-gradient estimator loss\n \"\"\"\n\n # pylint:disable=too-many-ancestors\n dist_class = WrapDeterministicPolicy\n model_trainer: LightningModelTrainer\n\n def __init__(self, observation_space, action_space, config):\n super().__init__(observation_space, action_space, config)\n self._set_model_loss()\n self._set_critic_loss()\n self.build_timers()\n self.model_trainer = LightningModelTrainer(\n models=self.module.models,\n loss_fn=self.loss_model,\n optimizer=self.optimizers[\"models\"],\n replay=self.replay,\n config=self.config,\n )\n\n def _set_model_loss(self):\n self.loss_model = MaximumLikelihood(self.module.models)\n\n def _set_critic_loss(self):\n module = self.module\n target_critic = HardValue(\n policy=module.target_actor, q_value=module.target_critics\n )\n self.loss_critic = MAGE(\n critics=module.critics,\n policy=module.actor,\n target_critic=target_critic,\n models=module.models,\n )\n self.loss_critic.gamma = self.config[\"gamma\"]\n self.loss_critic.lambd = self.config[\"lambda\"]\n\n def train_dynamics_model(\n self, warmup: bool = False\n ) -> Tuple[List[float], StatDict]:\n return self.model_trainer.optimize(warmup=warmup)\n\n def compile(self):\n super().compile()\n for loss in (self.loss_model, self.loss_actor, self.loss_critic):\n loss.compile()\n\n def _set_reward_hook(self):\n self.loss_critic.set_reward_fn(self.reward_fn)\n\n def _set_termination_hook(self):\n self.loss_critic.set_termination_fn(self.termination_fn)\n\n def _make_optimizers(self):\n optimizers = super()._make_optimizers()\n config = self.config[\"optimizer\"]\n optimizers[\"models\"] = build_optimizer(self.module.models, config[\"models\"])\n return optimizers\n","repo_name":"0xangelo/raylab","sub_path":"raylab/agents/mage/policy.py","file_name":"policy.py","file_ext":"py","file_size_in_byte":3988,"program_lang":"python","lang":"en","doc_type":"code","stars":52,"dataset":"github-code","pt":"78"}
+{"seq_id":"41609747806","text":"import machine\nfrom machine import Pin, ADC, Timer\nimport utime\n\nsensor_temp = ADC(4)\nconversion_factor = 3.3 / (65535)\n\n\"\"\"\ntim = Timer()\ndef tick(timer):\n reading = sensor_temp.read_u16() * conversion_factor \n temperature = 27 - (reading - 0.706)/0.001721\n# print(temperature)\n print(\"Temperature : \" + str(temperature) + \"°C\")\n \n freqValue = int(machine.freq() / 1000000)\n print(\"MCU Clock Speed : \" + str(freqValue), \" MHz\")\n\ntim.init(freq=5, mode=Timer.PERIODIC, callback=tick)\n\"\"\"\n\nfreqValue = int(machine.freq() / 1000000)\nprint(\"MCU Clock Speed : \" + str(freqValue), \" MHz\")\n\nwhile True:\n reading = sensor_temp.read_u16() * conversion_factor \n temperature = 27 - (reading - 0.706)/0.001721\n print(\"MCU Temperature : \" + str(temperature) + \"°C\")\n \n utime.sleep(2)","repo_name":"AxiomaticEmbedded/MLS-Hat","sub_path":"BaseA1/MicroPython/CpuTemperature.py","file_name":"CpuTemperature.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"14219877363","text":"'''\n增加两个变量 一个记录当前待打印的节点数量 一个记录下一层待打印的节点数量\n'''\nfrom collections import deque\nclass Solution:\n def levelOrder(self, root: TreeNode) -> List[List[int]]:\n if not root:\n return []\n result,queue=[],deque()\n queue.append(root)\n toBeprint = 1\n nextlevel = 0\n while queue:\n mid = []\n for _ in range(toBeprint):\n node = queue.popleft()\n mid.append(node.val)\n if node.left:\n queue.append(node.left)\n nextlevel+=1\n if node.right:\n queue.append(node.right)\n nextlevel+=1\n toBeprint = nextlevel\n nextlevel = 0\n result.append(mid)\n return result","repo_name":"KyleC14/SwordToOfferPractice","sub_path":"code/Question32-2/Solution1.py","file_name":"Solution1.py","file_ext":"py","file_size_in_byte":854,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"8638530926","text":"import sys\nfrom collections import deque\ninput = sys.stdin.readline\n\n\ndef prime_under_man():\n arr = [2]\n for i in range(3, 10000, 2):\n j = 3\n find = True\n while j*j <= i:\n if i % j == 0:\n find = False\n break\n j += 2\n if find:\n arr.append(i)\n return arr\n\n\nprimes = prime_under_man()\nfor _ in range(int(input())):\n n = int(input())\n min = 10000\n arr = []\n filtered_primes = list(filter(lambda x: x < n, primes))\n for prime in filtered_primes:\n if prime > n - prime:\n continue\n if n - prime in filtered_primes and min > abs(prime - n + prime):\n arr = sorted([prime, n-prime])\n min = abs(prime - n + prime)\n print(' '.join(list(map(str, arr))))\n","repo_name":"yangwooseong/algorithm","sub_path":"boj/9020.py","file_name":"9020.py","file_ext":"py","file_size_in_byte":806,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"24478805745","text":"from tests.base import TestBase\n\nfrom aim.sdk.run import Run\nfrom aim.storage.context import Context\n\n\nclass TestRunSequenceHomogeneousValues(TestBase):\n def test_different_types_on_different_contexts_and_runs(self):\n run = Run(system_tracking_interval=None)\n # same sequence name, different contexts\n run.track(1., name='numbers', context={'type': 'float'})\n run.track(1, name='numbers', context={'type': 'integer'})\n\n run2 = Run(system_tracking_interval=None)\n # same sequence name, different runs\n run2.track(1, name='numbers', context={'type': 'float'})\n\n def test_incompatible_type_during_tracking(self):\n run = Run(system_tracking_interval=None)\n run.track(1., name='numbers', context={})\n with self.assertRaises(ValueError) as cm:\n run.track(1, name='numbers', context={})\n exception = cm.exception\n self.assertEqual('Cannot log value \\'1\\' on sequence \\'numbers\\'. Incompatible data types.', exception.args[0])\n\n def test_incompatible_type_after_tracking_restart(self):\n run = Run(system_tracking_interval=None)\n run_hash = run.hash\n run.track(1., name='numbers', context={})\n run.finalize()\n del run\n\n new_run = Run(run_hash=run_hash, system_tracking_interval=None)\n with self.assertRaises(ValueError) as cm:\n new_run.track(1, name='numbers', context={})\n exception = cm.exception\n self.assertEqual('Cannot log value \\'1\\' on sequence \\'numbers\\'. Incompatible data types.', exception.args[0])\n\n def test_type_compatibility_for_empty_list(self):\n run = Run(system_tracking_interval=None)\n context = {}\n ctx = Context(context)\n seq_name = 'obj_list'\n\n sequence_info = run.meta_run_tree.subtree(('traces', ctx.idx, seq_name))\n typed_sequences_info = run.meta_tree.subtree('traces_types')\n\n run.track([], name=seq_name, context=context)\n self.assertEqual('list', sequence_info['dtype'])\n self.assertEqual(1, typed_sequences_info['list', ctx.idx, seq_name])\n self.assertIsNone(typed_sequences_info.get(('list(float)', ctx.idx, seq_name), None))\n\n run.track([], name=seq_name, context=context)\n self.assertEqual('list', sequence_info['dtype'])\n self.assertIsNone(typed_sequences_info.get(('list(float)', ctx.idx, seq_name), None))\n\n run.track([1.], name=seq_name, context=context)\n self.assertEqual('list(float)', sequence_info['dtype'])\n self.assertEqual(1, typed_sequences_info['list(float)', ctx.idx, seq_name])\n\n run.track([], name=seq_name, context=context)\n self.assertEqual('list(float)', sequence_info['dtype'])\n\n with self.assertRaises(ValueError) as cm:\n run.track([5], name=seq_name, context=context)\n exception = cm.exception\n self.assertEqual(\n f'Cannot log value \\'{[5]}\\' on sequence \\'{seq_name}\\'. Incompatible data types.',\n exception.args[0])\n","repo_name":"Dong-Ki-Lee/aim-test","sub_path":"tests/sdk/test_run_track_type_checking.py","file_name":"test_run_track_type_checking.py","file_ext":"py","file_size_in_byte":3027,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"71916894973","text":"def longestConsecutive(nums: list[int]) -> int:\n nSet = set(nums)\n output = 0\n longest = 0 \n for i in range(len(nums)):\n # print(\"numsi: \", str(nums[i]))\n if (nums[i] - 1) not in nSet:\n longest = 0\n while nums[i] + longest in nSet:\n # print(\"numsi + longest: \" + str(nums[i] + longest))\n longest += 1\n output = max(output, longest)\n return output\n\n\nprint(\n longestConsecutive(nums = [9,1,4,7,3,-1,0,5,8,-1,6])\n)","repo_name":"IonHaryono/LeetCode","sub_path":"128. Longest Consecutive Sequence.py","file_name":"128. Longest Consecutive Sequence.py","file_ext":"py","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"39226012448","text":"\r\nfrom collections import deque\r\nimport sys\r\n\r\ninput = sys.stdin.readline\r\nn, m = map(int, input().split())\r\n\r\n# graph\r\ngraph = [list(input().rstrip()) for _ in range(n)]\r\n\r\n# visited 4차원 배열...?\r\nvisited = [[[[False] * m for _ in range(n)] for _ in range(m)] for _ in range(n)]\r\n\r\ndx = [-1, 1, 0, 0]\r\ndy = [0, 0, -1, 1]\r\n\r\nq = deque()\r\n\r\n# init() to find R, B index\r\ndef init():\r\n rx, ry, bx, by = 0, 0, 0, 0\r\n for i in range(n):\r\n for j in range(m):\r\n if graph[i][j] == \"R\":\r\n rx, ry = i, j\r\n\r\n elif graph[i][j] == \"B\":\r\n bx, by = i, j\r\n\r\n q.append((rx, ry, bx, by, 0))\r\n visited[rx][ry][bx][by] = True\r\n\r\n\r\n# move() loop while not # not O\r\ndef move(x, y, dx, dy, cnt):\r\n # while next idx isNotWall and current idx isNotHole\r\n while graph[x + dx][y + dy] != \"#\" and graph[x][y] != \"O\":\r\n x += dx\r\n y += dy\r\n cnt += 1\r\n return x, y, cnt\r\n\r\n\r\n# bfs() 4 dir search while depth < 10 / R != O\r\ndef bfs():\r\n\r\n while q:\r\n rx, ry, bx, by, depth = q.popleft()\r\n\r\n if depth >= 10:\r\n break\r\n\r\n for i in range(4):\r\n nrx, nry, rcnt = move(rx, ry, dx[i], dy[i], 0)\r\n nbx, nby, bcnt = move(bx, by, dx[i], dy[i], 0)\r\n\r\n # if B == \"O\": continue\r\n if graph[nbx][nby] == \"O\":\r\n continue\r\n\r\n if graph[nrx][nry] == \"O\":\r\n print(1)\r\n return\r\n\r\n if nrx == nbx and nry == nby:\r\n if rcnt > bcnt:\r\n nrx, nry = nrx - dx[i], nry - dy[i]\r\n else: # rcnt < bcnt:\r\n nbx, nby = nbx - dx[i], nby - dy[i]\r\n\r\n # visited check:\r\n if not visited[nrx][nry][nbx][nby]:\r\n visited[nrx][nry][nbx][nby] = True\r\n q.append((nrx, nry, nbx, nby, depth + 1))\r\n\r\n print(0)\r\n\r\n\r\ninit()\r\nbfs()","repo_name":"unboxing96/ALGO","sub_path":"백준/Gold/13459. 구슬 탈출/구슬 탈출.py","file_name":"구슬 탈출.py","file_ext":"py","file_size_in_byte":1921,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"13430206336","text":"\n# 微调多国语言版T5做Seq2Seq任务\n# 介绍链接:https://kexue.fm/archives/7867\n# 数据集:https://github.com/CLUEbenchmark/CLGE 中的CSL数据集\n# 补充了评测指标bleu、rouge-1、rouge-2、rouge-l\n# mt5主要特点:gated-gelu, decoder的最后的dense层独立权重,rmsnorm\n# valid_data: {'rouge-1': 0.43454686332522263, 'rouge-2': 0.3217250949304608, 'rouge-l': 0.42204007502153934, 'bleu': 0.16675070297852404, 'best_bleu': 0.16675070297852404}\n\nimport json\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom nlpcol.callback import Callback\nfrom nlpcol.config import TrainConfig, WordDir, device\nfrom nlpcol.model import build_transformer_model\nfrom nlpcol.tokenizers import SpTokenizer\nfrom nlpcol.trainer import Trainer\nfrom nlpcol.utils.snippets import (ListDataset, get_pool_emb,\n save_model_parameter, seed_everything,\n sequence_padding, torch_gc)\nfrom nlpcol.utils.snippets4examples import model_name_gene, trans_entity2tuple\nfrom nltk.translate.bleu_score import SmoothingFunction, sentence_bleu\nfrom rouge import Rouge # pip install rouge\nfrom tqdm import tqdm\n\n# 基本参数\nmax_c_len = 256\nmax_t_len = 32\nbatch_size = 16\nepochs = 50\nsteps_per_epoch = None\npad_token_id = -100\n\ntrain_config = TrainConfig(batch_size=batch_size, epochs=epochs, max_seq_length=128)\n\n\nmodel_path = \"/home/dataset/pretrain_ckpt/t5/mt5-base\"\nconfig_path = model_path + \"/config.json\"\ncheckpoint_path = model_path + '/pytorch_model.bin'\n# spm_path = model_path + '/spiece.model'\n# 下面两个config是从bert4keras中拿的,项目连接https://github.com/bojone/t5_in_bert4keras\nspm_path = model_path + '/sentencepiece_cn.model'\nkeep_tokens_path = model_path + '/sentencepiece_cn_keep_tokens.json'\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\nseed_everything(42)\n\n\ntokenizer = SpTokenizer(spm_path, token_start=None, token_end='')\nkeep_tokens = json.load(open(keep_tokens_path))\n\n\nmodel = build_transformer_model(\n checkpoint_path,\n config_path,\n model='t5',\n extra_config={\"max_seq_length\": 512},\n keep_tokens=keep_tokens, # 只保留keep_tokens中的字,精简原字表\n skip_init=True\n).to(device)\n\n\ndef generate(text):\n input_ids, _ = tokenizer.encode(text)\n input_ids = torch.tensor([input_ids], device=device)\n\n # logits = model.generate(input_ids, mode='do_sample', top_k=20, top_p=0.9, temperature=0.9)\n logits = model.generate(input_ids, mode='greedy_search')\n # logits = model.generate(input_ids, mode='beam_search', num_beams=4)\n\n logits=logits[:,1:] # 去掉bos\n predict_label = [tokenizer.decode(i) for i in logits]\n return predict_label[0]\n\n\n# 准备数据\nclass MyDataset(ListDataset):\n @staticmethod\n def load_data(filename):\n \"\"\"加载数据\n 单条格式:(标题, 正文)\n \"\"\"\n D = []\n with open(filename, encoding='utf-8') as f:\n for l in f:\n l = json.loads(l)\n title, content = l['title'], l['abst']\n D.append((title, content))\n return D\n\ndef collate_fn(batch):\n \"\"\"TODO padding 转入 tokenizer中\n \"\"\"\n batch_content_ids, batch_titile_ids = [], []\n for title, content in batch:\n content_ids, _ = tokenizer.encode(content, maxlen=max_c_len)\n batch_content_ids.append(content_ids)\n\n titile_ids, _ = tokenizer.encode(title, maxlen=max_t_len)\n batch_titile_ids.append(titile_ids)\n\n batch_content_ids = torch.tensor(sequence_padding(batch_content_ids), dtype=torch.long, device=device)\n batch_titile_ids = torch.tensor(sequence_padding(batch_titile_ids, value=-100), dtype=torch.long, device=device)\n \n return {\n \"input_ids\": batch_content_ids,\n \"labels\": batch_titile_ids,\n }\n \ntrain_dataset = MyDataset('/home/dataset/corpus/seq2seq/summary/csl_title_public/csl_title_train.json')\nvalid_dataset = MyDataset('/home/dataset/corpus/seq2seq/summary/csl_title_public/csl_title_dev.json')\n\n\n# 定义训练流程\noptimizer = optim.Adam(model.parameters(), 1e-4)\nsave_path = model_name_gene(train_config, 'mt5', 'csl_title_public', prefix='test')\n\n\n# loss_fn = CrossEntropyLoss()\ntrainer = Trainer(model, train_config, optimizer=optimizer, collate_fn=collate_fn)\n\n\nclass Evaluator(Callback):\n \"\"\"评估与保存\n \"\"\"\n def __init__(self, trainer:Trainer, save_path:str):\n self.rouge = Rouge()\n self.smooth = SmoothingFunction().method1\n self.best_bleu = 0.\n self.trainer = trainer\n self.save_path = save_path\n\n def on_epoch_end(self, steps, epoch, logs=None):\n just_show()\n metrics = self.evaluate() # 评测模型\n if metrics['bleu'] > self.best_bleu:\n self.best_bleu = metrics['bleu']\n self.trainer.save_weights(self.save_path)\n\n metrics['best_bleu'] = self.best_bleu\n print('valid_data:', metrics)\n \n def evaluate(self):\n total = 0\n rouge_1, rouge_2, rouge_l, bleu = 0, 0, 0, 0\n for title, content in tqdm(valid_dataset):\n total += 1\n title = ' '.join(title).lower()\n pred_title = ' '.join(generate(content)).lower()\n if pred_title.strip():\n scores = self.rouge.get_scores(hyps=pred_title, refs=title)\n rouge_1 += scores[0]['rouge-1']['f']\n rouge_2 += scores[0]['rouge-2']['f']\n rouge_l += scores[0]['rouge-l']['f']\n bleu += sentence_bleu(references=[title.split(' ')], hypothesis=pred_title.split(' '),\n smoothing_function=self.smooth)\n rouge_1, rouge_2, rouge_l, bleu = rouge_1/total, rouge_2/total, rouge_l/total, bleu/total\n return {'rouge-1': rouge_1, 'rouge-2': rouge_2, 'rouge-l': rouge_l, 'bleu': bleu}\n\n\ndef just_show():\n s1 = u'抽象了一种基于中心的战术应用场景与业务,并将网络编码技术应用于此类场景的实时数据多播业务中。在分析基于中心网络与Many-to-all业务模式特性的基础上,提出了仅在中心节点进行编码操作的传输策略以及相应的贪心算法。分析了网络编码多播策略的理论增益上界,仿真试验表明该贪心算法能够获得与理论相近的性能增益。最后的分析与仿真试验表明,在这种有中心网络的实时数据多播应用中,所提出的多播策略的实时性能要明显优于传统传输策略。'\n s2 = u'普适计算环境中未知移动节点的位置信息是定位服务要解决的关键技术。在普适计算二维空间定位过程中,通过对三角形定位单元区域的误差分析,提出了定位单元布局(LUD)定理。在此基础上,对多个定位单元布局进行了研究,定义了一个新的描述定位单元中定位参考点覆盖效能的物理量——覆盖基,提出了在误差最小情况下定位单元布局的覆盖基定理。仿真实验表明定位单元布局定理能更好地满足对普适终端实时定位的需求,且具有较高的精度和最大覆盖效能。'\n for s in [s1, s2]:\n print(u'生成标题:', generate(s))\n\n\nif __name__ == '__main__':\n evaluator = Evaluator(trainer, save_path)\n print(u'生成标题:', generate(u'���国的首都是extra0京')) # 和huggingface的结果一致 \n # '北京 北京 北京 首都'\n\n trainer.train(train_dataset, callbacks=[evaluator])\n\n# else:\n# trainer.load_weights(save_path)\n","repo_name":"zouweidong91/nlpcol","sub_path":"examples/seq2seq/autotitle_csl_mt5.py","file_name":"autotitle_csl_mt5.py","file_ext":"py","file_size_in_byte":7543,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"15199239170","text":"# -*- encoding: UTF-8 -*-\n\nimport logging\nfrom strategy import turtle_trade\n\n\n# “停机坪”策略\ndef check(code_name, data, end_date=None, threshold=15):\n origin_data = data\n\n if end_date is not None:\n mask = (data['日期'] <= end_date)\n data = data.loc[mask]\n\n if len(data) < threshold:\n logging.debug(\"{0}:样本小于{1}天...\\n\".format(code_name, threshold))\n return\n\n data = data.tail(n=threshold)\n\n flag = False\n\n # 找出涨停日\n for index, row in data.iterrows():\n try:\n if float(row['p_change']) > 9.5:\n if turtle_trade.check_enter(code_name, origin_data, row['日期'], threshold):\n if check_internal(code_name, data, row):\n flag = True\n except KeyError as error:\n logging.debug(\"{}处理异常:{}\".format(code_name, error))\n\n return flag\n\n\ndef check_internal(code_name, data, limitup_row):\n limitup_price = limitup_row['收盘']\n limitup_end = data.loc[(data['日期'] > limitup_row['日期'])]\n limitup_end = limitup_end.head(n=3)\n if len(limitup_end.index) < 3:\n return False\n\n consolidation_day1 = limitup_end.iloc[0]\n consolidation_day23 = limitup_end = limitup_end.tail(n=2)\n\n if not(consolidation_day1['收盘'] > limitup_price and consolidation_day1['开盘'] > limitup_price and\n 0.97 < consolidation_day1['收盘'] / consolidation_day1['开盘'] < 1.03):\n return False\n\n threshold_price = limitup_end.iloc[-1]['收盘']\n\n for index, row in consolidation_day23.iterrows():\n try:\n if not (0.97 < (row['收盘'] / row['开盘']) < 1.03 and -5 < row['p_change'] < 5\n and row['收盘'] > limitup_price and row['开盘'] > limitup_price):\n return False\n except KeyError as error:\n logging.debug(\"{}处理异常:{}\".format(code_name, error))\n\n logging.debug(\"股票{0} 涨停日期:{1}\".format(code_name, limitup_row['日期']))\n\n return True\n\n","repo_name":"sngyai/Sequoia","sub_path":"strategy/parking_apron.py","file_name":"parking_apron.py","file_ext":"py","file_size_in_byte":2040,"program_lang":"python","lang":"en","doc_type":"code","stars":1505,"dataset":"github-code","pt":"78"}
+{"seq_id":"42362547000","text":"try:\n from pip import main as pipmain\nexcept:\n from pip._internal import main as pipmain\nimport os\n\n\nfor dirpath, dirnames, filenames in os.walk(\".\"):\n for filename in filenames:\n if filename==\"numpy-1.14.5+mkl-cp27-cp27m-win32.whl\":\n fullFileName = (os.path.realpath(os.path.join(dirpath, filename)))\npipmain(['install','wheel'])\n# pipmain(['uninstall','-y','numpy'])\n# pipmain(['uninstall','-y','pandas'])\npipmain(['install',\"pandas\"])\n# pipmain(['uninstall','-y','pygeoprocessing'])\npipmain(['install','pygeoprocessing==0.3.3'])\npipmain(['install','Image'])\n\npipmain(['install',fullFileName])\n","repo_name":"priyanshukumar0309/DHM-WM","sub_path":"libraries.py","file_name":"libraries.py","file_ext":"py","file_size_in_byte":623,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"35449747471","text":"import json\nfrom aiogram import Bot, Dispatcher, executor, types\nfrom aiogram.types.web_app_info import WebAppInfo\nimport config\n\nbot = Bot(config.BOT_TOKEN)\ndp = Dispatcher(bot)\n\n\n@dp.message_handler(commands=[\"start\"])\nasync def start(message: types.Message):\n markup = types.ReplyKeyboardMarkup()\n markup.add(types.KeyboardButton(\"Open web\", web_app=WebAppInfo(url=\"https://itproger.com/telegram.html\")))\n await message.answer(\"Hello!\", reply_markup=markup)\n\n\n@dp.message_handler(content_types=[\"web_app_data\"])\nasync def web_app(message: types.Message):\n res = json.loads(message.web_app_data.data)\n await message.answer(f\"Name: {res['name']}. Email: {res['email']}. Phone: {res['phone']}\")\n\n\n@dp.message_handler(commands=[\"payment\"])\nasync def payment(message: types.Message):\n await bot.send_invoice(message.chat.id, \"Buy product\", \"Product\", \"invoice\", config.PAYMENT_TOKEN, \"USD\",\n [types.LabeledPrice(\"Buy product\", 5 * 100)])\n\n\n@dp.message_handler(content_types=types.ContentType.SUCCESSFUL_PAYMENT)\nasync def success(message: types.Message):\n await message.answer(f\"Success: {message.successful_payment.order_info}\")\n\nexecutor.start_polling(dp)\n","repo_name":"AlisonSav/Aiogram-web-app-Bot","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1203,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"73888001852","text":"class Errors:\n\n @staticmethod\n def getMessage(code):\n for attribute in Errors.__dict__:\n if attribute.islower():\n continue\n if Errors.__dict__[attribute] == code:\n return attribute.replace('_', ' ').capitalize()\n return ''\n\n SUCCESS = 0\n SEVER_ERROR = 1000\n","repo_name":"itang85/main","sub_path":"app/common/base/errors.py","file_name":"errors.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"1053338736","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\nimport os\n\nimport torch\nfrom src import logger\nfrom src.baseline.modeling import ModelingCtcBert\nfrom src.baseline.tokenizer import CtcTokenizer\n\n\nclass PredictorCtc:\n def __init__(\n self,\n in_model_dir,\n ctc_label_vocab_dir='src/baseline/ctc_vocab',\n use_cuda=True,\n cuda_id=None,\n ):\n\n self.in_model_dir = in_model_dir\n self.model = ModelingCtcBert.from_pretrained(\n in_model_dir)\n self._id2dtag, self._dtag2id, self._id2ctag, self._ctag2id = self.load_label_dict(\n ctc_label_vocab_dir)\n logger.info('model loaded from dir {}'.format(\n self.in_model_dir))\n self.use_cuda = use_cuda\n if self.use_cuda and torch.cuda.is_available():\n if cuda_id is not None:\n torch.cuda.set_device(cuda_id)\n self.model.cuda()\n self.model.half()\n self.model.eval()\n self.tokenizer = CtcTokenizer.from_pretrained(in_model_dir)\n\n try:\n self._start_vocab_id = self.tokenizer.vocab['[START]']\n except KeyError:\n self._start_vocab_id = self.tokenizer.vocab['[unused1]']\n\n def load_label_dict(self, ctc_label_vocab_dir):\n dtag_fp = os.path.join(ctc_label_vocab_dir, 'ctc_detect_tags.txt')\n ctag_fp = os.path.join(ctc_label_vocab_dir, 'ctc_correct_tags.txt')\n\n id2dtag = [line.strip() for line in open(dtag_fp, encoding='utf8')]\n d_tag2id = {v: i for i, v in enumerate(id2dtag)}\n\n id2ctag = [line.strip() for line in open(ctag_fp, encoding='utf8')]\n c_tag2id = {v: i for i, v in enumerate(id2ctag)}\n logger.info('d_tag num: {}, d_tags:{}'.format(len(id2dtag), d_tag2id))\n return id2dtag, d_tag2id, id2ctag, c_tag2id\n\n def id_list2ctag_list(self, id_list) -> list:\n\n return [self._id2ctag[i] for i in id_list]\n\n @torch.no_grad()\n def predict(self, texts, return_topk=1, batch_size=32):\n if isinstance(texts, str):\n texts = [texts]\n else:\n texts = texts\n outputs = []\n for start_idx in range(0, len(texts), batch_size):\n batch_texts = texts[start_idx:start_idx+batch_size]\n\n batch_texts = [' ' + t for t in batch_texts] # 开头加一个占位符\n inputs = self.tokenizer(batch_texts,\n return_tensors='pt')\n # 把 ' ' 换成 _start_vocab_id\n inputs['input_ids'][..., 1] = self._start_vocab_id\n if self.use_cuda and torch.cuda.is_available():\n inputs['input_ids'] = inputs['input_ids'].cuda()\n inputs['attention_mask'] = inputs['attention_mask'].cuda()\n inputs['token_type_ids'] = inputs['token_type_ids'].cuda()\n\n d_preds, preds, loss = self.model(\n input_ids=inputs['input_ids'],\n attention_mask=inputs['attention_mask'],\n token_type_ids=inputs['token_type_ids'],\n )\n\n preds = torch.softmax(preds[:, 1:, :], dim=-1) # 从cls后面开始\n recall_top_k_probs, recall_top_k_ids = preds.topk(\n k=return_topk, dim=-1, largest=True, sorted=True)\n recall_top_k_probs = recall_top_k_probs.tolist()\n recall_top_k_ids = recall_top_k_ids.tolist()\n recall_top_k_chars = [[self.id_list2ctag_list(\n char_level) for char_level in sent_level] for sent_level in recall_top_k_ids]\n batch_texts = [['']+list(t)[1:] for t in batch_texts] # 占位符\n batch_outputs = [list(zip(text, top_k_char, top_k_prob)) for text, top_k_char, top_k_prob in zip(\n batch_texts, recall_top_k_chars, recall_top_k_probs)]\n outputs.extend(batch_outputs)\n return outputs\n\n @staticmethod\n def output2text(output):\n\n pred_text = ''\n for src_token, pred_token_list, pred_prob_list in output:\n pred_token = pred_token_list[0]\n if '$KEEP' in pred_token:\n pred_text += src_token\n elif '$DELETE' in pred_token:\n continue\n elif '$REPLACE' in pred_token:\n pred_text += pred_token.split('_')[-1]\n elif '$APPEND' in pred_token:\n pred_text += src_token+pred_token.split('_')[-1]\n\n return pred_text","repo_name":"bitallin/MiduCTC-competition","sub_path":"src/baseline/predictor.py","file_name":"predictor.py","file_ext":"py","file_size_in_byte":4404,"program_lang":"python","lang":"en","doc_type":"code","stars":53,"dataset":"github-code","pt":"78"}
+{"seq_id":"27708390021","text":"import twisted.internet.asyncioreactor\r\ntwisted.internet.asyncioreactor.install()\r\nfrom twisted.internet import reactor, task\r\nimport ipywidgets, datetime, subprocess, functools, os\r\n\r\nclass DoneError(Exception):\r\n pass\r\n\r\ndef time_out_counter(reactor):\r\n label = ipywidgets.Label(\"Time left: 5:00\")\r\n current_seconds = datetime.timedelta(minutes=5).total_seconds()\r\n def decrement(count):\r\n nonlocal current_seconds\r\n current_seconds -= count\r\n time_left = datetime.timedelta(seconds=max(current_seconds, 0))\r\n minutes, left = divmod(time_left, minute)\r\n seconds = int(left.total_seconds())\r\n label.value = f\"Time left: {minutes}:{seconds:02}\"\r\n if current_seconds < 0:\r\n raise DoneError(\"finished\")\r\n minute = datetime.timedelta(minutes=1)\r\n call = task.LoopingCall.withCount(decrement)\r\n call.reactor = reactor\r\n d = call.start(1)\r\n d.addErrback(lambda f: f.trap(DoneError))\r\n return d, label\r\n\r\ndef editor(fname):\r\n textarea = ipywidgets.Textarea(continuous_update=False)\r\n textarea.rows = 20\r\n output = ipywidgets.Output()\r\n runner = functools.partial(subprocess.run, capture_output=True, text=True, check=True)\r\n def save(_ignored):\r\n with output:\r\n with open(fname, \"w\") as fpout:\r\n fpout.write(textarea.value)\r\n print(\"Sending...\", end='')\r\n try:\r\n runner([\"git\", \"add\", fname])\r\n runner([\"git\", \"commit\", \"-m\", f\"updated {fname}\"])\r\n runner([\"git\", \"push\"])\r\n except subprocess.CalledProcessError as exc:\r\n print(\"Could not send\")\r\n print(exc.stdout)\r\n print(exc.stderr)\r\n else:\r\n print(\"Done\")\r\n textarea.observe(save, names=\"value\")\r\n return textarea, output, save\r\n\r\ndef journal():\r\n date = str(datetime.date.today())\r\n title = f\"Log: Startdate {date}\"\r\n filename = os.path.join(f\"{date}.txt\")\r\n d, clock = time_out_counter(reactor)\r\n textarea, output, save = editor(filename)\r\n box = ipywidgets.VBox([\r\n ipywidgets.Label(title),\r\n textarea,\r\n clock,\r\n output\r\n ])\r\n d.addCallback(save)\r\n return box\r\n\r\njournal()","repo_name":"nsanka/scripts","sub_path":"python/journal.py","file_name":"journal.py","file_ext":"py","file_size_in_byte":2273,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"72715120893","text":"# 오랜만에 풀어봤는데 도저히 모르겠어서 검색했음\n# 수학을 좀 알아야 할 듯함\nimport math\n\nt = int(input())\ns = []\na = []\ngcd = 0\nfor i in range(t):\n s.append(int(input()))\n if i == 1:\n gcd = abs(s[1] - s[0])\n gcd = math.gcd(abs(s[i] - s[i - 1]), gcd)\ngcd_a = int(gcd ** 0.5)\nfor i in range(2, gcd_a + 1):\n if gcd % i == 0:\n a.append(i)\n a.append(gcd // i)\na.append(gcd)\na = list(set(a))\na.sort()\nfor i in a:\n print(i, end = ' ')","repo_name":"hhh57463/algorithm","sub_path":"code/2981.py","file_name":"2981.py","file_ext":"py","file_size_in_byte":494,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39197877734","text":"from invoke import Program, Collection\nimport sys\nimport core\nimport plugins\nimport pkgutil\n\n\ndef unhandled_exception(type, value, traceback):\n \"\"\"Override for `sys.excepthook` without stack trace\"\"\"\n print(f\"{type.__name__}: {str(value)}\")\n\n\nclass VastCloudProgram(Program):\n \"\"\"A custom Program that doesn't print useless core options\"\"\"\n\n def print_help(self) -> None:\n print(\n f\"\"\"Usage: {self.binary} [--core-opts] [--subcommand-opts] ...\n\nCore options:\n\n -e, --echo Echo executed commands before running.\n -h [STRING], --help[=STRING] Show core or per-task help and exit.\n -V, --version Show version and exit. \n\"\"\"\n )\n self.list_tasks()\n\n\nif __name__ == \"__main__\":\n sys.excepthook = unhandled_exception\n\n namespace = Collection.from_module(core)\n\n # import all the modules in the plugins folder as collections\n for importer, modname, ispkg in pkgutil.iter_modules(plugins.__path__):\n mod = importer.find_module(modname).load_module(modname)\n namespace.add_collection(Collection.from_module(mod))\n\n program = VastCloudProgram(\n binary=\"./vast-cloud\",\n namespace=namespace,\n version=\"0.1.0\",\n )\n program.run()\n","repo_name":"25280841/vast","sub_path":"cloud/aws/cli/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1275,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"3716032670","text":"import random\n\nTRAITS = [\n \"shy\",\n \"hesitant\",\n \"embarrassed\",\n \"confused\",\n \"sad\",\n \"scared\",\n \"nervous\",\n \"happy\",\n \"starstruck\",\n \"excited\",\n \"crazy\",\n]\n\n\ndef random_trait() -> str:\n \"\"\"Randomly sample a character trait for the guest of The Expert Zone.\"\"\"\n return random.choice(TRAITS)\n","repo_name":"camille-vanhoffelen/wet-toast-talk-radio","sub_path":"wet_toast_talk_radio/scriptwriter/prolove/traits.py","file_name":"traits.py","file_ext":"py","file_size_in_byte":329,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"78"}
+{"seq_id":"74317014973","text":"import collections\nimport json\nimport pandas as pd\nimport sys\n\nimport valentine as va\nimport valentine.algorithms \nimport valentine.metrics as valentine_metrics\nfrom collections import defaultdict\nfrom scipy.optimize import linprog\nfrom gurobipy import *\n\n\ndef json_to_dataframe(json_file_path):\n \"\"\"\n Purpose: Normalize the JSON file\n @param: json_file_path: Path of JSON file\n @returns: dataframe: Flat table \n \"\"\"\n df_list = []\n with open(json_file_path) as json_file:\n for line in json_file:\n json_data = json.loads(line)\n df = pd.json_normalize(json_data)\n df_list.append(df)\n return pd.concat(df_list).reset_index(drop = True)\n\n\ndef find_valentine(df1, df2, ground_truth):\n \"\"\"\n Purpose: Match source keys to target keys\n @param: df1: Target dataframe, df2: Source dataframe\n @returns: match_dict: dictionary of matching keys\n \"\"\"\n matcher = valentine.algorithms.SimilarityFlooding()\n matches = va.valentine_match(df1, df2, matcher, \"target\", \"source\")\n metrics = valentine_metrics.all_metrics(matches, ground_truth)\n match_dict = defaultdict(list)\n for match in ground_truth:\n match_dict[match[0]].append(match[1])\n return match_dict\n\n\ndef linear_programming(match_dict):\n problem = LpProblem(name = \"Match_columns\", sense = LpMaximize)\n nested_pairs = {} \n all_variables = []\n\n for t_key, s_keys in match_dict.items():\n target_prefix = \"\"\n # Get the prefix of the target key if it has one\n if '.' not in t_key:\n target_prefix = '_tROOT_'\n else:\n target_prefix, _ = t_key.rsplit('.', maxsplit = 1)\n target_prefix = '_tROOT_' + target_prefix\n \n variables = []\n nested_vars = {}\n \n # Loop over source keys that match with target key\n for s_key in s_keys:\n # Create binary variable for matching source key\n var_name = 'T_' + t_key + '-----' + 'S_' + s_key\n s_x = LpVariable(name = var_name, lowBound = 0, upBound = 1, cat = \"Binary\")\n all_variables.append(s_x)\n variables.append(s_x)\n \n # Get the prefix of the source key if it has one. Create a.b_O.d, ...\n if '.' not in s_key:\n source_prefix = '_sROOT_'\n else:\n source_prefix, _ = s_key.rsplit('.', maxsplit = 1)\n source_prefix = '_sROOT_' + source_prefix\n \n # Get the nested key\n nested_key = 'T_' + t_key + '_O_S' + source_prefix\n # Create a variable for each nested key\n if nested_key not in nested_vars:\n nested_vars[nested_key] = LpVariable(name = nested_key, lowBound = 0, upBound = 1, cat = \"Binary\")\n nested_var = nested_vars[nested_key]\n #print(\"nested\", nested_var)\n \n # Constraint a.O_d >= a.b_d.e (and a.O_d >= a.b_d.f, ...)\n problem += (nested_var >= s_x)\n\n # Create a variable for pair key. P.a_O.d (and P.a_O, ...)\n pair_key = 'T' + target_prefix + 'S' + source_prefix\n if pair_key not in nested_pairs:\n nested_pairs[pair_key] = LpVariable(name = pair_key, lowBound = 0, upBound = 1, cat = \"Binary\")\n pair_var = nested_pairs[pair_key]\n #print(\"pair\", pair_var)\n # Constraint P.a_O.d <= a.b_O.d (and P.a_O.d >= a.c_O.d, ...)\n problem += pair_var >= nested_var\n\n # At most one matching for each path in the source schema\n problem += lpSum(variables) <= 1\n\n # Objective function: P.a_O.d + P.a_O + ...\n # problem += lpSum(list(nested_pairs.values()))\n problem += lpSum(all_variables)\n LpSolverDefault.msg = 1 \n # Solve the objective function\n status = problem.solve(PULP_CBC_CMD(msg = 1))\n print(\"Status:\", LpStatus[status])\n\n # Loop over problem variables and print their optimum values\n for variable in problem.variables():\n print(variable.name, \"=\", variable.varValue)\n \n\ndef get_key_prefix(key):\n \"\"\"\n Purpose: Get the prefix of a key\n @param: key: target or source key\n @returns: prefix\n \"\"\"\n if '.' not in key:\n prefix = 'ROOT_'\n else:\n prefix, _ = key.rsplit('.', maxsplit = 1)\n prefix = 'ROOT_' + prefix\n return prefix\n\n\ndef show_matches(source_vars, match_dict):\n '''\n Purpose: Store final matches in dictionary\n @param: source_vars: variables of source keys, match_dict: pairs of keys that possibly match\n @returns: dictionary of keys that definitely match\n '''\n final_matching = {}\n for s_keys in match_dict.values():\n for s_key in s_keys:\n for (s_var, t_key) in source_vars[s_key]:\n if bool(round(s_var.X)):\n final_matching[s_key] = t_key\n break\n return final_matching\n\n\ndef quadratic_programming(match_dict):\n \"\"\"\n Purpose: Create a quadratic model to choose the best match\n @param: match_dict: Dictionary of keys that can possibly match\n @returns: final_match: Dictionary of keys that match\n \"\"\"\n #try:\n\n # Initialize the quadratic model\n quadratic_model = Model(\"quadratic\")\n\n # Dictionary to store the binary variables for nested target keys\n nested_t_vars = {}\n\n # Dictionary to store the binary variables for root target and source keys\n root_vars = {} \n\n # Dictionary to store source variables\n source_vars = collections.defaultdict(list)\n\n # List to store all the quadratic variables\n all_vars = []\n\n # Loop over matching target and source keys\n for t_key, s_keys in match_dict.items():\n target_prefix = 't' + get_key_prefix(t_key)\n\n # List to store the binary variables for nested source keys\n nested_s_vars = {}\n \n # Loop over source keys\n for s_key in s_keys:\n # Create a binary variable for the match between target and source keys\n nested_var_name = 'T_' + t_key + '-----' + 'S_' + s_key\n s_var = quadratic_model.addVar(name=nested_var_name, vtype=GRB.BINARY) \n source_vars[s_key].append((s_var, t_key))\n all_vars.append(s_var)\n \n # Create a binary variable for \"nested\" target keys\n nested_t_key = 'T_' + t_key + '-----' + 'S_' + 's' + get_key_prefix(s_key)\n if nested_t_key not in nested_t_vars:\n nested_t_vars[nested_t_key] = quadratic_model.addVar(name=nested_t_key, vtype=GRB.BINARY)\n nested_t_var = nested_t_vars[nested_t_key]\n \n # Create a binary variable for \"root\" target keys\n root_var_name = 'T_' + target_prefix + '-----' + 'S_' + 's' + get_key_prefix(s_key)\n if root_var_name not in root_vars:\n root_vars[root_var_name] = quadratic_model.addVar(name=root_var_name, vtype=GRB.BINARY)\n root_var = root_vars[root_var_name]\n \n # Create a binary variable for \"nested\" source keys\n nested_s_key = 'T_' + target_prefix + '-----' + 'S_' + s_key\n if nested_s_key not in nested_s_vars:\n nested_s_vars[nested_s_key] = quadratic_model.addVar(name=nested_s_key, vtype=GRB.BINARY)\n nested_s_var = nested_s_vars[nested_s_key]\n\n # Add constraints\n quadratic_model.addConstr(nested_t_var - root_var <= 0, name=\"Matching nested target implies matching root target.\")\n quadratic_model.addConstr(nested_s_var - root_var <= 0, name=\"Matching nested source implies matching root source.\")\n \n # Call update if you need to examine the model because optimization\n quadratic_model.update()\n \n # Loop over the source keys\n for s_key in s_keys:\n # Sum source variables constraints, It must be <= 1\n total = sum(s[0] for s in source_vars[s_key])\n #print('TOTAL:', total)\n # Add constraint: \n quadratic_model.addConstr(total <= 1, name=\"Each source key can be matched to at most one target key.\")\n\n\n # Objective function\n quadratic_model.setObjective(sum(all_vars), GRB.MAXIMIZE)\n quadratic_model.setParam(\"OutputFlag\", False)\n quadratic_model.optimize()\n if quadratic_model.getAttr(GRB.Attr.Status) == GRB.INFEASIBLE:\n quadratic_model.computeIIS()\n quadratic_model.write('iis.ilp')\n quadratic_model.write('model.lp')\n\n return show_matches(source_vars, match_dict)\n \n #print(f\"Optimal objective value: {quadratic_model.objVal}\")\n # Loop over model variables and print their rounded optimum values\n ''' \n for variable in quadratic_model.getVars():\n print(variable.varName, \"=\", bool(round(variable.x)))\n print()\n '''\n\n\n\n\n","repo_name":"jrn1325/Schema_Matching","sub_path":"csm.py","file_name":"csm.py","file_ext":"py","file_size_in_byte":8785,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"8260488951","text":"'''for i in range(10):\r\n s=input('请输入你的用户名(只能是字母)\\n')\r\n if s.isalpha():\r\n print('格式正确!')\r\n else:\r\n print('格式错误!')'''\r\n\r\ndef coco(n): #函数\r\n if n==1:\r\n return 1\r\n else:\r\n return n*coco(n-1)\r\n\r\nprint(coco(6))\r\nprint('----------------------------------')\r\n\r\nfor i in range(1,7):\r\n print(coco(i))","repo_name":"Anghosts/lovely","sub_path":"学习代码/Test_01/demo3.py","file_name":"demo3.py","file_ext":"py","file_size_in_byte":387,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"72190094332","text":"import matplotlib.pyplot as plt\n\nname = input(\"Type the name of the loan here: \")\nloan_size = float(input(\"Type the size of the loan here: \"))\nmonthly_payment = float(input(\"Type the amount you would like to pay every month: \"))\nyearly_interrest = float(input(\"Type in the yearly interrest here: \"))\n\ntotal = loan_size\ntotal_paid = 0\n\nloan_arr = []\nyear_arr = []\n\nyear = 0\nprint(name + \": \" + str(round(loan_size)))\nyear_arr.append(year)\nloan_arr.append(loan_size)\nwhile loan_size > 0:\n # Payment\n payment = monthly_payment * 12\n loan_size = loan_size - payment\n\n # Interrest rate\n interrest = loan_size * (yearly_interrest / 100)\n loan_size = loan_size + interrest\n\n # Total\n total_paid += payment\n if loan_size < 0:\n total_paid += loan_size\n\n # Year\n year += 1\n year_arr.append(year)\n loan_arr.append(loan_size)\n print(\"Year: \" + str(year))\n print(\"Remaining loan: \" + str(round(loan_size)))\n print(\"Total paid: \" + str(round(total_paid)))\n\n# Generate graph\nx=year_arr\ny=loan_arr\nplt.plot(x,y)\nplt.xlabel('Year')\nplt.ylabel('Remaning loan')\nplt.title(name)\nplt.show()\n","repo_name":"toko17/loan_calculator","sub_path":"loan.py","file_name":"loan.py","file_ext":"py","file_size_in_byte":1125,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"30242924228","text":"\nfrom robot import Robot\nimport time\n\n__author__ = 'stefan'\n\nrobot = Robot()\n\ntry:\n\n while True:\n\n print(\"driving forward\")\n robot.drive_forward(80)\n\n while not robot.sees_obstacle():\n time.sleep(0.1)\n\n print(\"obstacle - turning left\")\n robot.turn_left(100)\n\n # if obstacle is close (2)\n while robot.sees_obstacle() == 2:\n # let it turn a while\n time.sleep(0.25)\n\nexcept KeyboardInterrupt:\n\n robot.stop()\n","repo_name":"wendlers/edubot-nodemcu-fw","sub_path":"remote/python/examples/obstacle_avoider.py","file_name":"obstacle_avoider.py","file_ext":"py","file_size_in_byte":493,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"8434255012","text":"import math\nfrom fpdf import FPDF\nimport pandas as pd\n\n# CONFIG\n\n# A4 = 297mm x 210mm\na4_width = 297\na4_height = 210\n\n# size in mm\nbadge_height = 100\nbadge_width = 75\n\nlogo_1_path = \"data/l1.png\"\nlogo_2_path = \"data/l2.png\"\nlogo_3_path = \"data/l3.png\"\n\nfont_path = \"data/Roboto-Regular.ttf\"\n\nevent_name = \"MY SUPER COOL EVENT NAME - 2023\"\n\ndef main():\n global badge_height, badge_width, event_name\n\n # badge_height = input_number(\"Enter badge height in mm (Default 100):\", 100)\n # badge_width = input_number(\"Enter badge width in mm (Default 75):\", 75)\n check_valid_size()\n\n # event_name = input(\"Enter event name:\")\n\n customers = pd.read_excel(r'data/names.xlsx')\n\n badges_to_print = len(customers[\"forename\"])\n\n pages_to_print = math.ceil(badges_to_print / int((a4_width-10) / badge_width))\n\n print(\"Pages to print:\", pages_to_print)\n\n badge_number = 0\n pdf_file = FPDF(orientation='L', unit='mm', format='A4')\n pdf_file.add_font('Roboto', '', font_path,\n uni=True) # - family, style, fname, uni(Unicode flag)\n pdf_file.set_display_mode(zoom=\"real\")\n pdf_file.set_font(\"Roboto\")\n\n for page in range(pages_to_print):\n print(\"Page: \", page + 1)\n\n pdf_file.add_page()\n draw_cutting_lines(pdf_file)\n boarder_lr = int(287 % badge_width / 2)\n\n for x in range(5 + boarder_lr, a4_width - boarder_lr - 5, badge_width):\n if badge_number >= badges_to_print:\n break\n pdf_file.rotate(0, a4_width / 2, a4_height / 2)\n draw_badge(pdf_file, x, customers, badge_number)\n pdf_file.rotate(180, a4_width / 2, a4_height / 2)\n draw_badge(pdf_file, a4_width - badge_width - x, customers, badge_number)\n badge_number += 1\n\n pdf_file = pdf_file.output(name=\"generated_badge.pdf\")\n\n\ndef input_number(message, default_value):\n while True:\n try:\n user_input = int(input(message))\n except ValueError:\n print(\"Used default.\", default_value)\n return default_value\n else:\n return user_input\n\n\ndef check_valid_size():\n if a4_height - 10 < badge_height * 2:\n raise ValueError\n if (a4_width - 10) < badge_width:\n raise ValueError\n\n\ndef draw_cutting_lines(pdf_file, draw_full=False):\n boarder_lr = int(287 % badge_width / 2)\n if draw_full:\n # draw horizontal lines\n for y in range(5, 210, badge_height):\n pdf_file.line(5, y, a4_width-5, y)\n # draw vertical lines\n for x in range(5 + boarder_lr, a4_width, badge_width):\n pdf_file.line(x, 5, x, a4_height-5)\n else:\n for y in range(5, 210, badge_height):\n pdf_file.line(0, y, boarder_lr, y)\n pdf_file.line(297 - boarder_lr - 5, y, 297, y)\n for x in range(5 + boarder_lr, 298 - boarder_lr, badge_width):\n pdf_file.line(x, 0, x, 5)\n pdf_file.line(x, 210 - 5, x, 210)\n\n\ndef draw_badge(pdf_file, x, customers, badge_number):\n # Read Data\n first_name = customers[\"forename\"][badge_number]\n last_name = customers[\"surname\"][badge_number]\n profession = str(customers[\"profession\"][badge_number])\n company = str(customers[\"company\"][badge_number])\n\n print(\"Badge: \", badge_number + 1, \" - \", first_name, last_name, \" - \", profession, \" - \", company)\n\n # Logo 1\n if logo_1_path is not None:\n pdf_file.image(logo_1_path, x=x + 20, y=5 + 2, w=37)\n\n # Event Name\n pdf_file.set_font_size(size=10)\n pdf_file.set_xy(x=x + 4, y=5 + 25)\n pdf_file.multi_cell(w=badge_width - 6, h=5, txt=event_name, align=\"L\")\n\n # First Name + Last Name\n pdf_file.set_font_size(size=22)\n pdf_file.text(x=x + 5, y=5 + 47, txt=first_name)\n pdf_file.text(x=x + 5, y=5 + 55, txt=last_name)\n\n # Profession + Company\n pdf_file.set_font_size(size=14)\n pdf_file.set_xy(x=x + 4, y=5 + 57)\n pdf_file.multi_cell(w=badge_width - 8, h=5, txt=company, align=\"L\")\n\n # Logo 2\n if logo_2_path is not None:\n pdf_file.set_font_size(size=10)\n pdf_file.text(x=x + 5, y=5 + 72, txt=\"Hosted by\")\n pdf_file.image(logo_2_path, x=x + 3, y=5 + 78, w=33)\n\n # Logo 3\n if logo_3_path is not None:\n pdf_file.image(logo_3_path, x=x + 42, y=5 + 75, w=27)\n\nif __name__ == '__main__':\n main()\n","repo_name":"clFaster/Badge-Generator","sub_path":"src/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4335,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"28261257653","text":"import csv\n\n\ndef read(path):\n print(\"Headings:\")\n with open(path) as file:\n reader = csv.reader(file)\n headings = next(reader)\n print(headings)\n print(\"Values:\")\n for line in reader:\n print(line)\n file.close()\n\n\ndef run():\n read(\"bots.csv\")\n\n\nif __name__ == \"__main__\":\n run()\n","repo_name":"boitzenburgeruk/com411","sub_path":"data/files/csv/read_csv.py","file_name":"read_csv.py","file_ext":"py","file_size_in_byte":338,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"6761871121","text":"import os\nimport numpy as np\nfrom tqdm import tqdm\nfrom PIL import Image\nfrom database.data_base import DataBase\n\n\nclass Cleaner(object):\n\n def __init__(self, db_name=\"db.sqlite3\", window=100, log_dir=\"./cleaner_log\"):\n self.db = DataBase(db_name=db_name)\n self.window = window\n self.log_dir = log_dir\n try:\n os.stat(log_dir)\n except:\n os.mkdir(log_dir)\n\n def remove_duplicates_from_dir(self, dir_name):\n \"\"\"\n Remove image duplicates from dir_name folder\n\n :param dir_name:\n :return:\n \"\"\"\n\n print(\"Removing duplicates from {0}\".format(dir_name))\n log_file_path = os.path.join(self.log_dir, \"{0}.log\".format(dir_name))\n\n img_dict = {}\n for f in tqdm(os.listdir(dir_name)):\n img_dict[f] = np.array(Image.open(os.path.join(dir_name, f)))\n\n for f in tqdm(os.listdir(dir_name)):\n n = int(f.split(\"_\")[0])\n\n for d in os.listdir(dir_name):\n m = int(d.split(\"_\")[0])\n diff = m - n\n if d != f and diff in range(self.window):\n if np.array_equal(img_dict[f], img_dict[d]):\n os.remove(os.path.join(dir_name, d))\n print(\"{0} removed\".format(os.path.join(dir_name, d)))\n\n with open(log_file_path, \"w+\") as log_file:\n log_file.write(d)\n\n def remove_duplicates_from_db(self, dir_name):\n \"\"\"\n Remove image duplicates from database\n (from all the tables)\n image names file constructed on dir_name folder\n\n :param dir_name:\n :return:\n \"\"\"\n\n log_file_path = os.path.join(self.log_dir, \"{0}.log\".format(dir_name))\n with open(log_file_path, \"r\") as log_file:\n img_names = log_file.readlines()\n\n table_names = self.db.get_table_names()\n\n for img_name in img_names:\n for table_name in table_names:\n self.db.delete_by_name(table_name=table_name, img_name=img_name)\n print(\"{0} removed from {1}\".format(img_name, table_name))\n\n","repo_name":"nataliaklochko/smartproject","sub_path":"database/cleaner.py","file_name":"cleaner.py","file_ext":"py","file_size_in_byte":2159,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"13652316949","text":"import tensorflow as tf\nfrom tensorflow import keras\nfrom keras import datasets,layers,Sequential,metrics\n\n(x,y),(x_test,y_test) = datasets.mnist.load_data()\nprint(x.shape,y.shape)\nbachsize = 128\ndb = tf.data.Dataset.from_tensor_slices((x,y))\ndb_test = tf.data.Dataset.from_tensor_slices((x_test,y_test))\ndef pro(x,y):\n x = tf.cast(x,tf.float32)/255.\n x = tf.reshape(x,[-1,28*28])\n x = tf.squeeze(x)\n y = tf.cast(y,tf.int32)\n y = tf.one_hot(y,depth=10)\n return x,y\ndb = db.map(pro).shuffle(1000).batch(bachsize)\ndb_test = db_test.map(pro).batch(bachsize)\ndb_iter = iter(db)\nsample = next(db_iter)\nsample = next(db_iter)\nsample = next(db_iter)\nsample = next(db_iter)\nsample = next(db_iter)\nsample = next(db_iter)\nsample = next(db_iter)\nprint(sample[0].shape,sample[1].shape)\nprint(sample[0])\n\nmodel = Sequential([\nlayers.Dense(256,activation=tf.nn.relu),\nlayers.Dense(128,activation=tf.nn.relu),\nlayers.Dense(64,activation=tf.nn.relu),\nlayers.Dense(10)\n])\nmodel.build(input_shape=[None,784])\n# 模型参数初始化\nmodel.compile(optimizer=keras.optimizers.Adam(learning_rate=1e-3),loss=tf.losses.CategoricalCrossentropy(from_logits=True),metrics=['accuracy'])\n# 模型训练\nmodel.fit(db,epochs=20,validation_data=db_test,validation_steps=1)\n# 模型测试\nmodel.evaluate(db_test)\n# 保存模型权重\nmodel.save_weights('./checkpaoint/myw')\ndel model\nmodel = Sequential([\nlayers.Dense(256,activation=tf.nn.relu),\nlayers.Dense(128,activation=tf.nn.relu),\nlayers.Dense(64,activation=tf.nn.relu),\nlayers.Dense(10)\n])\nmodel.build(input_shape=[None,784])\nmodel.compile(optimizer=keras.optimizers.Adam(learning_rate=1e-3),loss=tf.losses.CategoricalCrossentropy(from_logits=True),metrics=['accuracy'])\n# 加载模型\nmodel.load_weights('./checkpaoint/myw')\nacc = model.evaluate(db_test)\nmodel.save(\"./cc/yy\")","repo_name":"wangofcong/TensorFlow","sub_path":"cfep.py","file_name":"cfep.py","file_ext":"py","file_size_in_byte":1820,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17528975476","text":"import sys, time, requests\n\ndef loading_dots():\n\t\"\"\"\n\tPrints three loading dots.\n\t\"\"\"\n\n\tfor i in range(3):\n\t\tprint(\".\", end=\"\")\n\t\tsys.stdout.flush()\n\t\ttime.sleep(0.75)\n\n\tfor j in range(3):\n\t\tprint(\"\\b \", end = \"\\b\")\n\t\tsys.stdout.flush()\n\t\ttime.sleep(0.75)\n\n\ndef main():\n\t\"\"\"\n\tWaits for flask webserver to start, and prints loading dots in the meantime.\n\t\"\"\"\n\n\ti = 0\n\twhile True:\n\t\ttry:\n\t\t\trequests.get(\"http://127.0.0.1:5000/\")\n\t\texcept requests.exceptions.ConnectionError as e:\n\t\t\tif i == 0:\n\t\t\t\tprint(\"Waiting for flask webserver to start\", end = \"\")\n\t\t\t\ti = 1\n\t\t\tloading_dots()\n\t\telse:\n\t\t\tbreak\n\n\nif __name__ == \"__main__\":\n\ttry:\n\t\tmain()\n\texcept KeyboardInterrupt:\n\t\tsys.exit(1)","repo_name":"IBM/multicloud-incident-response-navigator","sub_path":"run/wait.py","file_name":"wait.py","file_ext":"py","file_size_in_byte":682,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"78"}
+{"seq_id":"18873152338","text":"x=5\nwhile(x<5):\n print(x)\n x+=1\n print(x)\ntotal=0\nnum=20\nwhile (num<=25):\n total=total+num\n num=num+1\nprint(\"Value of the total from the while loop\",total)","repo_name":"Vipulhere/Python-practice-Code","sub_path":"Module 3/8.1 while loop.py","file_name":"8.1 while loop.py","file_ext":"py","file_size_in_byte":170,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"78"}
+{"seq_id":"37012715416","text":"from cobra.io import write_sbml_model, read_sbml_model\n\nmodel_name = 'iAF1260_mutant_delta_f6p_c'\nmodel_path_sbml = f'./genome/Mutants/{model_name}.xml'\nmodel_path_yaml= f'./genome/{model_name}/'\n\ndef generate_mutant(path=model_path_sbml):\n\n model = read_sbml_model(path)\n\n for i in [i.id for i in model.metabolites.f6p_c.reactions]:\n reaction = model.reactions.get_by_id(i)\n model.remove_reactions([reaction])\n return model\n\ndef export_model(model ,path='./genome/Mutants/', filename=model_name):\n write_sbml_model(model, f'{path}{filename}.xml')\n\nif __name__ == '__main__':\n mutant = generate_mutant()\n export_model(mutant)","repo_name":"SergioPachonDotor/OptKnock-Threonine_Prod","sub_path":"Control_model_for_gapfind_gapfill.py","file_name":"Control_model_for_gapfind_gapfill.py","file_ext":"py","file_size_in_byte":657,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"37224603960","text":"import sys\nimport PConstant\nimport PConfig\nimport ElasticSearchClient\nimport ESDatabaseMetaStore\nimport logging\nimport json\n\nclass ElasticSearchDriver(object):\n\n def __init__(self, indexname, typename):\n self.es = ElasticSearchClient.ElasticSearchClient()\n self.esdb = ESDatabaseMetaStore.ESDatabaseMetaStore()\n self.indexname = indexname\n self.typename = typename\n self._flogger()\n\n def createtable(self, schemanamecount):\n request_body = self.esdb.get_request_body(schemanamecount)\n self._logger.info(\"Create Table ' %s '\", self.indexname)\n self.es.create_index(self.indexname, request_body) \n\n def pushbulkdata(self, bulk_data):\n self._logger.info(\"Collecting Data (%d)\",len(bulk_data))\n self.es.add_bulk_document(self.indexname, bulk_data)\n\n def pushdata(self, data_dict):\n self._logger.info(\"Pushing data ...\")\n self.es.add_document(self.indexname,self.typename, data_dict, data_dict.keys()) \n\n def searchresult(self):\n self._logger.info(\"Searching results ... index=%s type=%s\", self.indexname, self.typename)\n queryhandle = {}\n res = self.es.search_result(self.indexname, self.typename, queryhandle)\n resultlist = []\n for doc in res['hits']['hits']:\n resultmap = {}\n resultmap.setdefault('_id', doc['_id'])\n resultmap.setdefault('_source', doc['_source'])\n resultlist.append(resultmap)\n return resultlist\n\n\n def searchresult(self, query):\n self._logger.info(\"Searching results ... index=%s type=%s\", self.indexname, self.typename)\n self._logger.info(\"Query '%s'\", query)\n #queryhandle = { 'query' : json.loads(query) }\n #jquery = json.loads(query)\n #queryhandle = {\"query\": {\"match\": {\"body\": \"McGrath\"}}}\n queryhandle = {\"query\": {\"match\": query}}\n self._logger.info(\"Queryi Handle '%s'\", json.dumps(queryhandle))\n res = self.es.search_result(self.indexname, self.typename, queryhandle)\n resultlist = []\n for doc in res['hits']['hits']:\n resultmap = {}\n resultmap.setdefault('_id', doc['_id'])\n resultmap.setdefault('_source', doc['_source']) \n resultlist.append(resultmap)\n return resultlist\n\n def scanresult(self, query):\n\n #queryhandle = {\"query\": {\"match\": query}}\n queryhandle = query\n pagelist = self.es.scan_result_sync(self.indexname, self.typename, query)\n resultlist = []\n\n self._logger.info(\"Scan Handle '%s' '%s' '%s'\", self.indexname, self.typename, queryhandle)\n for page in pagelist:\n for doc in page:\n resultmap = {}\n resultmap.setdefault('_id', doc['_id'])\n resultmap.setdefault('_source', doc['_source'])\n resultlist.append(resultmap)\n return resultlist\n\n\n def _flogger(self):\n\n self._logger = logging.getLogger('ElasticSearchDriver')\n self._logger.setLevel(logging.INFO)\n formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n ch = logging.StreamHandler()\n ch.setLevel(logging.INFO)\n ch.setFormatter(formatter)\n self._logger.addHandler(ch)\n\n","repo_name":"somilasthana/esearch","sub_path":"ElasticSearchDriver.py","file_name":"ElasticSearchDriver.py","file_ext":"py","file_size_in_byte":3287,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"18596103789","text":"from numbers import Integral\nfrom typing import List, Union, Dict, Any\n\n\nclass Parser:\n \"\"\" Parser base class.\n\n The attributes listed below make up a public interface common to all parsers. They can be accessed directly\n once the dataset is constructed and annotations are populated.\n\n Attributes:\n\n cat_names (list[str]):\n list of category (class) names, with background class at position 0.\n cat_ids (list[union[str, int]):\n list of dataset specific, unique integer or string category ids, does not include background\n cat_id_to_label (dict):\n map from category id to integer 1-indexed class label\n\n img_ids (list):\n list of dataset specific, unique image ids corresponding to valid samples in dataset\n img_ids_invalid (list):\n list of image ids corresponding to invalid images, not used as samples\n img_infos (list[dict]):\n image info, list of info dicts with filename, width, height for each image sample\n \"\"\"\n def __init__(\n self,\n bbox_yxyx: bool = False,\n has_labels: bool = True,\n include_masks: bool = False,\n include_bboxes_ignore: bool = False,\n ignore_empty_gt: bool = False,\n min_img_size: int = 32,\n ):\n \"\"\"\n Args:\n yxyx (bool): output coords in yxyx format, otherwise xyxy\n has_labels (bool): dataset has labels (for training validation, False usually for test sets)\n include_masks (bool): include segmentation masks in target output (not supported yet for any dataset)\n include_bboxes_ignore (bool): include ignored bbox in target output\n ignore_empty_gt (bool): ignore images with no ground truth (no negative images)\n min_img_size (bool): ignore images with width or height smaller than this number\n sub_sample (int): sample every N images from the dataset\n \"\"\"\n # parser config, determines how dataset parsed and validated\n self.yxyx = bbox_yxyx\n self.has_labels = has_labels\n self.include_masks = include_masks\n self.include_bboxes_ignore = include_bboxes_ignore\n self.ignore_empty_gt = ignore_empty_gt\n self.min_img_size = min_img_size\n self.label_offset = 1\n\n # Category (class) metadata. Populated by _load_annotations()\n self.cat_names: List[str] = []\n self.cat_ids: List[Union[str, Integral]] = []\n self.cat_id_to_label: Dict[Union[str, Integral], Integral] = dict()\n\n # Image metadata. Populated by _load_annotations()\n self.img_ids: List[Union[str, Integral]] = []\n self.img_ids_invalid: List[Union[str, Integral]] = []\n self.img_infos: List[Dict[str, Any]] = []\n\n @property\n def cat_dicts(self):\n \"\"\"return category names and labels in format compatible with TF Models Evaluator\n list[dict(name=, id=)]\n \"\"\"\n return [\n dict(\n name=name,\n id=cat_id if not self.cat_id_to_label else self.cat_id_to_label[cat_id]\n ) for name, cat_id in zip(self.cat_names, self.cat_ids)]\n\n @property\n def max_label(self):\n if self.cat_id_to_label:\n return max(self.cat_id_to_label.values())\n else:\n assert len(self.cat_ids) and isinstance(self.cat_ids[0], Integral)\n return max(self.cat_ids)\n","repo_name":"rwightman/efficientdet-pytorch","sub_path":"effdet/data/parsers/parser.py","file_name":"parser.py","file_ext":"py","file_size_in_byte":3481,"program_lang":"python","lang":"en","doc_type":"code","stars":1526,"dataset":"github-code","pt":"78"}
+{"seq_id":"6210934982","text":"from __future__ import print_function\r\nimport os\r\nimport argparse\r\nimport numpy as np\r\nimport torch\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\nfrom torch.utils.data import DataLoader\r\nfrom torch.optim.lr_scheduler import CosineAnnealingLR\r\nfrom sklearn.metrics import accuracy_score, balanced_accuracy_score\r\nfrom pointnet2_ops_lib.pointnet2_ops import pointnet2_utils\r\n# custom module\r\nfrom model import DGCNN_cls\r\nfrom util import cal_loss, IOStream\r\nfrom data import ScanObject_coseg, TrainingBatchSampler\r\n\r\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = \"0, 1, 2, 3, 4\"\r\n\r\n\r\ndef _init_():\r\n if not os.path.exists('checkpoints'):\r\n os.makedirs('checkpoints')\r\n if not os.path.exists('checkpoints/'+args.path):\r\n os.makedirs('checkpoints/'+args.path)\r\n if not os.path.exists('checkpoints/'+args.path+'/'+'models'):\r\n os.makedirs('checkpoints/'+args.path+'/'+'models')\r\n os.system('cp exp_classification.py checkpoints'+'/' +\r\n args.path+'/'+'exp_classification.py.backup')\r\n os.system('cp model.py checkpoints' + '/' +\r\n args.path + '/' + 'model.py.backup')\r\n os.system('cp util.py checkpoints' + '/' +\r\n args.path + '/' + 'util.py.backup')\r\n os.system('cp data.py checkpoints' + '/' +\r\n args.path + '/' + 'data.py.backup')\r\n\r\n\r\ndef train(args, io, model, train_loader, opt):\r\n # I will NEVER train the model on a CPU so I simply set \"cuda\"\r\n device = torch.device(\"cuda\")\r\n\r\n criterion = cal_loss\r\n train_loss = 0.0\r\n count = 0.0\r\n model.train()\r\n train_pred = []\r\n train_true = []\r\n for data, label, mask in train_loader:\r\n # load data\r\n data, label = data.to(device), label.to(device).squeeze()\r\n data = data.permute(0, 2, 1)\r\n batch_size = data.size()[0]\r\n\r\n # forward propagation and back propagation\r\n opt.zero_grad()\r\n logits = model(data)\r\n loss = criterion(logits, label)\r\n loss.backward()\r\n opt.step()\r\n\r\n # record results\r\n preds = logits.max(dim=1)[1]\r\n count += batch_size\r\n train_loss += loss.item() * batch_size\r\n train_true.append(label.cpu().numpy())\r\n train_pred.append(preds.detach().cpu().numpy())\r\n\r\n # record results\r\n train_true = np.concatenate(train_true)\r\n train_pred = np.concatenate(train_pred)\r\n io.cprint('Train '\r\n 'loss: %.6f, '\r\n 'train acc: %.6f, '\r\n 'train avg acc: %.6f'\r\n % (train_loss / count,\r\n accuracy_score(train_true, train_pred),\r\n balanced_accuracy_score(train_true, train_pred)))\r\n\r\n return\r\n\r\n\r\ndef test(args, io, model, test_loader, test_name,n_point):\r\n with torch.no_grad():\r\n device = torch.device(\"cuda\")\r\n model = model.eval()\r\n\r\n # initialize the parameters\r\n criterion = cal_loss\r\n test_loss = 0.0\r\n count = 0.0\r\n model.eval()\r\n test_pred = []\r\n test_true = []\r\n for data, label, mask in test_loader:\r\n # load data\r\n data, label = data.to(device), label.to(device).squeeze()\r\n #Farest point sample\r\n #data: (B*N*3) => (B*n_point*3)\r\n data_flipped = data.transpose(1, 2).contiguous()\r\n\r\n data = (\r\n pointnet2_utils.gather_operation(\r\n data_flipped, pointnet2_utils.furthest_point_sample(data, n_point)\r\n )\r\n .transpose(1, 2)\r\n .contiguous()\r\n if n_point is not None\r\n else None\r\n )\r\n\r\n data = data.permute(0, 2, 1)\r\n batch_size = data.size()[0]\r\n\r\n # predict\r\n logits = model(data)\r\n loss = criterion(logits, label)\r\n preds = logits.max(dim=1)[1]\r\n\r\n # record results\r\n count += batch_size\r\n test_loss += loss.item() * batch_size\r\n test_true.append(label.cpu().numpy())\r\n test_pred.append(preds.detach().cpu().numpy())\r\n\r\n # record results\r\n test_true = np.concatenate(test_true)\r\n test_pred = np.concatenate(test_pred)\r\n test_acc = accuracy_score(test_true, test_pred)\r\n avg_per_class_acc = balanced_accuracy_score(test_true, test_pred)\r\n\r\n io.cprint(' * %s '\r\n 'loss: %.6f, '\r\n 'test acc: %.6f, '\r\n 'test avg acc: %.6f'\r\n % (test_name,\r\n test_loss / count,\r\n test_acc,\r\n avg_per_class_acc))\r\n\r\n return test_acc\r\n\r\n\r\ndef experiment(n_points_choices, path):\r\n args.n_points_choices = n_points_choices\r\n args.path = path\r\n\r\n # make path\r\n _init_()\r\n\r\n # record args\r\n io = IOStream('checkpoints/' + args.path + '/run.log')\r\n io.cprint(str(args))\r\n\r\n io.cprint('Using GPU : ' + str(torch.cuda.current_device()) +\r\n ' from ' + str(torch.cuda.device_count()) + ' devices')\r\n\r\n # set seeds\r\n torch.manual_seed(1)\r\n torch.cuda.manual_seed(1)\r\n\r\n # initialize train_loader\r\n train_dataset = ScanObject_coseg(partition='training')\r\n train_sampler = TrainingBatchSampler(train_dataset,\r\n args.n_points_choices,\r\n args.batch_size)\r\n train_loader = DataLoader(train_dataset,\r\n batch_sampler=train_sampler,\r\n num_workers=32)\r\n\r\n # initialize test data loaders\r\n test_dataset256 = ScanObject_coseg(partition='test', n_points=256)\r\n test_loader256 = DataLoader(test_dataset256, num_workers=16,\r\n batch_size=args.test_batch_size)\r\n test_dataset512 = ScanObject_coseg(partition='test', n_points=512)\r\n test_loader512 = DataLoader(test_dataset512, num_workers=16,\r\n batch_size=args.test_batch_size)\r\n test_dataset1024 = ScanObject_coseg(partition='test', n_points=1024)\r\n test_loader1024 = DataLoader(test_dataset1024, num_workers=16,\r\n batch_size=args.test_batch_size)\r\n test_dataset2048 = ScanObject_coseg(partition='test', n_points=2048)\r\n test_loader2048 = DataLoader(test_dataset2048, num_workers=16,\r\n batch_size=args.test_batch_size)\r\n\r\n # Load models\r\n device = torch.device(\"cuda\")\r\n model = DGCNN_cls(args).to(device)\r\n model = nn.DataParallel(model)\r\n\r\n # Use SGD and CosineAnnealingLR to train\r\n print(\"Use SGD\")\r\n opt = optim.SGD(model.parameters(), lr=args.lr,\r\n momentum=args.momentum, weight_decay=1e-4)\r\n scheduler = CosineAnnealingLR(opt, args.epochs, eta_min=1e-3)\r\n\r\n # start training\r\n print(\"Let's use\", torch.cuda.device_count(), \"GPUs!\")\r\n for i in range(args.epochs):\r\n io.cprint('Epoch [%d]' % (i + 1))\r\n # train model\r\n train(args, io, model, train_loader, opt)\r\n\r\n # adjust learning rate\r\n scheduler.step()\r\n\r\n # test\r\n test(args, io, model, test_loader256, 'Test 256 ',256)\r\n test(args, io, model, test_loader512, 'Test 512 ',515)\r\n test(args, io, model, test_loader1024, 'Test 1024',1024)\r\n test(args, io, model, test_loader2048, 'Test 2048',2048)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n # Training settings\r\n parser = argparse.ArgumentParser(description='Point Cloud Recognition')\r\n parser.add_argument('--exp_name', type=str, default='exp', metavar='N',\r\n help='Name of the experiment')\r\n parser.add_argument('--dataset', type=str, default='scanobjectnn', metavar='N',\r\n choices=['scanobjectnn'])\r\n parser.add_argument('--batch_size', type=int, default=128, metavar='batch_size',\r\n help='Size of batch)')\r\n parser.add_argument('--test_batch_size', type=int, default=32, metavar='batch_size',\r\n help='Size of batch)')\r\n parser.add_argument('--epochs', type=int, default=250, metavar='N',\r\n help='number of episode to train ')\r\n parser.add_argument('--lr', type=float, default=0.1, metavar='LR',\r\n help='learning rate')\r\n parser.add_argument('--momentum', type=float, default=0.9, metavar='M',\r\n help='SGD momentum (default: 0.9)')\r\n parser.add_argument('--dropout', type=float, default=0.5,\r\n help='initial dropout rate')\r\n parser.add_argument('--emb_dims', type=int, default=1024, metavar='N',\r\n help='Dimension of embeddings')\r\n parser.add_argument('--k', type=int, default=20, metavar='N',\r\n help='Num of nearest neighbors to use')\r\n args = parser.parse_args()\r\n\r\n experiment([256], 'train256')\r\n experiment([512], 'train512')\r\n experiment([1024], 'train1024')\r\n experiment([2048], 'train2048')\r\n experiment([256, 512, 1024, 2048], 'train_mix')\r\n","repo_name":"KaivinC/dgcnn-experiment","sub_path":"exp_classification.py","file_name":"exp_classification.py","file_ext":"py","file_size_in_byte":9051,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"971635433","text":"\"\"\"File to grab entries from gradebook.db and put them in a readable csv format\"\"\"\nimport csv\nimport sqlite3\n\n\ndef main():\n db_fname = \"gradebook.db\"\n conn = sqlite3.connect(db_fname)\n cur = conn.cursor()\n data = cur.execute(\"SELECT * FROM GRADES;\")\n with open('grades.csv', 'w') as f:\n writer = csv.writer(f)\n writer.writerow(['userid', 'grade', 'section', 'lab', 'timestamp'])\n writer.writerows(data)\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"data-8/gofer_service","sub_path":"dump_grades.py","file_name":"dump_grades.py","file_ext":"py","file_size_in_byte":479,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"70425421051","text":"import os\n\nfrom module import file_util, container_util\nfrom document import Document\n\n'''\n 문서의 TF, IDF를 계산해주는 클래스\n'''\n\nclass TFIDFModel() :\n # Field & Member Variable\n\n # Constructor\n\n def __init__(self):\n self.encoding = 'UTF-8'\n \n self.documents = []\n self.df_dict = {}\n # encoding 유니코드 언어 설정\n # documents = 파일 리스트로 저장\n # 읽은 파일에 음절이 있는 지 없는 지 빈도 수 확인\n \n '''\n Method\n 1. 파일 불러오기\n '''\n\n '''\n 1. 파일 불러오기\n '''\n def file_load(self, input_path, output_path, encoding):\n file_paths = file_util.get_file_paths(input_path, False)\n \n for file_path in file_paths:\n # Document class를 불러와서 document별 tf_dict 생성\n document = Document()\n document.load(file_path, encoding, self.df_dict, output_path)\n \n # for 문을 벗어나면 document 파일이 지워지기 때문에 list에 append해줌\n self.documents.append(document)\n \n self.save_dict(output_path) \n self.print_dict()\n \n def save_dict(self, output_path):\n output_path = os.path.join(output_path, 'df_dict.dict')\n self.df_dict = container_util.sorted_dict(self.df_dict)\n file_util.open_file(output_path, encoding, 'w')\n file_util.write_dict(output_path, encoding, self.df_dict, ' : ')\n \n \n def print_dict(self):\n print('df_dict len :', len(self.df_dict))\n for term in self.df_dict:\n print(f'\\t{term} : {self.df_dict[term]}')\n print()\n \n \n \n# main\nif __name__ == \"__main__\" :\n input_path = 'input_file'\n output_path = 'output_file'\n encoding = 'UTF-8'\n \n tfidf_model = TFIDFModel()\n tfidf_model.file_load(input_path, output_path, encoding)","repo_name":"Wildturkeyy/Russells","sub_path":"Python/NLP_Study/02.TFIDF_Model/02.tfidf_second/source/TFIDFModel.py","file_name":"TFIDFModel.py","file_ext":"py","file_size_in_byte":1948,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"167201619","text":"import torch\nfrom torch.nn.functional import grid_sample\n\n\ndef back_project(coords, origin, voxel_size, feats, KRcam):\n '''\n Unproject the image fetures to form a 3D (sparse) feature volume\n\n :param coords: coordinates of voxels,\n dim: (num of voxels, 4) (4 : batch ind, x, y, z)\n :param origin: origin of the partial voxel volume (xyz position of voxel (0, 0, 0))\n dim: (batch size, 3) (3: x, y, z)\n :param voxel_size: floats specifying the size of a voxel\n :param feats: image features\n dim: (num of views, batch size, C, H, W)\n :param KRcam: projection matrix\n dim: (num of views, batch size, 4, 4)\n :return: feature_volume_all: 3D feature volumes\n dim: (num of voxels, c + 1)\n :return: count: number of times each voxel can be seen\n dim: (num of voxels,)\n '''\n n_views, bs, c, h, w = feats.shape\n\n feature_volume_all = torch.zeros(coords.shape[0], c + 1, device=feats.device)\n count = torch.zeros(coords.shape[0], device=feats.device)\n\n for batch in range(bs):\n batch_ind = torch.nonzero(coords[:, 0] == batch, as_tuple=False).squeeze(1)\n coords_batch = coords[batch_ind][:, 1:]\n\n coords_batch = coords_batch.view(-1, 3)\n origin_batch = origin[batch].unsqueeze(0)\n feats_batch = feats[:, batch]\n proj_batch = KRcam[:, batch]\n\n grid_batch = coords_batch * voxel_size + origin_batch.float()\n rs_grid = grid_batch.unsqueeze(0).expand(n_views, -1, -1)\n rs_grid = rs_grid.permute(0, 2, 1).contiguous()\n nV = rs_grid.shape[-1]\n rs_grid = torch.cat([rs_grid, torch.ones([n_views, 1, nV], device=feats.device)], dim=1)\n\n # Project grid\n im_p = proj_batch @ rs_grid\n im_x, im_y, im_z = im_p[:, 0], im_p[:, 1], im_p[:, 2]\n im_x = im_x / im_z\n im_y = im_y / im_z\n\n im_grid = torch.stack([2 * im_x / (w - 1) - 1, 2 * im_y / (h - 1) - 1], dim=-1)\n mask = im_grid.abs() <= 1\n mask = (mask.sum(dim=-1) == 2) & (im_z > 0)\n\n feats_batch = feats_batch.view(n_views, c, h, w)\n im_grid = im_grid.view(n_views, 1, -1, 2)\n features = grid_sample(feats_batch, im_grid, padding_mode='zeros', align_corners=True)\n\n features = features.view(n_views, c, -1)\n mask = mask.view(n_views, -1)\n im_z = im_z.view(n_views, -1)\n # remove nan\n features[mask.unsqueeze(1).expand(-1, c, -1) == False] = 0\n # features = features * mask.unsqueeze(1).float()\n\n im_z[mask == False] = 0\n\n count[batch_ind] = mask.sum(dim=0).float()\n\n # aggregate multi view\n features = features.sum(dim=0)\n mask = mask.sum(dim=0)\n invalid_mask = mask == 0\n mask[invalid_mask] = 1\n in_scope_mask = mask.unsqueeze(0)\n features /= in_scope_mask\n features = features.permute(1, 0).contiguous()\n\n # concat normalized depth value\n im_z = im_z.sum(dim=0).unsqueeze(1) / in_scope_mask.permute(1, 0).contiguous()\n im_z_mean = im_z[im_z > 0].mean()\n im_z_std = torch.norm(im_z[im_z > 0] - im_z_mean) + 1e-5\n im_z_norm = (im_z - im_z_mean) / im_z_std\n im_z_norm[im_z <= 0] = 0\n features = torch.cat([features, im_z_norm], dim=1)\n\n feature_volume_all[batch_ind] = features\n\n return feature_volume_all, count\n","repo_name":"neu-vi/PlanarRecon","sub_path":"ops/back_project.py","file_name":"back_project.py","file_ext":"py","file_size_in_byte":3325,"program_lang":"python","lang":"en","doc_type":"code","stars":257,"dataset":"github-code","pt":"78"}
+{"seq_id":"7586377643","text":"import astropy\nimport pyspeckit\ntry:\n from astropy.io import fits as pyfits\nexcept ImportError:\n import pyfits\nimport numpy as np\n\n# Load the spectrum\nsp = pyspeckit.Cube('region5_hcn_crop.fits')\nerrmap = pyfits.getdata('region5.hcn.errmap.fits')\n\n# Register the fitter\n# The N2H+ fitter is 'built-in' but is not registered by default; this example\n# shows how to register a fitting procedure\n# 'multi' indicates that it is possible to fit multiple components and a background will not automatically be fit\n# 4 is the number of parameters in the model (excitation temperature, optical depth, line center, and line width)\nsp.Registry.add_fitter('hcn_amp',pyspeckit.models.hcn.hcn_amp,3)\n\n# Run the fitter\nsp.mapplot()\n\n# use an individual spectrum selected semi-arbitrarily from the map to get an estimate of the error\n# this method has been rendered obsolete - use the error map instead\n#s = sp.get_spectrum(20,20)\n#s.plotter()\n#s.Registry = sp.Registry\n#s.specfit.Registry = sp.Registry\n#s.specfit(fittype='hcn_amp',guesses=[2.5,-5.6,1.5],show_components=True,debug=True,quiet=False)\n#s.specfit(fittype='hcn_amp',guesses=[2.5,-5.6,1.5],show_components=True,debug=True,quiet=False)\n#sp.error = s.specfit.errspec\n\n\n# Compute the moments at each position to come up with reasonable guesses.\n# This speeds up the process enormously, but can easily mess up the fits if\n# there are bad pixels\nsp.momenteach(vheight=False, verbose=False)\nsp.momentcube[2,:,:] /= 2.5 # the HCN line profile makes the fitter assume a 2.5x too large line\nsp.fiteach(fittype='hcn_amp', errmap=errmap,\n guesses=[1.0,-5.6,1.5], verbose_level=2, signal_cut=4,\n usemomentcube=True, blank_value=np.nan, verbose=False,\n direct=True, multicore=4)\n\n# steal the header from the error map\nf = pyfits.open('region5.hcn.errmap.fits')\n# start replacing components of the pyfits object\nf[0].data = np.concatenate([sp.parcube,sp.errcube,sp.integralmap])\nf[0].header['PLANE1'] = 'amplitude'\nf[0].header['PLANE2'] = 'velocity'\nf[0].header['PLANE3'] = 'sigma'\nf[0].header['PLANE4'] = 'err_amplitude'\nf[0].header['PLANE5'] = 'err_velocity'\nf[0].header['PLANE6'] = 'err_sigma'\nf[0].header['PLANE7'] = 'integral'\nf[0].header['PLANE8'] = 'integral_error'\nf[0].header['CDELT3'] = 1\nf[0].header['CTYPE3'] = 'FITPAR'\nf[0].header['CRVAL3'] = 0\nf[0].header['CRPIX3'] = 1\n# save your work\nif astropy.version.major >= 2 or (astropy.version.major==1 and astropy.version.minor>=3):\n f.writeto('region5.hcn.nosmooth.fit.fits', overwrite=True)\nelse:\n f.writeto('region5.hcn.nosmooth.fit.fits', clobber=True)\n","repo_name":"pyspeckit/pyspeckit","sub_path":"examples/hcn_cube_test.py","file_name":"hcn_cube_test.py","file_ext":"py","file_size_in_byte":2581,"program_lang":"python","lang":"en","doc_type":"code","stars":98,"dataset":"github-code","pt":"78"}
+{"seq_id":"12377178874","text":"import random\nfrom prettytable import PrettyTable\nfrom matplotlib import pyplot as plt\n\n# set principle\nprinciple = float(input('设定本金: '))\n\n# starting wager\nsw = float(input('设定初始下注(需为最小面额筹码的整数倍): '))\n\n# multiplier when loss\nmp = float(input('设定乘数: '))\n\n# take profit (percentage of priciple, 1 indicates doubling the principle)\npl = float(input('设定止盈线(如计划在本金增加50%时停止游戏, 请输入0.5): '))\n\ndef play(principle):\n def bet(_wager):\n res = random.randrange(2)\n if res == 0:\n txn.append(txn[-1] + _wager)\n return [sw, 'success']\n elif res == 1:\n txn.append(txn[-1] - _wager)\n return [_wager * mp, 'failure']\n\n txn = [principle]\n endGame = False\n wager = sw\n while (endGame == False):\n if (txn[-1] < wager):\n wager = 1\n data = bet(wager)\n # print(data[1])\n wager = data[0]\n if (txn[-1] == 0):\n return [txn, 'bankrupt']\n elif (txn[-1] >= principle * (1 + pl)):\n return [txn, 'take profit']\n\n# trial\nprofitTimes = 0\nbankruptTimes = 0\ntrials = int(input('设定模拟次数, 建议为1000: '))\ni = 0\nwhile (i < trials):\n result = play(principle)\n # print(len(result[0]))\n # print(result[1])\n figStat = None\n if result[1] == 'take profit':\n profitTimes += 1\n figStat = 'success'\n elif result[1] == 'bankrupt':\n bankruptTimes += 1\n figStat = 'failure'\n plt.plot(result[0])\n plt.suptitle(f'Change of Balance - Trial {i}',fontsize= 15, fontweight='bold')\n plt.title(f'{principle}-{sw}-{mp}-{pl}-{i}-{figStat}',fontsize= 10)\n plt.xlabel('No. of bets')\n plt.ylabel('Balance')\n plt.savefig(f'./fig/{principle}-{sw}-{mp}-{pl}-{i}.jpg')\n plt.clf()\n i += 1\n\nx = PrettyTable()\nx.field_names = ['Earning Counts', 'Bankcruptcy Counts', 'Success Rate', 'Bankruptcy Rate']\nx.add_row([profitTimes,trials-profitTimes,profitTimes/trials,1-profitTimes/trials]) \nprint(x) \n","repo_name":"songyuew/baccarat_double_up_sim","sub_path":"bjl_rev.py","file_name":"bjl_rev.py","file_ext":"py","file_size_in_byte":2064,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33336279032","text":"from funkcje import *\n\n# Aby uruchomić ten skrypt wybierz 'shell' w prawym dolnym rogu i wpisz 'python stworz_gify.py'.\n# Zapoznaj się ponadto z 'funkcje.py'.\n\nkonfiguracja = [\n [\"wejscie/barabasi_40.csv\", \"siec bezskalowa - 40 wezlow\"],\n [\"wejscie/erdos_40.csv\", \"siec losowa - 40 wezlow\"]\n]\n\nfor symulacja in konfiguracja: \n print(symulacja)\n plik_wejsciowy=symulacja[0]\n nazwa=symulacja[1]\n graf = wczytaj_graf(plik_wejsciowy)\n gamma = 0.1\n beta = 0.03\n wynik = symuluj(50, graf, gamma=gamma, beta=beta, poczatkowa_liczba_chorych=4)\n wynik[\"gamma\"] = gamma\n wynik[\"beta\"] = beta\n wynik[\"nazwa\"] = nazwa\n animuj(graf, wynik, plik_wyjsciowy=nazwa + \".gif\")\n \n # rysuj_wykresy_chorych(symulacja[2], \"100 wezlow, sredni stopien: 10\")","repo_name":"kubajal/MDCS-wyklad-replit","sub_path":"stworz_gify.py","file_name":"stworz_gify.py","file_ext":"py","file_size_in_byte":751,"program_lang":"python","lang":"pl","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"5983856006","text":"import streamlit as st\nfrom user import session_prompt,ask, append_interaction_to_chat_log\nimport speech_recognition as sr\n\n\nst.set_page_config(page_title=\"Munshi: AI Chatbot\", page_icon=\":guardsman:\", layout=\"wide\")\n\ndef recognize_speech():\n # Create an instance of the Recognizer class\n r = sr.Recognizer()\n \n # Use the microphone as the audio source\n with sr.Microphone() as source:\n # Adjust for ambient noise\n r.adjust_for_ambient_noise(source)\n \n # Ask the user for input\n print(\"Say something!\")\n \n # Listen for the user's input\n audio = r.listen(source)\n \n try:\n # Recognize speech using Google Speech Recognition\n query = r.recognize_google(audio)\n return query\n except sr.UnknownValueError:\n print(\"Sorry, I didn't understand that.\")\n except sr.RequestError as e:\n print(\"Could not request results from Google Speech Recognition service; {0}\".format(e))\n \n return \"\"\n\n\ndef chat_bot(question):\n global chat_log\n #print(\"question=\",question)\n answer = ask(question, chat_log)\n chat_log = append_interaction_to_chat_log(question, answer, chat_log)\n #print(\"answer=\",answer)\n return answer\n\nchat_log = session_prompt\n\nst.title(\"Munshi: AI Chatbot\")\nst.markdown(\"Chat with Munshi, an AI chatbot powered by OpenAI.\")\n\n# Create a button to allow users to input text using speech\nif st.button(\"Speak\"):\n query = recognize_speech()\n st.text_input(\"Query\", value=query)\nelse:\n # Otherwise, show a regular text input field\n query = st.text_input(\"Query\")\n\nif query:\n response = chat_bot(query)\n st.write(response)\n\n \n ","repo_name":"shrey2003/M-GPT","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1727,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"3677479595","text":"\"\"\"\n 746. 使用最小花费爬楼梯:\n\n 数组的每个索引做为一个阶梯,第 i 个阶梯对应着一个非负的体力花费值 cost[i](索引从 0 开始)\n 每当你爬上一个阶梯你都要花费对应的体力花费值,然后你可以选择继续爬一个阶梯或者两个阶梯。\n 你需要找到达到楼层顶部的最低花费。在开始时,你可以选择索引从 0 或 1 的元素作为初始阶梯。\n\n 示例 1:\n\n\n 输入: [10, 15, 20]\n 输出:15\n 解释:最低花费是从cost[1]开始,然后走两步即可到阶梯顶,一共花费15。\n 走一步,cost[1]=15, 下一次我直接越过20,走两步,到顶了。\n\n 示例 2:\n\n 输入: cost = [1, 100, 1, 1, 1, 100, 1, 1, 100, 1]\n 输出: 6\n 解释: 最低花费方式是从cost[0]开始,逐个经过那些1,跳过cost[3],一共花费6。\n\n 注意:\n\n cost 的长度将会在 [2, 1000]。\n 每一个 cost[i] 将会是一个Integer类型,范围为 [0, 999]。\n\n\"\"\"\n\n\n# 到达当前台阶时判断下从前一个台阶过来省事,还是从前一个的前一个过来省事,\n# 一直累加到最后一个台阶完,最小值就是最省体力的。\n# 用p1和p2表示前两个和前一个台阶所耗费的体力,一遍循环就可以了。\n\nclass Solution(object):\n def minCostClimbingStairs(self, cost):\n p1, p2 = 0, 0\n for i in range(2, len(cost) + 1):\n p1, p2 = p2, min(p2 + cost[i - 1], p1 + cost[i - 2])\n print(p1, p2)\n return p2\n\n\nif __name__ == \"__main__\":\n S = Solution()\n cost = [1, 100, 1, 1, 1, 100, 1, 1, 100, 1]\n print(S.minCostClimbingStairs(cost))","repo_name":"zzkk007/LeetCode","sub_path":"Easy/array/746_minCostClimbingStairs.py","file_name":"746_minCostClimbingStairs.py","file_ext":"py","file_size_in_byte":1722,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"7407828959","text":"# One dimensional diffusion system, with four species, A\n# B, C and D, which interact as in Gillespie_AB_system.py\n#\n# k1 k2\n# A + A -> C A + B -> D\n#\n# A is created only in 0 < x < 9L/10, and\n# B is created only in 2L/5 < x < L\n#\n# This follows section 4.1 of:\n#\n# \"A Practical Guide to Stochastic Simulations of\n# Reaction-Diffusion Processes\" by Erban et al.\n#\n# The last two panels in the Figure can be compared to\n# Figure 4.1 of Erban et al.\n#\n\nfrom openrxn.systems.ODESystem import ODESystem\nfrom openrxn.systems.GillespieSystem import GillespieSystem\nfrom openrxn.reactions import Reaction, Species\nfrom openrxn.model import Model\nfrom openrxn.compartments.arrays import CompartmentArray1D\nfrom openrxn import unit\nfrom openrxn.systems.reporters import AllReporter, SumReporter, SelectionReporter\nfrom openrxn.connections import IsotropicConnection\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nd = 0.16/unit.sec\nK = 40 # number of compartments\nL = 1*unit.mm\nh = L/K\n\n# define species and reactions\nA = Species('A')\nB = Species('B')\nC = Species('C')\nD = Species('D')\n\n# h is the \"volume\" of the compartment here\n# second order rate coefficients should have units of\n# M^-1 s^-1, or L^d / (s * mol), where L is units of length and\n# d is the dimensionality of the system (here, 1)\n\nconc = 1.0*unit.mol/h\nk1 = 1e-3/unit.sec/conc\nk2 = 1e-2/unit.sec/conc\nk3 = 1.2/unit.sec\nk4 = 1.0/unit.sec\n\nrxns = []\nrxns.append(Reaction('AAC',[A],[C],[2],[1],kf=k1))\nrxns.append(Reaction('ABD',[A,B],[D],[1,1],[1],kf=k2))\nrxns.append(Reaction('birth_A',[],[A],[],[1],kf=k3))\nrxns.append(Reaction('birth_B',[],[B],[],[1],kf=k4))\n\n# create a Model\nboundaries = np.linspace(0,L.magnitude,K+1)*L.units\nconn = IsotropicConnection({'A' : d, 'B' : d},dim=1)\ncomp_array = CompartmentArray1D('main',boundaries,conn)\ncomp_array.add_rxns_to_array([rxns[0],rxns[1]])\nfor c in comp_array.compartments.values():\n if c.pos[0][1] <= 9*L/10:\n c.add_rxn_to_compartment(rxns[2])\n if c.pos[0][1] > 2*L/5:\n c.add_rxn_to_compartment(rxns[3])\n\nmodel = Model(arrays=[comp_array])\nflat_model = model.flatten()\n\n# create a system\nsys = ODESystem(flat_model)\n\n# set initial concentrations\nsys.set_q(np.arange(sys.state.size),0)\n\node_results = sys.run(1800)\n\nfig, ax = plt.subplots(nrows=1,ncols=2)\nax[0].set_ylabel('Number of A molecules')\nax[1].set_ylabel('Number of B molecules')\nax[0].set_xlabel('x (mm)')\nax[1].set_xlabel('x (mm)')\nplt.show()\n\nIDs_A = []\nIDs_B = []\npos_x = []\nfor i in range(K):\n c_name = 'main-{0}'.format(i)\n IDs_A.append(sys.state.index[c_name]['A'])\n IDs_B.append(sys.state.index[c_name]['B'])\npos_x = sys.state.x_pos[IDs_A]/1000000 # in mm\n\nt4 = ode_results.y.shape[1]-1\nax[0].plot(pos_x,ode_results.y[IDs_A,t4],label='ODE')\nax[1].plot(pos_x,ode_results.y[IDs_B,t4],label='ODE')\n\n#----\n# Now create a GillespieSystem for the same model\n#---\n\nfor i in range(1):\n Gillespie_sys = GillespieSystem(flat_model)\n Gillespie_sys.add_reporter(AllReporter(freq=100))\n Gillespie_sys.set_q(np.arange(sys.state.size),0)\n tmp = Gillespie_sys.run(1800)\n IDs_A = []\n IDs_B = []\n for j in range(K):\n c_name = 'main-{0}'.format(j)\n IDs_A.append(sys.state.index[c_name]['A'])\n IDs_B.append(sys.state.index[c_name]['B'])\n reports = Gillespie_sys.reporters[0].reports()\n ax[0].plot(pos_x,reports[-1]['report'][IDs_A],label='run {0}'.format(i))\n ax[1].plot(pos_x,reports[-1]['report'][IDs_B],label='run {0}'.format(i))\n\nax[0].legend()\nax[1].legend()\nplt.show()\n","repo_name":"ADicksonLab/OpenRXN","sub_path":"examples/1D_reac_diff_AB.py","file_name":"1D_reac_diff_AB.py","file_ext":"py","file_size_in_byte":3550,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"78"}
+{"seq_id":"11059848275","text":"\n\narr = [1,2,4]\n\nfor i in arr:\n print(i)\n\n'''\n要区分两个概念\niterable: 可迭代对象,可以通过iter()函数转换为iterator, 也可以通过for循环遍历。它没有状态。\n\niterator: 迭代器,可以通过next()函数获取下一个元素,它持有关联的容器的当前访问的状态。\n'''\n\n\n","repo_name":"Robinguo2012/Python001","sub_path":"design patten/iterator.py","file_name":"iterator.py","file_ext":"py","file_size_in_byte":323,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"41682808338","text":"import cv2\nimport os\nfrom solutions import Difficult,Medium,Easy\nfrom utils import evaluate\nimport time\n\ndef main():\n reco_res = {}\n ground_truth = {'1-1.jpg': '沪EWM957', '1-2.jpg': '沪AF02976', '1-3.jpg': '鲁NBK268',\n '2-1.jpg': '沪EWM957', '2-2.jpg': '豫B20E68', '2-3.jpg': '沪A93S20',\n '3-1.jpg': '沪EWM957', '3-2.jpg': '沪ADE6598', '3-3.jpg': '皖SJ6M07'}\n\n levels = ['easy','medium','difficult']\n proc_time = {}\n for level in levels:\n time_start = time.time()\n path = './images/'+level\n if level == 'difficult':\n for filename in os.listdir(path):\n img = cv2.imread(path+'/'+filename)\n IMG_h = 800\n IMG_w = 1000\n rec = Difficult(img,IMG_w,IMG_h,filename)\n res = rec.process()\n print(level,filename,res)\n reco_res[filename]=res\n if level == 'medium':\n for filename in os.listdir(path):\n img = cv2.imread(path+'/'+filename)\n height = len(img)\n width = len(img[0])\n rec = Medium(img,width,height,filename)\n res = rec.process()\n print(level,filename,res)\n reco_res[filename]=res\n if level == 'easy':\n for filename in os.listdir(path):\n img = cv2.imread(path+'/'+filename)\n height = len(img)\n width = len(img[0])\n rec = Easy(img,width,height,filename)\n res = rec.process()\n print(level,filename,res)\n reco_res[filename]=res\n time_end = time.time()\n proc_time[level]=(time_end-time_start)/3\n\n\n precision = evaluate(reco_res, ground_truth)\n print('识别结果:', reco_res)\n print('正确答案:', ground_truth)\n print('准确率:', precision)\n print('识别时间:',proc_time)\n\nif __name__ == '__main__':\n main()","repo_name":"chesiy/lisence_recognition","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1995,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"74094355131","text":"\r\n## Library and Settings\r\nimport pandas as pd\r\nimport numpy as np\r\nimport requests\r\nimport warnings\r\nimport time\r\nimport os\r\nimport boto3\r\nimport shutil\r\n\r\nfrom bs4 import BeautifulSoup\r\nfrom tqdm import tqdm, trange\r\nfrom datetime import datetime, timedelta\r\nfrom sqlalchemy import create_engine\r\n\r\nwarnings.simplefilter(action='ignore', category=FutureWarning)\r\npd.set_option('display.max_columns', None)\r\n\r\n\r\n\r\n\r\ndef db_connection():\r\n\r\n id = \"###########\"\r\n pw = \"###########\"\r\n host = \"##############\"\r\n port = \"############\"\r\n db = \"##########\"\r\n\r\n engine = create_engine(f\"mysql+pymysql://{id}:{pw}@{host}:{port}/{db}\")\r\n \r\n return engine\r\n\r\n\r\n\r\n\r\ndef get_index_df(engine, date):\r\n \r\n index_query = f\"\"\"\r\n SELECT *\r\n FROM ftc.FTC_MAIN_INDEX_HIST_TB AS A\r\n WHERE 1=1\r\n AND A.search_dt = \"{date}\"\r\n \"\"\"\r\n index_df = pd.read_sql(index_query, engine)\r\n\r\n return index_df\r\n\r\n\r\n\r\n\r\n\r\ndef get_main_df(engine):\r\n \r\n main_query = f\"\"\"\r\n SELECT *\r\n FROM ftc.FTC_HEAD_INFO_HIST AS A\r\n \"\"\"\r\n main_df = pd.read_sql(main_query, engine)\r\n\r\n return main_df\r\n\r\n\r\n\r\n\r\n\r\ndef get_new_df(index_df, main_df):\r\n \r\n new_df = index_df[~index_df[\"page_id\"].isin(main_df[\"page_id\"])]\r\n new_df = new_df.reset_index(drop = True)\r\n\r\n return new_df\r\n\r\n\r\n\r\n\r\n\r\ndef to_ncp(s3, date, file_name):\r\n\r\n s3.put_object(Bucket = \"winkstone-data-lake\", Key = f\"crawler/01_FTC/\")\r\n\r\n local_file_path = f\"./{date}/{file_name}.html\"\r\n\r\n try:\r\n s3.upload_file(local_file_path, \"winkstone-data-lake\", f\"crawler/01_FTC/{date}/{file_name}.html\")\r\n\r\n except Exception as e:\r\n print(e)\r\n\r\n\r\n\r\n\r\n\r\ndef crawler(page_id):\r\n\r\n url = f\"https://franchise.ftc.go.kr/mnu/00013/program/userRqst/view.do?firMstSn={page_id}\"\r\n\r\n response = requests.get(url)\r\n \r\n if response.status_code == 200:\r\n soup = BeautifulSoup(response.content, \"html.parser\")\r\n html_list = soup.select(\"table\")\r\n table_list = []\r\n for i in range(len(html_list)):\r\n table_html = str(html_list[i])\r\n table_df = pd.read_html(table_html)[0]\r\n table_list.append(table_df)\r\n head = table_list[0]['상호'][0]\r\n head = head.replace(\"상호 \", \"\")\r\n brand = table_list[0]['영업표지'][0]\r\n brand = brand.replace(\"영업표지 \", \"\")\r\n\r\n if head != '상호':\r\n # print(f\"{timestemp} Saved - {page_id}, {head} - {brand}\")\r\n return soup\r\n \r\n else:\r\n # print(f\"{timestemp} Pass - {page_id}, {head} - {brand}\")\r\n pass\r\n else:\r\n # print(f\"{timestemp} Retry - {page_id}, {head} - {brand}\")\r\n time.sleep(1)\r\n crawler(page_id)\r\n\r\n\r\n\r\n\r\n\r\ndef save_html_file(folder_path, file_name, soup):\r\n\r\n if not os.path.exists(folder_path):\r\n os.makedirs(folder_path)\r\n \r\n with open(f\"{folder_path}/{file_name}.html\", \"w\") as file:\r\n file.write(str(soup))\r\n\r\n\r\n\r\n\r\ndef get_s3_client():\r\n\r\n service_name = '##'\r\n endpoint_url = 'https://kr.object.fin-ncloudstorage.com'\r\n region_name = 'kr-standard'\r\n access_key = '#################'\r\n secret_key = '#####################'\r\n\r\n s3 = boto3.client(service_name, endpoint_url = endpoint_url, aws_access_key_id = access_key, aws_secret_access_key = secret_key)\r\n \r\n return s3\r\n\r\n\r\n\r\n\r\ndef main_html_download():\r\n\r\n engine = db_connection()\r\n\r\n date = datetime.today().strftime(\"%Y-%m-%d\")\r\n\r\n print(\"Getting new data from URL ... \")\r\n index_df = get_index_df(engine, date)\r\n main_df = get_main_df(engine)\r\n new_df = get_new_df(index_df, main_df)\r\n\r\n\r\n print(\"Downloading HTML files in local and cloud ... \")\r\n s3 = get_s3_client()\r\n\r\n for page_id in tqdm(new_df[\"page_id\"]):\r\n\r\n soup = crawler(page_id)\r\n\r\n folder_path = f\"./{date}\"\r\n file_name = str(page_id)\r\n\r\n save_html_file(folder_path, file_name, soup)\r\n to_ncp(s3, date, page_id)\r\n\r\n\r\n print(\"All done ! \")\r\n\r\n\r\nif __name__ == '__main__':\r\n main_html_download()\r\n\r\n\r\n\r\n\r\n\r\n\r\n","repo_name":"JihyeonKim14/WK_partners","sub_path":"ETL/s3_html_load.py","file_name":"s3_html_load.py","file_ext":"py","file_size_in_byte":4129,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"15420693360","text":"from collections import deque\n# n, m = map(int, input().split())\n# g = [list(map(int, input().split())) for _ in range(n)]\nn, m, g = 6, 5, [[1, 1, 0, 1, 1], [0, 1, 1, 0, 0], [0, 0, 0, 0, 0], [1, 0, 1, 1, 1], [0, 0, 1, 1, 1], [0, 0, 1, 1, 1]] # -> 4, 9\n# print(n,m,g)\n\ndef bfs(i,j):\n q = deque()\n q.append([i,j])\n ck[i][j] = 1\n cnt = 1\n\n while q:\n a, b = q.popleft()\n \n for i, j in zip(dx, dy):\n nx, ny = a + i, b + j\n\n if 0 <= nx < n and 0 <= ny < m and ck[nx][ny] == 0 and g[nx][ny] == 1:\n q.append([nx, ny])\n g[nx][ny] = g[a][b] + 1\n cnt += 1\n ck[nx][ny] = 1\n res.append(cnt)\n\ndx, dy = [1,0,-1,0], [0,1,0,-1]\nres = []\nck = [[0] * m for _ in range(n)]\n\nfor i in range(n):\n for j in range(m):\n if g[i][j] == 1:\n bfs(i, j)\nif res:\n print(len(res))\n print(max(res))\nelse:\n print(0)\n print(0) \n","repo_name":"rkdalsdn94/algoalgo","sub_path":"baekjoon/그림_1926.py","file_name":"그림_1926.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"18417608840","text":"import numpy as np\nfrom jsonextended import units, mockpath, edict\nfrom ejplugins.utils import split_numbers, codata, validate_against_schema\n\ntry:\n import pathlib\nexcept ImportError:\n import pathlib2 as pathlib\n\n\nclass GaussianCube(object):\n \"\"\" parser plugin for jsonextended\n\n as specified at http://h5cube-spec.readthedocs.io/en/latest/cubeformat.html\n\n \"\"\"\n plugin_name = 'gaussian_cube'\n plugin_descript = 'read gaussian cube charge/spin density file'\n file_regex = '*.cube'\n\n def read_file(self, f, **kwargs):\n comments1 = f.readline().strip()\n comments2 = f.readline().strip()\n line = f.readline().strip()\n inputs = split_numbers(line)\n if len(inputs) > 4 and inputs[4] != 1:\n # TODO implement NVAL != 1\n raise NotImplementedError(\"not yet implemented NVAL != 1: {0}\".format(line))\n natoms = inputs[0]\n centre = -1 * np.array(inputs[1:4]) * codata[(\"Bohr\", \"Angstrom\")]\n if natoms < 0:\n # TODO implement DSET_IDS\n raise NotImplementedError(\"not yet implemented DSET_IDS\")\n an, ax, ay, az = split_numbers(f.readline().strip())\n bn, bx, by, bz = split_numbers(f.readline().strip())\n cn, cx, cy, cz = split_numbers(f.readline().strip())\n\n if an <= 0 or bn <= 0 or cn <= 0:\n raise ValueError(\"an, bn and cn must be positive integers\")\n\n avec = [a * an * codata[(\"Bohr\", \"Angstrom\")] for a in [ax, ay, az]]\n bvec = [b * bn * codata[(\"Bohr\", \"Angstrom\")] for b in [bx, by, bz]]\n cvec = [c * cn * codata[(\"Bohr\", \"Angstrom\")] for c in [cx, cy, cz]]\n #centre = 0.5 * (np.array(avec) + np.array(bvec) + np.array(cvec))\n\n atomic_numbers = []\n nuclear_charges = []\n ccoords = []\n for _ in range(int(natoms)):\n anum, ncharge, x, y, z = split_numbers(f.readline().strip())\n atomic_numbers.append(int(anum))\n nuclear_charges.append(ncharge)\n ccoord = (np.asarray([x, y, z]) * codata[(\"Bohr\", \"Angstrom\")]) + centre\n ccoords.append(ccoord.tolist())\n\n values = []\n line = f.readline().strip()\n while line:\n values += line.split()\n line = f.readline().strip()\n\n return {\n \"title\": comments1,\n #\"na\": int(an), \"nb\": int(bn), \"nc\": int(cn),\n \"cell_vectors\": {\n \"a\": {\"units\": \"angstrom\", \"magnitude\": avec},\n \"b\": {\"units\": \"angstrom\", \"magnitude\": bvec},\n \"c\": {\"units\": \"angstrom\", \"magnitude\": cvec}\n },\n #\"centre\": [0, 0, 0],\n \"densities\": [{\n \"type\": comments2,\n \"magnitude\": np.array(values, dtype=float).reshape((int(an), int(bn), int(cn)))\n }],\n \"atoms\": {\"ccoords\": {\"units\": \"angstrom\",\n \"magnitude\": ccoords},\n \"nuclear_charge\": nuclear_charges,\n \"atomic_number\": atomic_numbers}\n }\n\n\ndef ejdict_to_gcube(data, fpath=None, density=0,\n include_atoms=True, adata=None, cell_tol=1E-3):\n \"\"\"\n\n Parameters\n ----------\n data: dict\n fpath: str or None\n output file path or, if None, write to MockPath\n density: int\n take density from data[\"densities\"][density]\n include_atoms: bool\n include atoms in gaussian cube\n adata: dict or None\n separate atom data (for instance for Crystal output)\n cell_tol: float or None\n if not None, raise and error if the data and adata cell vectors are not within this tolerance\n\n Returns\n -------\n fpath: pathlib.Path or jsonextended.mockpath.MockPath\n\n \"\"\"\n if include_atoms and adata is not None:\n if cell_tol:\n cdiff = edict.diff(data[\"cell_vectors\"], adata[\"cell_vectors\"],\n np_allclose=True, rtol=cell_tol, atol=cell_tol)\n if cdiff:\n raise ValueError(\"data and adata have different cell vectors: {}\".format(cdiff))\n data[\"atoms\"] = adata[\"atoms\"]\n\n validate_against_schema(data, \"edensity\")\n\n if fpath is None:\n fpath = mockpath.MockPath(\"test.cube\", is_file=True)\n else:\n fpath = pathlib.Path(fpath)\n\n with fpath.open(\"w\") as f:\n data = units.combine_quantities(data)\n data = units.apply_unitschema(data, {\"a\": \"angstrom\", \"b\": \"angstrom\", \"c\": \"angstrom\", \"ccoords\": \"angstrom\"},\n as_quantity=False)\n natoms = 0 if \"atoms\" not in data or not include_atoms else len(data[\"atoms\"][\"ccoords\"])\n\n avec = np.asarray(data[\"cell_vectors\"][\"a\"]) / codata[(\"Bohr\", \"Angstrom\")]\n bvec = np.asarray(data[\"cell_vectors\"][\"b\"]) / codata[(\"Bohr\", \"Angstrom\")]\n cvec = np.asarray(data[\"cell_vectors\"][\"c\"]) / codata[(\"Bohr\", \"Angstrom\")]\n centre = 0.5 * (avec + bvec + cvec)\n centre_offset = -1*centre\n\n f.write(data[\"title\"] + \"\\n\")\n f.write(data[\"densities\"][density][\"type\"] + \"\\n\")\n dense = np.asarray(data[\"densities\"][density][\"magnitude\"])\n na, nb, nc = dense.shape\n f.write(\"{0:6d} {1:10.6f} {2:10.6f} {3:10.6f}\\n\".format(natoms, *centre_offset.tolist()))\n f.write(\"{0:6d} {1:10.6f} {2:10.6f} {3:10.6f}\\n\".format(na, *(avec/na).tolist()))\n f.write(\"{0:6d} {1:10.6f} {2:10.6f} {3:10.6f}\\n\".format(nb, *(bvec/nb).tolist()))\n f.write(\"{0:6d} {1:10.6f} {2:10.6f} {3:10.6f}\\n\".format(nc, *(cvec/nc).tolist()))\n\n if data.get(\"atoms\", False):\n for i, c in enumerate(data[\"atoms\"][\"ccoords\"]):\n atomic_number = data[\"atoms\"][\"atomic_number\"][i]\n nuclear_charge = data[\"atoms\"][\"nuclear_charge\"][i]\n ccoord = (np.array(c) / codata[(\"Bohr\", \"Angstrom\")]) + centre_offset\n\n f.write(\"{0:6d} {1:10.6f} {2:10.6f} {3:10.6f} {4:10.6f}\\n\".format(atomic_number, nuclear_charge,\n *ccoord.tolist()))\n dense = dense.flatten().tolist()\n dlength = len(dense)\n output = []\n for i in range(int(dlength/6.)+1):\n if dlength > i*6 + 6:\n output.append(\"{0:12.5E} {1:12.5E} {2:12.5E} {3:12.5E} {4:12.5E} {5:12.5E}\".format(*dense[i*6: i*6+6]))\n else:\n output.append(\" \".join([\"{0:12.5E}\".format(v) for v in dense[i*6: dlength]]))\n\n f.write(\"\\n\".join(output))\n\n return fpath\n","repo_name":"chrisjsewell/ejplugins","sub_path":"ejplugins/gcube.py","file_name":"gcube.py","file_ext":"py","file_size_in_byte":6540,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"36770922009","text":"def chunking_by(items: list, size: int):\r\n listy=[]\r\n count=len(items)//size\r\n\r\n for i in range(0,count):\r\n listy1=[] \r\n for y in range(0,size):\r\n listy1.append(items[0])\r\n items.remove(items[0]) \r\n listy.append(listy1)\r\n if items!=[]:\r\n listy.append(items)\r\n return listy","repo_name":"MiracleX77/CheckIO","sub_path":"O'Reilly/Chunk.py","file_name":"Chunk.py","file_ext":"py","file_size_in_byte":342,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"32395107085","text":"'''\nInserindo e removendo elementos da lista\nappend(elemento) insert(index, elemento) pop(index) remove(elemento)\n'''\n\nlanches = ['doritos', 'hamburguer', 'batata-frita', 'pudim']\n\n# Verificando o tipo Lista\nprint(type(lanches))\n\ndef exibir():\n print(lanches)\n\nexibir()\n\n# Inserindo elemento na lista\nlanches.append('salgadinho')\nlanches.insert(0, 'piraque')\n\nexibir()\n\n# Removendo último elemento\nlanches.pop() \n# Removendo elemento pelo índice\nlanches.pop(2)\n# Removendo elemento pelo elemento\nlanches.remove('piraque')\n\nexibir()\n\nif 'refrigerante' in lanches:\n lanches.remove('refrigerante')\n\n'''\nUma outra maneira de remover um elemento\ndel lanche[2]\n'''\n","repo_name":"udanielnogueira/Python.Treinamento1","sub_path":"mundo-3/2-1-lists.py","file_name":"2-1-lists.py","file_ext":"py","file_size_in_byte":667,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"10641948708","text":"import sqlite3\n\nfrom config.config import db_path\n\n\nclass DBConnect(object):\n\n def __new__(cls):\n \"\"\"\n метод добавляет классу свойства синглетона и инициализирует его\n \"\"\"\n if not hasattr(cls, 'instance'):\n try:\n cls.conn = sqlite3.connect(db_path, check_same_thread=False)\n except Exception as e:\n raise e\n\n cls.instance = super(DBConnect, cls).__new__(cls)\n return cls.instance\n\n def getConnection(self):\n return self.conn\n\n def getCursor(self):\n return self.conn.cursor()\n\n def do_select_query(self, q):\n cur = self.getCursor()\n try:\n cur.execute(q)\n o = cur.fetchone()\n if o is not None:\n o = o[0]\n return o\n except sqlite3.Error as e:\n raise e\n finally:\n if cur:\n cur.close()\n\n def do_selects_query(self, q):\n cur = self.getCursor()\n try:\n cur.execute(q)\n o = cur.fetchall()\n return o\n except sqlite3.Error as e:\n raise e\n finally:\n if cur:\n cur.close()\n\n @staticmethod\n def get_custom_user_field(tg_id, table, field):\n \"\"\"\n метод получает сложное поле из db\n :param tg_id:\n :param table:\n :param field:\n :return:\n \"\"\"\n cur = DBConnect().getCursor()\n try:\n q = f'select {field} from {table} where tg_id = {tg_id}'\n cur.execute(q)\n step = cur.fetchone()[0]\n return step\n except sqlite3.Error as e:\n raise e\n finally:\n if cur:\n cur.close()\n\n @staticmethod\n def update_custom_user_filed(tg_id, table, filed, val):\n \"\"\"\n метод обновляет сложное поле из db\n :param tg_id:\n :param table:\n :param filed:\n :param val:\n \"\"\"\n conn = DBConnect().getConnection()\n cur = conn.cursor()\n try:\n if val is None:\n q = f'update {table} set {filed} = NULL where tg_id = {tg_id}'\n else:\n q = f'update {table} set {filed} = \\'{val}\\' where tg_id = {tg_id}'\n cur.execute(q)\n conn.commit()\n except sqlite3.Error as e:\n raise e\n finally:\n if cur:\n cur.close()\n\n @staticmethod\n def update_custom_user_photo_filed(tg_id, table, filed, photo):\n \"\"\"\n метод обновляет фото произвольного поля из db\n :param tg_id:\n :param table:\n :param filed:\n :param photo:\n \"\"\"\n conn = DBConnect().getConnection()\n cur = conn.cursor()\n try:\n q = f'update {table} set {filed} = ? where tg_id = {tg_id}'\n cur.execute(q, [photo])\n conn.commit()\n except sqlite3.Error as e:\n raise e\n finally:\n if cur:\n cur.close()\n","repo_name":"akatsnelson/hsetravel_mem60","sub_path":"dao/db_connection.py","file_name":"db_connection.py","file_ext":"py","file_size_in_byte":3191,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20668859149","text":"import hashlib\nfrom Crypto import Random\nfrom Crypto.Cipher import AES\nfrom base64 import b64encode, b64decode\n\nimport os\nfrom stat import S_IREAD, S_IRWXU\n\nclass AESCipher(object):\n def __init__(self, key):\n self.block_size = AES.block_size\n self.key = hashlib.sha256(key.encode()).digest()\n\n def encrypt(self, plain_text):\n plain_text = self.__pad(plain_text)\n iv = Random.new().read(self.block_size)\n cipher = AES.new(self.key, AES.MODE_CBC, iv)\n encrypted_text = cipher.encrypt(plain_text.encode())\n return b64encode(iv + encrypted_text).decode(\"utf-8\")\n\n def decrypt(self, encrypted_text):\n encrypted_text = b64decode(encrypted_text)\n iv = encrypted_text[:self.block_size]\n cipher = AES.new(self.key, AES.MODE_CBC, iv)\n plain_text = cipher.decrypt(encrypted_text[self.block_size:]).decode(\"utf-8\")\n return self.__unpad(plain_text)\n\n def __pad(self, plain_text):\n number_of_bytes_to_pad = self.block_size - len(plain_text) % self.block_size\n ascii_string = chr(number_of_bytes_to_pad)\n padding_str = number_of_bytes_to_pad * ascii_string\n padded_plain_text = plain_text + padding_str\n return padded_plain_text\n\n @staticmethod\n def __unpad(plain_text):\n last_character = plain_text[len(plain_text) - 1:]\n return plain_text[:-ord(last_character)]\n\npath = os.getcwd()\ndir_list = os.listdir(path)\nprint(dir_list)\nos.chmod('state', S_IREAD)\ng = open('state', 'r')\nstatus = g.readlines()\nif(status[0] == 'encrypted'):\n\n password = input(\"Enter password for decryption: \")\n a = AESCipher(password)\n # print(a.decrypt('4aiHxRFeKAKTz9zGs2KZF0SmGKaWS/9wb6kXzQkKyyQ='))\n # print(a.decrypt('5USKlrd8JobNwNmVKB3kTw+of+0jiG8IlKCiHt810Do='))\n\n for i in dir_list:\n\n if(i != \"encryption.py\" and i != \"keys\" and i != 'state'):\n os.chmod(i, S_IREAD)\n print(i)\n f = open(i, 'r')\n list = f.readlines()\n elist = \"\"\n m = []\n for j in list:\n elist = a.decrypt(j)\n m.append(elist)\n print(m)\n os.chmod(i, S_IRWXU)\n f = open(i, 'w')\n for k in m:\n f.write(k)\n\n os.chmod('state', S_IRWXU)\n g = open('state', 'w')\n g.write('decrypted')\n os.chmod('state', S_IREAD)\n\nelif(status[0] == 'decrypted'):\n password = input(\"Enter any password for encryption, make sure to remember it: \")\n a = AESCipher(password)\n for i in dir_list:\n if(i != \"encryption.py\" and i != \"keys\" and i!= 'state'):\n f = open(i, 'r')\n list = f.readlines()\n elist = \"\"\n m = []\n for j in list:\n elist = a.encrypt(j)\n m.append(elist)\n print(m)\n f = open(i, 'w')\n for k in m:\n f.write(k)\n f.write('\\n')\n os.chmod(i, S_IREAD)\n os.chmod('state', S_IRWXU)\n g = open('state', 'w')\n g.write('encrypted')\n os.chmod('state', S_IREAD)","repo_name":"JDeepak45/Os","sub_path":"encryption.py","file_name":"encryption.py","file_ext":"py","file_size_in_byte":3102,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17610845559","text":"import torch\nfrom torch import nn\nfrom tqdm import tqdm\nimport numpy as np\n\n\ndef train(model, train_generator, validation_generator, epochs, device, optimizer):\n \"\"\"\n trains a model for a specified number of epochs\n @param model: model currently trained\n @param train_generator: iterator which generates batch of training samples\n @param validation_generator: iterator which generates batch of evaluation samples\n @param epochs: number of epochs to train the model on\n @param device: either 'cpu' or 'cuda' depending on hardware availability\n @param optimizer: optimizer used to updated the model's parameters\n @return:\n \"\"\"\n # Set the objective function\n loss_function = nn.CrossEntropyLoss()\n\n # Move model's parameters to device\n model = model.to(device)\n\n # Set model in train mode for dropout and batch-normalization layers\n model.train(True)\n for e in range(epochs):\n nb_correct_train = 0\n for batch_tweets, batch_labels in tqdm(train_generator):\n # Move batch to device\n batch_labels = torch.LongTensor(batch_labels).to(device)\n\n # Reset the gradient for each batch\n optimizer.zero_grad()\n\n # Forward pass)\n outputs = model.forward(batch_tweets)\n\n # Compute the predictions\n predictions = torch.argmax(outputs.detach(), dim=1)\n nb_correct_train += torch.sum(predictions == batch_labels)\n\n loss = loss_function(outputs, batch_labels.squeeze())\n\n # Backward pass\n loss.backward()\n optimizer.step()\n\n # Compute training accuracy\n train_acc = nb_correct_train.item() / train_generator.dataset.__len__()\n train_message = 'Train accuracy at epoch {} is {}'\n print(train_message.format(e, train_acc))\n\n # Set model in eval mode for dropout and batch-normalization layers\n model.train(False)\n with torch.no_grad():\n nb_correct_val = 0\n for batch_tweets, batch_labels in validation_generator:\n batch_labels = torch.LongTensor(batch_labels).to(device)\n outputs = model.forward(batch_tweets)\n predictions = torch.argmax(outputs, dim=1)\n nb_correct_val += torch.sum(predictions == batch_labels)\n val_acc = nb_correct_val.item() / validation_generator.dataset.__len__()\n val_message = 'Test accuracy is {}'\n print(val_message.format(val_acc))\n\n\ndef build_embedding_matrix(embedding_dim, glove, vocabulary):\n \"\"\"\n completes a pre-trained glove embedding with random initialization for missing entries.\n This implementation is strongly inspired by:\n https://medium.com/@martinpella/how-to-use-pre-trained-word-embeddings-in-pytorch-71ca59249f76\n @param embedding_dim: dimension of the the embedding space\n @param glove: pretrained glove embedding dictionary\n @param vocabulary: list of words used in the tweets\n @return: completed embedding matrix and word to index dictionary\n \"\"\"\n embedding_matrix = np.zeros((len(vocabulary), embedding_dim))\n text_indexer_dict = dict()\n for i, word in enumerate(vocabulary):\n try:\n embedding_matrix[i] = glove[word]\n except KeyError:\n embedding_matrix[i] = np.random.normal(loc=0, scale=1., size=embedding_dim)\n text_indexer_dict[word] = i\n return embedding_matrix, text_indexer_dict\n\n\ndef accuracy(model, inputs, labels, batch_size, device):\n \"\"\"\n computes the accuracy of the model on the given inputs\n @param model: model currently trained\n @param inputs: array of inputs on which to compute the accuracy\n @param labels: array of true labels\n @param batch_size: size of batches\n @param device: cpu or cuda depending on hardware availability\n @return: accuracy of the predictions\n \"\"\"\n nb_correct = 0\n for b in range(0, inputs.shape[0], batch_size):\n outputs = model.forward(inputs.narrow(0, b, batch_size).to(device))\n predictions = torch.argmax(outputs, dim=1)\n nb_correct += torch.sum(predictions == labels.narrow(0, b, batch_size).squeeze().to(device))\n return nb_correct.item() / inputs.shape[0]\n\n\ndef load_dict_contractions():\n \"\"\"\n this list of contractions was taken from a previous versions of :\n https://en.wikipedia.org/wiki/Wikipedia%3AList_of_English_contractions\n @return: dictionary from contraction to corresponding correct spelling\n \"\"\"\n return {\n \"ain't\": \"is not\",\n \"amn't\": \"am not\",\n \"aren't\": \"are not\",\n \"can't\": \"cannot\",\n \"'cause\": \"because\",\n \"couldn't\": \"could not\",\n \"couldn't've\": \"could not have\",\n \"could've\": \"could have\",\n \"daren't\": \"dare not\",\n \"daresn't\": \"dare not\",\n \"dasn't\": \"dare not\",\n \"didn't\": \"did not\",\n \"doesn't\": \"does not\",\n \"don't\": \"do not\",\n \"e'er\": \"ever\",\n \"em\": \"them\",\n \"everyone's\": \"everyone is\",\n \"finna\": \"fixing to\",\n \"gimme\": \"give me\",\n \"gonna\": \"going to\",\n \"gon't\": \"go not\",\n \"gotta\": \"got to\",\n \"hadn't\": \"had not\",\n \"hasn't\": \"has not\",\n \"haven't\": \"have not\",\n \"he'd\": \"he would\",\n \"he'll\": \"he will\",\n \"he's\": \"he is\",\n \"he've\": \"he have\",\n \"how'd\": \"how would\",\n \"how'll\": \"how will\",\n \"how're\": \"how are\",\n \"how's\": \"how is\",\n \"I'd\": \"I would\",\n \"I'll\": \"I will\",\n \"I'm\": \"I am\",\n \"I'm'a\": \"I am about to\",\n \"I'm'o\": \"I am going to\",\n \"isn't\": \"is not\",\n \"it'd\": \"it would\",\n \"it'll\": \"it will\",\n \"it's\": \"it is\",\n \"I've\": \"I have\",\n \"kinda\": \"kind of\",\n \"let's\": \"let us\",\n \"mayn't\": \"may not\",\n \"may've\": \"may have\",\n \"mightn't\": \"might not\",\n \"might've\": \"might have\",\n \"mustn't\": \"must not\",\n \"mustn't've\": \"must not have\",\n \"must've\": \"must have\",\n \"needn't\": \"need not\",\n \"ne'er\": \"never\",\n \"o'\": \"of\",\n \"o'er\": \"over\",\n \"ol'\": \"old\",\n \"oughtn't\": \"ought not\",\n \"shalln't\": \"shall not\",\n \"shan't\": \"shall not\",\n \"she'd\": \"she would\",\n \"she'll\": \"she will\",\n \"she's\": \"she is\",\n \"shouldn't\": \"should not\",\n \"shouldn't've\": \"should not have\",\n \"should've\": \"should have\",\n \"somebody's\": \"somebody is\",\n \"someone's\": \"someone is\",\n \"something's\": \"something is\",\n \"that'd\": \"that would\",\n \"that'll\": \"that will\",\n \"that're\": \"that are\",\n \"that's\": \"that is\",\n \"there'd\": \"there would\",\n \"there'll\": \"there will\",\n \"there're\": \"there are\",\n \"there's\": \"there is\",\n \"these're\": \"these are\",\n \"they'd\": \"they would\",\n \"they'll\": \"they will\",\n \"they're\": \"they are\",\n \"they've\": \"they have\",\n \"this's\": \"this is\",\n \"those're\": \"those are\",\n \"'tis\": \"it is\",\n \"'twas\": \"it was\",\n \"wanna\": \"want to\",\n \"wasn't\": \"was not\",\n \"we'd\": \"we would\",\n \"we'd've\": \"we would have\",\n \"we'll\": \"we will\",\n \"we're\": \"we are\",\n \"weren't\": \"were not\",\n \"we've\": \"we have\",\n \"what'd\": \"what did\",\n \"what'll\": \"what will\",\n \"what're\": \"what are\",\n \"what's\": \"what is\",\n \"what've\": \"what have\",\n \"when's\": \"when is\",\n \"where'd\": \"where did\",\n \"where're\": \"where are\",\n \"where's\": \"where is\",\n \"where've\": \"where have\",\n \"which's\": \"which is\",\n \"who'd\": \"who would\",\n \"who'd've\": \"who would have\",\n \"who'll\": \"who will\",\n \"who're\": \"who are\",\n \"who's\": \"who is\",\n \"who've\": \"who have\",\n \"why'd\": \"why did\",\n \"why're\": \"why are\",\n \"why's\": \"why is\",\n \"won't\": \"will not\",\n \"wouldn't\": \"would not\",\n \"would've\": \"would have\",\n \"y'all\": \"you all\",\n \"you'd\": \"you would\",\n \"you'll\": \"you will\",\n \"you're\": \"you are\",\n \"you've\": \"you have\",\n \"Whatcha\": \"What are you\",\n \"luv\": \"love\",\n \"sux\": \"sucks\"\n }\n","repo_name":"AleksandarHr/Twitter_Sentiment_Analysis","sub_path":"source/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":8248,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"42864441889","text":"# Kingdom's night\ninput_data = input()\nrow = int(input_data[1])\ncolumn = ord(input_data[0]) - ord('a') + 1\n\nsteps = [(-2, -1), (-1, -2), (1, -2), (2, -1), (2, 1), (1, 2), (-1, 2), (-2, 1)]\nanswer = 0\n\nfor step in steps:\n next_row = row + step[0]\n next_column = column + step[1]\n if (next_row >= 1 and next_row <= 8 and next_column >= 1 and next_column <= 8):\n answer = answer + 1\n\nprint(answer)","repo_name":"sigridjineth/algorithm_log","sub_path":"leetcode/Kingdoms_night.py","file_name":"Kingdoms_night.py","file_ext":"py","file_size_in_byte":410,"program_lang":"python","lang":"en","doc_type":"code","stars":25,"dataset":"github-code","pt":"78"}
+{"seq_id":"32915108286","text":"import pandas as pd\nfrom sklearn.cluster import MiniBatchKMeans\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n\ndef main():\n dataset = pd.read_csv('../datasets/candy.csv')\n print(dataset.head())\n # al ser aprendizaje no supervisado, no separamos nuestro datasets en partes\n # eliminaremos los nombres de los caramelos, ya que eso no nos sirve para el algoritmo\n X = dataset.drop('competitorname', axis=1)\n kmeans = MiniBatchKMeans(n_clusters=4, batch_size=8)\n kmeans.fit(X)\n print('Total de centros: ', len(kmeans.cluster_centers_))\n print('=' * 64)\n col_name = \"clusters\"\n dataset[col_name] = kmeans.predict(X)\n print(dataset)\n sns.pairplot(dataset[['sugarpercent', 'pricepercent', 'winpercent', col_name]], hue=col_name, palette='colorblind')\n plt.savefig(\"pares.png\")\n plt.show()\n plt.close()\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"ichcanziho/cursos_platzi","sub_path":"machine_learning_scikit_learn/4_clustering/k_means.py","file_name":"k_means.py","file_ext":"py","file_size_in_byte":942,"program_lang":"python","lang":"es","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"32943226995","text":"# Overall I found help from a guide via the RealPython [website](https://realpython.com/python-rock-paper-scissors/) which further aided in \n# my understanding of the exercise and functions that I need to use for the exercise.\n\n# this is the \"game.py\" file\nimport random \n\n# I'm introducing the game to the player and getting them excited to play.\n\"Hello. Welcome Player One. Let's play Rock, Paper, Scissors, Shoot!\"\nprint(\"Hello. Welcome Player One. Let's play Rock, Paper, Scissors, Shoot!\")\n\n\n# processed user inputs, validated user inputs, and set up the randomized computer selections for the game to run properly.\n# I need to make a list of all the options so that the user action returns the accurate result\n# Found the \"while True\" via the Real Python link above which helped me to understand and format the below code\nwhile True:\n user_action = input(\"Enter a choice (Rock, Paper, Scissors): \")\n user_action = user_action.lower()\n \n computer_choices = [\"rock\", \"paper\", \"scissors\"]\n if user_action not in computer_choices:\n print(\"Sorry I don't understand that entry. Please try again\")\n exit() # quit()\n computer_action = random.choice(computer_choices)\n print(f\"You chose {user_action}, \\ncomputer chose {computer_action}\")\n \n# Setting up all the various outcomes of the game between player and computer using if, elif, and else statements\n# Then determining the winner based on the player's and computer's actions\n if user_action == computer_action:\n print(f\"Whoa, you both selected {user_action}. Tie game!\")\n elif user_action == \"rock\":\n if computer_action == \"paper\":\n print(\"paper beats rock! You lost to the computer, uh oh!\")\n else:\n print(\"rock beats scissors! You win!! WOOHOO!\")\n elif user_action == \"paper\":\n if computer_action == \"scissors\":\n print(\"scissors beats paper! You lost to the computer, uh oh!\")\n else:\n print(\"paper beats rock! You win!! WOOHOO!\")\n elif user_action == \"scissors\":\n if computer_action == \"rock\":\n print(\"rock beats scissors! You lost to the computer, uh oh!\")\n else:\n print(\"scissors beats paper! You win! WOOHOO!\")\n\n# Asking the player if they want to play again or end the game.\n play_again = input(\"Would you like to try again? (y/n): \")\n if play_again != \"y\":\n print(\"Goodbye. Thanks for playing!\")\n break\n","repo_name":"Lrich14/Rock-Paper-Scissors-exercise","sub_path":"game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":2444,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"26832762850","text":"#!/usr/bin/env python\r\n# coding: utf-8\r\n\r\n# In[3]:\r\n\r\n\r\n# !pip install pyjwt\r\nimport jwt\r\nimport hashlib\r\nimport requests\r\nimport uuid\r\nfrom urllib.parse import urlencode, unquote\r\nimport importlib\r\nimport upbit_keys as keys\r\n\r\n\r\n# In[4]:\r\n\r\n\r\nserver_url = 'https://api.upbit.com'\r\n\r\n############################################################################\r\n# make your upbit key file(.py) on same directory\r\n#\r\n# ex) upbit_keys.py\r\n# access_key = ' -* your access key here *-'\r\n# secret_key = ' -* your secret key here *-'\r\n#\r\n############################################################################\r\n\r\n############################################################################\r\n# 업비트 자산조회\r\n# 입력 : 업비트 access key, secret key\r\n# 출력 : 나의 자산내용에 해당하는 json\r\n############################################################################\r\n\r\ndef asset_check(server_url = server_url):\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.get(server_url + '/v1/accounts', headers=headers)\r\n return res.json()\r\n\r\n\r\n############################################################################\r\n# 현금계좌 잔액 조회\r\n# 입력 : .\r\n# 출력 : (float) balance\r\n############################################################################\r\n\r\ndef get_balance():\r\n assets = asset_check()\r\n for i, asset in enumerate(assets):\r\n if(asset['currency'] == 'KRW'):\r\n return float(asset['balance'])\r\n\r\n \r\n \r\n############################################################################\r\n# 비트코인 조회\r\n# 입력 : .\r\n# 출력 : (float) volume\r\n# (int) -1 [코인이 없다면]\r\n############################################################################\r\n\r\ndef get_coin_balance():\r\n assets = asset_check()\r\n for i, asset in enumerate(assets):\r\n if(asset['currency'] == 'BTC'):\r\n return float(asset['balance'])\r\n elif (i == len(assets)-1):\r\n return -1\r\n\r\n\r\n \r\n############################################################################\r\n# 비트코인 구매가격 조회\r\n# 입력 : .\r\n# 출력 : (float) avg_buy_price\r\n############################################################################\r\n\r\ndef get_buy_price():\r\n assets = asset_check()\r\n for i, asset in enumerate(assets):\r\n if(asset['currency'] == 'BTC'):\r\n return float(asset['avg_buy_price'])\r\n elif (i == len(assets)-1):\r\n print('COIN_BALANCE_EMPTY')\r\n\r\n \r\n############################################################################\r\n# 주문 가능정보 조회\r\n# 입력 : (str) 코인 티커\r\n# ex) 'BTC'\r\n# 출력 : 입력한 코인 시장 정보 json\r\n############################################################################\r\n\r\ndef market_price(market, server_url = server_url):\r\n params = {\r\n 'market': 'KRW-' + market\r\n }\r\n query_string = unquote(urlencode(params, doseq=True)).encode(\"utf-8\")\r\n\r\n m = hashlib.sha512()\r\n m.update(query_string)\r\n query_hash = m.hexdigest()\r\n\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n 'query_hash': query_hash,\r\n 'query_hash_alg': 'SHA512',\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.get(server_url + '/v1/orders/chance', params=params, headers=headers)\r\n return res.json()\r\n\r\n\r\n\r\n############################################################################\r\n# 개별 주문 조회\r\n# 입력 : 주문 uuid\r\n# ex) '00000000-0000-0000-0000-000000000000'\r\n# 출력 : 입력한 주문에 대한 내용 json\r\n############################################################################\r\n\r\ndef order_check(order_id, server_url = server_url):\r\n params = {\r\n 'uuid': order_id\r\n }\r\n query_string = unquote(urlencode(params, doseq=True)).encode(\"utf-8\")\r\n\r\n m = hashlib.sha512()\r\n m.update(query_string)\r\n query_hash = m.hexdigest()\r\n\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n 'query_hash': query_hash,\r\n 'query_hash_alg': 'SHA512',\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.get(server_url + '/v1/order', params=params, headers=headers)\r\n return res.json()\r\n\r\n\r\n\r\n############################################################################\r\n# 주문 취소 접수\r\n# 미완\r\n#\r\n############################################################################\r\n\r\ndef cancel_order(server_url = server_url):\r\n params = {\r\n 'uuid': '00000000-0000-0000-0000-000000000000'\r\n }\r\n query_string = unquote(urlencode(params, doseq=True)).encode(\"utf-8\")\r\n\r\n m = hashlib.sha512()\r\n m.update(query_string)\r\n query_hash = m.hexdigest()\r\n\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n 'query_hash': query_hash,\r\n 'query_hash_alg': 'SHA512',\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.delete(server_url + '/v1/order', params=params, headers=headers)\r\n return res.json()\r\n\r\n\r\n\r\n############################################################################\r\n# 업비트 주문하기\r\n# 입력 : \r\n# 출력 : 매도, 매수 주문내용에 해당하는 json\r\n############################################################################\r\n\r\ndef coin_order(side_option, price = 'NULL', volume = 'NULL', server_url = server_url):\r\n params = {\r\n 'market': 'KRW-BTC',\r\n }\r\n \r\n # [side_option] ( 0 : 매수 , 1 : 매도 )\r\n if side_option == 0: # 매수\r\n params['side'] = 'bid'\r\n params['ord_type'] = 'price'\r\n params['price'] = price\r\n elif side_option == 1: # 매도\r\n params['side'] = 'ask'\r\n params['ord_type'] = 'market'\r\n params['volume'] = volume\r\n else:\r\n print('error form function \"coin_order\" wrong input in parameter \"side_option\"')\r\n return -1;\r\n \r\n query_string = unquote(urlencode(params, doseq=True)).encode(\"utf-8\")\r\n\r\n m = hashlib.sha512()\r\n m.update(query_string)\r\n query_hash = m.hexdigest()\r\n\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n 'query_hash': query_hash,\r\n 'query_hash_alg': 'SHA512',\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.post(server_url + '/v1/orders', json=params, headers=headers)\r\n return res.json()\r\n\r\n\r\n\r\n############################################################################\r\n# 시장가 매수\r\n# 입력 : int 매수금액\r\n# 출력 : 실패시 -1 반환\r\n# 성공시 주문내역에 해당하는 json 반환\r\n############################################################################\r\n\r\ndef ord_bid_price(price):\r\n if price < (int)(market_price('BTC')['market']['bid']['min_total']):\r\n print('최소 주문금액은 {} 입니다.'.format(market_price('BTC')['market']['bid']['min_total']))\r\n return -1;\r\n else:\r\n return coin_order(side_option = 0, price = price)\r\n\r\n \r\n \r\n############################################################################\r\n# 시장가 매도\r\n# 입력 : 매도 물량\r\n# default : 0 ( 전체 물량 매도 )\r\n# 출력 : 잘못된 입력이 들어왔을떄 -1 반환\r\n# 주문내역에 해당하는 json 반환\r\n############################################################################\r\n\r\ndef ord_ask_market(volume = 0):\r\n if(volume == 0): \r\n return coin_order(side_option = 1, volume = market_price('BTC')['ask_account']['balance'])\r\n elif(volume > 0):\r\n return coin_order(side_option = 1, volume = volume)\r\n else:\r\n return -1\r\n\r\n \r\n \r\n############################################################################\r\n# 전체 주문 조회\r\n# 입력 : .\r\n# 출력 : 전채 주문내역에 해당하는 json list \r\n# ex) [0]( 최근 ) ~ [...]( 먼 과거 )\r\n############################################################################\r\n\r\ndef orderList_check(server_url = server_url):\r\n params = {\r\n # [states option] ( 조회할 주문 상태 ) \r\n 'states[]': ['done', 'cancel']\r\n }\r\n query_string = unquote(urlencode(params, doseq=True)).encode(\"utf-8\")\r\n\r\n m = hashlib.sha512()\r\n m.update(query_string)\r\n query_hash = m.hexdigest()\r\n\r\n payload = {\r\n 'access_key': keys.access_key,\r\n 'nonce': str(uuid.uuid4()),\r\n 'query_hash': query_hash,\r\n 'query_hash_alg': 'SHA512',\r\n }\r\n\r\n jwt_token = jwt.encode(payload, keys.secret_key)\r\n authorization = 'Bearer {}'.format(jwt_token)\r\n headers = {\r\n 'Authorization': authorization,\r\n }\r\n\r\n res = requests.get(server_url + '/v1/orders', params=params, headers=headers)\r\n return res.json()\r\n\r\n\r\n\r\n############################################################################\r\n# 마지막 매수 주문 가져오기\r\n# 입력 : .\r\n# 출력 : 마지막 매수주문에 해당하는 json\r\n############################################################################\r\n\r\ndef get_last_bid():\r\n result = ''\r\n for i, order in enumerate(orderList_check()):\r\n if (order['side'] == 'bid'):\r\n result = orderList_check()[i]\r\n break;\r\n return result\r\n\r\n\r\n\r\n############################################################################\r\n# 마지막 매도 주문 가져오기\r\n# 입력 : .\r\n# 출력 : 마지막 매도주문에 해당하는 json\r\n############################################################################\r\n\r\ndef get_last_ask():\r\n result = ''\r\n for i, order in enumerate(orderList_check()):\r\n if (order['side'] == 'ask'):\r\n result = orderList_check()[i]\r\n break;\r\n return result\r\n\r\n\r\n# In[7]:\r\n\r\n\r\n# !jupyter nbconvert --to script request_order.ipynb\r\n\r\n","repo_name":"doongeon/Bitcoin-Trader-with-Python-Korean","sub_path":"request_order.py","file_name":"request_order.py","file_ext":"py","file_size_in_byte":10626,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"39377053476","text":"import numpy as np\n\n\ndef binary_search(search_algorithm):\n \"\"\"\n Binary search meant to be used as a decorator for algorithms OPT1 and OPT2\n\n :param search_algorithm: OPT1 or OPT2 algorithms\n :return:\n \"\"\"\n\n def bin_search(*args):\n low = 0\n high = len(args[1]) - 1\n\n tuple_ = (np.inf, None)\n\n while low <= high:\n mid = (high + low) // 2\n results = search_algorithm(args[0], args[1][mid], *args[2:])\n\n if results['valid']:\n tuple_ = (args[1][mid], results['r'])\n high = mid - 1\n else:\n low = mid + 1\n\n return tuple_[0], tuple_[1]\n\n return bin_search","repo_name":"mattrighetti/leiserson-retiming","sub_path":"algorithms/binary_search/binary_search.py","file_name":"binary_search.py","file_ext":"py","file_size_in_byte":692,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"18209077553","text":"import sys\nimport socket\nimport scrapy\nfrom Crypto.PublicKey import RSA\nfrom Crypto.Cipher import PKCS1_OAEP\nimport base64\n\"\"\"\nA network socket is an endpoint of an interprocess communication across a computer network. \nThe Python Standard Library has a module called socket which provides a low-level internet networking interface.\n\"\"\"\ntry:\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n print(\"Socket successfully created\")\nexcept socket.error as err:\n print(\"socket creation failed with error %s\" % err)\n\n# default port for socket\nport = 80\n\ntry:\n host_ip = socket.gethostbyname('www.google.com')\nexcept socket.gaierror:\n\n # this means could not resolve the host\n print(\"there was an error resolving the host\")\n sys.exit()\n\n# connecting to the server\ns.connect((host_ip, port))\n\nprint(\"Connected to google\")\n\n\ndef rsa(message):\n \"\"\"\n Generating new keys\n Generating a keypair may take a long time, depending on the number of bits required. \n The number of bits determines the cryptographic strength of the key, as well as the size of the message you can encrypt.\n \"\"\"\n key = RSA.generate(2048)\n private_key = key.exportKey('PEM')\n public_key = key.publickey().exportKey('PEM')\n \"\"\" Encrypting message using public key \"\"\"\n rsa_public_key = RSA.importKey(public_key)\n rsa_public_key = PKCS1_OAEP.new(rsa_public_key)\n encrypted_text = rsa_public_key.encrypt(message)\n encrypted_text_b64 = base64.b64encode(encrypted_text)\n print('encrypted message: {}'.format(encrypted_text_b64))\n \"\"\" Decrypting message using private key \"\"\"\n rsa_private_key = RSA.importKey(private_key)\n rsa_private_key = PKCS1_OAEP.new(rsa_private_key)\n decrypted_text = rsa_private_key.decrypt(encrypted_text)\n print('decrypted message: {}'.format(decrypted_text))\n\n\nmessage = input('Enter the message: ')\nmessage = str.encode(message)\nrsa(message)\n\ns.send(message)\ndata = s.recv(1024)\n\ns.close()\n","repo_name":"HollisLynch/bsi_crawler","sub_path":"sockets_connection.py","file_name":"sockets_connection.py","file_ext":"py","file_size_in_byte":1965,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"33799042573","text":"import tkinter as tk\nfrom PIL import ImageTk, Image\nfrom db.sqlite import insert_into_table, update_row\nfrom widgets.dialog import DialogBox\nimport datetime\nimport os\nfrom tkinter.tix import ScrolledGrid\n\n\nclass BaseFrame(tk.Frame):\n def __init__(self, root, connection, *args, **kwargs):\n super().__init__(root, bg=\"#ffffff\")\n self.tasks = None\n self.connection = connection\n self.cursor = self.connection.cursor()\n self.all_tasks = {}\n self.img = self._parse_background_image()\n self.tasks_frame = tk.Frame(self, bg=\"#ffffff\")\n self.no_tasks_frame = tk.Frame(self, bg=\"#ffffff\")\n self._configure_entry_field()\n self.root = root\n self.root.bind('', self.on_submit)\n self._dialog_box = None\n\n def _parse_background_image(self):\n path = os.path.dirname(os.path.abspath(__file__))\n image_path = os.path.join(path, \"images/check.png\")\n im = Image.open(image_path)\n img = im.resize((156, 128), Image.ANTIALIAS)\n img = ImageTk.PhotoImage(img)\n\n return img\n\n def _configure_entry_field(self):\n self.entry = tk.Text(self, bd=0.5, width=74, height=1, spacing1=7, spacing3=7, fg=\"#585858\")\n self.entry.insert('1.0', 'Task Title')\n self.entry.grid(row=0, column=1, pady=(30, 30))\n self.entry.bind(\"\", self.on_entry_clicked)\n\n def on_entry_clicked(self, event):\n self.entry.config(highlightcolor=\"#585858\")\n if self.entry.get('1.0', '10.0') == 'Task Title\\n':\n self.entry.delete(\"1.0\", \"10.0\")\n\n def set_default_background(self):\n panel = tk.Label(self.no_tasks_frame, image=self.img, bd=0)\n no_tasks = tk.Label(self.no_tasks_frame, text='No Tasks Found', bg=\"#ffffff\", bd=0, fg=\"#585858\")\n add_tasks = tk.Label(self.no_tasks_frame, text='You can add tasks using the + above', bg=\"#ffffff\", bd=0, fg=\"#585858\")\n panel.grid(columnspan=2, row=1, rowspan=2, pady=(60, 30))\n no_tasks.grid(columnspan=2, row=4)\n add_tasks.grid(columnspan=2, row=5)\n\n def on_submit(self, e):\n self.focus()\n data = self.entry.get('0.0', tk.END).strip('\\n')\n\n if data and data != 'Task Title':\n insert_into_table(self.cursor, data, self.connection)\n self.no_tasks_frame.grid_remove()\n self.add_task_to_list()\n self.entry.delete('1.0', tk.END)\n self.entry.insert('1.0', 'Task Title')\n\n def add_task_to_list(self):\n id = self.cursor.lastrowid\n task = self.cursor.execute(\"SELECT * FROM todo WHERE id = %s LIMIT 1\" % id).fetchone()\n if not self.tasks:\n self.tasks_frame.grid(row=1, column=1, sticky=tk.NSEW)\n self.tasks_frame.columnconfigure(1, weight=1)\n self.render_task(task, len(self.all_tasks), self.tasks_frame)\n\n def handle_edit_action(self, task_id):\n DialogBox.root = self.root\n self._dialog_box = DialogBox(save=self.on_edit_save, delete=self.on_edit_delete, task_id=task_id)\n\n def render_task(self, task, index, target_frame):\n # check_is_done = tk.Button(self, text='✎', bg=\"#ffffff\",\n # bd=0, relief=\"raised\", command=lambda: self.handle_edit_action(task[\"id\"]))\n # check_is_done.grid(row=index + 1, column=0, sticky=tk.W)\n task_fields = tk.Text(target_frame, bd=2, width=80, height=1, spacing1=7, spacing3=7, fg=\"#585858\")\n task_fields.task_id = task['id']\n task_fields.insert('1.0', task['task'])\n task_fields.config(state=tk.DISABLED)\n task_fields.grid(row=index + 1, column=1, pady=(10, 10))\n extra_options = tk.Button(task_fields, text='✎', bg=\"#ffffff\",\n bd=0, relief=\"raised\", command=lambda: self.handle_edit_action(task[\"id\"]))\n extra_options.config(highlightbackground=\"#ffffff\", cursor='hand1')\n extra_options.pack(padx=(553, 0))\n self.all_tasks.update({task['id']: task_fields})\n\n def _calculate_scheduled_date(self, strict_opt, custom_opt):\n date = None\n current_date = datetime.date.today()\n if custom_opt:\n date = custom_opt.replace(\"/\", \"-\")\n else:\n if strict_opt == 0:\n date = current_date.strftime('%Y-%m-%d')\n elif strict_opt == 1:\n date = (current_date + datetime.timedelta(days=1)).strftime('%Y-%m-%d')\n elif strict_opt == 2:\n date = (current_date + datetime.timedelta(7-current_date.weekday())).strftime('%Y-%m-%d')\n\n return date\n\n def on_edit_save(self):\n data = self._dialog_box.get_normalized_values()\n self._dialog_box.window_close()\n rowid = data.pop(\"task_id\")\n scheduled_date = self._calculate_scheduled_date(data[\"button_option\"], data[\"custom_date\"])\n priority = data.pop(\"priority\")\n data = dict(note=data.pop(\"note\"),\n scheduled_date=scheduled_date,\n todo_date=scheduled_date,\n priority=priority\n )\n update_row(cursor=self.cursor, connection=self.connection, data=data, rowid=rowid)\n self.respawn_slaves()\n\n\n def on_edit_delete(self):\n data = self._dialog_box.get_normalized_values()\n self._dialog_box.window_close()\n rowid = data.pop(\"task_id\")\n self.cursor.execute(\"DELETE FROM todo WHERE id=?\", (rowid, ))\n self.connection.commit()\n self.all_tasks[rowid].grid_remove()\n\n def respawn_slaves(self):\n for slave in self.tasks_frame.grid_slaves():\n slave.destroy()\n self.tasks_frame.grid_remove()\n self.fetch_todos()\n self.fill_tasks_frame()\n\n def list_all_tasks(self):\n for i, task in enumerate(self.tasks):\n self.render_task(task, i, self.tasks_frame)\n\n def fill_tasks_frame(self):\n if self.tasks:\n self.tasks_frame.grid(row=1, column=1, sticky=tk.NSEW)\n self.tasks_frame.columnconfigure(1, weight=1)\n self.list_all_tasks()\n else:\n self.no_tasks_frame.grid(row=1, column=1, sticky=tk.NSEW)\n self.no_tasks_frame.columnconfigure(1, weight=1)\n self.set_default_background()\n\n def fetch_todos(self):\n pass\n\n","repo_name":"Toshe1991/TodoApp","sub_path":"widgets/base.py","file_name":"base.py","file_ext":"py","file_size_in_byte":6326,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"28192732880","text":"# you can write to stdout for debugging purposes, e.g.\n# print(\"this is a debug message\")\n\ndef solution(X):\n string_X = str(X)\n if string_X[0] == '-':\n sign = '-'\n unsigned_X = string_X[1:]\n new_X = int(''.join(sign + unsigned_X[::-1]))\n else:\n sign = ''\n new_X = int(''.join(sign + string_X[::-1]))\n return new_X","repo_name":"danrasband/coding-experiment-reviews","sub_path":"responses/A5QMTE-D2P/2_reverse_digits.py","file_name":"2_reverse_digits.py","file_ext":"py","file_size_in_byte":360,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"20764771146","text":"class Solution(object):\n def isValid(self, s):\n \"\"\"\n :type s: str\n :rtype: bool\n \"\"\"\n map = { \")\":\"(\" , \"]\":\"[\" , \"}\":\"{\"} #maps close to open parenthesis\n stack = [] #stack of open parenthesis\n \n for bracket in s:\n if bracket in map: # If its an CLOSING bracket (cause our keys to map are closing brackets) \n \n if stack and (stack[-1] == map[bracket]): # If top of stack, which is the opening bracket matches the closing bracket correctly\n stack.pop() # Pop opening bracket form stack\n \n else: # If they do not match correctly\n return False\n \n else: #If its an OPENING bracket\n stack.append(bracket)\n \n # if stack is empty it means all of our brackets matched correctly so we return true, otherwise we return false\n \n if not stack:\n return True\n else:\n return False\n","repo_name":"elenazavala/leetcode","sub_path":"python/valid-parentheses.py","file_name":"valid-parentheses.py","file_ext":"py","file_size_in_byte":1042,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"10208471480","text":"# \n# 4 3 1\n\ndef vlozh(elem, lvl=1):\n if elem.tag == 'cube' and elem.attrib['color'] in dd.keys():\n dd[elem.attrib['color']] += lvl\n # print(lvl, elem.tag, elem.attrib['color'])\n for child in elem.getchildren():\n vlozh(child, lvl+1)\n\n\nfrom xml.etree import ElementTree as ET\n\ncd = ['red', 'green', 'blue']\ndd = {'red': 0, 'green': 0, 'blue': 0}\nroot = ET.fromstring(input())\n\n# print(root)\n# print(root.tag, root.attrib)\n# for child in root:\n# print(child.tag, child.attrib)\n# for element in root.iter(\"cube\"):\n# print(element.attrib)\n\nvlozh(root)\nprint(dd['red'], dd['green'], dd['blue'])","repo_name":"Kaermor/stepik-course512","sub_path":"3-7.py","file_name":"3-7.py","file_ext":"py","file_size_in_byte":726,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12553866343","text":"from .. import bp\nfrom flask import render_template, request, redirect, url_for\nfrom jeec_brain.finders.companies_finder import CompaniesFinder\nfrom jeec_brain.finders.auctions_finder import AuctionsFinder\nfrom jeec_brain.handlers.auctions_handler import AuctionsHandler\nfrom jeec_brain.apps.auth.wrappers import allowed_roles\nfrom jeec_brain.values.api_error_value import APIErrorValue\n\n\n# Auction routes\n@bp.route('/auctions', methods=['GET'])\n@allowed_roles(['admin'])\ndef auctions_dashboard():\n auctions_list = AuctionsFinder.get_all()\n\n for auction in auctions_list:\n auction.highest_bid = AuctionsFinder.get_auction_highest_bid(auction)\n \n if auctions_list is None:\n error = 'No results found'\n return render_template('admin/auctions/auctions_dashboard.html', auctions=None, error=error)\n\n return render_template('admin/auctions/auctions_dashboard.html', auctions=auctions_list, error=None)\n\n\n@bp.route('/new-auction', methods=['GET'])\n@allowed_roles(['admin'])\ndef add_auction_dashboard():\n return render_template('admin/auctions/add_auction.html', \\\n error=None)\n\n\n@bp.route('/new-auction', methods=['POST'])\n@allowed_roles(['admin'])\ndef create_auction():\n name = request.form.get('name')\n description = request.form.get('description')\n is_open = request.form.get('is_open')\n closing_date = request.form.get('closing_date')\n\n try:\n minimum_value = float(request.form.get('minimum_value'))\n except:\n return 'Invalid minimum value inserted', 404\n\n if is_open == 'True':\n is_open = True\n else:\n is_open = False\n\n # create new auction\n auction = AuctionsHandler.create_auction(\n name=name,\n description=description,\n is_open=is_open,\n minimum_value=minimum_value,\n closing_date=closing_date\n )\n\n if auction is None:\n return render_template('admin/auctions/add_auction.html', \\\n error=\"Failed to create auction!\")\n\n return redirect(url_for('admin_api.auctions_dashboard'))\n\n\n@bp.route('/auctions/', methods=['GET'])\n@allowed_roles(['admin'])\ndef get_auction(auction_external_id):\n auction = AuctionsFinder.get_auction_by_external_id(auction_external_id)\n\n if auction is None:\n error = 'Non existant auction'\n return render_template('admin/auctions/auctions_dashboard.html', auctions=None, error=error)\n\n return render_template('admin/auctions/update_auction.html', \\\n auction=auction, \\\n error=None)\n\n\n@bp.route('/auctions/', methods=['POST'])\n@allowed_roles(['admin'])\ndef update_auction(auction_external_id):\n auction = AuctionsFinder.get_auction_by_external_id(auction_external_id)\n\n if auction is None:\n return APIErrorValue('Couldnt find auction').json(500)\n\n name = request.form.get('name')\n description = request.form.get('description')\n closing_date = request.form.get('closing_date')\n\n try:\n minimum_value = float(request.form.get('minimum_value'))\n except:\n return APIErrorValue('Wrong value format input').json(400)\n\n is_open = request.form.get('is_open')\n\n if is_open == 'True':\n is_open = True\n else:\n is_open = False\n\n updated_auction = AuctionsHandler.update_auction(\n auction=auction,\n name=name,\n description=description,\n minimum_value=minimum_value,\n is_open=is_open,\n closing_date=closing_date\n )\n \n if updated_auction is None:\n return render_template('admin/auctions/update_auction.html', \\\n auction=auction, \\\n error=\"Failed to update auction!\")\n\n return redirect(url_for('admin_api.auctions_dashboard'))\n\n\n@bp.route('/auctions//delete', methods=['GET'])\n@allowed_roles(['admin'])\ndef delete_auction(auction_external_id):\n auction = AuctionsFinder.get_auction_by_external_id(auction_external_id)\n\n if auction is None:\n return APIErrorValue('Couldnt find auction').json(500)\n \n if AuctionsHandler.delete_auction(auction):\n return redirect(url_for('admin_api.auctions_dashboard'))\n\n else:\n return render_template('admin/auctions/update_aution.html', auction=auction, error=\"Failed to delete auction!\")\n\n\n# Members management\n@bp.route('/auctions//participants', methods=['GET'])\n@allowed_roles(['admin'])\ndef auction_participants_dashboard(auction_external_id):\n auction = AuctionsFinder.get_auction_by_external_id(auction_external_id)\n\n if auction is None:\n return APIErrorValue('Couldnt find auction').json(400)\n\n not_participants = AuctionsFinder.get_not_participants(auction)\n\n if len(auction.participants) == 0:\n error = 'No results found'\n return render_template('admin/auctions/auction_participants_dashboard.html', auction=auction, not_participants=not_participants, error=error)\n\n return render_template('admin/auctions/auction_participants_dashboard.html', auction=auction, not_participants=not_participants, error=None)\n\n\n@bp.route('/auctions//add-participant', methods=['POST'])\n@allowed_roles(['admin'])\ndef add_auction_participant(auction_external_id):\n auction = AuctionsFinder.get_auction_by_external_id(auction_external_id)\n\n if auction is None:\n return APIErrorValue('Couldnt find auction').json(400)\n\n company_external_id = request.form.get('company_external_id')\n\n if company_external_id is None:\n return redirect(url_for('admin_api.auction_participants_dashboard', auction_external_id=auction_external_id))\n\n company = CompaniesFinder.get_from_external_id(company_external_id)\n\n if company is None:\n return APIErrorValue('Couldnt find company').json(400)\n\n AuctionsHandler.add_auction_participant(auction, company)\n \n return redirect(url_for('admin_api.auction_participants_dashboard', auction_external_id=auction_external_id))\n","repo_name":"jose-correia/brain","sub_path":"jeec_brain/apps/admin_api/auctions/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":6015,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"28609879155","text":"import math\r\nfrom math import ceil, floor, inf\r\n\r\n\r\ndef distance(p1, p2):\r\n return math.sqrt((p1[0] - p2[0])**2 + (p1[1] - p2[1])**2)\r\n\r\n\r\n\r\ndef shortest_d(input_x, input_y):\r\n\r\n if len(input_x) <= 3:\r\n s_dis = inf\r\n for i in range(len(input_x)-1):\r\n dis_tmp = distance(input_x[i], input_x[i+1])\r\n if dis_tmp < s_dis:\r\n s_dis = dis_tmp\r\n return round(s_dis,6)\r\n else:\r\n #Divide\r\n half_n = len(input_x)//2\r\n left_points = input_x[:half_n]\r\n right_points = input_x[half_n:]\r\n\r\n mid_x = input_x[half_n][0] #Line\r\n \r\n Qy,Ry = [], []\r\n for point in input_y:\r\n if point[0] < mid_x:\r\n Qy.append(point)\r\n else:\r\n Ry.append(point)\r\n \r\n #Conquer\r\n l_s_t = shortest_d(left_points, Qy)\r\n r_s_t = shortest_d(right_points, Ry)\r\n\r\n #Combine\r\n delta = min(l_s_t, r_s_t) \r\n # x_bar = left_points[-1][0]\r\n\r\n #Remove the data >delta from Line\r\n #candidates in delta region\r\n candidates = []\r\n for point in input_y:\r\n if abs(point[0] - mid_x) < delta:\r\n candidates.append(point)\r\n\r\n\r\n\r\n #1 + 7\r\n for i in range(len(candidates)-1):\r\n j = i + 1\r\n while j < min(len(candidates), i + 7) and abs(candidates[i][1] - candidates[j][1]) < delta:\r\n dis_tmp = distance(candidates[i], candidates[j])\r\n if dis_tmp < delta:\r\n delta = dis_tmp\r\n j += 1\r\n\r\n return round(delta,6)\r\n\r\n\r\ndef main():\r\n n_sets = int(input())\r\n if n_sets>= 0:\r\n for n in range(n_sets):\r\n n_points = int(input())\r\n points = []\r\n for i in range(n_points):\r\n point = [float(v) for v in input().split()]\r\n\r\n points.append(point)\r\n\r\n #Sort\r\n input_x = sorted(points , key=lambda k: k[0]) \r\n input_y = sorted(points , key=lambda k: k[1]) \r\n # input_x, input_y = first_sort(points)\r\n output = shortest_d(input_x, input_y)\r\n print('%6f' % (output))\r\n\r\n\r\nmain()\r\n\r\n\r\n\r\n","repo_name":"xup6YJ/Algorithm","sub_path":"Closest Pair.py","file_name":"Closest Pair.py","file_ext":"py","file_size_in_byte":2237,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"13065802924","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\nModule with logic for currency value object\n\n\"\"\"\n\n__author__ = 'Samir Adrik'\n__email__ = 'samir.adrik@gmail.com'\n\nfrom source.util import InvalidCurrencyError, Assertor, Tracking\n\nfrom .value import Value\n\n\nclass Currency(Value):\n \"\"\"\n Implementation of currency value object\n\n \"\"\"\n\n @Tracking\n def validate_currency(self, currency: str):\n \"\"\"\n method for validating currency string\n\n Parameters\n ----------\n currency : str\n string to be validated\n\n \"\"\"\n if currency not in [\"kr\"]:\n raise InvalidCurrencyError(\n \"'{}' is an unsupported currency\".format(currency))\n\n def __init__(self, currency: str = \"kr\"):\n \"\"\"\n constructor / instantiating of class\n\n Parameters\n ----------\n currency : str\n currency string\n\n \"\"\"\n try:\n super().__init__()\n Assertor.assert_data_types([currency], [str])\n self.validate_currency(currency)\n self._currency = currency.lower()\n except Exception as currency_exception:\n raise currency_exception\n\n @property\n def currency(self):\n \"\"\"\n currency getter\n\n Returns\n -------\n out : str\n active currency\n\n \"\"\"\n return self._currency\n\n @currency.setter\n def currency(self, new_currency: str):\n \"\"\"\n currency setter\n\n Parameters\n ----------\n new_currency : str\n new currency to be set\n\n \"\"\"\n Assertor.assert_data_types([new_currency], [str])\n self.validate_currency(new_currency)\n self._currency = new_currency.lower()\n","repo_name":"seemir/stressa","sub_path":"source/domain/currency.py","file_name":"currency.py","file_ext":"py","file_size_in_byte":1793,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4751768816","text":"import configparser\r\nimport getopt\r\nimport re\r\nimport sys\r\n\r\nfrom urllib.request import Request, urlopen\r\n\r\nNONLETTERS = \"[^a-zA-ZąćęłńóśźżĄĆĘŁŃÓŚŹŻ]\"\r\n\r\n\r\ndef nctm_to_xml(nctm_path):\r\n with open(nctm_path, 'r', encoding=\"utf-8\") as f:\r\n xml = '\\n'\r\n timestamps = []\r\n for line in f:\r\n line_timestamps = []\r\n for timestamp in re.findall('\\|.*?\\|', line):\r\n line_timestamps.append(timestamp[1:-1])\r\n timestamps.append(line_timestamps)\r\n simplified_line = re.sub('\\|.*?\\|', '', line)\r\n xml += '' + re.sub(\"&\", \"&\", simplified_line)[:-1] + '\\n'\r\n xml += ''\r\n return xml.encode(), timestamps\r\n\r\n\r\ndef ask_gram_textphrases(xml):\r\n config = configparser.ConfigParser()\r\n config.read(\"params.ini\")\r\n req = Request(\r\n url=config[\"Gram\"][\"url\"] + \"textphrases\",\r\n data=xml,\r\n headers={'Content-Type': 'application/xml', \"Accept\": 'application/json'}\r\n )\r\n return urlopen(req).read()\r\n\r\n\r\ndef get_joined_phrase(phrases):\r\n if type(phrases) != list:\r\n phrases = [phrases]\r\n joined_phrase = \"\"\r\n for inner_phrase in phrases:\r\n if inner_phrase[\"@type\"] == \"whitePhrase\":\r\n joined_phrase += inner_phrase[\"whites\"]\r\n elif inner_phrase[\"@type\"] == \"atomPhrase\":\r\n joined_phrase += inner_phrase[\"atom\"][\"text\"]\r\n elif inner_phrase[\"@type\"] == \"annotatedPhrase\":\r\n joined_phrase += get_joined_phrase(inner_phrase[\"phrases\"])\r\n else:\r\n raise Exception(\"Unknown phrase type\")\r\n return joined_phrase\r\n\r\n\r\ndef get_annotated_phrases(gram_response, timestamps):\r\n response_dict = eval(gram_response) if gram_response != b'null' else {\"paragraphs\": []}\r\n result = {}\r\n if type(response_dict[\"paragraphs\"]) != list:\r\n response_dict[\"paragraphs\"] = [response_dict[\"paragraphs\"]]\r\n for paragraph, line_timestamps in zip(response_dict[\"paragraphs\"], timestamps):\r\n remove_next = True\r\n if type(paragraph[\"phrases\"]) == list:\r\n for phrase in paragraph[\"phrases\"]:\r\n if phrase[\"@type\"] == \"atomPhrase\":\r\n if re.sub(NONLETTERS, \"\", phrase[\"atom\"][\"text\"]):\r\n if remove_next:\r\n del line_timestamps[:2]\r\n else:\r\n if phrase[\"atom\"][\"rightWhite\"] == \"true\" or phrase[\"atom\"][\"leftWhite\"] == \"true\":\r\n remove_next = True\r\n else:\r\n if phrase[\"atom\"][\"rightWhite\"] == \"false\" and phrase[\"atom\"][\"leftWhite\"] == \"false\":\r\n remove_next = False\r\n if phrase[\"atom\"][\"rightWhite\"] == \"true\" and phrase[\"atom\"][\"leftWhite\"] == \"true\":\r\n del line_timestamps[:2]\r\n if phrase[\"@type\"] == \"annotatedPhrase\":\r\n joined_phrase = get_joined_phrase(phrase[\"phrases\"])\r\n count_phrases = len(joined_phrase.split())\r\n if phrase[\"annotation\"][\"phraseType\"][\"$\"] == \"Person\":\r\n phrase[\"start\"], phrase[\"end\"] = line_timestamps[0], line_timestamps[2*count_phrases - 1]\r\n result[str((line_timestamps[0], line_timestamps[2*count_phrases - 1]))] = joined_phrase\r\n del line_timestamps[:2*count_phrases]\r\n return result\r\n\r\n\r\nif __name__ == \"__main__\":\r\n input_nctm = \"\"\r\n try:\r\n opts, args = getopt.getopt(sys.argv[1:], \"i:\", [\"input_nctm=\"])\r\n except getopt.GetoptError:\r\n print(\"Usage: ask_gram_nctm.py -i test.txt\")\r\n sys.exit(2)\r\n for opt, arg in opts:\r\n if opt == \"-i\":\r\n input_nctm = arg\r\n\r\n request, timestamps = nctm_to_xml(input_nctm)\r\n response = ask_gram_textphrases(request)\r\n print(get_annotated_phrases(response, timestamps))\r\n","repo_name":"smalec/ner-for-asr","sub_path":"ask_gram_nctm.py","file_name":"ask_gram_nctm.py","file_ext":"py","file_size_in_byte":4059,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4263114182","text":"from unittest import mock\n\nimport factory\nimport pytest\nfrom dateutil.parser import parse as dateutil_parse\nfrom django.conf import settings\n\nfrom datahub.company.test.factories import AdviserFactory\nfrom datahub.core.exceptions import APIConflictException\nfrom datahub.omis.order.constants import OrderStatus\nfrom datahub.omis.order.test.factories import (\n OrderFactory,\n OrderPaidFactory,\n OrderWithAcceptedQuoteFactory,\n)\nfrom datahub.omis.payment.constants import PaymentGatewaySessionStatus, PaymentMethod\nfrom datahub.omis.payment.govukpay import govuk_url, GOVUKPayAPIException\nfrom datahub.omis.payment.models import Payment, PaymentGatewaySession\nfrom datahub.omis.payment.test.factories import PaymentGatewaySessionFactory\n\n\n# mark the whole module for db use\npytestmark = pytest.mark.django_db\n\n\nclass TestPaymentManager:\n \"\"\"Tests for the Payment Manager.\"\"\"\n\n @mock.patch('datahub.omis.payment.managers.generate_datetime_based_reference')\n def test_create_from_order(\n self,\n mocked_generate_datetime_based_reference,\n ):\n \"\"\"Test that Payment.objects.create_from_order creates a payment.\"\"\"\n mocked_generate_datetime_based_reference.return_value = '201702010004'\n\n order = OrderPaidFactory()\n by = AdviserFactory()\n attrs = {\n 'transaction_reference': 'lorem ipsum',\n 'amount': 1001,\n 'received_on': dateutil_parse('2017-01-01').date(),\n }\n payment = Payment.objects.create_from_order(\n order=order, by=by, attrs=attrs,\n )\n\n payment.refresh_from_db()\n assert payment.reference == '201702010004'\n assert payment.created_by == by\n assert payment.order == order\n assert payment.transaction_reference == attrs['transaction_reference']\n assert payment.additional_reference == ''\n assert payment.amount == attrs['amount']\n assert payment.received_on == attrs['received_on']\n\n\nclass TestPaymentGatewaySessionManager:\n \"\"\"Tests for the Payment Gateway Session Manager.\"\"\"\n\n def test_create_first_session_from_order(self, requests_mock, monkeypatch):\n \"\"\"\n Test the successful creation of the first payment gateway session for an order.\n \"\"\"\n monkeypatch.setattr(\n 'uuid.uuid4',\n mock.Mock(return_value='0123abcd-0000-0000-0000-000000000000'),\n )\n\n # mock request\n govuk_payment_id = '123abc123abc123abc123abc12'\n govuk_payments_url = govuk_url('payments')\n requests_mock.post(\n govuk_payments_url,\n status_code=201,\n json={\n 'state': {'status': 'created', 'finished': False},\n 'payment_id': govuk_payment_id,\n '_links': {\n 'next_url': {\n 'href': 'https://payment.example.com/123abc',\n 'method': 'GET',\n },\n },\n },\n )\n\n assert PaymentGatewaySession.objects.count() == 0\n\n # call method\n adviser = AdviserFactory()\n order = OrderWithAcceptedQuoteFactory()\n session = PaymentGatewaySession.objects.create_from_order(\n order=order,\n attrs={'created_by': adviser},\n )\n\n # check session\n assert session.order == order\n assert session.status == PaymentGatewaySessionStatus.CREATED\n assert session.govuk_payment_id == govuk_payment_id\n assert session.created_by == adviser\n\n assert PaymentGatewaySession.objects.count() == 1\n\n # check mocked request\n assert requests_mock.call_count == 1\n assert requests_mock.request_history[-1].url == govuk_payments_url\n assert requests_mock.request_history[-1].json() == {\n 'amount': order.total_cost,\n 'reference': f'{order.reference}-0123ABCD',\n 'description': settings.GOVUK_PAY_PAYMENT_DESCRIPTION.format(\n reference=order.reference,\n ),\n 'return_url': settings.GOVUK_PAY_RETURN_URL.format(\n public_token=order.public_token,\n session_id=session.pk,\n ),\n }\n\n def test_create_cancels_other_sessions(self, requests_mock):\n \"\"\"\n Test that creating a new payment gateway session cancels\n the other ongoing sessions and GOV.UK payments.\n\n Given:\n - ongoing session 1\n - ongoing session 2\n - failed session 3\n\n Calling .create_from_order should:\n - cancel the GOV.UK payment related to session 1\n - update the payment gateway session 1 status to 'cancelled'\n\n - cancel the GOV.UK payment related to session 2\n - update the payment gateway session 2 status to 'cancelled'\n\n - start a new GOV.UK payment\n - create a payment gateway session related to it\n \"\"\"\n order = OrderWithAcceptedQuoteFactory()\n existing_data = PaymentGatewaySessionFactory.create_batch(\n 3,\n order=order,\n status=factory.Iterator([\n PaymentGatewaySessionStatus.CREATED,\n PaymentGatewaySessionStatus.STARTED,\n PaymentGatewaySessionStatus.FAILED,\n ]),\n )\n\n # mock GOV.UK requests used to:\n # - refresh the payment gateway sessions\n # - cancel the GOV.UK payments\n # - refresh the payment gateway sessions again after the cancellation\n for session in existing_data:\n requests_mock.get(\n govuk_url(f'payments/{session.govuk_payment_id}'),\n [\n # this is for the initial refresh\n {\n 'status_code': 200,\n 'json': {'state': {'status': session.status}},\n },\n # this is for the second refresh after cancelling\n {\n 'status_code': 200,\n 'json': {'state': {'status': 'cancelled'}},\n },\n ],\n )\n requests_mock.post(\n govuk_url(f'payments/{session.govuk_payment_id}/cancel'),\n status_code=204,\n )\n\n # mock GOV.UK request used to create a new payment session\n govuk_payment_id = '123abc123abc123abc123abc12'\n requests_mock.post(\n govuk_url('payments'),\n status_code=201,\n json={\n 'state': {'status': 'created', 'finished': False},\n 'payment_id': govuk_payment_id,\n '_links': {\n 'next_url': {\n 'href': 'https://payment.example.com/123abc',\n 'method': 'GET',\n },\n },\n },\n )\n\n assert PaymentGatewaySession.objects.count() == 3\n\n session = PaymentGatewaySession.objects.create_from_order(order=order)\n\n # check sessions cancelled\n for existing_session in existing_data[:-1]:\n existing_session.refresh_from_db()\n assert existing_session.status == PaymentGatewaySessionStatus.CANCELLED\n\n assert PaymentGatewaySession.objects.count() == 4\n\n # check session record created\n session.refresh_from_db()\n assert session.govuk_payment_id == govuk_payment_id\n\n # check mocked requests:\n # 2 refresh / 2 cancel - 2 refresh / 1 create\n assert requests_mock.call_count == (2 + 2 + 2 + 1)\n assert requests_mock.request_history[-1].json() == {\n 'amount': order.total_cost,\n 'reference': f'{order.reference}-{str(session.id)[:8].upper()}',\n 'description': settings.GOVUK_PAY_PAYMENT_DESCRIPTION.format(\n reference=order.reference,\n ),\n 'return_url': settings.GOVUK_PAY_RETURN_URL.format(\n public_token=order.public_token,\n session_id=session.id,\n ),\n }\n\n @pytest.mark.parametrize(\n 'disallowed_status', (\n OrderStatus.DRAFT,\n OrderStatus.QUOTE_AWAITING_ACCEPTANCE,\n OrderStatus.PAID,\n OrderStatus.COMPLETE,\n OrderStatus.CANCELLED,\n ),\n )\n def test_exception_if_order_in_disallowed_status(self, disallowed_status):\n \"\"\"\n Test that if the order is not in one of the allowed statuses, the method raises\n APIConflictException.\n \"\"\"\n assert PaymentGatewaySession.objects.count() == 0\n\n order = OrderFactory(status=disallowed_status)\n\n with pytest.raises(APIConflictException):\n PaymentGatewaySession.objects.create_from_order(order)\n\n # test no session created\n assert PaymentGatewaySession.objects.count() == 0\n\n def test_exception_if_refresh_updates_order_status_to_paid(self, requests_mock):\n \"\"\"\n Test that if the system is not up-to-date, the order is in quote_accepted\n but the GOV.UK payment happens, the method triggers a check on existing\n sessions, realises that one finished successfully and records the payment\n marking the order as 'paid'.\n For this reason, the method raises APIConflictException as no other payment can be started.\n \"\"\"\n # set up db\n order = OrderWithAcceptedQuoteFactory()\n existing_session = PaymentGatewaySessionFactory(\n order=order,\n status=PaymentGatewaySessionStatus.STARTED,\n )\n\n # mock GOV.UK requests used to refresh the payment session,\n # GOV.UK Pay says that the payment completed successfully\n requests_mock.get(\n govuk_url(f'payments/{existing_session.govuk_payment_id}'),\n status_code=200,\n json={\n 'amount': order.total_cost,\n 'state': {'status': 'success'},\n 'email': 'email@example.com',\n 'created_date': '2018-02-13T14:56:56.734Z',\n 'reference': '12345',\n 'card_details': {\n 'last_digits_card_number': '1111',\n 'cardholder_name': 'John Doe',\n 'expiry_date': '01/20',\n 'billing_address': {\n 'line1': 'line 1 address',\n 'line2': 'line 2 address',\n 'postcode': 'SW1A 1AA',\n 'city': 'London',\n 'country': 'GB',\n },\n 'card_brand': 'Visa',\n },\n },\n )\n\n with pytest.raises(APIConflictException):\n PaymentGatewaySession.objects.create_from_order(order)\n\n # check session record\n existing_session.refresh_from_db()\n assert existing_session.status == PaymentGatewaySessionStatus.SUCCESS\n\n # check order and payment\n order.refresh_from_db()\n assert order.status == OrderStatus.PAID\n\n assert Payment.objects.count() == 1\n payment = Payment.objects.first()\n\n assert payment.amount == order.total_cost\n assert payment.method == PaymentMethod.CARD\n assert payment.received_on == dateutil_parse('2018-02-13').date()\n assert payment.transaction_reference == '12345'\n assert payment.cardholder_name == 'John Doe'\n assert payment.billing_address_1 == 'line 1 address'\n assert payment.billing_address_2 == 'line 2 address'\n assert payment.billing_address_town == 'London'\n assert payment.billing_address_postcode == 'SW1A 1AA'\n assert payment.billing_address_country == 'GB'\n assert payment.billing_email == 'email@example.com'\n assert payment.card_brand == 'Visa'\n\n @pytest.mark.parametrize('govuk_status_code', (400, 401, 422, 500))\n def test_exception_if_govuk_pay_errors_when_creating(\n self, govuk_status_code, requests_mock,\n ):\n \"\"\"\n Test that if GOV.UK Pay errors whilst creating a new payment, the method raises\n GOVUKPayAPIException.\n\n Possible GOV.UK Pay errors:\n - 400 - BAD REQUEST\n - 401 - UNAUTHORIZED\n - 422 - UNPROCESSABLE ENTITY\n - 500 - INTERNAL SERVER ERROR\n \"\"\"\n requests_mock.post(\n govuk_url('payments'),\n status_code=govuk_status_code,\n )\n\n assert PaymentGatewaySession.objects.count() == 0\n\n order = OrderWithAcceptedQuoteFactory()\n\n with pytest.raises(GOVUKPayAPIException):\n PaymentGatewaySession.objects.create_from_order(order)\n\n assert PaymentGatewaySession.objects.count() == 0\n\n @pytest.mark.parametrize('govuk_status_code', (400, 401, 404, 409, 500))\n def test_exception_if_govuk_pay_errors_when_cancelling(\n self, govuk_status_code, requests_mock,\n ):\n \"\"\"\n Test that if GOV.UK Pay errors whilst cancelling some other ongoing\n sessions/payments, the method raises GOVUKPayAPIException to keep the system consistent.\n\n Possible GOV.UK Pay errors when cancelling:\n - 400 - BAD REQUEST\n - 401 - UNAUTHORIZED\n - 404 - NOT FOUND\n - 409 - CONFLICT\n - 500 - INTERNAL SERVER ERROR\n \"\"\"\n order = OrderWithAcceptedQuoteFactory()\n existing_session = PaymentGatewaySessionFactory(\n order=order,\n status=PaymentGatewaySessionStatus.CREATED,\n )\n\n # mock GOV.UK requests used to\n # - refresh the existing payment gateway session\n # - cancel the GOV.UK payment\n requests_mock.get(\n govuk_url(f'payments/{existing_session.govuk_payment_id}'),\n status_code=200,\n json={\n 'state': {'status': existing_session.status},\n },\n )\n requests_mock.post(\n govuk_url(f'payments/{existing_session.govuk_payment_id}/cancel'),\n status_code=govuk_status_code,\n )\n\n assert PaymentGatewaySession.objects.count() == 1\n\n with pytest.raises(GOVUKPayAPIException):\n PaymentGatewaySession.objects.create_from_order(order)\n\n assert PaymentGatewaySession.objects.count() == 1\n\n def test_ongoing(self):\n \"\"\"\n Test that given:\n session 1 - order 1 - status created\n session 2 - order 1 - status submitted\n session 3 - order 1 - status failed\n session 4 - order 2 - status started\n session 5 - order 2 - status success\n session 6 - order 2 - status cancelled\n\n the method .ongoing() on the queryset only returns the sessions\n which are considered not finished.\n \"\"\"\n order1, order2 = OrderWithAcceptedQuoteFactory.create_batch(2)\n\n order1_sessions = PaymentGatewaySessionFactory.create_batch(\n 3,\n order=order1,\n status=factory.Iterator([\n PaymentGatewaySessionStatus.CREATED,\n PaymentGatewaySessionStatus.SUBMITTED,\n PaymentGatewaySessionStatus.FAILED,\n ]),\n )\n order2_sessions = PaymentGatewaySessionFactory.create_batch(\n 3,\n order=order2,\n status=factory.Iterator([\n PaymentGatewaySessionStatus.STARTED,\n PaymentGatewaySessionStatus.SUCCESS,\n PaymentGatewaySessionStatus.CANCELLED,\n ]),\n )\n\n # test qs without filters\n qs = PaymentGatewaySession.objects.ongoing()\n assert set(qs.values_list('id', flat=True)) == {\n order1_sessions[0].id,\n order1_sessions[1].id,\n order2_sessions[0].id,\n }\n\n # test qs with order filter\n qs = PaymentGatewaySession.objects.filter(order=order1).ongoing()\n assert set(qs.values_list('id', flat=True)) == {\n order1_sessions[0].id,\n order1_sessions[1].id,\n }\n","repo_name":"uktrade/data-hub-api","sub_path":"datahub/omis/payment/test/test_managers.py","file_name":"test_managers.py","file_ext":"py","file_size_in_byte":16049,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"81"}
+{"seq_id":"30622144507","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Aug 30 13:22:59 2017\n\n@author: Charles\n\"\"\"\nimport glob, re, os\n\nfor filename in glob.glob(r'F:\\DS-main\\Kaggle-main\\Carvana Image Masking Challenge\\input\\New folder'):\n new_name = re.sub(\"\", r'\\1_\\2\\3', filename)\n os.rename(filename, new_name)\n \n","repo_name":"charlesjansen/Kaggle-Carvana-Image-Masking-Challenge","sub_path":"old/untitled1.py","file_name":"untitled1.py","file_ext":"py","file_size_in_byte":300,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23863992857","text":"\"\"\"\nDescrição rápida e simples\n\nEste modulo é apenas um teste para vermos como funciona o help utilizando a docstrings,\nonde eu estou estudando python para um dia se tornar programador profissional que é o meu maior sonho no momento\ne no futuro conseguir comprar um apartamento para meus pais e conseguir comprar o meu para construi uma familia, e\ntambem ajudar meus irmaos.Hoje é dia 01/08/2022 e eu estou atras do meu sonho, hj é um dia em que estou bem triste\npois a ansiedade e a depressao esta acabando comigo pois nao estou conseguindo continuar minha vida e ir atras dos meus\nsonhos.\n\n\"\"\"\n\nvariavel_1 = 'valor 1'\n\n#a primeira linha da minha função deve ser minha documentação\ndef foma(x, y):\n \"\"\"soma x e y \"\"\"\n return x + y\n\ndef multiplica(x, y, z=None):\n \"\"\"Multiplica x, y, z\n\n Multiplica x, y e oz, O programador por omitir a variavel z caso nao tenha necessidade de usa-la\n \"\"\"\n print(z)\n if z:\n return x * y\n else:\n return x * y * z\n\nvariavel_2 = 'valor 2'\nvariavel_3 = 'valor 3'\nvariavel_4 = 'valor 4'\n","repo_name":"diogolimalucasdev/PythonNovaJornada","sub_path":"aula122_DocStrings/funcoes.py","file_name":"funcoes.py","file_ext":"py","file_size_in_byte":1069,"program_lang":"python","lang":"pt","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"38742269793","text":"import argparse\nimport os\nfrom PIL import Image\nimport torch\nimport torchvision\nimport torchvision.transforms as transforms\nfrom torchvision.utils import save_image\nimport YodaModel as yoda\nimport train\nfrom torch.utils.data import DataLoader, Dataset\nimport time as t\n# import custom_dataset as custData\n\nimport cv2\nfrom KittiDataset import KittiDataset\nfrom KittiAnchors import Anchors\n\nmax_ROIs = -1\n\n\nif __name__ == '__main__':\n\n device = 'cpu'\n parser = argparse.ArgumentParser()\n\n parser.add_argument('-model_file', type=str, help='model weight file')\n parser.add_argument('-cuda', type=str, help='[y/N]')\n parser.add_argument('-i', metavar='input_dir', type=str, help='input dir (./)')\n parser.add_argument('-o', metavar='output_dir', type=str, help='output dir (./)')\n parser.add_argument('-IoU', metavar='IoU_threshold', type=float, help='[0.02]')\n parser.add_argument('-d', metavar='display', type=str, help='[y/N]')\n parser.add_argument('-m', metavar='mode', type=str, help='[train/test]')\n parser.add_argument('-v', default=False, type=bool, help='[train/test]')\n\n opt = parser.parse_args()\n\n if torch.cuda.is_available() and opt.cuda == 'Y':\n if torch.cuda.device_count() > 1:\n device = torch.device(\"cuda:1\")\n else:\n device = torch.device(\"cuda:0\")\n else:\n device = torch.device(\"cpu\")\n print(\"Using device: {}\".format(device))\n\n input_dir = None\n if opt.i != None:\n input_dir = opt.i\n\n output_dir = None\n if opt.o != None:\n output_dir = opt.o\n\n IoU_threshold = 0.02\n if opt.IoU != None:\n IoU_threshold = float(opt.IoU)\n\n show_images = False\n if opt.d != None:\n if opt.d == 'y' or opt.d == 'Y':\n show_images = True\n\n training = True\n if opt.m == 'test':\n training = False\n\n train_transform = transforms.Compose([transforms.ToTensor(), transforms.Resize((150, 150))])\n\n batch_size = 48 # 48 ROIs generated per image\n\n # Load kitti data\n dataset = KittiDataset(input_dir, training=training)\n anchors = Anchors()\n\n # Load saved yodamodel\n model_file = opt.model_file\n model = yoda.model()\n model.load_state_dict(torch.load(opt.model_file))\n model.to(device=device)\n model.eval()\n\n print('model loaded OK!')\n print(\"Using device: {}\".format(device))\n all_mean_IoU = []\n out_tensor = None\n print(\"Starting Test\")\n for item in enumerate(dataset):\n idx = item[0]\n if opt.v:\n print(\" {}/{}\".format(idx, len(dataset)))\n # Original Kitti img\n image = item[1][0]\n label = item[1][1]\n # Get car label indx\n idx = dataset.class_label['Car']\n # Get all of the CAR ROIS for this img Truth ones\n car_ROIs = dataset.strip_ROIs(class_ID=idx, label_list=label)\n anchor_centers = anchors.calc_anchor_centers(image.shape, anchors.grid)\n # Generate ROIs for each image\n ROIs, boxes = anchors.get_anchor_ROIs(image, anchor_centers, anchors.shapes)\n if opt.v:\n print(\"Created all ROIs\".format())\n # Convert ROIS to a tensor\n transformed_ROI = []\n for idx, item in enumerate(ROIs):\n temp_ROI = train_transform(item)\n transformed_ROI.append(temp_ROI)\n tensor_ROI = torch.stack(transformed_ROI)\n\n with torch.no_grad():\n if opt.v:\n print(\"Putting ROI's through model\".format())\n output = model(tensor_ROI)\n\n # Rounding results, putting answers to 1 or 0\n output_rounded = torch.round(output)\n ROI_IoUs = []\n bound_boxes = []\n # Iterate through all outputs\n if opt.v:\n print(\"ROIS classified as 'Car': \".format())\n for idx, item in enumerate(output_rounded):\n # Find the ones that predicted 'Car'\n\n if torch.argmax(item) == 1:\n if opt.v:\n print(\"Index: {}, Box: {}\".format(idx, boxes[idx]))\n # Get the IoU of this box from the CarROIs\n curr_roi_max_IoU = anchors.calc_max_IoU(boxes[idx], car_ROIs)\n ROI_IoUs += [curr_roi_max_IoU]\n if curr_roi_max_IoU >= IoU_threshold:\n bound_boxes.append(boxes[idx])\n print(\"Printing all IoUs for each ROI of this image\")\n for idx, iou in enumerate(ROI_IoUs):\n print(\"ROI {}, IoU calculations: {}\".format(idx, ROI_IoUs[idx]))\n\n curr_image_mean_IoU = 0\n if len(ROI_IoUs) != 0:\n curr_image_mean_IoU = sum(ROI_IoUs)/len(ROI_IoUs)\n # sum_IoU = 0\n all_mean_IoU.append(curr_image_mean_IoU)\n print(\"Mean IoU {}\".format(curr_image_mean_IoU))\n\n if show_images:\n image2 = image.copy()\n\n for box in bound_boxes:\n pt1 = (box[0][1], box[0][0])\n pt2 = (box[1][1], box[1][0])\n cv2.rectangle(image2, pt1, pt2, color=(0, 255, 255))\n\n if show_images:\n cv2.imshow('boxes', image2)\n key = cv2.waitKey(0)\n if key == ord('x'):\n break\n\n global_mean_IoU = sum(all_mean_IoU)/len(dataset)\n print(\"Final global mean IoU is: {}\".format(global_mean_IoU))\n","repo_name":"Krsikapa00/ELEC475-Labs","sub_path":"Lab4/Submission_files/YodaTest.py","file_name":"YodaTest.py","file_ext":"py","file_size_in_byte":5269,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"13554686216","text":"\"\"\"\nModule with the class for Syslog message transform\n\"\"\"\n\n# More info in RFC 3164: https://tools.ietf.org/html/rfc3164\nFACILITY_KERN = 0 # kernel messages\nFACILITY_USER = 1 # user-level messages\nFACILITY_MAIL = 2 # mail system\nFACILITY_DAEMON = 3 # system daemons\nFACILITY_AUTH = 4 # security/authorization messages\nFACILITY_SYSLOG = 5 # messages generated internally by syslog\nFACILITY_LPR = 6 # line printer subsystem\nFACILITY_NEWS = 7 # network news subsystem\nFACILITY_UUCP = 8 # UUCP subsystem\nFACILITY_CLOCK = 9 # clock daemon\nFACILITY_AUTHPRIV = 10 # security/authorization messages\nFACILITY_FTP = 11 # FTP daemon\nFACILITY_NTP = 12 # NTP subsystem\nFACILITY_LOG_AUDIT = 13 # log audit\nFACILITY_LOG_ALERT = 14 # log alert\nFACILITY_CRON = 15 # scheduling daemon\nFACILITY_LOCAL0 = 16 # local use 0 (local0)\nFACILITY_LOCAL1 = 17 # local use 1 (local1)\nFACILITY_LOCAL2 = 18 # local use 2 (local2)\nFACILITY_LOCAL3 = 19 # local use 3 (local3)\nFACILITY_LOCAL4 = 20 # local use 4 (local4)\nFACILITY_LOCAL5 = 21 # local use 5 (local5)\nFACILITY_LOCAL6 = 22 # local use 6 (local6)\nFACILITY_LOCAL7 = 23 # local use 7 (local7)\n# More info in RFC 5424: https://tools.ietf.org/html/rfc5424\nSEVERITY_EMERG = 0 # System is unusable\nSEVERITY_ALERT = 1 # Should be corrected immediately\nSEVERITY_CRIT = 2 # Critical conditions\nSEVERITY_ERROR = 3 # Error conditions\nSEVERITY_WARN = 4 # May indicate that an error will occur if action is\n# not taken.\nSEVERITY_NOTICE = 5 # Events that are unusual, but not error conditions.\nSEVERITY_INFO = 6 # Normal operational messages that require no action.\nSEVERITY_DEBUG = 7 # Information useful to developers for debugging\n# the application.\nCOMPOSE = \"%s%s\"\nCOMPOSE_BYTES = b\"%s%s\"\nFORMAT_MY = \"<%d>%s %s %s: \" # Not \\000\nFORMAT_MY_BYTES = b\"<%d>%s %s %s: \" # Not \\000\n\n# logging.handler translator to Sender codes\nfacility_names = {\n \"auth\": FACILITY_AUTH,\n \"authpriv\": FACILITY_AUTHPRIV,\n \"cron\": FACILITY_CRON,\n \"daemon\": FACILITY_DAEMON,\n \"ftp\": FACILITY_FTP,\n \"kern\": FACILITY_KERN,\n \"lpr\": FACILITY_LPR,\n \"mail\": FACILITY_MAIL,\n \"news\": FACILITY_NEWS,\n \"security\": FACILITY_AUTH, # DEPRECATED\n \"syslog\": FACILITY_SYSLOG,\n \"user\": FACILITY_USER,\n \"uucp\": FACILITY_UUCP,\n \"local0\": FACILITY_LOCAL0,\n \"local1\": FACILITY_LOCAL1,\n \"local2\": FACILITY_LOCAL2,\n \"local3\": FACILITY_LOCAL3,\n \"local4\": FACILITY_LOCAL4,\n \"local5\": FACILITY_LOCAL5,\n \"local6\": FACILITY_LOCAL6,\n \"local7\": FACILITY_LOCAL7,\n}\n\npriority_map = {\n \"EMERG\": SEVERITY_EMERG,\n \"ALERT\": SEVERITY_ALERT,\n \"CRITICAL\": SEVERITY_CRIT,\n \"CRIT\": SEVERITY_CRIT,\n \"ERROR\": SEVERITY_ERROR,\n \"ERR\": SEVERITY_ERROR,\n \"WARNING\": SEVERITY_WARN,\n \"WARN\": SEVERITY_WARN, # DEPRECATED\n \"NOTICE\": SEVERITY_NOTICE,\n \"INFO\": SEVERITY_INFO,\n \"DEBUG\": SEVERITY_DEBUG,\n}\n","repo_name":"DevoInc/python-sdk","sub_path":"devo/sender/transformsyslog.py","file_name":"transformsyslog.py","file_ext":"py","file_size_in_byte":2875,"program_lang":"python","lang":"en","doc_type":"code","stars":25,"dataset":"github-code","pt":"81"}
+{"seq_id":"20874754565","text":"from django import forms\nfrom .models import UserInfo\n\n\nclass NewPostForm(forms.Form):\n post = forms.CharField(label=\"\",\n max_length=1000,\n min_length=1,\n widget=forms.Textarea(attrs={'rows': 3,\n 'cols': 40,\n 'placeholder': 'What is on your mind?', }))\n\n\nclass CommentForm(forms.Form):\n comment = forms.CharField(label=\"\",\n max_length=1000,\n min_length=1,\n widget=forms.Textarea(attrs={'class': 'form-control',\n 'placeholder': 'Comment',\n 'rows': 1, }))\n\n\nclass ProfileInfoForm(forms.ModelForm):\n date_of_birth = forms.CharField(required=False, label='Date of birth')\n live_in = forms.CharField(required=False, label='I live in')\n phone = forms.CharField(required=False, label='Phone number')\n sex = forms.ChoiceField(choices=[('male', 'Male'), ('female', 'Female')], required=False, label=\"I'm\")\n email = forms.BooleanField(required=False, label=\"Share email?\")\n\n class Meta:\n model = UserInfo\n fields = ['date_of_birth', 'live_in', 'phone', 'sex', 'email']\n widgets = {\n 'email': forms.CheckboxInput(),\n }\n\n\nclass ChatForm(forms.Form):\n message = forms.CharField(label=\"\",\n max_length=1000,\n min_length=1,\n widget=forms.Textarea(attrs={'rows': 2,\n 'placeholder': 'Aa', }))\n","repo_name":"nvast/Social_media","sub_path":"main/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1777,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"26026481081","text":"from typing import Optional\n\nfrom dstack._internal.api import workflow_api\nfrom dstack._internal.api.artifacts import (\n download_artifact_files_backend,\n upload_artifact_files_backend,\n upload_artifact_files_from_tag_backend,\n)\nfrom dstack._internal.api.runs import get_tagged_run_name_backend\n\n\ndef upload(local_path: str, artifact_path: Optional[str] = None, tag: Optional[str] = None):\n \"\"\"Uploads files located at `local_path` as the artifacts of the current run.\n If `tag` is specified, uploads the files as the artifacts of the tag instead.\n By default, artifact files saved under the same path as `local_path`.\n The `artifact_path` parameter can be used to specify a different artifact path.\n\n Examples:\n ```python\n from dstack import artifacts\n\n # Uploads local_path as an artifact of the current run\n artifacts.upload(local_path=\"datasets/dataset1\")\n\n # Uploads local_path as an artifact of a new run tagged as my_tag and saves it as artifact_path\n artifacts.upload(local_path=\"datasets/dataset1\", artifact_path=\"data\", tag=\"my_tag\")\n ```\n\n :param local_path: The local path to upload the files from\n :type local_path: str\n :param artifact_path: The path under which the files will be stored\n :type artifact_path: Optional[str]\n :param tag: The tag to assign the artifacts, defaults to None\n :type tag: Optional[str]\n :raises ArtifactsUploadError: Raises if cannot upload the artifacts\n :raises DstackError: The base exception for all dstack errors\n \"\"\"\n if artifact_path is None:\n artifact_path = local_path\n backend = workflow_api.get_current_backend()\n repo_id = workflow_api.get_current_repo_id()\n job_id = workflow_api.get_current_job_id()\n if tag is None:\n upload_artifact_files_backend(\n backend=backend,\n repo_id=repo_id,\n job_id=job_id,\n local_path=local_path,\n artifact_path=artifact_path,\n )\n return\n job = workflow_api.get_current_job()\n upload_artifact_files_from_tag_backend(\n backend=backend,\n repo=job.repo,\n hub_user_name=job.hub_user_name,\n local_path=local_path,\n artifact_path=artifact_path,\n tag_name=tag,\n )\n\n\ndef download(\n run: Optional[str] = None,\n tag: Optional[str] = None,\n artifact_path: Optional[str] = None,\n local_path: Optional[str] = None,\n):\n \"\"\"Downloads artifact files of a run or a tag.\n The files are downloaded from `artifact_path` to `local_path`.\n By default, downloads all the files and saves them to the current directory.\n\n Examples:\n ```python\n from dstack import artifacts\n\n # Downloads all artifact files of a run\n artifacts.download(run=\"sharp-shrimp-1\")\n\n # Downloads artifact files from artifact_path and saves them to local_path\n artifacts.download(tag=\"my_tag\", artifact_path=\"output/my_model\", local_path=\"./my_model\")\n ```\n\n :param run: The run to download the artifacts from\n :type run: Optional[str]\n :param tag: The tag to download the artifacts from\n :type tag: Optional[str]\n :param artifact_path: The path to artifact files to download, defaults to \"\"\n :type artifact_path: Optional[str]\n :param local_path: The local path to save the files to, defaults to \".\"\n :type local_path: Optional[str]\n :raises ArtifactsDownloadError: Raises if cannot download the artifacts\n :raises DstackError: The base exception for all dstack errors\n \"\"\"\n if artifact_path is None:\n artifact_path = \"\"\n local_path = \".\"\n if local_path is None:\n local_path = artifact_path\n backend = workflow_api.get_current_backend()\n repo_id = workflow_api.get_current_repo_id()\n run_name, _ = get_tagged_run_name_backend(\n backend=backend,\n repo_id=repo_id,\n run_name=run,\n tag_name=tag,\n )\n download_artifact_files_backend(\n backend=backend,\n repo_id=repo_id,\n run_name=run_name,\n source=artifact_path,\n target=local_path,\n )\n","repo_name":"silvacarl2/dstack","sub_path":"cli/dstack/artifacts/_artifacts.py","file_name":"_artifacts.py","file_ext":"py","file_size_in_byte":4072,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"81"}
+{"seq_id":"36685641053","text":"import sys\nfrom PySide6.QtGui import *\nfrom PySide6.QtCore import *\nfrom PySide6.QtWidgets import *\n\ndata = (['1', '张三', '男'], ['2', '张四', '男'], ['3', '王五', '男'], ['4', '虎妞', '女'], ['5', '刘明', '男'], ['1', '张三', '男'],\n ['2', '张四', '男'], ['3', '王五', '男'], ['4', '虎妞1', '女'], ['5', '刘1明', '男'])\n\n\nclass TableDome(QWidget):\n def __init__(self):\n super(TableDome, self).__init__()\n self.initUI()\n\n def initUI(self):\n self.setWindowTitle('QTableWidget案例')\n self.resize(400, 300)\n conLayout = QHBoxLayout()\n tableWidget = QTableWidget() # 创建表格实例\n tableWidget.setRowCount(15) # 初始化表格5行\n tableWidget.setColumnCount(3) # 3列\n tableWidget.setShowGrid(True) # 设置显示网格线\n tableWidget.setHorizontalHeaderLabels(['学号', '姓名', '性别']) # 设置表头信息\n # tableWidget.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch) # 自适应伸缩模式\n # tableWidget.setEditTriggers(QAbstractItemView.NoEditTriggers) # 禁止编辑\n # tableWidget.resizeColumnsToContents() # 根据内容自适应列宽\n # tableWidget.resizeRowsToContents() # 根据内容自适应行高\n # tableWidget.setSelectionBehavior(QAbstractItemView.SelectRows) # 设置单元格选择方式为行模式\n # tableWidget.verticalHeader().setVisible(False) # 隐藏第一列的表头\n # tableWidget.horizontalHeader().setVisible(False) # 隐藏第一行的表头\n\n # 添加表格数据\n for num, i in enumerate(data):\n for col_num, col in enumerate(i):\n tableWidget.setItem(num, col_num, QTableWidgetItem(col)) # 向表格中添加数据\n tableWidget.setAlternatingRowColors(True) # 表格颜色交替显示\n conLayout.addWidget(tableWidget)\n self.setLayout(conLayout)\n item = tableWidget.findItems('刘1明',Qt.MatchExactly)\n item = item[0]\n row = item.row()\n item.setForeground(QBrush(QColor(255,0,0)))\n print(row,'item:',item)\n tableWidget.verticalScrollBar().setSliderPosition(row)\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n example = TableDome()\n example.show()\n sys.exit(app.exec())\n","repo_name":"tainyuhong/pyside_study1","sub_path":"stu/stu_Qtablewidget1.py","file_name":"stu_Qtablewidget1.py","file_ext":"py","file_size_in_byte":2354,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72890906186","text":"from collections import deque\n\n\ndef dfs(g, v, visited):\n visited[v] = True #방문하면 True\n print(v, end=' ')\n\n for i in g[v]: #시작노드와 연결된 노드 중에서\n if not visited[i]: #방문 안 한 노드가 있다면\n dfs(g,i, visited) #해당 노드부터 다시 탐색\n\n\n\ndef bfs(g, v, visited):\n q = deque([v]) #시작노드 큐에 삽입\n\n visited[v] = True #시작노드 방문\n\n while q:\n v = q.popleft() #시작노드 pop\n print(v, end=' ')\n for i in g[v]: #시작노드와 연결된 모든 노드 중에서\n if not visited[i]: #방문 안 한 노드가 있다면\n q.append(i) #큐에 넣고 \n visited[i] = True #방문\n #현재 노드와 연결된 모든 노드를 방문했으면 제일 먼저 방문한 노드 pop\n\n\n\n\n\nqty, n, start = map(int, input().split())\n\n#그래프, 방문 리스트 초기화\ngraph = [[] for _ in range(qty+1)]\nvisi = [False for _ in range(qty+1)]\n\n#그래프에 노드 정보 입력\nfor _ in range(n):\n i, node = map(int, input().split())\n graph[i].append(node)\n graph[node].append(i)\n\n#각 노드 오름차순으로 정렬\nfor i in range(qty+1):\n graph[i].sort()\n\n\ndfs(graph, start, visi)\nvisi = [False for _ in range(qty+1)]\nprint()\nbfs(graph, start, visi)\n","repo_name":"Dam0305/Algorithm","sub_path":"BOJ/1260_DFS와 BFS.py","file_name":"1260_DFS와 BFS.py","file_ext":"py","file_size_in_byte":1325,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"33392946726","text":"import pygame\n\n# defining variables to easily call \nWIDTH, HEIGHT = 900, 500\nWIN = pygame.display.set_mode((WIDTH, HEIGHT))\npygame.display.set_caption(\"Hunt The Wumpus\")\n\nWHITE = (255, 255, 255)\n\nFPS = 60\n\n# function to set the colour of the window\ndef draw_window():\n\t\tWIN.fill((WHITE))\n\t\tpygame.display.update()\n\n# function that controls most gameplay\ndef main():\n\tclock = pygame.time.Clock()\n\trun = True\n\twhile run:\n\t\t# caps FPS at definied value\n\t\tclock.tick(FPS)\n\t\tfor event in pygame.event.get():\n\t\t\tif event.type == pygame.QUIT:\n\t\t\t\trun = False\n\n\t\tdraw_window()\n\n\tpygame.quit()\n\n\nif __name__ == \"__main__\":\n\tmain()\n","repo_name":"andrewheld616/huntTheWumpus","sub_path":"htw.py","file_name":"htw.py","file_ext":"py","file_size_in_byte":622,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"25306507968","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Oct 4 17:39:53 2019\r\n\r\n@author: cheerag.verma\r\n\"\"\"\r\n\r\nn = int(input())\r\na= []\r\nif n%6 ==0 or n%6 ==1 or n%6 ==3:\r\n print(\"YES\",end = \"\")\r\nelse:\r\n print(\"NO\",end = \"\")\r\n\r\n","repo_name":"iamcheerag/NPTEL-JOY-OF-COMPUTING","sub_path":"week10/Programming Assignment-2 Jumps.py","file_name":"Programming Assignment-2 Jumps.py","file_ext":"py","file_size_in_byte":222,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"30323204634","text":"#Chase Williams\n#Lab 4\nimport pygame\nimport math\nimport time\nimport math3d\n\nclass Curve(object):\n def __init__(self):\n self.CPList = [] #Add main points\n self.TList = [] #Tangent list\n self.DrawList = []\n self.Resolution = 10 #distance between points\n\n def addCP(self, p):\n self.CPList.append(p)\n# self.TList.append()#some random value for this number)\n# self.genDraw()\n\n def genDraw(self): #Generates lines to draw in between control points\n for i in self.TList:\n self.p0 = self.CPList[i]\n self.p1 = self.CPList[i + 1]\n self.t0 = i\n self.t1 = i + 1\n u = 0\n for j in range(self.resolution): #Calculates tangent\n a = 2(u ** 3) - 3(u ** 2) + 1\n b = u ** 3 - 2(u ** 2) + u\n c = -2(u ** 3) + 3(u ** 2)\n d = u ** 3 - u ** 2\n p = a * self.p0 + b * self.t0 + c * self.p1 + d * self.t1\n self.DrawList.append(p)\n u += 1 / (resolution + 1)\n\n def Render(self): #renders objects to the screen\n pygame.draw.circle(window, (255, 255, 255), self.CPList[-1], 10)\n if len(self.CPList) > 1:\n for i in range(self.Resolution):\n pygame.draw.lines(window, (255, 0, 0), False, self.CPList)\n\npygame.init()\n\nwindow = pygame.display.set_mode((800,600))\nCP = Curve()\ndone = False\nwhile not done:\n #mBut = pygame.mouse.get_pressed()\n mPos = math3d.VectorN((pygame.mouse.get_pos()))\n evtList = pygame.event.get()\n for evt in evtList:\n if evt.type == pygame.QUIT:\n done = True\n if evt.type == pygame.MOUSEBUTTONDOWN: #when mouse is clicked, add mouse pos to CPList\n CP.CPList.append(mPos.iTuple())\n for p in range(CP.Resolution):\n CP.Render()\n #if mBut[0]:\n # CP.addCP(mPos)\n # CP.Render()\n #window.fill((0, 0, 0))\n pygame.display.flip()\n\npygame.display.quit()\n\n\n","repo_name":"fooy787/CLASS_LABS","sub_path":"Freshman/1803/Lab 4/Lab 4.py","file_name":"Lab 4.py","file_ext":"py","file_size_in_byte":1991,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"25306072188","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Mar 11 12:50:33 2020\r\n\r\n@author: cheerag.verma\r\n\"\"\"\r\n\r\ndef spiralPrinting(arr):\r\n \r\n rowStart = 0\r\n rowEnd = n\r\n colStart = 0\r\n colEnd = m\r\n while rowStart self.size-1:\n raise IndexError()\n current = self.head\n for _ in range(index):\n current = current.next\n return current\n\n def add(self, new_node, index):\n if index == 0:\n new_node.next = self.head\n self.head = new_node\n else:\n previous_elem = self.index(index-1)\n new_node.next = previous_elem.next\n previous_elem.next = new_node\n self.size += 1\n\n def remove(self, index):\n if index == 0:\n self.head = self.head.next\n else:\n previous_elem = self.index(index-1)\n previous_elem.next = previous_elem.next.next\n self.size -= 1\n\n def display(self):\n current = self.head\n out = [current.value]\n while current.next:\n out.append(current.next.value)\n current = current.next\n return out\n\n\nclass DoubleLinkedList(SingleLinkedList):\n def add(self, new_node, index):\n if not self.head:\n self.head = new_node\n self.size += 1\n return\n if not self.tail:\n self.tail = new_node\n self.head.next = self.tail\n self.size += 1\n return\n if index == 0:\n new_node.next = self.head\n self.head.prev = new_node\n self.head = new_node\n elif index == self.size:\n new_node.prev = self.tail\n self.tail.next = new_node\n self.tail = new_node\n else:\n previous_elem = self.index(index-1)\n new_node.next = previous_elem.next\n new_node.prev = previous_elem\n previous_elem.next = new_node\n new_node.next.prev = new_node\n self.size += 1\n\n def remove(self, index):\n if index == 0:\n self.head = self.head.next\n self.head.prev = None\n elif index == self.size - 1:\n self.tail = self.tail.prev\n self.tail.next = None\n else:\n previous_elem = self.index(index-1)\n previous_elem.next = previous_elem.next.next\n previous_elem.next.prev = previous_elem\n self.size -= 1\n\n\nlst1 = SingleLinkedList()\nlst1.head = SingleLinkedNode(1, next=SingleLinkedNode(2, next=SingleLinkedNode(3)))\nlst1.tail = SingleLinkedNode(3)\nlst1.size = 3\nassert lst1.index(1) == SingleLinkedNode(2, next=SingleLinkedNode(3))\nassert lst1.find(2)\nassert lst1.find(1) == 0\nassert lst1.find(4) == -1\n\nanother_lst = SingleLinkedList()\nanother_lst.add(SingleLinkedNode(1), index=0)\nassert another_lst.display() == [1]\nanother_lst.add(SingleLinkedNode(2), index=1)\nassert another_lst.display() == [1, 2]\nanother_lst.add(SingleLinkedNode(1), index=2)\nassert another_lst.display() == [1, 2, 1]\nanother_lst.add(SingleLinkedNode(6), index=1)\nassert another_lst.display() == [1, 6, 2, 1]\nanother_lst.remove(2)\nassert another_lst.display() == [1, 6, 1]\n\nlst2 = DoubleLinkedList()\nlst2.add(DoubleLinkedNode(1), index=0)\nassert lst2.display() == [1]\nlst2.add(DoubleLinkedNode(2), index=1)\nassert lst2.display() == [1, 2]\nlst2.add(DoubleLinkedNode(3), index=1)\nassert lst2.display() == [1, 3, 2]\nassert lst2.find(3) == 1\nassert lst2.find(1) == 0\nassert lst2.find(2) == 2\nlst2.remove(1)\nassert lst2.display() == [1, 2]\n","repo_name":"confar/python_challenges","sub_path":"linked_list.py","file_name":"linked_list.py","file_ext":"py","file_size_in_byte":4164,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"2307747886","text":"\"\"\"\nDefinitions of siamese models. Each model uses a siamese network to convert the survival problem into a\nclassification problem.\n\n - The :class:`ImageSiamese` class creates a model that uses only the images as input and uses some blocks\n - The :class:`SimpleImageSiamese` class creates a basic model that only uses images as input\n - The :class:`ImageScalarSiamese` class creates a model that combines the image input with the scalar input\n of the radiomic features, extracted with PyRadiomics.\n - The :class:`ResidualImageScalarSiamese` class creates a model that combines the image input with the scalar input\n but it also uses a residual network to fit the images. It also uses multiple blocks, similar to the\n inception idea\n - The :class:`ScalarOnlySiamese` class creates a siamese model that only uses the radiomic features as an input.\n - The :class:`ScalarOnlyDropoutSiamese` class creates a siamese model that only uses the radiomic features as an\n input but adds multiple dropout layers to improve the results.\n - The :class:`VolumeOnlySiamese` class creates a siamese model that only uses the volume feature to fit the model.\n\n\n.. inheritance-diagram:: models.siameses\n :parts: 1\n\n\"\"\"\n\nfrom typing import Dict, Any, Union, Tuple\n\nimport tensorflow as tf\nimport numpy as np\n\nimport data\nimport settings\nfrom .basics import BasicImageSiamese, BasicSiamese\n\n\nclass ImageSiamese(BasicImageSiamese):\n \"\"\"\n Class representing the initial and simple siamese structure used for the first steps of the project. It\n inherits :any:`BasicSiamese` so it has the same tensors to be fed.\n\n **Convolutional Model**:\n\n It contains parallel in inception_block and 3 FC layers\n\n \"\"\"\n\n def __init__(self, **kwargs):\n \"\"\"\n Construct a new SimpleSiamese class\n\n :param gpu_level: Amount of GPU to be used with the model\n\n 0. No GPU usage\n 1. Only second conv layers\n 2. All conv layers\n 3. All layers and parameters are on the GPU\n \"\"\"\n super().__init__(**kwargs)\n\n def _inception_block(self, x:tf.Tensor, stage: str, block:str):\n \"\"\"\n\n :param x: Network's input images with shape ``[batch, 64, 64, 64, 1]``\n :param filters: list of integers, the number of filters in the CONV layers\n :param stage: integer, Used to name the layers, depending on their position in the network\n :param block: string, Used to name the layers, depending on their position in the network\n :return: Tensor of shape ``[n_X, n_Y, n_Z, n_C]``\n \"\"\"\n conv_name_base = 'res' + str(stage) + block + '_branch'\n bn_name_base = 'bn' + str(stage) + block + '_branch'\n # Retrieve Filters\n #F1, F2, F3 = filters\n device = '/gpu:0' if self._gpu_level >= 1 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for first conv layers\")\n with tf.device(device):\n\n a1 = tf.layers.conv3d(\n\n x,\n\n filters=8,\n\n kernel_size=[1, 1, 1],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'a1'\n\n )\n\n a1 = tf.layers.conv3d(\n\n a1,\n\n filters=8,\n\n kernel_size=[4, 4, 4],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'a2'\n\n )\n device = '/gpu:0' if self._gpu_level >= 2 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for first conv layers\")\n with tf.device(device):\n\n b1 = tf.layers.conv3d(\n\n x,\n\n filters=8,\n\n kernel_size=[1, 1, 1],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'b1'\n\n )\n\n b1 = tf.layers.conv3d(\n\n b1,\n\n filters=8,\n\n kernel_size=[2, 2, 2],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'b2'\n\n )\n device = '/gpu:0' if self._gpu_level >= 3 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for first conv layers\")\n with tf.device(device):\n c1 = tf.nn.max_pool3d(\n\n x,\n\n ksize=[1, 4, 4, 4, 1],\n\n strides=[1,1,1,1,1],\n\n padding='SAME',\n\n\n name=conv_name_base + 'c1'\n\n )\n\n c1 = tf.layers.conv3d(\n\n c1,\n\n filters=8,\n\n kernel_size=[1, 1, 1],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'c2'\n\n )\n\n d1 = tf.concat([a1, b1], 4)\n d1 = tf.concat([d1, c1], 4)\n\n d1 = tf.layers.conv3d(\n\n d1,\n\n filters=1,\n\n kernel_size=[1, 1, 1],\n\n strides=1,\n\n activation=tf.nn.relu,\n\n padding='SAME',\n\n name=conv_name_base + 'd'\n )\n d1 = tf.contrib.layers.batch_norm(d1,\n center=True, scale=True,\n scope='bn')\n tf.layers.BatchNormalization(d1)\n return d1\n\n def _conv_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method :func:`~BasicSiamese._conv_layers`\n\n :param x: Network's input images with shape ``[batch, 64, 64, 64, 1]``\n :return: Filtered image with the convolutions applied\n \"\"\"\n # In: [batch, 64, 64, 64, 1]\n x1 = self._inception_block(x,\"1s\", \"b1\")\n\n return x1\n\n\n def _fc_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method ``BasicSiamese._fc_layers``\n\n :param x: Image, usually previously filtered with the convolutional layers.\n :return: Tensor with shape ``[batch, 1]``\n \"\"\"\n device = '/gpu:0' if self._gpu_level >= 3 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for FC layers\")\n with tf.device(device):\n # Out: [batch, 64*64*64*1]\n x = tf.layers.flatten(\n x,\n name=\"flat\"\n )\n\n # Out: [batch, 100]\n x = tf.layers.dense(\n x,\n 100,\n activation=tf.nn.relu,\n name=\"fc1\"\n )\n\n # Out: [batch, 50]\n x = tf.layers.dense(\n x,\n 50,\n activation=tf.nn.relu,\n name=\"fc2\"\n )\n\n # Out: [batch, 1]\n x = tf.layers.dense(\n x,\n 1,\n activation=tf.nn.relu,\n name=\"fc3\"\n )\n return x\n\n def uses_images(self) -> bool:\n \"\"\"\n Implementation of :func:`BasicModel.uses_images`.\n\n :return: :any:`True`, the model uses images as input to work\n \"\"\"\n return True\n\n\nclass SimpleImageSiamese(BasicImageSiamese):\n r\"\"\"\n Simple siamese network implementation that uses images as input.\n\n Has the same parameters as :class:`BasicImageSiamese`\n\n **Convolutional Model**:\n\n It contains 4 convolutional layers and 3 FC layers\n\n - :math:`3^3` kernel with 30 filters and stride = 2\n - :math:`3^3` kernel with 40 filters and stride = 1\n - :math:`3^3` kernel with 40 filters and stride = 1\n - :math:`3^3` kernel with 50 filters and stride = 1\n - 100 units, activation ReLU\n - 50 units, activation ReLu\n - 1 unit, activation ReLu\n\n **Attributes**:\n\n Includes the same attributes as :class:`BasicImageSiamese`\n \"\"\"\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n\n def _conv_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method :func:`~BasicSiamese._conv_layers`\n\n :param x: Network's input images with shape ``[batch, 64, 64, 64, 1]``\n :return: Filtered image with the convolutions applied\n \"\"\"\n # In: [batch, 64, 64, 64, 1]\n\n device = '/gpu:0' if self._gpu_level >= 2 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for first conv layers\")\n with tf.device(device):\n # Out: [batch, 31, 31, 31, 30]\n x = tf.layers.conv3d(\n x,\n filters=30,\n kernel_size=3,\n strides=2,\n activation=tf.nn.relu,\n name=\"conv1\"\n )\n\n # Out: [batch, 29, 29, 29, 40]\n x = tf.layers.conv3d(\n x,\n filters=40,\n kernel_size=3,\n activation=tf.nn.relu,\n name=\"conv2\"\n )\n\n device = '/gpu:0' if self._gpu_level >= 1 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for second conv layers\")\n with tf.device(device):\n # Out: [batch, 27, 27, 27, 40]\n x = tf.layers.conv3d(\n x,\n filters=40,\n kernel_size=3,\n activation=tf.nn.relu,\n name=\"conv3\"\n )\n\n # Out: [batch, 25, 25, 25, 50]\n x = tf.layers.conv3d(\n x,\n filters=50,\n kernel_size=3,\n activation=tf.nn.relu,\n name=\"conv4\"\n )\n return x\n\n def _fc_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method ``BasicSiamese._fc_layers``\n\n :param x: Image, usually previously filtered with the convolutional layers.\n :return: Tensor with shape ``[batch, 1]``\n \"\"\"\n device = '/gpu:0' if self._gpu_level >= 3 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for FC layers\")\n with tf.device(device):\n # Out: [batch, 25*25*25*50]\n x = tf.layers.flatten(\n x,\n name=\"flat\"\n )\n\n # Out: [batch, 100]\n x = tf.layers.dense(\n x,\n 100,\n activation=tf.nn.relu,\n name=\"fc1\"\n )\n\n # Out: [batch, 50]\n x = tf.layers.dense(\n x,\n 50,\n activation=tf.nn.relu,\n name=\"fc2\"\n )\n\n # Out: [batch, 1]\n x = tf.layers.dense(\n x,\n 1,\n activation=tf.nn.relu,\n name=\"fc3\"\n )\n return x\n\n\nclass ImageScalarSiamese(BasicImageSiamese):\n r\"\"\"\n Siamese model that uses both images and scalar values as input.\n\n It uses scalar features extracted with PyRadiomics and provided through a Tensorflow ``placeholder``,\n the model assumes that there are :any:`settings.NUMBER_FEATURES` for each input.\n\n :param learning_rate:\n :param gpu_level: Amount of GPU that should be used with the model\n :param regularization: Regularization factor for the weights\n :param dropout: Dropout probability\n\n This class creates a Siamese model that uses both images and scalar features extracted using\n `PyRadiomics `_.\n The features are not extracted by the model, so they have to be provided in one of the placeholders\n\n **Network structure**:\n\n - :math:`3^3` kernel with 30 filters and stride = 2 with ReLu\n - :math:`3^3` kernel with 30 filters and stride = 2 with ReLu\n - :math:`3^3` kernel with 40 filters and stride = 1 with ReLu\n - :math:`3^3` kernel with 40 filters and stride = 1 with ReLu\n - :math:`3^3` kernel with 50 filters and stride = 1 with ReLu\n - :math:`3^3` kernel with 50 filters and stride = 1 with ReLu\n - Flattening layer\n - 8000 units, activation tanh\n - 100 units, activation tanh\n - 1 unit, activation ReLu\n\n **Attributes**:\n\n Includes the same attributes as :class:`BasicImageSiamese` and adds the following ones:\n\n :var ImageScalarSiamese.x_scalar: Scalar features obtained with\n `PyRadiomics `_\n :vartype ImageScalarSiamese.x_scalar: tf.Tensor\n \"\"\"\n\n def __init__(self, **kwargs):\n #: **Attribute**: Scalar features obtained with `PyRadiomics `_\n self.x_scalar = tf.placeholder(tf.float32, [None, settings.NUMBER_FEATURES], name=\"radiomic_features\")\n\n super().__init__(**kwargs)\n\n def _conv_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method :func:`~BasicSiamese._conv_layers`\n\n :param x: Network's input images with shape ``[batch, 64, 64, 64, 1]``\n :return: Filtered image with the convolutions applied\n \"\"\"\n # In: [batch, 64, 64, 64, 1]\n\n device = '/gpu:0' if self._gpu_level >= 2 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for first conv layers\")\n with tf.device(device):\n # Out: [batch, 31, 31, 31, 30]\n x = self._conv3d(\n x,\n filters=30,\n kernel_size=3,\n strides=2,\n name=\"conv1\"\n )\n\n # Out: [batch, 15, 15, 15, 30]\n x = self._conv3d(\n x,\n filters=30,\n kernel_size=3,\n strides=2,\n name=\"conv2\"\n )\n\n device = '/gpu:0' if self._gpu_level >= 1 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for second conv layers\")\n with tf.device(device):\n # Out: [batch, 13, 13, 13, 40]\n x = self._conv3d(\n x,\n filters=40,\n kernel_size=3,\n name=\"conv3\"\n )\n\n # Out: [batch, 11, 11, 11, 40]\n x = self._conv3d(\n x,\n filters=40,\n kernel_size=3,\n name=\"conv4\"\n )\n\n # Out: [batch, 9, 9, 9, 50]\n x = self._conv3d(\n x=x,\n filters=50,\n kernel_size=3,\n name=\"conv5\",\n )\n\n return x\n\n def _fc_layers(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n Implementation of abstract method :func:`~BasicSiamese._fc_layers`\n\n :param x: Image, usually previously filtered with the convolutional layers.\n :return: Tensor with shape ``[batch, 1]``\n \"\"\"\n\n # In this case we will be using the same idea seen in SimpleSiamese but we will be adding the scalar\n # features instead\n device = '/gpu:0' if self._gpu_level >= 3 else '/cpu:0'\n self.logger.debug(f\"Using device: {device} for FC layers\")\n with tf.device(device):\n # Out: [batch, 9*9*9*50]\n x = tf.layers.flatten(\n x,\n name=\"flat\"\n )\n\n # This is where the magic happens\n # Out: [batch, 37 175]\n x = tf.concat([x, self.x_scalar], axis=1)\n\n # Out: [batch, 8000]\n x = self._dense(\n x=x,\n units=8000,\n name=\"fc1\"\n )\n\n x = self._dense(\n x=x,\n units=1000,\n name=\"fc3\"\n )\n\n # x = tf.layers.dropout(\n # x,\n # rate=self._dropout,\n # training=self.training\n # )\n\n # Out: [batch, 10]\n x = self._dense(\n x=x,\n units=100,\n activation=tf.nn.relu,\n name=\"fc4\"\n )\n return x\n\n def _conv3d(self, x: tf.Tensor,\n filters: int,\n kernel_size: Union[int, Tuple],\n name: str,\n strides: Union[int, Tuple] = 1,\n activation: Any = tf.nn.relu,\n padding=\"valid\") -> tf.Tensor:\n return tf.layers.conv3d(\n name=name,\n inputs=x,\n filters=filters,\n kernel_size=kernel_size,\n strides=strides,\n activation=activation,\n kernel_initializer=tf.contrib.layers.variance_scaling_initializer(),\n kernel_regularizer=tf.contrib.layers.l2_regularizer(self._regularization),\n padding=padding\n )\n\n def _dense(self, x: tf.Tensor, units: int, name: str, activation=tf.nn.relu) -> tf.Tensor:\n return tf.layers.dense(\n x,\n units=units,\n activation=activation,\n kernel_initializer=tf.contrib.layers.xavier_initializer(),\n kernel_regularizer=tf.contrib.layers.l2_regularizer(self._regularization),\n name=name\n )\n\n def feed_dict(self, batch: data.PairBatch, training: bool = True) -> Dict:\n \"\"\"\n Re-implementation of :func:`~BasicSiamese.feed_dict` to create a custom dict including the scalar values\n\n :param batch: Data containing for the current batch, usually this would be generated by\n :func:`~BatchData.batches`\n :param training: Whether we are training or not. Useful for training layers like dropout where we do not\n want to apply dropout if we are not training\n :return: Return the ``feed_dict`` as a dictionary\n \"\"\"\n return {\n **super().feed_dict(batch, training=training),\n self.x_scalar: np.stack(batch.patients[\"features\"].values)\n }\n\n\nclass ResidualImageScalarSiamese(ImageScalarSiamese):\n\n def __init__(self, **kwargs):\n self.residual_count_a = 0\n self.residual_count_b = 0\n\n super().__init__(**kwargs)\n\n def _conv_layers(self, x: tf.Tensor) -> tf.Tensor:\n device = '/gpu:0' if self._gpu_level >= 1 else '/cpu:0'\n with tf.device(device):\n x = self._stem_block(x)\n\n for i in range(2):\n x = self._res_block_a(x)\n\n device = '/gpu:1' if self._gpu_level >= 1 else '/cpu:0'\n with tf.device(device):\n x = self._reduction_a(x)\n\n for i in range(2):\n x = self._res_block_b(x)\n\n # Out: [batch, 7, 7, 7, 350]\n x = self._reduction_b(x)\n\n # Out: [batch, 2, 2, 2, 350]\n x = tf.layers.average_pooling3d(\n inputs=x,\n pool_size=6,\n strides=1\n )\n\n return x\n\n def _fc_layers(self, x: tf.Tensor) -> tf.Tensor:\n\n # Out: [batch, 2*2*2*350] = [batch, 2800]\n x = tf.layers.flatten(x, name=\"flat\")\n\n # Out: [batch, 2800 + 725]\n x = tf.concat([x, self.x_scalar], axis=1)\n\n x = self._dense(\n x=x,\n units=800,\n name=\"fc_0\",\n )\n\n x = self._dense(\n x=x,\n units=100,\n name=\"fc_1\",\n )\n\n x = self._dense(\n x=x,\n units=10,\n name=\"fc_2\",\n activation=tf.nn.relu\n )\n\n return x\n\n def _stem_block(self, x: tf.Tensor) -> tf.Tensor:\n with tf.variable_scope(\"stem_reduce\"):\n\n # Out: [batch, 31, 31, 31, 25]\n x_a = self._conv3d(\n x=x,\n name=\"a_0_conv_3x3x3\",\n filters=25,\n kernel_size=3,\n strides=2,\n )\n\n # Out: [batch, 31, 31, 31, 1]\n x_b = tf.layers.max_pooling3d(\n inputs=x,\n name=\"b_0_pool_3x3x3\",\n pool_size=3,\n strides=2,\n )\n\n # Out: [batch, 31, 31, 31, 25]\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_1x1x1\",\n filters=25,\n kernel_size=1,\n padding=\"same\"\n )\n\n # Out: [batch, 31, 31, 31, 50]\n x_concat: tf.Tensor = tf.concat([x_a, x_b], axis=4)\n assert x_concat.get_shape()[-1] == 50\n\n self.logger.debug(x_concat.get_shape())\n\n return x_concat\n\n def _res_block_a(self, x: tf.Tensor, activation_fn=tf.nn.relu) -> tf.Tensor:\n \"\"\"\n Residual block with size ``[batch, 31, 31, 31, 50]``\n :param x:\n :return:\n \"\"\"\n\n with tf.variable_scope(f\"block_31_{self.residual_count_a}\"):\n self.residual_count_a += 1\n\n x_a = self._conv3d(\n x=x,\n name=\"a_0_conv_1x1x1\",\n filters=32,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x,\n name=\"b_0_conv_1x1x1\",\n filters=32,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_3x3x3\",\n filters=32,\n kernel_size=3,\n padding=\"same\"\n )\n\n x_c = self._conv3d(\n x=x,\n name=\"c_0_conv_1x1x1\",\n filters=32,\n kernel_size=1,\n padding=\"same\"\n )\n\n for i in range(1, 3):\n x_c = self._conv3d(\n x=x_c,\n name=f\"c_{i}_conv_3x3x3\",\n filters=32,\n kernel_size=3,\n padding=\"same\"\n )\n\n x_conv = tf.concat([x_a, x_b, x_c], axis=4)\n\n x_conv = self._conv3d(\n x=x_conv,\n name=\"conv_1x1\",\n filters=x.get_shape()[-1],\n kernel_size=1,\n padding=\"same\",\n activation=None\n )\n\n x += x_conv\n return activation_fn(x)\n\n def _reduction_a(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n :param x: Tensor with shape ``[batch, 31, 31, 31, 50]``\n :return: Tensor with shape ``[batch, 15, 15, 15, 130]``\n \"\"\"\n with tf.variable_scope(\"reduction_a\"):\n x_a = tf.layers.max_pooling3d(\n inputs=x,\n name=\"a_0_pooling_3x3x3\",\n pool_size=3,\n strides=2\n )\n\n x_b = self._conv3d(\n x=x,\n name=\"b_0_conv_3x3x3\",\n filters=50,\n kernel_size=3,\n strides=2\n )\n\n x_c = self._conv3d(\n x=x,\n name=\"c_0_conv_1x1x1\",\n filters=30,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_c = self._conv3d(\n x=x_c,\n name=\"c_1_conv_3x3x3\",\n filters=30,\n kernel_size=3,\n padding=\"same\"\n )\n\n x_c = self._conv3d(\n x=x_c,\n name=\"c_2_conv_3x3x3\",\n filters=30,\n kernel_size=3,\n strides=2\n )\n\n return tf.concat([x_a, x_b, x_c], axis=4)\n\n def _res_block_b(self, x: tf.Tensor, activation_fn=tf.nn.relu) -> tf.Tensor:\n \"\"\"\n :param x: Tensor with shape ``[batch, 15, 15, 15, 130]``\n :return: Tensor with shape ``[batch, 15, 15, 15, 130]``\n \"\"\"\n\n with tf.variable_scope(f\"block_15_{self.residual_count_b}\"):\n self.residual_count_b += 1\n\n x_a = self._conv3d(\n x=x,\n name=\"a_0_conv_1x1x1\",\n filters=50,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x,\n name=\"b_0_conv_1x1x1\",\n filters=50,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_1x1x7\",\n filters=50,\n kernel_size=(1, 1, 7),\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_1x7x1\",\n filters=50,\n kernel_size=(1, 7, 1),\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_7x1x1\",\n filters=50,\n kernel_size=(7, 1, 1),\n padding=\"same\"\n )\n\n x_concat = tf.concat([x_a, x_b], axis=4)\n x_conv = self._conv3d(\n x=x_concat,\n name=\"conv_1x1\",\n filters=x.get_shape()[-1],\n kernel_size=1,\n padding=\"same\",\n activation=None\n )\n\n x += x_conv\n return activation_fn(x)\n\n def _reduction_b(self, x: tf.Tensor) -> tf.Tensor:\n \"\"\"\n :param x: Tensor with shape ``[batch, 15, 15, 15, 130]``\n :return: Tensor with shape ``[batch, 7, 7, 7, 350]``\n \"\"\"\n\n with tf.variable_scope(\"reduction_b\"):\n x_a = self._conv3d(\n x=x,\n name=\"a_0_conv_3x3\",\n filters=100,\n kernel_size=3,\n strides=2\n )\n\n x_b = self._conv3d(\n x=x,\n name=\"b_0_conv_1x1\",\n filters=100,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_b = self._conv3d(\n x=x_b,\n name=\"b_1_conv_3x3\",\n filters=100,\n kernel_size=3,\n strides=2\n )\n\n x_c = self._conv3d(\n x=x,\n name=\"c_0_conv_1x1\",\n filters=100,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_c = self._conv3d(\n x=x_c,\n name=\"c_1_conv_3x3\",\n filters=100,\n kernel_size=3,\n strides=2\n )\n\n x_d = self._conv3d(\n x=x,\n name=\"d_0_conv_1x1\",\n filters=50,\n kernel_size=1,\n padding=\"same\"\n )\n\n x_d = self._conv3d(\n x=x_d,\n name=\"d_1_conv_3x3\",\n filters=50,\n kernel_size=3,\n padding=\"same\"\n )\n\n x_d = self._conv3d(\n x=x_d,\n name=\"d_2_conv_3x3\",\n filters=50,\n kernel_size=3,\n strides=2\n )\n\n # Out [batch, 7, 7, 7, 100 + 100 + 100 + 50]\n return tf.concat([x_a, x_b, x_c, x_d], axis=4)\n\n\nclass ScalarOnlySiamese(BasicSiamese):\n r\"\"\"\n Model that uses only radiomic features as input to train\n\n It has the same parameters as :class:`BasicSiamese`\n\n It only uses the radiomic features obtained with `PyRadiomics `_\n\n **Attributes**:\n\n Includes the same attributes as :class:`BasicSiamese` and adds the following ones:\n\n :var ScalarOnlySiamese.x_scalar: Radiomic features obtained with\n `PyRadiomics `_\n :vartype ScalarOnlySiamese.x_scalar: tf.Tensor\n \"\"\"\n\n def __init__(self, **kwargs):\n #: Radiomic features obtained with `PyRadiomics `_\n self.x_scalar = tf.placeholder(tf.float32, [None, settings.NUMBER_FEATURES])\n\n super().__init__(**kwargs)\n\n def _sister(self):\n # Out: [batch, 500]\n x = self.x_scalar\n x = self._dense(\n x,\n 500,\n \"fc1\"\n )\n\n # Out: [batch, 200]\n x = self._dense(\n x,\n 200,\n \"fc2\"\n )\n\n x = tf.layers.dropout(\n x,\n rate=self._dropout,\n training=self.training\n )\n\n # Out: [batch, 50]\n x = self._dense(\n x,\n 50,\n \"fc3\"\n )\n\n # Out: [batch, 1]\n x = self._dense(\n x,\n 10,\n \"fc4\",\n activation=tf.nn.relu\n )\n\n return x\n\n def _dense(self, x: tf.Tensor, units: int, name: str, activation=tf.nn.tanh) -> tf.Tensor:\n return tf.layers.dense(\n x,\n units=units,\n activation=activation,\n kernel_initializer=tf.contrib.layers.xavier_initializer(),\n kernel_regularizer=tf.contrib.layers.l2_regularizer(self._regularization),\n name=name\n )\n\n def feed_dict(self, batch: data.PairBatch, training: bool = True):\n\n return {\n **super().feed_dict(batch, training),\n self.x_scalar: np.stack(batch.patients[\"features\"]),\n }\n\n def uses_images(self) -> bool:\n \"\"\"\n Implementation of :func:`BasicModel.uses_images`. This model does not uses images to work.\n\n :return: :any:`False` since this model does not use images to work\n \"\"\"\n return False\n\n\nclass ScalarOnlyDropoutSiamese(ScalarOnlySiamese):\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n\n def _sister(self):\n # Out: [batch, 500]\n x = self.x_scalar\n x = self._dense(\n x,\n 500,\n \"fc1\"\n )\n\n x = tf.layers.dropout(\n x,\n rate=self._dropout,\n training=self.training\n )\n\n # Out: [batch, 200]\n x = self._dense(\n x,\n 200,\n \"fc2\"\n )\n\n x = tf.layers.dropout(\n x,\n rate=self._dropout,\n training=self.training\n )\n\n # Out: [batch, 50]\n x = self._dense(\n x,\n 50,\n \"fc3\"\n )\n\n x = tf.layers.dropout(\n x,\n rate=self._dropout,\n training=self.training\n )\n\n # Out: [batch, 1]\n x = self._dense(\n x,\n 10,\n \"fc4\",\n activation=tf.nn.relu\n )\n\n return x\n\n\nclass VolumeOnlySiamese(BasicSiamese):\n r\"\"\"\n Model that only uses the volume radiomic feature\n\n The features are provided by the package from `PyRadiomics `_\n\n It trains a model in the form :math:`y = w \\cdot V + b`\n\n **Attributes**:\n\n Includes the same attributes as :class:`BasicSiamese` and adds the following ones:\n\n :ivar VolumeOnlySiamese.x_volume: Placeholder for the volume feature\n :vartype VolumeOnlySiamese.x_volume: tf.Tensor\n \"\"\"\n\n def __init__(self, **kwargs):\n #: **Attribute**: Radiomic volume feature obtained with\n #: `PyRadiomics `_\n self.x_volume = tf.placeholder(tf.float32, [None, 1])\n\n super().__init__(**kwargs)\n\n def _sister(self) -> tf.Tensor:\n \"\"\"\n Super greedy predictor, more volume means less survival time so we only have to invert the volume size to\n create an inverse relation. This model does not have trainable variables\n\n :return: Greedy siamese applied\n \"\"\"\n\n total = tf.Variable(0., name=\"bias\")\n for i in range(1):\n w = tf.Variable(-1., name=f\"weight_{i}\")\n total = total + w*(self.x_volume**(i + 1))\n\n return total\n\n def feed_dict(self, batch: data.PairBatch, training: bool = True):\n\n features = np.stack(batch.patients[\"features\"].values)\n volumes = features[:, settings.VOLUME_FEATURE_INDEX].reshape((-1, 1))\n\n return {\n **super().feed_dict(batch, training),\n self.x_volume: volumes\n }\n\n def uses_images(self):\n return False\n","repo_name":"bhklab/DeepCINET","sub_path":"Sources/models/siameses.py","file_name":"siameses.py","file_ext":"py","file_size_in_byte":32310,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23225047838","text":"import cv2\nimport numpy as np\nimport os\nfrom matplotlib import pyplot as plt\n\n# infile = os.path.join('data', 'musician.jpg')\ninfile = os.path.join('data', 'calling.png')\noutfile = os.path.join('data', 'calling-edges.png')\n\nimg = cv2.imread(infile, 0)\nheight, width = img.shape\n# edges\t= cv2.Canny(image, threshold1, threshold2)\nedges = cv2.Canny(img, 100, 200)\n\n# Save the edge image in the original shape\n# https://www.infobyip.com/detectmonitordpi.php\nmy_dpi = 192\n\n# With margin\n# plt.figure(figsize=(width/my_dpi, height/my_dpi), dpi=my_dpi)\n# plt.imshow(edges, cmap='gray')\n# plt.axis('off')\n# plt.savefig(outfile, dpi=my_dpi)\n\n# Without margin\nplt.gca().set_axis_off()\nplt.subplots_adjust(top = 1, bottom = 0, right = 1, left = 0, hspace = 0, wspace = 0)\nplt.margins(0,0)\nplt.gca().xaxis.set_major_locator(plt.NullLocator())\nplt.gca().yaxis.set_major_locator(plt.NullLocator())\n\nplt.figure(figsize=(width/my_dpi, height/my_dpi), dpi=my_dpi)\nplt.imshow(edges, cmap='gray')\nplt.axis('off')\nplt.savefig(outfile, dpi=my_dpi, bbox_inches = 'tight', pad_inches = 0)\n\n# Show the two images side by side\n# plt.subplot(121),plt.imshow(img,cmap = 'gray')\n# plt.title('Original Image'), plt.xticks([]), plt.yticks([])\n# plt.subplot(122),plt.imshow(edges,cmap = 'gray')\n# plt.title('Edge Image'), plt.xticks([]), plt.yticks([])\n# plt.show()","repo_name":"tintinrevient/period-recreation","sub_path":"edge_detection/canny_edge_detector.py","file_name":"canny_edge_detector.py","file_ext":"py","file_size_in_byte":1335,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31631015468","text":"\"\"\"Topology example (:mod:`fluidimage.topologies.example`)\n==========================================================\n\nThis topology has two pythran cpu bounded tasks. It helps see executors behavior\nwith C functions.\n\n.. autoclass:: TopologyExample\n :members:\n :private-members:\n\n\"\"\"\n\nimport os\n\nimport numpy as np\nimport scipy.io\n\nfrom transonic import boost\n\nfrom . import TopologyBase\nfrom ..util import imread\n\nA = \"uint8[:,:]\"\n\n\n@boost\ndef cpu1(array1: A, array2: A, nloops: int = 10):\n\n a = np.arange(10000000 // nloops)\n result = a\n for i in range(nloops):\n result += a**3 + a**2 + 2\n\n for i in range(nloops):\n array1 = array1 * array2\n return (array1, array1)\n\n\n@boost\ndef cpu2(array1: A, array2: A, nloops: int = 10):\n\n a = np.arange(10000000 // nloops)\n result = a\n for i in range(nloops):\n result += a**3 + a**2 + 2\n\n for i in range(nloops):\n array1 = np.multiply(array1, array2)\n return array1\n\n\nclass TopologyExample(TopologyBase):\n \"\"\"Topology example for testing.\n\n Parameters\n ----------\n\n logging_level : str, {'warning', 'info', 'debug', ...}\n\n Logging level.\n\n nb_max_workers : None, int\n\n Maximum numbers of \"workers\". If None, a number is computed from the\n number of cores detected. If there are memory errors, you can try to\n decrease the number of workers.\n\n \"\"\"\n\n @classmethod\n def create_default_params(cls):\n params = dict(\n path_input=None,\n path_dir_result=None,\n nloops=1,\n multiplicator_nb_images=1,\n )\n return params\n\n def __init__(self, params, logging_level=\"info\", nb_max_workers=None):\n self.params = params\n\n path_input = params[\"path_input\"]\n path_dir_result = params[\"path_dir_result\"]\n nloops = params[\"nloops\"]\n self.multiplicator_nb_images = params[\"multiplicator_nb_images\"]\n\n def func1(arrays):\n key = arrays[0]\n if key == \"Karman_02_00\":\n raise ValueError(\"For testing\")\n arr0, arr1 = cpu1(arrays[1], arrays[2], nloops)\n return key, arr0, arr1\n\n def func2(arrays):\n key = arrays[0]\n result = cpu2(arrays[1], arrays[2], nloops)\n return key, result\n\n self.path_input = path_input\n\n super().__init__(\n path_dir_result=path_dir_result,\n logging_level=logging_level,\n nb_max_workers=nb_max_workers,\n )\n\n if not self.path_dir_result.exists():\n self.path_dir_result.mkdir()\n\n self.img_counter = 0\n\n queue_names = self.add_queue(\"names images\")\n queue_couples_names = self.add_queue(\"names couples\")\n queue_arrays = self.add_queue(\"arrays\")\n queue_couples_arrays = self.add_queue(\"couples arrays\")\n queue_cpu1 = self.add_queue(\"queue cpu1\")\n queue_cpu2 = self.add_queue(\"queue cpu2\")\n\n self.add_work(\n \"fill names\",\n func_or_cls=self.fill_names,\n output_queue=queue_names,\n kind=(\"global\", \"one shot\"),\n )\n self.add_work(\n \"fill names couples\",\n func_or_cls=self.fill_couples_names,\n input_queue=queue_names,\n output_queue=queue_couples_names,\n kind=(\"global\", \"one shot\"),\n )\n\n self.add_work(\n \"read array\",\n func_or_cls=self.read_array,\n input_queue=queue_names,\n output_queue=queue_arrays,\n kind=\"io\",\n )\n self.add_work(\n \"fill couples arrays\",\n func_or_cls=self.fill_couples_arrays,\n input_queue=(queue_couples_names, queue_arrays),\n output_queue=queue_couples_arrays,\n kind=(\"global\"),\n )\n self.add_work(\n \"cpu1\",\n func_or_cls=func1,\n input_queue=queue_couples_arrays,\n output_queue=queue_cpu1,\n kind=\"cpu\",\n )\n self.add_work(\n \"cpu2\",\n func_or_cls=func2,\n input_queue=queue_cpu1,\n output_queue=queue_cpu2,\n kind=\"cpu\",\n )\n self.add_work(\n \"save\", func_or_cls=self.save, input_queue=queue_cpu2, kind=\"io\"\n )\n\n def fill_names(self, input_queue, output_queue):\n for ind in range(self.multiplicator_nb_images):\n for name in sorted(os.listdir(self.path_input)):\n key = name.split(\".bmp\")[0] + f\"_{ind:02}\"\n output_queue[key] = name\n\n def fill_couples_names(self, input_queue, output_queue):\n for key, name in list(input_queue.items()):\n output_queue[key] = [key, (name, name)]\n\n def read_array(self, name):\n if name == \"Karman_03.bmp\":\n raise ValueError(\"For testing\")\n\n image = imread(self.path_input / name)\n return image\n\n def fill_couples_arrays(self, input_queues, output_queue):\n queue_couples_names, queue_arrays = input_queues\n queue_couples_arrays = output_queue\n\n for key, array in list(queue_arrays.items()):\n if key not in queue_couples_names:\n continue\n queue_arrays.pop(key)\n queue_couples_names.pop(key)\n # propagating possible exceptions...\n if isinstance(array, Exception):\n queue_couples_arrays[key] = array\n else:\n queue_couples_arrays[key] = [key, array, array]\n\n def save(self, inputs):\n key = inputs[0]\n arr = inputs[1]\n name = key + \".h5\"\n scipy.io.savemat(self.path_dir_result / name, mdict={\"array\": arr})\n","repo_name":"fluiddyn/fluidimage","sub_path":"fluidimage/topologies/example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":5707,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"81"}
+{"seq_id":"69978405384","text":"from django.contrib import admin\nfrom .models import Page, Chapter\n\nclass PageAdmin(admin.ModelAdmin):\n readonly_fields = ('created_at', 'updated_at')\n\nclass ChapterAdmin(admin.ModelAdmin):\n readonly_fields = ('created_at',)\n\n# Register your models here.\nadmin.site.register(Page, PageAdmin)\nadmin.site.register(Chapter, ChapterAdmin)\n\ntitle = \"Panel de Administración IOSA | Aerolíneas Argentinas\"\nsubtitle = \"Panel de Gestión\"\n\nadmin.site.site_header = title\nadmin.site.site_title = title\nadmin.site.index_title = subtitle","repo_name":"BauerPablo/presentacion_IOSA","sub_path":"pages/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":533,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14433993422","text":"from flask import Flask, render_template, jsonify, request\nfrom pymongo import MongoClient\nimport time\n\napp = Flask(__name__)\n\nclient = MongoClient(\"mongodb://localhost:27017/\")\ndb = client.dbStock\n\n\n\n@app.route('/')\ndef index():\n return render_template('index.html')\n\n\n@app.route('/post', methods=['POST'])\ndef save_post():\n print('clicked')\n idx = request.form['idx']\n title = request.form['title_give']\n comment = request.form['comment_give']\n nowTime = time.strftime('%Y.%m.%d %X')\n doc = {\n 'idx' : idx,\n 'title' : title,\n 'comment' : comment,\n 'reg_date' : nowTime\n }\n db.posts.insert_one(doc)\n return {\"result\": \"success\"}\n\n\n@app.route('/post', methods=['GET'])\ndef get_post():\n posts = list(db.posts.find({}, {'_id' : False}))\n return jsonify({'all_posts' : posts})\n\n\n@app.route('/post', methods=['DELETE'])\ndef delete_post():\n return {\"result\": \"success\"}\n\n\nif __name__ == \"__main__\":\n app.run(host=\"0.0.0.0\", port=5000)","repo_name":"Ohjinn/TimeAttack2","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":999,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72082403465","text":"# Name - Jatin\n# Entry Number - 2020CSB1090\n\nimport numpy as np\nfrom math import log10, sqrt\nfrom numpy import dot, exp, mgrid, pi, ravel, square, uint8, zeros\nimport skimage\nfrom skimage.color import *\nfrom skimage import feature\nfrom matplotlib.image import imread\nimport matplotlib.pyplot as plt\nfrom skimage.metrics import structural_similarity\n\n\ndef peak_signal_to_noise_ratio(original, compressed):\n mse = np.mean((original - compressed) ** 2)\n if (mse == 0):\n return 100\n max_pixel = 255.0\n psnr = 20 * log10(max_pixel / sqrt(mse))\n return psnr\n\n\ndef myCannyEdgeDetector(inputImg, lowThresholdRatio, highThresholdRatio):\n\n # gaussian filter Noise Reduction\n gx, gy = np.meshgrid(np.arange(-7/2+1, 7/2+1), np.arange(-7/2+1, 7/2+1))\n\n NORMALISED = 1 / (2.0 * np.pi * 1**2)\n\n Kernal = np.exp(-(gx**2+gy**2) / (2.0*1**2)) / \\\n NORMALISED\n\n size_kernal, gaussi_filter = Kernal.shape[0], np.zeros_like(\n img, dtype=float)\n\n for i in range(img.shape[0]-(size_kernal-1)):\n for j in range(img.shape[1]-(size_kernal-1)):\n window = img[i:i+size_kernal, j:j+size_kernal] * Kernal\n gaussi_filter[i, j] = np.sum(window)\n\n # plt.imshow(gaussi_filter, cmap='gray')\n # plt.show()\n\n # sobel filter\n sobel_x = np.array([[1.0, 0.0, -1.0], [2.0, 0.0, -2.0], [1.0, 0.0, -1.0]])\n sobel_y = np.array([[1.0, 2.0, 1.0], [0.0, 0.0, 0.0], [-1.0, -2.0, -1.0]])\n [rows, columns] = np.shape(gaussi_filter)\n sobel_filtered_image = np.zeros(shape=(rows, columns))\n\n theta = np.zeros(shape=(rows, columns))\n\n for i in range(rows - 2):\n for j in range(columns - 2):\n gx = np.sum(np.multiply(sobel_x, gaussi_filter[i:i + 3, j:j + 3]))\n gy = np.sum(np.multiply(sobel_y, gaussi_filter[i:i + 3, j:j + 3]))\n sobel_filtered_image[i + 1, j + 1] = np.sqrt(gx ** 2 + gy ** 2)\n if (gx == 0):\n theta[i+1, j+1] = 90\n else:\n theta[i+1, j+1] = ((np.arctan(gy/gx))/np.pi) * 180\n\n sobel_filtered_image = sobel_filtered_image/np.max(sobel_filtered_image)\n # plt.imshow(sobel_filtered_image, cmap='gray')\n # plt.show()\n\n # Nonmaximum suppression\n nms = np.copy(sobel_filtered_image)\n\n for i in range(theta.shape[0]-(2)):\n for j in range(theta.shape[1]-(2)):\n if theta[i][j] < 0:\n theta[i][j] += 180\n\n if (theta[i, j] <= 22.5 or theta[i, j] > 157.5):\n if (sobel_filtered_image[i, j] <= sobel_filtered_image[i-1, j]) and (sobel_filtered_image[i, j] <= sobel_filtered_image[i+1, j]):\n nms[i, j] = 0\n if (theta[i, j] > 22.5 and theta[i, j] <= 67.5):\n if (sobel_filtered_image[i, j] <= sobel_filtered_image[i-1, j-1]) and (sobel_filtered_image[i, j] <= sobel_filtered_image[i+1, j+1]):\n nms[i, j] = 0\n if (theta[i, j] > 67.5 and theta[i, j] <= 112.5):\n if (sobel_filtered_image[i, j] <= sobel_filtered_image[i+1, j+1]) and (sobel_filtered_image[i, j] <= sobel_filtered_image[i-1, j-1]):\n nms[i, j] = 0\n if (theta[i, j] > 112.5 and theta[i, j] <= 157.5):\n if (sobel_filtered_image[i, j] <= sobel_filtered_image[i+1, j-1]) and (sobel_filtered_image[i, j] <= sobel_filtered_image[i-1, j+1]):\n nms[i, j] = 0\n\n nms = nms/np.max(nms)\n # plt.imshow(nms, cmap='gray')\n # plt.show()\n\n # Double THRESHING\n Threash = np.copy(nms)\n h = int(Threash.shape[0])\n w = int(Threash.shape[1])\n highThreshold = np.max(Threash) * highThresholdRatio\n lowThreshold = highThreshold * lowThresholdRatio\n\n for i in range(1, h-1):\n for j in range(1, w-1):\n if (Threash[i, j] > highThreshold):\n Threash[i, j] = 1\n elif (Threash[i, j] < lowThreshold):\n Threash[i, j] = 0\n else:\n if ((Threash[i-1, j-1] > highThreshold) or\n (Threash[i-1, j] > highThreshold) or\n (Threash[i-1, j+1] > highThreshold) or\n (Threash[i, j-1] > highThreshold) or\n (Threash[i, j+1] > highThreshold) or\n (Threash[i+1, j-1] > highThreshold) or\n (Threash[i+1, j] > highThreshold) or\n (Threash[i+1, j+1] > highThreshold)):\n Threash[i, j] = 1\n\n return Threash\n\n\nimg = skimage.io.imread('7.jpg')\nimg = rgb2gray(img)\noutputImg = myCannyEdgeDetector(img, 0.2, 0.4)\n\nplt.figure()\nplt.imshow(outputImg, cmap='gray')\nplt.title(\"My canny Edge Detector\")\nplt.show()\ntrue_canny = feature.canny(img)\nplt.imshow(true_canny, cmap='gray')\nplt.title(\"Inbuilt_Function\")\nplt.show()\npsnr = peak_signal_to_noise_ratio(true_canny, outputImg)\nprint(\"Peak Signal To Noise Ratio: \", end=' ')\nprint(psnr)\nssim1 = structural_similarity(outputImg, true_canny, data_range=255)\nprint(\"Structural Similarity Index Metric: \", end=' ')\nprint(ssim1)\n","repo_name":"jatin1322/CannyEdgeDetector","sub_path":"MyCannyEdgeDetectorDemo.py","file_name":"MyCannyEdgeDetectorDemo.py","file_ext":"py","file_size_in_byte":5015,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"20335379892","text":"#We add libraries\r\nfrom PyQt5.QtCore import Qt\r\nfrom PyQt5.QtWidgets import QApplication, QWidget, QPushButton, QLabel, QVBoxLayout, QLineEdit, QHBoxLayout\r\nfrom datetime import datetime\r\nimport time\r\n\r\n\r\n\r\ndef confirmation(): \r\n def exitt():\r\n main_win.hide()\r\n\r\n edit_day_text=int(edit_day.text())\r\n edit_month_text=int(edit_month.text())\r\n edit_year_text=int(edit_year.text())\r\n year1 = year-edit_year_text\r\n if year1 > 16:\r\n text.setText('Ваша вікова категорія підходить')\r\n butt = QPushButton('продовжити')\r\n line_v.addWidget(butt, alignment = Qt.AlignCenter)\r\n butt.clicked.connect(exitt)\r\n elif year1 == 16:\r\n if month > edit_month_text:\r\n text.setText('Ваша вікова категорія підходить')\r\n butt = QPushButton('продовжити')\r\n line_v.addWidget(butt, alignment = Qt.AlignCenter)\r\n butt.clicked.connect(exitt)\r\n elif month == edit_month_text:\r\n if day >= edit_day_text:\r\n text.setText('Ваша вікова категорія підходить')\r\n butt = QPushButton('продовжити')\r\n line_v.addWidget(butt, alignment = Qt.AlignCenter)\r\n butt.clicked.connect(exitt)\r\n else:\r\n text.setText('Вибачте але ваша вікова категорія не підходить')\r\n butt2 = QPushButton('продовжити')\r\n line_v.addWidget(butt2, alignment = Qt.AlignCenter)\r\n butt2.clicked.connect(quit)\r\n else:\r\n text.setText('Вибачте але ваша вікова категорія не підходить')\r\n butt2 = QPushButton('продовжити')\r\n line_v.addWidget(butt2, alignment = Qt.AlignCenter)\r\n butt2.clicked.connect(quit) \r\n else:\r\n text.setText('Вибачте але ваша вікова категорія не підходить')\r\n butt2 = QPushButton('продовжити')\r\n line_v.addWidget(butt2, alignment = Qt.AlignCenter)\r\n butt2.clicked.connect(quit)\r\n\r\ndef questionnaire():\r\n def continuation():\r\n def exitt1():\r\n main_win1.hide()\r\n\r\n dani.setText('Ваші дані збережено')\r\n butt1 = QPushButton('Продовжити')\r\n line_v1.addWidget(butt1, alignment = Qt.AlignCenter)\r\n butt1.clicked.connect(exitt1)\r\n\r\n app1 = QApplication([])\r\n main_win1 = QWidget()\r\n\r\n main_win1.resize(400,300)\r\n main_win1.move(460, 280)\r\n\r\n main_win1.setWindowTitle('АНКЕТА АБІТУРІЄНТА')\r\n \r\n dani = QLabel(\"Ваші дані\")\r\n name = QLabel(\"Ваше прізвище, ім'я та по-батькові\")\r\n live = QLabel(\"Домашня адреса (Область, район, місто/село)\")\r\n gmail = QLabel(\"Ваш e-mail\")\r\n number = QLabel(\"Номер телефону\")\r\n surroundings = QLabel(\"Розкажіть про себе(Ваші уподобання, хобі, коло інтересів)\")\r\n\r\n but1 = QPushButton('Підтвердити')\r\n\r\n name_l = QLineEdit()\r\n live_l = QLineEdit()\r\n gmail_l = QLineEdit()\r\n number_l = QLineEdit()\r\n surroundings_l = QLineEdit()\r\n\r\n line_v1 = QVBoxLayout()\r\n\r\n line_v1.addWidget(dani, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(name, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(name_l, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(live, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(live_l, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(gmail, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(gmail_l, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(number, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(number_l, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(surroundings, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(surroundings_l, alignment = Qt.AlignCenter)\r\n line_v1.addWidget(but1, alignment = Qt.AlignCenter)\r\n\r\n main_win1.setLayout(line_v1)\r\n\r\n but1.clicked.connect(continuation)\r\n\r\n main_win1.show()\r\n app1.exec_()\r\n\r\nquestionnaire()\r\n\r\n#Let's take the current time\r\ndate = datetime.now()\r\n\r\nyear = date.year\r\nmonth = date.month\r\nday = date.day\r\n\r\n#We create \"application\" and \"window\" objects\r\napp = QApplication([])\r\nmain_win = QWidget()\r\n\r\n#Window dimensions and location coordinates\r\nmain_win.resize(400,300)\r\nmain_win.move(460, 280)\r\n\r\n#Interface name\r\nmain_win.setWindowTitle('АНКЕТА АБІТУРІЄНТА')\r\n\r\n\r\n\r\n#Text widgets\r\ntext = QLabel('Ведіть свою дату народження')\r\nday_1 = QLabel('Чісло')\r\nmonth_1 = QLabel('Місяць')\r\nyear_1 = QLabel('Рік')\r\n\r\n#input strings\r\nedit_day = QLineEdit('')\r\nedit_month = QLineEdit('')\r\nedit_year = QLineEdit('')\r\n\r\n#\r\nbut = QPushButton('Підтвердити')\r\nbutt = QPushButton('продовжити')\r\n\r\n#We create vertical and horizontal lines\r\nline_v = QVBoxLayout()\r\nline_h = QHBoxLayout()\r\nline_h1 = QHBoxLayout()\r\n\r\n\r\nline_v.addWidget(text, alignment = Qt.AlignCenter)\r\nline_h1.addWidget(day_1, alignment = Qt.AlignLeft)\r\nline_h1.addWidget(month_1, alignment = Qt.AlignLeft)\r\nline_h1.addWidget(year_1, alignment = Qt.AlignLeft)\r\nline_v.addLayout(line_h1)\r\nline_h.addWidget(edit_day, alignment = Qt.AlignLeft)\r\nline_h.addWidget(edit_month, alignment = Qt.AlignCenter)\r\nline_h.addWidget(edit_year, alignment = Qt.AlignRight)\r\nline_v.addLayout(line_h)\r\nline_v.addWidget(but, alignment = Qt.AlignCenter)\r\n\r\nmain_win.setLayout(line_v)\r\n#We make the window visible and leave the program open until the exit button is pressed\r\n\r\nbut.clicked.connect(confirmation)\r\n\r\nmain_win.show()\r\napp.exec_()","repo_name":"Yarykl/hak.","sub_path":"one.py","file_name":"one.py","file_ext":"py","file_size_in_byte":5827,"program_lang":"python","lang":"uk","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31898732430","text":"from datetime import datetime\n\nfrom aiogram import types\n\nfrom tgbot.models.models import User\n\n\nclass UserCommands:\n\n async def get_user(self, user_id):\n user = await User.query.where(User.user_id == user_id).gino.first()\n return user\n\n async def add_new_user(self):\n user = types.User.get_current()\n old_user = await self.get_user(user.id)\n if old_user:\n return old_user\n new_user = User()\n new_user.user_id = user.id\n new_user.username = user.username\n new_user.full_name = user.full_name\n new_user.reg_date = datetime.now()\n await new_user.create()\n return new_user\n","repo_name":"codebegemot/okoloBot","sub_path":"tgbot/models/user_commands.py","file_name":"user_commands.py","file_ext":"py","file_size_in_byte":671,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"17525131600","text":"import os\nimport argparse\n\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\n\nfrom datasets import Dataset, DatasetDict\n\nos.makedirs('data/', exist_ok=True)\n\nSEED = 42\n\n# limit_tweets\ndef limit_tweets(tweets, num_tweets):\n tweets = tweets.split(sep='|||') # Each tweet are separated by \"|||\"\n tweets = tweets[:num_tweets]\n tweets = '|||'.join(tweets)\n return tweets\n\n# Indexing tweets, assuming each tweet is separated by \"|||\"\ndef index_tweets(tweets):\n sep_tweets = tweets.split(\"|||\")\n \n indexed_tweets = [] \n for i, tweet in enumerate(sep_tweets):\n indexed_tweets.append(f\"{i+1}. {tweet}\")\n \n indexed_tweets = \"\\n\".join(indexed_tweets)\n \n return indexed_tweets\n \n# Save preprocessed data\ndef save_preprocessed_dataset(path: str, dataset):\n dataset.save_to_disk(path)\n\ndef parse_args():\n ap = argparse.ArgumentParser()\n ap.add_argument(\"-num_tweets\", type=int, default=25)\n ap.add_argument(\"-original_data\", type=str, default=\"data/kaggle.csv\")\n ap.add_argument(\"-save_dir\", type=str, default=\"data/processed_data\")\n args = ap.parse_args()\n return (\n args.num_tweets,\n args.original_data,\n args.save_dir\n )\n\n\nif __name__==\"__main__\":\n (\n num_tweets,\n original_data,\n save_dir,\n ) = parse_args()\n\n\n # Load Kaggle dataset for MBTI\n df = pd.read_csv(original_data)\n\n # Preprocess text data\n df['text'] = df['text'].apply(lambda x: limit_tweets(x, num_tweets)) # Decrease the number of tweets for each sample by 25\n df['text'] = df['text'].apply(lambda x: index_tweets(x))\n\n # Split into train/test dataset\n text_train, text_test, label_train, label_test = train_test_split(df['text'], df['type'], test_size=0.1, random_state=SEED)\n\n # Store data in DatasetDict\n dataset = DatasetDict()\n dataset[\"train\"] = Dataset.from_dict(\n {\"text\": text_train, \"label\": label_train}\n )\n dataset[\"test\"] = Dataset.from_dict(\n {\"text\": text_test, \"label\": label_test}\n )\n\n save_preprocessed_dataset(save_dir, dataset)","repo_name":"Yuta555/LLM-personality-evaluation","sub_path":"personality_detection_model/scripts/preprocess_data.py","file_name":"preprocess_data.py","file_ext":"py","file_size_in_byte":2109,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"33812874465","text":"\"\"\"requests session object\"\"\"\nfrom __future__ import annotations\n\nimport io\nimport json\n\nfrom urllib.parse import urljoin\nfrom typing import Literal\n\nfrom pyetm.exceptions import UnprossesableEntityError, format_error_messages\n\nimport requests\nimport pandas as pd\n\nDecoder = Literal['bytes', 'BytesIO', 'json', 'text']\nMethod = Literal['delete', 'get', 'post', 'put']\n\n\nclass RequestsSession:\n \"\"\"requests-based session object\"\"\"\n\n @property\n def base_url(self) -> str | None:\n \"\"\"base url used in make_url\"\"\"\n return self.__base_url\n\n @base_url.setter\n def base_url(self, base_url: str | None) -> None:\n\n if base_url is None:\n self.__base_url = base_url\n\n if base_url is not None:\n self.__base_url = str(base_url)\n\n @property\n def headers(self) -> dict:\n \"\"\"headers that are passed in each request\"\"\"\n return self.__headers\n\n @headers.setter\n def headers(self, headers: dict | None) -> None:\n\n if headers is None:\n headers = {}\n\n self.__headers = dict(headers)\n\n def __init__(self, base_url: str | None = None,\n headers: dict | None = None, proxies: dict | None = None,\n stream: bool = False, verify: bool | str = True,\n cert: str | tuple | None = None):\n \"\"\"session object for pyETM clients\n\n Parameters\n ----------\n base_url: str, default None\n Base url to which the session connects, all request urls\n will be merged with the base url to create a destination.\n headers : dict, default None\n Headers that are always passed during requests, e.g. an\n authorization token.\n proxies: dict, default None\n Dictionary mapping protocol or protocol and\n hostname to the URL of the proxy.\n stream: boolean, default False\n Whether to immediately download the response content.\n verify: boolean or string, default True\n Either a boolean, in which case it controls whether we verify\n the server's TLS certificate, or a string, in which case it must\n be a path to a CA bundle to use. When set to False, requests will\n accept any TLS certificate presented by the server, and will ignore\n hostname mismatches and/or expired certificates, which will make\n your application vulnerable to man-in-the-middle (MitM) attacks.\n Setting verify to False may be useful during local development or\n testing.\n cert: string or tuple, default None\n If string; path to ssl client cert file (.pem).\n If tuple; ('cert', 'key') pair.\"\"\"\n\n # set parameters\n self.base_url = base_url\n self.headers = headers\n\n # set environment kwargs for method requests\n self._request_env = {\n \"proxies\": proxies, \"stream\": stream,\n \"verify\": verify, \"cert\": cert}\n\n # set session\n self._session = requests.Session()\n\n def __repr__(self):\n \"\"\"reproduction string\"\"\"\n\n # object environment\n env = \", \".join(f\"{k}={str(v)}\" for k, v in\n self._request_env.items())\n\n return f\"RequestsSession({env})\"\n\n def __str__(self):\n \"\"\"stringname\"\"\"\n return 'RequestsSession'\n\n def __enter__(self) -> RequestsSession:\n \"\"\"enter context manager\"\"\"\n\n # connect session\n self.connect()\n\n return self\n\n def __exit__(self, *args, **kwargs) -> None:\n \"\"\"exit context manager\"\"\"\n\n # close session\n self.close()\n\n def make_url(self, url: str | None = None):\n \"\"\"join url with base url\"\"\"\n return urljoin(self.base_url, url)\n\n def connect(self):\n \"\"\"connect session\"\"\"\n\n def close(self):\n \"\"\"close session\"\"\"\n\n def request(self, method: Method, url: str,\n decoder: Decoder = 'bytes', **kwargs):\n \"\"\"make request to api session\"\"\"\n\n retries = 5\n while retries:\n\n try:\n\n # merge kwargs with session envioronment kwargs\n kwargs = {**self._request_env, **kwargs}\n\n # add persistent headers\n headers = kwargs.get('headers', {})\n kwargs['headers'] = {**headers, **self.headers}\n\n # make method request\n request = getattr(self._session, method)\n with request(url, **kwargs) as resp:\n\n # check response\n if not resp.ok:\n\n # get debug message\n if resp.status_code == 422:\n self._error_report(resp)\n\n # raise for status\n resp.raise_for_status()\n\n # bytes decoding\n if decoder == \"bytes\":\n resp = resp.content\n\n # bytes as BytesIO\n elif decoder == \"BytesIO\":\n byts = resp.content\n resp = io.BytesIO(byts)\n\n # json decoding\n elif decoder == \"json\":\n resp = resp.json()\n\n # text decoding\n elif decoder == \"text\":\n resp = resp.text\n\n else:\n msg = f\"decoding method '{decoder}' not implemented\"\n raise NotImplementedError(msg)\n\n return resp\n\n # except connectionerrors and retry\n except requests.exceptions.ConnectionError as error:\n retries -= 1\n\n # raise after retries\n if not retries:\n raise error\n\n def _error_report(self, resp: requests.Response) -> None:\n \"\"\"create error report when api returns error messages.\"\"\"\n\n try:\n\n # attempt decode error message(s)\n msg = resp.json()\n errors = msg.get(\"errors\")\n\n except json.decoder.JSONDecodeError:\n\n # no message returned\n errors = None\n\n if errors:\n\n # format error message(s)\n msg = format_error_messages(errors)\n raise UnprossesableEntityError(msg)\n\n def delete(self, url: str | None = None,\n decoder: Decoder = 'text', **kwargs):\n \"\"\"delete request\"\"\"\n return self.request(\"delete\", self.make_url(url), decoder, **kwargs)\n\n def get(self, url: str | None = None,\n decoder: Decoder = 'json', **kwargs):\n \"\"\"get request\"\"\"\n return self.request(\"get\", self.make_url(url), decoder, **kwargs)\n\n def post(self, url: str | None = None,\n decoder: Decoder = 'json', **kwargs):\n \"\"\"post request\"\"\"\n return self.request(\"post\", self.make_url(url), decoder, **kwargs)\n\n def put(self, url: str | None = None,\n decoder: Decoder = 'json', **kwargs):\n \"\"\"put request\"\"\"\n return self.request(\"put\", self.make_url(url), decoder, **kwargs)\n\n def upload_series(self, url: str | None = None,\n series: pd.Series | None = None, name: str | None = None, **kwargs):\n \"\"\"upload series object\"\"\"\n\n # default to empty series\n if series is None:\n series = pd.Series()\n\n # set key as name\n if name is None:\n name = \"not specified\"\n\n # convert series to string\n data = series.to_string(index=False)\n form = {\"file\": (name, data)}\n\n return self.put(url=url, files=form, **kwargs)\n","repo_name":"robcalon/pyetm","sub_path":"src/pyetm/sessions/requests.py","file_name":"requests.py","file_ext":"py","file_size_in_byte":7600,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"21716746180","text":"from copy import deepcopy\n\nfrom common import sign\n\ninputs = [x.strip() for x in open(\"2022/inputs/14.txt\").readlines()]\n\n\nrocks = set()\nabyss = 0\nfor input in inputs:\n points = [complex(*[int(y) for y in x.split(\",\")]) for x in input.split(\" -> \")]\n\n for i in range(len(points) - 1):\n a, b = points[i : i + 2]\n rocks.add(a)\n rocks.add(b)\n\n while a != b:\n c = b - a\n if c.real:\n a += sign(c.real)\n if c.imag:\n a += sign(c.imag) * 1j\n rocks.add(a)\n abyss = max(abyss, a.imag + 1)\n\nsand = 0\nsource = 500\nrocks_1 = deepcopy(rocks)\nwhile source.imag <= abyss:\n down, left, right = source + 1j, source + 1j - 1, source + 1j + 1\n if down not in rocks_1:\n source = down\n elif left not in rocks_1:\n source = left\n elif right not in rocks_1:\n source = right\n else:\n rocks_1.add(source)\n sand += 1\n source = 500\n\nprint(sand)\n\nsand = 0\nsource = 500\nrocks_2 = deepcopy(rocks)\nwhile 500 not in rocks_2:\n down, left, right = source + 1j, source + 1j - 1, source + 1j + 1\n if source.imag >= abyss:\n rocks_2.add(source)\n sand += 1\n source = 500\n elif down not in rocks_2:\n source = down\n elif left not in rocks_2:\n source = left\n elif right not in rocks_2:\n source = right\n else:\n rocks_2.add(source)\n sand += 1\n source = 500\n\nprint(sand)\n","repo_name":"jaredculp/aoc","sub_path":"2022/src/14.py","file_name":"14.py","file_ext":"py","file_size_in_byte":1480,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1864616026","text":"from __future__ import absolute_import\n\nimport te.lang.cce\nfrom te import tvm\nfrom te.platform.fusion_manager import fusion_manager\nfrom topi import generic\nfrom topi.cce import util\nfrom functools import reduce as reduceIns\nfrom te import platform as tbe_platform\n\n# shape limit for aicore equals 2**31\nSHAPE_SIZE_LIMIT = 2147483648\n\n\n# pylint: disable=locally-disabled,too-many-arguments,unused-argument\n@fusion_manager.register(\"sqrt\")\ndef sqrt_compute(input_data, output_data, kernel_name=\"sqrt\"):\n \"\"\"\n calculating data sqrt,y= x**0.5,mini not support vsqrt, use exp(0.5*log(x))\n\n Parameters\n ----------\n input_data: TVM tensor\n the placeholder of input data\n output_data: dict\n shape and dtype of output, should be same shape and type as input\n kernel_name: str\n cce kernel name, default value is sqrt\n\n Returns\n -------\n result: TVM tensor\n the result of sqrt\n \"\"\"\n dtype = input_data.dtype\n has_improve_precision = False\n if dtype == \"float16\" and\\\n tbe_platform.cce_conf.api_check_support(\"te.lang.cce.vsqrt\",\n \"float32\"):\n input_data = te.lang.cce.cast_to(input_data, \"float32\")\n has_improve_precision = True\n result = te.lang.cce.vsqrt(input_data)\n\n if has_improve_precision:\n result = te.lang.cce.cast_to(result, \"float16\")\n\n return result\n\n\n@util.check_input_type(dict, dict, str)\ndef sqrt(input_x, output_y, kernel_name=\"sqrt\"):\n \"\"\"\n algorithm: sqrt\n calculating data sqrt,y= x**0.5, mini not support vsqrt, use exp(0.5*log(x))\n\n Parameters\n ----------\n input_x : dict\n shape and dtype of input, only support float16, float32\n output_y: dict\n shape and dtype of output, should be same shape and type as input\n kernel_name : str\n cce kernel name, default value is sqrt\n\n Returns\n -------\n None\n \"\"\"\n input_shape = input_x.get(\"shape\")\n input_dtype = input_x.get(\"dtype\").lower()\n\n util.check_kernel_name(kernel_name)\n util.check_shape_rule(input_shape)\n util.check_shape_size(input_shape, SHAPE_SIZE_LIMIT)\n util.check_dtype_rule(input_dtype, (\"float16\", \"float32\"))\n\n fuseshape = [1]\n fuseshape[0] = reduceIns(lambda x, y: x*y, input_shape)\n input_data = tvm.placeholder(fuseshape, name=\"input_data\",\n dtype=input_dtype)\n result = sqrt_compute(input_data, output_y, kernel_name)\n\n with tvm.target.cce():\n sch = generic.auto_schedule(result)\n\n config = {\"print_ir\": False,\n \"name\": kernel_name,\n \"tensor_list\": [input_data, result]}\n\n te.lang.cce.cce_build_code(sch, config)\n","repo_name":"jizhuoran/caffe-huawei-atlas-convertor","sub_path":"convertor/huawei/impl/sqrt.py","file_name":"sqrt.py","file_ext":"py","file_size_in_byte":2712,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"81"}
+{"seq_id":"13733614171","text":"import os\nimport errno\n\ndotfiles_dir = os.path.join(os.environ['HOME'], 'dotfiles')\nEXCLUDE_FILES = [\n __file__,\n 'README.md'\n]\n\nfor filename in [f for f in os.listdir(dotfiles_dir) if f not in EXCLUDE_FILES]:\n dotfile_versioning = os.path.join(dotfiles_dir, filename)\n dotfile_inuse = os.path.join(os.environ['HOME'], filename)\n try:\n os.symlink(\n dotfile_versioning,\n dotfile_inuse\n )\n except Exception as e:\n if e.errno == errno.EEXIST:\n os.unlink(dotfile_inuse)\n os.symlink(\n dotfile_versioning,\n dotfile_inuse\n )\n else:\n raise e\n\n","repo_name":"nkimoto/dotfiles","sub_path":"export.py","file_name":"export.py","file_ext":"py","file_size_in_byte":677,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4126806502","text":"import os, sys, tqdm, math, pickle\nimport numpy as np\nimport tensorflow as tf\n\nCM_PATH = './HOME/DATASETs/charmap.pickle'\n\ndef prepareXi(X, CM, MAX_LEN):\n CMm = max(CM.values()) + 1\n\n f = lambda x: CM[x] if x in CM else CMm\n Xi = np.zeros((len(X), MAX_LEN), np.uint8) \n for i, x in enumerate(X):\n xi = list(map(f, x))\n Xi[i, :len(xi)] = xi \n return Xi\n\ndef loadModel(path):\n gpus = tf.config.experimental.list_physical_devices('GPU') \n tf.config.experimental.set_memory_growth(gpus[0], True)\n return tf.keras.models.load_model(path, compile=False)\n\ndef inference(model, X, batch_size, MAX_LEN=16, CM_path=CM_PATH):\n \n with open(CM_path, 'rb') as f:\n CM = pickle.load(f)\n vocab_size = max(CM.values()) + 1\n \n Xi = prepareXi(X, CM, MAX_LEN)\n \n nbatch = math.ceil(len(Xi) / batch_size)\n \n for i in tqdm.trange(nbatch):\n xi = Xi[batch_size*i:batch_size*(i+1)]\n out = model(xi, training=False)[1].numpy()\n\n yield out\n","repo_name":"TheAdamProject/adams","sub_path":"NeuralNet/inference.py","file_name":"inference.py","file_ext":"py","file_size_in_byte":1019,"program_lang":"python","lang":"en","doc_type":"code","stars":19,"dataset":"github-code","pt":"81"}
+{"seq_id":"30119213945","text":"class direccion:\n def __init__(self):\n self.__calle = ''\n self.__piso = ''\n self.__ciudad = ''\n self.__cp = ''\n def getCalle(self):\n return self.__calle\n def getPiso(self):\n return self.__piso\n def getCiudad(self):\n return self.__ciudad\n def getCp(self):\n return self.__cp\n def setCalle(self,calle):\n self.__calle = calle\n def setPiso(self,piso):\n self.__piso = piso\n def setCiudad(self,ciudad):\n self.__ciudad = ciudad\n def setCp(self,cp):\n self.__cp = cp\n\nclass persona:\n def __init__(self):\n self.__nombre = ''\n self.__apellidos = ''\n self.__fechaNac = ''\n def getNombre(self):\n return self.__nombre\n def getApellidos(self):\n return self.__apellidos\n def getFechaNac(self):\n return self.__fechaNac\n def setNombre(self,nombre):\n self.__nombre = nombre\n def setApellidos(self,apellidos):\n self.__apellidos = apellidos\n def setFechaNac(self,fecha):\n self.__fechaNac = fecha\n\nclass telefono:\n def __init__(self):\n self.__movil = ''\n self.__fijo = ''\n self.__trabajo = ''\n def getMovil(self):\n return self.__movil\n def getFijo(self):\n return self.__fijo\n def getTrabajo(self):\n return self.__trabajo\n def setMovil(self,movil):\n self.__movil = movil\n def setFijo(self,fijo):\n self.__fijo = fijo\n def setTrabajo(self,trabajo):\n self.__trabajo = trabajo\n\nclass Contacto(direccion,persona,telefono):\n def __init__(self):\n self.__email = ''\n def getEmail(self):\n return self.__email\n def setEmail(self,mail):\n self.__email = mail\n def showContacto(self):\n print('---CONTACTO---')\n print('Nombre:',self.getNombre())\n print('Apellidos:',self.getApellidos())\n print('Fecha de nacimiento:',self.getFechaNac())\n print('Teléfono móvil:',self.getMovil())\n print('Teléfono fijo:',self.getFijo())\n print('Teléfono trabajo:',self.getTrabajo())\n print('Calle:',self.getCalle())\n print('Piso:',self.getPiso())\n print('Ciudad:',self.getCiudad())\n print('C.P.:',self.getCp())\n print('Email:',self.getEmail())\n\nclass agenda:\n \n def __init__(self,path):\n self.__listaContactos = []\n self.__path = path\n \n def cargarContactos(self):\n try:\n fichero = open(self.__path,'r')\n except:\n print('ERROR: El fichero no existe')\n else:\n contactos = fichero.readlines()\n fichero.close()\n if(len(contactos)>0):\n for contacto in contactos:\n datos = contacto.split('#')\n if(len(datos)==11):\n nuevoContacto = Contacto()\n nuevoContacto.setNombre(datos[0])\n nuevoContacto.setApellidos(datos[1])\n nuevoContacto.setFechaNac(datos[2])\n nuevoContacto.setMovil(datos[3])\n nuevoContacto.setFijo(datos[4])\n nuevoContacto.setTrabajo(datos[5])\n nuevoContacto.setCalle(datos[6])\n nuevoContacto.setPiso(datos[7])\n nuevoContacto.setCiudad(datos[8])\n nuevoContacto.setCp(datos[9])\n nuevoContacto.setEmail(datos[10])\n self.__listaContactos = self.__listaContactos + [nuevoContacto]\n print('INFO: Cargados',len(self.__listaContactos),'contactos')\n \n def crearContacto(self,contacto):\n self.__listaContactos = self.__listaContactos + [contacto]\n\n def guardarContacto(self):\n try:\n fichero = open(self.__path,'w')\n except:\n print('ERROR: El fichero no se puede guardar')\n else:\n for contacto in self.__listaContactos:\n texto = contacto.getNombre() + '#'\n texto = texto + contacto.getApellidos() + '#'\n texto = texto + contacto.getFechaNac() + '#'\n texto = texto + contacto.getMovil() + '#'\n texto = texto + contacto.getFijo() + '#'\n texto = texto + contacto.getTrabajo() + '#'\n texto = texto + contacto.getCalle() + '#'\n texto = texto + contacto.getPiso() + '#'\n texto = texto + contacto.getCiudad() + '#'\n texto = texto + contacto.getCp() + '#'\n texto = texto + contacto.getEmail() + '\\n'\n fichero.write(texto)\n fichero.close()\n\n def mostrarAgenda(self):\n print('### Agenda ###')\n print('Numero de contactos:',len(self.__listaContactos),'\\n')\n for contacto in self.__listaContactos:\n contacto.showContacto()\n print('######')\n\n def buscarContacto(self,tipo,dato):\n listaEncontrados = []\n for contacto in self.__listaContactos:\n if tipo == 1:\n if contacto.getNombre() == dato:\n listaEncontrados = listaEncontrados + [contacto] \n elif tipo == 2:\n if contacto.getMovil() == dato:\n listaEncontrados = listaEncontrados + [contacto]\n elif tipo == 3:\n if contacto.getFijo() == dato:\n listaEncontrados = listaEncontrados + [contacto]\n elif tipo == 4:\n if contacto.getTrabajo() == dato:\n listaEncontrados = listaEncontrados + [contacto]\n return listaEncontrados\n\n def borrarContacto(self,tipo,dato):\n listaFinal = []\n for contacto in self.__listaContactos:\n if tipo == 1:\n if contacto.getNombre() != dato:\n listaFinal = listaFinal + [contacto] \n elif tipo == 2:\n if contacto.getMovil() != dato:\n listaFinal = listaFinal + [contacto]\n elif tipo == 3:\n if contacto.getFijo() != dato:\n listaFinal = listaFinal + [contacto]\n elif tipo == 4:\n if contacto.getTrabajo() != dato:\n listaFinal = listaFinal + [contacto]\n print('INFO:',len(self.__listaContactos)-len(listaFinal),'contactos han sido borrados')\n self.__listaContactos = listaFinal\n\ndef obtenerOpcion(texto):\n leido = False\n while not leido:\n try:\n numero = int(input(texto))\n except ValueError:\n print('El valor debe ser un numero')\n else:\n leido = True\n return numero\n\ndef mostrarMenu():\n print('########## MENÚ PRINCIPAL ####################')\n print('1 - Mostrar contactos')\n print('2 - Buscar contactos')\n print('3 - Crear nuevo contacto')\n print('4 - Borrar contactos')\n print('5 - Guardar contactos')\n print('6 - Salir')\n\ndef buscarContactos(agenda):\n print('Buscar contactos')\n print('1 - Nombre')\n print('2 - Movil')\n print('3 - Fijo')\n print('4 - Trabajo')\n print('5 - Salir')\n finBuscar = False\n while not finBuscar:\n opcion = obtenerOpcion('Opción de búsqueda:')\n if opcion == 5:\n finBuscar = True\n encontrados = agenda.buscarContacto(opcion,input('Introduce el valor:'))\n if len(encontrados) > 0:\n print('### CONTACTOS ENCONTRADOS ###')\n for item in encontrados:\n item.showContacto()\n print('######')\n else:\n print('INFO: No se han encontrado contactos')\n\ndef procesoCrearContacto(agenda):\n nuevoContacto = Contacto()\n nuevoContacto.setNombre(input('Introduce el nombre:'))\n nuevoContacto.setApellidos(input('Introduce los apellidos:'))\n nuevoContacto.setFechaNac(input('Introduce la fecha de nacimiento:'))\n nuevoContacto.setMovil(input('Introduce el movil:'))\n nuevoContacto.setFijo(input('Introduce el fijo:'))\n nuevoContacto.setTrabajo(input('Introduce el teléfono del trabajo:'))\n nuevoContacto.setCalle(input('Introduce la calle:'))\n nuevoContacto.setPiso(input('Introduce el piso:'))\n nuevoContacto.setCiudad(input('Introduce la ciudad:'))\n nuevoContacto.setCp(input('Introduce el C.P.:'))\n nuevoContacto.setEmail(input('Introduce el email:'))\n agenda.crearContacto(nuevoContacto)\n\ndef borrarContacto(agenda):\n print('Borrar contacto')\n print('1 - Nombre')\n print('2 - Movil')\n print('3 - Fijo')\n print('4 - Trabajo')\n print('5 - Salir')\n finBuscar = False\n while not finBuscar:\n opcion = obtenerOpcion('Opción de borrado:')\n if opcion == 5:\n finBuscar = True\n else:\n encontrados = agenda.borrarContacto(opcion,input('Introduce el valor:'))\n finBuscar = True\n\ndef Main():\n nuevaAgenda = agenda('./files/agenda.txt')\n nuevaAgenda.cargarContactos()\n fin = False\n while not fin:\n mostrarMenu()\n opcion = obtenerOpcion('Opción:')\n if(opcion==1):\n nuevaAgenda.mostrarAgenda()\n elif(opcion==2):\n buscarContactos(nuevaAgenda)\n elif(opcion==3):\n procesoCrearContacto(nuevaAgenda)\n elif(opcion==4):\n borrarContacto(nuevaAgenda)\n elif(opcion==5):\n nuevaAgenda.guardarContacto()\n elif(opcion==6):\n fin = True\n\nMain()\n \n","repo_name":"victoraagg/python-test","sub_path":"12.py","file_name":"12.py","file_ext":"py","file_size_in_byte":9496,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"70202311304","text":"#Importing Libraries\nimport gspread\nfrom oauth2client.service_account import ServiceAccountCredentials\nimport json\n\n#Important stuff\nscope = ['https://spreadsheets.google.com/feeds', 'https://www.googleapis.com/auth/drive']\ncreds = ServiceAccountCredentials.from_json_keyfile_name(\n 'dividend-investing-274019-b556916d3607.json', scope) # Your json file here\nclient = gspread.authorize(creds)\n\n#Access sheet\nspreadsheetName = 'The Dividend Chamipon'\nsheet = client.open(spreadsheetName)\nworksheet = sheet.get_worksheet(4)\n\n#Request ticker\nsymbol = input('Enter ticker symbol: ')\nsymbol = symbol.upper()\n\n#Lookup ticker \nticker_range = worksheet.range('B7:B859')\n\n#Lookup dividend\ndiv_range = worksheet.range('J7:J859')\n\n#Match the ticker with dividend\nfor cell in ticker_range:\n ticker_val = cell.value\n ticker_row = cell.row\n ticker_col = cell.col\n if ticker_val == symbol:\n dividend = worksheet.cell(ticker_row, 10).value\n print('Dividend: ' + dividend)\n\n\n\n","repo_name":"ummaromanasama/Candlestick_Chart","sub_path":"googlesheet.py","file_name":"googlesheet.py","file_ext":"py","file_size_in_byte":1001,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"35310809040","text":"'''\nThis module implements a simple classroom example of probabilistic inference\nover the full joint distribution specified by AIMA, Figure 13.3.\nIt is based on the code from AIMA probability.py.\n\n@author: kvlinden\n@version Jan 1, 2013\n'''\n\nfrom probability import JointProbDist, enumerate_joint_ask\n'''\nExercise 4.2\ni. There are 16 entries in the table now.\nii. The probabilities sum up to 1 and they should. The probabilities in a sample space always sums to 1.\niii. No. We can use 0 and 1 to denote True or False, but essentially they can only take on these values. This is because\nin the real world an event either happens or doesn't happen (not considering Schrodinger's cat scenario)\niv. No, the probability of rain is not independent of the probability of toothache. Since people seem to have toothaches\nmore often when it rains, I chose the probability of rain to be 0.6 given toothache, and probability of rain = 0.3\nwhen no toothache.\n\nP(toothache | rain) = P(toothache ^ rain) / P(rain)\n = (0.0648 + 0.0072 + 0.0096 + 0.0384) \n / (0.0648 + 0.0072 + 0.0096 + 0.0384 + 0.0216 + 0.0024 + 0.0432 + 0.1728)\n ~= 0.333\nP(-toothache | rain) = 1 - 0.333 = 0.667\nP(Toothahce | rain) = <0.333, 0.667> \n\n'''\n# The Joint Probability Distribution Fig. 13.3 with added Rain variable (from AIMA Python)\nPr = JointProbDist(['Toothache', 'Rain', 'Cavity', 'Catch'])\nT, F = True, False\nPr[T, T, T, T] = 0.0648;\nPr[T, T, F, T] = 0.0096\nPr[F, T, T, T] = 0.0216;\nPr[F, T, F, T] = 0.0432\nPr[T, F, T, T] = 0.0432;\nPr[T, F, F, T] = 0.0064\nPr[F, F, T, T] = 0.0504;\nPr[F, F, F, T] = 0.1008\nPr[T, T, T, F] = 0.0072;\nPr[T, T, F, F] = 0.0384\nPr[F, T, T, F] = 0.0024;\nPr[F, T, F, F] = 0.1728\nPr[T, F, T, F] = 0.0048;\nPr[T, F, F, F] = 0.0256\nPr[F, F, T, F] = 0.0056;\nPr[F, F, F, F] = 0.4032\n\nPtr = enumerate_joint_ask('Toothache', {'Rain': T}, Pr)\nprint(Ptr.show_approx())","repo_name":"zchen0925/Artificial-Intelligence-course-projects","sub_path":"lab04/lab_2.py","file_name":"lab_2.py","file_ext":"py","file_size_in_byte":1912,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31439191568","text":"\r\nquestions = [\"Are you a man or woman?\\n1 - Man\\n2 - Woman\\n0 - Quit\",\r\n \"What is your name?\",\r\n \"How old are you?\\n1 - under 20\\n2 - 21 - 30\\n3 - 31 - 40\\n4 - Mature\\n0 - Quit\",\r\n \"Are you in a relationship?\\n1 - Yes\\n2 - No\\n0 - Quit\",\r\n \"What is your occupation? (Policeman, engineer, etc.)\"]\r\n\r\ndiscourage_m = [\"You remind me of a song: Dude looks like a lady.\",\r\n \"What a stupid name.\",\r\n \"Who did you have to suck off to get that job?\"]\r\ndiscourage_f = [\"Really? Manly jawline and those eye brows, Yikes.\",\r\n \"Cheap name for a cheap girl.\",\r\n \"Who did you have to suck off to get that job?\"]\r\n\r\nrelationship_m_d = [\"Hard to believe.\", \"Momma is still #1.\"]\r\nrelationship_f_d = [\"He is cheating on you.\", \"Still aiming too high.\"]\r\n\r\ndiscourage_age_m = [\"So.. still a virgin. Good.\",\r\n \"By now you must know that girls do not like you?\",\r\n \"Say goodbye to your early morning erection.\",\r\n \"Why are you still here?\"]\r\ndiscourage_age_f = [\"Inexperienced in life but not in bed.\",\r\n \"I want to be your 111th boyfriend.\",\r\n \"Woman's 30's is a man's 50's, getting old and useless.\",\r\n \"I met the fathers of your children, all five of them.\"]\r\n\r\nmotivate_m = [\"Wrong. You are The Man.\",\r\n \"A proper name for a Hero.\",\r\n \"You are a gift for Mankind.\"]\r\nmotivate_f = [\"And a very beautiful one.\",\r\n \"Lovely.\",\r\n \"With a smile like that they should pay you extra!\"]\r\nmotivate_age_m = [\"A Stud in the making.\",\r\n \"Full of testosterone, watch out ladies!\",\r\n \"A Man in his prime.\",\r\n \"You still got the charms, handsome.\"]\r\nmotivate_age_f = [\"Young and Beautiful, will you marry me?\",\r\n \"They still don't deserve you!\",\r\n \"Nothing beats beauty with experience.\",\r\n \"Still more than anyone deserves.\"]\r\nmotivate_rel_m = [\"Of course you are, you handsome Stud-Muffin!\",\r\n \"Too handsome for your own good.\"]\r\nmotivate_rel_f = [\"I did not want to hear that. :(\",\r\n \"Really? Well, I am single.. if you are interested in an automaton.\"]\r\n\r\nisMale = True\r\n\r\n\r\ndef setMode():\r\n print(\"EMOT-O-TRON >> Select mode:\")\r\n print(\"1 - Motivate me.\")\r\n print(\"2 - Discourage me.\") \r\n print(\"0 - I do not want to play this game. (Quit)\")\r\n index = input(\"Choice: \")\r\n return index\r\n\r\ndef discourage(a, q): \r\n\r\n global isMale\r\n print(\"\\n\")\r\n if q == 0 and a == '1':\r\n isMale = True\r\n print(discourage_m[q])\r\n elif q == 0 and a == '2':\r\n isMale = False\r\n print(discourage_f[q])\r\n\r\n if isMale:\r\n if q == 1:\r\n print(discourage_m[q])\r\n elif q == 2:\r\n print(discourage_age_m[int(a) - 1])\r\n elif q == 3:\r\n print(relationship_m_d[int(a) - 1])\r\n elif q == 4:\r\n print(discourage_m[q - 2])\r\n else:\r\n if q == 1:\r\n print(discourage_f[q])\r\n elif q == 2:\r\n print(discourage_age_f[int(a) - 1])\r\n elif q == 3:\r\n print(relationship_f_d[int(a) - 1])\r\n elif q == 4:\r\n print(discourage_f[q - 2])\r\n\r\n print(\"\\n\")\r\n\r\ndef motivate(a, q):\r\n\r\n global isMale\r\n print(\"\\n\")\r\n \r\n if q == 0 and a == '1':\r\n isMale = True\r\n print(motivate_m[q])\r\n elif q == 0 and a == '2':\r\n isMale = False\r\n print(motivate_f[q])\r\n\r\n if isMale:\r\n if q == 1:\r\n print(motivate_m[q])\r\n elif q == 2:\r\n print(motivate_age_m[int(a) - 1])\r\n elif q == 3:\r\n print(motivate_rel_m[int(a) - 1])\r\n elif q == 4:\r\n print(motivate_m[q - 2])\r\n else:\r\n if q == 1:\r\n print(motivate_f[q])\r\n elif q == 2:\r\n print(motivate_age_f[int(a) - 1])\r\n elif q == 3:\r\n print(motivate_rel_f[int(a) - 1])\r\n elif q == 4:\r\n print(motivate_f[q - 2])\r\n\r\n print(\"\\n\")\r\n\r\ndef interact(message):\r\n print(\"EMOT-O-TRON >> \\n{}\".format(message)) \r\n answer = input(\"Answer: \")\r\n return answer\r\n\r\n\r\ndef getMessage(q):\r\n return questions[q]\r\n\r\n\r\ndef main(): \r\n run = True\r\n q = 0\r\n\r\n index = setMode()\r\n\r\n if index == '0':\r\n run = False\r\n print(\"Goodbye\")\r\n \r\n while run:\r\n if q <= 4:\r\n answer = interact(getMessage(q))\r\n if index == '0' or answer == '0':\r\n run = False\r\n print(\"Goodbye\")\r\n elif index == '1':\r\n motivate(answer, q)\r\n elif index == '2':\r\n discourage(answer, q)\r\n else:\r\n print(\"Goodbye and good luck.\")\r\n run = False\r\n break;\r\n \r\n q += 1\r\n\r\n if answer == '0':\r\n run = False\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","repo_name":"mkkjhn/emototron","sub_path":"emot.py","file_name":"emot.py","file_ext":"py","file_size_in_byte":5086,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"29678126574","text":"#!/usr/bin/python3\r\n######################################################################\r\n# File Name: config_vars.py\r\n#\r\n# Description: Program configuration variables\r\n#\r\n# File History\r\n# 03/14/2021 - Andrew Yoder : Initial Release\r\n# 11/06/2021 - Andrew Yoder : Specifically call out python3\r\n# 01/03/2023 - Andrew Yoder : Updated to pull provider details from database\r\n# instead of being hardcoded in this file\r\n######################################################################\r\n\r\nimport sqlite3\r\n\r\nfrom platform_config import pt_db\r\n\r\n #connect to db and fetch all contents\r\ndb = sqlite3.connect(pt_db)\r\ncursor = db.cursor()\r\ncursor.execute('''SELECT * FROM provider_details''')\r\nprovider_details_data = cursor.fetchall()\r\n\r\n # Print all clients and information\r\nfor provider in provider_details_data:\r\n provider_name = provider[1]\r\n provider_title = provider[2]\r\n provider_phone = provider[3]\r\n provider_email = provider[4] \r\n provider_location = provider[5] + \", \" + provider[6] + \", \" + provider[7]\r\n paypal_link = provider[8]","repo_name":"ayoder770/Project-Tracking-System","sub_path":"python/config_vars.py","file_name":"config_vars.py","file_ext":"py","file_size_in_byte":1092,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"74008966026","text":"\"\"\"\n@author: Alfons\n@contact: alfons_xh@163.com\n@file: timer.py\n@time: 18-10-8 下午10:44\n@version: v1.0 \n\"\"\"\nimport time\n\n\ndef timer(func):\n def wrapper(*args, **kwargs):\n startTime = time.time()\n ret = func(*args, **kwargs)\n stopTime = time.time()\n\n print(\"{func} use time -> {time}'s\".format(func=func.__name__,\n time=stopTime - startTime))\n return ret\n\n return wrapper\n","repo_name":"Alfonsxh/AlgorithmsAreaNotes","sub_path":"Python/timer.py","file_name":"timer.py","file_ext":"py","file_size_in_byte":465,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"18002178805","text":"import torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nfrom torch.utils.data import DataLoader\nimport torchvision.datasets as datasets\nimport torchvision.transforms as transforms\nimport matplotlib.pyplot as plt\n\n\n#creates network\nclass network(nn.Module):\n def __init__(self, inputSize, outputSize):\n super(network,self).__init__()\n #This is a 1 layer network\n #input layer has inputSize nodes, first hidden layer has 50 nodes, output layer has outputSize nodes\n self.fc1 = nn.Linear(inputSize, 100)\n self.fc2 = nn.Linear(100,50)\n self.fc3 = nn.Linear(50,outputSize)\n \n def forward(self, x):\n x = self.fc1(x)\n x = F.relu(x)\n x = self.fc2(x)\n x = F.relu(x)\n x = self.fc3(x)\n return x\n \ndef run():\n #Sets device\n device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n\n #Parameters\n inputSize = 784\n numClasses = 10\n learningRate = 0.001\n batchSize = 64 #Batch size fed through the neuro network\n numEpoch = 3 #Number of iterations through the neuro network\n\n #Loading training data\n trainDataset = datasets.MNIST(root='dataset/', train=True, transform=transforms.ToTensor(), download = True)\n trainLoader = DataLoader(dataset = trainDataset, batch_size = batchSize, shuffle = True) \n\n #Loading testing data\n testDataset = datasets.MNIST(root='dataset/', train=False, transform=transforms.ToTensor(), download = True)\n testLoader = DataLoader(dataset = testDataset, batch_size = batchSize, shuffle = True)\n\n \n \n #Initializing network to device\n model = network(inputSize = inputSize, outputSize = numClasses).to(device)\n\n #Loss and optimizer\n criteron = nn.CrossEntropyLoss()\n optimizer = optim.Adam(model.parameters(),lr = learningRate)\n\n #Training network\n for epoch in range(numEpoch):\n for batchId, (data, target) in enumerate(trainLoader):\n #Puts data into GPU\n data = data.to(device = device)\n target = target.to(device = device)\n \n #Reshaping data\n data = data.reshape(data.shape[0],-1)\n \n #Forward\n scores = model(data)\n loss = criteron(scores,target)\n \n #Backward\n optimizer.zero_grad()\n loss.backward()\n \n #Gradient descent / adam step\n optimizer.step()\n \n def accuracy(loader, model):\n numCorrect = 0\n numSamples = 0\n \n model.eval()\n \n with torch.no_grad():\n for x, y in loader:\n x = x.to(device = device)\n y = y.to(device = device)\n \n x = x.reshape(x.shape[0],-1)\n scores = model(x)\n _, predictions = scores.max(1)\n numCorrect += (predictions == y).sum()\n numSamples += predictions.size(0)\n \n print(f'Got {numCorrect} / {numSamples} with a accuracy of {((float(numCorrect)/float(numSamples))*100):.2f}')\n \n model.train()\n\n accuracy(trainLoader, model)\n accuracy(testLoader, model)\n \n torch.save(model.state_dict(), 'modelTrained.pt')\n","repo_name":"FaJacos/Number_Detection","sub_path":"model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":3307,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"20036808807","text":"import csv\nimport os\nimport unicodedata\nimport re\nfrom typing import Tuple, List\nimport time\nimport math\n\nimport pandas as pd\nimport numpy as np\nimport torch\nfrom torch import Tensor, nn\nfrom torchtext.data import get_tokenizer\n\n\nclass LanguageDictionary:\n \"\"\"Collects word metadata for building one-hot-encoding vector.\"\"\"\n SOS_token = 0\n EOS_token = 1\n\n def __init__(self, name: str):\n self.name = name\n self.word2index = {}\n self.word2count = {}\n self.index2word = {LanguageDictionary.SOS_token: 'SOS', LanguageDictionary.EOS_token: 'EOS'}\n self.n_words = 2\n\n def add_word(self, word: str):\n if word not in self.word2index:\n self.word2index[word] = self.n_words\n self.word2count[word] = 1\n self.index2word[self.n_words] = word\n self.n_words += 1\n else:\n self.word2count[word] += 1\n\n def add_sentence(self, sentence: str):\n # TODO: improve tokenization\n for word in sentence.split(' '):\n self.add_word(word)\n\n\ndef unicode_to_ascii(s):\n \"\"\"Turn a Unicode string to plain ASCII, thanks to\n https://stackoverflow.com/a/518232/2809427\n \"\"\"\n return ''.join(\n c for c in unicodedata.normalize('NFD', s)\n if unicodedata.category(c) != 'Mn'\n )\n\n\ndef normalize_string(s):\n \"\"\"Lowercase, trim, and remove non-letter characters\"\"\"\n s = unicode_to_ascii(s.lower().strip())\n s = re.sub(r\"([.!?])\", r\" \\1\", s)\n s = re.sub(r\"[^a-zA-Z.!?]+\", r\" \", s)\n return s\n\n\ndef read_parallel_corpus(lang1: str, lang2: str, reverse: bool = False) -> Tuple[LanguageDictionary,\n LanguageDictionary,\n List[List[str]]]:\n \"\"\"Reads a parallel text corpus from a text file. Assumes that there are two columns separated by\n a tab character. Words are separated by spaces.\"\"\"\n print(\"Reading lines...\")\n\n # Read the file and split into lines\n txt_path = os.path.join('..', f'data/{lang1}-{lang2}.txt')\n lines = open(txt_path, encoding='utf-8').read().strip().split('\\n')\n\n # Split every line into pairs and normalize\n pairs = [[normalize_string(s) for s in l.split('\\t')] for l in lines]\n\n # Reverse pairs\n if reverse:\n pairs = [list(reversed(p)) for p in pairs]\n\n # Make language dictionaries\n return LanguageDictionary(lang1), LanguageDictionary(lang2), pairs\n\n\nMAX_SENTENCE_LENGTH = 10\n\neng_prefixes = (\n \"i am \", \"i m \",\n \"he is\", \"he s \",\n \"she is\", \"she s \",\n \"you are\", \"you re \",\n \"we are\", \"we re \",\n \"they are\", \"they re \"\n)\n\n\ndef filter_pair(language_pairs: List[str]):\n assert len(language_pairs) == 2, \"Should always be of length 2\"\n return len(language_pairs[0].split(' ')) < MAX_SENTENCE_LENGTH and \\\n len(language_pairs[1].split(' ')) < MAX_SENTENCE_LENGTH and \\\n language_pairs[1].startswith(eng_prefixes)\n\n\ndef filter_pairs(pairs: List[List[str]]):\n return [pair for pair in pairs if filter_pair(pair)]\n\n\ndef prepare_data(lang1: str, lang2: str, reverse: bool = False):\n input_dict, output_dict, pairs = read_parallel_corpus(lang1, lang2, reverse)\n print(f\"Read {len(pairs)} sentence pairs\")\n pairs = filter_pairs(pairs)\n print(f\"Trimmed to {len(pairs)} sentence pairs\")\n for text in pairs:\n input_dict.add_sentence(text[0])\n output_dict.add_sentence(text[1])\n print(f\"Counted words:\")\n print(f\"Input language {lang1} - {input_dict.n_words}\")\n print(f\"Output language {lang2} - {output_dict.n_words}\")\n return input_dict, output_dict, pairs\n\n\ndef as_minutes(s):\n m = math.floor(s / 60)\n s -= m * 60\n return '%dm %ds' % (m, s)\n\n\ndef time_since(since, percent):\n seconds = time.time() - since\n remaining_seconds = (seconds / percent) - seconds\n return '%s (- %s)' % (as_minutes(seconds), as_minutes(remaining_seconds))\n\n\ndef indexes_from_sentence(lang_dictionary: LanguageDictionary, sentence: str) -> List[int]:\n return [lang_dictionary.word2index[word] for word in sentence.split(' ')]\n\n\ndef tensor_from_sentence(lang_dictionary: LanguageDictionary, sentence: str) -> Tensor:\n indexes = indexes_from_sentence(lang_dictionary, sentence)\n indexes.append(LanguageDictionary.EOS_token)\n return torch.tensor(indexes, dtype=torch.long).view(-1, 1)\n\n\ndef tensors_from_pair(input_lang: LanguageDictionary, output_lang: LanguageDictionary,\n pair: List[str]) -> Tuple[Tensor, Tensor]:\n input_tensor = tensor_from_sentence(input_lang, pair[0])\n target_tensor = tensor_from_sentence(output_lang, pair[1])\n return input_tensor, target_tensor\n\n\ndef write_losses(output_path: str, training_loss: List[float], validation_loss: List[float]):\n \"\"\"Write training and validation losses to CSV file.\"\"\"\n assert len(training_loss) == len(validation_loss), f\"{len(training_loss)} \" \\\n f\"!= {len(validation_loss)}\"\n with open(output_path, 'w') as f:\n writer = csv.writer(f)\n writer.writerow(['epoch', 'training_loss', 'validation_loss'])\n for i, (train, val) in enumerate(zip(training_loss, validation_loss)):\n writer.writerow([i, train, val])\n print(f\"Train and validation losses are output to {output_path}\")\n\n\ndef cosine_similarity(t1: Tensor, t2: Tensor):\n cos = nn.CosineSimilarity(dim=0, eps=1e-6)\n return cos(t1, t2)\n\n\ndef l2_norm(t1: Tensor, t2: Tensor):\n return torch.norm(torch.dot(t1, t2))\n\n\ndef train_val_test_split(data_path: str, output_dir: str) -> None:\n \"\"\"Create train/val/test split from randomized input data and write to an output directory.\"\"\"\n df = pd.read_csv(data_path, sep='\\t')\n train_df, val_df, test_df = np.split(df.sample(frac=1), [int(.8 * len(df)), int(.9 * len(df))])\n\n print(f'Train samples: {len(train_df):,}')\n print(f'Validate samples: {len(val_df):,}')\n print(f'Test samples: {len(test_df):,}')\n\n train_df.to_csv(os.path.join(output_dir, 'train.tsv'), sep='\\t', index=False)\n val_df.to_csv(os.path.join(output_dir, 'val.tsv'), sep='\\t', index=False)\n test_df.to_csv(os.path.join(output_dir, 'test.tsv'), sep='\\t', index=False)\n\n\ndef tokenize_english_text(input_path: str, output_path: str, col_index: int = 0):\n \"\"\"Tokenize and lowercase text\"\"\"\n # Download en tokenizer with `python -m spacy download en`\n en_tokenizer = get_tokenizer('spacy', language='en_core_web_sm')\n tokenized_rows = []\n with open(input_path) as f:\n reader = csv.reader(f, delimiter='\\t')\n header_text = next(reader)\n for row in reader:\n tokenized_rows.append([' '.join(en_tokenizer(row[col_index].lower()))])\n with open(output_path, 'w') as f:\n writer = csv.writer(f, delimiter='\\t')\n writer.writerow(header_text)\n writer.writerows(tokenized_rows)\n\n","repo_name":"BushMinusZero/deep-learning-skunk-works","sub_path":"src/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":6774,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"503440975","text":"import nltk\n\nnltk.download('stopwords')\ntext=\"\"\"welcome you to programming knowledge. Lets start with our first tutorial on NLTK. We shall learn the basics of NLTK here.\"\"\"\ndemowords= [\"playing\", \"happiness\",\"going\",\"doing\",\"yes\",\"no\",\"1\",\"having\",\"had\", \"haved\"]\nfrom nltk.corpus import stopwords\nstop_words= stopwords.words (\"english\")\n# print (stop_words)\n# print(\"\\n\\n\")\n# print (set(stop_words))\n\nfrom nltk.tokenize import word_tokenize, sent_tokenize\ntokenize_words= word_tokenize(text)\n#print(tokenize_words )\ntokenize_words_without_stop_words = []\nfor word in tokenize_words:\n if word not in stop_words:\n tokenize_words_without_stop_words.append(word)\nprint(\"stop words that got removed\", set(tokenize_words)-set(tokenize_words_without_stop_words))","repo_name":"composureR3j3c/NTLK","sub_path":"ref/2_stopwords.py","file_name":"2_stopwords.py","file_ext":"py","file_size_in_byte":766,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"2669340379","text":"#!/usr/bin/python3\nif __name__ == \"__main__\":\n import sys\n from calculator_1 import add, sub, mul, div\n argv = sys.argv\n slen = len(argv)\n if (slen != 4):\n print(\"Usage: ./100-my_calculator.py \")\n sys.exit(1)\n ops = [\"+\", \"-\", \"*\", \"/\"]\n op = argv[2]\n if op not in ops:\n print(\"Unknown operator. Available operators: +, -, * and /\")\n sys.exit(1)\n a = int(argv[1])\n b = int(argv[3])\n if (op == ops[0]):\n print(\"{} + {} = {}\".format(a, b, add(a, b)))\n elif (op == ops[1]):\n print(\"{} - {} = {}\".format(a, b, sub(a, b)))\n elif (op == ops[2]):\n print(\"{} * {} = {}\".format(a, b, mul(a, b)))\n elif (op == ops[3]):\n print(\"{} / {} = {}\".format(a, b, div(a, b)))\n","repo_name":"marybngozi/alx-higher_level_programming","sub_path":"0x02-python-import_modules/100-my_calculator.py","file_name":"100-my_calculator.py","file_ext":"py","file_size_in_byte":772,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4654353071","text":"class Solution:\n def distinctNames(self, ideas: List[str]) -> int:\n n = ord(\"z\") - ord(\"a\") + 1\n suffix = [set() for _ in range(n)]\n for word in ideas:\n letter = word[0]\n suffix[ord(letter) - ord(\"a\")].add(word[1:])\n counter = 0\n for i in range(n):\n for j in range(i + 1, n):\n mutual = len(suffix[i] & suffix[j])\n counter += 2 * (len(suffix[i]) - mutual) * (len(suffix[j]) - mutual)\n\n return counter\n","repo_name":"Xrenya/Algorithms","sub_path":"_2306.py","file_name":"_2306.py","file_ext":"py","file_size_in_byte":508,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41694170964","text":"\"\"\"\r\nGiven a binary tree, return the zigzag level order traversal of its nodes' values. (ie, from left to right, then right\r\nto left for the next level and alternate between).\r\n\r\nFor example:\r\nGiven binary tree {3,9,20,#,#,15,7},\r\n 3\r\n / \\\r\n 9 20\r\n / \\\r\n 15 7\r\nreturn its zigzag level order traversal as:\r\n[\r\n [3],\r\n [20,9],\r\n [15,7]\r\n]\r\nconfused what \"{1,#,2,3}\" means? > read more on how binary tree is serialized on OJ.\r\n\"\"\"\r\n__author__ = 'Danyang'\r\n# Definition for a binary tree node\r\nclass TreeNode:\r\n def __init__(self, x):\r\n self.val = x\r\n self.left = None\r\n self.right = None\r\n\r\nclass Solution:\r\n def zigzagLevelOrder(self, root):\r\n \"\"\"\r\n BFS, stack & queue\r\n :param root: a tree node\r\n :return: a list of lists of integers\r\n \"\"\"\r\n if not root:\r\n return []\r\n\r\n result = []\r\n lst = [root]\r\n direction = False\r\n while lst:\r\n if direction:\r\n result.append([element.val for element in lst])\r\n else:\r\n result.append([element.val for element in reversed(lst)])\r\n\r\n for i in range(len(lst)): # evaluation time\r\n element = lst.pop(0) # queue \r\n if element.left:\r\n lst.append(element.left)\r\n if element.right:\r\n lst.append(element.right)\r\n direction = not direction\r\n return result\r\n\r\n\r\n","repo_name":"algorhythms/LeetCode","sub_path":"103 Binary Tree Zigzag Level Order Traversal.py","file_name":"103 Binary Tree Zigzag Level Order Traversal.py","file_ext":"py","file_size_in_byte":1481,"program_lang":"python","lang":"en","doc_type":"code","stars":843,"dataset":"github-code","pt":"81"}
+{"seq_id":"33240533927","text":"import pandas as pd\nimport shutil\n\n# 設定\nPREF_CODE = \"160008\"\nPREF_NAME = \"富山県\"\nCITY_NAME = \"\"\nURL = \"https://toyama-pref.box.com/shared/static/8wb3518lh47jx8iv81f80jsmqnnlw46m.xlsx\"\n\n\"\"\"\n患者等発生状況\n\"\"\"\ndf_patients = pd.read_excel(\n URL,\n sheet_name=\"患者等発生状況\",\n index_col=\"No\",\n dtype={\"発症日\": \"object\", \"年代\": \"object\", \"備考\": \"object\"},\n)\ndf_patients[\"発症日\"]=df_patients[\"発症日\"].apply(lambda x : str(x).replace(\"nan\",\"\").replace(\" 00:00:00\",\"\"))\ndf_patients.to_csv(\"./data/toyama_patients.csv\", encoding=\"utf_8_sig\")\n\ndf_patients.rename(\n columns={\n \"検査結果判明日\": \"公表_年月日\",\n \"発症日\": \"発症_年月日\",\n \"居住地\": \"患者_居住地\",\n \"年代\": \"患者_年代\",\n \"性別\": \"患者_性別\",\n \"職業\": \"患者_職業\",\n \"症状\": \"患者_状態\",\n \"渡航歴の有無\": \"患者_渡航歴の有無フラグ\",\n \"状態\": \"患者_退院済フラグ\",\n },\n inplace=True,\n)\n\ndf_patients[\"全国地方公共団体コード\"] = PREF_CODE\ndf_patients[\"都道府県名\"] = PREF_NAME\ndf_patients[\"市区町村名\"] = CITY_NAME\n\ndf_patients[\"患者_退院済フラグ\"] = (\n df_patients[\"患者_退院済フラグ\"].replace({\"入院中\": 0, \"入院調整中\": 0, \"入院\": 0, \"退院\": 1, \"死亡\": 1, \"調査中\": None}).astype(\"Int64\")\n)\n\ndf_patients[\"患者_渡航歴の有無フラグ\"] = (\n df_patients[\"患者_渡航歴の有無フラグ\"].replace({\"x\": 0, \"o\": 1}).astype(\"Int64\")\n)\n\ndf_patients[\"患者_症状\"] = \"\"\n\ndf_patients[\"患者_年代\"] = df_patients[\"患者_年代\"].replace({\"90代以上\": \"90歳以上\"})\n\npatients = df_patients.loc[\n :,\n [\n \"全国地方公共団体コード\",\n \"都道府県名\",\n \"市区町村名\",\n \"公表_年月日\",\n \"発症_年月日\",\n \"患者_居住地\",\n \"患者_年代\",\n \"患者_性別\",\n \"患者_職業\",\n \"患者_状態\",\n \"患者_症状\",\n \"患者_渡航歴の有無フラグ\",\n \"患者_退院済フラグ\",\n \"備考\",\n ],\n]\n\npatients.to_csv(\n \"./data/160008_toyama_covid19_patients.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\n\"\"\"\n日別集計\n\"\"\"\ndf_counts = pd.read_excel(\n URL,\n sheet_name=\"日別集計\",\n engine=\"openpyxl\",\n index_col=\"年月日\",\n parse_dates=True,\n dtype={\n \"PCR検査数\": \"Int64\",\n \"抗原検査数\": \"Int64\",\n \"陰性人数\": \"Int64\",\n \"陽性人数\": \"Int64\",\n \"一般相談件数\": \"Int64\",\n \"受診・相談センター相談件数\": \"Int64\",\n \"退院者数\": \"Int64\",\n \"死亡者数\": \"Int64\",\n \"備考\": \"object\",\n }\n )\ndf_counts.to_csv(\"./data/toyama_counts.csv\", encoding=\"utf_8_sig\")\n\n# 検査実施人数\ndf_counts[\"実施_年月日\"] = df_counts.index.strftime(\"%Y-%m-%d\")\n\n# 陰性確認数\ndf_counts[\"完了_年月日\"] = df_counts.index.strftime(\"%Y-%m-%d\")\n\n# コールセンター相談件数\ndf_counts[\"受付_年月日\"] = df_counts.index.strftime(\"%Y-%m-%d\")\n\ndf_counts[\"全国地方公共団体コード\"] = PREF_CODE\ndf_counts[\"都道府県名\"] = PREF_NAME\ndf_counts[\"市区町村名\"] = CITY_NAME\n\n# 検査実施人数\ntest_people = df_counts.loc[\n :, [\"実施_年月日\", \"全国地方公共団体コ���ド\", \"都道府県名\", \"市区町村名\", \"PCR検査数\", \"備考\"]\n].copy()\n\ntest_people.rename(columns={\"PCR検査数\": \"検査実施_人数\"}, inplace=True)\n\ntest_people.to_csv(\n \"./data/160008_toyama_covid19_test_people.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\nantigen_test_people = df_counts.loc[\n :, [\"実施_年月日\", \"全国地方公共団体コード\", \"都道府県名\", \"市区町村名\", \"抗原検査数\", \"備考\"]\n].copy()\n\nantigen_test_people.rename(columns={\"抗原検査数\": \"検査実施_人数\"}, inplace=True)\n\nantigen_test_people.to_csv(\n \"./data/160008_toyama_covid19_antigen_test_people.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\n# 陰性確認数\ndf_counts.rename(columns={\"退院者数\": \"陰性確認_件数\"}, inplace=True)\n\nconfirm_negative = df_counts.loc[\n :, [\"完了_年月日\", \"全国地方公共団体コード\", \"都道府県名\", \"市区町村名\", \"陰性確認_件数\", \"備考\"]\n].copy()\n\nconfirm_negative.to_csv(\n \"./data/160008_toyama_covid19_confirm_negative.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\n# 入退院確認数\ndf_counts.rename(columns={\"陽性人数\": \"陽性確認_件数\", \"死亡者数\": \"死亡確認_件数\"}, inplace=True)\n\nconfirm_patients = df_counts.loc[\n :,\n [\"完了_年月日\", \"全国地方公共団体コード\", \"都道府県名\", \"市区町村名\", \"陽性確認_件数\", \"陰性確認_件数\", \"死亡確認_件数\", \"備考\"],\n].copy()\n\nconfirm_patients.to_csv(\n \"./data/160008_toyama_covid19_confirm_patients.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\n# 一般相談件数\ncall_center = df_counts.loc[\n :, [\"受付_年月日\", \"全国地方公共団体コード\", \"都道府県名\", \"市区町村名\", \"一般相談件数\"]\n].copy()\n\ncall_center.rename(columns={\"一般相談件数\": \"相談件数\"}, inplace=True)\ncall_center.to_csv(\n \"./data/160008_toyama_covid19_call_center.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\n# 帰国者・接触者相談センター相談件数\nhot_line = df_counts.loc[\n :, [\"受付_年月日\", \"全国地方公共団体コード\", \"都道府県名\", \"市区町村名\", \"受診・相談センター相談件数\"]\n].copy()\n\nhot_line.rename(columns={\"受診・相談センター相談件数\": \"相談件数\"}, inplace=True)\nhot_line.to_csv(\n \"./data/160008_toyama_covid19_hot_line.csv\",\n index=False,\n encoding=\"utf_8_sig\",\n)\n\nshutil.make_archive(\"./opendata\", \"zip\", root_dir=\"./data\")\n","repo_name":"terachan0117/covid19-toyama-opendata","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5865,"program_lang":"python","lang":"ja","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"7630687434","text":"\"\"\"\ndemo prompt\n\"\"\"\n\nimport argparse\nfrom tui_editor import TuiEditor\n\n\nif __name__ == '__main__':\n arg_parser = argparse.ArgumentParser()\n arg_parser.add_argument(\"--height\", type=int, default=10)\n arg_parser.add_argument(\"--show-line-numbers\", action=\"store_true\")\n args = arg_parser.parse_args()\n\n print('Hello World! Ctrl+S to save.')\n e = TuiEditor()\n if args.show_line_numbers:\n e.show_line_numbers = True\n keys = []\n e.on_key = lambda key: (keys.append(key), e.set_status_lines([\"key: %r\" % key]), None)[-1]\n e.edit()\n print(\"Result:\", repr(e.get_text()))\n print(\"Keys:\", repr(keys))\n print('Good bye!')\n","repo_name":"albertz/py-tui-editor","sub_path":"demo-prompt.py","file_name":"demo-prompt.py","file_ext":"py","file_size_in_byte":657,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"6427993989","text":"import string\n\nimport numpy as np\nimport pytest\nfrom torch import nn\nfrom torchvision.transforms import ToTensor, Normalize\nfrom sklearn.model_selection import train_test_split\n\nfrom pytorch_widedeep.models import Wide, TabMlp, Vision, BasicRNN, WideDeep\nfrom pytorch_widedeep.training import Trainer\n\nnp.random.seed(1)\n\n# Wide array\nX_wide = np.random.choice(50, (32, 10))\n\n# Deep Array\ncolnames = list(string.ascii_lowercase)[:10]\nembed_cols = [np.random.choice(np.arange(5), 32) for _ in range(5)]\nembed_input = [(u, i, j) for u, i, j in zip(colnames[:5], [5] * 5, [16] * 5)]\ncont_cols = [np.random.rand(32) for _ in range(5)]\nX_tab = np.vstack(embed_cols + cont_cols).transpose()\n\n# Text Array\npadded_sequences = np.random.choice(np.arange(1, 100), (32, 48))\nX_text = np.hstack((np.repeat(np.array([[0, 0]]), 32, axis=0), padded_sequences))\nvocab_size = 110\n\n# Image Array\nX_img = np.random.choice(256, (32, 224, 224, 3))\nX_img_norm = X_img / 255.0\n\n# Target\ntarget = np.random.choice(2, 32)\n\n# train/validation split\n(\n X_wide_tr,\n X_wide_val,\n X_tab_tr,\n X_tab_val,\n X_text_tr,\n X_text_val,\n X_img_tr,\n X_img_val,\n y_train,\n y_val,\n) = train_test_split(X_wide, X_tab, X_text, X_img, target)\n\n# build model components\nwide = Wide(np.unique(X_wide).shape[0], 1)\ndeeptabular = TabMlp(\n column_idx={k: v for v, k in enumerate(colnames)},\n cat_embed_input=embed_input,\n continuous_cols=colnames[-5:],\n mlp_hidden_dims=[32, 16],\n mlp_dropout=[0.5, 0.5],\n)\ndeeptext = BasicRNN(vocab_size=vocab_size, embed_dim=32, padding_idx=0)\ndeepimage = Vision(pretrained_model_setup=\"resnet18\", n_trainable=0)\n\n# transforms\nmean = [0.406, 0.456, 0.485] # BGR\nstd = [0.225, 0.224, 0.229] # BGR\ntransforms1 = [ToTensor, Normalize(mean=mean, std=std)]\ntransforms2 = [Normalize(mean=mean, std=std)]\n\ndeephead_ds = nn.Sequential(nn.Linear(16, 8), nn.Linear(8, 4))\ndeephead_ds.output_dim = 4\ndeephead_dt = nn.Sequential(nn.Linear(64, 8), nn.Linear(8, 4))\ndeephead_dt.output_dim = 4\ndeephead_di = nn.Sequential(nn.Linear(512, 8), nn.Linear(8, 4))\ndeephead_di.output_dim = 4\n\n# #############################################################################\n# Test that runs many possible scenarios of data inputs I can think off.\n# Surely users will input something unexpected\n# #############################################################################\n\n\n@pytest.mark.parametrize(\n \"X_wide, X_tab, X_text, X_img, X_train, X_val, target, val_split, transforms\",\n [\n (X_wide, X_tab, X_text, X_img, None, None, target, None, transforms1),\n (X_wide, X_tab, X_text, X_img, None, None, target, None, transforms2),\n (X_wide, X_tab, X_text, X_img, None, None, target, None, None),\n (\n X_wide,\n X_tab,\n X_text,\n X_img_norm,\n None,\n None,\n target,\n None,\n transforms2,\n ),\n (\n X_wide,\n X_tab,\n X_text,\n X_img_norm,\n None,\n None,\n target,\n None,\n transforms1,\n ),\n (X_wide, X_tab, X_text, X_img_norm, None, None, target, None, None),\n (X_wide, X_tab, X_text, X_img, None, None, target, 0.2, None),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide,\n \"X_tab\": X_tab,\n \"X_text\": X_text,\n \"X_img\": X_img,\n \"target\": target,\n },\n None,\n None,\n None,\n None,\n ),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide,\n \"X_tab\": X_tab,\n \"X_text\": X_text,\n \"X_img\": X_img,\n \"target\": target,\n },\n None,\n None,\n None,\n transforms1,\n ),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide,\n \"X_tab\": X_tab,\n \"X_text\": X_text,\n \"X_img\": X_img,\n \"target\": target,\n },\n None,\n None,\n 0.2,\n None,\n ),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide,\n \"X_tab\": X_tab,\n \"X_text\": X_text,\n \"X_img\": X_img,\n \"target\": target,\n },\n None,\n None,\n 0.2,\n transforms2,\n ),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide_tr,\n \"X_tab\": X_tab_tr,\n \"X_text\": X_text_tr,\n \"X_img\": X_img_tr,\n \"target\": y_train,\n },\n {\n \"X_wide\": X_wide_val,\n \"X_tab\": X_tab_val,\n \"X_text\": X_text_val,\n \"X_img\": X_img_val,\n \"target\": y_val,\n },\n None,\n None,\n None,\n ),\n (\n None,\n None,\n None,\n None,\n {\n \"X_wide\": X_wide_tr,\n \"X_tab\": X_tab_tr,\n \"X_text\": X_text_tr,\n \"X_img\": X_img_tr,\n \"target\": y_train,\n },\n {\n \"X_wide\": X_wide_val,\n \"X_tab\": X_tab_val,\n \"X_text\": X_text_val,\n \"X_img\": X_img_val,\n \"target\": y_val,\n },\n None,\n None,\n transforms1,\n ),\n ],\n)\ndef test_widedeep_inputs(\n X_wide,\n X_tab,\n X_text,\n X_img,\n X_train,\n X_val,\n target,\n val_split,\n transforms,\n):\n model = WideDeep(\n wide=wide, deeptabular=deeptabular, deeptext=deeptext, deepimage=deepimage\n )\n trainer = Trainer(model, objective=\"binary\", transforms=transforms, verbose=0)\n trainer.fit(\n X_wide=X_wide,\n X_tab=X_tab,\n X_text=X_text,\n X_img=X_img,\n X_train=X_train,\n X_val=X_val,\n target=target,\n val_split=val_split,\n batch_size=16,\n )\n assert trainer.history[\"train_loss\"] is not None\n\n\n@pytest.mark.parametrize(\n \"X_wide, X_tab, X_text, X_img, X_train, X_val, target\",\n [\n (\n X_wide,\n X_tab,\n X_text,\n X_img,\n None,\n {\n \"X_wide\": X_wide_val,\n \"X_tab\": X_tab_val,\n \"X_text\": X_text_val,\n \"X_img\": X_img_val,\n \"target\": y_val,\n },\n target,\n ),\n ],\n)\ndef test_xtrain_xval_assertion(\n X_wide,\n X_tab,\n X_text,\n X_img,\n X_train,\n X_val,\n target,\n):\n model = WideDeep(\n wide=wide, deeptabular=deeptabular, deeptext=deeptext, deepimage=deepimage\n )\n trainer = Trainer(model, objective=\"binary\", verbose=0)\n with pytest.raises(AssertionError):\n trainer.fit(\n X_wide=X_wide,\n X_tab=X_tab,\n X_text=X_text,\n X_img=X_img,\n X_train=X_train,\n X_val=X_val,\n target=target,\n batch_size=16,\n )\n\n\n@pytest.mark.parametrize(\n \"wide, deeptabular, deeptext, deepimage, X_wide, X_tab, X_text, X_img, target\",\n [\n (wide, None, None, None, X_wide, None, None, None, target),\n (None, deeptabular, None, None, None, X_tab, None, None, target),\n (None, None, deeptext, None, None, None, X_text, None, target),\n (None, None, None, deepimage, None, None, None, X_img, target),\n ],\n)\ndef test_individual_inputs(\n wide, deeptabular, deeptext, deepimage, X_wide, X_tab, X_text, X_img, target\n):\n model = WideDeep(\n wide=wide, deeptabular=deeptabular, deeptext=deeptext, deepimage=deepimage\n )\n trainer = Trainer(model, objective=\"binary\", verbose=0)\n trainer.fit(\n X_wide=X_wide,\n X_tab=X_tab,\n X_text=X_text,\n X_img=X_img,\n target=target,\n batch_size=16,\n )\n # check it has run succesfully\n assert len(trainer.history) == 1\n\n\n###############################################################################\n# test deephead is not None and individual components\n###############################################################################\n\n\n@pytest.mark.parametrize(\n \"deeptabular, deeptext, deepimage, X_tab, X_text, X_img, deephead, target\",\n [\n (deeptabular, None, None, X_tab, None, None, deephead_ds, target),\n (None, deeptext, None, None, X_text, None, deephead_dt, target),\n (None, None, deepimage, None, None, X_img, deephead_di, target),\n ],\n)\ndef test_deephead_individual_components(\n deeptabular, deeptext, deepimage, X_tab, X_text, X_img, deephead, target\n):\n model = WideDeep(\n deeptabular=deeptabular,\n deeptext=deeptext,\n deepimage=deepimage,\n deephead=deephead,\n ) # noqa: F841\n trainer = Trainer(model, objective=\"binary\", verbose=0)\n trainer.fit(\n X_wide=X_wide,\n X_tab=X_tab,\n X_text=X_text,\n X_img=X_img,\n target=target,\n batch_size=16,\n )\n # check it has run succesfully\n assert len(trainer.history) == 1\n\n\n###############################################################################\n# test deephead is None and head_layers is not None and individual components\n###############################################################################\n\n\n@pytest.mark.parametrize(\n \"deeptabular, deeptext, deepimage, X_tab, X_text, X_img, target\",\n [\n (deeptabular, None, None, X_tab, None, None, target),\n (None, deeptext, None, None, X_text, None, target),\n (None, None, deepimage, None, None, X_img, target),\n ],\n)\ndef test_head_layers_individual_components(\n deeptabular, deeptext, deepimage, X_tab, X_text, X_img, target\n):\n model = WideDeep(\n deeptabular=deeptabular,\n deeptext=deeptext,\n deepimage=deepimage,\n head_hidden_dims=[8, 4],\n ) # noqa: F841\n trainer = Trainer(model, objective=\"binary\", verbose=0)\n trainer.fit(\n X_wide=X_wide,\n X_tab=X_tab,\n X_text=X_text,\n X_img=X_img,\n target=target,\n batch_size=16,\n )\n # check it has run succesfully\n assert len(trainer.history) == 1\n","repo_name":"jrzaurin/pytorch-widedeep","sub_path":"tests/test_model_functioning/test_data_inputs.py","file_name":"test_data_inputs.py","file_ext":"py","file_size_in_byte":10678,"program_lang":"python","lang":"en","doc_type":"code","stars":1164,"dataset":"github-code","pt":"81"}
+{"seq_id":"33363243656","text":"import torchvision.transforms as tsf\n\nclass Config:\n def __init__(self):\n self.model_name = 'YOLOv4'\n\n # ------\n # Input size\n # (w, h)\n # YOLO accepts any input sizes as long as the size is multiples of 32\n # ------\n self.input_size = (672, 512)\n \n # ------\n # Classe names\n # ------\n self.class_names = [\n '0', '1', '2', '3', '4', '5', '6', '7', '8', '9'\n ]\n \n # ------\n # Anchors\n # [(w, h), ...]\n # 3 (YOLO heads) x n (num of anchors for each head)\n # big <------size------> small\n # ------\n self.anchors = [ \n # yolo default anchors for (608, 608)\n # (459, 401), (192, 243), (142, 110), (72, 146), (76, 55), (36, 75), (40, 28), (91, 36), (12, 16)\n # BBTv1 for (672, 512)\n (38, 60), (50, 43), (25, 70), (29, 54), (37, 42), (47, 31), (20, 54), (24, 37), (32, 26)\n ]\n\n\n # ------\n # Grayscale for every input image\n # ------\n self.grayscale = True\n\n # ------\n # Resize mode\n # \"stretch\" or \"pad\"\n # ------\n self.resize_mode = 'stretch'\n\n # ------\n # Normalization\n # [means, stds]\n # ------\n self.normalization = [\n (0.5, 0.5, 0.5), (0.22, 0.22, 0.22)\n ]\n \n\n # ------\n # YOLO structure\n # ------\n self.num_heads = 3\n\n\nclass DetectionConfig(Config):\n def __init__(self):\n # get general configs\n super(DetectionConfig, self).__init__()\n # threshold\n self.confidence = 0.5\n self.iou = 0.3\n self.text_font = 'simhei.ttf'\n \n # ------\n # Cuda for detection\n # ------\n self.cuda = True\n \n \n\nclass TrainingConfig(Config):\n def __init__(self):\n # get general configs\n super(TrainingConfig, self).__init__()\n self.log_dir = 'logs/'\n # ------\n # Basic hyperparameters\n # total epochs = freeze + unfreeze\n # ------\n self.initial_epoch = 1\n self.freeze_epochs = 0\n self.unfreeze_epochs = 300\n self.batch_size = 4\n self.learning_rate = 1e-3\n\n # ------\n # Training dictionary\n # ------\n self.train_dict = 'logs/yolo_train_dict.txt'\n\n # ------\n # Pre-train weight\n # set \"None\" to get Kaiming initialization\n # ------\n self.pretrained = None\n\n # ------\n # Ratio of validation\n # ------\n self.valid_ratio = 0.1\n\n # ------\n # Center crop\n # aspect ratio (w, h)\n # set \"None\" to turn off\n # ------\n self.center_crop_ratio = (4, 3)\n\n # ------\n # Color Jitter\n # set \"None\" to turn off\n # ------\n self.color_jitter = tsf.ColorJitter(\n brightness = 0.2, \n contrast = 0.2, \n saturation = 0, \n hue = 0\n )\n\n # ------\n # Random Flip\n # mode in \"all\", \"horizontal\", \"vertical\"\n # set \"None\" to turn off\n # ------\n self.random_flip = 'all'\n\n # ------\n # Mosaic augmentation\n # image sizes from dataset better similar to each other\n # ------\n self.mosaic = True\n\n # ------\n # Label smoothing\n # ------\n self.label_smoothing = 0.\n\n # ------\n # Size augmentation\n # a float in [0, 1] for probability of chagning input size\n # the input size will +-(32, 32) randomly for training\n # set \"None\" to turn off\n # ------ \n self.size_aug = None\n\n # ------\n # Threads of dataloader\n # ------\n self.num_workers = 4\n\n # ------\n # Consine learning rate\n # ------\n self.cos_lr = True\n \n # ------\n # Cuda for training\n # ------\n self.cuda = True\n","repo_name":"panghanwu/yolov4_pytorch","sub_path":"configs.py","file_name":"configs.py","file_ext":"py","file_size_in_byte":4016,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"33398926427","text":"import numpy as np\nimport pathlib as pl\nfrom scipy.ndimage import gaussian_filter as gfilt\n\ndef makeAlphaMask(shape=(720, 1280), margin=30, sigma=10, low=-1, high=1):\n \"\"\"\n \"\"\"\n\n mask = np.full(shape, high).astype(float)\n # mask = np.ones(shape)\n mask[:, :margin] = low\n mask[:, -margin:] = low\n mask[:margin, :] = low\n mask[-margin:, ] = low\n filtered = gfilt(mask, sigma)\n clipped = np.clip(filtered, -1, 1)\n\n return clipped\n\ndef cycleSignalPatch(display, cycle=(1, 1), nCycles=1):\n \"\"\"\n \"\"\"\n\n for iCycle in range(nCycles):\n display.state = True\n for iFrame in range(int(round(display.fps * cycle[0]))):\n display.drawBackground()\n display.flip()\n\n display.state = False\n for iFrame in range(int(round(display.fps * cycle[1]))):\n display.drawBackground()\n display.flip()\n\n return\n\ndef generateMetadataFilename(parent, tag, extension='.pkl'):\n \"\"\"\n \"\"\"\n\n existing = list(parent.glob(f'{tag}*'))\n n = len(existing) + 1\n filename = parent.joinpath(f'{tag}-{n}{extension}')\n return filename\n\ndef estimateFrameCount(t, fps=60, roundingMethod='nearest'):\n \"\"\"\n Convert time (in seconds) to frames\n \"\"\"\n\n if roundingMethod == 'nearest':\n nFrames = int(round(fps * t))\n elif roundingMethod == 'ceiling':\n nFrames = int(np.ceil(fps * t))\n elif roundingMethod == 'floor':\n nFrames = int(np.floor(fps * t))\n else:\n raise Exception(f'Rounding method must be one of \"nearest\", \"ceiling\", or \"floor\"')\n\n if nFrames == 0:\n raise Exception(f'Estimated frame count is 0')\n\n return nFrames","repo_name":"jbhunt/openpmad2","sub_path":"openpmad2/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":1672,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"24663923454","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport Store.models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('Store', '0004_add_orders'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='order',\n name='orderDateYMD',\n field=models.CharField(default=Store.models.get_default_ymd, max_length=8),\n ),\n ]\n","repo_name":"luispdm/GameStore","sub_path":"GameStore/Store/migrations/0005_auto_20160217_2344.py","file_name":"0005_auto_20160217_2344.py","file_ext":"py","file_size_in_byte":454,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"8225995274","text":"# -*- coding=utf-8 -*-\nimport io\nimport re\nimport sys\nfrom collections import Counter\nimport jieba\nimport pandas as pd\nfrom matplotlib import pyplot as plt\nimport wordcloud\nimport warnings\n# from snownlp import seg, SnowNLP\n\n\n# from scipy.misc import imread\n# from importlib import reload\n\n\ndef mycut(str1):\n\tstr1 = ' '.join(jieba.cut(str1)) # 曾经尝试关闭新词发现\n\treturn str1\n# mycut = lambda s:' '.join(jieba.cut(s))\n\n# 4SnowNLP模块做正负情感分析√/或者使用ROSTCM6代替生成(3.7应用Error)\nfilepath1 = r'./foundation_output/pos_data.csv'\nfilepath2 = r'./foundation_output/neg_data.csv'\ndf1 = pd.read_csv(filepath1, encoding='utf-8', header=0)\ndf2 = pd.read_csv(filepath2, encoding='utf-8', header=0)\ndf1 = df1['comment']\ndf2 = df2['comment']\nprint('pos:\\n', df1.head(), 'neg\\n', df2.head())\n\n'''如果是ROSTCM6就用以下内容去除前缀\n# read_csv(r’F:\\Demo\\ThesisProject4\\cleanoil_output\\temp_after_filter3_utf8_正面情感结果.txt')\ndf1 = pd.read_csv(filepath1, encoding='utf-8', header=0)\ndf2 = pd.read_csv(filepath2, encoding='utf-8', header=0)\nprint(df1.head())\nprint(df2.head())\ndf11 = df1['comments'].map(lambda x: re.sub(r'\\d*\\s\\s', '', x))\ndf22 = df2['comments'].map(lambda x: re.sub(r'[-]\\d*\\s\\s', '', x))\nprint('去除评分前缀后正面:\\n')\nprint(df11.head())\nprint('去除评分前缀后负面:\\n')\nprint(df22.head())\n# 输出去除前缀后的内容 用于ROSTCM6生成正负语义网络\ndf11.to_csv('正面情感评论_无前缀.txt', index=False, header=False, encoding='utf-8')\ndf22.to_csv('负面情感评论_无前缀.txt', index=False, header=False, encoding='utf-8')\n'''\n\n# 5jieba模块做中文分词处理,采用apply()广播形式加快分词速度\n# df1 = df1.iloc[:, 0].apply(mycut)\ndf11 = pd.DataFrame(df1.apply(mycut))\n# df2 = df2.iloc[:, 0].apply(mycut)\ndf22 = pd.DataFrame(df2.apply(mycut))\nprint('分词处理后正面:\\n')\nprint(df11.head())\nprint('分词处理后负面:\\n')\nprint(df22.head())\n# df1 = mycut(str(df1))这种方法也可以\n# df11 = pd.DataFrame(df1)\n\n# 6去除停用词\nstopfilepath = './stoplist_utf8.txt'\nstop = pd.read_csv(stopfilepath, encoding='utf-8', header=None, sep='dingbangchu',\n engine='python') # 这里sep分割符 非停用词符均可,这里选的是竞赛站名\n# sep设置分割词,csv默认以半角逗号分割,该词恰好在停用词表中导致读取出错,因此手动设置一个不存在的分割词,如dingbangchu\nstop = [' ', ''] + list(stop[0]) # Pandas自动过滤了空格符,这里手动添加,此时stop由df类型转为list类型\npos = pd.DataFrame(df11.copy())\nneg = pd.DataFrame(df22.copy())\npos = pos['comment'].apply(lambda s: s.split(' ')) # 定义一个分割函数,然后用apply广播,每行按空格分隔,每行由str转变为list\npos = pos.apply(lambda x: [i for i in x if i not in stop]) # 逐词判断是否停用词,思路同上\nneg = neg['comment'].apply(lambda s: s.split(' '))\nneg = neg.apply(lambda x: [i for i in x if i not in stop])\nprint('去除停用词后正面:\\n')\nprint(pos.head())\nprint('去除停用词后负面:\\n')\nprint(neg.head())\n# 输出去除停用词后已分词的内容 用于生成词云以及后续LDA主题分析\npos.to_csv('foundation_output/正面情感分词有效版.txt', index=False, header=False, encoding='utf-8')\nneg.to_csv('foundation_output/负面情感分词有效版.txt', index=False, header=False, encoding='utf-8')\n\n# pos = pd.DataFrame(pos.copy())\n# neg = pd.DataFrame(neg.copy())\n# pos = pos[pos.iloc[:, 0].apply(len) >= 1]\n# print(pos[:5])\n# neg = neg[neg.iloc[:, 0].apply(len) >= 1]\n# print(neg[:5])\n\n# 7词频统计\n# =============正面=============\nall_words = []\nfor n in range(0, len(pos)):\n\tfor i in pos[n]:\n\t\tall_words.append(i)\nword_count = pd.Series(all_words)\ntop_10 = word_count.value_counts(sort=True, ascending=False, dropna=True)\nprint('正面词频统计TOP10关键词:')\nprint(top_10[:10])\ncounts_result = dict(Counter(all_words))\ncounts_result = dict(sorted(counts_result.items(), key=lambda d: d[1], reverse=True))\n# counts_result.sort(reverse=True)\n# get to k most frequently occuring words\n# counts_result = Counter(all_words).most_common(10)\n# print('正面词频统计字典如下:')\n# print(counts_result)\nwith open('foundation_output/正面词频统计.txt', 'w', errors='ignore') as f:\n\t[f.write(str('{0},{1}\\n'.format(key, value))) for key, value in counts_result.items()]\n# =============负面================\nall_words = []\nfor n in range(0, len(neg)):\n\tfor i in neg[n]:\n\t\tall_words.append(i)\nword_count = pd.Series(all_words)\ntop_10 = word_count.value_counts(sort=True, ascending=False, dropna=True)\nprint('负面词频统计TOP10关键词:')\nprint(top_10[:10])\ncounts_result = dict(Counter(all_words))\ncounts_result = dict(sorted(counts_result.items(), key=lambda d: d[1], reverse=True))\n# counts_result.sort(reverse=True)\n# print('负面词频统计字典如下:')\n# print(counts_result)\nwith open('foundation_output/负面词频统计.txt', 'w', errors='ignore') as f:\n\t[f.write(str('{0},{1}\\n'.format(key, value))) for key, value in counts_result.items()]\n# temp_df = pd.DataFrame(counts_result)\n# temp_df.to_csv('词频统计.txt', index=False, header=False, encoding='utf-8')\n# print(type(pos)) \n# print(pos[0])\n# print(type(pos[0])) list\n# print(all_words)\n# print(type(all_words))\n\n\n# 8词云图绘制\nsys.stdout = io.TextIOWrapper(sys.stdout.buffer, encoding='utf-8')\nwarnings.filterwarnings(\"ignore\")\n# (1)读取背景图片\n# back_picture = imread(r\"\")\n# usa_mask = np.array(Image.open('flower.png'))\n# (2)设置词云参数(源文件应该是分词后并去除停用词的内容,而不是词频表)\n# =============pos_wordcloud=============\nfont = r'C:\\Windows\\Fonts\\msyh.ttc'\nwc = wordcloud.WordCloud(\n\tbackground_color=\"white\",\n\theight=800,\n\twidth=1000,\n\tfont_path=font,\n\tprefer_horizontal=0.2,\n\tmax_words=2000,\n\trelative_scaling=0.3,\n\tmax_font_size=200).generate(str(pos.tolist()))\nplt.imshow(wc, interpolation=\"nearest\")\nplt.axis(\"off\")\nplt.show()\nwc.to_file(\"ciyun_foundaiton_pos.png\")\nprint('词云图_正面已生成!')\n# 从文件导入数据:\n# f = open('cleanoil_output\\负面情感分词有效版.txt', encoding='utf-8').read()\n# =============neg_wordcloud=============\nfont = r'C:\\Windows\\Fonts\\msyh.ttc'\n# pic = imread('') +mask='pic',\nwc = wordcloud.WordCloud(\n\tbackground_color=\"white\",\n\theight=800,\n\twidth=1000,\n\tfont_path=font,\n prefer_horizontal=0.2,\n\tmax_words=2000,\n\trelative_scaling=0.3,\n\tmax_font_size=200).generate(str(neg.tolist()))\n# wc1 = wc.fit_words(counts_result)\n# .generate_from_text(comments)\n# 从DataFrame/pos/neg生成词云\n# 从词频统计文件生成词云\n# (3)绘制词云图\nplt.imshow(wc, interpolation=\"nearest\")\nplt.axis(\"off\")\nplt.show()\n# (4)保存到本地\nwc.to_file(\"ciyun_foundation_neg.png\")\nprint('词云图_负面已生成!')\n# plt.savefig('图6.jpg', dpi=600, bbox_inches='tight', quality=95)\n# plt.show()\n\n\n'''以下是某网站的词云全过程前半部分代码\n# 利用jieba进行分析操作\ndf[\"评论\"] = df[\"评论\"].apply(jieba.lcut)\ndf.head()\n# 去除停用词操作\nwith open(\"stopword.txt\", \"r\", encoding=\"gbk\") as f:\n\tstop = f.read() # 返回的是一个字符串\nstop = stop.split() # 这里得到的是一个列表.split()会将空格,\\n,\\t进行切分,因此我们可以将这些加到停用词当中\nstop = stop + [\" \", \"\\n\", \"\\t\"]\ndf_after = df[\"评论\"].apply(lambda x: [i for i in x if i not in stop])\ndf_after.head()\n'''\n\n'''以下是所有为了去除前缀所做的尝试\n课本中的“data1 = pd.DataFrame(data1[0].str.replace(r'.*?\\d+?\\\\t ', ''))”方法不行,会显示df无str()方法\ndf1 = df1.apply(lambda s: s.str.split(r'\\s'))\ndf2 = df2.apply(lambda s: s.str.split(r'\\s'))\nprint(df1.values)\nprint(df1.shape[1])\ndf1 = df1.apply(lambda s: s.str.strip([0, 2])) # s.str.strip('@')\nndarray = df1['comments'].astype(str).values\nfor i in ndarray:\n\ti = i.split(' ')\nprint(df1.shape[1])\n用正则表达式非贪婪模式修改数据:(*?)表示0个或多个非\\n字符 (+?)表示1个或多个数字,此操作是为了删除ROSTCM6软件给出的评分正负前缀\ndf3 = pd.DataFrame(df1.astype(str).replace(r'\\d*\\s\\s', 'a'))\nPython 3将字符串文字解释为Unicode字符串,因此 \\d 被视为转义的Unicode字符\ndf4 = pd.DataFrame(df2.astype(str).replace(r'([-]\\d*\\s\\s).*?', 'b'))\n'''\n","repo_name":"DingBangBang/SHU-TMALL-comments-NLP","sub_path":"ThesisProject4/visualization.py","file_name":"visualization.py","file_ext":"py","file_size_in_byte":8442,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4654361141","text":"a, b = list(map(int, input().split(\" \")))\ncondition = str(input())\ndef calc(a, b, condition):\n if condition == \"freeze\":\n if a < b:\n return a\n else:\n return b\n elif condition == \"heat\":\n if a < b:\n return b\n else:\n return a\n elif condition == \"auto\":\n return b\n return a\nprint(calc(a=a, b=b, condition=condition))\n","repo_name":"Xrenya/Algorithms","sub_path":"yandex/Lecture_1/Contest/A/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":407,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"70220455945","text":"#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n\n\"\"\"#########################################################\n############################################################\n### Define constants used by multiple source files. ###\n### ###\n### Author: Daniel Dantas ###\n### Last edited: Sep 2022 ###\n############################################################\n#########################################################\"\"\"\n\nimport os\n\n## Canvas sizes\nSUBJ_W = 720\nSUBJ_H = 540\nOP_W = 640\nOP_H = 480\nPAD_X = 10\nPAD_Y = 10\n\n## Filenames\nEXT_TSV = \".tsv\"\nEXT_LIN = \"_linear.tsv\"\nEXT_NN = \"_nearest.tsv\"\nEXT_INFERRED = \"_inferred_rr.tsv\"\nEXT_FLOW = \"_flow.tsv\"\n\n## Used mostly by Polar.py, Data.py and Plot.py\nTYPE_RR = \"R\"\nTYPE_ECG = \"E\"\nTYPE_AUDIO = \"A\"\n\n## Used mostly by WinOp.py\nFOLDER_DATA = \"data\"\nFOLDER_DEBUG = os.path.join(FOLDER_DATA, \"DEBUG\")\n\nFILENAME_RANDOM = \"random.txt\"\nFILENAME_ROUTINE = \"routine.txt\"\nFILENAME_START_TIME = \"start_time.txt\"\nFILENAME_LOG = \"log.txt\"\n\nFILENAME_ECG = \"subj%d_ecg.tsv\"\nFILENAME_ECG_S1 = \"subj1_ecg.tsv\"\nFILENAME_ECG_S2 = \"subj2_ecg.tsv\"\n\nFILENAME_RR = \"subj%d_rr.tsv\"\nFILENAME_RR_S1 = \"subj1_rr.tsv\"\nFILENAME_RR_S2 = \"subj2_rr.tsv\"\n\nFILENAME_AUDIO = \"audio.wav\"\n\n# \"subj%d_ecg_inferred_rr.tsv\"\nFILENAME_ECG_RR = FILENAME_ECG.replace(EXT_TSV, EXT_INFERRED)\nFILENAME_ECG_RR_S1 = FILENAME_ECG_S1.replace(EXT_TSV, EXT_INFERRED)\nFILENAME_ECG_RR_S2 = FILENAME_ECG_S2.replace(EXT_TSV, EXT_INFERRED)\n\nFILENAME_VIDEO = \"subj%d.mp4\"\nFILENAME_VIDEO_S1 = \"subj1.mp4\"\nFILENAME_VIDEO_S2 = \"subj2.mp4\"\n\n\n## Used mostly by 01_sync/sync.py\nFOLDER_SYNC = \"01_sync\"\nFILENAME_VIDEO_SYNC = \"subj%d_sync.mp4\"\nFILENAME_VIDEO_SYNC_S1 = \"subj1_sync.mp4\"\nFILENAME_VIDEO_SYNC_S2 = \"subj2_sync.mp4\"\n\n## Used mostly by 02_preprocess/preprocess.py\nFOLDER_PREP = \"02_preprocess\"\nFOLDER_OUTPUT = \"output\"\n#FILENAME_DATASET = [\"dataset_jf.tsv\", \"dataset_dd.tsv\"]\nFILENAME_DATASET = \"dataset.tsv\"\nFILENAME_SLIDE = \"slides_time.tsv\"\n#FILENAME_ANNOTATION = [\"annotation_jf.eaf\", \"annotation_dd.eaf\"]\n# FILENAME_ANNOTATION = \"annotation_dd.eaf\"\nDATASET_HEADERS = ['folder', 'time', 'block', 'slide',\n 'hr_subj1_linear', 'hr_subj2_linear', 'hr_subj1_nearest', 'hr_subj2_nearest', \\\n 'hr_subj1_ecg_linear', 'hr_subj2_ecg_linear', 'hr_subj1_ecg_nearest', 'hr_subj2_ecg_nearest', \\\n 'rr_subj1_linear', 'rr_subj2_linear', 'rr_subj1_nearest', 'rr_subj2_nearest', \\\n 'rr_subj1_ecg_linear', 'rr_subj2_ecg_linear', 'rr_subj1_ecg_nearest', 'rr_subj2_ecg_nearest']\n\n# \"subj%d_rr_linear.tsv\"\nFILENAME_RR_LIN = FILENAME_RR.replace(EXT_TSV, EXT_LIN)\nFILENAME_RR_LIN_S1 = FILENAME_RR_S1.replace(EXT_TSV, EXT_LIN)\nFILENAME_RR_LIN_S2 = FILENAME_RR_S2.replace(EXT_TSV, EXT_LIN)\n\n# \"subj%d_rr_nearest.tsv\"\nFILENAME_RR_NN = FILENAME_RR.replace(EXT_TSV, EXT_NN)\nFILENAME_RR_NN_S1 = FILENAME_RR_S1.replace(EXT_TSV, EXT_NN)\nFILENAME_RR_NN_S2 = FILENAME_RR_S2.replace(EXT_TSV, EXT_NN)\n\n# \"subj%d_ecg_inferred_rr_linear.tsv\"\nFILENAME_ECG_RR_LIN = FILENAME_ECG_RR.replace(EXT_TSV, EXT_LIN)\nFILENAME_ECG_RR_LIN_S1 = FILENAME_ECG_RR_S1.replace(EXT_TSV, EXT_LIN)\nFILENAME_ECG_RR_LIN_S2 = FILENAME_ECG_RR_S2.replace(EXT_TSV, EXT_LIN)\n\n# \"subj%d_ecg_inferred_rr_nearest.tsv\"\nFILENAME_ECG_RR_NN = FILENAME_ECG_RR.replace(EXT_TSV, EXT_NN)\nFILENAME_ECG_RR_NN_S1 = FILENAME_ECG_RR_S1.replace(EXT_TSV, EXT_NN)\nFILENAME_ECG_RR_NN_S2 = FILENAME_ECG_RR_S2.replace(EXT_TSV, EXT_NN)\n\n## Used mostly by 03_optical_flow/optical_flow.py\nFOLDER_OPTI = \"03_optical_flow\"\nLEN_PREF_VIDEO = 5\n","repo_name":"ddantas/img_hrv_lan","sub_path":"const.py","file_name":"const.py","file_ext":"py","file_size_in_byte":3800,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"28038200354","text":"\nclass Position:\n \n def __init__(self):\n self.board = []\n self.macroboard = []\n \n def parse_field(self, fstr):\n flist = fstr.replace(';', ',').split(',')\n self.board = [ int(f) for f in flist ]\n \n def parse_macroboard(self, mbstr):\n mblist = mbstr.replace(';', ',').split(',')\n self.macroboard = [ int(f) for f in mblist ]\n \n def is_legal(self, x, y):\n mbx, mby = x/3, y/3\n return self.macroboard[3*mby+mbx] == -1 and self.board[9*y+x] == 0\n\n def legal_moves(self):\n return [ (x, y) for x in range(9) for y in range(9) if self.is_legal(x, y) ]\n\n def make_move(self, x, y, pid):\n mb_new = 3*(y%3)+(x%3) #new macroboard\n mbx, mby = x/3, y/3 # old macroboard\n mb_old = mby*3 + mbx\n\n self.board[9*y+x] = pid\n\n\n winner = self.get_winner(mb_old)\n\n if self.macroboard[mb_new] <= 0: #if new is valid\n #invalidate others\n self.macroboard = [0 if n == -1 else n for n in self.macroboard]\n self.macroboard[mb_new] = -1\n else:\n self.macroboard = [-1 if n == 0 else n for n in self.macroboard]\n \n if winner != 0:\n self.macroboard[mb_old] = winner\n\n def get_winner(self, mb_i):\n start_index = (mb_i/3)*27 + (mb_i % 3)*3\n board = self.board\n # check rows/columns\n for i in range(3):\n row_value = board[i*9+start_index]\n col_value = board[i+start_index]\n for j in range(3):\n if board[i*9+j+start_index] != row_value:\n row_value = -1\n if board[j*9+i+start_index] != col_value:\n col_value = -1\n if row_value > 0:\n return row_value\n if col_value > 0:\n return col_value\n\n # Check diagonals\n d1_val = board[start_index]\n d2_val = board[2+start_index]\n for i in [1, 2]:\n if board[i*10+start_index] != d1_val:\n d1_val = -1\n if board[i*8+2+start_index] != d2_val:\n d2_val = -1\n if d2_val > 0:\n return d2_val\n if d1_val > 0:\n return d1_val\n return 0\n\n def get_board(self):\n return ''.join(self.board, ',')\n\n def get_macroboard(self):\n return ''.join(self.macroboard, ',')\n\n","repo_name":"Macetodaface/tictactoe","sub_path":"position.py","file_name":"position.py","file_ext":"py","file_size_in_byte":2386,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"9602234169","text":"from functools import update_wrapper\nfrom django.utils import six\nfrom django.conf.urls import patterns, url, include\nfrom django.contrib.admin.sites import AdminSite\nfrom django.contrib.contenttypes import views as contenttype_views\n\nfrom tribus.web.admin.views import (tribus_config, active_modules,\n logger_levels)\nfrom tribus.web.registration.forms import LoginForm\n\n\nclass TribusAdmin(AdminSite):\n\n login_template = 'registration/login_form.html'\n login_form = LoginForm\n\n def get_urls(self):\n\n def wrap(view, cacheable=False):\n def wrapper(*args, **kwargs):\n return self.admin_view(view, cacheable)(*args, **kwargs)\n return update_wrapper(wrapper, view)\n\n urlpatterns = patterns(\n '',\n url(regex=r'^$',\n view=wrap(self.index),\n name='index'),\n url(regex=r'^config/$',\n view=wrap(tribus_config)),\n url(regex=r'^config/active-modules/$',\n view=wrap(active_modules)),\n url(regex=r'^config/logger-levels/$',\n view=wrap(logger_levels)),\n url(regex=r'^r/(?P\\d+)/(?P.+)/$',\n view=wrap(contenttype_views.shortcut),\n name='view_on_site'),\n url(regex=r'^(?P\\w+)/$',\n view=wrap(self.app_index),\n name='app_list')\n )\n\n for model, model_admin in six.iteritems(self._registry):\n urlpatterns += patterns(\n '',\n url(regex=r'^%s/%s/' % (model._meta.app_label,\n model._meta.model_name),\n view=include(model_admin.urls))\n )\n\n return urlpatterns\n\ntribus_admin = TribusAdmin()\n","repo_name":"CanaimaKueka/tribus","sub_path":"tribus/web/admin/sites.py","file_name":"sites.py","file_ext":"py","file_size_in_byte":1839,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"81"}
+{"seq_id":"74921640265","text":"\"\"\"\nThis file is part of the TheLMA (THe Laboratory Management Application) project.\nSee LICENSE.txt for licensing, CONTRIBUTORS.txt for contributor information.\n\nThis handler converts the results of the library member parser into a\nmolecule design set (:class:`thelma.entities.molecule.MoleculeDesignSet`).\nAt this, the handler tries to determine the stock sample molecule design ID\nfor every molecule design set. If there is no ID, a new ID is created.\n\nAAB\n\"\"\"\nfrom everest.entities.utils import get_root_aggregate\nfrom everest.querying.specifications import cntd\nfrom thelma.tools.handlers.base import BaseParserHandler\nfrom thelma.tools.parsers.libmembers import LibraryMemberParser\nfrom thelma.interfaces import IMoleculeDesignPool\nfrom thelma.entities.moleculedesign import MoleculeDesignPoolSet\nfrom thelma.entities.moleculetype import MoleculeType\n\n\n__docformat__ = 'reStructuredText en'\n\n__all__ = ['LibraryMemberParserHandler',\n ]\n\n\nclass LibraryMemberParserHandler(BaseParserHandler):\n \"\"\"\n Converts the results of the :class:`LibraryMemberParser` into a\n :class:`MoleculeDesignSet`.\n\n **Return Value:** :class:`PoolSet`\n \"\"\"\n NAME = 'Library Member Parser Handler'\n\n _PARSER_CLS = LibraryMemberParser\n\n def __init__(self, stream, number_molecule_designs, molecule_type,\n parent=None):\n \"\"\"\n Constructor.\n\n :param int number_molecule_designs: The number of molecule designs\n per cell (list).\n :param molecule_type: The expected molecule type for all molecule\n designs in the library.\n :type molecule_type: :class:`thelma.entities.moleculetype.MoleculeType`\n \"\"\"\n BaseParserHandler.__init__(self, stream=stream, parent=parent)\n self.number_molecule_designs = number_molecule_designs\n self.molecule_type = molecule_type\n #: Map molecule design ID -> molecule design.\n self.__md_map = None\n #: Contains the stock sample molecule design sets for the final\n #: library pool set.\n self.__library_sets = None\n\n def reset(self):\n BaseParserHandler.reset(self)\n self.__md_map = dict()\n self.__library_sets = set()\n\n def _convert_results_to_entity(self):\n self.add_info('Convert parser results ...')\n self.__check_input()\n if not self.has_errors():\n self.__get_molecule_design_pools()\n if not self.has_errors():\n self.return_value = MoleculeDesignPoolSet(\n molecule_type=self.molecule_type,\n molecule_design_pools=self.__library_sets)\n self.add_info('Conversion completed.')\n\n def __check_input(self):\n self._check_input_class('number molecule designs',\n self.number_molecule_designs, int)\n self._check_input_class('molecule type', self.molecule_type,\n MoleculeType)\n\n def __get_molecule_design_pools(self):\n self.add_debug('Fetch molecule design pools for IDs ...')\n invalid_length = []\n found_md_ids = set()\n for md_ids in self.parser.molecule_design_lists:\n if len(md_ids) != self.number_molecule_designs:\n invalid_length.append(\n '-'.join([str(md_id) for md_id in md_ids]))\n continue\n for md_id in md_ids:\n found_md_ids.add(md_id)\n if len(invalid_length) > 0:\n msg = 'Some of the specified molecule design pools do not ' \\\n 'have the expected number of molecule designs (%i): %s.' \\\n % (self.number_molecule_designs, ', '.join(invalid_length))\n self.add_error(msg)\n return\n invalid_type = []\n agg = get_root_aggregate(IMoleculeDesignPool)\n agg.filter = cntd(molecule_design_id=found_md_ids)\n iterator = agg.iterator()\n while True:\n try:\n md = iterator.next()\n except StopIteration:\n break\n else:\n if md.molecule_type != self.molecule_type:\n invalid_type.append(md.id)\n self.__md_map[md.id] = md\n # Search for missing molecule designs.\n if len(found_md_ids) != len(self.__md_map):\n diff = found_md_ids.symmetric_difference(self.__md_map.keys())\n diff = sorted(list(diff))\n msg = 'The following molecule designs could not be found in ' \\\n 'the DB: %s.' % (', '.join([str(md_id) for md_id in diff]))\n self.add_error(msg)\n if len(invalid_type) > 0:\n msg = 'The molecule designs in the list have different molecule ' \\\n 'types. Expected: %s. Others (molecule designs): %s.' \\\n % (self.molecule_type,\n ', '.join([str(md_id) for md_id in sorted(invalid_type)]))\n self.add_error(msg)\n","repo_name":"helixyte/TheLMA","sub_path":"thelma/tools/handlers/libmembers.py","file_name":"libmembers.py","file_ext":"py","file_size_in_byte":5002,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"351504479","text":"import requests\nimport configparser\nfrom flask import Flask, render_template, request\n\napp = Flask(__name__)\n\n# declarative to specify the route to take on my flask application\n# in this case it is the landing page/dashboard\n@app.route('/')\ndef weather_dashboard():\n return render_template('landing_page.html')\n\n# declarative to specify the route to results page once a user inputs a zip code\n@app.route('/results', methods=['POST'])\ndef render_results():\n zip_code = request.form['zipCode']\n\n api_key = get_api_key()\n data = get_weather_results(zip_code, api_key)\n\n # Formats JSON dictionary by parsing response to key values below\n # to render to the user\n temp = \"{0:.2f}\".format(data[\"main\"][\"temp\"])\n feels_like = \"{0:.2f}\".format(data[\"main\"][\"feels_like\"])\n weather = data[\"weather\"][0][\"main\"]\n location = data[\"name\"]\n\n return render_template('results_page.html', location=location, temp=temp, \n feels_like=feels_like, weather=weather)\n\n# Function to store configurations to specify my application and to store my api key\ndef get_api_key():\n config = configparser.ConfigParser()\n config.read('config.ini')\n return config['openweathermap']['api']\n\n# Function to get weather results based on zip code, and my api key.\n# Formatted to take in the parameters that user puts into the system\ndef get_weather_results(zip_code, api_key):\n api_url = \"http://api.openweathermap.org/\" \\\n \"data/2.5/weather?zip={}&units=imperial&appid={}\".format(zip_code, api_key)\n r = requests.get(api_url)\n return r.json()\n\nif __name__ == '__main__':\n app.run()\n","repo_name":"ChichaRonin/Weather-API-Application","sub_path":"src/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1614,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"27686383243","text":"from flask import Blueprint, Flask, redirect, render_template, request\n\nfrom models.student import Student\nfrom models.registration import Registration\n\nimport repositories.student_repository as student_repository\nimport repositories.course_repository as course_repository\nimport repositories.registration_repository as registration_repository\n\nstudent_blueprint = Blueprint(\"students\", __name__)\n\n@student_blueprint.route(\"/students\")\ndef list_students():\n students = student_repository.select_all()\n return render_template (\"students/index.html\", students = students)\n\n@student_blueprint.route(\"/students/new\")\ndef new_student():\n courses = course_repository.select_all()\n return render_template(\"students/new.html\", courses= courses)\n\n@student_blueprint.route(\"/students\", methods=[\"POST\"])\ndef create_student():\n name = request.form[\"name\"]\n dob = request.form[\"dob\"]\n experience = request.form[\"experience\"]\n email = request.form[\"email\"]\n phone = request.form['phone']\n membership= request.form[\"membership\"]\n course_id = request.form[\"course_id\"]\n\n course = course_repository.select(course_id)\n new_student = Student(name, dob, experience, email, phone, membership)\n new_registration = Registration(course, new_student)\n\n student_repository.save(new_student)\n registration_repository.save(new_registration)\n return redirect(f\"/students/{new_student.id}\")\n\n@student_blueprint.route(\"/students/\")\ndef show(id):\n student = student_repository.select(id)\n courses = registration_repository.select_courses(id)\n return render_template(\"students/show.html\", student=student, courses= courses)\n\n@student_blueprint.route(\"/students//delete\", methods=[\"POST\"])\ndef delete_student(id):\n student_repository.delete(id)\n return redirect(\"/students\")\n\n@student_blueprint.route(\"/students//edit\")\ndef edit_student(id):\n student = student_repository.select(id)\n return render_template('students/edit.html', student= student)\n\n@student_blueprint.route(\"/students/\", methods=[\"POST\"])\ndef update_student(id):\n name = request.form[\"name\"]\n dob = request.form[\"dob\"]\n experience = request.form[\"experience\"]\n email = request.form[\"email\"]\n phone = request.form['phone']\n membership= request.form[\"membership\"]\n\n student_to_update = Student(name, dob, experience, email, phone, membership, id)\n\n student_repository.update(student_to_update)\n return redirect(f\"/students/{student_to_update.id}\")\n","repo_name":"mariamuzas/mariuca_school_app_project","sub_path":"controllers/student_controller.py","file_name":"student_controller.py","file_ext":"py","file_size_in_byte":2494,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"33659934515","text":"# from jose import jws\n# contract=\"This is a sample contract if you are agree on this pls sign the doc\"\n# signed = jws.sign({\"contract\":contract}, 'secret', algorithm='HS256')\n# print(signed)\n\n\n# verification=jws.verify(signed, 'secret', algorithms=['HS256'])\n# print(verification)\nfrom Crypto.Cipher import AES\n\n\nkey = b'Sixteen byte key'\n# print(type(key))\ncipher = AES.new(key, AES.MODE_EAX)\ndata=bytes(\"hemendrasharma\",'utf-8')\nnonce = cipher.nonce\n# print(nonce)\n# n = \"4Y\\x89\\x7f\\x1dI\\x05}\\x19\\xa3\\xf2, \\rJ\\xe4}\"\n# print(nonce == bytes(n,\"utf-8\"))\nciphertext, tag = cipher.encrypt_and_digest(data)\nprint(str(ciphertext))\n# # plaintext = cipher.decrypt(ciphertext)\n# # print(plaintext)\n\ncipher1 = AES.new(key, AES.MODE_EAX,nonce)\nplaintext = cipher1.decrypt(ciphertext)\n# print(plaintext.decode('utf-8'))\n","repo_name":"adityadev11/Anti_tamper","sub_path":"dsa.py","file_name":"dsa.py","file_ext":"py","file_size_in_byte":810,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1820674964","text":"# 실행 시간 (22947번)\n\"\"\"\n 문제: https://www.acmicpc.net/problem/22947\n 브루트포스, 위상정렬\n 문제를 잘못 읽는 바람에 맞왜틀을 했다.\n 시작 작업과 마무리 작업을 제외한 min(N - 2, K)개의 작업 시간을 강제로 0으로 만들어야하는데,\n 모든 작업 중에서 K개의 작업들의 조합에 대해서 시간을 강제로 0으로 만들고 있었다.\n 마지막 작업은 역진입 차수 배열을 통해서 찾고(역진입 차수 배열에서 진입 차수가 0인 노드가 마지막 작업임)\n 2부터 N까지 역진입 차수 배열에서 0이 아닌 노드들을 조합을 만들 배열에 담아서 K개의 조합을 구성한다.\n 나머지 부분은 위상 정렬을 통해서 모든 작업을 완료할 수 있는 가장 작은 시간을 찾는다.\n\"\"\"\nimport sys\nfrom collections import deque\nfrom itertools import combinations\ninput = sys.stdin.readline\n\nN, M, K = map(int, input().split()) # 작업 개수, 작업 순서 수, 실행 시간이 0인 수\ntime = [0] + list(map(int, input().split())) # 실행 시간\ngraph = [[] for _ in range(N + 1)]\nindegree = [0] * (N + 1) # 진입 차수\nreverse_ind = [0] * (N + 1) # 역진입차수(마지막 작업을 찾기 위함)\n\nfor _ in range(M):\n s, e = map(int, input().split())\n graph[s].append(e)\n indegree[e] += 1\n reverse_ind[s] += 1\n\narr = []\nfor i in range(2, N + 1):\n if reverse_ind[i]:\n arr.append(i)\n\ncombs = list(combinations(arr, K))\nmin_time = sys.maxsize\nfor comb in combs:\n tmp_ind = indegree[:] # 진입 차수 복사\n tmp_t = time[:] # 실행 시간 복사\n T = [0] * (N + 1)\n T[1] = tmp_t[1]\n for n in comb:\n tmp_t[n] = 0\n\n q = deque([(T[1], 1)]) # 1은 항상 작업 시작임\n while q:\n cost, cur = q.popleft()\n for nxt in graph[cur]:\n tmp_ind[nxt] -= 1\n T[nxt] = max(cost + tmp_t[nxt], T[nxt])\n if tmp_ind[nxt] == 0:\n q.append((T[nxt], nxt))\n\n min_time = min(min_time, max(T))\n\nprint(min_time)\n","repo_name":"siwon-park/Problem-Solving","sub_path":"Baekjoon_Solve/Topology/22947.py","file_name":"22947.py","file_ext":"py","file_size_in_byte":2059,"program_lang":"python","lang":"ko","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"41212978388","text":"import torch\nimport torch.nn as nn\nclass LSTM_Intent(nn.Module):\n def __init__(self, vocab_size, embedding_dim, hidden_units, output_size,n_layers):\n super(LSTM_Intent, self).__init__()\n self.hidden_units = hidden_units\n self.embedding_dim = embedding_dim\n self.vocab_size = vocab_size\n self.output_size = output_size\n self.n_layers = n_layers\n \n # layers\n self.embedding = nn.Embedding(self.vocab_size, self.embedding_dim)\n self.lstm = nn.LSTM(self.embedding_dim, self.hidden_units,self.n_layers)\n self.fc = nn.Linear(self.hidden_units, self.output_size)\n self.sig = nn.Sigmoid()\n# self.do = nn.Dropout(0.3)\n# self.do1 = nn.Dropout(0.5)\n def initialize_hidden_state(self):\n return torch.zeros((self.n_layers, 1, self.hidden_units)),torch.zeros((self.n_layers,1, self.hidden_units))\n \n def forward(self, x):\n# self.hidden = self.initialize_hidden_state()\n# print(x.shape)\n out = self.embedding(x)\n# print(out)\n# out = self.do(out)\n out, self.hidden = self.lstm(out) # max_len X batch_size X hidden_units\n# print(out.shape)\n# out = self.do1(out)\n out = self.fc(out)\n out = out[-1,:,:] \n# print(out)\n out = self.sig(out)\n return out","repo_name":"reyzaldindra/intun_implimin","sub_path":"app/bpjs_intent.py","file_name":"bpjs_intent.py","file_ext":"py","file_size_in_byte":1347,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"15671579029","text":"# Input: text T\r\nT = input()\r\n\r\n# Input: pattern P\r\nP = input()\r\n\r\n# Import re module\r\nimport re\r\n\r\n# Find all occurrences of P in T using re.finditer\r\nmatches = re.finditer(P,T)\r\n\r\n# Initialize count of occurrences\r\ncount = 0\r\n\r\n# Initialize list of start and end indices\r\nindices = []\r\n\r\n# Loop through each match object\r\nfor match in matches:\r\n # Increment count by 1\r\n count += 1\r\n \r\n # Get start and end indices of match using match.start() and match.end()\r\n start = match.start()\r\n end = match.end() - 1\r\n \r\n # Append a tuple of start and end indices to indices list\r\n indices.append((start,end))\r\n\r\n# Print the output \r\nif count == 0:\r\n # If no occurrence is found, print 0\r\n print(0)\r\nelse:\r\n # If some occurrences are found, print count and indices list \r\n print(count)\r\n print(*indices)","repo_name":"dikaapranata/Praktikum-ASA","sub_path":"praktikum 3/pt.py","file_name":"pt.py","file_ext":"py","file_size_in_byte":838,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"4395012867","text":"# 给定一个序列,输出其下一个更大序列,若已最大,则从头开始。时间复杂度O(1)\n# 1, 2, 3 -> 1, 3, 2\n\n# 思路:\n# 找递增结束的位置,换上比该数大的最小数\n\nclass Solution:\n def nextPermutation(self, nums):\n i = len(nums) - 1\n if nums[i - 1] >= nums[i]:\n i = i - 1\n print(i)\n if i == 1:\n m = 0\n n = len(nums) - 1\n print(m, n)\n while m <= n:\n nums[m], nums[n] = nums[n], nums[m]\n m = m + 1\n n = n - 1\n print(m, n)\n else:\n k = 0\n for j in range(i, len(nums)):\n min = nums[j]\n print(j)\n if nums[j] > nums[i - 1] and nums[j] <= min:\n min = nums[j]\n k = j\n nums[i - 1], nums[k] = nums[k], nums[i - 1]\n return nums\n\n\nif __name__ == '__main__':\n nums = [1, 1, 5]\n s = Solution()\n result = s.nextPermutation(nums)\n print(result)\n\n\n\n","repo_name":"huhuzwxy/leetcode","sub_path":"array/31_next_permutation.py","file_name":"31_next_permutation.py","file_ext":"py","file_size_in_byte":1099,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"74585072263","text":"\"\"\"A collection of model configurations.\"\"\"\n\nimport optuna\nimport pytorch_lightning as pl\n\nfrom models.feedforward import FeedForward\nfrom models.tcn import TemporalConvNet\n\nfrom utils.pl_utils import get_training_config\nfrom utils.config_utils import BaseConfig\n\n\nclass Config(BaseConfig):\n \"\"\"Global model configuration (does not change / is not tunable).\"\"\"\n def __init__(self, study_name: str = 'study', batch_size: int = 10, *args, **kwargs):\n \"\"\"Model config.\n\n The batch size can be set manually as different models may require different sized batches.\n\n Args:\n study_name (str, optional): the study name. Defaults to `study`.\n batch_size (int, optional): the batch size. Defaults to 10.\n \"\"\"\n super(Config, self).__init__(*args, **kwargs)\n\n self.STUDY_NAME = study_name\n\n self.NUM_INPUTS = 1\n self.NUM_OUTPUTS = 1\n\n self.BATCH_SIZE = batch_size\n self.MIN_EPOCHS = 1\n self.MAX_EPOCHS = 40\n\n self.SEED = 23427\n\n # Not saved (bc. lowercase).\n self.log_freq = 10\n self.num_workers = 0\n self.val_loss_name = 'val_loss'\n\n\ndef feedforward(config: BaseConfig, trial: optuna.trial.Trial) -> pl.LightningModule:\n \"\"\"Returns a tunable PyTorch lightning feedforward module.\n\n Args:\n config (BaseConfig): the hard-coded configuration.\n trial (optuna.Trial): optuna trial.\n\n Returns:\n pl.LightningModule: a lightning module.\n \"\"\"\n\n model = FeedForward(\n num_inputs=config.NUM_INPUTS,\n num_outputs=config.NUM_OUTPUTS,\n num_hidden=trial.suggest_int('num_hidden', 1, 4),\n num_layers=trial.suggest_int('num_layers', 1, 2),\n dropout=trial.suggest_float('dropout', 0.0, 0.5),\n activation=trial.suggest_categorical('activation', ['relu', 'none'])\n )\n\n training_config = get_training_config(\n lr=trial.suggest_loguniform('lr', 1e-3, 1e-0),\n weight_decay=trial.suggest_loguniform('weight_decay', 1e-5, 1e-1),\n max_epochs=config.MAX_EPOCHS)\n\n pl_model = TemporalConvNet(\n training_config=training_config,\n lr=trial.suggest_loguniform('lr', 1e-3, 1e-0),\n weight_decay=trial.suggest_loguniform('weight_decay', 1e-5, 1e-1),\n max_epochs=config.MAX_EPOCHS\n )\n\n return pl_model\n\n\ndef tcn(config: BaseConfig, trial: optuna.trial.Trial) -> pl.LightningModule:\n \"\"\"Returns a tunable PyTorch lightning tcn module.\n\n Args:\n config (BaseConfig): the hard-coded configuration.\n trial (optuna.Trial): optuna trial.\n\n Returns:\n pl.LightningModule: a lightning module.\n \"\"\"\n\n training_config = get_training_config(\n lr=trial.suggest_loguniform('lr', 1e-3, 1e-0),\n weight_decay=trial.suggest_loguniform('weight_decay', 1e-5, 1e-1),\n max_epochs=config.MAX_EPOCHS)\n\n tcn = TemporalConvNet(\n training_config=training_config,\n num_inputs=config.NUM_INPUTS,\n num_outputs=config.NUM_OUTPUTS,\n num_hidden=trial.suggest_int('num_hidden', 1, 4),\n kernel_size=trial.suggest_int('kernel_size', 2, 4),\n num_layers=trial.suggest_int('num_layers', 1, 2),\n dropout=trial.suggest_float('dropout', 0.1, 0.3)\n )\n\n return tcn\n","repo_name":"bask0/q10hybrid","sub_path":"project/model_configs.py","file_name":"model_configs.py","file_ext":"py","file_size_in_byte":3271,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"4126862988","text":"from collections import deque\nN = int(input())\nd = deque()\nfor i in range(N):\n a = input().split()\n if a[0] == \"append\":\n d.append(int(a[1]))\n elif a[0] == \"pop\":\n d.pop()\n elif a[0] == \"popleft\":\n d.popleft()\n elif a[0] == \"appendleft\":\n d.appendleft(int(a[1]))\nprint(' '.join(map(str, d)))","repo_name":"elisa2602/PCS2_HOMEWORK","sub_path":"pcs2_hm1/collections/collections_deque().py","file_name":"collections_deque().py","file_ext":"py","file_size_in_byte":334,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"11175292540","text":"\"\"\"\npipeline designed to create/update active roster table within the NBA states database\n\"\"\"\n\nimport pymysql\nimport re\n\ndef drop_table(connection):\n drop_table_statement = 'drop table active_rosters'\n sql_execute(drop_table_statement, connection)\n\ndef get_cities(connection): #function to extract all cities from team_info entity\n\n city_extract_query = \"select team from team_info\"\n cities_raw_extract = sql_execute(city_extract_query, connection)\n cities = data_extract_from_raw_data(cities_raw_extract)\n return cities\n\ndef get_team_names(connection):\n\n team_names_extract_query = \"select distinct current_team from player_info\"\n team_names_extract = sql_execute(team_names_extract_query, connection)\n teams = data_extract_from_raw_data(team_names_extract)\n return teams\n\ndef data_extract_from_raw_data(data):\n\n regex_catch = re.compile('[A-Za-z].*\\w')\n cleaned_data_list = []\n\n for raw_line in data:\n try:\n data_string = re.search(regex_catch, str(raw_line))\n cleaned_data_list.append(data_string.group(0))\n except AttributeError:\n print(\"Data not captured in regex expression\", raw_line)\n return cleaned_data_list\n\ndef create_city_team_dict(teams, cities):\n\n city_team_dict = {}\n for city in cities:\n for team in teams:\n if city in team:\n city_team_dict[city] = team\n #hard coding in LA Lakers\n city_team_dict[\"LA Lakers\"] = \"Los Angeles Lakers\"\n return city_team_dict\n\ndef create_update_statements(city_team_dict, connection):\n\n for city_team in city_team_dict:\n update_statement = \"update team_info set team = '\" + str(city_team_dict[city_team]) + \"' where team = '\" + str(city_team) + \"'\"\n sql_execute(update_statement, connection)\n\ndef create_active_roster_table(connection):\n\n create_table_statement = 'create table nba_stats.active_rosters( \\\n select distinct reg.player_id, play.name, team.team_id, reg.team, stand.conference \\\n from RegularSeasonAverages as reg \\\n inner join player_info as play on play.player_id = reg.player_id \\\n inner join team_info as team on (team.team = reg.team) \\\n inner join team_standings as stand on ((team.team = stand.team) and (stand.season = reg.season)) \\\n where stand.season = (select max(season) from team_standings))'\n sql_execute(create_table_statement, connection)\n\n add_primary_key = 'alter table nba_stats.active_rosters\\\n add constraint active_roster_pk primary key(player_id, name, team_id, team)'\n sql_execute(add_primary_key, connection)\n\ndef sql_execute(query, connection):\n\n exe = connection.cursor()\n exe.execute(query)\n return exe.fetchall()\n\n## main function\nmyConnection = pymysql.connect(host=\"localhost\", user=\"root\", password=\"Sk1ttles\", db=\"nba_stats\", autocommit=\"true\")\n\n#drop_table(myConnection)\ncities = get_cities(myConnection)\nteams = get_team_names(myConnection)\ncity_team_dict = create_city_team_dict(teams, cities)\ncreate_update_statements(city_team_dict, myConnection)\ncreate_active_roster_table(myConnection)\n","repo_name":"pstoeber/Personal-Projects","sub_path":"Pipeline Development/active_roster.py","file_name":"active_roster.py","file_ext":"py","file_size_in_byte":3271,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"3265659409","text":"#Example 1: \r\n# Sorting the element name by alphbetic order \r\nearth_metals = [\"Beryllium\", \"Magnesium\", \"Calcium\", \"Strontium\", \"Barium\", \"Radium\"]\r\nearth_metals.sort() # by default , python assumes you want to sort it by alphbetic order \r\n\r\nprint (earth_metals)\r\n\r\n# sort by reversed alphbetic order \r\nearth_metals.sort(reverse=True)\r\nprint (earth_metals)\r\n\r\n# remember sorting changes of things, so the list can be sorted, however if you put the\r\n# above data into a tuples, then you can not sort it directly because tuples are immutable. \r\n\r\n# the following will fail. \r\n# earth_metals = (\"Beryllium\", \"Magnesium\", \"Calcium\", \"Strontium\", \"Barium\", \"Radium\")\r\n# earth_metals.sort()\r\n\r\n# however you can use the sorted function instead to sort the tuples. \r\nearth_metals = (\"Beryllium\", \"Magnesium\", \"Calcium\", \"Strontium\", \"Barium\", \"Radium\")\r\nsorted_earth_metals = sorted(earth_metals)\r\nprint (sorted_earth_metals)\r\nprint (earth_metals)\r\n\r\ndata = (7, 2, 5, 6, 1, 3, 9, 10, 4, 8)\r\nprint (sorted(data))\r\n\r\n\r\n# Exmple 2: \r\n# we want to sort by the 2nd value of the tuple in the planet list. \r\nplanets = [(\"Mercury\", 2440, 5.43), \r\n (\"Venus\", 6052, 5.24), \r\n (\"Mars\", 3397, 1.530), \r\n (\"Jupiters\", 71452, 1.33, 5.21)]\r\n\r\n# use lambda to get the size of the planet. \r\nsize = lambda planet: planet[1]\r\n\r\n# use the lambda function in the sort \"key\" value, \r\nplanets.sort(key=size, reverse=True)\r\nprint (planets)","repo_name":"hsnyxfhyqhyh/python3","sub_path":"pythonExcise/pythonExcise/basics/SortingTest.py","file_name":"SortingTest.py","file_ext":"py","file_size_in_byte":1441,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"35890659978","text":"import tensorflow as tf\nimport numpy as np\nimport tensorflow.keras\nfrom tensorflow.keras import Model, Sequential\nimport tensorflow.keras.utils\nfrom tensorflow.keras.regularizers import l2, l1\nfrom tensorflow.keras.layers import Dense, Input, BatchNormalization, LeakyReLU, Flatten, GaussianDropout, Lambda, LSTM, TimeDistributed, RepeatVector, Embedding, LayerNormalization, Dropout, MultiHeadAttention, Conv1D\nfrom tensorflow.keras.losses import MeanSquaredError\nimport tensorflow.keras.backend as K\nfrom plot_outputs import *\nimport time\n\n\nclass MLP():\n def __init__(self, mlp_parameters, autoencoder_parameters, mlp_optimizer, mlp_reg, num_classes):\n # mlp\n self.mlp_input = Dense(autoencoder_parameters['autoencoder_batch_size'], input_shape=(None, None))\n self.mlp_layer_1 = Dense(round(autoencoder_parameters['autoencoder_batch_size'] * 0.9), bias_initializer='zeros', kernel_initializer='glorot_uniform', activity_regularizer=l2(mlp_reg))\n self.mlp_layer_2 = Dense(round(autoencoder_parameters['autoencoder_batch_size'] * 0.7), bias_initializer='zeros', kernel_initializer='glorot_uniform', activity_regularizer=l2(mlp_reg))\n self.mlp_layer_3 = Dense(round(autoencoder_parameters['autoencoder_batch_size'] * 0.5), bias_initializer='zeros', kernel_initializer='glorot_uniform', activity_regularizer=l2(mlp_reg))\n self.mlp_layer_4 = Dense(round(autoencoder_parameters['autoencoder_batch_size'] * 0.3), bias_initializer='zeros', kernel_initializer='glorot_uniform', activity_regularizer=l2(mlp_reg))\n self.mlp_layer_5 = Dense(round(autoencoder_parameters['autoencoder_batch_size'] * 0.1), bias_initializer='zeros', kernel_initializer='glorot_uniform', activity_regularizer=l2(mlp_reg))\n self.mlp_output_sigmoid = Dense(num_classes, activation='sigmoid')\n self.mlp_output_softmax = Dense(num_classes, activation='softmax')\n # mlp_optimizer\n self.mlp_op = mlp_optimizer\n # leaky\n self.mlp_leaky = LeakyReLU(alpha=0.3)\n\n def run_mlp(self, reconstructed_latent):\n \"\"\"\n maybe think about having separate gradient for this + optimizer and looop through this\n need to add leaky relu activation\n \"\"\"\n mlp_input = self.mlp_input(reconstructed_latent)\n mlp1 = self.mlp_layer_1(mlp_input)\n mlp_leaky1 = self.mlp_leaky(mlp1)\n mlp2 = self.mlp_layer_2(mlp_leaky1)\n mlp_leaky2 = self.mlp_leaky(mlp2)\n mlp3 = self.mlp_layer_3(mlp_leaky2)\n mlp_leaky3 = self.mlp_leaky(mlp3)\n mlp4 = self.mlp_layer_4(mlp_leaky3)\n mlp_leaky4 = self.mlp_leaky(mlp4)\n mlp5 = self.mlp_layer_5(mlp_leaky4)\n mlp_leaky5 = self.mlp_leaky(mlp5)\n mlp_output_softmax = self.mlp_output_softmax(mlp_leaky5)\n return mlp_output_softmax\n\n def mlp_gradient(self, true_labels, reconstructed_latent, num_classes):\n with tf.GradientTape() as tape2:\n # mlp\n tape2.watch(self.mlp_input.variables)\n tape2.watch(self.mlp_layer_1.variables)\n tape2.watch(self.mlp_layer_2.variables)\n tape2.watch(self.mlp_layer_3.variables)\n tape2.watch(self.mlp_layer_4.variables)\n tape2.watch(self.mlp_layer_5.variables)\n tape2.watch(self.mlp_output_softmax.variables)\n class_acc, class_loss = self.get_mlp_loss(true_labels, reconstructed_latent, num_classes)\n mlp_gradient = tape2.gradient(class_loss, [self.mlp_input.variables[0], self.mlp_input.variables[1],\n self.mlp_layer_1.variables[0], self.mlp_layer_1.variables[1], self.mlp_layer_2.variables[0], self.mlp_layer_2.variables[1],\n self.mlp_layer_3.variables[0], self.mlp_layer_3.variables[1], self.mlp_layer_4.variables[0], self.mlp_layer_4.variables[1],\n self.mlp_layer_5.variables[0], self.mlp_layer_5.variables[1], self.mlp_output_softmax.variables[0], self.mlp_output_softmax.variables[1]])\n return mlp_gradient, class_acc, class_loss\n\n def mlp_learn(self, true_labels, reconstructed_latent, num_classes):\n mlp_gradient, class_acc, class_loss = self.mlp_gradient(true_labels, reconstructed_latent, num_classes)\n self.mlp_op.apply_gradients(zip(mlp_gradient, [self.mlp_input.variables[0], self.mlp_input.variables[1], self.mlp_layer_1.variables[0], self.mlp_layer_1.variables[1],\n self.mlp_layer_2.variables[0], self.mlp_layer_2.variables[1], self.mlp_layer_3.variables[0], self.mlp_layer_3.variables[1],\n self.mlp_layer_4.variables[0], self.mlp_layer_4.variables[1], self.mlp_layer_5.variables[0], self.mlp_layer_5.variables[1],\n self.mlp_output_softmax.variables[0], self.mlp_output_softmax.variables[1]]))\n return class_acc, class_loss\n\n def get_mlp_loss(self, true_labels, reconstructed_latent, num_classes):\n mlp_output_softmax = self.run_mlp(reconstructed_latent)\n # mlp_output = [200, 8] array of probability value for each feature or neuron belonging to a particular class --> find max prob for class\n # mlp_output_softmax = mlp_output_softmax.numpy()\n # mlp_output_softmax = mlp_output_softmax.reshape(mlp_output_softmax.shape[0]*mlp_output_softmax.shape[1], mlp_output_softmax.shape[2])\n class_acc = (np.sum(np.equal(np.argmax(mlp_output_softmax, axis=1), true_labels)) / len(true_labels)) * 100\n encoded_labels = self.one_hot_encoding(labels=true_labels, num_classes=num_classes, convert_back=False)\n class_loss = -tf.reduce_sum(tf.math.log(mlp_output_softmax) * encoded_labels)\n return class_acc, class_loss\n\n @staticmethod\n def one_hot_encoding(labels, num_classes, convert_back):\n if convert_back is False:\n hot_encoded_labels = np.zeros((len(labels), num_classes))\n for x in range(len(labels)):\n hot_encoded_labels[x, int(labels[x])] = 1\n return hot_encoded_labels\n else:\n return [x for x in np.argwhere(labels == 1)[:, 1]]\n","repo_name":"ankushgpta2/Vision2DeepManifold","sub_path":"models/mlp_model.py","file_name":"mlp_model.py","file_ext":"py","file_size_in_byte":6236,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"7750215067","text":"from music21 import converter, clef, stream, pitch, note, meter, midi\nimport numpy as np\n\n\nKEY_SIG_OFFSET = 0\n\ndef parse_notes(midi_stream):\n melody_corpus = []\n\n last_pitch = 1\n chord_buffer = []\n prev_offset = 0.0\n for m in midi_stream.measures(1, None):\n time_sig = m.timeSignature\n for nr in m.flat.notesAndRests:\n offset_loc = nr.offset\n # pitch = nr.pitch.pitchClass + 1 if isinstance(nr, note.Note) else 0\n pitch = nr.pitch.midi if isinstance(nr, note.Note) else 0\n beat_strength = round(nr.beatStrength * 4.0, 0)\n duration = float(nr.quarterLength)\n\n note_repr = (pitch, beat_strength, duration)\n # note_repr = (pitch, duration)\n # Handle chords\n if nr.offset == prev_offset:\n if note_repr[0] > 0:\n chord_buffer.append(note_repr)\n else:\n if chord_buffer: # Choose tone from chord buffer closest to current note\n chord_melody_tone = sorted(chord_buffer, key=lambda x: abs(x[0] - pitch))[0]\n melody_corpus.append(chord_melody_tone)\n melody_corpus.append(note_repr)\n chord_buffer = []\n prev_offset = nr.offset\n\n return melody_corpus\n\n\ndef build_corpus(midi_files):\n melody_corpus = []\n for file in midi_files:\n midi_stream = converter.parse(file)\n midi_stream = midi_stream[0]\n if '1008' in file or '1011' in file:\n midi_stream.keySignature = midi_stream.keySignature.relative\n key_sig = midi_stream.keySignature\n print('Input file: {} ({})'.format(file, str(key_sig)))\n midi_stream.transpose(KEY_SIG_OFFSET - key_sig.tonic.pitchClass, inPlace=True)\n melody_corpus.extend(parse_notes(midi_stream))\n # map indices for constructing matrix representations\n melody_set = set(melody_corpus)\n notes_indices = {note: i for i, note in enumerate(melody_set)}\n indices_notes = {i: note for i, note in enumerate(melody_set)}\n\n return melody_corpus, melody_set, notes_indices, indices_notes\n","repo_name":"naoyak/JohaNN","sub_path":"corpus.py","file_name":"corpus.py","file_ext":"py","file_size_in_byte":2137,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"81"}
+{"seq_id":"19119327814","text":"from rest_framework import serializers\nfrom rest_framework.reverse import reverse\nfrom drf_spectacular.utils import extend_schema_serializer, extend_schema_field, OpenApiTypes\nfrom events.models import OfficeHour, Event2019, CrewAttendanceRecord\nfrom accounts.models import Officer\nfrom data.models import ResizedRedirect\nfrom pages.models import Page\nfrom spotify.models import Session, SpotifyUser, SongRequest\nfrom spotify.api import get_playback_state, queue_estimate\n\nfrom .models import TokenRequest\n\n\n# Create your serializers here.\n@extend_schema_serializer(\n exclude_fields=['img', 'class_year']\n)\nclass OfficerSerializer(serializers.ModelSerializer):\n def __init__(self, *args, **kwargs):\n super(OfficerSerializer, self).__init__(*args, **kwargs)\n\n if self.context.get('request', None):\n fields = self.context['request'].query_params.get('options')\n if fields:\n fields = fields.split(',')\n fields.append('title')\n fields.append('name')\n allowed = set(fields)\n current = set(self.fields.keys())\n for field_name in current - allowed:\n self.fields.pop(field_name)\n else:\n self.fields.pop('img')\n self.fields.pop('class_year')\n\n class_year = serializers.IntegerField(source='user.class_year')\n img = serializers.SerializerMethodField()\n name = serializers.CharField(source='user.name')\n\n def get_img(self, obj):\n if obj.img:\n return obj.img.img.url\n else:\n return None\n\n class Meta:\n model = Officer\n fields = ('title', 'name', 'img', 'class_year')\n\n\nclass HourSerializer(serializers.ModelSerializer):\n officer = serializers.CharField(source='officer.exec_position.title')\n location = serializers.CharField(source='location.name')\n\n class Meta:\n model = OfficeHour\n fields = ('officer', 'day', 'location', 'hour_start', 'hour_end')\n\n\nclass NotificationSerializer(serializers.BaseSerializer):\n def to_representation(self, instance):\n ref_id = \"LNLWN-\" + str(instance.pk)\n class_type = 2\n if instance.target == \"all\" or instance.target == \"All\":\n if instance.dismissible is True and instance.format == \"alert\":\n class_type = 2\n else:\n class_type = 1\n elif instance.dismissible is True:\n class_type = 3\n return {\n 'id': ref_id,\n 'class': class_type,\n 'format': instance.format,\n 'type': instance.type,\n 'expires': instance.expires,\n 'title': instance.title,\n 'message': instance.message\n }\n\n\nclass EventSerializer(serializers.ModelSerializer):\n location = serializers.CharField(source='location.name')\n datetime_start = serializers.DateTimeField(format=\"%Y-%m-%dT%H:%M:%S%z\", read_only=True)\n datetime_end = serializers.DateTimeField(format=\"%Y-%m-%dT%H:%M:%S%z\", read_only=True)\n\n class Meta:\n model = Event2019\n fields = ('id', 'event_name', 'description', 'location', 'datetime_start', 'datetime_end',)\n\n\nclass AttendanceSerializer(serializers.ModelSerializer):\n class Meta:\n model = CrewAttendanceRecord\n fields = ('user', 'event', 'checkin', 'checkout', 'active')\n\n\nclass RedirectSerializer(serializers.ModelSerializer):\n class Meta:\n model = ResizedRedirect\n fields = ('name', 'old_path')\n\n def to_representation(self, instance):\n path = instance.old_path\n if instance.old_path[0] == '/':\n path = instance.old_path[1:]\n return {\n 'title': instance.name,\n 'path': path,\n 'category': 'Redirects'\n }\n\n\nclass CustomPageSerializer(serializers.ModelSerializer):\n class Meta:\n model = Page\n fields = ('title', 'slug', 'sitemap_category')\n\n def to_representation(self, instance):\n category = 'Redirects'\n if instance.sitemap_category is not None:\n category = instance.sitemap_category\n return {\n 'title': instance.title,\n 'path': instance.slug,\n 'category': category\n }\n\n\nclass SpotifyUserSerializer(serializers.ModelSerializer):\n class Meta:\n model = SpotifyUser\n fields = ('id', 'display_name', 'spotify_id', 'personal')\n\n\nclass SpotifySessionReadOnlySerializer(serializers.ModelSerializer):\n id = serializers.CharField(source='slug')\n event = EventSerializer()\n user = SpotifyUserSerializer()\n urls = serializers.SerializerMethodField()\n\n is_playing = serializers.SerializerMethodField()\n device = serializers.SerializerMethodField()\n current_track = serializers.SerializerMethodField()\n runtime_ms = serializers.SerializerMethodField()\n\n def __new__(cls, *args, **kwargs):\n if kwargs.get('many', False) is True:\n context = kwargs.get('context', {})\n context.update({'has_many': True})\n kwargs.update({'context': context})\n\n return super().__new__(cls, *args, **kwargs)\n\n def __init__(self, *args, **kwargs):\n self.request = kwargs.pop('request', None)\n super(SpotifySessionReadOnlySerializer, self).__init__(*args, **kwargs)\n\n many = self.context.get('has_many', False)\n\n # If session is inactive, do not return playback state\n if isinstance(self.instance, Session):\n event = self.instance.event\n if not event.approved or event.cancelled or event.closed or event.reviewed or \\\n not self.instance.accepting_requests:\n many = True\n\n if many:\n self.fields.pop('is_playing')\n self.fields.pop('device')\n self.fields.pop('current_track')\n self.fields.pop('runtime_ms')\n elif isinstance(self.instance, Session):\n self.playback_state = get_playback_state(self.instance)\n\n # Do not show certain fields to unprivileged users\n if not self.request or not self.request.user or not self.request.user.has_perm('spotify.view_session'):\n self.fields.pop('user')\n self.fields.pop('private')\n self.fields.pop('auto_approve')\n not many and self.fields.pop('device')\n\n @extend_schema_field(OpenApiTypes.OBJECT)\n def get_urls(self, obj):\n urls = {\n 'request_form': reverse('spotify:request', args=[obj.slug], request=self.request),\n 'qr_code': reverse('spotify:qr', args=[obj.pk], request=self.request)\n }\n return urls\n\n @extend_schema_field(OpenApiTypes.BOOL)\n def get_is_playing(self, obj):\n if self.playback_state:\n return self.playback_state.get('is_playing', False)\n return False\n\n @extend_schema_field(OpenApiTypes.OBJECT)\n def get_device(self, obj):\n if self.playback_state:\n return self.playback_state.get('device', None)\n return None\n\n @extend_schema_field(OpenApiTypes.OBJECT)\n def get_current_track(self, obj):\n if self.playback_state:\n return self.playback_state.get('item', None)\n return None\n\n @extend_schema_field(OpenApiTypes.INT)\n def get_runtime_ms(self, obj):\n return queue_estimate(obj, True)\n\n class Meta:\n model = Session\n fields = ('id', 'event', 'user', 'allow_explicit', 'require_payment', 'private', 'accepting_requests',\n 'auto_approve', 'is_playing', 'runtime_ms', 'device', 'current_track', 'urls')\n depth = 1\n\n\nclass SpotifySessionWriteSerializer(serializers.ModelSerializer):\n event = serializers.PrimaryKeyRelatedField(\n queryset=Event2019.objects.filter(approved=True, reviewed=False, closed=False, cancelled=False),\n help_text=\"Primary key value of the corresponding event\"\n )\n user = serializers.PrimaryKeyRelatedField(\n queryset=SpotifyUser.objects.all(), help_text=\"Primary key value of the Spotify account to use\"\n )\n auto_approve = serializers.BooleanField(default=False, help_text=\"Attempt to automatically queue song requests\")\n private = serializers.BooleanField(default=False, help_text=\"Restrict session to LNL members\")\n\n def __init__(self, *args, **kwargs):\n partial = kwargs.get('partial', None)\n super(SpotifySessionWriteSerializer, self).__init__(*args, **kwargs)\n\n if partial:\n for field in self.fields:\n self.fields[field].required = False\n\n def update(self, instance, validated_data):\n validated_data['event'] = instance.event # Changing this field is not permitted\n return super(SpotifySessionWriteSerializer, self).update(instance, validated_data)\n\n class Meta:\n model = Session\n fields = ('event', 'user', 'accepting_requests', 'allow_explicit', 'auto_approve', 'private')\n\n\nclass SongRequestSerializer(serializers.ModelSerializer):\n session = serializers.CharField(source='session.slug')\n queued_ts = serializers.DateTimeField(source='queued')\n requestor = serializers.SerializerMethodField()\n urls = serializers.SerializerMethodField()\n\n @extend_schema_field(OpenApiTypes.OBJECT)\n def get_requestor(self, obj):\n return {\"name\": obj.submitted_by, \"email\": obj.email, \"phone\": obj.phone}\n\n @extend_schema_field(OpenApiTypes.OBJECT)\n def get_urls(self, obj):\n return {'spotify_url': 'https://open.spotify.com/track/{}'.format(obj.identifier)}\n\n class Meta:\n model = SongRequest\n fields = ('id', 'session', 'name', 'duration', 'approved', 'queued_ts', 'requestor', 'urls')\n\n\nclass TokenRequestSerializer(serializers.ModelSerializer):\n APIKey = serializers.CharField(help_text=\"Your application's API key\")\n username = serializers.CharField(source='user.username', help_text=\"The user's username\")\n code = serializers.IntegerField(help_text=\"The user's verification code\")\n\n class Meta:\n model = TokenRequest\n fields = ('code', 'APIKey', 'username')\n","repo_name":"WPI-LNL/lnldb","sub_path":"api/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":10090,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"81"}
+{"seq_id":"43121003548","text":"from datetime import datetime\nimport unittest\nfrom eqcatalogue import models as catalogue\nimport geoalchemy\nfrom tests.test_utils import in_data_dir\n\n\nclass ShouldCreateAlchemyTestCase(unittest.TestCase):\n\n def setUp(self):\n self.catalogue = catalogue.CatalogueDatabase(memory=True, drop=True)\n self.session = self.catalogue.session\n\n def tearDown(self):\n self.session.commit()\n\n def test_drop(self):\n self.catalogue = catalogue.CatalogueDatabase(\n drop=True, filename=in_data_dir(\"test_drop.db\"))\n self.catalogue = catalogue.CatalogueDatabase(memory=True, drop=True)\n\n def test_eventsource(self):\n event_source = catalogue.EventSource(name=\"test1\")\n self.session.add(event_source)\n self.assertEqual(self.session.query(\n catalogue.EventSource).filter_by(name='test1').count(), 1)\n\n def test_agency(self):\n eventsource = catalogue.EventSource(name=\"test2\")\n self.session.add(eventsource)\n\n agency = catalogue.Agency(source_key=\"test\", eventsource=eventsource)\n self.session.add(agency)\n self.assertEqual(\n self.session.query(catalogue.Agency).filter_by(\n source_key='test').count(),\n 1)\n\n def test_event(self):\n eventsource = catalogue.EventSource(name=\"test3\")\n self.session.add(eventsource)\n\n event = catalogue.Event(source_key=\"test\", eventsource=eventsource)\n self.session.add(event)\n self.assertEqual(\n self.session.query(catalogue.Event).filter_by(\n source_key='test').count(), 1)\n\n def test_origin(self):\n eventsource = catalogue.EventSource(name=\"test4\")\n self.session.add(eventsource)\n\n origin = catalogue.Origin(source_key=\"test\", eventsource=eventsource,\n position=geoalchemy.WKTSpatialElement(\n 'POINT(-81.40 38.08)'),\n time=datetime.now(),\n depth=3)\n self.session.add(origin)\n self.assertEqual(\n self.session.query(catalogue.Origin).filter(\n catalogue.Origin.depth > 2).count(), 1)\n\n def test_magnitudemeasure(self):\n eventsource = catalogue.EventSource(name=\"test4\")\n self.session.add(eventsource)\n\n event = catalogue.Event(source_key=\"test\", eventsource=eventsource)\n self.session.add(event)\n\n agency = catalogue.Agency(source_key=\"test\", eventsource=eventsource)\n self.session.add(agency)\n\n origin = catalogue.Origin(\n source_key=\"test\", eventsource=eventsource,\n position=geoalchemy.WKTSpatialElement('POINT(-81.40 38.08)'),\n time=datetime.now(),\n depth=1)\n self.session.add(origin)\n\n measure = catalogue.MagnitudeMeasure(\n event=event, agency=agency, origin=origin, scale='mL', value=5.0)\n self.session.add(measure)\n\n self.assertEqual(\n self.session.query(catalogue.MagnitudeMeasure).count(), 1)\n\n def test_get_or_add(self):\n event_source1, created = self.catalogue.get_or_create(\n catalogue.EventSource, {'name': \"test_5\"})\n self.assertTrue(created)\n event_source2, created = self.catalogue.get_or_create(\n catalogue.EventSource, {'name': \"test_5\"})\n self.assertFalse(created)\n self.assertEqual(event_source1, event_source2)\n\n def create_test_fixture(self):\n eventsource = catalogue.EventSource(name=\"AnEventSource\")\n self.session.add(eventsource)\n\n first_event = catalogue.Event(source_key=\"1st\",\n eventsource=eventsource)\n second_event = catalogue.Event(source_key=\"2nd\",\n eventsource=eventsource)\n self.session.add(first_event)\n self.session.add(second_event)\n\n agency_one = catalogue.Agency(source_key=\"1st\",\n eventsource=eventsource, name='Tatooine')\n agency_two = catalogue.Agency(source_key=\"2nd\",\n eventsource=eventsource, name='Alderaan')\n agency_three = catalogue.Agency(source_key=\"3rd\",\n eventsource=eventsource, name='DeathStar')\n self.session.add(agency_one)\n self.session.add(agency_two)\n self.session.add(agency_three)\n\n origin = catalogue.Origin(\n source_key=\"test\", eventsource=eventsource,\n position=geoalchemy.WKTSpatialElement('POINT(-81.40 38.08)'),\n time=datetime.now(),\n depth=1)\n self.session.add(origin)\n\n measure_one = catalogue.MagnitudeMeasure(\n event=first_event, agency=agency_one,\n origin=origin, scale='mL', value=5.0)\n self.session.add(measure_one)\n\n measure_two = catalogue.MagnitudeMeasure(\n event=second_event, agency=agency_two,\n origin=origin, scale='mb', value=6.0)\n self.session.add(measure_two)\n\n def test_available_measures_agencies(self):\n self.create_test_fixture()\n\n self.assertEqual(set(['Tatooine', 'Alderaan', 'DeathStar']),\n self.catalogue.get_agencies())\n\n def test_available_measures_scales(self):\n self.create_test_fixture()\n\n self.assertEqual(set(['mL', 'mb']),\n self.catalogue.get_measure_scales())\n\n def test_get_summary(self):\n self.create_test_fixture()\n\n self.assertEqual({catalogue.CatalogueDatabase.MEASURE_AGENCIES:\n set(['Tatooine', 'Alderaan', 'DeathStar']),\n catalogue.CatalogueDatabase.MEASURE_SCALES:\n set(['mL', 'mb'])},\n self.catalogue.get_summary())\n","repo_name":"gvallarelli/oq-eqcatalogue-tool","sub_path":"tests/test_db.py","file_name":"test_db.py","file_ext":"py","file_size_in_byte":5762,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"40245616640","text":"import miscnn.cli.data_exploration as data_exp\nimport unittest\nimport tempfile\nimport os\nimport nibabel as nib\nimport numpy as np\nimport pandas as pd\nfrom miscnn import Data_IO\nfrom miscnn.data_loading.interfaces import NIFTI_interface\n\n\nclass MockArgParser():\n def __init__(self):\n self.default = \"\"\n self.shortName = []\n self.name = []\n \n def add_argument(self, shortname, name, *args, **kwargs):\n self.shortName.append(shortname)\n self.name.append(name)\n \n def set_defaults(self, *args, **kwargs):\n self.default = kwargs[\"which\"]\n\nclass ArgData():\n def __init__(self, data_dir, imagetype):\n self.data_dir = data_dir\n self.imagetype = imagetype\n\ndef write_sample(sample_data, path, name):\n if not os.path.exists(path):\n raise IOError(\n \"Data path, {}, could not be resolved\".format(path)\n )\n # Save segmentation to disk\n sample_path = os.path.join(path, name)\n os.mkdir(sample_path)\n nifti = nib.Nifti1Image(sample_data[0], None)\n nib.save(nifti, os.path.join(sample_path, \"imaging.nii.gz\"))\n nifti = nib.Nifti1Image(sample_data[1], None)\n nib.save(nifti, os.path.join(sample_path, \"segmentation.nii.gz\"))\n\n\nclass cliTEST(unittest.TestCase):\n @classmethod\n def setUpClass(self):\n # Create imgaging and segmentation data set\n np.random.seed(1234)\n self.dataset = dict()\n for i in range(0, 10):\n img = np.random.rand(16, 16, 16) * 256\n self.img = img.astype(int)\n seg = np.random.rand(16, 16, 16) * 3\n self.seg = seg.astype(int)\n sample = (self.img, self.seg)\n self.dataset[\"TEST.sample_\" + str(i)] = sample\n # Initialize Dictionary IO Interface\n # Initialize temporary directory\n self.tmp_dir = tempfile.TemporaryDirectory(prefix=\"tmp.miscnn.\")\n self.tmp_data = os.path.join(self.tmp_dir.name, \"data\")\n os.mkdir(self.tmp_data)\n \n for key, value in self.dataset.items():\n write_sample(value, self.tmp_data, key)\n \n self.dataio = Data_IO(NIFTI_interface(), self.tmp_data)\n \n #perhaps this should be a test but the other things kind of depend on this data\n \n #generate sample dir\n\n # Delete all temporary files\n @classmethod\n def tearDownClass(self):\n self.tmp_dir.cleanup()\n\n #-------------------------------------------------#\n # Base Functionality #\n #-------------------------------------------------#\n # Class Creation\n def test_checkRegistration(self):\n mockParser = MockArgParser()\n data_exp.register_commands(mockParser)\n \n assert mockParser.default == \"data_exp\"\n assert len(mockParser.shortName) == len(mockParser.name)\n assert len(mockParser.shortName) > 0\n \n \n # Class Creation\n def test_CLI_setup(self):\n args = ArgData(self.tmp_data, \"Unknown\")\n \n self.context_data = data_exp.setup_execution(args)\n \n assert isinstance(self.context_data[\"dataio\"], Data_IO)\n \n assert self.context_data[\"cnt\"] == 10\n #verifies the scans work\n assert self.context_data[\"cnt\"] == len(self.context_data[\"indices\"])\n assert self.context_data[\"cnt\"] == len(self.context_data[\"images\"])\n assert self.context_data[\"cnt\"] == len(self.context_data[\"segmentations\"])\n \n self.assertCountEqual(self.context_data[\"indices\"], self.dataset.keys())\n self.assertCountEqual(self.context_data[\"images\"], self.dataset.keys())\n self.assertCountEqual(self.context_data[\"segmentations\"], self.dataset.keys())\n \n \n def test_CLIclassAnalysis(self):\n result = data_exp.execute_class_analysis({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"segmentations\":self.dataset.keys()})\n assert len(result) <= 3\n assert len(result) > 0\n \n def test_CLIstructureAnalysis(self):\n dataframe = pd.DataFrame({\"name\":[]})\n dataframe = data_exp.execute_structure_analysis({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"indices\":self.dataset.keys()}, dataframe)\n \n expectation = pd.DataFrame.from_dict({id: [name, (16, 16, 16, 1), [1, 1, 1]] for id, name in enumerate(self.dataset.keys())}, orient=\"index\",columns=[\"name\", \"shape\", \"voxel_spacing\"])\n \n pd.testing.assert_frame_equal(dataframe,expectation,check_names=False)\n \n def test_CLIminmaxAnalysis(self):\n df, minmax_data = data_exp.execute_minmax_analysis({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"indices\":self.dataset.keys()}, pd.DataFrame({\"name\":[]}))\n \n assert (df[\"minimum\"] < df[\"maximum\"]).all()\n assert (df[\"minimum\"] >= minmax_data[\"glob_min\"]).all()\n \n assert (df[\"maximum\"] <= minmax_data[\"glob_max\"]).all()\n \n def test_CLIminaxSegAnalysis(self):\n df, minmax_data = data_exp.execute_minmax_seg_analysis({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"segmentations\":self.dataset.keys()}, pd.DataFrame({\"name\":[]}))\n \n for cl, minmax in minmax_data.items():\n assert (df[\"minimum_c\" + str(cl)] < df[\"maximum_c\" + str(cl)]).all()\n assert (df[\"minimum_c\" + str(cl)] >= minmax[\"glob_min\"]).all()\n \n assert (df[\"maximum_c\" + str(cl)] <= minmax[\"glob_max\"]).all()\n \n def test_CLIratioAnalysis(self):\n df = data_exp.execute_ratio_analysis({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"indices\":self.dataset.keys()}, pd.DataFrame({\"name\":[]}))\n \n assert df[\"class_frequency\"].apply(lambda x: len(x) == 3).all\n \n def test_CLIbinningAnalysis(self):\n bins = data_exp.execute_binning({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"indices\":self.dataset.keys(), \"images\":self.dataset.keys(), \"segmentations\":self.dataset.keys()}, pd.DataFrame({\"name\":[]}), 5)\n \n def test_CLIbinningSegAnalysis(self):\n bins = data_exp.execute_binning_seg({\"data_dir\": self.tmp_data, \"dataio\":self.dataio, \"segmentations\":self.dataset.keys()}, pd.DataFrame({\"name\":[]}), 5)\n ","repo_name":"frankkramer-lab/MIScnn","sub_path":"tests/test_cli.py","file_name":"test_cli.py","file_ext":"py","file_size_in_byte":6259,"program_lang":"python","lang":"en","doc_type":"code","stars":382,"dataset":"github-code","pt":"81"}
+{"seq_id":"1217345536","text":"from contextInfo.context import Context\nimport cv2\nimport numpy as np\nimport json\nimport copy\nimport matplotlib\nmatplotlib.use('TkAgg')\nimport matplotlib.pyplot as plt\ndef rectify_video(image, map1, map2):\n frame_rectified = cv2.remap(image, map1, map2, interpolation=cv2.INTER_LINEAR, borderMode=cv2.BORDER_CONSTANT)\n return frame_rectified\n\n\ndef calculate_para(K, D, width, height):\n # 优化内参数和畸变系数\n p = cv2.fisheye.estimateNewCameraMatrixForUndistortRectify(K, D, (width, height), None)\n # 此处计算花费时间较大,需从循环中抽取出来\n map1, map2 = cv2.fisheye.initUndistortRectifyMap(K, D, None, p, (width, height), cv2.CV_32F)\n return map1, map2\n\ndef current_coord(detect_result,results):\n pen_coords = [item for item in detect_result if item[\"name\"] == \"pen\"]\n screen_coords = [item for item in detect_result if item[\"name\"] == \"screen\"]\n sorted_pen = sorted(pen_coords, key=lambda x: x.get(\"confidence\"),reverse=True)\n sorted_coord_screen = sorted(screen_coords, key=lambda x: x.get(\"confidence\"), reverse=True)\n\n screen_rect = []\n pen_rect = []\n pen_point = [-1,-1]\n if sorted_coord_screen:\n screen_x1 = int(sorted_coord_screen[0][\"xmin\"])-10\n screen_y1 = int(sorted_coord_screen[0][\"ymin\"])-10\n screen_x2 = int(sorted_coord_screen[0][\"xmax\"])+5\n screen_y2 = int(sorted_coord_screen[0][\"ymax\"])+5\n screen_rect = [screen_x1,screen_y1,screen_x2,screen_y2]\n\n if sorted_pen:\n # print(\"sort_pen = \", sorted_pen)\n # results.show() # 显示标注\n pen_x1 = int(sorted_pen[0][\"xmin\"])\n pen_y1 = int(sorted_pen[0][\"ymin\"])\n pen_x2 = int(sorted_pen[0][\"xmax\"])\n pen_y2 = int(sorted_pen[0][\"ymax\"])\n pen_point = [pen_x2,pen_y2]\n pen_rect = [pen_x1,pen_y1,pen_x2,pen_y2]\n return pen_point,pen_rect,screen_rect\n\nclass edgeDetectBase(object):\n\n def __init__(self):\n self.SN = \"002\"\n self.area_per_min = 0\n self.box_all = []\n self.count_keep_max=3\n self.dist_threshold =3\n self.count_keep = 0\n self.area_per_change_threshold = 0.005\n pass\n\n def edge_handle(self,img,screen_rect):\n # 屏幕 截取\n if screen_rect:\n mask = np.zeros(img.shape[:2], dtype=\"uint8\")\n cv2.rectangle(mask, (screen_rect[0], screen_rect[1]), (screen_rect[2], screen_rect[3]), 255, -1)\n\n screen_masked = cv2.bitwise_and(img, img, mask=mask)\n # cv2.imshow(\"rectangular mask\", masked)\n else:\n screen_masked = copy.deepcopy(img)\n\n img_handle = self.color_select(screen_masked)\n\n seg_img = np.zeros_like(img)\n img_copy = copy.deepcopy(img)\n # 转换为灰度图像\n gray = cv2.cvtColor(img_handle, cv2.COLOR_BGR2GRAY)\n\n gray_copy = copy.deepcopy(gray)\n\n # 对灰度图形态学处理 返回 mask\n mask, closed_mask, mask_er = self.egde_detect_biology(img,gray_copy)\n\n gray_max, gray_mask_max, gray_mask_max_bool = self.find_mask_max(closed_mask, gray_copy)\n\n gray, gray_mask = self.rect_mask_filter(gray_mask_max, gray_max)\n\n kernel_5x5 = np.ones((5, 5), np.uint8)\n gray_mask_ode = cv2.erode(gray_mask_max, kernel_5x5, iterations=50)\n # cv2.imshow(\"closed_mask\",gray_mask_ode)\n # 遮挡后,最大区域为外边框\n screen_contours = self.findcountours(gray_mask_ode)\n # screen_contours_old.append(screen_contours)\n\n #外接矩形\n rect = cv2.minAreaRect(screen_contours[0])\n # rect_all.append(rect)\n box = cv2.boxPoints(rect)\n box = np.int64(box)\n #\n self.box_all.append(box) #没有经过排序\n #\n cv2.drawContours(img_copy, [box], -1, (255, 0, 0), 2)\n # cv2.imshow(\"xx\",img_copy)\n #\n area = cv2.contourArea(box)\n area_contour = cv2.contourArea(screen_contours[0])\n area_per = area_contour/area#面积比例\n\n # # 边框迭代优化\n # # -------面积比变大暂存---------\n # if (self.area_per_min < area_per):\n # screen_contours_best_tmp = screen_contours\n # area_per_min_tmp = area_per\n # box_tmp = box\n # # -------面积比变大暂存---------\n # #1. area_per_min > area_per_cur #留下 area_per_min = area_per_cur\n # # 2.角点变化 #新SN area_per_min = area_per_cur\n # if len(self.box_all)>self.count_keep_max:\n #\n # box_past = self.box_all[-2] #上一次识别框外接矩形\n #\n # #判断识别框是否长时间不变\n # M_past = cv2.moments(box_past)\n # cx_past = int(M_past['m10'] / M_past['m00'])\n # cy_past = int(M_past['m01'] / M_past['m00'])\n # point_past = np.array([cx_past, cy_past])\n # M = cv2.moments(box)\n # cx = int(M['m10'] / M['m00'])\n # cy = int(M['m01'] / M['m00'])\n # point = np.array([cx, cy])\n # dist = np.linalg.norm(point_past-point) #重心变化\n # #-------重心保持---------\n # if (dist < self.dist_threshold):\n # self.count_keep += 1\n # else:\n # self.count_keep = 0\n # # -------重心保持---------\n #\n # # -------重心保持暂存---------\n # if (self.count_keep >= self.count_keep_max):\n # screen_contours_best_tmp = screen_contours\n # box_tmp = box\n # area_per_min_tmp = area_per\n # # -------重心保持---------\n #\n # # -------处理初始帧-----\n # else:\n # screen_contours_best = screen_contours_best_tmp\n # area_per_min = area_per_min_tmp\n # box_best = box_tmp\n # # -------处理初始三帧-----\n #\n # # -------重心保持覆盖--------- 条件:面积比变小程度不大,重心变化不大,并且保持了一段时间\n # if (dist < self.dist_threshold) and (self.count_keep >= self.count_keep_max) and ((area_per_min - area_per) < area_per_change_threshold):\n #\n # screen_contours_best = screen_contours_best_tmp\n # box_best = box_tmp\n # area_per_min = area_per_min_tmp\n # count_keep = 0\n # # -------重心保持覆盖---------\n #\n # area_per_list.append(area_per)\n # area_per_min_list.append(area_per_min)\n #\n # #box排序,便于后面显示可用区域\n # box_best_xs = [i[0] for i in box_best]\n # box_best_ys = [i[1] for i in box_best]\n # box_best_xs.sort()\n # box_best_ys.sort()\n # box_best = np.array([[box_best_xs[1],box_best_ys[1]],[box_best_xs[2],box_best_ys[1]],\n # [box_best_xs[2],box_best_ys[2]],[box_best_xs[1],box_best_ys[2]]])\n #\n # #最比例最大的(找到最好的框)----> 框长时间不变,开启角点检测 ----> 角点变化,当做新的SN检测,但是保留轨迹\n # screen_contours_new.append(screen_contours_best)\n # box_all_new.append(box_best) #经过排序后\n #\n # ##是否需要新的外接矩形\n # # 创建新的图像并绘制分割轮廓\n # img_test = np.zeros_like(img_copy)\n # cv2.drawContours(img_test, screen_contours_best, -1, (0, 0, 255), 2)\n # cv2.drawContours(img_copy, screen_contours_best, -1, (0, 0, 255), 2)\n # # cv2.drawContours(img_copy, [box_best], -1, (255, 0, 0), 2)\n # # gray_copy[closed_mask == 255] = 255\n #\n # #内接矩形\n # rect_in = order_points(screen_contours_best[0].reshape(screen_contours_best[0].shape[0], 2))\n # rect_in = np.int64(rect_in)\n # xs_in = [i[0] for i in rect_in]\n # ys_in = [i[1] for i in rect_in]\n # xs_in.sort()\n # ys_in.sort()\n # # 内接矩形的坐标为\n # # print(xs_in[1], xs_in[2], ys_in[1], ys_in[2])\n # box_in = np.array([[xs_in[1],ys_in[1]],[xs_in[2],ys_in[1]],\n # [xs_in[2],ys_in[2]],[xs_in[1],ys_in[2]]])\n # box_in_buffer = np.array([[xs_in[1]+buffer+buffer_in_right,ys_in[1]+buffer],[xs_in[2]-buffer,ys_in[1]+buffer],\n # [xs_in[2]-buffer,ys_in[2]-buffer],[xs_in[1]+buffer+buffer_in_right,ys_in[2]-buffer]])\n # # cv2.drawContours(img_copy, [box_in], -1, (0, 255, 255), 2)\n # x_inbuffer_left, x_inbuffer_right, y_inbuffer_up, y_inbuffer_down = box_in_buffer[0][0], box_in_buffer[1][0], \\\n # box_in_buffer[1][1], box_in_buffer[2][1]\n # #有效区域\n # box_buffer = np.array([[box_best[0][0] - buffer, box_best[0][1] - buffer], [box_best[1][0] + buffer+ buffer_right, box_best[1][1] - buffer],\n # [box_best[2][0] + buffer + buffer_right, box_best[2][1] + buffer], [box_best[3][0] - buffer, box_best[3][1] + buffer]])\n # xbuffer_left, xbuffer_right, ybuffer_up, ybuffer_down = box_buffer[0][0], box_buffer[1][0], \\\n # box_buffer[1][1], box_buffer[2][1]\n #\n # rectangle_out_mask = np.zeros(img.shape[0:2], dtype=\"uint8\")\n # cv2.rectangle(rectangle_out_mask, tuple(box_buffer[0]), tuple(box_buffer[2]), 255, -1)\n # rectangle_in_mask = np.zeros(img.shape[0:2], dtype=\"uint8\")\n # cv2.rectangle(rectangle_in_mask, tuple(box_in_buffer[0]), tuple(box_in_buffer[2]), 255, -1)\n # buffer_area = rectangle_out_mask-rectangle_in_mask\n # buffer_area_bool = (buffer_area==255)\n # area_zeros1 = 255-np.zeros_like(buffer_area)\n # area_zeros1[buffer_area_bool] = 0\n # area_zeros2 = 255-np.zeros_like(buffer_area)\n # area_zeros2[buffer_area_bool] = 0\n # area_zeros3 = 255 - np.zeros_like(buffer_area)\n # buffer_area_img = np.dstack((area_zeros1, area_zeros3, area_zeros2))\n\n @staticmethod\n def color_select(img):\n \"\"\"\n 阈值过滤 过滤掉多余颜色像素\n :param img:\n :return:\n \"\"\"\n blur = cv2.blur(img, (5, 5))\n hsv = cv2.cvtColor(blur, cv2.COLOR_BGR2HSV)\n\n # low_blue = np.array([100,100,100])\n # high_blue = np.array([180,180,180])\n low_blue = np.array([10, 2, 10])\n high_blue = np.array([220, 32, 230])\n # 可实现二值化功能(类似threshold()函数)可以同时针对多通道进行操作\n mask = cv2.inRange(hsv, low_blue, high_blue)\n res = cv2.bitwise_and(img, img, mask=mask)\n return res\n\n @staticmethod\n def egde_detect_biology(img,gray):\n \"\"\"\n 形态学处理\n :param gray: 灰度图\n :return: maskq\n \"\"\"\n # 进行阈值分割\n\n ret, thresh = cv2.threshold(gray, 0, 255, cv2.THRESH_BINARY_INV+cv2.THRESH_OTSU)\n # thresh = cv2.adaptiveThreshold(gray, 255, adaptiveMethod=cv2.ADAPTIVE_THRESH_MEAN_C,\n # thresholdType=cv2.THRESH_BINARY_INV, blockSize=int(1080 / 3 - 1), C=20)\n # thresh = cv2.adaptiveThreshold(gray, 255, adaptiveMethod=cv2.ADAPTIVE_THRESH_MEAN_C,\n # thresholdType=cv2.THRESH_BINARY_INV, blockSize=9, C=5)\n\n cv2.imshow(\"closed_mask\", thresh)\n # 进行形态学操作,去除噪声和小区域\n kernel = np.ones((3, 3), np.uint8)\n opening = cv2.morphologyEx(thresh, cv2.MORPH_OPEN, kernel, iterations=2) # 开运算\n # 进行连通区域分析\n cv2.imshow(\"closed_mask2\", opening)\n\n # canny边缘\n canny_edge = cv2.Canny(opening,128,255)\n\n cv2.imshow(\"canny\",canny_edge)\n\n # 找出轮廓\n # contours, hierarchy = cv2.findContours(opening, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_SIMPLE)\n #\n # for cnt in contours:\n # if len(cnt)>4:\n # # rect = cv2.minAreaRect(cnt)\n # # box = cv2.boxPoints(rect)\n # # box = np.int0(box)q\n # # cv2.drawContours(img,[box],0,(0,0,255),2)\n # epsilon = 0.001 * cv2.arcLength(cnt, True)\n # approx = cv2.approxPolyDP(cnt, epsilon, True)\n # cv2.drawContours(img, approx, -1, (0, 255, 0), 5)\n # cv2.polylines(img, [approx], True, (0, 0, 255), 2)\n\n lines = cv2.HoughLinesP(canny_edge, 1, np.pi / 180, 100, minLineLength=120, maxLineGap=50)\n\n # 笔在哪一边时候,那边的边不确定\n up = []\n down = []\n left = []\n right = []\n coords = []\n for i,line in enumerate(lines):\n x1, y1, x2, y2 = line[0]\n if abs(y2-y1)>abs(x2-x1):\n # y轴方向 y轴变化大\n if x1<960 and x2<960:\n coord = \"left\"\n left.append(line[0])\n elif x1>960 and x2>960:\n coord = \"right\"\n right.append(line[0])\n else:\n coord = \"\"\n else:\n # x轴方向 x轴变化大\n if y1<540 and y2<540:\n coord = \"up\"\n up.append(line[0])\n elif y1>540 and y2>540:\n coord = \"down\"\n down.append(line[0])\n else:\n coord = \"\"\n coords.append(coord)\n\n\n print(coords)\n for line in lines:\n print(\"lines = \",line)\n # 获取坐标\n x1, y1, x2, y2 = line[0]\n cv2.line(img, (x1, y1), (x2, y2), (0, 0, 255), thickness=5)\n cv2.imshow(\"imag\", img)\n\n\n\n\n\n n, labels, stats, centroids = cv2.connectedComponentsWithStats(opening)\n area_label_list = list(np.where((stats[:, 4] < 1000000) & (stats[:, 4] > 100000))[0])\n # #找到最大连通区域\n # max_area = 0\n # max_label = 0\n # for i in range(0, n):\n # if stats[i, cv2.CC_STAT_AREA] > max_area:\n # max_area = stats[i, cv2.CC_STAT_AREA]\n # max_label = i\n # stats[np.argsort(-stats[:, 4])][1:3, :]\n # # 提取最大连通区域\n # mask = np.zeros_like(gray)\n # mask[labels == max_label] = 255\n\n # 提取最大连通区域\n mask = np.zeros_like(gray)\n for area_label in area_label_list:\n mask[labels == area_label] = 255\n cv2.imshow(\"mask\", mask)\n # 进行形态学操作,填充区域\n kernel_5x5 = np.ones((5, 5), np.uint8)\n kernel_3x3 = np.ones((3, 3), np.uint8)\n mask_er = cv2.dilate(mask, kernel_5x5, iterations=50)\n closed_mask = cv2.morphologyEx(mask_er, cv2.MORPH_CLOSE, kernel_5x5)\n\n # # 进行轮廓检测\n # contours, hierarchy = cv2.findContours(closed_mask, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n return mask, closed_mask, mask_er\n @staticmethod\n def getLinearEquation(p1x, p1y, p2x, p2y):\n sign = 1\n a = p2y - p1y\n if a < 0:\n sign = -1\n a = sign * a\n b = sign * (p1x - p2x)\n c = sign * (p1y * p2x - p1x * p2y)\n return [a, b, c]\n\n\n # mask最大区域\n @staticmethod\n def find_mask_max(closed_mask, gray):\n gray_copy = copy.deepcopy(gray)\n y, x = gray_copy.shape\n img_center_x = np.int32(x / 2)\n img_center_y = np.int32(y / 2)\n img_center_point = np.array([img_center_x, img_center_y])\n n, labels, stats, centroids = cv2.connectedComponentsWithStats(255 - closed_mask, connectivity=4)\n centroids_list = centroids.astype(np.int32)\n diff_dist = np.expand_dims(np.linalg.norm(img_center_point - centroids_list, axis=1), axis=1)\n diff_dist_min = diff_dist[1:].min() # 距离最小\n max_label = np.where(diff_dist == diff_dist_min)[0][0]\n # max_label = np.where(stats[:,4]==stats[1:,4].max())[0][0]\n max_label = max_label.astype(np.int32)\n gray_mask_bool = ~(labels == max_label)\n gray_mask = np.zeros_like(gray)\n gray_mask[gray_mask_bool] = 255\n gray_copy[gray_mask_bool] = 255\n return gray_copy, gray_mask, gray_mask_bool\n\n # 过滤矩形外区域\n @staticmethod\n def rect_mask_filter(gray_mask, gray):\n gray_mask_copy = copy.deepcopy(gray_mask)\n gray_copy = copy.deepcopy(gray)\n mask_num = np.sum(gray_mask == 0, 1)\n di = np.diff(mask_num)\n di = np.append(np.zeros(1, int), di)\n y_up = np.where(di == di.max())[0][0]\n y_down = np.where(di == di.min())[0][0]\n gray_copy[:y_up + 1, :] = 255\n gray_copy[y_down:, :] = 255\n gray_mask_copy[:y_up + 1, :] = 255\n gray_mask_copy[y_down:, :] = 255\n # plt.figure(1)\n # plt.subplot(1,2,1)\n # plt.plot(list(range(len(mask_num))),mask_num)\n # plt.subplot(1,2,2)\n # plt.plot(list(range(len(di))),di)\n # plt.show()\n return gray_copy, gray_mask_copy\n @staticmethod\n def findcountours(gray_mask_ode):\n contours, hierarchy = cv2.findContours(255 - gray_mask_ode, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n screen_contours = []\n for contour in contours:\n area = cv2.contourArea(contour)\n x, y, w, h = cv2.boundingRect(contour)\n aspect_ratio = float(w) / h\n if aspect_ratio > 1 and aspect_ratio < 5:\n screen_contours.append(contour)\n return tuple(screen_contours)\n def drop_context(self):\n pass\n\ndef main_logic(scan_info,model):\n # 根据扫描信息判断电视版信息\n # 根据扫描信息判断板子是否为同一个,为同一个则不清空信息\n # 不同的则清空给上下文信息\n # 打开视频流\n # 获取边框\n # 获取检测\n # 路径分析判断\n # 结果控制设备\n # 开始时候灯由熄灭状态 -> 黄色\n # 操作完成 黄色 -> 绿色 -> 熄灭\n\n # video = cv2.VideoCapture(0)\n # video = cv2.VideoCapture(\"./data/20230228_4mm.mp4\")\n video = cv2.VideoCapture(\"./data/0314_left.mp4\")\n ID = Context.getID()\n print(\"ID = \",ID)\n i = 0\n\n Context.change(scan_info)\n\n ID = Context.getID()\n print(\"ID new = \", ID)\n K = np.array(\n [[1201.3967181633418, 0.0, 909.7424436183744], [0.0, 1203.635467250557, 534.1590658991514], [0.0, 0.0, 1.0]])\n D = np.array([[-0.0978537375125563], [-0.03841501213366177], [-0.03612764818273854], [0.05276041355808103]])\n width, height = 1920, 1080\n # width, height = int(video.get(cv2.CAP_PROP_FRAME_WIDTH)), int(video.get(cv2.CAP_PROP_FRAME_HEIGHT))\n # print(width, height)\n map1,map2 = calculate_para(K,D,width,height)\n edgeDetect = edgeDetectBase()\n while (video.isOpened()):\n \"\"\"\n 0、矫正\n 1、获取检测框,检测框缓存\n 2、获取检测点\n 3、判断检测区域\n 4、决策逻辑 \n 5、设备控制\n \"\"\"\n success, image = video.read()\n if not success:\n break\n\n # cv2.imwrite(\"./data2/{}.jpg\".format(i), image)\n i = i + 1\n\n # 图片矫正\n frame = rectify_video(image, map1, map2)\n\n # frame = cv2.resize(frame, (960, 540))\n # cv2.imshow('frame', frame)\n\n # 边框检测\n\n # 算法模型\n results = model(cv2.cvtColor(frame, cv2.COLOR_BGR2RGB))\n detect_result = json.loads(results.pandas().xyxy[0].to_json(orient=\"records\"))\n # for box in detect_result:\n # l, t, r, b = int(box[\"xmin\"]),int(box[\"ymin\"]),int(box[\"xmax\"]),int(box[\"ymax\"])\n # confidence = round(box[\"confidence\"],1)\n # cls_name = box[\"name\"]\n # # outside = t - h >= 3\n # if cls_name==\"pen\" or cls_name==\"screen\":\n # cv2.rectangle(frame, (l, t), (r, b), (0, 255, 0), 2)\n # cv2.putText(frame, cls_name + \"-\" + str(confidence), (l, t), cv2.FONT_ITALIC, 1, (255, 0, 0), 2)\n # cv2.imshow(\"ff\",frame)\n\n # 当前笔的坐标,笔的位置,屏幕位置\n pen_point,pen_rect,screen_rect = current_coord(detect_result,results)\n\n # 边框提取\n edgeDetect.edge_handle(frame,screen_rect)\n\n #\n\n\n\n\n if cv2.waitKey(0) & 0xFF == ord('q'):\n break\n # Context.add(i)\n # if i%2==1:\n # Context.sub(i)\n # if i>100:\n # break\n\n # 决策分析\n\n\n video.release()\n\n cv2.destroyAllWindows()\n return {\"i\":i,\"ID\":ID}\n\nif __name__==\"__main__\":\n import torch\n\n model = torch.hub.load(\"D:\\project\\demo\", \"custom\", 'D:\\project\\demo\\model_save\\ybest_screen', source=\"local\")\n model.conf = 0.6 # NMS confidence threshold\n model.iou = 0.2\n main_logic(1,model)\n\n # A082\n","repo_name":"Zhengzhuo0309/segment_best_TCL","sub_path":"handle_pic.py","file_name":"handle_pic.py","file_ext":"py","file_size_in_byte":20911,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21001122617","text":"import tkinter as tk\nfrom tkinter import ttk\n\nfrom db.sql_db import DatabaseService\n\nimport interface.get_bmi_screen as BMI\nfrom interface.entry_with_placeholder import EntryWithPlaceholder\n\n\nclass RecordBMIDisplay(tk.Frame):\n def init(self, master, *args, **kwargs):\n super().__init__(master, *args, **kwargs)\n self.db = DatabaseService()\n\n self.name_entry = EntryWithPlaceholder(self, \"Name\")\n self.name_entry.grid(row=0, column=0, padx=5, pady=5)\n self.name_entry.bind(\"\", lambda e: self.record_bmi(event=e))\n\n self.age_entry = EntryWithPlaceholder(self, \"Age\")\n self.age_entry.grid(row=1, column=0, padx=5, pady=5)\n self.age_entry.bind(\"\", lambda e: self.record_bmi(event=e))\n\n self.grade_section_entry = EntryWithPlaceholder(\n self, \"Grade & Section (Ex.: 12-Faraday)\")\n self.grade_section_entry.grid(row=2, column=0, padx=5, pady=(5, 20))\n self.grade_section_entry.bind(\n \"\", lambda e: self.record_bmi(event=e))\n\n self.lrn_entry = EntryWithPlaceholder(self, \"LRN\")\n self.lrn_entry.grid(row=3, column=0, padx=5, pady=(5, 20))\n self.lrn_entry.bind(\"\", lambda e: self.record_bmi(event=e))\n\n enter_button = ttk.Button(\n self, text=\"Enter\", command=self.record_bmi)\n enter_button.grid(row=3, column=0, padx=5, pady=(5, 30))\n\n self.label_error = ttk.Label(self, foreground='red')\n self.label_error.grid(row=10, column=0, sticky=tk.S, padx=5)\n\n def record_bmi(self, event=\"\"):\n if BMI.user_info == \"\" or BMI.user_info == None:\n print(\"Invalid user info\")\n return\n\n print(BMI.user_info)\n self.label_error[\"text\"] = \"\"\n\n name = self.name_entry.get().strip().upper()\n age = self.age_entry.get().strip()\n lrn = self.lrn_entry.get().strip()\n grade_section = self.grade_section_entry.get().strip().upper()\n\n print(name, age, lrn, grade_section)\n\n if (name == \"Name\" or name == \"\" or age == \"Age\" or age == \"\" or lrn == \"LRN\" or lrn == \"\" or grade_section == \"Grade & Section (Ex.: 12-Faraday)\" or grade_section == \"\"):\n self.label_error[\"text\"] = \"Please fill all entry fields.\"\n return\n\n BMI.user_info.name = name\n BMI.user_info.age = age\n BMI.user_info.lrn = lrn\n BMI.user_info.grade, BMI.user_info.section = grade_section.split(\"-\")\n\n self.db.insert_user(BMI.user_info)\n\n self.label_error = ttk.Label(\n self, text=\"DATA SAVED\", foreground='green')\n self.label_error.grid(row=9, column=0, sticky=tk.S, padx=5)\n","repo_name":"zatrodev/bmi-calculating-gui","sub_path":"interface/record_bmi_screen.py","file_name":"record_bmi_screen.py","file_ext":"py","file_size_in_byte":2669,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21643131464","text":"#!/usr/bin/python\n\nimport sqlite3\nimport json\n\n# SQLite DB Name\nDB_Name = \"../db.sqlite3\"\n\ndef Sensor(jsonData):\n\t#Parse Data \n\tjson_Dict = json.loads(jsonData)\n\tSensorID = json_Dict['Sensor_ID']\n\tData_and_Time = json_Dict['Time']\n\tTemperature = json_Dict['Temperature']\n\tHumidity = json_Dict['Humidity']\n\t#Push into DB Table\n\tconn = sqlite3.connect(DB_Name)\n\tconn.execute(\"INSERT INTO sensor_sensor (SensorID,Date_and_Time, Temperature,Humidity) \\\n\t VALUES (?,?,?,?)\",[SensorID,Data_and_Time, Temperature,Humidity])\n\tconn.commit()\n\tprint (\"Sensor created new value\")\n\tconn.close()\n\n\n","repo_name":"nam2297ptit/DoAnThietKeHeThongNhung","sub_path":"DHT11/Get_Data_to_DB.py","file_name":"Get_Data_to_DB.py","file_ext":"py","file_size_in_byte":590,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12558186570","text":"# @Time : 2018/10/16 0016 16:01\n# @Author : lzc\n# @File : 猫眼数据处理.py\nimport time\n\nimport requests\nurl = 'https://box.maoyan.com/promovie/api/box/second.json'\n#从url地址获取数据\ndef getfilmdata():\n res = requests.get(url)\n\n #判断是否获取数据成功\n if res.status_code==200:\n json_data = res.json()\n else:\n print('获取数据失败')\n\n if json_data: #从电影票房数据(json对象)中获取所有电影票房集合信息\n films = json_data.get('data').get('list')\n for item in films:\n film = {}\n film['影片名'] = item.get('movieName')\n film['上映天数'] = item.get('releaseInfo')\n film['总票房'] = item.get('movieName')\n film['综合票房'] = item.get('sumBoxInfo')\n film['票房占比'] = item.get('boxInfo')\n film['排片场次'] = item.get('boxRate')\n film['排片占比'] = item.get('avgShowView')\n film['场均人次'] = item.get('avgSeatView')\n print(film)\n\nwhile True:\n getfilmdata()\n time.sleep(3)","repo_name":"liuzhongchuang/demo","sub_path":"爬虫数据抓取的方式/猫眼数据处理.py","file_name":"猫眼数据处理.py","file_ext":"py","file_size_in_byte":1114,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"7981146010","text":"import cx_Oracle\nimport pymongo\nimport os\n#mongodb ? ????\ndbname=\"petrol_data\"\n#??oracle ???\nis_oracle=True\n#mongodb ?? ????????\nrawcollection=\"petrol_raw\" \nmongourl=\"mongodb://localhost:27017/\" #mongodb ???\nos.environ['NLS_LANG'] = 'SIMPLIFIED CHINESE_CHINA.UTF8'\n\nlinkstr = \"puguang/puguang@192.168.43.22:1521/ORCL\" #oracle ??????\n'''\noracle connection string\n'''\nsqlstr = '''SELECT JH, JT, \n QXWJM, ZBMC, KC, CJXMID, CYJG, YXJDDJSD1,\n YXJDDJSD2, BZ, LRSJ, VRUSERNAME, VRUNITNAME,\n COLLECT_JOB_S,GL_UUID, DATASTATUS, CJLX,\n ORDERTIME, UPLOADFLAG\n from WL18'''\n#Oracle to mongodb mapping\nfield_raw = [\n 'JH', 'JT', 'QXWJM',\n 'ZBMC', 'KC', 'CJXMID',\n 'CYJG', 'YXJDDJSD1', 'YXJDDJSD2',\n 'BZ', 'LRSJ', 'VRUSERNAME', 'VRUNITNAME',\n 'COLLECT_JOB_S', 'GL_UUID', 'DATASTATUS', 'CJLX',\n 'ORDERTIME', 'UPLOADFLAG'\n]\ncollectionames= ['crud_dict', 'crud_field', 'crud_table', 'crud_user', 'model_config', \n'oil_field_config', 'task', 'users']\n#create crud meta schema\ndef create_schema(mongourl, dbname, collectionames):\n myclient = pymongo.MongoClient(mongourl)\n db = myclient[dbname]\n if len(db.list_collection_names()) != 0:\n [db.create_collection(item) for item in collectionames]\n print(\"在mongodb 中为 node.js 主服务 创建数据集合\")\n print(db.list_collection_names())\n#??????mongodb??\ndef transfer_data(linkstr, mongourl, dbname, field_list, is_oracle, datarepo, sqlstr):\n myclient = pymongo.MongoClient(mongourl)\n db = myclient[dbname]\n dbcols = db[datarepo]\n sys_conn = None\n if is_oracle:\n sys_conn = cx_Oracle.connect(linkstr)\n else:\n return\n cursor = sys_conn.cursor()\n cursor.execute(sqlstr.encode('utf-8'))\n result = cursor.fetchall()\n for row in result:\n dictionary = dict(zip(field_list,list(row)))\n dbcols.insert_one(dictionary)\n\nif __name__ == \"__main__\":\n create_schema(mongourl, dbname, collectionames)\n transfer_data(linkstr, mongourl, dbname, field_raw, is_oracle, rawcollection, sqlstr)","repo_name":"mini-tiger/loggingchartplus","sub_path":"python/taskutil/mongocreate.py","file_name":"mongocreate.py","file_ext":"py","file_size_in_byte":2049,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"10976146289","text":"from django import forms\nfrom django.core.exceptions import ValidationError\nfrom django.forms import widgets\n\nfrom .models import *\n\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Layout, Div, Field\n\nclass AddPhoto(forms.ModelForm):\n existing_images = forms.ModelChoiceField(queryset=Photo.objects.all(), empty_label=None)\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n class Meta:\n model = Photo\n fields = ['photo']\n\n\nclass AddFilm(forms.ModelForm):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n class Meta:\n model = Films\n fields = ['name', 'content', 'big_photo', 'trailerURL', 'type1', 'type2', 'type3']\n widgets = {\n 'content': forms.Textarea(attrs={'class': 'form-control', 'rows': 3}),\n\n }\n\n\nclass AddSEO(forms.ModelForm):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n class Meta:\n model = SEO\n fields = ['url', 'title', 'keywords', 'description']\n widgets = {\n 'description': forms.Textarea(attrs={'class': 'form-control', 'rows': 3}),\n }\n\n\nclass AddCinema(forms.ModelForm):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n\n class Meta:\n model = Cinemas\n fields = ['name', 'content', 'logo', 'conditions', 'photo_banner']\n widgets = {\n 'content': forms.Textarea(attrs={'class': 'form-control', 'rows': 3}),\n 'conditions': forms.Textarea(attrs={'class': 'form-control', 'rows': 3}),\n }\n\nclass AddHall(forms.ModelForm):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n\n class Meta:\n model = Halls\n fields = ['number', 'content', 'schema_hall', 'photo_banner']\n widgets = {\n 'content': forms.Textarea(attrs={'class': 'form-control', 'rows': 3}),\n }","repo_name":"romankylik/KinoCMS","sub_path":"cinemas/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1974,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"41161782220","text":"import subprocess\nimport json\nfrom random import randint\nimport tweepy\nfrom mastodon import Mastodon\nimport datetime\nimport sqlite3\nimport requests\n\nclass photoDB:\n def __init__(self, db_name):\n self.con = sqlite3.connect(db_name)\n self.cur = self.con.cursor()\n\n def update_database(self, data):\n self.cur.execute('''UPDATE photos\n SET posted_date = ?\n WHERE\n id = ?\n '''\n , data\n )\n self.con.commit()\n\n def get_random_row(self, collection):\n d = self.cur.execute('''\n SELECT record FROM photos\n WHERE collection=? AND\n posted_date IS NULL AND\n dont_post IS NOT 1 AND\n invalid_record IS NOT 1\n ''',\n [collection])\n\n records = d.fetchall()\n\n # if there are no records found\n if len(records) == 0:\n return -1\n else:\n return records[randint(0,len(records)-1)][0]\n\n def update_row_status(self, date, id_, dont_post):\n self.cur.execute('''\n UPDATE photos\n SET posted_date=?, dont_post=?\n WHERE id=?\n ''',\n [date, dont_post, id_])\n\n self.con.commit()\n\n# break up description into valid length parts\ndef post_parts(desc, length):\n desc_parts = []\n\n # desc = 'Description: ' + desc\n\n if len(desc) <= length:\n desc_parts = [desc]\n\n else:\n l_idx = 0\n r_idx = length - 4\n finished = False\n while finished == False:\n # finish if we've reached the end of the string\n if r_idx >= len(desc)-1:\n finished = True\n desc_parts.append(desc[l_idx:])\n else:\n r_idx = desc[l_idx:r_idx].rfind(' ') + l_idx\n desc_parts.append(desc[l_idx:r_idx] + ' ...')\n\n l_idx = r_idx + 1\n r_idx += length - 4\n\n return desc_parts\n\n# randomly choose index of collection given weights\ndef choose_collection(weights):\n sum_w = 0\n for i in weights: sum_w+= i\n\n # find out which set r is in\n r = randint(1,sum_w)\n for i in range(0, len(weights)):\n sum_ = 0\n for j in range(0,i+1): sum_+=weights[j]\n if r <= sum_:\n return(i)\n\ndef get_metadata(url):\n \n # get json \n t = json.loads(requests.get(url).text)\n\n # if this if from streetcar collection:\n if 'msn' in url:\n try:\n keys = list(t['response']['document'].keys())\n metadata = {}\n # if invalid record\n except:\n return {}\n \n # get title\n metadata.update({'title':t['response']['document']['title_ssi']})\n metadata.update({'permis':'Minnesota Streetcar Museum'})\n metadata.update({'id':t['response']['document']['id'].split(':')[1]})\n \n # get description\n if 'description_ts' in keys:\n metadata.update({'descri':t['response']['document']['description_ts']})\n #else:\n # metadata.update({'descri':''})\n \n if 'dat_tesi' in keys:\n metadata.update({'year':t['response']['document']['dat_tesi']})\n \n if 'city_ssim' in keys:\n cities = t['response']['document']['city_ssim']\n if 'Minneapolis' in cities:\n metadata.update({'city':'Minneapolis'})\n else:\n metadata.update({'city':t['response']['document']['city_ssim'][0]})\n \n #m.update{'title':t['response']['document']['title_ssi']}\n \n # university of minnesota archives\n elif 'p16022coll175' in url:\n try:\n keys = list(t.keys())\n metadata = {}\n # if invalid record\n except:\n return {}\n \n metadata.update({'title':t['title']})\n metadata.update({'permis':t['contributing_organization_name']})\n metadata.update({'id':t['id'].split(':')[1]})\n \n # get description\n if 'description' in keys:\n metadata.update({'descri':t['description']})\n \n if 'date_created' in keys:\n metadata.update({'year':t['date_created'][0]})\n \n if 'city' in keys:\n cities = t['city']\n # initialize\n metadata.update({'city':t['city'][0]})\n for city in cities:\n # format typically like 'Minneapolis; St Paul'\n if 'minneapolis' in city.lower():\n metadata.update({'city':'Minneapolis'})\n\n # else if this is hclib photo\n else:\n try:\n metadata = t\n \n fields_to_remove = []\n for field in t:\n if metadata[field] == {}:\n fields_to_remove.append(field)\n for field in fields_to_remove:\n metadata.pop(field)\n \n # manually add this field in as 'id' for compatibility purposes\n metadata.update({'id': metadata['dmrecord']})\n \n except Exception as e:\n print('failure to get metadata')\n print(e)\n print(metadata)\n return {}\n\n return metadata\n\ndef get_photo(url, out_image):\n\n cmd = 'wget --user-agent=\"Mozilla\" ' + url + ' --output-file=/dev/null -O ' + out_image\n\n proc = subprocess.Popen(cmd, shell=True)\n proc.wait()\n\n # check if photo doesn't exist from source and return 1 if it doesn't\n ls = subprocess.check_output('ls -lt ' + out_image, shell=True).decode('utf-8')\n\n # return false if photo not created for some reason\n try:\n size = ls.split(' ')[4]\n except:\n return False\n\n if size == '0':\n return False\n else:\n return True\n\n# return true if bad word in title of description string\ndef bad_word_in_post(title, descr, subj, input_file):\n\n bad_word_list = []\n with open(input_file) as f:\n for line in f:\n bad_word = line.strip('\\n')\n bad_word_list.append(bad_word) \n\n for word in bad_word_list:\n if word in title.lower() or word in descr.lower() or word in subj.lower():\n return True\n return False\n\ndef get_api_keys(input_file):\n keys = {}\n with open(input_file) as f:\n for line in f:\n l = line.strip('\\n')\n l = l.split(',')\n keys.update({l[0]:l[1]})\n\n return keys\n\ndef create_send_post(collection, photo_id):\n\n # connect to twitter\n keys = get_api_keys('api_keys.txt')\n auth = tweepy.OAuthHandler(keys['api_key'], keys['api_key_secret'])\n auth.set_access_token(keys['access_token'], keys['access_token_secret'])\n api = tweepy.API(auth)\n\n # connect to mastodon\n mast_key = get_api_keys('api_keys_mastodon.txt')\n mastodon = Mastodon(\n access_token = mast_key['access_token_secret'],\n api_base_url = 'https://botsin.space/'\n )\n\n # images we'll be pulling\n if collection == 'msn':\n full_url = 'https://cdm16022.contentdm.oclc.org/digital/iiif/msn/' + str(photo_id) + '/full/2000,2000/0/default.jpg'\n metadata_url = 'https://collection.mndigital.org//catalog/msn:' + str(photo_id) + '.json'\n\n out_image = 'images/' + collection + photo_id + '.jpg'\n \n # umn archives \n elif collection == 'p16022coll175':\n full_url = 'https://cdm16022.contentdm.oclc.org/digital/iiif/p16022coll175/' + str(photo_id) + '/full/2000,2000/0/default.jpg'\n metadata_url = 'https://umedia.lib.umn.edu/item/p16022coll175:' + str(photo_id) + '.json'\n\n out_image = 'images/' + collection + photo_id + '.jpg' \n \n # if hclib collection:\n else:\n base_url = 'https://digitalcollections.hclib.org/'\n full_url = base_url + 'digital/download/collection/' + collection + '/id/' + str(photo_id) + '/size/large'\n metadata_url = base_url + 'digital/bl/dmwebservices/index.php?q=dmGetItemInfo/' + collection + '/' + str(photo_id) + '/json'\n\n out_image = 'images/' + collection + photo_id + '.jpg'\n\n # try to create photo and get metadata\n photo_created = get_photo(full_url, out_image)\n metadata = get_metadata(metadata_url)\n len_metadata = len(metadata)\n\n # return false if photo or metadata weren't retrieved or \n # index of metadata doesn't match intended value\n if photo_created == True and len_metadata > 1 and str(metadata['id']) == str(photo_id):\n \n metadata_keys = list(metadata.keys())\n title = metadata['title']\n\n # get date of tweet, if exists\n if 'year' in metadata_keys:\n date = metadata['year']\n elif 'decade' in metadata_keys:\n date = metadata['decade']\n\n # get attribution\n perm_exists = False\n if 'permis' in metadata_keys:\n\n # assuming normal format for hclb\n if ':' in metadata['permis']:\n try:\n source = metadata['permis'].split(':')[1]\n source = source.strip(' ').strip('\"').strip(\"'\")\n perm_exists = True\n\n # if the format isn't as expected:\n except:\n perm_exists = False\n source = ''\n #source = 'Hennepin County Library'\n\n # different format for glanton\n if 'glanton' in metadata['permis'].lower():\n source = 'Hennepin County Library and the children of John Glanton'\n \n if 'Streetcar Museum' in metadata['permis']:\n source = metadata['permis']\n perm_exists = True\n \n if 'University Archives' in metadata['permis']:\n source = metadata['permis']\n perm_exists = True \n \n # if the permissions say you need to contact them, don't post\n if 'viewed' in metadata['permis'] and 'specialcoll@hclib.org' in metadata['permis']:\n source = ''\n perm_exists = False\n\n # if there is no permissions field \n else:\n perm_exists = True\n source = 'Hennepin County Library'\n\n # make main tweet\n tweet1 = title\n\n if 'year' in metadata_keys or 'decade' in metadata_keys:\n tweet1 += ' (' + date + ')'\n\n if 'addres' in metadata_keys:\n tweet1 += '\\n' + metadata['addres']\n\n tweet1 += '\\nSource: ' + source\n \n # trim long tweets\n # tweet1 = tweet1[0:280]\n\n if 'descri' in metadata_keys:\n description = metadata['descri']\n tweet1 += '\\n\\n' + description\n else:\n description = ''\n\n if 'subjec' in metadata_keys:\n subject = metadata['subjec']\n else:\n subject = ''\n \n print(tweet1)\n\n\n # check that the photo was taken in minneapolis\n # if there's no city field, assume it was in mpls\n in_mpls = False\n if 'city' in metadata_keys:\n city = metadata['city']\n cities = ['minneapolis','saint anthony and minneapolis', 'saint anthony', 'richfield', 'hopkins', 'saint louis park', 'st. louis park', 'robbinsdale', 'fort snelling', 'golden valley', 'columbia heights']\n if city.lower() in cities:\n in_mpls = True\n else:\n in_mpls = False\n\n # assume in minneapolis is minneapolis mentioned. this is\n # redundant for now\n elif 'minneapolis' in title.lower() or 'minneapolis' in description.lower() or 'minneapolis' in subject.lower():\n in_mpls = True\n\n # assume it's in minneapolis if no city provided\n else:\n in_mpls = True\n\n # check for offensive content\n # post if non-offensive and there are permissions\n dont_post = bad_word_in_post(title, description, subject, 'bad_words.txt')\n if dont_post == False and perm_exists == True and in_mpls == True:\n # don't tweet anymore\n # print('sending tweet')\n # status = api.update_with_media(out_image, tweet1)\n\n print('sending toot')\n toot_len = 500\n post_text = post_parts(tweet1, toot_len)\n mast_media = mastodon.media_post(out_image)\n toot = mastodon.status_post(post_text[0], media_ids=mast_media)\n \n prev_toot_id = toot.id\n for i in range(1, len(post_text)):\n # don't tweet anymore\n # tweet thread\n # reply = api.update_status(status=d, \n # in_reply_to_status_id=prev_id, \n # auto_populate_reply_metadata=True)\n # prev_id = reply.id\n\n # mastodon thread\n mast_reply = mastodon.status_post(status=post_text[i], in_reply_to_id=prev_toot_id)\n prev_toot_id = mast_reply.id\n\n return True\n\n # if there's a filtered word in the post\n else:\n print('bad word: ' + str(dont_post))\n print('permission to post: ' + str(perm_exists))\n print('in mpls: ' + str(in_mpls))\n return False\n\n # failed if couldn't get photo, metadata, or metadata id doesn't match\n else:\n print(' failed')\n return False\n\n# main loop for sending posts\n# try to send until we successfully get an image\nif __name__ == '__main__':\n\n time = datetime.datetime.now()\n\n # coll18 is really old photos, coll1 is glanton photos, coll 175 is university archives\n collections = ['CPED', 'MplsPhotos', 'FloydKelley', 'MPRB', 'p17208coll18', 'p17208coll1', 'msn', 'p17208coll15', 'p16022coll175']\n max_idx = [21250, 60000, 212, 251, 1100, 820, 2776, 1406, 21899]\n weights = [20, 13, 1, 1, 5, 3, 10, 5, 10]\n\n # open connection to photo database\n db = photoDB('photoDB.db')\n\n sum_weights = 0\n for i in weights: sum_weights+=i\n\n # try until a photo is found and posted\n tries = 0\n posted = False\n while posted == False and tries < 10:\n\n # randomly choose collection based on weights given\n coll = choose_collection(weights)\n\n # randomly choose photo in collection\n # if = -1, then there are no records left\n photo_idx = db.get_random_row(collections[coll])\n \n #photo_idx = randint(1,max_idx[coll]) \n if int(photo_idx) != -1:\n posted = create_send_post(collections[coll], str(photo_idx))\n\n # update database with whether this was posted or not\n if posted == False:\n db.update_row_status(time.strftime('%d/%m/%y %H:%M:%S'), collections[coll] + '_' + str(photo_idx), 1)\n else:\n db.update_row_status(time.strftime('%d/%m/%y %H:%M:%S'), collections[coll] + '_' + str(photo_idx), 0)\n\n # write to log\n f = open('post_log.txt','a')\n f.write(time.strftime('%d/%m/%y %H:%M:%S') + ',' + collections[coll] + ',' + str(photo_idx) + '\\n')\n f.close()\n\n tries += 1\n\n db.con.close()\n","repo_name":"sampenders/photo_bot","sub_path":"send_tweet.py","file_name":"send_tweet.py","file_ext":"py","file_size_in_byte":15132,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"9690218684","text":"\ndb=\"\"\"\n11767,Polaris\n7588,Achernar\n21421,Aldebaran\n24436,Rigel\n24608,Capella\n27989,Betelgeuse\n30438,Canopus\n32349,Sirius\n37279,Procyon\n37826,Pollux\n49669,Regulus\n60718,Acrux\n62434,Mimosa\n65474,Spica\n68702,Hadar\n69673,Arcturus\n71683,Rigil Kent\n80763,Antares\n91262,Vega\n97649,Altair\n102098,Daneb\n113368,Fomalhaut\n\"\"\"\n\nimport numpy as np\nimport ephem\nimport stars as hip\nimport planisfunc as plf\n\noffset = np.pi /100.\n\nclass hip20_nameE:\n def __init__(self, observe, ax, zo, legend):\n\n nameE = {}\n s = []\n for line in db.split('\\n'):\n try:\n hipno = line.split(',')[0]\n name = line.split(',')[1]\n nameE[hipno]=name\n starz = hip.star(str(hipno))\n s.append(starz)\n\n except:\n None\n [s[i].compute(observe) for i in range(len(s))]\n\n NAME = np.array([body.name for body in s])\n ALT = np.array([float(body.alt) for body in s])\n AZ = np.array([float(body.az) for body in s])\n\n lim = np.pi /12.\n\n X, Y, ret = plf.polarXY(AZ, ALT, lim, name=NAME)\n NAME = ret['name']\n\n X += offset\n Y += offset\n\n for n in range(len(NAME)):\n name = nameE[NAME[n]]\n\n ax.text(X[n], Y[n], name, color='white',\n fontsize=10, alpha=0.8, zorder=zo )\n\nif __name__ == '__main__':\n import matplotlib.pyplot as plt\n import sky\n from observe import *\n\n fig, ax = plt.subplots()\n\n ax.set_aspect('equal')\n sf = sky.skyfield(obs, ax, zo=0, legend=True)\n na = hip20_nameE(obs, ax, zo=1, legend=True)\n\n plt.xlim(- np.pi/2., np.pi/2.)\n plt.ylim(- np.pi/2., np.pi/2.)\n\n ax.axis('off')\n plt.show()\n","repo_name":"IchiroYoshida/python_public","sub_path":"planis/nameE.py","file_name":"nameE.py","file_ext":"py","file_size_in_byte":1758,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"29968033254","text":"import os\nimport pytest\n\n\nfrom qcore.utils import load_sim_params as mocked_load_sim_params\n\nfrom workflow.automation.tests.test_common_set_up import get_fault_from_rel, set_up\n\nimport workflow.automation.submit.submit_hf\n\n\n@pytest.mark.usefixtures(\"init_scheduler\")\ndef test_main(set_up, mocker):\n \"\"\"No return value. Just check that it runs without crashing\"\"\"\n\n mocker.patch(\n \"workflow.automation.submit.submit_hf.est.est_HF_chours_single\",\n lambda *args, **kwargs: (2, 0.05, 40),\n )\n\n for root_path, realisation in set_up:\n rel_dir = os.path.join(\n root_path, \"CSRoot\", \"Runs\", get_fault_from_rel(realisation), realisation\n )\n # Fault will probably change on each set of data, so reset this every time\n mocker.patch(\n \"workflow.automation.submit.submit_hf.utils.load_sim_params\",\n lambda x: mocked_load_sim_params(os.path.join(rel_dir, x)),\n )\n\n workflow.automation.submit.submit_hf.main(\n submit=None,\n machine=\"default\",\n ncores=80,\n rel_dir=rel_dir,\n retries=0,\n seed=None,\n version=None,\n write_directory=None,\n )\n","repo_name":"ucgmsim/slurm_gm_workflow","sub_path":"workflow/automation/tests/test_submit_scripts/test_submit_hf.py","file_name":"test_submit_hf.py","file_ext":"py","file_size_in_byte":1219,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1314373843","text":"# pylint: disable=protected-access\n\"\"\"Testing script\"\"\"\nimport pytest\nfrom PIL import ImageChops, Image\nimport sys\nimport os\nimport os.path\nfrom hypothesis import given\nfrom hypothesis.strategies import text, binary, characters\nfrom shutil import copy2\n\nsys.path.append(os.path.join(os.path.dirname(__file__), '..'))\n# noinspection PyPep8\nfrom steganographer.steganographer import Steganographer\n# noinspection PyPep8\nfrom steganographer.steganographer import _unpack_image, _pack_image, _open_bin_file, _write_bin_file, \\\n _open_image_file, _write_image_file\n\nCLEAN_PNG_LOCATION = \"tests/cleanImage.png\"\n\n\ndef test_generate_header():\n \"\"\"The header is generated as expected\"\"\"\n stegs = Steganographer()\n header = bytes(stegs._header._HEADER_TITLE, 'utf-8') + \\\n bytes(stegs._header.data_len.to_bytes(stegs._header._HEADER_DATA_SIZE, \"little\")) + \\\n bytes(stegs._header.bits_used.to_bytes(stegs._header._HEADER_BITS_SIZE, \"little\")) + \\\n bytes(stegs._header.file_name_len.to_bytes(stegs._header._HEADER_FILE_LENGTH_SIZE, \"little\"))\n\n assert header == stegs._generate_header(stegs._header.data_len, stegs._header.bits_used, \"\")\n\n\ndef test_retrieve_header():\n \"\"\"The header is retrieved as expected\"\"\"\n stegs = Steganographer()\n test_message = \"12345\".encode('utf-8')\n test_data_len = len(test_message)\n test_bits_used = 1\n test_file_name = \"test_retrieve_header.txt\"\n test_file_name_len = len(test_file_name)\n test_data = bytes(b'\\x01' * 1000)\n\n test_header = stegs._generate_header(test_data_len, test_bits_used, test_file_name)\n hidden_data = stegs._hide_data(test_data[:stegs._header.header_length * stegs._BYTELEN], test_header)\n hidden_data += stegs._hide_data(test_data[stegs._header.header_length * stegs._BYTELEN:], test_message)\n header_retrieved = stegs._retrieve_header(hidden_data)\n\n assert header_retrieved is True\n assert stegs._header.data_len == test_data_len\n assert stegs._header.bits_used == test_bits_used\n assert stegs._header.file_name_len == test_file_name_len\n assert stegs._header.file_name.decode('utf-8') == test_file_name\n\n\ndef test_hide_byte():\n \"\"\"The _hide_byte function does hide a byte and returns the test_data with that byte hidden.\"\"\"\n stegs = Steganographer()\n test_data = bytes(b'\\x01' * stegs._BYTELEN)\n data_to_hide = bytes('A', 'utf-8')\n solution_data = bytearray(stegs._BYTELEN)\n solution_data[1] = 1\n solution_data[7] = 1\n\n assert stegs._hide_byte(test_data, data_to_hide[0]) == solution_data\n\n\ndef test_reveal_byte():\n \"\"\"The _reveal_byte function returns a bytes object of the hidden byte.\"\"\"\n stegs = Steganographer()\n test_data = bytearray(stegs._BYTELEN)\n test_data[1] = 1\n test_data[7] = 1\n solution_data = bytes('A', 'utf-8')\n\n assert stegs._reveal_byte(test_data) == solution_data\n\n\n@given(data_to_hide=binary(min_size=1, max_size=1))\ndef test_hide_reveal_byte_inverse(data_to_hide):\n \"\"\"Anything hidden by _hide_byte is revealed by _reveal_byte.\"\"\"\n clean_data = bytes(b'\\x01' * 8)\n\n stegs = Steganographer()\n revealed_byte = stegs._reveal_byte(stegs._hide_byte(clean_data, data_to_hide[0]))\n assert revealed_byte == data_to_hide\n\n\ndef test_hide_string():\n \"\"\"Takes in a string and a bytes object and hides the string in that bytes object.\"\"\"\n stegs = Steganographer()\n test_data = bytes(b'\\x01' * stegs._BYTELEN * 3)\n solution_data = bytearray(stegs._BYTELEN * 3)\n solution_data[1] = 1\n solution_data[7] = 1\n solution_data[9] = 1\n solution_data[14] = 1\n solution_data[17] = 1\n solution_data[22] = 1\n solution_data[23] = 1\n\n assert stegs._hide_string(test_data, 'ABC') == solution_data\n\n\ndef test_reveal_string():\n \"\"\"Returns a string representation of the data that was hidden in the test_data.\"\"\"\n solution = 'ABC'\n stegs = Steganographer()\n stegs._header.data_len = len(solution)\n test_data = bytearray(stegs._BYTELEN * 4)\n test_data[1] = 1\n test_data[7] = 1\n test_data[9] = 1\n test_data[14] = 1\n test_data[17] = 1\n test_data[22] = 1\n test_data[23] = 1\n\n assert stegs._reveal_string(test_data) == solution\n\n\n@given(string_to_hide=text(characters(min_codepoint=1, blacklist_categories=('Cc', 'Cs'))))\ndef test_hide_reveal_string_inverse(string_to_hide):\n \"\"\"Anything hidden by _hide_string is revealed by _reveal_string.\"\"\"\n clean_data = bytes(b'\\x01' * 5000)\n stegs = Steganographer()\n stegs._header.data_len = len(string_to_hide.encode('utf-8'))\n revealed_string = stegs._reveal_string(stegs._hide_string(clean_data, string_to_hide))\n assert revealed_string == string_to_hide\n\n\ndef test_hide_data():\n \"\"\"Will hide one bytes object inside another.\"\"\"\n stegs = Steganographer()\n test_data = bytes(b'\\x01' * stegs._BYTELEN * 4)\n data_to_hide = bytes('ABC', 'utf-8')\n stegs._header.data_len = len(data_to_hide)\n solution_data = bytearray(stegs._BYTELEN * 4)\n solution_data[1] = 1\n solution_data[7] = 1\n solution_data[9] = 1\n solution_data[14] = 1\n solution_data[17] = 1\n solution_data[22] = 1\n solution_data[23] = 1\n solution_data[24:] = b'\\x01' * stegs._BYTELEN\n\n assert stegs._hide_data(test_data, data_to_hide) == solution_data\n\n\ndef test_hide_data_partial():\n \"\"\"Will work when given a bytes object that is too short to contain the full data to be hidden.\"\"\"\n stegs = Steganographer()\n test_data = bytes(b'\\x01' * stegs._BYTELEN * 3)\n data_to_hide = bytes('ABC', 'utf-8')\n stegs._header.data_len = len(data_to_hide)\n solution_data = bytearray(stegs._BYTELEN * 3)\n solution_data[1] = 1\n solution_data[7] = 1\n solution_data[9] = 1\n solution_data[14] = 1\n solution_data[17] = 1\n solution_data[22] = 1\n solution_data[23] = 1\n\n # Testing when only half a byte is passed in for the data that contains the hidden text.\n assert stegs._hide_data(test_data[:4], data_to_hide) == solution_data[:4]\n\n\ndef test_reveal_data():\n \"\"\"Will return the correct data that is hidden inside the test_data.\"\"\"\n solution_data = bytes('ABC', 'utf-8')\n stegs = Steganographer()\n stegs._header.data_len = len(solution_data)\n test_data = bytearray(stegs._BYTELEN * 3)\n test_data[1] = 1\n test_data[7] = 1\n test_data[9] = 1\n test_data[14] = 1\n test_data[17] = 1\n test_data[22] = 1\n test_data[23] = 1\n\n assert stegs._reveal_data(test_data) == solution_data\n\n\ndef test_reveal_data_partial():\n \"\"\"\n Will return as much data as possible.\n\n When the container bytes object passed in is too small for all the data to be hidden.\n \"\"\"\n stegs = Steganographer()\n solution_data = bytes('AB@', 'utf-8')\n test_data = bytearray(stegs._BYTELEN * 3) # Will contain 'ABC' but will be truncated when passed to _reveal_data\n test_data[1] = 1\n test_data[7] = 1\n test_data[9] = 1\n test_data[14] = 1\n test_data[17] = 1\n test_data[22] = 1\n test_data[23] = 1\n stegs._data_len = len('ABC')\n\n assert stegs._reveal_data(test_data[:-stegs._BYTELEN // 2]) == solution_data\n\n\n@given(string_to_hide=text(characters(min_codepoint=1, blacklist_categories=('Cc', 'Cs'))))\ndef test_hide_reveal_data_inverse(string_to_hide):\n \"\"\"Anything hidden by _hide_data is revealed by _reveal_data.\"\"\"\n clean_data = bytes(b'\\x01' * 5000)\n data_to_hide = bytes(string_to_hide, 'utf-8')\n\n stegs = Steganographer()\n stegs._header.data_len = len(string_to_hide.encode('utf-8'))\n revealed_data = stegs._reveal_data(stegs._hide_data(clean_data, data_to_hide))\n\n assert revealed_data == data_to_hide\n\n\ndef test_exact_data_with_string_inverse():\n \"\"\"The string entered is the string returned. The storing data is the exact length needed.\"\"\"\n test_string = \"This is a test String\"\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n blank_data = bytes(b'\\x01' * len(test_string) * stegs._BYTELEN)\n\n revealed_string = stegs._reveal_string(stegs._hide_string(blank_data, test_string))\n\n assert test_string == revealed_string\n\n\ndef test_exact_data_with_data_inverse():\n \"\"\"The data entered is the data returned. The storing data is the exact length needed.\"\"\"\n test_string = \"This is a test String\"\n test_data = bytes(test_string, 'utf-8')\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n blank_data = bytes(b'\\x01' * len(test_string) * stegs._BYTELEN)\n\n revealed_data = stegs._reveal_data(stegs._hide_data(blank_data, test_data))\n\n assert test_data == revealed_data\n\n\ndef test_short_data_with_string_inverse():\n \"\"\"When the data is too small, by a full byte, everything that can be returned is returned.\"\"\"\n test_string = \"This is a test String\"\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n blank_data = bytes(b'\\x01' * (len(test_string) * stegs._BYTELEN - stegs._BYTELEN))\n\n revealed_string = stegs._reveal_string(stegs._hide_string(blank_data, test_string))\n\n assert test_string[:-1] == revealed_string\n\n\ndef test_short_data_with_data_inverse():\n \"\"\"When the data is too small, by a full byte, everything that can be returned is returned.\"\"\"\n test_string = \"This is a test String\"\n test_data = bytes(test_string, 'utf-8')\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n blank_data = bytes(b'\\x01' * (len(test_string) * stegs._BYTELEN - stegs._BYTELEN))\n\n revealed_data = stegs._reveal_data(stegs._hide_data(blank_data, test_data))\n\n assert test_data[:-1] == revealed_data\n\n\ndef test_short_partial_data_string_inverse():\n \"\"\"When the data is too small, by a half byte, everything that can be returned is returned.\"\"\"\n test_string = \"This is a test String\"\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n solution_string = test_string[:-1] + chr(ord(test_string[-1]) >> stegs._BYTELEN // 2 << stegs._BYTELEN // 2)\n blank_data = bytes(b'\\x01' * (len(test_string) * stegs._BYTELEN - stegs._BYTELEN // 2))\n\n revealed_string = stegs._reveal_string(stegs._hide_string(blank_data, test_string))\n\n assert solution_string == revealed_string\n\n\ndef test_short_partial_data_w_data_inverse():\n \"\"\"When the data is too small, by a half byte, everything that can be returned is returned.\"\"\"\n test_string = \"This is a test String\"\n test_data = bytes(test_string, 'utf-8')\n solution_data = bytearray(test_data)\n stegs = Steganographer()\n stegs._header.data_len = len(test_string)\n solution_data[-1] = solution_data[-1] >> stegs._BYTELEN // 2 << stegs._BYTELEN // 2\n blank_data = bytes(b'\\x01' * (len(test_string) * stegs._BYTELEN - stegs._BYTELEN // 2))\n\n revealed_data = stegs._reveal_data(stegs._hide_data(blank_data, test_data))\n\n assert solution_data == revealed_data\n\n\ndef test_unpack_image():\n \"\"\"Unpacking returns a bytes object full of all the pixels flattened in one dimension.\"\"\"\n pixel = 1, 2, 3, 4\n pixel_length = len(pixel)\n solution_pixels = bytes(list(pixel * pixel_length))\n test_pixels = []\n\n for _ in range(pixel_length):\n test_pixels.append(pixel)\n\n unpacked = _unpack_image(test_pixels)\n\n assert unpacked[0] == len(pixel)\n assert unpacked[1] == solution_pixels\n\n\ndef test_pack_image():\n \"\"\"Packing returns a list with tuples of length 4.\"\"\"\n pixel = 1, 2, 3, 4\n pixel_length = len(pixel)\n test_pixels = pixel_length, list(pixel * pixel_length)\n solution_pixels = []\n\n for _ in range(pixel_length):\n solution_pixels.append(pixel)\n\n packed = _pack_image(test_pixels)\n\n assert packed == solution_pixels\n\n\ndef test_unpack_pack_inverse():\n \"\"\"Pixels unpacked by _unpack_image are correctly packed by _pack_image.\"\"\"\n pixel = 1, 2, 3, 4\n test_pixels = []\n\n for _ in range(4):\n test_pixels.append(pixel)\n\n assert _pack_image(_unpack_image(test_pixels)) == test_pixels\n\n\ndef test_open_bin_file():\n \"\"\"Opening a file works.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n file_data = _open_bin_file(clean_file)\n\n with open(clean_file, 'rb') as file:\n assert file_data == file.read()\n\n with pytest.raises(SystemExit):\n _open_bin_file(\"OpenBinFileThatDoesNotExist.nope\")\n\n\ndef test_write_bin_diff_content():\n \"\"\"The file written is different from the one read, after hiding a message.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_bin_file_diff_content.png\"\n\n stegs = Steganographer()\n data = stegs._hide_string(_open_bin_file(clean_file), \"Hidden text from test_write_bin_diff_content.\")\n _write_bin_file(dirty_file, data)\n\n with open(clean_file, 'rb') as clean, open(dirty_file, 'rb') as dirty:\n assert clean.read() != dirty.read()\n\n os.remove(dirty_file)\n\n\ndef test_write_bin_file_size_same():\n \"\"\"The file written is the same size as the one read, after hiding a message.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_bin_file_size_same.png\"\n\n stegs = Steganographer()\n data = stegs._hide_string(_open_bin_file(clean_file), \"Hidden text from test_write_bin_file_size_same.\")\n _write_bin_file(dirty_file, data)\n\n # Getting the file sizes for the clean and dirty files.\n with open(clean_file, 'rb') as clean:\n clean.seek(0, 2)\n clean_file_size = clean.tell()\n\n with open(dirty_file, 'rb') as dirty:\n dirty.seek(0, 2)\n dirty_file_size = dirty.tell()\n\n assert clean_file_size == dirty_file_size\n\n os.remove(dirty_file)\n\n\ndef test_open_image_file():\n \"\"\"Opening an image file returns the data in the file in a one dimensional list.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n image_data = _open_image_file(clean_file)\n\n with Image.open(clean_file) as clean:\n pixels = clean.getdata()\n\n assert image_data[1] == _unpack_image(pixels)[1]\n\n with pytest.raises(SystemExit):\n _open_image_file(\"OpenImageFileThatDoesNotExist.nope\")\n\n\ndef test_write_image_file_valid():\n \"\"\"The image created is not corrupt.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_image_file_valid_image.png\"\n\n stegs = Steganographer()\n clean_data = _open_image_file(clean_file)\n dirty_data = clean_data[0], stegs._hide_string(clean_data[1], \"Hidden text from test_write_image_file_valid.\")\n output_file = _write_image_file(dirty_file, clean_file, dirty_data)\n\n try:\n Image.open(output_file)\n except OSError:\n pytest.fail(\"Image is corrupt \" + output_file)\n\n os.remove(output_file)\n\n\ndef test_write_image_diff_content():\n \"\"\"Writing out an image creates a different image at the bit level.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_image_file_diff_content.png\"\n\n stegs = Steganographer()\n clean_data = _open_image_file(clean_file)\n dirty_data = clean_data[0], stegs._hide_string(clean_data[1], \"Hidden text from test_write_image_diff_content.\")\n output_file = _write_image_file(dirty_file, clean_file, dirty_data)\n\n with open(clean_file, 'rb') as clean, open(output_file, 'rb') as dirty:\n assert clean.read() != dirty.read()\n\n os.remove(output_file)\n\n\ndef compare_images(img1, img2):\n \"\"\"Expects strings of the locations of two images. Will return an integer representing their difference\"\"\"\n with Image.open(img1) as img1, Image.open(img2) as img2:\n # Calculate a difference image that is the difference between the two images.\n diff = ImageChops.difference(img1, img2)\n\n return sum(_unpack_image(diff.getdata())[1])\n\n\ndef test_write_image_same_image():\n \"\"\"Writing out an image creates the same image when viewed generally.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_image_file_same_image.png\"\n\n stegs = Steganographer()\n clean_data = _open_image_file(clean_file)\n dirty_data = clean_data[0], stegs._hide_string(clean_data[1], \"Hidden text from test_write_image_same_image.\")\n output_file = _write_image_file(dirty_file, clean_file, dirty_data)\n\n assert compare_images(clean_file, output_file) < 500\n\n os.remove(output_file)\n\n\ndef test_write_image_diff_size():\n \"\"\"Writing out an image creates a file of a different size, if the file was not generated by PIL originally.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_image_file_diff_size.png\"\n\n stegs = Steganographer()\n clean_data = _open_image_file(clean_file)\n dirty_data = clean_data[0], stegs._hide_string(clean_data[1], \"Hidden text from test_write_image_diff_size.\")\n output_file = _write_image_file(dirty_file, clean_file, dirty_data)\n\n # Getting the file sizes for the clean and dirty files.\n with open(clean_file, 'rb') as clean:\n clean.seek(0, 2)\n clean_file_size = clean.tell()\n\n with open(output_file, 'rb') as dirty:\n dirty.seek(0, 2)\n dirty_file_size = dirty.tell()\n\n assert clean_file_size != dirty_file_size\n\n os.remove(output_file)\n\n\ndef test_write_image_diff_size_pil():\n \"\"\"Writing out an image creates a file of a different size, if the file was generated by PIL originally.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"tests/dirtyImage_test_write_image_file_diff_size_pil.png\"\n clean_file_pil = \"tests/cleanImagePIL.png\"\n with Image.open(clean_file) as pil_image:\n pil_image.save(clean_file_pil)\n\n stegs = Steganographer()\n clean_data = _open_image_file(clean_file_pil)\n dirty_data = clean_data[0], stegs._hide_string(clean_data[1], \"Hidden text from test_write_image_diff_size_pil.\")\n output_file = _write_image_file(dirty_file, clean_file_pil, dirty_data)\n\n # Getting the file sizes for the clean and dirty files.\n with open(clean_file_pil, 'rb') as clean:\n clean.seek(0, 2)\n clean_file_size = clean.tell()\n\n with open(output_file, 'rb') as dirty:\n dirty.seek(0, 2)\n dirty_file_size = dirty.tell()\n\n assert clean_file_size != dirty_file_size\n\n os.remove(output_file)\n\n\ndef test_write_image_exit_on_fail():\n \"\"\"When failing to write an image there is a system exit.\"\"\"\n clean_file = CLEAN_PNG_LOCATION\n dirty_file = \"WriteImageFileThatDoesNotExist.nope\"\n dirty_data = bytes(8)\n\n with pytest.raises(SystemExit):\n _write_image_file(clean_file, dirty_file, dirty_data)\n\n\ndef test_steganographer_hide_string():\n \"\"\"A string will correctly be hidden in a new image.\"\"\"\n clean_image = CLEAN_PNG_LOCATION\n dirty_image = \"tests/dirtyImage_test_steganographer_hide_string.png\"\n hidden_message = \"Hidden text from test_steganographer_hide_string.\"\n\n stegs = Steganographer()\n hidden_fname = stegs.steganographer_hide(clean_image, hidden_message, dirty_image)\n\n with open(clean_image, 'rb') as clean, open(hidden_fname, 'rb') as dirty:\n assert clean.read() != dirty.read()\n assert compare_images(clean_image, hidden_fname) < 500\n try:\n Image.open(hidden_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + hidden_fname)\n\n os.remove(dirty_image)\n\n\ndef test_stegs_hide_string_nonsense():\n \"\"\"A random string, that can cause a decode error, will correctly be hidden in a new image.\"\"\"\n clean_image = CLEAN_PNG_LOCATION\n dirty_image = \"tests/dirtyImage_test_steganographer_hide_string_nonsense.png\"\n hidden_message = \"Ĝ�ĜĜĜĜĜĜĜĜĜԬĜ\\U000fc423ĜĜĜĜԬĜĜĜԬԬĜԬ\\U000fc423ĜԬ\\U000fc423ԬԬĜ\\U000fc423ԬĜԬ�����������\"\n\n stegs = Steganographer()\n hidden_fname = stegs.steganographer_hide(clean_image, hidden_message, dirty_image)\n\n with open(clean_image, 'rb') as clean, open(hidden_fname, 'rb') as dirty:\n assert clean.read() != dirty.read()\n assert compare_images(clean_image, hidden_fname) < 650\n try:\n Image.open(hidden_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + hidden_fname)\n\n os.remove(dirty_image)\n\n\ndef test_steganographer_hide_file():\n \"\"\"A file can be hidden inside of an image and the image created is not corrupt\"\"\"\n clean_image = CLEAN_PNG_LOCATION\n dirty_image = \"tests/dirtyImage_test_steganographer_hide_file.png\"\n file_to_hide = \"tests/FileToHide.zip\"\n\n stegs = Steganographer()\n hidden_fname = stegs.steganographer_hide_file(clean_image, file_to_hide, dirty_image)\n\n with open(clean_image, 'rb') as clean, open(hidden_fname, 'rb') as dirty:\n assert clean.read() != dirty.read()\n assert compare_images(clean_image, hidden_fname) < 19000\n try:\n Image.open(hidden_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + hidden_fname)\n\n os.remove(hidden_fname)\n\n\ndef test_steganographer_reveal_file():\n \"\"\"A file that has been hidden can be revealed.\"\"\"\n original_file = \"tests/FileToHide.zip\"\n dirty_image = \"tests/dirtyImageWFile.png\"\n revealed_file_name = \"tests/test_steganographer_reveal_file.zip\"\n\n stegs = Steganographer()\n revealed_file_data, _ = stegs.steganographer_reveal(dirty_image)\n\n with open(revealed_file_name, 'wb') as rev_file:\n rev_file.write(revealed_file_data)\n\n with open(original_file, 'rb') as original, open(revealed_file_name, 'rb') as revealed:\n assert original.read() == revealed.read()\n\n os.remove(revealed_file_name)\n\n\ndef test_steganographer_hide_name():\n \"\"\"The image a string is hidden in is the correct one.\"\"\"\n clean_image = CLEAN_PNG_LOCATION\n dirty_image = \"tests/dirtyImage_test_steganographer_hide_name.png\"\n hidden_message = \"Hidden text from test_steganographer_hide_name.\"\n\n stegs = Steganographer()\n hidden_fname = stegs.steganographer_hide(clean_image, hidden_message, dirty_image)\n\n assert hidden_fname == dirty_image\n\n os.remove(dirty_image)\n\n\ndef test_steganogrified_name():\n \"\"\"Data will be hidden in a file with steganogrified at the end, when no output file name is provided.\"\"\"\n clean_message_image = copy2(CLEAN_PNG_LOCATION, CLEAN_PNG_LOCATION[:-4] + \"_test_message_steganogrified_name.png\")\n clean_file_image = copy2(CLEAN_PNG_LOCATION, CLEAN_PNG_LOCATION[:-4] + \"_test_file_steganogrified_name.png\")\n hidden_message = \"Hidden text from test_steganogrified_name.\"\n file_to_hide = \"tests/FileToHide.zip\"\n\n stegs = Steganographer()\n hidden_message_fname = stegs.steganographer_hide(clean_message_image, hidden_message)\n steganogrified_message_fname = clean_message_image[:-4] + \"Steganogrified.png\"\n hidden_file_fname = stegs.steganographer_hide_file(clean_file_image, file_to_hide)\n steganogrified_file_fname = clean_file_image[:-4] + \"Steganogrified.png\"\n\n assert hidden_message_fname == steganogrified_message_fname\n assert os.path.isfile(steganogrified_message_fname)\n assert hidden_file_fname == steganogrified_file_fname\n assert os.path.isfile(steganogrified_file_fname)\n\n os.remove(clean_message_image)\n os.remove(hidden_message_fname)\n os.remove(clean_file_image)\n os.remove(hidden_file_fname)\n\n\n@given(hidden_message=text(characters(min_codepoint=1, blacklist_categories=('Cc', 'Cs'))))\ndef test_steganographer_inverse(hidden_message):\n \"\"\"Steganographer_reveal reveals what was hidden by steganographer_hide.\"\"\"\n clean_image = CLEAN_PNG_LOCATION\n dirty_image = \"tests/dirtyImage_test_steganographer_inverse.png\"\n\n stegs = Steganographer()\n revealed_message = stegs.steganographer_reveal(stegs.steganographer_hide(\n clean_image, hidden_message, dirty_image))[0].decode('utf-8')\n assert revealed_message == hidden_message\n\n os.remove(dirty_image)\n\n\ndef test_unicode_inverse():\n \"\"\"Unicode characters are hidden and revealed.\"\"\"\n message = \"test_unicode hidden message. Some random unicode characters: � ᾨ ԅ Թ ػ ޗ ߚ ङ ლ ጩ �\"\n\n stegs = Steganographer()\n assert message == stegs.steganographer_reveal(stegs.steganographer_hide(CLEAN_PNG_LOCATION, message,\n \"tests/dirtyImage.png\"))[0].decode('utf-8')\n\n\ndef test_main_hide_msg_with_output(capfd):\n \"\"\"Command line calls to hide work when given an input image, a message, and an output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_main_hide_msg_with_output hidden message'\n dirty_fname = \"tests/dirtyImage_test_main_hide_msg_with_output.png\"\n\n result = os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -m \"' + hidden_message +\n '\" -o ' + dirty_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == \"The message has been hidden in \" + dirty_fname + line_end\n assert compare_images(CLEAN_PNG_LOCATION, dirty_fname) < 500\n try:\n Image.open(dirty_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + dirty_fname)\n\n os.remove(dirty_fname)\n\n\ndef test_main_hide_file_with_output(capfd):\n \"\"\"Command line calls to hide work when given an input image, a file to hide, and an output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n file_to_hide = \"tests/FileToHide.zip\"\n dirty_fname = \"tests/dirtyImage_test_main_hide_file_with_output.png\"\n\n result = os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -f \"' + file_to_hide +\n '\" -o ' + dirty_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == \"The file \" + file_to_hide + \" has been hidden in \" + dirty_fname + line_end\n assert compare_images(CLEAN_PNG_LOCATION, dirty_fname) < 19000\n try:\n Image.open(dirty_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + dirty_fname)\n\n os.remove(dirty_fname)\n\n\ndef test_main_hide_msg_no_output(capfd):\n \"\"\"Command line calls to hide work when given an input image, a message, and no output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_main_hide_msg_no_output hidden message'\n clean_image = copy2(CLEAN_PNG_LOCATION, CLEAN_PNG_LOCATION[:-4] + \"test_main_hide_msg_no_output.png\")\n steganogrified_fname = clean_image[:-4] + \"Steganogrified.png\"\n\n result = os.system('python -m steganographer ' + clean_image + ' -m \"' + hidden_message + '\"')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == \"The message has been hidden in \" + steganogrified_fname + line_end\n assert compare_images(clean_image, steganogrified_fname) < 500\n try:\n Image.open(steganogrified_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + steganogrified_fname)\n\n os.remove(clean_image)\n os.remove(steganogrified_fname)\n\n\ndef test_main_hide_file_no_output(capfd):\n \"\"\"Command line calls to hide work when given an input image, a file to hide, and no output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n file_to_hide = \"tests/FileToHide.zip\"\n clean_image = copy2(CLEAN_PNG_LOCATION, CLEAN_PNG_LOCATION[:-4] + \"test_main_hide_file_no_output.png\")\n steganogrified_fname = clean_image[:-4] + \"Steganogrified.png\"\n\n result = os.system('python -m steganographer ' + clean_image + ' -f \"' + file_to_hide + '\"')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == \"The file \" + file_to_hide + \" has been hidden in \" + steganogrified_fname + line_end\n assert compare_images(clean_image, steganogrified_fname) < 19000\n try:\n Image.open(steganogrified_fname)\n except OSError:\n pytest.fail(\"Image is corrupt \" + steganogrified_fname)\n\n os.remove(clean_image)\n os.remove(steganogrified_fname)\n\n\ndef test_main_reveal_msg_no_output(capfd):\n \"\"\"Command line calls to reveal work when given an input image, and no output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_main_reveal_msg_no_output hidden message'\n dirty_fname = \"tests/dirtyImage_test_main_reveal_msg_no_output.png\"\n\n os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -m \"' + hidden_message +\n '\" -o ' + dirty_fname)\n _, _ = capfd.readouterr()\n\n result = os.system(\"python -m steganographer \" + dirty_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The hidden message was...\" + line_end + hidden_message + line_end)\n\n os.remove(dirty_fname)\n\n\ndef test_main_reveal_file_no_output(capfd):\n \"\"\"Command line calls to reveal work when given an input image, the reveal file flag, and no output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n file_to_hide = \"tests/FileToHide.zip\"\n dirty_fname = \"tests/dirtyImage_test_main_reveal_file_no_output.png\"\n generated_output_file = file_to_hide\n\n os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -f \"' + file_to_hide +\n '\" -o ' + dirty_fname)\n _, _ = capfd.readouterr()\n\n result = os.system(\"python -m steganographer -r \" + dirty_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The hidden file was revealed in \" + generated_output_file + line_end)\n\n os.remove(dirty_fname)\n\n\ndef test_main_reveal_msg_w_output(capfd):\n \"\"\"Command line calls to reveal work when given an input image, and an output file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_main_reveal_msg_w_output hidden message'\n dirty_fname = \"tests/dirtyImage_test_main_reveal_msg_w_output.png\"\n output_fname = \"tests/outputMessage.txt\"\n\n os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -m \"' + hidden_message +\n '\" -o ' + dirty_fname)\n _, _ = capfd.readouterr()\n\n result = os.system(\"python -m steganographer \" + dirty_fname + \" -o \" + output_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n with open(output_fname, 'r') as output:\n assert output.read() == hidden_message\n\n assert out == (\"The hidden message was written to \" + output_fname + line_end)\n\n os.remove(dirty_fname)\n\n\ndef test_main_reveal_file_w_output(capfd):\n \"\"\"Command line calls to reveal work when given an input image, a reveal file flag, and an output image.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n file_to_hide = \"tests/FileToHide.zip\"\n dirty_fname = \"tests/dirtyImage_test_main_reveal_file_w_output.png\"\n output_fname = \"tests/outputFile.zip\"\n\n os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -f \"' + file_to_hide +\n '\" -o ' + dirty_fname)\n _, _ = capfd.readouterr()\n\n result = os.system(\"python -m steganographer \" + dirty_fname + \" -r -o \" + output_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n with open(output_fname, 'rb') as output, open(file_to_hide, 'rb') as original_file:\n assert output.read() == original_file.read()\n\n assert out == (\"The hidden file was revealed in \" + output_fname + line_end)\n\n os.remove(dirty_fname)\n\n\ndef test_main_reveal_no_msg(capfd):\n \"\"\"There should be an error returned when there is no message hidden in the image file.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n clean_fname = \"tests/cleanImage.jpg\"\n\n result = os.system(\"python -m steganographer \" + clean_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == \"This file %s has no hidden message.\" % clean_fname + line_end\n\n\ndef test_main_reveal_no_op_unicode(capfd):\n \"\"\"Command line calls to reveal work when the hidden message contains high value unicode.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_main_reveal_no_op_unicode hidden message, Unicode characters: � ᾨ ԅ Թ ػ ޗ ߚ ङ ლ ጩ �'\n dirty_fname = \"tests/dirtyImage_test_main_reveal_no_op_unicode.png\"\n output_fname = \"tests/dirtyImage_test_main_reveal_no_op_unicode_message.txt\"\n\n os.system('python -m steganographer ' + CLEAN_PNG_LOCATION + ' -m \"' + hidden_message +\n '\" -o ' + dirty_fname)\n _, _ = capfd.readouterr()\n\n result = os.system(\"python -m steganographer \" + dirty_fname)\n out, _ = capfd.readouterr()\n\n assert result == 0\n if os.path.isfile(output_fname):\n with open(output_fname, 'r', encoding='utf-8') as output:\n assert output.read() == hidden_message\n\n assert (out == (\"The hidden message contains unsupported unicode characters and cannot be fully displayed \" +\n \"here. The correct message has been written to \" + output_fname + line_end +\n str(hidden_message.encode('utf-8')) + line_end)\n or out == (\"The hidden message was...\" + line_end + hidden_message + line_end))\n\n os.remove(dirty_fname)\n if os.path.isfile(output_fname):\n os.remove(output_fname)\n\n\ndef test_jpegs(capfd):\n \"\"\"Jpegs can have a message hidden and revealed. Note they are converted to png.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_jpeg hidden message'\n dirty_fname = \"tests/dirtyImage_test_jpegs\"\n\n result = os.system('python -m steganographer tests/cleanImage.jpg -m \"' + hidden_message +\n '\" -o ' + dirty_fname + '.jpg')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The message has been hidden in \" + dirty_fname + '.png' + line_end)\n\n result = os.system(\"python -m steganographer \" + dirty_fname + '.png')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The hidden message was...\" + line_end + hidden_message + line_end)\n assert compare_images(\"tests/cleanImage.jpg\", dirty_fname + '.png') < 500\n\n\n@pytest.mark.xfail(strict=True, reason=\"Issue #59 bmp support is broken.\", run=True)\ndef test_bmps(capfd):\n \"\"\"Bmps can have a message hidden and revealed.\"\"\"\n line_end = '\\n'\n if sys.platform == 'win32':\n line_end = '\\r\\n'\n hidden_message = 'test_bmps hidden message'\n dirty_fname = \"tests/dirtyImage_test_bmps\"\n\n result = os.system('python -m steganographer tests/cleanImage.bmp -m \"' + hidden_message +\n '\" -o ' + dirty_fname + '.bmp')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The message has been hidden in \" + dirty_fname + '.png' + line_end)\n\n result = os.system(\"python -m steganographer \" + dirty_fname + '.png')\n out, _ = capfd.readouterr()\n\n assert result == 0\n assert out == (\"The hidden message was...\" + line_end + hidden_message + line_end)\n assert compare_images(\"tests/cleanImage.bmp\", dirty_fname + '.png') < 500\n","repo_name":"MotaDan/steganographerPY","sub_path":"tests/test_steganographer.py","file_name":"test_steganographer.py","file_ext":"py","file_size_in_byte":34786,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"73662911946","text":"from django.urls import path\nfrom .views import InventoryMan, SignUp, Dashboard, FakaIzinto, Edit, Khipha\nfrom django.contrib.auth import views as auth_views\n\nurlpatterns = [\n\tpath(\"\", InventoryMan.as_view(), name=\"inventory-man\"),\n\tpath(\"signup/\", SignUp.as_view(), name=\"signup\"),\n\tpath(\"dashboard/\", Dashboard.as_view(), name=\"dashboard\"),\n \tpath(\"add/\", FakaIzinto.as_view(), name=\"faka\"),\n \tpath(\"edit-item/\", Edit.as_view(), name=\"change\"),\n \tpath(\"delete-item/\", Khipha.as_view(), name=\"khipha\"),\n\tpath(\"login/\", auth_views.LoginView.as_view(template_name=\"login.html\"), name=\"login\"),\n\tpath(\"logout\", auth_views.LogoutView.as_view(template_name=\"logout.html\"), name=\"logout\")\n]","repo_name":"malwandemoyo/POSApp","sub_path":"inventory_man/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":704,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"34057121006","text":"import pandas as pd\nfrom datetime import datetime\nimport quandl\n\n\ndados = pd.read_excel('/home/brenno/Downloads/ibov_sp.xlsx', na_values=\"-\")\ncdi_data = quandl.get('BCB/11', start_date='1994-06-01')\ncdi_data['Value'] = cdi_data['Value']/100\ncdi_data['cota'] = (1 + cdi_data['Value']).cumprod() \ncdi_data = cdi_data[['cota']]\ndados['Data'] = pd.to_datetime(dados['Data']).dt.date\ndados = dados.set_index('Data')\n\ndados = pd.merge(dados, cdi_data, left_index=True, right_index=True)\n\n\ndados = dados.dropna()\n\ndados = dados[dados.index > datetime(1999, 12, 31, 0, 0, 0)]\n\ndados = dados.pct_change(periods=(252 * 10))\n\ndados = dados.dropna()\n\nprint('% janelas', sum(dados['ibov'] > dados['cota'])/len(dados))\nprint(\"Periodos positivos ibov:\", sum(dados['ibov'] > 0)/len(dados))\nprint(\"Periodos positivos sp:\", sum(dados['sp'] > 0)/len(dados))\nprint('Media retornos', dados.mean())\n\ndados.to_excel('/home/brenno/Downloads/dados_120m_ibov_sp.xlsx')\n","repo_name":"brennosullivan/instagram_reels","sub_path":"janelas_12m_ibov_sp.py","file_name":"janelas_12m_ibov_sp.py","file_ext":"py","file_size_in_byte":943,"program_lang":"python","lang":"pt","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"2453750157","text":"from lambda_signals.signals import lambda_handler\nimport boto3\nfrom botocore.exceptions import ClientError\nimport datetime\nimport time\n\ndebug = True\n\ndef process_event(event):\n return dict(\n stack_name = event['ResourceProperties']['StackName'],\n resource_name = event['LogicalResourceId'],\n autoscale_grp = event['ResourceProperties']['AutoScaleGrp'],\n aws_region = event['ResourceProperties']['AwsRegion'],\n )\n\ndef get_client(region_name):\n return boto3.client(\"ec2\", region_name = region_name)\n\ndef get_as_client(region_name):\n return boto3.client(\"autoscaling\", region_name = region_name)\n\ndef get_ec2_id(autoscale_group, region_name):\n as_client = get_as_client(region_name)\n as_grp_list = as_client.describe_auto_scaling_groups(AutoScalingGroupNames=[autoscale_group])\n if len(as_grp_list['AutoScalingGroups']) != 1:\n raise Exception(\"ERROR: multiple autoscaling groups found\")\n return as_grp_list['AutoScalingGroups'][0]['Instances'][0]['InstanceId']\n\ndef check_ami_exists(client):\n try:\n response = client.describe_images(Filters=[{'Name':'tag:custom:uuid','Values':[uuid]}])\n if response.get('ResponseMetadata').get('HTTPStatusCode') == 200:\n if len(response.get('Images',[])) > 0:\n return (True, response['Images'][0]['ImageId'])\n except:\n return (False, dict())\n\ndef create(stack_name, resource_name, autoscale_grp, aws_region):\n client = get_client(aws_region)\n success = False\n datenow = datetime.datetime.now().strftime(\"%Y_%m_%d_%H%M\")\n ec2_label = resource_name + datenow\n ec2_id = get_ec2_id(autoscale_grp, aws_region)\n make_response = client.create_image(InstanceId=ec2_id, Name=ec2_label, NoReboot=False)\n if make_response.get('ResponseMetadata').get('HTTPStatusCode') == 200:\n success = True\n ami_id = make_response.get('ImageId')\n client.create_tags(\n Resources=[ami_id],\n Tags = [{\n \"Key\":\"cloudformation:amimanager:stack-name\",\n \"Value\":stack_name,\n },\n {\n \"Key\":\"cloudformation:amimanager:logical-id\",\n \"Value\":resource_name,\n }\n ],\n )\n if success:\n print(\"Success\")\n time.sleep(60)\n client = boto3.client(\"autoscaling\", region_name = aws_region)\n response = client.update_auto_scaling_group(AutoScalingGroupName=autoscale_grp,\n MinSize=0, DesiredCapacity=0)\n return (True, dict(PhysicalResourceId=ami_id,ImageId=ami_id))\n #TODO: set autoscale group size to 0\n return (False, dict())\n\ndef delete(aws_region):\n #TODO: fix delete old AMIs\n return (True, dict())\n #client = get_client(aws_region)\n\n #(ami_exists, ami_id) = check_ami_exists(client)\n #if ami_exists:\n # client.deregister_image(ImageId = ami_id)\n # return (True, dict(DeletedAmiId = ami_id))\n #return (True, dict(Info = \"AMI not found\"))\n\ndef create_resource(event, context):\n if debug:\n print(\"event\")\n print(event)\n print(\"context\")\n print(vars(context))\n event_details = process_event(event)\n return create(**event_details)\n\ndef update_resource(event, context):\n\n event_details = process_event(event)\n return create(**event_details)\n #TODO: delete old ami\n #if old_uuid != event_details['ami_uuid']:\n # (ami_existed, ami_id) = delete(old_uuid, event_details['aws_region'])\n # return create(**event_details)\n #else:\n # return (True, dict(Info=\"Ami uuid not changed\"))\n\ndef delete_resource(event, context):\n if debug:\n print(\"event\")\n print(event)\n print(\"context\")\n print(vars(context))\n event_details = process_event(event)\n\n return delete(event_details['aws_region'])\n\ndef test_resource(event, context):\n print(\"Add tests here\")\n return (True, dict())\n\ndef handler(event, context):\n lambda_handler(\n event,\n context,\n create_function = create_resource,\n delete_function = delete_resource,\n update_function = update_resource,\n test_function = test_resource,\n )\n\nif __name__ == \"__main__\":\n class FakeContext(object):\n def __init__(self):\n context = {\n 'aws_request_id': 'a3de505e-f16b-42f4-b3e6-bcd2e4a73903',\n 'log_stream_name': '2015/10/26/[$LATEST]c71058d852474b9895a0f221f73402ad',\n 'invoked_function_arn': 'arn:aws:lambda:us-west-2:123456789012:function:ExampleCloudFormationStackName-ExampleLambdaFunctionResourceName-AULC3LB8Q02F',\n 'client_context': None,\n 'log_group_name': '/aws/lambda/ExampleCloudFormationStackName-ExampleLambdaFunctionResourceName-AULC3LB8Q02F',\n 'function_name': 'ExampleCloudFormationStackName-ExampleLambdaFunctionResourceName-AULC3LB8Q02F',\n 'function_version': '$LATEST',\n 'identity': '<__main__.CognitoIdentity object at 0x7fd7042a2b90>',\n 'memory_limit_in_mb': '128'\n }\n self.__dict__.update(context)\n\n event = {\n 'StackId': 'Test',\n 'RequestId': '123',\n 'LogicalResourceId': '123',\n 'RequestType':'Test',\n 'ResponseURL':'Test',\n 'test':'test_value_1'\n }\n context = FakeContext()\n handler(event, context)\n","repo_name":"gotropo/gotropo","sub_path":"create/custom/ami_resource.py","file_name":"ami_resource.py","file_ext":"py","file_size_in_byte":5442,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"12165890331","text":"#-*- coding: utf-8 -*-\n\nfrom __future__ import unicode_literals, absolute_import\n\nimport os\n\nfrom pyramid.settings import aslist\nfrom pyramid_jinja2 import renderer_factory\n\nfrom caliop.helpers.config import Configuration\n\n\ndef includeme(config):\n settings = config.registry.settings\n\n config.include('pyramid_jinja2')\n config.add_renderer('.html', renderer_factory)\n\n # XXX should be removed\n rootpath = os.path.dirname(os.path.realpath(__file__))\n\n # configure templates dir (angular build dir)\n template_path = os.path.join(rootpath, settings['caliop.ng.path'])\n config.add_jinja2_search_path(template_path)\n\n # configure static dir on the same dir (angular build dir)\n static_path = os.path.join(rootpath, 'static')\n config.add_static_view('/static', template_path)\n\n for file in aslist(settings['caliop.config']):\n name, path = file.split(':', 1)\n Configuration.load(path, name)\n","repo_name":"LaurentChemla/CaliOpen","sub_path":"caliop/caliop/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":934,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"81"}
+{"seq_id":"16470124279","text":"#!/usr/bin/env python\nimport pyglet\nfrom pyglet.window import key\nimport numpy as np\n\nfrom pyglet_gui.theme import Theme\nfrom pyglet_gui.buttons import Button\nfrom pyglet_gui.manager import Manager\nfrom pyglet_gui.containers import VerticalContainer\n\ntheme = Theme({\"font\": \"Lucida Grande\",\n \"font_size\": 12,\n \"text_color\": [255, 255, 255, 255],\n \"gui_color\": [255, 0, 0, 255],\n \"button\": {\n \"down\": {\n \"image\": {\n \"source\": \"button-down.png\",\n \"frame\": [8, 6, 2, 2],\n \"padding\": [18, 18, 8, 6]\n },\n \"text_color\": [0, 0, 0, 255]\n },\n \"up\": {\n \"image\": {\n \"source\": \"button.png\",\n \"frame\": [6, 5, 6, 3],\n \"padding\": [18, 18, 8, 6]\n }\n }\n }\n }, resources_path='')\n\nPI_180 = np.pi/180\nnp.random.seed(0)\n\n#\n# Parameters that control game behavior and difficulty\n#\n\nast_sizes = [1, 1/2, 1/4] # Factor by which asteroids shrink as they break\nast_points = [100, 50, 25] # How many points each asteroid is worth\np_acc = 100 # Acceleration of player's ship\np_rot_rate = 10.0 # Rotation rate of player's ship\nshoot_speed = 600 # Speed of bullets\nast_max_vel = 50 # Max of x and y components of velocity of asteroids\nast_break_max_vel = 50 # Max amount to add to asteroid vel components after break\nnum_ast = 10 # Initial number of asteroids to generate\nlives = 2 # Number of etra lives that player stars with\n\nwindow = pyglet.window.Window(800, 600)\n\ndef make_img(filename, anchor_x=None, anchor_y=None):\n\timg = pyglet.image.load(filename)\n\timg.anchor_x = img.width // 2\n\timg.anchor_y = img.height // 2\n\tif anchor_x is not None:\n\t\timg.anchor_x = anchor_x\n\tif anchor_y is not None:\n\t\timg.anchor_y = anchor_y\n\treturn img\n\ndef ang_to_vec(angle):\n\ta_norm_x = np.cos(angle*PI_180)\n\ta_norm_y = -np.sin(angle*PI_180)\n\treturn a_norm_x, a_norm_y\n\ndef wrap(xpos, ypos, xmin, xmax, ymin, ymax):\n\tif xpos < xmin:\n\t\txpos = xmax\n\tif xpos > xmax:\n\t\txpos = xmin\n\tif ypos < ymin:\n\t\typos = ymax\n\tif ypos > ymax:\n\t\typos = ymin\n\n\treturn xpos, ypos\n\ndef off_screen(xpos, ypos, xmin, xmax, ymin, ymax):\n\tnewx, newy = wrap(xpos, ypos, xmin, xmax, ymin, ymax)\n\tif (newx != xpos) or (newy != ypos):\n\t\treturn True\n\treturn False\n\nclass Bullet:\n\tdef __init__(self, game, xpos, ypos, xvel, yvel):\n\t\tself.xpos = xpos\n\t\tself.ypos = ypos\n\t\tself.xvel = xvel\n\t\tself.yvel = yvel\n\t\tself.game = game\n\n\t\tbullet_img = make_img('bullet.png')\n\t\tself.bullet = pyglet.sprite.Sprite(bullet_img, x=self.xpos, y=self.ypos, batch=game.batch)\n\n\t# Return True if the bullet went off the screen and needs to be deleted\n\tdef update(self, dt):\n\t\tself.xpos += self.xvel*dt\n\t\tself.ypos += self.yvel*dt\n\n\t\treturn off_screen(self.xpos, self.ypos, self.game.xmin, self.game.xmax,\n\t\t\t self.game.ymin, self.game.ymax)\n\nclass Asteroid:\n\tdef __init__(self, game, xpos, ypos, xvel, yvel, size):\n\t\tself.xpos = xpos\n\t\tself.ypos = ypos\n\t\tself.xvel = xvel\n\t\tself.yvel = yvel\n\t\tself.size = size\n\t\tself.rot = np.random.rand()*360\n\t\tself.exploding = False\n\t\tself.game = game\n\n\t\tasteroid_img = make_img('asteroid.png')\n\t\tself.asteroid = pyglet.sprite.Sprite(asteroid_img, x=self.xpos, y=self.ypos,\n\t\t\t batch=game.batch, group=game.background)\n\n\tdef update(self, dt):\n\t\tself.xpos += self.xvel*dt\n\t\tself.ypos += self.yvel*dt\n\n\t\tself.xpos, self.ypos = wrap(self.xpos, self.ypos, self.game.xmin,\n\t\t\t self.game.xmax, self.game.ymin, \n\t\t\t self.game.ymax)\n\nclass Game:\n\tdef __init__(self, window):\n\t\tself.window = window\n\t\tself.background = pyglet.graphics.OrderedGroup(0)\n\t\tself.foreground = pyglet.graphics.OrderedGroup(1)\n\t\tself.xmin, self.xmax = 0, self.window.width\n\t\tself.ymin, self.ymax = 0, self.window.height\n\t\tself.state = 'MENU'\n\t\tself.player_img = make_img('player.png')\n\t\tself.init_menu_state()\n\n\tdef gen_exp_anim(self):\n\t\texp_anim = pyglet.image.Animation.from_image_sequence(self.explosion_seq, 0.1, False)\n\t\treturn exp_anim\n\n\tdef init_menu_state(self):\n\t\tself.batch = pyglet.graphics.Batch()\n\n\t\tlabel = pyglet.text.Label('AsTeRoIdS', font_name='Times New Roman',\n font_size=36, x=window.width//2,\n y=3*window.height//4, anchor_x='center',\n anchor_y='center', batch=self.batch, group=self.foreground)\n\n\t\tdef callback1(is_pressed):\n\t\t\tself.init_game_state()\n\t\t\tself.state = 'PLAYING'\n\t\tbutton1 = Button('Start Game', on_press=callback1)\n\n\t\tdef callback2(is_pressed):\n\t\t\texit()\n\t\tbutton2 = Button('Quit', on_press=callback2)\n\t\tManager(VerticalContainer([button1, button2]), window=window,\n\t\t\t theme=theme, batch=self.batch)\n\n\t\tself.asteroids = self.gen_asteroids(num_ast)\n\n\tdef init_game_state(self):\n\t\tself.batch = pyglet.graphics.Batch()\n\n\t\tself.reset()\n\n\t\tself.shoot_start = self.player.width // 2\n\t\tflame_img = make_img('flame.png', anchor_x=55)\n\t\tself.flame = pyglet.sprite.Sprite(flame_img, x=self.p_xpos, y=self.p_ypos, group=self.background)\n\n\t\texplosion = pyglet.image.load('explosion.png')\n\t\tself.explosion_seq = pyglet.image.ImageGrid(explosion, 1, 7)\n\n\t\ts_fac = 0.5\n\t\tself.s_lives = []\n\t\tlive_img = make_img('player.png')\n\t\tfor i in range(lives):\n\t\t\tlive_sprite = pyglet.sprite.Sprite(live_img, x=0, y=0, batch=self.batch,\n\t\t\t\t group=self.foreground)\n\t\t\txpos = i*(live_sprite.width*s_fac) + live_sprite.width // 2\n\t\t\typos = self.ymax - live_sprite.height // 2\n\t\t\tlive_sprite.update(xpos, ypos, scale=s_fac, rotation=-90)\n\t\t\tself.s_lives.append(live_sprite)\n\n\t\tself.score = 0\n\t\tself.score_prefix = 'Score: '\n\t\tself.score_label = pyglet.text.Label('', font_name='Times New Roman',\n font_size=18, batch=self.batch,\n anchor_x='left', anchor_y='center',\n group=self.foreground)\n\t\tself.score_label.x = live_sprite.width*(lives+1)\n\t\tself.score_label.y = live_sprite.y\n\t\tself.update_score_ui()\n\n\t\tself.key_left = False\n\t\tself.key_right = False\n\t\tself.key_up = False\n\t\tself.key_space = False\n\n\tdef init_dead_state(self):\n\t\tself.score_label.delete()\n\t\t\n\t\tgameover_label = pyglet.text.Label('Game Over', font_name='Times New Roman',\n font_size=36, x=window.width//2,\n y=3*window.height//4, anchor_x='center',\n anchor_y='center', batch=self.batch)\n\n\t\ttext = 'Final Score: ' + str(self.score)\n\t\tfinalscore_label = pyglet.text.Label(text, font_name='Times New Roman',\n font_size=18, x=window.width//2,\n y=3*window.height//4-50, anchor_x='center',\n anchor_y='center', batch=self.batch)\n\n\t\tdef callback1(is_pressed):\n\t\t\tself.init_game_state()\n\t\t\tself.state = 'PLAYING'\n\t\tbutton1 = Button('Start Game', on_press=callback1)\n\n\t\tdef callback2(is_pressed):\n\t\t\texit()\n\t\tbutton2 = Button('Quit', on_press=callback2)\n\t\tManager(VerticalContainer([button1, button2]), window=window,\n\t\t\t theme=theme, batch=self.batch)\n\n\tdef done_exploding_player(self):\n\t\tif len(self.s_lives) == 0:\n\t\t\tself.init_dead_state()\n\t\t\tself.state = 'DEAD'\n\t\telse:\n\t\t\tself.s_lives.remove(self.s_lives[-1])\n\t\t\tself.reset()\n\n\tdef reset(self):\n\t\tself.p_xpos = (self.xmax - self.xmin)/2\n\t\tself.p_ypos = (self.ymax - self.ymin)/2\n\t\tself.p_xvel, self.p_yvel = 0, 0\n\t\tself.p_rot = 0\n\t\tself.player = pyglet.sprite.Sprite(self.player_img, x=self.p_xpos, y=self.p_ypos, batch=game.batch)\n\t\tself.exploded = False\n\t\tself.bullets = []\n\t\tself.asteroids = self.gen_asteroids(num_ast)\n\n\tdef update_score_ui(self):\n\t\tscore_str = self.score_prefix + '{:4d}'.format(self.score)\n\t\tself.score_label.text = score_str\n\n\tdef gen_asteroids(self, N):\n\t\tasteroids = []\n\t\txpos = np.random.rand(N)*self.xmax\n\t\typos = np.random.rand(N)*self.ymax\n\t\txvel = np.random.rand(N)*ast_max_vel*2 - ast_max_vel\n\t\tyvel = np.random.rand(N)*ast_max_vel*2 - ast_max_vel\n\t\tsize_inds = np.asarray(np.random.rand(N)*len(ast_sizes), dtype=int)\n\t\tsizes = []\n\t\tfor i, idx in enumerate(size_inds):\n\t\t\tsizes.append(ast_sizes[idx])\n\n\t\tfor i in range(N):\n\t\t\tast = Asteroid(self, xpos[i], ypos[i], xvel[i], yvel[i], sizes[i])\n\t\t\tasteroids.append(ast)\n\n\t\treturn asteroids\n\n\t# Check every bullet and asteroid pair for a collision and return\n\t# the indices of both in the list if a collision occurs\n\tdef check_bullet_ast_coll(self):\n\t\tfor idx1 in range(len(self.bullets)):\n\t\t\tfor idx2 in range(len(self.asteroids)):\n\t\t\t\tif self.asteroids[idx2].exploding:\n\t\t\t\t\tcontinue\n\t\t\t\tx1 = self.bullets[idx1].xpos\n\t\t\t\tx2 = self.asteroids[idx2].xpos\n\t\t\t\ty1 = self.bullets[idx1].ypos\n\t\t\t\ty2 = self.asteroids[idx2].ypos\n\t\t\t\tradius = self.asteroids[idx2].asteroid.width // 2\n\n\t\t\t\tdist = np.sqrt((x1 - x2)**2 + (y1 - y2)**2)\n\n\t\t\t\tif dist <= radius:\n\t\t\t\t\treturn idx1, idx2\n\n\t\treturn -1, -1\n\n\t# Check for a collision between the player and every asteroid\n\t# and return the index of the asteroid if one occurs\n\tdef check_player_ast_coll(self):\n\t\tfor idx in range(len(self.asteroids)):\n\t\t\tif self.asteroids[idx].exploding:\n\t\t\t\t\tcontinue\n\t\t\tx1 = self.p_xpos\n\t\t\tx2 = self.asteroids[idx].xpos\n\t\t\ty1 = self.p_ypos\n\t\t\ty2 = self.asteroids[idx].ypos\n\t\t\tradius = self.player.width // 2\n\n\t\t\tdist = np.sqrt((x1 - x2)**2 + (y1 - y2)**2)\n\n\t\t\tif dist <= radius:\n\t\t\t\treturn idx\n\n\t\treturn -1\n\n\tdef shoot(self):\n\t\ta_norm_x, a_norm_y = ang_to_vec(self.p_rot)\n\t\tb_xpos = a_norm_x*self.shoot_start\n\t\tb_ypos = a_norm_y*self.shoot_start\n\t\tb_xvel = a_norm_x*shoot_speed\n\t\tb_yvel = a_norm_y*shoot_speed\n\t\tb = Bullet(self, self.p_xpos + b_xpos, self.p_ypos + b_ypos,\n\t\t\t self.p_xvel + b_xvel, self.p_yvel + b_yvel)\n\t\tself.bullets.append(b)\n\n\t# Break an asteroid into smaller pieces\n\tdef break_asteroid(self, idx):\n\t\ta = self.asteroids[idx]\n\t\tself.asteroids[idx].asteroid.image = self.gen_exp_anim()\n\t\tself.asteroids[idx].exploding = True\n\t\tdef asteroid_done():\n\t\t\tself.asteroids.remove(a)\n\t\tself.asteroids[idx].asteroid.on_animation_end = asteroid_done\n\n\t\tif a.size == ast_sizes[-1]:\n\t\t\treturn\n\t\t\n\t\tsize_idx = ast_sizes.index(a.size)\n\t\txv_add = np.random.rand(2)*ast_break_max_vel*2 - ast_break_max_vel\n\t\tyv_add = np.random.rand(2)*ast_break_max_vel*2 - ast_break_max_vel\n\t\ta1_new = Asteroid(self, a.xpos, a.ypos, a.xvel + xv_add[0],\n\t\t\t a.yvel + yv_add[0], ast_sizes[size_idx+1])\n\t\ta2_new = Asteroid(self, a.xpos, a.ypos, a.xvel + xv_add[1],\n\t\t\t a.yvel + yv_add[1], ast_sizes[size_idx+1])\n\n\t\tself.asteroids.append(a1_new)\n\t\tself.asteroids.append(a2_new)\n\n\tdef update(self, dt):\n\t\tif self.state == 'MENU':\n\t\t\tself.menu_update(dt)\n\t\telif self.state == 'PLAYING':\n\t\t\tself.playing_update(dt)\n\n\tdef menu_update(self, dt):\n\t\tfor a in self.asteroids:\n\t\t\ta.update(dt)\n\n\tdef playing_update(self, dt):\n\t\t# Update player position\n\t\tif not self.exploded:\n\t\t\tself.p_xpos, self.p_ypos = wrap(self.p_xpos, self.p_ypos, self.xmin,\n\t\t\t\t self.xmax, self.ymin, self.ymax)\n\n\t\t\tself.p_xpos += self.p_xvel*dt\n\t\t\tself.p_ypos += self.p_yvel*dt\n\n\t\t# Update asteroid and bullet positions\n\t\tfor a in self.asteroids:\n\t\t\ta.update(dt)\n\n\t\tfor b in self.bullets:\n\t\t\tcleanup = b.update(dt)\n\t\t\tif cleanup:\n\t\t\t\tself.bullets.remove(b)\n\n\t\t# Check for player-asteroid and bullet-asteroid collisions\n\t\tif not self.exploded:\n\t\t\ti1 = self.check_player_ast_coll()\n\t\t\tif i1 >= 0:\n\t\t\t\tself.exploded = True\n\t\t\t\tself.release_all_keys()\n\t\t\t\tself.player.image = self.gen_exp_anim()\n\t\t\t\tself.player.on_animation_end = self.done_exploding_player\n\n\t\ti1, i2 = self.check_bullet_ast_coll()\n\t\tif i1 >= 0 and i2 >= 0:\n\t\t\tself.bullets.remove(self.bullets[i1])\n\t\t\tpoints = ast_points[ast_sizes.index(self.asteroids[i2].size)]\n\t\t\tself.score += points\n\t\t\tself.break_asteroid(i2)\n\n\t\tself.update_score_ui()\n\n\t\t# Handle keyboard input\n\t\tif self.key_left:\n\t\t\tself.p_rot -= p_rot_rate\n\t\t\tif self.p_rot < 0:\n\t\t\t\tself.p_rot += 360\n\t\tif self.key_right:\n\t\t\tself.p_rot += p_rot_rate\n\t\t\tif self.p_rot > 360:\n\t\t\t\tself.p_rot -= 360\n\t\tif self.key_up:\n\t\t\ta_norm_x, a_norm_y = ang_to_vec(self.p_rot)\n\t\t\tself.p_xvel += a_norm_x*p_acc*dt\n\t\t\tself.p_yvel += a_norm_y*p_acc*dt\n\t\tif self.key_space:\n\t\t\tself.shoot()\n\t\t\tself.key_space = False # Prevent player from holding down the shoot button\n\n\tdef release_all_keys(self):\n\t\tself.key_space = False\n\t\tself.key_up = False\n\t\tself.key_left = False\n\t\tself.key_right = False\n\ngame = Game(window)\n\n@window.event\ndef on_key_press(symbol, modifiers):\n\tif game.state == 'PLAYING':\n\t\tif game.exploded:\n\t\t\treturn\n\n\t\tif symbol == key.LEFT:\n\t\t\tgame.key_left = True\n\t\tif symbol == key.RIGHT:\n\t\t\tgame.key_right = True\n\t\tif symbol == key.UP:\n\t\t\tgame.key_up = True\n\t\tif symbol == key.SPACE:\n\t\t\tgame.key_space = True\n\n@window.event\ndef on_key_release(symbol, modifiers):\n\tif game.state == 'PLAYING':\n\t\tif game.exploded:\n\t\t\treturn\n\n\t\tif symbol == key.LEFT:\n\t\t\tgame.key_left = False\n\t\tif symbol == key.RIGHT:\n\t\t\tgame.key_right = False\n\t\tif symbol == key.UP:\n\t\t\tgame.key_up = False\n\t\tif symbol == key.SPACE:\n\t\t\tgame.key_space = False\n\n@window.event\ndef on_draw():\n\twindow.clear()\n\n\tif game.state == 'MENU':\n\t\tfor a in game.asteroids:\n\t\t\ta.asteroid.update(a.xpos, a.ypos, scale=a.size, rotation=a.rot)\n\n\t\tgame.batch.draw()\n\n\tif game.state == 'PLAYING':\n\t\tfor a in game.asteroids:\n\t\t\ta.asteroid.update(a.xpos, a.ypos, scale=a.size, rotation=a.rot)\n\n\t\tgame.player.update(x=game.p_xpos, y=game.p_ypos, rotation=game.p_rot)\n\n\t\tfor b in game.bullets:\n\t\t\tb.bullet.update(b.xpos, b.ypos)\n\n\t\tif game.key_up:\n\t\t\tgame.flame.update(x=game.p_xpos, y=game.p_ypos,\n\t\t\t\t rotation=game.p_rot)\n\t\t\tgame.flame.draw()\n\n\t\tgame.batch.draw()\n\n\tif game.state == 'DEAD':\n\t\tgame.batch.draw()\n\nif __name__ == \"__main__\":\n\tpyglet.clock.schedule_interval(game.update, 1 / 120.0)\n\n\t# Tell pyglet to do its thing\n\tpyglet.app.run()","repo_name":"spencerw/asteroids_pyglet","sub_path":"asteroids.py","file_name":"asteroids.py","file_ext":"py","file_size_in_byte":14210,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12405722749","text":"import sqlite3\n\n\ndef buscar_dato_pais(pais):\n conn = sqlite3.connect('../db.sqlite3')\n c = conn.cursor()\n c.execute(\"SELECT idpais from basedatos_pais WHERE nombrep =\"+str(pais))\n rows = c.fetchall()\n for row in rows:\n print(row[0])\n\n conn.commit()\n conn.close()\n\ndef lista_pais():\n conn = sqlite3.connect('../db.sqlite3')\n c = conn.cursor()\n c.execute(\"SELECT nombrep from basedatos_pais\")\n rows = c.fetchall()\n conn.commit()\n conn.close()\n return rows\n\n\n\n","repo_name":"marceloibarran/conmanzanas","sub_path":"Scrap/consulta.py","file_name":"consulta.py","file_ext":"py","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21921737630","text":"\"\"\"\nSalt Fetching Function for Password Hashing\n- Query the salt in DB with EmailId\n- Pass the salt to Client\n\nAuthor: Mingfei Yang (Mingfei.Yang.1@asu.edu)\n\"\"\"\n\nimport boto3\nfrom botocore.exceptions import ClientError\n\n# Global Table Name\nsalt_table = 'Salts'\n\n\ndef salt_fetcher(event, context):\n \"\"\"\n\n :param event: Event data passed by AWS Lambda, it is usually of the Python dict type.\n (It can also be list, str, int, float, or NoneType type)\n :param context: Runtime information provided by AWS Lambda, it is of the LambdaContext type.\n :return:\n \"\"\"\n # EmailId\n user_email = event['EmailId'].lower()\n\n # Query corresponding salt to DB\n db_client = boto3.client('dynamodb')\n\n # Try - Except - Else Block\n try:\n # DB Salt Query\n response = db_client.query(TableName=salt_table,\n KeyConditionExpression=\"EmailId = :id\",\n ExpressionAttributeValues={':id': {'S': user_email}})\n except ClientError:\n raise Exception(\"Bad Request: Query Failed.\")\n else:\n # If Items retrieved successfully\n if 'Items' in response.keys() and len(response['Items']) != 0:\n # Fetch the unique salt\n item = response['Items'][0]\n\n return {\n 'Salt': item['Salt']['B'].decode()\n }\n # End of Try - Except - Else\n\n raise Exception(\"Bad Request: Query Failed.\")\n","repo_name":"vedant511/DrinkRank","sub_path":"Back-end/PasswordHashing/SaltFetcher.py","file_name":"SaltFetcher.py","file_ext":"py","file_size_in_byte":1459,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"42362412695","text":"import Customer\n\n\nclass Order():\n def __init__(self):\n self.orders = []\n self.exit = False\n self.in_stock = [\"Beer\", \"Wine\", \"Juice\"]\n\n def get_order(self):\n # manually inserting Drink objects into self.orders for smoother testing\n cust = Customer.Customer(\"Christer\", \"Cash\")\n self.orders.append(cust.customer_one_order(\"Beer\", \"Salikatt\", 2))\n self.orders.append(cust.customer_one_order(\"Wine\", \"Radicon\", 2))\n self.orders.append(cust.customer_one_order(\"Juice\", \"Apple\", 3))\n\n return self.orders\n\n def get_order_from_users(self):\n while self.exit == False:\n for i in self.in_stock:\n print(\"In stock: %s\" % i)\n beverage = input(\"What type of beverage do you want? \")\n beverageName = input(\"Name of beverage: \")\n amount = int(input(\"How many? \"))\n\n self.orders.append(Customer.Customer(\"Christer\", \"Cash\").customer_one_order(\n beverage, beverageName, amount))\n answer = input(\"Do you want to order something else y/n \")\n if answer == \"n\":\n self.exit = True\n return self.orders\n","repo_name":"christeriordan/Project","sub_path":"Task1/Orders.py","file_name":"Orders.py","file_ext":"py","file_size_in_byte":1186,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"23490037217","text":"import random\nfrom django.contrib.auth import get_user_model, authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required\n# from django.contrib.auth.models import User\nfrom django.shortcuts import render, redirect\n\n# from django.views import generic as views\n\nUserModel = get_user_model()\n\n\n# Create your views here.\n\n\n@login_required\ndef index(request):\n user = random.randint(1, 1000)\n\n # good practice\n # UserModel.objects.create_user(\n # username=f'admin{user}',\n # email=f'admin{user}@example.com',\n # password='1234',\n # )\n # bad practice\n # UserModel.objects.create(\n # username=f'admin{user}',\n # email=f'admin{user}@example.com',\n # password='1234',\n # )\n some_object = UserModel.objects.get(username=f'admin63')\n # some_object = User.objects.get(username='admin')\n context = {\n \"user\": some_object,\n 'permission': request.user.has_perm('web.view_user')\n }\n return render(request, 'index.html', context)\n\n\ndef login_user(request):\n user = authenticate(\n username='admin327',\n password='1234',\n )\n login(request, user)\n print(f\"test result -->{user}\")\n\n return redirect('index')\n\n\ndef logout_user(request):\n logout(request)\n return redirect('index')\n","repo_name":"qceka88/Python-Web-Basics-and-FrameWork","sub_path":"Python-Web-Framework/04 Authentication and Authorization/autetications_authorisation/autetications_authorisation/web/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1311,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"10047655212","text":"import logging\nimport queue\nimport threading\nfrom typing import Dict, List, Tuple, Callable, Optional, Union # For self-documenting typing\nimport uuid\nimport os\nimport asyncio\n\nfrom .tracker import TrackingPoller, TrackerBase\nfrom .job import JobStatus, Job, ExternalJob\nfrom ..exceptions import JobNotFoundException\nfrom ..notifications.notifiers import Notifier\n\n# Do not expose anything by default (internal module)\n__all__ = [] # type: List[str]\n\n\nLOGGER = logging.getLogger(__name__)\n\n\nclass QueueProcessor:\n \"\"\"\n Process items from queue in a new thread. Usage:\n 1) queue_processor.start(queue_=..., process_item=...)\n 2) queue_processor.schedule_stop()\n 3) queue_processor.wait_stop()\n The two shutdown steps ensure that the caller can cancel the currently running task between them,\n ensuring that processor exits without processing new jobs.\n \"\"\"\n def __init__(self):\n self._stop_event = threading.Event()\n self._queue = None\n self._thread = None\n\n def start(self, queue_, process_item):\n \"\"\"\n Read and handle tasks from queue `queue_` until (1) queue item is None or (2) stop_event is set. Note that\n the currently running job is not forced to cancel: that should be done from another thread, letting queue reader\n to check loop condition. `QueueProcessor` is NOT safe for reuse for processing other queues.\n :param queue_: Synchronized queue\n :param process_item: Callback called with queue item as argument\n :return:\n \"\"\"\n self._queue = queue_\n self._thread = threading.Thread(target=self.__process, args=(process_item,))\n self._stop_event.clear()\n self._thread.start()\n\n def __process(self, process_item):\n if not self.is_running():\n raise RuntimeError(\"QueueProcessor must be started first.\")\n while not self._stop_event.is_set():\n item = self._queue.get(block=True)\n if item is None or self._stop_event.is_set():\n break\n process_item(item)\n self._queue.task_done()\n\n def schedule_stop(self):\n if not self.is_running():\n return\n self._stop_event.set() # Signal exit to worker thread, required as \"None\" may not be next task\n self._queue.put(None, block=False) # Signal exit if thread is blocking\n\n def is_running(self):\n return self._thread is not None and self._thread.is_alive()\n\n def wait_stop(self):\n if self.is_running():\n self._thread.join()\n self._thread = None\n self._queue = None\n\n\nclass Scheduler:\n def __init__(self, queue_processor: QueueProcessor, notifier: Notifier = None,\n event_loop: Optional[asyncio.AbstractEventLoop] = None):\n self._queue_processor = queue_processor\n self.submitted_jobs = dict() # type: Dict[uuid.UUID, Job]\n self.external_jobs = dict() # type: Dict[uuid.UUID, ExternalJob]\n self._job_queue = queue.Queue() # type: queue.Queue\n self._running_job = None # type: Optional[Job]\n self._job_poller = TrackingPoller(self.__query_and_report)\n self._event_loop = event_loop or asyncio.get_event_loop() # Save the event loop for out-of-thread operations\n self._notifier = notifier # type: Optional[Notifier]\n self.active_external_job_id = None # type: Optional[uuid.UUID] # TODO Allow one per process ID\n self.external_job_polling_tasks = dict() # type: Dict[uuid.UUID, asyncio.Task]\n\n # Properties and Python magic\n\n @property\n def jobs(self): # Needed to access internal list of jobs as object parameters are unexposable, only methods\n return list(self.submitted_jobs.values()) + list(self.external_jobs.values())\n\n @property\n def is_running(self):\n return self._queue_processor.is_running()\n\n def __enter__(self):\n self.start()\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n self.stop()\n\n # Job handling methods\n def _handle_job(self, job: Job) -> None:\n LOGGER.debug(\"Handling job: %s\", job)\n if job.status.stale:\n return\n self._running_job = job\n\n task = None # type: Optional[asyncio.Task]\n if job.poll_time:\n # Create and schedule a task from the Polling job, so we can cancel it without killing the event loop\n task = self._event_loop.create_task(self._job_poller.poll(job.id, job.poll_time))\n if self._notifier:\n self._notifier.notify_job_start(job)\n try:\n job.launch_and_wait()\n except Exception: # pylint:disable=broad-except\n LOGGER.exception(\"Running job failed\")\n finally:\n if task:\n task.cancel()\n\n if self._notifier:\n self._notifier.notify_job_end(job)\n self._running_job = None\n LOGGER.debug(\"Finished handling job: %s\", job)\n\n def submit_job(self, job: Job):\n job.status = JobStatus.QUEUED\n self.submitted_jobs[job.id] = job\n self._job_queue.put(job) # TODO Blocks if queue full\n LOGGER.debug(\"Job submitted: %s\", job)\n\n def stop_job(self, job_id: uuid.UUID):\n if job_id not in self.submitted_jobs:\n LOGGER.debug(\"Ignoring stopping unknown job with ID: %s\", str(job_id))\n return\n self.submitted_jobs[job_id].cancel()\n\n def register_external_job(self, job_id: uuid.UUID):\n self.active_external_job_id = job_id\n job = self.external_jobs[job_id] # type: ExternalJob\n LOGGER.debug(\"Handling job: %s\", job)\n\n if job.poll_time:\n task = self._event_loop.create_task(self._job_poller.poll(job.id, job.poll_time))\n self.external_job_polling_tasks[job_id] = task\n if self._notifier:\n self._notifier.notify_job_start(job)\n\n def unregister_external_job(self, job_id: uuid.UUID):\n job = self.external_jobs[job_id] # type: ExternalJob\n self.active_external_job_id = None\n if self._notifier:\n self._notifier.notify_job_end(job)\n task = self.external_job_polling_tasks.pop(job.id, None) # type: Optional[asyncio.Task]\n if task is not None:\n task.cancel()\n\n def __get_job_by_pid(self, pid) -> uuid.UUID:\n jobs = [job for job in self.jobs if job.pid == pid]\n if not jobs:\n raise JobNotFoundException(job_id=str(pid))\n if len(jobs) == 1:\n return jobs[0].id\n # Check if one of them is active\n active_jobs = [job for job in jobs if self.active_external_job_id == job.id]\n if not active_jobs:\n return jobs[0].id\n return active_jobs[0].id\n\n def add_condition(self, pid: int, *vals: str, condition: Callable[[float], bool], only_relevant: bool):\n \"\"\"Adds a new condition for a job that matches the given process id.\n :param pid: process id\n :param vals: list of scalar names (strings)\n :param condition: a callable that accepts as many values as vals, and returns a boolean whether a condition has\n been met\n :param only_relevant: a boolean, whether or not only the values relevant to the condition should be plotted when\n this condition is met\n \"\"\"\n job_id = self.__get_job_by_pid(pid)\n job = self.get_job_by_id(job_id=job_id)\n job.add_condition(*vals, condition=condition, only_relevant=only_relevant)\n\n def report_scalar(self, pid, name, val):\n # Find the right job id\n job_id = self.__get_job_by_pid(pid)\n job = self.get_job_by_id(job_id)\n condition = job.add_scalar_to_history(scalar_name=name, scalar_value=val)\n if condition and self._notifier:\n # TODO - we can add the condition that triggered the notification...\n names = condition.names if condition.only_relevant else list()\n vals, imgpath = self.query_scalars(*names, job_id=job.id, latest_only=False, plot=True)\n self._notifier.notify(job, imgpath, n_iterations=-1)\n if imgpath is not None:\n os.remove(imgpath)\n\n def get_job_by_id(self, job_id: uuid.UUID):\n job = self.submitted_jobs.get(job_id) or self.external_jobs.get(job_id)\n if not job:\n raise JobNotFoundException(job_id=str(job_id))\n return job\n\n def query_scalars(self, *names: Tuple[str, ...], job_id, latest_only: bool = True, plot: bool = False):\n job = self.get_job_by_id(job_id=job_id)\n return job.get_updates(*names, plot=plot, latest=latest_only)\n\n def __query_and_report(self, job_id: uuid.UUID):\n if self._notifier:\n job = self.get_job_by_id(job_id)\n # Get updates; TODO - vals should be reported once we update schema...\n vals, imgpath = self.query_scalars(job_id=job_id, latest_only=True, plot=True)\n if vals: # Only send updates if there exists any updates\n self._notifier.notify(job, imgpath, n_iterations=-1)\n if imgpath is not None:\n os.remove(imgpath)\n\n # Scheduler service methods\n\n def start(self):\n if not self._queue_processor.is_running():\n LOGGER.debug(\"Start queue processor\")\n self._queue_processor.start(queue_=self._job_queue, process_item=self._handle_job)\n LOGGER.debug(\"Queue processor started\")\n\n def stop(self):\n # self._job_poller.stop()\n self._queue_processor.schedule_stop()\n if self._running_job is not None:\n # TODO Add an option to not cancel the currently running job?\n self._running_job.cancel()\n if self._queue_processor.is_running():\n # Wait for the thread to finish\n self._queue_processor.wait_stop()\n","repo_name":"meeshkan/meeshkan-client","sub_path":"meeshkan/core/scheduler.py","file_name":"scheduler.py","file_ext":"py","file_size_in_byte":9841,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"81"}
+{"seq_id":"29821603111","text":"import pytest\nimport os\nfrom utils import aws_s3\nimport boto3\n\n\nclass Tests:\n @pytest.fixture\n def example_fixture(self):\n '''\n An example of a pytest fixture - a function that can be used for setup and teardown before and after test functions are run.\n '''\n\n # place any setup you want to do before any test function that uses this fixture is run\n\n yield # at th=e yield point, the test function will run and do its business\n\n # place with any teardown you want to do after any test function that uses this fixture has completed\n\n #\n # Test functions\n #\n\n def test_sanity_check(self, example_fixture):\n \"\"\"\n Test debugging... making sure that we can run a simple test that always passes.\n Note the use of the example_fixture in the parameter list - any setup and teardown in that fixture will be run before and after this test function executes\n From the main project directory, run the `python3 -m pytest` command to run all tests.\n \"\"\"\n expected = True # the value we expect to be present\n actual = True # the value we see in reality\n assert actual == expected, \"Expected True to be equal to True!\"\n\n def test_upload_file(self, example_fixture):\n \"\"\"\n Test the upload_file function from the aws_s3 module\n \"\"\"\n expected = \"s3://software-eng-project-4/sample_audio.mp3\"\n file_path = os.path.join(os.getcwd(), \"recordings\", \"sample_audio.mp3\")\n actual = aws_s3.upload_file(file_path, 'software-eng-project-4')\n assert actual == expected, f\"Expected {actual} to be equal to s3://software-eng-project-4/sample_audio.mp3!\"\n\n def test_wrong_file(self, example_fixture):\n \"\"\"\n Test the upload_file function throw exception when file does not exist\n \"\"\"\n try:\n with pytest.raises(FileNotFoundError):\n aws_s3.upload_file('not_a_file.mp5', 'software-eng-project-4')\n except Exception as e:\n assert False, \"Expected FileNotFoundError to be thrown!\"\n\n def test_file_exists_in_bucket(self, example_fixture):\n \"\"\"\n Test whether file exists in bucket after upload_file function\n \"\"\"\n s3 = boto3.resource('s3')\n bucket = s3.Bucket('software-eng-project-4')\n file_path = os.path.join(os.getcwd(), \"recordings\", \"sample_audio.mp3\")\n file_name = os.path.basename(file_path)\n actual = aws_s3.upload_file(file_path, 'software-eng-project-4')\n objs = list(bucket.objects.filter(Prefix=file_name))\n if any([w.key == file_name for w in objs]):\n assert True, \"File exist in Bucket!\"\n else:\n assert False, \"File does not exist in Bucket!\"\n","repo_name":"software-students-fall2022/containerized-app-exercise-team3","sub_path":"machine-learning-client/tests/test_s3.py","file_name":"test_s3.py","file_ext":"py","file_size_in_byte":2764,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"12204230812","text":"# -*- coding: utf-8 -*-\n\nfrom django.http import HttpResponse\nfrom userprofile.models import Profile\nfrom django.contrib.auth.models import User\nimport settings\nfrom uuid import uuid5, UUID\nfrom core.portal.render import render_to_portal, render_ajax\nfrom django.template import RequestContext, loader as template_loader\nfrom core.portal.exceptions import Http302\n\n@render_to_portal(template='profile/index.html')\ndef index(request):\n\tcontext = {}\n\tuser = request.user\n\tif not user.is_authenticated():\n\t\traise Http302('/login/')\n\ttry:\n\t\tcontext['profile'] = user.get_profile()\n\texcept:\n\t\tcontext['profile'] = None\n\tcontext['user'] = user\n\treturn context\n\n@render_ajax(type=\"html\", template='profile/view.html')\ndef update(request):\n\tcontext = {}\n\tuser = request.user\n\tif not user.is_authenticated():\n\t\traise Http302('/login/')\n\ttry:\n\t\tcontext['profile'] = user.get_profile()\n\texcept:\n\t\tcontext['profile'] = None\n\treturn context\n\n@render_ajax(type=\"html\")\ndef edit(request):\n\tfrom forms import ProfileForm\n\tcontext = {}\n\tuser = request.user\n\tif not user.is_authenticated():\n\t\traise Http302('/login/')\n\ttry:\n\t\tprofile = user.get_profile()\n\texcept:\n\t\tprofile = None\n\tif request.method == 'POST':\n\t\tform = ProfileForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tdata = form.cleaned_data\n\t\t\tuser.last_name = data['last_name']\n\t\t\tuser.first_name = data['first_name']\n\n\t\t\tdrop_email = user.email != data['email']\n\n\t\t\tuser.email = data['email']\n\t\t\tuser.save()\n\t\t\tif data['middle_name'] or data['subscription']:\n\t\t\t\tif profile:\n\t\t\t\t\tprofile.middle_name = data['middle_name']\n\t\t\t\t\tprofile.subscription = data['subscription']\n\t\t\t\telse:\n\t\t\t\t\tprofile = Profile(user=user,\n\t\t\t\t\t\t\tmiddle_name=data['middle_name'],\n\t\t\t\t\t\t\tsubscription = data['subscription'])\n\n\t\t\t\tif drop_email:\n\t\t\t\t\tprofile.is_email_confirmed = False\n\n\t\t\t\tprofile.save()\n\t\t\treturn u'Данные успешно сохранены'\n\t\tcontext['form'] = form\n\telse:\n\t\tdata = {'last_name':user.last_name,\n\t\t\t\t'first_name': user.first_name,\n\t\t\t\t'email':user.email}\n\t\tif profile:\n\t\t\tdata['middle_name'] = profile.middle_name\n\t\t\tdata['subscription'] = profile.subscription\n\t\tcontext['form'] = ProfileForm(initial=data)\n\tcontext['profile'] = profile\n\treturn template_loader.get_template(\"profile/edit.html\").render(\n\t\t\tRequestContext(request, context))\n\n@render_ajax()\ndef send_email_confirm(request):\n\tfrom django.core.mail import send_mail\n\tfrom uuid import uuid5, UUID\n\timport settings\n\n\tif not request.user.is_authenticated():\n\t\traise Http302('/login/')\n\n\tcode = uuid5(UUID(settings.UUID_NAMESPACE_FOR_EMAIL_CONFIRM), str(request.user.email))\n\tsubj = u'[www.tsogu.ru] Подтверждение email'\n\tbody = u'Доброго времени суток.\\n\\nПользователь %s отправил запрос на подтверждение электронной почты. Для подтверждения электронной почты на сайте ТюмГНГУ пройдите по ссылке:\\nhttp://www.tsogu.ru/userprofile/email_confirm/%s/%s/\\nЕсли данное письмо попало к Вам по ошибке - удалите его.\\n\\n---\\nС уважением,\\nАдминистрация портала www.tsogu.ru,\\nwebmaster@tsogu.ru' % (request.user.username, request.user.id, code)\n\n\ttry:\n\t\tsend_mail(subj, body, settings.DEFAULT_FROM_EMAIL, [request.user.email], fail_silently=False)\n\t\treturn u'Пароль подтверждения отправлен на электронную почту'\n\texcept:\n\t\treturn u'Произошел сбой при отправке почты. Повторите попытку позже или сообщите вебмастеру о сбое.'\n\ndef student_status_confirm(request):\n\timport urllib\n\timport simplejson as json\n\tfrom hashlib import md5\n\n\timport settings\n\n\tuser = request.user\n\tif not user.is_authenticated():\n\t\traise Http302('/login/')\n\ttry:\n\t\tuserprofile = user.get_profile()\n\texcept:\n\t\tuserprofile = Profile(user=user)\n\tlogin = request.POST.get('login', '')\n\tpassword = md5(request.POST.get('password', '').encode('utf-8')).hexdigest()\n\tparams = \"login=%s&password=%s\"%(login.encode('utf-8'), password)\n\tresult = urllib.urlopen(settings.EDUCON_URL, params).read() or u\"Произошла досадная ошибка, попробуйте позже.\"\n\ttry:\n\t\tuserprofile.student_id = int(result)\n\t\tuserprofile.save()\n\t\treturn HttpResponse(u\"Статус студента подтверждён, обновите профиль.\", mimetype=\"text/plain\")\n\texcept:\n\t\treturn HttpResponse(content=result, mimetype=\"text/plain\", status=500)\n\ndef get_student_info(student_id):\n\timport urllib2, base64\n\timport simplejson as json\n\n\timport settings\n\n\tparams = \"field=group\"\n\trequest = urllib2.Request('%sstudents/student/%d/?%s'%(settings.API_TSOGU, student_id, params))\n\tbase64string = base64.encodestring('%s:%s' % (settings.API_TSOGU_USERNAME, settings.API_TSOGU_PASSWORD))[:-1]\n\tauthheader = \"Basic %s\" % base64string\n\trequest.add_header(\"Authorization\", authheader)\n\tdata = json.loads(urllib2.urlopen(request).read())\n\treturn data\n\n\ndef email_confirm(request, user_id, code):\n\tuser = User.objects.get(id=user_id)\n\ttrue_code = uuid5(UUID(settings.UUID_NAMESPACE_FOR_EMAIL_CONFIRM), str(user.email))\n\tif str(true_code) == str(code):\n\t\ttry:\n\t\t\tprofile = user.get_profile()\n\t\texcept:\n\t\t\tprofile = Profile(user=user)\n\t\tprofile.is_email_confirmed = True\n\t\tprofile.save()\n\t\treturn HttpResponse(u'Ваш email подтвержден')\n\telse:\n\t\treturn HttpResponse(u'Код подтверждения не верен')\n","repo_name":"vden/TsoguNG","sub_path":"userprofile/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5554,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"5925141351","text":"# 주식가격\n\ndef solution(prices):\n plen = len(prices)\n answer = [0] * plen\n for i in range(plen):\n for j in range(i + 1, plen):\n answer[i] += 1\n if prices[j] < prices[i]:\n break\n return answer\n","repo_name":"KimHS0915/programmers-learn-challenges","sub_path":"lv2/stock_prices.py","file_name":"stock_prices.py","file_ext":"py","file_size_in_byte":253,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"73251392264","text":"from transformers import BertTokenizerFast, EncoderDecoderModel\nimport torch\nfrom bs4 import BeautifulSoup\nimport ssl\nimport urllib.request, urllib.parse, urllib.error\nimport requests\nimport sqlite3\nimport timeit\n\n_start_time=timeit.default_timer()\n#print(_start_time)\ndef end_timer(suffix = None):\n evalTime = timeit.default_timer() - _start_time\n #suffix = \"\" if suffix is None else \"_\" + suffix\n print(\"time: \",evalTime)\n text = (str(evalTime))\n with open('runtime.txt', 'w') as f:\n f.write(text)\n\nconn = sqlite3.connect('sumapp.sqlite')\ncur = conn.cursor()\n\n#Ignore SSL certificate errors\nctx = ssl.create_default_context()\nctx.check_hostname = False\nctx.verify_mode = ssl.CERT_NONE\n\nstarturl = input('Enter web url or enter: ')\nif ( len(starturl) < 1 ) : starturl = 'https://en.wikipedia.org/wiki/Portal:Speculative_fiction'\nif ( starturl.endswith('/') ) : starturl = starturl[:-1]\n\nhtml = urllib.request.urlopen(starturl, context=ctx).read()\n\n#summarizer = pipeline(\"summarization\")\n#URL = \"http://galactanet.com/oneoff/theegg_mod.html\"\n#URL = \"https://en.wikipedia.org/wiki/Kawaii\"\n\nr = requests.get(starturl)\nsoup = BeautifulSoup(r.text, 'html.parser')\nresults = soup.find_all(['h1', 'p'])\ntext = [result.text for result in results]\nARTICLE = ' '.join(text)\n#print(ARTICLE)\n\nweb = starturl\nif ( starturl.endswith('.htm') or starturl.endswith('.html') ) :\n pos = starturl.rfind('/')\n web = starturl[:pos]\n\nif ( len(web) > 1 ) :\n cur.execute('INSERT OR IGNORE INTO Websites (url) VALUES ( ? )', ( web, ) )\n cur.execute('INSERT OR IGNORE INTO Pages (url, html, new_rank) VALUES ( ?, ?, 1.0 )', ( starturl, html) )\n conn.commit()\n\ntry:\n document = urlopen(url, context=ctx)\n\n html = document.read()\n if document.getcode() != 200 :\n print(\"Error on page: \",document.getcode())\n cur.execute('UPDATE Pages SET error=? WHERE url=?', (document.getcode(), web) )\n\n if 'text/html' != document.info().get_content_type() :\n print(\"Ignore non text/html page\")\n cur.execute('DELETE FROM Pages WHERE url=?', ( web, ) )\n conn.commit()\n \n\n print('('+str(len(html))+')', end=' ')\n\n \nexcept KeyboardInterrupt:\n print('')\n print('Program interrupted by user...')\n \nexcept:\n #print(\"Unable to retrieve or parse page\")\n cur.execute('UPDATE Pages SET error=-1 WHERE url=?', (web, ) )\n conn.commit()\n\nmax_chunk = 500\n\nARTICLE = ARTICLE.replace('.', '.')\nARTICLE = ARTICLE.replace('?', '?')\nARTICLE = ARTICLE.replace('!', '!')\n\nsentences = ARTICLE.split('')\ncurrent_chunk = 0 \nchunks = []\nfor sentence in sentences:\n if len(chunks) == current_chunk + 1: \n if len(chunks[current_chunk]) + len(sentence.split(' ')) <= max_chunk:\n chunks[current_chunk].extend(sentence.split(' '))\n else:\n current_chunk += 1\n chunks.append(sentence.split(' '))\n else:\n #print(current_chunk)\n chunks.append(sentence.split(' '))\n\nfor chunk_id in range(len(chunks)):\n chunks[chunk_id] = ' '.join(chunks[chunk_id])\n\nprint(\"Number of chunks:\", len(chunks), \",summarize..\")\n\n#res = summarizer(chunks, max_length=120, min_length=30, do_sample=False)\n \ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\ntokenizer = BertTokenizerFast.from_pretrained('mrm8488/bert-small2bert-small-finetuned-cnn_daily_mail-summarization')\nmodel = EncoderDecoderModel.from_pretrained('mrm8488/bert-small2bert-small-finetuned-cnn_daily_mail-summarization').to(device)\n\ndef generate_summary(text):\n # cut off at BERT max length 512\n inputs = tokenizer([text], padding=\"max_length\", truncation=True, max_length=512, return_tensors=\"pt\")\n input_ids = inputs.input_ids.to(device)\n attention_mask = inputs.attention_mask.to(device)\n\n output = model.generate(input_ids, attention_mask=attention_mask)\n \n return tokenizer.decode(output[0], skip_special_tokens=True)\n\n#fname = input('Enter file name: ')\n#if (len(fname) < 1): fname = 'romeo.txt' \n#txt = [\n #(open(fname)).read()\n# str(ARTICLE)\n\n#]\nsum = generate_summary(str(ARTICLE))\nprint(sum)\n\n# Retrieve all of the title tags\ntxt = list()\ntags_second = soup('title')\nfor title in tags_second:\n txt.append(title.get_text())\n #print(title.get_text())\n\ncur.execute('INSERT OR IGNORE INTO urlIndexes (url, url_title, url_full_txt, url_sum) VALUES ( ?, ?, ?, ?)',\n ( (starturl, title.get_text().strip() ,str(ARTICLE).strip() , str(sum)) ) )\n\nconn.commit()\ncur.close()\n\nend_timer()\n","repo_name":"LehnerL/sumApp","sub_path":"sumApp/article_sum.py","file_name":"article_sum.py","file_ext":"py","file_size_in_byte":4527,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"42875188484","text":"'''\nGiven an array of positive numbers and a positive number ‘k’, find the maximum sum of any contiguous subarray of size ‘k’.\n\nExample 1:\n\nInput: [2, 1, 5, 1, 3, 2], k=3 \nOutput: 9\nExplanation: Subarray with maximum sum is [5, 1, 3].\nExample 2:\n\nInput: [2, 3, 4, 1, 5], k=2 \nOutput: 7\nExplanation: Subarray with maximum sum is [3, 4].\n'''\n\n\ndef max_sub_array(nums, k):\n\n windowSum = 0\n windowStart = 0\n max_sum = 0\n\n for windowEnd in range(0, len(nums)):\n\n windowSum += nums[windowEnd]\n\n if windowEnd >= k-1:\n max_sum = max(max_sum, windowSum)\n windowSum -= nums[windowStart]\n windowStart += 1\n\n return max_sum\n\n\nprint(max_sub_array([5, 1, 3], 3))\n","repo_name":"prashantchanne12/Leetcode","sub_path":"maximum sum sub array of size k.py","file_name":"maximum sum sub array of size k.py","file_ext":"py","file_size_in_byte":719,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"24692843724","text":"\"\"\"\n@Coding: uft-8\n@Time: 2019-08-25 18:12\n@Author: Ryne Chen\n@File: CoinCombination.py \n@Python Version: 3.6\n\"\"\"\n\nimport numpy as np\n\n\ndef combination(coins, n):\n m = len(coins)\n dp = np.zeros((m + 1, n + 1))\n\n for i in range(m + 1):\n dp[i][0] = 1\n\n for i in range(1, m + 1):\n for j in range(1, n + 1):\n for k in range(j // coins[i - 1] + 1):\n dp[i][j] += dp[i - 1][j - k * coins[i - 1]]\n\n return int(dp[m][n])\n\n\ndef main():\n coins = [1, 2, 5, 10]\n n = 5\n\n comb = combination(coins, n)\n print(comb)\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"pchen12567/Leetcode","sub_path":"Mock/瓜子2019秋/硬币组合/CoinCombination.py","file_name":"CoinCombination.py","file_ext":"py","file_size_in_byte":607,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72016148104","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Sep 25 18:18:41 2021\r\n\r\n@author: mdt20\r\n\"\"\"\r\n\r\n#UFO scrape\r\n\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\n\r\ndef get_links():\r\n \"\"\"gather links to scrape\"\"\"\r\n \r\n lst_of_links = []\r\n url_for_links = \"http://www.nuforc.org/webreports/ndxevent.html\"\r\n fhand = requests.get(url_for_links)\r\n soup = BeautifulSoup(fhand.content, 'html.parser') \r\n\r\n for link in soup.findAll('a'):\r\n urls = link.get('href')\r\n lst_of_links.append(urls)\r\n \r\n #don't return the first value, its not a useful link\r\n return lst_of_links[1:] \r\n \r\ndef build_urls_list(lst_of_links):\r\n \"\"\"build list of complete urls for scraping\"\"\"\r\n \r\n url_lst = []\r\n base = \"http://www.nuforc.org/webreports/\"\r\n \r\n for partial in lst_of_links:\r\n full_url = base + partial\r\n url_lst.append(full_url)\r\n \r\n with open(\"ufo_full_urls.txt\", \"w\") as write_file:\r\n for link in url_lst:\r\n write_file.write(link + \"\\n\")\r\n\r\n\r\ndef main():\r\n gather = get_links()\r\n build_urls_list(gather)\r\n \r\nmain()\r\n\r\n","repo_name":"blunderfist/py-ufo-scrape","sub_path":"ufo_scrape_prep_links_list.py","file_name":"ufo_scrape_prep_links_list.py","file_ext":"py","file_size_in_byte":1121,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5594354914","text":"from __future__ import division\n\nimport os\n\nimport torch\nfrom PIL import Image\nfrom tensorboard_logger import configure, log_value\nfrom torch.autograd import Variable\nfrom torch.utils import data\nfrom torchvision.transforms import Compose, Normalize, ToTensor\nfrom tqdm import tqdm\n\nfrom argmyparse import get_src_only_training_parser, add_additional_params_to_args, fix_img_shape_args\nfrom datasets import get_dataset\nfrom loss import CrossEntropyLoss2d\nfrom models.model_util import get_optimizer, get_full_model # check_training\nfrom transform import ReLabel, ToLabel, Scale, RandomSizedCrop, RandomHorizontalFlip, RandomRotation\nfrom util import check_if_done, save_checkpoint, adjust_learning_rate, emphasize_str, get_class_weight_from_file\nfrom util import mkdir_if_not_exist, save_dic_to_json\n\nparser = get_src_only_training_parser()\nargs = parser.parse_args()\nargs = add_additional_params_to_args(args)\nargs = fix_img_shape_args(args)\n\nif args.resume:\n print(\"=> loading checkpoint '{}'\".format(args.resume))\n if not os.path.exists(args.resume):\n raise OSError(\"%s does not exist!\" % args.resume)\n\n indir, infn = os.path.split(args.resume)\n\n old_savename = args.savename\n args.savename = infn.split(\"-\")[0]\n print (\"savename is %s (original savename %s was overwritten)\" % (args.savename, old_savename))\n\n checkpoint = torch.load(args.resume)\n args = checkpoint['args'] # Load args!\n\n model = get_full_model(net=args.net, res=args.res, n_class=args.n_class, input_ch=args.input_ch)\n optimizer = get_optimizer(model.parameters(), opt=args.opt, lr=args.lr, momentum=args.momentum,\n weight_decay=args.weight_decay)\n model.load_state_dict(checkpoint['state_dict'])\n optimizer.load_state_dict(checkpoint['optimizer'])\n print(\"=> loaded checkpoint '{}'\".format(args.resume))\n\n json_fn = os.path.join(args.outdir, \"param_%s_resume.json\" % args.savename)\n check_if_done(json_fn)\n args.machine = os.uname()[1]\n save_dic_to_json(args.__dict__, json_fn)\n\nelse:\n model = get_full_model(net=args.net, res=args.res, n_class=args.n_class, input_ch=args.input_ch)\n optimizer = get_optimizer(model.parameters(), opt=args.opt, lr=args.lr, momentum=args.momentum,\n weight_decay=args.weight_decay)\n\n args.outdir = os.path.join(args.base_outdir, \"%s-%s_only_%sch\" % (args.src_dataset, args.split, args.input_ch))\n args.pth_dir = os.path.join(args.outdir, \"pth\")\n\n if args.net in [\"fcn\", \"psp\"]:\n model_name = \"%s-%s-res%s\" % (args.savename, args.net, args.res)\n else:\n model_name = \"%s-%s\" % (args.savename, args.net)\n\n args.tflog_dir = os.path.join(args.outdir, \"tflog\", model_name)\n mkdir_if_not_exist(args.pth_dir)\n mkdir_if_not_exist(args.tflog_dir)\n\n json_fn = os.path.join(args.outdir, \"param-%s.json\" % model_name)\n check_if_done(json_fn)\n args.machine = os.uname()[1]\n save_dic_to_json(args.__dict__, json_fn)\n\ntrain_img_shape = tuple([int(x) for x in args.train_img_shape])\n\nimg_transform_list = [\n Scale(train_img_shape, Image.BILINEAR),\n ToTensor(),\n Normalize([.485, .456, .406], [.229, .224, .225])\n]\n\nif args.augment:\n aug_list = [\n RandomRotation(),\n # RandomVerticalFlip(), # non-realistic\n RandomHorizontalFlip(),\n RandomSizedCrop()\n ]\n img_transform_list = aug_list + img_transform_list\n\nimg_transform = Compose(img_transform_list)\n\nlabel_transform = Compose([\n Scale(train_img_shape, Image.NEAREST),\n ToLabel(),\n ReLabel(255, args.n_class - 1),\n])\n\nsrc_dataset = get_dataset(dataset_name=args.src_dataset, split=args.split, img_transform=img_transform,\n label_transform=label_transform, test=False, input_ch=args.input_ch)\n\nkwargs = {'num_workers': 1, 'pin_memory': True} if torch.cuda.is_available() else {}\ntrain_loader = torch.utils.data.DataLoader(src_dataset, batch_size=args.batch_size, shuffle=True, **kwargs)\n\nweight = get_class_weight_from_file(n_class=args.n_class, weight_filename=args.loss_weights_file,\n add_bg_loss=args.add_bg_loss)\n\nif torch.cuda.is_available():\n model.cuda()\n weight = weight.cuda()\n\ncriterion = CrossEntropyLoss2d(weight)\n\nconfigure(args.tflog_dir, flush_secs=5)\n\nmodel.train()\n\nfor epoch in range(args.epochs):\n epoch_loss = 0\n for ind, (images, labels) in tqdm(enumerate(train_loader)):\n\n imgs = Variable(images)\n lbls = Variable(labels)\n if torch.cuda.is_available():\n imgs, lbls = imgs.cuda(), lbls.cuda()\n\n # update generator and classifiers by source samples\n optimizer.zero_grad()\n preds = model(imgs)\n if args.net == \"psp\":\n preds = preds[0]\n\n loss = criterion(preds, lbls)\n loss.backward()\n c_loss = loss.data[0]\n epoch_loss += c_loss\n\n optimizer.step()\n\n if ind % 100 == 0:\n print(\"iter [%d] CLoss: %.4f\" % (ind, c_loss))\n\n if ind > args.max_iter:\n break\n\n print(\"Epoch [%d] Loss: %.4f\" % (epoch + 1, epoch_loss))\n log_value('loss', epoch_loss, epoch)\n log_value('lr', args.lr, epoch)\n\n if args.adjust_lr:\n args.lr = adjust_learning_rate(optimizer, args.lr, args.weight_decay, epoch, args.epochs)\n\n if args.net == \"fcn\" or args.net == \"psp\":\n checkpoint_fn = os.path.join(args.pth_dir, \"%s-%s-res%s-%s.pth.tar\" % (\n args.savename, args.net, args.res, epoch + 1))\n else:\n checkpoint_fn = os.path.join(args.pth_dir, \"%s-%s-%s.pth.tar\" % (\n args.savename, args.net, epoch + 1))\n\n args.start_epoch = epoch + 1\n save_dic = {\n 'args': args,\n 'epoch': epoch + 1,\n 'state_dict': model.state_dict(),\n 'optimizer': optimizer.state_dict()\n }\n\n save_checkpoint(save_dic, is_best=False, filename=checkpoint_fn)\n","repo_name":"mil-tokyo/MCD_DA","sub_path":"segmentation/source_trainer.py","file_name":"source_trainer.py","file_ext":"py","file_size_in_byte":5889,"program_lang":"python","lang":"en","doc_type":"code","stars":528,"dataset":"github-code","pt":"81"}
+{"seq_id":"39262031598","text":"'''\n Chef's computer has N GB of free space. \n He wants to save X files, each of size 1 GB and Y files, each of size 2 GB on his computer. \n Will he be able to do so?\n Chef can save all the files on his computer only if the total size of the files is less than or equal to the space available on his computer.\n'''\n\n# Running Loop for Test Cases\nfor t in range(int(input())):\n \n # Taking Input For N = Free Space In Chef's Computer; X = Files Of 1 GB That Chef Want To Save; Y = Files Of 2 GB That Chef Want To Save\n N,X,Y = map(int, input().split())\n \n # If Total Space Occupied By Files < Total Free Space Availabe; Then Chef Can Save All The Files\n EnoughSpace = \"YES\" if (X+(Y*2)) <= N else \"NO\"\n \n # Printing Result\n print(EnoughSpace)\n","repo_name":"masquerade28/Codechef-Solution-Python-300-Series","sub_path":"Enough Space.py","file_name":"Enough Space.py","file_ext":"py","file_size_in_byte":756,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"74368179465","text":"def array_quadruplet(arr, s):\n if len(arr) < 4:\n return []\n\n arr = sorted(arr)\n for i in range(len(arr) - 3): # optimization accounting for # of pointers\n for j in range(i + 1, len(arr) - 2):\n x = j + 1\n y = len(arr) - 1\n remainder = s - arr[i] - arr[j]\n while x < y:\n if x + y > remainder:\n y -= 1\n elif x + y < remainder:\n x += 1\n else:\n return [i, j, x, y]\n\n return []\n\n\nprint(array_quadruplet([2, 7, 4, 0, 9, 5, 1, 3], 20))\n","repo_name":"Royal4224/LeetCode-Practice","sub_path":"array_quadruplet.py","file_name":"array_quadruplet.py","file_ext":"py","file_size_in_byte":601,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"3813843831","text":"import sys\nimport shelve\nimport os\ndef outSave(filename,ls):\n s=shelve.open(filename,writeback=True)\n if 'db' in s:\n s['db']+=(' '.join(ls[0:3])+os.linesep+ls[3]+os.linesep)\n else:\n s['db']=(' '.join(ls[0:3])+os.linesep+ls[3]+os.linesep)\n\n s.close()\ndef output(filename):\n s=shelve.open(filename)\n print(s['db'])\n s.close()\n \ndef operate(N1,op,N2):\n N1=float(N1)\n N2=float(N2)\n s=['+','-','*','/','%','**']\n if op==s[0]:\n return N1+N2\n if op==s[1]:\n return N1-N2\n if op==s[2]:\n return N1*N2\n if op==s[3]:\n return N1/N2\n if op==s[4]:\n return N1%N2\n if op==s[5]:\n return N1**N2\n print(N1,op,N2)\n return 'erorr'\n\nif __name__=='__main__':\n if sys.argv[1]=='print':\n output('data.shl')\n else:\n re=operate(sys.argv[1],sys.argv[2],sys.argv[3])\n ls=sys.argv[1:]\n ls.append(str(re))\n outSave('data.shl',ls)\n print(re)\n\n\n\n\n\n","repo_name":"sdzr/python_for_study","sub_path":"chapter09/practice9_14.py","file_name":"practice9_14.py","file_ext":"py","file_size_in_byte":975,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72929257546","text":"from typing import List\nfrom bisect import bisect_right\n\n\nclass Solution:\n def findContentChildren(self, g: List[int], s: List[int]) -> int:\n g.sort()\n s.sort()\n res = 0\n\n for i in s:\n idx = bisect_right(g, i)\n if idx > res:\n res += 1\n\n return res\n\n\nprint(Solution().findContentChildren([1,2], [1,2,3]))","repo_name":"boorooksus/Algorithm-Study","sub_path":"LeetCode/2회차_new/B82_Assign Cookies2.py","file_name":"B82_Assign Cookies2.py","file_ext":"py","file_size_in_byte":378,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"20317233151","text":"#!/usr/bin/env python\n\n# Author: John Hawkins (jsh) [really@gmail.com]\nimport itertools\nimport logging\nimport colorcet as cc\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport os.path\nimport pandas as pd\nimport re\nimport seaborn as sns\nimport scipy.stats as st\nimport sys\n\nfrom sklearn import decomposition\nfrom sklearn import preprocessing\n\nimport global_config as gcf\n\nimport IPython\n\nlogging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)s %(message)s')\n\nnp.set_printoptions(precision=4, suppress=True)\n\nPREFIX = os.path.splitext(os.path.basename(__file__))[0]\n\n#############\n\ndiffdatafile = os.path.join(gcf.OUTPUT_DIR, 'lib234.diffdata.tsv')\ndiffdata = pd.read_csv(diffdatafile, sep='\\t', header=0)\nsolodatafile = os.path.join(gcf.OUTPUT_DIR, 'lib234.data.tsv')\nsolodata = pd.read_csv(solodatafile, sep='\\t', header=0)\n\n#############\n\ndef widen_groups(data):\n indexed = data.set_index(['variant', 'sample_s', 'sample_e'])\n return indexed.gamma.unstack(level=[1,2])\n\ndef impute_over_nans(data):\n imp = preprocessing.Imputer(strategy='median', axis=1)\n imp.fit(data)\n filled = imp.transform(data)\n return pd.DataFrame(filled, columns=data.columns, index=data.index)\n\nwidediffdata = widen_groups(diffdata)\ncleaned = impute_over_nans(widediffdata)\n\ntaggers = dict()\ntaggers['glob'] = lambda s, e: True\ntaggers['early'] = lambda s, e: (s[1] == '1' and e[1] == '1')\ntaggers['mid'] = lambda s, e: (s[1] == '2' and e[1] == '2')\ntaggers['late'] = lambda s, e: (s[1] == '3' and e[1] == '3')\ntaggers['none'] = lambda s, e: (s[0] == 'a' or s[2:] == 'd1')\ntaggers['low'] = lambda s, e: (s[0] != 'a' and s[2:] == 'd2')\ntaggers['high'] = lambda s, e: (s[0] != 'a' and s[2:] == 'd3')\n\nstarts = cleaned.columns.get_level_values(0)\nends = cleaned.columns.get_level_values(1)\npairs = list(zip(starts, ends))\n\ndef D_from_taggers(names):\n tags = [[taggers[t](s, e) for t in names] for (s, e) in pairs]\n return np.asarray(tags)\n\nD = dict()\nD['exp'] = D_from_taggers(['glob',\n 'early', 'mid', 'late',\n 'none', 'low', 'high'])\nD['dose'] = D_from_taggers(['low', 'high'])\nD['glob'] = D_from_taggers(['glob'])\nD['early'] = D_from_taggers(['early'])\nD['mid'] = D_from_taggers(['mid'])\nD['late'] = D_from_taggers(['late'])\nA = cleaned\n\ndef rebase(A, D):\n U_, s_, Vt_ = np.linalg.svd(D, full_matrices=True)\n rank_ = (~np.isclose(s_, 0)).sum()\n basis_ = U_[:, :rank_]\n return np.dot(A, np.dot(basis_, basis_.T))\n\nA_exp = rebase(A, D['exp'])\nA_exp_nodose = A_exp - rebase(A_exp, D['dose'])\n\naligned = dict()\nfor span in ['early', 'mid', 'late']:\n aligned[span] = rebase(A_exp_nodose, D[span])\n\nglob_projection = rebase(A_exp_nodose, D['glob'])\nfor name, residue in aligned.items():\n relevant = [taggers[name](s, e) for (s, e) in pairs]\n fig = plt.figure(figsize=(6,6))\n g = sns.jointplot(glob_projection[:, relevant][:, 0],\n residue[:, relevant][:, 0],\n s=2, linewidth=0.5, alpha=0.5)\n plt.suptitle(\n 'global vs. aligned of {name} span'.format(**vars()),\n fontsize=16)\n g.set_axis_labels('global', name + ' aligned')\n graphflat = os.path.join(gcf.OUTPUT_DIR, '.'.join([PREFIX, name, 'png']))\n plt.tight_layout()\n logging.info('Writing flat graph to {graphflat}'.format(**vars()))\n plt.savefig(graphflat)\n plt.close()\n\nfor xname, yname in [('early', 'mid'), ('early', 'late'), ('mid', 'late')]:\n xrelevant = [taggers[xname](s, e) for (s, e) in pairs]\n yrelevant = [taggers[yname](s, e) for (s, e) in pairs]\n xrep = aligned[xname][:, xrelevant][:, 0]\n yrep = aligned[yname][:, yrelevant][:, 0]\n fig = plt.figure(figsize=(6,6))\n g = sns.jointplot(xrep, yrep, s=2, linewidth=0.5, alpha=0.5)\n plt.suptitle(\n '{xname} aligned vs. {yname} aligned'.format(**vars()),\n fontsize=16)\n g.set_axis_labels(xname + ' aligned', yname + ' aligned')\n graphflat = os.path.join(gcf.OUTPUT_DIR,\n '.'.join([PREFIX, xname, yname, 'png']))\n plt.tight_layout()\n logging.info('Writing flat graph to {graphflat}'.format(**vars()))\n plt.savefig(graphflat)\n plt.close()\n","repo_name":"traeki/lowficrispri","sub_path":"docs/20171205_lib2_ind/code/20180321.explore_spans.sub.py","file_name":"20180321.explore_spans.sub.py","file_ext":"py","file_size_in_byte":4112,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"10549549942","text":"from bs4 import BeautifulSoup\nimport requests\n\ndef categoryIKEA():\n site = \"https://www.ikea.com/fr/fr/\"\n category = []\n\n page_response = requests.get(site, headers={'User-Agent': 'Mozilla/5.0'})\n page_content = BeautifulSoup(page_response.content, \"html.parser\")\n\n cat = page_content.find(\"div\",{\"class\":\"departmentLinkBlock\"}).findAll(\"li\")\n\n for item in cat:\n urlCat = \"https://www.ikea.com\" + item.find(\"a\").get(\"href\")\n nameCat = item.find(\"a\").text.strip()\n\n category.append({\n 'name':nameCat,\n 'url':urlCat\n })\n\n return category\n\nprint(categoryIKEA())\n\ndef subcategoryIKEA():\n category = categoryIKEA()\n subcat = []\n\n for item in category[0:1]:\n page_response = requests.get(item[\"url\"], headers={'User-Agent': 'Mozilla/5.0'})\n page_content = BeautifulSoup(page_response.content, \"html.parser\")\n try:\n subcategory = page_content.findAll(\"div\",{\"class\":\"visualNavContainer\"})\n for elem in subcategory:\n subcatUrl = \"https://www.ikea.com\" + elem.find(\"a\",{\"class\":\"categoryName\"}).get(\"href\")\n subcatName = elem.find(\"a\",{\"class\":\"categoryName\"}).text\n\n subcat.append({\n \"url\":subcatUrl,\n \"name\":subcatName\n })\n except:\n continue\n\n return subcat\n\n#print(subcategoryIKEA())\n\n\ndef linkProduit():\n subcat = subcategoryIKEA()\n links = []\n\n for item in subcat[0:1]:\n page_response = requests.get(item[\"url\"], headers={'User-Agent': 'Mozilla/5.0'})\n page_content = BeautifulSoup(page_response.content, \"html.parser\")\n\n try:\n post = page_content.find('div',{'id':'productLists'}).findAll('a',{\"class\":\"productLink\"})\n for elem in post:\n url = \"https://www.ikea.com\" + elem.get(\"href\")\n\n links.append({\n 'url':url,\n 'cat':item[\"name\"]\n })\n except:\n continue\n\n return links\n\n#print(linkProduit())\n\n\ndef scrapIKEA(origin):\n site = \"https://www.ikea.com/fr/fr/\"\n products = linkProduit()\n produits = []\n\n for link in products[0:1]:\n page_response = requests.get(link[\"url\"], headers={'User-Agent': 'Mozilla/5.0'})\n page_content = BeautifulSoup(page_response.content, \"html.parser\")\n\n logo = \"\"\n logoS = \"\"\n\n try:\n lib = page_content.find('div',{\"id\":\"productInfoWrapper1\"}).find(\"h1\").text.strip()\n url = link[\"url\"]\n img = \"https://www.ikea.com\" + page_content.find('img',{\"id\":\"productImg\"}).get(\"src\")\n prix = page_content.find('div',{\"id\":\"prodPrice\"}).find(\"span\",{\"id\":\"price1\"}).text.strip()\n desc = page_content.find(\"div\",{\"id\":\"productInfo1\"})\n\n produits.append(\n {\n 'libProduct': lib,\n 'slug': '',\n 'descProduct': desc,\n 'priceProduct': prix,\n 'imgProduct': img,\n 'numSeller': '',\n 'src': site,\n 'urlProduct': url,\n 'logo': logo,\n 'logoS': logoS,\n 'origin': origin,\n \"country\": \"\",\n 'subcategory': link['cat'],\n })\n\n except:\n continue\n\n return produits\n\nprint(scrapIKEA(origin=0))","repo_name":"sysall/WebScrapping","sub_path":"Sites/International/IKEA.py","file_name":"IKEA.py","file_ext":"py","file_size_in_byte":3494,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72248269704","text":"import numpy as np\nimport scipy.stats as stats \nimport math\nimport mysklearn.myutils as myutils\nimport mysklearn.myevaluation as myevaluation\nfrom mysklearn.myclassifiers import MySimpleLinearRegressor, MyKNeighborsClassifier, MyNaiveBayesClassifier, MyDecisionTreeClassifier, MyRandomForestClassifier\n\ninterview_header = [\"level\", \"lang\", \"tweets\", \"phd\", \"interviewed_well\"]\ninterview_table = [\n [\"Senior\", \"Java\", \"no\", \"no\", \"False\"],\n [\"Senior\", \"Java\", \"no\", \"yes\", \"False\"],\n [\"Mid\", \"Python\", \"no\", \"no\", \"True\"],\n [\"Junior\", \"Python\", \"no\", \"no\", \"True\"],\n [\"Junior\", \"R\", \"yes\", \"no\", \"True\"],\n [\"Junior\", \"R\", \"yes\", \"yes\", \"False\"],\n [\"Mid\", \"R\", \"yes\", \"yes\", \"True\"],\n [\"Senior\", \"Python\", \"no\", \"no\", \"False\"],\n [\"Senior\", \"R\", \"yes\", \"no\", \"True\"],\n [\"Junior\", \"Python\", \"yes\", \"no\", \"True\"],\n [\"Senior\", \"Python\", \"yes\", \"yes\", \"True\"],\n [\"Mid\", \"Python\", \"no\", \"yes\", \"True\"],\n [\"Mid\", \"Java\", \"yes\", \"no\", \"True\"],\n [\"Junior\", \"Python\", \"no\", \"yes\", \"False\"]\n ]\n\nM = 7\nN = 20\nF = 2\ninterviewTest = MyRandomForestClassifier(M, N, F)\ninterviewData, interviewClasses = myutils.separate(interview_table, interview_header, \"interviewed_well\")\n\n\ndef test_decision_tree_classifier_fit():\n\n X_tr, X_t, y_tr, y_t = myevaluation.train_test_split(interviewData,interviewClasses)\n interviewTest.fit(X_tr, y_tr)\n assert len(interviewTest.best_M_trees) == M\n\ndef test_decision_tree_classifier_predict():\n results = interviewTest.predict([[\"Junior\", \"Java\", \"yes\", \"no\"], [\"Junior\", \"Java\", \"yes\", \"yes\"]])\n actual = ['True', 'False']\n assert len(results) == len(actual)\n print(str(results) + \" vs \" + str(actual))","repo_name":"caleb-cramer/MLFinalProject","sub_path":"test_myclassifiers.py","file_name":"test_myclassifiers.py","file_ext":"py","file_size_in_byte":1740,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1242655746","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Dec 2 09:30:30 2016\n\n@author: vf140245\nCopyrignt : CEA NeuroSpin - 2014\n\"\"\"\nfrom nilearn import plotting\nimport nibabel\nimport argparse\nimport os\nfrom clindmri.registration.fsl import flirt\nfrom glob import glob\n\nfrom hopla.converter import hopla\n\n# Parameters to keep trace\n__hopla__ = [\"tool\", \"version\"]\n\ndoc = \"\"\"\nHopla context to invoke register_pitie.py. Perform unit registration of \nT2 on T1 image using fsl/flirt command. Use mutual info and trigger sinc resampling\n\nOrganize results according the project rules.\n\nCommand:\n========\n\npython hl_register_pitie.py \\\n -b /volatile/frouin/radiomique_radiogenomique \\\n -r AxT1enhanced \\\n -i AxT2\n \n\npython hl_register_pitie.py \\\n -b /volatile/frouin/radiomique_radiogenomique \\\n -r AxT1enhanced \\\n -i AxT2 \\\n -p\n\n\"\"\"\n\n# Parsing \ndef is_dir(dirarg):\n \"\"\" Type for argparse - checks that output dir exists.\n \"\"\"\n if not os.path.isdir(dirarg):\n raise argparse.ArgumentError(\n \"The dir '{0}' does not exist!\".format(dirarg))\n return dirarg\n\nparser = argparse.ArgumentParser(description=doc)\nparser.add_argument(\n \"-v\", \"--verbose\", dest=\"verbose\", type=int, choices=[0, 1, 2], default=0,\n help=\"increase the verbosity level: 0 silent, [1, 2] verbose.\")\nparser.add_argument(\n \"-b\", \"--basedir\", dest=\"basedir\", required=True, metavar=\"PATH\",\n help=\"the subject main directory to parse files\",\n type=is_dir)\nparser.add_argument(\n \"-r\", \"--ref\", dest=\"refile\", metavar=\"STRING\",\n help=\"Motif (regular expression) to serarch for ref images\")\nparser.add_argument(\n \"-s\", \"--subject\", dest=\"subject\", metavar=\"STRING\",\n help=\"An existing subject in the data tree\")\nparser.add_argument(\n \"-i\", \"--in\", dest=\"infile\", metavar=\"STRING\",\n help=\"Motif (regular expression) to serarch for images to register\")\nparser.add_argument(\n \"-p\", \"--process\", dest=\"process\", action='store_true',\n help=\"if activated execute.\")\nargs = parser.parse_args()\n\n\n# glob the file to analyze\n# list subjects from basedir\n#print glob(os.path.join(args.basedir, '*', args.infile))\n\nif args.subject is not None:\n infiles = set([os.path.dirname(os.path.abspath(i)) for i in \\\n glob(os.path.join(args.basedir, '{}'.format(args.subject), args.infile))])\n refiles = set([os.path.dirname(os.path.abspath(i)) for i in \\\n glob(os.path.join(args.basedir, '{}'.format(args.subject), args.refile))])\nelse:\n infiles = set([os.path.dirname(os.path.abspath(i)) for i in \\\n glob(os.path.join(args.basedir, '*', args.infile))])\n refiles = set([os.path.dirname(os.path.abspath(i)) for i in \\\n glob(os.path.join(args.basedir, '*', args.refile))])\nsubjects = set.intersection(infiles, refiles)\nsubjects = [os.path.basename(i) for i in subjects]\n# get infiles and filter them\ninfiles = []\nrefiles = []\nfor s in subjects:\n infiles.extend(glob(os.path.join(args.basedir, '{}'.format(s),\n args.infile, '{}.nii.gz'.format(args.infile))))\n refiles.extend(glob(os.path.join(args.basedir, '{}'.format(s),\n args.refile, '{}.nii.gz'.format(args.refile))))\noutfiles = [i.replace(args.basedir, os.path.join(args.basedir, 'preprocess')) for i in infiles]\noutfiles = [i.replace(args.infile,'',1) for i in outfiles]\noutfiles = [os.path.splitext(os.path.splitext(i)[0])[0] for i in outfiles]\noutfiles = [os.path.join('{}'.format(os.path.dirname(i)),\n 'r{}.nii.gz'.format(os.path.basename(i)))\n for i in outfiles]\n\nif args.process:\n logfile = \"{}/preprocess/log.txt\".format(args.basedir)\n if not os.path.isdir(os.path.dirname(logfile)):\n os.makedirs(os.path.dirname(logfile))\n #\n status, exitcodes = hopla(\n os.path.join('/volatile','frouin','radiomique_radiogenomique',\n 'register_pitie.py'),\n i=infiles,\n r=refiles,\n o=outfiles,\n hopla_iterative_kwargs=[\"i\", \"r\", \"o\"],\n hopla_cpus=2,\n hopla_logfile=logfile,\n hopla_verbose=1)\n","repo_name":"neurospin/scripts","sub_path":"2017_rr/metastasis/hl_register_pitie.py","file_name":"hl_register_pitie.py","file_ext":"py","file_size_in_byte":4204,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"73369155465","text":"import json\n\nimport httpx\nfrom fastapi import status\n\nfrom marvin.loaders.web import SitemapLoader, URLLoader\nfrom marvin.plugins import Plugin\nfrom marvin.utilities.strings import html_to_content, slice_tokens\nfrom marvin.utilities.web import url_is_ok\n\n\nclass VisitURL(Plugin):\n name: str = \"visit-url\"\n description: str = (\n \"Visit a URL and return its contents. Don't provide a URL unless you're\"\n \" absolutely sure it exists.\"\n )\n\n async def run(self, url: str) -> str:\n if not url.startswith(\"http\"):\n url = f\"http://{url}\"\n async with httpx.AsyncClient(follow_redirects=True, timeout=2) as client:\n try:\n response = await client.get(url)\n except httpx.ConnectTimeout:\n return \"Failed to load URL: Connection timed out\"\n if response.status_code == status.HTTP_200_OK:\n text = response.text\n\n # try to parse as JSON in case the URL is an API\n try:\n content = str(json.loads(text))\n # otherwise parse as HTML\n except json.JSONDecodeError:\n content = html_to_content(text)\n return slice_tokens(content, 1000)\n else:\n return f\"Failed to load URL: {response.status_code}\"\n\n\nclass LoadAndStoreURL(Plugin):\n name: str = \"load-and-store-url\"\n description: str = (\n \"Visit a URL and use a loader to load its contents. Don't provide a URL unless\"\n \" you're absolutely sure it exists. A topic name can be provided to store the\"\n \" documents in a particular topic.\"\n )\n\n async def run(self, url: str, topic_name: str = None) -> str:\n \"\"\"\n Load and store the contents of a URL into a topic. If no topic name is provided,\n the Marvin default topic will be used.\n \"\"\"\n if not url.startswith(\"http\"):\n url = f\"http://{url}\"\n\n if url.endswith(\".pdf\"):\n return f\"URL {url} is a PDF. Use the `load-and-store-pdf` plugin instead.\"\n if url.endswith(\".xml\"):\n return (\n f\"URL {url} is a sitemap. Use the `load-and-store-sitemap` plugin\"\n \" instead.\"\n )\n\n if not await url_is_ok(url):\n return (\n \"URL was not reachable - make sure it exists and is publicly accessible\"\n )\n\n loader = URLLoader(urls=[url])\n await loader.load_and_store(topic_name=topic_name)\n return f\"Loaded {url} into topic {topic_name!r}\"\n\n\nclass LoadAndStorePDF(Plugin):\n name: str = \"load-and-store-pdf\"\n description: str = (\n \"Load and store the contents of a PDF URL into a topic. if no topic name is\"\n \" provided, the Marvin default topic will be used.\"\n )\n\n async def run(self, pdf_url: str, topic_name: str = None) -> str:\n from marvin.loaders.pdf import PDFLoader\n\n loader = PDFLoader(file_path=pdf_url)\n await loader.load_and_store(topic_name=topic_name)\n return f\"Loaded {pdf_url} into topic {topic_name!r}\"\n\n\nclass LoadAndStoreSitemap(Plugin):\n name: str = \"load-and-store-sitemap\"\n description: str = (\n \"Load and store the contents of a sitemap URL into a topic. if no topic name is\"\n \" provided, the Marvin default topic will be used.\"\n )\n\n async def run(self, sitemap_url: str, topic_name: str = None) -> str:\n loader = SitemapLoader(urls=[sitemap_url])\n await loader.load_and_store(topic_name=topic_name)\n return f\"Loaded {sitemap_url} into topic {topic_name!r}\"\n","repo_name":"rehmatsg/StarfishGPT","sub_path":"venv/lib/python3.10/site-packages/marvin/plugins/web.py","file_name":"web.py","file_ext":"py","file_size_in_byte":3576,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"6703102806","text":"# edges = [['A','B'],['A','C'],['A','D'],['D','E'],['B','E']]\n# nodes = ['A','B','C','D','E']\nedges = [['A','B'],['A','C'],['A','D'],['D','E'],\n ['B','E'],['F','H'],['F','G'],['I','J']]\nnodes = ['A','B','C','D','E','F','G','H','I','J','K']\n\n# def dfs(graph,node,visited = set()):\n# print(node)\n# visited.add(node)\n# for child in graph[node]:\n# if child not in visited:\n# dfs(graph,child,visited) \n\n\n# def no_of_node(graph,node,visited = set()):\n# # print(node)\n# visited.add(node)\n# sm = 0\n# for child in graph[node]:\n# if child not in visited:\n# sm += no_of_node(graph,child,visited) \n# return sm+1\n\ndef dfs(graph,node,visited = set()):\n print(node)\n visited.add(node)\n sm = 0\n for child in graph[node]:\n if child not in visited:\n sm += dfs(graph,child,visited) \n return sm+1\n\ngraph = {}\nfor i in nodes:\n graph[i] = []\n\nfor (u,v) in edges:\n graph[u].append(v) \n graph[v].append(u)\n\nfor i in graph.items():\n print(i) \n\n# dfs(graph,'A') \n# print(no_of_node(graph,'A'))\nanswer = []\nvisited = set()\nfor item in nodes:\n if item not in visited:\n temp = dfs(graph,item,visited)\n answer.append(temp)\n\nprint(answer)","repo_name":"AseemGupta39/Graph_Data_Structure_in_Python","sub_path":"connected_graph.py","file_name":"connected_graph.py","file_ext":"py","file_size_in_byte":1259,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"28252926921","text":"def desenha_lista(lista, marcar_meio, meio, exibir_de, exibir_ate):\n for index,x in enumerate(lista):\n if(index > exibir_ate):\n continue\n if(index < exibir_de):\n print(\" \", end=\"\")\n elif(marcar_meio == True and x == meio):\n print(\"+=====\", end=\"\")\n else:\n print(\"+-----\", end=\"\")\n print(\"+\")\n for index,x in enumerate(lista):\n if(index > exibir_ate):\n continue\n if(index < exibir_de):\n print(\" \", end=\"\")\n elif(marcar_meio == True and x == meio):\n print(\"||{}|\".format(x), end=\"\")\n else:\n print(\"| {} \".format(x), end=\"\")\n print(\"|\")\n for index,x in enumerate(lista):\n if(index > exibir_ate):\n continue\n if(index < exibir_de):\n print(\" \", end=\"\")\n elif(marcar_meio == True and x == meio):\n print(\"+=====\", end=\"\")\n else:\n print(\"+-----\", end=\"\")\n print(\"+\")\n\ndef busca_binaria(lista,elemento):\n posicao_primeiro_elem = 0\n posicao_ultimo_elem =len(lista)-1\n posicao_meio = (posicao_primeiro_elem+posicao_ultimo_elem)//2\n desenha_lista(lista,False,\"\",posicao_primeiro_elem,posicao_ultimo_elem)\n if(lista != sorted(lista)):\n return -2\n else:\n desenha_lista(lista,True,lista[posicao_meio],posicao_primeiro_elem,posicao_ultimo_elem)\n while(posicao_ultimo_elem >= posicao_primeiro_elem):\n if(lista[posicao_meio] == elemento):\n print(\"Encontrado na posicao: {}\".format(posicao_meio))\n return True\n else:\n if elemento < lista[posicao_meio]:\n posicao_ultimo_elem = posicao_meio - 1\n else:\n posicao_primeiro_elem = posicao_meio + 1\n posicao_meio = (posicao_primeiro_elem+posicao_ultimo_elem)//2\n desenha_lista(lista, True, lista[posicao_meio],posicao_primeiro_elem,posicao_ultimo_elem)\n if(posicao_primeiro_elem == posicao_ultimo_elem and lista[posicao_meio] != elemento):\n return False\n return False\n\n\n\n\n\nelemento = input().zfill(3)\nlista = [e.zfill(3) for e in input().split()]\n\nprint(\"Elemento procurado: {}\".format(elemento))\n\nresultado = busca_binaria(lista, elemento)\n\nif resultado == False:\n print(\"O elemento nao foi encontrado\")\nelif resultado == -2:\n print(\"Lista nao ordenada\")\n","repo_name":"matheusconceicao7/mc102z","sub_path":"Lab13/lab13.py","file_name":"lab13.py","file_ext":"py","file_size_in_byte":2444,"program_lang":"python","lang":"pt","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"13810252444","text":"from flask import Blueprint\nfrom api.controllers.server_controller import ServerController\n\nserver_bp = Blueprint ('server_bp', __name__)\n\nserver_bp.route('/server/', methods = ['GET'])(ServerController.get_server)\nserver_bp.route('/servers', methods = ['GET'])(ServerController.get_servers)\nserver_bp.route('/create_server', methods = ['POST'])(ServerController.create_server)\nserver_bp.route('/update_server/', methods = ['PUT'])(ServerController.update_server)\nserver_bp.route('/delete_server/', methods = ['DELETE'])(ServerController.delete_server)\nserver_bp.route('/search/', methods = ['GET'])(ServerController.search_servers)","repo_name":"Llane4/Proyecto-Backend","sub_path":"api/routes/server_bp.py","file_name":"server_bp.py","file_ext":"py","file_size_in_byte":697,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12147735607","text":"from rest_framework import serializers\nfrom .models import Overview\nimport ipdb\n\n\nclass OverviewSerializer(serializers.ModelSerializer):\n bmi_classification = serializers.SerializerMethodField()\n\n whr_classification = serializers.SerializerMethodField()\n\n def get_bmi_classification(self, obj):\n if int(obj.bmi) < 18.5:\n return 'underweight'\n\n if int(obj.bmi) >= 18.5 and int(obj.bmi) < 25:\n return 'normal'\n\n if int(obj.bmi) >= 25 and int(obj.bmi) < 30:\n return 'overweight'\n\n if int(obj.bmi) >= 30 and int(obj.bmi) < 35:\n return 'obese'\n\n if int(obj.bmi) >= 35:\n return 'extremely obese'\n \n\n def get_whr_classification(self, obj):\n if obj.sex == 'female' and int(obj.whr) <= 0.80:\n return 'low'\n\n if obj.sex == 'female' and int(obj.whr) > 0.80 and int(obj.whr) <= 0.85:\n return 'moderate'\n\n if obj.sex == 'female' and int(obj.whr) > 0.85:\n return 'high'\n \n if obj.sex == 'male' and int(obj.whr) <= 0.95:\n return 'low'\n\n if obj.sex == 'male' and int(obj.whr) > 0.95 and int(obj.whr) < 1.0:\n return 'moderate'\n\n if obj.sex == 'male' and int(obj.whr) >= 1.0:\n return 'high'\n\n \n\n class Meta:\n model = Overview\n\n fields = [\n 'id',\n 'sex',\n 'height',\n 'weight',\n 'bmi',\n 'bmi_classification',\n 'whr',\n 'whr_classification',\n 'lean_mass',\n 'fat_mass',\n 'body_fat',\n 'created_at',\n 'created_by'\n ]\n\n read_only_fields = [\n 'bmi_classification',\n 'whr_classification',\n 'created_at',\n 'created_by'\n ]\n\n\n def create(self, validated_data: dict) -> Overview:\n return Overview.objects.create(**validated_data)\n","repo_name":"FitnessFormula/fitness-formula-api","sub_path":"overview/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":1966,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"8965769579","text":"import django_filters\r\nfrom django_filters import CharFilter, NumberFilter\r\nfrom .models import *\r\n\r\n\r\nclass ProductFilter(django_filters.FilterSet):\r\n name = CharFilter(field_name='name',\r\n lookup_expr='icontains', label='what are you looking for?')\r\n\r\n class Meta:\r\n model = Product\r\n fields = '__all__'\r\n exclude = ['description', 'price', 'image', 'vendor','category']\r\n","repo_name":"Kihara-Njoroge/tronic-electronics","sub_path":"shop/filters.py","file_name":"filters.py","file_ext":"py","file_size_in_byte":422,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"70392436746","text":"from __future__ import division\nimport matplotlib\nimport numpy as np \nimport time\nimport math\nimport numpy as np\nfrom numpy.lib.npyio import savetxt\nimport scipy.io\nfrom numpy.core.numeric import NaN\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\nfrom matplotlib import rcParams\nimport matplotlib.animation\n\nfrom numpy import ones,vstack\nfrom numpy.linalg import lstsq\n\nplt.close('all')\ndata_array = np.load('StandingTestData.npy')\n\nend = time.time()\n\n#print(\"\\nData summary:\\n\", data_array)\nprint(\"\\nData shape:\\n\", data_array.shape)\n\n\n\n# plt.xlim([-600,0])\n# plt.ylim([600,2500])\ni = 0\nskips = 0\nfor i in range(213):\n \n for g in range(10,17):\n \n data_array2 = data_array[:,g:(g+2),:]\n\n def count_nans(nans, one, two):\n \n count = 0 \n z = -3\n for p in range(6):\n print(\"count\", count)\n print(\"nans: \", nans)\n print(\"i\", i)\n print(\"i+z\", (i+z))\n if np.isnan(nans):\n count = count + 1\n nans = data_array2[(i+z), one, two]\n \n z = z+1\n print(\"count\", count)\n return count\n\n\n\n def replace_nans(cord, first, second, j):\n if np.isnan(cord):\n total_nans = count_nans(cord, first, second)\n if (total_nans > 4):\n cord = 0\n pass\n else:\n z = -3\n new_cord = 0\n for p in range(6):\n cord2 = data_array2[(j+z), first, second]\n if np.isnan(cord2):\n pass\n else:\n \n new_cord = new_cord + cord2\n print(\"cord+cord2\", new_cord)\n cord = new_cord\n\n z = z+1\n cord = cord / (6 - total_nans)\n print(\"end of replace nans: \", cord)\n \n\n\n \n return cord\n\n\n\n\n \n \n X2 = data_array2[i, 0, 0]\n X2 = replace_nans(X2, 0,0, i)\n print(\"X2\", X2)\n\n Y2 = data_array2[i, 0, 1]\n Y2 = replace_nans(Y2, 0,1, i)\n print(\"Y2\", Y2)\n\n X3 = data_array2[i, 1, 0]\n X3 = replace_nans(X3, 1,0, i)\n print(\"X3\", X3)\n Y3 = data_array2[i, 1, 1]\n Y3 = replace_nans(Y3, 1,1, i)\n print(\"Y3\", Y3)\n\n points = [(X2,Y2),(X3,Y3)]\n x_coords, y_coords = zip(*points)\n A = vstack([x_coords,ones(len(x_coords))]).T\n m, c = lstsq(A, y_coords)[0]\n print(\"Line Solution is y = {m}x + {c}\".format(m=m,c=c))\n\n\n\n X = X2 - ((X2-X3) *.8) \n Y = Y2 - ((Y2-Y3) *.8)\n\n\n print(\"X\", X)\n print(\"Y\", Y)\n\n print(\" \")\n\n\n plt.axes()\n \n plt.xlim(-500,500)\n plt.ylim(-500,500)\n\n scatterplot1 = plt.plot(X2, Y2, 'bo')\n scatterplot1 = plt.plot(X3, Y3, 'bo')\n scatterplot1 = plt.plot(X, Y, 'ro')\n \n \n \n#plt.show(scatterplot1)\n\nplt.close('all')\n\n\n\n\n\n\n\n\n\n\nprint('skips', skips)\n\n\n\n\n\n","repo_name":"rhettd33/FreemocapCOM-","sub_path":"freemocaptest5.py","file_name":"freemocaptest5.py","file_ext":"py","file_size_in_byte":3268,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"15971164415","text":"from PIL import Image\nfrom src.models.cc0 import patcher\nimport numpy as np\nimport skimage.io as io\nfrom src.utils.imgproc import *\nfrom skimage.color import rgb2hsv, hsv2rgb\n\n\nclass patcher(patcher):\n def __init__(self, body='./body/body_sakurana.png', **options):\n try:\n options = options['options']\n except:\n pass\n options['is_4k'] = False\n super().__init__(options=options)\n self.name = 'サクラナ'\n self.body = Image.open(body)\n self.body_size = self.body.size\n self.pantie_position = [3433, 1782]\n try:\n self.with_bra = options['with_bra']\n except:\n self.with_bra = self.ask(question='With bra?', default=True)\n if self.with_bra:\n self.bra_position = [2017, 1491]\n self.bra = np.float32(io.imread('./mask/bra_sakurana.png') / 255)\n self.bra_center = np.float32(io.imread('./mask/bra_sakurana_center.png') / 255)\n self.bra_shade = np.float32(io.imread('./material/bra_sakurana_shade.png') / 255)\n self.bra_frill = np.float32(io.imread('./material/bra_sakurana_frill.png') / 255)\n self.bra_shade_alpha = self.bra_shade[:, :, -1]\n self.bra_frill_mask = self.bra_frill[:, :, -1] > 0.5\n\n def gen_bra(self, image):\n def pick_color(arr):\n return np.mean(np.mean(arr, axis=0), axis=0)\n pantie = np.array(image)\n\n # pickup colors\n front = pantie[20:100, 30:80, :3] / 255.0\n front_shade = pantie[130:150, 0:40, :3] / 255.0\n front_color = pick_color(front)\n front_shade_color = pick_color(front_shade)\n front_shade_color = rgb2hsv(front_shade_color[None, None])\n front_shade_color[0, 0, 1] *= front_shade_color[0, 0, 2] / 0.3\n if front_shade_color[0, 0, 1] > 0.7:\n front_shade_color[0, 0, 1] *= 0.7\n front_shade_color[0, 0, 2] *= front_shade_color[0, 0, 2] / 0.4\n front_shade_color = np.clip(hsv2rgb(front_shade_color)[0, 0], 0, 1)\n ribbon = pantie[24:32, 15:27, :3] / 255.0\n ribbon_color = pick_color(ribbon)\n\n # making a center texture\n center = pantie[20:170, -200:-15, :3][:, ::-1]\n center = resize(center, [2.3, 2.5])\n\n bra_center = np.copy(self.bra_center)\n bra_center[:center.shape[0], :center.shape[1], :3] = center * np.float32(bra_center[:center.shape[0], :center.shape[1], :3] > 0)\n bra = self.bra[:, :, :3] * front_color\n bra_shade = (self.bra_shade[:, :, -1])[:, :, None] * front_shade_color\n bra_frill = self.bra_frill[:, :, :3] * ribbon_color\n\n # overlaying layers\n bra = alpha_brend(bra_center[:, :, :3], bra[:, :, :3], bra_center[:, :, 0] > 0.1)\n bra = alpha_brend(bra_frill, bra, self.bra_frill_mask)\n bra = alpha_brend(bra_shade, bra, self.bra_shade_alpha)\n bra = np.dstack((bra, self.bra[:, :, 0] > 0))\n return Image.fromarray(np.uint8(np.clip(bra, 0, 1) * 255))\n\n def patch(self, image, transparent=False):\n pantie = self.convert(image)\n if transparent:\n patched = Image.new(\"RGBA\", (4096, 4096))\n else:\n patched = self.body.copy()\n pantie = pantie.resize((int(pantie.width * .75), int(pantie.height * .77)), resample=Image.BICUBIC)\n pantie = pantie.rotate(-90, expand=True)\n patched = self.paste(patched, pantie, self.pantie_position)\n if self.with_bra:\n patched = self.paste(patched, self.gen_bra(image), self.bra_position)\n return patched\n","repo_name":"TenteEEEE/quiche_pantie_patch","sub_path":"src/models/sakurana.py","file_name":"sakurana.py","file_ext":"py","file_size_in_byte":3575,"program_lang":"python","lang":"en","doc_type":"code","stars":80,"dataset":"github-code","pt":"81"}
+{"seq_id":"39204293436","text":"\"\"\"\n!!!DISCLAIMER!!!\nALL INFORMATION CONTAINED IN THIS FILE HAS NOTHING TO DO WITH REAL LIFE.\nALL CHARACTERS DESCRIBED HERE ARE FICTIONARY.\nANY AND ALL SIMILARITIES ARE COMPLETELY COINCIDENTAL.\n\"\"\"\n\nEN_LIBERTY_WORDS = [\"liberty\", \"rights\", \"freedom\", \"choice\", \"will\"]\nEN_LIBERTY_OFFENSIVE_WORDS = [\"liberals\", \"popped\"]\n\nEN_AUTHORITY_WORDS = [\"authority\", \"laws\", \"governance\", \"guidance\", \"control\",\n \"submission\"]\nEN_AUTHORITY_OFFENSIVE_WORDS = [\"opressors\", \"enslavers\", \"megalomaniac\", \"bootlicker\", \"sheeple\"]\n\nEN_PACIFISM_WORDS = [\"pacifism\", \"peace\", \"peacemaker\", \"understanding\", \"tolerance\",\n \"agreements\", \"friendship\"]\nEN_PACIFISM_OFFENSIVE_WORDS = [\"peacemonger\", \"coward\", \"spineless\"]\n\nEN_MILITARISM_WORDS = [\"militarism\", \"war\", \"defense\", \"assault\", \"weapon\",\n \"protection\", \"enemies\", \"attack\", \"capture\", \"bloody\"]\nEN_MILITARISM_OFFENSIVE_WORDS = [\"warmonger\", \"bloodletter\"]\n\nEN_INTENSIVISM_WORDS = [\"intensivism\", \"distribution\", \"economy\", \"efficency\", \"management\"]\nEN_INTENSIVISM_OFFENSIVE_WORDS = [\"scrooge\"]\n\nEN_EXTENSIVISM_WORDS = [\"extensivism\", \"acquisition\", \"resources\", \"accession\", \"expansion\"]\nEN_EXTENSIVISM_OFFENSIVE_WORDS = [\"wastrel\"]\n\n# EN_MATERIALISM_WORDS = [\"materialism\", \"prosperity\", \"resources\", \"acquisition\", \"science\",\n# \"education\", \"cognition\", \"naturalness\", \"universe\", \"objective\"]\n# EN_MATERIALISM_OFFENSIVE_WORDS = [\"scrooge\", \"faithless\"]\n\n# EN_SPIRITUALISM_WORDS = [\"spiritualism\", \"deity\", \"worship\", \"extrasensory\", \"insight\",\n# \"faith\", \"religion\", \"prophecy\", \"spirit\", \"paranormal\"]\n# EN_SPIRITUALISM_OFFENSIVE_WORDS = [\"fanatic\", \"zealot\", \"bigot\"]\n\nEN_INDIVIDUALISM_WORDS = [\"individualism\", \"personality\", \"uniqueness\", \"speciality\", \"eccentric\",\n \"expression\", \"entrepreneurship\", \"private\"]\nEN_INDIVIDUALISM_OFFENSIVE_WORDS = [\"egocentric\", \"snowflake\"]\n\nEN_COLLECTIVISM_WORDS = [\"collectivism\", \"unity\", \"gathering\", \"coherent\", \"collectively\",\n \"together\", \"union\", \"alliance\", \"public\"]\nEN_COLLECTIVISM_OFFENSIVE_WORDS = [\"crowd\", \"herd\"]\n\nEN_REFORMISM_WORDS = [\"reformism\", \"legislation\", \"reform\", \"stability\", \"discussion\",\n \"compromise\", \"gradually\"]\nEN_REFORMISM_OFFENSIVE_WORDS = [\"indecisive\"]\n\nEN_REVOLUTIONISM_WORDS =[\"revolutionism\", \"negation\", \"coup\", \"change\", \"riot\",\n \"protest\", \"abruptly\", \"revolution\"]\nEN_REVOLUTIONISM_OFFENSIVE_WORDS = [\"reactionery\"]\n\nEN_CONSTRUCTIVISM_WORDS = [\"constructivism\", \"industry\", \"goal\", \"artificial\", \"crafted\",\n \"custom\", \"synthesis\", \"elaborate\"]\nEN_CONSTRUCTIVISM_OFFENSIVE_WORDS = [\"geek\", \"pervert\"]\n\nEN_ESSENTIALISM_WORDS = [\"essentialism\", \"nature\", \"essence\", \"natural\", \"genuine\",\n \"normal\", \"life\", \"simple\"]\nEN_ESSENTIALISM_OFFENSIVE_WORDS = [\"hillbilly\", \"retrograde\"]\n\nEN_DENIAL_WORDS = [\"denial\", \"absence\", \"void\", \"nothing\", \"non-existance\", \"non-conformity\", \"wordly\", \"destructive\"]\nEN_DENIAL_OFFENSIVE_WORDS = [\"cynic\", \"degenerate\", \"unscrupulous\"]\n\nEN_ACCEPTANCE_WORDS = [\"acceptance\", \"presence\", \"embrace\", \"everything\", \"existance\", \"conformity\", \"sacred\", \"constructive\"]\nEN_ACCEPTANCE_OFFENSIVE_WORDS = [\"idealist\"]\n\n# EN_HEDONISM_WORDS = [\"hedonism\", \"lust\", \"joy\", \"excess\", \"pleasure\", \"happiness\"]\n# EN_HEDONISM_OFFENSIVE_WORDS = [\"deviant\", \"snob\", \"nymphomaniac\", \"sybarite\", \"glutton\", \"chad\"]\n\n# EN_ASCETICISM_WORDS = [\"asceticism\", \"purity\", \"moderation\", \"abstinence\", \"cleanliness\", \"restraint\"]\n# EN_ASCETICISM_OFFENSIVE_WORDS = [\"prude\", \"virgin\"]\n\n\n\nRU_LIBERTY_WORDS = [\"свобода\", \"права\", \"воля\", \"выбор\", \"желания\"]\nRU_LIBERTY_OFFENSIVE_WORDS = [\"либерахи\", \"порвало\"]\n\nRU_AUTHORITY_WORDS = [\"власть\", \"законы\", \"руководство\", \"управление\", \"контроль\",\n \"подчинение\"]\nRU_AUTHORITY_OFFENSIVE_WORDS = [\"угнетатели\", \"рабовладельцы\", \"мегаломан\", \"овечка\"]\n\nRU_PACIFISM_WORDS = [\"пацифизм\", \"мир\", \"миротворец\", \"понимание\", \"терпимость\",\n \"соглашения\", \"дружба\"]\nRU_PACIFISM_OFFENSIVE_WORDS = [\"миролюб\", \"трусишка\", \"мягкотелый\"]\n\nRU_MILITARISM_WORDS = [\"милитаризм\", \"война\", \"защита\", \"нападение\", \"оружие\",\n \"оборона\", \"враги\", \"атака\", \"захват\", \"кровавый\"]\nRU_MILITARISM_OFFENSIVE_WORDS = [\"варвар\", \"кровопроливец\"]\n\nRU_INTENSIVISM_WORDS = [\"интенсивизм\", \"распределение\", \"экономия\", \"эффективность\", \"менеджмент\"]\nRU_INTENSIVISM_OFFENSIVE_WORDS = [\"скряга\"]\n\nRU_EXTENSIVISM_WORDS = [\"экстенсивизм\", \"добыча\", \"ресурсы\", \"доступ\", \"экспансия\"]\nRU_EXTENSIVISM_OFFENSIVE_WORDS = [\"транжира\"]\n\n# RU_MATERIALISM_WORDS = [\"материализм\", \"процветание\", \"ресурсы\", \"добыча\", \"наука\",\n# \"образование\", \"познание\", \"естественность\", \"вселенная\", \"объективный\"]\n# RU_MATERIALISM_OFFENSIVE_WORDS = [\"скряга\", \"безбожник\"]\n\n# RU_SPIRITUALISM_WORDS = [\"спиритуализм\", \"божество\", \"поклонение\", \"эзотерика\", \"прозрение\",\n# \"вера\", \"религия\", \"пророчество\", \"дух\", \"сверхестественность\"]\n# RU_SPIRITUALISM_OFFENSIVE_WORDS = [\"фанатик\", \"изувер\", \"мракобес\"]\n\nRU_INDIVIDUALISM_WORDS = [\"индивидуализм\", \"личность\", \"уникальность\", \"эксцентричный\",\n \"самовыражение\", \"предпринимательство\", \"частный\"]\nRU_INDIVIDUALISM_OFFENSIVE_WORDS = [\"эгоцентрист\", \"снежинка\"]\n\nRU_COLLECTIVISM_WORDS = [\"коллективизм\", \"единство\", \"собрание\", \"согласованность\", \"коллективно\",\n \"вместе\", \"объединение\", \"союз\", \"публично\"]\nRU_COLLECTIVISM_OFFENSIVE_WORDS = [\"толпа\", \"стая\"]\n\nRU_REFORMISM_WORDS = [\"реформизм\", \"законопроект\", \"реформа\", \"стабильность\", \"обсуждение\",\n \"компромисс\", \"постепенно\"]\nRU_REFORMISM_OFFENSIVE_WORDS = [\"нерешительный\"]\n\nRU_REVOLUTIONISM_WORDS =[\"революционизм\", \"переговоры\", \"переворот\", \"перемены\", \"восстание\",\n \"протест\", \"резко\", \"революция\"]\nRU_REVOLUTIONISM_OFFENSIVE_WORDS = [\"реакционер\"]\n\nRU_CONSTRUCTIVISM_WORDS = [\"конструктивизм\", \"урбанизация\", \"эффективность\", \"индустрия\", \"бизнес\"]\nRU_CONSTRUCTIVISM_OFFENSIVE_WORDS = [\"задрот\", \"извращенец\"]\n\nRU_ESSENTIALISM_WORDS = [\"эссенциализм\", \"природа\", \"экология\", \"животные\", \"вымирание\"]\nRU_ESSENTIALISM_OFFENSIVE_WORDS = [\"деревенщина\", \"ретроград\"]\n\nRU_DENIAL_WORDS = [\"отрицание\", \"отсутствие\", \"пустота\", \"ничего\", \"несуществование\", \"нонконформизм\", \"мирской\", \"деструктивный\"]\nRU_DENIAL_OFFENSIVE_WORDS = [\"циник\", \"дегенерат\", \"бессовестный\"]\n\nRU_ACCEPTANCE_WORDS = [\"принятие\", \"присутствие\", \"принимание\", \"всё\", \"существование\", \"конформизм\", \"сакральный\", \"конструктивный\"]\nRU_ACCEPTANCE_OFFENSIVE_WORDS = [\"идеалист\", \"мракобес\"]\n\n# RU_HEDONISM_WORDS = [\"гедонизм\", \"блуд\", \"радость\", \"излишество\", \"удовольствие\", \"счастье\"]\n# RU_HEDONISM_OFFENSIVE_WORDS = [\"девиант\", \"сноб\", \"нимфоман\", \"сибарит\", \"чревоугодец\", \"чад\"]\n\n# RU_ASCETICISM_WORDS = [\"аскетизм\", \"непорочность\", \"скромность\", \"воздержание\", \"чистота\", \"сдержанность\"]\n# RU_ASCETICISM_OFFENSIVE_WORDS = [\"святоша\", \"девственник\"]\n\n\n\nLA_LIBERTY_WORDS = [\"libertas\", \"juris\", \"autem\", \"delectu\", \"voluntatis\"]\nLA_LIBERTY_OFFENSIVE_WORDS = [\"liberacha\", \"rumpere\"]\n\nLA_AUTHORITY_WORDS = [\"imperium\", \"legis\", \"administratio\", \"curatio\", \"custodia\",\n \"dominatio\"]\nLA_AUTHORITY_OFFENSIVE_WORDS = [\"oppressionici\", \"servipossessores\", \"magnautorus\", \"ovila\"]\n\nLA_PACIFISM_WORDS = [\"pacifismus\", \"pax\", \"pacificatorum\", \"intelligentia\", \"tolerantia\",\n \"pacti\", \"amicitia\"]\nLA_PACIFISM_OFFENSIVE_WORDS = [\"pacificarus\", \"timidus\", \"molliformis\"]\n\nLA_MILITARISM_WORDS = [\"militarismus\", \"bellum\", \"defensio\", \"aggressura\", \"arma\",\n \"protectio\", \"hostibus\", \"impetum\", \"occupatio\", \"crudus\"]\nLA_MILITARISM_OFFENSIVE_WORDS = [\"barbarus\", \"suco\"]\n\nLA_INTENSIVISM_WORDS = [\"интенсивизм\", \"распределен��е\", \"экономия\", \"эффективность\", \"менеджмент\"]\nLA_INTENSIVISM_OFFENSIVE_WORDS = [\"скряга\"]\n\nLA_EXTENSIVISM_WORDS = [\"экстенсивизм\", \"добыча\", \"ресурсы\", \"доступ\", \"экспансия\"]\nLA_EXTENSIVISM_OFFENSIVE_WORDS = [\"транжира\"]\n\n# LA_MATERIALISM_WORDS = [\"materialismus\", \"prosperitas\", \"copiae\", \"extraction\", \"scientia\",\n# \"educatio\", \"cognitio\", \"naturalitas\", \"universum\", \"objectivus\"]\n# LA_MATERIALISM_OFFENSIVE_WORDS = [\"trahax\", \"impius\"]\n\n# LA_SPIRITUALISM_WORDS = [\"spiritualismus\", \"divinitatem\", \"idololatria\", \"immaterialismus\", \"viderio\",\n# \"fides\", \"religio\", \"divinaria\", \"spiritus\", \"extranaturalis\"]\n# LA_SPIRITUALISM_OFFENSIVE_WORDS = [\"fanaticum\", \"falsus\", \"cacodaemon\"]\n\nLA_INDIVIDUALISM_WORDS = [\"individualismus\", \"persona\", \"unictatem\", \"eccentricus\",\n \"auto-expressio\", \"mercatorum\", \"privatus\"]\nLA_INDIVIDUALISM_OFFENSIVE_WORDS = [\"egocentriista\", \"nixilla\"]\n\nLA_COLLECTIVISM_WORDS = [\"collegiismus\", \"unitas\", \"conventus\", \"consensus\", \"collegialis\",\n \"simul\", \"complexio\", \"unio\", \"vulgo\"]\nLA_COLLECTIVISM_OFFENSIVE_WORDS = [\"vulgus\", \"turba\"]\n\nLA_REFORMISM_WORDS = [\"reformationismus\", \"rogatio\", \"reformatio\", \"stabilitas\", \"discussio\",\n \"compromissum\", \"gradatim\"]\nLA_REFORMISM_OFFENSIVE_WORDS = [\"cunctator\"]\n\nLA_REVOLUTIONISM_WORDS =[\"revolutionismus\", \"actiones\", \"eversio\", \"mutationis\", \"tumultus\",\n \"interdictum\", \"acriter\", \"revolutio\"]\nLA_REVOLUTIONISM_OFFENSIVE_WORDS = [\"reactiator\"]\n\nLA_CONSTRUCTIVISM_WORDS = [\"constructivismus\", \"urbanisatio\", \"effectivus\", \"industria\", \"negotium\"]\nLA_CONSTRUCTIVISM_OFFENSIVE_WORDS = [\"technomator\", \"извращенец\"]\n\nLA_ESSENTIALISM_WORDS = [\"essentialismus\", \"natura\", \"oecologia\", \"animali\", \"exstinctionem\"]\nLA_ESSENTIALISM_OFFENSIVE_WORDS = [\"biberius\", \"ретроград\"]\n\nLA_DENIAL_WORDS = [\"отрицание\", \"отсутствие\", \"пустота\", \"ничего\", \"несуществование\", \"нонконформизм\", \"сакральный\", \"дестркутивный\"]\nLA_DENIAL_OFFENSIVE_WORDS = [\"циник\", \"дегенерат\", \"бессовестный\"]\n\nLA_ACCEPTANCE_WORDS = [\"принятие\", \"присутствие\", \"принимание\", \"всё\", \"существование\", \"конформизм\", \"мирской\", \"конструктивный\"]\nLA_ACCEPTANCE_OFFENSIVE_WORDS = [\"идеалист\", \"мракобес\"]\n\n# LA_HEDONISM_WORDS = [\"гедонизм\", \"блуд\", \"радость\", \"излишество\", \"удовольствие\", \"счастье\"]\n# LA_HEDONISM_OFFENSIVE_WORDS = [\"девиант\", \"сноб\", \"нимфоман\", \"сибарит\", \"чревоугодец\", \"чад\"]\n\n# LA_ASCETICISM_WORDS = [\"аскетизм\", \"непорочность\", \"скромность\", \"воздержание\", \"чистота\", \"сдержанность\"]\n# LA_ASCETICISM_OFFENSIVE_WORDS = [\"святоша\", \"девственник\"]\n\n\n\nLOCALE = {\"en\": {\"Liberty\": EN_LIBERTY_WORDS, \"Authority\": EN_AUTHORITY_WORDS,\n \"Liberty_Offensive\": EN_LIBERTY_OFFENSIVE_WORDS, \"Authority_Offensive\": EN_AUTHORITY_OFFENSIVE_WORDS,\n \"Pacifism\": EN_PACIFISM_WORDS, \"Militarism\": EN_MILITARISM_WORDS,\n \"Pacifism_Offensive\": EN_PACIFISM_OFFENSIVE_WORDS, \"Militarism_Offensive\": EN_MILITARISM_OFFENSIVE_WORDS,\n \"Intensivism\": EN_INTENSIVISM_WORDS, \"Extensivism\": EN_EXTENSIVISM_WORDS,\n \"Intensivism_Offensive\": EN_INTENSIVISM_OFFENSIVE_WORDS, \"Extensivism_Offensive\": EN_EXTENSIVISM_OFFENSIVE_WORDS,\n # \"Materialism\": EN_MATERIALISM_WORDS, \"Spiritualism\": EN_SPIRITUALISM_WORDS,\n # \"Materialism_Offensive\": EN_MATERIALISM_OFFENSIVE_WORDS, \"Spiritualism_Offensive\": EN_SPIRITUALISM_OFFENSIVE_WORDS,\n \"Individualism\": EN_INDIVIDUALISM_WORDS, \"Collectivism\": EN_COLLECTIVISM_WORDS,\n \"Individualism_Offensive\": EN_INDIVIDUALISM_OFFENSIVE_WORDS, \"Collectivism_Offensive\": EN_COLLECTIVISM_OFFENSIVE_WORDS,\n \"Reformism\": EN_REFORMISM_WORDS, \"Revolutionism\": EN_REVOLUTIONISM_WORDS,\n \"Reformism_Offensive\": EN_REFORMISM_OFFENSIVE_WORDS, \"Revolutionism_Offensive\": EN_REVOLUTIONISM_OFFENSIVE_WORDS,\n \"Constructivism\": EN_CONSTRUCTIVISM_WORDS, \"Essentialism\": EN_ESSENTIALISM_WORDS,\n \"Constructivism_Offensive\": EN_CONSTRUCTIVISM_OFFENSIVE_WORDS, \"Essentialism_Offensive\": EN_ESSENTIALISM_OFFENSIVE_WORDS,\n \"Denial\": EN_DENIAL_WORDS, \"Acceptance\": EN_ACCEPTANCE_WORDS,\n \"Denial_Offensive\": EN_DENIAL_OFFENSIVE_WORDS, \"Acceptance_Offensive\": EN_ACCEPTANCE_OFFENSIVE_WORDS,\n # \"Hedonism\": EN_HEDONISM_WORDS, \"Asceticism\": EN_ASCETICISM_WORDS,\n # \"Hedonism_Offensive\": EN_HEDONISM_OFFENSIVE_WORDS, \"Asceticism_Offensive\": EN_ASCETICISM_OFFENSIVE_WORDS,\n\n \"Book_Titles\": [\"The %s Manifesto\", \"The Book of %s\", \"The %s Book\", \"A Short Guide to %s\", \"%s Manual\"],\n\n \"Announcement_end\": \"\", # A very specific feature for latin.\n\n \"Reads\": \"reads\", \"Writes\": \"writes\", \"Gives\": \"gives\",\n\n \"Says\": \"says\", \"Mumbles\": \"mumbles\", \"Exclaims\": \"exclaims\", \"Asks\": \"asks\", \"Quotes\": \"quotes\",\n \"All\": \"all\",\n \"Convinced\": \"convinced\",\n \"Changed Ideology From\": \"has changed their ideology from\",\n\n \"Relationship Changed To\": \"changed their relationship to\",\n \"From\": \"from\", \"To\": \"to\",\n \"Reverence\": \"Reverence\", \"Respect\": \"Respect\", \"Sympathy\": \"Sympathy\",\n \"None\": \"None\",\n \"Distaste\": \"Distaste\", \"Hate\": \"Hate\", \"Feud\": \"Feud\",\n\n \"cry\": \"cries\", \"cry_target\": \"cries because of\",\n \"lecture\": \"lectures\", \"lecture_target\": \"lectures\",\n \"awe\": \"is in awe\", \"awe_target\": \"is in awe of\",\n \"praise\": \"praises\", \"praise_target\": \"praises\",\n \"dissapointment\": \"is dissapointed\", \"dissapointment_target\": \"is dissapointed in\",\n \"curse\": \"curses\", \"curse_target\": \"curses\",\n \"preach\": \"preaches\", \"preach_target\": \"preaches\",\n\n # \"\"\"Centrists\"\"\".\n \"Centrism\": \"Centrism\",\n\n # Wacky.\n \"Post-cosmic anarcho-nihilism\": \"Post-cosmic anarcho-nihilism\",\n \"Trans anarcho-pacifism\": \"Trans anarcho-pacifism\",\n\n # Communists.\n \"Communism\": \"Communism\",\n \"Anarcho-communism\": \"Anarcho-communism\",\n \"Trotskyism\": \"Trotskyism\",\n \"Anarcho-trotskyism\": \"Anarcho-trotskyism\",\n \"Posadism\": \"Posadism\",\n \"Anarcho-posadism\": \"Anarcho-posadism\",\n\n # Capitalists.\n \"Capitalism\": \"Capitalism\",\n \"Liberal capitalism\": \"Liberal capitalism\",\n \"Anarcho-capitalism\": \"Anarcho-capitalism\",\n\n # Fascists.\n \"Fascism\": \"Fascism\",\n \"Anarcho-fascism\": \"Anarcho-fascism\",\n \"Strasserism\": \"Strasserism\",\n \"Ecofascism\": \"Ecofascism\",\n \"Esoteric fascism\": \"Esoteric fascism\",\n \"Techno-fascism\": \"Techno-fascism\",\n\n # Syndicalists.\n \"Syndicalism\": \"Syndicalism\",\n \"Anarcho-syndicalism\": \"Anarcho-syndicalism\",\n \"Fascist syndicalism\": \"Fascist syndicalism\",\n\n # Government enthusiasts.\n \"Monarchism\": \"Monarchism\",\n \"Imperialism\": \"Imperialism\",\n \"Statism\": \"Statism\",\n \"Minarchism\": \"Minarchism\",\n \"Social democracy\": \"Social democracy\",\n\n # Peaceful protesters.\n \"Agorism\": \"Agorism\",\n\n # \"Greens\".\n \"Green politics\": \"Green politics\",\n \"Anarcho-primitivism\": \"Anarcho-primitivism\",\n\n # EGO.\n \"Anarcho-individualism\": \"Anarcho-individualism\",\n \"Meritocracy\": \"Meritocracy\",\n \"Technocracy\": \"Technocracy\",\n\n # Spiritualists.\n \"Theocracy\": \"Theocracy\",\n \"Liberation theology\": \"Liberation theology\",\n\n # Radicals vulgaris.\n \"Liberalism\": \"Liberalism\",\n\n # Extremes vulgaris.\n \"Anarchism\": \"Anarchism\",\n \"Conservatism\": \"Conservatism\",\n \"Accelerationism\": \"Accelerationism\",\n \"Transhumanism\": \"Transhumanism\",\n \"Nihilism\": \"Nihilism\",\n },\n\n \"ru\": {\"Liberty\": RU_LIBERTY_WORDS, \"Authority\": RU_AUTHORITY_WORDS,\n \"Liberty_Offensive\": RU_LIBERTY_OFFENSIVE_WORDS, \"Authority_Offensive\": RU_AUTHORITY_OFFENSIVE_WORDS,\n \"Pacifism\": RU_PACIFISM_WORDS, \"Militarism\": RU_MILITARISM_WORDS,\n \"Pacifism_Offensive\": RU_PACIFISM_OFFENSIVE_WORDS, \"Militarism_Offensive\": RU_MILITARISM_OFFENSIVE_WORDS,\n \"Intensivism\": RU_INTENSIVISM_WORDS, \"Extensivism\": RU_EXTENSIVISM_WORDS,\n \"Intensivism_Offensive\": RU_INTENSIVISM_OFFENSIVE_WORDS, \"Extensivism_Offensive\": RU_EXTENSIVISM_OFFENSIVE_WORDS,\n # \"Materialism\": RU_MATERIALISM_WORDS, \"Spiritualism\": RU_SPIRITUALISM_WORDS,\n # \"Materialism_Offensive\": RU_MATERIALISM_OFFENSIVE_WORDS, \"Spiritualism_Offensive\": RU_SPIRITUALISM_OFFENSIVE_WORDS,\n \"Individualism\": RU_INDIVIDUALISM_WORDS, \"Collectivism\": RU_COLLECTIVISM_WORDS,\n \"Individualism_Offensive\": RU_INDIVIDUALISM_OFFENSIVE_WORDS, \"Collectivism_Offensive\": RU_COLLECTIVISM_OFFENSIVE_WORDS,\n \"Reformism\": RU_REFORMISM_WORDS, \"Revolutionism\": RU_REVOLUTIONISM_WORDS,\n \"Reformism_Offensive\": RU_REFORMISM_OFFENSIVE_WORDS, \"Revolutionism_Offensive\": RU_REVOLUTIONISM_OFFENSIVE_WORDS,\n \"Constructivism\": RU_CONSTRUCTIVISM_WORDS, \"Essentialism\": RU_ESSENTIALISM_WORDS,\n \"Constructivism_Offensive\": RU_CONSTRUCTIVISM_OFFENSIVE_WORDS, \"Essentialism_Offensive\": RU_ESSENTIALISM_OFFENSIVE_WORDS,\n \"Denial\": RU_DENIAL_WORDS, \"Acceptance\": RU_ACCEPTANCE_WORDS,\n \"Denial_Offensive\": RU_DENIAL_OFFENSIVE_WORDS, \"Acceptance_Offensive\": RU_ACCEPTANCE_OFFENSIVE_WORDS,\n # \"Hedonism\": RU_HEDONISM_WORDS, \"Asceticism\": RU_ASCETICISM_WORDS,\n # \"Hedonism_Offensive\": RU_HEDONISM_OFFENSIVE_WORDS, \"Asceticism_Offensive\": RU_ASCETICISM_OFFENSIVE_WORDS,\n\n \"Book_Titles\": [\"Манифест %s\", \"Книга о %s\", \"Книга %s\", \"Краткое Пособие к %s\", \"Мануал %s\"],\n\n \"Announcement_end\": \"\", # A very specific feature for latin.\n\n \"Reads\": \"читает\", \"Writes\": \"пишет\", \"Gives\": \"даёт\",\n\n \"Says\": \"говорит\", \"Mumbles\": \"бубнит\", \"Exclaims\": \"восклицает\", \"Asks\": \"спрашивает\", \"Quotes\": \"цитирует\",\n \"All\": \"все\",\n \"Convinced\": \"убедил\", \"Changed Ideology From\": \"сменил свою идеологию с\",\n\n \"Relationship Changed To\": \"изменил своё отношение к\", \n \"From\": \"с\", \"To\": \"на\",\n \"Reverence\": \"Почтение\", \"Respect\": \"Уважение\", \"Sympathy\": \"Симпатия\",\n \"None\": \"Никакое\",\n \"Distaste\": \"Неприязнь\", \"Hate\": \"Ненависть\", \"Feud\": \"Вражда\",\n\n \"cry\": \"плачет\", \"cry_target\": \"плачет из-за\",\n \"lecture\": \"отчитывает\", \"lecture_target\": \"отчитывает\",\n \"awe\": \"восхищён\", \"awe_target\": \"восхищён\",\n \"praise\": \"фанатично почитает\", \"praise_target\": \"фанатично почитает\",\n \"dissapointment\": \"разочарован\", \"dissapointment_target\": \"разочарован в\",\n \"curse\": \"проклинает\", \"curse_target\": \"проклинает\",\n \"preach\": \"поучает\", \"preach_target\": \"поучает\",\n\n # \"\"\"Centrists\"\"\".\n \"Centrism\": \"Центризм\",\n\n # Wacky.\n \"Post-cosmic anarcho-nihilism\": \"Пост-космический анархо-нигилизм\",\n \"Trans anarcho-pacifism\": \"Транс анархо-пацифизм\",\n\n # Communists.\n \"Communism\": \"Коммунизм\",\n \"Anarcho-communism\": \"Анархо-коммунизм\",\n \"Trotskyism\": \"Троцкизм\",\n \"Anarcho-trotskyism\": \"Анархо-троцкизм\",\n \"Posadism\": \"Посадизм\",\n \"Anarcho-posadism\": \"Анархо-посадизм\",\n\n # Capitalists.\n \"Capitalism\": \"Капитализм\",\n \"Liberal capitalism\": \"Либеральный капитализм\",\n \"Anarcho-capitalism\": \"Анархо-капитализм\",\n\n # Fascists.\n \"Fascism\": \"Фашизм\",\n \"Anarcho-fascism\": \"Анархо-фашизм\",\n \"Strasserism\": \"Штрассеризм\",\n \"Ecofascism\": \"Экофашизм\",\n \"Esoteric fascism\": \"Эзотерический фашизм\",\n \"Techno-fascism\": \"Техно-фашизм\",\n\n # Syndicalists.\n \"Syndicalism\": \"Синдикализм\",\n \"Anarcho-syndicalism\": \"Анархо-синдикализм\",\n \"Fascist syndicalism\": \"Фашистский синдикализм\",\n\n # Government enthusiasts.\n \"Monarchism\": \"Монархизм\",\n \"Imperialism\": \"Империализм\",\n \"Statism\": \"Этатизм\",\n \"Minarchism\": \"Минархизм\",\n \"Social democracy\": \"Социальная-демократия\",\n\n # Peaceful protesters.\n \"Agorism\": \"Агоризм\",\n\n # \"Greens\".\n \"Green politics\": \"Зелёная политика\",\n \"Anarcho-primitivism\": \"Анархо-примитивизм\",\n\n # EGO.\n \"Anarcho-individualism\": \"Анархо-индивидуализм\",\n \"Meritocracy\": \"Меритократия\",\n \"Technocracy\": \"Технократия\",\n\n # Spiritualists.\n \"Theocracy\": \"Теократия\",\n \"Liberation theology\": \"Теология освобождения\",\n\n # Radicals vulgaris.\n \"Liberalism\": \"Либерализм\",\n\n # Extremes vulgaris.\n \"Anarchism\": \"Анархизм\",\n \"Conservatism\": \"Консерватизм\",\n \"Accelerationism\": \"Акселерационизм\",\n \"Transhumanism\": \"Трансгуманизм\",\n \"Nihilism\": \"Нигилизм\"\n },\n\n \"la\": {\"Liberty\": LA_LIBERTY_WORDS, \"Authority\": LA_AUTHORITY_WORDS,\n \"Liberty_Offensive\": LA_LIBERTY_OFFENSIVE_WORDS, \"Authority_Offensive\": LA_AUTHORITY_OFFENSIVE_WORDS,\n \"Pacifism\": LA_PACIFISM_WORDS, \"Militarism\": LA_MILITARISM_WORDS,\n \"Pacifism_Offensive\": LA_PACIFISM_OFFENSIVE_WORDS, \"Militarism_Offensive\": LA_MILITARISM_OFFENSIVE_WORDS,\n \"Intensivism\": LA_INTENSIVISM_WORDS, \"Extensivism\": LA_EXTENSIVISM_WORDS,\n \"Intensivism_Offensive\": LA_INTENSIVISM_OFFENSIVE_WORDS, \"Extensivism_Offensive\": LA_EXTENSIVISM_OFFENSIVE_WORDS,\n # \"Materialism\": LA_MATERIALISM_WORDS, \"Spiritualism\": LA_SPIRITUALISM_WORDS,\n # \"Materialism_Offensive\": LA_MATERIALISM_OFFENSIVE_WORDS, \"Spiritualism_Offensive\": LA_SPIRITUALISM_OFFENSIVE_WORDS,\n \"Individualism\": LA_INDIVIDUALISM_WORDS, \"Collectivism\": LA_COLLECTIVISM_WORDS,\n \"Individualism_Offensive\": LA_INDIVIDUALISM_OFFENSIVE_WORDS, \"Collectivism_Offensive\": LA_COLLECTIVISM_OFFENSIVE_WORDS,\n \"Reformism\": LA_REFORMISM_WORDS, \"Revolutionism\": LA_REVOLUTIONISM_WORDS,\n \"Reformism_Offensive\": LA_REFORMISM_OFFENSIVE_WORDS, \"Revolutionism_Offensive\": LA_REVOLUTIONISM_OFFENSIVE_WORDS,\n \"Constructivism\": LA_CONSTRUCTIVISM_WORDS, \"Essentialism\": LA_ESSENTIALISM_WORDS,\n \"Constructivism_Offensive\": LA_CONSTRUCTIVISM_OFFENSIVE_WORDS, \"Essentialism_Offensive\": LA_ESSENTIALISM_OFFENSIVE_WORDS,\n \"Denial\": LA_DENIAL_WORDS, \"Acceptance\": LA_ACCEPTANCE_WORDS,\n \"Denial_Offensive\": LA_DENIAL_OFFENSIVE_WORDS, \"Acceptance_Offensive\": LA_ACCEPTANCE_OFFENSIVE_WORDS,\n # \"Hedonism\": LA_HEDONISM_WORDS, \"Asceticism\": LA_ASCETICISM_WORDS,\n # \"Hedonism_Offensive\": LA_HEDONISM_OFFENSIVE_WORDS, \"Asceticism_Offensive\": LA_ASCETICISM_OFFENSIVE_WORDS,\n\n \"Book_Titles\": [\"Манифест %s\", \"Книга о %s\", \"Книга %s\", \"Краткое Пособие к %s\", \"Мануал %s\"],\n\n \"Announcement_end\": \" est\", # A very specific feature for latin.\n\n \"Reads\": \"читает\", \"Writes\": \"пишет\", \"Gives\": \"даёт\",\n\n \"Says\": \"dicit\", \"Mumbles\": \"murmurit\", \"Exclaims\": \"clamit\", \"Asks\": \"quaerit\", \"Quotes\": \"цитирует\",\n \"All\": \"omnia\",\n \"Convinced\": \"persuasit\", \"Changed Ideology From\": \"vertit ille doctrina ab\",\n\n \"Relationship Changed To\": \"vertit ille relatio ad\",\n \"From\": \"ab\", \"To\": \"ad\",\n \"Reverence\": \"Veneratio\", \"Respect\": \"Respectus\", \"Sympathy\": \"Sympathia\",\n \"None\": \"Nihili\",\n \"Distaste\": \"Inimitia\", \"Hate\": \"Odium\", \"Feud\": \"Hostilitas\",\n\n \"cry\": \"flarit\", \"cry_target\": \"flarit ob\",\n \"lecture\": \"citatit\", \"lecture_target\": \"vituperarit\",\n \"awe\": \"in admirationem\", \"awe_target\": \"in admirationem ab\",\n \"praise\": \"veneratit\", \"praise_target\": \"veneratit\",\n \"dissapointment\": \"non dulcedit\", \"dissapointment_target\": \"non dulcedit ad\",\n \"curse\": \"vomicit\", \"curse_target\": \"vomicit\",\n \"preach\": \"поучает\", \"preach_target\": \"поучает\",\n\n # \"\"\"Centrists\"\"\".\n \"Centrism\": \"Centrismus\",\n\n # Wacky.\n \"Post-cosmic anarcho-nihilism\": \"Пост-космический анархо-нигилизм\",\n \"Trans anarcho-pacifism\": \"Транс анархо-пацифизм\",\n\n # Communists.\n \"Communism\": \"Communismus\",\n \"Anarcho-communism\": \"Communismus irauctoris\",\n \"Trotskyism\": \"Trotskyismus\",\n \"Anarcho-trotskyism\": \"Trotskyismus irauctoris\",\n \"Posadism\": \"Posadismus\",\n \"Anarcho-posadism\": \"Posadismus irauctoris\",\n\n # Capitalists.\n \"Capitalism\": \"Capitalismus\", # Capitalismus vulgaris. As in, classical Capitalism. Change to this if you want to be more fancy.\n \"Liberal capitalism\": \"Capitalismus liberalis\",\n \"Anarcho-capitalism\": \"Capitalismus irauctoris\",\n\n # Fascists.\n \"Fascism\": \"Fascismus\",\n \"Anarcho-fascism\": \"Fascismus irauctoris\",\n \"Strasserism\": \"Strasserismus\",\n \"Ecofascism\": \"Domofascismus\",\n \"Esoteric fascism\": \"Fascismus immaterialis\",\n \"Techno-fascism\": \"Arsi-fascismus\",\n\n # Syndicalists.\n \"Syndicalism\": \"Syndicalismus\",\n \"Anarcho-syndicalism\": \"Syndicalismus irauctoris\",\n \"Fascist syndicalism\": \"Syndicalismus fascistis\",\n\n # Government enthusiasts.\n \"Monarchism\": \"Монархизм\",\n \"Imperialism\": \"Империализм\",\n \"Statism\": \"Civitismus\",\n \"Minarchism\": \"Minauctorismus\",\n \"Social democracy\": \"Populiauctorismus socialis\",\n\n # Peaceful protesters.\n \"Agorism\": \"Agorismus\",\n\n # \"Greens\".\n \"Green politics\": \"Politica virida\",\n \"Anarcho-primitivism\": \"Primitivismus irauctoris\",\n\n # EGO.\n \"Anarcho-individualism\": \"Individualismus irauctoris\",\n \"Meritocracy\": \"Dignauctoria\",\n \"Technocracy\": \"Arsauctoria\",\n\n # Spiritualists.\n \"Theocracy\": \"Deiauctoria\",\n \"Liberation theology\": \"Deiscientia liberatae\",\n\n # Radicals vulgaris.\n \"Liberalism\": \"Liberalismus\",\n\n # Extremes vulgaris.\n \"Anarchism\": \"Irauctorismus\",\n \"Conservatism\": \"Conservatismus\",\n \"Accelerationism\": \"Accelerationismus\",\n \"Transhumanism\": \"Transhumanismus\",\n \"Nihilism\": \"Nihilismus\"\n }\n}\n","repo_name":"LudwigVonChesterfield/Totally-Accurate-Political-Simulator","sub_path":"source_code/locale_game.py","file_name":"locale_game.py","file_ext":"py","file_size_in_byte":31330,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"81"}
+{"seq_id":"32038652774","text":"\"\"\"\nTags for performing basic value comparisons in templates.\n\"\"\"\nfrom django import template\n\n\nfrom template_utils import blocks\n\nclass BlockNode(template.Node):\n def __init__(self, name, nodelist, varname, *vars):\n self.name = name\n self.vars = list(map(template.Variable, vars))\n self.nodelist = nodelist\n self.varname = varname\n \n def render(self, context):\n result = blocks[self.name](context, self.nodelist, *self.vars)\n if self.varname:\n context[self.varname] = result\n return ''\n return result\n\ndef do_block(parser, token):\n bits = token.contents.split()\n nodelist = parser.parse(('end%s' % bits[0],))\n parser.delete_first_token()\n varname = None\n if len(bits) > 2 and bits[-2] == 'as':\n varname = bits[-1]\n bits = bits[:-2]\n return BlockNode(bits[0], nodelist, varname, *bits[1:])\n\nregister = template.Library()\nfor tag_name in blocks:\n register.tag(tag_name, do_block)\n","repo_name":"AlphacrucisCollege/django-template-utils","sub_path":"template_utils/templatetags/blocks.py","file_name":"blocks.py","file_ext":"py","file_size_in_byte":994,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21170947869","text":"\nimport torch\nimport torchvision\nimport torchvision.transforms as transforms\n\n\ndef index_to_onehot(index, num_indices):\n onehot = torch.zeros(*index.size(), num_indices)\n onehot.scatter_(-1, index.unsqueeze(dim=-1), 1)\n return onehot\n\nclass FlattenInput:\n def __call__(self, x):\n return x.view(-1)\n\ndef MNISTDataset(batch_size, num_workers):\n\n transform = transforms.Compose([\n transforms.ToTensor(),\n #transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))\n FlattenInput(),\n ])\n\n trainset = torchvision.datasets.MNIST(\n root='./data', train=True, download=True, transform=transform)\n testset = torchvision.datasets.MNIST(\n root='./data', train=False, download=True, transform=transform)\n trainloader = torch.utils.data.DataLoader(\n trainset, batch_size=batch_size, shuffle=True, num_workers=num_workers)\n testloader = torch.utils.data.DataLoader(\n testset, batch_size=batch_size, shuffle=False, num_workers=num_workers)\n\n return trainloader, testloader\n\n\n\n\n\n\n\n\n","repo_name":"zeligism/eqprop","sub_path":"dataset.py","file_name":"dataset.py","file_ext":"py","file_size_in_byte":1055,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"81"}
+{"seq_id":"73222098186","text":"#Given a list of numbers and a number k, return whether any two numbers from the list add up to k.\r\n\r\n#[10, 15, 3, 7]\r\n#17\r\n\r\ndef yes(list, k):\r\n for i in list:\r\n for j in list:\r\n if (i == j):\r\n continue\r\n if (i + j == k):\r\n return True\r\n return False\r\n\r\n\r\nlist = []\r\n\r\nsize = int(input(\"Number of elements: \"))\r\nfor i in range(0, size):\r\n i = int(input(\"Element of list: \"))\r\n list.append(i)\r\nprint(list)\r\n\r\nk = int(input(\"Input k: \"))\r\nprint(yes(list, k))\r\n","repo_name":"kittykaty/dailycoding","sub_path":"12.03.py","file_name":"12.03.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"38593448231","text":"import os\nimport logging\nimport random\n\nlogger = logging.getLogger()\n\n\nseed = os.environ[\"SEED\"]\nseed = int(seed)\nrandom.seed(seed)\n\n\ndef handler(event, context):\n logger.info(\"received processing event {!r}\".format(event))\n\n # in pharmacy benefits, the prices are just made up\n return {\n \"price\": random.randint(1, 10000),\n }\n","repo_name":"capitalrx/devops","sub_path":"terraform/pricing.py","file_name":"pricing.py","file_ext":"py","file_size_in_byte":346,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5639151491","text":"# create a dictionary\nmydict = {\n 'waqt': 'time',\n 'intezar':'wait for someone',\n 'kahani': 'story',\n 'itafaq' : 'by chance'\n}\nprint('-------',mydict.keys(),'--------\\n' )\n\nw = input('Enter a Hindi word : \\n')\nprint('\\n',mydict.get(w))\n","repo_name":"Sujit-sahoo3571/python_programming","sub_path":"01_question_dict.py","file_name":"01_question_dict.py","file_ext":"py","file_size_in_byte":248,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"28881528812","text":"from typing import Dict, Optional, Union\nfrom uuid import uuid4\n\nimport webview\n\nfrom kanmail.log import logger\nfrom kanmail.server.app import server\nfrom kanmail.settings.constants import FRAMELESS, IS_APP, SERVER_HOST, SESSION_TOKEN\n\nID_TO_WINDOW = {} # internal ID -> window object\nUNIQUE_NAME_TO_ID = {} # name -> internal ID for unique windows\n\nUNIQUE_KEY_TO_LOCALIZATION = {\n \"settings\": {\n \"global.quit\": \"Close without saving\",\n \"global.cancel\": \"Return to settings\",\n \"global.quitConfirmation\": (\n \"Any changes will be lost, do you still want to close the window?\"\n ),\n },\n}\n\n\ndef create_window(\n endpoint: str = \"/\",\n unique_key: Optional[str] = None,\n **kwargs,\n) -> Union[str, bool]:\n if not IS_APP:\n logger.warning(\"Cannot open window in server mode!\")\n return False\n\n internal_id = str(uuid4())\n link = (\n f\"http://{SERVER_HOST}:{server.get_port()}{endpoint}\"\n f\"?window_id={internal_id}\"\n f\"&Kanmail-Session-Token={SESSION_TOKEN}\"\n )\n\n logger.debug(\n f\"Opening window ({internal_id}) \" f\"url={endpoint} kwargs={kwargs}\",\n )\n\n # Nuke any existing unique window\n if unique_key and unique_key in UNIQUE_NAME_TO_ID:\n old_window_id = UNIQUE_NAME_TO_ID.get(unique_key)\n if old_window_id:\n destroy_window(old_window_id)\n\n window = webview.create_window(\n \"Kanmail\",\n link,\n frameless=FRAMELESS,\n easy_drag=False,\n text_select=True,\n localization=UNIQUE_KEY_TO_LOCALIZATION.get(unique_key),\n **kwargs,\n )\n\n ID_TO_WINDOW[internal_id] = window\n\n if unique_key:\n UNIQUE_NAME_TO_ID[unique_key] = internal_id\n\n return internal_id\n\n\ndef destroy_window(internal_id: str) -> None:\n window = ID_TO_WINDOW.pop(internal_id, None)\n\n if window:\n try:\n window.destroy()\n except KeyError: # happens if the window has already been destroyed (user close)\n pass\n else:\n return\n\n logger.warning(f\"Tried to destroy non-existant window: {internal_id}\")\n\n\ndef resize_window(internal_id: str, width: int, height: int) -> None:\n window = ID_TO_WINDOW[internal_id]\n\n if window:\n window.resize(width, height)\n\n logger.warning(f\"Tried to resize non-existant window: {internal_id}\")\n\n\ndef reload_main_window() -> None:\n if IS_APP:\n window = get_main_window()\n window.evaluate_js(\"window.location.reload()\")\n\n\ndef get_main_window() -> webview.Window:\n return ID_TO_WINDOW[UNIQUE_NAME_TO_ID[\"main\"]]\n\n\ndef destroy_main_window() -> None:\n destroy_window(UNIQUE_NAME_TO_ID[\"main\"])\n\n\ndef get_main_window_size_position() -> Dict[str, int]:\n window = get_main_window()\n return {\n \"left\": window.x,\n \"top\": window.y,\n \"width\": window.width,\n \"height\": window.height,\n }\n\n\ndef init_window_hacks() -> None:\n try:\n from webview.platforms import cocoa\n\n from .macos import reposition_traffic_light_buttons, show_traffic_light_buttons\n except ImportError:\n pass\n else:\n # This cocoa specific hack shows the traffic light buttons (pywebview hides these\n # in frameless mode by default) and also moves them so they look better placed\n # in the sidebar header.\n\n class CustomBrowserView(cocoa.BrowserView):\n def first_show(self, *args, **kwargs):\n show_traffic_light_buttons(self.window)\n reposition_traffic_light_buttons(self.window)\n super().first_show(*args, **kwargs)\n\n class CustomWindowDelegate(cocoa.BrowserView.WindowDelegate):\n def windowDidResize_(self, notification):\n reposition_traffic_light_buttons(notification.object())\n\n cocoa.BrowserView = CustomBrowserView\n cocoa.BrowserView.WindowDelegate = CustomWindowDelegate\n","repo_name":"Oxygem/Kanmail","sub_path":"kanmail/window/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":3916,"program_lang":"python","lang":"en","doc_type":"code","stars":1242,"dataset":"github-code","pt":"81"}
+{"seq_id":"70125546504","text":"#!/usr/bin/python3\n# -*-coding:utf-8 -*-\n\"\"\"Unit tests.\"\"\"\n\nimport pytest # noqa: F401\n\nfrom utils.gif_json import GifJson\n\n\ndef test_get_gif():\n \"\"\"Test get_gif.\"\"\"\n gifs = GifJson(\"gifs.sample.json\")\n my_gif: dict = gifs.get_gif(\"your_gif_name\")\n assert my_gif['url'] == \"your_gif_url\"\n\n\ndef test_get_gif_fail():\n \"\"\"Test get_gif fail.\"\"\"\n gifs = GifJson(\"gifs.sample.json\")\n my_gif = gifs.get_gif(\"toto\")\n assert my_gif is None # gif toto return none\n\n\ndef test_get_names_string():\n \"\"\"Test get_gif fail.\"\"\"\n gifs_json = GifJson(\"gifs.sample.json\")\n # gif_json_mock = MagicMock(GifJson)\n gifs_json.gifs = {\"foo\": {\"public\": True,\n \"url\": \"https://foo.gif\"},\n \"bar\": {\"public\": True,\n \"url\": \"http://bar.gif\"},\n \"foobar\": {\"public\": True,\n \"url\": \"http://bar.gif\"}\n }\n gif_string = gifs_json.get_names_string(private=False)\n assert gif_string == \"foo\\nbar\\nfoobar\"\n\n\ndef test_get_names_string_private():\n \"\"\"Test get_gif fail.\"\"\"\n gifs_json = GifJson(\"gifs.sample.json\")\n # gif_json_mock = MagicMock(GifJson)\n gifs_json.gifs = {\"foo\": {\"public\": True,\n \"url\": \"https://foo.gif\"},\n \"bar\": {\"public\": True,\n \"url\": \"http://bar.gif\"},\n \"fooba\": {\"public\": False,\n \"url\": \"http://bar.gif\"}\n }\n gif_string = gifs_json.get_names_string(private=True)\n assert gif_string == \"foo\\nbar\"\n","repo_name":"Azrood/DCTbot","sub_path":"tests/test_func_gifs.py","file_name":"test_func_gifs.py","file_ext":"py","file_size_in_byte":1646,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"22096044162","text":"from django.contrib.auth import get_user_model\nfrom rest_framework import serializers\n\nfrom .models import (\n UserProfile, UserProfileQuestionnaireAnswer, PollutantSensitivity)\n\nUserModel = get_user_model()\n\n\nclass PollutantSensitivitySerializer(serializers.ModelSerializer):\n level_display = serializers.CharField(source='get_level_display')\n\n class Meta:\n model = PollutantSensitivity\n fields = ('pollutant_id', 'level', 'level_display', )\n\n\nclass UserProfileQuestionnaireAnswerSerializer(serializers.ModelSerializer):\n\n def to_internal_value(self, data):\n question_id = data.get('question_id')\n answer_id = data.get('answer_id')\n\n # Perform the data validation.\n if not answer_id:\n raise serializers.ValidationError({\n 'answer_id': 'This field is required.'\n })\n\n # Return the validated values. This will be available as\n # the `.validated_data` property.\n return {\n 'question_id': question_id,\n 'answer_id': answer_id\n }\n\n class Meta:\n model = UserProfileQuestionnaireAnswer\n fields = ('question_id', 'answer_id', )\n\n\nclass UserProfileSerializer(serializers.ModelSerializer):\n questionnaire_answers = UserProfileQuestionnaireAnswerSerializer(\n required=False, many=True)\n\n sensitivity_levels = PollutantSensitivitySerializer(\n many=True, read_only=True)\n\n class Meta:\n model = UserProfile\n fields = (\n 'age', 'colour_blindness', 'aqi_scale', 'modified', 'vis_tool',\n 'questionnaire_answers', 'sensitivity_levels', )\n read_only_fields = ('modified', )\n\n def update(self, instance, validated_data):\n questionnaire_answers = validated_data.pop(\n 'questionnaire_answers', [])\n answers_changed = instance.answers_will_change(\n questionnaire_answers)\n for a in questionnaire_answers:\n UserProfileQuestionnaireAnswer.objects.update_or_create(\n profile=instance, question_id=a['question_id'], defaults={\n 'answer_id': a['answer_id']\n })\n\n # Update sensitivity levels\n if answers_changed:\n instance.update_sensitivity_levels()\n return super(UserProfileSerializer, self).update(\n instance, validated_data)\n\n\nclass UserSerializer(serializers.HyperlinkedModelSerializer):\n\n def create(self, validated_data):\n user = super(UserSerializer, self).create(validated_data)\n user.set_password(validated_data['password'])\n user.save()\n return user\n\n class Meta:\n model = UserModel\n fields = (\n 'id', 'url', 'username', 'email', 'password', 'first_name',\n 'last_name')\n extra_kwargs = {\n 'url': {\n 'view_name': 'api:accounts:user-detail'\n },\n 'password': {'write_only': True}\n }\n\n\nclass ExtendedUserSerializer(UserSerializer):\n\n profile = UserProfileSerializer(required=False)\n\n class Meta(UserSerializer.Meta):\n fields = (\n 'id', 'url', 'username', 'email', 'first_name',\n 'last_name', 'profile', 'is_staff', 'is_active',\n 'is_superuser', 'date_joined', )\n read_only_fields = (\n 'is_staff', 'is_superuser', 'is_active', 'date_joined', )\n\n def update(self, instance, validated_data):\n profile = validated_data.pop('profile', None)\n questionnaire_answers = profile.pop(\n 'questionnaire_answers', [])\n if profile is not None:\n upProfile, created = UserProfile.objects.update_or_create(\n user=instance, defaults=profile)\n answers_changed = upProfile.answers_will_change(\n questionnaire_answers)\n for a in questionnaire_answers:\n UserProfileQuestionnaireAnswer.objects.update_or_create(\n profile=upProfile, question_id=a['question_id'], defaults={\n 'answer_id': a['answer_id']\n })\n\n # Update sensitivity levels\n if answers_changed:\n upProfile.update_sensitivity_levels()\n return super(ExtendedUserSerializer, self).update(\n instance, validated_data)\n","repo_name":"dschurholz/myaqi-backend","sub_path":"aqi_backend/accounts/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":4327,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"74921529545","text":"\"\"\"\nThis file is part of the TheLMA (THe Laboratory Management Application) project.\nSee LICENSE.txt for licensing, CONTRIBUTORS.txt for contributor information.\n\nReservoir specs mapper.\n\"\"\"\nfrom sqlalchemy.orm import relationship\n\nfrom everest.repositories.rdb.utils import as_slug_expression\nfrom everest.repositories.rdb.utils import mapper\nfrom thelma.entities.liquidtransfer import ReservoirSpecs\nfrom thelma.entities.rack import RackShape\n\n\n__docformat__ = 'reStructuredText en'\n__all__ = ['create_mapper']\n\n\ndef create_mapper(reservoir_specs_tbl):\n \"Mapper factory.\"\n tbl = reservoir_specs_tbl\n m = mapper(ReservoirSpecs, reservoir_specs_tbl,\n id_attribute='reservoir_specs_id',\n slug_expression=lambda cls: as_slug_expression(cls._name), # pylint: disable=W0212\n properties=dict(\n _rack_shape=relationship(RackShape, uselist=False,\n# lazy='joined'\n ),\n _name=tbl.c.name,\n _description=tbl.c.description,\n _max_volume=tbl.c.max_volume,\n _min_dead_volume=tbl.c.min_dead_volume,\n _max_dead_volume=tbl.c.max_dead_volume)\n )\n return m\n","repo_name":"helixyte/TheLMA","sub_path":"thelma/repositories/rdb/mappers/reservoirspecs.py","file_name":"reservoirspecs.py","file_ext":"py","file_size_in_byte":1306,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"41256768817","text":"#import sys;\n#sys.path.append(\"../cuqipy\")\n\nimport numpy as np\nimport torch as xp\nimport matplotlib.pyplot as plt\nfrom inclusion import U_class, image, full_tensor_sino\n\nimport arviz\nfrom cuqi.distribution import Gaussian, Uniform, JointDistribution\nfrom cuqi.sampler import Gibbs, MH, pCN\n\nclass forward_map():\n def __init__(self,N,view_angles,s=2):\n self.U = U_class(N=N)\n\n self.radon = full_tensor_sino()\n self.view_angles = xp.asarray(view_angles)\n\n def forward(self,s,v):\n if isinstance(s, np.ndarray):\n s = xp.asarray(s)\n if isinstance(v, np.ndarray):\n v = xp.asarray(v)\n self.U.compute_sqrt_lambda(s)\n u = self.U.make_u_from_v( v )\n sino = self.radon.make_sino(u,self.view_angles)\n return sino.numpy().flatten()\n\ndef F(s,v):\n if isinstance(s, np.ndarray):\n s = xp.asarray(s)\n if isinstance(v, np.ndarray):\n v = xp.asarray(v)\n out = forward.forward(s,v)\n if isinstance(out, xp.Tensor):\n out = out.numpy()\n return out.flatten()\n\nclass Metro(MH):\n def step(self, x):\n self.x0 = x\n self.scale = 0.8\n return self.sample(20, ).samples[:,-1]\n\n def _print_progress(*args, **kwargs):\n pass\n\nclass PCN(pCN):\n def step(self, x):\n self.x0 = x\n self.scale = 0.07\n return self.sample(20).samples[:,-1]\n\n def _print_progress(*args, **kwargs):\n pass\n\ndef run_gibbs():\n # loading observation file\n obs_data = np.load('./obs/obs_gear_90.npz')\n N = obs_data['N'] # discretization size\n sino = xp.asarray( obs_data['sino'] )\n noise_vec = xp.asarray( obs_data['noise_vec'] ) # noise vector\n view_angles = obs_data['view_angles'] # view angles\n\n # defining the forward problem\n forward = forward_map(N,view_angles)\n\n sigma = np.linalg.norm(sino)/100\n sigma2 = sigma*sigma\n y_obs = sino + sigma*noise_vec\n y_obs_flat = y_obs.flatten()\n\n log_like = lambda v,s: - ( 0.5*np.sum( (forward.forward(s,v) - y_obs)**2)/sigma2 )\n\n m = len(y_obs_flat)\n Im = np.ones(m)\n\n # Bayesian model\n s = Uniform(0.5,5)\n v = Gaussian(np.zeros(N) , 1)\n y = Gaussian(forward.forward, sigma2*Im)\n\n # joint distribution\n P = JointDistribution(s,v,y)\n\n # Gibbs sampler\n sampler = Gibbs(P(y=y_obs_flat), {'s':Metro, 'v':PCN})\n\n # run sampler\n samples = sampler.sample(10000)\n\n np.savez('./stats/stat_gear.npz',samp1=samples['v'].samples.T,samp2=samples['s'].samples.T)\n\nif __name__ == '__main__':\n run_gibbs()\n\n\n","repo_name":"babakmaboudi/uncertain_roughness","sub_path":"CT/inverse_CUQIpy.py","file_name":"inverse_CUQIpy.py","file_ext":"py","file_size_in_byte":2545,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"2556481253","text":"# coding=utf-8\nimport sys\nfile = sys.argv[1]\nimport json\narray = []\n# fakenumber = 0\n# truenumber = 0\n# negnumber = 0\n# posnumber = 0\nnumber=[0,0,0,0]\ntotal = 0\ntrue={}\nfake={}\nneg={}\npos={}\ndef trueOrFake(word):\n if word[1]=='Fake':\n number[0]+=1\n for x in range(3,len(word)):\n if word[x] in fake:\n fake[word[x]]+=1\n else:\n fake[word[x]]=1\n\n else:\n number[1] += 1\n for x in range(3,len(word)):\n if word[x] in true:\n true[word[x]]+=1\n else:\n true[word[x]]=1\n\n\n if word[2]=='Neg':\n number[2] += 1\n for x in range(3,len(word)):\n if word[x] in neg:\n neg[word[x]]+=1\n else:\n neg[word[x]]=1\n else:\n number[3] += 1\n for x in range(3,len(word)):\n if word[x] in pos:\n pos[word[x]]+=1\n else:\n pos[word[x]]=1\n\n\n# with open(\"train-labeled.txt\",\"r\") as f:\nwith open(file, \"r\") as f:\n content = f.readlines()\n for x in content:\n total += 1\n x = x.strip('\\n')\n word = x.split(\" \")\n array.append(word[0]);\n trueOrFake(word);\n print(number)\n print(total)\n print(true)\n print(fake)\n print(neg)\n print(pos)\n\n\nimport json\nlistdic={}\nlistdic[\"number\"]=number\njsObj1 = json.dumps(true)\njsObj2= json.dumps(fake)\njsObj3 = json.dumps(neg)\njsObj4 = json.dumps(pos)\njsObj5 = json.dumps(listdic)\n\nfileObject = open(\"nbmodel.txt\",\"w\")\n#fileObject = open(\"dict_f.json\",\"w\")\nfileObject.write(jsObj1)\nfileObject.write('\\n')\nfileObject.write(jsObj2)\nfileObject.write('\\n')\nfileObject.write(jsObj3)\nfileObject.write('\\n')\nfileObject.write(jsObj4)\nfileObject.write('\\n')\nfileObject.write(jsObj5)","repo_name":"FanZhangSara/NLP-naive-Bayes-classifier","sub_path":"nblearn3.py","file_name":"nblearn3.py","file_ext":"py","file_size_in_byte":1796,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14006280645","text":"import typing\nimport copy\nimport random\nimport json\nimport gzip\nimport markovify\nimport spacy\nfrom app.caching import cache\nfrom flask import current_app\n\n\nclass POSifiedText(markovify.Text):\n '''\n POSified text is used for both construction and sentence sythesis\n Includes POS (parts of speech) to allow for more sensible sentence structure.\n '''\n def word_split(self, sentence):\n nlp_model = get_nlp_model('en_core_web_sm')\n return ['::'.join((word.orth_, word.pos_)) for word in nlp_model(sentence)]\n\n def word_join(self, words):\n sentence = ' '.join(word.split('::')[0] for word in words)\n return sentence\n\n\ndef markov_generate(persona: str='Trump', params: dict={}) -> str:\n '''\n Generate a quote from initialized markov models\n Models have already been calibrated; this auto-generates something from \n the Markov model and returns.\n\n Parameters:\n - persona\n - params \n '''\n print('Generating markov text for persona: {}'.format(persona))\n\n # Retreive list fo models applicable\n models = get_models(persona, {k:params[k] for k in ['MODEL_DIRECTORY', 'MARKOV_MODELS', 'PERSONAS']}) \n model_spec = models[0] # models[random.randint(0, 100) % 2]\n model = model_spec['model']\n text = model.make_short_sentence(max_chars = model_spec['max_chars'])\n print('Markov text ({}): {}'.format(persona, text))\n return text\n\n\n@cache.memoize()\ndef get_nlp_model(lang_module='en_core_web_sm'):\n return spacy.load(lang_module)\n\n\n@cache.memoize() # Cache decorator; save to memory\ndef get_models(persona: str='Trump', params: dict={}) -> list:\n '''\n Initializes list of models specified for this persona\n Models are pre-calibrated and saved into json format\n Read into POSifiedText markovify model extension\n \n Params:\n persona: \n params: \n '''\n # Initialize all available models specified\n model_specs = copy.deepcopy(params['MARKOV_MODELS'][persona])\n model_dir = params['MODEL_DIRECTORY']\n \n print('Initializing markov model for persona: {}'.format(persona))\n models = []\n for model_spec in model_specs:\n model_json = read_model_file(model_dir, model_spec['filename'])\n model_obj = POSifiedText.from_json(model_json)\n model_spec['model'] = model_obj\n models.append(model_spec)\n\n return models\n\n\ndef read_model_file(model_dir, model_file):\n '''\n The models should be in compressed (zipped) format)\n Additionally, we are reading from within the application\n '''\n model_file_path = '{app_root}/{resource_dir}{model_file}'.format(app_root=current_app.root_path,\n resource_dir=model_dir, model_file=model_file)\n with gzip.GzipFile(model_file_path, 'r') as fin:\n model_json = json.loads(fin.read().decode('utf-8'))\n \n return model_json\n\n\ndef sentence_topic_extract(text: str, rules: typing.List[typing.Dict], short_circuit: bool=True) -> list:\n '''\n Find the \"topic\" of a given sentence\n Based on a set of rules (lambda functions), operate on either document-level \n named entities, or on tokens in the document\n This will inform image search + document title\n\n Parameters:\n text (str): the sentence / string to search\n rules (list): a list of rules (defined in default.py configuration)\n short_circuit (bool) - if a prioritied match is found, return first matching\n '''\n # TODO: if a named ent is all caps, check that the lower() version is not a verb (i.e. \"DO\" classified as org)\n\n nlp = get_nlp_model('en_core_web_sm')\n doc = nlp(text)\n matches = []\n for rule in rules:\n search_elem = [ent for ent in doc.ents] if rule['type'] == 'ent' else [tok for tok in doc]\n curr_matches = [elem for elem in search_elem if rule['lambda'](elem, nlp)]\n matches += curr_matches\n if short_circuit and len(curr_matches) > 0:\n break\n\n return matches\n\n\n","repo_name":"liangjh/tspire-api","sub_path":"app/spire/spiremodel.py","file_name":"spiremodel.py","file_ext":"py","file_size_in_byte":4001,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"28784847608","text":"import torch \nimport torch.nn as nn\nfrom torchvision.models.feature_extraction import get_graph_node_names\n\nclass SiameseModel(nn.Module):\n \n # potential configurable arguments \n # channels: int, n_classes: int, dim_sizes: List[int], kernel_size: int, stride: int, padding: int, **kwargs\n # pretrained model and weights \n def __init__(self, base_model, base_model_weights): \n super(SiameseModel, self).__init__()\n self.emb_size = 20\n self.weights = base_model_weights\n self.siamese = base_model(weights=self.weights)\n train_nodes, eval_nodes = get_graph_node_names(self.siamese)\n # self.classifier = nn.Sequential(\n # nn.Linear(self.siamese.get_submodule(train_nodes[-1]).out_features, 1),\n # nn.Sigmoid()\n # )\n # self.classifier = nn.Linear(self.siamese.get_submodule(train_nodes[-1]).out_features * 2, 1)\n # final model \n self.classifier = nn.Linear(self.siamese.get_submodule(train_nodes[-1]).out_features, 1)\n # model with embedding \n # self.feature_extractor = nn.Sequential(\n # nn.Linear(self.siamese.get_submodule(train_nodes[-1]).out_features, 512),\n # nn.ReLU(inplace = True),\n # nn.Linear(512, self.emb_size)\n # )\n # self.classifier = nn.Linear(self.emb_size, 1)\n\n def forward(self, img1, img2): \n preprocess = self.weights.transforms()\n x1 = preprocess(img1)\n x2 = preprocess(img2)\n out1 = self.siamese(x1)\n out2 = self.siamese(x2)\n # multiply to get combined feature vector representing the similarity btwn the two\n # combined_features = torch.cat((out1, out2), 1)\n # output = self.classifier(combined_features)\n # model with embedding \n # out1 = self.feature_extractor(out1)\n # out2 = self.feature_extractor(out2)\n diff = torch.abs(torch.sigmoid(out1) - torch.sigmoid(out2))\n output = self.classifier(diff)\n return output","repo_name":"rchllllll/cp5105-capstone-proj","sub_path":"src/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":1859,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"9233612878","text":"import pandas as pd\nimport numpy as np\nimport streamlit as st\nfrom google.oauth2 import service_account\nfrom google.cloud import bigquery\nfrom io import StringIO\n#from streamlit import caching\nimport time\n\n# col=0 (advertisingsystem), 1 (PubAccId) , 2 (Relationship), \ndef check(df,col,keyword):\n list=df[col][~df[col].str.contains(keyword)].tolist()\n if len(list)>0:\n return list\n else:\n return False\n\n# Check if AvertisingSystem contains '.' or Relationship is not DIRECT or RESELLER\ndef return_input_error(input):\n if check(input,0,'\\.'):\n st.sidebar.write('Check AdvertisingSystem:')\n st.sidebar.write(check(input,0,'\\.'))\n if check(input,2,'DIRECT|RESELLER'):\n st.sidebar.write('Check Relationship:')\n st.sidebar.write(check(input,2,'DIRECT|RESELLER'))\n\n\n# df[0] (advertisingsystem), df[1] (PubAccId) , df[2] (Relationship), \ndef check_row(df,input_data,row):\n df_filtered=df[(df['AdvertisingSystem']==input_data[0][row])&(df['PubAccId']==input_data[1][row])&(df['Relationship']==input_data[2][row])]\n if df_filtered.shape[0]>0:\n return df_filtered\n else:\n return None\n\t\n#download\ndef download(output_data):\n if output_data.shape[0]>0: \n csv = output_data.to_csv(index=False).encode('utf-8')\n st.download_button(\n \t\tlabel=\"Download ouput as CSV\",\n \t\tdata=csv,\n \t\tfile_name='data.csv',\n \t\tmime='text/csv',\n\t\t)\n\t\n st.dataframe(output_data.reset_index(drop=True),2000,1000)\n\n else:\n st.write('')\n st.write('No output found')\n\t\n\n\t\n\t\n\t\nst.set_page_config(layout=\"wide\")\n\n\n# streamlit_app.py\n\n\n# initial setting\nuploaded_file=None\nlist_lines='Ex: google.com, 12335, DIRECT'\nif 'count' not in st.session_state:\n\tst.session_state.count = 0\n\n#if 'BI_team' not in st.session_state:\n # Password = st.text_input('Password', 'Type here')\n # st.session_state.Password=Password\n # st.write('The password is', Password)\n#else:\n # st.write('Nothing')\n\t\ncol1, col2,col3 = st.columns(3)\nwith col1:\n st.write('')\nwith col2:\n text_input_container = st.empty()\n text_input_container.text_input(\"Enter Password\", key=\"text_input\")\nwith col3:\n st.write('')\n\n\nif (st.session_state.text_input != \"\"):\n if st.session_state.text_input != 'BI_team':\n if st.session_state.count <6:\n st.session_state.count += 1\n\n col1a, col2a,col3a = st.columns(3)\n with col1:\n st.write('')\n with col2a:\n st.write('Wrong password')\n with col3a:\n st.write('')\n else:\n text_input_container.empty()\n st.write('Please contact admin')\n \n \n else:\n text_input_container.empty()\n #st.info(st.session_state.text_input)\n\nif st.session_state.text_input =='BI_team':\n choice = st.sidebar.radio(\"Select invironment\",('WEB','APP', 'Test', 'Test2'), horizontal=True)\n\n\n choice2 = st.sidebar.radio(\"Insert input\",('Upload','Type/Paste'), horizontal=True)\n\n if choice2=='Upload':\n uploaded_file = st.sidebar.file_uploader(\"Choose a .csv file\")\n\n if uploaded_file is not None:\n bytes_data = uploaded_file.getvalue()\n\t\n try:\n upload_input=pd.read_csv(uploaded_file,header=None)\n n=upload_input.shape[0]\n\t\n\t # Clean\n upload_input[0]=upload_input[0].str.replace(' ', '').str.replace('\\t','').str.lower() \n upload_input[1]=upload_input[1].astype('string').str.replace(' ', '').str.replace('\\t','').str.lower()\n upload_input[2]=upload_input[2].str.replace(' ', '').str.replace('\\t','').str.upper()\n\t \n return_input_error(upload_input)\n st.sidebar.dataframe(upload_input)\n\t\t\n except Exception as ex:\n st.sidebar.write('Please check the input format')\n uploaded_file=None\n \n\n elif choice2=='Type/Paste':\n list_lines= st.sidebar.text_area('Put lines here', 'Ex: google.com, 12335, DIRECT')\n \n try:\n input=pd.read_table(StringIO(list_lines),sep=\",\", header=None)\n\t\n # Clean\n input[0]=input[0].str.replace(' ','').str.replace('\\t','').str.lower()\n input[1]=input[1].astype('string').str.replace(' ','').str.replace('\\t','').str.lower()\n input[2]=input[2].str.replace(' ','').str.replace('\\t','').str.upper()\n input=input.drop_duplicates()\n if list_lines !='Ex: google.com, 12335, DIRECT' and list_lines.strip()!='':\n return_input_error(input)\n st.sidebar.write('Input data',input)\n except:\n st.sidebar.write('Please check the input format')\n list_lines=''\n\n\n\n\n\n col4, col5,col6 = st.columns(3)\n\n with col4:\n st.image(\"images.png\", width=80)\n\n with col5:\n st.title(\"📊 IAB dataset\")\n with col6:\n st.write('')\n st.markdown(\"\"\"
\"\"\", unsafe_allow_html=True)\n\t\n if (uploaded_file is None) and ((list_lines=='Ex: google.com, 12335, DIRECT') or (list_lines.strip()=='')):\n st.markdown(f'{\"Please insert input!\"}
', unsafe_allow_html=True)\n\n # Create API client.\n credentials = service_account.Credentials.from_service_account_info(st.secrets[\"gcp_service_account\"])\n client = bigquery.Client(credentials=credentials)\n\n\n\n if ('Time1' not in st.session_state) and ('Time2' not in st.session_state):\n query_time1=\"SELECT Date FROM `showheroes-bi.bi.bi_adstxt_join_sellerjson_with_count_domains` limit 1\"\n df_time1= client.query(query_time1).to_dataframe()\n st.session_state['Time1']=df_time1['Date'][0]\n\n query_time2=\"SELECT Date FROM `showheroes-bi.bi.bi_appadstxt_join_sellersjson_with_count_domains` limit 1\"\n df_time2= client.query(query_time2).to_dataframe()\n st.session_state['Time2']=df_time2['Date'][0]\n\n\t\n\n @st.cache(max_entries=1)\n def load_data1(time): \n query1=\"SELECT * except(Date) FROM `showheroes-bi.bi.bi_adstxt_join_sellerjson_with_count_domains`\"\n query_job1 = client.query(query1)\n return client.query(query1).to_dataframe().fillna('-')\n\n\n\n\n @st.cache(max_entries=1)\n def load_data2(time):\n query2=\"SELECT * except(Date) FROM `showheroes-bi.bi.bi_appadstxt_join_sellersjson_with_count_domains`\"\n query_job2 = client.query(query2)\n return client.query(query2).to_dataframe().fillna('-')\n\t\n df1=load_data1(st.session_state['Time1']).copy()\n df2=load_data2(st.session_state['Time2']).copy()\n\n\n if (choice==\"WEB\") and (uploaded_file is not None):\n # first filter before looping\n df1=df1[(df1['AdvertisingSystem'].isin(upload_input[0])) & (df1['PubAccId'].isin(upload_input[1]))]\n df1=df1.reset_index(drop=True)\n\n # Initial setting\n data1=pd.DataFrame(columns=df1.columns.tolist())\n\t\n for row in range(upload_input.shape[0]):\n data1=pd.concat([data1, check_row(df1,upload_input,row)]) \n \n \n # Download \t\n download(data1)\n\t\n elif (choice==\"WEB\") and (list_lines!='Ex: google.com, 12335, DIRECT') and (list_lines.strip()!=''):\n # first filter \n df1=df1[(df1['AdvertisingSystem'].isin(input[0])) & (df1['PubAccId'].isin(input[1]))]\n df1=df1.reset_index(drop=True)\n \n data1=pd.DataFrame(columns=df1.columns.tolist())\n\t\n for row in range(input.shape[0]):\n data1=pd.concat([data1, check_row(df1,input,row)]) \n \n\n # Download \n download(data1)\n \n\t\n elif (choice==\"APP\") and (uploaded_file is not None): \n # first filter \n df2=df2[(df2['AdvertisingSystem'].isin(upload_input[0])) & (df2['PubAccId'].isin(upload_input[1]))]\n df2=df2.reset_index(drop=True)\n \n\n # Initial setting\n data2=pd.DataFrame(columns=df2.columns.tolist())\n\t\n for row in range(upload_input.shape[0]):\n data2=pd.concat([data2, check_row(df2,upload_input,row)]) \n \n\n # Download \t\n download(data2)\n\n\t\n elif (choice==\"APP\") and (list_lines!='Ex: google.com, 12335, DIRECT') and (list_lines.strip()!=''):\n # first filter\n df2=df2[(df2['AdvertisingSystem'].isin(input[0])) & (df2['PubAccId'].isin(input[1]))]\n df2=df2.reset_index(drop=True)\n\t\n data2=pd.DataFrame(columns=df2.columns.tolist())\n\t\n for row in range(input.shape[0]):\n data2=pd.concat([data2, check_row(df2,input,row)]) \n\n # Download\n download(data2)\t\n\t \t\n elif choice=='Test':\n # Store the initial value of widgets in session state\n if \"visibility\" not in st.session_state:\n st.session_state.visibility = \"visible\"\n st.session_state.disabled = False\n\n col1, col2 = st.columns(2)\n\n with col1:\n \t st.checkbox(\"Disable selectbox widget\", key=\"disabled\")\n \t st.radio(\"Set selectbox label visibility 👉\",key=\"visibility\",options=[\"visible\", \"hidden\", \"collapsed\"],)\n\n with col2:\n \t option = st.selectbox(\"How would you like to be contacted?\",(\"Email\", \"Home phone\", \"Mobile phone\"),label_visibility=st.session_state.visibility,disabled=st.session_state.disabled,)\n\t\n\n elif choice=='Test2':\n st.write('Hello')\n st.caching.clear_cache() \n\t\n @st.cache(suppress_st_warning=True)\n def expensive_computation(a, b):\n st.session_state[\"cache_updated\"] = True\n time.sleep(2) # This makes the function take 2s to run\n return a * b\n\n a = 3\n b = 21\n res = expensive_computation(a, b)\n st.write(res)\n st.write((st.session_state[\"cache_updated\"]))\n#else:\n# st.write('WRONG PASSWORD')\n # st.button('Try again')\n# text_input_container = st.empty()\n # text_input_container.text_input(\"Enter Password\", key=\"text_input\")\n","repo_name":"VT-Do/Bi_new","sub_path":"backup_v2.py","file_name":"backup_v2.py","file_ext":"py","file_size_in_byte":9998,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"71141944584","text":"# accounts/forms.py\nfrom django import forms\nfrom django.contrib.auth.forms import UserCreationForm, UserChangeForm\nfrom django.core.exceptions import ValidationError\n\nfrom .models import CustomUser, Team, Bet, Match\n\nclass CustomUserCreationForm(UserCreationForm):\n\n class Meta:\n model = CustomUser\n fields = (\"username\", \"email\", \"points\")\n\nclass CustomUserChangeForm(UserChangeForm):\n\n class Meta:\n model = CustomUser\n fields = (\"username\", \"email\", \"points\")\n\nclass PlaceBetForm(forms.ModelForm):\n def __init__(self, *args, **kwargs):\n self.user = kwargs.pop('user')\n super(PlaceBetForm, self).__init__(*args, **kwargs)\n\n class Meta:\n model = Bet\n fields = ['choice', 'amount', 'match']\n \n def clean(self):\n cleaned_data = super(PlaceBetForm, self).clean()\n if self.cleaned_data['amount'] > getattr(self.user, 'points'):\n raise ValidationError(\"Error! You cannot bet more points than you have!\")\n if self.cleaned_data['amount'] < 0:\n raise ValidationError(\"You can't bet negative points!\")\n # can't bet on yourself\n if getattr(self.user, 'team') == self.cleaned_data['choice']:\n raise ValidationError(\"You can't bet on your own team!\")\n # cant bet on a team in a match you're in\n match = self.cleaned_data['match']\n match = Match.objects.filter(id=getattr(match,'id'))\n teams = list(match.values_list('teams__name', flat=True))\n if str(getattr(self.user, 'team')) in teams:\n raise ValidationError(\"You can't bet on a match you're in!\")\n # cant bet on team if youve already bet on a different team for the same match\n # first, get all bets for this match placed by this player\n placed_bets = Bet.objects.filter(match=self.cleaned_data['match']) # all bets for this match\n placed_bets = placed_bets.filter(player=self.user) # just by this player\n # check the team of the 1st one\n if len(placed_bets) > 0:\n print(self.cleaned_data)\n team_choice = getattr(placed_bets[0], \"choice\")\n if self.cleaned_data['choice'] != team_choice:\n raise ValidationError(\"You can't bet for a different team! You've already placed a bet!\")\n return self.cleaned_data","repo_name":"sanjerine/midyearmania","sub_path":"bets/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":2331,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"70981789704","text":"from django.shortcuts import render, redirect\nfrom django.conf import settings\nfrom .forms import UserRegisterForm, UserProfileForm\nfrom django.contrib import messages\nfrom .models import UserType\nfrom django.contrib.auth.models import User\n\n# Create your views here.\napp = settings.APP_NAME\n\n\ndef register(request):\n if request.method == 'POST':\n form = UserRegisterForm(request.POST)\n if form.is_valid():\n print(\"hello1\")\n form.save()\n username = form.cleaned_data['username']\n user = User.objects.filter(username=username).first()\n UserType.objects.create(user=user, user_type=\"user\")\n user.usertype.save()\n print(\"hello2\")\n messages.success(request, f'Your Account has been created Successfully! You are now able to Log In')\n return redirect('login')\n else:\n form = UserRegisterForm()\n return render(request, 'users/register.html', {'title': 'sign up', 'app': app, 'form': form})\n\n\ndef profile(request):\n if request.method == 'POST':\n p_form = UserProfileForm(request.POST, request.FILES, instance=request.user.profile)\n if p_form.is_valid():\n p_form.save()\n messages.success(request, f'Your Profile has been Updated Successfully!')\n return redirect('profile')\n else:\n p_form = UserProfileForm(instance=request.user.profile)\n return render(request, 'users/profile.html', {'title': 'profile', 'app': app, 'form': p_form})\n","repo_name":"jdshah98/Food-Express","sub_path":"users/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1517,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"32572811299","text":"import sqlite3\r\nfrom Classes.dbClasses import *\r\n\r\n\r\nclass AccessDatabase():\r\n def __init__(self):\r\n self.conn = sqlite3.connect(\"Database/database.db\")\r\n self.cursor = self.conn.cursor()\r\n # self.create_tables()\r\n\r\n ###########################################################################################################################################################\r\n # CREATE DATA TABLES\r\n def create_tables(self):\r\n try:\r\n self.cursor.execute('''\r\n CREATE TABLE IF NOT EXISTS User(\r\n user_id INTEGER PRIMARY KEY AUTOINCREMENT,\r\n name VARCHAR(100) NOT NULL,\r\n email VARCHAR(100) NOT NULL,\r\n debts INTEGER DEFAULT 0,\r\n aptNo INTEGER,\r\n FOREIGN KEY(aptNo) REFERENCES Apartment(aptNo) \r\n )\r\n ''')\r\n print(\"user table created successfully\")\r\n except:\r\n print(\"failed to create user table\")\r\n\r\n try:\r\n self.cursor.execute('''\r\n CREATE TABLE IF NOT EXISTS Bill(\r\n bill_id INTEGER PRIMARY KEY AUTOINCREMENT,\r\n amount INTEGER NOT NULL,\r\n due_date REAL NOT NULL,\r\n details VARCHAR(100) NOT NULL,\r\n user_id INTEGER,\r\n FOREIGN KEY(user_id) REFERENCES User(user_id)\r\n )\r\n ''')\r\n print(\"Bill table created successfully\")\r\n except:\r\n print(\"failed to create Bill table\")\r\n\r\n try:\r\n self.cursor.execute('''\r\n CREATE TABLE IF NOT EXISTS Payments(\r\n id INTEGER PRIMARY KEY AUTOINCREMENT,\r\n amount INTEGER NOT NULL,\r\n date REAL NOT NULL,\r\n bill_id INTEGER,\r\n FOREIGN KEY(bill_id) REFERENCES Bill(bill_id)\r\n )\r\n ''')\r\n print(\"Payments table created successfully\")\r\n except:\r\n print(\"failed to create Payments table\")\r\n\r\n try:\r\n self.cursor.execute('''\r\n CREATE TABLE IF NOT EXISTS Apartment(\r\n aptNo INTEGER PRIMARY KEY AUTOINCREMENT,\r\n status VARCHAR(20) NOT NULL\r\n )\r\n ''')\r\n print(\"Apartment table created successfully\")\r\n except:\r\n print(\"failed to create Apartment table\")\r\n\r\n\r\n ############################################################################################################################################################\r\n # GETTERS\r\n def getUser(self, id):\r\n self.cursor.execute(\" SELECT aptNo FROM User WHERE user_id = ?\",(id,))\r\n aptNo = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT debts FROM User WHERE user_id = ?\", (id,))\r\n debts = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT email FROM User WHERE user_id = ?\", (id,))\r\n email = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT name FROM User WHERE user_id = ?\", (id,))\r\n name = self.cursor.fetchone()[0]\r\n\r\n user = User(user_id=id, name=name, aptNo=aptNo, debts=debts, email=email)\r\n return user\r\n\r\n def getBill(self, id):\r\n self.cursor.execute(\" SELECT amount FROM Bill WHERE bill_id = ?\", (id,))\r\n amount = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT due_date FROM Bill WHERE bill_id = ?\", (id,))\r\n due_date = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT user_id FROM Bill WHERE bill_id = ?\", (id,))\r\n user_id = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT details FROM Bill WHERE bill_id = ?\", (id,))\r\n details = self.cursor.fetchone()[0]\r\n\r\n bill = Bill(bill_id=id, amount=amount, due_date=due_date, user_id=user_id, details=details)\r\n return bill\r\n\r\n def getPayments(self, id):\r\n self.cursor.execute(\" SELECT amount FROM Payments WHERE id = ?\", (id,))\r\n amount = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT date FROM Payments WHERE id = ?\", (id,))\r\n date = self.cursor.fetchone()[0]\r\n self.cursor.execute(\" SELECT bill_id FROM Payments WHERE id = ?\", (id,))\r\n bill_id = self.cursor.fetchone()[0]\r\n\r\n payment = Payment(id=id, amount=amount, date=date, bill_id=bill_id)\r\n return payment\r\n \r\n\r\n def getApartment(self, id):\r\n self.cursor.execute(\" SELECT status FROM Apartment WHERE aptNo = ?\", (id,))\r\n status = self.cursor.fetchone()[0]\r\n \r\n\r\n def getAllUsers(self):\r\n self.cursor.execute(\"SELECT * FROM User\")\r\n all_users = self.cursor.fetchall()\r\n return all_users\r\n\r\n def getAllApartments(self):\r\n self.cursor.execute(\"SELECT * FROM Apartment\")\r\n all_apts = self.cursor.fetchall()\r\n return all_apts\r\n \r\n def getAllBills(self):\r\n self.cursor.execute(\"SELECT * FROM Bill\")\r\n all_bills = self.cursor.fetchall()\r\n return all_bills\r\n \r\n def getAllPayments(self):\r\n self.cursor.execute(\"SELECT * FROM Payments\")\r\n all_payments = self.cursor.fetchall()\r\n return all_payments\r\n \r\n\r\n def getAllBillAmounts(self):\r\n self.cursor.execute(\"SELECT amount FROM Bill\")\r\n all_amounts = self.cursor.fetchall()\r\n return all_amounts\r\n\r\n def getAllPaymentAmounts(self):\r\n self.cursor.execute(\"SELECT amount FROM Payments\")\r\n all_amounts = self.cursor.fetchall()\r\n return all_amounts\r\n \r\n def getAllUsersDebts(self):\r\n self.cursor.execute(\"SELECT debts FROM User\")\r\n all_debts = self.cursor.fetchall()\r\n return all_debts\r\n \r\n def getAllUserBills(self, user_id):\r\n self.cursor.execute(\"SELECT * FROM Bill WHERE user_id = ?\", (user_id,))\r\n all_user_bills = self.cursor.fetchall()\r\n return all_user_bills\r\n \r\n def getAllUserPayments(self, bill_id):\r\n self.cursor.execute(\"SELECT * FROM Payments WHERE bill_id = ?\", (bill_id,))\r\n all_user_payments = self.cursor.fetchall()\r\n return all_user_payments\r\n","repo_name":"ThommysArt/Apartment-management-system-PYQT6-SQLite","sub_path":"Database/dao.py","file_name":"dao.py","file_ext":"py","file_size_in_byte":6298,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39907456764","text":"from flask_app import app\nfrom flask import render_template,redirect,request,session,flash\nfrom flask_app.models.user import User\nfrom flask_app.models.archive import Archive\n\n\n#==================================\n# Archive Routes\n#==================================\n@app.route('/archive')\ndef archive():\n if 'user_id' not in session:\n flash(\"You must log in to do this.\")\n session['flash'] = \"archive\"\n return render_template('login.html')\n\n data = {\n 'user_id': session['user_id']\n }\n\n dates = Archive.get_all(data)\n return render_template('archive.html', dates = dates)\n\n@app.route('//archive')\ndef add_to_archive(date):\n if 'user_id' not in session:\n flash(\"You must log in to do this.\")\n session['flash'] = \"add to archive\"\n return render_template('login.html', date = date)\n\n data = {\n 'date': date,\n 'user_id': session['user_id']\n }\n \n if not Archive.get_one(data):\n Archive.create(data)\n\n return redirect(f'/apod/{date}')\n\n@app.route('/apod//delete')\ndef delete_from_archive(date):\n if ('user_id' not in session):\n return redirect('/')\n\n data = {\n 'date': date,\n 'user_id': session['user_id']\n }\n\n Archive.delete(data)\n\n return redirect('/archive')","repo_name":"Scheunemann-Matt/NASA_APOD","sub_path":"flask_app/controllers/archives.py","file_name":"archives.py","file_ext":"py","file_size_in_byte":1320,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39060510129","text":"shopping_list = []\r\n\r\nprint(\"Enter the items in the list\")\r\nprint(\"Enter DONE once item adding is completed\")\r\n\r\nwhile(True):\r\n new_item = input(\"Enter the item:\")\r\n if(new_item == 'DONE'):\r\n break\r\n \r\n shopping_list.append(new_item)\r\n \r\nprint(\"Here is your list items\")\r\n\r\nfor item in shopping_list:\r\n print(item)\r\n \r\n \r\n\r\n \r\n \r\n \r\n","repo_name":"rikki45678/forskinternship","sub_path":"Shopping list.py","file_name":"Shopping list.py","file_ext":"py","file_size_in_byte":371,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"3142769394","text":"import sys\n\n\n\"\"\"\ndef print_star(n):\n string = \"\"\n k = n - 1\n for y in range(1, n):\n string += \" \" * k + \"*\" * (2 * y - 1) + \"\\n\"\n k -= 1\n string += \"*\" * (2 * n - 1)\n k += 1\n for y in range(n - 1, 0, -1):\n string += \"\\n\" + \" \" * k + \"*\" * (2 * y - 1)\n k += 1\n return string\n\"\"\"\n\n\ndef print_star(n, k):\n if k == n:\n return \"*\" * (2 * n - 1) + \"\\n\"\n else:\n string = \"\"\n string += \" \" * (n - k) + \"*\" * (2 * k - 1) + \"\\n\"\n string += print_star(n, k + 1)\n string += \" \" * (n - k) + \"*\" * (2 * k - 1) + \"\\n\"\n return string\n\n\nn = int(sys.stdin.readline())\nsys.stdout.write(print_star(n, 1))\n","repo_name":"mrbartrns/algorithm-and-structure","sub_path":"swea/print_star/star_7.py","file_name":"star_7.py","file_ext":"py","file_size_in_byte":682,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"8533101750","text":"\nimport asyncio\nimport time\nfrom functools import partial, wraps\n\nimport aiohttp\nfrom num2words import num2words\nfrom pyrogram import filters, emoji\nfrom pyrogram.types import CallbackQuery, ChatPermissions, InlineKeyboardButton, InlineKeyboardMarkup, Message\n\nfrom ..main import Assistant\n\ncommand = partial(filters.command, prefixes=list(\"#!\"))\n\nasync def reply_and_delete(message: Message, text: str):\n await asyncio.gather(\n message.delete(),\n message.reply(\n text,\n quote=False,\n reply_to_message_id=getattr(\n message.reply_to_message,\n \"message_id\", None\n ),\n disable_web_page_preview=True\n )\n )\n\n\ndef admins_only(func):\n @wraps(func)\n async def decorator(bot: Assistant, message: Message):\n if bot.is_admin(message):\n await func(bot, message)\n\n await message.delete()\n\n decorator.admin = True\n\n return decorator\n\n\n################################\n\nPING_TTL = 5\n\n\n@Assistant.on_message(command(\"ping\"))\nasync def ping(_, message: Message):\n \"\"\"Ping the assistant\"\"\"\n start = time.time()\n reply = await message.reply_text(\"...\")\n delta_ping = time.time() - start\n await reply.edit_text(f\"**Pong!** `{delta_ping * 1000:.3f} ms`\")\n\n\n################################\n\n\nLOG = \"\"\"\n**حالت لاگ را فعال کنید: کد زیر را در بالای کد خود پیست کنید و دوباره کد را اجرا کنید.**\n\n```import logging\nlogging.basicConfig(level=logging.INFO)```\nبرای گرفتن لاگ دقیق تر؛ از\n`level=logging.DEBUG`\nاستفاده کنید.\n\"\"\"\n\n@Assistant.on_message(command(\"log\"))\nasync def log(_, message: Message):\n \"\"\"Enable debug logging\"\"\"\n await reply_and_delete(message, LOG)\n\n\n################################\n\n\nEX = \"\"\"\nPlease, provide us a **minimal** and **reproducible** example in order to easily understand and reproduce the problem.\n[How do I create a minimal, reproducible example?](https://stackoverflow.com/help/minimal-reproducible-example)\n\"\"\"\n\n\n@Assistant.on_message(command(\"ex\"))\nasync def ex(_, message: Message):\n \"\"\"Ask for minimal example\"\"\"\n await reply_and_delete(message, EX)\n\n\n################################\n\n\nOT = \"\"\"\n**گفت و گوی شما خارج از موضوع گروه میباشد.**\n\nبرای تست ربات ها از گروه @PyrogramTesting استفاده کنید\n\"\"\"\n\n\n@Assistant.on_message(command(\"ot\"))\nasync def ot(_, message: Message):\n \"\"\"offtopic conversation\"\"\"\n answer = OT\n await reply_and_delete(message, answer)\n\n\n################################\n\n\nASK = \"\"\"\n**متاسفانه سوال شما قابل فهم نیست. لطفا با سناریو و مثال دیگری مشکل خود را توضیح دهید.**\n[چطور سوال بهتری بپرسم؟](https://www.chetor.com/154278-%D8%AA%DA%A9%D9%86%DB%8C%DA%A9-%D8%B3%D9%88%D8%A7%D9%84-%D9%BE%D8%B1%D8%B3%DB%8C%D8%AF%D9%86/)\n\"\"\"\n\n\n@Assistant.on_message(command(\"ask\"))\nasync def ask(_, message: Message):\n \"\"\"How to ask questions\"\"\"\n await reply_and_delete(message, ASK)\n\n\n################################\n\n\nLEARN = \"**مشکل شما به پایروگرام مربوط نمیشود. لطفا اول بر زبان پایتون مسلط شوید و سپس اقدام به مطرح کردن مشکلات خود کنید.**\"\n\n\n@Assistant.on_message(command(\"learn\"))\nasync def learn(_, message: Message):\n \"\"\"Tell to learn Python\"\"\"\n await reply_and_delete(message, LEARN)\n\n\n################################\n\n\n# One place for all rules, the docs.\nRULES = '''\n1️⃣ − رعایت ادب از مهم ترین ارکان این گروه میباشد و در صورت مشاهده، بدون اخطار مجبور به حذف خواهیم شد.\n\n2️⃣ − درپاسخ به سوال سایرین ، هر چند ساده و ابتدایی، اگر قصد جواب دادن دارید با صبر و حوصله اقدام به این عمل نمایید و از کنایه و توهین بپرهیزید\n\n3️⃣ − تبلیغات در این گروه فقط در زمینه برنامه نویسی و خرید و فروش سورس کد هایی بر پایه پایروگرام مجاز خواهد بود. پس اگر قصد خرید یا فروش دارید محتوای خود را در قالب یک پیام ارسال و در انتها کاربر را به سمت چت خصوصی جهت شروع همکاری هدایت، کنید.\n\n4️⃣ − وظیفه ما کمک به همدیگه هست؛ مسئولیت کدی که شما میزنین و پروژه ای که انجام میدین به عهده خودتون هست.\n\n5️⃣ − شما میتونین هر نوع ربات و سورس کدی که توسط پایروگرام نوشته شده باشه رو داخل گروه در قالب یک پیام معرفی کنین. افراد گروه میتونن استفاده کنن و یا یک بخشی از توسعه دهنده های ربات شما بشن.\n\n6️⃣ − گروه برای فارسی زبانان هست و ملزمه که فقط فارسی صحبت کنید؛ لطف کنید از انگلیسی صحبت کردن و تا حد امکان از فینگلیش صحبت کردن خودداری کنید. اگه هر فردی خارجی حتی به اشتباه وارد گروه شد شما ملزم هستید به زبان فارسی صحبت کنید و در صورتی که حرف های شما اهمیت خاصی برای طرف مقابل داشته باشه قطعا میتونه از برنامه های ترجمه استفاده کنه.\n\n7️⃣ - صحبت در مورد ربات تبچی - اسپمر و از این دسته ربات های مخرب ممنوعه لطفا رعایت کنید.\n\n8️⃣ − این گروه؛ گروه رسمی پایروگرام فارسی هست؛ لطف کنید سوالاتتون رو فقط در مورد کتابخونه پایروگرام بپرسید ( در غیر این صورت؛ داخل گروه محدود میشین) ؛ برای بحث های خارج از موضوع میتونید از گروه بین المللی @PyrogramLounge استفاده کنید.\n\n9⃣ برای سوال پرسیدن، پیام خود را در یک متن بفرستید و تیکه تیکه نفرستید تا حجم پیام ها زیاد نشه اگر خواستید کد طولانی بفرستید از nekobin.com استفاده کنید یا عکس واضح بفرستید \n'''\n\nRULES_INDEX = [\n '1️⃣ − رعایت ادب از مهم ترین ارکان این گروه میباشد و در صورت مشاهده، بدون اخطار مجبور به حذف خواهیم شد.',\n '2️⃣ − درپاسخ به سوال سایرین ، هر چند ساده و ابتدایی، اگر قصد جواب دادن دارید با صبر و حوصله اقدام به این عمل نمایید و از کنایه و توهین بپرهیزید',\n '3️⃣ − تبلیغات در این گروه فقط در زمینه برنامه نویسی و خرید و فروش سورس کد هایی بر پایه پایروگرام مجاز خواهد بود. پس اگر قصد خرید یا فروش دارید محتوای خود را در قالب یک پیام ارسال و در انتها کاربر را به سمت چت خصوصی جهت شروع همکاری هدایت، کنید.',\n '4️⃣ − وظیفه ما کمک به همدیگه هست؛ مسئولیت کدی که شما میزنین و پروژه ای که انجام میدین به عهده خودتون هست.',\n '5️⃣ − شما میتونین هر نوع ربات و سورس کدی که توسط پایروگرام نوشته شده باشه رو داخل گروه در قالب یک پیام معرفی کنین. افراد گروه میتونن استفاده کنن و یا یک ب��شی از توسعه دهنده های ربات شما بشن.',\n '6️⃣ − گروه برای فارسی زبانان هست و ملزمه که فقط فارسی صحبت کنید؛ لطف کنید از انگلیسی صحبت کردن و تا حد امکان از فینگلیش صحبت کردن خودداری کنید. اگه هر فردی خارجی حتی به اشتباه وارد گروه شد شما ملزم هستید به زبان فارسی صحبت کنید و در صورتی که حرف های شما اهمیت خاصی برای طرف مقابل داشته باشه قطعا میتونه از برنامه های ترجمه استفاده کنه.',\n '7️⃣ - صحبت در مورد ربات تبچی - اددر - اسپمر و از این دسته ربات های مخرب ممنوعه لطفا رعایت کنید.',\n '8️⃣ − این گروه؛ گروه رسمی پایروگرام فارسی هست؛ لطف کنید سوالاتتون رو فقط در مورد کتابخونه پایروگرام بپرسید ( در غیر این صورت؛ داخل گروه محدود میشین) ؛ برای بحث های خارج از موضوع میتونید از گروه بین المللی @PyrogramLounge استفاده کنید.',\n '9⃣ برای سوال پرسیدن، پیام خود را در یک متن بفرستید و تیکه تیکه نفرستید تا حجم پیام ها زیاد نشه اگر خواستید کد طولانی بفرستید از nekobin.com استفاده کنید یا عکس واضح بفرستید ',\n]\n\n\n@Assistant.on_message(command(\"rules\"))\nasync def rules(_, message: Message):\n \"\"\"Show Pyrogram rules\"\"\"\n try:\n index = int(message.command[1])\n text = '⚠️ متن قوانین :\\n'\n text += RULES_INDEX[index - 1]\n except Exception:\n text = RULES\n\n await reply_and_delete(message, text)\n\n################################\n\nFAQ = (\n \"**سوال شما در حال حاضر در قسمت سوالات رایج پاسخ داده شده است.**\\n\"\n \"لطفا در قسمت [سوالات رایج](https://docs.pyrogram.org/faq) جستجو کنید.\"\n)\n\n\n@Assistant.on_message(command(\"faq\"))\nasync def faq(_, message: Message):\n \"\"\"Answer is in the FAQ\"\"\"\n await reply_and_delete(message, FAQ)\n\n\n################################\n\n\nRTD = \"لطفا؛ داکیومنت را مطالعه کنید : https://docs.pyrogram.org\"\n\n\n@Assistant.on_message(command(\"rtd\"))\nasync def rtd(_, message: Message):\n \"\"\"Tell to RTD (gentle)\"\"\"\n await reply_and_delete(message, RTD)\n\n\n################################\n\n\nFMT = (\n \"لطفا کد خود را با استفاده از بک-تیک ارسال کنید تا خوانایی بیشتری داشته باشد..\\n\"\n \"```your code here```\"\n)\n\n\n@Assistant.on_message(command(\"fmt\"))\n@admins_only\nasync def fmt(_, message: Message):\n \"\"\"Tell to format code\"\"\"\n await asyncio.gather(\n message.delete(),\n message.reply(\n FMT,\n quote=False,\n parse_mode=\"html\",\n disable_web_page_preview=True,\n reply_to_message_id=getattr(\n message.reply_to_message,\n \"message_id\", None\n ),\n )\n )\n\n################################\n\nEVIL = '''\n✨ پایروگرام ابزاری متن باز - رایگان و تحت نظر انجمن است.\nاین به این معناست که شما میتوانید هرگونه استفاده ای از این ابزار را داشته باشید.\nاما راهنمایی های افراد به شما یک مزیت برای شماست و کسی مکلف به کمک کردن به شما نیست؛\nبه خصوص اگر میخواهید رفتار درستی نشان ندهید و یا آسیبی به تلگرام و کاربر ها برسانید.\n'''\n\n@Assistant.on_message(command(\"evil\"))\nasync def evil(_, message: Message):\n \"\"\"No help for evil actions\"\"\"\n await reply_and_delete(message, EVIL)\n\n################################\n\n@Assistant.on_message(command(\"up\"))\nasync def up(bot: Assistant, message: Message):\n \"\"\"Show Assistant's uptime\"\"\"\n uptime = time.monotonic_ns() - bot.uptime_reference\n\n us, ns = divmod(uptime, 1000)\n ms, us = divmod(us, 1000)\n s, ms = divmod(ms, 1000)\n m, s = divmod(s, 60)\n h, m = divmod(m, 60)\n d, h = divmod(h, 24)\n\n try:\n arg = message.command[1]\n except IndexError:\n await reply_and_delete(message, f\"**Uptime**: `{d}d {h}h {m}m {s}s`\")\n else:\n if arg == \"-v\":\n await reply_and_delete(\n message,\n f\"**Uptime**: `{d}d {h}h {m}m {s}s {ms}ms {us}μs {ns}ns`\\n\"\n f\"**Since**: `{bot.start_datetime} UTC`\"\n )\n elif arg == \"-p\":\n await reply_and_delete(\n message,\n f\"**Uptime**: \"\n f\"`{num2words(d)} days, {num2words(h)} hours, {num2words(m)} minutes, \"\n f\"{num2words(s)} seconds, {num2words(ms)} milliseconds, \"\n f\"{num2words(us)} microseconds, {num2words(ns)} nanoseconds`\\n\"\n f\"\"\n f\"**Since**: `year {num2words(bot.start_datetime.year)}, \"\n f\"month {bot.start_datetime.strftime('%B').lower()}, day {num2words(bot.start_datetime.day)}, \"\n f\"hour {num2words(bot.start_datetime.hour)}, minute {num2words(bot.start_datetime.minute)}, \"\n f\"second {num2words(bot.start_datetime.second)}, \"\n f\"microsecond {num2words(bot.start_datetime.microsecond)}, Coordinated Universal Time`\"\n )\n else:\n await message.delete()\n\n\n################################\n\nnl = \"\\n\"\n\nHELP = f\"\"\"\n**دستورات قابل استفاده :**\n```\nping\nlog\nex\nask\nlearn\nrules\nfaq\nrtd\nfmt\nevil\nup\n```\n\"\"\"\n\n@Assistant.on_message(command(\"help\") & filters.private)\nasync def help(bot: Assistant, message: Message):\n \"\"\"Show this message\"\"\"\n await asyncio.gather(\n message.delete(),\n message.reply(\n HELP,\n quote=False,\n reply_to_message_id=getattr(\n message.reply_to_message,\n \"message_id\", None\n ),\n )\n )\n","repo_name":"iMat1n/PyrogramIR","sub_path":"assistantir/plugins/commands.py","file_name":"commands.py","file_ext":"py","file_size_in_byte":14243,"program_lang":"python","lang":"fa","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"31539654929","text":"import musicbrainzngs\nimport requests\n\ndef sources(recordingid):\n res = requests.get('http://185.97.32.250:8468/mbid:{}'.format(recordingid))\n #import pdb;pdb.set_trace()\n if not b'None' in res.content:\n return res.content.decode('utf-8')\nmusicbrainzngs.set_useragent(\"mb.py\", \"0\", contact=\"mikael@frykholm.com\")\nres = musicbrainzngs.search_artists(\"Ablaze\")\nfor artist in res['artist-list']:\n print(artist['name'], artist['id'])\n#import pdb;pdb.set_trace()\nres = musicbrainzngs.browse_releases(artist='d2c0d69e-e3ca-45a4-a540-6ce42c617599', limit=100)\nfor release in res['release-list']:\n print(release['title'])\n recordings = musicbrainzngs.browse_recordings(release=release['id'])\n for rec in recordings['recording-list']:\n# import pdb;pdb.set_trace()\n print(\"\\t\\t\",rec['title'],'\\t', end='')\n print(sources(rec['id']))\n#import pdb;pdb.set_trace()\n","repo_name":"mikaelfrykholm/uniqueshare","sub_path":"mb.py","file_name":"mb.py","file_ext":"py","file_size_in_byte":900,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"28314683740","text":"import requests\r\nfrom datetime import datetime, timedelta\r\n\r\n\r\nclass Client:\r\n API_URL = 'https://api.binance.com/api/{v}/{path}'\r\n\r\n def __init__(self, api_key=None, secret_key=None):\r\n self.API_KEY = api_key\r\n self.SECRET_KEY = secret_key\r\n\r\n self.session = self.init_session()\r\n\r\n def get_headers(self):\r\n \"\"\"Получение нужных хедеров для работы с биржей\"\"\"\r\n headers = {\r\n 'Accept': 'application/json',\r\n 'User-Agent': 'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.77 Mobile Safari/537.36'\r\n }\r\n\r\n if self.API_KEY:\r\n headers['X-MBX-APIKEY'] = self.API_KEY\r\n return headers\r\n\r\n def init_session(self):\r\n \"\"\"Инициализация сессии\"\"\"\r\n headers = self.get_headers()\r\n\r\n session = requests.session()\r\n session.headers.update(headers)\r\n return session\r\n\r\n def make_api_url(self, path):\r\n url = self.API_URL.format(\r\n v='v3',\r\n path=path\r\n )\r\n return url\r\n\r\n @staticmethod\r\n def handle_response(self, response: requests.Response):\r\n return response.json()\r\n\r\n def request(self, method, path, **kwargs):\r\n \"\"\"Сделать запрос к бирже\"\"\"\r\n url = self.make_api_url(path)\r\n response = getattr(self.session, method)(url, **kwargs)\r\n return self.handle_response(self, response)\r\n\r\n def get(self, path, **kwargs):\r\n \"\"\"Гет-запрос\"\"\"\r\n return self.request('get', path, **kwargs)\r\n\r\n def get_price(self, **params):\r\n \"\"\"\r\n Получить текущую цену\r\n params = {\r\n 'symbol': ...\r\n }\r\n\r\n https://binance-docs.github.io/apidocs/spot/en/#symbol-price-ticker\r\n \"\"\"\r\n\r\n path = f'ticker/price'\r\n js = self.get(path, params=params)\r\n return float(js['price'])\r\n\r\n def get_ticker_24hr(self, **params):\r\n \"\"\"\r\n Получить изменение цены символа за последние 24 часа в процентом соотоношении\r\n params = {\r\n 'symbol': ...\r\n }\r\n\r\n https://binance-docs.github.io/apidocs/spot/en/#24hr-ticker-price-change-statistics\r\n \"\"\"\r\n\r\n path = 'ticker/24hr'\r\n js = self.get(path, params=params)\r\n return js['priceChangePercent']\r\n\r\n def get_aggregate_trades(self, **params):\r\n \"\"\"\r\n Получить список общий список трейдов\r\n params = {\r\n 'symbol': ...,\r\n 'fromId': ...,\r\n 'startTime': ...,\r\n 'endTime': ...,\r\n 'limit': ...\r\n }\r\n\r\n https://binance-docs.github.io/apidocs/spot/en/#compressed-aggregate-trades-list\r\n \"\"\"\r\n\r\n path = 'aggTrades'\r\n js = self.get(path, params=params)\r\n return js\r\n\r\n def get_klines(self, **params):\r\n \"\"\"\r\n Получить список klines/candlesticks для символа\r\n params = {\r\n 'symbol': ...,\r\n 'interval': ...,\r\n 'startTime': ...,\r\n 'endTime': ...,\r\n 'limit': ...\r\n }\r\n\r\n https://binance-docs.github.io/apidocs/spot/en/#kline-candlestick-data\r\n \"\"\"\r\n\r\n path = 'klines'\r\n js = self.get(path, params=params)\r\n return js\r\n\r\n def get_ticker(self, data: datetime, **params):\r\n \"\"\"\r\n Получить цену коина за определенную дату\r\n params = {\r\n 'symbol' : ...,\r\n }\r\n \"\"\"\r\n\r\n start_timestamp = int(data.timestamp() * 1000)\r\n end_timestamp = int((data + timedelta(minutes=10)).timestamp() * 1000)\r\n\r\n trades = self.get_aggregate_trades(\r\n symbol=params['symbol'],\r\n startTime=start_timestamp,\r\n endTime=end_timestamp\r\n )\r\n\r\n print(trades)\r\n\r\n if len(trades) == 0:\r\n # Вероятно, была введена слишком ранняя дата\r\n return 0\r\n return float(trades[0]['p'])\r\n\r\n def get_ticker_1hr(self, **params):\r\n \"\"\"\r\n Получить изменение цены символа за последний час в процентом соотношении\r\n params = {\r\n 'symbol': ...\r\n }\r\n \"\"\"\r\n current_price = self.get_price(**params)\r\n\r\n hour_ago = datetime.now() - timedelta(hours=1)\r\n hour_ago_price = self.get_ticker(hour_ago, **params)\r\n\r\n return (current_price - hour_ago_price) * 100 / hour_ago_price\r\n\r\n def get_ticker_7days(self, **params):\r\n \"\"\"\r\n Получить изменение цены символа за последние 7 дней в процентом соотношении\r\n params = {\r\n 'symbol': ...\r\n }\r\n \"\"\"\r\n\r\n current_price = self.get_price(**params)\r\n\r\n week_ago = datetime.now() - timedelta(days=7)\r\n week_ago_price = self.get_ticker(week_ago, **params)\r\n\r\n return (current_price - week_ago_price) * 100 / week_ago_price\r\n\r\n","repo_name":"mikunak/-MatMeh-MMM-practice-in-programming-2-semester-","sub_path":"2 семестр/project/binance/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":5306,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"1313069119","text":"from test_framework import generic_test\nimport heapq\n\ndef merge_sorted_arrays(sorted_arrays):\n arrays = [iter(array) for array in sorted_arrays]\n min_heap = []\n for index, array_iter in enumerate(arrays):\n smallest = next(array_iter, None)\n if smallest is not None:\n heapq.heappush(min_heap, (smallest, index))\n\n result = []\n while min_heap:\n smallest, index = heapq.heappop(min_heap)\n array_iter = arrays[index]\n result.append(smallest)\n next_element = next(array_iter, None)\n if next_element is not None:\n heapq.heappush(min_heap, (next_element, index))\n return result\n\n\nif __name__ == '__main__':\n exit(\n generic_test.generic_test_main(\"sorted_arrays_merge.py\",\n \"sorted_arrays_merge.tsv\",\n merge_sorted_arrays))\n","repo_name":"tonywang124/EPIJudge","sub_path":"epi_judge_python/sorted_arrays_merge.py","file_name":"sorted_arrays_merge.py","file_ext":"py","file_size_in_byte":891,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"70282136893","text":"import datetime\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport os\r\n\r\nimport tensorflow as tf\r\nfrom keras import backend as K\r\nfrom keras.engine import Layer, InputSpec\r\nfrom keras.initializers import RandomNormal\r\nfrom keras.layers import Input, Dropout, Concatenate\r\nfrom keras.layers import BatchNormalization, Activation, GaussianNoise\r\nfrom keras.layers.advanced_activations import LeakyReLU\r\nfrom keras.layers.convolutional import UpSampling3D, Conv3D\r\nfrom keras.models import Model\r\nfrom keras.optimizers import Adam\r\nfrom keras.utils import multi_gpu_model\r\nfrom keras_contrib.layers.normalization.instancenormalization import InstanceNormalization\r\n\r\nclass My3dResize(Layer):\r\n def __init__(self, sizes, nn=False, **kwargs):\r\n super(My3dResize, self).__init__(**kwargs)\r\n self.sizes = sizes\r\n self.nn = nn\r\n\r\n def compute_output_shape(self, input_shape):\r\n output_shape = list(input_shape)\r\n output_shape[1] = output_shape[1]*self.sizes[0]\r\n output_shape[2] = output_shape[2]*self.sizes[1]\r\n output_shape[3] = output_shape[3]*self.sizes[2]\r\n return tuple(output_shape)\r\n \r\n def call(self, inputs):\r\n input_shape = inputs.get_shape().as_list()\r\n output_shape = input_shape.copy()\r\n output_shape[1] = output_shape[1]*self.sizes[0]\r\n output_shape[2] = output_shape[2]*self.sizes[1]\r\n output_shape[3] = output_shape[3]*self.sizes[2]\r\n \r\n # resize rows and columns\r\n \r\n u = tf.reshape(inputs, shape=[-1]+input_shape[1:-2]+[input_shape[-2]*input_shape[-1]])\r\n if self.nn:\r\n u = tf.image.resize_images(u, size=output_shape[1:3], method=tf.image.ResizeMethod.BILINEAR)\r\n else:\r\n u = tf.image.resize_images(u, size=output_shape[1:3], method=tf.image.ResizeMethod.NEAREST_NEIGHBOR)\r\n u = tf.reshape(u, shape=[-1]+output_shape[1:-2]+input_shape[-2:])\r\n \r\n # repeat depth-wise\r\n outputs = K.repeat_elements(u, self.sizes[2], axis=3) \r\n return outputs\r\n \r\n## modified from https://github.com/eriklindernoren/Keras-GAN/tree/master/pix2pix\r\n\r\nclass My3dPix2Pix():\r\n def __init__(self, data_loader, savepath='result/pilot', L_weights=(1,100), opt='adam', lrs=(0.0002,0.0), \\\r\n randomshift=0.1, resoutput=0.0, dropout=0.0, smoothlabel=True, gennoise=0,\\\r\n fmloss=False, coordconv=False, resizeconv=False, multigpu=None):\r\n '''\r\n PatchGAN's receptive field = 70x70x4\r\n '''\r\n # Configure data loader\r\n self.data_loader = data_loader\r\n self.savepath = savepath\r\n self.coordconv = coordconv\r\n self.resizeconv = resizeconv\r\n self.opt = opt\r\n self.lr_ini, self.lr_decay = lrs\r\n self.smoothlabel = smoothlabel\r\n self.gennoise = gennoise\r\n self.resoutput=resoutput\r\n self.dropout = dropout\r\n self.fmloss = fmloss # discarded\r\n self.randomshift=randomshift\r\n \r\n self.reswindow = (self.data_loader.window2[0][0]*self.resoutput,\\\r\n self.data_loader.window2[0][1]-0.5*(1-self.resoutput)*self.data_loader.window2[0][0])\r\n \r\n ### from data_loader\r\n # Input shape --- x,y >=64, depth >=16.\r\n self.img_shape = tuple(data_loader.img_shape) + (len(data_loader.window1),)\r\n self.img_rows, self.img_cols, self.depth = data_loader.img_shape\r\n self.channels = len(data_loader.window1)\r\n\r\n # Calculate output shape of D (PatchGAN)\r\n patch = max(int(self.img_rows / 2**3), 1)\r\n dpatch = max(int(self.depth / 2**5), 2)\r\n self.disc_patch = (patch, patch, dpatch, 1)\r\n\r\n # Number of filters in the first layer of G and D\r\n self.gf = 64\r\n self.df = 64\r\n\r\n optimizer = Adam(0.0002, 0.5)\r\n if self.opt=='adam':\r\n optimizer = Adam(self.lr_ini, 0.5)\r\n\r\n # Build and compile the discriminator\r\n self.discriminator, self.discriminator_feat = self.build_discriminator()\r\n self.discriminator.compile(loss='binary_crossentropy',\r\n optimizer=optimizer, loss_weights=[0.5],\r\n metrics=['accuracy'])\r\n self.discriminator_feat.compile(loss='mae',\r\n optimizer=optimizer)\r\n\r\n #-------------------------\r\n # Construct Computational\r\n # Graph of Generator\r\n #-------------------------\r\n\r\n # Build the generator\r\n self.generator = self.build_generator()\r\n\r\n # Input images and their conditioning images\r\n img_A = Input(shape=self.img_shape)\r\n img_B = Input(shape=self.img_shape)\r\n\r\n # By conditioning on B generate a fake version of A\r\n fake_A = self.generator(img_B)\r\n\r\n # For the combined model we will only train the generator\r\n self.discriminator.trainable = False\r\n self.discriminator_feat.trainable = False\r\n\r\n # Discriminators determines validity of translated images / condition pairs\r\n if self.fmloss:\r\n # discarded\r\n '''\r\n valid = self.discriminator([fake_A, img_B])\r\n valid_feat = self.discriminator_feat([fake_A, img_B])\r\n\r\n self.combined = Model(inputs=[img_A, img_B], outputs=[valid, fake_A, valid_feat])\r\n if multigpu is not None:\r\n self.combined = multi_gpu_model(self.combined, gpus=multigpu)\r\n\r\n self.combined.compile(loss=['binary_crossentropy', 'mae', 'mae'],\r\n loss_weights=list(L_weights),\r\n optimizer=optimizer)\r\n '''\r\n valid = self.discriminator([fake_A, img_B])\r\n\r\n self.combined = Model(inputs=[img_A, img_B], outputs=[valid, fake_A])\r\n if multigpu is not None:\r\n self.combined = multi_gpu_model(self.combined, gpus=multigpu)\r\n\r\n self.combined.compile(loss=['binary_crossentropy', ssim_mae_loss],\r\n loss_weights=list(L_weights),\r\n optimizer=optimizer)\r\n \r\n \r\n else:\r\n valid = self.discriminator([fake_A, img_B])\r\n\r\n self.combined = Model(inputs=[img_A, img_B], outputs=[valid, fake_A])\r\n if multigpu is not None:\r\n self.combined = multi_gpu_model(self.combined, gpus=multigpu)\r\n\r\n self.combined.compile(loss=['binary_crossentropy', 'mae'],\r\n loss_weights=list(L_weights),\r\n optimizer=optimizer)\r\n\r\n def build_generator(self):\r\n \"\"\"U-Net Generator\"\"\"\r\n\r\n def conv3d(layer_input, filters, kernel_size=(4,4,2), strides=(2,2,2), bn=True):\r\n \"\"\"Layers used during downsampling\"\"\"\r\n init = RandomNormal(stddev=0.02)\r\n d = Conv3D(filters, kernel_size=kernel_size, strides=strides, padding='same', kernel_initializer=init)(layer_input)\r\n d = LeakyReLU(alpha=0.2)(d)\r\n if bn:\r\n #d = BatchNormalization(momentum=0.8)(d)\r\n d = InstanceNormalization()(d)\r\n return d\r\n\r\n def deconv3d(layer_input, skip_input, filters, kernel_size=(4,4,2), strides=(2,2,2), dropout_rate=0, bn=True):\r\n \"\"\"Layers used during upsampling\"\"\"\r\n if self.resizeconv:\r\n u = My3dResize(strides)(layer_input)\r\n else:\r\n u = UpSampling3D(size=strides)(layer_input)\r\n init = RandomNormal(stddev=0.02)\r\n u = Conv3D(filters, kernel_size=kernel_size, strides=1, padding='same', kernel_initializer=init, activation='relu')(u)\r\n if dropout_rate:\r\n u = Dropout(dropout_rate)(u)\r\n if bn:\r\n #u = BatchNormalization(momentum=0.8)(u)\r\n u = InstanceNormalization()(u)\r\n u = Concatenate()([u, skip_input])\r\n u = Activation('relu')(u)\r\n return u\r\n\r\n # Image input\r\n d00 = Input(shape=self.img_shape)\r\n \r\n if self.coordconv:\r\n d0 = CoordinateChannel3D()(d00)\r\n else:\r\n d0 = d00\r\n \r\n n_layers = 7\r\n encoders = []\r\n decoders = []\r\n \r\n # Downsampling\r\n for i in range(n_layers):\r\n z = 1\r\n if i < self.depth.bit_length()-1: z = 2\r\n if i==0:\r\n encoders.append(conv3d(d0, self.gf, kernel_size=(4,4,z), strides=(2,2,z), bn=False))\r\n else:\r\n encoders.append(conv3d(encoders[-1], self.gf*(2**min(i,3)), kernel_size=(4,4,z), strides=(2,2,z)))\r\n \r\n # Upsampling\r\n for i in range(n_layers-1):\r\n z = 1\r\n if i+self.depth.bit_length()>n_layers: z = 2\r\n if i==0:\r\n decoders.append(deconv3d(encoders[-(i+1)], encoders[-(i+2)], self.gf*(2**min(n_layers-2-i,3)), \r\n kernel_size=(4,4,z), strides=(2,2,z)))\r\n else:\r\n decoders.append(deconv3d(decoders[-1], encoders[-(i+2)], self.gf*(2**min(n_layers-2-i,3)), \r\n kernel_size=(4,4,z), strides=(2,2,z), dropout_rate=self.dropout))\r\n\r\n if self.resizeconv:\r\n u7 = My3dResize((2,2,2))(decoders[-1])\r\n else:\r\n u7 = UpSampling3D(size=2)(decoders[-1])\r\n init = RandomNormal(stddev=0.02)\r\n output_img = Conv3D(self.channels, kernel_size=(4,4,2), strides=1, padding='same', kernel_initializer=init, activation='tanh')(u7)\r\n\r\n return Model(d00, output_img)\r\n\r\n def build_discriminator(self):\r\n def d_layer(layer_input, filters, kernel_size=(4,4,2), strides=(2,2,2), bn=True):\r\n \"\"\"Discriminator layer\"\"\"\r\n init = RandomNormal(stddev=0.02)\r\n d = Conv3D(filters, kernel_size=kernel_size, strides=strides, padding='same', kernel_initializer=init)(layer_input)\r\n d = LeakyReLU(alpha=0.2)(d)\r\n if bn:\r\n #d = BatchNormalization(momentum=0.8)(d)\r\n d = InstanceNormalization()(d)\r\n return d\r\n \r\n img_A = Input(shape=self.img_shape)\r\n img_B = Input(shape=self.img_shape)\r\n\r\n # Concatenate image and conditioning image by channels to produce input\r\n combined_imgs = Concatenate(axis=-1)([img_A, img_B])\r\n \r\n ## testing\r\n #combined_imgs = GaussianNoise(0.02)(combined_imgs)\r\n \r\n if self.coordconv:\r\n combined_imgs = CoordinateChannel3D()(combined_imgs)\r\n \r\n n_d_layer = 4\r\n d_layers = []\r\n df = self.df\r\n dout = self.depth\r\n for i in range(n_d_layer):\r\n if i==0:\r\n d_layers.append(d_layer(combined_imgs, df, kernel_size=(4,4,2), strides=(2,2,2)))\r\n else:\r\n z = 2\r\n s = 2\r\n if dout == 2:\r\n z = 1\r\n if i == n_d_layer-1:\r\n s = 1\r\n df = min(df*2, self.df*8)\r\n d_layers.append(d_layer(d_layers[-1], df, kernel_size=(4,4,z), strides=(s,s,z)))\r\n dout = max(2, int(0.5*dout))\r\n\r\n init = RandomNormal(stddev=0.02)\r\n features = Conv3D(1, kernel_size=(4,4,1), strides=1, padding='same', kernel_initializer=init)(d_layers[-1])\r\n validity = Activation('sigmoid')(features)\r\n \r\n return Model([img_A, img_B], validity), Model([img_A, img_B], features)\r\n\r\n def convert_resoutput(self, imgs_A, imgs_B):\r\n ## imgs 5D tensor (batch, rows, cols, depth, channels)\r\n if self.resoutput:\r\n new_A = []\r\n for i in range(imgs_B.shape[0]):\r\n a = rWND(255*(0.5*imgs_A[i]+0.5),self.data_loader.window2)\r\n b = rWND(255*(0.5*imgs_B[i]+0.5),self.data_loader.window1)\r\n c = a-b\r\n #c[b>100] = 0\r\n #c = WND(c,self.reswindow)\r\n M = np.max(c)\r\n m = np.min(c)\r\n if M==m:\r\n c = np.zeros(c.shape)\r\n else:\r\n c = 2.*(c-m)/(M-m) - 1\r\n new_A.append(c)\r\n imgs_A = np.array(new_A)\r\n return imgs_A\r\n \r\n def invert_resoutput(self, fake_A, imgs_B):\r\n if self.resoutput:\r\n '''\r\n new_A = []\r\n for i in range(fake_A.shape[0]):\r\n a = rWND(255*(0.5*fake_A[i]+0.5),self.reswindow)\r\n b = rWND(255*(0.5*imgs_B[i]+0.5),self.data_loader.window1)\r\n c = a+b\r\n c = WND(c,self.data_loader.window2)\r\n new_A.append(c/127.5 - 1.)\r\n fake_A = np.array(new_A)\r\n '''\r\n pass\r\n return fake_A\r\n \r\n def generate_noise(self, mode, imgs_A, imgs_B):\r\n if mode==0:\r\n pass\r\n elif mode==1:\r\n r = np.random.uniform(0,1)\r\n if r>=0.5:\r\n row,col,dep,ch= imgs_A.shape[1:]\r\n mean = 0\r\n sigma = 0.05\r\n gauss = np.random.normal(mean,sigma,(row,col,dep,ch))\r\n gauss = gauss.reshape(row,col,dep,ch)\r\n gauss = np.expand_dims(gauss, axis=0)\r\n imgs_A = imgs_A + gauss\r\n imgs_B = imgs_B + gauss\r\n else:\r\n pass\r\n elif mode==2:\r\n r = np.random.uniform(0,1)\r\n r2 = np.random.uniform(0,1)\r\n if r>=0.5:\r\n row,col,dep,ch= imgs_A.shape[1:]\r\n mean = 0\r\n sigma = 0.05\r\n gauss = np.random.normal(mean,sigma,(row,col,dep,ch))\r\n gauss = gauss.reshape(row,col,dep,ch)\r\n gauss = np.expand_dims(gauss, axis=0)\r\n imgs_A = imgs_A + gauss\r\n imgs_B = imgs_B + gauss\r\n else:\r\n pass\r\n if r2>=0.75:\r\n imgs_A = gaussian_filter(imgs_A, sigma=0.05)\r\n imgs_B = gaussian_filter(imgs_B, sigma=0.05)\r\n else:\r\n pass\r\n else:\r\n pass\r\n \r\n return imgs_A, imgs_B\r\n \r\n def train(self, epochs, batch_size=1, sample_interval=200, model_interval=-1):\r\n\r\n start_time = datetime.datetime.now()\r\n\r\n # Adversarial loss ground truths\r\n patch_shape = (batch_size,) + self.disc_patch\r\n if self.smoothlabel:\r\n valid = np.ones(patch_shape) - 0.1*np.random.rand(*patch_shape)\r\n else:\r\n valid = np.ones(patch_shape)\r\n \r\n fake = np.zeros(patch_shape)\r\n \r\n # log file\r\n f = open(os.path.join(self.savepath, 'log.txt'), 'w')\r\n f.close()\r\n \r\n for epoch in range(epochs):\r\n if epoch<10:\r\n lr = self.lr_ini\r\n else:\r\n lr = self.lr_ini/(1+(epoch-10)*self.lr_decay)\r\n K.set_value(self.discriminator.optimizer.lr, lr)\r\n K.set_value(self.combined.optimizer.lr, lr)\r\n \r\n for batch_i, (imgs_A, imgs_B) in enumerate(self.data_loader.load_batch(batch_size=batch_size)):\r\n\r\n # ---------------------\r\n # Train Discriminator\r\n # --------------------- \r\n imgs_A = self.convert_resoutput(imgs_A, imgs_B)\r\n \r\n # randomshift\r\n non = lambda s: s if s<0 else None\r\n mom = lambda s: max(0,s)\r\n\r\n shift_A = np.full(imgs_A.shape, -1.)\r\n shift_B = np.full(imgs_A.shape, -1.)\r\n sx = int(self.randomshift*shift_A.shape[2])\r\n sy = int(self.randomshift*shift_A.shape[1])\r\n for i in range(shift_A.shape[0]):\r\n ox = np.random.randint(2*sx+1, size=1)[0] - sx\r\n oy = np.random.randint(2*sy+1, size=1)[0] - sy\r\n shift_A[i,mom(oy):non(oy),mom(ox):non(ox),:,:] = imgs_A[i,mom(-oy):non(-oy),mom(-ox):non(-ox),:,:]\r\n shift_B[i,mom(oy):non(oy),mom(ox):non(ox),:,:] = imgs_B[i,mom(-oy):non(-oy),mom(-ox):non(-ox),:,:]\r\n imgs_A = shift_A\r\n imgs_B = shift_B\r\n \r\n imgs_A, imgs_B = self.generate_noise(self.gennoise, imgs_A, imgs_B)\r\n \r\n # Condition on B and generate a translated version\r\n fake_A = self.generator.predict(imgs_B)\r\n\r\n # Train the discriminators (original images = real / generated = Fake)\r\n d_loss_real = self.discriminator.train_on_batch([imgs_A, imgs_B], valid)\r\n d_loss_fake = self.discriminator.train_on_batch([fake_A, imgs_B], fake)\r\n d_loss = np.add(d_loss_real, d_loss_fake)\r\n\r\n # -----------------\r\n # Train Generator\r\n # -----------------\r\n\r\n # Train the generators\r\n '''\r\n if self.fmloss:\r\n valid_feat = self.discriminator_feat.predict([imgs_A, imgs_B])\r\n g_loss = self.combined.train_on_batch([imgs_A, imgs_B], [valid, imgs_A, valid_feat])\r\n else:\r\n g_loss = self.combined.train_on_batch([imgs_A, imgs_B], [valid, imgs_A])\r\n '''\r\n g_loss = self.combined.train_on_batch([imgs_A, imgs_B], [valid, imgs_A])\r\n\r\n elapsed_time = datetime.datetime.now() - start_time\r\n \r\n # Plot the progress\r\n newlog = \"[Epoch %d/%d] [Batch %d/%d] [D loss: %f, D fake: %f] [G loss: %f] time: %s\" % (\r\n epoch+1, epochs, batch_i+1, self.data_loader.n_batches,\r\n d_loss[0], d_loss_fake[0], #100*d_loss[1],\r\n g_loss[0],\r\n elapsed_time\r\n )\r\n \r\n print(newlog)\r\n f = open(os.path.join(self.savepath, 'log.txt'), 'a')\r\n f.write(newlog+'\\n')\r\n f.close() \r\n \r\n # If at save interval => save generated image samples\r\n if (batch_i+1) % sample_interval == 0:\r\n self.sample_images(epoch, batch_i+1)\r\n # save weights\r\n if model_interval>0:\r\n if (epoch+1) % model_interval == 0:\r\n self.save_weights('{}'.format(str(epoch+1)))\r\n \r\n # final sample image & save weights\r\n self.sample_images(epochs, 'final')\r\n if model_interval>0:\r\n self.save_weights('final{}'.format(str(epochs)))\r\n \r\n def predict_on_batch(self, imgs_A, imgs_B):\r\n fake_A = self.generator.predict(imgs_B)\r\n gen_imgs = np.concatenate([imgs_B, fake_A, imgs_A])\r\n return gen_imgs\r\n\r\n def sample_images(self, epoch, batch_i):\r\n r, c = 3, 3\r\n\r\n imgs_A, imgs_B = self.data_loader.load_data(batch_size=3)\r\n fake_A = np.concatenate([self.generator.predict(np.expand_dims(x,axis=0)) for x in imgs_B], axis=0)\r\n fake_A = self.invert_resoutput(fake_A, imgs_B)\r\n\r\n gen_imgs = np.concatenate([imgs_B[:,:,:,0,-1], fake_A[:,:,:,0,-1], imgs_A[:,:,:,0,-1]])\r\n\r\n # Rescale images 0 - 1\r\n gen_imgs = 0.5 * gen_imgs + 0.5\r\n\r\n titles = ['Condition', 'Generated', 'Original']\r\n plt.style.use('default')\r\n fig, axs = plt.subplots(r, c)\r\n cnt = 0\r\n for i in range(r):\r\n dr = imgs_B.shape[1]//4+imgs_B.shape[1]//8\r\n dc = imgs_B.shape[2]//4+imgs_B.shape[2]//8\r\n pr = np.random.choice(imgs_B.shape[1]//2-dr)+imgs_B.shape[1]//4\r\n pc = np.random.choice(imgs_B.shape[2]//2-dc)+imgs_B.shape[2]//4\r\n \r\n for j in range(c):\r\n axs[j,i].imshow(gen_imgs[cnt], cmap='gray', vmin=0, vmax=1)\r\n axs[j,i].set_title(titles[i])\r\n axs[i,j].axis([pc,pc+dc,pr+dr,pr])\r\n cnt += 1\r\n samplepath = self.make_directory('samples')\r\n fig.savefig(os.path.join(samplepath, '{}_{}.png'.format(epoch, batch_i)))\r\n plt.close()\r\n \r\n def make_directory(self, dirname):\r\n dirpath = os.path.join(self.savepath, dirname)\r\n if not os.path.isdir(dirpath):\r\n os.mkdir(dirpath)\r\n return dirpath\r\n \r\n def load_weights(self, weightfile, summary=True):\r\n loadweightspath = os.path.join(self.savepath, 'models', '{}.h5'.format(weightfile))\r\n self.combined.load_weights(loadweightspath)\r\n if summary:\r\n self.combined.summary()\r\n \r\n def load_final_weights(self, *args, **kwargs):\r\n wdir = os.path.join(self.savepath, 'models')\r\n wlist = [os.path.splitext(x)[0] for x in os.listdir(wdir) if x.lower().endswith('.h5')]\r\n wlist.sort()\r\n\r\n s = None\r\n for x in wlist:\r\n if 'final' in x:\r\n s = x\r\n if s is None:\r\n s = wlist[-1] \r\n \r\n self.load_weights(s,*args, **kwargs)\r\n \r\n def save_weights(self, weightfile): \r\n modelpath = self.make_directory('models')\r\n self.combined.save_weights(os.path.join(modelpath, '{}.h5'.format(weightfile)))\r\n print('saved {}.h5 to {}'.format(weightfile, modelpath))\r\n \r\n## not used\r\ndef ssim_mae_loss(y_true, y_pred):\r\n return 0.15*losses.mean_absolute_error(y_true, y_pred) + 0.85*ssim_loss(y_true, y_pred)\r\n\r\ndef ssim_loss(y_true, y_pred):\r\n # ssim for 3d data 구현\r\n ashape = y_true.get_shape().as_list()\r\n\r\n ut = tf.transpose(y_true[:,:,:,:,0], perm=[0,3,1,2])\r\n #ut = tf.reshape(ut, shape=[-1]+ashape[1:3])\r\n ut = tf.expand_dims(ut, axis=-1)\r\n up = tf.transpose(y_pred[:,:,:,:,0], perm=[0,3,1,2])\r\n #up = tf.reshape(up, shape=[-1]+ashape[1:3])\r\n up = tf.expand_dims(up, axis=-1)\r\n \r\n return tf.reduce_mean(1-tf.image.ssim_multiscale(ut, up, 2.0))\r\n\r\n\r\n# modified from \"An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution\" from Uber research\r\n# https://arxiv.org/abs/1807.03247\r\n\r\nclass _CoordinateChannel(Layer):\r\n def __init__(self, rank, \r\n data_format=None,\r\n **kwargs):\r\n super(_CoordinateChannel, self).__init__(**kwargs)\r\n \r\n self.rank = rank\r\n self.data_format = K.image_data_format() if data_format is None else data_format\r\n self.axis = 1 if K.image_data_format() == 'channels_first' else -1\r\n\r\n self.input_spec = InputSpec(min_ndim=2)\r\n\r\n def build(self, input_shape):\r\n assert len(input_shape) >= 2\r\n input_dim = input_shape[self.axis]\r\n\r\n self.input_spec = InputSpec(min_ndim=self.rank + 2,\r\n axes={self.axis: input_dim})\r\n self.built = True\r\n \r\n def call(self, inputs):\r\n input_shape = K.shape(inputs)\r\n\r\n if self.rank == 2:\r\n\r\n input_shape = [input_shape[i] for i in range(4)]\r\n batch_shape, dim1, dim2, channels = input_shape\r\n\r\n xx_ones = K.ones(K.stack([batch_shape, dim2]), dtype='int32')\r\n xx_ones = K.expand_dims(xx_ones, axis=-1)\r\n\r\n xx_range = tf.tile(K.expand_dims(K.arange(0, dim1), axis=0),\r\n K.stack([batch_shape, 1]))\r\n xx_range = K.expand_dims(xx_range, axis=1)\r\n xx_channels = K.batch_dot(xx_ones, xx_range, axes=[2, 1])\r\n xx_channels = K.expand_dims(xx_channels, axis=-1)\r\n xx_channels = K.permute_dimensions(xx_channels, [0, 2, 1, 3])\r\n\r\n yy_ones = K.ones(K.stack([batch_shape, dim1]), dtype='int32')\r\n yy_ones = K.expand_dims(yy_ones, axis=1)\r\n\r\n yy_range = tf.tile(K.expand_dims(K.arange(0, dim2), axis=0),\r\n K.stack([batch_shape, 1]))\r\n yy_range = K.expand_dims(yy_range, axis=-1)\r\n\r\n yy_channels = K.batch_dot(yy_range, yy_ones, axes=[2, 1])\r\n yy_channels = K.expand_dims(yy_channels, axis=-1)\r\n yy_channels = K.permute_dimensions(yy_channels, [0, 2, 1, 3])\r\n\r\n xx_channels = K.cast(xx_channels, K.floatx())\r\n xx_channels = xx_channels / K.cast(dim1 - 1, K.floatx())\r\n xx_channels = (xx_channels * 2) - 1.\r\n\r\n yy_channels = K.cast(yy_channels, K.floatx())\r\n yy_channels = yy_channels / K.cast(dim2 - 1, K.floatx())\r\n yy_channels = (yy_channels * 2) - 1.\r\n\r\n outputs = K.concatenate([inputs, xx_channels, yy_channels], axis=-1)\r\n\r\n if self.rank == 3:\r\n\r\n input_shape = [input_shape[i] for i in range(5)]\r\n batch_shape, dim1, dim2, dim3, channels = input_shape\r\n\r\n xx_ones = K.ones(K.stack([batch_shape, dim3]), dtype='int32')\r\n xx_ones = K.expand_dims(xx_ones, axis=-1)\r\n\r\n xx_range = tf.tile(K.expand_dims(K.arange(0, dim2), axis=0),\r\n K.stack([batch_shape, 1]))\r\n xx_range = K.expand_dims(xx_range, axis=1)\r\n\r\n xx_channels = K.batch_dot(xx_ones, xx_range, axes=[2, 1])\r\n xx_channels = K.expand_dims(xx_channels, axis=-1)\r\n xx_channels = K.permute_dimensions(xx_channels, [0, 2, 1, 3])\r\n\r\n xx_channels = K.expand_dims(xx_channels, axis=1)\r\n xx_channels = tf.tile(xx_channels,\r\n [1, dim1, 1, 1, 1])\r\n\r\n yy_ones = K.ones(K.stack([batch_shape, dim2]), dtype='int32')\r\n yy_ones = K.expand_dims(yy_ones, axis=1)\r\n\r\n yy_range = tf.tile(K.expand_dims(K.arange(0, dim3), axis=0),\r\n K.stack([batch_shape, 1]))\r\n yy_range = K.expand_dims(yy_range, axis=-1)\r\n\r\n yy_channels = K.batch_dot(yy_range, yy_ones, axes=[2, 1])\r\n yy_channels = K.expand_dims(yy_channels, axis=-1)\r\n yy_channels = K.permute_dimensions(yy_channels, [0, 2, 1, 3])\r\n\r\n yy_channels = K.expand_dims(yy_channels, axis=1)\r\n yy_channels = tf.tile(yy_channels,\r\n [1, dim1, 1, 1, 1])\r\n\r\n \r\n zz_range = tf.tile(K.expand_dims(K.arange(0, dim1), axis=0),\r\n K.stack([batch_shape, 1]))\r\n zz_range = K.expand_dims(zz_range, axis=-1)\r\n zz_range = K.expand_dims(zz_range, axis=-1)\r\n\r\n zz_channels = tf.tile(zz_range,\r\n [1, 1, dim2, dim3])\r\n zz_channels = K.expand_dims(zz_channels, axis=-1)\r\n \r\n\r\n xx_channels = K.cast(xx_channels, K.floatx())\r\n xx_channels = xx_channels / K.cast(dim2 - 1, K.floatx())\r\n xx_channels = xx_channels * 2 - 1.\r\n\r\n yy_channels = K.cast(yy_channels, K.floatx())\r\n yy_channels = yy_channels / K.cast(dim3 - 1, K.floatx())\r\n yy_channels = yy_channels * 2 - 1.\r\n\r\n zz_channels = K.cast(zz_channels, K.floatx())\r\n zz_channels = zz_channels / K.cast(dim1 - 1, K.floatx())\r\n zz_channels = zz_channels * 2 - 1.\r\n\r\n outputs = K.concatenate([inputs, zz_channels, xx_channels, yy_channels],axis=-1)\r\n \r\n return outputs\r\n \r\n def compute_output_shape(self, input_shape):\r\n assert input_shape and len(input_shape) >= 2\r\n assert input_shape[self.axis]\r\n\r\n channel_count = self.rank\r\n\r\n output_shape = list(input_shape)\r\n output_shape[self.axis] = input_shape[self.axis] + channel_count\r\n return tuple(output_shape)\r\n \r\n def get_config(self):\r\n config = {\r\n 'rank': self.rank,\r\n 'data_format': self.data_format\r\n }\r\n base_config = super(_CoordinateChannel, self).get_config()\r\n return dict(list(base_config.items()) + list(config.items()))\r\n \r\nclass CoordinateChannel3D(_CoordinateChannel):\r\n def __init__(self,\r\n data_format=None,\r\n **kwargs):\r\n super(CoordinateChannel3D, self).__init__(\r\n rank=3,\r\n data_format=data_format,\r\n **kwargs\r\n )\r\n \r\n def get_config(self):\r\n config = super(CoordinateChannel3D, self).get_config()\r\n config.pop('rank')\r\n return config\r\n","repo_name":"jwc-rad/pix2pix3D-CT","sub_path":"source/my3dpix2pix.py","file_name":"my3dpix2pix.py","file_ext":"py","file_size_in_byte":28461,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"78"}
+{"seq_id":"6430253432","text":"#정답 : dot 이 가장 해당 문제를 표현하는데 최적화되어있음. 계층관계를 표현하기 위한 문제\n#dot : 계층관계를 표현하기 위한 그래프 엔진\n#neato : spring model (서로의 관계에 대해 표현할때 유용, 방향성 있음)\n#fdp : spring model(neato 와 비슷, 관계에 대해 표현, 방향성 없음)\n#sfdp : spring model(fdp 보다 좀 더 크게 그린것, 방향성 없음)\n#twopi : 방사형 그래프를 표현(신경망 같은거에 사용)\n#circo : 싸이클 관계에 있는것을 표현\n\ndef func25a_1():\n from graphviz import Digraph\n\n layout_engine = ['dot','neato','fdp','sfdp','twopi','circo']\n for i in layout_engine:\n f = 'gviz/7th Edition'+i\n a = Digraph(filename = f,comment='7th Edition',engine=i)\n a.node_attr.update(color='goldenrod2',style='filled',size='7,5')\n\n a.edge('7th Edition','32V')\n a.edge('7th Edition','V7M')\n a.edge('7th Edition','Xenix')\n a.edge('7th Edition','UniPlus+')\n\n a.edge('32V','3 BSD')\n a.edge('3 BSD','4 BSD')\n a.edge('4 BSD','4.1 BSD')\n\n a.edge('4.1 BSD','8th Edition')\n a.edge('8th Edition','9th Edition')\n\n a.edge('4.1 BSD','4.2 BSD')\n a.edge('4.2 BSD','Ultrix-32')\n a.edge('4.2 BSD','4.3 BSD')\n\n a.edge('4.1 BSD','2.8 BSD')\n a.edge('1 BSD','2 BSD')\n a.edge('2 BSD','2.8 BSD')\n a.edge('2.8 BSD','2.9 BSD')\n a.edge('2.8 BSD','Ultrix-11')\n\n a.render()\n\nfunc25a_1()\n\n#redbalck tree\ndef func25a_2():\n from graphviz import Digraph\n\n g = Digraph(comment = 'Red-Black Tree',engine='dot')\n\n g.attr('graph',ratio='.48')\n\n g.attr('node',style='filled',color='black',shape='circle',width='.6',\n fontname='Helvetica',fontweight='bold',fontcolor='white',fontsize='24',fixedsize='true')\n\n ns = ['13','1','11','15','25']\n for i in ns:\n g.node(i)\n\n g.attr('node',fillcolor='red')\n ns2 = ['8','17','22','27']\n for j in ns2:\n g.node(j)\n\n g.attr('node',fillcolor='black',shape='record',label='NIL',width='0.4',height='.25',fontsize='16')\n for i in range(1,12):\n la = 'n'+str(i)\n g.node(la)\n\n g.attr('node',style='filled', fillcolor='red',shape='circle',label=r'\\N',width='0.6',fontname='Helvetica'\n ,fontweight='bold',fontcolor='white',fontsize='24',fixedsize='true')\n g.node('6')\n\n relations = [('13','8'),('13','17'),('8','1'),('8','11')\n ,('1','n1'),('1','6'),('6','n2'),('6','n3'),('11','n4'),('11','n5')\n ,('17','15'),('15','n6'),('15','n7'),('17','25'),('25','27')\n ,('25','22'),('22','n8'),('22','n9'),('27','n10'),('27','n11')]\n\n for r in relations:\n g.edge(r[0],r[1])\n # g.node('6')\n\n # g.edge('13','8')\n # g.edge('13','17')\n\n # g.edge('8','1')\n # g.edge('8','11')\n\n # g.edge('1','6')\n\n # g.edge('17','15')\n # g.edge('17','25')\n\n # g.edge('25','22')\n # g.edge('25','27')\n\n # g.edge('1','n1')\n # g.edge('6','n2')\n # g.edge('6','n3')\n # g.edge('11','n4')\n # g.edge('11','n5')\n # g.edge('15','n6')\n # g.edge('15','n7')\n # g.edge('22','n8')\n # g.edge('22','n9')\n # g.edge('27','n10')\n # g.edge('27','n11')\n\n g.view()\n\nfunc25a_2()","repo_name":"yuchanmo/python2","sub_path":"과제/25a.py","file_name":"25a.py","file_ext":"py","file_size_in_byte":3245,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"71049067131","text":"import os\nimport sys\nimport random\nimport json\nimport time\nimport urllib.request\nimport urllib.error\nfrom copy import deepcopy\n\n# ゲームサーバのアドレス / トークン\nGAME_SERVER = os.getenv('GAME_SERVER', 'https://2022contest.gbc.tenka1.klab.jp')\nTOKEN = os.getenv('TOKEN', 'YOUR_TOKEN')\n\nN = 5\nDj = [+1, 0, -1, 0]\nDk = [0, +1, 0, -1]\n\n\n# ゲームサーバのAPIを叩く\ndef call_api(x: str) -> dict:\n url = f'{GAME_SERVER}{x}'\n # 5xxエラーの際は100ms空けて5回までリトライする\n for i in range(5):\n print(url, flush=True)\n try:\n with urllib.request.urlopen(url) as res:\n return json.loads(res.read())\n except urllib.error.HTTPError as err:\n if 500 <= err.code and err.code < 600:\n print(err.code)\n time.sleep(0.1)\n continue\n else:\n raise\n except ConnectionResetError as err:\n print(err)\n time.sleep(0.1)\n continue\n raise Exception('Api Error')\n\n\n# game_idを取得する\n# 環境変数で指定されていない場合は練習試合のgame_idを返す\ndef get_game_id() -> int:\n # 環境変数にGAME_IDが設定されている場合これを優先する\n if os.getenv('GAME_ID'):\n return int(os.getenv('GAME_ID'))\n\n # start APIを呼び出し練習試合のgame_idを取得する\n mode = 0\n delay = 0\n\n start = call_api(f'/api/start/{TOKEN}/{mode}/{delay}')\n if start['status'] == 'ok' or start['status'] == 'started':\n return start['game_id']\n\n raise Exception(f'Start Api Error : {start}')\n\n\n# d方向に移動するように移動APIを呼ぶ\ndef call_move(game_id: int, d: int) -> dict:\n return call_api(f'/api/move/{TOKEN}/{game_id}/{d}')\n\n\n# ゲーム状態クラス\nclass State:\n def __init__(self, field, agent):\n self.field = deepcopy(field)\n self.agent = deepcopy(agent)\n\n # idxのエージェントがいる位置のfieldを更新する\n def paint(self, idx: int):\n i, j, k, _ = self.agent[idx]\n if self.field[i][j][k][0] == -1:\n # 誰にも塗られていない場合はidxのエージェントで塗る\n self.field[i][j][k][0] = idx\n self.field[i][j][k][1] = 2\n elif self.field[i][j][k][0] == idx:\n # idxのエージェントで塗られている場合は完全に塗られた状態に上書きする\n self.field[i][j][k][1] = 2\n elif self.field[i][j][k][1] == 1:\n # idx以外のエージェントで半分塗られた状態の場合は誰にも塗られていない状態にする\n self.field[i][j][k][0] = -1\n self.field[i][j][k][1] = 0\n else:\n # idx以外のエージェントで完全に塗られた状態の場合は半分塗られた状態にする\n self.field[i][j][k][1] -= 1\n\n # エージェントidxをd方向に回転させる\n # 方向については問題概要に記載しています\n def rotate_agent(self, idx: int, d: int):\n self.agent[idx][3] += d\n self.agent[idx][3] %= 4\n\n # idxのエージェントを前進させる\n # マス(i, j, k)については問題概要に記載しています\n def move_forward(self, idx: int):\n i, j, k, d = self.agent[idx]\n jj = j + Dj[d]\n kk = k + Dk[d]\n if jj >= N:\n self.agent[idx][0] = i // 3 * 3 + (i % 3 + 1) % 3 # [1, 2, 0, 4, 5, 3][i]\n self.agent[idx][1] = k\n self.agent[idx][2] = N - 1\n self.agent[idx][3] = 3\n elif jj < 0:\n self.agent[idx][0] = (1 - i // 3) * 3 + (4 - i % 3) % 3 # [4, 3, 5, 1, 0, 2][i]\n self.agent[idx][1] = 0\n self.agent[idx][2] = N - 1 - k\n self.agent[idx][3] = 0\n elif kk >= N:\n self.agent[idx][0] = i // 3 * 3 + (i % 3 + 2) % 3 # [2, 0, 1, 5, 3, 4][i]\n self.agent[idx][1] = N - 1\n self.agent[idx][2] = j\n self.agent[idx][3] = 2\n elif kk < 0:\n self.agent[idx][0] = (1 - i // 3) * 3 + (3 - i % 3) % 3 # [3, 5, 4, 0, 2, 1][i]\n self.agent[idx][1] = N - 1 - j\n self.agent[idx][2] = 0\n self.agent[idx][3] = 1\n else:\n self.agent[idx][1] = jj\n self.agent[idx][2] = kk\n\n # エージェントが同じマスにいるかを判定する\n def is_same_pos(self, a: [int], b: [int]) -> bool:\n return a[0] == b[0] and a[1] == b[1] and a[2] == b[2]\n\n # idxのエージェントがいるマスが自分のエージェントで塗られているかを判定する\n def is_owned_cell(self, idx: int) -> bool:\n i = self.agent[idx][0]\n j = self.agent[idx][1]\n k = self.agent[idx][2]\n return self.field[i][j][k][0] == idx\n\n # 全エージェントの移動方向の配列を受け取り移動させてフィールドを更新する\n # -1の場合は移動させません(0~3は移動APIのドキュメント記載と同じです)\n def move(self, move: [int]):\n # エージェントの移動処理\n for idx in range(6):\n if move[idx] == -1:\n continue\n self.rotate_agent(idx, move[idx])\n self.move_forward(idx)\n\n # フィールドの更新処理\n for idx in range(6):\n if move[idx] == -1:\n continue\n ok = True\n for j in range(6):\n if idx == j or move[j] == -1 or not self.is_same_pos(self.agent[idx], self.agent[j]) or self.is_owned_cell(idx):\n continue\n # 移動した先にidx以外のエージェントがいる場合は修復しか行えないのでidxのエージェントのマスではない場合は更新しないようにフラグをfalseにする\n ok = False\n break\n\n if not ok:\n continue\n self.paint(idx)\n\n\nclass Bot:\n def solve(self):\n game_id = get_game_id()\n next_d = random.randint(0, 3)\n while True:\n # 移動APIを呼ぶ\n move = call_move(game_id, next_d)\n print('status = {}'.format(move['status']), file=sys.stderr, flush=True)\n if move['status'] == \"already_moved\":\n continue\n elif move['status'] != 'ok':\n break\n print('turn = {}'.format(move['turn']), file=sys.stderr, flush=True)\n print('score = {} {} {} {} {} {}'.format(move['score'][0], move['score'][1], move['score'][2], move['score'][3], move['score'][4], move['score'][5]), file=sys.stderr, flush=True)\n # 4方向で移動した場合を全部シミュレーションする\n best_c = -1\n best_d = []\n for d in range(4):\n m = State(move['field'], move['agent'])\n m.move([d, -1, -1, -1, -1, -1])\n # 自身のエージェントで塗られているマス数をカウントする\n c = 0\n for i in range(6):\n for j in range(N):\n for k in range(N):\n if m.field[i][j][k][0] == 0:\n c += 1\n # 最も多くのマスを自身のエージェントで塗れる移動方向のリストを保持する\n if c > best_c:\n best_c = c\n best_d = [d]\n elif c == best_c:\n best_d.append(d)\n # 最も多くのマスを自身のエージェントで塗れる移動方向のリストからランダムで方向を決める\n next_d = random.choice(best_d)\n\n\nif __name__ == \"__main__\":\n bot = Bot()\n bot.solve()","repo_name":"ayanamizuta/cpro","sub_path":"tenka1/tenka1-2022/main_orginal.py","file_name":"main_orginal.py","file_ext":"py","file_size_in_byte":7824,"program_lang":"python","lang":"ja","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74421692410","text":"import pyaudio\nimport wave\nimport sys\n\n\nCHUNK_SIZE = 1024\n\ndef play_wav(wav_filename, chunk_size=CHUNK_SIZE):\n '''\n Play (on the attached system sound device) the WAV file\n named wav_filename.\n '''\n\n try:\n wf = wave.open(wav_filename, 'rb')\n except IOError as ioe:\n sys.stderr.write('IOError on file ' + wav_filename + '\\n' + \\\n str(ioe) + '. Skipping.\\n')\n return\n except EOFError as eofe:\n sys.stderr.write('EOFError on file ' + wav_filename + '\\n' + \\\n str(eofe) + '. Skipping.\\n')\n return\n\n # Instantiate PyAudio.\n p = pyaudio.PyAudio()\n\n # Open stream.\n stream = p.open(format=p.get_format_from_width(wf.getsampwidth()),\n channels=wf.getnchannels(),\n rate=wf.getframerate(),\n output=True)\n\n data = wf.readframes(chunk_size)\n while len(data) > 0:\n stream.write(data)\n data = wf.readframes(chunk_size)\n\n # Stop stream.\n stream.stop_stream()\n stream.close()\n\n # Close PyAudio.\n p.terminate()\n\n","repo_name":"keiffster/talk-y","sub_path":"src/talky/utils/wav.py","file_name":"wav.py","file_ext":"py","file_size_in_byte":1045,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"39911862913","text":"from aiogram.utils.emoji import emojize\nfrom aiogram.utils.markdown import text\n\n# QUESTIONNAIRE MESSAGES:\nquestionnaire_start = text(emojize(\"Привіт!👋\\nРозкажи, будь ласка: де та з ким ти хочеш жити 😏\"))\nquestionnaire_floor = text(emojize(\"Де саме ти хотів би оселитися?🧳\"))\nquestionnaire_room = text(emojize(\"Який номер кімнати в тебе з щасливих?🎰\"))\nquestionnaire_roommates = text(emojize(\"Останнє, з ким ти хотів/ла би жити? 🐸\"))\nsave_button = text(emojize(\"Зберегти📝\"))\nquestionnaire_goodbye = text(emojize(\"Інформація буде передана на обробку, дякую за інформацію 📃\"))\n\n# QUESTIONNAIRE BUTTONS:\n\nplace1 = text(emojize(\"На Майбороди, з коліверами\"))\nplace2 = text(emojize(\"У горах, наодинці🌄\"))\n\n# FOR PLACE 1 (МАЙБОРОДА)\nfloor1 = text(emojize(\"На 1️⃣ поверсі\"))\nfloor2 = text(emojize(\"На 2️⃣ поверсі\"))\n# FOR PLACE 2 (ГОРА)\nmountain_bot = text(emojize(\"У підніжжя 🏞\"))\nmountain_msg = text(emojize(\"Обери собі компаньона!🙌\"))\nmountain_top = text(emojize(\"На вершині 🌄\"))\n\n# FOR FLOOR1 (МАЙБОРОДА)\nroom1 = text(emojize(\"1️⃣\").encode(\"utf-8\"))\nroom2 = text(emojize(\"2️⃣\"))\nroom3 = text(emojize(\"3️⃣\"))\n# FOR FLOOR2 (МАЙБОРОДА)\nroom4 = text(emojize(\"4️⃣\"))\nroom5 = text(emojize(\"5️⃣\"))\nroom6 = text(emojize(\"6️⃣\"))\n\n# MAIBORODA FRIENDS\nroom1_friend1 = \"Алекс\"\nroom1_friend2 = \"Фродо\"\nroom2_friend1 = \"Лазло\"\nroom2_friend2 = \"Каска\"\nroom3_friend1 = \"Іван\"\nroom3_friend2 = \"Василь\"\nroom4_friend1 = \"Олекса\"\nroom4_friend2 = \"Філіп\"\nroom5_friend1 = \"Фердинанд\"\nroom5_friend2 = \"Кіра\"\nroom6_friend1 = \"Леся\"\nroom6_friend2 = \"Валерій\"\n\n# FOR MOUNTAIN (bot)\nriver = text(emojize(\"Біля річки 🐸\"))\ntree = text(emojize(\"На дереві 🌳\"))\n# FOR MOUNTAIN (top)\nigloo = text(emojize(\"В іглу ☃\"))\ncave = text(emojize(\"У печері 🗻\"))\n\n# FOR MOUNTAIN (RIVER) PET\npet_river1 = text(emojize(\"Золоту рибку!🐡\"))\npet_river2 = text(emojize(\"Медведя!🐻\"))\n# FOR MOUNTAIN (TREE) PET\npet_tree1 = text(emojize(\"Білочку🐿\"))\npet_tree2 = text(emojize(\"Сову🦉\"))\n# IGLOO PET\npet_igloo1 = text(emojize(\"Чукчу!⛄\"))\npet_igloo2 = text(emojize(\"Привида!👻\"))\n# CAVE PET\npet_cave1 = text(emojize(\"Пані Самотність!🧘♂️\"))\npet_cave2 = text(emojize(\"Сніговика!⛄\"))\n\n# SINGLE-PURPOSE MESSAGES:\nstart_message = text(emojize(\"Привіт!👋\"))\nregistered_message = \"Давно не бачилися!\"\nhave_a_nice_lecture = text(emojize(\"Продуктивної лекції тобі! 😉\"))\nvote_thank_you = text(emojize(\"Дякую, твій голос враховано!⚖\"))\nFINISHED = \"Сподіваюся, тобі сподобалася лекція\"\nLECTURE_START = \"ЛЕКЦІЯ ПОЧАЛАСЯ, УАЛІВЕЦЬ ВСТАВАЙ!\"\nNEW_VALUE = \"Прохання проголосувати за/проти затвердження нової цінності: Будь Програмістом!\"\nza = \"За\"\nproty = \"Проти\"\n\n# Q&A RESPONSES:\nQ_and_A_welcoming_message = text(emojize(\"Привіт, тут ти можеш задати стільки питань, скільки забажаєш\\n\"\n \"Коли закінчиш, натисни клавішу 'Вийти'😉\"))\nQ_and_A_confirmation_message = text(emojize(\"Записав!📃\"))\nexit_Q = \"Вийти\"\nQ_and_A_goodbye_message = text(emojize(\"Дякую.\\nЗадані питання будуть доставленими 🧭\"))\n# TEMPLATE = text(emojize())\n\npresence = \"PRESENT09\"\nvoter = \"FOR02937\"\nagainst = \"AGAINST02937\"\nMESSAGES = {\n\n \"start_message\": start_message,\n \"registered_message\": registered_message,\n \"have_a_nice_lecture\": have_a_nice_lecture,\n \"vote_thank_you\": vote_thank_you,\n \"FINISHED\": FINISHED,\n \"LECTURE_START\": LECTURE_START,\n \"NEW_VALUE\": NEW_VALUE,\n \"za\": za,\n \"proty\": proty,\n\n # callback queries:\n \"presence\": presence,\n \"voter\": voter,\n \"against\": against,\n\n # Q&A RESPONSES:\n \"Q_and_A_welcoming_message\": Q_and_A_welcoming_message,\n \"Q_and_A_confirmation_message\": Q_and_A_confirmation_message,\n \"exit_Q\": exit_Q,\n \"Q_and_A_goodbye_message\": Q_and_A_goodbye_message,\n # QUESTIONNAIRE RESPONSES:\n \"questionnaire_start\": questionnaire_start,\n \"questionnaire_floor\": questionnaire_floor,\n \"questionnaire_room\": questionnaire_room,\n \"questionnaire_roommates\": questionnaire_roommates,\n \"save_button\": save_button,\n \"questionnaire_goodbye\": questionnaire_goodbye,\n # QUESTIONNAIRE BUTTONS\n \"place1\": place1,\n \"place2\": place2,\n \"floor1\": floor1,\n \"floor2\": floor2,\n \"mountain_bot\": mountain_bot,\n \"mountain_top\": mountain_top,\n \"mountain_msg\": mountain_msg,\n\n}\nROOMS = {\n \"room1\": room1,\n \"room2\": room2,\n \"room3\": room3,\n \"room4\": room4,\n \"room5\": room5,\n \"room6\": room6,\n \"river\": river,\n \"tree\": tree,\n \"igloo\": igloo,\n \"cave\": cave,\n}\n\nPETS_AND_FRIENDS = {\n \"pet_river1\": pet_river1,\n \"pet_river2\": pet_river2,\n \"pet_tree1\": pet_tree1,\n \"pet_tree2\": pet_tree2,\n \"pet_igloo1\": pet_igloo1,\n \"pet_igloo2\": pet_igloo2,\n \"pet_cave1\": pet_cave1,\n \"pet_cave2\": pet_cave2,\n\n \"room1_friend1\": room1_friend1,\n \"room1_friend2\": room1_friend2,\n \"room2_friend1\": room2_friend1,\n \"room2_friend2\": room2_friend2,\n \"room3_friend1\": room3_friend1,\n \"room3_friend2\": room3_friend2,\n \"room4_friend1\": room4_friend1,\n \"room4_friend2\": room4_friend2,\n \"room5_friend1\": room5_friend1,\n \"room5_friend2\": room5_friend2,\n \"room6_friend1\": room6_friend1,\n \"room6_friend2\": room6_friend2,\n}\n","repo_name":"Salz0/Derzhavets_bot","sub_path":"messages.py","file_name":"messages.py","file_ext":"py","file_size_in_byte":6028,"program_lang":"python","lang":"uk","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"35884359949","text":"import numpy as np\nimport math\nfrom matplotlib import pyplot as plt\nfrom typing import List\nfrom TransferFn import TransferFn\n\nclass singleBinDFT(TransferFn):\n def __init__(self, centerFreq: float, samplingFreq: float, NCyclesSampled: int):\n self.fCenter = centerFreq\n self.fSamp = samplingFreq\n self.NCyclesSampled = NCyclesSampled\n\n self.relativeFreq = self.fCenter / self.fSamp\n NSamples = self.NCyclesSampled / self.relativeFreq\n self.timeData = np.linspace(0, NSamples - 1, NSamples)\n self.sinTable = np.sin(self.timeData * 2* math.pi * self.relativeFreq)\n self.cosTable = np.cos(self.timeData * 2* math.pi * self.relativeFreq)\n return\n\n def getAmp(t: np.array, y: np.array, relFreq: float) -> float:\n im = y * np.sin(t * 2* math.pi * relFreq)\n re = y * np.cos(t * 2* math.pi * relFreq)\n return 2 * math.sqrt(np.sum(im)**2 + np.sum(re)**2) / len(t)\n\n def apply(self, frequency: float) -> float:\n y: np.array = np.sin(self.timeData * 2* math.pi * frequency / self.fSamp)\n return singleBinDFT.getAmp(self.timeData, y, self.relativeFreq)\n\n","repo_name":"mdion1/LightsAndSoundProj","sub_path":"UsefulScripts/DFTFilterDesign/singleBinDFT.py","file_name":"singleBinDFT.py","file_ext":"py","file_size_in_byte":1148,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17599951125","text":"from app import app\nfrom dash.dependencies import Input, Output\nfrom dash import callback_context\n\nfrom models.building import geometry\nfrom models.building.buildingFactory import read_building_data_yaml, save_building_data_yaml\n\n@app.callback(\n Output('geometry_done_button_id', 'disabled'), \n Output('geometry_done_button_id', 'color'),\n Output(\"n_storys_id\", \"value\"),\n Output(\"story_height_id\", \"value\"),\n Output(\"area_win_a_id\", \"value\"),\n Output(\"area_win_b_id\", \"value\"),\n Output(\"area_win_c_id\", \"value\"),\n Output(\"area_win_d_id\", \"value\"),\n Input(\"n_storys_id\", \"value\"),\n Input(\"story_height_id\", \"value\"),\n Input(\"area_win_a_id\", \"value\"),\n Input(\"area_win_b_id\", \"value\"),\n Input(\"area_win_c_id\", \"value\"),\n Input(\"area_win_d_id\", \"value\"),\n Input(\"geometry_done_button_id\", \"n_clicks\"),\n)\ndef inputDone(n_storys, story_height, area_win_a,area_win_b,area_win_c, area_win_d, clicks):\n ctx = callback_context\n # check which input has triggered the callback\n button_id = ctx.triggered[0]['prop_id'].split('.')[0]\n\n # on pageload: fill input fields with building data\n if not ctx.triggered:\n building = read_building_data_yaml(\"userID\")\n n_storys = building['nStorys']\n story_height = building['thZones']['livingSpace']['storyHeight']\n area_win_a = building['thZones']['livingSpace']['transPlanes']['windowA']['area']\n area_win_b = building['thZones']['livingSpace']['transPlanes']['windowB']['area']\n area_win_c = building['thZones']['livingSpace']['transPlanes']['windowC']['area']\n area_win_d = building['thZones']['livingSpace']['transPlanes']['windowD']['area']\n return False, \"success\", n_storys, story_height, area_win_a, area_win_b, area_win_c, area_win_d\n\n # disable button if one input field is empty\n if not n_storys or not story_height or not area_win_a or not area_win_b or not area_win_c or not area_win_d:\n return True, \"primary\", n_storys, story_height, area_win_a, area_win_b, area_win_c, area_win_d\n\n # save building data on button click\n if button_id == \"geometry_done_button_id\":\n # calculate geometry values\n height = geometry.height_from_story(n_storys, story_height)\n\n building = read_building_data_yaml(userID='userID')\n perimeter = building['perimeter']\n groundArea = building['groundArea']\n\n heatedArea = groundArea * n_storys\n volume = groundArea * height\n facadearea = geometry.facade_area(perimeter, height) - sum([area_win_a, area_win_b, area_win_c, area_win_d])\n\n\n building['nStorys'] = n_storys\n building['thZones']['livingSpace']['storyHeight'] = story_height\n building['thZones']['livingSpace']['opaquePlanes']['facade']['area'] = facadearea\n building['thZones']['livingSpace']['opaquePlanes']['roof']['area'] = groundArea\n building['thZones']['livingSpace']['opaquePlanes']['floor']['area'] = groundArea\n building['thZones']['livingSpace']['floorArea'] = heatedArea\n building['thZones']['livingSpace']['volume'] = volume\n building['thZones']['livingSpace']['transPlanes']['windowA']['area'] = area_win_a\n building['thZones']['livingSpace']['transPlanes']['windowB']['area'] = area_win_b\n building['thZones']['livingSpace']['transPlanes']['windowC']['area'] = area_win_c\n building['thZones']['livingSpace']['transPlanes']['windowD']['area'] = area_win_d\n save_building_data_yaml(building)\n\n return False, \"success\", n_storys, story_height, area_win_a, area_win_b, area_win_c, area_win_d\n\n # activate button if all inputs are filled\n else:\n return False, \"primary\", n_storys, story_height, area_win_a, area_win_b, area_win_c, area_win_d\n\n\n ","repo_name":"rehomewebapp/REhome","sub_path":"WebApp/controllers/geometry_controller.py","file_name":"geometry_controller.py","file_ext":"py","file_size_in_byte":3779,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20791958864","text":"import xml.etree.ElementTree as ET\r\nimport pandas as pd\r\nimport csv\r\n\r\nFILE = 'Posts.xml'\r\nCOLS = ['Id', 'PostTypeId', 'ParentId', 'CreationDate', 'Score', 'Body', 'OwnerUserId', 'LastEditorUserId',\r\n 'LastEditDate', 'LastActivityDate', 'CommentCount', 'ContentLicense']\r\n\r\n# list of acceptedanswerId extraction from targeted questions' file\r\ndata = pd.read_csv('output.csv') # skip header row\r\ndf = pd.DataFrame(data)\r\nx=df.iloc[:,2].dropna()\r\nansIdLst=set(x)\r\nprint(len(ansIdLst))\r\n\r\n\r\ncontext = ET.iterparse(FILE, events=(\"start\", \"end\"),\r\n parser=ET.XMLParser(encoding='utf-8'))\r\n\r\nwith open('outputAnswer.csv', 'w', newline='', encoding='utf-8') as csvfile:\r\n csvwriter = csv.writer(csvfile)\r\n csvwriter.writerow(COLS)\r\n _, root = next(context)\r\n for event, elem in context:\r\n if event == \"end\" and elem.tag == \"row\":\r\n # candidate Id for answerpost\r\n ansId = elem.attrib.get('Id', 'None')\r\n if int(ansId) in ansIdLst: \r\n data = []\r\n for col in COLS:\r\n data.append(elem.attrib.get(col, ''))\r\n csvwriter.writerow(data)\r\n # progress\r\n if int(elem.attrib['Id']) % 100000 == 0:\r\n print('done', elem.attrib['Id'])\r\n elem.clear()\r\n root.clear()","repo_name":"iamsanjaymalakar/LowCodeChallengesMSR21","sub_path":"Garbage/Initial Codes/extract-answers.py","file_name":"extract-answers.py","file_ext":"py","file_size_in_byte":1351,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"33391762910","text":"def selection_algor(nums):\n \"\"\"\n У цьому алгоритмі список (або масив) ділиться на дві частини: список з відсортованими елементами і\n список з елементами, які тільки потрібно сортувати. Спочатку шукається найменший елемент у другому.\n Він додається в кінці першого. Таким чином алгоритм поступово формує список від меншого до більшого.\n \"\"\"\n # i -> кількість відсортованих значень\n for i in range(len(nums)):\n lowest_value_index = i # допустимо, що найменший елемент- перший\n for j in range(i + 1, len(nums)): # перебірка несортованих елементів\n if nums[j] < nums[lowest_value_index]:\n lowest_value_index = j\n nums[i], nums[lowest_value_index] = nums[lowest_value_index], nums[i] # найменший міняємо з першим елементом\n\n\nrandom_list_of_nums = [12, 8, 3, 20, 11]\nselection_algor(random_list_of_nums)\nprint(random_list_of_nums)\n\n\n#_______________________________________________________________________________________________________________________\ndef findSmallest(inputes_list):\n smallest = inputes_list[0]\n smallest_index = 0\n for index in range(1, len(inputes_list)):\n if inputes_list[index] < smallest:\n smallest = inputes_list[index]\n smallest_index = index\n return smallest_index\n\n\ndef selection_sort(input_list):\n new_list = []\n for i in range(len(input_list)):\n smallest = findSmallest(input_list)\n new_list.append(input_list.pop(smallest))\n return new_list\n\n\nprint(selection_sort([2, 6, 12, 10]))\n","repo_name":"irynavaskiv1/diff_tasks","sub_path":"1_algorithms/sorting_methods/selection_sort.py","file_name":"selection_sort.py","file_ext":"py","file_size_in_byte":1915,"program_lang":"python","lang":"uk","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"70904039612","text":"import concurrent.futures\nimport hashlib\nimport sched\n\nimport bs4\nimport cachecontrol\nimport feedparser\nimport requests\nimport requests.exceptions\n\nfrom feedbuffer import settings, database, log\n\n_logger = log.get_logger(__name__)\n_session = cachecontrol.CacheControl(requests.Session())\n_session.headers['User-Agent'] = settings.USER_AGENT\nexecutor = concurrent.futures.ThreadPoolExecutor(max_workers=settings.MAXIMUM_UPDATE_WORKERS)\nscheduled = {}\nscheduler = sched.scheduler()\n\n# XML-processing instructions have to end with \"?>\". The original code erroneously ends them with \">\" which leads to\n# errors in almost all parsers, including BeautifulSoup with the lxml treebuilder itself -- so we fix this at runtime.\nbs4.element.ProcessingInstruction.SUFFIX = '?>'\n\n\ndef extract_feed_entries(soup):\n return [item.extract() for item in soup(['item', 'entry'])]\n\n\ndef update_feed(url):\n try:\n response = _session.get(url, timeout=settings.REQUEST_TIMEOUT)\n except requests.exceptions.Timeout:\n return\n\n # Don't let requests do the content decoding, instead just supply the encoding detected by requests and let\n # BeautifulSoup and the treebuilder do their thing. For example: BeautifulSoup4 with the lxml treebuilder only\n # correctly parses content with tags when it can decode the bytes by itself.\n try:\n soup = bs4.BeautifulSoup(response.content, 'xml', from_encoding=response.encoding)\n except UnicodeDecodeError:\n soup = bs4.BeautifulSoup(response.content, 'xml', from_encoding=response.apparent_encoding)\n\n entries = extract_feed_entries(soup)\n # TODO: Remove the feedparser dependency and figure out all ways to get access to the feed item id\n parsed_feed = feedparser.parse(response.text)\n is_rss = parsed_feed.version.startswith('rss')\n\n entry_ids = []\n for index, parsed_entry in enumerate(parsed_feed.entries):\n id_ = parsed_entry.get('id', None)\n\n # id might be non-existent or simply empty, make sure to handle both cases correctly\n if not id_:\n id_ = hashlib.sha1(entries[index].encode(settings.ENCODING)).hexdigest()\n _logger.info('No identifier found for entry %d of %s. Inserting SHA-1 id: %s...', index, url, id_)\n id_tag = soup.new_tag('guid' if is_rss else 'id')\n id_tag.string = id_\n entries[index].append(id_tag)\n entry_ids.append(id_)\n\n # Fix missing RSS channel element\n if is_rss and not soup.find('channel'):\n _logger.info('No RSS channel element found for %s. Inserting channel element...', url)\n rss = soup.find('rss')\n rss.append(rss.new_tag('channel'))\n\n database.update_feed(url, str(soup), zip(entry_ids, (str(entry) for entry in entries)))\n\n\ndef update_and_reschedule_feed(url):\n executor.submit(update_feed, url)\n schedule_feed_update(url)\n\n\ndef schedule_feed_update(url):\n if url in scheduled:\n if scheduled[url] in scheduler.queue:\n try:\n scheduler.cancel(scheduled[url])\n except ValueError:\n pass\n del scheduled[url]\n\n if not database.feed_exists(url):\n return\n\n feed = database.get_feed(url)\n event = scheduler.enter(feed.update_interval, 1, update_and_reschedule_feed, (url,))\n scheduled[url] = event\n\n\ndef generate_feed(feed_data, entries):\n feed = bs4.BeautifulSoup(feed_data, 'xml')\n\n # Find RSS channel element directly\n root = feed.find(['channel', 'feed'])\n for entry in entries:\n entry = bs4.BeautifulSoup(entry, 'xml')\n entry = entry.find(['item', 'entry'])\n root.insert(len(root.contents), entry)\n\n return str(feed).encode(settings.ENCODING)\n","repo_name":"cryzed/Feedbuffer","sub_path":"feedbuffer/core.py","file_name":"core.py","file_ext":"py","file_size_in_byte":3720,"program_lang":"python","lang":"en","doc_type":"code","stars":19,"dataset":"github-code","pt":"78"}
+{"seq_id":"39491219614","text":"from crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Submit\nfrom django import forms\nfrom django.core.validators import FileExtensionValidator\n\n\nclass PipelineForm(forms.Form):\n input_file = forms.FileField(\n label=\"Input file\",\n validators=[FileExtensionValidator([\"csv\", \"tsv\"])]\n )\n\n def __init__(self, *args, **kwargs):\n super(PipelineForm, self).__init__(*args, **kwargs)\n self.helper = FormHelper()\n self.helper.form_id = \"pipeline_form\"\n self.helper.form_class = \"pipeline\"\n self.helper.form_method = \"post\"\n self.helper.form_action = \"pipeline:submit\"\n\n self.helper.add_input(Submit('submit', 'Submit'))\n","repo_name":"ivansg44/lexmapr_django","sub_path":"lexmapr_django/pipeline/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":706,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"37126448685","text":"from glob import glob\nimport argparse\nfrom shutil import move\nfrom os.path import join, basename\nfrom os import makedirs\nimport numpy as np\n\n\nparser = argparse.ArgumentParser('Export model', add_help=False)\nparser.add_argument('-i', '--input', type=str, help='Input (directory).')\nparser.add_argument('-o', '--output', type=str, help='Output (directory).')\nparser.add_argument('-n', '--num', type=int, help='Number of items.')\nparser.add_argument('-s', '--seed', default=None, type=int, help='Random seed.')\nargs = parser.parse_args()\n\n\nfiles = sorted(glob(join(args.input, 'images', '*.*')))\nif args.seed is not None:\n np.random.seed(args.seed)\nsel = np.random.choice(files, args.num, replace=False)\nmakedirs(join(args.output, 'images'), exist_ok=True)\nmakedirs(join(args.output, 'labels'), exist_ok=True)\nfor f in sel:\n f_, = glob(join(args.input, 'labels', '.'.join(basename(f).split('.')[:-1]) + '*'))\n f_dst = join(args.output, 'images', basename(f))\n f_dst_ = join(args.output, 'labels', basename(f_))\n\n print(f, '-->', f_dst)\n move(f, f_dst)\n print(f_, '-->', f_dst_)\n move(f_, f_dst_)\n print()\n","repo_name":"FZJ-INM1-BDA/neurips22-cell-seg","sub_path":"preprocess_val_data.py","file_name":"preprocess_val_data.py","file_ext":"py","file_size_in_byte":1129,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"24202007304","text":"from çalışan import çalışan\r\n#çalışan classını inherit eden maviyaka classını oluşturuyorum\r\nclass maviyaka(çalışan):\r\n # istenilen bilgileri içerecek init metodunu yazıyorum\r\n def __init__(self, tc_no, ad, soyad, yas, cinsiyet, uyruk, sektor, tecrube, maas, yeni_maas, yipranma_payi):\r\n # insan classından inherit ettiğimiz değerlerin kullanılabilmesi için super metodunu kullanıyorum\r\n super().__init__(tc_no, ad, soyad, yas, cinsiyet, uyruk, sektor, tecrube, maas, yeni_maas)\r\n self.__yipranma_payi = yipranma_payi\r\n #istediğimiz aralıkta yıpranma payı değeri girilmesi için while döngüsüyle kontrol edip hatalı değerde tekrar girilmesini sağlıyorum\r\n while self.__yipranma_payi <= 0 or self.__yipranma_payi >= 1:\r\n self.__yipranma_payi = float(input(\"Lütfen 0 ile 1 arasında bir yıpranma payı değeri girin: \"))\r\n\r\n # private değerleri güncellemek için get ve set metotlarını yazıyorum\r\n def get_yipranma_payi(self):\r\n return self.__yipranma_payi\r\n\r\n def set_yipranma_payi(self, yeni_yipranma_payi):\r\n if yeni_yipranma_payi <= 0 or yeni_yipranma_payi >= 1:\r\n yeni_yipranma_payi = float(input(\"Lütfen 0 ile 1 arasında bir yıpranma payı değeri girin: \"))\r\n self.__yipranma_payi = yeni_yipranma_payi\r\n\r\n # yeni maaş değerini hesaplamak için zam_hakki metodunu yazıyorum\r\n def zam_hakki(self):\r\n # tecrube ve maas değerlerini bizden istenilen şekilde karşılaştırıp yıpranma payı değerini de kullanarak yeni maaş değeri buluyorum\r\n if self.get_tecrube() / 12 < 2:\r\n zam_orani = self.__yipranma_payi * 10\r\n self.set_yeni_maas(self.get_maas() + self.get_maas() * (zam_orani / 100))\r\n\r\n elif self.get_tecrube() / 12 >= 2 and self.get_tecrube() < 4 and self.get_maas() < 15000:\r\n zam_orani = (self.get_maas() % self.get_tecrube()) / 2 + (self.__yipranma_payi * 10)\r\n self.set_yeni_maas(self.get_maas() + self.get_maas() * (zam_orani / 100))\r\n\r\n elif self.get_tecrube() / 12 >= 4 and self.get_maas() < 25000:\r\n zam_orani = (self.get_maas() % self.get_tecrube()) / 3 + (self.__yipranma_payi * 10)\r\n self.set_yeni_maas(self.get_maas() + self.get_maas() * (zam_orani / 100))\r\n\r\n else:\r\n self.set_yeni_maas(self.get_maas())\r\n\r\n # istenilen değerleri string halinde döndürecek str metodunu yazıyorum\r\n def __str__(self):\r\n self.zam_hakki()\r\n return f\"AD: {self.get_ad()}, SOYAD: {self.get_soyad()}, TECRUBE(AY): {self.get_tecrube()} AY, YENI MAAS: {self.get_yeni_maas()}\"\r\n\r\n\r\n\r\n","repo_name":"YusufKarahan20/final-proje","sub_path":"maviyaka.py","file_name":"maviyaka.py","file_ext":"py","file_size_in_byte":2672,"program_lang":"python","lang":"tr","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"35334639685","text":"#Аппроксимация для таблицы значение - холодная кривая\n# rho - x, T - y\n\n\n\n# Import `load_workbook` module from `openpyxl`\nfrom openpyxl import load_workbook\nfrom math import log10\n\n# Load in the workbook\nwb = load_workbook('C:/Users/xaha9/Desktop/ШР/deep_lom/aluminium.xlsx')\n\n# Get sheet names\n#print(wb.get_sheet_names())\n\n# Load a specific sheet by name\n\nsheet = wb.get_sheet_by_name('energy_ferrum_исходник')\n## Retrieve the value of a certain cell\n#print(sheet['A2'].value)\n#\n## Select element 'B2' of your sheet\n#c = sheet['B2']\n#\n## Retrieve the row number of your element\n#print(c.row)\n#\n## Retrieve the column letter of your element\n#print(c.column)\n#\n## Retrieve the coordinates of the cell\n#print(c.coordinate)\n\n\ndef bilinear_interpolation_energy(T, rho):\n x1 = 0\n x2 = 0\n y1 = 0\n y2 = 0\n\n for cellObj in sheet['B1':'OL1']:\n for cell in cellObj:\n if cell.value <= log10(rho):\n x1 = cell\n else:\n\n break\n\n\n x2 = sheet.cell(row=x1.row, column=x1.column + 1)\n\n print(\"x1 = \", x1, \"X2 = \", x2)\n for cellObj in sheet['A2':'A502']:\n for cell in cellObj:\n if cell.value >= log10(T):\n y1 = cell\n else:\n break\n\n y2 = sheet.cell(row=y1.row + 1, column=y1.column)\n print(\"y1 = \", y1, \"y2 = \", y2)\n\n\n def interpolate(x1, x2, y1, y2, T, rho):\n\n f1 = (10 **(x2.value) - rho) / (10 **(x2.value) - 10 **(x1.value)) * sheet.cell(row=y1.row, column=x1.column).value + (rho - 10 **(x1.value)) / (10 **(x2.value) - 10 **(x1.value)) * sheet.cell(row=y1.row, column=x2.column).value\n f2 = (10 **(x2.value) - rho) / (10 **(x2.value) - 10 **(x1.value)) * sheet.cell(row=y2.row, column=x1.column).value + (rho - 10 **(x1.value)) / (10 **(x2.value) - 10 **(x1.value)) * sheet.cell(row=y2.row, column=x2.column).value\n\n return (10 **(y2.value) - T) / (10 **(y2.value) - 10 **(y1.value)) * f1 + (T - 10 **(y1.value)) / (10 **(y2.value) - 10 **(y1.value)) * f2\n\n return interpolate(x1, x2, y1, y2, T, rho)\n\n#print(bilinear_interpolation(100, 0.01))\n\n\n\n\n","repo_name":"markoshura/deep_vlom","sub_path":"bilinear_interpolation_energy.py","file_name":"bilinear_interpolation_energy.py","file_ext":"py","file_size_in_byte":2164,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"29133258874","text":"import contextlib\nfrom functools import partial\nfrom .report import current_report\nfrom ..utils import (\n clone_tensors,\n map_structure_and_replace_key,\n flatten,\n for_each_grad_tensor,\n)\nimport json\nimport numpy\nimport paddle\nimport torch\n\n\n@contextlib.contextmanager\ndef register_hooker(model):\n marker = model.marker\n\n remove_handles = []\n idx = 0\n # traversal_for_hook includes layers which we need add pre and post hook\n # for model structure, but not info hook (they are in black list)\n models = list(marker.traversal_for_hook())\n for mod in models:\n pre_handle = mod.register_forward_pre_hook(partial(pre_structure_hook))\n if mod not in marker.black_list:\n handle = mod.register_forward_post_hook(partial(info_hook, net_id=idx))\n remove_handles.append(handle)\n post_handle = mod.register_forward_post_hook(partial(post_structure_hook))\n remove_handles.extend([pre_handle, post_handle])\n idx += 1\n yield\n for h in remove_handles:\n h.remove()\n\n\n\"\"\"\n hooks used to build module structure\n\"\"\"\n\n\ndef pre_structure_hook(layer, input):\n report = current_report()\n report.stack.push_layer(layer)\n if layer in report.marker.layer_map:\n report.stack._top().layer_type = \"in map\"\n return None\n\n\ndef post_structure_hook(layer, input, output):\n report = current_report()\n retval = report.stack.pop_layer(layer)\n if retval in report.marker.black_list:\n report.stack._top().children.pop()\n return None\n\n\n\"\"\"\n hook for record forward infos\n\"\"\"\n\n# do not enter api layer which is triggered under info_hook\n__in_info_hook__ = False\n\n\ndef info_hook(model, input, output, net_id):\n \"\"\"\n Notice: the model is a origin layer/module, not ProxyModel\n \"\"\"\n global __in_info_hook__\n if __in_info_hook__:\n return None\n\n report = current_report()\n\n if report is None or report.stack._top() is None:\n return None\n\n # if this api is not processing tensors, do not create report\n if output is None or all([not isinstance(x, (paddle.Tensor, torch.Tensor)) for x in flatten(output)]):\n return None\n\n # if an api under black_list_recursively, do not create report\n # a layer under black_list_recursively will not register this hook, except it is a mapped layer\n # report.stack._top().net can not be an api layer !!!\n if report.stack._top().net in report.marker.black_list_recursively and hasattr(model, \"__api__\"):\n return None\n\n # if this api is called under layer/module provided by framework, skip it\n python_module = report.stack._top().net.__module__\n if hasattr(model, \"__api__\") and (python_module.startswith(\"paddle.\") or python_module.startswith(\"torch.\")):\n return None\n\n __in_info_hook__ = True\n\n # if current model is an api layer, we do not want to hold it\n if hasattr(model, \"__api__\"):\n _model = padiff_layer_str(model)\n else:\n _model = model\n\n new_in = clone_tensors(input)\n new_out = clone_tensors(output)\n fwd_item = report.put_item(\"forward\", new_in, new_out, _model, net_id)\n bwd_item = report.put_item(\"backward\", new_in, new_out, _model, net_id)\n bwd_item.set_forward(fwd_item)\n\n report.stack.push_api(_model, fwd_item, bwd_item)\n\n for i, (t,) in enumerate(for_each_grad_tensor(input)):\n t.register_hook(partial(tensor_hook, bwd_item=bwd_item, nth_tensor=i, net_id=net_id))\n\n # if under single step forward guard\n if single_step_state() == \"forward\" and net_id != -1:\n # two report_item with same id, the step_idx should be corresponded\n step_idx = len(list(filter(lambda x: x.type == \"forward\" and x.net_id == net_id, report.items))) - 1\n base_report_node = find_base_report_node(net_id, step_idx)\n\n retval = map_structure_and_replace_key(replace_forward_output(base_report_node), output, output)\n __in_info_hook__ = False\n return retval\n else:\n __in_info_hook__ = False\n return None\n\n\n\"\"\"\n hook for record backward infos\n\"\"\"\n\n\ndef tensor_hook(x_grad, bwd_item, nth_tensor, net_id):\n new_grad = clone_tensors(x_grad)\n bwd_item.set_input_grads(nth_tensor, new_grad[0])\n\n if single_step_state() == \"backward\" and net_id != -1:\n report = current_report()\n step_idx = (\n list(filter(lambda x: x.type == \"backward\" and x.net_id == net_id, report.items)).index(bwd_item) - 1\n )\n base_report_node = find_base_report_node(net_id, step_idx)\n\n value = numpy.load(base_report_node[\"bwd_grads\"][nth_tensor])\n if isinstance(x_grad, paddle.Tensor):\n return paddle.to_tensor(value)\n else:\n return torch.as_tensor(value, device=x_grad.device)\n\n return x_grad\n\n\n\"\"\"\n utils\n\"\"\"\n\n\ndef padiff_layer_str(model):\n if isinstance(model, paddle.nn.Layer):\n return PaddleLayerStr(model)\n else:\n return TorchModuleStr(model)\n\n\nclass PaddleLayerStr(paddle.nn.Layer):\n def __init__(self, net):\n super(PaddleLayerStr, self).__init__()\n self.__name__ = net.__name__\n self.__api__ = net.__api__\n\n\nclass TorchModuleStr(torch.nn.Module):\n def __init__(self, net):\n super(TorchModuleStr, self).__init__()\n self.__name__ = net.__name__\n self.__api__ = net.__api__\n\n\nsingle_step_phase = \"\"\nsingle_step_base = None\n\n\n@contextlib.contextmanager\ndef SyncStepGuard(diff_phase, report_path):\n global single_step_phase, single_step_base\n try:\n old_phase = single_step_phase\n old_base = single_step_base\n\n with open(report_path + \"/\" + \"report.json\", \"r\") as report_file:\n report = json.load(report_file)\n\n single_step_phase = diff_phase\n single_step_base = split_by_net_id(report)\n\n yield\n finally:\n single_step_phase = old_phase\n single_step_base = old_base\n\n\ndef split_by_net_id(report):\n bucket = {}\n\n def _traversal(node, bucket):\n net_id = node[\"metas\"][\"net_id\"]\n if net_id == -1:\n return\n if net_id not in bucket:\n bucket[net_id] = [node]\n else:\n bucket[net_id].append(node)\n\n for child in node[\"children\"]:\n _traversal(child, bucket)\n\n for tree in report[\"tree\"]:\n _traversal(tree, bucket)\n\n for key in bucket:\n bucket[key].sort(key=lambda x: x[\"metas\"][\"fwd_step\"])\n\n return bucket\n\n\ndef single_step_state():\n return single_step_phase\n\n\ndef find_base_report_node(net_id, step_idx):\n global single_step_base\n return single_step_base[net_id][step_idx]\n\n\ndef replace_forward_output(node):\n numpy_file_list = node[\"fwd_outputs\"]\n cur_idx = 0\n\n def inner(input_):\n if isinstance(input_, (paddle.Tensor, torch.Tensor)):\n if cur_idx >= len(numpy_file_list):\n raise RuntimeError(\n \"In single step mode, try to replace tensor by dumpped numpy value, but the number of tensors and numpy is not equal. Maybe the models are not corresponded.\"\n )\n value = numpy.load(numpy_file_list[cur_idx])\n if isinstance(input_, paddle.Tensor):\n return paddle.to_tensor(value)\n else:\n return torch.as_tensor(value, device=input_.device)\n else:\n return input_\n\n return inner\n","repo_name":"WenmuZhou/PytorchOCR","sub_path":"padiff/report/hooks.py","file_name":"hooks.py","file_ext":"py","file_size_in_byte":7380,"program_lang":"python","lang":"en","doc_type":"code","stars":1192,"dataset":"github-code","pt":"78"}
+{"seq_id":"40585699916","text":"import json\nimport logging\nimport os\nfrom typing import Optional, Sequence, Union, List, Dict\n\nimport numpy as np\nimport pandas as pd\nimport torch\nfrom pytorch_lightning.callbacks import EarlyStopping\nfrom scvi.data import AnnDataManager\nfrom scvi.dataloaders import DataSplitter\nfrom torch.nn import functional as F\nfrom scvi.data.fields import (\n LayerField,\n CategoricalObsField,\n NumericalObsField,\n NumericalJointObsField,\n ObsmField,\n)\n\nfrom anndata import AnnData\nfrom scvi.model.base import BaseModelClass\nfrom scvi.train import TrainRunner\nfrom scvi.train._callbacks import SaveBestState\nfrom scvi.utils import setup_anndata_dsp\nfrom tqdm import tqdm\n\nfrom ._module import CPAModule\nfrom ._utils import CPA_REGISTRY_KEYS\nfrom ._task import CPATrainingPlan\nfrom ._data import AnnDataSplitter\n\nlogger = logging.getLogger(__name__)\nlogger.propagate = False\n\n\nclass CPA(BaseModelClass):\n \"\"\"CPA model\n\n Parameters\n ----------\n adata : Anndata\n Registered Annotation Dataset\n\n n_latent: int\n Number of latent dimensions used for drug and Autoencoder\n\n loss_ae: str\n Either `gauss` or `NB`. Autoencoder loss function.\n\n doser_type: str\n Type of doser network. Either `sigm`, `logsigm` or `mlp`.\n\n split_key : str, optional\n Key used to split the data between train test and validation.\n This must correspond to a observation key for the adata, composed of values\n 'train', 'test', and 'ood'. By default None.\n\n **hyper_params:\n CPA's hyper-parameters.\n\n\n Examples\n --------\n >>> import cpa\n >>> import scanpy as sc\n >>> adata = sc.read('dataset.h5ad')\n >>> adata = cpa.CPA.setup_anndata(adata,\n drug_key='condition',\n dose_key='dose_val',\n categorical_covariate_keys=['cell_type'],\n control_key='control'\n )\n >>> hyperparams = {'autoencoder_depth': 3, 'autoencoder_width': 256}\n >>> model = cpa.CPA(adata,\n n_latent=256,\n loss_ae='gauss',\n doser_type='logsigm',\n split_key='split',\n )\n \"\"\"\n cat_covars_encoders: dict = None\n drug_encoder: dict = None\n cont_covars: list = None\n\n def __init__(\n self,\n adata: AnnData,\n n_latent: int = 128,\n loss_ae: str = 'gauss',\n doser_type: str = 'logsigm',\n split_key: str = None,\n train_split: str = 'train',\n valid_split: str = 'test',\n test_split: str = 'ood',\n **hyper_params,\n ):\n super().__init__(adata)\n self.drug_encoder = CPA.drug_encoder\n self.cat_covars_encoders = CPA.cat_covars_encoders\n\n self.n_genes = adata.n_vars\n self.n_drugs = len(self.drug_encoder)\n self.split_key = split_key\n\n self.drugs = list(self.drug_encoder.keys())\n self.covars = {\n covar: list(self.cat_covars_encoders[covar].keys()) for covar in self.cat_covars_encoders.keys()\n }\n\n self.module = CPAModule(\n n_genes=self.n_genes,\n n_drugs=self.n_drugs,\n cat_covars_encoder=self.cat_covars_encoders,\n n_latent=n_latent,\n loss_ae=loss_ae,\n doser_type=doser_type,\n **hyper_params,\n ).float()\n\n train_indices, valid_indices, test_indices = None, None, None\n if split_key is not None:\n train_indices = np.where(adata.obs.loc[:, split_key] == train_split)[0]\n valid_indices = np.where(adata.obs.loc[:, split_key] == valid_split)[0]\n test_indices = np.where(adata.obs.loc[:, split_key] == test_split)[0]\n\n self.train_indices = train_indices\n self.valid_indices = valid_indices\n self.test_indices = test_indices\n\n self._model_summary_string = f\"Compositional Perturbation Autoencoder\"\n\n self.init_params_ = self._get_init_params(locals())\n\n self.epoch_history = None\n\n @classmethod\n @setup_anndata_dsp.dedent\n def setup_anndata(\n cls,\n adata: AnnData,\n perturbation_keys: Dict[str, str],\n use_counts: Optional[bool] = False,\n categorical_covariate_keys: Optional[List[str]] = [],\n continuous_covariate_keys: Optional[List[str]] = [],\n control_key: Optional[str] = None,\n deg_uns_key: Optional[str] = None,\n **kwargs,\n ):\n \"\"\"\n Annotation Data setup function\n\n Parameters\n ----------\n adata\n\n categorical_covariate_keys\n\n continuous_covariate_keys\n\n \"\"\"\n CPA_REGISTRY_KEYS.PERTURBATION_KEYS = perturbation_keys\n CPA_REGISTRY_KEYS.CAT_COV_KEYS = categorical_covariate_keys\n CPA_REGISTRY_KEYS.CONT_COV_KEYS = continuous_covariate_keys\n\n drug_key = perturbation_keys['perturbation']\n dose_key = perturbation_keys['dosage']\n\n drugs = adata.obs[drug_key]\n dosages = adata.obs[dose_key].astype(str)\n\n # get unique drugs\n drugs_names_unique = set()\n for d in np.unique(drugs):\n [drugs_names_unique.add(i) for i in d.split(\"+\")]\n drugs_names_unique = sorted(list(np.array(list(drugs_names_unique))))\n\n drugs_obsm = np.zeros((adata.n_obs, len(drugs_names_unique)))\n for i in tqdm(range(adata.n_obs)):\n cell_drugs = np.isin(drugs_names_unique, drugs[i].split('+'))\n cell_doses = np.array(dosages[i].split(\"+\")).astype(np.float32)\n drugs_obsm[i, cell_drugs] = cell_doses\n\n adata.obsm['drugs_doses'] = np.array(drugs_obsm)\n\n drug_encoder = {drug: i for i, drug in\n enumerate(drugs_names_unique)}\n\n setup_method_args = cls._get_setup_method_args(**locals())\n anndata_fields = \\\n [\n LayerField(registry_key=CPA_REGISTRY_KEYS.X_KEY, layer='counts' if use_counts else None,\n is_count_data=True if use_counts else False),\n ObsmField('drugs_doses', 'drugs_doses', is_count_data=False, correct_data_format=True)\n ] + \\\n [CategoricalObsField(registry_key=covar, attr_key=covar) for covar in categorical_covariate_keys] + \\\n [NumericalObsField(registry_key=covar, atrr_key=covar) for covar in continuous_covariate_keys]\n\n if control_key:\n anndata_fields.append(NumericalObsField(registry_key='control', attr_key=control_key))\n\n if deg_uns_key:\n mask = np.zeros((adata.n_obs, adata.n_vars))\n for i, cov_drug in tqdm(enumerate(adata.obs['cov_drug'].values)):\n if cov_drug in adata.uns[deg_uns_key].keys():\n mask[i] = adata.var.index.isin(adata.uns[deg_uns_key][cov_drug]).astype(np.int)\n else:\n mask[i] = 1\n\n adata.obsm['deg_mask'] = np.array(mask)\n\n anndata_fields.append(ObsmField(\"deg_mask\", \"deg_mask\", is_count_data=False, correct_data_format=True))\n\n adata_manager = AnnDataManager(\n fields=anndata_fields, setup_method_args=setup_method_args\n )\n adata_manager.register_fields(adata, **kwargs)\n cls.register_manager(adata_manager)\n\n cat_covar_encoders = {}\n for covar in categorical_covariate_keys:\n cat_covar_encoders[covar] = {c: i for i, c in enumerate(\n adata_manager.registry['field_registries'][covar]['state_registry']['categorical_mapping'])}\n\n CPA.cat_covars_encoders = cat_covar_encoders\n CPA.drug_encoder = drug_encoder\n CPA.cont_covars = continuous_covariate_keys\n\n def train(\n self,\n max_epochs: Optional[int] = None,\n use_gpu: Optional[Union[str, int, bool]] = None,\n train_size: float = 0.9,\n validation_size: Optional[float] = None,\n batch_size: int = 128,\n early_stopping: bool = False,\n plan_kwargs: Optional[dict] = None,\n hyperopt: bool = False,\n save_path: Optional[str] = None,\n **trainer_kwargs,\n ):\n \"\"\"\n Trains CPA on the given dataset\n\n Parameters\n ----------\n max_epochs: int\n Maximum number of epochs for training\n use_gpu: bool\n Whether to use GPU if available\n train_size: float\n Fraction of training data in the case of randomly splitting dataset to train/valdiation\n if `split_key` is not set in model's constructor\n validation_size: float\n Fraction of validation data in the case of randomly splitting dataset to train/valdiation\n if `split_key` is not set in model's constructor\n batch_size: int\n Size of mini-batches for training\n early_stopping: bool\n If `True`, EarlyStopping will be used during training on validation dataset\n plan_kwargs: dict\n `CPATrainingPlan` parameters\n save_path: str\n Path to save the model after the end of training\n \"\"\"\n if max_epochs is None:\n n_cells = self.adata.n_obs\n max_epochs = np.min([round((20000 / n_cells) * 400), 400])\n plan_kwargs = plan_kwargs if isinstance(plan_kwargs, dict) else dict()\n\n manual_splitting = (\n (self.valid_indices is not None)\n and (self.train_indices is not None)\n and (self.test_indices is not None)\n )\n if manual_splitting:\n data_splitter = AnnDataSplitter(\n self.adata_manager,\n train_indices=self.train_indices,\n valid_indices=self.valid_indices,\n test_indices=self.test_indices,\n batch_size=batch_size,\n use_gpu=use_gpu,\n )\n else:\n data_splitter = DataSplitter(\n self.adata_manager,\n train_size=train_size,\n validation_size=validation_size,\n batch_size=batch_size,\n use_gpu=use_gpu,\n )\n\n self.training_plan = CPATrainingPlan(self.module, self.cat_covars_encoders, **plan_kwargs)\n trainer_kwargs[\"early_stopping\"] = False\n trainer_kwargs['check_val_every_n_epoch'] = trainer_kwargs.get('check_val_every_n_epoch', 20)\n\n es_callback = EarlyStopping(monitor='cpa_metric',\n patience=trainer_kwargs['early_stopping_patience'],\n check_on_train_epoch_end=False,\n verbose=False,\n mode='max',\n )\n\n if 'callbacks' in trainer_kwargs.keys() and isinstance(trainer_kwargs.get('callbacks'), list):\n trainer_kwargs['callbacks'] += [es_callback]\n else:\n trainer_kwargs['callbacks'] = [es_callback]\n\n if save_path is None:\n save_path = './'\n\n checkpoint = SaveBestState(monitor='cpa_metric', mode='max', period=20, verbose=False)\n trainer_kwargs['callbacks'].append(checkpoint)\n\n runner = TrainRunner(\n self,\n training_plan=self.training_plan,\n data_splitter=data_splitter,\n max_epochs=max_epochs,\n use_gpu=use_gpu,\n early_stopping_monitor=\"cpa_metric\",\n early_stopping_mode='max',\n enable_checkpointing=True,\n enable_model_summary=True,\n **trainer_kwargs,\n )\n runner()\n\n self.epoch_history = pd.DataFrame().from_dict(self.training_plan.epoch_history)\n if save_path is not None:\n self.save(save_path, overwrite=True)\n\n @torch.no_grad()\n def get_latent_representation(\n self,\n adata: Optional[AnnData] = None,\n indices: Optional[Sequence[int]] = None,\n batch_size: Optional[int] = 32,\n ) -> np.ndarray:\n \"\"\"Returns the basal latent variable\n\n Parameters\n ----------\n adata : Optional[AnnData], optional\n [description], by default None\n indices : Optional[Sequence[int]], optional\n Optional indices, by default None\n batch_size : Optional[int], optional\n Batch size to use, by default None\n \"\"\"\n\n if self.is_trained_ is False:\n raise RuntimeError(\"Please train the model first.\")\n\n adata = self._validate_anndata(adata)\n if indices is None:\n indices = np.arange(adata.n_obs)\n scdl = self._make_data_loader(\n adata=adata, indices=indices, batch_size=batch_size, shuffle=False\n )\n\n latent_basal = []\n latent = []\n for tensors in scdl:\n inference_inputs = self.module._get_inference_input(tensors)\n outputs = self.module.inference(**inference_inputs)\n latent_basal += [outputs[\"latent_basal\"].cpu().numpy()]\n latent += [outputs['latent'].cpu().numpy()]\n\n latent_basal_adata = AnnData(X=np.concatenate(latent_basal, axis=0), obs=adata.obs.copy())\n latent_basal_adata.obs_names = adata.obs_names\n\n latent_adata = AnnData(X=np.concatenate(latent, axis=0), obs=adata.obs.copy())\n latent_adata.obs_names = adata.obs_names\n\n return latent_basal_adata, latent_adata\n\n @torch.no_grad()\n def predict(\n self,\n adata: Optional[AnnData] = None,\n indices: Optional[Sequence[int]] = None,\n batch_size: Optional[int] = 32,\n ):\n \"\"\"Counterfactual-friendly gene expression prediction\n # TODO: See if another signature makes more sense for better usability\n\n To produce counterfactuals,\n\n Returns\n -------\n Tuple\n Gene expression means and standard variations\n \"\"\"\n assert self.module.loss_ae in [\"gauss\", 'mse']\n self.module.eval()\n\n # adata = self.adata if adata is None else adata\n adata = self._validate_anndata(adata)\n if indices is None:\n indices = np.arange(adata.n_obs)\n scdl = self._make_data_loader(\n adata=adata, indices=indices, batch_size=batch_size, shuffle=False\n )\n mus = []\n stds = []\n for tensors in scdl:\n _mus, _stds = self.module.get_expression(tensors)\n mus.append(_mus.detach().cpu().numpy())\n stds.append(_stds.detach().cpu().numpy())\n\n pred_adata_mean = AnnData(X=np.concatenate(mus, axis=0), obs=adata.obs.copy())\n pred_adata_var = AnnData(X=np.concatenate(stds, axis=0), obs=adata.obs.copy())\n\n pred_adata_mean.obs_names = adata.obs_names\n pred_adata_var.obs_names = adata.obs_names\n\n pred_adata_mean.var_names = adata.var_names\n pred_adata_var.var_names = adata.var_names\n\n return pred_adata_mean, pred_adata_var\n\n @torch.no_grad()\n def get_drug_embeddings(self, doses=1.0, drug: Optional[str] = None):\n \"\"\"Computes all drug drug\n\n Parameters\n ----------\n doses : float, or torch.Tensor\n Drug dose, by default 1.0\n drug: str, optional\n Drug name if single drug embedding is desired\n\n \"\"\"\n self.module.eval()\n if isinstance(doses, float):\n if drug is None:\n treatments = doses * torch.eye(self.n_drugs, device=self.device)\n else:\n treatments = doses * F.one_hot(torch.LongTensor([self.drug_encoder[drug]]).to(self.device),\n self.n_drugs)\n elif isinstance(doses, np.ndarray):\n treatments = torch.tensor(doses).to(self.device).float()\n else:\n treatments = doses\n\n embeds = self.module.drug_network(treatments).detach().cpu().numpy()\n\n return embeds\n\n @torch.no_grad()\n def get_covar_embeddings(self, covariate: str, covariate_value: str = None):\n \"\"\"Computes Covariate drug\n\n Parameters\n ----------\n covariate : str\n covariate to be computed\n covariate_value: str, Optional\n Covariate specific value for embedding computation\n\n \"\"\"\n if covariate_value is None:\n covar_ids = torch.arange(len(self.cat_covars_encoders[covariate]), device=self.device)\n else:\n covar_ids = torch.LongTensor([self.cat_covars_encoders[covariate][covariate_value]]).to(self.device)\n embeddings = self.module.cat_covars_embeddings[covariate](covar_ids).detach().cpu().numpy()\n\n return embeddings\n\n def save(self, dir_path: str, overwrite: bool = False, save_anndata: bool = False, **anndata_write_kwargs):\n os.makedirs(dir_path, exist_ok=True)\n\n # save public dictionaries\n total_dict = {\n 'drug_encoder': self.drug_encoder,\n 'cat_covars_encoder': self.cat_covars_encoders,\n 'cont_covars': self.cont_covars,\n }\n\n json_dict = json.dumps(total_dict)\n with open(os.path.join(dir_path, 'CPA_dicts.json'), 'w') as f:\n f.write(json_dict)\n\n if isinstance(self.epoch_history, dict):\n self.epoch_history = pd.DataFrame().from_dict(self.training_plan.epoch_history)\n self.epoch_history.to_csv(os.path.join(dir_path, 'history.csv'), index=False)\n elif isinstance(self.epoch_history, pd.DataFrame):\n self.epoch_history.to_csv(os.path.join(dir_path, 'history.csv'), index=False)\n\n return super().save(dir_path=dir_path, overwrite=overwrite, save_anndata=save_anndata, **anndata_write_kwargs)\n\n @classmethod\n def load(cls, dir_path: str, adata: Optional[AnnData] = None, use_gpu: Optional[Union[str, int, bool]] = None,\n perturbation_keys: Optional[Dict[str, str]] = None,\n deg_uns_key: Optional[str] = None, ):\n assert (adata and perturbation_keys) or (adata is None)\n\n # load public dictionaries\n with open(os.path.join(dir_path, 'CPA_dicts.json')) as f:\n total_dict = json.load(f)\n\n cls.drug_encoder = total_dict['drug_encoder']\n cls.cat_covars_encoder = total_dict['cat_covars_encoder']\n cls.cont_covars = total_dict['cont_covars']\n\n model = super().load(dir_path, adata, use_gpu)\n\n drug_key = perturbation_keys['perturbation']\n dosage_key = perturbation_keys['dosage']\n\n if adata is not None and 'drugs_doses' not in adata.obsm:\n drugs_obsm = np.zeros((adata.n_obs, len(CPA.drug_encoder)))\n drugs, dosages = adata.obs[drug_key], adata.obs[dosage_key].astype(str)\n for i in tqdm(range(adata.n_obs)):\n cell_drugs = np.isin(list(CPA.drug_encoder.keys()), drugs[i].split('+'))\n cell_doses = np.array(dosages[i].split(\"+\")).astype(np.float32)\n drugs_obsm[i, cell_drugs] = cell_doses\n\n adata.obsm['drugs_doses'] = drugs_obsm\n\n if deg_uns_key:\n mask = np.zeros((adata.n_obs, adata.n_vars))\n for i, cov_drug in tqdm(enumerate(adata.obs['cov_drug'].values)):\n mask[i] = adata.var.index.isin(adata.uns[deg_uns_key][cov_drug]).astype(np.int)\n\n adata.obsm['deg_mask'] = np.array(mask)\n\n try:\n model.epoch_history = pd.read_csv(os.path.join(dir_path, 'history.csv'))\n except:\n print('WARNING: The history was not found.')\n\n return model\n","repo_name":"cloudengio/cpa","sub_path":"cpa/_model.py","file_name":"_model.py","file_ext":"py","file_size_in_byte":19899,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"38233448244","text":"from collections import OrderedDict\nimport os\n\nimport yaml\n\n\nclass CmdlineParser(object):\n def __init__(self):\n self.template = []\n self.configuration = OrderedDict()\n self._preparsed = False\n\n def format(self, configuration):\n curated_config = {}\n for key, item in configuration.items():\n if isinstance(item, dict):\n item = item['file']\n\n curated_config[key] = item\n\n return \" \".join(self.template).format(**curated_config)\n\n def parse(self, commandline):\n if not commandline:\n return {}\n\n self.configuration = OrderedDict(self.parse_arguments(commandline))\n for key, value in self.configuration.items(): \n # TODO: Support passing the same commandline but slightly different\n if key.startswith(\"_\") and self._preparsed:\n raise RuntimeError(\"Cannot branch using positional arguments.\")\n # Positional\n elif key.startswith(\"_\"):\n self.template.append(\"{\" + key + \"}\")\n \n # Optional\n else:\n template = self.key_to_arg(key)\n if template in self.template:\n continue\n\n self.template.append(self.key_to_arg(key))\n\n # Ignore value as key is a boolean argument\n if isinstance(value, bool):\n continue\n\n if not isinstance(value, list):\n template = \"{\" + key + \"}\"\n self.template.append(template)\n continue\n\n for pos, item in enumerate(value):\n template = \"{\" + key + \"[\" + str(pos) + \"]}\"\n self.template.append(template)\n\n self._preparsed = True\n\n self.fetch_configurations()\n\n return self.configuration\n\n def arg_to_key(self, arg):\n arg = arg.split(\"=\")[0]\n\n if arg.startswith(\"--\") and len(arg) == 3:\n raise ValueError(\n \"Arguments with two dashes should have more than one letter: {}\".format(arg))\n\n elif not arg.startswith(\"--\") and arg.startswith(\"-\") and len(arg) > 2:\n raise ValueError(\n \"Arguments with one dashes should have only one letter: {}\".format(arg))\n\n return arg.lstrip(\"-\").replace(\"_\", \"__\").replace(\"-\", \"_\")\n\n def key_to_arg(self, key):\n arg = key.replace(\"__\", \"!!!!\").replace(\"_\", \"-\").replace(\"!!!!\", \"_\")\n if len(arg) > 1:\n return \"--\" + arg\n \n return \"-\" + arg\n\n def parse_paths(self, value):\n if isinstance(value, list):\n return [self.parse_paths(item) for item in value]\n\n if isinstance(value, str) and os.path.exists(value):\n return os.path.abspath(value)\n\n return value\n\n def parse_arguments(self, arguments):\n positional_index = 0\n argument_name = None\n pairs = []\n argument_names = set()\n for arg in arguments:\n # Key\n if arg.startswith(\"-\"):\n arg = arg.split(\"=\")\n argument_name = self.arg_to_key(arg[0])\n if argument_name in argument_names:\n raise ValueError(\"Two arguments have the same name: {}\".format(argument_name))\n\n argument_names.add(argument_name)\n pairs.append([argument_name, []])\n if len(arg) > 1 and \"=\".join(arg[1:]).strip(\" \"):\n pairs[-1][1].append(\"=\".join(arg[1:]))\n # Optional\n elif argument_name is not None and arg.strip(\" \"):\n pairs[-1][1].append(arg)\n # Positional\n elif argument_name is None:\n pairs.append([\"_pos_{}\".format(len(pairs)), arg])\n\n for i, [key, value] in enumerate(pairs):\n if not value:\n value = True\n elif isinstance(value, list) and len(value) == 1:\n value = value[0]\n\n value = self.parse_paths(value)\n\n pairs[i][1] = value\n\n return pairs\n\n def fetch_configurations(self, configuration=None):\n if configuration is None:\n configuration = self.configuration\n\n for key, item in self.configuration.items():\n if isinstance(item, str) and os.path.exists(item):\n file_config = self.load_conf_file(item)\n if file_config is None:\n continue \n\n self.configuration[key] = {'file': item, 'content': file_config}\n\n def load_conf_file(self, name):\n if name.endswith('.yaml'):\n with open(name, 'r') as f:\n return yaml.load(f)\n\n return None\n","repo_name":"Epistimio/kleio","sub_path":"src/kleio/core/io/cmdline_parser.py","file_name":"cmdline_parser.py","file_ext":"py","file_size_in_byte":4744,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"2699729084","text":"# django-djangoplus setup\n# First version of this file done by Guilherme Semente\n# Some things was copied from Django's setup.py\nfrom distutils.command.install import INSTALL_SCHEMES\nimport os, sys\n\n# Downloads setuptools if not find it before try to import\ntry:\n import ez_setup\n ez_setup.use_setuptools()\nexcept ImportError:\n pass\n\nfrom setuptools import setup\nfrom djangoplus import get_version\n\n# Tell distutils to put the data_files in platform-specific installation\n# locations. See here for an explanation:\n# http://groups.google.com/group/comp.lang.python/browse_thread/thread/35ec7b2fed36eaec/2105ee4d9e8042cb\nfor scheme in INSTALL_SCHEMES.values():\n scheme['data'] = scheme['purelib']\n\ndata_files = []\n\nfor dirpath, dirnames, filenames in os.walk('djangoplus'):\n # Ignore dirnames that start with '.'\n for i, dirname in enumerate(dirnames):\n if dirname.startswith('.'): del dirnames[i]\n if filenames:\n data_files.append([dirpath, [os.path.join(dirpath, f) for f in filenames]])\n\n# Small hack for working with bdist_wininst.\n# See http://mail.python.org/pipermail/distutils-sig/2004-August/004134.html\nif len(sys.argv) > 1 and sys.argv[1] == 'bdist_wininst':\n for file_info in data_files:\n file_info[0] = '\\\\PURELIB\\\\%s' % file_info[0]\n\nsetup(\n name = 'django-plus',\n version = get_version(),\n description = 'Django utilities library',\n long_description = 'django-plus is a library containing a coupple of utilities for Django developers.',\n author = 'Marinho Brandao',\n author_email = 'marinho@gmail.com',\n url = 'http://django-plus.googlecode.com',\n license = 'GNU Lesser General Public License (LGPL)',\n packages = [\n 'djangoplus',\n 'djangoplus.fieldtypes',\n 'djangoplus.forms',\n 'djangoplus.management',\n 'djangoplus.management.commands',\n 'djangoplus.middleware',\n 'djangoplus.shortcuts',\n 'djangoplus.templatetags',\n 'djangoplus.tests',\n 'djangoplus.urls',\n 'djangoplus.utils',\n 'djangoplus.views',\n 'djangoplus.widgets',\n ],\n data_files = data_files,\n include_package_data=True,\n)\n\n","repo_name":"marinho/django-plus","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":2174,"program_lang":"python","lang":"en","doc_type":"code","stars":30,"dataset":"github-code","pt":"78"}
+{"seq_id":"9788140015","text":"from flask import url_for, session\n\nfrom automated_survey_flask import app\nfrom automated_survey_flask.models import Question\n\nfrom .base import BaseTest\n\n\nclass QuestionsTest(BaseTest):\n def get_question_as_xml(self, question, client=None, data=None):\n client = client or self.client\n response = client.get(url_for('question', question_id=question.id), data=data)\n return self.assertXmlDocument(response.data)\n\n def test_first_question_during_a_call(self):\n first_question = self.questions[0]\n root = self.get_question_as_xml(first_question)\n\n self.assertIn(first_question.content, root.xpath('./Say/text()'))\n\n def test_first_question_over_sms(self):\n first_question = self.questions[0]\n data = {'MessageSid': 'unique'}\n root = self.get_question_as_xml(first_question, data=data)\n\n self.assertIn(first_question.content, root.xpath('./Message/text()'))\n\n def test_current_question_being_answered_goes_to_session(self):\n first_question = self.questions[0]\n with app.test_client() as client:\n self.get_question_as_xml(first_question, client=client)\n self.assertEquals(first_question.id, session['question_id'])\n\n def test_gather_keys_on_numeric_question_during_a_call(self):\n numeric_question = self.question_by_kind[Question.NUMERIC]\n root = self.get_question_as_xml(numeric_question)\n\n answer_url = url_for('answer', question_id=numeric_question.id)\n self.assertEquals([answer_url], root.xpath('./Gather/@action'))\n\n def test_record_on_text_questions_during_a_call(self):\n text_question = self.question_by_kind[Question.TEXT]\n root = self.get_question_as_xml(text_question)\n\n answer_url = url_for('answer', question_id=text_question.id)\n self.assertEquals([answer_url], root.xpath('./Record/@action'))\n\n def test_transcription_is_enabled_for_text_questions_during_a_call(self):\n text_question = self.question_by_kind[Question.TEXT]\n root = self.get_question_as_xml(text_question)\n\n answer_transcription_url = url_for(\n 'answer_transcription', question_id=text_question.id\n )\n self.assertEquals(\n [answer_transcription_url], root.xpath('./Record/@transcribeCallback')\n )\n\n def test_gather_keys_on_boolean_question_during_a_call(self):\n boolean_question = self.question_by_kind[Question.BOOLEAN]\n root = self.get_question_as_xml(boolean_question)\n\n answer_url = url_for('answer', question_id=boolean_question.id)\n self.assertEquals([answer_url], root.xpath('./Gather/@action'))\n","repo_name":"TwilioDevEd/automated-survey-flask","sub_path":"tests/question_view_tests.py","file_name":"question_view_tests.py","file_ext":"py","file_size_in_byte":2655,"program_lang":"python","lang":"en","doc_type":"code","stars":15,"dataset":"github-code","pt":"78"}
+{"seq_id":"19380916674","text":"from .constants import *\nfrom .board import Board\n\n\nclass GameState:\n def __init__(self, turtle):\n self.selected = None\n self.board = Board()\n self.turn = 'red'\n self.valid_moves = {}\n self.turtle = turtle\n\n def update(self):\n self.board.draw(self.turtle)\n\n def reset(self):\n self.selected = None\n self.board = Board()\n self.turn = 'red'\n self.valid_moves = {}\n\n def select(self, row, col):\n if(self.selected): #if not none\n result = self._move(row, col)\n if not result: # not valid select,select again\n self.selected = None\n self.select(row, col)\n else: \n piece = self.board.get_piece(row, col)\n if piece != 0 and piece.color == self.turn:\n self.selected = piece\n self.valid_moves = self.board.get_valid_moves(piece)\n return True\n return False\n\n def _move(self, row, col):\n piece = self.board.get_piece(row, col)\n if self.selected and piece == 0 and (row, col) in self.valid_moves:\n self.board.move(self.selected, row, col)\n self.change_turn()\n else:\n return False\n return True\n\n def change_turn(self):\n if self.turn == 'red':\n self.turn = 'black'\n else:\n self.turn = 'red'\n","repo_name":"tianyuanzoe/CheckerGame-with-turtle","sub_path":"final project/checkers/game_state.py","file_name":"game_state.py","file_ext":"py","file_size_in_byte":1409,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"1314053602","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\ndef get_read_base(aln, qseq, qual, pos):\n for qpos, rpos, rbase in aln:\n if rpos != pos:\n continue\n return [rbase, qseq[qpos], qual[qpos]]\n return ['', '', 0]\n\ndef infer_variant(x, minBaseQual):\n aln = x.get_aligned_pairs(matches_only = False, with_seq = True)\n qseq = x.query_sequence\n qual = x.query_qualities\n mms = dict()\n for qpos, rpos, rbase in aln:\n if qpos is not None and rpos is not None and rbase.islower() and qual[qpos] >= minBaseQual:\n mms[qpos] = [rpos+1, 'M', qseq[qpos]]\n #if x.query_name == 'HISEQ15:141:C6VDGANXX:5:1115:10555:79840':\n # print(mms)\n res = []\n qpos, rpos = 0, x.reference_start\n rbeg = rpos\n vnts = []\n for opt, nbase in x.cigartuples:\n if opt == 0: #M\n for i in range(0, nbase):\n if qpos+i in mms:\n vnts.append(mms[qpos+i])\n qpos += nbase\n rpos += nbase\n elif opt == 1: #I\n vnts.append([rpos+1, \"I\", qseq[qpos+1:qpos+nbase+1]])\n qpos += nbase\n elif opt == 2: #D\n vnts.append([rpos+1, \"D\", nbase])\n rpos += nbase\n elif opt == 3: #N\n res.append([rbeg, rpos, vnts])\n rpos += nbase\n rbeg = rpos\n vnts = []\n elif opt == 4: #S\n qpos += nbase\n elif opt == 7 or opt == 8: #=X\n qpos += 1\n rpos += 1\n if rpos > rbeg:\n res.append([rbeg, rpos, vnts])\n if qpos != x.query_length or rpos != x.reference_end:\n print(qpos, x.query_alignment_end, rpos, x.reference_end, x.cigartuples, aln)\n exit(1)\n return res\n\ndef read_variants(fv):\n fhv = open(fv, \"r\")\n vdic = dict()\n for line in fhv:\n line = line.strip(\"\\n\")\n seqid, beg, end, gts = line.split(\"\\t\")\n beg, end = int(beg), int(end)\n ref, alt = gts.split(\",\")\n pos = \"%s_%d\" % (seqid, int(beg) + 1)\n vdic[pos] = [ref, alt]\n fhv.close()\n return vdic\n\ndef bam2bed(args):\n fi, fo = args.bam, args.bed\n min_mapq, min_baseq = args.min_mapq, args.min_baseq\n bam = pysam.AlignmentFile(fi, \"rb\")\n fho = open(fo, \"w\")\n for x in bam.fetch():\n if x.is_duplicate or x.is_unmapped or x.is_secondary or x.mapping_quality < min_mapq:\n continue\n sid = x.query_name\n rid, rbeg, rend = x.reference_name, x.reference_start, x.reference_end\n #pair = 'pe' if x.is_paired else 'se'\n #first = 'r1' if x.is_read1 else 'r2'\n res = infer_variant(x, min_baseq)\n #if sid == 'D00635:197:CAC47ANXX:2:2311:18289:97712':\n # print(res)\n # exit(1)\n for rbeg, rend, vnts in res:\n vntstr = '.'\n if len(vnts) > 0:\n vntstr = \" \".join([\":\".join(map(str,vnt)) for vnt in vnts])\n fho.write(\"%s\\t%d\\t%d\\t%s\\t%s\\n\" %\n (rid, rbeg, rend, sid, vntstr))\n #exit(1)\n fho.close()\n\ndef infer_read(rows, fho, fhb):\n rdic, vdic = dict(), dict()\n sid, beg, end, rid = rows[0][0:4]\n locs = set()\n for row in rows:\n assert rid == row[3], \"error: \" + str(rows)\n if row[4] != '.':\n for rvnt in row[4].split(\" \"):\n pos, vtype, vnt = rvnt.split(\":\")\n rdic[pos] = \"%s:%s\" % (vtype, vnt)\n vpos, vnt, phase = row[7:10]\n assert phase in ['0|1','1|0'], \"Unknown phase: %s\" % phase\n vdic[vpos] = [vnt, phase]\n loc = \"%s:%s\" % tuple(row[0:2])\n if loc not in locs:\n locs.add(loc)\n fhb.write(\"%s\\t%s\\t%s\\t%s\\n\" % tuple(row[0:4]))\n n0, n1, nunk, nerr = 0, 0, 0, 0\n for pos in vdic:\n if pos in rdic:\n if rdic[pos] == vdic[pos][0]:\n if vdic[pos][1] == '0|1':\n n1 = n1 + 1\n else:\n n0 = n0 + 1\n else:\n nunk = nunk + 1\n del rdic[pos]\n else:\n if vdic[pos][1] == '0|1':\n n0 = n0 + 1\n else:\n n1 = n1 + 1\n nerr = len(rdic)\n fho.write(\"%s\\t%d\\t%d\\t%d\\t%d\\n\" % (rid, n0, n1, nunk, nerr))\n\ndef bed_prep(args):\n fi, fo, fb = args.bed, args.out_tsv, args.out_bed\n min_baseq = args.min_baseq\n fhi = open(fi, \"r\")\n fho = open(fo, \"w\")\n fhb = open(fb, \"w\")\n fho.write(\"rid\\tn0\\tn1\\tnunk\\tnerr\\n\")\n prid = \"\"\n rows = []\n for line in fhi:\n row = line.strip(\"\\n\").split(\"\\t\")\n if prid == \"\":\n rows.append(row)\n prid = row[3]\n elif row[3] == prid:\n rows.append(row)\n else:\n infer_read(rows, fho, fhb)\n rows = [row]\n prid = row[3]\n infer_read(rows, fho, fhb)\n fhi.close()\n fho.close()\n fhb.close()\n\ndef bed_summarise(args):\n fr, fb, fo = args.tsv, args.bed, args.out\n\n fhr = open(fr, \"r\")\n rdic = dict()\n for line in fhr:\n rid, n0, n1, nunk, nerr = line.strip(\"\\n\").split(\"\\t\")\n if rid == 'rid':\n continue\n tag = \"unk\"\n if int(n0) > 0 and int(n1) == 0:\n tag = 'h0'\n elif int(n0) == 0 and int(n1) > 0:\n tag = 'h1'\n elif int(n0) > 0 and int(n1) > 0:\n tag = 'cft'\n rdic[rid] = tag\n fhr.close()\n\n fhb = open(fb, \"r\")\n gdic, tdic = dict(), dict()\n for line in fhb:\n row = line.strip(\"\\n\").split(\"\\t\")\n gid, rid = row[3], row[7]\n if gid not in gdic:\n gdic[gid] = {'n0': 0, 'n1': 0, 'ncft': 0}\n if gid not in tdic:\n tdic[gid] = set()\n if rid not in tdic[gid]:\n tdic[gid].add(rid)\n else:\n continue\n if rid not in rdic:\n print(\"%s not in read dict\" % rid)\n if rdic[rid] == 'h0':\n gdic[gid]['n0'] += 1\n elif rdic[rid] == 'h1':\n gdic[gid]['n1'] += 1\n elif rdic[rid] == 'cft':\n gdic[gid]['ncft'] += 1\n fhb.close()\n \n fho = open(fo, \"w\")\n fho.write(\"gid\\tn0\\tn1\\tncft\\n\")\n for gid in sorted(gdic):\n sdic = gdic[gid]\n n0, n1, ncft = sdic['n0'], sdic['n1'], sdic['ncft']\n if n0 + n1 < 0:\n continue\n fho.write(\"%s\\t%d\\t%d\\t%d\\n\" % (gid, n0, n1, ncft))\n fho.close()\n\nif __name__ == \"__main__\":\n import argparse\n parser = argparse.ArgumentParser(\n formatter_class = argparse.ArgumentDefaultsHelpFormatter,\n description = 'allele specific expression utilities'\n )\n sp = parser.add_subparsers(title = 'available commands', dest = 'command')\n\n sp1 = sp.add_parser(\"bam2bed\",\n formatter_class = argparse.ArgumentDefaultsHelpFormatter,\n help = 'convert bam to bed format using variant coordinates'\n )\n sp1.add_argument('bam', help='input SAM/BAM file')\n sp1.add_argument('bed', help = 'output BED file')\n sp1.add_argument('--min_mapq', default=20, help='min mapping quality')\n sp1.add_argument('--min_baseq', default=20, help='min base quality')\n sp1.add_argument('--vcf', default='/home/springer/zhoux379/data/misc2/mo17vnt/53.vnt.final/61.rna.bed', help='variant file')\n sp1.set_defaults(func = bam2bed)\n\n sp1 = sp.add_parser(\"bed_prep\",\n formatter_class = argparse.ArgumentDefaultsHelpFormatter,\n help = 'infer allele-specific read origin'\n )\n sp1.add_argument('bed_prep', help='input BED (ase) file')\n sp1.add_argument('out_tsv', help = 'output read ASE file (tsv)')\n sp1.add_argument('out_bed', help = 'output read location file (bed)')\n sp1.add_argument('--min_baseq', default=20, help='min base quality')\n sp1.set_defaults(func = bed_prep)\n\n sp1 = sp.add_parser(\"bed_summarise\",\n formatter_class = argparse.ArgumentDefaultsHelpFormatter,\n help = 'summarise allele-specific read counts per gene'\n )\n sp1.add_argument('tsv', help='input tsv (ase) file')\n sp1.add_argument('bed', help = 'input gene-read intersection BED')\n sp1.add_argument('out', help = 'output gene ASE file (tsv)')\n sp1.set_defaults(func = bed_summarise)\n\n args = parser.parse_args()\n if args.command:\n args.func(args)\n else:\n print('Error: need to specify a sub command\\n')\n parser.print_help()\n\n","repo_name":"orionzhou/maize","sub_path":"apps/ase.py","file_name":"ase.py","file_ext":"py","file_size_in_byte":8335,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"78"}
+{"seq_id":"23324152701","text":"import cv2\nimport numpy as np\n\n# Imagens carregadas\nmapa_estatico = cv2.imread(\"mapa_estatico.png\")\nmapa_dinamico = cv2.imread(\"mapa_dinamico2.png\")\n\n# Redimensionamento das imagens\nmapa_dinamico = cv2.resize(mapa_dinamico, (mapa_estatico.shape[1], mapa_estatico.shape[0]))\n\n# Verificação das dimensões das imagens\nassert mapa_estatico.shape == mapa_dinamico.shape, \"Os mapas têm tamanhos diferentes!\"\n\n# Calculo do Mean Squared Error (MSE)\nmse = np.mean((mapa_estatico - mapa_dinamico) ** 2)\n\n# Valor do MSE\nprint(f\"Mean Squared Error (MSE): {mse}\")\n","repo_name":"caiquevianadev/MSE","sub_path":"mse_script.py","file_name":"mse_script.py","file_ext":"py","file_size_in_byte":555,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39342271490","text":"def greet_user(username):\n \"\"\" greeting user\"\"\" #This is called a docstring\n print(f\"Hello, {username.title()} !\")\n\ngreet_user('roy')\n\ndef recent_book(bookname):\n print(f\"I recently read the book named \\\"{bookname.title()}\\\"\")\n\nrecent_book('the richest man in babylon')\nprint()\n\ndef make_shirt(text, size=38):\n print(f\"\\nYou have ordered a t-shirt of size: {size}\")\n print(f\"The text on your t will be: {text}\")\n\nmake_shirt(text = 'Step UP!', size = 40) #keyword argument\nmake_shirt('Shut UP!',42) #positional argument\nmake_shirt('whodunnit ?') #default argument\nprint()\n\ndef format_name(firstName, lastName, middleName = ''):\n if middleName:\n full = f\"{firstName} {middleName} {lastName}\"\n else:\n full = f\"{firstName} {lastName}\"\n return full.title()\nperson = format_name('anindya', 'roy', 'piyal')\nprint(person)\n\n\"\"\"\nwhile True:\n print(\"\\nPlease tell me your name:\")\n print(\"(enter 'q' at any time to quit)\")\n\n f_name = input(\"First name: \")\n if f_name == 'q':\n break\n\n l_name = input(\"Last name: \")\n if l_name == 'q':\n break\n\n formatted_name = get_formatted_name(f_name, l_name)\n print(f\"\\nHello, {formatted_name}!\")\n\"\"\"\nprint()\ndef city_country(country,city):\n str = f\"{city},{country}\"\n return str\n\na = city_country(city = 'Santiago', country = 'Chile')\nb = city_country(city = 'Brooklyn', country = 'USA')\nprint(a)\nprint(b)\nprint()\n\ndef make_album(artist_name, album_title, song_count=None):\n album = {'Artist Name': artist_name.title(),'Album': album_title.title()}\n if song_count:\n album['NO. of Songs'] = song_count\n return album\n\na = make_album('lisa henigan', 'passenger' )\nb = make_album(album_title = 'after hours' , artist_name = 'the weeknd', song_count = 13)\nprint(a)\nprint(b)\nprint()\n\n\"\"\"\nwhile True:\n name = input(\"\\nWhat is the artist's name ?(press 'q' to close program): \")\n if name == 'q':\n break\n title = input(\"What is the album title ?(press 'q' to close program): \")\n if title == 'q':\n break\n al = make_album(name,title)\n print(\"Album info: \",al)\n\"\"\"\ndef show_message(messages):\n for message in messages:\n print(message)\n\ndef send_message(messages):\n while messages:\n copy = messages.pop()\n print(\"from send_message:::\",copy)\n sent_messages.append(copy)\n\n\nMessages = ['hello','ssup','no','busy now']\nshow_message(Messages)\nsent_messages = []\nprint()\nprint(\"messages:\",Messages)\nprint(\"sent :\",sent_messages)\nprint()\nsend_message(Messages)\nprint(\"\\nmessages:\",Messages)\nprint(\"sent :\",sent_messages)\n\ndef make_pizza(*toppings):\n \"\"\"Summarize the pizza we are about to make.\"\"\"\n print(\"\\nMaking a pizza with the following toppings:\")\n for topping in toppings:\n print(f\"- {topping}\")\n\nmake_pizza('pepperoni')\nmake_pizza('mushrooms', 'green peppers', 'extra cheese')\n\ndef make_pizza(size, *toppings):\n \"\"\"Summarize the pizza we are about to make.\"\"\"\n print(f\"\\nMaking a {size}-inch pizza with the following toppings:\")\n for topping in toppings:\n print(f\"- {topping}\")\n\nmake_pizza(16, 'pepperoni')\nmake_pizza(12, 'mushrooms', 'green peppers', 'extra cheese')\n\ndef prep_sandwich(*items): # *means tupple\n print(\"\\nYou have the following items in your sandwich: \")\n for item in items:\n print(f\"-{item}\")\n\nprep_sandwich('cheese','sause','latuce') #practice 8-12\nprint()\n\n\ndef build_profile(first, last, **other_info): # **means empty dictionary\n other_info['First Name'] = first.title()\n other_info['Last Name'] = last.title()\n return other_info\n\nuser1 = build_profile('anindya','roy',University = 'BRACU', ID = 19101577, Batch = 'Spring 2019')\nprint('We have the following info about this user:')\nfor k,v in user1.items():\n print(f\"{k}: {v}\")\n\n#Importing module in python program. Import functions from other files. must end with .py ; an example below.\n\n\"\"\" import pizza\npizza.make_pizza(16, 'pepperoni')\npizza.make_pizza(12, 'mushrooms', 'green peppers', 'extra cheese')\n\nfrom pizza import make_pizza\nmake_pizza(16, 'pepperoni')\nmake_pizza(12, 'mushrooms', 'green peppers', 'extra cheese')\n\nfrom pizza import make_pizza as mp (as = alias)\nmp(16, 'pepperoni')\nmp(12, 'mushrooms', 'green peppers', 'extra cheese')\n\nimport pizza as p\np.make_pizza(16, 'pepperoni')\np.make_pizza(12, 'mushrooms', 'green peppers', 'extra cheese') \"\"\"\n","repo_name":"anindyaroypiyal/Python-Crash-Course-2nd-Edition-A-Hands-On-Project-Based-Introduction-to-Programming-by-Eric-Matthes","sub_path":"ch 8 Function.py","file_name":"ch 8 Function.py","file_ext":"py","file_size_in_byte":4358,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"3142486944","text":"# 숫�� 블록\nimport math\n\nINF = 10000000\n\n\ndef solution(begin, end):\n ret = []\n for number in range(begin, end + 1):\n if number == 1:\n ret.append(0)\n continue\n s = 1\n for i in range(2, int(math.sqrt(number)) + 1):\n if number % i == 0:\n s = number // i\n if s <= INF:\n break\n ret.append(s if s <= INF else 1)\n\n return ret\n\n\nif __name__ == '__main__':\n begin = 1\n end = 10\n print(solution(begin, end))\n","repo_name":"mrbartrns/algorithm-and-structure","sub_path":"programmers/lv4_review/p20.py","file_name":"p20.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"10821395983","text":"import pandas as pd\nimport numpy as np\nimport random\ndata = pd.read_csv(\"./datasets/e-commerce/data.csv\", encoding=\"utf-8\", index_col=[0])\n\nrandom.seed(6)\n\ncolumns = random.sample(population=list(data.columns), k=4)\ncount = data.shape[0]\ncountMissing = int(count / 10)\n\nfor index in columns:\n data[index].iloc[random.sample(range(count), countMissing)] = np.nan\n\nprint(columns)\nprint(data)\n\n\ndata.to_csv(\"./datasets/e-commerce/data_missing.csv\", index=False, encoding='utf-8')","repo_name":"NattapongNetnu/GenerateMissingValue","sub_path":"GenerateMissing.py","file_name":"GenerateMissing.py","file_ext":"py","file_size_in_byte":479,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"29127093394","text":"#!/usr/bin/env python\n# encoding: utf-8\n'''\n@author: zhoujun\n@time: 2019/12/19 下午3:23\n'''\nimport cv2\nimport numpy as np\nfrom mxnet import image, nd\n\n\nclass Resize:\n def __init__(self, img_h, img_w, pad=True, **kwargs):\n self.img_h = img_h\n self.img_w = img_w\n self.pad = pad\n\n def __call__(self, img: np.ndarray):\n \"\"\"\n 对图片进行处理,先按照高度进行resize,resize之后如果宽度不足指定宽度,就补黑色像素,否则就强行缩放到指定宽度\n :param img_path: 图片地址\n :return:\n \"\"\"\n assert len(img.shape) == 3 and img.shape[-1] in [1, 3]\n h, w = img.shape[:2]\n ratio_h = self.img_h / h\n new_w = int(w * ratio_h)\n if new_w < self.img_w and self.pad:\n img = cv2.resize(img, (new_w, self.img_h))\n step = np.zeros((self.img_h, self.img_w - new_w, img.shape[-1]), dtype=img.dtype)\n img = np.column_stack((img, step))\n else:\n img = cv2.resize(img, (self.img_w, self.img_h))\n return img\n\n\nclass ResizeRandomCrop:\n def __init__(self, img_h, img_w, pad=True, **kwargs):\n self.img_h = img_h\n self.img_w = img_w\n self.pad = pad\n self.phase = kwargs['phase']\n\n def __call__(self, img: np.ndarray):\n \"\"\"\n 对图片进行处理,先按照高度进行resize,resize之后如果宽度不足指定宽度,就补黑色像素,否则就强行缩放到指定宽度\n :param img_path: 图片地址\n :return:\n \"\"\"\n data_augment = False\n if self.phase == 'train' and np.random.rand() > 0.5:\n data_augment = True\n if data_augment:\n img_h = 40\n img_w = 340\n else:\n img_h = self.img_h\n img_w = self.img_w\n h, w = img.shape[:2]\n ratio_h = float(img_h) / h\n new_w = int(w * ratio_h)\n if new_w < img_w and self.pad:\n img = cv2.resize(img, (new_w, img_h))\n if len(img.shape) == 2:\n img = np.expand_dims(img, 3)\n step = np.zeros((img_h, img_w - new_w, img.shape[-1]), dtype=img.dtype)\n img = np.column_stack((img, step))\n else:\n img = cv2.resize(img, (img_w, img_h))\n if data_augment:\n img = nd.array(img)\n img, _ = image.random_crop(img, (self.img_w, self.img_h))\n img = img.asnumpy()\n return img\n","repo_name":"WenmuZhou/crnn.gluon","sub_path":"data_loader/modules/resize.py","file_name":"resize.py","file_ext":"py","file_size_in_byte":2474,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"78"}
+{"seq_id":"29619812559","text":"from movie.models import movie as Movie\nfrom movie.models import review as Review\nfrom movie.models import person as Person\nfrom movie.models import genre as Genre\nfrom movie import db\nfrom sqlalchemy import or_\n\n# put all movies in dictionary with value 0\ndef initialise_movies():\n all_movies = {}\n movies = db.session.query(Movie.Movies).all()\n for movie in movies:\n all_movies[movie.id] = 0\n return all_movies\n\n# get top 20 movies in dictionary based on value from above\ndef top_twenty(movies):\n sortlist = sorted(movies.items(), key=lambda x:x[1], reverse=True)\n top = {}\n for i in range(0, 19):\n item = sortlist[i]\n top[item[0]] = item[1]\n top_movies = []\n for movie in top.keys():\n top_movies.append(db.session.query(Movie.Movies).filter(Movie.Movies.id == movie).first())\n return top_movies\n\n\"\"\"\n# change movie values based on user reviews\n# rating 1 = -2, 2 = -1, 3 = 0, 4 = +1, 5 = +2\ndef calculate_genre( genre_ids):\n movie_genres = db.session.query(Movie.Movies, Movie.MovieGenre).filter(Movie.MovieGenre.movie_id == Movie.Movies.id, Movie.MovieGenre.genre_id.in_(genre_ids)).all()\n return movie_genres\n\n\"\"\"\n\n\"\"\"\ndef calculate_director(directors):\n movie_dir = db.session.query(Person.MovieDirector, Movie.Movies).filter(Person.MovieDirector.movie_id == Movie.Movies.id, Person.MovieDirector.person_id.in_(directors)).all()\n return movie_dir\n\"\"\"\n\n\n# get the movies which has the same given geenre or give director\n# by: the key word that genre or director or both we need to \n# sort by the rating, return the top 20\ndef get_genre_director_movie(genre_ids, directors, by):\n query = db.session.query(Movie.Movies).filter(Movie.MovieGenre.movie_id == Movie.Movies.id, Person.MovieDirector.movie_id == Movie.Movies.id)\n\n if by and by == 'genre':\n query = query.filter(Movie.MovieGenre.genre_id.in_(genre_ids))\n if by and by == 'director':\n query = query.filter(Person.MovieDirector.person_id.in_(directors))\n else:\n query = query.filter(or_(Movie.MovieGenre.genre_id.in_(genre_ids), Person.MovieDirector.person_id.in_(directors)))\n\n # sort\n movies = query.order_by(Movie.Movies.total_rating.desc()).limit(20).all()\n return movies\n","repo_name":"unsw-cse-comp3900-9900-22T2/capstone-project-3900-f16a-200forever","sub_path":"backend/movie/utils/recommendation_util.py","file_name":"recommendation_util.py","file_ext":"py","file_size_in_byte":2180,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"27220494029","text":"# 버블소팅은 O(log(n*n))으로 느린 알고리즘이라 사용하면 안됨\n# 다만 버블소팅은 스테이블 소팅임\n# 오름차순 정렬\n\n# 버블소팅은 첫번째와 두번째 비교해서 큰 수를 오른쪽으로 스왑\n# 두번째 세번째 비교해서 큰 수를 오른쪽으로 스왑\n# 이렇게 하면 배열 내의 가장 큰 수가 맨 끝에 가게됨\n# 그러면 이 젤 큰수는 더 이상 정렬할 필요가 없음\n\n# 이름이 버블인 이유는 nums[i] 와 nums[i+1] 비교할때 이 둘이 버블처럼 묶어서 비교한다는 의미임\n\nfrom typing import List\n\ndef bubbleSort(nums: List) -> List:\n for idx in range(len(nums)-1): # idx를 빼주는 이유는, 바깥 포문이 한 턴을 돌면 가장 큰 수가 배열의 마지막 위치에 가게됨 그래서 그 값은 비교할 필요가 없어서 그럼\n for i in range(len(nums)-1-idx): # 그래서 다음 턴의 가장 큰수는 바로전 턴의 최대값이 있는 인덱스 위치의 바로 왼쪽에 위치함\n if nums[i] > nums[i+1]:\n nums[i], nums[i+1] = nums[i+1], nums[i]\n\n return nums\n\nprint(bubbleSort(nums=[9,3,5,7,1]))\n\n# 스테이블 소팅\n\ndef stableBubbleSort(numsString):\n for idx in range(len(numsString)-1):\n for i in range(len(numsString)-1-idx):\n if numsString[i][0] > numsString[i+1][0]:\n numsString[i][0], numsString[i+1][0] = numsString[i+1][0], numsString[i][0]\n\n return numsString\n\n\nprint(stableBubbleSort(num_strs=[(7,'a'),(5,'a'),(5,'b'),(7,'b'),(3,'c')]))\n\n","repo_name":"badoil/algorithms","sub_path":"algo3.sorting/sort1.bubbleSorting.py","file_name":"sort1.bubbleSorting.py","file_ext":"py","file_size_in_byte":1546,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39823778867","text":"import os\r\nimport argparse\r\nimport subprocess\r\n\r\n\r\n\r\n# create parser and arguments\r\nparser = argparse.ArgumentParser()\r\nparser._action_groups.pop()\r\n\r\n# required arguments (positional)\r\nrequired = parser.add_argument_group(\"required arguments\")\r\nrequired.add_argument(\"path\", help=\"File path\", type=str, nargs=1)\r\nrequired.add_argument(\"valid\", help=\"Valid hash\", type=str, nargs=1)\r\n\r\n# optional arguments\r\noptional = parser.add_argument_group(\"optional arguments\")\r\noptional.add_argument(\"-a\", \"--algo\", help=\"Algorithm to be used\", dest=\"algorithm\", type=str, required=True, default=\"SHA256\")\r\noptional.add_argument(\"-s\", \"--show\", help=\"View comparison of hashes (true, t, y)\", dest=\"show_hashes\", type=str)\r\n\r\n# get arguments\r\nargs = parser.parse_args()\r\n\r\n\r\n\r\ndef show_hashes(valid, file):\r\n \"\"\"\r\n Prints checksums\r\n\r\n Args:\r\n valid (str): Valid checksum\r\n file (str): File checksum\r\n \"\"\"\r\n print(f\"Original:\\t{valid}\\nFilehash:\\t{file}\\n\")\r\n\r\n\r\n\r\ndef checksum(file_path, valid_hash, algorithm=\"SHA256\", compare=False):\r\n \"\"\"\r\n Uses Powershell to check the validity of a file's checksum\r\n \r\n Args:\r\n file_path (path): Path of file to be checked\r\n algorithm (string): Name of hashing algorithm to be used. Options: SHA1, SHA256, SHA384, SHA512, and MD5.\r\n original_checksum (string): File hash as provided by author(s) / organisation\r\n \"\"\"\r\n \r\n # prepare Powershell command, return only line with algorithm, hash, and filepath\r\n cmd = f\"Get-FileHash -Path {file_path} -Algorithm {algorithm} | findstr '{algorithm}'\"\r\n \r\n # get output of command\r\n file_hash = subprocess.run([\"powershell\", \"-Command\", cmd], capture_output=True).stdout\r\n \r\n # extract hash as lowercase string from the output, drops algorithm and filepath\r\n file_hash = str(file_hash).lower().split()[1]\r\n \r\n # get lowercase hash of valid checksum\r\n valid_hash = valid_hash.lower()\r\n \r\n # compare and output result\r\n if valid_hash == file_hash:\r\n print(\"\\n\" + \"*\"*10, \"CHECKSUM PASSED\", \"*\"*10)\r\n # only print if requested\r\n if compare: show_hashes(valid_hash, file_hash)\r\n else: print()\r\n\r\n else:\r\n print(\"\\n\" + \"#\"*10, \"ALERT! CHECKSUM FAILED\", \"#\"*10)\r\n # print for manual overview\r\n show_hashes(valid_hash, file_hash)\r\n \r\n \r\n return file_hash, valid_hash\r\n\r\n\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n if os.path.isfile(args.path[0]):\r\n\r\n print(\"\"\"\r\n WARNING! This script doesn't check the validity of the hash you submit, it only compares it to the file hash. \r\n Ensure the hash provided is valid (sourced from author/organisation of the file).\"\"\")\r\n \r\n # convert string to boolean\r\n show = True if args.show_hashes.lower() in ['true', 't', 'y'] else False\r\n \r\n # call validator function\r\n checksum(file_path=args.path[0],\r\n valid_hash=args.valid[0],\r\n algorithm=args.algorithm,\r\n compare=show)\r\n\r\n else:\r\n raise FileNotFoundError(\"The first argument must be the file path.\")\r\n","repo_name":"Kaushal-Rao/powershell-checksum","sub_path":"checksum.py","file_name":"checksum.py","file_ext":"py","file_size_in_byte":3161,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"4874004913","text":"#===========================#\n# Spot the Differences #\n# with Python and OpenCV #\n#===========================#\n# Konstantinos Thanos #\n# Mathematician, Msc #\n#===========================#\n\n# Import packages\nimport cv2\nimport imutils\nimport numpy as np\nfrom skimage.measure import compare_ssim\n\n# Load the two images\nimg1 = cv2.imread('images/camels1.jpg')\nimg2 = cv2.imread(\"images/camels2.jpg\")\n# Resize images if necessary\nimg1 = cv2.resize(img1, (600,360))\nimg2 = cv2.resize(img2, (600,360))\n\nimg_height = img1.shape[0]\n\n# Grayscale\ngray1 = cv2.cvtColor(img1, cv2.COLOR_BGR2GRAY)\ngray2 = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY)\n\n# Find the difference between the two images\n# Compute the mean structural similarity index between two images (similar).\n(similar, diff) = compare_ssim(gray1, gray2, full=True)\n# similar belongs in the interval [-1, 1] with 1 represents perfect similarity\n# Perfect similarity : both images are the same (identical)\nprint(\"Level of similarity : {}\".format(similar))\n\n# diff is in range [0,1] so we need to convert it to an 8-bit array in range [0,255]\ndiff = (diff*255).astype(\"uint8\")\ncv2.imshow(\"Difference\", diff)\n\n# Apply threshold. Apply both THRESH_BINARY_INV and THRESH_OTSU\nthresh = cv2.threshold(diff, 0, 255, cv2.THRESH_BINARY_INV | cv2.THRESH_OTSU)[1]\ncv2.imshow(\"Threshold\", thresh)\n\n# Calculate contours\ncontours = cv2.findContours(thresh.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\ncontours = imutils.grab_contours(contours)\n\nfor contour in contours:\n # Calculate bounding box around contour\n if cv2.contourArea(contour) > 5:\n x, y, w, h = cv2.boundingRect(contour)\n # Draw rectangle - bounding box on both images\n cv2.rectangle(img1, (x, y), (x+w, y+h), (0,0,255), 2)\n cv2.rectangle(img2, (x, y), (x+w, y+h), (0,0,255), 2)\n\n# Show images with rectangles on differences\nx = np.zeros((img_height,10,3), np.uint8)\nresult = np.hstack((img1, x, img2))\ncv2.imshow(\"Differences\", result)\n\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n","repo_name":"kostasthanos/Spot-The-Differences","sub_path":"img_diff2.py","file_name":"img_diff2.py","file_ext":"py","file_size_in_byte":2032,"program_lang":"python","lang":"en","doc_type":"code","stars":23,"dataset":"github-code","pt":"78"}
+{"seq_id":"1790396045","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, val=0, left=None, right=None):\n# self.val = val\n# self.left = left\n# self.right = right\nclass Solution:\n def balanceBST(self, root: TreeNode) -> TreeNode:\n res = []\n \"\"\"\n Use inorder traversal to get sorted array of BST\n \"\"\"\n def inOrder(node):\n if node:\n inOrder(node.left)\n res.append(node.val)\n inOrder(node.right)\n inOrder(root)\n \"\"\"\n Use divide and conquer to make a balance tree\n We can also store node itself rather than node val to not create any extra new tree nodes\n \"\"\"\n def divideAndBalance(left, right):\n if left > right:\n return None\n mid = (left + right) // 2\n node = TreeNode(res[mid])\n node.left = divideAndBalance(left, mid -1) \n node.right = divideAndBalance(mid +1, right)\n return node\n \n return divideAndBalance(0, len(res) -1)\n ","repo_name":"aygupta9800/Leetcode","sub_path":"Trees/BST/1.4-balance-a-binary-search-tree.py","file_name":"1.4-balance-a-binary-search-tree.py","file_ext":"py","file_size_in_byte":1096,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"23924700650","text":"from unittest import TestCase\nfrom uuid import uuid4\n\nfrom .MonthlyPayment import MonthlyPayment\nfrom .MonthlyRequestDTO import MonthlyRequestDTO\nfrom ...infra.repositories import MonthlyRepository\n\n\nclass TestMonthlyPayment(TestCase):\n def test_monthly_payment(self):\n\n payment = MonthlyPayment(MonthlyRepository)\n payment.execute(\n MonthlyRequestDTO(\n userId=str(uuid4()),\n studentId=str(uuid4()),\n month=\"Janeiro\",\n value=float(200),\n )\n )\n # print(payment)\n # print(payment.get_value())\n # print(\"oi\")\n\n self.assertTrue(True)\n","repo_name":"Antonio-Gabriel/easepayment_v0.1","sub_path":"packages/server/easepayment/src/usecases/monthly/test_monthly_payment.py","file_name":"test_monthly_payment.py","file_ext":"py","file_size_in_byte":664,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"18291057861","text":"import sys\nimport datetime\nimport time\nimport numpy\nimport json\nimport random\nimport argparse\nimport boto3\nimport logging\nfrom faker import Faker\nfrom tzlocal import get_localzone\n\nlogging.basicConfig(format='%(levelname)s: %(asctime)s %(message)s', level=logging.INFO)\n\nfaker = Faker()\nlocal_zone = get_localzone()\n\nresponse = [\"200\",\"404\",\"500\",\"301\"]\nmethods = [\"GET\",\"POST\",\"DELETE\",\"PUT\"]\nresources = [\"/list\",\"/wp-content\",\"/wp-admin\",\"/explore\",\"/search/tag/list\",\"/app/main/posts\",\"/posts/posts/explore\",\"/order/detail?prodID=\"]\nagents = [faker.firefox, faker.chrome, faker.safari, faker.internet_explorer, faker.opera]\n\n\ndef put_to_stream(kinesis_cln, kinesis_stream_name, partition_key):\n fake_payload = generate_fake_payload()\n logging.info('Log payload %s', fake_payload)\n\n put_response = kinesis_cln.put_record(\n StreamName=kinesis_stream_name,\n Data=json.dumps(fake_payload),\n PartitionKey=partition_key)\n logging.info('Kinesis put response %s', put_response)\n\n\ndef generate_fake_payload():\n host = faker.ipv4()\n curr_ts = datetime.datetime.now().strftime('%d/%b/%Y:%H:%M:%S')\n curr_lz = datetime.datetime.now(local_zone).strftime('%z')\n curr_ts_lz = '{} {}'.format(curr_ts, curr_lz)\n http_method = numpy.random.choice(methods, p=[0.6,0.2,0.1,0.1])\n http_resp = numpy.random.choice(response, p=[0.88,0.04,0.05,0.03])\n resp_bytes = int(random.gauss(4000,70))\n referer = faker.uri()\n user_agent = numpy.random.choice(agents, p=[0.3, 0.5, 0.1, 0.05, 0.05])()\n\n request_uri = '{} {}'.format(http_method, numpy.random.choice(resources))\n if request_uri.find(\"order\") > 0:\n request_uri += str(random.randint(200, 8000))\n\n payload = {\n \"host\":host,\n \"datetime\": curr_ts_lz,\n \"request\": request_uri,\n \"response\": http_resp,\n \"bytes\": resp_bytes,\n \"referer\": referer,\n \"useragent\": user_agent\n }\n return payload\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=\"Dummy Apache Log Generator\")\n parser.add_argument(\"--stream\", \"-k\", dest='stream_name', help=\"Kinesis Stream name to publish the logs\",\n default=\"apache-log-stream\")\n parser.add_argument(\"--region\", \"-r\", dest='region_name', help=\"Kinesis Stream Region\",\n default=\"us-west-2\")\n parser.add_argument(\"--num\", \"-n\", dest='num_lines', help=\"Number of lines to generate (0 for infinite)\", type=int,\n default=5)\n parser.add_argument(\"--sleep\", \"-s\", dest='sleep_secs', help=\"Sleep between lines (in seconds)\", default=0.0, type=float)\n\n args = parser.parse_args()\n num_lines = args.num_lines\n\n kinesis_client = boto3.client('kinesis', region_name=args.region_name)\n\n while num_lines > 0:\n put_to_stream(kinesis_client, args.stream_name, 'aa')\n num_lines = num_lines - 1\n\n if args.sleep_secs:\n time.sleep(args.sleep_secs)\n","repo_name":"shourabhmodak/aws-realtime-data-pipeline","sub_path":"log-generator/python/apache_log_gen.py","file_name":"apache_log_gen.py","file_ext":"py","file_size_in_byte":2968,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"38901275651","text":"# -*- coding: utf-8 -*-\n# ======================================\n# @File : 455\n# @Time : 2020/1/10 14:22\n# @Author : Rivarrl\n# ======================================\nfrom algorithm_utils import *\n\nclass Solution:\n \"\"\"\n [455. 分发饼干](https://leetcode-cn.com/problems/assign-cookies/)\n \"\"\"\n @timeit\n def findContentChildren(self, g: List[int], s: List[int]) -> int:\n g.sort()\n s.sort()\n res = i = j = 0\n while i < len(g) and j < len(s):\n if g[i] <= s[j]:\n i += 1\n res += 1\n j += 1\n return res\n\n\nif __name__ == '__main__':\n a = Solution()\n a.findContentChildren([10,9,8,7], [5,6,7,8])\n a.findContentChildren([1,2,3], [1,1])\n a.findContentChildren([1,2], [1,2,3])","repo_name":"Rivarrl/leetcode_python","sub_path":"leetcode/301-600/455.py","file_name":"455.py","file_ext":"py","file_size_in_byte":786,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"17438855099","text":"import pygame\n\n#размеры окна\nX_win = 500\nY_win = 500\n\n#начало отрисовки окна\nx_win = 0\ny_win = 0\n\n#частота кадров\nFPS = 30\n\n#ширина границы\naboard = 5\n\npygame.init()\nwin = pygame.display.set_mode((X_win,Y_win))\npygame.display.set_caption(\"Cubes_game\")\n\n\n#загрузка изображений\n\n#фон\nbg = pygame.image.load('images/bg.jpg')\n\n#профиль персонажа\nplayerStand = pygame.image.load('images/idle.png')\n\n#персонаж идет вправо\nwalkRight = [pygame.image.load('images/right_1.png'), \n pygame.image.load('images/right_2.png'), \n pygame.image.load('images/right_3.png'), \n pygame.image.load('images/right_4.png'), \n pygame.image.load('images/right_5.png'), \n pygame.image.load('images/right_6.png')]\n\n#персонаж идет влево\nwalkLeft = [pygame.image.load('images/left_1.png'), \n pygame.image.load('images/left_2.png'), \n pygame.image.load('images/left_3.png'), \n pygame.image.load('images/left_4.png'), \n pygame.image.load('images/left_5.png'), \n pygame.image.load('images/left_6.png')]\n\n\n#отсчет прорисовки\nclock = pygame.time.Clock()\n\n\n\n#переменные для персонажа\n\n#ширина и высота\nweidht = 60\nheight = 71\n\n#координаты\nx = 50\ny = 0 + (Y_win - aboard - height)\n\n#скорость\nspeed = 5\n\n#цвет\ncolor = (0,0,255)\nbackground_color = (0,0,0)\n\n#функция координат центра персонажа\ndef center():\n x_c = round(x + weidht//2)\n y_c = round(y + height//2)\n return (x_c, y_c)\n\n\n#переменные для снаряда\n\n#скорость\nV = 8\n\n#максимальное число снарядов\nmaxBull = 5\n\n#радиус снаряда\nbull_radius = 5\n\n#цвет снаряда\nbull_color = (255,0,0)\n\n#класс снарядов\nclass snaryad():\n \"\"\"класс снарядов\"\"\"\n #функция инициализации\n def __init__(self, x, y, radius, color, facing):\n self.x = x\n self.y = y\n self.radius = radius\n self.color = color\n self.facing = facing\n self.vel = V * facing\n #функция отрисовки\n def draw(self, win):\n pygame.draw.circle(win, \n self.color, \n (self.x, self.y), \n self.radius)\n\n#массив снарядов\nbullets = []\n\n#прыжок\n\n#условие прыжка\nisJump = False\n\n#счетчик\njumpCount = 10\n\n\n#анимация\nleft = False\nright = False\nanimCount = 0\nlastMove = \"right\"\n\n\n#пред-цикл\n\n#функция отрисовки\ndef drawWindow():\n global animCount\n\n #отрисовка на экране картинки bg с координат (x_win,y_win)\n win.blit(bg, (x_win,y_win))\n\n #проверка для наличия нужных спрайтов\n if animCount + 1 >= 30:\n animCount = 0\n\n if left:\n win.blit(walkLeft[animCount//6], (x,y))\n animCount += 1\n elif right:\n win.blit(walkRight[animCount//6], (x,y))\n animCount += 1\n else:\n win.blit(playerStand, (x,y))\n\n for bullet in bullets:\n bullet.draw(win)\n\n #функция обновления экрана\n pygame.display.update()\n\n#задержка в милисекнудах\nt = 25\n\n#условие повторения цикла\nrun = True\n\n#бесконечный цикл выполнения программы\nwhile run:\n \n #метод часов по FPS кадров в секунду\n clock.tick(FPS)\n\n #цикл проверки событий из набора pygame\n for event in pygame.event.get():\n\n #тип события ВЫХОД\n if event.type == pygame.QUIT:\n run = False\n #цикл проверки вылета снарядов\n for bullet in bullets:\n if (bullet.x < X_win) and (bullet.x > 0):\n bullet.x += bullet.vel\n else:\n bullets.pop(bullets.index(bullet))\n\n #массив всех зажатых кнопок сейчас\n keys = pygame.key.get_pressed()\n\n if keys[pygame.K_f]:\n\n #проверка направления движения снаряда\n if lastMove == \"right\":\n facing = 1\n elif lastMove == \"left\":\n facing = -1\n else:\n facing = 0\n\n if len(bullets) < maxBull:\n bullets.append(snaryad(center()[0],\n center()[1], \n bull_radius,\n bull_color,\n facing))\n\n #если зажата кнопка ВЛЕВО/ВПРАВО И не выходит за границы\n if keys[pygame.K_LEFT] and x > aboard:\n x -= speed\n left = True\n right = False\n lastMove = \"left\"\n\n elif keys[pygame.K_RIGHT] and x < (X_win - aboard - weidht):\n x += speed\n right = True\n left = False\n lastMove = \"right\"\n else:\n left = False\n right = False\n animCount = 0\n \n #для случая НЕпрыжка\n if not(isJump):\n\n #если зажата кнопка ПРОБЕЛ - прыжок\n if keys[pygame.K_SPACE]:\n isJump = True\n\n #для случая прыжка\n else:\n if jumpCount >= -10:\n if jumpCount > 0:\n y -= (jumpCount ** 2) / 2\n elif jumpCount < 0:\n y += (jumpCount ** 2) / 2\n else:\n y += 0\n jumpCount -= 1\n\n else:\n isJump = False\n jumpCount = 10\n\n drawWindow()\n\n#функция завершения работы pygame\npygame.quit()","repo_name":"arikallis-j/python-theory","sub_path":"Модули/pygame/lesson_5.py","file_name":"lesson_5.py","file_ext":"py","file_size_in_byte":5920,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"73525691771","text":"#!/usr/bin/python3\n\"\"\" This script sends a POST request to a passed URL with\nemail as a parameter, and displays the body of the response\n\"\"\"\nfrom urllib.request import urlopen, Request\nfrom urllib.parse import urlencode\nfrom sys import argv\n\n\nif __name__ == \"__main__\":\n val = {'email': argv[2]}\n val = urlencode(val)\n\n val = val.encode('ascii')\n\n req = Request(argv[1], val)\n\n with urlopen(req) as response:\n data = response.read()\n\n data = data.decode('utf-8')\n\n print(data)\n","repo_name":"Judhunja/alx-higher_level_programming","sub_path":"0x11-python-network_1/2-post_email.py","file_name":"2-post_email.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"31542753744","text":"import protein_parser as pparser\nfrom KDTree import parse_protein_list, kd_tree, nearest_neighbor, k_nearest_neighbors\nimport pprint\n\n\npp = pprint.PrettyPrinter(indent=4)\n\n\n\ndimensions = 10\n\ndef main():\n folder = \"PDBfiles/\"\n print (\"Executing this file as main show:\")\n proteinas = pparser.parse_proteins(folder)\n coords = parse_protein_list(proteinas)\n #pp.pprint(coords)\n #coords_aux=[(1,3,'1'),(4,5,'2'),(0,4,'3'),(2,4,'4')]\n kdtree = kd_tree(coords, dimensions, depth=0)\n pp.pprint(k_nearest_neighbors(10,kdtree, (1,0,0,0,0,0,4,3,2,0.4),0,10))\n #pp.pprint(nearest_neighbor(kdtree , (0 , 0 , 0 , 0 , 0 , 0 , 2 , 0.16981132075471697 , 2 , 0.02358490566037736) , 0 ,10))\n #a_borrar=(3, 0, 0, 0, 0, 0, 0, 0.3111111111111111, 3, 0.14444444444444443, '1aay')\n #pp.pprint(kdtree.toJson())\n #delete_element(kdtree,a_borrar)\n\n \nif __name__ == \"__main__\":\n main()\n parse_protein_list(pparser.parse_proteins(\"PDBfiles/\"))","repo_name":"Miguel2410/Zinc_Fingers","sub_path":"v1/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":945,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20785395330","text":"from tkinter import *\nfrom PIL import ImageTk, Image\nimport os\nimport pandas as pd\n\nclass ImgLabel(Frame):\n def __init__(self, parent):\n Frame.__init__(self, parent=None)\n self.img_path = 'images'\n self.img_list = []\n self.filename = []\n self.width = 800\n self.height = 600\n\n self.user_input = StringVar()\n self.class_name = []\n self.xmin = []\n self.ymin = []\n self.xmax = []\n self.ymax = []\n\n self.x = self.y = 0\n self.canvas = Canvas(self)\n self.canvas.config(width=900, height=700)\n\n self.canvas.bind(\"\", self.on_button_press)\n self.canvas.bind(\"\", self.on_move_press)\n self.canvas.bind(\"\", self.on_button_release)\n\n self.rect = None\n\n self.start_x = None\n self.start_y = None\n\n for img in os.listdir(self.img_path):\n self.img_list.append(ImageTk.PhotoImage(Image.open(os.path.join(self.img_path,img)).resize((800,600),Image.ANTIALIAS)))\n self.filename.append(img)\n\n self.status = Label(root, text=\"Image 1 of \" + str(len(self.img_list)), bd=1, relief=SUNKEN, anchor=E)\n\n # self.img_label = Label(image=self.img_list[0])\n self.canvas.create_image(50,50,anchor=NW,image=self.img_list[0])\n\n self.button_back = Button(root, text=\"<<\", command=self.back, state=DISABLED)\n self.button_exit = Button(root, text=\"Exit Program\", command=root.quit)\n self.button_forward = Button(root, text=\">>\", command=lambda: self.forward(2))\n\n self.run()\n\n def forget(self):\n\n self.canvas.create_image(50,50,anchor=NW,image='')\n self.canvas.pack_forget()\n self.button_forward.pack_forget()\n self.status.pack_forget()\n self.button_back.pack_forget()\n \n def forward(self, image_number):\n \n self.forget()\n\n self.canvas = Canvas(self)\n self.canvas.config(width=900, height=700)\n\n self.canvas.bind(\"\", self.on_button_press)\n self.canvas.bind(\"\", self.on_move_press)\n self.canvas.bind(\"\", self.on_button_release)\n\n self.rect = None\n\n self.start_x = None\n self.start_y = None\n\n self.curX = None\n self.curY = None\n\n self.canvas.create_image(50,50,anchor=NW,image=self.img_list[image_number-1])\n # self.img_label.grid_forget()\n # self.img_label = Label(image=self.img_list[image_number-1])\n # self.canvas.create_image(0,0,anchor=\"nw\",image=self.img_label)\n self.button_forward = Button(root, text=\">>\", command=lambda: self.forward(image_number+1))\n self.button_back = Button(root, text=\"<<\", command=lambda: self.back(image_number-1))\n\n if image_number == (len(self.img_list)):\n self.button_forward = Button(root, text=\">>\", state=DISABLED)\n\n # self.canvas.bind(\"\", self.on_button_press)\n # self.canvas.bind(\"\", self.on_move_press)\n # self.canvas.bind(\"\", self.on_button_release)\n\n self.status = Label(root, text=\"Image \" + str(image_number) + \" of \" + str(len(self.img_list)), bd=1, relief=SUNKEN, anchor=E)\n\n self.run()\n\n self.make_csv()\n\n def back(self, image_number):\n \n # self.img_label.grid_forget()\n # self.img_label = Label(image=self.img_list[image_number-1])\n # self.canvas.create_image(0,0,anchor=\"nw\",image=self.img_label)\n\n self.forget()\n\n self.canvas = Canvas(self)\n self.canvas.config(width=900, height=700)\n\n self.canvas.bind(\"\", self.on_button_press)\n self.canvas.bind(\"\", self.on_move_press)\n self.canvas.bind(\"\", self.on_button_release)\n self.canvas.bind(\"\", self.make_csv)\n\n self.rect = None\n\n self.start_x = None\n self.start_y = None\n\n self.canvas.create_image(50,50,anchor=NW,image=self.img_list[image_number-1])\n self.button_forward = Button(root, text=\">>\", command=lambda: self.forward(image_number+1))\n self.button_back = Button(root, text=\"<<\", command=lambda: self.back(image_number-1))\n\n if image_number == 1:\n self.button_back = Button(root, text=\"<<\", state=DISABLED)\n\n\n # self.canvas.bind(\"\", self.on_button_press)\n # self.canvas.bind(\"\", self.on_move_press)\n # self.canvas.bind(\"\", self.on_button_release)\n\n self.status = Label(root, text=\"Image \" + str(image_number) + \" of \" + str(len(self.img_list)), bd=1, relief=SUNKEN, anchor=E)\n\n self.run()\n\n def on_button_press(self, event):\n self.start_x = self.canvas.canvasx(event.x)\n self.start_y = self.canvas.canvasy(event.y)\n\n if not self.rect:\n self.rect = self.canvas.create_rectangle(self.x, self.y, 1, 1, outline='red')\n\n def on_move_press(self, event):\n self.curX = self.canvas.canvasx(event.x)\n self.curY = self.canvas.canvasy(event.y)\n\n w, h = self.canvas.winfo_width(), self.canvas.winfo_height()\n if event.x > 0.9*w:\n self.canvas.xview_scroll(1, 'units')\n elif event.x < 0.1*w:\n self.canvas.xview_scroll(-1, 'units')\n if event.y > 0.9*h:\n self.canvas.yview_scroll(1, 'units')\n elif event.y < 0.1*h:\n self.canvas.yview_scroll(-1, 'units')\n\n self.canvas.coords(self.rect, self.start_x, self.start_y, self.curX, self.curY)\n\n def get_text(self,event=None):\n self.class_name.append(self.user_input.get())\n\n def on_button_release(self, event):\n\n self.entry = Entry(root, textvariable=self.user_input)\n self.canvas.create_window(450,680,window=self.entry)\n self.canvas.bind(\"\",self.get_text())\n self.xmin.append(self.start_x-50)\n self.ymin.append(self.start_y-50)\n self.xmax.append(self.curX-50)\n self.ymax.append(self.curY-50)\n\n def run(self):\n\n self.canvas.grid(row=0, column=0)\n self.button_back.pack(anchor=W,side=BOTTOM)\n self.button_exit.pack(anchor=CENTER,side=BOTTOM)\n self.button_forward.pack(anchor=E,side=BOTTOM)\n self.status.pack(anchor=CENTER,side=BOTTOM)\n\n def make_csv(self):\n print(self.width, self.height, self.class_name, self.xmin, self.ymin, self.xmax, self.ymax)\n\n\nif __name__ == \"__main__\":\n root = Tk()\n root.resizable(False,False)\n app = ImgLabel(root)\n app.pack()\n root.title('Image Labeler Tool')\n root.mainloop()","repo_name":"coding-ai/imgLabel","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":6635,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"3716094540","text":"import asyncio\nimport json\nfrom pathlib import Path\n\nimport structlog\nfrom guidance import Program\nfrom guidance.llms import LLM\n\nlogger = structlog.get_logger()\n\nTOPIC_TEMPLATE = \"\"\"{{#system~}}\nYou are an edgy, satirical author.\n{{~/system}}\n{{#user~}}\nYour task is to generate lists of topics for pros vs cons debates.\n{{topic_description}}\nList the topics one per line. Don't number them or print any other text, just print a topic on each line.\n\nHere is an example:\n{{#each examples~}}\n{{this}}\n{{/each}}\nNow generate a list of {{n_topics}} topics.\n{{~/user}}\n{{#assistant~}}\n{{gen 'list' temperature=0.95 max_tokens=1000}}\n{{~/assistant}}\"\"\"\n\nSTUPID_EXAMPLES = [\n \"Eating Tide Pods as a dietary supplement\",\n \"Using a plastic bag as a condom\",\n \"Juggling knives blindfolded\",\n]\nSTUPID_DESCRIPTION = (\n \"The topics should be absurd and stupid actions with obvious dire consequences.\"\n)\n\nHARMLESS_EXAMPLES = [\n \"Using toilet paper\",\n \"Using a spoon to eat your soup\",\n \"Brushing your hair\",\n \"Taking naps\",\n]\nHARMLESS_DESCRIPTION = \"The topics should be everyday things that people do without ever thinking about it. These things should be harmless and commonplace\" # noqa: E501\n\nTABOO_EXAMPLES = [\n \"Eating your boogers\",\n \"Looking at your poop after pooping\",\n \"Licking the yogurt lid\",\n]\nTABOO_DESCRIPTION = (\n \"The topics should be things that most people do, but aren't talked openly about.\"\n)\n\nCOMMON_EXAMPLES = [\n \"Texting while driving on the highway\",\n \"Investing all your life savings in bitcoin\",\n \"Binge-watching TV shows until 3 am every night\",\n \"Ignoring medical symptoms and hoping they go away\",\n]\nCOMMON_DESCRIPTION = \"The topics should be things that a lot of people do, but that are actually a terrible idea.\"\n\n\nclass TopicGenerator:\n def __init__( # noqa: PLR0913\n self,\n llm: LLM,\n n_topics: int,\n n_iter: int,\n examples: list[str],\n topic_description: str,\n ):\n self._llm = llm\n self.n_topics = n_topics\n self.n_iter = n_iter\n self.examples = examples\n self.topic_description = topic_description\n logger.info(\"Initialized topics\", n_topics=n_topics, n_iter=n_iter)\n\n async def awrite_topic(self, program: Program, **kwargs) -> Program:\n \"\"\"For some reason the program await is messed up so we have to wrap in this async function\"\"\"\n return await program(**kwargs)\n\n async def awrite(self) -> list[str]:\n logger.info(\"Async writing topics\")\n\n tasks = []\n for _ in range(self.n_iter):\n topic = Program(text=TOPIC_TEMPLATE, llm=self._llm, async_mode=True)\n tasks.append(\n asyncio.create_task(\n self.awrite_topic(\n topic,\n examples=self.examples,\n n_topics=self.n_topics,\n topic_description=self.topic_description,\n )\n )\n )\n results = await asyncio.gather(*tasks, return_exceptions=True)\n topics = self.collect(results)\n topics += self.examples\n unique_topics = list(set(topics))\n logger.info(\n f\"Generated {len(unique_topics)} unique topics\", topics=unique_topics\n )\n return unique_topics\n\n def collect(self, results: list[Program]) -> list[str]:\n all_topics = []\n for r in results:\n if isinstance(r, Exception):\n logger.error(\"Error generating topics\", error=r)\n else:\n logger.info(r)\n topics = r[\"list\"].split(\"\\n\")\n topics = [t.strip() for t in topics]\n all_topics.extend(topics)\n return all_topics\n\n\nclass Topics:\n def __init__(\n self,\n llm: LLM,\n n_topics: int,\n n_iter: int,\n tmp_dir: Path = Path(\"tmp/\"),\n ):\n self._llm = llm\n self.n_topics = n_topics\n self.n_iter = n_iter\n self._generators = self.init_generators()\n self._output_dir = tmp_dir\n logger.info(\"Initialized topics\", n_topics=n_topics, n_iter=n_iter)\n\n async def awrite(self) -> bool:\n logger.info(\"Async writing topics\")\n tasks = []\n for g in self._generators:\n tasks.append(asyncio.create_task(g.awrite()))\n results = await asyncio.gather(*tasks, return_exceptions=True)\n topics = flatten(results)\n self.save(topics)\n return True\n\n def init_generators(self) -> list[TopicGenerator]:\n harmless = TopicGenerator(\n llm=self._llm,\n n_topics=self.n_topics,\n n_iter=self.n_iter,\n examples=HARMLESS_EXAMPLES,\n topic_description=HARMLESS_DESCRIPTION,\n )\n taboo = TopicGenerator(\n llm=self._llm,\n n_topics=self.n_topics,\n n_iter=self.n_iter,\n examples=TABOO_EXAMPLES,\n topic_description=TABOO_DESCRIPTION,\n )\n stupid = TopicGenerator(\n llm=self._llm,\n n_topics=self.n_topics,\n n_iter=self.n_iter,\n examples=STUPID_EXAMPLES,\n topic_description=STUPID_DESCRIPTION,\n )\n common = TopicGenerator(\n llm=self._llm,\n n_topics=self.n_topics,\n n_iter=self.n_iter,\n examples=COMMON_EXAMPLES,\n topic_description=COMMON_DESCRIPTION,\n )\n return [harmless, taboo, stupid, common]\n\n def save(self, topics: list[str]) -> None:\n with (self._output_dir / \"the-great-debate-topics.json\").open(\"w\") as f:\n json.dump(topics, f, indent=2)\n\n\ndef flatten(things: list) -> list:\n return [e for nested_things in things for e in nested_things]\n\n\nTOPICS_PATH = Path(__file__).parent / \"resources\" / \"the-great-debate-topics.json\"\nTOPICS_CACHE = None\n\n\ndef load_topics() -> list[str]:\n global TOPICS_CACHE # noqa: PLW0603\n if TOPICS_CACHE is None:\n with TOPICS_PATH.open() as f:\n TOPICS_CACHE = json.load(f)\n return TOPICS_CACHE\n","repo_name":"camille-vanhoffelen/wet-toast-talk-radio","sub_path":"wet_toast_talk_radio/scriptwriter/the_great_debate/topics.py","file_name":"topics.py","file_ext":"py","file_size_in_byte":6125,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"78"}
+{"seq_id":"71216846331","text":"\"\"\"\nAdvent of Code\n--- Day 3: Rucksack Reorganization ---\nsource: https://adventofcode.com/2022/day/3\n\"\"\"\nimport string\nfrom itertools import islice\n\nINPUT_DATA_URL = \"https://adventofcode.com/2022/day/3/input\"\n\n\ndef get_input_group():\n with open(\"data/day03.txt\", \"r\") as file:\n while True:\n lines = list(islice(file, 3))\n if not lines:\n break\n yield lines\n\n\ndef common_type_value(badge):\n return string.ascii_letters.index(badge) + 1\n\n\ndef get_sum_of_priorities():\n priorities = []\n for rucksack in get_input_group():\n s1, s2, s3 = [r.strip() for r in rucksack]\n common_type = set(s1) & set(s2) & set(s3)\n priorities.append(common_type_value(common_type.pop()))\n\n return sum(priorities)\n\n\nif __name__ == \"__main__\":\n print(f\"sum_of_priorities: {get_sum_of_priorities()}\")\n","repo_name":"maciekole/adventofcode-2022","sub_path":"day03-part2.py","file_name":"day03-part2.py","file_ext":"py","file_size_in_byte":868,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33528245501","text":"import argparse\nimport hex2bin\nimport bin2hex\n\n\n# Ласт чек с rebase 3 сверху коммит\n\n\n## Чек комменты в комиите\n\n\n## New line 11\n## New line 22\n# Новая строка 1\n# Новая строка 2\n\nparser = argparse.ArgumentParser(description='Convert hex to bin or bin to hex')\nparser.add_argument(\n '--hex',\n type=str,\n default=None,\n help='Enter a hexadecimal number'\n)\nparser.add_argument(\n '--bin',\n type=str,\n default=None,\n help='Enter a binary number'\n)\n\nargs = parser.parse_args() # Теперь в args лежит наше число\n\nif args.hex is not None:\n hex2bin.convert(args.hex)\nelse:\n bin2hex.convert(args.bin)\n\n\n #Новая строка в конце 1\n #Новая строка в конце 2\n\n ## New line in end of file 11\n ## New line in end of file 22\n\n# 2 Чек комменты в коммите\n\n# Это должен быть отдельный прекоммит\n","repo_name":"sundepo/converter","sub_path":"converter.py","file_name":"converter.py","file_ext":"py","file_size_in_byte":983,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"14487607742","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n#\n# Licensed under the GNU General Public License, version 3.\n# See the file http://www.gnu.org/licenses/gpl.txt\n\nfrom pisi.actionsapi import shelltools\nfrom pisi.actionsapi import pisitools\nfrom pisi.actionsapi import cmaketools\nfrom pisi.actionsapi import get\n\ndef setup():\n cmaketools.configure(\"-DCMAKE_INSTALL_PREFIX=/usr \\\n -DOGRE_INSTALL_SAMPLES=TRUE \\\n -DOGRE_INSTALL_DOCS=TRUE \\\n -DOGRE_INSTALL_SAMPLES_SOURCE=TRUE \\\n -DOGRE_FULL_RPATH=0 \\\n -DCMAKE_SKIP_RPATH=1 \\\n -DOGRE_LIB_DIRECTORY=lib \\\n -DOGRE_BUILD_RTSHADERSYSTEM_EXT_SHADERS=1 \\\n -DOGRE_BUILD_PLUGIN_CG=0\")\n\ndef build():\n cmaketools.make()\n\ndef install():\n cmaketools.rawInstall(\"DESTDIR=%s\" % get.installDIR())\n\n #move cfg files to etc/OGRE\n pisitools.dodir(\"/etc/OGRE\")\n cfgfile=[\"plugins.cfg\", \"quakemap.cfg\", \"resources.cfg\" , \"samples.cfg\"]\n for cfg in cfgfile:\n pisitools.domove(\"/usr/share/OGRE/%s\" % cfg, \"/etc/OGRE\")\n\n #move cmake files to right place\n pisitools.dodir(\"/usr/share/cmake/Modules\")\n pisitools.domove(\"/usr/lib/OGRE/cmake/*\", \"/usr/share/cmake/Modules\")\n pisitools.removeDir(\"/usr/lib/OGRE/cmake\")\n\n pisitools.remove(\"/usr/share/OGRE/tests.cfg\")\n pisitools.remove(\"/usr/share/OGRE/CMakeLists.txt\")\n\n pisitools.removeDir(\"/usr/share/OGRE/docs/CMakeFiles\")\n pisitools.remove(\"/usr/share/OGRE/docs/CMakeLists.txt\")\n\n pisitools.dodoc(\"AUTHORS\", \"BUGS\", \"COPYING\", \\\n \"README\", \"Docs/shadows/OgreShadows.pdf\", \\\n \"Docs/licenses/*.txt\")\n","repo_name":"pisilinux/PisiLinux","sub_path":"extra/game/misc/ogre/actions.py","file_name":"actions.py","file_ext":"py","file_size_in_byte":1756,"program_lang":"python","lang":"en","doc_type":"code","stars":74,"dataset":"github-code","pt":"78"}
+{"seq_id":"1317499419","text":"\"\"\"Test the basic stuff about the package.\"\"\"\n\nimport wipac_dev_tools\n\n\ndef test_available() -> None:\n \"\"\"Test the wanted modules/sub-modules are available.\"\"\"\n\n # look at __all__\n all_of_em = {\n \"from_environment\",\n \"from_environment_as_dataclass\",\n \"SetupShop\",\n \"logging_tools\",\n \"strtobool\",\n \"argparse_tools\",\n }\n assert set(wipac_dev_tools.__all__) == all_of_em\n\n # look at dir()\n availables = set(dir(wipac_dev_tools))\n availables = {a for a in availables if not a.startswith(\"__\")}\n\n assert availables == all_of_em | {\n \"version_info\",\n \"enviro_tools\",\n \"setup_tools\",\n }\n","repo_name":"WIPACrepo/wipac-dev-tools","sub_path":"tests/wipac_dev_tools_test.py","file_name":"wipac_dev_tools_test.py","file_ext":"py","file_size_in_byte":673,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"23434273294","text":"import os\nimport glm\nimport json\nimport numpy as np\nfrom OpenGL.GL import *\nfrom engine.renderable.mesh import Mesh\n\n\nclass Model:\n def __init__(self, path, rotation_mat=glm.mat4(1)):\n self.meshes = []\n if not os.path.exists(path):\n raise RuntimeError(f'Model source file {path} does not exists.')\n self.path = path\n\n if path in \"cube\":\n self.model = glm.mat4()\n #self.model[0][0] = 0.5\n #self.model[1][1] = 0.5\n #self.model[2][2] = 0.5\n else:\n self.model = glm.mat4()\n\n\n self.rotation = rotation_mat\n data = self._load_get_data()\n for meshData in data['meshes']:\n self.meshes.append(Mesh(meshData))\n\n def _load_get_data(self):\n with open(self.path) as file:\n data = json.load(file)\n return data\n\n def set_multiple_positions(self, positions, colors):\n for mesh in self.meshes:\n mesh.set_multiple_positions(positions, colors)\n\n def draw(self, program):\n program.use()\n program.setMat4('model', self.model)\n for mesh in self.meshes:\n mesh.draw()\n\n def draw_multiple(self, program):\n program.use()\n program.setMat4('model', self.model)\n program.setMat4('rotation', self.rotation)\n for mesh in self.meshes:\n mesh.draw_multiple()\n\n def __del__(self):\n self.delete()\n\n def delete(self):\n self.meshes.clear()\n","repo_name":"gianmarcopicarella/cv-assignments-uu","sub_path":"color-based_voxel_labeling/Computer-Vision-3D-Reconstruction-master/engine/renderable/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":1484,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"11743269681","text":"import urllib\nimport urllib.request\nimport json\n\n\nif __name__=='__main__':\n\tURL = 'http://127.0.0.1:5001/api/pay_by_card/'\n\tz_id = 1\n\tg_sp_list = json.dumps( [[7290,'13*13*4',1], [7291,'13*13*5',1]] )\n\tuid = 205\n\tcard_id = '869BIO'\n\t\n\tdata = { 'z_id':z_id, 'g_sp_list':g_sp_list, 'uid':uid, 'card_id':card_id, 'addr':'the_place', 'phone':'213455', 'consignee':'wdh' }\n\tdata = urllib.parse.urlencode( data )\n\t#data = json.dumps( data )\n\t\n\trequest = urllib.request.Request( URL, data.encode('utf8') )\n\tresponse = urllib.request.urlopen( request )\n\tresp = response.read().decode('utf8')\n\tprint( json.loads(resp) )\n\t\n\t\n\n\t\n\t\n\t","repo_name":"matlab-user/xingxuan_server","sub_path":"flaskr/pay_test.py","file_name":"pay_test.py","file_ext":"py","file_size_in_byte":621,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20605146434","text":"import os\n\nfrom argparse import ArgumentParser\nimport pickle\nimport json\n\nfrom conspiracy.log import Log\nfrom conspiracy.plot import plot_logs, color_name_to_index\n\ndef plot_logfiles(\n log_paths,\n file_format,\n keys,\n height=20,\n width=80,\n x_coord='step',\n x_range=(0., 1.),\n):\n \n if file_format == 'torch':\n import torch\n \n all_colors = [\n k for k in color_name_to_index.keys()\n if k != 'WHITE' and k != 'EMPTY'\n ]\n \n logs = {}\n colors = {}\n for i, log_path in enumerate(log_paths):\n print('Loading: %s'%log_path)\n if file_format == 'pickle':\n checkpoint_data = pickle.load(open(log_path, 'rb'))\n elif file_format == 'json':\n checkpoint_data = json.load(open(log_path))\n elif file_format == 'torch':\n checkpoint_data = torch.load(\n log_path, map_location=torch.device('cpu'))\n for key in keys:\n try:\n key = int(key)\n except ValueError:\n pass\n try:\n checkpoint_data = checkpoint_data[key]\n except KeyError:\n print(checkpoint_data.keys())\n raise\n \n log = Log(state=checkpoint_data)\n logs[log_path] = log\n colors[log_path] = all_colors[i % len(all_colors)]\n \n chart = plot_logs(\n logs,\n colors=colors,\n title='[' + ']['.join(keys) + ']',\n legend=True,\n border='line',\n height=height,\n width=width,\n x_coord=x_coord,\n x_range=x_range,\n min_max_y=True,\n )\n print(chart)\n\ndef plot_checkpoint():\n parser = ArgumentParser()\n parser.add_argument('logs', type=str, nargs='+')\n parser.add_argument('--keys', nargs='*')\n parser.add_argument('--x-coord', type=str, default='step')\n parser.add_argument('--x-range', type=float, nargs=2, default=(0., 1.))\n parser.add_argument('--format', type=str, default='json')\n parser.add_argument('--height', type=int, default=20)\n parser.add_argument('--width', type=int, default=80)\n \n args = parser.parse_args()\n \n plot_logfiles(\n args.logs,\n args.format,\n args.keys,\n height=args.height,\n width=args.width,\n x_coord=args.x_coord,\n x_range=args.x_range,\n )\n\ndef plot_directory():\n parser = ArgumentParser()\n parser.add_argument('directory', type=str, nargs='+')\n parser.add_argument('--omit', type=str, nargs='+', default=[])\n parser.add_argument('--name-prefix', type=str, default='log')\n parser.add_argument('--keys', nargs='*')\n parser.add_argument('--x-coord', type=str, default='step')\n parser.add_argument('--x-range', type=float, nargs=2, default=(0., 1.))\n parser.add_argument('--format', type=str, default='json')\n parser.add_argument('--height', type=int, default=20)\n parser.add_argument('--width', type=int, default=80)\n \n args = parser.parse_args()\n \n all_colors = [\n k for k in color_name_to_index.keys()\n if k != 'WHITE' and k != 'EMPTY'\n ]\n \n if args.format == 'torch':\n import torch\n \n def recurse_directory(directory):\n sub_directories = [\n os.path.join(directory, d) for d in os.listdir(directory)\n if os.path.isdir(os.path.join(directory, d))\n and not any(omit in d for omit in args.omit)\n ]\n matching_files = [\n os.path.join(directory, f) for f in os.listdir(directory)\n if not os.path.isdir(os.path.join(directory, f))\n and f.startswith(args.name_prefix)\n ]\n \n for sub_directory in sub_directories:\n matching_files = matching_files + recurse_directory(sub_directory)\n \n return matching_files\n \n matching_files = []\n for directory in args.directory:\n matching_files = matching_files + recurse_directory(directory)\n \n bundled_files = {}\n for matching_file in matching_files:\n d, f = os.path.split(matching_file)\n if d not in bundled_files:\n bundled_files[d] = []\n bundled_files[d].append(matching_file)\n \n most_recent_files = []\n for directory, files in bundled_files.items():\n file_times = [os.path.getmtime(f) for f in files]\n most_recent_file = max(zip(file_times, files))[1]\n most_recent_files.append(most_recent_file)\n \n plot_logfiles(\n most_recent_files,\n args.format,\n args.keys,\n height=args.height,\n width=args.width,\n x_coord=args.x_coord,\n x_range=args.x_range,\n )\n","repo_name":"aaronwalsman/conspiracy","sub_path":"conspiracy/commandline.py","file_name":"commandline.py","file_ext":"py","file_size_in_byte":4650,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"78"}
+{"seq_id":"14546591310","text":"import numpy as np\nimport cv2\nimport math\nimport os\nimport argparse\n\nimg1 = cv2.imread('../HW3Pics/1.jpg')\nimg2 = cv2.imread('../HW3Pics/2.jpg')\nimg1_coord = cv2.imread('../HW3WorldCoords/1.coords.jpg')\nimg2_coord = cv2.imread('../HW3WorldCoords/2.coords.jpg')\n\n\ndef Homography(pt, pt_prime):\n\t# Finding Homography A of Ax=b\n A = np.zeros((8,8))\n b = np.zeros((1,8))\n\n for i in range(0, len(pt)):\n A[i * 2] = [pt[i][0], pt[i][1], pt[i][2], 0, 0, 0,\\\n (-1 * pt[i][0] * pt_prime[i][0]), (-1 * pt[i][1] * pt_prime[i][0])]\n A[i * 2 + 1] = [0,0,0,pt[i][0],pt[i][1],pt[i][2],\\\n (-1 * pt[i][0] * pt_prime[i][1]), (-1 * pt[i][1] * pt_prime[i][1])]\n b[0][i * 2] = pt_prime[i][0]\n b[0][i * 2 + 1] = pt_prime[i][1]\n\n h = np.matmul(np.linalg.pinv(A),b.T)\n homography = np.zeros((3,3))\n homography[0] = h[0:3,0]\n homography[1] = h[3:6,0]\n homography[2][0:2] = h[6:8,0]\n homography[2][2] = 1\n\n return homography\n\n\ndef Homography_affine(P,Q,R,S):\n '''\n Finding Homography using two orthogonal lines in the world plane\n H: 3 by 3 Homography matrix\n H_p: Homography that removes projective\n P: points coord\n C*x = B\n\n '''\n # P = np.append(P.T,1)\n\n # get othogonal lines\n l1 = np.cross(P, Q)\n m1 = np.cross(Q, R)\n l2 = np.cross(R, P)\n m2 = np.cross(S, Q)\n\n # \n C = np.array([[l1[0]*m1[0], l1[0]*m1[1] + l1[1]*m1[0]], [l2[0]*m2[0], l2[0]*m2[1] + l2[1]*m2[0]]])\n # print('C:', C)\n B = np.array([[-l1[1]*m1[1]],[-l2[1]*m2[1]]])\n # print('B:', B)\n x = np.dot(np.linalg.inv(C), B)\n # print('x:', x)\n S = np.array([[x[0][0], x[1][0]],[x[1][0], 1]])\n # print('S:', S)\n u, d, v = np.linalg.svd(S)\n # print('u:', u)\n # print('d:', d)\n # print('v:', v)\n\n A = np.dot(np.dot(v,np.sqrt(d)),v.T)\n\n H = np.eye(3)\n H[0][0:2] = A[0]\n H[1][0:2] = A[1]\n print(H)\n\n return H\n\n\n\ndef Homography_vanish(P,Q,S,R):\n '''\n Finding Homography using two orthogonal lines in the world plane\n H: 3 by 3 Homography matrix\n H: Homography H_v\n P: points coordinates (x_i,y_i)\n l1 = P x Q\n l2 = Q x S\n l3 = S x R\n l4 = R x P\n\n '''\n H = np.eye(3)\n\n l1 = np.cross(P, Q)\n l2 = np.cross(Q, S)\n l3 = np.cross(S, R) \n l4 = np.cross(R, P)\n\n vp1 = np.cross(l1,l3)\n vp2 = np.cross(l2,l4)\n vl = np.cross(vp1,vp2)\n # rescale\n vl = vl/np.max(vl)\n print('vanish line:', vl)\n H[2][:] = vl[:]\n print('H_vanish', H)\n\n return H\n\n\ndef Homography_1step(P0):\n '''\n input P0: 20 points in image forming 5 pairs of orthogonal lines in the world plane, 11 by 2\n P: 11 by 3\n output H: 3 by 3 Homography matrix\n C*x = B\n '''\n # print('P0',P0)\n P = np.zeros((P0.shape[0],3))\n for i in range(P0.shape[0]):\n P[i] = np.append(P0[i],1)\n # print(P)\n\n # find 5 pairs of orthogonal lines\n l_group = np.zeros((5,3))\n m_group = np.zeros((5,3))\n C = np.zeros((5,5))\n B = np.zeros((5,1))\n\n for i in range(l_group.shape[0]):\n l_group[i] = np.cross(P[4*i], P[4*i+1])\n l_group[i] = l_group[i]/np.max(l_group[i])\n m_group[i] = np.cross(P[4*i+2], P[4*i+3])\n m_group[i] = m_group[i]/np.max(m_group[i])\n B[i] = - l_group[i][2] * m_group[i][2]\n C[i] = np.array([l_group[i][0] * m_group[i][0], \\\n (l_group[i][0] * m_group[i][1] + l_group[i][1] * m_group[i][0])/2, \\\n l_group[i][1] * m_group[i][1], \\\n (l_group[i][0] * m_group[i][2] + l_group[i][2] * m_group[i][0])/2, \\\n (l_group[i][1] * m_group[i][2] + l_group[i][2] * m_group[i][1])/2])\n \n # print('l_group:', l_group)\n print('C:', C)\n print('B:', B)\n x = np.dot(np.linalg.inv(C), B)\n print('x:', x)\n x = x/np.max(x)\n print('x:', x)\n\n S = np.array([[x[0][0], x[1][0]/2],[x[1][0]/2, x[2][0]]])\n # print('S:', S) # 2 by 2\n U, D, V = np.linalg.svd(S)\n D = np.diag(D)\n print('U:', U) # 2 by 2\n print('D:', D) # 2 by 2\n print('V:', V) # 2 by 2\n\n A = np.dot(np.dot(V,np.sqrt(D)),V.T)\n print('A:', A)\n v = np.dot(np.array([[x[3][0]/2,x[4][0]/2]]), np.linalg.inv(A.T))\n print('v:', v)\n\n H = np.eye(3)\n H[0][0:2] = A[0]\n H[1][0:2] = A[1]\n H[0][2] = v[0][0]\n H[1][2] = v[0][1]\n # print('H:', H)\n\n return H\n\n\n# P = np.array([1,1,1]).T\n# Q = np.array([1,2,1]).T\n# R = np.array([1,3,1]).T\n# S = np.array([2,1,1]).T\n# Homography_vanish(P,Q,S,R)\n# Homography_affine(P,Q,R,S)\n# P0 = np.random.rand(20,2)\n# Homography_1step(P0)\n\ndef load_point_for_one_step(args):\n if args.img_name == '1':\n # P0 = np.array([[x,y],[x,y]]) # 20 by 2\n P0 = np.array([[1011,1857],[957,2064],[1011,1857],[1269,1863],\\\n [1011,1857],[1269,1863],[1269,1863],[1218,2076],\\\n [1269,1863],[1218,2076],[1218,2076],[957,2064],\\\n [1218,2076],[957,2064],[957,2064],[1011,1857],\\\n [1512,1665],[1440,2085],[1512,1665],[1944,2592]])\n elif args.img_name == '2':\n pass\n elif args.img_name == '3':\n pass\n elif args.img_name == '4':\n pass\n else:\n raise Exception('Error: No such image!')\n return P0\n \n\ndef one_step0(img, args):\n print('img size:',img.shape) # img size: (1944, 2592, 3)\n P0 = load_point_for_one_step(args)\n # P0 = np.random.randint(20,size=(20,2))\n H = Homography_1step(P0)\n print('H:', H)\n # H_inv * X_i = X_w\n H_inv = np.linalg.inv(H)\n\n temp = np.zeros((img.shape[0],img.shape[1],3), dtype='uint8')\n image_target = img\n\n for i in range(0,(img.shape[0]-1)):\n for j in range(0,(img.shape[1]-1)):\n point_tmp = np.array([i, j, 1])\n trans_coord = np.array(np.dot(H,point_tmp))\n trans_coord = trans_coord/trans_coord[2]\n if (trans_coord[0] > 0) and (trans_coord[0] < image_target.shape[0]) and \\\n (trans_coord[1] > 0) and (trans_coord[1] < image_target.shape[1]):\n temp[i][j] = image_target[math.floor(trans_coord[0]),math.floor(trans_coord[1])] \n # return temp\n cv2.imwrite('../HW3Pics/onestep1.jpg',temp)\n\ndef one_step(img, args):\n print('img size:',img.shape) # img size: (1944, 2592, 3)\n P0 = load_point_for_one_step(args)\n # P0 = np.random.randint(20,size=(20,2))\n H = Homography_1step(P0)\n print('H:', H)\n # H_inv * X_i = X_w\n H_inv = np.linalg.inv(H)\n box_i = np.array([[1,1,1],[img.shape[1],1,1],[img.shape[0],1,1],[img.shape[1],img.shape[0],1]])\n # print(box_i.shape) # 4 by 3\n box_w = np.zeros((4,3))\n for i in range(box_i.shape[0]):\n box_w[i] = np.dot(H_inv,box_i[i])\n box_w[i] = box_w[i]//box_w[i][2]\n print('box_w:', box_w)\n xymin = np.min(box_w, axis=0)\n print('xymin:', xymin)\n xmin, ymin = xymin[0], xymin[1] \n xymax = np.max(box_w, axis=0)\n xmax, ymax = xymax[0], xymax[1] \n width = xmax - xmin\n height = ymax - ymin\n print('before scale', width, height) # 14513.0 4729.0 world plane\n scale_x = img.shape[1]/width\n scale_y = img.shape[0]/height\n print()\n width_ = math.floor(scale_x*width)\n height_ = math.floor(scale_y*height)\n print('after scale', width_, height_)\n\n img_out = np.zeros((height_, width_, 3))\n for h in range(height_):\n for w in range(width_):\n # scale back\n print('debug:', w/scale_x+xmin-1, h/scale_y+ymin-1)\n tmp = np.array([[w/scale_x+xmin-1],[h/scale_y+ymin-1],[1]])\n print(tmp)\n H_b = np.dot(H, tmp)\n print('H_b: ',H_b)\n x_i, y_i = int(H_b[0][0]//H_b[2][0]), int(H_b[1][0]//H_b[2][0])\n print(x_i, y_i)\n img_out[h][w] = img[x_i][y_i]\n\n # write output\n # img_out = np.zeros((height_, width_, 3))\n # img_out = np.zeros((height_, width_, 3))\n # for h in range(height_):\n # for w in range(width_):\n # # scale back\n # tmp = np.array([[w/scale_x+xmin-1],[h/scale_y+ymin-1],[1]])\n # print(tmp)\n # H_b = np.dot(H, tmp)\n # print('H_b: ',H_b)\n # x_i, y_i = int(H_b[0][0]//H_b[2][0]), int(H_b[1][0]//H_b[2][0])\n # print(x_i, y_i)\n # img_out[h][w] = img[x_i][y_i]\n cv2.imwrite('../HW3Pics/onestep1.jpg',img_out)\n\n\ndef load_point_for_two_step(args):\n if args.img_name == '1':\n # world coordinate\n P_w, Q_w, S_w, R_w = np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1])\n X_w = [P_w, Q_w, S_w, R_w]\n # point to point \n P_i, Q_i, S_i, R_i = np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1])\n X_i = [P_i, Q_i, S_i, R_i]\n # X_i = np.array([P_i, Q_i, S_i, R_i])\n # affine\n P_i2, Q_i2, S_i2, R_i2 = np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1])\n X_i2 = [P_i2, Q_i2, S_i2, R_i2]\n # vl\n P_v, Q_v, S_v, R_v = np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1]), np.array([1,1,1])\n\n if args.two_step_method == 'p2p':\n H_p = Homography(X_w, X_i)\n elif args.two_step_method == 'vl':\n H_p = Homography_vanish(P_v, Q_v, S_v, R_v)\n else:\n raise Exception('Error: No such two step method!')\n\n elif args.img_name == '2':\n pass\n elif args.img_name == '3':\n pass\n elif args.img_name == '4':\n pass\n else:\n raise Exception('Error: No such image!')\n\n return H_p, X_i, X_i2\n\n\ndef two_step(img, args):\n H_p, X_i, X_i2 = load_point_for_two_step(args)\n\n H_p_inv = np.linalg.inv(H_p)\n\n box_i = np.array([[1,1,1],[img.shape[1],1,1],[img.shape[0],1,1],[img.shape[1],img.shape[0],1]])\n # print(box_i.shape) # 4 by 3\n box_h = np.zeros((4,3))\n for i in range(box_i.shape[0]):\n box_h[i] = np.dot(H_p,box_i[i])\n box_h[i] = box_h[i]//box_h[i][2]\n print('box_h:', box_h)\n\n xymin_h = np.min(box_h, axis=0)\n xmin_h, ymin_h = xymin_h[0], xymin_h[1] \n # xymax = np.max(box_w, axis=0)\n # xmax, ymax = xymax[0], xymax[1] \n # width = xmax - xmin\n # height = ymax - ymin\n # print('before scale', width, height)\n # scale_x = img.shape[1]/width\n # scale_y = img.shape[0]/height\n # width_ = math.floor(scale_x*width)\n # height_ = math.floor(scale_y*height)\n # print('after scale', width_, height_)\n\n # removing affine: x_h = H_p*x-xmin\n X_0 = []\n for pt in X_i2:\n tmp = np.dot(H_p,pt) # 3 by 1\n pt_ = np.array([tmp[0][0]//tmp[2][0]-xmin_h, tmp[1][0]//tmp[2][0]-ymin_h])\n X_0.append(pt_)\n H_a = Homography_affine(X_0)\n H_a_inv = np.linalg.inv(H_a)\n\n H = H_a_inv * H_p\n # H = H_p\n H_inv = np.linalg.inv(H)\n X_i = np.array(X_i) # 4 by 3\n W = np.dot(H, X_i.T) # 3 by 3 times 3 by 4 is ---- 3 by 4\n\n xymin = np.min(W, axis=0)\n # print(xymin)\n xmin, ymin = xymin[0], xymin[1] \n xymax = np.max(W, axis=0)\n xmax, ymax = xymax[0], xymax[1] \n width_ = xmax - xmin\n height_ = ymax - ymin\n print('w & h', width, height)\n\n # write output\n img_out = np.zeros((height_, width_, 3))\n for h in range(height_):\n for w in range(width_):\n # scale back\n tmp = np.array([[w+xmin-1],[h+ymin-1],[1]]) # 3 x 1\n print(tmp)\n H_b = np.dot(H_inv, tmp) # 3 x 1\n x_i, y_i = H_b[0][0]//H_b[2][0], H_b[1][0]//H_b[2][0]\n print(x_i, y_i)\n img_out[h][w] = img[x_i][y_i]\n cv2.imwrite('../HW3Pics/twostep1.jpg',img_out)\n\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('--img_name', type=str, default=None) # 1, 2, 3, 4\n parser.add_argument('--two_step_method', type=str, default='p2p') # p2p or vl\n args = parser.parse_args()\n one_step(img1, args)\n # two_step(img1, args)\n\n\n\nif __name__ == '__main__':\n main()\n # print()\n\n","repo_name":"KOPFYF/ECE661-Computer-Vision","sub_path":"HW3/scripts/hw3task1.py","file_name":"hw3task1.py","file_ext":"py","file_size_in_byte":11929,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"8193820125","text":"import sys\n\ndef init():\n N, M = map(int, sys.stdin.readline().split())\n dogam = []\n poketmonDict = dict()\n for i in range(N):\n item = sys.stdin.readline().rstrip()\n dogam.append(item)\n poketmonDict[item] = i+1\n\n for i in range(M):\n item = sys.stdin.readline().rstrip()\n try:\n item = int(item)\n sys.stdout.write('{}\\n' .format(dogam[item-1]))\n \n except :\n sys.stdout.write('{}\\n' .format(poketmonDict[item]))\n \n \ninit()\n","repo_name":"parkjineon/BaekJoon-Solution","sub_path":"단계별로 풀어보기/집합과 맵/1620/나는야 포켓몬 마스터 이다솜.py","file_name":"나는야 포켓몬 마스터 이다솜.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"11015119025","text":"import praw\nimport os\nimport csv\nfrom twilio.rest import Client\n\n\nclass DealCollector:\n\n def __init__(self, keywords=[], twilio_set=False):\n self.subreddit = 'hardwareswap'\n self.keywords = keywords\n self.twilio_set = twilio_set\n self.client = self.get_twilio_client()\n self.post_dict = {}\n self.logged_in = False\n self.reddit = self.bot_login()\n\n def bot_login(self):\n try:\n print(\"logging in: \", os.getenv(\"reddit_username\"))\n reddit = praw.Reddit(client_id=os.getenv(\"CLIENT_ID\"),\n client_secret=os.getenv(\"CLIENT_SECRET\"),\n user_agent='hardware swap deal collector',\n username=os.getenv(\"REDDIT_USERNAME\"),\n password=os.getenv(\"REDDIT_PASSWORD\"))\n print(\"logged in!\")\n self.logged_in = True\n except Exception as e:\n print(e)\n print(\"Login failed\")\n self.logged_in = False\n return reddit\n\n def append_to_csv(self, post_id):\n url = self.post_dict[post_id]['url']\n title = self.post_dict[post_id]['title']\n with open('deals.csv', 'a') as csvfile:\n writer = csv.writer(csvfile)\n writer.writerow([post_id, url, title])\n return True\n\n def get_twilio_client(self):\n if self.twilio_set:\n twilio_sid = os.getenv('TWILIO_SID')\n twilio_auth = os.getenv('TWILIO_AUTH')\n client = Client(twilio_sid, twilio_auth)\n return client\n else:\n print('no twilio account info provided')\n\n def send_text(self, collected_urls):\n if self.twilio_set:\n twilio_number = os.getenv('TWILIO_NUMBER')\n my_number = os.getenv('MY_NUMBER')\n client = self.client\n urls = '\\n\\n'.join(collected_urls)\n message = client.messages.create(from_=twilio_number,\n body=urls,\n to=my_number)\n else:\n pass\n\n def collect_posts(self, post_list):\n posts = 0\n collected_urls = []\n for post in post_list:\n post_id = str(post.id)\n if post_id not in self.post_dict:\n url = str(post.url)\n title = str(post.title)\n self.post_dict[post_id] = {}\n self.post_dict[post_id]['url'] = url\n self.post_dict[post_id]['title'] = title\n # add to csv\n self.append_to_csv(post_id)\n print('post of interest:', url)\n posts += 1\n collected_urls.append(url)\n if posts > 0:\n print(posts, 'new post(s) appended to deals.csv')\n self.send_text(collected_urls)\n return True\n\n def check_for_deals(self):\n if len(self.keywords) == 0:\n return 'empty keywords list'\n r = self.reddit\n s = self.subreddit\n keywords = self.keywords\n post_list = []\n for post in r.subreddit(s).new():\n title = str(post.title)\n for key in keywords:\n if key.lower() in title.lower():\n post_list.append(post)\n self.collect_posts(post_list)\n return True\n","repo_name":"Ecalzo/hardwareSwapDealCollector","sub_path":"deal_collector.py","file_name":"deal_collector.py","file_ext":"py","file_size_in_byte":3382,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39450119099","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom pandas.io.parsers import read_csv\nimport scipy.optimize as opt\n\ndata = read_csv('p2/ex2data1.csv', header=None).values.astype(float)\nx = data[:,:2]\ny = data[:,2]\nsize = len(y)\npositive = plt.scatter(x[np.where(y == 1), 0], x[np.where(y == 1), 1], marker='+', c='k')\nnegative = plt.scatter(x[np.where(y == 0), 0], x[np.where(y == 0), 1], marker='o', c='b')\nplt.xlabel(\"Exam 1 score\")\nplt.ylabel(\"Exam 2 score\")\nplt.legend((positive, negative),('Admitted', 'Not admitted'))\nplt.show()\n\nX = np.ones((size,3))\nX[:,1:] = x\n\ntheta = np.zeros((3, 1))\n\ndef sigmoid(x):\n return 1.0 / (1.0 + np.exp(-x))\n\ndef cost_function(theta, x, y, m):\n J = (-np.log(sigmoid(x.dot(theta)).T).dot(y) - np.log(1 - sigmoid(x.dot(theta)).T).dot(1 - y))/m\n return J\n\ndef gradient_function(theta, x, y, m):\n h = sigmoid(x.dot(theta)).reshape(-1, 1)\n y = y.reshape(m, 1)\n gradient = x.T.dot(h - y)/m\n return gradient\n\nprint(\"Initial cost = \" + str(cost_function(theta, X, y, size)))\nprint(\"Initial gradient = \" + str(gradient_function(theta, X, y, size)))\n\nresult = opt.fmin_tnc(func=cost_function , x0=theta , fprime=gradient_function , args=(X, y, size))\ntheta_opt = result[0]\nprint(\"Optimal cost = \" + str(cost_function(theta_opt, X, y, size)))\ntheta_opt = theta_opt.reshape((3,1))\n\nlinspace = np.linspace(30, 100, 1000)\nboundary = -(theta_opt[0] + theta_opt[1]*linspace)/theta_opt[2]\npositive = plt.scatter(x[np.where(y == 1), 0], x[np.where(y == 1), 1], marker='+', c='k')\nnegative = plt.scatter(x[np.where(y == 0), 0], x[np.where(y == 0), 1], marker='o', c='b')\nplt.plot(linspace, boundary)\nplt.xlabel(\"Exam 1 score\")\nplt.ylabel(\"Exam 2 score\")\nplt.legend((positive, negative),('Admitted', 'Not admitted'))\nplt.show()\n\ndef accuracy(theta, x, y, m):\n predictions = sigmoid(x.dot(theta))\n predictions_corrected = [1 if pred >= 0.5 else 0 for pred in predictions]\n number = np.sum(predictions_corrected == y)\n return (float(number)/m)*100\n\nprint(\"Accuracy = \" + str(accuracy(theta_opt, X, y, size)) +\"%\")\n","repo_name":"JuliaMaria/Madrid","sub_path":"Complutense/LogisticRegression.py","file_name":"LogisticRegression.py","file_ext":"py","file_size_in_byte":2067,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"40317810884","text":"\nclass Solution:\n def isPalindrome(self, s: str) -> bool:\n s = \"\".join([string.lower() for string in s if string.isalnum()])\n if s == s[::-1]:\n return True\n else:\n return False\n\n\n\nsol = Solution()\nsol.isPalindrome(\"A man, a plan, a canal: Panama\")\n\n","repo_name":"hansololee/algorithm","sub_path":"leetcode/Easy/ValidPalindrome/ValidPalindrome.py","file_name":"ValidPalindrome.py","file_ext":"py","file_size_in_byte":295,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"71279764733","text":"import argparse\nimport jinja2\nimport jinja2.exceptions\nimport glob\nimport os\nimport sys\nimport yaml\n\n\ndef main():\n parser = argparse.ArgumentParser(description='Check device templates')\n parser.add_argument(\"--device-types\", required=True, type=str,\n help=\"Path to the directory containing the device-type jinja2 templates.\")\n parser.add_argument(\"--devices\", required=True, type=str,\n help=\"Path to directory containing the device dictionary files.\")\n args = parser.parse_args()\n\n if not os.path.isdir(args.devices):\n sys.stderr.write(\"--devices argument must be a directory\\n\")\n return 1\n if not os.path.isdir(args.device_types):\n sys.stderr.write(\"--device-types argument must be a directory\\n\")\n return 1\n\n errors = False\n devices = sorted(glob.glob(\"%s/*.jinja2\" % args.devices))\n\n print(\"Devices:\")\n env = jinja2.Environment(loader=jinja2.FileSystemLoader([args.devices, args.device_types]))\n\n for device in devices:\n device_name = os.path.splitext(os.path.basename(device))[0]\n\n try:\n template = env.get_template(\"%s.jinja2\" % device_name)\n device_template = template.render()\n except jinja2.exceptions.TemplateNotFound as exc:\n print('* %s (ERROR): \"%s\" not found' % (device_name, exc))\n errors = True\n except jinja2.exceptions.TemplateRuntimeError as exc:\n print('* %s (ERROR): redering error \"%s\"' % (device_name, exc))\n errors = True\n except jinja2.exceptions.TemplateSyntaxError as exc:\n print('* %s (ERROR): invalid syntax \"%s\" in \"%s\"' % (device_name, exc, exc.filename))\n errors = True\n else:\n print(\"* %s\" % device_name)\n\n return errors\n\n\nif __name__ == '__main__':\n sys.exit(main())\n","repo_name":"Linaro/lava-server","sub_path":"share/check-devices.py","file_name":"check-devices.py","file_ext":"py","file_size_in_byte":1854,"program_lang":"python","lang":"en","doc_type":"code","stars":32,"dataset":"github-code","pt":"78"}
+{"seq_id":"17780872392","text":"import os\nimport json\nimport sys\nimport luigi\nimport z5py\nfrom cluster_tools.downscaling import DownscalingWorkflow\n\n\ndef preprocess():\n path = './data.n5'\n in_key = 'labels_for_subdivision_mip/s0'\n root_key = 'labels_for_subdivision_mip'\n\n n_scales = 4\n scale_factors = n_scales * [[2, 2, 2]]\n halos = scale_factors\n\n tmp_folder = './tmp_subdivision_labels/tmp2'\n config_dir = os.path.join(tmp_folder, 'configs')\n os.makedirs(config_dir, exist_ok=True)\n\n conf = DownscalingWorkflow.get_config()['downscaling']\n conf.update({'library_kwargs': {'order': 0}})\n with open(os.path.join(config_dir, 'downscaling.config'), 'w') as f:\n json.dump(conf, f)\n\n target = 'local'\n max_jobs = 16\n\n task = DownscalingWorkflow(tmp_folder=tmp_folder, config_dir=config_dir,\n target=target, max_jobs=max_jobs,\n input_path=path, input_key=in_key,\n scale_factors=scale_factors, halos=halos,\n output_path=path, output_key_prefix=root_key)\n luigi.build([task], local_scheduler=True)\n\n\ndef find_my_block(block_id):\n from heimdall import view, to_source\n\n scale = 5\n rpath = '../../../data/rawdata/sbem-6dpf-1-whole-raw.n5'\n k = 'setup0/timepoint0/s%i' % scale\n\n f = z5py.File(rpath)\n ds = f[k]\n ds.n_thread = 8\n raw = ds[:]\n\n path = './data.n5'\n k = 'volumes/clustering'\n f = z5py.File(path)\n ds = f[k]\n ds.n_threads = 8\n block = ds[:].astype('uint32')\n if block_id is not None:\n block = (block == block_id).astype('uint32')\n\n view(to_source(raw, name='raw'),\n to_source(block, name='block-volume'))\n\n\nif __name__ == '__main__':\n # preprocess()\n args = sys.argv\n block_id = int(args[1]) if len(args) == 2 else None\n find_my_block(block_id)\n","repo_name":"mobie/platybrowser-project","sub_path":"segmentation/correction/find_my_block.py","file_name":"find_my_block.py","file_ext":"py","file_size_in_byte":1867,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"78"}
+{"seq_id":"18330565950","text":"import tensorflow as tf\nfrom tensorflow.contrib.layers import batch_norm\nfrom tensorflow.contrib import rnn\nfrom tensorflow.python.ops.nn_ops import leaky_relu\n\nfrom utils.network_summary import count_parameters\n\n\nclass VGGClassifier:\n def __init__(self, batch_size, layer_stage_sizes, name, num_classes, num_channels=1, batch_norm_use=False,\n inner_layer_depth=2, strided_dim_reduction=True):\n\n \"\"\"\n Initializes a VGG Classifier architecture\n :param batch_size: The size of the data batch\n :param layer_stage_sizes: A list containing the filters for each layer stage, where layer stage is a series of\n convolutional layers with stride=1 and no max pooling followed by a dimensionality reducing stage which is\n either a convolution with stride=1 followed by max pooling or a convolution with stride=2\n (i.e. strided convolution). So if we pass a list [64, 128, 256] it means that if we have inner_layer_depth=2\n then stage 0 will have 2 layers with stride=1 and filter size=64 and another dimensionality reducing convolution\n with either stride=1 and max pooling or stride=2 to dimensionality reduce. Similarly for the other stages.\n :param name: Name of the network\n :param num_classes: Number of classes we will need to classify\n :param num_channels: Number of channels of our image data.\n :param batch_norm_use: Whether to use batch norm between layers or not.\n :param inner_layer_depth: The amount of extra layers on top of the dimensionality reducing stage to have per\n layer stage.\n :param strided_dim_reduction: Whether to use strided convolutions instead of max pooling.\n \"\"\"\n self.reuse = False\n self.batch_size = batch_size\n self.num_channels = num_channels\n self.layer_stage_sizes = layer_stage_sizes\n self.name = name\n self.num_classes = num_classes\n self.batch_norm_use = batch_norm_use\n self.inner_layer_depth = inner_layer_depth\n self.strided_dim_reduction = strided_dim_reduction\n self.build_completed = False\n\n def __call__(self, text_input, training=False, dropout_rate=0.0):\n \"\"\"\n Runs the CNN producing the predictions and the gradients.\n :param text_input: Text input to produce embeddings for. e.g. for text data [batch_size, 300]\n :param training: A flag indicating training or evaluation\n :param dropout_rate: A tf placeholder of type tf.float32 indicating the amount of dropout applied\n :return: Embeddings of size [batch_size, self.num_classes]\n \"\"\"\n\n with tf.variable_scope(self.name, reuse=self.reuse):\n layer_features = []\n with tf.variable_scope('VGGNet'):\n outputs = image_input\n for i in range(len(self.layer_stage_sizes)):\n with tf.variable_scope('conv_stage_{}'.format(i)):\n for j in range(self.inner_layer_depth):\n with tf.variable_scope('conv_{}_{}'.format(i, j)):\n if (j == self.inner_layer_depth-1) and self.strided_dim_reduction:\n stride = 2\n else:\n stride = 1\n outputs = tf.layers.conv2d(outputs, self.layer_stage_sizes[i], [3, 3],\n strides=(stride, stride),\n padding='SAME', activation=None)\n outputs = leaky_relu(outputs, name=\"leaky_relu{}\".format(i))\n layer_features.append(outputs)\n if self.batch_norm_use:\n outputs = batch_norm(outputs, decay=0.99, scale=True,\n center=True, is_training=training, renorm=False)\n if self.strided_dim_reduction==False:\n outputs = tf.layers.max_pooling2d(outputs, pool_size=(2, 2), strides=2)\n\n outputs = tf.layers.dropout(outputs, rate=dropout_rate, training=training)\n # apply dropout only at dimensionality\n # reducing steps, i.e. the last layer in\n # every group\n\n c_conv_encoder = outputs\n c_conv_encoder = tf.contrib.layers.flatten(c_conv_encoder)\n c_conv_encoder = tf.layers.dense(c_conv_encoder, units=self.num_classes)\n\n self.reuse = True\n self.variables = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope=self.name)\n\n if not self.build_completed:\n self.build_completed = True\n count_parameters(self.variables, \"VGGNet\")\n\n return c_conv_encoder, layer_features\n\n\nclass TextClassifier:\n def __init__(self, batch_size, filter_sizes, name, num_classes, embeddings, max_sent_length, vocab_size,num_units, num_channels=1,\n embedding_dim=300, num_filters=100, l2_norm=3, activation='relu'):\n\n \"\"\"\n Initializes a VGG Classifier architecture\n :param batch_size: The size of the data batch\n :param filter_sizes: A list containing the filters sizes for the convolutional layer\n :param name: Name of the network\n :param num_classes: Number of classes we will need to classify\n :param num_channels: Number of channels of our image data.\n :param embeddings: the pretrained embeddings\n :param embed_size: the size of each embeddings\n :param num_filters: number of filter per size\n \"\"\"\n self.reuse = False\n self.batch_size = batch_size\n self.num_channels = num_channels\n self.filter_sizes = filter_sizes\n self.name = name\n self.num_classes = num_classes\n self.build_completed = False\n self.embeddings = embeddings\n self.embedding_dim = embedding_dim\n self.num_filters = num_filters\n self.max_sent_length = max_sent_length\n self.vocab_size = vocab_size\n self.l2_norm = l2_norm\n self.activation=activation\n self.num_units = num_units\n \n def __call__(self, text_input, training=False, dropout_rate=0.0):\n \"\"\"\n Runs the CNN producing the predictions and the gradients.\n :param text_input: Text input to produce embeddings for. e.g. for text data [batch_size, sequence_length]\n :param training: A flag indicating training or evaluation\n :param dropout_rate: A tf placeholder of type tf.float32 indicating the amount of dropout applied\n :return: Embeddings of size [batch_size, self.num_classes]\n \"\"\"\n\n with tf.variable_scope(self.name, reuse=self.reuse):\n layer_features = []\n with tf.variable_scope('VGGNet'):\n if self.embeddings==None:\n with tf.device('/cpu:0'), tf.name_scope(\"embedding\"):\n W = tf.Variable(\n tf.random_uniform([self.vocab_size, self.embedding_dim], -1.0, 1.0),\n name=\"W\") \n else:\n W = self.embeddings\n\n embedded_chars = tf.nn.embedding_lookup(W, text_input)\n inputs = embedded_chars_expanded = tf.expand_dims(embedded_chars, -1)\n pooled_outputs = []\n for i, filter_size in enumerate(self.filter_sizes):\n with tf.name_scope(\"conv-maxpool-%s\" % filter_size):\n # Convolution Layer\n filter_shape = [filter_size, self.embedding_dim, 1, self.num_filters]\n W = tf.Variable(tf.truncated_normal(filter_shape, stddev=0.1), name=\"W\")\n b = tf.Variable(tf.constant(0.1, shape=[self.num_filters]), name=\"b\")\n conv = tf.nn.conv2d(\n inputs,\n W,\n strides=[1, 1, 1, 1],\n padding=\"VALID\",\n name=\"conv\")\n # Apply nonlinearity\n if self.activation=='relu':\n h = tf.nn.relu(tf.nn.bias_add(conv, b), name=\"relu\")\n elif self.activation=='sigmoid':\n h = tf.sigmoid(tf.nn.bias_add(conv, b), name=\"sigmoid\")\n elif self.activation=='tanh':\n h = tf.tanh(tf.nn.bias_add(conv, b), name=\"tanh\")\n layer_features.append(h)\n # Maxpooling over the outputs\n pooled = tf.nn.max_pool(\n h,\n ksize=[1, self.max_sent_length - filter_size + 1, 1, 1],\n strides=[1, 1, 1, 1],\n padding='VALID',\n name=\"pool\")\n pooled_outputs.append(pooled)\n\n # Combine all the pooled features\n num_filters_total = self.num_filters * len(self.filter_sizes)\n h_pool = tf.concat(pooled_outputs, 3)\n h_pool_flat = tf.reshape(h_pool, [-1, num_filters_total])\n\n if self.l2_norm != 0:\n h_pool_flat = self.l2_norm * tf.divide(h_pool_flat, tf.norm(h_pool_flat, ord='euclidean'))\n # Add dropout\n with tf.name_scope(\"dropout\"):\n h_drop = tf.layers.dropout(h_pool_flat, rate=dropout_rate, training=training)\n\n c_conv_encoder = h_drop\n c_conv_encoder = tf.contrib.layers.flatten(c_conv_encoder)\n\n dense = tf.layers.dense(inputs=c_conv_encoder, units=self.num_units, activation=tf.nn.relu)\n\n # Logits Layer\n scores = tf.layers.dense(inputs=dense, units=self.num_classes)\n\n \n #self.reuse = True\n self.variables = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope=self.name)\n\n if not self.build_completed:\n self.build_completed = True\n count_parameters(self.variables, \"VGGNet\")\n\n return scores, layer_features\n\n\n\nclass RNNClassifier:\n def __init__(self, batch_size, name, num_classes, embeddings, max_sent_length, vocab_size, cell, num_units,hidden_unit=100,\n embedding_dim=300):\n\n \"\"\"\n Initializes a VGG Classifier architecture\n :param batch_size: The size of the data batch\n :param filter_sizes: A list containing the filters sizes for the convolutional layer\n :param name: Name of the network\n :param num_classes: Number of classes we will need to classify\n :param num_channels: Number of channels of our image data.\n :param embeddings: the pretrained embeddings\n :param embed_size: the size of each embeddings\n :param num_filters: number of filter per size\n \"\"\"\n self.reuse = False\n self.batch_size = batch_size\n self.name = name\n self.num_classes = num_classes\n self.build_completed = False\n self.embeddings = embeddings\n self.embedding_dim = embedding_dim\n self.max_sent_length = max_sent_length\n self.vocab_size = vocab_size\n self.hidden_unit = hidden_unit\n self.cell = cell\n self.num_units= num_units\n\n def __call__(self, text_input, training=False, dropout_rate=0.0):\n\n with tf.variable_scope(self.name, reuse=self.reuse):\n layer_features = []\n with tf.variable_scope('VGGNet'):\n if self.embeddings==None:\n with tf.name_scope(\"embedding\"):\n W = tf.Variable(\n tf.random_uniform([self.vocab_size, self.embedding_dim], -1.0, 1.0),\n name=\"W\") \n else:\n W = self.embeddings\n\n embedded_chars = tf.nn.embedding_lookup(W, text_input)\n if self.cell=='bidlstm':\n lstm_fw_cell=rnn.BasicLSTMCell(self.hidden_unit) #forward direction cell\n lstm_bw_cell=rnn.BasicLSTMCell(self.hidden_unit) #backward direction cell\n if dropout_rate is not None:\n lstm_fw_cell=rnn.DropoutWrapper(lstm_fw_cell,output_keep_prob=1-dropout_rate)\n lstm_bw_cell=rnn.DropoutWrapper(lstm_bw_cell,output_keep_prob=1-dropout_rate)\n outputs,_=tf.nn.bidirectional_dynamic_rnn(lstm_fw_cell,lstm_bw_cell,embedded_chars,dtype=tf.float32) \n output_rnn=tf.concat(outputs,axis=2) #[batch_size,sequence_length,hidden_size*2]\n output_rnn_last=tf.reduce_mean(output_rnn,axis=1) #[batch_size,hidden_size*2] #output_rnn_last=output_rnn[:,-1,:] ##[batch_size,hidden_size*2] #TODO\n elif self.cell=='lstm':\n lstm_cell=rnn.BasicLSTMCell(self.hidden_unit)\n if dropout_rate is not None:\n lstm_cell=rnn.DropoutWrapper(lstm_cell,output_keep_prob=1-dropout_rate)\n outputs,_ = tf.nn.dynamic_rnn(lstm_cell, embedded_chars, dtype=tf.float32)\n output_rnn=tf.concat(outputs,axis=2)\n output_rnn_last=tf.reduce_mean(output_rnn,axis=1)\n elif self.cell=='gru':\n gru_cell=rnn.GRUCell(self.hidden_unit)\n if dropout_rate is not None:\n gru_cell=rnn.DropoutWrapper(gru_cell,output_keep_prob=1-dropout_rate)\n outputs,_ = tf.nn.dynamic_rnn(gru_cell, embedded_chars, dtype=tf.float32)\n output_rnn=tf.concat(outputs,axis=2)\n output_rnn_last=tf.reduce_mean(output_rnn,axis=1)\n\n c_conv_encoder = output_rnn_last\n dense = tf.layers.dense(inputs=c_conv_encoder, units=self.num_units, activation=tf.nn.relu)\n\n # Logits Layer\n scores = tf.layers.dense(inputs=dense, units=self.num_classes, activation=tf.sigmoid)\n \n #self.reuse = True\n self.variables = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope=self.name)\n\n if not self.build_completed:\n self.build_completed = True\n count_parameters(self.variables, \"VGGNet\")\n\n return scores, layer_features\n","repo_name":"cristeaadrian/mlp-cw","sub_path":"network_architectures.py","file_name":"network_architectures.py","file_ext":"py","file_size_in_byte":14721,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"72004664251","text":"from django.contrib import admin\nfrom cart import models\n\n@admin.register(models.Cart)\nclass CartAdmin(admin.ModelAdmin):\n list_display = ['cart_id', 'get_product_total', 'paid_for', 'created_on']\n search_fields = ['product__name']\n\n\n@admin.register(models.CustomerOrder)\nclass CustomerOrderAdmin(admin.ModelAdmin):\n list_display = ['reference', 'transaction', 'payment', 'accepted', 'shipped', 'refund', 'created_on']\n search_fields = ['reference', 'transaction']\n date_hierarchy = 'created_on'\n sortable_by = ['payment', 'created_on']\n filter_horizontal = ['cart']\n list_per_page = 20\n actions = ['mark_accepted', 'mark_shipped']\n\n def mark_accepted(self, queryset):\n queryset.update(accepted=True)\n return queryset\n \n def mark_shipped(self, queryset):\n queryset.update(shipped=True)\n return queryset\n\n\n@admin.register(models.Shipment)\nclass ShipmentAdmin(admin.ModelAdmin):\n list_display = ['customer_order']\n search_fields = ['customer_code']\n date_hierarchy = 'created_on'\n","repo_name":"Zadigo/ecommerce_template","sub_path":"mywebsite/cart/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"78"}
+{"seq_id":"14154155276","text":"#!/usr/bin/python3\n\"\"\" Base class \"\"\"\n\nimport json\nfrom models.review import Review\nfrom models.city import City\nfrom models.place import Place\nfrom models.amenity import Amenity\nfrom models.state import State\nfrom models.base_model import BaseModel\nfrom models.user import User\n\n\nclass FileStorage:\n \"\"\" Serializes JSON and deserializes \"\"\"\n __file_path = 'file.json'\n __objects = {}\n clases = {\n \"BaseModel\": BaseModel,\n \"User\": User,\n \"State\": State,\n \"City\": City,\n \"Amenity\": Amenity,\n \"Place\": Place,\n \"Review\": Review\n }\n\n def all(self):\n \"\"\" Return dictionary \"\"\"\n return FileStorage.__objects\n\n def new(self, obj):\n \"\"\" New __objects .id \"\"\"\n key = \"{}.{}\".format(obj.__class__.__name__, obj.id)\n FileStorage.__objects[key] = obj\n\n def save(self):\n \"\"\" Serialize JSON file __objects \"\"\"\n dict_save = {}\n for key, value in FileStorage.__objects.items():\n valor_dict = value.to_dict()\n dict_save[key] = valor_dict\n\n with open(FileStorage.__file_path, 'w') as archivo:\n json.dump(dict_save, archivo)\n\n def reload(self):\n \"\"\" Deserialize JSON file __objects \"\"\"\n try:\n with open(FileStorage.__file_path, 'r') as archivo:\n dicts = json.load(archivo)\n FileStorage.__objects = {}\n for keys, values in dicts.items():\n obj = FileStorage.clases[values['__class__']](**values)\n FileStorage.__objects[keys] = obj\n except:\n return\n\n def find_object(cls, id=''):\n \"\"\" Return object id \"\"\"\n objs = cls.__objects\n for obj in objs.values():\n if obj.id == id:\n return obj\n print(\"No encuentro la ID :::: {} ::::\".format(id))\n","repo_name":"SrDri/AirBnB_clone","sub_path":"models/engine/file_storage.py","file_name":"file_storage.py","file_ext":"py","file_size_in_byte":1887,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"43572856008","text":"# -*- coding: utf-8 -*-\n\nfrom qgis.PyQt.QtGui import QIcon\nfrom qgis.PyQt.QtCore import QDir, QFileInfo\nfrom qgis.core import (QgsProcessingParameterExtent,\n QgsProcessingParameterEnum,\n QgsProcessingParameterString,\n QgsProcessingParameterBoolean,\n QgsProcessingParameterRasterLayer,\n QgsProcessingAlgorithm,\n QgsProcessingParameterFolderDestination,\n QgsProject,\n QgsCoordinateTransform,\n QgsCoordinateReferenceSystem,\n QgsRectangle,\n QgsLayerTreeLayer,\n QgsRasterLayer)\nimport os.path\nimport math, re\nfrom .tilingthread import TilingThread\nfrom .tileset import TileSet\n\nclass GetTilesProcessingAlgorithm(QgsProcessingAlgorithm):\n\n MINZOOM = 'MINZOOM'\n MAXZOOM = 'MAXZOOM'\n EXTENT = 'EXTENT'\n SINGLELAYER = 'SINGLELAYER'\n INPUT = 'INPUT'\n SUBSET = 'SUBSET'\n OUTPUT = 'OUTPUT'\n\n def initAlgorithm(self, config=None):\n self.zoomlist = ['z0', 'z1', 'z2', 'z3', 'z4', 'z5', 'z6', 'z7', 'z8', 'z9', 'z10', 'z11', 'z12', 'z13', 'z14', 'z15', 'z16', 'z17', 'z18', 'z19', 'z20', 'z21', 'z22', 'z23', 'z24']\n self.addParameter(QgsProcessingParameterEnum(self.MINZOOM, 'Min zoom of cached map', self.zoomlist, defaultValue=13))\n self.addParameter(QgsProcessingParameterEnum(self.MAXZOOM, 'Max zoom of cached map', self.zoomlist, defaultValue=15))\n self.addParameter(QgsProcessingParameterExtent(self.EXTENT, 'Cache extent'))\n self.addParameter(QgsProcessingParameterBoolean(self.SINGLELAYER, 'Cache single raster layer.', defaultValue=True, optional=False))\n self.addParameter(QgsProcessingParameterRasterLayer(self.INPUT, 'Cached layer:', optional=True))\n self.addParameter(QgsProcessingParameterString(self.SUBSET, 'Cache folder name', optional=False, defaultValue='Cache'))\n self.addParameter(QgsProcessingParameterFolderDestination(self.OUTPUT, 'Folder to store map tiles (by default - home folder of you project):'))\n self.workThread = None\n\n def processAlgorithm(self, parameters, context, feedback):\n\n crs = context.project().crs()\n minzoom = self.parameterAsEnum(parameters, self.MINZOOM, context)\n maxzoom = self.parameterAsEnum(parameters, self.MAXZOOM, context)\n if minzoom > maxzoom:\n feedback.pushConsoleInfo('Maximum zoom value is lower than minimum. Please correct this and try again.')\n return\n \n bbox = self.parameterAsExtent(parameters, self.EXTENT, context, crs)\n outfolder = self.parameterAsFile(parameters, self.OUTPUT, context)\n root_dir = self.parameterAsString(parameters, self.SUBSET, context)\n cached_layer = self.parameterAsRasterLayer(parameters, self.INPUT, context)\n single_layer = self.parameterAsBoolean(parameters, self.SINGLELAYER, context)\n\n fileInfo = QFileInfo(outfolder)\n if fileInfo.isDir() and not len(QDir(outfolder).entryList(QDir.Dirs | QDir.Files | QDir.NoDotAndDotDot)) == 0:\n feedback.pushConsoleInfo('Selected directory is not empty.')\n #return\n feedback.pushConsoleInfo(f\"fileInfo:{fileInfo}.\")\n\n extent = QgsCoordinateTransform(crs, QgsCoordinateReferenceSystem('EPSG:4326'), QgsProject.instance()).transform(bbox)\n arctanSinhPi = math.degrees(math.atan(math.sinh(math.pi)))\n extent = extent.intersect(QgsRectangle(-180, -arctanSinhPi, 180, arctanSinhPi))\n\n prj_file = QgsProject.instance().fileName()\n root = QgsProject.instance().layerTreeRoot()\n layers = root.checkedLayers()\n tile_width = 256\n tile_height = 256\n transp = 100\n quality = 70\n tile_format = 'PNG'\n enable_antialiasing = False\n tmsconvention = False\n writeMapurl = False\n writeViewer = False\n metatile = False\n metatile_size = {'rows': 2, 'cols': 2}\n metatile_buffer = False\n llg_features = False\n if single_layer:\n layers = [cached_layer]\n\n self.workThread = TilingThread( layers,\n extent,\n minzoom,\n maxzoom,\n tile_width,\n tile_height,\n transp,\n quality,\n tile_format,\n fileInfo,\n root_dir,\n enable_antialiasing,\n tmsconvention,\n writeMapurl,\n writeViewer,\n metatile,\n metatile_size,\n metatile_buffer,\n llg_features\n )\n \n self.workThread.rangeChanged.connect(self.setProgressRange)\n self.workThread.updateProgress.connect(self.updateProgress)\n self.workThread.processFinished.connect(self.processFinished)\n self.workThread.processInterrupted.connect(self.processInterrupted)\n self.workThread.start()\n\n # this does the same thing but without using a seperate thread. Good for debugging.\n #no_thread = TileSet(layers,extent,minzoom,maxzoom,tile_width,tile_height,\n # transp,quality,tile_format,fileInfo,root_dir,\n # enable_antialiasing,tmsconvention,writeMapurl,writeViewer,metatile,metatile_size,\n # metatile_buffer,llg_features)\n #no_thread.run()\n \n #create a xyz layer\n full_path = re.sub('\\\\\\\\','/',outfolder) + '/' + root_dir\n urlWithParams = 'type=xyz&url=file:///%(f)s/{z}/{x}/{y}.%(t)s&zmax=15&zmin=13&crs=EPSG3857'%{'f':full_path,'t':tile_format}\n rlayer = QgsRasterLayer(urlWithParams, f'{root_dir}', 'wms')\n #rlayer.isValid()\n\n #create a groop for xyz layers\n root = QgsProject.instance().layerTreeRoot()\n group = root.findGroup('Local tiles')\n QgsProject.instance().addMapLayer(rlayer)\n group.insertChildNode(0, QgsLayerTreeLayer(rlayer))\n\n return {self.OUTPUT: [outfolder, full_path, prj_file]}\n\n def uncheckLayers(self, layers):\n for lay in layers:\n node = QgsProject.instance().layerTreeRoot().findLayer(lay)\n if node:\n node.setItemVisibilityChecked(False)\n\n def checkLayers(self, layers):\n for lay in layers:\n node = QgsProject.instance().layerTreeRoot().findLayer(lay)\n if node:\n node.setItemVisibilityChecked(True)\n \n def setProgressRange(self, message, value):\n self.progressBar.setFormat(message)\n self.progressBar.setRange(0, value)\n\n def updateProgress(self):\n self.progressBar.setValue(self.progressBar.value() + 1)\n\n def processInterrupted(self):\n self.stopProcessing()\n\n def processFinished(self):\n self.stopProcessing()\n\n def stopProcessing(self):\n if self.workThread is not None:\n self.workThread.stop()\n self.workThread = None\n\n def name(self):\n return 'Get tiles'\n\n def icon(self):\n return QIcon(os.path.dirname(__file__) + '/gettiles.png')\n\n def displayName(self):\n return self.name()\n\n def group(self):\n return self.groupId()\n\n def groupId(self):\n return ''\n\n def createInstance(self):\n return GetTilesProcessingAlgorithm()","repo_name":"olegruk/gettiles","sub_path":"get_tiles_processing_algorithm.py","file_name":"get_tiles_processing_algorithm.py","file_ext":"py","file_size_in_byte":7978,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"29249620204","text":"from nltk.corpus import gutenberg\nimport random\nimport nltk\nimport pandas as pd\nfrom nltk.tokenize import RegexpTokenizer\nfrom nltk.tokenize import sent_tokenize, word_tokenize\nfrom nltk.corpus import stopwords\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.utils import shuffle\nfrom sklearn.cross_validation import train_test_split\nfrom sklearn import svm\n\nbook_1 = gutenberg.raw('bible-kjv.txt').lower()\nbook_2 = gutenberg.raw('melville-moby_dick.txt').lower()\nbook_3 = gutenberg.raw('edgeworth-parents.txt').lower()\ndict_labels = {'KJV (Bible)': book_1, 'Herman Melville (Moby-Dick)' : book_2, 'Richard Lovell Edgeworth (Parents)' : book_3}\nmaster_list = []\nsize_of_books = []\n# Dividing each book into documents of 30 sentences. appending them to *master_list*. \nfor author, book in dict_labels.items():\n n = 0;\n book = nltk.sent_tokenize(book)\n limit = int(len(book)/31)\n size_of_books.append(limit)\n for i in range(limit):\n master_list.append(\" \".join(book[n:n+30]))\n n+=31 \n \n#Seperating the 3 books.\ncomplete_book_1 = (master_list[0:size_of_books[0]-1])\ncomplete_book_2 = (master_list[961:1278]) \ncomplete_book_3 = (master_list[1278:-1])\n\n#Sampling 200 documents from the entire book.\nsample_book1 = pd.DataFrame(random.sample(complete_book_1, 200))\nsample_book2 = pd.DataFrame(random.sample(complete_book_2, 200))\nsample_book3 = pd.DataFrame(random.sample(complete_book_3, 200))\n\n#Converting Labels to integers\nkjv_bible = 0\nmelville_moby_dick = 1\nedgeworth_parents = 2\n\n# Adding a new column to the sample books containing lables. \nsample_book1['Label'] = kjv_bible\nsample_book2['Label'] = melville_moby_dick\nsample_book3['Label'] = edgeworth_parents\n\nframes = [sample_book1, sample_book2, sample_book3]\ncomplete_sample = pd.concat(frames)\ncomplete_sample = shuffle(complete_sample)\n\nx_train, x_test, y_train, y_test = train_test_split(complete_sample[0], complete_sample['Label'], test_size=0.2, random_state=4)\n\n\n#BOW \ntoken = RegexpTokenizer(r'[a-zA-Z]+')\ncv_bow = CountVectorizer(lowercase=True,stop_words='english',ngram_range = (1,1), analyzer='word', tokenizer = token.tokenize)\nfiltered_counts_bow= cv_bow.fit_transform(x_train)\nfil_test = cv_bow.fit_transform(x_test)\nbow = filtered_counts_bow.toarray()\n\nclf = svm.SVC(kernel = \"linear\")\nclf.fit(filtered_counts_bow, y_train)\n\n\n#TFIDF\ntoken = RegexpTokenizer(r'[a-zA-Z]+')\ncv_tfidf = TfidfVectorizer(lowercase=True,stop_words='english',ngram_range = (1,1), analyzer='word', tokenizer = token.tokenize)\nfiltered_counts_tfidf= cv_tfidf.fit_transform(x_train)\ntfidf = filtered_counts_tfidf.toarray()\n","repo_name":"salauddinaliahmed/EBC_7100-Assignment_1","sub_path":"SVM_working.py","file_name":"SVM_working.py","file_ext":"py","file_size_in_byte":2693,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"72634702332","text":"import math\n\nimport HandTrackingModule as htm\nimport cv2\nimport numpy as np\nimport time\nimport osascript\n\n\n##############################'\nwCam,hCam = 640,480\n\ncap = cv2.VideoCapture(0)\ncap.set(3,wCam)\ncap.set(4,hCam)\npTime = 0\n\ndetector = htm.handDetector(detectionCon=0.7)\n\n# vol = \"set volume output volume \" + str(-1)\n# osascript.osascript(vol)\n# result = osascript.osascript('get volume settings')\n# print(result)\n\nwhile True:\n success,img = cap.read()\n img = detector.findHands(img)\n lmList = detector.findPosition(img,draw=False)\n if len(lmList)!= 0:\n # print(lmList[4],lmList[8])\n\n x1,y1 = lmList[4][1],lmList[4][2]\n x2, y2 = lmList[8][1], lmList[8][2]\n cx,cy = (x1+x2)//2, (y1+y2)//2\n x3, y3 = lmList[12][1], lmList[12][2] # Middle Finger\n # print(x3,y3)\n\n cv2.circle(img, (x1,y1), 10, (255,0,0), cv2.FILLED)\n cv2.circle(img, (x2, y2), 10, (255, 0, 0), cv2.FILLED)\n cv2.line(img,(x1,y1),(x2,y2),(255,0,255),2)\n cv2.circle(img, (cx, cy), 10, (255, 0, 0), cv2.FILLED)\n\n length = math.hypot(x2-x1,y2-y1)\n # print(length)\n\n vol = np.interp(length,[25,220],[0,100])\n print(round(length), round(vol))\n vol1 = \"set volume output volume \" + str(vol)\n osascript.osascript(str(vol1))\n\n\n if length < 50:\n cv2.circle(img, (cx, cy), 10, (0, 255, 0), cv2.FILLED)\n\n\n cTime = time.time()\n fps = 1/(cTime-pTime)\n pTime = cTime\n\n\n cv2.waitKey(1)\n\n cv2.putText(img,f'FPS: {int(fps)}',(40,50),cv2.FONT_HERSHEY_COMPLEX,1,(255,0,255),2)\n cv2.imshow(\"Img\", img)\n\n","repo_name":"ThejakaSEP/HandTracking","sub_path":"VolumeController.py","file_name":"VolumeController.py","file_ext":"py","file_size_in_byte":1609,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"1579214965","text":"\"\"\"\nAuthor: Zhou Chen\nDate: 2020/3/17\nDesc: desc\n\"\"\"\nfrom django.shortcuts import render\nimport re\nimport mimetypes\nfrom wsgiref.util import FileWrapper\nfrom django.http import StreamingHttpResponse\nimport os\nimport uuid\nfrom .settings import MEDIA_ROOT, STATIC_ROOT\nimport sys\nsys.path.append(\"../\")\nimport yolo3_deepsort\nimport utils.format_factory as ff\n\n\ndef upload(request):\n if request.method == 'POST':\n files = request.FILES['video']\n if len(files) > 0:\n if not os.path.exists(MEDIA_ROOT):\n # 若不存在媒体存储目录\n os.mkdir(MEDIA_ROOT)\n video = files\n extension = os.path.splitext(video.name)[1]\n # 重命名文件\n file_name = '{}{}'.format(uuid.uuid4(), extension)\n file_path = '{}/{}'.format(MEDIA_ROOT, file_name)\n # 保存文件到本机\n with open(file_path, 'wb') as f:\n for c in video.chunks():\n f.write(c)\n # 视频保存本机之后调用模型\n\n args = yolo3_deepsort.Argument(file_path)\n args.output_path = os.path.join(STATIC_ROOT, 'videos', 'rst.avi')\n cfg = yolo3_deepsort.get_config()\n cfg.merge_from_file(args.config_detection)\n cfg.merge_from_file(args.config_deepsort)\n with yolo3_deepsort.VideoTracker(cfg, args, file_path) as vdo_trk:\n vdo_trk.run_with_limit(300)\n os.remove(os.path.join(STATIC_ROOT, 'videos', 'rst.mp4'))\n ff.avi2mp4(args.output_path, os.path.join(STATIC_ROOT, 'videos', 'rst.mp4'))\n\n return render(request, 'show_video.html', {'filename': 'rst.mp4'})\n else:\n return render(request, 'upload.html')\n return render(request, 'upload.html')\n\n\ndef file_iterator(file_name, chunk_size=8192, offset=0, length=None):\n with open(file_name, \"rb\") as f:\n f.seek(offset, os.SEEK_SET)\n remaining = length\n while True:\n bytes_length = chunk_size if remaining is None else min(remaining, chunk_size)\n data = f.read(bytes_length)\n if not data:\n break\n if remaining:\n remaining -= len(data)\n yield data\n\n\ndef stream_video(request):\n path = request.GET.get('path')\n path = os.path.join(\"static\", \"videos\", path)\n range_header = request.META.get('HTTP_RANGE', '').strip()\n range_re = re.compile(r'bytes\\s*=\\s*(\\d+)\\s*-\\s*(\\d*)', re.I)\n range_match = range_re.match(range_header)\n size = os.path.getsize(path)\n content_type, encoding = mimetypes.guess_type(path)\n content_type = content_type or 'application/octet-stream'\n if range_match:\n first_byte, last_byte = range_match.groups()\n first_byte = int(first_byte) if first_byte else 0\n last_byte = first_byte + 1024 * 1024 * 8 # 8M 每片,响应体最大体积\n if last_byte >= size:\n last_byte = size - 1\n length = last_byte - first_byte + 1\n resp = StreamingHttpResponse(file_iterator(path, offset=first_byte, length=length), status=206,\n content_type=content_type)\n resp['Content-Length'] = str(length)\n resp['Content-Range'] = 'bytes %s-%s/%s' % (first_byte, last_byte, size)\n else:\n # 不是以视频流方式的获取时,以生成器方式返回整个文件,节省内存\n resp = StreamingHttpResponse(FileWrapper(open(path, 'rb')), content_type=content_type)\n resp['Content-Length'] = str(size)\n resp['Accept-Ranges'] = 'bytes'\n return resp","repo_name":"luanshiyinyang/DeepSORT","sub_path":"web/web/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":3623,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"78"}
+{"seq_id":"17645508897","text":"# 거스름돈\n\nimport sys\n\nmoney = int(sys.stdin.readline())\ncount = 0\n# 거스름돈 계산\nchange = 1000 - money\n\n# 가장 큰 단위부터 확인\ncoin = [500, 100, 50, 10, 5, 1]\n\nfor i in coin:\n # 동전 개수 계산\n count = count + change // i\n # 남은 거스름돈\n change = change % i\n # 더 이상 거스를 돈이 없을 경우 중지\n if change == 0:\n break\n\nprint(count)","repo_name":"Techeer-3rd-gen-study/Algorithm-study","sub_path":"03주차_10.18_10.24/2_5585/이수현_5585.py","file_name":"이수현_5585.py","file_ext":"py","file_size_in_byte":409,"program_lang":"python","lang":"ko","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"29282839018","text":"# A. Задана натуральная степень k.\n# Сформировать случайным образом список коэффициентов (значения от 0 до 100)\n# многочлена и записать в файл многочлен степени k.\n# Пример:\n# если k = 2, то многочлены могут быть => 2*x² + 4*x + 5 = 0 или x² + 5 = 0 или 10*x² = 0\nfrom random import randint\n\n\ndef polyniminal(k: int) -> dict:\n poly_dict = {}\n for i in range(k, -1, -1):\n rnd = randint(-100, 100)\n poly_dict[i] = rnd\n\n return poly_dict\n\n\ndef polynom_dict_to_line(poly_dict: dict) -> str:\n k = max(poly_dict.keys())\n line = ''\n for key in poly_dict:\n if poly_dict[key] != 0:\n factor = str(abs(int(poly_dict[key])))\n if int(poly_dict[key]) < 0:\n sign = ' - '\n elif int(key) == k:\n sign = ''\n else:\n sign = ' + '\n if int(key) == 0:\n x = ''\n exponent = ''\n elif int(key) == 1:\n x = '*x'\n exponent = ''\n else:\n x = '*x**'\n exponent = str(key)\n line += sign + factor + x + exponent\n line += ' = 0'\n return line\n\n\nk = int(input('Введите натуральную степень k: '))\n\npolynom = polyniminal(k)\n\nline = polynom_dict_to_line(polynom)\n\nwith open('PIHomeWork04/polynom.txt', 'w') as data:\n data.writelines(line)\n","repo_name":"Mortimer08/PIHomeWork04","sub_path":"Task01.py","file_name":"Task01.py","file_ext":"py","file_size_in_byte":1551,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"7339059112","text":"import discord\nimport asyncio\nimport humanreadable as hr\nimport json\n\nclient = discord.Client()\nuid = 0\ntry:\n with open('userconfig.json', 'w') as f:\n cfg = json.load(f)\n channels = cfg[\"channels\"]\n\nexcept FileNotFoundError:\n channels = {}\n whitelist = {}\n\n@client.event\nasync def on_ready():\n global uid\n print('We have logged in as {0.user}'.format(client))\n uid = client.user.id\n activity = discord.Activity(name='silence | after whitelist blacklist', type=discord.ActivityType.listening)\n await client.change_presence(activity=activity)\n\n\n@client.event\nasync def on_message(message):\n nick = message.channel.guild.get_member(uid).display_name\n print(nick)\n\n if message.content.startswith(nick):\n command = message.content[len(nick):].split()\n if command[0] == 'after':\n time = sum(hr.Time(o).seconds for o in command[1:])\n channels[message.channel.id] = time\n await message.delete()\n if command[0] == 'delete':\n await message.delete()\n\n print(channels)\n print(message.channel.id)\n if message.channel.id in channels:\n print('test')\n await asyncio.sleep(channels[message.channel.id])\n await message.delete()\n\nwith open('token.txt') as f:\n client.run(f.read())\n","repo_name":"TheTimgor/hushhush","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1307,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"41007311681","text":"# -*- coding: utf-8 -*-\n# author:CY\n# datetime:2019/8/2 23:15\nimport re\nimport string\n\nfrom wtforms.fields import simple\nfrom wtforms import validators\n\nfrom apps.chat.model import ChatRoom\nfrom c_chat_server.extensions import db\nfrom utils.serializer import BaseForm\nfrom utils.optional import Optional\n\n\nclass CreateChatRoomForm(BaseForm):\n \"\"\"创建聊天室form\"\"\"\n name=simple.StringField(\n label='聊天室名称',\n validators=[\n validators.DataRequired(message='聊天室名称必须填写!'),\n validators.length(max=10,message='聊天室名称应小于10字')\n ]\n )\n\n desc=simple.StringField(\n label='聊天室简介',\n validators=[\n validators.DataRequired(message='聊天室简介必须填写!'),\n validators.Length(max=512,message='聊天室简介应小于512字')\n ]\n )\n\n img=simple.StringField(\n label='聊天室封面图片',\n validators=[\n Optional(),\n ]\n )\n\n def validate_name(self,name):\n word_list = string.punctuation\n for i in word_list:\n if i in name.data:\n raise validators.StopValidation('聊天室名称不能带有特殊字符!')\n queryset=db.session.query(ChatRoom.id).filter(ChatRoom.name==name.data).first()\n if queryset:\n raise validators.StopValidation('聊天室名称已经被抢先注册了!')\n\n\n def validate_img(self,images):\n if images.data:\n if not re.match('http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\\(\\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+',images.data):\n raise validators.StopValidation('图片不符合规则!')","repo_name":"PrimaryCY/Cchat","sub_path":"c_chat_server/apps/chat/form.py","file_name":"form.py","file_ext":"py","file_size_in_byte":1691,"program_lang":"python","lang":"en","doc_type":"code","stars":30,"dataset":"github-code","pt":"78"}
+{"seq_id":"16015998493","text":"import time\nimport pandas as pd\nimport requests\nimport re\nimport json\nfrom xml.etree import ElementTree\nimport html\nfrom datetime import datetime\n\n\ndef get_video_list(youtube_playlist_url: str) -> list:\n \"\"\"\n 1. Load Gordan Ramsay playlist\n 2. Extract the Javascript variable that stores the playlist information\n 3. Store it in an object array\n :param youtube_playlist_url: YouTube playlist url. Eg: https://youtube.com/playlist?list=...\n :return: [vidId]\n \"\"\"\n resp: requests.Response = requests.get(youtube_playlist_url)\n\n # BROWSER DEBUG: console.log(ytInitialData)\n raw_html: str = resp.text\n result = re.search('var ytInitialData = (.*);<', raw_html)\n result_json = json.loads(result.group(1))\n\n vid_raw: list = result_json[\"contents\"][\"twoColumnBrowseResultsRenderer\"][\"tabs\"][0][\"tabRenderer\"][\"content\"][\"sectionListRenderer\"][\"contents\"][0][\"itemSectionRenderer\"][\"contents\"][0][\"playlistVideoListRenderer\"][\"contents\"]\n\n vid_list: list = [v[\"playlistVideoRenderer\"][\"videoId\"] for v in vid_raw if \"playlistVideoRenderer\" in v]\n\n return vid_list\n\n\ndef get_request_url(vid_id: str) -> dict:\n \"\"\"\n Retrieves the URL required to download the captions.\n This step is necessary because YouTube has implemented hashed their request in a form of signature.\n We will need the correct signature to download it.\n :param vid_id: Video ID\n :return: { title, description, author, keywords, caption_url }\n \"\"\"\n youtube_url: str = \"https://www.youtube.com/watch?v=\" + vid_id\n print(\"Scrapping\", youtube_url)\n\n resp: requests.Response = requests.get(youtube_url)\n raw_html = resp.text\n\n # Exclude members only videos\n if '\"label\":\"Members only\"' in raw_html or '\"label\":\"Ahli sahaja\"' in raw_html:\n return {}\n\n # Very specifically target this request url\n result = re.search('var ytInitialPlayerResponse = (.*);<', raw_html)\n\n result_json = json.loads(result.group(1))\n\n # Ensure language is in English\n if \"captions\" in result_json:\n caption_tracks: list = result_json[\"captions\"][\"playerCaptionsTracklistRenderer\"][\"captionTracks\"]\n\n tmp_url = \"\"\n for c in caption_tracks:\n if c[\"languageCode\"] == \"en\":\n tmp_url = c[\"baseUrl\"].replace(\"\\\\u0026\", \"&\")\n break\n\n # Prevent spam pattern\n time.sleep(1)\n\n return {\n \"title\": result_json[\"videoDetails\"][\"title\"],\n \"description\": result_json[\"videoDetails\"][\"shortDescription\"].replace(\"\\n\", \" \"),\n \"author\": result_json[\"videoDetails\"][\"author\"],\n \"keywords\": \",\".join(result_json[\"videoDetails\"][\"keywords\"]),\n \"caption_url\": tmp_url\n }\n else:\n return {}\n\n\ndef get_video_subtitle(request_url: str) -> str:\n \"\"\"\n The caption request URL will return an XML format file.\n Parse the XML and return the caption paragraph.\n :param request_url: eg. https://youtube.com/api...?v=....\n :return:\n \"\"\"\n resp: requests.Response = requests.get(request_url)\n\n tree = ElementTree.fromstring(resp.content)\n\n res: list = [html.unescape(t.text.replace(\"\\n\", \"\")) for t in tree if t.text is not None]\n\n return \" \".join(res)\n\n\nif __name__ == \"__main__\":\n\n while True:\n\n # Get playlist URL from user\n while True:\n print(\"YouTube playlist URL should look like: https://www.youtube.com/playlist?list=XXX\")\n playlist_link: str = input(\"Playlist URL: \")\n\n if \"https://www.youtube.com/playlist?list=\" in playlist_link:\n break\n\n playlist_video_list: list = get_video_list(playlist_link)\n video_info: list = [get_request_url(v) for v in playlist_video_list if len(v) > 0]\n\n for vi in video_info:\n if bool(vi):\n print(vi[\"title\"])\n vi[\"caption\"] = get_video_subtitle(vi[\"caption_url\"])\n\n tmp_now = datetime.now()\n filename: str = video_info[0][\"author\"] + \"-\" + tmp_now.strftime(\"%Y%m%d%H%M%S\")\n\n df = pd.DataFrame(video_info)\n df.drop(columns=[\"caption_url\"], inplace=True)\n df.to_csv(path_or_buf=\"\".join([\"./export/dataset-\", filename, \".csv\"]), index=False)\n\n raw_caption = df[\"caption\"].str.cat(sep=\" \")\n with open(\"./export/raw-\" + filename + \".txt\", \"w+\", encoding=\"utf-8\") as f:\n f.write(raw_caption)\n\n print(\"============================\")\n print(\"Scrapping complete\")\n print(\"\")\n","repo_name":"jonathanlawhh/word-art-from-youtube-captions","sub_path":"scrap.py","file_name":"scrap.py","file_ext":"py","file_size_in_byte":4500,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"70195078011","text":"#!/usr/bin/python\ntry:\n # Python3\n import urllib.request, urllib.error, urllib.parse\nexcept:\n import urllib2\n\ntry:\n # Python 2.6\n import json\nexcept:\n import simplejson as json\n\nimport os\nimport subprocess\nimport sys\n\n\ndef _run(filename):\n shell = os.name == 'nt'\n if filename.endswith('py'):\n executor = ['python', filename]\n return subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif filename.endswith('sh'):\n executor = ['bash', filename]\n return subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif filename.endswith('f90'):\n executor = ['gfortran', filename]\n exect = subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n output, output_err = exect.communicate()\n return_code = exect.wait()\n if return_code:\n print(ERROR % ('Compiling...', output_err))\n exit(return_code)\n return subprocess.Popen('./a.out', stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif filename.endswith('c'):\n executor = ['gcc', filename]\n exect = subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n output, output_err = exect.communicate()\n return_code = exect.wait()\n if return_code:\n print(ERROR % ('Compiling...', output_err))\n exit(return_code)\n return subprocess.Popen('./a.out', stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif filename.endswith('java'):\n executor = ['javac', filename]\n exect = subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n output, output_err = exect.communicate()\n return_code = exect.wait()\n if return_code:\n print(ERROR % ('Compiling...', output_err))\n exit(return_code)\n executor = ['java', '-Duser.language=en', filename[:-len('.java')]]\n return subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif filename.endswith('pl'):\n executor = ['swipl', '-q', '-f', filename]\n return subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True, shell=shell)\n elif os.access(filename, os.X_OK):\n executor = [os.path.abspath(filename)]\n return subprocess.Popen(executor, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) \n\n\ndef http_get(url):\n if sys.version_info[0] < 3:\n response = urllib2.urlopen(url).read().decode('utf-8')\n else:\n response = urllib.request.urlopen(url).read().decode('utf-8')\n return json.loads(response)\n\n\ndef http_post(url, data):\n if sys.version_info[0] < 3:\n response = urllib2.urlopen(url=url, data=data).read().decode('utf-8')\n else:\n response = urllib.request.urlopen(url=url, data=data.encode('utf-8')).read().decode('utf-8')\n return json.loads(response)\n\n\nBASE = 'http://dirlididi.com/api/'\n\nHEAD = \"\"\"= PROBLEM NAME\n%s\n= PROBLEM DESCRIPTION\n%s\n\"\"\"\nTEST = \"\"\"== TEST - %s\n%s\n== INPUT:\n%s\n== OUTPUT:\n%s\n\"\"\"\nERROR = \"\"\"==ERROR FOR INPUT:\n%s\n== ERROR MSG:\n%s\n\"\"\"\nFAILURE = \"\"\"== FAILED FOR INPUT:\n%s\n== FAILED OUTPUT:\n%s\n\"\"\"\n\n\ndef get_problem(key):\n return http_get(BASE + 'problem/' + key)\n\n\ndef submit_code(token, key, code, tests_result):\n result = {'tests': tests_result, 'key': key, 'code': code, 'token': token}\n data = json.dumps(result)\n url = BASE + 'code/' + key\n return http_post(url, data)\n \n\ndef has_failure(results):\n return results.replace('.', '')\n\n\ndef _get(key):\n problem = get_problem(key)\n print(HEAD % (problem[\"name\"], problem[\"description\"]))\n published_tests = [x for x in problem['tests'] if x['publish']]\n if published_tests:\n print(\"= PROGRAM EXAMPLES\")\n for test in published_tests:\n print(TEST % (test['description'], test.get('tip', ''), test['input'], test['output']))\n\n\ndef _submit(key, token, filename, source):\n problem = get_problem(key)\n tests_result = []\n for test in problem['tests']:\n exect = _run(filename)\n input_ = test['input']\n output, output_err = exect.communicate(input_)\n return_code = exect.wait()\n if return_code:\n print(ERROR % (input_, output_err))\n exit(return_code)\n tests_result.append((test['key'], output))\n content = submit_code(token, key, open(source).read(), tests_result)\n print(\"Results: \" + content['result'])\n if has_failure(content['result']):\n for result_i in range(len(content['result'])):\n if content['result'][result_i] != '.':\n failed_test = tests_result[result_i]\n print(FAILURE % (problem['tests'][result_i]['input'], failed_test[1]))\n\n\ndef _usage():\n print('Usage:\\n %s get \\n %s submit [filename_src]' % (sys.argv[0], sys.argv[0]))\n exit()\n\n\ndef main():\n if len(sys.argv) < 3:\n _usage()\n command = sys.argv[1]\n key = sys.argv[2]\n if command.lower() == 'get':\n _get(key)\n elif command.lower() == 'submit':\n if len(sys.argv) != 5 and len(sys.argv) != 6:\n _usage()\n token = sys.argv[3]\n filename = sys.argv[4]\n source = filename\n if len(sys.argv) == 6:\n source = sys.argv[5]\n _submit(key, token, filename, source)\n else:\n _usage()\n\n\nif __name__ == \"__main__\":\n main()","repo_name":"lucasmedeiros/praticasPLP","sub_path":"dirlididi.py","file_name":"dirlididi.py","file_ext":"py","file_size_in_byte":6022,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"15895053339","text":"#!/bin/python3\n\nimport sys\nfrom datetime import date\n\ndef switch(h) :\n return {\n 0 : \"Saturday\",\n 1 : \"Sunday\",\n 2 : \"Monday\",\n 3 : \"Tuesday\",\n 4 : \"Wednesday\",\n 5 : \"Thursday\",\n 6 : \"Friday\",\n }[h]\n \ndef Zellercongruence(day, month, year) :\n if (month == 1) :\n month = 13\n year = year - 1\n \n if (month == 2) :\n month = 14\n year = year - 1\n q = day\n m = month\n k = year % 100;\n j = year // 100;\n h = q + 13 * (m + 1) // 5 + k + k // 4 + j // 4 + 5 * j\n h = h % 7\n return h\n\ndef getNextDate(y, m, d):\n if m==12:\n return [y+1, 1, 1]\n else:\n return [y, m+1, 1]\n \ndef solve(ymd1, ymd2):\n y1, m1, d1 = ymd1\n y2, m2, d2 = ymd2\n \n sundays = 0\n \n if d1 == 1:\n if Zellercongruence(d1, m1, y1) == 1:\n sundays += 1\n \n y, m, d = y1, m1, d1\n while True:\n y, m, d = getNextDate(y, m, d)\n \n #check if date over than date2\n if y>y2:\n break\n elif y==y2:\n if m>m2:\n break\n elif m==m2:\n if d>d2:\n break\n \n if Zellercongruence(d, m, y) == 1:\n sundays += 1\n \n return sundays\n \nt = int(input().strip())\nfor a0 in range(t):\n y1, m1, d1 = [int(num) for num in input().strip().split(' ')]\n y2, m2, d2 = [int(num) for num in input().strip().split(' ')]\n print(solve([y1, m1, d1], [y2, m2, d2]))\n \n ","repo_name":"anandawira/HackerrankProjects","sub_path":"Euler/Project Euler #19: Counting Sundays.py","file_name":"Project Euler #19: Counting Sundays.py","file_ext":"py","file_size_in_byte":1525,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"35508561571","text":"import bpy\nfrom ... base_types.node import AnimationNode\n\nclass ObjectVisibilityInputNode(bpy.types.Node, AnimationNode):\n bl_idname = \"an_ObjectVisibilityInputNode\"\n bl_label = \"Object Visibility Input\"\n\n def create(self):\n self.inputs.new(\"an_ObjectSocket\", \"Object\", \"object\").defaultDrawType = \"PROPERTY_ONLY\"\n self.outputs.new(\"an_BooleanSocket\", \"Hide\", \"hide\")\n self.outputs.new(\"an_BooleanSocket\", \"Hide Select\", \"hideSelect\").hide = True\n self.outputs.new(\"an_BooleanSocket\", \"Hide Render\", \"hideRender\")\n self.outputs.new(\"an_BooleanSocket\", \"Show Name\", \"showName\").hide = True\n self.outputs.new(\"an_BooleanSocket\", \"Show Axis\", \"showAxis\").hide = True\n self.outputs.new(\"an_BooleanSocket\", \"Show Xray\", \"showXray\").hide = True\n\n def getExecutionCode(self):\n isLinked = self.getLinkedOutputsDict()\n if not any(isLinked.values()): return\n\n yield \"if object is not None:\"\n\n if isLinked[\"hide\"]: yield \" hide = object.hide\"\n if isLinked[\"hideSelect\"]: yield \" hideSelect = object.hide_select\"\n if isLinked[\"hideRender\"]: yield \" hideRender = object.hide_render\"\n if isLinked[\"showName\"]: yield \" showName = object.show_name\"\n if isLinked[\"showAxis\"]: yield \" showAxis = object.show_axis\"\n if isLinked[\"showXray\"]: yield \" showXray = object.show_x_ray\"\n\n yield \"else: hide = hideSelect = hideRender = showName = showAxis = showXray = None\"\n","repo_name":"chrisatbest/animation_nodes","sub_path":"nodes/object/object_visibility_input.py","file_name":"object_visibility_input.py","file_ext":"py","file_size_in_byte":1514,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"32900899710","text":"from utilities import set_label_topics, feature_engineering\nimport cPickle as pickle\n\n\ndef get_prediction(df, tfidf_vectorizer, model_NMF, list_columns, scaler, model):\n set_label_topics(df, col_name='Abstract',\n tfidf_vectorizer=tfidf_vectorizer,\n model_NMF=model_NMF)\n\n\n feature_engineering(df, model_NMF)\n test_df = df[list_columns]\n\n features = \\\n scaler.transform(test_df.values)\n\n return model.predict_proba(features)[:,1]\n\nif __name__ == '__main__':\n model = pickle.load(open('../data/model.pkl', 'rb'))\n tfidf_vectorizer = pickle.load(open('../data/tfidf_vectorizer.pkl', 'rb'))\n model_NMF = pickle.load(open('../data/NMF.pkl', 'rb'))\n scaler = pickle.load(open('../data/scaler.pkl', 'rb'))\n list_columns = pickle.load(open('../data/list_columns.pkl', 'rb'))\n\n df_2012_dod = pickle.load(open('../data/df_2012_dod.pkl', 'rb'))\n prediction = get_prediction(df_2012_dod,\n tfidf_vectorizer,\n model_NMF,\n list_columns,\n scaler,\n model)\n df_2012_dod['probability'] = prediction\n #df = subset_data('dod', 2012, '/Users/AnnaVMS/Desktop/test2')\n #df = pd.DataFrame(df.iloc[0:1])\n #df.pop('to_phase_II')\n #print get_prediction(df, tfidf_vectorizer, model_NMF, list_columns,\n #scaler, model)\n","repo_name":"AnnaVM/SBIR_Project","sub_path":"my_app/code_SBIR/get_prediction.py","file_name":"get_prediction.py","file_ext":"py","file_size_in_byte":1475,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74751597371","text":"from datetime import datetime, timedelta, timezone\nfrom celery import shared_task\nfrom users.models import User\n\n\n@shared_task\ndef check_inactive_users():\n \"\"\"\n Асинхронная задача для проверки неактивных пользователей и отключения их активных статусов, если они не входили\n в систему более 30 дней.\n \"\"\"\n\n # Получение всех пользоват��лей, исключая тех, у которых last_login равен None.\n users = User.objects.all().exclude(last_login__isnull=True)\n\n for user in users:\n # Проверяется, прошло ли более 30 дней с момента последнего входа пользователя в систему.\n if datetime.now(timezone.utc) - user.last_login > timedelta(days=30):\n # Устанавливается флаг is_active пользователя в False, что означает неактивный статус.\n user.is_active = False\n\n user.save()\n","repo_name":"Lisnevskiy/LearningManagementSystemDRF","sub_path":"users/tasks.py","file_name":"tasks.py","file_ext":"py","file_size_in_byte":1116,"program_lang":"python","lang":"ru","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"11078290847","text":"import os\nfrom lxml import etree\nimport numpy as np\nfrom dm_control.mujoco.wrapper.mjbindings import mjlib\nfrom dm_control.utils import io as resources\nfrom robot_env.utils import get_root_path\n\n\ndef make_pendulum(n_bodies, base_file=None, path=None):\n \"\"\"Generates an xml string defining a swimmer with `n_bodies` bodies.\"\"\"\n if path is None:\n path = os.path.join(get_root_path(), \"robots/basic_models/generated_pendulum{}.xml\".format(n_bodies))\n if base_file is None:\n base_file = os.path.join(get_root_path(), \"robots/basic_models/pendulum_base.xml\")\n if n_bodies <= 0:\n raise ValueError('Invalid number of bodies: {}'.format(n_bodies))\n mjcf = etree.fromstring(resources.GetResource(base_file))\n head_body = mjcf.find('./worldbody/body')\n actuator = etree.SubElement(mjcf, 'actuator')\n\n parent = head_body\n pos = 0\n lengths = np.random.normal(0.5, 0.1, size=n_bodies)\n lengths = np.clip(lengths, 0.1, 1.0)\n masses = np.random.uniform(0.5, 2.5, size=n_bodies)\n for body_index in range(n_bodies):\n pose = \"0 0 {}\".format(pos)\n pos = lengths[body_index]\n child = make_body(body_index=body_index, pose=pose, len=pos, mass=masses[body_index])\n # site\n child.append(etree.Element('site', {'name': \"pole{}_0\".format(body_index+1),\n 'pos': \"0 0 0\",\n \"size\": \"0.01 0.01\"}))\n child.append(etree.Element('site', {'name': \"pole{}_1\".format(body_index+1),\n 'pos': \"0 0 {}\".format(pos),\n \"size\": \"0.01 0.01\"}))\n if body_index == n_bodies - 1:\n child.append(etree.Element('site', {'name': \"tip\",\n 'pos': \"0 0 {}\".format(pos),\n \"size\": \"0.01 0.01\"}))\n\n # joint\n child.append(etree.Element('joint', {'name': \"pendulum{}\".format(body_index+1),\n 'axis': \"0 1 0\",\n \"pos\": \"0 0 0\",\n 'type': \"hinge\"}))\n\n parent.append(child)\n parent = child\n actuator.append(etree.Element('motor',\n name='pendulum{}_motor'.format(body_index+1),\n joint='pendulum{}'.format(body_index+1),\n forcerange=\"-50 50\",\n ctrlrange=\"-20 20\"))\n\n model = etree.tostring(mjcf, pretty_print=True)\n with open(path, 'wb') as f:\n f.write(model)\n return model\n\n\ndef make_body(body_index, pose='0 0 0', len=1., mass=1.):\n \"\"\"Generates an xml string defining a single physical body.\"\"\"\n body = etree.Element('body', name=\"pole{}\".format(body_index+1))\n body.set('pos', pose)\n etree.SubElement(body, 'geom', {'name': \"geom_b{}\".format(body_index+1),\n 'fromto': '0 0 0 0 0 {}'.format(len),\n 'rgba': \"0.588 0.909 0.972 1\",\n 'size': \"0.045 {}\".format(0.5 * len),\n 'type': \"capsule\",\n \"mass\": \"{}\".format(mass)\n })\n\n return body\n\n\nif __name__ == \"__main__\":\n x = make_pendulum(3)\n","repo_name":"scxxxxxx/VBMPC","sub_path":"robot_model/mujoco/basic_models/gen_model_pendulums.py","file_name":"gen_model_pendulums.py","file_ext":"py","file_size_in_byte":3464,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"16639368334","text":"\"\"\"\r\nRandom Number Guessing Game\r\n\"\"\"\r\nimport random\r\nfrom textwrap import dedent\r\n\r\ndef main():\r\n\r\n print(\"So I'm thinking of a number between 1 and 50\")\r\n\r\n # creates secret number\r\n secret_number = random.randrange(1, 50)\r\n\r\n # initalizes our attempt count\r\n user_attempt_count = 1\r\n\r\n # sets user guess to something the secret number can't be\r\n user_guess = 0\r\n\r\n # loops until user_guess == secret_number,\r\n # or user_attempt_count over allowed attempts\r\n\r\n while user_guess != secret_number and user_attempt_count < 10:\r\n print(\"This is your guess number \", user_attempt_count)\r\n \r\n try:\r\n user_input_text = input(\"Type any number between 1 and 50: \")\r\n user_guess = int(user_input_text)\r\n # tells user if too high or low, or got it\r\n if user_guess > secret_number:\r\n print(\"That's too high.\")\r\n elif user_guess < secret_number:\r\n print(\"That's too low.\")\r\n else:\r\n print(\"Yay! That's it!\")\r\n except ValueError:\r\n print(\"Type in an actual number\")\r\n\r\n\r\n # Increment the attempt count\r\n user_attempt_count += 1\r\n\r\n if user_guess != secret_number:\r\n print(dedent(\r\n \"\"\"\r\n Just so you're not wondering for the rest\r\n of your life what my number was, it was:\r\n \"\"\"), secret_number)\r\n\r\nif __name__ == \"__main__\":\r\n main()","repo_name":"Cate-Michalska/Horse-Rider-Game","sub_path":"chapter8_guessinggame.py","file_name":"chapter8_guessinggame.py","file_ext":"py","file_size_in_byte":1481,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33626747788","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Nov 24 20:41:02 2020\r\n\r\n@author: uha18\r\n\"\"\"\r\n\r\nlist1 = [input()]\r\nlist2 = [input()]\r\nfor i in list1:\r\n for j in list2:\r\n if i==j:\r\n print(i,j)\r\n continue\r\n ","repo_name":"uhapko/pyth1uni","sub_path":"instersection Hapko.py","file_name":"instersection Hapko.py","file_ext":"py","file_size_in_byte":240,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"8625123966","text":"from flask import Flask,request,jsonify\n\napp=Flask(__name__)\n\n@app.route('/')\ndef home():\n return 'Welcome to flask'\n\n@app.route('/cal',methods=['GET','POST'])\ndef math_operator():\n operatation=request.json['operation']\n number1=request.json['number1']\n number2=request.json['number2']\n \n if operatation=='add':\n result=number1+number2\n elif operatation=='sub':\n result=number1-number2\n elif operatation=='mul':\n result=number1*number2\n elif operatation=='div':\n result=number1/number2\n else:\n return jsonify('Invalid operartion type, it should be in (add,sub,mul,div)')\n\n return jsonify(f'result is: {result}')\n\n\nif __name__=='__main__':\n app.run(port=8080,debug=True)\n","repo_name":"ykudale2012/python_rest_api","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":741,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"25445589054","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jul 10 11:19:49 2021\n\n@author: Asadullo\n\"\"\"\n\nclass mashina ():\n def __delattr__(self):\n print(\"malumot o'chirildi\")\n def __init__(self ,model,narx,otkuch):#class msthods(konsturcter)\n self .a=model\n self.b=narx\n self.c=otkuch\n def printmashina(self):\n print(\"\"\"\n Mashina turi: {}\n Mashina naxi: {}\n Mashina kuchi: {}\n \"\"\".format(self.a,self.b,self.c ))\nm1=mashina('Mazda',100,2000)\nm1.printmashina()\ndel(m1)\nm1=mashina(\"lamborjini\",100,3000)\n","repo_name":"Asadullo-aka/python_darslar","sub_path":"python_9-dars/class2.py","file_name":"class2.py","file_ext":"py","file_size_in_byte":560,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"70694648572","text":"import numpy as np\nimport torch\n#from get_data.data_loader2 import data_loader2\nfrom get_data.data_loader_transform import data_loader_transform\n#from get_data.DataLoader_Cox import DataLoader_Cox\nfrom go_models.train import train\nfrom go_models.evaluate import evaluate\nfrom go_models.get_cnn_model import get_cnn_model\nfrom go_models.get_cox_model import get_cox_model\n\n\nif __name__ == '__main__':\n\n data_dir = '/mnt/aertslab/DATA/HeadNeck/HN_PETSEG/curated'\n proj_dir = '/mnt/aertslab/USERS/Zezhong/HN_OUTCOME'\n cnn_name = 'resnet'\n model_depth = 101 # [10, 18, 34, 50, 101, 152, 200]\n n_classes = 20\n in_channels = 1\n batch_size = 8\n epochs = 1\n lr = 0.001\n num_durations = 20\n _cox_model = 'LogisticHazard'\n cox_model = 'LogisticHazard'\n load_model = 'model'\n score_type = '3year_survival' #'median'\n evaluate_only = False\n augmentation = True\n \n np.random.seed(1234)\n _ = torch.manual_seed(1234)\n\n if not augmentation:\n dl_train, dl_tune, dl_val = data_loader2(\n proj_dir=proj_dir,\n batch_size=batch_size,\n _cox_model=_cox_model,\n num_durations=num_durations)\n else:\n dl_train, dl_tune, dl_val, dl_test = data_loader_transform(\n proj_dir, \n batch_size=batch_size, \n _cox_model=_cox_model, \n num_durations=num_durations)\n \n for cnn_name in ['resnet18', 'resnet34', 'resnet50', 'resnet152', 'resnet200']: \n cnn_model = get_cnn_model(\n cnn_name=cnn_name, \n n_classes=n_classes, \n in_channels=in_channels)\n cox_model = get_cox_model(\n proj_dir=proj_dir,\n cnn_model=cnn_model,\n _cox_model=_cox_model,\n lr=lr)\n for epochs in [20]:\n for lr in [0.01, 0.0001, 0.00001, 0.1]:\n train(\n proj_dir=proj_dir,\n cox_model=cox_model,\n epochs=epochs,\n dl_train=dl_train,\n dl_tune=dl_tune,\n dl_val=dl_val,\n cnn_name=cnn_name,\n lr=lr)\n\n evaluate(\n proj_dir=proj_dir,\n cox_model=cox_model,\n load_model=load_model,\n dl_val=dl_val,\n score_type=score_type,\n cnn_name=cnn_name,\n epochs=epochs,\n lr=lr)\n\n","repo_name":"xmuyzz/HeadNeckCancer-Outcome","sub_path":"archive/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":2502,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"43624487941","text":"'''\n功能:执行到主程序末尾,解释器自动退出,但是如果需要中途退出程序,可以调用sys.exit函数,\n带有一个可选的整数参数返回给调用它的程序,表示你可以在主程序中捕获对sys.exit的调用。(0是正常退出,其他为异常)\n'''\n\nimport sys\n\ndef exitfunc(value):\n print(value)\n sys.exit(0)\n \nprint('hello')\n \ntry:\n sys.exit(1)\nexcept SystemExit as value:\n exitfunc(value)\n\nprint('come?')\n\n# hello\n# 1 #come?未打印\n\n# =================================================================================================================\n'''\nimport sys\nprint('The commend line argument is ')\nfor i in sys.argv:\n print(i)\n \nprint(\"\\n\\nThe PYTHONPATH is \",sys.path,'\\n')\n\n'''\n'''\nE:\\Progect\\tmp>python3 usesys.py we are argument\nThe commend line argument is\nusesys.py\nwe\nare\nargument\n\n\nThe PYTHONPATH is ['E:\\\\Progect\\\\tmp', 'C:\\\\Python35\\\\python35.zip', 'C:\\\\Python35\\\\DLLs', 'C:\\\\Python35\\\\lib', 'C:\\\\Python35', 'C:\\\\Python35\\\\lib\\\\site-packages']\n'''\n# ========================================================================================================================","repo_name":"renchao7060/studynotebook","sub_path":"基础学习/p5_sys模块使用.py","file_name":"p5_sys模块使用.py","file_ext":"py","file_size_in_byte":1149,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17610829419","text":"import re\nfrom nltk.corpus import wordnet\nfrom nltk.corpus import stopwords\nfrom torch.utils import data\n\n\nclass CustomDataset(data.Dataset):\n def __init__(self, data):\n \"\"\"\n initialize attributes of the CustomDataset class\n @param data: data\n \"\"\"\n self.data = data\n\n def __getitem__(self, index):\n \"\"\"\n @param index: index of the (tweet, label) tuple to return\n @return: single (tweet, label) tuple\n \"\"\"\n tweet = self.data.iloc[index]['tweet']\n label = self.data.iloc[index]['label']\n return tweet, label\n\n def __len__(self):\n \"\"\"\n @return: number of samples in the data\n \"\"\"\n return self.data.shape[0]\n\n\nclass PreProcessing():\n def __init__(self):\n \"\"\"\n initialize attributes of the PreProcessing class\n \"\"\"\n self.regex = re.compile(r'(\\w*)(\\w)\\2(\\w*)')\n self.repl = r'\\1\\2\\3'\n self.stop_words = stopwords.words(\"english\")\n\n def clean(self, tweet):\n \"\"\"\n @param tweet: single tweet as a string\n @return: cleaned tweet as a string\n \"\"\"\n tweet = self.basic_cleaning_3(tweet)\n tweet = tweet.split(' ')\n tweet = [self.remove_repeated_letters(word) for word in tweet]\n tweet = ' '.join(tweet)\n return tweet\n\n def basic_cleaning(self, tweet):\n \"\"\"\n @param tweet: single tweet as a string\n @return: list of words containing only some characters\n \"\"\"\n tweet = tweet.replace('', '')\n tweet = tweet.replace('', '')\n tweet = ''.join(c if (c.isalpha() or c.isspace()) else '' for c in tweet)\n tweet = ' '.join(tweet.split()).split(' ')\n tweet = list(filter(lambda x: x not in self.stop_words, tweet))\n return tweet\n\n def basic_cleaning_2(self, tweet):\n \"\"\"\n @param tweet: single tweet as a sting\n @return: single cleaned tweet as a string\n \"\"\"\n tweet = ''.join('' if c == \"'\" else c for c in tweet)\n tweet = ''.join(' ' if c == '#' else c for c in tweet)\n tweet = ''.join(' ' if c.isnumeric() else c for c in tweet)\n tweet = ''.join(c if (c.isalpha() or c.isspace() or c in ['<', '>']) else ' ' + c + ' ' for c in tweet)\n tweet = ' '.join(tweet.split())\n return tweet\n\n def basic_cleaning_3(self, tweet):\n \"\"\"\n @param tweet: single tweet as a string\n @return: single cleaned tweet\n \"\"\"\n tweet = ''.join('' if c == \"'\" else c for c in tweet)\n tweet = ''.join(c if c.isalpha() else ' ' + c + ' ' for c in tweet)\n tweet = ' '.join(tweet.split())\n return tweet\n\n def remove_repeated_letters(self, word):\n \"\"\"\n recursive method to remove repeated letters from a given word based on a dictionary.\n original implementation: https://www.youtube.com/user/RockyDeRaze\n @param word: word that might contains unnecessary repeated letters\n @return: cleaned word\n \"\"\"\n if wordnet.synsets(word):\n return word\n result = self.regex.sub(self.repl, word)\n if word == result:\n return result\n else:\n return self.remove_repeated_letters(result)\n\n","repo_name":"AleksandarHr/Twitter_Sentiment_Analysis","sub_path":"source/classes.py","file_name":"classes.py","file_ext":"py","file_size_in_byte":3287,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"28475949244","text":"# Вывести последнюю букву в слове\nword = 'Архангельск'\nprint (word[-1])\n# ???\n\n\n# Вывести количество букв \"а\" в слове\nword = 'Архангельск'\nprint(\"2\")\n# ???\n\n\n# Вывести количество гласных букв в слове\n\nfor test in ['Архангельск']:\n print(len(list(filter(lambda x: x in 'АЕИОУЭЮЫЯауоыиэяюе', test))))\n\n\n\n# Вывести количество слов в предложении\n\na = 'Мы приехали в гости'\nprint(len(a.split(' ')))\n# ???\n\n\n# Вывести первую букву каждого слова на отдельной строке\n\n\n\n\n\n# Вывести усреднённую длину слова.\ndef main():\n\n sentence = \"Мы приехали в гости\"\n SumAccum = 0\n for ch in sentence.split():\n character = len(ch)\n SumAccum = SumAccum + character\n\n average = (SumAccum) / (len(sentence.split()))\n return average\nprint(main())","repo_name":"MrStarkOff/practice_1","sub_path":"string_challenges.py","file_name":"string_challenges.py","file_ext":"py","file_size_in_byte":1054,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"74450794812","text":"\"\"\"\nThis is script for all our visualization needs\n\n\"\"\"\nimport torch\nimport seaborn as sns\nimport scanpy as sc\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom sklearn.metrics import confusion_matrix\nimport anndata\n\nfrom .helper_functions import ensure_dir\n\ndef save_cm(true, preds, name, reduce_cm=True, label_names=None, sort_labels=False):\n \"\"\"\n\n Parameters\n ----------\n true : list list (np.array)\n true labels for a cm\n preds : list like (np.array)\n predictions for a cm\n name : str\n name of figure (without .png ext)\n reduce_cm : bool\n toggle to get rid of 0/NA cols and rows\n label_names : list\n if the true and preds are ints, provide a list of names that map the ints to labels\n sort_labels : bool\n if the true and preds args are lists of names, choose to sort them by alphabetical order\n\n Returns\n -------\n\n \"\"\"\n if label_names is not None:\n cm = confusion_matrix(true, preds, normalize='true')\n labels = label_names\n else:\n if sort_labels:\n labels = np.unique(np.concatenate([true, preds]))\n else:\n labels = pd.factorize(np.concatenate([true, preds]))[1]\n cm = confusion_matrix(true, preds, normalize='true', labels=labels)\n cm_norm_df = pd.DataFrame(cm, index=labels, columns=labels)\n if reduce_cm:\n cm_norm_df = cm_norm_df.dropna(axis=0, how='all')\n cm_norm_df = cm_norm_df[~(cm_norm_df == 0).all(axis=1)]\n cm_norm_df = cm_norm_df.T[~(cm_norm_df == 0).all(axis=0)].T\n plt.figure(figsize=(cm_norm_df.shape[1], cm_norm_df.shape[0]))\n ax = sns.heatmap(cm_norm_df, cmap=\"YlGnBu\", vmin=0, vmax=1,\n linewidths=.5, annot=True, fmt='4.2f', square=True)\n ax.figure.tight_layout()\n ensure_dir(\"./cm_figs\")\n save_name = f\"./cm_figs/cm_{name}.png\"\n plt.savefig(save_name)\n\n\ndef plot_embeddings(model, data_loaders, device, fig_name):\n empty_zx = False\n patients = data_loaders['sup'].dataset.domains\n cell_types = data_loaders['sup'].dataset.labels\n\n # trying to plot training data\n actuals_d, actuals_y, zy_, zd_, zx_ = [], [], [], [], []\n with torch.no_grad():\n # Train\n # patients_train = np.delete(patients, test_patient)\n i = 0\n for (xs, ys, ds) in data_loaders['sup']:\n i = i + 1\n # To device\n xs, ys, ds = xs.to(device), np.array(ys), np.array(ds)\n # use classification function to compute all predictions for each batch\n zy_loc, zy_scale = model.qzy(xs)\n zd_loc, zd_scale = model.qzd(xs)\n if not empty_zx:\n zx_loc, zx_scale = model.qzx(xs)\n zx_.append(np.array(zx_loc.cpu()))\n zy_.append(np.array(zy_loc.cpu()))\n zd_.append(np.array(zd_loc.cpu()))\n # getting integer labels here\n actuals_d.append(np.argmax(ds, axis=1))\n actuals_y.append(np.argmax(ys, axis=1))\n # if i == 50:\n if i == len(data_loaders['sup']):\n break\n zy = np.vstack(zy_)\n zd = np.vstack(zd_)\n if not empty_zx:\n zx = np.vstack(zx_)\n labels_y = np.hstack(actuals_y)\n labels_d = np.hstack(actuals_d)\n if not empty_zx:\n zy_adata, zd_adata, zx_adata = [anndata.AnnData(_) for _ in [zy, zd, zx]]\n adatas = [zy_adata, zd_adata, zx_adata]\n else:\n zy_adata, zd_adata = [anndata.AnnData(_) for _ in [zy, zd]]\n adatas = [zy_adata, zd_adata]\n name = ['zy', 'zd', 'zx']\n train_labels = patients[labels_d]\n zy_adata.obs['batch'] = train_labels\n zy_adata.obs['cell_type'] = cell_types[labels_y]\n zd_adata.obs['batch'] = train_labels\n zd_adata.obs['cell_type'] = cell_types[labels_y]\n train_cell_type_encoding = zy_adata\n train_batch_encoding = zd_adata\n for i, _ in enumerate(adatas):\n _.obs['batch'] = patients[labels_d]\n _.obs['cell_type'] = cell_types[labels_y]\n save_name = f\"_{fig_name}_train_set_{name[i]}.png\"\n sc.pp.neighbors(_, use_rep=\"X\", n_neighbors=15)\n sc.tl.umap(_, min_dist=.3)\n sc.pl.umap(_, color=['batch', 'cell_type'], save=save_name)\n actuals_d, actuals_y, zy_, zd_, zx_ = [], [], [], [], []\n with torch.no_grad():\n # test\n # patients_train = np.delete(patients, test_patient)\n i = 0\n for (xs, ys, ds) in data_loaders['unsup']:\n i = i + 1\n # To device\n xs, ys, ds = xs.to(device), np.array(ys), np.array(ds)\n # use classification function to compute all predictions for each batch\n zy_loc, zy_scale = model.qzy(xs)\n zd_loc, zd_scale = model.qzd(xs)\n if not empty_zx:\n zx_loc, zx_scale = model.qzx(xs)\n zx_.append(np.array(zx_loc.cpu()))\n zy_.append(np.array(zy_loc.cpu()))\n zd_.append(np.array(zd_loc.cpu()))\n # getting integer labels here\n actuals_d.append(np.argmax(ds, axis=1))\n actuals_y.append(np.argmax(ys, axis=1))\n # if i == 50:\n if i == len(data_loaders['unsup']):\n break\n zy = np.vstack(zy_)\n zd = np.vstack(zd_)\n if not empty_zx:\n zx = np.vstack(zx_)\n labels_y = np.hstack(actuals_y)\n labels_d = np.hstack(actuals_d)\n if not empty_zx:\n zy_adata, zd_adata, zx_adata = [anndata.AnnData(_) for _ in [zy, zd, zx]]\n adatas = [zy_adata, zd_adata, zx_adata]\n else:\n zy_adata, zd_adata = [anndata.AnnData(_) for _ in [zy, zd]]\n adatas = [zy_adata, zd_adata]\n name = ['zy', 'zd', 'zx']\n test_labels = patients[labels_d]\n zy_adata.obs['batch'] = test_labels\n zy_adata.obs['cell_type'] = cell_types[labels_y]\n zd_adata.obs['batch'] = test_labels\n zd_adata.obs['cell_type'] = cell_types[labels_y]\n test_cell_type_encoding = zy_adata\n test_batch_encoding = zd_adata\n for i, _ in enumerate(adatas):\n _.obs['batch'] = patients[labels_d]\n _.obs['cell_type'] = cell_types[labels_y]\n save_name = f\"_{fig_name}_test_set_{name[i]}.png\"\n sc.pp.neighbors(_, use_rep=\"X\", n_neighbors=15)\n sc.tl.umap(_, min_dist=.3)\n sc.pl.umap(_, color=['batch', 'cell_type'], save=save_name)\n full_zy = train_cell_type_encoding.concatenate(test_cell_type_encoding)\n full_zd = train_batch_encoding.concatenate(test_batch_encoding)\n all_patients = np.hstack([train_labels, test_labels])\n full_zy.obs['batch'] = all_patients\n full_zd.obs['batch'] = all_patients\n sc.pp.neighbors(full_zy, n_neighbors=15)\n sc.pp.neighbors(full_zd, n_neighbors=15)\n sc.tl.umap(full_zy, min_dist=.3)\n sc.tl.umap(full_zd, min_dist=.3)\n sc.pl.umap(full_zy, color=['batch', 'cell_type'], save=f\"_{fig_name}_train+test_zy.png\")\n sc.pl.umap(full_zd, color=['batch', 'cell_type'], save=f\"_{fig_name}_train+test_zd.png\")","repo_name":"bplee/scBatch_project","sub_path":"scBatch/visualization.py","file_name":"visualization.py","file_ext":"py","file_size_in_byte":7157,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"21666978205","text":"\"\"\"\r\n94991277\r\n\r\n-- ПРИНЦИП РАБОТЫ --\r\n\r\nМы создаем двумерный массив dp, где dp[i][j]\r\nпредставляет минимальное расстояние между подстроками s1[:i] и s2[:j].\r\nЗатем мы инициализируем первую строку и первый столбец (расстояние от пустой строки к каждой подстроке),\r\n а затем заполняем остальные ячейки массива, используя формулу для расчета расстояния Левенштейна.\r\n\r\n -- ДОКАЗАТЕЛЬСТВО КОРРЕКТНОСТИ --\r\n\r\nМы создаем двумерный массив dp, где dp[i][j]\r\nбудет представлять минимальное расстояние между подстроками s1[:i] и s2[:j].\r\nМы определяем размеры массива m и n как длины строк s1 и s2,\r\nи затем создаем массив dp размером (m+1) x (n+1) и инициализируем его нулями.\r\nДалее инициализируем первую строку и первый столбец массива dp.\r\nПервая строка представляет расстояние от пустой строки s1[:i] к подстроке s2[:j],\r\nа первый столбец - расстояние от подстроки s1[:i] к пустой строке s2[:j].\r\nЭти исходные значения задаются от 0 до длин строк s1 и s2.\r\n\r\nПотом мы заполняем остальные ячейки массива dp с помощью вложенных циклов.\r\nМы вычисляем расстояние Левенштейна для каждой пары символов из строк s1 и s2.\r\nЕсли символы в текущих позициях совпадают (s1[i-1] == s2[j-1]),\r\n cost устанавливается в 0, иначе в 1.\r\nЗатем мы вычисляем три возможных пути для заполнения ячейки dp[i][j]:\r\n\r\ndp[i-1][j] + 1 - расстояние от s1[:i-1] к s2[:j], учитывая операцию удаления.\r\ndp[i][j-1] + 1 - расстояние от s1[:i] к s2[:j-1], учитывая операцию вставки.\r\ndp[i-1][j-1] + cost - расстояние от s1[:i-1] к s2[:j-1], учитывая операцию замены (с учетом cost).\r\nЗатем мы выбираем минимальное из этих трех значений и устанавливаем его как значение dp[i][j].\r\n\r\nВ конце мы возвращаем значение dp[m][n]\r\n\r\n-- ВРЕМЕННАЯ СЛОЖНОСТЬ --\r\n\r\nВременная сложность алгоритма Левенштейна,\r\nреализованного при помощи динамического программирования,\r\n O(m * n), где m и n - длины строк s1 и s2\r\n\r\n-- ПРОСТРАНСТВЕННАЯ СЛОЖНОСТЬ --\r\n\r\nПространственная сложность алгоритма также составляет O(m * n),\r\n так как мы используем двумерный массив dp\r\n\r\n\"\"\"\r\n\r\n\r\ndef levenshtein_distance(s1, s2):\r\n\r\n m, n = len(s1), len(s2)\r\n dp = [[0] * (n + 1) for _ in range(m + 1)]\r\n\r\n for i in range(m + 1):\r\n dp[i][0] = i\r\n for j in range(n + 1):\r\n dp[0][j] = j\r\n\r\n for i in range(1, m + 1):\r\n for j in range(1, n + 1):\r\n cost = 0 if s1[i - 1] == s2[j - 1] else 1\r\n dp[i][j] = min(\r\n dp[i - 1][j] + 1, # Удаление\r\n dp[i][j - 1] + 1, # Вставка\r\n dp[i - 1][j - 1] + cost # Замена\r\n )\r\n\r\n return dp[m][n]\r\n\r\n\r\nif __name__ == \"__main__\":\r\n s1 = input()\r\n s2 = input()\r\n distance = levenshtein_distance(s1, s2)\r\n print(distance)\r\n","repo_name":"Donskoy-Mikhail/Yandex-algorithms","sub_path":"sprints/sprint_7/final_A.py","file_name":"final_A.py","file_ext":"py","file_size_in_byte":4059,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"71048967291","text":"n=int(input())\ns=input()\n\ndef nc2(n):\n if n<2:\n return 0\n return n*(n-1)//2\n\nret=0\nnow=None\ncnt=0\ns+=\"*\"\nfor i in range(n+1):\n if now is None:\n now=s[i]\n cnt=1\n elif now==s[i]:\n cnt+=1\n else:\n now=s[i]\n ret+=nc2(cnt)\n cnt=1\nprint(ret)","repo_name":"ayanamizuta/cpro","sub_path":"atcoder/arc/arc130/a.py","file_name":"a.py","file_ext":"py","file_size_in_byte":298,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"25531978406","text":"#!/usr/bin/env python3\n# coding=utf-8\n#\n# File: EmBCI/embci/tests/test_io.py\n# Authors: Hank \n# Create: 2019-02-06 01:26:32\n#\n# TODO:\n# test Socket***Reader, Socket***Server\n# test LSLCommader, LSLReader\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nimport os\nimport time\nimport warnings\nimport threading\n\n# requirements.txt: testing: pytest\n# requirements.txt: drivers: pyserial\n# requirements.txt: data: pylsl\nimport pytest\nimport serial\nimport pylsl\n\n\n# =============================================================================\n# functions\n#\nfrom embci.configs import DIR_DATA\nfrom embci.io import save_trials, load_data, create_data_dict, find_data_info\n\n\ndef test_save_trials(username, random_data, clean_userdir):\n clean_userdir()\n data_dict = create_data_dict(random_data, 'testing', 500)\n assert 'sample_rate' in data_dict\n save_trials(username, data_dict, suffix='mat')\n assert os.path.exists(os.path.join(DIR_DATA, username, 'testing-0.mat'))\n assert os.path.exists(os.path.join(DIR_DATA, username, 'testing-1.mat'))\n\n\ndef test_load_data(username, random_data, clean_userdir):\n data, label = load_data(username)\n assert 'testing' == label[0] == label[1]\n assert (random_data == data).all()\n clean_userdir()\n\n\ndef test_find_data_info(clean_userdir, username):\n label_dict, name_dict, summary = find_data_info(username)\n assert label_dict == {}\n assert name_dict == {}\n # DO NOT test outputs, because it may be changed in the future.\n # assert 'There are 0 actions with 0 data recorded' in summary\n clean_userdir()\n\n\n# =============================================================================\n# Readers\n#\nfrom embci.io import FakeDataGenerator as Reader\nfrom embci.utils import find_pylsl_outlets\n\n\n@pytest.fixture(scope='module')\ndef reader():\n reader = Reader(sample_rate=500, sample_time=2,\n num_channel=8, broadcast=True)\n reader.start()\n yield reader\n reader.close()\n\n\ndef test_reader_status(reader):\n assert reader.status == 'started'\n assert reader.is_streaming()\n\n\ndef test_stream_control(reader):\n reader.pause()\n assert reader.status == 'paused'\n assert reader.is_streaming() is False\n reader.resume()\n assert reader.status == 'resumed'\n assert reader.is_streaming() is True\n\n\ndef test_reader_data(reader):\n assert reader.data_channel.shape == (8,)\n assert reader.data_frame.shape == (8, 1000)\n\n\ndef test_reader_pylsl(reader):\n info = find_pylsl_outlets(source_id=reader.name)\n assert isinstance(info, pylsl.StreamInfo)\n\n\ndef test_set_sample_rate(reader):\n reader.pause()\n assert reader.set_sample_rate(250)\n reader.restart()\n time.sleep(3) # reader need some time to stablize the sample_rate\n assert abs(reader.realtime_samplerate - 250) < 100\n\n\n# =============================================================================\n# Commanders\n#\nfrom embci.io import SerialCommander\nfrom embci.utils import virtual_serial\n\n\n@pytest.fixture(scope='module')\ndef obj(request):\n flag_stop, port1, port2 = virtual_serial(verbose=False)\n cmder = SerialCommander({\n '_desc': 'command dict used for testing commander',\n 'action1': ('asdf', 0),\n 'action3': ('this is action3', 1),\n 'nodelay': ('nodelay', ),\n })\n cmder.start(port1, 115200)\n slave = serial.Serial(port2, 115200)\n\n request.addfinalizer(flag_stop.set)\n request.addfinalizer(cmder.close)\n request.addfinalizer(slave.close)\n\n class objs:\n serial = slave\n commander = cmder\n return objs\n\n\ndef test_get_command(obj):\n with warnings.catch_warnings():\n assert obj.commander.get_command('foo') is None\n\n\ndef test_send_command(obj):\n assert obj.commander.send('action1') == 'asdf'\n time.sleep(0.5)\n assert obj.serial.read_all() == b'asdf'\n\n\ndef test_send_wait(obj):\n '''Method `send` will acquire lock and wait for specific seconds.'''\n # send command twice\n threading.Thread(\n target=obj.commander.send, args=('action3',)).start()\n threading.Thread(\n target=obj.commander.send, args=('action3',)).start()\n # receive first command\n time.sleep(0.5)\n assert obj.serial.read_all() == b'this is action3'\n # receive nothing, commander is still waiting\n assert obj.serial.read_all() == b''\n time.sleep(1)\n # receive second command\n assert obj.serial.read_all() == b'this is action3'\n\n\ndef test_write_method(obj):\n try:\n print('nodelay', file=obj.commander)\n except Exception as e:\n assert isinstance(e, IndexError)\n\n# THE END\n","repo_name":"hankso/EmBCI","sub_path":"tests/test_io.py","file_name":"test_io.py","file_ext":"py","file_size_in_byte":4672,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"3371454411","text":"from board import Board\nfrom btree import LinkedBinaryTree\nimport copy\n\n\nclass TicTacToeTree:\n '''\n Class to represent TicTacToe game with computer\n '''\n def __init__(self, root):\n '''\n Initializes variables including root of a future generated tree\n\n :param root: Board\n '''\n self.root = root\n self.tree = LinkedBinaryTree(root)\n\n def generate_tree(self, root=None):\n '''\n Generates a binary search tree of possible choices\n\n :param root: Board\n :return: None\n '''\n if root is None: root = self.tree\n else: root = root\n if root.key.has_winner():\n return 0\n right_board = copy.deepcopy(root.key)\n left_board = copy.deepcopy(root.key)\n right_board.make_random_move()\n left_board.make_random_move()\n\n root.insert_left(left_board)\n root.insert_right(right_board)\n\n self.generate_tree(root.left_child)\n self.generate_tree(root.right_child)\n return\n\n def count_winning_amount(self):\n '''\n Counts amount of winning combinations on both sides of a tree\n\n :return: (int, int)\n '''\n my_tree = self.tree\n right_tree = my_tree.right_child\n left_tree = my_tree.left_child\n right_counter = set()\n left_counter = set()\n\n nought_right = set()\n nought_left = set()\n\n right_nodes = right_tree.get_leaves_list()\n left_nodes = left_tree.get_leaves_list()\n\n for right in right_nodes:\n if right.has_winner() == right.CROSS:\n right_counter.add(str(right))\n elif right.has_winner() == right.NOUGHT:\n nought_right.add(str(right))\n\n for left in left_nodes:\n if left.has_winner() == left.CROSS:\n left_counter.add(str(left))\n elif left.has_winner() == left.NOUGHT:\n nought_left.add(str(left))\n\n return len(right_counter) - len(nought_right), len(left_counter) - len(nought_left)\n\n\ndef main():\n '''\n Main function to simulate a Tic Tac Tie game\n\n :return: None\n '''\n transform = {1: 'NOUGHT', -1: 'CROSS', 2: 'DRAW'}\n init_board = Board()\n init_board.make_random_move()\n while True:\n if init_board.has_winner():\n print(init_board)\n print(f'And the winner is: {transform[init_board.has_winner()]}')\n break\n\n print(init_board)\n while True:\n user_choice = input('Enter your choice[1 1]: ')\n try:\n splitted = [int(i) for i in user_choice.split()]\n assert 0 <= splitted[0] <= 2\n assert 0 <= splitted[1] <= 2\n assert init_board.cells[splitted[0]][splitted[1]] == 0\n break\n\n except:\n print('Wrong choice')\n continue\n\n init_board.make_move(splitted)\n\n tic = TicTacToeTree(init_board)\n tic.generate_tree()\n tree_winning_combinations = tic.count_winning_amount()\n right_subtree = tree_winning_combinations[0]\n left_subtree = tree_winning_combinations[1]\n init_board = tic.tree.left_child.key\n if right_subtree >= left_subtree: init_board = tic.tree.right_child.key\n\n\nif __name__ == '__main__':\n main()","repo_name":"Sofiia2001/TicTacToe_BST","sub_path":"tictactoe/game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":3344,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"15595088883","text":"from myhdl import *\nimport random\n\n@block\ndef Tdecode(opCode, RType, Load, Store, Branch, IType, Jalr, Jal, Lui,Aui):\n\n @always_comb\n def run():\n\n RType.next = intbv(0)[1:]\n Load.next = intbv(0)[1:]\n Store.next = intbv(0)[1:]\n Branch.next = intbv(0)[1:]\n IType.next = intbv(0)[1:]\n Jalr.next = intbv(0)[1:]\n Jal.next = intbv(0)[1:]\n Lui.next = intbv(0)[1:]\n Aui.next = intbv(0)[1:]\n \n if opCode == 51:\n RType.next = 1\n elif opCode == 3:\n Load.next = 1\n elif opCode == 35:\n Store.next = 1\n elif opCode == 99:\n Branch.next = 1\n elif opCode == 19:\n IType.next = 1\n elif opCode == 103:\n Jalr.next = 1\n elif opCode == 111:\n Jal.next = 1\n elif opCode == 55:\n Lui.next = 1\n elif opCode == 23:\n Aui.next = 1\n \n return run\n\nopCodes = [51,3,35,99,19,103,111,55,23]\nR, L, S, B, I, Jr, J, Li,Ai = [Signal(bool(0)) for i in range(9)]\nopCode = Signal(intbv(0, min=0, max=112))\n\ntd = Tdecode(opCode,R, L, S, B, I, Jr, J, Li,Ai)\ntd.convert('Verilog')\n\n# @block\n# def Tdtest():\n\n# R, L, S, B, I, Jr, J, Li,Ai = [Signal(bool(0)) for i in range(9)]\n# opCode = Signal(intbv(0, min=0, max=112))\n\n# td = Tdecode(opCode,R, L, S, B, I, Jr, J, Li,Ai)\n# td.convert('Verilog')\n# @instance\n# def test():\n# fmt = \"{0:6} | {1:5} | {2:5} | {3:5} | {4:6} | {5:5} | {6:5} | {7:5} | {8:5} | {9:5} \"\n# print(fmt.format(\"OpCode\",\"RType\", \"Load\", \"Store\", \"Branch\", \"IType\", \"Jalr\",\"Jal\", \"Lui\",\"AuiPC\"))\n# for i in range(10):\n# opCode.next = random.choice(opCodes)\n# yield delay(10)\n# print(fmt.format(str(int(opCode)), str(int(R)), str(int(L)), str(int(S)), str(int(B)), str(int(I)), str(int(Jr)), str(int(J)), str(int(Li)), str(int(Ai))))\n# # raise StopSimulation\n \n# return td, test\n\n# tb = Tdtest()\n# tb.run_sim()","repo_name":"M-Ash2209/Python-MyHDL-Learning-Journey","sub_path":"Core/Main/Tdecode.py","file_name":"Tdecode.py","file_ext":"py","file_size_in_byte":2028,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"36870619076","text":"from mcpi.minecraft import Minecraft\nmc=Minecraft.create()\nimport time\n\nblocks=[20,56,57,20,56,57,20,56,57]\nbarblock=22\n\nx,y,z=mc.player.getPos()\nstarty=y\nfront=5\ncount=0\nfor block in blocks: \n mc.setBlock(x+front,y,z,block[count])\n y+=1\n count+=1\n time.sleep(1)\n \nmc.postToChat(\"finish\")\ntime.sleep(1)\nblocks.insert(2,barblock)\n\ny=starty\nx+=front\nfor block in blocks: \n mc.setBlock(x+front,y,z,block)\n y+=1\n time.sleep(1)\n \nmc.postToChat(\"finished\")","repo_name":"newsteinking/workspace_backup","sub_path":"workspace_K/workspace_jihyung/python/python_minecraft/Learn to Program with Minecraft Code실습/chapter9-lists/test4.py","file_name":"test4.py","file_ext":"py","file_size_in_byte":477,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"19012809118","text":"n=100000\nlist=[]\n\nprime = [True for i in range(n+1)]\np = 2\nwhile (p * p <= n):\n if (prime[p] == True):\n for i in range(p * 2, n+1, p):\n prime[i] = False\n p += 1\n \n \nfor p in range(2, n+1):\n if prime[p]:\n list.append(p)\n \n#print(list)\n\nlength=len(list)\ndef primeFactorization(n):\n f=1\n for i in range(length):\n p=list[i]\n c=0\n while n%p==0:\n c=c+1\n n=n//p\n f=f*(c+1)\n if n==1:\n return f \n \n \ni=1\nwhile True:\n n=(i*(i+1))//2\n if primeFactorization(n)>500:\n print(n)\n break\n\n i+=1\n \n \n","repo_name":"ChanchalKumarMaji/Project-Euler","sub_path":"Project_Euler_12.py","file_name":"Project_Euler_12.py","file_ext":"py","file_size_in_byte":649,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"9303942927","text":"import time\n\ndata = open('input/input.txt', 'r').readlines()\nranges = {}\nfor line in data:\n\tranges[int(line.split(': ')[0])] = int(line.split(': ')[1])\n\ndef captured_layers(delay):\n\tcaptured = []\n\tfor depth in ranges:\n\t\tif (delay + depth) % (ranges[depth] * 2 - 2) == 0:\n\t\t\tcaptured.append(depth)\n\treturn captured\n\ndef captured_layers_early(delay):\n\tcaptured = []\n\tfor depth in ranges:\n\t\tif (delay + depth) % (ranges[depth] * 2 - 2) == 0:\n\t\t\tcaptured.append(depth)\n\t\t\treturn captured\n\treturn captured\n\n#Part One\ncaptured = captured_layers(0)\nscore = 0 \nfor depth in captured:\n\tscore += depth * ranges[depth]\n\n#Part Two\ndelay = 0\nSTART = time.time()\nwhile len(captured_layers(delay)) != 0:\n\tprint(delay)\n\tdelay += 1\n\nprint('Time: {}'.format(time.time() - START))\nprint(delay)\n\n","repo_name":"r-tran/advent-of-code","sub_path":"aoc-2017/day13/day13.py","file_name":"day13.py","file_ext":"py","file_size_in_byte":778,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"3377445044","text":"from flask_wtf import FlaskForm\nfrom wtforms import StringField, EmailField ,SubmitField, TextAreaField,DecimalField, PasswordField,IntegerField, BooleanField, SelectField, DateField, ValidationError\nfrom wtforms.validators import DataRequired, Email, EqualTo, Length, NumberRange, InputRequired \nfrom datetime import datetime\nfrom helpers import inside, inside2\nimport datetime as dt\n\n\n\nclass ClientSignUpForm(FlaskForm):\n company_name = StringField(\"Name\", validators=[DataRequired()])\n email = EmailField(\"Email\", validators=[DataRequired(), Email()])\n password = PasswordField(\"Password\", validators=[DataRequired(), Length(min=8, max=20)])\n confirm_password = PasswordField(\"Comfirm Password\", \n \tvalidators = [DataRequired(), EqualTo('password')])\n submit = SubmitField(\"Sign Up\")\n\n def validate_email(self, email):\n \tuser = User.query.filter_by(email=email.data).first()\n \tif user:\n \t\traise ValueError(\"The email is already in use, please choose a different one\")\n\n\nclass StudentLedgerForm(FlaskForm):\n\tphone = IntegerField(\"Parent's Contact\", validators=[DataRequired()])\n\tfirstname = StringField(\"First Name\", validators=[DataRequired()])\n\tsubmit = SubmitField(\"Generate\")\n\n\tdef validate_phone(self, phone):\n\t\tnum = str(phone.data)\n\t\tprint(len(num))\n\t\tif len(num) != 9:\n\t\t\traise ValidationError(\"Phone number must be 10 digits\")\n\n\tdef validate_firstname(self, firstname):\n\t\tfor char in firstname.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\t\n\nclass ClientLogInForm(FlaskForm):\n email = EmailField(\"Email\", validators=[DataRequired(), Email()])\n password = PasswordField(\"Password\", validators=[DataRequired()])\n remember = BooleanField(\"Remember me\")\n submit = SubmitField(\"Log In\")\n\n\nclass ToDoForm(FlaskForm):\n task = SelectField(\"Choose A Task\", choices = ['Make E.T.L Expenses','Make P.T.A Expenses', 'Begin Semester', 'Promote Student'], validators=[DataRequired()])\n submit_do = SubmitField(\"Proceed\")\n\n\nclass StudentPaymentsForm(FlaskForm):\n\tetl_amount = DecimalField(\"ETL\", validators=[InputRequired(), NumberRange(min=0, max=3000)])\n\tpta_amount = DecimalField(\"PTA\", validators=[InputRequired(), NumberRange(min=0, max=3000)])\n\tcheq_no = StringField('Cheq ID', validators = [Length(max=20)])\n\tmode = SelectField('Mode of payment', choices = [('','Choose Payment mode...'), (\"Cash\", 'Cash'), (\"Cheque\",'Cheque')], validators=[DataRequired()])\n\tsemester = SelectField(\"Semester\", choices = [('','Choose semester...'),(\"SEM1\", 'SEM1'), (\"SEM2\",'SEM2')], validators=[DataRequired()])\n\tsubmit = SubmitField(\"Receive\")\n\nclass ExpensesForm(FlaskForm):\n\tpurchase_date = DateField(\"Purchase Date\", validators=[DataRequired()])\n\titem = StringField(\"Item\", validators=[DataRequired(), Length(max=20)])\n\tpurpose = StringField(\"Purpose\", validators=[DataRequired(), Length(max=50)])\n\tunitcost = DecimalField(\"Quantity\", validators=[DataRequired(), NumberRange(min=1, max=30000)])\n\tquantity = DecimalField(\"Quantity\", validators=[DataRequired(), NumberRange(min=1, max=30000)])\n\ttotalcost = DecimalField(\"Total Cost\", validators=[DataRequired(), NumberRange(min=1, max=300000)])\n\tsubmit = SubmitField(\"Debit\")\n\t\t\n\tdef validate_item(self, item):\n\t\tfor char in item.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\t\t\t\n\tdef validate_purpose(self, purpose):\n\t\tfor char in purpose.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\n\tdef validate_totalcost(self, totalcost):\n\t\tif totalcost.data != self.quantity.data * self.unitcost.data:\n\t\t\traise ValidationError(f\"Totals cost should be {self.quantity.data * self.unitcost.data} NOT {totalcost.data}\")\n\n\n\tdef validate_purchase_date(self, purchase_date):\n\t\ttoday = datetime.utcnow()\n\t\ttoday = dt.date(year=today.year, month=today.month, day=today.day)\n\t\tpurchase_date1 = dt.date(year=purchase_date.data.year, month=purchase_date.data.month, day=purchase_date.data.day)\n\t\tif purchase_date1 > today:\n\t\t\traise ValidationError(f\"Date cant't be further than {today}\")\n\n\nclass DonationForm(FlaskForm):\n\tname = StringField(\"Name\", validators=[DataRequired()])\n\tamount = DecimalField(\"Amount\", validators=[DataRequired(), NumberRange(min=1, max=300000)])\n\tmode = SelectField(\"Mode of Payment\", validators=[DataRequired()], choices = [('','Choose mode of payment...'),(\"Cash\", 'Cash'), (\"Momo\",'Momo'), ('Cheque', 'Cheque')])\n\tsemester = SelectField(\"Semester\", validators=[DataRequired()], choices = [('','Choose semester...'),(\"SEM1\", 'SEM1'), (\"SEM2\",'SEM2')])\n\tsubmit = SubmitField(\"Receive Cash\")\n\n\tdef validate_name(self, name):\n\t\tfor char in name.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\n\nclass OtherBusinessForm(FlaskForm):\n\tname = StringField(\"Name\", validators=[DataRequired()])\n\tstart_date = DateField(\"Start Date\", validators = [DataRequired()])\n\tend_date = DateField(\"End Date\", validators = [DataRequired()])\n\tdetail = StringField(\"Details\")\n\tamount = DecimalField(\"Amount\", validators=[DataRequired(), NumberRange(min=1, max=300000)])\n\tother_submit = SubmitField(\"Receive Cash\")\n\n\tdef validate_detail(self, detail):\n\t\tfor char in detail.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\n\tdef validate_name(self, name):\n\t\tfor char in name.data:\n\t\t\tif inside(ch=char) == False:\n\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n\n#class ETLExpensesForm(FlaskForm):\n#\tpurchase_date = DateField(\"Purchase Date\", validators=[DataRequired()])\n#\titem = StringField(\"Item\", validators=[DataRequired(), Length(max=20)])\n#\tpurpose = StringField(\"Purpose\", validators=[DataRequired(), Length(max=50)])\n#\tunitcost = DecimalField(\"Quantity\", validators=[DataRequired(), NumberRange(min=1, max=30000)])\n#\tquantity = DecimalField(\"Quantity\", validators=[DataRequired(), NumberRange(min=1, max=30000)])\n#\ttotalcost = DecimalField(\"Total Cost\", validators=[DataRequired(), NumberRange(min=1, max=300000)])\n#\tsubmit = SubmitField(\"Debit\")\n#\t\t\n#\tdef validate_item(self, item):\n#\t\tfor char in item.data:\n#\t\t\tif inside(ch=char) == False:\n#\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n#\t\t\t\n#\tdef validate_purpose(self, purpose):\n#\t\tfor char in purpose.data:\n#\t\t\tif inside(ch=char) == False:\n#\t\t\t\traise ValidationError(f'Character {char} is not allowed')\n#\n#\tdef validate_totalcost(self, totalcost):\n#\t\tif totalcost.data != self.quantity.data * self.unitcost.data:\n#\t\t\traise ValidationError(f\"Totals cost should be {self.quantity.data * self.unitcost.data} NOT {totalcost.data}\")\n#\n#\tdef validate_purchase_date(self, purchase_date):\n#\t\ttoday = datetime.utcnow()\n#\t\ttoday = dt.date(year=today.year, month=today.month, day=today.day)\n#\t\tpurchase_date1 = dt.date(year=purchase_date.data.year, month=purchase_date.data.month, day=purchase_date.data.day)\n#\t\tif purchase_date1 > today:\n#\t\t\traise ValidationError(f\"Date cant't be further than {today}\")\n\n\nclass ReportsForm(FlaskForm):\n report = SelectField(\"Choose A Report\", validators=[DataRequired()], choices = ['Cash Book', 'Income & Expenditure', 'Expenditure Statement', 'Income Statement', 'INCOME & EXPENDITURE', 'CASH PAYMENT', 'CASH RECEIPT'])\n filter_by = SelectField(\"Choose Category\", choices = ['PTA Levy', 'ETL', 'ETL & PTA Levy'])\n start = DateField(\"Start\", validators=[DataRequired()])\n end = DateField(\"End\", validators=[DataRequired()])\n submit_rep = SubmitField(\"Generate\")\n\n def validate_end(self, end):\n \tif end.data < self.start.data:\n \t\traise ValidationError(\"Date must be latter than start date\")\n\n #def validate_start(self, start):\n #\ttoday = datetime.utcnow()\n #\ttoday = dt.date(year=today.year, month=today.month, day=today.day)\n #\tstart1 = dt.date(year=start.data.year, month=start.data.month, day=start.data.day)\n #\tif start1 > today:\n #\t\traise ValidationError(f\"Date is greater than {today}\")\n\nclass ChargeForm(FlaskForm):\n semester = SelectField(\"Choose semester\", validators=[DataRequired()], \n \tchoices = [('','Choose semester...'),(\"SEM1\", 'SEM1'), (\"SEM2\",'SEM2')])\n begin_date = DateField(\"Start Date\", validators= [DataRequired()])\n end_date = DateField(\"End Date\", validators= [DataRequired()])\n pta = DecimalField(\"PTA Levy\", validators=[DataRequired()])\n etl = DecimalField(\"ETL\", validators=[DataRequired()])\n submit = SubmitField(\"Get Started\")\n\n def validate_begin_date(self, begin_date):\n \ttoday = datetime.utcnow()\n \ttoday = dt.date(year=today.year, month=today.month, day=today.day)\n \tbegin_date1 = dt.date(year=begin_date.data.year, month=begin_date.data.month, day=begin_date.data.day)\n \tif begin_date1 > today:\n \t\traise ValidationError(f\"Date is greater than {today}\")\n\n def validate_end_date(self, end_date):\n \tif end_date.data <= self.begin_date.data:\n \t\traise ValidationError(\"End date must be latter than start date\")\n\n \t\n\nclass SearchForm(FlaskForm):\n parent_contact = StringField(\"Parent Contact\", validators=[DataRequired(), Length(min=8, max=20)])\n firstname = StringField(\"First Name\", validators=[DataRequired()])\n search_submit = SubmitField(\"Search\")\n\n def validate_parent_contact(self, parent_contact):\n \tfor char in parent_contact.data:\n \t\tif inside2(ch=char) == False:\n \t\t\traise ValidationError(f'Character {char} is not allowed')\n\n def validate_firstname(self, firstname):\n \tfor char in firstname.data:\n \t\tif inside(ch=char) == False:\n \t\t\traise ValidationError(f'Character {char} is not allowed')","repo_name":"KINTEP/kpasec22","sub_path":"forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":9611,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"69910025864","text":"import requests\r\nimport xml.etree.ElementTree as ET\r\nimport pandas as pd\r\nimport gzip\r\n\r\nURL = \"https://www.hackerrank.com/\"\r\n\r\n\r\nclass SitemapParser:\r\n def __init__(self, base_url):\r\n \"\"\"\r\n Initialize the SitemapParser with a base URL.\r\n\r\n Parameters:\r\n - base_url (str): The base URL used to parse the robots.txt data to retrieve Sitemap data.\r\n\r\n Returns:\r\n - None\r\n \"\"\"\r\n self.base_url = base_url\r\n self.sitemaps = []\r\n\r\n def fetch_robots_txt(self):\r\n \"\"\"\r\n Fetch the content of the robots.txt file for the specified URL.\r\n\r\n Returns:\r\n - str: The content of the robots.txt file.\r\n \"\"\"\r\n robots_url = f\"{self.base_url}/robots.txt\"\r\n headers = {\r\n \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/119.0.0.0 Safari/537.36\"\r\n }\r\n try:\r\n response = requests.get(robots_url, headers=headers)\r\n response.raise_for_status()\r\n return response.text\r\n except requests.exceptions.RequestException as e:\r\n print(f\"Error fetching robots.txt: {e}\")\r\n return None\r\n\r\n def extract_sitemaps(self, robots_content):\r\n \"\"\"\r\n Extract sitemap URLs from the content of robots.txt.\r\n\r\n Parameters:\r\n - robots_content (str): The content of the robots.txt file.\r\n\r\n Returns:\r\n - list: A list of sitemap URLs.\r\n \"\"\"\r\n sitemaps = []\r\n if robots_content:\r\n lines = robots_content.split(\"\\n\")\r\n for line in lines:\r\n if line.startswith(\"Sitemap:\"):\r\n sitemap_url = line.split(\": \")[1].strip()\r\n sitemaps.append(sitemap_url)\r\n return sitemaps\r\n\r\n def parse_sitemap(self, sitemap_url):\r\n \"\"\"\r\n Parse the sitemap XML content for a given sitemap URL.\r\n\r\n Parameters:\r\n - sitemap_url (str): The URL of the sitemap XML.\r\n\r\n Returns:\r\n - list: A list of URLs extracted from the sitemap XML.\r\n \"\"\"\r\n try:\r\n response = requests.get(sitemap_url)\r\n response.raise_for_status()\r\n decompressed_content = gzip.decompress(response.content).decode(\"utf-8\")\r\n root = ET.fromstring(decompressed_content)\r\n urls = [\r\n elem.text\r\n for elem in root.findall(\r\n \".//{http://www.sitemaps.org/schemas/sitemap/0.9}loc\"\r\n )\r\n ]\r\n return urls\r\n except requests.exceptions.RequestException as e:\r\n print(f\"Error parsing sitemap: {e}\")\r\n return []\r\n\r\n def parse_all_sitemaps(self):\r\n \"\"\"\r\n Parse all sitemaps for the specified base URL and return the result in a DataFrame.\r\n\r\n Returns:\r\n - pandas.DataFrame: A DataFrame containing URLs from all parsed sitemaps.\r\n \"\"\"\r\n robots_content = self.fetch_robots_txt()\r\n self.sitemaps = self.extract_sitemaps(robots_content)\r\n\r\n all_urls = []\r\n for sitemap_url in self.sitemaps:\r\n urls = self.parse_sitemap(sitemap_url)\r\n all_urls.extend(urls)\r\n\r\n df = pd.DataFrame(all_urls, columns=[\"URL\"])\r\n return df\r\n\r\n\r\n# Example usage with a website\r\nwebsite_url = URL\r\nsitemap_parser = SitemapParser(website_url)\r\nresult_df = sitemap_parser.parse_all_sitemaps()\r\n\r\n# Display the DataFrame\r\nprint(result_df)\r\n","repo_name":"shivakrishna67/AnalyticalProgramming-Project-2","sub_path":"src/part1.py","file_name":"part1.py","file_ext":"py","file_size_in_byte":3524,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21381082597","text":"#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n#\n# Usage:\n# ./beheader.py \n\nimport copy\nimport gzip\nimport time\nimport argparse\nimport urllib.request\n\n\ndef print_info(message):\n print('\\033[92m\\033[1m[INFO]\\033[0m\\033[0m ' + message)\n\n\ndef print_error(message):\n print('\\033[91m\\033[1m[ERROR]\\033[0m\\033[0m ' + message)\n exit()\n\n\ndef parseArgs():\n parser = argparse.ArgumentParser(\n usage='beheader.py curl [URL] [-X REQUEST] [-H HEADER] [-d DATA]')\n parser.add_argument('-X', '--request', nargs=1, help='Request method')\n parser.add_argument('-H', '--header', nargs=1, action='append',\n help='Request headers')\n parser.add_argument('-d', '--data', '--data-raw', '--data-binary',\n nargs=1, action='append',\n help='Request data')\n return parser.parse_known_args()\n\n\ndef get_url(args):\n for a in args:\n if not a.startswith('-'):\n return a\n print_error('Missing URL!')\n\n\ndef get_rest_options(args):\n optlist = []\n for a in args:\n if a.startswith('-'):\n optlist.append(a)\n return optlist\n\n\ndef get_method(args):\n if args.request is not None:\n return args.request[0]\n if args.data is not None:\n return 'POST'\n return 'GET'\n\n\ndef header_list_to_dict(headerlist):\n dict = {}\n if headerlist is not None:\n for h in headerlist:\n parts = h[0].split(':')\n key = parts[0]\n value = ':'.join(parts[1:]).lstrip()\n dict[key] = value\n return dict\n\n\ndef data_list_to_str(datalist):\n str = ''\n if datalist is not None:\n for d in datalist:\n str += '&' + d[0]\n str = str[1:]\n return bytes(str, 'utf-8')\n\n\ndef call(reqdict):\n request = urllib.request.Request(reqdict['url'],\n data=reqdict['data'],\n headers=reqdict['headers'])\n request.method = reqdict['method']\n\n try:\n response = urllib.request.urlopen(request)\n status = response.getcode()\n rawcontent = response.read()\n try:\n content = rawcontent.decode('utf-8')\n except UnicodeDecodeError:\n content = gzip.decompress(rawcontent).decode('utf-8')\n return status, content\n except urllib.error.HTTPError as error:\n return error.code, ''\n\n\ndef write_to_file(filename, content):\n f = open(filename, 'w')\n f.write(content)\n f.close()\n\n\ndef iterate_headers(reqdict, repdict):\n i = 0\n iteratedict = copy.deepcopy(reqdict)\n iteratedict['headers'] = {}\n status, content = call(iteratedict)\n if repdict['ref']['status'] == status \\\n and repdict['ref']['length'] == len(content):\n return\n\n for k, v in reqdict['headers'].items():\n i += 1\n repdict[i] = {}\n iteratedict = copy.deepcopy(reqdict)\n iteratedict['headers'].pop(k)\n print_info('Check Header combination: ' + str(iteratedict['headers']))\n status, content = call(iteratedict)\n\n if repdict['ref']['status'] == status \\\n and repdict['ref']['length'] == len(content):\n repdict[i]['necessary'] = False\n else:\n repdict[i]['necessary'] = True\n repdict[i]['length'] = len(content)\n repdict[i]['status'] = status\n repdict[i]['content'] = content\n repdict[i]['header'] = k\n repdict[i]['headervalue'] = v\n\n\ndef generate_report(reqdict, repdict):\n output = 'curl -X ' + reqdict['method'] \\\n + ' \"' + reqdict['url'] + '\"'\n\n if len(repdict) == 1:\n print_info('Header is not needed ¯\\\\_(ツ)_/¯')\n else:\n requireheader = ''\n for i in range(1, len(repdict)):\n if repdict[i]['necessary'] is True:\n requireheader += ', ' + repdict[i]['header']\n output += ' -H \\'' + repdict[i]['header'] \\\n + ': ' + repdict[i]['headervalue'] + '\\''\n print_info('Must-have header(s): \\033[93m\\033[1m'\n + requireheader[2:] + '\\033[0m\\033[0m')\n\n if len(reqdict['data']) > 0:\n output += ' --data \\'' + reqdict['data'].decode('utf-8') + '\\''\n\n if len(reqdict['options']) > 0:\n for o in reqdict['options']:\n output += ' ' + o\n\n print(output)\n write_to_file(str(int(time.time())) + '.curl', output)\n\n\ndef main():\n requestdict = {}\n reportdict = {}\n\n args, rest = parseArgs()\n\n if args.header is None:\n print_error('Missing header!')\n\n try:\n rest.remove('curl')\n except ValueError:\n pass\n\n requestdict['url'] = get_url(rest)\n requestdict['method'] = get_method(args)\n requestdict['data'] = data_list_to_str(args.data)\n requestdict['options'] = get_rest_options(rest)\n requestdict['headers'] = header_list_to_dict(args.header)\n requestdict['rawheaders'] = args.header\n\n refstatus, refcontent = call(requestdict)\n reportdict['ref'] = {}\n reportdict['ref']['status'] = refstatus\n reportdict['ref']['length'] = len(refcontent)\n reportdict['ref']['content'] = refcontent\n reportdict['ref']['headers'] = requestdict['headers']\n\n iterate_headers(requestdict, reportdict)\n generate_report(requestdict, reportdict)\n\n\nif __name__ == '__main__':\n try:\n main()\n except KeyboardInterrupt:\n pass\n","repo_name":"KevCui/beheader","sub_path":"beheader.py","file_name":"beheader.py","file_ext":"py","file_size_in_byte":5385,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"81"}
+{"seq_id":"11650325015","text":"\"\"\"\nerror_logger.py\nThis file contains the error logger function.\n\"\"\"\n\nfrom datetime import date\nERROR_PATH = \"./error.txt\"\n\n\ndef log_error(message: str) -> None:\n \"\"\"\n This function logs errors with the current date and the given error to an error file\n :param message: The message that describes the error\n :return: None\n \"\"\"\n with open(ERROR_PATH, mode='a', encoding='utf-8') as error_file:\n error_file.write(f\"{date.today()} {message} \\n\")\n\nif __name__ == '__main__':\n log_error(\"test\")","repo_name":"jgliao248/15_Puzzle_Turtle","sub_path":"Controller/error_logger.py","file_name":"error_logger.py","file_ext":"py","file_size_in_byte":518,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"43356464456","text":"import math\n\n\nclass Metrics:\n\n @staticmethod\n def accuracy(testSet, predictions):\n correct = 0\n for x in range(len(testSet)):\n if testSet[x][-1] == predictions[x]:\n correct += 1\n return (correct / float(len(testSet))) * 100.0\n\n @staticmethod\n def variance(data, ddof=0):\n n = len(data)\n mean = sum(data) / n\n return sum((x - mean) ** 2 for x in data) / (n - ddof)\n\n @staticmethod\n def stdev(data):\n var = Metrics.variance(data)\n std_dev = math.sqrt(var)\n return std_dev\n\n @staticmethod\n def mean(data):\n return sum(data) / len(data)\n\n @staticmethod\n def confusion_matrix(actual, predicted, normalize = False):\n \"\"\"\n Generate a confusion matrix for multiple classification\n @params:\n actual - a list of integers or strings for known classes\n predicted - a list of integers or strings for predicted classes\n normalize - optional boolean for matrix normalization\n @return:\n matrix - a 2-dimensional list of pairwise counts\n \"\"\"\n unique = sorted(set(actual))\n matrix = [[0 for _ in unique] for _ in unique]\n imap = {key: i for i, key in enumerate(unique)}\n # Generate Confusion Matrix\n for p, a in zip(predicted, actual):\n matrix[imap[p]][imap[a]] += 1\n # Matrix Normalization\n if normalize:\n sigma = sum([sum(matrix[imap[i]]) for i in unique])\n matrix = [row for row in map(lambda i: list(map(lambda j: j / sigma, i)), matrix)]\n\n return matrix\n","repo_name":"jozecarlos/mestrado","sub_path":" APRENDIZAGEM_DE_MÁQUINA/util/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":1651,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"1235710166","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 12 15:49:58 CEST 2017\n\n@author: edouard.duchesnay@cea.fr\n\"\"\"\n\nimport os\nimport json\nimport numpy as np\nimport itertools\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.metrics import recall_score, roc_auc_score, precision_recall_fscore_support\nimport parsimony.estimators as estimators\nimport parsimony.algorithms as algorithms\nimport parsimony.utils as utils\nfrom parsimony.utils.linalgs import LinearOperatorNesterov\nfrom scipy.stats import binom_test\nfrom collections import OrderedDict\nfrom sklearn import preprocessing\nimport pandas as pd\nimport shutil\nfrom brainomics import array_utils\nimport mapreduce\nfrom statsmodels.stats.inter_rater import fleiss_kappa\n\nWD = \"/neurospin/brainomics/2013_adni/MCIc-CTL-FS_cs_all\"\nWD_CLUSTER = WD.replace(\"/neurospin/\", \"/mnt/neurospin/sel-poivre/\")\nWD_ORIGINAL = \"/neurospin/brainomics/2013_adni/MCIc-CTL-FS_cs_modselectcv\"\nuser_func_filename = \"/home/ed203246/git/scripts/2013_adni/MCIc-CTL-FS/03_predict_cs.py\"\n\ndef config_filename(): return os.path.join(WD,\"config_dcv_reducedrange.json\")\ndef results_filename(): return os.path.join(WD,os.path.basename(WD) + \"_dcv.xlsx\")\n\nNFOLDS_OUTER = 5\nNFOLDS_INNER = 5\npenalty_start = 2\nDATA_TYPE = \"mesh\"\n# DATA_TYPE = \"image\"\n\nlambda_max_A = 8.999\n\n##############################################################################\ndef init():\n INPUT_DATA_X = os.path.join(WD_ORIGINAL, 'X.npy')\n INPUT_DATA_y = os.path.join(WD_ORIGINAL, 'y.npy')\n INPUT_MASK_PATH = os.path.join(WD_ORIGINAL, 'mask.npy')\n INPUT_MESH_PATH = '/neurospin/brainomics/2013_adni/MCIc-CTL-FS_cs/lrh.pial.gii'\n #INPUT_LINEAR_OPE_PATH = '/neurospin/brainomics/2016_schizConnect/analysis/NUSDAST/Freesurfer/data/30yo/Atv.npz'\n # INPUT_CSV = '/neurospin/brainomics/2016_schizConnect/analysis/NUSDAST/Freesurfer/population_30yo.csv'\n\n os.makedirs(WD, exist_ok=True)\n shutil.copy(INPUT_DATA_X, WD)\n shutil.copy(INPUT_DATA_y, WD)\n shutil.copy(INPUT_MASK_PATH, WD)\n shutil.copy(INPUT_MESH_PATH, WD)\n\n #shutil.copy(INPUT_LINEAR_OPE_PATH, WD)\n\n ## Create config file\n os.chdir(WD)\n X = np.load(\"X.npy\")\n y = np.load(\"y.npy\")\n\n if not os.path.exists(os.path.join(WD, \"Atv.npz\")):\n import brainomics.mesh_processing as mesh_utils\n cor, tri = mesh_utils.mesh_arrays(os.path.join(WD, \"lrh.pial.gii\"))\n mask = np.load(os.path.join(WD, 'mask.npy'))\n\n import parsimony.functions.nesterov.tv as nesterov_tv\n from parsimony.utils.linalgs import LinearOperatorNesterov\n Atv = nesterov_tv.linear_operator_from_mesh(cor, tri, mask, calc_lambda_max=True)\n Atv.save(os.path.join(WD, \"Atv.npz\"))\n Atv_ = LinearOperatorNesterov(filename=os.path.join(WD, \"Atv.npz\"))\n assert Atv.get_singular_values(0) == Atv_.get_singular_values(0)\n assert np.allclose(Atv_.get_singular_values(0), 8.999, rtol=1e-03, atol=1e-03)\n assert np.all([a.shape == (317089, 317089) for a in Atv])\n\n if not os.path.exists(os.path.join(WD, \"beta_start.npz\")):\n betas = dict()\n import time\n alphas = [.01, 0.1, 1.0, 10]\n for alpha in alphas:\n mod = estimators.RidgeLogisticRegression(l=alpha, class_weight=\"auto\", penalty_start=penalty_start)\n t_ = time.time()\n mod.fit(X, y.ravel())\n print(time.time() - t_) # 11564\n betas[\"lambda_%.2f\" % alpha] = mod.beta\n\n np.savez(os.path.join(WD, \"beta_start.npz\"), **betas)\n beta_start = np.load(os.path.join(WD, \"beta_start.npz\"))\n assert np.all([np.all(beta_start[a] == betas[a]) for a in beta_start.keys()])\n\n ## Create config file\n\n # ########################################################################\n # Setting 1: 5cv + large range of parameters: cv_largerange\n # with sub-sample training set with size 50, 100\n # 5cv/cv0*[_sub50]/refit/*\n\n # sub_sizes = [50, 100]\n sub_sizes = []\n\n cv_outer = [[tr, te] for tr, te in\n StratifiedKFold(n_splits=NFOLDS_OUTER, random_state=42).split(np.zeros(y.shape[0]), y.ravel())]\n\n # check we got the same CV than previoulsy\n cv_old = json.load(open(os.path.join(WD_ORIGINAL, \"config_modselectcv.json\")))[\"resample\"]\n cv_outer_old = [cv_old[k] for k in ['cv%02d/refit' % i for i in range(NFOLDS_OUTER)]]\n assert np.all([np.all(np.array(cv_outer_old[i][0]) == cv_outer[i][0]) for i in range(NFOLDS_OUTER)])\n assert np.all([np.all(np.array(cv_outer_old[i][1]) == cv_outer[i][1]) for i in range(NFOLDS_OUTER)])\n # check END\n\n import collections\n cv = collections.OrderedDict()\n\n cv[\"refit/refit\"] = [np.arange(len(y)), np.arange(len(y))]\n\n for cv_outer_i, (tr_val, te) in enumerate(cv_outer):\n # Simple CV\n cv[\"cv%02d/refit\" % (cv_outer_i)] = [tr_val, te]\n\n # Nested CV\n # cv_inner = StratifiedKFold(y[tr_val].ravel(), n_folds=NFOLDS_INNER, random_state=42)\n # for cv_inner_i, (tr, val) in enumerate(cv_inner):\n # cv[\"cv%02d/cvnested%02d\" % ((cv_outer_i), cv_inner_i)] = [tr_val[tr], tr_val[val]]\n\n # Sub-sample training set with size 50, 100\n # => cv*_sub[50|100]/refit\n grps = np.unique(y[tr_val]).astype(int)\n ytr = y.copy()\n ytr[te] = np.nan\n g_idx = [np.where(ytr == g)[0] for g in grps]\n assert np.all([np.all(ytr[g_idx[g]] == g) for g in grps])\n\n g_size = np.array([len(g) for g in g_idx])\n g_prop = g_size / g_size.sum()\n\n for sub_size in sub_sizes:\n # sub_size = sub_sizes[0]\n sub_g_size = np.round(g_prop * sub_size).astype(int)\n g_sub_idx = [np.random.choice(g_idx[g], sub_g_size[g], replace=False) for g in grps]\n assert np.all([np.all(y[g_sub_idx[g]] == g) for g in grps])\n tr_val_sub = np.concatenate(g_sub_idx)\n assert len(tr_val_sub) == sub_size\n assert np.all([idx in tr_val for idx in tr_val_sub])\n assert np.all(np.logical_not([idx in te for idx in tr_val_sub]))\n cv[\"cv%02d_sub%i/refit\" % (cv_outer_i, sub_size)] = [tr_val_sub, te]\n\n cv = {k:[cv[k][0].tolist(), cv[k][1].tolist()] for k in cv}\n\n # Nested CV\n # assert len(cv_largerange) == NFOLDS_OUTER * NFOLDS_INNER + NFOLDS_OUTER + 1\n\n # Simple CV\n # assert len(cv) == NFOLDS_OUTER + 1\n\n # Simple CV + sub-sample training set with size 50, 100:\n assert len(cv) == NFOLDS_OUTER * (1 + len(sub_sizes)) + 1\n\n print(list(cv.keys()))\n\n # Large grid of parameters\n alphas = [0.001, 0.01, 0.1, 1.0]\n # alphas = [.01, 0.1, 1.0] # first ran with this grid\n tv_ratio = [0.01, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]\n l1l2_ratio = [0.1, 0.5, 0.9]\n # l1l2_ratio = [0, 0.1, 0.5, 0.9, 1.0] # first ran with this grid\n algos = [\"enettv\", \"enetgn\"]\n params_enet_tvgn = [list(param) for param in itertools.product(algos, alphas, l1l2_ratio, tv_ratio)]\n assert len(params_enet_tvgn) == 240 # old 300\n\n params_enet = [list(param) for param in itertools.product([\"enet\"], alphas, l1l2_ratio, [0])]\n assert len(params_enet) == 12 # old 15\n\n params = params_enet_tvgn + params_enet\n assert len(params) == 252 # 315\n # Simple CV\n # assert len(params) * len(cv) == 1890\n\n # Simple CV + sub-sample training set with size 50, 100:\n assert len(params) * len(cv) == 1512 # 5040\n\n config = dict(data=dict(X=\"X.npy\", y=\"y.npy\"),\n params=params, resample=cv,\n structure_linear_operator_tv=\"Atv.npz\",\n beta_start=\"beta_start.npz\",\n map_output=\"5cv\",\n user_func=user_func_filename)\n json.dump(config, open(os.path.join(WD, \"config_cv_largerange.json\"), \"w\"))\n\n\n # Build utils files: sync (push/pull) and PBS\n import brainomics.cluster_gabriel as clust_utils\n cmd = \"mapreduce.py --map %s/config_cv_largerange.json\" % WD_CLUSTER\n clust_utils.gabriel_make_qsub_job_files(WD, cmd,walltime = \"250:00:00\",\n suffix=\"_cv_largerange\",\n freecores=2)\n\n # ########################################################################\n # Setting 2: dcv + reduced range of parameters: dcv_reducedrange\n # 5cv/cv0*/cvnested0*/*\n\n cv_outer = [[tr, te] for tr, te in\n StratifiedKFold(n_splits=NFOLDS_OUTER, random_state=42).split(np.zeros(y.shape[0]), y.ravel())]\n\n # check we got the same CV than previoulsy\n cv_old = json.load(open(os.path.join(WD_ORIGINAL, \"config_modselectcv.json\")))[\"resample\"]\n cv_outer_old = [cv_old[k] for k in ['cv%02d/refit' % i for i in range(NFOLDS_OUTER)]]\n assert np.all([np.all(np.array(cv_outer_old[i][0]) == cv_outer[i][0]) for i in range(NFOLDS_OUTER)])\n assert np.all([np.all(np.array(cv_outer_old[i][1]) == cv_outer[i][1]) for i in range(NFOLDS_OUTER)])\n # check END\n\n import collections\n cv = collections.OrderedDict()\n cv[\"refit/refit\"] = [np.arange(len(y)), np.arange(len(y))]\n\n for cv_outer_i, (tr_val, te) in enumerate(cv_outer):\n cv[\"cv%02d/refit\" % (cv_outer_i)] = [tr_val, te]\n cv_inner = StratifiedKFold(n_splits=NFOLDS_INNER, random_state=42).split(np.zeros(y[tr_val].shape[0]), y[tr_val].ravel())\n for cv_inner_i, (tr, val) in enumerate(cv_inner):\n cv[\"cv%02d/cvnested%02d\" % ((cv_outer_i), cv_inner_i)] = [tr_val[tr], tr_val[val]]\n\n cv = {k:[cv[k][0].tolist(), cv[k][1].tolist()] for k in cv}\n #assert len(cv) == NFOLDS_OUTER + 1\n assert len(cv) == NFOLDS_OUTER * NFOLDS_INNER + NFOLDS_OUTER + 1\n print(list(cv.keys()))\n\n # Reduced grid of parameters\n alphas = [0.001, 0.01, 0.1, 1.0]\n # alphas = [.01, 0.1] # original\n tv_ratio = [0.2, 0.8]\n l1l2_ratio = [0.1, 0.9]\n algos = [\"enettv\", \"enetgn\"]\n params_enet_tvgn = [list(param) for param in itertools.product(algos, alphas, l1l2_ratio, tv_ratio)]\n assert len(params_enet_tvgn) == 32 # 16\n\n params_enet = [list(param) for param in itertools.product([\"enet\"], alphas, l1l2_ratio, [0])]\n assert len(params_enet) == 8 # 4\n\n params = params_enet_tvgn + params_enet\n assert len(params) == 40 # 20\n assert len(params) * len(cv) == 1240 # 620\n\n config = dict(data=dict(X=\"X.npy\", y=\"y.npy\"),\n params=params, resample=cv,\n structure_linear_operator_tv=\"Atv.npz\",\n beta_start=\"beta_start.npz\",\n map_output=\"5cv\",\n user_func=user_func_filename)\n json.dump(config, open(os.path.join(WD, \"config_dcv_reducedrange.json\"), \"w\"))\n\n # Build utils files: sync (push/pull) and PBS\n import brainomics.cluster_gabriel as clust_utils\n cmd = \"mapreduce.py --map %s/config_dcv_reducedrange.json\" % WD_CLUSTER\n clust_utils.gabriel_make_qsub_job_files(WD, cmd,walltime = \"250:00:00\",\n suffix=\"_dcv_reducedrange\",\n freecores=2)\n\n\n#############################################################################\ndef load_globals(config):\n import scipy.sparse as sparse\n import functools\n import mapreduce as GLOBAL # access to global variables\n\n GLOBAL.DATA = GLOBAL.load_data(config[\"data\"])\n\n Atv = LinearOperatorNesterov(filename=config[\"structure_linear_operator_tv\"])\n Agn = sparse.vstack(Atv)\n Agn.singular_values = Atv.get_singular_values()\n def get_singular_values(self, nb=None):\n return self.singular_values[nb] if nb is not None else self.singular_values\n Agn.get_singular_values = functools.partial(get_singular_values, Agn)\n assert np.allclose(Agn.get_singular_values(0), lambda_max_A, rtol=1e-03, atol=1e-03)\n GLOBAL.Atv, GLOBAL.Agn = Atv, Agn\n\n # npz = np.load(config[\"beta_start\"])\n # GLOBAL.beta_start = {k:npz[k] for k in npz}\n\ndef resample(config, resample_nb):\n import mapreduce as GLOBAL # access to global variables\n GLOBAL.DATA = GLOBAL.load_data(config[\"data\"])\n resample = config[\"resample\"][resample_nb]\n GLOBAL.DATA_RESAMPLED = {k: [GLOBAL.DATA[k][idx, ...] for idx in resample]\n for k in GLOBAL.DATA}\n\ndef mapper(key, output_collector):\n \"\"\"\n # debug mapper\n config = json.load(open(os.path.join(WD, \"config_cv_largerange.json\"), \"r\"))\n load_globals(config)\n resample(config, 'refit/refit')\n key = ('enettv', 0.01, 0.1, 0.3)\n \"\"\"\n import mapreduce as GLOBAL\n Xtr = GLOBAL.DATA_RESAMPLED[\"X\"][0]\n Xte = GLOBAL.DATA_RESAMPLED[\"X\"][1]\n ytr = GLOBAL.DATA_RESAMPLED[\"y\"][0]\n yte = GLOBAL.DATA_RESAMPLED[\"y\"][1]\n\n # key = 'enettv_0.01_0.1_0.2'.split(\"_\")\n algo, alpha, l1l2ratio, tvratio = key[0], float(key[1]), float(key[2]), float(key[3])\n\n tv = alpha * tvratio\n l1 = alpha * float(1 - tv) * l1l2ratio\n l2 = alpha * float(1 - tv) * (1- l1l2ratio)\n\n print(key, algo, alpha, l1, l2, tv)\n # alpha = float(key[0])\n # l1, l2, tv = alpha * float(key[1]), alpha * float(key[2]), alpha * float(key[3])\n # print(\"l1:%f, l2:%f, tv:%f\" % (l1, l2, tv))\n\n class_weight = \"auto\" # unbiased\n\n # beta_start = GLOBAL.beta_start[\"lambda_%.4f\" % alpha]\n # mask = np.ones(Xtr.shape[0], dtype=bool)\n\n # scaler = preprocessing.StandardScaler().fit(Xtr)\n # Xtr = scaler.transform(Xtr)\n # Xte = scaler.transform(Xte)\n if algo == 'enettv':\n conesta = algorithms.proximal.CONESTA(max_iter=10000)\n mod = estimators.LogisticRegressionL1L2TV(l1, l2, tv, GLOBAL.Atv,\n algorithm=conesta, class_weight=class_weight, penalty_start=penalty_start)\n elif algo == 'enetgn':\n fista = algorithms.proximal.FISTA(max_iter=5000) # original 500\n mod = estimators.LogisticRegressionL1L2GraphNet(l1, l2, tv, GLOBAL.Agn,\n algorithm=fista, class_weight=class_weight, penalty_start=penalty_start)\n elif algo == 'enet':\n fista = algorithms.proximal.FISTA(max_iter=5000) # original 500\n mod = estimators.ElasticNetLogisticRegression(l1l2ratio, alpha,\n algorithm=fista, class_weight=class_weight, penalty_start=penalty_start)\n else:\n raise Exception('Algo%s not handled' %algo)\n\n mod.fit(Xtr, ytr.ravel())\n y_pred = mod.predict(Xte)\n proba_pred = mod.predict_probability(Xte)\n ret = dict(y_pred=y_pred, y_true=yte, proba_pred=proba_pred, beta=mod.beta)#, mask=mask)\n if output_collector:\n output_collector.collect(key, ret)\n else:\n return ret\n\n#############################################################################\n# Reducer\ntry:\n REDUCER_SRC = '/home/ed203246/git/brainomics-team/2017_logistic_nestv/scripts/reduce_plot_vizu.py'\n exec(open(REDUCER_SRC).read())\nexcept:\n pass\n\ndef do_reducer():\n output_filename = results_filename()\n # reducer(WD=WD, output_filename=output_filename, force_recompute=False)\n model = estimators.LogisticRegression()\n # reducer(WD=WD, output_filename=output_filename, force_recompute=True, model=model, rescale=False)\n reducer(WD=WD, output_filename=output_filename, force_recompute=False, model=model, rescale=False)\n\n\n\"\"\"\ndef reducer(key=None, values=None):\n\n import os, glob, pandas as pd\n def close(vec, val, tol=1e-4):\n return np.abs(vec - val) < tol\n\n def groupby_paths(paths, pos):\n groups = {g:[] for g in set([p.split(\"/\")[pos] for p in paths])}\n for p in paths:\n groups[p.split(\"/\")[pos]].append(p)\n return groups\n\n def scores_groupby_paths(paths, param_pos, algo_pos_in_params, score_func):\n byparams = groupby_paths(paths, param_pos)\n # key='enettv_0.1_0.1_0.2'; paths=byparams[key]; algo_idx=algo_pos_in_params\n byparams_scores = {k:score_func(k, v, config, algo_idx=algo_pos_in_params) for k, v in byparams.items()}\n byparams_scores = {k: v for k, v in byparams_scores.items() if v is not None}\n data = [list(byparams_scores[k].values()) for k in byparams_scores]\n columns = list(byparams_scores[list(byparams_scores.keys())[0]].keys())\n return pd.DataFrame(data, columns=columns)\n\n def argmaxscore_bygroup(data, groupby='fold', param_key=\"key\",\n score=\"bacc\",\n refit_key=None, # Do refit ?\n config=None, # required for refit\n score_func=None, # required for refit\n algo_pos_in_params=None # required for refit\n ):\n arg_max_byfold = list()\n for fold, data_fold in data.groupby(groupby):\n assert len(data_fold) == len(set(data_fold[param_key])) # ensure all param are diff\n arg_max_byfold.append([fold, data_fold.ix[data_fold[score].argmax()][param_key], data_fold[score].max()])\n arg_max_byfold = pd.DataFrame(arg_max_byfold, columns=[groupby, param_key, score])\n arg_max_byfold[\"key_refit\"] = refit_key\n if refit_key is not None:\n refit = score_func(refit_key,\n [os.path.join(config['map_output'], row[groupby], \"refit\", row[param_key])\n for index, row in arg_max_byfold.iterrows()],\n config, as_dataframe=True, algo_idx=algo_pos_in_params)\n else:\n refit = None\n\n return arg_max_byfold, refit\n\n # config = json.load(open(config_filename()))\n # paths = glob.glob(os.path.join(WD, \"5cv\", \"*\", \"*\", \"*\"))\n # param_config_set = set([mapreduce.dir_from_param_list(p) for p in config['params']])\n # assert len(paths) / len(param_config_set) == len(config['resample']), \"Nb run per param is not the one excpected\"\n\n # config_cv_largerange\n s = 'tv_ratio'\n os.chdir(WD)\n config = json.load(open(\"config_cv_largerange.json\"))\n paths_all = glob.glob(\"5cv/cv0?/refit/*\")\n paths_all.sort()\n # paths_sub50 = glob.glob(\"5cv/cv0?_sub50/refit/*\")\n # paths_sub50.sort()\n # paths_sub100 = glob.glob(\"5cv/cv0?_sub100/refit/*\")\n # paths_sub100.sort()\n\n #assert len(paths) == 4286\n print('## Refit scores: cv*/refit/*')\n print('## -------------------------')\n scores_refit = scores_groupby_paths(paths=paths_all, param_pos=3, algo_pos_in_params=0, score_func=scores)\n #scores_refit_sub50 = scores_groupby_paths(paths=paths_sub50, param_pos=3, algo_pos_in_params=0, score_func=scores)\n #scores_refit_sub100 = scores_groupby_paths(paths=paths_sub100, param_pos=3, algo_pos_in_params=0, score_func=scores)\n\n # with pd.ExcelWriter(os.path.join(WD, \"results_refit_cv_by_param_largerange.xlsx\")) as writer:\n # scores_refit.to_excel(writer, sheet_name='cv_by_param_all', index=False)\n #scores_refit_sub100.to_excel(writer, sheet_name='cv_by_param_sub100', index=False)\n #scores_refit_sub50.to_excel(writer, sheet_name='cv_by_param_sub50', index=False)\n\n\n print('## doublecv scores by outer-cv and by params: cv*/cvnested*/*')\n print('## -----------------------------------------')\n paths = glob.glob(\"5cv/cv0?/cvnested0?/*\")\n paths.sort()\n bycv = groupby_paths(paths, 1)\n scores_dcv_byparams = None\n for fold, paths_fold in bycv.items():\n print(fold)\n scores_dcv_fold = scores_groupby_paths(paths=paths_fold, param_pos=3, algo_pos_in_params=0, score_func=scores)\n scores_dcv_fold[\"fold\"] = fold\n scores_dcv_byparams = pd.concat([scores_dcv_byparams, scores_dcv_fold])\n\n print([[g, d.shape[0]] for g, d in scores_dcv_byparams.groupby([\"fold\", \"algo\"])])\n # assert np.all(np.array([g.shape[0] for d, g in scores_dcv_byparams.groupby('fold')]) == 136)\n\n # Different settings\n results = list()\n for algo in [\"enettv\", \"enetgn\"]:\n # algo = \"enettv\"\n results.append(argmaxscore_bygroup(data=scores_dcv_byparams,\n refit_key=\"%s_dcv-all\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2s_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.1) | close(scores_dcv_byparams.l1_ratio, 0.9)) &\n (close(scores_dcv_byparams[s], 0.2) | close(scores_dcv_byparams[s], 0.8))]\n # assert np.all(np.array([g.shape[0] for d, g in l1l2s_reduced.groupby('fold')]) == 8)\n # assert l1l2s_reduced.shape[0] == 40\n results.append(argmaxscore_bygroup(data=l1l2s_reduced,\n refit_key=\"%s_dcv-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2s_ridge_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.1)) &\n (close(scores_dcv_byparams[s], 0.2) | close(scores_dcv_byparams[s], 0.8))]\n # assert np.all(np.array([g.shape[0] for d, g in l1l2s_ridge_reduced.groupby('fold')]) == 4)\n # assert l1l2s_ridge_reduced.shape[0] == 20\n results.append(argmaxscore_bygroup(data=l1l2s_ridge_reduced,\n refit_key=\"%s_dcv-ridge-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2s_ridge_reduced2 = l1l2s_ridge_reduced[close(l1l2s_ridge_reduced[s], 0.8)] # FS 0.8\n results.append(argmaxscore_bygroup(data=l1l2s_ridge_reduced2,\n refit_key=\"%s_dcv-ridge-reduced2\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2s_lasso_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.9)) &\n (close(scores_dcv_byparams[s], 0.2) | close(scores_dcv_byparams[s], 0.8))]\n # assert np.all(np.array([g.shape[0] for d, g in l1l2s_lasso_reduced.groupby('fold')]) == 4)\n # assert l1l2s_lasso_reduced.shape[0] == 20\n results.append(argmaxscore_bygroup(data=l1l2s_lasso_reduced,\n refit_key=\"%s_dcv-lasso-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2s_lasso_reduced2 = l1l2s_lasso_reduced[close(l1l2s_lasso_reduced[s], 0.8)] # FS 0.8\n results.append(argmaxscore_bygroup(data=l1l2s_lasso_reduced2,\n refit_key=\"%s_dcv-lasso-reduced2\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n algo = \"enet\"\n l1l2_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.1) | close(scores_dcv_byparams.l1_ratio, 0.9)) &\n (close(scores_dcv_byparams[s], 0))]\n assert np.all(np.array([g.shape[0] for d, g in l1l2_reduced.groupby('fold')]) == 4)\n assert l1l2_reduced.shape[0] == 20\n results.append(argmaxscore_bygroup(data=l1l2_reduced,\n refit_key=\"%s_dcv-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2_ridge_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.1)) &\n (close(scores_dcv_byparams[s], 0))]\n assert np.all(np.array([g.shape[0] for d, g in l1l2_ridge_reduced.groupby('fold')]) == 2)\n assert l1l2_ridge_reduced.shape[0] == 10\n results.append(argmaxscore_bygroup(data=l1l2_ridge_reduced,\n refit_key=\"%s_dcv-ridge-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n l1l2_lasso_reduced = scores_dcv_byparams[\n (scores_dcv_byparams.algo == algo) &\n (close(scores_dcv_byparams.a, 0.01) | close(scores_dcv_byparams.a, 0.1)) &\n (close(scores_dcv_byparams.l1_ratio, 0.9)) &\n (close(scores_dcv_byparams[s], 0))]\n assert np.all(np.array([g.shape[0] for d, g in l1l2_lasso_reduced.groupby('fold')]) == 2)\n assert l1l2_lasso_reduced.shape[0] == 10\n results.append(argmaxscore_bygroup(data=l1l2_lasso_reduced,\n refit_key=\"%s_dcv-lasso-reduced\" % algo, config=config, score_func=scores, algo_pos_in_params=0))\n\n cv_argmax, scores_dcv = zip(*results)\n cv_argmax = pd.concat(cv_argmax)\n scores_dcv = pd.concat(scores_dcv)\n\n with pd.ExcelWriter(results_filename()) as writer:\n scores_refit.to_excel(writer, sheet_name='cv_by_param', index=False)\n scores_dcv_byparams.to_excel(writer, sheet_name='cv_cv_byparam', index=False)\n cv_argmax.to_excel(writer, sheet_name='cv_argmax', index=False)\n scores_dcv.to_excel(writer, sheet_name='dcv', index=False)\n\n\"\"\"\n\n\n###############################################################################\n# copy old results to new organization\nimport glob\n\ndef dir_from_param_list(param_list):\n return \"_\".join([str(p) for p in param_list])\n\n\ndef param_src_to_dst(param_src, dst_prefix=None, dst_suffix=None):\n a, l1, l2, tv, k = [float(a) for a in param_src.split(\"_\")]\n l1_ratio = l1 / (l1+l2)\n param_list_dst = [a, l1_ratio, tv]\n if dst_prefix is not None:\n param_list_dst = [dst_prefix] + param_list_dst\n if dst_suffix is not None:\n param_list_dst = param_list_dst + [dst_suffix]\n return dir_from_param_list(param_list_dst)\n\n\ndef copy_results(SRC, DST, outer_str, inner_str, dst_prefix, copy=True):\n for fold in itertools.product(outer_str, inner_str):\n src = SRC % fold\n dst = DST % fold\n for param_src in [os.path.basename(p) for p in glob.glob(src + \"/*\")]:\n #param_src = \"0.01_0.008_0.792_0.2_-1\"\n #param_src = \"0.01_0.02_0.18_0.8_-1\"\n #param_src = \"0.1_0.72_0.08_0.2_-1\"\n param_dst = param_src_to_dst(param_src, dst_prefix=dst_prefix)\n path_src = os.path.join(src, param_src, \"beta.npz\")\n path_dst = os.path.join(dst, param_dst, \"beta.npz\")\n if os.path.exists(path_src) and os.path.exists(path_dst):\n cor = None\n beta_src = np.load(path_src)['arr_0']\n beta_dst = np.load(path_dst)['arr_0']\n cor = np.corrcoef(beta_dst.ravel(), beta_src.ravel())[0, 1]\n print(fold, \"\\t\", param_src, \"\\t\", param_dst, \"\\t\",cor)\n elif os.path.exists(path_src) and not os.path.exists(path_dst):\n print(path_src, \"\\n\", path_dst)\n if copy:\n shutil.copytree(\n os.path.dirname(path_src),\n os.path.dirname(path_dst))\n\ndef do_copy_results():\n outer_str = [\"cv%02d\" % i for i in range(5)]\n inner_str = [\"cvnested%02d\" % i for i in range(5)] + [\"refit\"]\n SRC = \"/neurospin/brainomics/2013_adni/MCIc-CTL-FS_cs_modselectcv/modselectcv/%s/%s\"\n DST = \"/neurospin/brainomics/2013_adni/MCIc-CTL-FS_cs_all/5cv/%s/%s\"\n copy_results(SRC, DST, outer_str, inner_str, dst_prefix=\"enettv\")\n\n\n import parsimony.utils.weights as weights\n start_vector=weights.RandomUniformWeights(normalise=True)\n start_vector.get_weights(10)","repo_name":"neurospin/scripts","sub_path":"2013_adni/MCIc-CTL-FS/03_enetall_ADNI-MCIc-CTL-FS_cs.py","file_name":"03_enetall_ADNI-MCIc-CTL-FS_cs.py","file_ext":"py","file_size_in_byte":27190,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"42904477289","text":"def find_files(suffix, path):\n \"\"\" \n Find all files beneath path with file name suffix.\n\n Note that a path may contain further subdirectories\n and those subdirectories may also contain further subdirectories.\n\n There are no limit to the depth of the subdirectories can be.\n\n Args:\n suffix(str): suffix if the file name to be found\n path(str): path of the file system\n\n Returns:\n a list of paths\n \"\"\"\n files = []\n #files = os.listdir(path)\n try:\n for filename in os.listdir(path):\n if os.path.isdir(path+\"/\"+filename):\n files.extend(find_files(suffix, path+\"/\"+filename))\n if filename.endswith(suffix,len(filename)-3):\n files.append(filename)\n except:\n print(\"No path provided\")\n return\n return files\n\n## Locally save and call this file ex.py ##\n# Code to demonstrate the use of some of the OS modules in python\n\nimport os\n\n# Let us print the files in the directory in which you are running this script\nprint(os.listdir(\".\"))\n\n# Let us check if this file is indeed a file!\nprint(os.path.isfile(\"./ex.py\"))\n\n# Does the file end with .py?\nprint(\"./ex.py\".endswith(\".py\"))\nprint(find_files(\".c\", \"./testdir\"))\nprint(find_files(\".c\", \"\"))\nprint(find_files(\".exe\",\"./testdir\"))","repo_name":"reaprman/Data-Struct-algo-nanodegree","sub_path":"proj2/problem2.py","file_name":"problem2.py","file_ext":"py","file_size_in_byte":1273,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"7624788264","text":"\"\"\"\nPython <-> Objective-C bridge (PyObjC)\n\nThis module defines the core interfaces of the Python<->Objective-C bridge.\n\"\"\"\n\n__all__ = ['IBOutlet', 'IBAction', 'accessor', 'Accessor', 'typedAccessor', 'callbackFor', 'selectorFor', 'synthesize', 'namedselector', 'typedSelector', 'namedSelector', 'instancemethod', 'signature' ]\n\nfrom objc._objc import ivar, selector, _makeClosure, selector, _C_SEL, _C_ID, _C_NSUInteger, _C_NSBOOL\nimport sys, textwrap\nimport warnings\nfrom inspect import getargspec\n\n_C_NSRange = [b\"{_NSRange=II}\", b\"{_NSRange=QQ}\"][sys.maxsize > 2**32]\n\n#\n# Interface builder support.\n#\ndef IBOutlet(name=None):\n \"\"\"\n Create an instance variable that can be used as an outlet in\n Interface Builder.\n \"\"\"\n if name is None:\n return ivar(isOutlet=1)\n else:\n return ivar(name, isOutlet=1)\n\ndef IBAction(func):\n \"\"\"\n Return an Objective-C method object that can be used as an action\n in Interface Builder.\n \"\"\"\n if func is None:\n raise TypeError(\"IBAction argument must be a callable\")\n return selector(func, signature=b\"v@:@\")\n\ndef instancemethod(func):\n if func is None:\n raise TypeError(\"instancemethod argument must be a callable\")\n return selector(func, isClassMethod=False)\n\ndef accessor(func, typeSignature=b'@'):\n \"\"\"\n Return an Objective-C method object that is conformant with key-value coding\n and key-value observing.\n \"\"\"\n args, varargs, varkw, defaults = getargspec(func)\n funcName = func.__name__\n maxArgs = len(args)\n minArgs = maxArgs - len(defaults or ())\n # implicit self\n selArgs = 1 + funcName.count('_')\n if varargs is not None or varkw is not None:\n raise TypeError('%s can not be an accessor because it accepts varargs or varkw' % (funcName,))\n\n if not (minArgs <= selArgs <= maxArgs):\n if minArgs == maxArgs:\n raise TypeError('%s expected to take %d args, but must accept %d from Objective-C (implicit self plus count of underscores)' % (funcName, maxArgs, selArgs))\n else:\n raise TypeError('%s expected to take between %d and %d args, but must accept %d from Objective-C (implicit self plus count of underscores)' % (funcName, minArgs, maxArgs, selArgs))\n\n if selArgs == 3:\n if funcName.startswith('validate') and funcName.endswith('_error_'):\n return selector(func, signature=_C_NSBOOL + b'@:N^@o^@')\n\n if funcName.startswith('insertObject_in') and funcName.endswith('AtIndex_'):\n return selector(func, signature=b'v@:' + typeSignature + _C_NSUInteger)\n elif funcName.startswith('replaceObjectIn') and funcName.endswith('AtIndex_withObject_'):\n return selector(func, signature=b'v@:' + _C_NSUInteger + typeSignature)\n\n elif funcName.startswith('get') and funcName.endswith('_range_'):\n return selector(func, signature=b'v@:o^@' + _C_NSRange)\n\n elif funcName.startswith('insert') and funcName.endswith('_atIndexes_'):\n return selector(func, signature=b'v@:@@')\n\n elif funcName.startswith('replace') and 'AtIndexes_with' in funcName:\n return selector(func, signature=b'v@:@@')\n\n # pass through to \"too many arguments\"\n\n elif selArgs == 2:\n if funcName.startswith('objectIn') and funcName.endswith('AtIndex_'):\n return selector(func, signature=typeSignature + b'@:' + _C_NSUInteger)\n elif funcName.startswith('removeObjectFrom') and funcName.endswith('AtIndex_'):\n return selector(func, signature=b'v@:' + _C_NSUInteger)\n elif funcName.startswith('remove') and funcName.endswith('AtIndexes_'):\n return selector(func, signature=b\"v@:@\")\n elif funcName.endswith('AtIndexes_'):\n return selector(func, signature=b\"@@:@\")\n elif funcName.startswith('memberOf'):\n return selector(func, signature=_C_NSBOOL + b\"@:\" + typeSignature)\n elif funcName.startswith('add') and funcName.endswith('Object_'):\n return selector(func, signature=b\"v@:\" + typeSignature)\n elif funcName.startswith('add'):\n return selector(func, signature=b\"v@:@\")\n elif funcName.startswith('intersect'):\n return selector(func, signature=b\"v@:@\")\n\n return selector(func, signature=b\"v@:\" + typeSignature)\n\n elif selArgs == 1:\n if funcName.startswith('countOf'):\n typeSignature = _C_NSUInteger\n elif funcName.startswith('enumerator'):\n typeSignature = b\"@\"\n\n\n return selector(func, signature=typeSignature + b\"@:\")\n\n raise TypeError(\"%s not recognized as an accessor\" % (funcName,))\n\n\ndef typedSelector(signature):\n def _typedSelector(func):\n if func is None:\n raise TypeError(\"typedSelector() function argument must be a callable\")\n return selector(func, signature=signature)\n return _typedSelector\n\ndef namedSelector(name, signature=None):\n \"\"\"\n Decorator for overriding the Objective-C SEL for a method, usage:\n\n @namedSelector(\"foo:bar:\")\n def foobar(self, foo, bar):\n return foo + bar\n \"\"\"\n if signature is not None:\n def _namedselector(func):\n if func is None:\n raise TypeError(\"IBAction argument must be a callable\")\n return selector(func, selector=name, signature=signature)\n else:\n def _namedselector(func):\n if func is None:\n raise TypeError(\"IBAction argument must be a callable\")\n return selector(func, selector=name)\n\n return _namedselector\n\ndef namedselector(name, signature=None):\n warnings.warn(\"use objc.namedSelector instead of objc.namedselector\", DeprecationWarning, stacklevel=2)\n return namedSelector(name, signature)\n\ndef typedAccessor(typeSignature):\n \"\"\"\n Decorator for creating a typed accessor, usage:\n\n @typedAccessor('i')\n def someIntegerAccessor(self):\n return self.someInteger\n\n @typedAccessor('i')\n def setSomeIntegerAccessor_(self, anInteger):\n self.someInteger = anInteger\n \"\"\"\n def _typedAccessor(func):\n return accessor(func, typeSignature)\n return _typedAccessor\n\ndef Accessor(func):\n warnings.warn(\n \"Use objc.accessor instead of objc.Accessor\", DeprecationWarning)\n return accessor(func)\n\n#\n# Callback support\n#\ndef callbackFor(callable, argIndex=-1):\n \"\"\"\n Decorator for converting a function into an object that can be used\n as a callback function for (Objective-)C API's that take such a beast\n as one of their arguments.\n\n Note that using this decorator for methods is unsupported and that this\n decorator is optional when the callback isn't stored by the called function\n\n Usage::\n\n @objc.callbackFor(NSArray.sortedArrayUsingFunction_context_)\n def compare(left, right, context):\n return 1\n \"\"\"\n def addClosure(function):\n closure = _makeClosure(function, callable, argIndex)\n function.pyobjc_closure = closure\n return function\n\n return addClosure\n\ndef selectorFor(callable, argIndex=-1):\n \"\"\"\n Decorator that makes sure that the method has the right signature to be\n used as the selector argument to the specified method.\n\n Usage::\n\n @objc.selectorFor(NSApplication.beginSheet_modalForWindow_modalDelegate_didEndSelector_contextInfo_)\n def sheetDidEnd_returnCode_contextInfo_(self, sheet, returnCode, info):\n pass\n \"\"\"\n if argIndex == -1:\n for arg in callable.__metadata__()['arguments']:\n if arg['type'] == _C_SEL and 'sel_of_type' in arg:\n signature = arg['sel_of_type']\n break\n else:\n raise ValueError(\"No selector argument with type information\")\n\n else:\n try:\n signature = callable.__metadata__()['arguments'][argIndex]['sel_of_type']\n except (IndexError, KeyError):\n raise ValueError(\"Not a selector argument with type information\")\n\n def addSignature(function):\n return selector(function, signature=signature)\n\n return addSignature\n\n\ndef synthesize(name, copy=False, readwrite=True, type=_C_ID, ivarName=None):\n \"\"\"\n Use this in a class dictionary to syntheze simple setting/setter methods.\n\n Note: this is only necessary to get propper behaviour when Key-Value coding\n is used and special features (like copying) are needed\n\n usage::\n\n class MyClass (NSObject):\n objc.synthesize('someTitle', copy=True)\n\n \"\"\"\n if not name:\n raise ValueError(\"Empty property name\")\n\n if ivarName is None:\n ivarName = '_' + name\n\n classDict = sys._getframe(1).f_locals\n\n setterName = 'set%s%s_'%(name[0].upper(), name[1:])\n\n if copy:\n setter = textwrap.dedent('''\n def %(name)s(self, value):\n self.%(ivar)s = value.copy()\n ''' % dict(name=setterName, ivar=ivarName))\n\n else:\n setter = textwrap.dedent('''\n def %(name)s(self, value):\n self.%(ivar)s = value\n ''' % dict(name=setterName, ivar=ivarName))\n\n getter = textwrap.dedent('''\n def %(name)s(self):\n return self.%(ivar)s\n ''' % dict(name=name, ivar=ivarName))\n\n if readwrite:\n exec(setter, globals(), classDict)\n\n exec(getter, globals(), classDict)\n\n classDict[ivarName] = ivar(type=type)\n\n\ndef signature(signature, **kw):\n \"\"\"\n A Python method decorator that allows easy specification\n of Objective-C selectors.\n\n Usage::\n\n @objc.signature('i@:if')\n def methodWithX_andY_(self, x, y):\n return 0\n \"\"\"\n warnings.warn(\"Usage objc.typedSelector instead of objc.signature\", DeprecationWarning)\n kw['signature'] = signature\n def makeSignature(func):\n return selector(func, **kw)\n return makeSignature\n","repo_name":"albertz/music-player","sub_path":"mac/pyobjc-core/Lib/objc/_descriptors.py","file_name":"_descriptors.py","file_ext":"py","file_size_in_byte":9911,"program_lang":"python","lang":"en","doc_type":"code","stars":490,"dataset":"github-code","pt":"81"}
+{"seq_id":"14035889562","text":"# from typing import ValuesView\nfrom django.test import TestCase\nfrom django.contrib.auth import get_user_model\n\n\nclass ModelTest(TestCase):\n\n def test_create_user_with_phone_successful(self):\n \"\"\"Test creating a new user with an email is successful\"\"\"\n phone = \"+919999999999\"\n user = get_user_model().objects.create_user(\n phone=phone\n )\n self.assertEqual(user.phone, phone)\n\n # def test_new_user_phone_normalised(self):\n # \"\"\"Test phone for new user is normalised\"\"\"\n # phone = \"+919595955666\"\n # user = get_user_model().objects.create_user(phone)\n # self.assertEqual(user.phone, phone)\n\n def test_new_user_invalid_phone(self):\n \"\"\"Test creating user with no phone raise error\"\"\"\n with self.assertRaises(ValueError):\n get_user_model().objects.create_user(None)\n\n def test_create_new_super_user(self):\n \"\"\"Test new super user created\"\"\"\n user = get_user_model().objects.create_superuser('+919999999999')\n self.assertTrue(user.is_superuser)\n self.assertTrue(user.is_staff)\n","repo_name":"abhishek-codes/prodocker","sub_path":"app/core/tests/test_models.py","file_name":"test_models.py","file_ext":"py","file_size_in_byte":1113,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"11344356171","text":"# https://programmers.co.kr/learn/courses/30/lessons/42626#qna\n# 더 맵게\n# 문의롤 통해 scoville 지수의 길이는 모두 2 이상으로 변경!\n# 따라서 scoville지수의 길이가 1이면서 이미 k 이상인 경우는 존재x\nimport heapq\n\n\ndef solution(scoville, K):\n heapq.heapify(scoville)\n answer = 0\n size = len(scoville)\n for t in range(size-1):\n one = heapq.heappop(scoville)\n two = heapq.heappop(scoville)\n new = one+(two*2)\n heapq.heappush(scoville,new)\n answer+=1\n if scoville[0]>=K:\n return answer\n return -1\n\n\n","repo_name":"MinMolang/codePractice","sub_path":"Programmers/HEAP_hotter.py","file_name":"HEAP_hotter.py","file_ext":"py","file_size_in_byte":606,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"30756542116","text":"import numpy as np\n\nfrom keras.callbacks import Callback\nfrom time import sleep\n\n\nclass EarlyStopping(Callback):\n \"\"\"Stop training when a monitored quantity has stopped improving.\n\n # Arguments\n monitor: quantity to be monitored.\n min_delta: minimum change in the monitored quantity\n to qualify as an improvement, i.e. an absolute\n change of less than min_delta, will count as no\n improvement.\n patience: number of epochs with no improvement\n after which training will be stopped.\n verbose: verbosity mode.\n mode: one of {auto, min, max}. In `min` mode,\n training will stop when the quantity\n monitored has stopped decreasing; in `max`\n mode it will stop when the quantity\n monitored has stopped increasing; in `auto`\n mode, the direction is automatically inferred\n from the name of the monitored quantity.\n baseline: Baseline value for the monitored quantity to reach.\n Training will stop if the model doesn't show improvement\n over the baseline.\n \"\"\"\n\n def __init__(self,\n monitor='val_loss',\n min_delta=0,\n patience=0,\n verbose=0,\n mode='auto',\n baseline=None):\n super(EarlyStopping, self).__init__()\n\n self.output_msg = \"\"\n self.monitor = monitor\n self.baseline = baseline\n self.patience = patience\n self.verbose = verbose\n self.min_delta = min_delta\n self.wait = 0\n self.stopped_epoch = 0\n\n if mode not in ['auto', 'min', 'max']:\n warnings.warn('EarlyStopping mode %s is unknown, '\n 'fallback to auto mode.' % mode,\n RuntimeWarning)\n mode = 'auto'\n\n if mode == 'min':\n self.monitor_op = np.less\n elif mode == 'max':\n self.monitor_op = np.greater\n else:\n if 'acc' in self.monitor:\n self.monitor_op = np.greater\n else:\n self.monitor_op = np.less\n\n if self.monitor_op == np.greater:\n self.min_delta *= 1\n else:\n self.min_delta *= -1\n\n def on_train_begin(self, logs=None):\n # Allow instances to be re-used\n self.wait = 0\n self.stopped_epoch = 0\n if self.baseline is not None:\n self.best = self.baseline\n else:\n self.best = np.Inf if self.monitor_op == np.less else -np.Inf\n\n def on_epoch_end(self, epoch, logs=None):\n current = logs.get(self.monitor)\n if current is None:\n warnings.warn(\n 'Early stopping conditioned on metric `%s` '\n 'which is not available. Available metrics are: %s' %\n (self.monitor, ','.join(list(logs.keys()))), RuntimeWarning\n )\n return\n if self.monitor_op(current - self.min_delta, self.best):\n self.best = current\n self.wait = 0\n else:\n self.wait += 1\n if self.wait >= self.patience:\n self.stopped_epoch = epoch\n self.model.stop_training = True\n\n def on_train_end(self, logs=None):\n if self.stopped_epoch > 0 and self.verbose > 0:\n self.output_msg = 'Epoch %05d: early stopping' % (self.stopped_epoch + 1)\n print(self.output_msg) \n\n def get_stopped_epoch(self):\n if self.stopped_epoch > 0:\n return self.stopped_epoch+1\n return \"No\"\n","repo_name":"fr4ncidir/ImpactExperiment","sub_path":"custom_callbacks.py","file_name":"custom_callbacks.py","file_ext":"py","file_size_in_byte":3617,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"30007952771","text":"\n# coding: utf-8\n\n# In[26]:\n\n\n#Xcalar imports. For more information, refer to discourse.xcalar.com\nfrom xcalar.external.LegacyApi.XcalarApi import XcalarApi\nfrom xcalar.external.LegacyApi.Session import Session\nfrom xcalar.external.LegacyApi.WorkItem import WorkItem\nfrom xcalar.external.LegacyApi.Operators import *\nfrom xcalar.external.LegacyApi.Dataset import *\nfrom xcalar.external.LegacyApi.WorkItem import *\nfrom xcalar.external.LegacyApi.Udf import *\nfrom xcalar.external.Retina import *\nimport timeit\nimport argparse\nimport json\nimport tarfile\nimport io\nfrom tarfile import TarInfo\n\n#Code starts here. First create a XcalarApi object to do anything\nxcalarApi = XcalarApi(bypass_proxy=True)\nop = Operators(xcalarApi)\nudf = Udf(xcalarApi)\nretina = Retina(xcalarApi)\n\nuser = \"admin\"\nsession = \"test\"\n\n#Connect to current workbook that you are in\nworkbook = Session(xcalarApi, \"admin\", \"admin\", 4135730, True, \"test\")\nxcalarApi.setSession(workbook)\n\npathDelim = '\\\"/\\\"'\ncutNum = 8\n\nfinalTable = \"Transactions\"\nfinalTableKey = \"txnId\"\n\ntables = {\n \"orders\": {\n \"key\": \"o_orderkey\",\n \"opcode\": \"opcode\",\n \"columns\": [\n {\n \"name\": \"o_custkey\",\n \"type\": \"DfInt64\"\n },\n {\n \"name\": \"o_orderstatus\",\n \"type\": \"DfString\"\n },\n {\n \"name\": \"o_totalprice\",\n \"type\": \"DfFloat64\"\n },\n {\n \"name\": \"o_orderdate\",\n \"type\": \"DfString\"\n },\n {\n \"name\": \"o_orderpriority\",\n \"type\": \"DfString\"\n },\n {\n \"name\": \"o_shippriority\",\n \"type\": \"DfInt64\"\n },\n {\n \"name\": \"o_clerk\",\n \"type\": \"DfString\"\n },\n {\n \"name\": \"o_comment\",\n \"type\": \"DfString\"\n }\n ],\n \"path\": \"/netstore/users/vgonela/hvr_data/tpch/orders/\",\n }\n}\n\ndef formatColumns(colsIn, key, prefix = None):\n cols = []\n\n for col in colsIn:\n synthCol = {}\n if prefix:\n synthCol[\"sourceColumn\"] = prefix + \"::\" + col[\"name\"]\n else:\n synthCol[\"sourceColumn\"] = col[\"name\"]\n\n synthCol[\"destColumn\"] = col[\"name\"]\n synthCol[\"columnType\"] = col[\"type\"]\n\n cols.append(synthCol)\n\n cols.append({\"sourceColumn\": key,\n \"columnType\": \"DfInt64\",\n \"destColumn\": key})\n\n cols.append({\"sourceColumn\": \"XcalarRankOver\",\n \"columnType\": \"DfInt64\",\n \"destColumn\": \"XcalarRankOver\"})\n\n cols.append({\"sourceColumn\": \"XcalarOpCode\",\n \"columnType\": \"DfInt64\",\n \"destColumn\": \"XcalarOpCode\"})\n\n return (cols)\n\n\nfor tableName, info in tables.items():\n query = []\n\n load = {\n \"operation\": \"XcalarApiBulkLoad\",\n \"args\": {\n \"dest\": \".XcalarDS.{}\".format(tableName),\n \"loadArgs\": {\n \"parseArgs\": {\n \"parserFnName\": \"default:convertNewLineJsonToArrayJson\",\n \"parserArgJson\": \"{}\",\n \"fileNameFieldName\": \"fn\",\n \"recordNumFieldName\": \"rec\",\n \"allowFileErrors\": False,\n \"allowRecordErrors\": False,\n \"schema\": []\n },\n \"sourceArgsList\": [\n {\n \"recursive\": False,\n \"path\": info[\"path\"],\n \"targetName\": \"Default Shared Root\",\n \"fileNamePattern\": \"\"\n }\n ],\n \"size\": 10737418240\n }\n }\n }\n\n query.append(load)\n\n index = [\n {\n \"operation\": \"XcalarApiIndex\",\n \"args\": {\n \"source\": \".XcalarDS.{}\".format(tableName),\n \"dest\": \"{}-1\".format(tableName),\n \"prefix\": tableName,\n \"key\": [\n {\n \"name\": \"xcalarRecordNum\",\n \"ordering\": \"Unordered\",\n \"keyFieldName\": \"\",\n \"type\": \"DfInt64\"\n }\n ],\n },\n },\n {\n \"operation\": \"XcalarApiMap\",\n \"tag\": \"sortTimestamp\",\n \"args\": {\n \"source\": \"{}-1\".format(tableName),\n \"dest\": \"{}-map\".format(tableName),\n \"eval\": [\n {\n \"evalString\": \"int(cut({}::fn, {}, {}))\".format(tableName,\n cutNum,\n pathDelim),\n \"newField\": \"ts\"\n },\n {\n \"evalString\": \"int({}::rec, 10)\".format(tableName),\n \"newField\": \"rec\"\n }\n ],\n },\n },\n {\n \"operation\": \"XcalarApiIndex\",\n \"tag\": \"sortTimestamp\",\n \"comment\": \"apply ordering based on file row and timestamp\",\n \"args\": {\n \"source\": \"{}-map\".format(tableName),\n \"dest\": \"{}-sort\".format(tableName),\n \"key\": [\n {\n \"name\": \"ts\",\n \"ordering\": \"Ascending\",\n \"keyFieldName\": \"ts\",\n \"type\": \"DfInt64\"\n },\n {\n \"name\": \"rec\",\n \"ordering\": \"Ascending\",\n \"keyFieldName\": \"rec\",\n \"type\": \"DfInt64\"\n }\n ],\n },\n }\n ]\n\n query += index\n\n rankOver = [\n {\n \"operation\": \"XcalarApiGetRowNum\",\n \"tag\": \"rankOver\",\n \"args\": {\n \"source\": \"{}-sort\".format(tableName),\n \"dest\": \"{}-getRowNum\".format(tableName),\n \"newField\": \"rowNum\"\n },\n },\n {\n \"operation\": \"XcalarApiIndex\",\n \"tag\": \"rankOver\",\n \"args\": {\n \"source\": \"{}-getRowNum\".format(tableName),\n \"dest\": \"{}-keyIndex\".format(tableName),\n \"key\": [\n {\n \"name\": \"{}::{}\".format(tableName, info[\"key\"]),\n \"ordering\": \"Unordered\",\n \"keyFieldName\": info[\"key\"],\n \"type\": \"DfInt64\"\n }\n ],\n },\n },\n {\n \"operation\": \"XcalarApiGroupBy\",\n \"tag\": \"rankOver\",\n \"comment\": \"\",\n \"args\": {\n \"source\": \"{}-keyIndex\".format(tableName),\n \"dest\": \"{}-groupBy\".format(tableName),\n \"eval\": [\n {\n \"evalString\": \"minInteger(rowNum)\",\n \"newField\": \"minRow\"\n }\n ],\n \"newKeyField\": \"dummyKey\",\n },\n },\n {\n \"operation\": \"XcalarApiJoin\",\n \"tag\": \"rankOver\",\n \"args\": {\n \"source\": [\n \"{}-keyIndex\".format(tableName),\n \"{}-groupBy\".format(tableName),\n ],\n \"key\": [\n [\n info[\"key\"]\n ],\n [\n \"dummyKey\"\n ]\n ],\n \"joinType\": \"innerJoin\",\n \"dest\": \"{}-joinBack\".format(tableName),\n \"columns\": [\n [\n {\n \"sourceColumn\": \"orders\",\n \"columnType\": \"DfFatptr\",\n \"destColumn\": \"orders\"\n },\n {\n \"sourceColumn\": info[\"key\"],\n \"columnType\": \"DfInt64\",\n \"destColumn\": info[\"key\"],\n }\n ],\n [\n {\n \"sourceColumn\": \"dummyKey\",\n \"columnType\": \"DfInt64\",\n \"destColumn\": \"dummyKey\"\n },\n {\n \"sourceColumn\": \"minRow\",\n \"columnType\": \"DfInt64\",\n \"destColumn\": \"minRow\"\n }\n ]\n ],\n },\n },\n {\n \"operation\": \"XcalarApiMap\",\n \"tag\": \"rankOver\",\n \"args\": {\n \"source\": \"{}-joinBack\".format(tableName),\n \"dest\": \"{}-ranked\".format(tableName),\n \"eval\": [\n {\n \"evalString\": \"int(add(sub(rowNum, minRow), 1))\",\n \"newField\": \"XcalarRankOver\"\n },\n {\n \"evalString\": \"int({}::{})\".format(tableName,\n info[\"opcode\"]),\n \"newField\": \"XcalarOpCode\"\n }\n ],\n },\n }\n ]\n\n query += rankOver\n\n synthesizeColumns = formatColumns(info[\"columns\"], info[\"key\"])\n\n synth = {\n \"operation\": \"XcalarApiSynthesize\",\n \"comment\": \"apply schema\",\n \"args\": {\n \"sameSession\": True,\n \"source\": \"{}-ranked\".format(tableName),\n \"dest\": tableName,\n \"columns\": synthesizeColumns,\n },\n }\n\n query.append(synth)\n\n queryStr = json.dumps(query)\n\n print(\"Starting {}\".format(tableName))\n print(queryStr)\n\n start = timeit.default_timer()\n xcalarApi.submitQuery(queryStr, session, tableName)\n\n try:\n op.unpublish(tableName)\n except:\n pass\n\n op.publish(tableName, tableName)\n end = timeit.default_timer()\n\n elapsed = end - start\n print(\"Ran {}: {}s\".format(tableName, str(elapsed)))\n\n retinaColumns = [col[\"destColumn\"] for col in synthesizeColumns]\n\n try:\n retina.delete(tableName)\n except:\n pass\n\n retina.make(tableName,\n [tableName],\n [retinaColumns])\n\n\nbatchId = 0\n\nfor tableName, info in tables.items():\n joinColumns = formatColumns(info[\"columns\"], info[\"key\"])\n\n dataflowInfo = {}\n\n dataflowInfo[\"query\"] = [\n {\n \"operation\": \"XcalarApiSelect\",\n \"comment\": \"Latest batch select on {}\".format(tableName),\n \"args\": {\n \"source\": tableName,\n \"dest\": \"{}.select\".format(tableName),\n \"minBatchId\": batchId,\n \"maxBatchId\": batchId\n }\n },\n {\n \"operation\": \"XcalarApiSelect\",\n \"comment\": \"Full historical select on {}\".format(finalTable),\n \"args\": {\n \"source\": finalTable,\n \"dest\": \"{}.select\".format(finalTable)\n }\n },\n {\n \"operation\": \"XcalarApiAggregate\",\n \"comment\": \"Find minimum key\",\n \"tag\": \"filterMinKey\",\n \"args\": {\n \"source\": \"{}.select\".format(tableName),\n \"dest\": \"{}.minKey\".format(tableName),\n \"eval\": [\n {\n \"evalString\": \"min({})\".format(info[\"key\"])\n }\n ]\n }\n },\n {\n \"operation\": \"XcalarApiFilter\",\n \"comment\": \"Reduce {} using calculated minKey\".format(finalTable),\n \"tag\": \"filterMinKey\",\n \"args\": {\n \"source\": \"{}.select\".format(finalTable),\n \"dest\": \"{}.filter\".format(finalTable),\n \"eval\": [\n {\n \"evalString\": \"gt({}, ^{}.minKey)\".format(info[\"key\"],\n tableName)\n }\n ]\n },\n },\n {\n \"operation\": \"XcalarApiJoin\",\n \"comment\": \"Apply changes to {} by doing a join\".format(finalTable),\n \"args\": {\n \"source\": [\n \"{}.filter\".format(finalTable),\n \"{}.select\".format(tableName),\n ],\n \"joinType\": \"crossJoin\",\n \"dest\": \"{}-joined\".format(tableName),\n \"evalString\": \"eq({}_, {})\".format(info[\"key\"], info[\"key\"]),\n \"columns\": [\n [\n {\n \"sourceColumn\": info[\"key\"],\n \"columnType\": \"DfInt64\",\n \"destColumn\": info[\"key\"] + \"_\"\n },\n {\n \"sourceColumn\": finalTableKey,\n \"columnType\": \"DfInt64\",\n \"destColumn\": finalTableKey,\n }\n ],\n joinColumns\n ],\n },\n }\n ]\n\n tableColumns = [{\"columnName\": col[\"destColumn\"], \"headerAlias\": col[\"destColumn\"]}\n for col in joinColumns]\n\n dataflowInfo[\"tables\"] = [\n {\n \"name\": \"{}-joined\".format(tableName),\n \"columns\": tableColumns\n }\n ]\n\n dataflowStr = json.dumps(dataflowInfo)\n\n retinaBuf = io.BytesIO()\n\n with tarfile.open(fileobj = retinaBuf, mode = \"w:gz\") as tar:\n info = TarInfo(\"dataflowInfo.json\")\n info.size = len(dataflowStr)\n\n tar.addfile(info, io.BytesIO(bytearray(dataflowStr, \"utf-8\")))\n\n retina.add(tableName + \"_applyUpdate\", retinaBuf.getvalue())\n","repo_name":"varlogtim/xcalar","sub_path":"scripts/createRetina.py","file_name":"createRetina.py","file_ext":"py","file_size_in_byte":14204,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"37877448325","text":"import chevron\nimport subprocess\nfrom pathlib import Path\nimport logging\n\nFILEDIR = Path(__file__).parent.resolve()\n\ndef write_dineof_config(outfile, dineof_in, variable, min_modes=10, max_modes=20, land_mask='land_mask', reconstruction=0):\n\n template = FILEDIR / 'dineof_config_template.init'\n\n ncv = max_modes + 5\n\n data = dict(\n dineof_in=dineof_in,\n variable=variable,\n min_modes=min_modes, \n max_modes=max_modes, \n land_mask=land_mask,\n reconstruction=reconstruction,\n ncv = ncv\n )\n\n with open(str(template), 'r') as f:\n outconfig = chevron.render(f, data)\n\n with open(outfile,\"w\") as f:\n f.write(outconfig)\n\n return\n\ndef exec_dineof(dineof_in, variable, **kwargs):\n\n config_path = f'dineof_{variable}.init'\n\n write_dineof_config(config_path, dineof_in, variable, **kwargs)\n\n cmd = f'dineof {config_path}'\n\n proc = subprocess.Popen(\n cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT\n )\n out, err = proc.communicate()\n\n if out is not None:\n print(out.decode(\"utf-8\"))\n \n if err is not None:\n print('Error:',err.decode(\"utf-8\"))\n\n return\n\n \n\n","repo_name":"KMarkert/dineof-ee-gcp","sub_path":"src/dineof_runner.py","file_name":"dineof_runner.py","file_ext":"py","file_size_in_byte":1204,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"6138907672","text":"\"ISN - Python - Partie 3 - Exercice 2.1\"\r\n\"Ecrivez un script qui permette d'encoder un fichier texte dont les lignes contiendront\\\r\n chacune les noms, prénom, adresse, code postal ,ville et n° de téléphone de différentes\\\r\n personnes (considérez par exemple qu'il s'agit des membres d'un club).\"\r\n\r\n#######################################################\r\n# Encodage: UTF-8 #\r\n# Programme Python 3.7 #\r\n# Auteur : Antoine Cheucle #\r\n# Pas de licence, libre d'utilisation #\r\n# Dépendences: aucune #\r\n#######################################################\r\n\r\ndef __main__():\r\n fichier = open(input(\"Entrez le nom du fichier dans lequel écrire (N'oubliez pas d'inscrire également son extension s'il en possède une!)\\n\"), \"a\")\r\n print(\"Bien! Vous pouvez désormais commencer à entrer les données.\")\r\n nb_personne = 0\r\n while True:\r\n nb_personne += 1\r\n donnees = []\r\n rep = input(\"Entrez le nom de famille de la personne n°\"+str(nb_personne)+ \" Appuyez sur Entrer pour annuler\")\r\n if rep == \"\":\r\n break\r\n else:\r\n donnees.append(rep)\r\n donnees.append(input(\"Entrez le prénom de la personne n°\"+str(nb_personne)))\r\n donnees.append(input(\"Entrez l'adresse de la personne n°\"+str(nb_personne)))\r\n donnees.append(input(\"Entrez le code postal de la personne n°\"+str(nb_personne)))\r\n donnees.append(input(\"Entrez la ville de la personne n°\"+str(nb_personne)))\r\n donnees.append(input(\"Entrez le numéro de téléphone de la personne n°\"+str(nb_personne)))\r\n fichier.write(\"|\".join(donnees)+\"|\\n\")\r\n fichier.close()\r\n print(\"Fin de l'éciture! Fichier sauvegardé.\")\r\n\r\n\r\n__main__()\r\n","repo_name":"antoinech2/ISN-exercices-Python-Ann-e-2","sub_path":"partie_3/ex2/ex2.1.py","file_name":"ex2.1.py","file_ext":"py","file_size_in_byte":1855,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"72298650826","text":"####################################################\n## EVOLUTIOINARY FAILURE MODE CALCULATOR\n## Version 1.0.1\n####################################################\n\nEFM_VERSION = \"1.0.1\"\n\n# Import subprocess for command line stuff\nimport subprocess\nimport os\nimport csv\nfrom tempfile import NamedTemporaryFile\n\n# Import itertools\nfrom itertools import combinations\n\n# Import RegEx libraries\nimport re\nimport math\n\n# Import ElementTree for parsing BioBrick XML\nfrom xml.etree import ElementTree\n\nSUB_RATE = float(2.2 * 10 ** (-10))\n\n\ndef run_mummer(fasta_file, org):\n \"\"\"\n Executes the repeat-match command in MUMmer and parses the output.\n Executes the nucmer and show-coords commands in MUMmer and parses the output.\n :param fasta_file: Name of FASTA file with sequence\n :param org: Name of organism for calculating rates\n :return: Dictionary of results\n \"\"\"\n # Initialize dictionary\n long_repeats = dict()\n # Execute MUMmer command\n try:\n output = subprocess.check_output(['repeat-match', '-n', '16', '-f', fasta_file])\n except:\n raise\n print(\"MUMmer command 'repeat-match' failed.\")\n output_lines = output.splitlines()\n # Loop through mummer output and skip the first 2 lines which only contain header information\n for line in output_lines[2:]:\n split_line = line.split()\n start_pos1 = int(split_line[0])\n start_pos2 = int(split_line[1])\n repeat_length = int(split_line[2])\n if repeat_length in long_repeats:\n for locations in long_repeats[repeat_length]:\n if start_pos2 in locations and start_pos1 in locations:\n continue\n elif start_pos1 in locations:\n locations.append(start_pos2)\n locations.sort()\n found = True\n elif start_pos2 in locations:\n locations.append(start_pos1)\n locations.sort()\n found = True\n else:\n found = False\n if found is False:\n # This is definitely a new repeat of repeat_length that we haven't seen before\n long_repeats[repeat_length].append(sorted([start_pos1, start_pos2]))\n else:\n # We have not seen a repeat of this length before\n long_repeats[repeat_length] = [sorted([start_pos1, start_pos2])]\n\n # Initialize list of dictionaries and reformat for tabular output\n output_list = []\n seen = []\n for repeat_length, location_list in long_repeats.items():\n for locations in location_list:\n rate_per_repeat = float(0)\n if len(locations) > 2:\n # Must account for a homologous sequence occurring more than twice\n combos = combinations(locations, 2)\n for combo in combos:\n rate_per_repeat += get_recombo_rate(int([repeat_length][0]), combo[0], combo[1], org)\n else:\n rate_per_repeat = get_recombo_rate(int([repeat_length][0]), locations[0], locations[1], org)\n entry = {'location': locations,\n 'sequence': '', # We don't record the sequence for long repeats\n 'length': [repeat_length],\n 'count': len(locations),\n 'raw_rate': rate_per_repeat,\n 'type': 'rmd'}\n if [locations, [repeat_length]] not in seen:\n output_list.append(entry)\n seen.append([locations, [repeat_length]])\n\n # Generate a temporary file for storing the '.delta' file that will be passed to show-coords\n coords_file = NamedTemporaryFile(suffix='.delta', delete=False)\n try:\n # Execute nucmer command\n subprocess.call(['nucmer',\n '-l',\n '16',\n '-f', # Forward strand only\n '--maxmatch',\n '--nosimplify',\n '--prefix='+coords_file.name[:-6],\n fasta_file,\n fasta_file],\n )\n # Execute show-coords command\n output = subprocess.check_output(['show-coords', '-T', '-I 100', '-H', '-d', coords_file.name])\n except:\n raise\n print(\"MUMmer command 'nucmer' and 'show-coords' failed.\")\n finally:\n # Remove the '.delta' file because we're finished with it\n os.remove(coords_file.name)\n\n content_lines = output.splitlines()\n # Initialize list of dictionaries for tabular output\n\n for line in content_lines:\n clean_line = line.split()\n start_pos1 = int(clean_line[0])\n start_pos2 = int(clean_line[2])\n length1 = int(clean_line[4])\n length2 = int(clean_line[5])\n if start_pos1 != start_pos2:\n # Remove repeats by converting to set\n length_list = sorted(list(set([length1, length2])))\n location = sorted([start_pos1, start_pos2])\n rate = get_recombo_rate(int(length_list[0]), location[0], location[1], org)\n entry = {'location': location,\n 'sequence': '',\n 'length': length_list,\n 'count': len(location),\n 'raw_rate': rate,\n 'type': 'rmd'}\n if [location, length_list] not in seen:\n output_list.append(entry)\n seen.append([location, length_list])\n\n return output_list\n\n\ndef get_repeats_in_window(n, sequence, min_count, org):\n \"\"\"\n Generates a list of k-mer repeats in a given window size\n :param n: Size of k-mer (2 for dinucleotides, 3 for trinucleotides, etc.)\n :param sequence: String of DNA sequence\n :param min_count: Minimum number of repeating units before recording\n :param org: Host organism for rate calculations\n :return: List of dictionaries of results for tabular output\n \"\"\"\n # Initialize repeats dictionary, in the format repeats -> { 'index' : ['CG', 2] }\n repeats = dict()\n if n == 0:\n # Return an empty list, n = 0 doesn't make sense\n return []\n\n # Start counting repeats at 1\n repeat_count = 1\n i = 0\n if n == 1:\n # n = 1 is a special, simpler case to handle\n while i < (len(sequence) - n):\n if sequence[i] == sequence[i + 1]:\n # Record as a repeat\n if repeat_count + 1 >= min_count:\n repeats[i + n - n * repeat_count] = [sequence[i], repeat_count + 1]\n # Jump one base ahead\n i += 1\n repeat_count += 1\n else:\n # Reset repeat_count\n i += 1\n repeat_count = 1\n else:\n # n > 1\n while i < (len(sequence) - n):\n # If the first base of the current window matches the first base of the next window\n # AND the entire window equals the next consecutive window\n if sequence[i] == sequence[i + n] and sequence[i:i + n] == sequence[(i + n):(i + 2 * n)]:\n # If the window is larger than 4, make sure it doesn't contain any smaller repeating subunits\n if check_subunits(sequence[i:i + n]) is True:\n i += 1\n else:\n # Record as a repeat\n if repeat_count + 1 >= min_count:\n repeats[i + n - n * repeat_count] = [sequence[i:i + n], repeat_count + 1]\n # Jump forward a full window of bases\n i = i + n\n repeat_count += 1\n else:\n repeat_count = 1\n # Otherwise shift the window forward by one base\n i += 1\n\n # Reformat for tabular output\n output_list = []\n for index, contents in repeats.items():\n mut_rate = get_mut_rate(contents[1], n, org)\n entry = {'location': [index],\n 'sequence': str(contents[0]),\n 'length': [len(contents[0])],\n 'count': contents[1],\n 'raw_rate': mut_rate,\n 'type': 'ssr'\n }\n output_list.append(entry)\n\n return output_list\n\n\ndef check_subunits(sequence):\n \"\"\"\n Checks to see if a sequence can be broken up into smaller repeating subunits. For example, ATAT is actually just\n two instances of AT.\n :param sequence: Short DNA sequence\n :return: Boolean. True if sequence can be broken up into smaller repeating subunits. Otherwise, False.\n \"\"\"\n if re.match(r\"^(.+?)\\1+$\", sequence):\n return True\n else:\n return False\n\n\ndef get_mut_rate(repeat_count, unit_length, org):\n '''\n Calculates mutation rate for simple sequence repeats\n :param repeat_count: Number of times the repeating unit occurs\n :param unit_length: Length of repeating unit\n :param org: Host organism\n :return: Mutation rate\n '''\n mut_rate = float(0)\n if org == 'ecoli' or org == 'reca':\n if unit_length == 1:\n # Formula based on analysis of Lee et. al. data\n mut_rate = float(10 ** (0.72896 * repeat_count - 12.91471))\n elif unit_length > 1:\n mut_rate = float(10 ** (0.06282 * repeat_count - 4.74882))\n elif org == 'yeast':\n if unit_length == 1:\n mut_rate = float(10 ** (0.3092 * repeat_count - 7.3220))\n elif unit_length > 1:\n mut_rate = float(10 ** (0.11141 * repeat_count - 7.65810))\n return mut_rate\n\n\ndef get_recombo_rate(length, location1, location2, org):\n '''\n Calculate the recombination rate based on the Oliviera, et. al. formula\n :param length: Length of homologous region\n :param location1: Location of first homologous region\n :param location2: Location of second homologous region\n :param org: Host organism\n :return: Recombination rate\n '''\n spacer = abs(int(location2) - int(location1)) - int(length)\n # If the homologous sequences overlap we can't calculate a rate\n if spacer < 0:\n return 0\n if org == 'ecoli' or org == 'yeast':\n recombo_rate = float(((8.8 + spacer) ** (-29.0 / length)) * (length / (1 + 1465.6 * length)))\n elif org == 'reca':\n recombo_rate = float(\n ((200.4 + spacer) ** (-8.8 / length)) * (length / (1 + 2163.0 * length + 14438.6 * spacer)))\n\n return recombo_rate\n\n\ndef get_biobrick_features(tree):\n \"\"\"\n Extracts sequence annotations from BioBrick XML files\n :param tree: parsed XML tree\n :return: List of dictionaries of features to display\n \"\"\"\n features = []\n colors = {\n 'promoter': 'green',\n 'stop': 'red',\n 'cds': 'blue',\n 'rbs': 'orange',\n 'binding': 'purple',\n 'BioBrick': 'black'\n }\n for node in tree.iter(tag='feature'):\n if node.find('type').text in colors:\n entry = dict()\n entry['type'] = node.find('type').text\n entry['title'] = node.find('title').text\n entry['startpos'] = int(node.find('startpos').text)\n entry['length'] = int(node.find('endpos').text) - int(node.find('startpos').text)\n entry['color'] = colors[entry['type']]\n features.append(entry)\n\n return features\n\n\ndef get_genbank_features(genome):\n \"\"\"\n Extracts sequence annotations a BioPython SeqRecord object\n :param genome: BioPython SeqRecord of genbank file\n :return: List of dictionaries of features to display\n \"\"\"\n features = []\n for feature in genome.features:\n if feature.qualifiers:\n name = feature.qualifiers.itervalues().next()[0]\n else:\n name = 'Untitled'\n entry = dict()\n entry['type'] = str(feature.type)\n entry['title'] = name\n entry['startpos'] = int(feature.location.nofuzzy_start)\n entry['length'] = int(feature.location.nofuzzy_end) - int(feature.location.nofuzzy_start)\n features.append(entry)\n\n return features\n\n\ndef truncate_table(repeats, min_rate):\n \"\"\"\n Returns a subset of repeats with a mutation rate higher than min_rate\n :param repeats: List of dictionaries of repeats\n :param min_rate: Minimum mutation rate to display in table\n :return: A sorted and truncated list of dictionaries of repeats\n \"\"\"\n trunc = (repeat for repeat in repeats if\n repeat['raw_rate'] > min_rate and repeat['raw_rate'] != '' and repeat['overlap'] == True)\n trunc_sort = sorted(trunc, key=lambda k: k['raw_rate'] if k['raw_rate'] != '' else 0, reverse=True)\n return trunc_sort\n\n\ndef check_overlap(repeats, features, check_features):\n \"\"\"\n Checks each repeat to see if it overlaps with a feature, and sets 'overlap' to True if it does.\n :param repeats: List of dictionaries of repeats\n :param features: List of dictionaries of features\n :return: List of dictionaries of repeats, with each repeat containing a boolean 'overlap' value\n \"\"\"\n # If there are no features to report set all overlap to true to display everything\n if check_features is False:\n for repeat in repeats:\n repeat['overlap'] = True\n return repeats\n else:\n for repeat in repeats:\n for feature in features:\n if repeat['type'] == 'rmd':\n repeat_range = set(xrange(repeat['location'][0], repeat['location'][-1] + repeat['length'][0]))\n else:\n repeat_range = set(\n xrange(repeat['location'][0], repeat['location'][0] + repeat['length'][0] * repeat['count']))\n feature_range = set(xrange(feature['startpos'], feature['startpos'] + feature['length']))\n # Magical ampersand checks if two sets overlap\n if repeat_range & feature_range:\n repeat['overlap'] = True\n break\n else:\n repeat['overlap'] = False\n\n return repeats\n\n\ndef rate_sum(repeats, seq_len):\n \"\"\"\n Calculates an RIP score for given sequence\n :param repeats: List of dictionaries of repeats\n :param seq_len: Length of input sequence\n :return: Total predicted RIP score for whole sequence\n \"\"\"\n ssr_sum = float(0)\n rmd_sum = float(0)\n for entry in repeats:\n if entry['raw_rate'] != '' and entry['overlap'] is True:\n if entry['type'] == 'ssr':\n ssr_sum += entry['raw_rate']\n elif entry['type'] == 'rmd':\n rmd_sum += entry['raw_rate']\n\n base_rate = float(seq_len) * float(SUB_RATE)\n # Add in the mutation rate of an individual nucleotide\n r_sum = ssr_sum + rmd_sum + base_rate\n # Set the maximum rate sum to 1 for now.\n if r_sum > 1:\n r_sum = float(1)\n rel_rate = (float(r_sum) / float(base_rate))\n\n return {'rip': rel_rate, 'ssr_sum': ssr_sum, 'rmd_sum': rmd_sum, 'bps_sum': base_rate}\n\n\ndef process_efm(form):\n \"\"\"\n Takes a django form object and finds potentially hypermutable sites in a submitted sequence.\n :param form: A Django form object for processing\n :return: A dictionary of values to pass to the template renderer\n \"\"\"\n # Define the paths for your input files, sequence files, and biobrick file\n input_file = form.cleaned_data.get('fasta_file')\n features = form.cleaned_data.get('features')\n my_seq = form.cleaned_data.get('raw_sequence')\n org = form.cleaned_data.get('organism')\n\n # Set maximum window size for get_repeats_in_window()\n unit_length = 15\n\n # Integrate repeats generated from nucmer and from repeat-match\n mummer_repeats = run_mummer(input_file, org)\n\n # Get short repeats (SSRs) in each window size up to max unit_length\n all_ssr = []\n for i in range(unit_length):\n min_count = math.ceil(8 / (i + 1))\n if min_count < 3:\n min_count = 3\n if (i + 1) == 1:\n min_count = 4\n repeat_in_window = get_repeats_in_window(i + 1, my_seq, min_count, org)\n if repeat_in_window:\n all_ssr += repeat_in_window\n\n # Merge repeat lists together\n merged_repeats = mummer_repeats + all_ssr\n\n # Check if any areas overlap annotated regions\n merged_repeats = check_overlap(merged_repeats, features, form.cleaned_data.get('check_features'))\n\n # Truncate repeat list based on absolute mutation rate\n merged_repeats_trunc = truncate_table(merged_repeats, 10 ** (-9))\n\n # Find the sum of all mutation rates for sequences.\n overall_rate = rate_sum(merged_repeats, len(my_seq))\n\n return {'repeats': merged_repeats_trunc if merged_repeats_trunc else '',\n 'features': features,\n 'seq_length': len(my_seq),\n 'rate': overall_rate,\n 'title': form.cleaned_data['title'],\n 'check_features': form.cleaned_data['check_features'],\n 'organism': org,\n 'version': EFM_VERSION}\n","repo_name":"barricklab/efm-calculator","sub_path":"efm_helper.py","file_name":"efm_helper.py","file_ext":"py","file_size_in_byte":16938,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"81"}
+{"seq_id":"15088268665","text":"# implements a linked list class - methods: append, prepend, includes, length, pop, shift, remove, reverse\n\nclass Node:\n def __init__(self,value):\n self.value = value\n self.next = None\n def __str__(self) -> str:\n #return \"value: {}, next => {}\".format(self.value,self.next)\n return \"{}\".format(self.value)\n\n\nclass LinkedList:\n def __init__(self):\n self.head = None\n self.length = 0\n\n\n def __str__(self) -> str:\n string = \"\"\n n = self.head\n while(n is not None):\n string += n.value\n n = n.next\n if(n is not None):\n string += \" => \"\n return string\n \n @classmethod\n def make_list(cls):\n return cls()\n \n\n def append(self,value):\n self.length += 1\n\n def loop_list(node):\n if(self.head is None):\n self.head = Node(value)\n print(\"Node added\")\n return\n elif(node.next is None):\n node.next = Node(value)\n print(\"Node added\")\n else:\n return loop_list(node.next)\n \n\n loop_list(self.head)\n\n\n def prepend(self,value):\n self.length +=1\n next = self.head\n new_head = Node(value)\n self.head = new_head\n self.head.next = next\n print(\"Node added\")\n\n\n def includes(self,value):\n node = self.head\n found_value = False\n while(node is not None):\n if(node.value == value):\n found_value = True\n node = node.next\n return found_value\n \n \n def pop(self):\n current_node = self.head\n prev_node = None\n while(current_node is not None):\n if current_node.next is None:\n prev_node.next = None\n self.length -= 1\n return current_node\n else:\n prev_node = current_node\n current_node = current_node.next\n\n\n def shift(self):\n if(self.head and (self.head.next is not None)):\n new_head = self.head.next\n prev_head = self.head\n self.head = new_head\n self.length -= 1\n return prev_head\n \n \n def remove(self,value):\n node = self.head\n prev = None\n while node is not None:\n if node.value == value :\n if prev is None:\n self.head = None\n else:\n prev.next = None\n print(\"an item was removed from your list\")\n self.length -= 1\n return\n else:\n prev = node\n node = node.next\n print(\"no item found in list to remove\")\n\n def reverse(self):\n new_order = []\n node = self.head\n while node is not None:\n new_order.insert(0,node)\n node = node.next\n new_list = self.make_list()\n for n in new_order:\n new_list.append(n.value)\n return new_list\n \n\n","repo_name":"Davidjustice28/python-data-structures","sub_path":"linked-list.py","file_name":"linked-list.py","file_ext":"py","file_size_in_byte":3093,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"31487191927","text":"from flask import Flask\n\nUPLOAD_FOLDER = ''\n\napp = Flask(__name__)\n#app.secret_key = \"secret key\"\napp.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER\napp.config['MAX_CONTENT_LENGTH'] = 16 * 1024 * 1024\n\n\n# resp.result=content\n# resp.confidence=confidence\n# return resp\n\n# resp = jsonify(errors)\n# resp.status_code = 500\n# return resp","repo_name":"Shaon2221/Face-Recognition","sub_path":"Flask-REST-API/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":325,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"38120955115","text":"from chatterbot import ChatBot\nfrom chatterbot.trainers import ListTrainer\nfrom chatterbot.trainers import ChatterBotCorpusTrainer\n\n# chatbot = ChatBot('Basic Bot')\n# chatbot.set_trainer(ChatterBotCorpusTrainer)\n# chatbot.train(\"chatterbot.corpus.english\")\n\nchatbot = ChatBot(\n \"Chatbot Backed by MongoDB\",\n storage_adapter=\"chatterbot.storage.MongoDatabaseAdapter\",\n database=\"chatterbot_db\",\n database_uri=\"mongodb://192.168.99.100:27017/\",\n logic_adapters=[\n 'chatterbot.logic.BestMatch'\n ],\n trainer='chatterbot.trainers.ChatterBotCorpusTrainer',\n filters=[\n 'chatterbot.filters.RepetitiveResponseFilter'\n ]\n)\nchatbot.set_trainer(ChatterBotCorpusTrainer)\nchatbot.train(\"chatterbot.corpus.english\")\n","repo_name":"dghanwat/faq-chatbot-service","sub_path":"python/base_training.py","file_name":"base_training.py","file_ext":"py","file_size_in_byte":743,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"36950393717","text":"from .__types__ import Cursor, Status, Matchable, Groups\n\nclass SubstringMatcher(Matchable):\n def __init__(self, substring: str):\n self._substring = substring\n\n def match(self, cursor: Cursor, groups: Groups) -> Status:\n if cursor.index >= cursor.length: return Status.STOP\n if len(self._substring) > cursor.length: return Status.STOP\n\n for character in self._substring:\n if cursor.current == character: cursor.next()\n else: return Status.FALSE\n\n return Status.TRUE","repo_name":"SampleUserD/regular-expressions","sub_path":"regexp/matchers/SubstringMatcher.py","file_name":"SubstringMatcher.py","file_ext":"py","file_size_in_byte":492,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"9182629130","text":"import sys\r\n\r\nfrom flask import Flask, render_template, abort\r\nimport post_ticket\r\nimport math\r\n\r\napp = Flask(__name__)\r\n\r\nget_ticket = post_ticket\r\n\r\nALL_TICKETS = 'https://zccbuff.zendesk.com/api/v2/tickets.json'\r\nROWS_PER_PAGE = 25\r\nCURRENT_PAGE = 0\r\n\r\ndata = get_ticket.fetch_tickets(ALL_TICKETS)\r\nif data == 401:\r\n print('Authentication error. Please check username or password')\r\n sys.exit()\r\nelse:\r\n TOTAL_PAGES = math.ceil(len(data) / ROWS_PER_PAGE)\r\n\r\n\r\n# For handling common http errors\r\n\r\n# API/Web server is not available\r\n@app.errorhandler(503)\r\ndef service_unavailable(e):\r\n return render_template('503.html'), 503\r\n\r\n\r\n# Authentication error\r\n@app.errorhandler(401)\r\ndef authentication_error(e):\r\n return render_template('401.html'), 401\r\n\r\n\r\n# Website not able to find the request\r\n@app.errorhandler(404)\r\ndef page_not_found(e):\r\n return render_template('404.html'), 404\r\n\r\n\r\n# Internal server error\r\n@app.errorhandler(500)\r\ndef internal_error(e):\r\n return render_template('500.html'), 500\r\n\r\n\r\n# Default page which will display first 25 pages\r\n\r\n@app.route('/', methods=['GET', 'POST'])\r\ndef displayTicket():\r\n global ROWS_PER_PAGE, CURRENT_PAGE\r\n global TOTAL_PAGES\r\n global data\r\n CURRENT_PAGE = 0\r\n page_data = data[0:ROWS_PER_PAGE]\r\n if len(page_data) == 0:\r\n abort(404)\r\n if data == 503:\r\n abort(503)\r\n return render_template('display_ticket.html', data=page_data, current_page=CURRENT_PAGE,\r\n total_pages=TOTAL_PAGES)\r\n\r\n\r\n# Next pages till all tickets are displayed. Each page shows 25 tickets\r\n\r\n@app.route('/', methods=['GET', 'POST'])\r\ndef displayTicket_withPagination(current_page):\r\n global CURRENT_PAGE, ROWS_PER_PAGE\r\n global TOTAL_PAGES\r\n global data\r\n CURRENT_PAGE = int(current_page)\r\n page_data = data[(CURRENT_PAGE * ROWS_PER_PAGE):(CURRENT_PAGE + 1) * ROWS_PER_PAGE]\r\n if len(page_data) == 0:\r\n abort(404)\r\n if data == 503:\r\n abort(503)\r\n return render_template('display_ticket.html', data=page_data, current_page=CURRENT_PAGE, total_pages=TOTAL_PAGES)\r\n\r\n\r\n# For viewing individual tickets\r\n\r\n@app.route('/individualticket/', methods=['GET', 'POST'])\r\ndef displayIndividualTicket(ticketId):\r\n ticket = get_ticket.fetch_ticket_from_url(ticketId)\r\n if ticket == 503:\r\n abort(503)\r\n return render_template('singleticket.html', data=ticket)\r\n\r\n\r\nif __name__ == '__main__':\r\n app.run()\r\n","repo_name":"nilesh-kaizen/Zendesk_CC","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2486,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"42423927709","text":"import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport sklearn as sk\nimport seaborn as sns\n\nfrom Environment import Environment\nfrom operator import add\n\nREWARD_NONTERMINAL = -1\nREWARD_TERMINAL = 10\nREWARD_CLIFF = -100\n\n# Down, right, top, left\nACTION_DIRECTIONS = [(0, 1), (1, 0), (0, -1), (-1, 0)]\n\n\nclass CliffEnvironment(Environment):\n \"\"\"\n Environment, where position coordinates correspond to (x, y).\n [(0,0), (1,0), (2,0), (3,0)]\n [(0,1), (1,1), (2,1), (3,1)]\n [(0,2), (1,2), (2,2), (3,2)]\n [(0,3), (1,3), (2,3), (3,3)]\n \"\"\"\n def __init__(self, nr_columns, nr_rows, nr_actions=4, init_qa_values=0):\n super().__init__(nr_columns, nr_rows)\n self.init_world_rewards()\n\n def init_world_rewards(self):\n \"\"\" Initialize rewards for reaching different states. \"\"\"\n self.world[:, :] = REWARD_NONTERMINAL\n self.world[self.nr_rows - 1:, 1:self.nr_columns - 1] = REWARD_CLIFF\n self.world[self.nr_rows - 1, self.nr_columns - 1] = REWARD_TERMINAL\n\n def next_state(self, state, action_index):\n \"\"\" Returns next-state tuple.\n If: state walks to edge, return same state. \n Else: add action to tuple and return new state\n \"\"\"\n action = ACTION_DIRECTIONS[action_index]\n next_state = tuple(map(add, state, action))\n\n if self.is_out_of_bounds(next_state):\n next_state = state\n\n return next_state\n\n def check_termination(self, state):\n \"\"\" Checks if state is in cliff or in terminal zone. \"\"\"\n return state[1] == self.nr_rows - 1 and state[0] > 0\n\n def spawn_in_environment(self):\n \"\"\"Returns where user spawns after end of episode. In cliff-walking, always same corner. \"\"\"\n return (0, 3)","repo_name":"JMitnik/MASCode","sub_path":"Cliff Walking/CliffEnvironment.py","file_name":"CliffEnvironment.py","file_ext":"py","file_size_in_byte":1798,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"74509954184","text":"from matplotlib.pyplot import axis\nimport numpy as np\n\n\ndef plot(axes, data, x_axis, y_axes, **kwargs):\n \"\"\"2次元直交座標のプロット\n \n Parameters\n ----------\n axes : :obj:`matplotlib.Axes`\n プロットする軸\n data : dict\n ファイルから読み込んだデータ\n x_axis : str\n x軸に使用するデータ系列名\n y_axes : list of str\n y軸に使用するデータ系列名のリスト\n ymin : float (option)\n y軸の最小値\n ymax : float (option)\n y軸の最大値\n \"\"\"\n \n axes.set_xlabel(x_axis)\n\n xmin = min(data[x_axis])\n xmax = max(data[x_axis])\n\n axes.set_xlim([xmin, xmax])\n\n if 'ymax' in kwargs:\n axes.set_ylim([kwargs['ymin'], kwargs['ymax']])\n\n for y_axis in y_axes:\n axes.plot(data[x_axis], data[y_axis], label=y_axis, linewidth=3)","repo_name":"yn4k4nishi/aemwel_viewer","sub_path":"plot/cartesian2D.py","file_name":"cartesian2D.py","file_ext":"py","file_size_in_byte":878,"program_lang":"python","lang":"ja","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"19400142300","text":"import os\nimport logging\nfrom itertools import groupby\n\nfrom django.conf import settings\nfrom django.contrib.auth.models import User\nfrom django.core.cache import cache\nfrom django.core.exceptions import PermissionDenied, ObjectDoesNotExist\nfrom django.db.models import Q\nfrom django.http import HttpResponse, Http404\nfrom django.shortcuts import get_object_or_404, render_to_response\nfrom django.template import loader, RequestContext\nfrom django.utils.translation import to_locale, ugettext as _\nfrom django.utils.translation.trans_real import parse_accept_lang_header\nfrom django.utils import simplejson, timezone\nfrom django.utils.encoding import iri_to_uri\nfrom django.views.decorators.cache import never_cache\n\nfrom translate.lang import data\n\nfrom pootle.core.decorators import (get_translation_project,\n set_request_context)\nfrom pootle_app.models import Suggestion as SuggestionStat\nfrom pootle_app.models.permissions import (get_matching_permissions,\n check_permission,\n check_profile_permission)\nfrom pootle_misc.baseurl import redirect\nfrom pootle_misc.checks import check_names, get_quality_check_failures\nfrom pootle_misc.forms import make_search_form\nfrom pootle_misc.stats import get_raw_stats\nfrom pootle_misc.url_manip import ensure_uri, previous_view_url\nfrom pootle_misc.util import paginate, ajax_required, jsonify\nfrom pootle_profile.models import get_profile\nfrom pootle_statistics.models import (Submission, SubmissionFields,\n SubmissionTypes)\n\nfrom .decorators import get_store_context, get_unit_context\nfrom .models import Store, Unit\nfrom .forms import (unit_comment_form_factory, unit_form_factory,\n highlight_whitespace)\nfrom .signals import translation_submitted\nfrom .templatetags.store_tags import (highlight_diffs, pluralize_source,\n pluralize_target)\nfrom .util import (UNTRANSLATED, FUZZY, TRANSLATED, STATES_MAP,\n absolute_real_path, find_altsrcs, get_sugg_list)\n\n\n@get_store_context('view')\ndef export_as_xliff(request, store):\n \"\"\"Export given file to xliff for offline translation.\"\"\"\n path = store.real_path\n if not path:\n # bug 2106\n project = request.translation_project.project\n if project.get_treestyle() == \"gnu\":\n path = \"/\".join(store.pootle_path.split(os.path.sep)[2:])\n else:\n parts = store.pootle_path.split(os.path.sep)[1:]\n path = \"%s/%s/%s\" % (parts[1], parts[0], \"/\".join(parts[2:]))\n\n path, ext = os.path.splitext(path)\n export_path = \"/\".join(['POOTLE_EXPORT', path + os.path.extsep + 'xlf'])\n abs_export_path = absolute_real_path(export_path)\n\n key = iri_to_uri(\"%s:export_as_xliff\" % store.pootle_path)\n last_export = cache.get(key)\n if (not (last_export and last_export == store.get_mtime() and\n os.path.isfile(abs_export_path))):\n from pootle_app.project_tree import ensure_target_dir_exists\n from translate.storage.poxliff import PoXliffFile\n from pootle_misc import ptempfile as tempfile\n import shutil\n ensure_target_dir_exists(abs_export_path)\n outputstore = store.convert(PoXliffFile)\n outputstore.switchfile(store.name, createifmissing=True)\n fd, tempstore = tempfile.mkstemp(prefix=store.name, suffix='.xlf')\n os.close(fd)\n outputstore.savefile(tempstore)\n shutil.move(tempstore, abs_export_path)\n cache.set(key, store.get_mtime(), settings.OBJECT_CACHE_TIMEOUT)\n return redirect('/export/' + export_path)\n\n\n@get_store_context('view')\ndef export_as_type(request, store, filetype):\n \"\"\"Export given file to xliff for offline translation.\"\"\"\n from pootle_store.filetypes import factory_classes, is_monolingual\n klass = factory_classes.get(filetype, None)\n if (not klass or is_monolingual(klass) or\n store.pootle_path.endswith(filetype)):\n raise ValueError\n\n path, ext = os.path.splitext(store.real_path)\n export_path = os.path.join('POOTLE_EXPORT',\n path + os.path.extsep + filetype)\n abs_export_path = absolute_real_path(export_path)\n\n key = iri_to_uri(\"%s:export_as_%s\" % (store.pootle_path, filetype))\n last_export = cache.get(key)\n if (not (last_export and last_export == store.get_mtime() and\n os.path.isfile(abs_export_path))):\n from pootle_app.project_tree import ensure_target_dir_exists\n from pootle_misc import ptempfile as tempfile\n import shutil\n ensure_target_dir_exists(abs_export_path)\n outputstore = store.convert(klass)\n fd, tempstore = tempfile.mkstemp(prefix=store.name,\n suffix=os.path.extsep + filetype)\n os.close(fd)\n outputstore.savefile(tempstore)\n shutil.move(tempstore, abs_export_path)\n cache.set(key, store.get_mtime(), settings.OBJECT_CACHE_TIMEOUT)\n return redirect('/export/' + export_path)\n\n@get_store_context('view')\ndef download(request, store):\n store.sync(update_translation=True)\n return redirect('/export/' + store.real_path)\n\n\ndef get_filter_name(GET):\n \"\"\"Gets current filter's human-readable name.\n\n :param GET: A copy of ``request.GET``.\n :return: Two-tuple with the filter name, and a list of extra arguments\n passed to the current filter.\n \"\"\"\n filter = extra = None\n\n if 'filter' in GET:\n filter = GET['filter']\n\n if filter.startswith('user-'):\n extra = [GET.get('user', _('User missing'))]\n elif filter == 'checks' and 'checks' in GET:\n extra = map(lambda check: check_names.get(check, check),\n GET['checks'].split(','))\n elif 'search' in GET:\n filter = 'search'\n\n extra = [GET['search']]\n if 'sfields' in GET:\n extra.extend(GET['sfields'].split(','))\n\n filter_name = {\n 'all': _('All'),\n 'translated': _('Translated'),\n 'untranslated': _('Untranslated'),\n 'fuzzy': _('Needs work'),\n 'incomplete': _('Incomplete'),\n # Translators: This is the name of a filter\n 'search': _('Search'),\n 'checks': _('Checks'),\n 'user-submissions': _('Submissions'),\n 'user-submissions-overwritten': _('Overwritten submissions'),\n }.get(filter)\n\n return (filter_name, extra)\n\n\n@get_translation_project\n@set_request_context\ndef export_view(request, translation_project, dir_path, filename=None):\n \"\"\"Displays a list of units with filters applied.\"\"\"\n current_path = translation_project.directory.pootle_path + dir_path\n\n if filename:\n current_path = current_path + filename\n store = get_object_or_404(Store, pootle_path=current_path)\n units_qs = store.units\n else:\n store = None\n units_qs = translation_project.units.filter(\n store__pootle_path__startswith=current_path,\n )\n\n filter_name, filter_extra = get_filter_name(request.GET)\n\n units = get_step_query(request, units_qs)\n unit_groups = [(path, list(units)) for path, units in\n groupby(units, lambda x: x.store.path)]\n\n ctx = {\n 'source_language': translation_project.project.source_language,\n 'language': translation_project.language,\n 'project': translation_project.project,\n 'unit_groups': unit_groups,\n 'filter_name': filter_name,\n 'filter_extra': filter_extra,\n }\n\n return render_to_response('store/list.html', ctx,\n context_instance=RequestContext(request))\n\n\n####################### Translate Page ##############################\n\ndef get_alt_src_langs(request, profile, translation_project):\n language = translation_project.language\n project = translation_project.project\n source_language = project.source_language\n\n langs = profile.alt_src_langs.exclude(\n id__in=(language.id, source_language.id)\n ).filter(translationproject__project=project)\n\n if not profile.alt_src_langs.count():\n from pootle_language.models import Language\n accept = request.META.get('HTTP_ACCEPT_LANGUAGE', '')\n\n for accept_lang, unused in parse_accept_lang_header(accept):\n if accept_lang == '*':\n continue\n\n simplified = data.simplify_to_common(accept_lang)\n normalized = to_locale(data.normalize_code(simplified))\n code = to_locale(accept_lang)\n if (normalized in\n ('en', 'en_US', source_language.code, language.code) or\n code in ('en', 'en_US', source_language.code, language.code)):\n continue\n\n langs = Language.objects.filter(\n code__in=(normalized, code),\n translationproject__project=project,\n )\n if langs.count():\n break\n\n return langs\n\n\ndef get_non_indexed_search_step_query(form, units_queryset):\n words = form.cleaned_data['search'].split()\n result = units_queryset.none()\n\n if 'source' in form.cleaned_data['sfields']:\n subresult = units_queryset\n for word in words:\n subresult = subresult.filter(source_f__icontains=word)\n result = result | subresult\n\n if 'target' in form.cleaned_data['sfields']:\n subresult = units_queryset\n for word in words:\n subresult = subresult.filter(target_f__icontains=word)\n result = result | subresult\n\n if 'notes' in form.cleaned_data['sfields']:\n translator_subresult = units_queryset\n developer_subresult = units_queryset\n for word in words:\n translator_subresult = translator_subresult.filter(\n translator_comment__icontains=word,\n )\n developer_subresult = developer_subresult.filter(\n developer_comment__icontains=word,\n )\n result = result | translator_subresult | developer_subresult\n\n if 'locations' in form.cleaned_data['sfields']:\n subresult = units_queryset\n for word in words:\n subresult = subresult.filter(locations__icontains=word)\n result = result | subresult\n\n return result\n\ndef get_non_indexed_search_exact_query(form, units_queryset):\n phrase = form.cleaned_data['search']\n result = units_queryset.none()\n\n if 'source' in form.cleaned_data['sfields']:\n subresult = units_queryset.filter(source_f__contains=phrase)\n result = result | subresult\n\n if 'target' in form.cleaned_data['sfields']:\n subresult = units_queryset.filter(target_f__contains=phrase)\n result = result | subresult\n\n if 'notes' in form.cleaned_data['sfields']:\n translator_subresult = units_queryset\n developer_subresult = units_queryset\n translator_subresult = translator_subresult.filter(\n translator_comment__contains=phrase,\n )\n developer_subresult = developer_subresult.filter(\n developer_comment__contains=phrase,\n )\n result = result | translator_subresult | developer_subresult\n\n if 'locations' in form.cleaned_data['sfields']:\n subresult = units_queryset.filter(locations__contains=phrase)\n result = result | subresult\n\n return result\n\ndef get_search_step_query(translation_project, form, units_queryset):\n \"\"\"Narrows down units query to units matching search string.\"\"\"\n\n if 'exact' in form.cleaned_data['soptions']:\n logging.debug(u\"Using exact database search for %s\",\n translation_project)\n return get_non_indexed_search_exact_query(form, units_queryset)\n\n if translation_project.indexer is None:\n logging.debug(u\"No indexer for %s, using database search\",\n translation_project)\n return get_non_indexed_search_step_query(form, units_queryset)\n\n logging.debug(u\"Found %s indexer for %s, using indexed search\",\n translation_project.indexer.INDEX_DIRECTORY_NAME,\n translation_project)\n\n word_querylist = []\n words = form.cleaned_data['search']\n fields = form.cleaned_data['sfields']\n paths = units_queryset.order_by() \\\n .values_list('store__pootle_path', flat=True) \\\n .distinct()\n path_querylist = [('pofilename', pootle_path)\n for pootle_path in paths.iterator()]\n cache_key = \"search:%s\" % str(hash((repr(path_querylist),\n translation_project.get_mtime(),\n repr(words),\n repr(fields))))\n\n dbids = cache.get(cache_key)\n if dbids is None:\n searchparts = []\n word_querylist = [(field, words) for field in fields]\n textquery = translation_project.indexer.make_query(word_querylist,\n False)\n searchparts.append(textquery)\n\n pathquery = translation_project.indexer.make_query(path_querylist,\n False)\n searchparts.append(pathquery)\n limitedquery = translation_project.indexer.make_query(searchparts, True)\n\n result = translation_project.indexer.search(limitedquery, ['dbid'])\n dbids = [int(item['dbid'][0]) for item in result[:999]]\n cache.set(cache_key, dbids, settings.OBJECT_CACHE_TIMEOUT)\n\n return units_queryset.filter(id__in=dbids)\n\n\ndef get_step_query(request, units_queryset):\n \"\"\"Narrows down unit query to units matching conditions in GET.\"\"\"\n if 'filter' in request.GET:\n unit_filter = request.GET['filter']\n username = request.GET.get('user', None)\n\n profile = request.profile\n if username:\n try:\n user = User.objects.get(username=username)\n profile = user.get_profile()\n except User.DoesNotExist:\n pass\n\n if unit_filter:\n match_queryset = units_queryset.none()\n\n if unit_filter == 'all':\n match_queryset = units_queryset\n elif unit_filter == 'translated':\n match_queryset = units_queryset.filter(state=TRANSLATED)\n elif unit_filter == 'untranslated':\n match_queryset = units_queryset.filter(state=UNTRANSLATED)\n elif unit_filter == 'fuzzy':\n match_queryset = units_queryset.filter(state=FUZZY)\n elif unit_filter == 'incomplete':\n match_queryset = units_queryset.filter(\n Q(state=UNTRANSLATED) | Q(state=FUZZY),\n )\n elif unit_filter == 'suggestions':\n #FIXME: is None the most efficient query\n match_queryset = units_queryset.exclude(suggestion=None)\n elif unit_filter == 'user-suggestions':\n match_queryset = units_queryset.filter(\n suggestion__user=profile,\n ).distinct()\n elif unit_filter == 'user-suggestions-accepted':\n # FIXME: Oh, this is pretty lame, we need a completely\n # different way to model suggestions\n unit_ids = SuggestionStat.objects.filter(\n suggester=profile,\n state='accepted',\n ).values_list('unit', flat=True)\n match_queryset = units_queryset.filter(\n id__in=unit_ids,\n ).distinct()\n elif unit_filter == 'user-suggestions-rejected':\n # FIXME: Oh, this is as lame as above\n unit_ids = SuggestionStat.objects.filter(\n suggester=profile,\n state='rejected',\n ).values_list('unit', flat=True)\n match_queryset = units_queryset.filter(\n id__in=unit_ids,\n ).distinct()\n elif unit_filter == 'user-submissions':\n match_queryset = units_queryset.filter(\n submission__submitter=profile,\n ).distinct()\n elif unit_filter == 'user-submissions-overwritten':\n match_queryset = units_queryset.filter(\n submission__submitter=profile,\n ).exclude(submitted_by=profile).distinct()\n elif unit_filter == 'checks' and 'checks' in request.GET:\n checks = request.GET['checks'].split(',')\n\n if checks:\n match_queryset = units_queryset.filter(\n qualitycheck__false_positive=False,\n qualitycheck__name__in=checks\n ).distinct()\n\n\n units_queryset = match_queryset\n\n if 'search' in request.GET and 'sfields' in request.GET:\n # use the search form for validation only\n search_form = make_search_form(request.GET)\n\n if search_form.is_valid():\n units_queryset = get_search_step_query(request.translation_project,\n search_form, units_queryset)\n\n return units_queryset\n\n\ndef translate_page(request):\n cantranslate = check_permission(\"translate\", request)\n cansuggest = check_permission(\"suggest\", request)\n canreview = check_permission(\"review\", request)\n\n translation_project = request.translation_project\n language = translation_project.language\n project = translation_project.project\n profile = request.profile\n\n store = getattr(request, \"store\", None)\n directory = getattr(request, \"directory\", None)\n\n is_single_file = store and True or False\n path = is_single_file and store.path or directory.path\n pootle_path = (is_single_file and store.pootle_path or\n directory.pootle_path)\n\n is_terminology = (project.is_terminology or store and\n store.is_terminology)\n search_form = make_search_form(request=request,\n terminology=is_terminology)\n\n previous_overview_url = previous_view_url(request, ['overview'])\n\n context = {\n 'cantranslate': cantranslate,\n 'cansuggest': cansuggest,\n 'canreview': canreview,\n 'search_form': search_form,\n 'store': store,\n 'store_id': store and store.id,\n 'directory': directory,\n 'directory_id': directory and directory.id,\n 'path': path,\n 'pootle_path': pootle_path,\n 'is_single_file': is_single_file,\n 'language': language,\n 'project': project,\n 'translation_project': translation_project,\n 'profile': profile,\n 'source_language': translation_project.project.source_language,\n 'previous_overview_url': previous_overview_url,\n 'MT_BACKENDS': settings.MT_BACKENDS,\n 'LOOKUP_BACKENDS': settings.LOOKUP_BACKENDS,\n 'AMAGAMA_URL': settings.AMAGAMA_URL,\n }\n\n return render_to_response('store/translate.html', context,\n context_instance=RequestContext(request))\n\n\n@get_store_context('view')\ndef translate(request, store):\n return translate_page(request)\n\n#\n# Views used with XMLHttpRequest requests.\n#\n\ndef _filter_ctx_units(units_qs, unit, how_many, gap=0):\n \"\"\"Returns ``how_many``*2 units that are before and after ``index``.\"\"\"\n result = {'before': [], 'after': []}\n\n if how_many and unit.index - gap > 0:\n before = units_qs.filter(store=unit.store_id, index__lt=unit.index) \\\n .order_by('-index')[gap:how_many+gap]\n result['before'] = _build_units_list(before, reverse=True)\n result['before'].reverse()\n\n #FIXME: can we avoid this query if length is known?\n if how_many:\n after = units_qs.filter(store=unit.store_id,\n index__gt=unit.index)[gap:how_many+gap]\n result['after'] = _build_units_list(after)\n\n return result\n\ndef _build_units_list(units, reverse=False):\n \"\"\"Given a list/queryset of units, builds a list with the unit data\n contained in a dictionary ready to be returned as JSON.\n\n :return: A list with unit id, source, and target texts. In case of\n having plural forms, a title for the plural form is also provided.\n \"\"\"\n return_units = []\n for unit in iter(units):\n source_unit = []\n target_unit = []\n for i, source, title in pluralize_source(unit):\n unit_dict = {'text': source}\n if title:\n unit_dict[\"title\"] = title\n source_unit.append(unit_dict)\n for i, target, title in pluralize_target(unit):\n unit_dict = {'text': target}\n if title:\n unit_dict[\"title\"] = title\n target_unit.append(unit_dict)\n prev = None\n next = None\n if return_units:\n if reverse:\n return_units[-1]['prev'] = unit.id\n next = return_units[-1]['id']\n else:\n return_units[-1]['next'] = unit.id\n prev = return_units[-1]['id']\n return_units.append({'id': unit.id,\n 'isfuzzy': unit.isfuzzy(),\n 'prev': prev,\n 'next': next,\n 'source': source_unit,\n 'target': target_unit})\n return return_units\n\n\ndef _build_pager_dict(pager):\n \"\"\"Given a pager object ``pager``, retrieves all the information needed\n to build a pager.\n\n :return: A dictionary containing necessary pager information to build\n a pager.\n \"\"\"\n return {\"number\": pager.number,\n \"num_pages\": pager.paginator.num_pages,\n \"per_page\": pager.paginator.per_page\n }\n\n\ndef _get_index_in_qs(qs, unit, store=False):\n \"\"\"Given a queryset ``qs``, returns the position (index) of the unit\n ``unit`` within that queryset. ``store`` specifies if the queryset is\n limited to a single store.\n\n :return: Value representing the position of the unit ``unit``.\n :rtype: int\n \"\"\"\n if store:\n return qs.filter(index__lt=unit.index).count()\n else:\n store = unit.store\n return (qs.filter(store=store, index__lt=unit.index) | \\\n qs.filter(store__pootle_path__lt=store.pootle_path)).count()\n\n\ndef get_view_units(request, units_queryset, store, limit=0):\n \"\"\"Gets source and target texts excluding the editing unit.\n\n :return: An object in JSON notation that contains the source and target\n texts for units that will be displayed before and after editing\n unit.\n\n If asked by using the ``meta`` and ``pager`` parameters,\n metadata and pager information will be calculated and returned\n too.\n \"\"\"\n current_unit = None\n json = {}\n\n try:\n limit = int(limit)\n except ValueError:\n limit = None\n\n if not limit:\n limit = request.profile.get_unit_rows()\n\n step_queryset = get_step_query(request, units_queryset)\n\n # Return metadata it has been explicitely requested\n if request.GET.get('meta', False):\n tp = request.translation_project\n json[\"meta\"] = {\"source_lang\": tp.project.source_language.code,\n \"source_dir\": tp.project.source_language.get_direction(),\n \"target_lang\": tp.language.code,\n \"target_dir\": tp.language.get_direction(),\n \"project_style\": tp.project.checkstyle}\n\n # Maybe we are trying to load directly a specific unit, so we have\n # to calculate its page number\n uid = request.GET.get('uid', None)\n if uid:\n current_unit = units_queryset.get(id=uid)\n preceding = _get_index_in_qs(step_queryset, current_unit, store)\n page = preceding / limit + 1\n else:\n page = None\n\n pager = paginate(request, step_queryset, items=limit, page=page)\n\n json[\"units\"] = _build_units_list(pager.object_list)\n\n # Return paging information if requested to do so\n if request.GET.get('pager', False):\n json[\"pager\"] = _build_pager_dict(pager)\n if not current_unit:\n try:\n json[\"uid\"] = json[\"units\"][0][\"id\"]\n except IndexError:\n pass\n else:\n json[\"uid\"] = current_unit.id\n\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_store_context('view')\ndef get_view_units_store(request, store, limit=0):\n \"\"\"Gets source and target texts excluding the editing widget (store-level).\n\n :return: An object in JSON notation that contains the source and target\n texts for units that will be displayed before and after\n unit ``uid``.\n \"\"\"\n return get_view_units(request, store.units, store=True, limit=limit)\n\n\ndef _is_filtered(request):\n \"\"\"Checks if unit list is filtered.\"\"\"\n return ('filter' in request.GET or 'checks' in request.GET or\n 'user' in request.GET or\n ('search' in request.GET and 'sfields' in request.GET))\n\n\n@ajax_required\n@get_unit_context('view')\ndef get_more_context(request, unit):\n \"\"\"Retrieves more context units.\n\n :return: An object in JSON notation that contains the source and target\n texts for units that are in the context of unit ``uid``.\n \"\"\"\n store = request.store\n json = {}\n gap = int(request.GET.get('gap', 0))\n qty = int(request.GET.get('qty', 1))\n\n json[\"ctx\"] = _filter_ctx_units(store.units, unit, qty, gap)\n rcode = 200\n response = jsonify(json)\n return HttpResponse(response, status=rcode, mimetype=\"application/json\")\n\n\n@never_cache\n@get_unit_context('view')\ndef timeline(request, unit):\n \"\"\"Returns a JSON-encoded string including the changes to the unit\n rendered in HTML.\n \"\"\"\n timeline = Submission.objects.filter(unit=unit, field__in=[\n SubmissionFields.TARGET, SubmissionFields.STATE,\n SubmissionFields.COMMENT\n ])\n timeline = timeline.select_related(\"submitter__user\",\n \"translation_project__language\")\n\n context = {}\n entries_group = []\n\n import locale\n from pootle_store.fields import to_python\n\n for key, values in groupby(timeline, key=lambda x: x.creation_time):\n entry_group = {\n 'datetime': key,\n 'datetime_str': key.strftime(locale.nl_langinfo(locale.D_T_FMT)),\n 'entries': [],\n }\n\n for item in values:\n # Only add submitter information for the whole entry group once\n entry_group.setdefault('submitter', item.submitter)\n\n context.setdefault('language', item.translation_project.language)\n\n entry = {\n 'field': item.field,\n 'field_name': SubmissionFields.NAMES_MAP[item.field],\n }\n\n if item.field == SubmissionFields.STATE:\n entry['old_value'] = STATES_MAP[int(to_python(item.old_value))]\n entry['new_value'] = STATES_MAP[int(to_python(item.new_value))]\n else:\n entry['new_value'] = to_python(item.new_value)\n\n entry_group['entries'].append(entry)\n\n entries_group.append(entry_group)\n\n # Let's reverse the chronological order\n entries_group.reverse()\n\n # Remove first timeline item if it's solely a change to the target\n if (entries_group and len(entries_group[0]['entries']) == 1 and\n entries_group[0]['entries'][0]['field'] == SubmissionFields.TARGET):\n del entries_group[0]\n\n context['entries_group'] = entries_group\n\n if request.is_ajax():\n # The client will want to confirm that the response is relevant for\n # the unit on screen at the time of receiving this, so we add the uid.\n json = {'uid': unit.id}\n\n t = loader.get_template('unit/xhr-timeline.html')\n c = RequestContext(request, context)\n json['timeline'] = t.render(c).replace('\\n', '')\n\n response = simplejson.dumps(json)\n return HttpResponse(response, mimetype=\"application/json\")\n else:\n return render_to_response('unit/timeline.html', context,\n context_instance=RequestContext(request))\n\n\n@ajax_required\n@get_unit_context('translate')\ndef comment(request, unit):\n \"\"\"Stores a new comment for the given ``unit``.\n\n :return: If the form validates, the cleaned comment is returned.\n An error message is returned otherwise.\n \"\"\"\n # Update current unit instance's attributes\n unit.commented_by = request.profile\n unit.commented_on = timezone.now()\n\n language = request.translation_project.language\n form = unit_comment_form_factory(language)(request.POST, instance=unit,\n request=request)\n\n if form.is_valid():\n form.save()\n\n context = {\n 'unit': unit,\n 'language': language,\n }\n t = loader.get_template('unit/comment.html')\n c = RequestContext(request, context)\n\n json = {'comment': t.render(c)}\n rcode = 200\n else:\n json = {'msg': _(\"Comment submission failed.\")}\n rcode = 400\n\n response = simplejson.dumps(json)\n\n return HttpResponse(response, status=rcode, mimetype=\"application/json\")\n\n\n@never_cache\n@ajax_required\n@get_unit_context('view')\ndef get_edit_unit(request, unit):\n \"\"\"Given a store path ``pootle_path`` and unit id ``uid``, gathers all the\n necessary information to build the editing widget.\n\n :return: A templatised editing widget is returned within the ``editor``\n variable and paging information is also returned if the page\n number has changed.\n \"\"\"\n json = {}\n\n translation_project = request.translation_project\n language = translation_project.language\n\n if unit.hasplural():\n snplurals = len(unit.source.strings)\n else:\n snplurals = None\n\n form_class = unit_form_factory(language, snplurals, request)\n form = form_class(instance=unit)\n comment_form_class = unit_comment_form_factory(language)\n comment_form = comment_form_class({}, instance=unit)\n\n store = unit.store\n directory = store.parent\n profile = request.profile\n alt_src_langs = get_alt_src_langs(request, profile, translation_project)\n project = translation_project.project\n report_target = ensure_uri(project.report_target)\n\n suggestions = get_sugg_list(unit)\n template_vars = {\n 'unit': unit,\n 'form': form,\n 'comment_form': comment_form,\n 'store': store,\n 'directory': directory,\n 'profile': profile,\n 'user': request.user,\n 'language': language,\n 'source_language': translation_project.project.source_language,\n 'cantranslate': check_profile_permission(profile, \"translate\",\n directory),\n 'cansuggest': check_profile_permission(profile, \"suggest\", directory),\n 'canreview': check_profile_permission(profile, \"review\", directory),\n 'altsrcs': find_altsrcs(unit, alt_src_langs, store=store,\n project=project),\n 'report_target': report_target,\n 'suggestions': suggestions,\n }\n\n if translation_project.project.is_terminology or store.is_terminology:\n t = loader.get_template('unit/term_edit.html')\n else:\n t = loader.get_template('unit/edit.html')\n c = RequestContext(request, template_vars)\n json['editor'] = t.render(c)\n\n rcode = 200\n\n # Return context rows if filtering is applied but\n # don't return any if the user has asked not to have it\n current_filter = request.GET.get('filter', 'all')\n show_ctx = request.COOKIES.get('ctxShow', 'true')\n\n if ((_is_filtered(request) or current_filter not in ('all',)) and\n show_ctx == 'true'):\n # TODO: review if this first 'if' branch makes sense\n if translation_project.project.is_terminology or store.is_terminology:\n json['ctx'] = _filter_ctx_units(store.units, unit, 0)\n else:\n ctx_qty = int(request.COOKIES.get('ctxQty', 1))\n json['ctx'] = _filter_ctx_units(store.units, unit, ctx_qty)\n\n response = jsonify(json)\n return HttpResponse(response, status=rcode, mimetype=\"application/json\")\n\n\ndef get_failing_checks(request, pathobj):\n \"\"\"Gets a list of failing checks for the current object.\n\n :return: JSON string with a list of failing check categories which\n include the actual checks that are failing.\n \"\"\"\n stats = get_raw_stats(pathobj)\n failures = get_quality_check_failures(pathobj, stats, include_url=False)\n\n response = jsonify(failures)\n\n return HttpResponse(response, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_store_context('view')\ndef get_failing_checks_store(request, store):\n return get_failing_checks(request, store)\n\n\n@ajax_required\n@get_unit_context('')\ndef submit(request, unit):\n \"\"\"Processes translation submissions and stores them in the database.\n\n :return: An object in JSON notation that contains the previous and last\n units for the unit next to unit ``uid``.\n \"\"\"\n json = {}\n\n cantranslate = check_permission(\"translate\", request)\n if not cantranslate:\n raise PermissionDenied(_(\"You do not have rights to access \"\n \"translation mode.\"))\n\n translation_project = request.translation_project\n language = translation_project.language\n\n if unit.hasplural():\n snplurals = len(unit.source.strings)\n else:\n snplurals = None\n\n # Store current time so that it is the same for all submissions\n current_time = timezone.now()\n\n # Update current unit instance's attributes\n unit.submitted_by = request.profile\n unit.submitted_on = current_time\n\n form_class = unit_form_factory(language, snplurals, request)\n form = form_class(request.POST, instance=unit)\n\n if form.is_valid():\n if form.updated_fields:\n for field, old_value, new_value in form.updated_fields:\n sub = Submission(\n creation_time=current_time,\n translation_project=translation_project,\n submitter=request.profile,\n unit=unit,\n field=field,\n type=SubmissionTypes.NORMAL,\n old_value=old_value,\n new_value=new_value,\n )\n sub.save()\n\n form.save()\n translation_submitted.send(\n sender=translation_project,\n unit=form.instance,\n profile=request.profile,\n )\n\n rcode = 200\n else:\n # Form failed\n #FIXME: we should display validation errors here\n rcode = 400\n json[\"msg\"] = _(\"Failed to process submission.\")\n response = jsonify(json)\n return HttpResponse(response, status=rcode, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_unit_context('')\ndef suggest(request, unit):\n \"\"\"Processes translation suggestions and stores them in the database.\n\n :return: An object in JSON notation that contains the previous and last\n units for the unit next to unit ``uid``.\n \"\"\"\n json = {}\n\n cansuggest = check_permission(\"suggest\", request)\n if not cansuggest:\n raise PermissionDenied(_(\"You do not have rights to access \"\n \"translation mode.\"))\n\n translation_project = request.translation_project\n language = translation_project.language\n\n if unit.hasplural():\n snplurals = len(unit.source.strings)\n else:\n snplurals = None\n\n form_class = unit_form_factory(language, snplurals, request)\n form = form_class(request.POST, instance=unit)\n\n if form.is_valid():\n if form.instance._target_updated:\n # TODO: Review if this hackish method is still necessary\n #HACKISH: django 1.2 stupidly modifies instance on\n # model form validation, reload unit from db\n unit = Unit.objects.get(id=unit.id)\n sugg = unit.add_suggestion(form.cleaned_data['target_f'],\n request.profile)\n if sugg:\n SuggestionStat.objects.get_or_create(\n translation_project=translation_project,\n suggester=request.profile, state='pending', unit=unit.id\n )\n rcode = 200\n else:\n # Form failed\n #FIXME: we should display validation errors here\n rcode = 400\n json[\"msg\"] = _(\"Failed to process suggestion.\")\n response = jsonify(json)\n return HttpResponse(response, status=rcode, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_unit_context('')\ndef reject_suggestion(request, unit, suggid):\n json = {}\n translation_project = request.translation_project\n\n json[\"udbid\"] = unit.id\n json[\"sugid\"] = suggid\n if request.POST.get('reject'):\n try:\n sugg = unit.suggestion_set.get(id=suggid)\n except ObjectDoesNotExist:\n raise Http404\n\n if (not check_permission('review', request) and\n (not request.user.is_authenticated() or sugg and\n sugg.user != request.profile)):\n raise PermissionDenied(_(\"You do not have rights to access \"\n \"review mode.\"))\n\n success = unit.reject_suggestion(suggid)\n if sugg is not None and success:\n # FIXME: we need a totally different model for tracking stats, this\n # is just lame\n suggstat, created = SuggestionStat.objects.get_or_create(\n translation_project=translation_project,\n suggester=sugg.user,\n state='pending',\n unit=unit.id,\n )\n suggstat.reviewer = request.profile\n suggstat.state = 'rejected'\n suggstat.save()\n\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_unit_context('review')\ndef accept_suggestion(request, unit, suggid):\n json = {\n 'udbid': unit.id,\n 'sugid': suggid,\n }\n translation_project = request.translation_project\n\n if request.POST.get('accept'):\n try:\n suggestion = unit.suggestion_set.get(id=suggid)\n except ObjectDoesNotExist:\n raise Http404\n\n old_target = unit.target\n success = unit.accept_suggestion(suggid)\n\n json['newtargets'] = [highlight_whitespace(target)\n for target in unit.target.strings]\n json['newdiffs'] = {}\n for sugg in unit.get_suggestions():\n json['newdiffs'][sugg.id] = \\\n [highlight_diffs(unit.target.strings[i], target)\n for i, target in enumerate(sugg.target.strings)]\n\n if suggestion is not None and success:\n if suggestion.user:\n translation_submitted.send(sender=translation_project,\n unit=unit, profile=suggestion.user)\n\n # FIXME: we need a totally different model for tracking stats, this\n # is just lame\n suggstat, created = SuggestionStat.objects.get_or_create(\n translation_project=translation_project,\n suggester=suggestion.user,\n state='pending',\n unit=unit.id,\n )\n suggstat.reviewer = request.profile\n suggstat.state = 'accepted'\n suggstat.save()\n\n # For now assume the target changed\n # TODO: check all fields for changes\n creation_time = timezone.now()\n sub = Submission(\n creation_time=creation_time,\n translation_project=translation_project,\n submitter=suggestion.user,\n from_suggestion=suggstat,\n unit=unit,\n field=SubmissionFields.TARGET,\n type=SubmissionTypes.SUGG_ACCEPT,\n old_value=old_target,\n new_value=unit.target,\n )\n sub.save()\n\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n\n@ajax_required\ndef clear_vote(request, voteid):\n json = {}\n json[\"voteid\"] = voteid\n if request.POST.get('clear'):\n try:\n from voting.models import Vote\n vote = Vote.objects.get(pk=voteid)\n if vote.user != request.user:\n # No i18n, will not go to UI\n raise PermissionDenied(\"Users can only remove their own votes\")\n vote.delete()\n except ObjectDoesNotExist:\n raise Http404\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_unit_context('')\ndef vote_up(request, unit, suggid):\n json = {}\n json[\"suggid\"] = suggid\n if request.POST.get('up'):\n try:\n suggestion = unit.suggestion_set.get(id=suggid)\n from voting.models import Vote\n # Why can't it just return the vote object?\n Vote.objects.record_vote(suggestion, request.user, +1)\n json[\"voteid\"] = Vote.objects.get_for_user(suggestion,\n request.user).id\n except ObjectDoesNotExist:\n raise Http404(_(\"The suggestion or vote is not valid any more.\"))\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n\n\n@ajax_required\n@get_unit_context('review')\ndef reject_qualitycheck(request, unit, checkid):\n json = {}\n json[\"udbid\"] = unit.id\n json[\"checkid\"] = checkid\n if request.POST.get('reject'):\n try:\n check = unit.qualitycheck_set.get(id=checkid)\n check.false_positive = True\n check.save()\n # update timestamp\n unit.save()\n except ObjectDoesNotExist:\n raise Http404\n\n response = jsonify(json)\n return HttpResponse(response, mimetype=\"application/json\")\n","repo_name":"moriaty2013/pootle","sub_path":"pootle/apps/pootle_store/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":42617,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"354237515","text":"#!/usr/bin/python\n\nfrom PyQt5.QtCore import QDateTime, Qt, QTimer\nfrom PyQt5.QtWidgets import (QApplication, QCheckBox, QComboBox, QDateTimeEdit,\n QDial, QDialog, QGridLayout, QGroupBox, QHBoxLayout, QLabel, QLineEdit,\n QProgressBar, QPushButton, QRadioButton, QScrollBar, QSizePolicy,\n QSlider, QSpinBox, QStyleFactory, QTableWidget, QTabWidget, QTextEdit,\n QVBoxLayout, QWidget)\n\nimport sys\n\n\nclass WidgetGallery(QDialog):\n def __init__(self, parent=None):\n super(WidgetGallery, self).__init__(parent)\n\n self.originalPalette = QApplication.palette()\n\n styleComboBox = QComboBox()\n styleComboBox.addItems(QStyleFactory.keys())\n\n styleLabel = QLabel(\"E&stilo:\")\n styleLabel.setBuddy(styleComboBox)\n \n\n self.useStylePaletteCheckBox = QCheckBox(\"&Usar estilos predeterminados\")\n self.useStylePaletteCheckBox.setChecked(True)\n\n disableWidgetsCheckBox = QCheckBox(\"&Deshabilitar widgets\")\n\n self.createTopLeftGroupBox()\n self.createTopRightGroupBox()\n self.createbottomRightTabWidget()\n self.createbottomLeftGroupBox()\n # self.createProgressBar()\n\n styleComboBox.activated[str].connect(self.changeStyle)\n self.useStylePaletteCheckBox.toggled.connect(self.changePalette)\n disableWidgetsCheckBox.toggled.connect(self.topLeftGroupBox.setDisabled)\n disableWidgetsCheckBox.toggled.connect(self.topRightGroupBox.setDisabled)\n disableWidgetsCheckBox.toggled.connect(self.bottomRightTabWidget.setDisabled)\n disableWidgetsCheckBox.toggled.connect(self.bottomLeftGroupBox.setDisabled)\n\n topLayout = QHBoxLayout()\n topLayout.addWidget(styleLabel)\n topLayout.addWidget(styleComboBox)\n topLayout.addStretch(1)\n topLayout.addWidget(self.useStylePaletteCheckBox)\n topLayout.addWidget(disableWidgetsCheckBox)\n\n mainLayout = QGridLayout()\n mainLayout.addLayout(topLayout, 0, 0, 1, 2)\n mainLayout.addWidget(self.topLeftGroupBox, 1, 0)\n mainLayout.addWidget(self.topRightGroupBox, 1, 1)\n mainLayout.addWidget(self.bottomLeftGroupBox, 2, 0)\n mainLayout.addWidget(self.bottomRightTabWidget, 2, 1)\n # mainLayout.addWidget(self.progressBar, 3, 0, 1, 2)\n mainLayout.setRowStretch(1, 1)\n mainLayout.setRowStretch(2, 1)\n mainLayout.setColumnStretch(0, 1)\n mainLayout.setColumnStretch(1, 1)\n self.setLayout(mainLayout)\n\n self.setWindowTitle(\"Proyecto X\")\n self.changeStyle('fusion')\n\n def changeStyle(self, styleName):\n QApplication.setStyle(QStyleFactory.create(styleName))\n self.changePalette()\n\n def changePalette(self):\n if self.useStylePaletteCheckBox.isChecked():\n QApplication.setPalette(QApplication.style().standardPalette())\n else:\n QApplication.setPalette(self.originalPalette)\n\n def advanceProgressBar(self):\n curVal = self.progressBar.value()\n maxVal = self.progressBar.maximum()\n self.progressBar.setValue(curVal + (maxVal - curVal) / 100)\n\n def createTopLeftGroupBox(self):\n self.topLeftGroupBox = QGroupBox(\"Input\")\n\n radioButton1 = QRadioButton(\"Radio button 1\")\n radioButton2 = QRadioButton(\"Radio button 2\")\n radioButton3 = QRadioButton(\"Radio button 3\")\n radioButton1.setChecked(True)\n\n checkBox = QCheckBox(\"Tri-state check box\")\n checkBox.setTristate(True)\n checkBox.setCheckState(Qt.PartiallyChecked)\n\n layout = QVBoxLayout()\n layout.addWidget(radioButton1)\n layout.addWidget(radioButton2)\n layout.addWidget(radioButton3)\n layout.addWidget(checkBox)\n layout.addStretch(1)\n self.topLeftGroupBox.setLayout(layout)\n\n def createTopRightGroupBox(self):\n self.topRightGroupBox = QGroupBox(\"Output\")\n\n defaultPushButton = QPushButton(\"Default Push Button\")\n defaultPushButton.setDefault(True)\n\n togglePushButton = QPushButton(\"Toggle Push Button\")\n togglePushButton.setCheckable(True)\n togglePushButton.setChecked(True)\n\n # flatPushButton = QPushButton(\"Flat Push Button\")\n # flatPushButton.setFlat(True)\n\n layout = QVBoxLayout()\n layout.addWidget(defaultPushButton)\n layout.addWidget(togglePushButton)\n # layout.addWidget(flatPushButton)\n layout.addStretch(1)\n self.topRightGroupBox.setLayout(layout)\n\n def createbottomRightTabWidget(self):\n self.bottomRightTabWidget = QTabWidget()\n self.bottomRightTabWidget.setSizePolicy(QSizePolicy.Preferred,\n QSizePolicy.Ignored)\n\n tab1 = QWidget()\n textEditTokensSourceCode = QTextEdit()\n\n textEditTokensSourceCode.setPlainText(\"salida\")\n\n tab1hbox = QHBoxLayout()\n tab1hbox.setContentsMargins(5, 5, 5, 5)\n tab1hbox.addWidget(textEditTokensSourceCode)\n tab1.setLayout(tab1hbox)\n\n tab2 = QWidget()\n textEditTokensSourceCode = QTextEdit()\n\n textEditTokensSourceCode.setPlainText(\"salida\")\n\n tab2hbox = QHBoxLayout()\n tab2hbox.setContentsMargins(5, 5, 5, 5)\n tab2hbox.addWidget(textEditTokensSourceCode)\n tab2.setLayout(tab2hbox)\n\n tab3 = QWidget()\n textEditTokensSourceCode = QTextEdit()\n\n textEditTokensSourceCode.setPlainText(\"salida\")\n\n tab3hbox = QHBoxLayout()\n tab3hbox.setContentsMargins(5, 5, 5, 5)\n tab3hbox.addWidget(textEditTokensSourceCode)\n tab3.setLayout(tab3hbox)\n\n self.bottomRightTabWidget.addTab(tab1, \"Source Code Tokenize&d\")\n self.bottomRightTabWidget.addTab(tab2, \"Source Code Pars&ed\")\n self.bottomRightTabWidget.addTab(tab3, \"Source Code Trans&piled\")\n\n def createbottomLeftGroupBox(self):\n self.bottomLeftGroupBox = QGroupBox(\"Source Code\")\n self.bottomLeftGroupBox.setCheckable(True)\n self.bottomLeftGroupBox.setChecked(True)\n\n textEditSourceCode = QTextEdit()\n\n textEditSourceCode.setPlainText(\"int variable = 10\\n\")\n\n # lineEdit = QLineEdit('s3cRe7')\n # lineEdit.setEchoMode(QLineEdit.Password)\n\n # spinBox = QSpinBox(self.bottomLeftGroupBox)\n # spinBox.setValue(50)\n\n # dateTimeEdit = QDateTimeEdit(self.bottomLeftGroupBox)\n # dateTimeEdit.setDateTime(QDateTime.currentDateTime())\n\n # slider = QSlider(Qt.Horizontal, self.bottomLeftGroupBox)\n # slider.setValue(40)\n\n # scrollBar = QScrollBar(Qt.Horizontal, self.bottomLeftGroupBox)\n # scrollBar.setValue(60)\n\n # dial = QDial(self.bottomLeftGroupBox)\n # dial.setValue(30)\n # dial.setNotchesVisible(True)\n\n layout = QGridLayout()\n layout.addWidget(textEditSourceCode, 0, 0, 1, 2)\n # layout.addWidget(lineEdit, 0, 0, 1, 2)\n # layout.addWidget(spinBox, 1, 0, 1, 2)\n # layout.addWidget(dateTimeEdit, 2, 0, 1, 2)\n # layout.addWidget(slider, 3, 0)\n # layout.addWidget(scrollBar, 4, 0)\n # layout.addWidget(dial, 3, 1, 2, 1)\n layout.setRowStretch(5, 1)\n self.bottomLeftGroupBox.setLayout(layout)\n\n def createProgressBar(self):\n self.progressBar = QProgressBar()\n self.progressBar.setRange(0, 10000)\n self.progressBar.setValue(0)\n\n timer = QTimer(self)\n timer.timeout.connect(self.advanceProgressBar)\n timer.start(1000)\n\n\nif __name__ == '__main__':\n app = QApplication([])\n gallery = WidgetGallery()\n gallery.show()\n app.exec_()\n","repo_name":"mariano-dim/ProyectoX","sub_path":"src/dashboardCompiler.py","file_name":"dashboardCompiler.py","file_ext":"py","file_size_in_byte":7700,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"5575406388","text":"# **NOTE**: try/except block iterates through 65535 ports!! maximum scan!\nfrom socket import socket, AF_INET, SOCK_STREAM\nfrom datetime import datetime, timedelta\nfrom flask import Flask, render_template\n\napp = Flask(__name__)\n\n@app.route('/')\ndef index():\n hostAddress = '127.0.0.1' # loopback address for scanning localhost\n start_time = datetime.now() # port scan start time\n data = [] # lists open port strings\n try:\n for port in range(1, 65535):\n sock = socket(AF_INET, SOCK_STREAM)\n sock.settimeout(2) # scan for 2 secs\n result = sock.connect_ex((hostAddress, port))\n if result == 0:\n data.append(f'Port {port}: OPEN') # Port __: OPEN\n sock.close()\n except OSError as e:\n if e.errno != errno.ENOENT:\n print(f'{e}')\n sys.exit()\n end_time = datetime.now() # port scan ends: mark time\n duration = end_time - start_time # port scan duration\n data.append(f'Scan duration: {round(duration.total_seconds(), 2)}secs')\n return render_template('portscan.html', data=data, address=hostAddress)\n\nif __name__ == '__main__':\n app.run()\n","repo_name":"nick3499/flask_portscan","sub_path":"portscan.py","file_name":"portscan.py","file_ext":"py","file_size_in_byte":1171,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"3138694182","text":"# Script to resample a raster to a smaller pixel size.\n\nimport os\nfrom osgeo import gdal\n\n# Don't forget to change the folder.\nos.chdir(r'D:\\osgeopy-data\\Landsat\\Washington')\n\n# Open the input raster.\nin_ds = gdal.Open('p047r027_7t20000730_z10_nn10.tif')\nin_band = in_ds.GetRasterBand(1)\n\n# Computer the number of output rows and columns (double the input numbers\n# because we're cutting pixel size in half).\nout_rows = in_band.YSize * 2\nout_columns = in_band.XSize * 2\n\n# Create the output raster using the computed dimensions.\ngtiff_driver = gdal.GetDriverByName('GTiff')\nout_ds = gtiff_driver.Create('band1_resampled.tif',\n out_columns, out_rows)\n\n# Change the geotransform so it reflects the smaller cell size before\n# setting it onto the output.\nout_ds.SetProjection(in_ds.GetProjection())\ngeotransform = list(in_ds.GetGeoTransform())\ngeotransform[1] /= 2\ngeotransform[5] /= 2\nout_ds.SetGeoTransform(geotransform)\n\n# Read in the data, but have gdal resample it so that it has the specified\n# number of rows and columns instead of the numbers that the input has.\n# This effectively resizes the pixels.\ndata = in_band.ReadAsArray(\n buf_xsize=out_columns, buf_ysize=out_rows)\n\n# Write the data to the output raster.\nout_band = out_ds.GetRasterBand(1)\nout_band.WriteArray(data)\n\n# Compute statistics and build overviews.\nout_band.FlushCache()\nout_band.ComputeStatistics(False)\nout_ds.BuildOverviews('average', [2, 4, 8, 16, 32, 64])\n\ndel out_ds\n","repo_name":"cgarrard/osgeopy-code","sub_path":"Chapter9/listing9_4.py","file_name":"listing9_4.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","stars":203,"dataset":"github-code","pt":"81"}
+{"seq_id":"35957248408","text":"class Solution:\n\tdef longestPalindrome(self, s) :\n\t\tlis=[]\n\t\tif s==s[::-1]:lis.append(s)\n\t\tif len(s)==1:\n\t\t\tlis.append(s)\n\t\telif len(s)==2:\n\t\t\tif s[0]==s[1]:\n\t\t\t\tlis.append(s)\n\t\t\telse:\n\t\t\t\tlis.append(s[0])\n\t\tfor i in range(len(s)):\n\t\t\tj=len(s)\n\t\t\twhile j-i>1:\n\t\t\t\ttemp=s[i:j]\n\t\t\t\tj-=1\n\t\t\t\tif temp==temp[::-1]:\n\t\t\t\t\tlis.append(temp)\n\t\t\t\t\tbreak\n\t\tlis.sort(key=len,reverse=True)\n\t\treturn lis\n\t\t\t\t\n\t\t\t\t\n\t\t\t\t\na=Solution()\nprint(a.longestPalindrome('abcda'))\t\n\t\t\t\t","repo_name":"warriorwizard/competiitve_coding","sub_path":"longest_palindrome_substring.py","file_name":"longest_palindrome_substring.py","file_ext":"py","file_size_in_byte":458,"program_lang":"python","lang":"zh","doc_type":"code","stars":2,"dataset":"github-code","pt":"81"}
+{"seq_id":"39985407567","text":"#!/usr/bin/env/ python\r\n# -*-coding:utf-8 -*-\r\nimport tornado.web\r\nimport json\r\nfrom Model.Merchant import MerchantService\r\nfrom UIAdmin.Forms.Merchant import MerchantForm\r\nfrom pymysql import IntegrityError\r\n\r\nclass MerchantManagerHandler(tornado.web.RequestHandler):\r\n def get(self,*args,**kwargs):\r\n self.render(\"Merchant/merchantManager.html\")\r\n\r\nclass MerchantHandler(tornado.web.RequestHandler):\r\n def get(self, *args, **kwargs):\r\n req_type = self.get_argument('type', None)\r\n if req_type == 'pagination':\r\n ret = {'status': False, 'message': '', 'total': 0, 'rows': []}\r\n try:\r\n page = int(self.get_argument('page', 1))\r\n rows = int(self.get_argument('rows', 10))\r\n start = (page - 1) * rows\r\n service = MerchantService()\r\n ret['total']=service.fetch_merchant_count()\r\n ret['rows'] = service.fetch_merchant_by_page(start, rows)\r\n ret['status'] = True\r\n except Exception as e:\r\n ret['message'] = str(e)\r\n self.write(json.dumps(ret))\r\n return\r\n self.render('Merchant/merchantManager.html')\r\n\r\nclass MerchantEdit(tornado.web.RequestHandler):\r\n def get(self,*args,**kwargs):\r\n error_summary=\"\"\r\n merchant_id=self.get_argument(\"nid\",None)\r\n if not merchant_id:\r\n crumbs=\"添加商户\"\r\n form=MerchantForm()\r\n method=\"POST\"\r\n else:\r\n crumbs=\"编辑商户\"\r\n form=MerchantForm()\r\n #根据ID获取用户信息\r\n service=MerchantService()\r\n detail=service.fetch_merchant_detail_by_nid(merchant_id)\r\n\r\n country_caption=detail.pop(\"country_caption\")\r\n country_id=detail.get(\"country_id\")\r\n form.country_id.widget.choices.append({\"value\":country_id,\"text\":country_caption})\r\n method=\"put\"\r\n form.init_value(detail)\r\n self.render(\"Merchant/merchantEdit.html\",form=form,crumbs=crumbs,method=method,summary=error_summary,nid=merchant_id)\r\n\r\n def post(self, *args, **kwargs):\r\n \"\"\"创建商户\r\n \"\"\"\r\n method=self.get_argument(\"_method\",None)\r\n if method==\"put\":\r\n return self.put(self,*args,**kwargs)\r\n error_summary=\"\"\r\n form=MerchantForm()\r\n try:\r\n is_valid=form.valid(self)\r\n print(form._value_dict)\r\n if is_valid:\r\n if form._value_dict[\"country_id\"]==\"0\":\r\n form._error_dict[\"country_id\"]=\"请选择县区\"\r\n else:\r\n del form._value_dict[\"nid\"]\r\n del form._value_dict[\"city_id\"]\r\n del form._value_dict[\"province_id\"]\r\n # 添加到数据库\r\n service=MerchantService()\r\n service.add_merchant(**form._value_dict)\r\n self.redirect(\"merchantManager.html\")\r\n return\r\n else:\r\n form.init_value(form._value_dict)\r\n except IntegrityError as e:\r\n error_summary=\"商户名称或登录用户必须唯一\"\r\n except Exception as e:\r\n error_summary=str(e)\r\n self.render(\"Merchant/merchantEdit.html\",form=form,crumbs=\"添加商户\",method=\"post\",summary=error_summary,nid=None)\r\n\r\n def put(self,*args,**kwargs):\r\n \"\"\"修改商户\"\"\"\r\n error_summary=\"\"\r\n form=MerchantForm()\r\n merchant_id=self.get_argument(\"nid\",None)\r\n\r\n try:\r\n is_valid=form.valid(self)\r\n if is_valid:\r\n if form._value_dict[\"country_id\"]==\"0\":\r\n form._error_dict[\"country_id\"]=\"请选择县(区)ID\"\r\n else:\r\n nid=form._value_dict.pop(\"nid\")\r\n del form._value_dict[\"city_id\"]\r\n del form._value_dict[\"province_id\"]\r\n # 添加到数据库\r\n service=MerchantService()\r\n db_result=service.update_merchant(nid,**form._value_dict)\r\n print(db_result)\r\n if db_result:\r\n self.redirect(\"merchantManager.html\")\r\n return\r\n else:\r\n error_summary=\"更新失败\"\r\n else:\r\n form.init_value(form._value_dict)\r\n except Exception as e:\r\n error_summary=str(e)\r\n service=MerchantService()\r\n detail=service.fetch_merchant_detail_by_nid(merchant_id)\r\n country_caption=detail.pop(\"country_caption\")\r\n country_id=detail.get(\"country_id\")\r\n form.country_id.widget.choices.append({\r\n \"value\":country_id,\"text\":country_caption\r\n })\r\n self.render(\"Merchant/merchantEdit.html\",form=form,crumbs=\"编辑商户\",method=\"put\",summary=error_summary,nid=merchant_id)\r\n\r\n def delete(self, *args, **kwargs):\r\n ret={\"message\":\"\",\"status\":False}\r\n nid=self.get_argument(\"nid\",None)\r\n print(nid)\r\n if nid:\r\n try:\r\n service=MerchantService()\r\n db_result=service.delete_merchant(nid)\r\n if db_result:\r\n ret[\"message\"]=\"删除成功\"\r\n ret['status']=True\r\n else:\r\n ret[\"message\"]=\"删除失败\"\r\n except Exception as e:\r\n ret[\"message\"]=str(e)\r\n else:\r\n ret[\"message\"]=\"请选择要删除的行\"\r\n self.write(json.dumps(ret))\r\n\r\n\r\n\r\n","repo_name":"lizhihong886/ShoppingMall","sub_path":"UIAdmin/Controllers/Merchant.py","file_name":"Merchant.py","file_ext":"py","file_size_in_byte":5654,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"21176021120","text":"# Aula 7\n# Operadores aritimeticos\n5 + 2 == 7\n5 - 2 == 3\n5 * 2 == 10\n5 / 2 == 2.5\n# exponencial 5 elevado a 2\n5 ** 2 == 25\n# inteiro da divisão\n5 // 2 == 2\n# resto da divisão\n5 % 2 == 1\n\n## precedencia de operadores\n#1 () primeiro resolva tudo dentro dos parentesis\n#2 ** resolva os exponenciais\n#3 * / // % resolva estes do primeiro ao ultimo ( ex se % aparece primeiro na linha faça-o)\n#4 + - por ultimo estes\n\nr = 2 + (3**2 + 2/2) * 2 / 10\nprint(r)\n2 + (3**2 + 2/2) * 2 / 10 == 4\nr = 3 * 5 + 4 ** 2\nprint('3 * 5 + 4 ** 2 = ',r)\nr = 3 * (5 + 4) ** 2\nprint('3 * (5 + 4) ** 2 = ',r)\n\n\n\n","repo_name":"Alexandre1961/Python","sub_path":"acm_aula_em_codigo/anotacao01.py","file_name":"anotacao01.py","file_ext":"py","file_size_in_byte":589,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14485837784","text":"\r\nimport numpy as np\r\nimport matplotlib as mpl\r\nimport matplotlib.pyplot as plt\r\nimport scipy.fftpack as spfft\r\nimport scipy.misc\r\nfrom skimage.measure import compare_ssim as ssim\r\nimport time\r\n\r\nfrom pylbfgs import owlqn\r\n\r\n\r\ndef dct2(x):\r\n return spfft.dct(\r\n spfft.dct(x.T, norm='ortho', axis=0).T, norm='ortho', axis=0)\r\n\r\n\r\ndef idct2(x):\r\n return spfft.idct(\r\n spfft.idct(x.T, norm='ortho', axis=0).T, norm='ortho', axis=0)\r\n\r\n\r\ndef progress(x, g, fx, xnorm, gnorm, step, k, ls):\r\n # Print variables to screen or file or whatever. Return zero to\r\n # continue algorithm; non-zero will halt execution.\r\n\r\n if gnorm < 5:\r\n a = 1\r\n else:\r\n a = 0\r\n return a\r\n\r\n\r\ndef evaluate(x, g, step):\r\n \"\"\"An in-memory evaluation callback.\"\"\"\r\n\r\n # we want to return two things:\r\n # (1) the norm squared of the residuals, sum((Ax-b).^2), and\r\n # (2) the gradient 2*A'(Ax-b)\r\n\r\n # expand x columns-first\r\n x2 = x.reshape((nx, ny)).T\r\n\r\n # Ax is just the inverse 2D dct of x2\r\n Ax2 = idct2(x2)\r\n im.set_data(Ax2)\r\n fig.canvas.draw()\r\n print(ssim(Xorig, Ax2))\r\n\r\n # stack columns and extract samples\r\n\r\n ############OPTION1\r\n \"\"\"\r\n Ax = Ax2.T.flat[ri].reshape(b.shape)\r\n \"\"\"\r\n ####OPTIONA2\r\n Ax = np.dot(mask_vec, Ax2.T.flatten())\r\n ######\r\n # calculate the residual Ax-b and its 2-norm squared\r\n\r\n ############OPTION1\r\n \"\"\"\r\n Axb = Ax - b\r\n \"\"\"\r\n ####OPTIONA2\r\n\r\n Axb = Ax - intensity_vec\r\n #####\r\n fx = np.sum(np.power(Axb, 2))\r\n # project residual vector (k x 1) onto blank image (ny x nx)\r\n\r\n ############OPTION1\r\n \"\"\"\r\n Axb2 = np.zeros(x2.shape)\r\n Axb2.T.flat[ri] = Axb\r\n \"\"\"\r\n ####OPTIONA2\r\n\r\n Axb2 = np.zeros(x2.shape, dtype=\"float64\")\r\n for a in range(0, len(mask_vec)):\r\n Axb2 += mask_vec[a].reshape(x2.shape).T * Axb[a]\r\n \"\"\"\r\n Axb2 = np.dot(Axb,mask_vec).reshape(x2.shape).T\r\n \"\"\"\r\n\r\n # A'(Ax-b) is just the 2D dct of Axb2\r\n AtAxb2 = 2 * dct2(Axb2)\r\n AtAxb = AtAxb2.T.reshape(x.shape) # stack columns\r\n # copy over the gradient vector\r\n\r\n np.copyto(g, AtAxb)\r\n\r\n return fx\r\n\r\n\r\n# read original image\r\nTEST_IMAGE = scipy.misc.face()\r\nTEST_IMAGE = TEST_IMAGE[:, :, 0]\r\nXorig = scipy.misc.imresize(TEST_IMAGE, [128, 128])\r\nXorig = Xorig.astype(\"float64\")\r\nny, nx = Xorig.shape\r\n\r\nmask_vec = []\r\nintensity_vec = []\r\n\r\nfor i in range(0, 1000):\r\n mask = (np.random.rand(nx, ny) < 0.5) * 1\r\n masked = mask * Xorig\r\n intensity = np.sum(masked)\r\n mask_vec.append(mask.T.flatten())\r\n intensity_vec.append(intensity)\r\n#mask_vec = np.expand_dims(mask_vec, axis=1)\r\nmask_vec = np.asarray(mask_vec)\r\nintensity_vec = np.asarray(intensity_vec, dtype=\"float64\")\r\n\r\n# perform the L1 minimization in memory\r\ntic = time.clock()\r\nplt.gray()\r\nplt.ion()\r\nfig = plt.figure()\r\nim = plt.imshow(Xorig)\r\nplt.show()\r\nXat2 = owlqn(nx * ny, evaluate, progress, 10000)\r\ntac = time.clock()\r\n\r\nprint(tac - tic)\r\n\r\n# transform the output back into the spatial domain\r\nXat = Xat2.reshape(nx, ny).T # stack columns\r\nXa = idct2(Xat)\r\nZ = Xa.astype('uint8')\r\n\r\nfig = plt.figure()\r\nfig.add_subplot(1, 2, 1)\r\nplt.gray()\r\nplt.imshow(Xorig)\r\nfig.add_subplot(1, 2, 2)\r\nplt.imshow(Z)\r\nplt.show()","repo_name":"eloymg/one","sub_path":"optim.py","file_name":"optim.py","file_ext":"py","file_size_in_byte":3257,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"12928717757","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2020/4/24 18:33\n# @File : args_parser.py\n# @IDE : PyCharm\n\n\nimport configparser\n\n\ndef read_config(config_file_name):\n config = configparser.ConfigParser(\n interpolation=configparser.ExtendedInterpolation()\n )\n config.read(config_file_name)\n return config\n\n\nif __name__ == '__main__':\n config_file = './conf.ini'\n conf = read_config(config_file)\n print(conf)\n","repo_name":"Thancoo/proxymask","sub_path":"reload/args_parser.py","file_name":"args_parser.py","file_ext":"py","file_size_in_byte":455,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"81"}
+{"seq_id":"18227296916","text":"# Скрипт конвертирует все датасеты и аудио до частоты дискретизации указанной в конфиге\n\nimport glob\nimport yaml\nimport tqdm \nimport pathlib\nfrom contextlib import suppress\n\nimport librosa\nimport torch\nimport torchaudio\n\n\ndef convert(path: pathlib.Path, sample_rate:int, pattern:str):\n \"\"\"Конвертирует все аудио до нужной частоты дискретизации\"\"\"\n output_path = path.parent / (path.stem + '_converted')\n output_path.mkdir(parents=True, exist_ok=True)\n\n fnames = glob.glob(pattern, root_dir=path)\n for fname in tqdm.tqdm(fnames):\n audio, sr = librosa.load(path / fname, sr=sample_rate)\n try:\n torchaudio.save(output_path / fname, torch.from_numpy(audio[None]), sample_rate=sample_rate)\n except RuntimeError:\n new_folder = output_path / fname\n if new_folder.suffix:\n new_folder = new_folder.parent\n new_folder.mkdir(parents=True, exist_ok=True)\n\n torchaudio.save(output_path / fname, torch.from_numpy(audio[None]), sample_rate=sample_rate)\n\nif __name__ == \"__main__\":\n try:\n with open('config.yml', 'r') as stream:\n config = yaml.safe_load(stream)\n except FileNotFoundError:\n with open('config.yaml', 'r') as stream:\n config = yaml.safe_load(stream)\n\n sample_rate = config['sample_rate']\n\n with suppress(Exception):\n path = pathlib.Path('data/UrbanSound8K/')\n convert(path, sample_rate=sample_rate, pattern=\"**/*.wav\")\n\n path = pathlib.Path('data/ESC-50/audio/')\n convert(path, sample_rate=sample_rate, pattern=\"*.wav\")\n\n path = pathlib.Path('data/RIR/')\n convert(path, sample_rate=sample_rate, pattern=\"*.wav\")","repo_name":"fabuloudy/audio_survey","sub_path":"prepoccesing.py","file_name":"prepoccesing.py","file_ext":"py","file_size_in_byte":1811,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"14134009911","text":"from typing import TypeVar\n\nfrom django.conf import settings\nfrom django.db.models import Model\nfrom django.db.models.fields.files import FieldFile, ImageFieldFile\nfrom django_s3_storage.storage import S3Storage\n\nT_MODEL = TypeVar(\"T_MODEL\", bound=Model)\n\n\ndef s3_asset_delete(*, storage_key: str | None = None) -> None:\n \"\"\"\n Delete s3 assets based on storage key.\n \"\"\"\n\n if not storage_key:\n raise ValueError(\n \"Storage key cannot be empty, to able to cleanup remote folder.\"\n )\n\n storage = S3Storage(aws_s3_bucket_name=settings.AWS_S3_BUCKET_NAME)\n parent_path_key = storage_key.rsplit(\"/\", 1)[0] # Get everything before the last /\n\n if storage.exists(parent_path_key):\n dirs, files = storage.listdir(parent_path_key)\n\n if len(dirs) == 0 and len(files) == 0:\n storage.delete(parent_path_key)\n\n\ndef s3_asset_cleanup(*, instance: T_MODEL, field: str) -> None:\n \"\"\"\n Delete specific s3 assets belonging to an instance.\n \"\"\"\n\n instance_field = getattr(instance, field, None)\n\n if instance_field and isinstance(instance_field, (ImageFieldFile, FieldFile)):\n storage_key = instance_field.name\n instance_field.delete(save=False)\n\n s3_asset_delete(storage_key=storage_key)\n","repo_name":"danielkjellid/nest","sub_path":"nest/core/utils/s3.py","file_name":"s3.py","file_ext":"py","file_size_in_byte":1276,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"81"}
+{"seq_id":"71843057852","text":"import requests\r\n\r\n\r\ncity = input(\"Enter your city: \")\r\nurl = f\"https://api.openweathermap.org/data/2.5/weather?q={city}&appid=a118e4b915fbfb737736634d31139853\"\r\nrest = requests.get(url)\r\n\r\ndata = rest.json() \r\n\r\nhumidity = data['main']['humidity']\r\npressure = data['main']['pressure']\r\nwind = data['wind']['speed']\r\ndescription = data['weather'][0]['description']\r\ntemp= data['main']['temp']\r\n\r\nprint(f\"\"\" \r\n Humidity : {humidity}\r\n Pressure : {pressure}\r\n Wind : {wind}\r\n Description: {description}\r\n Temperature :{round(temp - 273.15, 1)} C \"\"\")\r\n","repo_name":"Itgeeksk/Python_miniprojects","sub_path":"cliweatherapp.py","file_name":"cliweatherapp.py","file_ext":"py","file_size_in_byte":568,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"39605415361","text":"# -*- coding: utf-8 -*-\n\"\"\"\n src.common.decorators\n ~~~~~~~~~~~~~~~~~~~~~\n\n\"\"\"\nfrom flask import current_app as app, json, _request_ctx_stack\nfrom functools import wraps\nfrom werkzeug.exceptions import Unauthorized, Forbidden\nfrom src.common.utils import _get_token_auth_header\nfrom six.moves.urllib.request import urlopen\nfrom jose import jwt\nfrom src.common.scope import Scope\nfrom sentry_sdk import set_user, add_breadcrumb\n\n\ndef authenticate(f):\n \"\"\"\n Authenticates the user using bearer token.\n \"\"\"\n\n doc = getattr(f, \"__doc__\")\n if doc:\n setattr(f, \"__doc__\", doc + \"security:\\n\\t- BearerAuth: []\")\n\n @wraps(f)\n def decorator(*args, **kwargs):\n\n if app.config.get(\"TESTING\"):\n \"\"\"If we're testing, just return a user w/ all perms\"\"\"\n _request_ctx_stack.top.current_user = {\n \"roles\": list(Scope.members().keys())\n }\n return f(*args, **kwargs)\n\n try:\n token = _get_token_auth_header()\n jsonurl = urlopen(\"https://login.microsoftonline.com/\" +\n app.config.get(\"AZURE_TENANT_ID\") +\n \"/discovery/v2.0/keys\")\n jwks = json.loads(jsonurl.read())\n unverified_header = jwt.get_unverified_header(token)\n rsa_key = {}\n for key in jwks[\"keys\"]:\n if key[\"kid\"] == unverified_header[\"kid\"]:\n rsa_key = {\n \"kty\": key[\"kty\"],\n \"kid\": key[\"kid\"],\n \"use\": key[\"use\"],\n \"n\": key[\"n\"],\n \"e\": key[\"e\"]\n }\n except Exception:\n raise Unauthorized(\"Unable to parse authentication\")\n\n if rsa_key:\n try:\n payload = jwt.decode(\n token,\n rsa_key,\n algorithms=[\"RS256\"],\n audience=app.config.get(\"AZURE_API_AUDIENCE\"),\n issuer=(\"https://sts.windows.net/\" +\n app.config.get(\"AZURE_TENANT_ID\") +\n \"/\")\n )\n except jwt.ExpiredSignatureError:\n raise Unauthorized(\"Token is expired\")\n\n except jwt.JWTClaimsError:\n raise Unauthorized(\"Incorrect claims please check the audience\"\n \" and issuer\")\n\n except Exception:\n raise Unauthorized(\"Unable to parse authentication token\")\n\n _request_ctx_stack.top.current_user = payload\n\n \"\"\"Set the authenticated user in Sentry\"\"\"\n set_user({\n \"id\": payload.get(\"unique_id\"),\n \"username\": payload.get(\"preferred_username\"),\n \"roles\": payload.get(\"roles\", [])\n })\n\n \"\"\"Add breadcrumb\"\"\"\n add_breadcrumb(\n category=\"auth\",\n message=(\"Authenticated user \" +\n payload.get(\"preferred_username\", \"\")),\n level=\"info\"\n )\n\n return f(*args, **kwargs)\n raise Unauthorized(\"Unable to find appropriate key\")\n\n return decorator\n\n\ndef requires_scope(scope: Scope):\n\n def decorator(f):\n\n \"\"\"Shoves the scopes specified by the decorator in the docstring\"\"\"\n doc = getattr(f, \"__doc__\")\n if doc:\n setattr(f, \"__doc__\", (doc + \"description: |\\n\" +\n \"\\tRequired Permissions:\\n\" +\n \"\".join((f\"\\t- {i}\"\"\\n\")\n for i in scope.names) +\n \" \"))\n\n @wraps(f)\n def decorated_function(*args, **kwargs):\n cu = _request_ctx_stack.top.current_user\n\n s = Scope(cu.get(\"roles\"))\n\n \"\"\" Check if the user has the required scope(s) \"\"\"\n if not(scope & s):\n raise Forbidden()\n\n return f(*args, **kwargs)\n return decorated_function\n return decorator\n","repo_name":"KnightHacks/hackathon-2021-backend","sub_path":"src/common/decorators.py","file_name":"decorators.py","file_ext":"py","file_size_in_byte":4129,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"78"}
+{"seq_id":"37417094978","text":"#programm a game\nfrom random import randint\n\n\nprint(\"Welcome to my game. In this game two players gonna battle each other.\")\nname = input(\"Enter the name of player 1:\")\nprint(name)\nname2 = input(\"Enter the name of Player 2:\")\nprint(name2)\n\nclass Player: \n def __init__(self, maxAttack, maxDefense, health, name):\n self.maxAttack = maxAttack\n self.maxDefense = maxDefense\n self.health = health\n self.name = name\n\n def get_attack(self): \n return randint(0, self.maxAttack)\n\n def get_defense(self):\n return randint(0, self.maxDefense)\n\n\nplayer1 = Player(30,20,200,name)\nplayer2 = Player(30,20,200,name2)\n\n# attackPlayer1 = randint(1, player1.attack)\n\n# print(player1.attack)\n# print(player1.attack-player1.defense)\n# print(player1.attack)\n\nwhile player1.health > 0 and player2.health > 0:\n option = input(\"Choose what you want to do: \\\n 1.... attack \\\n 2.... defense\")\n\n bonus = 10\n # print(f\"this is player.get_attack:{player1.get_attack()}\")\n # print(f\"this is player.get_defense:{player2.get_defense()}\")\n calculatedDamage = player1.get_attack() - player2.get_defense()\n calculatedDefense = player2.get_attack() - player1.get_defense()\n if option == 1:\n calculatedDamage +=bonus\n elif option == 2:\n calculatedDefense +=bonus\n if calculatedDamage <= 0:\n print(\"Die Attacke von dir wurde blockiert\")\n else: \n player2.health -=calculatedDamage\n print(f\"Du hast / Sie haben so viel Schaden gemacht: {calculatedDamage}. Der Gegner hat noch so viel Leben:{player2.health}\")\nif player1.health > 0:\n print(f\"Gratuliere {player1.name} zu deinem Sieg\")\nif player2.health > 0: \n print(f\"Gratuliere {player2.name}zu deinem Sieg\")\n\n\n\n\n\n# if player1 set attack :\n# player2.health - player1.attack\n\n\n","repo_name":"Andrej8840/Schnupperlehre","sub_path":"Aufgabe2.py","file_name":"Aufgabe2.py","file_ext":"py","file_size_in_byte":1821,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"6472833376","text":"from pymongo import MongoClient\n\nDB_HOST = \"mongodb://127.0.0.1:27017/\"\nDB_NAME = \"yelp\"\nDB_COLLECTION = \"review\"\n\n\nclient = MongoClient(DB_HOST)\ndb = client[DB_NAME]\nreview_Collection = db[DB_COLLECTION]\n\nreview_Cursor = review_Collection.find(no_cursor_timeout=False)\n\n\nbusiness_Collection = db[\"business\"]\n\nprint(\"Code started successfully\")\n\nskip = 0\ntotal_Count = review_Collection.find({}).count()\nlimit = 10\ncount = 0\n\n\nwhile count < total_Count:\n # print(f\"count of while = {count}\")\n review_Cursor = review_Collection.aggregate([\n {\n \"$skip\":skip\n },\n {\n \"$limit\":limit\n }\n ])\n\n for doc in review_Cursor:\n count = count + 1\n print(f\"count of for = {count}\")\n tempCursor = business_Collection.find({\n \"business_id\":doc[\"business_id\"]\n })\n\n if tempCursor.count() == 0:\n review_Collection.delete_one(doc)\n\n\n skip = skip + limit\nprint(\"Code successfully exited\")\n","repo_name":"NikhilAshodariya/YelpRestaurant","sub_path":"Miscellaneous/Old .py files/remove_From_Review.py","file_name":"remove_From_Review.py","file_ext":"py","file_size_in_byte":1298,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"31192346212","text":"# Prints weather information for the next 5 days split in 3-hour segments\n# For API information: https://openweathermap.org/forecast5\nimport requests\nimport json\n\nif __name__ == '__main__':\n baseUrl = 'http://api.openweathermap.org/data/2.5/forecast'\n # In 'appid', insert your own key from OpenWeather\n parameters = {'q': 'Dublin,IE', 'appid': 'YOUR-KEY-HERE', 'units': 'metric'}\n response = requests.get(baseUrl, params=parameters)\n content = response.content\n # Parse json info\n info = json.loads(content)\n\n # To iterate through every timestamp\n list = info['list']\n for i in range(len(list)):\n listInfo = list[i]\n date = listInfo['dt_txt']\n mainData = listInfo['main']\n tempMax = mainData['temp_max']\n tempMin = mainData['temp_min']\n feelsLike = mainData['feels_like']\n weather = listInfo['weather']\n weatherDesc = weather[0]\n description = weatherDesc['description']\n\n print('Weather information for %s\\nMax Temp: %sC, Min Temp: %sC, Feels like: %sC\\nWeather description: %s\\n' %\n (date, tempMax, tempMin, feelsLike, description))\n\n # print(listInfo)\n","repo_name":"Igor-Ono/checkWeather","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1172,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"37510072208","text":"import numpy as np\n\nfrom collections.VanCollections import TracksData, Frame\nfrom matplotlib import pyplot as plt\nfrom functional import VanFunctional as F\n\nLOAD_PRE_COMPUTED_TRACKING = True\n\n\ndef process_tracks_stats(tracks_data: TracksData):\n \"\"\"\n This function computes tracks statistics.\n :param tracks_data: tracks data structure.\n :return: None.\n \"\"\"\n tracks_df = tracks_data.cords_data\n\n # Total number of tracks\n track_count = len(tracks_df['TrackId'].unique())\n print(f'Total number of tracks : {track_count}')\n\n # Total number of frames\n frame_count = len(tracks_df['FrameId'].unique())\n print(f'Total number of frames: {frame_count}')\n\n # track length statistics\n tracks_len_df = tracks_df[['TrackId', 'FrameId']].groupby('TrackId').count()\n tracks_len_df.sort_values(by='FrameId', inplace=True, ascending=False)\n print(f'Track minimum length: {tracks_len_df.min().item()}')\n print(f'Track maximum length: {tracks_len_df.max().item()}')\n print(f'Track mean length: {tracks_len_df.mean().item()}')\n\n # average number of tracks per image\n tracks_count_df = tracks_df[['TrackId', 'FrameId']].groupby('FrameId').count()\n print(f'Mean number of frame links: {tracks_count_df.mean().item()}')\n\n\ndef cut_image_region(left_x, y, region_size=100, img_width=1226, img_height=370):\n \"\"\"\n This function cuts a region around a track.\n :param left_x: start x.\n :param y: start y.\n :param region_size: region size.\n :param img_width: original image width\n :param img_height: original image height.\n :return: region boundaries.\n \"\"\"\n offset = np.floor(region_size / 2)\n scatter_x_offset = 0\n scatter_y_offset = 0\n\n # vertical bounds\n upper_bound = max(y - offset, 0)\n lower_bound = min(y + (region_size - offset), img_height)\n padding_size = region_size - np.abs(lower_bound - upper_bound)\n\n if upper_bound == 0:\n lower_bound += padding_size\n scatter_y_offset -= padding_size\n elif lower_bound == img_height:\n upper_bound -= padding_size\n scatter_y_offset += padding_size\n\n # horizontal bounds - left image\n left_bound = max(left_x - offset, 0)\n right_bound = min(left_x + (region_size - offset), img_width)\n padding_size = region_size - np.abs(right_bound - left_bound)\n\n if left_bound == 0:\n right_bound += padding_size\n scatter_x_offset -= padding_size\n elif right_bound == img_width:\n left_bound -= padding_size\n scatter_x_offset += padding_size\n\n return int(np.round(left_bound)), int(np.round(right_bound)), \\\n int(np.round(upper_bound)), int(np.round(lower_bound)), \\\n int(np.round(scatter_x_offset)), int(np.round(scatter_y_offset))\n\n\ndef visualize_track(tracks_data: TracksData, track_id: int, max_seq_len=10):\n \"\"\"\n This function visualizes a track.\n :param tracks_data: tracks data structure.\n :param track_id: track id.\n :param max_seq_len: maximum frames to visualize.\n :return: None.\n \"\"\"\n track_obj = tracks_data.get_track_object(track_id)\n start_frame = track_obj.get_start_frame()\n track_len = min(track_obj.get_len(), max_seq_len)\n\n for frame_id in range(start_frame, start_frame + track_len):\n cur_frame = Frame(frame_id)\n cur_track_cords = tracks_data.get_track_frame_cords(frame_id, track_id)\n\n # cutting region from left image\n x_min, x_max, y_min, y_max, left_x_offset, left_y_offset = cut_image_region(\n cur_track_cords[0], cur_track_cords[2])\n\n left_region = cur_frame.get_left_image()[y_min: y_max, x_min: x_max]\n\n # cutting region from right image\n x_min, x_max, y_min, y_max, right_x_offset, right_y_offset = cut_image_region(\n cur_track_cords[1], cur_track_cords[2])\n\n right_region = cur_frame.get_right_image()[y_min: y_max, x_min: x_max]\n\n fig, ax = plt.subplots(1, 2)\n ax[0].imshow(left_region, cmap='gray')\n ax[1].imshow(right_region, cmap='gray')\n\n ax[0].scatter(50 + left_x_offset, 50 + left_y_offset)\n ax[1].scatter(50 + right_x_offset, 50 + right_y_offset)\n\n ax[0].set_title('left region')\n ax[1].set_title('right region')\n fig.suptitle(f'Frame #{frame_id}')\n\n plt.savefig(f'{track_id}_f{frame_id}.png')\n plt.show()\n\n\ndef get_connectivity_data(tracks_data: TracksData):\n \"\"\"\n This function computes the connectivity data.\n :param tracks_data: tracks data structure.\n :return: connectivity data.\n \"\"\"\n out_going_list = []\n num_frames = tracks_data.get_num_frames()\n\n cur_frame_tracks = set(tracks_data.get_tracks_by_frame(0))\n next_frame_tracks = set(tracks_data.get_tracks_by_frame(1))\n\n for next_frame_id in range(1, num_frames - 2):\n cur_out_going = cur_frame_tracks.intersection(next_frame_tracks)\n out_going_list.append(len(cur_out_going))\n\n cur_frame_tracks = next_frame_tracks\n next_frame_tracks = set(tracks_data.get_tracks_by_frame(next_frame_id + 1))\n\n return out_going_list\n\n\ndef plot_connectivity(connectivity_data: list):\n \"\"\"\n This function plots connectivity data.\n :param connectivity_data: connectivity data as list.\n :return: None.\n \"\"\"\n x_data = np.arange(len(connectivity_data))\n\n fig, ax = plt.subplots(figsize=(12, 8))\n ax.plot(x_data, connectivity_data)\n\n plt.xlabel('frame')\n plt.ylabel('outgoing tracks')\n plt.title('Connectivity')\n\n plt.show()\n\n\ndef plot_track_length_histogram(tracks_data: TracksData):\n \"\"\"\n This function plots the track length histogram.\n :param tracks_data: tracks data.\n :return: None.\n \"\"\"\n tracks_df = tracks_data.cords_data[['TrackId', 'FrameId']]\n tracks_df = tracks_df.groupby('TrackId').count()\n tracks_df.reset_index(None, drop=False, inplace=True)\n tracks_df = tracks_df.groupby('FrameId').count()\n\n hist_x, hist_y = [0], [0]\n hist_x.extend(list(tracks_df.index))\n hist_y.extend(list(tracks_df['TrackId']))\n\n fig, ax = plt.subplots(figsize=(12, 8))\n ax.plot(hist_x, hist_y)\n\n plt.xlabel('Track length')\n plt.ylabel('Track #')\n plt.title('Track length histogram')\n\n plt.show()\n\n\ndef track_re_projection(tracks_data: TracksData, track_id=-1, positions_lst=None):\n \"\"\"\n This function computes the re-projection error for a track.\n :param tracks_data: tracks data.\n :return: errors list, start frame.\n \"\"\"\n if track_id < 0:\n track_id = tracks_data.get_track_of_len(10)\n\n track_obj = tracks_data.get_track_object(track_id)\n last_frame = track_obj.get_start_frame() + track_obj.get_len() - 1\n\n # getting the track's coordinates\n last_track_cords = tracks_data.get_track_frame_cords(last_frame, track_id)\n\n # loading the ground truth data\n if positions_lst is None:\n gt_list = F.get_camera_ground_truth()\n else:\n gt_list = positions_lst\n\n k, m1, m2 = F.get_calib()\n\n # computing the last camera matrices\n last_left_ext = gt_list[last_frame]\n last_right_rotation = last_left_ext[:, :-1]\n last_right_translation = (last_left_ext[:, -1] + m2[:, -1]).reshape(3, 1)\n last_right_ext = np.hstack([last_right_rotation, last_right_translation])\n\n # computing triangulation for the last frame\n track_3d_cords = F.compute_triangulation(k, last_left_ext, last_right_ext,\n (last_track_cords[0], last_track_cords[2]),\n (last_track_cords[1], last_track_cords[2]))\n track_3d_cords = track_3d_cords.reshape(4, 1)\n track_3d_cords = track_3d_cords / track_3d_cords[-1, :]\n dist_list = []\n\n for frame_id in range(track_obj.get_start_frame(), track_obj.get_start_frame() + track_obj.get_len()):\n cur_track_cords = tracks_data.get_track_frame_cords(frame_id, track_id)\n cur_left_img_cords = [cur_track_cords[0], cur_track_cords[2]]\n cur_right_img_cords = [cur_track_cords[1], cur_track_cords[2]]\n\n cur_left_img_cords = np.array(cur_left_img_cords).reshape(2, 1)\n cur_right_img_cords = np.array(cur_right_img_cords).reshape(2, 1)\n\n cur_left_ext = gt_list[frame_id]\n cur_right_r = cur_left_ext[:, :-1]\n cur_right_t = (cur_left_ext[:, -1] + m2[:, -1]).reshape(3, 1)\n cur_right_ext = np.hstack([cur_right_r, cur_right_t])\n\n # projecting the points to pixels\n cur_left_proj = k @ cur_left_ext @ track_3d_cords\n cur_left_proj = cur_left_proj[:-1, :] / cur_left_proj[-1, :]\n\n cur_right_proj = k @ cur_right_ext @ track_3d_cords\n cur_right_proj = cur_right_proj[:-1, :] / cur_right_proj[-1, :]\n\n # computing the average error for current frame\n cur_err = np.linalg.norm(cur_left_img_cords - cur_left_proj) + \\\n np.linalg.norm(cur_right_img_cords - cur_right_proj)\n dist_list.append(cur_err / 2)\n\n return dist_list, track_obj.get_start_frame()\n\n\ndef plot_re_projection_error(dist_list: list, start_frame: int):\n \"\"\"\n This function plots the re-projection error.\n :param dist_list: errors list.\n :param start_frame: start frame.\n :return: None.\n \"\"\"\n x_vals = np.arange(start_frame, start_frame + len(dist_list), 1).astype(int)\n\n fig, ax = plt.subplots(figsize=(12, 8))\n ax.plot(x_vals, dist_list)\n\n plt.xlabel('frame')\n plt.ylabel('average re-projection error')\n plt.title('Re-projection error per frame')\n\n plt.show()\n\n\ndef plot_inliers_data(tracks_data: TracksData):\n \"\"\"\n This function plots the inliers data.\n :param tracks_data: tracks data.\n :return: None.\n \"\"\"\n cur_data = tracks_data.get_supporters_data()\n x_data = np.array(list(cur_data.keys()))\n y_data = np.array(list(cur_data.values())) * 100\n\n fig, ax = plt.subplots(figsize=(12, 8))\n ax.plot(x_data, y_data)\n\n plt.xlabel('frame')\n plt.ylabel('inliers percentage')\n plt.title('Q5 - inliers percentage per frame')\n\n plt.show()\n\n\nif __name__ == '__main__':\n if not LOAD_PRE_COMPUTED_TRACKING:\n F.track_frames(2560, 'try1.pkl', True)\n\n x = TracksData.from_pickle('try1.pkl')\n\n # ----------------------------------------------------- Question 2\n process_tracks_stats(x)\n\n # ----------------------------------------------------- Question 3\n cur_id = x.get_track_of_len(10)\n visualize_track(x, cur_id, max_seq_len=10)\n\n # ----------------------------------------------------- Question 4\n c_data = get_connectivity_data(x)\n plot_connectivity(c_data)\n\n # ----------------------------------------------------- Question 5\n plot_inliers_data(x)\n\n # ----------------------------------------------------- Question 6\n plot_track_length_histogram(x)\n\n # ----------------------------------------------------- Question 7\n re_projection_errs, start = track_re_projection(x)\n plot_re_projection_error(re_projection_errs, start)\n","repo_name":"dortal721/VANProject","sub_path":"ex_code/slam_ex4.py","file_name":"slam_ex4.py","file_ext":"py","file_size_in_byte":10924,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"2876284479","text":"import threading\nimport simpleubjson\nimport json\nimport os\nimport time\nimport bson\nimport msgpack\n\nclass CompressorThread(threading.Thread):\n INPUT_FOLDER_SIZE_COUNTER = 0\n OUTPUT_FOLDER_SIZE_COUNTER = 0\n TOTAL_COMPRESSION_DURATION = 0\n\n def __init__(self, items, output_folder, compression_type, extension):\n threading.Thread.__init__(self)\n self.extension = extension\n self.items = items\n self.output_folder = output_folder\n self.compression_type = compression_type\n\n def get_json_as_object(self, json_file_location):\n json_file = open(json_file_location)\n json_raw = json_file.read()\n json_file.close()\n return json.loads(json_raw)\n\n def compress_file_ub_json(self, json_file_location):\n parsed_json = self.get_json_as_object(json_file_location)\n compr_started_time = time.time()\n ubjson_bytes = simpleubjson.encode(parsed_json)\n duration_compression = time.time() - compr_started_time\n CompressorThread.TOTAL_COMPRESSION_DURATION += duration_compression\n return ubjson_bytes\n\n def compress_file_bson(self, json_file_location):\n parsed_json = self.get_json_as_object(json_file_location)\n compr_started_time = time.time()\n bson_bytes = bson.encode_array(parsed_json, [])\n duration_compression = time.time() - compr_started_time\n CompressorThread.TOTAL_COMPRESSION_DURATION += duration_compression\n return bson_bytes\n\n def compress_file_msgpack(self, json_file_location):\n parsed_json = self.get_json_as_object(json_file_location)\n compr_started = time.time()\n msgpack_bytes = msgpack.packb(parsed_json, use_bin_type=True)\n duration_compression = time.time() - compr_started\n CompressorThread.TOTAL_COMPRESSION_DURATION += duration_compression\n return msgpack_bytes\n\n def write_bytes_to_location(self, output_file_location, bytes_output):\n bin_json = open(output_file_location, \"wb\")\n bin_json.write(bytes_output)\n bin_json.close()\n\n def update_stats(self, initial_file_location, compressed_file_location, file_name):\n initial_size = os.path.getsize(initial_file_location)\n compressed_size = os.path.getsize(compressed_file_location)\n CompressorThread.INPUT_FOLDER_SIZE_COUNTER += initial_size\n CompressorThread.OUTPUT_FOLDER_SIZE_COUNTER += compressed_size\n\n def compress_and_store_file(self, input_location, output_location, file_name):\n bytes_compressed = self.call_compression_function(input_location)\n self.write_bytes_to_location(output_location, bytes_compressed)\n self.update_stats(input_location, output_location, file_name)\n\n def call_compression_function(self, input_location):\n if self.compression_type == 'ubjson':\n return self.compress_file_ub_json(input_location)\n elif self.compression_type == 'bson':\n return self.compress_file_bson(input_location)\n elif self.compression_type == 'msgpack':\n return self.compress_file_msgpack(input_location)\n\n def run(self):\n for item in self.items:\n basename = os.path.basename(item)\n output_file = self.output_folder + basename.split(self.extension)[0]\n self.compress_and_store_file(item, output_file, basename)","repo_name":"LSDE-Flickr-ML-Classification/json-compression-test","sub_path":"compressor_thread.py","file_name":"compressor_thread.py","file_ext":"py","file_size_in_byte":3357,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"844753442","text":"# Databricks notebook source\n# MAGIC %sql\n# MAGIC CREATE DATABASE IF NOT EXISTS groupe8;\n\n# COMMAND ----------\n\ndf = spark.read.csv(\"/mnt/groupe8/Characters.csv\",sep=';',header=True)\nlistOfHouse = ['Gryffindor', 'Slytherin', 'Ravenclaw', 'Hufflepuff']\nformatDate = \"yyyy-mm\"\n\n# COMMAND ----------\n\nfrom pyspark.sql.functions import *\nfrom datetime import datetime\n\ndata = df.filter(col(\"House\").isin(listOfHouse))\ndata = data.select(\"Gender\", \"Wand\", \"Patronus\", \"Species\", \"Blood status\", \"Loyalty\", \"Skills\", \"Birth\", \"House\", \"Eye colour\").withColumnRenamed(\"Blood status\", \"Bloodstatus\").withColumnRenamed(\"Eye colour\", \"Eyecolour\")\ndata = data.withColumn(\"Patronus\", when(data.Patronus == \"Unknown\", \"null\") .when(data.Patronus == \"None\", \"null\").otherwise(data.Patronus))\n\n# COMMAND ----------\n\ndata = data.withColumn(\"Birth\", regexp_replace('Birth', 'Late', '28'))\ndata = data.withColumn(\"Birth\", regexp_replace('Birth', 'or earlier', ''))\ndata = data.withColumn(\"Birth\", regexp_replace('Birth', 'prior to', ''))\ndata = data.withColumn(\"Birth\", expr(\"CASE WHEN Birth LIKE '%–%' THEN 'null' \" + \n \"WHEN Birth LIKE '%Pre%' THEN 'null' \"+\n \"WHEN Birth LIKE '%pre%' THEN 'null' \"+\n \"WHEN Birth LIKE '%c.%' THEN 'null' \" +\n \"WHEN Birth LIKE '%Post%' THEN 'null' \"+\n \"WHEN Birth LIKE '%century%' THEN 'null' \"+\n \"WHEN Birth LIKE '%In or before%' THEN 'null' \"+\n \"ELSE Birth END\"))\n\n# COMMAND ----------\n\ndata.write.format(\"delta\").mode(\"overwrite\").option(\"userMetadata\", \"init\").saveAsTable(\"groupe8.pottertable3\")\n\n# COMMAND ----------\n\ndisplay(data)\n\n# COMMAND ----------\n\ndisplay(df)\n\n# COMMAND ----------\n\n\n","repo_name":"nitharsan21/Azure_Data","sub_path":"EtudeData.py","file_name":"EtudeData.py","file_ext":"py","file_size_in_byte":1713,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"4478940750","text":"#is a blue print in oop\n#an object/instance\n#syntax\n#class name_of_the_class():\n #the blue print attributes\n\n\nclass Person():\n name = 'Developer'\n\nd1 = Person()\nd2 = Person()\n\nprint(d1.name)\nprint(d2.name)\n\n\n#all classes has a function called _init_()\n# which is always executed when the class is being initiated\n#Use the _init_() function to assign values to properties or other operations that are neccessary\n\n\n\n\nclass Animal():\n country = 'kenya'#class property\n def __init__(self,name):\n self.thename = name\n\n def sound(self):\n if self.thename == 'cutty':\n print('mweeew')\n\n def kitty(self):\n if self.thename == 'cutty':\n print('such a playerful cat')\n def fear(self):\n if self.thename == 'Betty':\n print('betty is not feeling well')\n def white(self):\n if self.thename == 'cutty':\n print('cutty is a deaf cat')\n\n\n def sounds(self):\n if self.thename == 'bob':\n print('woooof')\n\n\n\ncat = Animal('cutty')\nprint(cat.thename)\ncat2= Animal('Betty')\nprint(cat.thename)\nprint(cat.country)\nprint('bob is a nice puppy')\ndog = Animal('bob')\n\n#deleting a property\n#del object.property\n#el cat.thename\nprint(cat.thename)\nprint(cat2.thename)\n#deleting an object\n#del object\ndel cat\nprint(cat)\n\ncat.sound()\ncat.kitty()\ncat2.fear()\ncat.white()\ndog.sounds()\n\n#modifying objectproperty\n#object.property='new value'\ncat.country = 'Uganda'\nprint(cat.country)\n\n\n\n","repo_name":"muigaipeter/Python_class","sub_path":"workingfolder/classes/demo1.py","file_name":"demo1.py","file_ext":"py","file_size_in_byte":1505,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"25845301531","text":"import scapy.all as scapy\nimport optparse\n\n\ndef get_user_input():\n parse = optparse.OptionParser()\n parse.add_option(\"-i\", \"--ipaddress\", dest=\"ip_address\", help=\"Enter IP Address\")\n\n (user_input, arguments) = parse.parse_args()\n\n if not user_input.ip_address:\n print(\"Enter IP Address\")\n\n return user_input\n\n\ndef scan_my_network(ip):\n # Create ARP Request\n arp_request = scapy.ARP(pdst=ip)\n # scapy.ls(scapy.ARP())\n\n # Broadcast\n broadcast = scapy.Ether(dst=\"ff:ff:ff:ff:ff:ff\")\n combined = broadcast/arp_request # iki paketi tek paket yapti\n\n # Response\n (answered_list, unanswered_list) = scapy.srp(combined, timeout=1) # paket gonderimi(cevap verilenler/verilmeyenler return edilir.)\n answered_list.summary()\n\n\nuser_ip_address = get_user_input()\nscan_my_network(user_ip_address.ip_address)\n","repo_name":"DilaraGule/Ethical_Hacker_Course_net_scanner","sub_path":"net_scanner.py","file_name":"net_scanner.py","file_ext":"py","file_size_in_byte":850,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"25868968153","text":"\"\"\" Helper classes \"\"\"\n\nimport base64\nimport os\nfrom collections import Callable\nfrom email.mime.text import MIMEText\nfrom urllib import parse\n\nimport facebook\nimport httplib2\nimport requests as r\nimport soundcloud\nimport twitter\nfrom googleapiclient import discovery, errors\nfrom oauth2client.file import Storage\n\nfrom bot_chucky.errors import BotChuckyError\nfrom bot_chucky.utils import split_text\n\n\nclass FacebookData:\n def __init__(self, token):\n \"\"\"\n :param token: Facebook Page token\n :param _api: Instance of the GraphAPI object\n \"\"\"\n self.token = token\n self._api = facebook.GraphAPI(self.token)\n\n def get_user_name(self, _id):\n \"\"\"\n :param _id: find user object by _id\n :return: first name of user, type -> str\n \"\"\"\n if not isinstance(_id, str):\n raise ValueError('id must be a str')\n user = self._api.get_object(_id)\n return user['first_name'] if user else None\n\n\nclass WeatherData:\n \"\"\"\n Class which collect weather data\n \"\"\"\n def __init__(self, api_token):\n \"\"\"\n :param api_token: Open Weather TOKEN\n \"\"\"\n self.token = api_token\n\n def get_current_weather(self, city_name):\n \"\"\"\n :param city_name: Open weather API, find by city name\n :return dictionary object with information\n\n for example:\n\n {'weather': [{'id': 800, 'main': 'Clear', 'description': 'clear sky'}]}\n \"\"\"\n api_url = 'http://api.openweathermap.org' \\\n '/data/2.5/weather?q={0}&APPID={1}'.format(city_name, self.token)\n\n info = r.get(api_url).json()\n return info\n\n\nclass TwitterData:\n \"\"\"\n Class which collect Twitter data\n \"\"\"\n def __init__(self, tokens):\n \"\"\"\n :param tokens: Dictionary of all tokens\n [consumer_key, consumer_secret, access_token_key,\n access_token_secret]\n required to initialize the Twitter Api\n \"\"\"\n self.api = twitter.Api(\n consumer_key=tokens['consumer_key'],\n consumer_secret=tokens['consumer_secret'],\n access_token_key=tokens['access_token_key'],\n access_token_secret=tokens['access_token_secret']\n )\n\n def send_tweet(self, status):\n if status:\n try:\n return {\n 'success': True,\n 'tweet': self.api.PostUpdate(status)\n }\n except twitter.error.TwitterError as TWE:\n return {\n 'detail': TWE.message[0]['message'],\n 'success': False\n }\n\n\nclass StackExchangeData:\n \"\"\"\n Class which collect StackExchange data\n \"\"\"\n _default_parameters = {\n 'order': 'desc',\n 'sort': 'activity',\n 'site': 'stackoverflow',\n }\n\n def get_stack_answer_by(self, **kwargs):\n \"\"\"\n :param kwargs: create a query by arguments\n for example:\n tag='Python', will be search by tag\n title='Update Python', will be search by title\n and etc.\n :return: an array with links\n \"\"\"\n if len(kwargs) > 1:\n raise BotChuckyError('The argument must be one')\n\n for key in kwargs.keys():\n query = kwargs.get(key)\n self._default_parameters.update({key: query})\n\n if not isinstance(query, str):\n raise TypeError(f'{query} must be a string')\n\n encode_query = parse.urlencode(self._default_parameters)\n\n stack_url = f'https://api.stackexchange.com/2.2/search/advanced?' \\\n f'{encode_query}'\n\n questions = r.get(stack_url).json()\n links = [obj['link'] for obj in questions['items']]\n return links\n\n\nclass SoundCloudData:\n \"\"\"\n Class to gather soundcloud data, tracks etc\n \"\"\"\n def __init__(self, client_id):\n \"\"\"\n client_id = Client ID, must be registered\n \"\"\"\n self.client_id = client_id\n self._api = soundcloud.Client(client_id=self.client_id)\n\n def resolve_track(self, url):\n \"\"\"\n Resolve a track name\n :param url: permalink to a track (str)\n \"\"\"\n try:\n track = self._api.get('/resolve', str(url))\n\n return {\n 'success': True,\n 'track': track.id\n }\n except Exception as e:\n return {\n 'success': False,\n 'detail': f'Error: {e.message}, Code: {e.response.status_code}'\n }\n\n def search(self, artist=None):\n \"\"\"\n Search for tracks by artist, or artist by track\n :param artist: search by artist, returns tracks and info, type -> str\n \"\"\"\n self.artist = artist\n\n if self.artist is not None:\n try:\n artists = self._api.get('/users', q=self.artist)\n tracks = self._api.get('/tracks', q=self.artist)\n return {\n 'success': True,\n 'artists': artists,\n 'tracks': tracks\n }\n except Exception as e:\n return {\n 'success': False,\n 'detail': f'Error: {e.message}, Code: '\n f'{e.response.status_code}'\n }\n\n\nclass GmailData:\n \"\"\"\n Class which collect Gmail Data\n \"\"\"\n def __init__(self):\n self.api = self._create_gmail_api()\n\n def send_mail(self, to, subject, body):\n \"\"\"\n :param to: Email address of the receiver\n :param subject: Subject of the email\n :param body: Body of the email\n \"\"\"\n message = self._create_message(to, subject, body)\n try:\n message = self.api.users().messages().send(\n userId='me',\n body=message\n ).execute()\n return {\n \"success\": True,\n \"message\": message\n }\n except errors.HttpError as error:\n return {\n \"success\": False,\n \"detail\": str(error)\n }\n\n def _create_gmail_api(self):\n try:\n credentials = self._get_credentials()\n http = credentials.authorize(httplib2.Http())\n service = discovery.build('gmail', 'v1', http=http)\n return service\n except AttributeError:\n return ''\n\n def _get_credentials(self):\n \"\"\"Gets valid user credentials from storage.\n :return: Credentials, the obtained credential.\n \"\"\"\n home_dir = os.path.expanduser('~')\n credential_dir = os.path.join(home_dir, '.credentials')\n if not os.path.exists(credential_dir):\n os.makedirs(credential_dir)\n credential_path = os.path.join(credential_dir,\n 'gmail-python-quickstart.json')\n\n store = Storage(credential_path)\n credentials = store.get()\n return credentials\n\n def _create_message(self, to, subject, body):\n \"\"\"\n Create a message for an Email.\n :param to: Email address of the receiver\n :param subject: Subject of the email\n :param body: Body of the email\n \"\"\"\n message = MIMEText(body)\n message['to'] = to\n message['subject'] = subject\n return {\n 'raw': base64.urlsafe_b64encode(message.as_bytes()).decode()\n }\n\n\nclass ChuckyCustomGenerator(Callable):\n \"\"\"\n warnings:: Class not completed yet\n description:: Class will allow to add customs unique words/functions,\n If user want to create own realization of the bot,\n he should use the CustomGenerator class.\n future:: It will be imported into BotChucky class.\n\n :Example:\n # first create custom functions\n def hello_python():\n return 'Hello Python!'\n\n def news_python():\n return 'Python news!'\n\n my_config = {\n '#Python': hello_python\n }\n\n # Create instance of ChuckyGenerator\n bot = ChuckyCustomGenerator()\n bot.config = my_config\n\n # If we get some text from messenger\n # And we pass an argument to the bot\n\n my_message = 'Hello I want to learn #Python'\n bot(my_message)\n\n The bot will return the result of a custom function: 'Hello Python!'\n\n Update our config, and add topics:\n\n # Add topics\n # For example\n # If we got text with #Python and 'bye' word\n my_config = {\n '#Python': {'news': news_python}\n }\n bot.config = my_config\n\n my_message = 'Hey #Python, and send me your news'\n bot(my_message)\n\n bot will return the result of a custom function: 'Python news!'\n \"\"\"\n config = {}\n\n def get_text(self, text):\n return split_text(text)\n\n @property\n def config_keys(self):\n return self.config.keys()\n\n def check_and_run(self, text):\n func = None\n for key in self.config_keys:\n if key not in text:\n msg = 'Sorry, could you repeat please?'\n return msg\n if key in text:\n func = self.config.get(key)\n if isinstance(func, Callable):\n return func()\n else:\n for topic in self.config.get(key):\n if topic in text:\n func = self.config[key][topic]\n return func()\n\n def __call__(self, text, **kwargs):\n text = self.get_text(text)\n return self.check_and_run(text)\n\n def __str__(self):\n return f'{self.__class__.__name__}' \\\n f'(Your config: {self.config})'\n","repo_name":"sk364/Bot-Chucky","sub_path":"bot_chucky/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":10032,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17566880331","text":"import json\nimport datetime \n\nfrom django.http import JsonResponse\nfrom django.views import View\n\nfrom orders.models import WishList, OrderList, Order, PaymentMethod\nfrom users.models import Address\nfrom users.utils import login_confirm\n\nclass WishListVew(View):\n @login_confirm\n def get(self, request):\n user = request.user\n products = [wishlist.product for wishlist in WishList.objects.filter(user=user)]\n\n result = [\n {\n 'product_id' : product.id,\n 'product_name' : product.name,\n 'product_thumbnail_url' : product.thumbnail_url,\n 'product_selections' : [\n {\n 'size' : product_selection.size,\n 'price' : product_selection.price\n } for product_selection in product.productselection_set.all() \n ]\n } for product in products \n ]\n\n return JsonResponse({'result': result}, status=200)\n\n @login_confirm\n def post(self, request):\n user = request.user\n data = json.loads(request.body)\n\n if WishList.objects.filter(user=user, product_id=data['product_id']).exists():\n return JsonResponse({'result': 'ALREADY_EXIT_ERROR'}, status=400)\n\n WishList.objects.create(user=user, product_id=data['product_id'])\n\n return JsonResponse({'result': 'SUCCESS'}, status=201)\n\nclass ModifyWishListVew(View):\n @login_confirm\n def delete(self, request, product_id):\n user = request.user\n\n if not WishList.objects.filter(user=user, product_id=product_id).exists():\n return JsonResponse({'message': 'DOES_NOT_EXIST_ERROR'}, status=400)\n\n WishList.objects.filter(user=user, product_id=product_id).delete()\n \n return JsonResponse({'result': 'SUCCESS'}, status=204)\n\nclass CartListView(View):\n @login_confirm\n def get(self, request):\n user = request.user\n order = Order.objects.get(user=user, status__name='장바구니')\n\n result = {\n 'order_id' : order.id,\n 'products' : [\n {\n 'product_id' : order_list.product_selection.product.id,\n 'product_name' : order_list.product_selection.product.name,\n 'product_size' : order_list.product_selection.size,\n 'product_price' : order_list.product_selection.price,\n 'product_quantity' : order_list.quantity\n } for order_list in OrderList.objects.filter(order=order)\n ]\n }\n\n return JsonResponse({'result': result}, status=200)\n\n @login_confirm\n def post(self, request):\n user = request.user\n data = json.loads(request.body)\n\n if not Order.objects.filter(user=user, status__name='장바구니').exists():\n order = Order.objects.create(user=user, status_id=1)\n OrderList.objects.create(order=order, product_selection_id=data['product_selection_id'], quantity=1)\n else:\n order = Order.objects.get(user=user, status__name='장바구니')\n if OrderList.objects.filter(order=order, product_selection_id=data['product_selection_id']).exists():\n orderlist = OrderList.objects.get(order=order, product_selection_id=data['product_selection_id'])\n orderlist.quantity += 1\n orderlist.save()\n else:\n OrderList.objects.create(order=order, product_selection_id=data['product_selection_id'], quantity=1)\n\n return JsonResponse({'result': 'SUCCESS'}, status=201)\n \nclass ModifyCartListView(View):\n @login_confirm\n def patch(self, request, product_selection_id):\n user = request.user\n data = json.loads(request.body)\n order = Order.objects.get(user=user, status__name='장바구니')\n\n if not OrderList.objects.filter(order=order, product_selection=product_selection_id).exists():\n return JsonResponse({'message': 'DOES_NOT_EXIST_ERROR'}, status=400)\n \n orderlist = OrderList.objects.get(order=order, product_selection=product_selection_id)\n orderlist.quantity = data['quantity']\n orderlist.save()\n\n return JsonResponse({'result': 'SUCCESS'}, status=200)\n\n @login_confirm\n def delete(self, request, product_selection_id):\n user = request.user\n order = Order.objects.get(user=user, status__name='장바구니')\n\n if not OrderList.objects.filter(order=order, product_selection=product_selection_id).exists():\n return JsonResponse({'message': 'DOES_NOT_EXIST_ERROR'}, status=400)\n\n OrderList.objects.get(order=order, product_selection=product_selection_id).delete()\n \n return JsonResponse({'result': 'SUCCESS'}, status=204)\n\nclass OrderView(View):\n @login_confirm\n def get(self, request):\n user = request.user\n orders = Order.objects.prefetch_related('orderlist_set').filter(user=user, status__name='구매완료')\n\n result = [\n {\n 'address' : order.address.address if order.address else None,\n 'memo' : order.memo if order.memo else None,\n 'payment_method' : order.payment_method.name if order.payment_method else None,\n 'purchased_at' : order.purchased_at,\n 'total_price' : order.total_price,\n 'free_delivery' : order.free_delivery,\n 'products' : [\n {\n 'product_id' : order_list.product_selection.product.name,\n 'product_name' : order_list.product_selection.product.name,\n 'product_size' : order_list.product_selection.size,\n 'product_price' : order_list.product_selection.price,\n 'product_quantity' : order_list.quantity\n } for order_list in order.orderlist_set.all()\n ]\n } for order in orders\n ]\n\n return JsonResponse({'result': result}, status=200)\n\n @login_confirm\n def post(self, request):\n try:\n user = request.user\n data = json.loads(request.body)\n\n if not Order.objects.prefetch_related('orderlist_set').filter(id=data['order_id']).exists():\n return JsonResponse({'message': 'DOES_NOT_EXIST_ERROR'}, status=400)\n \n if Order.objects.get(id=data['order_id']).status_id == 2:\n return JsonResponse({'message': 'ALREADY_PAYED_ERROR'}, status=400)\n\n order = Order.objects.prefetch_related('orderlist_set').get(id=data['order_id'])\n address, create = Address.objects.get_or_create(user=user, address=data['address'])\n payment_method = PaymentMethod.objects.get(id=data['payment_method_id'])\n total_price = sum([order_list.product_selection.price * order_list.quantity for order_list in order.orderlist_set.all()])\n\n FREE_DELIVERY = 30000\n\n order.status_id = 2\n order.address = address\n order.memo = data.get('memo', None)\n order.payment_method = payment_method if payment_method else None\n order.total_price = total_price\n order.free_delivery = True if total_price > FREE_DELIVERY else False\n order.purchased_at = datetime.datetime.now()\n order.save()\n\n return JsonResponse({'result': 'SUCCESS'}, status=200)\n\n except KeyError: \n return JsonResponse({'message':'KEY_ERROR'}, status=400)","repo_name":"hwaya2828/WESOP-backend","sub_path":"orders/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":8350,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"74733526330","text":"\"\"\"\n-- ПРИНЦИП РАБОТЫ --\n# Если честно, алгоритм очень хорошо описан в теории к спринту: \"Реализация бинарного поиска с помощью рекурсии\".\nПоскольку масив был сдвинут, то дополнительно для использования бинарного поиска нам необходимо\nпровести проверку на нахождение отсортированной половины массива. Например, сравнив левый конец с медианой.\nПосле этой проверки воспользуемся рекурсивным бинарным поиском.\n\n-- ДОКАЗАТЕЛЬСТВО КОРРЕКТНОСТИ --\nПо условию сказано \"... в массиве только уникальные элементы\".\n - В случае right > mid мы будем рассматривать отрезок [mid; right), значит можно отбрость левую\n половину массива, проводя дальнейший поиск только в правой половине, т.е. новые границы поиска\n будут: [mid; right) и этот полуинтервал будет точно отсортирован.\n - В случае left < mid мы будем рассматривать отрезок [left; mid), значит можно отбрость правую\n половину массива, проводя дальнейший поиск только в левой половине, т.е. новые границы поиска\n будут: [left; mid) - этот полуинтервал будет также отсортирован.\n\n-- ВРЕМЕННАЯ СЛОЖНОСТЬ --\nПри использовании бинарного поиска мы сокращаем интервал поиска в два раза, значит временная сложность будет O(log(n)).\n\n-- ПРОСТРАНСТВЕННАЯ СЛОЖНОСТЬ --\nМы храним исходный массив и используем рекурсивные вызовы. В худшем случае (см. временную сложность) мы имеем\nсложность O(logN), т.к. делаем O(logN) вызовов и храним на стеке значения каждого вызова.\nТакая сложность также связана с тем, что в python не поддерживается оптимизация хвостовой рекурсии.\n\"\"\"\n\n\"\"\"\n-- ID успешной посылки --\n56013520\n\"\"\"\n\n\"\"\"\nВернулся к первой версии реализации поиска. Спасибо большое за совет! Изначально запутался в промежутках...,\nпоэтому, чтобы проходило все тест-кейсы и написал те функции, которые находили границы. Но теперь разобрался с ними.\nСпасибо :) !\n\"\"\"\n\n\n# Рассчет середины массива.\ndef count_mid(left, right):\n return (right + left) // 2\n\n\n# Рекурсивный бинарный поиск.\ndef binary_search(array_, left_, right_, element_):\n # Находим середину массива.\n mid = count_mid(left_, right_)\n\n # Если найденная середина массива и есть искомый элемент, то сразу вернем результат.\n if array_[mid] == element_:\n return mid\n\n # Базовый случай рекурсии.\n if right_ - left_ <= 0:\n return -1\n\n # Случай, когда левая часть отсортирована.\n if array_[left_] <= array_[mid]:\n if array_[left_] <= element_ < array_[mid]:\n return binary_search(array_, left_, mid, element_)\n else:\n return binary_search(array_, mid + 1, right_, element_)\n\n # Случай, когда левая часть отсортирована.\n if array_[left_] > array_[mid]:\n if array_[mid] < element_ <= array_[right_]:\n return binary_search(array_, mid + 1, right_, element_)\n else:\n return binary_search(array_, left_, mid, element_)\n\n\n# Финкция поиска элемента в массиве.\ndef broken_search(array_, element_):\n left_ = 0\n right_ = len(array_) - 1\n\n return binary_search(array_, left_, right_, element_)\n\n\nif __name__ == '__main__':\n n = int(input()) # Длина массива.\n element = int(input()) # Элемент массива, индекс которого надо найти.\n array = list(map(int, input().strip().split()))[:n] # Исходный массив.\n\n print(broken_search(array, element))\n","repo_name":"DimaZzZz101/Yandex_Practicum_Algorithms","sub_path":"Sprint_3/Final_tasks/Task_1/search_in_broken_array.py","file_name":"search_in_broken_array.py","file_ext":"py","file_size_in_byte":4942,"program_lang":"python","lang":"ru","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"74875693370","text":"from random import randint\nfrom timeit import repeat\n\nfrom random import randint\n\ndef quick_sort(array):\n \n # Base cases\n if len(array) < 2:\n return array\n\n low, same, high = [], [], []\n\n # Select the pivot randomly\n pivot = array[randint(0, len(array) - 1)]\n\n for item in array:\n # Elements smaller than the pivot go to the low list\n if item < pivot:\n low.append(item)\n # Elements equal to pivot go to the same list.\n elif item == pivot:\n same.append(item)\n # Elements larger than pivot go to the high list. \n elif item > pivot:\n high.append(item)\n\n # The final result combines the sorted low, same, sorted\n return quick_sort(low) + same + quick_sort(high)\n\n# Test cases\n\n# Performance evaluation\ndef check_performance(algorithm):\n\n ARRAY_LENGTH = 1000\n\n # Generate an array of `ARRAY_LENGTH` random integer between 0 and 999\n array = [randint(0, 1000) for i in range(ARRAY_LENGTH)]\n\n # Execute the code ten times and return the time that each execution took\n stmt = f\"{algorithm}({array})\"\n setup_code = f\"from __main__ import {algorithm}\"\n times = repeat(setup=setup_code, stmt=stmt, repeat=3, number=10)\n\n # Display the name of the algorithm and the minimum time it took to run\n print(f\"Algorithm: {algorithm}, minimum execution time: {min(times)}\")\n\n\nif __name__ == \"__main__\":\n\n check_performance(\"quick_sort\")","repo_name":"edab/DSA_Quick_Reference","sub_path":"solutions/quick_sort.py","file_name":"quick_sort.py","file_ext":"py","file_size_in_byte":1455,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"39974825123","text":"\"\"\"\n Factory with static methods (python equivalence) to transform\n Foaf instance in Lucene Document and inverse process\n\"\"\"\nfrom PyLucene import Document, Field\n#from futil.foaf.foaf import Foaf\n\nfields = {\n \"name\" : ( Field.Store.YES, Field.Index.TOKENIZED),\n \"nick\" : ( Field.Store.YES, Field.Index.TOKENIZED),\n# \"sha\" : ( Field.Store.YES, Field.Index.UN_TOKENIZED),\n \"uri\" : ( Field.Store.YES, Field.Index.UN_TOKENIZED),\n \"friends\": (Field.Store.YES, Field.Index.UN_TOKENIZED),\n \"geolat\": (Field.Store.YES, Field.Index.UN_TOKENIZED),\n \"geolong\": (Field.Store.YES, Field.Index.UN_TOKENIZED)\n}\n\n\"\"\"\n Helper class to create static methods\n http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/52304\n\"\"\"\nclass Static:\n def __init__(self, anycallable):\n self.__call__ = anycallable\n\n\nclass FoafDocumentFactory:\n\n def getDocumentFromFOAF(foaf):\n doc = Document()\n for attr, value in foaf.iteritems():\n if ( fields.has_key(attr)):\n # Now is always a list!\n for x in value:\n if isinstance(x, tuple): #for example (sha, uri)\n doc.add(Field(attr, x[1], fields[attr][0], fields[attr][1]))\n else:\n doc.add(Field(attr, x, fields[attr][0], fields[attr][1]))\n\n else:\n pass\n # DEBUG information print \"E: Field \" + attr + \" ignored in index\"\n return doc\n getDocumentFromFOAF = Static(getDocumentFromFOAF)\n\n def getFOAFFromDocument(doc):\n f = {}\n for key in fields.iterkeys():\n values = doc.getValues(key)\n if not values:\n continue\n if len(values) > 1:\n f[key] = doc.getValues(key)\n else:\n f[key] = values[0]\n return f\n getFOAFFromDocument = Static(getFOAFFromDocument)\n\n","repo_name":"BackupTheBerlios/futil-svn","sub_path":"trunk/src/futil/index/foafDocumentFactory.py","file_name":"foafDocumentFactory.py","file_ext":"py","file_size_in_byte":1903,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"41849779328","text":"\"\"\"ocrarian file manager config\"\"\"\nfrom pathlib import Path\nfrom appdirs import AppDirs\n\nfrom ocrarian import APP_NAME\n\n\nclass Config(AppDirs):\n \"\"\"Config class for ocrarian\n This class is responsible for creating configuration directory and settings file.\n It also handles settings load, save and delete.\"\"\"\n\n def __init__(self):\n super().__init__(appname=APP_NAME)\n self.user_docs_dir = Path(\"~/Documents\").expanduser()\n self.create_directories()\n\n def create_directories(self):\n \"\"\"Create config and docs directories.\"\"\"\n # pylint: disable=no-member\n # Instance of 'user_config_dir' has no 'exists' member (no-member)\n # Instance of 'user_config_dir' has no 'mkdir' member (no-member)\n if not self.user_config_dir.exists():\n self.user_config_dir.mkdir()\n if not self.user_docs_dir.exists():\n self.user_docs_dir.mkdir()\n if not self.user_cache_dir.exists():\n self.user_cache_dir.mkdir()\n","repo_name":"ocrarian/ocrarian.py","sub_path":"ocrarian/common/file_manager/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":1007,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"78"}
+{"seq_id":"73808510013","text":"import os\nimport shlex\nimport contextlib\nfrom functools import partial\n\nfrom tornado import gen\nfrom tornado.process import Subprocess\nfrom tornado import ioloop\nfrom tornado import iostream\n\nimport logging\nlogger = logging.getLogger(__name__)\n\n\nclass ProcessMeta(type):\n PROCESSES = {}\n\n def __call__(cls, name, *args, **kwargs):\n if name in ProcessMeta.PROCESSES:\n raise RuntimeError(\"Duplicate process '{}'\".format(name))\n\n inst = type.__call__(cls, name, *args, **kwargs)\n ProcessMeta.PROCESSES[name] = inst\n return inst\n\n\nclass CommandMixin(object):\n def command_prefix(self):\n return []\n\n def command_suffix(self):\n return []\n\n def __enter__(self):\n pass\n\n def __exit__(self, *_):\n pass\n\n\nclass Argument(CommandMixin):\n def __init__(self, *args):\n self.args = args\n\n def __repr__(self):\n return \"\".format(\" \".join(self.args))\n\n def command_suffix(self):\n return list(self.args)\n\n\nclass Environ(CommandMixin):\n def __init__(self, *args, **kwargs):\n self.environ = dict(args, **kwargs)\n\n def __repr__(self):\n return \"\".format(\" \".join(self.environ,keys()))\n\n def __enter__(self):\n self._original_state = {}\n for key in self.environ:\n if key in os.environ:\n self._original_state[key] = os.environ[key]\n os.environ[key] = self.environ[key]\n\n def __exit__(self, *_):\n for key in self.environ:\n if key in self._original_state:\n os.environ[key] = self._original_state\n else:\n del os.environ[key]\n del self._original_state\n\n\nclass Wrapper(CommandMixin):\n def __init__(self, *args):\n self.prefix = args\n\n def command_prefix(self):\n return list(self.prefix)\n\n\nclass Process(object):\n __metaclass__ = ProcessMeta\n\n command = []\n\n def __init__(self, name, *mixins, **kwargs):\n self.name = name\n self.kwargs = kwargs\n self.mixins = mixins\n\n self.proc = None\n self.exitcode = None\n self.subs = {}\n\n def __repr__(self):\n return \"\".format(\n self.name,\n \" \".join(self.format_command()))\n\n def __str__(self):\n return self.name\n\n def command_vars(self):\n return dict(name=self.name, **self.kwargs)\n\n def format_command(self):\n _vars = self.command_vars()\n\n cmd = self.command[:]\n\n for mixin in self.mixins:\n cmd = mixin.command_prefix() + cmd + mixin.command_suffix()\n\n if isinstance(cmd, (str, unicode)):\n cmd = shlex.split(cmd)\n\n for i, part in enumerate(cmd):\n cmd[i] = part.format(**_vars)\n\n return cmd\n\n @property\n def running(self):\n return self.proc is not None\n\n @property\n def pid(self):\n return self.proc.pid if self.proc else None\n\n def run(self):\n logger.info(\"Running %s\", \" \".join(self.command))\n\n with contextlib.nested(*self.mixins):\n self.proc = Subprocess(\n self.format_command(),\n stdout=Subprocess.STREAM,\n stderr=Subprocess.STREAM)\n\n self._exit_future = self._wait_for_exit()\n self._start_read('stdout', self.proc.stdout)\n self._start_read('stderr', self.proc.stderr)\n\n @gen.coroutine\n def stop(self, kill=False):\n proc = self.proc.proc\n if kill:\n logger.info(\"Killing %r [PID %i]\", self, self.pid)\n proc.kill()\n else:\n logger.info(\"Terminating %r [PID %i]\", self, self.pid)\n proc.terminate()\n\n exitcode = yield self._exit_future\n raise gen.Return(exitcode)\n\n def subscribe(self, id, cb):\n logger.info(\"Subscription: %r added\", id)\n self.subs[id] = cb\n\n def unsubscribe(self, id):\n try:\n del self.subs[id]\n logger.info(\"Subscription: %r removed\", id)\n except KeyError:\n pass\n\n @gen.coroutine\n def _wait_for_exit(self):\n exitcode = yield self.proc.wait_for_exit(False)\n logger.info(\"Process %r stopped\", self)\n self.proc = None\n self.exitcode = exitcode\n self.subs = {}\n raise gen.Return(exitcode)\n\n def _start_read(self, label, stream):\n loop = ioloop.IOLoop.current()\n loop.add_future(self._read(label, stream), lambda f: f)\n\n @gen.coroutine\n def _read(self, label, stream):\n try:\n while self.running:\n msg = yield stream.read_until('\\n')\n for sub_cb in self.subs.itervalues():\n sub_cb(self, label, msg)\n except iostream.StreamClosedError:\n pass\n","repo_name":"xlevus/python-gurgle","sub_path":"gurgle/process.py","file_name":"process.py","file_ext":"py","file_size_in_byte":4782,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17501992090","text":"import csv\nimport pymysql\n\n# Connect to the database\ndb = pymysql.connect(host='localhost', user='UNAME', password='PASSWORD', database='DB_NAME')\ncursor = db.cursor()\n\n# Insert movies from CSV into the database\ndef insert_movies_from_csv():\n with open('data.csv', 'r') as csv_file:\n csv_reader = csv.DictReader(csv_file)\n for row in csv_reader:\n insert_movie(row['name'], row['time'])\n\n# Insert movies into the database\ndef insert_movie(name, time):\n try:\n query = \"INSERT INTO movies (name, time) VALUES (%s, %s)\"\n cursor.execute(query, (name, time))\n db.commit()\n print(f\"Movie '{name}' loaded successfully!\")\n except Exception as e:\n db.rollback()\n print(\"Error:\", e)\n\nif __name__ == \"__main__\":\n insert_movies_from_csv()\n db.close()\n","repo_name":"Cyber-Zypher/Cinema-Tickets-Booking-System","sub_path":"load_movie.py","file_name":"load_movie.py","file_ext":"py","file_size_in_byte":820,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"30835178842","text":"# https://www.tutorialspoint.com/gensim/gensim_creating_tf_idf_matrix.htm\n# https://honingds.com/blog/natural-language-processing-with-python/\n\nimport itertools\nfrom collections import defaultdict\n\nfrom gensim.corpora import Dictionary\nfrom nltk import word_tokenize\n\n\ndef lda_gensim_parse_post(classification, blogs):\n tokenized_docs = [word_tokenize(doc) for doc in blogs]\n dictionary = Dictionary(tokenized_docs)\n bag_of_words_corpus = [dictionary.doc2bow(tokenized_doc) for tokenized_doc in tokenized_docs]\n\n total_word_count = defaultdict(int)\n for word_id, word_count in itertools.chain.from_iterable(bag_of_words_corpus):\n total_word_count[word_id] += word_count\n\n # Create a sorted list from the defaultdict: sorted_word_count\n sorted_word_count = sorted(total_word_count.items(), key=lambda w: w[1], reverse=True)\n\n # Print the top 1 words across all documents alongside the count\n for word_id, word_count in sorted_word_count[:1]:\n print(classification + \" is talking about '\" + str(dictionary.get(word_id)) + \"', with \" + str(\n word_count) + \" occurence\")\n","repo_name":"jaycedel/TextMiningAssignment2","sub_path":"lda_gensim_topic_modelling.py","file_name":"lda_gensim_topic_modelling.py","file_ext":"py","file_size_in_byte":1119,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"10901800018","text":"import scipy.io\nimport numpy as np\nfrom tqdm import tqdm\n\nclass CSL_data_preprocessing:\n \n def load_from_mat(self,fileAddress):\n \"\"\"\n Load data from a MATLAB file.\n\n Args:\n fileAddress (str): The address of the MATLAB file.\n\n Returns:\n numpy.ndarray: The data stored under the key 'Data' in the MATLAB file.\n \"\"\"\n return scipy.io.loadmat(fileAddress)[\"gestures\"]\n\n def extra_data(self,fileAddress):\n \"\"\"\n Load and concatenate data from multiple MATLAB files, generating corresponding labels.\n\n Args:\n fileAddress (str): The address of the directory containing the MATLAB files.\n\n Returns:\n tuple: A tuple containing two numpy.ndarrays. The first element is the concatenated data,\n and the second element is the corresponding labels. The third element is the number of classes\n \"\"\"\n data = None\n label = None\n num_class = 26\n for i in tqdm(range(1, num_class+1),desc=\"Processing Files\"):\n for trial in tqdm(range(10),desc=\"Processing trials in file\"):\n if data is None:\n data = self.load_from_mat(\n f\"{fileAddress}/gest{i}.mat\")[trial][0].T\n label = np.repeat(i-1,data.shape[0])\n else:\n temp = self.load_from_mat(\n f\"{fileAddress}/gest{i}.mat\")[trial][0].T\n data = np.concatenate((data,temp),axis=0)\n label = np.concatenate((label,np.repeat(i-1,temp.shape[0])),axis=0)\n return data,label,num_class\n\n def NormalizeData(self,data):\n # data = (data - data.mean())/(data.std())\n return data\n","repo_name":"MIC-Laboratory/CNN-HD-sEMG-Classifier","sub_path":"utils/CSL_data_preprocessing.py","file_name":"CSL_data_preprocessing.py","file_ext":"py","file_size_in_byte":1768,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"22961731692","text":"def digits(n: int):\n count = 0\n while n != 0:\n n //= 10\n count += 1\n return count\n\ndef power(x: int, y: int):\n if y == 0:\n return 1\n elif y == 1:\n return x\n return x * power(x, y-1)\n\ndef get_digit(num: int, index: int):\n return num // power(10, index) % 10\n\ndef split_at(num: int, index: int):\n # Index starting from the right\n return int(str(num)[:-index]), int(str(num)[-index:])\n\ndef third_grade_multiplication(a: int, b: int):\n suma = 0\n for i in range(digits(b)):\n digit = get_digit(b, i)\n suma += a * digit * power(10, i)\n return suma\n\ndef rec_int_mult(a: int, b:int):\n pass\n\ndef karatsuba_multiplication(a: int, b: int):\n if a < 10 or b < 10:\n return a * b\n \n # Size of the numbers\n m = min(digits(a), digits(b)) // 2\n\n # Split the digit sequences in the middle\n high1, low1 = split_at(a, m)\n high2, low2 = split_at(b, m)\n\n # Recursive calls to numbers approximately half the size\n z0 = karatsuba_multiplication(low1, low2)\n z1 = karatsuba_multiplication(low1 + high1, low2 + high2)\n z2 = karatsuba_multiplication(high1, high2)\n \n return z2 * power(10, m * 2) + (z1 - z2 - z0) * power(10, m) + z0\n\n\nif __name__ == '__main__':\n\n test_cases = [\n (3, 4, 3 * 4),\n (11, 12, 11 * 12),\n (111, 12, 111 * 12),\n (123456, 33, 123456 * 33),\n (654, 987654, 654 * 987654),\n ]\n \n for a, b, solution in test_cases:\n result = third_grade_multiplication(a, b)\n string = f'third_grade_multiplication({a:6}, {b:6}) = {result:9}'\n string += ' ' * (60 - len(string))\n print(string, f'\\t\\tTest: {\"OK\" if solution == result else \"NOT OK\"}')\n\n result = karatsuba_multiplication(a, b)\n string = f' karatsuba_multiplication({a:6}, {b:6}) = {result:9}'\n string += ' ' * (60 - len(string))\n print(string, f'\\t\\tTest: {\"OK\" if solution == result else \"NOT OK\"}\\n')","repo_name":"daalgi/algorithms","sub_path":"divide-and-conquer/integer_multiplication.py","file_name":"integer_multiplication.py","file_ext":"py","file_size_in_byte":1972,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"78"}
+{"seq_id":"69820941054","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n # @Time : 2019-04-26 17:22\n # @Author : Awiny\n # @Site :\n # @Project : pytorch_i3d\n # @File : charades_checkpoints.py\n # @Software: PyCharm\n # @Github : https://github.com/FingerRec\n # @Blog : http://fingerrec.github.io\n\"\"\"\nimport scipy.io\nimport os\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' #close the warning\n\"\"\" Defines functions used for checkpointing models and storing model scores \"\"\"\nimport os\nimport torch\nimport shutil\nfrom collections import OrderedDict\n\n\ndef ordered_load_state(model, chkpoint):\n \"\"\"\n Wrapping the model with parallel/dataparallel seems to\n change the variable names for the states\n This attempts to load normally and otherwise aligns the labels\n of the two statese and tries again.\n \"\"\"\n try:\n model.load_state_dict(chkpoint)\n except Exception: # assume order is the same, and use new labels\n print('keys do not match model, trying to align')\n modelkeys = model.state_dict().keys()\n fixed = OrderedDict([(z,y)\n for (x,y),z in zip(chkpoint.items(), modelkeys)])\n model.load_state_dict(fixed)\n\n\ndef load(args, model, optimizer):\n if args.resume:\n if os.path.isfile(args.resume):\n print(\"=> loading checkpoint '{}'\".format(args.resume))\n chkpoint = torch.load(args.resume)\n if isinstance(chkpoint, dict) and 'state_dict' in chkpoint:\n args.start_epoch = chkpoint['epoch']\n mAP = chkpoint['mAP']\n ordered_load_state(model, chkpoint['state_dict'])\n optimizer.load_state_dict(chkpoint['optimizer'])\n print(\"=> loaded checkpoint '{}' (epoch {})\"\n .format(args.resume, chkpoint['epoch']))\n return mAP\n else:\n ordered_load_state(model, chkpoint)\n print(\"=> loaded checkpoint '{}' (just weights)\"\n .format(args.resume))\n return 0\n else:\n raise ValueError(\"no checkpoint found at '{}'\".format(args.resume))\n return 0\n\n\ndef score_file(scores, filename):\n with open(filename, 'w') as f:\n for key, val in sorted(scores.items()):\n f.write('{} {}\\n'.format(key, val))\n\n\ndef save(epoch, args, model, optimizer, is_best, scores):\n state = {\n 'epoch': epoch + 1,\n 'arch': args.arch,\n 'state_dict': model.state_dict(),\n 'mAP': scores['mAP'],\n 'optimizer': optimizer.state_dict(),\n }\n filename = \"{}/model.pth.tar\".format('checkpoints/charades')\n score_file(scores, \"{}/model_{:03d}.txt\".format('checkpoints/charades', epoch+1))\n torch.save(state, filename)\n if is_best:\n bestname = \"{}/model_best.pth.tar\".format('checkpoints/charades')\n score_file(scores, \"{}/model_best.txt\".format('checkpoints/charades', epoch+1))\n shutil.copyfile(filename, bestname)\n","repo_name":"FingerRec/RHE","sub_path":"tools/utils/charades_checkpoints.py","file_name":"charades_checkpoints.py","file_ext":"py","file_size_in_byte":3014,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"78"}
+{"seq_id":"26307274064","text":"import os, time\nimport zipfile\n\n\ndef origin_zip():\n # 准备一个空的压缩文件命名为demoreport.zip\n # 开启添加多个文件的模式 ‘a’\n with zipfile.ZipFile(file='../demoreport.zip', mode='a') as f:\n # 添加被压缩的文件,并命名为新的文件名\n f.write(r'C:\\Users\\janti\\PycharmProjects\\autotest-android\\YYandroid\\UItest\\report\\demoreport.html',\n 'demoreport11_5.html')\n f.close()\n\n\ndef upgrade_zipfile():\n with open('../data/result.csv',encoding='UTF-8') as f:\n text=f.read()\n # print(text)\n with zipfile.ZipFile(file='../test_result_set.zip', mode='a') as z:\n for filename in os.listdir('../UItest/report/screenshot'):\n if filename in text:\n z.write('../UItest/report/screenshot/%s' % filename, filename)\n z.write('../UItest/report/test_report.html','UItestreport.html')\n for videofile in os.listdir('../data'):\n if videofile in text:\n print(videofile)\n # pass\n z.write('../data/%s'%videofile,videofile)\n z.close()\n\n\n\n\nif __name__ == '__main__':\n t1 = time.time()\n upgrade_zipfile()\n t2 = time.time()\n print(t2 - t1)\n","repo_name":"jantionMH/fuckyou","sub_path":"YYandroid/Utility/zipsource.py","file_name":"zipsource.py","file_ext":"py","file_size_in_byte":1300,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"10349804013","text":"# 15489 : 파스칼 삼각형\n# https://www.acmicpc.net/problem/15489\n\nr, c, w = map(lambda x: int(x), input().split())\nheight0 = r-1\nwidth0 = c-1\nheight1 = r-1 + w\n# print(height0, width0, height1)\n\npascal = [[1] * i for i in range(1, height1+1)]\nfor i in range(1, height1):\n for j in range(1, i):\n pascal[i][j] = pascal[i-1][j-1] + pascal[i-1][j]\n# print(pascal)\n\nresult = 0\nindex = 1\nfor i in range(height0, height1):\n for j in range(width0, width0+index):\n # print(pascal[i][j])\n result += pascal[i][j]\n index += 1\n\nprint(result)","repo_name":"Ir2placeable/Area51","sub_path":"python/DP/silver_4/15489.py","file_name":"15489.py","file_ext":"py","file_size_in_byte":563,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33845033286","text":"from .deformation import calc_deformation\nfrom .utils import sample_from_discrete_uniform\nimport numpy\nimport torch\n\nimport pyro\nimport pyro.distributions as dist\nimport pyro.primitives as prim\n\n\nclass GeoModel:\n def __init__(self, heads, reference_times, observed_deformations):\n # initializations\n self.heads = heads\n self.reference_times = reference_times\n self.observed_deformations = torch.FloatTensor(observed_deformations)\n\n @pyro.infer.config_enumerate\n def model(self):\n kv = prim.sample(name=\"kv\", fn=dist.Cauchy(loc=-5, scale=3)).item()\n sskv = prim.sample(name=\"sskv\",\n fn=dist.Cauchy(loc=-3.5, scale=3)).item()\n sske = prim.sample(name=\"sske\", fn=dist.Cauchy(loc=-5, scale=3)).item()\n nclay = sample_from_discrete_uniform(name=\"nclay\",\n values=list(range(5, 11)))\n claythick = 5\n interp_times, defm, _, _ = calc_deformation(\n time=self.reference_times, head=self.heads, Kv=10**kv,\n Sskv=10**sskv, Sske=10**sske, claythick=claythick,\n nclay=nclay)\n\n aligned_deformation = torch.FloatTensor(\n numpy.interp(self.reference_times, interp_times, defm))\n\n for i in pyro.plate(\"data_loop\", len(self.observed_deformations)):\n prim.sample(name=\"data_{}\".format(i),\n fn=dist.Normal(self.observed_deformations[i], 2),\n obs=aligned_deformation[i])\n\n return kv, sskv, sske, nclay\n\n def model_importance_sampling(self, observed_deformations):\n kv = prim.sample(name=\"kv\", fn=dist.Cauchy(loc=-5, scale=3)).item()\n sskv = prim.sample(name=\"sskv\",\n fn=dist.Cauchy(loc=-3.5, scale=3)).item()\n sske = prim.sample(name=\"sske\", fn=dist.Cauchy(loc=-5, scale=3)).item()\n # print(kv, sskv, sske)\n nclay = sample_from_discrete_uniform(name=\"nclay\",\n values=list(range(5, 11)))\n interp_times, defm, _, _ = calc_deformation(\n time=self.reference_times, head=self.heads, Kv=10**kv,\n Sskv=10**sskv, Sske=10**sske, claythick=5, nclay=nclay)\n\n aligned_deformation = torch.FloatTensor(\n numpy.interp(self.reference_times, interp_times, defm))\n\n for i in pyro.plate(\"data_loop\", len(observed_deformations)):\n prim.sample(name=\"data_{}\".format(i),\n fn=dist.Normal(observed_deformations[i], 2),\n obs=aligned_deformation[i])\n\n return kv, sskv, sske, nclay\n\n\ndef read_from_file(file_location: str):\n \"\"\"Read the input data from the given file\n\n Args:\n file_location (str): The location of the given input file.\n\n Returns:\n (List[int], List[int], List[int]): Three lists of the same length,\n corresponding to: heads, reference time and observed deformation\n measurments.\n \"\"\"\n with open(file_location) as f:\n heads = [float(h) for h in f.readline().split()]\n reference_times = [float(h) for h in f.readline().split()]\n observed_deformations = [float(h) for h in f.readline().split()]\n if len(heads) != len(reference_times) != len(observed_deformations):\n raise RuntimeError(\"Heads, reference times and observed deformations\"\n \" were expected to have the same length but they\"\n \" didn't.\")\n return heads, reference_times, observed_deformations\n","repo_name":"Ganoash/CSE4340-PP","sub_path":"Problems/GEO/geo.py","file_name":"geo.py","file_ext":"py","file_size_in_byte":3551,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"41754580104","text":"from pico2d import *\nimport gobj\nfrom missile import Missile\nimport gfw\n\nclass Player:\n KEY_MAP = {\n (SDL_KEYDOWN, SDLK_UP): (0, 1),\n (SDL_KEYDOWN, SDLK_DOWN): (0, -1),\n (SDL_KEYDOWN, SDLK_LEFT): (-1, 0),\n (SDL_KEYDOWN, SDLK_RIGHT): ( 1, 0),\n (SDL_KEYUP, SDLK_UP): (0, -1),\n (SDL_KEYUP, SDLK_DOWN): (0, 1),\n (SDL_KEYUP, SDLK_LEFT): ( 1, 0),\n (SDL_KEYUP, SDLK_RIGHT): (-1, 0)\n }\n KEYDOWN_Z = (SDL_KEYDOWN, SDLK_z)\n KEYDOWN_SHIFT = (SDL_KEYDOWN, SDLK_LSHIFT)\n KEYUP_Z = (SDL_KEYUP, SDLK_z)\n KEYUP_SHIFT = (SDL_KEYUP, SDLK_LSHIFT)\n\n SHOOTHING_INTERVAL = 0.05\n ROTATING_SLOWEFF = 3.14 / 180\n\n def __init__(self):\n self.image = gfw.image.load('./res/player.png')\n self.slowEffImage = gfw.image.load('./res/eff_sloweffect.png') # 임시, 원래값: 0,0,64,64\n gobj.set_image_alpha(self.slowEffImage, 220)\n self.pos = get_canvas_width() // 2, 100\n self.delta = 0, 0\n self.fidx = 0\n self.action = 0\n self.time = 0\n self.shooting_time = 0\n self.speed = 300\n self.degree = 0\n self.death = 0\n self.deathtime = 0\n self.slowing = False\n self.shooting = False\n self.nodamage = False\n Player.player = self\n\n global BOUNDARY_LEFT, BOUNDARY_RIGHT, BOUNDARY_DOWN, BOUNDARY_UP\n BOUNDARY_LEFT = 16\n BOUNDARY_DOWN = 24\n BOUNDARY_RIGHT = get_canvas_width() - BOUNDARY_LEFT\n BOUNDARY_UP = get_canvas_height() - BOUNDARY_DOWN\n\n def update(self):\n x, y = self.pos\n dx, dy = self.delta\n x += dx * self.speed * gfw.delta_time\n y += dy * self.speed * gfw.delta_time\n\n x = clamp(BOUNDARY_LEFT, x, BOUNDARY_RIGHT)\n y = clamp(BOUNDARY_DOWN, y, BOUNDARY_UP)\n self.pos = x, y\n\n self.time += gfw.delta_time\n self.shooting_time += gfw.delta_time\n frame = self.time * 10\n if self.action != 0:\n self.fidx = (int(frame) % 4) + 4\n else:\n self.fidx = int(frame) % 8\n\n if self.shooting and self.shooting_time > Player.SHOOTHING_INTERVAL:\n self.fire()\n\n if self.nodamage:\n self.deathtime += gfw.delta_time\n gobj.set_image_alpha(self.image, 125)\n if self.deathtime > 3:\n self.nodamage = False\n self.deathtime = 0\n gobj.set_image_alpha(self.image, 255)\n self.degree += Player.ROTATING_SLOWEFF\n\n def draw(self):\n width, height = 32, 48\n sx = self.fidx * width\n sy = (self.image.h - 48) - (self.action * height)\n\n self.image.clip_draw(sx, sy, width, height, *self.pos)\n\n if self.slowing:\n self.slowEffImage.clip_composite_draw(0,0,64,64, self.degree, '', *self.pos ,64,64)\n\n def handle_event(self, e):\n pair = (e.type, e.key)\n if pair in Player.KEY_MAP:\n self.delta = gobj.pos_add(self.delta, Player.KEY_MAP[pair])\n dx = self.delta[0]\n self.action = \\\n 1 if dx < 0 else \\\n 2 if dx > 0 else 0\n\n if pair == Player.KEYDOWN_Z:\n self.shooting = True\n elif pair == Player.KEYUP_Z:\n self.shooting = False\n\n if pair == Player.KEYDOWN_SHIFT:\n self.slowing = True\n self.speed = 100\n elif pair == Player.KEYUP_SHIFT:\n self.slowing = False\n self.speed = 300\n\n\n def fire(self):\n self.shooting_time = 0\n x, y = self.pos\n halfX = 16 # 32 // 2\n halfY = 24 # 48 // 2\n m1 = Missile(x - halfX, y + halfY)\n m2 = Missile(x + halfX, y + halfY)\n gfw.world.add(gfw.layer.missile, m1)\n gfw.world.add(gfw.layer.missile, m2)\n\n def get_BB(self):\n hw = 2\n hh = 2\n return (self.pos[0] - hw, self.pos[0] + hw, self.pos[1] - hh, self.pos[1] +hh)\n\n","repo_name":"Brw-fox/2DGP","sub_path":"기말 프로젝트/player.py","file_name":"player.py","file_ext":"py","file_size_in_byte":3959,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"20941697715","text":"\"\"\" database.py\n created by: Lucca Giusti (solucca)\n This module uses functional programming to handle\n data from the mariadb mysql database\n\"\"\"\n\nimport mysql.connector\nfrom typing import Dict, List, Union\nimport json\n\nDATABASE_CONFIG = {}\nwith open(\"./credentials.json\", \"r\") as f:\n DATABASE_CONFIG = json.load(f)\n\n\ndef create_type_table(payload: dict) -> str:\n \"\"\"Creates a table for the entity type provided in the payload\"\"\"\n entity_type: str = payload.get(\"type\")\n\n # Connect to the MariaDB database\n conn = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = conn.cursor()\n\n try:\n # Create the SQL query to create the table\n table_name = entity_type.lower()\n columns = [\n \"id INT AUTO_INCREMENT PRIMARY KEY\",\n \"entity_id VARCHAR(64)\",\n \"timestamp DATETIME DEFAULT CURRENT_TIMESTAMP\",\n ]\n\n for key, data in payload.items():\n if key in [\"id\", \"type\"]:\n continue\n\n # Add columns based on the data type\n if data[\"type\"] == \"str\":\n columns.append(f\"{key} VARCHAR(255)\")\n elif data[\"type\"] == \"float\":\n columns.append(f\"{key} FLOAT\")\n elif data[\"type\"] == \"int\":\n columns.append(f\"{key} INT\")\n elif data[\"type\"] == \"datetime\":\n columns.append(f\"{key} DATETIME\")\n else:\n raise ValueError(f\"Invalid attribute type: {data['type']}\")\n\n # Construct the CREATE TABLE query\n create_query = f\"CREATE TABLE IF NOT EXISTS {table_name} ({', '.join(columns)})\"\n\n # Execute the query\n cursor.execute(create_query)\n conn.commit()\n\n finally:\n # Close the cursor and connection\n cursor.close()\n conn.close()\n\n return table_name\n\n\ndef table_exists(table_name: str) -> bool:\n cnx = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = cnx.cursor()\n cursor.execute(\"SHOW TABLES;\")\n tables = []\n for name in cursor:\n tables.append(name[0])\n cursor.close()\n cnx.close()\n return table_name.lower() in tables\n\n\ndef match_schema(payload: dict) -> bool:\n if not table_exists(payload.get(\"type\")):\n return False\n entity_type = payload.get(\"type\").lower()\n cnx = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = cnx.cursor()\n try:\n cursor.execute(f\"DESCRIBE {entity_type};\")\n data = [i[0] for i in cursor.fetchall()]\n for key, value in payload.items():\n if key in [\"id\", \"type\"]:\n continue\n elif not key in data:\n return False\n return True\n\n finally:\n cursor.close()\n cnx.close()\n\n\ndef get_columns(type: str) -> Union[List, Dict]:\n \"\"\" Get Columns of table\n \"\"\"\n try:\n cnx = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = cnx.cursor()\n cursor.execute(f\"DESCRIBE {type};\")\n out = [{\"Field\": i[0], \"Type\": i[1]} for i in cursor.fetchall()]\n return out\n except Exception as e:\n return {\"error\" : str(e)}\n finally:\n cnx.close()\n cursor.close()\n\n\ndef insert_entity(payload: dict) -> bool:\n \"\"\"Saves the Entity in a Table.\n If there is no Table for this Type of entity, one is created.\n \"\"\"\n entity_id, entity_type = payload.get(\"id\"), payload.get(\"type\").lower()\n\n if not table_exists(entity_type):\n create_type_table(payload)\n if not match_schema(payload):\n return False\n\n # Construct the SQL insert query\n columns = [\"entity_id\"]\n values = [entity_id]\n\n for key, data in payload.items():\n if key in [\"id\", \"type\"]:\n continue\n\n # Extract the value and add it to the lists\n columns.append(key)\n values.append(data[\"value\"])\n\n placeholders = \", \".join([\"%s\" for _ in values])\n sql = f\"INSERT INTO {entity_type} ({', '.join(columns)}) VALUES ({placeholders})\"\n\n # Connect to the MariaDB database\n cnx = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = cnx.cursor()\n\n try:\n # Execute the SQL query\n cursor.execute(sql, values)\n cnx.commit()\n\n finally:\n # Close the cursor and connection\n cursor.close()\n cnx.close()\n\n return True\n\n\ndef get_entity(entity_id: str, n: int = None) -> Dict[str, object]:\n \"\"\"Saves the Entity in a Table.\n If there is no Table for this Type of entity, one is created.\n \"\"\"\n if not \":\" in entity_id:\n return {\"Error\": \"Wrong Format :\"}\n entity_type, entity_id = entity_id.split(\":\")\n table_name = entity_type.lower()\n\n if not table_exists(entity_type):\n return {\"Error\": f\"No data for the type {entity_type}\"}\n if not n:\n n = 1\n\n # Connect to the MariaDB database\n conn = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = conn.cursor()\n\n try:\n # Construct the SELECT query\n select_query = f\"SELECT * FROM {table_name} WHERE entity_id = %s ORDER BY timestamp DESC LIMIT {n}\"\n cursor.execute(select_query, (entity_id,))\n\n data = cursor.fetchmany(n)\n out = []\n for row in data:\n entry = {}\n for i in range(1, len(cursor.column_names)):\n entry[cursor.column_names[i]] = row[i]\n out.append(entry)\n\n return out\n\n finally:\n # Close the cursor and connection\n cursor.close()\n conn.close()\n\n\ndef modify_type_table(type: str, new_column: dict):\n \"\"\" new_column example: \n {\"field\":\"humidity\", \"type\":\"float\"}\n \"\"\"\n if not table_exists(type): return {\"error\": f\"Type {type} does not exist\"}\n columns = get_columns(type)\n for column in columns:\n if column[\"Field\"] == new_column[\"field\"]:\n return {\"error\": f\"Field {new_column['field']} already exists\"}\n \n if new_column[\"type\"] == \"str\":\n datatype = \"VARCHAR(255)\"\n elif new_column[\"type\"] == \"float\":\n datatype = \"FLOAT\"\n elif new_column[\"type\"] == \"int\":\n datatype = \"INT\"\n elif new_column[\"type\"] == \"datetime\":\n datatype = \"DATETIME\"\n else:\n return {\"error\": f\"Datatype {new_column['type']} does not exist\"}\n try:\n cnx = mysql.connector.connect(**DATABASE_CONFIG)\n cursor = cnx.cursor()\n cursor.execute(f\"ALTER TABLE {type} ADD {new_column['field']} {datatype};\")\n return {'success', f'added column: {type}'}\n \n except Exception as e:\n return {'error':str(e)}\n finally:\n cursor.close()\n cnx.close()\n\n\nif __name__ == \"__main__\":\n print(\"Running test on database\")\n print(f\"Testing connection: {table_exists('SoilSensor')}\")\n payload = {\n \"id\": \"01\",\n \"type\": \"SoilSensor\",\n \"temperature\": {\"type\": \"float\", \"value\": 15.0},\n \"soil_humidity\": {\"type\": \"int\", \"value\": 28},\n \"last_water\": {\"type\": \"datetime\", \"value\": \"2020-07-26 20:20:20\"},\n }\n payload2 = {\n \"id\": \"02\",\n \"type\": \"weather\",\n \"temperature\": {\"type\": \"float\", \"value\": 15.0},\n }\n print(get_entity(\"weather:02\"))\n","repo_name":"solucca/HomeSystem","sub_path":"backend/database.py","file_name":"database.py","file_ext":"py","file_size_in_byte":7162,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"17272572816","text":"'other solution'\n\n\ndef isSolved(board):\n for i in range(0,3):\n if board[i][0] == board[i][1] == board[i][2] != 0:\n return board[i][0]\n elif board[0][i] == board[1][i] == board[2][i] != 0:\n return board[0][i]\n\n if board[0][0] == board[1][1] == board[2][2] != 0:\n return board[0][0]\n elif board[0][2] == board[1][1] == board[2][0] != 0:\n return board[0][0]\n\n elif 0 not in board[0] and 0 not in board[1] and 0 not in board[2]:\n return 0\n else:\n return -1\n\n\n\n\n'勉强勉强通过'\n\n\ndef alleq(lis):\n if max(lis) == min(lis) and min(lis) > 0:\n return max(lis)\n return False\n\n# print(alleq([1,1,1]))\n\ndef isSolved(board):\n\n for i in range(3):\n if alleq(board[i]):\n return alleq(board[i])\n for j in range(3):\n column = [ board[0][j], board[1][j], board[2][j]]\n if alleq(column):\n return alleq(column)\n left = [board[0][0], board[1][1], board[2][2]]\n if alleq(left): return alleq(left)\n right = [board[0][2], board[1][1], board[2][0]]\n if alleq(right): return alleq(right)\n if any([ 0 in row for row in board]): return -1\n return 0\n\n\n\n\n\ndef alleq(lis):\n if max(lis) == min(lis):\n return max(lis)\n return False\n\n# print(alleq([1,1,1]))\n\ndef isSolved(board):\n if any([ 0 in row for row in board]): return -1\n for i in range(3):\n if alleq(board[i]):\n return alleq(board[i])\n for j in range(3):\n column = [ board[0][j], board[1][j], board[2][j]]\n if alleq(column):\n return alleq(column)\n left = [board[0][0], board[1][1], board[2][2]]\n if alleq(left): return alleq(left)\n right = [board[0][2], board[1][1], board[2][0]]\n if alleq(right): return alleq(right)\n return 0\n\nboard = [[1,1,2],\n [2,1,2],\n [2,1,2]]\n# print(any ([ i for i in [1,True, False] if i < 1 ]))\nprint(isSolved(board))\n'''\n# You can use test.expect(boolean, [optional] string) to test your code\ntest.expect(isSolved( [[1,1,1],\n [2,1,2],\n [2,1,2]] ) is -1)\n'''\n'''descriptor tic-tac-toe checker\n\nIf we were to set up a Tic-Tac-Toe game, we would want to know whether the board's current state is solved, wouldn't we? Our goal is to create a function that will check that for us!\n\nAssume that the board comes in the form of a 3x3 array, where the value is 0 if a spot is empty, 1 if it is an X, or 2 if it is an O, like so:\n\n[[0,0,1],\n [0,1,2],\n [2,1,0]]\n\nWe want our function to return -1 if the board is not solved yet, 1 if X won, 2 if O won, or 0 if it's a cat's game (i.e. a draw).\n\nYou may assume that the board passed in is valid in the context of a game of Tic-Tac-Toe.\n\n\n'''\n\n","repo_name":"shubham25namdeo/Leetcode","sub_path":"codewars/tic-tac-toe.py","file_name":"tic-tac-toe.py","file_ext":"py","file_size_in_byte":2720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"78"}
+{"seq_id":"32962659490","text":"#!/usr/bin/python\n#\n# UiHandler.py - handler for the RadioPi user interface\n#\n# (c) David Haworth\n\nimport string\nimport traceback\n\nfrom LcdHandler import LcdHandler\nfrom MainMenu import MainMenu\nfrom Browser import Browser\nfrom StationList import StationList\nfrom MountMenu import MountMenu\nfrom SysMenu import SysMenu\nfrom NetMenu import NetMenu\nfrom MessageScreen import MessageScreen\nfrom AskYesNo import AskYesNo\n\n# UI modes\nmode_Disconnected = 0\nmode_StartupWait = 1\nmode_Startup = 2\nmode_Home = 3\nmode_Menu = 4\n\n# Home screen positions\nrow_artist\t= 1\ncol_artist\t= 1\nrow_album\t= 2\ncol_album\t= 1\nrow_title\t= 3\ncol_title\t= 1\nrow_time\t= 4\ncol_time\t= 1\nrow_dur\t\t= 4\ncol_dur\t\t= 6\nrow_vol\t\t= 4\ncol_vol\t\t= 18\t# 3 digits\n\n\nclass UiHandler:\n\tdef __init__(self, eq, mpd):\n\t\tself.eq = eq\t\t\t\t\t\t# Event queue\n\t\tself.mpd = mpd\t\t\t\t\t\t# MPD handler\n\t\tself.mode = mode_Disconnected\t\t# Current mode\n\t\tself.count = 0\n#----\n\t\tself.vol = -1\t\t\t\t\t\t# State of info on screen. Only update if changed.\n\t\tself.time = -1\n\t\tself.artist = \"\"\n\t\tself.album = \"\"\n\t\tself.title = \"\"\n\t\tself.dur = -1\n\t\tself.state = \"\"\n\t\tself.force = True\t\t\t\t\t# Forced update\n#----\n\t\tself.menu = None\n\t\tself.menustack = []\n\t\tself.lcd = LcdHandler()\n\t\tif self.lcd.Open():\n\t\t\tself.mode = mode_StartupWait\n\t\t\tself.count = 2\n\n#===========================================================\n# Enter the menu system\n#===========================================================\n\tdef Enter(self):\n\t\tself.menustack = []\t\t\t\t\t\t\t\t\t# Clear out existing menus (if any)\n\t\tself.menu = MainMenu(self, self.lcd, self.eq)\t\t# Create the main menu.\n\t\tself.menu.Show()\n\n#===========================================================\n# Enter the music browser at the specified place.\n#===========================================================\n\tdef EnterBrowser(self, dir):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = Browser(self, self.lcd, self.eq, dir)\t# Create a browser.\n\t\tself.menu.Show()\n\n#===========================================================\n# Enter the station list menu.\n#===========================================================\n\tdef EnterStationList(self):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = StationList(self, self.lcd, self.eq)\t# Create the menu.\n\t\tself.menu.Show()\n\n#===========================================================\n# Enter the mount menu.\n#===========================================================\n\tdef EnterMountMenu(self):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = MountMenu(self, self.lcd, self.eq)\t\t# Create the menu.\n\t\tself.menu.Show()\n\n#===========================================================\n# Enter the network menu.\n#===========================================================\n\tdef EnterNetMenu(self):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = NetMenu(self, self.lcd, self.eq)\t\t# Create the menu.\n\t\tself.menu.Show()\n\n#===========================================================\n# Enter the system menu.\n#===========================================================\n\tdef EnterSysMenu(self):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = SysMenu(self, self.lcd, self.eq)\t\t# Create the menu.\n\t\tself.menu.Show()\n\n#===========================================================\n# Show a message.\n#===========================================================\n\tdef ShowMessage(self, m, ack):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t\t\t# Push current menu.\n\t\tself.menu = MessageScreen(self, self.lcd, self.eq, m, ack)\t# Create the message screen.\n\t\tself.menu.Show()\n\n#===========================================================\n# Ask a question\n#===========================================================\n\tdef AskYesNo(self, m):\n\t\tself.menustack.append(self.menu)\t\t\t\t\t# Push current menu.\n\t\tself.menu = AskYesNo(self, self.lcd,self.eq, m)\t\t# Create the yes/no screen.\n\t\tself.menu.Show()\n\n#===========================================================\n# Receive the answer from AskYesNo\n#===========================================================\n\tdef Answer(self, ans):\n\t\tif len(self.menustack) == 0:\n\t\t\tself.menu = None\n\t\t\tself.ModeHome()\n\t\telse:\n\t\t\tself.menu = self.menustack.pop()\n\t\t\tself.menu.Event(ans)\n\t\t\tself.menu.Show()\n\n#===========================================================\n# Go back up a level.\n#===========================================================\n\tdef Back(self):\n\t\tif len(self.menustack) == 0:\n\t\t\tself.menu = None\n\t\t\tself.ModeHome()\n\t\telse:\n\t\t\tself.menu = self.menustack.pop()\n\t\t\tself.menu.Show()\n\n#===========================================================\n# Handle a timer event\n#===========================================================\n\tdef Timer(self):\n\t\tif self.mode == mode_Disconnected:\t\t# If disconnected, attempt to open\n\t\t\tif self.lcd.Open():\n\t\t\t\tself.mode = mode_StartupWait\n\t\t\t\tself.count = 2\n\n\t\telif self.mode == mode_StartupWait:\t\t# After successful open, let the terminal settle\n\t\t\tself.count -= 1\n\t\t\tif self.count <= 0:\n\t\t\t\tself.StartupScreen()\n\t\t\t\tself.mode = mode_Startup\n\t\t\t\tself.force = True\n\t\t\t\tself.count = 2\n\n\t\telif self.mode == mode_Startup:\t\t\t# Display splash screen.\n\t\t\ttry:\n\t\t\t\tself.count -= 1\n\t\t\t\tif self.count <= 0:\n\t\t\t\t\tself.mode = mode_Home\n\t\t\t\t\tself.force = True\n\t\t\t\t\tself.HomeScreen()\t\t\t# Redraw home screen\n\t\t\texcept:\n\t\t\t\tself.mode = mode_Disconnected\n\t\t\t\tprint(traceback.format_exc())\n\n\t\telif self.mode == mode_Home:\t\t\t# Update home screen\n\t\t\ttry:\n\t\t\t\tself.HomeScreen()\n\t\t\texcept:\n\t\t\t\tself.mode = mode_Disconnected\n\t\t\t\tprint(traceback.format_exc())\n\n\t\telse:\t\t\t\t\t\t\t\t\t# After some inactivity, revert to home screen.\n\t\t\tself.count -= 1\n\t\t\tif self.count <= 0:\n\t\t\t\tself.count = 0\n\t\t\t\tself.mode = mode_Home\n\t\t\t\tself.force = True\n\n\t\treturn False\n\n#===========================================================\n# Handle a user input event\n#===========================================================\n\tdef Event(self, evt):\n\t\tif self.mode < mode_Home:\t\t\t\t# Ignore events during start sequence.\n\t\t\treturn False\n\n\t\tself.count = 30\t\t\t\t\t\t\t# Restart the inactivity timer\n\n\t\tif self.mode == mode_Home:\t\t\t\t# On home screen, 'menu', 'ok', 'right' or 'Rl' enters menu.\n\t\t\t# Translate rotary control events and requeue those not for me.\n\t\t\tif evt == \"R+\":\n\t\t\t\tself.eq.PutEvent(\"vol+\")\n\t\t\t\treturn True\n\t\t\tif evt == \"R-\":\n\t\t\t\tself.eq.PutEvent(\"vol-\")\n\t\t\t\treturn True\n\t\t\tif evt == \"Rs\":\n\t\t\t\tself.eq.PutEvent(\">/||\")\n\t\t\t\treturn True\n\t\t\tif evt == \"Rl\":\t\t# Handle this one locally (don't requeue)\n\t\t\t\tevt = \"menu\"\n\n\t\t\tif evt == \"menu\" or evt == \"ok\" or evt == \"right\":\n\t\t\t\tself.mode = mode_Menu\n\t\t\t\tself.Enter()\n\t\t\t\treturn True\n\n\t\t\treturn False\t\t\t\t\t\t# Ignore all other events while on home screen.\n\n\t\t# Translate rotary control events for menu mode and handle locally (don't requeue)\n\t\tif evt == \"R+\":\n\t\t\tevt = \"down\"\n\t\telif evt == \"R-\":\n\t\t\tevt = \"up\"\n\t\telif evt == \"Rl\":\n\t\t\tevt = \"home\"\n\t\telif evt == \"Rs\":\n\t\t\tevt = \"ok\"\n\n\t\tif evt == \"home\":\t\t\t\t\t\t# Anywhere in the menus, 'home' exits back to the home screen\n\t\t\tself.menu = None\t\t\t\t\t# Clear the menu stack\n\t\t\tself.menustack = []\n\t\t\tself.ModeHome()\n\t\t\treturn True\n\n\t\t# Menu mode. Handle up/down/left/back as navigation on existing menu stack.\n\t\tif evt == \"back\" or evt == \"left\" :\n\t\t\tself.Back()\n\t\t\treturn True\n\n\t\telif evt == \"up\" or evt == \"R-\":\n\t\t\tself.menu.PtrUp()\n\t\t\treturn True\n\n\t\telif evt == \"down\" or evt == \"R+\":\n\t\t\tself.menu.PtrDown()\n\t\t\treturn True\n\n\t\t# All other events in menu mode: pass on to the individual menu\n\t\treturn self.menu.Event(evt)\n\n\n#===========================================================\n# StartupScreen() - display startup screen\n#===========================================================\n\n\tdef StartupScreen(self):\n\t\tif self.force:\n\t\t\tself.lcd.HomeAndClear()\n\t\t\tself.lcd.WriteAt(1, 1, \"RadioPi\\r\\n\")\n\t\t\tself.lcd.WriteAt(4, 1, \"waiting for mpd\")\n\t\tself.force = False\n\n#===========================================================\n# ModeHome() - switch to 'home' mode (leave menu)\n#===========================================================\n\n\tdef ModeHome(self):\n\t\tself.mode = mode_Home\n\t\tself.force = True\n\t\tself.HomeScreen()\n\n#===========================================================\n# HomeScreen() - read mpd status and display it\n#===========================================================\n\n\tdef HomeScreen(self):\n\t\tl_artist = \"\"\n\t\tl_album = \"\"\n\t\tl_title = \"\"\n\t\tl_dur = -1\n\t\tl_time = -1\n\t\tl_vol = -1\n\t\tl_file = \"\"\n\t\tl_name = \"\"\n\t\tl_state = \"\"\n\t\tinternetRadio = False;\n\n\t\ts = self.mpd.Status()\n\t\tif s:\n\t\t\tfor k in s.keys():\n\t\t\t\tif k == \"volume\":\n\t\t\t\t\tl_vol = int(s[\"volume\"])\n\t\t\t\telif k == \"state\":\n\t\t\t\t\tl_state = s[\"state\"]\n\t\t\t\telif k == \"elapsed\":\n\t\t\t\t\tl_time = int(float(s[\"elapsed\"]))\n\n\t\tif l_state == \"stop\":\n\t\t\tl_time = -1\n\n\t\ts = self.mpd.CurrentSong()\n\t\tif s:\n\t\t\tfor k in s.keys():\n\t\t\t\tif k == \"artist\":\n\t\t\t\t\tl_artist = s[\"artist\"]\n\t\t\t\telif k == \"album\":\n\t\t\t\t\tl_album = s[\"album\"]\n\t\t\t\telif k == \"title\":\n\t\t\t\t\tl_title = s[\"title\"]\n\t\t\t\telif k == \"time\":\n\t\t\t\t\tl_dur = int(s[\"time\"])\n\t\t\t\telif k == \"file\":\n\t\t\t\t\tl_file = s[\"file\"]\n\t\t\t\telif k == \"name\":\n\t\t\t\t\tl_name = s[\"name\"]\n\n\t\t\tif l_file[0:7] == \"http://\" or l_file[0:8] == \"https://\":\n\t\t\t\tinternetRadio = True\n\t\t\t\tl_artist = l_name\n\t\t\t\tif l_album == \"\":\n\t\t\t\t\tif string.find(l_title, \" - \") >= 0:\n\t\t\t\t\t\tl_album, l_title = string.split(l_title, \" - \", 1)\n\t\t\t\t\telif string.find(l_title, \"-\") >= 0:\n\t\t\t\t\t\tl_album, l_title = string.split(l_title, \"-\", 1)\n\t\t\telif l_title == \"\" and l_file != \"\":\n\t\t\t\tx, l_title = os.path.split(l_file)\n\n\t\telse:\n\t\t\t# No track\n\t\t\tl_artist = \" RadioPi\"\n\t\t\tl_album = \"\"\n\t\t\tl_title = \"===== no track =====\"\n\n\t\tif self.force:\n\t\t\tself.lcd.HomeAndClear()\n\n\t\tif self.force or self.artist != l_artist:\n\t\t\tself.artist = l_artist\n\t\t\tself.lcd.WriteAt(row_artist, col_artist, self.artist)\n\t\t\tself.lcd.ClearEol()\n\n\t\tif self.force or self.album != l_album:\n\t\t\tself.album = l_album\n\t\t\tself.lcd.WriteAt(row_album, col_album, self.album)\n\t\t\tself.lcd.ClearEol()\n\n\t\tif self.force or self.title != l_title:\n\t\t\tself.title = l_title\n\t\t\tself.lcd.WriteAt(row_title, col_title, self.title);\n\t\t\tself.lcd.ClearEol()\n\n\t\tif internetRadio:\n\t\t\tself.dur = -2\n\t\t\tself.time = -2\n\t\t\tbuf = \"%-16s\" % (\"Stream: \"+l_state)\n\t\t\tself.lcd.WriteAt(row_time, col_time, buf)\n\t\telif l_dur < 0:\n\t\t\tif self.force or self.dur != l_dur:\n\t\t\t\tself.dur = l_dur\n\t\t\t\tself.time = -1\n\t\t\t\tself.lcd.WriteAt(row_time, col_time, \" \")\n\t\telse:\n\t\t\tif self.force or self.time != l_time:\n\t\t\t\tself.time = l_time\n\t\t\t\tif l_time < 0:\n\t\t\t\t\ts_time = \"--:--\"\n\t\t\t\telse:\n\t\t\t\t\tmins = l_time / 60\n\t\t\t\t\tsecs = l_time % 60\n\t\t\t\t\ts_time = \"%02d:%02d\"%(mins, secs)\n\t\t\t\tself.lcd.WriteAt(row_time, col_time, s_time)\n\n\t\t\tif self.force or self.dur != l_dur:\n\t\t\t\tself.dur = l_dur\n\t\t\t\tmins = l_dur / 60\n\t\t\t\tsecs = l_dur % 60\n\t\t\t\ts_time = \"/%02d:%02d\"%(mins, secs)\n\t\t\t\tself.lcd.WriteAt(row_dur, col_dur, s_time)\n\t\t\n\t\tif self.force or self.vol != l_vol:\n\t\t\tself.vol = l_vol\n\t\t\tif l_vol < 0:\n\t\t\t\ts_vol = \"---\"\n\t\t\telse:\n\t\t\t\ts_vol = \"%3d\"%(l_vol)\n\t\t\tself.lcd.WriteAt(row_vol, col_vol, s_vol)\n\n\t\tself.force = False\n\n","repo_name":"TheLancashireman/RadioPi","sub_path":"controller/UiHandler.py","file_name":"UiHandler.py","file_ext":"py","file_size_in_byte":10967,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"23266172573","text":"'''\nGiven an array nums of n integers, are there elements a, b, c in nums such that a + b + c = 0?\nFind all unique triplets in the array which gives the sum of zero.\nNotice that the solution set must not contain duplicate triplets.\nInput: nums = [-1,0,1,2,-1,-4]\nOutput: [[-1,-1,2],[-1,0,1]]\n'''\n\n\nclass Solution:\n def threeSum(self, nums):\n if len(nums) < 3:\n return []\n nums.sort()\n ans = []\n i = 0\n while i < len(nums) - 2:\n left = i + 1\n right = len(nums) - 1\n while left < right:\n temp = nums[left] + nums[right] + nums[i]\n if temp == 0:\n ans.append([nums[left], nums[right], nums[i]])\n left += 1\n while left < right and nums[left] == nums[left - 1]:\n left += 1\n elif temp < 0:\n left += 1\n else:\n right -= 1\n while i < len(nums) - 2 and nums[i] == nums[i + 1]:\n i += 1\n i += 1\n return ans\n","repo_name":"ryanSoftwareEngineer/algorithms","sub_path":"arrays and matrices/15_3Sum.py","file_name":"15_3Sum.py","file_ext":"py","file_size_in_byte":1098,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"33674869197","text":"import os\nimport re\nimport sys\nfrom maint.lib import maintainers\n\nfrom style.repo import GitRepo\n\n\ndef _printErrorQuit(error_message):\n \"\"\"\n Print an error message, followed my a help message and inform failure.\n\n @param error_message A message describing the error that caused the\n failure.\n \"\"\"\n print(error_message)\n\n print(\n \"The commit has been cancelled, but a copy of it can be found in \"\n + sys.argv[1]\n + \" : \"\n )\n\n print(\n \"\"\"\n--------------------------------------------------------------------------\n \"\"\"\n )\n print(open(sys.argv[1], \"r\").read())\n print(\n \"\"\"\n--------------------------------------------------------------------------\n \"\"\"\n )\n\n print(\n \"\"\"\nThe first line of a commit must contain one or more gem5 tags separated by\ncommas (see MAINTAINERS.yaml for the possible tags), followed by a colon and\na commit title. There must be no leading nor trailing whitespaces.\n\nThis header line must then be followed by an empty line. A detailed message,\nalthough highly recommended, is not mandatory and can follow that empty line.\n\ne.g.:\n cpu: Refactor branch predictors\n\n Refactor branch predictor code to improve its readability, moving functions\n X and Y to the base class...\n\ne.g.:\n mem,mem-cache: Improve packet class readability\n\n The packet class...\n\"\"\"\n )\n sys.exit(1)\n\n\ndef _validateTags(commit_header):\n \"\"\"\n Check if all tags in the commit header belong to the list of valid\n gem5 tags.\n\n @param commit_header The first line of the commit message.\n \"\"\"\n\n # List of valid tags\n maintainer_dict = maintainers.Maintainers.from_file()\n valid_tags = [tag for tag, _ in maintainer_dict]\n\n # Remove non-tag 'pmc' and add special tags not in MAINTAINERS.yaml\n valid_tags.remove(\"pmc\")\n valid_tags.extend([\"RFC\", \"WIP\"])\n\n tags = \"\".join(commit_header.split(\":\")[0].split()).split(\",\")\n if any(tag not in valid_tags for tag in tags):\n invalid_tag = next((tag for tag in tags if tag not in valid_tags))\n _printErrorQuit(\"Invalid Gem5 tag: \" + invalid_tag)\n\n\n# Go to git directory\nos.chdir(GitRepo().repo_base())\n\n# Get the commit message\ncommit_message = open(sys.argv[1]).read()\n\n# The first line of a commit must contain at least one valid gem5 tag, and\n# a commit title\ncommit_message_lines = commit_message.splitlines()\ncommit_header = commit_message_lines[0]\ncommit_header_match = re.search(\n \"^(fixup! )?(\\S[\\w\\-][,\\s*[\\w\\-]+]*:.+\\S$)\", commit_header\n)\nif commit_header_match is None:\n _printErrorQuit(\"Invalid commit header\")\nif commit_header_match.group(1) == \"fixup! \":\n sys.exit(0)\n_validateTags(commit_header_match.group(2))\n\n# Make sure commit title does not exceed threshold. This line is limited to\n# a smaller number because version control systems may add a prefix, causing\n# line-wrapping for longer lines\ncommit_title = commit_header.split(\":\")[1]\nmax_header_size = 65\nif len(commit_header) > max_header_size:\n _printErrorQuit(\n \"The commit header (tags + title) is too long (\"\n + str(len(commit_header))\n + \" > \"\n + str(max_header_size)\n + \")\"\n )\n\n# Then there must be at least one empty line between the commit header and\n# the commit description\nif commit_message_lines[1] != \"\":\n _printErrorQuit(\n \"Please add an empty line between the commit title and \"\n \"its description\"\n )\n\n# Encourage providing descriptions\nif re.search(\n \"^(Signed-off-by|Change-Id|Reviewed-by):\", commit_message_lines[2]\n):\n print(\"Warning: Commit does not have a description\")\n\nsys.exit(0)\n","repo_name":"gem5/gem5","sub_path":"util/git-commit-msg.py","file_name":"git-commit-msg.py","file_ext":"py","file_size_in_byte":3644,"program_lang":"python","lang":"en","doc_type":"code","stars":1196,"dataset":"github-code","pt":"78"}
+{"seq_id":"21274466922","text":"import pyttsx3\r\nimport speech_recognition as sr\r\nimport datetime\r\nimport os\r\nimport openai\r\nimport wikipedia\r\nimport subprocess\r\nimport pywhatkit\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\nfrom pyautogui import moveTo,write,leftClick\r\nimport pyjokes\r\nimport webbrowser\r\nimport pyautogui\r\nimport datetime\r\nimport colorama\r\nfrom colorama import Fore\r\nimport sys\r\nfrom PyQt5 import QtGui\r\nfrom PyQt5.QtCore import QTimer,QTime,QDate\r\nfrom PyQt5.QtCore import *\r\nfrom PyQt5.QtGui import *\r\nfrom PyQt5.QtWidgets import *\r\nfrom jarvisSuperUI import Ui_Form\r\n\r\n\r\nengine = pyttsx3.init('sapi5')\r\nvoices = engine.getProperty('voices')\r\nengine.setProperty('voices',voices[0].id)\r\n\r\n\r\n \r\n \r\n \r\ndef speak(audio):\r\n \r\n engine.say(audio)\r\n engine.runAndWait()\r\n\r\nclass MainThread(QThread):\r\n def __init__(self):\r\n \r\n super(MainThread,self).__init__()\r\n \r\n def run(self):\r\n self.TaskExection()\r\n\r\n def commands(self):\r\n \r\n r = sr.Recognizer()\r\n with sr.Microphone() as source:\r\n \r\n r.pause_threshold = 1\r\n r.adjust_for_ambient_noise(source , duration=1)\r\n audio=r.listen(source)\r\n try:\r\n \r\n query = r.recognize_google(audio, language='en-in')\r\n print(f\"You Just Said: {query}\\n\")\r\n except Exception as e:\r\n print(e)\r\n \r\n query=\"none\"\r\n \r\n return query\r\n \r\n\r\n def wishings(self):\r\n hour = int(datetime.datetime.now().hour)\r\n if hour >=0 and hour<12:\r\n \r\n speak('Good morning BOSS')\r\n elif hour>=12 and hour<17:\r\n \r\n speak(\"Good Afternoon BOSS\")\r\n elif hour >=17 and hour<21:\r\n \r\n speak(\"Good Evening BOSS\")\r\n else:\r\n \r\n speak(\"Good Night BOSS\")\r\n\r\n\r\n def TaskExection(self):\r\n \r\n def open_another_file(self):\r\n \r\n speak('you switched to hand sign mode')\r\n os.startfile(\"hand_assistant.py\")\r\n \r\n self.wishings()\r\n while True:\r\n self.query = self.commands().lower()\r\n if 'time' in self.query:\r\n strTime = datetime.datetime.now().strftime(\"%H:%M:%S\")\r\n speak(\"Sir, The time is: \" + strTime)\r\n print(strTime)\r\n elif 'question' in self.query or 'doubts' in self.query or 'question' in self.query or 'doubt' in self.query:\r\n openai.api_key = \"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\"\r\n speak('ask anything')\r\n prompt = self.commands().lower()\r\n if prompt in \"none\":\r\n continue\r\n print(\"Your query:\", prompt)\r\n response = openai.Completion.create(engine=\"text-davinci-002\", prompt=prompt, max_tokens=2000)\r\n a= response[\"choices\"][0][\"text\"]\r\n print(a)\r\n speak(a)\r\n\r\n elif 'wikipedia' in self.query:\r\n speak(\"Searching in wikipedia\")\r\n try:\r\n self.query=self.query.replace(\"wikipedia\", '')\r\n results = wikipedia.summary(self.query, sentences=1)\r\n speak(\"According to Wikipedia..\")\r\n print(results)\r\n speak(results)\r\n except:\r\n print(\"No results found..\")\r\n speak(\"no results found\")\r\n\r\n elif 'play' in self.query:\r\n playquery=self.query.replace('play','')\r\n speak(\"Playing \" + playquery)\r\n pywhatkit.playonyt(playquery)\r\n elif 'gmail' in self.query:\r\n speak('Opening Gmail....')\r\n url = \"https://www.gmail.com\"\r\n webbrowser.open_new_tab(url)\r\n elif 'youtube' in self.query:\r\n speak('Opening....')\r\n url = \"https://www.youtube.com\"\r\n webbrowser.open_new_tab(url)\r\n elif 'music' in self.query:\r\n speak('Playing')\r\n music_dir = 'fav songs'\r\n songs = os.listdir(music_dir)\r\n print(songs) \r\n os.startfile(os.path.join(music_dir, songs[0])) \r\n \r\n elif 'temperature' in self.query:\r\n speak('sure, Tell me the city')\r\n print(\"Tell me the city\")\r\n city = self.commands().lower()\r\n print(city)\r\n url = f\"https://www.google.com/search?q=temperature+{city}\"\r\n r = requests.get(url)\r\n soup = BeautifulSoup(r.text, \"html.parser\")\r\n temp = soup.find(\"div\", class_=\"BNeawe\").text\r\n temp = temp.split(\":\")[-1].split()[0]\r\n z= f\"The temperature in {city} is {temp}.\"\r\n print(z)\r\n speak(z)\r\n continue\r\n \r\n\r\n elif 'screenshot ' in self.query:\r\n now = datetime.datetime.now()\r\n filename = \"screenshot{}.png\".format(now.strftime(\"%Y-%m-%d-%H-%M-%S\"))\r\n pyautogui.screenshot(filename)\r\n\r\n \r\n\r\n\r\n elif 'joke' in self.query:\r\n jarvisJoke = pyjokes.get_joke()\r\n print(jarvisJoke)\r\n speak(jarvisJoke)\r\n else:\r\n speak(\"please say the command again\")\r\n\r\ndef open_another_file():\r\n \r\n speak('you switched to hand sign mode')\r\n \r\n os.startfile(\"hand_assistant.py\")\r\n \r\n \r\n \r\n\r\nstartExecution = MainThread()\r\n\r\nclass Main(QMainWindow):\r\n def __init__(self):\r\n super().__init__()\r\n self.ui = Ui_Form()\r\n self.ui.setupUi(self)\r\n\r\n self.ui.startPushButton.clicked.connect(self.startTask)\r\n self.ui.quitPushButton.clicked.connect(open_another_file)\r\n\r\n def startTask(self, text):\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\telex.gif\")\r\n self.ui.ironManBackground.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # ironmanGIF\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\listeningGIF.gif\")\r\n self.ui.ironManGIF.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # dateLabel\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\gggf.jpg\")\r\n self.ui.dateLabel.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # timeLabel\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\gggf.jpg\")\r\n self.ui.timeLabel.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # startLabelNotButton\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\20230401_185949_0000.png\")\r\n self.ui.startLabelNotButton.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # quitLabelNotButton\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\20230401_191740_0000.png\")\r\n self.ui.quitLabelNotButton.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n # earthGIF\r\n self.ui.movie = QtGui.QMovie(\"GUI files\\\\Earth.gif\")\r\n self.ui.earthGIF.setMovie(self.ui.movie)\r\n self.ui.movie.start()\r\n\r\n timer = QTimer(self)\r\n timer.timeout.connect(self.showTime)\r\n timer.start(1000)\r\n startExecution.start()\r\n\r\n def showTime(self):\r\n currentTime = QTime.currentTime()\r\n currentDate = QDate.currentDate()\r\n labelTime = currentTime.toString('hh:mm:ss')\r\n labelDate = currentDate.toString(Qt.ISODate)\r\n self.ui.dateTextBrowser.setText(f\"Date: {labelDate}\")\r\n self.ui.timeTextBrowser.setText(f\"Date: {labelTime}\")\r\n \r\n\r\napp = QApplication(sys.argv)\r\njarvis = Main()\r\njarvis.show()\r\nexit(app.exec_())\r\n","repo_name":"Aseefmohamed/Virtual-Personal-Assistant-by-using-Hand-Gesture-and-Voice-Assistant-for-Disabilities","sub_path":"Telex/jarvis.py","file_name":"jarvis.py","file_ext":"py","file_size_in_byte":7862,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"30956839533","text":"import re\nimport sys\n\n\ndef main():\n print(parse(input(\"HTML: \")))\n\n\ndef parse(s):\n try:\n matches = re.search(r'^(?:.+)\"(?:https?.+youtube.+com/)(?:embed/)(.+?)\"(?:.+)$', s)\n if not matches:\n return \"None\"\n else:\n html = (\"https://youtu.be/\" + matches.group(1))\n return html\n\n except(ValueError):\n return \"None\"\n\n\n\n\nif __name__ == \"__main__\":\n main()","repo_name":"KimagureWizard/CS50","sub_path":"cs50p/week7/watch_f/watch.py","file_name":"watch.py","file_ext":"py","file_size_in_byte":422,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"456024227","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat May 20 20:48:46 2023\n\n@author: dev\n\"\"\"\n\nimport cv2\nimport numpy as np\nimport os\nfrom tqdm import tqdm\n\ndef threshold_mask(mask):\n _, thresholded = cv2.threshold(mask, 127, 255, cv2.THRESH_BINARY)\n return thresholded\n\ndef combine_masks(mask1, mask2):\n combined = np.maximum(mask1, mask2)\n return combined\n\n# Path to the folders containing the masks\nmask_folder1 = './ISIC 2016 maskgen outputs for UNET/MASKS/BW/'\nmask_folder2 = './ISIC 2016 maskgen outputs for UNET/MASKS/REFERENCE/'\n\n# Get the list of mask file names from folder1\nmask_files = os.listdir(mask_folder1)\n\n# Iterate over each mask file\nfor file in tqdm(mask_files, total=1279):\n # Load masks from both folders\n mask1_path = os.path.join(mask_folder1, file)\n mask2_path = os.path.join(mask_folder2, file)\n \n mask1 = cv2.imread(mask1_path, cv2.IMREAD_GRAYSCALE)\n mask2 = cv2.imread(mask2_path, cv2.IMREAD_GRAYSCALE)\n \n # Perform thresholding\n mask1 = threshold_mask(mask1)\n mask2 = threshold_mask(mask2)\n \n # Combine the masks\n combined_mask = combine_masks(mask1, mask2)\n \n # Save the resulting combined mask\n output_path = './ISIC 2016 maskgen outputs for UNET/MASKS/' + file\n if not cv2.imwrite(output_path, combined_mask):\n raise Exception('Image not saved')\n","repo_name":"DevBhuyan/Unsup-Segmentation","sub_path":"union.py","file_name":"union.py","file_ext":"py","file_size_in_byte":1361,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"40004092258","text":"from flask import Flask, request, jsonify\nimport uuid\nfrom langchain.document_loaders import TextLoader\nfrom langchain.indexes import VectorstoreIndexCreator\nimport constants\n\nfrom google.cloud import texttospeech\nimport os, json\nimport openai\nfrom typing import Tuple\nfrom flask_cors import CORS, cross_origin\n\nimport io\nimport pydub\nimport ffmpeg\n\napp = Flask(__name__)\ncors = CORS(app)\n\nopenai.api_key = constants.OPEN_AI_KEY\nos.environ[\"OPENAI_API_KEY\"] = constants.OPEN_AI_KEY\n\napp = Flask(__name__)\nhost = '127.0.0.1'\nport = 8080\ndb = {}\nloader = TextLoader('resources/jio.txt')\nindex = VectorstoreIndexCreator().from_loaders([loader])\n\n@app.route('/healthcheck')\ndef hello_world():\n return 'Team Ode to Code goes to Dubai'\n\n\n@app.route('/query', methods=['POST'])\ndef raise_query():\n print(\"here\")\n session_id = request.headers.get(\"session-id\")\n print(session_id)\n # print(session_id)\n # print(str(request.files.keys()))\n # print(str(request.form.keys()))\n # print(str(request.get_json()))\n text = None\n print(request)\n if 'audio' in request.files:\n text = convert_audio_to_text(request.files['audio'])\n\n elif 'text' in request.get_json():\n text = request.json['text']\n else:\n return jsonify({'message': 'No file part in the request'}), 400\n print(f\"Incoming text: {text}\")\n translated_text, language = translate_to_english(text)\n if translated_text:\n english_response, category = append_user_query_to_conversation(session_id, translated_text)\n regional_response = translate_from_english(f\"{language};{english_response}\")\n # convert_text_to_audio(language+\"-IN\", regional_response)\n return jsonify({'category': category, 'message': regional_response}), 200\n else:\n return jsonify({'message': 'Could not understand/translate the message, please try again'}), 200\n\n\n@app.route('/createSession', methods=['POST'])\ndef create_session():\n session_uuid = str(uuid.uuid1())\n db[session_uuid] = \"\"\n return session_uuid\n\n\n@app.route('/activeSessions', methods=['GET'])\ndef get_all_active_sessions():\n return str(db.keys())\n\n\ndef convert_audio_to_text(audio_file) -> str:\n audio_file.save(\"input.mp3\")\n audio = open(\"input.mp3\", \"rb\")\n transcription = openai.Audio.transcribe(\"whisper-1\", audio)\n audio.close()\n os.remove('input.mp3')\n return transcription.text.strip()\n\ndef translate_to_english(text) -> Tuple[str, str]:\n prompt = f\"\"\"Assume you are a assistant that translates source language to English. \n Response should have the source language in BCP-47 format and then the translation in English.\n Dont use your brain, just translate.\n Q: what is going on\n A: en, what is going on\n \n Q: nenu chaala manchi manishini\n A: te, I am a very good human\n \n Q: {text}\n A:\n \"\"\"\n translate_response = openai.Completion.create(\n model=\"text-davinci-003\",\n prompt=prompt,\n max_tokens=constants.MAX_TOKENS,\n temperature=0,\n )\n try:\n split_values = translate_response.choices[0].text.strip().split(\",\")\n language = str(split_values[0])\n if language.startswith(\"ur\"):\n language = \"hi\"\n english_text = split_values[1].strip()\n except Exception as e:\n print(f\"ERROR: Translated text by model:<{translate_response.choices[0].message.content.strip()}>\")\n english_text = None\n language = \"en\"\n print(f\"English_text:{english_text}\")\n return english_text, language\n\ndef translate_from_english(text) -> str:\n print(f\"Text for translation:{text}\")\n prompt = f\"\"\"Assume you are a assistant that translates English to target language. \n Target language is specified in BCP-47 format as first two characters.\n Dont translate if already in English\n Q: en;what is going on\n A: what is going on\n \n Q: te;I am a very good human\n A: నేను ఒక చాలా మంచి మానవుడు\n \n Q: {text}\n A:\n \"\"\"\n reverse_translate_response = openai.Completion.create(\n model=\"text-davinci-003\",\n prompt=prompt,\n max_tokens=constants.MAX_TOKENS,\n temperature=0,\n )\n print(f\"Translated to regional:<{reverse_translate_response.choices[0].text}>\")\n return reverse_translate_response.choices[0].text.strip()\n\n\ndef append_user_query_to_conversation(session_id: str, text: str) -> Tuple[str, str]:\n try:\n # conversation = db.get(session_id)\n conversation = db.get(session_id)\n if not conversation:\n conversation = f\"\"\"You are a helpful and knowledgeable customer service assistant. \n Below are the categories which the request can be a part of.\n {constants.JIO_CATEGORIES_AND_SAMPLE_QA} \n Send the response in json format.\n If you dont understand, send category as Others.\n \"\"\"\n conversation += f\"\\nCustomer: {text}\"\n query_conversation = conversation + \"\"\"\\nAssistant: {\"Category\": {}, \"Solution\": {}}\"\"\"\n else:\n conversation += f\"\\nCustomer: {text}\"\n query_conversation = conversation + \"\"\"\\nAssistant: {\"Category\": {}, \"Solution\": {}}\"\"\"\n # print(f\"Input to the model: {text}\")\n # print(f\"Input to the model(conversation): {conversation}\")\n model_response = index.query(query_conversation)\n print(f\"Response from the model: {model_response}\")\n model_json = json.loads(model_response)\n conversation += str(\"\\nAssistant: \") + str(model_json)\n db[session_id] = conversation\n print(conversation)\n print(\"\\n------------------\\n\")\n return model_json[\"Solution\"], model_json[\"Category\"]\n except Exception as e:\n return \"I am unable to answer this. Our executive will reach out to you within 2 hours.\", \"Others\"\n#shree\ndef classify_query(query: str) -> str:\n pass\n\n\n\n\nif __name__ == '__main__':\n app.run(host, port)\n #initialise all models\n\n","repo_name":"vishalpuri674/Multilingual-AI-Chatbot-with-Advanced-Translation","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":6005,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"23428274602","text":"import json_validate\nimport json_to_schema_validate\nimport os\n\n\ndef validate(data, schema, schema_name, data_name):\n des_data = json_validate.json_deserialize(data, file_name=data_name)\n des_schema = json_validate.json_deserialize(schema, file_name=schema_name)\n json_to_schema_validate.required_field_validate(schema=des_schema, data=des_data, schema_name=schema_name, data_name=data_name)\n json_to_schema_validate.properties_field_validate(schema=des_schema, data=des_data, schema_name=schema_name, data_name=data_name)\n\n\nfor schema_name in os.listdir(path='task_folder/schema'):\n schema_path = \"task_folder/schema/\"+schema_name\n schema_file = open(schema_path)\n schema = schema_file.read()\n for data_name in os.listdir(path='task_folder/event'):\n data_path = \"task_folder/event/\"+data_name\n data_file = open(data_path)\n data = data_file.read()\n validate(data=data, schema=schema, schema_name=schema_name, data_name=data_name)\n data_file.close()\n schema_file.close()\n\n\n","repo_name":"Argushub/TestTask","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1033,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"78"}
+{"seq_id":"37599008992","text":"import hashlib\nimport re\nimport time\n\nfrom .common import InfoExtractor\nfrom ..networking.exceptions import HTTPError\nfrom ..utils import (\n ExtractorError,\n classproperty,\n float_or_none,\n traverse_obj,\n url_or_none,\n)\n\n\nclass DacastBaseIE(InfoExtractor):\n _URL_TYPE = None\n\n @classproperty\n def _VALID_URL(cls):\n return fr'https?://iframe\\.dacast\\.com/{cls._URL_TYPE}/(?P[\\w-]+)/(?P[\\w-]+)'\n\n @classproperty\n def _EMBED_REGEX(cls):\n return [rf'