diff --git "a/2223.jsonl" "b/2223.jsonl" new file mode 100644--- /dev/null +++ "b/2223.jsonl" @@ -0,0 +1,418 @@ +{"seq_id":"74753593155","text":"#!/usr/bin/env python3 Line 1\n# -*- coding: utf-8 -*- Line 2\n#----------------------------------------------------------------------------\n# Created By : davoc\n# Created: Clase 3 ciclo 1 mision tic 2022. Parte 5 , Tarea para el 4 Mayo 2022\n# version ='2.0'\n\n#Se trata de escribir el algoritmo que permita calcular el valor a pagar para una compra de un articulo determinado \n#del que se adquieren una o varias unidades. El iva a aplicar es del 19% y si el precio bruto(precio de venta mas IVA)\n#es mayor de 500.000 COP, se aplicará un descuento del 6.5% sobre el total.\n#Se debe pedir al usuario que ingrese el valor del articulo y la cantidad.\n\nnumero_producto = int(input(\"Cuantos productos compró?: \")) #solo admite enteros de la cantidad total de articulos. #!!restringir a enteros positivos\nprecio_articulo = float(input(\"Cuanto costó el producto?: \")) #funciona con float usando . pero no con comas\n#Como puedo hacer para que reconozca el formato del usuario, ya que en unos paises\n#el punto es usado como separador decimal, mientras que la coma como separador de miles, y viceversa?\n\ntotal_sin_iva = precio_articulo * numero_producto\nprecio_bruto = precio_articulo * numero_producto * 0.19 + precio_articulo * numero_producto\ndiferencia_promo = 500000 - total_sin_iva\n\n#el precio bruto antes del + solo calcularia el iva de esos producto, entonces le sumo de nuevo las var\n#podria almacenar el total del iva en una var nueva, y en la sumarla a precio sin iva\nif precio_bruto >= 500000:\n precio_promo = precio_articulo * numero_producto * 0.065 + precio_articulo * numero_producto\n ahorro_compra = precio_bruto - precio_promo\n print(f\"El total a pagar con la promo es de: {precio_promo} , y su ahorro en esta compra es de {ahorro_compra}\")\nelse:\n print(f\"Lo sentimos, te faltó : {diferencia_promo} para aplicar la promo\")\n #print(f\"Lo sentimos, te faltó {total_sin_iva}-{tope_promo} para aplicar la promo}\") Se pueden hacer operaciones de resta en un print? \n\nprint(f\"El precio total sin iva es de : {total_sin_iva}\")\nprint(f\"El total a pagar con iva del 19% sin la promo es de : {precio_bruto}\")\n","repo_name":"labsigco/Mision51_2022","sub_path":"Ciclo1/Unidad1/Scripts/Clase3_5EjercicioMañana.py","file_name":"Clase3_5EjercicioMañana.py","file_ext":"py","file_size_in_byte":2129,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"14226018233","text":"#entering the file name to open the file\n#file name =searching from file.txt\n\n#we are going to count word from file and put them into dictionary so\n#that i count the occurence of word in file\nfname = input(\"Enter file name: \")\n#opening the file usinf open functin\nfh = open(fname)\n\n# empty dictionary\nd={}\nfor i in fh:\n #split word on the basis of spaces\n wordsList=i.split()\n #counting the element of list and addedto dictionary \n for j in wordsList:\n d[j]=d.get(j,0)+1\nprint(d)\n\n","repo_name":"harshittaneja090/mywork.github.io","sub_path":"python/file handling in python examples/beggining codes of file handling/code 13 couting word from file.py","file_name":"code 13 couting word from file.py","file_ext":"py","file_size_in_byte":500,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"21260489282","text":"import os\nimport numpy as np\n\nnumimages = 869\nsizeimages = 265 * 91\ni = 0\nmatrix = np.zeros([numimages, sizeimages * 8])\n\nwith os.scandir('C:/Users/Ext1306/Desktop/00') as entries:\n for entry in entries:\n img = np.load(entry)\n matrix[i,:] = img.flatten()\n i += 1\nprint(np.shape(matrix))\nmatrix = matrix.T\nnp.save('pca_matrix', matrix)","repo_name":"Eortvald/Foss-autoencoder","sub_path":"preprocess/PCA.py","file_name":"PCA.py","file_ext":"py","file_size_in_byte":358,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"16524026051","text":"import sys\nfrom cpyutils.parameters import CmdLineParser, Flag, Parameter, Argument, Operation\nimport cpyutils.eventloop as eventloop\nimport cpyutils.db as db\nimport cpyutils.log\nimport cpyutils.xmlrpcutils as xmlrpcutils\nimport logging\nimport version\n\n_LOGGER = cpyutils.log.Log(\"IPFLOATER\")\n\ndef main_function():\n logging.basicConfig(filename=None,level=logging.DEBUG)\n eventloop.create_eventloop(True)\n\n class IPFloaterCmdLine(CmdLineParser):\n def preops(self, result, error):\n SERVER=result.values['--server-ip'][0]\n PORT=result.values['--server-port'][0]\n self._XMLRPC_SERVER = xmlrpcutils.ServerProxy(\"http://%s:%d\" % (SERVER, PORT))\n\n def ippool(self, parse_result, error):\n try:\n _, ips = self._XMLRPC_SERVER.get_public_ips()\n return True, \"IP Pool:\\n%s\\n%s\" % (\"-\"*40, \", \".join(ips))\n except:\n return False, \"Could not contact the server\"\n\n def getip(self, parse_result, error):\n result, ep = self._XMLRPC_SERVER.create_public_redirection(\"\", -1, parse_result.values['private ip'][0], 0)\n if result:\n return True, \"Public IP obtained: %s\" % str(ep)\n else:\n return False, \"Could not obtain a redirection (server responded: %s)\" % ep\n\n def redirect(self, parse_result, error):\n result, ep = self._XMLRPC_SERVER.create_public_redirection(parse_result.values['public ip'][0], 0, parse_result.values['private ip'][0], 0)\n if result:\n return True, \"Public IP obtained: %s\" % str(ep)\n else:\n return False, \"Could not obtain a redirection (server responded: %s)\" % ep\n \n def releaseip(self, parse_result, error):\n ip = parse_result.values['public ip'][0]\n result, ep = self._XMLRPC_SERVER.clean_public_ip(ip)\n if result:\n return True, \"Released the redirection from IP %s\" % (ip)\n else:\n return False, \"Could not release the redirection from IP %s (server responded: %s)\" % (ip, ep)\n \n def status(self, result, error):\n try:\n return True, \"Table of redirections:\\n%s\\n%s\" % (\"-\"*40, self._XMLRPC_SERVER.get_redirections())\n except:\n return False, \"Could not contact the server\"\n\n def version(self, result, error):\n try:\n server_version = self._XMLRPC_SERVER.get_version()\n return True, \"Client version: %s\\nServer version: %s\" % (version.get(), server_version)\n except:\n return True, \"Client version: %s\\nCould not contact server\" % version.get()\n def arp(self, parse_result, error):\n mac = parse_result.values['mac'][0]\n result, ip = self._XMLRPC_SERVER.arp(mac)\n if result:\n return True, \"%s\" % (ip)\n else:\n return False, \"Failed to get the ip address for %s (server responded: %s)\" % (mac, ip)\n \n ap = IPFloaterCmdLine(\"ipfloater\", \"This the client for ipfloaterd, which is a server that deals with iptables to enable floating IPs in private networks\", [\n Parameter(\"--server-ip\", \"-i\", \"The ip adress in which ipfloater listens\", 1, False, [\"127.0.0.1\"]),\n Parameter(\"--server-port\", \"-p\", \"The ip port in which ipfloater listens\", 1, False, [7000]),\n Operation(\"getip\", desc = \"Requests a floating IP for a private IP\", arguments = [\n Argument(\"private ip\", \"private ip address to which is requested the floating ip\", mandatory = True, count = 1),\n ]),\n Operation(\"redirect\", desc = \"Redirects a floating IP to a private IP\", arguments = [\n Argument(\"public ip\", \"floating ip address\", mandatory = True, count = 1),\n Argument(\"private ip\", \"private ip address to which is requested the floating ip\", mandatory = True, count = 1),\n ]),\n Operation(\"releaseip\", desc = \"Releases a floating IP\", arguments = [\n Argument(\"public ip\", \"public ip address (the floating ip)\", mandatory = True, count = 1),\n ]),\n Operation(\"status\", desc = \"Gets the status of the redirections\"),\n Operation(\"version\", desc = \"Gets the version of the client and the server\"),\n Operation(\"ippool\", desc = \"Gets the public ip addresses in the pool\"),\n Operation(\"arp\", desc = \"Requests the IP for a MAC address\", arguments = [\n Argument(\"mac\", \"the mac address for which is requested the ip\", mandatory = True, count = 1),\n ]),\n ])\n \n ap.self_service(True)\n \nif __name__ == '__main__':\n main_function()","repo_name":"grycap/ipfloater","sub_path":"ipfloater.py","file_name":"ipfloater.py","file_ext":"py","file_size_in_byte":4792,"program_lang":"python","lang":"en","doc_type":"code","stars":19,"dataset":"github-code","pt":"61"} +{"seq_id":"37301548398","text":"import urllib\nimport urllib2\nimport json\n\nimport requests\nimport collections\n\n\nGET_ACCESS_TOKEN_URL = 'https://api.weixin.qq.com/cgi-bin/token'\n\nCREATE_MENU_URL = 'https://api.weixin.qq.com/cgi-bin/menu/create'\n\nDELETE_MENU_URL = 'https://api.weixin.qq.com/cgi-bin/menu/delete'\n\nGET_MENU_URL = 'https://api.weixin.qq.com/cgi-bin/menu/get'\n\nCREATE_CHN_QRCODE_URL = 'https://api.weixin.qq.com/cgi-bin/qrcode/create'\n\n\ndef post_json(url, data_map):\n req = urllib2.Request(url)\n req.add_header('Content-Type', 'application/json')\n resp = urllib2.urlopen(req, json.dumps(data_map, ensure_ascii=False))\n return resp.read()\n\n\ndef get(url, data=None, user_agent=None):\n if data:\n encode_params = urllib.urlencode(data)\n url = url + \"?\" + encode_params\n\n request = urllib2.Request(url)\n if user_agent:\n request.add_header('User-Agent', user_agent)\n\n resp = urllib2.urlopen(request)\n return resp.read()\n\n\ndef get_access_token(app_id, app_secret):\n \"\"\"\n 获取凭证\n @see http://mp.weixin.qq.com/wiki/index.php?title=%E9%80%9A%E7%94%A8%E6%8E%A5%E5%8F%A3%E6%96%87%E6%A1%A3\n \"\"\"\n data = {\n 'grant_type': 'client_credential',\n 'appid': app_id,\n 'secret': app_secret,\n }\n resp = get(GET_ACCESS_TOKEN_URL, data)\n return resp\n\n\ndef create_menu(menu_map, access_token):\n \"\"\"\n 创建菜单\n \"\"\"\n\n url = CREATE_MENU_URL + '?access_token=%s' % access_token\n # resp = urllib2.urlopen(url, menu_json_str.encode('utf-8'))\n resp = post_json(url, menu_map)\n return resp\n\n\ndef delete_menu(access_token):\n \"\"\"\n 删除菜单,建议现考虑是否直接关闭开发者模式\n \"\"\"\n url = DELETE_MENU_URL + '?access_token=%s' % access_token\n resp = post_json(url, None)\n return resp\n\n\ndef get_current_menu(access_token):\n \"\"\"\n 获取当前菜单配置\n \"\"\"\n data = {\n 'access_token': access_token,\n }\n resp = get(GET_MENU_URL, data)\n return resp\n\n\ndef create_chn_qrcode(scene_str, access_token):\n \"\"\"\n 创建渠道永久二维码, 需要订阅号才能使用\n scene_str: 渠道标识字符串\n @see http://mp.weixin.qq.com/wiki/18/28fc21e7ed87bec960651f0ce873ef8a.html\n \"\"\"\n url = CREATE_CHN_QRCODE_URL + '?access_token=%s' % access_token\n data = {\n \"action_name\": \"QR_LIMIT_STR_SCENE\",\n \"action_info\": {\n \"scene\": {\n \"scene_str\": scene_str\n }\n }\n }\n resp = post_json(url, data)\n return resp\n\n\ndef auth_url(appid, redirect_uri, state):\n url = \"https://open.weixin.qq.com/connect/oauth2/authorize\"\n data = collections.OrderedDict()\n data['appid'] = appid\n data['redirect_uri'] = redirect_uri\n data['response_type'] = 'code'\n data['scope'] = 'snsapi_userinfo'\n data['state'] = state\n ans = \"%s?%s%s\" % (url, urllib.urlencode(data), \"#wechat_redirect\")\n return ans\n\ndef get_access_token_by_code(appid, secret, code):\n url = \"https://api.weixin.qq.com/sns/oauth2/access_token\"\n params = {\n 'appid': appid,\n 'secret': secret,\n 'code': code,\n 'grant_type': 'authorization_code'\n }\n try:\n res = requests.get(url, params=params)\n res.raise_for_status()\n return res.json()\n except requests.RequestException as e:\n return {'e': e.message}\n\n\ndef get_userinfo(access_token, openid):\n url = \"https://api.weixin.qq.com/sns/userinfo\"\n params = {\n 'access_token': access_token,\n 'openid': openid,\n 'lang': 'zh_CN',\n }\n\n try:\n res = requests.get(url, params=params)\n res.raise_for_status()\n return res.json()\n except requests.RequestException as e:\n return {'e': e.message}\n","repo_name":"xym2010/wechat_lottery","sub_path":"app/wechat/wechat_api.py","file_name":"wechat_api.py","file_ext":"py","file_size_in_byte":3770,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"1345569242","text":"\nimport cv2\nimport mediapipe as mp\n\ncam = cv2.VideoCapture(0) # não reconheceu minha camera\nmodulo = mp.solutions.face_detection\nface_detection = modulo.FaceDetection()\nshow = mp.solutions.drawing_utils\n\nwhile cam.isOpened():\n x, frame = cam.read()\n if not x:\n print(\"fafa\")\n break\n\n faces = face_detection.process(frame)\n\n if faces.detections:\n for rosto in faces.detections:\n show.draw_detection(frame, rosto)\n\n cv2.imshow(\"Rostos na sua webcam\", frame)\n\n if cv2.waitKey(5) == 27:\n break\n\ncam.release()\ncv2.destroyAllWindows()","repo_name":"Marcos-VM-1708/cam_detention_","sub_path":"face_track.py","file_name":"face_track.py","file_ext":"py","file_size_in_byte":586,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"42262613171","text":"import json\nimport numpy\nfrom numpy import *\nimport scipy\nimport scipy.stats\nimport pymultinest\nimport argparse\n\nfrom interp_atm_pdf import Initialize_Atmospheric_PDFs\nfrom interp_astro_pdf import Initialize_Interpolator_Astrophysical_PDF\nfrom full_likelihood import Log_Likelihood\n\n# Recommended run:\n# python likelihood_analysis_parser.py --verbose=1 --n_live_points=200 --evidence_tolerance=0.01\n\n# After MultiNest finishes, run this to analyse the results:\n# multinest_marginals.py out/likelihood/\n\nparser = argparse.ArgumentParser(description='Likelihood analysis')\n\nparser.add_argument(\"--n_live_points\", help=\"Default: 100\",\n\ttype=int, default=100)\n\nparser.add_argument(\"--evidence_tolerance\", help=\"Default: 0.1\",\n\ttype=float, default=0.1)\n\nparser.add_argument(\"--resume\", help=\"Resume MultiNest run [default: False]\",\n\taction=\"store_true\")\n\nparser.add_argument(\"--verbose\", help=\"Default: 0\",\n\ttype=int, default=0)\n\nargs = parser.parse_args()\n\nn_live_points = args.n_live_points\nevidence_tolerance = args.evidence_tolerance\nresume = args.resume\nverbose = args.verbose\n\n\ndef Prior(cube, ndim, nparams):\n\n\t# Spectral index. Uniform prior between 1.8 and 3.\n\tcube[0] = 2.0+cube[0] #1.8+cube[0]*1.2\n\n\t# Log10 of mass of mediator [GeV]. Log uniform prior between -3.0 and -1.0.\n\tcube[1] = -3.0+2.0*cube[1]\n\n\t# Log10 of coupling constant. Log uniform prior between -3.0 and -1.0\n\tcube[2] = -3.0+2.0*cube[2]\n\n\t# Expected number of astrophysical neutrinos. Uniform distribution between 0 and 80.\n\tcube[3] = cube[3]*80\n\t# cube[1] = cube[1]*80\n\t# cube[3] = 10.**(cube[3]*(1.9084+3.0)-3.0)\n\n\t# Expected number of conv. atm. neutrinos. Uniform distribution between 0 and 80.\n\tcube[4] = cube[4]*80\n\t# cube[4] = 10.**(cube[4]*(1.9084-3.0)+3.0)\n\n\t# Expected number of prompt atm. neutrinos. Uniform distribution between 0 and 80.\n\tcube[5] = cube[5]*1\n\t# cube[5] = 10.**(cube[5]*(1.9084-3.0)+3.0)\n\n\t# Expected number of atm. muons. Uniform distribution between 0 and 80.\n\tcube[6] = cube[6]*80\n\t# cube[6] = 10.**(cube[6]*(1.9084-3.0)+3.0)\n\n\treturn 0\n\n\ndef Log_Likelihood_MultiNest(cube, ndim, nparams):\n\n\tgamma = cube[0]\n\tlog10_g = cube[1]\n\tlog10_M = cube[2]\n\tN_a = cube[3]\n\tN_conv = cube[4]\n\tN_pr = cube[5]\n\tN_mu = cube[6]\n\n\tll = Log_Likelihood(gamma, log10_g, log10_M, N_a, N_conv, N_pr, N_mu,\n interp_astro_pdf_sh, pdf_atm_conv_sh, pdf_atm_pr_sh,\n interp_astro_pdf_tr, pdf_atm_conv_tr, pdf_atm_pr_tr,\n pdf_atm_muon_tr, num_ic_sh=58, num_ic_tr=22, verbose=verbose)\n\n\treturn ll\n\n\n# Initialize the atmospheric PDFs for all of the IceCube events\npdf_atm_conv_sh, pdf_atm_pr_sh, pdf_atm_pr_sh, pdf_atm_muon_sh, \\\n pdf_atm_conv_tr, pdf_atm_pr_tr, pdf_atm_muon_tr = \\\n Initialize_Atmospheric_PDFs(verbose=verbose)\n\n# Initialize the astrophysical PDF interpolators for all of the IceCube events\ninterp_astro_pdf_sh, interp_astro_pdf_tr = \\\n Initialize_Interpolator_Astrophysical_PDF(verbose=verbose)\n\n\nparameters = [\"gamma\", \"log10_g\", \"log10_M\", \"N_a\", \"N_conv\", \"N_pr\", \"N_mu\"]\nn_params = len(parameters)\n\n# Run MultiNest\npymultinest.run(Log_Likelihood_MultiNest, Prior, n_params,\n\t outputfiles_basename='out/likelihood/',\n\t\t\t\tresume=resume, verbose=verbose, n_live_points=n_live_points,\n\t\t\t\tseed=-1, evidence_tolerance=evidence_tolerance,\n\t\t\t\tsampling_efficiency=0.8,\n\t\t\t\timportance_nested_sampling=True,\n\t\t\t\tconst_efficiency_mode=False)\n\t\t\t\t#, log_zero=-300.0)\n# const_efficiency_mode=True, sampling_efficiency=1)\n\njson.dump(parameters, open('out/likelihood/params.json', 'w')) # Save parameter names\n\n\n","repo_name":"mbustama/secret-nu-int","sub_path":"dev-likelihood-sl/likelihood_analysis_parser.py","file_name":"likelihood_analysis_parser.py","file_ext":"py","file_size_in_byte":3563,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"15445567267","text":"class Shape:\n def __init__(self, l, b, h, r):\n # self.l = None\n self.length = l\n self.breath = b\n self.height = h\n self.radios = r\n\n # def getData(self):\n # self.l = float(input(\"Enter The length: \"))\n # self.b =\n #\n","repo_name":"myproject2022/MyPythonProject","sub_path":"takeopython/ClassInherit.py","file_name":"ClassInherit.py","file_ext":"py","file_size_in_byte":274,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"39501239281","text":"class Solution:\n # @param A, a list of integers\n # @return a boolean\n\n # dp[i]: if you are able to reach the last index with i elements\n # dp[i] = (or dp[k]) for all k which A[k] + k >= i\n # initial: dp[0] = True\n # ans: dp[n-1]\n # LTE!\n def canJump(self, A):\n # write your code here\n if not A:\n return False\n dp = [False for i in xrange(len(A))]\n dp[0] = True\n for i in xrange(1, len(A)):\n for j in xrange(i):\n if dp[j] and j + A[j] >= i:\n dp[i] = True\n break\n return dp[len(A) - 1]\n\n # Greedy\n def canJump(self, A):\n if not A:\n return False\n farthest = A[0]\n for i in xrange(1, len(A)):\n # If i can be reached from beginning and\n # we can reach farther from i\n if i <= farthest and i + A[i] > farthest:\n farthest = A[i] + i\n return farthest >= len(A) - 1\n\n","repo_name":"jwyx3/practices","sub_path":"python/jump-game.py","file_name":"jump-game.py","file_ext":"py","file_size_in_byte":992,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"86341367462","text":"import numpy as np\nimport pandas as pd\n\ndf = pd.read_csv('../Brazil.csv').drop(['COUNTRY', 'PERWT'], axis=1)\ndf['constant'] = 1\ndf['female'] = (df['SEX'] == 'Female').astype(float)\ndf.drop('SEX', axis=1, inplace=True)\n\ndummies = pd.get_dummies(df[['MARST', 'NATIVITY', 'EDATTAIN', 'EMPSTAT', 'OCCISCO', 'INDGEN']], drop_first=True).astype(np.int8)\ndf = pd.concat([df[['INCTOT']], dummies, df[['female', 'AGE', 'constant']]], axis=1)\ndf.to_csv('brazil_dummies.csv', index=False)","repo_name":"chrisdfong/Gender-Pay-Gap-Analysis","sub_path":"py_files/first.py","file_name":"first.py","file_ext":"py","file_size_in_byte":477,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"61"} +{"seq_id":"21297665023","text":"import pandas as pd\nimport pyodbc\nfrom datetime import date\nfrom datetime import datetime\nfrom datetime import timedelta\nimport requests as rs\nimport sqltdb as sqdb\nimport sitecount as st\nimport omfn as fn\nimport urllib3\nimport urllib.parse\n\n\ntday = date.today()\ntmdlta = datetime.now() + timedelta(minutes=1)\ntmnw = datetime.now() - timedelta(minutes=1)\nqryst = tmnw.strftime('%Y-%m-%d %H:%M:%S')\nqryend = tmdlta.strftime('%Y-%m-%d %H:%M:%S')\n\n\ndef handdler(ussd,msg,msisdn):\n nw = datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n rval = \"\"\n if msg !=\"\":\n ms = msisdn[-10:len(msisdn)]\n ms4sms = msisdn[-11:len(msisdn)]\n code = fn.sitecode_pick(msg)\n if \"ALL\" in msg or '2G' in msg or \"3G\" in msg or \"SC\" in msg or \"4G\" in msg:\n xx = st.siteinfo(msg)\n print(nw, xx)\n yy = st.sms(ms4sms,xx)\n rval = \"S\"\n elif \"PGSTART\" in msg and code != 'NA':\n xx = st.roc(ussd,code,ms,'PGSTART')\n print(nw,xx)\n if 'PGON_DONE' in xx:\n rval = \"S\"\n else:\n rval = \"F\"\n elif \"PGSTOP\" in msg and code != 'NA':\n xx = st.roc(ussd,code,ms,'PGSTOP')\n print(nw,xx)\n if 'PGOFF_DONE' in xx:\n rval = \"S\"\n else:\n rval = \"F\"\n else:\n rval = \"Not Related Query\"\n return rval\n\ndef main():\n nww = datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n df = st.smscheck()\n if df.shape[0] != 0:\n for i in range(len(df)):\n msg1 = df.loc[i,\"MESSAGE\"]\n if isinstance(msg1, str):\n ussd = df.loc[i,\"USDLogId\"]\n msg = msg1.upper()\n msisdn = df.loc[i,\"DESTADDR\"]\n sqret = sqdb.queryussd(ussd)\n if sqret == 0:\n st.general_qry()\n rval = handdler(ussd,msg,msisdn)\n st.general_qry()\n if rval == 'S':\n rv2 = sqdb.insertussd(ussd)\n if rv2 == \"S\":\n print('Cycle Complete for::::: ', nww, ussd, msg, msisdn)\n else:\n print(\"Cycle failed:::\", nww, ussd, msg, msisdn)\n else:\n print(rval)\n else:\n print('already served::', nww, ussd, msg, msisdn)\n else:\n print('no sms')\n return \"done at \" + nww","repo_name":"FuckBrains/omEngin","sub_path":"Z_ALL_FILE/Py1/10262020-221-XAQ-main.py","file_name":"10262020-221-XAQ-main.py","file_ext":"py","file_size_in_byte":2494,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"72287262915","text":"#!/usr/bin/env python3\nimport tweepy\nfrom sense_hat import SenseHat\nfrom datetime import datetime\nsense = SenseHat()\nclient = tweepy.Client(consumer_key='YOUR API KEY',\n consumer_secret='YOUR API KEY SECRET',\n access_token='YOUR ACCESS TOKEN',\n access_token_secret='YOUR ACCESS TOKEN SECRET')\n\ntemp = round(sense.get_temperature(),1)\nnow = datetime.now()\ncurrent_time = now.strftime(\"%H:%M\")\nfunny = [\"Right now the temperature at my desk is \", \"Oh my word it is cold at my desk, it is only \", \"Feeling hot, hot, hot, the temperature is \"]\nif temp < 22.0:\n msg = current_time+\" \"+[1]+str(temp)+\" Celsius\"\nelif temp > 23.0:\n msg = current_time+\" \"+funny[2]+str(temp)+\" Celsius\"\nelse:\n msg = current_time+\" \"+funny[0]+str(temp)+\" Celsius\"\n\nresponse = client.create_tweet(text=msg)\nprint(response)\nprint(msg)\nsense.show_message(str(temp)+\"C\")\n","repo_name":"lesp/LXF-Twitter-Sense","sub_path":"send_tweet.py","file_name":"send_tweet.py","file_ext":"py","file_size_in_byte":915,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23641012851","text":"#!/usr/bin/env python\nimport sys\nimport re\nimport urllib\nimport math\nimport os\nfrom timeit import Timer\n\n#def subset = lambda x: [[y for j, y in enumerate(set(x)) if (i >> j) & 1] for i in range(2**len(set(x)))]\n\ndef output(case, answer):\n\treturn \"Case #%d: %d\\n\" % (case, answer)\n\ndef func(n):\n\tif n == 0: return 0\n\tif n == 1 or n == 2: return 1\n\tif n % 3 == 0: return n/3\n\telse: return int(math.floor(float(n)/3)) + 1\n\ndef determine(l,s,p):\n\ttotal = 0\n\tmaxnorm = [func(val) for val in l]\n\tcandidates = len([val for index, val in enumerate(maxnorm) if val == p-1 and l[index] % 3 != 1 and l[index] > 1])\n\twhile candidates and s:\n\t\ttotal += 1\n\t\tcandidates -= 1\n\t\ts -= 1\n\ttotal += len([val for val in maxnorm if val >= p])\n\treturn total\n\t\ndef main(filein, fileout):\n\tcase = 0\n\tf = open(filein, 'r')\n\to = open(fileout, 'w')\n\ttimes = int(f.readline())\n\twhile case < times:\n\t\tanswer = 0\n\t\tvalues = f.readline().split(\" \")\n\t\tnumPlayers = int(values[0])\n\t\tsurprising = int(values[1])\n\t\tlimit = int(values[2])\n\t\tvalues = [int(number) for number in values[-numPlayers:]]\n\t\tanswer = determine(values,surprising,limit)\n\t\tcase += 1\n\t\to.write(output(case, answer))\n\tf.close()\n\to.close()\n\nif __name__ == \"__main__\":\n\tmain(sys.argv[1], sys.argv[1][:-2]+'out')\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_96/596.py","file_name":"596.py","file_ext":"py","file_size_in_byte":1246,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"1373607119","text":"import pygame # importing pygame library\nimport random\nfrom sys import exit # Importing 'sys' important for various functions\n\npygame.init() \n\nscreen = pygame.display.set_mode((400, 400)) #showing the screen\nstarted = False #declaring a variable to show the state of game\n\nclass Player: \n speed = 0.5 #speed of the spaceship on press of button\n #initial x and y position of the player\n player_x = 170 #170 to keep it in center of screen as our spaceship is 60px wide\n player_y = 350 #setting y coordinate of spaceship\n player_x_change = 0\n\n #adding the spaceship image in code\n spaceship = pygame.image.load(\n 'C:\\Projects\\Space Invaders\\space-invaders\\Elements\\player.png')\n \n def player(x, y):\n screen.blit(Player.spaceship, (x, y)) #showing the image on screen\n\n def player_movement():\n Player.player_x += Player.player_x_change #this will increase or decrease (depending when the function is called) the x coordinate of position by continuosly adding 0.5 (speed variable) to it.\n\n def player_boundary():\n if Player.player_x < 0: # adding boundary to the game\n Player.player_x = 0 # adding boundary to the game\n elif Player.player_x >= 340: # adding boundary to the game\n Player.player_x = 340 # adding boundary to the game\n \n def line():\n rect1 = ((0 ,330 , 400, 200)) #making a rectangle the endline of the screen (0,330) : coordinates of top left corner of rectangle\n #(400,200) size of rectangle\n pygame.draw.rect(screen, (40,40,40), rect1) #dsiplaying the rectangle on 'screen' and color is (40,40,40)\n\nclass Enemy:\n\n enemy_speed = 0.2 #speed at which enemy moves\n enemy_fasten = 0.2 #speed at which enemy moves when it hit the wall\n\n enemy_red = [] #list to store enemy_images as there are multiple\n\n enemy_x = [] #list to store the x coordinates of enemy_image corresponding to the image stored in enemy_red list\n enemy_y = [] #list to store the y coordinates of enemy_image corresponding to the image stored in enemy_red list\n\n enemy_x_change = [] #list to store the change in x coordinates of enemy_image corresponding to the image stored in enemy_red list\n enemy_y_change = [] #list to store the change in y coordinates of enemy_image corresponding to the image stored in enemy_red list\n\n number_of_enemies = 4 #number of enemies\n\n #making/adding the enemies in game\n for enemies in range(number_of_enemies):\n #adding the enemies to the list enemy_red\n enemy_red.append(pygame.image.load(\n 'C:\\Projects\\Space Invaders\\space-invaders\\Elements\\enemy_red.png')) \n # locates the enemy on a random point between 0, 360\n enemy_x.append(random.randint(0, 360))\n # locates the enemy on a random point between 32, 64\n enemy_y.append(random.randint(32, 64))\n #adding enemy to the corresponding list so that each sprite moves individually at its own speed\n enemy_x_change.append(enemy_speed)\n enemy_y_change.append(32) #change in y coordinate, it is 32 because the heigth of enemy is 32px. As we want the enemy to move to next row on hitting wall, we kept the change to 32\n\n\n def enemy1(x,y,i):\n #showing enemies on screen\n #here x and y are the variable coordinates\n #and i is a variable which will be used in a future loop \n screen.blit(Enemy.enemy_red[i], (x,y))\n\nclass Laser:\n laser_x = Player.player_x + 30 #spaceship initial posn + 30\n laser_y = Player.player_y + 5 #spaceship initial posn + 5\n laser_x_change = 0\n laser_y_change = 1\n laser_state = 'rest'\n #addiing and resizing laser image \n laser = pygame.image.load(\n 'C:\\Projects\\Space Invaders\\space-invaders\\Elements\\laser_bullet.jpg')\n laser = pygame.transform.scale(laser, (4, 20))\n\n def laser_fire(x, y):\n global laser_state #this is same laser state shown in laser.py\n laser_state = 'fired' #changing the state, will change the value to 'fired' when this function is called\n screen.blit(Laser.laser, (x, y)) #displaying image on screen when function is called\n\n def laserstate():\n #when this function is called\n if Laser.laser_state == 'fired' : #it will check the state of laser, if it is fired then\n #it will display the laser and change its coordinate by 'laser_y_change' value, so that it appears that laser is moving upwards. \n Laser.laser_fire(Laser.laser_x + 26, Laser.laser_y) \n Laser.laser_y -= Laser.laser_y_change\n\n def laser_boundary():\n #this function will reset the state of the laser once it is out of window, so that we can shoot the laser again\n if Laser.laser_y <= 0 : #if the coordinate of laser goes out of screen then\n Laser.laser_y = Player.player_y #reset the y coordinate of the laser to the y coordinate of player (spaceship)\n Laser.laser_state = 'rest' #and change the laser state to rest so that we can fire it again\n\nclass Controls:\n def player_control():\n #this function has the controls of player\n if pygame.key.get_pressed()[pygame.K_LEFT] or pygame.key.get_pressed()[pygame.K_a]:\n Player.player_x_change -= Player.speed #will decrease the x coordinate by 'speed' (0.5px) so that it can move to left\n if pygame.key.get_pressed()[pygame.K_RIGHT] or pygame.key.get_pressed()[pygame.K_d]:\n Player.player_x_change += Player.speed #will increase the x coordinate by 'speed' (0.5px) so that it can move to right\n #laser will fire when upper arrow is pressed\n if pygame.key.get_pressed()[pygame.K_UP] or pygame.key.get_pressed()[pygame.K_w]:\n Laser.laser_x = Player.player_x\n if Laser.laser_state == 'rest': #checking if laser state is in rest state \n #this condition above is important becase, without it if we accidentally clicked the up_arrow, the position of laser will reset to spaceship position\n #and it will refire from the spaceship position\n Laser.laser_state = 'fired' #so that it can be changes to fire state\n Laser.laser_fire(Laser.laser_x, Laser.laser_y) #and can be fired from the position of the spaceship (laser_x and laser_y are equals to spaceship position)\n\nclass Score:\n score = 0 #keeping the track of score which will increase upon collision\n #declaring the font used. \n font = pygame.font.Font('C:\\Projects\\Space Invaders\\space-invaders\\Fonts\\VCR_OSD_MONO_1.001.ttf', 16)\n\n def show_score(x, y): #to display the score count using blit()\n score_count = Score.font.render('Score : ' + str(Score.score), True, (255,255,255))\n screen.blit(score_count, (x,y))\n\n def show_title(x, y): #to show 'space invaders title' using (blit)\n score_count = Score.font.render('Space Invaders', True, (255,255,255))\n screen.blit(score_count, (x,y)) \n\nclass StartScreen:\n start = True #variable to store the state of start window, if its true that means start window is displayed on screen. If false, then its not.\n font = pygame.font.Font(r'Fonts\\upheavtt.ttf', 56) #declaring the font\n\n def start_screen(): \n #adding and displaying (blit function) the start image on screen.\n start_image = pygame.image.load(\"C:\\Projects\\Space Invaders\\space-invaders\\Elements\\Space Invaders Start.png\")\n start_image = pygame.transform.scale(start_image, (400,400)) #resizing the screen so that it fits the window\n\n screen.blit(start_image, (0,0))\n \n def show_start():\n while StartScreen.start: #while the start variable in StartScreen class is True (it becomes true when the function is called)\n for event in pygame.event.get(): \n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n if event.type == pygame.KEYDOWN :#checks if SPACE key is pressed\n if pygame.key.get_pressed()[pygame.K_SPACE]:\n StartScreen.start = False #sets the start variable to false, hence start screen is closed \n \n StartScreen.start_screen() #calling the start_screen function to display the intro image\n pygame.display.update()\n\nclass Collision:\n def collision():\n laser_rect = Laser.laser.get_rect(center=(Laser.laser_x, Laser.laser_y)) #getting rectangle of size and width of laser image, and mapping it to position of laser\n enemy_rect = [] #making an empty list to store the rectangles we will get of the enemy images stored in enemy_rect list\n\n for i in range(Enemy.number_of_enemies): #iterating over enemies\n Enemy.enemy_x[i] += Enemy.enemy_x_change[i] #this will constantly change the enemy's (at index 'i' in the enemy_rect list) x coordinate by enemy_x_change\n Enemy.enemy_y[i] = Enemy.enemy_y_change[i] #this will constantly change the enemy's (at index 'i' in the enemy_rect list) y coordinate by enemy_x_change\n \n #this is enemy boundary, if enemy at say index 'i' hits the wall ie goes outside the window it will be reset to a new coordinate as per the code\n if Enemy.enemy_x[i] < 0: #if enemy at index 'i' hits the left wall\n Enemy.enemy_x_change[i] = Enemy.enemy_fasten #change the speed of enemy and hence its coordinates by logic mentioned earlier\n Enemy.enemy_y_change[i] += 32 #to change the y position of alien, to make it come one row down if it touches walls\n if Enemy.enemy_x[i] >= 340: #if enemy at index 'i' hits the right wall\n Enemy.enemy_x_change[i] = -Enemy.enemy_fasten #change the speed of enemy and hence its coordinates by logic mentioned earlier\n Enemy.enemy_y_change[i] += 32 #to change the y position of alien, to make it come one row down if it touches walls\n \n if Enemy.enemy_y[i] >= 300 : #if player hits the y boundary that is 300 \n GameOver.show_game_over() #game over screen will be displayed\n break\n \n Enemy.enemy1(Enemy.enemy_x[i], Enemy.enemy_y[i], i) #showing the enemies on screen\n\n #now we will add the rectangles of enemy at index i to the enemy_rect list.\n enemy_rect.append(Enemy.enemy_red[i].get_rect(center=(Enemy.enemy_x[i], Enemy.enemy_y[i]))) \n\n if laser_rect.colliderect(enemy_rect[i]): #colliderect is an inbuilt function which will return True if the two given rectangles collide each other\n print(Score.score)\n Enemy.enemy_x[i] = 10 #set the enemy's x coordinate to 10\n Enemy.enemy_y[i] = 60 #set the enemy's y coordinate to 60\n Enemy.enemy_y_change[i] = random.randint(32, 150) #set the enemy_y_change to any random int between 32 and 150 so that the y coordinate of that specific enemy is changed.\n Score.score += 1 #increase the score by 1 \n Laser.laser_state = 'rest' #change the state of laser to rest so that it can be fired again\n Laser.laser_y = Player.player_y #change the y coordinate of laser to player's y coordinate. \n \n Laser.laser_fire(Player.player_x + 300, Player.player_y + 300) #make the laser dissapear once it hits the spaceship\n\nclass GameScreen :\n def gameplay():\n\n Player.player_movement() #adding the movement logic of player\n Player.player_boundary() #adding the boundary by calling the Boundary() function from Player Class\n\n Laser.laserstate() #checking the state of the laser by calling the laserstate() function from Laser Class\n Laser.laser_boundary() #adding the condition to check if laser is outside the window or not by calling laser_boundary() function from Laser Class\n \n Collision.collision() #checking for collision of enemy and bullet calling collision() function from Collision class\n\n Player.line() #displaying the end line by calling line function from Player class\n Player.player(Player.player_x, Player.player_y) #displaying spaceship by calling player function from Player class\n\n Score.show_score(10,10) #displaying the score by calling show_score function from Score class. (10,10) is the position.\n Score.show_title(260,10) #displaying the title by calling show_title func from Score Class. (260, 10) is the position.\n\nclass Paused :\n pause = True\n\n def paused():\n paused = GameOver.font.render('PAUSED', True, (255,255,255))\n paused_rect = paused.get_rect(center=(400/2, 400/2))\n end_score = Score.font.render('score : '+ str(Score.score), True, (255,255,255))\n rect1 = pygame.Rect((0 ,0 , 400, 400))\n pygame.draw.rect(screen, (40,40,40), rect1)\n screen.blit(paused, paused_rect)\n screen.blit(end_score, (150,230))\n \n while Paused.pause:\n for event in pygame.event.get():\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n if event.type == pygame.KEYDOWN: # whem key is pressed\n if pygame.key.get_pressed()[pygame.K_SPACE]:\n Paused.pause = False\n GameScreen.gameplay()\n\n pygame.display.update()\n\n def pause_control():\n\n if pygame.key.get_pressed()[pygame.K_SPACE]:\n Paused.pause = True\n Paused.paused()\n\nclass GameOver:\n font = pygame.font.Font('C:\\Projects\\Space Invaders\\space-invaders\\Fonts\\VCR_OSD_MONO_1.001.ttf', 40)\n stopped = True\n\n def gameover():\n game_over = GameOver.font.render('GAME OVER', True, (255,255,255))\n game_over_rect = game_over.get_rect(center=(400/2, 400/2))\n end_score = Score.font.render('score : '+ str(Score.score), True, (255,255,255))\n rect1 = pygame.Rect((0 ,0 , 400, 400))\n pygame.draw.rect(screen, (40,40,40), rect1)\n screen.blit(game_over,game_over_rect)\n screen.blit(end_score, (150,230))\n \n def show_game_over():\n while GameOver.stopped:\n for event in pygame.event.get():\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit() \n \n GameOver.gameover()\n pygame.display.update()\n\ndef small_text(text) :\n Score.font.render(text, True, (255,255,255))\n\ndef main():\n clock = pygame.time.Clock()\n while True:\n \n screen.fill((30, 30, 30)) # Set the Backgroud Color\n\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n exit() \n if event.type == pygame.KEYDOWN: # whem key is pressed\n Controls.player_control()\n Paused.pause_control() \n if event.type == pygame.KEYUP:\n Player.player_x_change = 0\n\n \n StartScreen.show_start()\n GameScreen.gameplay()\n clock.tick(3000)\n pygame.display.update()\n\nmain()","repo_name":"ayushxpatne/space-invaders","sub_path":"Code/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":15161,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"34321504525","text":"import dht\nimport network\nimport time\nfrom config import *\nfrom machine import Pin\nfrom micropyserver import MicroPyServer\ngc.collect()\n\n\nrelay1 = Pin(0, Pin.OUT)\nrelay2 = Pin(13, Pin.OUT)\nrelay3 = Pin(12, Pin.OUT)\nrelay4 = Pin(4, Pin.OUT)\n\nrelay1.on()\nrelay2.on()\nrelay3.on()\nrelay4.on()\n\nblueled = Pin(2, Pin.OUT)\nblueled.on()\n\nadc = machine.ADC(0)\nm_vin = Pin(5, Pin.OUT)\nm_vin.off()\n\nwlan_id = ssid\nwlan_pass = password\nwlan = network.WLAN(network.STA_IF)\nwlan.active(True)\nserver = MicroPyServer()\n\n# if wlan.isconnected() == False:\n # wlan.connect(wlan_id, wlan_pass)\n # while wlan.isconnected() == False:\n # time.sleep(1)\n \nwhile wlan.isconnected() == False:\n print(\"trying hui\")\n wlan.connect(ssid2, password2)\n time.sleep(15)\n if wlan.isconnected() == False:\n print(\"trying random\")\n wlan.connect(ssid, password)\n time.sleep(15)\n\nprint('IP by DHCP:', wlan.ifconfig()[0])\n\nvarVolt = 4.1339\nvarProcess = 0.05\nPc = 0.0\nG = 0.0\nP = 1.0\nXp = 0.0\nZp = 0.0\nXe = 0.0\n\ndef relay_state(n):\n if n == 1:\n if relay1.value() == 1:\n rez = 0\n elif relay1.value() == 0:\n rez = 1\n elif n == 2:\n if relay2.value() == 1:\n rez = 0\n elif relay2.value() == 0:\n rez = 1\n elif n == 3:\n if relay3.value() == 1:\n rez = 0\n elif relay3.value() == 0:\n rez = 1\n elif n == 4:\n if relay4.value() == 1:\n rez = 0\n elif relay4.value() == 0:\n rez = 1\n return rez\n\ndef kalman(var):\n global varVolt\n global varProcess\n global Pc\n global G\n global P\n global Xp\n global Zp\n global Xe\n Pc = P + varProcess\n G = Pc / (Pc + varVolt)\n P = (1 - G) * Pc\n Xp = Xe\n Zp = Xp\n Xe = G * (var - Zp) + Xp # \"фильтрованное\" значение\n return Xe\n\ndef show_data(request):\n blueled.on()\n d = dht.DHT11(Pin(14))\n d.measure()\n hum = round(kalman(d.humidity()))\n server.send(str(d.temperature()) + \",\" + str(hum))\n blueled.off()\n\ndef show_moisture(request):\n blueled.on()\n m_vin.on()\n s = str(adc.read())\n time.sleep(1)\n m_vin.off()\n server.send(s);\n blueled.off()\n\ndef reboot(request):\n machine.reset()\n\ndef relay1_on(request):\n relay1.off()\n relay1_status(request)\n\ndef relay1_off(request):\n relay1.on()\n relay1_status(request)\n\ndef relay2_on(request):\n relay2.off()\n relay2_status(request)\n\ndef relay2_off(request):\n relay2.on()\n relay2_status(request)\n\ndef relay3_on(request):\n relay3.off()\n relay3_status(request)\n\ndef relay3_off(request):\n relay3.on()\n relay3_status(request)\n\ndef relay4_on(request):\n relay4.off()\n relay4_status(request)\n\ndef relay4_off(request):\n relay4.on()\n relay4_status(request)\n\ndef relay1_status(request):\n server.send(str(relay_state(1)))\n\ndef relay2_status(request):\n server.send(str(relay_state(2)))\n\ndef relay3_status(request):\n server.send(str(relay_state(3)))\n\ndef relay4_status(request):\n server.send(str(relay_state(4)))\n\n''' add request handler '''\nserver.add_route(\"/data\", show_data)\nserver.add_route(\"/moisture\", show_moisture)\nserver.add_route(\"/reboot\", reboot)\nserver.add_route(\"/relay1_on\", relay1_on)\nserver.add_route(\"/relay1_off\", relay1_off)\nserver.add_route(\"/relay2_on\", relay2_on)\nserver.add_route(\"/relay2_off\", relay2_off)\nserver.add_route(\"/relay3_on\", relay3_on)\nserver.add_route(\"/relay3_off\", relay3_off)\nserver.add_route(\"/relay4_on\", relay4_on)\nserver.add_route(\"/relay4_off\", relay4_off)\nserver.add_route(\"/relay1_status\", relay1_status)\nserver.add_route(\"/relay2_status\", relay2_status)\nserver.add_route(\"/relay3_status\", relay3_status)\nserver.add_route(\"/relay4_status\", relay4_status)\n\nprint (\"starting http server\")\n''' start server '''\nserver.start()\n\nblueled.off()\n","repo_name":"makeinstall77/strawberry_monitoring","sub_path":"app/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3861,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23502432121","text":"class PancakeStack:\n def __init__(self, stack = []):\n s = []\n for p in stack:\n if p == '+':\n s.append(1)\n else:\n s.append(-1)\n self.stack = s\n def __str__(self):\n return str(self.stack)\n def flip(self, n):\n self.stack[:n] = [p*-1 for p in self.stack[:n]][::-1]\n def is_good(self):\n for p in self.stack:\n if p == -1:\n return False\n return True\n def solve(self, acc):\n if self.is_good():\n return acc\n if self.stack == []:\n return acc\n if self.stack[-1] == 1:\n s = PancakeStack()\n s.stack = self.stack[:-1]\n return s.solve(acc)\n elif self.stack[0] == -1:\n self.flip(len(self.stack))\n return self.solve(acc + 1)\n else:\n count = 0\n for p in self.stack:\n if p == 1:\n count += 1\n else:\n break\n self.flip(count)\n return self.solve(acc + 1)\n\ndef read_data(filename):\n with open(filename) as f:\n num_test_cases = int(f.readline())\n test_cases = []\n for _ in range(num_test_cases):\n test_case = PancakeStack(f.readline().strip())\n test_cases.append(test_case)\n return num_test_cases, test_cases\n\nif __name__ == \"__main__\":\n num_test_cases, test_cases = read_data(\"input.in\")\n for it in range(num_test_cases):\n test_case = test_cases[it]\n print(\"Case #{}:\".format(it + 1), test_case.solve(0))\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_178/795.py","file_name":"795.py","file_ext":"py","file_size_in_byte":1616,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23564842811","text":"def last_number(x):\n i = 0\n while i + 1 < len(x):\n if x[i] > x[i + 1]:\n x[i] -= 1\n for j in range(i + 1, len(x)):\n x[j] = 9\n return x\n elif x[i] == x[i + 1]:\n j = i + 1\n fl = False\n while j + 1 < len(x):\n if x[j] < x[j + 1]:\n i = j + 1\n fl = True\n break\n elif x[j] > x[j + 1]:\n x[i] -= 1\n for k in range(i + 1, len(x)):\n x[k] = 9\n return x\n else:\n j += 1\n if fl == False:\n return x\n else:\n i += 1\n return x\n\n\ndef str2list(x):\n return [int(xx) for xx in x if xx != '\\n']\n\ndef list2str(x):\n strconv = ''.join([str(xx) for xx in x])\n if strconv[0] == '0' and len(strconv) > 1:\n return strconv[1:]\n else:\n return strconv\n\n\nif __name__ == '__main__':\n responses = []\n inf = 'tidy2.in'\n outf = 'tidy2.out'\n with open(inf, 'r') as f:\n cases = int(f.readline())\n for i in range(cases):\n numb = f.readline()\n x = str2list(numb)\n res = last_number(x)\n responses.append(list2str(res))\n with open(outf, 'w') as f:\n for i, r in enumerate(responses):\n f.write('Case #{}: {}\\n'.format(i + 1, r))\n\n #\n # x = '20'\n # x = str2list(x)\n # x = last_number(x)\n # print list2str(x)","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_200/601.py","file_name":"601.py","file_ext":"py","file_size_in_byte":1539,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"43009241375","text":"import os\nimport hashlib\nimport subprocess\nimport requests\nimport threading\nimport time\nimport signal\n\nfrom watchfiles import watch, Change\n\nimport psycopg2\nimport psycopg2.extras\n\ndef log(msg: str):\n print(msg, flush=True)\n\nclass ImmichDatabase:\n def __init__(self, host: str, database: str, user: str, password: str, port: int):\n self.conn = psycopg2.connect(host=host, database=database, user=user, password=password, port=port)\n self.conn.set_client_encoding('UTF8')\n\n def last_removed_asset(self, user_id: str) -> list[psycopg2.extras.RealDictRow]:\n with self.conn.cursor(cursor_factory=psycopg2.extras.RealDictCursor) as cur:\n cur.execute(\"\"\"\n SELECT\n assets_filesync_lookup.asset_path,\n assets_delete_audits.asset_id\n FROM assets_delete_audits\n INNER JOIN assets_filesync_lookup\n ON assets_delete_audits.checksum = assets_filesync_lookup.checksum\n AND assets_delete_audits.user_id = assets_filesync_lookup.user_id\n WHERE assets_filesync_lookup.user_id = %s\n AND assets_delete_audits.file_removed = 'false'\n ORDER BY changed_on desc\n LIMIT 1\n \"\"\", (user_id,))\n\n return cur.fetchall()\n\n def set_asset_removed(self, asset_id: str) -> None:\n with self.conn.cursor() as cur:\n cur.execute(\"\"\"\n UPDATE assets_delete_audits\n SET file_removed = 'true'\n WHERE asset_id = %s\n \"\"\", (asset_id,))\n self.conn.commit()\n\n def save_hash(self, user_id: str, asset_path: str, checksum: bytes) -> None:\n with self.conn.cursor() as cur:\n cur.execute(\"\"\"\n INSERT INTO\n assets_filesync_lookup(user_id, asset_path, checksum)\n VALUES(%s, %s, %s)\n ON CONFLICT (user_id, asset_path) DO\n UPDATE SET checksum = %s\n WHERE assets_filesync_lookup.asset_path = %s\n AND assets_filesync_lookup.user_id = %s;\n \"\"\",\n (user_id, asset_path, checksum,\n checksum, asset_path, user_id))\n self.conn.commit()\n\n def get_asset_id_by_path(self, user_id: str, asset_path: str) -> psycopg2.extras.RealDictRow | None:\n with self.conn.cursor(cursor_factory=psycopg2.extras.RealDictCursor) as cur:\n cur.execute(\"\"\"\n SELECT assets.id\n FROM assets\n INNER JOIN assets_filesync_lookup\n ON assets.checksum = assets_filesync_lookup.checksum\n WHERE assets_filesync_lookup.asset_path = %s\n AND assets_filesync_lookup.user_id = %s\n \"\"\", (asset_path, user_id))\n return cur.fetchone()\n\n def close(self):\n self.conn.commit()\n self.conn.close()\n\nclass ImmichAPI:\n def __init__(self, host: str, api_key: str):\n self.host = host\n self.headers = {\n \"Accept\": \"application/json\",\n \"Content-Type\": \"application/json\",\n \"x-api-key\": api_key\n }\n\n def get_user_id(self) -> str:\n r = requests.get(f\"{self.host}/user/me\", headers=self.headers)\n return r.json()[\"id\"]\n \n def delete_asset(self, asset_id: str) -> dict:\n data = { \"ids\": [ asset_id ] }\n r = requests.delete(f\"{self.host}/asset\", headers=self.headers, json=data)\n return r.json()\n\ndef hash_file(path: str) -> bytes:\n file_hash = hashlib.sha1()\n with open(path, \"rb\") as f:\n fb = f.read(2048)\n while len(fb) > 0:\n file_hash.update(fb)\n fb = f.read(2048)\n return file_hash.digest()\n\ndef ignored_paths(path: str) -> bool:\n if os.path.basename(path).startswith(\".\"):\n return True\n\n if os.path.isdir(path):\n return True\n \n return False\n\ndef hash_all_files(db: ImmichDatabase, user_id: str, path: str) -> None:\n for root, _, files in os.walk(path):\n for file in files:\n if ignored_paths(file):\n continue\n\n file_path = os.path.join(root, file)\n relative_path = os.path.relpath(file_path, path)\n db.save_hash(user_id, relative_path, hash_file(file_path))\n log(f\"Hash {file_path} and store in database\")\n\ndef import_asset(db: ImmichDatabase, api: ImmichAPI, key: str, base_path: str, asset_path: str) -> None:\n snap_path = os.getenv(\"SNAP\")\n relative_path = os.path.relpath(asset_path, base_path)\n import_command = [\n f\"{snap_path}/bin/immich-cli\", \"upload\",\n \"--server\", os.getenv(\"IMMICH_SERVER_ADDRESS\"),\n \"--key\", key,\n \"--yes\",\n asset_path\n ]\n\n if snap_path:\n result = subprocess.run(import_command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n else:\n result = subprocess.CompletedProcess([], 0)\n log(f\"MOC: {import_command}\")\n\n if result and result.returncode != 0:\n log(f\"Error: Failed to import {asset_path}\")\n log(f\"CLI (stdout): {result.stdout.decode('utf-8')}\")\n log(f\"CLI (stderr): {result.stderr.decode('utf-8')}\")\n else:\n checksum = hash_file(asset_path)\n user_id = api.get_user_id()\n db.save_hash(user_id, relative_path, checksum)\n log(f\"Hash {relative_path} and store in database for user {user_id})\")\n\ndef delete_asset(db: ImmichDatabase, api: ImmichAPI, asset_path: str, base_path: str) -> None:\n relative_path = os.path.relpath(asset_path, base_path)\n user_id = api.get_user_id()\n asset = db.get_asset_id_by_path(user_id, relative_path)\n if asset:\n log(f\"Asset {asset['id']} removed from database\")\n api.delete_asset(asset[\"id\"])\n else:\n log(f\"Asset {relative_path} not found in database\")\n\ndef file_watcher(event: threading.Event, db: ImmichDatabase, api: ImmichAPI, api_key: str, user_path: str) -> None:\n log(\"File watcher thread running...\")\n for changes in watch(user_path, recursive=True, stop_event=event):\n for c_type, c_path in changes:\n\n if ignored_paths(c_path):\n continue\n\n if c_type == Change.added:\n log(f\"{c_path} added, import asset to Immich\")\n import_asset(db, api, api_key, user_path, c_path)\n elif c_type == Change.modified:\n log(f\"{c_path} modified, re-import asset to Immich\")\n import_asset(db, api, api_key, user_path, c_path)\n elif c_type == Change.deleted:\n log(f\"{c_path} deleted, mark asset as removed\")\n delete_asset(db, api, c_path, user_path)\n\ndef database_watcher(event: threading.Event, db: ImmichDatabase, api: ImmichAPI, user_path: str) -> None:\n log(\"Database watcher thread running...\")\n user_id = api.get_user_id()\n while not event.is_set():\n for record in db.last_removed_asset(user_id):\n asset_id = record['asset_id']\n asset_path = record['asset_path']\n full_path = f\"{user_path}/{asset_path}\"\n if os.path.exists(full_path):\n log(f\"Remove asset {asset_id} user {user_id} path {asset_path}\")\n os.remove(full_path)\n else:\n log(f\"Asset {asset_id} user {user_id} path {asset_path} already removed\")\n log(f\"Mark asset {asset_id} as removed\")\n db.set_asset_removed(asset_id)\n time.sleep(5)\n\ndef main():\n db = ImmichDatabase(\n host=os.environ[\"DB_HOSTNAME\"],\n database=os.environ[\"DB_DATABASE_NAME\"],\n user=os.environ[\"DB_USERNAME\"],\n password=os.environ[\"DB_PASSWORD\"],\n port=5432\n )\n\n api_key = os.environ[\"IMMICH_API_KEY\"]\n immich = ImmichAPI(os.environ[\"IMMICH_SERVER_URL\"], api_key)\n snap_common = os.environ[\"SNAP_COMMON\"]\n user_id = immich.get_user_id()\n user_path = f\"{snap_common}/sync/{user_id}\"\n\n log(f\"Starting sync for user {user_id} at {user_path}\")\n\n log(f\"Initial file hash import of all files in {user_path}\")\n hash_all_files(db, user_id, user_path)\n\n stop_event = threading.Event()\n\n watch_thread = threading.Thread(\n target=file_watcher,\n args=(stop_event, db, immich, api_key, user_path)\n )\n\n database_thread = threading.Thread(\n target=database_watcher,\n args=(stop_event, db, immich, user_path)\n )\n\n watch_thread.start()\n database_thread.start()\n\n signal.signal(signal.SIGTERM, lambda signum, frame: stop_event.set())\n\n while True:\n if not watch_thread.is_alive():\n log(\"Critical: Thread watch is not alive\")\n if not database_thread.is_alive():\n log(\"Critical: Thread database is not alive\")\n time.sleep(10)\n\nif __name__ == '__main__':\n main()\n","repo_name":"nsg/immich-distribution","sub_path":"src/bin/sync-service.py","file_name":"sync-service.py","file_ext":"py","file_size_in_byte":8852,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"61"} +{"seq_id":"39904203962","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nArxiv Telegram Bot - Base Program\n\nProgram is used to start up the telegram bot\n\"\"\"\n\nimport logging\nimport os\n\nimport dotenv\n\nfrom telegram.ext import (\n Updater,\n MessageHandler,\n CommandHandler,\n Dispatcher,\n PicklePersistence,\n)\n\nfrom arxiv_telegram_bot.functions.handlers import (\n start,\n fetch,\n preference_conversation_handler,\n error,\n schedule,\n unschedule,\n)\n\n\nlogging.basicConfig(\n format=\"%(asctime)s - %(name)s - %(levelname)s - %(message)s\", level=logging.INFO\n)\nlogger = logging.getLogger(__name__)\n\ndotenv.load_dotenv()\nPORT = int(os.environ.get(\"PORT\", 8443))\nTOKEN = os.environ.get(\"TOKEN\")\nHEROKU_URL = os.environ.get(\"HEROKU_URL\")\n\n\ndef main():\n \"\"\"Start the bot.\"\"\"\n\n # Create the Updater and pass it your bot's token.\n persistence = PicklePersistence(filename=\"/tmp/arxivTelegramBot\")\n updater = Updater(TOKEN, use_context=True, persistence=persistence)\n dispatcher: Dispatcher = updater.dispatcher\n\n dispatcher.add_handler(CommandHandler(\"test\", start))\n dispatcher.add_handler(CommandHandler(\"latest\", fetch))\n dispatcher.add_handler(preference_conversation_handler())\n dispatcher.add_handler(CommandHandler(\"schedule\", schedule))\n dispatcher.add_handler(CommandHandler(\"unschedule\", unschedule))\n dispatcher.add_error_handler(error)\n\n if os.environ.get(\"ENV\") == \"HEROKU\":\n updater.start_webhook(\n listen=\"0.0.0.0\",\n port=int(PORT),\n url_path=TOKEN,\n webhook_url=f\"{HEROKU_URL}/{TOKEN}\",\n )\n else:\n updater.start_polling()\n\n updater.idle()\n","repo_name":"sonaalPradeep/arxiv-telegram-bot","sub_path":"arxiv_telegram_bot/base.py","file_name":"base.py","file_ext":"py","file_size_in_byte":1659,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"72237813633","text":"\"\"\"empty message\n\nRevision ID: 7f30c93eb9e2\nRevises: 7b6ade3a0ce9\nCreate Date: 2023-05-10 20:46:02.842541\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '7f30c93eb9e2'\ndown_revision = '7b6ade3a0ce9'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n with op.batch_alter_table('group', schema=None) as batch_op:\n batch_op.add_column(sa.Column('start_time', sa.Time(), nullable=True))\n\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n with op.batch_alter_table('group', schema=None) as batch_op:\n batch_op.drop_column('start_time')\n\n # ### end Alembic commands ###\n","repo_name":"Cpierswim/SwimTeamManager","sub_path":"backend/migrations/versions/7f30c93eb9e2_.py","file_name":"7f30c93eb9e2_.py","file_ext":"py","file_size_in_byte":789,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"41352690959","text":"# -*- coding: utf-8 -*-\nfrom __future__ import print_function\n\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Activation, Flatten\nfrom keras.layers import Convolution2D, MaxPooling2D\nfrom keras.optimizers import RMSprop\nimport argparse\nimport datetime\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport kml_utils\n\nSLASH = 0.2 # percentage of test(validation) data\n\n# parsing arguments\ndef parse_args():\n parser = argparse.ArgumentParser(description='image classifier')\n parser.add_argument('--data', dest='data_dir', default='data')\n parser.add_argument('--list', dest='list_dir', default='list')\n args = parser.parse_args()\n return args\n\nargs = parse_args()\nif kml_utils.exist_list(args.list_dir):\n print('Lists already exist in ./{0}. Use these lists.'.format(args.list_dir))\n classes, train_list, test_list = kml_utils.load_lists(args.list_dir)\nelse:\n print('Lists do not exist. Create list from ./{0}.'.format(args.data_dir))\n classes, train_list, test_list = kml_utils.create_list(args.data_dir, args.list_dir, SLASH)\n\ntrain_image, train_label = kml_utils.load_images(classes, train_list)\ntest_image, test_label = kml_utils.load_images(classes, test_list)\n\n# convert to numpy.array\nx_train = np.asarray(train_image)\ny_train = np.asarray(train_label)\nx_test = np.asarray(test_image)\ny_test = np.asarray(test_label)\n\nprint('train samples: ', len(x_train))\nprint('test samples: ', len(x_test))\n\nNUM_CLASSES = len(classes)\nBATCH_SIZE = 32\nEPOCH = 100\n\n# building the model\nprint('building the model ...')\n\nmodel = Sequential()\n\nmodel.add(Convolution2D(32, 3, 3, border_mode='valid',\n input_shape=x_train.shape[1:]))\nmodel.add(Activation('relu'))\nmodel.add(Convolution2D(32, 3, 3))\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\nmodel.add(Convolution2D(64, 3, 3, border_mode='valid'))\nmodel.add(Activation('relu'))\nmodel.add(Convolution2D(64, 3, 3))\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\nmodel.add(Flatten())\nmodel.add(Dense(256))\nmodel.add(Activation('relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(NUM_CLASSES))\nmodel.add(Activation('softmax'))\n\nrmsplop = RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0)\nmodel.compile(loss='categorical_crossentropy', optimizer=rmsplop, metrics=['accuracy'])\n\n# training\nhist = model.fit(x_train, y_train,\n batch_size=BATCH_SIZE,\n verbose=1,\n nb_epoch=EPOCH,\n validation_data=(x_test, y_test)) \n\n# save model\ndate_str = datetime.datetime.now().strftime('%Y%m%d%H%M%S')\nmodel.save('kml_' + date_str + '.model')\n\n# plot loss\nprint(hist.history.keys())\nloss = hist.history['loss']\nval_loss = hist.history['val_loss']\nacc = hist.history['acc']\nval_acc = hist.history['val_acc']\n\nnb_epoch = len(loss)\nfig, ax1 = plt.subplots()\nax1.plot(range(nb_epoch), loss, label='loss', color='b')\nax1.plot(range(nb_epoch), val_loss, label='val_loss', color='g')\nleg = plt.legend(loc='upper left', fontsize=10)\nleg.get_frame().set_alpha(0.5)\nax2 = ax1.twinx()\nax2.plot(range(nb_epoch), acc, label='acc', color='r')\nax2.plot(range(nb_epoch), val_acc, label='val_acc', color='m')\nleg = plt.legend(loc='upper right', fontsize=10)\nleg.get_frame().set_alpha(0.5)\nplt.grid()\nplt.xlabel('epoch')\nplt.savefig('graph_' + date_str + '.png')\nplt.show()\n","repo_name":"domkade/kill_me_learning","sub_path":"kml_train.py","file_name":"kml_train.py","file_ext":"py","file_size_in_byte":3429,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"61"} +{"seq_id":"13406273012","text":"import glob\nimport os\nimport tarfile\nimport torch.utils.data as data\nfrom PIL import Image\nimport numpy as np\nfrom torchvision.datasets.utils import download_url\n\n\nclass DomainDataset(data.Dataset):\n \"\"\"\n An abstract class for dataset that can be download as a zip file.\n The dataset can be kept in the zip file.\n \n Args:\n data_dir (string): Root directory of dataset where directory exists or will be saved to if download is set to\n True.\n url (str): url to the compressed dataset file\n md5_file (str): md5 checksum of the file\n filename (str): which name to save the downloaded file\n base_folder (str): where the images are located after the extraction\n num_images (int): how many images are located inside the base folder\n transform (callable, optional): A function/transform that takes in an\n PIL image and returns a transformed version. E.g\n ``transforms.RandomCrop``\n download (bool, optional): If true, downloads the dataset from the\n internet and puts it in data_dir directory. If dataset is already\n downloaded, it is not downloaded again.\n is_zip: if `True`, reads the images directly from compressed archive\n \"\"\"\n \n def __init__(self,\n data_dir,\n url=None,\n md5_file=None,\n filename='dataset.tar.gz',\n num_images=10000,\n transform=None,\n download=False,\n is_zip=True):\n\n super().__init__()\n\n self.data_dir = data_dir\n self.transform = transform\n\n self.url = url\n self.filename = filename\n self.zipped_filepath = os.path.join(self.data_dir, self.filename)\n\n self.md5_file = md5_file\n\n self.num_images = num_images\n self.is_zip = is_zip\n self.zfile = None\n\n if download:\n self.download()\n\n if not self._check_integrity():\n raise RuntimeError('Dataset not found or corrupted.' + ' You can use download=True to download it')\n\n def __getitem__(self, index):\n \"\"\"\n Args:\n index (int): Index\n\n Returns:\n image\n \"\"\"\n\n if self.zfile is None:\n self.zfile = tarfile.open(self.zipped_filepath)\n\n img = self.data[index]\n\n # doing this so that it is consistent with all other datasets\n # to return a PIL Image\n if self.is_zip:\n with self.zfile.extractfile(img) as f:\n img = Image.open(f)\n img = img.convert('RGB')\n elif isinstance(img, str):\n img = Image.open(img)\n img = img.convert('RGB')\n else:\n img = Image.fromarray(img)\n\n if self.transform is not None:\n img = self.transform(img)\n\n return img\n\n def __len__(self):\n return len(self.data)\n\n def _check_integrity(self):\n if self.is_zip and os.path.exists(self.zipped_filepath):\n return True\n\n if not self.is_zip and os.path.isdir(self.data_dir) and ((self.num_images is None or self.data is None)\n or len(self.data) == self.num_images):\n return True\n\n return False\n\n @property\n def data(self):\n if not hasattr(self, '_data'):\n if self.is_zip:\n self._data = list(f.name for f in tarfile.open(self.zipped_filepath, 'r').getmembers() if f.isfile())\n else:\n self._data = [\n f for f in glob.iglob(os.path.join(self.data_dir, '**', '*'), recursive=True)\n if DomainDataset.__is_image_file(f)\n ]\n\n return self._data\n\n def download(self):\n\n if self._check_integrity():\n return\n\n download_url(self.url, self.data_dir, self.filename, self.md5_file)\n\n if not self.is_zip:\n # extract file\n tar = tarfile.open(self.zipped_filepath, 'r')\n tar.extractall(self.data_dir)\n tar.close()\n\n def __repr__(self):\n fmt_str = 'Dataset ' + self.__class__.__name__ + '\\n'\n fmt_str += ' Number of datapoints: {}\\n'.format(self.__len__())\n fmt_str += ' Root Location: {}\\n'.format(self.data_dir)\n tmp = ' Transforms (if any): '\n fmt_str += '{0}{1}\\n'.format(tmp, self.transform.__repr__().replace('\\n', '\\n' + ' ' * len(tmp)))\n return fmt_str\n \n @staticmethod\n def __is_image_file(filename):\n \"\"\"Checks if a file is an allowed image extension.\n Args:\n filename (string): path to a file\n Returns:\n bool: True if the filename ends with a known image extension\n \"\"\"\n IMG_EXTENSIONS = ['.jpg', '.jpeg', '.png', '.ppm', '.bmp', '.pgm', '.tif']\n filename_lower = filename.lower()\n return any(filename_lower.endswith(ext) for ext in IMG_EXTENSIONS)\n","repo_name":"PatrickgHayes/gmm-dnn-for-interpretability","sub_path":"datasets/domain_dataset.py","file_name":"domain_dataset.py","file_ext":"py","file_size_in_byte":5009,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"39297617035","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# \n# Vamos a utilizar las herramientas de Web Scrapping para la siguiente página y poder obtener la lista de los paises Suramericanos y las variables claves de su economía: https://en.wikipedia.org/wiki/South_America\n\n# In[210]:\n\n\n# Importamos librerías\nimport requests\n\n\n# In[211]:\n\n\n# Escribimos la página web que vamos a scrapear\nwebsite_url = 'https://en.wikipedia.org/wiki/South_America'\npage = requests.get(website_url)\n\nprint(page.text)\n\n\n# In[212]:\n\n\n# Hacemos simulación de que alguien entra a la pagina web. Podemos encontrarlo en la pagina: https://developers.whatismybrowser.com/useragents/explore/software_name/chrome/\n# Y pasamos el \"User-agent\" para que pueda simular interacción con la página usando Navegador web y evite que nos bloqueen\ninter = {\"User-agent\": 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/74.0.3729.157 Safari/537.36'}\n\nanswer = requests.get(website_url, headers=inter)\n\n# Si se obtiene el código de respuesta \"200\", significa que no hemos tenido problema\nprint(answer)\n\n\n# Empezamos leyendo el código fuente de la página web y creando un objeto BeautifulSoup.\n# \n# BeautifulSoup permite crear un árbol de análisis para las páginas que se buscan analizar y que se pueden usar para extraer datos de HTML.\n# \n# La función prettify() en BeautifulSoup nos permitirá ver cómo se estructuran las etiquetas en el documento.\n\n# In[213]:\n\n\nfrom bs4 import BeautifulSoup\n\nsoup = BeautifulSoup(page.content, \"html.parser\")\nprint(soup.prettify())\n\n\n# In[214]:\n\n\ntitle = soup.find(id=\"firstHeading\")\ntitle\n\n\n# In[215]:\n\n\n# Imprimimos el titulo de la página web\nprint(title.text)\n\n\n# In[216]:\n\n\ntables = soup.find_all(\"table\", {\"class\":\"wikitable sortable\"})\n\n\n# In[217]:\n\n\n#En la página web podemos observar que la tabla que se desea extraer la información es la número 8\nmy_table = soup.find_all('table')[8]\nprint(my_table)\n\n\n# In[218]:\n\n\n# Buscamos todos los elementos 'th' en el cuerpo de la tabla, usando find_all():\nth = my_table.find_all(\"th\")\nprint(th)\n\n\n# In[219]:\n\n\n# Extraemos el título de los enlaces, para conocer los nombres de las columnas:\nfor i in range(len(th)):\n link = th[i].find(\"a\")\n\n if link != None:\n print(link.get(\"title\"))\n\n\n# In[220]:\n\n\n# Extraemos solo los valores de la tabla\nmytable=soup.find_all('table')[8]\nrows=mytable.find_all('tr')\nrows=rows[1:-1]\n\nc1=[]\nc2=[]\nc3=[]\nc4=[]\nc5=[]\nc6=[]\nc7=[]\n\nfor row in rows:\n x=row.find_all('td')\n# x=x[1].text\n# c1.append(x[:-1])\n x1=x[0]\n x2=x[1]\n x3=x[2]\n x4=x[3]\n x5=x[4]\n x6=x[5]\n x7=x[6]\n c1.append(x1.text[:-1])\n c2.append(x2.text[:-1])\n c3.append(x3.text[:-1])\n c4.append(x4.text[:-1])\n c5.append(x5.text[:-1])\n c6.append(x6.text[:-1])\n c7.append(x7.text[:-1])\n \nprint(c2)\n \n\n\n# In[221]:\n\n\nimport pandas as pd\n\ndf=pd.DataFrame()\n\ndf['Country']=c1\ndf['GPD_nominal']=c2\ndf['GDP_PPP']=c3\ndf['GDP_PPP_per_capita']=c4\ndf['Merchandise_exports']=c5\ndf['HDI']=c6\ndf['Percent_less_than_2']=c7\n\ndf\n\n\n# In[222]:\n\n\n# Exportamos la tabla a un archivo csv\ndf.to_csv(\"South_America_Economy.csv\", index = False)\n\n\n# In[ ]:\n\n\n\n\n","repo_name":"danamejia1810/Web-Scrapping-Economia-Paises-Sudamericanos","sub_path":"Código Práctica Tipología .py","file_name":"Código Práctica Tipología .py","file_ext":"py","file_size_in_byte":3175,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"74820125315","text":"# coding=utf-8\n\n\nfrom excel_generator.Style import Style\n# if you need add style for single gird, add it as need_merge cell\nfrom excel_generator.Common import bg_color, alignment, side_style, font_style, number_format\n\n# RSM and SD\n\ncommon1_matrix = [\n [None, None],\n [None, 'Product'],\n ['PCMO', 'WT'],\n ['PCMO', 'Ultra'],\n ['PCMO', 'HX8'],\n ['PCMO', 'HX7'],\n ['PCMO', 'HX6'],\n ['PCMO', 'HX5'],\n ['PCMO', 'HX3'],\n ['PCMO', 'HX2'],\n ['PCMO', 'Other'],\n ['CRTO', 'R6'],\n ['CRTO', 'R5'],\n ['CRTO', 'R4 Plus'],\n ['CRTO', 'R4'],\n ['CRTO', 'R3'],\n ['CRTO', 'R2'],\n ['CRTO', 'Gadus'],\n ['CRTO', 'Spirax'],\n ['CRTO', 'Other'],\n ['Sum Total', None]\n]\n\ncommon1_header_product = ['WT', 'Ultra', 'HX8', 'HX7', 'HX6', 'HX5', 'HX3', 'HX2', 'Other',\n 'R6', 'R5', 'R4 Plus', 'R4', 'R3', 'R2', 'Gadus', 'Spirax', 'Other']\n\ncommon1_need_merge = [\n {'coordinate': [2, 0, 10, 0], 'style': Style(bg_color[4], al=alignment[5])},\n {'coordinate': [11, 0, 19, 0], 'style': Style(bg_color[4], al=alignment[5])},\n {'coordinate': [20, 0, 20, 1],\n 'style': Style(bg_color[4], border=side_style[3], font=font_style[2], al=alignment[5])},\n\n {'coordinate': [1, 1, 1, 1], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [2, 1, 2, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [3, 1, 3, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [4, 1, 4, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [5, 1, 5, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [6, 1, 6, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [7, 1, 7, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [8, 1, 8, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [9, 1, 9, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [10, 1, 10, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [11, 1, 11, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [12, 1, 12, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [13, 1, 13, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [14, 1, 14, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [15, 1, 15, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [16, 1, 16, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [17, 1, 17, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [18, 1, 18, 1], 'style': Style(bg_color[4], al=alignment[1])},\n {'coordinate': [19, 1, 19, 1], 'style': Style(bg_color[4], al=alignment[1])}\n]\n\ncommon_header1 = {'matrix': common1_matrix, 'merge': common1_need_merge, 'product': common1_header_product,\n 'row': 21, 'col': 2, 'owner': 1, 'a_column': 7, 'owner_width': 3, 'follower_width': 6}\n\n# SGM\n\ncommon_header_sgm_1_0 = [\n ['SD', 'RSM', 'Ref Target KL', 'RSM&SD Submitted Target KL', 'Target Volume KL',\n 'Target C3 $', 'Target Proceed $']\n]\n\ncommon_header_sgm_1_0_merge = [\n {'coordinate': [0, 0, 0, 0], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 1, 0, 1], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])}\n]\n\ncommon_header_sgm_1_1 = [\n ['SD', 'RSM', 'Province', 'City', 'LE KL', 'Market size KL(this year)', 'Market Share %',\n 'Market size KL(last year)', 'Market Growth %', 'Platform', 'Market Share Score',\n 'Market Growth Score', 'Platform Score', 'Market Share Score(0.75)',\n 'Market Growth Score(0.15)',\n 'Platform Score(0.1)', 'Total Score', 'Increase %', 'Ref Target KL', 'Target KL']\n]\n\ncommon_header_sgm_1_1_merge = [\n {'coordinate': [0, 0, 0, 0], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 1, 0, 1], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 2, 0, 2], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 3, 0, 3], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])}\n]\n\ncommon_header_sgm_1_1_formula = [0, 0, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, -1]\ncommon_header_sgm_1_1_total = [0, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]\ncommon_header_sgm_1_1_number = [None, None, None, None, None, None, number_format['percent'], None,\n number_format['percent'], None,\n None, None, None, None, None, None, None, number_format['percent'], None, None]\n\ncommon_header_sgm_2_0 = [\n ['SD', 'RSM', 'Province', 'City', 'Ref Volume KL', 'Ref C3 $', 'Ref Proceed $',\n 'Target Volume KL', 'Target C3 $', 'Target Proceed $']\n]\n\ncommon_header_sgm_2_1 = [\n ['SD', 'RSM', 'Province', 'City', 'UC3 $', 'UC3 $', 'UC3 $', 'UC3 $', 'UC3 $',\n 'UC3 $', 'UC3 $', 'UC3 $', 'UC3 $', 'UNP $', 'UNP $', 'UNP $', 'UNP $', 'UNP $',\n 'UNP $', 'UNP $', 'UNP $', 'UNP $', 'Portfolio %', 'Portfolio %', 'Portfolio %',\n 'Portfolio %', 'Portfolio %', 'Portfolio %', 'Portfolio %', 'Portfolio %', 'Portfolio %'],\n ['SD', 'RSM', 'Province', 'City', 'WT', 'Ultra', 'HX8', 'HX7', 'HX6', 'HX5', 'HX3', 'HX2', 'Other',\n 'WT', 'Ultra', 'HX8', 'HX7', 'HX6', 'HX5', 'HX3', 'HX2', 'Other',\n 'WT', 'Ultra', 'HX8', 'HX7', 'HX6', 'HX5', 'HX3', 'HX2', 'Other']\n]\ncommon_header_sgm_2_1_merge = [\n {'coordinate': [0, 0, 1, 0], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 1, 1, 1], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 2, 1, 2], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 3, 1, 3], 'style': Style(bg_color[4], font=font_style[2], al=alignment[1])},\n {'coordinate': [0, 4, 0, 12], 'style': Style(bg_color[4], font=font_style[2], al=alignment[2])},\n {'coordinate': [0, 13, 1, 21], 'style': Style(bg_color[4], font=font_style[2], al=alignment[2])},\n {'coordinate': [0, 22, 1, 30], 'style': Style(bg_color[4], font=font_style[2], al=alignment[2])}\n]\n\ncommon_header_sgm_1 = {\n 0: {'data': common_header_sgm_1_0, 'scale': [1, 7], 'merge': common_header_sgm_1_0_merge, 'formula': None,\n 'number_format': None},\n 1: {'data': common_header_sgm_1_1, 'scale': [1, 20], 'merge': common_header_sgm_1_1_merge,\n 'formula': common_header_sgm_1_1_formula, 'number_format': common_header_sgm_1_1_number,\n 'total': common_header_sgm_1_1_total}\n}\n\ncommon_header_sgm_2 = {\n 0: {'data': common_header_sgm_2_0, 'scale': [1, 10], 'merge': None, 'formula': None, 'number_format': None},\n 1: {'data': common_header_sgm_2_1, 'scale': [2, 31], 'merge': common_header_sgm_2_1_merge, 'formula': [],\n 'number_format': None}\n}\n\nheader_index = {\n 'RSM': common_header1,\n 'SGM': {1: common_header_sgm_1,\n 2: common_header_sgm_2,\n 3: common_header_sgm_1,\n 4: common_header_sgm_2}\n}\n","repo_name":"intwzt/ShellExcel","sub_path":"template/HeaderTemplate.py","file_name":"HeaderTemplate.py","file_ext":"py","file_size_in_byte":7099,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"19927543175","text":"from django.conf import settings\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\n\nclass Migration(migrations.Migration):\n\n initial = True\n\n dependencies = [\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ('sessions', '0001_initial'),\n ('core', '0001_initial'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='article',\n name='owner',\n field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL),\n ),\n migrations.AddField(\n model_name='article',\n name='owner_sessions',\n field=models.ManyToManyField(blank=True, db_table='article_owners', to='sessions.Session'),\n ),\n ]\n","repo_name":"Flaiers/flatype","sub_path":"src/apps/core/migrations/0002_initial.py","file_name":"0002_initial.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"61"} +{"seq_id":"27344817364","text":"import copy\nimport math\nfrom functools import reduce\nfrom model.Cluster import Cluster\nfrom model.PointsWorker import PointsWorker\n\nEXTERNAL_K_MAX = 20\nEXTERNAL_K_MIN = 10\nINTERNAL_K_MAX = 10\nINTERNAL_K_MIN = 4\n\n\nclass UnionClusterizationWorker:\n\n def __init__(self, points):\n self.worker = PointsWorker()\n self.worker.points = points\n\n def make_clusterization(self, k_max, k_min, clusters=None):\n\n k_cluster_dict = {}\n distance_k_dict = {}\n clusters = clusters\n\n for k in range(k_max, k_min - 1, -1):\n print(f\"Получаем {k}-кластаризацию\")\n # получение и разметка кластеров\n clusters = [*self.worker.make_union_clustering(k, clusters)]\n print(f\"Получаем знаки на {k}-кластаризацию\")\n signs = [f\"{i + 1} class\" for i in range(len(clusters))]\n for cl, sign in zip(clusters, signs):\n cl.accept_class_sign(sign)\n\n # оценка кластеризации\n internal = reduce(lambda a, cl: a + cl.get_internal_cluster_distance(), clusters, 0) / len(clusters)\n external = Cluster([cl.get_center() for cl in clusters]).get_internal_cluster_distance()\n k_cluster_dict[k] = copy.deepcopy(clusters)\n distance_k_dict[math.fabs(internal - external)] = k\n\n return k_cluster_dict[distance_k_dict[min(distance_k_dict.keys())]]\n","repo_name":"makdim5/SpaceGeometry3DAnalizer","sub_path":"ML_Union_Algorithm/model/clusterization.py","file_name":"clusterization.py","file_ext":"py","file_size_in_byte":1479,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23414918351","text":"#!/bin/python2\n\nimport fileinput\n\ndef read_int(lines, cur):\n return (int(lines[cur]), cur + 1)\n\ndef read_matr(lines, cur):\n matr = [map(int, lines[cur+i].split()) for i in xrange(4)]\n return (matr, cur + 4)\n\nlines = [line for line in fileinput.input()]\n\nT = int(lines[0])\ncur = 1\nfor i in xrange(1, T+1):\n n1, cur = read_int(lines, cur)\n m1, cur = read_matr(lines, cur)\n n2, cur = read_int(lines, cur)\n m2, cur = read_matr(lines, cur)\n\n ans = set(m1[n1-1]).intersection(set(m2[n2-1]))\n ans = list(ans)\n if len(ans) == 1:\n msg = str(ans[0])\n elif len(ans) == 0:\n msg = \"Volunteer cheated!\"\n elif len(ans) > 1:\n msg = \"Bad magician!\"\n\n print (\"Case #%i: \" + msg) % (i,)\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_135/2823.py","file_name":"2823.py","file_ext":"py","file_size_in_byte":690,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"74628377153","text":"\"\"\"\nThread Pool:\nReusing existing Threads, Because creating threads are expensive.\nAlso, Most of the computer OS caps the number of Thread that they can run.\nAnd program could be crash if you will try to create more number of threads.\n\nWhy Creating Threads are Expensive ?\nA: For that we need to look at the structure of Thread Pool:\n 1. components:\n 1. Work Producers (tasks like Network call, i/o task, db connections, interrupts, rw to a file, etc)\n 2. Job Queue (intermediate b/w producer and pool)\n 3. Thread Pool (list of threads of fixed len)\n 2. lot of calls to OS and OS need to allocate OS and CPU\n\nSo, ThreadPoolExecutor comes to rescue, an approach to keep up the throughput is to create & instantiate\na pool of idle threads beforehand and reuse the threads from this pool until all the threads are exhausted.\nAlso, the pool keeps track and manages the threads lifecycle and schedules them on the programmer’s behalf\nthus making the code much simpler and less buggy.\n\nwe can use 3 methods to spawn threads from ThreadPoolExecutor:\n1. map(fn, *iterables, timeout = None, chunksize = 1)\n2. submit(fn, *args, **kwargs) -> Future:\n3. shutdown(wait = True, *, cancel_futures = False)\n i. It must be called before executor.submit() and executor.map() method else it would throw RuntimeError.\n ii. It signals the executor to free up all resources when the futures are done executing.\n iii. wait=True makes the method not to return until execution of all threads is done and resources are freed up.\n iv. cancel_futures=True then the executor will cancel all the future threads that are yet to start.\n\n\n`workers` are just MAX number of running tasks on parallel threads, hence a thread is nothing but worker !\n\n\"\"\"\n\nimport time\n\nimport logging\nimport random\n\nfrom threading import Thread, get_ident, current_thread\nfrom threading import Timer # utilized to run a code after a specified time period\n\n\n# concurrent is the high level version of Threading to hide all the ugly working of thread details\nfrom concurrent.futures import Future # The upcoming proxy object\nfrom concurrent.futures import ThreadPoolExecutor # the Thread Pool Executor, Python 3.2+\n# from concurrent.futures import ProcessPoolExecutor # the Process Pool Executor\n# from concurrent.futures import as_completed\n\n\n# to see the concept of thread reusing we need to make uneven time period for each tasks\nwait_time = 10\n\n\ndef some_task(item):\n \"\"\"This Function will take 14 sec to complete\"\"\"\n # no_tasks = random.randrange(start=0, stop=10, step=1)\n logging.info(f\"Task: {item} started!\")\n # id of current Thread, is created by OS and id belongs to the worker\n logging.info(f'Thread {item}: id = {get_ident()}')\n logging.info(f'Thread {item}: name = {current_thread().name}')\n logging.info(f'Thread {item}: sleeping for {wait_time}')\n time.sleep(random.randrange(wait_time))\n logging.info(f'Thread {item}: finished')\n\n\n# Main function\ndef main():\n logging.basicConfig(\n format='%(levelname)s - %(asctime)s: %(message)s',\n datefmt='%H:%M:%S',\n level=logging.DEBUG\n )\n logging.info('App Start')\n\n cores = 4 # MacBook Pro cores\n workers = 2*cores + 1\n items = 20\n\n # No need to Join the Threads\n # No need to Monitor or Handle the Threads\n # automatically spawn a new worker when there is\n # Said objects use significant amount of memory and for last project uses the large memory.\n # To reduce this memory management overhead (allocating and deallocating many threads)\n with ThreadPoolExecutor(max_workers=workers) as executor:\n executor.map(some_task, range(0, items))\n\n # some of the ids will gets repeated in the terminal that depicts the reuse of Threads\n logging.info('App Finished')\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"deepanshumehtaa/Concurrency-Python","sub_path":"tut4_ThreadPoolExecutor_map.py","file_name":"tut4_ThreadPoolExecutor_map.py","file_ext":"py","file_size_in_byte":3843,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"28034874997","text":"import torch\nimport numpy as np\nfrom torch.autograd import Variable\nimport torch.nn as nn\nimport torch.nn.functional as F\n\n\nbnmom = 1e-3\n\n\nclass mie(nn.Module):\n\n\tdef __init__(self,modim,tardim,dp):\n\t\t\n\t\tsuper(mie, self).__init__()\n\t\ttctrl = modim[-1]\n\t\tself.modim = modim\n\t\tself.tardim = tardim\n\t\tself.dp = dp\n\n\t\tself.moac = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tif idx == len(modim)-2:\n\t\t\t\tself.moac.append(nn.Softplus(modim[idx+1]).cuda())\n\t\t\telse:\t\t\t\t\n\t\t\t\tself.moac.append(nn.PReLU(modim[idx+1]).cuda())\n\n\t\tself.xmoac = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tself.xmoac.append(nn.PReLU(modim[len(modim)-idx-2]).cuda())\n\n\t\tself.ac = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tself.ac.append(nn.PReLU(tardim[idx+1]).cuda())\n\n\t\tself.xac = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tself.xac.append(nn.PReLU(tardim[len(tardim)-idx-2]).cuda())\n\n\t\tself.net = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tx = nn.ModuleList([])\n\t\t\tfor j in range(tctrl):\n\t\t\t\tx.append(nn.Linear(tardim[idx], tardim[idx+1]).cuda())\n\t\t\tself.net.append(x)\n\n\t\tself.xnet = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tx = nn.ModuleList([])\n\t\t\tfor j in range(tctrl):\n\t\t\t\tx.append(nn.Linear(tardim[len(tardim)-idx-1], tardim[len(tardim)-idx-2]).cuda())\n\t\t\tself.xnet.append(x)\n\n\t\tself.monet = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tx = nn.ModuleList([])\n\t\t\tx.append(nn.Linear(modim[idx], modim[idx+1]).cuda())\n\t\t\tself.monet.append(x)\n\n\t\tself.xmonet = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tx = nn.ModuleList([])\n\t\t\tx.append(nn.Linear(modim[len(modim)-idx-1], modim[len(modim)-idx-2]).cuda())\n\t\t\tself.xmonet.append(x)\n\n\t\t\t\n\t\tself.d = nn.Dropout(p=dp)\n\t\tbnmom=1e-3\n\n\t\tself.bn = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tself.bn.append(nn.BatchNorm1d(tardim[idx], affine=True, momentum=bnmom).cuda())\n\n\t\tself.xbn = nn.ModuleList([])\n\t\tfor idx in range(len(tardim)-1):\n\t\t\tself.xbn.append(nn.BatchNorm1d(tardim[len(tardim)-idx-2], affine=True, momentum=bnmom).cuda())\n\n\t\tself.mobn = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tself.mobn.append(nn.BatchNorm1d(modim[idx], affine=True, momentum=bnmom).cuda())\n\n\t\tself.xmobn = nn.ModuleList([])\n\t\tfor idx in range(len(modim)-1):\n\t\t\tself.xmobn.append(nn.BatchNorm1d(modim[len(modim)-idx-2], affine=True, momentum=bnmom).cuda())\n\n\n\tdef m(self, x0): #x0:bsxdatadim\n\t\t\n\t\tdobn = 1>2\n\n\t\tfor idx in range(len(self.modim)-1):\n\t\t\tif dobn:\n\t\t\t\tx0 = self.mobn[idx](x0)\n\t\t\tx0 = self.monet[idx][0](x0)\n\t\t\tx0 = self.moac[idx](x0)\n\n\t\treturn x0\n\n\tdef xm(self, x0): #x0:bsxdatadim\n\t\t\n\t\tdobn = 1>2\n\n\t\tfor idx in range(len(self.modim)-1):\n\t\t\tx0 = self.xmonet[idx][0](x0)\n\t\t\tx0 = self.xmoac[idx](x0)\n\t\t\tif dobn and idx < len(self.modim)-2:\n\t\t\t\tx0 = self.xmobn[idx](x0)\n\t\treturn x0\n\t\t\n\tdef g(self, x0,weights): #x0:bsxdatadim\n\t\t\n\t\ttctrl = len(weights)\n\t\t\n\t\tdobn = 1>2\n\n\t\tfor idx in range(len(self.tardim)-1):\n\t\t\tif dobn:\n\t\t\t\tx0 = self.bn[idx](x0)\n\t\t\ttx = 0\n\t\t\ttx_test = 0\n\t\t\tw_sum = 0\n\t\t\tfor j in range(tctrl):\n\t\t\t\tv = self.net[idx][j](x0)\n\t\t\t\tw = weights[j]/torch.sum(weights)\n\t\t\t\tw_sum += w\n\t\t\t\twv = w*v\n\t\t\t\ttx_test = tx_test + v\n\t\t\t\ttx = tx + wv\n\t\t\t\t#tx = tx + weights[j]/torch.sum(weights) * self.net[idx][j](x0)\n\t\t\tx0 = tx\n\t\t\tx0 = self.ac[idx](x0)\n\n\t\treturn x0\n\n\tdef h(self, x0, weights): #x0:bsxdatadim\n\n\t\ttemp = int(x0.shape[0])\n\t\tx0 = x0.reshape((1, temp))\n\n\t\ttctrl = len(weights)\n\n\t\tdobn = 1>2\n\n\t\tfor idx in range(len(self.tardim)-1):\n\t\t\ttx = 0\n\t\t\tfor j in range(tctrl):\n\t\t\t\ttx = tx + weights[j]/torch.sum(weights)* self.xnet[idx][j](x0)\n\t\t\tx0 = tx\n\t\t\tx0 = self.xac[idx](x0)\n\t\t\tif dobn and idx < len(self.tardim)-2:\n\t\t\t\tx0 = self.xbn[idx](x0)\n\t\t\t\n\t\treturn x0\n\n\n\n","repo_name":"YuanBoot/Intrinsic_Garment_Space","sub_path":"scripts/mie_model.py","file_name":"mie_model.py","file_ext":"py","file_size_in_byte":3700,"program_lang":"python","lang":"en","doc_type":"code","stars":229,"dataset":"github-code","pt":"61"} +{"seq_id":"5699659606","text":"# Baekjoon Online Judge - 2776번. 암기왕\n\nimport sys\ninput = sys.stdin.readline\n\n\nT = int(input())\n\nfor _ in range(T):\n N = int(input())\n numbers = list(map(int, input().split()))\n numbers.sort()\n M = int(input())\n target = list(map(int, input().split()))\n answer = []\n for num in target:\n left, right = 0, N - 1\n found = False\n while left <= right:\n mid = (left + right) // 2\n if num == numbers[mid]:\n found = True\n break\n # 찾고자 하는 값이 현재 값 보다 작다면 값의 범위를 줄인다.\n if num < numbers[mid]:\n right = mid - 1\n # 찾고자 하는 값이 현재 값 보다 크다면 값의 범위를 늘린다.\n else:\n left = mid + 1\n\n if found:\n answer.append(1)\n else:\n answer.append(0)\n\n for i in answer:\n print(i)\n","repo_name":"wnstj-yang/Algorithm","sub_path":"BOJ/BOJ_2776.py","file_name":"BOJ_2776.py","file_ext":"py","file_size_in_byte":955,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"29837844737","text":"import re\nfrom BeautifulSoup import BeautifulSoup\n\nfrom fiftystates.scrape import NoDataForPeriod\nfrom fiftystates.scrape.legislators import Legislator, LegislatorScraper\n\ndef split_name(full_name):\n last_name = full_name.split(',')[0]\n rest = ','.join(full_name.split(',')[1:])\n\n m = re.search('(\\w+)\\s([A-Z])\\.$', rest)\n if m:\n first_name = m.group(1)\n middle_name = m.group(2)\n else:\n first_name = rest\n middle_name = ''\n\n if last_name.endswith(' Jr.'):\n first_name += ' Jr.'\n last_name = last_name.replace(' Jr.', '')\n\n return (first_name.strip(), last_name.strip(), middle_name.strip())\n\nclass KYLegislatorScraper(LegislatorScraper):\n state = 'ky'\n\n def scrape(self, chamber, year):\n if year != '2009':\n raise NoDataForPeriod(year)\n\n if chamber == 'upper':\n leg_list_url = 'http://www.lrc.ky.gov/senate/senmembers.htm'\n else:\n leg_list_url = 'http://www.lrc.ky.gov/house/hsemembers.htm'\n\n with self.urlopen(leg_list_url) as leg_list:\n leg_list = BeautifulSoup(leg_list)\n leg_table = leg_list.find(id=\"table2\")\n\n for row in leg_table.findAll('tr')[1:]:\n leg_link = row.findAll('td')[1].font\n if leg_link: leg_link = leg_link.a\n if not leg_link:\n # Vacant seat\n continue\n\n full_name = leg_link.contents[0].strip()\n\n district = \"\"\n for text in row.findAll('td')[2].findAll(text=True):\n district += text.strip()\n district = district.strip()\n\n self.parse_legislator(chamber, year, full_name,\n district, leg_link['href'])\n\n def parse_legislator(self, chamber, year, full_name, district, url):\n with self.urlopen(url) as leg_page:\n leg_page = BeautifulSoup(leg_page)\n name_str = leg_page.find('strong').contents[0].strip()\n\n if name_str.endswith('(D)'):\n party = 'Democrat'\n elif name_str.endswith('(R)'):\n party = 'Republican'\n elif name_str.endswith('(I)'):\n party = 'Independent'\n else:\n party = 'Other'\n\n full_name = full_name.replace('\\n', '').replace('"', '\"')\n full_name = full_name.replace('\\t', '').replace('\\r', '')\n (first_name, last_name, middle_name) = split_name(full_name)\n\n legislator = Legislator(year, chamber, district, full_name,\n first_name, last_name, middle_name, party)\n legislator.add_source(url)\n\n self.save_legislator(legislator)\n","repo_name":"runderwood/fiftystates","sub_path":"fiftystates/scrape/ky/legislators.py","file_name":"legislators.py","file_ext":"py","file_size_in_byte":2774,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"61"} +{"seq_id":"36183420302","text":"import pygame, random\nimport wizards.constants, wizards.monster_ai\n\nclass BaseMonster(pygame.sprite.Sprite):\n def __init__(self, mid, x, y, name, level, m_type):\n super().__init__()\n self.monster_id = mid\n self.x = x\n self.y = y \n self.name = name\n self.level = level\n self.orig_hp = self.get_hp(level)\n self.hp = self.orig_hp \n self.dead = False\n self.weapon = None\n self.current_weapon = None\n self.inventory = []\n self.m_type = m_type\n self.fleeing = False\n self.morale = 6\n self.save_magic = 16\n self.undead = False\n self.never_surrender = False\n self.hit_chance = wizards.bags.NumberBag(1, 20, 2)\n\n self.weight = 0\n\n self.charmable = True\n self.charmed = False\n self.charmed_by = None\n self.charm_duration = 0\n self.charm_gfx = None\n\n self.asleep = False\n self.asleep_for = 0\n\n # TODO Monster AI\n self.ai = None\n #self.ai = self.set_ai(wizards.monster_ai.PassiveAI(coll_map))\n self.level_seen = {}\n self.moved = False\n self.player_seen = False\n\n def get_id(self):\n return self.monster_id\n \n def updatePosition(self,direction,col_map):\n \n new_x = self.x\n new_y = self.y\n if direction == 0:\n new_y = self.y - 1\n elif direction == 1:\n new_x = self.x + 1\n elif direction == 2:\n new_y = self.y + 1\n elif direction == 3:\n new_x = self.x - 1 \n \n if self.is_valid_move(new_x, new_y, col_map):\n self.x = new_x\n self.y = new_y\n \n self.rect.x = self.x * wizards.constants.CHAR_SIZE\n self.rect.y = self.y * wizards.constants.CHAR_SIZE\n\n def __str__(self):\n return self.name + \": \" + str(self.monster_id) + \" >> \"+ str(self.y) + \"_\" + str(self.x)\n\n def __eq__(self, other):\n return self.monster_id == other.monster_id\n\n def __lt__(self, other):\n return self.monster_id < other.monster_id\n\n def __hash__(self):\n return self.monster_id\n\n def set_ai(self, a):\n self.ai = a\n\n def set_position(self, x, y, monster_map):\n\n if (self.x, self.y) in monster_map:\n del monster_map[(self.x, self.y)]\n\n self.x = x\n self.y = y\n monster_map[(self.x, self.y)] = self.monster_id\n self.rect.x = self.x * wizards.constants.CHAR_SIZE\n self.rect.y = self.y * wizards.constants.CHAR_SIZE\n \n def is_valid_move(self, x, y, col_map):\n if col_map[y][x] == 0:\n return True\n else:\n return False\n\n def take_damage(self, dmg):\n self.hp -= dmg\n if self.hp < 1:\n self.dead = True\n\n def get_hp(self, num_of_dice):\n \"\"\"Get initial hitpoints, level * D8\"\"\"\n total = 0\n for i in range(num_of_dice):\n total += random.randrange(1,9)\n return total\n\n def in_panic(self):\n roll = (random.randrange(6) + 1) + (random.randrange(6) + 1)\n if roll > self.morale:\n return True\n else:\n return False\n\n def get_weapon_damage(self):\n if self.current_weapon is not None:\n return self.current_weapon.max_damage\n else:\n return 0\n\n def do_turn(self, player, player_map, collision_map, monster_map, combat_resovler):\n if self.ai is not None:\n self.ai.update(self, player, player_map, collision_map, monster_map, combat_resovler)\n print(self.name + \" \" + str(self.monster_id) + \" has moved\")\n\n def add_item_to_inventory(self, i):\n self.inventory.append(i)","repo_name":"Grufferz/wizards-of-twiddly","sub_path":"wizards/base_monster.py","file_name":"base_monster.py","file_ext":"py","file_size_in_byte":3752,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"14868653851","text":"import unittest\nfrom montyhall import MontyHall\n\nclass MontyHallTest(unittest.TestCase):\n\n def test_make_choice(self):\n num_doors = 3\n num_rounds = 1000\n one_third = float(1) / 3\n two_thirds = float(2) / 3\n accepted_delta = 0.05\n\n correct_guesses_initial = [MontyHall(num_doors).make_choice(keep_initial_choice=True) for i in range(num_rounds)].count(True)\n correct_guesses_changed = [MontyHall(num_doors).make_choice(keep_initial_choice=False) for i in range(num_rounds)].count(True)\n\n # Keeping initial guess should give us 1/3 correct correct guesses\n self.assertAlmostEqual(correct_guesses_initial, one_third * num_rounds, delta=accepted_delta * num_rounds)\n # Changing initial guess should give us 2/3 correct correct guesses\n self.assertAlmostEqual(correct_guesses_changed, two_thirds * num_rounds, delta=accepted_delta * num_rounds)\n\nif __name__ == '__main__':\n unittest.main()\n ","repo_name":"fsto/pyMontyHall","sub_path":"test_montyhall.py","file_name":"test_montyhall.py","file_ext":"py","file_size_in_byte":990,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"24974687886","text":"#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\n# @FileName : demo_2_4.py\n# @Time : 2022/9/26 19:28\n# @Author : liang.lian\nimport re\n\ntext1 = \"11/27/2012\"\ntext2 = \"Nov 27, 2012\"\ntext3 = 'Today is 11/27/2012. PyCon starts 3/13/2013.'\n\ndatepat = re.compile(r'\\d+/\\d+/\\d+')\n\nif datepat.match(text1):\n print('yes')\nelse:\n print('no')\n\nif datepat.match(text2):\n print('yes')\nelse:\n print('no')\n\ndatepat1 = re.compile(r'(\\d+)(/\\d+)(/\\d+)')\nprint(datepat1.findall(text3))\n\nfor m in datepat1.finditer(text3):\n print(m.groups())","repo_name":"comeonlian/code-dev","sub_path":"python-cookbook/chapter02/demo_2_4.py","file_name":"demo_2_4.py","file_ext":"py","file_size_in_byte":541,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"75144368834","text":"import cv2\nimport os\n\n#Initialize camera\ncam1 = cv2.VideoCapture(0);\ncam2 = cv2.VideoCapture(0);\n\nret1, image1 = cam1.read();\nret2, image2 = cam2.read();\n\nif ret1:\n os.system('mkdir ./pics/')\n cv2.imwrite(\"./pics/ex1.jpg\", image1)\n\nif ret2:\n os.system('mkdir ./pics/')\n cv2.imwrite(\"./pics/ex2.jpg\", image2)\n\ncam1.release()\ncam2.release()","repo_name":"RexGoliath1/Argyle","sub_path":"examples/webcam.py","file_name":"webcam.py","file_ext":"py","file_size_in_byte":342,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"75237175554","text":"import yaml\nimport os\nfrom utils import recreate_dirs\n\n\nclass Config:\n\n def __init__(self, cfg_id, create_dirs=False):\n self.id = cfg_id\n cfg_name = 'config/statereg/%s.yml' % cfg_id\n if not os.path.exists(cfg_name):\n print(\"Config file doesn't exist: %s\" % cfg_name)\n exit(0)\n cfg = yaml.safe_load(open(cfg_name, 'r'))\n\n # create dirs\n self.base_dir = 'results'\n self.cfg_dir = '%s/statereg/%s' % (self.base_dir, cfg_id)\n self.model_dir = '%s/models' % self.cfg_dir\n self.result_dir = '%s/results' % self.cfg_dir\n self.log_dir = '%s/log' % self.cfg_dir\n self.tb_dir = '%s/tb' % self.cfg_dir\n os.makedirs(self.model_dir, exist_ok=True)\n os.makedirs(self.result_dir, exist_ok=True)\n if create_dirs:\n recreate_dirs(self.log_dir, self.tb_dir)\n\n # training config\n self.meta_id = cfg['meta_id']\n self.seed = cfg['seed']\n self.fr_num = cfg['fr_num']\n self.v_net = cfg.get('v_net', 'lstm')\n self.v_net_param = cfg.get('v_net_param', None)\n self.v_hdim = cfg['v_hdim']\n self.mlp_dim = cfg['mlp_dim']\n self.cnn_fdim = cfg['cnn_fdim']\n self.lr = cfg['lr']\n self.num_epoch = cfg['num_epoch']\n self.iter_method = cfg['iter_method']\n self.shuffle = cfg.get('shuffle', False)\n self.num_sample = cfg.get('num_sample', 20000)\n self.save_model_interval = cfg['save_model_interval']\n self.fr_margin = cfg['fr_margin']\n self.pose_only = cfg.get('pose_only', False)\n self.causal = cfg.get('causal', False)\n self.cnn_type = cfg.get('cnn_type', 'mlp')\n\n # misc config\n self.humanoid_model = cfg['humanoid_model']\n self.vis_model = cfg['vis_model']\n","repo_name":"Garfield-kh/PoseTriplet","sub_path":"imitator/pose_imitation/utils/statereg_config.py","file_name":"statereg_config.py","file_ext":"py","file_size_in_byte":1815,"program_lang":"python","lang":"en","doc_type":"code","stars":293,"dataset":"github-code","pt":"61"} +{"seq_id":"24798336808","text":"from contextlib import contextmanager\nfrom dataclasses import dataclass\nfrom typing import Iterable, Iterator, List, Optional, Sequence, Collection, TypeVar, Union\n\nfrom rich.live import Live\nfrom rich.progress import Progress, BarColumn, TimeElapsedColumn, TimeRemainingColumn, TextColumn\nfrom rich.traceback import install\ninstall(show_locals=False)\n\nV = TypeVar('V')\n\n@dataclass\nclass Bar:\n progress: Progress\n job_id: int\n\n def iter(\n self,\n it: Iterable[V],\n description: str = \"\") -> Iterator[V]:\n try:\n total = len(it)\n except:\n total = -1\n\n self.progress.reset(self.job_id)\n self.progress.update(\n self.job_id,\n total=total if total > 0 else 3,\n description=description)\n if total < 0:\n self.progress.advance(self.job_id)\n for i in it:\n yield i\n if total > 0:\n self.progress.advance(self.job_id)\n\n def range(\n self,\n *args,\n description=\"\"):\n for arg in args:\n assert not isinstance(arg, str), f\"'{arg}' is not int, it's maybe description\"\n return self.iter(\n range(*args),\n description=description)\n\n def update(self, description: str = None):\n self.progress.update(self.job_id, description=description)\n\n\n@contextmanager\ndef progress_bar(num: int = 1, refresh_hz=1):\n job_progress = Progress(\n \"{task.description}\",\n TimeElapsedColumn(),\n BarColumn(),\n TextColumn(\"[progress.percentage]{task.percentage:>3.0f}%\"),\n TimeRemainingColumn()\n )\n\n bars = tuple(Bar(job_progress, job_progress.add_task(\"\"))\n for _ in range(num))\n\n with Live(job_progress, refresh_per_second=refresh_hz):\n yield bars\n","repo_name":"liyihc/clad","sub_path":"clad/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1852,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"41787239065","text":"# -*- coding: utf-8 -*-\n\nimport json\nimport traceback\nimport math\nfrom scrapy.conf import settings\n\nfrom product_ranking.guess_brand import guess_brand_from_first_words\nfrom product_ranking.items import Price, SiteProductItem\nfrom product_ranking.spiders import BaseProductsSpider, cond_set_value\nfrom product_ranking.powerreviews import parse_powerreviews_buyer_reviews\n\nfrom scrapy import Request\nfrom scrapy.log import DEBUG\n\n\nclass BJSProductsSpider(BaseProductsSpider):\n name = 'bjs_products'\n allowed_domains = ['bjs.com', 'bjswholesale-cors.groupbycloud.com', 'readservices-b2c.powerreviews.com']\n\n SEARCH_URL = \"https://bjswholesale-cors.groupbycloud.com/api/v1/search\"\n PRODUCT_URL = \"https://api.bjs.com/digital/live/api/v1.0/pdp/10201?productId={product_id}&pageName=PDP&clubId=0096\"\n REVIEW_URL = \"http://readservices-b2c.powerreviews.com/m/9794/l/en_US/product/{part_num}/reviews?\"\n\n payload = {\n \"area\": \"BCProduction\",\n \"biasing\": {\"biases\": []},\n \"collection\": \"productionB2CProducts\",\n \"excludedNavigations\": ['visualVariant.nonvisualVariant.availability'],\n \"fields\": ['*'],\n \"pageSize\": 40,\n \"query\": \"\",\n \"refinements\": [],\n \"skip\": 0,\n \"sort\": {\n \"field\": \"_relevance\",\n \"order\": \"Descending\"\n }\n }\n\n headers = {\n 'Accept-Encoding': 'gzip, deflate, br',\n 'Content-Type': 'application/json',\n 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 '\n '(KHTML, like Gecko) Chrome/57.0.2987.133 Safari/537.36'\n }\n\n def __init__(self, *args, **kwargs):\n self.total_matches = None\n super(BJSProductsSpider, self).__init__(\n site_name=self.allowed_domains[0],\n *args,\n **kwargs)\n self.user_agent = 'Slackbot-LinkExpanding 1.0 (+https://api.slack.com/robots)'\n\n settings.overrides['DOWNLOADER_CLIENTCONTEXTFACTORY'] = 'product_ranking.utils.TLSFlexibleContextFactory'\n\n def start_requests(self):\n for request in super(BJSProductsSpider, self).start_requests():\n if not self.product_url:\n data = self.payload.copy()\n data['query'] = self.searchterms[0]\n data['skip'] = 0\n\n request = request.replace(url=self.SEARCH_URL, method=\"POST\", body=json.dumps(data),\n headers=self.headers,\n meta={'search_term': self.searchterms[0], 'remaining': self.quantity})\n if self.product_url:\n prod = SiteProductItem()\n prod['is_single_result'] = True\n prod['url'] = self.product_url\n prod['search_term'] = ''\n\n product_id = self.product_url.split('/')[-1]\n url = self.PRODUCT_URL.format(product_id=product_id)\n request = request.replace(url=url, callback=self._parse_single_product, meta={'product': prod})\n\n yield request\n\n def _parse_single_product(self, response):\n return self.parse_product(response)\n\n def _scrape_total_matches(self, response):\n if self.total_matches:\n return self.total_matches\n try:\n contents = json.loads(response.body)\n self.total_matches = int(contents.get('totalRecordCount'))\n return self.total_matches\n except Exception as e:\n self.log(\"Exception looking for total_matches {}\".format(e), DEBUG)\n finally:\n self.total_matches = 0\n\n def _scrape_next_results_page_link(self, response):\n meta = response.meta\n current_page = meta.get('current_page', 1)\n total_matches = self._scrape_total_matches(response)\n results_per_page = self._scrape_results_per_page(response)\n if not results_per_page:\n results_per_page = 40\n if total_matches and current_page < math.ceil(total_matches / float(results_per_page)):\n current_page += 1\n st = response.meta['search_term']\n data = self.payload.copy()\n data['query'] = st\n data['skip'] = (current_page - 1) * 40\n meta['current_page'] = current_page\n return Request(\n url=self.SEARCH_URL, method=\"POST\", body=json.dumps(data), headers=self.headers, meta=meta)\n\n def _scrape_product_links(self, response):\n links = []\n try:\n contents = json.loads(response.body)\n for record in contents.get('records', []):\n link = 'https://www.bjs.com' + record.get('allMeta', {}).get('visualVariant')[0].get('nonvisualVariant', [])[0].get('product_url')\n links.append(link)\n except Exception as e:\n self.log(\"Exception looking for product links {}\".format(e), DEBUG)\n finally:\n for link in links:\n prod = SiteProductItem()\n prod['url'] = link\n prod_id = link.split('/')[-1]\n link = self.PRODUCT_URL.format(product_id=prod_id)\n yield link, prod\n\n @staticmethod\n def _parse_title(data):\n title = data.get('description', {}).get('name')\n return title\n\n def _parse_price(self, data):\n price = data.get('maximumItemPrice', {})\n if not price:\n price = data.get('bjsClubProduct', [{}])[0].get('clubItemStandardPrice', {})\n try:\n return Price(price=float(price.get('amount')), priceCurrency='USD') if price else None\n except:\n self.log('Error Parsing Price: {}'.format(traceback.format_exc()))\n\n @staticmethod\n def _parse_image(data):\n images = data.get('productImages', {}).get('fullImage')\n return images\n\n def _parse_categories(self, data):\n category_list = []\n try:\n categories_info = data.get('breadCrumbDetail')\n category_level = categories_info.get('Levels')\n for index in range(1, category_level + 1):\n category = categories_info.get('Level{}'.format(index)).split('||')[-1]\n category_list.append(category)\n return category_list\n except:\n self.log(\"Error while parsing categories {}\".format(traceback.format_exc()))\n\n @staticmethod\n def _search_attribute(attribute_name, data):\n if data.get('descriptiveAttributes'):\n for attr in data.get('descriptiveAttributes'):\n if attr.get('name') == attribute_name:\n return attr.get('attributeValueDataBeans', [{}])[0].get('value')\n\n def parse_product(self, response):\n meta = response.meta.copy()\n product = meta['product']\n\n try:\n data = json.loads(response.body_as_unicode())\n except:\n self.log('JSON not found or invalid JSON: {}'\n .format(traceback.format_exc()))\n product['not_found'] = True\n return product\n\n title = self._parse_title(data)\n if title is None:\n product[\"no_longer_available\"] = True\n return product\n cond_set_value(product, 'title', title)\n\n price = self._parse_price(data)\n cond_set_value(product, 'price', price)\n\n image_url = self._parse_image(data)\n cond_set_value(product, 'image_url', image_url)\n\n brand = guess_brand_from_first_words(product['title'])\n cond_set_value(product, 'brand', brand)\n\n if data.get('bjsitems', []):\n sku = data.get('bjsitems', [])[0].get('articleId')\n cond_set_value(product, 'sku', sku)\n cond_set_value(product, 'reseller_id', sku)\n\n model = data.get('manufacturerPartNumber')\n cond_set_value(product, 'model', model)\n\n upc = self._search_attribute('upc', data)\n cond_set_value(product, 'upc', upc)\n\n categories = self._parse_categories(data)\n cond_set_value(product, 'categories', categories)\n\n if categories:\n cond_set_value(product, 'department', categories[-1])\n\n # Available Online: 1 or 0 (1 = yes, 0 = no)\n if data.get('bjsClubProduct', []):\n online_avail = data.get('bjsClubProduct', [])[0].get('itemAvailableOnline', 'N')\n product['available_online'] = 1 if online_avail == 'Y' else 0\n\n # Available In-club(store): 1 or 0 (1 = yes, 0 = no)\n if data.get('bjsClubProduct', []):\n club_avail = data.get('bjsClubProduct', [])[0].get('itemAvailableInClub', 'N')\n product['available_store'] = 1 if club_avail == 'Y' else 0\n\n product['is_out_of_stock'] = str(data.get('description', {}).get('available')) == '0'\n\n product['is_in_store_only'] = str(product.get('available_online', None)) == '0' and str(\n product.get('available_store', None)) == '1'\n\n product['locale'] = \"en-US\"\n\n part_number = data.get('partNumber')\n\n if part_number:\n url = self.REVIEW_URL.format(part_num=part_number)\n return Request(url=url,\n callback=self._parse_reviews,\n meta={'product': product},\n headers={'authorization': '7c12e7e9-fe30-4e7a-bcb8-8376b9117a6b'},\n dont_filter=True)\n\n return product\n\n @staticmethod\n def _parse_reviews(response):\n meta = response.meta\n product = meta.get('product')\n cond_set_value(product, 'buyer_reviews', parse_powerreviews_buyer_reviews(response))\n\n return product\n","repo_name":"aprosdev/ecom-predictor","sub_path":"product-ranking/product_ranking/spiders/bjs.py","file_name":"bjs.py","file_ext":"py","file_size_in_byte":9618,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"61"} +{"seq_id":"22518833564","text":"import argparse\nimport sys\nfrom peheader import *\n\ndef start():\n parser = argparse.ArgumentParser()\n parser.add_argument('-v', '--version', action='version', version='version 0.0.1')\n parser.add_argument('-i', '--info',action=\"store_true\", help='PE文件头信息')\n parser.add_argument('filename')\n args = parser.parse_args()\n return args\n\nif __name__ == \"__main__\":\n args = start()\n if args.info or (len(sys.argv) == 2 and args.filename != None):\n d = {}\n r = open(args.filename,'rb')\n dosheader = r.read(0x40)\n ImageDosHeader = ImageDosHeader(dosheader)\n r.seek(ImageDosHeader.PEoffser(),0)\n ntheader = r.read(0xf0)\n ImageNtHeader = ImageNtHeader(ntheader)\n ImageNtHeader.show()","repo_name":"yifeng-lee/PEParser","sub_path":"PEparser.py","file_name":"PEparser.py","file_ext":"py","file_size_in_byte":761,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"74203215233","text":"#!/usr/local/bin/python3\n\n\ndef tag_bloco(conteudo, *args, classe='success', inline=False):\n tag = 'span' if inline else 'div'\n conteudo = conteudo if not callable(conteudo) else conteudo(*args)\n return f'<{tag} class={classe}>{conteudo}'\n\n\ndef tag_lista(*itens):\n lista = ''.join(f'
  • {item}
  • ' for item in itens)\n return f''\n\n\nif __name__ == '__main__':\n print(tag_bloco('teste1'))\n print(tag_bloco('teste2', inline=True))\n print(tag_bloco('teste3', classe='danger'))\n print(tag_lista('teste1', 'teste2'))\n print(tag_bloco(tag_lista, 'teste1', 'teste2', classe='danger'))\n","repo_name":"flaviogf/courses","sub_path":"coders/curso_python/funcoes/v4.py","file_name":"v4.py","file_ext":"py","file_size_in_byte":631,"program_lang":"python","lang":"fr","doc_type":"code","stars":4,"dataset":"github-code","pt":"61"} +{"seq_id":"34564710118","text":"import argparse\r\nimport os\r\nfrom dartel_pipeline import batched_spm12_dartel\r\n\r\n\r\nif __name__=='__main__':\r\n\r\n parser = argparse.ArgumentParser()\r\n parser.add_argument('--num_encoding_layers', type = int, default = 2, help = 'keep the default setting')\r\n parser.add_argument('--num_filters', type = int, default = 64, help = 'keep the default setting')\r\n parser.add_argument('--num_subjects', type = int, default = 2, help = 'keep the default setting')\r\n parser.add_argument('--num_voxels_per_subject', type = int , default = 1, help = 'keep the default setting')\r\n parser.add_argument('--filepath_csv', type = str, help = 'the location of the .csv file containing the meta-data assocated to the dataset in cause')\t\r\n parser.add_argument('--dirpath_raw_data', type = str, help = 'the location of the directory containing the raw T1 nifti files')\r\n parser.add_argument('--dataset_name', type = str, help = 'the name of the dataset in cause, it will influence where the results are written')\r\n parser.add_argument('--size_batch_preprocessing', type = int, help = 'how many nifti files to process at the same time')\r\n args = parser.parse_args()\r\n\r\n ##### spm12 pre-processing #####\r\n batched_spm12_dartel(img_dir = args.dirpath_raw_data, name_of_dataset = args.dataset_name, size_batch = args.size_batch_preprocessing)\r\n\r\n\r\n ##### getting LocalBrainAge predictions #####\r\n cmd = 'python3 ./LocalBrainAge_testing.py --filepath_csv='+str(args.filepath_csv)+' --dirpath_gm='+str(args.dirpath_raw_data)+'/gm_data --dirpath_wm='+str(args.dirpath_raw_data)+'/wm_data --dataset_name='+str(args.dataset_name)","repo_name":"SebastianPopescu/U-NET-for-LocalBrainAge-prediction","sub_path":"full_testing_script.py","file_name":"full_testing_script.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"61"} +{"seq_id":"32759738878","text":"##parameters=datastructure\n\"\"\"\nDo the necessary rendering or redirection after an entry has been\nsuccessfully created and filled with the initial values by the user.\n\nThe context is the directory.\n\nMay return a rendered document, or do a redirect.\n\"\"\"\n\nfrom urllib import urlencode\n\ndirname = context.getId()\nid_field = context.id_field\nid = datastructure.getDataModel()[id_field]\n\nportal_url = context.portal_url()\nargs = urlencode({'dirname': dirname,\n 'id': id,\n 'portal_status_message': 'psm_entry_created',\n })\naction_path = 'taskdirectory_entry_edit_form?'+args\ncontext.REQUEST.RESPONSE.redirect('%s/%s' % (portal_url, action_path))\n","repo_name":"nuxeo-cps/products--CPSTaskManager","sub_path":"skins/task_directory/taskdirectory_entry_created.py","file_name":"taskdirectory_entry_created.py","file_ext":"py","file_size_in_byte":690,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"21872380766","text":"#Document Enhancer\r\n\r\nimport docx\r\nfrom docx import Document\r\nfrom docx.shared import Inches\r\nfrom docx.shared import Pt\r\nimport random\r\nfrom docx.shared import RGBColor\r\nfrom docx.enum.text import WD_ALIGN_PARAGRAPH\r\nfrom docx.enum.style import WD_STYLE_TYPE\r\nimport re\r\n\r\nimport os\r\nimport sys\r\nimport torch\r\nimport argparse\r\nimport numpy as np\r\nfrom gpt2Pytorch.GPT2.model import (GPT2LMHeadModel)\r\nfrom gpt2Pytorch.GPT2.utils import load_weight\r\nfrom gpt2Pytorch.GPT2.config import GPT2Config\r\nfrom gpt2Pytorch.GPT2.sample import sample_sequence\r\nfrom gpt2Pytorch.GPT2.encoder import get_encoder\r\n\r\nfrom gpt2Pytorch.mainLib import *\r\n\r\n\r\ndocument = Document(\"test.docx\")\r\ndocument = Document()\r\n\r\ndef formatStyles():\r\n style = document.styles['Normal']\r\n font = style.font\r\n font.name = 'Times New Roman'\r\n font.size = Pt(12)\r\n font.color.rgb = RGBColor(0,0,0)\r\n font.underline = False\r\n style.paragraph_format.line_spacing = 2\r\n\r\n styles = document.styles\r\n styles['Title'].delete()\r\n style = styles.add_style('Title', WD_STYLE_TYPE.PARAGRAPH)\r\n\r\n style = document.styles['Title']\r\n font = style.font\r\n font.name = 'Times New Roman'\r\n font.size = Pt(20)\r\n font.color.rgb = RGBColor(0,0,0)\r\n font.underline = False\r\n style.paragraph_format.line_spacing = 2\r\n\r\n\r\ndef addheader():\r\n section = document.sections[0]\r\n header = section.header\r\n head = header.paragraphs[0]\r\n list = readParagraph(0)\r\n head.text = list.split( )[1]\r\n head.alignment = WD_ALIGN_PARAGRAPH.RIGHT\r\n\r\n\r\ndef addheading():\r\n list = readHeading()\r\n heading = document.add_paragraph(list[0])\r\n heading = document.add_paragraph(list[1])\r\n heading = document.add_paragraph(headingDate(list[2]))\r\n heading = document.add_paragraph(list[3])\r\n\r\n heading.alignment = WD_ALIGN_PARAGRAPH.LEFT\r\n heading.paragraph_format.line_spacing = 2\r\n\r\n\r\ndef addtitle():\r\n title = document.add_paragraph(readParagraph(4))\r\n title.alignment = WD_ALIGN_PARAGRAPH.CENTER\r\n\r\n\r\ndef addbody():\r\n list = readBody()\r\n for i in list:\r\n paragraph = document.add_paragraph('\\t' + i)\r\n for paragraph_text in AIconverter(readParagraph(5)).split('\\n\\n'):\r\n paragraph = document.add_paragraph(\"\\t\"+paragraph_text.strip())\r\n\r\n paragraph_format = paragraph.paragraph_format\r\n paragraph_format.line_spacing = 2\r\n\r\n\r\ndef readFile():\r\n file = open(\"test.txt\", \"r\")\r\n content = file.read()\r\n\r\n\r\ndef readHeading():\r\n doc = docx.Document('test.docx')\r\n headingList = [doc.paragraphs[0].text, doc.paragraphs[1].text, doc.paragraphs[2].text, doc.paragraphs[3].text]\r\n return headingList;\r\n\r\n\r\ndef readParagraph(paragraph):\r\n doc = docx.Document('test.docx')\r\n return doc.paragraphs[paragraph].text\r\n\r\n\r\ndef headingDate(date):\r\n try:\r\n day = re.search(\"([^\\d])([0-2]|[0-2][0-9])([^\\d])\" , \" \"+date+\" \")\r\n print(day.group())\r\n year = re.search(\"[2-9][0-9][0-9][0-9]\" , date)\r\n print(year.group())\r\n month = re.search(\"[^\\s\\d][^\\s\\d][^\\s\\d]\" , date)\r\n print(month.group())\r\n newdate=day.group()[1:-1]+\" \"+month.group().capitalize()+\". \"+year.group()\r\n print(newdate)\r\n except:\r\n print(\"error\")\r\n newdate=date\r\n return newdate\r\n\r\n\r\ndef readBody():\r\n doc = docx.Document('test.docx')\r\n list = []\r\n i=len(doc.paragraphs)-1\r\n while i < len(doc.paragraphs):\r\n list.append(doc.paragraphs[i].text)\r\n i+=1\r\n return list\r\n\r\n\r\n\r\n\r\n#MAIN\r\ndef main():\r\n formatStyles()\r\n addheader()\r\n addheading()\r\n addtitle()\r\n addbody()\r\n document.save('main2.docx')\r\n\r\nmain()\r\n\r\n\r\n\r\n\r\n\r\ndocumentold = Document()\r\nparagraph = 0\r\nword = 0\r\n\r\n#returns true if a word is ignorable and false if important\r\ndef ignorable(word):\r\n ignore = [\"The\", \"the\", \"To\", \"to\", \"Of\", \"of\", \"Be\", \"be\", \"and\", \"A\", \"a\", \"That\", \"that\", \"Have\", \"have\", \"I\",\r\n \"It\", \"it\", \"For\", \"for\", \"Not\", \"not\", \"With\", \"with\", \"You\", \"you\", \"As\", \"as\", \"Do\", \"do\", \"At\", \"at\"\r\n \"This\", \"this\", \"By\", \"by\", \"or\", \"An\", \"an\", \"From\", \"from\", \"Will\", \"will\", \"Is\", \"is\"]\r\n for x in range( 0, len(ignore) ):\r\n if(word == ignore[x]):\r\n return True\r\n return False\r\n\r\n#takes in a string and returns a dictionary on the word count of each word\r\ndef getRepetitive( text ):\r\n unique = {}\r\n\r\n for word in text:\r\n if ignorable(word) == False:\r\n if len(unique) == 0:\r\n unique[word] = 1\r\n else:\r\n for y in list(unique):\r\n if(word == y):\r\n unique[word] += 1\r\n break\r\n unique[word] = 1\r\n return unique\r\n","repo_name":"sethrodg/doc-enhancer","sub_path":"main2.py","file_name":"main2.py","file_ext":"py","file_size_in_byte":4730,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"2077531569","text":"# multiAgents.py\n# --------------\n# Licensing Information: You are free to use or extend these projects for\n# educational purposes provided that (1) you do not distribute or publish\n# solutions, (2) you retain this notice, and (3) you provide clear\n# attribution to UC Berkeley, including a link to http://ai.berkeley.edu.\n# \n# Attribution Information: The Pacman AI projects were developed at UC Berkeley.\n# The core projects and autograders were primarily created by John DeNero\n# (denero@cs.berkeley.edu) and Dan Klein (klein@cs.berkeley.edu).\n# Student side autograding was added by Brad Miller, Nick Hay, and\n# Pieter Abbeel (pabbeel@cs.berkeley.edu).\n\nimport sys\nfrom util import manhattanDistance\nfrom game import Directions\nfrom game import Actions\nimport random, util\n\nfrom game import Agent\n\ndef closestFoodOrComputeDistance(curP, destP, gameState):\n if curP == destP: return None, 0\n walls = gameState.getWalls()\n foodList = gameState.getFood().asList()\n count = 0\n points = [curP,]\n queue = util.Queue()\n queue.push(curP)\n while not queue.isEmpty():\n count += 1\n qsize = len(queue.list)\n for i in range(qsize):\n temp = queue.pop()\n for direction in [Directions.WEST, Directions.NORTH, Directions.EAST, Directions.SOUTH]:\n x,y = temp\n dx, dy = Actions.directionToVector(direction)\n nextx, nexty = int(x + dx), int(y + dy)\n if not walls[nextx][nexty]:\n if (nextx, nexty) not in points:\n queue.push((nextx, nexty))\n points.append((nextx, nexty))\n \n if destP == None:\n if (nextx, nexty) in foodList:\n return (nextx, nexty), count\n elif (nextx, nexty) == destP:\n return None, count\n return None, None\n\nclass ReflexAgent(Agent):\n \"\"\"\n A reflex agent chooses an action at each choice point by examining\n its alternatives via a state evaluation function.\n\n The code below is provided as a guide. You are welcome to change\n it in any way you see fit, so long as you don't touch our method\n headers.\n \"\"\"\n\n\n def getAction(self, gameState):\n \"\"\"\n You do not need to change this method, but you're welcome to.\n\n getAction chooses among the best options according to the evaluation function.\n\n Just like in the previous project, getAction takes a GameState and returns\n some Directions.X for some X in the set {North, South, West, East, Stop}\n \"\"\"\n # Collect legal moves and successor states\n legalMoves = gameState.getLegalActions()\n\n # Choose one of the best actions\n scores = [self.evaluationFunction(gameState, action) for action in legalMoves]\n bestScore = max(scores)\n bestIndices = [index for index in range(len(scores)) if scores[index] == bestScore]\n chosenIndex = random.choice(bestIndices) # Pick randomly among the best\n\n \"Add more of your code here if you want to\"\n return legalMoves[chosenIndex]\n\n def evaluationFunction(self, currentGameState, action):\n \"\"\"\n Design a better evaluation function here.\n\n The evaluation function takes in the current and proposed successor\n GameStates (pacman.py) and returns a number, where higher numbers are better.\n\n The code below extracts some useful information from the state, like the\n remaining food (newFood) and Pacman position after moving (newPos).\n newScaredTimes holds the number of moves that each ghost will remain\n scared because of Pacman having eaten a power pellet.\n\n Print out these variables to see what you're getting, then combine them\n to create a masterful evaluation function.\n \"\"\"\n # Useful information you can extract from a GameState (pacman.py)\n successorGameState = currentGameState.generatePacmanSuccessor(action)\n newPos = successorGameState.getPacmanPosition()\n newFood = successorGameState.getFood()\n newGhostStates = successorGameState.getGhostStates()\n newScaredTimes = [ghostState.scaredTimer for ghostState in newGhostStates]\n\n \"*** YOUR CODE HERE ***\"\n return successorGameState.getScore()\n\ndef scoreEvaluationFunction(currentGameState):\n \"\"\"\n This default evaluation function just returns the score of the state.\n The score is the same one displayed in the Pacman GUI.\n\n This evaluation function is meant for use with adversarial search agents\n (not reflex agents).\n \"\"\"\n return currentGameState.getScore()\n\nclass MultiAgentSearchAgent(Agent):\n \"\"\"\n This class provides some common elements to all of your\n multi-agent searchers. Any methods defined here will be available\n to the MinimaxPacmanAgent, AlphaBetaPacmanAgent & ExpectimaxPacmanAgent.\n\n You *do not* need to make any changes here, but you can if you want to\n add functionality to all your adversarial search agents. Please do not\n remove anything, however.\n\n Note: this is an abstract class: one that should not be instantiated. It's\n only partially specified, and designed to be extended. Agent (game.py)\n is another abstract class.\n \"\"\"\n\n def __init__(self, evalFn = 'scoreEvaluationFunction', depth = '2'):\n self.index = 0 # Pacman is always agent index 0\n self.evaluationFunction = util.lookup(evalFn, globals())\n self.depth = int(depth)\n\nclass MinimaxAgent(MultiAgentSearchAgent):\n \"\"\"\n Your minimax agent (question 2)\n \"\"\"\n def maxVal(self,state,agent,depth): #pacman's turn get the max val\n val = -999999\n actions = state.getLegalActions(agent)\n if Directions.STOP in actions:\n actions.remove(Directions.STOP)\n for action in actions:\n val = max(val,self.minimax(state.generateSuccessor(agent,action),agent+1,depth+1))\n return val\n\n def minVal(self,state,agent,depth): #ghosts' turn to get the min val\n val = 999999\n for action in state.getLegalActions(agent):\n val = min(val,self.minimax(state.generateSuccessor(agent,action),agent+1,depth+1))\n return val\n\n def minimax(self,gameState,agent,depth):\n score = 0\n if agent == self.agentCount: ##it's pacman's turn\n agent = self.index\n if depth == self.depth*self.agentCount or gameState.isWin() or gameState.isLose():## finish?\n score = self.evaluationFunction(gameState)\n elif agent == self.index: ## judge again, is pacman's turn?\n score = self.maxVal(gameState,agent,depth)\n else: ##ghosts's turn\n score = self.minVal(gameState,agent,depth)\n return score\n\n def getAction(self, gameState):\n \"\"\"\n Returns the minimax action from the current gameState using self.depth\n and self.evaluationFunction.\n\n Here are some method calls that might be useful when implementing minimax.\n\n gameState.getLegalActions(agentIndex):\n Returns a list of legal actions for an agent\n agentIndex=0 means Pacman, ghosts are >= 1\n\n gameState.generateSuccessor(agentIndex, action):\n Returns the successor game state after an agent takes an action\n\n gameState.getNumAgents():\n Returns the total number of agents in the game\n \"\"\"\n depth = 0\n agentIndex = self.index\n Dict = {}\n self.agentCount = gameState.getNumAgents()\n actions = gameState.getLegalActions(agentIndex)\n if Directions.STOP in actions:\n actions.remove(Directions.STOP)\n for action in actions:\n eval_f = self.minimax(gameState.generateSuccessor(agentIndex,action),agentIndex+1,depth+1)\n Dict[eval_f] = action\n choices = Dict[max(Dict)]\n return choices\n\n\n# class MinimaxAgent(MultiAgentSearchAgent):\n# \"\"\"\n# Your minimax agent (question 2)\n# \"\"\"\n\n# def getAction(self, gameState):\n# \"\"\"\n# Returns the minimax action from the current gameState using self.depth\n# and self.evaluationFunction.\n\n# Here are some method calls that might be useful when implementing minimax.\n\n# gameState.getLegalActions(agentIndex):\n# Returns a list of legal actions for an agent\n# agentIndex=0 means Pacman, ghosts are >= 1\n\n# gameState.generateSuccessor(agentIndex, action):\n# Returns the successor game state after an agent takes an action\n\n# gameState.getNumAgents():\n# Returns the total number of agents in the game\n# \"\"\"\n# \"*** YOUR CODE HERE ***\"\n# # return self.dfMiniMaxSearch(self.depth, gameState, True)[1]\n\n# def dfMiniMax(depth, curState, agentId):\n# if depth == 0 or curState.isWin() or curState.isLose():\n# return self.evaluationFunction(curState)\n\n# legalActions = curState.getLegalActions(agentId)\n# if agentId == 0:\n# maxScore = -sys.maxint\n# for action in legalActions:\n# nextState = curState.generateSuccessor(agentId, action)\n# maxScore = max(maxScore, dfMiniMax(depth, nextState, agentId+1))\n# return maxScore\n# else :\n# minScore = sys.maxint\n# for action in legalActions:\n# nextState = curState.generateSuccessor(agentId, action)\n# if agentId+1 == curState.getNumAgents():\n# minScore = min(minScore, dfMiniMax(depth-1, nextState, 0))\n# else: minScore = min(minScore, dfMiniMax(depth, nextState, agentId+1))\n# return minScore\n\n# legalActions = gameState.getLegalActions(0)\n# childList = [gameState.generateSuccessor(0, action) for action in legalActions]\n# scores = [dfMiniMax(self.depth, nextState, 1) for nextState in childList]\n# bestScore = max(scores)\n# bestIndices = [index for index in range(len(scores)) if scores[index] == bestScore]\n# # print dir(gameState.problem)\n# # closestFood = closestFoodOrComputeDistance(gameState.getPacmanPosition(), None, gameState)\n# # dists = [closestFoodOrComputeDistance(nextState.getPacmanPosition(), closestFood[0], nextState)[1] \\\n# # for nextState in childList]\n# # chosenIndex = min(bestIndices, key=lambda x: dists[x])\n# random.seed()\n# chosenIndex = random.choice(bestIndices)\n# return legalActions[chosenIndex]\n\n# util.raiseNotDefined()\n\n# # consider only pacman, see all ghosts as min\n# # def dfMiniMaxSearch(self, depth, curState, pacmanTurn):\n# # if depth == 0 or curState.isWin() or curState.isLose():\n# # return curState.getScore(), None\n# # legalActions = curState.getLegalActions()\n# # successors = [curState.generateSuccessor(0, action) for action in legalActions]\n# # scores = [self.dfMiniMaxSearch(depth-1, nextState, not pacmanTurn)[0] for nextState in successors]\n# # bestScore = max(scores) if pacmanTurn else min(scores)\n# # bestIndices = [index for index in range(len(scores)) if scores[index] == bestScore]\n# # closestFood = closestFoodOrComputeDistance(curState.getPacmanPosition(), None, curState)\n# # dists = [closestFoodOrComputeDistance(nextState.getPacmanPosition(), closestFood[0], nextState)[1] \\\n# # for nextState in successors]\n# # chosenIndex = min(bestIndices, key=lambda x: dists[x])\n# # # chosenIndex = random.choice(bestIndices) # Pick randomly among the best\n# # return bestScore, legalActions[chosenIndex]\n\n\nclass AlphaBetaAgent(MultiAgentSearchAgent):\n \"\"\"\n Your minimax agent with alpha-beta pruning (question 3)\n \"\"\"\n \n def getAction(self, gameState):\n \"\"\"\n Returns the minimax action using self.depth and self.evaluationFunction\n \"\"\"\n \"*** YOUR CODE HERE ***\"\n # return self.alphabeta(self.depth, gameState, -sys.maxint, sys.maxint, True)[1]\n \n def dfAlphaBeta(depth, curState, agentId, alpha, beta):\n legalActions = curState.getLegalActions(agentId)\n if depth == 0 or curState.isWin() or curState.isLose():\n return curState.getScore()\n if agentId == 0:\n v = -sys.maxint\n for action in legalActions:\n nextState = curState.generateSuccessor(agentId, action)\n v = max(v, dfAlphaBeta(depth, nextState, agentId+1, alpha, beta))\n alpha = max(alpha, v)\n if alpha > beta: break\n return v\n else :\n v = sys.maxint\n for action in legalActions:\n nextState = curState.generateSuccessor(agentId, action)\n if agentId+1 == curState.getNumAgents():\n v = min(v, dfAlphaBeta(depth-1, nextState, 0, alpha, beta))\n else: v = min(v, dfAlphaBeta(depth, nextState, agentId+1, alpha, beta))\n beta = min(beta, v)\n if alpha > beta: break\n return v\n\n bestIndices = []\n legalActions = gameState.getLegalActions(0)\n alpha, beta = -sys.maxint, sys.maxint\n for i, action in enumerate(legalActions):\n nextState = gameState.generateSuccessor(0, action)\n newAlpha = dfAlphaBeta(self.depth, nextState, 1, alpha, beta)\n if alpha < newAlpha:\n alpha, bestIndices = newAlpha, [i]\n elif alpha == newAlpha:\n bestIndices.append(i)\n \n random.seed()\n chosenIndex = random.choice(bestIndices)\n return legalActions[chosenIndex]\n\n util.raiseNotDefined()\n\n # consider only pacman, see all ghosts as min\n # def alphabeta(self, depth, curState, alpha, beta, pacmanTurn):\n # if depth == 0 or curState.isWin() or curState.isLose():\n # return curState.getScore(), None\n # legalActions = curState.getLegalActions()\n # successors = [curState.generateSuccessor(0, action) for action in legalActions]\n # bestIndices = []\n # if pacmanTurn:\n # for i, nextState in enumerate(successors):\n # newAlpha = self.alphabeta(depth-1,nextState,alpha,beta,not pacmanTurn)[0]\n # if alpha < newAlpha:\n # alpha, bestIndices = newAlpha, [i]\n # elif alpha == newAlpha:\n # bestIndices.append(i)\n # if beta < alpha: break\n # if depth != self.depth: return alpha, None\n # else :\n # for i, nextState in enumerate(successors):\n # newBeta = self.alphabeta(depth-1,nextState,alpha,beta,not pacmanTurn)[0]\n # if beta > newBeta:\n # beta, bestIndices = newBeta, [i]\n # elif beta == newBeta:\n # bestIndices.append(i)\n # if beta < alpha: break\n # if depth != self.depth: return beta, None\n # closestFood = closestFoodOrComputeDistance(curState.getPacmanPosition(), None, curState)\n # dists = [closestFoodOrComputeDistance(nextState.getPacmanPosition(), closestFood[0], nextState)[1] \\\n # for nextState in successors]\n # if len(bestIndices) == 0:\n # bestIndices = [i for i in range(len(legalActions))]\n # chosenIndex = min(bestIndices, key=lambda x: dists[x])\n # if pacmanTurn: return alpha, legalActions[chosenIndex]\n # else : return beta, legalActions[chosenIndex]\n\n\nclass ExpectimaxAgent(MultiAgentSearchAgent):\n \"\"\"\n Your expectimax agent (question 4)\n \"\"\"\n\n def getAction(self, gameState):\n \"\"\"\n Returns the expectimax action using self.depth and self.evaluationFunction\n\n All ghosts should be modeled as choosing uniformly at random from their\n legal moves.\n \"\"\"\n \"*** YOUR CODE HERE ***\"\n util.raiseNotDefined()\n\ndef betterEvaluationFunction(currentGameState):\n \"\"\"\n Your extreme ghost-hunting, pellet-nabbing, food-gobbling, unstoppable\n evaluation function (question 5).\n\n DESCRIPTION: \n \"\"\"\n \"*** YOUR CODE HERE ***\"\n util.raiseNotDefined()\n\n# Abbreviation\nbetter = betterEvaluationFunction\n\n","repo_name":"shinegrin/Pacman","sub_path":"multiagent/multiAgents.py","file_name":"multiAgents.py","file_ext":"py","file_size_in_byte":16702,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"3735117762","text":"# -*- coding: utf-8 -*-\nfrom setuptools import setup, find_packages\n\nwith open('requirements.txt') as f:\n\tinstall_requires = f.read().strip().split('\\n')\n\n# get version from __version__ variable in purpletheme/__init__.py\nfrom purpletheme import __version__ as version\n\nsetup(\n\tname='purpletheme',\n\tversion=version,\n\tdescription='ERPNext app for Vezolve',\n\tauthor='Ayeshka Abeysinghe',\n\tauthor_email='ayeshka@vezolve.com',\n\tpackages=find_packages(),\n\tzip_safe=False,\n\tinclude_package_data=True,\n\tinstall_requires=install_requires\n)\n","repo_name":"ayeshkaVezolve/vezolve_purpletheme","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":532,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"3058001602","text":"import inflection\n\n\ndef create_expressions(requested_fields: list) -> tuple[str, dict]:\n \"\"\"\n Creates expression components for a dynamo query\n :param requested_fields: List of enum fields names\n\n example usage:\n requested_fields = [\"ID\", \"Created\", \"FileName\"]\n projection_expression, expression_attribute_names = create_expressions(requested_fields)\n\n result:\n [\n \"#ID_attr,#Created_attr,#FileName_attr\",\n {\"#ID_attr\": \"ID\", \"#Created_attr\": \"Created\", \"#FileName_attr\": \"FileName\"}\n ]\n \"\"\"\n projection_expression = \"\"\n expression_attribute_names = {}\n\n for field_definition in requested_fields:\n field_placeholder = create_expression_attribute_placeholder(field_definition)\n if len(projection_expression) > 0:\n projection_expression = f\"{projection_expression},{field_placeholder}\"\n else:\n projection_expression = field_placeholder\n\n expression_attribute_names[field_placeholder] = field_definition\n return projection_expression, expression_attribute_names\n\n\ndef create_attribute_filter(filtered_fields: dict) -> str:\n \"\"\"\n Creates a filter for dynamodb queries for existing and non-existing attributes\n :param filtered_fields: Dictionary of filtered fields\n\n example usage:\n fields_filter = {\n DocumentReferenceMetadataFields.DELETED.value: \"\",\n DocumentReferenceMetadataFields.FILENAME.value: \"Test Filename\"\n }\n attribute_filter = create_attribute_filter(fields_filter)\n\n result:\n \"attribute_not_exists(Deleted) OR Deleted = :Deleted_val AND \"\n \"Filename = :Filename_val\"\n\n \"\"\"\n attr_filter = \"\"\n\n for field_name, field_value in filtered_fields.items():\n base_filter_string = (\n f\"{field_name} = {create_expression_value_placeholder(field_name)}\"\n )\n if not field_value:\n filter_string = (\n f\"attribute_not_exists({field_name}) OR {base_filter_string}\"\n )\n else:\n filter_string = base_filter_string\n\n if not attr_filter:\n attr_filter = filter_string\n else:\n attr_filter += f\" AND {filter_string}\"\n\n return attr_filter\n\n\ndef create_update_expression(field_names: list) -> str:\n \"\"\"\n Creates an expression for dynamodb queries to SET a new value for an item\n :param field_names: List of fields to update\n\n example usage:\n field_names = [\"Name\", \"Age\"...]\n fields_filter = create_update_expression(field_names)\n\n result:\n \"SET #Name_attr = :Name_val, #Age_attr = :Age_val\"\n\n \"\"\"\n update_expression = \"SET\"\n for field in field_names:\n expression = f\" {create_expression_attribute_placeholder(field)} = {create_expression_value_placeholder(field)}\"\n if update_expression == \"SET\":\n update_expression += expression\n else:\n update_expression += f\",{expression}\"\n\n return update_expression\n\n\ndef create_expression_attribute_values(attribute_field_values: dict) -> dict:\n \"\"\"\n Maps a dict of expression names and expression values to create a dictionary to pass into query\n :param attribute_field_values: Dictionary of attribute field names and values\n\n example usage:\n attribute_field_values = {\n DocumentReferenceMetadataFields.DELETED.value: \"\",\n DocumentReferenceMetadataFields.FILENAME.value: \"Test Filename\"\n }\n expression_attribute_values = create_expression_attribute_values(attribute_field_values)\n\n result:\n {\n \":Deleted_val\" : \"\"\n \":FileName_val\" : \"Test Filename\"\n }\n \"\"\"\n expression_attribute_values = {}\n for field_name, field_value in attribute_field_values.items():\n expression_attribute_values[\n f\"{create_expression_value_placeholder(field_name)}\"\n ] = field_value\n\n return expression_attribute_values\n\n\ndef create_expression_value_placeholder(value: str) -> str:\n \"\"\"\n Creates a placeholder value for an expression attribute name\n :param value: Value to change into a placeholder\n\n example usage:\n placeholder = create_expression_value_placeholder(\"VirusScanResult\")\n\n result:\n \":VirusScanResult_val\"\n \"\"\"\n return f\":{inflection.camelize(value, uppercase_first_letter=True)}_val\"\n\n\ndef create_expression_attribute_placeholder(value: str) -> str:\n \"\"\"\n Creates a placeholder value for a projection attribute name\n :param value: Value to change into a placeholder\n\n example usage:\n placeholder = create_expression_attribute_placeholder(\"VirusScanResult\")\n\n result:\n \"#VirusScanResult_attr\"\n \"\"\"\n return f\"#{inflection.camelize(value, uppercase_first_letter=True)}_attr\"\n","repo_name":"nhsconnect/national-document-repository","sub_path":"lambdas/utils/dynamo_utils.py","file_name":"dynamo_utils.py","file_ext":"py","file_size_in_byte":4849,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"32626067466","text":"import httplib2\r\nfrom json import dumps\r\n\r\nh = httplib2.Http(disable_ssl_certificate_validation=True)\r\nh.add_credentials('scoblrtest1@cldc.shoretel.com', 'Shor1234')\r\n#h.add_certificate(\"89458dc4c867bd29dbd76bab0570af6ab3b968ed\", \"C:\\\\Users\\\\pkumarasami\\\\Desktop\\\\ShoreTel\\\\Work Space\\\\Sprint 1\\\\shoretel_engineering_ca.crt\", \"cldc.shoretel.com\")\r\n#body = {'USERNAME': 'scoblrtest1@cldc.shoretel.com', 'PASSWORD': 'Shore1234'}\r\nheaders = {'Content-type': 'application/json;charset=utf-8','Accept':'application/json,text/plain'}\r\n\r\nbody = {\"fn\": \"pannaga\",\"n\": \"jayaram\"}\r\n#response, content = h.request(\"https://buddycloud.cldc.shoretel.com/api/profile\", \"PUT\")\r\nurl = 'https://buddycloud.cldc.shoretel.com/api/workspaces'\r\ndictionary = {\"channel\": \"Parasuram221\",\"title\": \"scoe parasu workspace2 2\"}\r\n\r\nresp, content = h.request(\r\n uri=url,\r\n method='POST',\r\n headers={'Content-Type': 'application/json'},\r\n body=dumps(dictionary),\r\n)\r\nprint (resp)\r\nprint (content)","repo_name":"parasuramankumarasami/Examples","sub_path":"Workspace REST API/CreateWorkspace.py","file_name":"CreateWorkspace.py","file_ext":"py","file_size_in_byte":994,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"17199692189","text":"from PyQt5.QtWidgets import QWidget\nfrom PyQt5.QtGui import (QColor, QFont, QPainter, QPalette, QPen, QBrush)\nfrom PyQt5.QtCore import Qt\n\nclass PowerShow(QWidget):\n \"\"\"docstring for PowerShow\"\"\"\n def __init__(self):\n super(PowerShow, self).__init__()\n # self.arg = arg\n\n self.setPalette(QPalette(QColor(239,246,250)))\n self.setAutoFillBackground(True)\n self.setGeometry(100,100,100,100)\n self.setMinimumSize(100, 100)\n # self.show()\n self.pter = QPainter(self)\n self.powerList = {'logNumber':0,\n 'currentPower':0,\n 'averagePower':0,\n 'variancePower':0,\n 'maxPower':0,\n 'minPower':0}\n\n def paintEvent(self,event):\n pter = self.pter\n self.text = '最大功率:'+self.__Power2str(self.powerList['maxPower'])+'\\n\\\n最小功率:'+self.__Power2str(self.powerList['minPower'])+'\\n\\\n平均功率:'+self.__Power2str(self.powerList['averagePower'])+'\\n\\\n功率方差:'+str(round(self.powerList['variancePower'],2))+'\\n'\n self.textshow = self.__Power2str(self.powerList['currentPower'])\n pter.begin(self)\n # print('PowerShowlist',self.text ,'\\n',self.textshow)\n pter.setPen(QPen(Qt.black,0.1))\n pter.setBrush(QBrush(QColor(125,185,222)))\n pter.drawRoundedRect(event.rect(), 10, 10)\n pter.translate(10,10)\n self.drawPowerText(event,pter)\n # pter.drawRoundedRect(20,20, 210, 160,50,50)\n pter.translate(130,2)\n self.drawPowershishiText(event, pter)\n pter.translate(-5,20)\n self.drawPowerCurrentText(event, pter)\n\n pter.end()\n\n def drawPowerText(self,event,qp):\n qp.setPen(Qt.white)\n qp.setFont(QFont('微软雅黑', 10))\n qp.drawText(event.rect(), Qt.RightToLeft, self.text)\n\n def drawPowerCurrentText(self,event,qp):\n qp.setPen(Qt.white)\n qp.setFont(QFont('微软雅黑', 25))\n qp.drawText(event.rect(), Qt.RightToLeft, self.textshow)\n\n def drawPowershishiText(self,event,qp):\n qp.setPen(Qt.white)\n qp.setFont(QFont('微软雅黑', 8))\n qp.drawText(event.rect(), Qt.RightToLeft, '实时:')\n\n def updateFigure(self):\n self.update()\n\n def __Power2str(self,data):\n if data > 0.1:\n return str(round(data,2))+'W'\n else:\n return str(round(data*1000,2)) + 'mW'\n\n\nif __name__ == '__main__':\n import sys\n from PyQt5.QtWidgets import QApplication\n app = QApplication(sys.argv)\n addressBook = PowerShow()\n addressBook.show()\n\n sys.exit(app.exec_())\n","repo_name":"lidingke/photodarkening","sub_path":"view/powershow.py","file_name":"powershow.py","file_ext":"py","file_size_in_byte":2609,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"12476372232","text":"def sevens_in_a_row(arr, n):\n\tif n==1:\n\t\tfor item in arr:\n\t\t\tif item==7:\n\t\t\t\treturn True\n\t\t\telse:\n\t\t\t\treturn False\n\tbr=1\n\tflag=False\n\tfor i in range(0,len(arr)-1):\n\t\tif arr[i]==7:\n\t\t\tif arr[i]==arr[i+1]:\n\t\t\t\tbr+=1\n\t\t\t\tif br>=n:\n\t\t\t\t\tflag = True\n\t\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tbr=1\n\treturn flag\n","repo_name":"mileto94/HackBulgaria","sub_path":"week0/First/6-SevensRow/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":287,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23560901961","text":"\r\n\"\"\"\r\nGiven a number N, break it down into an array\r\n\r\nN = '999' outputs\r\na = [[1,1,1,1,1,1,1,1,1],\r\n [1,1,1,1,1,1,1,1,1],\r\n [1,1,1,1,1,1,1,1,1]]\r\nIt's just magic\r\n\"\"\"\r\ndef magic(N):\r\n\r\n bar = []\r\n\r\n for i in range(len(N)):\r\n bar.append([0]*9)\r\n for j in range(0,len(N)):\r\n x = int(N[j])\r\n for i in range(0,x):\r\n bar[j][i] = 1\r\n return bar, len(N)\r\n\r\ndef help(s): # reduces the order by one decimal, for example 100 to 10\r\n bar = [0]*9\r\n if s == 0:\r\n bar = [1]*9\r\n else:\r\n for i in range(0,s-1):\r\n bar[i]=1\r\n return bar\r\n\r\ndef unmagic(bar):\r\n s = 0\r\n tidy_N = \"\"\r\n tiding_N = \"\"\r\n for i in range(0, len(bar)):\r\n s = sum(bar[i])\r\n tiding_N += str(s)\r\n tidy_N = int(tiding_N)\r\n return tidy_N\r\n\r\ndef wizard(bar, index):\r\n aux_bar = bar[:]\r\n s=0\r\n flag = 0\r\n while(flag < index ):\r\n for i in range(0, index):\r\n if aux_bar[index-i-1] < aux_bar[index-i-2] and i!=index-1:\r\n aux_bar[index-i-1] = [1]*9\r\n aux_bar[index-i-2] = help(sum(bar[index-i-2]))\r\n flag = 0\r\n elif aux_bar[index-i-1] >= aux_bar[index-i-2]:\r\n flag += 1\r\n\r\n return aux_bar\r\n\r\n\r\norder = []\r\nwith open(\"input.txt\") as f:\r\n file_o = open('output.txt','w') #GET IT OPEN HERE\r\n input = f.readlines()\r\n input_x = [i.replace(\"\\n\", \"\") for i in input]\r\n test_size = int( input_x[0] )\r\n\r\n if test_size == len(input_x[1:]):\r\n for j in range(0, test_size):\r\n order.extend(input_x[j+1].split(\" \"))\r\n print(\"test \",order , len(order))\r\n bar = []\r\n index = []\r\n for i in range(0, len(order)): # The input is now ready for some magic\r\n bar, index = magic(order[i])\r\n untidy = wizard(bar, index)\r\n tidy = unmagic(untidy)\r\n print(\"tidy\", tidy)\r\n file_o.write(\"Case #\"+ str(i+1) +\": \"+ str(tidy) +\"\\n\")\r\n else:\r\n print (\"Input count error\")\r\n file_o.close()\r\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_200/4210.py","file_name":"4210.py","file_ext":"py","file_size_in_byte":2056,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"17481563832","text":"from preprocessing import *\nimport os\nimport numpy as np\n\n\"\"\"\n Before running the script, make sure to have the scans in the structure\n below. Names may vary:\n\n input:\n training scans - ./data/name_of_data/train/*.nii.gz\n test scans - ./data/name_of_data/test/*.nii.gz\n annotation masks - ./data/annotations/*.nii.gz (optional)\n crop_file - *.npz (optional, manual crops of scan of region of interest) \n\n output:\n training - ./data/name_of_data/train/trainA\n - ./data/name_of_data/train/trainB\n - ./data/name_of_data/train/annotations\n \n testing - ./data/name_of_data/train/testA\n - ./data/name_of_data/train/testB\n - ./data/name_of_data/train/annotations\n\n Then transfer trainA, trainB, testA, testB to CycleGAN/datasets/name_of_data\n\"\"\"\n\ndef get_patches(scan_path, scan_name, side='c', patch_size=256, patch_step=(128, 128)):\n scan = np.load(scan_path)['data']\n\n # crop scan using segmentation - currently not in use\n # seg = np.load(seg_path)['data']\n # cropped_scan = crop_volume(scan, seg, is_mr)\n \n # get all patches\n all_patches = get_all_patches(scan, side=side, dim=patch_size, step=patch_step)\n \n print(all_patches.shape)\n\n return all_patches\n\n\ndef prepare_data(root_path, crops, is_train = True, is_prep_npz=True, is_prep_seg=False, side='c', patch_size=256, patch_step=(128, 128)):\n\n data_type = 'train'\n if is_train is False:\n data_type = 'test'\n\n # root_path = './data/visceral_full'\n \n train_path = root_path + '/' + data_type\n dom_a_path = train_path + '/{}A'.format(data_type) # CT\n dom_b_path = train_path + '/{}B'.format(data_type) # MR\n \n train_seg_path = train_path + '/annotations'\n seg_root_path = root_path + '/annotations'\n \n nii_ext_name = '.nii.gz'\n scan_paths_train = get_image_paths_given_substr(train_path, '.nii')\n scan_names = [ p.split('/')[-1].strip('.nii.gz') for p in scan_paths_train ]\n\n os.makedirs(train_seg_path, exist_ok=True)\n os.makedirs(dom_a_path, exist_ok=True)\n os.makedirs(dom_b_path, exist_ok=True)\n \n if is_prep_npz is True:\n print(\"Converting zipped nii to npz with crops\")\n os.makedirs(dom_a_path, exist_ok=True)\n os.makedirs(dom_b_path, exist_ok=True)\n prepare_volume_as_npz(scan_paths_train, nii_ext_name, train_path, crops)\n \n if is_prep_seg is True:\n print(\"Getting all segmentations\")\n os.makedirs(train_seg_path, exist_ok=True)\n prepare_seg_as_npz(seg_root_path, scan_names, train_seg_path, crops)\n\n # only generate slices when preparing training data!\n if is_train is True:\n \n print(\"Processing npz volume files to npz image slices\")\n \n npz_file_paths = get_image_paths_given_substr(train_path, '.npz')\n\n for scan_path in npz_file_paths:\n scan_name = scan_path.replace(\".npz\", \"\").split('/')[-1]\n print(scan_name)\n seg_path = train_seg_path + '/' + scan_name + '.npz'\n is_ct = is_ct_file(scan_path)\n is_mr = not is_ct\n\n # get all patches\n all_patches = get_patches(scan_path, scan_name, side=side, patch_size=patch_size, patch_step=patch_step)\n\n for i, patch in enumerate(all_patches):\n dom_path = dom_b_path\n \n if (is_ct):\n dom_path = dom_a_path\n\n save_path = dom_path + '/' + scan_name + '_' + str(i) + '.npz'\n \n # patch = resize_img(patch, 128)\n np.savez(save_path, data=patch)\n\n\nif __name__ == '__main__':\n data_path = './data/visceral_full'\n crop_path = './visceral_crops.npz'\n \n crops = np.load(crop_path, allow_pickle=True)['data']\n\n # prepare train data here\n prepare_data(data_path, crops, is_train=True, is_prep_npz=True, is_prep_seg=False, side='c', patch_size=256, patch_step=(64, 64))\n\n # prepare test data here\n prepare_data(data_path, crops, is_train=False, is_prep_npz=False, is_prep_seg=True)\n\n","repo_name":"momenator/ct-mr-translation","sub_path":"prep_data.py","file_name":"prep_data.py","file_ext":"py","file_size_in_byte":4082,"program_lang":"python","lang":"en","doc_type":"code","stars":14,"dataset":"github-code","pt":"61"} +{"seq_id":"8096850483","text":"import unittest\nimport os\nfrom flask import Flask\nfrom dotenv import load_dotenv\nfrom src import db\nfrom src.repositories.user_repository import user_repository as ur\nfrom src.entities.user import User\nfrom src.tools.db_tools import clear_database\n\n\nclass TestUserRepository(unittest.TestCase):\n\n def setUp(self):\n load_dotenv()\n self.app = Flask(__name__)\n self.app.config[\"SECRET_KEY\"] = os.getenv(\"SECRET_KEY\")\n self.app.config[\"SQLALCHEMY_DATABASE_URI\"] = os.getenv(\"DATABASE_URL\")\n db.init_app(self.app)\n\n self.app_context = self.app.app_context()\n self.app_context.push()\n clear_database()\n self.ur = ur\n self.user1 = User(\"Tiina Testiopettaja\", \"tiina.testiope@email.com\", True)\n self.ur.register(self.user1)\n\n def tearDown(self):\n db.drop_all()\n self.app_context.pop()\n\n def test_get_user_by_email_invalid(self):\n \"\"\"\n Test that an invalid email returns False\n \"\"\"\n user = ur.get_user_by_email(\"moti@motivaatio.com\")\n self.assertFalse(user)\n\n def test_get_user_by_email(self):\n \"\"\"\n Test that a user is returned with the correct email\n \"\"\"\n user = ur.get_user_by_email(\"tiina.testiope@email.com\")\n self.assertEqual(user.name, \"Tiina Testiopettaja\")\n self.assertEqual(user.email, \"tiina.testiope@email.com\")\n self.assertTrue(user.isteacher)\n","repo_name":"piryopt/pienryhmien-optimointi","sub_path":"tests/user_repository_test.py","file_name":"user_repository_test.py","file_ext":"py","file_size_in_byte":1435,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"9799566616","text":"from django.db import migrations\n\n\ndef rename_food_names(apps, schema_editor):\n db_alias = schema_editor.connection.alias\n Food = apps.get_model(\"product\", \"Food\")\n foods = Food.objects.using(db_alias).all()\n for food in foods:\n food.name = food.name.lower()\n food.save()\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('product', '0022_add_default_scorecreator'),\n ]\n\n operations = [\n migrations.RunPython(rename_food_names),\n ]\n","repo_name":"mathijsromans/consupedia","sub_path":"product/migrations/0023_rename_food_names_lowercase.py","file_name":"0023_rename_food_names_lowercase.py","file_ext":"py","file_size_in_byte":497,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"16385396781","text":"import json\nimport re\nimport ast\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport operator\nimport collections\n\njson_contents = open('output/msd_fit_categories0.2.txt','r').read()[1:-1]\njson_contents_split = [int(a) for a in json_contents.split()]\ncluster_sizes = sorted(collections.Counter(json_contents_split).items())\n\ncluster_years = collections.defaultdict(list)\njson_contents = open('output/song_groupings0.2.txt','r').read()\nfor g in re.finditer('(\\d{1,2}): \\[.*?(\\)\\])',json_contents):\n cluster_num = g.group(1)\n for year in re.finditer(', (\\d{4})\\),',g.group(0)):\n cluster_years[cluster_num].append(int(year.group(1)))\n\nall_song_dists = {}\nall_song_nums = {}\nfor key in cluster_years.keys():\n song_dist = cluster_years[key]\n all_songs_dists_raw = sorted(collections.Counter(song_dist).items())\n total_songs_num = sum([tup[1] for tup in all_songs_dists_raw])\n all_song_nums[key] = total_songs_num\n all_song_dists[key] = [(tup[0],float(tup[1])/total_songs_num) for tup in all_songs_dists_raw]\n\nplt.switch_backend('agg')\nfor idx, key in enumerate(cluster_years):\n plt.xlim([1970,2010])\n plt.hist(cluster_years[key])\n plt.title('Group {}'.format(key))\n plt.savefig('output/songdist_group' + str(key) + '_0.2.png')\n plt.close()\n","repo_name":"matthewsilver/msd-parser","sub_path":"group_interpreter.py","file_name":"group_interpreter.py","file_ext":"py","file_size_in_byte":1283,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"35658638835","text":"\"\"\"Main experiment file.\"\"\"\n\nimport warnings; warnings.filterwarnings(\"ignore\")\n\n# reproducibility bit ----------------\nfrom random import seed; seed(42)\nfrom numpy.random import seed as np_seed; np_seed(42)\nfrom tensorflow.compat.v1 import set_random_seed; set_random_seed(42)\nimport os; os.environ['PYTHONHASHSEED'] = str(42)\n# -----------------------------------\n\nimport argparse\n\nfrom sklearn.feature_extraction.text import CountVectorizer, TfidfVectorizer\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import LinearSVC\nfrom sklearn.pipeline import Pipeline\n\nfrom evaluation import Evaluation\nfrom models import (BayesFeatures, BertFeatures, WordEmbeddings)\nfrom reader import Reader, merge_datasets\nfrom utils import debug_tests\n\n\nclass EnglishCompare(object):\n\n def __init__(self, pipeline: Pipeline, datasets: list = None,\n merge: bool = False, cross: bool = True, neural: bool = False,\n clean: bool = True, preprocess: bool = False,\n multi_read: int = 0) -> None:\n # NOTE: comment out those unavailable, or provide own list\n self.datasets = [\n ('bretschneider', 'agg_set'),\n ('kaggle', 'kag_set'),\n ('kontostathis', 'msp_set'),\n ('maral', 'ytb_set'),\n ('vanhee', 'asken_set'),\n ('xu', 'xu_set'),\n ('kaggle', 'kag_conv'),\n ('vanhee', 'asken_conv'),\n ('toxic', 'toxic_set')\n ] if not datasets else datasets\n self.data = Reader(clean=True, preprocess=False, language='en',\n multi_threading=multi_read)\n self.eval = Evaluation(pipeline,\n headers=['_'.join(x) for x in self.datasets],\n merge=merge, cross=cross, neural=neural)\n self.merge = merge\n\n def _cross_data(self) -> (list, list):\n train = [data for data in self.data.subset(self.datasets)]\n test = [_data for _data in self.data.subset(self.datasets)]\n if self.merge:\n train = train[:-1]\n train = [merge_datasets(train)]\n return train, test\n\n def run(self, nest: bool = False, store: bool = False,\n report: bool = False) -> None:\n print(f\"\\n> Merging: {self.merge}\")\n train, test = self._cross_data()\n self.eval.score(train, test=test, nest=nest, store=store,\n report=report)\n\n\nclass DutchCompare(object):\n\n def __init__(self, pipeline: Pipeline, datasets: list = None,\n merge: bool = False, cross: bool = True, neural: bool = False,\n clean: bool = True, preprocess: bool = False,\n multi_read: int = 0) -> None:\n if merge:\n raise(ValueError(\"Sorry, the NL data manually does merging.\"))\n self.datasets = [\n ('vanhee', 'asknl_set'),\n ('vanhee', 'simnl_set'),\n ('vanhee', 'donnl_set'),\n ('vanhee', 'cnvnl_set'),\n ('vanhee', 'cnsnl_set')\n ] if not datasets else datasets\n self.data = Reader(clean=True, preprocess=False, language='nl')\n headers = ['_'.join(x) for x in self.datasets]\n headers += [f'{headers[0]}+{headers[1]}']\n self.eval = Evaluation(pipeline, headers=headers, cross=cross,\n neural=neural)\n\n def _combine_sets(self) -> dict:\n sets = {'ask': '', 'sim': '', 'don': '', 'cnv': '', 'cns': ''}\n train = [data for data in self.data.subset(self.datasets)]\n\n for data in train:\n for key in sets:\n if key in data.id:\n sets[key] = data\n return sets\n\n def run(self, nest: bool = False, store: bool = False,\n report: bool = False) -> None:\n dsets = self._combine_sets()\n config = [\n (dsets['ask'], dsets['ask']),\n (dsets['sim'], dsets['sim']),\n (dsets['cnv'], dsets['cnv']),\n (dsets['cns'], dsets['cns']),\n (dsets['ask'], dsets['sim']),\n (dsets['sim'], dsets['ask']),\n (dsets['ask'], dsets['don']),\n (dsets['sim'], dsets['don']),\n (dsets['ask'], dsets['cnv']),\n (dsets['ask'], dsets['cns']),\n (dsets['sim'], dsets['cnv']),\n (dsets['sim'], dsets['cns']),\n (dsets['cnv'], dsets['ask']),\n (dsets['cnv'], dsets['sim']),\n (dsets['cnv'], dsets['don']),\n (dsets['cnv'], dsets['cns']),\n (dsets['cns'], dsets['cnv']),\n (dsets['cns'], dsets['ask']),\n (dsets['cns'], dsets['sim']),\n (dsets['cns'], dsets['don']),\n (merge_datasets([dsets['ask'], dsets['sim']]), dsets['ask']),\n (merge_datasets([dsets['ask'], dsets['sim']]), dsets['sim']),\n (merge_datasets([dsets['ask'], dsets['sim']]), dsets['don']),\n (merge_datasets([dsets['ask'], dsets['sim']]), dsets['cnv']),\n (merge_datasets([dsets['ask'], dsets['sim']]), dsets['cns']),\n (merge_datasets([dsets['ask'], dsets['sim'], dsets['cnv'],\n dsets['cns']]), dsets['ask']),\n (merge_datasets([dsets['ask'], dsets['sim'], dsets['cnv'],\n dsets['cns']]), dsets['sim']),\n (merge_datasets([dsets['ask'], dsets['sim'], dsets['cnv'],\n dsets['cns']]), dsets['don']),\n (merge_datasets([dsets['ask'], dsets['sim'], dsets['cnv'],\n dsets['cns']]), dsets['cnv']),\n (merge_datasets([dsets['ask'], dsets['sim'], dsets['cnv'],\n dsets['cns']]), dsets['cns'])\n ]\n for train, test in config:\n self.eval.score([train], test=[test], nest=nest, df=False)\n\n\ndef select_model(key: str) -> Pipeline:\n \"\"\"Select the model to use based on argparse input.\"\"\"\n\n # NOTE: all these if statements don't look particularly nice, but we also\n # don't want to load a bunch of models we're not gonna use now, do we?\n\n # Simple Default Test\n if key == 'debug':\n return {\n ('vect', TfidfVectorizer(ngram_range=(1, 2), min_df=3,\n max_df=0.9, use_idf=1, smooth_idf=1,\n sublinear_tf=1)): {},\n ('nbf', BayesFeatures()): {},\n ('lr', LogisticRegression(dual=True, random_state=42,\n class_weight=\"balanced\")): {}\n }\n\n # FINAL BINARY SVM BASELINE\n elif key == 'baseline':\n return {\n ('vect', CountVectorizer(binary=True)): {\n 'vect__ngram_range': [(1, 1), (1, 2), (1, 3)]\n },\n ('svc', LinearSVC(random_state=42)): {\n 'svc__C': [1e-3, 1e-2, 1e-1, 1e-0, 1e1, 1e2, 1e3],\n 'svc__loss': ['hinge', 'squared_hinge'],\n 'svc__class_weight': [None, \"balanced\"]\n }\n }\n\n elif key == 'debug-baseline':\n return {\n ('vect', CountVectorizer(binary=True)): {},\n ('svc', LinearSVC(random_state=42)): {}\n }\n\n # NB-SVM Model\n elif key == 'nbsvm':\n return {\n ('vect', TfidfVectorizer(ngram_range=(1, 2), min_df=3, max_df=0.9,\n use_idf=1, smooth_idf=1,\n sublinear_tf=1)): {},\n ('nbf', BayesFeatures()): {},\n ('lr', LogisticRegression(dual=True, solver='liblinear',\n random_state=42,\n class_weight=\"balanced\")): {\n 'lr__C': [1, 2, 3, 4, 5, 10, 25, 50, 100, 200, 500]\n }\n }\n\n elif key == 'debug-nbsvm':\n return {\n ('vect', TfidfVectorizer()): {},\n ('nbf', BayesFeatures()): {},\n ('lr', LogisticRegression()): {}\n }\n\n # LR over Embeddings\n elif key == 'w2v':\n return {\n # NOTE: cow.nl.kv for Dutch\n ('vct', WordEmbeddings(pre_trained=\"w2v.kv\")): {},\n ('lr', LogisticRegression(class_weight=\"balanced\",\n solver='liblinear', random_state=42)): {\n 'lr__C': [1, 2, 3, 4, 5, 10, 25, 50, 100, 200, 500],\n }\n }\n\n elif key == 'debug-w2v':\n return {\n # NOTE: cow.nl.kv for Dutch\n ('vct', WordEmbeddings(pre_trained=\"w2v.kv\")): {},\n ('lr', LogisticRegression()): {}\n }\n\n # LR over DistilBert\n elif key == 'bert':\n return {\n ('vect', BertFeatures()): {},\n ('lr', LogisticRegression(class_weight=\"balanced\",\n solver='liblinear', random_state=42)): {\n 'lr__C': [1, 2, 3, 4, 5, 10, 25, 50, 100, 200, 500],\n }\n }\n\n elif key == 'debug-bert':\n return {\n ('vect', BertFeatures()): {},\n ('lr', LogisticRegression()): {}\n }\n\n # Reproduction B-LSTM\n elif key == 'blstm':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='blstm', inp_dim=128, num_classes=2, learn_rate=0.01,\n batch_size=128, epochs=10, embed_size=50)): {}\n }\n\n elif key == 'debug-blstm':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='blstm', inp_dim=128, num_classes=2, learn_rate=1,\n batch_size=32, epochs=1, embed_size=50)): {}\n }\n\n # Reproduction CNN\n elif key == 'cnn':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='cnn', inp_dim=128, num_classes=2, learn_rate=0.01,\n batch_size=128, epochs=10, embed_size=50)): {}\n }\n\n elif key == 'debug-cnn':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='cnn', inp_dim=128, num_classes=2, learn_rate=1,\n batch_size=32, epochs=1, embed_size=50)): {}\n }\n\n # Own NN Grid\n elif key == 'nn':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='clstm', inp_dim=128, num_classes=2, learn_rate=0.01,\n batch_size=128, epochs=10, embed_size=50, character_level=True,\n early_stop=3)\n ): {\n # NOTE: roughly optimal params: 128 batch / 100, 50 embeddings\n 'neur__batch_size': [32, 64, 128, 256],\n 'neur__embed_size': [50, 100, 200, 300],\n 'neur__learn_rate': [0.1, 0.01, 0.05, 0.001, 0.005]\n }\n }\n\n elif key == 'debug-nn':\n from neural import ReproductionNeuralNetwork\n return {\n ('neur', ReproductionNeuralNetwork(\n m_type='clstm', inp_dim=128, num_classes=2, learn_rate=1,\n batch_size=32, epochs=1, embed_size=50)): {}\n }\n\n else:\n raise(KeyError(f\"Sorry, `{key}` is not a valid --model name.\"))\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(\n description='Cyberbullying detection replication environment.',\n formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n\n parser.add_argument('model',\n help=\"\"\"debug | baseline | nbsvm | w2v | bert |\n blstm | cnn | nn -- Debug will run all possible\n configurations!\"\"\", type=str)\n parser.add_argument('--language', default='en', type=str, help=\"\"\"Run on\n English (en) or Dutch (nl) data.\"\"\")\n parser.add_argument('--preprocessing', default='clean', type=str,\n help=\"none | clean | preprocess\")\n parser.add_argument('--merge', default=False, type=bool, help=\"\"\"Merge all\n training sets (D_All in paper).\"\"\")\n parser.add_argument('--nest', default=False, type=bool, help=\"\"\"Report\n nested cross-validation scores (only relevant when\n using GridSearch).\"\"\")\n parser.add_argument('--single_domain', default=False, type=bool,\n help=\"Don't run eval cross-domain.\")\n parser.add_argument('--multi_read', default=0, help=\"\"\"Number of cores the\n _reader_ should use for multi-threading.\"\"\")\n parser.add_argument('--store', default=False, type=bool, help=\"\"\"Save the\n best model in a pickle file under /results.\"\"\")\n parser.add_argument('--report', default=False, type=bool, help=\"\"\"Report\n the most important features for SVM/LR models.\"\"\")\n args = parser.parse_args()\n\n Experiment = EnglishCompare if args.language == 'en' else DutchCompare\n if args.model == 'debug':\n debug_tests(args, EnglishCompare, select_model)\n else:\n Experiment(pipeline=select_model(args.model), merge=args.merge,\n datasets=None, cross=args.single_domain,\n neural=args.model in ['blstm', 'cnn', 'nn'],\n clean='clean' in args.preprocessing,\n preprocess='preprocess' in args.preprocessing,\n multi_read=args.multi_read).run(nest=args.nest,\n store=args.store,\n report=args.report)\n","repo_name":"cmry/amica","sub_path":"experiments.py","file_name":"experiments.py","file_ext":"py","file_size_in_byte":13653,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"61"} +{"seq_id":"41192601053","text":"import FWCore.ParameterSet.Config as cms\n\nhgcalMultiClusters = cms.EDProducer('HGCalMultiClusterProducer',\n HGCLayerClusters = cms.InputTag('hgcalLayerClusters'),\n verbosity = cms.untracked.uint32(3),\n doSharing = cms.bool(False),\n HGCEEInput = cms.InputTag('HGCalRecHit', 'HGCEERecHits'),\n HGCFHInput = cms.InputTag('HGCalRecHit', 'HGCHEFRecHits'),\n multiclusterRadii = cms.vdouble(\n 2,\n 5,\n 5\n ),\n HGCBHInput = cms.InputTag('HGCalRecHit', 'HGCHEBRecHits'),\n HGCLayerClustersSharing = cms.InputTag('hgcalLayerClusters', 'sharing'),\n minClusters = cms.uint32(3),\n mightGet = cms.optional.untracked.vstring\n)\n","repo_name":"cms-sw/cmssw-cfipython","sub_path":"RecoLocalCalo/HGCalRecProducers/hgcalMultiClusters_cfi.py","file_name":"hgcalMultiClusters_cfi.py","file_ext":"py","file_size_in_byte":628,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"3586780282","text":"from .historyitem_subcomponent import historyitem_module as HistoryItem\nfrom ....common_components.datetime_datatypes import datetime_module as DateTime\nfrom ....common_components.datetime_datatypes import eras_module as EraFunctions\nfrom .graphing_subcomponent import graphing_module as Graphing\n\n\nclass DefineHistory:\n\n\tdef __init__(self):\n\n\t\t# An array of historic monitor history\n\t\tself.monitorhistory = []\n\n\t\t# Defines the granularity of display of monitor data\n\t\tself.erasize = 4 # Ten minute intervals\n\t\tself.longerasize = 5 # Hour intervals\n\n\t\t# Graphing module\n\t\tself.graphs = Graphing.creategraphing(self.erasize, self.longerasize)\n\n# =========================================================================================\n\n\tdef addhistoryentry(self, colourcounts, networkstatus, uploadedtotal, temperature):\n\n\t\tcurrentdatetime = DateTime.getnow()\n\t\tnewhistoryitem = HistoryItem.createhistoryitem(currentdatetime, colourcounts, networkstatus, uploadedtotal,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttemperature)\n\t\tself.monitorhistory.append(newhistoryitem)\n\t\tself.clearuphistory(currentdatetime)\n\t\t#print(\"NEW MONITOR HISTORY ITEM: \", newhistoryitem.getsavedata())\n\t\treturn newhistoryitem.getsavedata()\n\n\n\n\tdef restorehistory(self, saveddatalist):\n\n\t\tfor dataitem in saveddatalist:\n\t\t\tself.monitorhistory.append(HistoryItem.createfromfile(dataitem))\n\n\n\n\tdef gethistorygraphics(self, historyperiod):\n\n\t\tif historyperiod == \"Latest\":\n\t\t\toutput = self.graphs.drawgraphs(False, self.monitorhistory)\n\t\telif historyperiod == \"Recent\":\n\t\t\toutput = self.graphs.drawgraphs(True, self.getlonghistory())\n\t\telse:\n\t\t\toutput = {}\n\n\t\treturn output\n\n\n\n\tdef clearuphistory(self, currentdatetime):\n\n\t\tif currentdatetime.gettimevalue() < 600:\n\t\t\tprint(\"Before clean up: \", len(self.monitorhistory))\n\t\t\tthreshold = DateTime.createfromobject(currentdatetime)\n\t\t\tthreshold.adjustdays(-11)\n\t\t\tnewhistorylist = []\n\t\t\tfor historyitem in self.monitorhistory:\n\t\t\t\tif DateTime.isfirstlaterthansecond(historyitem.getdatetime(), threshold) == True:\n\t\t\t\t\tnewhistorylist.append(historyitem)\n\n\t\t\tself.monitorhistory = newhistorylist.copy()\n\t\t\tprint(\"After clean up: \", len(self.monitorhistory))\n\n\n\n\n\tdef getlonghistory(self):\n\n\t\toutcome = []\n\t\tcurrentlonghistoryitem = HistoryItem.createblank(DateTime.createfromiso(\"20100101000000\"))\n\t\tfor historyitem in self.monitorhistory:\n\t\t\tnewhour = historyitem.getdatetime()\n\t\t\tif EraFunctions.compareeras(newhour, currentlonghistoryitem.getdatetime(), 5) == True:\n\t\t\t\tcurrentlonghistoryitem.cumulate(historyitem)\n\t\t\telse:\n\t\t\t\toutcome.append(currentlonghistoryitem)\n\t\t\t\tcurrentlonghistoryitem = HistoryItem.createblank(EraFunctions.geteraasobject(newhour, 5))\n\t\t\t\tcurrentlonghistoryitem.cumulate(historyitem)\n\t\tif EraFunctions.compareeras(currentlonghistoryitem.getdatetime(), DateTime.getnow(), 5) == False:\n\t\t\toutcome.append(currentlonghistoryitem)\n\t\treturn outcome\n\n\n\n","repo_name":"johnpcole/Download-Manager","sub_path":"codebase/manager_component/monitoring_subcomponent/history_subcomponent/history_class.py","file_name":"history_class.py","file_ext":"py","file_size_in_byte":2886,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"9161716571","text":"#! /usr/bin/env python\n\"\"\"\nSetup for Robbie\n\"\"\"\nimport os\nfrom setuptools import setup\nfrom subprocess import check_output\n\n\n# The following two functions were taken from the repo: https://github.com/pyfidelity/setuptools-git-version/blob/master/setuptools_git_version.py\n\ndef format_version(version, fmt=\"{tag}_{gitsha}\"):\n parts = version.split(\"-\")\n if len(parts) < 4:\n return parts[0]\n assert len(parts) in (3, 4)\n dirty = len(parts) == 4\n tag, count, sha = parts[:3]\n if count == \"0\" and not dirty:\n return tag\n return fmt.format(tag=tag, commitcount=count, gitsha=sha.lstrip(\"g\"))\n\n\ndef get_git_version():\n git_version = check_output(\"git describe --tags --long --dirty --always\".split()).decode('utf-8').strip()\n return format_version(version=git_version)\n\n\nrobbie_version = get_git_version()\n#m ake a temporary version file to be installed then delete it\nwith open(\"robbie_version.sh\", \"a\") as the_file:\n the_file.write(f\"#!/bin/bash -l\\necho {robbie_version}\")\n\nsetup(\n name=\"Robbie\",\n version=robbie_version,\n description=\"A batch processing work-flow for the detection of radio transients and variables\",\n url=\"https://github.com/PaulHancock/Robbie\",\n python_requires=\">=3.6\",\n packages=['robbie'],\n scripts=[\n \"robbie_version.sh\",\n # python\n \"scripts/auto_corr.py\",\n \"scripts/calc_var.py\",\n \"scripts/collect_transients.py\",\n \"scripts/filter_transients.py\",\n \"scripts/get_epoch.py\",\n \"scripts/get_lc_from_vot.py\",\n \"scripts/join_catalogues.py\",\n \"scripts/plot_variables.py\",\n \"scripts/make_weights.py\",\n \"scripts/reprojection.py\",\n \"scripts/convol_common_resolution.py\",\n # nextflow\n \"robbie.nf\",\n \"nextflow.config\",\n ],\n)\nos.remove(\"robbie_version.sh\")\n","repo_name":"PaulHancock/Robbie","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1845,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"8627165092","text":"upp, low, dig, pct = 0, 0, 0, 0\n\npswd = input('암호 입력: ')\nif pswd.isalnum() == False : pct = 1 # 특수문자 있으면 pct = 1\nfor k in pswd:\n if k.isupper() : upp = 1 # 대문자 있으면 upp = 1\n elif k.islower() : low = 1 # 소문자 있으면 low = 1\n elif k.isdigit() : dig = 1 # 숫자 있으면 dig = 1\n\nif low + upp + dig + pct >= 3 :\n print('사용 가능')\nelse : print('!!불가능한 암호!!')","repo_name":"hellen1221/2023pythonclass","sub_path":"class6/passw.py","file_name":"passw.py","file_ext":"py","file_size_in_byte":459,"program_lang":"python","lang":"ko","doc_type":"code","stars":2,"dataset":"github-code","pt":"61"} +{"seq_id":"19053414160","text":"#coding:utf-8\n\"\"\"\nWrite a Python program to find the roots of a quadratic function. Go to the editor\n\nExpected Output :\n\nQuadratic function : (a * x^2) + b*x + c \na: 25 \nb: 64 \nc: 36 \nThere are 2 roots: -0.834579 and -1.725421\n\"\"\"\nfrom math import sqrt\ndef roots_quadratic():\n a = float(input(\"a: \"))\n b = float(input(\"b: \"))\n c = float(input(\"c: \"))\n\n discriminant=b**2 - 4*a*c\n if discriminant > 0:\n num_roots = 2\n x1 = (((-b) + sqrt(discriminant))/(2*a)) \n x2 = (((-b) - sqrt(discriminant))/(2*a))\n print(\"There are 2 roots: %f and %f\" % (x1, x2))\n elif discriminant == 0:\n num_roots = 1\n x = (-b) / 2*a\n print(\"There is one root: \", x)\n else:\n num_roots = 0\n print(\"No roots, discriminant < 0.\")\n exit()\n\nprint(roots_quadratic())","repo_name":"DonaFidele/PythonExercices","sub_path":"math/exo_30.py","file_name":"exo_30.py","file_ext":"py","file_size_in_byte":1057,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"8059203934","text":"from ...models import Move\nfrom ...models import Game\nfrom rest_framework import serializers\nfrom django.http import Http404\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom rest_framework.views import APIView\nfrom ...utils import get_neighbors\n\n\ndef tile_coordinates(value):\n valid_values = range(0, 9)\n if value not in valid_values:\n raise serializers.ValidationError('Not a valid value.')\n\n\ndef sudoku_value(value):\n valid_values = [None]\n valid_values.extend(range(1, 10))\n if value not in valid_values:\n raise serializers.ValidationError('Not a valid value.')\n\n\nclass MoveSerializer(serializers.ModelSerializer):\n x = serializers.IntegerField(validators=[tile_coordinates])\n y = serializers.IntegerField(validators=[tile_coordinates])\n value = serializers.IntegerField(validators=[sudoku_value], allow_null=True)\n\n class Meta:\n model = Move\n fields = ('id', 'game', 'previous_move', 'x', 'y', 'value')\n read_only_fields = ('id', 'previous_move',)\n\n\nclass MoveAPIListViewV1(APIView):\n def get(self, request, format=None):\n game_id = request.GET.get('game', None)\n try:\n game_id = int(game_id)\n except (TypeError, ValueError):\n return Response({'detail': 'The query parameter game must be a positive integer.'}, status.HTTP_400_BAD_REQUEST)\n\n try:\n game = Game.objects.get(id=game_id)\n except Game.DoesNotExist:\n raise Http404\n\n moves = Move.objects.filter(game=game)\n serializer = MoveSerializer(moves, many=True)\n return Response(serializer.data)\n\n def post(self, request, format=None):\n serializer = MoveSerializer(data=request.data)\n if serializer.is_valid():\n game = serializer.validated_data.get('game')\n\n # Ensure the tile can be edited\n if not game.is_tile_editable(serializer.validated_data.get('x'), serializer.validated_data.get('y')):\n return Response({'detail': 'The given coordinate is not editable.'}, status=status.HTTP_400_BAD_REQUEST)\n\n # Ensure the tile doesn't conflict with neighbors\n rendered_game = game.render_game()\n if serializer.validated_data['value'] in get_neighbors(serializer.validated_data.get('y'), serializer.validated_data.get('x'), rendered_game):\n return Response({'detail': 'Bad move! Conflict with our neighbors.'}, status=status.HTTP_400_BAD_REQUEST)\n\n # Determine the last move dynamically\n previous_move = Move.objects.filter(game=request.data.get('game')).order_by('-id').first()\n serializer.save(previous_move=previous_move)\n\n # Update the game state\n game.user_input[serializer.validated_data['y']][serializer.validated_data['x']] = serializer.validated_data['value']\n game.save()\n\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n","repo_name":"ryanisnan/sudoku_be","sub_path":"src/sudoku/sudoku/api/v1/move.py","file_name":"move.py","file_ext":"py","file_size_in_byte":3067,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"23426202901","text":"\r\nout_file = open('b_out.txt', 'w')\r\n\r\ndef out(case_idx, num):\r\n\tprint(\"Case #%d: %.7f\" % (case_idx + 1, num), file=out_file)\r\n\r\ndef get_time(target, rate):\r\n\treturn target / rate\r\n\r\nwith open('B-large.in') as f:\r\n\tnum_cases = int(f.readline())\r\n\tfor i in range(num_cases):\r\n\t\t[cost, increase, target] = [float(x) for x in f.readline().split(\" \")]\r\n\r\n\t\tcur_rate = 2\r\n\t\tbest_time = get_time(target, cur_rate)\r\n\t\ttime_to_add = 0.0\r\n\t\twhile True:\r\n\t\t\ttime_to_add += get_time(cost, cur_rate)\r\n\t\t\tcur_rate += increase\r\n\t\t\talt_time = get_time(target, cur_rate) + time_to_add\r\n\r\n\t\t\tif alt_time < best_time:\r\n\t\t\t\tbest_time = alt_time\r\n\t\t\telse:\r\n\t\t\t\tbreak\r\n\r\n\t\talt_time = get_time(target, 2 + increase) + cost\r\n\r\n\t\tout(i, best_time)\r\n\r\n\t\t\r\n\r\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_136/2878.py","file_name":"2878.py","file_ext":"py","file_size_in_byte":733,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"37802762647","text":"from tweepy.streaming import StreamListener\nfrom tweepy import OAuthHandler\nfrom tweepy import Stream\nimport json\n\n# access tokens for twitter\naccess_token = \"981309028490981376-YoTeZT7xidSEg1euxFbIzYshY16sVpI\"\naccess_token_secret = \"BK9P4xUY3zPgVZw6wpQxJkmjsmH4NsM8wJ2rEVOkH5Mv3\"\nconsumer_key = \"fg4pBbU44JUMRnDRbjtOmKCRR\"\nconsumer_secret = \"iN3k4V7MuA7M12rwZOvmyAfJudd4qm4mnt68kSTJuOSkfP0K7X\"\n\n\ntweets_file = None\ntoxic_key_words_file = \"../../../../resources/toxic-keywords.txt\"\nnon_toxic_key_words_file = \"../../../../resources/non-toxic-keywords.txt\"\n\n\"\"\"\n Returns toxic keywords\n\"\"\"\n\n\ndef keywords(key_words_file):\n fp = open(key_words_file, \"r\")\n words = [w.strip() for w in fp.readlines() if w.strip() is not None and len(w.strip()) > 0]\n return words\n\n\n\"\"\"\nThis is a basic listener that just prints received tweets to stdout.\n\"\"\"\n\n\nclass StdOutListener(StreamListener):\n def on_status(self, status):\n with open(tweets_file,'a',encoding=\"utf-8\") as tf:\n tf.write('\\n')\n if hasattr(status, 'retweeted_status'):\n try:\n tweet = status.retweeted_status.extended_tweet['full_text']\n tf.write(tweet.strip().replace('\\n', \" \").replace('\\r', ''))\n except:\n tweet = status.retweeted_status.text\n tf.write(tweet.strip().replace('\\n', \" \").replace('\\r', ''))\n else:\n try:\n tweet = status.extended_tweet[\"full_text\"]\n tf.write(tweet.strip().replace('\\n', \" \").replace('\\r', ''))\n except:\n tweet = status.text\n tf.write(tweet.strip().replace('\\n', \" \").replace('\\r', ''))\n return True\n\n def on_error(self, status):\n print(status)\n\n\nif __name__ == '__main__':\n # This handles Twitter authentication and the connection to Twitter Streaming API\n listener = StdOutListener()\n auth = OAuthHandler(consumer_key, consumer_secret)\n auth.set_access_token(access_token, access_token_secret)\n stream = Stream(auth, listener)\n\n # This line filter Twitter Streams to capture data\n # change file name (keywords and output file) to track the respective\n tweets_file = '../../../../resources/toxic-tweets.txt'\n stream.filter(languages=[\"en\"], track=keywords(toxic_key_words_file))\n\n","repo_name":"KavyaJampani/toxic_tweet","sub_path":"src/edu/utdallas/dc/tweetCollect.py","file_name":"tweetCollect.py","file_ext":"py","file_size_in_byte":2369,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"39501535181","text":"class Solution:\n \"\"\"\n @param matrix, a list of lists of integers\n @param target, an integer\n @return a boolean, indicate whether matrix contains target\n \"\"\"\n def searchMatrix(self, matrix, target):\n if not matrix or not matrix[0]:\n return False\n n, m = len(matrix), len(matrix[0])\n left, right = 0, n - 1\n while left + 1 < right:\n mid = left + (right - left) / 2\n if matrix[mid][0] > target:\n right = mid - 1\n else:\n left = mid\n row = right\n if matrix[right][0] > target:\n row = left\n\n left, right = 0, m - 1\n while left + 1 < right:\n mid = left + (right - left) / 2\n if matrix[row][mid] > target:\n right = mid - 1\n else:\n left = mid\n if matrix[row][right] == target or matrix[row][left] == target:\n return True\n return False\n\n","repo_name":"jwyx3/practices","sub_path":"python/search-a-2d-matrix.py","file_name":"search-a-2d-matrix.py","file_ext":"py","file_size_in_byte":974,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"23552033031","text":"# Leon Xueliang Liu 2017\r\n\r\nwith open('B-large.in', 'r') as f:\r\n\tcontent = f.readlines()\r\n\r\nT = int(content[0]) # # of cases\r\ndata = [line.split()[0] for line in content[1:]]\r\n\r\nresult = [] # list of results\r\n\r\nfor n in range(T):\r\n\tans = []\r\n\tN = str(data[n])\r\n\tfor char in N:\r\n\t\tans.append(char)\r\n\tL = len(ans)\r\n\r\n\t# forward pass, if next # is smaller, decrement current, set all later to 9\r\n\tfor i in range(L-1):\r\n\t\tif int(ans[i]) > int(ans[i+1]):\r\n\t\t\tans[i] = str(int(ans[i])-1)\r\n\t\t\tfor j in range(i+1, L):\r\n\t\t\t\tans[j] = '9'\r\n\t\t\tbreak\r\n\r\n\t# reverse pass, if next # is bigger, decrement next, set current to 9\r\n\tfor i in range(L-1):\r\n\t\tm = int(ans[L-1-i])\r\n\t\tn = int(ans[L-2-i])\r\n\t\tif m < n:\r\n\t\t\tans[L-1-i] = '9'\r\n\t\t\tans[L-2-i] = str(n-1)\r\n\r\n\tans = int(\"\".join(ans))\r\n\tresult.append(ans)\t\t\r\n\r\n#write to output\r\nwith open('B-large.txt','w+') as f:\r\n\tfor count, ans in enumerate(result):\r\n\t\tf.write(\"Case #{}: {}\\n\".format(count+1, ans))\r\n\r\n","repo_name":"dr-dos-ok/Code_Jam_Webscraper","sub_path":"solutions_python/Problem_200/1243.py","file_name":"1243.py","file_ext":"py","file_size_in_byte":941,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"34211402822","text":"import uproot\nimport awkward as ak\nimport matplotlib.pyplot as plt\nimport numpy as np\n\n\n# read the ttree\ntree = uproot.open(\"myNanoProdMc2016_NANO_1.root\")[\"Events\"]\n# read the pf candidates information into akward arrays\npfcands = tree.arrays(tree.keys('PF_*') + ['nPF'], entry_start=0, entry_stop=1000)\n\n# plot the number of PF candidates per event\nplt.figure()\nplt.hist( ak.to_numpy(pfcands['nPF']), bins=50, range=(0,3000), histtype='step')\nplt.savefig(\"nPF.png\")\n\n# get the unique pdgIds\npdgIds = ak.to_numpy(ak.flatten(pfcands['PF_pdgId']))\nprint(\"unique pdgIds: \", np.unique(pdgIds))\nprint(\"unique abs(pdgIds): \", np.unique(abs(pdgIds)))\n\n# calculate the average of abs(PF_dxy) for charged pf candidates \ndxyavg_PV = ak.to_numpy(ak.mean(abs(pfcands['PF_dxy']),weight=((pfcands['PF_charge']!=0) * (pfcands['PF_puppiWeight']==1)), axis=1), allow_missing=False)\ndxyavg_PU = ak.to_numpy(ak.mean(abs(pfcands['PF_dxy']),weight=((pfcands['PF_charge']!=0) * (pfcands['PF_puppiWeight']==0)), axis=1), allow_missing=False)\n\nplt.figure()\nplt.hist(dxyavg_PU, bins=50, range=(0,1.0), histtype='step', label='Charged PU')\nplt.hist(dxyavg_PV, bins=50, range=(0,1.0), histtype='step', label='Charged PV')\nplt.legend(loc=\"upper right\")\nplt.xlabel('Average |PF_dxy| for charged particles [cm]')\nplt.savefig('PF_absdxy_charged.png')\n","repo_name":"yongbinfeng/PFCands","sub_path":"demo.py","file_name":"demo.py","file_ext":"py","file_size_in_byte":1321,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"73849201473","text":"from libqtile import bar\nfrom libqtile.config import Screen\nfrom qtile_extras import widget\nfrom modules.widgets import *\nfrom utils.settings import colors, two_monitors, wallpaper_main, wallpaper_sec\n\ncolor_alert = '#ee9900'\ncolor_frame = '#808080'\n\nwidget_defaults = dict(\n font=\"FiraCode Nerd Font\",\n fontsize=14,\n padding=2,\n)\n\nextension_defaults = widget_defaults.copy()\n\ndef create_bar(extra_bar = False):\n \"\"\"Create top bar, defined as function to allow duplication in other monitors\"\"\"\n return bar.Bar(\n [\n gen_separator(25,50),\n w_bar_icon,\n w_window_name,\n gen_spacer(),\n *gen_groupbox(),\n gen_spacer(),\n *((w_systray,) if not extra_bar else ()),\n gen_separator(15,50),\n vol_icon, w_vol,\n gen_separator(5,50),\n network_icon, w_network,\n gen_separator(11,50),\n clock_icon, w_clock,\n gen_separator(25,50),\n ],\n 36,\n margin=[0, 0, 4, 0],\n background=\"#000000\", opacity=0.8,\n )\n\nmain_screen_bar = create_bar()\nif two_monitors:\n secondary_screen_bar = create_bar(True)\n\nscreens = [\n Screen(\n wallpaper=wallpaper_main,\n wallpaper_mode=\"fill\",\n top=main_screen_bar,\n bottom=bar.Gap(4),\n left=bar.Gap(4),\n right=bar.Gap(4),\n ),\n]\n\nif two_monitors:\n screens.append(\n Screen(\n wallpaper=wallpaper_sec,\n wallpaper_mode=\"fill\",\n top=secondary_screen_bar,\n bottom=bar.Gap(4),\n left=bar.Gap(4),\n right=bar.Gap(4),\n ),\n )\n","repo_name":"fantasy0x1/dotfiles","sub_path":".config/qtile/modules/screens.py","file_name":"screens.py","file_ext":"py","file_size_in_byte":1661,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"61"} +{"seq_id":"12044794007","text":"# Parser class that will consume a list of token to build the grammar\nfrom lox.token import LoxToken\nfrom lox.tokentype import TokensDic as Tk\nfrom lox.constants import LoxConstant\nfrom lox.stmt import *\nfrom lox.expr import *\nfrom lox.visitor import Visitor\nfrom lox.error import ParserError, LoxError\nfrom lox.functiontypes import FunctionType\n\nfrom typing import List\n\n\nclass Parser:\n #\n # The parser is initialized with the list of tokens to parse\n def __init__(self, tokens: LoxToken):\n self.tokens = tokens\n self.current = 0\n\n #\n # get the current token without moving forward\n def peek(self) -> LoxToken:\n return self.tokens[self.current]\n\n #\n # Get previous token\n def previous(self) -> LoxToken:\n if self.current > 0:\n return self.tokens[self.current-1]\n else:\n return None\n\n #\n # Get next token\n def next(self) -> LoxToken:\n if self.is_at_end():\n raise ParserError(\n self.peek(), \"reached end of file, cannot get next token.\")\n return self.tokens[self.current+1]\n\n #\n # Check we did not reach the end of the file\n def is_at_end(self) -> bool:\n return self.peek().type == Tk.EOF\n\n #\n # Check the token type whithout moving forward\n def check(self, ttype: str) -> bool:\n if self.is_at_end():\n return False\n return self.peek().type == ttype\n\n #\n # Check the next token type whithout moving forward\n def checknext(self, ttype: str) -> bool:\n if self.is_at_end():\n return False\n return self.next().type == ttype\n\n #\n # To check if the current token type is any of the input list.\n # If so consume the token (advance) that can be retrieved with previous()\n def match(self, *tokentypes: List[str]) -> bool:\n for t in tokentypes:\n if self.check(t):\n self.advance()\n return True\n return False\n\n #\n # Consume and return token\n def advance(self) -> LoxToken:\n if not self.is_at_end():\n self.current += 1\n return self.previous()\n\n #\n # Check if the next token is of the specified type and return it, otherwise return an error\n # moves forward if so.\n def consume(self, ttype: str, message: str) -> LoxToken:\n # print(\"--> in parserconsume: Token:\",\n # self.tokens[self.current].type, \" to compare with: \", ttype)\n if self.check(ttype):\n return self.advance()\n raise ParserError(self.peek(), message)\n\n #\n # -------------------------------------------------\n # Statement functions\n # -------------------------------------------------\n #\n def declaration(self) -> Stmt:\n try:\n # variable declaration statement\n if self.match(Tk.VAR):\n return self.vardeclaration()\n # function declaration statement\n elif self.check(Tk.FUN): # and self.checknext(Tk.IDENTIFIER):\n return self.fundeclaration()\n # class declaration statement\n elif self.match(Tk.CLASS):\n return self.classdeclaration()\n # generic statement\n else:\n return self.statement()\n\n except ParserError as err:\n print(\"error in parsing at token \" +\n str(self.previous()) + \" \" + err.message)\n return None\n # self.synchronize()\n\n def vardeclaration(self) -> Stmt:\n # Let's get the identifier\n tk_varname = self.consume(\n Tk.IDENTIFIER, \"Expect a variable identifier\")\n # Retrieve the initializing expression\n initializer = None\n if self.match(Tk.EQUAL):\n initializer = self.expression()\n # Make sure the statement has a termination statement\n self.consume(Tk.SEMICOLON, \"Expect ';' after variable declaration.\")\n return Var(tk_varname, initializer)\n\n def fundeclaration(self) -> Stmt:\n self.consume(Tk.FUN, \"Expect 'fun' for function statement.\")\n # a function statement should not be anonymous and is not a lambda\n return Function(self.functionbody(FunctionType.FUNCTION))\n\n def classdeclaration(self) -> Stmt:\n superclass = None\n name = self.consume(\n Tk.IDENTIFIER, \"Expect a class name after 'class'.\")\n if self.match(Tk.LESS):\n self.consume(\n Tk.IDENTIFIER, \"Expect a super class name after '<'.\")\n superclass = Variable(self.previous())\n self.consume(Tk.LEFT_BRACE, \"Expect a '{' name after class name.\")\n methods = []\n while not self.check(Tk.RIGHT_BRACE) and not self.is_at_end():\n methods.append(self.functionbody(FunctionType.METHOD))\n self.consume(Tk.RIGHT_BRACE, \"Expect '}' after class body.\")\n return Class(name, superclass, methods)\n\n def statement(self) -> Stmt:\n if self.match(Tk.PRINT):\n return self.printstatement()\n if self.match(Tk.LEFT_BRACE):\n return Block(self.blockstatement())\n if self.match(Tk.IF):\n return self.ifstatement()\n if self.match(Tk.WHILE):\n return self.whilestatement()\n if self.match(Tk.FOR):\n return self.forstatement()\n if self.match(Tk.BREAK):\n return self.breakstatement()\n if self.match(Tk.RETURN):\n return self.returnstatement()\n return self.expressionstatement()\n\n def ifstatement(self) -> Stmt:\n self.consume(Tk.LEFT_PAREN,\n \"expect a ( at the start of the if condition.\")\n condition = self.expression()\n self.consume(Tk.RIGHT_PAREN,\n \"expect a ) at the end of the if condition.\")\n thenbranch = self.statement()\n elsebranch = None\n if self.match(Tk.ELSE):\n elsebranch = self.statement\n return If(condition, thenbranch, elsebranch)\n\n def whilestatement(self) -> Stmt:\n self.consume(Tk.LEFT_PAREN,\n \"expect a ( at the start of the 'while' condition.\")\n condition = self.expression()\n self.consume(Tk.RIGHT_PAREN,\n \"expect a ) at the end of the 'while' condition.\")\n body = self.statement()\n return While(condition, body)\n\n def forstatement(self) -> Stmt:\n self.consume(Tk.LEFT_PAREN, \"expect a ( after a 'for'.\")\n initializer = None\n if self.match(Tk.VAR):\n initializer = self.vardeclaration()\n elif not self.match(Tk.SEMICOLON):\n initializer = self.expressionstatement()\n condition = None\n if not self.check(Tk.SEMICOLON):\n condition = self.expression()\n self.consume(Tk.SEMICOLON, \"expect a ; after 'for' condition.\")\n increment = None\n if not self.check(Tk.RIGHT_PAREN):\n increment = self.expression()\n self.consume(Tk.RIGHT_PAREN,\n \"expect a ) at the end of the 'for' increment.\")\n body = self.statement()\n # Build the equivalent while loop\n # --> first create a while loop with condition + (body, increment)\n body = Block([body, Expression(increment)])\n if condition is None:\n condition = Literal(True)\n body = While(condition, body)\n # --> add the initializer\n if initializer is not None:\n body = Block([initializer, body])\n return body\n\n def blockstatement(self) -> List[Stmt]:\n statements = []\n while not self.is_at_end() and not self.check(Tk.RIGHT_BRACE):\n statements.append(self.declaration())\n self.consume(Tk.RIGHT_BRACE, \"expect } at the end of block.\")\n return statements\n\n def breakstatement(self) -> Stmt:\n self.consume(Tk.SEMICOLON, \"expect a ';' after a 'break'.\")\n return Break(self.previous())\n\n # def varstatement(self) -> Stmt:\n # try:\n # if self.match(Tk.VAR):\n # return self.vardeclaration()\n # return self.statement()\n # except ParserError as err:\n # pass\n # # self.synchronize(err)\n\n def printstatement(self) -> Stmt:\n value = self.expression()\n # consume the semicolon that should be at the end of the statement\n self.consume(Tk.SEMICOLON, \"expecting a ';' at the end of the line.\")\n return Print(value)\n\n def returnstatement(self) -> Stmt:\n keyword = self.previous()\n value = None\n if not self.check(Tk.SEMICOLON):\n value = self.expression()\n self.consume(\n Tk.SEMICOLON, \"expect a ';' at the end of the 'return' statement.\")\n return Return(keyword, value)\n\n def expressionstatement(self) -> Stmt:\n # Retrieve the expression in the statement\n expr = self.expression()\n # Consume the ; token\n self.consume(Tk.SEMICOLON, \"expecting a ; at the end of the line\")\n return Expression(expr)\n #\n # --------------------------------------\n # Expressions functions\n # --------------------------------------\n #\n\n def expression(self) -> Expr:\n return self.assignment()\n\n def assignment(self) -> Expr:\n # Consider the possible left hand side of the assignment as any expression\n expr = self.logic_or()\n # if we match the = we know it is a candidate for assignment\n if self.match(Tk.EQUAL):\n tk_equals = self.previous()\n # assignment are right associative, so we neeed to parse them right recursively\n right = self.assignment()\n # if the l-value is a possible variable, return the assignment\n if isinstance(expr, Variable):\n return Assign(expr.name, right)\n # we can also assign instances property\n elif isinstance(expr, Get):\n return Set(expr.getobject, expr.name, right)\n # we do not have a variable like, error\n raise ParserError(tk_equals, \"invalid assignment target\")\n # the expression is not an assignment\n return expr\n\n def logic_or(self) -> Expr:\n expr = self.logic_and()\n while (self.match(Tk.OR)):\n logic_op = self.previous()\n right = self.logic_and()\n expr = Logical(expr, logic_op, right)\n return expr\n\n def logic_and(self) -> Expr:\n expr = self.equality()\n while (self.match(Tk.AND)):\n logic_op = self.previous()\n right = self.equality()\n expr = Logical(expr, logic_op, right)\n return expr\n\n def equality(self) -> Expr:\n expr = self.comparison()\n\n while self.match(Tk.EQUAL_EQUAL, Tk.BANG_EQUAL):\n operator = self.previous()\n right = self.comparison()\n expr = Binary(expr, operator, right)\n return expr\n\n def comparison(self) -> Expr:\n expr = self.addition()\n\n while self.match(Tk.GREATER_EQUAL, Tk.GREATER,\n Tk.LESS, Tk.LESS_EQUAL):\n operator = self.previous()\n right = self.addition()\n expr = Binary(expr, operator, right)\n return expr\n\n def addition(self) -> Expr:\n expr = self.multiplication()\n\n while self.match(Tk.PLUS, Tk.MINUS):\n operator = self.previous()\n right = self.multiplication()\n expr = Binary(expr, operator, right)\n return expr\n\n def multiplication(self) -> Expr:\n expr = self.unary()\n\n while self.match(Tk.SLASH, Tk.STAR):\n operator = self.previous()\n right = self.unary()\n expr = Binary(expr, operator, right)\n return expr\n\n def unary(self) -> Expr:\n # this time we have a right associative operator\n while self.match(Tk.MINUS, Tk.BANG):\n operator = self.previous()\n right = self.unary()\n # we loop until all operators are consumed\n return Unary(operator, right)\n # no more operator, we have a call expr\n return self.call()\n\n def call(self) -> Expr:\n expr = self.primary()\n # Let's parse function call (passing the callee) as long as we have paranthesis\n while True:\n if self.match(Tk.LEFT_PAREN):\n expr = self.finishcall(expr)\n elif self.match(Tk.DOT):\n name = self.consume(\n Tk.IDENTIFIER, \"Expect a property name after the '.'\")\n expr = Get(expr, name)\n else:\n break\n return expr\n\n def finishcall(self, callee: Expr) -> Expr:\n arguments = []\n if not self.check(Tk.RIGHT_PAREN):\n while \"let's parse arguments as long as we have commas\":\n arguments.append(self.expression())\n if not self.match(Tk.COMMA):\n break\n if len(arguments) > LoxConstant.max_param:\n LoxError.error(self.peek(),\n \"function cannot have more than 8 arguments\")\n call_left_paren = self.consume(\n Tk.RIGHT_PAREN, \"expect a ')' at the end of a function call.\")\n return Call(callee, call_left_paren, arguments)\n\n def functionbody(self, kind: FunctionType):\n funcid = None\n functiontype = kind\n if kind is FunctionType.FUNCTION:\n funcid = self.consume(\n Tk.IDENTIFIER, \"Expect a function name after 'fun'.\")\n self.consume(Tk.LEFT_PAREN, \"expect a '(' after function name.\")\n elif kind is FunctionType.LAMBDA:\n self.consume(Tk.LEFT_PAREN, \"expect a '(' after 'fun' for lambda.\")\n elif kind is FunctionType.METHOD:\n funcid = self.consume(Tk.IDENTIFIER, \"Expect a method name.\")\n if funcid is LoxConstant.init_method:\n functiontype = FunctionType.INIT\n self.consume(Tk.LEFT_PAREN, \"expect a '(' after method name.\")\n else:\n raise ParserError(\n self.previous(), \"Unexpected function type in parser.\")\n parameters = []\n # All parameters should be identifiers\n if not self.check(Tk.RIGHT_PAREN):\n while \"we have comma after the parameter\":\n parameters.append(\n self.consume(Tk.IDENTIFIER, \"expect identifiers as function parameters.\"))\n if len(parameters) > LoxConstant.max_param:\n LoxError.error(self.previous(\n ), \"function can take at most \" + LoxConstant.max_param + \" parameters.\")\n if not self.match(Tk.COMMA):\n break\n self.consume(Tk.RIGHT_PAREN,\n \"expect a ) at the end of the function parameters.\")\n # Parse the body of the function\n self.consume(\n Tk.LEFT_BRACE, \"expect a { after function parameters list.\")\n body = self.blockstatement()\n return FunctionExp(funcid, parameters, body, functiontype)\n\n def primary(self) -> Expr:\n if self.match(Tk.FALSE):\n return Literal(False)\n\n if self.match(Tk.TRUE):\n return Literal(True)\n\n if self.match(Tk.NIL):\n return Literal(None)\n\n if self.match(Tk.NUMBER, Tk.STRING):\n return Literal(self.previous().literal)\n\n if self.match(Tk.IDENTIFIER):\n return Variable(self.previous())\n\n # I consider function definition in expression as lambda. Open to discussion\n if self.match(Tk.FUN):\n return self.functionbody(FunctionType.LAMBDA)\n\n if self.match(Tk.THIS):\n return This(self.previous())\n\n if self.match(Tk.SUPER):\n keyword = self.previous()\n self.consume(Tk.DOT, \"Expect a '.' after 'super'.\")\n method = self.consume(\n Tk.IDENTIFIER, \"Expect superclass method name after 'super.'.\")\n return Super(keyword, method)\n\n if self.match(Tk.LEFT_PAREN):\n expr = self.expression()\n self.consume(Tk.RIGHT_PAREN, \"Expect ')' after expr.\")\n return Grouping(expr)\n\n raise ParserError(self.peek(), \"expecting an expr.\")\n #\n # Parsing the list of statements\n\n def parse(self) -> List[Stmt]:\n try:\n statements = []\n while not self.is_at_end():\n statements.append(self.declaration())\n\n return statements\n except ParserError as e:\n print(e.message)\n","repo_name":"marcjourneux/pylox","sub_path":"lox/parser.py","file_name":"parser.py","file_ext":"py","file_size_in_byte":16570,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"61"} +{"seq_id":"43144010380","text":"'''\nLicensed under the Apache License, Version 2.0. See License.txt in the project root for license information.\n\nTests some different ways to Bootstrap nodes. \n(Note: The APITest does some of this type also!)\n\nCreated on Oct 17, 2014\n\n@author: dfleck\n'''\n\nfrom twisted.trial import unittest\nfrom twisted.internet import task, defer, reactor\nfrom twisted.python import log, failure\n\n\nfrom gmu.chord import NetworkUtils, Config\nfrom gmu.chord.NodeLocation import NodeLocation\nfrom gmu.chord.MetricsMessageObserver import MetricsMessageObserver\n\nfrom gmu.netclient.classChordNetworkChord import classChordNetworkChord\n\n# Testing Modules\nfrom ConnectivityCounter import ConnectivityCounter\nfrom SampleClient import SampleClient\nimport TestUtils\nimport sys, random\nfrom TestMessageObserver import TestMessageObserver\n\n\nnumNodes = 5 # Number of nodes per enclave\nnumMessages=5 # Total number of messages to send to each node\nstartingPort = 12350\n\n\nclass BootstrapTests(unittest.TestCase):\n\n @classmethod\n def setUpClass(cls):\n super(BootstrapTests, cls).setUpClass()\n BootstrapTests.logObs = log.startLogging(sys.stdout)\n \n \n @classmethod\n def tearDownClass(cls):\n super(BootstrapTests, cls).tearDownClass()\n if BootstrapTests.logObs is not None:\n BootstrapTests.logObs.stop()\n \n def setUp(self):\n '''Start the reactor so we don't have to do it in the nodes.'''\n\n # Turn ON warning for this test!\n Config.WARN_NO_MESSAGE_AUTHENTICATOR = True\n Config.ALLOW_NO_AUTHENTICATOR = False\n\n # This is the IP of the node. Note: This MUST be \n # an external ID or the code won't work!\n self.myIP = NetworkUtils.getNonLoopbackIP (None, None)\n log.msg('Got IP: %s:%s' % (self.myIP, type(self.myIP)))\n\n return\n \n \n \n\n\n \n \n @defer.inlineCallbacks\n def testReBootstrap(self):\n '''Create a BS node, then a client node, then kill the BS Node, wait, restart a BS node and \n check for re-Bootstrap\n '''\n global startingPort\n \n # Create Bootstrap\n port = 12345\n bootstrapNodeLocation = NodeLocation(None, self.myIP, port)\n self.allNodes = []\n self.allMetricsObservers = []\n self.allTestObservers = []\n\n # Build the BS node\n log.msg(\"building BS node...\")\n (status, bsClientAPI, bsNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port, None, 'theEnclave', False, True)\n self.assertTrue(status, 'Could not build bootstrap node')\n \n # Build the client node\n log.msg(\"building client node...\")\n (status, clClientAPI, clNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port+1, bootstrapNodeLocation, 'theEnclave', False, False)\n self.assertTrue(status, 'Could not build client node')\n \n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n self.assertTrue(connected, \"Client did not connect to the bootstrap node in testReBootstrap!\")\n \n # Now kill the BS node\n yield bsNetworkAPI.disconnect()\n bsNetworkAPI = None\n bsClientAPI = None\n \n # Gotta wait for disconnect to really finish\n yield TestUtils.waitForConnectionCache()\n \n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n self.assertTrue(not connected, \"Client remains connected to the bootstrap node in testReBootstrap after killing BS node!\")\n \n # Now startup another bootstrap node\n log.msg(\"building BS node...\")\n (status, bsClientAPI, bsNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port, None, 'theEnclave', False, True)\n self.assertTrue(status, 'Could not build second bootstrap node')\n \n # Wait for it to connect or fail -- basically waiting for the\n # maintenance call to run correctly.\n for _ in range(10):\n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n if connected:\n break\n yield TestUtils.wait(1)\n \n self.assertTrue(connected, \"Client could not re-bootstrap!\")\n\n \n # Now shut everything down\n \n log.msg(\"\\n\\ntestReBootstrap: Shutting down now...\\n\\n\")\n yield clNetworkAPI.disconnect()\n yield bsNetworkAPI.disconnect()\n\n if Config.USE_CONNECTION_CACHE:\n yield TestUtils.waitForConnectionCache()\n else: \n yield TestUtils.wait(5)\n\n defer.returnValue(True)\n \n \n @defer.inlineCallbacks\n def testReBootstrapNewNodeLoc(self):\n '''Create a BS node, then a client node, then kill the BS Node, wait, restart a BS node at a different location and \n check for re-Bootstrap\n '''\n global startingPort\n \n # Create Bootstrap\n port = 12345\n bootstrapNodeLocation = NodeLocation(None, self.myIP, port)\n self.allNodes = []\n self.allMetricsObservers = []\n self.allTestObservers = []\n\n # Build the BS node\n log.msg(\"building BS node...\")\n (status, bsClientAPI, bsNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port, None, 'theEnclave', False, True)\n self.assertTrue(status, 'Could not build bootstrap node')\n \n # Build the client node\n log.msg(\"building client node...\")\n (status, clClientAPI, clNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port+1, bootstrapNodeLocation, 'theEnclave', True, False)\n self.assertTrue(status, 'Could not build client node')\n \n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n self.assertTrue(connected, \"Client did not connect to the bootstrap node in testReBootstrap!\")\n \n # Now kill the BS node\n yield bsNetworkAPI.disconnect()\n bsNetworkAPI = None\n bsClientAPI = None\n \n \n # Gotta wait for network timeout disconnect to really finish.\n yield TestUtils.wait(30)\n \n \n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n self.assertTrue(not connected, \"Client remains connected to the bootstrap node in testReBootstrap after killing BS node!\")\n \n # Now startup another bootstrap node\n log.msg(\"building BS node...\")\n (status, bsClientAPI, bsNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port+3, None, 'theEnclave', False, True)\n self.assertTrue(status, 'Could not build second bootstrap node')\n \n # Wait for it to connect or fail -- basically waiting for the\n # maintenance call to run correctly.\n for _ in range(30):\n # Check that the client is connected to something\n connected = yield clNetworkAPI.isConnected()\n if connected:\n break\n yield TestUtils.wait(1)\n \n\n \n # Now shut everything down\n \n log.msg(\"\\n\\ntestReBootstrap: Shutting down now...\\n\\n\")\n yield clNetworkAPI.disconnect()\n yield bsNetworkAPI.disconnect()\n \n if Config.USE_CONNECTION_CACHE:\n yield TestUtils.waitForConnectionCache()\n else: \n yield TestUtils.wait(5)\n \n \n # This is the final question\n self.assertTrue(connected, \"Client could not re-bootstrap!\")\n\n defer.returnValue(True)\n \n \n @defer.inlineCallbacks\n def testDisconnectedBootstraps(self):\n '''Create a BS node and some clients. Create another bootstrap node and some clients (so we essentially have two rings). \n Verify, that the bootstrap nodes autodiscover each other and connect together \n '''\n global startingPort\n \n # Create Bootstrap\n port = 12345\n bootstrapNodeLocation = NodeLocation(None, self.myIP, port)\n bootstrapNodeLocation2 = NodeLocation(None, self.myIP, port+1)\n self.allNodes = []\n self.allMetricsObservers = []\n self.allTestObservers = []\n\n # Build the BS node\n (status, bsClientAPI, bsNetworkAPI) = yield TestUtils.startNodeUsingAPI(bootstrapNodeLocation.ip, bootstrapNodeLocation.port, None, 'theEnclave', True, True)\n self.allMetricsObservers.append(MetricsMessageObserver(bsNetworkAPI.chordNode))\n self.assertTrue(status, 'Could not build bootstrap node')\n\n # Build second BS node\n (status, bsClientAPI2, bsNetworkAPI2) = yield TestUtils.startNodeUsingAPI(bootstrapNodeLocation2.ip, bootstrapNodeLocation2.port, None, 'theEnclave', True, True)\n self.allMetricsObservers.append(MetricsMessageObserver(bsNetworkAPI2.chordNode))\n self.assertTrue(status, 'Could not build bootstrap node 2')\n\n \n # Build the client node\n (status, clClientAPI, clNetworkAPI) = yield TestUtils.startNodeUsingAPI(self.myIP, port+2, bootstrapNodeLocation, 'theEnclave', False, False)\n self.allMetricsObservers.append(MetricsMessageObserver(clNetworkAPI.chordNode)) \n self.assertTrue(status, 'Could not build client node')\n \n # Build the client node\n (status, clClientAPI2, clNetworkAPI2) = yield TestUtils.startNodeUsingAPI(self.myIP, port+3, bootstrapNodeLocation2, 'theEnclave', False, False)\n self.allMetricsObservers.append(MetricsMessageObserver(clNetworkAPI2.chordNode))\n self.assertTrue(status, 'Could not build client node')\n\n\n # Wait for flooding to reach all the nodes\n waiter = ConnectivityCounter()\n yield waiter.waitForConnectivity(3, clNetworkAPI.chordNode) # Does not count clNode itself.\n \n \n # Now shut everything down\n yield clNetworkAPI.disconnect()\n yield clNetworkAPI2.disconnect()\n yield bsNetworkAPI.disconnect()\n yield bsNetworkAPI2.disconnect()\n \n if Config.USE_CONNECTION_CACHE:\n yield TestUtils.waitForConnectionCache()\n else: \n yield TestUtils.wait(5)\n \n defer.returnValue(True)\n \n \n \n \n \n \n ","repo_name":"danfleck/Class-Chord","sub_path":"network-client/src/tests/BootstrapTests.py","file_name":"BootstrapTests.py","file_ext":"py","file_size_in_byte":10560,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"28533466495","text":"import sqlite3\nimport time\nfrom flask import Flask, jsonify, request, abort\nfrom argparse import ArgumentParser\n\n\nDB = 'db.sqlite'\n\n \napp = Flask(__name__)\n\n@app.route('/', methods=['GET'])\ndef home():\n return \"\"\"

    It works!

    \"\"\"\n\n@app.route('/api/login', methods=['POST'])\ndef login():\n if not request.json:\n abort(404)\n \n user_id = fetch_user_id(request.json[\"email\"], request.json[\"password\"])\n\n session_id = int(round(time.time() * 1000))\n changeData(\"\"\"\n INSERT INTO session (id, user_id)\n VALUES(?,?)\n \"\"\", (session_id, user_id))\n\n return jsonify({\n \"matching_user_id\": user_id,\n \"session_id\": session_id\n }), 200\n\n@app.route('/api/logout', methods=['POST'])\ndef logout():\n if not request.json:\n abort(404)\n \n return jsonify(changeData(\"\"\"\n DELETE FROM session\n WHERE id = ?\n \"\"\", (request.json[\"session_id\"],))\n ) , 200\n\n\n@app.route('/api/admin/see_table/', methods=['GET'])\ndef admin_get_table(table):\n parser = {\n 'user': parseUser,\n 'task': parseTask,\n 'reminder': parseReminder\n }\n data = fetchData(parser[table], f'SELECT * FROM {table}')\n return jsonify(data), 200\n\n \n@app.route('/api/task/', methods=['GET'])\ndef retrieve_task(session_id):\n user_id = fetch_user_id_using_session(session_id)\n tasks = fetch_task(user_id)\n for task in tasks:\n task[\"reminder\"] = fetch_reminder(task[\"id\"])\n return jsonify(tasks), 200\n\n@app.route('/api/task', methods=['POST'])\ndef create_task():\n if not request.json:\n abort(404)\n \n user_id = fetch_user_id_using_session(request.json[\"session_id\"])\n\n new_task = (\n user_id,\n request.json['title'],\n request.json['content'],\n request.json['pinned']\n )\n \n response = changeData(\"\"\"\n INSERT INTO task (user_id,title,content,pinned)\n VALUES(?,?,?,?)\n \"\"\", new_task)\n\n task_id = response[\"id\"]\n\n # Insert reminders\n reminders = request.json[\"reminders\"]\n for r in reminders:\n changeData(\"\"\"\n INSERT INTO reminder(task_id, date)\n VALUES(?,?)\n \"\"\", (task_id, r[\"date\"]))\n\n return jsonify(response), 200 \n\n@app.route('/api/task', methods=['DELETE'])\ndef delete_task():\n if not request.json:\n abort(404)\n \n if not fetch_user_id_using_session(request.json[\"session_id\"]):\n abort(404)\n\n response = changeData(\"\"\"\n DELETE FROM task WHERE id=?\n \"\"\", (request.json['task_id'],))\n \n return jsonify(response), 200\n\n@app.route('/api/task', methods=['PUT'])\ndef update_task():\n print(request.json[\"session_id\"])\n if not fetch_user_id_using_session(request.json[\"session_id\"]):\n abort(404)\n\n new_task = (\n request.json['title'],\n request.json['content'],\n request.json['pinned'],\n request.json['task_id'],\n )\n response = changeData(\"\"\"\n UPDATE task \n SET\n title = ?,\n content = ?,\n pinned = ?\n WHERE id = ?\n \"\"\", new_task)\n\n\n task_id = request.json['task_id']\n\n # Delete related reminders\n changeData(\"\"\"\n DELETE FROM reminder WHERE task_id=?\n \"\"\", (task_id,))\n\n # Update reminders\n reminders = request.json[\"reminders\"]\n\n for r in reminders:\n changeData(\"\"\"\n INSERT INTO reminder(task_id, date)\n VALUES(?,?)\n \"\"\", (task_id, r[\"date\"]))\n\n return jsonify(response), 200 \n\n\ndef fetch_reminder(task_id):\n return fetchData(parseReminder, \n \"\"\"\n SELECT * FROM reminder\n WHERE task_id = ?\n \"\"\", (task_id,))\n \ndef fetch_task(user_id):\n return fetchData(parseTask, \n \"\"\"\n SELECT * FROM task\n WHERE user_id = ?\n \"\"\", (user_id,))\n\ndef fetch_user_id_using_session(session_id):\n result = fetchData(parseSession,\n \"\"\"\n SELECT * FROM session\n WHERE id = ?\n \"\"\", (session_id,))\n\n if len(result) > 0:\n return result[0][\"user_id\"]\n else:\n return None\n\ndef fetch_user_id(email, password):\n result = fetchData(parseUser, \n \"\"\"\n SELECT * FROM user\n WHERE email = ?\n AND password = ?\n \"\"\", (email,password))\n \n if len(result) > 0:\n return result[0][\"id\"]\n else:\n return None\n\n\ndef parseSession(row):\n return {\n 'id': row[0],\n 'user_id': row[1],\n }\n\ndef parseUser(row):\n return {\n 'id': row[0],\n 'email': row[1],\n }\n \ndef parseTask(row):\n return {\n 'id': row[0],\n 'user_id': row[1],\n 'title': row[2],\n 'content': row[3],\n 'pinned': row[4],\n }\n\ndef parseReminder(row):\n return {\n 'id': row[0],\n 'task_id': row[1],\n 'date': row[2],\n }\n \ndef fetchData(parser, query, queryParam=None):\n db = sqlite3.connect(DB)\n cursor = db.cursor()\n if queryParam:\n cursor.execute(query, queryParam)\n else:\n cursor.execute(query)\n rows = cursor.fetchall()\n db.close()\n result = []\n for row in rows:\n result.append(parser(row))\n return result\n\n# Change means : INSERT, UPDATE or DELETE\ndef changeData(query, queryParam):\n db = sqlite3.connect(DB)\n cursor = db.cursor()\n cursor.execute(query, queryParam)\n id = cursor.lastrowid\n db.commit()\n response = {\n 'id': id,\n 'affected': db.total_changes,\n }\n db.close()\n return response\n \n\nif __name__ == '__main__':\n parser = ArgumentParser()\n parser.add_argument('-p', '--port', default=5000, type=int, help='port to listen on')\n args = parser.parse_args()\n port = args.port\n\n app.run(host='0.0.0.0', port=port)\n","repo_name":"wongjiahau/wireless-app-assignment-backend","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5792,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"29495187222","text":"import unittest\nfrom rnamake import sqlite_library, settings, motif, residue_type, util, motif_state_tree\nfrom rnamake.unittests import build\n\n\nclass BasicLibrariesUnittests(unittest.TestCase):\n\n def setUp(self):\n self.rts = residue_type.ResidueTypeSet()\n\n path = settings.RESOURCES_PATH + \"/motifs/base.motif\"\n self.base_motif = motif.file_to_motif(path)\n self.added_motif = motif.file_to_motif(path)\n\n self.base_ms = self.base_motif.get_state()\n self.added_ms = self.added_motif.get_state()\n\n def _test_correct_build(self, mlib, ms_lib):\n\n for m in mlib.all():\n m1 = motif.get_aligned_motif(self.base_motif.ends[1], m.ends[0], m)\n m2 = motif.get_aligned_motif(m1.ends[1],\n self.added_motif.ends[0],\n self.added_motif)\n\n ms = ms_lib.get(name=m.name,\n end_name=m.ends[0].name(),\n end_id=m.end_ids[0])\n\n ms1 = motif.get_aligned_motif_state_single(self.base_ms.end_states[1], ms)\n ms2 = motif.get_aligned_motif_state_single(ms1.end_states[1], self.added_ms)\n\n #print m2.ends[1].d()\n #print self.added_ms.end_states[1].d\n diff = ms2.end_states[1].diff(m2.ends[1].state())\n if diff > 0.01:\n self.fail(m.name + \" did not give the same answer as its state\")\n\n\n def test_correct_build_twoway(self):\n mlib = sqlite_library.MotifSqliteLibrary(\"twoway\")\n mlib.load_all()\n ms_lib = sqlite_library.MotifStateSqliteLibrary(\"twoway\")\n ms_lib.load_all()\n self._test_correct_build(mlib, ms_lib)\n\n def test_correct_build_nway(self):\n mlib = sqlite_library.MotifSqliteLibrary(\"nway\")\n mlib.load_all()\n ms_lib = sqlite_library.MotifStateSqliteLibrary(\"nway\")\n ms_lib.load_all()\n self._test_correct_build(mlib, ms_lib)\n\n\nclass LargeBuildUnittests(unittest.TestCase):\n\n def test_large_random_builds(self):\n for i in range(100):\n builder = build.BuildMotifTree()\n mt = builder.build(10)\n\n mst = motif_state_tree.MotifStateTree(mt=mt)\n mt_end= mt.last_node().data.ends[1].state()\n mst_end = mst.last_node().data.cur_state.end_states[1]\n\n diff = mst_end.diff(mt_end)\n if diff > 0.1:\n self.fail(\" did not give the same answer as its state\")\n\ndef main():\n unittest.main()\n\nif __name__ == '__main__':\n main()","repo_name":"zhuoyuzhang/RNAMake","sub_path":"rnamake/unittests/intergration/motif_compared_to_ms.py","file_name":"motif_compared_to_ms.py","file_ext":"py","file_size_in_byte":2568,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"20055595116","text":"from DB import DataBase\nfrom tkinter import END, ACTIVE, filedialog\nfrom copy import copy\nfrom PIL import ImageTk, Image\nfrom sqlite3 import IntegrityError\nfrom Tagger import Tagger\nfrom collections.abc import Iterable\nfrom os.path import isdir, join\nfrom threading import Thread\n\n\nclass Intermediary:\n\n def __init__(self, ui):\n\n self.ui = ui\n self.any = True\n self.queue = None\n self.curr_img = 0\n\n def confirmed(self, event):\n\n #Listbox items\n tags = self.ui.builder.get_object(\"ListSelected\").get(0, END)\n# Unique images' ids\n res = set()\n for tg in tags:\n ids = DataBase.get_tagged_items(tg)\n #TODO maybe sth better if nth found\n if len(ids) < 1:\n return\n# Add else intersection\n if self.any:\n res |= set(ids)\n else:\n res &= set(ids)\n\n lt = list()\n for r in res:\n lt.append(DataBase.get_path(r))\n self.queue_images(lt)\n\n def clear(self, event):\n \"\"\"Clears tags listbox\"\"\"\n\n #Listbox clearing\n self.ui.builder.get_object(\"ListSelected\").delete(0, END)\n\n def list_tag(self, event):\n \"\"\"Add tag to listbox ListSelected\"\"\"\n\n eadd = self.ui.builder.get_object(\"EAdd\")\n val = eadd.get()\n eadd.delete(0, END)\n\n self.ui.builder.get_object(\"ListSelected\").insert(END, val)\n\n def remove_tag(self, event):\n \"\"\"Remove tag from listbox ListSelected\"\"\"\n\n event.widget.delete(ACTIVE)\n\n def rany(self, event):\n \"\"\"Changes search method\"\"\"\n\n var = self.ui.builder.get_variable(\"VarAny\").get()\n if var != \"Any\":\n self.any = True\n else:\n self.any = False\n\n def queue_images(self, ids):\n \"\"\"Queue images to display\"\"\"\n\n# Converting to list from different types of arguments\n if isinstance(ids, str):\n ids = [ids]\n elif isinstance(ids, Iterable):\n ids = list(ids)\n\n self.queue = copy(ids)\n self.curr_img = -1\n self.list_queue()\n self.show_image()\n\n def show_image(self, pth=None):\n \"\"\"\n Display image, if no argument is present get from queue. Called automatically when queue changes\n :arg pth: Path to image\n \"\"\"\n\n# Next image from queue\n if pth is None:\n self.curr_img += 1\n self.curr_img %= len(self.queue)\n pth = self.queue[self.curr_img]\n else:\n self.queue_images(list(pth))\n\n# Wrong path\n if pth is None:\n return\n\n# Prepare image\n img = Image.open(pth)\n\n if img.width > 800 or img.height > 450:\n factor = min(800/img.width, 450/img.height)\n img = img.resize((int(img.width*factor), int(img.height*factor)))\n\n img = ImageTk.PhotoImage(img)\n\n# Display image\n label = self.ui.builder.get_object(\"LImage\")\n label.config(image=img)\n label.image = img\n\n# Mark current image in listbox\n lb = self.ui.builder.get_object(\"ListResults\")\n lb.selection_clear(0, END)\n lb.selection_set(self.curr_img)\n\n# List tags\n self.list_image_tags()\n\n def list_queue(self):\n \"\"\"Called by queue_images. Lists queued paths in ListResults listbox\"\"\"\n\n lb = self.ui.builder.get_object(\"ListResults\")\n self.clear_results()\n for pth in self.queue:\n lb.insert(END, pth.split(sep=\"\\\\\")[-1])\n\n def list_image_tags(self):\n \"\"\"Called by show image. Adds current tags to ListTags listbox\"\"\"\n\n tags = DataBase.get_image_tags(pth=self.queue[self.curr_img])\n lt = self.ui.builder.get_object(\"ListTags\")\n self.clear_tags()\n for tag in tags:\n lt.insert(END, tag)\n\n def clear_results(self):\n\n self.ui.builder.get_object(\"ListResults\").delete(0, END)\n\n def clear_tags(self):\n\n self.ui.builder.get_object(\"ListTags\").delete(0, END)\n\n def path_input(self, pth):\n \"\"\"Handle request for new input file. If new tag and display else display\"\"\"\n\n self.clear_results()\n tfiles = None\n\n# Single image path\n if not isdir(pth):\n new = True\n# SQL exception if path is not unique\n try:\n DataBase.add_image(pth)\n except IntegrityError:\n new = False\n\n# Tag new\n if new:\n tags = Tagger.tag_file(pth)\n for tag in tags:\n DataBase.tag_image(tag, pth=pth)\n# Directory path\n else:\n tags, tfiles = Tagger.tag_dir(pth)\n for i in range(len(tfiles)):\n f = tfiles[i]\n# Full path to image\n fpth = join(pth, f)\n tfiles[i] = fpth\n# Continue if already present\n if DataBase.exists(pth=fpth):\n continue\n else:\n DataBase.add_image(fpth)\n# Tuple results\n if not isinstance(tags[i], str):\n for t in tags[i]:\n DataBase.tag_image(t, pth=fpth)\n# String result\n else:\n DataBase.tag_image(tags[i], pth=fpth)\n\n L = 1\n# Display\n if tfiles is None:\n self.queue_images(pth)\n else:\n #Number of listbox for results length\n L = len(tfiles)\n nlb = max(3, L)\n nlb = min(nlb, 12)\n self.ui.builder.get_object(\"ListResults\").config(height=nlb)\n\n self.queue_images(tfiles)\n\n self.update_info(\"Processed \" + str(L) + \" images\")\n\n def choose_dir(self, event):\n\n directory = filedialog.askdirectory()\n self.path_input(directory)\n\n def choose_file(self, event):\n\n file = filedialog.askopenfilename(title=\"Select file\", filetypes=((\"jpeg files\", \"*.jpg\"),\n (\"jpeg files\", \"*.jpeg\")))\n self.path_input(file)\n\n def next_image(self, event):\n\n self.show_image()\n\n def prev_image(self, event):\n\n self.curr_img -= 2\n if self.curr_img < 0:\n self.curr_img = -1\n self.show_image()\n\n def listbox_image(self, event):\n\n idx = event.widget.index(ACTIVE)\n self.curr_img = idx - 1\n self.show_image()\n\n def update_info(self, info=\" \"):\n\n linfo = self.ui.builder.get_object(\"LInfo\")\n linfo.configure(text=info)\n","repo_name":"Maciej-R/CV_Object_Detection","sub_path":"Intermediary.py","file_name":"Intermediary.py","file_ext":"py","file_size_in_byte":6656,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"15650192941","text":"# otvoriti romeo.txt file\nfhand = open('romeo.txt')\n\n# ucitati sve rijeci u dict i prikazati koliko puta se rijec ponavlja\ncounts = dict()\n\nfor line in fhand:\n words = line.split()\n # varijabla words je objekt list sastavljen od stringova iz te linije\n for word in words:\n counts[word] = counts.get(word, 1) + 1\n\n# da bi vrijednosti u dictionary sortirali, moramo od njega napraviti listu\n\nprint ( sorted( [ (v,k) for k,v in counts.items() ] ) )\n'''\n[(1, 'b'), (10,'b'), (22, 'c')]\n'''\n","repo_name":"thecodereule/exercises","sub_path":"Tuples2.py","file_name":"Tuples2.py","file_ext":"py","file_size_in_byte":502,"program_lang":"python","lang":"hr","doc_type":"code","stars":1,"dataset":"github-code","pt":"61"} +{"seq_id":"11775942028","text":"import numpy as np\nimport util\nimport sys\nfrom random import random\n\nsys.path.append('../linearclass')\n\n# NOTE : You need to complete logreg implementation first!\n\nfrom logreg_loop_version import LogisticRegression\n\n# Character to replace with sub-problem letter in plot_path/save_path\nWILDCARD = 'X'\n# Ratio of class 0 to class 1\nkappa = 0.1\n\n\ndef main(train_path, validation_path, save_path):\n \"\"\"Problem 2: Logistic regression for imbalanced labels.\n\n Run under the following conditions:\n 1. naive logistic regression\n 2. upsampling minority class\n\n Args:\n train_path: Path to CSV file containing training set.\n validation_path: Path to CSV file containing validation set.\n save_path: Path to save predictions.\n \"\"\"\n output_path_naive = save_path.replace(WILDCARD, 'naive')\n output_path_upsampling = save_path.replace(WILDCARD, 'upsampling')\n\n # *** START CODE HERE ***\n # Part (b): Vanilla logistic regression\n # Make sure to save predicted probabilities to output_path_naive using np.savetxt()\n\n x_train, y_train = util.load_dataset(train_path, add_intercept=True)\n x_valid, y_valid = util.load_dataset(validation_path, add_intercept=True)\n y_train = np.expand_dims(y_train, 1)\n #\"\"\"\n vanilla = LogisticRegression()\n vanilla.fit(x_train, y_train)\n y_prediction = vanilla.predict(x_valid)\n np.savetxt(output_path_naive, y_prediction)\n util.plot(x_valid, y_valid, vanilla.theta, '/Users/cindyxu/Documents/cs229/assignments/ps1/src/imbalanced/vanilla.png')\n return confusionMatrix(y_valid, y_prediction)\n #\"\"\"\n # Part (d): Upsampling minority class\n # Make sure to save predicted probabilities to output_path_upsampling using np.savetxt()\n # Repeat minority examples 1 / kappa times\n\n \"\"\"\n m = x_train.shape[0]\n for i in range(m):\n if y_train[i, 0] == 1.0:\n for j in range(int(1/kappa)):\n x_train = np.vstack((x_train, x_train[i, :]))\n y_train = np.vstack((y_train, y_train[i, :]))\n\n upsampling = LogisticRegression()\n upsampling.fit(x_train, y_train)\n y_prediction_up = upsampling.predict(x_valid)\n np.savetxt(output_path_upsampling, y_prediction_up)\n util.plot(x_valid, y_valid, upsampling.theta, '/Users/cindyxu/Documents/cs229/assignments/ps1/src/imbalanced/upsampling.png')\n return confusionMatrix(y_valid, y_prediction_up)\n \"\"\"\n # *** END CODE HERE\n\n\ndef confusionMatrix(y_true, y_pred):\n TN, FP, FN, TP = 0, 0, 0, 0\n for i in range(y_pred.shape[0]):\n if y_pred[i, 0] > 0.5:\n y_pred[i, 0] = 1.0\n if y_pred[i, 0] < 0.5:\n y_pred[i, 0] = 0.0\n\n if y_pred[i, 0] == y_true[i]:\n if y_pred[i, 0] == 0.0:\n TN += 1\n if y_pred[i, 0] == 1.0:\n TP += 1\n if y_pred[i, 0] - y_true[i] == 1.0:\n FP += 1\n if y_pred[i, 0] - y_true[i] == -1.0:\n FN += 1\n\n A = (TP + TN) / y_true.shape[0]\n A0, A1 = TN / (TN + FP), TP / (TP + FN)\n A_balanced = 1 / 2 * (A0 + A1)\n print(\"A: \", A)\n print(\"A0: \", A0)\n print(\"A1: \", A1)\n print(\"A_balanced: \", A_balanced)\n return A, A_balanced, A0, A1\n\n\nif __name__ == '__main__':\n main(train_path='/Users/cindyxu/Documents/cs229/assignments/ps1/src/imbalanced/train.csv',\n validation_path='/Users/cindyxu/Documents/cs229/assignments/ps1/src/imbalanced/validation.csv',\n save_path='/Users/cindyxu/Documents/cs229/assignments/ps1/src/imbalanced/imbalanced_X_pred.txt')\n","repo_name":"xingzix/cs229hw","sub_path":"ps1/src/imbalanced/imbalanced.py","file_name":"imbalanced.py","file_ext":"py","file_size_in_byte":3548,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"35380592369","text":"import cv2 as cv\r\nimport numpy as np\r\n\r\n# img_path = 'C:/Users/Sam_She/Desktop/dinasaur.jpg'\r\nimg_path = './yun.png'\r\n\r\n\r\n# img = np.reshape(cv.imread(img_path), (80, 80, 3))\r\nimg = cv.imread(img_path, -1) \r\nmBackground = img[:,:,3] == 0\r\nimg[mBackground] = (255,255,255,255)\r\n\r\nimg = img[:,:,0:3]\r\n\r\n\r\n# fix rgba color, refer to stackoverflow.com/questions/3803888/how-to-load-png-images-with-4-channels\r\nh, w = img.shape[0:2]\r\nh = h//2\r\nw = w//2\r\n\r\n# img = np.reshape(img, (h//2, w//2, 3))\r\n# h, w = img.shape[0:2]\r\n# img = np.reshape(img, (h//2, w//2, 3))\r\n# h, w = img.shape[0:2]\r\n\r\nresult_array = [0 for i in range(h*w)]\r\ndemo = np.zeros([h,w,3],dtype='uint8')\r\n\r\n\r\nfor i in range(h*w):\r\n y = (i//w) * 2\r\n x = (i%w) * 2\r\n bgr = img[y, x]\r\n u16 = (int(bgr[2]/255*31)<<11) + (int(bgr[1]/255*63)<<5) +( int(bgr[0]/255*31))\r\n result_array[i] = u16 # draw point\r\n demo[i//w, i%w] = bgr\r\n # result_array[x+y*80] = u16\r\n # print(y,x)\r\nresult_array = [str(hex(i)) for i in result_array]\r\nresult = ','.join(result_array)\r\nprint(result)\r\n\r\n\r\ncv.imwrite('img.png', demo)\r\nfor i in range(h):\r\n\tfor j in range(w):\r\n\t\tif (result_array[j + w * i]) == result_array[0]:\r\n\t\t\tprint(\"-\", end=\"\")\r\n\t\telse:\r\n\t\t\tprint(\"0\", end=\"\")\r\n\tprint(\"\")\r\ncv.imwrite('img.png', demo)\r\n\r\n# for i in range(img.shape[0]):\r\n# \tfor j in range(img.shape[1]):\r\n# \t\tif (img[i,j,1]) >= 100:\r\n# \t\t\tprint(\"-\", end=\"\")\r\n# \t\telse:\r\n# \t\t\tprint(\"0\", end=\"\")\r\n# \tprint(\"\")\r\n","repo_name":"wangsy503/Dinosaur-Game","sub_path":"ref_pics/imgProcess.py","file_name":"imgProcess.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"71026378114","text":"import os\nimport subprocess \nimport json\nimport smtplib\nimport datetime\nimport time\nfrom threading import Thread\nfrom email.mime.text import MIMEText\nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.base import MIMEBase\nfrom email import encoders\n\n#Extract properties \ndef getProperties():\n with open('configP2P.json', 'r') as file:\n properties = json.load(file)\n return properties\n\ndef run_torrent(torrentName):\n os.system('deluged')\n time.sleep(5)\n os.system('deluge-console add ' + torrentName)\n\ndef unlink_torrent(fileName):\n os.system('deluge-console rm ' + fileName)\n os.system('rm /home/maro96leon/Downloads/*')\n\ndef check_status():\n status = subprocess.check_output(\"deluge-console info\", shell=True).decode()\n if \"Seeding\" not in status:\n return False\n return True\n\ndef gitUpdate():\n os.system('git pull')\n\ndef logStartNetstat(i, n):\n os.system(\"netstat -s | grep segments >> Logs/Netstat_T\" + str(n) + \"_C\" + str(i) + \"_Start.log\")\n\ndef logEndNetstat(i, n):\n os.system(\"netstat -s | grep segments >> Logs/Netstat_T\" + str(n) + \"_C\" + str(i) + \"_End.log\")\n\ndef handleIfTop(t, i, n):\n os.system(\"sudo iftop -t -s \" + str(t) + \" >> Logs/P2P_T\" + str(n) + \"_C\" + str(i) + \"_traffic.log\")\n\ndef makeDirFile():\n os.system('rm -rf Logs')\n os.system('mkdir Logs')\n\ndef sout(l):\n log.write(l + '\\n')\n log.flush()\n print(l)\n\ndef getId():\n with open('id','r') as f:\n nId = json.load(f)['id']\n return nId\n\ndef send_mail_gmail(username,password,sender,toaddrs_list,msg_text,subject,attachment_path_list):\n s = smtplib.SMTP('smtp.gmail.com:587')\n s.starttls()\n s.login(username, password)\n #s.set_debuglevel(1)\n msg = MIMEMultipart()\n recipients = toaddrs_list\n msg['Subject'] = subject\n msg['From'] = sender\n msg['To'] = \", \".join(recipients)\n if attachment_path_list is not None:\n for each_file_path in attachment_path_list:\n try:\n file_name=each_file_path.split(\"/\")[-1]\n part = MIMEBase('application', \"octet-stream\")\n part.set_payload(open(each_file_path, \"rb\").read())\n\n encoders.encode_base64(part)\n part.add_header('Content-Disposition', 'attachment' ,filename=file_name)\n msg.attach(part)\n except:\n print(\"could not attach file\")\n msg.attach(MIMEText(msg_text,'html'))\n s.sendmail(sender, recipients, msg.as_string())\n\ngitUpdate()\nmakeDirFile()\nproperties = getProperties()\ni = getId()\nprint(\"Running client #\" + i)\nt = properties['runtime']\nn = properties['numberClients']\nunlink_torrent(properties['fileName'])\nthread = Thread(target=handleIfTop, args=[t, i, n])\nwith open('Logs/P2P_T' + str(n) + \"_C\" + str(i) + \".log\", 'w') as log:\n thread.start()\n logStartNetstat(i, n)\n run_torrent(properties['torrentName'])\n tStart = datetime.datetime.now()\n done = False\n while not done:\n done = check_status()\n summary = str(datetime.datetime.now() - tStart) + \"s\"\n logEndNetstat(i, n)\n sout(\"C: Transfered in \" + summary)\nthread.join()\nunlink_torrent(properties['fileName'])\nlogs = ['Logs/P2P_T' + str(n) + \"_C\" + str(i) + \".log\",\n\"Logs/P2P_T\" + str(n) + \"_C\" + str(i) + \"_traffic.log\",\n\"Logs/Netstat_T\" + str(n) + \"_C\" + str(i) + \"_End.log\",\n\"Logs/Netstat_T\" + str(n) + \"_C\" + str(i) + \"_Start.log\"]\nsend_mail_gmail(properties['email'], properties['passwd'],properties['email'],properties['dest'],\n\"Working OK\", \"Logs for T\" + str(n) + \" C\" + str(i) + \" \" + properties['fileName'], logs)\n","repo_name":"JkRuiz/RedesLab4","sub_path":"P2P/p2pRunner.py","file_name":"p2pRunner.py","file_ext":"py","file_size_in_byte":3591,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"61"} +{"seq_id":"39032700583","text":"#!/usr/bin/env python\nimport os\nimport sqlite3\nimport xml.etree.ElementTree as ET # for xml\n\n# 读取数据库并存入 whitelist_dics {{{\n# whitelist_dics = {table_name_list[i]:{key_domain:(prefix, suffix, score, description)}}\n\nconn = sqlite3.connect('whitelists.db') # 如果文件不存在,会自动在当前目录创建:\ncursor = conn.cursor() # 创建一个Cursor:\n\n\n# get all table_name from database and save it to the variable {table_name_list} {{{\ntable_name_list = []\n\ncursor.execute(\"SELECT name FROM sqlite_master WHERE type='table';\")\ntable_names = cursor.fetchall() \n# print(table_names) # [('wiki',), ('blogs',), ('library',), ('software',), ('video',), ('repository',), ('bbs',)]\nfor _ in table_names:\n table_name_list.append(_[0])\n\n#print(table_name_list)\n# }}}\n\n\nwhitelist_dics = {} # {table_name_list[i]:{key_domain:(prefix, suffix, score, description)}}\n\nfor i in range(len(table_name_list)):\n sq = f\"select * from {table_name_list[i]};\"\n cursor.execute(sq)\n data_all = cursor.fetchall()\n \n #print(data_all) #[(1,2,3,4,5),(1,2,3,4,5),]\n\n # 存入 whitelist_dics {{{\n tmp_dic = {}\n\n key_domain = ''\n tmp_lis = []\n\n\n for data in data_all:\n key_domain = data[0]\n tmp_lis = data[1:]\n\n tmp_dic[key_domain] = tmp_lis\n\n whitelist_dics[table_name_list[i]] = tmp_dic\n\n # }}}\n \n \n\n#print(whitelist_dics)\n\n\ncursor.close() # 关闭Cursor:\nconn.commit() # 提交事务:\nconn.close() # 关闭Connection:\n\n# }}}\n\nlis = [] # 临时列表\nlis_total = [] # 总列表,递增添加所有名单,不减少\n\n\n# generate urls_list {{{\n\ndef gen_urls_list(whitelist_dic, startwith_at=False):\n if not startwith_at : # @ 符号是 uBlacklist 的白名单的前缀\n '''\n uBlacklist whitelist rule:\n with prefix\n @*://*.prefix.domain_name.suffix/*\n no prefix\n @*://*.domain_name.suffix/*\n no prefix and no domain name, only the suffix\n @*://*.suffix/*\n '''\n # k,v 即 domain:[prefix, suffix, score, description]\n # 将内容存到临时列表 lis = [url, url,,,], url='@*://prefix.domain_name.suffix'\n for k,v in whitelist_dic.items(): \n url = '@*://*.'\n # 加前缀\n if v[0] != '':\n if v[0].startswith('http://') or v[0].startswith('https://'):\n url = '@'+v[0]+'.' # uBlacklist suppot rules like \"@https://www.cnblogs.com/*\"\n \n elif v[0].startswith('www'): # www.cnblogs.com/*\n # 在 cse.google.com 中,\"*.www.cnblogs.com/*\" 不会匹配 https://www.cnblogs.com/*\n # 但是 \"*.my.oschina.net/*\" 能匹配到 https://my.oschina.net/*\n url = \"@https://\"+v[0]+'.'\n else:\n url+=v[0] + '.'\n # 加域名\n if k != '':\n url += k.lower() # uBlacklist 对域名区分大小写 @*://*.stackoverflow.com/* 与 @*://*.StackOverflow.com/* 拦截效果不同\n else:\n url = url[:-1] # change url('@*://*.') to '@*://*', 为了添加指定后缀的域名,如 @*://*.edu\n\n # 加后缀\n if v[1] != '': \n # @*://*.docin.com/p-* , 后缀以 \"p-\" 开头,如 “https://www.docin.com/p-1706944942.html”\n if '/' in v[1]:\n url+='.'+v[1]+'*' # 这个可以取代下面的写法\n # 添加完全后缀, @*://*.mathsisfun.com/*\n else: \n url+='.'+v[1]+'/*'\n\n #print(url)\n lis.append(url)\n lis_total.append(url)\n else:\n # for google cse annotations\n # lis = [[url,score,description], [url,score,description],,, ]\n '''\n cse whitelist rule:\n with prefix\n *.prefix.domain_name.suffix/*\n https://prefix.domain_name.suffix/*\n\n no prefix\n *.domain_name.suffix/*\n no prefix and no domain name, only the suffix\n *.suffix/*\n '''\n for k,v in whitelist_dic.items(): \n url = '*.'\n # 加前缀\n if v[0] != '':\n if v[0].startswith('http://') or v[0].startswith('https://'): # http(s)://www\n url = v[0]+'.' \n elif v[0].startswith('www'): # www.cnblogs.com/*\n # 在 cse.google.com 中,\"*.www.cnblogs.com/*\" 不会匹配 https://www.cnblogs.com/*\n # 但是 \"*.my.oschina.net/*\" 能匹配到 https://my.oschina.net/*\n url = \"https://\"+v[0]+'.'\n else:\n url+=v[0] + '.'\n # 加域名\n if k != '':\n url += k.lower()\n else:\n url = url[:-1] # 为了添加指定后缀的域名,如 *.edu\n\n # 加后缀, *.docin.com/p-* , 后缀以 \"p-\" 开头,如 “https://www.docin.com/p-1706944942.html”\n if v[1] != '': \n if '/' in v[1]:\n url+='.'+v[1]+'*'\n # 添加完全后缀, *.mathsisfun.com/*\n else: \n url+='.'+v[1]+'/*'\n\n #print(url)\n # annotations.xml 中的 score 是字符串格式: https://developers.google.com/custom-search/docs/annotations\n lis.append([url, str(v[2]), v[3]]) # lis = [[url,str(score),description], [url,str(score),description],,, ]\n lis_total.append(url)\n ...\n# }}}\n\n# uBlacklist txt subscription txt {{{\ndef gen_subscription_txt():\n # generate whitelist rule text\n with open( output + '/whitelist.txt', 'w') as f:\n f.write(r'*://*/*')\n\n for k,v in whitelist_dics.items():\n #print(k, v)\n gen_urls_list(v)\n filename = output + '/' + k + '.txt'\n with open(filename, 'w') as f:\n for each in lis:\n f.write(each+'\\n')\n\n lis.clear()\n# }}}\n\n# 汇总 txt {{{\n# 汇总列表,for uBlacklist\ndef gen_subscription_combined_txt():\n with open(output + '/whitelists_combined.txt', 'w') as f:\n for each in lis_total:\n f.write(each+'\\n')\n\n# 汇总域名列表,for other ways:\"cse.google.com\",油猴插件\ndef gen_domain_name_txt():\n with open(output + '/domain_name.txt', 'w') as f:\n for each in lis_total:\n if each.startswith('@http'): # @http(s)://www.cnblogs.com/*\n f.write(each[1:]+'\\n') # http(s)://www.cnblogs.com/*\n else:\n f.write(each[5:]+'\\n') # *.prefix.domain_name.suffix\n #print(each[5:])\n\n# }}}\n\n# 增加换行符 {{{\n# https://vae-0118.github.io/2017/11/06/Python%E4%B8%ADXML%E7%9A%84%E8%AF%BB%E5%86%99%E6%80%BB%E7%BB%93/\ndef __indent(elem, level=0):\n i = \"\\n\" + level*\"\\t\"\n if len(elem):\n if not elem.text or not elem.text.strip():\n elem.text = i + \"\\t\"\n if not elem.tail or not elem.tail.strip():\n elem.tail = i\n for elem in elem:\n __indent(elem, level+1)\n if not elem.tail or not elem.tail.strip():\n elem.tail = i\n else:\n if level and (not elem.tail or not elem.tail.strip()):\n elem.tail = i\n# }}}\n\n# facet_items {{{\n# weight from -1.0 to 1.0\nfacet_items = {\n 'wiki':{'Label_name':'wiki',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.9',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_','/m/05z1_','/m/01mkq','/m/04rjg']},\n 'bbs':{'Label_name':'bbs',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.8',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_','/m/05z1_','/m/01mkq','/m/04rjg']},\n 'repository':{'Label_name':'repository',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.8',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_','/m/05z1_','/m/01mkq','/m/04rjg']},\n 'blogs':{'Label_name':'blogs',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.7',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_','/m/05z1_','/m/01mkq','/m/04rjg']},\n 'library':{'Label_name':'library',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.4',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_','/m/05z1_','/m/01mkq','/m/04rjg']},\n 'software':{'Label_name':'software',\n 'Label_mode':'FILTER',\n 'Label_weight':'0.5',\n 'Label_enable_for_facet_search':'false',\n 'Rewrite':'',\n 'entities':[]},\n 'pdf':{'Label_name':'pdf',\n 'Label_mode':'BOOST',\n 'Label_weight':'0.5',\n 'Label_enable_for_facet_search':'false',\n 'Rewrite':'filetype:pdf',\n 'entities':[]},\n 'video':{'Label_name':'video',\n 'Label_mode':'FILTER',\n 'Label_weight':'0',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'',\n 'entities':['/m/01mf_']},\n 'edu':{'Label_name':'edu',\n 'Label_mode':'BOOST',\n 'Label_weight':'0.1',\n 'Label_enable_for_facet_search':'true',\n 'Rewrite':'site:.edu',\n 'entities':[]},\n}\n# }}}\n\n# 这个文件或许手修改更方便, 所以只生成该文件中的标签部分 {{{\ndef gen_cse_xml():\n root = ET.Element('Facet') # 创建根节点\n tree = ET.ElementTree(root) # 创建文档\n\n\n for facet in list(facet_items.values()):\n FacetItem = ET.Element('FacetItem') # 子节点\n\n Label = ET.SubElement(FacetItem, 'Label')\n Label.set('name', facet['Label_name'])\n Label.set('mode', facet['Label_mode'])\n Label.set('weight',facet['Label_weight'])\n Label.set('enable_for_facet_search',facet['Label_enable_for_facet_search'])\n\n Rewrite = ET.SubElement(Label, 'Rewrite')\n rewrite_text = facet['Rewrite']\n if rewrite_text != '':\n Rewrite.text = rewrite_text\n\n entities = ET.SubElement(Label, 'entities')\n for mid in facet['entities']:\n entity = ET.SubElement(entities, 'entity')\n entity.set('mid', mid)\n\n Title = ET.SubElement(FacetItem, 'Title')\n Title.text = facet['Label_name']\n\n root.append(FacetItem) # 放到根节点下\n\n __indent(root) # 增加换行符\n tree.write(output + '/cse_FacetLabels.xml', encoding='utf-8', xml_declaration=True)\n ...\n\n# }}}\n\n# generate annotations.xml {{{\n# https://vae-0118.github.io/2017/11/06/Python%E4%B8%ADXML%E7%9A%84%E8%AF%BB%E5%86%99%E6%80%BB%E7%BB%93/\ndef gen_annotations_xml():\n total_length = 0\n\n root = ET.Element('Annotations') # 创建根节点\n tree = ET.ElementTree(root) # 创建文档\n\n # add Annotation {{{\n\n # whitelists 总字典 { 'bbs':{'':[],'':[]}, 'blogs':{} }\n for k,v in whitelist_dics.items():\n #print('k = {}, v={}'.format(k, v)) # 'wiki':{'domain':(prefix,suffix,score,description)}\n # 将 k 对应的字典转为存储到临时列表 lis\n gen_urls_list(v, True)\n # print(lis) # [[url, score, description], ['*.stackoverflow.com/*', '0.8', description]]\n\n # 遍历列表\n for each in lis:\n # each 的属性,权重\n element = ET.Element('Annotation') # 子节点\n element.set('about', each[0]) # about 存 url pattern\n element.set('score', each[1]) # str(score)\n\n Label = ET.SubElement(element, 'Label')\n Label.set('name', '_include_')\n Comment = ET.SubElement(element, 'Comment')\n Comment.text = each[2] # description\n\n # 添加标签,如