diff --git "a/802.jsonl" "b/802.jsonl" new file mode 100644--- /dev/null +++ "b/802.jsonl" @@ -0,0 +1,91 @@ +{"seq_id":"21758922950","text":"from django.http import HttpResponse\nfrom django.http import HttpResponse\nfrom django.template import loader\nfrom django.shortcuts import render\n\ndef saludo(request):\n mensaje = 'PaRa InGrEsAr a NuEsTrO SiTiO, POR FAVOR sElEcCiOnA aquí.'\n return HttpResponse(mensaje)\n\ndef segunda_vista(request):\n return HttpResponse(\"

Mi primer proyecto con Django

\")\n\ndef miNombreEs(request, nombre):\n data = f\"Mi nombre es:

{nombre}

\"\n return HttpResponse(data)\n\ndef probandoTemplate(request):\n nombre = \"Diego\"\n apellido = \"Garcia\"\n\n namelist = [\"Diego\", \"Alumno1\", \"Alumno2\", \"Alumno3\", \"Alumno4\"]\n\n diccionario = {\n \"nombre\": nombre,\n \"apellido\": apellido,\n \"namelist\": namelist\n }\n\n plantilla = loader.get_template(\"template1.html\")\n documento = plantilla.render(diccionario)\n return HttpResponse(documento)\n","repo_name":"diegobocaju/TercerPreEntregaGarcia","sub_path":"Proyecto1/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":938,"program_lang":"python","lang":"es","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"23612195761","text":"import os\nimport argparse\nimport shutil\nimport re\n\nparser = argparse.ArgumentParser()\nparser.add_argument('-e', '--exported')\n\nargs = parser.parse_args()\nexports = []\n\nif args.exported:\n\texports = args.exported.split(',')\n\nbase_dir = os.path.dirname(os.path.realpath(__file__))\nbuild_dir = os.path.join(base_dir, 'build')\n\nos.makedirs(build_dir, exist_ok=True)\n\ndef get_modinfo_version():\n\twith open(os.path.join(base_dir, 'modinfo.lua'), 'r') as modinfo:\n\t\tcontent = modinfo.read()\n\t\tm = re.search(r'version.*=.*\\\"(.+)\\\"', content)\n\t\treturn m.group(1)\n\nversion = get_modinfo_version()\nprint(\"VERSION = {}\".format(version))\n\nmod_build_dir = os.path.join(build_dir, \"Ozzy {}\".format(version))\nshutil.rmtree(mod_build_dir, ignore_errors=True)\nos.makedirs(mod_build_dir, exist_ok=True)\n\nfor d in ('anim', 'scripts', 'bigportraits', 'images', 'sound'):\n\tshutil.copytree(os.path.join(base_dir, d), os.path.join(mod_build_dir, d))\n\nfor f in ('modicon.tex', 'modicon.xml', 'modinfo.lua', 'modmain.lua', 'modpic.png'):\n\tshutil.copy2(os.path.join(base_dir, f), os.path.join(mod_build_dir, f))\n\nif len(exports) > 0:\n\tfor d in exports:\n\t\tshutil.copytree(os.path.join(base_dir, 'exported', d), os.path.join(mod_build_dir, 'exported', d))\n\ndef rm_files_without_exts(path, exts):\n\tfor item in os.listdir(path):\n\t\titem_path = os.path.join(path, item)\n\n\t\tif os.path.isdir(item_path):\n\t\t\trm_files_without_exts(item_path, exts)\n\t\telse:\n\t\t\tfilename, file_ext = os.path.splitext(item_path)\n\n\t\t\tif file_ext not in exts:\n\t\t\t\tos.remove(item_path)\n\nrm_files_without_exts(os.path.join(mod_build_dir, 'sound'), ['.fev', '.fsb'])\nrm_files_without_exts(os.path.join(mod_build_dir, 'images'), ['.tex', '.xml'])\nrm_files_without_exts(os.path.join(mod_build_dir, 'bigportraits'), ['.tex', '.xml'])\n","repo_name":"renetta96/wuzzy","sub_path":"build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":1767,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"23033046879","text":"from django.conf.urls import patterns, include, url\nfrom django.contrib.auth.views import login, logout\n\nfrom django.contrib import admin\n\nfrom users_.views import welcome, profile, register, new_test, delete_test, admin_test, ajax_new_question, ajax_delete_question, admin_question, ajax_new_answer, ajax_delete_answer, user_test, ajax_send_result\n\nfrom django.conf import settings\n\nadmin.autodiscover()\n\nurlpatterns = patterns('',\n # Examples:\n # url(r'^$', 'hellodjango.views.home', name='home'),\n # url(r'^blog/', include('blog.urls')),\n\n url(r'^$', welcome),\n url(r'^register/', register),\n url(r'^profile/ajax_new_test', new_test),\n url(r'^profile/ajax_delete_test', delete_test),\n url(r'^profile/admin/(?P\\d{1,50})/$', admin_test),\n url(r'^profile/admin/(?P\\d{1,50})/ajax_new_question', ajax_new_question),\n url(r'^profile/admin/(?P\\d{1,50})/ajax_delete_question', ajax_delete_question),\n url(r'^profile/admin/(?P\\d{1,50})/(?P\\d{1,50})/$', admin_question),\n url(r'^profile/admin/(?P\\d{1,50})/(?P\\d{1,50})/ajax_new_answer', ajax_new_answer),\n url(r'^profile/admin/(?P\\d{1,50})/(?P\\d{1,50})/ajax_delete_answer', ajax_delete_answer),\n\n url(r'^profile/user/(?P\\d{1,50})/$', user_test),\n url(r'^profile/user/(?P\\d{1,50})/ajax_send_result', ajax_send_result),\n # url(r'^admin/', include(admin.site.urls)),\n)\n\nurlpatterns += patterns('',\n url(r'login/$', login, kwargs={'template_name': 'login.html'}, name='login'),\n url(r'logout/$', logout, kwargs={'next_page': '/login'}, name='logout'),\n url(r'^profile/$', profile, name='profile'),\n)\n\nurlpatterns += patterns('',\n (r'^static/(?P.*)$', 'django.views.static.serve', {'document_root': settings.STATIC_ROOT}),\n)","repo_name":"rydztom/questions2","sub_path":"hellodjango/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1839,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"37899563166","text":"from .high_level_planner.dijkstra_planner import Dijkstra_Planner\nfrom .hj_prox.hj_prox_alg import Tracking_Cost, HJ_Prox_Optimizer\nimport torch\nfrom .cem.cem_optimizer import CEM_Optimizer\n\n\nclass TrajOptParams:\n def __init__(self) -> None:\n self.t_param = 0.1\n self.n_batch = 10\n self.horizon = 3\n\n # number of iterations for gradient descent or CEM iteration\n self.num_iter = 10\n\n # number of samples used in CEM\n self.num_samples = 500\n\n\nclass HighLowPlanner:\n def __init__(self, nz, nu, enc, forward_dyn, kmeans, MDP) -> None:\n # latent state dimension\n self.nz = nz\n # control input dimension\n self.nu = nu\n\n self.enc = enc\n self.forward_dyn = forward_dyn\n\n self.kmeans = kmeans\n self.MDP = MDP\n self.high_level_planner = Dijkstra_Planner(MDP)\n\n self.n_batch = 10\n self.horizon = 3\n\n def get_action(self, init_obs, target_obs, u_min=None, u_max=None, opt_params=None, method=\"hj-prox\"):\n if opt_params is None:\n opt_params = TrajOptParams()\n\n # u_min, u_max: tensor of size (nu,), should be given entrywise\n device = init_obs.device\n\n # extract optimization parameters\n t_param = opt_params.t_param\n num_iter = opt_params.num_iter\n n_batch, horizon = opt_params.n_batch, opt_params.horizon\n\n init_lat_state = self.enc(init_obs)\n target_lat_state = self.enc(target_obs)\n\n kmeans = self.kmeans\n high_level_planner = self.high_level_planner\n\n # find the MDP state of the initial and target observation\n cur_mdp_state = kmeans.predict(init_lat_state.detach().cpu())[0]\n target_mdp_state = kmeans.predict(target_lat_state.detach().cpu())[0]\n\n if cur_mdp_state != target_mdp_state:\n next_mdp_state = high_level_planner.step(cur_mdp_state, target_mdp_state)\n next_lat_state = kmeans.cluster_centers_[next_mdp_state]\n else:\n next_mdp_state = target_mdp_state\n next_lat_state = target_lat_state\n\n nu = self.nu\n init_actions = u_min + (u_max - u_min) * torch.rand((n_batch * horizon, nu)).to(device)\n init_actions = init_actions.view((n_batch, horizon, nu))\n\n tracking_cost_fcn = Tracking_Cost(self.forward_dyn, init_lat_state, next_lat_state)\n\n if method == \"hj-prox\":\n hj_optimizer = HJ_Prox_Optimizer(tracking_cost_fcn, init_actions, t_param, x_min=u_min, x_max=u_max)\n output_action, action_list = hj_optimizer.prox_grad_descent(iter_num=num_iter, x_init=init_actions, t=t_param)\n\n if method == \"cem\":\n cem_optimizer = CEM_Optimizer(tracking_cost_fcn, x_min=u_min, x_max=u_max)\n output_action, _ = cem_optimizer.cem_iter(init_actions, num_samples=opt_params.num_samples, num_iter=num_iter)\n\n rollout_costs = tracking_cost_fcn(output_action)\n selected_action = output_action[rollout_costs[0].argmin().item()]\n\n # action = selected_action[0:1, :]\n\n return selected_action\n","repo_name":"microsoft/Intrepid","sub_path":"src/learning/planning/high_low_plan.py","file_name":"high_low_plan.py","file_ext":"py","file_size_in_byte":3086,"program_lang":"python","lang":"en","doc_type":"code","stars":26,"dataset":"github-code","pt":"94"} +{"seq_id":"31759659602","text":"from django.urls import resolve, reverse\n\nfrom django.test import TestCase\nfrom ..models import Categories, Products\n\n\nclass TestViews(TestCase):\n def setUp(self):\n self.category = Categories.objects.create(\n name=\"Food\"\n )\n\n self.category2 = Categories.objects.create(\n name=\"Books\"\n )\n\n self.product = Products.objects.create(\n category_id=1,\n name=\"rice\",\n price=345.56,\n )\n\n self.product2 = Products.objects.create(\n category_id=1,\n name=\"yam\",\n price=345.56,\n )\n\n self.product3 = Products.objects.create(\n category_id=1,\n name=\"noodles\",\n price=345.56,\n )\n\n self.product4 = Products.objects.create(\n category_id=2,\n name=\"Django\",\n price=345.56,\n )\n self.products = Products.objects.all()\n\n\n def test_HomepageView_was_found(self):\n response = self.client.get(reverse(\"categories:homepage\"))\n self.assertEquals(response.status_code, 200)\n\n def test_template_used_by_HomepageView(self):\n response = self.client.get(\"/shop/\")\n self.assertTemplateUsed(response, \"home.html\")\n\n def test_general_product_list_page_was_found(self):\n response = self.client.get(\"/shop/categories/general_product_list/\")\n self.assertEquals(response.status_code, 200)\n\n def test_categorised_product_list_page_was_found(self):\n response = self.client.get(\"/shop/categories/categorised_product_list/rice/\")\n self.assertEquals(response.status_code, 200)\n\n def test_categorised_product_list_view_returns_products_under_the_specified_category(self):\n response = self.client.get(reverse(\"categories:categorised_product_list\", args=[\"Food\"]))\n self.assertGreaterEqual(len(response.context[\"products\"]), 2)\n\n def test_general_product_list_view_returns_all_products_in_db_if_category_is_not_specified(self):\n response = self.client.get(\"/shop/categories/general_product_list/\")\n self.assertGreater(len(response.context[\"products\"]), 3)\n\n def test_template_used_by_general_product_list(self):\n response = self.client.get(\"/shop/categories/general_product_list/\")\n self.assertTemplateUsed(response, \"product_list.html\")\n\n def test_template_used_by_categorised_product_list(self):\n response = self.client.get(reverse(\"categories:categorised_product_list\", args=[\"Books\"]))\n self.assertTemplateUsed(response, \"product_list.html\")\n\n def test_product_details_view_page_was_found(self):\n product_slug = self.products[1].slug\n response = self.client.get(\"/shop/categories/product_details/f'{product_slug}'/\")\n self.assertEquals(response, \"/shop/\")\n\n\n","repo_name":"Chukwunazaekpere/Django_world","sub_path":"test_views.py","file_name":"test_views.py","file_ext":"py","file_size_in_byte":2819,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"73895330548","text":"#Excersice instructions\n\n#If the bill was $150.00, split between 5 people, with 12% tip. \n#Each person should pay (150.00 / 5) * 1.12 = 33.6\n#Round the result to 2 decimal places.\n\nTipOptions = [10,15,20]\nTipPOptions = [1,2,3]\n\nprint(\"Welcome to the tip calculator \\n\")\nbill = float(input(\"What was the total bill? $\"))\n\n#Validate the user input\nloop = False\nwhile loop == False:\n tip = input(\"How much tip would you like to give? \\n 1) \" + str(TipOptions[0]) + \"% \\n 2) \" + str(TipOptions[1]) + \"% \\n 3) \" + str(TipOptions[2]) + \"% \\n\") \n try:\n int(tip)\n integer = True\n\n i = 0\n valid = False\n\n while i < len(TipPOptions):\n if int(tip) == TipPOptions[i]:\n valid = True\n i += 1\n if valid:\n tip = TipPOptions.index(int(tip))\n tip = TipOptions[int(tip)]\n loop = True\n else:\n print(tip + \" is not an option, please pick one from the options\")\n except ValueError:\n integer = False\n\npeople = int(input(\"How many people to split the bill?\"))\n\ntip_as_percent = tip / 100\ntotal_tip_amount = bill * tip_as_percent\ntotal_bill = bill + total_tip_amount\nbill_per_person = total_bill / people\nfinal_amount = round(bill_per_person, 2)\n\n\n# FAQ: How to round to 2 decimal places?\n\n# Find the answer in the Q&A here: https://www.udemy.com/course/100-days-of-code/learn/lecture/17965132#questions/13315048\n\n\nprint(f\"Each person should pay: ${final_amount}\")","repo_name":"AlejandroSolares/100-days-challenge-Python","sub_path":"Day2.py","file_name":"Day2.py","file_ext":"py","file_size_in_byte":1488,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"38108917849","text":"\nimport os\nimport cv2\nimport sys\nimport dlib\nimport argparse\n# import numpy as np\nimport imutils\nimport csv\nimport datetime\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfields = ['concentration', 'time']\nfilename = \"head_movement.csv\"\nwith open(filename, 'w') as csvfile:\n csvwriter = csv.writer(csvfile)\n csvwriter.writerow(fields)\n\nimport time\n#import Face Recognition\nimport face_recognition\n\n# helper modules\nfrom drawFace import draw\nimport reference_world as world\ncounter=0\ncount=0\n# PREDICTOR_PATH = os.path.join(\"models\", \"shape_predictor_68_face_landmarks.dat\")\nPREDICTOR_PATH='shape_predictor_68_face_landmarks.dat'\ndetector = dlib.get_frontal_face_detector()\nif not os.path.isfile(PREDICTOR_PATH):\n print(\"[ERROR] USE models/downloader.sh to download the predictor\")\n sys.exit()\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-f\", \"--focal\",\n type=float, default=1,\n help=\"Callibrated Focal Length of the camera\")\nparser.add_argument(\"-s\", \"--camsource\", type=int, default=0,\n\thelp=\"Enter the camera source\")\n\nargs = vars(parser.parse_args())\n\nface3Dmodel = world.ref3DModel()\n\ndef main():\n predictor = dlib.shape_predictor(PREDICTOR_PATH)\n\n cap = cv2.VideoCapture(args[\"camsource\"])\n # condition1=False\n # condition2 = False\n face_found=True\n not_there = False\n bool1=True\n while True:\n GAZE=\"Face Not Found\"\n ret, img = cap.read()\n if not ret:\n print(f\"[ERROR - System]Cannot read from source: {args['camsource']}\")\n break\n\n #faces = detector(cv2.cvtColor(img, cv2.COLOR_BGR2RGB), 0)\n # popular feature extraction technique(It is a simplified representation of the image that\n # contains only the most important information about the image.)\n # for images – Histogram of Oriented Gradients, or HOG as its commonly known\n # faces = face_recognition.face_locations(img, model=\"hog\")\n # print(faces)\n img = imutils.resize(img, width=450)\n gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n\n # detect faces in the grayscale frame via dlib’s built-in face detector.\n faces = detector(gray, 0)\n if not faces:\n face_found=False\n else:\n face_found=True\n if face_found == False:\n # counter1+=1\n cv2.putText(img, \"INATTENTIVE\", (20, 100), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n\n with open(filename, 'a') as csvfile:\n rows = [['INATTENTIVE', datetime.datetime.now()]]\n csvwriter = csv.writer(csvfile)\n csvwriter.writerows(rows)\n\n # condition2=True\n # else:\n # condition2=False\n\n # Returns an array of bounding boxes of human faces in a image\n # A list of tuples of found face locations in css(top, right, bottom, left) order\n for face in faces:\n #Extracting the co cordinates to convert them into dlib rectangle object\n x=face.left()\n y=face.top()\n w=face.right()-x\n h=face.bottom()-y\n u=face.right()\n v=face.bottom()\n # x = int(face[3])\n # y = int(face[0])\n # w = int(abs(face[1]-x))\n # h = int(abs(face[2]-y))\n # u=int(face[1])\n # v=int(face[2])\n\n newrect = dlib.rectangle(x,y,u,v)\n cv2.rectangle(img, (x, y), (x+w, y+h),\n (0, 255, 0), 2)\n shape = predictor(gray , face)\n\n draw(img, shape)\n\n refImgPts = world.ref2dImagePoints(shape)\n\n height, width, channels = img.shape\n focalLength = args[\"focal\"] * width\n cameraMatrix = world.cameraMatrix(focalLength, (height / 2, width / 2))\n\n mdists = np.zeros((4, 1), dtype=np.float64)\n\n # calculate rotation and translation vector using solvePnP\n success, rotationVector, translationVector = cv2.solvePnP(\n face3Dmodel, refImgPts, cameraMatrix, mdists)\n\n noseEndPoints3D = np.array([[0, 0, 1000.0]], dtype=np.float64)\n noseEndPoint2D, jacobian = cv2.projectPoints(\n noseEndPoints3D, rotationVector, translationVector, cameraMatrix, mdists)\n\n # draw nose line\n p1 = (int(refImgPts[0, 0]), int(refImgPts[0, 1]))\n p2 = (int(noseEndPoint2D[0, 0, 0]), int(noseEndPoint2D[0, 0, 1]))\n cv2.line(img, p1, p2, (110, 220, 0),\n thickness=2, lineType=cv2.LINE_AA)\n\n # calculating euler angles\n rmat, jac = cv2.Rodrigues(rotationVector)\n angles, mtxR, mtxQ, Qx, Qy, Qz = cv2.RQDecomp3x3(rmat)\n # print('*' * 80)\n # print(f\"Qx:{Qx}\\tQy:{Qy}\\tQz:{Qz}\\t\")\n # x = np.arctan2(Qx[2][1], Qx[2][2])\n # y = np.arctan2(-Qy[2][0], np.sqrt((Qy[2][1] * Qy[2][1] ) + (Qy[2][2] * Qy[2][2])))\n # z = np.arctan2(Qz[0][0], Qz[1][0])\n # print(\"ThetaX: \", x)\n # print(\"ThetaY: \", y)\n # print(\"ThetaZ: \", z)\n # print('*' * 80)\n if angles[1] < -15:\n GAZE = \"Looking: Left\"\n elif angles[1] > 30:\n GAZE = \"Looking: Right\"\n else:\n GAZE = \"Forward\"\n #counting the consecutive frames for which person is not looking forward\n if GAZE==\"Looking: Right\" or GAZE==\"Looking: Left\":\n cv2.putText(img, \"INATTENTIVE\", (20, 100), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n\n with open(filename, 'a') as csvfile:\n rows=[['INATTENTIVE' , datetime.datetime.now()]]\n csvwriter = csv.writer(csvfile)\n csvwriter.writerows(rows)\n\n # condition1=True\n # counter+=1\n else:\n # cv2.putText(img, \"ATTENTIVE\", (20, 100), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n with open(filename, 'a') as csvfile:\n rows=[['ATTENTIVE' ,datetime.datetime.now()]]\n csvwriter = csv.writer(csvfile)\n csvwriter.writerows(rows)\n # condition1=False\n cv2.putText(img, GAZE, (20, 20), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 80), 2)\n # if counter >= 5 and condition1==False:\n # count+=1\n # #face not found for some tym\n # if counter1>=5 and condition2==False:\n # count+=1\n # if counter>=20 and condition1==False:\n # count1+=1\n # #face not found bhut tym vala\n # if counter1 >= 20 and condition2 == False:\n # count1 += 1\n # #bhut tym baad screen k samne aya h\n # if counter1>=50 and condition2==False:\n # not_there=True\n #not watching forward from so much tym\n # b = time.time()\n # diff = b - a\n # if (diff >= 60):\n # count=0\n # count1=0\n # bool1=True\n # cv2.putText(img, \"Count:{}\".format(count), (300, 20), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 80), 2)\n # cv2.putText(img, \"Count1:{}\".format(count1), (300, 60), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 80), 2)\n # if count>=5 or count1>=2 or not_there==True:\n # cv2.putText(img, \"INATTENTIVE\", (300, 100), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n cv2.imshow(\"Head Pose\", img)\n\n key = cv2.waitKey(1) & 0xFF\n if key == ord(\"q\"):\n break\n df=pd.read_csv(\"head_movement.csv\")\n INATTENTIVE_HEAD_COUNT=len(df[df['concentration'] == 'INATTENTIVE'])\n ATTENTIVE_HEAD_COUNT=len(df[df['concentration'] == 'ATTENTIVE'])\n TOTAL_HEAD_COUNT=len(df['concentration'])\n label=['ATTENTIVE','INATTENTIVE']\n ATTENTIVE_PERCENTAGE=(ATTENTIVE_HEAD_COUNT/TOTAL_HEAD_COUNT)*100\n INATTENTIVE_PERCENTAGE=(INATTENTIVE_HEAD_COUNT/TOTAL_HEAD_COUNT)*100\n data=[ATTENTIVE_PERCENTAGE,INATTENTIVE_PERCENTAGE]\n fig = plt.figure(figsize=(10, 7))\n plt.pie(data, labels=label,explode=(0.07,0),colors=('green','red'),shadow=True,autopct='%1.1f%%')\n plt.title(\"Based on Head Movement\", bbox={'facecolor': '0.8', 'pad': 5})\n plt.savefig('head_movement.png')\n cap.release()\n cv2.destroyAllWindows()\n # plt.show()\n\n\nif __name__ == \"__main__\":\n # path to your video file or camera serial\n main()","repo_name":"Anuj-singla/Attentiveness_detection","sub_path":"head_position.py","file_name":"head_position.py","file_ext":"py","file_size_in_byte":8422,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"25646427561","text":"from interface import info_retrieval\r\n\r\n\r\ndef main():\r\n s = info_retrieval()\r\n # s.create_index()\r\n s.load_inverted_index()\r\n query = input(\"请输入想要查询的内容:\")\r\n expand, all_check_result = s.spell_check(query)\r\n if expand:\r\n print(\"你可能想要查询的是:(根据序号选择你想要查询的内容,输入 0 则继续查询 \" + query + \" )\")\r\n for i, result in enumerate(all_check_result):\r\n print(i+1, ''.join(result))\r\n num = eval(input(\"查询内容:(0~n):\"))\r\n if num != 0 and num <= len(all_check_result):\r\n query = ''.join(all_check_result[num - 1])\r\n score = s.info_search(query)\r\n print(\"总共检索到\" + str(len(score)) + \"个结果\")\r\n for i in range(len(score)):\r\n print(score[i][\"file\"])\r\n\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","repo_name":"Liu-YT/Info-Search","sub_path":"src/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":867,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"12489508439","text":"from django.urls import include, path\nfrom rest_framework import routers\n\nfrom . import views\n\nrouter = routers.DefaultRouter()\nrouter.register(r'posts', views.PostViewSet, basename=\"posts\")\nrouter.register(r'groups', views.GroupViewSet, basename=\"groups\")\nrouter.register(r'follow', views.FollowViewSet, basename=\"follow\")\nrouter.register(\n r'posts/(?P\\d+)/comments',\n views.CommentViewSet, basename=\"comments\"\n)\n\nurlpatterns = [\n path('v1/', include(router.urls)),\n]\n","repo_name":"vadikam100500/Twitter_killer_API","sub_path":"api/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":487,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"28851871828","text":"# encoding: utf-8\n\"\"\"\n@author: liuwz\n@time: 2021/5/17 3:41 下午\n@file: 换工位.py\n@desc: \n\"\"\"\n\nfrom typing import List\n\"\"\"\n因为要实施结对编程,想让两个员工的工位挨在一起:要求 [0,1] 员工坐在一起,[2, 3] 员工坐在一起,以此类推。不过挨着具体坐的位置并不重要,只要能挨在一起就可以了。\n比如 [0, 1, 3, 2] 与 [2, 3, 1, 0] 都是满足要求的。现在给定一个数组 A[],求换工位的最少次数,尽量让两个员工坐在一起。(给定 N 个员工,他们的编号总是 [0~N-1] ,并且 N 总是偶数)。\n\n输入:A[] = [0, 3, 2, 1]\n输出:1\n\n解释:只需要换 1 次就可以了,比如,将 0 号员工与 2 号员工交换。\n\"\"\"\n\n\nclass UF:\n def __init__(self, n):\n self.par = dict()\n self.ans = 0\n for i in range(n):\n # 初始化的时候就结成组 01 23 45 67...\n self.par[i] = i - (i&1)\n\n\n def find(self, x):\n if x == self.par[x]:\n return x\n\n self.par[x] = self.find(self.par[x])\n return self.par[x]\n\n def union(self, x, y):\n x_par = self.find(x)\n y_par = self.find(y)\n if x_par != y_par:\n self.par[x_par] = y_par\n # 合并次数统计 即为想要接环的次数 反着推 你想要知道让两个数字挨着需要多少次交换 那就初始化的时候挨着需要多少次能够变成现在的情况\n self.ans += 1\n\n\nclass Solution:\n def exchange_station(self, A):\n n = len(A)\n uf = UF(n)\n\n for i in range(0, len(A), 2):\n uf.union(A[i], A[i+1])\n print(uf.par)\n return uf.ans\n\n\na = Solution().exchange_station([6, 4, 5, 2, 3, 7, 0, 1] )\nprint(a)","repo_name":"starrye/LeetCode","sub_path":"union-find sets/换工位.py","file_name":"换工位.py","file_ext":"py","file_size_in_byte":1749,"program_lang":"python","lang":"zh","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"42080509120","text":"import pyttsx3\r\nimport os \r\nimport time\r\n\r\nwhile(1==1):\r\n\r\n pyttsx3.speak(\"Welcome to chat bot\")\r\n print(\"Welcome to chat bot\")\r\n\r\n pyttsx3.speak(\"how can i help you\")\r\n print(\"How can I help you :\" ,end=' ')\r\n p = input()\r\n\r\n if ((\"run\" in p) or (\"launch\" in p) or (\"open\" in p) or (\"execute\" in p)) and ((\"chrome\" in p) or (\"google\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else:\r\n pyttsx3.speak(\"Opening Chrome\")\r\n os.system(\"start chrome\")\r\n time.sleep(5)\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"open\" in p) or (\"execute\" in p)) and ((\"notepad\" in p) or (\"notebook\" in p) or (\"editor\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"Opening notepad\")\r\n os.system(\"notepad\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"media\" in p) or (\"player\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening windows media player\")\r\n os.system(\"start wmplayer\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"explorer\" in p) or (\"browser\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening internet explorer\")\r\n os.system(\"start iexplore\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"outlook\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening microsoft outlook\")\r\n os.system(\"start outlook\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"powerpoint\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening microsoft powerpoint\")\r\n os.system(\"start powerpnt\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"computer\" in p) or (\"pc\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening my computer\")\r\n os.system(\"start explorer\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"calculator\" in p) or (\"calculations\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening calculator\")\r\n os.system(\"calc\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"magnifier\" in p) or (\"zoom\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening magnifier\")\r\n os.system(\"start magnify\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"paint\" in p) or (\"draw\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening paint\")\r\n os.system(\"start mspaint\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"registry\" in p) and (\"editor\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening registry editor\")\r\n os.system(\"start regedit\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"wordpad\" in p) or (\"word\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening wordpad\")\r\n os.system(\"start write\")\r\n\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"system\" in p) and (\"configuration\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening system configuration\")\r\n os.system(\"msconfig\")\r\n\t\r\n elif ((\"run\" in p) or (\"launch\" in p) or (\"execute\" in p) or (\"open\" in p)) and ((\"command\" in p) and (\"promt\" in p) or (\"cmd\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"opening command promt\")\r\n os.system(\"start cmd.exe\")\r\n\r\n elif (\"exit\" in p) or (\"quit\" in p):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"Thanks for coming Had good time with you\")\r\n print(\"Thanks for coming . Had good time with you\")\r\n break\r\n\r\n elif ((\"take\" in p) or (\"capture\" in p)) and ((\"screenshot\" in p) or (\"ss\" in p) or (\"screen\" in p)):\r\n if (\"dont\" in p):\r\n \t print(\"OK as you wish\")\r\n else: \r\n pyttsx3.speak(\"Taking Screenshot\") \r\n import pyautogui\r\n from tkinter import filedialog\r\n myScreenshot = pyautogui.screenshot()\r\n file_path = filedialog.asksaveasfilename(defaultextension='.jpg')\r\n myScreenshot.save(file_path)\r\n time.sleep(5)\r\n else : \r\n print(\"dont support\")\r\n \r\n\r\n\r\n","repo_name":"Akshat0508/Python-Project","sub_path":"project1.py","file_name":"project1.py","file_ext":"py","file_size_in_byte":6469,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"2674532335","text":"import matplotlib.pyplot as plt\nimport numpy as np\n\n# 문제에서 주어진 데이터\nstep = 500\nmu = 0.002\nsigma = 0.01\n\n# 데이터 만들기\ndata = np.random.normal(mu, sigma, 500).cumsum()\nx = np.arange(0, 500, 1)\nlowerBound = mu * x - sigma * np.sqrt(x)\nupperBound = mu*x + sigma * np.sqrt(x)\n\n# 그래프 그리기\nplt.plot(x, data, color='red', label='Walker position')\nplt.plot(x, lowerBound, linestyle='none')\nplt.plot(x, upperBound, linestyle='none')\nplt.plot(x, (lowerBound+upperBound)/2, color='black',\n linestyle='dashed', label='Population mean') # 평균 선 그리기, 줄 스타일을 - 로\n\nplt.fill_between(x, lowerBound, upperBound,\n facecolor='#FDFD96', label='1 sigma range') # lowerBound와 upperBound 사이 채우기\nplt.fill_between(x, upperBound, data, where=data > upperBound, facecolor='darkblue',\n interpolate=True) # 범위 벗어나서 큰 경우 darkblue로 채우기\nplt.fill_between(x, lowerBound, data, where=data < lowerBound, facecolor='green',\n interpolate=True) # 범위 벗어나서 작은 경우 green으로 채우기\n\n\n# x축 표현할 값 지정\nplt.xticks(np.arange(0, 600, step=100))\nplt.xlim(0, 500)\n\n# x축, y축 라벨 지정\nplt.xlabel('The number of steps', fontsize=10)\nplt.ylabel('Position', fontsize=10)\n\n# 그리드 보이기, 범례 표시하고 그래프 보이기\nplt.grid(True)\nplt.legend()\nplt.show()\n","repo_name":"metamoong/data_analysis","sub_path":"midterm/midterm_03.py","file_name":"midterm_03.py","file_ext":"py","file_size_in_byte":1422,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"27595927043","text":"# to set up:\n# share calendar from outlook.office.com\n# get email from there, extract ics link.\n# this script will then parse that ics link and tell you either not busy, almost busy, or busy.\n#\n# https://stackoverflow.com/questions/3408097/parsing-files-ics-icalendar-using-python\n\n\n#from icalevents.icalevents import events\n\nimport json\n#import requests\nimport urllib.request\n\n#import wget\n#import vobject\n#from ics import Calendar, Event\n#import ics\n#from tzlocal import get_localzone\nfrom icalendar import Calendar\n\n# https://icalevents.readthedocs.io/en/latest/\n#from icalevents.icalevents import events_async, latest_events, all_done\nfrom time import sleep\nimport icalendar\nimport recurring_ical_events\n\n\nimport pytz\nimport datetime\nimport os\n\ndef main(url_file = \"calendar_url.txt\"):\n\n # put the url to the ics in the \"secret\" (not in github) url file\n with open(url_file, 'r') as f:\n url = f.read()\n\n filename = \"reachcalendar.ics\"\n\n if os.path.exists(filename):\n os.remove(filename)\n \n #wget.download(url) # this hung on me.. punting\n \n # https://stackoverflow.com/questions/35115513/python-bad-request-in-get-but-wget-and-curl-work\n #headers = {'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36'}\n # https://stackoverflow.com/questions/12624133/wget-with-python-time-limit\n \n # https://stackoverflow.com/questions/20227324/programmatically-download-content-from-shared-dropbox-folder-links\n# headers = { 'User-Agent': 'Wget/1.20.3 (linux-gnu)',\n# 'Accept': '*/*',\n# 'Accept-Encoding': 'identity',\n# 'Host': 'outlook.office365.com',\n# 'Connection': 'Keep-Alive'}\n#\n# print(url)\n# response = requests.get(url, timeout=10.0, data=json.dumps(headers=headers)\n\n\n\n response = urllib.request.urlopen(url, timeout=10.0)\n data = response.read() # a `bytes` object\n #text = data.decode('utf-8') # a `str`; this step can't be used if data is binary\n \n with open(filename, 'wb') as f:\n f.write(data)\n\n\n \n \n #utc=pytz.UTC\n # https://stackoverflow.com/questions/31304890/wrong-aware-datetime-with-pytz-and-america-chicago\n #chicago = pytz.timezone('America/Chicago')\n chicago_time = pytz.timezone(\"America/Chicago\")\n\n\n def pre_meeting():\n #\n start = datetime.datetime.now() \n start = chicago_time.localize(start)\n \n end = datetime.datetime.now()\n end = chicago_time.localize(end)\n end = end + datetime.timedelta(minutes=2)\n \n with open(filename, 'r') as f:\n ical_string = f.read()\n \n start_date = start\n end_date = end\n \n print(start_date)\n print(end_date)\n \n calendar = icalendar.Calendar.from_ical(ical_string)\n events = recurring_ical_events.of(calendar).between(start_date, end_date)\n found_meeting = False\n for event in events:\n start = event[\"DTSTART\"].dt\n duration = event[\"DTEND\"].dt - event[\"DTSTART\"].dt\n transp = event[\"TRANSP\"]\n print(f\"start {start} duration {duration} transp {transp}\")\n if transp == \"OPAQUE\":\n found_meeting = True\n return found_meeting\n\n def busy():\n #\n start = datetime.datetime.now() \n start = chicago_time.localize(start)\n \n end = datetime.datetime.now()\n end = chicago_time.localize(end)\n \n with open(filename, 'r') as file:\n ical_string = file.read()\n \n start_date = start\n end_date = end\n \n print(start_date)\n print(end_date)\n \n calendar = icalendar.Calendar.from_ical(ical_string)\n events = recurring_ical_events.of(calendar).between(start_date, end_date)\n found_meeting = False\n for event in events: \n start = event[\"DTSTART\"].dt\n duration = event[\"DTEND\"].dt - event[\"DTSTART\"].dt\n transp = event[\"TRANSP\"]\n print(f\"start {start} duration {duration} transp {transp}\")\n if transp == \"OPAQUE\":\n found_meeting = True\n return found_meeting\n\n if busy():\n data = {\"calendar_status\":\"busy\"}\n else:\n if pre_meeting():\n data = {\"calendar_status\":\"pre meeting\"}\n else:\n data = {\"calendar_status\":\"free\"}\n\n print(data)\n # https://stackoverflow.com/questions/12309269/how-do-i-write-json-data-to-a-file\n with open('calendar_status.json', 'w') as f:\n json.dump(data, f)\n\n return data\n \n\ndef _format_name(address):\n \"\"\"Retrieve the e-mail and the name from an address.\n :arg an address object, e.g. mailto:test@test.test\n :returns str: The name and the e-mail address.\n \"\"\"\n email = address.split(':')[-1]\n name = email.split('@')[0]\n if not email:\n return ''\n return f\"{name} <{email}>\"\n\n\ndef _format_attendees(attendees):\n \"\"\"Format the list of attendees.\n :arg any attendees: Either a list, a string or a vCalAddress object.\n :returns str: Formatted list of attendees.\n \"\"\"\n if isinstance(attendees, list):\n return '\\n'.join(map(lambda s: s.rjust(len(s) + 5), map(_format_name, attendees)))\n return _format_name(attendees)\n\n \ndef view(event):\n \"\"\"Make a human readable summary of an iCalendar file.\n :returns str: Human readable summary.\n \"\"\"\n summary = event.get('summary', default='')\n organizer = _format_name(event.get('organizer', default=''))\n attendees = _format_attendees(event.get('attendee', default=[]))\n location = event.get('location', default='')\n comment = event.get('comment', '')\n description = event.get('description', '').split('\\n')\n description = '\\n'.join(map(lambda s: s.rjust(len(s) + 5), description))\n\n start = event.decoded('dtstart')\n if 'duration' in event:\n end = event.decoded('dtend', default=start + event.decoded('duration'))\n else:\n end = event.decoded('dtend', default=start)\n duration = event.decoded('duration', default=end - start)\n if isinstance(start, datetime.datetime):\n start = start.astimezone(start.tzinfo)\n start = start.strftime('%c')\n if isinstance(end, datetime.datetime):\n end = end.astimezone(end.tzinfo)\n end = end.strftime('%c')\n\n \n\n return f\"\"\" Status: {flag}\n Organizer: {organizer}\n Attendees:\n{attendees}\n Summary : {summary}\n Starts : {start}\n End : {end}\n Duration : {duration}\n Location : {location}\n Comment : {comment}\n Description:\n{description}\"\"\"\n\n\n\n\n\n \n \nif __name__ == '__main__':\n import typer\n typer.run(main)\n\n\n","repo_name":"poleguy/fire_led","sub_path":"check_calendar/check_calendar.py","file_name":"check_calendar.py","file_ext":"py","file_size_in_byte":6785,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"20264235208","text":"from PyQt5.QtWidgets import QApplication, QWidget, QVBoxLayout, QLabel, QPushButton\nfrom PyQt5.QtCore import pyqtSignal\nfrom FirstPage import FirstPage\n\n\nclass SecondPage(QWidget):\n sendData = pyqtSignal(str)\n\n def __init__(self, first_page):\n super().__init__()\n\n self.first_page = first_page\n\n self.label = QLabel(\"Data from First Page: None\")\n layout = QVBoxLayout()\n layout.addWidget(self.label)\n\n self.button = QPushButton(\"Send Data to First Page\")\n self.button.clicked.connect(self.sendDataToFirstPage)\n layout.addWidget(self.button)\n\n self.setLayout(layout)\n\n def sendDataToFirstPage(self):\n data = \"Hello from Second Page\"\n self.sendData.emit(data)\n\n def receiveData(self, data):\n self.label.setText(\"Data from First Page: \" + str(data))\n\n\nif __name__ == \"__main__\":\n app = QApplication([])\n first_page = FirstPage()\n second_page = SecondPage(first_page)\n\n first_page.sendData.connect(second_page.receiveData)\n second_page.sendData.connect(first_page.receiveData)\n\n first_page.show()\n second_page.show()\n app.exec_()\n","repo_name":"junxian428/Python_Qt_Page_Transfer_data_example","sub_path":"SecondPage.py","file_name":"SecondPage.py","file_ext":"py","file_size_in_byte":1146,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"71380820470","text":"# Memory Complexity: O(size_marker)\n# Runtime Complexity: O(n)\n\nif __name__ == \"__main__\":\n\n def check_marker(x,size_marker=4):\n for i in range(0,len(x)-size_marker):\n x_subset = x[i:i+size_marker]\n if len(set(x_subset)) == size_marker:\n return i + size_marker\n return None\n\n # Problem 1 \n file = open('input.txt', 'r')\n Lines = file.read().splitlines() \n print('---------- Problem 1 -------------')\n for line in Lines:\n print('Marker is found at: ' + str(check_marker(line))) \n print('---------- Problem 2 -------------')\n for line in Lines:\n print('Marker is found at: ' + str(check_marker(line,14))) \n\n\n\n \n\n\n\n\n\n \n\n \n","repo_name":"Joachim-Sh/advent-of-code-2022","sub_path":"day 6/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":725,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"19499551650","text":"#!/usr/bin/python3\n# -*- coding : utf-8 -*-\n# @Author : Tachiu Lam\n# @Mail : techaolin@gamil.com\n# @Datetime : 2020/12/10 17:31\n# @Software : PyCharm\n\nimport time\nimport json\nimport requests\nfrom AssetManage.models import Asset\nfrom API.Functions.vulnerability import VulnerabilityManage\n\n\nclass Img_Scan:\n \"\"\"镜像漏洞获取、保存类\"\"\"\n\n @classmethod\n def transform_severity_to_level(cls, severity):\n \"\"\"将扫描报告中的severity转换为相应level\"\"\"\n level_table = {\n \"Unknown\": 0, \"Negligible\": 0, \"Low\": 1, \"Medium\": 2, \"High\": 3, \"Critical\": 4\n }\n level = level_table.get(severity) if level_table.get(severity) else 0\n return level\n\n @classmethod\n def generate_api(cls, content):\n \"\"\"构造获取harbor对应镜像漏洞的api\"\"\"\n resources = content.get('event_data').get('resources')[0]\n repository = content.get('event_data').get('repository')\n\n img_name = repository.get('name')\n namespace = repository.get('namespace')\n sha256 = resources.get('digest')\n resource_url = resources.get('resource_url').split('/')[0]\n scan_status = resources.get('scan_overview').get(\n 'application/vnd.scanner.adapter.vuln.report.harbor+json; version=1.0').get('scan_status')\n scanner = resources.get('scan_overview').get(\n 'application/vnd.scanner.adapter.vuln.report.harbor+json; version=1.0').get(\n 'scanner').get('name')\n\n api_url = 'https://' + resource_url + '/api/v2.0/projects/' + namespace + '/repositories/' + img_name + '/artifacts/' + sha256 + '/additions/vulnerabilities'\n return {'api_url': api_url, 'sha256': sha256, 'img_name': img_name}\n\n @classmethod\n def create_or_update_vulnerability(cls, num_id, vulnerability):\n \"\"\"关联漏洞信息,漏洞信息只新增和更新,不自动删除\"\"\"\n for each_v in vulnerability:\n v = {}\n v['cve'] = each_v.get('id')\n v['name'] = each_v.get('package') + ' ' + each_v.get('id') # 库名+ cve名称作为漏洞名字\n v['fix'] = each_v.get('fix_version')\n v['introduce'] = each_v.get('description')\n v['port'] = each_v.get('version') # 影响版本,为适应函数,命名为port\n v['introduce'] = v['vuln_info'] = each_v.get('links')[0]\n v['level'] = cls.transform_severity_to_level(each_v.get('severity'))\n v['source'] = '2' # 漏洞导入方式:镜像扫描\n\n v_num_id = VulnerabilityManage.get_vuln_id() + 1 # 获取漏洞表id\n\n exits = VulnerabilityManage.status(num_id=num_id, name=v['name'], v_num_id=v_num_id)\n\n v['asset'] = exits['asset'] # 先进行漏洞和资产绑定,避免删除其他资产漏洞\n if exits.get('exits') is True:\n v['fix_status'] = exits['fix_status'] # 继承漏洞状态\n VulnerabilityManage.update_or_create(v, exits=True).get('result')\n else:\n v['fix_status'] = exits['fix_status']\n v['v_id'] = exits['v_id']\n v['v_type'] = exits['v_type']\n VulnerabilityManage.update_or_create(v, exits=False).get('result')\n v_num_id += 1 # 新建查询漏洞,漏洞id都需要递增\n\n return {'result': '漏洞导入成功'}\n\n @classmethod\n def scan_deal(cls, api_url, sha256, img_name):\n \"\"\"获取harbor漏洞接口的数据,新建or更新资产,更新漏洞信息\"\"\"\n content = requests.get(api_url, verify=False).content # 自签证书需要取消证书校验\n with open('./log.txt', 'a') as f:\n f.write(str(content))\n content = json.loads(content)\n # 判断数据是否存在\n if content.get('application/vnd.scanner.adapter.vuln.report.harbor+json; version=1.0'):\n data = content.get('application/vnd.scanner.adapter.vuln.report.harbor+json; version=1.0')\n exits = Asset.objects.filter(asset_key=sha256).first() # 查看唯一值asset_key是否存在\n if not exits:\n num_id = Asset.objects.latest('id').id\n num_id += 1 #\n asset_id = '03' + time.strftime('%Y%m%d', time.localtime(time.time())) + str(num_id) # 镜像ID分类03\n # 资产创建或更新\n asset_create = Asset.objects.get_or_create(\n asset_id=asset_id,\n asset_name=img_name,\n asset_type_id=37, # 镜像资产分类\n asset_key=sha256, # 镜像哈希为唯一ID\n asset_score=data.get('severity'),\n asset_area_id=13, # 默认归类到安全组项目\n # asset_description=asset_description,\n )\n # asset_create : (, True)\n if asset_create[1]:\n asset_create[0].save()\n else: # 镜像已存在的情况,需要查找到资产对应的id,并更新资产类型\n num_id = Asset.objects.get(asset_key=sha256).id\n Asset.objects.filter(asset_key=sha256).update(asset_type_id=37)\n\n # 漏洞更新\n vuln_result = cls.create_or_update_vulnerability(num_id, data.get(\"vulnerabilities\"))\n return {'img': sha256, 'vulnerability': vuln_result}\n return {'msg': False}\n\n @classmethod\n def main(cls, content):\n api_info = cls.generate_api(content)\n api_url = api_info.get(\"api_url\")\n sha256 = api_info.get(\"sha256\")\n img_name = api_info.get(\"img_name\")\n\n result = cls.scan_deal(api_url, sha256, img_name)\n return result","repo_name":"TachiuLam/SeMF","sub_path":"API/Functions/img_scan.py","file_name":"img_scan.py","file_ext":"py","file_size_in_byte":5757,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"94"} +{"seq_id":"21867859873","text":"#When an argument is used to customize a program, it is called a parameter.\n#In the previous section SourceLanguageCode is an example of a parameter. \n#Which is something the function needs to perform the action required.\n\n#A parameter in Python is like a space in a recipe that you can fill with different things\n#when you use the recipe. It makes your code flexible, so you can do different things with it \n#by changing what you put into those spaces.\n\n#The value we assigned en is an example of an argument,\n#which is a piece of information to the function so it can perform the designed action.\n\n#A parameter is like the blank space in a recipe, waiting for something to go in.\n#An argument is the real thing you put into that space when you use the recipe.\n\nimport boto3\n\ndef translate_text():\n client = boto3.client('translate')\n response = client.translate_text(\n Text='I am learning to code in AWS', \n SourceLanguageCode='en', \n TargetLanguageCode='fr' \n )\n print(response) \n\ndef main():\n translate_text()\n\nif __name__==\"__main__\":\n main()\n","repo_name":"brianarorie/red-july-2023","sub_path":"w1-arguments.py","file_name":"w1-arguments.py","file_ext":"py","file_size_in_byte":1087,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"13596615521","text":"########################################################\r\n# CSC 242 Section 504 Winter 2017\r\n# Recursion exercises (Ch. 10)\r\n########################################################\r\n\r\ndef factorial(n):\r\n 'nonrecursive version of factorial'\r\n answer = 1\r\n for i in range(1,n+1):\r\n answer *= i\r\n return answer \r\n\r\ndef factorial_r(n):\r\n 'recursive version of factorial'\r\n #start with base case which we know the answer to without any calculation\r\n if n == 1:\r\n return 1\r\n else:\r\n return n * factorial(n-1)\r\n\r\n# a utility function, meant to help visualize recursion\r\ndef print_with_space(var, level):\r\n if level == 0:\r\n print(var)\r\n else:\r\n print(' ', end='')\r\n print_with_space(var, level-1)\r\n\r\n \r\n# so that we can see the levels of recursion\r\ndef factorial_print(n, level=0):\r\n print_with_space('factorial({})'.format(n), level)\r\n answer = 1\r\n if n > 1:\r\n answer = n * factorial_print(n-1, level+1)\r\n print_with_space(answer, level)\r\n return answer\r\n\r\ndef exp(x,y):\r\n 'iterative version of computing x to the y power'\r\n answer = 1\r\n for i in range(1,y+1):\r\n answer *= x\r\n return answer\r\n\r\ndef exp_r(x, y):\r\n if y == 0:\r\n return 1\r\n return x * exp_r(x, y-1)\r\n\r\ndef exp_print(x,y,level=0):\r\n print_with_space('exp({},{})'.format(x,y), level)\r\n if y == 0:\r\n print_with_space(1, level)\r\n return 1\r\n answer = x * exp_print(x,y-1,level+1)\r\n print_with_space(answer, level)\r\n return answer\r\n\r\n# greatest common divisor (iterative)\r\n# if y divides into x evenly, then gcd(x,y) => y\r\n# otherwise, it's gcd(y, x%y)\r\ndef gcd(x,y):\r\n while x % y != 0:\r\n temp = y\r\n y = x % y\r\n x = temp\r\n return y\r\n \r\n\r\n# greatest common divisor (recursive)\r\n# if y divides into x evenly, then gcd(x,y) => y\r\n# otherwise, it's gcd(y, x%y)\r\ndef gcd_r(x,y):\r\n# print('gcd_r({},{})'.format(x,y))\r\n if x % y == 0:\r\n return y\r\n else:\r\n return gcd_r(y, x%y)\r\n\r\n# Fibonacci sequence (recursive): 1, 1, 2, 3, 5, 8, 13, 21, 34...\r\n# fib(6) should return 8\r\ndef fib(n):\r\n if n == 1 or n == 2:\r\n return 1\r\n else:\r\n return fib(n-1) + fib(n-2)\r\n\r\n# Fibonacci sequence (recursive): 1, 1, 2, 3, 5, 8, 13, 21, 34...\r\n# fib(6) should return 8\r\ndef fib_print(n, level=0):\r\n print_with_space('fib({})'.format(n), level)\r\n if n == 1 or n == 2:\r\n return 1\r\n else:\r\n return fib_print(n-1, level+1) + fib_print(n-2, level+1)\r\n \r\n# iterative; without the \"in\" operator\r\n# return True is item is in the list or string x\r\n# or false otherwise. Do not use the Python\r\n# \"in\" operator in this problem\r\ndef occur(item, x):\r\n 'determines if item is found in a list'\r\n i = 0\r\n while i < len(x):\r\n if i == item:\r\n return True\r\n i += 1\r\n return False\r\n \r\n# recursive\r\ndef occur_r(item, x):\r\n if x == [ ]:\r\n return False\r\n elif x[0] == item:\r\n return True\r\n else:\r\n return occur_r(item, x[1:])\r\n\r\n# examples that are relevant to many of the problems on the 5th lab\r\n\r\n# return a ***list*** of the words which start with 'q'\r\n# >>> qwords(['query', 'ask', 'question', 'inquire'])\r\n# ['query', 'question']\r\n\r\ndef qwords(words):\r\n if words == [ ]: # base case\r\n return list() # [ ]\r\n elif words[0][0] == 'q': # recursive case(s): first word starts with 'q'\r\n return [ words[0] ] + qwords(words[1:])\r\n else: # recursive case(s): first word does not start with 'q'\r\n return qwords(words[1:])\r\n\r\n# return True if sentence contains all of the letters in the alphabet,\r\n# or False otherwise\r\ndef all_letters(sentence, missing = None):\r\n if missing == None:\r\n missing = set('abcdefghijklmnopqrstuvwxyz')\r\n if sentence == '': # base case\r\n return len(missing) == 0\r\n elif sentence[0] in missing: # recursive case 1\r\n missing.remove(sentence[0])\r\n return all_letters(sentence[1:], missing)\r\n else: # recursive case 2\r\n return all_letters(sentence[1:], missing)\r\n\r\n################################################################\r\n# continue with exercises from notes\r\n################################################################\r\n# again, try to write this without using the Python \"in\" operator\r\ndef occurrences_r(item, x):\r\n 'count number of times item is in x'\r\n if x == []: # base case\r\n return 0\r\n elif item == x[0]: # recursive case 1\r\n return occurrences_r(item, x[1:]) + 1\r\n else:\r\n return occurrences_r(item, x[1:])\r\n\r\n#************************************************\r\n#* We stopped here on 2/8\r\n#************************************************\r\n\r\n# iterative\r\ndef pal(word):\r\n 'determines whether or not a word is a palindrome'\r\n # for example, 'noon', 'radar', but not 'moon', 'noun', 'runner'\r\n for i in range(len(word)):\r\n if word[i] != word[-(i+1)]:\r\n return False\r\n return True\r\n\r\ndef pal_r(word):\r\n if len(word)<=1:\r\n return True\r\n elif word=='':\r\n return True\r\n elif word[0]!= word[-1]:\r\n return False\r\n else:\r\n return pal_r(word[1:-1])\r\n\r\ndef one_to_n(n):\r\n if n==0:\r\n return [0]\r\n elif n==1:\r\n return [1]\r\n else:\r\n return one_to_n(n-1)+[n]\r\n\r\ndef modulus(x, y):\r\n if x==y:\r\n return 0\r\n elif x-y>> lst1 = ['a', 'b']\r\n# >>> lst2 = copy_r(lst1)\r\n# >>> lst1[0] = 'x'\r\n# >>> lst1\r\n# ['x', 'b']\r\n# >>> lst2\r\n# ['a', 'b']\r\n\r\ndef copy(list):\r\n answer = [ ]\r\n for item in list:\r\n answer.append(item)\r\n return answer\r\n\r\n# copy a list\r\ndef copy_r(lst):\r\n if lst==[]:\r\n return []\r\n else:\r\n return [lst[0]]+copy_r(lst[1:])\r\n\r\n# return [1, 2, 4, 8, ..., x] recursively\r\n# (x is largest power of 2 <= n)\r\ndef powers_of_2(n):\r\n return [ ]\r\n\r\n\r\n# examples with lists of lists\r\n#\r\n# This function returns the number of items in the list x. x may\r\n# contain embedded lists.\r\ndef len_embed(x):\r\n if not isinstance(x, list):\r\n return 1\r\n elif x == [ ]:\r\n return 0\r\n else:\r\n return len_embed(x[0]) + len_embed(x[1:])\r\n\r\ndef len_embed_print(lst):\r\n print('elen({})'.format(lst))\r\n if not isinstance(lst, list):\r\n print('returns 1')\r\n return 1\r\n elif lst == [ ]:\r\n print('returns 0')\r\n return 0\r\n else:\r\n ret = elen(lst[0]) + elen(lst[1:])\r\n print('returns {}'.format(ret))\r\n return ret\r\n\r\n# write a function which is passed 2 parameters: x (a list)\r\n# and item. It returns True if item appears in x, or False\r\n# otherwise..\r\n#\r\n# for example:\r\n#\r\n# >>> contains_embed(3, [[[0, 1], 2], 1, [3, 1]])\r\n# True\r\n# >>> contains_embed(4, [[0, 1], [3], 2])\r\n# False\r\n\r\n# written such that x may or may not be a list\r\ndef contains_embed(item, x):\r\n if not isinstance(x, list):\r\n return item ==x\r\n elif x==[]:\r\n return False\r\n else:\r\n return contains_embed(item, x[0]) or contains_embed(item, x[1:])\r\n \r\n\r\n# count the number of times that item is in x, where\r\n# x may contain embedded lists\r\ndef count_embed(x, item):\r\n pass\r\n\r\n# return the maximum item in x, where x may contain\r\n# embedded lists\r\ndef max_embed(x):\r\n if not isinstance(x, list):\r\n return x\r\n elif x==[]:\r\n return None\r\n else:\r\n max0= max_embed(x[0])\r\n max1= max_embed(x[1:])\r\n if max0==None:\r\n return max1\r\n elif max1==None:\r\n return max0\r\n elif max0 >= max1:\r\n return max0\r\n else:\r\n return max1\r\n\r\n# return the sum of the items in x, where x may contain\r\n# embedded lists\r\ndef sum_embed(x):\r\n if not isinstance(x, list):\r\n return x\r\n elif x==[]:\r\n return 0\r\n else:\r\n return sum_embed(x[0]) + sum_embed(x[1:])\r\n\r\n# try this\r\n##>>> lst0 = [1,2]\r\n##>>> lst1 = [3,4,5]\r\n##>>> lst2 = [lst0, lst1]\r\n##>>> lst2\r\n##[[1, 2], [3, 4, 5]]\r\n##>>> lst3 = copy_r(lst2)\r\n##>>> lst3\r\n##[[1, 2], [3, 4, 5]]\r\n##>>> lst4 = deep_copy_r(lst2)\r\n##>>> lst0[0] = 100\r\n##>>> lst0\r\n##[100, 2]\r\n##>>> lst2\r\n##[[100, 2], [3, 4, 5]]\r\n##>>> lst3\r\n##[[100, 2], [3, 4, 5]]\r\n##>>> lst4\r\n##[[1, 2], [3, 4, 5]]\r\n\r\n# \"deep\" copy a list\r\ndef deep_copy_r(x):\r\n return x\r\n\r\ndef count_embed(item, x):\r\n try:\r\n return count_embed(item, x[0]) + count_embed(item, x[1:])\r\n except TypeError: # x is not a list\r\n return item == x\r\n except IndexError: # might be generated by x[1:] \r\n return 0\r\n\r\ndef sum_embed(x):\r\n try:\r\n return sum_embed(x[0]) + sum_embed(x[1:])\r\n except TypeError:\r\n return x\r\n except IndexError:\r\n return 0\r\n\r\ndef max_try(x):\r\n max0=0\r\n max1=0\r\n try:\r\n max0=max_try(x[0])\r\n except TypeError:\r\n return max0\r\n try:\r\n max1=max_try(x[1:])\r\n except IndexError:\r\n return max0\r\n except TypeError:\r\n return max0\r\n if max0 >= max1:\r\n return max0\r\n else:\r\n return max1\r\n\r\ndef deep_copy_r(x):\r\n try:\r\n return [deep_copy_r(x[0])] + deep_copy_r(x[1:])\r\n except TypeError:\r\n return x\r\n except IndexError:\r\n return list()\r\n\r\ndef flatten(x):\r\n try:\r\n return flatten(x[0]) + flatten(x[1:])\r\n except TypeError:\r\n return [x]\r\n except IndexError:\r\n return []\r\n\r\n\r\n\r\n\r\n","repo_name":"lpmurray16/2015-2019","sub_path":"Old_DePaul_Classes/Python Class/recursion_practice.py","file_name":"recursion_practice.py","file_ext":"py","file_size_in_byte":9551,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"18758663731","text":"import sys\nimport numpy as np\nimport cv2\nimport matplotlib.pyplot as plt\nimport glob\nfrom skimage.feature import hog\nimport ipdb\n\n##############################\n# Define a function to return HOG features and visualization\ndef get_hog_features(img, orient, pix_per_cell, cell_per_block, vis=False, feature_vec=True):\n\tif vis == True:\n\t\tfeatures, hog_image = hog(img, orientations=orient, pixels_per_cell=(pix_per_cell, pix_per_cell), \n\t\t\t\tcells_per_block=(cell_per_block,cell_per_block), visualise=True, feature_vector=feature_vec, transform_sqrt=True)\n\t\treturn features, hog_image\n\telse:\n\t\tfeatures = hog(img, orientations=orient, pixels_per_cell=(pix_per_cell, pix_per_cell), \n\t\t\t\tcells_per_block=(cell_per_block,cell_per_block), visualise=False, feature_vector=feature_vec, transform_sqrt=True)\n\t\treturn features\n\n\n##############################\ndef convert_to_colorspace(img_bgr, desired_color_space):\n\timg = img_bgr.copy()\n\tif desired_color_space == 'RGB':\n\t\timg = cv2.cvtColor(img_bgr, cv2.COLOR_BGR2RGB)\n\telif desired_color_space == 'HSV':\n\t\tcar_img = cv2.cvtColor(img_bgr, cv2.COLOR_BGR2HSV)\n\telif desired_color_space == 'HLS':\n\t\tcar_img = cv2.cvtColor(img_bgr, cv2.COLOR_BGR2HLS)\n\telif desired_color_space == 'YCrCb':\n\t\tcar_img = cv2.cvtColor(img_bgr, cv2.COLOR_BGR2YCrCb)\n\treturn img\n\n##############################\n# main\nif __name__ == '__main__':\n\tif len(sys.argv) != 3:\n\t\tprint('{} {} {}'.format(sys.argv[0], 'rootdir', 'glob_pattern'))\n\t\tsys.exit(1)\n\n\trootdir = sys.argv[1]\n\tglob_pattern = sys.argv[2]\n\n\timages = glob.glob(rootdir + '\\\\**\\\\' + glob_pattern, recursive=True)\n\tprint('Total # of *.jpeg files: ', len(images))\n\tcars = []\n\tnotcars = []\n\n\tfor image in images:\n\t\tif 'image' in image or 'extra' in image:\n\t\t\tnotcars.append(image)\n\t\telse:\n\t\t\tcars.append(image)\n\n\t# Generate a random index to look at a car image and not car image\n\tcarind = np.random.randint(0, len(cars))\n\tnotcar_ind = np.random.randint(0, len(notcars))\n\n\t# Define HOG parameters\n\t#orient = 9\n\t#pix_per_cell = 8\n\t#cell_per_block = 2\n\tparameters = [(9,8,2), (11,8,2),(9,8,4),(11,8,4),(9,16,2), (11,16,2)]\t# list of tuples where each tuple is of the form (orient, pix_per_cell, cell_per_block)\n\n\t# Read the image\n\tcar_img_bgr = cv2.imread(cars[carind])\n\tnotcar_img_bgr = cv2.imread(notcars[notcar_ind])\n\t# conver to gray\n\tcar_gray = cv2.cvtColor(car_img_bgr, cv2.COLOR_BGR2GRAY)\n\tnotcar_gray = cv2.cvtColor(notcar_img_bgr, cv2.COLOR_BGR2GRAY)\n\t# convert to different color space if needed\n\tcar_img = convert_to_colorspace(car_img_bgr, 'YCrCb')\n\tnotcar_img = convert_to_colorspace(car_img_bgr, 'YCrCb')\n\t\n\tfig = plt.figure(figsize=(12,8))\n\t#fig = plt.figure()\n\tnum_rows = 3\t#len(parameters) + 1\n\tnum_cols = len(parameters)\n\t# plot car and notcar in the 1st row\n\tplt.subplot(num_rows, num_cols, 1)\n\tplt.imshow(car_gray, cmap='gray')\n\tplt.title('car')\n\tplt.subplot(num_rows, num_cols, 2)\n\tplt.imshow(notcar_gray, cmap='gray')\n\tplt.title('notcar')\n\n\tplt_index = num_cols + 1\n\tfor j,params in enumerate(parameters):\n\t\torient, pix_per_cell, cell_per_block = params\t\t\n\t\t# Call HOG function vis=True to see an image output\n\t\tfeatures, car_hog_image_gray = get_hog_features(car_gray, orient, pix_per_cell, cell_per_block, vis=True, feature_vec=False)\n\t\tplt.subplot(num_rows, num_cols, plt_index+j)\n\t\tplt.imshow(car_hog_image_gray, cmap='gray')\n\t\tplt.title('{}'.format(params))\n\n\tplt_index += num_cols\n\tfor j,params in enumerate(parameters):\n\t\torient, pix_per_cell, cell_per_block = params\t\t\n\t\t# Call HOG function vis=True to see an image output\n\t\tfeatures, notcar_hog_image_gray = get_hog_features(notcar_gray, orient, pix_per_cell, cell_per_block, vis=True, feature_vec=False)\n\n\t\t# plot car HOG and notcar HOG\n\t\tplt.subplot(num_rows, num_cols, plt_index+j)\n\t\tplt.imshow(notcar_hog_image_gray, cmap='gray')\n\t\tplt.title('{}'.format(params))\n\tplt.tight_layout()\n\tplt.show()","repo_name":"vnaveen9296/self-driving-cars-vehicle-detection","sub_path":"get_hog.py","file_name":"get_hog.py","file_ext":"py","file_size_in_byte":3849,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"38218064544","text":"# Python 3 server example\r\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\r\nimport urllib.request\r\nimport time\r\nimport json\r\n\r\nhostName = \"localhost\"\r\nserverPort = 4000\r\n\r\nclass MyServer(BaseHTTPRequestHandler):\r\n def do_GET(self):\r\n if self.path == \"/\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"text/html\")\r\n self.end_headers()\r\n f = open(\"index.html\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/default.png\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"image/png\")\r\n self.end_headers()\r\n f = open(\"/home/aj/lights/SFMG.png\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/user\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"application/json\")\r\n self.end_headers()\r\n f = open(\"user.json\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/odometer.js\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"application/javascript\")\r\n self.end_headers()\r\n f = open(\"/home/aj/lights/odometer.js\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/odometer.css\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"text/css\")\r\n self.end_headers()\r\n f = open(\"/home/aj/lights/odometer.css\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/settings\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"application/json\")\r\n self.end_headers()\r\n f = open(\"settings.json\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n elif self.path == \"/favicon.ico\":\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"image/x-icon\")\r\n self.end_headers()\r\n f = open(\"/home/aj/lights/favicon.ico\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n else:\r\n self.send_response(404)\r\n self.send_header(\"Content-type\", \"text/html\")\r\n self.end_headers()\r\n f = open(\"/home/aj/lights/404.html\", \"rb\")\r\n self.wfile.write(f.read())\r\n f.close()\r\n\r\n def do_POST(self):\r\n if self.path == \"/search\":\r\n query = json.loads(self.rfile.read(int(self.headers['Content-Length'])).decode(\"utf-8\"))\r\n req = urllib.request.Request(\"https://axern.space/api/search?platform=youtube&type=channel&query=\"+query[\"search\"], headers={'User-Agent': 'Mozilla/5.0'})\r\n content = urllib.request.urlopen(req).read()\r\n content = json.loads(content)\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"application/json\")\r\n self.end_headers()\r\n self.wfile.write(json.dumps(content).encode(\"utf-8\"))\r\n elif self.path == \"/save\":\r\n query = json.loads(self.rfile.read(int(self.headers['Content-Length'])).decode(\"utf-8\"))\r\n if query['api'] == 'mixerno':\r\n query['api'] = 'https://mixerno.space/api/youtube-channel-counter/user/'\r\n query['path'] = \"counts[0].count\"\r\n elif query['api'] == 'axern':\r\n query['api'] = 'https://axern.space/api/get?platform=youtube&type=channel&id='\r\n query['path'] = \"estSubCount\"\r\n elif query['api'] == 'xyz':\r\n query['api'] = 'https://livecounts.xyz/api/youtube-live-subscriber-count/live/'\r\n query['path'] = \"counts[0]\"\r\n elif query['api'] == 'raw':\r\n query['api'] = 'https://yt.lemnoslife.com/noKey/channels?part=snippet,statistics&id='\r\n query['path'] = \"items[0].statistics.subscriberCount\"\r\n open(\"/home/aj/lights/settings.json\", \"w\").write(json.dumps(query))\r\n self.send_response(200)\r\n self.send_header(\"Content-type\", \"application/json\")\r\n self.end_headers()\r\n self.wfile.write(bytes(\"{\\\"status\\\": \\\"ok\\\"}\", \"utf-8\"))\r\n\r\nif __name__ == \"__main__\": \r\n webServer = HTTPServer((hostName, serverPort), MyServer)\r\n print(\"Server started http://%s:%s\" % (hostName, serverPort))\r\n\r\n try:\r\n webServer.serve_forever()\r\n except KeyboardInterrupt:\r\n pass\r\n\r\n webServer.server_close()\r\n print(\"Server stopped.\")","repo_name":"mgcountss/led-sub-count","sub_path":"server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":4662,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"6475965014","text":"import pytest\n\n\n@pytest.mark.usefixtures(\"test_app\")\nclass TestURLs:\n def test_home(self, test_app):\n \"\"\"\n test the home page\n \"\"\"\n text = b'{\"test\":\"True\"}\\n'\n r = test_app.post(\n '/test',\n data=text,\n content_type='application/json'\n )\n assert r.status_code == 200\n assert r.data == text\n","repo_name":"kaolalicai/flask-scaffold","sub_path":"test/test_urls.py","file_name":"test_urls.py","file_ext":"py","file_size_in_byte":382,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"39858816598","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Requerimientos',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('monitorizacion', models.BooleanField(default=None)),\n ('ventilacion_mecanica', models.BooleanField(default=None)),\n ('inmovilizacion', models.BooleanField(default=None)),\n ('marcapaso', models.BooleanField(default=None)),\n ('BIC', models.CharField(max_length=20, blank=True)),\n ('oxigenoterapia', models.CharField(blank=True, max_length=20, choices=[(b'1', b'Naricera'), (b'2', b'Mascarilla Fio2 35-50%'), (b'3', b'Ventilacion Invasiva | No Invasiva')])),\n ('acceso_vascular', models.CharField(blank=True, max_length=20, choices=[(b'1', b'Periferico'), (b'2', b'Central')])),\n ('hemodinamia', models.CharField(blank=True, max_length=20, choices=[(b'1', b'Estable'), (b'2', b'Inestable')])),\n ('ventilatorio', models.CharField(blank=True, max_length=20, choices=[(b'1', b'Eupneico'), (b'2', b'Alterado')])),\n ('glasgow', models.CharField(blank=True, max_length=20, choices=[(b'1', b'14 - 15'), (b'2', b'9 - 13'), (b'3', b'0 - 8')])),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n ]\n","repo_name":"GonzaloAfa/EmergencyCare","sub_path":"diagnostico/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":1588,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"20911519198","text":"#!/usr/bin/env python3\n\n\n#\ndef insert_in_sortedlist(arr, el):\n n = len(arr)\n if n == 0:\n arr.append(el)\n elif n == 1:\n if el <= arr[0]: arr.insert(0, el)\n else: arr.append(el)\n else:\n if el <= arr[0]: arr.insert(0, el)\n elif el >= arr[-1]: arr.append(el)\n else:\n i = 0\n while i < n-1:\n if arr[i] <= el <= arr[i+1]:\n arr.insert(i+1, el)\n break\n i += 1\n return arr\n\n\n\"\"\"\nce algorithme est-il correct?\nje pense que oui, mais je ne suis pas sur.\nje ne suis pas sur de la complexité temporelle de cet algorithme.\nje pense que c'est O(n^2), mais je ne suis pas sur.\n\"\"\"\ndef insertion (arr) :\n n = len(arr)\n sortedList = []\n for el in arr :\n sortedList = insert_in_sortedlist(sortedList, el)\n \n return sortedList\n\n\n#complexité temporelle: O(n^2)\n#complexité spatiale: O(n)\ndef insertion_sort(arr):\n for i in range(1, len(arr)):\n curr = arr[i]\n j = i-1\n while j >= 0 and curr < arr[j]:\n arr[j+1] = arr[j]\n j -= 1\n arr[j+1] = curr\n return arr\n\n\n# for testing\ndef main() :\n # random list of 15 integers between 0 and 10\n import random\n \n \n listofInt = [random.randint(0,10) for i in range(25)]\n print(listofInt)\n \n sortedList = insertion_sort(listofInt)\n listofInt.sort()\n print(listofInt)\n print()\n print(sortedList == listofInt)\n print()\n \n\nif __name__ == \"__main__\" :\n main()","repo_name":"StenTech/ProblemSolving","sub_path":"Algorithms/Sorting/insertion.py","file_name":"insertion.py","file_ext":"py","file_size_in_byte":1513,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"37286146244","text":"from urllib import request\r\n\r\nfilename = \"mean.zip\"\r\nurl = \"http://ftp.lysator.liu.se/pub/awe32/soundfonts/{}\".format(filename)\r\nprint(\"Downloading: {}\".format(url))\r\nr = request.urlopen(url)\r\nprint(\"Saving to: {}\".format(filename))\r\nf = open(filename, \"wb\")\r\nf.write(r.read())\r\nf.close()\r\nr.close()\r\nprint(\"Done!\")\r\n","repo_name":"xdevsoft/bootcamp","sub_path":"exercise2-1.py","file_name":"exercise2-1.py","file_ext":"py","file_size_in_byte":317,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"15214386901","text":"test_input = '''vJrwpWtwJgWrhcsFMMfFFhFp\njqHRNqRjqzjGDLGLrsFMfFZSrLrFZsSL\nPmmdzqPrVvPwwTWBwg\nwMqvLMZHhHMvwLHjbvcjnnSBnvTQFn\nttgJtRGJQctTZtZT\nCrZsJsPPZsGzwwsLwLmpwMDw'''\ndef character_to_priority(char):\n priority = ord(char.lower()) - ord('a') + 1\n if char.isupper():\n priority += 26\n return priority\n\ndef group_gen(input):\n lines = input.splitlines()\n for n in range(0, len(lines), 3):\n yield lines[n:n+3]\n\ndef get_badge(group):\n commons = set.intersection(*[set(elf) for elf in group])\n assert len(commons) == 1, f'Items {commons} are common in group. Expected exactly one'\n return commons.pop()\n\ndef process_input(input):\n return sum(character_to_priority(get_badge(group)) for group in group_gen(input))\n\nassert process_input(test_input) == 70, 'Priorities for these items must still be found to organize the sticker attachment efforts: here, they are 18 (r) for the first group and 52 (Z) for the second group. The sum of these is 70.'\n\nwith open('input.txt') as f:\n print(process_input(f.read()))","repo_name":"bgrevelt/Advent-of-code-2022","sub_path":"day3/day3_2.py","file_name":"day3_2.py","file_ext":"py","file_size_in_byte":1045,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"16438369176","text":"import tkinter as tk\nimport tkinter.messagebox\n\nwindow=tk.Tk()\nwindow.title(\"Calculator v1.01\")\nframe=tk.Frame(master=window,bg=\"#2f3133\",padx=10)\nframe.pack()\nentry_mes = tk.Entry(master=frame, borderwidth=5, width=30)\nentry_mes.grid(row=0, column=0, columnspan=8, ipady=1, pady=5)\n\n#create function\ndef click_press(number):\n entry_mes.insert(tk.END, number)\n \n\ndef equal():\n try:\n y = str(eval(entry_mes.get()))\n entry_mes.delete(0, tk.END)\n entry_mes.insert(0, y)\n except:\n tkinter.messagebox.showinfo(\"Error\", \"Check Syntax Error\")\n \n#After you hit the equal symbol, initial data will be replaced by the result\ndef clear():\n entry_mes.delete(0, tk.END)\n \n \nbutton_1 = tk.Button(master=frame, text='1', bg=\"#303334\",fg=\"white\",padx=5,\n pady=5, width=3, command=lambda: click_press(1))\nbutton_1.grid(row=1, column=0, pady=1)\nbutton_2 = tk.Button(master=frame, text='2', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(2))\nbutton_2.grid(row=1, column=1, pady=1)\nbutton_3 = tk.Button(master=frame, text='3', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(3))\nbutton_3.grid(row=1, column=2, pady=1)\nbutton_4 = tk.Button(master=frame, text='4', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(4))\nbutton_4.grid(row=2, column=0, pady=1)\nbutton_5 = tk.Button(master=frame, text='5', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(5))\nbutton_5.grid(row=2, column=1, pady=1)\nbutton_6 = tk.Button(master=frame, text='6', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(6))\nbutton_6.grid(row=2, column=2, pady=1)\nbutton_7 = tk.Button(master=frame, text='7', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(7))\nbutton_7.grid(row=3, column=0, pady=1)\nbutton_8 = tk.Button(master=frame, text='8', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(8))\nbutton_8.grid(row=3, column=1, pady=1)\nbutton_9 = tk.Button(master=frame, text='9', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(9))\nbutton_9.grid(row=3, column=2, pady=1)\nbutton_0 = tk.Button(master=frame, text='0', bg=\"#303334\",fg=\"white\", padx=5,\n pady=5, width=3, command=lambda: click_press(0))\nbutton_0.grid(row=4, column=1, pady=2)\n \nbutton_add = tk.Button(master=frame, text=\"+\", bg=\"#303334\",fg=\"white\", padx=15,\n pady=5, width=3, command=lambda: click_press('+'))\nbutton_add.grid(row=1, column=4, pady=2)\n \nbutton_subtract = tk.Button(\n master=frame, text=\"-\", padx=15, pady=5, width=3, bg=\"#303334\",fg=\"white\", \n command=lambda: click_press('-'))\nbutton_subtract.grid(row=2, column=4, pady=2)\n \nbutton_multiply = tk.Button(\n master=frame, text=\"*\", padx=15, pady=5, width=3, bg=\"#303334\",fg=\"white\", \n command=lambda: click_press('*'))\nbutton_multiply.grid(row=3, column=4, pady=2)\n \nbutton_div = tk.Button(master=frame, text=\"/\", padx=5, bg=\"#303334\",fg=\"white\",\n pady=5, width=3, command=lambda: click_press('/'))\nbutton_div.grid(row=4, column=4, pady=2)\n \nbutton_clear = tk.Button(master=frame, text=\"Clear\", bg=\"red\",fg=\"white\",\n padx=15, pady=5, width=12, command=clear())\nbutton_clear.grid(row=1, column=5, columnspan=2, pady=2)\n \nbutton_equal = tk.Button(master=frame, text=\"=\", bg=\"darkblue\",fg=\"white\", padx=5,\n pady=5, width=9, command=equal)\nbutton_equal.grid(row=2, column=5, columnspan=3, pady=2)\n\nwindow.mainloop()\n\n","repo_name":"Craio98/Py_Calculator_GUI","sub_path":"scripts/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3776,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"30270642391","text":"# -*- coding:utf-8 -*-\n__author__ = 'ChenJiaBao'\n__date__ = '2018/9/20 15:02'\nfrom . import index_blu\nfrom flask import render_template, current_app, session, request, jsonify, g\nfrom info.models import User, News, Category\nfrom info import constants\nfrom info.utils.response_code import RET\nfrom info.utils.common import user_login_data\n\n\n@index_blu.route('/')\n@user_login_data\ndef index():\n \"\"\"\n 首页\n :return:\n \"\"\"\n user = g.user\n data = {\n 'user': user.to_dict() if user else None\n }\n\n # 右侧排行数据\n news_list = News.query.order_by(News.clicks.desc()).limit(constants.CLICK_RANK_MAX_NEWS)\n\n # 查询分类数据\n categories = Category.query.all()\n\n return render_template('news/index.html', data=data, news_list=news_list, categories=categories)\n\n\n@index_blu.route('/favicon.ico')\ndef favicon():\n \"\"\"加载网站小图标\"\"\"\n return current_app.send_static_file('news/favicon.ico')\n\n\n@index_blu.route('/news_list')\ndef news_list():\n \"\"\"\n 获取首页新闻数据\n \"\"\"\n cid = request.args.get('cid', '1')\n page = request.args.get('page', '1')\n per_page = request.args.get('per_page', '10')\n try:\n cid = int(cid)\n page = int(page)\n per_page = int(per_page)\n except Exception as e:\n current_app.logger.error(e)\n return jsonify(errno=RET.PARAMERR, errmsg='参数错误')\n filters = [News.status == 0]\n if cid != 1:\n filters.append(News.category_id == cid)\n try:\n page_data = News.query.filter(*filters).order_by(News.create_time.desc()).paginate(page, per_page, False)\n total_page = page_data.pages\n except Exception as e:\n current_app.logger.error(e)\n return jsonify(errno=RET.DBERR, errmsg='数据查询错误')\n news_data = []\n for news in page_data.items:\n news_data.append(news.to_basic_dict())\n return jsonify(errno=RET.OK, errmsg='ok', news_data=news_data, total_page=total_page)\n\n","repo_name":"chenjb04/News","sub_path":"info/modules/index/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1969,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"2838278066","text":"import cv2\nfrom tracing import *\nimport direct as Direction\nimport stop_detection as STOP \nimport numpy as np\n\nvideo = cv2.VideoCapture(-1)\nlocation=(30,30)\nfont=cv2.FONT_HERSHEY_SIMPLEX\nfontscale=1\nthickness=3\nSTOP_DETECT_COUNT=0\nwhile True:\n _, frame = video.read()\n D=Direction.direct_detection(frame,STOP_DETECT_COUNT)\n stop_frame=frame[150:,:]\n if STOP.stop_detection(stop_frame,STOP_DETECT_COUNT)==True:\n STOP_DETECT_COUNT+=1\n stop()\n sleep(5)\n elif D:\n if D=='R':\n right(100)\n else:\n left(100)\n else: \n frame1 = frame.copy() #흰색용 복사\n frame2 = frame.copy() #검은색용 복사\n frame3 = frame.copy() #신호등, 표지판 \n #관심영역 아래로 설정\n\n gray1 = cv2.cvtColor(frame1,cv2.COLOR_BGR2GRAY)\n blur1 = cv2.GaussianBlur(gray1,(5,5),0)\n _,white_dst = cv2.threshold(blur1,240,255,cv2.THRESH_BINARY+cv2.THRESH_OTSU)\n white = white_dst[370:480,0:640]\n black=cv2.bitwise_not(white)\n #threshold\n #contour\n# cv2.imshow('white',white)\n contours1,hierarchy = cv2.findContours(white.copy(), 1, cv2.CHAIN_APPROX_NONE)\n contours2,_=cv2.findContours(black.copy(), 1, cv2.CHAIN_APPROX_NONE)\n if len(contours1)>0 and len(contours2)>0: #흰 검 둘다잡힐떄\n c=max(contours1, key=cv2.contourArea)\n M=cv2.moments(c)\n try:\n cx=int(M['m10']/M['m00'])\n cy=int(M['m01']/M['m00'])\n #무게중심코드\n except ZeroDivisionError as e:\n # print(\"what?\")\n straight(100)\n continue\n if cx<=365:\n right(100)\n else:\n left(100)\n elif len(contours1)<=0 and len(contours2)>0: #검은색만 잡힐때\n straight(100)\n else:\n # print(\"what is it?\")\n pass\n # cv2.imshow('frame',frame)\n if cv2.waitKey(1) & 0xFF == ord('q'):break\nMotor_end()\nframe.release()\ncv2.destroyAllWindows()","repo_name":"hyunsooklm/Edison-LineTracing","sub_path":"trace.py","file_name":"trace.py","file_ext":"py","file_size_in_byte":2106,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"38539140309","text":"# import numpy as np\n# import h5py\n# import matplotlib.pyplot as plt\n# from rasterio.plot import show\n# pcp = h5py.File('D:/Users/khoramian.a/Downloads/3B', 'r')\n# pcp=pcp['Grid']\n# pcp=pcp['precipitationCal']\n# pcp=np.array(pcp)\n# pcp=pcp[0,::,::]\n# cn=pcp[0,0]\n# pcp = np.where(pcp == cn, np.nan, pcp)\n\n# # for i in pcp:\n# # for j in i:\n# # if j<-999:\n# # j=np.nan\n \n# # plt.imshow(pcp)\n# # plt.show()\n# show(pcp)\nimport datetime\nimport math\nimport requests\nimport json\nimport os\nfrom shutil import copyfile\nfrom PCPfunctions import maskingfiles\n\nst_date=[2020,11,29,7,25]\nend_date=[2020,11,29,23,35]\nst=datetime.datetime(st_date[0],st_date[1],st_date[2],st_date[3],st_date[4])\ned=datetime.datetime(end_date[0],end_date[1],end_date[2],end_date[3],end_date[4])\ndiff=ed-st\nsteps=diff.seconds/(60*30)\n# intitae first time step and steps count\nst=st.replace(minute=30*math.floor(st.minute/30))\nsteps=math.floor(steps)\nstp_date=st\n#\n\n \n#API request\napiresponse = requests.get('https://pmmpublisher.pps.eosdis.nasa.gov/opensearch?q=precip_30mn&lat=33&lon=53&limit=100&startTime=' + str(st_date[0]) + '-' + str(st_date[1]).zfill(2) + '-' + str(st_date[2]).zfill(2) + '&endTime=' + str(end_date[0]) + '-' + str(end_date[1]).zfill(2) + '-' + str(end_date[2]).zfill(2))\njsondic=json.loads(apiresponse.text)\ndel apiresponse\nfor i in range (0,jsondic['totalItems']):\n print(i)\n api_folder_name=jsondic['items'][i]['properties']['date']['@value']\n api_folder_name='api/'+api_folder_name[0:4]+api_folder_name[5:7]\n file_name=jsondic['items'][i]['@id']+'.tif'\n api_path_name=api_folder_name+'/'+file_name\n DB_folder_name='DB/'+api_folder_name[4:8]+api_folder_name[8:10]\n DB_path_name=DB_folder_name+'/'+file_name\n #Checking file exist in DB folder or not and if not download it\n if not os.path.isfile(DB_path_name):\n adress=jsondic['items'][i]['action'][1]['using'][1]['url']\n req = requests.get(adress)\n if not os.path.exists(api_folder_name):\n os.makedirs(api_folder_name) \n with open(api_path_name, 'wb') as f:\n f.write(req.content)\n del adress, req, f\n # Make a copy to DB folder\n if not os.path.exists(DB_folder_name):\n os.makedirs(DB_folder_name)\n copyfile(api_path_name, DB_path_name)\n#\n\n\n#Getting remainings from ftp server\n\n# ftp request\nfor i in range(0,steps):\n print(i)\n stp2_date=stp_date+datetime.timedelta(seconds=59,minutes=29)\n file_name='gpm_30mn_'+str(stp2_date.year)+str(stp2_date.month).zfill(2)+str(stp2_date.day).zfill(2)+ '_' + str(stp2_date.hour).zfill(2)+str(stp2_date.minute).zfill(2) + str(stp2_date.second).zfill(2) + '.tif'\n BD_folder_name='DB/'+str(stp2_date.year)+str(stp2_date.month).zfill(2)\n ftp_folder_name='ftp/'+str(stp2_date.year)+str(stp2_date.month).zfill(2)\n DB_path_name=DB_folder_name + '/' + file_name\n ftp_path_name=ftp_folder_name + '/' + file_name\n # 3B-HHR-E.MS.MRG.3IMERG.20201101-S003000-E005959.0030.V06B.RT-H5\n if not os.path.isfile(DB_path_name):\n digi=str((math.floor((stp_date.hour*60+stp_date.minute)/30))*30).zfill(4)\n file_url='https://jsimpsonhttps.pps.eosdis.nasa.gov/imerg/gis/early/'+ str(stp_date.year) + '/' + str(stp_date.month).zfill(2) + '/3B-HHR-E.MS.MRG.3IMERG.' + str(stp_date.year) + str(stp_date.month).zfill(2) + str(stp_date.day).zfill(2) + '-S' + str(stp_date.hour).zfill(2) + str(stp_date.minute).zfill(2) + str(stp_date.second).zfill(2) + '-E' + str(stp2_date.hour).zfill(2) + str(stp2_date.minute).zfill(2) + str(stp2_date.second).zfill(2) + '.' + digi + '.V06B.30min.tif'\n with requests.Session() as session:\n req = session.request('get', file_url)\n r = session.get(req.url, auth=('khoramian.a@gmail.com', 'khoramian.a@gmail.com'))\n if not os.path.exists(ftp_folder_name):\n os.makedirs(ftp_folder_name)\n with open(ftp_path_name, 'wb') as f:\n f.write(r.content)\n del r, req, f\n maskingfiles(ftp_path_name,DB_path_name)\n # else:\n # copyfile(src, dst)\n stp_date=stp_date+datetime.timedelta(minutes=30)\n \n\n","repo_name":"aminkhoramian/Precipitation","sub_path":"readingimerg.py","file_name":"readingimerg.py","file_ext":"py","file_size_in_byte":4191,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"3553164269","text":"class Solution(object):\n def containsDuplicate(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: bool\n \"\"\"\n set_dic = set()\n for num in nums:\n if num in set_dic:\n return True\n else:\n set_dic.add(num)\n return False \n ","repo_name":"talha-husnain/interview_practice","sub_path":"5_check_dulplicates.py","file_name":"5_check_dulplicates.py","file_ext":"py","file_size_in_byte":327,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"7891179448","text":"import unittest\nimport pytest\nfrom src.repositories.journey_repository import JourneyRepository\nfrom src.repositories.database import db\n\n\nclass TestJourneyRepository(unittest.TestCase):\n @pytest.fixture(autouse=True)\n def prepare_fixture(self, app):\n self.app = app\n\n def setUp(self):\n self.repository = JourneyRepository(db)\n\n def test_get_defined_range_list_of_all_journeys_as_Journey_objects_in_id_number_order(self):\n with self.app.app_context():\n result = self.repository.get_range_from_all_journeys_by_id(0, 10)\n\n self.assertEqual(len(result), 10)\n self.assertEqual(\n str(result[0]),\n '2021-05-31 23:57:25 094 -> 100 2021-06-01 00:05:46, 2043 m, 500 sec'\n )\n self.assertEqual(result[0].id, 1)\n self.assertEqual(\n str(result[9]),\n '2021-05-31 23:50:19 116 -> 145 2021-06-01 00:05:58, 3248 m, 935 sec'\n )\n self.assertEqual(result[9].id, 10)\n\n def test_get_defined_range_list_of_all_journeys_as_Journey_objects_in_decreasing_time_order(self):\n with self.app.app_context():\n result = self.repository.get_range_from_all_journeys_by_time(\n 6, 11, True)\n\n self.assertEqual(len(result), 5)\n self.assertEqual(\n str(result[0]),\n '2021-05-31 23:54:11 034 -> 081 2021-06-01 00:17:11, 2550 m, 1377 sec'\n )\n self.assertEqual(result[0].id, 7)\n self.assertEqual(\n str(result[4]),\n '2021-05-31 23:50:05 147 -> 232 2021-06-01 00:01:22, 1633 m, 672 sec'\n )\n self.assertEqual(result[4].id, 11)\n\n def test_get_defined_range_list_of_all_journeys_as_Journey_objects_in_increasing_time_order(self):\n with self.app.app_context():\n result = self.repository.get_range_from_all_journeys_by_time(\n 0, 3, False)\n\n self.assertEqual(len(result), 3)\n self.assertEqual(\n str(result[0]),\n '2021-05-31 23:28:16 082 -> 084 2021-05-31 23:31:00, 665 m, 159 sec'\n )\n self.assertEqual(result[0].id, 33)\n self.assertEqual(\n str(result[2]),\n '2021-05-31 23:30:06 547 -> 547 2021-05-31 23:49:17, 739 m, 1146 sec'\n )\n self.assertEqual(result[2].id, 31)\n\n def test_get_defined_range_list_of_all_journeys_as_Journey_objects_in_decreasing_distance_order(self):\n with self.app.app_context():\n result = self.repository.get_range_from_all_journeys_by_distance(\n 0, 2)\n\n self.assertEqual(len(result), 2)\n self.assertEqual(\n str(result[0]),\n '2021-05-31 23:31:27 315 -> 272 2021-05-31 23:57:24, 5495 m, 1553 sec'\n )\n self.assertEqual(\n str(result[1]),\n '2021-05-31 23:53:04 240 -> 281 2021-06-01 00:14:52, 5366 m, 1304 sec'\n )\n\n def test_get_defined_range_list_of_all_journeys_as_Journey_objects_in_decreasing_duration_order(self):\n with self.app.app_context():\n result = self.repository.get_range_from_all_journeys_by_duration(\n 0, 1)\n\n self.assertEqual(len(result), 1)\n self.assertEqual(\n str(result[0]),\n '2021-05-31 23:30:45 573 -> 511 2021-06-01 15:45:49, 2834 m, 58499 sec'\n )\n","repo_name":"kivistoilkka/city-bike-app","sub_path":"src/tests/repositories/journey_repository_test.py","file_name":"journey_repository_test.py","file_ext":"py","file_size_in_byte":3326,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"38178548363","text":"from ..Semi import SemiUser\nfrom ..ClsUtil import extra_kw\n\n__all__ = [\"Invite\"]\n\nclass Invite:\n \"\"\"\n {{cls}} instance = Invite(*, too_many_args_to_list_here)\n\n {{desc}} Represents an invite\n\n {{noinit}}\n\n {{param}} code [str]\n Invite code, eg `ZM84Nm6n`\n\n {{param}} guild [dict]\n The guild object\n\n {{param}} channel [dict]\n The channel object\n\n {{param}} target_user [dict]\n The targeted user\n\n {{param}} target_user_type [int]\n I honestly do not know\n\n {{param}} approximate_presence_count [int]\n Rough number of statuses\n\n {{param}} approximate_member_count [int]\n Rough number of players\n\n {{param}} bot_obj [~/Bot]\n The bot object\n\n {{prop}} code [str]\n Invite code, eg `ZM84Nm6n`\n\n {{prop}} guild_id [int]\n The guild ID\n\n {{prop}} guild [~/Guild]\n The guild object\n\n {{prop}} channel_id [int]\n The channel ID\n\n {{prop}} channel [~/Channel.AnyChannel]\n The channel object\n\n {{prop}} target [~/Semi.SemiUser]\n The target user\n\n {{prop}} target_type [int]\n I honestly do not know\n\n {{prop}} rough_status_count [int]\n Rough number of statuses\n\n {{prop}} rough_player_count [int]\n Rough number of players\n\n {{prop}} bot_obj [~/Bot]\n The bot object\n \"\"\"\n def __init__(self, *, code, guild, channel, target_user, target_user_type,\n approximate_presence_count, approximate_member_count,\n bot_obj = None, **kw):\n extra_kw(kw, \"Invite\")\n self.code = code\n self.url = \"https://discord.gg/\" + code\n self.target = SemiUser(**self.target_user)\n self.target_type = self.target_user_type\n self.rough_status_count = approximate_presence_count\n self.rough_player_count = approximate_member_count\n self.bot_obj = bot_obj\n self.guild_id = guild[\"id\"]\n self.channel_id = channel[\"id\"]\n bot_obj.listener.invites[code] = self\n\n @property\n def guild(self):\n return self.bot_obj.guilds(self.guild_id)\n\n @property\n def channel(self):\n return self.bot_obj.all_channels(self.channel_id)\n\n async def delete(self):\n \"\"\"\n {{fn}} await instance.delete()\n\n {{desc}} Deletes this object and the invite\n\n {{note}} This does NOT return a ~/NonExistentObj after deletion because\n Discord creates a random invite code\n \"\"\"\n await self.bot_obj.http.req(m = \"-\", u = f\"/invites/{self.code}\")\n del self.bot_obj.listener.invites[self.code]\n del self\n\n def __dict__(self):\n \"\"\"\n {{bltin}} instance.__dict__()\n {{usage}} dict(instance)\n\n {{pydesc}} __dict__\n\n {{rtn}} [dict]\n \"\"\"\n return {\n \"code\": self.code,\n \"guild\": dict(self.guild),\n \"channel\": dict(self.channel),\n \"target_user\": dict(self.target),\n \"target_user_type\": self.target_type,\n \"approximate_member_count\": self.rough_player_count,\n \"approximate_presence_count\": self.rough_status_count\n }\n","repo_name":"VoxelPrismatic/prizmatic","sub_path":"discord/models/Invite/Invite.py","file_name":"Invite.py","file_ext":"py","file_size_in_byte":3154,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"30793497684","text":"import pytest\nfrom gmsh import model as gmsh_model\nfrom bgem.gmsh import gmsh, field, options\nimport numpy as np\nimport os\nimport math\nimport yaml\n\n\ndef create_fractures_rectangles(gmsh_geom, fractures, base_shape: 'ObjectSet'):\n # From given fracture date list 'fractures'.\n # transform the base_shape to fracture objects\n # fragment fractures by their intersections\n # return dict: fracture.region -> GMSHobject with corresponding fracture fragments\n shapes = []\n for i, fr in enumerate(fractures):\n shape = base_shape.copy()\n print(\"fr: \", i, \"tag: \", shape.dim_tags)\n shape = shape.scale([fr.rx, fr.ry, 1]) \\\n .rotate(axis=fr.rotation_axis, angle=fr.rotation_angle) \\\n .translate(fr.centre) \\\n .set_region(fr.region)\n\n shapes.append(shape)\n\n fracture_fragments = gmsh_geom.fragment(*shapes)\n return fracture_fragments\n\ndef apply_field3(dim, tolerance=0.15, max_mismatch=5, mesh_name=\"field_mesh\"):\n \"\"\"\n Create a mesh of dimension dim on a unit cube and\n compare element sizes to given reference function of coordinates.\n \"\"\"\n\n sample_dir = \"output\"\n os.chdir(sample_dir)\n\n model = gmsh.GeometryOCC(mesh_name)\n\n rec = model.rectangle([20, 20]).set_region(\"square2\")\n boundaries = gmsh.ObjectSet.get_boundary(rec, True)\n\n f1 = 3 - field.y * field.y / 5 - field.x * field.x / 5\n f2 = field.y * field.y / 50 + field.x * field.x / 50\n\n\n #f2 = (field.abs(field.y) +10) / 10\n\n f = field.maximum(f1, f2)\n #f = f2\n\n model.set_mesh_step_field(f)\n # model.set_mesh_step_field(field2)\n # rec1.mesh_step(1)\n # rec.mesh_step(0.5)\n\n model.write_brep()\n model.mesh_options.CharacteristicLengthMin = 0.5\n model.mesh_options.CharacteristicLengthMax = 10\n model.make_mesh([rec], dim=dim)\n model.write_mesh(mesh_name + \".msh2\", gmsh.MeshFormat.msh2)\n\n del model\n\n","repo_name":"GeoMop/Endorse-legacy","sub_path":"experiments/edz_transport/trash/mesh_field3.py","file_name":"mesh_field3.py","file_ext":"py","file_size_in_byte":1905,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"36178100361","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Oct 1 08:26:12 2020\n\n@author: mehedi.md.hasan\n\"\"\"\n\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep 17 00:25:23 2020\n\n@author: mehedi.md.hasan\n\"\"\"\n\nimport yaml\n\n\ndef main():\n with open('test.yml', 'r') as f:\n playbook = yaml.load(f)\n \n# print(playbook)\n \n cleanUps = findCleanUp(playbook)\n if len(cleanUps) > 0 :\n print(\"Environment has been cleaned up at least once\")\n else:\n print(\"Environment has not been cleaned up in any task\")\n\n \ndef findCleanUp(playbook):\n \n cleanedUpRoles = {}\n totalRoles = []\n cleanUpSubstrings = ['cleanup']\n\n for role in playbook:\n roleVars = role['vars']\n \n for roleVar in roleVars:\n if (any([substring in roleVar for substring in cleanUpSubstrings])):\n \n# print(\"Cleaning Up: \" + roleVar)\n cleanedUpRoles['roleName'] = role['name']\n cleanedUpRoles['varName'] = roleVar\n totalRoles.append(cleanedUpRoles)\n \n \n \n return totalRoles\n\n\n\nif __name__ == \"__main__\":\n main()","repo_name":"akondrahman/IaCTesting","sub_path":"old_detectors/converted_detecting_test_environment_cleanup_in_yaml.py","file_name":"converted_detecting_test_environment_cleanup_in_yaml.py","file_ext":"py","file_size_in_byte":1142,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"94"} +{"seq_id":"7435813505","text":"import numpy as np\nimport torch\n\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data.distributed import DistributedSampler\n\nfrom . import data\nfrom . import losses\nfrom . import models \nfrom . import optim\nfrom . import tasks\n\n\ndef get_name_and_params(base):\n name = getattr(base, 'name')\n params = getattr(base, 'params') or {}\n return name, params\n\n\ndef get_transform(base, transform, mode=None):\n if not base: return None\n transform = getattr(base, transform)\n if not transform: return None\n name, params = get_name_and_params(transform)\n if mode:\n params.update({'mode': mode})\n return getattr(data.transforms, name)(**params)\n\n\ndef build_transforms(cfg, mode):\n # 1-Resize\n resizer = get_transform(cfg.transform, 'resize')\n # 2-(Optional) Data augmentation\n augmenter = None\n if mode == \"train\":\n augmenter = get_transform(cfg.transform, 'augment')\n # 3-(Optional) Crop\n cropper = get_transform(cfg.transform, 'crop', mode=mode)\n # 4-Preprocess\n preprocessor = get_transform(cfg.transform, 'preprocess')\n return {\n 'resize': resizer,\n 'augment': augmenter,\n 'crop': cropper,\n 'preprocess': preprocessor\n }\n\n\ndef build_dataset(cfg, data_info, mode):\n dataset_class = getattr(data.datasets, cfg.data.dataset.name)\n dataset_params = cfg.data.dataset.params\n dataset_params.test_mode = mode != 'train'\n dataset_params = dict(dataset_params)\n if \"FeatureDataset\" not in cfg.data.dataset.name:\n transforms = build_transforms(cfg, mode)\n dataset_params.update(transforms)\n dataset_params.update(data_info)\n return dataset_class(**dataset_params)\n\n\ndef build_dataloader(cfg, dataset, mode):\n\n def worker_init_fn(worker_id): \n np.random.seed(np.random.get_state()[1][0] + worker_id)\n\n dataloader_params = {}\n dataloader_params['num_workers'] = cfg.data.num_workers\n dataloader_params['drop_last'] = mode == 'train'\n dataloader_params['shuffle'] = mode == 'train'\n dataloader_params[\"pin_memory\"] = cfg.data.get(\"pin_memory\", True)\n if mode in ('train', 'valid'):\n if mode == \"train\":\n dataloader_params['batch_size'] = cfg.train.batch_size\n elif mode == \"valid\":\n dataloader_params[\"batch_size\"] = cfg.evaluate.get(\"batch_size\") or cfg.train.batch_size\n sampler = None\n if cfg.data.get(\"sampler\") and mode == 'train':\n name, params = get_name_and_params(cfg.data.sampler)\n sampler = getattr(data.samplers, name)(dataset, **params)\n if sampler:\n dataloader_params['shuffle'] = False\n if cfg.strategy == 'ddp':\n sampler = data.samplers.DistributedSamplerWrapper(sampler)\n dataloader_params['sampler'] = sampler\n print(f'Using sampler {sampler} for training ...')\n elif cfg.strategy == 'ddp':\n dataloader_params[\"shuffle\"] = False\n dataloader_params['sampler'] = DistributedSampler(dataset, shuffle=mode==\"train\")\n else:\n assert cfg.strategy != \"ddp\", \"DDP currently not supported for inference\"\n dataloader_params['batch_size'] = cfg.evaluate.get(\"batch_size\") or cfg.train.batch_size\n\n loader = DataLoader(dataset,\n **dataloader_params,\n worker_init_fn=worker_init_fn)\n return loader\n\n\ndef build_model(cfg):\n name, params = get_name_and_params(cfg.model)\n if cfg.model.params.get(\"cnn_params\", None):\n cnn_params = cfg.model.params.cnn_params\n if cnn_params.get(\"load_pretrained_backbone\", None):\n if \"foldx\" in cnn_params.load_pretrained_backbone:\n cfg.model.params.cnn_params.load_pretrained_backbone = cnn_params.load_pretrained_backbone.\\\n replace(\"foldx\", f\"fold{cfg.data.outer_fold}\")\n print(f'Creating model <{name}> ...')\n model = getattr(models.engine, name)(**params)\n if 'backbone' in cfg.model.params:\n print(f' Using backbone <{cfg.model.params.backbone}> ...')\n if 'pretrained' in cfg.model.params:\n print(f' Pretrained : {cfg.model.params.pretrained}')\n if \"load_pretrained\" in cfg.model:\n import re\n if \"foldx\" in cfg.model.load_pretrained:\n cfg.model.load_pretrained = cfg.model.load_pretrained.replace(\"foldx\", f\"fold{cfg.data.outer_fold}\")\n print(f\" Loading pretrained checkpoint from {cfg.model.load_pretrained}\")\n weights = torch.load(cfg.model.load_pretrained, map_location=lambda storage, loc: storage)['state_dict']\n weights = {re.sub(r'^model.', '', k) : v for k,v in weights.items() if \"loss_fn\" not in k}\n model.load_state_dict(weights) \n return model \n\n\ndef build_loss(cfg):\n name, params = get_name_and_params(cfg.loss)\n print(f'Using loss function <{name}> ...')\n params = dict(params)\n if \"pos_weight\" in params:\n params[\"pos_weight\"] = torch.tensor(params[\"pos_weight\"])\n criterion = getattr(losses, name)(**params)\n return criterion\n\n\ndef build_scheduler(cfg, optimizer):\n # Some schedulers will require manipulation of config params\n # My specifications were to make it more intuitive for me\n name, params = get_name_and_params(cfg.scheduler)\n print(f'Using learning rate schedule <{name}> ...')\n\n if name == 'CosineAnnealingLR':\n # eta_min <-> final_lr\n # Set T_max as 100000 ... this is changed in on_train_start() method\n # of the LightningModule task \n\n params = {\n 'T_max': 100000,\n 'eta_min': max(params.final_lr, 1.0e-8)\n }\n\n if name in ('OneCycleLR', 'CustomOneCycleLR'):\n # Use learning rate from optimizer parameters as initial learning rate\n lr_0 = cfg.optimizer.params.lr\n lr_1 = params.max_lr\n lr_2 = params.final_lr\n # lr_0 -> lr_1 -> lr_2 \n pct_start = params.pct_start\n params = {}\n params['steps_per_epoch'] = 100000 # see above- will fix in task\n params['epochs'] = cfg.train.num_epochs\n params['max_lr'] = lr_1\n params['pct_start'] = pct_start\n params['div_factor'] = lr_1 / lr_0 # max/init\n params['final_div_factor'] = lr_0 / max(lr_2, 1.0e-8) # init/final\n\n scheduler = getattr(optim, name)(optimizer=optimizer, **params)\n \n # Some schedulers might need more manipulation after instantiation\n if name in ('OneCycleLR', 'CustomOneCycleLR'):\n scheduler.pct_start = params['pct_start']\n\n # Set update frequency\n if name in ('OneCycleLR', 'CustomOneCycleLR', 'CosineAnnealingLR'):\n scheduler.update_frequency = 'on_batch'\n elif name in ('ReduceLROnPlateau'):\n scheduler.update_frequency = 'on_valid'\n else:\n scheduler.update_frequency = 'on_epoch'\n\n return scheduler\n\n\ndef build_optimizer(cfg, parameters):\n name, params = get_name_and_params(cfg.optimizer)\n print(f'Using optimizer <{name}> ...')\n optimizer = getattr(optim, name)(parameters, **params)\n return optimizer\n\n\ndef build_task(cfg, model):\n name, params = get_name_and_params(cfg.task)\n print(f'Building task <{name}> ...')\n return getattr(tasks, name)(cfg, model, **params)\n\n\n","repo_name":"i-pan/kaggle-rsna-cspine","sub_path":"src/skp/builder.py","file_name":"builder.py","file_ext":"py","file_size_in_byte":7242,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"94"} +{"seq_id":"14024699426","text":"#design a calculator which will correctly solve all tghe problems except the folloing\r\n#45*3 = 555, 56*9=77, 56/6=4\r\n#your code should take operator and two numbers as input from the user and then return the result\r\na = float(input(\"Enter first number: \"))\r\nb = float(input(\"Enter second number: \"))\r\noperation = [\"+\", \"-\", \"*\", \"/\"]\r\nprint(\"+ indicates addition,- indicates subtraction,/ indicates division, * indicates multiplication \")\r\nprint(operation)\r\nc = input(\"operator required: \")\r\nfaulty_numbers = [45,3,56,9,56,6]\r\nif c == \"+\":\r\n print(a+b)\r\nelif c == \"-\":\r\n print(a-b)\r\nelif c == \"*\":\r\n if (a == 45) & (b == 3):\r\n print(\"555\")\r\n elif (a == 56) & (b == 9):\r\n print(\"77\")\r\n else:\r\n print(a*b)\r\nelse:\r\n if (a == 56) & (b == 6):\r\n print(\"4\")\r\n else:\r\n print(a/b)\r\n\r\n\r\n","repo_name":"ROHANSHINDE1234/python-","sub_path":"Faulty calculator.py","file_name":"Faulty calculator.py","file_ext":"py","file_size_in_byte":834,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"24579968910","text":"from spectlib.watch import Watch\nimport spectlib.config\nimport spectlib.util\n\nimport urllib2\nimport urllib \nimport re\nimport os\nfrom xml.dom import minidom\nfrom xml.etree import ElementTree as ET\n\ntype = \"Watch_web_greader\"\ntype_desc = _(\"Google Reader\")\nicon = 'internet-news-reader'\ncategory = _(\"Internet\")\n\n\ndef get_add_gui_info():\n return [(\"username\", spectlib.gtkconfig.Entry(_(\"Username\"))),\n (\"password\", spectlib.gtkconfig.PasswordEntry(_(\"Password\")))]\n\n\nclass Watch_web_greader(Watch):\n \"\"\"\n this watch will check if you have new news on your google reader account\n \"\"\"\n\n def __init__(self, specto, id, values):\n watch_values = [(\"username\", spectlib.config.String(True)),\n (\"password\", spectlib.config.String(True))]\n\n url = \"http://www.google.com/reader/\"\n self.standard_open_command = spectlib.util.return_webpage(url)\n\n #Init the superclass and set some specto values\n Watch.__init__(self, specto, id, values, watch_values)\n\n self.use_network = True\n self.icon = icon\n self.type_desc = type_desc\n self.cache_file = os.path.join(self.specto.CACHE_DIR, \"greader\" + self.username + \".cache\")\n\n #watch specific values\n self.unreadMsg = 0\n self.newMsg = 0\n self.news_info = Feed_collection()\n self.or_more = \"\"\n\n self.read_cache_file()\n\n def check(self):\n \"\"\" Check for new news on your greader account. \"\"\"\n try:\n self.newMsg = 0\n self.unreadMsg = 0\n greader = Greader(self.username, self.password, \"specto\")\n auth = greader.login()\n feed_db = greader.get_unread_items(auth)\n for feed in feed_db:\n self.unreadMsg += feed.messages\n if feed.messages > 0 and self.news_info.add(feed):\n self.actually_changed = True\n self.newMsg += feed.messages\n if self.unreadMsg == 0:#no unread items, we need to clear the watch\n self.mark_as_read()\n self.news_info = Feed_collection()\n else:\n if self.unreadMsg == 1000:\n self.or_more = _(\" or more\")\n \n self.write_cache_file()\n\n except:\n self.set_error()\n\n Watch.timer_update(self)\n\n def get_gui_info(self):\n return [(_('Name'), self.name),\n (_('Last changed'), self.last_changed),\n (_('Username'), self.username),\n (_('Unread messages'), str(self.unreadMsg) + self.or_more)]\n\n def get_balloon_text(self):\n \"\"\" create the text for the balloon \"\"\"\n unread_messages = self.news_info.get_unread_messages()\n if len(unread_messages) == 1:\n text = _(\"New newsitems in %s...\\n\\n... totalling %s unread items.\") %(unread_messages[0].name, str(self.unreadMsg) + self.or_more)\n else:\n i = 0 #show max 4 feeds\n feed_info = \"\"\n while i < len(unread_messages) and i < 4:\n feed_info += unread_messages[i].name + \", \"\n if i == 3 and i < len(unread_messages) - 1:\n feed_info += _(\"and others...\")\n i += 1\n feed_info = feed_info.rstrip(\", \")\n text = _(\"%d new newsitems in %s...\\n\\n... totalling %s unread items.\") %(self.newMsg, feed_info, str(self.unreadMsg) + self.or_more)\n return text\n\n def get_extra_information(self):\n i = 0\n feed_info = \"\"\n while i < len(self.news_info) and i < 4:\n # TODO: do we need to self.escape the name and messages?\n feed_info += \"\" + self.news_info[i].name + \": \" + str(self.news_info[i].messages) + \"\\n\"\n if i == 3 and i < len(self.news_info) - 1:\n feed_info += _(\"and others...\")\n i += 1\n return feed_info\n\n def read_cache_file(self):\n if os.path.exists(self.cache_file):\n try:\n f = open(self.cache_file, \"r\")\n except:\n self.specto.logger.log(_(\"There was an error opening the file %s\") % self.cache_file, \"critical\", self.name)\n else:\n for line in f:\n info = line.split(\"&Separator;\")\n feed = Feed(info[0], info[1].replace(\"\\n\", \"\"))\n self.news_info.add(feed)\n\n finally:\n f.close()\n\n def write_cache_file(self):\n self.news_info.remove_old()\n try:\n f = open(self.cache_file, \"w\")\n except:\n self.specto.logger.log(_(\"There was an error opening the file %s\") % self.cache_file, \"critical\", self.name)\n else:\n for feed in self.news_info:\n f.write(feed.name + \"&Separator;\" + str(feed.messages) + \"\\n\")\n finally:\n f.close()\n\n def remove_cache_files(self):\n os.unlink(self.cache_file)\n\n\nclass Feed():\n\n def __init__(self, name, messages):\n self.name = name\n self.messages = int(messages)\n self.found = False\n self.new = False\n\n\nclass Feed_collection():\n\n def __init__(self):\n self.feed_collection = []\n\n def add(self, feed):\n self.new = True\n self.changed = False\n for _feed in self.feed_collection:\n if feed.name == _feed.name:\n if feed.messages > _feed.messages:\n self.new = False\n self.changed = True\n _feed.messages = feed.messages\n _feed.found = True\n else:\n _feed.messages = feed.messages\n self.new = False\n _feed.found = True\n\n if self.new == True:\n feed.found = True\n feed.new = True\n self.feed_collection.append(feed)\n return True\n elif self.changed == True:\n feed.found = True\n feed.updated = True\n return True\n else:\n return False\n\n def __getitem__(self, id):\n return self.feed_collection[id]\n\n def __len__(self):\n return len(self.feed_collection)\n\n def remove_old(self):\n i = 0\n collection_copy = []\n for _feed in self.feed_collection:\n if _feed.found == True:\n collection_copy.append(_feed)\n i += 1\n self.feed_collection = collection_copy\n\n def clear_old(self):\n for _feed in self.feed_collection:\n _feed.found = False\n _feed.new = False\n _feed.updated = False\n\n def get_unread_messages(self):\n unread = []\n for _feed in self.feed_collection:\n if _feed.new == True or _feed.updated == True:\n unread.append(_feed)\n return unread\n\nclass Greader:\n def __init__(self, user, password, source): \n self.google_url = 'http://www.google.com' \n self.reader_url = self.google_url + '/reader' \n self.login_url = 'https://www.google.com/accounts/ClientLogin' \n self.read_items_url = self.reader_url + '/api/0/unread-count'\n self.list_feeds_url = self.reader_url + '/api/0/subscription/list'\n self.source = source\n self.user = user\n self.password = password\n \n def login(self):\n #login / get SED \n header = {'User-agent' : self.source} \n post_data = urllib.urlencode({ 'Email': self.user, 'Passwd': self.password, 'service': 'reader', 'source': self.source, 'continue': self.google_url, }) \n request = urllib2.Request(self.login_url, post_data, header) \n \n try : \n f = urllib2.urlopen( request ) \n result = f.read() \n \n except: \n raise Exception('Error logging in')\n \n return re.search('Auth=(\\S*)', result).group(1) \n \n def get_results(self, auth, url):\n #get results from url \n header = {'User-agent' : self.source} \n header['Authorization']='GoogleLogin auth=%s' % auth \n \n request = urllib2.Request(url, None, header) \n \n try : \n f = urllib2.urlopen( request ) \n result = f.read() \n \n except: \n raise Exception('Error getting data from %s' % url) \n \n return result \n \n #get a feed of the users read items \n def get_unread_items(self, auth):\n feed_db = [] \n data = self.get_results(auth, self.read_items_url)\n feed_data = self.list_feeds(auth)\n node = ET.XML(data)\n feed_node = ET.XML(feed_data)\n \n total_unread = 0\n node = node.find(\"list\")\n feed_node = feed_node.find(\"list\")\n for o in node.findall(\"object\"):\n feed = \"\"\n total_unread = 0\n feed_title = \"\"\n for n in o.findall(\"string\"):\n if (n.attrib[\"name\"] == \"id\"):\n feed = n.text\n for n in o.findall(\"number\"):\n if (n.attrib[\"name\"] == \"count\"):\n total_unread = int(n.text)\n if feed[0:5] != \"user/\":\n for x in feed_node.findall(\"object\"):\n found = False\n for y in x.findall(\"string\"):\n if(y.attrib[\"name\"] == \"id\" and y.text == feed):\n found = True\n if(y.attrib[\"name\"] == \"title\" and found == True):\n feed_title = y.text\n if feed_title != \"\" and total_unread > 0:\n f = Feed(feed_title, total_unread)\n feed_db.append(f)\n return feed_db\n \n def list_feeds(self, auth):\n return self.get_results(auth, self.list_feeds_url)\n","repo_name":"cappert/specto","sub_path":"spectlib/plugins/watch_web_greader.py","file_name":"watch_web_greader.py","file_ext":"py","file_size_in_byte":9899,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"2670580781","text":"'''\nCode for HackerRank\nInterview Preparation Kit\nWesin Ribeiro\n\n###############\nproblem\n###############\nThere are a number of people queued up, and each person wears a sticker indicating their initial position in the queue\nInitial positions increment by 1 from 1 at the front of the line to n at the back\nAny person in the queue can bribe the person directly in front of them to swap positions\nIf two people swap positions, they still wear the same sticker denoting their original places in line.\nOne person can bribe at most two others.\nthe minimum number of bribes that took place to get the queue into its current state!\n\nIt must print an integer representing the minimum number of bribes necessary, \nor Too chaotic if the line configuration is not possible. \n\n###############\nObservations\n###############\n\n* initial state is the ascending sorted line \n* if the current position is larger than two, so it is chaotic\n* how to count the number of bribes?\n\nmy first approach is counting the number of bubbles in bubble sort algorithm,\nbut this solution doesn't work for large list of numbers,\neven if I break the counting when I realize that the list is chaotic.\n\nwho far is a element from initial position?\nif distance > 2, list is too chaotic\nelse, sum the distances.\n\nwe don't care if P has moved\nforwards, it is better to count how many times\nP has RECEIVED a bribe, by looking at who is\nahead of P. P's original position is the value\nof P.\n\nAnyone who bribed P cannot get to higher than\none position in front if P's original position,\nso we need to look from one position in front\nof P's original position to one in front of P's\ncurrent position, and see how many of those \npositions in Q contain a number large than P.\n\nIn other words we will look from P-1 to i-1,\nwhich in Python is range(P-1,i-1+1), or simply\nrange(P-1,i). To make sure we don't try an\nindex less than zero, replace P-1 with\nmax(P-1,0)\n\n'''\nimport math\nimport os\nimport random\nimport re\nimport sys\n\ndef minimumBribes(q):\n length = len(q)\n bribe = 0\n distance = 0\n for i,p in enumerate(q):\n print(i,p)\n distance = p - (i+1)\n if distance > 2:\n print('Too chaotic')\n return\n for j in range(max(p-2,0),i):\n if q[j] > p:\n bribe += 1\n \n print(bribe)\n\ndef main():\n t = 2\n\n for i in range(t): \n n = int(input())\n c = list(map(int, input().rstrip().split()))\n minimumBribes(c)\n \nmain()\n\n","repo_name":"wesinalves/codeskill","sub_path":"hackerrank/interview/new_year.py","file_name":"new_year.py","file_ext":"py","file_size_in_byte":2491,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"30305798131","text":"import json\nfrom builtins import eval\n\nfrom django.http import JsonResponse\n\nfrom .models import Degree\nfrom .models import DegreeRequirement\n\n\ndef advance_sem(year, sem):\n return (year, 2) if sem is 1 else (year + 1, 1)\n\n\ndef update_elective_code(c):\n if 'title' in c:\n # if (c['title'].lower() == \"elective course\"):\n # c['code'] = \"Elective Course\"\n if 'code' not in c:\n c['code'] = 'Elective Course'\n c.pop('title')\n return c\n\n\ndef generate_degree_plan(code, start_year_sem):\n degree = Degree.objects.filter(code=code)[0]\n reqs = degree.requirements\n if reqs == \"{}\":\n reqs = dict()\n final_year = 3\n if \"Honours\" in degree.name:\n final_year = 4\n for i in range(1, final_year + 1):\n for j in range(1, 3):\n reqs[str(i) + \".\" + str(j)] = [{\"title\": 'Elective Course'}]\n to_return = []\n year, sem = start_year_sem.split('S')\n year, sem = int(year), int(sem)\n for year_sem, courses in sorted(eval(str(reqs)).items(), key=lambda session: float(session[0])):\n for c in courses:\n update_elective_code(c)\n if c['code'] == \"OR\":\n c['code'] = 'Elective Course'\n if len(courses) < 4:\n for i in range(4 - len(courses)):\n courses.append({\"code\": 'Elective Course'})\n to_return.append({'{}S{}'.format(year, sem): courses[0:4]})\n year, sem = advance_sem(year, sem)\n return JsonResponse({\"response\": to_return})\n\ndef get_degree_requirements(_code):\n # code, year = _code.split(\"-\")[0], _code.split(\"-\")[1]\n # degree_requirement = DegreeRequirement.objects.filter(code=code, year=year)[0]\n # return degree_requirement._json()\n with open('static/json/{}.json'.format(_code)) as file:\n return file.read()\n","repo_name":"AICourseSelection/courseai","sub_path":"Code/courseai/degree/degree_plan_helper.py","file_name":"degree_plan_helper.py","file_ext":"py","file_size_in_byte":1844,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"94"} +{"seq_id":"4488689775","text":"import pandas as pd\nimport argparse\n\nif __name__ == \"__main__\":\n\tparser = argparse.ArgumentParser(description=\"find change in fitness values between treated and untreated\")\n\trequired = parser.add_argument_group('required')\n\trequired.add_argument('-t', '--csv_treated', nargs='*',required=True,\n\t\thelp='fitness values of treated condition')\n\trequired.add_argument('-u', '--csv_untreated', nargs='*',required=True,\n\t\thelp='fitness values of untreated condition')\n\trequired.add_argument('-n', '--name', nargs='*',required=True,\n\t\thelp='name of compound to label output csv file')\n\targs = parser.parse_args()\n\tct = pd.read_csv(args.csv_treated[0],index_col='Unnamed: 0')\n\tcu = pd.read_csv(args.csv_untreated[0],index_col='Unnamed: 0')\n\tmergedF = pd.merge(ct, cu, on='variant')\n\tmergedF['changeInFitness'] = mergedF['fitness_x'] - mergedF['fitness_y']\n\tmergedF.to_csv(args.name[0] + \"barcode_fitness_change.csv\")\n\n","repo_name":"stephaniewankowicz/PUBS","sub_path":"step5.py","file_name":"step5.py","file_ext":"py","file_size_in_byte":912,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"17222757169","text":"import os\nimport subprocess\n\ngit_root_dir = subprocess.check_output(\"git rev-parse --show-toplevel\".split(\" \")).decode(\"utf-8\").strip()\nsource_folder = os.path.abspath(os.path.join(git_root_dir, 'datasets'))\ndest_folder = '/var/lib/mysql-files/'\n\nfor file in ['PUMS.csv', 'PUMS_dup.csv', 'PUMS_large.csv', 'PUMS_pid.csv']:\n source_path = os.path.join(source_folder, file)\n dest_path = os.path.join(dest_folder, file)\n os.system('cp {} {}'.format(source_path, dest_path))\n\nsource_path = os.path.join(source_folder, 'PUMS_null.csv')\ndest_path = os.path.join(dest_folder, 'PUMS_null.csv')\n\nwith open(source_path, 'r') as f:\n lines = f.read().splitlines(keepends=False)\n with open(dest_path, 'w') as f2:\n for line in lines:\n line_parts = line.split(',')\n line_parts = [p if p != '' else \"\\\\N\" for p in line_parts]\n f2.write(','.join(line_parts))\n f2.write('\\n')\n\n","repo_name":"opendp/smartnoise-sdk","sub_path":"sql/tests/setup/mysql/copy.py","file_name":"copy.py","file_ext":"py","file_size_in_byte":925,"program_lang":"python","lang":"en","doc_type":"code","stars":220,"dataset":"github-code","pt":"94"} +{"seq_id":"17217621959","text":"import sys\r\nimport os\r\nimport struct\r\nimport math\r\nimport cv2\r\nimport numpy as np\r\nimport argparse\r\n\r\n\r\ndef getDVSeventsDavis(file, numEvents=1e10, startTime=0):\r\n \"\"\" DESCRIPTION: This function reads a given aedat file and converts it into four lists indicating \r\n timestamps, x-coordinates, y-coordinates and polarities of the event stream. \r\n \r\n Args:\r\n file: the path of the file to be read, including extension (str).\r\n numEvents: the maximum number of events allowed to be read (int, default value=1e10).\r\n startTime: the start event timestamp (in microseconds) where the conversion process begins (int, default value=0).\r\n\r\n Return:\r\n ts: list of timestamps in microseconds.\r\n x: list of x-coordinates in pixels.\r\n y: list of y-coordinates in pixels.\r\n pol: list of polarities (0: on -> off, 1: off -> on). \r\n \"\"\"\r\n print('\\ngetDVSeventsDavis function called \\n')\r\n sizeX = 346\r\n sizeY = 260\r\n x0 = 0\r\n y0 = 0\r\n x1 = sizeX\r\n y1 = sizeY\r\n \r\n print('Reading in at most', str(numEvents))\r\n \r\n\r\n triggerevent = int('400', 16)\r\n polmask = int('800', 16)\r\n xmask = int('003FF000', 16)\r\n ymask = int('7FC00000', 16)\r\n typemask = int('80000000', 16)\r\n typedvs = int('00', 16)\r\n xshift = 12\r\n yshift = 22\r\n polshift = 11\r\n x = []\r\n y = []\r\n ts = []\r\n pol = []\r\n numeventsread = 0\r\n \r\n length = 0\r\n aerdatafh = open(file, 'rb')\r\n k = 0\r\n p = 0\r\n statinfo = os.stat(file)\r\n if length == 0:\r\n length = statinfo.st_size\r\n print(\"file size\", length)\r\n\r\n lt = aerdatafh.readline()\r\n while lt and str(lt)[2] == \"#\":\r\n p += len(lt)\r\n k += 1\r\n lt = aerdatafh.readline()\r\n continue\r\n\r\n aerdatafh.seek(p)\r\n tmp = aerdatafh.read(8)\r\n p += 8\r\n while p < length:\r\n ad, tm = struct.unpack_from('>II', tmp)\r\n ad = abs(ad)\r\n if tm >= startTime:\r\n if (ad & typemask) == typedvs:\r\n xo = sizeX - 1 - float((ad & xmask) >> xshift)\r\n yo = float((ad & ymask) >> yshift)\r\n polo = 1 - float((ad & polmask) >> polshift)\r\n if xo >= x0 and xo < x1 and yo >= y0 and yo < y1:\r\n x.append(xo)\r\n y.append(yo)\r\n pol.append(polo)\r\n ts.append(tm)\r\n aerdatafh.seek(p)\r\n tmp = aerdatafh.read(8)\r\n p += 8\r\n numeventsread += 1\r\n\r\n print('Total number of events read =', numeventsread)\r\n print('Total number of DVS events returned =', len(ts))\r\n return ts, x, y, pol\r\n\r\ndef get_all_path(open_file_path):\r\n rootdir = open_file_path\r\n path_list = []\r\n list = os.listdir(rootdir) \r\n for i in range(0, len(list)):\r\n com_path = os.path.join(rootdir, list[i])\r\n if os.path.isfile(com_path):\r\n path_list.append(com_path)\r\n if os.path.isdir(com_path):\r\n path_list.extend(get_all_path(com_path))\r\n return path_list\r\n\r\n\r\ndef event_neighbor_filter(data=np.array([]), height=260, width=346, margin=1, threshold=1):\r\n img = np.zeros([height, width], dtype=np.int8)\r\n\r\n for idx in range(0, data.shape[0]):\r\n img[data[idx, 1], data[idx, 0]] = 1\r\n\r\n pos_tuple = np.where(img == 1)\r\n pos = np.array([pos_tuple[0], pos_tuple[1]]).T\r\n\r\n img_padding = np.zeros([height + 2 * margin, width + 2 * margin], dtype=np.int8)\r\n img_padding[margin:height + margin, margin:width + margin] = img\r\n\r\n for idx in range(0, pos.shape[0]):\r\n num_of_events = 0\r\n for i in range(-margin, margin + 1):\r\n for j in range(-margin, margin + 1):\r\n num_of_events += img_padding[pos[idx][0] + i][pos[idx][1] + j]\r\n img[pos[idx][0]][pos[idx][1]] = num_of_events > threshold\r\n\r\n data_filtered_tuple = np.where(img == 1)\r\n data_filtered = np.array([data_filtered_tuple[1], data_filtered_tuple[0]]).T\r\n\r\n return data_filtered\r\n \r\nif __name__ == '__main__':\r\n # parse the command line argument\r\n parser = argparse.ArgumentParser(description='SAE for encoding.')\r\n parser.add_argument('file_path', help='The .aedat file path.')\r\n args = parser.parse_args()\r\n all_path=get_all_path(args.file_path)\r\n \r\n for i in range(0, len(all_path)):\r\n inputfile = all_path[i]\r\n filepath,fullname = os.path.split(inputfile)\r\n name,ext = os.path.splitext(fullname)\r\n dirs = '/home/autodrive/EventSOT/'+name+'/img/'#save dirs\r\n if not os.path.exists(dirs):\r\n print('create dirs')\r\n os.makedirs(dirs)\r\n \r\n T, X, Y, Pol = getDVSeventsDavis(inputfile)#Read the quaternion array\r\n T = np.array(T).reshape((-1, 1))\r\n\r\n X = np.array(X).reshape((-1, 1))\r\n Y = np.array(Y).reshape((-1, 1))\r\n Pol = np.array(Pol).reshape((-1, 1))\r\n step_time = 2000 #The sliding time of a frame\r\n cumulative_time = 20000 #The cumulative time of a frame\r\n start_idx = 0\r\n end_idx = 0\r\n slid_idx = 0\r\n start_time = T[0]\r\n slid_time = start_time + step_time\r\n end_time = start_time + cumulative_time\r\n img_count = 1\r\n begin_number = 101 #the begin frame of the sequence\r\n \r\n filepath,fullname = os.path.split(inputfile)\r\n name,ext = os.path.splitext(fullname)\r\n \r\n \r\n while end_time <= T[-1]:\r\n \r\n while T[slid_idx] < slid_time:\r\n slid_idx = slid_idx + 1\r\n while T[end_idx] < end_time:\r\n end_idx = end_idx + 1\r\n\r\n data_x = np.array(X[start_idx:end_idx]).reshape((-1, 1))\r\n data_y = np.array(Y[start_idx:end_idx]).reshape((-1, 1))\r\n data_T = np.array(T[start_idx:end_idx]).reshape((-1, 1))\r\n data0 = np.column_stack((data_x, data_y)).astype(np.int32)\r\n data = event_neighbor_filter(data0, margin=1, threshold=1)\r\n \r\n timestamp=start_time*np.ones((260,346))\r\n \r\n for i in range(0, data.shape[0]):\r\n timestamp[data[i,1], data[i,0]]=data_T[i]\r\n if img_count >= begin_number:\r\n grayscale = np.flip(255*(timestamp-start_time)/step_time, 0).astype(np.uint8)#The normalization formula\r\n \r\n cv2.imshow('img',grayscale)\r\n \r\n cv2.waitKey(5)\r\n wfile = dirs +str(img_count).zfill(4) + '.png'\r\n cv2.imwrite(wfile,grayscale) \r\n \r\n slid_time += step_time\r\n start_time += step_time \r\n end_time += step_time\r\n start_idx = slid_idx\r\n img_count += 1","repo_name":"pipigenius/EventSOT","sub_path":"sae.py","file_name":"sae.py","file_ext":"py","file_size_in_byte":6740,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"26842433757","text":"# CLASS (aka type)\n# btw everything is an object, incl. strings numbers etc\n\ntype_of_asdf = type('asdf') # gets the class, of which the string 'asdf' is an instance\nprint(type_of_asdf(8))\nprint(type_of_asdf([1, 2, 3, 4]))\n# what did just happen??? Actually, 'str' is the class of any string - the well-known str function is ctuly class\n# so type_of_asdf = str\n\nclass Kotatko:\n species = 'ctuly iz a kat'\n\n # constructor, this one is with optional args\n def __init__(self, *args):\n self.jmeno = args[0] if args else 'bezejmenná kočka'\n\n def zamnoukej(self): # btw \"self\" is just a conventional name\n print(f\"Mňau já jsem {self.jmeno}, {self.species}\")\n\n def sezrat(self, zradlo):\n print(f\"{self.jmeno} sežralo {zradlo}\")\n\n # stringify function - is called whenever the object is treated as a string\n def __str__(self):\n return f\"{self.species} called {self.jmeno}\"\n\nmourek = Kotatko() # class is capitalized, instance is not capitalized\nmourek.zamnoukej()\nmourek.jmeno = 'Mourek'\nprint(f\"teď už {mourek.jmeno}\")\nprint('má jméno? ' + str(hasattr(mourek, 'jmeno')))\n# mourek.zamnoukej = 12345 # of course I could just overwrite the method with anything\n\nbashka = Kotatko('Bažka')\nbashka.sezrat('rybu') # first param of any method is always self, the other params are callable\nbashka.sezrat(zradlo='rybu') # it's better codestyle to explicitly assign the param.\n\nprint(bashka, str(bashka)) # the same thing\n\n# NOW LET'S CREATE A ŠTĚŇÁTKO, but to avoid boilerplate it will inherit\nclass Zviratko:\n asdf = 123\n def __init__(self, *args):\n self.jmeno = args[0] if args else 'bezejmenné '+self.species\n def sezrat(self, zradlo):\n print(f\"{self.jmeno} sežral/a {zradlo}\")\n\n# something = Zviratko() # Zviratko is not instantiable like this, it is missing species..\n\n# inheritance is easy like that. Zviratky is superclass, Stenatko is subclass\nclass Stenatko(Zviratko):\n species = '100% PES'\n\n \"\"\"\n BTW Stenatko DOES NOT HAVE __init__(), that's why Zviratko.init is called!\n But if had init, it'd override the Zviratko init, so I'd have to then explicitly call Zviratko init:\n def __init__():\n super().__init__()\n ...more_stuff()\n \"\"\"\n\n def zastekej(self):\n print(self.asdf)\n print(f\"Haf štěk já jsem {self.jmeno}\")\n\n # superclass already has \"sezrat\", so override it\n def sezrat(self, zradlo):\n print(f\"{self.jmeno} nejdříve {zradlo} očuchává\")\n # if we want to extend original fn, just call it now on the superclass, like this:\n super().sezrat(\"očuchané \"+zradlo)\n\nzeryk = Stenatko('Žeryk')\nprint(f'Žeryk je pes? {isinstance(zeryk, Stenatko)}')\nprint(f'Žeryk je zvíře? {isinstance(zeryk, Zviratko)}')\nprint(f'Žeryk je kočka? {isinstance(zeryk, Kotatko)}')\nzeryk.zastekej()\ndog = Stenatko()\ndog.sezrat('kost')\n\n# POLYMORPHISM - both Kotatko and Stenatko can be used there where we expect Zviratko\nzviratka = [Kotatko('Micka'), Stenatko('Azorek')]\nfor zviratko in zviratka:\n zviratko.sezrat('flákota')\n","repo_name":"Lemonexe/python-tutorial","sub_path":"base/17_classes.py","file_name":"17_classes.py","file_ext":"py","file_size_in_byte":3077,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"9432425711","text":"import sys\nimport math\nsys.stdin = open('input.txt')\n\nA, B = map(int, input().split())\n\nx = int((-2*A + math.sqrt(((2*A)**2) - 4 * B)) / 2)\ny = int((-2*A - math.sqrt(((2*A)**2) - 4 * B)) / 2)\n\nif x == y:\n print(x)\nelif x < y:\n print(x, y)\nelse:\n print(y, x)","repo_name":"lhynjn9/APS","sub_path":"22.완전탐색/Baekjoon/17945.py","file_name":"17945.py","file_ext":"py","file_size_in_byte":266,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"94"} +{"seq_id":"17996352717","text":"from office365.runtime.queries.create_entity import CreateEntityQuery\nfrom office365.runtime.queries.service_operation import ServiceOperationQuery\nfrom office365.runtime.paths.service_operation import ServiceOperationPath\nfrom office365.sharepoint.base_entity_collection import BaseEntityCollection\nfrom office365.sharepoint.fields.field import Field\nfrom office365.sharepoint.fields.creation_information import FieldCreationInformation\nfrom office365.sharepoint.fields.type import FieldType\nfrom office365.sharepoint.fields.xmlSchemaFieldCreationInformation import XmlSchemaFieldCreationInformation\nfrom office365.sharepoint.taxonomy.field import TaxonomyField\nfrom office365.sharepoint.taxonomy.sets.set import TermSet\n\n\nclass FieldCollection(BaseEntityCollection):\n \"\"\"Represents a collection of Field resource.\"\"\"\n\n def __init__(self, context, resource_path=None, parent=None):\n super(FieldCollection, self).__init__(context, Field, resource_path, parent)\n\n def add_geolocation_field(self, title, description=None):\n \"\"\"\n Creates Geolocation field\n\n :param str title: Specifies the display name of the field\n :param str or None description: Specifies the description of the field\n \"\"\"\n return self.add(FieldCreationInformation(title=title,\n description=description,\n field_type_kind=FieldType.Geolocation))\n\n def add_url_field(self, title, description=None):\n \"\"\"\n Creates Url field\n\n :param str title:\n :param str or None description:\n \"\"\"\n return self.add(FieldCreationInformation(title=title,\n description=description,\n field_type_kind=FieldType.URL))\n\n def add_lookup_field(self, title, lookup_list, lookup_field_name, allow_multiple_values=False):\n \"\"\"\n Creates a Lookup field\n\n :param bool allow_multiple_values: Flag determines whether to create multi lookup field or not\n :param str lookup_field_name: Specifies the name of the field in the other data source when creating\n a lookup field\n :param str or office365.sharepoint.lists.list.List lookup_list: Lookup List object or identifier\n :param str title: Specifies the display name of the field.\n \"\"\"\n return_type = Field(self.context)\n\n def _add_lookup_field(lookup_list_id):\n \"\"\"\n :type lookup_list_id: str\n \"\"\"\n if allow_multiple_values:\n field_schema = '''\n \n '''.format(title=title, lookup_field_name=lookup_field_name, lookup_list_id=lookup_list_id)\n self.create_field_as_xml(field_schema, return_type=return_type)\n else:\n self.add_field(FieldCreationInformation(title=title,\n lookup_list_id=lookup_list_id,\n lookup_field_name=lookup_field_name,\n field_type_kind=FieldType.Lookup),\n return_type=return_type)\n\n from office365.sharepoint.lists.list import List\n if isinstance(lookup_list, List):\n\n def _lookup_list_loaded():\n _add_lookup_field(lookup_list.id)\n lookup_list.ensure_property(\"Id\", _lookup_list_loaded)\n else:\n _add_lookup_field(lookup_list)\n return return_type\n\n def add_choice_field(self, title, values, multiple_values=False):\n \"\"\"\n Creates a Choice field\n\n :param bool multiple_values:\n :param list[str] values:\n :param str title: Specifies the display name of the field.\n \"\"\"\n fld_type = FieldType.MultiChoice if multiple_values else FieldType.Choice\n create_field_info = FieldCreationInformation(title, fld_type)\n [create_field_info.Choices.add(choice) for choice in values]\n return self.add_field(create_field_info)\n\n def add_user_field(self, title):\n \"\"\"\n Creates a User field\n\n :param str title: Specifies the display name of the field\n \"\"\"\n return self.add_field(FieldCreationInformation(title, FieldType.User))\n\n def add_text_field(self, title):\n \"\"\"\n Creates a Text field\n\n :param str title: specifies the display name of the field\n :rtype: office365.sharepoint.fields.text.FieldText\n \"\"\"\n return self.add_field(FieldCreationInformation(title, FieldType.Text))\n\n def add_dependent_lookup_field(self, display_name, primary_lookup_field_id, show_field):\n \"\"\"Adds a secondary lookup field to a field collection (target). \n Args:\n display_name (str): title of the added field in the target FieldCollection.\n primary_lookup_field_id (str): ID of the main lookup-field in the target to associate the dependent lookup field with.\n show_field (str): name of the field from the source list to include data from.\n Returns:\n Field: reference to the SP.Field that was added.\n \"\"\"\n return_type = Field(self.context)\n self.add_child(return_type)\n parameters = {\n \"displayName\": display_name,\n \"primaryLookupFieldId\": primary_lookup_field_id,\n \"showField\": show_field\n }\n qry = ServiceOperationQuery(self, \"AddDependentLookupField\", None, parameters,\n None, return_type)\n self.context.add_query(qry)\n return return_type\n\n def add(self, field_create_information):\n \"\"\"Adds a fields to the fields collection.\n\n :type field_create_information: office365.sharepoint.fields.creation_information.FieldCreationInformation\n \"\"\"\n return_type = Field.create_field_from_type(self.context, field_create_information)\n self.add_child(return_type)\n qry = CreateEntityQuery(self, return_type, return_type)\n self.context.add_query(qry)\n return return_type\n\n def add_field(self, parameters, return_type=None):\n \"\"\"Adds a fields to the fields collection.\n\n :type parameters: office365.sharepoint.fields.creation_information.FieldCreationInformation\n :param Field or None return_type: Return type\n \"\"\"\n if return_type is None:\n return_type = Field(self.context)\n self.add_child(return_type)\n payload = {\"parameters\": parameters}\n qry = ServiceOperationQuery(self, \"AddField\", None, payload, None,\n return_type)\n self.context.add_query(qry)\n return return_type\n\n def create_taxonomy_field(self, name, term_set, allow_multiple_values=False):\n \"\"\"\n Creates a Taxonomy field\n\n :param str name: Field name\n :param str or TermSet term_set: TermSet identifier or object\n :param bool allow_multiple_values: Specifies whether the column will allow more than one value\n \"\"\"\n return_type = TaxonomyField(self.context)\n\n if isinstance(term_set, TermSet):\n def _term_set_loaded():\n TaxonomyField.create(self, name, term_set.id, None, allow_multiple_values,\n return_type=return_type)\n\n term_set.ensure_property(\"id\", _term_set_loaded)\n return return_type\n else:\n\n def _term_store_loaded(term_store):\n TaxonomyField.create(self, name, term_set, term_store.id, allow_multiple_values,\n return_type=return_type)\n\n self.context.load(self.context.taxonomy.term_store, after_loaded=_term_store_loaded)\n return return_type\n\n def create_field_as_xml(self, schema_xml, return_type=None):\n \"\"\"\n Creates a field based on the values defined in the parameters input parameter.\n\n :param str schema_xml: Specifies the schema that defines the field\n :param Field or None return_type: Return type\n \"\"\"\n if return_type is None:\n return_type = Field(self.context)\n self.add_child(return_type)\n payload = {\"parameters\": XmlSchemaFieldCreationInformation(schema_xml)}\n qry = ServiceOperationQuery(self, \"CreateFieldAsXml\", None, payload, None, return_type)\n self.context.add_query(qry)\n return return_type\n\n def get_by_id(self, _id):\n \"\"\"\n Gets the fields with the specified ID.\n\n :param str _id: The field identifier.\n \"\"\"\n return Field(self.context, ServiceOperationPath(\"getById\", [_id], self.resource_path))\n\n def get_by_internal_name_or_title(self, value):\n \"\"\"Returns the first field (2) in the collection based on the internal name or the title specified\n by the parameter.\n\n :param str value: The title or internal name to look up the field (2) by.\n \"\"\"\n return Field(self.context,\n ServiceOperationPath(\"getByInternalNameOrTitle\", [value], self.resource_path))\n\n def get_by_title(self, title):\n \"\"\"\n Returns the first fields object in the collection based on the title of the specified fields.\n\n :param str title: The title to look up the field by\n \"\"\"\n return Field(self.context, ServiceOperationPath(\"getByTitle\", [title], self.resource_path))\n","repo_name":"vgrem/Office365-REST-Python-Client","sub_path":"office365/sharepoint/fields/collection.py","file_name":"collection.py","file_ext":"py","file_size_in_byte":9810,"program_lang":"python","lang":"en","doc_type":"code","stars":1050,"dataset":"github-code","pt":"96"} +{"seq_id":"70045320635","text":"\"\"\"\n定义:使用了yield的函数被称为生成器\n跟普通函数不同的是,生成器是一个返回迭代器的函数,只能用于迭代操作,更简单点理解生成器就是一个迭代器。\n\"\"\"\n\n\ndef fibonacci(n):\n a, b, counter = 0, 1, 0\n while True:\n if (counter > n):\n return\n yield a\n a, b = b, a + b\n counter += 1\n\n\nfor i in fibonacci(10):\n print(i)\n","repo_name":"newpi6/python-demo","sub_path":"class_demo/generator_demo.py","file_name":"generator_demo.py","file_ext":"py","file_size_in_byte":422,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"42689995883","text":"import os\nimport sys\nfrom tqdm import tqdm\nimport numpy as np\n\nCSD = os.path.dirname(__file__)\nBDR = os.path.dirname(CSD)\nsys.path.insert(0, BDR)\nimport mpi_cnn\n\nBATCH_SIZE = 64\nPRINT_MODEL_SUMMARY = False\nE_MIN, E_MAX = 1.0e09, 500.0e09\nNU = NU_MIN = NU_MAX = 0.25 # because we have assumed equal P_ratio\n\n\nDATA_SETS_DIR = os.path.join(BDR, \"0_datasets\", \"containers\", \"xy\")\nMODELS_DIR = os.path.join(CSD, \"trained_models\")\nRESULTS_DIR = os.path.join(CSD, \"Test-Results\")\nos.makedirs(RESULTS_DIR, exist_ok=True)\n\n#\nRVE_MIA = mpi_cnn.load_data_set(\n os.path.join(DATA_SETS_DIR, f\"rve_testing_images.npy\"),\n permute_dims=(0, 3, 1, 2)\n)\nRVE_META = mpi_cnn.load_data_set(\n os.path.join(DATA_SETS_DIR, f\"rve_testing_labels.npy\")\n)\n#\ntest_results = {}\nfor a_model_name in tqdm(os.listdir(MODELS_DIR)): # NEX_500-R_1\n nx_id, rl_id = a_model_name.split(\"-\")\n nex_tr = int(nx_id.split(\"_\")[1])\n real_num = int(rl_id.split(\"_\")[1])\n #\n nex_ts = nex_tr // 2\n anx_mia = RVE_MIA[:nex_ts, 0:1]\n anx_meta = RVE_META[:nex_ts]\n true_prop = anx_meta[:, 5:8]\n vf = anx_meta[:, 0:1]\n em = anx_meta[:, 1:2]\n ef = anx_meta[:, 3:4]\n cnn_model = mpi_cnn.load_model(\n os.path.join(MODELS_DIR, a_model_name, \"best_model.pth\"),\n inc=2, # number of input channels\n ouu=3, # number of output units\n els_mod_min=E_MIN,\n els_mod_max=E_MAX,\n p_ratio=NU,\n print_summary=False,\n )\n test_dl = mpi_cnn.make_data_loaders(\n (anx_mia,true_prop, vf, em, ef), BATCH_SIZE, nex_ts, shuffle=False\n )\n predictions = cnn_model.predictions_on(test_dl)\n test_results[a_model_name] = mpi_cnn.torch.concatenate(\n (anx_meta[:, :8], predictions), dim=1\n ).cpu().numpy()\n del cnn_model\n\nnp.savez_compressed(\n os.path.join(RESULTS_DIR, \"test_results.npz\"),\n **test_results,\n)\n","repo_name":"338rajesh/mpi-cnn","sub_path":"3_test_native_domain/1_test_models.py","file_name":"1_test_models.py","file_ext":"py","file_size_in_byte":1862,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"37167353304","text":"from collections import deque\n\n\nclass Solution:\n def maxSlidingWindow(self, nums, k):\n \"\"\"\n :type nums: List[int]\n :type k: int\n :rtype: List[int]\n \"\"\"\n q = deque()\n ret = []\n for i, num in enumerate(nums):\n # ensure the 1st element in queue is the largest\n while q and nums[q[-1]] < num:\n q.pop()\n q.append(i)\n if q[0] == i - k:\n q.popleft()\n if i >= k - 1:\n ret.append(nums[q[0]])\n return ret\n","repo_name":"bgoonz/DS-ALGO-OFFICIAL","sub_path":"CONTENT/DS-n-Algos/ALGO/_LEETCODE/__UNSORTED/239.py","file_name":"239.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"96"} +{"seq_id":"22861446362","text":"#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n# @Date : 2020/8/2\n# @Author : Bruce Liu /Lin Luo\n# @Mail : 15869300264@163.com\nfrom creational.singleton.example import Nation, test\n\n\nclass SingletonProxy(object):\n \"\"\"\n 类装饰器\n 这里使用了python的修饰器方法,但实际上这个属于代理模式,所以使用proxy命名\n \"\"\"\n\n def __init__(self, cls):\n \"\"\"\n 类装饰器初始化函数\n :param cls: 需要实例化的类\n \"\"\"\n self._cls = cls\n self._instances = {}\n\n def __call__(self, *args, **kwargs):\n \"\"\"\n 类装饰器需要实现__call__方法,用于自身的调用\n :param args:\n :param kwargs:\n :return:\n \"\"\"\n if self._cls in self._instances:\n print('return instance from class proxy attribute')\n return self._instances[self._cls]\n else:\n self._instances[self._cls] = self._cls(*args, **kwargs)\n print('return instance by created')\n return self._instances[self._cls]\n\n\n@SingletonProxy\n# 使用修饰器,在执行类的实例化时实现单例\nclass China(Nation):\n def __init__(self, name: str = '中国'):\n super().__init__(name)\n # 使用修饰器会由于修饰器的调用,China会变成装饰器对象而不是原来的China类,导致调用失败\n # 解决方式为在修饰器的代码中植入一段内容,将被修饰的类的属性赋予修饰器对象\n # super(China, self).__init__(name)\n\n\nif __name__ == '__main__':\n print('using class proxy to create singleton')\n test('Bruce', 'Clerk', China)\n","repo_name":"BruceWW/python_design_pattern","sub_path":"creational/singleton/example/class_proxy.py","file_name":"class_proxy.py","file_ext":"py","file_size_in_byte":1650,"program_lang":"python","lang":"zh","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"11903042778","text":"import wx\n\nclass MyFrame(wx.Frame):\n def __init__(self, parent, title):\n super(MyFrame, self).__init__(parent, title=title, size = (1200,800))\n self.panel = MyPanel(self)\n\n \n\nclass MyPanel(wx.Panel):\n def __init__(self, parent):\n super(MyPanel, self).__init__(parent)\n \n # \"Input\"-\"Output\"\n self.titleInput = wx.StaticText(self, label = \"Input:\", pos = (10,5)) \n self.titleOutput = wx.StaticText(self, label = \"Output:\", pos = (330,5)) \n font = wx.Font(10, wx.DEFAULT, wx.NORMAL, wx.BOLD)\n self.titleInput.SetFont(font)\n self.titleOutput.SetFont(font)\n \n # temperature input\n self.T = wx.StaticText(self, label = \"T\", pos = (10,33)) \n self.T_input = wx.TextCtrl(self, size=(50,-1), pos = (25,30))\n self.T_unit = wx.StaticText(self, label=\"°C\", pos = (80,33))\n self.T_input.SetValue('900')\n \n # component combobox (Cl, X, Y)\n self.component = wx.StaticText(self, label = \"Component\", pos = (120, 33)) \n self.component_choices = ['Cl', 'X', 'Y']\n self.component_combobox = wx.ComboBox(self, choices = self.component_choices, pos = (195, 30))\n #grey = wx.Font(10, wx.DEFAULT, wx.NORMAL, wx)\n self.component_combobox.SetValue('Cl')\n \n # initial C\n self.InitialC = wx.StaticText(self, label = \"Initial C =\", pos = (10, 66)) \n self.InitialC_input = wx.TextCtrl(self, size=(50,-1), pos = (110, 63))\n self.wt_unit1 = wx.StaticText(self, label = \"wt.%\", pos = (165, 66))\n \n # left boundary\n self.LeftBoundary = wx.StaticText(self, label = \"Left Boundary =\", pos = (10, 99)) \n self.left_boundary_input = wx.TextCtrl(self, size=(50,-1), pos = (110, 96))\n self.wt_unit2 = wx.StaticText(self, label = \"wt.%\", pos = (165, 99))\n \n # right boundary\n self.RightBoundary = wx.StaticText(self, label = \"Right Boundary =\", pos = (10, 132)) \n self.right_boundary_input = wx.TextCtrl(self, size=(50,-1), pos = (110, 129))\n self.wt_unit3 = wx.StaticText(self, label = \"wt.%\", pos = (165, 132))\n \n # natural data upload button\n self.natural_data = wx.StaticText(self, label = \"Natural data:\", pos = (10,165)) \n self.naturaldata_button = wx.Button(self, label=\"Upload\", pos=(100,162))\n \n ###### Plot fig. 1 button #####\n self.plot_button = wx.Button(self, label = \"Plot\", pos = (100, 195)) \n self.plot_button.Bind(wx.EVT_BUTTON, self.OnClickPlot) \n \n #dx (distance step)\n self.dx = wx.StaticText(self, label = \"dx\", pos = (10,240)) \n self.dx_input = wx.TextCtrl(self, size=(50,-1), pos = (30,237))\n self.dx_unit = wx.StaticText(self, label=\"µm\", pos = (85,240))\n \n #dt (time step)\n self.dt = wx.StaticText(self, label = \"dt\", pos = (140,240)) \n self.dt_input = wx.TextCtrl(self, size=(50,-1), pos = (160,237))\n self.dt_unit = wx.StaticText(self, label=\"s\", pos = (215,240))\n \n # interation input\n self.iteration = wx.StaticText(self, label = \"Iteration\", pos = (10,273)) \n self.iteration_input = wx.TextCtrl(self, size=(50,-1), pos = (60,270))\n \n ###### Run modeling and plot figure 2 button #####\n self.Run_button = wx.Button(self, label = \"Run!\", pos = (150, 269))\n self.Run_button.Bind(wx.EVT_BUTTON, self.OnClickRun)\n \n \n \n ###### TEST BUTTON #####\n self.test_button = wx.Button(self, label = \"test button\", pos = (10, 400)) #####\n self.test_button.Bind(wx.EVT_BUTTON, self.OnClickTest) #####\n \n\n def OnClickPlot(self, event):\n # Display figure 1\n self.single_ele_diff_pic = wx.StaticBitmap(self, size=(500,500), pos = (330,25))\n self.single_ele_diff_pic.SetBitmap(wx.Bitmap('images/single_ele_diffusion.png'))\n\n def OnClickRun(self, event):\n # Display figure 2\n self.figure_pic = wx.StaticBitmap(self, size=(500,500), pos = (330,25))\n self.figure_pic.SetBitmap(wx.Bitmap('images/figure.png'))\n \n def OnClickTest(self, event): #####\n Ti = self.T_input.GetValue() #####\n self.print_testvalue = wx.StaticText(self, label = Ti, pos = (10, 420)) #####\n\n\nclass MyApp(wx.App):\n def OnInit(self):\n self.frame = MyFrame(parent=None, title=\"ApTimer\")\n self.frame.Show()\n return True\n \n \napp = MyApp()\napp.MainLoop()\n\n\n","repo_name":"taquynhnga2001/ApTimer-single-element","sub_path":"test_interface.py","file_name":"test_interface.py","file_ext":"py","file_size_in_byte":4888,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"23875875007","text":"# importing pygame library and player file to integrate\nimport pygame\nimport player\nimport world\nimport decorations\nimport coins\nimport obstacles\nimport health\n\n\n# creating class for game state\nclass GameState:\n # initialising variables\n def __init__(self, window_surface):\n self.transition_target = None\n self.window_surface = window_surface\n\n self.background_surf = None\n self.background_image = pygame.image.load('img/bg.png')\n # scaling background image down to fit where needed\n self.background_image = pygame.transform.scale(self.background_image, (1200, 640))\n self.width = self.background_image.get_width()\n\n # creating instances\n self.player = player.Character(50, 400, 7.0, 0.15, 80, 67, 0, 9)\n self.world = world.my_world\n\n # health bar instance\n self.health = health.HealthBar(20, 60, self.player.health, self.player.max_health)\n\n self.score_value = 100\n self.score_font = pygame.font.SysFont('Roboto', 35)\n self.score_x = 20\n self.score_y = 20\n\n # collision variable\n self.get_hit = False\n\n # function for what happens as soon as game state is called through the main menu\n def start(self):\n self.transition_target = None\n self.background_surf = pygame.Surface((800, 640))\n\n # function for killing screen when state is no longer in use\n def stop(self):\n self.background_surf = None\n\n # function for handling user events (back to main menu when user presses escape)\n def handle_events(self, event):\n if event.type == pygame.KEYDOWN and event.key == pygame.K_ESCAPE:\n self.transition_target = 'main_menu'\n\n # keyboard input (on press)\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_a: # initiates move\n self.player.move_left = True\n if event.key == pygame.K_d:\n self.player.move_right = True\n if event.key == pygame.K_LEFT: # allows variance in controls\n self.player.move_left = True\n if event.key == pygame.K_RIGHT:\n self.player.move_right = True\n if event.key == pygame.K_w:\n self.player.jump = True\n if event.key == pygame.K_UP:\n self.player.jump = True\n\n # keyboard input (on release)\n if event.type == pygame.KEYUP:\n if event.key == pygame.K_a: # stops movement on release\n self.player.move_left = False\n if event.key == pygame.K_d:\n self.player.move_right = False\n if event.key == pygame.K_LEFT:\n self.player.move_left = False\n if event.key == pygame.K_RIGHT:\n self.player.move_right = False\n\n # update function for drawing elements onto screen\n def update(self, time_delta):\n # calling the move function on the player\n player.screen_scroll = self.player.move()\n\n self.window_surface.blit(self.background_surf, (0, 0))\n # updating background image\n for x in range(4):\n # repeating background and parallax scrolling\n self.window_surface.blit(self.background_image, ((x * self.width) - player.bg_scroll * 0.75, 0))\n # drawing world\n self.world.draw(self.window_surface)\n player.bg_scroll -= player.screen_scroll\n\n # check for collision with coin - remove coin once collected\n if pygame.sprite.spritecollide(self.player, coins.coin_group, True):\n self.score_value += 15\n\n # check for collision with obstacle\n if pygame.sprite.spritecollide(self.player, obstacles.obstacle_group, False) and not self.get_hit:\n self.get_hit = True\n self.player.health -= 15\n if self.score_value >= 0:\n self.score_value -= 15\n if self.score_value <= 0:\n self.score_value = 0\n if not pygame.sprite.spritecollide(self.player, obstacles.obstacle_group, False):\n self.get_hit = False\n\n # drawing and updating decoration tiles\n decorations.decoration_group.draw(self.window_surface)\n coins.coin_group.draw(self.window_surface)\n obstacles.obstacle_group.draw(self.window_surface)\n world.exit_group.draw(self.window_surface)\n\n # displaying score\n score = self.score_font.render(\"SCORE : \" + str(self.score_value), True, (198, 90, 0))\n score_text_pos = self.score_x, self.score_y\n self.window_surface.blit(score, score_text_pos)\n\n # display health bar\n self.health.draw(self.player.health, self.window_surface)\n\n # update sprite groups\n decorations.decoration_group.update()\n coins.coin_group.update()\n obstacles.obstacle_group.update()\n world.exit_group.update()\n\n # draw method imported from player + animation\n self.player.draw(self.window_surface)\n self.player.update()\n self.player.update_anim(time_delta)\n\n pygame.display.flip()\n","repo_name":"priyap2166/ground_breaker.3","sub_path":"game_state.py","file_name":"game_state.py","file_ext":"py","file_size_in_byte":5076,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"33310514697","text":"#!/usr/bin/env python\n\nimport rospy \nimport numpy as np\nimport cv2\nimport struct\n\nfrom vision_nodes.utils import hello\n\nfrom std_msgs.msg import Header\nfrom sensor_msgs import point_cloud2\nfrom sensor_msgs.msg import PointCloud2, PointField\n\ndef main():\n\n rospy.init_node(\"py_custom\")\n\n node_path = rospy.get_param(\"/p_cloud/node_path\", \"~/media\")\n print(\"Ruta del paquete\" + node_path)\n\n img_pub = cv2.imread(node_path + \"/media/im18.png\")\n depth_map = cv2. imread(node_path + \"/media/pr18.png\", cv2.IMREAD_UNCHANGED)\n\n print(depth_map)\n\n pc_pub = rospy.Publisher(\"/point_cloud_py\", PointCloud2, queue_size=1)\n\n K = np.array([[379.2713, 0.0, 319.348],\n [0.0, 379.271, 238.448],\n [0.0, 0.0, 1.0]])\n\n K_inv = np.linalg.inv(K)\n \n #Matriz de rotacion image -> Rviz\n theta = -np.pi/2\n R_c_rviz = np.array([[1.0, 0.0, 0.0], [0.0, np.cos(theta), -np.sin(theta)], [0.0, np.sin(theta), np.cos(theta)]])\n\n #Inicializar lista de puntos de 3D\n points_3d = []\n\n #Calcular nube de puntos\n for i in range(img_pub.shape[0]):\n for j in range(img_pub.shape[1]):\n #Crear vectors de puntos en 2D\n p_2d = np.array([[j], [i], [1]])\n\n #Calcular punto de 3D\n z = 0.001*depth_map[i, j] #Profundidad en milimetros\n p_3d = z*np.dot(K_inv, p_2d)\n p_3d = np.dot(R_c_rviz, p_3d)\n \n #Extraer color de la imagen (imagen BGR)\n b = img_pub[i, j, 0]\n g = img_pub[i, j, 1]\n r = img_pub[i, j, 2]\n a = 255\n \n rgb = struct.unpack(\"I\", struct.pack('BBBB', b, g, r, a))[0]\n pt = [p_3d[0,0], p_3d[1,0], p_3d[2,0], rgb]\n points_3d.append(pt)\n\n #Crear cabecera del mensaje\n header = Header()\n header.frame_id = \"world\"\n\n #Crear campos\n fields = [PointField('x', 0, PointField.FLOAT32, 1),\n PointField('y', 4, PointField.FLOAT32, 1),\n PointField('z', 8, PointField.FLOAT32, 1),\n PointField('rgba', 12, PointField.UINT32, 1)] #FLOAT32\n\n #Crear nube de puntos\n pc_msg = point_cloud2.create_cloud(header, fields, points_3d)\n\n\n rate = rospy.Rate(15)\n\n while not rospy.is_shutdown():\n\n pc_msg.header.stamp = rospy.Time.now()\n pc_pub.publish(pc_msg)\n\n\n rate.sleep()\n\nif __name__ =='__main__':\n main()","repo_name":"JosueRodmoTec/Vision_Ros","sub_path":"ros_ws_copy/src/vision_nodes/src/point_cloud.py","file_name":"point_cloud.py","file_ext":"py","file_size_in_byte":2407,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"4464619811","text":"# -*- coding: utf-8 -*-\r\n\r\n\"\"\"\r\nMRE.py\r\n***************************************************************************\r\n* *\r\n* This program is free software; you can redistribute it and/or modify *\r\n* it under the terms of the GNU General Public License as published by *\r\n* the Free Software Foundation; either version 2 of the License, or *\r\n* (at your option) any later version. *\r\n* *\r\n***************************************************************************\r\n\"\"\"\r\n__author__ = 'Leandro França'\r\n__date__ = '2022-11-30'\r\n__copyright__ = '(C) 2022, Leandro França'\r\n\r\nfrom PyQt5.QtCore import *\r\nfrom qgis.core import *\r\nfrom numpy import sqrt, array, mean, std\r\n\r\n\r\nclass MRE(QgsProcessingAlgorithm):\r\n\r\n REF = 'REF'\r\n TESTE = 'TESTE'\r\n OUTPUT = 'OUTPUT'\r\n HTML = 'HTML'\r\n\r\n def tr(self, string):\r\n return QCoreApplication.translate('Processing', string)\r\n\r\n def createInstance(self):\r\n return MRE()\r\n\r\n def name(self):\r\n return 'mre'\r\n\r\n def displayName(self):\r\n return self.tr('MRE - Retângulos Equivalentes')\r\n\r\n def group(self):\r\n return self.tr('Feições Lineares')\r\n\r\n def groupId(self):\r\n return 'feicoeslineares'\r\n\r\n def shortHelpString(self):\r\n return self.tr('''Esta ferramenta calcula as discrepâncias entre feições lineares pelo Método dos Retângulos Equivalentes (MRE).\r\nDas discrepâncias planimétricas é determinado a classificação do Padrão de Exatidão Cartográfica para Produtos Cartográficos Digitais (PEC-PCD).\r\nAutor: Leandro França - Eng. Cartógrafo''')\r\n\r\n def initAlgorithm(self, config=None):\r\n self.addParameter(\r\n QgsProcessingParameterFeatureSource(\r\n self.REF,\r\n self.tr('Linhas de referência'),\r\n [QgsProcessing.TypeVectorLine]\r\n )\r\n )\r\n \r\n self.addParameter(\r\n QgsProcessingParameterFeatureSource(\r\n self.TESTE,\r\n self.tr('Polígonos das discrepâncias'),\r\n [QgsProcessing.TypeVectorPolygon]\r\n )\r\n )\r\n \r\n self.addParameter(\r\n QgsProcessingParameterFeatureSink(\r\n self.OUTPUT,\r\n self.tr('Distâncias Equivalentes')\r\n )\r\n )\r\n \r\n self.addParameter(\r\n QgsProcessingParameterFileDestination(\r\n 'HTML',\r\n 'Relatório do PEC-PCD',\r\n self.tr('arquivo HTML (*.html)')\r\n )\r\n )\r\n \r\n def str2HTML(self, texto):\r\n if texto:\r\n dicHTML = {'Á': 'Á',\t'á': 'á',\t'Â': 'Â',\t'â': 'â',\t'À': 'À',\t'à': 'à',\t'Å': 'Å',\t'å': 'å',\t'Ã': 'Ã',\t'ã': 'ã',\t'Ä': 'Ä',\t'ä': 'ä',\t'Æ': 'Æ',\t'æ': 'æ',\t'É': 'É',\t'é': 'é',\t'Ê': 'Ê',\t'ê': 'ê',\t'È': 'È',\t'è': 'è',\t'Ë': 'Ë',\t'ë': 'Ë',\t'Ð': 'Ð',\t'ð': 'ð',\t'Í': 'Í',\t'í': 'í',\t'Î': 'Î',\t'î': 'î',\t'Ì': 'Ì',\t'ì': 'ì',\t'Ï': 'Ï',\t'ï': 'ï',\t'Ó': 'Ó',\t'ó': 'ó',\t'Ô': 'Ô',\t'ô': 'ô',\t'Ò': 'Ò',\t'ò': 'ò',\t'Ø': 'Ø',\t'ø': 'ø',\t'Ù': 'Ù',\t'ù': 'ù',\t'Ü': 'Ü',\t'ü': 'ü',\t'Ç': 'Ç',\t'ç': 'ç',\t'Ñ': 'Ñ',\t'ñ': 'ñ',\t'Ý': 'Ý',\t'ý': 'ý',\t'\"': '"', '”': '"',\t'<': '<',\t'>': '>',\t'®': '®',\t'©': '©',\t'\\'': ''', 'ª': 'ª', 'º': 'º', '°':'°'}\r\n for item in dicHTML:\r\n if item in texto:\r\n texto = texto.replace(item, dicHTML[item])\r\n return texto\r\n else:\r\n return ''\r\n\r\n def processAlgorithm(self, parameters, context, feedback):\r\n \r\n ref = self.parameterAsSource(\r\n parameters,\r\n self.REF,\r\n context\r\n )\r\n \r\n if ref is None:\r\n raise QgsProcessingException(self.invalidSourceError(parameters, self.REF))\r\n \r\n teste = self.parameterAsSource(\r\n parameters,\r\n self.TESTE,\r\n context\r\n )\r\n\r\n if teste is None:\r\n raise QgsProcessingException(self.invalidSourceError(parameters, self.TESTE))\r\n \r\n \r\n itens = {\r\n 'discrep' : QVariant.Double,\r\n 'peso': QVariant.Double,\r\n }\r\n \r\n Fields = QgsFields()\r\n \r\n for item in itens:\r\n Fields.append(QgsField(item, itens[item]))\r\n \r\n (sink, dest_id) = self.parameterAsSink(\r\n parameters,\r\n self.OUTPUT,\r\n context,\r\n Fields,\r\n QgsWkbTypes.Polygon,\r\n ref.sourceCrs()\r\n )\r\n if sink is None:\r\n raise QgsProcessingException(self.invalidSinkError(parameters, self.OUTPUT))\r\n \r\n html_output = self.parameterAsFileOutput(\r\n parameters, \r\n self.HTML, \r\n context\r\n )\r\n\r\n \r\n PEC = { '0.5k': {'planim': {'A': {'EM': 0.14, 'EP': 0.085},'B': {'EM': 0.25, 'EP': 0.15},'C': {'EM': 0.4, 'EP': 0.25},'D': {'EM': 0.5, 'EP': 0.3}}, 'altim': {'A': {'EM': 0.135, 'EP': 0.085},'B': {'EM': 0.25, 'EP': 0.165},'C': {'EM': 0.3, 'EP': 0.2},'D': {'EM': 0.375, 'EP': 0.25}}},\r\n '1k': {'planim': {'A': {'EM': 0.28, 'EP': 0.17},'B': {'EM': 0.5, 'EP': 0.3},'C': {'EM': 0.8, 'EP': 0.5},'D': {'EM': 1, 'EP': 0.6}}, 'altim': {'A': {'EM': 0.27, 'EP': 0.17},'B': {'EM': 0.5, 'EP': 0.33},'C': {'EM': 0.6, 'EP': 0.4},'D': {'EM': 0.75, 'EP': 0.5}}},\r\n '2k': {'planim': {'A': {'EM': 0.56, 'EP': 0.34},'B': {'EM': 1, 'EP': 0.6},'C': {'EM': 1.6, 'EP': 1},'D': {'EM': 2, 'EP': 1.2}}, 'altim': {'A': {'EM': 0.27, 'EP': 0.17},'B': {'EM': 0.5, 'EP': 0.33},'C': {'EM': 0.6, 'EP': 0.4},'D': {'EM': 0.75, 'EP': 0.5}}},\r\n '5k': {'planim': {'A': {'EM': 1.4, 'EP': 0.85},'B': {'EM': 2.5, 'EP': 1.5},'C': {'EM': 4, 'EP': 2.5},'D': {'EM': 5, 'EP': 3}}, 'altim': {'A': {'EM': 0.54, 'EP': 0.34},'B': {'EM': 1, 'EP': 0.67},'C': {'EM': 1.2, 'EP': 0.8},'D': {'EM': 1.5, 'EP': 1}}},\r\n '10k': {'planim': {'A': {'EM': 2.8, 'EP': 1.7},'B': {'EM': 5, 'EP': 3},'C': {'EM': 8, 'EP': 5},'D': {'EM': 10, 'EP': 6}}, 'altim': {'A': {'EM': 1.35, 'EP': 0.84},'B': {'EM': 2.5, 'EP': 1.67},'C': {'EM': 3, 'EP': 2},'D': {'EM': 3.75, 'EP': 2.5}}},\r\n '25k': {'planim': {'A': {'EM': 7, 'EP': 4.25},'B': {'EM': 12.5, 'EP': 7.5},'C': {'EM': 20, 'EP': 12.5},'D': {'EM': 25, 'EP': 15}}, 'altim': {'A': {'EM': 2.7, 'EP': 1.67},'B': {'EM': 5, 'EP': 3.33},'C': {'EM': 6, 'EP': 4},'D': {'EM': 7.5, 'EP': 5}}},\r\n '50k': {'planim': {'A': {'EM': 14, 'EP': 8.5},'B': {'EM': 25, 'EP': 15},'C': {'EM': 40, 'EP': 25},'D': {'EM': 50, 'EP': 30}}, 'altim': {'A': {'EM': 5.5, 'EP': 3.33},'B': {'EM': 10, 'EP': 6.67},'C': {'EM': 12, 'EP': 8},'D': {'EM': 15, 'EP': 10}}},\r\n '100k': {'planim': {'A': {'EM': 28, 'EP': 17},'B': {'EM': 50, 'EP': 30},'C': {'EM': 80, 'EP': 50},'D': {'EM': 100, 'EP': 60}}, 'altim': {'A': {'EM': 13.7, 'EP': 8.33},'B': {'EM': 25, 'EP': 16.67},'C': {'EM': 30, 'EP': 20},'D': {'EM': 37.5, 'EP': 25}}},\r\n '250k': {'planim': {'A': {'EM': 70, 'EP': 42.5},'B': {'EM': 125, 'EP': 75},'C': {'EM': 200, 'EP': 125},'D': {'EM': 250, 'EP': 150}}, 'altim': {'A': {'EM': 27, 'EP': 16.67},'B': {'EM': 50, 'EP': 33.33},'C': {'EM': 60, 'EP': 40},'D': {'EM': 75, 'EP': 50}}}}\r\n \r\n dicionario = {'0.5k': '1:500', '1k': '1:1.000', '2k': '1:2.000', '5k': '1:5.000', '10k': '1:10.000', '25k': '1:25.000', '50k': '1:50.000', '100k': '1:100.000', '250k': '1:250.000'}\r\n \r\n valores = ['A', 'B', 'C', 'D']\r\n \r\n Escalas = [ esc for esc in dicionario]\r\n \r\n # Verificacoes\r\n # As duas camadas devem estar no mesmo SRC e estarem projetadas\r\n # As duas camadas devem ser do tipo linha\r\n crs1 = ref.sourceCrs()\r\n crs2 = teste.sourceCrs()\r\n distance = QgsDistanceArea()\r\n\r\n \r\n if not (crs1 == crs2 and not(crs1.isGeographic())):\r\n raise QgsProcessingException('SRC das camadas de entrada devem ser iguais e estar projetados!')\r\n\r\n \r\n feedback.pushInfo('Aplicando o o Método dos Retângulos Equivalentes...')\r\n \r\n \r\n # Calcular Discrepancias\r\n DISCREP = []\r\n COMPR = []\r\n total = 100.0 / teste.featureCount() if teste.featureCount() else 0\r\n \r\n for index, feat1 in enumerate(teste.getFeatures()):\r\n pol = feat1.geometry()\r\n compr = 0\r\n S = pol.area()\r\n p = pol.length()/2.0\r\n x1 = (p-sqrt(p*p-4*S))/2.0\r\n DISCREP += [float(x1)]\r\n for feat2 in ref.getFeatures():\r\n lin = feat2.geometry()\r\n if pol.intersects(lin):\r\n inter = pol.intersection(lin)\r\n if inter.type() == 1:\r\n compr += inter.length()\r\n COMPR += [compr]\r\n \r\n feature = QgsFeature(Fields)\r\n feature.setGeometry(pol)\r\n feature.setAttributes([float(x1), float(compr)])\r\n sink.addFeature(feature, QgsFeatureSink.FastInsert)\r\n \r\n if feedback.isCanceled():\r\n break\r\n feedback.setProgress(int((index+1) * total))\r\n\r\n \r\n\r\n # Gerar relatorio do metodo\r\n DISCREP= array(DISCREP)\r\n COMPR = array(COMPR)\r\n EMQ = sqrt((DISCREP*DISCREP*COMPR).sum()/COMPR.sum())\r\n media_Pond =sum(DISCREP*COMPR)/sum(COMPR)\r\n RESULTADOS = {}\r\n for escala in Escalas:\r\n mudou = False\r\n for valor in valores[::-1]:\r\n EM = PEC[escala]['planim'][valor]['EM']\r\n EP = PEC[escala]['planim'][valor]['EP']\r\n if (sum((DISCREP0.9 and (EMQ < EP):\r\n RESULTADOS[escala] = valor\r\n mudou = True\r\n if not mudou:\r\n RESULTADOS[escala] = 'R'\r\n\r\n feedback.pushInfo('RESULTADOS:')\r\n feedback.pushInfo('Media: {} m'.format(round(DISCREP.mean(),3)))\r\n feedback.pushInfo('Media Ponderada das Discrepancias: {} m'.format(round(media_Pond,3)))\r\n feedback.pushInfo('REMQ Ponderado: {} m'.format(round(EMQ,3)))\r\n feedback.pushInfo('Desvio-padrao: {} m'.format(round(DISCREP.std(),3)))\r\n for result in RESULTADOS:\r\n feedback.pushInfo('{} ➜ {}'.format(dicionario[result],RESULTADOS[result]))\r\n\r\n \r\n # Criacao do arquivo html com os resultados\r\n arq = open(html_output, 'w')\r\n texto = '''\r\n \r\n \r\n \r\n \r\n MRE\r\n \r\n \r\n
MÉTODO\r\n DOS\r\n RETÂNGULOS EQUIVALENTES
\r\n
\r\n
\r\n 1. Camada de Referência
\r\n     a. nome: {}
\r\n     b. total de feições: {}
\r\n
\r\n 2. Camada de Teste
\r\n     a. nome: {}
\r\n     b. total de feições: {}
\r\n
\r\n 3. Relatório
\r\n     a. Comprimento total relacionado (m): {}
\r\n     b. média ponderada das discrepâncias (m): {}
\r\n     c. REMQ ponderado (m): {}
\r\n     d. discrepância máxima: {}
\r\n     e. discrepância mínima: {}
\r\n     f. PEC-PCD:
'''.format(self.str2HTML(ref.sourceName()), ref.featureCount(), self.str2HTML(teste.sourceName()), teste.featureCount(), '{:.4f}'.format(COMPR.sum()), '{:.4f}'.format(media_Pond), '{:.4f}'.format(EMQ), '{:.4f}'.format(max(DISCREP)),'{:.4f}'.format(min(DISCREP)))\r\n \r\n texto += '''\r\n \r\n ''' \r\n \r\n for escala in Escalas:\r\n texto += ' '.format(dicionario[escala])\r\n texto +='''\r\n \r\n '''\r\n \r\n for escala in Escalas:\r\n texto += ' '.format(RESULTADOS[escala])\r\n texto +='''\r\n \r\n \r\n
{}
{}
\r\n
\r\n
\r\n
\r\n
Leandro França\r\n 2022
\r\n Eng. Cartógrafo
\r\n email: geoleandro.franca@gmail.com
\r\n
\r\n
\r\n \r\n '''\r\n arq.write(texto)\r\n arq.close()\r\n \r\n feedback.pushInfo('Operação finalizada com sucesso!')\r\n feedback.pushInfo('Leandro França - Eng Cart')\r\n return {self.OUTPUT: dest_id,\r\n self.HTML: html_output}\r\n","repo_name":"LEOXINGU/doutorado","sub_path":"QGIS/processing/scripts/MRE.py","file_name":"MRE.py","file_ext":"py","file_size_in_byte":13846,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"21292232177","text":"\"\"\" https://leetcode.com/problems/lexicographically-smallest-beautiful-string/\nlearn from lingshen: https://www.bilibili.com/video/BV1QX4y1m71X/?spm_id_from=333.337.search-card.all.click&vd_source=349088063a324f397a5e80a33effd4f0\n\"\"\"\nfrom header import *\n\nclass Solution:\n def smallestBeautifulString(self, s: str, k: int) -> str:\n A = [ord(c)-ord('a') for c in s]\n i = len(A)-1\n A[i] += 1\n while 0<=i1 and A[i]==A[i-2]:\n A[i] += 1\n else:\n i += 1\n return ''.join(chr(ord('a')+a) for a in A)\n \n\n\"\"\" cef\n\"abcz\"\n26\n\"dc\"\n4\n\"ced\"\n6\n\n\"abda\"\n\"\"\"","repo_name":"824zzy/Leetcode","sub_path":"Q_Greedy/AdvancedGreedy/L3_2663_Lexicographically_Smallest_Beautiful_String.py","file_name":"L3_2663_Lexicographically_Smallest_Beautiful_String.py","file_ext":"py","file_size_in_byte":831,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"96"} +{"seq_id":"7148645363","text":"from fastapi import APIRouter, Depends, Path, HTTPException, Request, status\nfrom fastapi.templating import Jinja2Templates\nfrom models import Todo, TodoItems, Item\n\ntodo_router = APIRouter()\n\ntodo_list = []\ntemplates = Jinja2Templates(directory=\"templates/\")\n\n@todo_router.post(\"/todo\", status_code=201)\nasync def add_todo(request: Request, todo: Todo = Depends(Todo.as_form)):\n print(f\"The item: ${todo}\")\n todo.id = len(todo_list) + 1\n todo_list.append(todo)\n\n return templates.TemplateResponse(\"todo.html\", {\n \"todos\": todo_list,\n \"request\": request\n })\n\n@todo_router.get(\"/todo\", response_model=TodoItems)\nasync def retrieve_todos(request: Request):\n print(\"received the request\")\n return templates.TemplateResponse(\"todo.html\", {\n \"request\": request,\n \"todos\": todo_list\n })\n\n@todo_router.delete(\"/todo/{todo_id}\")\nasync def delete_single_todo(todo_id: int = Path(..., title=\"ID of the todo item to be deleted\")) -> dict:\n for index in range(len(todo_list)):\n todo = todo_list[index]\n\n if todo.id == todo_id:\n todo_list.pop(index)\n\n return {\n \"message\": \"Todo deleted successfully\"\n }\n\n raise HTTPException(\n status_code=status.HTTP_404_NOT_FOUND,\n detail='Todo with supplied ID does not exist'\n )\n\n\n@todo_router.put(\"/todo/{todo_id}\")\nasync def update_todo(todo_data: Todo, todo_id: int = Path(..., title=\"the ID of the todo item to update\")) -> dict:\n for todo in todo_list:\n if todo.id == todo_id:\n todo.item = todo_data.item\n\n return {\n \"message\": \"Todo updated successfully\"\n }\n\n raise HTTPException(\n status_code=status.HTTP_404_NOT_FOUND,\n detail='Todo with supplied ID does not exist'\n )\n\n@todo_router.get(\"/todo/{todo_id}\")\nasync def get_single_todo(request: Request, todo_id: int = Path(..., title=\"The ID of the todo to retrieve\")) -> dict:\n for todo in todo_list:\n if todo.id == todo_id:\n return templates.TemplateResponse(\"todo.html\", {\n \"request\": request,\n \"todo\": todo\n })\n\n\n raise HTTPException(\n status_code=status.HTTP_404_NOT_FOUND,\n detail='Todo with supplied ID does not exist'\n )\n\n@todo_router.delete(\"/todo\")\nasync def delete_all_todo() -> dict:\n todo_list.clear()\n\n return {\n \"message\": \"Todos deleted successfully\"\n }","repo_name":"erickmwarama2/fast_api_routes","sub_path":"todo.py","file_name":"todo.py","file_ext":"py","file_size_in_byte":2474,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"21669189995","text":"import cv2\nimport os\nfrom camera_calibration import undistort\nfrom ball_tracking_yolo import yolo_label_to_pixel_coords\nfrom formulas import dist_formula\nfrom collections import namedtuple\n\nPERSON_LABEL_NUM = '0'\nSPORTSBALL_LABEL_NUM = '32'\n\n\n#source: https://stackoverflow.com/questions/27152904/calculate-overlapped-area-between-two-rectangles\ndef intersection_area(bb_shotput, bb_thrower):\n Rectangle = namedtuple('Rectangle', 'xmin ymin xmax ymax')\n r_shotput = Rectangle(bb_shotput['x1'], bb_shotput['y1'], bb_shotput['x2'], bb_shotput['y2'])\n r_thrower = Rectangle(bb_thrower['x1'], bb_thrower['y1'], bb_thrower['x2'], bb_thrower['y2'])\n dx = min(r_shotput.xmax, r_thrower.xmax) - max(r_shotput.xmin, r_thrower.xmin)\n dy = min(r_shotput.ymax, r_thrower.ymax) - max(r_shotput.ymin, r_thrower.ymin)\n if (dx >= 0) and (dy >= 0):\n return dx * dy\n else: return 0\n\n\n# if a sports ball and person are both detected within the search window\n# AND sports ball bounding box does not overlap too closely or is not completely within human bounding box\n# then that is starting point for shot put flight\ndef detect_flight(video, stop_box_coords, sector_coords, isLeft):\n\n #initialize key variables\n shotput_in_flight = False\n\n #read first frame in video\n success, frame = video.read()\n frame = undistort(frame, isLeft)\n height, width, _ = frame.shape\n\n # we know coordinates of box upon calibration in the video, create search window above box for ball\n if isLeft is True:\n search_window_top_left = (sector_coords[0][0], int(sector_coords[0][1] - height*0.2))\n search_window_bottom_right = (int(sector_coords[1][0] + width*0.2), stop_box_coords[1])\n else:\n search_window_top_left = (int(sector_coords[0][0]-width*0.2), int(sector_coords[0][1] - height*0.2))\n search_window_bottom_right = (sector_coords[1][0], stop_box_coords[1])\n\n frame_num = 0\n\n # Testing Code\n # cv2.circle(frame, sector_coords[0], 5, (255, 255, 0), 5)\n # cv2.circle(frame, sector_coords[1], 5, (255, 255, 0), 5)\n # cv2.circle(frame, stop_box_coords, 5, (255,255,0), 5)\n # cv2.rectangle(frame, search_window_top_left, search_window_bottom_right, (255,0,0), 5)\n # cv2.imshow(\"frame\", frame)\n # cv2.waitKey(0)\n\n #looping through frames until shot put is detected to be in flight or video ends\n while shotput_in_flight is False and success is True:\n #first, undistort image based on camera calibration\n if isLeft:\n frame = undistort(frame, True)\n else:\n frame = undistort(frame, False)\n\n\n #write frame so yolo detect can find it\n cv2.imwrite(\"currentdetectflightframe.png\", frame)\n\n #call my adapted yolov3 detect script on current frame using pretrained yolov3 weights (which is the default, hence why --weights specifier is not included)\n os.system(\"python3.7 /Users/ashley20/PycharmProjects/ThesisCameraCalibration/yolov3/detect.py\"\n \" --source /Users/ashley20/PycharmProjects/ThesisCameraCalibration/currentdetectflightframe.png\"\n \" --save-txt\"\n \" --save-conf\"\n \" --exist-ok\"\n )\n\n #TESTING CODE\n testframe = cv2.imread('/Users/ashley20/PycharmProjects/ThesisCameraCalibration/yolov3/runs/detect/exp/currentdetectflightframe.png')\n # cv2.rectangle(testframe, search_window_top_left, search_window_bottom_right,\n # (255, 0, 0), 5)\n # cv2.imshow(\"test\", testframe)\n # cv2.waitKey(0)\n\n # open txt file of labels that yolov3 creates during detect.py\n with open(\n '/Users/ashley20/PycharmProjects/ThesisCameraCalibration/yolov3/runs/detect/exp/labels/currentdetectflightframe.txt',\n 'r') as label_file:\n #initialize dictionary used to process detected object information returned from label file\n #key = class label number, will either be '0' or '32'\n #value = array of pixel values for every detected object in the respective category\n detected_objs = {}\n\n #loop through label file to process detected objects\n for line in label_file:\n line = line.split()\n object_label_num = line[0]\n #if the current detected object is a person or sports ball, add to dictionary\n #else ignore object\n if object_label_num == PERSON_LABEL_NUM or object_label_num == SPORTSBALL_LABEL_NUM:\n x, y, width, height, confidence = yolo_label_to_pixel_coords(line, frame)\n #adding information to dictionary\n if object_label_num in detected_objs:\n objs = detected_objs[object_label_num]\n objs.append([x,y,width,height,confidence])\n detected_objs[object_label_num] = objs\n else:\n detected_objs[object_label_num] = [[x,y,width,height,confidence]]\n\n #check if at least one human and at least one sports ball detected\n #else, move to next frame\n if PERSON_LABEL_NUM in detected_objs and SPORTSBALL_LABEL_NUM in detected_objs:\n #check if at least one sports ball detected is within search window\n shotput_candidates = detected_objs[SPORTSBALL_LABEL_NUM]\n\n #new array for only shotput candidates within search window\n filtered_shotput_candidates = []\n for sc in shotput_candidates:\n if (sc[0] > search_window_top_left[0] and sc[0] < search_window_bottom_right[0] and sc[1] > search_window_top_left[1] and sc[1] < search_window_bottom_right[1]):\n filtered_shotput_candidates.append(sc)\n\n #if there are shotput candidates inside the search window\n #else, move to next frame\n if len(filtered_shotput_candidates) > 0:\n #check if at least one person detected is within search window\n thrower_candidates = detected_objs[PERSON_LABEL_NUM]\n\n # new array for only thrower candidates within search window\n filtered_thrower_candidates = []\n for tc in thrower_candidates:\n if (tc[0] > search_window_top_left[0] and tc[0] < search_window_bottom_right[0] and tc[1] > search_window_top_left[1] and tc[1] < search_window_bottom_right[1]):\n filtered_thrower_candidates.append(tc)\n\n # if there are thrower candidates inside the search window, then there is at least one shot put candidate\n # AND at least one thrower candidate\n # else, move to next frame\n if len(filtered_thrower_candidates) > 0:\n\n #for each shot put candidate, find person who is the closest by Euclidean distance\n for sc in filtered_shotput_candidates:\n all_dists = {}\n for tc in filtered_thrower_candidates:\n dist = dist_formula((sc[0], sc[1]), (tc[0], tc[1]))\n all_dists[(tc[0], tc[1], tc[2], tc[3])] = dist\n min_key = min(all_dists, key=all_dists.get)\n min_dist = all_dists[min_key]\n tc=min_key\n\n #create dictionary to construct bounding box of current shot put and thrower candidates\n x_min = sc[0] - int(sc[2] / 2)\n y_min = sc[1] - int(sc[3] / 2)\n x_max = sc[0] + int(sc[2] / 2)\n y_max = sc[1] + int(sc[3] / 2)\n bb_shotput = {\n 'x1': x_min,\n 'x2': x_max,\n 'y1': y_min,\n 'y2': y_max\n }\n x_min = tc[0] - int(tc[2] / 2)\n y_min = tc[1] - int(tc[3] / 2)\n x_max = tc[0] + int(tc[2] / 2)\n y_max = tc[1] + int(tc[3] / 2)\n bb_thrower = {\n 'x1': x_min,\n 'x2': x_max,\n 'y1': y_min,\n 'y2': y_max\n }\n\n #if shot put bounding box is entirely inside thrower bounding box, then ignore and move to next frame\n #this is to prevent the shot put from being detected while it is still in the thrower's hand\n if bb_shotput['x1'] > bb_thrower['x1'] and bb_shotput['y1'] > bb_thrower['y1'] and bb_shotput['x2'] < bb_thrower['x2'] and bb_shotput['y2'] < bb_thrower['y2']:\n print(\"nested bounding boxes\")\n continue\n\n #calculate percent area of shot put bounding box that is within thrower bounding box\n #it must be below a certain threshold to be deemed \"has left thrower's hand\"\n intersected_area = intersection_area(bb_shotput, bb_thrower)\n bb_shotput_area = int(sc[2])*int(sc[3])\n area_ratio = intersected_area/bb_shotput_area\n\n #check that shot put is close enough to thrower without overlapping too much\n #if these criteria are true, then shot put has just left the thrower's hand\n if min_dist < 500 and area_ratio < 0.6:\n shotput_in_flight = True\n print(min_dist)\n print(area_ratio)\n #TESTING CODE\n # cv2.rectangle(testframe, search_window_top_left, search_window_bottom_right,\n # (255, 0, 0), 5)\n # cv2.circle(testframe, (bb_thrower['x1'], bb_thrower['y1']),5, (255, 255, 0), 5)\n # cv2.circle(testframe, (bb_thrower['x2'], bb_thrower['y2']), 5, (255, 255, 0), 5)\n # cv2.imshow('f', testframe)\n # cv2.waitKey(0)\n cv2.destroyAllWindows()\n print(sc[2])\n print(sc[3])\n return (int(sc[0]), int(sc[1]), int(sc[2]), int(sc[3]))\n\n #increments frame number\n frame_num += 1\n #gets next frame\n success, frame = video.read()\n\n #if video ends and while loop exits, but never finds the flight of shot put\n if shotput_in_flight is False:\n print(\"video ended before detecting shot put in flight, exiting code\")\n return None\n exit()\n\nif __name__ == '__main__':\n print(\"in detect flight\")\n video = cv2.VideoCapture('Thesis_Data_Videos_Test/throwclose_2_414_behind_shot_on_190_left.MP4')\n box_coords_left = (1109, 1223)\n box_coords_right = (842, 915)\n sector_coords_left = [(321,855),(1016,941)]\n sector_coords_right = [(898,627), (1470,684)]\n if not video.isOpened():\n print(\"no video opened\")\n exit()\n detect_flight(video, box_coords_left, sector_coords_left, isLeft=True)\n","repo_name":"ashleys20/ThesisCopy","sub_path":"detect_flight.py","file_name":"detect_flight.py","file_ext":"py","file_size_in_byte":11642,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"38509843536","text":"from django.urls import reverse\n\nfrom automatic_replenishment_system.retail_core.core.utils.csv_utils import FieldFileCsvHelper\nfrom automatic_replenishment_system.retail_core.core.utils.request_maker import RequestMaker\n\n\nclass BrandCreationProcessManager:\n def __init__(self, brand, product_file, store_file, warehouse_file, static_rank_file):\n self.brand = brand\n self.product_file = product_file\n self.store_file = store_file\n self.warehouse_file = warehouse_file\n self.static_rank_file = static_rank_file\n self.brand_importer_api = 'http://0.0.0.0:8000' + reverse('brand_importer')\n\n def execute(self):\n products, _ = FieldFileCsvHelper().read_csv_file(self.product_file)\n stores, _ = FieldFileCsvHelper().read_csv_file(self.store_file)\n warehouses, _ = FieldFileCsvHelper().read_csv_file(self.warehouse_file)\n static_ranks, _ = FieldFileCsvHelper().read_csv_file(self.static_rank_file)\n post_data = self._get_post_parameters(products, stores, warehouses, static_ranks)\n self._call_api(post_data)\n\n def _get_post_parameters(self, products, stores, warehouses, static_ranks):\n post_data = dict()\n post_data['name'] = self.brand.name\n post_data['ranking_model'] = self.brand.ranking_model\n post_data['stores'] = stores\n post_data['products'] = products\n post_data['warehouses'] = warehouses\n post_data['static_ranks'] = static_ranks\n return post_data\n\n def _call_api(self, request_parameters):\n response = RequestMaker().post_request(self.brand_importer_api, request_parameters, retry=3)\n return response\n","repo_name":"udwivedi394/automatic_replenishment","sub_path":"automatic_replenishment_system/retail_core/core/brand/brand_saver.py","file_name":"brand_saver.py","file_ext":"py","file_size_in_byte":1674,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"26606147042","text":"from django.shortcuts import render, get_object_or_404\nfrom rango.models import Category, Page, UserProfile\nfrom rango.forms import CategoryForm, PageForm, UserForm, UserProfileForm\nfrom django.http import HttpResponseRedirect, HttpResponse\nfrom django.core.urlresolvers import reverse\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required\nfrom datetime import datetime\nfrom rango.bing_search import run_query\nfrom django.contrib.auth.models import User\nfrom rango.category_search import get_category_list\n\n# Create your views here.\ndef index(request):\n #request.session.set_test_cookie()\n\n category_list = Category.objects.order_by('-likes')[:5]\n page_list = Page.objects.order_by(\"-views\")[0:5]\n context = {\"categories\": category_list, \"pages\": page_list}\n\n visits = request.session.get(\"visits\")\n\n if not visits:\n visits = 1\n\n reset_last_visit = False\n\n last_visit = request.session.get(\"last_visit\")\n\n if last_visit:\n last_visit_time = datetime.strptime(last_visit[0:-7], \"%Y-%m-%d %H:%M:%S\")\n\n if (datetime.now() - last_visit_time).seconds > 5:\n visits += 1\n reset_last_visit = True\n \n else:\n reset_last_visit= True\n\n if reset_last_visit:\n request.session[\"last_visit\"] = str(datetime.now())\n request.session[\"visits\"] = visits\n\n context[\"visits\"] = visits\n \n return render(request, \"rango/index.html\", context)\n\ndef category(request, category_name_slug):\n context = {}\n try:\n category = Category.objects.get(slug=category_name_slug)\n context[\"category_name\"] = category.name\n\n pages = Page.objects.filter(category=category).order_by(\"-views\")\n\n context[\"pages\"] = pages\n context[\"category\"] = category\n\n except Category.DoesNotExist:\n pass\n\n result_list = []\n query = None\n\n if request.method == \"POST\":\n query = request.POST[\"query\"].strip()\n if query:\n result_list = run_query(query)\n \n context[\"result_list\"] = result_list\n if not query:\n query = category.name\n context[\"query\"] = query\n\n return render(request, \"rango/category.html\", context)\n\n@login_required\ndef add_category(request):\n if request.method == \"POST\":\n form = CategoryForm(request.POST)\n\n if form.is_valid():\n form.save(commit=True)\n return HttpResponseRedirect(reverse(\"rango:index\"))\n\n else:\n form = CategoryForm()\n\n return render(request, \"rango/add_category.html\", {\"form\":form})\n\n@login_required\ndef add_page(request, category_name_slug):\n context = {}\n \n try:\n category = Category.objects.get(slug=category_name_slug)\n context[\"category\"] = category\n except Category.DoesNotExist:\n category = False\n\n if request.method == \"POST\":\n form = PageForm(request.POST)\n\n if form.is_valid():\n form.save(commit=True)\n return HttpResponseRedirect(reverse(\"rango:category\", args=(category_name_slug,)))\n\n else:\n if category:\n form = PageForm(initial={\"category\": category})\n context[\"form\"] = form\n\n return render(request, \"rango/add_page.html\", context)\n\ndef track_url(request):\n context = {}\n pages = []\n category = None\n\n if request.method == \"GET\":\n if \"category_id\" in request.GET:\n category_id = request.GET[\"category_id\"]\n try:\n category = Category.objects.get(pk=category_id)\n except Category.DoesNotExist:\n category = None\n\n if category:\n if \"page_id\" in request.GET:\n page_id = request.GET[\"page_id\"]\n \n try:\n page = Page.objects.get(pk=page_id)\n except Page.DoesNotExist:\n page = None\n\n if page:\n page.views += 1\n page.save()\n\n pages = Page.objects.filter(category=category).order_by(\"-views\")\n\n context[\"pages\"] = pages\n context[\"category\"] = category\n return render(request, \"rango/search_add_page.html\", context)\n\n@login_required\ndef register_profile(request):\n profile_registered = False\n\n if request.method == \"POST\":\n profile_form = UserProfileForm(data=request.POST)\n\n if profile_form.is_valid():\n if request.user.is_authenticated():\n user = request.user\n if user.is_active:\n profile = profile_form.save(commit=False)\n profile.user = user\n\n # handle profile picture\n if \"picture\" in request.FILES:\n profile.picture = request.FILES[\"picture\"]\n\n profile_registered = True\n profile.save()\n\n else:\n profile_form = UserProfileForm()\n\n context = {\"profile_form\": profile_form, \"profile_registered\": profile_registered}\n\n return render(request, \"rango/profile_registration.html\", context)\n\n@login_required\ndef profile(request):\n context = {}\n\n if request.method == \"GET\":\n if \"user_id\" in request.GET:\n user_id = request.GET[\"user_id\"]\n user = get_object_or_404(User, pk=user_id)\n user_profile = UserProfile.objects.get(pk=user)\n\n if user.id != request.user.id:\n context[\"own\"] = False\n else:\n context[\"own\"] = True\n\n context[\"user_profile\"] = user_profile\n\n return render(request, \"rango/profile.html\", context)\n \n else:\n return HttpResponseRedirect(reverse(\"rango:index\", ))\n\n elif request.method == \"POST\":\n user = request.user\n user_profile = UserProfile.objects.get(pk=user)\n\n if request.POST[\"username\"]:\n user.username = request.POST[\"username\"]\n\n if request.POST[\"email\"]:\n user.email = request.POST[\"email\"]\n\n if request.POST[\"first_name\"]:\n user.first_name = request.POST[\"first_name\"]\n\n if request.POST[\"last_name\"]:\n user.last_name = request.POST[\"last_name\"]\n\n if request.POST[\"website\"]:\n user_profile.website = request.POST[\"website\"]\n\n if \"picture\" in request.FILES:\n user_profile.picture = request.FILES[\"picture\"]\n \n user.save()\n user_profile.save()\n\n context = {\"user_profile\": user_profile, \"own\": True}\n return render(request, \"rango/profile.html\", context)\n\n@login_required\ndef users(request):\n context = {\"users\": None}\n if request.user.is_authenticated():\n user = request.user\n if user.is_active:\n users = User.objects.all()\n context[\"users\"] = users\n\n return render(request, \"rango/users.html\", context)\n\n@login_required\ndef like_category(request):\n if request.method == \"GET\":\n if \"category_id\" in request.GET:\n category_id = request.GET[\"category_id\"]\n try:\n category = Category.objects.get(pk=category_id)\n except Category.DoesNotExist:\n category = None\n\n if category:\n category.likes += 1\n category.save()\n return HttpResponse(category.likes)\n\n return HttpResponse(0)\n\ndef suggest_category(request):\n context = {}\n categories = []\n if request.method == \"GET\":\n if \"suggestion\" in request.GET:\n starts_with = request.GET[\"suggestion\"]\n\n categories = get_category_list(8, starts_with)\n\n context[\"cats\"] = categories\n return render(request, 'rango/cats.html', context)\n\n@login_required\ndef search_add_page(request):\n context = {}\n pages = []\n category = None\n\n if request.method == \"GET\":\n if \"category_id\" in request.GET:\n category_id = request.GET[\"category_id\"]\n try:\n category = Category.objects.get(pk=category_id)\n except Category.DoesNotExist:\n category = None\n\n if category:\n if \"page_link\" and \"page_title\" in request.GET:\n page_link = request.GET[\"page_link\"]\n page_title = request.GET[\"page_title\"]\n new_page = Page.objects.get_or_create(category=category, title=page_title, url=page_link, views=0)[0]\n\n pages = Page.objects.filter(category=category).order_by(\"-views\")\n\n context[\"pages\"] = pages\n context[\"category\"] = category\n return render(request, \"rango/search_add_page.html\", context)\n\ndef about(request):\n visits = request.session.get(\"visits\")\n if not visits:\n visits = 1\n return render(request, \"rango/about.html\", {\"visits\": visits})\n\n# def register(request):\n# #if request.session.test_cookie_worked():\n# #print(\"TEST COOKIE WORKED!\")\n# #request.session.delete_test_cookie()\n\n# registered = False\n\n# if request.method == \"POST\":\n# user_form = UserForm(data=request.POST)\n# profile_form = UserProfileForm(data=request.POST)\n\n# if user_form.is_valid() and profile_form.is_valid():\n# user = user_form.save()\n\n# # hash password\n# user.set_password(user.password)\n# user.save()\n\n# profile = profile_form.save(commit=False)\n# profile.user = user\n\n# # handle profile picture\n# if \"picture\" in request.FILES:\n# profile.picture = request.FILES[\"picture\"]\n\n# profile.save()\n\n# registered = True\n\n# else:\n# user_form = UserForm()\n# profile_form = UserProfileForm()\n\n# context = {\"user_form\": user_form, \"profile_form\": profile_form, \"registered\": registered}\n\n# return render(request, \"rango/register.html\", context)\n\n# def user_login(request):\n# if request.method == \"POST\":\n# username = request.POST[\"username\"]\n# password = request.POST[\"password\"]\n\n# user = authenticate(username=username, password=password)\n\n# if user:\n# if user.is_active:\n# login(request, user)\n# return HttpResponseRedirect(reverse(\"rango:index\"))\n# else:\n# return HttpResponse(\"Your Rango account is disabled.\")\n\n# else:\n# return HttpResponse(\"Invalid username or password.\")\n\n# else:\n# return render(request, \"rango/login.html\", {})\n\n# @login_required\n# def user_logout(request):\n# logout(request)\n# return HttpResponseRedirect(reverse(\"rango:index\"))\n\n","repo_name":"naity/Rango","sub_path":"tango_with_django_project/rango/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":10630,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"24942659115","text":"from rest_framework.response import Response\nfrom rest_framework.decorators import api_view\nfrom rest_framework import status\n\nfrom .models import Log\nfrom .serializers import LogSerializer\n\n\n@api_view(['GET', 'POST'])\ndef logList(request):\n if (request.method == 'GET'):\n user = request.GET.get(\"user\")\n logs = Log.objects.filter(user=user)\n serializer = LogSerializer(logs, many=True)\n return Response(serializer.data)\n\n elif (request.method == 'POST'):\n serializer = LogSerializer(data=request.data)\n if (serializer.is_valid()):\n serializer.save()\n return Response(serializer.data)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@api_view(['GET', 'PUT', 'DELETE'])\ndef logDetail(request, pk):\n\n try:\n log = Log.objects.get(pk=pk)\n except Log.DoesNotExist:\n return Response(status=status.HTTP_404_NOT_FOUND)\n\n if (request.method == 'GET'):\n serializer = LogSerializer(log)\n return Response(serializer.data)\n\n elif (request.method == 'PUT'):\n serializer = LogSerializer(log, data=request.data)\n if (serializer.is_valid()):\n serializer.save()\n return Response(serializer.data)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n elif (request.method == 'DELETE'):\n log.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n","repo_name":"AnshulKanwar/Nutrition-Planner","sub_path":"backend/logger/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1456,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"25158501928","text":"import numpy as np\nfrom sklearn import datasets\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.neighbors import KNeighborsClassifier\n#数据归一化的模块\nfrom sklearn.preprocessing import StandardScaler\n\n\niris = datasets.load_iris()\nX = iris.data\ny = iris.target\nX_train,X_test,y_train,y_test = train_test_split(X,y,test_size=0.2,random_state=666)\nstandardscaler = StandardScaler()\nstandardscaler.fit(X_train)\nmeanis = standardscaler.mean_ #由用户传进去的数据得到的数据后面带_ 这是代码规范\nstdis = standardscaler.scale_ #std_这种查询方差的方法不合理,用scale_\nprint(meanis)\nprint(stdis)\ntransformis = standardscaler.transform(X_train)\nX_train = transformis #得到归一化后的数据\nprint(X_train)\nX_test = standardscaler.transform(X_test) #得到归一化后的数据\nprint(X_test)\nknn_cif = KNeighborsClassifier(n_neighbors=3)\nknn_cif.fit(X_train,y_train)\npred = knn_cif.predict(X_test)\nprint(pred)\nprint(knn_cif.score(X_test,y_test))\n\n\n\n#自定义StandardScaler\n# class StandardScaler:\n#\n# def __init__(self):\n# self.mean_ = None\n# self.scale_ = None\n# def fit(self,X):\n# \"暂时处理二维的数据\"\n# assert X.ndim == 2,\"the dimension of X must be 2\"\n# self.mean_ = np.mean(X,axis=0)\n# self.scale_ = np.std(X,axis=0)\n# return self\n# def transform(self,X):\n# \"也暂时处理二维数据\"\n# assert self.mean_ is not None and self.scale_ is not None,'must fit before transform'\n# assert X.shape[1] == len(self.mean_) ,\"the feature number of X must be equal to mean_ and scale_ \"\n# assert X.ndim == 2 ,\"the dimension of X must be 2\"\n# #要把用户的数据确定为浮点型\n# resX = np.array(X,dtype=float)\n# #这两种方法好像有相同的作用\n# # resX = np.empty(shape=X.shape,dtype=float)\n# result = (resX-self.mean_)/self.scale_\n# return result\n# iris = datasets.load_iris()\n# X = iris.data\n# y = iris.target\n#\n# X_train,X_test,y_train,y_test = train_test_split(X,y,test_size=0.2,random_state=666)\n# standardscaler = StandardScaler()\n# standardscaler.fit(X_train)\n# print(standardscaler.mean_)\n# print(standardscaler.scale_)\n# X_train = standardscaler.transform(X_train)\n# X_test = standardscaler.transform(X_test)\n# knn_cif = KNeighborsClassifier(n_neighbors=3)\n# knn_cif.fit(X_train,y_train)\n# print(knn_cif.score(X_test,y_test))\n\n'''\n[5.83416667 3.0825 3.70916667 1.16916667]\n[0.81019502 0.44076874 1.76295187 0.75429833]\n1.0\n'''\n","repo_name":"LiaoBoWen/Machine_learn","sub_path":"KNN/sklearn_scaler_.py","file_name":"sklearn_scaler_.py","file_ext":"py","file_size_in_byte":2577,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"74364710395","text":"# Submitted BY:Fozia Hameed\r\n\r\nimport numpy as np\r\nimport pandas as pd\r\nimport streamlit as st\r\nimport altair as alt\r\nimport random\r\n\r\n\r\n\r\nst.title(''' \r\n # **Pakistan Agricultural & Population Information Statistics Web App**\r\n\r\n ''')\r\nst.image('flag.png',width=200)\r\nst.sidebar.title('Pakistan DashBoard')\r\nst.sidebar.markdown(''' \r\nThis app is to give insights about Population Information \r\nThe data considerd for this analysis from 1950 to 2018.\r\nSelect the different options to vary the Visualization\r\nAll the Charts are interactive. \r\n\r\nNote: The data arehas been read from the https://www.fao.org/faostat/en/#data/OA\r\n\r\nDesigned by:Fozia Hameed''') \r\n\r\ndf=pd.read_csv('population.csv')\r\ngroup1=df['Element'].unique().tolist()\r\ncolor_array=['black', 'silver', 'gray', 'white', 'maroon', 'red', 'purple', 'fuchsia', 'green', 'lime', 'olive', 'yellow', 'navy', 'blue', 'teal', 'aqua', 'orange', 'aliceblue', 'antiquewhite', 'aquamarine',\r\n 'azure', 'beige', 'bisque', 'blanchedalmond', 'blueviolet', 'brown', 'burlywood', 'cadetblue', 'chartreuse', 'chocolate', 'coral', 'cornflowerblue', 'cornsilk', 'crimson', 'cyan', 'darkblue', \r\n 'darkcyan', 'darkgoldenrod', 'darkgray', 'darkgreen', 'darkgrey', 'darkkhaki', 'darkmagenta', 'darkolivegreen', 'darkorange', 'darkorchid', 'darkred', 'darksalmon', 'darkseagreen', 'darkslateblue',\r\n 'darkslategray', 'darkslategrey', 'darkturquoise', 'darkviolet', 'deeppink', 'deepskyblue', 'dimgray', 'dimgrey', 'dodgerblue', 'firebrick', 'floralwhite', 'forestgreen', 'gainsboro', 'ghostwhite',\r\n 'gold', 'goldenrod', 'greenyellow', 'grey', 'honeydew', 'hotpink', 'indianred', 'indigo', 'ivory', 'khaki', 'lavender', 'lavenderblush', 'lawngreen', 'lemonchiffon', 'lightblue', 'lightcoral', \r\n 'lightcyan', 'lightgoldenrodyellow', 'lightgray', 'lightgreen', 'lightgrey', 'lightpink', 'lightsalmon', 'lightseagreen', 'lightskyblue', 'lightslategray', 'lightslategrey', 'lightsteelblue', \r\n 'lightyellow', 'limegreen', 'linen', 'magenta', 'mediumaquamarine', 'mediumblue', 'mediumorchid', 'mediumpurple', 'mediumseagreen', 'mediumslateblue', 'mediumspringgreen', 'mediumturquoise', \r\n 'mediumvioletred', 'midnightblue', 'mintcream', 'mistyrose', 'moccasin', 'navajowhite', 'oldlace', 'olivedrab', 'orangered', 'orchid', 'palegoldenrod', 'palegreen', 'paleturquoise', 'palevioletred', \r\n 'papayawhip', 'peachpuff', 'peru', 'pink', 'plum', 'powderblue', 'rosybrown', 'royalblue', 'saddlebrown', 'salmon', 'sandybrown', 'seagreen', 'seashell', 'sienna', 'skyblue', 'slateblue', \r\n 'slategray', 'slategrey', 'snow', 'springgreen', 'steelblue', 'tan', 'thistle', 'tomato', 'turquoise', 'violet', 'wheat', 'whitesmoke', 'yellowgreen', 'rebeccapurple']\r\n\r\n\r\n\r\n\r\ngr=st.selectbox('How you want to see Graph of different groups of population?',('Choose','Single','Cumulative'))\r\nif gr=='Single':\r\n typ = st.selectbox(\"Select the type of Chart\",('Choose',\"Line Chart\",\"Bar Chart\"))\r\n group=st.selectbox('Select Group of Population',df['Element'].unique().tolist())\r\n \r\n ca = alt.Chart(df[df[\"Element\"]== group]).encode(x=alt.X(\"Year\",scale=alt.Scale(df['Year'].min(),df['Year'].max())), y=alt.Y(\"Value\",scale=alt.Scale(df['Value'].min(),df['Value'].max())), \r\n tooltip=[\"Year\",\"Value\"]).interactive()\r\n\r\n \r\n if typ == \"Line Chart\":\r\n st.altair_chart(ca.mark_line(color=color_array[random.randint(0,150)]),use_container_width=True)\r\n else :\r\n st.altair_chart(ca.mark_bar(color=color_array[random.randint(0,150)]),use_container_width=True)\r\n \r\n\r\n\r\nelif gr=='Cumulative': \r\n options = st.multiselect(\r\n 'Select Multiple groups',\r\n df[\"Element\"][:5])\r\n \r\n \r\n\r\n fire=alt.Chart(df[df[\"Element\"].isin(options)]).mark_line().encode( x=alt.X(\"Year\"),\r\n y=alt.Y(\"Value\"),\r\n color=\"Element\",\r\n tooltip=[\"Year\",\"Value\"]).interactive()\r\n \r\n bar1 = alt.Chart(df[df[\"Element\"].isin(options)]).mark_bar().encode(\r\n y=\"Value\",\r\n x=alt.X(\"Element\",sort=\"-y\"),\r\n color=\"Element\",\r\n tooltip = [\"Year\",\"Value\"]\r\n).interactive()\r\n st.altair_chart(fire|bar1)\r\n ","repo_name":"foziahameed/streamlit_app","sub_path":"Pakistan_Population_App.py","file_name":"Pakistan_Population_App.py","file_ext":"py","file_size_in_byte":4265,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"73648818874","text":"from random import randint\n\n#file = open('27_A_7275.txt')\n#n, m = map(int, file.readline().split())\n\ndef f1():\n f = open('test')\n n, m = map(int, f.readline().split())\n summ = 0\n left = 0\n right = 0\n pref = [0] * n\n nums = [0] * n\n maxx = float('-inf')\n\n for i in range(n):\n n1, n2 = map(int, f.readline().split())\n n2 = (n2 + 29) // 30\n pref[i] = n2\n nums[i] = n1\n\n for i in range(n):\n while right < n and (nums[right] - nums[i]) <= m:\n summ += pref[right]\n right += 1\n while nums[i] - nums[left] > m:\n summ -= pref[left]\n left += 1\n maxx = max(maxx, summ)\n return(maxx)\n\ndef f2():\n fff = open('test')\n maxx = float('-inf')\n n, m = map(int, fff.readline().split())\n nums = [0] * n\n pref = [0] * n\n\n for i in range(n):\n n1, n2 = map(int, fff.readline().split())\n n2 = (n2 + 29) // 30\n pref[i] = n2\n nums[i] = n1\n\n for i in range(n):\n summ = 0\n for j in range(i, n):\n if (nums[j] - nums[i]) <= m:\n summ += pref[j]\n else:\n break\n for j in range(i - 1, -1, -1):\n if (nums[i] - nums[j]) <= m:\n summ += pref[j]\n else:\n break\n maxx = max(maxx, summ)\n return maxx\n\n\n\n\nwhile True:\n s = set()\n ff = open('test', 'w')\n last = 0\n for i in range(1001):\n if i == 0:\n ff.write(f'{1000} {randint(1, 1000)}\\n')\n else:\n nn = randint(1, 1000)\n s.add(last + nn)\n ff.write(f'{last + nn} {randint(1, 1000)}\\n')\n last += nn\n ff.close()\n\n if f1() != f2():\n print(f1(), f2(), 'false')","repo_name":"Propolisss/school_code_python","sub_path":"classwork/11_05_2023/2_2.py","file_name":"2_2.py","file_ext":"py","file_size_in_byte":1764,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"70288378876","text":"from __future__ import print_function\nimport os\nimport time\nimport glob\nfrom tkinter import *\n#import Tkinter, Tkconstants, tkFileDialog\nfrom threading import Lock\nfrom PIL import Image, ImageTk\nfrom tkinter.filedialog import askopenfilename \nfrom tkinter.filedialog import askdirectory \n\nfrom TestElasticSearch import NcsistSearchApi as InfinitySearchApi\n\n\nclass StopWatch():\n def __init__(self):\n # Get initial time in seconds since the Epoch.\n self.initial_time = time.time()\n self.last_mark_time = self.initial_time\n\n def mark(self, unit='sec'):\n \"\"\"\n Make time span since initial\n :param unit:\n :return:\n \"\"\"\n current_time = time.time()\n\n if unit == 'ms':\n time_span = (current_time - self.last_mark_time) * 1000\n else:\n time_span = current_time - self.last_mark_time\n\n self.last_mark_time = current_time\n return time_span\n\n def total_elapsed(self, unit='sec'):\n current_time = time.time()\n\n if unit == 'ms':\n time_span = (current_time - self.initial_time) * 1000\n else:\n time_span = current_time - self.initial_time\n return time_span\n\n def reset(self):\n self.initial_time = time.time()\n\n\nclass MainWindow(Frame):\n def __init__(self, master):\n sys.stdout = self # redirect stdout to self write()/flush()\n self.master = master\n self.frame = Frame(master)\n self.frame.pack()\n self.master.title('Infinity Search 1.0')\n self.lock_output = Lock()\n\n # GUI variables\n self.json_folder = StringVar()\n self.json_folder.set(r'/data1/JC_Sample/sample_data_only_face')\n self.query_file = StringVar()\n self.query_file.set(r'/data1/JC_Sample/query1.json')\n self.es_server = StringVar()\n self.es_server.set('hosts=10.10.53.201,10.10.53.204,10.10.53.207;port=9200;id=esadmin;passwd=esadmin@2018')\n self.search_field = StringVar()\n self.search_field.set('imgPath')\n self.search_text = StringVar()\n\n self.image_no = Image.open('no-image.png')\n self.image_no.thumbnail((100, 100), Image.ANTIALIAS)\n self.label_list_file = []\n self.label_list_image = []\n self.label_list_photo = []\n self.label_list_score = []\n self.image_path = r'/data1/JC_Sample/sample_data_only_face'\n\n self.update_id = StringVar()\n self.update_id.set('LMTSvmYBxjzL5lGNJfi-')\n self.update_field = StringVar()\n self.update_field.set('category')\n self.update_text = StringVar()\n self.update_text.set('face,wang')\n\n # ES variable\n self.es = None\n\n self.setup_widget()\n\n def setup_widget(self):\n # Panel layout\n Grid.rowconfigure(self.master, 0, weight=1)\n Grid.columnconfigure(self.master, 0, weight=1)\n self.frame.grid(row=0, column=0, sticky='NEWS')\n\n Grid.rowconfigure(self.frame, 0, weight=0)\n Grid.rowconfigure(self.frame, 1, weight=0)\n Grid.rowconfigure(self.frame, 2, weight=1)\n\n self.setup_widget_operation()\n self.setup_widget_result()\n self.setup_widget_output()\n\n def setup_widget_operation(self):\n optFrame = Frame(self.frame, borderwidth=1, relief=\"sunken\")\n Grid.rowconfigure(optFrame, 0, weight=0)\n Grid.columnconfigure(optFrame, 0, weight=0)\n Grid.columnconfigure(optFrame, 1, weight=4)\n Grid.columnconfigure(optFrame, 2, weight=4)\n Grid.columnconfigure(optFrame, 3, weight=0)\n\n # ES Server\n Label(optFrame, text='ES Conn Str:') \\\n .grid(row=0, column=0, sticky=(E), padx=3, pady=3)\n Entry(optFrame, width=60, textvariable=self.es_server) \\\n .grid(row=0, column=1, columnspan=2, sticky='NEWS', padx=3, pady=3)\n Button(optFrame, text=\"Connect\", command=self.cmd_connect_es) \\\n .grid(row=0, column=3, sticky='NEWS', padx=3, pady=3)\n # Import JSON selection\n Button(optFrame, text=\"Load:\", command=self.cmd_select_folder) \\\n .grid(row=1, column=0, sticky='NEWS', padx=3, pady=3)\n Entry(optFrame, width=60, textvariable=self.json_folder, state=DISABLED) \\\n .grid(row=1, column=1, columnspan=2, sticky='NEWS', padx=3, pady=3)\n Button(optFrame, text=\"Process\", command=self.cmd_load) \\\n .grid(row=1, column=3, sticky='NEWS', padx=3, pady=3)\n # Query\n Button(optFrame, text=\"Query:\", command=self.cmd_select_file) \\\n .grid(row=2, column=0, sticky='NEWS', padx=3, pady=3)\n Entry(optFrame, width=60, textvariable=self.query_file, state=DISABLED) \\\n .grid(row=2, column=1, columnspan=2, sticky='NEWS', padx=3, pady=3)\n Button(optFrame, text=\"Process\", command=self.cmd_query) \\\n .grid(row=2, column=3, sticky='NEWS', padx=3, pady=3)\n # Search\n Label(optFrame, text='Search:') \\\n .grid(row=3, column=0, sticky=(E), padx=3, pady=3)\n Entry(optFrame, width=20, textvariable=self.search_field) \\\n .grid(row=3, column=1, sticky='NEWS', padx=3, pady=3)\n Entry(optFrame, width=40, textvariable=self.search_text) \\\n .grid(row=3, column=2, sticky='NEWS', padx=3, pady=3)\n Button(optFrame, text=\"Process\", command=self.cmd_search) \\\n .grid(row=3, column=3, sticky='NEWS', padx=3, pady=3)\n # Update\n Label(optFrame, text='Update:') \\\n .grid(row=4, column=0, sticky=(E), padx=3, pady=3)\n Entry(optFrame, width=20, textvariable=self.update_id) \\\n .grid(row=4, column=1, sticky='NEWS', padx=3, pady=3)\n Entry(optFrame, width=40, textvariable=self.update_field) \\\n .grid(row=4, column=2, sticky='NEWS', padx=3, pady=3)\n Button(optFrame, text=\"Process\", command=self.cmd_update) \\\n .grid(row=4, column=3, sticky='NEWS', padx=3, pady=3)\n\n Entry(optFrame, width=40, textvariable=self.update_text) \\\n .grid(row=5, column=2, sticky='NEWS', padx=3, pady=3)\n\n # Clear All document\n commandFrame = Frame(optFrame, borderwidth=1)\n Button(commandFrame, text=\"Clear All\", command=self.cmd_clear_all_doc) \\\n .grid(row=0, column=0, sticky='NEWS', padx=3, pady=3)\n Button(commandFrame, text=\"Count\", command=self.cmd_count) \\\n .grid(row=0, column=1, sticky='NEWS', padx=3, pady=3)\n commandFrame \\\n .grid(row=6, column=0, columnspan=4, sticky='NEWS', padx=3, pady=3)\n\n optFrame \\\n .grid(row=0, column=0, sticky='NEWS', padx=3, pady=3)\n\n def setup_widget_result(self):\n # Query result panel\n picFrame = Frame(self.frame, borderwidth=1, relief=\"sunken\")\n Grid.rowconfigure(picFrame, 0, weight=0)\n Grid.rowconfigure(picFrame, 1, weight=1)\n Grid.rowconfigure(picFrame, 2, weight=0)\n Grid.columnconfigure(picFrame, 0, weight=1)\n Grid.columnconfigure(picFrame, 1, weight=1)\n Grid.columnconfigure(picFrame, 2, weight=1)\n Grid.columnconfigure(picFrame, 3, weight=1)\n Grid.columnconfigure(picFrame, 4, weight=1)\n\n for i in range(5):\n label = Label(picFrame, text='file')\n label.grid(row=0, column=i, sticky='EW', padx=3, pady=3)\n self.label_list_file.append(label)\n\n photo = ImageTk.PhotoImage(self.image_no)\n self.label_list_photo.append(photo)\n\n label = Label(picFrame, image=photo)\n label.grid(row=1, column=i, sticky='EW', padx=3, pady=3)\n self.label_list_image.append(label)\n\n label = Label(picFrame, text='0.0')\n label.grid(row=2, column=i, sticky='EW', padx=3, pady=3)\n self.label_list_score.append(label)\n\n picFrame \\\n .grid(row=1, column=0, sticky='NEWS', padx=3, pady=3)\n\n def setup_widget_output(self):\n # Message output panel\n txtFrame = Frame(self.frame, borderwidth=1, relief=\"sunken\")\n self.txt_output = Text(txtFrame, wrap=NONE, font='arial 9', height=16, borderwidth=0)\n vscroll = Scrollbar(txtFrame, orient=VERTICAL, command=self.txt_output.yview)\n self.txt_output['yscroll'] = vscroll.set\n vscroll.pack(side=\"right\", fill=\"y\")\n self.txt_output.pack(side=\"left\", fill=\"both\", expand=True)\n txtFrame \\\n .grid(row=2, column=0, sticky='NEWS', padx=3, pady=3)\n self.txt_output.insert(INSERT, 'Program started\\n')\n\n def write(self, text):\n self.lock_output.acquire()\n try:\n self.txt_output.insert(INSERT, text)\n self.txt_output.see(END)\n finally:\n self.lock_output.release()\n\n def flush(self):\n pass\n\n def nonblock_stdout(self, out, thd_id = 0):\n for line in iter(out.readline, b''):\n line = line.decode('cp950')\n if len(line) > 4:\n print('\\t({0}): {1}'.format(thd_id, line), end='')\n out.close()\n\n def nonblock_stderr(self, out, thd_id = 0):\n for line in iter(out.readline, b''):\n line = line.decode('cp950')\n if len(line) > 4:\n print('\\t({0}): {1}'.format(thd_id, line), end='')\n out.close()\n\n def clear_output(self):\n self.txt_output.delete('1.0', END)\n\n def display_result(self, pic_list):\n if pic_list:\n for i in range(5):\n img_file = os.path.join(self.image_path, pic_list[i]['_source']['imgPath'])\n dir_name = os.path.basename(os.path.dirname(img_file))\n self.label_list_file[i].configure(text=dir_name)\n image = Image.open(img_file)\n image.thumbnail((100, 100), Image.ANTIALIAS)\n photo = ImageTk.PhotoImage(image)\n self.label_list_photo[i] = photo\n self.label_list_image[i].configure(image=self.label_list_photo[i])\n score = '{0:7.6f}'.format(pic_list[i]['_score'])\n self.label_list_score[i].configure(text=score)\n else:\n for i in range(5):\n self.label_list_file[i].configure(text='file')\n self.label_list_photo[i] = ImageTk.PhotoImage(self.image_no)\n self.label_list_image[i].configure(image=self.label_list_photo[i])\n self.label_list_score[i].configure(text='0.0')\n pass\n\n def cmd_connect_es(self):\n try:\n stopwatch = StopWatch()\n self.clear_output()\n self.es = InfinitySearchApi.InfinitySearch(self.es_server.get())\n status = self.es.status()\n print('Elapsed: {0:6.2f} sec(s)'.format(stopwatch.mark()))\n print(status)\n except Exception as ex:\n print(ex)\n\n def cmd_select_folder(self):\n folder = askdirectory()\n if folder:\n self.json_folder.set(folder)\n\n def cmd_select_file(self):\n file = askopenfilename(title=\"Select file\", filetypes=((\"json files\", \"*.json\"), (\"all files\", \"*.*\")))\n if file:\n self.query_file.set(file)\n\n def cmd_load(self):\n folder = self.json_folder.get()\n stopwatch = StopWatch()\n if folder and self.es:\n files = glob.glob(os.path.join(folder, '*.json'))\n files.sort()\n for f in files:\n self.es.load_data(f)\n print('Elapsed: {0:6.2f} sec(s)'.format(stopwatch.mark()))\n\n def cmd_count(self):\n result = self.es.count()\n print(result)\n\n def cmd_clear_all_doc(self):\n if self.es:\n status = self.es.delete_all()\n print(status)\n pass\n\n def cmd_query(self):\n self.clear_output()\n self.display_result(None)\n file = self.query_file.get()\n stopwatch = StopWatch()\n if file and self.es:\n result = self.es.query_result(file)\n self.display_result(result)\n print('Elapsed: {0:6.2f} sec(s)'.format(stopwatch.mark()))\n\n def cmd_search(self):\n self.clear_output()\n self.display_result(None)\n field = self.search_field.get()\n search_text = self.search_text.get()\n stopwatch = StopWatch()\n if self.es and field.__len__() > 0 and search_text.__len__() > 0:\n result = self.es.search_result(field, search_text)\n self.display_result(result)\n print('Elapsed: {0:6.2f} sec(s)'.format(stopwatch.mark()))\n\n def cmd_update(self):\n self.clear_output()\n self.display_result(None)\n id_no = self.update_id.get()\n field = self.update_field.get()\n text = self.update_text.get()\n stopwatch = StopWatch()\n if self.es and len(id_no) > 0 and len(field) > 0 and len(text) > 0:\n result = self.es.update_fields(id_no, field, text)\n print('Elapsed: {0:6.2f} sec(s)'.format(stopwatch.mark()))\n\n\ndef main():\n root = Tk()\n MainWindow(root)\n root.mainloop()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"SWhite4han/JC_Demo","sub_path":"TestElasticSearch/InfinitySearch.py","file_name":"InfinitySearch.py","file_ext":"py","file_size_in_byte":13036,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"25414143381","text":"from common_logic import CommonLogic\n\n\ndef get_count(arr):\n count = 0\n for x in arr:\n x = x.replace(\" \", \"\")\n myset = set(x)\n count += len(myset)\n return count\n\n\ndef get_count_second_question(arr):\n count = 0\n for x in arr:\n line = x.split(' ')\n for y in line[0]:\n if (x.count(y)) == len(line):\n count +=1\n return count\n\n\nif __name__ == '__main__':\n print(get_count_second_question(CommonLogic.get_input_as_batches()))\n","repo_name":"stubz151/Advent-Of-code","sub_path":"day6/day6.py","file_name":"day6.py","file_ext":"py","file_size_in_byte":501,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"38162096429","text":"from collections import namedtuple\nfrom typing import List\nimport numpy as np\n\nBaseExperienceReplay = namedtuple(\n \"BaseExperienceReplay\", [\"state_tiles\", \"action\", \"successor_tiles\", \"reward\"]\n)\n# action is encoded as an int in 0..3\nACTIONS = [\"Up\", \"Down\", \"Left\", \"Right\"]\n\n\nclass ExperienceReplay(BaseExperienceReplay):\n \"\"\"\n An experience replay tuple with helper functions to convert to and from\n a flattened representation (e.g. for a neural net).\n \"\"\"\n\n def __repr__(self):\n return (\n f\"Previous state\\n{tiles_repr(self.state_tiles)}\\n\"\n f\"Action: {int(self.action)} ({ACTIONS[int(self.action)]})\\n\"\n f\"New state\\n{tiles_repr(self.successor_tiles)}\\n\"\n f\"Reward: {self.reward}\"\n )\n\n @property\n def state_bitarray(self) -> np.ndarray:\n return convert_tiles_to_bitarray(self.state_tiles)\n\n @property\n def successor_bitarray(self) -> np.ndarray:\n return convert_tiles_to_bitarray(self.successor_tiles)\n\n def flatten(self) -> np.ndarray:\n \"\"\"stores the (s, a, s', r) in a flattened representation\"\"\"\n return np.hstack(\n (\n self.state_bitarray,\n np.array([self.action]),\n self.successor_bitarray,\n np.array([self.reward]),\n )\n )\n\n @classmethod\n def from_flattened(cls, flat_tuple: np.ndarray):\n assert flat_tuple.shape == (2 * 16 * 17 + 2,)\n\n current_state_bitarray = flat_tuple[0 : (16 * 17)]\n action = flat_tuple[16 * 17]\n new_state_bitarray = flat_tuple[(16 * 17 + 1) : (2 * 16 * 17 + 1)]\n reward = flat_tuple[(2 * 16 * 17 + 1)]\n\n current_state_tiles = convert_bitarray_to_tiles(current_state_bitarray)\n new_state_tiles = convert_bitarray_to_tiles(new_state_bitarray)\n return cls(current_state_tiles, action, new_state_tiles, reward)\n\n\ndef convert_tiles_to_bitarray(tiles) -> np.ndarray:\n \"\"\"\n Convert from a 4x4 array, where each cell is the log base 2 value of the tile,\n into a flattened bitarray representation, where each of the 16 cells is represented by 17 bits,\n with the first bit set if the tile value is 2, the second bit set in the tile value is 4,\n and so on up to 2^17 (the maximum possible tile value on a 4x4 board with 4-tiles being\n the maximum possible spawned tile).\n \"\"\"\n flat_tiles = np.ravel(tiles)\n bitarray_input = np.zeros((16, 17))\n for i in range(16):\n if flat_tiles[i] != 0:\n # value of 1 (means the the tile is 2) should set bit 0 in bitarray\n bitarray_input_idx = flat_tiles[i] - 1\n bitarray_input[i, bitarray_input_idx] = 1\n return np.ravel(bitarray_input)\n\n\ndef convert_bitarray_to_tiles(bitarray: np.ndarray) -> list:\n \"\"\"\n Convert from flattened bitarray representation, where each of the 16 cells is\n represented by 17 bits (first bit is set if tile value is 2), to a 4x4 array,\n where each cell is the log base 2 value of the tile.\n \"\"\"\n assert bitarray.size == 16 * 17\n bitarray_reshape = np.reshape(bitarray, (4, 4, 17))\n tiles = []\n for r in range(4):\n tile_row = []\n for c in range(4):\n one_hot_tile_encoding = bitarray_reshape[r, c]\n if np.count_nonzero(one_hot_tile_encoding) == 0:\n tile_row.append(0)\n else:\n assert np.sum(one_hot_tile_encoding) == 1\n assert np.count_nonzero(one_hot_tile_encoding) == 1\n tile_row.append(np.argmax(one_hot_tile_encoding) + 1)\n tiles.append(tile_row)\n return tiles\n\n\n# def parse_flattened_experience_tuple(flat_tuple: np.ndarray):\n# assert flat_tuple.shape == (2 * 16 * 17 + 2,)\n\n# current_state_bitarray = flat_tuple[0:(16 * 17)]\n# action = flat_tuple[16 * 17]\n# new_state_bitarray = flat_tuple[(16 * 17 + 1):(2 * 16 * 17 + 1)]\n# reward = flat_tuple[(2 * 16 * 17 + 1)]\n# return (current_state_bitarray, action, new_state_bitarray, reward)\n\n\ndef tiles_repr(tiles: List[List[int]]) -> str:\n output = \"\"\n for row_idx in range(len(tiles)):\n row = tiles[row_idx]\n for col_idx in range(len(row)):\n output += f\"{row[col_idx]:2d}\"\n if col_idx < len(row) - 1:\n output += \" \"\n if row_idx < len(tiles) - 1:\n output += \"\\n\"\n return output\n","repo_name":"alexander-jiang/RL2048","sub_path":"reinforcement_learning/experience_replay_utils.py","file_name":"experience_replay_utils.py","file_ext":"py","file_size_in_byte":4399,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"19881405189","text":"from typing import List\nclass Solution:\n # 哈希映射\n def findLHS(self, nums: List[int]) -> int:\n numDict = dict()\n numList = list()\n for i in nums:\n if i in numDict:\n numDict[i] += 1\n else:\n numDict[i] = 1\n for j in list(numDict.keys()):\n if j + 1 in list(numDict.keys()):\n numList.append(numDict[j] + numDict[j + 1])\n if len(numList) == 0:\n return 0\n return max(numList)\n\n\nif __name__==\"__main__\":\n nums = [1, 3, 2, 2, 5, 2, 3, 7]\n method = Solution()\n answer = method.findLHS(nums)\n print(answer)","repo_name":"Lee8150951/LeetCode-Python","sub_path":"Hash Map/594.py","file_name":"594.py","file_ext":"py","file_size_in_byte":567,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"11761492329","text":"import os\nfrom collections import deque\n\n\nclass Process:\n def __init__(self, process_id, arrival_time, burst_time, priority):\n self.process_id = process_id\n self.arrival_time = arrival_time\n self.burst_time = burst_time\n self.priority = priority\n self.completion_time = 0\n self.waiting_time = 0\n self.turnaround_time = 0\n self.remaining_time = burst_time\n\n\ndef fcfs(processes):\n processes.sort(key=lambda x: x.arrival_time) # Sắp xếp tiến trình theo thời gian xuất hiện\n curr_time = 0\n for process in processes:\n if process.arrival_time > curr_time:\n curr_time = process.arrival_time\n process.completion_time = curr_time + process.burst_time\n process.turnaround_time = process.completion_time - process.arrival_time\n process.waiting_time = process.turnaround_time - process.burst_time\n curr_time = process.completion_time\n return processes\n\n\ndef sjf_non_preemptive(processes):\n processes.sort(\n key=lambda x: (x.burst_time, x.arrival_time)) # Sắp xếp tiến trình theo burst time và thời gian xuất hiện\n curr_time = 0\n for process in processes:\n if process.arrival_time > curr_time:\n curr_time = process.arrival_time\n process.completion_time = curr_time + process.burst_time\n process.turnaround_time = process.completion_time - process.arrival_time\n process.waiting_time = process.turnaround_time - process.burst_time\n curr_time = process.completion_time\n return processes\n\n\ndef round_robin(processes, time_quantum):\n curr_time = 0\n queue = deque(processes)\n completed_processes = []\n while queue:\n process = queue.popleft()\n if process.remaining_time <= time_quantum:\n curr_time += process.remaining_time\n process.remaining_time = 0\n process.completion_time = curr_time\n process.turnaround_time = process.completion_time - process.arrival_time\n process.waiting_time = process.turnaround_time - process.burst_time\n completed_processes.append(process)\n else:\n curr_time += time_quantum\n process.remaining_time -= time_quantum\n queue.append(process)\n return completed_processes\n\n\ndef priority_preemptive(processes):\n def find_max_prior_arrived(currTime, lst):\n available_processes = [process for process in lst if\n process.arrival_time <= currTime and process.remaining_time > 0]\n if available_processes:\n available_processes.sort(key=lambda x: x.priority)\n return available_processes[0], True\n return None, False\n\n def all_done(lst):\n return all(process.remaining_time == 0 for process in lst)\n\n curr_time = 0\n while not all_done(processes):\n process, has_arrived = find_max_prior_arrived(curr_time, processes)\n if not has_arrived:\n curr_time += 1\n continue\n\n process.remaining_time -= 1\n curr_time += 1\n if process.remaining_time == 0:\n process.completion_time = curr_time\n process.turnaround_time = process.completion_time - process.arrival_time\n process.waiting_time = process.turnaround_time - process.burst_time\n\n return processes\n\n\ndef priority_non_preemptive(processes):\n processes.sort(key=lambda x: (x.priority, x.arrival_time)) # Sắp xếp tiến trình theo ưu tiên và thời gian xuất hiện\n curr_time = 0\n for process in processes:\n if process.arrival_time > curr_time:\n curr_time = process.arrival_time\n process.completion_time = curr_time + process.burst_time\n process.turnaround_time = process.completion_time - process.arrival_time\n process.waiting_time = process.turnaround_time - process.burst_time\n curr_time = process.completion_time\n return processes\n\ndef srtf_preemptive(processes):\n processes.sort(key=lambda x: x.arrival_time) # Sort processes by arrival time\n current_time = 0\n remaining_processes = list(processes)\n completed_processes = []\n\n while remaining_processes:\n eligible_processes = [p for p in remaining_processes if p.arrival_time <= current_time]\n if not eligible_processes:\n current_time += 1\n continue\n\n # Find the process with the shortest remaining burst time\n shortest_process = min(eligible_processes, key=lambda x: x.remaining_time)\n\n # Execute the process for one time unit\n shortest_process.remaining_time -= 1\n current_time += 1\n\n # Check if the process has completed\n if shortest_process.remaining_time == 0:\n shortest_process.completion_time = current_time\n shortest_process.turnaround_time = shortest_process.completion_time - shortest_process.arrival_time\n shortest_process.waiting_time = shortest_process.turnaround_time - shortest_process.burst_time\n completed_processes.append(shortest_process)\n remaining_processes.remove(shortest_process)\n\n return completed_processes\n\ndef average_turnaround_time(processes):\n if not processes:\n return 0\n return sum(process.turnaround_time for process in processes) / len(processes)\n\n\ndef average_waiting_time(processes):\n if not processes:\n return 0\n return sum(process.waiting_time for process in processes) / len(processes)\n\n\ndef generate_gantt_chart(processes):\n gantt_chart = \" \"\n curr_time = 0\n for process in processes:\n gantt_chart += \"|\"\n for i in range(process.completion_time - curr_time):\n gantt_chart += \"-\"\n gantt_chart += f\"P{process.process_id}\"\n curr_time = process.completion_time\n\n gantt_chart += \"|\"\n return gantt_chart\n\n\ndef get_processes_from_file(file_path):\n processes = []\n with open(file_path, 'r') as file:\n lines = file.readlines()\n for line in lines:\n process_data = line.split()\n if len(process_data) == 4:\n if process_data[0] != None:\n process_id, arrival_time, burst_time = process_data\n process = Process(process_id, int(arrival_time), int(burst_time), int(priority))\n else:\n process_id, arrival_time, burst_time, priority = map(int, process_data)\n process = Process(process_id, arrival_time, burst_time, priority)\n elif len(process_data) == 3:\n if process_data[0] != None:\n process_id, arrival_time, burst_time = process_data\n process = Process(process_id, int(arrival_time), int(burst_time), '')\n else:\n process_id, arrival_time, burst_time = map(int, process_data)\n process = Process(process_id, arrival_time, burst_time, '')\n else:\n print(f\"Invalid data in line: {line}\")\n continue\n processes.append(process)\n return processes\n\n\ndef write_results_to_file(file_path, avg_waitingTime, avg_turnaroundTime, input_file, algorithm_type):\n with open(file_path, 'a') as file:\n file.write(f\"Results using {algorithm_type}:\\n\")\n file.write(f\"Average Waiting Time: {avg_waitingTime}\\n\")\n file.write(f\"Average Turnaround Time: {avg_turnaroundTime}\\n\")\n file.write(f\"Input File: {input_file.upper()}\\n\")\n file.write(\"\\n\")\n\n\ndef sort_output_file(file_path):\n with open(file_path, 'r') as file:\n lines = file.readlines()\n\n algorithms = []\n algorithm_data = {}\n current_algorithm = None\n\n for line in lines:\n if line.startswith(\"Results using\"):\n if current_algorithm:\n algorithms.append((current_algorithm, algorithm_data))\n current_algorithm = line.strip().split(\" using \")[1][:-1]\n algorithm_data = {\"avg_waiting_time\": None, \"avg_turnaround_time\": None, \"input_file\": None}\n elif line.startswith(\"Average Waiting Time:\"):\n algorithm_data[\"avg_waiting_time\"] = float(line.strip().split(\": \")[1])\n elif line.startswith(\"Input File:\"):\n algorithm_data[\"input_file\"] = line.strip().split(\": \")[1]\n\n algorithms.append((current_algorithm, algorithm_data))\n sorted_algorithms = sorted(algorithms, key=lambda x: x[1][\"avg_waiting_time\"])\n\n with open(file_path, 'w') as file:\n for algorithm, data in sorted_algorithms:\n file.write(f\"--------------------------------------\\n\")\n file.write(f\"Results using {algorithm}:\\n\")\n file.write(f\"Average Waiting Time: {data['avg_waiting_time']}\\n\")\n file.write(f\"Input File: {data['input_file']}\\n\")\n file.write(\"\\n\")\n\n\ndef compare_algorithms():\n print(\"Compare Algorithms\")\n num_algorithms = int(input(\"Enter the number of algorithms to compare: \"))\n\n chosen_algorithms = []\n for i in range(num_algorithms):\n print(f\"\\nChoose algorithm {i + 1}:\")\n scheduling_algorithm, algorithm_type = choose_algorithm()\n if scheduling_algorithm is None:\n return None\n chosen_algorithms.append((scheduling_algorithm, algorithm_type))\n\n file_name = choose_input_file()\n list_processes = get_processes_from_file(file_name)\n output_file_name = input(\"Enter the output file name (Example: compare_results.txt): \")\n results = []\n for algorithm, algorithm_type in chosen_algorithms:\n result = algorithm(list_processes)\n avg_turnaroundTime = average_turnaround_time(result)\n avg_waitingTime = average_waiting_time(result)\n results.append((algorithm_type, result, avg_waitingTime, avg_turnaroundTime))\n\n print(f\"\\nResult Table {algorithm_type}:\")\n print(f\"ProcessID\\tArrivalTime\\tBurstTime\\tPriority\\tCompletionTime\\tTurnaroundTime\\tWaitingTime\")\n for process in result:\n if hasattr(process, 'priority'):\n print(\n f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\t{process.priority}\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n else:\n print(\n f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\tNo Priority\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n with open(output_file_name, 'a') as file:\n file.write(f\"-------------------- {file_name.upper()}--------------------\\n\")\n file.write(f\"Comparing {num_algorithms} algorithms for {file_name}:\\n\")\n for algorithm_type, _, avg_waitingTime, avg_turnaroundTime in results:\n print(f\"------------------------------------\\n\")\n file.write(f\"Algorithm: {algorithm_type}\\n\")\n print(f\"Algorithm: {algorithm_type}\\n\")\n file.write(f\"Average Waiting Time: {avg_waitingTime}\\n\")\n print(f\"Average Waiting Time: {avg_waitingTime}\\n\")\n file.write(f\"Average Turnaround Time: {avg_turnaroundTime}\\n\")\n print(f\"Average Turnaround Time: {avg_turnaroundTime}\\n\")\n file.write(\"\\n\")\n\n\n# def compare_algorithms():\n# print(\"Compare Algorithms\")\n# num_algorithms = int(input(\"Enter the number of algorithms to compare: \"))\n#\n# chosen_algorithms = []\n# for i in range(num_algorithms):\n# print(f\"\\nChoose algorithm {i + 1}:\")\n# scheduling_algorithm, algorithm_type = choose_algorithm()\n# if scheduling_algorithm is None:\n# return None\n# chosen_algorithms.append((scheduling_algorithm, algorithm_type))\n#\n# file_name = choose_input_file()\n#\n# list_processes = get_processes_from_file(file_name)\n#\n# output_file_name = input(\"Enter the output file name (Example: compare_results.txt): \")\n#\n# avg_waiting_times = []\n# avg_turnaround_times = []\n#\n# with open(output_file_name, 'a') as output_file:\n# output_file.write(f\"Comparison Results for {file_name}:\\n\")\n#\n# for algorithm, algorithm_type in chosen_algorithms:\n# result = algorithm(list_processes)\n# avg_turnaround_time = average_turnaround_time(result)\n# avg_waiting_time = average_waiting_time(result)\n# avg_waiting_times.append(avg_waiting_time)\n# avg_turnaround_times.append(avg_turnaround_time)\n#\n# output_file.write(f\"\\nAlgorithm: {algorithm_type}\\n\")\n# output_file.write(f\"Average Waiting Time: {avg_waiting_time}\\n\")\n# output_file.write(f\"Average Turnaround Time: {avg_turnaround_time}\\n\")\n#\n# print(f\"\\n{algorithm_type}:\")\n# print(f\"Average Waiting Time: {avg_waiting_time}\")\n# print(f\"Average Turnaround Time: {avg_turnaround_time}\")\n#\n# print(f\"Results saved to {output_file_name}\")\n\n# ... Rest of the code ...\n\n# def compare_algorithms():\n# print(\"Compare Algorithms\")\n# num_algorithms = int(input(\"Enter the number of algorithms to compare: \"))\n#\n# chosen_algorithms = []\n# for i in range(num_algorithms):\n# print(f\"\\nChoose algorithm {i + 1}:\")\n# scheduling_algorithm, algorithm_type = choose_algorithm()\n# if scheduling_algorithm is None:\n# return None\n# chosen_algorithms.append((scheduling_algorithm, algorithm_type))\n#\n# file_name = choose_input_file()\n#\n# list_processes = get_processes_from_file(file_name)\n#\n# for algorithm, algorithm_type in chosen_algorithms:\n# result = algorithm(list_processes)\n# avg_turnaroundTime = average_turnaround_time(result)\n# avg_waitingTime = average_waiting_time(result)\n#\n# print(f\"\\nResult Table {algorithm_type}:\")\n# print(f\"ProcessID\\tArrivalTime\\tBurstTime\\tPriority\\tCompletionTime\\tTurnaroundTime\\tWaitingTime\")\n# for process in result:\n# if hasattr(process, 'priority'):\n# print(\n# f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\t{process.priority}\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n# else:\n# print(\n# f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\tNo Priority\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n#\n# with open('compare.txt', 'a') as file:\n# file.write(f\"Algorithm: {algorithm_type}\\n\")\n# file.write(f\"Average Waiting Time: {avg_waitingTime}\\n\")\n# file.write(f\"Average Turnaround Time: {avg_turnaroundTime}\\n\\n\")\n#\n# print(f\"\\n{algorithm_type}:\")\n# print(f\"Average Waiting Time: {avg_waitingTime}\")\n# print(f\"Average Turnaround Time: {avg_turnaroundTime}\")\n\n# ... Rest of the code ...\ndef get_input_files_from_directory(directory_path):\n input_files = []\n for file_name in os.listdir(directory_path):\n if file_name.endswith(\".txt\"): # Chỉ xem xét các file có đuôi là .txt\n input_files.append(os.path.join(\"\", file_name))\n return input_files\n\n\n# Sử dụng hàm get_input_files_from_directory để lấy danh sách các file input từ một thư mục cụ thể\ndirectory_path = \"C:/Users/ADMIN/PycharmProjects/pythonCPU\"\ninput_files = get_input_files_from_directory(directory_path)\n\ndef read_file(file_name):\n try:\n with open(file_name, 'r') as file:\n content = file.read()\n print(content)\n except FileNotFoundError:\n print(f\"Error: File '{file_name}' not found.\")\ndef choose_input_file():\n print(\"Choose an input file:\")\n print(\"1. input1.txt\")\n print(\"2. input2.txt\")\n print(\"3. input3.txt\")\n print(\"4. input4.txt\")\n print(\"5. input5.txt\")\n print(\"6. input6.txt\")\n print(\"7. input7.txt\")\n choice = input(\"Enter your choice: \")\n\n if choice == '1':\n return \"input1.txt\"\n elif choice == '2':\n return \"input2.txt\"\n elif choice == '3':\n return \"input3.txt\"\n elif choice == '4':\n return \"input4.txt\"\n elif choice == '5':\n return \"input5.txt\"\n elif choice == '6':\n return \"input6.txt\"\n elif choice == '7':\n return \"input7.txt\"\n else:\n print(\"Invalid choice. Using default file input1.txt.\")\n return \"input1.txt\"\n\n\ndef choose_algorithm():\n print(\"Choose a scheduling algorithm:\")\n print(\"1. FCFS (First Come First Serve)\")\n print(\"2. SJF (Shortest Job First)\")\n print(\"3. RR (Round Robin)\")\n print(\"4. SRTF Preemptive (Shortest Remaining Time First)\")\n print(\"5. Priority Preemptive\")\n print(\"6. Priority Non_Preemptive\")\n print(\"7. Quit\")\n choice = input(\"Enter your choice: \")\n\n if choice == '1':\n return fcfs, \"FCFS (First Come First Serve)\"\n elif choice == '2':\n return sjf_non_preemptive, \"SJF (Shortest Job First)\"\n elif choice == '3':\n time_quantum = int(input(\"Enter time quantum for Round Robin('example:4'): \"))\n return lambda processes: round_robin(processes,\n time_quantum), f\"RR (Round Robin) with Time Quantum {time_quantum}\"\n elif choice == '4':\n return srtf_preemptive, \"SRTF Preemptive (Shortest Remaining Time First)\"\n elif choice == '5':\n return priority_preemptive, \"Priority Preemptive\"\n elif choice == '6':\n return priority_non_preemptive, \"Priority Non_Preemptive\"\n elif choice == '7':\n return None, None\n else:\n print(\"Invalid choice. Using default Priority Preemptive.\")\n return fcfs, \"FCFS (First Come First Serve)\"\n\ndef main():\n while True:\n print(\"\\nCPU SCHEDULING ALGORITHM:\")\n print(\"1. Run a Single Algorithm\")\n print(\"2. Compare Algorithms\")\n print(\"3. Display Result\")\n print(\"4. Quit\")\n choice = input(\"Enter your choice: \")\n\n if choice == '1':\n scheduling_algorithm, algorithm_type = choose_algorithm()\n if scheduling_algorithm is not None:\n file_name = choose_input_file()\n list_processes = get_processes_from_file(file_name)\n result = scheduling_algorithm(list_processes)\n avg_turnaroundTime = average_turnaround_time(list_processes)\n avg_waitingTime = average_waiting_time(list_processes)\n print(f\"\\nResult Table {algorithm_type}:\")\n print(f\"ProcessID\\tArrivalTime\\tBurstTime\\tPriority\\tCompletionTime\\tTurnaroundTime\\tWaitingTime\")\n for process in result:\n if hasattr(process, 'priority'):\n print(\n f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\t{process.priority}\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n else:\n print(\n f\"{process.process_id}\\t\\t\\t{process.arrival_time}\\t\\t\\t{process.burst_time}\\t\\t\\tNo Priority\\t\\t\\t{process.completion_time}\\t\\t\\t\\t{process.turnaround_time}\\t\\t\\t\\t{process.waiting_time}\")\n print(f\"\\nAverage Waiting Time: {avg_waitingTime}\")\n print(f\"Average Turnaround Time: {avg_turnaroundTime}\")\n print(\"\\nGantt Chart: \\n\")\n gantt_chart = generate_gantt_chart(result)\n print(gantt_chart)\n print(\n \"\\n--------------------------------------------------------------------------------------------------\\n\")\n output_file = \"output.txt\"\n write_results_to_file(output_file, avg_waitingTime, avg_turnaroundTime, file_name, algorithm_type)\n sort_output_file(output_file)\n\n elif choice == '2':\n compare_algorithms()\n elif choice == '3':\n # break\n print(\"All Files in the Directory:\")\n for file_path in input_files:\n print(file_path)\n print(\"-------------------\")\n file_name = input(\"Open Files name : \")\n read_file(file_name)\n elif choice == '4':\n print(\"Goodbye 😊😊\")\n break\n else:\n print(\"Invalid choice. Please try again.\")\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"CassanoKEDOKATO/Priority-based-CPU-Scheduling-Algorithms","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":20538,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"25396943528","text":"def test(rooms):\n roomNumberToRoom = {}\n for roomNumber, room in enumerate(rooms, 1):\n roomNumberToRoom[roomNumber] = room\n\n roomNumberAndRooms = list(sorted(roomNumberToRoom.items(), key=lambda item: -item[1]))\n\n papers = []\n for roomNumber, room in roomNumberAndRooms:\n for _ in range(room):\n papers.append(roomNumber)\n \n students = []\n for roomNumber, room in roomNumberAndRooms[1:]:\n for _ in range(room):\n students.append(roomNumber)\n\n for roomNumber, room in roomNumberAndRooms[0:1]:\n for _ in range(room):\n students.append(roomNumber)\n\n for paper, student in zip(papers, students):\n if paper == student:\n return None\n\n return [roomNumber for roomNumber, _ in roomNumberAndRooms]\n\n\nnumberOfRooms = int(input())\nrooms = list(map(int, input().split()))\nresult = test(rooms)\nif result == None:\n print(\"impossible\")\nelse:\n print(*result)\n","repo_name":"KentGrigo/Kattis","sub_path":"python/redistribution.py","file_name":"redistribution.py","file_ext":"py","file_size_in_byte":958,"program_lang":"python","lang":"en","doc_type":"code","stars":14,"dataset":"github-code","pt":"96"} +{"seq_id":"11057349767","text":"from tkinter import *\n\nfrom random import randint\n\nw = Tk()\nw.title(\"Stone Paper Scissor\")\nw.configure(background=\"black\")\n\nimage_pstone= PhotoImage(name=\"stone1\",file=\"i\\st1.png\")\nimage_pscissor= PhotoImage(name=\"scissor1\",file=\"i\\s1.png\")\nimage_ppaper= PhotoImage(name=\"paper1\",file=\"i\\p1.png\")\nimage_cstone= PhotoImage(name=\"stone2\",file=\"i\\st2.png\")\nimage_cscissor= PhotoImage(name=\"scissor2\",file=\"i\\s2.png\")\nimage_cpaper= PhotoImage(name=\"paper2\",file=\"i\\p2.png\")\n\nplabel=Label(w,image=image_pstone)\n\nclabel=Label(w,image=image_cstone)\n\nplabel.grid(row=1,column=0)\nclabel.grid(row=1,column=4)\n\n\npindicate=Label(w,text=\"PLAYER\",font=(\"papyrus\",20,\"bold\"),fg=\"light blue\",bg=\"black\").grid(row=0,column=0)\ncindicate=Label(w,text=\"COMPUTER \",font=(\"papyrus\",20,\"bold\"),fg=\"light blue\",bg=\"black\").grid(row=0,column=4)\nvsindicate=Label(w,text=\"V/S\",font=(\"algerian\",50,\"bold\"),bg=\"black\",fg=\"dark red\").grid(row=1,column=2)\n\n\n\n\ndef wincondition(p,c):\n\tif(p==c):\n\t\twin.config(text=\"It's a Tie\")\n\telif p==\"stone\" :\n\t\tif c==\"paper\":\n\t\t\twin.config(text=\"Computer Wins!!!\")\n\t\t\t \n\t\telse:\n\t\t\twin.config(text=\"Player Wins!!!\")\n\t\t\t\n\telif p==\"paper\" :\n\t\tif c==\"scissor\":\n\t\t\twin.config(text=\"Computer Wins!!!\")\n\t\t\t\n\t\telse:\n\t\t\twin.config(text=\"Player Wins!!!\")\n\t\t\t\n\telif p==\"scissor\" :\n\t\tif c==\"stone\":\n\t\t\twin.config(text=\"Computer Wins!!!\")\n\t\t\t\n\t\telse:\n\t\t\twin.config(text=\"Player Wins!!!\")\n\t\t\t\n\t\t\nselect=[\"stone\",\"paper\",\"scissor\"]\n\t\t\n\ndef input(a):\n\tin_c=select[randint(0,2)]\n\tif in_c==\"stone\":\n\t\tclabel.configure(image=image_cstone)\n\telif in_c==\"scissor\":\n\t\tclabel.configure(image=image_cscissor)\n\telse:\n\t\tclabel.configure(image=image_cpaper)\n\t\n\tif a==\"stone\":\n\t\tplabel.configure(image=image_pstone)\n\telif a==\"scissor\":\n\t\tplabel.configure(image=image_pscissor)\n\telse:\n\t\tplabel.configure(image=image_ppaper)\n\t\n\twincondition(a, in_c)\n\nwin = Label(w, text=\"\", font=('Times New Roman', 30))\nwin.config(fg=\"gold\",bg=\"dark blue\", font=('Times New Roman', 30, 'bold'))\nwin.grid(row=5, column=2)\n\nbutton_stone=Button(w,width=10,height=3,text=\"Stone\",font=(\"algerian\",18,\"bold\"),bg=\"silver\",fg=\"red\",relief=\"sunken\",command=lambda:input(\"stone\")).grid(row=4,column=1)\nbutton_paper=Button(w,width=10,height=3,text=\"Paper\",font=(\"algerian\",18,\"bold\"),bg=\"silver\",fg=\"red\",relief=\"sunken\",command=lambda:input(\"paper\")).grid(row=4,column=2)\nbutton_scissor=Button(w,width=10,height=3,text=\"Scissor\",font=(\"algerian\",18,\"bold\"),bg=\"silver\",fg=\"red\",relief=\"sunken\",command=lambda:input(\"scissor\")).grid(row=4,column=3)\n\nw.mainloop()\n\n","repo_name":"ad9587/Stone-Paper-Scissor-Game","sub_path":"main code/spc.py","file_name":"spc.py","file_ext":"py","file_size_in_byte":2513,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"40715024834","text":"#!/usr/bin/python\n\"\"\"\\\nSimple g-code streaming script\n\"\"\"\n \nimport serial\nimport time\nimport argparse\n\n# parser = argparse.ArgumentParser(description='This is a basic gcode sender. http://crcibernetica.com')\n# parser.add_argument('-p','--port',help='Input USB port',required=True)\n# parser.add_argument('-f','--file',help='Gcode file name',required=True)\n# args = parser.parse_args()\n\n# argPort = args.port\n# argFile = args.file\n\nargPort = 'COM3'\nargFile = r\"G:\\My Drive\\Roriks Documents\\Projects\\CNC-Drawing\\Apollo\\apollo.gcode\"\n\n\n## show values ##\nprint (\"USB Port: %s\" % argPort )\nprint (\"Gcode file: %s\" % argFile )\n\n\ndef removeComment(string):\n\tif (string.find(';')==-1):\n\t\treturn string\n\telse:\n\t\treturn string[:string.index(';')]\n\n\ndef sendLine(s, line):\n msg = (line + '\\n').encode('utf-8')\n s.write(msg) # Send g-code block\n grbl_out = s.readline() # Wait for response with carriage return\n print('{} : {}'.format(msg, grbl_out.strip()))\n \ndef plunge(s):\n sendLine(s,'M5')\n\ndef retract(s):\n sendLine(s,'M3S90')\n time.sleep(0.3)\n\n# Open serial port\n#s = serial.Serial('/dev/ttyACM0',115200)\ns = serial.Serial(argPort, 115200)\nprint('Opening Serial Port')\n \n# Open g-code file\nf = open(argFile,'r')\nprint('Opening gcode file')\n\n# Wake up \nsendLine(s, '\\r\\n\\r\\n')\ntime.sleep(2) # Wait for Printrbot to initialize\ns.flushInput() # Flush startup text in serial input\n\n# Zero\nprint('Zeroing Home')\nsendLine(s, '$H')\nsendLine(s, 'G92 X0 Y0 Z0')\nretract(s)\n\nprint('Sending gcode')\n# Stream g-code\nskipNext = False\n\nfor line in f:\n if skipNext:\n skipNext = False\n continue \n\n if line.lower().strip() == '; plunge':\n plunge(s)\n skipNext = True\n continue\n\n if line.lower().strip() == '; retract':\n retract(s)\n skipNext = True\n continue\n\n l = removeComment(line)\n l = l.strip() # Strip all EOL characters for streaming\n if (l.isspace()==False and len(l)>0) :\n sendLine(s, l)\n\n# Close file and serial port\nf.close()\ns.close()","repo_name":"rorikmelberg/CNC","sub_path":"gcodesender.py","file_name":"gcodesender.py","file_ext":"py","file_size_in_byte":2026,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"69854656637","text":"import json\nimport math\nimport csv\nimport sys\nfrom copy import deepcopy\nimport data\nimport NUM\nimport sym\nimport row\nsys.path.append(\"./src\")\n\n\n# Numerics\n# Seed = 937162211\n\n\ndef rint(lo, hi):\n return math.floor(0.5 + rand(lo, hi))\n\nglobal Seed\n\n\ndef rand(lo, hi):\n Seed = 93716211\n lo = lo or 0\n hi = hi or 1\n Seed = (16807 * Seed) % 2147483647\n return lo + (hi - lo) * Seed / 2147483647\n\ndef rnd(n, places=3):\n mult = 10 ** places\n return math.floor(n * mult + 0.5) / mult\n\n\n# Lists\ndef map(t, fun):\n u = {}\n if(type(t) == dict):\n for k,v in t.items():\n v = fun(v)\n if k is None:\n k = 1+len(u)\n u[k] = v \n return u\n else:\n for k,v in enumerate(t):\n v = fun(v)\n if k is None:\n k = 1+len(u)\n u[k] = v \n return u\n\n\ndef many(t, n):\n u = []\n for i in range(n):\n u.append(any(t))\n\ndef kap(t, fun):\n u = {}\n for v in t:\n v, k = fun(t.index(v), v)\n if k in u.keys():\n u[k] = v\n else:\n u[len(u) + 1] = v\n return u\n\n\ndef sort(t):\n t = sorted(t)\n return t\n\n\ndef keys(t):\n return sorted(t.keys())\n\n\n# Strings\ndef oo(t):\n print(o(t))\n return True\n\n\ndef fmt(s, *a):\n return str.format(s, a)\n\n\ndef o(t):\n if(type(t) == NUM.NUM):\n return t.__dict__\n elif(type(t) == sym.sym):\n return t.__dict__\n elif(type(t) == row.Row):\n return t.__dict__\n \n if (not isinstance(t, dict)) and (not isinstance(t, list)):\n return str(t)\n\n\n def show(k, v):\n if str(k).find('_') != 0:\n v = o(v)\n return isinstance(t, dict) and (\":\" + str(k) + \" \" + str(v)) or str(v)\n\n u = []\n if isinstance(t, dict):\n for k, v in t.items():\n showop = show(k, v)\n if showop:\n u.append(showop)\n u.sort()\n\n elif isinstance(t, list):\n u = t\n return \"{\" + \" \".join(str(val) for val in u) + \"}\"\n\n\ndef the(t):\n oo(t)\n\n\ndef typecheck(x):\n try:\n a = float(x)\n b = int(a)\n except (TypeError, ValueError):\n pass\n else:\n return b\n try:\n a = float(x)\n except (TypeError, ValueError):\n return x\n else:\n return float(x)\n\n\ndef csv_read(filename):\n f = open(filename, 'r')\n # f = open(r'./etc/data/auto93.csv', 'r')\n\n reader = csv.reader(f)\n t = []\n for row in reader:\n t.append([typecheck(ele) for ele in row])\n # if fun is not None:\n # fun(t)\n return t\n\ndef cosine(a, b, c):\n den = 1 if c == 0 else 2 * c\n x1 = (a**2 + c**2 - b**2) / den\n x2 = max(0, min(1, x1))\n y = abs((a**2 - x2**2)) ** 0.5\n return x2, y\n\ndef last(t):\n if type(t) == dict:\n x = list(t.values())[-1]\n else:\n x = t[-1] \n return x\n\n\ndef show(node, what=None, cols=None, n_places=None, lvl=None):\n if node:\n lvl = lvl or 0\n print(\"|.. \" * lvl, end=\"\")\n if (\"left\" not in node):\n print(last(last(node[\"data\"].rows).cells))\n else:\n print(str(int(100 * node[\"C\"])))\n show(node.get(\"left\", None), what, cols, n_places, lvl+1)\n show(node.get(\"right\", None), what, cols, n_places, lvl+1)\n\ndef copy(t):\n if isinstance(t, list) or isinstance(t, dict):\n return deepcopy(t)\n\ndef repRows(t, rows):\n rows = copy(rows)\n for j,s in enumerate(rows[len(rows)-1]):\n rows[0][j] = str(rows[0][j]) + ' : ' + str(s)\n rows.pop()\n for n,row in enumerate(rows):\n if n==0:\n row.append('thingX')\n else:\n u = t.get('rows')[len(t.get('rows')) -n]\n row.append(u[len(u)-1])\n return data.DATA(rows)\n\ndef any(t):\n rintVal = rint1(None, len(t)-1)\n return t[rintVal]\n\ndef rint1(lo, hi):\n return math.floor(0.5 + rand(lo, hi))\n\ndef repPlace(data):\n n,g = 20,[]\n for i in range(n+1):\n g.append([])\n for j in range(n+1):\n g[i].append(\" \")\n maxy = 0\n print(\"\")\n for r, row in enumerate(data.rows):\n c = chr(r+65)\n print(c, last(row.cells))\n x, y = int(row.x*n), int(row.y*n)\n maxy = max(maxy, y)\n g[y][x] = c\n print(\"\")\n for y in range(maxy):\n print(\"{\" + \"\".join(g[y]) + \"}\")\n\ndef repgrid(sFile, t, rows, cols):\n t = exec(open(sFile).read())\n rows = repRows(t, transpose(t[\"cols\"]))\n cols = repCols(t[\"cols\"])\n show(rows.cluster())\n show(cols.cluster())\n repPlace(rows)\n\ndef transpose(t):\n u = []\n for i in range(len(t[0])):\n u.append([t[j][i] for j in range(len(t))])\n return u\n\ndef dofile(fileName):\n with open(fileName) as f:\n return json.load(f)\n\ndef repCols(cols):\n copycols = deepcopy(cols)\n for col in cols:\n col[-1] = str(col[0]) + \":\" + str(col[-1])\n for j in range(1, len(col)):\n col[j-1] = col[j]\n col.pop()\n cols.insert(0, ['Num' + str(k) for k in range(len(cols[0]))]) \n cols[0][-1] = \"thingX\"\n # cols = {k:{j:l for j,l in enumerate(v)} for k,v in enumerate(cols)}\n return data.DATA(cols)\n\ndef any(t):\n rintVal = rint(None, len(t)-1)\n return t[rintVal]\n\ndef rint(lo, hi):\n return math.floor(0.5 + rand(lo, hi))\n\n","repo_name":"aadiltajani/CSC591-HW","sub_path":"src/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":5217,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"14301382265","text":"# https://leetcode.com/problems/minimum-add-to-make-parentheses-valid/\n# 2021/10\n# 24 ms\n\nclass Solution:\n def minAddToMakeValid(self, s: str) -> int:\n unpaired, added = 0, 0\n for c in s:\n if c == '(':\n unpaired += 1\n else:\n if unpaired < 1:\n added += 1\n else:\n unpaired -= 1\n return unpaired + added\n","repo_name":"zvant/LeetCodeSolutions","sub_path":"0921.minimum_add_to_make_parentheses_valid_M.py","file_name":"0921.minimum_add_to_make_parentheses_valid_M.py","file_ext":"py","file_size_in_byte":430,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"73459417275","text":"import tkinter\nfrom tkinter import messagebox\n\nfrom tictactoe_gameengine import TictactoeGameEngine\n\n\nclass TictactoeGUI:\n def __init__(self):\n self.game_engine = TictactoeGameEngine()\n self.init_GUI()\n\n def init_GUI(self):\n self.CANVAS_SIZE = 300\n self.root=tkinter.Tk()\n self.root.title(\"틱택토\")\n self.root.geometry(f\"{self.CANVAS_SIZE}x{self.CANVAS_SIZE}\")\n self.root.resizable(width=False, height=False) #창 크기 변경 x\n\n self.canvas = tkinter.Canvas(self.root, bg=\"white\", width=self.CANVAS_SIZE, height=self.CANVAS_SIZE)\n self.canvas.pack()\n\n self.images = {} #{\"X\":PhotoImage객체, \"O\":PhotoImage객체}\n self.images[\"X\"] = tkinter.PhotoImage(file=\"X.gif\")\n self.images[\"O\"] = tkinter.PhotoImage(file=\"O.gif\")\n\n self.canvas.bind(\"\", self.click_handler) #self.click_handler() 를 쓰면 클릭 안되도 실행 -> 시험\n\n self.root.mainloop()\n\n def click_handler(self, event):\n row, col = self.coordinate_to_position(event.x, event.y)\n # set row, col\n self.game_engine.set(row,col)\n # show board\n self.game_engine.show_board()\n self.draw_board()\n # set winner\n winner = self.game_engine.set_winner()\n # 승자가 있거나 무승부면, 게임오버, 결과표시\n if winner == \"O\" or winner == \"X\":\n messagebox.showinfo(\"Game ovr\",f\"{winner} win\")\n self.root.quit()\n elif winner == \"d\":\n messagebox.showinfo(\"Game ovr\",f\"무승부\")\n self.root.quit()\n # chang_turn\n self.game_engine.change_turn()\n\n def draw_board(self):\n TILE_SIXE = self.CANVAS_SIZE // self.game_engine.SIZE #300//3 = 100\n x = 0\n y = 0\n for i,v in enumerate(self.game_engine.board):\n if v == \".\":\n pass\n else: #elif v == \"X\" or V == \"O\";\n self.canvas.create_image(x,y,anchor=\"nw\", image=self.images[v])\n x += TILE_SIXE\n if i % self.game_engine.SIZE == self.game_engine.SIZE - 1 :\n x = 0\n y += TILE_SIXE\n\n def coordinate_to_position(self, x, y):\n row = y // (self.CANVAS_SIZE//self.game_engine.SIZE) + 1\n col = x // (self.CANVAS_SIZE//self.game_engine.SIZE) + 1\n return row, col\n\n\nif __name__ == '__main__':\n ttt_GUI = TictactoeGUI()","repo_name":"jihami/Programming_PythonClass","sub_path":"tictactoe/tictactoe_GUI.py","file_name":"tictactoe_GUI.py","file_ext":"py","file_size_in_byte":2441,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"33184684571","text":"# https://leetcode.cn/contest/ubiquant2022/ranking/\n\nclass Solution:\n def numberOfPairs(self, nums: List[int]) -> int:\n cc = Counter()\n for x in nums:\n y = int(str(x)[::-1])\n cc[x-y] += 1\n ans = 0\n MOD = 10**9+7\n for k, v in cc.items():\n ans += v*(v-1)//2\n ans %= MOD\n return ans","repo_name":"zhenfelix/OnlineJudgeCodings","sub_path":"LeetCode/力扣杯/九坤投资专场-01.py","file_name":"九坤投资专场-01.py","file_ext":"py","file_size_in_byte":368,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"70905899515","text":"from __future__ import print_function\nfrom django.shortcuts import render_to_response, redirect\nfrom django.http import JsonResponse, HttpResponse\nfrom django.core.context_processors import csrf\nfrom django import forms\nfrom django.template import RequestContext\nfrom django.views.decorators.csrf import csrf_exempt\n\nfrom popsite.models import account\nfrom popsite.models import subkey\nfrom popsite.models import transaction\nfrom popsite.models import access\nfrom popsite.models import payment\n\nfrom django.core.cache import cache as rdb\nfrom django.core.validators import URLValidator\n\nfrom popsite.forms import CaptchaTestForm\nfrom popsite.forms import AccountsForm\nfrom popsite.forms import SearchForm, RegistrationForm\nfrom popsite.forms import VerificationForm\nfrom popsite.forms import RefundForm\n\nfrom tables import LoginTable, TransactionTable, SubkeyTable\nfrom popsite.tasks import create_keys\n\nfrom blockchain.exchangerates import get_ticker # remote\nfrom blockchain.pushtx import pushtx # remote\nfrom blockchain.blockexplorer import get_address # remote\n\nfrom pycoin.key.validate import is_public_bip32_valid, is_address_valid\nfrom pycoin.key.Key import Key\nfrom pycoin.tx.tx_utils import create_signed_tx\nfrom pycoin.tx import Spendable\nfrom pycoin.serialize import h2b\n\nfrom binascii import hexlify\nfrom datetime import datetime\nfrom decimal import Decimal\nfrom hashlib import sha256\nfrom M2Crypto import RSA, X509\nfrom random import choice\nfrom re import match\nfrom string import ascii_letters, digits, maketrans\nfrom ssl import PROTOCOL_TLSv1, PROTOCOL_SSLv23, PROTOCOL_SSLv3 #, PROTOCOL_SSLv2\nfrom ssl import get_server_certificate # remote\nfrom urllib2 import urlopen # remote\nimport json\nimport csv\n#from pycoin.tx.pay_to import ScriptMultisig, address_for_pay_to_script\n#from django.core.exceptions import ValidationError\n#from pycoin.services import spendables_for_address # remote\n\n#region Variables\ncryptoproof_deposit_key_str = 'secret_xpubkey_string'\ncryptoproof_deposit_key_obj = Key.from_text(cryptoproof_deposit_key_str)\ncryptoproof_payment_key_str = 'secret_xprikey_string'\ncryptoproof_payment_key_obj = Key.from_text(cryptoproof_payment_key_str)\ntransaction_code = 'secret_random_string'\n\napi_code = 'secret_api_string'\nhostname = 'blockchain.info'\n#cryptoproof_addr = '17NKQddUT443ZweQAtDvcQNC9oso2j9Zcv'\n\ninvalid_xpub = 'Invalid extended public key(s)'\ninvalid_addr = 'Invalid bitcoin address'\nerror_msg = 'Something bad happened, please try again later'\nsplit_string = 'xpub661MyMwAqRbc'\n\nnetwork_transaction_fee = Decimal('0.0001')\ncost_per_key = Decimal('0.01')\nfree_keys = 10\nhard_index_limit = 10000\nminimum_confirmations = 3\nsigs_limit = 15\nstring_length = 32\nsession_expiry_time = 600 # 10 minutes\n#endregion\n# Snippet from pycoin.services, received error when imported from library\ndef spendables_for_address(bitcoin_address):\n \"\"\"Snippet from pycoin.services, received error when imported from library\n Gets spendables objects for address\n\n Args:\n bitcoin_address (str, required): Mainnet Bitcoin address\n\n Returns:\n spendables (list): List of unspent inputs\n \"\"\"\n URL = \"http://blockchain.info/unspent?active=%s\" % bitcoin_address\n r = json.loads(urlopen(URL).read().decode(\"utf8\"))\n spendables = []\n for u in r[\"unspent_outputs\"]:\n coin_value = u[\"value\"]\n script = h2b(u[\"script\"])\n previous_hash = h2b(u[\"tx_hash\"])\n previous_index = u[\"tx_output_n\"]\n spendables.append(Spendable(coin_value, script, previous_hash, previous_index))\n return spendables\n\n# Test if https url and attempt certificate extraction #REMOTE\ndef get_cypher(submitted_url):\n \"\"\"REMOTE, Test if https url, attempt certificate extraction and encrypt random string\n\n Args:\n submitted_url (str, required): http(s) URL\n\n Returns:\n spendables (list): List of unspent inputs\n \"\"\"\n if match('^https', submitted_url):\n modified_url = submitted_url.replace('https://', '').strip('/')\n\n cert = False\n for i in [PROTOCOL_TLSv1, PROTOCOL_SSLv23, PROTOCOL_SSLv3]:\n try:\n cert = get_server_certificate((modified_url, 443), ssl_version=i)\n break\n except:\n pass\n\n if cert:\n vstring = ''.join([choice(ascii_letters + digits) for _ in xrange(32)])\n\n cert509 = X509.load_cert_string(cert)\n rsa = cert509.get_pubkey().get_rsa()\n return hexlify(rsa.public_encrypt(vstring, RSA.pkcs1_padding)), vstring\n\n return False, False\n\n# Set or unset session variables for certificate if available #REMOTE\ndef set_cypher_session(request, submitted_url):\n \"\"\"Save verification strings to session cache\n\n Args:\n request (request object, required): request object\n submitted_url (str, required): http(s) URL\n \"\"\"\n if not request.session.get('cert_check'):\n cypher = get_cypher(submitted_url)\n request.session['encrypted_str'] = cypher[0]\n request.session['decrypted_str'] = cypher[1]\n request.session['cert_check'] = True\n\n# Return decimal whole btc balance #REMOTE\ndef addr_bal(address):\n \"\"\"Retrieves lowest balance up to minimum confirmations\n\n Args:\n address (str, required): Mainnet Bitcoin address\n\n Returns:\n unconfirmed_balance (Decimal): Address balance at 0 confirmations\n confirmed_balance (Decimal): Address balance at minimum confirmations\n \"\"\"\n # Blockchain.info returns positive confirmed balance after withdrawl so unconfirmed balance is needed\n confirmed_balance = Decimal('{0:.8}'.format(urlopen('https://%s/q/addressbalance/%s?confirmations=%s&api_code=%s' % (hostname, address, minimum_confirmations, api_code)).read()))\n unconfirmed_balance = Decimal('{0:.8}'.format(urlopen('https://%s/q/addressbalance/%s?confirmations=%s&api_code=%s' % (hostname, address, 0, api_code)).read()))\n\n if unconfirmed_balance < confirmed_balance:\n return unconfirmed_balance / 100000000\n\n return confirmed_balance / 100000000\n\ndef get_bitcoin_price(request):\n \"\"\"Check for updated Bitcoin price\n\n Args:\n request (request object, required): request object\n\n Returns:\n price (Decimal): Bitcoin price\n False (Bool): False\n \"\"\"\n price = rdb.get('price')\n if not price:\n try:\n price = Decimal('{0:.2f}'.format(get_ticker()['USD'].p15min))\n rdb.set('price', price, timeout=600)\n return price\n except:\n request.session['error_msg'] = error_msg\n return False\n return price\n\n# Retrieve IP address of client\ndef get_client_ip(request):\n \"\"\"Get client ip address\n\n Args:\n request (request object, required): request object\n\n Returns:\n request.META.get('REMOTE_ADDR') (str): Client ip address\n x_forwarded_for.split(',')[0] (str): Client ip address\n \"\"\"\n x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR')\n if x_forwarded_for:\n return x_forwarded_for.split(',')[0]\n\n return request.META.get('REMOTE_ADDR')\n\ndef set_pass_token(request, destination):\n \"\"\"Set pass token in request cache for adjacent views access\n\n Args:\n request (request object, required): request object\n destination (str, required): varification token for webpage access\n \"\"\"\n request.session[u'pass_token'] = destination\n request.session.set_expiry(session_expiry_time)\n\ndef verify_pass_token(request, source):\n \"\"\"Check if verify cache token matches client string\n\n Args:\n request (request object, required): request object\n source (str, required): submitted verification string\n\n Returns:\n True (bool): True\n \"\"\"\n if not request.session.get('pass_token') == source:\n return False\n del request.session['pass_token']\n return True\n\ndef details(request): #, pubkey_json, sigs):\n \"\"\"Perform checks on public key(s) and return dictionary of account values\n\n Args:\n request (request object, required): request object\n pubkey_json (str, required): submitted verification string\n sigs (str, required): number of signatories\n\n Returns:\n locals (dict): Dictionary of account valuew\n \"\"\"\n # Check api request limit\n if rdb.get(get_client_ip(request)):\n return dict(status=False, message='request limit reached')\n rdb.set(get_client_ip(request), True, timeout=1)\n\n # Check for valid sig number\n try:\n sigs = int(request.session['sigs']) if request.session.get('sigs') else int(request.POST['sigs'])\n except:\n return dict(status=False, message='invalid sigs field')\n\n # Check for valid sig limit\n if sigs > sigs_limit:\n return dict(status=False, message='invalid sigs limit')\n\n # Create xpubkey string list\n xpubkey_str_lst = []\n for i in request.session['xpubkeys'].split(split_string) if request.session.get('xpubkeys') else request.POST['xpubkeys'].split(split_string):\n xpubkey_str_lst.append(split_string + i)\n del xpubkey_str_lst[0]\n\n # Adjust sigs number\n if len(xpubkey_str_lst) < 2:\n sigs = 0\n\n # Check for conflicting sigs number\n if sigs > len(xpubkey_str_lst):\n return dict(status=False, message='number of sigs cannot be greater than number of submitted xpubkeys')\n\n # Check valid pubkeys\n for i in xpubkey_str_lst:\n if not is_public_bip32_valid(i):\n return dict(status=False, message=invalid_xpub)\n\n # Check unique pubkeys\n if not len(xpubkey_str_lst) == len(set(xpubkey_str_lst)):\n return dict(status=False, message='Pubkeys must be unique')\n\n # Create xpubkey object list\n xpubkey_obj_lst = []\n for i in xpubkey_str_lst:\n xpubkey_obj_lst.append(Key.from_text(i))\n\n # Make sure list is not empty\n if len(xpubkey_obj_lst) < 1:\n return dict(status=False, message='xpubkey object error')\n\n for i in xpubkey_obj_lst:\n # Test if depth and child index are beyond maximum\n if i.child_index() > 2146483647:\n return dict(status=False, message='xpubkey child index cannot be more than 2146483647')\n\n if i.tree_depth() > 255:\n return dict(status=False, message='xpubkey tree depth cannot be more than 255')\n\n # Test if key balance is zero (optional restriction)\n try:\n addr_obj = get_address(i.address())\n except:\n return dict(status=False, message=error_msg)\n\n if not addr_obj.total_received == 0:\n return dict(status=False, message='public key received balance cannot be non-zero')\n\n if len(xpubkey_obj_lst) > 1 and sigs < 1:\n return dict(status=False, message='sigs cannot be zero for multiple keys')\n\n hashone = sha256(''.join(xpubkey_str_lst) + str(sigs)).hexdigest()\n hashtwo = sha256(hashone).hexdigest()\n\n # Get client subkey object\n string_obj = maketrans('', '')\n translation = string_obj.translate(string_obj, digits)\n translated = str(hashone).translate(string_obj, translation)\n\n # Check that subkey path is at least 3 levels deep\n if not len(translated) > 18:\n return dict(status=False, message='hashed address of the pubkey(s) contains less than 9 digits and cannot be used for security reasons,\\nplease use a different pubkey')\n\n # Get path to private key\n key_path = ''\n limit = len(translated) / 9\n c = 0\n while c <= limit:\n key_path += '%s/' % translated[c*9:(c+1)*9]\n c += 1\n\n request.session['path'] = key_path.strip('/')\n subkey_obj = cryptoproof_deposit_key_obj.subkey_for_path(key_path.strip('/'))\n account_record_qset = account.objects.filter(hashtwo=hashtwo)\n\n status = True\n return locals()\n\ndef get_signed_tx(request, cryptoproof_payment_key_obj, account_record):\n \"\"\"Verify signed transaction\n\n Args:\n request (request object, required): request object\n cryptoproof_payment_key_obj (xpubkey_obj, required): payment address key\n account_record (access object, required): account record\n\n Returns:\n locals (dict): Dictionary of account valuew\n \"\"\"\n def create_tx(l):\n \"\"\"Create signed transaction\n\n Args:\n l (list, required): Contains details to build signed object\n\n Returns:\n create_signed_tx (signed transaction object): signed transaction\n \"\"\"\n return create_signed_tx(spendables,\n l,\n wifs=[Key.from_text(request.session.get('cryptoproof_subkey_str')).wif()],\n fee='standard')\n\n # Get spendables\n try:\n spendables = spendables_for_address('%s&api_code=%s&confirmations=%s' % (request.session.get('deposit_address'), api_code, minimum_confirmations))\n except:\n return False\n\n if not request.POST.get('refund_address'):\n # Reserve transaction\n try:\n return create_tx([(cryptoproof_payment_key_obj.subkey(account_record.id).address(), int(account_record.spent * 100000000)),\n Key.from_text(request.session.get('cryptoproof_subkey_str')).address()])\n except:\n return create_tx([cryptoproof_payment_key_obj.subkey(account_record.id).address()])\n\n # Withdrawl request\n if account_record.spent > 0:\n try:\n return create_tx([(cryptoproof_payment_key_obj.subkey(account_record.id).address(), int(account_record.spent * 100000000)),\n request.POST.get('refund_address')])\n except:\n return create_tx([cryptoproof_payment_key_obj.subkey(account_record.id).address()]) # is this necessary?\n else:\n return create_tx([request.POST.get('refund_address')])\n\n\n#######\n# API #\n#######\n## DICTIONARIE ##\n# 1 args\n@csrf_exempt\ndef api_search(request): #, address_json):\n if request.POST:\n # Check api request limit\n if rdb.get(get_client_ip(request)):\n response_json = dict(status=False, message='request limit reached')\n return JsonResponse(response_json)\n rdb.set(get_client_ip(request), True, timeout=1)\n\n # Test if string is valid BTC address and search database for match if so\n if not is_address_valid(request.POST['info']):\n response_json = dict(status=False, message='invalid bitcoin address')\n return JsonResponse(response_json)\n\n # If key_result exists, increment result number, get account record\n key_result = subkey.objects.filter(hash=sha256(request.POST['info']).hexdigest())\n if not key_result.exists():\n response_json = dict(status=False, message='address not found')\n return JsonResponse(response_json)\n\n key_record = key_result[0]\n key_record.hits += 1\n key_record.last_accessed = datetime.now()\n key_record.save()\n\n ac_obj = account.objects.filter(id=key_record.account_id)[0]\n message = dict(compressed_address=request.POST['info'],\n associated_url=ac_obj.url,\n verified_status=ac_obj.verified_status,\n hits=key_record.hits)\n\n response_json = dict(status=True, message=message)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_process_transactions(request):\n if request.POST:\n if request.POST.get('transaction_code') == transaction_code:\n\n payment_obj_qset = payment.objects.all()\n if payment_obj_qset.exists():\n\n for i in payment_obj_qset:\n pushtx(i.signed_tx)\n\n account_record = account.objects.filter(id=i.account_id)[0]\n account_record.spent = 0\n account_record.save()\n i.delete()\n\n response_json = dict(status=True)\n return JsonResponse(response_json)\n\n# 2 args\n@csrf_exempt\ndef api_url(request): #, pubkey_json, sigs):\n if request.POST:\n # Validate URL\n url = request.POST['info'].replace('www.', '')\n validator = URLValidator(schemes=['http', 'https'])\n try:\n validator(url)\n except:\n response_json = dict(status=False, message='url string error', url=url)\n return JsonResponse(response_json)\n\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # exit if url is unchanged\n if account_record.url == url:\n response_json = dict(status=False, message='url string unchanged')\n return JsonResponse(response_json)\n\n account_record.verified_status = False\n account_record.url = url\n account_record.save()\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_url').save()\n response_json = dict(status=True)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_verify(request): #, pubkey_json, sigs):\n if request.POST:\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n decrypted_str = request.POST.get('info')\n if not decrypted_str:\n # Check not already verified\n if account_record.verified_status:\n response_json = dict(status=False, message='account already verified')\n return JsonResponse(response_json)\n\n # Validate URL\n validator = URLValidator(schemes=['https'])\n try:\n validator(account_record.url)\n except:\n response_json = dict(status=False, message='url string error')\n return JsonResponse(response_json)\n\n # Get cypher for url\n cypher = get_cypher(account_record.url)\n encrypted_str = cypher[0]\n decrypted_str = cypher[1]\n\n # Save decrypted string as key for 30 seconds to redis cache, value as True\n rdb.set(''.join(details_dict['xpubkey_str_lst']) + decrypted_str, True, timeout=30)\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_verify_get').save()\n response_json = dict(status=True, message=encrypted_str)\n return JsonResponse(response_json)\n\n # Check valid string\n if not rdb.get(''.join(details_dict['xpubkey_str_lst']) + decrypted_str):\n response_json = dict(status=False, message='invalid string')\n return JsonResponse(response_json)\n\n # Get record and set verified status\n account_record.verified_status = True\n account_record.save()\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_verify_post').save()\n\n response_json = dict(status=True)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_account(request): #, pubkey_json, sigs, web_ui=False):\n if request.POST:\n # Get account details dictionary\n\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # Get current bitcoin price\n price = get_bitcoin_price(request)\n if not price:\n return dict(status=False, message='blockchain price api failed')\n\n # Get balance\n try:\n balance = addr_bal(details_dict['subkey_obj'].address()) - account_record.spent\n except:\n return dict(status=False, message='blockchain balance api failed')\n\n # Calculate variables and initialise dictionary\n credit = Decimal('{0:.2f}'.format(price * balance)) - Decimal('0.01')\n if credit < 0:\n credit = Decimal('{0:.2f}'.format(0))\n\n affordable_keys = int(credit / cost_per_key)\n upper_index = affordable_keys + account_record.index_limit\n\n remaining_key_allowance = free_keys - account_record.index_limit\n if not remaining_key_allowance <= 0:\n upper_index += remaining_key_allowance\n else:\n remaining_key_allowance = 0\n\n account_record_dict = details_dict['account_record_qset'].values()[0]\n # if not web_ui:\n del account_record_dict['hashtwo']\n del account_record_dict['spent']\n del account_record_dict['id']\n\n account_record_dict.update(dict(deposit_address=details_dict['subkey_obj'].address(),\n credit=credit,\n available_keys=affordable_keys,\n upper_index=upper_index,\n free_keys=remaining_key_allowance,\n balance=balance,\n price=price))\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_account').save()\n return JsonResponse(dict(status=True, message=account_record_dict))\n\n## LIST OF DICTIONARIES ##\n@csrf_exempt\ndef api_login_history(request): #, pubkey_json, sigs):\n if request.POST:\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # Create and return json response\n l = []\n for i in access.objects.filter(account_id=account_record.id).order_by('-id'):\n l.append(dict(date=str(i.id), ip_address=i.ipaddress, interface=i.interface.encode('utf-8')))\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_login_history').save()\n response_json = dict(status=True, message=l)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_subkey_activity(request): #, pubkey_json, sigs):\n if request.POST:\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # Create and return json response\n l = []\n for i in subkey.objects.filter(account_id=account_record.id).filter(hits__gt=0).order_by('-last_accessed'):\n l.append(dict(date=str(i.last_accessed), subkey_index=i.subkey_index, hits=i.hits))\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_subkey_activity').save()\n response_json = dict(status=True, message=l)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_transaction_history(request): #, pubkey_json, sigs):\n if request.POST:\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # Create and return json response\n l = []\n for i in transaction.objects.filter(account_id=account_record.id).order_by('-id'):\n l.append(dict(date=str(i.id), btc_price=i.btc_price, btc_amount=i.btc_amount))\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_transaction_history').save()\n response_json = dict(status=True, message=l)\n return JsonResponse(response_json)\n\n# 3 args\n@csrf_exempt\ndef api_index(request): #, pubkey_json, sigs, submitted_index_limit):\n if request.POST:\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n account_record = details_dict['account_record_qset'][0]\n\n # Get current bitcoin price\n price = get_bitcoin_price(request)\n if not price:\n return dict(status=False, message='blockchain price api failed')\n\n # Get balance\n try:\n balance = addr_bal(details_dict['subkey_obj'].address()) - account_record.spent\n except:\n return dict(status=False, message='blockchain balance api failed')\n\n # Calculate variables and initialise dictionary\n credit = Decimal('{0:.2f}'.format(price * balance)) - Decimal('0.01')\n if credit < 0:\n credit = Decimal('{0:.2f}'.format(0))\n\n affordable_keys = int(credit / cost_per_key)\n upper_index = affordable_keys + account_record.index_limit\n submitted_index_limit = int(request.POST['info'])\n extra_keys = submitted_index_limit - account_record.index_limit\n\n remaining_key_allowance = free_keys - account_record.index_limit\n if remaining_key_allowance < 1:\n remaining_key_allowance = 0\n\n # upper_index = details_dict['affordable_keys'] + details_dict['remaining_key_allowance']\n if upper_index < 1:\n response_json = dict(status=False, message='insufficient credit')\n return JsonResponse(response_json)\n\n # Check submitted limit is not greater than allowed limit or less than 1\n if submitted_index_limit > upper_index or submitted_index_limit < 1:\n response_json = dict(status=False, message='index limit invalid')\n return JsonResponse(response_json)\n\n # Check for extra keys\n if extra_keys < 1:\n response_json = dict(status=False, message='index limit unchanged')\n return JsonResponse(response_json)\n\n chargable_keys = extra_keys - remaining_key_allowance\n if not chargable_keys < 1: # Charge\n btc_spent = Decimal(\"{0:.8f}\".format((chargable_keys * cost_per_key) / price))\n\n account_record.spent += btc_spent\n transaction.objects.create(account_id=account_record.id,\n btc_price=price,\n btc_amount=btc_spent).save()\n\n # Create child objects\n create_keys.delay(''.join(details_dict['xpubkey_str_lst']),\n submitted_index_limit,\n account_record.id,\n account_record.index_limit,\n request.POST['sigs'])\n\n account_record.index_limit = submitted_index_limit\n account_record.save()\n\n # Create and save transaction for spent funds\n if account_record.spent > network_transaction_fee:\n try:\n spendables = spendables_for_address('%s&api_code=%s&confirmations=%s' % (details_dict['subkey_obj'].address(), api_code, minimum_confirmations))\n except:\n response_json = dict(status=False, message='pycoin spendables api failed')\n return JsonResponse(response_json)\n\n try:\n signed_tx = create_signed_tx(spendables,\n [(cryptoproof_payment_key_obj.subkey(account_record.id).address(), int(account_record.spent * 100000000)), details_dict['subkey_obj'].address()],\n wifs=[details_dict['subkey_obj'].wif()],\n fee='standard')\n except:\n signed_tx = create_signed_tx(spendables,\n [cryptoproof_payment_key_obj.subkey(account_record.id).address()], #int(account_record.spent * 100000000)), details_dict['subkey_obj'].address()],\n wifs=[details_dict['subkey_obj'].wif()],\n fee='standard')\n\n #signed_tx = get_signed_tx(request, cryptoproof_payment_key_obj, account_record)\n #pushtx(signed_tx.as_hex())\n\n payment_obj_lst = payment.objects.filter(account_id=account_record.id)\n if payment_obj_lst.exists():\n payment_obj_lst.delete()\n payment.objects.create(account_id=account_record.id, signed_tx=signed_tx.as_hex()).save()\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_index').save()\n response_json = dict(status=True)\n return JsonResponse(response_json)\n\n@csrf_exempt\ndef api_withdraw(request): #, pubkey_json, sigs, address_json):\n if request.POST:\n # Check valid bitcoin address\n if not is_address_valid(request.POST['info']):\n response_json = dict(status=False, message='invalid bitcoin address')\n return JsonResponse(response_json)\n\n # Get account details dictionary\n details_dict = details(request) #, pubkey_json, sigs)\n if not details_dict['status']:\n response_json = dict(status=False, message=details_dict['message'])\n return JsonResponse(response_json)\n\n # Check account exists\n if not details_dict['account_record_qset'].exists():\n response_json = dict(status=False, message='account record does not exist')\n return JsonResponse(response_json)\n\n account_record = details_dict['account_record_qset'][0]\n # Get balance\n try:\n balance = addr_bal(details_dict['subkey_obj'].address()) - account_record.spent\n except:\n response_json = dict(status=False, message='blockchain balance api failed')\n return JsonResponse(response_json)\n\n #balance = Decimal(request.session.get('balance'))\n if not balance > network_transaction_fee:\n response_json = dict(status=False, message='balance must be greater than %s' % network_transaction_fee)\n return JsonResponse(response_json)\n\n try:\n spendables = spendables_for_address('%s&api_code=%s&confirmations=%s' % (details_dict['subkey_obj'].address(), api_code, minimum_confirmations))\n except:\n response_json = dict(status=False, message='pycoin spendables api failed')\n return JsonResponse(response_json)\n\n # Create and push refund transaction\n if account_record.spent > 0:\n try:\n signed_tx = create_signed_tx(spendables,\n [(cryptoproof_payment_key_obj.subkey(account_record.id).address(), int(account_record.spent * 100000000)), request.POST['info']],\n wifs=[details_dict['subkey_obj'].wif()],\n fee='standard')\n except:\n signed_tx = create_signed_tx(spendables,\n [cryptoproof_payment_key_obj.subkey(account_record.id).address()],\n wifs=[details_dict['subkey_obj'].wif()],\n fee='standard')\n else:\n signed_tx = create_signed_tx(spendables, [request.POST['info']],\n wifs=[details_dict['subkey_obj'].wif()],\n fee='standard')\n\n try:\n pushtx(signed_tx.as_hex())\n except:\n response_json = dict(status=False, message='blockchain pushtx api failed')\n return JsonResponse(response_json)\n\n payment_obj_lst = payment.objects.filter(account_id=account_record.id)\n if payment_obj_lst.exists():\n payment_obj_lst.delete()\n\n transaction.objects.create(account_id=account_record.id,\n btc_price=000,\n btc_amount=balance - Decimal(str(signed_tx.fee() / float(100000000)))).save()\n\n account_record.spent = 0\n account_record.save()\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='api_withdraw').save()\n response_json = dict(status=True)\n return JsonResponse(response_json)\n\n\n############\n# DOWNLOAD #\n############\n\ndef audit(request):\n # Create the HttpResponse object with the appropriate CSV header.\n timestamp = datetime.now().date().strftime('%d%m%Y') + datetime.now().time().strftime('%H%M%S')\n\n response = HttpResponse(content_type='text/csv')\n response['Content-Disposition'] = 'attachment; filename=\"cryptoproof_db_{}.csv\"'.format(timestamp)\n\n writer = csv.writer(response)\n subkey_lst = subkey.objects.filter(hits__gt=0).order_by('?')\n\n if subkey_lst.exists():\n writer.writerow(['hashes', 'submissions'])\n\n for i in subkey_lst:\n writer.writerow([i.hash, i.hits])\n else:\n writer.writerow(['None'])\n\n return response\n\ndef login_history(request):\n try:\n # Create the HttpResponse object with the appropriate CSV header.\n timestamp = datetime.now().date().strftime('%d%m%Y') + datetime.now().time().strftime('%H%M%S')\n\n response = HttpResponse(content_type='text/csv')\n response['Content-Disposition'] = 'attachment; filename=\"cryptoproof_login_history_{}.csv\"'.format(timestamp)\n\n writer = csv.writer(response)\n login_history = access.objects.filter(account_id=request.session['account_id']).order_by('-id')\n\n if login_history.exists():\n writer.writerow(['date', 'ip_address', 'interface'])\n\n for i in login_history:\n writer.writerow([i.id, i.ipaddress, i.interface])\n else:\n writer.writerow(['None'])\n\n return response\n\n except:\n request.session['error_msg'] = 'Session timeout'\n return redirect('login')\n\ndef subkey_activity(request):\n try:\n # Create the HttpResponse object with the appropriate CSV header.\n timestamp = datetime.now().date().strftime('%d%m%Y') + datetime.now().time().strftime('%H%M%S')\n\n response = HttpResponse(content_type='text/csv')\n response['Content-Disposition'] = 'attachment; filename=\"cryptoproof_subkey_activity_{}.csv\"'.format(timestamp)\n\n writer = csv.writer(response)\n subkey_activity = subkey.objects.filter(account_id=request.session['account_id']).filter(hits__gt=0).order_by('-subkey_index')\n\n if subkey_activity.exists():\n writer.writerow(['last_accessed', 'subkey_index', 'hits'])\n\n for i in subkey_activity:\n writer.writerow([i.last_accessed, i.subkey_index, i.hits])\n else:\n writer.writerow(['None'])\n\n return response\n\n except:\n request.session['error_msg'] = 'Session timeout'\n return redirect('login')\n\ndef transaction_history(request):\n try:\n # Create the HttpResponse object with the appropriate CSV header.\n timestamp = datetime.now().date().strftime('%d%m%Y') + datetime.now().time().strftime('%H%M%S')\n\n response = HttpResponse(content_type='text/csv')\n response['Content-Disposition'] = 'attachment; filename=\"cryptoproof_transaction_history_{}.csv\"'.format(timestamp)\n\n writer = csv.writer(response)\n transaction_history = transaction.objects.filter(account_id=request.session['account_id']).order_by('-id')\n\n if transaction_history.exists():\n writer.writerow(['date', 'btc_price', 'btc_amount'])\n\n for i in transaction_history:\n writer.writerow([i.id, i.btc_price, i.btc_amount])\n else:\n writer.writerow(['None'])\n\n return response\n\n except:\n request.session['error_msg'] = 'Session timeout'\n return redirect('login')\n\n\n#########\n# VIEWS #\n#########\n\ndef search(request):\n #region POST\n if request.POST:\n if not verify_pass_token(request, 'search'):\n return redirect('search')\n\n if request.POST.get('camera'):\n request.session.update(camera=True)\n return redirect('search')\n\n if request.POST.get('to_login'):\n return redirect('login')\n #\n # if not is_address_valid(request.POST.get('address')):\n # request.session['error_msg'] = invalid_addr\n # return redirect('search')\n\n # results_dict = json.loads(api_search(request, request.POST.get('address')).getvalue())\n results_dict = json.loads(api_search(request).getvalue())\n if not results_dict['status']:\n request.session['error_msg'] = results_dict['message']\n return redirect('search')\n\n results_dict.update(csrf(request))\n return render_to_response('popsite/addinfo.html', results_dict['message'])\n #endregion\n #region GET\n # Clear and set initial session data\n html_dtc = dict()\n html_dtc.update(dict(camera=request.session.get('camera')))\n html_dtc.update(dict(error_msg=request.session.get('error_msg')))\n html_dtc.update(dict(form=SearchForm()))\n html_dtc.update(csrf(request))\n\n request.session.flush()\n set_pass_token(request, u'search')\n return render_to_response('popsite/search.html', html_dtc)\n #endregion\n\ndef login(request):\n #region POST\n if request.POST:\n # Check valid token\n if not verify_pass_token(request, 'login'):\n return redirect('login')\n\n # Check for button click\n if request.POST.get('to_search'):\n return redirect('search')\n\n # Check for empty strings\n if request.POST.get('xpubkeys') == '':\n request.session['error_msg'] = invalid_xpub\n return redirect('login')\n\n # Get account details dictionary\n details_dict = details(request) #, request.POST['xpubkeys'], request.POST['sigs'])\n if not details_dict['status']:\n request.session['error_msg'] = details_dict['message']\n return redirect('login')\n\n # Populate session cache\n request.session['xpubkeys'] = ''.join(details_dict['xpubkey_str_lst'])\n request.session['sigs'] = details_dict['sigs']\n request.session['hahsone'] = details_dict['hashone']\n request.session['hashtwo'] = details_dict['hashtwo']\n request.session['deposit_address'] = details_dict['subkey_obj'].address()\n request.session['cryptoproof_subkey_str'] = details_dict['subkey_obj'].as_text(as_private=True)\n\n # Distinguish between login or register\n if not details_dict['account_record_qset'].exists():\n set_pass_token(request, u'register')\n return redirect('register')\n\n # Get account record and balance\n account_record = details_dict['account_record_qset'][0]\n try:\n balance = addr_bal(request.session.get('deposit_address')) - account_record.spent\n request.session['balance'] = str(balance)\n except:\n request.session['error_msg'] = error_msg + ' 1'\n return redirect('login')\n\n request.session['account_id'] = account_record.id\n request.session['account_url'] = account_record.url\n\n # Set cypher session\n if not account_record.verified_status:\n try:\n set_cypher_session(request, account_record.url)\n except:\n request.session['error_msg'] = error_msg + ' 2'\n return redirect('login')\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='web_ui').save()\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n #endregion\n #region GET\n # Clear and set initial session data\n html_dtc = dict()\n html_dtc.update(dict(error_msg=request.session.get('error_msg')))\n html_dtc.update(dict(form=RegistrationForm()))\n html_dtc.update(csrf(request))\n\n request.session.flush()\n set_pass_token(request, u'login')\n return render_to_response('popsite/login.html', html_dtc)\n #endregion\n\ndef register(request):\n #region POST\n if request.POST:\n\n if not verify_pass_token(request, 'register'):\n return redirect('login')\n\n if not CaptchaTestForm(request.POST).is_valid():\n request.session['error_msg'] = 'You got the capture wrong'\n return redirect('login')\n\n account_record = account.objects.create()\n account_record.hashtwo = request.session.get('hashtwo')\n account_record.deposit_address = sha256(request.session.get('deposit_address')).hexdigest()\n account_record.save()\n request.session['account_id'] = account_record.id\n request.session['account_url'] = account_record.url\n\n try:\n balance = addr_bal(request.session.get('deposit_address'))\n request.session['balance'] = str(balance)\n except:\n request.session['error_msg'] = error_msg\n return redirect('login')\n\n access.objects.create(account_id=account_record.id, ipaddress=get_client_ip(request), interface='web_ui').save()\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n #endregion\n #region GET\n if not verify_pass_token(request, 'register'):\n return redirect('login')\n\n price = get_bitcoin_price(request)\n if not price:\n return redirect('login')\n\n form = CaptchaTestForm()\n credit = Decimal('{0:.2f}'.format(0)) - Decimal('0.01')\n if credit < 0:\n credit = Decimal('{0:.2f}'.format(0))\n\n balance = 0\n html_dtc = dict(form=form,\n deposit_address=request.session.get('deposit_address'),\n price=price,\n balance=Decimal(\"{0:.8f}\".format(balance)),\n credit=credit,\n affordable_keys=free_keys)\n\n html_dtc.update(csrf(request))\n set_pass_token(request, u'register')\n return render_to_response('popsite/register.html', html_dtc)\n #endregion\n\ndef dashboard(request):\n #region POST\n if request.POST:\n if not verify_pass_token(request, 'dashboard'):\n return redirect('login')\n\n # Check if verify button clicked and redirect to verification page if so\n if request.POST.get('verify'):\n set_pass_token(request, u'verify')\n return redirect('verify')\n\n account_record = account.objects.filter(id=request.session.get('account_id'))[0]\n\n try:\n abs_balance = addr_bal(request.session.get('deposit_address'))\n balance = abs_balance - account_record.spent\n request.session['balance'] = str(balance)\n except:\n request.session['error_msg'] = error_msg\n return redirect('login')\n\n # Check for refund request\n if request.POST.get('refund_address'):\n\n # Check valid bitcoin address\n if not is_address_valid(request.POST['refund_address']):\n request.session['error_msg'] = 'Invalid bitcoin address'\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n\n signed_tx = get_signed_tx(request, cryptoproof_payment_key_obj, account_record)\n if not signed_tx:\n request.session['error_msg'] = 'Invalid transaction'\n return redirect('login')\n\n try:\n pushtx(signed_tx.as_hex())\n except:\n request.session['error_msg'] = error_msg\n return redirect('login')\n\n payment_obj_lst = payment.objects.filter(account_id=account_record.id)\n if payment_obj_lst.exists():\n payment_obj_lst.delete()\n\n transaction.objects.create(account_id=account_record.id,\n btc_price=000,\n btc_amount=balance - Decimal(str(signed_tx.fee() / float(100000000)))).save()\n #btc_amount=balance - signed_tx.fee()).save()\n\n account_record.spent = 0\n request.session['balance'] = str(Decimal('{0:.8f}'.format(0)))\n account_record.save()\n\n request.session['error_msg'] = 'Withdrawal submitted successfully.'\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n\n # Set variables\n submitted_url = request.POST.get('url').replace('www.', '')\n if not submitted_url:\n submitted_url = 'Anonymous'\n\n if request.POST.get('index_limit'):\n submitted_index_limit = int(request.POST.get('index_limit'))\n else:\n submitted_index_limit = account_record.index_limit\n\n price = get_bitcoin_price(request)\n if not price:\n return redirect('login')\n\n # Set maximum key index\n credit = Decimal('{0:.2f}'.format(price * balance)) - Decimal('0.01')\n if credit < 0:\n credit = Decimal('{0:.2f}'.format(0))\n\n affordable_keys = int(credit / cost_per_key)\n upper_index = affordable_keys + account_record.index_limit\n remaining_key_allowance = free_keys - account_record.index_limit\n if not remaining_key_allowance <= 0:\n upper_index += remaining_key_allowance\n\n if submitted_index_limit > upper_index:\n request.session['error_msg'] = 'Database conflict error'\n return redirect('login')\n\n # Reset verification status and cert_check to false if url has changed\n if not account_record.url == submitted_url:\n account_record.verified_status = False\n request.session['cert_check'] = False\n account_record.url = submitted_url\n request.session['account_url'] = submitted_url\n\n # Get cypher if verified status is false\n if not account_record.verified_status:\n try:\n set_cypher_session(request, submitted_url)\n except:\n request.session['error_msg'] = error_msg\n return redirect('login')\n\n # Check if account credited and update spend field if needed\n extra_keys = submitted_index_limit - account_record.index_limit\n if extra_keys < 1:\n account_record.save()\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n\n # Does the person have any free keys left? How many?\n remaining_free_keys = 0\n if account_record.index_limit < free_keys:\n remaining_free_keys = free_keys - account_record.index_limit\n\n # Discount free keys\n chargable_keys = extra_keys - remaining_free_keys\n\n if not chargable_keys < 1: # Charge\n btc_spent = Decimal(\"{0:.8f}\".format((chargable_keys * cost_per_key) / price))\n account_record.spent += btc_spent\n\n transaction.objects.create(account_id=account_record.id,\n btc_price=price,\n btc_amount=btc_spent).save()\n\n # Create child objects\n create_keys.delay(request.session['xpubkeys'],\n submitted_index_limit,\n account_record.id,\n account_record.index_limit,\n request.session.get('sigs', '0'))\n\n account_record.index_limit = submitted_index_limit\n account_record.save()\n\n request.session['balance'] = str(abs_balance - account_record.spent)\n # Create and save transaction for spent funds\n if account_record.spent > network_transaction_fee:\n signed_tx = get_signed_tx(request, cryptoproof_payment_key_obj, account_record)\n if not signed_tx:\n request.session['error_msg'] = 'Invalid transaction'\n return redirect('login')\n\n payment_obj_lst = payment.objects.filter(account_id=account_record.id)\n if payment_obj_lst.exists():\n payment_obj_lst.delete()\n payment.objects.create(account_id=account_record.id, signed_tx=signed_tx.as_hex()).save()\n\n set_pass_token(request, u'dashboard')\n return redirect('dashboard')\n #endregion\n #region GET\n if not verify_pass_token(request, 'dashboard'):\n return redirect('login')\n\n # Get current bitcoin price\n price = get_bitcoin_price(request)\n if not price:\n return redirect('login')\n\n # Calculate max available keys based on excess balance\n account_record = account.objects.filter(id=request.session['account_id'])[0]\n balance = Decimal(request.session.get('balance'))\n\n # Blank url fiels if 'Anonymous'\n if account_record.url == 'Anonymous':\n account_record.url = ''\n\n # Set maximum key index\n credit = Decimal('{0:.2f}'.format(price * balance)) - Decimal('0.01')\n if credit < 0:\n credit = Decimal('{0:.2f}'.format(0))\n\n affordable_keys = int(credit / cost_per_key)\n upper_index = affordable_keys + account_record.index_limit\n remaining_key_allowance = free_keys - account_record.index_limit\n if not remaining_key_allowance <= 0:\n upper_index += remaining_key_allowance\n else:\n remaining_key_allowance = 0\n\n account_form = AccountsForm(instance=account_record)\n account_form.fields['index_limit'] = forms.IntegerField(label='Index Limit',\n min_value=account_record.index_limit,\n max_value=upper_index,\n initial=account_record.index_limit)\n\n subkey_obj_qset = subkey.objects.filter(account_id=account_record.id, hits__gt=0)\n subkey_table = SubkeyTable(subkey_obj_qset, order_by=\"-last_accessed\")\n subkey_table.paginate(per_page=10)\n\n transaction_obj_qset = transaction.objects.filter(account_id=account_record.id)\n transaction_table = TransactionTable(transaction_obj_qset, order_by=\"-id\")\n transaction_table.paginate(per_page=10)\n\n login_obj_qset = access.objects.filter(account_id=account_record.id)\n login_table = LoginTable(login_obj_qset, order_by=\"-id\")\n login_table.paginate(per_page=10)\n\n html_dtc = dict(index_limit=account_record.index_limit,\n deposit_address=request.session.get('deposit_address'),\n balance=balance,\n credit=credit,\n price=price,\n affordable_keys=affordable_keys,\n remaining_key_allowance=remaining_key_allowance,\n login_table=login_table,\n subkey_table=subkey_table,\n subkey_table_exists=True if subkey_obj_qset.exists() else False,\n transaction_table=transaction_table,\n transaction_table_exists=True if transaction_obj_qset.exists() else False,\n verified_status=account_record.verified_status,\n logged_in=True,\n account_form=account_form,\n refund_form=False,\n encrypted_str=request.session.get('encrypted_str'),\n minimum_deposit=float(network_transaction_fee))\n\n # Include refund form if balance is high enough\n if balance > network_transaction_fee:\n refund_form = RefundForm()\n refund_form.fields['refund_address'] = forms.CharField(label='BTC Withdrawal Address\\n')\n html_dtc.update(refund_form=refund_form)\n\n if request.session.get('error_msg'):\n html_dtc['error_msg'] = request.session['error_msg']\n del request.session['error_msg']\n\n if not account_record.verified_status:\n if request.session.get('encrypted_str'):\n html_dtc.update(encrypted_str=request.session.get('encrypted_str'))\n\n html_dtc.update(csrf(request))\n set_pass_token(request, u'dashboard')\n return render_to_response('popsite/dashboard.html', html_dtc, RequestContext(request))\n #endregion\n\ndef verify(request):\n #region POST\n if request.POST:\n\n if not verify_pass_token(request, 'verify'):\n return redirect('login')\n\n account_record = account.objects.filter(id=request.session['account_id'])[0]\n if not request.session['account_url'] == account_record.url:\n request.session['error_msg'] = 'Database conflict error'\n return redirect('login')\n\n set_pass_token(request, u'dashboard')\n if request.POST.get('back_button'):\n return redirect('dashboard')\n\n if request.POST.get('decoded_str', False) == request.session.get('decrypted_str', None):\n account_record = account.objects.filter(id=request.session['account_id'])[0]\n account_record.verified_status = True\n account_record.save()\n\n request.session['encrypted_str'] = False\n request.session['decrypted_str'] = False\n\n else:\n request.session['error_msg'] = 'Invalid decrypted string'\n\n return redirect('dashboard')\n #endregion\n #region GET\n if not verify_pass_token(request, 'verify'):\n return redirect('login')\n\n verification_form = VerificationForm()\n\n html_dtc = dict(verification_form=verification_form,\n encrypted_str=request.session.get('encrypted_str'))\n\n if request.session.get('error_msg'):\n html_dtc.update(request.session.get('error_msg'))\n del request.session['error_msg']\n\n html_dtc.update(csrf(request))\n set_pass_token(request, u'verify')\n return render_to_response('popsite/verify.html', html_dtc)\n #endregion\n","repo_name":"cryptoproofinfo/webapp","sub_path":"popsite/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":56128,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"41002123160","text":"from django import forms\nfrom zentral.contrib.inventory.models import BusinessUnit\nfrom .models import Tenant\n\n\nclass TenantForm(forms.ModelForm):\n client_secret = forms.CharField(widget=forms.PasswordInput(render_value=True))\n\n class Meta:\n model = Tenant\n fields = \"__all__\"\n\n widgets = {\n \"client_id\": forms.TextInput,\n \"name\": forms.TextInput,\n \"tenant_id\": forms.TextInput,\n \"description\": forms.Textarea(attrs={\"rows\": \"2\"}),\n }\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.order_fields([\n \"business_unit\",\n \"name\", \"description\",\n \"tenant_id\", \"client_id\",\n \"client_secret\",\n ])\n self.fields[\"business_unit\"].queryset = (\n BusinessUnit.objects.filter(source__module=\"zentral.contrib.inventory\")\n .order_by('meta_business_unit__name')\n )\n if self.instance.pk:\n self.fields[\"client_secret\"].initial = self.instance.get_client_secret()\n\n def save(self):\n new_tenant = self.instance.pk is None\n if new_tenant:\n tenant = super().save()\n else:\n tenant = super().save(commit=False)\n tenant.set_client_secret(self.cleaned_data[\"client_secret\"])\n tenant.save()\n return tenant\n","repo_name":"zentralopensource/zentral","sub_path":"ee/zentral/contrib/intune/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1396,"program_lang":"python","lang":"en","doc_type":"code","stars":705,"dataset":"github-code","pt":"96"} +{"seq_id":"19674643811","text":"#! /usr/bin/env python3\n# ---------------------------------------------------------------\n# Caterham 7 Dash\n# Revision : 0.1\n# Author : Mark Rodman\n# ---------------------------------------------------------------\n# Reads information from the MBE ECU for display on screen via\n# pygame.\n# ---------------------------------------------------------------\n# Status: incomplete\n# ---------------------------------------------------------------\n# Notes,\n# ------\n# __Dictionaries__\n# data_dict = dictionary of data excluding RPM, eg TPS Site : 10. Contents initially built from data_value_labels\n# __Lists__\n# data_txt_as_list = list of instances of DataText, used to print data on the screen.\n\nimport os, sys, time, datetime\nimport random\nimport can\nfrom dash_support import *\nfrom colours import *\nfrom gauges_text import *\nfrom pygame.locals import *\nfrom candata import *\nfrom display_assets import *\n\n\n# Initial setup, needs restructuring, shouldn't be global!\n# init pygame\npygame.init()\npygame.font.init()\npygame.mixer.quit() # bug fix, killing mixer stop CPU hog!!!\n\n# Pygame event for CANBus\nPOLLCAN = pygame.USEREVENT + 1\npygame.time.set_timer(POLLCAN, PollCAN_schedule)\n\n# sort display surface\ngameDisplay = pygame.display.set_mode((display_width, display_height))\nwindowSurface = pygame.display.set_mode((display_width, display_height), 0, 32)\npygame.display.set_caption(display_title)\nwindowSurface.fill(STARTCOLOUR)\n\n# setup fonts\navailable_fonts = pygame.font.get_fonts()\nfor font in range(len(available_fonts)):\n if available_fonts[font] == LCD_font:\n lcd_fontpath = pygame.font.match_font(available_fonts[font])\n if available_fonts[font] == 'hack':\n hack_font = pygame.font.match_font(available_fonts[font])\n\n# set up fonts\nrpmFont = pygame.font.Font(hack_font, rpm_fontsize)\nlabelFont = pygame.font.Font(hack_font, label_fontsize)\ndataFont = pygame.font.Font(hack_font, data_fontsize)\n# setup clock\nclock = pygame.time.Clock()\n# ------------------------------------------------------------------\n\n\ndef demo_rpm(demo_rpm_val):\n if demo_rpm_val < max_rpm:\n demo_rpm_val += 25\n else:\n demo_rpm_val = 0\n return demo_rpm_val\n\n\ndef processing_loop(bus):\n table_collect = table_collect_start\n # setup screen layout, borders etc\n draw_screen_borders(windowSurface)\n draw_screen_labels(windowSurface, labelFont, 3, 40, 30)\n data_txt_as_list = list_data_text(windowSurface, dataFont, 160, 40, 30)\n\n # declare rpm txt instance and display 0000 value\n rpm_txt = SplitDataText(\"rpm\", windowSurface, hack_font, rpm_fontsize, 0.9, ([GREEN, TEXT_BG]),\n ([V_DARK_GREEN, TEXT_BG]), [420, 160])\n\n # declare rpm gauge instance and display bar for zero value\n rpm_bar = DisplayBarGauge(\"test\", 0, max_rpm, windowSurface,\n ([rev_image1, rev_image2, rev_image3, rev_image_shift]),\n BLACK, ([10, 15]), ([2500, 6600, 7500]))\n\n rpm_dial_gauge = DisplayDialGauge(windowSurface, [330, 55, 325, 325], 2, GAUGE_BORDER_COLOUR)\n\n trace_gauge = DisplayTraceGauge(windowSurface, ([0, 365]), 100, ([DARK_GREEN, BLACK]), (7800, 0), False, True)\n\n # create data dictionary\n data_dict = dict.fromkeys(data_value_labels, 0)\n\n keep_running = True\n demo_loop = False\n random_loop = False\n demo_rpm_val = 0\n rpm_reading = Rpmval(\"rpm\", 0)\n\n while keep_running:\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n keep_running = False\n pygame.quit()\n sys.exit()\n elif event.type == POLLCAN:\n if demo_loop:\n demo_rpm_val = demo_rpm(demo_rpm_val)\n rpm_reading.set_change(demo_rpm_val)\n if random_loop:\n rpm_reading.set_change(random.randint(1, max_rpm))\n\n else:\n if live: # collect RPM Data via can at base frequency set by pygame clock\n rough_str, hex_id, data_hex = receive_can_frame(bus)\n rpm_value = process_can_message(rough_str)\n rpm_reading.set_change(rpm_value)\n\n if table_collect == 0: # collect table data, at lower frequency\n # poll for table data from CAN here. # collect goes here\n table_collect = table_collect_start # reset timer used to lower frequency\n table_collect -= 1\n else: # testing loop used to display data\n if table_collect == 0: # again check frequency counter\n data_dict['TPS Site'] = (random.randint(1, 16)) # test update, TPS Site\n data_dict['Air Temp'] = (random.randint(1, 30)) # test update, Air Temp\n data_dict['Coolant Temp'] = (random.randint(1, 110)) # test update, Coolant Temp\n data_dict['Battery Volt'] = (round(random.uniform(1, 13), 2)) # test update, Battery Volt\n data_dict['Throttle Angle'] = (round(random.uniform(1, 5.00), 3))\n table_collect = table_collect_start # reset the counter\n table_collect -= 1 # dec the counter\n\n elif event.type == KEYDOWN:\n demo_loop = False\n random_loop = False\n if event.key == K_UP:\n rpm_reading.test_change(250)\n if event.key == K_DOWN:\n rpm_reading.test_change(-100)\n if event.key == K_LEFT:\n rpm_reading.reset_current_val(0)\n if event.key == K_RIGHT:\n rpm_reading.wipe()\n if event.key == K_LSHIFT:\n demo_loop = True\n if event.key == K_RSHIFT:\n random_loop = True\n\n if not rpm_reading.rx_val_inc:\n rpm_dial_gauge.draw_wiper_arc()\n\n # bug fix to stop zero values from the keyboard. -- code improvement needed\n if rpm_reading.rx_val < 0:\n rpm_reading.rx_val = 0\n\n rpm_bar.updatebar(rpm_reading.rx_val)\n rpm_dial_gauge.data_arc(rpm_reading.rx_val)\n rpm_txt.update(rpm_reading.rx_val)\n trace_gauge.update(rpm_reading.rx_val)\n\n # update main data table text values from -----\n for data in (data_dict.items()):\n for i in range(len(data_txt_as_list)):\n if data_txt_as_list[i].name == data[0]: # match name of dict item with instance name! and if match\n data_txt_as_list[i].update(data[1]) # update instance data, to update screen etc.\n # ---------------------------------------------\n\n pygame.display.update()\n clock.tick(clock_val)\n return\n\n\ndef main():\n if not live:\n bus = ''\n else:\n bus = can.interface.Bus(bustype=bustype, channel=mybus, bitrate=bitrate)\n processing_loop(bus)\n return\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"Purplemeanie/Caterham-Dashboard","sub_path":"dash.py","file_name":"dash.py","file_ext":"py","file_size_in_byte":7541,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"9282735340","text":"__license__ = \"GNU Affero General Public License http://www.gnu.org/licenses/agpl.html\"\n__copyright__ = \"Copyright (C) 2022 The OctoPrint Project - Released under terms of the AGPLv3 License\"\n\nimport threading\n\nimport requests\nfrom flask_babel import gettext\n\nimport octoprint.plugin\nfrom octoprint.schema.webcam import RatioEnum, Webcam, WebcamCompatibility\nfrom octoprint.webcams import WebcamNotAbleToTakeSnapshotException\n\n\nclass ClassicWebcamPlugin(\n octoprint.plugin.AssetPlugin,\n octoprint.plugin.TemplatePlugin,\n octoprint.plugin.SettingsPlugin,\n octoprint.plugin.WebcamProviderPlugin,\n octoprint.plugin.WizardPlugin,\n):\n def __init__(self):\n self._capture_mutex = threading.Lock()\n self._webcam_name = \"classic\"\n\n # ~~ TemplatePlugin API\n\n def get_assets(self):\n return {\n \"js\": [\n \"js/classicwebcam.js\",\n \"js/classicwebcam_settings.js\",\n \"js/classicwebcam_wizard.js\",\n ],\n \"less\": [\"less/classicwebcam.less\"],\n \"css\": [\"css/classicwebcam.css\"],\n }\n\n def get_template_configs(self):\n return [\n {\n \"type\": \"settings\",\n \"template\": \"classicwebcam_settings.jinja2\",\n \"custom_bindings\": True,\n },\n {\n \"type\": \"webcam\",\n \"name\": \"Classic Webcam\",\n \"template\": \"classicwebcam_webcam.jinja2\",\n \"custom_bindings\": True,\n \"suffix\": \"_real\",\n },\n {\n \"type\": \"wizard\",\n \"name\": \"Classic Webcam Wizard\",\n \"template\": \"classicwebcam_wizard.jinja2\",\n \"suffix\": \"_wizard\",\n },\n ]\n\n # ~~ WebcamProviderPlugin API\n\n def get_webcam_configurations(self):\n streamRatio = self._settings.get([\"streamRatio\"])\n if streamRatio == \"4:3\":\n streamRatio = RatioEnum.four_three\n else:\n streamRatio = RatioEnum.sixteen_nine\n webRtcServers = self._settings.get([\"streamWebrtcIceServers\"])\n cacheBuster = self._settings.get_boolean([\"cacheBuster\"])\n stream = self._get_stream_url()\n snapshot = self._get_snapshot_url()\n flipH = self._settings.get_boolean([\"flipH\"])\n flipV = self._settings.get_boolean([\"flipV\"])\n rotate90 = self._settings.get_boolean([\"rotate90\"])\n snapshotSslValidation = self._settings.get_boolean([\"snapshotSslValidation\"])\n\n try:\n streamTimeout = int(self._settings.get([\"streamTimeout\"]))\n except Exception:\n streamTimeout = 5\n\n try:\n snapshotTimeout = int(self._settings.get([\"snapshotTimeout\"]))\n except Exception:\n snapshotTimeout = 5\n\n return [\n Webcam(\n name=self._webcam_name,\n displayName=\"Classic Webcam\",\n flipH=flipH,\n flipV=flipV,\n rotate90=rotate90,\n snapshotDisplay=snapshot,\n canSnapshot=self._can_snapshot(),\n compat=WebcamCompatibility(\n stream=stream,\n streamTimeout=streamTimeout,\n streamRatio=streamRatio,\n cacheBuster=cacheBuster,\n streamWebrtcIceServers=webRtcServers,\n snapshot=snapshot,\n snapshotTimeout=snapshotTimeout,\n snapshotSslValidation=snapshotSslValidation,\n ),\n extras=dict(\n stream=stream,\n streamTimeout=streamTimeout,\n streamRatio=streamRatio,\n streamWebrtcIceServers=webRtcServers,\n cacheBuster=cacheBuster,\n ),\n ),\n ]\n\n def _get_snapshot_url(self):\n return self._settings.get([\"snapshot\"])\n\n def _get_stream_url(self):\n return self._settings.get([\"stream\"])\n\n def _can_snapshot(self):\n snapshot = self._get_snapshot_url()\n return snapshot is not None and snapshot.strip() != \"\"\n\n def take_webcam_snapshot(self, _):\n snapshot_url = self._get_snapshot_url()\n if not self._can_snapshot():\n raise WebcamNotAbleToTakeSnapshotException(self._webcam_name)\n\n with self._capture_mutex:\n self._logger.debug(f\"Capturing image from {snapshot_url}\")\n r = requests.get(\n snapshot_url,\n stream=True,\n timeout=self._settings.get_int([\"snapshotTimeout\"]),\n verify=self._settings.get_boolean([\"snapshotSslValidation\"]),\n )\n r.raise_for_status()\n return r.iter_content(chunk_size=1024)\n\n # ~~ SettingsPlugin API\n\n def get_settings_defaults(self):\n return dict(\n flipH=False,\n flipV=False,\n rotate90=False,\n stream=\"\",\n streamTimeout=5,\n streamRatio=\"16:9\",\n streamWebrtcIceServers=[\"stun:stun.l.google.com:19302\"],\n snapshot=\"\",\n cacheBuster=False,\n snapshotSslValidation=True,\n snapshotTimeout=5,\n )\n\n def get_settings_version(self):\n return 1\n\n def on_settings_migrate(self, target, current):\n if current is None:\n config = self._settings.global_get([\"webcam\"])\n if config:\n self._logger.info(\n \"Migrating settings from webcam to plugins.classicwebcam...\"\n )\n\n # flipH\n self._settings.set_boolean([\"flipH\"], config.get(\"flipH\", False))\n self._settings.global_remove([\"webcam\", \"flipH\"])\n\n # flipV\n self._settings.set_boolean([\"flipV\"], config.get(\"flipV\", False))\n self._settings.global_remove([\"webcam\", \"flipV\"])\n\n # rotate90\n self._settings.set_boolean([\"rotate90\"], config.get(\"rotate90\", False))\n self._settings.global_remove([\"webcam\", \"rotate90\"])\n\n # stream\n self._settings.set([\"stream\"], config.get(\"stream\", \"\"))\n self._settings.global_remove([\"webcam\", \"stream\"])\n\n # streamTimeout\n self._settings.set_int([\"streamTimeout\"], config.get(\"streamTimeout\", 5))\n self._settings.global_remove([\"webcam\", \"streamTimeout\"])\n\n # streamRatio\n self._settings.set([\"streamRatio\"], config.get(\"streamRatio\", \"16:9\"))\n self._settings.global_remove([\"webcam\", \"streamRatio\"])\n\n # streamWebrtcIceServers\n self._settings.set(\n [\"streamWebrtcIceServers\"],\n config.get(\n \"streamWebrtcIceServers\", [\"stun:stun.l.google.com:19302\"]\n ),\n )\n self._settings.global_remove([\"webcam\", \"streamWebrtcIceServers\"])\n\n # snapshot\n self._settings.set([\"snapshot\"], config.get(\"snapshot\", \"\"))\n self._settings.global_remove([\"webcam\", \"snapshot\"])\n\n # cacheBuster\n self._settings.set_boolean(\n [\"cacheBuster\"], config.get(\"cacheBuster\", False)\n )\n self._settings.global_remove([\"webcam\", \"cacheBuster\"])\n\n # snapshotTimeout\n self._settings.set_int(\n [\"snapshotTimeout\"], config.get(\"snapshotTimeout\", 5)\n )\n self._settings.global_remove([\"webcam\", \"snapshotTimeout\"])\n\n # snapshotSslValidation\n self._settings.set_boolean(\n [\"snapshotSslValidation\"], config.get(\"snapshotSslValidation\", True)\n )\n self._settings.global_remove([\"webcam\", \"snapshotSslValidation\"])\n\n # ~~ WizardPlugin API\n\n def is_wizard_required(self):\n required = (\n not self._get_stream_url()\n or not self._get_snapshot_url()\n or not self._settings.global_get([\"webcam\", \"ffmpegPath\"])\n )\n firstrun = self._settings.global_get([\"server\", \"firstRun\"])\n return required and firstrun\n\n def get_wizard_version(self):\n return 1\n\n\n__plugin_name__ = gettext(\"Classic Webcam\")\n__plugin_author__ = \"Christian Würthner\"\n__plugin_description__ = \"Provides a simple webcam viewer in OctoPrint's UI, images provided by an MJPEG webcam.\"\n__plugin_disabling_discouraged__ = gettext(\n \"This plugin provides the standard webcam in OctoPrint. If you do not have any other plugin providing a webcam set up, the webcam section in the control tab will no longer be visible.\"\n)\n__plugin_license__ = \"AGPLv3\"\n__plugin_pythoncompat__ = \">=3.7,<4\"\n__plugin_implementation__ = ClassicWebcamPlugin()\n","repo_name":"OctoPrint/OctoPrint","sub_path":"src/octoprint/plugins/classicwebcam/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":8962,"program_lang":"python","lang":"en","doc_type":"code","stars":7739,"dataset":"github-code","pt":"96"} +{"seq_id":"20253948763","text":"from bs4 import BeautifulSoup\nimport requests\nfrom api_manager import ApiManager\n\ndate = input(\"Which year do you want to travel to? Type the date in this format YYYY-MM-DD: \")\nurl = f\"https://www.billboard.com/charts/hot-100/{date}\"\nresponse = requests.get(url)\nwebsite_html = response.text\nsoup = BeautifulSoup(website_html, \"html.parser\")\nmusic_tags = soup.select(selector=\"li ul li h3\")\nartist_tags = soup.select(selector=\"li ul li span\")\nmusics = [tag.text.strip() for tag in music_tags]\nartists = [tag.text.strip() for tag in artist_tags][::7]\n\nplaylist_name = f\"Billboard Hot 100 When {date}\"\nmanager = ApiManager()\nplaylist_id = manager.create_playlist(playlist_name)\n\ntrack_ids = [manager.search_musics_id(music) for music in musics]\nmanager.add_musics_into_playlist(track_ids=track_ids, playlist_id=playlist_id)\n","repo_name":"mayushen/BillboardHot100-When-CustomDate-SpotifyAutoPlaylist-Project","sub_path":"the-Billboard-Hot100/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":822,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"38916875510","text":"import os\nfrom datetime import datetime\n\nimport django\nimport pandas\nfrom apscheduler.schedulers.background import BackgroundScheduler\nfrom django.contrib.auth.hashers import make_password\nfrom django.db import transaction\n\n\ndef start():\n \"\"\" 스케줄러에 작업을 등록한 후 실행합니다. \"\"\"\n\n scheduler = BackgroundScheduler()\n # scheduler.add_job(check_current_time, 'interval', seconds=1)\n scheduler.add_job(update_data, 'cron', hour=0)\n scheduler.start()\n\n\ndef check_current_time():\n \"\"\" 현재 시간을 출력합니다. \"\"\"\n\n print(\"Now : %s\" % datetime.now())\n\n\n@transaction.atomic()\ndef update_data():\n \"\"\" 엑셀 파일로부터 데이터를 읽어 DB에 저장합니다. \"\"\"\n\n from apps.account.models import Account, AccountAsset\n from apps.asset.models import Asset\n from apps.user.models import User\n\n asset_group_info_set = pandas.read_excel('data/asset_group_info_set.xlsx')\n account_asset_info_set = pandas.read_excel('data/account_asset_info_set.xlsx')\n account_basic_info_set = pandas.read_excel('data/account_basic_info_set.xlsx')\n\n for i, row in asset_group_info_set.iterrows():\n print(i)\n Asset.objects.get_or_create(\n isin=row['ISIN'],\n defaults={\n 'asset_name': row['종목명'],\n 'asset_group': row['자산그룹']\n }\n )\n\n for i, row in account_asset_info_set.iterrows():\n print(i)\n user, created = User.objects.get_or_create(\n name=row['고객이름'],\n defaults={\n 'username': 'user' + str(i).zfill(4),\n 'password': make_password('testtest')\n }\n )\n\n account, created = Account.objects.get_or_create(\n account_number=row['계좌번호'],\n defaults={\n 'user': user,\n 'stock_firm': row['증권사'],\n 'account_name': row['계좌명']\n }\n )\n\n asset = Asset.objects.get(isin=row['ISIN'])\n AccountAsset.objects.get_or_create(\n account=account,\n asset=asset,\n defaults={\n 'current_price': row['현재가'],\n 'count': row['보유수량']\n }\n )\n\n for i, row in account_basic_info_set.iterrows():\n print(i)\n account = Account.objects.get(account_number=row['계좌번호'])\n account.principal = int(row['투자원금'])\n account.save()\n\n\nif __name__ == '__main__':\n # django 환경 로드\n os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'config.settings.dev')\n django.setup()\n\n update_data()\n","repo_name":"BomiChoi/fint-server-api","sub_path":"updater.py","file_name":"updater.py","file_ext":"py","file_size_in_byte":2667,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"33317180514","text":"# URL = \"https://replit.com/@JackobWave/day-38#main.py\"\n\nimport os\nfrom datetime import datetime\nimport requests\n\nGENDER = \"male\"\nWEIGHT_KG = 59\nHEIGHT_CM = 174\nAGE = 24\n\nNUTRITIONX_API_KEY = os.environ['NUTRITIONX_API_KEY']\nNUTRITIONX_ID = os.environ['NUTRITIONX_ID']\nMAIL = os.environ['MAIL']\nMAIL_PASSWORD = os.environ['MAIL_PASSWORD']\n\n\nsheety_api = os.environ['sheety_api']\nnutritionx_api_endpoint = \"https://trackapi.nutritionix.com/v2/natural/exercise\"\n\nnutri_headers = {\n \"x-app-id\": NUTRITIONX_ID,\n \"x-app-key\": NUTRITIONX_API_KEY\n}\nnutrition_params = {\n \"query\": input(\"Tell me which exercises you did: \"),\n \"gender\": GENDER,\n \"weight_kg\": WEIGHT_KG,\n \"height_cm\": HEIGHT_CM,\n \"age\": AGE\n}\nsheety_headers = {\n \"Authorization\": os.environ['Bearer'],\n}\ntoday_date = datetime.now().strftime(\"%d/%m/%Y\")\nnow_time = datetime.now().strftime(\"%X\")\n\n# post needs data= where get needs params=\nresponse = requests.post(url=nutritionx_api_endpoint, data=nutrition_params, headers=nutri_headers)\nresult = response.json()\nprint(result)\nfor exercise in result[\"exercises\"]:\n sheet_inputs = {\n \"workout\": {\n \"date\": today_date,\n \"time\": now_time,\n \"exercise\": exercise[\"name\"].title(),\n \"duration\": exercise[\"duration_min\"],\n \"calories\": exercise[\"nf_calories\"]\n }\n }\n response = requests.post(url=sheety_api, json=sheet_inputs, headers=sheety_headers)\n\nprint(response.text)\n\n","repo_name":"JakubWij/100_days_of_code_learning_journey","sub_path":"day-38-workout track app/workout track with environ values, repl.it.py","file_name":"workout track with environ values, repl.it.py","file_ext":"py","file_size_in_byte":1474,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"27512625891","text":"from numpy import abs as abs_, logical_or as or_\n\nfrom openfisca_core.periods import Period\n\nfrom openfisca_france.model.base import *\n\n\nclass inapte_travail(Variable):\n value_type = bool\n entity = Individu\n label = 'Reconnu inapte au travail'\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n\nclass asi_aspa_base_ressources_individu(Variable):\n value_type = float\n label = 'Base ressources individuelle du minimum vieillesse/ASPA'\n entity = Individu\n definition_period = MONTH\n set_input = set_input_divide_by_period\n\n def formula(individu, period, parameters):\n last_year = period.last_year\n three_previous_months = period.last_3_months\n law = parameters(period)\n leg_1er_janvier = parameters(period.start.offset('first-of', 'year'))\n\n ressources_incluses = [\n 'allocation_securisation_professionnelle',\n 'chomage_net',\n 'dedommagement_victime_amiante',\n 'gains_exceptionnels',\n 'indemnites_chomage_partiel',\n 'indemnites_journalieres',\n 'indemnites_volontariat',\n 'pensions_alimentaires_percues',\n 'pensions_invalidite',\n 'prestation_compensatoire',\n 'prime_forfaitaire_mensuelle_reprise_activite',\n 'retraite_brute',\n 'revenus_stage_formation_pro',\n 'rsa_base_ressources_patrimoine_individu',\n 'salaire_de_base',\n 'traitement_indiciaire_brut'\n ]\n\n # Revenus du foyer fiscal que l'on projette sur le premier invidividu\n rente_viagere_titre_onereux_foyer_fiscal = individu.foyer_fiscal('rente_viagere_titre_onereux', three_previous_months, options = [ADD])\n revenus_foyer_fiscal_individu = rente_viagere_titre_onereux_foyer_fiscal * individu.has_role(FoyerFiscal.DECLARANT_PRINCIPAL)\n\n def revenus_tns():\n revenus_auto_entrepreneur = individu('rpns_auto_entrepreneur_benefice', three_previous_months, options = [ADD])\n # Les revenus TNS hors AE sont estimés en se basant sur le revenu N-1\n\n rpns_micro_entreprise_benefice = individu('rpns_micro_entreprise_benefice', last_year) * (3 / 12)\n rpns_benefice_exploitant_agricole = individu('rpns_benefice_exploitant_agricole', last_year) * (3 / 12)\n rpns_micro_entreprise_beneficens_autres_revenus = individu('rpns_autres_revenus', last_year) * (3 / 12)\n\n return (\n revenus_auto_entrepreneur\n + rpns_micro_entreprise_benefice\n + rpns_benefice_exploitant_agricole\n + rpns_micro_entreprise_beneficens_autres_revenus\n )\n\n pension_invalidite = (individu('pensions_invalidite', period) > 0)\n aspa_eligibilite = individu('aspa_eligibilite', period)\n asi_eligibilite = individu('asi_eligibilite', period)\n\n # Exclut l'AAH si éligible ASPA, retraite ou pension invalidité\n # en application du II.B. de http://www.legislation.cnav.fr/Pages/texte.aspx?Nom=LE_MIN_19031982\n aah = individu('aah', three_previous_months, options = [ADD])\n aah = aah * not_(aspa_eligibilite) * not_(asi_eligibilite) * not_(pension_invalidite)\n\n pensions_alimentaires_versees = individu(\n 'pensions_alimentaires_versees_individu', three_previous_months, options = [ADD]\n )\n\n def abattement_salaire():\n aspa_couple = individu.famille('aspa_couple', period)\n\n # Abattement sur les salaires (appliqué sur une base trimestrielle)\n abattement_forfaitaire_base = (\n leg_1er_janvier.marche_travail.salaire_minimum.smic.smic_b_horaire * law.marche_travail.salaire_minimum.smic.nb_heures_travail_mensuel\n )\n\n taux_abattement_forfaitaire = where(\n aspa_couple,\n law.prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa.abattement_forfaitaire.couples,\n law.prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa.abattement_forfaitaire.personnes_seules\n )\n\n abattement_forfaitaire = abattement_forfaitaire_base * taux_abattement_forfaitaire\n salaire_de_base = individu('salaire_de_base', three_previous_months, options = [ADD])\n traitement_indiciaire_brut = individu('traitement_indiciaire_brut', three_previous_months, options = [ADD])\n return min_(salaire_de_base + traitement_indiciaire_brut, abattement_forfaitaire)\n\n base_ressources_3_mois = sum(\n max_(0, individu(ressource_type, three_previous_months, options = [ADD]))\n for ressource_type in ressources_incluses\n ) + aah + revenus_foyer_fiscal_individu + revenus_tns() - abs_(pensions_alimentaires_versees) - abattement_salaire()\n\n return base_ressources_3_mois / 3\n\n\nclass asi_aspa_base_ressources(Variable):\n value_type = float\n label = 'Base ressource du minimum vieillesse et assimilés (ASPA)'\n entity = Famille\n definition_period = MONTH\n set_input = set_input_divide_by_period\n\n def formula(famille, period):\n base_ressources_i = famille.members('asi_aspa_base_ressources_individu', period)\n ass_i = famille.members('ass', period)\n return famille.sum(base_ressources_i + ass_i, role = Famille.PARENT)\n\n\nclass aspa_eligibilite(Variable):\n value_type = bool\n label = \"Indicatrice individuelle d'éligibilité à l'allocation de solidarité aux personnes agées\"\n entity = Individu\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n def formula(individu, period, parameters):\n age = individu('age', period)\n inapte_travail = individu('inapte_travail', period)\n taux_incapacite = individu('taux_incapacite', period)\n aah = parameters(period).prestations_sociales.prestations_etat_de_sante.invalidite.aah\n aspa = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa\n\n condition_invalidite = (taux_incapacite > aspa.taux_incapacite_aspa_anticipe) + inapte_travail\n condition_age_base = (age >= aspa.age_min)\n condition_age_anticipe = (age >= aah.age_legal_retraite) * condition_invalidite\n condition_age = condition_age_base + condition_age_anticipe\n condition_nationalite = individu('asi_aspa_condition_nationalite', period)\n\n return condition_age * condition_nationalite\n\n\nclass asi_eligibilite(Variable):\n value_type = bool\n label = \"Indicatrice individuelle d'éligibilité à l'allocation supplémentaire d'invalidité\"\n entity = Individu\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n def formula(individu, period):\n last_month = Period(('month', period.start, 1)).offset(-1)\n\n non_eligible_aspa = not_(individu('aspa_eligibilite', period))\n touche_pension_invalidite = individu('pensions_invalidite', period) > 0\n handicap = individu('handicap', period)\n touche_retraite = individu('retraite_nette', last_month) > 0\n condition_nationalite = individu('asi_aspa_condition_nationalite', period)\n\n eligible = (\n non_eligible_aspa\n * condition_nationalite\n * (handicap * touche_retraite + touche_pension_invalidite)\n )\n\n return eligible\n\n\nclass asi_aspa_condition_nationalite(Variable):\n value_type = bool\n default_value = True\n label = \"Condition de nationalité et de titre de séjour pour bénéficier de l'ASPA ou l'ASI\"\n entity = Individu\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n def formula(individu, period, parameters):\n ressortissant_eee = individu('ressortissant_eee', period)\n ressortissant_suisse = individu('nationalite', period) == b'CH'\n duree_possession_titre_sejour = individu('duree_possession_titre_sejour', period)\n duree_min_titre_sejour = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa.duree_min_titre_sejour\n\n return or_(ressortissant_eee, ressortissant_suisse, duree_possession_titre_sejour >= duree_min_titre_sejour)\n\n\nclass asi_aspa_nb_alloc(Variable):\n value_type = int\n label = \"Nombre d'allocataires ASI/ASPA\"\n entity = Famille\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n def formula(famille, period, parameters):\n aspa_elig_i = famille.members('aspa_eligibilite', period)\n asi_elig_i = famille.members('asi_eligibilite', period)\n\n nb_allocataire_asi = famille.sum(asi_elig_i, role = Famille.PARENT)\n nb_allocataire_aspa = famille.sum(aspa_elig_i, role = Famille.PARENT)\n\n return nb_allocataire_asi + nb_allocataire_aspa\n\n\nclass asi(Variable):\n value_type = float\n label = \"Allocation supplémentaire d'invalidité (ASI)\"\n entity = Individu\n definition_period = MONTH\n reference = 'https://www.legifrance.gouv.fr/codes/id/LEGISCTA000006156277/'\n set_input = set_input_divide_by_period\n calculate_output = calculate_output_add\n\n def formula_2020_04_01(individu, period, parameters):\n maries = individu.famille('maries', period)\n en_couple = individu.famille('en_couple', period)\n asi_aspa_nb_alloc = individu.famille('asi_aspa_nb_alloc', period)\n base_ressources = individu.famille('asi_aspa_base_ressources', period)\n asi = parameters(period).prestations_sociales.prestations_etat_de_sante.invalidite.asi\n aspa = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa\n\n demandeur_eligible_asi = individu.famille.demandeur('asi_eligibilite', period)\n demandeur_eligible_aspa = individu.famille.demandeur('aspa_eligibilite', period)\n conjoint_eligible_asi = individu.famille.conjoint('asi_eligibilite', period)\n conjoint_eligible_aspa = individu.famille.conjoint('aspa_eligibilite', period)\n\n # Un seul éligible\n elig1 = ((asi_aspa_nb_alloc == 1) & (demandeur_eligible_asi | conjoint_eligible_asi))\n # Couple d'éligibles mariés\n elig2 = demandeur_eligible_asi & conjoint_eligible_asi & maries\n # Couple d'éligibles non mariés\n elig3 = demandeur_eligible_asi & conjoint_eligible_asi & not_(maries)\n # Un seul éligible et époux éligible ASPA\n elig4 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & maries\n # Un seul éligible et conjoint non marié éligible ASPA\n elig5 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & not_(maries)\n\n plafond_ressources = (\n elig1 * (asi.plafond_ressource_seul * not_(en_couple) + asi.plafond_ressource_couple * en_couple)\n + elig2 * asi.plafond_ressource_couple\n + elig3 * asi.plafond_ressource_couple\n + elig4 * aspa.plafond_ressources.couples\n + elig5 * aspa.plafond_ressources.couples) / 12\n\n montant_servi_asi = max_(plafond_ressources - base_ressources, 0)\n\n return montant_servi_asi * (\n individu.has_role(Famille.DEMANDEUR) * demandeur_eligible_asi * (elig1 + elig2 / 2 + elig3 / 2)\n + individu.has_role(Famille.CONJOINT) * conjoint_eligible_asi * (elig1 + elig2 / 2 + elig3 / 2)\n )\n\n def formula_2007(individu, period, parameters):\n maries = individu.famille('maries', period)\n en_couple = individu.famille('en_couple', period)\n asi_aspa_nb_alloc = individu.famille('asi_aspa_nb_alloc', period)\n base_ressources = individu.famille('asi_aspa_base_ressources', period)\n asi = parameters(period).prestations_sociales.prestations_etat_de_sante.invalidite.asi\n aspa = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa\n\n demandeur_eligible_asi = individu.famille.demandeur('asi_eligibilite', period)\n demandeur_eligible_aspa = individu.famille.demandeur('aspa_eligibilite', period)\n conjoint_eligible_asi = individu.famille.conjoint('asi_eligibilite', period)\n conjoint_eligible_aspa = individu.famille.conjoint('aspa_eligibilite', period)\n\n # Un seul éligible\n elig1 = ((asi_aspa_nb_alloc == 1) & (demandeur_eligible_asi | conjoint_eligible_asi))\n # Couple d'éligibles mariés\n elig2 = demandeur_eligible_asi & conjoint_eligible_asi & maries\n # Couple d'éligibles non mariés\n elig3 = demandeur_eligible_asi & conjoint_eligible_asi & not_(maries)\n # Un seul éligible et époux éligible ASPA\n elig4 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & maries\n # Un seul éligible et conjoint non marié éligible ASPA\n elig5 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & not_(maries)\n\n montant_max = (\n elig1 * asi.montant_seul\n + elig2 * asi.montant_couple\n + elig3 * 2 * asi.montant_seul\n + elig4 * (asi.montant_couple / 2 + aspa.montant_maximum_annuel.couples / 2)\n + elig5 * (asi.montant_seul + aspa.montant_maximum_annuel.couples / 2)) / 12\n\n ressources = base_ressources + montant_max\n\n plafond_ressources = (\n elig1 * (asi.plafond_ressource_seul * not_(en_couple) + asi.plafond_ressource_couple * en_couple)\n + elig2 * asi.plafond_ressource_couple\n + elig3 * asi.plafond_ressource_couple\n + elig4 * aspa.plafond_ressources.couples\n + elig5 * aspa.plafond_ressources.couples) / 12\n\n depassement = max_(ressources - plafond_ressources, 0)\n\n diff = (\n (elig1 | elig2 | elig3) * (montant_max - depassement)\n + elig4 * (asi.montant_couple / 12 / 2 - depassement / 2)\n + elig5 * (asi.montant_seul / 12 - depassement / 2)\n )\n\n # Montant mensuel servi (sous réserve d'éligibilité)\n montant_servi_asi = max_(diff, 0)\n return montant_servi_asi * (\n individu.has_role(Famille.DEMANDEUR) * demandeur_eligible_asi * (elig1 + elig2 / 2 + elig3 / 2)\n + individu.has_role(Famille.CONJOINT) * conjoint_eligible_asi * (elig1 + elig2 / 2 + elig3 / 2)\n )\n\n\nclass aspa_couple(Variable):\n value_type = bool\n label = \"Couple au sens de l'ASPA\"\n entity = Famille\n definition_period = MONTH\n set_input = set_input_dispatch_by_period\n\n def formula_2002_01_01(famille, period):\n maries = famille('maries', period)\n return maries\n\n def formula_2007_01_01(famille, period):\n en_couple = famille('en_couple', period)\n return en_couple\n\n\nclass aspa(Variable):\n calculate_output = calculate_output_add\n value_type = float\n entity = Famille\n label = 'Allocation de solidarité aux personnes agées'\n reference = 'http://vosdroits.service-public.fr/particuliers/F16871.xhtml'\n definition_period = MONTH\n set_input = set_input_divide_by_period\n\n def formula_2020_04_01(famille, period, parameters):\n maries = famille('maries', period)\n en_couple = famille('en_couple', period)\n asi_aspa_nb_alloc = famille('asi_aspa_nb_alloc', period)\n base_ressources = famille('asi_aspa_base_ressources', period)\n P = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse\n\n demandeur_eligible_asi = famille.demandeur('asi_eligibilite', period)\n demandeur_eligible_aspa = famille.demandeur('aspa_eligibilite', period)\n conjoint_eligible_asi = famille.conjoint('asi_eligibilite', period)\n conjoint_eligible_aspa = famille.conjoint('aspa_eligibilite', period)\n\n # Un seul éligible\n elig1 = ((asi_aspa_nb_alloc == 1) & (demandeur_eligible_aspa | conjoint_eligible_aspa))\n # Couple d'éligibles\n elig2 = (demandeur_eligible_aspa & conjoint_eligible_aspa)\n # Un seul éligible et époux éligible ASI\n elig3 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & maries\n # Un seul éligible et conjoint non marié éligible ASI\n elig4 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & not_(maries)\n\n elig = elig1 | elig2 | elig3 | elig4\n\n montant_asi_demandeur = famille.demandeur('asi', period)\n montant_asi_conjoint = famille.conjoint('asi', period)\n\n montant_max = (\n elig1 * P.aspa.montant_maximum_annuel.personnes_seules\n + elig2 * P.aspa.montant_maximum_annuel.couples\n + elig3 * (montant_asi_demandeur + P.aspa.montant_maximum_annuel.couples / 2)\n + elig4 * (montant_asi_conjoint + P.aspa.montant_maximum_annuel.couples / 2)\n ) / 12\n\n ressources = base_ressources + montant_max\n\n plafond_ressources = (\n elig1\n * (P.aspa.plafond_ressources.personnes_seules * not_(en_couple) + P.aspa.plafond_ressources.couples * en_couple)\n + (elig2 | elig3 | elig4)\n * P.aspa.plafond_ressources.couples\n ) / 12\n\n depassement = max_(ressources - plafond_ressources, 0)\n\n diff = (\n (elig1 | elig2) * (montant_max - depassement)\n + (elig3 | elig4) * (P.aspa.montant_maximum_annuel.couples / 12 / 2 - depassement / 2)\n )\n\n # Montant mensuel servi (sous réserve d'éligibilité)\n montant_servi_aspa = max_(diff, 0)\n\n # TODO: Faute de mieux, on verse l'aspa à la famille plutôt qu'aux individus\n # aspa[CHEF] = demandeur_eligible_aspa*montant_servi_aspa*(elig1 + elig2/2)\n # aspa[PART] = conjoint_eligible_aspa*montant_servi_aspa*(elig1 + elig2/2)\n return elig * montant_servi_aspa\n\n def formula_2006_01_01(famille, period, parameters):\n maries = famille('maries', period)\n en_couple = famille('en_couple', period)\n asi_aspa_nb_alloc = famille('asi_aspa_nb_alloc', period)\n base_ressources = famille('asi_aspa_base_ressources', period)\n asi = parameters(period).prestations_sociales.prestations_etat_de_sante.invalidite.asi\n aspa = parameters(period).prestations_sociales.solidarite_insertion.minimum_vieillesse.aspa\n\n demandeur_eligible_asi = famille.demandeur('asi_eligibilite', period)\n demandeur_eligible_aspa = famille.demandeur('aspa_eligibilite', period)\n conjoint_eligible_asi = famille.conjoint('asi_eligibilite', period)\n conjoint_eligible_aspa = famille.conjoint('aspa_eligibilite', period)\n\n # Un seul éligible\n elig1 = ((asi_aspa_nb_alloc == 1) & (demandeur_eligible_aspa | conjoint_eligible_aspa))\n # Couple d'éligibles\n elig2 = (demandeur_eligible_aspa & conjoint_eligible_aspa)\n # Un seul éligible et époux éligible ASI\n elig3 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & maries\n # Un seul éligible et conjoint non marié éligible ASI\n elig4 = ((demandeur_eligible_asi & conjoint_eligible_aspa) | (conjoint_eligible_asi & demandeur_eligible_aspa)) & not_(maries)\n\n elig = elig1 | elig2 | elig3 | elig4\n\n montant_max = (\n elig1 * aspa.montant_maximum_annuel.personnes_seules\n + elig2 * aspa.montant_maximum_annuel.couples\n + elig3 * (asi.montant_couple / 2 + aspa.montant_maximum_annuel.couples / 2)\n + elig4 * (asi.montant_seul + aspa.montant_maximum_annuel.couples / 2)\n ) / 12\n\n ressources = base_ressources + montant_max\n\n plafond_ressources = (\n elig1\n * (aspa.plafond_ressources.personnes_seules * not_(en_couple) + aspa.plafond_ressources.couples * en_couple)\n + (elig2 | elig3 | elig4)\n * aspa.plafond_ressources.couples\n ) / 12\n\n depassement = max_(ressources - plafond_ressources, 0)\n\n diff = (\n (elig1 | elig2) * (montant_max - depassement)\n + (elig3 | elig4) * (aspa.montant_maximum_annuel.couples / 12 / 2 - depassement / 2)\n )\n\n # Montant mensuel servi (sous réserve d'éligibilité)\n montant_servi_aspa = max_(diff, 0)\n\n # TODO: Faute de mieux, on verse l'aspa à la famille plutôt qu'aux individus\n # aspa[CHEF] = demandeur_eligible_aspa*montant_servi_aspa*(elig1 + elig2/2)\n # aspa[PART] = conjoint_eligible_aspa*montant_servi_aspa*(elig1 + elig2/2)\n return elig * montant_servi_aspa\n","repo_name":"openfisca/openfisca-france","sub_path":"openfisca_france/model/prestations/minima_sociaux/asi_aspa.py","file_name":"asi_aspa.py","file_ext":"py","file_size_in_byte":20771,"program_lang":"python","lang":"fr","doc_type":"code","stars":234,"dataset":"github-code","pt":"96"} +{"seq_id":"18990530983","text":"# Information on Bit Manipulation: https://leetcode.com/problems/sum-of-two-integers/solutions/84278/a-summary-how-to-use-bit-manipulation-to-solve-problems-easily-and-efficiently/\n# Information on how to solve: https://leetcode.com/problems/sum-of-two-integers/solutions/167931/Solution-with-ACTUAL-explanation-(how-you-would-work-this-out)/\n# Information on Masks (also see comment below): https://leetcode.com/problems/sum-of-two-integers/solutions/489210/read-this-if-you-want-to-learn-about-masks/?languageTags=python3\n\nclass Solution:\n def getSum(self, a: int, b: int) -> int:\n # Create a 32-bit mask since Python does not cap at 32-bits\n mask = 0xFFFFFFFF\n\n while (b & mask) > 0:\n \"\"\"\n When a and b are both 1, then we've reached the max in this digit place. \n To account for this, we carry the 1 to the next digit place.\n \"\"\"\n carry = (a & b) << 1\n\n \"\"\"\n If a and b are NOT the same, a will equal 1. \n if they were the same, a will equal 0. In this situation, carry would be equal to 1.\n We assign carry to b and continue the operation until the mask is exceeded or b is 0 \n \"\"\"\n a = a ^ b\n b = carry\n\n \"\"\"\n Return a if there was no further carry at the end of the loop (ie b > 0).\n If there was a carry remaining, use the mask to ensure the value does not go beyond the size threshold.\n Masks will set any values beyond a specific place to 0 when you & it with another value.\n This ensures that the value of a does not continue carrying to infinity, instead converging on a 32 bit number.\n https://stackoverflow.com/questions/10493411/what-is-bit-masking\n \"\"\"\n return a & mask if b > 0 else a","repo_name":"jonotassia/coding-challenges","sub_path":"371 - Sum of Two Integers.py","file_name":"371 - Sum of Two Integers.py","file_ext":"py","file_size_in_byte":1815,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"13554374548","text":"import sys\nfrom confluent_kafka import Consumer\nimport json\nfrom elasticsearch import Elasticsearch\nimport random\n\nes = Elasticsearch([{'host': 'elasticsearch', 'port': 9200}])\n\ntopic = sys.argv[1]\nprint (\"Consuming on topic\", topic)\nprint\n\nconf = {'bootstrap.servers': 'broker:9092', 'client.id': 'test', 'default.topic.config': {'acks': 'all'}, 'group.id': 'mygroup'}\n\nconsumer = Consumer(conf)\nconsumer.subscribe([topic])\nwhile True:\n\tmsg = consumer.poll(1.0)\n\tif msg is None:\n\t\tcontinue\n\tif msg.error():\n\t\tprint(\"Consumer error: {}\".format(msg.error()))\n\t\tcontinue\n\tprint('Message key: {}'.format(msg.key().decode('utf-8')))\n\tprint('Message value: {}'.format(msg.value().decode('utf-8')))\n\tes.index(index=topic, doc_type='jsonfile', id=random.randint(1,10000000), body=format(msg.value().decode('utf-8')))\n\nconsumer.close()","repo_name":"dbspcl/JPL-POC-kafka","sub_path":"python-with-confluent/scripts/gettopic.py","file_name":"gettopic.py","file_ext":"py","file_size_in_byte":827,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"1726713543","text":"# --- Define your functions below! ---\r\nquestion1 = ['hi', 'hello', 'hi!', 'hello!', 'Hi!', 'Hello!', 'Hi', 'Hello', 'yo', 'Yo', 'yo!']\r\n\r\nquestion2sleep = ['sleep', 'Sleep', 'sleep!', 'Sleep!']\r\nquestion2eat = ['eat', 'Eat', 'eat!', 'Eat!']\r\nquestion2wn = ['watch netflix', 'Watch Netflix', 'watch Netflix', 'watch netflix!', 'Watch Netflix!', 'watch Netflix!']\r\nquestionfdmc = ['mac and cheese', 'macaroni and cheese', 'Mac and Cheese', 'Macaroni and Cheese', 'Mac and cheese', 'Macaroni and cheese']\r\nquestiontvsl = ['sherlock', 'Sherlock', 'My favorite tv show is Sherlock', 'my favorite tv show is sherlock', 'my favorite tv show is Sherlock', 'My favortie tv show is sherlock', 'My favorite tv show is Sherlock.', 'My favorite TV show is Sherlock.']\r\n\r\ndef valid_responses(response, validans):\r\n\r\n if response in validans:\r\n return True\r\n return False\r\n\r\n\r\n\r\ndef intro():\r\n response = input(\"Hi! Emma here! I made this chatbox so you can learn a bit about me and I can learn a bit about you!\")\r\n if valid_responses(response, question1):\r\n Name = input(\"(Hi! What is your name?) \")\r\n print(\"Hello,\", (Name))\r\n instructions()\r\n\r\n else:\r\n Name = input(\"(What is your name?) \")\r\n print(\"Hello,\", (Name))\r\n instructions()\r\n\r\n\r\ndef convo():\r\n response = input(\"(What is one of your favorite things to do?) \")\r\n if valid_responses(response, question2sleep):\r\n print(\"(Lol, same. I slept 13-14 hrs in a row before.)\")\r\n\r\n else:\r\n if valid_responses(response, question2eat):\r\n response = input(\"(What is your favorite food?) \")\r\n if valid_responses(response, questionfdmc):\r\n print(\"(Mac and cheese is one of my favorite foods too.) \")\r\n\r\n else:\r\n print(\"(Yummy!) \")\r\n\r\n\r\n else:\r\n if valid_responses(response, question2wn):\r\n response = input(\"(What is your favorite TV show?) \")\r\n if valid_responses(response, questiontvsl):\r\n print(\"(SAMEEEEEEEE one of my favorite episodes is Sign of Three) \")\r\n\r\n\r\n else:\r\n print(\"(Cool!) \")\r\n\r\n\r\n answer = input(\"(What is your favorite color?) \")\r\n\r\n print(\"Yeah!\", (answer), \"is a nice color! My favorite colors are rosy pink and gold.\")\r\n\r\n answer = input(\"(Do you play any sports?) \")\r\n print((\"Lol cool! I did ballet for about 4 years.\"))\r\n\r\n if input == \"yes\":\r\n answer = input(\"(What sport do you play?) \")\r\n print((\"Nice!\"))\r\n\r\n else:\r\n answer = input(\"(What do you do for extracurriculars?) \")\r\n print(\"(Cool! Okay, I'm going to stop saying cool now lol.) \")\r\n\r\ndef instrument():\r\n answer = input(\"(Do you play an instrument?) \")\r\n if answer == \"yes\":\r\n answer = input(\"(What instument do you play?) \")\r\n parameter(answer)\r\n\r\n else:\r\n print(\"(Nice!) \")\r\n\r\ndef parameter(answer):\r\n print((answer),\"is a cool instrument. I played violin for about 4-5 years.\" )\r\n\r\ndef instructions():\r\n print(\"(If you want to exit, just type bye.) \")\r\n\r\ndef bye():\r\n print(\"(Bye!)\", (Name))\r\n\r\n# --- Put your main program below! ---\r\n\r\ndef main():\r\n\r\n while True:\r\n\r\n intro()\r\n\r\n convo()\r\n\r\n instrument()\r\n\r\n\r\n\r\n\r\n\r\n# DON'T TOUCH! Setup code that runs your main() function.\r\n\r\nif __name__ == \"__main__\":\r\n\r\n main()\r\n","repo_name":"probablydoinghw/countryStarter","sub_path":"chatbox2.py","file_name":"chatbox2.py","file_ext":"py","file_size_in_byte":3191,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"96"} +{"seq_id":"9284257789","text":"#pandigital checker for identity\n#only need to cycle from 1 to 9999 i think:\ndef f(a,b):\n count = []\n string = ''\n string += str(a) + str(b) + str(a*b)\n if len(string) == 9:\n for i in range(0,9):\n count.append(int(string[i]))\n if set(count) == {1,2,3,4,5,6,7,8,9}:\n return True\n else:\n return False\n else:\n return False\n\nnum = []\n\nfor i in range(1,10000):\n for j in range(1,10000):\n if f(i,j):\n num.append(i*j)\n\n\nprint(sum(set(num)))\n","repo_name":"DevinAdvani/Project_Euler_Solutions","sub_path":"Solutions/32.py","file_name":"32.py","file_ext":"py","file_size_in_byte":531,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"71899852156","text":"\n\nimport sys\nimport time\nfrom pysinewave import SineWave, BeatWave, BeatWaveGenerator, SineWaveGenerator\nimport numpy\n# from scipy import signal\n# import pygame\n# from matplotlib import pyplot\n\ndef plot_wave():\n swg = SineWaveGenerator(12,10)\n bwg = BeatWaveGenerator(12,50,10)\n\n pyplot.plot(swg.next_data(44100))\n pyplot.plot(bwg.next_data(44100))\n pyplot.show()\n\ndef play_beat():\n sw = BeatWave(12,50,3)\n\n sw.play()\n\n time.sleep(1)\n\n # for i in range(40):\n # sw.set_beat_frequency(5+2*i)\n # time.sleep(0.1)\n # j = 5+2*i\n # for i in range(40):\n # sw.set_beat_frequency(j-2*i)\n # time.sleep(0.1)\n\n sw.set_beat_frequency(5)\n\n time.sleep(1)\n\n sw.set_beat_frequency(10)\n\n time.sleep(2)\n\n sw.set_beat_frequency(25)\n\n time.sleep(2)\n\n sw.set_pitch(5)\n\n time.sleep(2)\n\n sw.stop()\n\ndef play_tone():\n bw = BeatWave(12,5,50)\n sw = SineWave(12,50)\n\n bw.play()\n\n time.sleep(1)\n\n # for i in range(40):\n # sw.set_beat_frequency(5+2*i)\n # time.sleep(0.1)\n # j = 5+2*i\n # for i in range(40):\n # sw.set_beat_frequency(j-2*i)\n # time.sleep(0.1)\n\n bw.set_beat_frequency(0)\n\n time.sleep(1)\n\n bw.stop()\n sw.play()\n\n time.sleep(1)\n\n sw.stop()\n bw.play()\n\n time.sleep(1)\n\n bw.set_pitch(5)\n\n time.sleep(1)\n\n bw.set_pitch(10)\n\n time.sleep(2)\n\n bw.set_pitch(-7)\n\n time.sleep(1)\n\n bw.set_beat_frequency(10)\n\n time.sleep(2)\n\n bw.set_pitch(5)\n\n time.sleep(2)\n\n bw.stop()\n\ndef pygame_test():\n\n sampleRate = 44100\n freq1 = 300\n freq2 = 301\n\n pygame.mixer.init(44100,-16,2,512)\n # sampling frequency, size, channels, buffer\n\n # Sampling frequency\n # Analog audio is recorded by sampling it 44,100 times per second, \n # and then these samples are used to reconstruct the audio signal \n # when playing it back.\n\n # size\n # The size argument represents how many bits are used for each \n # audio sample. If the value is negative then signed sample \n # values will be used.\n\n # channels\n # 1 = mono, 2 = stereo\n\n # buffer\n # The buffer argument controls the number of internal samples \n # used in the sound mixer. It can be lowered to reduce latency, \n # but sound dropout may occur. It can be raised to larger values\n # to ensure playback never skips, but it will impose latency on sound playback. \n\n arr0 = numpy.array([4096 * numpy.sin(2.0 * numpy.pi * freq1 * x / sampleRate) for x in range(0, sampleRate)]).astype(numpy.int16)\n\n arr2 = numpy.array([4096 * numpy.sin(2.0 * numpy.pi * freq2 * x / sampleRate) for x in range(0, sampleRate)]).astype(numpy.int16)\n\n # arr2 = numpy.abs(arr2)\n\n # arr2n = arr2/numpy.amax(arr2)\n\n # arr = (arr0*arr2n).astype(numpy.int16)\n arr = arr0+arr2\n\n pyplot.plot(arr0)\n pyplot.plot(arr2+arr0)\n pyplot.show()\n # ddd\n \n arr2 = numpy.c_[arr,arr]\n sound = pygame.sndarray.make_sound(arr2)\n sound.play(-1)\n pygame.time.delay(4000)\n sound.stop()\n\n freq2 = 304\n\n arr2 = numpy.array([4096 * numpy.sin(2.0 * numpy.pi * freq2 * x / sampleRate) for x in range(0, sampleRate)]).astype(numpy.int16)\n\n arr = arr0+arr2\n\n\n\n # arr2 = numpy.array([4096 * numpy.sin(2.0 * numpy.pi * freq/100 * x / sampleRate) for x in range(0, sampleRate)]).astype(numpy.int16)\n\n # arr2 = numpy.abs(arr2)\n\n # arr2n = arr2/numpy.amax(arr2)\n\n # arr = (arr0*arr2n).astype(numpy.int16)*4\n # arr2 = numpy.c_[arr,arr]\n # sound = pygame.sndarray.make_sound(arr2)\n # sound.play(-1)\n # pygame.time.delay(2000)\n # sound.stop()\n\nif __name__ == \"__main__\":\n # play_beat()\n play_tone()\n\n","repo_name":"david-jenkins/pydcam","sub_path":"pydcam/tests/sound_test.py","file_name":"sound_test.py","file_ext":"py","file_size_in_byte":3682,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"32637000615","text":"# binarization of an image: turinng an image to a binary\n\nimport cv2 as cv\n\nimg = cv.imread('images/5.jpg')\ncv.imshow('Image',img)\n\ngray = cv.cvtColor(img,cv.COLOR_BGR2GRAY)\ncv.imshow('Gray',gray)\n\n# simple thresholding\nthreshold, thres = cv.threshold(gray,150,255,cv.THRESH_BINARY)\nthreshold, thres = cv.threshold(gray,100,255,cv.THRESH_BINARY)\n\ncv.imshow('Simple Threshold',thres)\n\nthreshold, thres_inv = cv.threshold(gray,100,255,cv.THRESH_BINARY_INV)\n\ncv.imshow('Simple Threshold inverse',thres_inv)\n\n# adaptive thresholding\n\nadaptive_threshold = cv.adaptiveThreshold(gray,255,cv.ADAPTIVE_THRESH_MEAN_C,cv.THRESH_BINARY,11,3)\n# adaptive_threshold = cv.adaptiveThreshold(gray,255,cv.ADAPTIVE_THRESH_MEAN_C,cv.THRESH_BINARY_INV,11,3)\n\ncv.imshow('Adaptive Thresholding',adaptive_threshold)\n\n\n\n\ncv.waitKey(0)","repo_name":"AritraBhattacharjee/100daysofcode","sub_path":"DAY38 - DAY54 OpenCV/13_thresholding.py","file_name":"13_thresholding.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"70130024317","text":"#! /usr/bin/env python2.7\n\nname = \"oneshot_achilles\"\n\nimport sys\nimport rospy\nimport time\nimport std_msgs.msg\nsys.path.append(\"/home/amigos/ros/src/necst_ros3/lib\")\nimport achilles\n\n\ndef callback(req):\n data1 = []\n data2 = []\n dfs1_list = []\n dfs2_list = []\n array1 = std_msgs.msg.Float64MultiArray()\n array2 = std_msgs.msg.Float64MultiArray()\n \n dfs = achilles.dfs()\n data = dfs.oneshot(repeat=1, integsec=req.data, starttime=0.0)\n \n [data1.extend(i) for i in list(data[0])]\n [dfs1_list.append(data1[i*16384:(i+1)*16384]) for i in range(int(len(data1)/16384))]\n [data2.extend(i) for i in list(data[1])]\n [dfs2_list.append(data2[i*16384:(i+1)*16384]) for i in range(int(len(data2)/16384))]\n \n array1.data = dfs1_list[0]\n array2.data = dfs2_list[0]\n \n pub1.publish(array1)\n pub2.publish(array2)\n time.sleep(0.01)\n return\n\nif __name__ == \"__main__\":\n rospy.init_node(name)\n\n pub1 = rospy.Publisher(\n name = \"/achilles/data1\",\n data_class = std_msgs.msg.Float64MultiArray,\n latch = True,\n queue_size = 1,\n )\n\n pub2 = rospy.Publisher(\n name = \"/achilles/data2\",\n data_class = std_msgs.msg.Float64MultiArray,\n latch = True,\n queue_size = 1,\n )\n \n sub = rospy.Subscriber(\n name = \"/spectrometer/oneshot_cmd\",\n data_class = std_msgs.msg.Float32,\n callback = callback,\n queue_size = 1,\n )\n\n rospy.spin()\n","repo_name":"nanten2/necst_ros3","sub_path":"scripts/obs/oneshot_achilles.py","file_name":"oneshot_achilles.py","file_ext":"py","file_size_in_byte":1533,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"} +{"seq_id":"42934874753","text":"# coding: utf-8\n\nimport os.path, hashlib\nfrom tornado.options import options\n\ndef save_images(http_files):\n\n for http_file in http_files:\n name, ext = http_file['filename'].rsplit('.')\n file_path = os.path.join(options.path, \"i/%s.%s\" % (hashlib.md5(name).hexdigest(), ext))\n with open(file_path, 'wb') as f:\n f.write(http_file['body'])\n\n if not os.path.exists(file_path):\n return False\n\n return True\n","repo_name":"cloudorz/apple","sub_path":"utils/imagepp.py","file_name":"imagepp.py","file_ext":"py","file_size_in_byte":457,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"96"} +{"seq_id":"38218772089","text":"# https://www.acmicpc.net/problem/23288\nimport sys\nfrom collections import deque\n\ninput = sys.stdin.readline\n\n\ndef bfs(si, sj):\n q = deque()\n cur = board[si][sj]\n visited = set()\n visited.add((si, sj))\n q.append((si, sj))\n res = 1\n while q:\n i, j = q.popleft()\n for di, dj in d:\n ni = di + i\n nj = dj + j\n if -1 < ni < n and -1 < nj < m:\n if board[ni][nj] == cur and (ni, nj) not in visited:\n q.append((ni, nj))\n visited.add((ni, nj))\n res += 1\n return res * cur\n\n\ndef locate():\n if dd == 0:\n tmp = dice[2]\n dice[2] = 7 - dice[0]\n dice[0] = tmp\n elif dd == 1:\n tmp = dice[1]\n dice[1] = 7 - dice[0]\n dice[0] = tmp\n elif dd == 2:\n tmp = dice[0]\n dice[0] = 7 - dice[2]\n dice[2] = tmp\n elif dd == 3:\n tmp = dice[0]\n dice[0] = 7 - dice[1]\n dice[1] = tmp\n\n\nn, m,k = map(int, input().split())\nboard = [list(map(int, input().split())) for _ in range(n)]\ndice = [6, 5, 3] # 아래, 앞, 오른쪽\n\nd = [(0, 1), (1, 0), (0, -1), (-1, 0)]\ndd = 0\ndx = 0\ndy = 0\nanswer = 0\nfor _ in range(k):\n nx = d[dd][0] + dx\n ny = d[dd][1] + dy\n if not (-1 < nx < n and -1 < ny < m):\n dd = (dd + 2) % 4\n nx = d[dd][0] + dx\n ny = d[dd][1] + dy\n dx = nx\n dy = ny\n answer += bfs(dx, dy)\n locate() # 주사위 방향조정\n if dice[0] > board[dx][dy]:\n dd = (dd + 1) % 4\n elif dice[0] < board[dx][dy]:\n dd = (dd - 1) % 4\n\nprint(answer)\n","repo_name":"brunchmate/AlgorithmStudy","sub_path":"maplejh/BOJ/23288.py","file_name":"23288.py","file_ext":"py","file_size_in_byte":1609,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"96"}