diff --git "a/3195.jsonl" "b/3195.jsonl"
new file mode 100644--- /dev/null
+++ "b/3195.jsonl"
@@ -0,0 +1,35 @@
+{"seq_id":"13713775044","text":"import pytest\nfrom selenium import webdriver\nfrom selenium.webdriver.chrome.options import Options as chrome_options\n\n\n# MVP. В этом файле задаются настройки тестирования\n\n\n@pytest.fixture()\ndef get_chrome_options():\n # получение опций браузера\n options = chrome_options()\n options.add_argument('chrome') # use headless if u don't need a browser UI\n options.add_argument('--start-maximized') # окно открывается. как F11\n options.add_argument('--window-size=1280,720') # разрешение окна\n return options\n\n\n@pytest.fixture()\ndef get_webdriver(get_chrome_options):\n # получение драйвера браузера при заданных опциях\n options = get_chrome_options\n driver = webdriver.Chrome(options=options)\n return driver\n\n\n@pytest.fixture(scope='function') # function отвечает за исполнение фикстуры каждый тест\n# можно поставить аргумент session - раз в сессию\ndef setup(request, get_webdriver):\n # получение драйвера и отправление запроса на сайт\n driver = get_webdriver # драйвер получается из метода\n url = 'https://www.wikipedia.org/' # задается url сайта, на котором будет проводиться тест.\n if request.cls is not None:\n request.cls.driver = driver\n driver.get(url)\n # driver.delete_all_cookies() # используется, потому что сайт ограничивает доступ к ресурсам из-за автоматизированного тестирования\n yield driver # возврат driver, но не знаю почему используется yield\n driver.quit() # закрытие всех окон, можно использовать driver.close(), но тогда закроется только одна вкладка теста\n","repo_name":"reennn/MySeleniumTry","sub_path":"tests/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":2046,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"23581362283","text":"from sis_provisioner.events import SISProvisionerProcessor, ProcessorException\nfrom sis_provisioner.models.events import GroupLog\nfrom sis_provisioner.events.group.dispatch import (\n AffiliateLoginGroupDispatch, SponsoredLoginGroupDispatch,\n StudentLoginGroupDispatch, UWGroupDispatch)\nfrom aws_message.crypto import aes128cbc, CryptoException\nfrom base64 import b64decode\nimport json\n\nQUEUE_SETTINGS_NAME = 'GROUP'\n\n\nclass GroupProcessor(SISProvisionerProcessor):\n \"\"\"\n UW GWS Group Event Processor\n \"\"\"\n _logModel = GroupLog\n\n # What we expect in a UW Group event message\n _eventMessageType = 'gws'\n _eventMessageVersion = 'UWIT-1'\n\n def __init__(self):\n super(GroupProcessor, self).__init__(\n queue_settings_name=QUEUE_SETTINGS_NAME, is_encrypted=True)\n\n def validate_message_body(self, message):\n header = message['header']\n if header['messageType'] != self._eventMessageType:\n raise ProcessorException(\n 'Unknown Message Type: {}'.format(header['messageType']))\n\n if header['version'] != self._eventMessageVersion:\n raise ProcessorException(\n 'Unknown Message Version: {}'.format(header['version']))\n\n context = json.loads(b64decode(header['messageContext']))\n self._action = context['action']\n self._groupname = context['group']\n self._dispatch = None\n\n for dispatch_class in [\n StudentLoginGroupDispatch, SponsoredLoginGroupDispatch,\n AffiliateLoginGroupDispatch, UWGroupDispatch]:\n dispatch = dispatch_class(self.settings)\n if dispatch.mine(self._groupname):\n self._dispatch = dispatch\n break\n\n return (self._dispatch is not None)\n\n def _parse_signature(self, message):\n header = message['header']\n\n to_sign = '{}\\n'.format(header[u'contentType'])\n if 'keyId' in header:\n to_sign += '{}\\n{}\\n'.format(header[u'iv'], header[u'keyId'])\n to_sign += (\n '{context}\\n{msgid}\\n{msgtype}\\n{sender}\\n{cert}\\n'\n '{timestamp}\\n{version}\\n{body}\\n').format(\n context=header[u'messageContext'], msgid=header[u'messageId'],\n msgtype=header[u'messageType'], sender=header[u'sender'],\n cert=header[u'signingCertUrl'], timestamp=header[u'timestamp'],\n version=header[u'version'], body=message['body'])\n\n sig_conf = {\n 'cert': {\n 'type': 'url',\n 'reference': header[u'signingCertUrl']\n }\n }\n\n return (sig_conf, to_sign, header['signature'])\n\n def decrypt_message_body(self, message):\n header = message['header']\n body = message['body']\n try:\n if set(['keyId', 'iv']).issubset(header):\n key = header['keyId']\n keys = self.settings.get('BODY_DECRYPT_KEYS', {})\n\n cipher = aes128cbc(\n b64decode(keys[key]), b64decode(header['iv']))\n body = cipher.decrypt(b64decode(body))\n return body\n\n except KeyError as ex:\n raise ProcessorException('Invalid keyId: {}'.format(key))\n except CryptoException as ex:\n raise ProcessorException('Cannot decrypt: {}'.format(ex))\n except Exception as ex:\n raise ProcessorException('Cannot read: {}'.format(ex))\n\n def process_message_body(self, json_data):\n if json_data is not None:\n n = self._dispatch.run(self._action, self._groupname, json_data)\n if n:\n self.record_success_to_log(n)\n","repo_name":"uw-it-aca/canvas-sis-provisioner","sub_path":"sis_provisioner/events/group/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":3662,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"73007951895","text":"# Application of moving a ball, and you should try to attach the bomb\n\n# import module\nimport simpleguitk as simplegui\nimport random\nimport math\n\n# define globals\nwidth = 500\t\t# frame width\nheight = 500\t# frame height\nspeed = 10\t\t# the ball's default speed\ncurrent_key = ' '\t\t# the default current key\nrebound = False\t\t# disable rebound for default\ncurrent_score = 0\t# current score you get\nrequired_score = 20\t\t# required score that you have to get\nscore_per_green = 1\t\t# score once you attach a green bomb\nscore_per_yellow = 2\t# score once you attach a yellow bomb\nscore_per_blue = 3\t\t# score once you attach a blue bomb\nscore_string = str(current_score) + \"/\" + str(required_score)\t# string format of score\n\n# define class\nclass Ball:\n\tpos = [width / 2, height / 2]\t\t# the ball's current position in the frame\n\tradius = 20\t\t# radius of ball\n\tline_width = 2\t\t# line width of ball\n\tline_color = \"Red\"\t\t# line color of ball\n\tfill_color = \"White\"\t# filling color of ball\n\t# ball move left\n\tdef move_left(self):\n\t\tif self.pos[0] > 0:\n\t\t\tself.pos[0] -= speed\n\n\t# ball move right\n\tdef move_right(self):\n\t\tif self.pos[0] < width:\n\t\t\tself.pos[0] += speed\n\n\t# ball move up\n\tdef move_up(self):\n\t\tif self.pos[1] > 0:\n\t\t\tself.pos[1] -= speed\n\n\t# ball move down\n\tdef move_down(self):\n\t\tif self.pos[1] < height:\n\t\t\tself.pos[1] += speed\n\nclass Bomb_Green:\n\tpos = [0, 0]\n\tradius = 4\n\tline_width = 4\n\tline_color = \"Green\"\n\tfill_color = \"Green\"\n\nclass Bomb_Yellow:\n\tpos = [0, 0]\n\tradius = 3\n\tline_width = 3\n\tline_color = \"Yellow\"\n\tfill_color = \"Yellow\"\n\nclass Bomb_Blue:\n\tpos = [0, 0]\n\tradius = 2\n\tline_width = 2\n\tline_color = \"Blue\"\n\tfill_color = \"Blue\"\n\n# define instance of class Ball and class Bomb\nball = Ball()\n\n# define bomb list\nbomb_green_list = Bomb_Green()\nbomb_yellow_list = Bomb_Yellow()\nbomb_blue_list = Bomb_Blue()\n\n# define event handlers\n# generate a bomb\ndef generate_bomb():\n\tglobal bomb_green_list, bomb_yellow_list, bomb_blue_list\n\tbomb_green_list.pos[0] = random.randrange(0, width)\n\tbomb_green_list.pos[1] = random.randrange(0, height)\n\tbomb_yellow_list.pos[0] = random.randrange(0, width)\n\tbomb_yellow_list.pos[1] = random.randrange(0, height)\n\tbomb_blue_list.pos[0] = random.randrange(0, width)\n\tbomb_blue_list.pos[1] = random.randrange(0, height)\n\n# condition of catch bomb\ndef catch(ball, bomb):\n\tdistance = (ball.pos[0] - bomb.pos[0]) ** 2 + (ball.pos[1] - bomb.pos[1]) ** 2\n\tlimit = (ball.radius + bomb.radius) ** 2\n\tif distance <= limit:\n\t\treturn True\n\telse:\n\t\treturn False\n\n# calculator the score\ndef get_score():\n\tglobal current_score\n\tif catch(ball, bomb_green_list) == True:\n\t\tcurrent_score += score_per_green\n\tif catch(ball, bomb_yellow_list) == True:\n\t\tcurrent_score += score_per_yellow\n\tif catch(ball, bomb_blue_list) == True:\n\t\tcurrent_score += score_per_blue\n\n# speed up the ball's movement\ndef speed_up():\n\tglobal speed\n\tspeed += 1\n\tspeed_label.set_text(\"speed = \" + str(speed))\n\n# speed down the ball's movement\ndef speed_down():\n\tglobal speed\n\tif speed > 0:\n\t\tspeed -= 1\n\tspeed_label.set_text(\"speed = \" + str(speed))\n\n# press the key \"up\" or \"down\" or \"left\" or \"right\" in the label\ndef keydown(key):\n\tglobal current_key\n\tcurrent_key = key\n\n# handler of timer\ndef time_handler_ball():\n\tglobal current_key\n\tif current_key == simplegui.KEY_MAP[\"left\"]:\n\t\tif rebound == True and ball.pos[0] == 0:\n\t\t\tcurrent_key = simplegui.KEY_MAP[\"right\"]\n\t\t\tball.move_right()\n\t\telse:\n\t\t\tball.move_left()\n\telif current_key == simplegui.KEY_MAP[\"right\"]:\n\t\tif rebound == True and ball.pos[0] == width:\n\t\t\tcurrent_key = simplegui.KEY_MAP[\"left\"]\n\t\t\tball.move_left()\n\t\telse:\n\t\t\tball.move_right()\n\telif current_key == simplegui.KEY_MAP[\"up\"]:\n\t\tif rebound == True and ball.pos[1] == 0:\n\t\t\tcurrent_key = simplegui.KEY_MAP[\"down\"]\n\t\t\tball.move_down()\n\t\telse:\n\t\t\tball.move_up()\n\telif current_key == simplegui.KEY_MAP[\"down\"]:\n\t\tif rebound == True and ball.pos[1] == height:\n\t\t\tcurrent_key = simplegui.KEY_MAP[\"up\"]\n\t\t\tball.move_up()\n\t\telse:\n\t\t\tball.move_down()\n\tget_score()\n\tif current_score >= required_score:\n\t\tstop()\n\ndef time_handler_bomb():\n\tgenerate_bomb()\n\n# reset: make the speed and ball's position default value\ndef reset():\n\tglobal speed, ball, rebound, score_string, current_score\n\t# global bomb_green_list, bomb_yellow_list, bomb_blue_list\n\t# stop()\n\tspeed = 10\n\tball.pos = [width / 2, height / 2]\n\tbomb_green_list.pos = [0, 0]\n\tbomb_yellow_list.pos = [0, 0]\n\tbomb_blue_list.pos = [0, 0]\n\trebound = False\n\trebound_label.set_text(\"rebound = \" + str(rebound))\n\tcurrent_score = 0\n\tscore_string = str(current_score) + \"/\" + str(required_score)\n\t# frame.set_draw_handler(draw)\t# draws 50 times per second\n\t# stop()\n\n# start movement\ndef start():\n\ttimer_ball.start()\n\ttimer_bomb.start()\n\n# stop movement\ndef stop():\n\ttimer_ball.stop()\n\ttimer_bomb.stop()\n\n# change the set of current rebound\ndef set_rebound():\n\tglobal rebound\n\tif rebound == True:\n\t\trebound = False\n\telse:\n\t\trebound = True\n\trebound_label.set_text(\"rebound = \" + str(rebound))\n\n# draw the frame\ndef draw(canvas):\n\t# global bomb_green_list, bomb_yellow_list, bomb_blue_list\n\tcanvas.draw_circle(ball.pos, ball.radius, ball.line_width, ball.line_color, ball.fill_color)\n\tcanvas.draw_circle(bomb_green_list.pos, bomb_green_list.radius, bomb_green_list.line_width, bomb_green_list.line_color, bomb_green_list.fill_color)\n\tcanvas.draw_circle(bomb_yellow_list.pos, bomb_yellow_list.radius, bomb_yellow_list.line_width, bomb_yellow_list.line_color, bomb_yellow_list.fill_color)\n\tcanvas.draw_circle(bomb_blue_list.pos, bomb_blue_list.radius, bomb_blue_list.line_width, bomb_blue_list.line_color, bomb_blue_list.fill_color)\n\tglobal score_string\n\tscore_string = str(current_score) + \"/\" + str(required_score)\n\tcanvas.draw_text(score_string, [400, 50], 20, \"White\")\n\n# create frame\nframe = simplegui.create_frame(\"Move Ball\", width, height)\ntimer_ball = simplegui.create_timer(100, time_handler_ball)\ntimer_bomb = simplegui.create_timer(1000, time_handler_bomb)\n\n# register event handlers into frame\nframe.set_draw_handler(draw)\t# draws 50 times per second\nspeed_label = frame.add_label(\"speed = \" + str(speed))\nframe.add_button(\"speed up\", speed_up, 50)\nframe.add_button(\"speed down\", speed_down, 50)\nframe.add_button(\"reset\", reset, 50)\nframe.add_button(\"start\", start, 50)\nframe.add_button(\"stop\", stop, 50)\nrebound_label = frame.add_label(\"rebound = \" + str(rebound))\nframe.add_button(\"rebound\", set_rebound, 50)\nframe.set_keydown_handler(keydown)\n\n# start frame\nframe.start()\n","repo_name":"lookfiresu123/Interacive_python","sub_path":"auto_move_ball.py","file_name":"auto_move_ball.py","file_ext":"py","file_size_in_byte":6444,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"12475083968","text":"class AoCVM:\n\n def reset(self):\n self.pos = 0\n self.accumulator = 0\n self.counter = 0\n self.execlines = set()\n\n def __init__(self):\n self.reset()\n\n def fromtxt(self, filepath):\n self.code = []\n with open('day08input.txt') as inpfile:\n while True:\n line = inpfile.readline()\n if line.strip() == '':\n break\n operation = line.strip().split()[0]\n argument = int(line.strip().split()[1])\n self.code.append([operation, argument])\n inpfile.close()\n self.maxpos = len(self.code)-1\n\n def step(self):\n op, arg = self.code[self.pos][0], self.code[self.pos][1]\n self.execlines.add(self.pos)\n if op == 'nop':\n self.pos += 1\n elif op == 'acc':\n self.accumulator += arg\n self.pos += 1\n elif op == 'jmp':\n self.pos += arg\n\n def loopcheck(self, report=False):\n if self.pos in self.execlines:\n if report:\n print('Loop start at line {}'.format(self.pos))\n return True\n else:\n return False\n\n def outcheck(self):\n return (self.pos < 0) or (self.pos > self.maxpos)\n\n def endofcode(self):\n return self.pos == self.maxpos\n\n\nvm = AoCVM()\nvm.fromtxt('day08input.txt')\n\n# part 1\nwhile True:\n if vm.loopcheck(report=True):\n print('Accumulator value before loop is {}'.format(vm.accumulator))\n break\n vm.step()\n if vm.outcheck():\n print('Out of code!')\n break\n\nvm.reset()\n\n# part 2\nmaybugs = []\nfor i in range(len(vm.code)):\n if vm.code[i][0] in ['nop', 'jmp']:\n maybugs.append(i)\nfor i in maybugs:\n # flip\n if vm.code[i][0] == 'nop':\n vm.code[i][0] = 'jmp'\n else:\n vm.code[i][0] = 'nop'\n # trying\n vm.reset()\n while True:\n if vm.loopcheck():\n break\n vm.step()\n if vm.outcheck():\n print('Out of code!')\n break\n if vm.endofcode():\n break\n if vm.endofcode():\n vm.step()\n print('Accumulator value at end of code is {}'.format(vm.accumulator))\n break\n # reverse flip\n if vm.code[i][0] == 'nop':\n vm.code[i][0] = 'jmp'\n else:\n vm.code[i][0] = 'nop'\n","repo_name":"aldragon-net/AoC","sub_path":"2020/day08/day08.py","file_name":"day08.py","file_ext":"py","file_size_in_byte":2361,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"32168821315","text":"import torch\r\nimport torch.nn as nn\r\n\r\nlayer = nn.Conv2d(1,3,kernel_size=3,stride=1,padding=0)\r\nx = torch.rand(1,1,28,28)\r\nout = layer(x)\r\nprint(out.shape) #torch.Size([1, 3, 26, 26])\r\n\r\nlayer = nn.Conv2d(1,3,kernel_size=3,stride=1,padding=1)\r\nx = torch.rand(1,1,28,28)\r\nout = layer(x)\r\nprint(out.shape) #torch.Size([1, 3, 28, 28])\r\n\r\nlayer = nn.Conv2d(1,3,kernel_size=3,stride=2,padding=1)\r\nx = torch.rand(1,1,28,28)\r\nout = layer(x)\r\nprint(out.shape) #torch.Size([1, 3, 14, 14])\r\n\r\nprint(layer.weight)\r\nprint(layer.weight.shape)\r\nx = out\r\nlayer = nn.MaxPool2d(2,stride=2)\r\nlayer = nn.AvgPool2d(2,stride=2)\r\n\r\nx = torch.rand(100,16,784)\r\nlayer = nn.BatchNorm1d(16)\r\nout = layer(x)\r\n\r\nprint(layer.running_mean)\r\nprint(layer.running_var)","repo_name":"TodayUp/Pytorch_Course","sub_path":"12CNN.py","file_name":"12CNN.py","file_ext":"py","file_size_in_byte":742,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"68"}
+{"seq_id":"22945790475","text":"from django.conf import settings\nfrom django.core.mail import send_mail\nfrom django.shortcuts import render, redirect\nfrom datetime import datetime, date, timedelta\nfrom .models import advancepayment, paydowncreditcard, salesrecpts, timeact, timeactsale, Cheqs, suplrcredit, addac, \\\n bills, invoice, expences, payment, credit, delayedcharge, estimate, service, noninventory, bundle, employee, \\\n payslip, inventory, customer, supplier, company, accounts, ProductModel, ItemModel, accountype, \\\n expenseaccount, incomeaccount, accounts1, recon1, recordpay, addtax1, bankstatement, customize\nfrom django.contrib.auth.models import auth, User\nfrom django.contrib import messages\nfrom django.db.models import Sum, Q\nfrom django.db.models.functions import Coalesce\nimport json\nfrom django.http.response import JsonResponse\nfrom django.contrib.auth.decorators import login_required\nimport itertools\n\n\ndef index(request):\n return render(request, 'app1/index.html')\n\n\ndef go(request):\n return render(request, 'app1/login.html')\n\n\ndef create(request):\n try:\n if request.method == 'POST':\n firstname = request.POST['first_name']\n lastname = request.POST['last_name']\n email = request.POST['email']\n username = request.POST['username']\n password = request.POST['password']\n conformpassword = request.POST['conformpassword']\n if password == conformpassword:\n if User.objects.filter(username=username).exists():\n messages.info(\n request, 'This username already exists. Sign up again')\n return render(request, 'app1/login.html')\n else:\n user = User.objects.create_user(first_name=firstname, last_name=lastname, email=email,\n username=username, password=password)\n user.save()\n return render(request, 'app1/company.html', {'member': user})\n else:\n return redirect('')\n else:\n return redirect('')\n except:\n return redirect('')\n\n\ndef regcomp(request):\n return render(request, 'app1/login.html')\n\ndef register(request, id):\n try:\n if request.method == 'POST':\n cname = request.POST.get('cname')\n caddress = request.POST.get('caddress')\n city = request.POST.get('city')\n state = request.POST.get('state')\n pincode = request.POST.get('pincode')\n cemail = request.POST.get('cemail')\n phone = request.POST.get('phone')\n try:\n img1 = request.FILES['img1']\n except:\n img1 = 'default'\n bname = request.POST.get('bname')\n industry = request.POST.get('industry')\n ctype = request.POST.get('ctype')\n abt = request.POST.get('abt')\n paid = request.POST.get('paid')\n new_id = User.objects.get(id=id)\n companys = company(id=new_id, cname=cname, caddress=caddress, city=city, state=state, pincode=pincode,\n cemail=cemail,\n phone=phone, bname=bname, industry=industry, ctype=ctype, abt=abt, paid=paid)\n if img1 != 'default':\n companys.cimg = img1\n companys.save()\n toda = date.today()\n tod = toda.strftime(\"%Y-%m-%d\")\n comp = company.objects.get(id=new_id)\n accountsecond = [\n ['Account Receivable(Debtors)', 'Account Receivable(Debtors)',\n 'Account Receivable(Debtors)'],\n ['Current Assets', 'Deferred Service Tax Input Credit', 'Deferred CGST'],\n ['Current Assets', 'Deferred Service Tax Input Credit',\n 'Deferred GST Input Credit'],\n ['Current Assets', 'Deferred Service Tax Input Credit', 'Deferred IGST'],\n ['Current Assets', 'Deferred Service Tax Input Credit',\n 'Deferred Krishi Kalyan Cess Input Credit'],\n ['Current Assets', 'Prepaid Expenses', 'Prepaid Expenses'],\n ['Current Assets', 'Deferred Service Tax Input Credit',\n 'Deferred Service Tax Input Credit'],\n ['Current Assets', 'Deferred Service Tax Input Credit', 'Deferred SGST'],\n ['Current Assets', 'Deferred Service Tax Input Credit',\n 'Deferred VAT Input Credit'],\n ['Current Assets', 'Service Tax Refund', 'GST Refund'],\n ['Current Assets', 'Inventory', 'Inventory Asset'],\n ['Current Assets', 'Service Tax Refund',\n 'Krishi Kalyan Cess Refund'],\n ['Current Assets', 'Prepaid Expenses', 'Prepaid Insurance'],\n ['Current Assets', 'Service Tax Refund', 'Service Tax Refund'],\n ['Current Assets', 'Other Current Assets', 'TDS Receivable'],\n ['Current Assets', 'Other Current Assets', 'Uncategorised Asset'],\n ['Current Assets', 'Undeposited Fund', 'Undeposited Fund'],\n ['Fixed Assets', 'Accumulated Depreciation',\n 'Accumulated Depreciation'],\n ['Fixed Assets', 'Buildings', 'Buildings and Improvements'],\n ['Fixed Assets', 'Furniture and fixtures', 'Furniture and Equipment'], [\n 'Fixed Assets', 'Land', 'Land'],\n ['Fixed Assets', 'Leasehold Improvements', 'Leasehold Improvements'],\n ['Fixed Assets', 'Vehicles', 'Vehicles'],\n ['Accounts Payable(Creditors)', 'Accounts Payable(Creditors)',\n 'Accounts Payable(Creditors)'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'CGST Payable'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'CST Payable'],\n ['Current Liabilities', 'Tax Suspense', 'CST Suspense'],\n ['Current Liabilities', 'Sales And Service Tax Payable', 'GST Payable'],\n ['Current Liabilities', 'Tax Suspense', 'GST Suspense'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'IGST Payable'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input CGST'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input CGST Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input IGST'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input IGST Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input Krishi Kalyan Cess'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input Krishi Kalyan Cess RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input Service Tax'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Input Service Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input SGST'],\n ['Current Liabilities', 'Sales asnd Service Tax Payable',\n 'Input SGST Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input VAT 14%'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input VAT 4%'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Input VAT 5%'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Krishi Kalyan Cess Payable'],\n ['Current Liabilities', 'Tax Suspense',\n 'Krishi Kalyan Cess Suspense'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output CGST'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Output CGST Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output CST 2%'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output IGST'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Output IGST Tax RCM'],\n ['Current Liabilties', 'Sales and Service Tax Payable',\n 'Output Krishi Kaylan Cess'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Output Krishi Kalyan Cess RCM'],\n ['Current Liabilties', 'Sales and Service Tax Payable',\n 'Output Service Tax'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Output Service Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output SGST'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Output SGST Tax RCM'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output VAT 14%'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output VAT 4%'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'Output VAT 5%'],\n ['Current Liabilties', 'Sales and Service Tax Payable',\n 'Service Tax Payable'],\n ['Current Liabilities', 'Tax Suspense', 'Service Tax Suspense'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'SGST Payable'],\n ['Current Liabilities', 'Sales and Service Tax Payable',\n 'Swachh Barath Cess Payable'],\n ['Current Liabilities', 'Tax Suspense',\n 'Swachh Barath Cess Suspense'],\n ['Current Liabilities', 'Current Liabilities', 'TDS Payable'],\n ['Current Liabilities', 'Sales and Service Tax Payable', 'VAT Payable'],\n ['Current Liabilities', 'Tax Suspense', 'VAT Suspense'],\n ['Equity', 'Opening Balance Equity', 'Opening Balance Equity'],\n ['Equity', 'Retained Earnings', 'Retained Earnings'],\n ['Income', 'Service/Fee Income', 'Billable Expense Income'],\n ['Income', 'Service/Fee Income', 'Consulting Income'],\n ['Income', 'Sales of Product Income', 'Product Sales'], [\n 'Income', 'Sales of Product Income', 'Sales'],\n ['Income', 'Sales of Product Income', 'Sales-Hardware'],\n ['Income', 'Sales of Product Income', 'Sales-Software'],\n ['Income', 'Sales of Product Income',\n 'Sales-Support and Maintanance'],\n ['Income', 'Discount/Refund Given', 'Sales Discounts'],\n ['Income', 'Sales of Product Income', 'Sales of Product Income'],\n ['Income', 'Service/Fee Income', 'Services'],\n ['Income', 'Unapplied Cash Payment Income',\n 'Unapplied Cash Payment Income'],\n ['Income', 'Service/Fee Income', 'Uncategorised Income'],\n ['Cost of Goods Sold', 'Suppliers and Materials-COS', 'Cost of Sales'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Equipment Rental for Jobs'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Freight and Shipping Cost'],\n ['Cost of Goods Sold', 'Suppliers and Materials-COS',\n 'Inventory Shrinkage'],\n ['Cost of Goods Sold', 'Cost of Goods Sold', 'Merchant Account Fees'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Purchases-Hardware for Resale'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Purchases-Software for Resale'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Subcontracted Services'],\n ['Cost of Goods Sold', 'Cost of Goods Sold',\n 'Tools and Craft Suppliers'],\n ['Expenses', 'Advertising/Promotional', 'Advertising/Promotional'],\n ['Expenses', 'Bank Charges', 'Bank Charges'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Business Licenses and Permitts'],\n ['Expenses', 'Charitable Contributions', 'Charitable Contributions'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Computer and Internet Expense'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Continuing Education'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Depreciation Expense'],\n ['Expenses', 'Dues and Subscriptions', 'Dues and Subscriptions'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Housekeeping Charges'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Insurance Expenses'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Insurance Expenses-General Liability Insurance'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Insurance Expenses-Health Insurance'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Insurance Expenses-Life and Disability Insurance'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Insurance Expenses-Professional Liability'],\n ['Expenses', 'Interest Paid', 'Interest Expenses'],\n ['Expenses', 'Meals and Entertainment', 'Meals and Entertainment'],\n ['Expenses', 'Office/General Administrative Expenses', 'Office Supplies'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Postage and Delivery'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Printing and Reproduction'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Professional Fees'],\n ['Expenses', 'Suppliers and Materials', 'Purchases'],\n ['Expenses', 'Office/General Administrative Expenses', 'Rent Expense'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Repair and Maintanance'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Small Tools and Equipments'],\n ['Expenses', 'Tax Expense', 'Swachh Barath Cess Expense'],\n ['Expense', 'Office/General Administrative Expenses', 'Taxes-Property'],\n ['Expenses', 'Office/General Administrative Expenses',\n 'Telephone Expense'],\n ['Expenses', 'Office/General Administrative Expenses', 'Travel Expense'],\n ['Expenses', 'Other Miscellaneous Service Cost',\n 'Uncategorised Expense'],\n ['Expenses', 'Utilities', 'Utilities'],\n ['Other Income', 'Other Miscellaneous Income',\n 'Finance Charge Income'],\n ['Other Income', 'Other Miscellaneous Income',\n 'Insurance Proceeds Received'],\n ['Other Income', 'Interest Earned', 'Interest Income'],\n ['Other Income', 'Other Miscellaneous Income',\n 'Proceeds From Sale of Assets'],\n ['Other Income', 'Other Miscellaneous Income',\n 'Shipping and Delivery Income'],\n ['Other Expenses', 'Other Expenses', 'Ask My Accountant'],\n ['Other Expenses', 'Other Expenses', 'CGST Write-Off'],\n ['Other Expense', 'Other Expense', 'GST Write-Off'],\n ['Other Expenses', 'Other Expenses', 'IGST Write-Off'],\n ['Other Expenses', 'Other Expenses', 'Miscellaneous Expense'],\n ['Other Expenses', 'Other Expenses', 'Political Contributions'],\n ['Other Expenses', 'Other Expenses',\n 'Reconciliation Discrepancies'],\n ['Other Expenses', 'Other Expenses', 'SGST Write-Off'],\n ['Other Expenses', 'Other Expenses', 'Tax Write-Off'],\n ['Other Expenses', 'Other Expenses', 'Vehicle Expenses']]\n\n accounype = [['Deferred CGST'], ['Deferred GST Input Credit'], ['Deferred IGST'],\n ['Deferred Krishi Kalyan Cess Input Credit'],\n ['Deferred Service Tax Input Credit'], [\n 'Deferred SGST'], ['Deferred VAT Input Credit'],\n ['GST Refund'],\n ['Inventory Asset'], ['Paid Insurance'], [\n 'Service Tax Refund'], ['TDS Receivable'],\n ['Uncategorised Asset'],\n ['Accumulated Depreciation'], ['Buildings and Improvements'], [\n 'Furniture and Equipments'],\n ['Land'],\n ['Leasehold Improvements'], ['Vehicles'], [\n 'CGST Payable'], ['CST Payable'], ['CST Suspense'],\n ['GST Payable'],\n ['GST Suspense'], ['IGST Payable'], ['Input CGST'], [\n 'Input CGST Tax RCM'], ['Input IGST'],\n ['Input IGST Tax RCM'],\n ['Input Krishi Kalyan Cess'], [\n 'Input Krishi Kalyan Cess RCM'], ['Input Service Tax'],\n ['Input Service Tax RCM'],\n ['Input SGST'], ['Input SGST Tax RCM'], [\n 'Input VAT 14 %'], ['Input VAT 4%'], ['Input VAT 5%'],\n ['Krishi Kalyan Cess Payable'], [\n 'Krishi Kalyan Cess Suspense'], ['Output CGST'],\n ['Output CGST Tax RCM'],\n ['Output CST 2%'], ['Output IGST'], [\n 'Output IGST Tax RCM'], ['Output Krishi Kalyan Cess'],\n ['Output Krishi Kalyan Cess RCM'], [\n 'Output Service Tax'], ['Output Service Tax RCM'],\n ['Output SGST'],\n ['Output SGST Tax RCM'], ['Output VAT 14%'], [\n 'Output VAT 4%'], ['Output VAT 5%'],\n ['Service Tax Payable'],\n ['Service Tax Suspense'], ['SGST Payable'], [\n 'SGST Suspense'], ['Swachh Barath Cess Payable'],\n ['Swachh Barath Cess Suspense'], ['TDS Payable'], ['VAT Payable'], ['VAT Suspense']]\n for i in range(len(accountsecond)):\n for j in range(1):\n accounts1model = accounts1(cid=comp, acctype=accountsecond[i][0],\n detype=accountsecond[i][1], name=accountsecond[i][2], description='',\n gst='', deftaxcode='', balance=0.0,\n asof=tod)\n accounts1model.save()\n for i in range(len(accounype)):\n for j in range(1):\n accoutype = accountype(\n cid=comp, accountname=accounype[i][0])\n accoutype.save()\n return redirect('regcomp')\n else:\n return redirect('regcomp')\n except:\n return redirect('regcomp')\n\n\ndef login(request):\n try:\n if request.method == 'POST':\n try:\n username = request.POST['username']\n password = request.POST['password']\n user = auth.authenticate(username=username, password=password)\n cmp1 = company.objects.get(id=user.id)\n request.session[\"uid\"] = user.id\n if user is not None:\n auth.login(request, user)\n return redirect('/app1/godash')\n else:\n messages.info(request, 'Invalid username or password')\n return redirect('regcomp')\n except:\n messages.info(request, 'Invalid username or password')\n return render(request, 'app1/login.html')\n else:\n messages.info(request, 'Invalid username or password')\n return render(request, 'app1/login.html')\n except:\n messages.info(request, 'Invalid username or password')\n return render(request, 'app1/login.html')\n\n\n@login_required(login_url='regcomp')\ndef godash(request):\n try:\n cmp1 = company.objects.get(id=request.session[\"uid\"])\n request.session[\"invcol\"] = \" \"\n request.session[\"noninvcol\"] = \" \"\n request.session[\"buncol\"] = \" \"\n context = {'cmp1': cmp1}\n return render(request, 'app1/dashbord.html', context)\n except:\n cmp1 = company.objects.get(id=request.session[\"uid\"])\n request.session[\"invcol\"] = \" \"\n request.session[\"noninvcol\"] = \" \"\n request.session[\"buncol\"] = \" \"\n context = {'cmp1': cmp1}\n return render(request, 'app1/dashbord.html', context)\n return redirect('/')\n\n\ndef logout(request):\n request.session[\"uid\"] = \"\"\n auth.logout(request)\n return redirect('/')\n\n\n@login_required(login_url='regcomp')\ndef userprofile(request, id):\n try:\n user1 = User.objects.get(id=id)\n cmp1 = company.objects.get(id=request.session[\"uid\"])\n return render(request, 'app1/userprofile.html', {'users': user1, 'cmp1': cmp1})\n except:\n return redirect('godash')\n\n\n@login_required(login_url='regcomp')\ndef edituserprofile(request):\n try:\n user1 = User.objects.get(id=request.session[\"uid\"])\n cmp1 = company.objects.get(id=request.session[\"uid\"])\n context = {'users': user1, 'cmp1': cmp1}\n return render(request, 'app1/edituserprofile.html', context)\n except:\n return redirect('godash')\n\n\n@login_required(login_url='regcomp')\ndef updateuserprofile(request):\n try:\n user = User.objects.get(id=request.session[\"uid\"])\n comp = company.objects.get(id=user.id)\n\n user.first_name = request.POST[\"first_name\"]\n user.last_name = request.POST[\"last_name\"]\n user.email = request.POST[\"email\"]\n p1 = request.POST[\"newpassword\"]\n p2 = request.POST[\"newpassword1\"]\n\n comp.cname = request.POST[\"cname\"]\n comp.caddress = request.POST[\"caddress\"]\n comp.city = request.POST[\"city\"]\n comp.state = request.POST[\"state\"]\n comp.pincode = request.POST[\"pincode\"]\n comp.cemail = request.POST[\"cemail\"]\n comp.phone = request.POST[\"phone\"]\n comp.bname = request.POST[\"bname\"]\n comp.industry = request.POST[\"industry\"]\n comp.ctype = request.POST[\"ctype\"]\n try:\n img1 = request.FILES[\"img1\"]\n comp.cimg = img1\n except:\n img2 = request.POST[\"img2\"]\n comp.cimg = img2\n\n comp.save()\n user.save()\n\n if p1 == p2:\n if p1 != \"\":\n user.set_password(p1)\n user.save()\n logout(request)\n return redirect('/app1/go')\n elif p1 == \"\":\n return redirect('/app1/godash')\n\n return redirect('/app1/godash')\n except:\n return redirect('godash')\n\n\n@login_required(login_url='regcomp')\ndef editsettings(request):\n return render(request, 'app1/editsettings.html')\n\n\n\n@login_required(login_url='regcomp')\ndef cashposition(request):\n try:\n label_1 = []\n data_1 = []\n cmp1 = company.objects.get(id=request.session['uid'])\n cashpo = accounts.objects.filter(\n detype='Cash and Cash Equivalents', cid=cmp1)\n balance = accounts.objects.order_by('balance')[:10]\n for bala in balance:\n if bala.cid == cmp1:\n label_1.append(bala.name)\n data_1.append(bala.balance)\n context = {'cmp1': cmp1,\n 'cashpo': cashpo,\n 'label_1': label_1,\n 'data_1': data_1}\n return render(request, 'app1/cashposition.html', context)\n except:\n return redirect('godash')\n\n\n@login_required(login_url='regcomp')\ndef editaccounts(request):\n try:\n user1 = User.objects.get(id=request.session[\"uid\"])\n cmp1 = company.objects.get(id=request.session[\"uid\"])\n context = {'users': user1, 'cmp1': cmp1}\n return render(request, 'app1/accountssettings.html', context)\n except:\n return redirect('godash')\n\n\n@login_required(login_url='regcomp')\ndef updateaccounts(request):\n try:\n user = User.objects.get(id=request.session[\"uid\"])\n comp = company.objects.get(id=user.id)\n\n user.first_name = request.POST[\"first_name\"]\n user.last_name = request.POST[\"last_name\"]\n user.email = request.POST[\"email\"]\n p1 = request.POST[\"newpassword\"]\n p2 = request.POST[\"newpassword1\"]\n\n comp.cname = request.POST[\"cname\"]\n comp.caddress = request.POST[\"caddress\"]\n comp.city = request.POST[\"city\"]\n comp.state = request.POST[\"state\"]\n comp.pincode = request.POST[\"pincode\"]\n comp.cemail = request.POST[\"cemail\"]\n comp.phone = request.POST[\"phone\"]\n comp.bname = request.POST[\"bname\"]\n comp.industry = request.POST[\"industry\"]\n comp.ctype = request.POST[\"ctype\"]\n try:\n img1 = request.FILES[\"img1\"]\n comp.cimg = img1\n except:\n img2 = request.POST[\"img2\"]\n comp.cimg = img2\n\n comp.save()\n user.save()\n\n if p1 == p2:\n if p1 != \"\":\n user.set_password(p1)\n user.save()\n logout(request)\n return redirect('/app1/go')\n elif p1 == \"\":\n return redirect('/app1/godash')\n\n return redirect('/app1/godash')\n except:\n return redirect('godash')\n","repo_name":"adwaithp/Finsysapp","sub_path":"app1/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":26390,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"42862959424","text":"import os.path\nimport re\n\nfrom uttl.buildout.command_recipe import CommandRecipe\nfrom zc.buildout import UserError\n\nclass DevenvRecipe(CommandRecipe):\n\tdef __init__(self, buildout, name, options):\n\t\tsuper().__init__(buildout, name, options, executable='devenv.com')\n\n\t\t# synonyms\n\n\t\tif 'solution' in self.options:\n\t\t\tself.options['solution-path'] = self.options['solution']\n\n\t\t# solution\n\n\t\tif not 'solution-path' in self.options:\n\t\t\traise UserError('Missing mandatory \"solution-path\" option.')\n\n\t\tself.args += [ os.path.abspath(self.options['solution-path']) ]\n\n\t\t# actions on a project\n\n\t\tif 'build' in self.options:\n\t\t\tself.args += [ '/Build', self.options['build'] ]\n\n\t\tif 'rebuild' in self.options:\n\t\t\tself.args += [ '/Rebuild', self.options['rebuild'] ]\n\n\t\tif 'clean' in self.options:\n\t\t\tself.args += [ '/Clean', self.options['clean'] ]\n\n\t\tif 'deploy' in self.options:\n\t\t\tself.args += [ '/Deploy', self.options['deploy'] ]\n\n\t\tif 'project' in self.options:\n\t\t\tif not any(action in ['build', 'rebuild', 'clean', 'deploy'] for action in self.options):\n\t\t\t\traise UserError('Missing a \"build\", \"rebuild\", \"clean\", or \"deploy\" option in order to use \"project\".')\n\n\t\t\tself.args += [ '/Project', self.options['project'] ]\n\n\t\t# commands\n\n\t\tif 'command' in self.options:\n\t\t\tself.args += [ '/Command', '\"%s\"' % self.options['command'] ]\n\n\t\tself.options['args'] = ' '.join(str(e) for e in self.args)\n\n\tdef command_install(self):\n\t\tself.runCommand(self.args, parseLine=self.parseLine)\n\n\tcheck_errors = re.compile(r'.*Error: (.*)')\n\tcheck_failed = re.compile(r'.*(Build FAILED).*')\n\tcheck_artefacts = re.compile(r'.*(.+) -> (.+)')\n\n\tdef parseLine(self, line):\n\t\t# check for errors\n\n\t\tif self.check_errors.match(line) or self.check_failed.match(line):\n\t\t\treturn False\n\n\t\t# add artefacts to options\n\n\t\tmatch = self.check_artefacts.match(line)\n\t\tif match:\n\t\t\tself.options.created(match.group(2))\n\n\t\treturn True\n\ndef uninstall(name, options):\n\tpass","repo_name":"MrHands/uttl-buildout","sub_path":"uttl/buildout/devenv/devenv_recipe.py","file_name":"devenv_recipe.py","file_ext":"py","file_size_in_byte":1936,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"74622569176","text":"#!/usr/bin/env python\r\n# -*- coding: utf-8 -*-\r\n# -*- 三思网络(r34.cc) -*-\r\nfrom urllib import request\r\nimport urllib.parse\r\nimport os,os.path,json,sys\r\nimport requests\r\nimport time\r\nimport base64\r\nimport configparser\r\nimport chardet,time,subprocess\r\nimport os\r\nfrom datetime import datetime\r\n\r\n\r\nMBPATH=r'D://老黄牛小工具//word模板'\r\nJSONFILE=\"D://老黄牛小工具//ExcelQuery//temp//temp.json\"\r\n\r\nTEMPTEXT=r'..//..//ExcelQuery//temp//temp.txt'\r\nINIPATH=r\"..//..//配置文件//myconf.ini\"\r\nJBPATH=r'..//..//脚本文件'\r\n\r\nFK='临时'\r\nINIUPDATEFLAG='GetFromIni'\r\n\r\n \r\ndef savestr(str):\r\n with open(JSONFILE, 'w') as file_object:\r\n file_object.write(str)\r\n\r\ndef savesttxt(str,path=TEMPTEXT):\r\n with open(path, 'w') as file_object:\r\n file_object.write(str)\r\n\r\n\"\"\" 读取图片 \"\"\"\r\ndef get_file_content(filePath):\r\n with open(filePath, 'rb') as fp:\r\n return fp.read()\r\n\r\n\r\ndef readini(sec,key):\r\n config = configparser.ConfigParser()\r\n config.read(INIPATH,encoding='utf-8')\r\n appid = config[sec][key]\r\n return appid\r\n\r\ndef readiniconfig():\r\n config = configparser.ConfigParser()\r\n config.read(INIPATH, encoding='utf-8')\r\n return config\r\n\r\ndef updatearrfromini(arr,config=\"\"):\r\n newarr={}\r\n for inkey in arr.keys():\r\n if inkey != \"\" :\r\n newarr[inkey]=arr[inkey]\r\n\r\n if config !=\"\":\r\n # titlepr(\"config更新中:\",config)\r\n for inkey in arr.keys():\r\n if inkey != \"\" :\r\n if arr[inkey] ==INIUPDATEFLAG:\r\n try:\r\n #titlepr(\"config['ApiKeys'][inkey]\",config['ApiKeys'][inkey])\r\n newarr[inkey]=config['ApiKeys'][inkey]\r\n except:\r\n print(\"从config强制更新配置[\"+inkey+ \"]失败!~\")\r\n return newarr\r\n\r\ndef firstkey(arr):\r\n for k,v in arr.items():\r\n #print(k)\r\n return k\r\n\r\n#读取json数据\r\ndef readjson(key=\"contents\"):\r\n listarr={}\r\n with open(JSONFILE,'r',encoding='utf8') as f:\r\n ct = f.read()\r\n if ct.startswith(u'\\ufeff'):\r\n ct = ct.encode('utf8')[3:].decode('utf8')\r\n #ct = ct.replace('\\\\','\\\\\\\\')\r\n jsonarr = json.loads(ct)\r\n #print(key in jsonarr.keys())\r\n if key in jsonarr.keys():\r\n listarr = jsonarr[key]\r\n else:\r\n listarr=jsonarr\r\n return listarr\r\n\r\n#格式化文本的回车\r\ndef textgx(t,flag=\"text\"):\r\n if flag==\"text\":\r\n t=t.replace('
','')\r\n t=t.replace('
','')\r\n return t\r\n\r\n\r\n#读取txt文档,可以兼容不同的编码方式\r\ndef readtxt(path):\r\n f = open(path, 'rb')\r\n r = f.read()\r\n f_charInfo = chardet.detect(r)\r\n geshi=f_charInfo['encoding']\r\n ct=r.decode(f_charInfo['encoding']) # 通过取得的文本格式读取txt\r\n f.close()\r\n return ct\r\n\r\n#读取二进制的文件\r\ndef readfile(path):\r\n f = open(path, 'rb')\r\n r = f.read()\r\n return r\r\n\r\n\r\n#读取二进制的文件并转化为f64\r\ndef readfile2f64(path):\r\n if os.path.exists(path) :\r\n with open(path,'rb') as f:\r\n f64=base64.b64encode(f.read()).decode(\"utf-8\")\r\n else:\r\n f64=\"\"\r\n return f64\r\n#读取二进制的文件并转化为f64,然后删除\r\ndef readfile2f64anddel(path):\r\n aa=readfile2f64(path)\r\n os.remove(path)\r\n return aa\r\n#读取二进制的文件并转化为f64\r\ndef writefile64(f64,path=\"\"):\r\n if path !=\"\" and f64 !=\"\":\r\n with open(path, 'wb') as f:\r\n ff=base64.b64decode(f64)\r\n f.write(ff)\r\n\r\n return path\r\n\r\ndef runcmdstr(cmdstr):\r\n tmp = os.popen(cmdstr)\r\n res = tmp.read()# 要用read()方法读取后才是文本对象\r\n tmp.close()# 需将对象关闭\r\n return res\r\n\r\n\r\ndef randfile(valarr=\"\",fkeyold=\"\",flag=\"new\"):\r\n #print(\"randfile---\"+\"fkeyold:\"+fkeyold+\"flag:\"+flag)\r\n #print(valarr)\r\n #print(\"valarr---\")\r\n nowpath= str(os.getcwd()) + \"/\"\r\n \r\n if fkeyold==\"\":\r\n return nowpath+ str(int(round(time.time() * 1000)))+\".tmp\"\r\n \r\n try:\r\n f=valarr[fkeyold]\r\n except:\r\n return nowpath + str(int(round(time.time() * 1000)))+\".tmp\"\r\n\r\n\r\n fex=os.path.splitext(f)[-1]\r\n #print(\"fex---\"+\" fex:\"+fex+\" flag:\"+flag)\r\n if flag==\"new\":\r\n #print(\"flag in \")\r\n return nowpath + str(int(round(time.time() * 1000)))+\".tmpnew\"+fex\r\n else:\r\n #print(\"flag not in \")\r\n return nowpath + str(int(round(time.time() * 1000)))+\".tmp\"+fex\r\n\r\n path =nowpath + str(int(round(time.time() * 1000)))+\".tmp\"\r\n return path\r\n \r\n\r\n\"\"\" 等待脚本执行结束 \"\"\"\r\ndef runwait(cmd):\r\n p=subprocess.Popen(cmd,shell=True)\r\n return_code=p.wait() #等待子进程结束,并返回状态码;\r\n return return_code\r\n \r\n\r\n\r\n\r\n#安全的获得键值\r\ndef safegetkey(arr,key):\r\n try:\r\n s=arr[key]\r\n except:\r\n s=\"\"\r\n return s\r\n\r\n\r\n#删除文件,文件可以不存在\r\ndef safedel(path):\r\n if os.path.exists(path):\r\n os.remove(path)\r\n safedeltmpfile()\r\n return 0\r\n\r\n#删除当前文件夹的临时文件tt=300秒\r\ndef safedeltmpfile(tt=300):\r\n errstr=\"\"\r\n root = os.getcwd()\r\n for dirpath, dirnames, filenames in os.walk(root):\r\n for filepath in filenames:\r\n if \".tmp\" in filepath:\r\n try:\r\n mtime = os.path.getmtime(filepath) #修改时间\r\n nowtime = time.time()\r\n if nowtime-mtime >tt:\r\n if os.path.exists(filepath):\r\n os.remove(filepath)\r\n except Exception as e:# 保存函数出错后的执行结果\r\n errstr = '错误类型:'+ e.__class__.__name__+\"\\n\"+ '错误明细:'+str(e)\r\n #print(errstr)\r\n return errstr\r\n\r\n\"\"\" 字符串变base64的字符串 \"\"\"\r\ndef str2f64(s):\r\n try:\r\n s=str.encode(s)\r\n except:\r\n a=1\r\n json64=base64.b64encode(s).decode(\"utf-8\")\r\n return json64\r\n\r\n\"\"\" base64转换为字符 \"\"\"\r\ndef f64tostr(f64):\r\n json64=base64.b64decode(f64)\r\n return json64\r\n\r\n#读取json数据\r\ndef readjsonarr(key=\"\",key2=\"\"):\r\n #print(key+\"--\"+key2)\r\n listarr={}\r\n with open(JSONFILE,'r',encoding='utf8') as f:\r\n ct = f.read()\r\n if ct.startswith(u'\\ufeff'):\r\n ct = ct.encode('utf8')[3:].decode('utf8')\r\n #ct = ct.replace('\\\\','\\\\\\\\')\r\n jsonarr = json.loads(ct)\r\n \r\n \r\n\r\n if key2==\"\" and key==\"\":\r\n return jsonarr[key]\r\n elif key != \"\":\r\n return jsonarr[key][key2]\r\n elif key==\"\":\r\n return jsonarr\r\n else:\r\n return \"error\"\r\n\r\n#字典合并\r\ndef merge(dict1, dict2):\r\n newdict={}\r\n for key in dict1.keys():\r\n if key !=\"\":\r\n newdict[key]=dict1[key]\r\n for key in dict2.keys():\r\n if key !=\"\":\r\n if key not in dict1.keys():\r\n newdict[key]=dict2[key]\r\n\r\n return newdict \r\n#字典相减\r\ndef dictdiff(dict1, dict2):\r\n newdict={}\r\n for key in dict1.keys():\r\n newdict[key]=dict1[key]\r\n for key in dict2.keys():\r\n if key in dict1.keys():\r\n del newdict[key]\r\n\r\n return newdict\r\n\r\n#备用的统计函数等 \r\ndef tj():\r\n return \"\"\r\n\r\n\r\n\r\n#这里替换成常用函数\r\ndef arr2dict(arr2):\r\n n=0\r\n i=0\r\n k={}\r\n dictarr={}\r\n dictarrs={}\r\n for arr1 in arr2:\r\n i=0\r\n for arrval in arr1:\r\n #print(arrval)\r\n if n==0:\r\n k[i]=arrval\r\n else:\r\n dictarr[k[i]]=arrval\r\n i=i+1\r\n dictarrs[n]=dictarr\r\n n=n+1\r\n\r\n return dictarrs\r\n\r\n\r\ndef firstval(arr):\r\n for index in arr:\r\n return arr[index]\r\n\r\n\r\ndef savesttxt(str,path=TEMPTEXT):\r\n with open(path, 'w') as file_object:\r\n file_object.write(str)\r\n\r\n\r\ndef startstr(str):\r\n savesttxt(str)\r\n os.system(\"cmd.exe /c start \" +TEMPTEXT)\r\n\r\ndef startarr(arr):\r\n str=json.dumps(arr)\r\n startstr(str)\r\n\r\n\r\n\r\n#保证键值存在,不存在退出\r\ndef checkkey(arr,key):\r\n if key in arr.keys():\r\n return True\r\n else:\r\n print(\"Error,请确保标题行中有【\" + key + \"】\")\r\n time.sleep(10)\r\n sys.exit()\r\n \r\n\"\"\" 打印输出信�� \"\"\"\r\ndef pr(str,prflag=\"true\"):\r\n if prflag==\"true\":\r\n print(str)\r\n\"\"\" 打印输出信息 \"\"\"\r\ndef titlepr(title,arr,prflag=\"true\"):\r\n if prflag==\"true\":\r\n if title !=\"\":\r\n print(\" ----------- \" + title + \" -----------\")\r\n print(arr)\r\n else:\r\n print(arr)\r\n\r\n\r\n\"\"\" 字典保证值的存在 \"\"\"\r\ndef arrstr(arr,key,s=\"\"):\r\n if key in arr.keys():\r\n return arr\r\n else:\r\n arr[key]=s\r\n return arr\r\n\r\n\r\n\r\n\r\n#这里替换成常用函数\r\ndef savearr2json(arr,wtype=\"all\",code=\"0\"):\r\n arr2=arr2jsonarr(arr,wtype,code)\r\n with open(JSONFILE, 'w') as file_object:\r\n file_object.write(json.dumps(arr2))\r\n return arr2\r\n\r\n\r\n\r\n\"\"\" 错误执行时存到json \"\"\"\r\ndef saveerr2json(arr,s=\"\",key=\"执行结果\",):\r\n arr[key]=s\r\n savearr2json(arr,\"all\")\r\n print(\"函数savearr2json print: \"+s)\r\n return arr\r\n \r\n\"\"\" 数组生成json对应的数组 \"\"\"\r\ndef arr2jsonarr(arr,wtype=\"all\",code=\"0\"):\r\n sc=os.path.splitext(os.path.basename(__file__))[0]\r\n arr2={}\r\n arr2['script']=sc\r\n arr2['w']=wtype\r\n arr2['code']=code\r\n arr2['contents']=arr\r\n return arr2\r\n\r\n\"\"\" 数组转成json字符串 \"\"\"\r\ndef arr2json64str(arr,wtype=\"all\",code=\"0\"):\r\n arr=arr2jsonarr(arr,wtype,code)\r\n jsonstr=json.dumps(arr)\r\n return str2f64(jsonstr) \r\n\r\n\"\"\" base64转成jsonarr \"\"\"\r\ndef json64tojsonarr(json64):\r\n json64str=base64.b64decode(json64) #??原来写成fd2.json64\r\n jsonarr = json.loads(json64str)\r\n return jsonarr\r\n\r\n\r\n#list可以转字符串\r\ndef list2str3(listorstr,cols=\"\"):\r\n txt =\"\"\r\n flag = isinstance(listorstr,list)\r\n print(flag)\r\n return listorstr\r\n\r\n#list可以转字符串\r\ndef list2str(listorstr,cols=\"\"):\r\n txt =\"\"\r\n flag = isinstance(listorstr,list)\r\n\r\n if flag ==True:\r\n for v in listorstr:\r\n txt = txt + cols + list2str(v)\r\n else:\r\n txt =listorstr\r\n return str(txt)\r\n\r\n\r\n\r\n\"\"\" 数组变成base64的字符串 \"\"\"\r\ndef arr2json64(ret2arr):\r\n jsonarr=arr2jsonarr(ret2arr) #结构化数据\r\n jsonarrstr=json.dumps(jsonarr) #编码为字符串\r\n return str2f64(jsonarrstr)\r\n\r\n\r\ndef sysargvandjson(keyword=\"webppath\",inputstr=\"\"):\r\n inputpatharr={}\r\n #print(inputstr !=\"\")\r\n if inputstr !=\"\": #1.如果有输入,就强制跳转\r\n inputpatharr[0]=inputstr\r\n return inputpatharr\r\n #print(keyword)\r\n if len(sys.argv)>1: #2.判断有没有参数\r\n n = len(sys.argv)\r\n temp=\"\"\r\n for i in range(1, n):\r\n inputpatharr[i]=sys.argv[i]\r\n else: #3.读取json中的配置,看看有没有keyword\r\n jsonarr=readjson()\r\n #print(jsonarr)\r\n if keyword in jsonarr.keys():\r\n inputpatharr[0] = jsonarr[keyword]\r\n else:\r\n inputpatharr = {}\r\n\r\n print(inputpatharr)\r\n return inputpatharr\r\n\r\ndef sysargv1(inputstr=r\"C:\\Users\\lilyhcn\\Desktop\\4\\1.webp\"): #只有一个参数\r\n if len(sys.argv)>1: #判断有没有参数\r\n n = len(sys.argv)\r\n temp=\"\"\r\n for i in range(1, n):\r\n if i == 1 :\r\n temp = temp + sys.argv[i]\r\n else:\r\n temp = temp + sys.argv[i] +\" \"\r\n inputstr=temp\r\n #print(\"input: \"+ inputstr)\r\n return inputstr\r\n\r\n\r\ndef readimg2base64(filepath=\"\"):\r\n with open(filepath, 'rb') as fp:\r\n data = fp.read()\r\n return base64.b64encode(data)\r\n\r\n\r\ndef upload(path):\r\n headers = {'Authorization': 'W2v89ImTEPqkCVr03dpRfW3wvjwbHyjj'}\r\n files = {'smfile': open(path, 'rb')}\r\n url = 'https://sm.ms/api/v2/upload'\r\n res = requests.post(url, files=files, headers=headers).json()\r\n aa=json.dumps(res, indent=4)\r\n aa=json.loads(aa)\r\n successstr = aa['success']\r\n #print(successstr)\r\n if successstr == False:\r\n imageurl=aa['images']\r\n elif successstr == True:\r\n imageurl=aa['data']['url']\r\n else:\r\n print('error')\r\n return imageurl\r\n\r\ndef test():\r\n aa='this is test!~'\r\n print(aa)\r\n return aa\r\n\r\n\r\n#把相对路径转成绝对路径\r\ndef getwnewfilepath(output1,listarr):\r\n return 11\r\n\r\n#把相对路径转成绝对路径\r\n#raltiveapth 相对路径\r\n#folder 文件夹\r\ndef getwholepath(raltiveapth,folder=\"\"):\r\n if folder ==\"\":\r\n folder =os.getcwd()\r\n if \":\" in raltiveapth:\r\n new_file = raltiveapth\r\n else:\r\n new_file = folder+\"\\\\\"+raltiveapth\r\n #print(new_file)\r\n return new_file\r\n\r\n\r\ndef mkdir(path):\r\n folder=os.path.dirname(path)\r\n print(folder)\r\n folderexist = os.path.exists(folder)\r\n if not folderexist:\r\n os.makedirs(folder)\r\n\r\n\r\ndef GetDesktopPath():\r\n return os.path.join(os.path.expanduser(\"~\"), 'Desktop')\r\n\r\n#这里是专用函数,其它地方用不着\r\n\r\ndef getexepath(jbname):\r\n jbstr = \"ahk,vbs,py\" #可能的后缀\r\n for jbext in jbstr.split(\",\"): #遍历可能的后缀\r\n jbpath=JBPATH + '/' + jbext +'/' + jbname + '.' + jbext\r\n #print(jbpath)\r\n if os.path.exists(jbpath):\r\n return jbpath\r\n return \"JbNotExist\" #都找不到,就返回空\r\n\r\n\r\n\r\ndef printvalarr(valarr,errstr=\"错误输出值\",prflag=True):\r\n valarr[\"执行结果\"]=errstr\r\n if prflag==True :\r\n print(errstr)\r\n return valarr\r\n\r\n#读取输入的数组\r\ndef getvalarr(jsonarr,inarr,outarr,prflag=\"false\"):\r\n valarr,contarr,temparr={},{},{}\r\n #计算出contarr\r\n contarr=jsonarr[\"contents\"]\r\n contarr=dictdiff(contarr,outarr)\r\n temparr[\"执行结果\"]=\"\"\r\n contarr=dictdiff(contarr,temparr)\r\n \r\n key=getonlykey(contarr)\r\n if len(contarr)==1 and len(inarr)==1:\r\n key=getonlykey(inarr)\r\n key2=getonlykey(contarr)\r\n\r\n if key2 == FK:\r\n key2=getonlykey(contarr)\r\n valarr[key]=contarr[key2]\r\n else:\r\n for inkey in inarr.keys():\r\n valarr[inkey]=contarr[inkey]\r\n\r\n else:\r\n for inkey in inarr.keys():\r\n valarr[inkey]=contarr[inkey]\r\n valarr=merge(valarr,contarr)\r\n \r\n return valarr\r\n\r\n#读取输入的数组\r\ndef getonlykey(arr):\r\n keylen=len(arr)\r\n if keylen==1:\r\n for key in arr:\r\n key=key\r\n else:\r\n key=\"error,getonlykey\"\r\n return key\r\n\r\n\r\n\r\n#处理计算得到的数组\r\ndef mboutputarr(fd2,prflag=\"true\",arr2ret={},f64=\"\",wpath=\"\",keyflag=\"all\"):\r\n #一、 是否有函数的输入\r\n #titlepr(\"运行后的返回值:\",arr2ret,prflag) \r\n if \"执行结果\" not in arr2ret: \r\n arr2ret[\"执行结果\"] = \"√\"\r\n \r\n if fd2=={}:\r\n savearr2json(arr2ret, keyflag)\r\n\r\n arr2ret64 = arr2json64str(arr2ret,keyflag)\r\n \r\n fd2new={}\r\n fd2new[\"json64\"]=arr2ret64\r\n fd2new[\"f64\"]=f64\r\n\r\n if f64!=\"\" and wpath !=\"\":\r\n writefile64(f64,wpath)\r\n \r\n #titlepr(\"最终输出的数组为:\",fd2new,prflag)\r\n #print(\"mboutputarr 运行结束\")\r\n return fd2new\r\n\r\n\r\n\r\n\r\n\r\n\r\n#获取fd2的字典\r\ndef getfd2(fd2={},flag=\"json64\",path=\"\"):\r\n\r\n if flag==\"json64\":\r\n if fd2=={}:#看是否函数传入\r\n\r\n json64=readfile2f64(JSONFILE)\r\n else:\r\n #按照默认标题读取\r\n try:\r\n json64=fd2[\"json64\"]\r\n except:\r\n json64=fd2.json64\r\n return json64\r\n elif flag==\"f64\":\r\n\r\n if fd2=={} and path !=\"\":#看是否函数传入\r\n f64=readfile2f64(path)\r\n else:\r\n try:\r\n f64=fd2[\"f64\"]\r\n except:\r\n try:\r\n f64=fd2.f64\r\n except:\r\n f64=\"\"\r\n\r\n return f64\r\n else:\r\n return \"getfd2arr未知错误。\"\r\n\r\n#获取fd2的字典\r\ndef getfd2_f64(fd2={},fkeyold=\"\",jsonarr={}):\r\n f64=\"\"\r\n if fd2=={} and fkeyold !=\"\":#看是否函数传入\r\n inpath=jsonarr[\"contents\"][fkeyold]\r\n f64=readfile2f64(inpath)\r\n return f64\r\n elif fd2!={}:\r\n try:\r\n f64=fd2[\"f64\"]\r\n except:\r\n try:\r\n f64=fd2.f64\r\n except:\r\n f64=\"\"\r\n\r\n return f64\r\n elif fd2=={} and fkeyold ==\"\":#看是否函数传入\r\n return \"\"\r\n else:\r\n return \"getfd2arr未知错误。\"\r\n\r\n","repo_name":"lilyhcn1/ExcelRunAnyCode","sub_path":"老黄牛小工具/脚本文件/py/lilyfun.py","file_name":"lilyfun.py","file_ext":"py","file_size_in_byte":16636,"program_lang":"python","lang":"en","doc_type":"code","stars":26,"dataset":"github-code","pt":"68"}
+{"seq_id":"31053229902","text":"import sys\nsys.stdin = open('2669.txt')\n\ncanvas = [[0]*100 for _ in range(100)]\ncnt = 0\nfor i in range(4):\n x1, y1, x2, y2 = map(int, input().split())\n for x in range(x1, x2):\n for y in range(y1, y2):\n if canvas[x][y] == 1:\n continue\n canvas[x][y] = 1\n cnt += 1\n\nprint(cnt)","repo_name":"gyur1kim/APS-SWEA","sub_path":"IM시험대비/BOJ/0825/2669_직사각형.py","file_name":"2669_직사각형.py","file_ext":"py","file_size_in_byte":334,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"35447027476","text":"# Imports\nfrom random import choice\nfrom pickle import Pickler, Unpickler\nimport os\n\nos.chdir(\"./Python/pendu\")\n\n# Liste mots\nwords_list = [\"banane\", \"citron\", \"ananas\", \"orange\", \"fraise\", \"cerise\"]\n\n# Initialise le score \ndef init_score(file):\n try:\n with open(file, \"rb\") as scores_file:\n score_dict = Unpickler(scores_file).load()\n except:\n score_dict = {}\n return score_dict\n\n# Récupère un mot au hasard\ndef get_random_elt(elts_list):\n return choice(elts_list)\n\n# Génére un mot masqué à partir d'un mot et d'une liste de lettre à affiché\ndef gen_masked_word(word, letters_list):\n word_masked = \"\"\n for letter in word:\n if letter in letters_list:\n word_masked += letter\n else:\n word_masked += \"*\"\n return word_masked\n\n# Récupère l'input du joueur, vérifie si l'input est une lettre et retourne cette lettre\ndef get_player_input():\n player_input = input(\"Tapez une lettre : \")\n if not(player_input.isalpha()) or len(player_input) > 1:\n print(\"Valeur invalide\")\n get_player_input()\n return player_input.lower()\n\n# Compare le nouveau score avec l'ancien et garde le plus élevé. Retourne le dico mis à jour \ndef get_player_score(pseudo, score, dico):\n if pseudo in dico:\n score = max(dico[pseudo], score)\n dico.update([(pseudo, score)])\n return (dico)\n\n# Initialisation\ngame_over = False\nlives_left = 8\nword_random = get_random_elt(words_list)\nletters_guessed = []\nword_masked = gen_masked_word(word_random, letters_guessed)\nscores = init_score(\"scores\")\nplayer_pseudo = input(\"Entre votre pseudo : \")\n\n# Game start\nwhile not(game_over):\n print(word_masked)\n player_input = get_player_input()\n if player_input in letters_guessed:\n print(\"Vous avez déjà essayé la lettre \" + player_input)\n elif player_input in word_random:\n letters_guessed.append(player_input)\n else: \n letters_guessed.append(player_input)\n lives_left -= 1\n print(\"{} ne se trouve pas dans le mot, il vous reste {} essai(s)\".format(player_input, lives_left))\n\n word_masked = gen_masked_word(word_random, letters_guessed)\n \n if word_masked == word_random:\n print(\"Bien joué ! Le mot était {}.\".format(word_random))\n game_over = True\n \n player_score = lives_left + len(word_random)\n scores = get_player_score(player_pseudo, player_score, scores)\n \n with open(\"scores\", \"wb\") as scores_file:\n Pickler(scores_file).dump(scores) \n elif lives_left <= 0:\n print(\"Vous avez perdu :/ Le mot était {}.\".format(word_random))\n game_over = True\n\nprint(scores)","repo_name":"Stephane0713/html-css-js","sub_path":"Python/pendu/pendu.py","file_name":"pendu.py","file_ext":"py","file_size_in_byte":2695,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"28370522554","text":"from pytranslations import translations\n\ndef test_no_users():\n tr = translations.Translations()\n with open(\"src/tests/test_data.json\", 'r') as f:\n tr.get_translations_from_json(f.read())\n tr.default_language = \"en\"\n assert tr.t_lang(\"ru\", \"translation_1\", \"abc\", \"cde\") == \"Result abc abc phrase cde ru\"\n assert tr.t_lang(\"ru\", \"translation_3\", \"abc\") == \"Result phrase abc en 3\"\n\ndef test_users():\n tr = translations.Translations()\n with open(\"src/tests/test_data.json\", 'r') as f:\n tr.get_translations_from_json(f.read())\n tr.default_language = \"en\"\n tr.add_user(\"1\", \"en\")\n tr.add_user(\"2\", \"en\")\n tr.add_user(\"3\", \"ru\")\n assert tr.t_id(\"3\", \"translation_3\", \"qweqwe\") == \"Result phrase qweqwe en 3\"\n \n","repo_name":"YLashko/translations","sub_path":"src/tests/test_translations.py","file_name":"test_translations.py","file_ext":"py","file_size_in_byte":758,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"68"}
+{"seq_id":"40521069636","text":"\r\n\r\nfrom pywifi import const\r\nimport time\r\nimport tkinter.filedialog #\r\nimport tkinter.messagebox #\r\n\r\nimport random\r\nfrom tkinter import *\r\nfrom tkinter import ttk\r\nimport pywifi\r\n\r\nclass MY_GUI():\r\n def __init__(self, init_window_name):\r\n self.init_window_name = init_window_name\r\n #\r\n self.get_value = StringVar() #\r\n #\r\n self.get_wifi_value = StringVar()\r\n #\r\n self.get_wifimm_value = StringVar()\r\n #\r\n self.wifi = pywifi.PyWiFi()\r\n #\r\n self.iface = self.wifi.interfaces()[0]\r\n #\r\n self.iface.disconnect()\r\n time.sleep(1) #\r\n #\r\n assert self.iface.status() in \\\r\n [const.IFACE_DISCONNECTED, const.IFACE_INACTIVE]\r\n\r\n def __str__(self):\r\n #\r\n return '(WIFI:%s,%s)' % (self.wifi, self.iface.name())\r\n\r\n #\r\n def set_init_window(self):\r\n self.init_window_name.title(\"WIFI CRACKING TOOL\")\r\n self.init_window_name.geometry('+500+200')\r\n labelframe = LabelFrame(width=400, height=200, text=\"Menu-Broke WIFI automatically\") # \r\n labelframe.grid(column=0, row=0, padx=10, pady=10)\r\n self.search1 = Label(labelframe, text=\"Broke WIFI\").grid(column=0, row=0)\r\n self.search = Button(labelframe, text=\"Searching WiFi\", command=self.scans_wifi_list).grid(column=0, row=0)\r\n self.pojie = Button(labelframe, text=\"Cracking\", command=self.readPassWord).grid(column=1, row=0)\r\n self.label = Label(labelframe, text=\"Path:\").grid(column=0, row=1)\r\n self.label1 = Button(labelframe, text=\"Improve Security\", command=self.improvePassWord).grid(column=1, row=0)\r\n self.path = Entry(labelframe, width=12, textvariable=self.get_value).grid(column=1, row=1)\r\n self.file = Button(labelframe, text=\"add pwd txt\", command=self.add_mm_file).grid(column=2, row=1)\r\n self.wifi_text = Label(labelframe, text=\"WiFiID:\").grid(column=0, row=2)\r\n self.wifi_input = Entry(labelframe, width=12, textvariable=self.get_wifi_value).grid(column=1, row=2)\r\n self.wifi_mm_text = Label(labelframe, text=\"WiFiPWD:\").grid(column=2, row=2)\r\n self.wifi_mm_input = Entry(labelframe, width=10, textvariable=self.get_wifimm_value).grid(column=3, row=2,sticky=W)\r\n self.wifi_labelframe = LabelFrame(text=\"wifiLIST\")\r\n self.wifi_labelframe.grid(column=0, row=3, columnspan=4, sticky=NSEW)\r\n #\r\n self.wifi_tree = ttk.Treeview(self.wifi_labelframe, show=\"headings\", columns=(\"a\", \"b\", \"c\", \"d\"))\r\n self.vbar = ttk.Scrollbar(self.wifi_labelframe, orient=VERTICAL, command=self.wifi_tree.yview)\r\n self.wifi_tree.configure(yscrollcommand=self.vbar.set)\r\n #\r\n self.wifi_tree.column(\"a\", width=50, anchor=\"center\")\r\n self.wifi_tree.column(\"b\", width=100, anchor=\"center\")\r\n self.wifi_tree.column(\"c\", width=100, anchor=\"center\")\r\n self.wifi_tree.column(\"d\", width=100, anchor=\"center\")\r\n self.wifi_tree.heading(\"a\", text=\"WiFiID\")\r\n self.wifi_tree.heading(\"b\", text=\"SSID\")\r\n self.wifi_tree.heading(\"c\", text=\"BSSID\")\r\n self.wifi_tree.heading(\"d\", text=\"signal\")\r\n self.wifi_tree.grid(row=4, column=0, sticky=NSEW)\r\n self.wifi_tree.bind(\"\", self.onDBClick)\r\n self.vbar.grid(row=4, column=1, sticky=NS)\r\n\r\n #\r\n def scans_wifi_list(self): #\r\n #\r\n print(\"^_^ Start scaning wifi...\")\r\n self.iface.scan()\r\n time.sleep(9)\r\n #\r\n scanres = self.iface.scan_results()\r\n #\r\n nums = len(scanres)\r\n print(\"Amount: %s\" % (nums))\r\n #\r\n self.show_scans_wifi_list(scanres)\r\n return scanres\r\n\r\n #\r\n def show_scans_wifi_list(self, scans_res):\r\n for index, wifi_info in enumerate(scans_res):\r\n self.wifi_tree.insert(\"\", 'end', values=(index + 1, wifi_info.ssid, wifi_info.bssid, wifi_info.signal))\r\n\r\n #\r\n def add_mm_file(self):\r\n self.filename = tkinter.filedialog.askopenfilename()\r\n self.get_value.set(self.filename)\r\n\r\n #\r\n def onDBClick(self, event):\r\n self.sels = event.widget.selection()\r\n self.get_wifi_value.set(self.wifi_tree.item(self.sels, \"values\")[1])\r\n\r\n #\r\n def readPassWord(self):\r\n self.getFilePath = self.get_value.get()\r\n self.get_wifissid = self.get_wifi_value.get()\r\n pwdfilehander = open(self.getFilePath, \"r\", errors=\"ignore\")\r\n while True:\r\n try:\r\n self.pwdStr = pwdfilehander.readline()\r\n if not self.pwdStr:\r\n break\r\n self.bool1 = self.connect(self.pwdStr, self.get_wifissid)\r\n if self.bool1:\r\n self.res = \"[*] pwd correct!wifi:%s,pwd:%s \" % (self.get_wifissid, self.pwdStr)\r\n self.get_wifimm_value.set(self.pwdStr)\r\n tkinter.messagebox.showinfo('attention', 'crack successful!!!')\r\n print(self.res)\r\n break\r\n else:\r\n self.res = \"[*] pwd wrong!wifi:%s,pwd:%s\" % (self.get_wifissid, self.pwdStr)\r\n print(self.res)\r\n time.sleep(4)\r\n except:\r\n continue\r\n\r\n def improvePassWord(self):\r\n self.getPath = self.get_value.get()\r\n self.get_wifissid = self.get_wifi_value.get()\r\n pwdfilehander = open(self.getPath, \"r\", errors=\"ignore\")\r\n while True:\r\n try:\r\n self.pwdStr = pwdfilehander.readline()\r\n if not self.pwdStr:\r\n break\r\n self.bool1 = self.connect(self.pwdStr, self.get_wifissid)\r\n if self.bool1:\r\n self.res = \"[*] pwd improve!wifi:%s,pwd:%s \" % (self.get_wifissid+\"123\", self.pwdStr)\r\n self.get_wifimm_value.set(self.pwdStr)\r\n tkinter.messagebox.showinfo('attention', 'improve successful!!!')\r\n print(self.res)\r\n break\r\n else:\r\n self.res = \"[*] pwd wrong!wifi:%s,pwd:%s\" % (self.get_wifissid, self.pwdStr)\r\n print(self.res)\r\n time.sleep(4)\r\n except:\r\n continue\r\n #\r\n def connect(self, pwd_Str, wifi_ssid):\r\n # 创建wifi链接文件\r\n self.profile = pywifi.Profile()\r\n self.profile.ssid = wifi_ssid #\r\n self.profile.auth = const.AUTH_ALG_OPEN #\r\n self.profile.akm.append(const.AKM_TYPE_WPA2PSK) #\r\n self.profile.cipher = const.CIPHER_TYPE_CCMP #\r\n self.profile.key = pwd_Str #\r\n self.iface.remove_all_network_profiles() #\r\n self.tmp_profile = self.iface.add_network_profile(self.profile) #\r\n self.iface.connect(self.tmp_profile) #\r\n time.sleep(5)\r\n if self.iface.status() == const.IFACE_CONNECTED: #\r\n isOK = True\r\n else:\r\n isOK = False\r\n self.iface.disconnect() #\r\n time.sleep(1)\r\n #\r\n assert self.iface.status() in \\\r\n [const.IFACE_DISCONNECTED, const.IFACE_INACTIVE]\r\n return isOK\r\n\r\n\r\ndef gui_start():\r\n init_window = Tk()\r\n ui = MY_GUI(init_window)\r\n print(ui)\r\n ui.set_init_window()\r\n init_window.mainloop()\r\n\r\n\r\nif __name__ == \"__main__\":\r\n gui_start()\r\n\r\n","repo_name":"hwen554/Crack-Pwd","sub_path":"WIFI CRACKING.py","file_name":"WIFI CRACKING.py","file_ext":"py","file_size_in_byte":7429,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"68"}
+{"seq_id":"25932667021","text":"# -*- coding: utf-8 -*-\n\nimport datetime\nimport json\n\nfrom decimal import Decimal\nfrom collections import OrderedDict, namedtuple\nfrom collections.abc import Iterable\nfrom itertools import chain as iter_chain\n\nfrom ..domobj import dset, DSetBase, dobject, reshape\nfrom .dtable import json_object, dtable, dsequence\n\nfrom sqlblock import SQL\nfrom sqlblock.asyncpg import transaction\n\n\n_EntryTuple = namedtuple('EntryTuple', ['pkey_attrs', 'pkey_values',\n 'attrs', 'values'])\n\ndef _dtable_diff(current, past):\n \"\"\"diff dtable object, return the delta information.\n\n The delta information is a tuple, the data is added, changged and removed.\n \"\"\"\n\n inslst = [] # [obj] the objects to be inserted\n dellst = [] # [objid] the objid to be deleted\n chglst = [] # [{attr: (current, past)}], modified\n\n item_type = current.__dset_item_class__\n pkey_attrs = item_type.__dobject_key__\n value_attrs = item_type.__dobject_att__\n for curr_obj in current:\n if curr_obj not in past:\n inslst.append(curr_obj)\n continue\n\n past_obj = past[curr_obj]\n\n modified = OrderedDict()\n for attr_name in value_attrs:\n if not hasattr(past_obj, attr_name):\n continue\n\n newval = getattr(curr_obj, attr_name)\n oldval = getattr(past_obj, attr_name)\n if newval == oldval:\n continue\n\n modified[attr_name] = (newval, oldval)\n\n if modified:\n chglst.append((curr_obj.__dobject_key__, modified))\n\n for past_obj in past:\n if past_obj not in current:\n dellst.append(past_obj)\n\n # inserted data tuple\n pkvals, values = [], []\n for obj in inslst: # objects to be inserted\n pkvals.append(tuple(getattr(obj, f) for f in pkey_attrs))\n\n attr_values = []\n for attr_name, attr in value_attrs.items():\n attr_value = getattr(obj, attr_name)\n\n if (issubclass(attr.type, json_object)\n and not isinstance(attr_value, str)): # json cast\n attr_value = json.dumps(attr_value)\n\n attr_values.append(attr_value)\n\n values.append(tuple(attr_values))\n\n dt_ins = _EntryTuple(pkey_attrs, pkvals, value_attrs, values)\n\n # deleted data tuple\n item_type = past.__dset_item_class__\n pkvals = []\n for obj in dellst:\n pkvals.append(tuple(getattr(obj, f) for f in pkey_attrs))\n\n dt_del = _EntryTuple(pkey_attrs, pkvals, None, None)\n\n # modified data tuple\n\n pkvals = []\n values = []\n for objid, modified in chglst:\n pkvals.append(tuple(getattr(objid, f) for f in pkey_attrs))\n values.append(modified)\n\n dt_chg = _EntryTuple(pkey_attrs, pkvals, value_attrs, values)\n\n return dt_ins, dt_chg, dt_del\n\n@transaction._dsn_db\nasync def pq_dtable_merge(current, past, _dsn_db=None):\n\n dins, dchg, ddel = _dtable_diff(current, past)\n table_name = current.__dset_item_class__.__name__\n\n dobj_cls = current.__dset_item_class__\n attrs = OrderedDict((attr_name, attr) for attr_name, attr in\n iter_chain(dobj_cls.__dobject_key__.items(),\n dobj_cls.__dobject_att__.items()))\n\n seq_attrs = {}\n for n, attr in attrs.items():\n if issubclass(attr.type, dsequence):\n seq_attrs[n] = attr\n\n dobj_cls = current.__dset_item_class__\n attrs = OrderedDict((attr_name, attr) for attr_name, attr in\n iter_chain(dobj_cls.__dobject_key__.items(),\n dobj_cls.__dobject_att__.items()))\n seq_attrs = {}\n for n, attr in attrs.items():\n if issubclass(attr.type, dsequence):\n seq_attrs[n] = attr\n\n if dins.values:\n await _do_insert(table_name, seq_attrs, dins, _dsn_db)\n\n if dchg.values:\n await _do_update(table_name, seq_attrs, dchg, _dsn_db)\n\n if ddel.pkey_values:\n await _do_delete(table_name, ddel, _dsn_db)\n\nasync def _do_insert(table_name, seq_attrs, dins, _dsn_db=None):\n\n cols = tuple(iter_chain(dins.pkey_attrs.keys(), dins.attrs.keys()))\n values = [k + v for k, v in zip(dins.pkey_values, dins.values)]\n\n\n\n # If there are new sequence objects,\n # get next values of them in a batch\n if seq_attrs:\n seq_cols = [] # (col_idx, col_name, [seq_value])\n for i, colname in enumerate(cols):\n if colname in seq_attrs:\n seq_cols.append((i, colname, []))\n\n for record in values:\n for seq_col in seq_cols:\n seq_val = record[seq_col[0]]\n if seq_val is not None:\n seq_col[2].append(seq_val)\n\n for colidx, colname, seqvals in seq_cols:\n allocate_sequence(seq_attrs[colname], seqvals)\n\n _dsn_db << f\"\"\"\n INSERT INTO {table_name}\n ({', '.join(cols)}) VALUES\n ({','.join([ '{' + c + '}' for c in cols])})\n \"\"\"\n params_list = list(\n dict(t for t in zip(cols, col_values))\n for col_values in values\n )\n await _dsn_db(params_list)\n\nasync def _do_update(table_name, seq_attrs, dchg, _dsn_db=None):\n if seq_attrs:\n # for attrname in dchg.values.items():\n seq_cols = {}\n for record in dchg.values:\n for colname in record:\n if colname in seq_attrs:\n try:\n seqvals = seq_cols[colname]\n except KeyError:\n seq_cols[colname] = seqvals = []\n\n seqvals.append(record[colname][0])\n\n for colname, seqvals in seq_cols.items():\n allocate_sequence(seq_attrs[colname], seqvals)\n\n pk_attrs = list(dchg.pkey_attrs.keys())\n pk_expr = SQL(*(f\"{n}={{{n}}}\" for n in pk_attrs), sep=' AND ')\n for i in range(len(dchg.values)):\n params = OrderedDict(p for p in zip(pk_attrs, dchg.pkey_values[i]))\n for f, v in dchg.values[i].items():\n params[f] = v[0]\n\n chg_attrs = list(dchg.values[i].keys())\n val_expr = SQL(*(f\"{n}={{{n}}}\" for n in chg_attrs), sep=', ')\n\n _dsn_db << f\"\"\"\n UPDATE {table_name} SET\n {{val_expr}}\n WHERE {{pk_expr}}\n \"\"\"\n await _dsn_db(**params)\n\nasync def _do_delete(table_name, ddel, _dsn_db=None):\n values = []\n pk_attrs = list(ddel.pkey_attrs.keys())\n for i in range(len(ddel.pkey_values)):\n r = OrderedDict(p for p in zip(pk_attrs, ddel.pkey_values[i]))\n values.append(r)\n\n pk_expr = SQL(*(f\"{n}={{{n}}}\" for n in pk_attrs), sep=' AND ')\n\n _dsn_db << f\"\"\"\n DELETE FROM {table_name} WHERE {{pk_expr}}\n \"\"\"\n await _dsn_db(values)\n\n\ndef allocate_sequence(attr, seq_values):\n pending = [i for i, s in enumerate(seq_values) if not s.allocated]\n if not pending:\n return\n\n # nextval of sequence\n seqname = attr.type.__name__\n newvals = dbc.nextval(seqname, batch_cnt=len(pending))\n\n for idx, value in zip(pending, newvals):\n seq_values[idx].value = value\n\n@transaction._dsn_db\nasync def dmerge(current, origin=None, _dsn_db=None):\n \"\"\"\n Merge the current change of object into the origin.\n\n \"\"\"\n if current is None and origin is None:\n return\n\n if current is not None and not isinstance(current, DSetBase):\n if not isinstance(current, dobject):\n err = 'The current object should be dobject or dset type: %s'\n err %= current.__class__.__name__\n raise TypeError(err)\n\n dos = dset(current.__class__)()\n dos._add(current)\n current = dos\n\n if origin is not None and not isinstance(origin, DSetBase):\n if not isinstance(origin, dobject):\n err = 'The origin object should be dobject or dset type: %s'\n err %= origin.__class__.__name__\n raise TypeError(err)\n\n dos = dset(origin.__class__)()\n if origin:\n dos._add(origin)\n\n origin = dos\n\n if current is None:\n current = origin.__class__()\n\n if origin is None:\n origin = current.__class__()\n\n await pq_dtable_merge(current, origin, _dsn_db=_dsn_db)\n","repo_name":"lcgong/domainics","sub_path":"domainics/asyncdb/dmerge.py","file_name":"dmerge.py","file_ext":"py","file_size_in_byte":8174,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"68"}
+{"seq_id":"26488803056","text":"import numpy as np\nimport torch\nfrom torch.autograd import Function\n\n\ndef _process_params(new_params):\n if hasattr(new_params, 'cpu'):\n new_params = new_params.cpu().detach().numpy()\n return new_params\n\n\nclass QuantumFunction(Function):\n \"\"\"\n The class is for constructing the Pytorch Quantum model with our own quantum circuit.\n \"\"\"\n\n @staticmethod\n def forward(ctx, params: torch.Tensor, expval_fn, state):\n ctx.torch_device = None\n if params.is_cuda:\n ctx.torch_device = params.get_device()\n\n _params = _process_params(params)\n\n expval_fn.update(_params)\n expval = torch.as_tensor(expval_fn.forward(state), device=ctx.torch_device)\n\n ctx.fn = expval_fn\n ctx.state = state\n ctx.params = _params\n return expval\n\n @staticmethod\n def backward(ctx, grad_output):\n expval_fn = ctx.fn\n _params = ctx.params\n state = ctx.state\n\n expval_fn.update(_params)\n _, gradients = expval_fn.backward(state)\n g_params = grad_output * torch.as_tensor(gradients, device=ctx.torch_device)\n return g_params, None, None\n\n\nclass QuantumModel(torch.nn.Module):\n def __init__(self,\n expvalcost):\n super().__init__()\n init_params = expvalcost.params\n if isinstance(init_params, np.ndarray):\n init_params = torch.from_numpy(init_params).requires_grad_(True)\n else:\n init_params = init_params.clone().detach().requires_grad_(True)\n\n self.params = torch.nn.Parameter(init_params)\n expvalcost.update(self.params)\n self.quantum_fn = expvalcost\n\n def forward(self, state=None, circuit=None):\n if circuit is not None:\n self.quantum_fn.circuit = circuit\n if self.quantum_fn.backend_mode == 'spinq':\n\n if state is not None and len(state.size()) > 1:\n loss = torch.zeros(state.size(0))\n for i in range(state.size(0)):\n loss[i] += QuantumFunction.apply(self.params, self.quantum_fn, state[i, :])\n else:\n loss = QuantumFunction.apply(self.params, self.quantum_fn, state)\n return loss\n else:\n return self.quantum_fn.forward(state)","repo_name":"SpinQTech/SpinQit","sub_path":"spinqit/algorithm/torch_interface.py","file_name":"torch_interface.py","file_ext":"py","file_size_in_byte":2282,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"68"}
+{"seq_id":"29070624579","text":"# SuperFastPython.com\n# example of using a pipe between processes\nfrom time import sleep\nfrom random import random\nfrom multiprocessing import Process\nfrom multiprocessing import Pipe\n\n# custom function generate work items (sender)\ndef sender(connection):\n print('Sender: Running', flush=True)\n # generate work\n for _ in range(10):\n # generate a value\n value = random()\n # block\n sleep(value)\n # send data\n connection.send(value)\n # all done, signal to expect no further messages\n connection.send(None)\n print('Sender: Done', flush=True)\n\n# custom function to consume work items (receiver)\ndef receiver(connection):\n print('Receiver: Running', flush=True)\n # consume work\n while True:\n # get a unit of work\n item = connection.recv()\n # report\n print(f'>receiver got {item}', flush=True)\n # check for stop\n if item is None:\n break\n # all done\n print('Receiver: Done', flush=True)\n\n# protect the entry point\nif __name__ == '__main__':\n # create the pipe\n conn1, conn2 = Pipe()\n # start the sender\n sender_p = Process(target=sender, args=(conn2,))\n sender_p.start()\n # start the receiver\n receiver_p = Process(target=receiver, args=(conn1,))\n receiver_p.start()\n # wait for all processes to finish\n sender_p.join()\n receiver_p.join()\n","repo_name":"SuperFastPython/PythonMultiprocessingJumpStart","sub_path":"src/lesson05_pipe.py","file_name":"lesson05_pipe.py","file_ext":"py","file_size_in_byte":1392,"program_lang":"python","lang":"en","doc_type":"code","stars":18,"dataset":"github-code","pt":"68"}
+{"seq_id":"40095810227","text":"import hangarEvents as he\nimport pyautogui as py\n\ninventoryFull = 'Inv_Full.png'\ninventoryPicture = 'Inventory.png'\nSearch = 'INV_Search.png'\nDroneBay = 'Scripts/Screen/Images/Ship/Inventory/Inventory_DroneBay.png'\nItemHangar = 'Scripts/Screen/Images/Ship/Inventory/Inventory_ItemHangar.png'\nMiningHold = 'Scripts/Screen/Images/Ship/Inventory/Inventory_MiningHold.png'\nShipHangar = 'Scripts/Screen/Images/Ship/Inventory/Inventory_ShipHangar.png'\n\n\nclass Inventory:\n isFull = True\n isOpen = False\n position = ''\n\n def __init__(self):\n self.isFull = True\n self.isOpen = True\n self.position = None\n\n def openInventory(self):\n if not self.isOpen:\n py.keyDown('alt')\n py.press('c')\n py.keyUp('alt')\n\n def isFulled(self):\n self.isFull = he.isFoundInAnArea(inventoryFull)\n\n def isOpened(self):\n self.isOpen = he.isFoundInAnArea(inventoryPicture)\n\n def positionOf(self):\n self.position = he.findInAnAreaPos(inventoryPicture)\n\n def checkInventory(self):\n self.positionOf()\n self.isFulled()\n self.positionOf()\n\n\n\n\n","repo_name":"AnarShia/EveOnlineBot","sub_path":"Scripts/Images/Ship/Inventory/InventoryEvents.py","file_name":"InventoryEvents.py","file_ext":"py","file_size_in_byte":1134,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"68"}
+{"seq_id":"24745497385","text":"import pygame\nfrom pygame.sprite import Sprite\n\n\nclass Alien(Sprite):\n \"\"\"A class to represent a single alien in the fleet.\"\"\"\n\n def __init__(self, ai_settings, screen, type_alien):\n \"\"\"Initialize the alien and set its starting position.\"\"\"\n super(Alien, self).__init__()\n self.screen = screen\n self.ai_settings = ai_settings\n self.type = type_alien\n self.alien_images = []\n self.load_alien_images()\n self.index = 0\n self.image = self.alien_images[self.index] # Default\n self.frame_counter = 0\n self.will_fire = False\n self.rect = self.alien_images[0].get_rect()\n\n # Start each new alien near the top left of the screen.\n self.rect.x = self.rect.width\n self.rect.y = self.rect.height\n\n # Store the alien's exact position\n self.x = float(self.rect.x)\n\n def blitme(self):\n \"\"\"Draw the alien at its current location.\"\"\"\n self.screen.blit(self.image, self.rect)\n\n def check_edges(self):\n \"\"\"Return True if alien is at edge of screen.\"\"\"\n screen_rect = self.screen.get_rect()\n if self.rect.right >= screen_rect.right:\n return True\n elif self.rect.left <= 0:\n return True\n\n def update(self):\n \"\"\"Move the alien right or left.\"\"\"\n self.frame_counter += 1\n if self.frame_counter <= 120:\n self.index = 0\n elif self.frame_counter <= 240:\n self.index = 1\n else:\n self.frame_counter = 0\n # Change picture for animation\n if self.type != 4:\n self.image = self.alien_images[self.index]\n self.x += (self.ai_settings.alien_speed_factor * self.ai_settings.fleet_direction)\n else:\n self.x += self.ai_settings.ufo_speed_factor\n self.rect.x = self.x\n\n def load_alien_images(self):\n if self.type == 1:\n self.alien_images.append(pygame.image.load('images/Alien1.1.bmp'))\n self.alien_images.append(pygame.image.load('images/Alien1.2.bmp'))\n elif self.type == 2:\n self.alien_images.append(pygame.image.load('images/Alien2.1.bmp'))\n self.alien_images.append(pygame.image.load('images/Alien2.2.bmp'))\n elif self.type == 3:\n self.alien_images.append(pygame.image.load('images/Alien3.1.bmp'))\n self.alien_images.append(pygame.image.load('images/Alien3.2.bmp'))\n elif self.type == 4:\n self.alien_images.append(pygame.image.load('images/Alien4.bmp'))\n","repo_name":"beaubadilla/cpsc386_intro_game_design","sub_path":"alien-invaders/AlienInvaders/alien.py","file_name":"alien.py","file_ext":"py","file_size_in_byte":2562,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"36482658775","text":"# Definition for a binary tree node.\nclass TreeNode:\n def __init__(self, val=0, left=None, right=None):\n self.val = val\n self.left = left\n self.right = right\n\n\nclass Solution:\n def recoverTree(self, root: TreeNode) -> None:\n \"\"\"\n Do not return anything, modify root in-place instead.\n \"\"\"\n if not root:\n return None\n\n box = {}\n\n def check(node: TreeNode):\n if \"prev\" in box and node.val < box[\"prev\"].val:\n if \"big\" not in box:\n box[\"big\"] = box[\"prev\"]\n box[\"prev\"] = node\n box[\"small_cand\"] = node\n else:\n box[\"small\"] = node\n else:\n box[\"prev\"] = node\n \n\n def preorder(node: TreeNode):\n if not node or \"small\" in box:\n return\n preorder(node.left)\n check(node)\n preorder(node.right)\n\n preorder(root)\n if \"small\" not in box:\n if \"big\" not in box:\n raise Exception(\"invalid input\")\n else:\n box[\"small\"] = box[\"small_cand\"]\n\n box[\"big\"].val ^= box[\"small\"].val\n box[\"small\"].val ^= box[\"big\"].val\n box[\"big\"].val ^= box[\"small\"].val\n\n return root\n\nfn = Solution().recoverTree\n","repo_name":"kapppa-joe/leetcode-practice","sub_path":"daily/20201031_recover_bst.py","file_name":"20201031_recover_bst.py","file_ext":"py","file_size_in_byte":1373,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"17245161436","text":"import re\nfrom complex_domain.scrap_news.domain.entities.urls import Url, UrlCollection, TargetUrl\n\ndef test_criacao_url_com_erro():\n url = Url(\"I'm not an URL. \")\n \n assert url.valid == False\n\n\ndef test_url_protocol():\n url = Url(\"https://www.domain.com\")\n assert url.protocol.protocol == 'https' \n\ndef test_default_protocol():\n url = Url(\"www.noprotocol.com\")\n assert url.protocol.protocol == 'http'\n \ndef test_url_domain():\n url = Url(\"https://www.domain.com\")\n url2 = Url(\"www.domain.com\")\n assert url.domain.domain == 'domain.com' \n assert url2.domain.domain == 'domain.com'\n\ndef test_equals_url():\n url = Url(\"https://www1.domain1.com\")\n url1 = Url(\"https://www1.domain1.com\")\n url2 = Url(\"http://www1.domain1.com\")\n url3 = Url(\"http://br.domain1.com\")\n\n targets = [Url(\"http://www.folha.uol.com.br\"), Url(\"https://www.folha.uol.com.br\"), Url(\"http://www1.folha.uol.com.br\")]\n folha2 = Url(\"https://www.folha.uol.com.br\")\n folha3 = Url(\"https://top-of-mind.folha.uol.com.br/2021/\")\n folha4 = Url(\"http://transparencia.folha.uol.com.br \")\n\n assert url == url1\n assert url1 != url2\n assert url1 != url3\n assert folha2 in targets\n assert folha3 not in targets\n assert folha4 not in targets\n\ndef test_if_a_domain_is_contained_in_a_url():\n url1 = Url(\"https://www.domain1.com\")\n domain = url1.domain\n\n url2 = Url(\"https://subdomain.domain1.com\")\n url3 = Url(\"https://subdomain.domain2.com\")\n\n assert url1.contains(domain) == True\n assert url2.contains(domain) == True\n assert url3.contains(domain) == False\n\ndef test_url_collection():\n url1 = Url(\"http://www.domain.com\")\n url2 = Url(\"http://www.domain1.com\")\n url_collection1 = UrlCollection([url1, url2])\n url_collection2 = UrlCollection([url1])\n\n url = url_collection2.exclude(url_collection1)[0]\n\n assert url == url2\n\ndef test_real_url():\n url1 = Url(\"https://www1.folha.uol.com.br/credibilidade/folha-no-projeto-credibilidade.shtml\")\n\n assert url1.ignored == False\n assert url1.contains(domain=TargetUrl('folha.uol.com.br').domain) == True\n\ndef test_apply_rules():\n url_str = \"http://www1.folha.uol.com.br/livrariadafolha/2016/10/1580436-receitas-para-dormir-bem-sugere-como-ter-uma-noite-tranquila.shtml\"\n url1 = Url(url_str)\n pattern = r'.+\\/2016\\/\\d{2}[\\/0-9\\-a-z]+\\.shtml$'\n assert re.match(pattern, url_str) != None\n assert url1.is_accepted({'pattern': pattern})\n\n","repo_name":"luisfcaldeira/web_crawler_with_beatiful_soup","sub_path":"tests/domain/entities/test_urls.py","file_name":"test_urls.py","file_ext":"py","file_size_in_byte":2466,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"19594747254","text":"from django.urls import path\nfrom .views import InboxList, DeleteInboxThread, CreateInboxForm, Message\nfrom .views import CreateMessage\n\nurlpatterns = [\n path('chats/', InboxList.as_view(), name='inbox'),\n path('chats/delete/', DeleteInboxThread.as_view(),\n name='delete-thread'),\n path('chats/new-thread/', CreateInboxForm.as_view(), name='new-thread'),\n path('chats//', Message.as_view(), name='message'),\n path('chats//send/', CreateMessage.as_view(), name='send'),\n\n]\n","repo_name":"WisamTa/portfolio-project-4","sub_path":"priv_message/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":519,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"68"}
+{"seq_id":"804777202","text":"import os\r\nimport sqlalchemy\r\nimport config\r\n# import pandas as pd\r\nimport json\r\nfrom flask import jsonify\r\n\r\n# Sample use\r\n# import postgresql\r\n# postgres_obj = postgresql.PostgreSQL()\r\nclass PostgreSQL:\r\n def __init__(self):\r\n self.con, self.meta = self._connect()\r\n self.food_calories_table = self.meta.tables['food_calories']\r\n self.activity_calories_table = self.meta.tables['activity_calories']\r\n self.users_table = self.meta.tables['users']\r\n\r\n @staticmethod\r\n def _connect():\r\n # url = config.Config.DATABASE_URL\r\n url = 'postgresql://eatfitadmin@eatfit-db:Lerpekadut_82@eatfit-db.postgres.database.azure.com/eatfit'\r\n con = sqlalchemy.create_engine(url, client_encoding='utf8')\r\n meta = sqlalchemy.MetaData(bind=con, reflect=True)\r\n\r\n return con, meta\r\n\r\n # Insert\r\n\r\n # Sample use\r\n # postgres_obj.insert_food_calory({\r\n # 'food_name' : 'Nasi Goreng',\r\n # 'calory_amount' : 250\r\n # })\r\n def insert_user(self, user_dictionary):\r\n clause = self.users_table.insert().values(user_dictionary)\r\n self.con.execute(clause)\r\n user_json = json.dumps(user_dictionary)\r\n return jsonify (\r\n msg='Signup succesful',\r\n user=user_json\r\n )\r\n\r\n # Sample use\r\n # postgres_obj.insert_food_calory({\r\n # 'username' : 'geraldzakwan',\r\n # 'email' : 'geraldi.dzakwan@gmail.com',\r\n # 'password' : 'lerpekadutanjing',\r\n # })\r\n def insert_food_calory(self, food_calory_dictionary):\r\n clause = self.food_calories_table.insert().values(food_calory_dictionary)\r\n self.con.execute(clause)\r\n\r\n # Sample use\r\n # postgres_obj.batch_insert_food_calories('List Makanan.csv')\r\n # Csv header : no, food_meal, calory_amount\r\n # def batch_insert_food_calories(self, csv_path):\r\n # csv_dataframe = pd.read_csv(csv_path)\r\n #\r\n # for index, series in csv_dataframe.iterrows():\r\n # food_calory_dict = {}\r\n # for elem in series.iteritems():\r\n # if(elem[0] != 'no'):\r\n # food_calory_dict[elem[0]] = elem[1]\r\n # self.insert_food_calory(food_calory_dict)\r\n\r\n # Login sample use\r\n # postgres_obj.authenticate {\r\n # 'username' : 'geraldzakwan'\r\n # 'password' : 'lerpekadutanjing'\r\n # }\r\n def authenticate(self, login_dictionary):\r\n # if('username' in login_dictionary):\r\n # clause = self.users_table.select().where(\r\n # self.users_table.c.username == login_dictionary['username']\r\n # )\r\n # user = self.con.execute(clause).fetchone()\r\n # elif('email' in login_dictionary):\r\n # clause = self.users_table.select().where(\r\n # self.users_table.c.email == login_dictionary['email']\r\n # )\r\n # user = self.con.execute(clause).fetchone()\r\n # else:\r\n # return 'Authentication error - missing username/email'\r\n #\r\n # user_json = json.dumps(user.items())\r\n # if(user['password'] == login_dictionary['password']):\r\n # return jsonify(\r\n # msg='Authentication succesful',\r\n # user=user_json\r\n # )\r\n # else:\r\n # return jsonify(\r\n # msg='Authentication failed - wrong password'\r\n # )\r\n return jsonify(\r\n msg='Authentication failed - wrong password'\r\n )\r\n","repo_name":"zhorifiandi/Eatfit-Backend","sub_path":"postgresql.py","file_name":"postgresql.py","file_ext":"py","file_size_in_byte":3489,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"937117011","text":"import sys\r\nimport random\r\nimport string\r\nfrom time import sleep\r\nfrom selenium import webdriver\r\nfrom selenium.common.exceptions import NoSuchElementException\r\nfrom selenium.webdriver.common.action_chains import ActionChains\r\nfrom selenium.webdriver.support.ui import Select\r\n\r\n\r\nclass GeneralFunctions:\r\n # setting up the webdriver\r\n # param: website (str) pass the webadress e.g.: https://twitch.com\r\n @staticmethod\r\n def openWindow(driver, website, streamer=''):\r\n driver.get(website + '/' + streamer)\r\n driver.maximize_window()\r\n sleep(3)\r\n return\r\n\r\n # finds element with its tag-name and \"data-test-selector\"\r\n # param:\r\n # + elementType (str) e.g.: 'button', 'textarea', usw\r\n # + elementAttribute (str) pass the 'data-a-target' e.g.: 'chat-send-button', 'chat-send-button'\r\n @staticmethod\r\n def findElementByTagName(driver, elementType, attributeValue, attribute):\r\n wantedElements = driver.find_elements_by_tag_name(elementType)\r\n for idx in wantedElements:\r\n if idx.get_attribute(attribute) == attributeValue:\r\n return idx\r\n return None\r\n\r\n # finds elements with its tag-name and \"data-test-selector\"\r\n # param:\r\n # + elementType (str) e.g.: 'button', 'textarea', usw\r\n # + elementAttribute (str) pass the 'data-a-target' e.g.: 'chat-send-button', 'chat-send-button'\r\n @staticmethod\r\n def findElementsByTagName(driver, elementType, attributeValue, attribute):\r\n allElements = driver.find_elements_by_tag_name(elementType)\r\n wantedElements = []\r\n for idx in allElements:\r\n if idx.get_attribute(attribute) == attributeValue:\r\n wantedElements.append(idx)\r\n return wantedElements\r\n\r\n # find the code in the email\r\n # param: elementType (str) e.g.: 'p'\r\n @staticmethod\r\n def findCode(driver, elementType):\r\n wantedElement = None\r\n allOfType = driver.find_elements_by_tag_name(elementType)\r\n for element in allOfType:\r\n if len(element.text) == 6:\r\n wantedElement = element\r\n if wantedElement is None:\r\n print(\"Code not found\")\r\n return None\r\n return wantedElement\r\n\r\n # clicks the wanted button\r\n # param: buttonAttribute (str) pass the 'data-a-target' e.g.: 'chat-send-button'\r\n @staticmethod\r\n def clickButton(driver, buttonAttribute, elementType='button', attribute='data-a-target'):\r\n element = GeneralFunctions.findElementByTagName(driver, elementType, buttonAttribute, attribute)\r\n if element is None:\r\n print(\"Button not found\")\r\n return\r\n element.click()\r\n\r\n # passes the wanted text to the textbox\r\n # param:\r\n # + areaAttribute (str)\r\n # + text (str) pass the 'data-a-target' of the textbox e.g.: 'chat-input'\r\n @staticmethod\r\n def insertText(driver, attributeValue, text, elementType='textarea', attribute='data-a-target'):\r\n element = GeneralFunctions.findElementByTagName(driver, elementType, attributeValue, attribute)\r\n if element is None:\r\n print(\"Textbox not found\")\r\n return\r\n element.send_keys(text)\r\n\r\n # generate 3 digit random number\r\n @staticmethod\r\n def randomNumber(minDigits=1, maxDigits=4, lower=0, upper=9):\r\n number = ''\r\n for i in range(minDigits, maxDigits):\r\n number += str(random.randrange(lower, upper))\r\n return number\r\n\r\n # generate username\r\n @staticmethod\r\n def generateUsername(firstName, lastName):\r\n letters = string.ascii_lowercase\r\n username = firstName + lastName\r\n for i in range(3):\r\n username += str(random.choice(letters))\r\n username += GeneralFunctions.randomNumber()\r\n return username\r\n\r\n # generate 8 digit password\r\n @staticmethod\r\n def generatePassword(pwLenght=8):\r\n letters = string.ascii_lowercase\r\n letters += string.ascii_uppercase\r\n specialChar = [\"!\", \"$\", \"#\", \"+\", \"%\"]\r\n pwstring = ''.join(random.choice(letters) for i in range(pwLenght - 1))\r\n return pwstring + random.choice(specialChar)\r\n\r\n # safe userdata to .txt file\r\n @staticmethod\r\n def storeUserdata(self, file='userdata.txt'):\r\n userdata = open(file, \"w\")\r\n userdata.write(self.firstName + ', ')\r\n userdata.write(self.lastName + ', ')\r\n userdata.write(self.username + '@protonmail.com, ')\r\n userdata.write(self.password + ', ')\r\n userdata.write(self.birthday + ', ')\r\n userdata.write(self.birthmonth + ', ')\r\n userdata.write(self.birthyear + '\\n')\r\n userdata.close()\r\n\r\n # get last email\r\n @staticmethod\r\n def getLastUser():\r\n userdatafile = open(\"userdata.txt\", \"r\")\r\n userdata = userdatafile.readlines()\r\n userdatafile.close()\r\n lastUserdata = userdata[-1].split(', ')\r\n lastUserEmail = lastUserdata[2]\r\n lastUserPw = lastUserdata[3]\r\n return lastUserEmail, lastUserPw\r\n\r\n\r\n","repo_name":"AndreasOA/TwitchModMenu","sub_path":"python/GeneralFunctions.py","file_name":"GeneralFunctions.py","file_ext":"py","file_size_in_byte":5099,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"8582806","text":"import serial # Serial library use to connect python3 and arduino\nfrom serial import Serial\nimport pyautogui # pyautogui library will be use to perform arrow function\n\narduino = serial.Serial('/dev/ttyUSB0', 9600)\n\ni = 0\n\nwhile True:\n incoming_Data = arduino.readline() #readline() function will get data from serial monitor line by line\n\n if 'up' in incoming_Data.decode('utf-8'): \n \tpyautogui.press('up')\n \tincoming_Data = \"\"\n \tcontinue\n elif 'down' in incoming_Data.decode('utf-8'):\n \tpyautogui.keyDown('down')\n \tincoming_Data = \"\"\n \ti = 1\n \tcontinue\n\n if i == 1:\n \tpyautogui.press('up')\n\n i = 0\n\n incoming_Data = \"\"\n","repo_name":"gabrielgeier/arduino-chrome-dinossaur","sub_path":"script.py","file_name":"script.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"41905427026","text":"##parameters=post_id=None, REQUEST=None\n\n# $Id$\n\n\"\"\"Default view for the forum. If post_id is not None, post with id = \nwill be displayed, otherwise first post will be displayed.\"\"\"\n\nif post_id is None:\n threads = context.getContent().getThreads(proxy=context)\n if len(threads) > 0:\n return context.forum_view_main(post_id=threads[0]['id'])\n\nreturn context.forum_view_main()\n","repo_name":"nuxeo-cps/products--CPSForum","sub_path":"skins/forum_default/forum_view.py","file_name":"forum_view.py","file_ext":"py","file_size_in_byte":397,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"73605629018","text":"import keras.backend as K\nimport tensorflow.compat.v1 as tf\n\ntf.disable_v2_behavior()\nimport numpy as np\n\n\nclass AI:\n def __init__(self, center, radius, errors, same_shape):\n self.c = center\n self.r = radius\n if not same_shape:\n self.e = errors if errors is not None and errors.shape[-1] != 0 else None\n if self.e is not None:\n num_e = self.e.shape[-1] // self.c.shape[-1]\n mid = []\n for x in self.c.shape[1:]:\n mid.append(int(x))\n self.e = K.reshape(self.e, [-1] + mid + [num_e]) # batch, c[1], ..., c[-1], num_e\n input_perm = [len(self.e.shape) - 1] + list(np.arange(len(self.e.shape) - 1))\n self.e = tf.transpose(self.e, perm=input_perm) # num_e, batch, c[1], ..., c[-1]\n else:\n self.e = errors\n\n # print(\"init\", self.c.shape, self.r.shape, None if self.e is None else self.e.shape)\n\n def matmul(self, W):\n # with tf.name_scope(\"matmal\"):\n self.c = K.dot(self.c, W)\n self.r = K.dot(self.r, K.abs(W))\n self.e = None if self.e is None else K.dot(self.e, W)\n return self\n\n def bias_add(self, b):\n self.c = K.bias_add(self.c, b)\n return self\n\n def __add__(self, other):\n # print(\"add\")\n if isinstance(other, AI):\n if self.e is None:\n return AI(self.c + other.c, self.r + other.r, other.e, True)\n elif other.e is None:\n return AI(self.c + other.c, self.r + other.r, self.e, True)\n else:\n return AI(self.c + other.c, self.r + other.r, K.concatenate([self.e, other.e], axis=0), True)\n else:\n return AI(self.c + other, self.r, self.e, True)\n\n def __mul__(self, other):\n # print(\"mul\")\n if isinstance(other, AI):\n # raise NotImplementedError()\n lower_1, upper_1 = self.get_lu()\n lower_2, upper_2 = other.get_lu()\n all = K.stack([lower_1 * lower_2, lower_1 * upper_2, upper_1 * lower_2, upper_1 * upper_2], axis=0)\n lower = K.min(all, axis=0)\n upper = K.max(all, axis=0)\n return AI((lower + upper) / 2, (upper - lower) / 2, None, True)\n else:\n return AI(self.c * other, self.r * abs(other), None if self.e is None else self.e * other, True)\n\n def __truediv__(self, k):\n k = 1.0 / k\n return self * k\n\n def __sub__(self, other):\n return self + other * (-1)\n\n def get_lu(self):\n if self.e is None:\n return self.c - self.r, self.c + self.r\n else:\n return self.c - self.r - K.sum(K.abs(self.e), axis=0), self.c + self.r + K.sum(K.abs(self.e), axis=0)\n\n def activation(self, act):\n lower, upper = self.get_lu()\n lower = act(lower)\n upper = act(upper)\n self.c = (lower + upper) / 2\n self.r = (upper - lower) / 2\n self.e = None\n return self\n\n # def squeeze(self, axis):\n # self.c = K.squeeze(self.c, axis=axis)\n # self.r = K.squeeze(self.r, axis=axis)\n # if self.e is not None:\n # self.e = K.squeeze(self.e, axis=axis if axis < 0 else axis + 1)\n # return self\n\n def portion_sum(self, other):\n lower_1, upper_1 = self.get_lu()\n lower_2, upper_2 = other.get_lu()\n all = K.stack([lower_1 * lower_2, lower_1 * upper_2, upper_1 * lower_2, upper_1 * upper_2], axis=0)\n lower = K.sum(K.min(all, axis=0), axis=1)\n upper = K.sum(K.max(all, axis=0), axis=1)\n self.c = (lower + upper) / 2\n self.r = (upper - lower) / 2\n self.e = None\n return self\n\n def softmax(self, axis):\n lower, upper = self.get_lu()\n max_upper = K.max(upper, axis=axis, keepdims=True)\n max_lower = K.max(lower, axis=axis, keepdims=True)\n new_lower = K.exp(lower - max_upper) / K.sum(K.exp(upper - max_upper), axis=axis, keepdims=True)\n new_upper = K.minimum(1.0, K.exp(upper - max_lower) / K.sum(K.exp(lower - max_lower), axis=axis, keepdims=True))\n self.c = (new_lower + new_upper) / 2\n self.r = (new_upper - new_lower) / 2\n self.e = None\n return self\n\n def to_state(self):\n if self.e is None:\n # print(\"to_state\", self.c.shape, self.r.shape)\n return K.concatenate([self.c, self.r], axis=-1)\n else:\n # num_e, batch, c[1], ..., c[-1]\n input_perm = list(np.arange(1, len(self.e.shape))) + [0]\n errors = tf.transpose(self.e, perm=input_perm) # batch, c[1], ..., c[-1], num_e\n num_e = errors.shape[-1]\n mid = []\n for x in self.c.shape[1:-1]:\n mid.append(int(x))\n # print(\"to_state\", self.c.shape, self.e.shape)\n return K.concatenate(\n [self.c, self.r, K.reshape(errors, [-1] + mid + [num_e * self.c.shape[-1]])],\n axis=-1) # batch, c[1], ..., c[-1]*num_e\n\n def GRU_merge(self, self_act, a, b, act):\n lower_a, upper_a = a.get_lu()\n lower_b, upper_b = b.get_lu()\n fa_lower, fa_upper = lower_a, upper_a\n fb_lower, fb_upper = act(lower_b), act(upper_b)\n lower_x, upper_x = self.get_lu()\n fx_lower, fx_upper = self_act(lower_x), self_act(upper_x)\n partial_fx_lower = tf.gradients(fx_lower, lower_x)[0]\n partial_fx_upper = tf.gradients(fx_upper, upper_x)[0]\n\n def lower_a_greater_zero():\n uz_x_Phi = K.minimum(partial_fx_upper * fa_upper, (fx_upper - fx_lower) * fa_upper / (upper_x - lower_x))\n ax_right_upper = fx_upper * fa_upper\n ax_left_upper = uz_x_Phi * (lower_x - upper_x) + ax_right_upper\n lz_x_Phi = K.minimum(partial_fx_lower * fa_lower, (fx_lower - fx_upper) * fa_lower / (lower_x - upper_x))\n ax_left_lower = fx_lower * fa_lower\n ax_right_lower = lz_x_Phi * (upper_x - lower_x) + ax_left_lower\n return [ax_left_lower, ax_left_upper, ax_right_lower, ax_right_upper]\n\n def lower_b_greater_zero():\n uz_x_Phi = K.maximum(-partial_fx_lower * fb_upper, (-fx_upper + fx_lower) * fb_upper / (upper_x - lower_x))\n bx_left_upper = (1 - fx_lower) * fb_upper\n bx_right_upper = uz_x_Phi * (upper_x - lower_x) + bx_left_upper\n lz_x_Phi = K.maximum(-partial_fx_upper * fb_lower, (-fx_lower + fx_upper) * fb_lower / (lower_x - upper_x))\n bx_right_lower = (1 - fx_upper) * fb_lower\n bx_left_lower = lz_x_Phi * (lower_x - upper_x) + bx_right_lower\n return [bx_left_lower, bx_left_upper, bx_right_lower, bx_right_upper]\n\n def upper_a_less_zero():\n uz_x_Phi = K.maximum(partial_fx_lower * fa_upper, (fx_lower - fx_upper) * fa_upper / (lower_x - upper_x))\n ax_left_upper = fx_lower * fa_upper\n ax_right_upper = uz_x_Phi * (upper_x - lower_x) + ax_left_upper\n lz_x_Phi = K.maximum(partial_fx_upper * fa_lower, (fx_upper - fx_lower) * fa_lower / (upper_x - lower_x))\n ax_right_lower = fx_upper * fa_lower\n ax_left_lower = lz_x_Phi * (lower_x - upper_x) + ax_right_lower\n return [ax_left_lower, ax_left_upper, ax_right_lower, ax_right_upper]\n\n def upper_b_less_zero():\n uz_x_Phi = K.minimum(-partial_fx_upper * fb_upper, (-fx_upper + fx_lower) * fb_upper / (upper_x - lower_x))\n bx_right_upper = (1 - fx_upper) * fb_upper\n bx_left_upper = uz_x_Phi * (lower_x - upper_x) + bx_right_upper\n lz_x_Phi = K.minimum(-partial_fx_lower * fb_lower, (-fx_lower + fx_upper) * fb_lower / (lower_x - upper_x))\n bx_left_lower = (1 - fx_lower) * fb_lower\n bx_right_lower = lz_x_Phi * (upper_x - lower_x) + bx_left_lower\n return [bx_left_lower, bx_left_upper, bx_right_lower, bx_right_upper]\n\n def otherwise_a():\n uz_x_Phi = K.minimum(partial_fx_upper * fa_upper, (fx_upper - fx_lower) * fa_upper / (upper_x - lower_x))\n ax_right_upper = fx_upper * fa_upper\n ax_left_upper = uz_x_Phi * (lower_x - upper_x) + ax_right_upper\n lz_x_Phi = K.maximum(partial_fx_upper * fa_lower, (fx_upper - fx_lower) * fa_lower / (upper_x - lower_x))\n ax_right_lower = fx_upper * fa_lower\n ax_left_lower = lz_x_Phi * (lower_x - upper_x) + ax_right_lower\n return [ax_left_lower, ax_left_upper, ax_right_lower, ax_right_upper]\n\n def otherwise_b():\n uz_x_Phi = K.maximum(-partial_fx_lower * fb_upper, (-fx_upper + fx_lower) * fb_upper / (upper_x - lower_x))\n bx_left_upper = (1 - fx_lower) * fb_upper\n bx_right_upper = uz_x_Phi * (upper_x - lower_x) + bx_left_upper\n lz_x_Phi = K.minimum(-partial_fx_lower * fb_lower, (-fx_lower + fx_upper) * fb_lower / (lower_x - upper_x))\n bx_left_lower = (1 - fx_lower) * fb_lower\n bx_right_lower = lz_x_Phi * (upper_x - lower_x) + bx_left_lower\n return [bx_left_lower, bx_left_upper, bx_right_lower, bx_right_upper]\n\n a_anchors = otherwise_a()\n anchors_lower_a_greater_zero = lower_a_greater_zero()\n anchors_upper_a_less_zero = upper_a_less_zero()\n for i in range(4):\n a_anchors[i] = K.switch(K.greater(lower_a, K.zeros_like(lower_a)), anchors_lower_a_greater_zero[i],\n a_anchors[i])\n a_anchors[i] = K.switch(K.less(upper_a, K.zeros_like(upper_a)), anchors_upper_a_less_zero[i], a_anchors[i])\n\n b_anchors = otherwise_b()\n anchors_lower_b_greater_zero = lower_b_greater_zero()\n anchors_upper_b_less_zero = upper_b_less_zero()\n for i in range(4):\n b_anchors[i] = K.switch(K.greater(lower_b, K.zeros_like(lower_b)), anchors_lower_b_greater_zero[i],\n b_anchors[i])\n b_anchors[i] = K.switch(K.less(upper_b, K.zeros_like(upper_b)), anchors_upper_b_less_zero[i], b_anchors[i])\n\n for i in range(4):\n a_anchors[i] += b_anchors[i]\n lower_z = K.minimum(a_anchors[0], a_anchors[2])\n upper_z = K.maximum(a_anchors[1], a_anchors[3])\n return AI((lower_z + upper_z) / 2, (upper_z - lower_z) / 2, None, True)\n\n def GRU_merge1(self, self_act, a, b, act):\n lower_a, upper_a = a.get_lu()\n lower_b, upper_b = b.get_lu()\n fa_lower, fa_upper = lower_a, upper_a\n fb_lower, fb_upper = act(lower_b), act(upper_b)\n lower_x, upper_x = self.get_lu()\n fx_lower, fx_upper = self_act(lower_x), self_act(upper_x)\n\n def lower_a_greater_zero():\n return [fx_lower * fa_lower, fx_lower * fa_upper, fx_upper * fa_lower, fx_upper * fa_upper]\n\n def lower_b_greater_zero():\n return [(1 - fx_lower) * fb_lower, (1 - fx_lower) * fb_upper, (1 - fx_upper) * fb_lower, (\n 1 - fx_upper) * fb_upper]\n\n abounds = [fa_lower, fa_upper]\n bbounds = [fb_lower, fb_upper]\n xbounds = [fx_lower, fx_upper]\n rets = []\n for ab in abounds:\n for bb in bbounds:\n for xb in xbounds:\n rets.append(K.expand_dims(ab * xb + (1 - xb) * bb, axis=-1))\n\n rets = K.concatenate(rets, axis=-1)\n lower_z = K.min(rets, axis=-1)\n upper_z = K.max(rets, axis=-1)\n return AI((lower_z + upper_z) / 2, (upper_z - lower_z) / 2, None, True)\n\n # def mul(self, other, f1, f2, f1_dec=False):\n # # f1, f2 are two monotonic functions, and f1(x) > 0\n # # f1_dec is True means that f1 decreases with respect to x, otherwise increases\n # # f2 always increases with respect to y\n #\n # lower_1, upper_1 = self.get_lu()\n # if f1_dec:\n # f1_lower = f1(lower_1)\n # partial_f1_lower = tf.gradients(f1_lower, lower_1)\n # f1_upper = f1(upper_1)\n # partial_f1_upper = tf.gradients(f1_upper, upper_1)\n # else:\n # f1_lower = f1(upper_1)\n # partial_f1_lower = tf.gradients(f1_lower, upper_1)\n # f1_upper = f1(lower_1)\n # partial_f1_upper = tf.gradients(f1_upper, lower_1)\n # lower_2, upper_2 = other.get_lu()\n # f2_lower = f2(lower_2)\n # partial_f2_lower = tf.gradients(f2_lower, lower_2)\n # f2_upper = f2(upper_2)\n # partial_f2_upper = tf.gradients(f2_upper, upper_2)\n #\n # def lower_2_greater_zero():\n # if not f1_dec:\n # uz_x_Phi = K.minimum(partial_f1_upper * f2_upper,\n # (f1_upper - f1_lower) * f2_upper / (upper_1 - lower_1))\n # uz_x_ax = upper_1\n # uz_x_b = f1_upper * f2_upper\n # lz_x_Phi = K.minimum(partial_f1_lower * f2_lower,\n # (f1_lower - f1_upper) * f2_lower / (lower_1 - upper_1))\n # lz_x_ax = lower_1\n # lz_x_b = f1_lower * f2_lower\n # z_x_Phi = K.minimum(uz_x_Phi, lz_x_Phi)\n # volx = (uz_x_b - lz_x_b) * (upper_1 - lower_1) - (upper_1 - lower_1) * (upper_1 - lower_1) * z_x_Phi * 2\n # else:\n # uz_x_Phi = K.maximum(partial_f1_upper * f2_upper,\n # (f1_upper - f1_lower) * f2_upper / (lower_1 - upper_1))\n # uz_x_ax = lower_1\n # uz_x_b = f1_upper * f2_upper\n # lz_x_Phi = K.maximum(partial_f1_lower * f2_lower,\n # (f1_lower - f1_upper) * f2_lower / (upper_1 - lower_1))\n # lz_x_ax = upper_1\n # lz_x_b = f1_lower * f2_lower\n # z_x_Phi = K.maximum(uz_x_Phi, lz_x_Phi)\n # volx = (uz_x_b - lz_x_b) * (upper_1 - lower_1) + (upper_1 - lower_1) * (upper_1 - lower_1) * z_x_Phi * 2\n #\n # uz_y_Phi = K.minimum(partial_f2_upper * f1_upper, (f2_upper - f2_lower) * f1_upper / (upper_2 - lower_2))\n # uz_y_ay = upper_2\n # uz_y_b = f1_upper * f2_upper\n # lz_y_Phi = K.minimum(partial_f2_lower * f1_lower, (f2_lower - f2_upper) * f1_lower / (lower_2 - upper_2))\n # lz_y_ay = lower_2\n # lz_y_b = f1_lower * f2_lower\n # z_y_Phi = K.minimum(uz_y_Phi, lz_y_Phi)\n # voly = (uz_y_b - lz_y_b) * (upper_2 - lower_2) - (upper_2 - lower_2) * (upper_2 - lower_2) * z_y_Phi * 2\n # ret_x = K.concatenate([z_x_Phi, uz_x_b, ])\n #\n # def upper_2_less_zero():\n # pass\n #\n # def otherwise():\n # pass\n\n def __getitem__(self, item):\n errors = None\n if self.e is not None:\n # print(self.e.shape)\n errors = [K.expand_dims(self.e[i][item], axis=0) for i in range(int(self.e.shape[0]))]\n errors = K.concatenate(errors, axis=0)\n\n # print(\"getitem\")\n return AI(self.c[item], self.r[item], errors, True)\n","repo_name":"ForeverZyh/robust_rnn","sub_path":"AI.py","file_name":"AI.py","file_ext":"py","file_size_in_byte":15002,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"68"}
+{"seq_id":"43923260591","text":"__author__ = \"John McGehee (jmcgehee@altera.com)\"\n__revision__ = \"$Revision: #1 $\"\n__date__ = \"$Date: 2022/12/13 $\"\n__copyright__ = \"Copyright 2013 Altera Corporation\"\n\nimport datetime\nimport getpass\nfrom xml.etree.ElementTree import Comment, Element, SubElement\n\nfrom dmx.dmlib.templateset.alias import Alias\nfrom dmx.dmlib.templateset.template import Template\nfrom dmx.dmlib.templateset.successor import Successor\nfrom dmx.dmlib.templateset.xmlbase import XmlBase\nfrom dmx.dmlib.dmError import dmError\n\n\nclass Templateset(XmlBase):\n '''Construct a deliverable templateset of the specified name (id) and version.\n You can optionally specify:\n\n * `aliasNames`, The aliases to output. The default is to output all aliases.\n * `deliverableNames`, The deliverables to output. The default is to output all deliverables.\n '''\n\n #base __init__ not called: pylint: disable = W0231\n def __init__(self, id_, version, aliasNames=None, deliverableNames=None):\n self._id = id_ \n if not isinstance(version, basestring):\n raise dmError(\"Deliverable templateset version '{}' is not a string\".format(version))\n self._version = version \n self._date = datetime.date.today()\n self._author = getpass.getuser()\n\n if aliasNames is not None:\n self._aliasNames = aliasNames\n self._aliasNames.sort()\n else:\n # getNames() returns a sorted list\n self._aliasNames = Alias.getNames()\n\n if deliverableNames is not None:\n self._deliverableNames = deliverableNames\n self._deliverableNames.sort()\n else:\n self._deliverableNames = Template.getNames()\n\n @property\n def tagName(self):\n '''The tag name for this XML element.'''\n return 'templateset'\n \n @property\n def reportName(self):\n '''The natural language name for this object for use in reports and messages.'''\n return 'templateset'\n \n @property\n def deliverableNames(self):\n '''The list of deliverable names in this deliverable templateset.'''\n return self._deliverableNames\n \n def element(self, parent=None):\n '''Return an XML ElementTree representing this instance.\n\n If a `parent` Element is specified, make the ElementTree a SubElement of `parent`.\n However, since `` is the root element, it makes little sense to specify\n `parent`.\n \n >>> ts = Templateset('draft', '2.0')\n >>> ts.element().tag\n 'templateset'\n '''\n if parent:\n ts = SubElement(parent, self.tagName)\n else:\n ts = Element(self.tagName)\n ts.set('id', self._id)\n ts.set('version', self._version)\n ts.set('date', str(self._date))\n ts.set('author', self._author)\n titleComment = Comment(\"Altera Data Management templateset\\n\"\n \" This file is generated by the `templateset` program.\\n\"\n \" See http://sw-web2/tools.ice/icd_cad/dm/{}/doc/templateset.html\\n\"\n \"\\n\"\n \" Altera Confidential and Proprietary.\\n\"\n \" Copyright 2013-2014 Altera Corporation.\\n\".format(self._version))\n ts.append(titleComment)\n \n for deliverableName in self._deliverableNames:\n template = Template(deliverableName)\n template.element(ts)\n for deliverableName in self._deliverableNames:\n successor = Successor(deliverableName)\n successor.element(ts)\n for aliasName in self._aliasNames:\n alias = Alias(aliasName)\n alias.element(ts)\n return ts\n\n def report(self, ipName='ip1', cellName='ip1'):\n '''Return a human-readable report of this templateset.\n \n >>> ts = Templateset('draft', '2.0')\n >>> ts.report('ipName', 'cellName') #doctest: +ELLIPSIS +NORMALIZE_WHITESPACE\n \"Report on templateset ...\"\n '''\n ret = 'Report on {} {}, version {}\\n\\n'.format(self.reportName, self._id, self._version)\n \n # The method continues after this long explanation string\n ret +='''For other ways to visualize and access the templateset, see:\n http://sw-wiki.altera.com/twiki/bin/view/DesignAutomation/Templateset\n http://sw-wiki.altera.com/twiki/bin/view/DesignAutomation/WhereToPutDesignDataFiles\n \n'''\n for deliverableName in self._deliverableNames:\n template = Template(deliverableName)\n ret += template.report(ipName, cellName) + '\\n'\n \n for deliverableName in self._deliverableNames:\n successor = Successor(deliverableName)\n ret += successor.report() + '\\n'\n\n for aliasName in self._aliasNames:\n alias = Alias(aliasName)\n ret += alias.report() + '\\n'\n \n ret += '\\n\\n'\n return ret\n \n def reportReadiness(self):\n '''Return a report of how many deliverables in this templateset are ready.\n \n >>> ts = Templateset('draft', '2.0')\n >>> ts.reportReadiness() #doctest: +ELLIPSIS +NORMALIZE_WHITESPACE\n 'Progress report on templateset ...'\n '''\n ret = 'Progress report on templateset {}, version {}\\n\\n'.format(self._id, self._version)\n \n # The method continues after this long explanation string\n ret += 'This is the report of which deliverables are ready, and not ready.\\n'\n\n # Report aliases\n readyAliases = []\n unreadyAliases = []\n for aliasName in self._aliasNames:\n alias = Alias(aliasName)\n if alias.isReady:\n readyAliases.append(alias)\n else:\n unreadyAliases.append(alias)\n\n ret +='\\n{} design style aliases are ready for action:\\n'.format(len(readyAliases))\n ret += '-------------------------------------------------------------\\n'\n for alias in readyAliases:\n if alias.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(alias.id, alias.caseid )\n else:\n ret += '{}\\n'.format(alias.id)\n \n ret +='\\n{} design style aliases are NOT ready for action:\\n'.format(len(unreadyAliases))\n ret += '-----------------------------------------------------------------\\n'\n for alias in unreadyAliases:\n if alias.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(alias.id, alias.caseid )\n else:\n ret += '{}\\n'.format(alias.id)\n \n # Report deliverables\n readyTemplates = []\n unreadyTemplates = []\n for deliverableName in self._deliverableNames:\n template = Template(deliverableName)\n if template.isReady:\n readyTemplates.append(template)\n else:\n unreadyTemplates.append(template)\n\n ret += '\\n{} templates are ready for action:\\n'.format(len(readyTemplates))\n ret += '------------------------------------\\n'\n for template in readyTemplates:\n if template.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(template.id_, template.caseid )\n else:\n ret += '{}\\n'.format(template.id_)\n \n ret +='\\n{} templates are NOT ready for action:\\n'.format(len(unreadyTemplates))\n ret += '---------------------------------------\\n'\n for template in unreadyTemplates:\n if template.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(template.id_, template.caseid )\n else:\n ret += '{}\\n'.format(template.id_)\n \n ret += '\\n\\n'\n \n # Report successors\n readySuccessors = []\n unreadySuccessors = []\n for deliverableName in self._deliverableNames:\n successor = Successor(deliverableName)\n if successor.isReady:\n readySuccessors.append(successor)\n else:\n unreadySuccessors.append(successor)\n\n ret +='\\n{} predecessor-successor relationships are ready for action:\\n'.format(len(readySuccessors))\n ret += '-------------------------------------------------------------\\n'\n for successor in readySuccessors:\n if successor.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(successor.id, successor.caseid )\n else:\n ret += '{}\\n'.format(successor.id)\n\n ret +='\\n{} predecessor-successor relationships are NOT ready for action:\\n'.format(len(unreadySuccessors))\n ret += '-----------------------------------------------------------------\\n'\n for successor in unreadySuccessors:\n if successor.caseid :\n ret += '{}, Fogbugz case {}\\n'.format(successor.id, successor.caseid )\n else:\n ret += '{}\\n'.format(successor.id)\n\n return ret\n\n\nif __name__ == \"__main__\":\n # Running DeliverableFTemplateSet_test.py is the preferred test method,\n # but run doctest alone if the user requests.\n import doctest\n doctest.testmod()\n","repo_name":"lionelta/dmx_main_gdpxl_py23_cth","sub_path":"lib/python/dmx/dmlib/templateset/templateset.py","file_name":"templateset.py","file_ext":"py","file_size_in_byte":9183,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"2814365786","text":"import os\n\nimport ram\nimport Grind\n\n\njumps = {'JMP': '00C0',\n 'JZ': '00C1',\n 'JNZ': '00C2',\n 'JS': '00C3',\n 'JNS': '00C4',\n 'JO': '00C5',\n 'JNO': '00C6'}\ntable = {\n 'arithmetic': {\n 'ADD': ['00A0', '00B0'],\n 'SUB': ['00A1', '00B1'],\n 'MUL': ['00A2', '00B2'],\n 'DIV': ['00A3', '00B3'],\n 'MOD': ['00A6', '00B6'],\n 'XOR': ['00A7', '00B7']},\n 'INC': '00A4',\n 'DEC': '00A5',\n 'jumps': jumps,\n 'MOV': ['00D0', '00D1', '00D2', '00D3', '00D4'],\n 'CMP': ['00DC', '00DA', '00DB', '00DD'],\n 'PUSH': '00E0',\n 'POP': '00E1',\n 'PUSHG': '00EA',\n 'POPG': '00EB',\n 'OUT': ['00F0', '00F1', '00F2'],\n 'IN': ['00FF', '00FE', '00FD'],\n 'CALL': '0A00',\n 'RET': '0B00',\n 'SWP': '0C00'\n}\n\n\nregisters = {'AH': '0010',\n 'AL': '0001',\n 'AX': '0100',\n 'BH': '0020',\n 'BL': '0002',\n 'BX': '0200',\n 'CH': '0030',\n 'CL': '0003',\n 'CX': '0300',\n 'DH': '0040',\n 'DL': '0004',\n 'DX': '0400',\n 'ABX': '1000',\n 'BCX': '2000',\n 'CDX': '3000',\n 'DAX': '4000'\n }\n\ndata_table = {\n 'global': 0,\n 'private': 1\n}\n\nd = []\n\n\ndef data(line):\n t, var = line.split('>')\n var, d = var.split('=')\n if t == 'literal':\n d = [d]\n elif t == 'list':\n d = list(d[1:-1].split(','))\n elif t == 'string':\n d = [tohex(ord(v)) for v in d[1:-1]] + ['000A']\n return var, d\n\n\ndef tokenize(lines):\n mode = 0\n for line in lines:\n op, args = None, None\n line = line.strip('\\n')\n if '.data_' in line:\n mode = 1\n continue\n elif '.exec_' in line:\n mode = 2\n continue\n elif '%' in line:\n continue\n else:\n if mode == 1:\n d.append(data(line.strip('\\t').strip(' ')))\n continue\n elif mode == 2:\n if line.strip('\\t').strip(' ')[0] == ';':\n continue\n line = line.split(';')[0].strip('\\t').split('|')\n if ':' in line[0] and line[0] != '.':\n yield line[0].strip('\\t').strip(' '), []\n continue\n elif 'END' in line[0]:\n yield 'END', []\n continue\n elif line[0][0] == '.':\n yield line[0].strip('\\t').strip(' '), []\n continue\n line = line[0].split()\n op, args = line[0], line[1].split(',')\n yield op, args\n\n\ndef parse(lines):\n is_mem = lambda arg: True if arg[0] == '[' and arg[-1] == ']' else False\n for line in lines:\n op, args = line\n if args:\n if op == 'DB':\n if len(args[0]) == 4:\n op, args = args, []\n else:\n args = [tohex(ord(n)) for n in args[0][1:-1]]\n op, args = args[0], args[1:]\n elif op == 'CMP':\n if is_mem(args[1]):\n op = table[op][0]\n args[0] = registers[args[0]]\n args[1] = args[1][1:-1]\n elif args[1] in registers:\n op = table[op][1]\n args = [registers[arg] for arg in args]\n elif is_mem(args[0]):\n op = table[op][3]\n args[0] = registers[args[0][1:-1]]\n else:\n op = table[op][2]\n args[0] = registers[args[0]]\n elif op in table['arithmetic']:\n if args[1] in registers:\n op = table['arithmetic'][op][0]\n args = [registers[arg] for arg in args]\n else:\n op = table['arithmetic'][op][1]\n args[0] = registers[args[0]]\n elif op in table['jumps']:\n op = table['jumps'][op]\n elif len(args) == 1 and op in table:\n if op == 'OUT':\n if is_mem(args[0]):\n op = table[op][2]\n args[0] = registers[args[0][1:-1]]\n elif args[0] in registers:\n op = table[op][1]\n args[0] = registers[args[0]]\n else:\n op = table[op][0]\n elif op == 'IN':\n if args[0] in registers:\n op = table[op][0]\n args[0] = registers[args[0]]\n elif is_mem(args[0]):\n if args[0][1:-1] in registers:\n op = table[op][1]\n args[0] = registers[args[0][1:-1]]\n else:\n op = table[op][2]\n args[0] = args[0][1:-1]\n elif op in ['CALL', 'RET']:\n op, args = table[op], args\n elif args[0] in registers:\n op = table[op]\n args = [registers[args[0]]]\n else:\n op = table[op]\n elif op == 'MOV':\n op, args = mov(args)\n elif op == 'OUT':\n if args[0][0] == '[' and args[0][-1] == ']':\n op = table[op][2]\n args[0] = args[0][1:-1]\n elif args[0] in registers:\n op = table[op][1]\n args[0] = registers[args[0]]\n else:\n op = table[op][0]\n elif op == 'END':\n op, args = '0000', []\n yield op, args\n\n\ndef mov(args):\n if args[0] in registers:\n if len(args[1]) == 4 and args[1][1:-1] in registers:\n return '00D3', [registers[args[0]], registers[args[1][1:-1]]]\n elif len(args[1]) == 6:\n return '00D1', [registers[args[0]], args[1][1:-1]]\n else:\n return '00D0', [registers[args[0]], args[1]]\n else:\n if args[0][1:-1] in registers:\n return '00D4', [registers[args[0][1:-1]], registers[args[1]]]\n else:\n return '00D2', [args[0][1:-1], registers[args[1]]]\n\n\ndef flatten(foo):\n for x in foo:\n if hasattr(x, '__iter__') and not isinstance(x, str):\n for y in flatten(x):\n yield y\n else:\n yield x\n\n\ndef tohex(val):\n return hex(val & 0xFFFF)[2:].zfill(4).upper()\n\n\ndef twos_comp(val, bits):\n \"\"\"compute the 2's compliment of int value val\"\"\"\n if val & (1 << (bits - 1)) != 0:\n val -= (1 << bits)\n return val\n\n\ndef find_subs(program: list):\n sub_found = False\n subs = {}\n sub = []\n main = []\n name = ''\n for lineno in range(len(program)):\n if '.sub(' in program[lineno][0]:\n name = program[lineno][0][5:-2]\n sub = []\n sub_found = True\n sub.append((program[lineno][0][5:-2] + ':', []))\n elif sub_found and program[lineno][0] != '.endsub:':\n sub.append(program[lineno])\n elif program[lineno][0] == '.endsub:':\n subs[name] = sub\n sub_found = False\n else:\n main.append(program[lineno])\n return main, subs\n\n\ndef find_jumps(program):\n matches = {}\n l = program[:]\n for item in l:\n if item.endswith(':'):\n matches[item[:-1]] = l.index(item)\n l.remove(item)\n for i in range(len(l)):\n if l[i] in jumps.values():\n if l[i + 1] in matches.keys():\n print('Jump to', l[i + 1], end='')\n l[i + 1] = tohex(matches[l[i + 1]] - i)\n print(' =', twos_comp(int(l[i + 1], 16), 16))\n return l\n\n\ndef make_callable(program, calls):\n call = False\n out = []\n count = 0\n for i in range(len(program)):\n if program[i] == '0A00':\n call = True\n continue\n elif call:\n out.append(calls[program[i]])\n program[i] = out[-1]\n count += 1\n call = False\n return program, out\n\n\ndef load(program, fname):\n memory = ram.RAM()\n settings = None\n if fname + '.dripc' in os.listdir('.'):\n grind = Grind.Grind(file=fname + '.drip', config=fname + '.dripc')\n print(grind.subs, grind.switches)\n grind.sub()\n program = grind.out()\n if grind.switches:\n settings = grind.switches\n program = list(tokenize(program))\n var_table = {}\n for var in d:\n name, val = var\n var_table[name] = val\n program, subs = find_subs(program)\n subs2 = []\n sub_locs = {}\n sub_loc = 0\n for sub in subs:\n subs2.append(find_jumps(list(flatten(parse(subs[sub])))))\n sub_locs[sub] = tohex(sub_loc)\n sub_loc += len(subs2[-1])\n subs = subs2\n del subs2\n program, call_table, = make_callable(find_jumps(list(flatten(parse(program)))), sub_locs)\n sub_list = list(flatten(subs))\n program = sub_list + program\n base = int('FF00', 16)\n var_table_ref = {}\n address = int('FE00', 16)\n for varname in var_table:\n save = address\n var_table_ref[varname] = address\n print(varname, '-->', var_table[varname])\n for item in var_table[varname]:\n memory.put(tohex(address), item)\n address += 1\n address = save - 256\n final = program[:]\n for i in range(len(final)):\n if final[i][0] == '[':\n final[i] = final[i][1:-1]\n if final[i] in var_table_ref:\n final[i] = tohex(var_table_ref[final[i]])\n for i in range(len(final)):\n memory.put(tohex(i), final[i])\n for address, call in enumerate(call_table):\n memory.put(tohex(base + address), call)\n memory.show()\n return memory, len(list(flatten(subs)))\n","repo_name":"BytesAndCoffee/QuadShot","sub_path":"Drip.py","file_name":"Drip.py","file_ext":"py","file_size_in_byte":10011,"program_lang":"python","lang":"en","doc_type":"code","stars":14,"dataset":"github-code","pt":"68"}
+{"seq_id":"71635984217","text":"from data import DICTIONARY, LETTER_SCORES\n\ndef load_words():\n \"\"\"Load dictionary into a list and return list\"\"\"\n dictionary_file_obj = open(DICTIONARY,\"r\")\n dictionary_list = dictionary_file_obj.readlines()\n dictionary_list = [each_word[:-1] for each_word in dictionary_list]\n return dictionary_list\n\ndef calc_word_value(input_word):\n \"\"\"Calculate the value of the word entered into function\n using imported constant mapping LETTER_SCORES\"\"\"\n input_word_value = 0\n for each_letter in input_word:\n try:\n input_word_value += LETTER_SCORES[each_letter.upper()]\n except Exception as e:\n print(\"Exception Obtained in word: \" + \"'\" + input_word + \"'\" +\" and exception is \" + str(e))\n\n return input_word_value\n\ndef max_word_value(words_list = load_words()):\n \"\"\"Calculate the word with the max value, can receive a list\n of words as arg, if none provided uses default DICTIONARY\"\"\"\n max_value = 0\n max_word = \"\"\n for each_word in words_list:\n curr_word_value = calc_word_value(each_word)\n if curr_word_value > max_value:\n max_value = curr_word_value\n max_word = each_word\n return max_word\n\nif __name__ == \"__main__\":\n dictionary_list = load_words()\n max_word = max_word_value(('bob', 'julian', 'pybites', 'quit', 'barbeque'))\n print(\"Word with Maximum Value : \" + max_word)\n\n ","repo_name":"pybites/challenges","sub_path":"01/jatinsinghal1921/wordvalue.py","file_name":"wordvalue.py","file_ext":"py","file_size_in_byte":1404,"program_lang":"python","lang":"en","doc_type":"code","stars":683,"dataset":"github-code","pt":"68"}
+{"seq_id":"7294531201","text":"# create your Song class in this file\n\"\"\"\nThis class only interacts with song\n\"\"\"\n\nclass Song:\n #Determine items a song\n def __init__(self, title=\"\", artist=\"\", year=0, required=\"\"):\n self.artist = artist\n self.title = title\n self.year = year\n self.required = required\n\n #Display inputted songs\n def __str__(self):\n if self.required == \"n\":\n required = \"learned\"\n return (\"You have learned {} by {} ({})\".format(self.title,self.artist, self.year))\n else:\n required = \"y\"\n return (\"You have not learned {} by {} ({})\".format(self.title,self.artist, self.year))\n\n def mark_learned(self):\n #Mark the song learned\n self.required = 'n'\n return self.required\n","repo_name":"jc474652/leminhcp1404","sub_path":"song.py","file_name":"song.py","file_ext":"py","file_size_in_byte":775,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"11043998530","text":"from django.test import SimpleTestCase, TestCase as DjangoTestCase, LiveServerTestCase\nfrom django.urls import reverse\nfrom django.test.utils import override_settings\nfrom unittest import skipUnless\nfrom ackrep_core.test._test_utils import load_repo_to_db_for_ut, reset_repo\nfrom ackrep_core.util import run_command, utf8decode, strip_decode\nimport re\nimport json\nimport os\nfrom ackrep_core_django_settings import settings\nimport time\nimport subprocess\n\ntry:\n # noinspection PyPackageRequirements\n from splinter import Browser\nexcept ImportError:\n Browser = None\n splinter_available = False\nelse:\n splinter_available = True\n\nfrom ackrep_core import core\n\n# noinspection PyUnresolvedReferences\nfrom ipydex import IPS\n\n\nurl_of_external_test_repo = \"https://codeberg.org/cknoll/ackrep_data_demo_fork.git\"\n\n\"\"\"\nThis module contains the tests for the web application module (not ackrep_core)\n\n\npython3 manage.py test --keepdb --nocapture --rednose --ips ackrep_web.test.test_web:TestCases1\npython manage.py test --keepdb -v 2 --nocapture ackrep_web.test.test_web\n\nFor more infos see doc/devdoc/README.md.\n\"\"\"\n\n# inform the core module which path it should consinder as data repo\nackrep_data_test_repo_path = core.data_path = os.path.join(core.root_path, \"ackrep_data_for_unittests\")\n# this must also be set as env var because the tests will call some functions of ackrep\n# via command line\nos.environ[\"ACKREP_DATA_PATH\"] = ackrep_data_test_repo_path\n\n# due to the command line callings we also need to specify the test-database\nos.environ[\"ACKREP_DATABASE_PATH\"] = os.path.join(core.root_path, \"ackrep_core\", \"db_for_unittests.sqlite3\")\n\n# prevent cli commands to get stuck in unexpected IPython shell on error\n# (comment out for debugging)\nos.environ[\"NO_IPS_EXCEPTHOOK\"] = \"True\"\n\n# inform the core module which path it should consinder as results repo\nackrep_ci_results_test_repo_path = core.ci_results_path = os.path.join(\n core.root_path, \"ackrep_ci_results_for_unittests\"\n)\nos.environ[\"ACKREP_CI_RESULTS_PATH\"] = ackrep_ci_results_test_repo_path\n\n\nclass TestCases1(DjangoTestCase):\n def test_00(self):\n # for debugging\n pass\n # IPS()\n\n def test_landing_page(self):\n url = reverse(\"landing-page\")\n response = self.client.get(url)\n self.assertEqual(url, \"/\")\n\n # this should be a standard response for successful HTTP requests\n self.assertEqual(response.status_code, 200)\n\n bad_url = \"foo-bar/baz\"\n response = self.client.get(bad_url)\n self.assertEqual(response.status_code, 404)\n\n def test_entity_list(self):\n url = reverse(\"entity-list\")\n response = self.client.get(url)\n\n # this should be a standard response for successful HTTP requests\n self.assertEqual(response.status_code, 200)\n\n\nclass TestCases2(SimpleTestCase):\n \"\"\"\n These tests expect the database to be loaded\n \"\"\"\n\n databases = \"__all__\"\n\n def setUp(self):\n reset_repo(ackrep_data_test_repo_path)\n self.load_db()\n\n def load_db(self):\n load_repo_to_db_for_ut(ackrep_data_test_repo_path)\n\n def test_entity_detail(self):\n url = reverse(\"entity-detail\", kwargs={\"key\": \"UKJZI\"})\n response = self.client.get(url)\n self.assertEqual(response.status_code, 200)\n\n self.assertContains(response, \"utc_entity_full\")\n self.assertContains(response, \"Success\")\n\n url = reverse(\"entity-detail\", kwargs={\"key\": \"UXMFA\"})\n response = self.client.get(url)\n self.assertEqual(response.status_code, 200)\n\n self.assertContains(response, \"utc_entity_full\")\n self.assertContains(response, \"Success\")\n\n # skip test if done in CI, see https://ackrep-doc.readthedocs.io/en/latest/devdoc/design_considerations.html#ci\n # if os.environ.get(\"CI\") != \"true\":\n # self.assertContains(response, \"utc_img_url\")\n\n # regex = re.compile(\"utc_img_url:<(.*?)>\")\n # img_url = regex.findall(response.content.decode(\"utf8\"))\n\n # response = self.client.get(img_url)\n\n # # TODO test that this url returns a file\n\n @override_settings(DEBUG=True)\n def test_debug_message_printing(self):\n\n # broken lorenz system\n url = reverse(\"entity-detail\", kwargs={\"key\": \"LRHZX\"})\n\n # prevent expected error logs from showing during test\n loglevel = core.logger.level\n # core.logger.setLevel(50)\n\n # first: check if debug message shows when it should\n settings.DEBUG = True\n response = self.client.get(url)\n\n expected_error_infos = [\"utc_debug\", \"SyntaxError\", \"parameters.py\", \"line\"]\n for info in expected_error_infos:\n self.assertContains(response, info)\n self.assertNotContains(response, \"utc_output\")\n\n # second: check if debug message shows when it shouldn't\n settings.DEBUG = False\n response = self.client.get(url)\n\n expected_error_infos = [\"utc_debug\", \"SyntaxError\"]\n for info in expected_error_infos:\n self.assertNotContains(response, info)\n self.assertNotContains(response, \"utc_output\")\n\n core.logger.setLevel(loglevel)\n\n def test_show_last_passing(self):\n url = reverse(\"entity-detail\", kwargs={\"key\": \"LRHZX\"})\n response = self.client.get(url)\n infos = [\"Entity passed last:\", \"2022-06-24 00:00:00\"]\n for info in infos:\n self.assertContains(response, info)\n\n def test_sparql_query(self):\n\n url = reverse(\"search-sparql\")\n query = (\n \"query=%23+example+query%3A+select+all+possible+tags%0D%0A%0D%0APREFIX+P%3A+\"\n \"%0D%0A++++++++++++SELECT\"\n \"+%3Fentity%0D%0A++++++++++++WHERE+{%0D%0A++++++++++++++%3Fentity+P%3Ahas_ontology_based_tag\"\n \"+P%3AiTransfer_Function.%0D%0A++++++++++++}%0D%0A\"\n )\n response = self.client.get(f\"{url}?{query}\")\n\n self.assertEqual(response.status_code, 200)\n self.assertContains(response, \"utc_template_name=ackrep_web/search_sparql.html\")\n\n def tearDown(self) -> None:\n reset_repo(ackrep_data_test_repo_path)\n return super().tearDown()\n\n\nclass TestBugs(DjangoTestCase):\n \"\"\"\n Test for specific bugs\n \"\"\"\n\n def setUp(self):\n load_repo_to_db_for_ut(ackrep_data_test_repo_path)\n\n def test_entity_detail(self):\n url = reverse(\"entity-detail\", kwargs={\"key\": \"YJBOX\"})\n response = self.client.get(url)\n self.assertEqual(response.status_code, 200)\n\n self.assertContains(response, \"utc_entity_full\")\n\n\n@skipUnless(splinter_available, reason=\"browser automation is not installed\")\nclass TestUI(LiveServerTestCase):\n \"\"\"\n Itegration tests via browser automation (package: splinter)\n \"\"\"\n\n live_server_url = \"http://127.0.0.1:8000\"\n\n def setUp(self):\n d = dict()\n d[\"loggingPrefs\"] = {\"browser\": \"ALL\"}\n self.options_for_browser = dict(driver_name=\"chrome\", headless=True, desired_capabilities=d)\n\n self.browsers = []\n\n def tearDown(self):\n # quit all browser instances (also those which where not created by setUp)\n for browser in self.browsers:\n browser.quit()\n\n def local_reverse(self, *args, **kwargs):\n return f\"{self.live_server_url}{reverse(*args, **kwargs)}\"\n\n # noinspection PyMethodMayBeStatic\n def get_status_code(self, browser):\n \"\"\"\n for design reasons splinter does not grant access to http-status-codes.\n Thus, we write them to the base-template via a middleware, fetch them from the html source in the test\n\n :return:\n \"\"\"\n\n elt_list = browser.find_by_xpath('//script[@id=\"http_status_code\"]')\n if len(elt_list) == 0:\n # the http_status_code was not in the recieved page\n return None\n elif len(elt_list) == 1:\n raw_data = elt_list.first.html\n return json.loads(raw_data)\n else:\n msg = \"Multiple http_status_code-tags were found unexpectedly. Check template!\"\n raise ValueError(msg)\n\n @staticmethod\n def get_browser_log(browser):\n res = browser.driver.get_log(\"browser\")\n browser.logs.append(res)\n return res\n\n def new_browser(self):\n \"\"\"\n create and register a new browser\n\n :return: browser object and its index\n \"\"\"\n browser = Browser(**self.options_for_browser)\n browser.logs = []\n self.browsers.append(browser)\n\n return browser\n\n def test_list_entities(self):\n b = self.new_browser()\n url1 = self.local_reverse(\"landing-page\")\n b.visit(url1)\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n link = b.find_by_id(\"link_entity_list\")\n link.click()\n\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n def test_list_merge_requests(self):\n b = self.new_browser()\n url1 = self.local_reverse(\"landing-page\")\n b.visit(url1)\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n link = b.find_by_id(\"link_merge_request_list\")\n link.click()\n\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n def test_open_new_merge_request_form(self):\n b = self.new_browser()\n url1 = self.local_reverse(\"landing-page\")\n b.visit(url1)\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n link = b.find_by_id(\"link_new_merge_request\")\n link.click()\n\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n def test_sparql_search_dialog(self):\n b = self.new_browser()\n url1 = self.local_reverse(\"landing-page\")\n b.visit(url1)\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n\n button = b.find_by_id(\"link_search_sparql\")\n button.click()\n\n status_code = self.get_status_code(b)\n self.assertEqual(status_code, 200)\n","repo_name":"ackrep-org/ackrep_core","sub_path":"ackrep_web/test/test_web.py","file_name":"test_web.py","file_ext":"py","file_size_in_byte":10197,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"68"}
+{"seq_id":"6248592016","text":"from selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nimport time\nimport ctypes\nimport os\nimport sys\n\n\ndef EnterTextBoxWait(Id, text):\n element = wait.until(EC.presence_of_element_located((By.ID, Id))) #Ожидаем пока найдём id\n element.send_keys(text) #Отправляем ввод данных в тестовое поле\n\ndef ClickButtonById(Id):\n element = wait.until(EC.presence_of_element_located((By.ID, Id)))\n element.click() #Кликаем\nscript_path = os.path.dirname(os.path.abspath(sys.argv[0]))\nProcess = ctypes.CDLL(script_path + '\\Process.dll') # Замените на имя вашей DLL\ndriver = webdriver.Chrome()\n\n\nurl = \"https://www.roblox.com/login\" \n\n# Переход по URL\ndriver.get(url)\nwait = WebDriverWait(driver, 10) # Ожидаем максимум 10 секунд загрузки веб страницы\n\n#Начать цикл здесь\nEnterTextBoxWait(\"login-username\", \"r1l3g7x4o7b4e2\")\nEnterTextBoxWait(\"login-password\", \"s0qvdyemtc\")\nClickButtonById(\"login-button\")\nurl = \"https://www.roblox.com/home\"\nwait.until(EC.url_to_be(url)) #Ожидаем пока перекинет на новый url \nRobloxGame = \"https://www.roblox.com/games/13484061921/Eternal-Dream\"\ndriver.get(RobloxGame)\nwait.until(EC.url_to_be(RobloxGame)) #Ожидаем пока перекинет на новый url \nelement = wait.until(EC.presence_of_element_located((By.CLASS_NAME, \"icon-favorite\")))\nelement.click() \nelement = wait.until(EC.presence_of_element_located((By.CLASS_NAME, \"icon-common-play\")))\nelement.click() \nProcess.WaitProcessRoblox()\nelement = wait.until(EC.presence_of_element_located((By.CLASS_NAME, \"icon-like\")))\nelement.click() \nelement = wait.until(EC.presence_of_element_located((By.ID, \"nav-settings\")))\nelement.click() \ntime.sleep(1)\nelement = wait.until(EC.presence_of_element_located((By.CLASS_NAME, \"logout-menu-item\")))\nelement.click()\ntime.sleep(1)\nurl = \"https://www.roblox.com/login\"\ndriver.get(url)\nwait.until(EC.url_to_be(url))\n\nwhile True:\n time.sleep(1)","repo_name":"Gensokyo-Mangekyo/RobloxLikesTouhou","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2193,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"43059964269","text":"import collections\nimport struct\n\n\n# 2.1.2 EU863-870 ISM Band channel frequencies\n# The network channels can be freely attributed by the network operator. However the three\n# following default channels must be implemented in every EU868MHz end-device. Those\n# channels are the minimum set that all network gateways should always be listening on.\nMIN_LORA_FREQ = ({\"freq\": 868.1, \"max_dr\": 5, \"min_dr\": 0, \"mandatory\": True},\n {\"freq\": 868.3, \"max_dr\": 5, \"min_dr\": 0, \"mandatory\": True},\n {\"freq\": 868.5, \"max_dr\": 5, \"min_dr\": 0, \"mandatory\": True},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False},\n {\"freq\": 0, \"max_dr\": 0, \"min_dr\": 0, \"mandatory\": False})\n\nVALID_FREQ = [(8631+2*i)/10 for i in range(0, 28)] + [869.525, 869.850] + [(8701+2*i)/10 for i in range(0, 15)]\n\n\ndef get_min_dr(frequency):\n \"\"\" Returns the MIN data rate for the provided frequency.\"\"\"\n # if frequency in VALID_FREQ:\n # return 0\n # else:\n # return 0\n return 0\n\n\ndef get_max_dr(frequency):\n \"\"\" Returns the MAX data rate for the provided frequency.\"\"\"\n if frequency in VALID_FREQ:\n return 5\n else:\n return 0\n\n\n# 2.1.3 EU863-870 Data Rate and End-device Output Power encoding\nLoraDr = collections.namedtuple(\"LoraDr\", \"DR0 DR1 DR2 DR3 DR4 DR5 DR6\")\nLORA_DR = LoraDr(\n DR0='SF12BW125',\n DR1='SF11BW125',\n DR2='SF10BW125',\n DR3='SF9BW125',\n DR4='SF8BW125',\n DR5='SF7BW125',\n DR6='SF7BW250'\n)\n\nDrOffset = collections.namedtuple(\"DrOffset\", \"MIN MAX RX1_DEFAULT\")\nDR_OFFSET = DrOffset(\n MIN=0,\n MAX=5,\n RX1_DEFAULT=0\n)\n\n# 2.1.7 EU863-870 Receive windows\n# The RX1 receive window uses the same channel than the preceding uplink. The data rate is\n# a function of the uplink data rate and the RX1DROffset as given by the following table. The\n# allowed values for RX1DROffset are in the [0:5] range. Values in the [6:7] range are reserved\n# for future use.\n# The RX2 receive window uses a fixed frequency and data rate. The default lorawan_parameters are\n# 869.525 MHz / DR0 (SF12, 125 kHz)\nJoinAcceptDlSettings = collections.namedtuple(\"JoinAcceptDlSettings\",\n \"RX1OFFSET0_RX2DR0 \\\n RX1OFFSET1_RX2DR0 \\\n RX1OFFSET2_RX2DR0 \\\n RX1OFFSET3_RX2DR0 \\\n RX1OFFSET4_RX2DR0 \\\n RX1OFFSET5_RX2DR0 \\\n RX1OFFSET0_RX2DR1 \\\n RX1OFFSET1_RX2DR1 \\\n RX1OFFSET2_RX2DR1 \\\n RX1OFFSET3_RX2DR1 \\\n RX1OFFSET4_RX2DR1 \\\n RX1OFFSET5_RX2DR1 \\\n RX1OFFSET0_RX2DR2 \\\n RX1OFFSET1_RX2DR2 \\\n RX1OFFSET2_RX2DR2 \\\n RX1OFFSET3_RX2DR2 \\\n RX1OFFSET4_RX2DR2 \\\n RX1OFFSET5_RX2DR2 \\\n RX1OFFSET0_RX2DR3 \\\n RX1OFFSET1_RX2DR3 \\\n RX1OFFSET2_RX2DR3 \\\n RX1OFFSET3_RX2DR3 \\\n RX1OFFSET4_RX2DR3 \\\n RX1OFFSET5_RX2DR3\"\n )\nDLSETTINGS = JoinAcceptDlSettings(\n RX1OFFSET0_RX2DR0=b'\\x00',\n RX1OFFSET1_RX2DR0=b'\\x10',\n RX1OFFSET2_RX2DR0=b'\\x20',\n RX1OFFSET3_RX2DR0=b'\\x30',\n RX1OFFSET4_RX2DR0=b'\\x40',\n RX1OFFSET5_RX2DR0=b'\\x50',\n RX1OFFSET0_RX2DR1=b'\\x01',\n RX1OFFSET1_RX2DR1=b'\\x11',\n RX1OFFSET2_RX2DR1=b'\\x21',\n RX1OFFSET3_RX2DR1=b'\\x31',\n RX1OFFSET4_RX2DR1=b'\\x41',\n RX1OFFSET5_RX2DR1=b'\\x51',\n RX1OFFSET0_RX2DR2=b'\\x02',\n RX1OFFSET1_RX2DR2=b'\\x12',\n RX1OFFSET2_RX2DR2=b'\\x22',\n RX1OFFSET3_RX2DR2=b'\\x32',\n RX1OFFSET4_RX2DR2=b'\\x42',\n RX1OFFSET5_RX2DR2=b'\\x52',\n RX1OFFSET0_RX2DR3=b'\\x03',\n RX1OFFSET1_RX2DR3=b'\\x13',\n RX1OFFSET2_RX2DR3=b'\\x23',\n RX1OFFSET3_RX2DR3=b'\\x33',\n RX1OFFSET4_RX2DR3=b'\\x43',\n RX1OFFSET5_RX2DR3=b'\\x53'\n)\n\n# 2.1.9 EU863-870 Default Settings\n# If the actual parameter values implemented in the end-device are different from those default\n# values (for example the end-device uses a longer RECEIVE_DELAY1 and\n# RECEIVE_DELAY2 latency), those lorawan_parameters must be communicated to the network server\n# using an out-of-band channel during the end-device commissioning process. The network\n# server may not accept lorawan_parameters different from those default values.\n\n# 5 EU863-870 MAC Commands\n#\nMacCommands = collections.namedtuple(\"MacCommands\",\n \"LinkADRReq \\\n DevStatusReq\")\nMAC_COMMANDS = MacCommands(\n LinkADRReq=b'\\x03',\n DevStatusReq=b'\\x06'\n)\n\n\nDefault = collections.namedtuple(\"Default\", \"RECEIVE_DELAY1 RECEIVE_DELAY2 JOIN_ACCEPT_DELAY1 \\\n JOIN_ACCEPT_DELAY2 MAX_FCNT_GAP ADR_ACK_LIMIT ADR_ACK_DELAY ACK_TIMEOUT \\\n RX2_DEFAULT_FREQ\")\nDEFAULT_SETTINGS = Default(\n RECEIVE_DELAY1=1000000,\n RECEIVE_DELAY2=2000000,\n JOIN_ACCEPT_DELAY1=5000000,\n JOIN_ACCEPT_DELAY2=6000000,\n MAX_FCNT_GAP=16384,\n ADR_ACK_LIMIT=64,\n ADR_ACK_DELAY=32,\n ACK_TIMEOUT=2000000,\n RX2_DEFAULT_FREQ=869.525\n)\n\n\ndef get_cflist(*frequencies_to_add):\n \"\"\" Converts up to 5 frequency values into a 16 byte sequence. Each frequency is passed as an argument and\n is encoded as a 24 bits unsigned integer (three octets). All this channels have to be usable for DR0 to DR5\n 125kHz LoRa modulation. The frequency values must be provided in MHz (e.g. 868.1).\"\"\"\n cflist = b''\n assert 0 <= len(frequencies_to_add) <= 5\n for frequency in frequencies_to_add:\n if frequency in VALID_FREQ:\n # cflist += struct.pack('>BH', *(divmod(int(frequency * 10000), 1 << 16)))\n cflist += struct.pack('I\", b'\\x00' + freq_bytes[::-1])[0] / 10000\n\n\ndef parse_cflist(cflist_bytes):\n assert len(cflist_bytes) == 16\n # return [struct.unpack(\">I\", b'\\x00' + cflist_bytes[3 * i:3 * i + 3])[0] / 10000 for i in range(0, 5) if\n # not cflist_bytes[3 * i:3 * i + 3] == b'\\x00\\x00\\x00']\n return [struct.unpack(\">I\", b'\\x00' + cflist_bytes[3 * i:3 * i + 3][::-1])[0] / 10000 for i in range(0, 5) if\n not cflist_bytes[3 * i:3 * i + 3] == b'\\x00\\x00\\x00']\n\n","repo_name":"pablomodernell/lorawan_conformance_testing","sub_path":"lorawan/lorawan_parameters/region_eu868.py","file_name":"region_eu868.py","file_ext":"py","file_size_in_byte":8005,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"68"}
+{"seq_id":"8640316803","text":"from fastapi import FastAPI, UploadFile, File\r\nimport uvicorn\r\nimport argparse\r\nimport io\r\nimport cv2, pytesseract\r\nimport matplotlib.pyplot as plt\r\nfrom PIL import Image\r\nimport numpy as np\r\nfrom pytesseract import Output\r\nfrom starlette.responses import StreamingResponse\r\n\r\napp = FastAPI()\r\n\r\n\r\n# need to work in that part\r\ndef new_func(image):\r\n img_list = []\r\n opencvImage = cv2.imread(image)\r\n opencvImage = cv2.cvtColor(opencvImage, cv2.COLOR_BGR2RGB)\r\n opencvImage = cv2.cvtColor(opencvImage, cv2.COLOR_RGB2GRAY)\r\n pytesseract.pytesseract.tesseract_cmd = 'C:\\\\Program Files\\\\Tesseract-OCR\\\\tesseract.exe'\r\n ty = pytesseract.image_to_string(opencvImage)\r\n height, width = opencvImage.shape\r\n print(height,width)\r\n boxes = pytesseract.image_to_data(opencvImage)\r\n # print(boxes)\r\n for count, i in enumerate(boxes.splitlines()):\r\n d = {}\r\n if count!=0:\r\n i= i.split()\r\n \r\n if len(i)==12:\r\n x,y,w,h = int(i[6]),int(i[7]),int(i[8]), int(i[9])\r\n cv2.rectangle(opencvImage,(x,y),(x+w,h+y),(0,0,255),3)\r\n # x= left, y= top , x+w = right, h+y = bottom\r\n d['left'] = x\r\n d['top'] = y\r\n d['right'] = x+w\r\n d['bottom'] = h+y\r\n d['text'] = i[-1] \r\n img_list.append(d)\r\n cv2.putText(opencvImage, i[11],(x,y-5),cv2.FONT_HERSHEY_COMPLEX,0.5,(50,50,255),1)\r\n # b = i.split(' ')\r\n # print(b)\r\n print(img_list)\r\n cv2.imshow('',opencvImage)\r\n cv2.waitKey(0)\r\n return ty\r\n\r\n\r\n# new_func('C:\\\\Users\\\\cis\\\\Desktop\\\\fastapi_python\\\\fastapi_python\\\\serwis_detekcji_tekstu_image.jpg')\r\nnew_func(\"C:\\\\Users\\\\cis\\\\Downloads\\\\Release\\\\images\\\\download (1).png\")","repo_name":"sanjaysheel/text_detection_with_fastapi","sub_path":"fastapi_python/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":1796,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"724456371","text":"from django.urls import path\nfrom crm import views\n\n\nurlpatterns=[\n path('course',views.course,name='courseurl'),\n path('coursedetail/ ',views.course_details,name='coursedetailsurl'),\n path('enroll',views.enroll,name='enrollurl'),\n path('coursesearch',views.coursessearch,name='coursessearchurl'),\n path('adminhome',views.adminhome,name='adminhomeurl'),\n path('enrolledlist',views.enrolledlist,name='enrolledlisturl'),\n path('loginpage',views.loginpage,name='loginurl'),\n path('logout',views.logoutpage,name='logouturl'),\n path('coursesearch/',views.coursesearch,name='coursesearchurl'),\n path('adddemo',views.adddemo,name='adddemourl'),\n path('demoshow',views.demo_show,name='demoshowurl'),\n path('demoattend/',views.moverecords,name='demoattendurl'),\n path('demoattended_students',views.demoattended_students,name='demoattendstudentsurl'),\n path('demosearch/coursewise',views.demosearch,name='demosearchurl'),\n path('register',views.register,name='registerurl'),\n path('addstudent',views.add_student,name='addstudenturl'),\n path('studenthome',views.student_home,name='studenthomeurl'),\n]","repo_name":"sivamani159/CRM","sub_path":"crm/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1166,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}
+{"seq_id":"24293553174","text":"#DISCLAIMER\n#these tests work when I comment out the lines saying @login_required in the views.py file\n#I have uncommented those lines though as I want the edit and create CV features to only be available to me\n\nfrom django.urls import resolve\nfrom django.test import TestCase\nfrom django.http import HttpRequest\nfrom blog.views import post_list\nfrom blog.models import CV\nimport time\n\nclass CVListTest(TestCase):\n\n def test_returns_correct_html(self):\n response = self.client.get('/')\n self.assertTemplateUsed(response, 'blog/post_list.html', 'blog/base.html')\n\n def test_new_cv_returns_correct_html(self):\n response = self.client.get('/cv/new/')\n self.assertTemplateUsed(response, 'cv/cv_edit.html', 'blog/base.html')\n\n def test_can_save_a_POST_request(self):\n self.client.post('/cv/new/', data={'name': 'Matt', 'location': 'a', 'age': '1', 'contact': 'a', 'profile': 'a', 'experience': 'a', 'qualifications': 'a', 'hobbies': 'a', 'references': 'a'})\n self.assertEqual(CV.objects.count(), 1)\n new_cv = CV.objects.first()\n self.assertEqual(new_cv.name, 'Matt')\n\n def test_redirects_after_POST(self):\n response = self.client.post('/cv/new/', data={'name': 'Matt', 'location': 'a', 'age': '1', 'contact': 'a', 'profile': 'a', 'experience': 'a', 'qualifications': 'a', 'hobbies': 'a', 'references': 'a'})\n self.assertEqual(response.status_code, 302)\n self.assertEqual(response['location'], '/cv/1/')\n\n def test_only_saves_cv_when_necessary(self):\n self.client.get('/cv/')\n self.assertEqual(CV.objects.count(), 0)","repo_name":"mattledden/my-first-blog","sub_path":"blog/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1616,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"68"}