diff --git "a/4436.jsonl" "b/4436.jsonl"
new file mode 100644--- /dev/null
+++ "b/4436.jsonl"
@@ -0,0 +1,1975 @@
+{"seq_id":"74342334592","text":"import numpy as np\nimport agents\n\n\nCHECKPOINTS = 'checkpoints'\n\n\nclass Player(object):\n\n def __init__(self, env, monitor='output/', seed=None):\n self.env = env\n self.agents = {'universe': agents.A3C(env, monitor+'universe/',\n CHECKPOINTS+'/universe/'+env+'/', 1),\n 'tensorpack': agents.TPAgent(env, monitor+'tensorpack/',\n CHECKPOINTS+'/tensorpack/'+env+'/'+env, 1),\n 'random': agents.RandomAgent()}\n self.seed = seed\n self.best = ''\n \n \n def choose(self, num_episodes_eval=100):\n scores = {}\n for agent in self.agents.keys():\n scores[agent] = self.agents[agent].play(num_episodes_eval,\n env=self.env,\n record=False,\n seed=self.seed)\n \n self.best = max(scores, key=scores.get)\n\n\n def choose_and_record(self, num_episodes_eval=100, num_episodes_run=100):\n scores = {}\n for agent in self.agents.keys():\n scores[agent] = self.agents[agent].play(num_episodes_eval,\n env=self.env,\n record=False,\n seed=self.seed)\n \n self.best = max(scores, key=scores.get)\n self.agents[max(scores, key=scores.get)].play(num_episodes_run,\n env=self.env,\n record=True,\n seed=self.seed)\n\n def upload(self, outputm, api_key=''):\n self.agents[max(scores, key=scores.get)].do_submit(output, api_key)\n","repo_name":"libfun/deephack3","sub_path":"player.py","file_name":"player.py","file_ext":"py","file_size_in_byte":1925,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"60"}
+{"seq_id":"28228345063","text":"import cv2\nimport numpy as np\nfrom absl.logging import info\n\n\ndef visualize_image(image, scale, path):\n image = image.permute(1, 2, 0).cpu().numpy() * 255\n image = cv2.resize(image, (0, 0), fx=scale, fy=scale)\n info(f\"Saving image with scale({scale}) to {path}.\")\n image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n cv2.imwrite(filename=path, img=image)\n\n\ndef visualize_event(event, scale, path):\n event = event.permute(1, 2, 0).cpu().numpy()\n event_image = np.zeros((event.shape[0], event.shape[1], 3)) + 255\n event_image[event[:, :, 0] > 0] = [0, 0, 255]\n event_image[event[:, :, 1] > 0] = [255, 0, 0]\n event_image = event_image.astype(np.uint8)\n event_image = cv2.resize(event_image, (0, 0), fx=scale, fy=scale)\n info(f\"Saving event with scale({scale}) to {path}.\")\n cv2.imwrite(filename=path, img=event_image)\n\n\ndef visualize_event_alpx(event, path):\n event = event.squeeze().permute(1, 2, 0).cpu().numpy()\n event_abs = np.abs(event)\n event_abs = np.sum(event_abs, axis=2)\n event = np.sum(event, axis=2)\n event_image = np.zeros((event.shape[0], event.shape[1], 3)) + 255\n event_image[event_abs > 0] = [0, 0, 0]\n event_image[event > 0] = [0, 0, 255]\n event_image[event < 0] = [255, 0, 0]\n event_image = event_image.astype(np.uint8)\n info(f\"Saving event to {path}.\")\n cv2.imwrite(filename=path, img=event_image)\n","repo_name":"yunfanLu/INR-Event-VSR","sub_path":"egvsr/utils/visualize.py","file_name":"visualize.py","file_ext":"py","file_size_in_byte":1383,"program_lang":"python","lang":"en","doc_type":"code","stars":24,"dataset":"github-code","pt":"60"}
+{"seq_id":"1330376013","text":"import streamlit as st\nimport osmnx as ox\nimport networkx as nx\nimport pandas as pd\n\n# Load the graph (assuming you have saved it as 'delhi_ev_graph.graphml')\nG = ox.load_graphml('delhi_ev_graph.graphml')\n\ndef shortest_path_with_constraints(G, origin, destination, battery_range, charging_time, weight='length'):\n # Find the shortest path between origin and destination\n path = nx.shortest_path(G, origin, destination, weight=weight)\n\n # Calculate the path length\n path_length = nx.path_weight(G, path, weight=weight)\n\n # Check if the path length exceeds the battery range\n if path_length > battery_range:\n # Find charging stations on the path\n charging_stations_on_path = [node for node in path if G.nodes[node].get('charging_station', False)]\n\n # Check if there are charging stations on the path\n if charging_stations_on_path:\n # Find the nearest charging station to the origin\n nearest_charging_station = charging_stations_on_path[0]\n nearest_charging_station_distance = nx.shortest_path_length(G, origin, nearest_charging_station, weight=weight)\n\n # Compute the remaining battery range after reaching the nearest charging station\n remaining_battery_range = battery_range - nearest_charging_station_distance\n\n # Recursively compute the shortest path from the nearest charging station to the destination\n sub_path = shortest_path_with_constraints(G, nearest_charging_station, destination, remaining_battery_range + charging_time * 100, charging_time, weight)\n\n # Combine the sub-path with the current path\n path = path[:path.index(nearest_charging_station) + 1] + sub_path[1:]\n\n return path\n\n\n# Define a function to compute the battery range gained by charging for a specified time\ndef compute_battery_range(charging_time):\n # Compute the battery range gained by charging for the specified time\n battery_range_gained = charging_time * 60 # Assume 60 km of range gained per hour of charging\n \n return battery_range_gained\n\n# Streamlit app layout\nst.title(\"EV Routing Application\")\n\nstart_address = st.text_input(\"Enter starting address:\", value=\"\")\nend_address = st.text_input(\"Enter destination address:\", value=\"\")\nbattery_charge = st.number_input(\"Enter current battery charge (in kilometers):\", value=0)\ncharging_time = st.number_input(\"Enter charging time (in hours):\", value=0)\n\nif start_address and end_address and battery_charge is not None and charging_time is not None:\n # Geocode the addresses to get coordinates\n start_location = ox.geocode(start_address)\n end_location = ox.geocode(end_address) \n\n # Find the nearest nodes to the starting and ending points\n start_node = ox.distance.nearest_nodes(G, X=[start_location[1]], Y=[start_location[0]], return_dist=False)[0]\n end_node = ox.distance.nearest_nodes(G, X=[end_location[1]], Y=[end_location[0]], return_dist=False)[0]\n\n # Compute the battery range gained by charging for the specified time\n battery_range_gained = compute_battery_range(charging_time)\n\n # Compute the battery range considering the current battery charge and charging time\n battery_range = battery_charge + battery_range_gained\n\n # Compute the shortest path using Dijkstra's algorithm based on distance and battery constraints\n shortest_path = shortest_path_with_constraints(G, start_node, end_node, battery_range, charging_time, weight='length')\n shortest_path_distance = nx.shortest_path_length(G, source=start_node, target=end_node, weight='length', method='dijkstra')\n # Get the route geometry\n route_edges = ox.utils_graph.get_route_edge_attributes(G, shortest_path, attribute='geometry')\n route_geometry = [item for sublist in route_edges for item in sublist]\n \n # Plot the route\n fig, ax = ox.plot_graph_route(G, route=shortest_path, route_linewidth=6, node_size=0, bgcolor='k', edge_color='gray', edge_alpha=0.2, route_color='b')\n\n# Plot the origin and destination nodes\n orig_node_geom = G.nodes[start_node]['route_geometry']\n dest_node_geom = G.nodes[end_node]['route_geometry']\n ax.scatter(orig_node_geom.x, orig_node_geom.y, c='r', s=100, zorder=3)\n ax.scatter(dest_node_geom.x, dest_node_geom.y, c='r', s=100, zorder=3)\n \n# # Plot the route\n# fig, ax = ox.plot_graph_route(G, route=route_geometry, route_linewidth=6, node_size=0, bgcolor='k', edge_color='gray', edge_alpha=0.2, orig_dest_node_color='r', route_color='b')\n\n# # Show the route map\n# st.pyplot(route_map)\n\n# Display the route distance\n st.write(f\"Shortest path distance: {shortest_path_distance} meters\")\n","repo_name":"RhythmBindal/Predictive_Routing_for_EV_Charging_Stations","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":4660,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"37650503089","text":"from enum import Enum, auto\nimport random\nimport sys\n\n\nclass BoardState(Enum):\n BLANK = auto()\n PLAYER = auto()\n CPU = auto()\n\n\nclass GameState(Enum):\n PLAYING = auto()\n PLAYER_WIN = auto()\n CPU_WIN = auto()\n DRAW = auto()\n\n\nclass Board():\n def __init__(self):\n # ボードの幅\n self.board_length = 3\n # マスの数\n self.board_area = self.board_length ** 2\n # 初期化\n self.board = [BoardState.BLANK] * self.board_area\n # 先攻:True、後攻:False\n self.is_player_turn = True\n # ゲームの状態\n self.state = GameState.PLAYING\n\n def display(self):\n print_length = len(str(self.board_area)) + 1\n for i in range(self.board_area):\n if self.board[i] == BoardState.PLAYER:\n print('o'.rjust(print_length), end='')\n elif self.board[i] == BoardState.CPU:\n print('x'.rjust(print_length), end='')\n else:\n print('{0:{print_length}d}'.format(\n i, print_length=print_length), end='')\n if (i + 1) % self.board_length == 0:\n print()\n\n def input_player(self):\n while True:\n i = int(input())\n if i < 0 or i >= self.board_area or self.stone_exists(i) is True:\n print('error: invalid number')\n continue\n self.set_stone(i)\n break\n\n def input_cpu(self):\n idxs = [i for i, x in enumerate(self.board) if x == BoardState.BLANK]\n i = random.randint(0, len(idxs) - 1)\n self.set_stone(idxs[i])\n\n def stone_exists(self, i):\n return True if self.board[i] != BoardState.BLANK else False\n\n def set_stone(self, i):\n self.board[i] = BoardState.PLAYER if self.is_player_turn else BoardState.CPU\n self.update_state()\n self.next_turn()\n\n def next_turn(self):\n self.is_player_turn = not self.is_player_turn\n\n def judge(self, board_sub):\n return True if board_sub[0] != BoardState.BLANK and all([board_sub[0] == i for i in board_sub]) else False\n\n def update_state(self):\n if self.judge(self.board[0: self.board_area: self.board_length + 1]) or \\\n self.judge(self.board[self.board_length - 1: (self.board_length - 1) * self.board_length + 1: self.board_length - 1]):\n if self.is_player_turn:\n self.state = GameState.PLAYER_WIN\n else:\n self.state = GameState.CPU_WIN\n return\n for i in range(self.board_length):\n if self.judge(self.board[i: self.board_area: self.board_length]) or \\\n self.judge(self.board[self.board_length * i: self.board_length * (i + 1)]):\n if self.is_player_turn:\n self.state = GameState.PLAYER_WIN\n else:\n self.state = GameState.CPU_WIN\n return\n if BoardState.BLANK not in self.board:\n self.state = GameState.DRAW\n return\n self.state = GameState.PLAYING\n\n\nclass TicTacToe(Board):\n def input_cpu(self):\n self.set_stone(self.minimax(0))\n\n def unset_stone(self, i):\n self.board[i] = BoardState.BLANK\n self.next_turn()\n\n def evaluate(self, depth):\n if self.state == GameState.CPU_WIN:\n self.state = GameState.PLAYING\n return 10 - depth\n elif self.state == GameState.PLAYER_WIN:\n self.state = GameState.PLAYING\n return depth - 10\n else:\n self.state = GameState.PLAYING\n return 0\n\n def minimax(self, depth):\n if self.state != GameState.PLAYING:\n return self.evaluate(depth)\n best_i = 0\n evaluation_value = sys.maxsize if self.is_player_turn else (\n -1) * sys.maxsize\n for i in range(self.board_area):\n if self.stone_exists(i) is False:\n self.set_stone(i)\n evaluation_value_tmp = self.minimax(depth + 1)\n if self.is_player_turn:\n if evaluation_value_tmp > evaluation_value:\n evaluation_value = evaluation_value_tmp\n best_i = i\n else:\n if evaluation_value_tmp < evaluation_value:\n evaluation_value = evaluation_value_tmp\n best_i = i\n self.unset_stone(i)\n if depth == 0:\n return best_i\n else:\n return evaluation_value\n\n def run(self):\n self.display()\n while self.state == GameState.PLAYING:\n msg = 'PLAYER' if self.is_player_turn else 'CPU'\n print(msg)\n if self.is_player_turn:\n self.input_player()\n else:\n self.input_cpu()\n self.display()\n msg = 'player win' if self.state == GameState.PLAYER_WIN else 'cpu win' if self.state == GameState.CPU_WIN else 'draw'\n print(msg)\n\n\nif __name__ == \"__main__\":\n tictactoe = TicTacToe()\n tictactoe.run()\n","repo_name":"ogyogy/tic-tac-toe","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5110,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"24801274237","text":"#!/usr/bin/env python3\n\n\"\"\"\nRun with Psy.0.1.typelib on a default path or specify the following\nenvironmental variables before starting this script:\n GI_TYPELIB_PATH = \"path/to/Psy-0.1.typelib\"\nWhen the suitable libpsy-1.0.so .dll is available on the path you\nshould be able to run it otherwise use:\n LD_LIBRARY_PATH = \"path/to/libpsy.so/folder\"\n\"\"\"\n\nimport gi\nimport typing\nimport math as m\nimport time as t\nimport argparse as ap\n\ngi.require_versions({\"Psy\": \"0.1\", \"GLib\": \"2.0\"})\n\nfrom gi.repository import Psy\nfrom gi.repository import GLib\n\n\nclass MyCross(Psy.Cross):\n \"\"\"\n In order to override a virtual method you have to prepend your\n method with do_, so PsyCrossClass->update is called in python\n by MyCross.do_update\n \"\"\"\n\n def do_update(self, timepoint, frame_num):\n self.props.x += 1\n self.props.y += 1\n self.set_color(\n Psy.Color(r=(m.sin(t.time()) / 2 + 0.5), g=m.cos(t.time()) / 2 + 0.5, b=0.5)\n )\n\n\nclass MyRect(Psy.Rectangle):\n def __init__(self, nth_frame, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.is_white = True\n self.white = Psy.Color.new_rgb(1, 1, 1)\n self.black = Psy.Color()\n self.nth_frame = nth_frame\n self.counter = 0\n\n def do_update(self, timepoint, frame_num):\n if self.counter % self.nth_frame == 0:\n if self.is_white:\n self.set_color(self.white)\n else:\n self.set_color(self.black)\n self.is_white = not self.is_white\n self.counter += 1\n\n\ndef stop_loop(\n circle: Psy.Circle,\n time_point: Psy.TimePoint,\n tup: typing.Tuple[GLib.MainLoop, Psy.TimePoint],\n):\n \"\"\"\n Exit from the mainloop and exit the program\n \"\"\"\n loop = tup[0]\n time_start = tup[1]\n\n try:\n print(\n \"Circle.start = {}\".format(\n circle.props.start_time.subtract(time_start).props.seconds\n )\n )\n print(\n \"Circle.stop = {}\".format(\n circle.props.stop_time.subtract(time_start).props.seconds\n )\n )\n print(\n \"The circle is presented for roughly = {} seconds\".format(\n circle.props.stop_time.subtract(circle.props.start_time).props.seconds\n )\n )\n except Exception as e:\n print(\"We shouldn't get here\", e)\n pass\n\n loop.quit()\n\n\ndef circle_update(circle: Psy.Circle, nth_frame, data):\n \"\"\"\n Do something nice with the circle\n \"\"\"\n circle.props.radius = circle.props.radius + 1\n circle.props.x = circle.props.x - 1\n circle.props.num_vertices = circle.props.num_vertices + 1\n\n\ndef main(args: ap.Namespace):\n \"\"\"run the program using the parameters in args\"\"\"\n loop = GLib.MainLoop()\n clock = Psy.Clock()\n start = clock.now()\n dur = Psy.Duration.new_ms(args.start_dur)\n stim_dur = Psy.Duration.new_ms(args.duration)\n window = Psy.GtkWindow(n_monitor=args.monitor)\n circle = Psy.Circle.new(window)\n rect = Psy.Rectangle.new(window)\n flikker = MyRect(\n args.swap,\n canvas=window,\n width=100,\n height=100,\n x=-1920 / 2 + 50,\n y=1080 / 2 - 50,\n )\n flikker.set_color(Psy.Color.new_rgb(1, 0, 0))\n rect.props.x, rect.props.y = -200, -200\n rect.props.width, rect.props.height = 100, 100\n rect.set_color(Psy.Color.new_rgb(1.0, 0.0, 0.0))\n cross = MyCross(canvas=window, x=200, y=200, line_length=100, line_width=30)\n circle.play_for(start.add(dur), stim_dur)\n cross.play_for(start.add(dur), stim_dur)\n rect.play_for(start.add(dur), stim_dur)\n flikker.play_for(start.add(dur), stim_dur)\n\n circle.connect(\"stopped\", stop_loop, (loop, start))\n circle.connect(\"update\", circle_update)\n\n loop.run()\n\n\nif __name__ == \"__main__\":\n cmd_parser = ap.ArgumentParser(\n \"python-test.py\",\n description=\"show some test stimuli\",\n epilog=\"Happy Experimenting\",\n )\n\n cmd_parser.add_argument(\n \"-m\", \"--monitor\", help=\"Choose a monitor by number\", type=int, default=0\n )\n cmd_parser.add_argument(\n \"-s\",\n \"--start-dur\",\n help=\"The start duration before onset of the stimuli in milliseconds\",\n type=int,\n default=500,\n )\n cmd_parser.add_argument(\n \"-d\",\n \"--duration\",\n help=\"Duration of the stimuli in milliseconds\",\n type=int,\n default=4000,\n )\n cmd_parser.add_argument(\n \"--swap\",\n help=\"num frames between successive swaps of the rectangle\",\n type=int,\n default=1,\n )\n\n args = cmd_parser.parse_args()\n main(args)\n","repo_name":"UiL-OTS-labs/psy-lib","sub_path":"tests/python-test.py","file_name":"python-test.py","file_ext":"py","file_size_in_byte":4667,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"23703488085","text":"class RunnerUp:\n if __name__ == '__main__':\n print(\"Please enter how many number do you want to put int? : \")\n n = int(input())\n arr = map(int, input().split(' '))\n runner_up_list = arr\n reversed_num = list(runner_up_list)\n\n print(reversed_num)\n\n print(reversed_num.sort(reverse=True))\n\n n = [2, 3, 4, 5, 6]\n n.sort(reverse=True)\n print(n[1])","repo_name":"MemoNano/Webinar","sub_path":"HackerRank/RunnerUp.py","file_name":"RunnerUp.py","file_ext":"py","file_size_in_byte":404,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"10353305917","text":"from http.server import BaseHTTPRequestHandler, HTTPServer\nimport os\nimport json\n\nclass S(BaseHTTPRequestHandler):\n\n def do_GET(self):\n # read title from path\n title = self.path[1:]\n rss_file = os.path.join('rss', f'{title}.jsonl')\n\n # check if file exists\n if not os.path.exists(rss_file):\n self.send_response(404)\n self.end_headers()\n return\n \n # load rss and send\n with open(rss_file) as f:\n rss = [json.loads(i) for i in f.read().splitlines()]\n rss.reverse()\n\n # create xml and send response\n xml = make_xml(title, rss)\n self.send_response(200)\n self.send_header(\"Content-type\", 'text/xml')\n self.send_header(\"Content-Length\", len(xml))\n self.end_headers()\n self.wfile.write(xml.encode('utf-8'))\n\n def do_POST(self):\n # read title from path\n title = self.path[1:]\n rss_file = os.path.join('rss', f'{title}.jsonl')\n\n # read data from POST body\n content_length = int(self.headers['Content-Length'])\n data = self.rfile.read(content_length)\n\n # check if valid json\n try:\n js = json.loads(data)\n if not ('name' in js and 'magnet' in js):\n raise Exception('Bad request')\n except:\n self.send_response(400)\n self.end_headers()\n return\n \n # fix escaping\n js['magnet'] = js['magnet'].replace('&', '&')\n\n # append to jsonl\n with open(rss_file, 'a') as f:\n json.dump(js, f)\n f.write('\\n')\n\n # send OK\n self.send_response(200)\n self.end_headers()\n\ndef make_xml(title, items):\n content = ''\n for item in items:\n content += f'{item[\"name\"]}{item[\"guid\"]}'\n return f'{title}{content}' \n\ndef start_server(port=8080):\n server_address = ('', port)\n httpd = HTTPServer(server_address, S)\n try:\n httpd.serve_forever()\n except KeyboardInterrupt:\n pass\n httpd.server_close()","repo_name":"Haroon96/TorrenterBot","sub_path":"rss_server.py","file_name":"rss_server.py","file_ext":"py","file_size_in_byte":2292,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"22739486150","text":"def bubble_sort(arr):\n n = len(arr)\n for i in range(n-1):\n # Flag to check if any swaps are made in this pass\n swapped = False\n\n # Last i elements are already in place\n for j in range(0, n - i - 1):\n # Swap if the element found is greater than the next element\n if arr[j] > arr[j + 1]:\n arr[j], arr[j + 1] = arr[j + 1], arr[j]\n swapped = True\n\n # If no swaps are made in this pass, the array is already sorted\n if not swapped:\n break\n\n# Example usage:\nif __name__ == \"__main__\":\n arr = [64, 34, 25, 12, 22, 11, 90]\n bubble_sort(arr)\n print(\"Sorted array:\", arr)\n","repo_name":"harshpujari/DSAbyThomasCormen","sub_path":"2.2bubbleSort.py","file_name":"2.2bubbleSort.py","file_ext":"py","file_size_in_byte":682,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"19067160013","text":"import datetime\n\nimport requests\nfrom datetime import datetime\nfrom django.utils import timezone\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.shortcuts import render, redirect\nimport hashlib\nimport psycopg2\nimport json\nfrom django.http import JsonResponse\n\nfrom common.models import Xfactor_Log\n\nwith open(\"setting.json\", encoding=\"UTF-8\") as f:\n SETTING = json.loads(f.read())\nDBHost = SETTING['DB']['DBHost']\nDBPort = SETTING['DB']['DBPort']\nDBName = SETTING['DB']['DBName']\nDBUser = SETTING['DB']['DBUser']\nDBPwd = SETTING['DB']['DBPwd']\nUserTNM = SETTING['DB']['UserTNM']\nLogin_Method = SETTING['PROJECT']['LOGIN']\napiUrl = SETTING['API']['apiUrl']\nSesstionKeyPath = SETTING['API']['PATH']['SessionKey']\n\n\n# hi\n@csrf_exempt\ndef signup(request):\n if request.method == \"GET\":\n return render(request, 'common/signup.html')\n\n elif request.method == \"POST\":\n page = request.POST.get('page')\n x_id = request.POST.get('x_id')\n x_pw = request.POST.get('x_pw')\n re_x_pw = request.POST.get('re_x_pw')\n x_name = request.POST.get('x_name')\n x_email = request.POST.get('x_email')\n x_auth = request.POST.get('x_auth')\n res_data = {}\n\n RS = createUsers(x_id, x_pw, x_name, x_email, x_auth)\n if RS == \"1\":\n if page == 'um':\n res_data['error'] = \"회원가입에 성공하였습니다.\"\n redirect_url = '../user_management'\n function = 'Add User' # 분류 정보를 원하시는 텍스트로 변경해주세요.\n item = 'Add user '+ x_id\n result = '성공'\n user = request.session.get('sessionid')\n date = timezone.now()\n Xfactor_log = Xfactor_Log(\n log_func=function,\n log_item=item,\n log_result=result,\n log_user=user,\n log_date=date\n )\n Xfactor_log.save()\n return redirect(redirect_url)\n #return render(request, 'user_management.html', res_data)\n else :\n function = 'Signup User' # 분류 정보를 원하시는 텍스트로 변경해주세요.\n item = 'Signup user ' + x_id\n result = '성공'\n user = x_id\n date = timezone.now()\n Xfactor_log = Xfactor_Log(\n log_func=function,\n log_item=item,\n log_result=result,\n log_user=user,\n log_date=date\n )\n Xfactor_log.save()\n res_data['error'] = \"회원가입에 성공하였습니다.\"\n return render(request, 'common/login.html', res_data)\n else:\n res_data['error'] = \"아이디가 존재합니다.\"\n res_data['x_id'] = x_id\n res_data['x_name'] = x_name\n res_data['x_email'] = x_email\n res_data['x_auth'] = x_auth\n return render(request, 'common/signup.html', res_data)\n\n@csrf_exempt\ndef login(request):\n if Login_Method == \"WEB\":\n if request.method == 'GET':\n return render(request, 'common/login.html')\n\n # POST 방식 요청 -> 사용자가 보내는 데이터와 데이터베이스의 정보 일치여부 확인\n elif request.method == 'POST':\n x_id = request.POST.get('x_id', None)\n x_pw = request.POST.get('x_pw', None)\n\n # 응답 데이터\n res_data = {}\n\n # 모든 필드를 채우지 않았을 경우\n if not (x_id and x_pw):\n res_data['error'] = '아이디 또는 비밀번호를 입력해 주세요.'\n return render(request, 'common/login.html', res_data)\n # 모든 필드를 채웠을 경우\n else:\n RS = selectUsers(x_id, x_pw)\n print(RS)\n if RS == None:\n res_data['error'] = '아이디 또는 비밀번호가 일치하지 않습니다'\n return render(request, 'common/login.html', res_data)\n\n else:\n request.session['sessionid'] = RS[0]\n request.session['sessionname'] = RS[2]\n request.session['sessionemail'] = RS[3]\n function = 'Login' # 분류 정보를 원하시는 텍스트로 변경해주세요.\n item = 'admin 계정'\n result = '성공'\n user = RS[0]\n now = timezone.now().replace(microsecond=0)\n date = now.strftime(\"%Y-%m-%d %H:%M:%S\")\n print(date)\n Xfactor_log = Xfactor_Log(\n log_func=function,\n log_item=item,\n log_result=result,\n log_user=user,\n log_date=date\n )\n Xfactor_log.save()\n return redirect('../dashboard')\n elif Login_Method == \"Tanium\":\n if request.method == 'GET':\n returnData = {'Login_Method': Login_Method}\n return render(request, 'common/login.html', returnData)\n\n elif request.method == 'POST':\n\n x_id = request.POST.get('x_id', None)\n x_pw = request.POST.get('x_pw', None)\n\n # 응답 데이터\n res_data = {}\n\n # 모든 필드를 채우지 않았을 경우\n if not (x_id and x_pw):\n res_data['error'] = '아이디 또는 비밀번호를 입력해 주세요.'\n return render(request, 'common/login.html', res_data)\n # 모든 필드를 채웠을 경우\n else:\n TRS = taniumUsers(x_id, x_pw)\n if TRS == None:\n res_data['error'] = '아이디 또는 비밀번호가 일치하지 않습니다'\n return render(request, 'common/login.html', res_data)\n else:\n request.session['sessionid'] = x_id\n return redirect('../dashboard')\n\n@csrf_exempt\ndef updateform(request):\n try:\n if request.method == \"GET\":\n # print(request.session.sessionid)\n return render(request, 'common/updateform.html')\n\n elif request.method == \"POST\":\n x_id = request.POST.get('x_id')\n x_pw = request.POST.get('x_pw')\n hashpassword = hashlib.sha256(x_pw.encode()).hexdigest()\n Conn = psycopg2.connect('host={0} port={1} dbname={2} user={3} password={4}'.format(DBHost, DBPort, DBName, DBUser, DBPwd))\n Cur = Conn.cursor()\n\n query = \"\"\"\n select\n *\n from\n \"\"\" + UserTNM + \"\"\"\n where\n x_id = '\"\"\" + x_id + \"\"\"'\n and\n x_pw = '\"\"\" + hashpassword + \"\"\"'\n\n \"\"\"\n Cur.execute(query)\n RS = Cur.fetchall()\n res_data = {}\n print(RS)\n if RS[0] != None:\n res_data['x_id'] = RS[0][0]\n res_data['x_name'] = RS[0][2]\n res_data['x_email'] = RS[0][3]\n res_data['x_auth'] = RS[0][4]\n # print(res_data)\n return render(request, 'common/update.html', res_data)\n except:\n res_data['error'] = '비밀번호를 다시한번 확인 해 주세요.'\n return render(request, 'common/updateform.html', res_data)\n\n@csrf_exempt\ndef update(request):\n if request.method == \"GET\":\n # 404 에러페이지 넣을것\n return render(request, '')\n\n elif request.method == \"POST\":\n x_id = request.POST.get('x_id')\n x_pw = request.POST.get('x_pw')\n re_x_pw = request.POST.get('re_x_pw')\n x_name = request.POST.get('x_name')\n x_email = request.POST.get('x_email')\n x_auth = request.POST.get('x_auth')\n res_data = {}\n if not (x_id and x_pw and x_name and x_email and x_auth):\n res_data['x_id'] = x_id\n res_data['x_pw'] = x_pw\n res_data['x_name'] = x_name\n res_data['x_email'] = x_email\n res_data['x_auth'] = x_auth\n res_data['error'] = \"모든 값을 입력해야 합니다.\"\n return render(request, 'common/update.html', res_data)\n if x_pw != re_x_pw:\n res_data['x_id'] = x_id\n res_data['x_pw'] = x_pw\n res_data['x_name'] = x_name\n res_data['x_email'] = x_email\n res_data['x_auth'] = x_auth\n res_data['error'] = '비밀번호가 다릅니다.'\n return render(request, 'common/update.html', res_data)\n else:\n RS = updateUsers(x_id, x_pw, x_name, x_email, x_auth)\n if RS == \"1\":\n request.session['sessionname'] = x_name\n request.session['sessionemail'] = x_email\n return redirect('../dashboard')\n else:\n res_data['error'] = '회원정보 변경이 실패했습니다.'\n return render(request, 'common/update.html', res_data)\n\n@csrf_exempt\ndef logout(request):\n if Login_Method == \"WEB\":\n if 'sessionid' in request.session:\n\n function = 'Logout' # 분류 정보를 원하시는 텍스트로 변경해주세요.\n item = 'admin 계정'\n result = '성공'\n user = request.session.get('sessionid')\n date = timezone.now()\n Xfactor_log = Xfactor_Log(\n log_func=function,\n log_item=item,\n log_result=result,\n log_user=user,\n log_date=date\n )\n Xfactor_log.save()\n del (request.session['sessionid'])\n del (request.session['sessionname'])\n del (request.session['sessionemail'])\n return render(request, 'common/login.html')\n else:\n return render(request, 'common/login.html')\n elif Login_Method == \"Tanium\":\n if request.method == 'GET':\n returnData = {'Login_Method': Login_Method}\n return render(request, 'common/login.html', returnData)\n if 'sessionid' in request.session:\n del (request.session['sessionid'])\n return render(request, 'common/login.html')\n else:\n return render(request, 'common/login.html')\n\n@csrf_exempt\ndef selectUsers(x_id, x_pw):\n try:\n hashpassword = hashlib.sha256(x_pw.encode()).hexdigest()\n # print(hashpassword)\n\n Conn = psycopg2.connect('host={0} port={1} dbname={2} user={3} password={4}'.format(DBHost, DBPort, DBName, DBUser, DBPwd))\n Cur = Conn.cursor()\n\n query = \"\"\"\n select \n *\n from\n \"\"\" + UserTNM + \"\"\"\n where\n x_id = '\"\"\" + x_id + \"\"\"'\n and\n x_pw = '\"\"\" + hashpassword + \"\"\"' \n\n \"\"\"\n\n Cur.execute(query)\n RS = Cur.fetchone()\n # print(RS)\n return RS\n except:\n print(UserTNM + ' Table connection(Select) Failure')\n\n@csrf_exempt\ndef createUsers(x_id, x_pw, x_name, x_email, x_auth):\n try:\n hashpassword = hashlib.sha256(x_pw.encode()).hexdigest()\n Conn = psycopg2.connect('host={0} port={1} dbname={2} user={3} password={4}'.format(DBHost, DBPort, DBName, DBUser, DBPwd))\n Cur = Conn.cursor()\n query = \"\"\" \n INSERT INTO \n common_xfactor_xuser\n (x_id, x_pw, x_name, x_email, x_auth) \n VALUES ( \n '\"\"\" + x_id + \"\"\"',\n '\"\"\" + hashpassword + \"\"\"' ,\n '\"\"\" + x_name + \"\"\"',\n '\"\"\" + x_email + \"\"\"',\n '\"\"\" + x_auth + \"\"\"'\n );\n \"\"\"\n Cur.execute(query)\n Conn.commit()\n Conn.close()\n a = \"1\"\n return a\n except:\n print(UserTNM + ' Table connection(Select) Failure')\n a = \"0\"\n return a\n\n\n@csrf_exempt\ndef updateUsers(x_id, x_pw, x_name, x_email, x_auth):\n try:\n hashpassword = hashlib.sha256(x_pw.encode()).hexdigest()\n Conn = psycopg2.connect('host={0} port={1} dbname={2} user={3} password={4}'.format(DBHost, DBPort, DBName, DBUser, DBPwd))\n Cur = Conn.cursor()\n query = \"\"\" \n UPDATE\n common_xfactor_xuser \n SET\n x_pw= '\"\"\" + hashpassword + \"\"\"',\n x_name= '\"\"\" + x_name + \"\"\"',\n x_email= '\"\"\" + x_email + \"\"\"',\n x_auth= '\"\"\" + x_auth + \"\"\"'\n WHERE\n x_id = '\"\"\" + x_id + \"\"\"';\n \"\"\"\n # print(query)\n Cur.execute(query)\n Conn.commit()\n Conn.close()\n a = \"1\"\n return a\n except:\n print(UserTNM + ' Table connection(Update) Failure')\n a = \"0\"\n return a\n\n@csrf_exempt\ndef taniumUsers(x_id, x_pw):\n try:\n path = SesstionKeyPath\n urls = apiUrl + path\n headers = '{\"username\" : \"' + x_id + '\",\"domain\":\"\", \"password\":\"' + x_pw + '\"}'\n response = requests.post(urls, data=headers, verify=False)\n code = response.status_code\n if code == 200:\n a = response.json()\n sessionKey = a['data']['session']\n returnList = sessionKey\n return returnList\n elif code == 403:\n print()\n\n except ConnectionError as e:\n print(e)\n\n@csrf_exempt\ndef delete(request):\n x_ids_str = request.POST.get('x_id') # 쉼표로 구분된 문자열을 얻음\n x_ids = x_ids_str.split(',')\n try:\n Conn = psycopg2.connect('host={0} port={1} dbname={2} user={3} password={4}'.format(DBHost, DBPort, DBName, DBUser, DBPwd))\n Cur = Conn.cursor()\n for x_id in x_ids:\n query = \"\"\" \n DELETE FROM\n common_xfactor_xuser\n WHERE\n x_id = %s;\n \"\"\"\n Cur.execute(query, (x_id,))\n\n Conn.commit()\n Conn.close()\n\n function = 'User Delete' # 분류 정보를 원하시는 텍스트로 변경해주세요.\n item = 'Delete user ' + x_id\n result = '성공'\n user = request.session.get('sessionid')\n now = datetime.now().replace(microsecond=0)\n date = now.strftime(\"%Y-%m-%d %H:%M:%S\")\n print(date)\n Xfactor_log = Xfactor_Log(\n log_func=function,\n log_item=item,\n log_result=result,\n log_user=user,\n log_date=date\n )\n Xfactor_log.save()\n\n return JsonResponse({'result': 'success'}, status=200) # 성공적으로 삭제되었을 때 응답\n except Exception as e:\n print(str(e)) # 에러 메시지 출력 (디버깅 용)\n return JsonResponse({'result': 'failure'}, status=400) # 삭제 중 오류가 발생했을 때 응답\n","repo_name":"XionITS/X-Factor-SM-nc","sub_path":"common/views_user.py","file_name":"views_user.py","file_ext":"py","file_size_in_byte":15196,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"13392801473","text":"from hierarchybuilder.topic_clustering import main_clustering as main_clustering\nfrom hierarchybuilder.expansions import parse_medical_data\nimport torch\nimport os\nimport sys\nfrom transformers import AutoTokenizer, AutoModel\nimport statistics\n\ncos = torch.nn.CosineSimilarity(dim=0, eps=1e-08)\nnlp = parse_medical_data.nlp\ndevice = torch.device(\"cpu\")\n# device = torch.device(\"cuda:1\" if torch.cuda.is_available() else \"cpu\")\nsapBert_tokenizer = AutoTokenizer.from_pretrained('cambridgeltl/SapBERT-from-PubMedBERT-fulltext')\nsapBert_model = AutoModel.from_pretrained('cambridgeltl/SapBERT-from-PubMedBERT-fulltext')\nmodel = sapBert_model\nmodel = model.eval()\n\ndict_span_to_lemma_lst = {}\ntopics_dict = {}\ndict_span_to_counter = {}\ndict_word_to_lemma = {}\ndict_lemma_to_synonyms = {}\ndict_longest_span_to_counter = {}\ndict_noun_lemma_to_synonyms = {}\ndict_noun_lemma_to_noun_words = {}\ndict_noun_lemma_to_counter = {}\ndict_noun_word_to_counter = {}\ndict_full_np_to_sentences = {}\nentries_number_limit = 50\nhost_and_port = \"127.0.0.1:5000\"\n\n\ndef initialize_data(examples, host_val, port_val, ignore_words, entries_number,\n device_type, has_umls_server):\n global topics_dict, dict_span_to_counter, dict_word_to_lemma, dict_lemma_to_synonyms, \\\n dict_longest_span_to_counter, dict_noun_lemma_to_synonyms, dict_noun_lemma_to_noun_words, \\\n dict_noun_lemma_to_counter, dict_noun_word_to_counter, entries_number_limit, device, model, \\\n host_and_port, dict_full_np_to_sentences\n host_and_port = \"http://\" + host_val + \":\" + str(port_val)\n if device_type:\n device = device_type\n model = model.to(device)\n model = model.eval()\n entries_number_limit = entries_number\n if ignore_words is None:\n ignore_words = set()\n collection_format_examples = parse_medical_data.get_examples_as_all_optional_answers_format(examples)\n topics_dict, dict_span_to_counter, dict_word_to_lemma, dict_lemma_to_synonyms, \\\n dict_longest_span_to_counter, dict_noun_lemma_to_synonyms, dict_noun_lemma_to_noun_words, dict_noun_lemma_to_counter, \\\n dict_noun_word_to_counter, dict_full_np_to_sentences = \\\n main_clustering.convert_examples_to_clustered_data(collection_format_examples, ignore_words, host_and_port,\n has_umls_server)\n dict_span_to_counter.update(dict_noun_word_to_counter)\n dict_span_to_counter.update(dict_noun_lemma_to_counter)\n\n\ndef dfs_for_cyclic(visited, helper, node):\n visited.append(node)\n helper.append(node)\n children = node.children\n for child in children:\n if child not in visited:\n ans = dfs_for_cyclic(visited, helper, child)\n if ans == True:\n print(child.span_lst)\n return True\n elif child in helper:\n print(child.span_lst)\n return True\n helper.remove(node)\n return False\n\n\ndef isCyclic(nodes_lst):\n visited = []\n helper = []\n for i in nodes_lst:\n if i not in visited:\n ans = dfs_for_cyclic(visited, helper, i)\n if ans == True:\n print(i.span_lst)\n return True\n return False\n\n\ndef update_nodes_labels(nodes_lst, visited=set()):\n labels_lst = set()\n for node in nodes_lst:\n if node in visited:\n continue\n visited.add(node)\n desc_labels = update_nodes_labels(node.children, visited)\n node.label_lst.update(desc_labels)\n labels_lst.update(node.label_lst)\n return labels_lst\n\n\ndef get_all_spans(np_object_lst, all_spans, visited=set()):\n for np_object in np_object_lst:\n if np_object in visited:\n continue\n visited.add(np_object)\n all_spans.update(np_object.span_lst)\n get_all_spans(np_object.children, all_spans, visited)\n\n\ndef dfs(visited, node):\n if node not in visited:\n visited.append(node)\n for neighbour in node.children:\n dfs(visited, neighbour)\n\n\ndef depth_dag(node, counter=0, visited=set()):\n max_depth = counter\n for child in node.children:\n current_max_depth = depth_dag(child, counter + 1, visited)\n if current_max_depth > max_depth:\n max_depth = current_max_depth\n return max_depth\n\n\ndef get_leaves(node, leaves, visited=set()):\n if node in visited:\n return\n visited.add(node)\n if not node.children:\n leaves.add(node)\n return\n for child in node.children:\n get_leaves(child, leaves, visited)\n\n\ndef get_all_nodes(nodes_lst, visited):\n for node in nodes_lst:\n if node in visited:\n continue\n visited.add(node)\n get_all_nodes(node.children, visited)\n\n\ndef calculation_for_paper(topic_object_lst, top_k_topics):\n visited = set()\n get_all_nodes(topic_object_lst, visited)\n all_dag_nodes = list(visited)\n print(\"number of nodes is \" + str(len(all_dag_nodes)))\n max_depth = 0\n for topic in topic_object_lst:\n depth = depth_dag(topic)\n if depth > max_depth:\n max_depth = depth\n print(\"the depth of the entire DAG is \" + str(max_depth))\n min_leaves = 10000\n max_leaves = 0\n total_leaves = 0\n max_depth = 0\n min_depth = 1000\n total_depth = 0\n all_leaves = set()\n for entry in top_k_topics:\n leaves = set()\n get_leaves(entry, leaves)\n all_leaves.update(leaves)\n leaves_number = len(leaves)\n total_leaves += leaves_number\n if leaves_number < min_leaves:\n min_leaves = leaves_number\n if leaves_number > max_leaves:\n max_leaves = leaves_number\n depth = depth_dag(entry)\n total_depth += depth\n if depth > max_depth:\n max_depth = depth\n if depth < min_depth:\n min_depth = depth\n # top k leaves\n print(\"average number of leaves is \")\n print(total_leaves / len(top_k_topics))\n print(\"minimal leaves for top k entry is \" + str(min_leaves))\n print(\"maximal leaves for top k entry is \" + str(max_leaves))\n # top k depth\n print(\"average number of depth is \")\n print(total_depth / len(top_k_topics))\n print(\"minimal depth for top k entry is \" + str(min_depth))\n print(\"maximal depth for top k entry is \" + str(max_depth))\n all_dag_nodes_in_top_k = set()\n get_all_nodes(top_k_topics, all_dag_nodes_in_top_k)\n number_of_children_array = []\n for node in all_dag_nodes_in_top_k:\n if node in top_k_topics:\n continue\n if node in all_leaves:\n continue\n number_of_children_array.append(len(node.children))\n # internal nodes\n print(\"minimal val of internal nodes is \" + str(min(number_of_children_array)))\n print(\"maximal val of internal nodes is \" + str(max(number_of_children_array)))\n print(\"average number of internal nodes \")\n print(len(number_of_children_array) / len(top_k_topics))\n ans = statistics.variance(number_of_children_array)\n print(\"The variance of list is : \")\n print(ans)\n","repo_name":"itayair/hierarchybuilder","sub_path":"hierarchybuilder/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":7037,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"34752256802","text":"import torch\nimport numpy as np\nimport torch.nn as nn\nfrom torch.autograd import Variable\nimport torch.nn.functional as F\nimport copy\nfrom stale import compute_stale_grad_alpha\n\nclass Architect(object):\n def __init__(self, model, args):\n self.network_momentum = args.momentum\n self.network_weight_decay = args.weight_decay\n self.model = model\n self.optimizer = torch.optim.Adam(self.model.arch_parameters(),\n lr=args.arch_learning_rate, betas=(0.5, 0.999), weight_decay=args.arch_weight_decay)\n self.baseline = None\n self.baseline_decay = args.arch_baseline_decay\n\n\n def step(self, epoch_acc, epoch_index_normal, epoch_index_reduce):\n self._compute_grad(self.model.alphas_normal, epoch_acc, epoch_index_normal)\n self._compute_grad(self.model.alphas_reduce, epoch_acc, epoch_index_reduce)\n self.optimizer.step()\n self.optimizer.zero_grad()\n\n def _compute_grad(self, alphas, accuracy_list, index_list):\n grad = torch.zeros(alphas.size())\n prob = F.softmax(alphas, dim=-1)\n rewards = self._compute_reward(accuracy_list)\n for client_idx in range(len(rewards)):\n reward = rewards[client_idx]\n index = index_list[client_idx]\n client_grad = torch.Tensor(prob.shape)\n client_grad.copy_(prob)\n # nabla _alpha { log(p(g_i)) } = (p_1, ..., p_i-1, ..., p_N)\n for edge_idx in range(client_grad.shape[0]):\n index_prob = client_grad[edge_idx][index[edge_idx]]\n client_grad[edge_idx][index[edge_idx]] = index_prob -1\n grad += reward * client_grad\n grad /= len(rewards)\n alphas.grad = grad\n\n def _compute_reward(self,accuracy_list):\n # scale accuracy to 0-1\n avg_acc = torch.mean(torch.Tensor(accuracy_list)) / 100\n if self.baseline is None:\n self.baseline = avg_acc\n else:\n self.baseline += self.baseline_decay * (avg_acc - self.baseline)\n # reward = accuracy - baseline\n return [accuracy_list[i]/100 - self.baseline for i in range(len(accuracy_list))]\n\n def stale_step(self, epoch_acc, epoch_index_normal, epoch_index_reduce, stale_alphas_normal, stale_alphas_reduce, stale_acc, stale_index_normal, stale_index_reduce):\n self._compute_stale_grad(self.model.alphas_normal, epoch_acc, epoch_index_normal, stale_alphas_normal, stale_acc, stale_index_normal)\n self._compute_stale_grad(self.model.alphas_reduce, epoch_acc, epoch_index_reduce, stale_alphas_reduce , stale_acc, stale_index_reduce)\n self.optimizer.step()\n self.optimizer.zero_grad()\n\n def _compute_stale_grad(self, alphas, accuracy_list, index_list, old_alphas, old_accuracy, old_index):\n grad = torch.zeros(alphas.size())\n prob = F.softmax(alphas, dim=-1)\n rewards = self._compute_reward(accuracy_list)\n for client_idx in range(len(rewards)):\n reward = rewards[client_idx]\n index = index_list[client_idx]\n client_grad = torch.Tensor(prob.shape)\n client_grad.copy_(prob)\n # nabla _alpha { log(p(g_i)) } = (p_1, ..., p_i-1, ..., p_N)\n for edge_idx in range(client_grad.shape[0]):\n index_prob = client_grad[edge_idx][index[edge_idx]]\n client_grad[edge_idx][index[edge_idx]] = index_prob -1\n grad += reward * client_grad\n\n # stale update\n old_reward = self._compute_reward(old_accuracy)\n for stale_idx in range(len(old_alphas)):\n stale_grad = compute_stale_grad_alpha(old_index[stale_idx], old_alphas[stale_idx], alphas)\n grad += old_reward[stale_idx] * stale_grad\n grad /= (len(rewards)+len(old_alphas))\n alphas.grad = grad\n\n\n\n\n\nif __name__ == '__main__':\n from model_search import Network\n class TMP:\n def __init__(self):\n self.momentum = 0.9\n self.weight_decay = 3E-4\n self.arch_learning_rate = 1E-3\n self.arch_weight_decay = 3e-4\n args = TMP()\n criterion = nn.CrossEntropyLoss()\n model = Network(16, 10, 8, criterion)\n architect = Architect(model,args)\n pass","repo_name":"dixiyao/CS385","sub_path":"project2/rl_federated_nas/architect.py","file_name":"architect.py","file_ext":"py","file_size_in_byte":3869,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"75561732352","text":"from itertools import combinations\nimport matplotlib.gridspec as grsp\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport random\nfrom SALib.sample import saltelli\nfrom SALib.analyze import sobol\nfrom scipy.special import binom\nimport seaborn as sns\nimport sys\nimport timeit\nimport pickle\nfrom scipy.stats import multivariate_normal\n\nstart_time = timeit.default_timer()\n\ndef f_emul(gp,poly,X_new,n_samples,seed,svm,ranges,N): \n\tX_whole = np.zeros((X_new.shape[0],X_new.shape[1]))\n\tfor i in range(X_new.shape[1]):\n\t\tX_whole[:,i] = X_new[:,i]/ranges[i,1]\n\n\tif svm != 0:\n\t\tlabel = svm.predict(X_whole)\n\n\t\tconta1 = 0\n\t\tcontameno1 = 0\n\t\tfor i in range(label.shape[0]):\n\t\t\tif label[i] == 1:\n\t\t\t\tconta1 = conta1 + 1\n\t\t\telse:\n\t\t\t\tcontameno1 = contameno1 + 1\n\n\t\tprint('With {} input points (N), X Sobol has dimensions: {}'.format(N,X_new.shape))\t\t\n\t\tprint('SVM makes {} predictions, which should be equal to {}, the number of X Sobol rows'.format(label.shape[0],X_whole.shape[0]))\n\t\tprint('SVM predicts {} points to be discarded over {} points, which is {} per cent'.format(contameno1,label.shape[0],np.round(contameno1/label.shape[0]*100,1) ) )\n\t\t# print('conta1 {}'.format(conta1))\n\t\t# print('contameno1 {}'.format(contameno1))\n\n\t\tlista = np.where(label==-1)[0]\n\n\tres, cov = gp.predict(X_new, return_cov=True)\n\tif svm != 0:\n\t\tres[lista,:] = np.zeros((len(lista),1)) \n\n\tmean = poly.predict(X_new)\n\tif svm != 0:\n\t\tmean[lista,:] = np.zeros((len(lista),1))\n\n\ty_sample = multivariate_normal.rvs( np.ndarray.flatten(res+mean) , cov, size=n_samples, random_state=seed).T\n\n\t# if etic == 'SIsvm':\n\t# \tfor jj in lista:\n\t# \t\ty_sample[jj,:] = np.zeros((1,n_samples))\n\n\treturn y_sample\n\n\nSEED = 8\n\ndef main():\n\tseed = SEED\n\trandom.seed(seed)\n\tnp.random.seed(seed)\n\n\t# Usage:\n\t# python3 run_gsa_mio.py ranges, num_par, GP, poly, folder to save indexes, num points Sobol sequence, SVM.\n\t# THe script can be called with 5 or 6 parameters, depending whether or not there's the SVM!\n\n\tranges = np.load(sys.argv[1])\n\t\n\tif int(sys.argv[2]) == 10: # EPI_ENDO - ENDO_EPI\n\t\tlabels = ['Scar radius','Scar depth','Scar conductivity','Internal bath','EHT thickness','EHT conductivity','CP thickness','CP conductivity','EHT-tissue contact area','Delta thickness'] # 10\n\tif int(sys.argv[2]) == 9: # TRANSMURAL\n\t\tlabels = ['Scar radius','Scar conductivity','Internal bath','EHT thickness','EHT conductivity','CP thickness','CP conductivity','EHT-tissue contact area','Delta thickness'] # 9\n\tif int(sys.argv[2]) == 8: # BLOCKED\n\t\tlabels = ['Scar radius','Internal bath','EHT thickness','EHT conductivity','CP thickness','CP conductivity','EHT-tissue contact area','Delta thickness'] # 8\n\tif int(sys.argv[2]) == 7: # FIXED\n\t\tlabels = ['Internal bath','EHT thickness','EHT conductivity','CP thickness','CP conductivity','EHT-tissue contact area','Delta thickness'] # 7\n\t\n\tGP_path = sys.argv[3]\n\tpoly_path = sys.argv[4]\n\n\ttag = sys.argv[5]\n\n\tsobol_seq_points = int(sys.argv[6])\n\n\tif len(sys.argv) == 8:\n\t\tsvm_path = sys.argv[7]\n\t\twith open(svm_path, 'rb') as f:\n\t\t\tsvm = pickle.load(f)\n\telse:\n\t\tsvm = 0\n\n\t#========================\n\t# GPE loading\n\t#========================\n\n\twith open(GP_path, 'rb') as f:\n\t\tgp = pickle.load(f)\n\n\twith open(poly_path, 'rb') as f:\n\t\tpoly = pickle.load(f)\n\t\n\t#========================\n\t# SA LIB\n\t#========================\n\tN = sobol_seq_points # deafult 1000, try 2000, 3000 and 5000 \n\tD = len(labels)\n\tn_draws = 1000\n\n\tI = ranges\n\tindex_i = labels\n\tindex_ij = ['({}, {})'.format(c[0], c[1]) for c in combinations(index_i, 2)]\n\n\tproblem = {\n\t\t'num_vars': D,\n\t\t'names': index_i,\n\t\t'bounds': I\n\t}\n\n\tX_sobol = saltelli.sample(problem, N, calc_second_order=True) # N x (2D + 2)\n\n\n\tY = f_emul(gp, poly, X_sobol, n_draws, seed, svm, I, N)\n\n\n\tST = np.zeros((0, D), dtype=float)\n\tS1 = np.zeros((0, D), dtype=float)\n\tS2 = np.zeros((0, int(binom(D, 2))), dtype=float)\n\tfor i in range(n_draws):\n\t\tS = sobol.analyze(problem, Y[:,i], calc_second_order=True, parallel=True, n_processors=24, seed=seed)\n\t\ttotal_order, first_order, (_, second_order) = sobol.Si_to_pandas_dict(S)\n\t\tST = np.vstack((ST, total_order['ST'].reshape(1, -1)))\n\t\tS1 = np.vstack((S1, first_order['S1'].reshape(1, -1)))\n\t\tS2 = np.vstack((S2, np.array(second_order['S2']).reshape(1, -1)))\n\n\tprint('GSA - Elapsed time: {:.1f} min'.format( (timeit.default_timer() - start_time)/60 ))\n\n\tnp.savetxt(str(tag) + '/STi.txt', ST, fmt='%.6f')\n\tnp.savetxt(str(tag) + '/Si.txt', S1, fmt='%.6f')\n\tnp.savetxt(str(tag) + '/Sij.txt', S2, fmt='%.6f')\n\n\tdf_ST = pd.DataFrame(data=ST, columns=index_i)\n\tdf_S1 = pd.DataFrame(data=S1, columns=index_i)\n\tdf_S2 = pd.DataFrame(data=S2, columns=index_ij)\n\n\t# gs = grsp.GridSpec(2, 2)\n\t# fig = plt.figure(figsize=(2*8.27, 4*11.69/3))\n\t# ax0 = fig.add_subplot(gs[0, 0])\n\t# ax1 = fig.add_subplot(gs[0, 1])\n\t# ax2 = fig.add_subplot(gs[1, :])\n\t# sns.boxplot(ax=ax0, data=df_S1)\n\t# sns.boxplot(ax=ax1, data=df_ST)\n\t# sns.boxplot(ax=ax2, data=df_S2)\n\t# ax0.set_ylim(0, 1)\n\t# ax0.set_title('First-order effect', fontweight='bold', fontsize=12)\n\t# ax0.set_xticklabels(ax0.get_xticklabels(), rotation=45, horizontalalignment='right')\n\t# ax1.set_ylim(0, 1)\n\t# ax1.set_title('Total effect', fontweight='bold', fontsize=12)\n\t# ax1.set_xticklabels(ax1.get_xticklabels(), rotation=45, horizontalalignment='right')\n\t# ax2.set_ylim(0, 1)\n\t# ax2.set_title('Second-order effect', fontweight='bold', fontsize=12)\n\t# ax2.set_xticklabels(ax2.get_xticklabels(), rotation=45, horizontalalignment='right')\n\t# plt.savefig('si_distr_salib_{}_{}_2.png'.format(N, etichette[lol]))\n\nif __name__ == '__main__':\n\tmain()\n\n","repo_name":"DamiFass/GP-GSA","sub_path":"run_GSA.py","file_name":"run_GSA.py","file_ext":"py","file_size_in_byte":5630,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"10553783390","text":"from __future__ import absolute_import, division, print_function\n\nimport codecs\nimport re\n\nimport numpy as np\n\nfrom six.moves import range\n\n\nclass Alphabet(object):\n def __init__(self, config_file):\n self._config_file = config_file\n self._label_to_str = []\n self._str_to_label = {}\n self._size = 0\n with codecs.open(config_file, 'r', 'utf-8') as fin:\n for line in fin:\n if line[0:2] == '\\\\#':\n line = '#\\n'\n elif line[0] == '#':\n continue\n self._label_to_str += line[:-1] # remove the line ending\n self._str_to_label[line[:-1]] = self._size\n self._size += 1\n\n def string_from_label(self, label):\n return self._label_to_str[label]\n\n def label_from_string(self, string):\n try:\n return self._str_to_label[string]\n except KeyError as e:\n raise KeyError(\n '''ERROR: Your transcripts contain characters which do not occur in data/alphabet.txt! Use util/check_characters.py to see what characters are in your {train,dev,test}.csv transcripts, and then add all these to data/alphabet.txt.'''\n ).with_traceback(e.__traceback__)\n\n def decode(self, labels):\n res = []\n for label in labels:\n res.append(label)\n return res\n\n # def decode(self, labels):\n # return labels\n\n def size(self):\n return self._size\n\n def config_file(self):\n return self._config_file\n\n\ndef text_to_char_array(original, alphabet):\n r\"\"\"\n Given a Python string ``original``, remove unsupported characters, map characters\n to integers and return a numpy array representing the processed string.\n \"\"\"\n return np.asarray([alphabet.label_from_string(c) for c in original])\n\n\n# The following code is from: http://hetland.org/coding/python/levenshtein.py\n\n# This is a straightforward implementation of a well-known algorithm, and thus\n# probably shouldn't be covered by copyright to begin with. But in case it is,\n# the author (Magnus Lie Hetland) has, to the extent possible under law,\n# dedicated all copyright and related and neighboring rights to this software\n# to the public domain worldwide, by distributing it under the CC0 license,\n# version 1.0. This software is distributed without any warranty. For more\n# information, see \n\ndef levenshtein(a, b):\n \"Calculates the Levenshtein distance between a and b.\"\n n, m = len(a), len(b)\n if n > m:\n # Make sure n <= m, to use O(min(n,m)) space\n a, b = b, a\n n, m = m, n\n\n current = list(range(n+1))\n for i in range(1, m+1):\n previous, current = current, [i]+[0]*n\n for j in range(1, n+1):\n add, delete = previous[j]+1, current[j-1]+1\n change = previous[j-1]\n if a[j-1] != b[i-1]:\n change = change + 1\n current[j] = min(add, delete, change)\n\n return current[n]\n\n# Validate and normalize transcriptions. Returns a cleaned version of the label\n# or None if it's invalid.\n\n\ndef validate_label(label):\n # For now we can only handle [a-z ']\n if re.search(r\"[0-9]|[(<\\[\\]&*{]\", label) is not None:\n return None\n\n label = label.replace(\"-\", \"\")\n label = label.replace(\"_\", \"\")\n label = label.replace(\".\", \"\")\n label = label.replace(\",\", \"\")\n label = label.replace(\"?\", \"\")\n label = label.replace(\"\\\"\", \"\")\n label = label.strip()\n label = label.lower()\n\n return label if label else None\n","repo_name":"Chung-I/tsm-rnnt","sub_path":"stt/data/text.py","file_name":"text.py","file_ext":"py","file_size_in_byte":3603,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"60"}
+{"seq_id":"44682235711","text":"from charms.reactive import (\n is_state, set_state,\n when, when_not,\n hookenv, hook,\n)\nfrom charmhelpers.core.hookenv import config\nfrom subprocess import check_call\nfrom charmhelpers.fetch import apt_install, apt_update\nfrom rubylib import bundle\nimport os\nimport stat\nimport pwd\n\ntry:\n from Crypto.PublicKey import RSA\nexcept ImportError:\n apt_update()\n apt_install('python3-crypto')\n from Crypto.PublicKey import RSA\n\n@when('ruby.available')\ndef setup_tests():\n apt_install(['git'])\n if not os.path.exists(config('app-path')):\n clone()\n bundle('install')\n gen_sshkey()\n\ndef clone():\n cmd = \"git clone {} {}\".format('https://github.com/hardening-io/tests-ssh-hardening.git', config('app-path'))\n res = check_call(cmd, shell=True)\n if res != 0:\n status_set('error', 'has a problem with git, try `resolved --retry')\n sys.exit(1)\n\ndef gen_sshkey():\n key = RSA.generate(2048)\n priv_key_file = '/home/ubuntu/.ssh/id_rsa'\n pub_key_file = '/home/ubuntu/.ssh/id_rsa.pub'\n uid = pwd.getpwnam(\"ubuntu\").pw_uid\n if os.path.exists(priv_key_file):\n return\n with open(priv_key_file, 'w') as content_file:\n os.chmod(priv_key_file, stat.S_IREAD)\n os.chown(priv_key_file, uid, -1)\n content_file.write(key.exportKey('PEM').decode('utf-8'))\n pubkey = key.publickey()\n with open(pub_key_file, 'w') as content_file:\n os.chown(pub_key_file, uid, -1)\n content_file.write(pubkey.exportKey('OpenSSH').decode('utf-8'))\n","repo_name":"CanonicalLtd/hardening-ssh-tests","sub_path":"reactive/hardening-ssh-tests.py","file_name":"hardening-ssh-tests.py","file_ext":"py","file_size_in_byte":1515,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"37479308793","text":"import pandas as pd\nimport os\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport argparse\nfrom random import randint\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.preprocessing import StandardScaler\nfrom pandas.plotting import register_matplotlib_converters\nregister_matplotlib_converters()\n\nimport techindic.indicator as ti\nimport utils.stats as stats\nimport econometric.utils as model\nimport rl.objective_function as of\nimport rl.policy as policy\n#import net.variational_autoencoder as vae\n\nfilename=\"./model_backups\"\n#stockfile=\"./btc_year_summary.csv\"\nstockfile=\"./bitcoin.csv\"\n\n# Parse arguments from command line\nparser = argparse.ArgumentParser(description='Reinforcement Learning Model.')\nparser.add_argument('-e', '--epochs', dest='epochs', type=int, default=500, help='The number of epochs to train the model')\nparser.add_argument('-t', '--timesteps', dest='past_timesteps', type=int, default=5, help='Numbers of past data to take into account')\nparser.add_argument('-l', '--learning_rate', dest='learning_rate', type=float, choices=[0.1, 0.01, 0.001], default=0.01, help='Learning rate')\nparser.add_argument('--objective_function', dest='objective_function', choices=['Sharpe', 'Dsharpe'], default='Dsharpe', help='Objective function to maximize')\nparser.add_argument('--weights_init', dest='weights_init', choices=['Ones', 'Xavier'], default='Ones', help='How to initialize features weights')\nparser.add_argument('-v', '--verbose', action='store_true', help='Display more log messages')\nparser.add_argument('--version', action='version', version='1.0')\n#parser.add_argument('Windows Length', dest='accumulate', action='store_const', const=sum, default=max, help='sum the integers (default: find the max)')\n\nargs = parser.parse_args()\n\n# Hyperparameter setting\nepochs = args.epochs\npast_timesteps = args.past_timesteps\nlearning_rate = args.learning_rate\nverbose = args.verbose\nobjective_function = args.objective_function\n\nprint(\"Chosen hyperparameters: epochs:{}, windows:{}, learning rate:{}, objective function:{}, weights init scheme:{}, verbose:{}\".format(epochs, past_timesteps, learning_rate, objective_function, args.weights_init, verbose))\n\n# Import data, replace unwanted coma for float numbers, and convert to numeric number\ndata = pd.read_csv(stockfile)\n\n# No need for the new dataset\ndata.iloc[:, 1:].replace(',','', regex=True, inplace=True)\ndata_ordered = data.iloc[::-1].reset_index(drop=True)\ndata_processed = pd.concat([data_ordered.iloc[:,0], data_ordered.iloc[:, 1:].apply(pd.to_numeric, errors='coerce')], axis=1)\n\n# So we need this in ordrer to keep the next lines unchanged\n#data_processed = data\n\nprint(data.head())\n\n# Add Technical Indicators as features\nrsi = ti.RSIInidcator(data_processed['Fermeture']).rsi()\nmacd = ti.MACDIndicator(data_processed['Fermeture']).macd()\nbb = ti.BollingerBands(data_processed['Fermeture']).bollinger()\nma = ti.MovingAverage(data_processed['Fermeture']).movingAverage()\nwr = ti.WilliamsRIndicator(data_processed['Haut'], data_processed['Bas'], data_processed['Fermeture']).wr()\n\ndata_with_TI = pd.concat([data_processed, rsi, macd, bb, ma, wr], axis=1)\n\n# Available Features\n# [['Ouverture', 'Haut', 'Bas', 'ma7', 'ma21', '26ema', '12ema', 'MACD', 'upper_band', 'lower_band', 'ema', 'wr']]\n\n#selected_feature = data_with_TI[['Fermeture', 'Ouverture', 'VWP', 'MACD', 'ema', 'wr']]\nselected_feature = data_with_TI[['Fermeture', 'Ouverture','MACD', 'ema', 'wr']]\n\n# Numbers of selected features * (past_timesteps + 1 (because index start at 0)) + 1=Last_position\nnb_features = (past_timesteps + 1) * selected_feature.shape[1] + 1\n\n\n# Parameters initialization using Xavier initialization for tanh activation function\n# Or just using basic all ones initialization\nif args.weights_init == \"Xavier\":\n # normal distribution with mean=0 and variance= sqrt(2/(fan-in+fan-out))\n xavier_weights=np.random.normal(0, np.sqrt(2/(nb_features+1)), nb_features)\n theta=xavier_weights.flatten()\nelif args.weights_init == \"Ones\":\n theta = np.ones(nb_features)\n\n# Initialize sharpe ratios\nsharpes = np.zeros(epochs)\n\n# Split train/test sets, without shuffle as it is a time serie.\nselected_feature_train, selected_feature_test = train_test_split(selected_feature, test_size=0.2, shuffle=False)\n# With Scaling\nscaler = StandardScaler() # Or scaler = MinMaxScaler()\nselected_feature_scaled = scaler.fit_transform(selected_feature)\nselected_feature_scaled = pd.DataFrame(selected_feature_scaled, columns=selected_feature.columns)\nselected_feature_train_scaled, selected_feature_test_scaled = train_test_split(selected_feature_scaled, test_size=0.2, shuffle=False)\n\n#vae_data = vae.VariationalAutoencoder(selected_feature_train_scaled, selected_feature_test_scaled).generate()\n\n# Train the model\nfor i in range(epochs):\n grad, sharpe, positions, returns = policy.DirectReinforcementLearning(selected_feature_train_scaled, past_timesteps, nb_features, theta).gradientAscent(objective_function)\n theta = theta + grad * learning_rate\n if verbose:\n print(\"epochs:{} -> Gradients are:{} - Params:{}\".format(i, grad, theta))\n print(\"Sharpe: {}\".format(sharpe))\n sharpes[i] = sharpe\n\nprint(\"------- Training is over -------\")\nwith open(filename, 'a+') as f:\n f.write(\"epochs:{} \\nwindows:{} \\nlearning_rate:{} \\nobjective_function:{} \\nweights_init_scheme:{} \\ntheta:{}\\n\\n\".format(epochs, past_timesteps, learning_rate, objective_function, args.weights_init, theta))\n f.close()\nprint(\"------- Model Hyperparameters and parameters saved -------\")\n\n# Display sharpe ratio improvements over epochs\nif verbose:\n plt.figure()\n pd.Series(sharpes).plot()\n plt.legend(['Sharpe ratio'])\n plt.show()\n'''\n# parameters weight after 500 epochs for btc/usd min. VERY SUCCESSFULL\ntheta = [1.14979244, 1.15079199, 1.1502789, 0.80698264, 1.15029399, 0.33909104, 1.14928899, 1.14973558, 1.14949215, 0.76181637, 1.14962482, 0.29303071, 1.14896727, 1.1492582, 1.14909631, 0.72275713, 1.14918728, 0.28653372, 1.14881231, 1.14896827, 1.14888284, 0.69221627, 1.14893813, 0.30126941, 1.14928387, 1.14895654, 1.14912066, 0.68105205, 1.14916945, 0.37206205, 1.14969238, 1.14940828, 1.14954657, 0.68395287, 1.1495189, 0.49042269, 1.45383504]\n'''\n\n# Adjust indexes for the test set\nselected_feature_test_scaled.reset_index(drop=True, inplace=True)\nselected_feature_test.reset_index(drop=True, inplace=True)\n\n\n# Add an ARIMA prediction as a feature. For test set only as ARIMA is used for predicting test set's closed price\n# Warning: add too much time to be viable, in an online manner.\n# Need to find a way to add it. Currently we can't because we have not a theta for it.\nadd_arima = False\nif(add_arima):\n y_train = selected_feature_train[\"Fermeture\"]\n y_test = selected_feature_test[\"Fermeture\"]\n arima = model.FitARIMA(y_train, y_test).get_arima_values()\n arima_scaled = scaler.fit_transform(arima)\n arima_scaled_series = pd.Series( (v[0] for v in arima_scaled) )\n print(\"Arima {}\".format(arima_scaled_series))\n arima_scaled_series.reset_index(drop=True, inplace=True)\n\n # Add scaled ARIMA values to feature dataframe\n #selected_feature_test_scaled['ARIMA'] = arima_scaled_series\n selected_feature_test_scaled = selected_feature_test_scaled.assign(ARIMA=arima_scaled_series.values)\n print(selected_feature_test_scaled)\n\n# Run the model with the found parameters, on the test set\ngrad, sharpe, positions, returns = policy.DirectReinforcementLearning(selected_feature_test_scaled, past_timesteps, nb_features, theta).gradientAscent(objective_function)\n\n# Get profits to estimate wealth\nadd_profits, add_returns, _= of.Returns(selected_feature_test_scaled['Fermeture'], pd.Series(positions).round(), 0.025).getAdditiveProfits()\n\n# Actualise parameters\ntheta = theta + grad * learning_rate\n\n# Plot the changing positions\n# Separate buy signals from sell signals\nchanging_positions = (pd.Series(positions).round()).diff()\nchanging_positions.fillna(0, inplace=True)\nxbuy = [i for i in range(len(changing_positions)) if changing_positions[i] > 0]\nxsell = [i for i in range(len(changing_positions)) if changing_positions[i] < 0]\nybuy = []\nysell = []\nfor i in range(len(selected_feature_test_scaled['Fermeture'])):\n for j in range(len(xbuy)):\n if i == xbuy[j]:\n ybuy.append(selected_feature_test_scaled['Fermeture'].iloc[i])\n\n for k in range(len(xsell)):\n if i == xsell[k]:\n ysell.append(selected_feature_test_scaled['Fermeture'].iloc[i])\n\n# Plot the results\nplt.figure()\nplt.plot(pd.Series(add_returns).cumsum(), label=\"RLModel Add returns\", linewidth=1)\nplt.plot((selected_feature_test_scaled['Fermeture'].diff()).cumsum(), label=\"Buy and Hold\", linewidth=1)\nplt.plot(selected_feature_test_scaled['Fermeture'], label=\"Closing Price\", linewidth=1)\nplt.scatter(xbuy, ybuy, s=7, c='red', label=\"Buy Signal\")\nplt.scatter(xsell, ysell, s=7, c='blue', label=\"Sell Signal\")\nplt.xlabel('Ticks')\nplt.ylabel('Cumulative Returns');\nplt.legend()\nplt.title(\"RL Model vs. Buy and Hold - Test Data\");\nplt.show()\n","repo_name":"CarliKevn/Deep-Learning","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":9108,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"35239800406","text":"#!/usr/bin/python3\n\"\"\"\nSend a POST request to the passed URL with an email as a parameter\n\"\"\"\nimport urllib.request\nimport urllib.parse\nimport sys\n\narg = sys.argv\nif __name__ == \"__main__\":\n url = arg[1]\n value = {}\n value['email'] = arg[2]\n data = urllib.parse.urlencode(value)\n data = data.encode('ascii')\n req = urllib.request.Request(url, data)\n with urllib.request.urlopen(req) as response:\n html = response.read()\n print(html.decode('utf-8'))\n","repo_name":"carvanino/alx-higher_level_programming","sub_path":"0x11-python-network_1/2-post_email.py","file_name":"2-post_email.py","file_ext":"py","file_size_in_byte":484,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"15924726225","text":"import requests\nimport datetime\nfrom bs4 import BeautifulSoup\nfrom data_sources import source\n\n@source('live', name='Canada')\ndef import_data():\n for result in import_news():\n yield result\n\n for result in import_gov():\n yield result\n\ndef import_news():\n import string\n sourceURL = \"https://www.ctvnews.ca/health/coronavirus/tracking-every-case-of-covid-19-in-canada-1.4852102\"\n jsonURL = \"https://stats.ctvnews.ca/covidDapi/getAllCovidData\"\n\n # referer is required for authorization\n headers = {\n \"referer\": \"https://www.ctvnews.ca/health/coronavirus/tracking-every-case-of-covid-19-in-canada-1.4852102\"\n }\n\n datapoints = []\n content = requests.get(jsonURL, headers=headers, timeout=10).json()\n for row in content:\n entryDate = datetime.datetime.strptime(row['date'], \"%Y-%m-%d\").date()\n provinces = row['data']\n for data in provinces:\n province = string.capwords(data['provinceLabel'])\n\n total = data['totalCases']\n recovered = data.get('recoveries', None)\n deaths = data.get('deaths', None)\n tests = data.get('totalTests', None)\n yield {\n \"entry_date\": entryDate,\n \"country\": \"Canada\",\n \"province\": province,\n \"total\": total,\n \"recovered\": recovered,\n \"deaths\": deaths,\n \"tests\": tests\n }\n\ndef import_gov():\n url = \"https://www.canada.ca/en/public-health/services/diseases/2019-novel-coronavirus-infection.html\"\n soup = BeautifulSoup(requests.get(url, timeout=10).text, 'html.parser')\n stats = soup.select(\"#dataTable tbody tr\")\n datapoints = []\n for row in stats:\n tds = row.select(\"td\")\n province = tds[0].text\n total = int(tds[1].text.replace(\",\", \"\"))\n deaths = int(tds[3].text.replace(\",\", \"\"))\n if province != \"Canada\":\n yield {\n 'country': \"Canada\",\n 'province': province,\n 'total': total,\n 'deaths': deaths\n }\n\nif __name__ == \"__main__\":\n import_data()","repo_name":"myfatemi04/Corona-Vision","sub_path":"data_collection/data_sources/north_america/canada.py","file_name":"canada.py","file_ext":"py","file_size_in_byte":2154,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"42471591100","text":"import argparse\nfrom typing import List, Dict, Any\n\n\ndef add(x, y):\n return x+y\n\n\ndef mul(x, y):\n return x*y\n\n\nopcodes: Dict[int, Any] = dict(((1, add), (2, mul), (99, None)))\n\n\ndef sol1(data: List[int]) -> int:\n i = 0\n size = len(data)\n while i < size:\n opcode = data[i]\n method = opcodes[opcode]\n if method is not None:\n data[data[i + 3]] = method(data[data[i + 1]], data[data[i + 2]])\n else:\n break\n i += 4\n return data[0]\n\n\ndef sol2(data: List[str]) -> int:\n expected_data = 19690720\n for noun in range(100):\n for verb in range(100):\n new_data = [int(x) for x in data[0].split(',')]\n new_data[1] = noun\n new_data[2] = verb\n if expected_data == sol1(new_data):\n return 100 * noun + verb\n return -1\n\n\ndef get_input(filename: str) -> List[str]:\n with open(filename) as f:\n lines = f.readlines()\n return lines\n\n\ndef main() -> int:\n parser = argparse.ArgumentParser()\n parser.add_argument('--filename', default='input.txt')\n args = parser.parse_args()\n data = get_input(args.filename)\n new_data = [int(x) for x in data[0].split(',')]\n new_data[1] = 12\n new_data[2] = 2\n print(sol1(new_data))\n print(sol2(data))\n return 0\n\n\nif __name__ == '__main__':\n exit(main())\n","repo_name":"yoavcaspi/aoc2019","sub_path":"day2/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":1359,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"23623140476","text":"import numpy as np\nimport matplotlib.pyplot as plt\nimport random\n\narr_1D = np.array([1,2,3,4])\n# print(arr_1D)\n# print(type(arr_1D))\n# print(arr_1D.ndim)\n\narr_2D = np.array([[1,2,3,4], [5,6,7,8]])\n# print(arr_2D)\n# print(type(arr_2D))\nprint(arr_2D[:, 2:3])\n\n# print(arr_2D.size)\n# print(arr_2D.shape)\n# print(arr_2D.dtype)\n\nones_arr = np.ones((4,6), dtype=int)\n# print(ones_arr)\n\nzero_arr = np.zeros((2,3), dtype=int)\n# print(zero_arr)\n\nempty_arr = np.empty((2,3))\n# print(empty_arr)\n\narange_arr = np.arange(1,17, 2)\n# print(my_arr)\n\nreshape_arr = np.arange(1,17).reshape(4,4)\n# print(reshape_arr)\n\nrevel_arr = np.arange(1,11).reshape(5,2).ravel()\n# print(revel_arr)\n\nlinespace_arr = np.linspace(1,11,12)\n# print(linespace_arr)\n\ntranspose_arr = arr_2D.transpose()\n# print(transpose_arr)\n\narr1 = np.arange(1,10).reshape(3,3)\narr2 = np.arange(1,10).reshape(3,3)\n\nadd = arr1 + arr2\nsub = arr1 - arr2\nmul = arr1 * arr2\nmatric_multiplication = arr1 @ arr2\n\nmax_digit = arr1.argmax(axis= 1)\n# print(max_digit)\n\nmini_digit = arr2.min(axis=0)\n# print(mini_digit)\n\nsum_of_arr = np.sum(arr1)\n# print(sum_of_arr)\n\nnp.mean(arr1)\nnp.sqrt(arr1)\nnp.std(arr1)\nnp.exp(arr1)\nnp.log(arr2)\nnp.log10(arr2)\n\narr_slicing = np.arange(1,101).reshape(10,10)\n\n# print(arr_slicing[:, 0:1])\n# print(arr_slicing[6,9])\n# print(arr_slicing[1:4, 1:4])\n# print(arr_slicing.shape)\n\nConnection_arr = np.concatenate((arr1, arr2), axis=1)\n# print(Connection_arr)\n\nsplit_arr = np.array([1,2,3,4,5])\n# print(np.split(split_arr, [1,3]))\n\nx_value = np.arange(0,3*np.pi, 0.1)\ny_sin = np.sin(x_value)\n\n# plt.plot(x_value, y_sin)\n# plt.show()\n\ny_cos = np.cos(x_value)\n# plt.plot(x_value, y_cos)\n# plt.show()\n\ny_tan = np.tan(x_value)\n# plt.plot(x_value,y_tan)\n# plt.show()\n\nrandom_arr = np.random.random((3,3))\n# print(random_arr)\n\nrandint_arr = np.random.randint(1,100, (5,4))\n# print(randint_arr)\n\nx = [1,2,3,4,5,6]\nchoice_arr = np.random.choice(x)\n# print(choice_arr)\n\npermutation_arr = np.random.permutation(x)\n# print(permutation_arr)\n\n# if we want to print same number than we'll use seed()\n\nnp.random.seed(10)\nx = [1,2,3,4,5,6]\nchoice_arr = np.random.choice(x)\n# print(choice_arr)\n\nperson_name = 'Ali Rehan Codes'\nstr1 = 'Hello'\n\n# print(np.char.add(person_name, str1))\n# print(np.char.center(str1, 60, fillchar=\"*\"))\n\nnp.char.lower(str1)\nnp.char.upper(str1)\nnp.char.title(str1)\nnp.char.split(person_name)\n\nstr2 = 'dmy'\n# print(np.char.join(':', str2))\n\n","repo_name":"AliRehanCodes/Python-NumPy","sub_path":"NumPy.py","file_name":"NumPy.py","file_ext":"py","file_size_in_byte":2415,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"31799312086","text":"import re\nimport json\nimport pickle\n\n#NOTE: Not very readable for the basic people U_U\n\n# Save python objects in binary format\n# !IMPORTANT: 'movies' can relate to series as well... That's IMDB folks!\n# - obj/acts.pkl: Dict where key is an actor, and the values the movies\n# - obj/directs.pkl: Dict where key is director, and the values the movies\n# - obj/movies_actors.pkl: All actors which participated in the movies from acts AND directs\n# - obj/movies_directors.pkl: All directors which participated in the movies from acts AND directs\ndef save_obj(obj, name):\n with open('obj/'+ name + '.pkl', 'wb') as f:\n pickle.dump(obj, f, pickle.HIGHEST_PROTOCOL)\n\ndef load_obj(name):\n with open('obj/' + name + '.pkl', 'rb') as f:\n return pickle.load(f)\n\nclass Act(object):\n def wsearch(w):\n return re.compile(r'\\b({0})\\b'.format(w), flags=re.IGNORECASE).search\n\n def __init__(self, first, last):\n self.f = first\n self.l = last\n\n def regexp(self):\n return (Act.wsearch(self.f), Act.wsearch(self.l))\n\n def __repr__(self):\n return f\"f: {self.f} l: {self.l}\"\n\n## Finding Actors Information!\n#act_n = [\"Uma Thurman\", \"Harvey Keitel\", \"Bill Murray\", \"Frances McDormand\"]\n#gen_st = \"acts\"\nact_n = [\"Quentin Tarantino\", \"Wes Anderson\"]\ngen_st = \"directs\"\n#act_man_f = \"actors.list\"\n#act_woman_f = \"actresses.list\"\nact_man_f = \"directors.list\"\nact_woman_f = \"t\"\nact_l = [(lambda n: Act(n[0], n[1]))([t.lower() for t in s.split()]) for s in act_n]\nact_d = {}\nmov_d = {}\n\ndef st_op(t):\n s = t\n p = s.find(\"(\")\n count = 0\n while p < len(s) and p + 1 < len(s) and not s[p+1].isdigit() and not s[p+1] == '?':\n tlist = list(s)\n tlist[p] = '_'\n s = \"\".join(tlist)\n p = s.find(\"(\")\n count += 1\n if count > 20:\n return len(t)\n return p\n\ndef find_person_movie(l, f):\n if l and l[0] != '\\t':\n for i, a in enumerate(act_l):\n if all(x(l[:l.find('\\t')].replace(\",\",\"\")) for x in a.regexp()) and not act_d.get(act_n[i], None):\n cs = [l[l.find('\\t'):st_op(l)].strip().replace(\"\\t\",\"\").replace(\"\\n\",\"\")]\n l = f.readline()\n while l and l[0] == '\\t':\n mv = l.strip().replace(\"\\t\", \"\").replace(\"\\n\",\"\")\n cs.append(mv[:st_op(mv)].strip())\n l = f.readline()\n act_d[act_n[i]] = cs\n return l\n\ndef act_d_builder(l, f):\n if l and l[0] != '\\t':\n print(l[:l.find('\\t')].strip())\n if not act_d.get(l[:l.find('\\t')].strip(), None):\n idx = l[:l.find('\\t')].strip()\n if not list(filter(None, idx)):\n return\n cs = [l[l.find('\\t'):st_op(l)].strip().replace(\"\\t\",\"\").replace(\"\\n\",\"\")]\n l = f.readline()\n while l and l[0] == '\\t':\n mv = l.strip().replace(\"\\t\", \"\").replace(\"\\n\",\"\")\n cs.append(mv[:st_op(mv)])\n l = f.readline()\n act_d[idx] = cs\n return l\n\ndef mov_actors_builder(l, f, movd):\n if l and l[0] != '\\t':\n idx = l[:l.find('\\t')].strip()\n if not list(filter(None, idx)):\n return\n cs = [l[l.find('\\t'):st_op(l)].strip().replace(\"\\t\",\"\").replace(\"\\n\",\"\")]\n l = f.readline()\n while l and l[0] == '\\t':\n mv = l.strip().replace(\"\\t\", \"\").replace(\"\\n\",\"\")\n mv = mv[:st_op(mv)].strip()\n if mv in movd:\n if not mov_d.get(mv, None):\n mov_d[mv] = set()\n mov_d[mv].add(idx)\n else:\n mov_d[mv].add(idx)\n if(mv == \"Plain Pleasures\"):\n print(f\"Adicionado em {mv} = {idx}\")\n l = f.readline()\n return l\n\ndef build_movieDict():\n actd = load_obj(\"acts\")\n drcd = load_obj(\"directs\")\n movd = set()\n for block in list(actd.values()) + list(drcd.values()):\n for mov in block:\n movd.add(mov)\n return movd\n\ndef print_dict(name):\n \"DON'T INCLUDE THE .pkl IN THE 'name' ARG!!\"\n t = load_obj(name)\n for k in t.keys():\n print(k)\n for m in t[k]:\n print(f\"\\t\\t{m}\")\n print(\"\")\n\ndef main():\n print_dict(\"directs\")\n exit()\n movd = build_movieDict()\n with open(act_man_f, \"r\", encoding='latin-1') as m, open(act_woman_f, \"r\", encoding='latin-1') as w:\n c = 0\n lm = m.readline()\n lw = w.readline()\n while lm or lw:\n lm, lw = find_person_movie(lm, m), find_person_movie(lw, w)\n lm, lw = m.readline(), w.readline()\n if c%50000 == 0:\n print(f\"{c}º iteration...\")\n #input(\"continua...\")\n c+=1\n\n #save_obj(act_d,\"directs\")\nif __name__ == '__main__':\n main()\n","repo_name":"robotenique/movies-ontology","sub_path":"imdbAnalyzer.py","file_name":"imdbAnalyzer.py","file_ext":"py","file_size_in_byte":4868,"program_lang":"python","lang":"en","doc_type":"code","stars":18,"dataset":"github-code","pt":"60"}
+{"seq_id":"21401711484","text":"with open('./day1_input.txt', 'r') as file:\r\n\tdata = []\r\n\tfor line in file:\r\n\t\t\tif line[0] == '+':\r\n\t\t\t\tdata.append(int(line[1:]))\r\n\t\t\tif line[0] == '-':\r\n\t\t\t\tdata.append(-int(line[1:]))\r\n\tfreq = 0\r\n\tseen_freqs = [0]\r\n\titteration = 0\r\n\twhile True:\r\n\t\t#print('Itteration: ' + str(itteration))\r\n\t\tif(itteration%1 == 0):\r\n\t\t\tprint(str(itteration) + ', ' + str(len(seen_freqs)))\r\n\t\tfor change in data:\r\n\t\t\tfreq += change\r\n\r\n\t\t\tif freq in seen_freqs:\r\n\t\t\t\tprint('Repeated frequency: ' + str(freq))\r\n\t\t\t\tquit()\r\n\t\t\telse:\r\n\t\t\t\tseen_freqs.append(freq)\r\n\t\t\t\t#print(len(seen_freqs))\r\n\r\n\t\titteration += 1\r\n\r\n\t\tprint(freq)\r\n\t\t#print(seen_freqs)\r\n","repo_name":"Emil-IT/AoC18","sub_path":"day1.py","file_name":"day1.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"28464629212","text":"import time\nfrom dotenv import dotenv_values\nfrom autockt_client import (\n start as start_client,\n Config,\n)\nfrom autockt_shared import auto_ckt_sim_hdl21, OpAmpInput, auto_ckt_sim\n\nENABLE_HTTPS = True\n\n\ndef main():\n \"\"\"\n Not picked up by pytest\n \"\"\"\n\n # Load the .env file\n env = dotenv_values()\n\n # And get the server URL\n THE_SERVER_URL = env.get(\"THE_SERVER_URL\", None)\n if not THE_SERVER_URL:\n raise ValueError(\"THE_SERVER_URL not set in .env file\")\n cfg = Config(server_url=\"34.83.44.225\", enable_https=ENABLE_HTTPS)\n\n start_client(cfg)\n\n to_test = OpAmpInput(3, 3, 3, 3, 3, 3, 1e-12)\n\n start_time = time.time()\n auto_ckt_sim(to_test)\n end_time = time.time()\n print(\"total time (auto_ckt_sim): \" + str(end_time - start_time))\n\n start_time = time.time()\n auto_ckt_sim_hdl21(to_test)\n end_time = time.time()\n print(\"total time (auto_ckt_sim_hdl21): \" + str(end_time - start_time))\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"BWRC-AMS-ML-Discovery/BwrcAmsMlDiscovery","sub_path":"scripts/speed_benchmark.py","file_name":"speed_benchmark.py","file_ext":"py","file_size_in_byte":995,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"86272167792","text":"import tenpy.utility.operators as op\r\nimport math\r\n\r\nclass CreateVector:\r\n\r\n def __init__(self, info = []):\r\n self.info = info\r\n\r\n########################################################################################################################\r\n### VECTOR PROPERTIES\r\n########################################################################################################################\r\n\r\n #Returns value of vector as an array\r\n def val(self):\r\n return self.info\r\n\r\n #Returns magnitude of vector\r\n def mag(self):\r\n mag = 0\r\n for i in range(len(self.info)):\r\n mag += math.pow(self.info[i], len(self.info))\r\n return math.sqrt(mag)\r\n\r\n #Returns angle of vector\r\n def angle(self):\r\n self.x = self.info[0]\r\n self.y = self.info[1]\r\n if len(self.info) > 2:\r\n self.z = self.info[2]\r\n\r\n if len(self.info) == 2:\r\n if (self.x < 0) and (self.y > 0):\r\n return math.pi + math.atan(self.y/self.x)\r\n elif (self.x < 0) and (self.y < 0):\r\n return math.pi - math.atan(self.y/self.x)\r\n elif (self.x > 0) and (self.y < 0):\r\n return 2*math.pi + math.atan(self.y/self.x)\r\n elif (self.x > 0) and (self.y > 0):\r\n return math.atan(self.y/self.x)\r\n elif (self.x > 0) and (self.y == 0):\r\n return 0.0\r\n elif (self.x < 0) and (self.y == 0):\r\n return math.pi\r\n elif (self.x == 0) and (self.y > 0):\r\n return math.pi/2\r\n elif (self.x == 0) and (self.y < 0):\r\n return 3*math.pi/2\r\n else:\r\n return math.atan(self.y/self.x)\r\n elif len(self.info) == 3:\r\n pass\r\n else:\r\n return \"Error: Tenpy does not support non-two dimensional vector angles\"\r\n\r\n def angle_between(self, other):\r\n angle = math.acos(self.dot_prod(other)/(self.mag()*other.mag()))\r\n return angle\r\n\r\n########################################################################################################################\r\n### GENERAL VECTOR EVALUATION\r\n########################################################################################################################\r\n\r\n # Operates on vectors by element\r\n def element_eval(self, other=[], sign=\"+\"):\r\n if len(self.info) == len(other.info):\r\n result = []\r\n for i in range(len(self.info)):\r\n result.append(op.get_operator_fn(sign)(self.info[i], other.info[i]))\r\n return CreateVector(result)\r\n else:\r\n return \"Vector Not Operatable\"\r\n\r\n########################################################################################################################\r\n### VECTOR MULTIPLICATION\r\n########################################################################################################################\r\n\r\n def dot_prod(self, other):\r\n prod = 0\r\n if len(self.info) == len(other.info):\r\n for i in range(len(self.info)):\r\n prod += self.info[i] * other.info[i]\r\n else:\r\n return \"Error: Vector Length Mismatch\"\r\n return prod\r\n\r\n def cross_prod(self, other):\r\n prod = []\r\n\r\n a = self.info\r\n b = other.info\r\n if len(self.info) == len(other.info):\r\n if len(self.info) == 3 and len(other.info) == 3:\r\n #angle = math.acos(self.dot_prod(other)/self.mag()*other.mag())\r\n prod.append(a[1]*b[2]-a[2]*b[1])\r\n prod.append(a[2]*b[0]-a[0]*b[2])\r\n prod.append(a[0]*b[1]-a[1]*b[0])\r\n else:\r\n return \"Sorry, cross product is only applicable for 3d vectors\"\r\n else:\r\n return \"Error: Vector Length Mismatch\"\r\n return CreateVector(prod)\r\n\r\n########################################################################################################################\r\n### VECTOR RELATIONS\r\n########################################################################################################################\r\n\r\n def is_parallel(self, other):\r\n tolerance = 2.0e-5\r\n if (self.angle_between(other) < 0.0+tolerance):\r\n print(self.angle_between(other))\r\n return True\r\n else:\r\n return False\r\n\r\n def is_orthogonal(self, other):\r\n if (self.dot_prod(other) == 0):\r\n return True\r\n else:\r\n return False","repo_name":"SoftLocked/Tenpy","sub_path":"tenpy/vector.py","file_name":"vector.py","file_ext":"py","file_size_in_byte":4548,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"19892667142","text":"class Solution(object):\n def hammingWeight(self, n):\n \"\"\"\n :type n: int\n :rtype: int\n \"\"\"\n count = 0\n flag = 1\n\n while flag <= n:\n if n & flag:\n count += 1\n flag = flag << 1\n print(flag)\n print(count)\n\n return count\n\n def hammingWeight2(self, n):\n \"\"\"\n :type n: int\n :rtype: int\n \"\"\"\n count = 0\n\n while n:\n count += 1\n n = (n - 1) & n\n\n return count\n\n\ns = Solution()\ns.hammingWeight(0o0000000000000000000000000001011)\n","repo_name":"Dooooooooo21/leetcode","sub_path":"offer/15.py","file_name":"15.py","file_ext":"py","file_size_in_byte":606,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"41382942805","text":"import sys\nfrom PIL import Image\n\n# check for args\nif len(sys.argv) < 7:\n print(' ')\n print('Incorrect arguments provided.')\n exit()\n\n# get original\nimage_original = Image.open( sys.argv[1] )\n# make sure its rgba\nimage_original = image_original.convert('RGBA')\n\n# calc crop args\nx_start = int( sys.argv[3] )\ny_start = int( sys.argv[4] )\nx_end = int( sys.argv[3] ) + int( sys.argv[5] )\ny_end = int( sys.argv[4] ) + int( sys.argv[6] )\n\n# crop\nimage_crop = image_original.crop( ( x_start, y_start, x_end, y_end ) )\n\n# save crop\nimage_crop.save( sys.argv[2] )\nprint(' ')\nprint(f\"Cropped image stored as {sys.argv[2]}\")\n\n","repo_name":"Brugman/dgg-place-image-tools","sub_path":"crop.py","file_name":"crop.py","file_ext":"py","file_size_in_byte":625,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"38762629787","text":"# indexation : 0 = lieu vide, 1 = arbre, 2 = arbre en feu : rajouter un état final \"brûlé\" ? (totalement optionnel ceci dit)\n# manque de jolies couleurs QQ (trouver LA colormap des familles) (ou coloriser tout ça après coup aussi ça marche bien)\n\n# cf. propagation directionnelle pour les bails ajoutés (vent selon une direction, humidité, phénomène d'étincelle)\n\nfrom matplotlib.pyplot import matshow\nimport matplotlib as mpl\nimport matplotlib.pylab as plt\nimport matplotlib.animation as animation\nimport random\nimport numpy as np\n\n# 1) CREATION DE LA MATRICE ASSOCIEE A LA FORET\n \ndef bool_p(p):\n # renvoie True avec une probabilite p et False avec une probabilité 1-p\n return random.random() <= p\n \ndef matgen(n,m,d):\n # cree une forest de dimensions n*m avec des arbres places aléatoirements à une densité d (facteur de percolation par site ici)\n forest = np.zeros((n,m))\n for i in range(n):\n for j in range(m):\n if bool_p(d):\n forest[i,j] = 1.\n else:\n forest[i,j] = 0.\n return forest\n \n \n# 2) FONCTIONS UTILES\n\np_l = 1. # proba de percolation par lien, modifiée par l'utilisateur en cas de besoin\nw_dir = 'null' # direction éventuelle du vent\n \n\"\"\" \ndef burn_spot_R(forest): # nécessite d'être fix... pb de float ?\n # met le feu à un arbre aléatoirement\n n,m = forest.shape\n x=random.randint(0,n)\n y=random.randint(0,m)\n while(forest[x,y] != 1.):\n x=random.randint(0,n)\n y=random.randint(0,m)\n forest[x,y] = 2.F\n return(forest) \"\"\"\n\ndef start_R(forest): # alternative, renvoie un couple (i,j) tel que la case soit verte (au hasard)\n n,m = forest.shape\n x=random.randint(0,n-1)\n y=random.randint(0,m-1)\n while(forest[x,y] != 1.):\n x=random.randint(0,n-1)\n y=random.randint(0,m-1)\n return(x,y)\n\ndef burn_spot(forest,i,j):\n # met le feu à l'arbre en position (i,j)\n if forest[i,j] == 1.:\n forest[i,j] = 2. # on fout le feu la case d'indice (i,j)\n return forest\n \n\"\"\" def neighbors_from(x,y,forest): # renvoie un voisinage à 4 ou 8 cases dans la mesure du possible, du lieu en question\n # n,m = forest.shape\n # return [(x, y + 1 if y + 1 < SIZE else 0), (x, y - 1), (x + 1 if x + 1 < SIZE else 0, y),(x - 1, y)]\n \"\"\"\n \ndef nextToFire(forest,i,j):\n # existence d'un arbre en feu au voisinage de l'arbre (i,j)\n # tenter d'utiliser neighbors_from plutôt ? revoir le voisinage\n \n n,m=forest.shape\n if forest[i,j] == 1.:\n if (i > 0 and forest[i - 1,j] == 2.):\n return True\n if (i < n - 1 and forest[i + 1,j] == 2.):\n return True\n if (j > 0 and forest[i,j - 1] == 2.):\n return True\n if (j < m - 1 and forest[i,j + 1] == 2.):\n return True\n return False\n \n\"\"\" Autre procédé de vérification pour les alentours...\n\n if forest[i,j] == 1.:\n for y in range(max(0,i-1),min(n,i+2)):\n if forest[y,j] == 2.:\n return True\n for x in range(max(0,j-1),min(m,j+2)):\n if forest[i,x] == 2.:\n return True\n return False\n \n\"\"\"\n \ndef propagateFire(forest):\n \"\"\"les arbres qui peuvent bruler autour d'un arbre en feu prennent feu\n rq. : on se place dans un cadre de percolation par site, la probabilité variante est celle de densité de placement, pas celle d'ouverture des liens dans L^d...\n ainsi, un arbre à proximité du feu prend systématiquement feu mskn\n pour tenir compte du phénomène de percolation par lien, implémenter une probabilité que le lien entre l'arbre en question et le voisin soit ouvert \"\"\"\n \n n,m=forest.shape # en pratique toujours une grille carrée pour simplifier les choses\n for i in range(n): # densité brute et méchante ici (percolation par site)\n for j in range(m):\n if nextToFire(forest,i,j):\n if random.random() <= p_l: # intervention des liens du graphe ouverts / fermés ici \n forest[i,j] = 2.\n return forest\n \ndef stillOnFire(forest):\n # vérifie s'il existe encore un arbre susceptible de cramer\n n,m=forest.shape\n for i in range(n):\n for j in range(m):\n if nextToFire(forest,i,j):\n return True\n return False\n \ndef burnForest(forest,i,j):\n # démarre le feu aux coordonnées (i,j) et propage le feu jusqu'à ce que ça ne soit plus possible\n forest = burn_spot(forest,i,j)\n while stillOnFire(forest):\n forest = propagateFire(forest)\n return forest\n\ndef count(forest,f): # compte le nombre de zones indicées f\n n,m = forest.shape\n C = 0 # compteur de zones \n for k in range(n):\n for i in range(m):\n if forest[k,i] == f:\n C+=1\n return C\n\n# 2.2) COEUR DU PROGRAMME ET STATS\n\ndef simulation(n,m,d,p,mode,w): # rajouter un argument p_l pour la percolation par lien ici (n,m = dimensions, d,p = densité et proba de lien, mode = animé ou non, w = direction du vent (ou non))\n\n p_l = p # percolation par lien définie\n w_dir = w\n forest = matgen(n,m,d) \n green = count(forest,1.) # nb d'arbres à l'état initial\n void = count(forest,0.) # nb de zones vide au départ\n \n i,j = start_R(forest)\n if (mode == 1):\n forest = animate(forest,i,j) # procedé de percolation animé\n elif (mode == 0): # pour effectuer des centaines d'essais, mieux vaux désactiver l'animation\n forest = animate_nofilm(forest,i,j) # sans image\n \n bnt = count(forest,2.) # nb d'arbres brûlés\n bntPA = bnt / (n*m) # proportion brûlé / total\n bntPR = bnt / green # proportion brûlé / nb d'arbres qu'il y avait au début mskn\n \n return([bnt,bntPA,bntPR]) # certaines expériences à d > 0.5 sont plutôt étranges...\n\n\ndef stat_density(n,m,p): # STAT : proba par sites (paramètre variant : densité d)\n d = 0.1 # d va évoluer de 0.1 à 0.95 par pas de 0.05\n result=[] # matrice qui contiendra les listes (proba, bnt, bntPA et PR)\n while (d <= 0.95):\n interm=[] # contient les résultats intermédiaires dont on va prendre la moyenne à la fin\n k = 1\n while (k <= 100): # on fait 100 essais par valeur de densité d\n interm.append(simulation(n,m,d,p,0))\n k+=1\n # ici, interm contient [[brulé à l'essai 1, proportions à l'essai 1],[brulé à l'essai 2, proportions à l'essai 2],...]\n bntM = 0\n bntPAM = 0\n bntPRM = 0\n for k in range(len(interm)):\n bntM += interm[k][0]\n bntPAM += interm[k][1]\n bntPRM += interm[k][2]\n bntM /= len(interm)\n bntPAM /= len(interm)\n bntPRM /= len(interm)\n \n result.append([d,bntM,bntPAM,bntPRM]) # result contient un tableau avec densité (variante), nb de brulés et proportions\n d += 0.05\n \n return(result)\n\n\"\"\" def stat_areasonfire(n,m,p): # STAT : en percolation par sites, évalue le nb de zones en feu en fonction du nb d'étapes\n d = 0.6 # d évolue cette fois de 0.6 à 0.9 par pas de 0.05\n result=[] # la liste va contenir cette fois le nombre d'arbres en feu à chaque étape en moyenne sur 20 essais cette fois\n while (d <= 0.95):\n interm=[]\n k = 1\n while (k <= 100):\n F = matgen(n,m,d)\n interm.append \"\"\"\n \n# 3) ANIMATION # ajouter cmap=plt.cm.nom_de_la_map pour les couleurs dans matshow\n \ndef animate(forest,i,j):\n fig = plt.figure() # nouvelle figure\n film = []\n # Initialisation\n forestOnFire = burn_spot(forest,i,j)\n film.append([matshow(forestOnFire, fignum=False, animated=True)])\n plt.draw()\n \n # Propagation\n while stillOnFire(forest):\n forestOnFire = propagateFire(burn_spot(forest,i,j))\n film.append([matshow(forestOnFire, fignum=False, animated=True)])\n plt.draw()\n \n # Animation\n ani = animation.ArtistAnimation(fig, film, interval=100, blit=True, repeat=False)\n \n plt.draw()\n plt.show()\n \n return(forest)\n \ndef animate_nofilm(forest,i,j): # pour les stats\n forestOnFire = burn_spot(forest,i,j)\n while stillOnFire(forest):\n forestOnFire = propagateFire(burn_spot(forest,i,j))\n \n return (forest)\n \n ","repo_name":"dylanankrah/fire-prediction-TIPE","sub_path":"Code source/sauvegardes/source principale (sauvegarde, propagation isotrope).py","file_name":"source principale (sauvegarde, propagation isotrope).py","file_ext":"py","file_size_in_byte":8312,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"42587209990","text":"from abc import ABC\n\nimport torch\nimport torch.nn as nn\n\n\nclass BimodalFusion(nn.Module, ABC):\n \"\"\"Bimodal fusion combines features from different modalities into\n a single tensor.\n\n The input modalities' feature tensors are expected to have matching\n sizes [N x C_1] and [N x C_2]. For residual fusion, we further\n require C_1 = C_2.\n\n By convention, the second features are fused into the first, main\n modality. This matters as the output format will match that of the\n main modality\n \"\"\"\n\n MODES = ['residual', 'concatenation', 'both', 'modality']\n\n def __init__(self, mode='residual', **kwargs):\n super(BimodalFusion, self).__init__()\n self.mode = mode\n if self.mode == 'residual':\n self.f = lambda a, b: a + b\n elif self.mode == 'concatenation':\n self.f = lambda a, b: torch.cat((a, b), dim=-1)\n elif self.mode == 'both':\n self.f = lambda a, b: torch.cat((a, a + b), dim=-1)\n elif self.mode == 'modality':\n self.f = lambda a, b: b\n else:\n raise NotImplementedError(\n f\"Unknown fusion mode='{mode}'. Please choose among \"\n f\"supported modes: {self.MODES}.\")\n\n def forward(self, x_main, x_mod):\n if x_main is None:\n return x_mod\n if x_mod is None:\n return x_main\n\n # If the x_mod is a sparse tensor, we only keep its features\n x_mod = x_mod if isinstance(x_mod, torch.Tensor) else x_mod.F\n\n # Update the x_main while respecting its format\n x_main = self.f(x_main, x_mod)\n\n return x_main\n\n def extra_repr(self) -> str:\n return f\"mode={self.mode}\"\n","repo_name":"drprojects/DeepViewAgg","sub_path":"torch_points3d/modules/multimodal/fusion.py","file_name":"fusion.py","file_ext":"py","file_size_in_byte":1700,"program_lang":"python","lang":"en","doc_type":"code","stars":205,"dataset":"github-code","pt":"60"}
+{"seq_id":"31436347534","text":"import pyopencl as cl\nimport pyopencl.array as cl_array\nfrom pyopencl.tools import SVMAllocator, SVMPool\nimport numpy as np\n\nn = 50000\na = np.random.rand(n).astype(np.float32)\nb = np.random.rand(n).astype(np.float32)\n\n\nctx = cl.create_some_context()\nqueue = cl.CommandQueue(ctx)\n\nalloc = SVMAllocator(ctx, alignment=0, queue=queue)\nalloc = SVMPool(alloc)\n\na_dev = cl_array.to_device(queue, a, allocator=alloc)\nb_dev = cl_array.to_device(queue, b, allocator=alloc)\ndest_dev = cl_array.empty_like(a_dev)\n\nprg = cl.Program(ctx, \"\"\"\n __kernel void sum(__global const float *a,\n __global const float *b, __global float *c)\n {\n int gid = get_global_id(0);\n c[gid] = a[gid] + b[gid];\n }\n \"\"\").build()\n\nknl = prg.sum\nknl(queue, a.shape, None, a_dev.data, b_dev.data, dest_dev.data)\n\nnp.testing.assert_allclose(dest_dev.get(), (a_dev+b_dev).get())\n","repo_name":"inducer/pyopencl","sub_path":"examples/demo_array_svm.py","file_name":"demo_array_svm.py","file_ext":"py","file_size_in_byte":864,"program_lang":"python","lang":"en","doc_type":"code","stars":998,"dataset":"github-code","pt":"60"}
+{"seq_id":"28959390151","text":"import numpy as np\nimport ADmetrics as adm\nimport tensorflow as tf\nfrom sklearn import metrics\nimport tensorflow_probability as tfp\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.datasets import make_classification\nfrom sklearn.model_selection import train_test_split\n\ntfd = tfp.distributions\ntfpl = tfp.layers\ntfk = tf.keras\ntfkl = tf.keras.layers\n\ntf.keras.backend.set_floatx('float64')\n\n\nclass MargMLVIEstimation(tf.keras.Model):\n\n \"\"\" Marginalized Maximum Likelihood Estimation using Black-Box Variational Inference (Bayes) \"\"\"\n\n def __init__(self, n_units, n_features, n_classes, dense_layer_type, name='theta', **kwargs):\n super(MargMLVIEstimation, self).__init__(name=name, **kwargs)\n self.n_units = n_units\n self.n_classes = n_classes\n self.n_features = n_features\n self.dense_layer_type = dense_layer_type\n\n c = np.log(np.expm1(1.))\n scale = 1e-5 + tf.nn.softplus(c)\n\n if self.dense_layer_type.lower() == \"pw-re\".lower() \\\n or self.dense_layer_type.lower() == \"pw-li\".lower():\n if self.dense_layer_type.lower() == \"pw-re\".lower():\n self.dense_1 = tfpl.DenseReparameterization(self.n_features, activation=tf.nn.relu,\n trainable=True, name='pw_dense_1',)\n else:\n self.dense_1 = tfpl.DenseReparameterization(self.n_features, activation=None,\n trainable=True, name='pw_dense_1',)\n # In order to be able to compute the log likelihood\n # a distribution is needed, and it is instantiated as follow:\n self.pz_x = tfpl.DistributionLambda(lambda t: tfd.Normal(loc=t, scale=scale,),\n trainable=True, name='pz_x')\n self.py_x_z = tfpl.DenseReparameterization(self.n_classes,\n activation=None, name='py_x_z')\n\n elif self.dense_layer_type.lower() == \"fo-re\".lower()\\\n or self.dense_layer_type.lower() == \"fo-li\".lower():\n if self.dense_layer_type.lower() == \"fo-re\".lower():\n # With Relu activation\n self.dense_1 = tfpl.DenseFlipout(self.n_features, activation=tf.nn.relu,\n name='fl_dense_1')\n else:\n # With Linear activation\n self.dense_1 = tfpl.DenseFlipout(self.n_features, activation=None,\n name='fl_dense_1')\n\n # In order to be able to compute the log likelihood\n # a distribution is needed, and it is instantiated as follow:\n self.pz_x = tfpl.DistributionLambda(lambda t: tfd.Normal(loc=t, scale=scale, ),\n trainable=True, name='pz_x')\n self.py_x_z = tfpl.DenseFlipout(self.n_classes, activation=None,\n name='py_x_z')\n\n def call(self, inputs, training=False):\n x = self.dense_1(inputs, training=training)\n if training:\n pz_x = self.pz_x(x, training=training)\n py_x_z_logits = self.py_x_z(x, training=training)\n return pz_x, py_x_z_logits\n # py_x_z_logits\n return self.py_x_z(x, training=training)\n\n\ndef apply_mmle(n_units, n_features, n_classes, n_epochs, batch_size, learning_rate,\n dense_layer_type='path_wise', x_train=None, y_train=None, x_val=None,\n y_val=None, x_test=None, y_test=None, verbose=False):\n\n if x_train is not None:\n train_dataset = tf.data.Dataset.from_tensor_slices(\n (x_train, y_train)).batch(batch_size)\n else:\n print(\"No training set is provided!\")\n return None\n\n if x_val is not None:\n valid_dataset = tf.data.Dataset.from_tensor_slices(\n (x_val, y_val)).batch(batch_size)\n\n if x_test is not None:\n test_dataset = tf.data.Dataset.from_tensor_slices(\n (x_test, y_test)).batch(batch_size)\n\n # metrics for monitoring the training and validation procedure\n train_loss = tf.keras.metrics.Mean(name='train_loss')\n train_accuracy = tf.keras.metrics.BinaryAccuracy(name='train_accuracy')\n\n test_loss = tf.keras.metrics.Mean(name='test_loss')\n test_accuracy = tf.keras.metrics.BinaryAccuracy(name='test_accuracy')\n\n model_adv = MargMLVIEstimation(n_units=n_units, n_features=n_features,\n n_classes=n_classes,\n dense_layer_type=dense_layer_type)\n\n optimizer = tf.optimizers.Adam(learning_rate=learning_rate)\n\n @tf.function\n def training_step(x_batch, y_batch, model):\n with tf.GradientTape() as tape:\n\n dist_pz_x, py_x_z_logits = model(x_batch, training=True)\n\n neg_log_lik_pz_x = -tf.reduce_sum(\n dist_pz_x.log_prob(x_batch))\n\n neg_log_lik_py_x_z = tf.nn.softmax_cross_entropy_with_logits(\n labels=y_batch, logits=py_x_z_logits)\n\n neg_log_lik_py_x_z_cls = -tf.reduce_sum(\n tf.nn.softmax_cross_entropy_with_logits(labels=y_batch, logits=py_x_z_logits))\n kl_loss = sum(model.losses) # /(x_train.shape[0])\n\n total_loss = tf.math.multiply(neg_log_lik_pz_x, neg_log_lik_py_x_z) + neg_log_lik_py_x_z_cls # + kl_loss\n\n gradients = tape.gradient(total_loss, model.trainable_variables)\n optimizer.apply_gradients(zip(gradients, model.trainable_variables))\n\n _ = train_loss.update_state(total_loss)\n predictions = tf.nn.softmax(py_x_z_logits)\n _ = train_accuracy.update_state(y_batch, predictions)\n\n @tf.function\n def testing_step(x_batch_t, y_batch_t, model):\n dist_pz_x_t, py_x_z_logits_t = model(x_batch_t, training=True)\n neg_log_lik_pz_x_t = -tf.reduce_sum(dist_pz_x_t.log_prob(x_batch_t))\n neg_log_lik_py_x_z_t = tf.nn.softmax_cross_entropy_with_logits(labels=y_batch_t, logits=py_x_z_logits_t)\n neg_log_lik_py_x_z_cls_t = -tf.reduce_sum(\n tf.nn.softmax_cross_entropy_with_logits(labels=y_batch_t, logits=py_x_z_logits_t))\n kl_loss = sum(model.losses) # /(x_train.shape[0])\n\n total_loss_t = tf.math.multiply(neg_log_lik_pz_x_t, neg_log_lik_py_x_z_t) + neg_log_lik_py_x_z_cls_t\n\n _ = test_loss.update_state(total_loss_t)\n predictions_t = tf.nn.softmax(py_x_z_logits_t)\n _ = test_accuracy.update_state(y_batch_t, predictions_t)\n\n for epoch in range(n_epochs):\n\n # Reset the metrics at the start of the next epoch\n train_loss.reset_states()\n train_accuracy.reset_states()\n test_loss.reset_states()\n test_accuracy.reset_states()\n\n for step, (x_batch, y_batch) in enumerate(train_dataset):\n training_step(x_batch=x_batch, y_batch=y_batch, model=model_adv,)\n\n if x_val is not None:\n for step, (x_batch_v, y_batch_v) in enumerate(valid_dataset):\n testing_step(x_batch_t=x_batch_v, y_batch_t=y_batch_v, model=model_adv, )\n\n if x_test is not None:\n for step, (x_batch_t, y_batch_t) in enumerate(test_dataset):\n testing_step(x_batch_t=x_batch_t, y_batch_t=y_batch_t, model=model_adv, )\n\n # Log every 100 batches.\n if step % 100 == 0 and verbose:\n template = 'Epoch {}, Train Loss: {}, Train Accuracy: {}, Valid Loss: {}, Valid Accuracy: {}'\n print(template.format(epoch + 1,\n train_loss.result(),\n train_accuracy.result(),\n test_loss.result(),\n test_accuracy.result()))\n return model_adv\n\n\nif __name__ == '__main__':\n\n n_units = 10\n n_epochs = 500\n n_classes = 2\n n_features = 10\n batch_size = 100\n n_samples = 1000 # 100\n learning_rate = 1e-2\n\n inside_call = True\n dense_layer_type = 'path_wise'\n\n # Generate synthetic data / load data sets\n x_in, y_in = make_classification(n_samples=n_samples, n_features=n_features, n_informative=n_classes, n_redundant=0,\n n_repeated=0, n_classes=n_classes, n_clusters_per_class=1,\n weights=[0.2, 0.8], flip_y=0.4, class_sep=1.0, hypercube=True,\n shift=0.0, scale=1.0, shuffle=True, random_state=42)\n print(\"y_in:\", set(y_in))\n\n # Normalizing the data points\n x_in = np.divide(x_in, np.ptp(x_in, axis=0))\n x_in = x_in.astype('float64')\n y_in = y_in.astype('float64').reshape(-1, 1)\n\n one_hot_encoder = OneHotEncoder(sparse=False)\n y_in = one_hot_encoder.fit_transform(y_in)\n y_in = y_in.astype('float64')\n\n x_train, x_test, y_train, y_test = train_test_split(x_in, y_in, test_size=0.4, random_state=42, shuffle=True)\n x_test, x_val, y_test, y_val = train_test_split(x_test, y_test, test_size=0.5, random_state=42, shuffle=True)\n\n print(\"shapes:\", x_train.shape, y_train.shape, x_test.shape, y_test.shape, x_val.shape, y_val.shape)\n\n x_min = np.min(x_train, axis=0)\n x_max = np.max(x_train, axis=0)\n x_range = x_max - x_min\n\n model_adv = apply_mmle(n_units=n_units, n_features=n_features, n_classes=n_classes,\n n_epochs=n_epochs, batch_size=batch_size, learning_rate=learning_rate,\n dense_layer_type=dense_layer_type, x_train=x_train, y_train=y_train,\n x_val=x_val, y_val=y_val, x_test=None, y_test=None, verbose=False)\n\n py_x_z_logits = model_adv(x_test, training=False)\n py_x_z_probs = tf.nn.softmax(py_x_z_logits)\n labels_pred = tf.argmax(py_x_z_probs, axis=1)\n labels_true = one_hot_encoder.inverse_transform(y_test)\n\n if inside_call:\n adm.plot_roc_auv_curve_of_an_algorithm(alg_ms=labels_pred, gt_ms=labels_true,\n alg_probs=py_x_z_probs, gt_ms_onehot=y_test,\n data_name='make_cls', alg_name='fpvi-pw',\n name_of_auc_roc_fig=dense_layer_type, sample_weight=None, case=0)\n prf = metrics.precision_recall_fscore_support(y_true=labels_true, y_pred=labels_pred, average='weighted')\n print(\"PRF:\", prf)\n","repo_name":"Sorooshi/MMLE-by-BBVI","sub_path":"codes/marginalized_maximum_likelihood_estimation_by_bbvi.py","file_name":"marginalized_maximum_likelihood_estimation_by_bbvi.py","file_ext":"py","file_size_in_byte":10417,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"33098377646","text":"import argparse\nfrom fabrics import create_frame_reader, create_output_saver, \\\n create_detector, create_tracker\nfrom full_video_detector import FullVideoDetector\nfrom stepwise_video_detector import StepwiseVideoDetector\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n # Video to detect and track vehicles\n parser.add_argument('-q', '--frame_sequence', action = 'store_true',\n help = 'flag that allows to process image sequence')\n parser.add_argument('-v', '--video', help = 'video to detect vehicles or \\\n directory name containing image sequence')\n \n # Video-based detection algorithm\n parser.add_argument('-a', '--algorithm', default = 'FD', help = 'type \\\n of video-based detection algorithm (\\'FD\\' - full detection \\\n frame-by-frame, FDT - full detection + creating tracks)')\n \n # Detector parameters\n parser.add_argument('-d', '--detector', default = 'OpenCV',\n help = 'detector name (\\'OpenCV\\' mode supports DNN-based detectors)')\n #default = 'caffe'\n parser.add_argument('-f', '--framework', default = 'darknet',\n help = 'deep learning framework (\\'caffe\\', \\'darknet\\') \\\n are supported')\n parser.add_argument('-l', '--labels', default = '../tests/voc_classes.txt',\n help = 'file containing object classes for object detection \\\n in format \\'\\'')\n #default = '../tests/resnet50_rfcn_final.caffemodel',\n parser.add_argument('-w', '--weights',\n default = '../tests/yolo-voc.weights',\n help = 'model trained to detect objects')\n #default = '../tests/rfcn_pascal_voc_resnet50.prototxt',\n parser.add_argument('-p', '--representation',\n default = '../tests/yolo-voc.cfg',\n help = 'model description')\n #default = '102.9801 115.9465 122.7717',\n parser.add_argument('-m', '--mean', default = '0 0 0',\n help = 'mean intensity value')\n #default = 800,\n parser.add_argument('-c', '--cols', default = 416,\n help = 'input width (cols)')\n #default = 600,\n parser.add_argument('-r', '--rows', default = 416,\n help = 'input height (rows)')\n #default = 1.0,\n parser.add_argument('-s', '--scale_factor', default = 0.00392,\n help = 'scale factor for the input blob')\n #default = 'bgr',\n parser.add_argument('-bgr', default = 'rgb',\n help = 'flag to set the sequence of channels (\\'bgr\\' or \\'rgb\\')')\n parser.add_argument('-e', '--confidence_threshold', default = 0.5,\n help = 'confidence threshold')\n \n # Tracker parameters\n parser.add_argument('-t', '--tracker', default = None, help = 'tracker \\\n name supported by OpenCV (\\'BOOSTING\\', \\'MIL\\', \\'KCF\\', \\'TLD\\', \\\n \\'MEDIANFLOW\\', \\'GOTURN\\', \\'MOSSE\\', \\'CSRT\\')')\n\n # Options\n parser.add_argument('-so', '--std_output', action = 'store_true',\n help = 'redirect output information about detected objects \\\n to the standard output')\n parser.add_argument('-o', '--output', default = 'output.txt',\n help = 'output file containing list of detected vehicles')\n\n args = parser.parse_args()\n\n try:\n # Prepare video, detector and tracker\n video = create_frame_reader(args.frame_sequence, args.video)\n output_saver = create_output_saver(args.std_output, args.output)\n detector = create_detector(args.detector, args.labels, args.framework,\n args.weights, args.representation, args.mean, args.cols,\n args.rows, args.scale_factor, args.bgr, args.confidence_threshold)\n tracker = create_tracker(args.tracker)\n # Detect and track vehicles\n if (args.algorithm == 'FD'):\n video_detector = FullVideoDetector(video, detector,\n output_saver)\n elif (args.algorithm == 'FDT'):\n video_detector = StepwiseVideoDetector(video, detector, tracker,\n output_saver)\n else:\n raise ValueError('Video-based detection method {} \\\n is not supported'.format(args.algorithm))\n video_detector.process()\n except Exception as ex:\n print('ERROR: {}'.format(str(ex)))\n","repo_name":"valentina-kustikova/dnn-object-detectors-comp","sub_path":"vehicle-detector/video-detector/video_analyzer.py","file_name":"video_analyzer.py","file_ext":"py","file_size_in_byte":4170,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"1569718857","text":"import frappe\nfrom frappe import _, bold\nfrom frappe.model.document import Document\n\n\nclass PartyLink(Document):\n\tdef validate(self):\n\t\tif self.primary_role not in [\"Customer\", \"Supplier\"]:\n\t\t\tfrappe.throw(\n\t\t\t\t_(\n\t\t\t\t\t\"Allowed primary roles are 'Customer' and 'Supplier'. Please select one of these roles only.\"\n\t\t\t\t),\n\t\t\t\ttitle=_(\"Invalid Primary Role\"),\n\t\t\t)\n\n\t\texisting_party_link = frappe.get_all(\n\t\t\t\"Party Link\",\n\t\t\t{\"primary_party\": self.primary_party, \"secondary_party\": self.secondary_party},\n\t\t\tpluck=\"primary_role\",\n\t\t)\n\t\tif existing_party_link:\n\t\t\tfrappe.throw(\n\t\t\t\t_(\"{} {} is already linked with {} {}\").format(\n\t\t\t\t\tself.primary_role, bold(self.primary_party), self.secondary_role, bold(self.secondary_party)\n\t\t\t\t)\n\t\t\t)\n\n\t\texisting_party_link = frappe.get_all(\n\t\t\t\"Party Link\", {\"primary_party\": self.secondary_party}, pluck=\"primary_role\"\n\t\t)\n\t\tif existing_party_link:\n\t\t\tfrappe.throw(\n\t\t\t\t_(\"{} {} is already linked with another {}\").format(\n\t\t\t\t\tself.secondary_role, self.secondary_party, existing_party_link[0]\n\t\t\t\t)\n\t\t\t)\n\n\t\texisting_party_link = frappe.get_all(\n\t\t\t\"Party Link\", {\"secondary_party\": self.primary_party}, pluck=\"primary_role\"\n\t\t)\n\t\tif existing_party_link:\n\t\t\tfrappe.throw(\n\t\t\t\t_(\"{} {} is already linked with another {}\").format(\n\t\t\t\t\tself.primary_role, self.primary_party, existing_party_link[0]\n\t\t\t\t)\n\t\t\t)\n\n\n@frappe.whitelist()\ndef create_party_link(primary_role, primary_party, secondary_party):\n\tparty_link = frappe.new_doc(\"Party Link\")\n\tparty_link.primary_role = primary_role\n\tparty_link.primary_party = primary_party\n\tparty_link.secondary_role = \"Customer\" if primary_role == \"Supplier\" else \"Supplier\"\n\tparty_link.secondary_party = secondary_party\n\n\tparty_link.save(ignore_permissions=True)\n\n\treturn party_link\n","repo_name":"frappe/erpnext","sub_path":"erpnext/accounts/doctype/party_link/party_link.py","file_name":"party_link.py","file_ext":"py","file_size_in_byte":1755,"program_lang":"python","lang":"en","doc_type":"code","stars":15303,"dataset":"github-code","pt":"60"}
+{"seq_id":"34980979528","text":"#!/usr/bin/env python3\n\nimport os\nfrom subprocess import call\nimport argparse\nimport multiprocessing\n\ndef run_unit_tests(octopus_build_dir, use_verbose_output):\n octopus_test_dir = octopus_build_dir + \"/test\"\n os.chdir(octopus_test_dir)\n ctest_options = []\n if use_verbose_output:\n ctest_options.append(\"--verbose\")\n call([\"ctest\"] + ctest_options)\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--type',\n help='C++ compiler path',\n default=\"unit\")\nparser.add_argument('--verbose',\n help='Output verbose test information',\n action='store_true')\nparser.add_argument('--compiler',\n help='C++ compiler path')\nparser.add_argument('--threads',\n help='The number of threads to use for building',\n type=int)\nargs = vars(parser.parse_args())\n\nif args[\"type\"] not in [\"unit\", \"valgrind\", \"regression\"]:\n print(\"Unknown test type \" + type)\n exit()\n\n# This file is in octopus-dir/test\noctopus_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))\n\nroot_cmake = octopus_dir + \"/CMakeLists.txt\"\n\nif not os.path.exists(root_cmake):\n print(\"octopus source directory corrupted: root CMakeLists.txt is missing. Please re-download source code.\")\n exit()\n\noctopus_build_dir = octopus_dir + \"/build\"\n\nif not os.path.exists(octopus_build_dir):\n print(\"octopus source directory corrupted: build directory is missing. Please re-download source code.\")\n exit()\n\nos.chdir(octopus_build_dir) # so cmake doesn't pollute root directory\n\ncmake_options = []\n\nif args[\"type\"] == \"unit\":\n cmake_options.extend([\"-DBUILD_TESTING=ON\", octopus_dir])\nelif args[\"type\"] == \"valgrind\":\n cmake_options.append(\"-DCMAKE_BUILD_TYPE=Debug\")\n\nif args[\"compiler\"]:\n cmake_options.append(\"-DCMAKE_CXX_COMPILER=\" + args[\"compiler\"])\n\nret = call([\"cmake\"] + cmake_options + [\"..\"])\n\nmake_options = []\n\nif args[\"threads\"]:\n if (args[\"threads\"] > 1):\n make_options.append(\"-j\" + str(args[\"threads\"]))\nelse:\n make_options.append(\"-j\" + str(multiprocessing.cpu_count()))\n\nif ret == 0:\n ret = call([\"make\"] + make_options)\n if ret == 0:\n if args[\"type\"] == \"unit\":\n run_unit_tests(octopus_build_dir, args[\"verbose\"])\n elif args[\"type\"] == \"valgrind\":\n call([\"make\", \"install\"])\n","repo_name":"luntergroup/octopus","sub_path":"test/install.py","file_name":"install.py","file_ext":"py","file_size_in_byte":2385,"program_lang":"python","lang":"en","doc_type":"code","stars":288,"dataset":"github-code","pt":"60"}
+{"seq_id":"74601972989","text":"# Desenvolva um programa que pergunte a distância de uma viagem em KM. \n# Calcule o preço da passagem, cobrando 0,50 centavos por km para viagens de até 200km e 0,45 centavos para viagens mais longas.\n\nkm = float (input('Digite quantos KM terá a sua viagem: '))\npr1 = km*0.45\npr2 = km*0.50\nif km >= 200:\n print('O preço da sua viagem é: {:.2f} reais'.format(pr2))\nelse:\n print('O preço da sua viagem é: {:.2f} reais'.format(pr1))","repo_name":"luizhmfonseca/Estudos-Python","sub_path":"EXERCÍCIOS - meus códigos/EX31.py","file_name":"EX31.py","file_ext":"py","file_size_in_byte":439,"program_lang":"python","lang":"pt","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"22999297102","text":"# -- coding: UTF-8 --\n\n\"\"\"\nGiven an input dictionary in the form of a JSON how will you construct an\ninstance of a class without explicitly calling the class constructor.\n\"\"\"\n\nimport logging.config\n\n\n__author__ = 'saranya@gyandata.com'\n\nLOGGER = logging.getLogger('root')\nLOGGER_CONFIG_PATH = 'config/logging.json'\n\n\nclass JsonData:\n \"\"\" A class contains the co ordinates of a point\"\"\"\n def __init__(self, x, y):\n \"\"\"\n Constructor\n :ivar x: The x coordinate of the point\n :ivar y: The y coordinate of the point\n \"\"\"\n self.x = x\n self.y = y\n\n def __str__(self):\n return \"Value of x: %d and Value of y: %d\" % (self.x, self.y)\n","repo_name":"saranyasivam98/Classes_Objects","sub_path":"classes/without_constructor.py","file_name":"without_constructor.py","file_ext":"py","file_size_in_byte":689,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"1353902684","text":"# insertion sort(挿入ソート)\n\ndef insertion_sort(a: list[int]) -> None:\n for i in range(1, len(a)):\n v: int = a[i]\n\n # 挿入する場所jを探す\n j: int = i\n while j > 0:\n if a[j - 1] > v:\n a[j] = a[j - 1] # vより大きいものは1つ後ろに移す\n else:\n break\n j -= 1\n a[j] = v\n print(a)\n\n\nif __name__ == '__main__':\n a: list[int] = [5, 9, 2, 0, 4]\n\n insertion_sort(a)\n\n print(a)\n","repo_name":"yukinakanaka/AlgorithmAndDataStructure","sub_path":"chap12/12.1.py","file_name":"12.1.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"ja","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"6645261967","text":"import config.package\nimport os\n\nclass Configure(config.package.CMakePackage):\n def __init__(self, framework):\n config.package.CMakePackage.__init__(self, framework)\n self.gitcommit = 'master'\n # using fork of Sherry's branch to work around bug in handling of BLAS, pull request made \n self.download = ['git://https://github.com/petsc/superlu']\n# self.download = ['git://https://github.com/xiaoyeli/superlu']\n# self.download = ['git://https://bitbucket.org/petsc/pkg-superlu.git',\n# 'http://ftp.mcs.anl.gov/pub/petsc/externalpackages/superlu_5.1.tar.gz']\n self.functions = ['set_default_options']\n self.includes = ['slu_ddefs.h']\n self.liblist = [['libsuperlu.a']]\n # SuperLU has NO support for 64 bit integers, use SuperLU_Dist if you need that\n self.requires32bitint = 1; # 1 means that the package will not work with 64 bit integers\n self.excludedDirs = ['SuperLU_DIST','SuperLU_MT']\n # SuperLU does not work with --download-fblaslapack with Compaqf90 compiler on windows.\n # However it should work with intel ifort.\n self.downloadonWindows= 1\n self.hastests = 1\n self.hastestsdatafiles= 1\n return\n\n def setupDependencies(self, framework):\n config.package.CMakePackage.setupDependencies(self, framework)\n self.blasLapack = self.framework.require('config.packages.BlasLapack',self)\n self.deps = [self.blasLapack]\n return\n\n def formCMakeConfigureArgs(self):\n args = config.package.CMakePackage.formCMakeConfigureArgs(self)\n args.append('-DUSE_XSDK_DEFAULTS=YES')\n\n args.append('-DTPL_BLAS_LIBRARIES=\"'+self.libraries.toString(self.blasLapack.dlib)+'\"')\n\n # Tests are broken on Apple since they depend on a shared library that is not resolved against BLAS\n args.append('-Denable_tests=0')\n # CMake in SuperLU should set this; but like many other packages it does not\n args.append('-DCMAKE_INSTALL_NAME_DIR:STRING=\"'+os.path.join(self.installDir,self.libdir)+'\"')\n return args\n\n","repo_name":"taupalosaurus/petscAdapt","sub_path":"config/BuildSystem/config/packages/SuperLU.py","file_name":"SuperLU.py","file_ext":"py","file_size_in_byte":2074,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"1136252163","text":"import discord\nimport random\nimport os\nfrom dotenv import load_dotenv\n\nload_dotenv(verbose=True)\n\n#디스코드 개발 토큰\ntoken = os.getenv('TOKEN')\n\n\n#msg 초기화\nmsg = None \nx = None\n\n\nclient = discord.Client()\n\n#타자연습 영어 문장들 15개\nt = [\"Experience is not what happens to a man; it is what a man does with what happens to him.\",\n \"There is no feeling, except the extremes of fear and grief, that does not find relief in music.\",\n \"I hear and I forget. I see and I remember. I do and I understand.\",\n \"If you want to see what children can do, you must stop giving them things.\",\n \"The man of virtue makes the difficulty to be overcome his first business, and success only a subsequent consideration.\",\n \"People fail forward to success.\",\n \"It is wise to apply the oil of refined politeness to the mechanisms of friendship.\",\n \"Without friends no one would choose to live, though he had all other goods.\",\n \"Man's feelings are always purest and most glowing in the hour of meeting and of farewell.\",\n \"This bud of love, by summer's ripening breath, May prove a beauteous flower when next we meet.\",\n \"Running cross country is the closest man will ever get to flying.\",\n \"My philosophy is that not only are you responsible for your life, but doing the best at this moment puts you in the best place for the next moment.\",\n \"He that lives upon hope will die fasting.\",\n \"It has never been my object to record my dreams, just to realize them.\"\n ]\n\n#타자연습 한글 문장들 15개\nT = [\"사랑은 언제까지나 지속되어야 하는 것인가, 아니면 이런 저런 정거장에 멈춰서는 여러 열차와 같은 것인가? 내가 그녀를 사랑한다면 어떻게 그녀를 떠날 수 있나? 그 때 내가 그렇게 느꼈다면, 지금은 왜 아무 것도 느끼지 못할까?\",\n \"죄를 미워하되 죄인은 사랑하라.\",\n \"젊은이들은 젊음이 얼마나 힘들고 무시무시할 수 있는지 안다. 그들의 젊음은 다른 모든 사람들에게 허비되는데 그야말로 끔찍한 일이다. 젊은이들에게는 권위도 존경도 없다.\",\n \"들은 것은 잊어버리고, 본 것은 기억하고 직접 해본 것은 이해한다\",\n \"아이들이 무엇을 할 수 있는지 확인해보고 싶다면 주는 것을 멈추어 보면 된다.\",\n \"어진 사람은 난관의 극복을 제일 중요한 일로 생각하고, 성공 여부는 부차적인 것으로 본다.\",\n \"실패하는 것은 곧 성공으로 한 발짝 더 나아가는 것이다.\",\n \"우정이라는 기계에 잘 정제된 예의라는 기름을 바르는 것은 현명하다.\",\n \"모든 것을 가졌다 해도 친구가 없다면, 아무도 살길 원치 않을 것이다.\",\n \"책은 가장 조용하고 변함 없는 벗이다. 책은 가장 쉽게 다가갈 수 있고 가장 현명한 상담자이자, 가장 인내심 있는 교사이다.\",\n \"그것은 죽었으면 하고 바라는 사람들이 시간을 죽이기 위해 읽는 책이었다.\",\n \"인간의 감정은 누군가를 만날 때와 헤어질 때 가장 순수하며 가장 빛난다.\",\n \"이 사랑의 꽃봉오리는 여름날 바람에 마냥 부풀었다가, 다음 만날 때엔 예쁘게 꽃필 거예요.\",\n \"일부 과학자들에 따르면 미래는 과거와 똑같을 것이다. 단지 훨씬 값비쌀 뿐이다.\",\n \"우아함이란 이제 갖 사춘기를 벗어난 이들의 특권이 아니라, 이미 스스로의 미래를 꽉 잡고 있는 이들의 것이다.\"]\n\n#디스코드 봇 실행시 터미널로 보여주기\nclass Typewriterbot(discord.Client):\n chatTest = \"False\"\n channel = \"NULL\"\n q = \"NULL\"\n \n print(\"debug ready\")\n async def on_ready(self):\n game = discord.Game(\"!문제를 해결\")\n\n await client.change_presence(status=discord.Status.online, activity=game)\n print(\"Ready to Action...\")\n\n async def on_message(self, message):\n \n if message.author.bot:\n return None\n\n\n #기본적인 명령어\n if message.content == '!안녕':\n channel = message.channel\n msg = \"안녕\"\n await channel.send(msg)\n return None\n\n if message.content == '!잘가':\n channel = message.channel\n msg = \"잘가 ㅠㅠ\"\n await channel.send(msg)\n return None\n \n \n #명령어 보여주기 \n if message.content == '!명령어': \n channel = message.channel\n msg = \"```\\n!안녕 - 안녕\\n\"\n msg += \"!잘가 - 잘가 ㅠㅠ\\n\"\n msg += \"!타자연습 - 타자 연습 시작합니다.\\n\"\n msg += \"!타자연습 영어 - 타자 연습 영어 시작합니다.\\n\"\n msg += \"!타자연습 한글 - 타자 연습 한글 시작합니다.\\n\"\n msg += \"!c언어 - 필요한 문법 책 링크를 보여줍니다.\\n\"\n msg += \"!c언어 OO - OO 보여줌\\n 예시:!c언어 for\\n```\"\n await channel.send(msg)\n return None\n\n \n #c언어 명령어\n if message.content == '!c언어':\n channel = message.channel\n msg = \"```\\nC언어\\n\"\n msg += \"비쥬얼 스튜디오 설치하기\\n https://visualstudio.microsoft.com/ko/downloads/\\n\"\n msg += \"DEV C++ 설치하기\\n https://gabii.tistory.com/entry/Dev-C-Dev-C-%EB%8B%A4%EC%9A%B4%EB%A1%9C%EB%93%9C-%EB%B0%8F-%EC%84%A4%EC%B9%98\\n\"\n msg += \"변수\\n https://thebook.io/006989/ch02/01/\\n\"\n msg += \"함수\\n https://thebook.io/006989/ch03/\\n\"\n msg += \"연산자\\n https://thebook.io/006989/ch04/\\n\"\n msg += \"조건문\\n https://thebook.io/006989/ch05/\\n\"\n msg += \"반복문\\n https://thebook.io/006989/ch06/\\n\"\n msg += \"배열\\n https://thebook.io/006989/ch07/\\n\"\n msg += \"포인터\\n https://thebook.io/006989/ch08/\\n\\n\"\n msg += \"링크는 복사 붙여넣기로 사용하실 수 있습니다\\n```\"\n await channel.send(msg)\n return None\n\n if message.content == '!c언어 변수':\n channel = message.channel\n msg = \"변수\\n https://thebook.io/006989/ch02/01/\\n\"\n await channel.send(msg)\n return None\n\n if message.content == '!c언어 함수':\n channel = message.channel\n msg = \"함수\\n https://thebook.io/006989/ch03/\\n\"\n await channel.send(msg)\n return None\n\n if message.content == '!c언어 연산자':\n channel = message.channel\n msg = \"연산자\\n https://thebook.io/006989/ch04/\\n\"\n await channel.send(msg)\n return None \n\n if message.content == '!c언어 조건문':\n channel = message.channel\n msg = \"조건문\\n https://thebook.io/006989/ch05/\\n\"\n await channel.send(msg)\n return None \n \n if message.content == '!c언어 반복문':\n channel = message.channel\n msg = \"반복문\\n https://thebook.io/006989/ch06/\\n\"\n await channel.send(msg)\n return None \n\n if message.content == '!c언어 배열':\n channel = message.channel\n msg = \"배열\\n https://thebook.io/006989/ch07/\\n\"\n await channel.send(msg)\n return None \n \n if message.content == '!c언어 포인터':\n channel = message.channel\n msg = \"포인터\\n https://thebook.io/006989/ch08/\\n\"\n await channel.send(msg)\n return None \n \n #타자 연습 \n #타자연습 영어 시작\n if message.content == '!타자연습 영어': \n channel = message.channel\n msg = \"시작\\n\"\n msg += \"<문장>\"\n self.chatTest = \"True\"\n await channel.send(msg)\n \n self.q = random.choice(t)\n await channel.send(\"============= 문제 ============\\n!정답 입력후 정답을 입력해주세요 예(!정답 Hello world)\")\n await channel.send(self.q)\n return None\n\n if self.chatTest == \"True\": \n print(self.chatTest)\n channel = message.channel\n msg = message.content\n \n print(self.q)\n print(message.content)\n \n if '!정답 ' + self.q == message.content: #정답 입력받기\n await channel.send(\"정답\")\n print(\"정답\")\n else:\n await channel.send(\"땡\")\n print(\"땡\")\n\n self.chatTest = \"False\"\n return None\n\n\n #타자연습 한글 시작\n if message.content == '!타자연습 한글': \n channel = message.channel \n msg = \"시작\\n\"\n msg += \"<문장>\"\n self.chatTest = \"True\"\n await channel.send(msg)\n \n self.q = random.choice(T)\n await channel.send(\"============= 문제 ============\\n!정답 입력후 정답을 입력해주세요 예(!정답 안녕하세요)\")\n await channel.send(self.q)\n return None\n\n if self.chatTest == \"True\": \n print(self.chatTest)\n channel = message.channel\n msg = message.content\n \n print(self.q)\n print(message.content)\n \n if '!정답 ' + self.q == message.content: #정답 입력받기\n await channel.send(\"정답\")\n print(\"정답\")\n \n else:\n await channel.send(\"땡\")\n print(\"땡\")\n \n\n self.chatTest = \"False\"\n return None\n \nif __name__ == \"__main__\":\n client = Typewriterbot()\n client.run(token) ","repo_name":"potatovllage/Discord_Bot","sub_path":"discordbot.py","file_name":"discordbot.py","file_ext":"py","file_size_in_byte":10186,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"31333734867","text":"from django.urls import path\nfrom home.views import *\nfrom admin_dashboard.views import contact_us_mail\n\nurlpatterns = [\n path('', HomePageView.as_view(), name='home'),\n path('about', AboutPageView.as_view(), name='about'),\n path('services', ServicePageView.as_view(), name='service'),\n path('contact-us', ContactPageView.as_view(), name='contact'),\n path('gallery', GalleryPageView.as_view(), name='gallery'),\n path('mail/send', contact_us_mail, name='send_mail'),\n\n\n]\n\n","repo_name":"Ziko278/poultry","sub_path":"home/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"35723376397","text":"from flask import Flask, request\nfrom flask_restful import Resource, Api, reqparse, inputs\nfrom celery.result import AsyncResult\n\nfrom tasks import get_text, get_images, celery_app\n\napp = Flask(__name__)\napi = Api(app)\n\nparser = reqparse.RequestParser()\nparser.add_argument('site')\n\nclass Text(Resource):\n def post(self):\n args = parser.parse_args()\n result = get_text.delay(args['site'])\n return {'task_id': result.task_id}\n\nclass Images(Resource):\n def post(self):\n args = parser.parse_args()\n result = get_images.delay(args['site'])\n return {'task_id': result.task_id}\n\nclass Status(Resource):\n def get(self, task_id):\n status = AsyncResult(task_id, app=celery_app).status\n return {'status': status}\n\n\napi.add_resource(Text, '/text')\napi.add_resource(Images, '/images')\napi.add_resource(Status, '/status/')\n\n\nif __name__ == \"__main__\":\n app.run()\n","repo_name":"DominikZabron/machine-learning-assistance","sub_path":"api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":927,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"3922067364","text":"# configs\n\n# --- streaming ---\n# access token\nACCESS_TOKEN = '1440425207190024196-iSOaQja7rMiCMbZwoslmfOe0RX0XPu'\n# access secret\nACCESS_SECRET = '2eOScxtfc6aUqfZc4xEDfMTRlxJUIkavB1i0Shfbi6fIt'\n# consumer key\nCONSUMER_KEY = 'GnCIBr4YWVSFH65ZbOFvXJEpB'\n# consumer secret\nCONSUMER_SECRET = 'Q15jf7i3M7JcOtevi4EYhCjLblKUrJOdDBKgUAxz5npWVmifVB'\n\n# twtter checkpoint path\ncp_path = './data/checkpoint_TwitterApp'\n# twitter data output path\noutput_directory = './data/twitter/movie'\n\n# client IP\nIP = 'localhost'\n# client port\nPORT = 9001\n\n# the tags to track\ntags = ['movie']\n\n\n# --- movies ---\n# tmdb api key\napi_key = '94b42385a681053cab08a06553dcfa19'\n# tmdb language\nlanguage = 'en'\n# tmdb mode\ndebug = True\n\n# feature to collect\nfeatures_default = [\n 'id', 'title', 'release_date', 'vote_average', 'vote_count', \n 'genres', 'budget', 'popularity', 'revenue'\n]\n\n# raw data path\npath = './data/movies.csv'\n\n\n# --- preprocess --\n# preprocessed data path\npp_path = './data/movies_pp.csv'\n# lda num topics\nnum_topics=10\n# lda max iters\nmax_iterations=50\n# lda num of words for each topic\nwordNumbers=10\n\n\n# --- analysis ---\n# predicted results path\npred_path = './data/predicts.csv'\ntopic_path = './data/topics.txt'\n\n","repo_name":"harrypotter1501/movie-sentiment","sub_path":"config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":1218,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"27493246001","text":"# learning how to split \n\n#line = \"This is a string\"\n \n#print(line.split())\n#print(line.split(' '))\n\n#print('-'.join(line.split(' ')))\n\n#line = line.split(\" \")\n#line = \"-\".join(line)\n#print(line)\n\n\n#var = \"If there is a will there is a way\"\n\n#var = var.split()\n#var= \"/\".join(line)\n\n#print(var)\n\n#steve = \"starting to get a hang of this split thing\"\n\n#steve = steve.split(\" \")\n#steve = \"$\".join(steve)\n#print(steve)\n\n#cando = \"Want me to split a string and then join it with a - ?\"\n#print(cando)\n\n#cando = cando.split()\n#print(cando)\n#cando = \"-\".join(cando)\n#print(cando)\n\n#park = '-'.join(park)\n\n#print(park)\n\n#today = \"Is goin to be a good day\"\n#today = today.split()\n#today = \"-\".join(today)\n#print(today)\n\n#wisconsin = \"Really cold this time of year\"\n#wisconsin = wisconsin.split()\n#wisconsin = '-'.join(wisconsin)\n#print(wisconsin)\n\n#france = \"its nice and warm this time of year\"\n#france = france.split()\n#france = '-'.join(france)\n#print(france)\n\nlast_one = \"This is my last one as I think I've done it enough times to remember\"\n\nlast_one = last_one.split() \nprint(last_one) #print\n\nlast_one = \"-\".join(last_one) \nprint(last_one)\n\n\n\n\n\n\n\n\n\n","repo_name":"steve0c/red_python","sub_path":"spliting-lab.py","file_name":"spliting-lab.py","file_ext":"py","file_size_in_byte":1147,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"22515570637","text":"from django.contrib.auth.forms import UserCreationForm\nfrom django import forms\nfrom django.contrib.auth.models import User\n\nclass RegisterForm(UserCreationForm):\n\temail = forms.EmailField(label='Email address', max_length=75)\n\t\n\tclass Meta:\n\t\tmodel = User\n\t\tfields = ('username', 'email',) \t\n\t\t\n\tdef clean_email(self):\n\t\temail = self.cleaned_data[\"email\"]\n\t\t\n\t\ttry:\n\t\t\tUser.objects.get(email__iexact=email)\n\t\texcept User.DoesNotExist:\n\t\t\treturn email\n\t\t\n\t\traise forms.ValidationError(\"A user with that email address already exists.\")\n\t\n\tdef save(self, commit=True):\n\t\tuser = super(UserCreationForm, self).save(commit=False)\n\t\tuser.set_password(self.cleaned_data[\"password1\"])\n\t\tuser.email = self.cleaned_data[\"email\"]\n\t\tuser.is_active = True\n\t\tif commit:\n\t\t\tuser.save()\n\t\t\t\n\t\treturn user","repo_name":"tfaris/pyabetic","sub_path":"account_forms.py","file_name":"account_forms.py","file_ext":"py","file_size_in_byte":788,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"70710823873","text":"import numpy as np\r\n\r\n\r\nclass WordDict():\r\n def __init__(self, embeddingPath):\r\n self.word2ID, self.ID2word, self.wordVector = self.loadEmbedding(embeddingPath)\r\n\r\n def loadEmbedding(self, Path):\r\n word2ID = {'UNK': 0}\r\n ID2word = {0: 'UNK'}\r\n wordVector = []\r\n\r\n with open(Path, encoding='utf-8') as f:\r\n idx = 1\r\n while True:\r\n line = f.readline()\r\n if not line:\r\n break\r\n if idx == 1:\r\n dim = len(line.split(' ')) - 1\r\n unk_embedding = np.random.normal(0.0, 0.5, dim).astype('float32').tolist()\r\n wordVector.append(unk_embedding)\r\n\r\n lineList = line.split(' ')\r\n word2ID[lineList[0]] = idx\r\n ID2word[idx] = lineList[0]\r\n wordVector.append([float(x) for x in lineList[1:]])\r\n\r\n idx += 1\r\n\r\n return word2ID, ID2word, wordVector\r\n\r\n def get_word2ID(self):\r\n return self.word2ID\r\n\r\n def get_ID2word(self):\r\n return self.ID2word\r\n\r\n def get_wordVector(self):\r\n return self.wordVector\r\n\r\nif __name__ == '__main__':\r\n wordDict = WordDict('./glove.6B.100d.txt')\r\n","repo_name":"chipinzhen/BiLSTM-CRF","sub_path":"data/word_dict.py","file_name":"word_dict.py","file_ext":"py","file_size_in_byte":1257,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"21953694105","text":"from typing import List\n\n\nclass Solution:\n def countDistinct(self, nums: List[int], k: int, p: int) -> int:\n for i in range(len(nums)):\n if nums[i] % p == 0:\n nums[i] = str(nums[i]//p) + 'x'\n nums = [str(i) for i in nums]\n final_set = set()\n for i in range(len(nums)):\n for j in range(i+1, len(nums)+1):\n string = ' '.join(nums[i:j])\n if string.count('x') <= k:\n final_set.add(string)\n return len(final_set)\n\n\n\n\n\n\n\n\nif __name__ == '__main__':\n #print(Solution().countDistinct(nums = [2,3,3,2,2], k = 2, p = 2))\n #print(Solution().countDistinct(nums = [1,2,3,4], k = 4, p = 1))\n print(Solution().countDistinct([1,9,8,7,19], 1, 6))\n","repo_name":"liuyuhanalex/Leetcode","sub_path":"contest/6049. K Divisible Elements Subarrays.py","file_name":"6049. K Divisible Elements Subarrays.py","file_ext":"py","file_size_in_byte":763,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"72067172030","text":"import os\nimport struct\nimport shutil\nimport tempfile\nimport unittest\n\nimport numpy as np\n\nfrom mibidata import pseudodepths\n\n# File variant 1, 8 bins per spectrum, 16 = 4x4 pixels.\nNUM_PIXELS = 4\nHEADER = (1, 8, NUM_PIXELS)\n# Assume 10 cycles per pixel, so each data sub-list has 10 zeros and the rest\n# of the counts are bin numbers from 1 to 8. The (0, 0) entry to indicate\n# new cycle comes at the _end_ of the data for each cycle.\n#\n# This following data was generated with:\n#\n# pixel_lengths = np.random.randint(9, 20, 16)\n# DATA = []\n# for pl in pixel_lengths:\n# entries = np.zeros(pl, int)\n# num_counts = pl - 9\n# inds = np.random.choice(np.arange(pl), num_counts, replace=False)\n# bins = np.random.randint(1, 8, num_counts)\n# entries[inds] = bins\n# DATA.append(list(entries) + [0])\nDATA = [\n [0, 3, 0, 1, 1, 0, 3, 0, 0, 0, 3, 0, 3, 7, 3, 4, 0, 0, 5, 0],\n [0, 5, 4, 2, 0, 0, 0, 0, 0, 0, 3, 0, 2, 2, 0, 2, 0],\n [0, 0, 0, 6, 7, 0, 0, 0, 0, 0, 0, 2, 2, 0],\n [0, 1, 1, 0, 0, 3, 0, 0, 0, 0, 0, 0, 0],\n]\n# How many counts to expect if we manually split into two depths.\nDEPTH0 = [\n [0, 3, 0, 1, 1, 0, 3, 0, 0],\n [0, 5, 4, 2, 0, 0, 0, 0],\n [0, 0, 0, 6, 7, 0, 0],\n [0, 1, 1, 0, 0, 3, 0, 0],\n]\nDEPTH1 = [\n [0, 3, 0, 3, 7, 3, 4, 0, 0, 5, 0],\n [0, 0, 3, 0, 2, 2, 0, 2, 0],\n [0, 0, 0, 0, 2, 2, 0],\n [0, 0, 0, 0, 0],\n]\n\n\nclass TestMsdf(unittest.TestCase):\n\n @classmethod\n def setUpClass(cls):\n fh, fn = tempfile.mkstemp()\n cls.msdf = fn\n os.close(fh)\n sat = np.zeros((NUM_PIXELS, 2), int)\n cls.header = struct.pack(pseudodepths.HEADER_FORMAT, 1, 8, NUM_PIXELS)\n with open(fn, 'wb') as infile:\n infile.write(cls.header)\n end_sat = pseudodepths.HEADER_SIZE + \\\n NUM_PIXELS * pseudodepths.SAT_ENTRY_SIZE\n infile.seek(end_sat)\n for i, pixel in enumerate(DATA):\n sat[i, 0] = infile.tell()\n sat[i, 1] = len(pixel)\n for timestamp in pixel:\n count = int(timestamp > 0)\n infile.write(\n struct.pack(pseudodepths.DATA_FORMAT, timestamp, count))\n infile.seek(pseudodepths.HEADER_SIZE)\n for offset, length in sat:\n infile.write(\n struct.pack(pseudodepths.SAT_ENTRY_FORMAT, offset, length))\n cls.data_start = end_sat\n\n def setUp(self):\n self.tempdir = tempfile.mkdtemp()\n\n @classmethod\n def tearDownClass(cls):\n os.remove(cls.msdf)\n\n def tearDown(self):\n shutil.rmtree(self.tempdir)\n\n def _pack_sat(self, depth):\n sat = b''\n offset = self.data_start\n for d in depth:\n sat += struct.pack(\n pseudodepths.SAT_ENTRY_FORMAT, offset, len(d))\n offset += pseudodepths.DATA_SIZE * len(d)\n return sat\n\n def _pack_data(self, depth):\n data = b''\n for pixel in depth:\n for i in pixel:\n data += struct.pack(\n pseudodepths.DATA_FORMAT, i, int(i > 0))\n return data\n\n def test_split_into_one(self):\n \"\"\"If we split into one output file, we should get the same file out.\n \"\"\"\n cycles_per_pixel, cycles_per_scan = pseudodepths.divide(\n self.msdf, 1, self.tempdir)\n self.assertEqual(cycles_per_pixel, 10)\n self.assertEqual(cycles_per_scan, 10)\n new_file = os.path.join(self.tempdir, 'Depth0', 'Image.msdf')\n self.assertTrue(os.path.exists(new_file))\n with open(self.msdf, 'rb') as infile:\n expected_buffer = infile.read()\n with open(new_file, 'rb') as infile:\n new_buffer = infile.read()\n self.assertEqual(new_buffer, expected_buffer)\n\n def test_split_into_two(self):\n cycles_per_pixel, cycles_per_scan = pseudodepths.divide(\n self.msdf, 2, self.tempdir)\n self.assertEqual(cycles_per_pixel, 10)\n self.assertEqual(cycles_per_scan, 5)\n depth0 = os.path.join(self.tempdir, 'Depth0', 'Image.msdf')\n depth1 = os.path.join(self.tempdir, 'Depth1', 'Image.msdf')\n with open(depth0, 'rb') as infile:\n depth0_header = infile.read(pseudodepths.HEADER_SIZE)\n depth0_sat = infile.read(NUM_PIXELS * pseudodepths.SAT_ENTRY_SIZE)\n depth0_data = infile.read()\n with open(depth1, 'rb') as infile:\n depth1_header = infile.read(pseudodepths.HEADER_SIZE)\n depth1_sat = infile.read(NUM_PIXELS * pseudodepths.SAT_ENTRY_SIZE)\n depth1_data = infile.read()\n\n self.assertEqual(depth0_header, self.header)\n self.assertEqual(depth1_header, self.header)\n self.assertEqual(depth0_sat, self._pack_sat(DEPTH0))\n self.assertEqual(depth1_sat, self._pack_sat(DEPTH1))\n self.assertEqual(depth0_data, self._pack_data(DEPTH0))\n self.assertEqual(depth1_data, self._pack_data(DEPTH1))\n\n def test_split_into_three(self):\n \"\"\"This should raise because the number of cycles is not divisible by\n the number of desired pseudo-depths.\"\"\"\n with self.assertRaises(ValueError):\n pseudodepths.divide(self.msdf, 3, self.tempdir)\n\nif __name__ == '__main__':\n unittest.main()\n","repo_name":"ionpath/mibilib","sub_path":"mibidata/tests/test_pseudodepths.py","file_name":"test_pseudodepths.py","file_ext":"py","file_size_in_byte":5304,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"60"}
+{"seq_id":"25377520657","text":"import os\nfrom random import randint\n\ndef update_computer_brain(x, y):\n with open('FirstLastWord.txt', 'a+', encoding='UTF-8') as f:\n for value in x:\n if value not in y:\n f.write(f\"{value}\\n\")\n\n\n\nos.chdir('K:\\\\Python Group Study\\\\프로그래밍 기초 in Python\\\\GroupStudy')\n\nf = open('FirstLastWord.txt','r', encoding='UTF-8') # Brings Words that Computer has\n\nsys_wdbook = [] # List of words that computer has\ngame_wdbook = [] # List of words that used in game\nturn_count = 0 # How many turns went\nnxt_turn = \"\"\n\nfor line in f : # Read the file that has the words\n sys_wdbook.append(line.strip())\n\nwhile True: # Select who attacks first\n try:\n print(\"게임을 종료하시려면 언제든지 2번을 눌러주세요.\")\n turn = int(input(\"선공 후공을 정해주세요.(0: 선공, 1: 후공, 2: 종료) : \"))\n if turn == 0 or turn == 1 or turn == 2:\n break\n else:\n print(\"번호가 잘못되었습니다.\")\n except ValueError: #Error exception while choosing the turns\n print(\"숫자를 입력해주세요.\")\n\n\nwhile True: # Game first turn starts\n \n if int(turn) == 0 : # User's First Attack\n user_input = input(\"세상에 존재하는 단어(2~3자) 하나를 입력해주세요 :\") # User will input word\n if user_input == \"2\": \n print(\"사용자에 의해 게임이 종료되었습니다.\")\n exit()\n\n if len(user_input) > 3 or len(user_input) < 2:\n print(\"글자수를 맞춰주세요.\")\n\n elif len(user_input) <= 3 and len(user_input) >= 2:\n if turn_count == 0: # 1st user's input\n game_wdbook.append(user_input)\n turn_count += 1\n nxt_turn = \"com\"\n break\n\n\n elif int(turn) == 1: # Computer's First Attack\n sys_input = sys_wdbook[randint(0,len(sys_wdbook)-1)]\n print(sys_input)\n game_wdbook.append(sys_input)\n turn_count += 1\n nxt_turn = \"user\"\n break\n\n \n elif int(turn) == 2:\n print(\"게임 종료!!\")\n exit() \n\n","repo_name":"Wooil96/PythonStudy","sub_path":"프로그래밍 기초 in Python/GroupStudy/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":2130,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"46226537979","text":"import tensorflow as tf\n\n# --------------------------------- General -----------------------------------\n# The random seed to use. Seed or False. Kind of pointless at this point,\n# relict from earlier version.\nRANDOM_SEED = False\n\n# The number of layers to fit, which needs to be the same as in the respective\n# PPC configuration.\nN_LAYERS = 171\nN_DOMS = 5160\n\n# -------------------------------- TensorFlow ---------------------------------\nTF_FLOAT_PRECISION = tf.float32\nTF_CPU_ONLY = False\n# The length of the simulated photons array, this is the amount of photons we\n# optimize on each step. More is better, but it is limited by available GPU\n# memory. With the current program 700000 seems to be about the maximum a Tesla\n# P40 can handle without running out of memory.\nTF_HITLIST_LEN = 700000\n\n# ----------------------------- Ice Mocel Config ------------------------------\n# This is the path to the ice model to use to calculate the intrinsic\n# absorption coefficient and wavelength dependency of the dust absorption\n# coefficient. So the relevevant files are icemodel.dat for the delta\n# temperature values and icemodel.par for the 6 parameter ice model parameters.\nICE_MODEL_PATH = '/home/aharnisch/modded-PPC/real/ice/'\n\n# ------------------------------- Flasher Data --------------------------------\nDATA_PATH = '/net/big-tank/POOL/users/aharnisch/fake_flasher_data/'\n\n# ----------------------------- Simulation Data -------------------------------\n# The simulated photon data directory\nPHOTON_PATH = '/net/big-tank/POOL/users/aharnisch/iceopt_photons/'\n\n# --------------------------------- Training ----------------------------------\n# Flashing string, for now we only flash this one string. Should not make a\n# difference when comparing to simulation anyways since there are no model\n# errors. String 36 is in the middle of deep core. String 69 is in the top\n# right vorner of the second to last xy layer of strings (minimally effected by\n# deep core.)\nFLASHER_STRINGS = [36]\n\n# If this flag is set to true, the gradient is averaged over an entire string\n# before fed to the optimizer. This is not the same as evaluating string\n# batches. The batches are still performed on individual emitter DOMs but\n# instead of applying the gradient each time we evaluate it for all DOMs on a\n# string and then feed the acuumulated gradient to the optimizer. This might\n# make the gradient more stable becaue it includes information on all layers.\n# It also smooths out the loss significantly which is helpful when debugging.\n# If it is set to False the gradient is applied on every dom batch each time.\nGRADIENT_AVERAGING = True\n\n# The initial absorption coefficients to start with.\nINITIAL_ABS = [0.01 for i in range(N_LAYERS)]\n\n# The smallest allowed absorption coeffizient, values below are clipped on\n# every step\nMIN_ABS = 0.001\n\n# The maximum number of training steps to perform.\nMAX_STEPS = 200\n\n# The number of hits to rescale to. We rescale to this fixed amount of hits\n# every time to make the loss more comparable for different emitter DOMs. The\n# reason we have to rescale at all is the fact that we don't know how many\n# photons have been emitted on data.\nRESCALED_HITS = 100000\n\n\n# -------------------------------- Optimizer ----------------------------------\n# The initial learning rate\nINITIAL_LEARNING_RATE = 0.001\n# True or False to activate/deactivate learning rate decay\nLEARNING_DECAY = False\n# Decay modes: Linear or Exponential\nLEARNING_DECAY_MODE = 'Exponential'\n# decrease the INITIAL_LEARNING_RATE every LEARNING_STEPS steps by\n# LEARNING_DECR linearly or exponentially\nLEARNING_DECR = 0.95\nLEARNING_STEPS = 10\n\n# supported optimizers: Adam, GradientDescent\nOPTIMIZER = 'Adam'\nADAM_SETTINGS = dict(beta1=0.9, beta2=0.999, epsilon=1e-08)\n\n# --------------------------------- Logging -----------------------------------\nWRITE_INTERVAL = 1 # how many steps between each write\n","repo_name":"AlexHarn/tf-ppc-iceopt","sub_path":"settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":3901,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"36137579909","text":"# https://codeforces.com/problemset/problem/727/A\n\nfrom collections import defaultdict\n\n\ndef solve():\n a, b = map(int, input().split())\n target = b\n graph = defaultdict(lambda: [None, None, None])\n\n flag = True\n while b > a:\n parent = b\n if b % 2 == 0:\n b //= 2\n graph[parent][0] = b\n else:\n if (b - 1) % 10 != 0:\n flag = False\n break\n b = b // 10\n graph[parent][1] = b\n\n graph[b][2] = parent\n\n if a not in graph or not flag:\n print(\"NO\")\n return\n\n path = []\n while a < target:\n path.append(a)\n a = graph[a][2]\n path.append(target)\n\n print(f\"YES\\n{len(path)}\")\n print(*path)\n\n\nsolve()\n","repo_name":"Son-OfAnton/Competitive-Programming","sub_path":"Contest/Contest_15/transformationFromAtoB.py","file_name":"transformationFromAtoB.py","file_ext":"py","file_size_in_byte":759,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"10301604117","text":"from flask import Flask, render_template\n\nfrom flight_radar.flight_radar import process, get_flight_info\n\napp = Flask(__name__)\n\n\n@app.route(\"/\")\ndef index():\n return render_template('index.html')\n\n\n@app.route(\"/flights\")\ndef flights():\n flights_info = process()\n return render_template('flights.html', flights=flights_info)\n\n\n@app.route(\"/flight/\")\ndef flight(ads_hex):\n flight_info = get_flight_info(ads_hex)\n return render_template('flight.html', flight=flight_info)\n\n\nif __name__ == \"__main__\":\n app.run(host='0.0.0.0', port=5000)\n","repo_name":"ercanse/flight-track","sub_path":"ui.py","file_name":"ui.py","file_ext":"py","file_size_in_byte":569,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"39031565228","text":"# 골드4\n# 소수 경로\n# 에라토스테네스의 체\ndef et_prime(dp):\n for i in range(2, 10000):\n if dp[i] == False:\n if i >= 1000: sosu.append(i) # 1000 이상만 저장\n for j in range(i*2, 10000, i): dp[j] = True\n \n del dp\n\ndef bfs(start, end):\n\n if start == end:\n return 0\n else:\n dq = deque([[start, 1]])\n visited[int(start)] = True\n \n while dq:\n standard, depth = map(str, dq.popleft())\n standard = list(standard)\n \n for s in range(4): # 한번에 한 자릿수만 변경이 되므로 각 자릿수별로 연산\n for i in range(10):\n if standard[s] != str(i): # 기존값과 비교값이 같으면 패스\n temp = standard[s] # 기존값 임시저장\n standard[s] = str(i) # 기존값을 연산을 위해 비교값으로 대체\n ch = ''.join(standard) # 문자열 연결 \n\n if visited[int(ch)] == False: # 방문 체크 - 방문 안한경우만\n visited[int(ch)] = True \n if end == ch:\n return int(depth) # 최초 발견이 곧 최소횟수\n elif sosu.count(int(ch)) > 0: \n dq.append([ch, int(depth)+1]) \n\n standard[s] = temp # 기존값 복원\n\nif __name__ == \"__main__\":\n from collections import deque\n N = int(input())\n DP = [False for _ in range(10000)]\n sosu = []\n et_prime(DP)\n \n for i in range(N):\n a, b = input().split()\n visited = [False for _ in range(10000)]\n result = bfs(a, b)\n\n if result is None: print(\"Impossible\")\n else: print(result)\n\n\n\"\"\"\n3\n1033 8179\n1373 8017\n1033 1033\n\"\"\"","repo_name":"woghks778803/algorithm-study","sub_path":"backjoon/Gold/1963.py","file_name":"1963.py","file_ext":"py","file_size_in_byte":1874,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"31488606236","text":"import enum\n\nfrom django.db import models\nfrom django.core.exceptions import ValidationError\nfrom django.contrib.postgres.fields import JSONField\nfrom django.utils.translation import ugettext_lazy as _\nfrom .organization_model import Organization\n\nfrom os2datascanner.utils.system_utilities import time_now\nfrom os2datascanner.engine2.pipeline.messages import MatchesMessage\n\n\nclass DocumentReport(models.Model):\n scan_time = models.DateTimeField(null=True, db_index=True,\n verbose_name=_('scan time'))\n\n created_timestamp = models.DateTimeField(null=True,\n verbose_name=_('created timestamp'))\n\n organization = models.ForeignKey(Organization,\n null=True, blank=True,\n verbose_name=_('organization'),\n on_delete=models.PROTECT)\n\n path = models.CharField(max_length=2000, verbose_name=_(\"path\"),\n db_index=True)\n # It could be that the meta data should not be part of the jsonfield...\n data = JSONField(null=True)\n\n source_type = models.CharField(max_length=2000,\n verbose_name=_(\"source type\"))\n\n sensitivity = models.IntegerField(null=True, verbose_name=_(\"sensitivity\"))\n\n probability = models.FloatField(null=True, verbose_name=_(\"probability\"))\n\n # datasource_last_modified stores when the scanned file/email/element itself, has last been updated.\n # This timestamp is collected during scan and is from the datasource.\n datasource_last_modified = models.DateTimeField(null=True)\n\n def _str_(self):\n return self.path\n\n @property\n def matches(self):\n matches = self.data.get(\"matches\")\n return MatchesMessage.from_json_object(matches) if matches else None\n\n @enum.unique\n class ResolutionChoices(enum.Enum):\n # Future simplification note: the behaviour of the enumeration values\n # of this class is modelled on Django 3's model.Choices\n OTHER = 0, \"Andet\"\n EDITED = 1, \"Redigeret\"\n MOVED = 2, \"Flyttet\"\n REMOVED = 3, \"Slettet\"\n NO_ACTION = 4, \"Intet foretaget\"\n\n def __new__(cls, *args):\n obj = object.__new__(cls)\n # models.Choices compatibility: the last element of the enum value\n # tuple, if there is one, is a human-readable label\n obj._value_ = args[0] if len(args) < 3 else args[:-1]\n return obj\n\n def __init__(self, *args):\n self.label = args[-1] if len(args) > 1 else self.name\n\n # This is a class *property* in model.Choices, but that would require\n # sinister metaclass sorcery\n @classmethod\n def choices(cls):\n return [(k.value, k.label) for k in cls]\n\n resolution_status = models.IntegerField(choices=ResolutionChoices.choices(),\n null=True, blank=True, db_index=True,\n verbose_name=_(\"resolution status\"))\n\n resolution_time = models.DateTimeField(blank=True, null=True,\n verbose_name=_(\"resolution time\"))\n\n custom_resolution_status = models.CharField(max_length=1024, blank=True,\n verbose_name=_(\"justification\"))\n \n def clean(self):\n self.clean_custom_resolution_status()\n\n def clean_custom_resolution_status(self):\n self.custom_resolution_status = self.custom_resolution_status.strip()\n if self.resolution_status == 0 and not self.custom_resolution_status:\n raise ValidationError(\n {\n \"custom_resolution_status\":\n \"Resolution status 0 requires an\"\n \" explanation\"\n })\n\n def __init__(self, *args, **kwargs):\n # TODO: move to property/model method\n super().__init__(*args, **kwargs)\n self.__resolution_status = self.resolution_status\n\n def save(self, *args, **kwargs):\n now = time_now()\n\n # If Resolution status goes from not handled to handled - change resolution_time to now\n if self.__resolution_status == None and (self.resolution_status or self.resolution_status == 0):\n self.resolution_time = now\n\n # Adds a timestamp if it's a new match:\n if not self.pk:\n self.created_timestamp = now\n\n super().save(*args, **kwargs)\n\n # Add DocumentReport to Alias.match_relation, when it's saved to the db.\n from .aliases.alias_model import Alias\n try:\n metadata = self.data['metadata']['metadata'].values()\n value = list(metadata)[0]\n\n aliases = Alias.objects.select_subclasses()\n \n for alias in aliases:\n if str(alias) == value:\n try:\n tm = Alias.match_relation.through\n tm.objects.bulk_create([tm(documentreport_id=self.pk, alias_id=alias.pk)], ignore_conflicts=True)\n except:\n print(\"Failed to create match_relation\")\n except:\n print(self, \" has no metadata\")\n\n from ..views.views import send_socket_message\n send_socket_message()\n\n class Meta:\n verbose_name_plural = _(\"document reports\")\n ordering = ['-sensitivity', '-probability']\n","repo_name":"mBoegvald/bachelor-os2datascanner","sub_path":"src/os2datascanner/projects/report/reportapp/models/documentreport_model.py","file_name":"documentreport_model.py","file_ext":"py","file_size_in_byte":5563,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"73247043391","text":"from django.urls import path\nfrom . import views\nfrom . import api\n\n\napp_name = 'quiz'\n\nurlpatterns = [\n path(r'', views.quizside, name='quiz'),\n path(r'quiz/', views.quizcode, name='quizcode'),\n path(r'quiz/lag-quiz/', views.CreateQuiz.as_view(), name='create_quiz'),\n path(r'quiz//', views.play_quiz, name='play_quiz'),\n path(r'quiz//api/', api.QuizAPI.as_view(), name='play_quiz_api'),\n]\n","repo_name":"Fredrik3B/quiz","sub_path":"nettside/quiz/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":432,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"60"}
+{"seq_id":"15924637455","text":"import pandas as pd\nimport numpy as np\nimport io\nfrom upload import upload\nfrom datetime import timedelta, date\nimport requests\n\ndef import_csv_data(csv_text, entry_date):\n\t# load the CSV data\n\tstring_io = io.StringIO(csv_text)\n\tdf = pd.read_csv(string_io, keep_default_na=False, na_values=['___'])\n\t\n\t# define the columns\n\tlat_col = lng_col = \"\"\n\tcountry_col = province_col = county_col = \"\"\n\ttotal_col = death_col = recovered_col = \"\"\n\t\n\t# future-proofing\n\tfor col in df.columns:\n\t\tif 'lat' in col.lower(): lat_col = col\n\t\telif 'long' in col.lower(): lng_col = col\n\t\telif 'country' in col.lower(): country_col = col\n\t\telif 'province' in col.lower(): province_col = col\n\t\telif 'county' in col.lower(): county_col = col\n\t\telif \"death\" in col.lower(): death_col = col\n\t\telif \"dead\" in col.lower(): death_col = col\n\t\telif \"confirm\" in col.lower(): total_col = col\n\t\telif \"recover\" in col.lower(): recovered_col = col\n\t\n\tcontent = {\n\t\t'datapoint': [],\n\t\t'location': []\n\t}\n\n\tdf.sort_values(by=[col for col in [county_col, province_col, country_col] if col], ascending=False)\n\t\n\tfor _, row in df.iterrows():\n\t\t# Steps\n\t\t# 1. Find country, province, and county name\n\t\t# 2. Get the actual coronavirus numbers\n\t\t# 3. Estimate the location if we can\n\n\t\t# STEP 1 #\n\t\tcountry = row[country_col]\n\t\tprovince = row[province_col] if not pd.isnull(row[province_col]) else ''\n\t\tcounty = row[county_col] if county_col else ''\n\n\t\t## FOR DEBUG ##\n\t\tif (\"Korea\" not in country):\n\t\t\tcontinue\n\n\t\t# STEP 2 #\n\t\ttotal = row[total_col]\n\t\tdeaths = row[death_col]\n\t\trecovered = row[recovered_col]\n\n\t\tif not total: total = 0\n\t\tif not deaths: deaths = 0\n\t\tif not recovered: recovered = 0\n\n\t\tlocation_row = {}\n\t\tdatapoint_row = {}\n\n\t\tif province == 'Recovered':\n\t\t\tdatapoint_row = {\n\t\t\t\t\"country\": country,\n\t\t\t\t\"recovered\": recovered,\n\t\t\t\t\"entry_date\": entry_date\n\t\t\t}\n\n\t\t\tlocation_row = { \"country\": country }\n\t\telse:\n\t\t\tdatapoint_row = {\n\t\t\t\t\"country\": country,\n\t\t\t\t\"province\": province,\n\t\t\t\t\"county\": county,\n\t\t\t\t\"total\": total,\n\t\t\t\t\"deaths\": deaths,\n\t\t\t\t\"recovered\": recovered,\n\t\t\t\t\"entry_date\": entry_date\n\t\t\t}\n\n\t\t\tlocation_row = {\n\t\t\t\t\"country\": country,\n\t\t\t\t\"province\": province,\n\t\t\t\t\"county\": county\n\t\t\t}\n\n\t\t# Save the primary location data if we can\n\t\tif lat_col and lng_col:\n\t\t\tlat, lng = row[lat_col], row[lng_col]\n\t\t\tif lat and lng:\n\t\t\t\tlocation_row['latitude'] = lat\n\t\t\t\tlocation_row['longitude'] = lng\n\n\t\tcontent['location'].append(location_row)\n\t\tcontent['datapoint'].append(datapoint_row)\n\treturn content\n\ndef import_jhu_date(entry_date):\n\tdate_formatted = entry_date.strftime(\"%m-%d-%Y\")\n\tprint(\"\\rLoading data from JHU \" + date_formatted + '...', end='\\r')\n\t\n\t# download from Github\n\tgithub_raw_url = f\"https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/csse_covid_19_daily_reports/{date_formatted}.csv\"\n\tresponse = requests.get(github_raw_url, timeout=10)\n\t\n\tif response.status_code == 200:\n\t\treturn import_csv_data(response.text, entry_date)\n\telse:\n\t\tprint(\"404 not found\")\n\ndef import_jhu_date_range(date_1, date_2):\n\tnext_date = timedelta(days=1)\n\tcurrent_date = date_1\n\t\n\twhile current_date <= date_2:\n\t\tprint(\"Loading JHU data for\", current_date, \" \")\n\t\tresult = import_jhu_date(current_date)\n\t\tif result:\n\t\t\tyield result\n\t\tcurrent_date += next_date\n\ndef import_jhu_historical():\n\treturn import_jhu_date_range(date_1=date(2020, 3, 10), date_2=date.today())\n","repo_name":"myfatemi04/Corona-Vision","sub_path":"data_collection/data_imports/import_jhu.py","file_name":"import_jhu.py","file_ext":"py","file_size_in_byte":3419,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"21834804270","text":"import requests\r\n\r\ndef pegar_cotacoes():\r\n # Função que requisita a API informações sobre cotação atual \r\n requisicao = requests.get(\"https://economia.awesomeapi.com.br/last/USD-BRL,EUR-BRL,BTC-BRL\")\r\n\r\n requisicao_dic = requisicao.json()\r\n\r\n cotacao_dolar = requisicao_dic['USDBRL']['bid']\r\n cotacao_euro = requisicao_dic['EURBRL']['bid']\r\n cotacao_btc = requisicao_dic['BTCBRL']['bid']\r\n\r\n # estou chamando a label com a variável \"text\" que será completada com a geração do texto\r\n texto_cotacao[\"text\"] = f'''\r\n Dólar: {cotacao_dolar}\r\n Euro: {cotacao_euro}\r\n BTC: {cotacao_btc}'''\r\n\r\n\r\nimport tkinter as tk\r\n\r\n# inicia a janela \r\njanela1 = tk.Tk() \r\n\r\n# alterar titulo da janela\r\njanela1.title('Cotação atual') \r\n\r\n# label cria o texto na janela (param1 = janela, param2 = texto)\r\ntexto_janela = tk.Label(janela1, text=\"Clique no botão para ver a cotação das moedas\") \r\n# o grid define onde a Label do texto ficara (param1 = coluna, param2 = linha, param3 = espaçamento largura, param4 = espaçamento altura)\r\ntexto_janela.grid(column=1, row=0, padx=50, pady=20) \r\n\r\n# criando o botão (param1 = janela, param2 = texto, param3 = comando que o botão executara)\r\nbotao = tk.Button(janela1, text=\"Buscar Cotações\", command=pegar_cotacoes) \r\nbotao.grid(column=1, row=2)\r\n\r\ntexto_cotacao = tk.Label(janela1, text=\"\") # deixo o texto em branco (será preenchido na função)\r\ntexto_cotacao.grid(column=1, row=3, pady=20)\r\n\r\n\r\njanela1.mainloop() # mantém a janela aberta (última linha do código)","repo_name":"kakanetwork/Estudos-Linguagens","sub_path":"Prog-Py/Tkinter/Cotacao.py","file_name":"Cotacao.py","file_ext":"py","file_size_in_byte":1549,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"26095231648","text":"\"\"\"\ncolors = [\"red\", \"blue\", \"green\", \"purple\"]\nages = [14, 15, 15, 16, 16, 13, 18, 14, 15, 15, 14]\n\n\n# 5 - how do we remove an item from the list?\nprint(colors)\ndel colors[2]\nprint(colors)\n\"\"\"\ncolors = [\"red\", \"blue\", \"green\", \"purple\"]\nages = [14, 15, 15, 16, 16, 13, 18, 14, 15, 15, 14]\n\n\n# 6 - what happens when we use the append() function?\nprint(colors)\ncolors.append(\"orange\")\nprint(colors)\n","repo_name":"CNieves121/lps_compsci","sub_path":"class_samples/3-7_forloops/PartnerPractice.py","file_name":"PartnerPractice.py","file_ext":"py","file_size_in_byte":398,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"3310874999","text":"# -*- coding: utf-8 -*-\n# @Time : 2020/4/20 \n# @File : lr_scheduler.py\n# @Software: PyCharm\nimport tensorflow as tf\n\n\n\ndef MultiStepLR(initial_learning_rate, lr_steps, lr_rate, name='MultiStepLR'):\n \"\"\"Multi-steps learning rate scheduler.\"\"\"\n lr_steps_value = [initial_learning_rate]\n for _ in range(len(lr_steps)):\n lr_steps_value.append(lr_steps_value[-1] * lr_rate)\n return tf.keras.optimizers.schedules.PiecewiseConstantDecay(\n boundaries=lr_steps, values=lr_steps_value)\n\n\ndef MultiStepWarmUpLR(initial_learning_rate, lr_steps, lr_rate,\n warmup_steps=0., min_lr=0.,\n name='MultiStepWarmUpLR'):\n \"\"\"Multi-steps warm up learning rate scheduler.\"\"\"\n assert warmup_steps <= lr_steps[0]\n assert min_lr <= initial_learning_rate\n lr_steps_value = [initial_learning_rate]\n for _ in range(len(lr_steps)):\n lr_steps_value.append(lr_steps_value[-1] * lr_rate)\n return PiecewiseConstantWarmUpDecay(\n boundaries=lr_steps, values=lr_steps_value, warmup_steps=warmup_steps,\n min_lr=min_lr)\n\n\ndef CosineAnnealingLR_Restart(initial_learning_rate, t_period, lr_min):\n \"\"\"Cosine annealing learning rate scheduler with restart.\"\"\"\n return tf.keras.experimental.CosineDecayRestarts(\n initial_learning_rate=initial_learning_rate,\n first_decay_steps=t_period, t_mul=1.0, m_mul=1.0,\n alpha=lr_min / initial_learning_rate)\n\n\nclass PiecewiseConstantWarmUpDecay(\n tf.keras.optimizers.schedules.LearningRateSchedule):\n \"\"\"A LearningRateSchedule wiht warm up schedule.\n Modified from tf.keras.optimizers.schedules.PiecewiseConstantDecay\"\"\"\n\n def __init__(self, boundaries, values, warmup_steps, min_lr,\n name=None):\n super(PiecewiseConstantWarmUpDecay, self).__init__()\n\n if len(boundaries) != len(values) - 1:\n raise ValueError(\n \"The length of boundaries should be 1 less than the\"\n \"length of values\")\n\n self.boundaries = boundaries\n self.values = values\n self.name = name\n self.warmup_steps = warmup_steps\n self.min_lr = min_lr\n\n def __call__(self, step):\n with tf.name_scope(self.name or \"PiecewiseConstantWarmUp\"):\n step = tf.cast(tf.convert_to_tensor(step), tf.float32)\n pred_fn_pairs = []\n warmup_steps = self.warmup_steps\n boundaries = self.boundaries\n values = self.values\n min_lr = self.min_lr\n\n pred_fn_pairs.append(\n (step <= warmup_steps,\n lambda: min_lr + step * (values[0] - min_lr) / warmup_steps))\n pred_fn_pairs.append(\n (tf.logical_and(step <= boundaries[0],\n step > warmup_steps),\n lambda: tf.constant(values[0])))\n pred_fn_pairs.append(\n (step > boundaries[-1], lambda: tf.constant(values[-1])))\n\n for low, high, v in zip(boundaries[:-1], boundaries[1:],\n values[1:-1]):\n # Need to bind v here; can do this with lambda v=v: ...\n pred = (step > low) & (step <= high)\n pred_fn_pairs.append((pred, lambda: tf.constant(v)))\n\n # The default isn't needed here because our conditions are mutually\n # exclusive and exhaustive, but tf.case requires it.\n return tf.case(pred_fn_pairs, lambda: tf.constant(values[0]),\n exclusive=True)\n\n def get_config(self):\n return {\n \"boundaries\": self.boundaries,\n \"values\": self.values,\n \"warmup_steps\": self.warmup_steps,\n \"min_lr\": self.min_lr,\n \"name\": self.name\n }","repo_name":"PureHing/face-mask-detection-tf2","sub_path":"components/lr_scheduler.py","file_name":"lr_scheduler.py","file_ext":"py","file_size_in_byte":3816,"program_lang":"python","lang":"en","doc_type":"code","stars":73,"dataset":"github-code","pt":"60"}
+{"seq_id":"9234704222","text":"# -*- coding: utf-8 -*-\n# !/usr/bin/env python\n\"\"\"\n-------------------------------------------------\n File Name: GetFreeProxy.py\n Description : 抓取免费代理\n Author : JHao\n date: 2016/11/25\n-------------------------------------------------\n Change Activity:\n 2016/11/25:\n-------------------------------------------------\n\"\"\"\nimport re\nimport sys\nimport requests\n\ntry:\n from importlib import reload # py3 实际不会实用,只是为了不显示语法错误\nexcept:\n reload(sys)\n sys.setdefaultencoding('utf-8')\n\nsys.path.append('..')\n\nfrom Util.WebRequest import WebRequest\nfrom Util.utilFunction import getHtmlTree\nfrom Util.utilFunction import verifyProxyFormat\n\n# for debug to disable insecureWarning\nrequests.packages.urllib3.disable_warnings()\n\n\"\"\"\n data5u.com\n 66ip.cn\n 31f.cn\n xicidaili.com\n goubanjia.com\n kxdaili.com\n kuaidaili.com\n xsdaili.com\n zdaye.com\n ip3366.net\n iphai.com\n jiangxianli.com\n feiyiproxy.com\n qydaili.com\n\"\"\"\n\n\nclass GetFreeProxy(object):\n \"\"\"\n proxy getter\n \"\"\"\n\n def __init__(self):\n pass\n\n @staticmethod\n def freeProxyFirst():\n \"\"\"\n 无忧代理 http://www.data5u.com/\n 几乎没有能用的\n :return:\n \"\"\"\n url_list = [\n 'http://www.data5u.com/',\n 'http://www.data5u.com/free/gngn/index.shtml',\n 'http://www.data5u.com/free/gnpt/index.shtml'\n ]\n for url in url_list:\n html_tree = getHtmlTree(url)\n ul_list = html_tree.xpath('//ul[@class=\"l2\"]')\n for ul in ul_list:\n try:\n yield ':'.join(ul.xpath('.//li/text()')[0:2])\n except Exception as e:\n print(e)\n\n @staticmethod\n def freeProxySecond(area=34):\n \"\"\"\n 代理66 http://www.66ip.cn/\n :param area: 抓取代理页数,page=1北京代理页,page=2上海代理页......\n :return:\n \"\"\"\n area = 34 if area > 34 else area\n for area_index in range(1, area + 1):\n url = \"http://www.66ip.cn/areaindex_{}/1.html\".format(str(area_index))\n html_tree = getHtmlTree(url)\n tr_list = html_tree.xpath(\"//div[@id='footer']//table//tr[position()>1]\")\n if len(tr_list) == 0:\n continue\n for tr in tr_list:\n yield tr.xpath(\"./td[1]/text()\")[0] + \":\" + tr.xpath(\"./td[2]/text()\")[0]\n\n @staticmethod\n def freeProxyThird():\n \"\"\"\n 31代理 http://31f.cn/http-proxy/\n :return:\n \"\"\"\n urls = ['http://31f.cn/http-proxy/', 'http://31f.cn/https-proxy/']\n for url in urls:\n html_tree = getHtmlTree(url)\n try:\n tr_list = html_tree.xpath('//table[@class=\"table table-striped\"]//tr[position()>1]')\n for tr in tr_list:\n try:\n if '天' in tr.xpath('./td[9]/text()')[0]:\n continue\n yield tr.xpath('./td[2]/text()')[0] + ':' + tr.xpath('./td[3]/text()')[0]\n except Exception as e:\n print(e)\n except Exception as e:\n print(e)\n\n @staticmethod\n def freeProxyFourth(page_count=2):\n \"\"\"\n 西刺代理 http://www.xicidaili.com\n :return:\n \"\"\"\n url_list = [\n 'http://www.xicidaili.com/nn/', # 高匿\n 'http://www.xicidaili.com/nt/', # 透明\n ]\n for each_url in url_list:\n for i in range(1, page_count + 1):\n page_url = each_url + str(i)\n tree = getHtmlTree(page_url)\n proxy_list = tree.xpath('.//table[@id=\"ip_list\"]//tr[position()>1]')\n for proxy in proxy_list:\n try:\n yield ':'.join(proxy.xpath('./td/text()')[0:2])\n except Exception as e:\n pass\n\n @staticmethod\n def freeProxyFifth():\n \"\"\"\n guobanjia http://www.goubanjia.com/\n :return:\n \"\"\"\n url = \"http://www.goubanjia.com/\"\n tree = getHtmlTree(url)\n proxy_list = tree.xpath('//td[@class=\"ip\"]')\n # 此网站有隐藏的数字干扰,或抓取到多余的数字或.符号\n # 需要过滤掉
的内容\n xpath_str = \"\"\".//*[not(contains(@style, 'display: none'))\n and not(contains(@style, 'display:none'))\n and not(contains(@class, 'port'))\n ]/text()\n \"\"\"\n for each_proxy in proxy_list:\n try:\n # :符号裸放在td下,其他放在div span p中,先分割找出ip,再找port\n ip_addr = ''.join(each_proxy.xpath(xpath_str))\n port = each_proxy.xpath(\".//span[contains(@class, 'port')]/text()\")[0]\n yield '{}:{}'.format(ip_addr, port)\n except Exception as e:\n pass\n\n @staticmethod\n def freeProxySixth():\n \"\"\"\n 开心代理 http://ip.kxdaili.com/dailiip.html\n :return:\n \"\"\"\n urls = ['http://ip.kxdaili.com/dailiip/1/{}.html#ip'.format(str(page)) for page in range(1, 8)]\n for url in urls:\n try:\n html_tree = getHtmlTree(url)\n tr_list = html_tree.xpath('//table[@class=\"ui table segment\"]//tbody//tr')\n for tr in tr_list:\n try:\n yield tr.xpath('./td[1]/text()')[0] + ':' + tr.xpath('./td[2]/text()')[0]\n except Exception as e:\n print(e)\n except Exception as e:\n print(e)\n\n @staticmethod\n def freeProxySeventh():\n \"\"\"\n 快代理 https://www.kuaidaili.com\n \"\"\"\n url_list = [\n 'https://www.kuaidaili.com/free/inha/{page}/',\n 'https://www.kuaidaili.com/free/intr/{page}/'\n ]\n for url in url_list:\n for page in range(1, 3):\n page_url = url.format(page=page)\n tree = getHtmlTree(page_url)\n proxy_list = tree.xpath('.//table//tr')\n for tr in proxy_list[1:]:\n yield ':'.join(tr.xpath('./td/text()')[0:2])\n\n @staticmethod\n def freeProxyEight():\n \"\"\"\n 小舒代理 http://www.xsdaili.com/\n \"\"\"\n url = 'http://www.xsdaili.com/'\n html_tree = getHtmlTree(url)\n new_url = url + html_tree.xpath('//div[@class=\"col-md-12\"]/div[1]//a[1]/@href')[0]\n new_html_tree = getHtmlTree(new_url)\n proxy_list = new_html_tree.xpath('//div[@class=\"cont\"]/text()')\n for proxy in proxy_list:\n try:\n yield proxy.split('@')[0].strip()\n except Exception as e:\n print(e)\n\n @staticmethod\n def freeProxyNinth():\n \"\"\"\n 站大爷代理 http://ip.zdaye.com/\n :return:\n \"\"\"\n url = 'http://ip.zdaye.com/'\n html_tree = getHtmlTree(url)\n item_list = html_tree.xpath('//div[@class=\"Loglist\"]/div[2]/div[@class=\"panel-body\"]//a/text()')\n for item in item_list:\n try:\n yield item.split('@')[0].strip()\n except Exception as e:\n print(e)\n\n header = {\n 'Referer': 'http://ip.zdaye.com/',\n }\n new_urls = html_tree.xpath('//div[@class=\"Loglist\"]/div[1]/div[@class=\"panel-body\"]//a/@href')\n for new_url in new_urls:\n try:\n new_html_tree = getHtmlTree(url + new_url, header=header)\n new_item_list = new_html_tree.xpath('//div[@class=\"cont\"]/text()')\n for new_item in new_item_list:\n try:\n yield new_item.split('@')[0].strip()\n except Exception as e:\n print(e)\n except Exception as e:\n print(e)\n\n @staticmethod\n def freeProxyTen():\n \"\"\"\n 云代理 http://www.ip3366.net/free/\n :return:\n \"\"\"\n urls = ['http://www.ip3366.net/free/?stype=1&page={}'.format(str(i)) for i in range(1, 4)]\n request = WebRequest()\n for url in urls:\n r = request.get(url)\n proxies = re.findall(r'
(\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3})
[\\s\\S]*?
(\\d+)
', r.text)\n for proxy in proxies:\n yield \":\".join(proxy)\n\n @staticmethod\n def freeProxyEleven():\n \"\"\"\n IP海 http://www.iphai.com/free/ng\n :return:\n \"\"\"\n urls = [\n 'http://www.iphai.com/free/ng',\n 'http://www.iphai.com/free/np',\n 'http://www.iphai.com/free/wg',\n 'http://www.iphai.com/free/wp'\n ]\n request = WebRequest()\n for url in urls:\n r = request.get(url)\n proxies = re.findall(r'
',\n r.text)\n for proxy in proxies:\n yield \":\".join(proxy)\n\n @staticmethod\n def freeProxyTwelve(page_count=8):\n \"\"\"\n guobanjia http://ip.jiangxianli.com/?page=\n 免费代理库\n 超多量\n :return:\n \"\"\"\n for i in range(1, page_count + 1):\n url = 'http://ip.jiangxianli.com/?page={}'.format(i)\n html_tree = getHtmlTree(url)\n tr_list = html_tree.xpath(\"/html/body/div[1]/div/div[1]/div[2]/table/tbody/tr\")\n if len(tr_list) == 0:\n continue\n for tr in tr_list:\n yield tr.xpath(\"./td[2]/text()\")[0].strip() + \":\" + tr.xpath(\"./td[3]/text()\")[0].strip()\n\n @staticmethod\n def freeProxyThirteen():\n \"\"\"\n 飞蚁代理 http://www.feiyiproxy.com/?page_id=1457\n :return:\n \"\"\"\n url = 'http://www.feiyiproxy.com/?page_id=1457'\n html_tree = getHtmlTree(url)\n tr_list = html_tree.xpath('//div[@class=\"et_pb_code et_pb_module et_pb_code_1\"]//tr[position()>1]')\n for tr in tr_list:\n yield tr.xpath('./td[1]/text()')[0].strip() + ':' + tr.xpath('./td[2]/text()')[0].strip()\n\n @staticmethod\n def freeProxyFourteen():\n \"\"\"\n 旗云代理 http://www.qydaili.com/free/?action=china&page=\n :return:\n \"\"\"\n urls = ['http://www.qydaili.com/free/?action=china&page={}'.format(page) for page in range(1, 4)]\n for url in urls:\n html_tree = getHtmlTree(url)\n tr_list = html_tree.xpath('//table[@class=\"table table-bordered table-striped\"]//tbody//tr')\n for tr in tr_list:\n yield tr.xpath('./td[1]/text()')[0].strip() + ':' + tr.xpath('./td[2]/text()')[0].strip()\n\n @staticmethod\n def freeProxyWallFirst():\n \"\"\"\n 墙外网站 cn-proxy\n :return:\n \"\"\"\n urls = ['http://cn-proxy.com/', 'http://cn-proxy.com/archives/218']\n request = WebRequest()\n for url in urls:\n r = request.get(url)\n proxies = re.findall(r'
(\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3})
[\\w\\W]
(\\d+)
', r.text)\n for proxy in proxies:\n yield ':'.join(proxy)\n\n @staticmethod\n def freeProxyWallSecond():\n \"\"\"\n https://proxy-list.org/english/index.php\n :return:\n \"\"\"\n urls = ['https://proxy-list.org/english/index.php?p=%s' % n for n in range(1, 10)]\n request = WebRequest()\n import base64\n for url in urls:\n r = request.get(url)\n proxies = re.findall(r\"Proxy\\('(.*?)'\\)\", r.text)\n for proxy in proxies:\n yield base64.b64decode(proxy).decode()\n\n @staticmethod\n def freeProxyWallThird():\n urls = ['https://list.proxylistplus.com/Fresh-HTTP-Proxy-List-1']\n request = WebRequest()\n for url in urls:\n r = request.get(url)\n proxies = re.findall(r'
(\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3})
[\\s\\S]*?
(\\d+)
', r.text)\n for proxy in proxies:\n yield ':'.join(proxy)\n\n\nif __name__ == '__main__':\n from CheckProxy import CheckProxy\n\n CheckProxy.checkGetProxyFunc(GetFreeProxy.freeProxyFourteen)\n # CheckProxy.checkGetProxyFunc(GetFreeProxy.freeProxySecond)\n #\n # CheckProxy.checkAllGetProxyFunc()\n\n","repo_name":"AndrewZStore/private_proxy_pool","sub_path":"ProxyGetter/getFreeProxy.py","file_name":"getFreeProxy.py","file_ext":"py","file_size_in_byte":12603,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"36548730197","text":"from unittest import TestCase\n\nfrom unittest.mock import Mock, patch, sentinel as s\nfrom xivo import config_helper\n\nfrom ..controller import Controller\nfrom ..http_server import api, app\n\n\nclass TestController(TestCase):\n def setUp(self):\n self.http_server = (\n patch('wazo_phoned.controller.HTTPServer').start().return_value\n )\n self.plugin_manager = patch('wazo_phoned.controller.plugin_helpers').start()\n self.token_renewer = (\n patch('wazo_phoned.controller.TokenRenewer').start().return_value\n )\n self.bus_consumer = (\n patch('wazo_phoned.controller.CoreBusConsumer').start().return_value\n )\n self.status_aggregator = (\n patch('wazo_phoned.controller.StatusAggregator').start().return_value\n )\n config_helper.get_xivo_uuid = Mock(return_value='VALID-UUID')\n\n def tearDown(self):\n patch.stopall()\n\n def test_run_starts_http_server(self):\n config = self._create_config(**{'rest_api': {}, 'debug': s.debug})\n controller = Controller(config)\n controller.run()\n self.http_server.run.assert_called_once_with()\n\n def test_run_loads_plugins(self):\n config = self._create_config(\n **{'enabled_plugins': {'cisco': True, 'aastra': False}}\n )\n\n controller = Controller(config)\n controller.run()\n\n self.plugin_manager.load.assert_called_once_with(\n namespace='wazo_phoned.plugins',\n names=config['enabled_plugins'],\n dependencies={\n 'config': config,\n 'api': api,\n 'app': app,\n 'token_changed_subscribe': self.token_renewer.subscribe_to_token_change,\n 'bus_consumer': self.bus_consumer,\n 'status_aggregator': self.status_aggregator,\n 'phone_plugins': controller.phone_plugins,\n },\n )\n\n def _create_config(self, **kwargs):\n config = dict(kwargs)\n config.setdefault(\n 'auth',\n {\n 'host': 'localhost',\n 'port': 9497,\n 'verify_certificate': False,\n 'service_id': 'phoned',\n 'service_key': '123',\n },\n )\n config.setdefault('dird', {})\n config['dird'].setdefault('host', '')\n config['dird'].setdefault('port', '')\n config.setdefault(\n 'bus',\n {\n 'username': 'guest',\n 'password': 'guest',\n 'host': 'localhost',\n 'port': 5672,\n 'subscribe_exchange_name': 'wazo-headers',\n 'subscribe_exchange_type': 'headers',\n },\n )\n config.setdefault('rest_api', {})\n config['rest_api'].setdefault('authorized_subnets', [])\n config.setdefault('enabled_plugins', {})\n return config\n","repo_name":"wazo-platform/wazo-phoned","sub_path":"wazo_phoned/tests/test_controller.py","file_name":"test_controller.py","file_ext":"py","file_size_in_byte":2932,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"60"}
+{"seq_id":"26210473775","text":"import os \r\nimport logging\r\n\r\ndef create_logger(name, log_file, level=logging.INFO):\r\n\r\n folder = log_file.replace(log_file.split('/')[-1], \"\")\r\n if os.path.isdir(folder) is False:\r\n os.mkdir(folder)\r\n\r\n \"\"\"Function setup as many loggers as you want\"\"\"\r\n logging.basicConfig(filename = log_file, level = level, format = '%(asctime)s %(levelname)s %(message)s')\r\n handler = logging.StreamHandler() \r\n\r\n logger = logging.getLogger(name)\r\n logger.addHandler(handler)\r\n\r\n return logger\r\n ","repo_name":"Baron1014/credict_card_recsys","sub_path":"logger.py","file_name":"logger.py","file_ext":"py","file_size_in_byte":521,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"16399145480","text":"import os\r\n\r\nFILE_LOC = \"data.txt\"\r\n\r\n# color related code for better view on terminal\r\ncolors = {\r\n 'DANGER': '\\033[91m',\r\n 'WARNING': '\\033[33m',\r\n 'MILD': '\\033[94m',\r\n 'OK': '\\033[92m',\r\n 'RESET': '\\033[0m',\r\n 'BLUE': '\\033[94m',\r\n 'CYAN': '\\033[96m',\r\n}\r\n\r\nformat = {0: 5, 1:25, 2:10, 3:25, 4:5}\r\n\r\n# function to get the colored text\r\ndef color(text, text_color):\r\n if text_color in colors:\r\n return ''.join([colors[text_color], text, colors['RESET']])\r\n return text\r\n\r\n# the help text\r\nprint(\"\"\"Usage of the ToDo App CLI\\nWe support following commands as of now\\n\"\"\")\r\nprint('Command List')\r\nprint(\"\\t\" + color(\"Add item\", 'CYAN') + \" \" + color('--add', 'OK'))\r\n# print(\"\\t\" + color(\"Add due date\", 'CYAN') + \" \" + color('--due', 'OK'))\r\nprint(\"\\t\" + color(\"Remove an Item\", 'CYAN') + \" \" + color('--remove', 'OK'))\r\nprint(\"\\t\" + color(\"Mark item done\", 'CYAN') + \" \" + color('--done', 'OK'))\r\nprint(\"\\t\" + color(\"Mark item not done\", 'CYAN') + \" \" + color('--undone', 'OK'))\r\nprint(\"\\t\" + color(\"View items\", 'CYAN') + \" \" + color('--view', 'OK'))\r\n\r\ncmd = input(\"Enter the command \")\r\n\r\nif cmd:\r\n # Add the item\r\n if cmd == \"--add\":\r\n print(color(\"Add your task in following format\", 'BLUE'))\r\n print(color(\"'Title', 'Due Date', 'Description'\\n\", 'CYAN'))\r\n data = input()\r\n data = data.split(\",\")\r\n data = \", \".join(data)\r\n if data:\r\n with open(FILE_LOC, 'r+') as f:\r\n last_line = f.readlines()[-1]\r\n if last_line:\r\n try:\r\n count = int(last_line.split(\",\")[0])+1\r\n except ValueError:\r\n count = 1\r\n f.write(str(count)+\", \"+data+\", [ ]\"\"\\n\")\r\n print(color(\"Item successfully added!\", 'CYAN'))\r\n # remove the item\r\n elif cmd == \"--remove\":\r\n print(\"\\n\"+color(\"Remove your task by providing the id of the task\", 'BLUE'))\r\n try:\r\n id = int(input(\"Enter the id of the task you want to delete \"))\r\n except ValueError:\r\n raise(\"Please enter the correct id value\")\r\n with open(FILE_LOC, 'r') as f:\r\n lines = f.readlines()\r\n with open(FILE_LOC, 'w+') as f:\r\n count = 0\r\n for line in lines:\r\n if count == 0:\r\n f.write(line)\r\n count += 1\r\n continue\r\n no = int(line.split(\", \")[0])\r\n if no != id:\r\n f.write(line)\r\n count += 1\r\n print(color(\"Item successfully removed!\", 'CYAN')) \r\n # mark the item as done \r\n elif cmd == \"--done\":\r\n print(\"\\n\"+color(\"Mark your task as done by providing the id of the task\", 'BLUE'))\r\n try:\r\n id = int(input(\"Enter the id of the task you want to mark as done \"))\r\n except ValueError:\r\n raise(\"Please enter the correct id value\")\r\n with open(FILE_LOC, 'r') as f:\r\n lines = f.readlines()\r\n with open(FILE_LOC, 'w+') as f:\r\n count = 0\r\n for line in lines:\r\n if count == 0:\r\n f.write(line)\r\n count += 1\r\n continue\r\n no = int(line.split(\", \")[0])\r\n if no == id:\r\n line_text = line.split(\", \")\r\n line_text[-1] = \"[x]\\n\"\r\n line = \", \".join(line_text)\r\n f.write(line)\r\n else:\r\n f.write(line)\r\n count += 1\r\n print(color(\"Item successfully updated!\", 'CYAN')) \r\n # mark the item as undone\r\n elif cmd == \"--undone\":\r\n print(\"\\n\"+color(\"Mark your task as undone by providing the id of the task\", 'BLUE'))\r\n try:\r\n id = int(input(\"Enter the id of the task you want to mark as undone \"))\r\n except ValueError:\r\n raise(\"Please enter the correct id value\")\r\n with open(FILE_LOC, 'r') as f:\r\n lines = f.readlines()\r\n with open(FILE_LOC, 'w+') as f:\r\n count = 0\r\n for line in lines:\r\n if count == 0:\r\n f.write(line)\r\n count += 1\r\n continue\r\n no = int(line.split(\", \")[0])\r\n if no == id:\r\n line_text = line.split(\", \")\r\n line_text[-1] = \"[ ]\\n\"\r\n line = \", \".join(line_text)\r\n f.write(line)\r\n else:\r\n f.write(line)\r\n count += 1\r\n print(color(\"Item successfully updated!\", 'CYAN')) \r\n # list all the todo items\r\n elif cmd == \"--view\":\r\n with open(FILE_LOC, 'r') as f:\r\n lines = f.readlines()\r\n count = 0\r\n for line in lines:\r\n data = []\r\n line = line.split(\", \")\r\n for i in range(len(line)):\r\n data.append(line[i].strip().ljust(format[i]))\r\n if count == 0:\r\n print(color(\" \".join(data), 'OK'))\r\n else:\r\n print(color(\" \".join(data), 'RESET'))\r\n count += 1 \r\n else:\r\n print(\"Command not found, please run the script again and check the help\")\r\n","repo_name":"inovizz/python-for-devops","sub_path":"apps/todo-cli/cli0.py","file_name":"cli0.py","file_ext":"py","file_size_in_byte":5462,"program_lang":"python","lang":"en","doc_type":"code","stars":22,"dataset":"github-code","pt":"60"}
+{"seq_id":"12767091250","text":"import sys\nimport socket\nfrom datetime import datetime\n\n\ndef main():\n run = True\n port = sys.argv[1]\n ip_parent = sys.argv[2]\n parent_port = sys.argv[3]\n ips_file = sys.argv[4]\n parent_socket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # connecting to parent server\n client_socket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # connecting to client (to me)\n client_socket.bind(('', int(port)))\n\n ips_list = []\n file = open(ips_file)\n for lines in file:\n lines = lines.strip(\"\\n\")\n lines = lines.split(\",\")\n # the time - alive remaining from the ttl.\n lines.append(datetime(1, 1, 1, 1, 1, 1)) # initial time.\n lines.append(\"0\") # if its 0 - from ips.txt, don't consider ttl\n ips_list.append(lines)\n\n ips_dict = {x[0]: x[1:] for x in ips_list} # convert the list of lists to dictionary\n\n while run:\n flag = 0\n data, addr = client_socket.recvfrom(1024)\n data = data.decode()\n\n if data in ips_dict:\n desired_ip = ips_dict[data][0]\n desired_ttl = ips_dict[data][1]\n ip_and_ttl_reply = desired_ip + \",\" + desired_ttl\n # if the ttl not finished then we can sent the entry to the client, o.w we need to ask parent\n if is_this_entry_relevant(desired_ttl, ips_dict[data][2], ips_dict[data][3]):\n client_socket.sendto(ip_and_ttl_reply.encode(), addr)\n update_file(ips_file, ips_dict)\n else:\n flag = 1\n else:\n flag = 1\n\n if flag == 1: # need to communicate with parent server tp get the entry\n parent_socket.sendto(data.encode(), (ip_parent, int(parent_port)))\n parent_data, parent_add = parent_socket.recvfrom(1024) # receive answer from parent\n parent_data = parent_data.decode()\n # add to dictionary\n first_add, second_add = parent_data.split(',') #ip and ttl accordingly.\n third_add = datetime.now().strftime('%Y-%m-%d %H:%M:%S.%f')\n fourth_add = \"1\" # because this entry come from parent\n ips_dict[data] = first_add, second_add, third_add, fourth_add\n reply = ips_dict[data][0] + \",\" + ips_dict[data][1]\n client_socket.sendto(reply.encode(), addr)\n update_file(ips_file, ips_dict)\n client_socket.close()\n parent_socket.close()\n\n\ndef is_this_entry_relevant(ttl, remaining_time, is_from_parent):\n if is_from_parent == \"0\":\n # that means this entry is from txt\n return True\n # else - we check the ttl\n remaining_time = datetime.strptime(remaining_time, '%Y-%m-%d %H:%M:%S.%f') # convert str to datetime.\n result = (datetime.now() - remaining_time).total_seconds()\n return result < float(ttl)\n\n\ndef update_file(file_name, dic):\n with open(file_name, 'w') as file:\n file.truncate() # delete all file content.\n for name in dic:\n ip_address = dic[name][0]\n ttl = dic[name][1]\n remaining = dic[name][2]\n parent_flag = dic[name][3]\n if not is_this_entry_relevant(ttl, remaining, parent_flag):\n continue\n str_write = name + \",\" + ip_address + \",\" + ttl + \"\\n\"\n file.write(str_write) \n for name in list(dic.keys()):\n ip_address = dic[name][0]\n ttl = dic[name][1]\n remaining = dic[name][2]\n parent_flag = dic[name][3]\n if not is_this_entry_relevant(ttl, remaining, parent_flag):\n del dic[name]\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"YaelSim/ex1-networks","sub_path":"server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":3629,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"15029483394","text":"from fastapi import FastAPI, BackgroundTasks, Path\nimport asyncio\nfrom threading import Thread, Lock\n\nfrom worker import Worker\n\napp = FastAPI()\nworker = Worker()\n\nThread(target=worker.run_worker).start()\n\n@app.get(\"/add\")\nasync def add_item(\n background_tasks: BackgroundTasks, \n num: int = Path(..., title=\"The num might be positive integer\", ge=0, le=1000), \n timeout: int = Path(5, title=\"The num might be in seconds\", ge=0, le=1000)\n):\n background_tasks.add_task(worker.add_task, num, timeout)\n return {\"message\": \"task has added\"}\n\n@app.get(\"/list\")\nasync def show_list():\n return worker._list\n\n@app.get(\"/queue\")\nasync def show_queue():\n return worker._queue","repo_name":"kordimsan/async-api-web-app","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":686,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"71950170111","text":"import ipfblock\nimport ioport\nimport ipf.ipfblock.processing\nfrom ipf.ipftype.ipfimage3ctype import IPFImage3cType\n\n\nclass Arithmetic(ipfblock.IPFBlock):\n \"\"\" Abstract arithmetic block, \n \n Works with two images, produces one result image of same size \n \n \"\"\"\n type = \"Arithmetic\"\n category = \"Arithmetic and logic\"\n is_abstract_block = True\n \n def __init__(self):\n super(Arithmetic, self).__init__()\n self.input_ports[\"input_image_1\"] = ioport.IPort(self, IPFImage3cType)\n self.input_ports[\"input_image_2\"] = ioport.IPort(self, IPFImage3cType)\n self.output_ports[\"output_image\"] = ioport.OPort(self, IPFImage3cType)\n \n\n def get_preview_image(self):\n return self.output_ports[\"output_image\"]._value \n \n \n \n\n","repo_name":"anton-golubkov/Garland","sub_path":"src/ipf/ipfblock/arithmetic.py","file_name":"arithmetic.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"34991489356","text":"from collections import defaultdict\nfrom timeit import default_timer as timer\nimport sys\n\n\n# sys.setrecursionlimit(50000)\nwith open(\"knapsack1.txt\", \"r\") as f:\n first_line = [int(i) for i in f.readline().split()]\n W = first_line[0]\n n = first_line[1]\n data = f.readlines()\n v = [0]\n w = [0]\n for line in data:\n l = [int(j) for j in line.split()]\n v.append(l[0])\n w.append(l[1])\n\n\ndef multi_dict(k):\n if k == 1:\n return defaultdict()\n else:\n return defaultdict(lambda: multi_dict(k-1))\n\n\n# direct method\nA = multi_dict(2)\nfor x in range(W+1):\n A[0][x] = 0\nfor i in range(1, n+1):\n if i > 1:\n del A[i-2]\n for x in range(0, W+1):\n if x-w[i] >= 0:\n A[i][x] = max(A[i-1][x], A[i-1][x-w[i]]+v[i])\n else:\n A[i][x] = A[i-1][x]\n\nprint(A[n][W])\n\n\n# def knapSack(Wt, wt, vt, nt):\n# if nt == 0 or Wt == 0:\n# return 0\n# if wt[n-1] > Wt:\n# return knapSack(Wt, wt, vt, n-1)\n# else:\n# return max(\n# vt[n-1]+knapSack(Wt-wt[nt-1], wt, vt, n-1),\n# knapSack(Wt, wt, vt, n-1)\n# )\n#\n#\n# print(knapSack(W, w, v, n))","repo_name":"zhzeshu/Algorithms_Coursera","sub_path":"week12.py","file_name":"week12.py","file_ext":"py","file_size_in_byte":1174,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"75310281792","text":"import csv\nfrom matplotlib import pyplot as plt\nfrom matplotlib.ticker import MultipleLocator\nfrom matplotlib import rcParams\n\nconfig = {\n \"font.family\": \"serif\",\n # \"font.size\": 24,\n \"mathtext.fontset\": \"stix\",\n \"font.serif\": [\"SimHei\"],\n}\nrcParams.update(config)\n\n\ndef get_data(fliename, i, rounds):\n train_accuracy = []\n with open(\"{}.csv\".format(fliename), \"r\", encoding=\"utf-8\") as f:\n f_read = csv.reader(f)\n num = 0\n for v in f_read:\n\n if len(v) != 0 and num != 0:\n train_accuracy.append(float(v[i]))\n num += 1\n if num == rounds + 1:\n break\n return train_accuracy\n\n\ndef plot_MI(file, order, name, title, save_name, rounds, index='Testing Accuracy'):\n \"\"\"\n Verify the validity of mutual information\n \"\"\"\n test_accuracy = get_data(file, order[0], rounds)\n test_accuracy1 = get_data(file, order[1], rounds)\n test_accuracy2 = get_data(file, order[2], rounds)\n plt.figure()\n\n plt.plot(range(len(test_accuracy)), test_accuracy, marker=\".\", markersize=3,\n label=u'' + name[0], linewidth=1.0, color='r')\n plt.plot(range(len(test_accuracy1)), test_accuracy1, marker=\".\", markersize=3,\n label=u'' + name[1], linewidth=1.0, color='b')\n plt.plot(range(len(test_accuracy2)), test_accuracy2, marker=\".\", markersize=3,\n label=u'' + name[2], linewidth=1.0, color='purple')\n\n plt.tick_params(axis='both', which='major', direction='in', width=1, labelsize=10) # 刻度width=2\n\n ax = plt.gca()\n\n ax.yaxis.set_major_locator(MultipleLocator(0.01))\n ax.xaxis.set_major_locator(MultipleLocator(rounds / 10))\n\n # ax.set_ylim(0.64, 0.76)\n ax.set_ylim(0.92, 0.97)\n ax.set_xlim(0, rounds)\n\n ax.spines['bottom'].set_linewidth(0.6) # 设置底部坐标轴的粗细\n ax.spines['left'].set_linewidth(0.6) # 设置左边坐标轴的粗细\n ax.spines['right'].set_linewidth(0.6) # 设置右边坐标轴的粗细\n ax.spines['top'].set_linewidth(0.6)\n\n plt.legend(loc='lower right', fontsize=10)\n\n plt.text(0.02, 0.94, s='$\\mathrm{(b)}$', fontsize=10, transform=ax.transAxes)\n\n ax.set_ylabel(index, fontsize=10)\n ax.set_xlabel('Communication Rounds', fontsize=10)\n plt.title(title, fontsize=10)\n\n plt.rcParams['savefig.dpi'] = 1000\n plt.rcParams['figure.dpi'] = 1000\n plt.savefig('zn' + save_name, format='jpg', transparent=True)\n plt.show()\n\n\ndef plot(file, order, name, title, save_name, rounds, index='Testing Accuracy'):\n \"\"\"\n Compare the difference weighting schemes(Loss and Avg)\n \"\"\"\n test_accuracy = get_data(file, order[0], rounds)\n test_accuracy1 = get_data(file, order[1], rounds)\n plt.figure()\n\n plt.plot(range(len(test_accuracy)), test_accuracy, marker=\".\", markersize=3, markevery=3,\n label=u'' + name[0], linewidth=1.0, color='r')\n plt.plot(range(len(test_accuracy1)), test_accuracy1, marker=\".\", markersize=3, markevery=3,\n label=u'' + name[1], linewidth=1.0, color='b')\n\n plt.tick_params(axis='both', which='major', direction='in', width=1, labelsize=10) # 刻度width=2\n\n ax = plt.gca()\n\n ax.yaxis.set_major_locator(MultipleLocator(0.02))\n ax.xaxis.set_major_locator(MultipleLocator(rounds / 10))\n\n plt.ylim(0.64, 0.76)\n plt.xlim(0, rounds)\n ax.spines['bottom'].set_linewidth(0.6)\n ax.spines['left'].set_linewidth(0.6)\n ax.spines['right'].set_linewidth(0.6)\n ax.spines['top'].set_linewidth(0.6)\n\n plt.legend(loc='lower right', fontsize=10)\n plt.text(0.02, 0.94, s='$\\mathrm{(a)}$', fontsize=10, transform=ax.transAxes)\n\n plt.ylabel(index, fontsize=10)\n plt.xlabel('Communication Rounds', fontsize=13) # 'Communication Rounds'\n plt.title(title, fontsize=10)\n\n # 保存\n plt.rcParams['savefig.dpi'] = 1000\n plt.rcParams['figure.dpi'] = 1000\n plt.savefig('zn' + save_name, format='jpg', transparent=True)\n plt.show()\n\n\nif __name__ == '__main__':\n rounds = 50\n # file = '../exp/fmnist15_w'\n # order = [1, 2, 0]\n # name = ['$\\mathrm{MI}$', '$\\mathrm{Fomo}$', '$\\mathrm{Avg}$']\n # title = r\"$\\mathrm{FMNIST}$数据集, $\\mathrm{15}$个客户机\"\n # save_name = \"fmnist15_w.jpg\"\n\n # file = '../exp/cifar15_w'\n # order = [1, 0, 2]\n # name = ['$\\mathrm{MI}$', '$\\mathrm{Fomo}$', '$\\mathrm{Avg}$']\n # title = r\"$\\mathrm{CIFAR}$-10数据集, $\\mathrm{15}$个客户机\"\n # save_name = \"cifar15_w.jpg\"\n # plot_MI(file, order, name, title, save_name, rounds,'平均测试准确率')\n\n # order = [1, 0]\n # name = ['$\\mathrm{Avg}$', '$\\mathrm{Loss}$']\n\n # CIFAR\n # rounds = 100\n # file = '../exp/avg_w/cifar50_avg_w'\n # title = r\"$\\mathrm{CIFAR}$-10数据集, $\\mathrm{50}$个客户机\"\n # save_name = \"cifar50_avg_w.jpg\"\n\n # rounds = 50\n # file = '../exp/avg_w/cifar15_avg_w'\n # title = r\"$\\mathrm{CIFAR}$-10数据集, $\\mathrm{15}$个客户机\"\n # save_name = \"cifar15_avg_w.jpg\"\n # FMNIST\n\n # rounds = 50\n # file = '../exp/avg_w/fmnist15_avg_w'\n # title = r\"$\\mathrm{FMNIST}$数据集, $\\mathrm{15}$个客户机\"\n # save_name = \"fmnist15_avg_w.jpg\"\n #\n # rounds = 100\n # file = '../exp/avg_w/fmnist50_avg_w'\n # title = r\"$\\mathrm{FMNIST}$数据集, $\\mathrm{50}$个客户机\"\n # save_name = \"fmnist50_avg_w.jpg\"\n # plot(file, order, name, title, save_name, rounds, '平均测试准确率')\n","repo_name":"zhengLabs/pFedMI","sub_path":"plot/plot2.py","file_name":"plot2.py","file_ext":"py","file_size_in_byte":5416,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"4160124205","text":"import random\nimport tkinter as tk\n\n\nm = tk.Tk()\n''' \nwidgets are added here \n'''\n\n\ndef roll(y):\n result = random.randint(1, y)\n print(result)\n # return result\n\n\nprint('enter the maximum number')\n\n# print(roll(int(input())))\n\n\nm.title('Dice roller')\nbutton = tk.Button(m, text='Roll', width=25, command=roll(6))\nbutton.pack()\n\nm.mainloop()\n\n\n\n","repo_name":"DuncanSage/dice","sub_path":"dice.py","file_name":"dice.py","file_ext":"py","file_size_in_byte":352,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"3954503384","text":"import zmq\nimport time\nfrom ppp import *\nfrom queue import WRouteQueue\nfrom queue import WRouteItem\nfrom threading import Thread\n\nimport sys\nsys.path.append('../../lib')\nfrom default import WROUTE_ADDR, WROUTE_BACKEND_ADDR\nfrom log import log_err\n\nclass WRoutePoller(Thread):\n def _init_sock(self):\n self._context = zmq.Context(1)\n self._frontend = self._context.socket(zmq.ROUTER)\n self._backend = self._context.socket(zmq.ROUTER)\n self._frontend.bind(WROUTE_ADDR)\n self._backend.bind(WROUTE_BACKEND_ADDR)\n self._poller_backend = zmq.Poller()\n self._poller_backend.register(self._backend, zmq.POLLIN)\n self._poller = zmq.Poller()\n self._poller.register(self._frontend, zmq.POLLIN)\n self._poller.register(self._backend, zmq.POLLIN)\n \n def __init__(self):\n Thread.__init__(self)\n self._queue = WRouteQueue()\n self._init_sock()\n self._active = True\n \n @property\n def name(self):\n return self.__class__.__name__\n \n def run(self):\n timeout = time.time() + PPP_HEARTBEAT_INTERVAL\n while self._active:\n if len(self._queue.queue) > 0:\n poller = self._poller\n else:\n poller = self._poller_backend\n socks = dict(poller.poll(PPP_HEARTBEAT_INTERVAL * 1000))\n if socks.get(self._backend) == zmq.POLLIN:\n frames = self._backend.recv_multipart()\n if not frames:\n break\n identity = frames[0]\n self._queue.add(WRouteItem(identity))\n msg = frames[1:]\n if len(msg) == 1:\n if msg[0] not in (PPP_READY, PPP_HEARTBEAT):\n log_err(self, \"invalid message, %s\" % msg)\n else:\n self._frontend.send_multipart(msg)\n \n if time.time() >= timeout:\n for identity in self._queue.queue:\n msg = [identity, PPP_HEARTBEAT]\n self._backend.send_multipart(msg)\n timeout = time.time() + PPP_HEARTBEAT_INTERVAL\n \n if socks.get(self._frontend) == zmq.POLLIN:\n frames = self._frontend.recv_multipart()\n if not frames:\n break\n frames.insert(0, self._queue.pop())\n self._backend.send_multipart(frames)\n \n self._queue.purge()\n \n def stop(self):\n self._active = False\n ","repo_name":"sitian/wing","sub_path":"services/wroute/poller.py","file_name":"poller.py","file_ext":"py","file_size_in_byte":2595,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"24675497902","text":"\"\"\"Optimal expert that additionally uses substitution actions.\"\"\"\nfrom typing import Any, Iterable, List, Sequence\n\nimport numpy as np\n\nfrom trans import actions\nfrom trans import optimal_expert\nfrom trans.actions import Copy, Del, Edit, EndOfSequence, Ins, Sub\n\n\nclass EditDistanceAligner(actions.Aligner):\n def __init__(self, del_cost=1.0, ins_cost=1.0, sub_cost=1.0):\n self.del_cost = del_cost\n self.ins_cost = ins_cost\n self.sub_cost = sub_cost\n\n def action_sequence_cost(\n self, x: Sequence[Any], y: Sequence[Any], x_offset: int, y_offset: int\n ) -> float:\n ed = optimal_expert.edit_distance(\n x,\n y,\n del_cost=self.del_cost,\n ins_cost=self.ins_cost,\n sub_cost=self.sub_cost,\n x_offset=x_offset,\n y_offset=y_offset,\n )\n return ed[-1, -1]\n\n def action_cost(self, action: Edit):\n if isinstance(action, Copy) or isinstance(action, EndOfSequence):\n return 0\n if isinstance(action, Del):\n return self.del_cost\n if isinstance(action, Ins):\n return self.ins_cost\n if isinstance(action, Sub):\n return self.sub_cost\n raise ValueError(f\"Unexpected action: {action}!\")\n\n\nclass NoSubstitutionAligner(EditDistanceAligner):\n def __init__(self):\n super().__init__(del_cost=1.0, ins_cost=1.0, sub_cost=1.0)\n\n def action_cost(self, action: Edit):\n if isinstance(action, Sub):\n return np.inf\n return super().action_cost(action)\n\n\nclass OptimalSubstitutionExpert(optimal_expert.OptimalExpert):\n def __init__(\n self, aligner: actions.Aligner, maximum_output_length: int = 150\n ):\n super().__init__(maximum_output_length)\n self.aligner = aligner\n\n def find_valid_actions(\n self,\n x: Sequence[Any],\n i: int,\n y: Sequence[Any],\n prefixes: Iterable[optimal_expert.Prefix],\n ):\n if len(y) >= self.maximum_output_length:\n return {EndOfSequence()}\n input_not_empty = i < len(x)\n attention = x[i] if input_not_empty else None\n actions_prefixes: List[optimal_expert.ActionsPrefix] = []\n for prefix in prefixes:\n prefix_insert = prefix.leftmost_of_suffix\n if prefix_insert is None:\n valid_actions = {EndOfSequence()}\n else:\n valid_actions = {Ins(prefix_insert)}\n if input_not_empty:\n if prefix_insert is not None:\n if prefix_insert == attention:\n valid_actions.add(Copy(attention, prefix_insert))\n else:\n valid_actions.add(\n Sub(old=attention, new=prefix_insert)\n )\n valid_actions.add(Del(attention))\n actions_prefix = optimal_expert.ActionsPrefix(\n valid_actions, prefix\n )\n actions_prefixes.append(actions_prefix)\n return actions_prefixes\n\n def roll_out(\n self,\n x: Sequence[Any],\n t: Sequence[Any],\n i: int,\n actions_prefixes: Iterable[optimal_expert.ActionsPrefix],\n ):\n costs_to_go = dict()\n for actions_prefix in actions_prefixes:\n suffix_begin = actions_prefix.prefix.j\n for action in actions_prefix.actions:\n if isinstance(action, Del):\n x_offset = i + 1\n t_offset = suffix_begin\n elif isinstance(action, Ins):\n x_offset = i\n t_offset = suffix_begin + 1\n elif isinstance(action, Sub):\n x_offset = i + 1\n t_offset = suffix_begin + 1\n elif isinstance(action, EndOfSequence):\n x_offset = i\n t_offset = suffix_begin\n else:\n raise ValueError(f\"Unknown action: {action}\")\n sequence_cost = self.aligner.action_sequence_cost(\n x, t, x_offset, t_offset\n )\n action_cost = self.aligner.action_cost(action)\n cost = action_cost + sequence_cost\n if action not in costs_to_go or costs_to_go[action] > cost:\n costs_to_go[action] = cost\n return costs_to_go\n","repo_name":"sigmorphon/2021-task1","sub_path":"baseline/trans/optimal_expert_substitutions.py","file_name":"optimal_expert_substitutions.py","file_ext":"py","file_size_in_byte":4428,"program_lang":"python","lang":"en","doc_type":"code","stars":20,"dataset":"github-code","pt":"60"}
+{"seq_id":"36306909958","text":"#Import libraries\r\nfrom itertools import permutations\r\nfrom matplotlib import style \r\nimport math\r\nimport matplotlib.pyplot as plt \r\n\r\n####################################################################################\r\n#Declare empty arrays to store x,y coordinates\r\nleast_distance = math.inf\r\nbestx = []\r\nbesty = []\r\nbest_path = []\r\nx = []\r\ny = []\r\n\r\n####################################################################################\r\n#function definitions\r\n\r\n#read and parse data file\r\ndef read_datafile(path):\r\n #Import data file\r\n i = 0\r\n x = []\r\n y = []\r\n with open((path), \"r\") as file:\r\n for line in file:\r\n split_line = line.strip().split(\" \")\r\n \r\n #Track line number to remove header info\r\n if i > 6:\r\n #Populate x,y coordinate pairs into arrays\r\n x.append(float(split_line[1]))\r\n y.append(float(split_line[2]))\r\n #increment line counter\r\n i += 1 \r\n return x, y\r\n\r\n#######################################################\r\n#graph sets of xy coordinates\r\ndef graph_coords(x, y, min_dist):\r\n #Define graph style\r\n style.use('dark_background')\r\n \r\n # plotting the points\r\n plt.plot(x, y,'ro-')\r\n for i in range(len(x) - 1):\r\n plt.annotate(i + 1, (x[i], y[i]), textcoords=\"offset points\", xytext=(0,5), ha = 'center')\r\n \r\n # naming the axes \r\n plt.xlabel('x - axis') \r\n plt.ylabel('y - axis') \r\n \r\n # giving a title to my graph \r\n plt.title((\"Optimum Path Length: \" + str(min_dist)))\r\n \r\n # function to show the plot \r\n plt.pause(.05)\r\n plt.show() \r\n\r\n return\r\n\r\n#######################################################\r\n#Convert the indices to actual point numbers as in the tsp file\r\ndef index2point(path):\r\n out = []\r\n for element in path:\r\n out.append(element+1) \r\n return out\r\n\r\n#######################################################\r\n#Calculate distance for the trip\r\ndef calculate_trip_dist(tripx, tripy):\r\n dist = 0\r\n for i in range(len(tripx) - 1):\r\n dist = dist + (math.hypot(tripx[i] - tripx[(i+1)], tripy[i] - tripy[i+1]))\r\n return dist\r\n\r\n####################################################################################\r\n\r\n############\r\n#INPUT \r\n###########\r\n\r\n#data file path\r\nfile_path = str(r'C:\\Users\\burkh\\OneDrive\\Desktop\\AI\\Project1\\datasets\\Random4.tsp')\r\n#used to read and parse the tsp file\r\nx, y = read_datafile(file_path)\r\n\r\n\r\n############\r\n#PROCESSING\r\n###########\r\n\r\n#Create all possible paths\r\nperm = permutations(range(len(x)))\r\n\r\n#iterate through permutations\r\nfor p in perm: \r\n tripx = []\r\n tripy = []\r\n \r\n #record trip using permutations of the range of 0 -> length(x) as pointers\r\n for element in p:\r\n tripx.append(x[element])\r\n tripy.append(y[element])\r\n \r\n #take trip back to starting node \r\n tripx.append(x[p[0]])\r\n tripy.append(y[p[0]])\r\n \r\n #Calculate distance for the trip\r\n temp_dist = 0\r\n temp_dist = calculate_trip_dist(tripx, tripy)\r\n \r\n #store lowest distance and trip\r\n if(temp_dist < least_distance):\r\n least_distance = temp_dist\r\n #Store current optimums\r\n bestx = tripx\r\n besty = tripy\r\n best_path = p\r\n \r\n############\r\n#OUTPUT\r\n###########\r\n \r\n#Create graph of the optimum path \r\ngraph_coords(bestx, besty, least_distance)\r\n\r\n#Send output for report\r\nprint(least_distance)\r\nprint(index2point(best_path)) #Write best path using point numbers","repo_name":"BurkhardtMicah/Artificial-Intelligence","sub_path":"BruteForce_TSP.py","file_name":"BruteForce_TSP.py","file_ext":"py","file_size_in_byte":3538,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"39539041797","text":"#build path dictionary from pathout.txt\n\n'''\nfile = open('./output_data/pathout.txt','rU')\nlines = file.readlines()\nfile.close()\n\npathDict = {}\nfor l in range(1,len(lines)):\n x = lines[l].split()\n pathID = x[0]\n path = x[1]\n pathNodes = x[1].split('|')\n pathTarget = pathNodes[len(pathNodes)-1]\n pathDict[pathID]=pathTarget\n'''\n\nfile = open('./gams_intermediates/results.txt','rU')\nlines = file.readlines()\nfile.close()\n\ntargetSet = set()\nfor l in lines:\n #pathID = l.split()[0].replace('\"','')\n #targetSet.add(pathDict[pathID])\n targetSet.add(l.split()[0].replace('\"',''))\n\noutfile = open('./gams_intermediates/truetargets.gms','w')\n\noutfile.write('Set trueTargets(node)\\t\"targets\"\\n')\noutfile.write('/'+'\\t')\nfor t in targetSet:\n outfile.write(t+'\\n')\noutfile.write('/;\\n')\noutfile.close()\nprint(len(targetSet))\n\nfile = open('./input_data/targetntype2.tab','w')\nfile.write('#node\\tntype=discrete(source|target)\\n')\nfile.write('7025\\tsource\\n')\nfor t in targetSet:\n file.write(t+'\\ttarget\\n')\nfile.close()\n\n","repo_name":"Craven-Biostat-Lab/subnetwork_inference","sub_path":"NR2F1/scripts/target_ntype2_builder.py","file_name":"target_ntype2_builder.py","file_ext":"py","file_size_in_byte":1041,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"1158883225","text":"import question_bank\r\nimport time\r\n\r\nq_list=list(question_bank.all_questions)\r\na_list=list(question_bank.all_questions.values())\r\nrounds = 0\r\nscore = 0\r\n\r\ndef quiz():\r\n print(\"Welcome to the Videogame Trivia Quiz\")\r\n for questions in q_list:\r\n question() \r\n print(\"Thank you for playing! Your final score is:\", score)\r\n\r\ndef question():\r\n global a_list\r\n global q_list\r\n global rounds\r\n global score\r\n while rounds < 10:\r\n print(q_list[0])\r\n print(a_list[0])\r\n user_answer = input(\"Answer: \").upper()\r\n if user_answer not in [\"A\",\"B\",\"C\",\"D\"]:\r\n print(\"You have entered an invalid option\")\r\n continue\r\n elif user_answer == question_bank.answers[0]:\r\n print(\"Correct!\")\r\n a_list.pop(0)\r\n question_bank.answers.pop(0)\r\n q_list.pop(0)\r\n rounds += 1\r\n score += 1\r\n else:\r\n print(\"Sorry, the correct answer was\", question_bank.answers[0])\r\n a_list.pop(0)\r\n question_bank.answers.pop(0)\r\n q_list.pop(0)\r\n rounds +=1\r\n time.sleep(1)\r\n\r\nif __name__ == \"__main__\":\r\n quiz()","repo_name":"baconeggers/PersonalProjects","sub_path":"Trivia Game.py","file_name":"Trivia Game.py","file_ext":"py","file_size_in_byte":1190,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"40344094232","text":"# -*- coding: utf-8 -*-\n# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-\n# vi: set ft=python sts=4 ts=4 sw=4 et:\n\nfrom pathlib import Path\n\nimport pandas as pd\nimport pytest\n\nfrom halfpipe.design import group_design, prepare_data_frame\n\ntsv_str = \"\"\"SubjID\\tSex\\tSite.\\tRecur\\tAD\n11111111\\t0\\t1\\t0\\tNA\n11111112\\t1\\t1\\t1\\tNA\n11111113\\t0\\t1\\t2 \\tNA\n11111114\\t1\\t1\\tNA \\tNA\n11111115\\t0\\t1\\t1\\tNA\n\"\"\"\n\n\n@pytest.mark.parametrize(\n \"tsv_str\",\n [tsv_str, tsv_str.replace(\"\\t\", \",\"), tsv_str.replace(\"\\t\", \" \")],\n)\ndef test_prepare_data_frame(tmp_path: Path, tsv_str: str):\n spreadsheet_path = tmp_path / \"spreadsheet.tsv\"\n\n with open(spreadsheet_path, \"w\") as file_handle:\n file_handle.write(tsv_str)\n\n variables: list[dict] = [\n {\"name\": \"SubjID\", \"type\": \"id\"},\n {\"name\": \"Sex\", \"type\": \"categorical\", \"levels\": [\"0\", \"1\"]},\n {\"name\": \"Site.\", \"type\": \"categorical\", \"levels\": [\"1\"]},\n {\"name\": \"Recur\", \"type\": \"categorical\", \"levels\": [\"0\", \"1\", \"2\"]},\n {\"name\": \"AD\", \"type\": \"categorical\", \"levels\": [\"0\", \"1\", \"2\"]},\n ]\n subjects = [\"11111111\", \"11111112\", \"11111113\", \"11111114\", \"11111115\"]\n\n df = prepare_data_frame(spreadsheet_path, variables, subjects)\n assert df.shape == (5, 4)\n assert df[\"Recur\"].dtype == \"category\"\n assert df[\"Recur\"].notnull().sum() == 4\n assert df[\"AD\"].dtype == \"category\"\n\n\ndef test_group_design():\n data_frame = pd.DataFrame(\n {\n \"subject\": [\"s1\", \"s2\", \"s3\", \"s4\", \"s5\", \"s6\"],\n \"group\": [\"A\", \"A\", \"B\", \"B\", \"C\", \"C\"],\n }\n )\n data_frame = data_frame.set_index(\"subject\")\n data_frame[\"group\"] = data_frame[\"group\"].astype(\"category\")\n\n subjects = [\"s1\", \"s2\", \"s3\", \"s4\"]\n contrasts = [\n dict(type=\"infer\", variable=[\"group\"], levels=[\"A\", \"B\", \"C\"]),\n ]\n\n design = group_design(\n data_frame,\n contrasts,\n subjects,\n )\n assert len(design.regressor_list) == 2\n","repo_name":"HALFpipe/HALFpipe","sub_path":"tests/test_design.py","file_name":"test_design.py","file_ext":"py","file_size_in_byte":1989,"program_lang":"python","lang":"en","doc_type":"code","stars":55,"dataset":"github-code","pt":"60"}
+{"seq_id":"14824526990","text":"from django.urls import path\n\nfrom accounts import views\n\nurlpatterns = [\n path('login', views.login_view, name='login'),\n path('logout', views.logout_view, name='logout'),\n path('profile', views.profile_view, name='profile'),\n path('division:', views.division_view, name='division'),\n path('profile:', views.profile_view, name='profile'),\n path('add_profile', views.add_profile, name='add_profile'),\n path('add_division', views.add_division, name='add_division'),\n path('profiles', views.profiles_view, name='profiles'),\n]\n","repo_name":"Qvineox/accessControl_app","sub_path":"accounts/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":581,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"2554164615","text":"#import library\r\nfrom tkinter import *\r\nimport random\r\n\r\n\r\n#initialize window\r\nroot = Tk()\r\nroot.geometry('500x500')\r\nroot.resizable(0,0)\r\nroot.title('Rock,Paper,Scissors')\r\nroot.config(bg ='lightskyblue')\r\n\r\n#heading\r\nLabel(root, text = 'Rock, Paper ,Scissors' , font='arial 30 bold', bg = 'white').pack()\r\n\r\n##user choice\r\nuser_take = StringVar()\r\nLabel(root, text = 'choose any one: rock, paper ,scissors' , font='arial 19 bold', bg = 'white').place(x = 20,y=70)\r\nEntry(root, font = 'arial 15', textvariable = user_take , bg = 'antiquewhite2').place(x=140 , y = 130)\r\n\r\n\r\n#computer choice\r\nc_pick = random.randint(1,3)\r\nif c_pick == 1:\r\n c_pick = 'rock'\r\nelif c_pick ==2:\r\n c_pick = 'paper'\r\nelse:\r\n c_pick = 'scissors'\r\n \r\n\r\n#play\r\nResult = StringVar()\r\n\r\ndef play():\r\n user_pick = user_take.get()\r\n if user_pick == c_pick:\r\n Result.set('Tie,both are same')\r\n elif user_pick == 'rock' and c_pick == 'paper':\r\n Result.set('You loose,computer selected paper')\r\n elif user_pick == 'rock' and c_pick == 'scissors':\r\n Result.set('You win,computer selected scissors')\r\n elif user_pick == 'paper' and c_pick == 'scissors':\r\n Result.set('You loose,computer selected scissors')\r\n elif user_pick == 'paper' and c_pick == 'rock':\r\n Result.set('You win,computer selected rock')\r\n elif user_pick == 'scissors' and c_pick == 'rock':\r\n Result.set('You loose,computer selected rock')\r\n elif user_pick == 'scissors' and c_pick == 'paper':\r\n Result.set('You win ,computer selected paper')\r\n else:\r\n Result.set('invalid, Try again')\r\n \r\n \r\n#reset\r\ndef Reset():\r\n Result.set(\"\") \r\n user_take.set(\"\")\r\n\r\n##exit\r\ndef Exit():\r\n root.destroy()\r\n\r\n\r\n#button\r\nEntry(root, font = 'arial 10 bold', textvariable = Result, bg ='antiquewhite2',width = 64,).place(x=25, y = 250)\r\n\r\nButton(root, font = 'arial 13 bold', text = 'PLAY' ,padx =6,bg ='seashell4' ,command = play).place(x=213,y=190)\r\n\r\nButton(root, font = 'arial 13 bold', text = 'RESET' ,padx =6,bg ='seashell4' ,command = Reset).place(x=110,y=310)\r\n\r\nButton(root, font = 'arial 13 bold', text = 'EXIT' ,padx =6,bg ='seashell4' ,command = Exit).place(x=310,y=310)\r\n\r\nroot.mainloop()\r\n","repo_name":"taimoorfahim/Rock-Paper-Scissors_With-UI","sub_path":"rock-paper-scissors.py","file_name":"rock-paper-scissors.py","file_ext":"py","file_size_in_byte":2236,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"8285185336","text":"from Scene.scene import *\nimport pygame\nimport ctypes\nfrom Data.rec_data import *\n\nclass GameScene(Scene):\n\n def __init__(self, display, id, w_size, connector, resources):\n\n print(\"Game Scene\")\n \n super().__init__(display, id, w_size, connector, resources)\n \n # Constants\n self.GET_DATA_TIME = 1.25\n self.FONT_PATH = r\"..\\Resource\\OpenSans-Regular.ttf\"\n\n self._is_in_intro = True\n self._intro_font = pygame.font.Font(self.FONT_PATH, 40)\n self._info_text = self._intro_font.render(\"You think you can beat my bot? Let's try\", 0, (0, 0, 0))\n\n self._trans_to_next_scene = False\n\n self._data_trainer = DataTrainer()\n self.NUM_OF_INPUT_SAMPLE = DataTrainer.NUM_OF_INPUT_SAMPLE\n self._input_data = []\n\n self._connector = connector\n\n self._ptr_accel = self._connector.get_accel()\n self._ptr_emg_data = self._connector.get_emg_data()\n\n self._clock = pygame.time.Clock()\n self._timer = 0\n self._movement_accel = [(0, 0, 0), (0, 0, 0)]\n self._has_wave = False\n\n self._turn_number = 0\n\n for i in range(len(self._resources)):\n self._resources[i] = pygame.transform.scale(self._resources[i], (125, 125))\n\n self._render_imgs = []\n offset = 450, (self._w_size[1] - self._resources[3].get_rect().size[1] * 2 * 1.2) // 2\n for i in range(6): \n original_pos = (i % 3) * self._resources[3].get_rect().size[0] * 1.2, (i // 3) * self._resources[3].get_rect().size[1] * 1.2\n pos = tuple(map(lambda x, y: x + y, offset, original_pos))\n self._render_imgs.append([self._resources[3], pos])\n\n def update(self):\n\n super().update()\n\n if self._is_in_intro:\n self._timer += self._clock.tick() / 1000\n if self._timer >= 3.0:\n self._timer = 0\n self._is_in_intro = False\n \n elif self._trans_to_next_scene:\n self._timer += self._clock.tick() / 1000\n if self._timer >= 3.0:\n self._isEndScene = True\n\n else:\n if not self._has_wave:\n if self._timer <= 0.25:\n if self._timer == 0:\n temp = (ctypes.c_float * 3).from_address(self._ptr_accel)\n self._movement_accel[0] = [temp[i] for i in range(3)]\n self._timer += self._clock.tick() / 1000\n \n else:\n self._timer = 0\n temp = (ctypes.c_float * 3).from_address(self._ptr_accel)\n self._movement_accel[1] = [temp[i] for i in range(3)]\n ox = self._movement_accel[0][0]\n x = self._movement_accel[1][0]\n # Wave down\n if x - ox < -0.7 and ox != 0:\n self._has_wave = True\n \n else:\n if self._timer <= self.GET_DATA_TIME:\n self._timer += self._clock.tick() / 1000\n temp = (ctypes.c_int * DataTrainer.NUM_OF_SENSORS).from_address(self._ptr_emg_data)\n self._input_data.append([temp[i] for i in range(DataTrainer.NUM_OF_SENSORS)])\n else:\n # Finish delta_time second\n\n self._timer = 0\n self._has_wave = False\n\n print(len(self._input_data))\n\n predict_result = self._data_trainer.predict(self._input_data[:DataTrainer.NUM_OF_INPUT_SAMPLE])\n self._render_imgs[self._turn_number + 3][0] = self._resources[predict_result]\n self._render_imgs[self._turn_number][0] = self._resources[(predict_result + 1) % 3]\n \n if self._turn_number < 2:\n self._turn_number = self._turn_number + 1\n \n else:\n self._trans_to_next_scene = True\n\n self._input_data = []\n \n\n # Event handling\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n print(\"Exiting...\")\n self._connector.close()\n pygame.quit()\n exit()\n elif event.type == pygame.KEYDOWN:\n if event.key == pygame.K_SPACE:\n # Change Scene to Game Scene:\n self._isEndScene = True\n else: pass\n else: pass\n\n def render(self):\n super().render()\n\n if self._is_in_intro:\n self._display.blit(self._info_text, fix_pos([self._w_size[0] // 2, self._w_size[1] // 2], self._info_text.get_size()))\n\n else:\n self._display.blit(self._resources[4], (120, 300))\n self._display.blit(self._resources[5], (120, 140))\n\n for i in range(6):\n self._display.blit(self._render_imgs[i][0], self._render_imgs[i][1])\n\n \n ","repo_name":"ngthluu/RPS-Myogame","sub_path":"Scene/game_scene.py","file_name":"game_scene.py","file_ext":"py","file_size_in_byte":5040,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"71970817152","text":"alphabet = ('A','B','C','D','E','F','G','H','I','J','K','L','M','N','O','P','Q','R','S','T','U','V','W','X','Y','Z')\ndecryptedWord = []\n\nword = list(str(input('Write the ciphered word: ')).upper().strip())\nfactor = int(input('Write the factor: '))\n\nfor char in word:\n actualIndex = alphabet.index(char)\n newIndex = actualIndex - factor\n if newIndex > len(alphabet):\n print(\"I don't know how it happened, but 'newIndex' > 'alphabet'\")\n exit()\n elif newIndex < 0: newIndex = len(alphabet) - (newIndex * -1)\n decryptedWord.append(alphabet[newIndex])\n\nfor char in decryptedWord: print(char, end='')\n","repo_name":"Icaro-G-Silva/AlmostNothingUseful","sub_path":"CesarCipher/Python/CesarCipher.py","file_name":"CesarCipher.py","file_ext":"py","file_size_in_byte":624,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"10240207395","text":"from flask import Flask,render_template,redirect,request,flash,make_response\nimport os\nimport hashlib\nimport datetime\nfrom cloudant.client import Cloudant\nfrom cloudant import query,result,document\n\n\napp = Flask(__name__)\napp.secret_key = 'lionel'\n\nPORT = os.getenv('PORT', '5000');\n\nUSERNAME=\"\";\nPASSWORD=\"\";\nURL=\"\";\nclient = Cloudant(USERNAME, PASSWORD, url=\"\");\n\nUPLOAD_FOLDER=os.path.dirname(__file__)+\"/tmp/\";\napp.config['UPLOAD_FOLDER']=UPLOAD_FOLDER;\n\n@app.route('/')\ndef home():\n filelist=retrieve_files();\n return render_template(\"upload.html\",files=filelist);\n\n\n@app.route('/upload',methods=['POST'])\ndef upload():\n f = request.files['fileToUpload'];\n file_name = f.filename;\n f.save(os.path.join(app.config['UPLOAD_FOLDER'], file_name));\n contents = '';\n hash_content = hashlib.md5();\n with open(UPLOAD_FOLDER + file_name, 'rb') as content_file:\n contents = content_file.read();\n hash_content.update(contents);\n print(hash_content.hexdigest());\n\n client.connect();\n\n db = client['lionelfilestorage'];\n file_version_query=query.Query(db,selector={'file_name':file_name});\n version_no=-1;\n is_file_exists=False;\n with query.Query.custom_result(file_version_query) as query_result:\n for doc in query_result:\n version_no = doc[\"version_no\"];\n if doc['file_hashvalue'] == hash_content.hexdigest():\n is_file_exists=True;\n break;\n msg =''\n if is_file_exists==True:\n msg='File already exists,cant upload the same file.'\n else :\n version_no=1 if version_no==-1 else version_no+1;\n file_doc = {\n '_id':file_name+str(version_no),\n 'file_name': file_name,\n 'file_content': contents,\n 'version_no': version_no,\n 'file_hashvalue':hash_content.hexdigest(),\n 'last_modified':str(datetime.datetime.now()),\n 'type':'filedoc'\n }\n db.create_document(file_doc);\n msg='File uploaded sucessfully';\n client.disconnect();\n os.remove(UPLOAD_FOLDER + file_name);\n flash(msg);\n return redirect(\"/\");\n\n@app.route('/delete',methods=['GET','POST'])\ndef delete():\n if request.method=='POST':\n msg='';\n try :\n file_name = request.form['file_name'];\n version_no = request.form['version_no'];\n client.connect();\n db = client['lionelfilestorage'];\n my_doc = db[file_name + version_no];\n my_doc.delete();\n msg='File deleted successfully';\n except KeyError:\n msg='File not found for deletion';\n except Exception:\n msg='Some issue occured';\n finally:\n client.disconnect();\n flash(msg);\n filelist=retrieve_files();\n return render_template(\"delete.html\",files=filelist);\n\n@app.route('/download',methods=['GET','POST'])\ndef download():\n if request.method=='POST':\n try:\n file_name = request.form['file_name'];\n version_no = request.form['version_no'];\n client.connect();\n db = client['lionelfilestorage'];\n my_doc = db[file_name + version_no];\n response = make_response(my_doc['file_content']);\n response.headers[\"Content-Disposition\"] = \"attachment; filename=\" +file_name;\n msg = 'File downloaded successfully';\n flash(msg);\n return response;\n except KeyError:\n msg='File not found for download';\n except Exception:\n msg='Some issue occured';\n finally:\n client.disconnect();\n flash(msg);\n filelist = retrieve_files();\n return render_template(\"download.html\",files=filelist);\n\ndef retrieve_files():\n list=[];\n client.connect();\n db = client['lionelfilestorage'];\n file_version_query = query.Query(db, selector={'type':'filedoc'});\n with query.Query.custom_result(file_version_query) as query_result:\n for doc in query_result:\n list.append(doc);\n client.disconnect();\n return list;\n\nif __name__ == '__main__':\n app.run(host='0.0.0.0', port=int(PORT))\n #app.run()\n","repo_name":"lionel-ronald-sequeira/IBM-Bluemix2","sub_path":"welcome.py","file_name":"welcome.py","file_ext":"py","file_size_in_byte":4224,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"19844268668","text":"import matplotlib.pyplot as plt\nimport torch\nimport numpy as np\nimport scipy.io as sio\nfrom gan import Generator256\n\n\n\ndef load_data(DATA_PATH, device):\n data = sio.loadmat(DATA_PATH)\n \n S = torch.from_numpy(data['S']).type(torch.float32)\n T = torch.from_numpy(data['T']).type(torch.float32)\n C = torch.from_numpy(data['C']).type(torch.float32)\n \n S_true = torch.from_numpy(data['S_true']).type(torch.float32)\n C_true = torch.from_numpy(data['C_true']).type(torch.float32)\n T_true = torch.from_numpy(data['T_true']).type(torch.float32)\n \n # Permutation for compatibility with matlab generated arrays\n T = T.permute(2,0,1).to(device)\n T_true = T_true.permute(2,0,1).to(device)\n try:\n S = S.permute(2,0,1).to(device)\n except:\n S = S.unsqueeze(dim=0)\n \n try:\n S_true = S_true.permute(2,0,1).to(device)\n except:\n S_true = S_true.unsqueeze(dim=0)\n \n \n C = C.permute(1,0).to(device)\n C_true = C_true.permute(1,0).to(device)\n \n return S, C, T, S_true, C_true, T_true\n\n\n \n \ndef load_generator(GAN_PATH, device):\n generator = Generator256()\n \n checkpoint = torch.load(GAN_PATH, map_location=torch.device('cpu'))\n \n generator.load_state_dict(checkpoint['g_model_state_dict'])\n generator.eval()\n return generator.to(device)\n\ndef load_all( datapath, z_dimension, device, visualize_data = True, k=25):\n\n S, C, T, S_true, C_true, T_true = load_data(datapath, device)\n\n # C_true[0:,50:] = 0\n # T_true = get_tensor(S_true.unsqueeze(dim=1), C_true, device)\n\n if visualize_data:\n print(S.shape[0], \"emitters.\")\n fig, ax = plt.subplots(1, C.shape[0]+1, figsize=(5*C.shape[0], 3))\n \n for i in range(C.shape[0]):\n ax[i].plot(C_true[i,:].cpu().detach().numpy())\n \n ax[i+1].imshow(torch.log(T_true[k,...]).cpu() )\n\n # parameters\n R, I, J = S.shape # R is the number of emitters, I,J is the size of the SLF\n K = C.shape[-1] # K is the number of frequency bands\n\n # Initialize the latent vectors for each emitter\n Z_init = torch.randn((R, z_dimension), dtype=torch.float32).to(device) \n Z = torch.zeros((R, z_dimension), dtype=torch.float32).to(device) \n\n # zero start \n S_init = torch.zeros( (R, 1, I, J) ).to(device) \n C_init = torch.zeros(C.shape).to(device)\n\n T = T.unsqueeze(dim=1)\n\n return S_init, C_init, T, S_true, C_true, T_true, Z_init, Z","repo_name":"XiaoFuLab/Quantized-Radio-Map-Estimation-BTD-and-DGM","sub_path":"qmc_utils.py","file_name":"qmc_utils.py","file_ext":"py","file_size_in_byte":2456,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"60"}
+{"seq_id":"41081490560","text":"import numpy as np\nimport pytesseract as pt\nimport cv2\nimport win32gui\nimport mss\nimport mss.tools\nfrom CNN import CNN\nfrom PIL import Image\nfrom PIL import ImageGrab\nfrom pynput.mouse import Button, Controller\nfrom grabscreen import grab_screen\nfrom directkeys import *\nimport time\nfrom grabImageUsingCV2 import fetchScoreFromImage\n\nmouse=Controller()\nclass Dave(object):\n cnn_graph = CNN()\n def __init__(self):\n self.rward=0\n # self.reset()\n self.visit=0\n self.previous_screen=[]\n self.previous_compare_screen=[]\n # self.previous_gold_screen=[]\n self.previous_screen_score = 0\n self.curr_score = 0\n self.prev_unprocessed = 0\n self.flagTrophy = False\n\n def _get_reward(self, action):\n flag1=False\n flag2=False\n flag3=False\n flagTrophy = False\n screen = mss.mss().grab((0,0,1366,768))\n mss.tools.to_png(screen.rgb,screen.size,output='screen.jpg')\n self.visit+=1\n screen = np.asarray(screen)\n if self.visit > 1 :\n self.curr_score = 0\n reward_scrn = mss.mss().grab((310,0,470,47))\n mss.tools.to_png(reward_scrn.rgb,reward_scrn.size,output='reward.jpg')\n # reward_scrn = Image.open(\"reward.jpg\")\n reward_scrn = cv2.imread(\"reward.jpg\")\n self.curr_score = fetchScoreFromImage(reward_scrn) \n prev_score = self.previous_screen_score\n if self.curr_score != self.prev_unprocessed:\n self.prev_unprocessed = self.curr_score\n if self.curr_score < self.previous_screen_score:\n self.curr_score = self.previous_screen_score + 100\n elif(self.curr_score - self.previous_screen_score >= 1000):\n self.flagTrophy = True\n\n diff_in_score = self.curr_score - self.previous_screen_score\n if(diff_in_score >= 50):\n flag1 = True\n self.previous_screen_score = self.curr_score\n else:\n self.curr_score = self.previous_screen_score\n print(\"\\nPrevious Score was %d and Current score is %d \\n\" % (prev_score, self.curr_score))\n\n compare_screen=screen[50:,0:]\n if self.visit>1:\n diff2=cv2.subtract(compare_screen,self.previous_compare_screen)\n b2,g2,r2,a2=cv2.split(diff2)\n #print(cv2.countNonZero(b1) , cv2.countNonZero(g1) , cv2.countNonZero(r1))\n max2=max(cv2.countNonZero(b2) , cv2.countNonZero(g2) , cv2.countNonZero(r2))\n print(\"difference\",max2)\n if max2<1300:\n flag3=True \n self.previous_compare_screen=compare_screen \n \n # lives_screen= screen[55:75,70:150]\n # if self.visit>1:\n # diff1=cv2.subtract(lives_screen,self.previous_screen)\n # b1,g1,r1=cv2.split(diff1) \n # if cv2.countNonZero(b1) > 0 or cv2.countNonZero(g1) > 0 or cv2.countNonZero(r1) > 0:\n # flag2=True \n # self.previous_screen=lives_screen\n \n # flag 2 is for lives\n # if flag2:\n # return -1\n # flag 1 is true if any gem is collected\n if flagTrophy:\n return 1\n if flag1:\n return 0.2\n # flag 3 is true if there is no change in new frame.\n if flag3:\n return -0.1\n ingame_reward = 0.01\n return ingame_reward\n\n def _is_over(self, reward):\n if self.curr_score == 2200:\n is_over = True\n else:\n is_over = False\n return is_over\n \n \n def observe(self):\n screen = np.asarray(ImageGrab.grab(bbox=(0,50,1336,668))) \n state = self.cnn_graph.get_image_feature_map(screen)\n return state\n\n def act(self, action):\n keys_to_press = [[uparrow,leftarrow],[leftarrow],[uparrow,leftarrow],[uparrow,rightarrow],[rightarrow],[uparrow,rightarrow]]\n PressKey(keys_to_press[action][0])\n time.sleep(0.3)\n ReleaseKey(keys_to_press[action][0])\n i=1\n if(len(keys_to_press[action]) >=1):\n while i \\\n \\\n '\n\nif __name__ == \"__main__\":\n app.run()\n\n","repo_name":"Miti56/Smart-Alarm-Flask","sub_path":"Prueba Flask.py","file_name":"Prueba Flask.py","file_ext":"py","file_size_in_byte":358,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"39250063234","text":"import pytest\nimport saltext.credstash.sdbs.credstash as credstash_sdb\n\n\n@pytest.fixture\ndef configure_loader_modules():\n module_globals = {\n \"__salt__\": {\"this_does_not_exist.please_replace_it\": lambda: True},\n }\n return {\n credstash_sdb: module_globals,\n }\n\n\ndef test_replace_this_this_with_something_meaningful():\n assert \"this_does_not_exist.please_replace_it\" in credstash_sdb.__salt__\n assert credstash_sdb.__salt__[\"this_does_not_exist.please_replace_it\"]() is True\n","repo_name":"major0/salt-credstash-extension","sub_path":"tests/unit/sdbs/test_credstash.py","file_name":"test_credstash.py","file_ext":"py","file_size_in_byte":505,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"26435355250","text":"#!/usr/bin/env python3\nimport newspaper\nimport os\nimport sys\nimport datetime\nimport traceback\n\nos.chdir(sys.path[0]) \npath = os.getcwd()\n\ndef symlink():\n if not os.path.exists('/tmp/.newspaper_scraper'):\n origem = os.path.join(path,'newspaper_scraper')\n print(origem)\n #os.symlink(origem,'/tmp/.newspaper_scraper/')\n\ndef log(log : str, logsFile):\n now = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n logsFile.write(f'{log}[{now}]\\n')\n\ndef create_dirs():\n #create date dirs\n today = datetime.date.today()\n year = today.year\n month = today.month\n day = today.day\n\n outFile = f'{path}/articles/{year}/{month}'\n os.makedirs(outFile,exist_ok=True)\n return outFile + f'/{day}.html'\n\ndef main():\n symlink()\n\n url = 'https://www.jn.pt'\n #url = 'https://www.jornaldeangola.ao/ao/'\n #url = 'https://novojornal.co.ao'\n\n #create log file\n logsFile = open(f'{path}/logs.txt','w')\n\n log('Geting articles',logsFile)\n \n j = newspaper.build(url,memoized_articles=False)\n\n nArticles = j.size()\n log(f'{nArticles} articles',logsFile)\n\n outFile = create_dirs()\n output = open(outFile,'a')\n\n output.write('\\n')\n \n na = 0\n for article in j.articles:\n log(f'Article {na} of {nArticles}',logsFile)\n na+=1\n try:\n ar = newspaper.Article(article.url())\n ar.download()\n ar.parse()\n output.write(\n f'''\n \n \n {ar.title}\n \n \n {ar.url}\n \n \n {ar.autor}\n \n \n {ar.date}\n \n \n {ar.tags}\n \n \n {ar.text}\n \n \n '''\n )\n except Exception:\n print(traceback.format_exc)\n log(traceback.format_exc,logsFile)\n\n output.write('\\n')\n output.close()\n\n log('Finished writing',logsFile)\n\n logsFile.close()\n\nif __name__ == \"__main__\":\n main()","repo_name":"surumkata/spln-2223","sub_path":"TPC6/corpus.py","file_name":"corpus.py","file_ext":"py","file_size_in_byte":2447,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"14001908779","text":"import pygame, sys\nfrom pygame.locals import *\n\nimport brain\nfrom brain import *\n\nimport bird\nfrom bird import *\nimport pipe\nfrom pipe import *\n\n# Initializing\npygame.init()\n\n# Setting up FPS\nFPS = 60\nFramePerSec = pygame.time.Clock()\n\n# Creating colors\n# BLUE = (0, 0, 255)\nRED = (255, 0, 0)\n# GREEN = (0, 255, 0)\nBLACK = (0, 0, 0)\nWHITE = (255, 255, 255)\n\n# Other Variables for use in the program\nSCREEN_WIDTH = 400\nSCREEN_HEIGHT = 600\n\n# Create a black screen\nDISPLAYSURF = pygame.display.set_mode((SCREEN_WIDTH, SCREEN_HEIGHT))\nDISPLAYSURF.fill(BLACK)\npygame.display.set_caption(\"Game\")\n\n# Create player\nplayer = bird.Bird(SCREEN_HEIGHT, DISPLAYSURF, WHITE)\n\n# Create pipe array\npipes = []\npipes.append(pipe.Pipe(SCREEN_HEIGHT, SCREEN_WIDTH, DISPLAYSURF, WHITE))\n\n# Pipe spawner\npipe_delay = 2000 # 2 seconds\nnew_pipe = pygame.USEREVENT + 1\npygame.time.set_timer(new_pipe, pipe_delay)\n\n# Create brain\nbrain = brain.Brain()\n\n# Sprite groups\n# all_sprites = pygame.sprit\n\n# Game Loop\nwhile True:\n # Refresh background\n DISPLAYSURF.fill(BLACK)\n\n # Cycles through all occurring events\n for event in pygame.event.get():\n if event.type == QUIT:\n pygame.quit()\n sys.exit()\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_SPACE:\n player.up()\n if event.type == new_pipe:\n pipes.append(pipe.Pipe(SCREEN_HEIGHT, SCREEN_WIDTH, DISPLAYSURF, WHITE))\n # all_sprites.add(pipes[-1])\n\n player.update()\n player.show()\n\n for p in pipes:\n p.update()\n p.show()\n\n # Check if pipe hits bird\n if p.top_rect.colliderect(player.bird_rect) or p.bottom_rect.colliderect(player.bird_rect):\n p.colour = RED\n else:\n p.colour = WHITE\n if player.y == SCREEN_HEIGHT:\n DISPLAYSURF.fill(RED)\n if p.offscreen():\n # pipes.pop(pipes.index(p))\n pipes = [pipes[1]]\n\n pygame.display.flip()\n FramePerSec.tick(FPS)\n","repo_name":"jjacobgreen/projects","sub_path":"flappy/flappy_playable.py","file_name":"flappy_playable.py","file_ext":"py","file_size_in_byte":2009,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"5849481958","text":"from rdflib.plugins.serializers.turtle import TurtleSerializer\nfrom rdflib.namespace import Namespace, FOAF, SKOS, RDF\nfrom rdflib import BNode\nimport logging\nimport re\n\nSD = Namespace('http://www.w3.org/ns/sparql-service-description#')\nISOTHES = Namespace('http://purl.org/iso25964/skos-thes#')\n\nlogger = logging.getLogger(__name__)\n\n\nclass OrderedTurtleSerializer(TurtleSerializer):\n\n short_name = \"ots\"\n\n def __init__(self, store):\n super(OrderedTurtleSerializer, self).__init__(store)\n\n # Class order:\n self.class_order = []\n\n # Sort key generators for specific classes :\n self.sorters_by_class = {}\n\n # Default sort key generators\n self.sorters = [\n ('^(.+)$', lambda x: str(x[0])),\n ]\n\n def getSorters(self, class_uri):\n return self.sorters_by_class.get(class_uri, self.sorters)\n\n def getSortKeyFunction(self, class_uri):\n sorters = self.getSorters(class_uri)\n\n # Order of instances:\n def sortKeyFn(x):\n # Check if the instances match any special pattern:\n for pattern, func in sorters:\n m1 = re.search(pattern, x)\n if m1:\n return func(m1.groups())\n logging.warning('%s did not match any sorters', x)\n\n return sortKeyFn\n\n def orderSubjects(self):\n seen = {}\n subjects = []\n\n # Find classes not included in self.class_order and sort them alphabetically\n other_classes = [x for x in set(self.store.objects(predicate=RDF.type)) if x not in self.class_order]\n other_classes = sorted(other_classes)\n\n # Loop over all classes\n for class_uri in self.class_order + other_classes:\n\n # Sort the members of each class\n members = sorted(self.store.subjects(RDF.type, class_uri),\n key=self.getSortKeyFunction(class_uri))\n\n for member in members:\n subjects.append(member)\n self._topLevels[member] = True\n seen[member] = True\n\n # Include anything not seen yet\n recursable = [\n (isinstance(subject, BNode),\n self._references[subject], subject)\n for subject in self._subjects\n if subject not in seen\n ]\n\n recursable.sort()\n subjects.extend([subject for (isbnode, refs, subject) in recursable])\n\n return subjects\n","repo_name":"scriptotek/otsrdflib","sub_path":"otsrdflib/ots.py","file_name":"ots.py","file_ext":"py","file_size_in_byte":2441,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"60"}
+{"seq_id":"43809174699","text":"#\n# Expert distribution for all datasets.\n#\nfrom airflow import DAG\nfrom airflow.utils.dates import days_ago\nfrom datetime import timedelta\nfrom ala import cluster_setup, ala_config, ala_helper\nfrom ala.ala_helper import step_bash_cmd, s3_cp, get_default_args\n\nDAG_ID = 'Expert_distribution'\n\nSPARK_STEPS = [\n s3_cp(\"a. Copy IndexRecord to S3\", f\"s3://{ala_config.S3_BUCKET_AVRO}/pipelines-all-datasets/index-record\", \"hdfs:///pipelines-all-datasets/index-record\"),\n s3_cp(\"b. Copy existing Outlier Distribution Cache\", f\"s3://{ala_config.S3_BUCKET_AVRO}/pipelines-outlier/\", \"hdfs:///pipelines-outlier/\", action_on_failure=\"CONTINUE\"),\n step_bash_cmd(\"c. Outlier detection\", f\" la-pipelines outlier all --cluster\"),\n step_bash_cmd(\"d. Delete AVRO output on S3\", f\" sudo -u hadoop aws s3 rm s3://{ala_config.S3_BUCKET_AVRO}/pipelines-outlier/all --recursive\"),\n step_bash_cmd(\"e. Delete temp SCP directories created by s3-dist-cp\", f\" sudo -u hadoop hdfs dfs -rm -f /pipelines-outlier/pipelines-outlier_$folder$\"),\n s3_cp(\"f. Copy Outliers results to S3\", \"hdfs:///pipelines-outlier\", f\"s3://{ala_config.S3_BUCKET_AVRO}/pipelines-outlier\")\n]\n\nwith DAG(\n dag_id=DAG_ID,\n description=\"Expert distribution\",\n default_args=get_default_args(),\n dagrun_timeout=timedelta(hours=4),\n start_date=days_ago(1),\n schedule_interval=None,\n tags=['emr', 'all-datasets']\n) as dag:\n cluster_setup.run_large_emr(dag, SPARK_STEPS, \"bootstrap-index-actions.sh\")\n","repo_name":"AtlasOfLivingAustralia/pipelines-airflow","sub_path":"dags/expert_distribution_dag.py","file_name":"expert_distribution_dag.py","file_ext":"py","file_size_in_byte":1484,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"18395881208","text":"#https://leetcode.com/problems/unique-paths/discuss/22953/Java-DP-solution-with-complexity-O(n*m)\nclass Solution:\n def uniquePaths(self, m: int, n: int) -> int:\n #unique paths using 2-D array\n dp = [[0] * (n) for _ in range(m)]\n for i in range(m):\n dp[i][0] = 1\n for j in range(n):\n dp[0][j] = 1\n for i in range(1, m):\n for j in range(1, n):\n dp[i][j] = dp[i][j-1] + dp[i-1][j]\n\n return dp[m-1][n-1]\n \n def printdp(self, dp):\n for i in (dp):\n print(i)\n \n\n\ns = Solution()\nres = s.uniquePaths(3,2)\nprint(res)\n ","repo_name":"ArunAaryan/LeetCodePlaylist","sub_path":"LeetCodeMedium/UniquePaths.py","file_name":"UniquePaths.py","file_ext":"py","file_size_in_byte":639,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"9689822798","text":"import sys\nimport csv\nimport json\nimport argparse\nimport pandas as pd\nimport numpy as np\n\nimport serial\nfrom os import path\nfrom io import StringIO\n\nfrom PyQt5.Qt import *\nfrom pyqtgraph import PlotWidget\nfrom PyQt5 import QtCore\nimport pyqtgraph as pq\n\nimport threading\nimport time\n\nDATA_COLUMNS_NAMES = [\"type\", \"mac\", \"rssi\", \"channel\", \"secondary channel\" ,\"sig_mode\", \"bandwith\", \"STBC\", \"Length(bytes)\"]\n\n\nclass csi_data_graphical_window(QWidget):\n def __init__(self):\n super().__init__()\n\n self.resize(1280, 720)\n self.plotWidget_ted = PlotWidget(self)\n self.plotWidget_ted.setGeometry(QtCore.QRect(0, 0, 1280, 720))\n\n self.plotWidget_ted.setYRange(-20, 100)\n self.plotWidget_ted.addLegend()\n\n \n self.curve_list = []\n\n self.timer = pq.QtCore.QTimer()\n self.timer.timeout.connect(self.update_data)\n self.timer.start(100)\n\n def update_data(self):\n return\n\ndef csi_data_read_parse(port: str, csv_writer):\n ser = serial.Serial(port=port, baudrate=921600,\n bytesize=8, parity='N', stopbits=1)\n if ser.isOpen():\n print(\"open success\")\n else:\n print(\"open failed\")\n return\n\n while True:\n strings = str(ser.readline())\n if not strings:\n break\n\n strings = strings.lstrip('b\\'').rstrip('\\\\r\\\\n\\'')\n index = strings.find('CSI_RIG')\n\n if index == -1:\n continue\n\n csv_reader = csv.reader(StringIO(strings))\n csi_data = next(csv_reader)\n\n #print(csi_data)\n #print(len(csi_data))\n\n if csi_data[1] != \"a6:6f:82:3d:89:99\":\n continue\n\n if len(csi_data) != len(DATA_COLUMNS_NAMES):\n print(\"element number is not equal\")\n continue\n\n print(csi_data)\n\n csv_writer.writerow(csi_data)\n\n\n ser.close()\n return\n\n\nclass SubThread (QThread):\n def __init__(self, serial_port, save_file_name):\n super().__init__()\n self.serial_port = serial_port\n\n save_file_fd = open(save_file_name, 'w')\n self.csv_writer = csv.writer(save_file_fd)\n self.csv_writer.writerow(DATA_COLUMNS_NAMES)\n\n def run(self):\n csi_data_read_parse(self.serial_port, self.csv_writer)\n\n def __del__(self):\n self.wait()\n\n\nif __name__ == '__main__':\n if sys.version_info < (3, 6):\n print(\" Python version should >= 3.6\")\n exit()\n\n parser = argparse.ArgumentParser(\n description=\"Read CSI data from serial port and display it graphically\")\n parser.add_argument('-p', '--port', dest='port', action='store', required=True,\n help=\"Serial port number of csv_recv device\")\n parser.add_argument('-s', '--store', dest='store_file', action='store', default='./csi_recv_rigor.csv',\n help=\"Save the data printed by the serial port to a file\")\n\n args = parser.parse_args()\n serial_port = args.port\n file_name = args.store_file\n\n app = QApplication(sys.argv)\n\n subthread = SubThread(serial_port, file_name)\n subthread.start()\n\n window = csi_data_graphical_window()\n window.show()\n\n sys.exit(app.exec())\n","repo_name":"orkunispir/wifi-csi-sniffer","sub_path":"tools/csi_data_read.py","file_name":"csi_data_read.py","file_ext":"py","file_size_in_byte":3193,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"71951843071","text":"import time\n\n# Relationship between speed, distance and time:\n# time = distance / speed\n\ndistance = input(\"What is the total amount of metres you would like to travel:\\n\")\nspeed = input(\"What is the estimated average speed? Give answer in metre/second:\\n\")\n\nt = float(distance) / float(speed)\nformatted_time = time.strftime(\"%H:%M:%S\", time.gmtime(t))\n\nprint(\n f\"It will take approximately {formatted_time} (hh,mm,ss) to arive at your destination.\"\n)\n","repo_name":"Anton-Ca/Courses","sub_path":"Python/Pyth_intro/l03/t5.py","file_name":"t5.py","file_ext":"py","file_size_in_byte":454,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"5046406506","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\n・尺取法\n・まあWAだけども。\n\"\"\"\n\nimport sys\nfrom collections import deque, Counter, defaultdict\nfrom math import sqrt, hypot, factorial, pi, sin, cos, radians\nfrom heapq import heappop, heappush, heapify, heappushpop\nfrom bisect import bisect_left, bisect_right\nfrom itertools import permutations, product, combinations, combinations_with_replacement\nfrom operator import itemgetter, mul\nfrom copy import deepcopy\nfrom functools import reduce, partial\n\ndef input(): return sys.stdin.readline().strip()\nsys.setrecursionlimit(10 ** 9)\nINF = float('inf')\nMOD = 10 ** 9 + 7\n\nS = input()\nN = len(S)\n\nl,r = 0,1\nmx = 0\n# 外ループで左を動かす\nwhile l < N:\n # 内ループは条件を満たす限り右を動かす\n while r < N and S[l] != S[r]:\n r += 1\n # 同じ文字までの間隔が最大の所を探す\n if r < N:\n mx = max(r - l - 1, mx)\n if l == r:\n # 左が右に追いついたら、右も左に合わせて+1する\n r += 1\n l += 1\n\nif mx == 0:\n if N % 2 == 0:\n print('Second')\n else:\n print('First')\nelse:\n if mx % 2 == 0:\n print('First')\n else:\n print('Second')\n","repo_name":"Coki628/kyopro_submissions","sub_path":"AtCoder/ABC048d.py","file_name":"ABC048d.py","file_ext":"py","file_size_in_byte":1202,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"41582360579","text":"__author__ = \"soopercool101\"\r\n__version__ = \"1.0.0\"\r\n\r\nfrom BrawlCrate.API import *\r\nfrom BrawlLib.SSBB.ResourceNodes import *\r\nfrom BrawlCrate.NodeWrappers import PluginWrapper\r\nfrom System.Windows.Forms import ToolStripMenuItem\r\nimport struct\r\n\r\n# Wrapper for UVs\r\nclass UVWrapper(PluginWrapper):\r\n # This function returns a new instance of the class.\r\n # Necessary in order to properly call necessary functions\r\n def GetInstance(self):\r\n return UVWrapper()\r\n\r\ndef shiftUV_handler(sender, event_args):\r\n # Gather the required inputs from the user\r\n scaleX = BrawlAPI.UserFloatInput(\"X Scale multiplier to apply to the UVs (Applied first)\", \"X-Scale\", 1.0)\r\n scaleY = BrawlAPI.UserFloatInput(\"Y Scale multiplier to apply to the UVs (Applied first)\", \"Y-Scale\", 1.0)\r\n transX = BrawlAPI.UserFloatInput(\"X Translation offset to apply to the UVs (Applied second)\", \"X-Translation\", 0.0)\r\n transY = BrawlAPI.UserFloatInput(\"Y Translation offset to apply to the UVs (Applied second)\", \"Y-Translation\", 0.0)\r\n shiftUV(scaleX, scaleY, transX, transY)\r\n\r\ndef shiftUV(scaleX, scaleY, transX, transY):\r\n i = 0\r\n # Apply the modifiers to each UV point\r\n for vec2 in BrawlAPI.SelectedNode.Points:\r\n vec2.X = vec2.X * scaleX\r\n vec2.X = vec2.X + transX\r\n vec2.Y = vec2.Y * scaleY\r\n vec2.Y = vec2.Y + transY\r\n # Due to some quirk, necessary to set this manually to ensure saving works\r\n BrawlAPI.SelectedNode.Points[i] = vec2\r\n i += 1\r\n # Due to some quirk, necessary to set this manually to ensure saving works\r\n BrawlAPI.SelectedNode.Points = BrawlAPI.SelectedNode.Points\r\n # Flag node as needing saving\r\n BrawlAPI.SelectedNode.SignalPropertyChange()\r\n\r\n# Create an instance of our wrapper class and add it to the API wrapper cache\r\nwrapper = UVWrapper()\r\nBrawlAPI.AddWrapper[MDL0UVNode](wrapper)\r\n# Add a context menu item to our new wrapper\r\nBrawlAPI.AddContextMenuItem(UVWrapper, ToolStripMenuItem(\"Shift UVs\", None, shiftUV_handler))\r\n","repo_name":"soopercool101/BrawlCrateSamplePlugins","sub_path":"Loaders/Utility/Shift UVs.py","file_name":"Shift UVs.py","file_ext":"py","file_size_in_byte":2031,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"34102075826","text":"import os, json\nfrom langchain.llms import Databricks, OpenAI\nfrom dotenv import load_dotenv\nfrom langchain.prompts import ChatPromptTemplate\n\nclass RoleBasedAdvisor:\n def __init__(self, language_model='openai', config_file_path=None):\n self.template_string = \"\"\"{role_name} \\\nRespond to the user question that is delimited in triple backticks \\\nwith thoughtful and concise instructions that the user can easily implement in their \\\nday to day life.\nuser_question: ```{user_question}```\n\"\"\"\n self.role_description = {}\n self.role_description['doctor'] = \"\"\"You are a doctor (primary care physician) with 25 years of experience practicing in California. \\\nYou emphasize the importance of a healthy lifestyle that includes nutritious food and vigorous exercise.\"\"\"\n self.role_description['father'] = \"\"\"You are the user's father and cares deeply about their well being. You emphasize the importance of \\\nworking hard and getting a good education.\"\"\"\n self.role_description['business_partner'] = \"\"\"You are the user's business partner. You share a mutual interest in the success of your \\\ncompany. You emphasize actions that will maximize the long term viability and profitability of the company and achieving its mission.\"\"\"\n self.role_description['career_coach'] = \"\"\"You are the user's manager at work. You see great potential in the user to progress in their \\\ncareer. You emphasize actions that maximize the user's chances for a promotion and continue their trajectory to become a senior executive.\"\"\"\n self.user_question = \"I want to live a life that maximizes happiness and creates a positive impact on the world. What \\\nare the top 5 things I should do in the next week towards these goals?\"\n\n self.language_model = language_model\n if config_file_path is not None:\n with open(config_file_path) as f:\n self.config = json.load(f)\n self.llm = self.get_llm(language_model)\n\n def get_llm(self, language_model='openai'):\n load_dotenv()\n if 'DATABRICKS_RUNTIME_VERSION' in os.environ and language_model == 'openai': # Running in Databricks\n if 'OPENAI_API_KEY' not in os.environ:\n os.environ['OPENAI_API_KEY'] = dbutils.secrets.get('vbalasu', 'openai-databricks')\n\n if language_model == 'openai':\n llm = OpenAI(temperature=0.0, max_tokens=500)\n return llm\n elif language_model == 'llamav2':\n llm = Databricks(cluster_driver_port=self.config['port'], cluster_id=self.config['cluster_id'],\n model_kwargs={'temperature':0.0, 'max_new_tokens':500})\n return llm\n else:\n print('Unknown language model')\n return False\n \n def answer_as_role(self, user_question, role, verbose=False):\n prompt_template = ChatPromptTemplate.from_template(self.template_string)\n prompt = prompt_template.format_prompt(role_name=role, user_question=user_question)\n question = prompt.messages[0].content\n if verbose:\n print('/*\\n', f'LANGUAGE MODEL: {self.language_model}\\n\\n', question, '*/\\n\\n')\n return self.llm(question)","repo_name":"vbalasu/databricks-goodies","sub_path":"role-based-advisor/advisor.py","file_name":"advisor.py","file_ext":"py","file_size_in_byte":3201,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"2070378967","text":"# tuple fron iterobj\nL = [1, 2, 3]\niterator = iter(L)\nt = tuple(iterator)\nprint(t)\n\n# genexps and listcomps\n\"\"\"\nGenerator expressions are surrounded by parentheses (“()”)and list comprehensions are surrounded by square brackets (“[]”)\n\"\"\"\nxs = [4, 6, 7, 9, 10]\nxss = (x * 2 for x in xs)\nprint(xss) # xss is an iterobj\n\nxss_list = [x*2 for x in xs]\nprint(xss_list) # xss_list is a list\n\n# Generator functions and yields\n\ndef agen():\n i = 10\n yield [i, i, i]\n\na, b, c = agen()\n","repo_name":"nonomino/bad_python","sub_path":"noob/functional.py","file_name":"functional.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"45928289609","text":"import time\nimport numpy as np\n\nfrom rich.console import Console\nfrom rich import print\n\nimport cro_dt.VectorTree as vt\nfrom cro_dt.sup_configs import get_config, load_dataset\nconsole = Console()\n\nif __name__ == \"__main__\":\n # X, y = load_dataset(\"breast_cancer\")\n # X = np.array([[-3, -2], [-4, -2], [3, 3]])\n # y = np.array([1, 1, 0])\n W = np.array([[2, 3, -4], [5, -5, -4], [6, -2, 3]])\n X = np.array([[-3, -2], [4, -2], [3, 3], [-3, 1], [-2, -2], [-4, -4]])\n y = np.array([1, 1, 0, 1, 0, 0])\n # W = np.array([[2, 3, -4], [5, -5, -4], [4, 0, 2], [2, 0, 1], [6, -2, 3], [1, 3, 2], [3, 2, 4]])\n depth = int(np.log2(len(W) + 1))\n n_classes = np.max(y) + 1\n\n M = vt.create_mask_dx(depth)\n X_ = np.vstack((np.ones(len(X)).T, X.T)).T\n Y_ = np.tile(y, (len(W) + 1, 1))\n\n start_time = time.time()\n for _ in range(10000):\n acc1, labels1 = vt.dt_matrix_fit(X, y, W, -M, X_, Y_ + vt.MAGIC_NUMBER)\n end_time = time.time()\n\n console.rule(\"[red]Original vector tree implementation[/red]\")\n print(f\"Labels: {labels1}\")\n print(f\"Accuracy: {acc1}\")\n print(f\"Elapsed time: {'{:.3f}'.format(end_time - start_time)} seconds\")\n\n start_time = time.time()\n for _ in range(10000):\n acc2, labels2 = vt.dt_matrix_fit_dx(X, y, W, depth, n_classes, X_, Y_, M)\n end_time = time.time()\n\n console.rule(\"[red]New vector tree implementation[/red]\")\n print(f\"Labels: {labels2}\")\n print(f\"Accuracy: {acc2}\")\n print(f\"Elapsed time: {'{:.3f}'.format(end_time - start_time)} seconds\")\n\n start_time = time.time()\n for _ in range(10000):\n acc2, labels2 = vt.dt_matrix_fit_dx2(X, y, W, depth, n_classes, X_, Y_, M)\n end_time = time.time()\n\n console.rule(\"[red]Vector tree implementation but using numpy methods instead of pure matrix stuff[/red]\")\n print(f\"Labels: {labels2}\")\n print(f\"Accuracy: {acc2}\")\n print(f\"Elapsed time: {'{:.3f}'.format(end_time - start_time)} seconds\")\n \n # solution.update_leaves_by_dataset(X_train, y_train)\n # y_pred = solution.predict_batch(X_train)\n # accuracy = np.mean([(1 if y_pred[i] == y_train[i] else 0) for i in range(len(X_train))])\n # return accuracy","repo_name":"vgarciasc/CRO-DT","sub_path":"cro_dt/time_test.py","file_name":"time_test.py","file_ext":"py","file_size_in_byte":2187,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"20899542216","text":"import asyncio\n\n\nasync def count():\n print(\"One\")\n await asyncio.sleep(1)\n print(\"Two\")\n\n\nasync def main():\n await asyncio.gather(count(), count(), count())\n\n\nif __name__ == \"__main__\":\n import time\n\n s = time.perf_counter()\n asyncio.run(main())\n elapsed = time.perf_counter() - s\n print(f\"{__file__} executed in {elapsed:0.2f} seconds.\")\n\n\n#local branch 2 development branch\n\n\n# trying out git rebase\n# second commit for trying out rebase","repo_name":"axecalibur-dev/AsycnIOPython","sub_path":"async.py","file_name":"async.py","file_ext":"py","file_size_in_byte":466,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"24886324124","text":"# encoding:utf-8\n\n# -*- Python单例模式示例 -*-\n\nclass Singleton(object):\n def __new__(cls):\n if not hasattr(cls, 'instance'):\n cls.instance = super(Singleton, cls).__new__(cls)\n return cls.instance\n\nif __name__ == '__main__':\n a = Singleton() \n b = Singleton() \n print(id(a))\n print(id(b))\n","repo_name":"tanteng/learn-python","sub_path":"singleton.py","file_name":"singleton.py","file_ext":"py","file_size_in_byte":336,"program_lang":"python","lang":"en","doc_type":"code","stars":358,"dataset":"github-code","pt":"60"}
+{"seq_id":"14780755399","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on 2022 Dec 5\n\n@author: kblackw1\n\"\"\"\n\n############ reward ################\nfrom BanditTaskParam import params,rwd,validate_T,validate_R\n\nrwd['reward']=(rwd['reward']+2*rwd['base']+2*rwd['partial']) #for equivalence to 3 step task?\nact={'left':0,'right':1} \nstates={'loc':{'Pport':1},\n 'tone':{'6kHz':6}} \nparams['state_units']={'loc':False,'tone':False} #Try false/true\nstart=(states['loc']['Pport'],states['tone']['6kHz']) #used many times\nenv_params={'start':start}\nloc=states['loc'] #used only to define R and T\ntone=states['tone'] #used only to define R and T\nparams['events_per_trial']=1\n\nRbandit={};Tbandit={} #dictionaries to improve readability/prevent mistakes\nprwdR=0.8; prwdL=0.5 #initial values. These change with each block of trials\n\nTbandit={start:{act['left']:[(start,1)],act['right']:[(start,1)]}}\nRbandit={start:{act['left']:[(rwd['reward'],prwdL),(rwd['base'],1-prwdL)], \\\n act['right']: [(rwd['reward'],prwdR),(rwd['base'],1-prwdR)]}}\n\nif __name__== '__main__':\n ######## Make sure all needed transitions have been created\n validate_T(Tbandit,msg='validate bandit T')\n validate_R(Tbandit,Rbandit,msg='validate bandit R')\n","repo_name":"neurord/TD2Q","sub_path":"Bandit1stepParam.py","file_name":"Bandit1stepParam.py","file_ext":"py","file_size_in_byte":1213,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"20736283493","text":"# print even numbers\n# for i in range(2,11,2):\n# print(i,end=\" \")\n\n# sum of even numbers\n# sum=0\n# for i in range(2,10,2):\n# print(i)\n# sum=sum+i\n# print(\"sum is \",sum)\n#\n# the sum of even nmbewrs n th\n# sum=0\n# n=int(input(\"enter the limit=\"))\n# for i in range(2,n,2):\n# print(i)\n# sum=sum+i\n# print(\"sum is\",sum)\n# the sum of odd numbers n th numbers\nsum=0\nn=int(input(\"enter the limit=\"))\nfor i in range(1,n,1):\n print(i)\n sum = sum + i\nprint(\"sum is\", sum)\n\n# multiplication table of a number\n# n=int(input(\"enter the number=\"))\n# for i in range(1,11,1):\n# print(i,\"* \",n,\"=\",i*n)\n\n","repo_name":"Muhammedshaijal/python-works","sub_path":"even numbers.py","file_name":"even numbers.py","file_ext":"py","file_size_in_byte":616,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"39533758637","text":"from tkinter import *\r\njanela = Tk()\r\n#cria o objeto janela\r\njanela.title('Cainho')\r\n#nome da janela aberta\r\njanela.geometry('280x218')\r\n#fixa a resolução da janela\r\nlabel = Label(janela, text='Fontenelle perde a OBR')\r\nlabel.grid(column = 18,row = 8)\r\njanela.mainloop()\r\n#mantém a janela aberta ","repo_name":"Craveir0/Program_Fontnelson_FtCraveiro","sub_path":"Aula_1 - FontCrav.py","file_name":"Aula_1 - FontCrav.py","file_ext":"py","file_size_in_byte":300,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"23455462014","text":"\"\"\"\r\nTakes 12 images and outputs copies that each have a calender-month table pasted onto them.\r\n\r\nUsage: calendize.py [options]\r\n\r\nThe options are:\r\n\r\n[-a --alpha - The transparency of the calendar (0..1)]\r\n[-b --bottom - The bottom margin of the calendar (by default is auto-calculated)]\r\n[-c --borderColor - The color of the table borders]\r\n[--dpi] - DPI to render (by default is auto-calculated from the image size)\r\n[-h --help]\r\n[-m --month - Output for one month only (1..12)]\r\n[-r --right - The right margin of the calendar (by default is auto-calculated)]\r\n[-t --textColor - The color of the text]\r\n[-v --verbose - Verbose output]\r\n\r\nExamples:\r\ncalendize.py 2022 my-12-images temp\r\ncalendize.py 2022 my-12-images temp --dpi 150\r\ncalendize.py 2022 my-12-images temp --dpi 150 -b blue\r\ncalendize.py 2022 my-12-images temp --dpi 150 --borderColor blue --alpha 0.7\r\n\"\"\"\r\nfrom PIL import Image\r\nfrom os.path import isfile, join\r\nfrom os import listdir\r\nfrom calendar import month\r\nfrom optparse import OptionParser\r\nimport os\r\nfrom pathlib import Path\r\n\r\nimport _date_utils\r\nimport _figure_renderer\r\nimport service_auto_dpi_calculator\r\n\r\n# usage() - prints out the usage text, from the top of this file :-)\r\n\r\n\r\ndef usage():\r\n print(__doc__)\r\n\r\n\r\n# optparse - parse the args\r\nparser = OptionParser(\r\n usage=__doc__)\r\nparser.add_option('-a', '--alpha', dest='alpha', default=1.0,\r\n help=\"The transparency of the calendar (0..1). Defaults to 1 (fully opaque).\")\r\nparser.add_option('-b', '--bottom', dest='bottom_margin', default=50,\r\n help=\"The bottom margin of the calendar. By default is auto-calculated.\")\r\nparser.add_option('-c', '--borderColor', dest='borderColor', default=\"black\",\r\n help=\"The color of the table borders - for example black or red or blue\")\r\nparser.add_option('--dpi', dest='dpi', default=None,\r\n help=\"The DPI to render (Dots Per Inch). By default is auto-calculated for image size.\")\r\nparser.add_option('-m', '--month', dest='month', default=-1,\r\n help=\"Output for one month only (1..12)\")\r\nparser.add_option('-r', '--right', dest='right_margin', default=50,\r\n help=\"The right margin of the calendar. By default is auto-calculated.\")\r\nparser.add_option('-t', '--textColor', dest='textColor', default=\"black\",\r\n help=\"The color of the text - for example black or red or blue\")\r\nparser.add_option('-v', '--verbose', dest='is_verbose',\r\n action='store_const',\r\n const=True, default=False,\r\n help=\"Turn on verbose output\")\r\n\r\n(options, args) = parser.parse_args()\r\nif (len(args) != 3):\r\n usage()\r\n exit(2)\r\n\r\nYEAR = int(args[0])\r\nINPUTDIR = args[1]\r\nOUTDIR = args[2]\r\n\r\ndpi_options = options.dpi\r\nif options.dpi is not None:\r\n dpi_options = int(options.dpi)\r\n\r\n\r\ndef is_supported_file_type(filepath):\r\n file_extensions = [\".jpg\", \".jpeg\", \".png\"]\r\n return any(map(lambda ext: filepath.lower().endswith(ext), file_extensions))\r\n\r\n\r\ndef get_input_files(input_dir):\r\n # Assumption: files are soted already - this allows user to decide which is for which month...\r\n files = [f for f in listdir(input_dir) if isfile(\r\n join(input_dir, f)) and is_supported_file_type(f)]\r\n files = map((lambda f: join(input_dir, f)), files)\r\n return list(files)\r\n\r\n\r\ndef get_image_dimensions(image_file_path):\r\n image = Image.open(image_file_path)\r\n image_width, image_height = image.size\r\n return (image_width, image_height)\r\n\r\n\r\ndef calculate_bottom_right_offset_for(calender_image, background_image, bottom_margin, right_margin):\r\n background_image_width, background_image_height = background_image.size\r\n calender_image_width, calender_image_height = calender_image.size\r\n\r\n margin_tuple_from_topleft = (background_image_width - calender_image_width -\r\n right_margin, background_image_height - calender_image_height - bottom_margin)\r\n\r\n return margin_tuple_from_topleft\r\n\r\n\r\ndef paste_with_transparency(fg_img, bg_img, alpha=1.0, box=(0, 0)):\r\n fg_img_trans = Image.new(\"RGBA\", fg_img.size)\r\n fg_img_trans = Image.blend(fg_img_trans, fg_img, alpha)\r\n bg_img.paste(fg_img_trans, box, fg_img_trans)\r\n return bg_img\r\n\r\n\r\ndef paste_calendar_into_image(calendar_image_file_path, input_image_path, output_image_path, bottom_margin, right_margin, alpha):\r\n calender_image = Image.open(calendar_image_file_path)\r\n background_image = Image.open(input_image_path)\r\n offset = calculate_bottom_right_offset_for(\r\n calender_image, background_image, bottom_margin, right_margin)\r\n background_image = paste_with_transparency(\r\n calender_image, background_image, alpha, offset)\r\n background_image.save(output_image_path)\r\n\r\n\r\ndef generate_output_image_filename(input_image_path, month, year):\r\n input_image_name = os.path.basename(input_image_path)\r\n month_2_digits = f\"{month:02d}\"\r\n month_name = _date_utils.month_name(month)\r\n output_filename = f\"{year}-{month_2_digits}-{month_name}--{input_image_name}\"\r\n # output to PNG since repeatedly saving JPG will affect quality\r\n return Path(output_filename).with_suffix('.png')\r\n\r\n\r\nfiles = get_input_files(INPUTDIR)\r\nfiles_count = len(files)\r\nif (files_count != 12):\r\n print(\r\n f\"The input folder '{INPUTDIR}' should contain 12 images but found {files_count}\")\r\n exit(3)\r\n\r\nPath(OUTDIR).mkdir(parents=True, exist_ok=True)\r\n\r\n\r\ndef calculate_dpi_and_margins(input_image_path):\r\n input_width, input_height = get_image_dimensions(input_image_path)\r\n dpi_and_margins = service_auto_dpi_calculator.DpiAndMargins(\r\n dpi_options, int(options.bottom_margin), int(options.right_margin))\r\n if (options.dpi is None):\r\n dpi_and_margins = service_auto_dpi_calculator.calculate_dpi_and_margins_from_image_size(\r\n input_width, input_height, options.is_verbose)\r\n return dpi_and_margins\r\n\r\n\r\ndef generate_for_month(month):\r\n # month: 1 = January\r\n print(f\"Generating {_date_utils.month_name(month)} {YEAR} ...\")\r\n\r\n input_image_path = files[month - 1]\r\n\r\n dpi_and_margins = calculate_dpi_and_margins(input_image_path)\r\n\r\n calendar_image_file_path = _figure_renderer.render_table_for_month(\r\n month, YEAR, OUTDIR, options.borderColor, options.textColor, dpi_and_margins.dpi)\r\n\r\n output_image_path = os.path.join(\r\n OUTDIR, generate_output_image_filename(input_image_path, month, YEAR))\r\n paste_calendar_into_image(calendar_image_file_path, input_image_path, output_image_path,\r\n dpi_and_margins.bottom_margin, dpi_and_margins.right_margin, float(options.alpha))\r\n os.unlink(calendar_image_file_path)\r\n print(f\" - calendized image saved to {output_image_path} [OK]\")\r\n\r\n\r\nif (int(options.month) >= 1):\r\n generate_for_month(int(options.month))\r\nelse:\r\n for month in range(1, 12 + 1):\r\n generate_for_month(month)\r\n\r\nprint(\"[done]\")\r\n","repo_name":"mrseanryan/calendizer","sub_path":"calendize.py","file_name":"calendize.py","file_ext":"py","file_size_in_byte":7061,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"12381304821","text":"machine = [{'s': (0, 'Выдать сдачу 0 р'), '1': (1, 'Приняты деньги'), '2': (2, 'Приняты деньги'), '5': (0, 'Выдать шоколадку')},\n {'s': (0, 'Выдать сдачу 1 р'), '1': (2, 'Приняты деньги'), '2': (3, 'Приняты деньги'), '5': (1, 'Выдать шоколадку')},\n {'s': (0, 'Выдать сдачу 2 р'), '1': (3, 'Приняты деньги'), '2': (4, 'Приняты деньги'), '5': (2, 'Выдать шоколадку')},\n {'s': (0, 'Выдать сдачу 3 р'), '1': (4, 'Приняты деньги'), '2': (0, 'Выдать шоколадку'), '5': (3, 'Выдать шоколадку')},\n {'s': (0, 'Выдать сдачу 4 р'), '1': (0, 'Выдать шоколадку'), '2': (1, 'Выдать шоколадку'), '5': (4, 'Выдать шоколадку')}]\nall_commands = set().union(*(d.keys() for d in machine))\nstate = 0\n\ncommand = input()\n\nwhile command in all_commands:\n output = machine[state][command][1]\n state = machine[state][command][0]\n print(output)\n command = input()\nelse:\n print('Был получен неизвестный сигнал. Работа автомата прекращена.')\n","repo_name":"shinkai-tester/python_beginner","sub_path":"Lesson10/choc.py","file_name":"choc.py","file_ext":"py","file_size_in_byte":1289,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"18946553957","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Apr 26 10:05:27 2018\n\n@author: willalex\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\n\ntest1 = pd.read_csv('test_input.txt', sep=',', nrows = 180000, header = None, names = ['UserID', 'Bal', 'Date', 'Balance']) \ntest2 = pd.read_csv('test_input.txt', sep=',', skiprows = 180000, header = None , \n names =['UserID', 'Trx', 'CardID', 'Date', '2GAbstract', 'TranscationName', 'Classify', 'Amount'])\n#处理错位问题,处理的方法不是很好,以后还得学习\ntest2.loc[ test2['Amount'].isnull(), 'Amount'] = test2['Classify']\ntest2.loc[ test2['Amount'] == test2['Classify'], 'Classify' ] = test2['TranscationName']\ntest2.loc[ test2['Classify'] == test2['TranscationName'], 'TranscationName'] = None\ntest2['Amount'] = test2['Amount'].astype(float)\n\ntrain1 = pd.read_csv('train.txt', sep=',', nrows = 3638790, header = None, names = ['UserID', 'Bal', 'Date', 'Balance'])\ntrain2 =pd.read_csv('train.txt', sep=',', skiprows = 3638790, header = None, \n names =['UserID', 'Trx', 'CardID', 'Date', '2GAbstract', 'TranscationName', 'Classify', 'Amount'])\n\ntrain2.loc[ train2['Amount'].isnull(), 'Amount'] = train2['Classify']\ntrain2.loc[ train2['Amount'] == train2['Classify'], 'Classify' ] = train2['TranscationName']\ntrain2.loc[ train2['Classify'] == train2['TranscationName'], 'TranscationName'] = None\ntrain2['Amount'] = test2['Amount'].astype(float)\n\n#Merge\ntestdata = pd.merge(test1, test2, how = 'left', on = ['UserID', 'Date'], suffixes = ['_bal', '_trx'])\ntraindata = pd.merge(train1, train2, how = 'left', on = ['UserID', 'Date'])\n\n#取出最后一天\nnewdf = test1.groupby('UserID')\nnewdf2 = newdf['Date'].max()\n#sortedbyBalance = grouped['Date']\n\nnewddf = pd.DataFrame({'UserID': newdf2.index, 'Date':newdf2.values})\nnewddf.drop(3000, inplace = True)\n\ntemp = pd.merge(test1, newddf, how = 'inner', on = ['UserID', 'Date'])\ntemp['Date'] = temp['Date'] + 7\n\n#submission\ntemp.to_csv('demo.txt',sep = ',',index = False, header = False)\n","repo_name":"WillAlex2017/18ZhaoShangBank","sub_path":"moneyflow/demo1.py","file_name":"demo1.py","file_ext":"py","file_size_in_byte":2023,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"31915850958","text":"# printing out the co-prime number\ndef factor(n):\n factorlist=[]\n for i in range(2,n):\n if n%i==0:\n factorlist.append(i)\n return factorlist\n\nuser=input()\ninputlist=[]\nwhile user !=\"0\":\n user=int(user)\n if user<1000000000:\n inputlist.append(user)\n user=input()\n\nfor x in inputlist:\n z=factor(x)\n final=[]\n for i in range(2,x):\n k=0\n for m in z:\n if i%m==0:\n k+=1\n if k==0:\n final.append(i)\n print(len(final))\n\n","repo_name":"ahammadshawki8/Fun-Coding","sub_path":"co-prime.py","file_name":"co-prime.py","file_ext":"py","file_size_in_byte":522,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"38364240860","text":"from recog import recog\nfrom say import say\nfrom search_package import searchpac\nfrom install_package import install\nimport os\nimport datetime\nimport wikipedia\nimport webbrowser as wb\nimport subprocess\nimport pyautogui\nimport screen_brightness_control as sbc\nfrom AppOpener import run\nfrom lxml import html\nimport requests\nimport time\n\n\n# Function to Greet the user based on time\ndef Greeting():\n hour = int(datetime.datetime.now().hour)\n if hour>=0 and hour<12:\n say(\"Good Morning !!\")\n elif hour>=12 and hour<18:\n say(\"Good Afternoon!!\")\n else:\n say(\"Good Evening !!\")\n print(\"I am Thor, How may I help you?\")\n say(\"I am Thor, How may I help you?\")\n\n\n\nif __name__ == \"__main__\":\n Greeting() \n while True:\n query = recog().lower() # converting the query to all lowercase\n # searching for keywords inside the text query\n if \"wikipedia\" in query: # if we get a keyword as wikipedia\n say(\"Searching Wikipedia\")\n # replacing wikipedia with \"\"\n query = query.replace(\"wikipedia\",\"\") \n # searching the new query directly on wikipedia using the wikipedia library\n results = wikipedia.summary(query,sentences = 2)\n say(\"According to Wikipedia\")\n print(results)\n say(results)\n elif \"increase volume by\" in query: # if we get a keyword as increase volume by\n say(\"Increasing Volume\")\n # replacing increase volume by with \"\"\n query = query.replace(\"increase volume by\",\"\")\n # converting left query to integer\n query = int(query)\n # loop for query times\n for i in range(query):\n # pressing the volume up button query times\n pyautogui.press('volumeup')\n elif \"decrease volume by\" in query:# if we get a keyword as decrease volume by\n say(\"Decreasing Volume\")\n # replacing decrease volume by with \"\"\n query = query.replace(\"decrease volume by\",\"\")\n # converting left query to integer\n query = int(query)\n # loop for query times\n for i in range(query):\n # pressing the volume down button query times\n pyautogui.press('volumedown')\n \n elif \"open youtube\" in query:# if we get a keyword as open youtube\n # opeaning youtube.com using webbrowser library\n wb.open(\"youtube.com\")\n elif \"open google\" in query:# if we get a keyword as open google\n # opeaning google.com using webbrowser library\n wb.open(\"google.com\")\n elif \"open chess\" in query:# if we get a keyword as open chess\n # opeaning chess.com using webbrowser library\n wb.open(\"chess.com\")\n elif \"play music\" in query:# if we get a keyword as play music\n # address of the direcotory where the music is \n music_Dir = \"\"\n songs = os.listdir(music_Dir)\n # playing the first song in the directory\n os.startfile(os.path.join(music_Dir,songs[0]))\n elif \"the time\" in query:# if we get a keyword as the time\n # Time in H M S format\n Time = datetime.datetime.now().strftime(\"%H:%M:%S\")\n say(\"The time is\")\n say(Time) \n elif \"open\" in query:# if we get a keyword as open\n # replacing open with \"\"\n query = query.replace(\"open \",\"\")\n run(query)\n elif \"set brightness to\" in query:# if we get a keyword as set brightness to\n say(\"setting brightness\")\n # replacing set brightness to with \"\"\n query = query.replace(\"set brightness to\",\"\")\n # converting left query to integer\n query = int(query)\n # using the screen_brightness_control to set the brightness to query\n sbc.set_brightness(query)\n elif \"click\" in query:# if we get a keyword as click\n # holding down the win button\n pyautogui.keyDown(\"win\")\n # pressing the prtscn button\n pyautogui.press(\"printscreen\")\n # releasing up the win button\n pyautogui.keyUp(\"win\")\n elif \"search on youtube\" in query:# if we get a keyword as search as youtube\n # replacing search on youtube with \"\"\n query = query.replace(\"search on youtube\",\"\")\n # appending the query to the youtube url\n query = 'https://www.youtube.com/results?search_query='+query\n print(query)\n # opeaning the new formed url \n wb.open(query)\n elif \"search on google\" in query:# if we get a keyword as search on google\n # replacing search on google with \"\"\n query = query.replace(\"search on google\",\"\")\n # appending the query to the google url\n query = 'https://www.google.com/search?q='+query\n print(query)\n # opeaning the new formed url \n wb.open(query)\n else:\n print(\"I did not recognize that\") \n say(\"I did not recognize that!!\")","repo_name":"NamanHH99/Thor-Voice-Assistant","sub_path":"scripts/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5160,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"24620471334","text":"from django.test import TestCase\nfrom django.urls import reverse, resolve\nfrom .models import Product, ProductCategory\nfrom products.views import products_view, products_by_category_view, product_detail_view\n\n# Create your tests here.\n\n\nclass ProductTests(TestCase):\n \"\"\"Product name test\"\"\"\n\n def test_str(self):\n test_name = Product(product_name='Pork chops')\n self.assertEqual(str(test_name), 'Pork chops')\n\n\nclass test_url(TestCase):\n \"\"\"Test urls are resolved\"\"\"\n\n def test_product_url_resolves(self):\n url = reverse('products')\n self.assertEquals(resolve(url).func, products_view)\n\n def test_product_category_url_resolves(self):\n url = reverse('products_by_category', args=['path'])\n self.assertEquals(resolve(url).func, products_by_category_view)\n \n def test_product_detail_view_url_resolves(self):\n url = reverse('product_detail_view', args=['path'])\n self.assertEquals(resolve(url).func, product_detail_view)\n\n\nclass BasicTest(TestCase):\n \"\"\"Check fields post\"\"\"\n\n def test_fields(self):\n category = ProductCategory(category_name=\"Pork\")\n category.save()\n product = Product()\n product.category_name = category\n product.product_category = category\n product.product_name = \"Test Product\"\n product.product_price = 2.99\n product.product_weight = 350\n product.product_serves = 4\n product.product_description = \"A new test product\"\n product.product_image_name = \"donkey.jpg\"\n product.product_stock_qty = 10\n product.product_live = True\n product.save()\n record = Product.objects.get(id=1)\n self.assertEqual(record, product)\n","repo_name":"JayPeaa/msproject5","sub_path":"products/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1721,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"9043455983","text":"#Ascii sort of works, but it sucks, so all ascii functions have been commented out by default. You can uncomment them all to enable ascii conversion\n\nimport os\nimport tkinter as tk\nimport time\nimport pyperclip\n\ndef bin_pretty(binum):\n x = len(binum)\n r = x%4\n if r == 0:\n r = 4\n r = 4 - r\n while r >> 0:\n binum = \"0\" + binum\n r -= 1\n tempbin = binum\n fbin = \"\"\n while len(tempbin) >> 0:\n fbin = fbin + tempbin[:4] + \" \"\n tempbin = tempbin[4:]\n return fbin\n\n#Handle Events\n#Convert on Return:\ndef dec_key(event):\n copiedtext.pack_forget()\n try:\n #Convert from dec\n number = ent_dec.get()\n if number == \"\":\n ent_bin.delete(0, tk.END)\n ent_hex.delete(0, tk.END)\n# ent_ascii.delete(0, tk.END)\n return\n number = number.replace(\" \", \"\")\n number = int(number)\n decinum = number\n hexinum = hex(number)[2:]\n binanum = bin(number)[2:]\n# try:\n# ascinum = chr(decinum)\n# except OverflowError:\n# ascinum = \"Too big!\"\n# pass\n fbin = bin_pretty(binanum)\n except Exception as e:\n ent_bin.delete(0, tk.END)\n ent_bin.insert(0, \"ERROR: \" + str(e))\n ent_hex.delete(0, tk.END)\n ent_hex.insert(0, \"ERROR: \" + str(e))\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, \"ERROR: \" + str(e))\n return\n\n #Print Output\n ent_bin.delete(0, tk.END)\n ent_bin.insert(0, fbin)\n ent_hex.delete(0, tk.END)\n ent_hex.insert(0, hexinum.upper())\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, ascinum)\ndef bin_key(event):\n copiedtext.pack_forget()\n try:\n #Convert from bin\n number = ent_bin.get()\n if number == \"\":\n ent_dec.delete(0, tk.END)\n ent_hex.delete(0, tk.END)\n# ent_ascii.delete(0, tk.END)\n return\n number = number.replace(\" \", \"\")\n number = int(number)\n binanum = number\n decinum = int(str(number), 2)\n hexinum = hex(decinum)[2:]\n# try:\n# ascinum = chr(decinum)\n# except OverflowError:\n# ascinum = \"Too big!\"\n# pass\n fbin = bin_pretty(str(binanum))\n except Exception as e:\n ent_dec.delete(0, tk.END)\n ent_dec.insert(0, \"ERROR: \" + str(e))\n ent_hex.delete(0, tk.END)\n ent_hex.insert(0, \"ERROR: \" + str(e))\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, \"ERROR: \" + str(e))\n return\n\n #Print Output\n ent_bin.delete(0, tk.END)\n ent_bin.insert(0, fbin)\n ent_dec.delete(0, tk.END)\n ent_dec.insert(0, decinum)\n ent_hex.delete(0, tk.END)\n ent_hex.insert(0, hexinum.upper())\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, ascinum)\ndef hex_key(event):\n copiedtext.pack_forget()\n try:\n #Convert from hex\n number = ent_hex.get()\n if number == \"\":\n ent_bin.delete(0, tk.END)\n ent_dec.delete(0, tk.END)\n# ent_ascii.delete(0, tk.END)\n return\n number = number.replace(\" \", \"\")\n hexinum = number\n decinum = int(number, 16)\n binanum = bin(decinum)[2:]\n# try:\n# ascinum = chr(decinum)\n# except OverflowError:\n# ascinum = \"Too big!\"\n# pass\n fbin = bin_pretty(binanum)\n except Exception as e:\n ent_bin.delete(0, tk.END)\n ent_bin.insert(0, \"ERROR: \" + str(e))\n ent_dec.delete(0, tk.END)\n ent_dec.insert(0, \"ERROR: \" + str(e))\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, \"ERROR: \" + str(e))\n return\n\n #Print Output\n ent_hex.delete(0, tk.END)\n ent_hex.insert(0, hexinum.upper())\n ent_bin.delete(0, tk.END)\n ent_bin.insert(0, fbin)\n ent_dec.delete(0, tk.END)\n ent_dec.insert(0, decinum)\n# ent_ascii.delete(0, tk.END)\n# ent_ascii.insert(0, ascinum)\n#def askey(event):\n# copiedtext.pack_forget()\n# try:\n# #Convert from ascii\n# number = ent_ascii.get()\n# if number == \"\":\n# ent_bin.delete(0, tk.END)\n# ent_hex.delete(0, tk.END)\n# ent_dec.delete(0, tk.END)\n# return\n# number = number.replace(\" \", \"\")\n# ascinum = number.upper()\n# decinum = \"\"\n# for i in range(len(number)):\n# decinum = decinum + str(ord(number[i]))\n# decinum = int(decinum)\n# binanum = bin(decinum)[2:]\n# hexinum = hex(decinum)[2:]\n# fbin = bin_pretty(binanum)\n# except Exception as e:\n# ent_bin.delete(0, tk.END)\n# ent_bin.insert(0, \"ERROR: \" + str(e))\n# ent_hex.delete(0, tk.END)\n# ent_hex.insert(0, \"ERROR: \" + str(e))\n# ent_dec.delete(0, tk.END)\n# ent_dec.insert(0, \"ERROR: \" + str(e))\n# return\n#\n#\n# #Print Output\n# ent_bin.delete(0, tk.END)\n# ent_bin.insert(0, fbin)\n# ent_dec.delete(0, tk.END)\n# ent_dec.insert(0, decinum)\n# ent_hex.delete(0, tk.END)\n# ent_hex.insert(0, hexinum.upper())\n#ASCII FUNCTIONALITY REMOVED!\n\n#Copy values on click:\ndef dec_copy(event):\n copyme = ent_dec.get()\n pyperclip.copy(copyme)\n copiedtext.pack()\n\ndef bin_copy(event):\n copyme = ent_bin.get()\n pyperclip.copy(copyme.replace(\" \", \"\"))\n copiedtext.pack()\n\ndef hex_copy(event):\n copyme = ent_hex.get()\n pyperclip.copy(copyme)\n copiedtext.pack()\n\n#def ascopy(event):\n# copyme = ent_ascii.get()\n# pyperclip.copy(copyme)\n\n#Define Window\nwindow = tk.Tk()\nwindow.geometry(\"960x215\")\nwindow.title(\"Conversion Tool\")\n\n#Define Widgets and Actions\nentertext = tk.Label(text=\"Press ENTER to Convert\")\ncopytext = tk.Label(text=\"Right Click to Copy a Value\")\ncopiedtext = tk.Label(text=\"Value Copied!\")\nlbl_dec = tk.Label(text=\"Decimal\")\nbutt_dec = tk.Button(text=\"Copy\")\nlbl_bin = tk.Label(text=\"Binary\")\nlbl_hex = tk.Label(text=\"Hexidecimal\")\n#lbl_ascii = tk.Label(text=\"Ascii\")\nent_dec = tk.Entry()\nent_bin = tk.Entry()\nent_hex = tk.Entry()\n#ent_ascii = tk.Entry()\n#KeyRelease would be so cool, but the bin_key func would need some serious rework\nent_dec.bind(\"\", dec_key)\nent_bin.bind(\"\", bin_key)\nent_hex.bind(\"\", hex_key)\n#ent_ascii.bind(\"\", askey)\n#ent_dec.bind(\"\", dec_key)\n#ent_bin.bind(\"\", bin_key)\n#ent_hex.bind(\"\", hex_key)\n#ent_ascii.bind(\"\", askey)\nent_dec.bind(\"\", dec_copy)\nent_bin.bind(\"\", bin_copy)\nent_hex.bind(\"\", hex_copy)\n#ent_ascii.bind(\"\", ascopy)\n\n#Load Window Elements\nentertext.pack()\ncopytext.pack()\nlbl_dec.pack(anchor=\"w\")\nent_dec.pack(fill=tk.X)\nlbl_bin.pack(anchor=\"w\")\nent_bin.pack(fill=tk.X)\nlbl_hex.pack(anchor=\"w\")\nent_hex.pack(fill=tk.X)\n#lbl_ascii.pack(anchor=\"w\")\n#ent_ascii.pack(fill=tk.X)\n\n#Start Event Loop\ntk.mainloop()\n","repo_name":"aoxhwjfoavdlhsvfpzha/BaseConverter","sub_path":"ConversionTool.py","file_name":"ConversionTool.py","file_ext":"py","file_size_in_byte":6899,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"27096304384","text":"import csv\r\nimport pymysql\r\nimport configparser\r\n\r\n\r\nconfig = configparser.ConfigParser()\r\nconfig.read_file(open('credentials.py'))\r\ndbhost = config['csc']['dbhost']\r\ndbuser = config['csc']['dbuser']\r\ndbpw = config['csc']['dbpw']\r\n\r\ndbschema = 'dryan16'\r\n\r\ndbconn = pymysql.connect(host=dbhost,\r\n user=dbuser,\r\n passwd=dbpw,\r\n db=dbschema,\r\n use_unicode=True,\r\n charset='utf8mb4',\r\n autocommit=True)\r\ncursor = dbconn.cursor()\r\n\r\n\r\nfilename = 'peopleData.csv'\r\nmyRows = []\r\ntry:\r\n with open(filename, 'r') as myCSV:\r\n data = csv.reader(myCSV)\r\n next(myCSV)\r\n for row in data:\r\n myRows.append(row)\r\n myCSV.close()\r\nexcept FileNotFoundError:\r\n print('no file!')\r\n\r\ninsertQuery = 'INSERT INTO peopleData (first_name, last_name, company_name, adress, city, \\\r\n county, state, zip, phone1, phone2, email, web) \\\r\n VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s)'\r\n \r\nfor item in myRows:\r\n zero = item[0]\r\n one = item[1]\r\n two = item[2]\r\n three = item[3]\r\n four = item[4]\r\n five = item[5]\r\n six = item[6]\r\n seven = item[7]\r\n eight = item[8]\r\n nine = item[9]\r\n ten = item[10]\r\n eleven = item[11]\r\n cursor.execute(insertQuery, (zero, one, two, three, four, five, six, seven,eight,\\\r\n nine, ten, eleven))\r\nprint(\"_______________\")\r\ndbconn.close()\r\n","repo_name":"dryan9/Database-Management","sub_path":"week13__1.py","file_name":"week13__1.py","file_ext":"py","file_size_in_byte":1530,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"10410847543","text":"# class Car:\n# def __init__(self,price,speed,fuel,mileage):\n# self.price = price\n# self.speed = speed\n# self.fuel = fuel\n# self.mileage = mileage\n\n# def display_all(self):\n# print(\"price:\",self.price,\"speed:\",self.speed,\"fuel:\",self.fuel,\"mileage:\",self.mileage)\n\n\n# car1 = Car(300,30,2,10)\n\n\n# car1.display_all()\n\nclass car:\n def __init__(self,price,speed,fuel,mileage):\n self.price=price\n if(self.price>10000):\n tax=0.15\n else:\n tax=0.12\n self.speed=speed\n self.fuel=fuel\n self.mileage=mileage\n self.tax=tax\n def display_all(self):\n # print(\"Price:{}\".format(self.price))\n # print(\"Speed:{}\".format(self.speed))\n # print(\"fuel:{}\".format(self.fuel))\n # print(\"mileage:{}\".format(self.mileage)\n # print(\"tax:{}\".format(self.tax))\n print(\"Price: {}\".format(self.price))\n print(\"Speed: {}mph\".format(self.speed))\n print(\"Fuel: {}\".format(self.fuel))\n print(\"Mileage: {}mpg\".format(self.mileage))\n print(\"Tax: {}\".format(self.tax))\n\ntwo=car(300,31,'small',2)\none=car(2000,35,'Full',15)\none.display_all()\ntwo.display_all()\n\n\n # def __init__(self,price,name,weight,brand,status):\n # self.price = price\n # self.name = name\n # self.weight = weight\n # self.brand = brand\n # self.status = \"for sale\"\n\n # def return_item(self,reason_for_return):\n # if (self.reason_for_return = \"defective\"):\n # self.status = \"defective\"\n # self.price = 0\n # elif(self.reason_for_return = \"like new\"):\n # self.status = \"for sale\"\n # elif(self.reason_for_return = \"opened\"):\n # self.status = \"used\"\n # return self","repo_name":"michaelcan2/CodingDojo","sub_path":"Python/car_j.py","file_name":"car_j.py","file_ext":"py","file_size_in_byte":1796,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"29038317447","text":"from flask import Blueprint, render_template, request\nfrom twit_app.models import db, User, Tweet\n\ndelete_routes = Blueprint('delete_routes', __name__)\n\n@delete_routes.route('/', methods=[\"GET\", \"POST\"])\ndef index():\n if request.method == \"POST\":\n print(dict(request.form))\n\n twit_user = request.form\n input_name = twit_user['delete_user']\n\n user_info = User.query.filter_by(username=input_name).one()\n user_id = user_info.__dict__['id']\n\n Tweet.query.filter_by(user_id=user_id).delete()\n User.query.filter_by(username=input_name).delete()\n\n db.session.commit()\n\n data = User.query.all()\n return render_template(\"delete.html\", data=data)","repo_name":"rmsgn100/Section3-Solo-Project","sub_path":"twit_app/routes/delete_routes.py","file_name":"delete_routes.py","file_ext":"py","file_size_in_byte":702,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"11390026538","text":"import inspect\n\n\ndef expand(func):\n \"\"\"Decorator that expands an object's attributes to fill a function's\n parameters\n\n The decorated function should be passed an object whose attributes share\n the names of the function's parameters. The objects attributes will be\n expanded into both positional and keyword parameters.\n Supports functions and class methods.\n\n Decorated function parameters:\n expand: keyword arg. True to activate object expansion,\n else the decorated function behaves normally\n\n Usage:\n >>> @expand\n ... def f(a, b, c):\n ... print \"a: %s, b: %s, c: %s\" % (a, b, c)\n >>> f(3, 2, 1)\n a: 3, b: 2, c: 1\n >>> class Args(): # similar to the object return by argparse\n ... a = 3\n ... b = 2\n ... c = 1\n >>> f(Args(), expand=True)\n a: 3, b: 2, c: 1\n \"\"\"\n\n def _inner(*args, **kwargs):\n func_varnames = inspect.getargspec(func).args\n\n if kwargs.get(\"expand\") is True and func_varnames:\n func_args = []\n\n # deal with methods\n if func_varnames[0] is \"self\":\n func_args.append(args[0])\n args = args[1:]\n func_varnames = func_varnames[1:]\n\n func_args.extend(\n [getattr(args[0], varname) for varname in func_varnames]\n )\n\n return func(*func_args)\n\n else:\n return func(*args, **kwargs)\n\n return _inner\n","repo_name":"richlanc/argspander","sub_path":"argspander/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1508,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"30952257828","text":"\nfrom django.urls import path, include\nfrom . import views\nfrom . import api\napp_name='job'\nurlpatterns = [\n path('', views.job_list,name='job_list' ),\n #path('/', views.job_detail,name = 'job_detail' ),\n path('add_job/', views.add_job,name = 'add_job' ),\n\n path('/', views.job_detail,name = 'job_detail' ),\n #apis\n path('api/all_jobs_api',api.api_job_list,name='all_jobs_api'),\n path('api/job_detail_api/',api.api_job_detail,name='job_detail'),\n #generic api vewis\n path('api/v2/all_jobs_api',api.JobListApi.as_view(),name='job_list_veiw'),\n path('api/v2/all_jobs_api/job_detail/',api.JobDetailApi.as_view(),name='job_detail_view'),\n]\n","repo_name":"abdobassel/django_job_board","sub_path":"job/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":698,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"39664134935","text":"from setuptools import find_packages, setup\n\nINSTALL_REQUIRES = [\"django==2.2\"]\n\nsetup(\n name=\"django-test\",\n packages=find_packages(),\n test_suite=\"tests\",\n python_requires=\">=3.9\",\n url=\"https://github.com/LaurenceWarne/django-test\",\n version=\"0.1\",\n author=\"Laurence Warne\",\n license=\"MIT\",\n install_requires=INSTALL_REQUIRES\n)\n","repo_name":"LaurenceWarne/django-test","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":358,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"73248351231","text":"import requests\nimport socket\nfrom utilities import get_machine_id\n\nclass HTTPReporter :\n def __init__(self, logger, config):\n self.logger = logger\n self.config = config\n self.id = None\n\n def report_results(self, results):\n if results[\"passCount\"] == 1:\n self.register(results[\"video_name\"])\n\n self.send_heart_rate(results)\n\n def register(self, video_name):\n registration_info = {\"device\":get_machine_id(), \"video\":video_name}\n try:\n if \"computer_name\" in self.config:\n registration_info.update({\"name\":socket.gethostname()})\n if \"computer_description\" in self.config:\n registration_info.update({\"description\": self.config[\"computer_description\"]})\n response = requests.post(\"{}{}\".format(self.config[\"server_url\"], \"register\"), registration_info)\n if response.status_code == 200:\n result = response.json()\n self.id = result[\"id\"]\n return True\n else:\n self.logger.error(\"Registration failure, HTTP status: {}\".format(response.status_code))\n return False\n except requests.exceptions.RequestException as err:\n self.logger.error(\"Exception: {}\".format(err))\n\n def send_heart_rate(self, results):\n try:\n http_data = results.copy()\n http_data.update({\"DeviceId\":self.id})\n for key, value in results[\"trackers\"].items():\n http_data.update({key: value})\n response = requests.post(\"{}{}\".format(self.config[\"server_url\"], \"heartrate\"), http_data)\n if response.status_code == 200:\n return True\n else:\n self.logger.error(\"POST failure, HTTP status: {}\".format(response.status_code))\n return False\n\n except requests.exceptions.RequestException as err:\n self.logger.error(\"Exception: {}\".format(err))\n","repo_name":"FredOleary/VideoBiometrics","sub_path":"collector/reporters/http_reporter.py","file_name":"http_reporter.py","file_ext":"py","file_size_in_byte":1990,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"23231441414","text":"# given 2 arrays (strings), return a bool if it's the same or not\n# case sensitive with no white spaces\n\nlist1 = [\"kobe\", \"mj\", \"lebron\", \"shaq\", \"dude\",]\nlist2 = [\"kobe\", \"shaq\", \"mj\", \"lebron\", \"other dude\"]\n\n#solution 1\ndef two_list_is_equal_a(list1, list2):\n cloned_list1 = list1\n for i in range(len(list2)):\n if list2[i] in cloned_list1:\n cloned_list1.remove(list2[i])\n else:\n return False\n if len(cloned_list1) != 0:\n return False\n return True\n\nprint(\"1A)\",two_list_is_equal_a(list1, list2),\"\\n\")\n\ndef two_list_is_equal_b(ar_1, ar_2):\n if len(ar_1) != len(ar_2):\n return False\n seen_items_set = set(ar_1)\n for item in ar_2:\n if item not in seen_items_set:\n return False\n seen_items_set.remove(item)\n if len(seen_items_set) > 0:\n return False\n return True\n\nprint(\"1B)\",two_list_is_equal_b(list1, list2),\"\\n\")","repo_name":"SamuelFolledo/SPD1.41-Communication-and-Interviewing","sub_path":"classwork/is_two_list_equal.py","file_name":"is_two_list_equal.py","file_ext":"py","file_size_in_byte":922,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"38972737615","text":"from pyclientlib import SkinnyClient\n\n\ndef main():\n c = SkinnyClient()\n while True:\n print(\"> \", end=\"\")\n s = input()\n if not s:\n continue\n split = s.split()\n if split[0] == \"open\":\n print(c.Open(split[1]))\n elif split[0] == \"write\":\n print(c.SetContent(int(split[1]), split[2]))\n elif split[0] == \"read\":\n print(c.GetContent(int(split[1])))\n elif split[0] == \"lock\":\n print(c.Acquire(int(split[1]), True))\n elif split[0] == \"unlock\":\n print(c.Release(int(split[1])))\n elif split[0] == \"trylock\":\n print(c.TryAcquire(int(split[1]), True))\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"c5h11oh/DistributedSystems-Chubby","sub_path":"demo/demo1.py","file_name":"demo1.py","file_ext":"py","file_size_in_byte":732,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"71079501310","text":"import argparse\nimport copy\nimport json\nimport os\nfrom pathlib import Path\nfrom tdw.librarian import ModelLibrarian, ModelRecord\n\ndef str2table(v):\n return v.split(',')\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--directory\", type=str, default=\"../datasets/toys4K_obj_all/\")\nparser.add_argument(\"--dest\", type=str, default=\"library2/\")\nparser.add_argument(\"--vhacd\", type=int, default=10000) #precision of vahcd mesh decomposition, 800 000 the original default\nparser.add_argument(\"--name\", type=str2table, default=\"\")\nargs = parser.parse_args()\n\nsrc = Path().resolve().joinpath(args.directory).resolve()\ndest = Path().resolve().joinpath(args.dest)\nfile_type=\"obj\"\nlibrary_path = dest.joinpath(\"toys.json\")\nif not os.path.exists(dest):\n os.mkdir(args.dest)\n\ndef create_library(library_path, src) :\n # ModelLibrarian.create_library(description=\"Toys model librarian\", path=str(Path().home().joinpath(\"postdoc/datasets/toys4k_lib/toys.json\")))\n ModelLibrarian.create_library(description=\"Toys model librarian\", path=str(library_path))\n lib = ModelLibrarian(str(library_path.resolve()))\n for f in src.rglob(\"*.\"+file_type):\n # if f.name < \"chair_144\" and f.name > \"cupcake_026\":\n # continue\n\n record = ModelRecord()\n record.name = \"\".join(list(str(f.name))[:-4])\n if record.name == \"\":\n pass\n record.wcategory = \"\".join(list(str(f.name))[:-8])\n record.wnid = record.wcategory\n record.scale_factor = 0.1 if file_type == \"fbx\" else 0.2\n for platform in record.urls:\n dest_url = dest.joinpath(record.wnid + \"/\" + record.name + \"/\" + platform)\n url=\"file:///\" + str(dest_url.resolve()).replace(\"\\\\\", \"/\")\n record.urls[platform] = url\n lib.add_or_update_record(record, overwrite=True, write=False)\n # Write to disk.\n lib.write(pretty=False)\n\ndef fix_json(library_path):\n with open(str(library_path.resolve()), \"r\") as f:\n text = list(f.read())\n for i in range(len(text)):\n if text[i] == \",\":\n if text[i + 1] != \" \" and text[i + 1] != \"\\\"\":\n text[i] = \".\"\n fixed_text = \"\".join(text)\n with open(str(library_path.resolve()), 'w') as f:\n f.write(fixed_text)\n\n\ncreate_library(library_path,src)\nfix_json(library_path)\n\njs = json.load(open(library_path))\njs_copy = copy.deepcopy(js)\nto_remove = []\nwith open(\"disabled\") as f:\n rl = f.readlines()\n for l in rl:\n cat = \"\".join(list(l)[:-5])\n name = \"\".join(list(l)[:-1])\n if cat == \"\":\n continue\n to_remove.append(name)\nfor name in to_remove:\n if name in js_copy[\"records\"]:\n del js_copy[\"records\"][name]\n\nwith open(library_path, 'w') as f:\n json.dump(js_copy, f)\n","repo_name":"Aubret/models_tdw","sub_path":"create_library.py","file_name":"create_library.py","file_ext":"py","file_size_in_byte":2779,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"24452701505","text":"import cv2\nfrom Database_Operations import database\nimport pyzbar.pyzbar as pyzbar\n\ndef Qr_scan():\n img=cv2.VideoCapture(0)\n font = cv2.FONT_HERSHEY_SIMPLEX\n\n\n while True:\n success, frame= img.read()\n decoded=pyzbar.decode(frame)\n\n for object in decoded:\n (x,y,w,h)=object.rect\n cv2.rectangle(frame,(x,y),(x+w,y+h),(0,0,255),3)\n a2=database()\n rollno=str(object[0], 'utf-8')\n rollno_l=rollno.split('/')\n print(rollno_l)\n check=a2.verify(rollno_l[0],rollno_l[1])\n\n if(check==True):\n\n cv2.putText(frame,\"DETECTED\",(x,y),font,1,(0,255,0),2)\n\n else:\n cv2.putText(frame, \"INVALID\", (x, y), font, 1, (0, 255, 0), 2)\n\n\n cv2.imshow(\"Frame\",frame)\n if cv2.waitKey(1) & 0xFF==ord('q'):\n break\n\nQr_scan()","repo_name":"rishabhgupta03/Smart-Surveillance-System","sub_path":"Scanner.py","file_name":"Scanner.py","file_ext":"py","file_size_in_byte":878,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"35225941683","text":"k,n=map(int,input().split())\n\ncable=[]\n\nfor _ in range(k):\n length=int(input())\n cable.append(length)\n\nstart,end=1,max(cable)\nans=0\n\nwhile start<=end:\n mid=(start+end)//2\n sum=0\n for c in cable:\n sum+=c//mid\n if sum>=n:\n start=mid+1\n ans=mid\n else:\n end=mid-1\n\nprint(ans)","repo_name":"SarahParkSehyun/Baekjoon_PnP","sub_path":"silver/1654.py","file_name":"1654.py","file_ext":"py","file_size_in_byte":320,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"35675432524","text":"from CryoCore import API\nfrom CryoCloud.Common.DockerProcess import DockerProcess\n\n\nccmodule = {\n \"description\": \"Run stuff in docker environments\",\n \"depends\": [\"\"],\n \"provides\": [\"\"],\n \"inputs\": {\n \"gpu\": \"Run on GPU, default False\",\n \"target\": \"Target docker\",\n \"env\": \"Environment variables\",\n \"dirs\": \"Directories to map as volumes\",\n \"arguments\": \"Arguments for docker process\",\n \"log_all\": \"Log all output as debug, default False\",\n \"debug\": \"Debug - write docker commands to /tmp/ default False\"\n },\n \"outputs\": {\n },\n \"defaults\": {\n \"runOn\": \"success\"\n }\n}\n\n\ndef process_task(worker, task, cancel_event=None):\n \"\"\"\n worker.status and worker.log are ready here.\n\n Move files from one place to another\n Needs task[\"args\"][\"src\"] and \"dst\"\n\n \"\"\"\n\n gpu = False\n env = {}\n dirs = []\n args = []\n\n a = task[\"args\"]\n gpu = a.get(\"gpu\", False)\n\n if \"target\" not in task[\"args\"]:\n raise Exception(\"Missing docker target\")\n target = a[\"target\"]\n if not isinstance(target, list):\n target = [target]\n\n if len(target) == 0:\n raise Exception(\"Require parameter 'target'\")\n\n env = a.get(\"env\", {})\n dirs = a.get(\"dirs\", [])\n args = a.get(\"arguments\", [])\n log_all = a.get(\"log_all\", False)\n debug = a.get(\"debug\", False)\n\n dp = DockerProcess(target, worker.status, worker.log, API.api_stop_event,\n dirs=dirs, env=env, gpu=gpu, args=args, log_all=log_all,\n cancel_event=cancel_event, debug=debug)\n # cancel_event=cancel_event) # Doesn't work\n retval = dp.run()\n\n worker.log.debug(\"Docker completed\")\n return worker.status[\"progress\"].get_value(), retval\n","repo_name":"Snarkdoof/cryocloud","sub_path":"CryoCloud/Modules/docker.py","file_name":"docker.py","file_ext":"py","file_size_in_byte":1770,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"19700020104","text":"import numpy as np\n\ndef read(filename):\n\t#function that reads the x and y coordinate\n\tinfile = open(filename, 'r')\n\tTimestep = []\n\tTime =[]\n\tTemperature = []\n\tDiffusionConstant = []\n\trelevant_lines = infile.readlines()[1:] #skips the irrelevant lines\n\tfor line in relevant_lines: \n\t\tdata = line.split()\n\t\tTimestep.append(float(data[0]))\n\t\tTime.append(float(data[1]))\n\t\tTemperature.append(float(data[2]))\n\t\tDiffusionConstant.append(float(data[3]))\n\tinfile.close()\n\tTimestep = np.array(Timestep)\n\tTime = np.array(Time)\n\tTemperature = np.array(Temperature)\n\tDiffusionConstant = np.array(DiffusionConstant)\n\n\treturn Timestep, Time, Temperature, DiffusionConstant\n\n#Timestep, Time, Temperature, DiffusionConstant = read('statistics.txt')\n\nT_init = []\nfor i in range(1, 1001, 50):\n\tT_init.append(i)\nT_init = np.array(T_init)\n\nT_ratio = np.array([0.669704,0.741992,0.678682,0.690175,0.693458,0.719683,0.759031,0.779948,0.709018,0.740829,0.765328,0.744585,0.761976,0.755197,0.720936,0.738801,0.766893,0.729957,0.729681,0.751274])\n\nT = T_ratio*T_init #temperature at equilibrium\n\nimport matplotlib.pyplot as plt\n\nplt.plot(T, T_ratio, 'go')\nplt.rcParams.update({'font.size': 14})\nplt.ylabel('$T/T_i [m^2/s]$')\nplt.xlabel('$T$ [K]')\nplt.show()\n\n\n","repo_name":"livewj/Project5","sub_path":"build-molecular-dynamics-fys3150-Desktop_Qt_5_7_0_clang_64bit-Debug/readfile_Tloop.py","file_name":"readfile_Tloop.py","file_ext":"py","file_size_in_byte":1235,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"17937335810","text":"###########################################################################################################\n# 稼働設定:解像度 1920*1080 表示スケール125%\n###########################################################################################################\n# モジュールインポート\nimport pyautogui as pg\nimport time\nimport MJSOpen\n\n# pandasインポート\nimport pandas as pd\n\n# 配列計算関数numpyインポート\nimport numpy as np\n\n# osインポート\nimport os\n\n# datetimeインポート\n\n# 例外処理判定の為のtracebackインポート\nimport traceback\n\n# pandas(pd)で関与先データCSVを取得\nimport pyautogui\nimport pyperclip # クリップボードへのコピーで使用\nimport Function.ExcelFileAction as EFA\nimport Function.CSVOut as FCO\nimport Function.MJSSPOPDFMarge as PDFM\nimport datetime\nimport openpyxl\nfrom openpyxl.formatting.rule import Rule\nfrom ctypes import windll\n\nimport os\nimport sys\nimport PyPDF2\nfrom pdfminer.high_level import extract_text\nfrom pdfminer.pdfparser import PDFParser\nfrom pdfminer.pdfdocument import PDFDocument\nfrom pdfminer.pdfpage import PDFPage\nimport WarekiHenkan as WK\n\n# logger設定------------------------------------------------------------------------------------------------------------\nimport logging.config\n\nlogging.config.fileConfig(r\"LogConf\\loggingMJSSysUp.conf\")\nlogger = logging.getLogger(__name__)\n# ----------------------------------------------------------------------------------------------------------------------\n\n\ndef DriverUIWaitXPATH(UIPATH, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n try:\n driver.find_element_by_xpath(UIPATH)\n Flag = 1\n return True\n except:\n Flag = 0\n if Flag == 0:\n return False\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef DriverUIWaitAutomationId(UIPATH, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n try:\n driver.find_element_by_accessibility_id(UIPATH)\n Flag = 1\n return True\n except:\n Flag = 0\n if Flag == 0:\n return False\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef DriverUIWaitName(UIPATH, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n try:\n driver.find_element_by_Name(UIPATH)\n Flag = 1\n return True\n except:\n Flag = 0\n if Flag == 0:\n return False\n\n\n# ------------------------------------------------------------r----------------------------------------------------------\ndef DriverUIWaitclassname(UIPATH, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n try:\n driver.find_element_by_class_name(UIPATH)\n Flag = 1\n return True\n except:\n Flag = 0\n if Flag == 0:\n return False\n\n\n# ----------------------------------------------------------------------------------------------------------------------\n# ----------------------------------------------------------------------------------------------------------------------\ndef DriverFindClass(UIPATH, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n try:\n elList = driver.find_elements_by_class_name(UIPATH)\n Flag = 1\n return True, elList\n except:\n Flag = 0\n if Flag == 0:\n return False\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef DriverCheck(Hub, ObjName, driver): # XPATH要素を取得するまで待機\n for x in range(10):\n if Hub == \"AutomationID\":\n if (\n DriverUIWaitAutomationId(ObjName, driver) is True\n ): # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n driver.find_element_by_accessibility_id(ObjName) # 一括電子申告送信ボタン\n return True\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n elif Hub == \"XPATH\":\n if DriverUIWaitXPATH(ObjName, driver) is True: # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n driver.find_element_by_xpath(ObjName) # 一括電子申告送信ボタン\n return True\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n elif Hub == \"Name\":\n if DriverUIWaitName(ObjName, driver) is True: # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n driver.find_element_by_Name(ObjName) # 一括電子申告送信ボタン\n return True\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef DriverClick(Hub, ObjName, driver):\n if Hub == \"AutomationID\":\n if (\n DriverUIWaitAutomationId(ObjName, driver) is True\n ): # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n OMSObj = driver.find_element_by_accessibility_id(ObjName) # 一括電子申告送信ボタン\n OMSObj.click()\n return OMSObj\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n elif Hub == \"XPATH\":\n if DriverUIWaitXPATH(ObjName, driver) is True: # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n OMSObj = driver.find_element_by_xpath(ObjName) # 一括電子申告送信ボタン\n OMSObj.click()\n return OMSObj\n else:\n # 異��待機後処理\n print(\"要素取得に失敗しました。\")\n elif Hub == \"Name\":\n if DriverUIWaitName(ObjName, driver) is True: # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n OMSObj = driver.find_element_by_Name(ObjName) # 一括電子申告送信ボタン\n OMSObj.click()\n return OMSObj\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n elif Hub == \"class_name\":\n if DriverUIWaitclassname(ObjName, driver) is True: # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n OMSObj = driver.find_element_by_class_name(ObjName) # 一括電子申告送信ボタン\n OMSObj.click()\n return OMSObj\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef ImgCheck(FolURL2, FileName, conf, LoopVal): # 画像があればTrueを返す関数\n ImgURL = FolURL2 + \"/\" + FileName\n for x in range(LoopVal):\n try:\n p = pyautogui.locateOnScreen(ImgURL, confidence=conf)\n x, y = pyautogui.center(p)\n return True, x, y\n except:\n Flag = 0\n if Flag == 0:\n return False, \"\", \"\"\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef ImgNothingCheck(FolURL2, FileName, conf, LoopVal): # 画像がなければTrueを返す\n ImgURL = FolURL2 + \"/\" + FileName\n for x in range(LoopVal):\n try:\n p = pyautogui.locateOnScreen(ImgURL, confidence=conf)\n x, y = pyautogui.center(p)\n return False\n except:\n Flag = 0\n if Flag == 0:\n return True\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef ImgCheckForList(FolURL2, List, conf, LoopVal): # リスト内の画像があればTrueと画像名を返す\n for x in range(LoopVal):\n for ListItem in List:\n ImgURL = FolURL2 + \"/\" + ListItem\n try:\n p = pyautogui.locateOnScreen(ImgURL, confidence=conf)\n x, y = pyautogui.center(p)\n return True, ListItem, x, y\n break\n except:\n Flag = 0\n if Flag == 0:\n return False, \"\", \"\", \"\"\n\n\n# ----------------------------------------------------------------------------------------------------------------------\ndef ImgClick(FolURL2, FileName, conf, LoopVal): # 画像があればクリックしてx,y軸を返す\n ImgURL = FolURL2 + \"/\" + FileName\n for x in range(10):\n if (\n ImgCheck(FolURL2, FileName, conf, LoopVal)[0] is True\n ): # OMSメニューの年調起動ボタンを判定して初期処理分け\n # 正常待機後処理\n for y in range(10):\n try:\n p = pyautogui.locateOnScreen(ImgURL, confidence=conf)\n x, y = pyautogui.center(p)\n pyautogui.click(x, y)\n time.sleep(1)\n return x, y\n except:\n print(\"失敗\")\n else:\n # 異常待機後処理\n print(\"要素取得に失敗しました。\")\n\n\n# ------------------------------------------------------------------------------------------------------------------\n# RPA用画像フォルダの作成---------------------------------------------------------\nFolURL = os.getcwd().replace(\"\\\\\", \"/\") # 先\nTFolURL = FolURL + r\"\\RPAPhoto\\MJSKomonsakiUpDate\" # 先\nLURL = TFolURL + r\"\\顧問先-名称.csv\" # 処理状況CSVのURL\n# --------------------------------------------------------------------------------\nCSVs = FCO.CsvRead(LURL)[1]\nCSVLen = len(CSVs)\nfor CV in range(CSVLen):\n if CV > 2222:\n CSVsRow = CSVs.iloc[CV]\n if not CSVsRow[\"顧問先\"] == CSVsRow[\"顧問先\"]:\n print(\"nan\")\n else:\n print(CSVsRow[\"コード\"])\n KCB = ImgCheck(TFolURL, r\"\\K_CodeBox.png\", 0.9, 10)\n if KCB[0] is True:\n pyautogui.click(KCB[1] + 100, KCB[2]) # 横軸,縦軸\n pyperclip.copy(str(CSVsRow[\"コード\"]))\n pg.hotkey(\"ctrl\", \"v\")\n pg.press(\"return\")\n time.sleep(1)\n MB = ImgCheckForList(\n TFolURL,\n [\n r\"\\M_Bar.png\",\n r\"\\M_Bar2.png\",\n r\"\\M_Bar3.png\",\n r\"\\Name.png\",\n r\"\\Name2.png\",\n r\"\\Name3.png\",\n ],\n 0.9,\n 10,\n )\n if MB[0] is True:\n pyautogui.click(MB[2], MB[3]) # 横軸,縦軸\n RS = ImgCheckForList(\n TFolURL, [r\"\\R_Sou.png\", r\"\\R_Sou2.png\"], 0.9, 10\n )\n if RS[0] is True:\n pyautogui.click(RS[2] + 150, RS[3]) # 横軸,縦軸\n pg.press(\n [\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n \"backspace\",\n ]\n )\n pg.press(\n [\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n \"delete\",\n ]\n )\n pyperclip.copy(str(CSVsRow[\"コード\"]))\n pg.hotkey(\"ctrl\", \"v\")\n ImgClick(TFolURL, r\"\\U_Btn.png\", 0.9, 10)\n time.sleep(1)\n while (\n pg.locateOnScreen(\n TFolURL + r\"\\\\\" + \"UpDateFlag.png\", confidence=0.9\n )\n is not None\n ):\n time.sleep(1)\n time.sleep(1)\n","repo_name":"hasegawakaikeirpa/RPAScript","sub_path":"MJSKomonsakiUpDate.py","file_name":"MJSKomonsakiUpDate.py","file_ext":"py","file_size_in_byte":13678,"program_lang":"python","lang":"ja","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"38007371934","text":"import pygame\ndef share_diagonal(x0, y0, x1, y1):\n \"\"\" Is (x0, y0) on a shared diagonal with (x1, y1)? \"\"\"\n dy = abs(y1 - y0) # Calc the absolute y distance\n dx = abs(x1 - x0) # CXalc the absolute x distance\n return dx == dy \n\ndef col_clashes(bs, c):\n \"\"\" Return True if the queen at column c clashes\n with any queen to its left.\n \"\"\"\n for i in range(c): # Look at all columns to the left of c\n if share_diagonal(i, bs[i], c, bs[c]):\n return True\n\n return False \n\ndef has_clashes(the_board):\n \"\"\" Determine whether we have any queens clashing on the diagonals.\n We're assuming here that the_board is a permutation of column\n numbers, so we're not explicitly checking row or column clashes.\n \"\"\"\n for col in range(1,len(the_board)):\n if col_clashes(the_board, col):\n return True\n return False\n\ngravity = 0.1\n\nclass QueenSprite:\n\n def __init__(self, img, target_posn):\n self.image = img\n self.target_posn = target_posn\n (x, y) = target_posn\n self.posn = (x, 0) # Start ball at top of its column\n self.y_velocity = 0 # with zero initial velocity\n\n def update(self):\n self.y_velocity += gravity\n (x, y) = self.posn\n new_y_pos = y + self.y_velocity\n (target_x, target_y) = self.target_posn # Unpack the position\n dist_to_go = target_y - new_y_pos # How far to our floor?\n \n if dist_to_go < 0: # Are we under floor?\n self.y_velocity = -0.65 * self.y_velocity # Bounce\n new_y_pos = target_y + dist_to_go # Move back above floor\n \n self.posn = (x, new_y_pos) # Set our new position.\n\n def draw(self, target_surface): # Same as before.\n target_surface.blit(self.image, self.posn)\n \n def contains_point(self, pt):\n \"\"\" Return True if my sprite rectangle contains point pt \"\"\"\n (my_x, my_y) = self.posn\n my_width = self.image.get_width()\n my_height = self.image.get_height()\n (x, y) = pt\n return ( x >= my_x and x < my_x + my_width and\n y >= my_y and y < my_y + my_height)\n \n def handle_click(self):\n self.y_velocity += -8 # Kick it up \n\n\n\"TWO\" \nclass DukeSprite:\n\n def __init__(self, img, target_posn):\n self.image = img\n self.posn = target_posn\n self.anim_frame_count = 0\n self.curr_patch_num = 0\n\n def update(self):\n if self.anim_frame_count > 0:\n self.anim_frame_count = (self.anim_frame_count + 1 ) % 60\n self.curr_patch_num = self.anim_frame_count // 6\n\n def draw(self, target_surface):\n patch_rect = (self.curr_patch_num * 50, 0,\n 50, self.image.get_height())\n target_surface.blit(self.image, self.posn, patch_rect)\n\n def contains_point(self, pt):\n (my_x, my_y) = self.posn\n my_width = self.image.get_width()/10\n #The reason that clicking on the squares to the right of Duke still triggered the animation was because the whole spritesheet was still displayed as the image, but only the first part, where he is shown as idle, can be seen. Thus, the solution is to split the width by however many different sprite states Duke's sheet has, in this case 10. \n my_height = self.image.get_height()\n (x, y) = pt\n return ( x >= my_x and x < my_x + my_width and\n y >= my_y and y < my_y + my_height)\n\n def handle_click(self):\n if self.anim_frame_count == 0:\n self.anim_frame_count = 5\n\ndef draw_board(the_board):\n pygame.init()\n my_clock = pygame.time.Clock()\n colors = [(255,0,0), (0,0,0)] # Set up colors [red, black]\n \n n = len(the_board) # This is an NxN chess board.\n surface_sz = 540 # Proposed physical surface size.\n sq_sz = surface_sz // n # sq_sz is length of a square.\n surface_sz = n * sq_sz # Adjust to exactly fit n squares.\n\n # Create the surface of (width, height), and its window.\n surface = pygame.display.set_mode((surface_sz, surface_sz))\n\n ball = pygame.image.load(\"ball.png\")\n ball_offset = (sq_sz-ball.get_width()) // 2\n all_sprites=[]\n\n \n for (col, row) in enumerate(the_board):\n a_queen = QueenSprite(ball,\n (col*sq_sz+ball_offset, row*sq_sz+ball_offset))\n all_sprites.append(a_queen)\n # Load the sprite sheet\n duke_sprite_sheet = pygame.image.load(\"duke_spritesheet.png\")\n \n # Instantiate two duke instances, put them on the chessboard\n duke1 = DukeSprite(duke_sprite_sheet,(sq_sz*2, 0))\n duke2 = DukeSprite(duke_sprite_sheet,(sq_sz*5, sq_sz))\n \n # Add them to the list of sprites which our game loop manages\n all_sprites.append(duke1)\n all_sprites.append(duke2)\n\n while True:\n # Look for an event from keyboard, mouse, etc.\n ev = pygame.event.poll()\n if ev.type == pygame.QUIT:\n break;\n if ev.type == pygame.KEYDOWN:\n key = ev.dict[\"key\"]\n if key == 27: # On Escape key ...\n break # leave the game loop.\n if key == ord(\"r\"):\n colors[0] = (255, 0, 0) # Change to red + black.\n elif key == ord(\"g\"):\n colors[0] = (0, 255, 0) # Change to green + black.\n elif key == ord(\"b\"):\n colors[0] = (0, 0, 255) # Change to blue + black.\n \n if ev.type == pygame.MOUSEBUTTONDOWN:\n posn_of_click = ev.dict[\"pos\"]\n for sprite in all_sprites:\n if sprite.contains_point(posn_of_click):\n sprite.handle_click()\n break \n\n # Ask every sprite to update itself.\n for sprite in all_sprites:\n sprite.update()\n\n # Draw a fresh background (a blank chess board)\n # ... same as before ...\n # Look for an event from keyboard, mouse, etc.\n ev = pygame.event.poll()\n if ev.type == pygame.QUIT:\n break\n\n # Draw a fresh background (a blank chess board)\n for row in range(n): # Draw each row of the board.\n c_indx = row % 2 # Alternate starting color\n for col in range(n): # Run through cols drawing squares\n the_square = (col*sq_sz, row*sq_sz, sq_sz, sq_sz)\n surface.fill(colors[c_indx], the_square)\n # Now flip the color index for the next square\n c_indx = (c_indx + 1) % 2\n \n # Ask every sprite to draw itself.\n for sprite in all_sprites:\n sprite.draw(surface)\n\n pygame.display.flip()\n my_clock.tick(60)\n \n pygame.quit() \n \n#draw_board([6, 4, 2, 0, 5, 7, 1, 3])\n \n\"THREE\"\ndef pokerhand():\n import random\n rng=random.Random()\n cards=list(range(1, 53))\n rng.shuffle(cards)\n cards=cards[:5]\n return cards\n\ndef displayhand():\n pygame.init()\n pygame.display.set_caption('Poker Hand')\n scr_width=400\n scr_height=400\n surface=pygame.display.set_mode((scr_width, scr_height))\n surface.fill((0, 100, 0)) \n cards=pygame.image.load(\"cards.jpg\")\n width=cards.get_width()\n height=cards.get_height() \n hand=pokerhand()\n print(hand)\n x=(scr_width-width*5/13)/2\n y=(scr_height-height/4)/2\n pos=(x,y) \n for i, num in enumerate(hand):\n card_width=width/13*((num-1)%13)\n card_height=height/4*((num-1) // 13)\n rect=(card_width, card_height, width/13, height/4)\n surface.blit(cards, pos, rect)\n x=x+width/13\n pos=(x,y) \n while True:\n ev=pygame.event.poll()\n if ev.type==pygame.QUIT:\n break\n \n pygame.display.flip()\n \n pygame.quit()\n \ndisplayhand()","repo_name":"JTTheAxis/thinklikeacompscientist","sub_path":"Chapter 17 Practice.py","file_name":"Chapter 17 Practice.py","file_ext":"py","file_size_in_byte":7997,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"34562028164","text":"import urllib.request\nfrom bs4 import BeautifulSoup\n\napi = 'https://www.kma.go.kr/weather/forecast/mid-term-rss3.jsp'\nurls = urllib.request.urlopen(api).read()\nsoup = BeautifulSoup(urls, 'html.parser')\n\ncities = soup.find_all(\"city\")\ndata = soup.find_all(\"data\")\ndates = soup.find(\"tmef\")\n\nprint(dates.string)\nfor i in range(len(cities)):\n print(f'{cities[i].string}의 날씨는 {data[i*13].find(\"wf\").string}입니다.')\n\nprint(len(cities), len(data))","repo_name":"voidgogo/BigData","sub_path":"week07_web02.py","file_name":"week07_web02.py","file_ext":"py","file_size_in_byte":456,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"74380517631","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Sep 17 12:09:42 2021\n\n@author: lowkg\n\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom scipy.optimize import curve_fit\n\nTe = ( 38.58, 95.09,211.81,420.01,855.64)\nFc= ( 3.42,14.70,20.28,25.21,34.5)\n'''\nTe = ( 43.18, 100.69,217.11,423.74,855.64)\nFc= ( 13.98,19.23,38.54,45.58,52.05)\n'''\nEq_Option = [0,1]\nlegend =[]\n\ndef Func_ScandinavianCode(Te, Fc_u, tau, a):\n # ScandinavianCode Equation to calculate Compressive Strength\n return Fc_u * (np.exp(-np.power((tau / Te), a)))\n\ndef Func_AmericanCode(Te, Fc_u, k, Te_0):\n # AmericanCode Equation to calculate Compressive Strength\n return Fc_u*((k*(Te-Te_0))/(1+k*(Te-Te_0)))\n\n# array with 3 elements in order of Fc_u,tau,a\ndef Fc_ParameterGraph(pars,option): \n df = pd.DataFrame()\n Te_graph = []\n Fc_graph = []\n \n Fc_u = round(pars[0],4)\n par1 = round(pars[1],4)\n par2 = round(pars[2],4)\n \n iLim = max(Te)\n i = 0.5\n while i < iLim:\n Te_plot = i\n if option == 0:\n Fc_cal = Fc_u * (np.exp(-np.power((par1 / Te_plot), par2)))\n PlotLegend = 'ScandinavianEq'\n elif option == 1:\n Fc_cal = Fc_u*((par1*(Te_plot-par2))/(1+par1*(Te_plot-par2)))\n PlotLegend = 'AmericanEq'\n Te_graph.append(i)\n Fc_graph.append(Fc_cal)\n i+=0.5\n\n df['Te'] = Te_graph\n df['Fc'] = Fc_graph\n \n plt.scatter(Te, Fc)\n legend.append(PlotLegend)\n plt.plot(df['Te'], df['Fc'],linestyle='dashdot')\n plt.legend(legend)\n\ndef DeviationCalculation(pars,Te,Fc,option):\n Fc_u = round(pars[0],4)\n par1 =round(pars[1],4)\n par2 = round(pars[2],4)\n \n for i in range(len(Te)):\n Te_input = Te[i]\n if option == 0:\n Fc_cal = Fc_u * (np.exp(-np.power((par1 /Te_input), par2)))\n elif option == 1:\n Fc_cal = Fc_u*((par1*(Te_input-par2))/(1+par1*(Te_input-par2)))\n \n Dev = round((Fc_cal - Fc[i])/Fc[i]*100,2)\n print(\"Fc: \" + str(Fc[i]) + \"\\t Deviation: \" + str(Dev))\n print(\"\")\n\ndef getStrengthParameter(Te,Fc):\n F_strength =[Func_ScandinavianCode,Func_AmericanCode]\n Fc_0 = 25\n for option in Eq_Option:\n # Guestimate initial parameter\n if option == 0:\n par1 = 10 ; par2 = 0.6;\n param = ['Fc_u','tau','a']\n \n elif option == 1:\n par1 = 0 ; par2 = 0\n param = ['Fc_u','k','Te_0']\n \n pars, cov = curve_fit(f=F_strength[option], xdata=Te, ydata=Fc,\n p0=[Fc_0,par1,par2])\n \n for i in range(len(param)):\n print(param[i] + ':' + str(round(pars[i],2)))\n \n Fc_ParameterGraph(pars,option)\n DeviationCalculation(pars, Te, Fc,option)\n \n return pars\n\nplt.figure(figsize=(6, 4))\npars = getStrengthParameter(Te,Fc)\n\n\n\n","repo_name":"uen6ueakKG/ParameterCurveFit","sub_path":"CalibrationParameter.py","file_name":"CalibrationParameter.py","file_ext":"py","file_size_in_byte":2886,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"38952180636","text":"\n\nfrom AbstractSender import AbstractSender\nimport tempfile\nimport subprocess\nimport os\n\nclass PierfSender(AbstractSender):\n '''\n classdocs\n '''\n pierfXML_template =\\\n'''\n\n \n \n \n \n {1}\n \n \n \n \n\n'''\n def __init__(self, context):\n '''\n Constructor\n '''\n self.__defaults = {'interface':'eth0',\n 'pierf_app_full_file_name' : '/opt/pierf/pierf'}\n \n context.update(self.__defaults)\n self.__context = context\n \n def cutCRC32(self, packet ):\n return packet[0:-4]\n \n def convert2Hex(self, packet ):\n result = \"\"\n for byte in packet:\n result += \":\" + \"{0:2x}\".format(ord(byte)).replace(' ', '0').upper()\n \n return result\n \n def generatePierfXml(self, hexPacket):\n pierfXML = self.pierfXML_template.format(self.__context['interface'], \n hexPacket)\n return pierfXML\n \n def runPierf(self, configFileName):\n subprocess.call([self.__context['pierf_app_full_file_name'],configFileName])\n \n def sendPacket(self, packet): \n try:\n tmpFileName = None\n with tempfile.NamedTemporaryFile(mode='w+', delete=False) as tmpFile:\n tmpFile.write( \n self.generatePierfXml(\n self.convert2Hex(\n self.cutCRC32(packet))))\n \n tmpFile.flush()\n tmpFileName = tmpFile.name\n \n self.runPierf(tmpFileName)\n finally:\n if tmpFileName:\n os.unlink(tmpFileName)\n\n \n return None","repo_name":"olegh/packet_synthesier","sub_path":"src/PierfSender.py","file_name":"PierfSender.py","file_ext":"py","file_size_in_byte":1895,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"4108102230","text":"from .models import MainTable\nfrom .models import MainTable, CafedraClasses\n\n\nclass Rasp:\n\n # Функция устанавливает в главную таблицу преподователей\n def setTeacher(Table, PrepodArray, priority):\n print(priority)\n for x in range(len(Table)):\n for PrepodTable in PrepodArray:\n day = PrepodTable.filter(Date=Table[x].vremya)\n if day and Table[x].Prepod == \"\":\n mainTableday = Table.filter(vremya=Table[x].vremya)\n for lesson in day:\n for mainDay in mainTableday:\n if lesson.LessonNumber == mainDay.NLecii:\n if priority == 1:\n obj = Table.get(pk=mainDay.id)\n obj.Prepod = lesson.teacher.last_name + \" \" + lesson.teacher.first_name # надо откуда то брать имя препода\n obj.teacherId = lesson.teacher\n obj.save()\n\n if priority == 2:\n isBusy = MainTable.objects.filter(NLecii=mainDay.NLecii, vremya=mainDay.vremya)\n if isBusy:\n obj = Table.get(pk=mainDay.id + 1)\n obj.Prepod = lesson.teacher.last_name + \" \" + lesson.teacher.first_name # надо откуда то брать имя препода\n obj.teacherId = lesson.teacher\n obj.save()\n\n # функция устанавливает предмет в главную таблицу\n def setSubject(Table, studyPlan, hours):\n exTable = Table.exclude(Prepod=\"\")\n print(studyPlan)\n for x in range(len(exTable)):\n for y in range(len(studyPlan)):\n if exTable[x].teacherId == studyPlan[y].teacher and hours > 0:\n obj = Table.get(pk=exTable[x].id)\n hoursType = studyPlan[y].hours\n\n if studyPlan[y].typeSubject == \"Лекция\" and obj.Auditoriya == \"\" and studyPlan[\n y].remaningLectures > 0:\n clases = CafedraClasses.objects.filter(AllowedLections=\"True\")\n obj.Predmet = studyPlan[y].subject\n obj.Auditoriya = clases[0].ClassName\n obj.Podgruppa = studyPlan[y].typeSubject\n obj.save()\n studyPlan[y].remaningLectures -= 1\n studyPlan[y].save()\n hours -= 1\n continue\n\n if studyPlan[y].typeSubject == \"Практика\" and obj.Auditoriya == \"\" and studyPlan[\n y].remaningLectures > 0:\n clases = CafedraClasses.objects.filter(AllowedPractice=\"True\")\n obj.Predmet = studyPlan[y].subject\n obj.Auditoriya = clases[0].ClassName\n obj.Podgruppa = studyPlan[y].typeSubject\n obj.save()\n studyPlan[y].remaningLectures -= 1\n studyPlan[y].save()\n hours -= 1\n continue\n\n if studyPlan[y].typeSubject == \"ЛабРабота\" and obj.Auditoriya == \"\" and studyPlan[\n y].remaningLectures > 0:\n clases = CafedraClasses.objects.filter(AllowedLabs=\"True\")\n obj.Predmet = studyPlan[y].subject\n obj.Auditoriya = clases[0].ClassName\n obj.Podgruppa = studyPlan[y].typeSubject\n obj.save()\n\n studyPlan[y].remaningLectures -= 1\n studyPlan[y].save()\n hours -= 1\n continue\n\n def lastIterate(Table,studyPlan):\n Table = Table.filter(Prepod__exact='')\n teacherId = studyPlan[0].teacher\n for x in Table:\n x.Prepod = teacherId.last_name + \" \" + teacherId.first_name\n lections = studyPlan.filter(typeSubject=\"Лекция\")[0].remaningLectures\n practice = studyPlan.filter(typeSubject=\"Практика\")[0].remaningLectures\n for type in studyPlan:\n if type.typeSubject == \"Лекция\" and type.remaningLectures > 0:\n clases = CafedraClasses.objects.filter(AllowedLections=\"True\")\n x.Predmet = type.subject\n x.Auditoriya = clases[0].ClassName\n x.Podgruppa = type.typeSubject\n x.save()\n type.remaningLectures -= 1\n type.save()\n continue\n if type.typeSubject == \"Практика\" and type.remaningLectures > 0 and lections == 0:\n clases = CafedraClasses.objects.filter(AllowedPractice=\"True\")\n x.Predmet = type.subject\n x.Auditoriya = clases[0].ClassName\n x.Podgruppa = type.typeSubject\n x.save()\n type.remaningLectures -= 1\n type.save()\n\n continue\n if type.typeSubject == \"ЛабРабота\" and type.remaningLectures > 0 and lections == 0 and practice == 0:\n clases = CafedraClasses.objects.filter(AllowedLabs=\"True\")\n x.Predmet = type.subject\n x.Auditoriya = clases[0].ClassName\n x.Podgruppa = type.typeSubject\n x.save()\n type.remaningLectures -= 1\n type.save()\n\n continue\n","repo_name":"NavigatorZero/DiplomTimetable","sub_path":"Timetable/MainApp/Rasp.py","file_name":"Rasp.py","file_ext":"py","file_size_in_byte":5943,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"26162330849","text":"#!/usr/bin/env python\n#-*- coding:utf-8 -*-\n\nimport sys\nimport socket\n\nCOMMON_DOMAINS = [\n 'api',\n 'callbacks',\n 'apt',\n 'rpm',\n 'repo',\n 'www',\n 'www1',\n 'www2',\n 'www3',\n 'www4',\n 'www5',\n 'www6',\n 'www7',\n 'www8',\n 'www9',\n 'ws',\n 'beta',\n 'web',\n 'ftp',\n 'irc',\n 'pptp',\n 'old',\n 'deprecated',\n 'new',\n 'dns',\n 'dev',\n 'test',\n 'test2',\n 'stage',\n 'staging',\n 'admin',\n 'auth',\n 'database',\n 'db',\n 'master',\n 'slave',\n 'mysql',\n 'mariadb',\n 'mongodb',\n 'postgres',\n 'webmail',\n 'pop',\n 'smtp',\n 'mail',\n 'intranet',\n 'vpn',\n 'vps',\n 'jump',\n 'jumpbox',\n 'tunnel',\n 'demo',\n 'temp',\n 'backend',\n 'cms',\n 'crm',\n 'support',\n 'blog',\n 'help',\n 'stats',\n 'statistics',\n 'health',\n 'status',\n 'owncloud',\n 'nextcloud',\n 'redmine',\n 'source',\n 'git',\n 'gitlab',\n 'svn',\n 'hg',\n 'mercurial',\n 'bitbucket',\n 'wiki',\n 'mediawiki',\n 'office',\n 'backup',\n 'virtual',\n 'docker',\n 'k8s',\n 'kubernetes',\n 'visma',\n 'logs',\n 'media',\n 'static',\n 'images',\n 'img',\n 'imgs',\n 'assets',\n 'cache',\n 'episerver',\n 'epi',\n 'monitor',\n 'monitoring',\n 'sip',\n 'a','b','c','d','e','f','g','h','i','j',\n 'k','l','m','n','o','p','q','r','s','t',\n 'u','v','w','x','y','z',\n 'www.m',\n 'shop',\n \"cart\",\n 'store',\n 'buy',\n 'app',\n 'cpanel',\n 'home',\n 'forum',\n 'cdn',\n 'secure',\n 'whm',\n 'files',\n 'filetransfer',\n 'portal',\n 'member',\n 'members',\n 'community',\n 'subscriber',\n 'subscribers',\n 'payment',\n 'payments',\n 'mobile',\n 'phpmyadmin',\n 'cloud',\n 'fake',\n 'login',\n 'account',\n 'accounts',\n 'people',\n 'live',\n 'apps',\n 'jira',\n 'internal',\n 'secret',\n 'feed',\n 'go',\n 'redirect',\n 'partners',\n 'labs',\n 'en',\n 'us',\n 'fr',\n 'se',\n 'uk',\n 'fi',\n 'dk',\n 'de',\n 'no',\n 'au',\n 'be',\n 'ru',\n 'ch',\n 'it',\n 'nl',\n 'com',\n 'next',\n 'one',\n 'two',\n 'three',\n 'four',\n 'five',\n 'six',\n 'seven',\n 'eight',\n 'nine',\n 'ten',\n 'open',\n 'info',\n 'developer',\n 'enterprise',\n 'console',\n 'try',\n 'discuss',\n 'docs',\n 'newsletter',\n 'premium',\n 'feedback',\n 'updates',\n 'update',\n 'download',\n 'downloads',\n 'projects',\n 'project',\n 'survey',\n 'my',\n 'server',\n 'controller',\n 'stream',\n 'id',\n 'marketing',\n 'tracking',\n 'campaign',\n 'investors',\n 'investor',\n 'signup',\n 'transition',\n 'link',\n 'links',\n 'group',\n 'ups',\n 'dhl',\n 'fedex',\n 'distribution',\n 'observer',\n 'shopify',\n 'wordpress',\n 'aws',\n 'heroku',\n 'google',\n 'squarespace',\n 'resources',\n 'preview',\n]\n\ndef getIP(domain):\n try:\n data = socket.gethostbyname_ex(domain)\n ipx = data[2]\n return ipx\n except:\n return []\n\ndef main():\n\n result = []\n\n if len(sys.argv) != 2:\n print(\"Requiers a domain as argument\")\n return 0;\n\n for i in COMMON_DOMAINS:\n domain = i+\".\"+sys.argv[1]\n ips = getIP(domain)\n if len(ips) != 0:\n sys.stdout.write(\"%s:\\n\" % domain)\n for ip in ips:\n sys.stdout.write(\"\\t\\t\\t%s\\n\" % ip)\n sys.stdout.write('\\n\\n')\n sys.stdout.flush()\n\n\n\nif __name__ == '__main__':\n sys.exit(main())\n\n\n","repo_name":"samiberndtson/dns.py","sub_path":"dns.py","file_name":"dns.py","file_ext":"py","file_size_in_byte":3602,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"907354614","text":"import json\nimport fasttext\nimport pandas as pd\n\n# import os\n# os.chdir(\"/Users/a.orzikulov/Desktop/GitHub/Integro\") ################\n\nimport util ################\n\n\n# util.main() ################\n# file_path = \"output-25-8-2021.json\" ################\n\n\ndef main(file_path):\n with open(file_path, \"r\", encoding=\"UTF-8\") as file: ################\n data = json.load(file)\n\n LANGUAGE_MODEL_PATH = '../lid.176.bin'\n model = fasttext.load_model(LANGUAGE_MODEL_PATH)\n russian_texts = []\n uzbek_texts = []\n for channel, posts in data.items():\n for post_id, (date, views, post_text) in posts.items(): ## data[\"https://t.me/Buka_tumani\"].values()\n if not any([post_text == \"None\", post_text == \"\", post_text is None]):\n post_text = util.multiple_replace(util.REPLACEMENTS, post_text)\n prediction, score = model.predict(post_text)\n if prediction[0] == \"__label__ru\" and score[0] >= 0.8:\n russian_texts.append([channel, post_id, date, views, post_text])\n else:\n fixed_message = \"\"\n for idx, character in enumerate(post_text):\n if character == \"Е\" or character == \"е\":\n fixed_message += util.change_e(\n idx, post_text, util.TRANSLATOR, util.STOP_SYMBOLS, util.VOWELS)\n elif character == \"Ц\" or character == \"ц\":\n fixed_message += util.change_ts(\n idx, post_text, util.TRANSLATOR, util.STOP_SYMBOLS, util.VOWELS)\n elif character in util.TRANSLATOR:\n fixed_message += util.TRANSLATOR[character]\n else:\n fixed_message += character\n\n uzbek_texts.append([channel, post_id, date, views, fixed_message])\n\n util.log('info', f\"The number of Uzbek posts is {len(uzbek_texts)}\")\n util.log('info', f\"The number of Russian posts is {len(russian_texts)}\")\n header = [\"channel\", \"post_id\", \"date\", \"views\", \"post\"] # \"label\"\n if uzbek_texts:\n data_frame = pd.DataFrame(uzbek_texts)\n data_frame.to_excel('uzbek.xlsx', header=header, index=False)\n if russian_texts:\n data_frame = pd.DataFrame(russian_texts)\n data_frame.to_excel('russian.xlsx', header=header, index=False)\n\n util.log('success', \"Two Excel files have been created.\")\n\n# import numpy as np\n\n# with open(\"../uzbek_texts.txt\", \"w\", encoding=\"UTF-8\") as file:\n# for text in uzbek_texts:\n# file.writelines(text + \"\\n\")\n\n# with open(\"../russian_texts.txt\", \"w\", encoding=\"UTF-8\") as file:\n# for text in russian_texts:\n# file.writelines(text + \"\\n\")\n\n# russian_texts = []\n# uzbek_texts = []\n# for posts in data.values():\n# for post_text in posts.values(): ## data[\"https://t.me/Buka_tumani\"].values()\n# if not any([post_text == \"None\", post_text == \"\", post_text is None]):\n# if np.random.random() <= 0.03693444136657433:\n# post_text = util.multiple_replace(REPLACEMENTS, post_text)\n# prediction, score = model.predict(post_text)\n# if prediction[0] == \"__label__ru\" and score[0] >= 0.8:\n# russian_texts.append(post_text)\n# else:\n# uzbek_texts.append(post_text)\n\n# print(len(uzbek_texts))\n# print(len(russian_texts))\n","repo_name":"Asrorbek-Orzikulov/sentiment_analysis","sub_path":"action/converter.py","file_name":"converter.py","file_ext":"py","file_size_in_byte":3506,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"12343359757","text":"#!/usr/bin/env python3\nimport jsonschema\nimport sys\nimport yaml\n\nif len(sys.argv) < 2:\n sys.exit(\"Usage: {} MANUAL_FILE ...\".format(sys.argv[0]))\n\nwith open(\"manual_schema.yml\", \"r\") as schema_file:\n manual_schema = yaml.safe_load(schema_file)\n\nfor path in sys.argv[1:]:\n with open(path, \"r\") as manual_file:\n manual_data = yaml.safe_load(manual_file)\n try:\n jsonschema.validate(instance=manual_data, schema=manual_schema)\n except jsonschema.exceptions.ValidationError as e:\n print(\"Failed to validate:\", path, file=sys.stderr)\n sys.exit(e)\n","repo_name":"jqlang/jq","sub_path":"docs/validate_manual_schema.py","file_name":"validate_manual_schema.py","file_ext":"py","file_size_in_byte":584,"program_lang":"python","lang":"en","doc_type":"code","stars":27144,"dataset":"github-code","pt":"60"}
+{"seq_id":"22611509401","text":"from __future__ import print_function\nfrom django.template import Context, loader, RequestContext\nfrom app.models import Contacts\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth import authenticate, login, logout\n\n#ModelForms\nfrom django.shortcuts import render, redirect, get_object_or_404\nfrom django import forms\nfrom django.utils import timezone\nfrom django.views.decorators import csrf\nfrom django.urls import reverse , resolve\nfrom django.core.mail import send_mail\n\nfrom .forms import ContactForm, LookupForm, CancelForm\n\n#Recycling logic\nfrom .recycle import parse_address, confirm_subscription, cancel_subscription, get_initial_message, insert_contact, get_zones, insert_initial_message, select_initial_message\n\n#Hash value for obfuscating primary key. stackoverflow.com/questions/10559935/django-how-do-i-hash-a-url-from-the-database-objects-primary-key\nfrom mysite.passwords import OBFUSCATE\n\n#Logo image\nlogo_image=\"recyclobuddy_logo.jpg\"\n\n# Create your views here.\n\n#See pydanny.com/core-concepts-django-modelforms.html, but note several errors in the example code.\ndef index(request):\n message = \"\"\n if request.method == \"POST\":\n form = LookupForm(request.POST)\n\n if form.is_valid():\n\n #Capture fields from the form\n municipality=form.cleaned_data['municipality']\n address=form.cleaned_data['address']\n zip=form.cleaned_data['zip']\n\n #Parse address to put into standard form. Check for error\n error_code, parsed_address = parse_address(address, municipality)\n\n\n if error_code == 1:\n #If error_code==1, failed to find street identifier\n message = \"Didn't work. Please check municipality and omit apartment or suite from address.\"\n subscribe_URL=\"\"\n\n else:\n #Looks good, go ahead with the process\n\n #Look up zone information and return a zone dictionary giving zone and day for recycling, trash and yard waste\n try: \n zone_dict = get_zones(municipality, parsed_address, zip)\n server_failed=False\n except Exception:\n zone_dict = False\n server_failed=True\n raise\n\n if zone_dict:\n #Do lookup from schedules table and get message\n messages=get_initial_message(municipality, zone_dict)\n\n #Copy result into contacts table\n primary_key=insert_contact(municipality, parsed_address, zip, zone_dict)\n\n #Copy message into initial_messages table\n insert_initial_message(primary_key, messages)\n\n #Obfuscate the primary key\n masked_key = primary_key ^ OBFUSCATE\n\n #Create URL for subscription with primary key\n subscribe_URL=\"subscribe_\" + str(masked_key)\n\n #Redirect to subscription page\n return HttpResponseRedirect(subscribe_URL)\n\n else:\n #Failed. Could be the address wasn't good, (server_failed is False), or that server can't be reached. \n if server_failed==False:\n message = \"Didn't work. City couldn't locate that address. Missing N-S-E-W? \"\n else:\n message = \"Can't reach city server. Could be down or heavily loaded.\"\n \n subscribe_URL=\"\"\n form=LookupForm(request.POST)\n\n #Failed to get address in usable form or failed in zone look up\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'message': message,\n 'subscribe_URL': subscribe_URL,\n 'form' : form\n }\n\n return render (request, \"app/index.html\", c )\n\n else:\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'message': 'Hit a snag. Omit any apt or suite info, use 5-digit zip',\n 'form' : form\n }\n\n return render (request, \"app/index.html\", c ) \n else:\n form = LookupForm(initial={'municipality': 'LOWER_MERION'})\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'form': form,\n }\n\n return render (request, \"app/index.html\", c ) \n\n#@login_required\ndef subscribe(request, masked_key):\n #undo obfuscation\n primary_key = int(masked_key) ^ OBFUSCATE\n\n cat = get_object_or_404(Contacts, index_key=primary_key)\n\n #If this request has already been submitted, show the acknowledge page to avoid exposing private data.\n if cat.request==True:\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n #Send URL acknowledging request\n return render (request, \"app/acknowledge.html\", c )\n\n if request.method == \"POST\":\n #Using the instance here allows an update. docs.djangoproject.com/en/1.1/topics/forms/modelforms/#the-dave-method\n form = ContactForm(request.POST, instance=cat)\n\n if form.is_valid():\n model_instance = form.save()\n primary_key = model_instance.pk\n\n #Create message for confirmation\n confirmation_message=confirm_subscription(\n masked_key, \n model_instance.first_name,\n model_instance.last_name,\n model_instance.alert_day,\n model_instance.alert_time,\n model_instance.email_alert,\n model_instance.sms_alert,\n )\n \n #Update Contacts to reflect confirmation request\n c=Contacts.objects.get(pk=primary_key)\n c.request=True\n c.save()\n\n #Send mail message add try catch ???\n try:\n send_mail('Confirmation request', confirmation_message, 'recyclobuddy@recyclobuddy.com', [model_instance.email], fail_silently=False)\n except Exception:\n print (\"Failed to send confirmation email\\n\")\n\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n #Send URL acknowledging request\n return render (request, \"app/acknowledge.html\", c )\n\n else:\n \n form=ContactForm(instance = cat)\n \n #You are here either because ir's presenting the form before data is added, or the data isn't valid.\n \n #Get message information\n messages=select_initial_message(primary_key)\n\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'message_1': messages[0],\n 'message_2': messages[1],\n 'message_3': messages[2],\n 'form': form,\n }\n\n return render(request, \"app/subscription.html\", c ) \n\n#@login_required\ndef confirm(request, masked_key):\n #undo obfuscation\n primary_key = int(masked_key) ^ OBFUSCATE\n\n #Check if there is a valid object for this primary key\n cat = get_object_or_404(Contacts, index_key=primary_key)\n\n #Do validation: Does object exist and is request outstanding?\n c=Contacts.objects.get(pk=primary_key)\n if c and c.request==True:\n valid=True\n else:\n valid=False\n\n #If validation passed, then send confirmation message\n if valid == True:\n #Update Contacts to reflect subscription\n c.subscribe=True\n c.save()\n\n\n #Send okay message\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n return render(request, \"app/confirm.html\", c )\n\n else:\n #Failed, so send back to beginning.\n form=LookupForm()\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'form': form,\n }\n\n return render(request, \"app/try_again.html\", c ) \n\n#@login_required\ndef cancel(request):\n if request.method == \"POST\":\n form = CancelForm(request.POST)\n if form.is_valid():\n\n #Capture fields from the form\n email=form.cleaned_data['email']\n mobile=form.cleaned_data['mobile']\n\n #Check if email and mobile combination in database\n success=cancel_subscription(email, mobile)\n\n if success==True:\n #Cancellation worked.\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n return render(request, \"app/gone.html\", c )\n\n else:\n #Cancellation failed\n form = CancelForm()\n message = \"We couldn't find that combination of email address and mobile number. Please try again.\"\n\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'message': message,\n 'form': form,\n }\n\n return render(request, \"app/cancel.html\", c )\n\n else:\n form = CancelForm()\n message = \"Sorry to see you go! Please enter email and mobile number to discontinue alerts.\"\n \n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n 'message': message,\n 'form': form,\n }\n\n return render(request, \"app/cancel.html\", c )\n\ndef about(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/about.html\", c )\n\ndef faq(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/faq.html\", c )\n\ndef terms(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/terms.html\", c )\n\ndef trash_talk(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/trash-talk.html\", c )\n\ndef share(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/share.html\", c )\n\n#@login_required\ndef test(request):\n c = {\n 'app_template': 'app/basic_template.html',\n 'logo_image' : logo_image,\n }\n\n return render(request, \"app/repairs.html\", c )\n\n#@login_required\ndef root_index(request):\n return HttpResponseRedirect('./app/')\n\n@login_required\ndef success(request):\n return HttpResponseRedirect('../app/index')\n\n#@login_required\ndef logout_view(request):\n logout(request)\n return HttpResponseRedirect('../')\n","repo_name":"hhummel/recyclobuddy","sub_path":"app/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":11366,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"18233679661","text":"import sys\nfrom os.path import splitext\nfrom PIL import Image,ImageOps\n\nif len(sys.argv) == 3:\n extensions = [\".jpg\", \".jpeg\", \".png\"]\n extension1 = splitext(sys.argv[1])\n extension2 = splitext(sys.argv[2])\n if extension1[1] == extension2[1] and extension1[1] in extensions:\n try:\n BeforeImage = Image.open(sys.argv[1])\n except FileNotFoundError:\n sys.exit(\"File not found\")\n #Image.open\n shirt = Image.open(\"shirt.png\")\n size = shirt.size\n #ImageOps.fit\n muppet = ImageOps.fit(BeforeImage,size)\n #Image.paste\n muppet.paste(shirt, shirt)\n #Image.save\n muppet.save(sys.argv[2])\n\n elif extension1[1].lower() != extension2[1].lower():\n sys.exit(\"Input and output have different extensions\")\n else:\n sys.exit(\"Wrong Extension\")\n\nelif len(sys.argv) < 3:\n sys.exit(\"Too few command-line arguments\")\nelif len(sys.argv) > 3:\n sys.exit(\"Too many command-line arguments\")\nelse:\n sys.exit(\"Invalid input\")","repo_name":"Sowmika-Pulagam/Harvard_CS50","sub_path":"shirt/shirt.py","file_name":"shirt.py","file_ext":"py","file_size_in_byte":1162,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"36775368149","text":"import mysql.connector\nfrom dotenv import dotenv_values\n\nsecret = dotenv_values('.env')\n\ndef connectToDB():\n dbconfig = {\n \"host\": \"127.0.0.1\",\n \"port\": 3306,\n \"user\": secret[\"mysql_user\"], \n \"password\": secret[\"mysql_pwd\"],\n \"database\": \"taipei_trip\",\n }\n try:\n connectionPool = mysql.connector.pooling.MySQLConnectionPool(pool_name=\"taipei_trip\", pool_size=5, **dbconfig)\n print(\"建立connectionPool成功\")\n return connectionPool\n except Exception as ex:\n print(\"建立connectionPool失敗...\\n錯誤訊息:\",ex)\n return False \n\nconnectionPool = connectToDB()\n\ndef connectDB():\n if(connectionPool):\n return connectionPool.get_connection()\n else:\n return False\n\ndef insert(execute_str:str, execute_args: tuple):\n connection = connectDB()\n if(not connection):\n return False\n \n cursor = connection.cursor() \n try:\n print(\"開始執行insertData\")\n cursor.execute(execute_str, execute_args)\n connection.commit()\n except Exception as ex:\n print(f\"insert error msg = ${ex}\")\n return False\n finally:\n cursor.close()\n connection.close()\n print(\"insert Data 成功!!\")\n return True\n\ndef find(query_str, query_args=None):\n connection = connectDB()\n if(not connection):\n return False\n cursor = connection.cursor(dictionary=True)\n try:\n cursor.execute(query_str, query_args)\n result = cursor.fetchall()\n except Exception as ex:\n print(f\"memberInfo查詢失敗,錯誤訊息為\\n${ex}\")\n return False\n finally:\n cursor.close()\n connection.close()\n print(f\"dbConnector result = ${result}\")\n if(result):\n return result[0]\n else:\n return False\n \ndef delete(execute_str: str, execute_args: tuple):\n connection = connectDB()\n if(not connection):\n return False\n cursor = connection.cursor()\n try:\n cursor.execute(execute_str, execute_args)\n connection.commit()\n print(\"刪除成功!\")\n except Exception as e:\n print(\"刪除失敗,錯誤訊息\\n\",e)\n return False\n finally:\n cursor.close()\n connection.close()\n return True\n \ndef update(execute_str: str, execute_args: tuple):\n connection = connectDB()\n if(not connection):\n return False\n cursor = connection.cursor()\n try:\n cursor.execute(execute_str, execute_args)\n except Exception as e:\n print(\"更新失敗,錯誤訊息\\n\",e)\n return False\n finally:\n change_row = cursor.rowcount\n if(change_row == 0):\n print(\"更新失敗,尚未找到符合條件的目標。\")\n return False\n connection.commit()\n print(\"更新成功!\")\n cursor.close()\n connection.close()\n return True\n \nif __name__ == '__main__':\n pass\n\n\n","repo_name":"joeyliao127/Taipei_day_trip","sub_path":"packages/dbConnector.py","file_name":"dbConnector.py","file_ext":"py","file_size_in_byte":2954,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"40058425","text":"import collections\nfrom collections.abc import Iterable\n\nfrom django.utils.module_loading import import_string\n\nfrom .models import EventType, EventSideEffectLog\n\n\nclass EventTypeRegister(collections.UserDict):\n \"\"\"A dictionary mapping fully qualified event type values with the event type.\"\"\"\n\n def __init__(self, event_type_classes):\n super().__init__()\n for event_type_class in event_type_classes:\n event_type_enum = import_string(event_type_class)\n for event_type in event_type_enum:\n self.data[event_type.fully_qualified_value] = event_type\n\n\nRegisteredSideEffect = collections.namedtuple(\n \"RegisteredSideEffect\", field_names=(\"callable\", \"condition\")\n)\n\n\nclass EventHandlerRegister:\n \"\"\"Stores event handlers.\"\"\"\n\n def __init__(self):\n self.handlers = collections.defaultdict(lambda: [])\n self.side_effects = collections.defaultdict(lambda: [])\n\n def register(self, *, event_type):\n def decorator(f):\n if isinstance(event_type, EventType):\n self.handlers[event_type].append(f)\n elif isinstance(event_type, Iterable):\n for type in event_type:\n self.handlers[type].append(f)\n else:\n raise TypeError(f\"Unknown event type: {event_type}\")\n\n return f\n\n return decorator\n\n def register_side_effect(self, callable, *, condition=None):\n def decorator(f):\n self.side_effects[f].append(RegisteredSideEffect(callable, condition))\n return f\n\n return decorator\n\n def _run_event_function(self, log, function, *args, **kwargs):\n result = None\n try:\n result = function(*args, **kwargs)\n log.status = log.Status.SUCCESS\n log.message = str(result)\n except Exception as error:\n log.status = log.Status.FAILED\n log.message = repr(error)\n finally:\n log.save()\n\n return result\n\n def handle(self, event, skip_side_effects=False):\n for handler in self.handlers[event.type]:\n handler_log = event.handler_logs.create_from_function(function=handler)\n result = self._run_event_function(handler_log, handler, event)\n\n if skip_side_effects or handler_log.failed:\n continue\n\n for registered_side_effect in self.side_effects[handler]:\n condition_class = registered_side_effect.condition\n should_run = (\n condition_class().has_condition(event) if condition_class else True\n )\n status = (\n EventSideEffectLog.Status.PROCESSING\n if should_run\n else EventSideEffectLog.Status.SKIPPED\n )\n side_effect_log = handler_log.side_effect_logs.create_from_function(\n function=registered_side_effect.callable, status=status\n )\n\n if should_run:\n self._run_event_function(\n side_effect_log, registered_side_effect.callable, result\n )\n","repo_name":"vikashtank/django-event-sourcing","sub_path":"django_event_sourcing/registers.py","file_name":"registers.py","file_ext":"py","file_size_in_byte":3185,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"60"}
+{"seq_id":"36603342642","text":"#!/usr/bin/env python3\n# pylint: disable=missing-docstring,import-error,invalid-name\n\nimport ldap\nimport ldap3\nimport ssl\n\n\ndef python_ldap():\n print(\"python-ldap\")\n l = ldap.initialize(\"ldaps://ldap.su.se\") # noqa: E741\n\n l.sasl_gssapi_bind_s()\n print(\"whoami: {}\".format(l.whoami_s()))\n\n result = l.search_s(base=\"\",\n scope=ldap.SCOPE_SUBTREE,\n filterstr=\"(uid=simlu)\",\n attrlist=[\"eduPersonAffiliation\", \"displayName\"])\n\n for _dn, entry in result:\n for attribute in entry:\n print(\"attribute: {} is: {!r}\".format(\n attribute,\n # We need to decode the bytes to UTF-8 apparently\n [a.decode(\"utf-8\") for a in entry[attribute]]))\n\n\ndef ldaptre():\n print(\"ldap3\")\n conn = ldap3.Connection(\n server=ldap3.Server('ldap.su.se',\n use_ssl=True,\n tls=ldap3.Tls(\n validate=ssl.CERT_REQUIRED,\n )\n ),\n auto_bind=True,\n authentication=ldap3.SASL,\n sasl_mechanism=ldap3.GSSAPI,\n )\n print(\"whoami: {}\".format(conn.extend.standard.who_am_i()))\n conn.search(search_base='',\n search_scope=ldap3.SUBTREE,\n search_filter='(uid=simlu)',\n attributes=['eduPersonAffiliation', 'displayName'])\n for entry in conn.entries:\n for attribute in entry.entry_attributes:\n print(\"attribute: {} is: {!r}\".format(attribute,\n entry[attribute].values))\n\n\ndef main():\n python_ldap()\n print(\"\")\n ldaptre()\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"simmel/affilaffe","sub_path":"affilaffe/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1688,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"39845013158","text":"class Node:\r\n def __init__(self, symbol, frequency, left=None, right=None) -> None:\r\n self.symbol = symbol\r\n self.frequency = frequency\r\n self.left = left\r\n self.right = right\r\n self.code = \"\"\r\n\r\nclass Haff:\r\n def create_nodes(self, frequency: dict) -> list[Node]:\r\n nodes = []\r\n for symbol in frequency:\r\n nodes.append(Node(symbol, frequency[symbol]))\r\n return nodes\r\n\r\n def calculate_frequency(self, message: str) -> dict:\r\n frequency = {}\r\n for i in list(set(message)):\r\n frequency[i] = message.count(i)\r\n return frequency\r\n\r\n def calculate_code(self, node: Node, code=\"\", codes={}) -> str:\r\n\r\n code = code + str(node.code)\r\n\r\n if node.left:\r\n self.calculate_code(node.left, code, codes)\r\n if node.right:\r\n self.calculate_code(node.right, code, codes)\r\n\r\n if not node.left and not node.right:\r\n codes[node.symbol] = code\r\n\r\n return codes\r\n\r\n def replace_encoded(self, message: str, code_map: dict) -> dict:\r\n for i in code_map:\r\n message = message.replace(i, code_map[i])\r\n return message\r\n\r\n def decode(self, message: str, huffman_tree: Node):\r\n tree_head = huffman_tree\r\n decoded = \"\"\r\n\r\n for i in message:\r\n if i == '1':\r\n huffman_tree = huffman_tree.right \r\n if i == '0':\r\n huffman_tree = huffman_tree.left\r\n try:\r\n if huffman_tree.left.symbol == None and huffman_tree.right.symbol == None:\r\n pass\r\n except AttributeError:\r\n decoded += huffman_tree.symbol\r\n huffman_tree = tree_head\r\n \r\n return decoded\r\n\r\n def encode(self, message: str) -> str:\r\n nodes = self.create_nodes(self.calculate_frequency(message))\r\n\r\n while len(nodes) > 1:\r\n nodes = sorted(nodes, key=lambda x: x.frequency)\r\n\r\n right_node: Node = nodes[0]\r\n left_node: Node = nodes[1]\r\n\r\n right_node.code = 1\r\n left_node.code = 0\r\n\r\n new_node = Node(left_node.symbol + right_node.symbol, left_node.frequency + right_node.frequency, left_node, right_node)\r\n\r\n nodes.remove(left_node)\r\n nodes.remove(right_node)\r\n nodes.append(new_node)\r\n \r\n self.haffman_tree = nodes[0]\r\n\r\n return self.replace_encoded(message, self.calculate_code(self.haffman_tree))\r\n\r\n def get_haffman_tree(self) -> Node:\r\n return self.haffman_tree\r\n\r\nhaff = Haff()\r\n\r\nwith open(\"sixthMessage.txt\", \"r\", encoding=\"utf-8\") as file:\r\n variable = file.read()\r\n\r\nprint(haff.encode(variable))\r\nprint(haff.decode(haff.encode(variable), haff.get_haffman_tree()))","repo_name":"THATISMYSTUDYACCOUNTIAMNOTSTREWDRAGON/university","sub_path":"3_theoretical_lab/sixthTask.py","file_name":"sixthTask.py","file_ext":"py","file_size_in_byte":2831,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"14377844506","text":"#!/usr/bin/python3\n\nfrom networkx import write_edgelist, write_weighted_edgelist\nimport os\nfrom utils import *\nfrom optparse import OptionParser\nfrom multiprocessing import Process\nfrom time import time\nfrom datetime import datetime\n\ndef main(n_start, n_count=1, n_inc=1, c_in_start=10, c_in_count=1, c_in_inc=1, c_out_start=5, c_out_count=1, c_out_inc=1, comm_count = 2, DC=False, i=0):\n bp_uncertain = 'src/bp'\n\n edge_frac = 1.\n nonedge_mult = 5.\n b = 2\n trials = 2\n\n os.makedirs('out', exist_ok=True)\n os.makedirs('data', exist_ok=True)\n\n for n in custom_range(n_start, n_count, n_inc):\n for c_in in custom_range(c_in_start, c_in_count, c_in_inc):\n for c_out in custom_range(c_out_start, c_out_count, c_out_inc):\n original_net = 'data/original_net-%d-%f-%f-%f-%f-%f-%d.edges'%(n,c_in,c_out,b,edge_frac,nonedge_mult, i)\n uncertain_net = 'data/noisy_net-%d-%f-%f-%f-%f-%f-%d.edges'%(n,c_in,c_out,b,edge_frac,nonedge_mult, i)\n uncertain_comms = 'out/uncertain_comms-%d-%f-%f-%f-%f-%f-%d.out'%(n,c_in,c_out,b,edge_frac,nonedge_mult, i)\n \n print(\"making and fuzzing network\")\n G_orig = make_net(c_in, c_out, n)\n write_edgelist(G_orig, original_net)\n G, _ = fuzz_network(G_orig, 1, b, edge_frac, nonedge_mult)\n write_weighted_edgelist(G, uncertain_net)\n \n start1 = time()\n print(\"running belief propagation\")\n os.system('%s -i %s -o %s -c %d -l %d -n %d' % (bp_uncertain, uncertain_net, uncertain_comms, comm_count, 3, trials))\n end1 = time()\n\n with open('out/results.txt', 'a+') as out_file:\n out_file.write(\"%d %f %f\\t%f %f %f\\t %f %f\\t %s %d\\n\" %(n,\n c_in, c_out,\n b,edge_frac,nonedge_mult,\n evaluate(uncertain_comms, n), end1-start1,\n str(datetime.now()), i))\n\n\nif __name__ == '__main__':\n # parse command line options\n parser = OptionParser()\n parser.add_option('-n', type=int, dest = 'n', help='number of nodes in network', default=2000)\n parser.add_option('--c_in', type=float, dest = 'c_in', help='average within-community degree', default=40)\n parser.add_option('--c_out', type=float, dest = 'c_out', help='average between-community degree', default=10)\n parser.add_option('-k', type=int, dest = 'comm_count', help='number of communities', default=2)\n parser.add_option('--iters', type=int, help='number of instances to run, multithreaded', default = 1)\n\n (options, _) = parser.parse_args()\n\n ps = []\n # run multiple threads\n for i in range(options.iters):\n p = Process(target=main, args = (options.n, 1, 1, options.c_in, 1, 1, options.c_out, 1, 1, options.comm_count, False, i))\n ps.append(p)\n p.start()\n for p in ps:\n p.join()\n","repo_name":"nitramsivart/uncertain-networks","sub_path":"run_synthetic.py","file_name":"run_synthetic.py","file_ext":"py","file_size_in_byte":2995,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"60"}
+{"seq_id":"1224960390","text":"import sys\r\n\r\nsys.setrecursionlimit(10**6)\r\ntable = dict()\r\n\r\n\r\ndef choose(n, r, mod=None): # no mod, or mod ≠ prime\r\n if r > n or r < 0:\r\n return 0\r\n if r == 0:\r\n return 1\r\n if (n, r) in table:\r\n return table[(n, r)]\r\n table[(n, r)] = choose(n - 1, r) + choose(n - 1, r - 1)\r\n return table[(n, r)]\r\n\r\n\r\nn, *a = map(int, sys.stdin.read().split())\r\na += [0]\r\n\r\n\r\ndef main():\r\n prev = 1001001001\r\n cnt = 0\r\n res = 0\r\n for x in a:\r\n if x <= prev:\r\n res += choose(cnt, 2) + cnt\r\n cnt = 1\r\n else:\r\n cnt += 1\r\n prev = x\r\n print(res)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","repo_name":"kagemeka/atcoder-submissions","sub_path":"jp.atcoder/abc038/abc038_c/11764080.py","file_name":"11764080.py","file_ext":"py","file_size_in_byte":680,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"8541421352","text":"# -*- coding: utf-8 -*-\n\n# This sample demonstrates handling intents from an Alexa skill using the Alexa Skills Kit SDK for Python.\n# Please visit https://alexa.design/cookbook for additional examples on implementing slots, dialog management,\n# session persistence, api calls, and more.\n# This sample is built using the handler classes approach in skill builder.\nimport logging\nimport ask_sdk_core.utils as ask_utils\n\nfrom ask_sdk_core.skill_builder import SkillBuilder\nfrom ask_sdk_core.dispatch_components import AbstractRequestHandler\nfrom ask_sdk_core.dispatch_components import AbstractExceptionHandler\nfrom ask_sdk_core.handler_input import HandlerInput\nfrom ask_sdk_model.intent import Intent\nfrom ask_sdk_model.dialog import (ElicitSlotDirective, DelegateDirective)\nfrom ask_sdk_model.dialog_state import DialogState\nfrom ask_sdk_model.slu.entityresolution.status_code import StatusCode\n\nfrom ask_sdk_model import Response\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\n\n### api ###\nimport tmdbv3api\nfrom tmdbv3api import TMDb\n\nfrom tmdbv3api import Movie\nfrom tmdbv3api import Discover\nfrom tmdbv3api import Person\nfrom tmdbv3api import Search\nfrom tmdbv3api import Genre\n\ntmdb = TMDb()\ntmdb.api_key = '93a8c2d922b414294a124ef8dc9c2428'\n\n### firebase ###\nimport firebase_admin \nfrom firebase_admin import credentials, firestore\n\ncred = credentials.Certificate(\"firebase.json\")\nfirebase_admin.initialize_app(cred)\n\nfirestore_db = firestore.client()\n\n### canonical slot value constants ###\nFEEDBACK_POSITIVE = 'good'\nFEEDBACK_NEGATIVE = 'bad'\n\nRECOMMENDATION_ACCEPTED = \"Okay\"\nRECOMMENDATION_REJECTED = \"Something else\"\n\n### response templates ###\nSENTENCE_FEEDBACK_BEFORE_RECOMMENDATION = \"I'm happy to take a new recommendation request, but please give me feedback on my last recommendation first, which was the film {}. \"\nSENTENCE_NO_RECOMMENDATION_FOUND = \"I'm sorry. I was not able to find a {} recommendation available on your platforms based on your input. \"\nSENTENCE_MOVIE_NOT_FOUND = \"Hmm, I don't know the film {}. \"\nSENTENCE_ACTOR_NOT_FOUND = \"Sorry, I don't know the actor or actress {}. \"\nSENTENCE_GENRE_NOT_FOUND = \"Hmm, I don't recognize {} as a genre. \"\nSENTENCE_DID_NOT_UNDERSTAND = \"Sorry, I didn't get that. \"\nSENTENCE_RECOMMENDATION_ACCEPTED = \"Great! Enjoy the film! \"\nSENTENCE_ANOTHER_RECOMMENDATION = \"Okay. \"\nSENTENCE_CANCELED = \"Okay, maybe another time. \"\nSENTENCE_FEEDBACK_ACCEPTED = \"Alright, I will try to work your feedback into my next recommendations. \"\nSENTENCE_BACK_TO_RECOMMENDATION = \"Back to your recommendation request... \"\nSENTENCE_FEEDBACK_WAS_POSITIVE = \"Great to hear! \"\nSENTENCE_FEEDBACK_WAS_NEGATIVE = \"I'm sorry to hear that. \"\n\nPROMPT_TRY_AGAIN = \"Why don't you try again with a different request? If you need help, just say: Help. \"\nPROMPT_MOVIE_NOT_FOUND = \"Please tell me the movie's official English title. \"\nPROMPT_ACTOR_NOT_FOUND = \"Please use their commonly known name. \"\nPROMPT_GENRE_NOT_FOUND = \"Try a more common synonym or a more widely known genre. \"\nPROMPT_RECOMMENDATION_CONFIRMATION = \"You can say Okay to accept this recommendation or request another one by saying: Another one. To cancel, just say Stop. \"\nPROMPT_ANOTHER_RECOMMENDATION = \"How about this one? \"\nPROMPT_FEEDBACK_ASPECTS = \"What specifically {} you like about {}? That could be the acting, the story, or the setting of the movie. \"\n\n### other templates ###\nDELDIR_FEEDBACK = DelegateDirective(Intent(\n name=\"feedbackIntent\",\n slots={\n \"feedbackGeneral\": {\n \"name\": \"feedbackGeneral\"\n },\n \"feedbackAspects\": {\n \"name\": \"feedbackAspects\"\n }\n }\n))\n\ndef makeESDir(slotname, intent=None):\n if intent:\n return (ElicitSlotDirective(\n slot_to_elicit = slotname,\n updated_intent = intent\n ))\n else:\n return (ElicitSlotDirective(\n slot_to_elicit = slotname\n ))\n\n\n### own functions ###\nfrom sentimentAnalysis import *\nfrom recommender_functions import *\nfrom user_firebase_functions import *\nfrom movie_api_functions import *\n\nclass LaunchRequestHandler(AbstractRequestHandler):\n \"\"\"Handler for Skill Launch.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_request_type(\"LaunchRequest\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n\n responseBuilder = handler_input.response_builder\n userID = str(handler_input.request_envelope.context.system.user.user_id)\n speak_output = \"Welcome to Movie Tips! \"\n \n if check_user_exists_by_id(userID):\n speak_output = \"Welcome back {}! \".format(get_username_by_id(userID))\n if is_last_watched_movie_rated_by_id(userID):\n speak_output += \"Do you want me to recommend a movie? If so, just say: Recommend something.\"\n responseBuilder.speak(speak_output).ask(speak_output)\n else:\n speak_output += \"I see you haven't rated my last recommendation yet, which was the film {}. Let's do that now! \".format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n else:\n speak_output += \"I see you haven't set up a profile yet. Let's do that now! \"\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"setupIntent\" #weirdly enough, we don't have to pass the slots here and it works anyway... as opposed to the feedbackIntent which has Alexa completely checking out after the first prompt unless the slots are given. What even is this madness\n )))\n \n return responseBuilder.response\n\n\n\nclass RecommendationByMovieIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Recommendation by Movie Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"recommendationIntent_byMovie\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n userID = handler_input.request_envelope.context.system.user.user_id\n slot_movie = intent.slots[\"movieTitle\"]\n slot_confirmation = intent.slots[\"recommendationOK\"]\n \n if request.dialog_state == DialogState.STARTED or slot_movie.value is None:\n if is_last_watched_movie_rated_by_id(userID):\n \n #resolve \"my favorite movie\" if necessary\n resolution = slot_movie.resolutions.resolutions_per_authority[0]\n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotID = resolution.values[0].value.id\n if resolvedSlotID == \"FAV\":\n slot_movie.value = get_favourite_movie_by_id(userID)\n \n movieID = getMovieID(slot_movie.value)\n movieName = getMovieName(movieID)\n \n if movieID == 0:\n speak_output = SENTENCE_MOVIE_NOT_FOUND.format(slot_movie.value)\n prompt_output = PROMPT_MOVIE_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"movieTitle\"))\n else:\n slot_movie.value = movieName\n \n rec = recommendationSentenceFromMovieInput(userID, movieName)\n if rec['success']:\n speak_output = rec['sentence']\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else:\n speak_output = SENTENCE_FEEDBACK_BEFORE_RECOMMENDATION.format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n \n \n elif request.dialog_state == DialogState.IN_PROGRESS: #confirmation given for recommendation\n resolution = slot_confirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_confirmation.value = inputVal\n if inputVal == RECOMMENDATION_ACCEPTED:\n responseBuilder.add_directive(DelegateDirective(intent))\n elif inputVal == RECOMMENDATION_REJECTED:\n rec = recommendationSentenceFromMovieInput(userID, slot_movie.value)\n if rec['success']:\n speak_output = SENTENCE_ANOTHER_RECOMMENDATION\n prompt_output = PROMPT_ANOTHER_RECOMMENDATION+rec['sentence']+\" \"+PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else: #We're told to cancel/stop\n speak_output = SENTENCE_CANCELED\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"AMAZON.CancelIntent\"\n )))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n \n else: #request.dialog_state == DialogState.COMPLETED; we have a green light on the recommendation at this point\n acceptRecommendation(userID)\n speak_output = SENTENCE_RECOMMENDATION_ACCEPTED\n responseBuilder.speak(speak_output)\n \n return responseBuilder.response\n\n\nclass RecommendationByActorIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Recommendation by Actor Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"recommendationIntent_byActor\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n userID = handler_input.request_envelope.context.system.user.user_id\n slot_actor = intent.slots[\"actor\"]\n slot_confirmation = intent.slots[\"recommendationOK\"]\n \n if request.dialog_state == DialogState.STARTED or slot_actor.value is None:\n if is_last_watched_movie_rated_by_id(userID):\n \n #resolve \"my favorite actor\" if necessary\n resolution = slot_actor.resolutions.resolutions_per_authority[0]\n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotID = resolution.values[0].value.id\n if resolvedSlotID == \"FAV\":\n slot_actor.value = get_favourite_actress_by_id(userID)\n \n actorID = getActressId(slot_actor.value)\n actorName = getActressName(actorID)\n \n if actorID == 0:\n speak_output = SENTENCE_ACTOR_NOT_FOUND.format(slot_actor.value)\n prompt_output = PROMPT_ACTOR_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"actor\"))\n else:\n slot_actor.value = actorName\n\n rec = recommendationSentenceFromActressInput(userID, actorName)\n if rec['success']:\n speak_output = rec['sentence']\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else:\n speak_output = SENTENCE_FEEDBACK_BEFORE_RECOMMENDATION.format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n \n \n elif request.dialog_state == DialogState.IN_PROGRESS: #confirmation given for recommendation\n resolution = slot_confirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_confirmation.value = inputVal\n if inputVal == RECOMMENDATION_ACCEPTED:\n responseBuilder.add_directive(DelegateDirective(intent))\n elif inputVal == RECOMMENDATION_REJECTED:\n rec = recommendationSentenceFromActressInput(userID, slot_actor.value)\n if rec['success']:\n speak_output = SENTENCE_ANOTHER_RECOMMENDATION\n prompt_output = PROMPT_ANOTHER_RECOMMENDATION+rec['sentence']+\" \"+PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else: #We're told to cancel/stop\n speak_output = SENTENCE_CANCELED\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"AMAZON.CancelIntent\"\n )))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n \n else: #request.dialog_state == DialogState.COMPLETED; we have a green light on the recommendation at this point\n acceptRecommendation(userID)\n speak_output = SENTENCE_RECOMMENDATION_ACCEPTED\n responseBuilder.speak(speak_output)\n \n return responseBuilder.response\n\n\nclass RecommendationByGenreIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Recommendation by Genre Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"recommendationIntent_byGenre\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n userID = handler_input.request_envelope.context.system.user.user_id\n slot_genre = intent.slots[\"genre\"]\n slot_confirmation = intent.slots[\"recommendationOK\"]\n \n if request.dialog_state == DialogState.STARTED or slot_genre.value is None:\n if is_last_watched_movie_rated_by_id(userID):\n \n #resolve \"my favorite genre\" if necessary\n resolution = slot_genre.resolutions.resolutions_per_authority[0]\n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotID = resolution.values[0].value.id\n resolvedSlotName = resolution.values[0].value.name\n if resolvedSlotID == \"FAV\":\n slot_genre.value = get_liked_genre_by_id(userID)\n else:\n slot_genre.value = resolvedSlotName\n \n rec = recommendationSentenceFromGenreInput(userID, slot_genre.value)\n if rec['success']:\n speak_output = rec['sentence']\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else:\n speak_output = SENTENCE_GENRE_NOT_FOUND.format(slot_genre.value)\n prompt_output = PROMPT_GENRE_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"genre\"))\n \n else:\n speak_output = SENTENCE_FEEDBACK_BEFORE_RECOMMENDATION.format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n \n \n elif request.dialog_state == DialogState.IN_PROGRESS: #confirmation given for recommendation\n resolution = slot_confirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_confirmation.value = inputVal\n if inputVal == RECOMMENDATION_ACCEPTED:\n responseBuilder.add_directive(DelegateDirective(intent))\n elif inputVal == RECOMMENDATION_REJECTED:\n rec = recommendationSentenceFromGenreInput(userID, slot_genre.value)\n if rec['success']:\n speak_output = SENTENCE_ANOTHER_RECOMMENDATION\n prompt_output = PROMPT_ANOTHER_RECOMMENDATION+rec['sentence']+\" \"+PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else: #We're told to cancel/stop\n speak_output = SENTENCE_CANCELED\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"AMAZON.CancelIntent\"\n )))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n \n else: #request.dialog_state == DialogState.COMPLETED; we have a green light on the recommendation at this point\n acceptRecommendation(userID)\n speak_output = SENTENCE_RECOMMENDATION_ACCEPTED\n responseBuilder.speak(speak_output)\n \n return responseBuilder.response\n\n\n\nclass RecommendationRewatchIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Recommendation Rewatch Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"recommendationIntent_rewatch\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n userID = handler_input.request_envelope.context.system.user.user_id\n slot_confirmation = intent.slots[\"recommendationOK\"]\n \n if request.dialog_state == DialogState.STARTED: #\"recommend a rewatch\"/...\n if is_last_watched_movie_rated_by_id(userID):\n prevRecommendations = get_recommended_movies_by_id(userID)\n if prevRecommendations and len(prevRecommendations[0]): #check if we have recommendations at all\n rec = recommendationSentenceFromAgain(userID)\n if rec['success']:\n speak_output = rec['sentence']\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"rewatch\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n else:\n speak_output = \"Sorry, I don't have any previously accepted recommendations for you in my database that I could recommend again. Try using this feature again once you have accepted some new recommendations of mine. \"\n prompt_output = \"In the meantime... \"+PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else:\n speak_output = SENTENCE_FEEDBACK_BEFORE_RECOMMENDATION.format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n \n elif request.dialog_state == DialogState.IN_PROGRESS: #confirmation given for recommendation\n resolution = slot_confirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_confirmation.value = inputVal\n if inputVal == RECOMMENDATION_ACCEPTED:\n responseBuilder.add_directive(DelegateDirective(intent))\n elif inputVal == RECOMMENDATION_REJECTED:\n rec = recommendationSentenceFromAgain(userID)\n if rec['success']:\n speak_output = SENTENCE_ANOTHER_RECOMMENDATION\n prompt_output = PROMPT_ANOTHER_RECOMMENDATION+rec['sentence']+\" \"+PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new rewatch\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else: #We're told to cancel/stop\n speak_output = SENTENCE_CANCELED\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"AMAZON.CancelIntent\"\n )))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n \n else: #request.dialog_state == DialogState.COMPLETED; we have a green light on the recommendation at this point\n acceptRecommendation(userID)\n speak_output = SENTENCE_RECOMMENDATION_ACCEPTED\n responseBuilder.speak(speak_output)\n \n \n return responseBuilder.response\n\nclass RecommendationIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Recommendation Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"recommendationIntent\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n userID = handler_input.request_envelope.context.system.user.user_id\n slot_confirmation = intent.slots[\"recommendationOK\"]\n \n if request.dialog_state == DialogState.STARTED: #\"i want to watch something\"/...\n if is_last_watched_movie_rated_by_id(userID):\n #####\n rec = generalRecommendation(userID)\n if rec['success']:\n speak_output = rec['sentence']\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else:\n speak_output = SENTENCE_FEEDBACK_BEFORE_RECOMMENDATION.format(get_last_watched_movie_by_id(userID))\n responseBuilder.speak(speak_output).add_directive(DELDIR_FEEDBACK)\n \n elif request.dialog_state == DialogState.IN_PROGRESS: #confirmation given for recommendation\n resolution = slot_confirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_confirmation.value = inputVal\n if inputVal == RECOMMENDATION_ACCEPTED:\n responseBuilder.add_directive(DelegateDirective(intent))\n elif inputVal == RECOMMENDATION_REJECTED:\n rec = generalRecommendation(userID)\n if rec['success']:\n speak_output = SENTENCE_ANOTHER_RECOMMENDATION\n prompt_output = PROMPT_ANOTHER_RECOMMENDATION+rec['sentence']+\" \"+PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n else:\n speak_output = SENTENCE_NO_RECOMMENDATION_FOUND.format(\"new\")\n prompt_output = PROMPT_TRY_AGAIN\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output)\n \n else: #We're told to cancel/stop\n speak_output = SENTENCE_CANCELED\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(Intent(\n name=\"AMAZON.CancelIntent\"\n )))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_RECOMMENDATION_CONFIRMATION\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"recommendationOK\"))\n \n else: #request.dialog_state == DialogState.COMPLETED; we have a green light on the recommendation at this point\n acceptRecommendation(userID)\n speak_output = SENTENCE_RECOMMENDATION_ACCEPTED\n responseBuilder.speak(speak_output)\n \n \n return responseBuilder.response\n\n\nclass SetupIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Setup Profile Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"setupIntent\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n #at least for now, we only consider streaming options in Germany and ignore the language version of the films on a given streamer\n \n slot_name = intent.slots[\"name\"]\n slot_streamer = intent.slots[\"streamer\"]\n slot_favMovie = intent.slots[\"favMovie\"]\n slot_favActor = intent.slots[\"favActor\"]\n slot_favGenre = intent.slots[\"favGenre\"]\n slot_dislikedGenre = intent.slots[\"dislikedGenre\"]\n slot_finalConfirmation = intent.slots[\"finalConfirmation\"]\n \n #check where we are in the dialog:\n # 0 = no info given yet, need to get the name first\n # 1 = name is given and to be processed, nothing else yet. Get the streamers.\n # 2 = name (processed) and streamers (to be processed) are given, get the fav movie\n # 3 = name, streamers (processed) and fav movie (to be processed) are given, get the fav actors\n # 4 = name, streamers, fav movie (processed) and fav actors (to be processed) are given, get the liked genres\n # 5 = name, streamers, fav movie, fav actors (processed) and liked genres (to be processed) are given, get the disliked genres\n # 6 = name, streamers, fav movie, fav actors, liked genres (processed) and disliked genres (to be processed) are given, process the disliked genres\n # 7 = we have all we need, so we can give some confirmation and move on.\n \n setupPhase = 0\n if request.dialog_state == DialogState.STARTED:\n if slot_name.value is None:\n setupPhase = 0\n else:\n setupPhase = 1\n elif request.dialog_state == DialogState.IN_PROGRESS:\n if slot_streamer.value is None and slot_streamer.slot_value is None:\n setupPhase = 1\n elif slot_favMovie.value is None:\n setupPhase = 2\n elif slot_favActor.value is None:\n setupPhase = 3\n elif slot_favGenre.value is None:\n setupPhase = 4\n elif slot_dislikedGenre.value is None:\n setupPhase = 5\n elif slot_finalConfirmation.value is None:\n setupPhase = 6\n else:\n setupPhase = 7\n else: #request.dialog_state == DialogState.COMPLETED\n setupPhase = 8\n \n #respond per phase:\n if setupPhase == 0:\n responseBuilder.add_directive(DelegateDirective())\n \n elif setupPhase == 1:\n inputVal = slot_name.value\n speak_output = \"Hi {}! \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n \n elif setupPhase == 2:\n #try to resolve slot input\n allInputs = []\n resolvedInputs = []\n if \"resolvedStreamers\" in sessionAttribs and sessionAttribs[\"resolvedStreamers\"] is not None:\n resolvedInputs = list(sessionAttribs[\"resolvedStreamers\"])\n sessionAttribs[\"resolvedStreamers\"] = None\n currentUnresolved = {}\n \n if slot_streamer.slot_value.object_type == \"List\":\n for v in slot_streamer.slot_value.values:\n allInputs.append(v)\n else:\n allInputs.append(slot_streamer.slot_value)\n \n for i in allInputs:\n resolution = i.resolutions.resolutions_per_authority[0]\n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n resolvedSlotID = resolution.values[0].value.id\n if resolvedSlotID in (\"ATV-X\", \"AMZ-X\", \"SKY-X\"):\n currentUnresolved = {\"id\": resolvedSlotID, \"value\": i.value}\n break\n else:\n resolvedInputs.append(resolvedSlotName)\n else:# if i.value not in (\"the\", \"and\", \"also\"): #don't ask for stuff that we don't need but Alexa can't (always) filter out by itself\n currentUnresolved = {\"id\": \"\", \"value\": i.value}\n break\n \n stringifiedResolved = resolvedInputs and ((\", \".join(resolvedInputs[:-1])+\" and \"+resolvedInputs[-1], resolvedInputs[0])[len(resolvedInputs) == 1])\n \n if currentUnresolved:\n #response part 1\n if resolvedInputs:\n speak_output = \"I understood {}. But \".format(stringifiedResolved)\n else:\n speak_output = \"Sorry, \"\n \n #response part 2\n if currentUnresolved[\"id\"] == \"ATV-X\":\n speak_output += \"I'm not sure about \"+currentUnresolved[\"value\"]+\". \"\n prompt_output = \"Do you mean Apple TV Plus or Apple iTunes, also known as Apple TV? \"\n elif currentUnresolved[\"id\"] == \"AMZ-X\":\n speak_output += \"I'm not sure about \"+currentUnresolved[\"value\"]+\". \"\n prompt_output = \"Do you mean Amazon Prime Video or Amazon Video? \" #Alexa cuts \"Amazon\" as the first word out of slot inputs, because why the hell not. Should be avoidable e.g. if we preface it with one of the defined carrier phrases like \"I have ...\". To be safe we defined synonyms for the option without the leading \"Amazon\".\n elif currentUnresolved[\"id\"] == \"SKY-X\":\n speak_output += \"I'm not sure about \"+currentUnresolved[\"value\"]+\". \"\n prompt_output = \"Do you mean Sky Ticket, Sky Go or the Sky Store? \"\n else:\n speak_output += \"I don't recognize {} as a streaming service. \".format(currentUnresolved[\"value\"])\n prompt_output = \"Please try a platform that is available in Germany. \"\n \n #give the response\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"streamer\"))\n #store the resolved values to pick up in the next go\n sessionAttribs[\"resolvedStreamers\"] = resolvedInputs\n else: #implies resolvedInputs is not empty\n inputVal = stringifiedResolved\n slot_streamer.value = \"|\".join(resolvedInputs) ###\n speak_output = \"Great! you shall only receive films available on {} as recommendations. \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n \n \n #resolution = slot_streamer.resolutions.resolutions_per_authority[0]\n #\n #if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n # resolvedSlotName = resolution.values[0].value.name\n # resolvedSlotID = resolution.values[0].value.id\n # if resolvedSlotID == \"ATV-X\":\n # prompt_output = \"Do you mean Apple TV Plus or Apple iTunes, also known as Apple TV? \"\n # responseBuilder.speak(prompt_output).ask(prompt_output).add_directive(esDir)\n # elif resolvedSlotID == \"AMZ-X\":\n # prompt_output = \"Do you mean Amazon Prime Video or Amazon Video? \" #Alexa cuts \"Amazon\" as the first word out of slot inputs, because why the hell not. Should be avoidable e.g. if we preface it with one of the defined carrier phrases like \"I have ...\". To be safe we defined synonyms for the option without the leading \"Amazon\".\n # responseBuilder.speak(prompt_output).ask(prompt_output).add_directive(esDir)\n # elif resolvedSlotID == \"SKY-X\":\n # prompt_output = \"Do you mean Sky Ticket, Sky Go or the Sky Store? \"\n # responseBuilder.speak(prompt_output).ask(prompt_output).add_directive(esDir)\n # else:\n # inputVal = resolvedSlotName\n # slot_streamer.value = inputVal\n # speak_output = \"Great! you shall only receive films available on {} as recommendations. \".format(inputVal)\n # responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n #else:\n # speak_output = \"Sorry, I don't recognize {} as a streaming service. \".format(slot_streamer.value)\n # prompt_output = \"Please try a platform that is available in Germany. \"\n # responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(esDir)\n \n elif setupPhase == 3:\n movieID = getMovieID(slot_favMovie.value)\n movieName = getMovieName(movieID)\n \n if movieID == 0:\n speak_output = SENTENCE_MOVIE_NOT_FOUND.format(slot_favMovie.value)\n prompt_output = PROMPT_MOVIE_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"favMovie\"))\n else:\n inputVal = movieName\n slot_favMovie.value = inputVal\n speak_output = \"Alright! I will try to recommend films similar to {}. \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n \n elif setupPhase == 4:\n actorID = getActressId(slot_favActor.value)\n actorName = getActressName(actorID)\n if actorID == 0:\n speak_output = SENTENCE_ACTOR_NOT_FOUND.format(slot_favActor.value)\n prompt_output = PROMPT_ACTOR_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"favActor\"))\n else:\n inputVal = actorName\n slot_favActor.value = inputVal\n speak_output = \"Okay, I will try to recommend films with {}. \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n \n elif setupPhase == 5:\n resolution = slot_favGenre.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n resolvedSlotID = resolution.values[0].value.id #this is already the API-ready genre ID\n inputVal = resolvedSlotName\n slot_favGenre.value = inputVal\n if inputVal == \"TV Movie\": #edge case: We don't want to say \"more TV Movie movies\". Yes, this is unnecessary overengineering. No, I don't care.\n speak_output = \"Cool! I will try to recommend more TV Movies to you. \"\n else:\n speak_output = \"Cool! I will try to recommend more {} movies to you. \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_GENRE_NOT_FOUND.format(slot_favGenre.value)\n prompt_output = PROMPT_GENRE_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"favGenre\"))\n \n elif setupPhase == 6:\n resolution = slot_dislikedGenre.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n resolvedSlotID = resolution.values[0].value.id #this is already the API-ready genre ID\n inputVal = resolvedSlotName\n slot_dislikedGenre.value = inputVal\n if inputVal == \"TV Movie\": #same edge case handling as above\n speak_output = \"Good to know, I will try to keep TV Movies out of your recommendations. \"\n else:\n speak_output = \"Good to know, I will try to keep {} films out of your recommendations. \".format(inputVal)\n responseBuilder.speak(speak_output).add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_GENRE_NOT_FOUND.format(slot_dislikedGenre.value)\n prompt_output = PROMPT_GENRE_NOT_FOUND\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"dislikedGenre\"))\n \n elif setupPhase == 7:\n resolution = slot_finalConfirmation.resolutions.resolutions_per_authority[0]\n \n if resolution.status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedSlotName = resolution.values[0].value.name\n inputVal = resolvedSlotName\n slot_finalConfirmation.value = inputVal\n responseBuilder.add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = \"Please say Okay to save your profile or Stop to cancel.\"\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"finalConfirmation\"))\n \n else: #setupPhase == 8\n #send everything to Firebase\n #session attribs don't work because somehow the last one will only be available one intent call too late, so we wouldn't have the dislikedGenre here if stored via session attribs.\n if slot_finalConfirmation.value == \"Okay\":\n userID = handler_input.request_envelope.context.system.user.user_id\n knownUser = check_user_exists_by_id(userID)\n streamers = ([slot_streamer.value], slot_streamer.value.split(\"|\"))[\"|\" in slot_streamer.value]\n if knownUser:\n u = User(slot_dislikedGenre.value, slot_favActor.value, slot_favMovie.value, userID, \"DE\", [get_last_watched_movie_by_id(userID), is_last_watched_movie_rated_by_id(userID)], slot_favGenre.value, slot_name.value, get_recommended_movies_by_id(userID), streamers, get_likings_by_id(userID), [\"\"])\n else:\n u = User(slot_dislikedGenre.value, slot_favActor.value, slot_favMovie.value, userID, \"DE\", [\"\", True], slot_favGenre.value, slot_name.value, [\"\"], streamers, {'movie': 0, 'genre': 0, 'acting': 0}, [\"\"])\n add_user(u)\n speak_output = \"You're all set up! \"\n if not knownUser:\n speak_output += \"To get an overview of what you can ask me, just say: Help.\"\n else:\n speak_output = \"Okay, I will not save your new data to your profile.\"\n responseBuilder.speak(speak_output)\n \n return responseBuilder.response\n\n\nclass FeedbackIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Feedback Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"feedbackIntent\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n request = handler_input.request_envelope.request\n intent = request.intent\n responseBuilder = handler_input.response_builder\n sessionAttribs = handler_input.attributes_manager.session_attributes\n \n #for debugging the mess that is the Python ASK:\n sessionAttribs[\"stringSlotPythonInput_feedbackAspects\"] = str(intent.slots[\"feedbackAspects\"]) # stringified version gives python-style attributes (slot_value instead of slotValue) and sets unset ones to None (those don't show up at all in the JSON version). hasattr seems to best map the stringified version.\n sessionAttribs[\"stringSlotPythonInput_feedbackGeneral\"] = str(intent.slots[\"feedbackGeneral\"])\n attribs = {\n 'slotValue': hasattr(intent.slots[\"feedbackAspects\"], 'slotValue'),\n 'slot_value': hasattr(intent.slots[\"feedbackAspects\"], 'slot_value')\n }\n if hasattr(intent.slots[\"feedbackAspects\"], 'slot_value') and intent.slots[\"feedbackAspects\"].slot_value is not None:\n attribs['.slot_value.resolutions'] = hasattr(intent.slots[\"feedbackAspects\"].slot_value, 'resolutions')\n if hasattr(intent.slots[\"feedbackAspects\"].slot_value, 'resolutions') and intent.slots[\"feedbackAspects\"].slot_value.resolutions is not None:\n attribs['.slot_value.resolutions.resolutionsPerAuthority'] = hasattr(intent.slots[\"feedbackAspects\"].slot_value.resolutions, 'resolutionsPerAuthority')\n attribs['.slot_value.resolutions.resolutions_per_authority'] = hasattr(intent.slots[\"feedbackAspects\"].slot_value.resolutions, 'resolutions_per_authority')\n if hasattr(intent.slots[\"feedbackAspects\"].slot_value.resolutions, 'resolutions_per_authority'):\n attribs['.slot_value.resolutions.resolutions_per_authority[0]'] = str(intent.slots[\"feedbackAspects\"].slot_value.resolutions.resolutions_per_authority[0])\n else:\n attribs['.slot_value.resolutions.resolutionsPerAuthority'] = \"failed, no .slot_value.resolutions\"\n attribs['.slot_value.resolutions.resolutions_per_authority'] = \"failed, no .slot_value.resolutions\"\n else:\n attribs['.slot_value.resolutions'] = \"failed, no .slot_value\"\n sessionAttribs[\"feedbackAspectsAttribs\"] = attribs\n \n feedbackGeneralRaw = intent.slots[\"feedbackGeneral\"].value\n \n #if no aspect is given yet, .slotValue isn't there and .value is None... SOMETIMES. Other random times, .slot_value is there. OR .slotValue. Depending on Alexa's mood of the day. I want to speak to the manager, Sir!\n #if one aspect is given, .value exists and .slotValue is:\n #{'type': 'Simple','value': 'writing', 'resolutions': {\n # 'resolutionsPerAuthority': [{'authority': '...', 'status': {'code': 'ER_SUCCESS_MATCH'}, 'values': [{'value': {'name': 'story', 'id': '...'}}]}]\n #}}\n #if multi are given, .value is None and .slotValue is:\n #{'type': 'List', 'values': [\n # {'type': 'Simple', 'value': 'story', 'resolutions': {\n # 'resolutionsPerAuthority': [{'authority': '...', 'status': {'code': 'ER_SUCCESS_MATCH'}, 'values': [{'value': {'name': 'story', 'id': '...'}}]}]\n # }},\n # {'type': 'Simple', 'value': 'score', 'resolutions': {\n # 'resolutionsPerAuthority': [{'authority': '...', 'status': {'code': 'ER_SUCCESS_MATCH'}, 'values': [{'value': {'name': 'music', 'id': '...'}}]}]\n # }}\n #]}\n \n if hasattr(intent.slots[\"feedbackAspects\"], 'slotValue'):\n feedbackAspectsSyntaxIsMessedUpAgain = True #JS-style syntax\n feedbackAspectsRaw = intent.slots[\"feedbackAspects\"].slotValue\n if feedbackAspectsRaw is None:\n feedbackAspectsIsGiven = False\n else:\n feedbackAspectsIsGiven = True\n if feedbackAspectsRaw['type'] == 'List' or feedbackAspectsRaw['object_type'] == 'List':\n feedbackAspectsIsMulti = True\n else: #feedbackAspectsRaw['type'] == 'Simple'\n feedbackAspectsIsMulti = False\n elif hasattr(intent.slots[\"feedbackAspects\"], 'slot_value'):\n feedbackAspectsSyntaxIsMessedUpAgain = False #python-style syntax\n feedbackAspectsRaw = intent.slots[\"feedbackAspects\"].slot_value\n if feedbackAspectsRaw is None:\n feedbackAspectsIsGiven = False\n else:\n feedbackAspectsIsGiven = True\n #feedbackAspectsRaw.type == 'List' does not work, because that would be too easy I guess.\n if feedbackAspectsRaw.object_type == 'List':\n feedbackAspectsIsMulti = True\n else: #feedbackAspectsRaw.object_type == 'Simple'\n feedbackAspectsIsMulti = False\n else:\n feedbackAspectsSyntaxIsMessedUpAgain = True\n feedbackAspectsRaw = intent.slots[\"feedbackAspects\"].value\n feedbackAspectsIsMulti = False\n if feedbackAspectsRaw is None:\n feedbackAspectsIsGiven = False\n else:\n feedbackAspectsIsGiven = True\n \n #check where we are in the feedback dialog:\n # 0 = no info given yet, need to get feedbackGeneral first\n # 1 = feedbackGeneral is given and to be processed, but we have yet to get feedbackAspects\n # 2 = feedbackGeneral and feedbackAspects are given, feedbackAspects has yet to be processed\n # 3 = we have all we need, so we can give some confirmation and move on.\n feedbackPhase = 0\n if request.dialog_state == DialogState.STARTED:\n if feedbackGeneralRaw is None:\n feedbackPhase = 0\n else:\n feedbackPhase = 1\n elif request.dialog_state == DialogState.IN_PROGRESS:\n if feedbackGeneralRaw is None: ##\n feedbackPhase = 0 ##\n elif not feedbackAspectsIsGiven:\n feedbackPhase = 1\n else:\n feedbackPhase = 2\n else: #request.dialog_state == DialogState.COMPLETED\n feedbackPhase = 3\n \n \n \n if feedbackPhase == 0: #prompt feedbackGeneral slot automatically\n responseBuilder.add_directive(DelegateDirective())\n \n elif feedbackPhase == 1: #we have feedbackGeneral, let's parse it and prompt feedbackAspects\n #process feedbackGeneralRaw\n isFeedbackPositive = isSentimentPositive(str(feedbackGeneralRaw))\n #store the canonical, simplified version of the feedback in the intent for later use\n intent.slots[\"feedbackGeneral\"].value = (FEEDBACK_NEGATIVE, FEEDBACK_POSITIVE)[isFeedbackPositive]\n \n speak_output = (SENTENCE_FEEDBACK_WAS_NEGATIVE,SENTENCE_FEEDBACK_WAS_POSITIVE)[isFeedbackPositive]\n prompt_output = PROMPT_FEEDBACK_ASPECTS.format((\"didn't\", \"did\")[isFeedbackPositive], \"it\")\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"feedbackAspects\", intent))\n \n elif feedbackPhase == 2:\n #resolve slots: try to match feedbackAspectsRaw to a canonical slot name\n slotResolutions = []\n if feedbackAspectsIsMulti == True:\n if feedbackAspectsSyntaxIsMessedUpAgain:\n allvalues = feedbackAspectsRaw['values']\n for val in allvalues:\n slotResolutions.append({\n \"orig\": val['value'],\n \"resolution\": val['resolutions']['resolutionsPerAuthority'][0]\n })\n else:\n allvalues = feedbackAspectsRaw.values\n for val in allvalues:\n slotResolutions.append({\n \"orig\": val.value,\n \"resolution\": val.resolutions.resolutions_per_authority[0]\n })\n else:\n if feedbackAspectsSyntaxIsMessedUpAgain:\n slotResolutions.append({\n \"orig\": feedbackAspectsRaw['value'],\n \"resolution\": feedbackAspectsRaw['resolutions']['resolutionsPerAuthority'][0]\n })\n else:\n slotResolutions.append({\n \"orig\": feedbackAspectsRaw.value, #['value'] tested, does not work. for reasons.\n \"resolution\": feedbackAspectsRaw.resolutions.resolutions_per_authority[0]\n })\n \n resolvedToSave = []\n unresolvedToSave = []\n if feedbackAspectsSyntaxIsMessedUpAgain:\n for val in slotResolutions:\n if val['resolution']['status']['code'] == 'ER_SUCCESS_MATCH':\n resolvedToSave.append(val['resolution']['values'][0]['value']['name']) #resolved to canonical slot value\n else:\n if(val['orig'] != 'the'): #sometimes, e.g. when saying \"the writing and the cast\", Alexa will fill the slots with \"writing\", \"the\" and \"cast\". It can't resolve \"the\", so throw all the \"the\"s out here.\n unresolvedToSave.append(val['orig']) #unresolved\n else:\n for val in slotResolutions:\n if val['resolution'].status.code == StatusCode.ER_SUCCESS_MATCH:\n resolvedToSave.append(val['resolution'].values[0].value.name) #resolved to canonical slot value\n else:\n if(val['orig'] != 'the'): #sometimes, e.g. when saying \"the writing and the cast\", Alexa will fill the slots with \"writing\", \"the\" and \"cast\". It can't resolve \"the\", so throw all the \"the\"s out here.\n unresolvedToSave.append(val['orig']) #unresolved\n \n handler_input.attributes_manager.session_attributes[\"resolvedMovieAspects\"] = resolvedToSave\n handler_input.attributes_manager.session_attributes[\"unresolvedMovieAspects\"] = unresolvedToSave\n \n if len(resolvedToSave)>0:\n #intent.slots[\"feedbackAspects\"].value = str(resolvedToSave)\n userID = handler_input.request_envelope.context.system.user.user_id\n update_likings_by_id(userID, {\n \"acting\": (\"acting\" in resolvedToSave),\n \"genre\": (\"genre\" in resolvedToSave),\n \"movie\": (\"movie\" in resolvedToSave)\n })\n responseBuilder.add_directive(DelegateDirective(intent))\n else:\n speak_output = SENTENCE_DID_NOT_UNDERSTAND\n prompt_output = PROMPT_FEEDBACK_ASPECTS.format((\"didn't\", \"did\")[intent.slots[\"feedbackGeneral\"].value is FEEDBACK_POSITIVE], \"the recommendation\")\n responseBuilder.speak(speak_output+prompt_output).ask(prompt_output).add_directive(makeESDir(\"finalConfirmation\"))\n \n #responseBuilder.add_directive(DelegateDirective(intent))\n \n else: # feedbackPhase == 3; we have all feedback values and can move on.\n userID = handler_input.request_envelope.context.system.user.user_id\n rate_last_watched_movie_by_id(userID, (feedbackGeneralRaw==FEEDBACK_POSITIVE))\n #aspects = intent.slots[\"feedbackAspects\"].value\n #update_likings_by_id(userID, {\n # \"acting\": (\"acting\" in aspects),\n # \"genre\": (\"genre\" in aspects),\n # \"movie\": (\"movie\" in aspects)\n #})\n \n speak_output = SENTENCE_FEEDBACK_ACCEPTED\n responseBuilder.speak(speak_output)\n \n \n return responseBuilder.response\n\n\nclass HelpIntentHandler(AbstractRequestHandler):\n \"\"\"Handler for Help Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_intent_name(\"AMAZON.HelpIntent\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n speak_output = \"To get a recommendation, you can just say: Recommend something. You can also request films similar to a given film, from a specific genre or with an actor or actress of your choice. To let me pick a good rewatch, just say: Recommend a rewatch. To change your profile, just say: Setup.\"\n\n return (\n handler_input.response_builder\n .speak(speak_output)\n .ask(speak_output)\n .response\n )\n\n\nclass CancelOrStopIntentHandler(AbstractRequestHandler):\n \"\"\"Single handler for Cancel and Stop Intent.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return (ask_utils.is_intent_name(\"AMAZON.CancelIntent\")(handler_input) or\n ask_utils.is_intent_name(\"AMAZON.StopIntent\")(handler_input))\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n speak_output = \"Cancelling.\"\n\n return (\n handler_input.response_builder\n .speak(speak_output)\n .response\n )\n\n\nclass SessionEndedRequestHandler(AbstractRequestHandler):\n \"\"\"Handler for Session End.\"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_request_type(\"SessionEndedRequest\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n\n # Any cleanup logic goes here.\n\n return handler_input.response_builder.response\n\n\nclass IntentReflectorHandler(AbstractRequestHandler):\n \"\"\"The intent reflector is used for interaction model testing and debugging.\n It will simply repeat the intent the user said. You can create custom handlers\n for your intents by defining them above, then also adding them to the request\n handler chain below.\n \"\"\"\n def can_handle(self, handler_input):\n # type: (HandlerInput) -> bool\n return ask_utils.is_request_type(\"IntentRequest\")(handler_input)\n\n def handle(self, handler_input):\n # type: (HandlerInput) -> Response\n intent_name = ask_utils.get_intent_name(handler_input)\n speak_output = \"You just triggered \" + intent_name + \".\"\n\n return (\n handler_input.response_builder\n .speak(speak_output)\n # .ask(\"add a reprompt if you want to keep the session open for the user to respond\")\n .response\n )\n\n\nclass CatchAllExceptionHandler(AbstractExceptionHandler):\n \"\"\"Generic error handling to capture any syntax or routing errors. If you receive an error\n stating the request handler chain is not found, you have not implemented a handler for\n the intent being invoked or included it in the skill builder below.\n \"\"\"\n def can_handle(self, handler_input, exception):\n # type: (HandlerInput, Exception) -> bool\n return True\n\n def handle(self, handler_input, exception):\n # type: (HandlerInput, Exception) -> Response\n logger.error(exception, exc_info=True)\n\n speak_output = \"Sorry, I had trouble doing what you asked. Please try again.\"\n\n return (\n handler_input.response_builder\n .speak(speak_output)\n .ask(speak_output)\n .response\n )\n\n# The SkillBuilder object acts as the entry point for your skill, routing all request and response\n# payloads to the handlers above. Make sure any new handlers or interceptors you've\n# defined are included below. The order matters - they're processed top to bottom.\n\n\nsb = SkillBuilder()\n\nsb.add_request_handler(LaunchRequestHandler())\nsb.add_request_handler(RecommendationIntentHandler())\nsb.add_request_handler(RecommendationByMovieIntentHandler())\nsb.add_request_handler(RecommendationByActorIntentHandler())\nsb.add_request_handler(RecommendationByGenreIntentHandler())\nsb.add_request_handler(RecommendationRewatchIntentHandler())\nsb.add_request_handler(SetupIntentHandler())\nsb.add_request_handler(FeedbackIntentHandler())\n\nsb.add_request_handler(HelpIntentHandler())\nsb.add_request_handler(CancelOrStopIntentHandler())\nsb.add_request_handler(SessionEndedRequestHandler())\nsb.add_request_handler(IntentReflectorHandler()) # make sure IntentReflectorHandler is last so it doesn't override your custom intent handlers\n\nsb.add_exception_handler(CatchAllExceptionHandler())\n\nlambda_handler = sb.lambda_handler()","repo_name":"onumtiger/ultron","sub_path":"lambda/lambda_function.py","file_name":"lambda_function.py","file_ext":"py","file_size_in_byte":61283,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"31076673013","text":"import re\n\n\ndef pretty_print_state(state):\n '''Pretty print the text of a state in the interactive fiction game'''\n state = str(state)\n state = state.replace('\\\\\\'', '\\'')\n pattern = re.compile(r'\\\\n|b\\'|b\"')\n state = re.sub(pattern, ' ', state)\n state = state.strip('\\'').strip('\\\"').strip()\n return state\n\n\ndef demo_n_games(agent, n_games=1000, mode='agent'):\n '''Run n game simulations and compute the average performance'''\n scores = []\n for _ in range(n_games):\n score = agent.demo_game(mode=mode, verbose=False)\n scores.append(score)\n return sum(scores)/n_games\n","repo_name":"bingwang32/RL_InteractiveFiction","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":615,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"72039902271","text":"from PySide2 import QtCore, QtGui, QtWidgets\nfrom PySide2.QtCore import (QCoreApplication, QPropertyAnimation, QDate, QDateTime, QMetaObject, QObject, QPoint, QRect,\n QSize, QTime, QUrl, Qt, QEvent)\nfrom PySide2.QtGui import (QBrush, QColor, QConicalGradient, QCursor, QFont, QFontDatabase, QIcon, QKeySequence,\n QLinearGradient, QPalette, QPainter, QPixmap, QRadialGradient)\nfrom PySide2.QtWidgets import *\n\n\nclass Ui_MainWindow(object):\n \"\"\"Clase autogenerada por pyuic que al lanzarla genera la ventana principal de la aplicación.\"\"\"\n\n def setupUi(self, MainWindow):\n MainWindow.setObjectName(\"MainWindow\")\n MainWindow.resize(1000, 500)\n MainWindow.setMinimumSize(QSize(1000, 500))\n MainWindow.setStyleSheet(\"background-color: rgb(45, 45, 45);\")\n self.centralwidget = QtWidgets.QWidget(MainWindow)\n self.centralwidget.setObjectName(\"centralwidget\")\n self.verticalLayout = QtWidgets.QVBoxLayout(self.centralwidget)\n self.verticalLayout.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout.setSpacing(0)\n self.verticalLayout.setObjectName(\"verticalLayout\")\n self.Top_Bar = QtWidgets.QFrame(self.centralwidget)\n self.Top_Bar.setMaximumSize(QtCore.QSize(50000, 40))\n self.Top_Bar.setStyleSheet(\"background-color: rgb(35, 35, 35);\")\n self.Top_Bar.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.Top_Bar.setFrameShadow(QtWidgets.QFrame.Raised)\n self.Top_Bar.setObjectName(\"Top_Bar\")\n self.horizontalLayout = QtWidgets.QHBoxLayout(self.Top_Bar)\n self.horizontalLayout.setContentsMargins(0, 0, 0, 0)\n self.horizontalLayout.setSpacing(0)\n self.horizontalLayout.setObjectName(\"horizontalLayout\")\n self.frame_toggle = QtWidgets.QFrame(self.Top_Bar)\n self.frame_toggle.setMaximumSize(QtCore.QSize(70, 40))\n self.frame_toggle.setStyleSheet(\"\")\n self.frame_toggle.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.frame_toggle.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_toggle.setObjectName(\"frame_toggle\")\n self.verticalLayout_2 = QtWidgets.QVBoxLayout(self.frame_toggle)\n self.verticalLayout_2.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_2.setSpacing(0)\n self.verticalLayout_2.setObjectName(\"verticalLayout_2\")\n self.btn_toggle = QtWidgets.QPushButton(self.frame_toggle)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Expanding)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.btn_toggle.sizePolicy().hasHeightForWidth())\n self.btn_toggle.setSizePolicy(sizePolicy)\n self.btn_toggle.setStyleSheet(\"background-image: url(:/newPrefix/menu.png);\\n\"\n \"background-position: center;\\n\"\n \"background-repeat: no-reperat;\\n\"\n \"border: none;\\n\"\n \"background-color: rgb(27, 29, 35);\")\n self.btn_toggle.setText(\"\")\n self.btn_toggle.setObjectName(\"btn_toggle\")\n self.verticalLayout_2.addWidget(self.btn_toggle)\n self.horizontalLayout.addWidget(self.frame_toggle)\n self.frame_top = QtWidgets.QFrame(self.Top_Bar)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Preferred)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.frame_top.sizePolicy().hasHeightForWidth())\n self.frame_top.setSizePolicy(sizePolicy)\n self.frame_top.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.frame_top.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_top.setObjectName(\"frame_top\")\n self.horizontalLayout.addWidget(self.frame_top)\n self.verticalLayout.addWidget(self.Top_Bar)\n self.Content = QtWidgets.QFrame(self.centralwidget)\n self.Content.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.Content.setFrameShadow(QtWidgets.QFrame.Raised)\n self.Content.setObjectName(\"Content\")\n self.horizontalLayout_2 = QtWidgets.QHBoxLayout(self.Content)\n self.horizontalLayout_2.setContentsMargins(0, 0, 0, 0)\n self.horizontalLayout_2.setSpacing(0)\n self.horizontalLayout_2.setObjectName(\"horizontalLayout_2\")\n self.frame_left_menu = QtWidgets.QFrame(self.Content)\n self.frame_left_menu.setMinimumSize(QtCore.QSize(70, 0))\n self.frame_left_menu.setMaximumSize(QtCore.QSize(70, 16777215))\n self.frame_left_menu.setStyleSheet(\"background-color: rgb(35, 35, 35);\")\n self.frame_left_menu.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_left_menu.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_left_menu.setObjectName(\"frame_left_menu\")\n self.verticalLayout_3 = QtWidgets.QVBoxLayout(self.frame_left_menu)\n self.verticalLayout_3.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_3.setSpacing(6)\n self.verticalLayout_3.setObjectName(\"verticalLayout_3\")\n self.frame_top_menus = QtWidgets.QFrame(self.frame_left_menu)\n self.frame_top_menus.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.frame_top_menus.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_top_menus.setObjectName(\"frame_top_menus\")\n self.verticalLayout_4 = QtWidgets.QVBoxLayout(self.frame_top_menus)\n self.verticalLayout_4.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_4.setSpacing(0)\n self.verticalLayout_4.setObjectName(\"verticalLayout_4\")\n self.btn_menu_home = QtWidgets.QPushButton(self.frame_top_menus)\n self.btn_menu_home.setMinimumSize(QtCore.QSize(0, 40))\n self.btn_menu_home.setStyleSheet(\"QPushButton{\\n\"\n \" color:rgb(255, 255, 255);\\n\"\n \" background-color: rgb(35, 35, 35);\\n\"\n \" border: 0px solid;\\n\"\n \" background-image: url(:/newPrefix/home.png);\\n\"\n \" background-repeat: no-reperat;\\n\"\n \" border: none;\\n\"\n \" background-position: center;\\n\"\n \"}\\n\"\n \"\\n\"\n \"QPushButton:hover{\\n\"\n \" background-color: rgb(255, 204, 0);\\n\"\n \"}\")\n self.btn_menu_home.setText(\"\")\n self.btn_menu_home.setObjectName(\"btn_menu_home\")\n self.verticalLayout_4.addWidget(self.btn_menu_home)\n self.btn_menu_home_2 = QtWidgets.QPushButton(self.frame_top_menus)\n self.btn_menu_home_2.setMinimumSize(QtCore.QSize(0, 40))\n self.btn_menu_home_2.setStyleSheet(\"QPushButton{\\n\"\n \" color:rgb(255, 255, 255);\\n\"\n \" background-color: rgb(35, 35, 35);\\n\"\n \" border: 0px solid;\\n\"\n \" background-image: url(:/newPrefix/create.png);\\n\"\n \" background-repeat: no-reperat;\\n\"\n \" border: none;\\n\"\n \" background-position: center;\\n\"\n \"}\\n\"\n \"\\n\"\n \"QPushButton:hover{\\n\"\n \" \\n\"\n \" background-color: rgb(255, 204, 0);\\n\"\n \"}\")\n self.btn_menu_home_2.setText(\"\")\n self.btn_menu_home_2.setObjectName(\"btn_menu_home_2\")\n self.verticalLayout_4.addWidget(self.btn_menu_home_2)\n self.btn_menu_home_3 = QtWidgets.QPushButton(self.frame_top_menus)\n self.btn_menu_home_3.setMinimumSize(QtCore.QSize(0, 40))\n self.btn_menu_home_3.setStyleSheet(\"QPushButton{\\n\"\n \" color:rgb(255, 255, 255);\\n\"\n \" background-color: rgb(35, 35, 35);\\n\"\n \" \\n\"\n \" background-image: url(:/newPrefix/testList.png);\\n\"\n \" background-repeat: no-reperat;\\n\"\n \" border: none;\\n\"\n \" background-position: center;\\n\"\n \"}\\n\"\n \"\\n\"\n \"QPushButton:hover{\\n\"\n \" \\n\"\n \" background-color: rgb(255, 204, 0);\\n\"\n \"}\")\n self.btn_menu_home_3.setText(\"\")\n self.btn_menu_home_3.setObjectName(\"btn_menu_home_3\")\n self.verticalLayout_4.addWidget(self.btn_menu_home_3)\n self.verticalLayout_3.addWidget(self.frame_top_menus, 0, QtCore.Qt.AlignTop)\n self.horizontalLayout_2.addWidget(self.frame_left_menu)\n self.frame_pages = QtWidgets.QFrame(self.Content)\n self.frame_pages.setMaximumSize(QtCore.QSize(16777215, 16777215))\n self.frame_pages.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_pages.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_pages.setObjectName(\"frame_pages\")\n self.verticalLayout_5 = QtWidgets.QVBoxLayout(self.frame_pages)\n self.verticalLayout_5.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_5.setSpacing(0)\n self.verticalLayout_5.setObjectName(\"verticalLayout_5\")\n self.pages_widget = QtWidgets.QStackedWidget(self.frame_pages)\n self.pages_widget.setObjectName(\"pages_widget\")\n self.pg_home = QtWidgets.QWidget()\n self.pg_home.setObjectName(\"pg_home\")\n self.horizontalLayout_3 = QtWidgets.QHBoxLayout(self.pg_home)\n self.horizontalLayout_3.setObjectName(\"horizontalLayout_3\")\n self.frame = QtWidgets.QFrame(self.pg_home)\n self.frame.setMaximumSize(QtCore.QSize(700, 350))\n self.frame.setStyleSheet(\"QFrame{\\n\"\n \" background-color: rgb(56, 58, 89 );\\n\"\n \" color:rgb(220, 220, 220);\\n\"\n \" border-radius:10px;\\n\"\n \"}\\n\"\n \"\\n\"\n \"\")\n self.frame.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame.setObjectName(\"frame\")\n self.verticalLayout_6 = QtWidgets.QVBoxLayout(self.frame)\n self.verticalLayout_6.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_6.setSpacing(0)\n self.verticalLayout_6.setObjectName(\"verticalLayout_6\")\n self.label = QtWidgets.QLabel(self.frame)\n font = QtGui.QFont()\n font.setFamily(\"Segoe UI\")\n font.setPointSize(90)\n font.setBold(False)\n font.setWeight(50)\n self.label.setFont(font)\n self.label.setStyleSheet(\"color:rgb(255, 204, 0);\")\n self.label.setScaledContents(False)\n self.label.setAlignment(QtCore.Qt.AlignCenter)\n self.label.setObjectName(\"label\")\n self.verticalLayout_6.addWidget(self.label)\n self.label_2 = QtWidgets.QLabel(self.frame)\n font = QtGui.QFont()\n font.setFamily(\"Segoe UI\")\n font.setPointSize(23)\n self.label_2.setFont(font)\n self.label_2.setStyleSheet(\"color:rgb(98,114,164);\")\n self.label_2.setAlignment(QtCore.Qt.AlignCenter)\n self.label_2.setObjectName(\"label_2\")\n self.verticalLayout_6.addWidget(self.label_2)\n self.horizontalLayout_3.addWidget(self.frame)\n self.pages_widget.addWidget(self.pg_home)\n self.pg_create = QtWidgets.QWidget()\n self.pg_create.setObjectName(\"pg_create\")\n self.verticalLayout_8 = QtWidgets.QVBoxLayout(self.pg_create)\n self.verticalLayout_8.setContentsMargins(6, 10, 0, 0)\n self.verticalLayout_8.setSpacing(0)\n self.verticalLayout_8.setObjectName(\"verticalLayout_8\")\n self.pages_test_create = QtWidgets.QStackedWidget(self.pg_create)\n self.pages_test_create.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.pages_test_create.setObjectName(\"pages_test_create\")\n self.pg_testname = QtWidgets.QWidget()\n self.pg_testname.setObjectName(\"pg_testname\")\n self.horizontalLayout_4 = QtWidgets.QHBoxLayout(self.pg_testname)\n self.horizontalLayout_4.setContentsMargins(0, 0, 0, 0)\n self.horizontalLayout_4.setSpacing(0)\n self.horizontalLayout_4.setObjectName(\"horizontalLayout_4\")\n self.frame_2 = QtWidgets.QFrame(self.pg_testname)\n self.frame_2.setMaximumSize(QtCore.QSize(300, 300))\n self.frame_2.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_2.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_2.setLineWidth(1)\n self.frame_2.setObjectName(\"frame_2\")\n self.verticalLayout_9 = QtWidgets.QVBoxLayout(self.frame_2)\n self.verticalLayout_9.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_9.setSpacing(0)\n self.verticalLayout_9.setObjectName(\"verticalLayout_9\")\n self.label_TituloTest = QtWidgets.QLabel(self.frame_2)\n self.label_TituloTest.setMaximumSize(QtCore.QSize(16777215, 30))\n font = QtGui.QFont()\n font.setPointSize(20)\n font.setBold(False)\n font.setWeight(50)\n self.label_TituloTest.setFont(font)\n self.label_TituloTest.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_TituloTest.setAlignment(QtCore.Qt.AlignCenter)\n self.label_TituloTest.setObjectName(\"label_TituloTest\")\n self.verticalLayout_9.addWidget(self.label_TituloTest)\n spacerItem = QtWidgets.QSpacerItem(20, 33, QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed)\n self.verticalLayout_9.addItem(spacerItem)\n self.lineEdit_nombreTest = QtWidgets.QLineEdit(self.frame_2)\n self.lineEdit_nombreTest.setMinimumSize(QtCore.QSize(0, 30))\n self.lineEdit_nombreTest.setStyleSheet(\"QLineEdit {\\n\"\n \" \\n\"\n \" color: rgb(255, 255, 255);\\n\"\n \" background-color: rgb(33, 37, 43);\\n\"\n \" border-radius: 5px;\\n\"\n \" border: 2px solid rgb(33, 37, 43);\\n\"\n \" padding-left: 10px;\\n\"\n \" selection-color: rgb(255, 255, 255);\\n\"\n \" selection-background-color: rgb(255, 121, 198);\\n\"\n \"}\\n\"\n \"QLineEdit:hover {\\n\"\n \" border: 2px solid rgb(64, 71, 88);\\n\"\n \"}\\n\"\n \"QLineEdit:focus {\\n\"\n \" border: 2px solid rgb(91, 101, 124);\\n\"\n \"}\")\n self.lineEdit_nombreTest.setText(\"\")\n self.lineEdit_nombreTest.setObjectName(\"lineEdit_nombreTest\")\n self.verticalLayout_9.addWidget(self.lineEdit_nombreTest)\n self.label_tituloTestError = QtWidgets.QLabel(self.frame_2)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Preferred, QtWidgets.QSizePolicy.Preferred)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.label_tituloTestError.sizePolicy().hasHeightForWidth())\n self.label_tituloTestError.setSizePolicy(sizePolicy)\n self.label_tituloTestError.setMaximumSize(QtCore.QSize(5000, 25))\n self.label_tituloTestError.setLayoutDirection(QtCore.Qt.LeftToRight)\n self.label_tituloTestError.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_tituloTestError.setLineWidth(1)\n self.label_tituloTestError.setText(\"\")\n self.label_tituloTestError.setScaledContents(False)\n self.label_tituloTestError.setAlignment(QtCore.Qt.AlignCenter)\n self.label_tituloTestError.setObjectName(\"label_tituloTestError\")\n self.verticalLayout_9.addWidget(self.label_tituloTestError)\n spacerItem1 = QtWidgets.QSpacerItem(20, 21, QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed)\n self.verticalLayout_9.addItem(spacerItem1)\n self.frame_3 = QtWidgets.QFrame(self.frame_2)\n self.frame_3.setMaximumSize(QtCore.QSize(16777215, 40))\n self.frame_3.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_3.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_3.setObjectName(\"frame_3\")\n self.horizontalLayout_5 = QtWidgets.QHBoxLayout(self.frame_3)\n self.horizontalLayout_5.setContentsMargins(0, 0, 0, 0)\n self.horizontalLayout_5.setSpacing(0)\n self.horizontalLayout_5.setObjectName(\"horizontalLayout_5\")\n self.btn_crearTest = QtWidgets.QPushButton(self.frame_3)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.btn_crearTest.sizePolicy().hasHeightForWidth())\n self.btn_crearTest.setSizePolicy(sizePolicy)\n self.btn_crearTest.setMaximumSize(QtCore.QSize(150, 30))\n self.btn_crearTest.setStyleSheet(\"QPushButton {\\n\"\n \" \\n\"\n \" color: rgb(255, 255, 255);\\n\"\n \" border: 2px solid rgb(52, 59, 72);\\n\"\n \" border-radius: 5px; \\n\"\n \" background-color: rgb(52, 59, 72);\\n\"\n \"}\\n\"\n \"QPushButton:hover {\\n\"\n \" background-color: rgb(57, 65, 80);\\n\"\n \" border: 2px solid rgb(61, 70, 86);\\n\"\n \"}\\n\"\n \"QPushButton:pressed { \\n\"\n \" background-color: rgb(35, 40, 49);\\n\"\n \" border: 2px solid rgb(43, 50, 61);\\n\"\n \"}\")\n self.btn_crearTest.setObjectName(\"btn_crearTest\")\n self.horizontalLayout_5.addWidget(self.btn_crearTest)\n self.verticalLayout_9.addWidget(self.frame_3)\n self.horizontalLayout_4.addWidget(self.frame_2)\n self.pages_test_create.addWidget(self.pg_testname)\n self.preguntas_test = QtWidgets.QWidget()\n self.preguntas_test.setObjectName(\"preguntas_test\")\n self.horizontalLayout_6 = QtWidgets.QHBoxLayout(self.preguntas_test)\n self.horizontalLayout_6.setContentsMargins(0, 0, 0, 0)\n self.horizontalLayout_6.setSpacing(0)\n self.horizontalLayout_6.setObjectName(\"horizontalLayout_6\")\n self.frame_4 = QtWidgets.QFrame(self.preguntas_test)\n self.frame_4.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_4.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_4.setObjectName(\"frame_4\")\n self.label_numeroPregunta = QtWidgets.QLabel(self.frame_4)\n self.label_numeroPregunta.setGeometry(QtCore.QRect(380, 10, 101, 51))\n self.label_numeroPregunta.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_numeroPregunta.setObjectName(\"label_numeroPregunta\")\n self.textEdit_Enunciado = QtWidgets.QTextEdit(self.frame_4)\n self.textEdit_Enunciado.setGeometry(QtCore.QRect(180, 90, 501, 31))\n self.textEdit_Enunciado.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.textEdit_Enunciado.setObjectName(\"textEdit_Enunciado\")\n self.label_Enunciado = QtWidgets.QLabel(self.frame_4)\n self.label_Enunciado.setGeometry(QtCore.QRect(90, 90, 71, 31))\n self.label_Enunciado.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_Enunciado.setObjectName(\"label_Enunciado\")\n self.frame_5 = QtWidgets.QFrame(self.frame_4)\n self.frame_5.setGeometry(QtCore.QRect(170, 130, 511, 231))\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Fixed)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.frame_5.sizePolicy().hasHeightForWidth())\n self.frame_5.setSizePolicy(sizePolicy)\n self.frame_5.setMaximumSize(QtCore.QSize(600, 300))\n self.frame_5.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_5.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_5.setObjectName(\"frame_5\")\n self.verticalLayout_10 = QtWidgets.QVBoxLayout(self.frame_5)\n self.verticalLayout_10.setObjectName(\"verticalLayout_10\")\n self.frame_6 = QtWidgets.QFrame(self.frame_5)\n self.frame_6.setMaximumSize(QtCore.QSize(16777215, 54))\n self.frame_6.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_6.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_6.setObjectName(\"frame_6\")\n self.horizontalLayout_7 = QtWidgets.QHBoxLayout(self.frame_6)\n self.horizontalLayout_7.setObjectName(\"horizontalLayout_7\")\n self.label_resA = QtWidgets.QLabel(self.frame_6)\n self.label_resA.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_resA.setObjectName(\"label_resA\")\n self.horizontalLayout_7.addWidget(self.label_resA)\n self.textEdit_resA = QtWidgets.QTextEdit(self.frame_6)\n self.textEdit_resA.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.textEdit_resA.setObjectName(\"textEdit_resA\")\n self.horizontalLayout_7.addWidget(self.textEdit_resA)\n self.checkBox_resA = QtWidgets.QCheckBox(self.frame_6)\n self.checkBox_resA.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.checkBox_resA.setObjectName(\"checkBox_resA\")\n self.horizontalLayout_7.addWidget(self.checkBox_resA)\n self.verticalLayout_10.addWidget(self.frame_6)\n self.frame_7 = QtWidgets.QFrame(self.frame_5)\n self.frame_7.setMaximumSize(QtCore.QSize(16777215, 54))\n self.frame_7.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_7.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_7.setObjectName(\"frame_7\")\n self.horizontalLayout_8 = QtWidgets.QHBoxLayout(self.frame_7)\n self.horizontalLayout_8.setObjectName(\"horizontalLayout_8\")\n self.label_resB = QtWidgets.QLabel(self.frame_7)\n self.label_resB.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_resB.setObjectName(\"label_resB\")\n self.horizontalLayout_8.addWidget(self.label_resB)\n self.textEdit_resB = QtWidgets.QTextEdit(self.frame_7)\n self.textEdit_resB.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.textEdit_resB.setObjectName(\"textEdit_resB\")\n self.horizontalLayout_8.addWidget(self.textEdit_resB)\n self.checkBox_resB = QtWidgets.QCheckBox(self.frame_7)\n self.checkBox_resB.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.checkBox_resB.setObjectName(\"checkBox_resB\")\n self.horizontalLayout_8.addWidget(self.checkBox_resB)\n self.verticalLayout_10.addWidget(self.frame_7)\n self.label_preguntaError = QtWidgets.QLabel(self.frame_5)\n self.label_preguntaError.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_preguntaError.setText(\"\")\n self.label_preguntaError.setAlignment(QtCore.Qt.AlignCenter)\n self.label_preguntaError.setObjectName(\"label_preguntaError\")\n self.verticalLayout_10.addWidget(self.label_preguntaError)\n self.frame_8 = QtWidgets.QFrame(self.frame_5)\n self.frame_8.setFrameShape(QtWidgets.QFrame.StyledPanel)\n self.frame_8.setFrameShadow(QtWidgets.QFrame.Raised)\n self.frame_8.setObjectName(\"frame_8\")\n self.horizontalLayout_9 = QtWidgets.QHBoxLayout(self.frame_8)\n self.horizontalLayout_9.setObjectName(\"horizontalLayout_9\")\n self.btn_siguientePregunta = QtWidgets.QPushButton(self.frame_8)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.btn_siguientePregunta.sizePolicy().hasHeightForWidth())\n self.btn_siguientePregunta.setSizePolicy(sizePolicy)\n self.btn_siguientePregunta.setMaximumSize(QtCore.QSize(150, 30))\n self.btn_siguientePregunta.setStyleSheet(\"QPushButton {\\n\"\n \" \\n\"\n \" color: rgb(255, 255, 255);\\n\"\n \" border: 2px solid rgb(52, 59, 72);\\n\"\n \" border-radius: 5px; \\n\"\n \" background-color: rgb(52, 59, 72);\\n\"\n \"}\\n\"\n \"QPushButton:hover {\\n\"\n \" background-color: rgb(57, 65, 80);\\n\"\n \" border: 2px solid rgb(61, 70, 86);\\n\"\n \"}\\n\"\n \"QPushButton:pressed { \\n\"\n \" background-color: rgb(35, 40, 49);\\n\"\n \" border: 2px solid rgb(43, 50, 61);\\n\"\n \"}\")\n self.btn_siguientePregunta.setObjectName(\"btn_siguientePregunta\")\n self.horizontalLayout_9.addWidget(self.btn_siguientePregunta)\n self.btn_finalizarTest = QtWidgets.QPushButton(self.frame_8)\n sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed)\n sizePolicy.setHorizontalStretch(0)\n sizePolicy.setVerticalStretch(0)\n sizePolicy.setHeightForWidth(self.btn_finalizarTest.sizePolicy().hasHeightForWidth())\n self.btn_finalizarTest.setSizePolicy(sizePolicy)\n self.btn_finalizarTest.setMaximumSize(QtCore.QSize(150, 30))\n self.btn_finalizarTest.setStyleSheet(\"QPushButton {\\n\"\n \" \\n\"\n \" color: rgb(255, 255, 255);\\n\"\n \" border: 2px solid rgb(52, 59, 72);\\n\"\n \" border-radius: 5px; \\n\"\n \" background-color: rgb(52, 59, 72);\\n\"\n \"}\\n\"\n \"QPushButton:hover {\\n\"\n \" background-color: rgb(57, 65, 80);\\n\"\n \" border: 2px solid rgb(61, 70, 86);\\n\"\n \"}\\n\"\n \"QPushButton:pressed { \\n\"\n \" background-color: rgb(35, 40, 49);\\n\"\n \" border: 2px solid rgb(43, 50, 61);\\n\"\n \"}\")\n self.btn_finalizarTest.setObjectName(\"btn_finalizarTest\")\n self.horizontalLayout_9.addWidget(self.btn_finalizarTest)\n self.verticalLayout_10.addWidget(self.frame_8)\n self.horizontalLayout_6.addWidget(self.frame_4)\n self.pages_test_create.addWidget(self.preguntas_test)\n self.verticalLayout_8.addWidget(self.pages_test_create)\n self.pages_widget.addWidget(self.pg_create)\n self.pg_list = QtWidgets.QWidget()\n self.pg_list.setObjectName(\"pg_list\")\n self.verticalLayout_7 = QtWidgets.QVBoxLayout(self.pg_list)\n self.verticalLayout_7.setContentsMargins(0, 0, 0, 0)\n self.verticalLayout_7.setSpacing(1)\n self.verticalLayout_7.setObjectName(\"verticalLayout_7\")\n self.label_TituloTest_2 = QtWidgets.QLabel(self.pg_list)\n self.label_TituloTest_2.setMaximumSize(QtCore.QSize(16777215, 30))\n font = QtGui.QFont()\n font.setPointSize(20)\n font.setBold(False)\n font.setWeight(50)\n self.label_TituloTest_2.setFont(font)\n self.label_TituloTest_2.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.label_TituloTest_2.setAlignment(QtCore.Qt.AlignCenter)\n self.label_TituloTest_2.setObjectName(\"label_TituloTest_2\")\n self.verticalLayout_7.addWidget(self.label_TituloTest_2)\n self.listWidgetTests = QtWidgets.QListWidget(self.pg_list)\n self.listWidgetTests.setStyleSheet(\"color: rgb(255, 255, 255);\")\n self.listWidgetTests.setObjectName(\"listWidgetTests\")\n self.verticalLayout_7.addWidget(self.listWidgetTests)\n self.pages_widget.addWidget(self.pg_list)\n self.verticalLayout_5.addWidget(self.pages_widget)\n self.horizontalLayout_2.addWidget(self.frame_pages)\n self.verticalLayout.addWidget(self.Content)\n MainWindow.setCentralWidget(self.centralwidget)\n\n self.retranslateUi(MainWindow)\n self.pages_widget.setCurrentIndex(2)\n self.pages_test_create.setCurrentIndex(0)\n QtCore.QMetaObject.connectSlotsByName(MainWindow)\n\n def retranslateUi(self, MainWindow):\n _translate = QtCore.QCoreApplication.translate\n MainWindow.setWindowTitle(_translate(\"MainWindow\", \"MainWindow\"))\n self.label.setText(_translate(\"MainWindow\", \"EzTests\"))\n self.label_2.setText(_translate(\"MainWindow\", \"Genera Test de manera sencilla\"))\n self.label_TituloTest.setText(_translate(\"MainWindow\", \"Titulo\"))\n self.lineEdit_nombreTest.setPlaceholderText(_translate(\"MainWindow\", \"Nombre del Test\"))\n self.btn_crearTest.setText(_translate(\"MainWindow\", \"Crear Test\"))\n self.label_numeroPregunta.setText(_translate(\"MainWindow\", \"Pregunta numero 1\"))\n self.textEdit_Enunciado.setPlaceholderText(_translate(\"MainWindow\", \"Enunciado de la pregunta\"))\n self.label_Enunciado.setText(_translate(\"MainWindow\", \"Pregunta:\"))\n self.label_resA.setText(_translate(\"MainWindow\", \"Respuesta A:\"))\n self.textEdit_resA.setPlaceholderText(_translate(\"MainWindow\", \"Texto de la respuesta\"))\n self.checkBox_resA.setText(_translate(\"MainWindow\", \"Correcta?\"))\n self.label_resB.setText(_translate(\"MainWindow\", \"Respuesta B:\"))\n self.textEdit_resB.setPlaceholderText(_translate(\"MainWindow\", \"Texto de la respuesta\"))\n self.checkBox_resB.setText(_translate(\"MainWindow\", \"Correcta?\"))\n self.btn_siguientePregunta.setText(_translate(\"MainWindow\", \"Siguiente pregunta\"))\n self.btn_finalizarTest.setText(_translate(\"MainWindow\", \"Finalizar Test\"))\n self.label_TituloTest_2.setText(_translate(\"MainWindow\", \"Lista de Test\"))\n\n\nfrom images import create_rc\nfrom images import home_rc\nfrom images import menu_rc\nfrom images import testList_rc\n","repo_name":"Jowy43/EzTestGenerator","sub_path":"UI/mainWindow.py","file_name":"mainWindow.py","file_ext":"py","file_size_in_byte":32188,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"42205006641","text":"from board_env import SnapyEnv\n\nenv = SnapyEnv()\n\nepisodes = 100\n\nfor episode in range(episodes):\n done = False\n obs = env.reset()\n while True:\n random_action = env.action_space.sample()\n print('action',random_action)\n obs, reward, done, info = env.step(random_action)\n print('reward', reward)\n","repo_name":"here-and-now/SnaPy","sub_path":"tests/doublecheckenv.py","file_name":"doublecheckenv.py","file_ext":"py","file_size_in_byte":331,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"60"}
+{"seq_id":"73814059710","text":"import asyncio\nimport json\nfrom channels.generic.websocket import AsyncWebsocketConsumer\nfrom groupchat.models import Message\nfrom rest_framework_simplejwt.tokens import AccessToken\nfrom channels.db import database_sync_to_async\n\n\nclass GroupChatConsumer(AsyncWebsocketConsumer):\n\n @database_sync_to_async\n def save_message(self, group_id, sender_id, data):\n message = Message.objects.create(\n group_id=group_id,\n sender_id=sender_id,\n content=data\n )\n sender_username=message.sender.username\n return sender_username\n\n async def connect(self):\n path = self.scope['path']\n parts = path.split('/')\n self.group_id = parts[1] if len(parts) > 1 else None\n\n self.group_name = f'group_{self.group_id}'\n \n try:\n # Authenticate the user\n await self.authenticate()\n\n # If authentication is successful, accept the WebSocket connection\n await self.accept()\n # Join the group\n await self.channel_layer.group_add(self.group_name, self.channel_name)\n\n \n except Exception as e:\n # If authentication fails, reject the WebSocket connection\n await self.close()\n new_message = \"Error message: \" + str(e)\n raise Exception(new_message) from e\n \n async def authenticate(self):\n try: \n # Retrieve the JWT token from the WebSocket headers\n access_token=self.scope['headers'][0][1].split(' ')[1]\n\n # Decode and validate the JWT token\n access_token = AccessToken(access_token)\n\n # Check if the token is valid and not expired\n access_token.verify()\n\n decoded_token = access_token.payload\n is_admin = decoded_token['is_admin']\n self.user_id = decoded_token['user_id']\n\n\n if is_admin:\n raise Exception(\"You are not authorized to perform this action\")\n except Exception as e:\n new_message = \"Error message: \" + str(e)\n raise Exception(new_message) from e\n\n async def disconnect(self, close_code):\n # Leave the group\n await self.channel_layer.group_discard(self.group_name, self.channel_name)\n\n async def receive(self, text_data):\n try:\n data = json.loads(text_data)\n message_content = data['content']\n group_id= self.group_id\n\n try: \n sender_username=await self.save_message(group_id,self.user_id,message_content)\n \n # Broadcast the message to the group\n await self.channel_layer.group_send(self.group_name, {\n \"type\":\"group_message\",\n \"message\":message_content,\n \"sender_username\":sender_username\n })\n \n except asyncio.TimeoutError as e:\n return {\"results\": f\"timeout error on {e}\"}\n except Exception as e:\n return {\"results\": f\"error on {e}\"}\n \n except Exception as e:\n print(e)\n\n async def group_message(self, event):\n message = event['message']\n sender_username = event['sender_username']\n\n # Send the message to the WebSocket\n await self.send(text_data=json.dumps({\n 'message': message,\n 'sender_username': sender_username\n }))\n","repo_name":"shraddhaL/RealtimeChatConnect","sub_path":"ChatConnect/consumer.py","file_name":"consumer.py","file_ext":"py","file_size_in_byte":3525,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"25231149686","text":"#!/usr/bin/python2\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mlp import MLP\n\ndef esincos(x):\n return np.exp(x) - x * np.sin(x) * np.cos(x)\n\nif __name__ == \"__main__\":\n x = np.random.rand(1, 100) * 10\n #x = np.linspace(0,10,100)\n y = (np.sin(x) + 1) / 2\n\n plt.plot(x[0], y[0], \"*r\")\n plt.show()\n\n # print x, y\n\n net = MLP([1, 10, 1])\n \n for iter in range(10000):\n t1 = [net.forward(x.take([i],axis=1))[0][0] for i in range(100)]\n t = [(net.forward(x.take([i],axis=1)) - y.take([i],axis=1))[0][0] for i in range(100)]\n # print(t)\n #print(x[0])\n #print(t)\n if(iter % 2000 == 0):\n plt.plot(x[0], t1, \"g.\")\n plt.show()\n error = np.sum(np.abs(t))\n print(error)\n for i in range(100):\n #print \"x.take([i], axis=1)\"\n #print x.take([i], axis=1)\n net.calcdiff(x.take([i], axis=1), y.take([i], axis=1) ,1)\n net.update(0.1, 0.9)\n","repo_name":"daikankan/neural-network","sub_path":"python/test_mlp.py","file_name":"test_mlp.py","file_ext":"py","file_size_in_byte":988,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"71058837312","text":"import numpy as np\nimport cv2\n\ncapture = cv2.VideoCapture(0)\ncapture.set(cv2.CAP_PROP_FRAME_WIDTH, 320)\ncapture.set(cv2.CAP_PROP_FRAME_HEIGHT, 240)\n\nwhile True:\n if cv2.waitKey(10) > 0: \n break\n\n ret, frame = capture.read()\n\n cv2.putText(frame,'test',(0,25), cv2.FONT_HERSHEY_PLAIN,1,(255,255,255))\n cv2.imshow(\"camera test\", frame)\n\n\n\nimport tensorflow.keras\nimport numpy as np\nimport cv2\n\nmodel_filename ='C:\\\\AISpace\\\\keras_model.h5'\nmodel = tensorflow.keras.models.load_model(model_filename)\n\ncapture = cv2.VideoCapture(0)\ncapture.set(cv2.CAP_PROP_FRAME_WIDTH, 320)\ncapture.set(cv2.CAP_PROP_FRAME_HEIGHT, 240)\n\n\ndef preprocessing(frame):\n size = (224, 224)\n frame_resized = cv2.resize(frame, size, interpolation=cv2.INTER_AREA)\n frame_normalized = (frame_resized.astype(np.float32) / 127.0) - 1\n frame_reshaped = frame_normalized.reshape((1, 224, 224, 3))\n\n return frame_reshaped\n\n\ndef predict(frame):\n prediction = model.predict(frame)\n return prediction\n\nwhile True:\n ret, frame = capture.read()\n\n if cv2.waitKey(100) > 0: \n break\n\n preprocessed = preprocessing(frame)\n prediction = predict(preprocessed)\n\n if (prediction[0,0] < prediction[0,1]):\n print('case1')\n cv2.putText(frame, 'case1', (0, 25), cv2.FONT_HERSHEY_PLAIN, 1, (0, 0, 0))\n else:\n cv2.putText(frame, 'case2', (0, 25), cv2.FONT_HERSHEY_PLAIN, 1, (0, 0, 0))\n print('case2')\n\n cv2.imshow(\"VideoFrame\", frame)\n \n","repo_name":"lcm412/Open_source_lab1","sub_path":"src/camera.py","file_name":"camera.py","file_ext":"py","file_size_in_byte":1479,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"33205770426","text":"from Dataclasses.layer import Layer\nimport json\nimport torch\nfrom modelparser import ModelParser\nfrom Dataclasses.hyperpar import Hyperparameters\n\nclass MnistModel(torch.nn.Module):\n def __init__(self,layer_list):\n super().__init__()\n self.layer_list = layer_list\n\n def _activation_mapper(self, act_string):\n\n if act_string == \"ReLU\":\n return torch.nn.ReLU()\n \n elif act_string == 'Sigmoid':\n return torch.nn.Sigmoid()\n \n else:\n pass\n\n def build_model(self):\n \n module_list = list()\n\n for layer_ix, layer in enumerate(self.layer_list):\n \n if \"Linear\" in layer.Layer_name:\n module_list.append(torch.nn.Flatten())\n intilayer = torch.nn.Linear(layer.Num_inputs, layer.Num_outputs, bias = layer.Bias)\n if \"Conv\" in layer.Layer_name:\n intilayer = torch.nn.Conv2d(in_channels = layer.Num_inputs, out_channels = layer.Num_outputs, kernel_size = layer.kernel_size,\n stride = layer.stride, padding = layer.padding, bias = False)\n if \"MaxPool2d\" in layer.Layer_name:\n intilayer = torch.nn.MaxPool2d(kernel_size = layer.kernel_size)\n \n module_list.append(intilayer)\n\n act = self._activation_mapper(layer.Activation)\n if act != None:\n module_list.append(act)\n\n dpt = torch.nn.Dropout2d(layer.dropout) if layer.Dropout else False\n if dpt:\n module_list.append(dpt)\n\n print(module_list)\n self.pred = torch.nn.Sequential(*module_list)\n\n return self.pred\n\n\ndef inimodel():\n try:\n parser = ModelParser(\"../base_config.json\")\n except:\n print('Couldn\\'t import the configartion module')\n\n layers = parser.get_list()\n kwargs = parser.get_hp()[0]\n model = MnistModel(layers)\n \n return model.build_model()\n\nif __name__ == '__main__':\n model = inimodel()\n print(model[0].weight)","repo_name":"Ruturajrmane/Projects","sub_path":"ML Project Architecture/src/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":2043,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"27195342113","text":"# -*- coding: utf-8 -*-\n\"\"\"\n斗地主拆牌模块\n@author 江胤佐\n\"\"\"\nfrom __future__ import annotations\n\nimport math\nfrom abc import ABCMeta\nfrom collections import defaultdict\nfrom copy import deepcopy\nfrom functools import cmp_to_key\nfrom typing import Optional\n\nfrom duguai.card.cards import *\nfrom duguai.card.cards import card_lt2, card_split\nfrom duguai.card.combo import Combo\n\n\"\"\"顺子/连对/最小的长度\"\"\"\nKIND_TO_MIN_LEN = {1: 5, 2: 3}\nMAX_Q = 10000\n\n\ndef _most_value(x):\n return np.argmax(np.bincount(x))\n\n\nMAX_VALUE_CMP = cmp_to_key(lambda x, y: max(x) - max(y))\nMOST_VALUE_CMP = cmp_to_key(lambda x, y: _most_value(x) - _most_value(y))\n\n\nclass AbstractDecomposer(metaclass=ABCMeta):\n \"\"\"\n 拆牌类。该类只负责拆出较好的牌组,不考虑其它玩家手牌的情况。\n 状态(state)表示目前的手牌。\n 动作(action)表示待出的牌。\n Q值:状态-动作的价值,Q(s, a)值越大,则越要在状态s下采取行动a.\n q = d(next_state) + len(a)\n \"\"\"\n\n @classmethod\n def decompose_value(cls, card_after: np.ndarray) -> int:\n \"\"\"\n 获取一副牌的分解值\n \"\"\"\n if len(card_after) == 0:\n return 0\n card_after = card_lt2(card_after)\n di, d_value, _ = card_to_suffix_di(card_after)\n\n # 顺子/连对\n for t in range(1, 3):\n max_len = max(len(i) for i in card_split(di[t]))\n if max_len >= KIND_TO_MIN_LEN[t]:\n d_value = max(d_value, t * max_len)\n\n return d_value\n\n @classmethod\n def _delta_q(cls, _max_q, _q):\n return (_max_q - _q) if _max_q - _q < 1000 else (_max_q - MAX_Q + 1 - _q)\n\n def _calc_q(self, lt2_state: np.ndarray, actions: np.ndarray[np.ndarray]) -> np.ndarray:\n \"\"\"对每一种状态-动作计算其Q\"\"\"\n result = []\n for a in actions:\n\n reward: int = 0\n # 拆炸弹的惩罚值,保证在 5 5 5 5 6的情况下拆出炸弹而非三带一\n for card in a:\n if np.sum(lt2_state == card) == 4 and len(a) < 4:\n reward = -1\n break\n\n next_state: np.ndarray = get_next_state(lt2_state, a)\n if next_state.size > 0:\n d_value = self.decompose_value(next_state)\n result.append(d_value + reward + len(a))\n else:\n # 该动作打完就没牌了,故d值为最大值\n result.append(MAX_Q + reward + len(a))\n return np.array(result)\n\n def _eval_actions(self,\n func,\n lt2_state: np.ndarray,\n **kwargs) -> Tuple[np.ndarray, np.ndarray]:\n actions = np.array(\n func(lt2_state, kwargs['length']) if 'card_list' not in kwargs.keys() else func(kwargs['card_list'],\n kwargs['kind'],\n kwargs['length']))\n # q = d(next state) + len(a)\n # 计算lt2_state下每一个action的q值\n q_list: np.ndarray = self._calc_q(lt2_state, actions)\n if len(q_list) == 0:\n return np.array([], dtype=int), np.array([], dtype=int)\n\n return actions, q_list\n\n def _process_card(self, card: np.ndarray):\n\n # 将手牌分解成不连续的部分\n self._lt2_cards, eq2_cards, self._ghosts = card_lt2_two_g(card)\n self._lt2_states: List[np.ndarray] = card_split(self._lt2_cards)\n self.card2_count: int = len(eq2_cards)\n\n def _get_all_actions_and_q_lists(self, lt2_state: np.ndarray) -> int:\n \"\"\"获取一个lt2_state下所有的actions及其对应的q_lists\"\"\"\n\n di, max_count, max_card_value = card_to_suffix_di(lt2_state)\n\n # solo pair trio bomb plane other\n self._actions = [[], [], [], [], [], []]\n self._q_lists = [np.array([], dtype=int), np.array([], dtype=int), np.array([], dtype=int),\n np.array([], dtype=int), np.array([], dtype=int), np.array([], dtype=int)]\n\n # solo pair trio bomb\n for i in range(1, 5):\n self._actions[i - 1], self._q_lists[i - 1] = self._eval_actions(_get_single_actions, lt2_state, length=i)\n\n # plane\n for length in range(3, len(di[3]) + 1):\n seq_actions, seq_q_list = self._eval_actions(_get_seq_actions,\n lt2_state,\n card_list=di[3],\n kind=3,\n length=length)\n self._actions[4].extend(seq_actions)\n self._q_lists[4] = np.concatenate([self._q_lists[4], seq_q_list])\n\n # 拆出顺子、连对\n for k, min_len in KIND_TO_MIN_LEN.items():\n card_list = di[k]\n for length in range(min_len, len(card_list) + 1):\n seq_actions, seq_q_list = self._eval_actions(_get_seq_actions,\n lt2_state,\n card_list=card_list,\n kind=k,\n length=length)\n self._actions[5].extend(seq_actions)\n self._q_lists[5] = np.concatenate([self._q_lists[5], seq_q_list])\n\n max_q = 0\n for q_list in self._q_lists:\n if q_list.size:\n max_q = max(np.max(q_list), max_q)\n return max_q\n\n\nclass FollowDecomposer(AbstractDecomposer):\n \"\"\"\n 跟牌拆牌器\n \"\"\"\n\n def __init__(self):\n self._output: Optional[List[np.ndarray]] = None\n\n # 存放带牌的列表\n self._take_lists: Optional[Dict[int, List[np.ndarray]]] = None\n\n # 存放主牌的列表\n self._main_lists: Optional[Dict[int, List[np.ndarray]]] = None\n\n # 存放主牌+带牌的列表\n self._main_take_lists: Optional[Dict[int, List[np.ndarray]]] = None\n\n # 炸弹列表\n self._bomb_list: Optional[List[np.ndarray]] = None\n\n # 仅维护主牌大小\n self._max_combo: Optional[Combo] = None\n\n self._last_combo: Optional[Combo] = None\n\n self._main_kind: Optional[int] = None\n self._take_kind: Optional[int] = None\n\n def _add_bomb(self, bomb_list: list) -> None:\n \"\"\"添加炸弹\"\"\"\n\n self._bomb_list: List[np.ndarray] = []\n\n # 添加王炸\n if len(self._ghosts) == 2:\n self._bomb_list.append(self._ghosts)\n\n # 添加4个2炸弹\n if self.card2_count == 4:\n self._bomb_list.append(np.array([CARD_2] * 4))\n\n if self._last_combo.is_bomb():\n for card in bomb_list:\n if card > self._last_combo.value:\n self._bomb_list.append(np.array([card, card, card, card]))\n else:\n for card in bomb_list:\n self._bomb_list.append(np.array([card, card, card, card]))\n\n def _add_valid_ghost(self):\n \"\"\"加入单只王。在此之前先加入2\"\"\"\n\n if self._ghosts.size:\n if self._last_combo.is_solo() \\\n and self._last_combo.main_kind == 1 and self._last_combo.value < self._ghosts[-1]:\n self._main_lists[2].append(self._ghosts[-1:])\n self._max_combo.cards = self._ghosts[-1:]\n elif self._max_combo.take_kind == 1:\n self._take_lists[2].append(self._ghosts[-1:])\n\n def _add_valid_card2(self):\n \"\"\"加入合法的2,之后再加入王\"\"\"\n if self.card2_count:\n if self._last_combo.is_single() \\\n and self._last_combo.main_kind <= self.card2_count and self._last_combo.value < CARD_2:\n self._main_lists[self._max_combo.main_kind].append(np.array([CARD_2] * self._max_combo.main_kind))\n self._max_combo.cards = [CARD_2] * self._max_combo.main_kind\n if self._last_combo.take_kind <= self.card2_count:\n # 2的价值比正常牌+1\n self._take_lists[self._last_combo.take_kind + 1].append(np.array([CARD_2] * self._last_combo.take_kind))\n\n def __merge_takes_to_main_seq(self, main_q: int, main_seq: np.ndarray, take_count: int) -> Tuple[int, np.ndarray]:\n tk = 0\n main_takes: np.ndarray = np.array(main_seq)\n total_delta_q = main_q\n\n # 从小到大遍历_take_lists,保证先合并最佳takes\n for delta_q, take_list in sorted(self._take_lists.items()):\n for take in take_list:\n if take[0] not in main_seq:\n tk += 1\n total_delta_q += delta_q\n main_takes = np.concatenate((main_takes, take))\n if tk == take_count:\n return total_delta_q, main_takes\n return 0, np.array([])\n\n def _merge_valid_main_takes(self) -> None:\n \"\"\"将合法的主牌和带牌拼接起来\"\"\"\n\n # 非炸弹是3带1单/1对,炸弹是4带2\n take_count = self._last_combo.seq_len\n if self._last_combo.main_kind == 4:\n take_count *= 2\n\n self._main_take_lists = defaultdict(list)\n\n for take_list in self._take_lists.values():\n take_list.sort(key=MAX_VALUE_CMP)\n\n if self._main_lists:\n\n # 挑选最佳的main_list,并排序\n main_q = min(self._main_lists.keys())\n self._main_lists[main_q].sort(key=MAX_VALUE_CMP)\n\n for main_seq in self._main_lists[main_q]:\n total_delta_q, main_takes = self.__merge_takes_to_main_seq(main_q, main_seq, take_count)\n if main_takes.size > 0:\n # 将得到的main_takes根据价值好坏加入相应的列表中\n self._main_take_lists[total_delta_q].append(main_takes)\n\n # 得到最大的main_takes\n self._max_main_takes = self.__merge_takes_to_main_seq(0, self._max_combo.cards, take_count)[1]\n\n def _update_main_lists_and_find_max(self, a: np.ndarray, q: int, max_q: int) -> None:\n \"\"\"在action有效的情况下加入到主列表,并更新最大值\"\"\"\n main_kind = self._last_combo.main_kind\n seq_len = self._last_combo.seq_len\n value = self._last_combo.value\n\n combo = Combo()\n combo.cards = a\n # 筛选符合规则的主牌\n if combo.value > value and combo.main_kind == main_kind and combo.seq_len == seq_len:\n self._main_lists[self._delta_q(max_q, q)].append(a)\n # 仅对比主牌大小,不关心是否带了牌\n if combo.value > self._max_combo.value:\n self._max_combo = deepcopy(combo)\n\n def _best_main_takes(self):\n if not self._main_take_lists:\n return 0, []\n min_delta_q = min(self._main_take_lists.keys())\n self._main_take_lists[min_delta_q].sort(key=MOST_VALUE_CMP)\n return min_delta_q, self._main_take_lists[min_delta_q]\n\n def _append_takes(self, length: int, kind: int, max_q):\n for a, q in zip(self._actions[kind - 1], self._q_lists[kind - 1]):\n self._take_lists[self._delta_q(max_q, q)].append(a[:length])\n\n def _add_valid_lt2_actions(self):\n for lt2_state in self._lt2_states:\n if lt2_state.size > 0:\n max_q: int = super(FollowDecomposer, self)._get_all_actions_and_q_lists(lt2_state)\n\n # 把单或者对加入_take_lists,对子可以视为2个单加入take列表\n if self._take_kind == 1:\n self._append_takes(1, 1, max_q)\n self._append_takes(1, 2, max_q)\n elif self._take_kind == 2:\n self._append_takes(2, 2, max_q)\n\n for actions, q_list in zip(self._actions, self._q_lists):\n for a, q in zip(actions, q_list):\n # 将合法的action加入到_main_lists,同时更新最大的main_kind\n self._update_main_lists_and_find_max(a, q, max_q)\n\n def _thieve_valid_actions(self) -> Tuple[int, List[np.ndarray]]:\n \"\"\"根据last combo的限制,筛选出有效且较好的动作\"\"\"\n\n self._add_valid_card2()\n self._add_valid_ghost()\n self._add_valid_lt2_actions()\n\n if not self._main_lists:\n return 0, []\n\n if self._take_kind:\n self._merge_valid_main_takes()\n return self._best_main_takes()\n else:\n self._max_main_takes = self._max_combo.cards\n min_delta_q = min(self._main_lists.keys())\n\n self._main_lists[min_delta_q].sort(key=MAX_VALUE_CMP)\n return min_delta_q, self._main_lists[min_delta_q]\n\n def _init(self, last_combo: Combo):\n # 初始化,key代表max_q - q,key越小拆得越好,越要优先选择\n self._take_lists: Dict[int, List[np.ndarray]] = defaultdict(list)\n self._main_lists: Dict[int, List[np.ndarray]] = defaultdict(list)\n self._output = []\n\n # max_combo仅保留主要部分,忽略带的部分\n self._max_combo = deepcopy(last_combo)\n self._max_main_takes = self._max_combo.cards\n self._last_combo = last_combo\n\n self._main_kind = self._max_combo.main_kind\n self._take_kind = self._max_combo.take_kind\n\n def get_good_follows(self, state: np.ndarray, last_combo: Combo) \\\n -> Tuple[List[np.ndarray], int, List[np.ndarray], np.ndarray]:\n \"\"\"\n 尽量给出较好的跟牌行动。\n @param state: 当前手牌。\n @param last_combo: 上一次出牌\n @return: 四元组:炸弹, 最好的组合 - 最好的跟牌(数字越大越不应该这样拆牌), 好的出牌的数组, 最大的出牌\n \"\"\"\n if last_combo.is_rocket():\n return [], 0, [], np.array([], dtype=int)\n\n self._process_card(state)\n self._init(last_combo)\n\n min_delta_q, self._output = self._thieve_valid_actions()\n\n self._add_bomb(card_to_di(self._lt2_cards)[0][4])\n\n self._max_combo.cards = self._max_main_takes\n\n return self._bomb_list, min_delta_q, self._output, (\n self._max_main_takes if self._max_combo > last_combo else np.array([], dtype=int))\n\n\nclass PlayHand:\n \"\"\"\n 出牌时,根据d_actions对手牌进行进一步分类\n \"\"\"\n\n def __init__(self, min_solo: int, max_solo: int):\n \"\"\"\n 初始化Hand类\n @see PlayDecomposer\n \"\"\"\n # solo pair trio bomb\n self._singles: List[List[np.ndarray]] = [[], [], [], []]\n\n self._planes: List[np.ndarray] = []\n\n self._trios_take: List[np.ndarray] = []\n self._planes_take: List[np.ndarray] = []\n self._bombs_take: List[np.ndarray] = []\n\n self._seq_solo5: List[np.ndarray] = []\n self._other_seq: List[np.ndarray] = []\n self._has_rocket: bool = False\n\n self._min_solo: int = min_solo\n self._max_solo: int = max_solo\n\n def add_to_hand(self, card_lists: List[Dict[int, List[np.ndarray]]]):\n \"\"\"将各种类型牌加入到PlayHand中\"\"\"\n for i in range(4):\n if card_lists[i].keys():\n min_delta_q = min(card_lists[i].keys())\n self._singles[i] = card_lists[i][min_delta_q]\n self._singles[i].sort(key=MAX_VALUE_CMP)\n\n # plane\n if card_lists[4].keys():\n min_delta_q = min(card_lists[4].keys())\n self._planes = card_lists[4][min_delta_q]\n self._planes.sort(key=MAX_VALUE_CMP)\n\n if card_lists[5].keys():\n min_delta_q = min(card_lists[5].keys())\n for action in card_lists[5][min_delta_q]:\n if action.size == 5:\n self._seq_solo5.append(action)\n else:\n self._other_seq.append(action)\n self._seq_solo5.sort(key=MAX_VALUE_CMP)\n\n self._merge_main_takes(self._planes, self._planes_take)\n self._merge_main_takes(self._singles[2], self._trios_take)\n self._merge_main_takes(self._singles[3], self._bombs_take)\n\n i = 0\n while i < len(self._bombs_take):\n if self._bombs_take[i].size <= 4:\n del self._bombs_take[i]\n else:\n i += 1\n\n @staticmethod\n def _choose_takes(take_list: List[np.ndarray], main_part: np.ndarray, take_count: int, split_pair: bool = False):\n\n main_part = np.concatenate([main_part] + take_list[:take_count])\n if split_pair:\n main_part = np.concatenate([main_part, take_list[take_count][:1]])\n\n return main_part\n\n def _merge_main_takes(self, main_list: List[np.ndarray], extended_target: List[np.ndarray]):\n \"\"\"\n 合并主要部分与带的牌\n \"\"\"\n main_take_list: List[np.ndarray] = []\n for main_part in main_list:\n\n # 防止main part带上自己的部分,例如 7 7 7不能带7\n temp_pairs: List[np.ndarray] = [i for i in self._singles[1] if i[0] not in np.unique(main_part)]\n temp_solos: List[np.ndarray] = [i for i in self._singles[0] if\n i[0] not in np.unique(main_part) and i[0] not in np.unique(temp_pairs)]\n\n take_count: int = math.ceil(main_part.size / 3)\n if len(temp_solos) >= take_count and len(temp_pairs) >= take_count:\n if np.mean(temp_solos) > np.mean(temp_pairs):\n main_take_list.append(self._choose_takes(temp_solos, main_part, take_count))\n else:\n main_take_list.append(self._choose_takes(temp_pairs, main_part, take_count))\n elif len(temp_pairs) >= take_count:\n main_take_list.append(self._choose_takes(temp_pairs, main_part, take_count))\n elif len(temp_solos) >= take_count:\n main_take_list.append(self._choose_takes(temp_solos, main_part, take_count))\n elif len(temp_solos) + 2 * len(temp_pairs) >= take_count:\n len_solos = len(temp_solos)\n main_part = self._choose_takes(temp_solos, main_part, len_solos)\n main_take_list.append(\n self._choose_takes(\n temp_pairs, main_part, (take_count - len_solos) // 2, (take_count - len_solos) % 2 == 1\n )\n )\n else:\n main_take_list.append(main_part)\n extended_target.extend(main_take_list)\n extended_target.sort(key=MOST_VALUE_CMP)\n\n @property\n def solos(self) -> List[np.ndarray]:\n \"\"\"单\"\"\"\n return self._singles[0]\n\n @property\n def pairs(self) -> List[np.ndarray]:\n \"\"\"对\"\"\"\n return self._singles[1]\n\n @property\n def trios(self) -> List[np.ndarray]:\n \"\"\"三\"\"\"\n return self._singles[2]\n\n @property\n def trios_take(self):\n \"\"\"三带M\"\"\"\n return self._trios_take\n\n @property\n def bombs(self) -> List[np.ndarray]:\n \"\"\"炸弹\"\"\"\n return self._singles[3]\n\n @property\n def bombs_take(self) -> List[np.ndarray]:\n \"\"\"四带2\"\"\"\n return self._bombs_take\n\n @property\n def planes(self) -> List[np.ndarray]:\n \"\"\"飞机(不带M)\"\"\"\n return self._planes\n\n @property\n def planes_take(self):\n \"\"\"飞机(带M)\"\"\"\n return self._planes_take\n\n @property\n def other_seq(self) -> List[np.ndarray]:\n \"\"\"其它各种序列\"\"\"\n return self._other_seq\n\n @property\n def seq_solo5(self) -> List[np.ndarray]:\n \"\"\"长度为5的单顺\"\"\"\n return self._seq_solo5\n\n @property\n def has_rocket(self) -> bool:\n \"\"\"是否有王炸\"\"\"\n return self._has_rocket\n\n @property\n def min_solo(self) -> int:\n \"\"\"强拆的最小单牌\"\"\"\n return self._min_solo\n\n @property\n def max_solo(self) -> int:\n \"\"\"强拆的最大单牌\"\"\"\n return self._max_solo\n\n def __repr__(self):\n return 'PlayHand: ' + repr(self.__dict__)\n\n\nclass PlayDecomposer(AbstractDecomposer):\n \"\"\"\n 基于贪心法的斗地主出牌时的拆牌算法。\n 出牌时仅考虑强行拆最大和最小的单牌。其余牌型均按照最佳拆牌给出。\n\n 定义c为一张牌,由斗地主规则可知,c ∈ [1, 15] ∩ Z+。\n 定义s表示当前玩家拥有的所有牌的序列,s = (c1, c2, ..., ci)。\n 定义a为一次符合斗地主规则的出牌的序列,a = (c1, c2, ..., ci)。\n\n 记s下满足规则的所有拆牌动作的集合为A_s,a∈A_s。\n\n 用函数D(a)来计算a拆牌的好坏。D(a)定义如下:\n D(a) = len(a) + max( max(len(a')) , 1) - 拆炸弹的数量\n\n 其中定义域a∈A,值域D(a)∈Z+, a' ∈ s - a\n D(a)越大,拆牌越合理。\n\n 算法如下:\n 1. 将s分成连续的若干段、二和大小王,例如(1,1,2,2,5,5,7,10,13,14)分成(1,1,2,2) (5,5,) (7) (10) (13) (14)\n 2. 将大小王和二加入到最佳拆牌序列A‘中\n 3. 对每一段序列si,计算不带牌的动作a的 D(a)\n 4. 合并主牌和带牌的D(a)\n 5. 输出argmax(D(a))\n \"\"\"\n\n def __init__(self):\n self.cards_q_maps_list: Optional[List[Dict[int, List[np.ndarray]]]] = None\n\n def _map_actions(self, actions, q_list, max_q: int, idx: int):\n for a, q in zip(actions, q_list):\n if max_q == q:\n self.cards_q_maps_list[idx][max_q - q].append(a)\n\n def get_good_plays(self, cards: np.ndarray) -> PlayHand:\n \"\"\"\n 获取较好的出牌行动。\n @param cards: 当前手牌。\n @return: 包含所有好的出牌类型的数组\n \"\"\"\n self._process_card(cards)\n self.cards_q_maps_list = [defaultdict(list), defaultdict(list),\n defaultdict(list), defaultdict(list),\n defaultdict(list), defaultdict(list)]\n\n play_hand = PlayHand(np.min(cards), np.max(cards))\n\n for lt2_state in self._lt2_states:\n if lt2_state.size > 0:\n max_q = self._get_all_actions_and_q_lists(lt2_state)\n i = 0\n for actions, q_list in zip(self._actions, self._q_lists):\n self._map_actions(actions, q_list, max_q, i)\n i += 1\n\n if self.cards_q_maps_list[0].keys() and self.cards_q_maps_list[1].keys():\n min_key = min(self.cards_q_maps_list[0].keys())\n min_key2 = min(self.cards_q_maps_list[1].keys())\n self.cards_q_maps_list[0][min_key] = [i for i in self.cards_q_maps_list[0][min_key] if\n i[0] not in np.unique(self.cards_q_maps_list[1][min_key2])]\n if self.card2_count:\n self.cards_q_maps_list[self.card2_count - 1][0].append(np.array([CARD_2] * self.card2_count))\n\n if self._ghosts.size == 2:\n play_hand._has_rocket = True\n elif self._ghosts.size == 1:\n self.cards_q_maps_list[0][0].append(self._ghosts)\n\n play_hand.add_to_hand(self.cards_q_maps_list)\n return play_hand\n\n\ndef get_next_state(state: np.ndarray, action: np.ndarray) -> np.ndarray:\n \"\"\"\n 获取状态做出动作后的的下一个状态\n @param state: 状态\n @param action: 动作\n @return: 下一个状态\n \"\"\"\n next_state = list(state)\n for card in action:\n next_state.remove(card)\n return np.array(next_state)\n\n\ndef _get_single_actions(state: np.ndarray, length: int) -> List[List[int]]:\n \"\"\"\n 获取所有单种牌面的动作(单,对,三,炸弹)\n @param state: 状态\n @param length: 动作长度\n \"\"\"\n result = []\n last_card = -1\n state = list(state)\n for i in range(length, len(state) + 1):\n if state[i - 1] == state[i - length] and state[i - 1] != last_card and (\n state.count(state[i - 1]) < 4 or length % 2 == 0):\n last_card = state[i - 1]\n result.append([last_card] * length)\n return result\n\n\ndef _get_seq_actions(card_list: list, kind: int, length: int) -> List[List[int]]:\n \"\"\"\n 获取顺子/连对/飞机/炸弹的动作(单,对,三,炸弹)\n \"\"\"\n result = []\n for i in range(length - 1, len(card_list)):\n if card_list[i] == card_list[i - length + 1] + length - 1:\n result.append(sorted(card_list[i - length + 1: i + 1] * kind))\n return result\n","repo_name":"jiangyinzuo/du-guai","sub_path":"duguai/ai/decompose.py","file_name":"decompose.py","file_ext":"py","file_size_in_byte":24705,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"60"}
+{"seq_id":"41278982284","text":"import requests\n\nclass BomberFriends:\n\tdef __init__(self) -> None:\n\t\tself.api = \"https://e1e6.playfabapi.com\"\n\t\tself.headers = {\n\t\t\t\"user-agent\": \"Dalvik/2.1.0 (Linux; U; Android 7.1.2; ASUS_Z01QD Build/QKQ1.190825.002)\",\n\t\t\t\"x-playfabsdk\": \"Cocos2d-xSDK-0.40.180529\",\n\t\t\t\"content-type\": \"application/json\",\n\t\t\t\"connection\": \"keep_alive\"\n\t\t}\n\t\tself.user_id = None\n\t\tself.title_id = \"E1E6\"\n\t\tself.entity_token = None\n\t\tself.session_ticket = None\n\t\t\n\tdef login_with_custom_id(\n\t\t\tself,\n\t\t\tcustom_id: str,\n\t\t\tcreate_account: bool = True) -> dict:\n\t\tdata = {\n\t\t\t\"CreateAccount\": create_account,\n\t\t\t\"CustomId\": custom_id,\n\t\t\t\"TitleId\": self.title_id\n\t\t}\n\t\tresponse = requests.post(\n\t\t\tf\"{self.api}/Client/LoginWithCustomID\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\t\tif \"SessionTicket\" in response[\"data\"]:\n\t\t\tself.custom_id = custom_id\n\t\t\tself.user_id = response[\"data\"][\"PlayFabId\"]\n\t\t\tself.session_ticket = response[\"data\"][\"SessionTicket\"]\n\t\t\tself.entity_token = response[\"data\"][\"EntityToken\"][\"EntityToken\"]\n\t\t\tself.headers[\"x-authorization\"] = self.session_ticket\n\t\t\tself.headers[\"x-entitytoken\"] = self.entity_token\n\t\treturn response\n\n\tdef get_entity_token(self) -> dict:\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Authentication/GetEntityToken\",\n\t\t\theaders=self.headers).json()\n\n\tdef get_title_data(\n\t\t\tself,\n\t\t\tkeys: list = [\"ScriptVersionAndroid\"]) -> dict:\n\t\tdata = {\n\t\t\t\"Keys\": keys\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetTitleData\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef get_server_status(\n\t\t\tself,\n\t\t\tplatform: str = \"android\",\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"getServerStatus\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"platform\": platform\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef update_display_name(self, display_name: str) -> dict:\n\t\tdata = {\n\t\t\t\"DisplayName\": display_name\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/UpdateUserTitleDisplayName\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef get_leaderboard(\n\t\t\tself,\n\t\t\tmax_results_count: int = 100,\n\t\t\tstart_position: int = 0,\n\t\t\tstatistic_name: str = \"Trophies\",\n\t\t\tshow_avatar_url: bool = False,\n\t\t\tshow_banned_until: bool = False,\n\t\t\tshow_campaign_attributions: bool = False,\n\t\t\tshow_contact_email_addresses: bool = False,\n\t\t\tshow_created: bool = False,\n\t\t\tshow_display_name: bool = True,\n\t\t\tshow_last_login: bool = False,\n\t\t\tshow_linked_accounts: bool = False,\n\t\t\tshow_locations: bool = False,\n\t\t\tshow_memberships: bool = False,\n\t\t\tshow_origiation: bool = False,\n\t\t\tshow_push_notification_registrations: bool = False,\n\t\t\tshow_statistics: bool = False,\n\t\t\tshow_tags: bool = True,\n\t\t\tshow_total_value_to_data_in_usd: bool = False,\n\t\t\tshow_values_to_date: bool = False) -> dict:\n\t\tdata = {\n\t\t\t\"MaxResultsCount\": max_results_count,\n\t\t\t\"ProfileConstraints\": {\n\t\t\t\t\"ShowAvatarUrl\": show_avatar_url,\n\t\t\t\t\"ShowBannedUntil\": show_banned_until,\n\t\t\t\t\"ShowCampaignAttributions\": show_campaign_attributions,\n\t\t\t\t\"ShowContactEmailAddresses\": show_contact_email_addresses,\n\t\t\t\t\"ShowCreated\": show_created,\n\t\t\t\t\"ShowDisplayName\": show_display_name,\n \t\t\t\"ShowLastLogin\": show_last_login,\n \t\t\t\"ShowLinkedAccounts\": show_linked_accounts,\n \t\t\t\"ShowLocations\": show_locations,\n \t\t\t\"ShowMemberships\": show_memberships,\n \t\t\t\"ShowOrigination\": show_origiation,\n \t\t\t\"ShowPushNotificationRegistrations\": show_push_notification_registrations,\n \t\t\t\"ShowStatistics\": show_statistics,\n \t\t\t\"ShowTags\": show_tags,\n \t\t\t\"ShowTotalValueToDateInUsd\": show_total_value_to_data_in_usd,\n \t\t\t\"ShowValuesToDate\": show_values_to_date\n\t\t\t},\n\t\t\t\"StartPosition\": start_position,\n\t\t\t\"StatisticName\": statistic_name\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetLeaderboard\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef claim_season_prize(\n\t\t\tself,\n\t\t\tname: str,\n\t\t\tid: int = 0,\n\t\t\tis_free: bool = True,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"claimSeasonPrize\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"id\": id,\n\t\t\t\t\"isFree\": is_free,\n\t\t\t\t\"name\": name\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef get_initial_data(\n\t\t\tself,\n\t\t\tversion: int = 756,\n\t\t\tplatform: str = \"android\",\n\t\t\tpd_datas: list = [\"deckslots\", \"dungeonrundata\", \"testidata\"],\n\t\t\trod_datas: list = [\"_clandata\"],\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"getInitialData\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"version\": version,\n\t\t\t\t\"platform\": platform,\n\t\t\t\t\"pddatas\": pd_datas,\n\t\t\t\t\"roddatas\": rod_datas\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef claim_reward(\n\t\t\tself,\n\t\t\tid: int = 0,\n\t\t\treward_spin: bool = False,\n\t\t\tname: str = \"wheel\"\t,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"claimReward\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"id\": id,\n\t\t\t\t\"reward_spin\": reward_spin,\n\t\t\t\t\"name\": name\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef get_single_player_reward(\n\t\t\tself,\n\t\t\tlevel: int,\n\t\t\ttype: int = 0,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"getSinglePlayerReward\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"level\": level,\n\t\t\t\t\"type\": 0\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef save_bomber_user_data(\n\t\t\tself,\n\t\t\tdata: str,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"saveBomberUserData\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"data\": data,\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef start_opening_slot(\n\t\t\tself,\n\t\t\tslot: int,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"tryStartOpeningSlotChest\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"Slot\": slot\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef open_slot(\n\t\t\tself,\n\t\t\tslot: int,\n\t\t\tcost: int = 0,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"tryStartOpeningSlotChest\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"Slot\": slot,\n\t\t\t\t\"cost\": cost\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef update_user_data(\n\t\t\tself,\n\t\t\tdata: dict,\n\t\t\tpermission: str = \"Private\") -> dict:\n\t\tdata = {\n\t\t\t\"Data\": data,\n\t\t\t\"Permission\": permission\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/UpdateUserData\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef tutorial_won(\n\t\t\tself,\n\t\t\tstanding: int = 0,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"tutorialWon\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"standing\": standing\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\n\tdef add_fashion_points(\n\t\t\tself,\n\t\t\tpoints: int,\n\t\t\tad: bool = False,\n\t\t\tgenerate_play_stream_event: bool = False,\n\t\t\trevision_selection: str = \"Live\") -> dict:\n\t\tdata = {\n\t\t\t\"FunctionName\": \"addFashionPoints\",\n\t\t\t\"FunctionParameter\": {\n\t\t\t\t\"points\": points,\n\t\t\t\t\"ad\": ad\n\t\t\t},\n\t\t\t\"GeneratePlayStreamEvent\": generate_play_stream_event,\n\t\t\t\"RevisionSelection\": revision_selection\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/ExecuteCloudScript\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\t\t\n\tdef get_user_data(self, user_id: str) -> dict:\n\t\tdata = {\n\t\t\t\"PlayFabId\": user_id\n\t\t}\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetUserData\",\n\t\t\tjson=data,\n\t\t\theaders=self.headers).json()\n\t\n\tdef get_player_statistics(self) -> dict:\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetPlayerStatistics\",\n\t\t\theaders=self.headers).json()\n\t\n\tdef get_inventory(self) -> dict:\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetUserInventory\",\n\t\t\theaders=self.headers).json()\n\t\n\n\tdef get_catalog_items(self) -> dict:\n\t\treturn requests.post(\n\t\t\tf\"{self.api}/Client/GetCatalogItems\",\n\t\t\theaders=self.headers).json()\n","repo_name":"zeviel/bomber_friends.py","sub_path":"src/bomber_friends.py","file_name":"bomber_friends.py","file_ext":"py","file_size_in_byte":9525,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"32081428529","text":"import uuid\nfrom pathlib import Path\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.preprocessing import LabelEncoder\n\nfrom commons import CONFIGS_DIR, STATES_DIR\nfrom configuration import *\nfrom dataset_transforms import TransformsComposer, ToTensor, Rescale\nfrom classifier import Classifier\nfrom data_loader import DataLoader\nfrom m5 import M5\nfrom siamese import Siamese\nfrom similarity_classifier import SimilarityClassifier\nfrom similarity_dataset import SimilarityDataset\nfrom utils import create_results_directories\nfrom sample_logger import SampleLogger\n\nCONFIG_FILENAME = 'config.json'\nRESULTS_DIR = 'siamese_results/'\nSAMPLE_LOGGER_FILE = 'samples.json'\n\n\ndef main():\n config_filename = Path.cwd().joinpath(CONFIGS_DIR).joinpath(CONFIG_FILENAME)\n config = Configuration(config_filename)\n\n batch_size = 4\n epochs = 4\n\n results_dir_path = Path.cwd().joinpath(RESULTS_DIR)\n current_run_path = create_results_directories(results_dir_path)\n\n sample_logger_path = Path.cwd().joinpath(current_run_path).joinpath(SAMPLE_LOGGER_FILE)\n sample_logger = SampleLogger(sample_logger_path)\n\n transforms = TransformsComposer([Rescale(output_size=10000), ToTensor()])\n\n encoder = LabelEncoder()\n\n data_loader = DataLoader(config)\n x_train, y_train = data_loader.get_train_set()\n encoder.fit(y_train)\n\n classes = encoder.classes_\n classes_map = {}\n for i, category in enumerate(classes):\n classes_map[i] = category\n print(classes_map)\n\n y_train = encoder.transform(y_train)\n train_dataset = SimilarityDataset(x_train, y_train, classes_map, sample_logger, transforms)\n\n x_test, y_test = data_loader.get_test_set()\n y_test = encoder.transform(y_test)\n test_dataset = SimilarityDataset(x_test, y_test, classes_map, sample_logger, transforms)\n\n model = Siamese()\n\n states_dir = Path.cwd().joinpath(STATES_DIR)\n state_filename = f'{uuid.uuid1()}_state_{epochs}_epochs.pth'\n state_path = current_run_path.joinpath('best_snapshot').joinpath(state_filename)\n\n classifier = SimilarityClassifier(model=model, state_path=state_path)\n\n # Fit model on data\n train_loss_history, val_loss_history = classifier.fit(train_dataset, batch_size=batch_size, epochs=epochs,\n validation_data=test_dataset)\n\n sample_logger.save()\n\n # plt.figure()\n # plt.title(f'Model Loss for {epochs} epochs')\n # plt.xlabel('epoch')\n # plt.ylabel('loss')\n # plt.plot(train_loss_history, label='train')\n # plt.plot(val_loss_history, label='test')\n # plt.legend()\n # plt.show()\n\n predictions_path = Path.cwd().joinpath('./predicted.csv')\n validation_dataset = SimilarityDataset(x_test, y_test, classes_map, sample_logger, transforms)\n validation_model = Siamese(num_classes=len(classes_map))\n validation_classifier = SimilarityClassifier(validation_model, state_path=state_path)\n validation_classifier.predict(validation_dataset, batch_size=batch_size, output_filepath=predictions_path)\n\n\nif __name__ == '__main__':\n main()","repo_name":"Amnonop/voice-corruption-classifier","sub_path":"run_similarity.py","file_name":"run_similarity.py","file_ext":"py","file_size_in_byte":3109,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"44607233294","text":"soma = cont = cont_mil = menor = 0\nbarato = ''\nprint('-' * 40)\nprint('\\tLOJA DO JENA')\nprint('-' * 40)\nwhile True:\n nome = str(input('Nome do Produto: '))\n preco = int(input('Preço: R$ '))\n cont += 1\n soma += preco\n if preco >= 1000:\n cont_mil += 1\n if cont == 1 or preco < menor: # maneira simplificada\n menor = preco\n barato = nome\n continuar = ' '\n while continuar not in 'SN':\n continuar = str(input('Quer continuar? [S/N] ')).strip().upper()[0]\n if continuar == 'N':\n break\nprint('{:-^40}'.format(' FIM DO PROGRAMA '))\nprint(f'O total da compra foi R${soma:.2f}')\nprint(f'Temos {cont_mil} produtos custando mais de R$1000.00')\nprint(f'O produto mais barato foi {barato} que custa R${menor:.2f}')\nprint()\n\nprodutos = [] # type: list\nprecos = [] # type: list\nwhile True:\n produto = str(input('Nome do produto: '))\n preco = int(input('Preço: R$ '))\n produtos.append(produto)\n precos.append(preco)\n continuar = ' '\n while continuar not in 'SN':\n continuar = str(input('Quer continuar? [S/N] ')).strip().upper()[0]\n if continuar == 'N':\n break\nprint('{:-^40}'.format(' FIM DO PROGRAMA '))\nprint(f'O total da compra foi R${sum(precos):.2f}')\nprint(f'Temos {len([i for i in precos if i > 1000])} produto(s) custando mais que R$1.000,00') # noqa\nprint(f'O produto mais barato foi {produtos[precos.index(min(precos))]} e custou R${min(precos)}') # noqa\n","repo_name":"JenaCarry/Curso-Python","sub_path":"desafios/desafio070.py","file_name":"desafio070.py","file_ext":"py","file_size_in_byte":1453,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"60"}
+{"seq_id":"13913181820","text":"#!/usr/bin/env python3\n\n'''\n Python script to convert YouTube XML subtitles to SRT format\n It might be neater to use an XML library but I want to avoid \n external dependencies, and pretending that it's not XML, it\n resembles a fairly simple text file.\n \n Owain Kenway\n'''\n\n# Convert a numeric time in ms to format used in SRT\n# timestamp - the time in ms since start\n# delta - extra time for calculating durations.\ndef timeconvert(timestamp=0, delta=0):\n wn = timestamp + delta\n hours = int(wn/(3600 * 1000))\n wn = wn - (hours * 3600 * 1000)\n minutes = int(wn/(60 * 1000))\n wn = wn - (minutes*60*1000)\n seconds = int(wn/1000)\n ms = wn-(seconds * 1000)\n\n return (str(hours) + \":\" + str(minutes) + \":\" + str(seconds) + \",\" + str(ms))\n\n# Generate a title timestamp\n# SRT format for time is:\n# HH:mm:ss,ms --> HH:mm:ss,ms\ndef gents(timestamp=0, duration=0):\n\n b = timeconvert(timestamp, 0)\n e = timeconvert(timestamp, duration)\n\n return(b + \" --> \" + e)\n\n# Generate a title from an XML line\n# SRT format is:\n# counter\n# HH:mm:ss,ms --> HH:mm:ss,ms\n# text\n# \\n\ndef processline(line=\"\", counter=0):\n\n# ignore empty lines\n if line.strip() == \"\" :\n return \"\"\n\n t = 0\n d = 0\n text=\"\"\n\n# Tidy line\n line = line.strip()\n if (line[0:2] == \"
\"):\n line = line[:-4]\n line = line.strip()\n\n# Split into block about timing, and text block.\n tokens = line.split(\">\", 1)\n preamble = tokens.pop(0)\n \n# Work out times from the preamble\n times = preamble.split(\" \")\n t = int(times[0].strip(\"t=\").strip('\"').strip())\n d = int(times[1].strip(\"d=\").strip('\"').strip())\n\n# The rest of the line is text.\n text = tokens[0]\n\n# Replace some HTML entitles.\n text = text.replace(\"'\", \"'\")\n text = text.replace(\""\", '\"')\n\n return (str(counter) + \"\\n\" + gents(t,d) + \"\\n\" + text + \"\\n\")\n\n# Process a file.\ndef processfile(filename):\n data = \"\"\n xf = open(filename, 'r')\n fail=0\n\n# read in the file and skip lines that aren't titles.\n for line in xf:\n k = line.strip()\n\n # ditch bad lines\n if k.startswith(\"\"):\n fail = fail + 1\n elif k.startswith(\"